diff --git "a/6362.jsonl" "b/6362.jsonl" new file mode 100644--- /dev/null +++ "b/6362.jsonl" @@ -0,0 +1,680 @@ +{"seq_id":"294317534","text":"import bullet_builder as bb\n\ncore = bb.BulletCore(400, 100)\ngame = bb.BulletSimulator((800, 600), core)\n\nclass BulletA(bb.Bullet):\n\n\tCOMPONENTS = [\n\t\tbb.component.draw.GlowingCircle((0xFF, 0x1A, 0x1A), 7),\n\t\tbb.component.check_alive.AliveIfOnScreen((800, 600))\n\t]\n\ngenerator = bb.Bullet(0, 0, parent=core)\ngenerator.add_component(\n\tbb.component.spawn.LinearRadialSpawner(20, 100, bb.functions.SinusoidalWave(30, 2, 0, 1), BulletA))\n\ngame.start()","sub_path":"examples/radial_1.py","file_name":"radial_1.py","file_ext":"py","file_size_in_byte":445,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"340138777","text":"# =============================================================================\n# Authors: PAR Government\n# Organization: DARPA\n#\n# Copyright (c) 2016 PAR Government\n# All rights reserved.\n#==============================================================================\nimport numpy as np\nimport cv2\nimport logging\nimport os\nfrom scipy import spatial\nfrom maskgen.cv2api import cv2api_delegate\nfrom maskgen.image_wrap import ImageWrapper\nfrom maskgen.tool_set import VidTimeManager, differenceInFramesBetweenMillisecondsAndFrame\n\nimport hashlib\n\n\ndef readFrames(in_file, start_time, end_time):\n \"\"\"\n Function to read in video frames and store them in an array for later use.\n This limits the size of the video to just few minutes at most.\n Build a concatenated histogram of the GBR of each frame.\n :param in_file:\n :param offset_seconds: number of seconds to start search\n :return:\n @type offset_seconds: float\n \"\"\"\n if not os.path.exists(in_file):\n raise ValueError(in_file + ' not found')\n cap = cv2api_delegate.videoCapture(in_file)\n frames = list()\n histograms = list()\n fps = 0.0\n startFrame = 0\n time_manager = VidTimeManager(startTimeandFrame=start_time, stopTimeandFrame=end_time)\n try:\n while (cap.grab()):\n fps = cap.get(cv2api_delegate.prop_fps)\n elapsed_time = float(cap.get(cv2api_delegate.prop_pos_msec))\n time_manager.updateToNow(elapsed_time)\n if not time_manager.isBeforeTime() and not time_manager.isPastTime():\n ret, frame = cap.retrieve()\n frames.append(frame)\n hist = np.asarray(np.histogram(frame[:, :, 0], 256, (0, 255)))[0]\n hist = np.append(hist, np.asarray(np.histogram(frame[:, :, 1], 256, (0, 255)))[0])\n hist = np.append(hist, np.asarray(np.histogram(frame[:, :, 2], 256, (0, 255)))[0])\n histograms.append(hist)\n else:\n startFrame += 1\n if time_manager.isPastTime():\n break\n finally:\n cap.release()\n if len(frames) == 0:\n raise ValueError(in_file + ' unreadable')\n return [frames, histograms, fps, startFrame]\n\n\ndef createOutput(in_file, out_file, timeManager, codec=None):\n \"\"\"\n\n :param in_file:\n :param out_file:\n :param timeManager:\n :param codec:\n :return:\n @type in_file: str\n @type out_file: str\n @type timeManager: VidTimeManager\n \"\"\"\n logger = logging.getLogger('maskgen')\n cap = cv2api_delegate.videoCapture(in_file)\n fourcc = cv2api_delegate.get_fourcc(str(codec)) if codec is not None else cap.get(cv2api_delegate.fourcc_prop)\n fps = cap.get(cv2api_delegate.prop_fps)\n height = int(np.rint(cap.get(cv2api_delegate.prop_frame_height)))\n width = int(np.rint(cap.get(cv2api_delegate.prop_frame_width)))\n out_video = cv2api_delegate.videoWriter(out_file, fourcc, fps, (width, height), isColor=1)\n if not out_video.isOpened():\n err = out_file + \" fourcc: \" + str(fourcc) + \" FPS: \" + str(fps) + \\\n \" H: \" + str(height) + \" W: \" + str(width)\n raise ValueError('Unable to create video ' + err)\n try:\n writecount = 0\n dropcount = 0\n while (cap.grab()):\n ret, frame = cap.retrieve()\n elapsed_time = float(cap.get(cv2api_delegate.prop_pos_msec))\n timeManager.updateToNow(elapsed_time)\n if timeManager.isBeforeTime() or timeManager.isPastTime():\n out_video.write(frame)\n writecount += 1\n else:\n dropcount += 1\n logger.debug('Drop {} frames; Wrote {} frames'.format(dropcount, writecount))\n finally:\n cap.release()\n out_video.release()\n return dropcount\n\n\ndef scanHistList(histograms, distance, offset, saveHistFile=None):\n \"\"\"\n Function to compare frame image histograms and produce an array of the\n standard deviation of the differences.\n :param histograms: an array of concatenated GBR histograms\n :param distance: minimum number of frames apart to start the comparison\n :param offset: Number of frames to skip at the start of the list\n :return: Nx4 matrix where each column in start,end,length and std_flow.\n \"\"\"\n import math\n if distance >= len(histograms):\n raise ValueError('Video is to short for the distance provided.')\n history = np.zeros(((len(histograms) - (offset + distance)) * (len(histograms) - (offset + distance) + 1) // 2,\n 4), np.int)\n h_count = 0\n # front frame to compare. skip the first 30\n for i in range(offset, len(histograms) - distance):\n for j in range(i + distance, len(histograms)):\n std_flow = np.std(histograms[i] - histograms[j])\n history[h_count, :] = [i, j, j - i, std_flow]\n h_count += 1\n\n if saveHistFile is not None:\n np.savetxt(saveHistFile, history, delimiter=\",\", fmt='%2.3f')\n return history\n\n\ndef computeNormalDiffs(histograms, num_frames, logger=None):\n \"\"\"\n Return the average and standard deviation for the first number of frame\n histogram differences\n :param histograms:\n :param num_frames:\n :return:\n \"\"\"\n flow_list = np.zeros(num_frames)\n for i in range(1, num_frames):\n flow_list[i] = np.std(histograms[i] - histograms[i + 1])\n\n avg_flow = np.mean(flow_list)\n sigma_flow = np.std(flow_list)\n if logger is not None:\n logger.debug(\"mean flow {} with {} sigma\".format(avg_flow, sigma_flow))\n return [avg_flow, sigma_flow]\n\n\ndef selectBestMatches(differences, selection=50):\n \"\"\"\n return the 'selection' best results. Needs to be updated to try to find the longest\n rop that should work\n :param differences: histogram difference matrix\n :param selection: how manny to return\n :return: selectionX4 matrix of the best (least different)\n \"\"\"\n sort = differences[:, 3].argsort(axis=None)\n return differences[sort[:selection]]\n\n\n# best flow defined as the lowest sigam of the optical flow between frames\ndef selectBestFlow(frames, best_matches, logger):\n flow_list = np.zeros(best_matches.shape[0])\n for i in range(best_matches.shape[0]):\n past = cv2.cvtColor(frames[best_matches[i, 0]], cv2.COLOR_BGR2GRAY)\n future = cv2.cvtColor(frames[best_matches[i, 1]], cv2.COLOR_BGR2GRAY)\n flow = cv2api_delegate.calcOpticalFlowFarneback(past, future,\n 0.8, 7, 15, 3, 7, 1.5)\n #flow_list[i] = np.std(flow)\n flow_list[i] = np.mean(flow**2)\n\n\n if logger.isEnabledFor(logging.DEBUG):\n logger.debug('FOR {} to {}, STD={}'.format(best_matches[i, 0], best_matches[i, 1], flow_list[i]))\n return np.argmin(flow_list)\n\n\ndef getNormalFlow(frames):\n flow_list = np.zeros(len(frames) - 1)\n future = cv2.cvtColor(frames[0], cv2.COLOR_BGR2GRAY)\n for i in range(1, len(frames)):\n past = future\n future = cv2.cvtColor(frames[i], cv2.COLOR_BGR2GRAY)\n flow = cv2api_delegate.calcOpticalFlowFarneback(past, future,\n 0.8, 7, 15, 3, 7, 1.5)\n flow_list[i - 1] = np.mean(flow**2)\n return np.mean(flow_list)\n\n\ndef calculateOptimalFrameReplacement(frames, start, stop):\n avg_flow = np.mean([getNormalFlow(frames[max(0,start-30): start]),getNormalFlow(frames[stop: min(stop+30,len(frames))])])\n\n prev_frame = cv2.cvtColor(frames[start], cv2.COLOR_BGR2GRAY)\n next_frame = cv2.cvtColor(frames[stop], cv2.COLOR_BGR2GRAY)\n jump_flow = cv2api_delegate.calcOpticalFlowFarneback(prev_frame, next_frame,\n 0.8, 7, 15, 3, 7, 1.5, flags=2)\n std_jump_flow = np.mean(jump_flow**2)\n frames_to_add = int(np.rint(std_jump_flow / avg_flow))\n return frames_to_add\n\n\ndef dumpFrames(frames, file):\n with open(file, 'w') as fp:\n i = 1\n for frame in frames:\n fp.write('{},{}\\n'.format(i, hashlib.sha256(frame).hexdigest()))\n i += 1\n\n\ndef smartDropFrames(in_file, out_file,\n start_time,\n end_time,\n seconds_to_drop,\n savehistograms=False,\n codec=None,\n drop=True):\n \"\"\"\n :param in_file: is the full path of the video file from which to drop frames\n :param out_file: resulting video file\n :param start_time: (milli,frame no) for search space\n :param end_time: (milli,frame no) for search space\n :param seconds_to_drop:\n :param savehistograms: save histograms differences to file\n :param codec:\n :return: first and last frame numbers dropped, and the optimal number to add back/replace\n \"\"\"\n logger = logging.getLogger('maskgen')\n logger.info('Read {} frames into memory'.format(in_file))\n frames, histograms, fps, start = readFrames(in_file, start_time, end_time)\n #dumpFrames(frames, in_file[0:in_file.rfind('.')] + '-frames.csv')\n distance = int(round(fps * seconds_to_drop))\n logger.info('Distance {} for {} frames to drop with {} fps'.format(distance, seconds_to_drop, fps))\n offset = int(round(fps * seconds_to_drop))\n # avg_diffs, sigma_diffs = computeNormalDiffs(histograms, 60)\n logger.info('starting histogram computational')\n differences = scanHistList(histograms, distance, offset,\n saveHistFile=in_file[0:in_file.rfind('.')] + '-hist.csv' if savehistograms else None)\n logger.info('Finding best matches')\n best_matches = selectBestMatches(differences, selection=50)\n logger.info('Starting optical flow search')\n if best_matches is not None:\n best_flow = selectBestFlow(frames, best_matches, logger)\n logger.info('best pair: {}'.format(str(best_matches[best_flow])))\n frames_to_add = calculateOptimalFrameReplacement(frames, best_matches[best_flow][0],\n best_matches[best_flow][1])\n # add 2: one to advance to frame no and one to advance to first dropped frame\n firstFrametoDrop = best_matches[best_flow][0] + start + 2\n lastFrametoDrop = best_matches[best_flow][1] + start\n if drop:\n time_manager = VidTimeManager(startTimeandFrame=(0, firstFrametoDrop),\n stopTimeandFrame=(0, lastFrametoDrop))\n createOutput(in_file, out_file, time_manager, codec=codec)\n return firstFrametoDrop, lastFrametoDrop, frames_to_add\n\n\ndef dropFrames(in_file, out_file,\n start_time,\n end_time,\n codec=None):\n \"\"\"\n :param in_file: is the full path of the video file from which to drop frames\n :param out_file: resulting video file\n :param start_time: (milli,frame no) for search space\n :param end_time: (milli,frame no) for search space\n :param codec:\n :return:\n \"\"\"\n time_manager = VidTimeManager(startTimeandFrame=start_time, stopTimeandFrame=end_time)\n dropped = createOutput(in_file, out_file, time_manager, codec=codec)\n return time_manager.getStartFrame(), time_manager.getEndFrame(), dropped\n\n\nclass OpticalFlow:\n def __init__(self, prvs_frame, next_frame, flow, bkflow):\n self.prvs_frame = prvs_frame\n self.next_frame = next_frame\n self.flow = flow\n self.bkflow = bkflow\n self.hight = flow.shape[0]\n self.width = flow.shape[1]\n h, w = flow.shape[:2]\n self.coords = (np.swapaxes(np.indices((w, h), np.float32), 0, 2))\n\n def setFrames(self, prvs_frame, next_frame, flow, bkflow):\n self.prvs_frame = prvs_frame\n self.next_frame = next_frame\n self.flow = flow\n self.bkflow = bkflow\n\n def warpFlow(self, img, flow):\n adj = self.coords + flow\n underoverflow_width = np.logical_or(adj[:, :, 0] >= self.coords.shape[1],\n adj[:, :, 0] < 0)\n underoverflow_height = np.logical_or(adj[:, :, 1] >= self.coords.shape[0],\n adj[:, :, 1] < 0)\n adj[underoverflow_width] = self.coords[underoverflow_width]\n adj[underoverflow_height] = self.coords[underoverflow_height]\n return cv2.remap(img, adj, None, cv2.INTER_LINEAR)\n\n def setTime(self, frame_time):\n forward_flow = np.multiply(self.flow, 1 - frame_time)\n backward_flow = np.multiply(self.bkflow, frame_time)\n from_prev = self.warpFlow(self.prvs_frame, backward_flow)\n from_next = self.warpFlow(self.next_frame, forward_flow)\n from_prev = np.multiply(from_prev, (1 - frame_time))\n from_next = np.multiply(from_next, frame_time)\n frame = (np.add(from_prev, from_next)).astype(np.uint8)\n\n return frame\n\n # return the average sigma(optical flow)\n\nclass kdtreeOpticalFlow:\n def __init__(self, prvs_frame, next_frame, flow, bkflow):\n self.logger = logging.getLogger('maskgen')\n self.prvs_frame = prvs_frame\n self.next_frame = next_frame\n self.flow = flow\n self.bkflow = bkflow\n self.hight = flow.shape[0]\n self.width = flow.shape[1]\n self.count = 0\n h, w = flow.shape[:2]\n self.coords = (np.swapaxes(np.indices((w, h), np.float32), 0, 2))\n\n def setFrames(self, prvs_frame, next_frame, flow, bkflow):\n self.prvs_frame = prvs_frame\n self.next_frame = next_frame\n self.flow = flow\n self.bkflow = bkflow\n\n def warpFlow(self, img, flow):\n #s = time.time()\n res = self.adjustFlow_G(flow)\n #d = time.time() - s\n #self.logger.info('Inverse Took {} seconds'.format(d))\n adj = res[0] + self.coords\n mp = res[1]\n return cv2.remap(img, adj, None, cv2.INTER_LINEAR), mp\n\n def adjustFlow_G(self, flow, p=3.0, k=5):\n p = p\n k = k\n h, w = flow.shape[:2]\n coord = (np.swapaxes(np.indices((w, h), np.float32), 0, 2))\n cpy = np.copy(flow)\n cpy += coord\n ktree = spatial.cKDTree(np.reshape(cpy, (w * h, 2)))\n reverse = np.swapaxes(np.indices((w, h), np.float32), 0, 2)\n reverse[:][:] = -1000.0\n nearest = ktree.query(coord, k=k) # ,distance_upper_bound=2.0**0.5)\n mp = np.any((nearest[0] < 1.0), axis=2)\n # identify points that have source points close enough to use\n close_enough = np.any((nearest[0] < 1.0), axis=2)\n # id points that have at least one match 0 away\n exact = np.any((nearest[0] == 0.0), axis=2)\n values = np.asarray([nearest[1] % w, nearest[1] / w])\n\n for x in range(h):\n for y in range(w):\n found = False\n dist = nearest[0][x, y]\n # skip points too far away\n if not close_enough[x, y]:\n continue\n\n # process exact matches\n if exact[x, y]:\n # find max distance of the k closest source points\n md_k = np.argmax(((values[1, x, y] - x) ** 2 + (values[0, x, y] - y) ** 2) ** .5)\n\n # if mapped distance ==0 and source distance is the greatest, use it\n if dist[md_k] == 0:\n reverse[x, y, :] = values[:, x, y, md_k]\n # reverse[x][y][0] = values[0,x,y,md_k]\n found = True\n\n # process interpolation points\n if not found:\n weight_accum = np.sum(1 / dist[np.where(dist[:] > 0)] ** p)\n w_xyk = np.sum((values[:, x, y, np.where(dist[:] > 0)]) /\n (dist[np.where(dist[:] > 0)] ** p), axis=2)\n reverse[x][y] = (w_xyk / weight_accum)[:, 0]\n\n return reverse - coord, mp\n\n def setTime(self, frame_time, truth=None):\n forward_flow = np.multiply(self.flow, 1 - frame_time)\n # cv2.imwrite('fwd_frame.png',imageafy(forward_flow,self.next_frame))\n backward_flow = np.multiply(self.bkflow, frame_time)\n # cv2.imwrite('bwd_frame.png', imageafy(backward_flow, self.next_frame))\n from_prev, mpp = self.warpFlow(self.prvs_frame, backward_flow)\n from_next, mpn = self.warpFlow(self.next_frame, forward_flow)\n # cv2.imwrite('bwd_frame_real' + 'new' + str(self.count) + '.png', from_prev)\n # cv2.imwrite('fwd_frame_real' + 'new' + str(self.count) + '.png', from_next)\n truth = self.next_frame if truth is None else truth\n f = self.frameadjust(from_next, self.prvs_frame, mpp)\n n = self.frameadjust(from_prev, truth, mpn)\n from_next = f\n from_prev = n\n from_prev = np.multiply(from_prev, (1 - frame_time))\n from_next = np.multiply(from_next, frame_time)\n\n frame = (np.add(from_prev, from_next)).astype(np.uint8)\n self.count += 1\n return frame\n\n def frameadjust(self, frame, alterframe, mp):\n cpy = np.copy(frame)\n for x in range(len(frame)):\n for y in range(len(frame[0])):\n if np.array_equal(frame[x][y], np.zeros(len(frame[x][y]))):\n cpy[x][y] = alterframe[x][y]\n return cpy\n\n\nclass FrameAnalyzer:\n def __init__(self, start_time, end_time, fps):\n self.last_frames = []\n self.start_time = start_time\n self.end_time = end_time\n self.fps = fps\n\n def addFrame(self, frame):\n if self.end_time is not None:\n return\n self.last_frames.append(frame)\n if len(self.last_frames) > 50:\n self.last_frames = self.last_frames[1:]\n\n def framesToAdd(self):\n if self.end_time is not None:\n return differenceInFramesBetweenMillisecondsAndFrame(self.end_time, self.start_time, self.fps) + 1\n avg_flow = getNormalFlow(self.last_frames)\n std_jump_flow = np.mean(self.jump_flow**2)\n return int(np.rint(std_jump_flow / avg_flow))\n\n def updateFlow(self, last_frame, next_frame, direction):\n if direction != 'forward':\n prev_frame_gray = cv2.cvtColor(next_frame, cv2.COLOR_BGR2GRAY)\n next_frame_gray = cv2.cvtColor(last_frame, cv2.COLOR_BGR2GRAY)\n else:\n prev_frame_gray = cv2.cvtColor(last_frame, cv2.COLOR_BGR2GRAY)\n next_frame_gray = cv2.cvtColor(next_frame, cv2.COLOR_BGR2GRAY)\n self.jump_flow = cv2api_delegate.calcOpticalFlowFarneback(prev_frame_gray,\n next_frame_gray,\n 0.8, 7, 15, 3, 7, 1.5, 2)\n self.back_flow =self.jump_flow\n self.jump_flow = cv2api_delegate.calcOpticalFlowFarneback(next_frame_gray,\n prev_frame_gray,\n 0.8, 7, 15, 3, 7, 1.5, 2)\n\n\ndef smartAddFrames(in_file,\n out_file,\n start_time,\n end_time,\n codec=None,\n direction='forward'):\n \"\"\"\n :param in_file: is the full path of the video file from which to drop frames\n :param out_file: resulting video file\n :param start_time: (milli,frame no) start to fill\n :param end_time: (milli,frame no) end to fill\n :param codec:\n :return:\n \"\"\"\n logger = logging.getLogger('maskgen')\n import time\n cap = cv2api_delegate.videoCapture(in_file)\n fourcc = cv2api_delegate.get_fourcc(str(codec)) if codec is not None else cap.get(cv2api_delegate.fourcc_prop)\n fps = cap.get(cv2api_delegate.prop_fps)\n height = int(np.rint(cap.get(cv2api_delegate.prop_frame_height)))\n width = int(np.rint(cap.get(cv2api_delegate.prop_frame_width)))\n out_video = cv2api_delegate.videoWriter(out_file, fourcc, fps, (width, height), isColor=1)\n time_manager = VidTimeManager(startTimeandFrame=start_time, stopTimeandFrame=end_time)\n if not out_video.isOpened():\n err = out_file + \" fourcc: \" + str(fourcc) + \" FPS: \" + str(fps) + \\\n \" H: \" + str(height) + \" W: \" + str(width)\n raise ValueError('Unable to create video ' + err)\n try:\n last_frame = None\n frame_analyzer = FrameAnalyzer(start_time, end_time, fps)\n written_count = 0\n while (cap.grab()):\n ret, frame = cap.retrieve()\n frame_analyzer.addFrame(frame)\n elapsed_time = float(cap.get(cv2api_delegate.prop_pos_msec))\n time_manager.updateToNow(elapsed_time)\n if not time_manager.isBeforeTime():\n break\n out_video.write(frame)\n written_count += 1\n last_frame = frame\n next_frame = frame\n if logger.isEnabledFor(logging.DEBUG):\n logger.debug('Written {} frames '.format(written_count))\n logger.debug('Smart Add Frames ' + str(start_time) + ' to ' + str(end_time))\n logger.debug(\"Selected Before {}\".format(hashlib.sha256(last_frame).hexdigest()))\n logger.debug(\"Selected After {}\".format(hashlib.sha256(next_frame).hexdigest()))\n ImageWrapper(last_frame).save('before_' + str(time.clock()) + '.png')\n ImageWrapper(next_frame).save('after_' + str(time.clock()) + '.png')\n logger.debug(\"STD after and before {}\".format(np.std(last_frame - next_frame)))\n frame_analyzer.updateFlow(last_frame, next_frame, direction)\n opticalFlow = kdtreeOpticalFlow(last_frame, next_frame, frame_analyzer.jump_flow, frame_analyzer.back_flow)\n frames_to_add = frame_analyzer.framesToAdd()\n lf = last_frame\n written_count = 0\n for i in range(1, int(frames_to_add + 1)):\n frame_scale = i / (1.0 + frames_to_add)\n frame = opticalFlow.setTime(frame_scale)\n if logger.isEnabledFor(logging.DEBUG):\n logger.debug(\"frame {}\".format(np.std(frame - lf)))\n out_video.write(frame)\n lf = frame\n out_video.write(next_frame)\n written_count+=1\n while (cap.grab()):\n ret, frame = cap.retrieve()\n out_video.write(frame)\n written_count += 1\n logger.debug('Written additioning {} frames '.format(written_count))\n finally:\n cap.release()\n out_video.release()\n return frames_to_add, frames_to_add * (1000.0 / fps)\n\n\ndef copyFrames(in_file,\n out_file,\n start_time,\n end_time,\n paste_time,\n codec=None):\n \"\"\"\n :param in_file: is the full path of the video file from which to drop frames\n :param out_file: resulting video file\n :param start_time: (milli,frame no) to start to fill\n :param end_time: (milli,frame no) end fil\n :param codec:\n :return:\n \"\"\"\n import time\n logger = logging.getLogger('maskgen')\n frames_to_write = []\n frames_to_copy = []\n cap = cv2api_delegate.videoCapture(in_file)\n fourcc = cv2api_delegate.get_fourcc(str(codec)) if codec is not None else cap.get(cv2api_delegate.fourcc_prop)\n fps = cap.get(cv2api_delegate.prop_fps)\n height = int(np.rint(cap.get(cv2api_delegate.prop_frame_height)))\n width = int(np.rint(cap.get(cv2api_delegate.prop_frame_width)))\n out_video = cv2api_delegate.videoWriter(out_file, fourcc, fps, (width, height), isColor=1)\n if not out_video.isOpened():\n err = out_file + \" fourcc: \" + str(fourcc) + \" FPS: \" + str(fps) + \\\n \" H: \" + str(height) + \" W: \" + str(width)\n raise ValueError('Unable to create video ' + err)\n copy_time_manager = VidTimeManager(startTimeandFrame=start_time, stopTimeandFrame=end_time)\n paste_time_manager = VidTimeManager(startTimeandFrame=paste_time)\n write_count = 0\n try:\n while (not copy_time_manager.isPastTime() and cap.grab()):\n ret, frame = cap.retrieve()\n elapsed_time = float(cap.get(cv2api_delegate.prop_pos_msec))\n copy_time_manager.updateToNow(elapsed_time)\n paste_time_manager.updateToNow(elapsed_time)\n if not copy_time_manager.isBeforeTime() and not copy_time_manager.isPastTime():\n frames_to_copy.append(frame)\n if not paste_time_manager.isBeforeTime():\n frames_to_write.append(frame)\n else:\n out_video.write(frame)\n write_count += 1\n if logger.isEnabledFor(logging.DEBUG):\n logger.debug(\"First to copy {}\".format(hashlib.sha256(frames_to_copy[0]).hexdigest()))\n logger.debug(\"Last to copy {}\".format(hashlib.sha256(frames_to_copy[-1]).hexdigest()))\n ImageWrapper(frames_to_copy[0]).save('first_' + str(time.clock()) + '.png')\n ImageWrapper(frames_to_copy[-1]).save('last_' + str(time.clock()) + '.png')\n if len(frames_to_write) > 0:\n # paste prior to copy\n for copy_frame in frames_to_copy:\n out_video.write(copy_frame)\n for write_frame in frames_to_write:\n out_video.write(write_frame)\n else:\n # paste after to copy\n frame = None\n while (paste_time_manager.isBeforeTime() and cap.grab()):\n ret, frame = cap.retrieve()\n elapsed_time = float(cap.get(cv2api_delegate.prop_pos_msec))\n paste_time_manager.updateToNow(elapsed_time)\n if paste_time_manager.isBeforeTime():\n out_video.write(frame)\n write_count += 1\n for copy_frame in frames_to_copy:\n out_video.write(copy_frame)\n if frame is not None:\n out_video.write(frame)\n while (cap.grab()):\n ret, frame = cap.retrieve()\n out_video.write(frame)\n finally:\n cap.release()\n out_video.release()\n return write_count\n","sub_path":"maskgen/algorithms/optical_flow.py","file_name":"optical_flow.py","file_ext":"py","file_size_in_byte":25993,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"512262834","text":"import pyautogui\nimport webbrowser\nimport time\nimport random\n\nprint(\"****Tinder swiper V2****\")\n\npyautogui.FAILSAFE = False\nprint('Make sure you are already logged in Tinder On your Browser')\na = int(input('How many people do you want to swipe?\\n'))\n\nwebbrowser.open('https://www.tinder.com',new = 1,autoraise = True)\ntime.sleep(10)\n\nfor i in range(0,a) :\n\tif random.randint(0,1) > 0.5:\n\t\tpyautogui.press('right')\n\telse:\n\t\tpyautogui.press('left')\n\ttime.sleep(1)\n\tprint('person',i+1,'swiped')\n","sub_path":"python/TinderBot/Bot.py","file_name":"Bot.py","file_ext":"py","file_size_in_byte":492,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"331372940","text":"__author__ = 'Carlos_Vaquero'\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport csv\nfrom itertools import *\n\ndef window(seq, n=2):\n \"Returns a sliding window (of width n) over data from the iterable\"\n \" s -> (s0,s1,...s[n-1]), (s1,s2,...,sn), ... \"\n it = iter(seq)\n result = tuple(islice(it, n))\n if len(result) == n:\n yield result\n for elem in it:\n result = result[1:] + (elem,)\n yield result","sub_path":"Polynomial_regression/windowing_timing.py","file_name":"windowing_timing.py","file_ext":"py","file_size_in_byte":461,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"418527838","text":"\"\"\"\nMichael S. Emanuel\nMon Jun 11 14:31:47 2018\n\nHexadecimal numbers\nProblem 162\n\nIn the hexadecimal number system numbers are represented using 16 different digits:\n\n0,1,2,3,4,5,6,7,8,9,A,B,C,D,E,F\nThe hexadecimal number AF when written in the decimal number system equals 10x16+15=175.\n\nIn the 3-digit hexadecimal numbers 10A, 1A0, A10, and A01 the digits 0,1 and A are all present.\nLike numbers written in base ten we write hexadecimal numbers without leading zeroes.\n\nHow many hexadecimal numbers containing at most sixteen hexadecimal digits exist with all of the\ndigits 0,1, and A present at least once?\nGive your answer as a hexadecimal number.\n\n(A,B,C,D,E and F in upper case, without any leading or trailing code that marks the number\nas hexadecimal and without leading zeroes , e.g. 1A3F and not: 1a3f and not 0x1a3f\nand not $1A3F and not #1A3F and not 0000001A3F)\n\"\"\"\n\nfrom Euler.Utility import range_inc, prod\nfrom Euler.Digits import getDigits\nfrom typing import List\n\n\ndef meetsCriteria(n: int):\n \"\"\"Does integer n meet the stated criteria?\"\"\"\n digitSet = set(getDigits(n, 16))\n return 0 in digitSet and 1 in digitSet and 10 in digitSet\n\n\ndef listAll(k: int):\n \"\"\"Enumerate hex numbers of this length by brute force.\"\"\"\n nMin: int = 16**(k-1)\n nMax: int = 16**k-1\n nums: List[int] = []\n for n in range(nMin, nMax):\n if meetsCriteria(n):\n nums.append(n)\n return nums\n\n\ndef count_bf(k: int):\n \"\"\"Count hex numbers of this length by brute force.\"\"\"\n nMin: int = 16**(k-1)\n nMax: int = 16**k-1\n total: int = 0\n for n in range(nMin, nMax):\n if meetsCriteria(n):\n total += 1\n return total\n\n\ndef hex2(n: int) -> str:\n \"\"\"Hex numbers formatted with guidelines for this problem\"\"\"\n return hex(n).upper()[2:]\n\n\ndef count(k: int):\n \"\"\"How many k digit hex numbers meet these criteria?\"\"\"\n # How many k digit numbers starting with 1 or A meet these criteria?\n # Let k1 be the position of the first digit that is a 0 or an A\n # Let k2 be the position of the second digit that is a 0 or an A\n term1: int = 0\n # How many choices are there for each digit position?\n slots = k * [0]\n # There are two choices for the first digit, 1 and A\n slots[0] = 2\n for k1 in range(1, k-1):\n # Digits between 1 and k1 have 14 choices: three special digits, one already used\n slots[1:k1] = (k1-1)*[14]\n # The digit in slot k1 has 2 choices: 0, and the opposite of 1 or A in slot 1\n slots[k1] = 2\n for k2 in range(k1+1, k):\n # Digits between k1 and k2 have 15 choices: three special digits, two already used\n slots[k1+1:k2] = (k2-k1-1)*[15]\n # The digit in slot k2 is fully specified by slot 1 and k1\n slots[k2] = 1\n # The remaining slots have 16 possible values (any digit is OK)\n slots[k2+1:] = (k-k2-1)*[16]\n # This pattern's contribution to numbers with a leading digit of 1 or A\n term1 += prod(slots)\n # print(f'k1={k1}, k2={k2}, slots={slots}.')\n\n # How many k digit numbers starting with a digit other than 1 or A meet these criteria?\n # Let k1 be the position of the first digit that is 0, 1, or A\n # Let k2 be the position of the second digit that is 0, 1, or A\n # Let k3 be the position of the third digit that is 0, 1, or A\n term2: int = 0\n # How many choices are there for each digit position?\n slots = k * [0]\n for k1 in range(1, k-2):\n # Digits before k1 have 13 choices: anything but 0, 1, A\n slots[0:k1] = k1*[13]\n # The digit in slot k1 has 3 choices: 0, 1, A\n slots[k1] = 3\n for k2 in range(k1+1, k-1):\n # Digits between k1 and k2 have 14 choices\n slots[k1+1:k2] = (k2-k1-1)*[14]\n # The digit in slot k2 has two choices\n slots[k2] = 2\n for k3 in range(k2+1, k):\n # Digits between k2 and k3 have 15 choices\n slots[k2+1:k3] = (k3-k2-1)*[15]\n # The digit in slot k3 has 1 choice\n slots[k3] = 1\n # The remaining slots have 16 possible values (any digit is OK)\n slots[k3+1:] = (k-k3-1)*[16]\n # This pattern's contribution to numbers with a leading digit of 1 or A\n term2 += prod(slots)\n return term1 + term2\n\n\ndef main() -> str:\n # Compute the answer\n ansInt: int = sum(count(n) for n in range_inc(16))\n # Convert the answer to a hex string\n ans: str = hex2(ansInt)\n # Print the answer\n print(f'There are {ansInt} hex numbers with length <= 16 and at least one 0, 1 or A.')\n print(f'The answer is {ans}.')\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"Prob162_HexadecimalNumbers.py","file_name":"Prob162_HexadecimalNumbers.py","file_ext":"py","file_size_in_byte":4749,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"298748122","text":"##Script (Python) \"add_submit_helper\"\n##bind container=container\n##bind context=context\n##bind namespace=\n##bind script=script\n##bind subpath=traverse_subpath\n##parameters=model, id, title, result\n##title=\n##\nmodel.manage_addProduct['SilvaUCLLifeLearningContentTypes'].manage_addAnouncement(id, title)\nAnouncement = getattr(model, id)\nversion = Anouncement.get_editable()\n#version.set_pagetitle(result['pagetitle'])\n#version.set_shortdescription(result['ShortDescription'])\n#version.set_adminname(result['AdminName'])\n# more to follow\nreturn Anouncement\n","sub_path":"views/add/Anouncement/add_submit_helper.py","file_name":"add_submit_helper.py","file_ext":"py","file_size_in_byte":554,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"139483732","text":"import torch\nimport os\nimport csv\nimport torch.nn as nn\nimport torch.optim as optim\nfrom dataloader import get_loader, get_loader2\nimport torchvision.transforms as transforms\nfrom models import classifier\n\ndevice = torch.device('cuda' if torch.cuda.is_available() else 'cpu')\n\nextractor = classifier.Extractor().to(device)\nemotion_classifier = classifier.Classifier(7).to(device)\ngender_classifier = classifier.Classifier(2).to(device)\n\nclass_criterion = nn.CrossEntropyLoss()\n\nparams = list(list(extractor.parameters()) + list(emotion_classifier.parameters()) + list(gender_classifier.parameters()))\noptimizer = torch.optim.Adam(params, lr=1e-3)\n\ntotal_epochs = 100\n\nbaseDataPath = './data'\n#data_path = f\"{baseDataPath}/file_names.csv\"\ntransform = transforms.Compose([transforms.Resize((64,64)), \n transforms.ToTensor()])\n\ntrain_data = get_loader2(f\"{baseDataPath}/train.csv\", transform, 128, 4,shuffle=True)\nval_data = get_loader2(f\"{baseDataPath}/val.csv\", transform, 128, 4,shuffle=True)\ntest_data = get_loader2(f\"{baseDataPath}/test.csv\", transform, 128, 4,shuffle=True)\n\n\nmin_val_loss = 10000.0\ntotal_examples = 0\ngender_correct = 0\nemotion_correct = 0\nfor epoch in range(total_epochs):\n\tprint('starting training')\n\textractor.train()\n\tgender_classifier.train()\n\temotion_classifier.train()\n\n\tfor batch_idx, (image_data, gender_label, emotion_label) in enumerate(train_data):\n\n\t\t#optimizer.zero_grad()\n\t\textractor.zero_grad()\n\t\tgender_classifier.zero_grad()\n\t\temotion_classifier.zero_grad()\n\t\ttotal_examples += len(image_data)\n\n\t\timage_data,gender_label,emotion_label = image_data[:,:3, :, :].to(device), gender_label.to(device), emotion_label.to(device)\n\n\t\tfeature = extractor.forward(image_data).squeeze()\n\n\t\tgender_output = gender_classifier.forward(feature)\n\t\temotion_output = emotion_classifier.forward(feature)\n\n\t\tgender_loss = class_criterion(gender_output, gender_label)\n\t\temotion_loss = class_criterion(emotion_output, emotion_label)\n\t\ttotal_loss = gender_loss + emotion_loss\n\n\t\ttotal_loss.backward()\n\t\toptimizer.step()\n\n\t\tmax_index = gender_output.max(dim = 1)\n\t\tmax_index = max_index[1]\n\t\tgender_correct += (max_index == gender_label.to(device)).sum().item()\n\n\t\tmax_index = emotion_output.max(dim = 1)\n\t\tmax_index = max_index[1]\n\t\temotion_correct += (max_index == emotion_label.to(device)).sum().item()\n\n\t\tif(batch_idx % 20 == 0):\n\t\t\tprint((epoch, batch_idx))\n\t\t\tprint((gender_loss.item(), emotion_loss.item(), total_loss.item(), gender_correct/total_examples, emotion_correct/total_examples))\n\n\tprint('starting validation')\n\n\textractor.eval()\n\tgender_classifier.eval()\n\temotion_classifier.eval()\n\n\ttotal_gender_loss = 0.0\n\ttotal_emotion_loss = 0.0\n\ttotal_examples = 0\n\tgender_correct = 0\n\temotion_correct = 0\n\n\tfor batch_idx, (image_data, gender_label, emotion_label) in enumerate(val_data):\n\n\t\t#optimizer.zero_grad()\n\t\textractor.zero_grad()\n\t\tgender_classifier.zero_grad()\n\t\temotion_classifier.zero_grad()\n\t\ttotal_examples += len(image_data)\n\n\t\timage_data,gender_label,emotion_label = image_data[:,:3, :, :].to(device), gender_label.to(device), emotion_label.to(device)\n\n\t\tfeature = extractor.forward(image_data).squeeze()\n\n\t\tgender_output = gender_classifier.forward(feature)\n\t\temotion_output = emotion_classifier.forward(feature)\n\n\t\tgender_loss = class_criterion(gender_output, gender_label)\n\t\temotion_loss = class_criterion(emotion_output, emotion_label)\n\n\t\tmax_index = gender_output.max(dim = 1)\n\t\tmax_index = max_index[1]\n\t\tgender_correct += (max_index == gender_label.to(device)).sum().item()\n\n\t\tmax_index = emotion_output.max(dim = 1)\n\t\tmax_index = max_index[1]\n\t\temotion_correct += (max_index == emotion_label.to(device)).sum().item()\n\n\t\ttotal_gender_loss += gender_loss.item()\n\t\ttotal_emotion_loss += emotion_loss.item()\n\n\ttotal_loss = total_emotion_loss + total_gender_loss\n\n\tif(total_loss < min_val_loss):\n\t\tmin_val_loss = total_loss\n\t\tprint((epoch, min_val_loss, gender_correct/total_examples, emotion_correct/total_examples))\n\t\ttorch.save({\n\t\t\t'extractor' : extractor.state_dict(), \n\t\t\t'gender_classifier' : gender_classifier.state_dict(), \n\t\t\t'emotion_classifier' : emotion_classifier.state_dict(), \n\t\t\t'optimizer' : optimizer.state_dict(),\n\t\t\t'stat' : (epoch, min_val_loss, gender_correct/total_examples, emotion_correct/total_examples)\n\t\t\t}, './results/classifier')\n\t\tprint('saved')\n\t\tprint('')\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\t\n\n\n\n","sub_path":"train_classifier.py","file_name":"train_classifier.py","file_ext":"py","file_size_in_byte":4379,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"356502801","text":"#coding=utf-8\n\nimport tornado.web\nimport time\nfrom markdown import Markdown\nimport re\nfrom tornado.escape import xhtml_escape\nfrom emoji import emojis\nfrom config import google_analytics,admin\n\nhtml_killer = re.compile('<[^>]*>')\n\n#https://github.com/lepture/june/blob/master/june/lib/filters.py\nurl_replace = re.compile(r'(?m)^((?:https?://|www\\d{0,3}[.]|[a-z0-9.\\-]+[.][a-z]{2,4}'\n r'/)(?:[^\\s()<>]+|\\(([^\\s()<>]+|(\\([^\\s()<>]+\\)))*\\))+(?:\\(([^\\s()<>]+'\n r'|(\\([^\\s()<>]+\\)))*\\)|[^\\s`!()\\[\\]{};:\\'\".,<>?«»“”‘’]))')\npattern = re.compile(\n r'(?i)(?:<)((?:https?://|www\\d{0,3}[.]|[a-z0-9.\\-]+[.][a-z]{2,4}'\n r'/)(?:[^\\s()<>]+|\\(([^\\s()<>]+|(\\([^\\s()<>]+\\)))*\\))+(?:\\(([^\\s()<>]+'\n r'|(\\([^\\s()<>]+\\)))*\\)|[^\\s`!()\\[\\]{};:\\'\".,<>?«»“”‘’]))(?:>)')\n\nyoutube = re.compile('http://youtu.be/(.+)')\nyouku = re.compile('http://v.youku.com/v_show/id_([a-zA-Z0-9\\=]+).html')\nyinyuetai = re.compile('http://www.yinyuetai.com/video/(\\d+)')\nusername_finder = re.compile(u'@(\\w{1,25})\\s')\nemoji_finder = re.compile(u'(:[^:]+:)')\n\nmd = Markdown(extensions=['fenced_code'])\n\nclass BaseHandler(tornado.web.RequestHandler):\n\n def __init__(self,*args,**kwargs):\n tornado.web.RequestHandler.__init__(self,*args,**kwargs)\n self.db = self.application.db\n self.mc = self.application.mc\n\n def get_current_user(self):\n password = self.get_secure_cookie('user')\n return password and self.db.users.find_one({'password':password}) or None\n\n def get_error_html(self,status_code, **kwargs):\n return self.render_string('404.html',google_analytics=google_analytics,admin_list=admin,unread=0)\n\n def parse_user_agent(self):\n ua = self.request.headers.get(\"User-Agent\", \"bot\").lower()\n sources = ('iPod','iPhone','iPad','Android','Kindle','Windows Phone','Symbian')\n source = None\n for type in sources:\n if type.lower() in ua:\n source = type\n return source\n\n def render(self, template_name, **kwargs):\n user = self.current_user\n unread = 0\n if user:\n for x in user['notification']:\n if not x['read']:\n unread += 1\n tornado.web.RequestHandler.render(self,template_name=template_name,admin_list=admin,db=self.db,\n unread=unread,mc=self.mc,google_analytics=google_analytics,**kwargs)\n\nclass HomeHandler(BaseHandler):\n def get(self):\n posts = self.db.posts.find({},sort=[('changedtime', -1)],limit=15)\n self.render('index.html',time_span=time_span,posts=posts)\n\nclass EditModule(tornado.web.UIModule):\n def render(self,db):\n return self.render_string('modules/markdown.html',db=db)\n\nclass FeedHandler(BaseHandler):\n def get(self):\n self.set_header(\"Content-Type\", \"application/atom+xml\")\n url = ''\n posts = [_ for _ in self.db.posts.find({},sort=[('changedtime', -1)],limit=20)]\n tornado.web.RequestHandler.render(self,'atom.xml',url=url,name='全站',\n time=time,posts=posts)\n\nclass ErrorHandler(BaseHandler):\n def get(self, *args, **kwargs):\n raise tornado.web.HTTPError(404)\n\ndef time_span(t):\n '''convert timestamp to readable string.'''\n t=time.gmtime(t)\n return '' % time.strftime('%Y-%m-%dT%H:%M:%SZ',t)\n\ndef md_convert(txt,notice=False,time=None,user=None,db=None,postid=None):\n '''convert md to html'''\n #escape html\n for x in set(html_killer.findall(txt)):\n txt = txt.replace(x,xhtml_escape(x))\n\n #https://github.com/livid/v2ex/blob/master/v2ex/templatetags/filters.py\n #support video\n for video_id in set(youtube.findall(txt)):\n txt = txt.replace('http://youtu.be/' + video_id,'')\n for video_id in set(youku.findall(txt)):\n txt = txt.replace('http://v.youku.com/v_show/id_' + video_id + '.html', '')\n for video_id in set(yinyuetai.findall(txt)):\n txt = txt.replace('http://www.yinyuetai.com/video/' + video_id,'')\n\n txt = url_replace.sub(make_link,txt)\n txt = pattern.sub(make_link, txt)\n\n mentions = []\n for u in set(username_finder.findall(txt + ' ')):\n mentions.append(u)\n txt = txt.replace(u'@'+u,u'@%s' % (u,u))\n\n for emoji in set(emoji_finder.findall(txt)):\n if emoji in emojis:\n txt = txt.replace(emoji,u'' % emojis[emoji])\n\n txt = md.convert(txt).replace('\\n','
')\n txt = txt.replace('&lt;','<').replace('&gt;','>')\n\n if notice:\n for u in mentions:\n db.users.update({'username_lower':u.lower()},\n {'$push':\n {'notification':\n {'from':user,\n 'content':txt,\n 'time':time,\n 'postid':postid,\n 'read':False,\n }\n },\n })\n\n return txt\n\ndef make_link(m):\n link = m.group(1)\n if link.startswith('https://gist.github.com/') or link.startswith('http://gist.github.com/'):\n return '' % link.replace('https','http')\n if '.jpg' in link or '.jpeg' in link or '.gif' in link or '.png' in link:\n return '' % link\n else:\n return '%s' % (link, link)","sub_path":"common.py","file_name":"common.py","file_ext":"py","file_size_in_byte":6315,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"288184769","text":"import traceback\r\nfrom collections import namedtuple\r\nfrom datetime import datetime\r\nfrom dateutil.parser import parse\r\nfrom flask import url_for, current_app, render_template\r\nfrom lbrc_flask.database import db\r\nfrom identity.celery import celery\r\nfrom identity.demographics.model import (\r\n DemographicsRequest,\r\n DemographicsRequestData,\r\n DemographicsRequestPmiData,\r\n DemographicsRequestDataMessage,\r\n DemographicsRequestDataResponse,\r\n)\r\nfrom identity.demographics.smsp import (\r\n get_demographics_from_search,\r\n get_demographics_from_nhs_number,\r\n SmspException,\r\n)\r\nfrom lbrc_flask.logging import log_exception\r\nfrom lbrc_flask.emailing import email\r\nfrom identity.services.pmi import get_pmi, PmiException\r\nfrom lbrc_flask.data_conversions import (\r\n convert_dob,\r\n convert_gender,\r\n convert_name,\r\n convert_nhs_number,\r\n convert_postcode,\r\n convert_uhl_system_number,\r\n)\r\n\r\n\r\nclass ScheduleException(Exception):\r\n pass\r\n\r\n\r\ndef schedule_lookup_tasks(demographics_request_id):\r\n do_lookup_tasks.delay(demographics_request_id)\r\n\r\n@celery.task()\r\ndef do_lookup_tasks(demographics_request_id):\r\n current_app.logger.info(f'schedule_lookup_tasks (demographics_request_id={demographics_request_id})')\r\n\r\n try:\r\n dr = DemographicsRequest.query.get(demographics_request_id)\r\n\r\n if dr is None:\r\n raise Exception('Request id={} not found'.format(demographics_request_id))\r\n\r\n if dr.paused or dr.deleted or dr.result_created or dr.in_error:\r\n raise ScheduleException(f'Request id={demographics_request_id} scheduled when status is \"{dr.status}\"\"')\r\n\r\n current_app.logger.info(f'Scheduling demographics_request_id={demographics_request_id} with status \"{dr.status}\"')\r\n\r\n if not dr.data_extracted:\r\n extract_data.delay(demographics_request_id)\r\n elif not dr.pmi_data_pre_completed and not dr.skip_pmi:\r\n extract_pre_pmi_details.delay(demographics_request_id)\r\n elif not dr.lookup_completed:\r\n process_demographics_request_data.delay(demographics_request_id)\r\n elif not dr.pmi_data_post_completed and not dr.skip_pmi:\r\n extract_post_pmi_details.delay(demographics_request_id)\r\n elif not dr.result_created_datetime:\r\n produce_demographics_result.delay(demographics_request_id)\r\n\r\n db.session.add(dr)\r\n db.session.commit()\r\n\r\n except ScheduleException as sde:\r\n current_app.logger.warning(sde)\r\n except Exception as e:\r\n log_exception(e)\r\n save_demographics_error(demographics_request_id, e)\r\n\r\n\r\nclass SpineParameters:\r\n\r\n Warning = namedtuple('Warning', ['scope', 'message', 'message_type', 'source'])\r\n\r\n def __init__(self):\r\n self.nhs_number = None\r\n self.family_name = None\r\n self.given_name = None\r\n self.gender = None\r\n self.dob = None\r\n self.postcode = None\r\n self.warnings = []\r\n\r\n @property\r\n def valid_nhs_number_lookup_parameters(self):\r\n return self.nhs_number and self.dob\r\n\r\n @property\r\n def valid_search_lookup_parameters(self):\r\n return self.dob and self.gender\r\n \r\n def add_warning(self, scope, message, message_type='warning', source='validation'):\r\n self.warnings.append(SpineParameters.Warning(scope=scope, message=message, message_type=message_type, source=source))\r\n\r\n\r\ndef spine_lookup(demographics_request_data):\r\n params = get_spine_parameters(demographics_request_data)\r\n\r\n try:\r\n if params.valid_nhs_number_lookup_parameters:\r\n demographics = get_demographics_from_nhs_number(\r\n nhs_number=params.nhs_number,\r\n dob=params.dob,\r\n )\r\n elif params.valid_search_lookup_parameters:\r\n if not params.gender:\r\n params.add_warning(scope='gender', message='Missing value')\r\n\r\n demographics = get_demographics_from_search(\r\n family_name=params.family_name,\r\n given_name=params.given_name,\r\n gender=params.gender,\r\n dob=params.dob,\r\n postcode=params.postcode,\r\n )\r\n else:\r\n params.add_warning(message_type='error', scope='request', message='Not enough values to perform Spine lookup')\r\n demographics = None\r\n\r\n if demographics:\r\n demographics_request_data.response = DemographicsRequestDataResponse(\r\n nhs_number=demographics.nhs_number,\r\n title=demographics.title,\r\n forename=demographics.forename,\r\n middlenames=demographics.middlenames,\r\n lastname=demographics.lastname,\r\n sex=demographics.sex,\r\n postcode=demographics.postcode,\r\n address=demographics.address,\r\n date_of_birth=parse(demographics.date_of_birth) if demographics.date_of_birth else None,\r\n date_of_death=parse(demographics.date_of_death) if demographics.date_of_death else None,\r\n is_deceased=demographics.is_deceased,\r\n current_gp_practice_code=demographics.current_gp_practice_code,\r\n )\r\n\r\n except SmspException as e:\r\n params.add_warning(message_type='error', source='spine', scope='request', message=e.message)\r\n except Exception as e:\r\n params.add_warning(message_type='unknown error', source='spine', scope='request', message=str(e))\r\n finally:\r\n for w in params.warnings:\r\n demographics_request_data.messages.append(\r\n DemographicsRequestDataMessage(\r\n type=w.message_type,\r\n source=w.source,\r\n scope=w.scope,\r\n message=w.message,\r\n )\r\n )\r\n\r\n\r\ndef get_spine_parameters(demographics_request_data):\r\n result = SpineParameters()\r\n\r\n error, v_nhs_number = convert_nhs_number(demographics_request_data.nhs_number)\r\n if error is not None:\r\n result.add_warning(scope='nhs_number', message=error)\r\n\r\n error, v_gender = convert_gender_with_default(demographics_request_data.demographics_request.column_definition, demographics_request_data.gender)\r\n if error is not None:\r\n result.add_warning(scope='gender', message=error)\r\n\r\n error, v_family_name = convert_name(demographics_request_data.family_name)\r\n if error is not None:\r\n result.add_warning(scope='family_name', message=error)\r\n\r\n error, v_given_name = convert_name(demographics_request_data.given_name)\r\n if error is not None:\r\n result.add_warning(scope='given_name', message=error)\r\n\r\n error, v_dob = convert_dob(demographics_request_data.dob)\r\n if error is not None:\r\n result.add_warning(scope='dob', message=error)\r\n\r\n error, v_postcode = convert_postcode(demographics_request_data.postcode)\r\n if error is not None:\r\n result.add_warning(scope='postcode', message=error)\r\n\r\n if demographics_request_data.pmi_data:\r\n error, v_pmi_nhs_number = convert_nhs_number(demographics_request_data.pmi_data.nhs_number)\r\n if error is not None:\r\n result.add_warning(scope='pmi_nhs_number', message=error)\r\n\r\n error, v_pmi_gender = convert_gender_with_default(demographics_request_data.demographics_request.column_definition, demographics_request_data.pmi_data.gender)\r\n if error is not None:\r\n result.add_warning(scope='pmi_gender', message=error)\r\n\r\n error, v_pmi_family_name = convert_name(demographics_request_data.pmi_data.family_name)\r\n if error is not None:\r\n result.add_warning(scope='pmi_family_name', message=error)\r\n\r\n error, v_pmi_given_name = convert_name(demographics_request_data.pmi_data.given_name)\r\n if error is not None:\r\n result.add_warning(scope='pmi_given_name', message=error)\r\n\r\n error, v_pmi_dob = convert_dob(demographics_request_data.pmi_data.date_of_birth)\r\n if error is not None:\r\n result.add_warning(scope='pmi_date_of_birth', message=error)\r\n\r\n error, v_pmi_postcode = convert_postcode(demographics_request_data.pmi_data.postcode)\r\n if error is not None:\r\n result.add_warning(scope='pmi_postcode', message=error)\r\n\r\n else:\r\n v_pmi_nhs_number = None\r\n v_pmi_gender = None\r\n v_pmi_family_name = None\r\n v_pmi_given_name = None\r\n v_pmi_dob = None\r\n v_pmi_postcode = None\r\n\r\n result.nhs_number=(v_nhs_number or v_pmi_nhs_number)\r\n result.dob=(v_dob or v_pmi_dob)\r\n result.family_name=(v_family_name or v_pmi_family_name)\r\n result.given_name=(v_given_name or v_pmi_given_name)\r\n result.gender=(v_gender or v_pmi_gender)\r\n result.postcode=(v_postcode or v_pmi_postcode)\r\n\r\n return result\r\n\r\ndef convert_gender_with_default(column_definition, value):\r\n value = value.lower()\r\n\r\n gender_female_value = (column_definition.gender_female_value or '').lower()\r\n gender_male_value = (column_definition.gender_male_value or '').lower()\r\n\r\n if len(gender_female_value) > 0:\r\n if value == gender_female_value:\r\n value = 'f'\r\n\r\n if len(gender_male_value) > 0:\r\n if value == gender_male_value:\r\n value = 'm'\r\n\r\n return convert_gender(value)\r\n\r\n@celery.task()\r\ndef process_demographics_request_data(request_id):\r\n current_app.logger.info(f'process_demographics_request_data: request_id={request_id})')\r\n\r\n try:\r\n dr = DemographicsRequest.query.get(request_id)\r\n\r\n if dr is None:\r\n raise Exception('request not found')\r\n\r\n drd = DemographicsRequestData.query.filter_by(demographics_request_id=request_id).filter(\r\n DemographicsRequestData.processed_datetime.is_(None)\r\n ).first()\r\n\r\n if drd is None:\r\n dr.lookup_completed_datetime = datetime.utcnow()\r\n db.session.add(dr)\r\n else:\r\n # if not drd.has_error:\r\n spine_lookup(drd)\r\n \r\n drd.processed_datetime = datetime.utcnow()\r\n\r\n db.session.add(drd)\r\n\r\n db.session.commit()\r\n\r\n schedule_lookup_tasks(request_id)\r\n\r\n except Exception as e:\r\n log_exception(e)\r\n save_demographics_error(request_id, e)\r\n\r\n\r\n@celery.task()\r\ndef extract_data(request_id):\r\n current_app.logger.info(f'extract_data (request_id={request_id})')\r\n\r\n try:\r\n dr = DemographicsRequest.query.get(request_id)\r\n\r\n if dr is None:\r\n raise Exception('request not found')\r\n\r\n cd = dr.column_definition\r\n\r\n if len(dr.data) > 0:\r\n raise Exception(\r\n 'Attempting to extract data from DemographicsRequest (\"{}\") '\r\n 'that has already had data extracted.'.format(request_id)\r\n )\r\n\r\n if cd is None:\r\n raise Exception(\r\n 'Attempting to extract data from DemographicsRequest (\"{}\") '\r\n 'that did not have a column definition.'.format(request_id)\r\n )\r\n\r\n for i, r in enumerate(dr.iter_rows()):\r\n uhl_system_number = get_column_value(r, cd.uhl_system_number_column)\r\n nhs_number = get_column_value(r, cd.nhs_number_column)\r\n family_name = get_column_value(r, cd.family_name_column)\r\n given_name = get_column_value(r, cd.given_name_column)\r\n gender = get_column_value(r, cd.gender_column)\r\n dob = get_column_value(r, cd.dob_column)\r\n postcode = get_column_value(r, cd.postcode_column)\r\n\r\n if any([uhl_system_number, nhs_number, family_name, given_name, gender, dob, postcode]):\r\n d = DemographicsRequestData(\r\n demographics_request=dr,\r\n row_number=i,\r\n uhl_system_number=uhl_system_number,\r\n nhs_number=nhs_number,\r\n family_name=family_name,\r\n given_name=given_name,\r\n gender=gender,\r\n dob=dob,\r\n postcode=postcode,\r\n )\r\n\r\n current_app.logger.info(f'Saving extracting data={d}')\r\n\r\n dr.data.append(d)\r\n else:\r\n current_app.logger.info(f'Skipping empty data')\r\n\r\n dr.data_extracted_datetime = datetime.utcnow()\r\n db.session.add(dr)\r\n db.session.commit()\r\n\r\n schedule_lookup_tasks(request_id)\r\n\r\n except Exception as e:\r\n db.session.rollback()\r\n log_exception(e)\r\n save_demographics_error(request_id, e)\r\n\r\n\r\ndef get_column_value(record, column):\r\n if column is None:\r\n return ''\r\n\r\n if record[column.name] is None:\r\n return ''\r\n else:\r\n return str(record[column.name]).strip()\r\n\r\n\r\n@celery.task()\r\ndef produce_demographics_result(demographics_request_id):\r\n current_app.logger.info(f'produce_demographics_result (demographics_request_id={demographics_request_id})')\r\n\r\n try:\r\n dr = DemographicsRequest.query.get(demographics_request_id)\r\n\r\n current_app.logger.info(f'produce_demographics_result: Creating result')\r\n dr.create_result()\r\n\r\n dr.result_created_datetime = datetime.utcnow()\r\n\r\n db.session.add(dr)\r\n\r\n email(\r\n subject='Identity Demographics Request Complete',\r\n recipients=[dr.owner.email],\r\n message='Your demographics request {} is complete.'.format(\r\n url_for('ui.demographics', _external=True),\r\n ),\r\n html=render_template('email/request_complete.html', request=dr),\r\n )\r\n db.session.commit()\r\n except Exception as e:\r\n current_app.logger.info('produce_demographics_result: Rolling Back')\r\n db.session.rollback()\r\n log_exception(e)\r\n save_demographics_error(demographics_request_id, e)\r\n\r\n\r\n@celery.task()\r\ndef extract_pre_pmi_details(request_id):\r\n current_app.logger.info(f'extract_pre_pmi_details (request_id={request_id})')\r\n\r\n extract_pmi_details(\r\n request_id=request_id,\r\n data_selection_condition=DemographicsRequestData.pmi_pre_processed_datetime.is_(None),\r\n request_completed_attribute='pmi_data_pre_completed_datetime',\r\n data_completed_attribute='pmi_pre_processed_datetime',\r\n )\r\n\r\n\r\n@celery.task()\r\ndef extract_post_pmi_details(request_id):\r\n current_app.logger.info(f'extract_pre_pmi_details (request_id={request_id})')\r\n\r\n extract_pmi_details(\r\n request_id=request_id,\r\n data_selection_condition=DemographicsRequestData.pmi_post_processed_datetime.is_(None),\r\n request_completed_attribute='pmi_data_post_completed_datetime',\r\n data_completed_attribute='pmi_post_processed_datetime',\r\n )\r\n\r\n\r\ndef extract_pmi_details(request_id, data_selection_condition, request_completed_attribute, data_completed_attribute):\r\n current_app.logger.info(f'extract_pmi_details (request_id={request_id})')\r\n\r\n try:\r\n dr = DemographicsRequest.query.get(request_id)\r\n\r\n if dr is None:\r\n raise Exception('request not found')\r\n\r\n drd = DemographicsRequestData.query.filter_by(demographics_request_id=request_id).filter(data_selection_condition).first()\r\n\r\n if drd is None:\r\n current_app.logger.info(f'extract_pmi_details (request_id={request_id}): Done')\r\n setattr(dr, request_completed_attribute, datetime.utcnow())\r\n db.session.add(dr)\r\n else:\r\n if not drd.has_error and drd.pmi_data is None:\r\n get_pmi_details(drd)\r\n\r\n setattr(drd, data_completed_attribute, datetime.utcnow())\r\n db.session.add(drd)\r\n\r\n db.session.commit()\r\n\r\n schedule_lookup_tasks(request_id)\r\n\r\n except Exception as e:\r\n db.session.rollback()\r\n log_exception(e)\r\n save_demographics_error(request_id, e)\r\n\r\n\r\ndef get_pmi_details(drd):\r\n current_app.logger.info(f'get_pmi_details (Data request Data={drd.id})')\r\n\r\n try:\r\n error, v_nhs_number = convert_nhs_number(drd.nhs_number)\r\n if error is not None:\r\n drd.messages.append(\r\n DemographicsRequestDataMessage(\r\n type='warning',\r\n source='pmi_details',\r\n scope='nhs_number',\r\n message=error,\r\n )\r\n )\r\n\r\n error, v_s_number = convert_uhl_system_number(drd.uhl_system_number)\r\n if error is not None:\r\n drd.messages.append(\r\n DemographicsRequestDataMessage(\r\n type='warning',\r\n source='pmi_details',\r\n scope='uhl_system_number',\r\n message=error,\r\n )\r\n )\r\n\r\n pmi = get_pmi(nhs_number=v_nhs_number, uhl_system_number=v_s_number)\r\n\r\n if pmi is not None:\r\n pmi_details = DemographicsRequestPmiData(**pmi._asdict())\r\n\r\n pmi_details.demographics_request_data_id = drd.id\r\n db.session.add(pmi_details)\r\n\r\n except PmiException as e:\r\n drd.messages.append(\r\n DemographicsRequestDataMessage(\r\n type='error',\r\n source='pmi_details',\r\n scope='pmi_details',\r\n message=e.message,\r\n ))\r\n except Exception as e:\r\n log_exception(e)\r\n drd.messages.append(\r\n DemographicsRequestDataMessage(\r\n type='error',\r\n source='pmi_details',\r\n scope='pmi_details',\r\n message=traceback.format_exc(),\r\n ))\r\n\r\n\r\ndef save_demographics_error(demographics_request_id, e):\r\n dr = DemographicsRequest.query.get(demographics_request_id)\r\n if dr is not None:\r\n dr = DemographicsRequest.query.get(demographics_request_id)\r\n dr.set_error(traceback.format_exc())\r\n db.session.add(dr)\r\n db.session.commit()\r\n","sub_path":"identity/demographics/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":18023,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"281000333","text":"from model import piece\nfrom model import work\nfrom model import author\nfrom config.setting import db\n\npieces = piece.get_all()\nfor p in pieces:\n data = piece.parse_content(p.content)\n\n author_name = data[\"author\"]\n work_title = data[\"work\"]\n\n if p[\"author\"]:\n data[\"author\"] = p[\"author\"]\n elif data[\"author\"]:\n # return author_id\n data[\"author\"] = author.add(author_name)\n \n if p[\"work\"]:\n data[\"work\"] = p[\"work\"]\n elif data[\"work\"]:\n # return work_id\n data[\"work\"] = work.add(work_title)\n\n # if author_name or work_title:\n # # print p.content\n # print \"origin:\\t\\t\" + p.content\n # print \"content:\\t\" + data[\"content\"]\n # print \"author:\\t\\t\" + (author_name and author_name or \"None\")\n # print \"work:\\t\\t\" + (work_title and work_title or \"None\")\n # print \"\"\n \n db.update(\"piece\",where=\"id=$id\",vars=p,**data)","sub_path":"parse_old_contents.py","file_name":"parse_old_contents.py","file_ext":"py","file_size_in_byte":927,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"374491709","text":"#%%\nimport csv\nimport pickle\nimport pathlib\nimport numpy as np\nfrom sklearn.decomposition import PCA\nfrom gensim.models import Word2Vec\n\nMODEL = pathlib.Path('model_aligned/')\nMODEL_EXPORT = pathlib.Path('model_export/')\nVOCAB = 'common_vocab.pkl'\nOUTPUT_PCA = \"aligned_wv_pca.csv\"\nwith open(VOCAB, \"rb\") as f:\n VOCAB = pickle.load(f)\nVOCAB.remove(\"\")\nVOCAB = list(VOCAB)\n\n#%%\nall_wv = []\nall_words = []\n\nfor fp in MODEL.glob(\"*.model\"):\n out_fp = MODEL_EXPORT / (fp.stem + \".txt\")\n\n # Load aligned model\n model = Word2Vec.load(str(fp)).wv\n \n with open(out_fp, \"w\", encoding=\"utf-8\") as f:\n # Select only words in common vocab\n for word in VOCAB:\n vec = model[word]\n # Write to plain text\n vec_str = '\\t'.join(str(n) for n in vec)\n f.write(f\"{word}\\t{vec_str}\\n\")\n\n # Save to big matrix\n src, ts = fp.stem.split(\"_\")\n all_wv.append(vec)\n all_words.append(f\"{word}_{src}_{ts}\")\n\n\n#%%\n# PCA reduce all words in corpus\nall_wv = np.array(all_wv)\ntwodim = PCA().fit_transform(all_wv)[:, :2]\n\n# Save PCA results\nwith open(OUTPUT_PCA, \"w\", encoding=\"utf-8\", newline=\"\") as f:\n writer = csv.writer(f)\n writer.writerow([\"word\", \"src\", \"timestep\", \"PC1\", \"PC2\"])\n\n for word, (pc1, pc2) in zip(all_words, twodim):\n word, src, ts = word.split(\"_\")\n writer.writerow([word, src, ts, pc1, pc2])\n","sub_path":"src/word2vec/export.py","file_name":"export.py","file_ext":"py","file_size_in_byte":1423,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"334875118","text":"#!/usr/bin/python\n#coding:utf-8\n\n# 2013/03/17\n\n'''\nPyGraphviz は Graphviz のPythonインタフェースである。\nGraphviz は グラフのレイアウトと可視化パッケージである。\nPyGraphvizを使うことで, Pythonを使ってグラフを生成、編集、読み込み、書き込み、描写しGraphviz グラフデータ構造とレイアウトアルゴリズムにアクセスすることができる。 \nPyGraphviz は Networkxとは独立したものだが, 似たようなプログラミングインタフェースを提供する。 \n'''\n\n# Quick Example\nimport pygraphviz as pgv\nG=pgv.AGraph()\nG.add_node('a')\nG.add_edge('b','c')\nG\n'''\nstrict graph {\n a;\n b -- c;\n}\n'''\n#To load a dot file use\nG=pgv.AGraph(\"file.dot\")\n\n\n","sub_path":"3rd_party/pygraphviz/index.py","file_name":"index.py","file_ext":"py","file_size_in_byte":762,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"17874231","text":"\"\"\"Test KFRA backpropagation\n\nH_in → 1/N ∑ₙ Jₙ^T H_out Jₙ\n\nNotes:\n - `Dropout` cannot be tested,as the `autograd` implementation does a forward\n pass over each sample, while the `backpack` implementation requires only\n one forward pass over the batched data. This leads to different outputs,\n as `Dropout` is not deterministic.\n\"\"\"\nimport pytest\nimport torch\nfrom torch.nn import AvgPool2d, Conv2d, Linear, MaxPool2d, Sigmoid, Tanh, ZeroPad2d\n\nfrom backpack import extend\nfrom backpack.core.derivatives import derivatives_for\nfrom backpack.hessianfree.lop import transposed_jacobian_vector_product\n\nfrom .automated_test import check_sizes, check_values\n\n\ndef make_id(layer, input_shape):\n return \"in{}-{}\".format(input_shape, layer)\n\n\ntorch.manual_seed(0)\nARGS = \"layer,input_shape\"\nSETTINGS = [\n # (layer, input_shape)\n [Linear(20, 10), (5, 20)],\n [MaxPool2d(kernel_size=2), (5, 3, 10, 8)],\n [MaxPool2d(kernel_size=2), (1, 2, 4, 4)],\n [MaxPool2d(kernel_size=3, stride=2, padding=1), (3, 2, 9, 11)],\n [AvgPool2d(kernel_size=3), (3, 2, 15, 13)],\n [AvgPool2d(kernel_size=3, stride=2), (3, 2, 15, 13)],\n [AvgPool2d(kernel_size=3, stride=2, padding=1), (3, 2, 15, 13)],\n [Sigmoid(), (6, 20)],\n [Sigmoid(), (6, 2, 7)],\n [Tanh(), (6, 20)],\n [Tanh(), (6, 2, 7)],\n [Conv2d(2, 3, kernel_size=2), (3, 2, 11, 13)],\n [Conv2d(2, 3, kernel_size=2, padding=1), (3, 2, 11, 13)],\n [Conv2d(2, 3, kernel_size=2, padding=1, stride=2), (3, 2, 11, 13)],\n [Conv2d(2, 3, kernel_size=2, padding=1, stride=2, dilation=2), (3, 2, 11, 13)],\n [ZeroPad2d(2), (4, 3, 4, 5)],\n]\nIDS = [make_id(layer, input_shape) for (layer, input_shape) in SETTINGS]\n\n\ndef get_output_shape(input, layer):\n return layer(input).shape\n\n\ndef autograd_ea_jac_t_mat_jac_prod(layer, input, mat):\n def sample_jac_t_mat_jac_prod(layer, sample, mat):\n assert sample.shape[0] == 1, \"input is not batch size 1: {}\".format(\n sample.shape\n )\n assert len(mat.shape) == 2\n\n def sample_jac_t_mat_prod(layer, sample, mat):\n result = torch.zeros(sample.numel(), mat.size(1))\n\n sample.requires_grad = True\n output = layer(sample)\n\n for col in range(mat.size(1)):\n column = mat[:, col].reshape(output.shape)\n result[:, col] = transposed_jacobian_vector_product(\n [output], [sample], [column], retain_graph=True\n )[0].reshape(-1)\n\n return result\n\n jac_t_mat = sample_jac_t_mat_prod(layer, sample, mat)\n mat_t_jac = jac_t_mat.t()\n jac_t_mat_t_jac = sample_jac_t_mat_prod(layer, sample, mat_t_jac)\n jac_t_mat_jac = jac_t_mat_t_jac.t()\n\n return jac_t_mat_jac\n\n N = input.shape[0]\n input_features = input.shape.numel() // N\n\n result = torch.zeros(input_features, input_features)\n\n for n in range(N):\n sample_n = input[n].unsqueeze(0)\n result += sample_jac_t_mat_jac_prod(layer, sample_n, mat)\n\n return result / N\n\n\ndef derivative_from_layer(layer):\n layer_to_derivative = derivatives_for\n\n for module_cls, derivative_cls in layer_to_derivative.items():\n if isinstance(layer, module_cls):\n return derivative_cls()\n\n raise RuntimeError(\"No derivative available for {}\".format(layer))\n\n\ndef backpack_ea_jac_t_mat_jac_prod(layer, input, mat):\n layer = extend(layer)\n derivative = derivative_from_layer(layer)\n\n # forward pass to initialize backpack buffers\n _ = layer(input)\n\n return derivative.ea_jac_t_mat_jac_prod(layer, None, None, mat)\n\n\ndef generate_data_ea_jac_t_mat_jac_prod(layer, input_shape):\n input = torch.rand(input_shape)\n out_features = get_output_shape(input, layer)[1:].numel()\n mat = torch.rand(out_features, out_features)\n return input, mat\n\n\n@pytest.mark.parametrize(ARGS, SETTINGS, ids=IDS)\ndef test_ea_jac_t_mat_jac_prod(layer, input_shape):\n torch.manual_seed(0)\n input, mat = generate_data_ea_jac_t_mat_jac_prod(layer, input_shape)\n return _compare_ea_jac_t_mat_jac_prod(layer, input, mat)\n\n\ndef _compare_ea_jac_t_mat_jac_prod(layer, input, mat):\n autograd_result = autograd_ea_jac_t_mat_jac_prod(layer, input, mat)\n backpack_result = backpack_ea_jac_t_mat_jac_prod(layer, input, mat)\n\n check_sizes(autograd_result, backpack_result)\n check_values(autograd_result, backpack_result)\n\n return backpack_result\n\n\ndef test_ea_jac_t_mat_jac_prod_linear_manual():\n # Linear with manual ea_jac_t_mat_jac_prod\n input_shape = (5, 13)\n layer = torch.nn.Linear(13, 10)\n\n input, mat = generate_data_ea_jac_t_mat_jac_prod(layer, input_shape)\n\n test_result = _compare_ea_jac_t_mat_jac_prod(layer, input, mat)\n manual_result = torch.einsum(\n \"ab,ac,cd->bd\", layer.weight.data, mat, layer.weight.data\n )\n\n check_sizes(test_result, manual_result)\n check_values(test_result, manual_result)\n","sub_path":"test/test_ea_jac_t_mat_jac_prod.py","file_name":"test_ea_jac_t_mat_jac_prod.py","file_ext":"py","file_size_in_byte":4943,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"501369868","text":"import logging\nimport numpy as np\nimport pandas as pd\n\nfrom countess.plugins.scoring import BaseScorerPlugin\nfrom countess.plugins.options import Options\nfrom countess.base.constants import WILD_TYPE_VARIANT\nfrom countess.base.utils import log_message\nfrom countess.base.constants import IDENTIFIERS, VARIANTS\n\n\noptions = Options()\noptions.add_option(\n name=\"Normalization Method\",\n varname=\"logr_method\",\n dtype=str,\n default=\"Wild Type\",\n choices={\"Wild Type\": \"wt\", \"Full\": \"full\", \"Complete\": \"complete\"},\n hidden=False,\n)\n\n\nclass RatiosScorer(BaseScorerPlugin):\n\n name = \"Ratios\"\n version = \"1.0\"\n author = \"Alan Rubin, Daniel Esposito\"\n\n def compute_scores(self):\n for label in self.store_labels():\n self.calc_ratios(label)\n\n def calc_ratios(self, label):\n \"\"\"\n Calculate frequency ratios and standard errors between the\n last timepoint and the input. Ratios can be calculated using\n one of three methods:\n - wt\n - complete\n - full\n \"\"\"\n if self.store_check(\"/main/{}/scores\".format(label)):\n return\n\n log_message(\n logging_callback=logging.info,\n msg=\"Calculating ratios ({})\".format(label),\n extra={\"oname\": self.name},\n )\n c_last = \"c_{}\".format(self.store_timepoints()[-1])\n df = self.store_select(\n key=\"/main/{}/counts\".format(label), columns=[\"c_0\", \"{}\".format(c_last)]\n )\n\n if self.logr_method == \"wt\":\n if VARIANTS in self.store_labels():\n wt_label = VARIANTS\n elif IDENTIFIERS in self.store_labels():\n wt_label = IDENTIFIERS\n else:\n raise ValueError(\n \"Failed to use wild type log \"\n \"ratio method, suitable data \"\n \"table not present [{}]\".format(self.name)\n )\n\n shared_counts = self.store_select(\n key=\"/main/{}/counts\".format(wt_label),\n columns=[\"c_0\", \"{}\".format(c_last)],\n where=\"index='{}'\".format(WILD_TYPE_VARIANT),\n )\n\n # wild type not found\n if len(shared_counts) == 0:\n raise ValueError(\n \"Failed to use wild type log \"\n \"ratio method, wild type \"\n \"sequence not present [{}]\".format(self.name)\n )\n\n shared_counts = shared_counts.values + 0.5\n\n elif self.logr_method == \"complete\":\n shared_counts = (\n self.store_select(\n key=\"/main/{}/counts\".format(label),\n columns=[\"c_0\", \"{}\".format(c_last)],\n )\n .sum(axis=\"index\")\n .values\n + 0.5\n )\n\n elif self.logr_method == \"full\":\n shared_counts = (\n self.store_select(\n key=\"/main/{}/counts_unfiltered\".format(label),\n columns=[\"c_0\", \"{}\".format(c_last)],\n )\n .sum(axis=\"index\", skipna=True)\n .values\n + 0.5\n )\n else:\n raise ValueError(\n 'Invalid log ratio method \"{}\" '\n \"[{}]\".format(self.logr_method, self.name)\n )\n\n ratios = np.log(df[[\"c_0\", c_last]].values + 0.5) - np.log(shared_counts)\n ratios = ratios[:, 1] - ratios[:, 0] # selected - input\n ratios = pd.DataFrame(ratios, index=df.index, columns=[\"logratio\"])\n\n shared_variance = np.sum(1.0 / shared_counts)\n summed = np.sum(1.0 / (df[[\"c_0\", c_last]].values + 0.5), axis=1)\n\n ratios[\"variance\"] = summed + shared_variance\n ratios[\"score\"] = ratios[\"logratio\"]\n ratios[\"SE\"] = np.sqrt(ratios[\"variance\"])\n\n # re-order columns\n ratios = ratios[[\"score\", \"SE\", \"logratio\", \"variance\"]]\n self.store_put(\n key=\"/main/{}/scores\".format(label),\n value=ratios,\n data_columns=ratios.columns,\n )\n","sub_path":"countess/tests/data/plugins/ratios_scorer.py","file_name":"ratios_scorer.py","file_ext":"py","file_size_in_byte":4144,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"223751298","text":"from API.Toolbar.GoldenPhysicalToolbar.GoldenPhysicalToolbarConst import GoldenPhysicalToolbarConst\nfrom API.Utility.Util import Util\n\n#Function initialization\nutil = Util()\n\n#Device initialization\n\ndef main():\n util.init()\n util.clickOnPhysical()\n checkPoint()\n\ndef checkPoint():\n util.clickButton(GoldenPhysicalToolbarConst.NEW_CITY)\n snooze(2)\n # Verification Point 'VP1'\n test.compare(findObject(\":CAppWindowBase.centralwidget.m_pWorkSpaceWnd.CViewArea_Window1.QStackedWidget1.CWorkspace1.CGeoView1.QGraphicsItem_2\").toolTip, \"City\")\n # Verification Point 'VP3'\n test.compare(findObject(\":CAppWindowBase.centralwidget.m_pWorkSpaceWnd.CViewArea_Window1.QStackedWidget1.CWorkspace1.CGeoView1.QGraphicsItem_2\").visible, True)\n \n","sub_path":"trunk/workspace/Squish/src/TestScript/UI/suite_UI_51/tst_UI_51_PhysicalToolbar_NewCityButton/test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":759,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"498049958","text":"import unittest\nimport wikichatter.comment as comment\nimport wikichatter.textblock as textblock\n\nLEVEL0 = \"\"\nLEVEL1 = \":\"\nLEVEL2 = \"::\"\nLEVEL3 = \":::\"\nLEVEL4 = \"::::\"\nSIGNATURE = \"[[User:Name|Name]] ([[User talk:Name|talk]]) 19:40, 18 September 2013 (UTC)\"\nFILLER = \"Text\"\nEL = \"\\n\"\n\n\nclass CommentTest(unittest.TestCase):\n\n def test_basic_linear_identification(self):\n text = (\n LEVEL0 + FILLER + EL +\n LEVEL1 + FILLER + SIGNATURE + EL +\n LEVEL0 + FILLER + SIGNATURE + EL\n )\n blocks = textblock.generate_textblock_list(text)\n\n comments = comment.identify_comments_linear_merge(blocks)\n\n self.assertEqual(len(comments), 2)\n\n def test_linear_identification_hierarchy(self):\n text = (\n LEVEL0 + FILLER + EL +\n LEVEL0 + FILLER + SIGNATURE + EL +\n LEVEL1 + FILLER + SIGNATURE + EL +\n LEVEL2 + FILLER + EL +\n LEVEL2 + FILLER + SIGNATURE + EL +\n LEVEL1 + FILLER + SIGNATURE + EL\n )\n blocks = textblock.generate_textblock_list(text)\n\n comments = comment.identify_comments_linear_merge(blocks)\n\n self.assertEqual(len(comments), 1)\n self.assertEqual(len(comments[0].comments), 2)\n self.assertEqual(len(comments[0].comments[0].comments), 1)\n self.assertEqual(len(comments[0].comments[1].comments), 0)\n","sub_path":"test/test_comment.py","file_name":"test_comment.py","file_ext":"py","file_size_in_byte":1387,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"297856072","text":"# reading whole file at a time\n# lines at a time using list\n# writing line by line / whole \n\n\n\n\n\n\nfw = open('sample.txt', 'w')\nfw.write('Writing some stuff in my text file\\n')\nfw.write('Another sample line\\n')\nfw.close()\n\nfr = open('sample.txt', 'r')\ntext = fr.read()\nprint(text)\nfr.close()","sub_path":"Python Practice/23_files.py","file_name":"23_files.py","file_ext":"py","file_size_in_byte":290,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"544380491","text":"import logging\nimport re\nfrom typing import Optional, Tuple\n\nlogger = logging.getLogger(__name__)\n\n\ndef extract_version(\n string: str, name: Optional[str] = None\n) -> Optional[Tuple[str, str]]:\n \"\"\"\n Heuristic to extract a version-number from a string.\n\n See test file for supported formats. Returns None if no unambiguously\n version number could be found.\n\n :param string: the string to search\n :param name: the name of the program\n :return: None or a tuple of two strings:\n - type of version (\"stable\", \"beta\", \"alpha\", \"rc\" or \"unstable\")\n - version number\n \"\"\"\n # Remove a prefix of the name of the program if existent\n if name:\n namere = re.compile(r\"^\" + re.escape(name) + r\"[ _-]?\", re.IGNORECASE)\n string = re.sub(namere, \"\", string)\n string = re.sub(\n r\"^(releases|release|rel|version|vers|v\\.)[ _/-]?\",\n \"\",\n string,\n flags=re.IGNORECASE,\n )\n\n # Replace underscore/hyphen with dots if only underscores/hyphens are used\n if re.fullmatch(r\"[0-9_]*\", string):\n string = string.replace(\"_\", \".\")\n if re.fullmatch(r\"[0-9-]*\", string):\n string = string.replace(\"-\", \".\")\n\n # Check for proper stable versions such as `v1.2.3`\n exact = re.compile(r\"[vV]?(\\d{1,3}(\\.\\d{1,3})*)\")\n match = exact.fullmatch(string)\n if match:\n return \"stable\", match.group(1)\n\n stable = re.compile(\n r\"(\\s|^|v)(\\d{1,3}(\\.\\d{1,3})+(-\\d\\d?|[a-z])?)(\\s|$)\", re.IGNORECASE\n )\n pre = re.compile(\n r\"(\\s|^|v)((\\d{1,3}(\\.\\d{1,3})+)[._ -]?(alpha|beta|pre|rc|b|preview)[._-]?\\d*)(\\s|$)\",\n re.IGNORECASE,\n )\n explicitstable = re.compile(\n r\"(\\s|^|v)(\\d{1,3}(\\.\\d{1,3})+)(-stable)(\\s|$)\", re.IGNORECASE\n )\n\n match_stable = list(stable.finditer(string))\n match_pre = list(pre.finditer(string))\n match_explicit = list(explicitstable.finditer(string))\n if len(match_stable) + len(match_pre) + len(match_explicit) > 1:\n logger.warning(\"Multiple versions found for {} in '{}'\".format(name, string))\n return None\n\n if match_stable:\n return \"stable\", match_stable[0].group(2).strip()\n\n if match_pre:\n state = re.search(\n r\"[^a-zA-Z](alpha|beta|rc|b)($|[^a-zA-Z])\", string, re.IGNORECASE\n )\n if state:\n statestr = state.group(1).lower()\n if statestr == \"b\":\n statestr = \"beta\"\n return statestr, match_pre[0].group(2).strip()\n else:\n return \"unstable\", match_pre[0].group(2).strip()\n\n if match_explicit:\n return \"stable\", match_explicit[0].group(2).strip()\n\n return None\n","sub_path":"versionhandler.py","file_name":"versionhandler.py","file_ext":"py","file_size_in_byte":2675,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"102401916","text":"class Contact(object):\n def __init__(self, age, gender, first_name, last_name, phone_number):\n self.age = age\n self.gender = gender\n self.first_name = first_name\n self.last_name = last_name\n self.phone_number = phone_number\n self.recent_calls = []\n self.favorite = False\n\n @property\n def full_name(self):\n return \"{0} {1}\".format(self.first_name, self.last_name)","sub_path":"PythonFlow_March2018/Homework/valeriybutorin/hw4/entities/contact.py","file_name":"contact.py","file_ext":"py","file_size_in_byte":427,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"441116002","text":"# url = \"http://www.tianqi.com/\" \"http://lishi.tianqi.com/jiande/index.html\"\n\nimport time\nimport sqlite3\nimport threading\nimport commom as cj\n\nzhejiangcities = {'杭州' : 'hangzhou', '湖州' : 'huzhou', '嘉兴' : 'jiaxing', '金华' : 'jinhua', '丽水' : 'lishui', '宁波' : 'ningbo', '衢州' : 'quzhou1', '绍兴' : 'shaoxing', '台州' :'taizhou', '温州' : 'wenzhou', '舟山' : 'zhoushan'}\n\n\nconn = None\n\nclass City(threading.Thread):\n def __init__(self, citypy, conn):\n super(City, self).__init__()\n self.citypy = citypy\n self.conn = conn\n\n def run(self):\n try:\n cur = self.conn.cursor()\n soup = cj.req('http://' + self.citypy + \".tianqi.com\")\n datas = soup.select('#zone > option')\n for data in datas:\n cur.execute('insert into cities (cityname, citypy) values (?,?)', (data.get_text(), data.get('py')))\n print(self.citypy, data.get_text(), data.get('py'))\n lishiSoup = cj.req('http://lishi.tianqi.com/' + data.get('py') + \"/index.html\")\n lishiDatas = lishiSoup.select('#tool_site > .tqtongji1 > ul > li > a')\n for lishiData in lishiDatas:\n cur.execute('insert into lishiUrls (citypy, lishiUrl) values (?,?)', (data.get('py'), lishiData.get('href')))\n print('\\t', data.get('py'), lishiData.get('href'))\n\n cur.close()\n self.conn.commit()\n except Exception as err:\n cj.error('City', err, self.citypy)\n print(err)\n\nclass Weather(threading.Thread):\n def __init__(self, citypy, conn):\n super(Weather, self).__init__()\n self.citypy = citypy\n self.conn = conn\n\n def run(self):\n try:\n cur = self.conn.cursor()\n cur.execute(\"select lishiUrl from lishiUrls where citypy=?\", (self.citypy,))\n for (lishiUrl,) in cur.fetchall():\n soup = cj.req(lishiUrl)\n datas = soup.select('#tool_site > .tqtongji2 > ul')\n del datas[0]\n for data in datas:\n t = [e.get_text() for e in data.children if type(e) == type(data)]\n t.insert(0, self.citypy)\n cur.execute('insert into weather (citypy,date,max_temperature,min_temperature,weather,wind_direction,wind_power) values (?,?,?,?,?,?,?)', tuple(t))\n print(self.citypy, lishiUrl)\n\n cur.close()\n self.conn.commit()\n except Exception as err:\n cj.error('Weather', err, self.citypy)\n print(err)\n\n\n\ndef today_weather(pydatas):\n for py in pydatas:\n soup = cj.req('http://' + py + 'hangzhou''.tianqi.com')\n datas = soup.select('#today > ul')\n\ntry:\n def req_cities():\n for cityname, citypy in zhejiangcities.items():\n t = City(citypy, conn)\n t.start()\n t.join()\n def req_weather():\n global conn\n cur = conn.cursor()\n cur.execute(\"select citypy from cities\")\n for (citypy,) in cur.fetchall():\n w = Weather(citypy, conn)\n w.start()\n w.join()\n def req_everyday():\n pass\n\n\n def main():\n cj.create_data_dir()\n global conn\n conn = sqlite3.connect(\"data/zhejiangweath\" + str(time.time()) + \".db\", check_same_thread=False)\n cur = conn.cursor()\n cur.execute('create table cities (id integer primary key,cityname varchar(20),citypy varchar(20))')\n cur.execute('create table lishiUrls (id integer primary key,citypy varchar(20),lishiUrl varchar(50))')\n cur.execute('create table weather (id integer primary key,citypy varchar(20),date varchar(20),max_temperature integer,min_temperature integer,weather varchar(20),wind_direction varchar(20),wind_power varchar(20))')\n cur.close()\n\n req_cities()\n req_weather()\n\n def test():\n s = '''\n conn = sqlite3.connect(\"data/zhejiangweath1467773928.460905.db\", check_same_thread=False)\n cur = conn.cursor()\n cur.execute(\"select lishiUrl from lishiUrls where citypy=?\", ('hangzhou',))\n for (lishiUrl,) in cur.fetchall():\n print(lishiUrl)\n cur.close()\n\n soup = cj.req('http://lishi.tianqi.com/zhoushan/201605.html')\n datas = soup.select('#tool_site > .tqtongji2 > ul')\n del datas[0]\n for data in datas:\n t = [e.get_text() for e in data.children if type(e) == type(data)]\n print(t)\n '''\n py = 'hangzhou'\n soup = cj.req('http://' + py + 'hangzhou''.tianqi.com')\n datas = [e.get_text().strip() for e in soup.select('#today > ul > li')]\n datas.extend(datas[-1].split(' '))\n datas[0] = py\n datas[1] = cj.now_time_str()\n datas.extend([e[:-1] for e in datas[2].split('~')])\n del datas[2],datas[3]\n #cur.execute('insert into weather (citypy,date,weather,wind_direction,wind_power,max_temperature,min_temperature) values (?,?,?,?,?,?,?)', tuple(t))\n print(tuple(datas))\n\n predicts = soup.select('#detail > div > ul')\n for index,p in enumerate(predicts):\n pres = [e.get_text() for e in p.children]\n pres.extend([e for e in pres[3].split(' ')])\n pres.extend([e[:-1] for e in pres[1].split('~')])\n pres[0] = py\n pres[1] = cj.now_time_str()\n pres[3] = index\n #pres.append(index)\n print(pres)\n\n\n\nexcept Exception as err:\n print(err)\n\nif __name__ == '__main__':\n #main()\n test()\n","sub_path":"Weather/test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":5610,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"111548149","text":"# xops URL Configuration\n\nfrom django.conf.urls import url, include\nfrom django.contrib import admin\n\n\nurlpatterns = [\n url(r'^admin/', admin.site.urls),\n url(r'^', include('dashboard.urls')),\n url(r'^alert/', include('alert.urls')),\n url(r'^cmdb/', include('cmdb.urls'))\n]","sub_path":"xops_m/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":285,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"477074352","text":"#!/usr/bin/python3\n\"\"\"New view for Amenity objects that handles all\"\"\"\n\nfrom models.amenity import Amenity\nfrom flask import Flask, abort, jsonify, make_response\nfrom flask import request\nfrom api.v1.views import app_views\nfrom models.state import State\nfrom models.city import City\nfrom models import storage\n\n\n@app_views.route('/amenities', methods=['GET'])\ndef all_amenities():\n \"\"\"Return all Amenities\"\"\"\n amenities = storage.all(Amenity).values()\n list_ame = []\n for amenity in amenities:\n list_ame.append(amenity.to_dict())\n return jsonify(list_ame)\n\n\n@app_views.route('/amenities/', methods=['GET'])\ndef amenity_id(amenity_id=None):\n \"\"\"Return amenity\"\"\"\n if amenity_id is not None:\n my_amenity_obj = storage.get(Amenity, amenity_id)\n if my_amenity_obj is None:\n abort(404)\n else:\n return jsonify(my_amenity_obj.to_dict())\n\n\n@app_views.route('/amenities/', methods=['DELETE'])\ndef amenity_delete(amenity_id=None):\n \"\"\"DELETE amenity\"\"\"\n if amenity_id is not None:\n my_amenity_obj = storage.get(Amenity, amenity_id)\n if my_amenity_obj is None:\n abort(404)\n else:\n storage.delete(my_amenity_obj)\n return make_response(jsonify({}), 200)\n\n\n@app_views.route('/amenities', methods=['POST'])\ndef amenity_post():\n \"\"\"POST amenity\"\"\"\n my_json = request.get_json(silent=True)\n if my_json is not None:\n if \"name\" in my_json:\n name = my_json[\"name\"]\n new_city = Amenity(name=name)\n new_city.save()\n return make_response(jsonify(new_city.to_dict()), 201)\n else:\n abort(400, \"Missing name\")\n else:\n abort(400, \"Not a JSON\")\n\n\n@app_views.route('amenities/', methods=['PUT'])\ndef update_amenity(amenity_id=None):\n \"\"\"PUT amenity\"\"\"\n if amenity_id is not None:\n my_amenity_obj = storage.get(Amenity, amenity_id)\n if my_amenity_obj is None:\n abort(404)\n else:\n update_ = request.get_json(silent=True)\n if update_ is not None:\n for key, value in update_.items():\n setattr(my_amenity_obj, key, value)\n my_amenity_obj.save()\n return make_response(jsonify(my_amenity_obj.to_dict()), 200)\n else:\n abort(400, \"Not a JSON\")\n","sub_path":"api/v1/views/amenities.py","file_name":"amenities.py","file_ext":"py","file_size_in_byte":2415,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"332519287","text":"import Tablet\nimport kitchen.Cook as Cook\nimport kitchen.Waiter as Waiter\n\n\"\"\"\nRestaurant application\n\"\"\"\n\n\nclass Restaurant:\n def __init__(self, table_number, cooker_name):\n self.__tablet = Tablet.Tablet(table_number)\n self.__cooker = Cook.Cook(cooker_name)\n self.__tablet.add_subscriber(self.__cooker)\n self.__waiter = Waiter.Waiter()\n self.__cooker.add_subscriber(self.__waiter)\n self.__tablet.create_order()\n\n\nif __name__ == '__main__':\n Restaurant(8, 'Gordon James Ramsay Jr')\n","sub_path":"src/Restaurant.py","file_name":"Restaurant.py","file_ext":"py","file_size_in_byte":530,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"4443569","text":"def prod2(u,v):\r\n threshold = 4\r\n n = max(len(str(u)), len(str(v)))\r\n\r\n if u == 0 or v == 0:\r\n return 0\r\n elif n <= threshold:\r\n return u*v\r\n else:\r\n m = n // 2\r\n x, y = divmod(u, pow(10,m))\r\n w, z = divmod(v, pow(10,m))\r\n\r\n r = prod2(x+y, w+z)\r\n p = prod2(x, w)\r\n q = prod2(y, z)\r\n\r\n return p*pow(10,2*m) + (r-p-q)*pow(10,m) + q\r\n \r\na=1234567812345678\r\nb=2345678923456789\r\n\r\nprint(prod2(a,b))\r\nprint(a*b)\r\n","sub_path":"Python/2020-2_알고리즘분석/2_DivideConquer/2_6_ProductofLargeInt.py","file_name":"2_6_ProductofLargeInt.py","file_ext":"py","file_size_in_byte":488,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"605048616","text":"#coding=utf-8\n\n# Y.YANG\n\nimport socket\nfrom socket import error\nimport time\nimport queue\nimport threading\nimport logging\nimport os\n\n#Library for interface\nimport matplotlib.pyplot as plt\nimport numpy as np \nimport datetime as dt \nimport matplotlib.dates as mdates\nfrom datetime import datetime\n\n\n\nlogging.basicConfig(level=logging.INFO,\n filename='client.log',\n filemode='w',\n format='%(asctime)s %(filename)s[line:%(lineno)d] %(levelname)s %(message)s',\n datefmt='%a, %d %b %Y %H:%M:%S',)\nlog = logging.getLogger('client')\n\n\ndef create_sock(host,port):\n sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n sock.setsockopt(socket.SOL_SOCKET, socket.SO_KEEPALIVE, 1)\n while True:\n try:\n sock.connect((host, port))\n\n break\n except error as e:\n log.error(e)\n print(e)\n print('wait 5 seconds, reconnect again...')\n log.error('wait 5 seconds, reconnect again...')\n time.sleep(5)\n return sock\n\n\n\ndef start():\n\n #threading.Thread(target=write_data_to_queue).start()\n host = \"192.168.137.1\" # the server external ip, this ip should be ping success from current computer\n port = 9999\n sock = create_sock(host,port)\n while True:\n try:\n #print('receve data from'),\n #print(host)\n data = sock.recv(4096)\n print(data)\n time.sleep(1)\n except Exception as e:\n print(e)\n print('send data error...., reconnect to server ')\n print(host)\n\n log.error(e)\n sock = create_sock(host,port)\n\nif __name__ == \"__main__\":\n start()\n\n\n","sub_path":"recv_client_s.py","file_name":"recv_client_s.py","file_ext":"py","file_size_in_byte":1733,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"401797728","text":"# coding=utf-8\n# --------------------------------------------------------------------------------------------\n# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License. See License.txt in the project root for license information.\n# --------------------------------------------------------------------------------------------\n# coding: utf-8\n# pylint: skip-file\nfrom msrest.serialization import Model\n\n\nclass ServiceRunner(Model):\n \"\"\"A container for a managed identity to execute DevTest lab services.\n\n :param identity: The identity of the resource.\n :type identity: :class:`IdentityProperties\n `\n :param id: The identifier of the resource.\n :type id: str\n :param name: The name of the resource.\n :type name: str\n :param type: The type of the resource.\n :type type: str\n :param location: The location of the resource.\n :type location: str\n :param tags: The tags of the resource.\n :type tags: dict\n \"\"\"\n\n _attribute_map = {\n 'identity': {'key': 'identity', 'type': 'IdentityProperties'},\n 'id': {'key': 'id', 'type': 'str'},\n 'name': {'key': 'name', 'type': 'str'},\n 'type': {'key': 'type', 'type': 'str'},\n 'location': {'key': 'location', 'type': 'str'},\n 'tags': {'key': 'tags', 'type': '{str}'},\n }\n\n def __init__(self, identity=None, id=None, name=None, type=None, location=None, tags=None):\n self.identity = identity\n self.id = id\n self.name = name\n self.type = type\n self.location = location\n self.tags = tags\n","sub_path":"src/command_modules/azure-cli-lab/azure/cli/command_modules/lab/sdk/devtestlabs/models/service_runner.py","file_name":"service_runner.py","file_ext":"py","file_size_in_byte":1640,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"323559881","text":"from django.shortcuts import render\nfrom django.http import HttpResponseRedirect\nfrom datetime import datetime\nfrom SitterService import models\n\n\ndef retreat_applications(request):\n token = request.COOKIES.get('token')\n user = models.User.objects.filter(token=token).first()\n\n if not user:\n return HttpResponseRedirect('/login')\n\n template = 'base.html'\n sexes = models.Sex.values\n if user.type == models.User.Type.SITTER:\n retreat_applications = list(models.RetreatApplication.objects.filter(sitter=user))\n applications = []\n for retreat_application in retreat_applications:\n applications.append((list(retreat_application.pets.all()), retreat_application, None))\n\n if request.method == 'POST':\n if 'retreatAccept' in request.POST:\n id = request.POST.get('retreatAccept')\n retreat_application = models.RetreatApplication.objects.filter(id=int(id)).first()\n retreat_application.status = models.Status.ACCEPTED\n retreat_application.save(update_fields=['status'])\n return HttpResponseRedirect('/retreat_applications')\n\n if 'retreatReject' in request.POST:\n id = request.POST.get('retreatReject')\n retreat_application = models.RetreatApplication.objects.filter(id=int(id)).first()\n retreat_application.status = models.Status.REJECTED\n retreat_application.save(update_fields=['status'])\n return HttpResponseRedirect('/retreat_applications')\n\n status = models.Status.PROCESSING\n return render(request, 'retreat_applications.html',\n {'retreat_applications': applications, 'sexes': sexes, 'status': status, 'owner': False,\n 'admin': False, 'template': template})\n\n elif user.type == models.User.Type.OWNER:\n retreat_applications = list(models.RetreatApplication.objects.filter(owner=user))\n applications = []\n for retreat_application in retreat_applications:\n applications.append((list(retreat_application.pets.all()), retreat_application,\n models.Profile.objects.filter(user=retreat_application.sitter).first()))\n\n if request.method == 'POST':\n if 'retreatDelete' in request.POST:\n id = request.POST.get('retreatDelete')\n models.RetreatApplication.objects.filter(id=int(id)).delete()\n\n return HttpResponseRedirect('/retreat_applications')\n\n status = models.Status.PROCESSING\n return render(request, 'retreat_applications.html',\n {'retreat_applications': applications, 'sexes': sexes, 'status': status, 'owner': True,\n 'admin': False, 'template': template})\n\n elif user.type == models.User.Type.ADMIN:\n template = 'base_admin.html'\n retreat_applications = list(models.RetreatApplication.objects.filter())\n applications = []\n for retreat_application in retreat_applications:\n applications.append((list(retreat_application.pets.all()), retreat_application,\n models.Profile.objects.filter(user=retreat_application.sitter).first()))\n\n status = models.Status.PROCESSING\n return render(request, 'retreat_applications.html',\n {'retreat_applications': applications, 'sexes': sexes, 'status': status, 'owner': True,\n 'admin': True, 'template': template})\n\n\ndef retreat_application_create(request, sitter_id):\n token = request.COOKIES.get('token')\n user = models.User.objects.filter(token=token).first()\n\n if not user:\n return HttpResponseRedirect('/login')\n\n template = 'base.html'\n\n sitter = models.Profile.objects.filter(id=sitter_id).first()\n petTypes = list(sitter.petTypes.all())\n pets = [pet for pet in list(models.Pet.objects.filter(user=user)) if pet.petType in petTypes]\n application = models.RetreatApplication(sitter=sitter.user, owner=user, status='')\n daysCount = 0\n selectedPets = []\n message = ''\n\n if request.POST:\n if 'refresh' in request.POST or 'createApplication' in request.POST:\n selectedPets = [models.Pet.objects.filter(id=int(pet_id)).first() for pet_id in\n request.POST.getlist('pets')]\n application.dateFrom = request.POST.get('dateFrom')\n application.dateTo = request.POST.get('dateTo')\n application.description = request.POST.get('description')\n daysCount = (datetime.strptime(application.dateTo, '%Y-%m-%d') -\n datetime.strptime(application.dateFrom, '%Y-%m-%d')).days\n if daysCount < 1:\n message = \"Выберите другие даты\"\n else:\n application.totalCost = daysCount * sitter.price * len(selectedPets)\n\n if 'createApplication' in request.POST and message == '':\n application.status = models.Status.PROCESSING\n application.save()\n for pet in selectedPets:\n application.pets.add(pet)\n\n return HttpResponseRedirect('/retreat_applications')\n\n return render(request, 'retreat_application.html',\n {'application': application, 'daysCount': daysCount, 'sitter': sitter, 'pets': pets,\n 'view': False, 'role': sitter, 'selectedPets': selectedPets, 'message': message, 'template':template})\n\n\ndef retreat_application_view(request, app_id):\n token = request.COOKIES.get('token')\n user = models.User.objects.filter(token=token).first()\n\n if not user:\n return HttpResponseRedirect('/login')\n\n template = 'base.html'\n if user.type == models.User.Type.ADMIN:\n template = 'base_admin.html'\n\n sexes = models.Sex.values\n petTypes = list(models.PetType.objects.all())\n\n application = models.RetreatApplication.objects.filter(id=app_id).first()\n sitter = models.Profile.objects.filter(user=application.sitter).first()\n pets = list(application.pets.all())\n\n daysCount = (application.dateTo - application.dateFrom).days\n application.dateFrom = application.dateFrom.strftime('%Y-%m-%d') if application.dateFrom else None\n application.dateTo = application.dateTo.strftime('%Y-%m-%d') if application.dateTo else None\n\n role = models.Profile.objects.filter(user=application.owner).first()\n if user.type == models.User.Type.OWNER:\n role = sitter\n\n return render(request, 'retreat_application.html',\n {'application': application, 'daysCount': daysCount, 'sitter': sitter, 'pets': pets,\n 'sexes': sexes, 'petTypes': petTypes, 'view': True, 'role': role, 'template': template})\n","sub_path":"SitterService/applicationView.py","file_name":"applicationView.py","file_ext":"py","file_size_in_byte":6802,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"220130266","text":"class Node(object):\n def __init__(self, value = None, next = None):\n self.value = value\n self.next = next\n\nclass LinkedList(object):\n def __init__(self, *values):\n\n self.first = None\n self.last = None\n self.length = 0\n self.counter = -1\n\n for i in values:\n self.add(i)\n\n def add(self, value):\n if self.first == None:\n self.first = Node(value, None)\n self.last = self.first\n self.length = 1\n else:\n self.last.next = self.last = Node(value, None)\n self.length += 1\n\n def insert(self, index, value):\n if index == 0:\n self.length += 1\n if self.first == None:\n self.first = Node(value, None)\n self.last = self.first\n else:\n self.first = Node(value, self.first)\n return\n\n if index >= self.length:\n self.add(value)\n else:\n counter = 0\n current = self.first\n while current != None:\n counter += 1\n if counter == index:\n current.next = Node(value, current.next)\n self.length += 1\n if current.next.next == None:\n self.last = current.next\n self.length += 1\n break\n current = current.next\n\n def get(self, index):\n if index >= self.length:\n raise IndexError\n\n counter = -1\n current = self.first\n while current != None:\n counter += 1\n if counter == index:\n return current.value\n current = current.next\n\n def remove(self, value):\n current = self.first\n while current != None:\n\n if current.value == value:\n if current.next == None:\n self.remove_at(self.length - 1)\n return\n\n current.value = current.next.value\n current.next = current.next.next\n self.length -= 1\n return\n current = current.next\n\n def remove_at(self, index):\n if index >= self.length:\n raise IndexError\n current = self.first\n\n if index == 0:\n self.length -= 1\n curr = self.first.value\n self.first = current.next\n return curr\n counter = -1\n\n while current != None:\n counter += 1\n if counter + 1 == index:\n curr = current.next.value\n current.next = current.next.next\n self.length -= 1\n return curr\n curr = curr.next\n\n def clear(self):\n self.__init__()\n\n def contains(self, value):\n current = self.first\n while current != None:\n\n if current.value == value:\n return True\n current = current.next\n return False\n\n def len(self):\n return self.length\n\n def is_empty(self):\n if self.first == None:\n return True\n else:\n return False\n\n def __str__(self):\n if self.first != None:\n current = self.first\n out = 'LinkedList(' + str(current.value) + ', '\n while current.next != None:\n current = current.next\n out += str(current.value) + ', '\n return out.rstrip(', ') + ')'\n return 'LinkedList()'\n\n def __iter__(self):\n return self\n\n def __next__(self):\n self.counter += 1\n if self.counter < self.length:\n\n return self.get(self.counter)\n else:\n raise StopIteration\n\nif __name__ == '__main__':\n ll = LinkedList(1, 2, 3, 4, 5)\n print(ll)\n print(ll.length)\n ll.insert(0, 11)\n print(ll)\n print(ll.length)\n print(ll.is_empty())\n print('\\n')\n for item in ll:\n print(item)\n","sub_path":"task_linked_list.py","file_name":"task_linked_list.py","file_ext":"py","file_size_in_byte":3991,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"487500603","text":"#!/usr/bin/python3\n\"\"\" State Module for HBNB project \"\"\"\nfrom os import getenv\nfrom models.base_model import BaseModel, Base\nfrom sqlalchemy import Column, String, Integer\nfrom sqlalchemy.orm import relationship\nfrom models.city import City\nstorage_type = getenv('HBNB_TYPE_STORAGE')\n\n\nclass State(BaseModel, Base):\n \"\"\" State class \"\"\"\n __tablename__ = 'states'\n if storage_type == \"db\":\n name = Column(String(128), nullable=False)\n cities = relationship('City',\n cascade=\"all, delete\",\n backref='state')\n else:\n @property\n def cities(self):\n from models import storage\n \"\"\" getter attribute cities returns the list of City\n instances with state_id equals to the current State.id\n \"\"\"\n list_cities = []\n all_cities = storage.all(City)\n for city in all_cities.values():\n \"\"\"Iterate through all cities and if the\n city.state_id is eqaul to the current state id\n add it to the list to be returned\"\"\"\n if city.state_id == self.id:\n list_cities.append(city)\n return list_cities\n","sub_path":"models/state.py","file_name":"state.py","file_ext":"py","file_size_in_byte":1243,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"535224658","text":"import sys\nimport argparse as argp\n\nfrom tester.utils.report import Reporter\nfrom tester.utils.runner import run_test\nfrom tester.config import (\n TEST_GROUPS,\n TEST_LEVELS\n)\n\n\ndef main(args):\n\n group = args.group if args.group else TEST_GROUPS[0]\n if group not in TEST_GROUPS:\n Reporter.error('Undefined testing group')\n return\n\n level = args.level if args.level else TEST_LEVELS[0]\n if level not in TEST_LEVELS:\n Reporter.error('Undefined testing level')\n return\n\n run_test(group.lower(), level.lower())\n\n\nif __name__ == '__main__':\n\n parser = argp.ArgumentParser(description=\"Testing client.\")\n parser.add_argument('-g', '--group', type=str, help='testing group.')\n parser.add_argument('-l', '--level', type=str, help='testing level.')\n\n if len(sys.argv[1:]) == 0:\n parser.print_help()\n parser.exit()\n\n main(parser.parse_args())\n","sub_path":"tester/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":910,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"94944569","text":"#!/usr/bin/env python3\n# Copyright (c) 2004-present Facebook All rights reserved.\n# Use of this source code is governed by a BSD-style\n# license that can be found in the LICENSE file.\n\n\nfrom pyinventory.api.equipment import (\n add_equipment,\n get_equipment,\n get_equipment_by_external_id,\n get_equipment_properties,\n get_equipments,\n get_equipments_by_location,\n get_equipments_by_type,\n get_or_create_equipment,\n)\nfrom pyinventory.api.equipment_type import add_equipment_type\nfrom pyinventory.api.location import add_location\nfrom pyinventory.api.location_type import add_location_type\nfrom pyinventory.api.port import edit_port_properties, get_port, get_ports\nfrom pyinventory.api.port_type import add_equipment_port_type\nfrom pyinventory.common.cache import EQUIPMENT_TYPES\nfrom pyinventory.common.data_class import PropertyDefinition\nfrom pyinventory.graphql.enum.property_kind import PropertyKind\nfrom pysymphony import SymphonyClient\n\nfrom ..utils.base_test import BaseTest\nfrom ..utils.grpc.rpc_pb2_grpc import TenantServiceStub\n\n\nclass TestEquipment(BaseTest):\n def __init__(\n self, test_name: str, client: SymphonyClient, stub: TenantServiceStub\n ) -> None:\n super().__init__(test_name, client, stub)\n\n def setUp(self) -> None:\n super().setUp()\n self.port_type1 = add_equipment_port_type(\n self.client,\n name=\"port type 1\",\n properties=[\n PropertyDefinition(\n property_name=\"port property\",\n property_kind=PropertyKind.string,\n default_raw_value=\"port property value\",\n is_fixed=False,\n )\n ],\n link_properties=[\n PropertyDefinition(\n property_name=\"link property\",\n property_kind=PropertyKind.string,\n default_raw_value=\"link property value\",\n is_fixed=False,\n )\n ],\n )\n add_location_type(\n client=self.client,\n name=\"City\",\n properties=[\n PropertyDefinition(\n property_name=\"Mayor\",\n property_kind=PropertyKind.string,\n default_raw_value=None,\n is_fixed=False,\n ),\n PropertyDefinition(\n property_name=\"Contact\",\n property_kind=PropertyKind.email,\n default_raw_value=None,\n is_fixed=False,\n ),\n ],\n )\n add_equipment_type(\n client=self.client,\n name=\"Tp-Link T1600G\",\n category=\"Router\",\n properties=[\n PropertyDefinition(\n property_name=\"IP\",\n property_kind=PropertyKind.string,\n default_raw_value=None,\n is_fixed=False,\n )\n ],\n ports_dict={\"tp_link_port\": \"port type 1\"},\n position_list=[],\n )\n self.location = add_location(\n client=self.client,\n location_hirerchy=[(\"City\", \"Lima\")],\n properties_dict={\"Mayor\": \"Bernard King\", \"Contact\": \"limacity@peru.pe\"},\n lat=10,\n long=20,\n )\n self.equipment = add_equipment(\n client=self.client,\n name=\"TPLinkRouter\",\n equipment_type=\"Tp-Link T1600G\",\n location=self.location,\n properties_dict={\"IP\": \"127.0.0.1\"},\n )\n self.equipment_with_external_id = add_equipment(\n client=self.client,\n name=\"TPLinkRouterExt\",\n equipment_type=\"Tp-Link T1600G\",\n location=self.location,\n properties_dict={\"IP\": \"127.0.0.1\"},\n external_id=\"12345\",\n )\n\n def test_equipment_created(self) -> None:\n\n fetched_equipment = get_equipment(\n client=self.client, name=\"TPLinkRouter\", location=self.location\n )\n self.assertEqual(self.equipment, fetched_equipment)\n\n def test_equipment_with_external_id_created(self) -> None:\n\n fetched_equipment = get_equipment(\n client=self.client, name=\"TPLinkRouterExt\", location=self.location\n )\n self.assertEqual(self.equipment_with_external_id, fetched_equipment)\n\n def test_get_or_create_equipment(self) -> None:\n equipment2 = get_or_create_equipment(\n client=self.client,\n name=\"TPLinkRouter\",\n equipment_type=\"Tp-Link T1600G\",\n location=self.location,\n properties_dict={\"IP\": \"127.0.0.1\"},\n )\n self.assertEqual(self.equipment, equipment2)\n\n def test_get_equipments(self) -> None:\n equipments = get_equipments(client=self.client)\n self.assertEqual(len(list(equipments)), 2)\n\n def test_equipment_properties(self) -> None:\n properties = get_equipment_properties(\n client=self.client, equipment=self.equipment\n )\n self.assertTrue(\"IP\" in properties)\n self.assertEquals(\"127.0.0.1\", properties[\"IP\"])\n\n def test_equipment_get_port(self) -> None:\n fetched_port = get_port(\n client=self.client, equipment=self.equipment, port_name=\"tp_link_port\"\n )\n self.assertEqual(self.port_type1.name, fetched_port.definition.port_type_name)\n\n def test_get_ports(self) -> None:\n ports = list(get_ports(client=self.client))\n self.assertEqual(len(ports), 2)\n\n def test_equipment_edit_port_properties(self) -> None:\n edit_port_properties(\n client=self.client,\n equipment=self.equipment,\n port_name=\"tp_link_port\",\n new_properties={\"port property\": \"test_port_property\"},\n )\n fetched_port = get_port(\n client=self.client, equipment=self.equipment, port_name=\"tp_link_port\"\n )\n port_properties = fetched_port.properties\n self.assertEqual(len(port_properties), 1)\n\n property_type = port_properties[0].propertyType\n self.assertEqual(property_type.name, \"port property\")\n self.assertEqual(port_properties[0].stringValue, \"test_port_property\")\n\n def test_get_equipments_by_type(self) -> None:\n equipment_type_id = EQUIPMENT_TYPES[\"Tp-Link T1600G\"].id\n equipments = list(\n get_equipments_by_type(\n client=self.client, equipment_type_id=equipment_type_id\n )\n )\n self.assertEqual(len(equipments), 2)\n self.assertEqual(equipments[0].name, \"TPLinkRouter\")\n\n def test_get_equipments_by_location(self) -> None:\n equipments = list(\n get_equipments_by_location(client=self.client, location_id=self.location.id)\n )\n self.assertEqual(len(equipments), 2)\n self.assertEqual(equipments[0].name, \"TPLinkRouter\")\n\n def test_get_equipment_by_external_id(self) -> None:\n equipment = get_equipment_by_external_id(\n client=self.client, external_id=\"12345\"\n )\n self.assertEqual(self.equipment_with_external_id, equipment)\n","sub_path":"symphony/cli/tests/pyinventory_tests/test_equipment.py","file_name":"test_equipment.py","file_ext":"py","file_size_in_byte":7202,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"79419401","text":"# -*- coding: utf-8 -*-\n\nimport pygame\nimport os, sys\n\n# My Modules\n# from bin.modules.globals import Loading\nfrom bin.modules.globals import Manager\n\n\n# Returned folder with 'step' move back\ndef backdir(path, step):\n\tstep = abs(step) * -1\n\treturn '\\\\'.join(path.split('\\\\')[:step])\n\n\nclass MainMenu(object):\n\tdef __init__(self, master, mouse):\n\t\tself.master = master\t\t\t\t\t\t\t# Parent window\n\t\tself.mouse = mouse\t\t\t\t\t\t\t\t# Cursor module\n\t\tself.widget_list = []\t\t\t\t\t\t\t# List with widget names\n\t\tself.widgets = {}\t\t\t\t\t\t\t\t# {name : image-object}\n\n\t\t''' Text font '''\n\t\tself.menu_font = pygame.font.Font(\n\t\t\tbackdir(os.path.dirname(__file__), 2) + '\\\\fonts\\\\times.ttf',\n\t\t\t20\n\t\t)\n\t\tw, h = pygame.display.get_surface().get_size()\t# Parent window sizes\n\n\t\t''' Widgets sizes '''\n\t\tself.size = {\n\t\t\t'bg' : (int(w), int(h)),\n\t\t\t'logo' : (int(w*.6), int(h*.15)),\n\t\t\t'button' : (int(w*.6), int(h*.2))\n\t\t}\n\n\t\t''' Widgets positions '''\n\t\tself.position = {\n\t\t\t'bg' : (0, 0),\n\t\t\t'logo' : (w*.2, h*.01),\n\t\t\t'button' : [\n\t\t\t\t(int(w*.2), int(h*.17)),\n\t\t\t\t(int(w*.2), int(h*.38)),\n\t\t\t\t(int(w*.2), int(h*.59)),\n\t\t\t\t(int(w*.2), int(h*.80))\n\t\t\t]\n\t\t}\n\n\t\tself.initUI()\n\n\n\tdef initUI(self):\n\t\t''' Background image '''\n\t\tbg = pygame.image.load(os.path.dirname(__file__) + '\\\\images\\\\1-bg.jpg')\n\t\tself.widgets.setdefault('bg', pygame.transform.scale(bg, self.size['bg']))\n\t\tself.widget_list.append('bg')\n\n\t\t''' Logo text '''\n\t\tlogo = self.menu_font.render('My Garden', False, (255, 255, 255))\n\t\tself.widgets.setdefault('logo', pygame.transform.scale(logo, self.size['logo']))\n\t\tself.widget_list.append('logo')\n\n\t\t''' Button image '''\n\t\tbutton = pygame.image.load(os.path.dirname(__file__) + '\\\\images\\\\2-button.png')\n\t\tself.widgets.setdefault('button', pygame.transform.scale(button, self.size['button']))\n\t\tself.widget_list.append('button')\n\t\t# Light version\n\t\tself.light_button = pygame.image.load(os.path.dirname(__file__) + '\\\\images\\\\2-button_light.png')\n\t\tself.light_button = pygame.transform.scale(self.light_button, self.size['button'])\n\n\n\t# Show interface\n\tdef run(self):\n\t\tfor widget in self.widget_list:\n\t\t\t''' If for widget 1 position: '''\n\t\t\tif not isinstance(self.position[widget], list):\n\t\t\t\t''' Show widget on him position '''\n\t\t\t\tself.master.blit(self.widgets[widget], self.position[widget])\n\t\t\telse:\n\t\t\t\t''' Get all widget positions '''\n\t\t\t\tfor num, pos in enumerate(self.position[widget]):\n\t\t\t\t\ttexts = ('New game', 'Load game', 'Options', 'Exit')\n\t\t\t\t\t'''\n\t\t\t\t\tCheck all position with cursor.position.\n\t\t\t\t\tIf cursor.position on widget:\n\t\t\t\t\t\tDraw ligth button\n\t\t\t\t\tElse:\n\t\t\t\t\t\tDraw default button\n\t\t\t\t\t'''\n\t\t\t\t\tif self.check_guidance(self.mouse.x, self.mouse.y, widget, num):\n\t\t\t\t\t\tself.master.blit(self.light_button, pos)\n\t\t\t\t\telse:\n\t\t\t\t\t\tself.master.blit(self.widgets[widget], pos)\n\n\t\t\t\t\t''' Create text widget '''\n\t\t\t\t\ttemp_text = self.menu_font.render(texts[num], False, (255, 255, 255))\n\t\t\t\t\t''' Draw text '''\n\t\t\t\t\tself.master.blit(temp_text, (\n\t\t\t\t\t\tpos[0] + self.size[widget][0]/2 - self.menu_font.size(texts[num])[0]/2,\n\t\t\t\t\t\tpos[1] + self.size[widget][1]/2 - self.menu_font.size(texts[num])[1]/2)\n\t\t\t\t\t)\n\n\n\tdef check_guidance(self, x, y, widget, num):\n\t\t'''\n\t\tx, y = cursor.position\n\t\twidget = widget.name\n\t\tnum = index of internal position entry\n\t\t'''\n\t\tbx, by = self.position[widget][num]\n\t\tfor _ in self.position[widget]:\n\t\t\tif x >= bx and x <= bx + self.size[widget][0] and \\\n\t\t\t\ty >= by and y <= by + self.size[widget][1]:\n\t\t\t\treturn True\n\t\treturn False\n\n\n\t# Check on press LeftMouseButton\n\tdef mouse_press(self, pos):\n\t\t'''\n\t\tpos = cursor.position\n\t\t'''\n\t\tfor num, btn_pos in enumerate(self.position['button']):\n\t\t\tif self.check_guidance(*pos, 'button', num):\n\t\t\t\tif num == 0:\t\t\t\t\t\t\t\t\t# New game\n\t\t\t\t\treturn Manager.STATUS.set_value(1)\n\t\t\t\telif num == 1:\t\t\t\t\t\t\t\t\t# Load game\n\t\t\t\t\tprint('You load old game')\n\t\t\t\telif num == 2:\t\t\t\t\t\t\t\t\t# Options\n\t\t\t\t\tprint('You open options')\n\t\t\t\telif num == 3:\t\t\t\t\t\t\t\t\t# Exit\n\t\t\t\t\tsys.exit()","sub_path":"bin/modules/menu/MainMenu.py","file_name":"MainMenu.py","file_ext":"py","file_size_in_byte":3930,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"545343434","text":"# -*- coding: utf-8 -*-\n\"\"\"\n Description: 使用__new__方法实现单例模式\n\"\"\"\n\n\nclass Singleton(object):\n\n def __new__(cls, *args, **kwargs):\n if not hasattr(cls, '_instance'):\n orig = super(Singleton, cls)\n cls._instance = orig.__new__(cls, *args, **kwargs)\n return cls._instance\n\n\nobj = Singleton()\nobj1 = Singleton()\nprint(obj, obj1)\n","sub_path":"DesignPattern/code/cjl_singleton/newmethod.py","file_name":"newmethod.py","file_ext":"py","file_size_in_byte":387,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"550930762","text":"#!/usr/bin/env python\n# -*- encoding:utf-8 -*-\n\n# Author: lixingtie\n# Email: lixingtie@barfoo.com.cn\n# Create Date: 2013-9-10\n\nfrom bson import ObjectId, DBRef\nfrom framework.util.parser import parse\nfrom framework.data.mongo import db\nfrom framework.data import metadb as mdb\n\ndef remove_object(id):\n \"\"\"\n 删除对象\n \"\"\"\n #删除对象属性\n db.property.remove({ \"object.$id\" : ObjectId(id) })\n\n\ndef get_parent_relations_data(data):\n\n def get_parent_data(item):\n __parent_data__ = dict()\n if \"_relation_\" not in item or not item[\"_relation_\"]:return __parent_data__\n relation = item[\"_relation_\"]\n\n for r in relation:\n objectname = relation[r].collection\n id = relation[r].id\n value = db[objectname].find_one({\"_id\":id})\n if not value:continue\n p = get_parent_data(value)\n if p:\n value[\"__parent_data__\"] = p\n\n __parent_data__[r] = value\n\n return __parent_data__\n\n\n if not data:return data\n for i in range(0,len(data)):\n parent = get_parent_data(data[i])\n\n if parent:\n data[i][\"__parent_data__\"] = parent\n\n return data\n\n\ndef get_object_propertys(objectname,systemid):\n \"\"\"\n 获取对象属性\n \"\"\"\n propertys = dict()\n obj = get_object(objectname,systemid)\n if obj:\n with mdb.property.find({\"object\" : DBRef(\"object\", obj._id)}) as cursor:\n for prop in cursor:\n propertys[prop[\"name\"]] = prop\n\n return propertys\n\n\ndef get_package_object_propertys(objectname,systemid):\n \"\"\"\n 获取对象属性 及其关系属性\n \"\"\"\n result = get_object_propertys(objectname,systemid)\n\n name = get_object_parent_relations(objectname,systemid)\n if name:\n for i in name:\n if \"__parent_propertys__\" not in result:\n result[\"__parent_propertys__\"] ={i: get_package_object_propertys(i,systemid)}\n\n else:\n result[\"__parent_propertys__\"][i] = get_package_object_propertys(i,systemid)\n\n return result\n\n\n\ndef get_object(objectname,systemid):\n \"\"\"\n 获取object\n \"\"\"\n obj = mdb.object.find_one({\"system\" : DBRef(\"system\", ObjectId(systemid)),\"name\":objectname})\n return obj\n\n\ndef get_object_info(objectname,systemid):\n \"\"\"\n 获取object (公开部分信息)\n \"\"\"\n obj = mdb.object.find_one({\"system\" : DBRef(\"system\", ObjectId(systemid)),\"name\":objectname},{\"_id\":1,\"name\":1,\"alias\":1})\n\n return obj if obj else {}\n\n\ndef get_object_parent_relations(objectname,systemid):\n \"\"\"\n 获取父级对象objectname\n \"\"\"\n autodesign = get_autodesign(systemid)\n links = autodesign.get(\"design\").get(\"relation\").get(\"main\") if autodesign else {}\n objs = autodesign.get(\"design\").get(\"relation\").get(\"objs\") if autodesign else {}\n objid = get_id(objs,objectname)\n parentid = list()\n for i in objid:\n node = find_parent_relation(links,i)\n if node:\n parentid += node\n\n name = list()\n for i in parentid:\n obj = objs[i]\n if obj[\"name\"] not in name:\n name.append(obj[\"name\"])\n\n return name\n\n\ndef get_object_relations(objectname,systemid,isObjectname = False):\n \"\"\"\n 获取对象关系\n \"\"\"\n autodesign = get_autodesign(systemid)\n links = autodesign.get(\"design\").get(\"relation\").get(\"main\") if autodesign else {}\n objs = autodesign.get(\"design\").get(\"relation\").get(\"objs\") if autodesign else {}\n objid = get_id(objs,objectname)\n relations = list()\n\n for i in objid:\n data = find_relation(links,i)\n for n in data.get(\"nodes\"):\n name = objs[n.get(\"id\")].get(\"name\")\n\n if isObjectname:\n relations.append(name)\n continue\n\n alias = objs[n.get(\"id\")].get(\"alias\")\n propertys = get_package_object_propertys(name,systemid)\n relation = n.get(\"relation\")\n pagename = \"__\"+name+\"_item__\" if relation == \"one\" else \"__\"+name+\"_list__\" if relation == \"many\" else \"\"\n page = None\n if pagename:\n page = mdb.page.find_one({\"name\" : pagename.lower(),\"system\" : DBRef(\"system\", ObjectId(systemid))})\n relations.append({\"objectname\":name,\"objectalias\":alias,\"relation\":relation,\"pageid\":str(page._id) if page else \"\",\"propertys\":propertys})\n\n return relations\n\n\ndef get_autodesign(systemid):\n \"\"\"\n 获取自动设计数据\n \"\"\"\n system = mdb.system.find_one({\"_id\":ObjectId(systemid)})\n autodesign = system.get(\"autodesign\") if system else {}\n return autodesign\n\n\ndef get_relation(systemid):\n autodesign = get_autodesign(systemid)\n links = autodesign.get(\"design\").get(\"relation\").get(\"main\") if autodesign else {}\n return links\n\n\ndef get_objs(systemid):\n autodesign = get_autodesign(systemid)\n objs = autodesign.get(\"design\").get(\"relation\").get(\"objs\") if autodesign else {}\n return objs\n\n\ndef get_id(objs,name,identity = None):\n ids = list()\n key = \"identity\" if identity else \"name\"\n value = identity if identity else name\n\n for k,v in objs.items():\n if v.get(\"type\") == \"object\" and v.get(key) == value:\n ids.append(k)\n\n return ids\n\n\ndef find_relation(relations,id):\n \"\"\"\n 根据id寻找关系数据\n \"\"\"\n if len(relations) <= 0: return None\n _type = type(relations).__name__\n\n if _type in [\"dict\",\"Document\"]:\n return (relations if id == relations.get(\"id\") else find_relation(relations.get(\"nodes\"),id))\n\n elif _type == \"list\":\n for i in range(0,len(relations)):\n result = find_relation(relations[i],id)\n if result: return result\n\n\ndef find_parent_relation(jsonlist,value,parentid = None):\n \"\"\"\n 获取父节点\n \"\"\"\n if len(jsonlist) <= 0: return None\n _type = type(jsonlist).__name__\n\n if _type in [\"dict\",\"Document\"]:\n if value != jsonlist.get(\"id\"):\n return find_parent_relation(jsonlist.get(\"nodes\"),value,jsonlist.get(\"id\"))\n else:\n return str(parentid)\n\n elif _type == \"list\":\n keys = list()\n for i in range(0,len(jsonlist)):\n result = find_parent_relation(jsonlist[i],value,parentid)\n if result:\n result_type = type(result).__name__\n if result_type == \"str\" and result != \"None\":\n keys.append(result)\n elif result_type == \"list\":\n keys += result\n\n return keys\n\n\ndef data2object(data, uid=None):\n\n uid = uid or data.get('_id')\n\n if not uid:\n return data\n\n with mdb.property.find({ \"object.$id\" : ObjectId(uid) }) as cursor:\n for p in cursor:\n\n if p in ('file', 'image'):\n continue\n\n if p.name in data:\n data[p.name] = parse(data[p.name], p.datatype)\n\n return data\n","sub_path":"core/logic/object.py","file_name":"object.py","file_ext":"py","file_size_in_byte":6932,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"545267995","text":"import matplotlib.pyplot as plt\nimport gym\nfrom Working.Cartpole.pytorchdqn.ai import Agent\nimport numpy as np\n\nENVNAME = 'CartPole-v0'\n\nclass Environment:\n\n def __init__(self, name):\n self.env = gym.make(name)\n self.envName = name\n\n\nrun = Environment(ENVNAME)\n\nactionCnt = run.env.action_space.n\nstateCnt = run.env.observation_space.shape[0]\nagent = Agent(stateCnt, actionCnt, 0.99)\n\n\nEPISODES = 1000\nSTEPS = 400\nres = []\n\ndef start():\n\n for n in range(EPISODES):\n\n s = run.env.reset()\n R = 0\n\n for i in range(STEPS):\n\n if n > 900:\n run.env.render()\n a = agent.select_action(s)\n _s, r, done, info = run.env.step(a)\n\n if done:\n _s = np.zeros(stateCnt)\n\n agent.update(s, _s, r, a, done)\n agent.train()\n\n R += r\n s = _s\n\n if done:\n break\n\n res.append(R)\n\n print(f\"Total Reward for episode {n}: {R}\")\n print('*************Training Done*******************')\n\ntry:\n start()\nfinally:\n plt.plot(res)\n plt.show()","sub_path":"Working/Cartpole/pytorchdqn/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1116,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"114683085","text":"useFixture(default)\r\n\r\ndef test():\r\n\tfrom Modules import commonBits\r\n\tjava_recorded_version = '1.6.0_22'\r\n\r\n\tif window(commonBits.applicationName()):\r\n\t\tselect('FileNameTxtFld', commonBits.sampleDir() + 'protoSales.bin')\r\n\t\tclick('Proto Search')\r\n\t\tkeystroke('EditorPane', 'Context Menu')\r\n\t\tselect('Table1', r'cell:Proto Filename,2(' + commonBits.stdCopybookDir() + 'Sales.proto)')\r\n\t\tkeystroke('Table1', 'Context Menu', 'Proto Filename,2')\r\n\t\tselect('Table1', r'cell:Proto Filename,2(' + commonBits.stdCopybookDir() + 'Sales.proto)')\r\n\t\tkeystroke('Table1', 'Context Menu', 'Proto Filename,2')\r\n\t\tselect('Table1', r'cell:Proto Filename,2(' + commonBits.stdCopybookDir() + 'Sales.proto)')\r\n\t\tkeystroke('Table1', 'Context Menu', 'Proto Filename,2')\r\n\t\tselect('Table1', r'cell:Proto Filename,2(' + commonBits.stdCopybookDir() + 'Sales.proto)')\r\n\t\tkeystroke('Table1', 'Context Menu', 'Proto Filename,2')\r\n\t\tselect('Table1', r'cell:Proto Filename,2(' + commonBits.stdCopybookDir() + 'Sales.proto)')\r\n\t\tassert_p('Table1', 'Content', r'[[' + commonBits.stdCopybookDir() + 'DTAR020.proto, DTAR020.proto, sale], [' + commonBits.stdCopybookDir() + 'DTAR020.protocomp, DTAR020.proto, sale], [' + commonBits.stdCopybookDir() + 'Sales.proto, Sales.proto, sale], [' + commonBits.stdCopybookDir() + 'sales.protocomp, sales.proto, sale]]')\r\n\t\tclick('BasicInternalFrameTitlePane$NoFocusButton2')\r\n\t\tassert_p('FileNameTxtFld1', 'Text', commonBits.stdCopybookDir() + 'sales.protocomp')\r\n\t\tassert_p('ComboBox2', 'Text', 'sales.proto')\r\n\t\tassert_p('ComboBox3', 'Text', 'sale')\r\n\t\tassert_p('ComboBox', 'Text', 'Delimited Messages')\r\n\t\tassert_p('ComboBox1', 'Text', 'Compiled Proto')\r\n##\t\tassert_p('ComboBox1', 'Text', 'Proto Definition')\r\n\t\tassert_p('ComboBox2', 'Text', 'sales.proto')\r\n\t\tassert_p('ComboBox2', 'Content', '[[sales.proto]]')\r\n\t\tassert_p('ComboBox3', 'Text', 'sale')\r\n\t\tassert_p('ComboBox3', 'Content', '[[sale]]')\r\n\tclose()\r\n","sub_path":"Build/Instalation/ProtoBuf/MarathonTests/Marathon 1.1/ProtoBufEditor/TestCases/ProtoSearch/ProtoSearch2.py","file_name":"ProtoSearch2.py","file_ext":"py","file_size_in_byte":1920,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"528072800","text":"#Universal Encoder is required in the parent folder before running the scraper\r\nimport scrapy\r\nfrom ..items import QuestionListItem\r\nfrom datetime import datetime\r\nfrom elasticsearch import Elasticsearch\r\nimport tensorflow as tf\r\nimport tensorflow_hub as hub\r\n\r\n#Connecting to elasticsearch\r\nes = Elasticsearch([{'host': 'localhost', 'port': 9200}])\r\nif es.ping():\r\n print('Connected to ES!')\r\nelse:\r\n print('Could not connect!')\r\n exit()\r\n\r\n#defining mapping\r\nstructure = {\r\n \"settings\": {\r\n \"analysis\": {\r\n \"filter\": {\r\n \"autocomplete_filter\": {\r\n \"type\": \"edge_ngram\",\r\n \"min_gram\": 1,\r\n \"max_gram\": 5\r\n }\r\n },\r\n \"analyzer\": {\r\n \"autocomplete_analyzer\": {\r\n \"type\": \"custom\",\r\n \"tokenizer\": \"standard\",\r\n \"filter\": [ \"lowercase\", \"autocomplete_filter\"]\r\n }\r\n }\r\n }\r\n }, \r\n \"mappings\": {\r\n \"properties\": {\r\n \"answers\": {\r\n \"type\": \"text\",\r\n \"fields\": {\r\n \"keyword\": {\r\n \"type\": \"keyword\",\r\n \"ignore_above\": 256\r\n }\r\n }\r\n },\r\n \"details\": {\r\n \"type\": \"text\",\r\n \"fields\": {\r\n \"keyword\": {\r\n \"type\": \"keyword\",\r\n \"ignore_above\": 256\r\n }\r\n }\r\n },\r\n \"question\": {\r\n \"type\": \"text\",\r\n \"fields\": {\r\n \"keyword\": {\r\n \"type\": \"keyword\",\r\n \"ignore_above\": 256\r\n }\r\n },\r\n \"analyzer\": \"autocomplete_analyzer\"\r\n },\r\n \"tags\": {\r\n \"type\": \"text\",\r\n \"fields\": {\r\n \"keyword\": {\r\n \"type\": \"keyword\",\r\n \"ignore_above\": 256\r\n }\r\n }\r\n },\r\n \"upvotes\": {\r\n \"type\": \"text\",\r\n \"fields\": {\r\n \"keyword\": {\r\n \"type\": \"keyword\",\r\n \"ignore_above\": 256\r\n }\r\n }\r\n },\r\n \"total_vectors\": {\r\n \"type\": \"dense_vector\",\r\n \"dims\": 512\r\n }\r\n }\r\n\r\n }\r\n}\r\n\r\n#Creating an index using the above defined mapping\r\nres = es.indices.create(index='ie-3', ignore=400, body=structure)\r\n\r\n#Loading the encoder model for generating text embeddings\r\nembed = hub.load('universal_encoder')\r\n\r\n#Function to create text embeddings\r\ndef make_vector(query):\r\n embeddings = embed([query])\r\n vector = []\r\n for i in embeddings[0]:\r\n vector.append(float(i))\r\n return vector\r\n\r\n#Main scraper code\r\nclass stackoverflow(scrapy.Spider):\r\n i = 0\r\n page_no = 20001\r\n name = 'stackoverflow'\r\n start_urls = [\r\n \"https://stackoverflow.com/questions?sort=MostVotes&edited=true&page={}\".format(page_no)\r\n ]\r\n\r\n def parse(self, response):\r\n base_url = 'https://stackoverflow.com'\r\n que_set = response.css('div.question-summary')\r\n for q in que_set:\r\n self.i += 1\r\n link = q.css('h3 a::attr(href)').get()\r\n yield response.follow(url=base_url + link, callback=self.parse_question)\r\n if self.page_no<40001:\r\n self.page_no += 1\r\n yield scrapy.Request(url=\"https://stackoverflow.com/questions?sort=MostVotes&edited=true&page={}\".format(self.page_no), callback=self.parse)\r\n\r\n def parse_question(self, response):\r\n items = QuestionListItem()\r\n\r\n data = response.css('div.inner-content')\r\n items['question'] = data.css(\r\n \"div h1 a.question-hyperlink::text\").extract_first()\r\n details = data.css(\"div.question div.s-prose\").extract()\r\n answers_raw = data.css('#answers')\r\n \r\n tags = response.css('a.post-tag::text').extract()\r\n tags_set = set(tags)\r\n tags = list(tags_set)\r\n upvotes = response.css(\"div.js-vote-count::text\").extract_first()\r\n \r\n if answers_raw:\r\n answers = ''\r\n raw_data = answers_raw.css(\"div.s-prose\").extract_first()\r\n for row in raw_data:\r\n answers = answers + row \r\n\r\n concat_details = ''\r\n \r\n for detail in details:\r\n concat_details = concat_details + detail\r\n \r\n items['details'] = concat_details\r\n items['answers'] = answers\r\n items['upvotes'] = upvotes\r\n items['tags'] = tags\r\n items['total_vectors'] = make_vector(items['question'])\r\n doc = {\r\n 'question': items['question'],\r\n 'details': items['details'],\r\n 'answers': items['answers'],\r\n 'tags': items['tags'],\r\n 'upvotes':items['upvotes'],\r\n 'total_vectors': items['total_vectors']\r\n }\r\n \r\n #Inserting the document into our ElasticSearch Index \r\n # res = es.index(index=\"ie-3\", body=doc)\r\n # print(res['result'])\r\n\r\n yield items\r\n\r\n\r\n\r\n","sub_path":"wsa_scraper/wsa_scraper/spiders/stackoverflow.py","file_name":"stackoverflow.py","file_ext":"py","file_size_in_byte":5397,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"158327235","text":"import numpy as np\nfrom scipy.optimize import minimize\nfrom scipy.io import loadmat\nfrom numpy.linalg import det, inv\nfrom math import sqrt, pi\nimport scipy.io\nimport matplotlib.pyplot as plt\nimport pickle\nimport sys\n\n\ndef ldaLearn(X,y):\n # Inputs\n # X - a N x d matrix with each row corresponding to a training example\n # y - a N x 1 column vector indicating the labels for each training example\n #\n # Outputs\n # means - A d x k matrix containing learnt means for each of the k classes\n # covmat - A single d x d learnt covariance matrix \n \n #Initialize empty means vector\n means = np.empty((2,0))\n\n for i in range(len(np.unique(y))):\n # loop through each unique val of y\n yi=(np.unique(y)[i])\n \n #filter x for indices where y = yi and calc mean\n a=np.mean(X[y[:,0]==yi],0)\n \n #append mean to mean vector\n means = np.c_[means, a]\n \n #calc covariance\n covmat = np.cov(X,rowvar=0,bias=1)\n \n\n \n # IMPLEMENT THIS METHOD \n return means,covmat\n\ndef qdaLearn(X,y):\n # Inputs\n # X - a N x d matrix with each row corresponding to a training example\n # y - a N x 1 column vector indicating the labels for each training example\n #\n # Outputs\n # means - A d x k matrix containing learnt means for each of the k classes\n # covmats - A list of k d x d learnt covariance matrices for each of the k classes\n \n #same as lda learn except calc cov in for loop and append to list\n means = np.empty((2,0))\n covmats = []\n\n for i in range(len(np.unique(y))):\n yi=(np.unique(y)[i])\n a=np.mean(X[y[:,0]==yi],0)\n means = np.c_[means, a]\n \n cov = np.cov(X[y[:,0]==yi],rowvar=0,bias=1)\n\n covmats.append(cov)\n \n # IMPLEMENT THIS METHOD\n return means,covmats\n\ndef ldaTest(means,covmat,Xtest,ytest):\n # Inputs\n # means, covmat - parameters of the LDA model\n # Xtest - a N x d matrix with each row corresponding to a test example\n # ytest - a N x 1 column vector indicating the labels for each test example\n # Outputs\n # acc - A scalar accuracy value\n # ypred - N x 1 column vector indicating the predicted labels\n cov = covmat\n def sqdist (x,mean,cov):\n diff = x - np.transpose(mean)\n inv_Sigma = np.linalg.inv(cov)\n sqdist1 = np.dot(np.dot(np.transpose(diff),inv_Sigma),diff)\n return sqdist1\n \n \n p = np.zeros((len(Xtest),len(means[0])))\n \n for i in range(len(Xtest),):\n x = Xtest[i]\n for j in range(len(means[0])):\n mean = means[:,j]\n sqdistij = sqdist(x,mean,cov)\n p[i,j] = np.exp(-sqdistij)\n pred = np.argmax(p, axis=1)+1\n \n acc = sum(pred==ytest[:,0])/len(ytest)\n ypred = pred\n\n # IMPLEMENT THIS METHOD\n return acc,ypred\n\ndef qdaTest(means,covmats,Xtest,ytest):\n # Inputs\n # means, covmats - parameters of the QDA model\n # Xtest - a N x d matrix with each row corresponding to a test example\n # ytest - a N x 1 column vector indicating the labels for each test example\n # Outputs\n # acc - A scalar accuracy value\n # ypred - N x 1 column vector indicating the predicted labels\n\n # function to calculate mah square distance\n def sqdist (x,mean,cov):\n diff = x - np.transpose(mean)\n diffT =np.transpose(diff)\n inv_Sigma = np.linalg.inv(cov)\n sqdist1 = np.dot(np.dot(diffT,inv_Sigma),diff)\n return sqdist1\n\n\n p = np.zeros((len(Xtest),len(means[0])))\n \n for i in range(len(Xtest),):\n xi = Xtest[i]\n for j in range(len(means[0])):\n covj = covmats[j]\n meanj = means[:,j]\n sqdistij = sqdist(xi,meanj,covj)\n p[i,j] = np.exp(-sqdistij/2)/sqrt(np.linalg.det(covj))\n pred = np.argmax(p, axis=1)+1\n \n acc = sum(pred==ytest[:,0])/len(ytest)\n ypred = pred\n # IMPLEMENT THIS METHOD\n return acc,ypred\n\ndef learnOLERegression(X,y):\n # Inputs: \n # X = N x d \n # y = N x 1 \n # Output: \n # w = d x 1 \n\n # IMPLEMENT THIS METHOD \n w=np.dot(inv(np.dot(np.transpose(X),X)),np.dot(np.transpose(X),y))\n \n return w\n\ndef learnRidgeRegression(X,y,lambd):\n # Inputs:\n # X = N x d \n # y = N x 1 \n # lambd = ridge parameter (scalar)\n # Output: \n # w = d x 1 \n d=int(np.size(X,1)) \n w=np.dot(inv(lambd*np.identity(d)+np.dot(np.transpose(X),X)),np.dot(np.transpose(X),y))\n # IMPLEMENT THIS METHOD \n return w\n\ndef testOLERegression(w,Xtest,ytest):\n # Inputs:\n # w = d x 1\n # Xtest = N x d\n # ytest = X x 1\n # Output:\n # mse\n N=len(Xtest)\n #mse=(np.dot(np.transpose(ytest-np.dot(Xtest,w)),(ytest-np.dot(Xtest,w))))/N\n mse = (np.sum((ytest-np.dot(Xtest,w))**2))/N\n # IMPLEMENT THIS METHOD\n return mse\n\ndef regressionObjVal(w, X, y, lambd):\n\n # compute squared error (scalar) and gradient of squared error with respect\n # to w (vector) for the given data X and y and the regularization parameter\n # lambda \n w1=w.reshape((len(X[1]),1))\n \n error=np.dot(np.transpose(y-np.dot(X,w1)),(y-np.dot(X,w1)))/2 + (lambd*(np.dot(np.transpose(w1),w1)))/2\n \n obj_grad = np.sum((np.dot(X_i,w1) -y)*X_i ,0 ) +np.array(lambd*w).flatten() \n #obj_grad = np.sum(np.matrix(np.dot(np.transpose(X_i),np.sum(np.dot(X_i,w1)-y)))+ np.matrix(lambd*w1),1)\n error_grad = np.array(obj_grad).flatten()\n # IMPLEMENT THIS METHOD \n return error, error_grad\n\ndef mapNonLinear(x,p):\n # Inputs: \n # x - a single column vector (N x 1) \n # p - integer (>= 0) \n # Outputs: \n # Xd - (N x (d+1)) \n r=len(x)\n c=p+1\n Xd=np.zeros((r,c)) \n for i in range(0,c):\n Xd[:,i]=x**i\n# IMPLEMENT THIS METHOD\n return Xd\n\n# Main script\n\n# Problem 1\n# load the sample data \nif sys.version_info.major == 2:\n X,y,Xtest,ytest = pickle.load(open('sample.pickle','rb'))\nelse:\n X,y,Xtest,ytest = pickle.load(open('sample.pickle','rb'),encoding = 'latin1')\n\n# LDA\nmeans,covmat = ldaLearn(X,y)\nldaacc,ldares = ldaTest(means,covmat,Xtest,ytest)\nprint('LDA Accuracy = '+str(ldaacc))\n# QDA\nmeans,covmats = qdaLearn(X,y)\nqdaacc,qdares = qdaTest(means,covmats,Xtest,ytest)\nprint('QDA Accuracy = '+str(qdaacc))\n\n# plotting boundaries\nx1 = np.linspace(-5,20,100)\nx2 = np.linspace(-5,20,100)\nxx1,xx2 = np.meshgrid(x1,x2)\nxx = np.zeros((x1.shape[0]*x2.shape[0],2))\nxx[:,0] = xx1.ravel()\nxx[:,1] = xx2.ravel()\n\nfig = plt.figure(figsize=[12,6])\nplt.subplot(1, 2, 1)\n\nzacc,zldares = ldaTest(means,covmat,xx,np.zeros((xx.shape[0],1)))\nplt.contourf(x1,x2,zldares.reshape((x1.shape[0],x2.shape[0])),alpha=0.3)\nplt.scatter(Xtest[:,0],Xtest[:,1],c=ytest)\nplt.title('LDA')\nplt.xlabel('attribute 1')\nplt.ylabel('attribute 2')\nplt.subplot(1, 2, 2)\n\nzacc,zqdares = qdaTest(means,covmats,xx,np.zeros((xx.shape[0],1)))\nplt.contourf(x1,x2,zqdares.reshape((x1.shape[0],x2.shape[0])),alpha=0.3)\nplt.scatter(Xtest[:,0],Xtest[:,1],c=ytest)\nplt.title('QDA')\nplt.xlabel('attribute 1')\nplt.ylabel('attribute 2')\n\nplt.show()\n# Problem 2\nif sys.version_info.major == 2:\n X,y,Xtest,ytest = pickle.load(open('diabetes.pickle','rb'))\nelse:\n X,y,Xtest,ytest = pickle.load(open('diabetes.pickle','rb'),encoding = 'latin1')\n\n# add intercept\nX_i = np.concatenate((np.ones((X.shape[0],1)), X), axis=1)\nXtest_i = np.concatenate((np.ones((Xtest.shape[0],1)), Xtest), axis=1)\n\nw = learnOLERegression(X,y)\nmle_tr = testOLERegression(w,X,y) # training mse without intercept\nmle = testOLERegression(w,Xtest,ytest) # test mse without intercept\n\nw_i = learnOLERegression(X_i,y)\nmle_i_tr = testOLERegression(w_i,X_i,y) # training mse with intercept\nmle_i = testOLERegression(w_i,Xtest_i,ytest) # test mse with intercept\n\n\nprint('MSE without intercept on testing data = '+str(mle))# test mse without intercept\nprint('MSE with intercept on testing data = '+str(mle_i)) # test mse with intercept\nprint('MSE without intercept on training data = '+str(mle_tr)) # training mse without intercept\nprint('MSE with intercept on training data = '+str(mle_i_tr)) # training mse with intercept\n\n# Problem 3\nk = 101\nlambdas = np.linspace(0, 1, num=k)\ni = 0\nmses3_train = np.zeros((k,1))\nmses3 = np.zeros((k,1))\nfor lambd in lambdas:\n w_l = learnRidgeRegression(X_i,y,lambd)\n mses3_train[i] = testOLERegression(w_l,X_i,y)\n mses3[i] = testOLERegression(w_l,Xtest_i,ytest)\n i = i + 1\nfig = plt.figure(figsize=[12,6])\nplt.subplot(1, 2, 1)\nplt.plot(lambdas,mses3_train)\nplt.title('MSE for Train Data')\nplt.xlabel('$\\lambda$')\nplt.ylabel('Mean Squared Error')\nplt.subplot(1, 2, 2)\nplt.plot(lambdas,mses3)\nplt.title('MSE for Test Data')\nlambda_opt_3 = lambdas[np.argmin(mses3)]\nprint('optimal value for lambda using Ridge Regression '+str(lambda_opt_3))\n\nplt.xlabel('$\\lambda$')\nplt.ylabel('Mean Squared Error')\n\nplt.show()\nprint(' OLE without intercept='+str(np.linalg.norm(w))) # OLE without intercept\nprint('OLE with intercept='+str(np.linalg.norm(w_i))) # OLE with intercept\nprint('ridge regression='+str(np.linalg.norm(w_l))) # ridge regression \n# Problem 4\nk = 101\nlambdas = np.linspace(0, 1, num=k)\ni = 0\nmses4_train = np.zeros((k,1))\nmses4 = np.zeros((k,1))\nopts = {'maxiter' : 100} # Preferred value. \nw_init = np.ones((X_i.shape[1],1))\nfor lambd in lambdas:\n args = (X_i, y, lambd)\n w_l = minimize(regressionObjVal, w_init, jac=True, args=args,method='CG', options=opts)\n w_l = np.transpose(np.array(w_l.x))\n w_l = np.reshape(w_l,[len(w_l),1])\n mses4_train[i] = testOLERegression(w_l,X_i,y)\n mses4[i] = testOLERegression(w_l,Xtest_i,ytest)\n i = i + 1\nfig = plt.figure(figsize=[12,6])\nplt.subplot(1, 2, 1)\nplt.plot(lambdas,mses4_train)\nplt.plot(lambdas,mses3_train)\nplt.title('MSE for Train Data')\nplt.xlabel('$\\lambda$')\nplt.ylabel('Mean Squared Error')\nplt.legend(['Using scipy.minimize','Direct minimization'])\n\nplt.subplot(1, 2, 2)\nplt.plot(lambdas,mses4)\nplt.plot(lambdas,mses3)\nplt.title('MSE for Test Data')\nplt.xlabel('$\\lambda$')\nplt.ylabel('Mean Squared Error')\nplt.legend(['Using scipy.minimize','Direct minimization'])\nplt.show()\n\n#Problem 5\npmax = 7\nlambda_opt = 0.06 # REPLACE THIS WITH lambda_opt estimated from Problem 3\n#lambda_opt = lambdas[np.argmin(mses3)] # 0.059999999999999998\nmses5_train = np.zeros((pmax,2))\nmses5 = np.zeros((pmax,2))\nfor p in range(pmax):\n Xd = mapNonLinear(X[:,2],p)\n Xdtest = mapNonLinear(Xtest[:,2],p)\n w_d1 = learnRidgeRegression(Xd,y,0)\n mses5_train[p,0] = testOLERegression(w_d1,Xd,y)\n mses5[p,0] = testOLERegression(w_d1,Xdtest,ytest)\n w_d2 = learnRidgeRegression(Xd,y,lambda_opt)\n mses5_train[p,1] = testOLERegression(w_d2,Xd,y)\n mses5[p,1] = testOLERegression(w_d2,Xdtest,ytest)\n\nfig = plt.figure(figsize=[12,6])\nplt.subplot(1, 2, 1)\nplt.plot(range(pmax),mses5_train)\nplt.title('MSE for Train Data')\nplt.legend(('No Regularization','Regularization'))\nplt.xlabel('Degree of Non-linearity(p)')\nplt.ylabel('Mean Squared Error')\nplt.subplot(1, 2, 2)\nplt.plot(range(pmax),mses5)\nplt.title('MSE for Test Data')\nplt.legend(('No Regularization','Regularization'))\nplt.xlabel('Degree of Non-linearity(p)')\nplt.ylabel('Mean Squared Error')\nplt.show()\n\n","sub_path":"script.py","file_name":"script.py","file_ext":"py","file_size_in_byte":11985,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"301337873","text":"from absl import app\nfrom absl import flags\nimport socketio\nimport eventlet\nimport time, sys\nfrom utils.logger import Logger\nfrom signal_processing.wheelchair_signals_monitor import Wheelchair_Signals_Monitor\n\n# CONSTANTS\nGSR_CONSTANT = 10000 # used in computing the skin conductance (10000 according to seedstudio)\nEPSILON = 1e-22 # used to prevent numerical divide by zero errors\nSIGNAL_SET = {\"ecg\", \"gsr\", \"pm\", \"wm\"}\n\n# abseil flags\nFLAGS = flags.FLAGS\nflags.DEFINE_list(\"log_set\", \"\", \"set of tags to log\")\nflags.DEFINE_boolean(\"record\", False, \"record samples\")\nflags.DEFINE_enum(\"mode\", \"inference\", \"[inference, data_gathering]\",\n \"inference: predict using model, \"\n \"data_gathering: save data to database\")\n\n# server initializations\nsio = socketio.Server(cors_allowed_origins=\"*\")\nserver = socketio.WSGIApp(sio)\nlogger = Logger()\nsignal_monitor = Wheelchair_Signals_Monitor()\ncurrent_features = None\n\n# @sio.on(\"connect\")\n# def on_connect(sid, environ):\n# time.sleep(5)\n# sio.emit(\"request-discomfort-level\", \"\")\n# sio.emit(\"receive-inference\", \"69:69pm,Severe\")\n\n@sio.on(\"signal-nodemcu\")\ndef receive_signal(sid, data):\n # parse data\n data = data.split(\":\")\n\n # assert data integrity\n if len(data) != 2:\n return\n if not all(list(map(len, data))):\n return\n signal_type, signal = data\n\n # discard signals not in SIGNAL_SET\n if signal_type not in SIGNAL_SET:\n return\n\n # handle signal\n if signal_type == \"pm\" or signal_type == \"wm\":\n signal = list(map(int, signal.split(\",\")))\n if signal[2] < 0: signal[2] = 0\n # assert data integrity\n if len(signal) != 3:\n return\n else:\n signal = float(signal)\n if signal_type == \"gsr\":\n signal = get_skin_conductance(signal)\n\n handle_data(signal_type, signal)\n\ndef handle_data(signal_type, signal):\n current_time = get_time()\n\n # logging\n if FLAGS.record:\n record_values(current_time, signal, f\"./{signal_type}.samples\")\n logger.log(signal_type, f\"{signal_type}:{current_time},{signal}\")\n \n # send signal to mobile\n sio.emit(f\"mobile-{signal_type}-signal\", f\"{get_time()},{signal}\")\n\n # send the signal to the signal monitor\n features = signal_monitor.add_point(signal_type=signal_type, t=current_time, y=signal)\n if features != None:\n logger.log(\"features\", features)\n if FLAGS.mode == \"inference\":\n # # use model to predict\n # prediction = model.predict(features)\n # logger.log(\"inference\", prediction)\n # # send prediction to server\n # sio.emit(\"receive-inference\", f\"{current_time},{prediction}\")\n pass\n elif FLAGS.mode == \"data_gathering\":\n current_features = features\n # request discomfort level from the mobile app\n sio.emit(\"request-discomfort-level\", \"\")\n\n\n@sio.on(\"discomfort-level\")\ndef recieve_discomfort_level(sid, data):\n discomfort_level = data\n logger.log(\"discomfort-level\", data)\n # save current_feaures and discomfort_level to database\n \n\ndef get_skin_conductance(gsr_value):\n # equation from https://wiki.seeedstudio.com/Grove-GSR_Sensor/\n resistance = ((1024 + 2.0 * gsr_value) * GSR_CONSTANT) / ((512 - gsr_value) + EPSILON)\n # conductance is the reciprocal of resistance\n conductance = (1 / resistance) * 1e6\n return conductance\n\ndef get_time():\n reference_time = time.localtime(time.time())\n reference_time = time.mktime((\n reference_time.tm_year, reference_time.tm_mon, reference_time.tm_mday, \n 0, 0, 0, \n reference_time.tm_wday, reference_time.tm_yday, reference_time.tm_isdst\n ))\n millis_today = int((time.time() - reference_time) * 1000)\n return millis_today\n\ndef record_values(t, y, path_to_samples):\n with open(path_to_samples, \"a\") as fout:\n fout.write(f\"{t},{y}\\n\")\n\ndef main(argv):\n logger.log_set = set(FLAGS.log_set)\n eventlet.wsgi.server(eventlet.listen(('', 3000)), server)\n\nif __name__ == \"__main__\":\n app.run(main)\n","sub_path":"server/server.py","file_name":"server.py","file_ext":"py","file_size_in_byte":4109,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"80108497","text":"from InputsGlobais import*\r\n\r\n\r\n\r\nclass Logistica_Setores:\r\n def __init__(self, perc_eixo, frequência, ciclo, fat_manobra, vel_produtiva):\r\n self.perc_eixo=perc_eixo\r\n self.frequência= frequência\r\n self.ciclo=ciclo\r\n self.fat_manobra=fat_manobra\r\n self.vel_produtiva=vel_produtiva\r\n \r\n def km_diária(self):\r\n km_plano=Km_eixo*self.perc_eixo/100\r\n km_diária_plano=km_plano*(self.frequência/self.ciclo)*(1+self.fat_manobra/100)\r\n return km_diária_plano\r\n\r\n def tempo_diário(self):\r\n tempo_diário_plano=self.km_diária()/self.vel_produtiva\r\n return tempo_diário_plano\r\n\r\n\r\n","sub_path":"2015/LogísticaSetores.py","file_name":"LogísticaSetores.py","file_ext":"py","file_size_in_byte":666,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"494211792","text":"from rest_framework import filters\nfrom rest_framework import viewsets\nfrom blog.models import Post\nfrom blog.serializers import PostSerializer, PostDetailSerializer\n\n\nclass PostViewSet(viewsets.ModelViewSet):\n \"\"\"\n A simple ViewSet for listing or retrieving posts.\n \"\"\"\n\n # serializer_class = PostSerializer\n http_method_names = ['get', ]\n\n filter_backends = [filters.SearchFilter, filters.OrderingFilter]\n\n search_fields = [\n 'title',\n 'description',\n 'created_date',\n 'tag__name',\n ]\n\n ordering_fields = [\n 'featured',\n 'publication_date'\n ]\n\n ordering = [\n '-featured',\n 'publication_date',\n ]\n\n def _get_ids_from_csv(self, csv):\n return [int(str_val) for str_val in csv.split(',')]\n \n def get_queryset(self):\n tags = self.request.query_params.get('tags')\n qs = Post.objects.all()\n \n if tags:\n tags = self._get_ids_from_csv(tags)\n qs = qs.filter(tag__in=tags)\n \n return qs\n \n def get_serializer_class(self):\n if self.action == 'list':\n return PostSerializer\n if self.action == 'retrieve':\n return PostDetailSerializer\n","sub_path":"backend/blog/views/post.py","file_name":"post.py","file_ext":"py","file_size_in_byte":1228,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"575631077","text":"from setuptools import setup, find_packages\nimport os\n\ndef is_optional_enabled(optional):\n return os.environ.get(optional, None) is not None\n\nVIZDOOM = 'URNAI_VIZDOOM'\nG2048 = 'URNAI_2048'\nDEEPRTS = 'URNAI_DEEPRTS'\n\n#git_url = '{package} @ git+https://github.com/{user}/{package}.git/@{version}#egg={package}-0'\ngit_url = 'https://github.com/{user}/{package}.git/@{version}#egg={package}-0'\ndep_links = []\ndep_list = []\n\nif is_optional_enabled(DEEPRTS):\n print(\"DeepRTS installation enabled.\")\n dep_links.append(git_url.format(user='UIA-CAIR', package='DeepRTS', version='e54dc6c'))\n\nif is_optional_enabled(VIZDOOM):\n print(\"VizDoom installation enabled.\")\n dep_list.append('vizdoom')\n\nif is_optional_enabled(G2048):\n print(\"Gym-2048 installation enabled.\")\n dep_list.append('gym-2048')\n dep_links.append(git_url.format(user='ntasfi', package='PyGame-Learning-Environment', version='master'))\n\nsetup(\n name = \"urnai\",\n packages = find_packages(),\n install_requires = [\n 'absl-py',\n 'gym',\n 'tensorflow',\n 'numpy',\n 'matplotlib',\n 'keras',\n 'pysc2',\n 'pandas',\n ] + dep_list,\n dependency_links=dep_links,\n entry_points = {\n \"console_scripts\": ['urnai=urnai.urnai_cmd:main']\n },\n version = \"0.0.1-3\",\n description = \"A modular Deep Reinforcement Learning library that supports multiple environments, such as PySC2, OpenAI Gym, and PyGame Learning Environment.\",\n long_description = \"URNAI Tools is a modular Deep Reinforcement Learning library that supports multiple environments, such as PySC2, OpenAI Gym, and PyGame Learning Environment. The main goal of URNAI Tools is to provide an easy way to develop DRL agents in a way that allows the developer to reuse as much code as possible when developing different agents, and that also allows him to reuse previously implemented models in different environments and to integrate new environments easily when necessary. The main advantage of using URNAI Tools is that the models you make for one environment will also work in other environments, so you can prototype new agents for different environments very easily.\",\n author = \"UFRN-IMD-URNAITeam\",\n author_email = \"urnaiteam@gmail.com\",\n url = \"https://github.com/pvnetto/URNAI-Tools\",\n)\n","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":2322,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"404059959","text":"# -*- coding: utf-8 -*-\n\nfrom django.conf.urls import patterns, include, url\n\nurlpatterns = patterns('deveco.views',\n url(r'^$', \"home\", name=\"home\"),\n url(r'^entreprise/(?P[^\\/]+).html$', 'entreprise', name='entreprise'),\n url(r'^entreprisevcard/(?P[^\\/]+).vcf', 'entreprisevcard', name='entreprisevcard')\n\n )\n","sub_path":"deveco/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":411,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"631831233","text":"\"\"\"nn_weights_analysis.py\n\nCS7641-HW2: Optimization\nPerforms ABAGAILs Random Hill Climb optimization\nCode referenced from https://github.com/mitian223/CS7641/blob/master/HW2/ANN_rhc.py\n\n\nMike Tong\n\ncreated: FEB 2019\n\"\"\"\n\nimport os\nimport sys\nsys.path.append('./ABAGAIL/ABAGAIL.jar')\n\nfrom func.nn.backprop import BackPropagationNetworkFactory\nfrom shared import SumOfSquaresError, DataSet\nfrom opt.example import NeuralNetworkOptimizationProblem\nfrom func.nn.backprop import RPROPUpdateRule, BatchBackPropagationTrainer\nimport opt.RandomizedHillClimbing as RandomizedHillClimbing\nimport opt.SimulatedAnnealing as SimulatedAnnealing\nimport opt.ga.StandardGeneticAlgorithm as StandardGeneticAlgorithm\nimport opt.prob.MIMIC as MIMIC\nfrom func.nn.activation import LogisticSigmoid, HyperbolicTangentSigmoid\nfrom ancilary import errorOnDataSet, init_instances, train\n\n\n# Network parameters found \"optimal\" in Assignment 1\nINPUT_LAYER = 19\nHIDDEN_LAYER1 = 12\nHIDDEN_LAYER2 = 12\nOUTPUT_LAYER = 1\nbp_iters = 2000\nrhc_iters = 15000\nsa_iters = 10000\nga_iters = 10000\n\nbp_iterations = 10\nrhc_iterations = 10\n\ndata_path = './assets/data/cleaned_student_data.csv'\nbp_path = './assets/results/nn_weight/BP/BP_LOG.csv'\nrhc_path = './assets/results/nn_weight/RHC/RHC_LOG.csv'\nsa_path = './assets/results/nn_weight/SA/SA_LOG.csv'\nga_path = './assets/results/nn_weight/GA/GA_LOG.csv'\n\n\ndef Backpropogation(out_path, train_inst, test_inst, repeats, training_iterations):\n for i in range(repeats):\n out_path_ = out_path.replace(\"BP_\", 'BP_{}'.format(str(i).zfill(3)))\n with open(out_path_, 'w') as f:\n f.write('{},{},{},{},{},{}\\n'.format('iteration','MSE_trg','MSE_tst','acc_trg','acc_tst','elapsed'))\n factory = BackPropagationNetworkFactory()\n measure = SumOfSquaresError()\n data_set = DataSet(train_inst)\n # acti = LogisticSigmoid()\n acti = HyperbolicTangentSigmoid()\n rule = RPROPUpdateRule()\n classification_network = factory.createClassificationNetwork([INPUT_LAYER, HIDDEN_LAYER1, HIDDEN_LAYER2, OUTPUT_LAYER],acti)\n train(\n BatchBackPropagationTrainer(data_set,classification_network,measure,rule),\n classification_network, 'Backprop', train_inst, test_inst, measure, training_iterations, out_path_\n )\n\ndef Random_hill_climb(out_path, train_inst, test_inst, repeats, training_iterations):\n \"\"\"Run this experiment\"\"\"\n for i in range(repeats):\n out_path_ = out_path.replace(\"RHC_\", 'RHC_{}'.format(str(i).zfill(3)))\n with open(out_path_, 'w') as f:\n f.write('{},{},{},{},{},{}\\n'.format('iteration','MSE_trg','MSE_tst','acc_trg','acc_tst','elapsed'))\n factory = BackPropagationNetworkFactory()\n measure = SumOfSquaresError()\n data_set = DataSet(train_inst)\n # acti = LogisticSigmoid()\n acti = HyperbolicTangentSigmoid()\n rule = RPROPUpdateRule()\n classification_network = factory.createClassificationNetwork([INPUT_LAYER, HIDDEN_LAYER1, HIDDEN_LAYER2, OUTPUT_LAYER],acti)\n nnop = NeuralNetworkOptimizationProblem(data_set, classification_network, measure)\n oa = RandomizedHillClimbing(nnop)\n train(oa, classification_network, 'RHC', train_inst, test_inst, measure, training_iterations, out_path_)\n\ndef Simulated_annealing(out_path, train_inst, test_inst, T, CE, training_iterations):\n \"\"\"Run this experiment\"\"\"\n factory = BackPropagationNetworkFactory()\n measure = SumOfSquaresError()\n data_set = DataSet(train_inst)\n # acti = LogisticSigmoid()\n acti = HyperbolicTangentSigmoid()\n rule = RPROPUpdateRule()\n\n oa_name = \"SA_T{}_CE{}\".format(int(T), str(CE).split('.')[-1])\n with open(out_path.replace('SA_', oa_name),'w') as f:\n f.write('{},{},{},{},{},{}\\n'.format('iteration','MSE_trg','MSE_tst','acc_trg','acc_tst','elapsed'))\n\n classification_network = factory.createClassificationNetwork([INPUT_LAYER, HIDDEN_LAYER1, HIDDEN_LAYER2, OUTPUT_LAYER],acti)\n nnop = NeuralNetworkOptimizationProblem(data_set, classification_network, measure)\n oa = SimulatedAnnealing(T, CE, nnop)\n train(oa, classification_network, oa_name, train_inst, test_inst, measure, training_iterations, out_path.replace('SA_',oa_name))\n\ndef Genetic_algorithm(out_path, train_inst, test_inst, P, mate, mutate, training_iterations):\n \"\"\"Run this experiment\"\"\"\n factory = BackPropagationNetworkFactory()\n measure = SumOfSquaresError()\n data_set = DataSet(train_inst)\n # acti = LogisticSigmoid()\n acti = HyperbolicTangentSigmoid()\n rule = RPROPUpdateRule()\n\n oa_name = \"GA_P{}_mate{}_mut{}\".format(P, mate, mutate)\n with open(out_path.replace('GA_', oa_name),'w') as f:\n f.write('{},{},{},{},{},{}\\n'.format('iteration','MSE_trg','MSE_tst','acc_trg','acc_tst','elapsed'))\n\n classification_network = factory.createClassificationNetwork([INPUT_LAYER, HIDDEN_LAYER1, HIDDEN_LAYER2, OUTPUT_LAYER], acti)\n nnop = NeuralNetworkOptimizationProblem(data_set, classification_network, measure)\n oa = StandardGeneticAlgorithm(P, mate, mutate, nnop)\n train(oa, classification_network, oa_name, train_inst, test_inst, measure, training_iterations, out_path.replace('GA_', oa_name))\n\n\nif __name__ == \"__main__\":\n train_inst, test_inst = init_instances(data_path)\n Backpropogation(bp_path, train_inst, test_inst, bp_iterations, bp_iters)\n Random_hill_climb(rhc_path, train_inst, test_inst, rhc_iterations, rhc_iters)\n #\n for T in [1e1, 1e3, 1e5, 1e7, 1e9, 1e11, 1e13]:\n for CE in [0.20, 0.40, 0.60, 0.80, 0.90, 0.99]:\n Simulated_annealing(sa_path, train_inst, test_inst, T, CE, sa_iters)\n\n for p in [100, 200, 300]:\n for mate in [20, 30, 40]:\n for mutate in [20, 30, 40]:\n Genetic_algorithm(ga_path, train_inst, test_inst, p, mate, mutate, ga_iters)\n","sub_path":"assignments/assignment_2-Randomized_Optimization/analysis/nn_weights_analysis.py","file_name":"nn_weights_analysis.py","file_ext":"py","file_size_in_byte":5865,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"185134222","text":"# -*- coding: utf-8 -*-\n\"\"\"\n blog.articles.views\n ~~~~~~~~~~~~~~~~~~~~\n\n Blog-articles.\n\n :copyright: (c) 2015 by Michael Hoyt.\n\"\"\"\nimport datetime\n\nfrom flask import (Blueprint, redirect, url_for, current_app,\n request, flash, render_template)\n\nfrom blog.articles.models import Category, Article\n\narticles = Blueprint(\"articles\", __name__)\n\n@articles.route('/')\ndef index():\n '''\n The Blog's Index\n\n Contains a list-view of sorted posts and categories/tags.\n '''\n\n articles = Article.query.order_by('id desc').limit(5)\n categories = Category.query.all()\n\n return render_template('articles/index.html',\n articles=articles,\n categories=categories,\n )\n\n@articles.route('/')\n@articles.route('/-')\ndef article_detail(article_id, slug=None):\n '''\n Article Detail-View\n '''\n\n article = Article.query.get(article_id)\n\n config = {\n 'title': article.title,\n 'crumbs': True\n }\n return render_template('articles/article_detail.html',\n article=article,\n **config\n )\n\n@articles.route('/category/')\n@articles.route('/category/-')\ndef category_detail(category_id, slug=None):\n '''\n Category List-View\n '''\n\n category = Category.query.get(category_id)\n\n config = {\n 'title': category.title,\n 'crumbs': True\n }\n return render_template('articles/category_detail.html',\n category=category,\n **config\n )","sub_path":"blog/articles/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1550,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"463104667","text":"from urllib.parse import urlencode\nimport requests\nimport random\nimport string\nimport urllib.request\nimport os\nfrom multiprocessing.pool import Pool\nimport json\nimport math\nimport re\nfrom pymongo import MongoClient\nimport uuid\n\n\nclient = MongoClient()\ndb = client['kobe']\ncollection = db['kobe']\n\nbase_url = 'https://image.baidu.com/search/acjson?'\nheaders = {\n 'Host': 'image.baidu.com',\n 'Pragma': 'no-cache',\n 'Referer': 'https://image.baidu.com/search/index?tn=baiduimage&ipn=r&ct=201326592&cl=2&lm=-1&st=-1&fm=index&fr=&hs=0&xthttps=111111&sf=1&fmq=&pv=&ic=0&nc=1&z=&se=1&showtab=0&fb=0&width=&height=&face=0&istype=2&ie=utf-8&word=%E7%A7%91%E6%AF%94&oq=kebi&rsp=0',\n 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/71.0.3578.98 Safari/537.36',\n 'X-Requested-With': 'XMLHttpRequest'\n}\nregex = re.compile(r'\\\\(?![/u\"])')\n\n\ndef get_page(pn):\n gsm = get_gsm(2)\n random_digits = get_random_digits(13)\n params = {\n 'tn': 'resultjson_com',\n 'ipn': 'rj',\n 'cv': 201326592,\n 'is': '',\n 'fp': 'result',\n 'queryWord': '科比',\n 'cl': 2,\n 'lm': -1,\n 'ie': 'utf-8',\n 'oe': 'utf-8',\n 'adpicid': '',\n 'st': -1,\n 'z': '',\n 'ic': 0,\n 'hd': '',\n 'latest': '',\n 'copyright': '',\n 'word': '科比',\n 's': '',\n 'se': '',\n 'tab': '',\n 'width': '',\n 'height': '',\n 'face': 0,\n 'istype': 2,\n 'qc': '',\n 'nc': 1,\n 'fr': '',\n 'expermode': '',\n 'cg': 'star',\n 'pn': pn,\n 'rn': 30,\n 'gsm': gsm,\n random_digits: '',\n }\n url = base_url + urlencode(params)\n try:\n response = requests.get(url, headers=headers)\n response.encoding = 'utf-8'\n if response.status_code == 200:\n response_content = regex.sub(r'\\\\\\\\', response.content.decode())\n return json.loads(response_content, strict=False)\n except requests.ConnectionError as e:\n print('Error:', e.args)\n\n\ndef parse_page(json, pn):\n if json:\n base_path = os.path.abspath('.')\n if not os.path.exists(os.path.join(base_path, 'kobe-images')):\n os.mkdir(os.path.join(base_path, 'kobe-images'))\n items = json.get('data')\n for i, item in enumerate(items):\n kobe = {}\n image_url = item.get('thumbURL')\n kobe['_id'] = uuid.uuid1()\n kobe['title'] = item.get('fromPageTitleEnc')\n if image_url:\n req = urllib.request.urlopen(image_url)\n buf = req.read()\n file_path = os.path.join(base_path, 'kobe-images', str(pn + i + 1) + '.jpg')\n kobe['path'] = file_path\n if not os.path.exists(file_path):\n with open(file_path, 'wb') as f:\n f.write(buf)\n save_to_mongo(kobe)\n\n\ndef save_to_mongo(result):\n if collection.insert(result):\n print('Saved to Mongo')\n\n\ndef get_gsm(n):\n random_str = ''.join(random.sample(string.ascii_letters[:26] + string.digits, n))\n return random_str\n\n\ndef get_random_digits(n):\n random_str = ''.join(random.sample(string.digits, int(n/2))) + \\\n ''.join(random.sample(string.digits, math.ceil(n/2)))\n return random_str\n\n\ndef main(pn):\n json = get_page(pn)\n parse_page(json, pn)\n\n\nGROUP_START = 0\nGROUP_END = 20\n\n\nif __name__ == '__main__':\n import datetime\n\n print(datetime.datetime.now())\n for i in range(GROUP_START, GROUP_END):\n main(i*30)\n print(datetime.datetime.now())\n\n # print(datetime.datetime.now())\n # pool = Pool()\n # group = [x * 30 for x in range(GROUP_START, GROUP_END)]\n # pool.map(main, group)\n # pool.close()\n # pool.join()\n # print(datetime.datetime.now())\n\n\n\n\n\n\n\n","sub_path":"crawl-kobe.py","file_name":"crawl-kobe.py","file_ext":"py","file_size_in_byte":3905,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"235353115","text":"import time\nfrom pymi.config import config\nfrom pymi.migrator import Migrator\n\n\nclass Rollback(Migrator):\n start = 0\n\n def __init__(self):\n Migrator.__init__(self)\n self.start = 0\n\n def run(self):\n microtime = lambda: int(round(time.time() * 1000))\n self.start = microtime()\n\n driver = self.driver_current()\n connection = self.connection_get(driver)\n ran = self.migrations_run(connection)\n\n last = ran[-1]\n\n batch = last['batch']\n\n to_rollback = []\n\n for item in ran:\n if item['batch'] == batch:\n to_rollback.append(item)\n\n amount = 0\n\n for migration in to_rollback:\n filename = migration['filename']\n filename = filename.replace('_up_', '_down_')\n\n print(\" Rolling back \" + filename + \"...\")\n\n path = config['migrations_folder'] + '/' + filename + '.sql'\n\n with open(path) as f:\n sql = f.read()\n\n with connection.cursor() as cursor:\n cursor.execute(sql)\n\n # Commit the changes.\n connection.commit()\n\n print(\" success!\")\n\n amount += 1\n\n with connection.cursor() as cursor:\n batch_sql = \"DELETE FROM `migrations` WHERE batch=\" + str(batch) + \";\"\n cursor.execute(batch_sql)\n\n connection.commit()\n\n ms = str(microtime() - self.start)\n print(\"Rolled back \" + str(amount) + \" migrations in \" + ms + \"ms.\")\n","sub_path":"pymi/commands/rollback.py","file_name":"rollback.py","file_ext":"py","file_size_in_byte":1527,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"639619707","text":"from django.conf.urls import url\nfrom . import views\n\nurlpatterns = [\n url(r'^$', views.home, name='library'),\n url(r'^add-entry$', views.add_entry, name='add-entry'),\n url(r'^edit-entry/(?P\\S+)$', views.edit_entry, name='edit-entry'),\n url(r'^add-picture/(?P\\S+)$', views.add_picture, name='add-picture'),\n url(r'^add-pictures/(?P\\S+)$', views.add_pictures, name='add-pictures'),\n url(r'^edit-picture/(?P\\S+)/$', views.edit_picture, name='edit-picture'),\n url(r'^save-order', views.save_order, name='save-order'),\n]\n","sub_path":"oklei/back/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":576,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"480416125","text":"import argparse\nimport sqlite3\nimport json\nimport datetime\n\n\ndef export_to_file(args):\n connection = sqlite3.connect(args.db)\n cur = connection.cursor()\n q_data = \"SELECT * FROM data where id = (SELECT min(id) FROM data)\"\n q_del = \"DELETE FROM data where id =? \"\n cur.execute(q_data)\n rec = cur.fetchall()\n data = json.loads(rec[0][3])\n ts = data[\"timestamp\"]\n date = datetime.datetime.strptime(ts, \"%Y-%m-%d %H:%M:%S\")\n today = datetime.datetime.today()\n if rec[0][4] != \"S\" or today - date < datetime.timedelta(days=1):\n print(\"No data to export\",rec[0][4], ts)\n connection.close()\n return\n print(date, today, today - date > datetime.timedelta(days=10))\n f_name = \"mqtt_{yyyy}-{mm}-{dd}.csv\".format(yyyy=date.year, mm=date.month, dd=date.day)\n fo = open(f_name, \"a\")\n rec_count = 0\n while True:\n cur.execute(q_data)\n rec = cur.fetchall()\n data = json.loads(rec[0][3])\n ts = data[\"timestamp\"]\n dt = datetime.datetime.strptime(ts, \"%Y-%m-%d %H:%M:%S\")\n if date.year != dt.year or date.month != dt.month or date.day != dt.day:\n print(\"Exported:\", rec_count, \"File:\", f_name)\n fo.close()\n connection.commit()\n connection.close()\n break\n rec_count += 1\n fo.write(rec[0][3])\n fo.write(\"\\n\")\n cur.execute(q_del, (rec[0][0],))\n\n connection.close()\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument(\"db\",\n help=\"Database file\")\n\n# parser.add_argument(\"file\",\n# help=\"input file\")\n args = parser.parse_args()\n export_to_file(args)\n\n","sub_path":"logger/exp_mqtt.py","file_name":"exp_mqtt.py","file_ext":"py","file_size_in_byte":1718,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"83891235","text":"import pyblish.api\n\n\nclass ValidateMindbenderLookdevSingleShape(pyblish.api.InstancePlugin):\n \"\"\"One mesh per transform\"\"\"\n\n label = \"Lookdev Member Shapes\"\n order = pyblish.api.ValidatorOrder\n hosts = [\"maya\"]\n families = [\"mindbender.lookdev\"]\n\n def process(self, instance):\n from maya import cmds\n\n has_multiple_shapes = list()\n for node in instance:\n\n children = cmds.listRelatives(node, allDescendents=True) or list()\n shapes = cmds.listRelatives(node, shapes=True) or list()\n\n # Ensure there is only one child; there could be many,\n # including other transform nodes.\n has_single_shape = len(children) == 1\n\n # Ensure the one child is a shape\n has_single_child = len(shapes) == 1\n\n # Ensure the one child is of type \"mesh\"\n has_single_mesh = cmds.nodeType(shapes[0]) == \"mesh\"\n\n if not all([has_single_child,\n has_single_shape,\n has_single_mesh]):\n has_multiple_shapes.append(node)\n\n assert not has_multiple_shapes, (\n \"\\\"%s\\\" has transforms with multiple shapes: %s\" % (\n instance, \", \".join(\n \"\\\"\" + member + \"\\\"\" for member in has_multiple_shapes))\n )\n","sub_path":"mindbender/plugins/validate_lookdev_.py","file_name":"validate_lookdev_.py","file_ext":"py","file_size_in_byte":1330,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"224507375","text":"import pandas as pd\n\nclass Data:\n \n CAMINHO_DB = '../DB/dim'\n DB_NOME = 'dim_data'\n EXTENS_DB = '.csv'\n SEP = ';'\n \n TBL_DIM_DATA = CAMINHO_DB + '/' + DB_NOME + EXTENS_DB\n \n def tabelaCalendario(self, inicio='1900-01-01', fim='2099-12-31'):\n df = pd.DataFrame({'data': pd.date_range(inicio, fim)})\n df['dia'] = df.data.dt.day\n df['dia_da_semana_nome'] = df.data.dt.weekday_name\n df['dia_da_semana_numero'] = df.data.dt.weekday\n df['semana_do_mes'] = (df.dia - 1) // 7 + 1\n df['semana_do_ano'] = df.data.dt.weekofyear\n df['mes_numero'] = df.data.dt.month\n df['mes_nome'] = df.data.dt.month_name()\n df['bimestre'] = df.mes_numero // 2\n df['trimestre'] = df.data.dt.quarter\n df['semestre'] = (df.trimestre + 1) // 2\n df['ano'] = df.data.dt.year\n df['ano_mes_dia'] = df.ano.astype(str) + df.mes_numero.astype(str) + df.dia.astype(str)\n df['ano_mes'] = df.ano.astype(str) + df.mes_numero.astype(str)\n return df\n \n def dimensaoData(self):\n df = self.tabelaCalendario('2016-01-01', '2025-12-31')\n df.to_csv(self.TBL_DIM_DATA, index=False, sep=self.SEP)\n \n ","sub_path":"ETL/dim_data.py","file_name":"dim_data.py","file_ext":"py","file_size_in_byte":1211,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"539981759","text":"# Solution : 처음 traverse하면서 target의 height와 root로부터의 위치를 기록해놓음.\n# 두번째 traverse시 target의 위와 아래로 나눠 distance에 따른 node들을 출력.\n# Time : O(N), Space : O(h)\n\nfrom typing import Tuple\n\nclass Solution:\n def distanceK(self, root: TreeNode, target: TreeNode, K: int) -> List[int]:\n def findTarget(node: TreeNode, target: int, h: int, path: int) -> Tuple[int, int]: # return height and path from root\n if not node:\n return None\n if node.val == target:\n return (h, path)\n res_l = findTarget(node.left, target, h+1, path+(1<n:\n break \n print(next(x))\n c+=1\n","sub_path":"fibo2.py","file_name":"fibo2.py","file_ext":"py","file_size_in_byte":561,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"419020818","text":"import FWCore.ParameterSet.Config as cms\nimport copy\nfrom DisappTrks.StandardAnalysis.Cuts import * # Put all the individual cuts in this file\nfrom DisappTrks.StandardAnalysis.EventSelections import * # Get the composite cut definitions##################################################\n\n## Fake track control sample: W->mu nu\n##################################################\nWtoMuNu = cms.PSet(\n name = cms.string(\"WtoMuNu\"),\n triggers = triggersSingleMu,\n cuts = cms.VPSet (\n cutMuonPt25,\n cutMuonExactlyOne,\n cutElectronExactlyZero,\n cutMuonEta21,\n cutMuonTightID,\n cutMuonTightPFIso,\n cutMuonHighMT,\n )\n)\n\n##################################################\n## Fake track control sample: W->mu nu + disappearing track\n##################################################\nWtoMuNuCandTrk = copy.deepcopy(WtoMuNu)\nWtoMuNuCandTrk.name = cms.string(\"WtoMuNuCandTrk\")\naddCuts(WtoMuNuCandTrk.cuts, [cutTrkPt55] + candTrkCuts)\n\n##################################################\n## Fake track control sample: W->mu nu + disappearing track\n##################################################\nWtoMuNuDisTrk = copy.deepcopy(WtoMuNu)\nWtoMuNuDisTrk.name = cms.string(\"WtoMuNuDisTrk\")\naddCuts(WtoMuNuDisTrk.cuts, [cutTrkPt55] + disTrkCuts)\n\n##################################################\n## Fake track control sample: W->mu nu + disappearing track with 3 hits\n##################################################\nWtoMuNuDisTrkNHits3 = copy.deepcopy(WtoMuNuDisTrk)\nWtoMuNuDisTrkNHits3.name = cms.string(\"WtoMuNuDisTrkNHits3\")\ncutsToRemove = [\n cutTrkNValidHits,\n]\ncutsToAdd = [\n cutTrkNValidHits3,\n]\nremoveCuts(WtoMuNuDisTrkNHits3.cuts, cutsToRemove)\naddCuts (WtoMuNuDisTrkNHits3.cuts, cutsToAdd)\n\n##################################################\n## Fake track control sample: W->mu nu + disappearing track with 4 hits\n##################################################\nWtoMuNuDisTrkNHits4 = copy.deepcopy(WtoMuNuDisTrk)\nWtoMuNuDisTrkNHits4.name = cms.string(\"WtoMuNuDisTrkNHits4\")\ncutsToRemove = [\n cutTrkNValidHits,\n]\ncutsToAdd = [\n cutTrkNValidHits4,\n]\nremoveCuts(WtoMuNuDisTrkNHits4.cuts, cutsToRemove)\naddCuts (WtoMuNuDisTrkNHits4.cuts, cutsToAdd)\n\n##################################################\n## Fake track control sample: W->mu nu + disappearing track with 5 hits\n##################################################\nWtoMuNuDisTrkNHits5 = copy.deepcopy(WtoMuNuDisTrk)\nWtoMuNuDisTrkNHits5.name = cms.string(\"WtoMuNuDisTrkNHits5\")\ncutsToRemove = [\n cutTrkNValidHits,\n]\ncutsToAdd = [\n cutTrkNValidHits5,\n]\nremoveCuts(WtoMuNuDisTrkNHits5.cuts, cutsToRemove)\naddCuts (WtoMuNuDisTrkNHits5.cuts, cutsToAdd)\n\n##################################################\n## Fake track control sample: W->mu nu + disappearing track with 6 hits\n##################################################\nWtoMuNuDisTrkNHits6 = copy.deepcopy(WtoMuNuDisTrk)\nWtoMuNuDisTrkNHits6.name = cms.string(\"WtoMuNuDisTrkNHits6\")\ncutsToRemove = [\n cutTrkNValidHits,\n]\ncutsToAdd = [\n cutTrkNValidHits6,\n]\nremoveCuts(WtoMuNuDisTrkNHits6.cuts, cutsToRemove)\naddCuts (WtoMuNuDisTrkNHits6.cuts, cutsToAdd)\n","sub_path":"BackgroundEstimation/python/WtoMuNuSelections.py","file_name":"WtoMuNuSelections.py","file_ext":"py","file_size_in_byte":3161,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"102183641","text":"#! -*- coding: utf-8 -*-\n# desc: Transformer 基础类\nfrom typing import List\nimport tensorflow as tf\nimport tensorflow.keras.backend as K\nfrom tensorflow.keras import initializers\nfrom tensorflow.keras.layers import Dropout, Add\nfrom tensorflow.keras.models import Model\nfrom bert4one.layers import MultiHeadAttention, Concatenate1D\nimport numpy as np\n\n\nclass Transformer(object):\n \"\"\"模型基类\n \"\"\"\n\n def __init__(\n self,\n vocab_size, # 词表大小\n hidden_size, # 编码维度\n num_hidden_layers, # Transformer总层数\n num_attention_heads, # Attention的头数\n intermediate_size, # FeedForward的隐层维度\n hidden_act, # FeedForward隐层的激活函数\n dropout_rate=None, # Dropout比例\n embedding_size=None, # 是否指定embedding_size\n attention_head_size=None, # Attention中V的head_size\n attention_key_size=None, # Attention中Q,K的head_size\n sequence_length=None, # 是否固定序列长度\n keep_tokens=None, # 要保留的词ID列表\n compound_tokens=None, # 扩展Embedding\n residual_attention_scores=False, # Attention矩阵加残差\n layers=None, # 外部传入的Keras层\n prefix=None, # 层名前缀\n name=None, # 模型名称\n **kwargs):\n if keep_tokens is not None:\n vocab_size = len(keep_tokens)\n if compound_tokens is not None:\n vocab_size += len(compound_tokens)\n self.vocab_size = vocab_size\n self.hidden_size = hidden_size\n self.num_hidden_layers = num_hidden_layers\n self.num_attention_heads = num_attention_heads\n self.attention_head_size = attention_head_size or hidden_size // num_attention_heads\n self.attention_key_size = attention_key_size or self.attention_head_size\n self.intermediate_size = intermediate_size\n self.dropout_rate = dropout_rate or 0\n self.hidden_act = hidden_act\n self.embedding_size = embedding_size or hidden_size\n self.sequence_length = sequence_length\n self.keep_tokens = keep_tokens\n self.compound_tokens = compound_tokens\n self.attention_bias = None\n self.position_bias = None\n self.attention_scores = None\n self.residual_attention_scores = residual_attention_scores\n self.layers = {} if layers is None else layers\n self.prefix = prefix or ''\n self.name = name\n self.built = False\n\n def build(\n self,\n attention_caches=None,\n layer_norm_cond=None,\n layer_norm_cond_hidden_size=None,\n layer_norm_cond_hidden_act=None,\n additional_input_layers=None,\n **kwargs):\n \"\"\"模型构建函数\n attention_caches:为Attention的K,V的缓存序列字典,格式为\n {Attention层名: [K缓存, V缓存]};\n layer_norm_*系列参数:实现Conditional Layer Normalization时使用,\n 用来实现以“固定长度向量”为条件的条件Bert。\n \"\"\"\n if self.built:\n return None\n # Input\n inputs = self.get_inputs()\n self.set_inputs(inputs, additional_input_layers)\n # Other\n self.attention_caches = attention_caches or {}\n self.layer_norm_conds = [\n layer_norm_cond,\n layer_norm_cond_hidden_size,\n layer_norm_cond_hidden_act or 'linear',\n ]\n # Call\n outputs = self.call(inputs)\n self.set_outputs(outputs)\n # Model\n self.model = Model(self.inputs, self.outputs, name=self.name)\n self.built = True\n\n def call(self, inputs):\n \"\"\"定义模型的执行流程\n \"\"\"\n # Embedding\n outputs = self.apply_embeddings(inputs)\n # Main\n for i in range(self.num_hidden_layers):\n outputs = self.apply_main_layers(outputs, i)\n # Final\n outputs = self.apply_final_layers(outputs)\n return outputs\n\n def prefixed(self, name):\n \"\"\"给名字加前缀\n \"\"\"\n if name is not None:\n return self.prefix + name\n\n def apply(self, inputs=None, layer=None, arguments=None, **kwargs):\n \"\"\"通过apply调用层会自动重用同名层\n inputs: 上一层的输出;\n layer: 要调用的层类名;\n arguments: 传递给layer.call的参数;\n kwargs: 传递给层初始化的参数。\n \"\"\"\n if layer is Dropout and self.dropout_rate == 0:\n return inputs\n\n if layer is MultiHeadAttention and self.residual_attention_scores:\n kwargs['return_attention_scores'] = True\n\n arguments = arguments or {}\n name = self.prefixed(kwargs.get('name'))\n kwargs['name'] = name\n if name not in self.layers:\n layer = layer(**kwargs)\n name = layer.name\n self.layers[name] = layer\n\n if inputs is None:\n return self.layers[name]\n else:\n if isinstance(self.layers[name], MultiHeadAttention):\n if name in self.attention_caches:\n # 如果检测到Cache的传入,那么自动在Key,Value处拼接起来\n k_cache, v_cache = self.attention_caches[name]\n k_name, v_name = name + '-Cached-Key', name + '-Cached-Value'\n k = Concatenate1D(name=k_name)([k_cache, inputs[1]])\n v = Concatenate1D(name=v_name)([v_cache, inputs[2]])\n inputs = inputs[:1] + [k, v] + inputs[3:]\n if self.residual_attention_scores:\n # 如果使用残差Attention矩阵,则给每个Attention矩阵加上前上一层的Attention\n # 矩阵,这对应RealFormer设计(https://arxiv.org/abs/2012.11747)。目前\n # 该实现还相对粗糙,可能欠缺通用性。\n if self.attention_scores is not None:\n if arguments.get('a_bias'):\n a_bias = Add(name=name + '-Attention-Bias'\n )([inputs[3], self.attention_scores])\n else:\n a_bias = self.attention_scores\n inputs = inputs[:3] + [a_bias] + inputs[4:]\n arguments['a_bias'] = True\n o, a = self.layers[name](inputs, **arguments)\n self.attention_scores = a\n return o\n return self.layers[name](inputs, **arguments)\n\n def get_inputs(self):\n raise NotImplementedError\n\n def apply_embeddings(self, inputs):\n raise NotImplementedError\n\n def apply_main_layers(self, inputs, index):\n raise NotImplementedError\n\n def apply_final_layers(self, inputs):\n raise NotImplementedError\n\n def compute_attention_bias(self, inputs=None):\n \"\"\"定义每一层的Attention Bias\n \"\"\"\n return self.attention_bias\n\n def compute_position_bias(self, inputs=None):\n \"\"\"定义每一层的Position Bias(一般相对位置编码用)\n \"\"\"\n return self.position_bias\n\n def set_inputs(self, inputs, additional_input_layers=None):\n \"\"\"设置input和inputs属性\n \"\"\"\n if inputs is None:\n inputs = []\n elif not isinstance(inputs, list):\n inputs = [inputs]\n\n inputs = inputs[:]\n if additional_input_layers is not None:\n if not isinstance(additional_input_layers, list):\n additional_input_layers = [additional_input_layers]\n inputs.extend(additional_input_layers)\n\n self.inputs = inputs\n if len(inputs) > 1:\n self.input = inputs\n else:\n self.input = inputs[0]\n\n def set_outputs(self, outputs):\n \"\"\"设置output和oututs属性\n \"\"\"\n if not isinstance(outputs, list):\n outputs = [outputs]\n\n outputs = outputs[:]\n self.outputs = outputs\n if len(outputs) > 1:\n self.output = outputs\n else:\n self.output = outputs[0]\n\n @property\n def initializer(self):\n \"\"\"默认使用截断正态分布初始化\n \"\"\"\n return initializers.TruncatedNormal(stddev=0.02)\n\n def simplify(self, inputs):\n \"\"\"将list中的None过滤掉\n \"\"\"\n inputs = [i for i in inputs if i is not None]\n if len(inputs) == 1:\n inputs = inputs[0]\n\n return inputs\n\n def load_embeddings(self, embeddings):\n \"\"\"处理Embedding层权重\n \"\"\"\n if self.keep_tokens is not None:\n embeddings = embeddings[self.keep_tokens]\n\n if self.compound_tokens is not None:\n ext_embeddings = []\n for item in self.compound_tokens:\n if isinstance(item, list):\n item = (item, [1] * len(item))\n ext_embeddings.append(\n np.average(embeddings[item[0]], 0, item[1])\n )\n embeddings = np.concatenate([embeddings, ext_embeddings], 0)\n\n return embeddings\n\n def load_variable(self, checkpoint, name):\n \"\"\"加载单个变量的函数\n \"\"\"\n if isinstance(checkpoint, dict):\n return checkpoint[name]\n else:\n return tf.train.load_variable(checkpoint, name)\n\n def create_variable(self, name, value):\n \"\"\"创建一个变量\n \"\"\"\n return K.variable(self.initializer(value.shape), name=name), value\n\n def variable_mapping(self):\n \"\"\"构建keras层与checkpoint的变量名之间的映射表\n \"\"\"\n return {}\n\n def load_weights_from_checkpoint(self, checkpoint, mapping=None):\n \"\"\"根据mapping从checkpoint加载权重\n \"\"\"\n mapping = mapping or self.variable_mapping()\n mapping = {self.prefixed(k): v for k, v in mapping.items()}\n mapping = {k: v for k, v in mapping.items() if k in self.layers}\n\n weight_value_pairs = []\n for layer, variables in mapping.items():\n layer = self.layers[layer]\n weights = layer.trainable_weights\n values = [self.load_variable(checkpoint, v) for v in variables]\n\n if isinstance(layer, MultiHeadAttention):\n \"\"\"如果key_size不等于head_size,则可以通过\n 正交矩阵将相应的权重投影到合适的shape。\n \"\"\"\n count = 2\n if layer.use_bias:\n count += 2\n heads = self.num_attention_heads\n head_size = self.attention_head_size\n key_size = self.attention_key_size\n W = np.linalg.qr(np.random.randn(key_size, head_size))[0].T\n if layer.attention_scale:\n W = W * key_size**0.25 / head_size**0.25\n for i in range(count):\n w, v = weights[i], values[i]\n w_shape, v_shape = K.int_shape(w), v.shape\n if w_shape[-1] != v_shape[-1]:\n pre_shape = w_shape[:-1]\n v = v.reshape(pre_shape + (heads, head_size))\n v = np.dot(v, W)\n v = v.reshape(pre_shape + (heads * key_size,))\n values[i] = v\n\n weight_value_pairs.extend(zip(weights, values))\n\n K.batch_set_value(weight_value_pairs)\n\n def save_weights_as_checkpoint(self, filename, mapping=None):\n \"\"\"根据mapping将权重保存为checkpoint格式\n \"\"\"\n mapping = mapping or self.variable_mapping()\n mapping = {self.prefixed(k): v for k, v in mapping.items()}\n mapping = {k: v for k, v in mapping.items() if k in self.layers}\n\n with tf.Graph().as_default():\n all_variables, all_values = [], []\n for layer, variables in mapping.items():\n layer = self.layers[layer]\n values = K.batch_get_value(layer.trainable_weights)\n for name, value in zip(variables, values):\n variable, value = self.create_variable(name, value)\n all_variables.append(variable)\n all_values.append(value)\n with tf.Session() as sess:\n K.batch_set_value(zip(all_variables, all_values))\n saver = tf.train.Saver()\n saver.save(sess, filename)\n","sub_path":"bert4one/modules/transformer.py","file_name":"transformer.py","file_ext":"py","file_size_in_byte":12546,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"84527016","text":"import numpy as np\nfrom detection_metrics.utils.bbox import jaccard\n\n\nclass Fbeta:\n def __init__(self, n_class, beta=1, overlap_threshold=0.5):\n self.n_class = n_class\n self.beta = beta\n self.iou_thresh = overlap_threshold\n\n def f_beta(self, tp, fp, fn):\n if tp == 0 and fp == 0 and fn == 0:\n return -1\n return (1 + self.beta ** 2) * tp / ((1 + self.beta ** 2) * tp + (self.beta ** 2) * fn + fp)\n\n def evaluate_dataset(self, pred_bb_list, gt_bb_list, pred_classes_list=None, gt_classes_list=None):\n\n if gt_classes_list is None or pred_classes_list is None:\n assert self.n_class == 1, f\"Classes are not provided though n_classes = {self.n_class}\"\n\n assert (len(pred_bb_list) == len(gt_bb_list))\n if pred_classes_list is not None:\n assert (len(pred_bb_list) == len(pred_classes_list))\n if gt_classes_list is not None:\n assert (len(gt_bb_list) == len(gt_classes_list))\n print(\"If -1 returned than tp, fp, fn = 0! So dont include thouse objects into metric calculations!\")\n\n out = []\n for i in range(len(pred_bb_list)):\n pred = np.array(pred_bb_list[i])\n gt = np.array(gt_bb_list[i])\n\n pred_cls = np.zeros((len(pred)))\n if pred_classes_list is not None:\n pred_cls = np.array(pred_classes_list[i])\n\n gt_cls = np.zeros((len(gt)))\n if gt_classes_list is not None:\n gt_cls = np.array(gt_classes_list[i])\n\n k = self.evaluate(pred, pred_cls, gt, gt_cls)\n out.append(k)\n out = np.array(out)\n\n mean_f_betas_per_class = []\n for i in range(self.n_class):\n x = out[:, i]\n x = x[x != -1]\n mean_f_betas_per_class.append(x.mean())\n return out, mean_f_betas_per_class\n\n def evaluate(self, pred_bb, pred_classes, gt_bb, gt_classes):\n f_beta_per_class = []\n if len(pred_bb) == 0 and len(gt_bb) == 0:\n return [-1] * 3\n elif len(pred_bb) == 0:\n for i in range(self.n_class):\n gt_number = np.sum(gt_classes == i)\n fn = gt_number\n tp = 0\n fp = 0\n f_beta = self.f_beta(tp, fp, fn)\n f_beta_per_class.append(f_beta)\n return f_beta_per_class\n\n elif len(gt_bb) == 0:\n for i in range(self.n_class):\n pred_number = np.sum(pred_classes == i)\n fp = pred_number\n tp = 0\n fn = 0\n f_beta = self.f_beta(tp, fp, fn)\n f_beta_per_class.append(f_beta)\n return f_beta_per_class\n\n IoUmask = self.compute_iou_mask(pred_bb, gt_bb, self.iou_thresh)\n\n for i in range(self.n_class):\n gt_number = np.sum(gt_classes == i)\n pred_mask = (pred_classes == i)\n pred_number = np.sum(pred_mask)\n if pred_number == 0:\n fp = gt_number\n tp = 0\n fn = 0\n else:\n IoU1 = IoUmask[pred_mask, :]\n mask = IoU1[:, gt_classes == i]\n\n tp = Fbeta.compute_true_positive(mask)\n fp = pred_number - tp\n fn = gt_number - tp\n\n f_beta = self.f_beta(tp, fp, fn)\n f_beta_per_class.append(f_beta)\n return f_beta_per_class\n\n @staticmethod\n def compute_iou_mask(prediction, gt, overlap_threshold):\n IoU = jaccard(prediction, gt)\n # for each prediction select gt with the largest IoU and ignore the others\n for i in range(len(prediction)):\n maxj = IoU[i, :].argmax()\n IoU[i, :maxj] = 0\n IoU[i, (maxj + 1):] = 0\n # make a mask of all \"matched\" predictions vs gt\n return IoU >= overlap_threshold\n\n @staticmethod\n def compute_true_positive(mask):\n # sum all gt with prediction of its class\n return np.sum(mask.any(axis=0))\n","sub_path":"detection_metrics/detection_fbeta.py","file_name":"detection_fbeta.py","file_ext":"py","file_size_in_byte":4028,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"153122060","text":"import os\n\nbind = '127.0.0.1:5000'\nworkers = 2\nbacklog = 2048\nworker_class = \"sync\"\ndebug = True\ndaemon = True\nproc_name = 'gunicorn.proc'\npidfile = '/tmp/gunicorn.pid'\nlogfile = '/var/tmp/gunicorn_debug.log'\nloglevel = 'debug'\n","sub_path":"python3/gunicorn.conf.py","file_name":"gunicorn.conf.py","file_ext":"py","file_size_in_byte":228,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"247552174","text":"from django.conf.urls import url\nfrom django.urls import path\nfrom .views import *\n\n\nurlpatterns = [\n url(r'^$', index, name='main_url'),\n path('create', CardCreate.as_view(), name='card_create_url'),\n path('edit/', CardEdit.as_view(), name='card_edit_url'),\n path('delete/', delete, name='card_delete_url'),\n path('base', base, name='base_url'),\n path('personal_view/', personal_view, name='personal_view_url'),\n path('search/', search, name='search_url'),\n]\n","sub_path":"cards/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":519,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"452283902","text":"#La empresa que administra los cajeros automáticos lo contrata para que programen la entrega de\n#billetes, el usuario ingresa la cantidad de dinero que desea y usted debe indicar cuantos billetes de cada\n#denominación se debe entregar. Es importante entregar siempre la menor cantidad de billetes. \n\ndineroPedido=int(input(\"Ingresar dinero requerido \"))\n\n##############################\nde1000=dineroPedido//1000\nresto= dineroPedido%1000\n##############################\nde500=resto//500\nresto%=500\n##############################\nde200=resto//200\nresto%=200\n##############################\nde100=resto//100\nresto%=100\n##############################\nde50=resto//50\nresto%=50\n##############################\n\nprint(\"Usted va a recibir\",de1000,\"billete/s de $1000.\\n\",de500,\"billete/s de $500.\\n\",de200,\"billete/s de $200.\\n\",de100,\"billete/s de $100.\\n\",de50,\"billete/s de $50.\\n A usted le va a quedar en su cuenta $\",resto)\nsalida=input(\"Ingrese una letra y oprima 'Enter'\")","sub_path":"Practica1/Ejercicio17.py","file_name":"Ejercicio17.py","file_ext":"py","file_size_in_byte":971,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"606215314","text":"file_name = \"game_stat.txt\"\nyear = 2006\ngenre = \"RPG\"\ntitle = \"Counter-Strike\"\n\n\ndef get_table(file_name):\n with open(file_name, \"r\") as myfile:\n lines = myfile.readlines()\n table = [element.replace(\"\\n\", \"\").split(\"\\t\") for element in lines]\n return table\n\n\n# How many games are in the file?\ndef count_games(file_name):\n table = get_table(file_name)\n count = len(table)\n return count\n\n\n# Is there a game from a given year?\ndef decide(file_name, year):\n table = get_table(file_name)\n release = []\n for row in table:\n if int(row[2]) == year:\n release.append(int(row[2]))\n return len(release) != 0\n\n\n# Which was the latest game?\ndef get_latest(file_name):\n table = get_table(file_name)\n latest_year = max([row[2] for row in table])\n latest_release = next(row for row in table if row[2] == str(latest_year))\n return latest_release[0]\n\n\n# How many games do we have by genre?\ndef count_by_genre(file_name, genre):\n table = get_table(file_name)\n num_of_same_genre = []\n for row in table:\n if genre == row[3]:\n num_of_same_genre.append(row[0])\n return len(num_of_same_genre)\n\n\n# What is the line number of the given game (by title)?\ndef get_line_number_by_title(file_name, title):\n table = get_table(file_name)\n titles = [row[0] for row in table]\n if title in titles:\n return (titles.index(title) + 1)\n else:\n raise ValueError(\"{} is not found in the game stats.\".format(title))\n\n\n# What is the alphabetical ordered list of the titles?\n# Expected name of the function: sort_abc(file_name)\n# Expected output of the function: a list of strings\n# Extra: do not use the builtin sort() or sorted() functions, but implement an easy sort algorythm by your own!\n\n# What are the genres?\n# Expected name of the function: get_genres(file_name)\n# Expected output of the function: a list of the genres (without duplicates, in alphabetical order)\n\n# What is the release date of the top sold \"First-person shooter\" game?\n# Expected name of the function: when_was_top_sold_fps(file_name)\n# Expected output of the function: year of the release, as integer\n# (if there is no game with genre \"First-person shooter\" then raises ValueError exception\n","sub_path":"reports.py","file_name":"reports.py","file_ext":"py","file_size_in_byte":2245,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"49828767","text":"# -*- coding: utf-8 -*-\n\"\"\"\n~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\nRender 3D and 4D Polytopes using Python and POV-Ray\n~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n\nUsage:\n 1. Make sure the free raytracer POV-Ray is installed on your\n computer and can be found in system environment variables.\n\n 2. Run python main.py and wait for amazing things to happen!\n\"\"\"\nimport subprocess\nimport models\n\n\n# use higher supersampling level (e.g. 5) and smaller antialiasing level\n# (e.g. 0.001) to get betweer images\nPOVRAY_EXE = \"povray\"\nIMAGE_SIZE = 600\nIMAGE_QUALITY_LEVEL = 11 # between 0-11\nSUPER_SAMPLING_LEVEL = 5 # between 1-9\nANTIALIASING_LEVEL = 0.001\n\nTEMPLATE = \"cd povray/ && \" + \\\n POVRAY_EXE + \\\n \" +I{}\"+ \\\n \" +W\" + str(IMAGE_SIZE) + \\\n \" +H\" + str(IMAGE_SIZE) + \\\n \" +Q\" + str(IMAGE_QUALITY_LEVEL) + \\\n \" +A\" + str(ANTIALIASING_LEVEL) + \\\n \" +R\" + str(SUPER_SAMPLING_LEVEL) + \\\n \" +O{}\"\n\n\ndef render_polyhedra(coxeter_diagram,\n trunc_type,\n render_file=\"polyhedra.pov\",\n description=None,\n snub=False):\n if snub:\n P = models.Snub(coxeter_diagram, trunc_type)\n else:\n P = models.Polyhedra(coxeter_diagram, trunc_type)\n\n if not description:\n description = render_file[:-4]\n\n P.build_geometry()\n P.export_pov()\n \n command = TEMPLATE.format(render_file, description)\n subprocess.call(command, shell=True)\n\n\ndef render_polychora(coxeter_diagram,\n trunc_type,\n render_file,\n description=None):\n if not description:\n description = render_file[:-4]\n\n P = models.Polychora(coxeter_diagram, trunc_type)\n P.build_geometry()\n P.export_pov()\n\n command = TEMPLATE.format(render_file, description)\n subprocess.call(command, shell=True)\n\n\nif __name__ == \"__main__\":\n # platonic solids\n \"\"\"\n render_polyhedra((3, 2, 3), (1, 0, 0), description=\"tetrahedron\")\n render_polyhedra((4, 2, 3), (1, 0, 0), description=\"cube\")\n render_polyhedra((3, 2, 4), (1, 0, 0), description=\"octahedron\")\n render_polyhedra((5, 2, 3), (1, 0, 0), description=\"dodecahedron\")\n \"\"\"\n render_polyhedra((3, 2, 5), (1, 0, 0), description=\"icosahedron\")\n\n # archimedean solids\n \"\"\"\n render_polyhedra((3, 2, 3), (1, 1, 0), description=\"truncated-tetrahedron\")\n render_polyhedra((4, 2, 3), (1, 1, 0), description=\"truncated-cube\")\n render_polyhedra((3, 2, 4), (1, 1, 0), description=\"truncated-octahedron\")\n render_polyhedra((5, 2, 3), (1, 1, 0), description=\"truncated-dodecahedron\")\n render_polyhedra((3, 2, 5), (1, 1, 0), description=\"truncated-icosahedron\")\n render_polyhedra((4, 2, 3), (0, 1, 0), description=\"cuboctahedron\")\n render_polyhedra((5, 2, 3), (0, 1, 0), description=\"icosidodecahedron\")\n render_polyhedra((4, 2, 3), (1, 0, 1), description=\"rhombicuboctahedron\")\n render_polyhedra((5, 2, 3), (1, 0, 1), description=\"rhombicosidodecahedron\")\n render_polyhedra((4, 2, 3), (1, 1, 1), description=\"truncated-cuboctahedron\")\n \"\"\"\n render_polyhedra((5, 2, 3), (1, 1, 1), description=\"truncated-icosidodecahedron\")\n render_polyhedra((4, 2, 3), (1, 1, 1), description=\"snub-cube\", snub=True)\n render_polyhedra((5, 2, 3), (1, 1, 1), description=\"snub-dodecahedron\", snub=True)\n\n # prism and antiprism\n render_polyhedra((7, 2, 2), (1, 0, 1), description=\"7-prism\")\n render_polyhedra((8, 2, 2), (1, 1, 1), description=\"8-antiprism\", snub=True)\n\n # regular polychora\n render_polychora((3, 2, 2, 3, 2, 3), (1, 0, 0, 0), \"5-cell-1000.pov\", \"5-cell\")\n \"\"\"\n render_polychora((4, 2, 2, 3, 2, 3), (1, 0, 0, 0), \"8-cell-1000.pov\", \"8-cell\")\n render_polychora((3, 2, 2, 3, 2, 4), (1, 0, 0, 0), \"16-cell-1000.pov\", \"16-cell\")\n render_polychora((3, 2, 2, 4, 2, 3), (1, 0, 0, 0), \"24-cell-1000.pov\", \"24-cell\")\n render_polychora((5, 2, 2, 3, 2, 3), (1, 0, 0, 0), \"120-cell-1000.pov\", \"120-cell\")\n \"\"\"\n render_polychora((3, 2, 2, 3, 2, 5), (1, 0, 0, 0), \"600-cell-1000.pov\", \"600-cell\")\n # some truncated polychora\n # for more examples see the .pov files in the povray directory\n \"\"\"\n render_polychora((3, 2, 2, 3, 2, 3), (1, 1, 1, 0), \"5-cell-1110.pov\", \"cantitruncated-5-cell\")\n render_polychora((4, 2, 2, 3, 2, 3), (1, 1, 0, 0), \"8-cell-1100.pov\", \"truncated-8-cell\")\n render_polychora((3, 2, 2, 3, 2, 4), (1, 0, 0, 1), \"16-cell-1001.pov\", \"runcinated-16-cell\")\n render_polychora((3, 2, 2, 4, 2, 3), (1, 1, 1, 0), \"24-cell-1110.pov\", \"cantitruncated-24-cell\")\n \"\"\"\n render_polychora((3, 2, 2, 3, 2, 5), (1, 1, 0, 0), \"600-cell-1100.pov\", \"truncated-600-cell\")\n\n # 4d prism and duoprism\n render_polychora((6, 2, 2, 2, 2, 8), (1, 0, 1.6, 0), \"duoprism.pov\", \"6-8-duoprism\")\n render_polychora((5, 2, 2, 3, 2, 2), (1, 1, 0, 1), \"prism.pov\", \"5-3-prism\")\n","sub_path":"src/polytopes/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":4963,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"222655188","text":"# Python 3.6.2\n# Script automating the starting of the client individually.\n# Initialize the client\nimport client\nimport json\n\nwith open(\"client_configuration.json\") as client_configuration:\n \"\"\"Sets each variable equal to the value given in the client_configuration.json file\"\"\"\n\n client_config_data = json.load(client_configuration)\n port = client_config_data[\"port\"]\n network_architecture = client_config_data[\"network_architecture\"]\n remote_addresses = client_config_data[\"remote_addresses\"]\n command_execution = client_config_data[\"command_execution\"]\n default_log_level = client_config_data[\"default_log_level\"]\n modules = client_config_data[\"modules\"]\n net_size = client_config_data[\"net_size\"]\n directory_server = client_config_data[\"directory_server\"]\n\n\ndef init():\n x = client.Client()\n x.initialize(port=port, net_architecture=network_architecture, remote_addresses=remote_addresses,\n command_execution=command_execution, default_log_level=default_log_level, modules=modules,\n net_size=net_size, input_directory_server=directory_server)\n\n\nif __name__ == \"__main__\":\n init()\n","sub_path":"src/client/init_client.py","file_name":"init_client.py","file_ext":"py","file_size_in_byte":1156,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"486015094","text":"#!/usr/bin/env python3\n\"\"\"\nOnline (mostly) Schema Changes.\n\nRewritten in Python from the shell script\n\n!!! NOTE !!! If using this on a master consider using --no-replicate.\nIf you cannot do so, then the schema change probably deserves\na master rotation instead :-)\n\n!!! NOTE !!! Go read about table metadata locking during DDL in MariaDB 5.5.\nBe afraid. Be very afraid...\n\"\"\"\nimport argparse\nimport re\nimport shutil\nimport subprocess\nimport sys\n\nfrom wmfmariadbpy.WMFMariaDB import WMFMariaDB\n\n\nclass OnlineSchemaChanger(object):\n \"\"\"Class to perform the online schema changes.\"\"\"\n\n def __init__(self, conf):\n \"\"\"Get the configuration.\"\"\"\n self.conf = conf\n self._conn = None\n self._osctool = \"\"\n self._ptrep = []\n self._ptargs = []\n self._ptdrargs = []\n self._ddlrep = []\n self._ddlargs = []\n\n def __del__(self):\n \"\"\"Tidy up before finishing.\"\"\"\n if self._conn:\n self._conn.disconnect()\n self._conn = None\n\n @property\n def connection(self):\n \"\"\"Return the connection to the database (and create if needed).\"\"\"\n if not self._conn:\n self._conn = WMFMariaDB(host=self.conf.host,\n port=self.conf.port,\n debug=self.conf.debug)\n if not self._conn or not self._conn.connection:\n print(\"Connect failed: {}@{}:{}\".format(self.conf.user,\n self.conf.host,\n self.conf.port))\n sys.exit(1)\n\n return self._conn\n\n @property\n def ddl_rep(self):\n \"\"\"Get the data definition language replication args.\"\"\"\n if not self._ddlrep:\n self._ddlrep = []\n\n if self.conf.replicate:\n self._ddlrep.append(\"set session sql_log_bin=1;\")\n\n if self.conf.no_replicate:\n self._ddlrep.append(\"set session sql_log_bin=0;\")\n\n if self.conf.gtid_domain_id is not None:\n gtid = self.conf.gtid_domain_id\n self._ddlrep.append(\"set session gtid_domain_id = {};\".format(gtid))\n\n return self._ddlrep\n\n @property\n def ddl_args(self):\n \"\"\"Get the data definition language args.\"\"\"\n if not self._ddlargs:\n self._ddlargs = [\"SET SESSION innodb_lock_wait_timeout=1;\",\n \"SET SESSION lock_wait_timeout=60;\"]\n\n self._ddlargs.extend(self.ddl_rep)\n\n return self._ddlargs\n\n @property\n def pt_osc_rep(self):\n \"\"\"Get the Percona Toolkit OSC replication args.\"\"\"\n if not self._ptrep:\n self._ptrep = [\"--recurse=0\"]\n\n if self.conf.replicate:\n self._ptrep = [\"--recurse=1\", \"--chunk-size-limit=10\"]\n\n if self.conf.no_replicate:\n self._ptrep = [\"--recurse=0\", \"--set-vars=sql_log_bin=off\"]\n\n res = self.connection.execute(\"show slave status\")\n slave = res.get('numrows', 0) > 0\n if slave:\n self._ptrep.append(\"--check-slave-lag={}\".format(self.conf.host))\n\n return self._ptrep\n\n @property\n def pt_osc_dry_run_args(self):\n \"\"\"Get the Percona Toolkit OSC args for the dry run.\"\"\"\n if not self._ptdrargs:\n self._ptdrargs = []\n\n if self.conf.primary_key:\n self._ptdrargs.append(\"--no-check-alter\")\n\n self._ptdrargs.extend(self.pt_osc_rep)\n\n return self._ptdrargs\n\n @property\n def pt_osc_args(self):\n \"\"\"Get the Percona Toolkit OSC args.\"\"\"\n if not self._ptargs:\n self._ptargs = []\n\n if self.conf.primary_key:\n self._ptargs.append(\"--no-check-alter\")\n\n if self.conf.no_cleanup:\n self._ptargs.extend([\"--no-swap-tables\",\n \"--no-drop-new-table\",\n \"--no-drop-old-table\",\n \"--no-drop-triggers\"])\n\n self._ptargs.extend(self.pt_osc_rep)\n\n return self._ptargs\n\n def confirm(self):\n \"\"\"Ask for confirmation if we want to be bugged.\"\"\"\n if not self.conf.warn:\n return\n answer = input(\"continue? yes/no \")\n if not answer.startswith('y'):\n print('abort')\n sys.exit(0)\n\n def _execute(self, sql):\n \"\"\"Run a query and return wether it was successfull or not.\"\"\"\n print(sql)\n res = self.connection.execute(sql)\n return res.get('success')\n\n def execute(self, sql, args=[]):\n \"\"\"Run a query and return wether it was successfull or not.\n\n An optional list of args can be provided and will be run before.\n \"\"\"\n for arg in args:\n if not self._execute(arg):\n return False\n\n return self._execute(sql)\n\n @property\n def osctool(self):\n \"\"\"Get the pt osc tool.\"\"\"\n if not self._osctool:\n self._osctool = shutil.which('pt-online-schema-change')\n if not self._osctool:\n print(\"Error: Could not find 'pt-online-schema-change'.\")\n sys.exit(1)\n\n return self._osctool\n\n def run_command(self, cmd):\n \"\"\"Run a command with Popen and return True if successful.\"\"\"\n stdout = None if self.conf.debug else subprocess.PIPE\n stderr = None if self.conf.debug else subprocess.PIPE\n process = subprocess.Popen(cmd,\n stdout=stdout,\n stderr=stderr)\n ret_code = process.wait()\n return ret_code == 0\n\n def run_pt_ost_alter(self, db, dry_run=False):\n \"\"\"Run the percona ost alter on the given db.\"\"\"\n cmd = [self.osctool,\n \"--critical-load\", \"Threads_running=400\",\n \"--max-load\", \"Threads_running=300\",\n \"--alter-foreign-keys-method=none\", \"--force\",\n \"--nocheck-replication-filters\",\n \"--no-version-check\"]\n\n if dry_run:\n cmd.append(\"--dry-run\")\n cmd.extend(self.pt_osc_dry_run_args)\n else:\n cmd.append(\"--execute\")\n cmd.extend(self.pt_osc_args)\n\n cmd.extend([\"--alter\", '{}'.format(self.conf.altersql),\n \"D={},t={},h={},P={},u={}\".format(db,\n self.conf.table,\n self.conf.host,\n self.conf.port,\n self.conf.user)])\n\n return self.run_command(cmd)\n\n def run_pt_cleanup(self, db):\n \"\"\"Run a cleanup on a given db after the percona osc has been run.\"\"\"\n print(\"Ready for cleanup. Will do this:\")\n sql = \"rename table {0} to _{0}_done, _{0}_new to {0}\".format(self.conf.table)\n print(sql)\n self.confirm()\n self.execute(sql, self.ddl_rep)\n\n sql = \"drop trigger if exists pt_osc_{}_{}_upd\".format(db, self.conf.table)\n self.execute(sql, self.ddl_rep)\n\n sql = \"drop trigger if exists pt_osc_{}_{}_del\".format(db, self.conf.table)\n self.execute(sql, self.ddl_rep)\n\n sql = \"drop table if exists _{}_done\".format(self.conf.table)\n self.execute(sql, self.ddl_rep)\n\n def show_conf(self):\n \"\"\"Show the configuration to be used.\"\"\"\n print(\"Host : {}\".format(self.conf.host))\n print(\"Port : {}\".format(self.conf.port))\n print(\"Databases : {}\".format(self.conf.dblist))\n print(\"Table : {}\".format(self.conf.table))\n print(\"Alter SQL : {}\".format(self.conf.altersql))\n print(\"method : {}\".format(self.conf.method))\n print(\"pt dry args : {}\".format(self.pt_osc_dry_run_args))\n print(\"pt args : {}\".format(self.pt_osc_args))\n print(\"ddl args : {}\".format(self.ddl_args))\n print(\"analyze : {}\".format(self.conf.analyze))\n\n def change_database(self, db):\n \"\"\"Change the database and exit on fail.\"\"\"\n self.connection.change_database(db)\n if self.connection.database != db:\n print(\"Error: Could not change to '{}'.\".format(db))\n sys.exit(1)\n print(\"host: {}, database: {}\".format(self.conf.host, db))\n\n def check_collision(self):\n \"\"\"Check for table collisions and ask for confirmation if so.\"\"\"\n new_table = '_{}_new'.format(self.conf.table)\n res = self.connection.execute(\"show tables like '{}'\".format(new_table))\n if res.get('numrows', 0) > 0:\n print(\"{} already exists!\".format(new_table))\n self.confirm()\n\n def run_percona(self, db):\n \"\"\"Perform the operation on the given db with the percona method.\"\"\"\n dry_run_ret = self.run_pt_ost_alter(db, dry_run=True)\n if dry_run_ret:\n actual_run_ret = self.run_pt_ost_alter(db)\n if actual_run_ret:\n if self.conf.no_cleanup:\n self.run_pt_cleanup(db)\n\n if self.conf.analyze:\n sql = \"analyze table {};\".format(self.conf.table)\n self.execute(sql, self.ddl_rep)\n else:\n print(\"WARNING {} : {} encountered problems\".format(db, self.conf.table))\n return False\n else:\n print(\"SKIPPING {} : {} dry-run encountered problems\".format(db, self.conf.table))\n return False\n\n return True\n\n def run_ddl(self, db):\n \"\"\"Perform the operation on the given db with the ddl or ddlonline method.\"\"\"\n table = self.conf.table\n alter = self.conf.altersql\n if self.conf.method == \"ddl\":\n sql = \"alter table `{}` {}\".format(table, alter)\n else:\n sql = \"alter online table `{}` {}\".format(table, alter)\n\n if not self.execute(sql, self.ddl_args):\n print(\"WARNING {} encountered problems while being executed at {}.{}\"\n .format(alter, db, table))\n return False\n\n return True\n\n def run(self):\n \"\"\"Perform the online schema change operation.\"\"\"\n self.show_conf()\n self.confirm()\n\n for db in self.conf.dblist:\n self.change_database(db)\n\n success = True\n if self.conf.method == 'percona':\n self.check_collision()\n success = self.run_percona(db)\n elif self.conf.method in (\"ddl\", \"ddlonline\"):\n\n success = self.run_ddl(db)\n\n if not success and self.conf.warn:\n self.confirm()\n\n\ndef parse_args():\n \"\"\"Parse the execution parameters and return an object with them.\"\"\"\n parser = argparse.ArgumentParser()\n\n parser.add_argument('--host',\n help=\"the host to connect to\",\n required=True)\n parser.add_argument('--port', type=int,\n help=\"the port to connect to\",\n default=3306)\n\n parser.add_argument('--user', help=\"username to use\",\n default='root')\n\n dblist_group = parser.add_mutually_exclusive_group(required=True)\n dblist_group.add_argument('--db', nargs='+',\n help=\"Database(s) to be altered\")\n dblist_group.add_argument('--dblist',\n help=\"File with the list of databases\")\n\n parser.add_argument('--table',\n help=\"Table to alter\",\n required=True)\n\n methods = ['percona', 'ddl', 'ddlonline']\n parser.add_argument('--method',\n help=\"Method to use ({})\".format(', '.join(methods)),\n choices=methods,\n default='percona')\n\n warn_group = parser.add_mutually_exclusive_group()\n warn_group.add_argument('--warn', action='store_true', dest='warn',\n help=\"Ask for confirmation after a problem\")\n warn_group.add_argument('--no-warn', action='store_false', dest='warn',\n help=\"Don't ask for confirmation after a problem\")\n\n analyze_group = parser.add_mutually_exclusive_group()\n analyze_group.add_argument('--analyze',\n action='store_true', dest='analyze',\n help=\"Analyze after the alter\")\n analyze_group.add_argument('--no-analyze',\n action='store_false', dest='analyze',\n help=\"Don't analyze after the alter\")\n\n parser.add_argument('altersql', nargs=\"+\",\n help=\"Modification to be applied\")\n\n parser.add_argument('--debug',\n action='store_true',\n help=\"Show debug info\")\n\n parser.set_defaults(warn=True)\n parser.set_defaults(analyze=False)\n\n replicate_group = parser.add_mutually_exclusive_group()\n replicate_group.add_argument('--replicate', action='store_true',\n help=\"Replicate the changes\")\n replicate_group.add_argument('--no-replicate', action='store_true',\n help=\"Don't replicate the changes\")\n\n def valid_gtid(value):\n pat = re.compile(r\"[0-9]+\")\n if not pat.match(value):\n error_msg = \"'{}' is an invalid gtid\".format(value)\n raise argparse.ArgumentTypeError(error_msg)\n return value\n\n parser.add_argument('--gtid_domain_id', type=valid_gtid,\n help=\"gtid domain id\")\n\n parser.add_argument('--primary-key', action='store_true',\n help=\"Don't panic when altering a primary key\")\n parser.add_argument('--no-cleanup', action='store_true',\n help=\"Don't actually switch the new and old tables on completion\")\n\n args = parser.parse_args()\n\n args.altersql = ' '.join(args.altersql)\n\n # Check list of databases is correct\n if args.dblist:\n if not args.dblist.endswith('.dblist'):\n print(\"'{}' doesn't have the 'dblist' extension\".format(args.dblist))\n sys.exit(1)\n\n try:\n with open(args.dblist) as f:\n args.dblist = [l.strip() for l in f if l.strip()]\n except IOError:\n print(\"Can't read '{}'\".format(args.dblist))\n sys.exit(1)\n else:\n args.dblist = args.db\n\n return args\n\n\ndef main():\n \"\"\"Parse de arguments and create the object to perform the operation.\"\"\"\n options = parse_args()\n osc = OnlineSchemaChanger(options)\n osc.run()\n\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"wmfmariadbpy/osc_host.py","file_name":"osc_host.py","file_ext":"py","file_size_in_byte":15070,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"240740325","text":"import os ,django\nos.environ.setdefault(\"DJANGO_SETTINGS_MODULE\",\"Excelproject.settings\")\ndjango.setup()\n\n\nfrom home.models import Student\nfrom django.utils import timezone\nfrom faker import Faker\nfake=Faker()\n\ndef populate(n):\n for _ in range(n):\n fakename=fake.name()\n fakeemail=fake.email()\n fakeadress= fake.address()\n stud=Student.objects.get_or_create(name=fakename, email=fakeemail, adress=fakeadress) \npopulate(20000)\nprint('data is populate succes')","sub_path":"populate_scripts.py","file_name":"populate_scripts.py","file_ext":"py","file_size_in_byte":517,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"33862552","text":"# MIT licensed\n# Copyright (c) 2013-2017 lilydjwg , et al.\n\nimport os\nimport pytest\npytestmark = [pytest.mark.asyncio,\n pytest.mark.skipif(\"NVCHECKER_GITHUB_TOKEN\" not in os.environ,\n reason=\"requires NVCHECKER_GITHUB_TOKEN, or it fails too much\")]\n\nasync def test_github(get_version):\n assert await get_version(\"example\", {\"github\": \"harry-sanabria/ReleaseTestRepo\"}) == \"20140122.012101\"\n\nasync def test_github_default_not_master(get_version):\n assert await get_version(\"example\", {\"github\": \"MariaDB/server\"}) is not None\n\nasync def test_github_latest_release(get_version):\n assert await get_version(\"example\", {\"github\": \"harry-sanabria/ReleaseTestRepo\", \"use_latest_release\": 1}) == \"release3\"\n\nasync def test_github_max_tag(get_version):\n assert await get_version(\"example\", {\"github\": \"harry-sanabria/ReleaseTestRepo\", \"use_max_tag\": 1}) == \"second_release\"\n\nasync def test_github_max_tag_with_ignored_tags(get_version):\n assert await get_version(\"example\", {\"github\": \"harry-sanabria/ReleaseTestRepo\", \"use_max_tag\": 1, \"ignored_tags\": \"second_release release3\"}) == \"first_release\"\n","sub_path":"tests/test_github.py","file_name":"test_github.py","file_ext":"py","file_size_in_byte":1171,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"569629902","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\nfrom setuptools import setup, find_packages\nimport os\n\nimport re\n\nverstr = \"unknown\"\ntry:\n verstrline = open('ando/_version.py', \"rt\").read()\nexcept EnvironmentError:\n pass # Okay, there is no version file.\nelse:\n VSRE = r\"^__version__ = ['\\\"]([^'\\\"]*)['\\\"]\"\n mo = re.search(VSRE, verstrline, re.M)\n if mo:\n verstr = mo.group(1)\n else:\n raise RuntimeError(\"unable to find version in yourpackage/_version.py\")\n\n\n\nsetup(\n name=\"ando\",\n version=verstr,\n packages=find_packages(),\n author=\"Jeremy Garcia, Sylvain Takerkart\",\n description=\"Checks the validity of a directory with respect to the ANimal Data Organization (ANDO) specifications \",\n license='MIT',\n install_requires=[]\n)\n","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":780,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"439289415","text":"from django.db import models\n\n# Create your models here.\n\n\nfrom django.db import models\n\n# Create your models here.\nfrom django.db import models\nfrom django.core import validators\nfrom django.conf import settings\n\n\n# создаем класс пользователя\n\nclass Profile(models.Model):\n JOB_TITLES_CHOICES = [\n ('director', 'Директор'),\n ('manager', 'Менеджер'),\n ('accountant', 'Бухгалтер'),\n ('cashier', 'Кассир'),\n ]\n\n name = models.CharField(max_length=100, verbose_name='Имя')\n user = models.OneToOneField(settings.AUTH_USER_MODEL, on_delete=models.CASCADE, null=False,\n verbose_name='Пользователь')\n phone = models.CharField(max_length=20, verbose_name='Телефон')\n email = models.EmailField(max_length=254, verbose_name='Электронная почта')\n job_title = models.CharField(max_length=100, verbose_name='Должность', choices=JOB_TITLES_CHOICES,\n default='cashier')\n is_active = models.BooleanField(default=True, verbose_name='Активный')\n\n class Meta:\n db_table = \"profiles\"\n verbose_name = 'Профиль'\n verbose_name_plural = 'Профили'\n\n def __str__(self):\n return f\"{self.name}\"\n\n\nclass OperationName(models.Model):\n OPERATION_TYPES = [\n ('plus', 'Дебит'),\n ('minus', 'Кредит'),\n ]\n\n objects = models.Manager()\n name = models.CharField(max_length=255, verbose_name='Название операции')\n type = models.CharField(\n max_length=5,\n choices=OPERATION_TYPES,\n default='plus',\n verbose_name='Тип операции'\n )\n\n class Meta:\n db_table = \"OperationName\"\n verbose_name = 'Имя операции'\n verbose_name_plural = 'Имена операций'\n\n def __str__(self):\n return f\"{self.name}\"\n\n\nclass PaymentMethod(models.Model):\n PAYMENT_METHOD_TYPES = [\n ('cash', 'Наличный'),\n ('account', 'Безналичный'),\n ]\n\n name = models.CharField(max_length=32, verbose_name='Метод оплаты')\n type = models.CharField(max_length=20,\n choices=PAYMENT_METHOD_TYPES,\n default='cash',\n verbose_name='Нал/Безнал')\n\n class Meta:\n db_table = \"payment_methods\"\n verbose_name = 'Способ оплаты'\n verbose_name_plural = 'Способы оплаты'\n\n def __str__(self):\n return f\"{self.name}\"\n\n\nclass Operation(models.Model):\n objects = models.Manager()\n sale = models.ForeignKey('Sale', on_delete=models.CASCADE, null=True, verbose_name='Продажа')\n operation_name = models.ForeignKey(OperationName, on_delete=models.PROTECT, null=True, verbose_name='Операция')\n amount = models.IntegerField(validators=[validators.MinValueValidator(0)], verbose_name='Количество')\n price = models.DecimalField(max_digits=8, decimal_places=2, verbose_name='Сумма')\n\n def __str__(self):\n return f\"{self.operation_name} {self.price} {self.amount}\"\n\n class Meta:\n db_table = \"operations\"\n verbose_name = 'Операция'\n verbose_name_plural = 'Операции'\n\n\nclass Sale(models.Model):\n objects = models.Manager()\n date = models.DateTimeField(auto_now_add=True, verbose_name='Дата создания')\n update_date = models.DateTimeField(auto_now=True, verbose_name='Дата последнего изменения')\n profile = models.ForeignKey(Profile, on_delete=models.PROTECT, verbose_name='Профиль')\n payment = models.ForeignKey(PaymentMethod, on_delete=models.PROTECT, null=False, verbose_name='Способ оплаты')\n notice = models.TextField(blank=True, null=True, verbose_name='Описание')\n total = models.DecimalField(max_digits=8, decimal_places=2, verbose_name='Сумма', null=True)\n verbose_date = models.CharField(max_length=100, verbose_name='Дата', null=True, blank=True)\n def calculate_total(self):\n total = 0\n for operation in self.operation_set.all():\n total += operation.price * operation.amount\n return total\n\n def calculate_verbose_date(self):\n date_string = str(self.date)\n result = date_string.split('.')[0]\n return result\n\n\n\n def __str__(self):\n return f\"{self.pk} - {self.verbose_date} / {self.total}\"\n\n class Meta:\n db_table = \"sales\"\n verbose_name = 'Продажа'\n verbose_name_plural = 'Продажи'\n","sub_path":"uchet/kassa/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":4700,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"186816625","text":"def rgb2hsv(rgb): #RGB转HSV,参数是序列\r\n r=rgb[0]/255\r\n g=rgb[1]/255\r\n b=rgb[2]/255\r\n rgb_max=max(r,g,b)\r\n rgb_min=min(r,g,b)\r\n\r\n s=0 #饱和度0-1\r\n if rgb_max==0:\r\n s=0\r\n else:\r\n s=(rgb_max-rgb_min)/rgb_max\r\n\r\n v=rgb_max #明度0-1\r\n\r\n h=1 #色相0-360\r\n if rgb_max==rgb_min:\r\n h=0\r\n elif rgb_max==r and g>=b:\r\n h=60*((g-b)/(rgb_max-rgb_min))\r\n elif rgb_max==r and g=min(scope) and A<=max(scope):\r\n return 1\r\n else: \r\n return 0\r\n\r\ndef isHueInHscopes(Hue,Hscopes): #Hscopes=[[a,b],[c,d]..]\r\n In=0\r\n for scope in Hscopes:\r\n if isAInscope(Hue,scope)!=0:\r\n In=1\r\n return In\r\n\r\ndef isHsvInScopes(hsv,scopes): #hsv=[0,0,0],scopes={'h':[[75,165],[280,340]],'s':[0,1],'v':[0,1]}\r\n h=hsv[0]\r\n s=hsv[1]\r\n v=hsv[2]\r\n In=0\r\n if isHueInHscopes(h,scopes['h']) and isAInscope(s,scopes['s']) and isAInscope(v,scopes['v']):\r\n In=1\r\n return In\r\n","sub_path":"tools.py","file_name":"tools.py","file_ext":"py","file_size_in_byte":1363,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"628004615","text":"import math\nwhile True:\n a = float(input('Input a = '))\n b = float(input('Input b = '))\n c = float(input('Input c = '))\n D = b*b - 4*a*c\n if D > 0:\n print('D = ', D, '. D > 0, 2 roots')\n D_up1 = -b + math.sqrt(D)\n D_up2 = -b - math.sqrt(D)\n D_down = 2*a\n print(D_up1, D_up2, D_down)\n x1 = float(D_up1 / D_down)\n x2 = float(D_up2 / D_down)\n print('Roots: ', float(x1), 'and ', float(x2))\n elif D == 0:\n print('D = 0, 1 root')\n x = -b / 2*a\n print('Root: ', x)\n elif D < 0:\n print('D = 0, no roots')\n \"\"\"cont = input('Got another equation? (y/n)')\n if cont == 'y' or cont == 'Y':\n print('Alright, type another one')\n elif cont == 'n' or cont == 'N':\n print('See ya')\n else:\n print(\"Unknown command. Back to equation's question (y/n)\")\n continue\"\"\"\ninput()\n","sub_path":"roots_v1.py","file_name":"roots_v1.py","file_ext":"py","file_size_in_byte":907,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"104229767","text":"import pygame, sys\nfrom classes import *\n\npygame.init()\nSCREENWIDTH, SCREENHEIGHT = 640, 320\nscreen = pygame.display.set_mode((SCREENWIDTH, SCREENHEIGHT))\nclock = pygame.time.Clock()\nFPS = 24\n\nbug = Bug(0, 100, 40, 40, \"resource/bug.png\")\nbug_2 = Bug(0, 200, 40, 40, \"resource/bug.png\")\n#--------- Main game loop -----------\nwhile True:\n\t#PROCESSING\n\tfor event in pygame.event.get():\n\t\tif event.type == pygame.QUIT:\n\t\t\tpygame.quit()\n\t\t\tsys.exit()\n\n\t#PROCESSING\n\t#LOGIC\n\tbug.motion()\n\t#LOGIC\n\t#DRAW\n\tscreen.fill((0,0,0))\n\tBaseClass.all_sprites.draw(screen)\n\tpygame.display.flip()\n\t#DRAW\n\n\tclock.tick(FPS)","sub_path":"tut_7_bug_class.py","file_name":"tut_7_bug_class.py","file_ext":"py","file_size_in_byte":603,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"300873521","text":"import unittest\nfrom frontend.object.ObjectFactory import ObjectFactory\n\n\nclass TestCase006(unittest.TestCase):\n def test_exception(self):\n attributes = {'name': 'osman1', 'type': 'osman'}\n with self.assertRaises(TypeError):\n ObjectFactory(attributes).create()\n\n\nif __name__ == '__main__':\n suite = unittest.TestLoader().loadTestsFromTestCase(TestCase006)\n unittest.TextTestRunner(verbosity=1).run(suite)","sub_path":"tests/python/testCase006.py","file_name":"testCase006.py","file_ext":"py","file_size_in_byte":438,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"235144030","text":"class Solution(object):\n\n def convert(self, s, numRows):\n \"\"\"\n The string \"PAYPALISHIRING\" is written in a zigzag pattern on a given number of rows like this: (you may want to display this pattern in a fixed font for better legibility)\n\n P A H N\n A P L S I I G\n Y I R\n And then read line by line: \"PAHNAPLSIIGYIR\"\n\n Write the code that will take a string and make this conversion given a number of rows.\n\n Constraints:\n 1 <= s.length <= 1000\n s consists of English letters (lower-case and upper-case), ',' and '.'.\n 1 <= numRows <= 1000\n\n Runtime: 116 ms, faster than 19.05% of Python3 online submissions for ZigZag Conversion.\n Memory Usage: 14.4 MB, less than 45.72% of Python3 online submissions for ZigZag Conversion.\n\n\n Parameters\n ----------\n s : str\n numRows : int\n\n\n Returns\n -------\n ret : str\n\n\n Examples\n --------\n >>> Solution().convert(\"PAYPALISHIRING\", numRows=3)\n \"PAHNAPLSIIGYIR\"\n\n >>> Solution().convert(\"PAYPALISHIRING\", numRows=4)\n \"PINALSIGYAHRPI\"\n\n >>> Solution().convert(\"A\", numRows=1)\n \"A\"\n\n >>> Solution().convert(\"AB\", numRows=1)\n \"AB\"\n\n \"\"\"\n\n if numRows >= len(s) or numRows == 1:\n return s\n\n # 14, 3 = 2*3-2=4\n # 14, 4 = 2*4-2=6\n step = 2 * numRows - 2\n\n # [ \"\",\n # \"\",\n # \"\",\n # \"\" ]\n lines = ['']*numRows\n for idx, char in enumerate(s):\n # 5 % 6 = 5\n # 6 - 5 = 1\n # if index is > numRows, get offset from numRows\n # 0%2, 2-0%2\n lines[min(idx % step, step - idx % step)] += char\n return ''.join(lines)\n","sub_path":"algorithms/6_ZigZag_Conversion.py","file_name":"6_ZigZag_Conversion.py","file_ext":"py","file_size_in_byte":1811,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"100649207","text":"from guizero import App, Combo\n\ndef selected(value):\n print(value)\n\napp = App()\ncombo = Combo(app, [\"Nothing\", \"Something\"], command = selected, selected=\"Something\")\ncombo.append(\"Everything\")\ncombo2 = Combo(app, [\"hi\", \"bye\"])\napp.display()","sub_path":"examples/combo.py","file_name":"combo.py","file_ext":"py","file_size_in_byte":245,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"370324012","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Dec 27 15:02:45 2018\n\n@author: deepak\n\"\"\"\n\nimport csv as csv\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.markers as markers\nfrom mpl_toolkits.mplot3d import Axes3D\nimport scipy.interpolate as interpolate\nimport scipy.signal as signal\nimport scipy.ndimage as ndimage\nimport matplotlib.patches as patches\nfrom matplotlib.lines import Line2D\nfrom numpy.polynomial import polynomial as Poly\nimport cmocean\nimport pickle\nimport math\nimport os\nimport pandas as pd\nimport math\nfrom mpl_toolkits.axes_grid1.anchored_artists import AnchoredSizeBar\nimport cycler\nplt.close(\"all\")\nimport imp\nimport Track\nimport GravityMachineTrack\nimport FigureParameters\nimp.reload(Track)\nimport seaborn as sns\nimport scipy.stats as stats\n#plt.style.use(os.path.join(os.getcwd(),'gravmachinestyle.mplstyle'))\n#==============================================================================\n# Plot Parameters and Functions \n#==============================================================================\nfrom matplotlib import rcParams\nfrom matplotlib import rc\n#rcParams['axes.titlepad'] = 20 \n#rc('font',**{'family':'sans-serif','sans-serif':['Helvetica']})\n### for Palatino and other serif fonts use:\n##rc('font',**{'family':'serif','serif':['Palatino']})\n#rc('text', usetex=False)\n#plt.rc('font', family='serif')\n\nrc('font', family='sans-serif') \nrc('font', serif='Helvetica') \nrc('text', usetex='false') \nrcParams.update({'font.size': 18})\n#==============================================================================\ndef find_nearest(array, value):\n array = np.asarray(array)\n idx = (np.abs(array - value)).argmin()\n return array[idx]\n\n#dataFolder ='/Volumes/GRAVMACH1/VolvoxPhotoresponse/vvx25'\n\ndataFolder = 'E:/VolvoxPhotoresponse/vvx25'\n\n\nfile = \"track.csv\"\n\n*rest,orgName = os.path.split(dataFolder)\nsaveFolder = '/Users/deepak/Dropbox/GravityMachine/GravityMachineManuscript/EnsembleTrackStatistics'\n\n\n\n\nprint(50*'*')\nprint('Folder in which to save data: {}'.format(dataFolder))\nprint(50*'*')\n\n\n#======================================================================\n# Plots of Tracks\n#======================================================================\nTmin = 0\nTmax = 4030\n#OrgTrack1 = Track.GravMachineTrack(dataFolder,file,Tmin,Tmax)\nOrgTrack1 = GravityMachineTrack.gravMachineTrack(trackFile = os.path.join(dataFolder, file), Tmin = Tmin, Tmax = Tmax)\n\n#ImageFolder = '/Volumes/GRAVMACH1/VolvoxPhotoresponse/vvx25/images00012'\n#saveFolder = '/Users/deepak/Dropbox/GravityMachine/GravityMachineManuscript/Figures/Figure6/Volvox_RisingEdge'\n\nImageFolder = 'E:/VolvoxPhotoresponse/vvx25/images00012'\n\n\n#if (not os.path.exists(saveFolder)):\n# os.makedirs(saveFolder)\n \n#OrgTrack1.makeMovie(ImageFolder, saveFolder)\n\n\n\ncustom_line = [Line2D([0], [0], color='gold', lw=4), Line2D([0], [0], color='k', lw=4) ]\n\n\nRisingEdge = []\nFallingEdge = []\ncounter = 0\nfor ii in range(OrgTrack1.trackLen - 1):\n if(OrgTrack1.LED_intensity[ii]==0 and OrgTrack1.LED_intensity[ii+1]>0):\n \n RisingEdge.append(ii)\n counter += 1\n \n elif(OrgTrack1.LED_intensity[ii]>0 and OrgTrack1.LED_intensity[ii+1]==0):\n FallingEdge.append(ii)\n \nRisingEdge = np.array(RisingEdge)\nFallingEdge = np.array(FallingEdge)\n\nIndex = []\nIndex_falling = []\nfor ii in range(len(RisingEdge)-1):\n \n if(RisingEdge[ii+1] - RisingEdge[ii] > 100):\n Index.append(ii+1)\n \n \nRisingEdge = RisingEdge[Index]\n\nfor ii in range(len(FallingEdge)):\n \n nearestRisingEdge = find_nearest(RisingEdge, FallingEdge[ii])\n \n if(abs(FallingEdge[ii]-nearestRisingEdge) > 200):\n Index_falling.append(ii)\n\n \nFallingEdge = FallingEdge[Index_falling]\n\nnBlinks = len(RisingEdge)\n# Mid piint indcices between Rising and Falling edges where there is no trabnsition\nnoTransitionIndex = [int((FallingEdge[ii+1] + RisingEdge[ii])/2) for ii in range(nBlinks-1)]\n\nprint(noTransitionIndex)\n\n\n\nZ_movingAvg = OrgTrack1.smoothSignal(OrgTrack1.ZobjWheel,20)\n\n\n## Moving avg of the vertical velocity\n#Tmin = 0\n#Tmax = 4030\n#mask = np.array(OrgTrack1.Time>=Tmin, dtype='bool') & np.array(OrgTrack1.Time<=Tmax, dtype = 'bool')\n\n#------------------------------------------------------------------------------\n# Volvox phototaxis plots\n#------------------------------------------------------------------------------\n# Plots of displacement vs time\n#------------------------------------------------------------------------------\n#fig, ax1 = plt.subplots(figsize=(6,4))\n#\n#ax1.fill_between(OrgTrack1.Time[mask],y1=np.max(OrgTrack1.ZobjWheel[mask])+0.1,y2 = np.min(OrgTrack1.ZobjWheel[mask])-0.1, where = OrgTrack1.LED_intensity[mask]>0,facecolor = 'gold', alpha = 0.4)\n#ax1.fill_between(OrgTrack1.Time[mask],y1=np.max(OrgTrack1.ZobjWheel[mask])+0.1,y2 = np.min(OrgTrack1.ZobjWheel[mask])-0.1,where = OrgTrack1.LED_intensity[mask]==0,facecolor = 'k', alpha = 0.4)\n#ax1.plot(OrgTrack1.Time[mask], OrgTrack1.ZobjWheel[mask] , 'b-', linewidth = 2)\n##ax1.plot(OrgTrack1.Time[0:-1], OrgTrack1.Vz , 'b', linewidth = 1)\n#\n#\n#\n# \n##ax1.set_xlabel('Time (s)')\n##ax1.set_ylabel('Z (mm)', color='b')\n#\n##ax1.legend(custom_line, ['Light ON', 'Light OFF'])\n#ax1.set_aspect(2.5)\n##ax1.set_xticks(range(Tmin, Tmax,5))\n#ax1.set_xlim([np.min(OrgTrack1.Time[mask]), np.max(OrgTrack1.Time[mask])])\n#ax1.set_ylim([np.min(OrgTrack1.ZobjWheel[mask])-0.1, np.max(OrgTrack1.ZobjWheel[mask])+0.1])\n#plt.show()\n# \n#------------------------------------------------------------------------------\n\nZ_fast = OrgTrack1.ZobjWheel - Z_movingAvg\n\n\n\n#------------------------------------------------------------------------------\n# Plot of the Signal - Moving average\n#------------------------------------------------------------------------------\nfig, ax1 = plt.subplots()\n\n \nax1.fill_between(OrgTrack1.T,y1=np.max(OrgTrack1.ZobjWheel),y2 =np.min(OrgTrack1.ZobjWheel), where = OrgTrack1.LED_intensity>0,facecolor = 'gold', alpha = 0.4)\nax1.fill_between(OrgTrack1.T,y1=np.max(OrgTrack1.ZobjWheel),y2 =np.min(OrgTrack1.ZobjWheel),where = OrgTrack1.LED_intensity==0,facecolor = 'k', alpha = 0.4)\nax1.plot(OrgTrack1.T, OrgTrack1.ZobjWheel , 'b-', linewidth = 2)\nax1.plot(OrgTrack1.T[FallingEdge], Z_fast[FallingEdge], 'ro')\nax1.plot(OrgTrack1.T[RisingEdge], Z_fast[RisingEdge], 'go')\nax1.plot(OrgTrack1.T[noTransitionIndex], Z_fast[noTransitionIndex], 'bo')\n\nax1.set_xlabel('Time (s)')\nax1.set_ylabel('Z (mm)', color='b')\n\nax1.legend(custom_line, ['Light ON', 'Light OFF'])\nax1.set_xlim([0, np.max(OrgTrack1.T)])\nax1.set_ylim([np.min(OrgTrack1.ZobjWheel), np.max(OrgTrack1.ZobjWheel)])\nplt.show()\n#\n##------------------------------------------------------------------------------\n## Plot of the Vertical velocity vs Time\n##------------------------------------------------------------------------------\n\n#Vz_smooth = OrgTrack1.smoothSignal(OrgTrack1.Vz,OrgTrack1.window_time)\n#\n#Vz_smooth = np.insert(Vz_smooth,0,np.nan)\n#\n#fig, ax1 = plt.subplots(figsize=(6,4))\n#\n#\n# \n#ax1.fill_between(OrgTrack1.Time[mask],y1=np.nanmax(Vz_smooth[mask])+0.1,y2 = np.nanmin(Vz_smooth[mask])-0.1, where = OrgTrack1.LED_intensity[mask]>0,facecolor = 'gold', alpha = 0.4)\n#ax1.fill_between(OrgTrack1.Time[mask],y1=np.nanmax(Vz_smooth[mask])+0.1,y2 = np.nanmin(Vz_smooth[mask])-0.1,where = OrgTrack1.LED_intensity[mask]==0,facecolor = 'k', alpha = 0.4)\n#ax1.plot(OrgTrack1.Time[mask], Vz_smooth[mask] , 'k', linewidth = 2)\n#ax1.plot(OrgTrack1.Time[0:-1], OrgTrack1.Vz , 'b', linewidth = 1)\n#\n#\n#\n# \n#ax1.set_xlabel('Time (s)')\n#ax1.set_ylabel('Vz (mm)', color='b')\n#\n##ax1.legend(custom_line, ['Light ON', 'Light OFF'])\n#ax1.set_xlim([np.nanmin(OrgTrack1.Time[mask]), np.nanmax(OrgTrack1.Time[mask])])\n#ax1.set_ylim([np.nanmin(Vz_smooth[mask])-0.1, np.nanmax(Vz_smooth[mask])+0.1])\n##ax1.set_xticks(range(Tmin, Tmax,5))\n#plt.show()\n\n#------------------------------------------------------------------------------\n# Create a dataframe with all the behavioral events aligned at the Rising Edge and Falling Edge\n#------------------------------------------------------------------------------\nT_window = 10\n\nnWindow = math.ceil(int(T_window*OrgTrack1.samplingFreq)/2.)*2\nT_blink = np.zeros(nWindow)\n\n\n\nprint('No:of blinks: {}'.format(nBlinks))\ncolor = plt.cm.inferno(np.linspace(0, 1,nBlinks))\n \n# color = cmocean.cm.algae(np.linspace(0, 1,nBlinks))\n\nplt.rcParams['axes.prop_cycle'] = cycler.cycler('color', color)\n\n\n##------------------------------------------------------------------------------\n## Create a dataframe with all the behavioral events aligned at the Rising Edge\n##------------------------------------------------------------------------------\n\nVz_blink = np.zeros((nBlinks,nWindow),dtype='float')\n# Save the mean and std of velocity in the neighbhourhood of the blink\nVz_DL_mean = np.zeros(nBlinks)\nVz_DL_std = np.zeros(nBlinks)\nplt.figure(3)\nfor i,peakIndex in enumerate(RisingEdge):\n \n lower_index = peakIndex\n upper_index = peakIndex + int(nWindow)\n if(lower_index>=0 and upper_index=0 and upper_index=0 and upper_index0,facecolor = 'gold', alpha = 0.5)\n## ax1.fill_between(OrgTrack1.Time,y1=np.max(Z_fast),y2 =np.min(Z_fast),where = OrgTrack1.LED_intensity==0,facecolor = 'k', alpha = 0.2)\n## ax1.plot(OrgTrack1.Time, OrgTrack1.ZobjWheel , 'b-', linewidth = 2)\n# \n# ax1.plot(OrgTrack1.Time, OrgTrack1.ZobjWheel , 'b-', linewidth = 2)\n#\n# ax2 = ax1.twinx()\n# \n# ax2.plot(OrgTrack1.Time, OrgTrack1.LED_intensity, color = 'gold')\n# \n# \n# ax1.set_xlabel('Time (s)')\n# ax1.set_ylabel('V_Z (mm/s)', color='b')\n# ax2.set_ylabel('LED intensity')\n# \n## ax1.legend(custom_line, ['Light ON', 'Light OFF'])\n## ax1.set_xlim([0, np.max(OrgTrack1.Time)])\n## ax1.set_ylim([np.min(Z_fast), np.max(Z_fast)])\n# plt.show()\n# \n# fig, ax1 = plt.subplots()\n#\n# \n## ax1.fill_between(OrgTrack1.Time,y1=np.max(Z_fast),y2 =np.min(Z_fast), where = OrgTrack1.LED_intensity>0,facecolor = 'gold', alpha = 0.5)\n## ax1.fill_between(OrgTrack1.Time,y1=np.max(Z_fast),y2 =np.min(Z_fast),where = OrgTrack1.LED_intensity==0,facecolor = 'k', alpha = 0.2)\n## ax1.plot(OrgTrack1.Time, OrgTrack1.Xobj , 'r-', linewidth = 2)\n# \n# ax1.plot(OrgTrack1.Time, OrgTrack1.Vx , 'b-', linewidth = 2)\n#\n# ax2 = ax1.twinx()\n# \n# ax2.plot(OrgTrack1.Time, OrgTrack1.LED_intensity, color = 'gold')\n# \n# \n# ax1.set_xlabel('Time (s)')\n# ax1.set_ylabel('V_x (mm/s)', color='b')\n# ax2.set_ylabel('LED intensity')\n# \n## ax1.legend(custom_line, ['Light ON', 'Light OFF'])\n## ax1.set_xlim([0, np.max(OrgTrack1.Time)])\n## ax1.set_ylim([np.min(Z_fast), np.max(Z_fast)])\n# plt.show()","sub_path":"DataAnalysisScripts/PhotoResponse.py","file_name":"PhotoResponse.py","file_ext":"py","file_size_in_byte":19368,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"155794511","text":"\"\"\"\nA class I created to construct my neural networks\n\"\"\"\n\nfrom __future__ import print_function\nfrom math import sqrt\nimport numpy as np\nimport tensorflow as tf\nimport csv\n\n# Defines MLP (MultiLayer Perceptron) using tensorflow\nclass MLP(object):\n\t# Network shape : [num_neurons_input_layer, num_neurons_hidden_layer_1, num_neurons_hidden_layer_2, ... ,num_neurons_output_layer]\n\tdef __init__(self, network_shape, initial_learning_rate, decay_steps, decay_rate, regularization_parameter, dropout_keep_prob = 0.5):\n\t\tself.network_shape = network_shape\n\t\tself.num_layers = len(self.network_shape)\n\t\tself.initial_learning_rate = initial_learning_rate\n\t\tself.decay_steps = decay_steps\n\t\tself.decay_rate = decay_rate\n\t\tself.regularization_parameter = regularization_parameter\n\t\tself.dropout_keep_prob = dropout_keep_prob\n\t\tself.create_network()\n\t\tself.output_file_path = '../results/results.csv'\n\t\t\n\tdef create_network(self):\n\t\tself.graph = tf.Graph()\n\t\twith self.graph.as_default():\n\t\t\t# Input\n\t\t\tself.tf_dataset = tf.placeholder(tf.float32, shape=(None, self.network_shape[0]))\n\t\t\tself.tf_labels = tf.placeholder(tf.float32, shape=(None, self.network_shape[-1]))\n\n\t\t\t# Dropout keep probability (set to 1.0 for validation and test)\n\t\t\tself.keep_prob = tf.placeholder(tf.float32)\n\n\t\t\t# Variables\n\t\t\tself.weights = []\n\t\t\tself.biases = []\n\n\t\t\t# Constructs the network according to the given shape array\n\t\t\tfor i in range(self.num_layers-1):\n\t\t\t\tself.weights.append(tf.Variable(tf.truncated_normal([self.network_shape[i], self.network_shape[i+1]],stddev=1.0/(self.network_shape[i]))))\n\t\t\t\tself.biases.append(tf.Variable(tf.zeros([self.network_shape[i+1]])))\n\n\t\t\t# Global Step for learning rate decay\n\t\t\tglobal_step = tf.Variable(0)\n\n\t\t\t# Training computation (with dropout)\n\t\t\tlogits = tf.matmul(self.tf_dataset, self.weights[0]) + self.biases[0]\n\t\t\tfor i in range(1,self.num_layers-2):\n\t\t\t\twith tf.name_scope(\"layer_\"+str(i)) as scope:\n\t\t\t\t\tlogits = tf.matmul(tf.nn.dropout(tf.nn.relu(logits), self.keep_prob), self.weights[i]) + self.biases[i]\n\t\t\t\n\t\t\tlogits = tf.matmul(logits, self.weights[-1]) + self.biases[-1]\n\n\t\t\t# Cross entropy loss\n\t\t\tself.loss = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(logits, self.tf_labels))\n\n\t\t\t# L2 Regularization\n\t\t\tregularizers = tf.nn.l2_loss(self.weights[0]) + tf.nn.l2_loss(self.biases[0])\n\t\t\tfor i in range(1,self.num_layers-1):\n\t\t\t\tregularizers += tf.nn.l2_loss(self.weights[i]) + tf.nn.l2_loss(self.biases[i])\n\n\t\t\tself.loss += self.regularization_parameter * regularizers\n\n\t\t\tlearning_rate = self.initial_learning_rate #tf.train.exponential_decay(self.initial_learning_rate, global_step, self.decay_steps, self.decay_rate)\n\n\t\t\t# Passing global_step to minimize() will increment it at each step.\n\t\t\tself.optimizer = tf.train.AdamOptimizer(learning_rate).minimize(self.loss, global_step=global_step)\n\n\t\t\t# Predictions for the training, validation, and test data.\n\t\t\tself.prediction = tf.nn.softmax(logits)\n\t\t\t\n\tdef train(self, train_dataset, train_labels, valid_dataset, valid_labels, test_dataset, test_labels, batch_size, num_steps):\n\t\told_valid_accuracy = None\n\t\twith tf.Session(graph=self.graph) as session:\n\t\t\ttf.initialize_all_variables().run()\n\t\t\tprint(\"Initialized \", str(self.network_shape))\n\t\t\tfor step in range(num_steps+1):\n\t\t\t\t# Pick an offset within the training data, which has been randomized.\n\t\t\t\toffset = (step * batch_size) % (train_labels.shape[0] - batch_size)\n\t\t\t\t# Generate a minibatch.\n\t\t\t\tbatch_data = train_dataset[offset:(offset + batch_size), :]\n\t\t\t\tbatch_labels = train_labels[offset:(offset + batch_size), :]\n\n\t\t\t\tfeed_dict = {self.tf_dataset : batch_data, self.tf_labels : batch_labels, self.keep_prob : self.dropout_keep_prob}\n\t\t\t\t_, l, predictions = session.run([self.optimizer, self.loss, self.prediction], feed_dict=feed_dict)\n\t\t\t\t\t\n\t\t\t\tif (step % 100 == 0):\n\t\t\t\t\tprint(\"Minibatch loss at step %d: %f\" % (step, l))\n\t\t\t\t\tprint(\"Minibatch accuracy: %.1f%%\" % MLP.accuracy(session.run(self.prediction, feed_dict={self.tf_dataset : batch_data, self.tf_labels : batch_labels, self.keep_prob : 1.0}), batch_labels))\n\n\t\t\t\t\tvalid_prediction = session.run(self.prediction, feed_dict={self.tf_dataset : valid_dataset, self.tf_labels : valid_labels, self.keep_prob : 1.0})\n\t\t\t\t\tvalid_accuracy = MLP.accuracy(valid_prediction, valid_labels)\n\t\t\t\t\tprint(\"Validation accuracy: %.1f%%\" % valid_accuracy)\n\n\t\t\ttest_prediction = session.run(self.prediction, feed_dict={self.tf_dataset : test_dataset, self.tf_labels : test_labels, self.keep_prob : 1.0})\n\t\t\ttest_accuracy = MLP.accuracy(test_prediction, test_labels)\n\t\t\tprint(\"Test accuracy: %.1f%%\" % test_accuracy)\n\t\t\tself.save_results(test_accuracy)\n\t\n\t@staticmethod\n\tdef accuracy(predictions, labels):\n\t\treturn (100.0 * np.sum(np.argmax(predictions, 1) == np.argmax(labels, 1)) / predictions.shape[0])\n\t\t\n\tdef save_results(self,test_accuracy):\n\t\tprint(\"Saving results to \", self.output_file_path)\n\t\twith open(self.output_file_path, 'a') as csvfile:\n\t\t\tspamwriter = csv.writer(csvfile, delimiter=';', quotechar='|', quoting=csv.QUOTE_MINIMAL)\n\t\t\tspamwriter.writerow([str(self.network_shape),str(self.initial_learning_rate),str(self.decay_steps),str(self.decay_rate),str(self.regularization_parameter),str(test_accuracy)])\n\t\tprint(\"Results saved\")","sub_path":"scripts/mlp.py","file_name":"mlp.py","file_ext":"py","file_size_in_byte":5259,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"215255042","text":"import tensorflow as tf\nimport gym\nimport ShAl\nimport numpy as np\nfrom AandC import *\n#from TrainOrTest import *\n\n\ndef testDDPG(sess, env, actor, critic, newth):\n\n\n # test for max_episodes number of episodes\n # for i in range(int(1)):\n\n s = env.reset()\n\n ep_reward = 0\n ep_ave_max_q = 0\n\n #if action =='store_true':\n #env.render()\n \n\n a = actor.predict(np.reshape(s, (1, actor.s_dim))) \n\n \n s2, r, terminal, info = env.step(a[0], newth)\n\n \n s = s2\n ep_reward += r\n\n # if terminal:\n # print('| Episode: {:d} | Reward: {:d} |'.format(i, int(ep_reward)))\n # break\n print(\"************************ new value****************\")\n print(a)\n print(\"************************ new value****************\")\n return (a)\n # return a\n \ndef test(newth):\n\n with tf.Session() as sess:\n\n env = gym.make('shal-v0')\n np.random.seed(258)\n tf.set_random_seed(258)\n env.seed(258)\n env._max_episode_steps = 1000\n\n state_dim = env.observation_space.shape[0]\n action_dim = env.action_space.shape[0]\n action_bound = env.action_space.high\n\n actor = ActorNetwork(sess, state_dim, action_dim, action_bound,\n float(0.0001), float(0.001), int(64))\n\n critic = CriticNetwork(sess, state_dim, action_dim,\n float(0.001), float(0.001), float(0.99), actor.get_num_trainable_vars())\n\n saver = tf.train.Saver()\n saver.restore(sess, \"ckpt/model\")\n\n #testDDPG(sess, env, args, actor, critic)\n ping = testDDPG(sess, env, actor, critic, newth)\n\n\n#test()\n\n\n\n'''if __name__ == '__main__':\n parser = argparse.ArgumentParser(description='provide arguments for DDPG agent')\n\n # agent parameters\n parser.add_argument('--actor-lr', help='actor network learning rate', default=0.0001)\n parser.add_argument('--critic-lr', help='critic network learning rate', default=0.001)\n parser.add_argument('--gamma', help='discount factor for Bellman updates', default=0.99)\n parser.add_argument('--tau', help='target update parameter', default=0.001)\n parser.add_argument('--buffer-size', help='max size of the replay buffer', default=1000000)\n parser.add_argument('--minibatch-size', help='size of minibatch', default=64)\n\n # run parameters\n parser.add_argument('--env', help='gym env', default='shal-v0')\n parser.add_argument('--random-seed', help='random seed', default=258)\n parser.add_argument('--max-episodes', help='max num of episodes', default=250)\n parser.add_argument('--max-episode-len', help='max length of each episode', default=1000)\n parser.add_argument('--render-env', help='render gym env', action='store_true')\n parser.add_argument('--mode', help='train/test', default='train')\n \n \n args = vars(parser.parse_args())\n \n pp.pprint(args)\n\n if (args['mode'] == 'train'):\n train(args)\n elif (args['mode'] == 'test'):\n test(args)\n'''\n\n# test()\n \n\n","sub_path":"LMS/testCKPT.py","file_name":"testCKPT.py","file_ext":"py","file_size_in_byte":2996,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"349348844","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Nov 28 11:36:52 2019\n\n@author: hmsuleiman\n\"\"\"\nimport os\nimport gzip\npath = 'C:\\\\Users\\\\hmsuleiman\\\\Desktop\\\\Courses\\\\Python programming course\\\\Day3\\\\Data\\\\challenge.fa.gz'\n\ndef common_kmer(path, kmer_size): # with kmer_size you can change which kmer length you seek\n with gzip.open( path, 'rt') as fh: # open gzip file\n name = '>\\n'\n total_kmers = dict()\n while name != '':\n name = fh.readline()\n seq = fh.readline().strip('\\n')\n all_kmers(seq, total_kmers, kmer_size)\n print(most_common(total_kmers))\n \ndef all_kmers(seq,kmers, k = 7): # function detect most common K-mer, default 7-mer \n for i in range(0,len(seq) - k + 1):\n kmer = seq[i:i+k]\n if kmer in kmers:\n kmers[kmer] = kmers[kmer] + 1\n else:\n kmers[kmer] = 1\n\ndef most_common(total_kmers): \n for kmer in total_kmers:\n if len(kmer) == 0:\n print('No kmers in the dictionary.') # is this empty dict?\n else:\n value = 0\n values_kmers = [] # multiple kmers?\n for kmer in total_kmers: # for every kmer in total_kmers \n if total_kmers[kmer] > value: # if kmer has a value > 0\n value = total_kmers[kmer] # add kmer \n for kmer in total_kmers:\n if total_kmers[kmer] == value: # the population is in the population list\n values_kmers.append(kmer) # add kmer to value list\n # Printing the results\n if len(values_kmers) == 1: # if there is 1 kmer with the largest value \n print('{} -> {}.'.format(values_kmers[0], value))\n else:\n for kmer in values_kmers:\n print('- {}: {}'.format(kmer, value))# if there are multiple kmers\n","sub_path":"assignments day3.py","file_name":"assignments day3.py","file_ext":"py","file_size_in_byte":1768,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"45400859","text":"# -*- coding: utf-8 -*-\n\nimport inspect\nimport os\nfrom collections import namedtuple\nfrom operator import itemgetter\n\nimport pytest\n\nfrom wemake_python_styleguide.options.config import Configuration\nfrom wemake_python_styleguide.violations import (\n annotations,\n best_practices,\n complexity,\n consistency,\n naming,\n oop,\n refactoring,\n)\nfrom wemake_python_styleguide.violations.base import (\n ASTViolation,\n BaseViolation,\n MaybeASTViolation,\n SimpleViolation,\n TokenizeViolation,\n)\n\n\ndef _is_violation_class(cls) -> bool:\n base_classes = {\n ASTViolation,\n BaseViolation,\n SimpleViolation,\n TokenizeViolation,\n MaybeASTViolation,\n }\n if not inspect.isclass(cls):\n return False\n\n return issubclass(cls, BaseViolation) and cls not in base_classes\n\n\ndef _load_all_violation_classes():\n modules = [\n naming,\n complexity,\n consistency,\n best_practices,\n refactoring,\n oop,\n annotations,\n ]\n\n classes = {}\n for module in modules:\n classes_names_list = inspect.getmembers(module, _is_violation_class)\n only_classes = map(itemgetter(1), classes_names_list)\n classes.update({module: list(only_classes)})\n return classes\n\n\n@pytest.fixture(scope='session')\ndef all_violations():\n \"\"\"Loads all violations from the package and creates a flat list.\"\"\"\n classes = _load_all_violation_classes()\n all_errors_container = []\n for module_classes in classes.values():\n all_errors_container.extend(module_classes)\n return all_errors_container\n\n\n@pytest.fixture(scope='session')\ndef all_module_violations():\n \"\"\"Loads all violations from the package.\"\"\"\n return _load_all_violation_classes()\n\n\n@pytest.fixture(scope='session')\ndef absolute_path():\n \"\"\"Fixture to create full path relative to `contest.py` inside tests.\"\"\"\n def factory(*files):\n dirname = os.path.dirname(__file__)\n return os.path.join(dirname, *files)\n\n return factory\n\n\n@pytest.fixture(scope='session')\ndef options():\n \"\"\"Returns the options builder.\"\"\"\n default_values = {\n option.long_option_name[2:].replace('-', '_'): option.default\n for option in Configuration._options # noqa: WPS437\n }\n\n Options = namedtuple('options', default_values.keys())\n\n def factory(**kwargs):\n final_options = default_values.copy()\n final_options.update(kwargs)\n return Options(**final_options)\n\n return factory\n\n\n@pytest.fixture(scope='session')\ndef default_options(options): # noqa: WPS442\n \"\"\"Returns the default options.\"\"\"\n return options()\n","sub_path":"tests/conftest.py","file_name":"conftest.py","file_ext":"py","file_size_in_byte":2661,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"491943263","text":"from __future__ import annotations\n\nfrom typing import TYPE_CHECKING\n\nfrom kivy.animation import Animation\n\nfrom configurables import gameData\nfrom graphics import graphicsConfig\nfrom graphics.customWidgets.betterScatter import BetterScatter\nfrom graphics.spaceBuilderMergeGameScreenManager import get_screen\n\nif TYPE_CHECKING:\n from graphics.buildings import ResourceMiner\n from kivy.input import MotionEvent\n\nfrom kivy.properties import StringProperty, NumericProperty\nfrom kivy.uix.image import Image\n\nfrom resources import Textures\nfrom kivy.uix.floatlayout import FloatLayout\n\nfrom lib.betterLogger import BetterLogger\nfrom lib.globalEvents import GlobalEvents\n\n\nclass ResourceMinerManager(FloatLayout, BetterLogger):\n resource_miner_finished_icons: dict[int, ResourceMinerFinishedIcon] = {}\n\n def __init__(self, **kwargs):\n BetterLogger.__init__(self)\n FloatLayout.__init__(self, **kwargs)\n\n GlobalEvents.bind(mine_batch_finished=self.batch_finished)\n GlobalEvents.bind(remove_mine_finished_icon=self.remove_finished_icon)\n GlobalEvents.bind(on_scatter_transformed=self.update_positions)\n\n def batch_finished(self, resourceMiner: ResourceMiner):\n if resourceMiner.id not in self.resource_miner_finished_icons:\n (x1, y1), (x2, y2) = resourceMiner.get_projected_tl_br()\n x = x1 + ((x2 - x1) / 2)\n y = y2\n\n finished_icon = ClickableResourceMinerFinishedIcon(pos=(x, y), resource_name=resourceMiner.mine_item)\n finished_icon.resource_miner_id = resourceMiner.id\n self.resource_miner_finished_icons[resourceMiner.id] = finished_icon\n self.add_widget(finished_icon)\n\n\n self.resource_miner_finished_icons[resourceMiner.id].amount_that_has_been_mined += \\\n resourceMiner.mine_batch_amount\n\n def remove_finished_icon(self, icon: ResourceMinerFinishedIcon):\n self.remove_widget(icon)\n del self.resource_miner_finished_icons[icon.resource_miner_id]\n\n\n def update_positions(self): # TODO: Do something building is moved\n scatter: BetterScatter = get_screen(\"BaseBuildScreen\").ids[\"scatter\"]\n\n for building_id, icon in self.resource_miner_finished_icons.items():\n building = get_screen(\"BaseBuildScreen\").ids[\"base_layout\"].buildings[int(building_id)]\n (x1, y1), (x2, y2) = building.get_projected_tl_br()\n x = x1 + ((x2 - x1) / 2)\n y = y2\n\n icon.pos = scatter.to_parent(x, y)\n\n\nclass ResourceMinerFinishedIcon(FloatLayout, BetterLogger):\n amount_that_has_been_mined: int = 0\n\n bg_image: Image = None\n fg_image: Image = None\n\n resource_name: str = StringProperty(\"None\")\n resource_miner_id: int = None\n\n float_up_amount: int = NumericProperty(0)\n\n def __init__(self, **kwargs):\n self.bg_image = Image(allow_stretch=True, keep_ratio=True)\n self.fg_image = Image(allow_stretch=True, keep_ratio=True)\n\n BetterLogger.__init__(self)\n FloatLayout.__init__(self, **kwargs)\n\n self.bg_image.texture = Textures.get(\"ResourceMiner\", \"finished_icon_bg\")\n\n self.size_hint_x = None\n self.size_hint_y = 0.1\n\n self.add_widget(self.bg_image)\n self.add_widget(self.fg_image)\n\n\n def on_resource_name(self, _instance, _value: str):\n if self.resource_name == \"None\":\n self.fg_image.opacity = 0\n else:\n self.fg_image.opacity = 1\n try:\n self.fg_image.texture = Textures.get(\"ResourceMiner\", str(self.resource_name))\n except KeyError:\n self.fg_image.opacity = 0\n self.log_critical(\"No know texture -\", \"ResourceMiner -\", str(self.resource_name))\n\n\n def on_height(self, _instance, height: int):\n self.width = height\n\n self.bg_image.width = height\n self.fg_image.width = height\n\n def on_pos(self, _instance, pos):\n self.bg_image.pos = pos[0], pos[1] + self.float_up_amount\n self.fg_image.pos = pos[0], pos[1] + self.float_up_amount\n\n def on_float_up_amount(self, _instance, value):\n self.bg_image.pos = self.pos[0], self.pos[1] + value\n self.fg_image.pos = self.pos[0], self.pos[1] + value\n\n def on_width(self, _instance, width):\n self.bg_image.width = width\n self.fg_image.width = width\n\n\n def remove(self):\n a = Animation(**graphicsConfig.getdict(\"ResourceMiner\", \"icon_remove_animation\"),\n duration=graphicsConfig.getfloat(\"ResourceMiner\", \"icon_remove_animation_duration\"))\n a.bind(on_complete=lambda _animation, _instance: self.done_removing())\n a.start(self)\n\n def done_removing(self):\n GlobalEvents.dispatch(\"remove_mine_finished_icon\", self)\n\n\nclass ClickableResourceMinerFinishedIcon(ResourceMinerFinishedIcon):\n def on_touch_down(self, touch: MotionEvent):\n if self.collide_point(*touch.pos):\n gameData.add_to_inventory(self.resource_name, self.amount_that_has_been_mined)\n self.log_deep_debug(\"Clicked on, added\", self.amount_that_has_been_mined, \"of\", self.resource_name,\n \"to the inventory\")\n self.remove()\n\n\n__all__ = [\"ResourceMinerManager\"]\n","sub_path":"graphics/customWidgets/resourceMinerManager.py","file_name":"resourceMinerManager.py","file_ext":"py","file_size_in_byte":5259,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"610575556","text":"\"\"\"\nProblem:\n\nGiven a string with repeated characters, rearrange the string so that no two adjacent characters are the same. \nIf this is not possible, return None.\n\nExample:\n\nInput = \"aaabbc\"\nOutput = \"ababac\"\n\nInput = \"aaab\"\nOutput = None\n\"\"\"\n\n\ndef get_unique_adjacent(string):\n length = len(string)\n freq = {}\n\n if length == 0:\n return string\n\n # getting the character frequency\n for i in range(length):\n if string[i] in freq:\n freq[string[i]] += 1\n else:\n freq[string[i]] = 1\n\n sorted_freq = sorted(freq.items(), key=lambda x: x[1], reverse=True)\n queue = list(sorted_freq)\n\n # checking if a desired string can be formed\n if length % 2 == 0:\n if sorted_freq[0][1] > length // 2:\n return None\n else:\n if sorted_freq[0][1] > (length // 2) + 1:\n return None\n\n res = \"\"\n\n # creating the required string\n while queue:\n if len(queue) == 1:\n if queue[0][1] == 2:\n res = queue[0][0] + res + queue[0][0]\n break\n elif queue[0][1] == 1:\n if res[-1] != queue[0][0]:\n res += queue[0][0]\n else:\n res = queue[0][0] + res\n break\n else:\n return None\n\n res += queue[0][0] + queue[1][0]\n\n queue[0] = queue[0][0], queue[0][1] - 1\n queue[1] = queue[1][0], queue[1][1] - 1\n\n if len(queue) > 1 and queue[1][1] == 0:\n queue.pop(1)\n if len(queue) > 0 and queue[0][1] == 0:\n queue.pop(0)\n\n return res\n\n\n# DRIVER CODE\nprint(get_unique_adjacent(\"aaabbc\"))\nprint(get_unique_adjacent(\"aaabbcc\"))\nprint(get_unique_adjacent(\"aaabbac\"))\n\n# cannot form a word of the desired form\nprint(get_unique_adjacent(\"aaab\"))\nprint(get_unique_adjacent(\"aaabbaa\"))\n","sub_path":"Solutions/231.py","file_name":"231.py","file_ext":"py","file_size_in_byte":1868,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"212623113","text":"# MIT License\n#\n# Copyright (C) IBM Corporation 2018\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated\n# documentation files (the \"Software\"), to deal in the Software without restriction, including without limitation the\n# rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit\n# persons to whom the Software is furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in all copies or substantial portions of the\n# Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE\n# WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,\n# TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n# SOFTWARE.\nfrom __future__ import absolute_import, division, print_function, unicode_literals\n\nfrom io import BytesIO\nimport logging\n\nimport numpy as np\nfrom PIL import Image\n\nfrom art.defences.preprocessor import Preprocessor\nfrom art import NUMPY_DTYPE\n\nlogger = logging.getLogger(__name__)\n\n\nclass JpegCompression(Preprocessor):\n \"\"\"\n Implement the jpeg compression defence approach. Some related papers: https://arxiv.org/pdf/1705.02900.pdf,\n https://arxiv.org/abs/1608.00853\n \"\"\"\n params = ['quality', 'channel_index']\n\n def __init__(self, quality=50, channel_index=3):\n \"\"\"\n Create an instance of jpeg compression.\n\n :param quality: The image quality, on a scale from 1 (worst) to 95 (best). Values above 95 should be avoided.\n :type quality: `int`\n :param channel_index: Index of the axis in data containing the color channels or features.\n :type channel_index: `int`\n \"\"\"\n super(JpegCompression, self).__init__()\n self._is_fitted = True\n self.set_params(quality=quality, channel_index=channel_index)\n\n def __call__(self, x, y=None, quality=None, clip_values=(0, 1)):\n \"\"\"\n Apply jpeg compression to sample `x`.\n\n :param x: Sample to compress with shape `(batch_size, width, height, depth)`.\n :type x: `np.ndarray`\n :param y: Labels of the sample `x`. This function does not affect them in any way.\n :type y: `np.ndarray`\n :param quality: The image quality, on a scale from 1 (worst) to 95 (best). Values above 95 should be avoided.\n :type quality: `int`\n :return: compressed sample\n :rtype: `np.ndarray`\n \"\"\"\n if quality is not None:\n self.set_params(quality=quality)\n\n assert self.channel_index < len(x.shape)\n\n # Swap channel index\n if self.channel_index < 3:\n x_ = np.swapaxes(x, self.channel_index, 3)\n else:\n x_ = x.copy()\n\n # Convert into `uint8`\n x_ = x_ * 255\n x_ = x_.astype(\"uint8\")\n\n # Convert to 'L' mode\n if x_.shape[-1] == 1:\n x_ = np.reshape(x_, x_.shape[:-1])\n\n # Compress one image per time\n for i, xi in enumerate(x_):\n if len(xi.shape) == 2:\n xi = Image.fromarray(xi, mode='L')\n elif xi.shape[-1] == 3:\n xi = Image.fromarray(xi, mode='RGB')\n else:\n logger.log(level=40, msg=\"Currently only support `RGB` and `L` images.\")\n raise NotImplementedError(\"Currently only support `RGB` and `L` images.\")\n\n out = BytesIO()\n xi.save(out, format=\"jpeg\", quality=self.quality)\n xi = Image.open(out)\n xi = np.array(xi)\n x_[i] = xi\n del out\n\n # Expand dim if black/white images\n if len(x_.shape) < 4:\n x_ = np.expand_dims(x_, 3)\n\n # Convert to old dtype\n x_ = x_ / 255.0\n x_ = x_.astype(NUMPY_DTYPE)\n\n # Swap channel index\n if self.channel_index < 3:\n x_ = np.swapaxes(x_, self.channel_index, 3)\n\n x_ = np.clip(x_, clip_values[0], clip_values[1])\n\n return x_\n\n def fit(self, x, y=None, **kwargs):\n \"\"\"\n No parameters to learn for this method; do nothing.\n \"\"\"\n pass\n\n def set_params(self, **kwargs):\n \"\"\"\n Take in a dictionary of parameters and applies defence-specific checks before saving them as attributes.\n\n :param quality: The image quality, on a scale from 1 (worst) to 95 (best). Values above 95 should be avoided.\n :type quality: `int`\n :param channel_index: Index of the axis in data containing the color channels or features.\n :type channel_index: `int`\n \"\"\"\n # Save defense-specific parameters\n super(JpegCompression, self).set_params(**kwargs)\n\n if type(self.quality) is not int or self.quality <= 0 or self.quality > 100:\n logger.error('Image quality must be a positive integer and smaller than 101.')\n raise ValueError('Image quality must be a positive integer and smaller than 101.')\n\n if type(self.channel_index) is not int or self.channel_index <= 0:\n logger.error('Data channel must be a positive integer. The batch dimension is not a valid channel.')\n raise ValueError('Image quality must be a positive integer and smaller than 101.')\n\n return True\n\n\n\n","sub_path":"art/defences/jpeg_compression.py","file_name":"jpeg_compression.py","file_ext":"py","file_size_in_byte":5579,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"535030417","text":"#!/usr/bin/env python3\n# google-search.py - Opens several google search results \n\nimport requests, sys, webbrowser, bs4, csv \n\n#---------------\nprint('Googling...')\n\nres = requests.get('http://google.com/search?q=' + ' '.join(sys.argv[1:]))\nres.raise_for_status() # Remember to always do this\n\n# Retrieve top search result links\nsoup = bs4.BeautifulSoup(res.text, 'html.parser') \n\n# Open browser tab for each result\nlinkElems = soup.select('.r a')\nlinkList = []\nfor element in linkElems:\n\t# print(element.getText())\n\tlinkList.append(element.getText())\n\nprint(linkList)\n\nnumOpen = min(5, len(linkElems)) # making sure you either open 5 links or, if there's less, the entire list\n\n# for i in range(numOpen):\n# \twebbrowser.open('http://google.com' + linkElems[i].get('href'))\n\nsecretary = open('google-search.csv', 'a', newline='') # HIRE YOUR SECRETARY\n\ncsvSecretary = csv.writer(secretary) # TRAIN YOUR SECRETARY\n\ncsvSecretary.writerow(linkList) # TELL HIM WHAT TO WRITE\n\nsecretary.close() # LET HIM OFF FOR THE DAY \n\n\n\n\n","sub_path":"google-search.py","file_name":"google-search.py","file_ext":"py","file_size_in_byte":1020,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"64508981","text":"import numpy as np\n\nfrom KinetiKit.units import MHz, fs, nm, uW\nimport KinetiKit.kit as kin_kit\n\n\n__all__ = ['Excitation']\n\n\nclass Excitation():\n \"\"\"define excitation parameters for simulation\"\"\"\n \n pulse_default = {'power': 1*uW,\n 'reprate': 80*MHz,\n 'fwhm': 100*fs,\n 'wavelength': 400*nm,\n 'pulse_window': None,\n }\n \n cw_default = {'power': 0,\n 'wavelength': 514.5*nm\n }\n \n def __init__(self, pulse={}, cw={}, numcycles=500):\n \n pulse_params = self.pulse_default.copy(); pulse_params.update(pulse)\n self.pulse = pulse_params\n self.pulse_power = self.pulse['power'] \n self.pulse_energy = self.pulse['power'] / self.pulse['reprate']\n self.pulse_wavelength = self.pulse['wavelength']\n self.pulse_carriers = self.pulse_energy \\\n * joules_to_photons(self.pulse_wavelength)\n self.pulse_fwhm = self.pulse['fwhm']\n self.pulse_window = self.pulse['pulse_window']\n if self.pulse_window is None:\n self.pulse_window = 20 * self.pulse_fwhm\n \n cw_params = self.cw_default.copy(); cw_params.update(cw)\n self.cw = cw_params\n self.cw_wavelength = self.cw['wavelength']\n self.cw_power = self.cw['power'] * joules_to_photons(self.cw_wavelength)\n \n self.numcycles = numcycles\n \n def gen_pulse(self, t, center=0, stepsize=None):\n \"\"\"\n generates a Gaussian pulse along a time array; if the time step is too large\n compared to the FWHM of the pulse, a 1x1 array is generated that contains the \n rate of arrival of photons.\n \"\"\"\n pulseCarriers = self.pulse_carriers\n sigma_to_fwhm = 2 * np.sqrt(2 * np.log(2))\n pulsePeak = pulseCarriers * sigma_to_fwhm \\\n / (np.sqrt(2 * np.pi) * self.pulse_fwhm)\n if len(t) > 50:\n return kin_kit.Gauss(t, pulsePeak, center, self.pulse_fwhm)\n else:\n return np.array([pulseCarriers/stepsize])\n \n def updated_with(self, pulse={}, cw={}, numcycles=None):\n \"\"\"returns a new light object based on the current light object, but\n with some modified arguments.\n \n \"\"\"\n new_pulse = self.pulse; new_pulse.update(pulse)\n new_cw = self.cw; new_cw.update(cw)\n if numcycles is None:\n numcycles = self.numcycles\n return Excitation(new_pulse, new_cw, numcycles)\n \n\ndef joules_to_photons(wavelength):\n \"\"\"convert Joules to photons\n\n arguments\n wavelength : units of meters\n \"\"\"\n h_Planck = 6.626e-34 # J s\n c = 2.9979e8 # m / s\n return wavelength / (h_Planck * c)\n","sub_path":"KinetiKit/sim/lib/_excitation.py","file_name":"_excitation.py","file_ext":"py","file_size_in_byte":2734,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"368809026","text":"import cv2\nimport argparse\nimport numpy as np\nimport time\nfrom pyzbar import pyzbar\n\n\ndef qc(data):\n \n gray = cv2.cvtColor(data, cv2.COLOR_BGR2GRAY)\n\n gray = cv2.GaussianBlur(gray, (5, 5), 0)\n gray = cv2.medianBlur(gray, 5)\n gray = cv2.adaptiveThreshold(gray, 255, cv2.ADAPTIVE_THRESH_GAUSSIAN_C,\n cv2.THRESH_BINARY, 11, 3.5)\n kernel = np.ones((2, 3), np.uint8)\n gray = cv2.erode(gray, kernel, iterations=1)\n gray = cv2.dilate(gray, kernel, iterations=1)\n circles = cv2.HoughCircles(gray, cv2.HOUGH_GRADIENT, 1, 20, param1=50, param2=25, minRadius=50, maxRadius=90)\n \n if circles is None:\n return False\n else:\n return True\n\n\nwhile True:\n cap = cv2.VideoCapture(0)\n cap.set(3, 320)\n cap.set(4, 320)\n \n\n retval, frame = cap.read()\n obj = pyzbar.decode(frame)\n if qc(frame):\n print(obj)\n print('circle')\n \n else:\n print(obj)\n print('error')\n cap.release()\n obj = None\n time.sleep(1)\n\"\"\"while True:\n\n retval, frame = cap.read()\n decoded_objects = pyzbar.decode(frame)\n retval, buffer = cv2.imencode('.jpg', frame)\n\n if qc(frame):\n print('qc:pass')\n else:\n print('qc:fail')\n for obj in decoded_objects:\n print(obj.data)\n\n\"\"\"\n","sub_path":"src/vision/qrtest.py","file_name":"qrtest.py","file_ext":"py","file_size_in_byte":1309,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"104861514","text":"from ..emily_bot import Emily\nimport asyncio\nimport random\n\nBALL_ANSWERS = [\"It is certain\", \"It is decidedly so\", \"Without a doubt\",\n \"Yes definitely\", \"You may rely on it\", \"As I see it yes\",\n \"Most likely\", \"Outlook good\", \"Yes\",\n \"Signs point to yes\", \"Reply hazy try again\",\n \"Ask again later\", \"Better not tell you now\",\n \"Cannot predict now\", \"Concentrate and ask again\",\n \"Don't count on it\", \"My reply is no\",\n \"My sources say no\", \"Outlook not so good\",\n \"Very doubtful\"]\n\nclass Ball:\t\n\n\tdef command(self):\n\t\treturn '8ball'\n\n\tdef help(self):\n\t\treturn '8ball X - Ask 8ball a question'\n\n\t@asyncio.coroutine\n\tdef doCommand(self, message):\n\t\ttry:\n\t\t\twords = message.content.split(' ')\n\n\t\t\tif len(words) <= 1:\n\t\t\t\tyield from Emily.client.send_message(message.channel, 'No question asked.')\n\t\t\t\treturn\n\t\t\telse:\n\t\t\t\tquestion = ' '.join(words[1:])\n\t\t\t\tquestion = question.strip('?')\n\t\t\t\tanswer = random.randint(0, len(BALL_ANSWERS) - 1)\n\t\t\t\tyield from Emily.client.send_message(message.channel, \"**Question:** \" + question +\":question: \\n**Answer:** \" + BALL_ANSWERS[answer] + \":8ball:\")\n\t\t\t\treturn\n\n\n\t\texcept Exception as e:\n\t\t\traise\n\n\tdef visible(self):\n\t\treturn True\n\t\t","sub_path":"emily/cmd/ball.py","file_name":"ball.py","file_ext":"py","file_size_in_byte":1326,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"455846161","text":"import pygame\nimport pygame as pg\nfrom pygame.sprite import Sprite\n\nimport globals as g\nimport ressources as r\nfrom gui import print_message\n\nimport colors\n\nfrom random import randint\nfrom random import choice\n\nimport math\n\n\n# TODO Scale Graphics to 32x32\nclass GameObject(Sprite):\n \"\"\"A class to represent a GameObject (Player, Monster, Item, etc\"\"\"\n\n def __init__(self, settings, name, x, y, color=colors.black,\n img=None,\n blocks=False,\n fighter=None,\n ai=None,\n item=None,\n equipment=None):\n\n \"\"\"Initialize GameObject (Player, Monster, Item, etc\"\"\"\n super(GameObject,self).__init__()\n self.settings = settings\n self.name = name\n self.screen = settings.screen\n self.screen_width = settings.screen_width\n self.screen_height = settings.screen_height\n self.tilesize = settings.tilesize\n\n self.color = color\n self.blocks = blocks\n\n self.image = pg.Surface((self.tilesize, self.tilesize))\n self.image.fill(self.color)\n self.rect = self.image.get_rect()\n self.x = x\n self.y = y\n\n\n # TODO check what this does\n self.rect.x = x * self.tilesize\n self.rect.y = y * self.tilesize\n\n # TODO useful?\n if img != None:\n self.image = img\n self.rect = self.image.get_rect()\n\n\n\n self.fighter = fighter\n if self.fighter: # let the fighter component know who owns it\n self.fighter.owner = self\n\n self.ai = ai\n if self.ai: # let the AI component know who owns it\n self.ai.owner = self\n\n self.item = item\n if self.item: # let the Item component know who owns it\n self.item.owner = self\n\n self.equipment = equipment\n if self.equipment: # let the Equipment component know who owns it\n self.equipment.owner = self\n\n self.item = Item()\n self.item.owner = self\n\n\n\n def move(self, dx=0, dy=0):\n \"\"\"Move GameObject\"\"\"\n #TODO Problem with: IndexError: list index out of range\n try:\n g.map_matrix[self.x+dx][self.y+dy]\n except IndexError:\n return None\n\n\n # TODO only checks BLOCKED for monster_group\n if not g.map_matrix[self.x + dx][self.y + dy].blocked \\\n and not self.out_of_bounds(dx, dy)\\\n and not is_blocked(self.x + dx, self.y + dy, g.monster_group):\n\n self.x += dx\n self.y += dy\n\n def move_or_attack(self, dx=0, dy=0):\n\n # the coordinates the player is moving to/attacking\n x = self.x + dx\n y = self.y + dy\n\n # try to find an attackable object there\n target = None\n for obj in g.monster_group:\n if obj.fighter and obj.x == x and obj.y == y:\n target = obj\n break\n\n # attack if target found, move otherwise\n if target is not None:\n g.player.fighter.attack(target)\n else:\n self.move(dx, dy)\n g.fov_recompute = True\n\n def move_towards(self, target_x, target_y):\n # vector from this object to the target, and distance\n dx = target_x - self.x\n dy = target_y - self.y\n distance = math.sqrt(dx ** 2 + dy ** 2)\n\n # normalize it to length 1 (preserving direction), then round it and\n # convert to integer so the movement is restricted to the map grid\n dx = int(round(dx / distance))\n dy = int(round(dy / distance))\n self.move(dx, dy)\n\n def distance_to(self, other):\n # return the distance to another object\n dx = other.x - self.x\n dy = other.y - self.y\n return math.sqrt(dx ** 2 + dy ** 2)\n\n def update(self):\n self.rect.x = self.x * self.tilesize\n self.rect.y = self.y * self.tilesize\n\n def blitme(self):\n \"\"\"Draw the alien at its current location\"\"\"\n self.screen.blit(self.image, self.rect)\n\n # TODO out_of_bounds useless?\n def out_of_bounds(self, dx=0, dy=0):\n if self.x + dx < 0 \\\n or self.x + dx > (self.settings.tile_width) \\\n or self.y + dy < 0 \\\n or self.y + dy > (self.settings.tile_height):\n return True\n return False\n\n\n \"\"\"TEST\"\"\"\n def draw_addons(self):\n \"\"\"Make Equipment and addons visible on Player\"\"\"\n #rect = self.rect\n self.screen.blit(r.player_gun, self.rect)\n #self.screen.blit(r.player_helmet, self.rect)\n\n\n\n\nclass Fighter:\n \"\"\"combat-related properties and methods (monster, player, NPC).\"\"\"\n def __init__(self, hp, defense, power, death_function=None):\n self.max_hp = hp\n self.hp = hp\n self.defense = defense\n # TODO power rename to attack?\n self.power = power\n self.death_function = death_function\n\n # Equipment Slots\n #self.eq_slot_body = None\n #self.eq_slot_head = None\n #self.eq_slot_feet = None\n #self.eq_slot_gun = None\n\n # TODO Arm Hand hand2 RENAME?\n # nothing is used for KeyError in Equipment.equip()\n # blocked is used for slot_blocked=blocked so nonblocked can be equipped\n self.equipment_slots = {\"head\": 'unequipped',\n \"body\": 'unequipped',\n \"feet\": 'unequipped',\n \"arm\": 'unequipped',\n \"hand\": 'unequipped',\n \"hand2\": 'unequipped',\n\n # No Use\n \"nothing\": 'nothing',\n \"blocked\": 'unequipped'}\n\n def take_damage(self, damage):\n # apply damage if possible\n if damage > 0:\n self.hp -= damage\n\n # check for death. if there's a death function, call it\n if self.hp <= 0:\n class_function = self.death_function\n if class_function is not None:\n class_function(self.owner)\n\n def attack(self, target):\n # a simple formula for attack damage\n # TODO damage_calc\n damage = self.power - target.fighter.defense\n\n if damage > 0:\n # make the target take some damage\n print_message(self.owner.name.capitalize() + ' attacks ' + target.name +\n ' for ' + str(damage) + ' hit points.')\n target.fighter.take_damage(damage)\n else:\n print_message(self.owner.name.capitalize() + ' attacks ' + target.name +\n ' but it has no effect!')\n\n\n def heal(self, amount):\n # TODO heal function redundant?\n #heal by the given amount, without going over the maximum\n self.hp += amount\n if self.hp > self.max_hp:\n self.hp = self.max_hp\n\n# TODO Equipment as Item class or its own class?\nclass Item:\n #an item that can be picked up and used.\n def __init__(self, use_function=None):\n self.use_function = use_function\n\n def pick_up(self):\n #add to the player's inventory and remove from the map\n if len(g.inventory) >= 26:\n print_message('Your inventory is full, cannot pick up ' + self.owner.name + '.', colors.red)\n else:\n g.inventory.add(self.owner)\n g.item_group.remove(self.owner)\n\n print_message('You picked up a ' + self.owner.name + '!', colors.green)\n\n def use(self):\n #just call the \"use_fucntion\" if it is defined\n if self.use_function is None:\n print_message('The ' + self.owner.name + ' cannot be used.')\n else:\n if self.use_function() != 'cancelled':\n g.inventory.remove(self.owner)\n\n def drop(self, new_x=None, new_y=None):\n # TODO Use new_x, new_y?\n # Drop Item\n\n self.owner.x = g.player.x\n self.owner.y = g.player.y\n\n g.item_group.add(self.owner)\n g.inventory.remove(self.owner)\n\n print_message(self.owner.name + \" dropped!\", colors.darkest_turquoise)\n\n\n\n\n\n\n# TODO use_function in Equipment, Item, equip in Controls?\nclass Equipment:\n\n def __init__(self, slot=None, slot2='nothing', slot_block='blocked',\n attack_bonus=0,\n defense_bonus=0,\n health_bonus=0):\n\n self.attack_bonus = attack_bonus\n self.defense_bonus = defense_bonus\n self.health_bonus = health_bonus\n self.slot = slot\n self.slot2 = slot2\n self.slot_block = slot_block\n\n self.equipped = False\n\n #self.owner.Item.use_function = self.toggle_equip()\n #self.item.use_function = self.toggle_equip()\n\n def toggle_equip(self):\n\n if self.equipped:\n self.unequip()\n else:\n self.equip()\n\n def equip(self):\n \"\"\"Equip item and add bonuses\"\"\"\n\n\n # Check if Slot is unequipped and blocked slot isnt already blocked\n if g.player.fighter.equipment_slots[self.slot] == 'unequipped' and \\\n g.player.fighter.equipment_slots[self.slot_block] == 'unequipped':\n\n g.player.fighter.equipment_slots[self.slot] = self #self.owner.name\n self.equipped = True\n\n # Add bonuses\n self.set_item_bonuses(g.player, 1)\n\n # If a slot will be blocked, add self to blocked slot\n if self.slot_block != 'blocked':\n g.player.fighter.equipment_slots[self.slot_block] = self\n\n print_message(\"item equipped\", colors.green)\n\n # Try Slot 2\n elif self.slot2 != 'nothing':\n if g.player.fighter.equipment_slots[self.slot2] == 'unequipped':\n g.player.fighter.equipment_slots[self.slot2] = self\n self.equipped = True\n\n # Add bonuses\n self.set_item_bonuses(g.player, 1)\n\n print_message(\"item equipped\", colors.green)\n\n else:\n print_message(\"slot occupied\", colors.red)\n\n #TEST Display Equipment\n #print(g.player.fighter.equipment_slots)\n\n\n\n def unequip(self):\n \"\"\"Unequip item and distract bonuses\"\"\"\n\n # TODO 2 If's HACKY\n # Try if Equipment self is in the first possible slot\n if g.player.fighter.equipment_slots[self.slot] == self:\n\n\n g.player.fighter.equipment_slots[self.slot] = 'unequipped'\n g.player.fighter.equipment_slots[self.slot_block] = 'unequipped'\n self.equipped = False\n\n # Distract bonuses\n self.set_item_bonuses(g.player, -1)\n\n\n if self.slot_block != 'nothing':\n g.player.fighter.equipment_slots[self.slot_block] = 'unequipped'\n\n print_message(\"item unequipped\", colors.yellow)\n\n # Try slot 2\n elif g.player.fighter.equipment_slots[self.slot2] == self:\n\n g.player.fighter.equipment_slots[self.slot2] = 'unequipped'\n self.equipped = False\n\n # Distract bonuses\n self.set_item_bonuses(g.player, -1)\n\n print_message(\"item unequipped\", colors.yellow)\n\n # TEST Display Equipment\n #print(g.player.fighter.equipment_slots)\n\n\n def set_item_bonuses(self, wearer, pos_neg):\n \"\"\"Add or Subtract Bonuses using pos_neg as multiplier\"\"\"\n\n wearer.fighter.defense += self.defense_bonus * pos_neg\n wearer.fighter.max_hp += self.health_bonus * pos_neg\n wearer.fighter.power += self.attack_bonus * pos_neg\n\n\n\n\n\n\n # TODO AI_BasicMonster\nclass BasicMonster:\n # TODO AI findet nur Player?\n #AI for a basic monster.\n def take_turn(self):\n # a basic monster takes its turn. If you can see it, it can see you\n monster = self.owner\n if (monster.x, monster.y) in g.visible_tiles:\n\n # move towards player if far away\n if monster.distance_to(g.player) >= 2:\n monster.move_towards(g.player.x, g.player.y)\n\n # close enough, attack! (if the player is still alive.)\n elif g.player.fighter.hp > 0:\n monster.fighter.attack(g.player)\n\n\nclass ConfusedMonster:\n # AI for a temporarily confused monster (reverts to previous AI after a while).\n def __init__(self, old_ai, num_turns=10):\n self.old_ai = old_ai\n self.num_turns = num_turns\n\n def take_turn(self):\n if self.num_turns > 0: # still confused...\n # move in a random direction, and decrease the number of turns confused\n self.owner.move(randint(-1, 1), randint(-1, 1))\n self.num_turns -= 1\n\n else: # restore the previous AI (this one will be deleted because it's not referenced anymore)\n self.owner.ai = self.old_ai\n print_message('The ' + self.owner.name + ' is no longer confused!', colors.red)\n\n# TODO AI-SYSTEM like this? ...\n\"\"\"\nThis seems like a neat way to interrupt a monster's actions in reaction to something. \nIt is, but not for every situation! You could create a full finite state machine by just \nswapping AI components, but such a system would be very hard to debug. Instead, for most \ntypes of AI that have different states, you can simply have a \"state\" property in the AI component, like this:\n\nclass DragonAI:\n def __init__(self):\n self.state = 'chasing'\n \n def take_turn(self):\n if self.state == 'chasing': ...\n elif self.state == 'charging-fire-breath': ...\n\"\"\"\n\nclass GraphicObject(Sprite):\n \"\"\" Simple GraphicObject for example Explosions etc.\"\"\"\n\n def __init__(self, settings, image, x, y,\n color=None,\n duration=None):\n\n \"\"\"Initialize \"\"\"\n super(GraphicObject,self).__init__()\n self.settings = settings\n self.screen = settings.screen\n self.screen_width = settings.screen_width\n self.screen_height = settings.screen_height\n\n\n self.image = image\n self.rect = self.image.get_rect()\n\n if color != None:\n self.image.fill(color)\n\n self.rect.centerx = x\n self.rect.centery = y\n\n\n\n # Time Variables\n self.last_update = pygame.time.get_ticks()\n self.duration = duration\n\n self.name = \"GraphObj\"\n\n\n\n\n def blitme(self):\n \"\"\"Draw the alien at its current location\"\"\"\n self.screen.blit(self.image, self.rect)\n\n def update(self):\n now = pygame.time.get_ticks()\n if now - self.last_update > self.duration:\n self.kill()\n\n# TODO ProjectileObjectX OLD DELETE\nclass ProjectileObjectX(Sprite):\n \"\"\"A class to represent a single alien in the fleet\"\"\"\n\n def __init__(self, settings, img, x, y,\n source=None,\n target=None,\n target_x=None, target_y=None,\n velX=None, velY=None,\n delay=None,\n hit=False):\n\n \"\"\"Initialize GameObject (Player, Monster, Item, etc\"\"\"\n super(ProjectileObject,self).__init__()\n self.settings = settings\n self.screen = settings.screen\n self.screen_width = settings.screen_width\n self.screen_height = settings.screen_height\n\n\n self.image = img\n\n self.target_x = target_x\n self.target_y = target_y\n self.target = target\n\n self.source = source\n\n self.rect = self.image.get_rect()\n\n self.rect.x = x\n self.rect.y = y\n\n #???\n self.x = x\n self.y = y\n\n #self.speedx = 5\n #self.speedy = 5\n\n\n self.pos = pygame.math.Vector2(self.rect.center)\n self.vel = pygame.math.Vector2(velX, velY)\n\n # Time Variables\n self.last_update = pygame.time.get_ticks()\n self.delay = delay\n\n self.hit = hit\n\n # rotate\n #self.image_orig = img.copy()\n self.image_orig = self.image.copy()\n\n\n\n def blitme(self):\n \"\"\"Draw the alien at its current location\"\"\"\n\n # TODO performance issues checking if for every blit?\n # Delay check\n now = pygame.time.get_ticks()\n if now - self.last_update > self.delay:\n\n self.screen.blit(self.image, self.rect)\n\n def update(self):\n #self.rect.x += self.speedx\n #self.rect.y += self.speedy\n\n # TODO performance issues checking if for every blit?\n # Delay check\n now = pygame.time.get_ticks()\n if now - self.last_update > self.delay:\n\n self.pos = self.pos + self.vel\n self.rect.center = (int(self.pos[0]), int(self.pos[1]))\n\n #print(self.rect.center, \" Target:\", self.target_x, \" \", self.target_y, \" \", self.rect)\n\n #Make a bigger target rect to scan for hits with target\n target_rect = pygame.Rect(0, 0, 10, 10) # 8, 8\n target_rect.center = (self.target_x, self.target_y)\n\n #Kill sprite if collided\n if self.rect.colliderect(target_rect):\n self.kill()\n if self.target != None and self.target in g.monster_group:\n g.player.fighter.attack(self.target)\n\n # if target in monster.group\n # source attack target\n\n\n # Kill Sprite if out of bounds\n if self.rect.centerx < 0 or \\\n self.rect.centerx > (self.settings.screen_width) or \\\n self.rect.centery < 0 or \\\n self.rect.centery > (self.settings.screen_height):\n\n self.kill()\n\n # TODO Projectile KILL ANIMATION\n # TODO Fix Collision Detection!\n\n\n def rotate(self, rotation_angle):\n # TODO destroys image when rotating\n #self.image = pygame.transform.rotate(self.image, self.rot_speed)\n\n new_image = pygame.transform.rotate(self.image, rotation_angle)\n old_center = self.rect.center\n self.image = new_image\n self.rect = self.image.get_rect()\n self.rect.center = old_center\n\n\nclass ProjectileObject(Sprite):\n \"\"\"A class to represent a single alien in the fleet\"\"\"\n\n def __init__(self, settings, source,\n img=None,\n color=None,\n target=None,\n target_x=None, target_y=None,\n velX=None, velY=None,\n delay=None,\n hit=False):\n\n \"\"\"Initialize GameObject (Player, Monster, Item, etc\"\"\"\n super(ProjectileObject,self).__init__()\n self.settings = settings\n self.screen = settings.screen\n self.screen_width = settings.screen_width\n self.screen_height = settings.screen_height\n\n self.color = color\n self.image = pygame.Surface((2,2))\n self.image.fill(color)\n\n if img != None:\n self.image = img\n\n self.target_x = target_x\n self.target_y = target_y\n self.target = target\n\n self.source = source\n\n self.rect = self.image.get_rect()\n\n self.rect.centerx = self.source.rect.centerx\n self.rect.centery = self.source.rect.centery\n\n self.pos = pygame.math.Vector2(self.rect.center)\n self.vel = pygame.math.Vector2(velX, velY)\n\n # Time Variables\n self.last_update = pygame.time.get_ticks()\n self.delay = delay\n\n self.hit = hit\n self.name = \"Projectile\"\n\n # rotate\n self.image_orig = self.image.copy()\n\n\n\n self.testlistx = []\n self.testlisty = []\n\n\n def blitme(self):\n \"\"\"Draw the alien at its current location\"\"\"\n\n # TODO performance issues checking if for every blit?\n # Delay check\n now = pygame.time.get_ticks()\n if now - self.last_update > self.delay:\n\n self.screen.blit(self.image, self.rect)\n\n\n def update(self):\n\n\n # TODO performance issues checking if for every blit?\n # Delay check\n now = pygame.time.get_ticks()\n if now - self.last_update > self.delay:\n\n \"\"\"TEST\"\"\"\n # Create Bullet Tail\n #self.create_bullet_tail()\n #pg.draw.line(self.screen, colors.white, (self.rect.centerx, self.rect.centery),\n # (self.target_x, self.target_y))\n\n line_x = self.source.rect.centerx\n line_y = self.source.rect.centery\n #\n # if self.line_counter == 10:\n # line_x = self.rect.centerx\n # line_y = self.rect.centery\n # line_counter = 0\n # self.line_counter +=1\n\n # if now - self.last_update > 100:\n # line_x =\n # line_y =\n\n if len(self.testlistx) > 4:\n line_x = self.testlistx.pop(2)#[5]\n line_y = self.testlisty.pop(2)#[5]\n\n\n\n pg.draw.line(self.screen, colors.white, (line_x, line_y),\n (self.rect.centerx, self.rect.centery))\n\n\n self.pos = self.pos + self.vel\n self.rect.center = (int(self.pos[0]), int(self.pos[1]))\n self.testlistx.append(int(self.pos[0]))\n self.testlisty.append(int(self.pos[1]))\n\n #print(self.rect.center, \" Target:\", self.target_x, \" \", self.target_y, \" \", self.rect)\n\n #Make a bigger target rect to scan for hits with target\n target_rect = pygame.Rect(0, 0, 10, 10) # 8, 8\n target_rect.center = (self.target_x, self.target_y)\n\n #Kill sprite if collided\n if self.rect.colliderect(target_rect):\n self.kill()\n \"\"\"TEST\"\"\"\n self.create_impact()\n\n\n if self.target != None and self.target in g.monster_group:\n g.player.fighter.attack(self.target)\n\n # if target in monster.group\n # source attack target\n\n\n # Kill Sprite if out of bounds\n if self.rect.centerx < 0 or \\\n self.rect.centerx > (self.settings.screen_width) or \\\n self.rect.centery < 0 or \\\n self.rect.centery > (self.settings.screen_height):\n\n self.kill()\n \"\"\"TEST\"\"\"\n self.create_impact()\n\n # TODO Projectile KILL ANIMATION\n # TODO Fix Collision Detection!\n\n def create_impact(self):\n impact = GraphicObject(g.settings, r.bullet_hit, self.rect.centerx, self.rect.centery,\n duration=50)\n g.graphics_group.add(impact)\n\n # TODO DELETE\n def create_bullet_tail(self):\n\n # dir_x, dir_y = self.direction\n # x = self.rect.centerx - (1 * dir_x)\n # y = self.rect.centery - (1 * dir_y)\n\n # trail_pos = self.pos + self.vel\n # print(\"1: \", trail_pos[0])\n # x = trail_pos[0] - 0.1\n # y = trail_pos[1] - 0.1\n # print(\"2: \", trail_pos[0])\n\n #self.pos = pygame.math.Vector2(self.rect.center)\n #self.vel = pygame.math.Vector2(velX, velY)\n\n\n print(\"1: \", self.pos[0])\n trail_vel = self.vel * 0.999999999999999999999999\n trail_pos = self.pos + trail_vel\n x = trail_pos[0]\n y = trail_pos[1]\n print(\"2: \", trail_pos[0])\n\n # for i in g.graphics_group:\n # print(i.name)\n # print()\n\n\n\n tail = GraphicObject(g.settings, self.image, x, y,\n duration=0)\n g.graphics_group.add(tail)\n\n\n def rotate(self, rotation_angle):\n # TODO destroys image when rotating\n #self.image = pygame.transform.rotate(self.image, self.rot_speed)\n\n #new_image = pygame.transform.rotate(self.image, rotation_angle)\n new_image = pygame.transform.rotozoom(self.image, rotation_angle, 1)\n old_center = self.rect.center\n self.image = new_image\n self.rect = self.image.get_rect()\n self.rect.center = old_center\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n#TODO where to put is_blocked?\ndef is_blocked(x, y, group):\n # first test the map tile\n if g.map_matrix[x][y].blocked:\n return True\n\n # now check for any blocking objects\n for obj in group:\n if obj.blocks and obj.x == x and obj.y == y:\n return True\n\n return False\n\n#TODO all_objects class with clear/update function\n\n# TODO send to back, dead spirtes (Or dead group??) in Gameobject\n# def send_to_back(self):\n# #make this object be drawn first, so all others appear above it if they're in the same tile.\n# global objects\n# objects.remove(self)\n# objects.insert(0, self)\n\n\n","sub_path":"objects.py","file_name":"objects.py","file_ext":"py","file_size_in_byte":24456,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"454430794","text":"# -*- coding: utf-8 -*-\nfrom abc import ABC, abstractmethod\nimport paramiko, socket\nimport threading as clients\nimport sys, time\nfrom pathlib import Path\nimport configparser\nconfig = configparser.ConfigParser()\nconfig.read('configuration.ini')\nserver_port = int(config.get('NETWORKING', 'SERVER_PORT').strip())\nrsa_key = \"{}/{}\".format(Path.home(), config.get('NETWORKING', 'RSA_KEY').strip())\naddresses = [e[:8].strip() for e in config.get('NETWORKING', 'ALLOW').split(',')]\n\n__all__ = ['Spana', ]\n\nclass Server(paramiko.ServerInterface):\n\tdef __init__(self, remote, log_book):\n\t\tself.remote = remote\n\t\tself.log_book = log_book\n\n\tdef check_channel_request(self, kind, chanid):\n\t\tif kind == 'session':\n\t\t\treturn paramiko.OPEN_SUCCEEDED\n\n\tdef check_auth_publickey(self, username, key):\n\t\tself.user = username\n\t\treturn paramiko.AUTH_SUCCESSFUL\n\n\tdef get_allowed_auths(self, username):\n\t\treturn 'publickey'\n\n\tdef check_channel_exec_request(self, channel, command):\n\t\tcommand = command.decode(\"utf-8\")\n\t\tif command in ['terminal']:\n\t\t\tself.log_book.info(\"[*] Incoming call from [{}] which is using the client: [{}]\".format(self.remote, command))\n\t\t\treturn True\n\t\telse:\n\t\t\tchannel.close()\n\t\t\tself.log_book.warning('Closed the channel for {}. Wrong kind of client..'.format(self.remote))\n\t\t\treturn False\n\n\nclass Spana(ABC):\n\tsock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n\tsock.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)\n\tsend_prompt = True\n\n\tdef __init__(self):\n\t\tself.ib_affair()\n\t\tself.log_setup()\n\t\tself.port = server_port\n\t\tself.addresses = addresses\n\t\tself.host_key = paramiko.RSAKey(filename=rsa_key)\n\t\tself.sock.bind((self.host, self.port))\n\t\tself.sock.listen(5)\n\t\twhile True:\n\t\t\tclient, address = self.sock.accept()\n\t\t\tclient.settimeout(30)\n\t\t\tclients.Thread(target=self.client_connection, args=(client, address)).start()\n\t\t\ttime.sleep(2)\n\n\t@abstractmethod\n\tdef log_setup(self):\n\t\tpass\n\t@abstractmethod\n\tdef ib_affair(self):\n\t\tpass\n\t@abstractmethod\n\tdef interests(self):\n\t\tpass\n\t@abstractmethod\n\tdef radio(self):\n\t\tpass\n\tdef user_rights(self, address):\n\t\tif(address[:7] in self.addresses):\n\t\t\tself.log_book.info(\"[*] [{}] have access to the server.\".format(address))\n\t\t\tself.remote_user = address\n\t\t\treturn True\n\t\telse:\n\t\t\tself.log_book.critical(\"[*] [{}] doesn't have access to the server!\".format(address))\n\t\t\treturn False\n\n\tdef hold_prompt(self, message, boolean, user, remote_address):\n\t\tif not boolean:\tself.user_prompt(message, user); self.ask_for_prompt(message, user, remote_address)\n\n\tdef ask_for_prompt(self, message, user, remote_address, **kwargs):\n\t\tinp = message.makefile('r+U'); cmd = inp.readline().strip('\\r\\n')\n\t\thostname = socket.gethostname()\n\n\t\tself.log_book.info(\"[*] {0}@{1} sent: {2}\".format(user, remote_address, cmd))\n\n\t\tif cmd in ['q', 'quit', 'exit']:\n\t\t\tmessage.send(\"Hamilton@{0}:~$ {1}@{2} is now being disconnected.\\n\".format(hostname, user, remote_address))\n\t\t\tself.hold_prompt(message, True, user, remote_address); message.close()\n\t\t\tself.log_book.info(\"[*] Disconnected {0}@{1}..\".format(user, remote_address))\n\t\telif cmd in ['-H', '-halt']:\n\t\t\tself.hold_prompt(message, False, user, remote_address)\n\t\telif cmd == \"interests\":\n\t\t\tmessage.send(\"[Hamilton@{0}] In the Interest of the Nation: {1}\\n\".format(hostname, str(self.interests('example'))))\n\t\t\tself.hold_prompt(message, False, user, remote_address)\n\t\telif cmd == \"add\":\n\t\t\tmessage.send(\"[Hamilton@{0}] In the Interest of the Nation: {1}\\n\".format(hostname, str(self.radio('example', 'NEW_DATA'))))\n\t\t\tself.hold_prompt(message, False, user, remote_address)\n\t\telse:\n\t\t\tmessage.send(\"[Hamilton@{0}] Couldn't recognize the command input from {1}\\n\".format(hostname, user))\n\t\t\tself.hold_prompt(message, False, user, remote_address)\n\n\tdef user_prompt(self, message, cmd):\n\t\tmessage.send('{0}@{1}:~$ '.format(cmd, socket.gethostname()))\n\n\tdef client_connection(self, client, address):\n\t\tif self.user_rights(address[0]):\n\t\t\tt = paramiko.Transport(client);t.set_gss_host(socket.getfqdn(\"\"))\n\t\t\tt.load_server_moduli(); t.add_server_key(self.host_key)\n\t\t\tt.use_compression(compress=True); server = Server(remote=address[0], log_book=self.log_book)\n\t\t\tt.start_server(server=server); message = t.accept(timeout=None)\n\t\t\tmessage.send('\\n')\n\t\t\tfor e in self.motd.split('\\\\n'):\n\t\t\t\tmessage.send(e+'\\n')\n\t\t\tmessage.send('\\n\\n')\n\t\t\tself.user_prompt(message, server.user); self.ask_for_prompt(message, server.user, self.remote_user)\n\t\telse: self.sock.close()","sub_path":"libs/interfaces/ssl/interface.py","file_name":"interface.py","file_ext":"py","file_size_in_byte":4454,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"375792315","text":"import sys\nimport argparse\nfrom pathlib import Path\nimport numpy as np\n\ndef do_work(input1_file, output1_file, parameter1):\n output_text = ''\n\n output1_file.write(output_text)\n\ndef main(args):\n parser = argparse.ArgumentParser(description='Time-Series-Decomposition Input and Output')\n\n # 算子输入参数\n parser.add_argument(\"--input1\", type=str, required=True)\n parser.add_argument(\"--parameter1\", type=str, required=True)\n\n # 算子输出, 系统会自动分配一个路径, 输出值必须写入这个路径才会被系统识别打包存到S3, 为其他算子输入引用做准备\n parser.add_argument(\"--output1\", type=str, required=True)\n args = parser.parse_args(args)\n\n # 创建输出路径,存放输出文件(路径可以存在,也可以不存在)\n Path(args.output1).parent.mkdir(parents=True, exist_ok=True)\n\n with open(args.input1, 'r') as input1_file:\n with open(args.output1, 'w') as output1_file:\n do_work(input1_file, output1_file, args.parameter1)\n\nif __name__ == '__main__':\n main(sys.argv[1:])\n","sub_path":"price-forecast/modules/Samples/src/Sample.py","file_name":"Sample.py","file_ext":"py","file_size_in_byte":1081,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"131510584","text":"import os\nimport sys\n\npath=input(\"Prosím zadejte pozici spojeneckých pozemních jednotek (winpath) -> \")\nos.chdir(path[:2])\nos.chdir(path[2:])\ncount=input(\"Zadejte počet pro odstranění -> \")\nfor i in range(1, int(count)):\n os.rmdir(str(i))\nsys.exit()\n","sub_path":"python/bomby/ANTIBOMBA.py","file_name":"ANTIBOMBA.py","file_ext":"py","file_size_in_byte":259,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"326955114","text":"import csv\r\n\r\ndef fix_nulls(s):\r\n for line in s:\r\n yield line.replace('\\0', ' ')\r\n\r\ncdata = []\r\nticket = []\r\nindividualcandidate = []\r\n#candidates\r\nwith open(\"SenateDivisionFirstPrefsByPollingPlaceDownload-20499-211.csv\", 'r') as csvfile:\r\n reader = csv.reader(csvfile)\r\n for row in reader:\r\n if row[4] == \"Balnarring\":\r\n if row[9] == \"\":\r\n #ticket\r\n if row[8] != \"INFORMAL\":\r\n ticket.append([row[8],[]])\r\n else:\r\n #individual\r\n individualcandidate.append([row[8] + \":\" + row[9],[]])\r\n \r\ncdata = ticket + individualcandidate\r\n##for key in cdata:\r\n## print(key)\r\n\r\n#prefs\r\nwith open(\"aec-senate-formalpreferences-20499-VIC.csv\", 'r') as csvfile:\r\n reader = csv.reader(csvfile)\r\n count = 0\r\n for row in reader:\r\n if count % 100000 == 0:\r\n for j in cdata:\r\n print(j[0] + \":\" + str(len(j[1])))\r\n print(count)\r\n if \"ElectorateNm\" not in row and \"------------\" not in row:\r\n i = 0\r\n for vote in row[5].split(\",\"):\r\n if vote == str(1):\r\n break\r\n i = i + 1\r\n if(i < len(cdata)):\r\n cdata[i][1].append(row[5].split(\",\"))\r\n count = count + 1\r\n\r\nfor j in cdata:\r\n print(j[0] + \":\" + str(len(j[1])))\r\n","sub_path":"senatetest.py","file_name":"senatetest.py","file_ext":"py","file_size_in_byte":1393,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"404239453","text":"import requests, json\nimport os\nimport pandas as pd\n# Imports the Google Cloud client library\nfrom google.cloud import language\nfrom google.cloud.language import enums\nfrom google.cloud.language import types\n\n# Try it out by opening Python and typing\n# from geoparse_documents import annotateFile; f=annotateFile(); from googleNLP import parseWithGoogle; parseWithGoogle(f.pdf_words_clean[0:10000], print_input=True, print_output=True)\n\ndef parseWithGoogle(text, print_output=False, print_input=False):\n \"\"\" Makes request to Google Cloud and returns a df\n\n Input\n text String. Do not pass lists, dictionaries, etc. Basic strings only.\n\n Returns\n Pandas dataframe with each extracted location\n \"\"\"\n client = language.LanguageServiceClient()\n\n if print_input:\n print(text)\n\n # Instantiates a plain text document.\n document = types.Document(\n content=text,\n type=enums.Document.Type.PLAIN_TEXT)\n\n entities = client.analyze_entities(document).entities\n\n # entity types from enums.Entity.Type\n entity_type = ('UNKNOWN', 'PERSON', 'LOCATION', 'ORGANIZATION',\n 'EVENT', 'WORK_OF_ART', 'CONSUMER_GOOD', 'OTHER')\n if print_output:\n for entity in entities:\n print('=' * 20)\n print(u'{:<16}: {}'.format('name', entity.name))\n print(u'{:<16}: {}'.format('type', entity_type[entity.type]))\n print(u'{:<16}: {}'.format('metadata', entity.metadata))\n print(u'{:<16}: {}'.format('salience', entity.salience))\n print(u'{:<16}: {}'.format('wikipedia_url',\n entity.metadata.get('wikipedia_url', '-')))\n\n # Convert output to a pandas dataframe\n df = pd.DataFrame(data=None, index=range(0, len(entities)), columns=['name', 'entity_type', 'salience', 'wikipedia_url', 'google_knowledge_graph'])\n for ix, entity in enumerate(entities):\n df.loc[ix] = [\n entity.name, \n entity_type[entity.type], \n entity.salience, \n entity.metadata.get('wikipedia_url', None),\n entity.metadata.get('mid', None)]\n\n # Only return location entities\n df = df[df.entity_type == 'LOCATION']\n return df\n\n\ndef _get_api_key():\n key = os.environ.get('googleCloud_key')\n if not key:\n raise PermissionError('No Geoparser IO key found. Request one on www.geoparser.io and set as environment variable.')\n return(key)\n\n\ndef _flatten_into_df(data_to_flatten):\n \"\"\" Takes data in list or dict format.\n\n Input\n data_to_flatten a dictionary or list containing the data to end up in the dataframe. \n \"\"\"\n \n # check if data is a dictionary - if it is, make it into a list\n if isinstance(data_to_flatten, dict):\n data_to_flatten = [data_to_flatten]\n\n for idx, item in enumerate(data_to_flatten):\n df_element = pd.io.json.json_normalize(item)\n if idx == 0:\n df_temp = df_element\n else:\n df_temp = df_temp.append(df_element)\n\n return df_temp","sub_path":"googleNLP.py","file_name":"googleNLP.py","file_ext":"py","file_size_in_byte":3030,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"337408681","text":"import os\nfrom os.path import join, exists\nimport numpy as np\nimport argparse\nimport json\n\nimport torch\n\nfrom model.sketch_rnn import SketchRNN\nfrom dataloader.dataset import build_line, process_label_file\n\n\ndef predict(config):\n device = torch.device(\"cuda\" if torch.cuda.is_available() else 'cpu')\n\n sketch = []\n key_id = []\n with open(config.test_file, 'r') as f:\n lines = f.readlines()\n for i in range(1, len(lines)):\n line = lines[i].strip()\n drawing = line.split('\\\"')[1]\n drawing = build_line(json.loads(drawing))\n sketch += [drawing]\n id = line.split(',')[0]\n key_id += [id]\n\n dictionary, reverse_dict = process_label_file(config.label_file)\n\n sketchrnn = SketchRNN(config.input_size, config.hidden_size, output_size=len(dictionary.keys()),\n n_layers=config.n_layers, device=device).to(device)\n\n if config.load_model:\n sketchrnn.load_state_dict(torch.load(config.load_model))\n\n submission = open(config.submission_file, 'w')\n submission.write('key_id,word\\n')\n\n sketchrnn.eval()\n for i in range(len(sketch)):\n s = torch.from_numpy(np.array([sketch[i]])).to(device)\n seq_len = torch.Tensor([s.shape[1]]).int().to(device)\n hidden = sketchrnn.initHidden(s.shape[0]).to(device)\n\n output, hidden, new_idx = sketchrnn(s, hidden, seq_len)\n output = np.argsort(output.detach().cpu().numpy()[0])[::-1][0:3]\n\n submission.write(key_id[i] + ',')\n for j, o in enumerate(output, 1):\n o = reverse_dict[o].replace(' ', '_')\n if j < 3:\n submission.write(o + ' ')\n else:\n submission.write(o + '\\n')\n\n submission.close()\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument('--test_file', type=str, default='data/test_simplified.csv')\n parser.add_argument('--submission_file', type=str, default='data/submission.csv')\n parser.add_argument('--label_file', type=str, default='/media/liuwq/data/Dataset/quick draw/label.csv')\n parser.add_argument('--input_size', type=int, default=3)\n parser.add_argument('--hidden_size', type=int, default=100)\n parser.add_argument('--n_layers', type=int, default=2)\n parser.add_argument('--load_model', type=str, default=None)\n\n config = parser.parse_args()\n predict(config)\n\n\n","sub_path":"pytorch/predict.py","file_name":"predict.py","file_ext":"py","file_size_in_byte":2409,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"88260140","text":"import numpy as np\nimport random\n\nfrom PIL import ImageOps\nimport torch\nfrom torchvision import transforms\nfrom skimage.transform import resize\n\n\nclass Compose(object):\n \"\"\"引数transformに格納された変形を順番に実行するクラス\n 対象画像とアノテーション画像を同時に変換させます。 \n \"\"\"\n\n def __init__(self, transforms):\n self.transforms = transforms\n\n def __call__(self, img, anno_class_img):\n for t in self.transforms:\n img, anno_class_img = t(img, anno_class_img)\n return img, anno_class_img\n\nclass RandomCrip(object):\n def __init__(self, cut_size):\n self.cut_size = cut_size\n def __call__(self, img, anno_class_img):\n width = img.shape[0]\n height = img.shape[1]\n\n start_w = random.randint(0, width - self.cut_size)\n start_h = random.randint(0, height - self.cut_size)\n\n img = img[start_w:start_w+self.cut_size, start_h:start_h+self.cut_size]\n anno_class_img = anno_class_img[start_w:start_w+self.cut_size, start_h:start_h+self.cut_size]\n\n return img, anno_class_img\n\nclass Mono2Color(object):\n def to8bit(self, img):\n img_8bit = img * (255/ (img.max()-img.min())) \n img_8bit = np.uint8(img_8bit)\n return img_8bit\n def padding_median(self,image):\n img = image.copy()\n med = np.median(img)\n \n img[img==0] = med\n return img \n def image_transform(self, img, mode):\n img = self.padding_median(img)\n if mode==0:\n img = self.to8bit(img)\n elif mode==1:\n img = np.log10(img+1e-2)\n img = self.to8bit(img)\n else:\n img_log = np.log10(img+1e-2)\n img_down = self.to8bit(img)\n img_down_log = self.to8bit(img_log)\n img = (img_down + img_down_log)/2\n return img \n def mono2color(self, img):\n img_1ch = self.image_transform(img,0)\n img_2ch = self.image_transform(img,1)\n img_3ch = self.image_transform(img,2)\n\n return np.array([img_1ch, img_2ch, img_3ch], dtype=np.uint8)\n\n def __call__(self, img, anno_class_img):\n img = self.mono2color(img)\n img = img.transpose(1,2,0)\n return img, anno_class_img\n\n\nclass Resize(object):\n def __init__(self, rate):\n self.rate = rate\n \n def resize_img(self, img, rate):\n height = int(img.shape[0] * rate)\n width = int(img.shape[1] * rate)\n img_resized = resize(img, (height, width))\n return img_resized \n\n def __call__(self, img, anno_class_img):\n # img : numpy\n # anno_class_img : pillow\n img = self.resize_img(img, self.rate)\n anno_class_img = anno_class_img.resize(\n (int(anno_class_img.size[0]*self.rate),int(anno_class_img.size[1]*self.rate)),\n Image.NEAREST)\n anno_class_img = np.array(anno_class_img)\n return img, anno_class_img\n\nclass RandomMirror(object):\n \"\"\"50%の確率で左右反転させるクラス\"\"\"\n\n def __call__(self, img, anno_class_img):\n if np.random.randint(2):\n img = np.fliplr(img)\n anno_class_img = np.fliplr(anno_class_img)\n return img, anno_class_img\n\t\t\t\nclass RandomRotation(object):\n def __call__(self, img, anno_class_img):\n if np.random.randint(2):\n i = np.random.randint(1,4)\n img = np.rot90(img, i)\n anno_class_img = np.rot90(anno_class_img, i)\n return img, anno_class_img\n\n\nclass Normalize_Tensor(object):\n def __init__(self, data_mean, data_std):\n self.data_mean = data_mean\n self.data_std = data_std\n\n def __call__(self, img, anno_class_img):\n # img を Tensor に変換\n img = transforms.functional.to_tensor(img.copy())\n\n # dataの標準化\n img = transforms.functional.normalize(\n img, self.data_mean, self.data_std\n )\n # annotation img を Tensor に変換\n anno_class_img = torch.from_numpy(anno_class_img.copy())\n\n return img, anno_class_img\n","sub_path":"utils/data_augumentation.py","file_name":"data_augumentation.py","file_ext":"py","file_size_in_byte":4084,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"99028147","text":"\"\"\"\n井字棋\n\nVersion: 1.0\nAuthor: large-rabbit\n\"\"\"\nimport os\n\n\ndef print_board(map_arr):\n \"\"\"\n 画出棋盘\n\n :param map_arr: 棋盘的地图数组\n :return: None\n \"\"\"\n os.system('cls')\n for col in range(3):\n print(map_arr[col * 3 + 0] + '|' + map_arr[col * 3 + 1] + '|' + map_arr[col * 3 + 2])\n if col != 2:\n print('-+-+-')\n\n\ndef decide_win(map_arr):\n \"\"\"\n 判断输赢\n\n :param map_arr: 棋盘的地图数组\n :return: 胜利方:'X'、'O'、'平局'\n \"\"\"\n who_win = decide_row(map_arr)\n if who_win == '平局':\n who_win = decide_col(map_arr)\n if who_win == '平局':\n who_win = decide_across(map_arr)\n return who_win\n\n\ndef decide_row(map_arr):\n \"\"\"\n 判断横向是否有胜负\n\n :param map_arr: 棋盘的地图数组\n :return: 胜利方:'X'、'O'、'平局'\n \"\"\"\n for col in range(3):\n tem_arr = map_arr[col*3:col*3+3]\n result = detect_win(tem_arr)\n if result != '平局':\n return result\n return '平局'\n\n\ndef decide_col(map_arr):\n \"\"\"\n 判断纵向是否有胜负\n\n :param map_arr: 棋盘的地图数组\n :return: 胜利方:'X'、'O'、'平局'\n \"\"\"\n\n for row in range(3):\n tem_arr = [map_arr[col * 3 + row] for col in range(3)]\n result = detect_win(tem_arr)\n if result != '平局':\n return result\n return '平局'\n\n\ndef decide_across(map_arr):\n \"\"\"\n 判断对角是否有胜负\n\n :param map_arr: 棋盘的地图数组\n :return: 胜利方:'X'、'O'、'平局'\n \"\"\"\n tem_arr = [map_arr[0], map_arr[4], map_arr[8]]\n result = detect_win(tem_arr)\n if result != '平局':\n return result\n tem_arr = [map_arr[2], map_arr[4], map_arr[6]]\n result = detect_win(tem_arr)\n if result != '平局':\n return result\n return '平局'\n\n\ndef detect_win(tem_arr):\n if tem_arr == ['X', 'X', 'X']:\n return 'X胜'\n elif tem_arr == ['O', 'O', 'O']:\n return 'O胜'\n else:\n return '平局'\n\n\ndef main():\n is_again = True\n while is_again:\n map_arr = [' '] * 9\n is_continue = True\n who_win = 0\n turn = 'X'\n count = 0\n print_board(map_arr)\n while is_continue:\n is_input = True\n while is_input:\n print('{0}出棋'.format(turn))\n try:\n x = int(input('请输入X坐标(从0开始,2结束):'))\n y = int(input('请输入y坐标(从0开始,2结束):'))\n if 0 <= x <= 2 and 0 <= y <= 2 and map_arr[3 * y + x] == ' ':\n map_arr[3 * y + x] = turn\n turn = 'O' if turn == 'X' else 'X'\n is_input = False\n else:\n print('坐标有误,请重新输入')\n except ValueError:\n print('坐标有误,请重新输入')\n print_board(map_arr)\n who_win = decide_win(map_arr)\n if who_win == '平局':\n if count == 8:\n is_continue = False\n else:\n is_continue = True\n else:\n is_continue = False\n count += 1\n os.system('cls')\n print('=' * 12)\n print('该局为:{0}'.format(who_win))\n print('=' * 12)\n choice = input('是否再来一局?yes|no:')\n is_again = choice == 'yes'\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"game/Tic-tac-toe-game.py","file_name":"Tic-tac-toe-game.py","file_ext":"py","file_size_in_byte":3549,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"609223547","text":"# show = '''\n# ====自动化部署监控与报警====\n# > 0. 批量免密登陆 shell\n# > 1. agent端分发并启动 fabric\n# > 2. server端启动 redis/监控/报警\n# > 3. onestep 查看agent端实时数据\n# > 4. 一键执行全部(1~4) 确认顺序后执行\n# ==========================='''\n# while 1:\n# print(show)\n# choice = int(input('>>').strip())\n# if not choice <= 4: continue\nfrom fabric.api import *\n\nenv.user = 'root'\nenv.hosts = ['10.0.0.11', '10.0.0.12']\n\n\ndef deploy():\n put('../agent/monitor.py', '/tmp/monitor.py') # 测试路径\n run('pip3 install psutil && nohup python3 /tmp/monitor.py &')\n\n\ndef show_data(): # 查询所有数据\n from db_links import redis_link\n for agent_ip in redis_link.keys():\n print(agent_ip, redis_link.hgetall(agent_ip))\n\n\ndef onestep():\n # local('yum install -y ') # ??? 失误,忘记了\n local('pip3 install redis wechatpy smtplib pycrypto cryptography ')\n local('nohup python3 ./recv.py &')\n local('nohup python3 ./alarm.py &')\n show_data()\n","sub_path":"server/start.py","file_name":"start.py","file_ext":"py","file_size_in_byte":1051,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"317537984","text":"from tkinter import *\n\ndef btn_soma():\n\tif (str(ed.get()).isnumeric() and str(ed1.get()).isnumeric()):\n\t\tsoma = float(ed.get()) + float(ed1.get())\n\t\tres['text'] = str(soma)\n\t\tres['fg'] = 'blue'\n\telse:\n\t\tres['text'] = 'Digite um número válido'\n\t\tres['fg'] = 'red'\n\t \t\njanela = Tk()\njanela.geometry('300x140+300+300')\n\ned = Entry(janela)\ned.place(x=90,y=10)\n\ned1 = Entry(janela)\ned1.place(x=90,y=40)\n\nres = Label(janela, text='>>')\nres.place(x=85,y=70)\n\nbtn = Button(janela, text='Somar', command = btn_soma)\nbtn.place(x=125,y=110)\n\njanela.mainloop()\n\n#soma = float(ed.get()) + float(ed1.get())\n\n#and ed1.get() != '': ","sub_path":"files/tkinter/calculadora.py","file_name":"calculadora.py","file_ext":"py","file_size_in_byte":618,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"141907239","text":"#!/usr/bin/env python\n\nimport os\nfrom bs4 import BeautifulSoup # BeautifulSoup4 package\nimport urllib.request\n\n# Grab the HTML from a web page just like we did\n# in the first example\nmy_address = \"https://docs.python.org/3.4/whatsnew/3.4.html\"\n\n# Open my_address, read page and decode from bytes to text\nwith urllib.request.urlopen(my_address) as html_page:\n html_text = html_page.read().decode(\"utf-8\")\n\n# Pass the HTML to the BeautifulSoup constructor.\n# The second argument tells beautiful soup which parser to use\nsoup = BeautifulSoup(html_text, \"lxml\")\n\nresult = soup.get_text()\ntext = os.linesep.join([s for s in result.splitlines() if s])\nprint(text)\n","sub_path":"Stream-2/Unit29-Web_Scraping/2-Scraping_With_an_HTML_Parser/beautiful_soup_test.py","file_name":"beautiful_soup_test.py","file_ext":"py","file_size_in_byte":662,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"403440673","text":"from selenium import webdriver\nfrom thonny import get_workbench\nfrom selenium.common import exceptions\nfrom selenium.webdriver.common.desired_capabilities import DesiredCapabilities\nfrom tkinter.simpledialog import askstring\nfrom threading import Thread\nimport time\n\n\nclass Singleton:\n \"\"\"This Singleton class is needed so you do not spawn more selenium contexts when\n opening a new link or working on a past context.\n\n __instance: The current instance of this class.\n closed: A variable that indicates if the current browser window has been closed by\n the user.\n sleeptime: The seconds to wait after every check if an HTML element has changed.\n observed_ids: The HTML element ids which are observed by the observe method.\n \"\"\"\n\n __instance = None\n closed = False\n sleeptime = 2\n observed_ids = []\n\n @staticmethod\n def getInstance():\n \"\"\"This is a modified Singleton.getInstance() method. It will create a new\n instance if the variable closed is set to true.\n\n Returns:\n Nested Singleton object with an initialized webdriver using a selenium\n context\n \"\"\"\n if Singleton.__instance == None or Singleton.closed:\n Singleton()\n return Singleton.__instance\n\n def __init__(self):\n \"\"\"This is a modified Singleton constructor. Just like Singleton.getInstance()\n it will create a new instance if the variable closed is set to true. Before\n creating the new context, security settings will also be set for Firefox. Other\n browsers only support the W3C compliant capabilities (\n https://www.w3.org/TR/webdriver/#capabilities).\n \"\"\"\n if Singleton.__instance != None and Singleton.closed == False:\n raise Exception(\"This class is a singleton!\")\n else:\n try:\n firefox_capabilities = DesiredCapabilities.FIREFOX\n # Only communicate over HTTPS and do not accept insecure Certificates\n firefox_capabilities[\"handleAlerts\"] = False\n firefox_capabilities[\"acceptSslCerts\"] = False\n firefox_capabilities[\"acceptInsecureCerts\"] = False\n profile = webdriver.FirefoxProfile()\n profile.accept_untrusted_certs = False\n profile.set_preference(\"network.http.use-cache\", False)\n profile.set_preference(\"dom.security.https_only_mode\", True)\n # Disable unsafe cipher\n profile.set_preference(\"security.ssl3.rsa_des_ede3_sha\", False)\n # Disable ciphers with no forward secrecy\n profile.set_preference(\"security.ssl3.dhe_rsa_aes_128_sha\", False)\n profile.set_preference(\"security.ssl3.dhe_rsa_aes_256_sha\", False)\n profile.set_preference(\"security.ssl3.rsa_aes_128_sha\", False)\n profile.set_preference(\"security.ssl3.rsa_aes_256_sha\", False)\n profile.set_preference(\"security.ssl3.ecdhe_ecdsa_aes_128_sha\", False)\n profile.set_preference(\"security.ssl3.ecdhe_ecdsa_aes_256_sha\", False)\n profile.set_preference(\"security.ssl3.ecdhe_rsa_aes_128_sha\", False)\n profile.set_preference(\"security.ssl3.ecdhe_rsa_aes_256_sha\", False)\n self.driver = webdriver.Firefox(\n capabilities=firefox_capabilities, firefox_profile=profile\n )\n except BaseException:\n try:\n # Chromium Command Line Switches https://peter.sh/experiments/chromium-command-line-switches/\n options = webdriver.ChromeOptions()\n options.set_capability(\n \"chrome.switches\",\n \"--cipher-suite-blacklist TLS_RSA_WITH_3DES_EDE3_SHA,\"\n + \"TLS_DH_RSA_WITH_AES_128_SHA,TLS_DH_RSA_WITH_AES_256_SHA,\"\n + \"TLS_RSA_WITH_AES_128_SHA,TLS_RSA_WITH_AES_256_SHA,\"\n + \"TLS_ECDH_ECDSA_WITH_AES_128_SHA,TLS_ECDH_ECDSA_WITH_AES_256_SHA,\"\n + \"TLS_ECDH_RSA_WITH_AES_128_SHA,TLS_ECDH_RSA_WITH_AES_256_SHA\",\n )\n self.driver = webdriver.Chrome(chrome_options=options)\n except BaseException:\n try:\n self.driver = webdriver.Safari()\n except BaseException:\n try:\n self.driver = webdriver.Edge()\n except BaseException:\n try:\n caps = DesiredCapabilities.OPERA[\n \"chrome.switches\"\n ] = \"--cipher-suite-blacklist TLS_RSA_WITH_3DES_EDE3_SHA,TLS_DH_RSA_WITH_AES_128_SHA,TLS_DH_RSA_WITH_AES_256_SHA,TLS_RSA_WITH_AES_128_SHA,TLS_RSA_WITH_AES_256_SHA,TLS_ECDH_ECDSA_WITH_AES_128_SHA,TLS_ECDH_ECDSA_WITH_AES_256_SHA,TLS_ECDH_RSA_WITH_AES_128_SHA,TLS_ECDH_RSA_WITH_AES_256_SHA\"\n self.driver = webdriver.Opera(desired_capabilities=caps)\n except BaseException:\n try:\n self.driver = webdriver.Ie()\n except Exception as e:\n print(str(e))\n print(\n \"You seem to be using a browser and webdriver \"\n + \"which are not supported by selenium. You can \"\n + \"find the list of supported browsers here: \"\n + \"https://www.selenium.dev/documentation/en/\"\n + \"getting_started_with_webdriver/browsers/\"\n )\n Singleton.closed = False\n Singleton.__instance = self\n\n def toggle_closed(self):\n \"\"\"This method toggles the closed variable.\n\n Returns:\n None\n \"\"\"\n Singleton.closed = not Singleton.closed\n\n def get_sleeptime(self):\n \"\"\"This method gets the current sleep time. The time to wait after every check\n if an HTML element has changed.\n\n Returns:\n float in seconds\n \"\"\"\n return Singleton.sleeptime\n\n def set_sleeptime(self, sleeptime):\n \"\"\"This method sets the sleep time. The time to wait after every check if an\n HTML element has changed.\n\n :param sleeptime: Seconds to sleep represented as float\n\n Returns:\n None\n \"\"\"\n Singleton.sleeptime = sleeptime\n\n def get_observed_ids(self):\n \"\"\"Get the list of HTML element ids which are currently observed.\n\n Returns:\n list of HTML element ids\n \"\"\"\n return Singleton.observed_ids\n\n def add_observed_id(self, html_id):\n \"\"\"This method adds one HTML element id to the list of currently observed ids.\n\n :param html_id: HTML element id that you wish to add\n\n Returns:\n None\n \"\"\"\n Singleton.observed_ids.append(html_id)\n\n def remove_observed_id(self, html_id):\n \"\"\"This method removes one HTML element id to the list of currently observed\n ids.\n\n :param html_id: HTML element id that you wish to remove\n\n Returns:\n None\n \"\"\"\n Singleton.observed_ids.remove(html_id)\n\n\ndef open_website():\n \"\"\"This method gets called if the \"Open Website\" command is clicked in the \"tools\"\n menu. It will get the Singleton instance and then tries to open the desired\n website. Insecure Certificates will be shown as an error page to the user. If the\n browser window has been closed by the user, a new selenium context will be created.\n\n Returns:\n None\n \"\"\"\n singleton = Singleton.getInstance()\n address = askstring(\"Website\", \"Which website would you like to visit?\")\n try:\n singleton.driver.get(address)\n except exceptions.InsecureCertificateException:\n return\n except exceptions.WebDriverException:\n singleton.toggle_closed()\n singleton = Singleton.getInstance()\n singleton.driver.get(address)\n\n\ndef observe_element_in_background():\n \"\"\"This method gets called by the start_observing_element_by_id method. If the text\n of the HTML element changes, the info will be printed to stdout.\n\n Returns:\n None\n \"\"\"\n singleton = Singleton.getInstance()\n driver = singleton.driver\n url = driver.current_url\n observed_ids = singleton.get_observed_ids()\n list_id = len(observed_ids) - 1\n html_id = observed_ids[list_id]\n observed_text = driver.find_element_by_id(html_id).text\n print(\"Start observing on the following URL: \" + url)\n print(\"And the following ID: \" + html_id)\n while html_id in singleton.get_observed_ids():\n try:\n element = driver.find_element_by_id(html_id)\n if element.text != observed_text:\n print(\"Text of \" + html_id + \" changed to: \" + element.text)\n observed_text = element.text\n except exceptions.NoSuchElementException as e:\n print(e)\n print(\"Observing this element is not possible as it does not exist.\")\n return\n except exceptions.WebDriverException as e:\n print(\n \"The Browser window was closed. Observing this element is not possible\"\n + \" anymore\"\n )\n print(e)\n return\n time.sleep(singleton.get_sleeptime())\n print(\"Observing was stopped as requested.\")\n\n\ndef start_observing_element_by_id():\n \"\"\"This method gets called if the \"Start observing element by id\" command is\n clicked in the \"tools\" menu. It will prompt the user to provide an HTML ID which\n will be added to the list of observed elements. After that, a new thread will be\n created to observe this HTML ID.\n\n Returns:\n None\n \"\"\"\n observe_id = askstring(\n \"ID to Observe\", \"Which HTML element ID would you like to observe?\"\n )\n singleton = Singleton.getInstance()\n singleton.add_observed_id(observe_id)\n t = Thread(target=observe_element_in_background)\n t.daemon = True\n t.start()\n\n\ndef stop_observing_element_by_id():\n \"\"\"This method gets called if the \"Stop observing element by id\" command is clicked\n in the \"tools\" menu. It will prompt the user to provide an HTML ID which will be removed\n from the list of observed elements.\n\n Returns:\n None\n \"\"\"\n observe_id = askstring(\n \"ID to Observe\", \"Which HTML element ID would you like to stop observing?\"\n )\n singleton = Singleton.getInstance()\n singleton.remove_observed_id(observe_id)\n\n\ndef load_plugin():\n \"\"\"This method gets called if this plugin is in the PYTHONPATH environment variable\n upon starting thonny. This code is executed before TK windows are drawn. That is\n why you should use add a command to the thonny GUI before running anything.\n\n Returns:\n None\n \"\"\"\n get_workbench().add_command(\n command_id=\"webview_open_website\",\n menu_name=\"tools\",\n command_label=\"Open website\",\n handler=open_website,\n )\n get_workbench().add_command(\n command_id=\"webview_observing_add\",\n menu_name=\"tools\",\n command_label=\"Start observing element by id\",\n handler=start_observing_element_by_id,\n )\n get_workbench().add_command(\n command_id=\"webview_observing_delete\",\n menu_name=\"tools\",\n command_label=\"Stop observing element by id\",\n handler=stop_observing_element_by_id,\n )\n","sub_path":"thonnycontrib/thonny-webdriver.py","file_name":"thonny-webdriver.py","file_ext":"py","file_size_in_byte":11840,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"395982936","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# Author: maplebeats\n# gtalk/mail: maplebeats@gmail.com\n\nimport logging\n\nlogger = logging.getLogger()\nFORMAT = '%(levelname)s %(module)s %(message)s'\nformatter = logging.Formatter(FORMAT)\nhdlr = logging.StreamHandler()\nhdlr.setFormatter(formatter)\nlogger.addHandler(hdlr)\nlogger.setLevel(logging.DEBUG)\n","sub_path":"logger.py","file_name":"logger.py","file_ext":"py","file_size_in_byte":349,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"630923828","text":"import codecs\r\n# for working with files\r\nclass File:\r\n # default constructor\r\n def __init__(self, name, mode):\r\n self.name = name\r\n self.mode = mode\r\n self.encoding = \"\"\r\n self.error = False\r\n self.errorMessage = \"\"\r\n \r\n # alternate constructor with ENCODING\r\n def File(self, name, mode, encoding):\r\n self.__init__(name, mode)\r\n self.encoding = encoding\r\n \r\n # gets probability ENCODING\r\n def getEncoding(self, encodings):\r\n result = \"\"\r\n for encoding in encodings:\r\n try:\r\n file1 = codecs.open(self.name, self.mode, encoding)\r\n file1.readline()\r\n self.error = False\r\n self.errorMessage = \"\"\r\n result = encoding\r\n except UnicodeDecodeError as e:\r\n self.error = True\r\n self.errorMessage = e\r\n except Exception as e:\r\n self.error = True\r\n self.errorMessage = e\r\n break\r\n finally:\r\n file1.close()\r\n if self.error == False:\r\n self.encoding = result\r\n return result\r\n # get first line for checking\r\n def getFirstLine(self):\r\n result = \"\"\r\n if len(self.encoding) == 0:\r\n return result\r\n try:\r\n file1 = codecs.open(self.name, self.mode, self.encoding)\r\n result = file1.readline()\r\n self.error = False\r\n self.errorMessage = \"\"\r\n except UnicodeDecodeError as e:\r\n self.error = True\r\n self.errorMessage = e\r\n except Exception as e:\r\n self.error = True\r\n self.errorMessage = e\r\n finally:\r\n file1.close()\r\n if self.error == False:\r\n return result\r\n ","sub_path":"lib/clFile.py","file_name":"clFile.py","file_ext":"py","file_size_in_byte":1885,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"648191225","text":"# -*- coding: utf-8 -*-\nimport os\nimport iso8601\n\nfrom copy import deepcopy\nfrom time import sleep\nfrom datetime import timedelta\n\nfrom openprocurement.api.tests.blanks.json_data import test_document_data\nfrom openprocurement.api.tests.base import PrefixedRequestClass\nfrom openprocurement.contracting.ceasefire.constants import ENDPOINTS\nfrom openprocurement.contracting.ceasefire.tests import base\nfrom openprocurement.contracting.ceasefire.tests.fixtures.data import contract_create_data\nfrom openprocurement.contracting.ceasefire.utils import view_milestones_by_type\n\n\nclass CeasefireTest(base.BaseWebTest):\n\n mock_config = base.MOCK_CONFIG\n record_http = True\n docservice = True\n\n def setUp(self):\n self.relative_to = os.path.dirname(base.__file__)\n self.app.RequestClass = PrefixedRequestClass\n self.app.authorization = ('Basic', ('broker', ''))\n self.couchdb_server = self.app.app.registry.couchdb_server\n self.db = self.app.app.registry.db\n if self.docservice:\n self.setUpDS()\n self.app.app.registry.docservice_url = 'http://public.docs-sandbox.ea.openprocurement.org'\n\n def endpoint(self, name, acc_token=None):\n prefixed = '/' + ENDPOINTS[name]\n if acc_token:\n return prefixed + '?acc_token=' + acc_token\n return prefixed\n\n @staticmethod\n def get_dateMet(dueDate_str, partiallyMet=False):\n due_date = iso8601.parse_date(dueDate_str)\n days_to_add = 2 if partiallyMet else -2\n tdelta = timedelta(days=days_to_add)\n date_met = due_date + tdelta\n return date_met.isoformat()\n\n def update_milestones(self):\n response = self.app.get(\n self.endpoint('contracts').format(\n contract_id=self.contract_id\n )\n )\n self.assertEqual(response.status, '200 OK')\n milestones = response.json['data']['milestones']\n self.milestones = view_milestones_by_type(milestones, 'type')\n\n def test_docs_tutorial(self):\n\n out_dir = 'docs/source/tutorial/'\n\n with open(out_dir + 'contracts-listing-empty.http', 'w') as self.app.file_obj:\n response = self.app.get(self.endpoint('contracts_collection'))\n self.assertEqual(response.status, '200 OK')\n\n # Create contract as a bot\n self.app.authorization = ('Basic', ('contracting', ''))\n response = self.app.post_json(\n self.endpoint('contracts_collection'),\n {'data': contract_create_data}\n )\n self.contract_id = response.json['data']['id']\n access = response.json['access']\n self.assertEqual(response.status, '201 Created')\n\n self.app.authorization = ('Basic', ('broker', ''))\n\n with open(out_dir + 'get-created-contract.http', 'w') as self.app.file_obj:\n response = self.app.get(\n self.endpoint('contracts').format(\n contract_id=self.contract_id\n )\n )\n self.assertEqual(response.status, '200 OK')\n\n with open(out_dir + 'create-transfer.http', 'w') as self.app.file_obj:\n transfer = self.app.post_json('/transfers', {'data': {}}).json\n\n with open(out_dir + 'use-transfer.http', 'w') as self.app.file_obj:\n self.app.post_json(\n '/contracts/{0}/ownership'.format(self.contract_id),\n {'data': {\n 'id': transfer['data']['id'],\n 'transfer': access['transfer']\n }}\n )\n\n acc_token = transfer['access']['token']\n\n with open(out_dir + 'patch-contract-to-active-payment.http', 'w') as self.app.file_obj:\n response = self.app.patch_json(\n self.endpoint('contracts', acc_token).format(\n contract_id=self.contract_id\n ),\n {'data': {'status': 'active.payment'}})\n\n# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n# ~~~~~~~~~~~~~~~~~~~~~Milestones~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~#\n# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n self.update_milestones()\n\n self.app.get(self.endpoint('contracts_collection'))\n sleep(0.31415926) # must use some delay to give DB time to build index\n\n with open(out_dir + 'view-contracts-after-create.http', 'w') as self.app.file_obj:\n response = self.app.get(self.endpoint('contracts_collection'))\n self.assertEqual(response.status, '200 OK')\n\n with open(out_dir + 'get-financing-milestone-processing.http', 'w') as self.app.file_obj:\n response = self.app.get(\n self.endpoint('milestones').format(\n contract_id=self.contract_id,\n milestone_id=self.milestones['financing']['id']))\n\n with open(out_dir + 'patch-financing-to-met.http', 'w') as self.app.file_obj:\n response = self.app.patch_json(\n self.endpoint('milestones', acc_token).format(\n contract_id=self.contract_id,\n milestone_id=self.milestones['financing']['id']\n ),\n {'data': {'dateMet': self.get_dateMet(self.milestones['financing']['dueDate'])}}\n )\n\n with open(out_dir + 'patch-financing-to-partially-met.http', 'w') as self.app.file_obj:\n response = self.app.patch_json(\n self.endpoint('milestones', acc_token).format(\n contract_id=self.contract_id,\n milestone_id=self.milestones['financing']['id']\n ),\n {'data': {'dateMet': self.get_dateMet(self.milestones['financing']['dueDate'], partiallyMet=True)}}\n )\n\n with open(out_dir + 'get-contract-after-financing-partiallyMet.http', 'w') as self.app.file_obj:\n response = self.app.get(\n self.endpoint('contracts').format(\n contract_id=self.contract_id\n )\n )\n self.assertEqual(response.status, '200 OK')\n\n with open(out_dir + 'get-approval-milestone.http', 'w') as self.app.file_obj:\n response = self.app.get(\n self.endpoint('milestones').format(\n contract_id=self.contract_id,\n milestone_id=self.milestones['approval']['id']))\n\n self.update_milestones()\n target_reporting_due_date = iso8601.parse_date(self.milestones['approval']['dueDate']) + timedelta(days=1200)\n\n with open(out_dir + 'patch-reporting-due-date.http', 'w') as self.app.file_obj:\n response = self.app.patch_json(\n self.endpoint('milestones', acc_token).format(\n contract_id=self.contract_id,\n milestone_id=self.milestones['reporting']['id']\n ),\n {'data': {'dueDate': target_reporting_due_date.isoformat()}}\n )\n\n with open(out_dir + 'patch-approval-to-met.http', 'w') as self.app.file_obj:\n response = self.app.patch_json(\n self.endpoint('milestones', acc_token).format(\n contract_id=self.contract_id,\n milestone_id=self.milestones['approval']['id']\n ),\n {'data': {'dateMet': self.get_dateMet(self.milestones['approval']['dueDate'])}}\n )\n\n with open(out_dir + 'get-contract-after-approval-met.http', 'w') as self.app.file_obj:\n response = self.app.get(\n self.endpoint('contracts').format(\n contract_id=self.contract_id\n )\n )\n self.assertEqual(response.status, '200 OK')\n\n with open(out_dir + 'get-reporting-processing.http', 'w') as self.app.file_obj:\n response = self.app.get(\n self.endpoint('milestones').format(\n contract_id=self.contract_id,\n milestone_id=self.milestones['reporting']['id']))\n\n self.update_milestones()\n\n with open(out_dir + 'patch-reporting-to-met.http', 'w') as self.app.file_obj:\n response = self.app.patch_json(\n self.endpoint('milestones', acc_token).format(\n contract_id=self.contract_id,\n milestone_id=self.milestones['reporting']['id']\n ),\n {'data': {'dateMet': self.get_dateMet(self.milestones['reporting']['dueDate'])}}\n )\n\n with open(out_dir + 'get-contract-after-reporting-met.http', 'w') as self.app.file_obj:\n response = self.app.get(\n self.endpoint('contracts').format(\n contract_id=self.contract_id\n )\n )\n self.assertEqual(response.status, '200 OK')\n\n# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n# ~~~~~~~~~~~~~~~~~~~~~Document operations~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~#\n# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n\n document_data = deepcopy(test_document_data)\n document_data.update({\n 'url': self.generate_docservice_url(),\n 'documentOf': 'milestone',\n 'relatedItem': self.milestones['financing']['id']\n })\n\n with open(out_dir + 'upload-document-to-milestone.http', 'w') as self.app.file_obj:\n response = self.app.post_json(\n self.endpoint('contracts_documents_collection', acc_token).format(\n contract_id=self.contract_id,\n milestone_id=self.milestones['financing']['id']\n ),\n {'data': document_data}\n )\n self.assertEqual(response.status, '201 Created')\n\n document_1_id = response.json['data']['id']\n\n with open(out_dir + 'get-contract-documents.http', 'w') as self.app.file_obj:\n response = self.app.get(\n self.endpoint('contracts_documents', acc_token).format(\n contract_id=self.contract_id,\n document_id=document_1_id\n )\n )\n","sub_path":"docs/update_tutorial_requests.py","file_name":"update_tutorial_requests.py","file_ext":"py","file_size_in_byte":10195,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"624096798","text":"def initialize_fitting_mat(s1, s2, indelPen):\n dp_mat = [[0] * (len(s2) + 1) for _ in range(len(s1) + 1)]\n\n for i in range(len(dp_mat[0])):\n dp_mat[0][i] = -(i*indelPen)\n for i in range(len(dp_mat)):\n dp_mat[i][0] = 0\n\n return dp_mat\n\n\ndef initialize_bt_mat(s1, s2):\n bt_mat = [[''] * (len(s2) + 1) for _ in range(len(s1) + 1)]\n for i in range(1, len(bt_mat)):\n bt_mat[i][0] = 'source'\n for i in range(1, len(bt_mat[0])):\n bt_mat[0][i] = 'left'\n\n return bt_mat\n\n\n\n\ndef findFittingAlignment(s1, s2):\n bt_dict = {0:'diag', 1:'left', 2:'up'}\n indelPen = 1\n dp_mat = initialize_fitting_mat(s1, s2, indelPen)\n\n bt_mat = initialize_bt_mat(s1, s2)\n\n maxVal = 0\n startPoint = []\n\n for i in range(1, len(dp_mat)):\n for j in range(1, len(dp_mat[0])):\n var = 1 if s1[i-1] == s2[j-1] else -1\n values = [dp_mat[i-1][j-1] + var, dp_mat[i][j-1] - indelPen, dp_mat[i-1][j] - indelPen]\n dp_mat[i][j] = max(values)\n bt_mat[i][j] = bt_dict[values.index(dp_mat[i][j])]\n if j == len(dp_mat[0]) - 1:\n if dp_mat[i][j] > maxVal:\n maxVal = dp_mat[i][j]\n startPoint = [i,j]\n\n print(dp_mat[startPoint[0]][startPoint[1]])\n\n return backtrack_local(startPoint, bt_mat, s1, s2)\n\n\ndef backtrack_local(startPoint, bt_mat, s1, s2):\n s1align = ''\n s2align = ''\n\n i = startPoint[0]\n j = startPoint[1]\n\n while bt_mat[i][j] != 'source':\n if bt_mat[i][j] == 'diag':\n s1align += s1[i-1]\n s2align += s2[j-1]\n\n elif bt_mat[i][j] == 'left':\n s2align += s2[j-1]\n s1align += '-'\n j -= 1\n continue\n\n elif bt_mat[i][j] == 'up':\n s2align += '-'\n s1align += s1[i-1]\n i -= 1\n continue\n\n i -= 1\n j -= 1\n\n return s1align[::-1], s2align[::-1]\n\n\n\nwith open('rosalind_ba5h.txt','r+') as file:\n s1 = file.readline().rstrip()\n s2 = file.readline().rstrip()\n\n\n alignment = findFittingAlignment(s1, s2)\n print('\\n'.join(alignment))\n","sub_path":"FittingAlignment.py","file_name":"FittingAlignment.py","file_ext":"py","file_size_in_byte":2142,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"67957450","text":"#!/usr/bin/env python2.7\n\nimport sys\nimport gzip\n\n## input VCF file\n\nprint(\"[HEADER]\")\nprint(\"[Data]\")\ncol_header = \",\".join([\"SNP Name\",\"Sample Name\",\"GC Score\",\"Allele1 - AB\", \"Allele2 - AB\",\"X\",\"Y\",\"X Raw\",\"Y Raw\",\"B Allele Freq\"])\nprint(col_header)\n\nfor line in sys.stdin:\n row = line.rstrip().split(\"\\t\")\n if line.startswith(\"#CHROM\"):\n sampleName = row[9]\n\n if not line.startswith(\"#\"):\n row = line.rstrip().split(\"\\t\")\n chrom = row[0]\n pos = row[1]\n ID = row[2]\n\n infoMap = {} \n for item in row[7].split(\";\"):\n #print(item)\n if(\"=\" in item):\n k,v = item.split(\"=\")\n infoMap[k] = v\n\n # Checking array ALLELE_A is the REF or ALT in the VCF\n # The VCF INFO tag ALLELE_A or ALLELE_B will have asterix after allele if it is REF \n if(\"*\" in infoMap[\"ALLELE_A\"]):\n is_ALLELE_A_ref = 1\n else:\n is_ALLELE_A_ref = 0\n\n # set default allele to \"N\" for final report alleles\n fr_allele1 = \"N\"\n fr_allele2 = \"N\"\n\n formatTags = row[8].split(\":\")\n genoData = row[9].split(\":\")\n genoMap = dict(zip(formatTags,genoData))\n\n geno = genoMap[\"GT\"]\n if(is_ALLELE_A_ref==1):\n if(geno==\"0/0\"):\n fr_allele1 = \"A\"\n fr_allele2 = \"A\"\n if(geno==\"0/1\"):\n fr_allele1 = \"A\"\n fr_allele2 = \"B\"\n if(geno==\"1/1\"):\n fr_allele1 = \"B\"\n fr_allele2 = \"B\"\n if(geno==\"./.\"):\n fr_allele1 = \"-\"\n fr_allele2 = \"-\"\n else:\n if(geno==\"0/0\"):\n fr_allele1 = \"B\"\n fr_allele2 = \"B\" \n if(geno==\"1/0\"):\n fr_allele1 = \"A\"\n fr_allele2 = \"B\"\n if(geno==\"1/1\"):\n fr_allele1 = \"A\"\n fr_allele2 = \"A\"\n if(geno==\"./.\"):\n fr_allele1 = \"-\"\n fr_allele2 = \"-\"\n\n outrow = [ID,sampleName,genoMap[\"IGC\"],fr_allele1,fr_allele2,genoMap[\"NORMX\"],genoMap[\"NORMY\"],genoMap[\"X\"],genoMap[\"Y\"],genoMap[\"BAF\"]]\n print(\",\".join(outrow))\n","sub_path":"parseVcfToBAFRegress.py","file_name":"parseVcfToBAFRegress.py","file_ext":"py","file_size_in_byte":1921,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"467756215","text":"#!/usr/bin/env python3\n# -*- coding:utf-8 -*-\n\nmouths = [\n 'January',\n 'February',\n 'March',\n 'April',\n 'may',\n 'July',\n 'August',\n 'September',\n 'October',\n 'November',\n 'December'\n]\n\nendings = ['st', 'nd', 'rd'] + 17 * ['th'] + ['st', 'nd', 'rd'] + 7 * ['th'] + ['st']\n\nyear = input('Year: ')\nmouth = input('Mouth: ')\nday = input('Day: ')\n\nmouth_number = int(mouth)\nday_number = int(day)\n\nmouth_name = mouths[mouth_number - 1]\nordinal = day + endings[day_number - 1]\n\nprint(mouth_name + ' ' + ordinal + ', ' + year)","sub_path":"YMD.py","file_name":"YMD.py","file_ext":"py","file_size_in_byte":550,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"92014619","text":"# -*- coding: utf-8 -*-\n\n\nclass NestedIterator(object):\n\n def __init__(self, nestedList):\n \"\"\"\n Initialize your data structure here.\n :type nestedList: List[NestedInteger]\n \"\"\"\n self.list = nestedList\n\n def next(self):\n \"\"\"\n :rtype: int\n \"\"\"\n if not self.hasNext():\n return\n v = self.list.pop(0)\n if isinstance(v, list):\n v = NestedIterator(v)\n index = 0\n while v.hasNext():\n # 生成的元素重新放回当前的 list\n self.list.insert(index, v.next())\n index += 1\n # 返回当前 list 的首位元素\n return self.list.pop(0)\n else:\n return v\n\n def hasNext(self):\n \"\"\"\n :rtype: bool\n \"\"\"\n return bool(len(self.list))\n\n\nif __name__ == '__main__':\n nestedList = [[1, 1], 2, [1, 1]]\n i, v = NestedIterator(nestedList), []\n while i.hasNext():\n v.append(i.next())\n\n print(v)\n nestedList = [1, [4, [6]]]\n i, v = NestedIterator(nestedList), []\n while i.hasNext():\n v.append(i.next())\n print(v)\n","sub_path":"leetcode/0341.flatten-nested-list-iterator/flatten-nested-list-iterator.py","file_name":"flatten-nested-list-iterator.py","file_ext":"py","file_size_in_byte":1171,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"369669158","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# All internal text data is stored as unicode, and will be compared by unicode\n\nimport re\nimport utils_helper as utils\n\n\nclass GsRewrite:\n def __init__(self, conf):\n self.conf = conf\n\n def process(self, area, name, addr, blng, blat, slng, slat, soso_st, soso_ba, soso_district):\n result = {'errno': 1, 'errmsg': 'fail to rewrite'}\n\n try:\n params = self.convert_param(area, name, addr, blng, blat, slng, slat)\n area, name, addr, blng, blat, slng, slat = (x for x in params)\n except:\n result['errno'] = 2\n return result\n\n # 非法过滤\n if utils.filter_request(area, name, addr, blng, blat, slng, slat):\n result['errno'] = 3\n return result\n\n # (name, addr) = self.split_long_name(name, addr) # ignored during small flow testing\n\n info = {}\n rwtype = self.match(area, name, addr, blng, blat, slng, slat, info, soso_st, soso_ba, soso_district)\n (area, name, addr, blng, blat, slng, slat) = self.fill(rwtype, info, area, name, addr, blng, blat, slng, slat)\n\n if not self.revert_param(area, name, addr, blng, blat, slng, slat):\n result['errno'] = 4\n return result\n\n result = {'errno': 0, 'name': name, 'addr': addr, 'lng': blng, 'lat': blat, 'rwtype': rwtype}\n return result\n\n def match(self, area, name, addr, blng, blat, slng, slat, info, soso_st, soso_ba, soso_district):\n if self.match_white(area, name, addr, blng, blat, slng, slat, info):\n return 1\n\n # 点击大于阈值\n click_threshold = self.conf.default_click_threshold\n if area in self.conf.click_threshold_by_area:\n click_threshold = self.conf.click_threshold_by_area[area]\n (click, score) = self.get_poi_score_click(area, name)\n if score > 50 or click >= click_threshold:\n return 10\n\n \"\"\"\n # name中含有站、桥等字样\n if re.search(u'桥|站|地铁|立交|机场|号线', name):\n return 20\n \"\"\"\n\n \"\"\"\n # name包含商圈名,且与POI在5km范围内\n if area in self.conf.ba:\n for ba_item in self.conf.ba[area]:\n if ba_item[0] in name and 5000 > get_distance(slng, slat, ba_item[1], ba_item[2]):\n info['ba'] = ba_item[0]\n return 50\n \"\"\"\n\n # name等于区县名\n if name in self.conf.region_list:\n info['region'] = name\n return 60\n\n # name等于省市名\n if name in self.conf.cities:\n info['city'] = name\n return 70\n\n # soso_data = utils.get_soso_rgeoc(slng, slat, self.conf.soso_srv_ip, self.conf.soso_srv_appkey, self.conf.soso_timeout)\n # soso_st = utils.get_soso_street(soso_data)\n # # soso_ba = utils.get_soso_ba(soso_data)\n # soso_district = utils.get_soso_district(soso_data)\n\n # name中已有路、街、道用区改写\n if len(addr) <= 5 and re.search(u'路|街|道', name) and soso_district and soso_district not in name and not any(region in name for region in self.conf.region_list):\n info['district'] = soso_district\n return 75\n\n \"\"\"\n # name大于8个字\n if len(re.sub(u'[a-zA-Z0-9()-]', '', name)) > 8:\n return 40\n\n # 用商圈改写\n if soso_ba:\n info['ba'] = soso_ba\n return 80 if soso_ba in name else 90\n \"\"\"\n\n # 没有商圈但有路\n if soso_st and not re.search(u'高速|快速|(一|二|三|四|五|六|七)环', soso_st):\n info['district_street'] = ''\n if soso_district and soso_district not in name and not any(region in name for region in self.conf.region_list):\n info['district_street'] = soso_district\n\n if soso_st and soso_st not in name:\n info['district_street'] += soso_st\n\n if len(addr) <= 5:\n return 100\n\n return -1\n\n def fill(self, rwtype, info, area, name, addr, blng, blat, slng, slat):\n if rwtype in [10, 60, 70]:\n addr = \"\"\n\n if rwtype == 75 and 'district' in info:\n addr = info['district']\n\n if rwtype == 100 and 'district_street' in info:\n addr = info['district_street']\n\n # extra short\n if addr in self.conf.cities:\n addr = \"\"\n\n addr = re.sub(u'\\d+(号|楼|层)*$', '', addr)\n\n return area, name, addr, blng, blat, slng, slat\n\n def get_poi_score_click(self, area, name):\n return 0, 0\n\n def match_white(self, area, name, addr, blng, blat, slng, slat, info):\n return False\n\n def split_long_name(self, name, addr):\n parts = name.split(u'区')\n if len(parts) != 2 or len(parts[0]) < 1:\n return name, addr\n\n # 区前一个字不能是下面任一\n if parts[0][-1] in u\"小一二三四五六七八九十东南西北中\":\n return name, addr\n\n name = parts[0] + u'区'\n addr = parts[1]\n\n return name, addr\n\n def convert_param(self, area, name, addr, blng, blat, slng, slat):\n area = int(area)\n blng, blat = (float(x) for x in (blng, blat))\n slng, slat = (float(x) for x in (slng, slat))\n # name和addr已经是unicode(ucs-2)\n return area, name, addr, blng, blat, slng, slat\n\n def revert_param(self, area, name, addr, blng, blat, slng, slat):\n try:\n name = name.encode('utf-8')\n addr = addr.encode('utf-8')\n return True\n except:\n return False\n","sub_path":"models/gsrewrite.py","file_name":"gsrewrite.py","file_ext":"py","file_size_in_byte":5689,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"523003520","text":"#!/usr/bin/env python\n\nimport unittest\nfrom dominion import Game, Card, Piles\nimport dominion.Card as Card\n\n\n###############################################################################\nclass Card_Mystic(Card.Card):\n def __init__(self):\n Card.Card.__init__(self)\n self.cardtype = Card.CardType.ACTION\n self.base = Card.CardExpansion.DARKAGES\n self.desc = \"+2 coin, +1 action; Name a card. Reveal the top card of your deck. If it's the named card, put it into your hand.\"\n self.name = \"Mystic\"\n self.actions = 1\n self.coin = 2\n self.cost = 5\n\n ###########################################################################\n def special(self, game, player):\n \"\"\" \" Name a card. Reveal the top card of your deck. If it's\n the named card, put it into your hand\"\"\"\n options = [{\"selector\": \"0\", \"print\": \"No guess\", \"card\": None}]\n index = 1\n for c in sorted(game.cardTypes()):\n sel = \"%s\" % index\n options.append({\"selector\": sel, \"print\": \"Guess %s\" % c.name, \"card\": c})\n index += 1\n o = player.user_input(options, \"Guess the top card\")\n if not o[\"card\"]:\n return\n c = player.next_card()\n player.reveal_card(c)\n if o[\"card\"].name == c.name:\n player.output(\"You guessed correctly\")\n player.add_card(c, Piles.HAND)\n else:\n player.output(\"You chose poorly - it was a %s\" % c.name)\n player.add_card(c, \"topdeck\")\n\n\n###############################################################################\nclass Test_Mystic(unittest.TestCase):\n def setUp(self):\n self.g = Game.TestGame(\n numplayers=1,\n initcards=[\"Mystic\"],\n badcards=[\"Tournament\", \"Fool's Gold\", \"Pooka\"],\n )\n self.g.start_game()\n self.plr = self.g.player_list(0)\n self.card = self.g[\"Mystic\"].remove()\n\n def test_play(self):\n \"\"\"No guess should still get results\"\"\"\n self.plr.add_card(self.card, Piles.HAND)\n self.plr.test_input = [\"0\"]\n self.plr.play_card(self.card)\n self.assertEqual(self.plr.actions.get(), 1)\n self.assertEqual(self.plr.coins.get(), 2)\n\n def test_good(self):\n \"\"\"When the guess is good the card should move to the hand\"\"\"\n self.plr.add_card(self.card, Piles.HAND)\n self.plr.piles[Piles.DECK].set(\"Province\")\n self.plr.test_input = [\"Province\"]\n self.plr.play_card(self.card)\n self.assertEqual(self.plr.actions.get(), 1)\n self.assertEqual(self.plr.coins.get(), 2)\n self.assertIn(\"Province\", self.plr.piles[Piles.HAND])\n self.assertTrue(self.plr.piles[Piles.DECK].is_empty())\n\n def test_bad(self):\n \"\"\"When the guess is bad the card should stay on the deck\"\"\"\n self.plr.add_card(self.card, Piles.HAND)\n self.plr.piles[Piles.DECK].set(\"Province\")\n self.plr.test_input = [\"Gold\"]\n self.plr.play_card(self.card)\n self.assertEqual(self.plr.actions.get(), 1)\n self.assertEqual(self.plr.coins.get(), 2)\n self.assertNotIn(\"Gold\", self.plr.piles[Piles.HAND])\n self.assertNotIn(\"Province\", self.plr.piles[Piles.HAND])\n self.assertEqual(self.plr.piles[Piles.DECK][-1].name, \"Province\")\n\n\n###############################################################################\nif __name__ == \"__main__\": # pragma: no cover\n unittest.main()\n\n# EOF\n","sub_path":"dominion/cards/Card_Mystic.py","file_name":"Card_Mystic.py","file_ext":"py","file_size_in_byte":3488,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"1327356","text":"\"\"\"hub\"\"\"\nimport os, time, sys\n# destinataire<>expéditeur<>message\n# 0 1 2\n\ndef envoi(param): # la direction est le destinataire !\n desti = param[0] # on cherche par qui il faut passer\n phrase = str(param[0]) +\"<>\"+ str(param[1]) +\"<>\"+ str(param[2])\n print('Le hub transmet : \"' + param[2] + '\" de ' + \"l'ordi \" + param[1] + \" à destination de l'ordi \" + param[0])\n f=open(str(desti)+\".txt\",\"a\")\n f.write(phrase + \"\\n\")\n f.close()\n return True\n \ndef traitement():\n f=open(\"hub.txt\", \"r\")\n tache = f.readline()\n contenu = f.read()\n f.close()\n f=open(\"hub.txt\", \"w\")\n f.write(contenu)\n f.close()\n param = tache.split(\"<>\")\n if param!=[\"\"]: envoi(param)\n\t\ndef main():\n\ttraitement()\n\ttime.sleep(1)\n\tmain()\n\nos.system('cls')\nfichier=open(\"hub.txt\",\"w\") #création\nfichier.close()\nmain()\n\n\"\"\"fin hub\"\"\"\n","sub_path":"etoile_hub.py","file_name":"etoile_hub.py","file_ext":"py","file_size_in_byte":879,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"405725228","text":"\n# The Natural Language Toolkit,\n# or NLTK for short, is a Python library written\n# for working and modeling text.\n\n# It provides a high-level api to flexibly \n# implement a variety of cleaning methods.\n\n# load data\nfilename = 'metamorphosis_clean.txt'\nfile = open(filename, 'rt')\ntext = file.read()\nfile.close()\n# split into words\nfrom nltk.tokenize import word_tokenize, sent_tokenize\n# next line does the same as split() .\ntokens = word_tokenize(text)\n\n# remove all tokens that are not alphabetic\nwords = [word for word in tokens if word.isalpha()]\n\n# filter out stop words\n# stop words are i me my meself we ......\nfrom nltk.corpus import stopwords\nstop_words = set(stopwords.words('english'))\nwords = [w for w in words if not w in stop_words]\n\n# Stemming is a process where words are reduced \n# to a root by removing inflection through dropping\n# unnecessary characters, usually a suffix.\n# There are several stemming models, including \n# Porter and Snowball. But there is a danger of \n# “over-stemming” were words like “universe” and \n# “university” are reduced to the same root of\n# “univers”.\n# stemming of words\nfrom nltk.stem.porter import PorterStemmer\nporter = PorterStemmer()\nstemmed = [porter.stem(word) for word in tokens]\n\n# After stemming, ['morning', 'troubled'] converted\n# to ['morn','troubl'] by PorterStemmer.\n\n# this split by sentences.\nsentences = sent_tokenize(text)\n\n\nprint(stemmed[:100])\n","sub_path":"text_cleaning2.py","file_name":"text_cleaning2.py","file_ext":"py","file_size_in_byte":1440,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"303884349","text":"import re\nfrom thefuck.utils import for_app\n\n\n@for_app(\"composer\")\ndef match(command):\n # determine error type\n # matching \"did you mean this\" is not enough as composer also gives spelling suggestions for mistakes other than mispelled commands\n is_undefined_command_error = \"CommandNotFoundException\" in command.output\n suggestions_present = \"Did you mean\" in command.output\n return is_undefined_command_error and suggestions_present\n\n\ndef get_new_command(command):\n # since the command class already tells us the original argument, we need not resort to regex\n broken_cmd = command.script_parts[1]\n one_suggestion_only = \"Did you mean this?\" in command.output\n if one_suggestion_only:\n new_cmd = (\n re.search(r\"Did you mean this\\?[^\\n]*\\n\\s*([^\\n]*)\", command.output)\n .group(1)\n .strip()\n )\n return command.script.replace(broken_cmd, new_cmd)\n # else there are multiple suggestions\n # trim output text to make it more digestable by regex\n trim_start_index = command.output.find(\"Did you mean\")\n short_output = command.output[trim_start_index:]\n stripped_lines = [line.strip() for line in short_output.split(\"\\n\")]\n # each of the suggested commands can be found from index 1 to the first occurrence of a blank string\n end_index = stripped_lines.index(\"\")\n suggested_commands = stripped_lines[1:end_index]\n return [\n command.script.replace(broken_cmd, cmd.strip()) for cmd in suggested_commands\n ]\n","sub_path":"thefuck/rules/composer_not_command.py","file_name":"composer_not_command.py","file_ext":"py","file_size_in_byte":1516,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"82006122","text":"# Copyright Swiss Data Science Center (SDSC). A partnership between\n# École Polytechnique Fédérale de Lausanne (EPFL) and\n# Eidgenössische Technische Hochschule Zürich (ETHZ).\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"Classes for integration with Calamus.\"\"\"\n\nimport copy\nimport inspect\n\nimport marshmallow\nfrom calamus import fields\nfrom calamus.schema import JsonLDSchema as CalamusJsonLDSchema\nfrom calamus.utils import normalize_type, normalize_value\nfrom marshmallow.base import SchemaABC\n\nfrom renku.core import errors\nfrom renku.domain_model.project_context import project_context\n\nprov = fields.Namespace(\"http://www.w3.org/ns/prov#\")\nrdfs = fields.Namespace(\"http://www.w3.org/2000/01/rdf-schema#\")\nrenku = fields.Namespace(\"https://swissdatasciencecenter.github.io/renku-ontology#\")\nschema = fields.Namespace(\"http://schema.org/\")\noa = fields.Namespace(\"http://www.w3.org/ns/oa#\")\ndcterms = fields.Namespace(\"http://purl.org/dc/terms/\")\n\n\nclass JsonLDSchema(CalamusJsonLDSchema):\n \"\"\"Base schema class for Renku.\"\"\"\n\n def __init__(self, *args, commit=None, **kwargs):\n \"\"\"Create an instance.\"\"\"\n self._commit = commit\n super().__init__(*args, **kwargs)\n\n def _deserialize(self, *args, **kwargs):\n data = super()._deserialize(*args, **kwargs)\n const_args = inspect.signature(self.opts.model)\n parameters = const_args.parameters.values()\n\n if any(p.name == \"commit\" for p in parameters):\n if self._commit:\n self._add_field_to_data(data, \"commit\", self._commit)\n elif (\n project_context.has_context()\n and \"_label\" in data\n and data[\"_label\"]\n and \"@UNCOMMITTED\" not in data[\"_label\"]\n and \"@\" in data[\"_label\"]\n ):\n try:\n self._add_field_to_data(\n data,\n \"commit\",\n project_context.repository.get_commit(data[\"_label\"].rsplit(\"@\", maxsplit=1)[-1]),\n )\n except errors.GitCommitNotFoundError:\n # NOTE: This means the commit does not exist in the local repository. Could be an external file?\n pass\n\n return data\n\n def _add_field_to_data(self, data, name, value):\n if value:\n if name in data:\n raise ValueError(f\"Field {name} is already in data {data}\")\n data[name] = value\n\n\nclass Uri(fields._JsonLDField, marshmallow.fields.String, marshmallow.fields.Dict):\n \"\"\"A Dict/String field.\"\"\"\n\n def __init__(self, *args, **kwargs):\n \"\"\"Create an instance.\"\"\"\n super().__init__(*args, **kwargs)\n\n def _serialize(self, value, attr, obj, **kwargs):\n if isinstance(value, str):\n value = super(fields._JsonLDField, self)._serialize(value, attr, obj, **kwargs)\n if self.parent.opts.add_value_types:\n value = {\"@value\": value, \"@type\": \"http://www.w3.org/2001/XMLSchema#string\"}\n elif isinstance(value, dict):\n value = super(marshmallow.fields.String, self)._serialize(value, attr, obj, **kwargs)\n\n return value\n\n def _deserialize(self, value, attr, data, **kwargs):\n value = normalize_value(value)\n if not value:\n return None\n elif isinstance(value, str):\n return value\n elif isinstance(value, dict):\n return super(marshmallow.fields.String, self)._deserialize(value, attr, data, **kwargs)\n else:\n raise ValueError(f\"Invalid type for field {self.name}: {type(value)}\")\n\n\nclass StringList(fields.String):\n \"\"\"A String field that might be a list when deserializing.\"\"\"\n\n def __init__(self, *args, return_max_value=True, **kwargs):\n \"\"\"Create an instance.\"\"\"\n super().__init__(*args, **kwargs)\n self.return_max_value = return_max_value\n\n def _deserialize(self, value, attr, data, **kwargs):\n value = normalize_value(value)\n\n if isinstance(value, (list, tuple, set)):\n value = sorted(value, reverse=self.return_max_value)\n value = value[0] if len(value) > 0 else None\n\n return super()._deserialize(value, attr, data, **kwargs)\n\n\nclass DateTimeList(fields.DateTime):\n \"\"\"A DateTime field that might be a list when deserializing.\"\"\"\n\n def __init__(self, *args, **kwargs):\n \"\"\"Create an instance.\"\"\"\n super().__init__(*args, **kwargs)\n\n def _deserialize(self, value, attr, data, **kwargs):\n value = normalize_value(value)\n\n if isinstance(value, (list, tuple, set)):\n value = sorted(value)\n value = value[0] if len(value) > 0 else None\n\n return super()._deserialize(value, attr, data, **kwargs)\n\n\nclass Nested(fields.Nested):\n \"\"\"Nested field that passes along commit info.\"\"\"\n\n def __init__(self, *args, **kwargs):\n \"\"\"Init method.\"\"\"\n super().__init__(*args, **kwargs)\n\n @property\n def schema(self):\n \"\"\"The nested ``calamus.Schema`` object.\n\n This method was copied from marshmallow and modified to support\n multiple different nested schemes.\n \"\"\"\n if not self._schema:\n # Inherit context from parent.\n context = getattr(self.parent, \"context\", {})\n self._schema = {\"from\": {}, \"to\": {}}\n for nest in self.nested:\n if isinstance(nest, SchemaABC):\n rdf_type = str(normalize_type(nest.opts.rdf_type))\n model = nest.opts.model\n if not rdf_type or not model:\n raise ValueError(\"Both rdf_type and model need to be set on the \" \"schema for nested to work\")\n _schema = copy.copy(nest)\n _schema.context.update(context)\n # Respect only and exclude passed from parent and\n # re-initialize fields\n set_class = _schema.set_class\n if self.only is not None:\n if self._schema.only is not None:\n original = _schema.only\n else: # only=None -> all fields\n original = _schema.fields.keys()\n _schema.only = set_class(self.only).intersection(original)\n if self.exclude:\n original = _schema.exclude\n _schema.exclude = set_class(self.exclude).union(original)\n _schema._init_fields()\n _schema._visited = self.root._visited\n self._schema[\"from\"][rdf_type] = _schema\n self._schema[\"to\"][model] = _schema\n else:\n if isinstance(nest, type) and issubclass(nest, SchemaABC):\n schema_class = nest\n elif not isinstance(nest, (str, bytes)):\n raise ValueError(\"Nested fields must be passed a \" \"Schema, not {}.\".format(nest.__class__))\n elif nest == \"self\":\n ret = self\n while not isinstance(ret, SchemaABC):\n ret = ret.parent\n schema_class = ret.__class__\n else:\n schema_class = marshmallow.class_registry.get_class(nest)\n\n rdf_type = str(normalize_type(schema_class.opts.rdf_type))\n model = schema_class.opts.model\n if not rdf_type or not model:\n raise ValueError(\"Both rdf_type and model need to be set on the \" \"schema for nested to work\")\n\n kwargs = {}\n\n self._schema[\"from\"][rdf_type] = schema_class(\n many=False,\n only=self.only,\n exclude=self.exclude,\n context=context,\n load_only=self._nested_normalized_option(\"load_only\"),\n dump_only=self._nested_normalized_option(\"dump_only\"),\n lazy=self.root.lazy,\n flattened=self.root.flattened,\n _visited=self.root._visited,\n **kwargs,\n )\n self._schema[\"to\"][model] = self._schema[\"from\"][rdf_type]\n return self._schema\n","sub_path":"renku/command/schema/calamus.py","file_name":"calamus.py","file_ext":"py","file_size_in_byte":8986,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"177921453","text":"import time\nimport torch\nimport torch.nn as nn\nimport torch.optim as optim\nfrom torch.utils.data import ConcatDataset\nfrom torch.utils.tensorboard import SummaryWriter\nfrom torch.utils.data import TensorDataset, DataLoader\nfrom torch.utils.data import random_split\nimport argparse\nimport os\n\nfrom utils import generate_from_labeled_hdfs_file, generate_from_labeled_openstack_file, create_cross_val_loader\nfrom model import Model\n\n\n# Device configuration\ndevice = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n\n\ndef generate(name):\n num_sessions = 0\n inputs = []\n outputs = []\n with open('data/' + name, 'r') as f:\n for line in f.readlines():\n num_sessions += 1\n line = tuple(map(lambda n: n - 1, map(int, line.strip().split())))\n for i in range(len(line) - window_size):\n inputs.append(line[i:i + window_size])\n outputs.append(line[i + window_size])\n print('Number of sessions({}): {}'.format(name, num_sessions))\n print('Number of seqs({}): {}'.format(name, len(inputs)))\n dataset = TensorDataset(torch.tensor(inputs, dtype=torch.float), torch.tensor(outputs))\n return dataset\n\n\ndef create_train_dataloader(normal_train_path, window_size, num_classes):\n normal_train_dataset = generate_from_labeled_hdfs_file(normal_train_path, window_size, num_classes)\n normal_train_dataloader = DataLoader(normal_train_dataset, batch_size=batch_size, shuffle=True, pin_memory=True)\n\n return normal_train_dataloader\n\n\ndef train(model, optimizer, scheduler, train_dataloader, window_size):\n for epoch in range(num_epochs): # Loop over the train and validation datasets multiple times\n train_loss = 0\n valid_loss = 0\n\n # Training\n for step, (seq, label) in enumerate(train_dataloader):\n # Forward pass\n seq = seq.clone().detach().view(-1, window_size).to(device)\n x_onehot = torch.nn.functional.one_hot(seq.long(), input_size).float()\n output = model(x_onehot)\n loss = criterion(output, label.to(device))\n\n # Backward and optimize\n optimizer.zero_grad()\n loss.backward()\n train_loss += loss.item()\n optimizer.step()\n\n print('Epoch [{}/{}], train_loss: {:.4f}'.format(epoch + 1, num_epochs, train_loss / train_total_step))\n\n scheduler.step(train_loss / train_total_step)\n\n if epoch % snapshot_period == 0:\n if not os.path.isdir(model_dir):\n os.makedirs(model_dir)\n torch.save(model.state_dict(), model_dir + '/' + log + \"_iteration=\" + str(epoch) + '.pt')\n\n return train_loss\n\n\ndef validate(model, normal_val_loader, abnormal_val_loader, window_size):\n with torch.no_grad():\n TP = 0\n FP = 0\n\n # Normal validation dataset\n for line in normal_val_loader:\n for i in range(len(line) - window_size):\n seq = line[i:i + window_size]\n label = line[i + window_size]\n seq = torch.tensor(seq, dtype=torch.float).view(-1, window_size).to(device)\n x_onehot = torch.nn.functional.one_hot(seq.long(), input_size).float()\n label = torch.tensor(label).view(-1).to(device)\n output = model(x_onehot)\n predicted = torch.argsort(output, 1)[0][-num_candidates:]\n if label not in predicted:\n FP += 1\n break\n\n # Abnormal validation dataset\n for line in abnormal_val_loader:\n for i in range(len(line) - window_size):\n seq = line[i:i + window_size]\n label = line[i + window_size]\n seq = torch.tensor(seq, dtype=torch.float).view(-1, window_size).to(device)\n x_onehot = torch.nn.functional.one_hot(seq.long(), input_size).float()\n label = torch.tensor(label).view(-1).to(device)\n output = model(x_onehot)\n predicted = torch.argsort(output, 1)[0][-num_candidates:]\n if label not in predicted:\n TP += 1\n break\n\n FN = len(abnormal_val_loader) - TP\n P = 100 * TP / (TP + FP)\n R = 100 * TP / (TP + FN)\n F1 = 2 * P * R / (P + R)\n\n print(\n 'false positive (FP): {}, false negative (FN): {}, Precision: {:.3f}%, Recall: {:.3f}%, F1-measure: {:.3f}%'.format(\n FP, FN, P, R, F1))\n\n return P, R, F1\n\n\nif __name__ == '__main__':\n\n # Hyperparameters\n num_classes = 31\n num_epochs = 1\n batch_size = 8192\n model_dir = 'model'\n log = 'cross_validation'\n parser = argparse.ArgumentParser()\n parser.add_argument('-num_layers', default=2, type=int)\n parser.add_argument('-hidden_size', default=64, type=int)\n parser.add_argument('-snapshot_period', default=10, type=int)\n parser.add_argument('-num_candidates', default=9, type=int)\n args = parser.parse_args()\n num_layers = args.num_layers\n hidden_size = args.hidden_size\n num_candidates = args.num_candidates\n snapshot_period = args.snapshot_period\n\n window_size_list = [5, 6]#, 7, 8, 9, 10, 11]\n\n model_dict = dict()\n\n for window_size in window_size_list:\n\n normal_train_dataloader = create_train_dataloader('/home/toni/Downloads/balanced/normal_train.txt', window_size, num_classes)\n\n normal_val_loader = create_cross_val_loader('/home/toni/Downloads/balanced/normal_test.txt', window_size, num_classes)\n abnormal_val_loader = create_cross_val_loader('/home/toni/Downloads/balanced/anomaly.txt', window_size, num_classes)\n\n writer = SummaryWriter(log_dir='log/' + log)\n\n input_size = num_classes + 1\n\n model = Model(num_classes+1, hidden_size, num_layers, device).to(device)\n\n # Loss and optimizer\n criterion = nn.CrossEntropyLoss()\n optimizer = torch.optim.SGD(model.parameters(), lr=0.01, momentum=0.9, nesterov=True)\n scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(optimizer, mode='min', factor=0.1, patience=10, verbose=True, threshold=0.001, threshold_mode='abs', cooldown=5, eps=0)\n\n # Train the model\n start_time = time.time()\n train_total_step = len(normal_train_dataloader)\n\n train_loss = train(model, optimizer, scheduler, normal_train_dataloader, window_size)\n\n P, R, F1 = validate(model, normal_val_loader, abnormal_val_loader, window_size)\n\n model_dict[window_size] = (model, P, R, F1, train_loss)\n writer.add_scalar('Precision', P, window_size)\n writer.add_scalar('Recall', R, window_size)\n writer.add_scalar('F1 score', F1, window_size)\n writer.add_scalar('train_loss', train_loss, window_size)\n\n elapsed_time = time.time() - start_time\n print('elapsed_time: {:.3f}s'.format(elapsed_time))\n writer.close()\n print('Finished Training')\n","sub_path":"LogKeyModel_train.py","file_name":"LogKeyModel_train.py","file_ext":"py","file_size_in_byte":6928,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"404105379","text":"#!/usr/bin/env python2\n\nimport rospy\nimport smach\nfrom std_msgs.msg import Bool, Empty, String\n\n\nclass WaitForMessage(smach.State):\n def __init__(self, timeout=10):\n super(WaitForMessage, self).__init__(\n outcomes=[\"timeout\", \"message_received\", \"preempted\"], output_keys=[\"msg\"]\n ) # all outcomes and user data are defined\n # initialise the subscriber\n self.sub = rospy.Subscriber(\"/message\", String, self.message_received_cb)\n # where we will put the content of the message\n self.msg = None\n # the timeout float\n self.timeout = timeout\n\n # callback of the subscriber\n def message_received_cb(self, msg):\n # we copy the content of the message inside our place holder\n self.msg = msg.data\n\n def execute(self, ud):\n # compute the timeout with rostime\n ros_timeout_ = rospy.Time().now() + rospy.Duration(self.timeout)\n while not self.msg: # an empty string is considered False, non-empty True\n # check if we are over the timeout rostime\n if rospy.Time.now() > ros_timeout_:\n rospy.logwarn(\"Timed Out\") # warning message\n return \"timeout\" # timeout\n # check if we are preempted or if ctrl+c\n elif self.preempt_requested() or rospy.is_shutdown():\n rospy.logwarn(\"Preempted ! (or shutdown)\") # waning message\n return \"preempted\" # preempted\n rospy.sleep(0.1) # free the thread for callback checking\n rospy.loginfo(\"Message received : '%s'\", self.msg) # information message\n ud.msg = self.msg # copy the message content inside the userdata\n return \"message_received\" # message received\n","sub_path":"src/smach_tutorial/solution/sol_AdvancedStates.py","file_name":"sol_AdvancedStates.py","file_ext":"py","file_size_in_byte":1744,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"301948000","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Dec 7 21:17:38 2016\n@author: nishant\n........................................................................................................................\nTask 2 \nYou are given a list of words a baby can speak. There is a recorder that can convert the baby's speech to text, but it cannot recognize pauses, so it does not put spaces in between the words.\nWrite a program that can convert the baby's recorded speech into words spoken by the baby (existing in the baby dictionary). Every part of the text is actually a word, so if the program outputs one word and the rest as garbled text, it won't be correct. Every test case will have only one correct answer (assume your test case wont contain any phrase that can have more than one correct answer)\n Input Variables\n1) baby_words = array containing the words the baby can speak.\n2) garbled_text = text given by the recorder that converted speech to text \nSample Input\ngarbled_text = gagagoogoo\nbaby_words = [gag goo gaga] \n Sample output : gaga goo goo\n The following outputs are wrong:\n1) gag\n2) gag goo\n3) gag a goo goo\n The output must be formatted in the following manner:\nEach word must be on the same line and separated by a single space. (The example above has the correct format)\n\"\"\"\nprint(\"Sample Input\")\ngarbled_text = input(\"Enter the garbled text : \")\nbaby_words=input(\"Enter baby words with spaces : \").strip().split(' ')\nbaby_words.sort(key=len)\nresult=[]\nli=[]\nwhile(len(garbled_text)!=0):\n for i in range(0,len(baby_words)):\n if(baby_words[i]==garbled_text[:len(baby_words[i])]):\n li=baby_words[i]\n result.append(li)\n garbled_text=garbled_text[len(li):]\nprint(\"Sample Output :\") \nfor j in result:\n print(j,sep=' ',end=' ')\n","sub_path":"Task2.py","file_name":"Task2.py","file_ext":"py","file_size_in_byte":1781,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"257006401","text":"from random import randint, sample\nimport random\nimport itertools\nfrom time import time\nimport math\nimport sys\nimport numpy as np #pip install numpy \nclass SSP():\n# Declare the class\n def __init__(self, S=[], t=0):\n# Initiation\n self.S = S\n# s is an array\n self.t = t\n# t is 0\n self.n = len(S)\n# n is the length of s\n #\n self.decision = False\n#the decision is false\n self.total = 0\n#total is 0\n self.selected = []\n# the selected is s\n\n def __repr__(self):\n# return a printable representation of the object\n return \"SSP instance: S=\"+str(self.S)+\"\\tt=\"+str(self.t)\n \n def random_instance(self, n, bitlength=10):\n max_n_bit_number = 2**bitlength-1\n# maximum bit number is the bitlength square - 1\n self.S = sorted( [ randint(0,max_n_bit_number) for i in range(n) ] , reverse=True)\n # s is sorted in a forloop in range of n and it is beiween the number of 0 to the maximum bit number\n self.t = randint(0,n*max_n_bit_number)\n# t is either 0 or n x the maximum bit number\n self.n = len( self.S )\n#n is the lenght of s\n\n def random_yes_instance(self, n, bitlength=10):\n max_n_bit_number = 2**bitlength-1\n# maximum bit number is the bitlength square - 1\n self.S = sorted( [ randint(0,max_n_bit_number) for i in range(n) ] , reverse=True)\n # s is sorted in a forloop in range of n and it is beiween the number of 0 to the maximum bit number\n self.t = sum( sample(self.S, randint(0,n)) )\n# t is the sum of s or the random instances 0 or n\n self.n = len( self.S )\n#n is the lenght of s\n\n ###\n\n def try_at_random(self):\n candidate = []\n# candidate is an array\n total = 0\n# total is 0\n while total != self.t:\n# while total doesn't equal to t\n candidate = sample(self.S, randint(0,self.n))\n# candidate equals to the \n total = sum(candidate)\n# total equals to the sum x candidate\n print( \"Trying: \", candidate, \", sum:\", total )\n# print statement\n\n def Exhaustive_Search(self):\n #declare the function Exhaustive Search\n subset = self.getSubset()\n total = 0\n #calls the function getSubsets to generate all the subsets for the initial set and return the subsets\n for set in subsets:\n #loop through every set of the subsets\n total = sum(set)\n #set the total to equal the sum of the current set\n print(\"Trying: \", set, \", sum:\", total )\n #print the set that is currently being tested\n if (total == self.t):\n #if the total does not match the value of t which is the target value then solution has been found\n print(\"Solution: \" +str(set))\n #print the solution out\n return Trues\n \ninstance = SSP() #instance equals the SSP class\ninstance.random_yes_instance(4)\ninstance.Exhaustive_Search()\nprint( instance )\n\n","sub_path":"Exhuastive_Search.py","file_name":"Exhuastive_Search.py","file_ext":"py","file_size_in_byte":2953,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"315827588","text":"#CalBmi.py\n'''height,weight=eval(input(\"请输入身高/(米)和体重/(公斤):[逗号隔开]\"))\nbmi=weight/pow(height,2)\nprint(\"BMI数值为:{:.2f}\".format(bmi))\nwho=\"\"\nif bmi<18.5:\n who=\"偏瘦\"\nelif bmi<25:\n who=\"正常\"\nelif bmi<30:\n who=\"偏胖\"\nelse:\n who=\"肥胖\"\nprint(\"BMI指标为:国际'{0}',\".format(who),end=\"\")\n\nif bmi<18.5:\n who=\"偏瘦\"\nelif bmi<24:\n who=\"正常\"\nelif bmi<28:\n who=\"偏胖\"\nelse:\n who=\"肥胖\"\nprint(\"国内'{0}'\".format(who))'''\n\n\nheight,weight=eval(input(\"请输入身高/(米)和体重/(公斤):[逗号隔开]\"))\nbmi=weight/pow(height,2)\nprint(\"BMI数值为:{:.2f}\".format(bmi))\nwho=\"\"\nif bmi<18.5:\n who,nat=\"偏瘦\",\"偏瘦\"\nelif bmi<24:\n who,nat=\"正常\",\"正��\"\nelif bmi<25:\n who,nat=\"正常\",\"偏胖\"\nelif bmi<28:\n who,nat=\"偏胖\",\"偏胖\"\nelif bmi<30:\n who,nat=\"偏胖\",\"肥胖\"\nelse:\n who,nat=\"肥胖\",\"肥胖\"\nprint(\"BMI指标为:国际'{}',国内'{}'\".format(who,nat))\n\n\n\n","sub_path":"CalBmi.py","file_name":"CalBmi.py","file_ext":"py","file_size_in_byte":985,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"580889891","text":"\"\"\"\n Defines **async** handlers for socket.io server\n\n\n SEE https://pypi.python.org/pypi/python-socketio\n SEE http://python-socketio.readthedocs.io/en/latest/\n\n\"\"\"\n# pylint: disable=C0111\n\n\nimport logging\nimport os\n\nimport socketio\n\nimport config\nimport interactive_services_manager\nfrom s3wrapper.s3_client import S3Client\nfrom simcore_sdk.config.s3 import Config as s3_config\n\n_LOGGER = logging.getLogger(__file__)\n\nSIO = socketio.AsyncServer(async_mode='aiohttp', logging=_LOGGER)\n\nCONFIG = config.CONFIG[os.environ.get('SIMCORE_WEB_CONFIG', 'default')]\n\n\n@SIO.on('connect')\ndef connect(sid, environ):\n # pylint: disable=W0613\n # environ = WSGI evnironment dictionary\n _LOGGER.debug(\"client %s connects\", sid)\n interactive_services_manager.session_connect(sid)\n return True\n\n\n@SIO.on('getInteractiveServices')\nasync def get_interactive_services_handler(sid, data):\n # pylint: disable=C0103\n # pylint: disable=W0613\n _LOGGER.debug(\"client %s gets interactive services\", sid)\n result = interactive_services_manager.retrieve_list_of_services()\n await SIO.emit('getInteractiveServices', data=result, room=sid)\n\n\n@SIO.on('startDynamic')\nasync def start_dynamic_service(sid, data):\n service_name = data['serviceName']\n node_id = data['nodeId']\n _LOGGER.debug(\"client %s requests start %s\", sid, service_name)\n result = interactive_services_manager.start_service(sid, service_name, node_id)\n # TODO: Connection failure raises exception that is not treated, which stops the webserver\n # Add mechanism to handle these situations (retry, abandon...)\n try:\n await SIO.emit('startDynamic', data=result, room=sid)\n except IOError as err:\n _LOGGER.exception(err)\n\n\n@SIO.on('stopDynamic')\ndef stop_dynamic_service(sid, data):\n node_id = data['nodeId']\n _LOGGER.debug(\"client %s requests stop %s\", sid, node_id)\n interactive_services_manager.stop_service(sid, node_id)\n\n\n@SIO.on('startModeler')\nasync def start_modeler_handler(sid, data):\n _LOGGER.debug(\"client %s requests start modeler %s\", sid, data)\n result = interactive_services_manager.start_service(sid, 'modeler', data)\n # TODO: Connection failure raises exception that is not treated, which stops the webserver\n # Add mechanism to handle these situations (retry, abandon...)\n try:\n await SIO.emit('startModeler', data=result, room=sid)\n except IOError as err:\n _LOGGER.exception(err)\n\n\n@SIO.on('stopModeler')\nasync def stop_modeler_handler(sid, data):\n _LOGGER.debug(\"client %s requests stop modeler %s\", sid, data)\n result = interactive_services_manager.stop_service(sid, data)\n await SIO.emit('stopModeler', data=result, room=sid)\n\n\n@SIO.on('startJupyter')\nasync def start_jupyter_handler(sid, data):\n _LOGGER.debug(\"client %s requests start jupyter %s\", sid, data)\n result = interactive_services_manager.start_service(\n sid, 'jupyter-base-notebook', data)\n _LOGGER.debug(\"director %s starts jupyter %s: %s\", sid, data, result)\n await SIO.emit('startJupyter', data=result, room=sid)\n\n\n@SIO.on('stopJupyter')\nasync def stop_jupyter_handler(sid, data):\n _LOGGER.debug(\"client %s requests stop jupyter %s\", sid, data)\n result = interactive_services_manager.stop_service(sid, data)\n await SIO.emit('stopJupyter', data=result, room=sid)\n\n\n@SIO.on('presignedUrl')\nasync def retrieve_url_for_file(sid, data):\n _LOGGER.debug(\"client %s requests S3 url for %s\", sid, data)\n _config = s3_config()\n _LOGGER.debug(\"S3 endpoint %s\", _config.endpoint)\n\n\n s3_client = S3Client(endpoint=_config.endpoint,\n access_key=_config.access_key, secret_key=_config.secret_key)\n url = s3_client.create_presigned_put_url(_config.bucket_name, data[\"fileName\"])\n #result = minioClient.presigned_put_object(data[\"bucketName\"], data[\"fileName\"])\n # Response error is still possible since internally presigned does get\n # bucket location.\n data_out = {}\n data_out[\"url\"] = url\n await SIO.emit('presignedUrl', data=data_out, room=sid)\n\n\n@SIO.on('listObjects')\nasync def list_S3_objects(sid, data):\n _LOGGER.debug(\"client %s requests objects in storage. Extra argument %s\", sid, data)\n _config = s3_config()\n\n s3_client = S3Client(endpoint=_config.endpoint,\n access_key=_config.access_key, secret_key=_config.secret_key)\n\n objects = s3_client.list_objects_v2(_config.bucket_name)\n data_out = []\n for obj in objects:\n obj_info = {}\n obj_info['path'] = obj.bucket_name + '/' + obj.object_name\n # @maiz: this does not work, please review\n #obj_info['lastModified'] = obj.last_modified.isoformat()\n obj_info['size'] = obj.size\n data_out.append(obj_info)\n await SIO.emit('listObjects', data=data_out, room=sid)\n\n\n@SIO.on('disconnect')\ndef disconnect(sid):\n _LOGGER.debug(\"client %s disconnected\", sid)\n interactive_services_manager.session_disconnected(sid)\n","sub_path":"services/web/server/src/async_sio.py","file_name":"async_sio.py","file_ext":"py","file_size_in_byte":4939,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"1064067","text":"# -*- coding: utf-8 -*-\n# @Time : 18-5-20 下午8:07\n# @Author : zhoujun\nimport time\nimport json\nfrom collections import OrderedDict\nfrom pathlib import Path\n\n\ndef setup_logger(log_file_path: str = None):\n import logging\n logger = logging.getLogger('crnn.gluon')\n logger.setLevel(logging.DEBUG)\n formatter = logging.Formatter('%(asctime)s %(name)s %(levelname)s: %(message)s')\n ch = logging.StreamHandler()\n ch.setFormatter(formatter)\n logger.addHandler(ch)\n if log_file_path is not None:\n file_handle = logging.FileHandler(log_file_path)\n file_handle.setFormatter(formatter)\n logger.addHandler(file_handle)\n logger.info('logger init finished')\n return logger\n\n\n# --exeTime\ndef exe_time(func):\n def newFunc(*args, **args2):\n t0 = time.time()\n back = func(*args, **args2)\n print(\"{} cost {:.3f}s\".format(func.__name__, time.time() - t0))\n return back\n\n return newFunc\n\n\ndef save_json(data, json_path):\n with open(json_path, mode='w', encoding='utf8') as f:\n json.dump(data, f, indent=4)\n\n\ndef load_json(json_path):\n with open(json_path, mode='r', encoding='utf8') as f:\n data = json.load(f)\n return data\n\n\ndef read_json(fname):\n if isinstance(fname, str):\n fname = Path(fname)\n with fname.open('rt') as handle:\n return json.load(handle)\n\n\ndef write_json(content, fname):\n if isinstance(fname, str):\n fname = Path(fname)\n with fname.open('wt') as handle:\n json.dump(content, handle, indent=4, sort_keys=False)\n\n\ndef get_ctx(gpus):\n import mxnet as mx\n from mxnet import nd\n \"\"\"If GPU is available, return mx.gpu(0); else return mx.cpu()\"\"\"\n try:\n ctx = []\n for gpu in gpus:\n ctx_i = mx.gpu(gpu)\n _ = nd.array([0], ctx=ctx_i)\n ctx.append(ctx_i)\n except:\n ctx = [mx.cpu()]\n return ctx\n\n\ndef punctuation_mend(string):\n # 输入字符串或者txt文件路径\n import unicodedata\n import pathlib\n\n table = {ord(f): ord(t) for f, t in zip(\n u',。!?【】()%#@&1234567890“”‘’',\n u',.!?[]()%#@&1234567890\"\"\\'\\'')} # 其他自定义需要修改的符号可以加到这里\n if pathlib.Path(string).is_file():\n with open(string, 'r', encoding='utf-8') as f:\n res = unicodedata.normalize('NFKC', f.read())\n res = res.translate(table)\n with open(string, 'w', encoding='utf-8') as f:\n f.write(res)\n else:\n res = unicodedata.normalize('NFKC', string)\n res = res.translate(table)\n return res\n\n\nif __name__ == '__main__':\n print(punctuation_mend('1'))","sub_path":"utils/util.py","file_name":"util.py","file_ext":"py","file_size_in_byte":2707,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"585911048","text":"'''\n Loads a saved model and uses it in 2 different modes:\n - Displays NxN random images\n - Displays N images with their reconstruction from the dataset.\n'''\n\nimport matplotlib.pyplot as plt\nfrom tqdm import trange, tqdm\nimport numpy as np\nimport argparse\n\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torch.optim as optim\nimport torchvision\nimport torchvision.transforms as transforms\nprint(torch.__version__)\nprint(torchvision.__version__)\n\nfrom vaegan import VAEGAN\n\n# Parse arguments\nparser = argparse.ArgumentParser(description='VAEGAN playing MNIST')\nparser.add_argument('--weights', type=str, default='weights/base.torch', help='Weight file to load.')\nparser.add_argument('--mode', type=str, default='generator', choices=['generator', 'reconstructor'], help='Mode for the script')\nparser.add_argument('--N', type=int, default=4, help='Images to display, depends on mode.')\nargs = parser.parse_args()\n\n# Find the available device\ndevice = torch.device(\"cuda:0\" if torch.cuda.is_available() else \"cpu\")\n\n# Load the model\nstate = torch.load(args.weights, map_location='cpu')\ninput_shape = (1, 28, 28)\nvaegan = VAEGAN(input_shape, latent_size=state['latent_size'], convs=state['convs']).to(device)\nvaegan.set_device(device)\nvaegan.load_state_dict(state['state_dict'])\nvaegan.eval()\n\nif args.mode == 'generator':\n # Create an NxN vector of latent_size\n x = np.random.randn(args.N, args.N, state['latent_size']).astype(np.float32)\n # Decode using the VAE decoder\n generated = vaegan.decode(torch.from_numpy(np.reshape(x, (-1, state['latent_size']))))\n # Reshape to display [has shape (batch, 1, 28, 28)]\n generated = generated.permute(0, 2, 3, 1) # shape: (batch, 28, 28, 1)\n generated = np.reshape(generated.detach().numpy(), (args.N, args.N, 28, 28))\n # Create a window with NxN subplots\n fig, ax = plt.subplots(args.N, args.N)\n for i in range(args.N):\n for k in range(args.N):\n ax[i, k].imshow(generated[i, k], cmap='gray')\n plt.show()\nelif args.mode == 'reconstructor':\n # Get N images from MNIST\n transform = transforms.Compose([transforms.ToTensor()])\n testset = torchvision.datasets.MNIST(root='./mnist_data', train=False, download=True, transform=transform)\n testloader = torch.utils.data.DataLoader(testset, batch_size=args.N, shuffle=False, num_workers=2)\n dataiter = iter(testloader)\n images, labels = dataiter.next()\n # Get the reconstructed images\n mu, log_sigma = vaegan.encode(images)\n rebuild = vaegan.decode(mu)\n # Transform for numpy\n ground = np.reshape(images.detach().numpy(), (args.N, 28, 28))\n rebuild = np.reshape(rebuild.detach().numpy(), (args.N, 28, 28))\n # Plot\n fig, ax = plt.subplots(args.N, 2)\n for i in range(args.N):\n ax[i, 0].imshow(ground[i], cmap='gray')\n ax[i, 1].imshow(rebuild[i], cmap='gray')\n plt.show()\nelse:\n raise Exception('Unrecognized mode.')\n","sub_path":"vaegan/play_mnist.py","file_name":"play_mnist.py","file_ext":"py","file_size_in_byte":2946,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"385396665","text":"def incr(s,k):\n for i in range(len(s)):\n if ord(s[i])== 32: #To convert space to @\n s[i]=\"@\"\n elif ord(s[i]) == 46:\n s[i]=\"#\" #To convert \".\" to #\n elif ord(s[i]) in range(97,123):\n s[i]=ord(s[i])+k #To encrypt A to Z \n if s[i]<123:\n s[i]=chr(s[i])\n else:\n s[i]=chr(s[i]%123+97)\n elif ord(s[i]) in range(65,91):\n s[i]=ord(s[i])+k #To encrypt a to z\n if s[i]<91:\n s[i]=chr(s[i])\n else:\n s[i]=chr(s[i]%91+65)\n elif ord(s[i]) in range(48,58):\n s[i]=ord(s[i])+k #To encrypt 0 to 9\n if s[i]<58:\n s[i]=chr(s[i])\n else:\n s[i]=chr(s[i]%58+48)\n \nprint(\"Hello MIT CELL:\")\nk=int(input(\"Please enter k value to shift tranform message: \")) #Taking the shift value key for encrypting\ns=list(input(\"Now please type your message: \")) #Taking the message\nl=s.copy() #Copying list to another list \nincr(l,k) #Calling Function to encrypt the message\nprint(\"You send-\",\"\".join(l)) \nprint(\"\\nHello Instructor:\\nYou got a message: \",\"\".join(l))\nm=int(input(\"To read this in original please type k value provided by Cell: \")) #Taking the key from the reciever\nif k==m:\n print(\"Original Message: \",\"\".join(s)) #If key matches print the original message\nelse:\n print(\"Error in k value\")\n","sub_path":"Secure_Communication.py","file_name":"Secure_Communication.py","file_ext":"py","file_size_in_byte":1600,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"358575702","text":"\"\"\"\nAsteroids!\n\"\"\"\n \nimport pygame\nimport math\nimport copy\n\ndef rotatePoint(centerPoint,point,angle):\n \"\"\"Rotates a point around another centerPoint. Angle is in degrees.\n Rotation is counter-clockwise\"\"\"\n angle = math.radians(angle)\n temp_point = point[0]-centerPoint[0] , point[1]-centerPoint[1]\n temp_point = ( temp_point[0]*math.cos(angle)-temp_point[1]*math.sin(angle) , temp_point[0]*math.sin(angle)+temp_point[1]*math.cos(angle))\n temp_point = temp_point[0]+centerPoint[0] , temp_point[1]+centerPoint[1]\n return temp_point\n\n\n# Define some colors\nBLACK = (0, 0, 0)\nWHITE = (255, 255, 255)\nGREEN = (0, 255, 0)\nRED = (255, 0, 0)\n \npygame.init()\n \n# Set the width and height of the screen [width, height]\nsize = (600, 600)\nscreen = pygame.display.set_mode(size)\n \npygame.display.set_caption(\"My Game\")\n\nshoot_sound = pygame.mixer.Sound('pew.wav')\n \n# Loop until the user clicks the close button.\ndone = False\n \n# Used to manage how fast the screen updates\nclock = pygame.time.Clock()\nall_sprites_list = pygame.sprite.Group()\nbullets = []\npygame.display.set_caption(\"Astroids\")\n\nclass Ship(pygame.sprite.Sprite):\n def __init__(self):\n pygame.sprite.Sprite.__init__(self)\n self.image = pygame.Surface([26,26])\n self.pts = [[15, 0],[5,25], [15,15], [25,25]]\n pygame.draw.lines(self.image, WHITE, True, self.pts,2)\n self.rect = self.image.get_rect()\n self.rect.x = 300\n self.rect.y = 300\n self.angle = 0\n self.speed = 0\n\n def rotate(self, angle):\n for i, pt in enumerate(self.pts):\n self.pts[i] = rotatePoint([13,13], pt, angle)\n self.image.fill(BLACK)\n pygame.draw.lines(self.image, WHITE, True, self.pts,2)\n\n def get_direction(self):\n # we know pts[0] and pts[2] give us a vector we are \"pointing\" at:\n vector = [self.pts[0][0] - self.pts[2][0], self.pts[0][1] - self.pts[2][1]]\n # normalize:\n norm = math.sqrt(vector[0] ** 2 + vector[1] ** 2)\n vector = [vector[0] / norm, vector[1] / norm]\n return vector\n\nclass Bullet(pygame.sprite.Sprite):\n def __init__(self):\n pygame.sprite.Sprite.__init__(self)\n self.image = pygame.Surface([2,2])\n self.image.fill(WHITE)\n self.speed = 5\n self.life = 0\n \n\nship = Ship()\nall_sprites_list.add(ship)\n\n# -------- Main Program Loop -----------\nwhile not done:\n # --- Main event loop\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n done = True\n if event.type == pygame.KEYDOWN:\n if event.key == pygame.K_RIGHT:\n ship.angle = 1\n if event.key == pygame.K_LEFT:\n ship.angle = -1\n if event.key == pygame.K_UP:\n ship.speed = 10\n if event.key == pygame.K_SPACE:\n shoot_sound.play()\n bullet = Bullet()\n bullet.rect = bullet.image.get_rect()\n bullet.rect = copy.copy(ship.rect)\n bullet.rect.x += ship.pts[0][0]\n bullet.rect.y += ship.pts[0][1]\n bullet.direction = ship.get_direction()\n bullets.append(bullet)\n all_sprites_list.add(bullet)\n \n if event.type == pygame.KEYUP:\n if event.key == pygame.K_RIGHT or event.key == pygame.K_LEFT:\n ship.angle = 0\n \n \n # --- Game logic should go here\n\n if ship.angle != 0:\n ship.rotate(ship.angle)\n\n vect = ship.get_direction()\n ship.rect.x += (vect[0]*ship.speed)\n ship.rect.y += (vect[1]*ship.speed)\n\n for bullet in bullets:\n bullet.rect.x += (bullet.direction[0] * bullet.speed)\n bullet.rect.y += (bullet.direction[1] * bullet.speed) \n \n if bullet.rect.x < 0:\n bullet.rect.x = size[0]\n elif bullet.rect.x > size[0]:\n bullet.rect.x = 0\n if bullet.rect.y < 0:\n bullet.rect.y = size[1]\n elif bullet.rect.y > size[1]:\n bullet.rect.y = 0\n bullet.life += 1\n if bullet.life > 100:\n bullets.remove(bullet)\n all_sprites_list.remove(bullet)\n\n \n\n # wrap ship:\n if ship.rect.x < 0:\n ship.rect.x = size[0]\n if ship.rect.x > size[0]:\n ship.rect.x = 0\n if ship.rect.y < 0:\n ship.rect.y = size[1]\n if ship.rect.y > size[1]:\n ship.rect.y = 0\n\n # decay speed:\n if ship.speed > 0:\n ship.speed -= 0.1\n if ship.speed < 0:\n ship.speed = 0\n \n \n \n # --- Screen-clearing code goes here\n \n # Here, we clear the screen to white. Don't put other drawing commands\n # above this, or they will be erased with this command.\n \n # If you want a background image, replace this clear with blit'ing the\n # background image.\n screen.fill(BLACK)\n \n # --- Drawing code should go here\n all_sprites_list.draw(screen)\n \n # --- Go ahead and update the screen with what we've drawn.\n pygame.display.flip()\n \n # --- Limit to 60 frames per second\n clock.tick(60)\n \n# Close the window and quit.\npygame.quit()\n","sub_path":"asteroids.py","file_name":"asteroids.py","file_ext":"py","file_size_in_byte":5148,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"274205975","text":"# Script for Tkinter GUI chat client.\n\nfrom socket import AF_INET, socket, SOCK_STREAM\nfrom threading import Thread\nimport tkinter\nfrom random import choice\n\nNUM_MESSAGES = 2\nPOSSIBLE_COLORS = [\"#0000bf\",\"#0000ff\",\"#00bf00\",\"#00ff00\",\"#bf0000\",\n\t\"#ff0000\",\"#00bfff\",\"#bf00ff\",\"#bfff00\",\"#00ffbf\",\"#ff00bf\",\"#ffbf00\"]\n\ndef receive():\n\t# Handles receiving of messages\n\twhile True:\n\t\ttry:\n\t\t\tmsg = client_socket.recv(BUFSIZ).decode(\"utf8\")\n\t\t\tmsg_list.insert(tkinter.END, msg)\n\t\t\tglobal NUM_MESSAGES\n\t\t\tNUM_MESSAGES += 1\n\t\t\tmsg_list.see(NUM_MESSAGES)\n\t\texcept OSError:\t# Possibly client has left the chat\n\t\t\tbreak\n\n\ndef send(event=None):\t# event is passed by binders \n\t# Handles sending of messages\n\tmsg = my_msg.get()\n\tmy_msg.set(\"\")\t# clears input field\n\tclient_socket.send(bytes(msg, \"utf8\"))\n\tif msg == \"*quit\":\n\t\t#client_socket.close()\n\t\ttop.quit()\n\n\ndef on_closing(event=None):\n\t# This function is to be called when the window is closed\n\tmy_msg.set(\"*quit\")\n\tsend()\n\n# top level widget:\ntop = tkinter.Tk()\ntop.title(\"Mesajlaşma\")\n\n# frame for holding list of messages and string variable for message\nmessages_frame = tkinter.Frame(top)\nmy_msg = tkinter.StringVar()\t# for the messages to be sent\nmy_msg.set(\"\")\nscrollbar = tkinter.Scrollbar(messages_frame)\t# to navigate through messages.\t\n\n# create message list and pack everything\nLISTBOX_HEIGHT = 20\nLISTBOX_WEIGHT = 100\nmsg_list = tkinter.Listbox(messages_frame, height=LISTBOX_HEIGHT,\n\twidth=LISTBOX_WEIGHT, yscrollcommand=scrollbar.set)\nscrollbar.pack(side=tkinter.RIGHT, fill=tkinter.Y)\nmsg_list.pack(side=tkinter.LEFT, fill=tkinter.BOTH)\nmsg_list.pack()\nmessages_frame.pack()\t\n\n\n# create input field and bind it to the string variable above\nentry_field = tkinter.Entry(top, textvariable=my_msg, width=30)\nentry_field.bind(\"\", send)\t# send message when pressed enter\nentry_field.pack()\nsend_button = tkinter.Button(top, text=\"Send\", command=send,\t# send message\n\theight=2, width=30)\t # when clicked button \n \nsend_button.pack()\n\ntop.protocol(\"WM_DELETE_WINDOW\", on_closing)\n\n\n# connecting to the server: ---sockets part---\nHOST = input(\"Enter host:\")\nPORT = input(\"Enter port:\")\nif not PORT:\n\tPORT = 33000\t# default value.\nelse:\n\tPORT = int(PORT)\nif not HOST:\n\tHOST = \"127.0.0.1\"\n\nBUFSIZ = 4096\nADDR = (HOST, PORT)\nclient_socket = socket(AF_INET, SOCK_STREAM)\nclient_socket.connect(ADDR)\n\nreceive_thread = Thread(target=receive)\nreceive_thread.start()\ntkinter.mainloop()\t# starts GUI execution\n\n","sub_path":"chat app/cli_exe/client_of_chat.py","file_name":"client_of_chat.py","file_ext":"py","file_size_in_byte":2551,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"5542226","text":"# python -m src_yml.preprocess_v2\n# %%\ntry:\n import warnings\n warnings.filterwarnings('ignore')\n import utils\nexcept Exception as e:\n print(e)\n pass\n# %%\nimport numpy as np\nimport pandas as pd\nfrom glob import glob\nimport os\nimport shutil\nimport matplotlib.pyplot as plt\nfrom sklearn.model_selection import StratifiedKFold\nfrom tqdm import tqdm, tqdm_gui, trange\nfrom PIL import Image, ImageEnhance, ImageFilter\n# %%\npath_data = 'garbage_classify/train_data/'\npath_data_train = 'tmp/data_train/'\npath_data_valid = 'tmp/data_valid/'\nlabels_file = 'tmp/labels_raw.csv'\nimg_size = 224\n# %%\n\ntry:\n labels = pd.read_csv(labels_file)\nexcept FileNotFoundError as e:\n print(e)\n labels = glob(f'{path_data}/*.txt')\n labels = pd.concat([pd.read_csv(label_f, header=None)\n for label_f in labels])\n labels.columns = ['fname', 'label']\n labels.to_csv(labels_file, index=None)\n\nlabels.head()\n\n# %%\nkfold = StratifiedKFold(n_splits=5, random_state=201908, shuffle=True)\n\n# %%\n# for train, valid in kfold.split(labels.fname, labels.label):\n# print(train.shape, valid.shape)\nidx_tr, idx_val = next(kfold.split(labels.fname, labels.label))\n# %%\nlabels_tr = labels.iloc[idx_tr]\nlabels_val = labels.iloc[idx_val]\n# labels_tr.to_csv('tmp/labels_train.csv',index=None)\nlabels_val.to_csv('tmp/labels_valid.csv', index=None)\n\n# %%\nlabels_tr.groupby(by='label').count().plot()\n\n# %%\nlabels_val.groupby(by='label').count().plot()\n# %%\nshutil.rmtree(path_data_train, True)\nshutil.rmtree(path_data_valid, True)\nos.mkdir(path_data_train)\nos.mkdir(path_data_valid)\n# # %%\n# for r in tqdm(labels_val.itertuples(), desc='Validation', total=labels_val.shape[0]):\n# img = Image.open(path_data+r.fname)\n# img_new = img.resize((img_size, img_size), Image.BICUBIC)\n# img_new.save(path_data_valid+r.fname)\n# # %%\n# for r in tqdm(labels_tr.itertuples(), desc='Train raw', total=labels_tr.shape[0]):\n# img = Image.open(path_data+r.fname)\n# img_new = img.resize((img_size, img_size), Image.BICUBIC)\n# img_new.save(path_data_train+r.fname)\n# %%\nflips = [-1, Image.FLIP_LEFT_RIGHT]\nrotates = [0, 30, 90, 120, 180, 210, 270, 300]\n# filters = [ImageFilter.BoxBlur(\n# 0), ImageFilter.GaussianBlur(7), ImageFilter.SHARPEN]\n# ehcs = [ImageEnhance.Sharpness(1),\n# ImageEnhance.Sharpness(0.7),\n# # ImageEnhance.Sharpness(1.4),\n# ImageEnhance.Brightness(0.6),\n# ImageEnhance.Brightness(1.4),\n# ImageEnhance.Contrast(0.6),\n# ImageEnhance.Contrast(1.4),\n# ]\nehcs = [(ImageEnhance.Sharpness, 1, 's1'),\n (ImageEnhance.Sharpness, 0, 's0'),\n (ImageEnhance.Brightness, 0.5, 'b0.5'),\n (ImageEnhance.Brightness, 1.4, 'b1.4'),\n (ImageEnhance.Contrast, 0.5, 'c0.5'),\n (ImageEnhance.Contrast, 1.5, 'c1.5'),\n ]\n# %%\n\n\ndef img_aug(r):\n data = []\n img_raw = Image.open(\n path_data+r.fname).resize((img_size, img_size), Image.BICUBIC)\n for ro in rotates:\n for flp in flips:\n for eh in ehcs:\n if (ro, flp, eh) == (0, -1, ehcs[0]):\n continue\n fname = f'{ro}_{eh[2]}_{flp}_{r.fname}'\n data.append([fname, r.label])\n # 数据增强\n if flp == -1:\n img_flp = img_raw\n else:\n img_flp = img_raw.transpose(flp)\n img_rotate = img_flp.rotate(ro)\n img_ehc = eh[0](img_rotate).enhance(eh[1])\n img_ehc.save(path_data_train+fname)\n\n return pd.DataFrame(data, columns=labels.columns)\n\n\n# %%\ntrain_lbs = []\naug_times = len(rotates)*len(ehcs)*len(flips) # 扩充倍数\nmin_smaples = labels_tr.groupby(by='label').count().min().fname\nfor label in labels.label.unique():\n label_imgs = labels_tr[labels_tr.label == label]\n c_samples = label_imgs.shape[0]\n alpha = (aug_times*min_smaples - c_samples)/((aug_times-1)*c_samples)\n alpha = alpha if alpha > 0 else 0\n print(f'Label {label:02d}, {c_samples:03d} samples, alpha={alpha:.4f}')\n imgs_aug = label_imgs.sample(frac=alpha)\n aug_rows = []\n for r in tqdm(imgs_aug.itertuples(), total=imgs_aug.shape[0]):\n new_rows = img_aug(r)\n aug_rows.append(new_rows)\n # break\n train_lbs.append(pd.concat(aug_rows))\n # break\nnew_lbs = pd.concat(train_lbs)\nlabels_train = pd.concat([labels_tr, new_lbs]) # 合并新旧数据\nlabels_train.to_csv('tmp/labels_train.csv', index=None)\n# %%\nlabels_train = pd.read_csv('tmp/labels_train.csv')\nlabels_train.groupby(by='label').count().plot()\n\n\n# %%\n","sub_path":"src_yml/preprocess_v2.py","file_name":"preprocess_v2.py","file_ext":"py","file_size_in_byte":4609,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"104760856","text":"#!/usr/bin/python3\n# File Name: streamer.py\n# Author: F.a.Leistra\n# Date: 05-02-2021\n# Desrc: Main file for the crypto trader (simulation)\n\n\nimport pandas as pd\nfrom binance.client import Client\nfrom datetime import datetime\nimport time\nimport sys\n\n\ndef check_candle(open, close):\n '''Function that checks the difference between the opening and closing value\n of a candle. The function returns 1 if the candle closed higher than opened and\n vice versa'''\n\n value = float(close) - float(open)\n if value > 0:\n return 1\n else:\n return -1\n\ndef get_candle_data(client, sign):\n '''Function that will make sure all data gets obtained from the binance API'''\n\n while True:\n candles = client.get_historical_klines(sign, Client.KLINE_INTERVAL_1MINUTE, \"2 min ago UTC\")\n if len(candles) == 2:\n break\n else:\n print('not all data obtained, waiting 2 secs')\n time.sleep(2)\n\n return candles\n\n\ndef main():\n\n client = Client('222f7FGVWJ6EVdwFA71gx67QuQK4p5pkU3oEgXZqgAmNGEcLmfhfZSxBtSrVG0aX',\n '9Ta87I5ky0lN0cDOGM1KrdGJlXGzVJ65dagEQySDY43yp9WHUhhIsXHtpnPjxD8K')\n\n # Testing the connectivity\n print(client.ping())\n\n # Assinging the BTC symbol\n BTC = 'BTCUSDT'\n\n # Obtaining the order book\n order_book = client.get_order_book(symbol=BTC)\n\n print(order_book.keys())\n\n candles = client.get_historical_klines(BTC, Client.KLINE_INTERVAL_1MINUTE, \"2 min ago UTC\")\n\n # infitine loop\n # - tijd 1 min verder -> candles ophalen\n # - afgelopen 9 candles vergelijken met stijns patronen\n # - 10 met 15 vergelijken\n # - wanneer een patroon wordt gevonden een simulatie bet plaatsen\n # - uitkomst van die bet opslaan\n\n patterns = []\n\n for c in candles:\n print(c)\n print(c[0] / 1000)\n print('Time:', datetime.utcfromtimestamp(c[0] / 1000).strftime('%Y-%m-%d %H:%M:%S'))\n print('Open:', c[1])\n print('High:', c[2])\n print('Low:', c[3])\n print('Close:', c[4])\n value = float(c[4]) - float(c[1])\n print(value)\n if value > 0:\n print('green', 1)\n else:\n print('red', -1)\n\n patterns.append(value)\n\n patterns = {\n '[1, -1, 1, -1, 1, -1, -1, 1, -1]': 'down', # 1\n '[-1, -1, -1, -1, 1, -1, 1, -1, -1]': 'down', # 2\n '[-1, -1, 1, -1, 1, -1, 1, -1, 1]': 'down', # 3\n '[1, -1, -1, 1, 1, -1, 1, -1, 1]': 'down', # 4\n '[1, -1, 1, -1, 1, -1, 1, -1, -1]': 'down', # 5\n '[1, -1, 1, -1, 1, 1, 1, -1, -1]': 'down', # 6\n '[1, -1, 1, 1, 1, 1, 1, -1, 1]': 'down', # 7\n '[1, 1, 1, -1, 1, -1, -1, -1, 1]': 'down', # 8\n '[1, 1, 1, -1, 1, -1, 1, 1, -1]': 'down', # 9\n '[1, 1, 1, 1, 1, -1, 1, -1, 1]': 'down', # 10\n '[-1, -1, -1, 1, -1, 1, -1, -1, 1]': 'up', # 11\n '[-1, 1, 1, -1, -1, 1, -1, 1, -1]': 'up', # 12\n '[-1, 1, -1, 1, 1, 1, -1, -1, 1]': 'up', # 13\n '[-1, -1, -1, -1, -1, 1, -1, -1, 1]': 'up', # 14\n '[-1, 1, -1, 1, -1, 1, -1, 1, -1]': 'up', # 15\n '[-1, 1, -1, 1, -1, 1, -1, -1, -1]': 'up', # 16\n '[1, 1, -1, 1, -1, 1, 1, 1, -1]': 'up', # 17\n '[1, 1, 1, 1, -1, -1, -1, -1, -1]': 'up', # 18\n '[-1, 1, 1, 1, -1, 1, -1, -1, -1]': 'up', # 19\n '[-1, 1, 1, 1, 1, 1, -1, 1, -1]': 'up'} # 20\n\n for i in patterns.keys():\n i = i.strip('][').split(', ')\n\n time_count = 0\n pattern_storage = []\n cash = 1000\n multiplier = 125\n\n\n with open('status_log.txt', 'w+') as f:\n f.write('Trade History:\\n')\n\n count = 0\n while True:\n print('--- Initiating Loop ---')\n print('This is wat the pattern storage looks like: {}'.format(pattern_storage))\n candles = get_candle_data(client, BTC)\n # Loop for checking the time\n while True:\n candles1 = get_candle_data(client, BTC)\n # if the close time of our initial candles matches the open time of our new candles, a minute has passed\n if candles[1][0] == candles1[0][0]:\n print('Binance - Minute has passed')\n count += 1\n break\n\n else:\n time.sleep(1)\n\n # Main loop for making the trades, this loop will only be reached if a minute has passed\n open_candle = candles1[0][1]\n close_candle = candles1[0][4]\n\n # Determining the situation\n print('checking outcome of {} and {}'.format(open_candle, close_candle))\n candle_outcome = check_candle(open_candle, close_candle)\n print('outcome is {}'.format(candle_outcome))\n pattern_storage.append(candle_outcome)\n print(pattern_storage)\n\n time_count += 1\n\n # Checking if the storgae contains enough values for a possible pattern\n if len(pattern_storage) == 9:\n # Checking if the pattern exists in our dictionary\n if str(pattern_storage) in patterns:\n outcome = patterns[str(pattern_storage)]\n print('------------------ A PATTERN HAS BEEN FOUND -------------------')\n\n bet_size = 0.1 * cash\n cash -= bet_size\n bet_count = 0\n\n while True:\n candles2 = get_candle_data(client, BTC)\n if candles1[1][0] == candles2[0][0]:\n # 1 minute has passed\n bet_count += 1\n if bet_count == 1:\n close_price = candles2[0][4]\n\n if bet_count == 5:\n final_price = candles2[0][4]\n final_outcome = float(final_price) - float(close_price)\n break\n\n else:\n time.sleep(5)\n\n # bet was down\n if outcome == 'down':\n if float(final_price) < float(close_price):\n # bet was correct\n correct = True\n cash += (-final_outcome * multiplier * bet_size) + bet_size\n else:\n # bet was incorrect\n correct = False\n cash += (final_outcome * multiplier * bet_size) + bet_size\n\n # bet was up\n else:\n if float(final_price) > float(close_price):\n # bet was corect\n correct = True\n cash += (final_outcome * multiplier * bet_size) + bet_size\n else:\n # bet was incorrect\n correct = False\n cash += (-final_outcome * multiplier * bet_size) + bet_size\n\n print('Bet: {}, Outcome: {}, Cash: {}'.format(outcome, correct, cash))\n\n with open('status_log.txt', 'a') as f:\n f.write('pattern detected {} - Bet outcome is {} - Correct: {} - Cash: {}\\n'\n .format(str(pattern_storage), outcome, correct, cash))\n pattern_storage.clear()\n else:\n time.sleep(1)\n\n if len(pattern_storage) > 0:\n pattern_storage.pop(0)\n\n print('-------- COMPLETED ONE ITERATION ---------')\n if count == 120:\n print('exiting...')\n break\n\n\nif __name__ == \"__main__\":\n main()","sub_path":"streamer.py","file_name":"streamer.py","file_ext":"py","file_size_in_byte":7486,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"604321803","text":"#coding=utf-8\nfrom time import sleep\nfrom appium import webdriver\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.support.ui import WebDriverWait\nfrom selenium.webdriver.support import expected_conditions\nimport time\ndef find_toast(message, timeout, poll_frequency, driver):\n message = '//*[@text=\\'{}\\']'.format(message)\n element = WebDriverWait(driver, timeout, poll_frequency).until(expected_conditions.presence_of_element_located((By.XPATH, message)))\n print(element.text) #获取元素文本\ndesired_caps = {}\ndesired_caps['platformName'] = 'Android'\ndesired_caps['platformVersion'] = '4.4.2'\ndesired_caps['deviceName'] = '127.0.0.1:62001'\ndesired_caps['appPackage'] = 'com.testerhome.webview'\ndesired_caps['appActivity'] = 'com.testerhome.webview.MainActivity'\ndesired_caps['noReset'] = 'True'\ndesired_caps['automationName'] = 'Uiautomator2' #这个条件必须加上\ndriver = webdriver.Remote('http://localhost:4723/wd/hub', desired_caps)\n# m=driver.find_element_by_android_uiautomator('new UiSelector().text(\"click toast\")').get_attribute('text')\n# m2=driver.find_element_by_android_uiautomator('new UiSelector().text(\"click toast\")').get_attribute('name')\n# driver.find_element_by_android_uiautomator('new UiSelector().text(\"click toast\")').click()\n# m=driver.find_element_by_android_uiautomator('new UiSelector().text(\"click toast\")').get_attribute('text')\n# print(m,m2)\n# driver.find_element_by_android_uiautomator('click toast').click()\n# driver.find_element_by_xpath(\"//android.widget.Button[contains(@text,'click toast')]\").click() #xpath定位\n# print(driver.contexts)\ntime.sleep(2)\nprint(driver.contexts)\ndriver.switch_to.context('WEBVIEW_com.testerhome.webview')\ntry:\n driver.find_element_by_accessibility_id('小说 Link').click()\nexcept Exception as e:\n print(e)\n\n# print(driver.page_source)\n# driver.find_element_by_partial_link_text(\"小说\").click()\n# driver.find_element_by_xpath(\"//[contains(text(),'小说')]\").click()\n# el=driver.find_elements_by_xpath('android.view.View')\n# for i in el:\n# print(i)\n# print(len(el))\n# for i in el:\n# if i.get_attribute('name')==\"小说 Link\":\n# i.click()\n# else:\n# print(i.get_attribute(\"name\"))\n\ndriver.switch_to.context('NATIVE_APP')\nel=driver.find_elements_by_class_name('android.widget.Button')\nfor i in el:\n print(i)\nprint(len(el))\n# driver.find_element_by_id('com.testerhome.webview:id/clickToToastActivity').click()\ndriver.find_element_by_xpath(\"//android.widget.Button[contains(@text,'click toast')]\").click()\ndriver.find_element_by_id(\"com.testerhome.webview:id/toast\").click()\nfind_toast('Toast Test',3,0.5,driver)\n# driver.switch_to.context('WEBVIEW_com.testerhome.webview')\n# print(driver.current_context)\n# driver.switch_to.context('NATIVE_APP')\n# print(driver.current_context)\n# driver.find_element_by_id('com.testerhome.webview:id/toast')\ndriver.quit()","sub_path":"table_test/old/webview_test.py","file_name":"webview_test.py","file_ext":"py","file_size_in_byte":2893,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"194992132","text":"#!../bin/python\n\nimport sys\nimport math\nimport string\nimport numpy as np\nimport random\n\nimport pygame\nfrom utils import Move, closest_distance\n\nbackground = 255, 226, 191\nwall = pygame.image.load('images/wall.png')\nfloor = pygame.image.load('images/floor.png')\nbox = pygame.image.load('images/box.png')\nbox_docked = pygame.image.load('images/box_docked.png')\nworker = pygame.image.load('images/worker.png')\nworker_docked = pygame.image.load('images/worker_dock.png')\ndocker = pygame.image.load('images/dock.png')\nclass Game:\n ACTION_SPACE_SIZE = 5\n \n convert = {\n \"#\": 0, # wall\n \" \": 0.2, # floor\n \".\": 0.3, # dock\n \"$\": 0.5, # box\n \"*\": 0.6, # box on dock\n \"+\": 0.7, # worker on dock\n \"@\": 0.8 # worker on floor\n # current robot 1\n }\n\n def is_valid_value(self, char):\n if (char == ' ' or # floor\n char == '#' or # wall\n char == '@' or # worker on floor\n char == '.' or # dock\n char == '*' or # box on dock\n char == '$' or # box\n char == '+'): # worker on dock\n return True\n else:\n return False\n\n def __init__(self, filename, level, n_robots = 1):\n level = int(level)\n self.filename = filename\n self.matrix = []\n self.robots = []\n self.index = 0\n self.episode_step = 0\n self.level = level\n self.n_robots = n_robots\n if level < 1:\n print(\"ERROR: Level \"+str(level)+\" is out of range\")\n sys.exit(1)\n else:\n file = open(filename, 'r')\n level_found = False\n for line in file:\n row = []\n if not level_found:\n if \"Level \"+str(level) == line.strip():\n level_found = True\n else:\n if line.strip() != \"\":\n row = []\n for c in line:\n if c != '\\n' and self.is_valid_value(c):\n row.append(c)\n elif c == '\\n': # jump to next row when newline\n continue\n else:\n print(\"ERROR: Level \"+str(level) +\n \" has invalid value \"+c)\n sys.exit(1)\n self.matrix.append(row)\n else:\n break\n self.robots = self.get_robots()\n if len(self.robots) == 0:\n self.robots = self.set_robots()\n\n def reset(self):\n self.matrix = []\n self.robots = []\n self.index = 0\n self.episode_step = 0\n level=self.level\n if level < 1:\n print(\"ERROR: Level \"+str(level)+\" is out of range\")\n sys.exit(1)\n else:\n file = open(self.filename, 'r')\n level_found = False\n for line in file:\n row = []\n if not level_found:\n if \"Level \"+str(level) == line.strip():\n level_found = True\n else:\n if line.strip() != \"\":\n row = []\n for c in line:\n if c != '\\n' and self.is_valid_value(c):\n row.append(c)\n elif c == '\\n': # jump to next row when newline\n continue\n else:\n print(\"ERROR: Level \"+str(level) +\n \" has invalid value \"+c)\n sys.exit(1)\n self.matrix.append(row)\n else:\n break\n self.robots = self.get_robots()\n if len(self.robots) == 0:\n self.robots = self.set_robots()\n return self.get_state()\n\n\n def get_robots(self):\n x = 0\n y = 0\n robots = []\n for row in self.matrix:\n for char in row:\n if char == '@': # robot on floor\n robots.append((x, y))\n elif char == '+': # robot on goal\n robots.append((x, y))\n x = x + 1\n x = 0\n y = y + 1\n return robots\n\n def set_robots(self):\n robots = []\n x = 0\n y = 0\n for row in self.matrix:\n for char in row:\n if char == \" \" :\n robots.append((x, y,\"@\"))\n if char == \".\":\n robots.append((x, y,\"+\"))\n x = x + 1\n x = 0\n y = y + 1\n\n robots = random.sample(robots, self.n_robots)\n for robot in robots:\n self.set_content(robot[0], robot[1], robot[2])\n return [(robot[0], robot[1]) for robot in robots]\n\n def load_size(self):\n x = 0\n y = len(self.matrix)\n for row in self.matrix:\n if len(row) > x:\n x = len(row)\n return (x * 32, y * 32)\n\n def get_matrix(self):\n return self.matrix\n\n def print_matrix(self):\n for row in self.matrix:\n for char in row:\n sys.stdout.write(char)\n sys.stdout.flush()\n sys.stdout.write('\\n')\n\n def get_content(self, x, y):\n return self.matrix[y][x]\n\n def set_content(self, x, y, content):\n if self.is_valid_value(content):\n self.matrix[y][x] = content\n else:\n print(\"ERROR: Value '\"+content+\"' to be added is not valid\")\n\n def can_move(self, x, y):\n return self.get_content(self.robots[self.index][0]+x, self.robots[self.index][1]+y) not in ['#', '*', '$', '@','+']\n\n def next(self, x, y):\n return self.get_content(self.robots[self.index][0]+x, self.robots[self.index][1]+y)\n\n def can_push(self, x, y):\n return (self.next(x, y) in ['*', '$'] and self.next(x+x, y+y) in [' ', '.'])\n\n def is_completed(self):\n for row in self.matrix:\n for cell in row:\n if cell == '$':\n return False\n return True\n\n def move_box(self, x, y, a, b):\n # (x,y) -> move to do\n # (a,b) -> box to move\n current_box = self.get_content(x, y)\n future_box = self.get_content(x+a, y+b)\n if current_box == '$' and future_box == ' ':\n self.set_content(x+a, y+b, '$')\n self.set_content(x, y, ' ')\n elif current_box == '$' and future_box == '.':\n self.set_content(x+a, y+b, '*')\n self.set_content(x, y, ' ')\n elif current_box == '*' and future_box == ' ':\n self.set_content(x+a, y+b, '$')\n self.set_content(x, y, '.')\n elif current_box == '*' and future_box == '.':\n self.set_content(x+a, y+b, '*')\n self.set_content(x, y, '.')\n\n def action(self, choice):\n if choice == 0:\n moves = self.move(0, -1, True)\n elif choice == 1:\n moves = self.move(0, 1, True)\n elif choice == 2:\n moves = self.move(-1, 0, True)\n elif choice == 3:\n moves = self.move(1, 0, True)\n elif choice == 4:\n moves = [Move((0,0),(0,0),\"\")]\n else:\n print(\"choice not supported\")\n print(choice)\n\n self.index += 1\n self.index = self.index % len(self.robots)\n return moves\n\n def move(self, x, y, save):\n moves = []\n if self.can_move(x, y):\n current = self.robots[self.index]\n char = self.get_content(current[0], current[1])\n future = self.next(x, y)\n moves.append(Move(\n (current[0], current[1]),\n (current[0]+x, current[1]+y),\n \"robot\"\n ))\n if char == '@' and future == ' ':\n # worker to floor\n self.set_content(current[0]+x, current[1]+y, '@')\n self.set_content(current[0], current[1], ' ')\n elif char == '@' and future == '.':\n # worker to goal\n self.set_content(current[0]+x, current[1]+y, '+')\n self.set_content(current[0], current[1], ' ')\n elif char == '+' and future == ' ':\n # worker on goal to floor\n self.set_content(current[0]+x, current[1]+y, '@')\n self.set_content(current[0], current[1], '.')\n elif char == '+' and future == '.':\n # worker on goal to goal\n self.set_content(current[0]+x, current[1]+y, '+')\n self.set_content(current[0], current[1], '.')\n self.robots[self.index] = (current[0]+x, current[1]+y)\n elif self.can_push(x, y):\n current = self.robots[self.index]\n char = self.get_content(current[0], current[1])\n future = self.next(x, y)\n future_box = self.next(x+x, y+y)\n moves.append(Move(\n (current[0], current[1]),\n (current[0]+x, current[1]+y),\n \"robot\"\n ))\n moves.append(Move(\n (current[0]+x, current[1]+y),\n (current[0]+2*x, current[1]+2*y),\n \"box\"\n ))\n if char == '@' and future == '$' and future_box == ' ':\n # worker push box to floor\n self.move_box(current[0]+x, current[1]+y, x, y)\n self.set_content(current[0], current[1], ' ')\n self.set_content(current[0]+x, current[1]+y, '@')\n elif char == '@' and future == '$' and future_box == '.':\n # worker push box to goal\n self.move_box(current[0]+x, current[1]+y, x, y)\n self.set_content(current[0], current[1], ' ')\n self.set_content(current[0]+x, current[1]+y, '@')\n elif char == '@' and future == '*' and future_box == ' ':\n # worker push box on goal to floor\n self.move_box(current[0]+x, current[1]+y, x, y)\n self.set_content(current[0], current[1], ' ')\n self.set_content(current[0]+x, current[1]+y, '+')\n elif char == '@' and future == '*' and future_box == '.':\n # worker push box on goal to goal\n self.move_box(current[0]+x, current[1]+y, x, y)\n self.set_content(current[0], current[1], ' ')\n self.set_content(current[0]+x, current[1]+y, '+')\n if char == '+' and future == '$' and future_box == ' ':\n # worker on goal push box to floor\n self.move_box(current[0]+x, current[1]+y, x, y)\n self.set_content(current[0], current[1], '.')\n self.set_content(current[0]+x, current[1]+y, '@')\n elif char == '+' and future == '$' and future_box == '.':\n # worker on goal push box to goal\n self.move_box(current[0]+x, current[1]+y, x, y)\n self.set_content(current[0], current[1], '.')\n self.set_content(current[0]+x, current[1]+y, '@')\n elif char == '+' and future == '*' and future_box == ' ':\n # worker on goal push box on goal to floor\n self.move_box(current[0]+x, current[1]+y, x, y)\n self.set_content(current[0], current[1], '.')\n self.set_content(current[0]+x, current[1]+y, '+')\n elif char == '+' and future == '*' and future_box == '.':\n # worker on goal push box on goal to goal\n self.move_box(current[0]+x, current[1]+y, x, y)\n self.set_content(current[0], current[1], '.')\n self.set_content(current[0]+x, current[1]+y, '+')\n self.robots[self.index] = (current[0]+x, current[1]+y)\n else:\n moves.append(Move(\n (0, 0),\n (0, 0),\n \"\"\n ))\n return moves\n\n def step(self, action):\n self.action(action)\n self.episode_step += 1\n if self.episode_step > 100:\n return self.get_state(), self.reward(), True\n else:\n return self.get_state(), self.reward(), self.is_completed()\n\n def get_state(self):\n state = np.zeros((10,11,1))\n x = 0\n y = 0\n for row in self.matrix:\n for char in row:\n state[y][x]=[self.convert[char]]\n x = x + 1\n x = 0\n y = y + 1\n if state[self.robots[self.index][1],self.robots[self.index][0]] != 0.7: #robot on dock\n state[self.robots[self.index][1],self.robots[self.index][0]] = [1]\n return state\n\n def reward(self):\n goals= []\n boxes = []\n available_boxes = []\n available_goals = []\n x = 0\n y = 0\n for row in self.matrix:\n for char in row:\n if char == \".\" or char == \"*\" or char ==\"+\":\n goals.append((x,y))\n if char == \".\" or char ==\"+\":\n available_goals.append((x,y))\n if char == '$' or char == \"*\":\n boxes.append((x,y))\n if char == '$':\n available_boxes.append((x,y))\n x = x + 1\n x = 0\n y = y + 1\n if(self.is_completed()):\n return 50\n else:\n total = 0\n bonus = 0\n for box in available_boxes:\n closest_dist = closest_distance(box,available_goals)\n total += closest_dist\n for robot in self.robots:\n total += closest_distance(robot,available_boxes)\n\n bonus = len(boxes) - len(available_boxes)\n return 4/total + bonus\n \ndef print_game(matrix, screen):\n screen.fill(background)\n x = 0\n y = 0\n for row in matrix:\n for char in row:\n if char == ' ': # floor\n screen.blit(floor, (x, y))\n elif char == '#': # wall\n screen.blit(wall, (x, y))\n elif char == '@': # worker on floor\n screen.blit(worker, (x, y))\n elif char == '.': # dock\n screen.blit(docker, (x, y))\n elif char == '*': # box on dock\n screen.blit(box_docked, (x, y))\n elif char == '$': # box\n screen.blit(box, (x, y))\n elif char == '+': # worker on dock\n screen.blit(worker_docked, (x, y))\n x = x + 32\n x = 0\n y = y + 32\n\n\ndef get_key():\n while 1:\n event = pygame.event.poll()\n if event.type == pygame.KEYDOWN:\n return event.key\n else:\n pass\n\n\ndef display_box(screen, message):\n \"Print a message in a box in the middle of the screen\"\n fontobject = pygame.font.Font(None, 18)\n pygame.draw.rect(screen, (0, 0, 0),\n ((screen.get_width() / 2) - 100,\n (screen.get_height() / 2) - 10,\n 200, 20), 0)\n pygame.draw.rect(screen, (255, 255, 255),\n ((screen.get_width() / 2) - 102,\n (screen.get_height() / 2) - 12,\n 204, 24), 1)\n if len(message) != 0:\n screen.blit(fontobject.render(message, 1, (255, 255, 255)),\n ((screen.get_width() / 2) - 100, (screen.get_height() / 2) - 10))\n pygame.display.flip()\n\n\ndef display_end(screen):\n message = \"Level Completed\"\n fontobject = pygame.font.Font(None, 18)\n pygame.draw.rect(screen, (0, 0, 0),\n ((screen.get_width() / 2) - 100,\n (screen.get_height() / 2) - 10,\n 200, 20), 0)\n pygame.draw.rect(screen, (255, 255, 255),\n ((screen.get_width() / 2) - 102,\n (screen.get_height() / 2) - 12,\n 204, 24), 1)\n screen.blit(fontobject.render(message, 1, (255, 255, 255)),\n ((screen.get_width() / 2) - 100, (screen.get_height() / 2) - 10))\n pygame.display.flip()\n\n\ndef ask(screen, question):\n \"ask(screen, question) -> answer\"\n pygame.font.init()\n current_string = []\n display_box(screen, question + \": \" + \"\".join(current_string))\n while 1:\n inkey = get_key()\n if inkey == pygame.K_BACKSPACE:\n current_string = current_string[0:-1]\n elif inkey == pygame.K_RETURN:\n break\n elif inkey == pygame.K_MINUS:\n current_string.append(\"_\")\n elif inkey <= 127:\n current_string.append(chr(inkey))\n display_box(screen, question + \": \" + \"\".join(current_string))\n return \"\".join(current_string)\n\n\ndef start_game():\n start = pygame.display.set_mode((320, 240))\n level = ask(start, \"Select Level\")\n if int(level) > 0:\n return level\n else:\n print(\"ERROR: Invalid Level: \"+str(level))\n sys.exit(2)\n\ndef checkSameBox(moves):\n for i, move in enumerate(moves):\n for move2 in moves[i:]:\n if(move.end[0] == move2.start[0] and move.end[1] == move2.start[1]):\n return True\n return False","sub_path":"sokoban.py","file_name":"sokoban.py","file_ext":"py","file_size_in_byte":17274,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"357777095","text":"# 0.\r\n# 使用递归编写一个十进制转换为二进制的函数(要求采用“取2取余”的方式,\r\n# 结果与调用bin()一样返回字符串形式)。\r\ndef Dec2Bin(x):\r\n if x:\r\n return x % 2 + Dec2Bin(x//2) * 10\r\n else:\r\n return 0\r\nprint(Dec2Bin(62))\r\n\r\n\r\n# def Dec2Bin(dec):\r\n# result = ''\r\n#\r\n# if dec:\r\n# result = Dec2Bin(dec // 2)\r\n# return result + str(dec % 2)\r\n# else:\r\n# return result\r\n#\r\n#\r\n# print(Dec2Bin(62))\r\n# 1.\r\n# 写一个函数get_digits(n),将参数n分解出每个位的数字并按顺序存放到列表中。\r\nlist1 = []\r\ndef get_digits(n):\r\n if n:\r\n list1.append(n%10)\r\n return get_digits(n//10)\r\n else:\r\n list1.reverse()\r\n print(list1)\r\nget_digits(12345)\r\n\r\nresult = []\r\n\r\n\r\n# def get_digits(n):\r\n# if n > 0:\r\n# result.insert(0, n % 10)\r\n# get_digits(n // 10)\r\n#\r\n#\r\n# get_digits(12345)\r\n# print(result)\r\n# 2.\r\n# 还记得求回文字符串那道题吗?现在让你使用递归的方式来求解,亲还能骄傲的说我可以吗?\r\n# str1 = input(\"请输入字符串\")\r\n# def couplet(n=0):\r\n# if n < len(str1) // 2:\r\n# if str1[n] != str1[-n-1]:\r\n# print(str1,\"不是回文数\")\r\n# return\r\n# else:\r\n# return couplet(n+1)\r\n# else:\r\n# print(str1,\"是回文联\")\r\n# couplet()\r\n\r\n\r\n# def is_palindrome(n, start, end):\r\n# if start > end:\r\n# return 1\r\n# else:\r\n# return is_palindrome(n, start + 1, end - 1)\\\r\n# if n[start] == n[end] else 0\r\n#\r\n#\r\n# string = input('请输入一串字符串:')\r\n# length = len(string) - 1\r\n#\r\n# if is_palindrome(string, 0, length):\r\n# print('\"%s\"是回文字符串!' % string)\r\n# else:\r\n# print('\"%s\"不是回文字符串!' % string)\r\n\r\n# 3.\r\n# 使用递归编程求解以下问题:\r\ndef year(n):\r\n if (n-1):\r\n return year(n-1) + 2\r\n else:\r\n return 10\r\nprint(year(5))\r\n\r\n\r\n# def age(n):\r\n# if n == 1:\r\n# return 10\r\n# else:\r\n# return age(n-1) + 2\r\n\r\n# 4.\r\n# 请写下这一节课你学习到的内容:格式不限,回忆并复述是加强记忆的好方式!\r\n# (1)斐波那契数列的递归实现:\r\n#\r\n# 1,1,2,3,5,8,13,21.....\r\n#\r\n# 我们可以用数学函数来定义:\r\ndef num(x = 1,y = 1): # 范围\r\n if x < 100:\r\n print(x,end=' ')\r\n return num(y,x + y)\r\n else:\r\n print()\r\nnum()\r\n\r\n# 分别用迭代和递归实现:\r\n#\r\n# 迭代\r\n#\r\n#\r\n# def fab(n):\r\n# n1 = 1\r\n# n2 = 1\r\n# n3 = 1\r\n# if n < 1:\r\n# print(\"输入错误!\")\r\n# return -1\r\n# while (n - 2) > 0:\r\n# print(n1, end=' ')\r\n# n3 = n2 + n1\r\n# n1 = n2\r\n# n2 = n3\r\n# n -= 1\r\n# return n3\r\n# fab(10)\r\n#\r\n#\r\n# 递归:\r\n#\r\n# def fab(n):\r\n# if n < 1:\r\n# print('输入有误!')\r\n# return -1\r\n# elif n == 1 or n == 2:\r\n# return 1\r\n# else:\r\n# return ferber(n - 1) + ferber(n - 2)\r\n#\r\n#\r\n# 递归算法称为分治思想。\r\n#\r\n# (2)递归实现汉诺塔\r\n#\r\n# 对于游戏的玩法,我们可以简单分解为三个步骤:\r\n#\r\n# 将前63个盘子从a移到b上\r\n#\r\n# 将最底下的第64个盘子从a移到c\r\n#\r\n# 将b上的63个盘子移到c\r\n#\r\n# 问题1:将a上的63个盘子借助c移到b\r\n#\r\n# 问题2:将b上的63个盘子借助a移到c\r\n#\r\n# 然后:\r\n#\r\n# 问题1拆解为:\r\n#\r\n#         将前62个盘子从a移到c上\r\n#\r\n#         将最底下的第63个盘子从a移到b\r\n#\r\n#         将c上的62个盘子移到b\r\n#\r\n# 问题2拆解为:\r\n#\r\n#         将前62个盘子从b移到a上\r\n#\r\n#         将最底下的第63个盘子从b移到c\r\n#\r\n#         将a上的62个盘子移到b\r\n#\r\ndef hanoi(n,a,b,c):\r\n 'n为盘子数,a,b,c,为三根柱子'\r\n if n==1:\r\n print(a,'->',c)\r\n else:\r\n hanoi(n-1,a,c,b) # 将前n-1个盘子从a 移到b\r\n print(a,'->',c) #将最底下最后一个盘子从a移到c\r\n hanoi(n-1,b,a,c) #将b上的n-1个盘子移到c上\r\n\r\nhanoi(2,'a','b','c')\r\nprint(\"---------------\")\r\nhanoi(3,'a','b','c')\r\n# 动动手\r\n# 0.\r\n# 使用递归编写一个十进��转换为二进制的函数(要求采用“取2取余”的方式,结果与调用bin()\r\n# 一样返回字符串形式)。\r\n#\r\n# def Dec2Bin(dec):\r\n# result = ''\r\n#\r\n# if dec:\r\n# result = Dec2Bin(dec // 2)\r\n# return result + str(dec % 2)\r\n# else:\r\n# return result\r\n#\r\n#\r\n# print(Dec2Bin(62))\r\n# 1.\r\n# 写一个函数get_digits(n),将参数n分解出每个位的数字并按顺序存放到列表中。\r\n# 举例:get_digits(12345) == > [1, 2, 3, 4, 5]\r\n#\r\n# 解题思路:利用除以10取余数的方式,每次调用get_digits(n // 10),并将余数存放到列表中即可。要注意的是结束条件设置正确。\r\n#\r\n# result = []\r\n#\r\n#\r\n# def get_digits(n):\r\n# if n > 0:\r\n# result.insert(0, n % 10)\r\n# get_digits(n // 10)\r\n#\r\n#\r\n# get_digits(12345)\r\n# print(result)\r\n# 2.\r\n# 还记得求回文字符串那道题吗?现在让你使用递归的方式来求解,亲还能骄傲的说我可以吗?\r\n# 解题思路:有好多种方法,不过综合效率来说,小甲鱼的实现方式比较朴素,利用递归每次索引前后两个字符进行对比,当start > end的时候,也正是首尾下标“碰面”的时候,即作为结束递归的条件。\r\n#\r\n# def is_palindrome(n, start, end):\r\n# if start > end:\r\n# return 1\r\n# else:\r\n# return is_palindrome(n, start + 1, end - 1) if n[start] == n[end] else 0\r\n#\r\n#\r\n# string = input('请输入一串字符串:')\r\n# length = len(string) - 1\r\n#\r\n# if is_palindrome(string, 0, length):\r\n# print('\"%s\"是回文字符串!' % string)\r\n# else:\r\n# print('\"%s\"不是回文字符串!' % string)\r\n# 3.\r\n# 使用递归编程求解以下问题:\r\n# 有5个人坐在一起,问第五个人多少岁?他说比第4个人大2岁。问第4个人岁数,他说比第3个人大2岁。问第三个人,又说比第2人大两岁。问第2个人,说比第一个人大两岁。最后问第一个人,他说是10岁。请问第五个人多大?\r\n#\r\n# 解题思路:利用递归的方法,递归分为回推和递推两个阶段。要想知道第五个人岁数,需知道第四人的岁数,依次类推,推到第一人(10\r\n# 岁),再往回推。\r\n#\r\n# def age(n):\r\n# if n == 1:\r\n# return 10\r\n# else:\r\n# return age(n - 1) + 2\r\n#\r\n#\r\n# print('哈哈,我知道了,第五个人的年龄是 %d 岁,啵啵脆!' % age(5))\r\n","sub_path":"ex23、24_recurrence2.py","file_name":"ex23、24_recurrence2.py","file_ext":"py","file_size_in_byte":6669,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"41068973","text":"from django.contrib import admin\nfrom django.urls import path,include\nfrom rest_framework_simplejwt.views import (\n TokenObtainPairView,\n TokenRefreshView,\n TokenVerifyView,\n)\nfrom user.views import RegisterView\n\nurlpatterns = [\n path('admin/', admin.site.urls),\n path('',include('organization.urls')),\n path('auth/',include('rest_framework.urls')),\n path('api/token/verify/', TokenVerifyView.as_view(), name='token_verify'),\n path('api/token/', TokenObtainPairView.as_view(), name='token_obtain_pair'),\n path('api/token/refresh/', TokenRefreshView.as_view(), name='token_refresh'),\n path('register/',RegisterView.as_view()),\n \n]\n","sub_path":"taskapi/taskapi/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":663,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"198140215","text":"'''\n实验名称:Table(表格)\n版本:v1.0\n日期:2020.7\n作者:01Studio【www.01Studio.org】\n'''\n\nimport lvgl as lv\nimport ujson\n\nfrom ili9341 import ili9341\nfrom xpt2046 import xpt2046\n\nTFT_IS_PORTRAIT =1 #竖屏:1 ,横屏:0 ;\nTOUCH_READY = 0 #用于检测触摸屏是否已经校准过;\n\n#LCD ili9341初始化\ndisp = ili9341(\n miso=12,\n mosi=13,\n clk=14,\n cs=15,\n dc=21,\n rst=33,\n power=50, #硬件不支持,随便配一个参数\n backlight=51, #硬件不支持,随便配一个参数\n backlight_on= 1,\n power_on= 1,\n width=240 if TFT_IS_PORTRAIT else 320,\n height=320 if TFT_IS_PORTRAIT else 240,\n rot=ili9341.PORTRAIT if TFT_IS_PORTRAIT else ili9341.LANDSCAPE #垂直方向PORTRAIT ;水平方向:LANDSCAPE\n)\n\n#触摸屏设置校准\nTOUCH_CS = 2 #触摸屏CS片选引脚\nTOUCH_INTERRUPT=0 #横屏\n\nif TFT_IS_PORTRAIT:\n TOUCH_CALI_FILE = \"touch_cali_PORTRAIT.json\" #保存为竖屏触摸参数\nelse:\n TOUCH_CALI_FILE = \"touch_cali_LANDSCAPE.json\" #保存为横屏触摸参数\n\n#从没做过触摸校准\nif TOUCH_CALI_FILE not in uos.listdir():\n touch = xpt2046(\n cs=TOUCH_CS,\n transpose=TFT_IS_PORTRAIT,\n )\n\n from touch_cali import TouchCali\n\n touch_cali = TouchCali(touch, TOUCH_CALI_FILE)\n touch_cali.start()\n\n#已经做过触摸校准,直接调用触摸参数文件\nelse:\n with open(TOUCH_CALI_FILE, 'r') as f:\n param = ujson.load(f)\n touch_x0 = param['cal_x0']\n touch_x1 = param['cal_x1']\n touch_y0 = param['cal_y0']\n touch_y1 = param['cal_y1']\n\n touch = xpt2046(\n cs=TOUCH_CS,\n transpose=TFT_IS_PORTRAIT,\n cal_x0=touch_x0,\n cal_x1=touch_x1,\n cal_y0=touch_y0,\n cal_y1=touch_y1,\n )\n\n TOUCH_READY = 1 #表示已经配置好触摸参数\n\n#############################################\n############### Table ################\n#############################################\n\nif TOUCH_READY:\n\n # Create a normal cell style\n style_cell1 = lv.style_t()\n lv.style_copy(style_cell1, lv.style_plain)\n style_cell1.body.border.width = 1\n style_cell1.body.border.color = lv.color_make(0,0,0)\n\n # Crealte a header cell style\n style_cell2 = lv.style_t()\n lv.style_copy(style_cell2, lv.style_plain)\n style_cell2.body.border.width = 1\n style_cell2.body.border.color = lv.color_make(0,0,0)\n style_cell2.body.main_color = lv.color_make(0xC0, 0xC0, 0xC0)\n style_cell2.body.grad_color = lv.color_make(0xC0, 0xC0, 0xC0)\n\n table = lv.table(lv.scr_act())\n table.set_style(lv.table.STYLE.CELL1, style_cell1)\n table.set_style(lv.table.STYLE.CELL2, style_cell2)\n table.set_style(lv.table.STYLE.BG, lv.style_transp_tight)\n table.set_col_cnt(2)\n table.set_row_cnt(4)\n table.align(None, lv.ALIGN.CENTER, 0, 0)\n\n # Make the cells of the first row center aligned\n table.set_cell_align(0, 0, lv.label.ALIGN.CENTER)\n table.set_cell_align(0, 1, lv.label.ALIGN.CENTER)\n\n # Make the cells of the first row TYPE = 2 (use `style_cell2`)\n table.set_cell_type(0, 0, 2)\n table.set_cell_type(0, 1, 2)\n\n # Fill the first column\n table.set_cell_value(0, 0, \"Name\")\n table.set_cell_value(1, 0, \"Apple\")\n table.set_cell_value(2, 0, \"Banana\")\n table.set_cell_value(3, 0, \"Citron\")\n\n # Fill the second column\n table.set_cell_value(0, 1, \"Price\")\n table.set_cell_value(1, 1, \"$7\")\n table.set_cell_value(2, 1, \"$4\")\n table.set_cell_value(3, 1, \"$6\")\n","sub_path":"esp32_resources/3.pyWiFi-ESP32/5.LVGL/2.小部件实验/21-Table(表格)/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":3507,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"37097980","text":"from flask import Flask, Blueprint\nfrom flask_restplus import Api\n\nimport yaml\nimport os\n\nblueprint = Blueprint(\"open_api\", __name__, url_prefix=\"/open_api\")\napi = Api(blueprint)\n\n\ndef register_api():\n from apps.users.views import DemoView, Servers, Server, UserListApi\n from apps.users import user_api\n api.add_resource(DemoView, \"/home\")\n user_api.add_resource(Servers, \"/servers\")\n user_api.add_resource(Server, \"/servers/<_id>\")\n user_api.add_resource(UserListApi, \"/user/<_id>\")\n\n\ndef read_yaml(config_name, config_path):\n \"\"\"\n config_name:需要读取的配置内容\n config_path:配置文件路径\n \"\"\"\n if config_name and config_path:\n with open(config_path, 'r', encoding='utf-8') as f:\n conf = yaml.safe_load(f.read())\n if config_name in conf.keys():\n return conf[config_name.upper()]\n else:\n raise KeyError('未找到对应的配置信息')\n else:\n raise ValueError('请输入正确的配置名称或配置文件路径')\n\n\ndef create_app(config_name=None, config_path=None):\n app = Flask(__name__, template_folder=\"templates\", static_folder=\"static\")\n app.register_blueprint(blueprint)\n register_api()\n # # 读取配置文件\n if not config_path:\n pwd = os.getcwd()\n config_path = os.path.join(pwd, 'config/config.yaml')\n if not config_name:\n config_name = 'PRODUCTION'\n\n # 读取配置文件\n conf = read_yaml(config_name, config_path)\n app.config.update(conf)\n\n return app\n","sub_path":"apps/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":1534,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"74442005","text":"class Node:\n\tdef __init__(self,dataval = None):\n\t\tself.dataval = dataval\n\t\tself.nextval = None\n\nclass SLinkedList:\n\tdef __init__(self):\n\t\tself.headval = None\n\n\tdef listprint(self):\n\t\tprintval = self.headval\n\t\twhile printval is not None:\n\t\t\tprint(printval.dataval)\n\t\t\tprintval = printval.nextval\n\tdef atbeg(self,newdat):\n\t\tf = self.headval\n\t\tself.headval = newdat\n\t\tself.headval.nextval = f\n\tdef atend(self,newdat):\n\t\tnewnode = Node(newdat)\n\t\tif self.headval is None:\n\t\t\tself.headval = newnode\n\t\t\treturn\n\t\tlaste = self.headval\n\t\twhile(laste.nextval):\n\t\t\tlaste = laste.nextval\n\t\tlaste.nextval = newnode\n\tdef inbetw(self,midnode,newdat):\n\t\tif midnode is None:\n\t\t\tprint(\"midnode is absent\")\n\t\t\treturn\n\t\tnewnode = Node(newdat)\n\t\tnewnode.nextval = midnode.nextval\n\t\tmidnode.nextval = newnode\n\tdef delitem(self,remkey):\n\t\thead = self.headval\n\t\tif (head is not None):\n\t\t\tif head.dataval == remkey:\n\t\t\t\tself.headval = head\n\t\t\t\thead = None\n\t\t\t\treturn\n\n\t\twhile (head is not None):\n\t\t\tif head.dataval == remkey:\n\t\t\t\tbreak\n\t\t\tprev = head\n\t\t\thead = head.nextval\n\t\tif (head == None):\n\t\t\treturn\n\t\tprev.nextval = head.nextval\n\n\t\thead = None","sub_path":"linkedlist.py","file_name":"linkedlist.py","file_ext":"py","file_size_in_byte":1123,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"143239549","text":"'''\n- Leetcode problem: 20\n\n- Difficulty: Easy\n\n- Brief problem description:\n\nGiven a string containing just the characters '(', ')', '{', '}', '[' and ']', determine if the input string is valid.\n\nAn input string is valid if:\n\nOpen brackets must be closed by the same type of brackets.\nOpen brackets must be closed in the correct order.\nNote that an empty string is also considered valid.\n\nExample 1:\n\nInput: \"()\"\nOutput: true\nExample 2:\n\nInput: \"()[]{}\"\nOutput: true\nExample 3:\n\nInput: \"(]\"\nOutput: false\nExample 4:\n\nInput: \"([)]\"\nOutput: false\nExample 5:\n\nInput: \"{[]}\"\nOutput: true\n\n- Solution Summary:\n\n- Used Resources:\n\n--- Bo Zhou\n'''\n\n\nclass Solution:\n def isValid(self, s: str) -> bool:\n pStack = []\n for c in s:\n if c == \"{\":\n pStack.append(\"}\")\n elif c == \"[\":\n pStack.append(\"]\")\n elif c == \"(\":\n pStack.append(\")\")\n elif len(pStack) == 0 or pStack.pop() != c:\n return False\n\n return len(pStack) == 0","sub_path":"p20_Valid_Parentheses.py","file_name":"p20_Valid_Parentheses.py","file_ext":"py","file_size_in_byte":1041,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"569700238","text":"import math\nimport numpy as np\nimport sys\nimport json\nimport sqlite3\nfrom sqlite3 import Error\nimport copy\n\nvideoNum = 50\nlabelNum = 16\n\nlabelPosInDB = 5\ndb_videos = []\ndb_users = []\ndb_history = []\nuser_history = []\ncurrentUser = 10\n\ndiversity_weight = 0.3\n\n\ndef read_in():\n line = sys.stdin.readline()\n return json.loads(line)\n\n\ndef create_connection(db):\n\n try:\n conn = sqlite3.connect(db)\n return conn\n except Error as e:\n print(e)\n\n return None\n\n\ndatabase = \"data_moe.db\"\nconn = create_connection(database)\n\n\ndef selectVideos(connect):\n cur = connect.cursor()\n cur.execute(\"SELECT * FROM Videos\")\n\n rows = cur.fetchall()\n return rows\n\n\ndef selectUsers(connect):\n cur = connect.cursor()\n cur.execute(\"SELECT * FROM Users\")\n\n rows = cur.fetchall()\n return rows\n\n\ndef selectHistory(connect):\n cur = connect.cursor()\n cur.execute(\"SELECT * FROM WatchHistory\")\n\n rows = cur.fetchall()\n return rows\n\n\ndef main():\n global userID\n for line in sys.stdin:\n userID = line\n\n return_to_js = []\n res_list = start()\n for res in res_list:\n return_to_js.append(str(res.video))\n\n# print(return_to_js)\n\n\nclass Video(object):\n def __init__(self, video, value, labels):\n self.video = video\n self.value = value\n self.labels = labels\n\n def __repr__(self):\n return '{}'.format(self.video)\n\n def getkey(self):\n return self.value\n\n def getlabel(self):\n return self.labels\n\n\ndef cosine(a, b):\n numer = sum(float(i) * float(j) for i, j in zip(a, b))\n denomin = rooted(a) * rooted(b)\n return numer / float(denomin)\n\n\ndef rooted(i):\n return math.sqrt(sum([float(a) * float(a) for a in i]))\n\n\ndef euclidean_simil(a, b):\n return math.sqrt(sum(pow(float(i) - float(j), 2) for i, j in zip(a, b)))\n\n\ndef content():\n with conn:\n db_videos = selectVideos(conn)\n db_users = selectUsers(conn)\n\n finalist = []\n\n for elems in db_users:\n if elems[0] == int(userID):\n global userP\n userP = elems[2]\n\n for elems in db_videos:\n userP0 = userP.split(\";\")\n labels = elems[labelPosInDB].split(\";\")\n labels[-1] = labels[-1].strip()\n finalist.append(Video(elems[0], euclidean_simil(userP0, labels), labels))\n\n finalist = sorted(finalist, key=lambda x: x.value, reverse=False)[:videoNum]\n\n return finalist\n\n\ndef diversity(contentlist):\n\n matrix = np.empty((0, labelNum))\n for each in contentlist:\n matrix = np.append(matrix, [each.labels], axis=0)\n\n reversed = matrix.transpose()\n\n totalsum = 0\n div = []\n\n for line in reversed:\n div.append(np.count_nonzero(line == '1') / videoNum)\n\n for each in matrix:\n for pos in range(labelNum):\n totalsum += (float(each[pos]) - div[pos]) ** 2\n\n std_dev = math.sqrt(totalsum / videoNum)\n\n return std_dev\n\n\nclass User(object):\n def __init__(self, user, value, videos):\n self.user = user\n self.value = value\n self.videos = videos\n\n def __repr__(self):\n return '{} {} {}'.format(self.user, self.value, self.videos)\n\n def getuser(self):\n return self.user\n\n def getkey(self):\n return self.value\n\n def getlist(self):\n return self.videos\n\n\ndef dis(a, b):\n numer = sum(float(i) * float(j) for i, j in zip(a, b))\n denomin = rooted(a) * rooted(b)\n return numer / float(denomin)\n\n\ndef rooted(i):\n return math.sqrt(sum([float(a) * float(a) for a in i]))\n\n\ndef collaborative():\n with conn:\n db_users = selectUsers(conn)\n db_history = selectHistory(conn)\n try:\n finalist = []\n userP0 = []\n\n for elems in db_users:\n if elems[0] == int(userID):\n userP0 = list(elems[2].split(\";\"))[:16]\n\n for elems in db_history:\n if elems[0] == int(userID):\n user_history.append(elems[1])\n\n for users in db_users:\n if users[0] != int(userID):\n others = list(users[2].split(\";\"))[:16]\n otherhistories = []\n for histories in db_history:\n if users[0] == histories[0]:\n otherhistories.append(histories[1])\n finalist.append(User(users[0], dis(userP0, others), otherhistories))\n\n userlist = sorted(finalist, key=lambda x: x.value, reverse=True)\n\n return userlist\n\n except:\n print(\"Collaborative Filtering is failed\")\n\n\ndef mix(cnt, topusers, contentlist):\n check = True\n userLabels = userP.split(\";\")\n\n for x in topusers:\n temp = x.videos\n for i in range(0, len(temp)):\n newV = str(temp[i])\n\n for y in contentlist:\n if str(y.video) == newV or newV == '\\n' or user_history.__contains__(newV) or len(newV) != 11:\n check = False\n if check:\n contentlist[cnt * 2 % videoNum].video = newV\n forlabels = findlabel(newV)\n contentlist[cnt * 2 % videoNum].value = euclidean_simil(userLabels, forlabels)\n contentlist[cnt * 2 % videoNum].labels = forlabels\n\n return contentlist\n check = True\n\n return contentlist\n\n\ndef Avg_Simil(list):\n newlist = []\n for each in list:\n newlist.append(each.value)\n return np.average(newlist)\n\n\ndef Clone(li0):\n li_copy = li0[:]\n return li_copy\n\n\ndef findlabel(new):\n label = \"\"\n fv = open('rawdata/RealVideoData_moe.csv')\n\n for line in fv:\n elems = line.split(\",\")\n if (new == elems[0]):\n labels = elems[1].split(\";\")\n labels[-1] = labels[-1].strip()\n # print(labels)\n # label = list(elems[1])[:labelNum]\n\n fv.close()\n return label\n\n\ndef start():\n contentlist = content()\n print(contentlist)\n new_contentlist = copy.deepcopy(contentlist)\n\n std_dev = diversity(contentlist)\n topusers = collaborative()\n newsum = 0\n\n # normalise\n for user in topusers:\n newsum += user.value\n for user in topusers:\n user.value = user.value/newsum\n\n # avg_simil = Avg_Simil(contentlist)\n # max_val = max(0, (avg_simil*(1-diversity_weight) + (diversity_weight*std_dev))/2)\n\n cnt = 1\n while cnt < ((videoNum/5)+1):\n mix(cnt, topusers, contentlist)\n # new_avg_simil = Avg_Simil(contentlist)\n # new_threshold_changed = (new_avg_simil*(1-diversity_weight) + (diversity_weight*std_dev))/2\n cnt += 1\n\n # if (new_threshold_changed > max_val):\n # new_contentlist = copy.deepcopy(contentlist)\n # print(new_contentlist)\n # max_val = max(max_val, new_threshold_changed)\n\n # print(contentlist)\n return contentlist\n\n\ndef run(thisUser):\n global userID\n userID = thisUser\n return start()\n\n\nmain()\n","sub_path":"FitnessRSWebsite/Adaptive_moe.py","file_name":"Adaptive_moe.py","file_ext":"py","file_size_in_byte":6922,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"461305013","text":"# Originally by Alexander (Max) deGroot\n# https://github.com/amdegroot/ssd.pytorch.git\n\n\nimport torch\nfrom enum import Enum\nimport math\nimport gc\n\ndef split(boxes):\n return boxes[..., :2], boxes[..., 2:]\n\ndef split4(boxes):\n return boxes[..., 0], boxes[..., 1], boxes[..., 2], boxes[..., 3]\n\ndef extents_form(boxes):\n lower, upper = split(boxes)\n return torch.cat([(lower + upper) * 0.5, upper - lower], 1)\n\ndef point_form(boxes):\n centre, size = split(boxes)\n extents = size * 0.5\n return torch.cat([centre - extents, centre + extents], 1)\n\n\n\ndef transform(boxes, offset, scale):\n lower, upper = boxes[:, :2], boxes[:, 2:]\n\n offset, scale = torch.Tensor(offset), torch.Tensor(scale)\n\n lower = lower.add(offset).mul(scale)\n upper = upper.add(offset).mul(scale)\n\n return torch.cat([lower.min(upper), lower.max(upper)], 1)\n\n\ndef transpose(boxes):\n if boxes.size(0) > 0:\n x1, y1, x2, y2 = split4(boxes)\n return torch.stack([y1, x1, y2, x2], boxes.dim() - 1)\n else:\n return boxes\n\n\ndef subset(boxes, labels, inds):\n if inds.dim() > 1:\n inds = inds.squeeze(1)\n return torch.index_select(boxes, 0, inds), torch.index_select(labels, 0, inds)\n\n return torch.Tensor(), torch.LongTensor()\n\n\ndef filter_invalid(boxes, labels):\n valid = (boxes[:, 2] - boxes[:, 0] > 0) & (boxes[:, 3] - boxes[:, 1] > 0)\n return subset(boxes, labels, valid.nonzero())\n\n\ndef area(boxes):\n x1, y1, x2, y2 = boxes[:,0], boxes[:,1], boxes[:,2], boxes[:,3]\n return (x2-x1) * (y2-y1)\n\n\ndef filter_hidden(boxes, labels, lower, upper, min_visible=0.0):\n bounds = torch.Tensor([[*lower, *upper]])\n overlaps = (intersect(bounds, boxes) / area(boxes)).squeeze(0)\n return subset(boxes, labels, overlaps.gt(min_visible).nonzero())\n\n\ndef clamp(boxes, lower, upper):\n\n boxes[:, 0].clamp_(min = lower[0])\n boxes[:, 1].clamp_(min = lower[1])\n boxes[:, 2].clamp_(max = upper[0])\n boxes[:, 3].clamp_(max = upper[1])\n\n return boxes\n\n\ndef intersect(box_a, box_b):\n \"\"\" Intersection of bounding boxes\n Args:\n box_a: (tensor) bounding boxes, Shape: [n,4].\n box_b: (tensor) bounding boxes, Shape: [m,4].\n Return:\n (tensor) intersection area, Shape: [n,m].\n \"\"\"\n n = box_a.size(0)\n m = box_b.size(0)\n\n max_xy = torch.min(box_a[:, 2:].unsqueeze(1).expand(n, m, 2),\n box_b[:, 2:].unsqueeze(0).expand(n, m, 2))\n min_xy = torch.max(box_a[:, :2].unsqueeze(1).expand(n, m, 2),\n box_b[:, :2].unsqueeze(0).expand(n, m, 2))\n inter = torch.clamp((max_xy - min_xy), min=0)\n return inter[:, :, 0] * inter[:, :, 1]\n\n\ndef iou(box_a, box_b):\n \"\"\"Compute the IOU of two sets of boxes in point form.\n Args:\n box_a, box b: Bounding boxes in point form. shapes ([n, 4], [m, 4])\n Return:\n jaccard overlap: (tensor) Shape: [n, m]\n \"\"\"\n inter = intersect(box_a, box_b)\n area_a = ((box_a[:, 2]-box_a[:, 0]) *\n (box_a[:, 3]-box_a[:, 1])).unsqueeze(1).expand_as(inter) # [n,m]\n area_b = ((box_b[:, 2]-box_b[:, 0]) *\n (box_b[:, 3]-box_b[:, 1])).unsqueeze(0).expand_as(inter) # [n,m]\n union = area_a + area_b - inter\n return inter / union # [n,m]\n\n\nnms_defaults = {\n 'nms_threshold':0.5,\n 'class_threshold':0.05,\n 'max_detections':100\n}\n\n\ndef nms(boxes, confs, nms_threshold=0.5, class_threshold=0.05, max_detections=100):\n '''Non maximum suppression.\n Args:\n boxes: (tensor) bounding boxes in point form, sized [n,4].\n confs: (tensor) confidence scores, sized [n,].\n nms_threshold: (float) overlap iou threshold.\n class_threshold: (float) absolute threshold for confidence.\n max_detections: (float) max detections (for efficiency)\n Returns:\n keep: indices of boxes to keep\n Reference:\n https://github.com/rbgirshick/py-faster-rcnn/blob/master/lib/nms/py_cpu_nms.py\n '''\n x1, y1, x2, y2 = boxes[:,0], boxes[:,1], boxes[:,2], boxes[:,3]\n areas = (x2-x1) * (y2-y1)\n\n _, order = confs.sort(0, descending=True)\n\n keep = []\n while order.numel() > 0 and len(keep) < max_detections:\n i = order[0].item()\n\n score = confs[i]\n if score < class_threshold:\n break\n\n keep.append(i)\n\n if order.numel() == 1:\n break\n\n xx1 = x1[order[1:]].clamp(min=x1[i])\n yy1 = y1[order[1:]].clamp(min=y1[i])\n xx2 = x2[order[1:]].clamp(max=x2[i])\n yy2 = y2[order[1:]].clamp(max=y2[i])\n\n w = (xx2-xx1).clamp(min=0)\n h = (yy2-yy1).clamp(min=0)\n inter = w * h\n ovr = inter / areas[order[1:]].clamp(max=areas[i])\n\n ids = (ovr <= nms_threshold).nonzero()\n if ids.numel() == 0:\n break\n\n ids = ids.squeeze(1)\n order = order[ids+1]\n\n\n return torch.LongTensor(keep)\n\n\n\n\n\ndef make_boxes(box_sizes, box_dim, image_dim):\n w, h = box_dim\n\n n = len(box_sizes)\n\n xs = torch.arange(0, w, dtype=torch.float).add_(0.5).view(1, w, 1, 1).expand(h, w, n, 1)\n ys = torch.arange(0, h, dtype=torch.float).add_(0.5).view(h, 1, 1, 1).expand(h, w, n, 1)\n\n xs = xs.mul(image_dim[0] / w)\n ys = ys.mul(image_dim[1] / h)\n\n box_sizes = torch.FloatTensor(box_sizes).view(1, 1, n, 2).expand(h, w, n, 2)\n boxes = torch.cat([xs, ys, box_sizes], 3).view(-1, 4)\n\n return boxes\n\n\ndef make_anchors(box_sizes, layer_dims, image_dim, crop_boxes=True):\n boxes = [make_boxes(boxes, box_dim, image_dim) for boxes, box_dim in zip(box_sizes, layer_dims)]\n boxes = torch.cat(boxes, 0)\n\n if crop_boxes:\n return extents_form(clamp(point_form(boxes), (0, 0), image_dim))\n\n return boxes\n\ndef anchor_sizes(size, aspects, scales):\n def anchor(s, ar):\n return (s * math.sqrt(ar), s / math.sqrt(ar))\n\n return [anchor(size * scale, ar) for scale in scales for ar in aspects]\n\ndef encode(boxes, labels, anchor_boxes, match_thresholds=(0.4, 0.5)):\n '''Encode target bounding boxes and class labels.\n We obey the Faster RCNN box coder:\n tx = (x - anchor_x) / anchor_w\n ty = (y - anchor_y) / anchor_h\n tw = log(w / anchor_w)\n th = log(h / anchor_h)\n Args:\n boxes: (tensor) ground truth bounding boxes in point form, sized [n, 4].\n labels: (tensor) object class labels, sized [n].\n anchor_boxes: (tensor) bounding boxes in extents form, sized [m, 4].\n Returns:\n loc_targets: (tensor) encoded bounding boxes, sized [m, 4].\n class_targets: (tensor) encoded class labels, sized [m].\n '''\n\n if boxes.dim() == 1:\n n = anchor_boxes.size(0)\n class_targets = torch.LongTensor(n).fill_(0) # all negative labels\n loc_targets = torch.FloatTensor(n, 4).fill_(0) # will be ignored for negative label anyway\n return loc_targets, class_targets\n\n match_neg, match_pos = match_thresholds\n\n assert match_pos >= match_neg\n\n ious = iou(point_form(anchor_boxes), boxes)\n max_ious, max_ids = ious.max(1)\n\n boxes = boxes[max_ids]\n\n boxes_pos, boxes_size = split(extents_form(boxes))\n anchor_pos, anchor_size = split(anchor_boxes)\n\n loc_pos = (boxes_pos - anchor_pos) / anchor_size\n loc_size = torch.log(boxes_size/anchor_size)\n loc_targets = torch.cat([loc_pos,loc_size], 1)\n\n class_targets = 1 + labels[max_ids]\n class_targets[max_ious <= match_neg] = 0 # negative label is 0\n\n ignore = (max_ious > match_neg) & (max_ious <= match_pos) # ignore ious between [0.4,0.5]\n class_targets[ignore] = -1 # mark ignored to -1\n\n return loc_targets, class_targets\n\n\n\ndef decode(loc_preds, class_preds, anchor_boxes):\n '''Decode (encoded) predictions and anchor boxes to give detected boxes.\n Args:\n loc_preds: (tensor) box predictions in encoded form, sized [n, 4].\n class_preds: (tensor) object class predictions, sized [n].\n anchor_boxes: (tensor) bounding boxes in extents form, sized [m, 4].\n Returns:\n boxes: (tensor) detected boxes in point form, sized [k, 4].\n labels: (tensor) detected class labels [k].\n '''\n\n #num_classes = class_preds.size(1)\n #confs = F.normalize(class_preds, dim=1).narrow(1, 1, num_classes)\n\n loc_pos, loc_size = split(loc_preds)\n anchor_pos, anchor_size = split(anchor_boxes)\n\n pos = loc_pos * anchor_size + anchor_pos\n sizes = loc_size.exp() * anchor_size\n\n boxes = point_form(torch.cat([pos, sizes], 1))\n confs, labels = class_preds.max(1)\n\n return (boxes, labels, confs)\n\nnms_defaults = {\n 'nms_threshold':0.5,\n 'class_threshold':0.05,\n 'max_detections':100\n}\n\ndef filter_preds(keep, boxes, labels, confs):\n if(keep.dim() > 0):\n return boxes[keep], labels[keep], confs[keep]\n else:\n return boxes.new(), labels.new(), confs.new()\n\ndef filter_nms(boxes, labels, confs, nms_threshold=0.5, class_threshold=0.05, max_detections=100):\n inds = nms(boxes, confs, nms_threshold=nms_threshold, class_threshold=class_threshold, \\\n max_detections=max_detections).type_as(labels)\n\n return filter_preds(inds, boxes, labels, confs)\n\n\ndef decode_nms(loc_preds, class_preds, anchor_boxes, nms_threshold=0.5, class_threshold=0.05, max_detections=100):\n assert loc_preds.dim() == 2 and class_preds.dim() == 2\n\n boxes, labels, confs = decode(loc_preds, class_preds, anchor_boxes)\n inds = nms(boxes, confs, nms_threshold=nms_threshold, class_threshold=class_threshold, \\\n max_detections=max_detections).type_as(labels)\n\n return filter_preds(inds, boxes, labels, confs)\n","sub_path":"detection/box.py","file_name":"box.py","file_ext":"py","file_size_in_byte":9506,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"552035042","text":"'''The definition of temperature scales is related to the linear expansion of certain liquids, such as mercury and alcohol. Originally, these scales were literally rulers for measuring length of fluid in the narrow marked or \"graduated\" tube as a proxy for temperature. The alcohol starts in a bulb, and then expands linearly into the tube, in response to increasing temperature of the bulb or whatever surrounds it.\n\nIn this exercise, we will explore the conversion between the Fahrenheit and Celsius temperature scales as a demonstration of interpreting slope and intercept of a linear relationship within a physical context.'''\n#TASK\n# Complete the function temps_F = convert_scale(temps_C) as a linear model where \"x\" is temps_C and \"y\" is temps_F.\n# Compute the intercept zero_offset as the difference between the freezing points freeze_F and freeze_C\n# Compute the slope as a unit_ratio, change in temps_F divided by change in temps_C\n# Use the predefined plot_temperatures() to plot the resulting model.\n\n# Complete the function to convert C to F\ndef convert_scale(temps_C):\n (freeze_C, boil_C) = (0, 100)\n (freeze_F, boil_F) = (32, 212)\n change_in_C = ____ - freeze_C\n change_in_F = ____ - freeze_F\n slope = ____ / ____\n intercept = ____ - freeze_C\n temps_F = ____ + (____ * temps_C)\n return temps_F\n\n# Use the convert function to compute values of F and plot them\ntemps_C = np.linspace(0, 100, 101)\ntemps_F = convert_scale(temps_C)\nfig = plot_temperatures(temps_C, temps_F)\n\n\n\n\n\n#SOLUTION\n# Complete the function to convert C to F\ndef convert_scale(temps_C):\n (freeze_C, boil_C) = (0, 100)\n (freeze_F, boil_F) = (32, 212)\n change_in_C = boil_C - freeze_C\n change_in_F = boil_F - freeze_F\n slope = change_in_F / change_in_C\n intercept = freeze_F - freeze_C\n temps_F = intercept + (slope * temps_C)\n return temps_F\n\n# Use the convert function to compute values of F and plot them\ntemps_C = np.linspace(0, 100, 101)\ntemps_F = convert_scale(temps_C)\nfig = plot_temperatures(temps_C, temps_F)","sub_path":"DataCamp_Introduction_to_Linear_Modeling_in_Python/2.2.1.Linear_Proportionality.py","file_name":"2.2.1.Linear_Proportionality.py","file_ext":"py","file_size_in_byte":2044,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"568032870","text":"import re\n\nurls = '(?: %s)' % '|'.join(\"\"\"http telnet gopher file wais\nftp\"\"\".split())\nltrs = r'\\w'\ngunk = r'/#~:.?+=&%@!\\-'\npunc = r'.:?\\-'\nany = \"%(ltrs)s%(gunk)s%(punc)s\" % { 'ltrs' : ltrs,\n 'gunk' : gunk,\n 'punc' : punc }\n\nurl = r\"\"\"\n \\b # start at word boundary\n %(urls)s : # need resource and a colon\n [%(any)s] +? # followed by one or more\n # of any valid character, but\n # be conservative and take only\n # what you need to....\n (?= # look-ahead non-consumptive assertion\n [%(punc)s]* # either 0 or more punctuation\n (?: [^%(any)s] # followed by a non-url char\n | # or end of the string\n $\n )\n )\n \"\"\" % {'urls' : urls,\n 'any' : any,\n 'punc' : punc }\n\nurl_re = re.compile(url, re.VERBOSE | re.MULTILINE)\nip_re = re.compile(r'\\b(25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\\.'+\n r'(25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\\.'+\n r'(25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\\.'+\n r'(25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\\b')\naddr_re = re.compile(r'[a-zA-Z0-9+_\\-\\.]+@[0-9a-zA-Z][.-0-9a-zA-Z]*.[a-zA-Z]+',\n re.IGNORECASE)\naddr_re = re.compile(r\"\"\"[a-z0-9!#$%&*+/=?^_`{|}~-]+(?:\\.[a-z0-9!#$%&'*+/=?^_`{|}~-]+)*@(?:[a-z0-9](?:[a-z0-9-]*[a-z0-9])?\\.)+(?:[A-Z]{2}|com|org|net|edu|gov|mil|biz|info|mobi|name|aero|asia|jobs|museum)\\b\"\"\",re.IGNORECASE)\ndomain_re = re.compile(r\"\"\"(?:[a-z0-9](?:[a-z0-9-]*[a-z0-9])?\\.)+(?:[A-Z]{2}|com|org|net|edu|gov|mil|biz|info|mobi|name|aero|asia|jobs|museum)\\b\"\"\",\n re.IGNORECASE)\n\ndef getUrls(text):\n \"\"\"Given a text string, returns all the urls we can find in it.\"\"\"\n if text is None: return []\n return url_re.findall(text)\n\ndef getIps(text):\n \"\"\"Given a text string, returns all the IP addresses we can find\n in it.\"\"\"\n if text is None: return []\n ips = ip_re.findall(text)\n if ips:\n for i,ip in enumerate(ips):\n ips[i] = '.'.join(ip)\n return ips\n\ndef getEmailAddrs(text):\n \"\"\"Given a text string, returns all the email addresses we can\n find in it.\"\"\"\n if text is None: return []\n addrs = []\n for a in addr_re.findall(text):\n if a[0] == \"'\": a = a[1:]\n addrs.append(a)\n return addrs\n\ndef getDomains(text):\n \"\"\"Given a text string, returns all the domains we can find in it.\"\"\"\n if text is None: return []\n return domain_re.findall(text)\n","sub_path":"spamRe.py","file_name":"spamRe.py","file_ext":"py","file_size_in_byte":2744,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"109846444","text":"import os\nimport pathlib\nimport random\n\nfrom PIL import Image\n\nfrom config import TARGET_SIZE\n\n\ndef preprocessing(p_image, outdir=None, rename=False):\n name = pathlib.Path(p_image).stem\n image = Image.open(p_image)\n\n images, names = modification(image, name)\n\n if outdir:\n for image, name in zip(images, names):\n ext = list(\"abcdefghijklmnopqrstuvwxyz\")\n name = name+'_{}'.format(random.choice(ext)) if rename else name\n image.save(os.path.join(outdir, name+'.jpg'))\n\n return images, names\n\n\ndef modification(image, name):\n \"\"\" 画像加工用の関数\n 例えばリサイズとか一枚の画像を4分割するとか、ぼかしをかけるとか\n ブートストラップで画像水増しとか,そういう処理を記述する\n デフォルトでは推論時もこの関数が呼ばれるので、とくに\n 画像水増しなんかするなら整合性に注意\n \"\"\"\n images = [image]\n names = [name]\n\n images = [image.resize(TARGET_SIZE) for image in images]\n names = [name for name in names]\n\n return images, names\n\n\ndef postprocessing(image, result):\n return image, result\n","sub_path":"absoluteCLF/src/processing.py","file_name":"processing.py","file_ext":"py","file_size_in_byte":1192,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"253445491","text":"import unittest\nfrom flask import Flask\nfrom flask.ext.testing import TestCase\nfrom sqlalchemy.exc import IntegrityError\nfrom app import app\nfrom database import db\nfrom models import User\n\nclass UserTest(TestCase):\n \n def create_dummy_user(self):\n user = User(name=\"Test User\", facebook_id=123456, gender=\"male\", username=\"testuser\", locale=\"pt_BR\")\n db.session.add(user)\n db.session.commit()\n return user\n\n def create_app(self):\n app = Flask(__name__)\n app.config['TESTING'] = True\n app.config['SQLALCHEMY_DATABASE_URI'] = 'sqlite:///db/app_test.db'\n db.init_app(app)\n return app\n\n def setUp(self):\n db.create_all()\n\n def tearDown(self):\n db.session.remove()\n db.drop_all()\n\n def test_should_create_a_new_user(self):\n \"\"\" It should be possible to create a new user in the database \"\"\"\n user = self.create_dummy_user()\n first_user = User.query.filter(User.facebook_id=='123456').first()\n assert first_user is not None\n\n def test_user_is_invalid(self):\n \"\"\" User should be invalid when facebook_id is missing\"\"\"\n user = User(name=\"Test User\", gender=\"male\", username=\"testuser\", locale=\"pt_BR\")\n self.assertFalse(user.is_valid())\n\n def test_user_should_be_unique(self):\n \"\"\" It should raise an error when a user with facebook_id already exists \"\"\"\n user = self.create_dummy_user()\n self.assertRaises(IntegrityError,self.create_dummy_user)\n\n def test_user_should_be_deleted(self):\n \"\"\" It should be possible to delete an user from the database \"\"\"\n user = self.create_dummy_user()\n first_user = User.query.filter(User.facebook_id=='123456').first()\n db.session.delete(first_user)\n db.session.commit()\n first_user = User.query.filter(User.facebook_id=='123456').first()\n assert first_user is None\n\nif __name__ == '__main__':\n unittest.main()","sub_path":"user_test.py","file_name":"user_test.py","file_ext":"py","file_size_in_byte":1836,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"234267735","text":"import subprocess\nimport platform\n\nplat=platform.platform()\nif 'Ubuntu' in plat:\n f = subprocess.Popen(['tail','-F','-n','1000','/var/log/syslog'],\\\n stdout=subprocess.PIPE,stderr=subprocess.PIPE)\nelse:\n f = subprocess.Popen(['tail','-F','-n','1000','/var/log/messages'],\\\n stdout=subprocess.PIPE,stderr=subprocess.PIPE)\n\nwhile True:\n line = f.stdout.readline()\n if \"NVRM:\" in str(line):\n file = open('/var/log/gpuerrors.log','a')\n print(line.decode('utf8', errors='strict').strip())\n file.write(line.decode('utf8', errors='strict').strip())\n file.write('\\n')\n file.close()\n","sub_path":"nvidia-efa-ami_base/cloudwatch/nvidia/aws-hwaccel-error-parser.py","file_name":"aws-hwaccel-error-parser.py","file_ext":"py","file_size_in_byte":630,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"111597392","text":"import os\nimport pandas as pd\nimport numpy as np\nimport gensim\nfrom sklearn.neighbors import NearestNeighbors\nimport networkx as nx\nimport regex as re\nfrom unidecode import unidecode\nfrom collections import Counter\nfrom scipy import sparse\nfrom sparsesvd import sparsesvd\n\nfrom regcomments.loaders import listCommentTextDockets, iterDocketCommentText\n\nfrom regcomments.dirs import *\n\n\n\ndocketsDF = pd.DataFrame(listCommentTextDockets(),columns=['docketId'])\n\n\n\nclass LsaModel():\n def __init__(self,tokenized,no_below=0,no_above=1,k=200):\n\n self.dictionary = gensim.corpora.Dictionary(tokenized)\n self.dictionary.filter_extremes(no_below=no_below,no_above=no_above)\n\n self.logEnt = gensim.models.LogEntropyModel([self.dictionary.doc2bow(tokens) for tokens in tokenized])\n\n self.lsaModel = gensim.models.LsiModel([self.logEnt[self.dictionary.doc2bow(tokens)] for tokens in tokenized],id2word=self.dictionary,num_topics=k)\n\n\n def vectorize(self,tokenized,index=None,normalize=True):\n V = gensim.matutils.corpus2dense([self.lsaModel[self.logEnt[self.dictionary.doc2bow(tokens)]] for tokens in tokenized], len(self.lsaModel.projection.s)).T\n V = V / self.lsaModel.projection.s\n\n if normalize:\n V = V / np.sqrt((V**2).sum(axis=1))[:,np.newaxis]\n\n if index is None:\n return V\n else:\n return pd.DataFrame(V,index=index)\n\n\n\ndef iterParagraphs(text):\n start = 0\n for m in re.finditer(r'(?= min_len:\n yield line\n\n# list(iterLines('''\n# Here is some text\n# with multiple lines\n#\n# and some big spaces\n#\n# '''))\n\ndef lineHash(line):\n line = re.sub(r'[^a-z0-9]','',line.lower())\n return line\n\ndef linesJaccard(text0,text1,min_len=20):\n lines0 = set(lineHash(line) for line in iterLines(text0,min_len))\n lines1 = set(lineHash(line) for line in iterLines(text1,min_len))\n\n denominator = len(lines0 | lines1)\n\n if denominator:\n return len(lines0 & lines1) / denominator\n else:\n return 0\n\n\n\n\n# Fit lsa vectors on paragraph sample ------------------------------------------\nprint('Building paragraph sample')\nsampleDF = pd.DataFrame()\nfor docketId in docketsDF['docketId'].sample(10000):\n documentsDF = pd.DataFrame(iterDocketCommentText(docketId),columns=['documentId','attachment_number','text'])\n if len(documentsDF):\n documentsDF['text'] = documentsDF['text'].apply(lambda b: unidecode(b.decode('utf8')))\n paragraphsDF = pd.DataFrame([p for text in documentsDF['text'] for p in iterParagraphs(text)],columns=['text'])\n\n paragraphsDF = paragraphsDF.drop_duplicates()\n\n sampleDF = sampleDF.append(paragraphsDF.sample(frac=0.01))\n\nif len(sampleDF) > 100000:\n sampleDF = sampleDF.sample(100000)\nprint('Sample size: {} paragraphs'.format(len(sampleDF)))\n\nprint('Fitting LSA model')\nmodel = LsaModel(sampleDF['text'].apply(lambda s: list(iterTokens(s))),no_below=5)\n\ndel sampleDF\n\n\n\n# Identify duplicate documents -------------------------------------------------\nprint('Identifying duplicates')\nn_neighbors = 5\nmin_cosine = 0.95\nmin_jaccard = 0.5\n\ngroupsDF = pd.DataFrame()\nfor i,docketId in enumerate(docketsDF['docketId']):\n print('{}/{} {}'.format(i+1,len(docketsDF),docketId))\n\n documentsDF = pd.DataFrame(iterDocketCommentText(docketId),columns=['documentId','attachment_number','text'])\n documentsDF['docketId'] = docketId\n\n if len(documentsDF) > 1:\n documentsDF['text'] = documentsDF['text'].str.decode('utf8')\n\n uniqueDF = documentsDF[['text']].drop_duplicates().reset_index(drop=True)\n\n V = model.vectorize(uniqueDF['text'].apply(lambda s: list(iterTokens(s))))\n V = V / np.sqrt((V**2).sum(axis=1))[:,np.newaxis]\n\n nearestNeighbors = NearestNeighbors(n_neighbors=min(len(uniqueDF),n_neighbors))\n nearestNeighbors.fit(V)\n\n distances,matches = nearestNeighbors.kneighbors(V)\n\n matchPairs = np.vstack([np.kron(np.arange(matches.shape[0]),np.ones(matches.shape[1]).astype(int)),matches.ravel()]).T\n\n matchPairs = np.sort(np.array(matchPairs),axis=1)\n\n matchDF = pd.DataFrame(matchPairs,columns=['doc_i','doc_j'])\n matchDF['distance'] = distances.ravel()\n matchDF['cosine'] = (V[matchPairs[:,0]] * V[matchPairs[:,1]]).sum(axis=1)\n\n matchDF = matchDF.drop_duplicates(['doc_i','doc_j'])\n\n matchDF = matchDF[matchDF['doc_i'] != matchDF['doc_j']]\n\n matchDF = matchDF[matchDF['cosine'] > min_cosine]\n\n for c in 'i','j':\n matchDF['text_'+c] = uniqueDF.loc[matchDF['doc_'+c],'text'].values\n\n matchDF['line_jaccard'] = [linesJaccard(text_i,text_j) for text_i,text_j in matchDF[['text_i','text_j']].itertuples(index=False)]\n\n matchDF = matchDF[matchDF['line_jaccard'] > min_jaccard]\n\n # matchDF = matchDF.sort_values('cosine',ascending=False)\n\n if len(matchDF) or (len(uniqueDF) < len(documentsDF)):\n\n G = nx.Graph()\n G.add_nodes_from(documentsDF.index.get_level_values(0))\n for i,text in documentsDF[['text']].itertuples():\n G.add_edge(i,text)\n\n for i,j in matchDF[['doc_i','doc_j']].itertuples(index=False):\n G.add_edge(i,j)\n\n componentMap = {i:c for c,component in enumerate(nx.connected_components(G)) for i in component}\n documentsDF['group'] = [componentMap[i] for i in documentsDF.index.get_level_values(0)]\n documentsDF['group_size'] = documentsDF.groupby('group')['documentId'].transform('count')\n\n uniqueDF['v_0'] = V[:,0]\n\n documentsDF = pd.merge(documentsDF,uniqueDF,'left')\n\n documentsDF = documentsDF.sort_values(['group_size','group','v_0'],ascending=[False,True,False])\n\n # Add groups information\n groupsDF = groupsDF.append(documentsDF.drop(['text','v_0'],axis=1),sort=False)\n\n continue\n\n\n # Cover the case where there are no duplicates found\n documentsDF['group'] = list(range(len(documentsDF)))\n documentsDF['group_size'] = 1\n\n groupsDF = groupsDF.append(documentsDF.drop('text',axis=1),sort=False)\n\n\ngroupsDF.to_csv(os.path.join(dataDir,'comments','duplicateDocuments.csv'))\n\n\n\n# Test code\n# documentsDF[documentsDF['group_size']>1].sort_values(['group_size','group'],ascending=[False,True])\n#\n# matchDF.describe()\n#\n#\n# examples = documentsDF[documentsDF['group']==10]['text']\n#\n# examples = uniqueDF.loc[(20,23),'text']\n#\n# print('\\n----------------------------------------------\\n'.join(examples))\n","sub_path":"dataProcessing/comments/identifyDuplicates_LSA.py","file_name":"identifyDuplicates_LSA.py","file_ext":"py","file_size_in_byte":7383,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"547692474","text":"# Assign spell power lists to variables\ngandalf = [10, 11, 13, 30, 22, 11, 10, 33, 22, 22]\nsaruman = [23, 66, 12, 43, 12, 10, 44, 23, 12, 17]\n\n# Assign 0 to each variable that stores the victories\nscore_gandalf = 0\nscore_saruman = 0\ntotal_matches = 0\n# Execution of spell clashes\nspell_options = range(len(gandalf)) # creates a range based on the total length of the gandalf list\nfor spell_power in spell_options: # for loop to start the clash\n if gandalf[spell_power] > saruman[spell_power]: # if gandalfs power is bigger than saurman, make gandalf the winner\n score_gandalf += 1 # add score to gandalfs total wins\n total_matches += 1 # counter for total matches\n print(\"Gandalf has won clash: {0}\\n- Gandalf's power: {1}\\n- Saruman's power: {2} \\n\".format(total_matches, gandalf[spell_power], saruman[spell_power])) # match status\n\n elif saruman[spell_power] > gandalf[spell_power]:# if saurman power is bigger than gandalf, make saurman the winner\n score_saruman += 1 # add score to saurman total wins\n total_matches += 1 # counter for total matches\n print(\"Saruman has won clash: {0}\\n- Gandalf's power: {1}\\n- Saruman's power: {2} \\n\".format(total_matches, gandalf[spell_power], saruman[spell_power])) # match status\n\n else:# otherwise its a draw\n total_matches += 1 # counter for total matches\n print(\"Clash was a draw!\") # match status\n \n# We check who has won, do not forget the possibility of a draw.\n# Print the result based on the winner.\nif score_gandalf > score_saruman: # assign the winner variable to who one based on total score ( Gandalf has more score here)\n winner = (\"Gandalf\")\nelif score_gandalf < score_saruman: # assign the winner variable to who one based on total score ( Saruman has more score here)\n winner = (\"Saruman\")\nelse: # otherwise print its a draw\n print(\"NO ONE!!! It is a draw :)\")\n\n\nprint(\"Total clashes: {0}\\nGandalf's Wins: {1}\\nSaurman's Wins: {2}\\nVictor is: {3}\".format(total_matches, score_gandalf, score_saruman, winner))# FINAL MATCH STATUS\n\n#3/22/19\n","sub_path":"duel/Raw Python/duel-of-sorcerers-solved-NoahBeckerman.py","file_name":"duel-of-sorcerers-solved-NoahBeckerman.py","file_ext":"py","file_size_in_byte":2032,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"550292564","text":"class Solution(object):\n def numberOfBoomerangs(self, points):\n\n \"\"\"\n\n :type points: List[List[int]]\n :rtype: int\n \"\"\"\n res = 0\n # 以每一个点为中心, 进行尝试, 看看有没有距离相同的点出现, 如果放在map中, 然后对map进行统计操作,\n # 在进行下一个点的时候 要将之前的map清空\n for i in range(len(points)):\n # 用一个字典, 记录下与某点相同距离的次数\n d = {}\n for j in range(len(points)):\n if i == j:\n continue\n dis = (points[i][0] - points[j][0]) ** 2 + (\n points[i][1] - points[j][1]) ** 2\n d[dis] = d.get(dis, 0) + 1\n for item in d.values():\n if item >= 2:\n res += item * (item - 1)\n d.clear()\n return res\n\n\ndef main():\n solution = Solution()\n ret = solution.numberOfBoomerangs([1, 2, 3, 1])\n print(ret)\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"simple/447_number_of_boomerangs.py","file_name":"447_number_of_boomerangs.py","file_ext":"py","file_size_in_byte":1060,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"299891228","text":"'''\nPlot the analytical derivative of a function, and compare it to its numerical derivative\nf(t) = exp(-2t) * sin(10t - 6)\n\nf'(t) = df /dt = -2 * exp(-2t) * sin(10t - 6) + 10 * exp(-2t) * cos(10t - 6)\n\ng(t) = [ f(t+dt) - f(t)] / dt\n'''\n\nimport os\nimport numpy as np\nimport sympy as sy\nimport matplotlib.pyplot as plt\nfrom matplotlib.pyplot import figure\n\nfigure(figsize=(10, 6), dpi=80)\n\n# Define the analytical functions:\nf_eq = lambda t: np.exp(-2*t) * np.sin(10*t - 6)\ndf_tru_eq = lambda t: -2 * np.exp(-2*t) * np.sin(10*t - 6) + 10 * np.exp(-2*t) * np.cos(10*t - 6)\n\n# Define the tru_time array, to plot the true derivative\nt_fin = 3\ntru_time = np.arange(0, t_fin, 0.01)\n\n# Evaluate the true derivative\ndf_tru_data = df_tru_eq(tru_time)\n\n# Generate array from 0 to 4, with step of dt\ndt = 0.1\ndsc_time = np.arange(0, t_fin, dt)\n\n# Calculate the approximate derivative\nf_data = f_eq(dsc_time)\ndf_apx_data = np.diff(f_data) / dt # np.diff(a) calculates the difference between one element and the next, within array a \n\n# Calculate the symbolic derivative and evaluate it over time_apx\nt = sy.symbols('t', real = True)\nf_sym = sy.exp(-2*t) * sy.sin(10*t-6) # Defines the symbolic equation\ndf_sym = sy.lambdify(t,sy.diff(f_sym,t),\"numpy\") # Finds the symbolic derivative and converts it to a lambda function\ndf_sym_data = df_sym(dsc_time)\n\nplt.plot(tru_time,df_tru_data)\nplt.scatter(dsc_time[:-1],df_apx_data, facecolors='none', edgecolors='r') # df_apx has n-1 elements, so we must also reduce dsc_time's length\nplt.plot(dsc_time, df_sym_data, 'yo', markersize=5)\nplt.xlabel('t')\nplt.ylabel('df/dt')\nplt.legend(['Analytical derivative, over continuous time',\\\n 'Numerical derivative, over discrete time',\\\n 'Symbolic derivative, over discrete time'])\n\nsave_path = os.path.dirname(os.path.abspath(__file__)) + '/img/'\n\nif not os.path.isdir(save_path):\n os.mkdir(save_path)\n\nfigure_name = 'fig_derivatives_comparison.png'\nplt.savefig(save_path + figure_name)\nplt.show()\n","sub_path":"03 Analytical vs Numerical derivatives/analytical_vs_numerical.py","file_name":"analytical_vs_numerical.py","file_ext":"py","file_size_in_byte":2019,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"68076830","text":"#!/usr/bin/env python\n\nimport argparse\nimport math\nimport os\nimport sys\n\nfrom lnd import Lnd\n\nfrom logic import Logic\n\nMAX_CHANNEL_CAPACITY = 16777215\nMAX_SATOSHIS_PER_TRANSACTION = 4294967\n\n\ndef main():\n parser = argparse.ArgumentParser()\n\n parser.add_argument(\"-l\", \"--listcandidates\", action=\"store_true\", default=False)\n parser.add_argument(\"channel\", help=(\"channel identifier, can be either the channel index as given by -l\"\n \" or the channel's pubkey\"), nargs=\"?\")\n # args.amount is essentially a list, and what matters to us is the first value it *may* have\n parser.add_argument(\"amount\", help=(\"amount of the rebalance, in satoshis. If not specified, the amount computed\"\n \" for a perfect rebalance will be used\"), nargs='?')\n args = parser.parse_args()\n\n if args.listcandidates or args.channel is None:\n list_candidates()\n sys.exit()\n\n # first we deal with the first argument, channel, to figure out what it means\n if args.channel and len(args.channel) < 4:\n # here we are in the \"channel index\" case\n index = int(args.channel) - 1\n candidates = get_rebalance_candidates()\n candidate = candidates[index]\n remote_pubkey = candidate.remote_pubkey\n else:\n # else the channel argument should be the node's pubkey\n remote_pubkey = args.channel\n # candidate is a channel -- we find it by filtering through all candidates\n candidate = [c for c in get_rebalance_candidates() if c.remote_pubkey == remote_pubkey][0]\n\n # then we figure out whether an amount was specified or if we compute it ourselves\n if args.amount:\n amount = int(args.amount)\n else:\n amount = int(math.ceil(float(get_remote_surplus(candidate)) / 2))\n if amount > MAX_SATOSHIS_PER_TRANSACTION:\n amount = MAX_SATOSHIS_PER_TRANSACTION\n\n response = Logic(lnd, remote_pubkey, amount).rebalance()\n if response:\n print(response)\n\n\ndef list_candidates():\n index = 0\n candidates = get_rebalance_candidates()\n for candidate in candidates:\n index += 1\n rebalance_amount = int(math.ceil(float(get_remote_surplus(candidate)) / 2))\n if rebalance_amount > MAX_SATOSHIS_PER_TRANSACTION:\n rebalance_amount = str(rebalance_amount) + \" (max per transaction: %d)\" % MAX_SATOSHIS_PER_TRANSACTION\n\n print(\"(%2d) Pubkey: \" % index + candidate.remote_pubkey)\n print(\"Local ratio: \" + str(get_local_ratio(candidate)))\n print(\"Capacity: \" + str(candidate.capacity))\n print(\"Remote balance: \" + str(candidate.remote_balance))\n print(\"Local balance: \" + str(candidate.local_balance))\n print(\"Amount for 50-50: \" + str(rebalance_amount))\n print(get_capacity_and_ratio_bar(candidate))\n print(\"\")\n print(\"\")\n print(\"Run with two arguments: 1) pubkey of channel to fill 2) amount\")\n\n\ndef get_rebalance_candidates():\n low_local = list(filter(lambda c: get_local_ratio(c) < 0.5, lnd.get_channels()))\n return sorted(low_local, key=get_remote_surplus, reverse=False)\n\n\ndef get_local_ratio(channel):\n remote = channel.remote_balance\n local = channel.local_balance\n return float(local) / (remote + local)\n\n\ndef get_remote_surplus(channel):\n return channel.remote_balance - channel.local_balance\n\n\ndef get_capacity_and_ratio_bar(candidate):\n columns = get_columns()\n columns_scaled_to_capacity = int(round(columns * float(candidate.capacity) / MAX_CHANNEL_CAPACITY))\n\n bar_width = columns_scaled_to_capacity - 2\n result = \"|\"\n ratio = get_local_ratio(candidate)\n length = int(round(ratio * bar_width))\n for x in range(0, length):\n result += \"=\"\n for x in range(length, bar_width):\n result += \" \"\n return result + \"|\"\n\n\ndef get_columns():\n return int(os.popen('stty size', 'r').read().split()[1])\n\n\nlnd = Lnd()\nmain()\n","sub_path":"rebalance.py","file_name":"rebalance.py","file_ext":"py","file_size_in_byte":3976,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"318749666","text":"#!/usr/bin/env python3\n# -*- coding:utf-8 -*-\n# author: bigfoolliu\n\n\n\"\"\"\npython类型注解\n\n- https://blog.csdn.net/Skr_Eric/article/details/83063971?utm_medium=distribute.pc_relevant.none-task-blog-title-2&spm=1001.2101.3001.4242\n\n- 用 : 类型 的形式指定函数的参数类型,用 -> 类型 的形式指定函数的返回值类型。\n- 然后特别要强调的是,Python 解释器并不会因为这些注解而提供额外的校验,没有任何的类型检查工作。也就是说,这\n- 些类型注解加不加,对你的代码来说没有任何影响:\n\n但这么做的好处是:\n\n- 让别的程序员看得更明白\n- 让 IDE 了解类型,从而提供更准确的代码提示、补全和语法检查(包括类型检查,可以看到 str 和 float 类型的参数被高亮提示)\n\n使用特定的库对文件进行类型检查\npip3 install mypy\nmypy test.py\n\"\"\"\n\n\ndef add(x: int = 1, y: int = 2) -> int:\n \"\"\"类型注解\"\"\"\n return x + y\n\n\ndef add2(x, y):\n return x + y\n\n\ndef add3():\n # 变量的类型注解释\n a: int = 1\n l: list = [1, 'a']\n\n from typing import List\n li: List[int] = [1, 3] # 指明一个全部由整数构成的列表\n print(a, l, li)\n\n\nif __name__ == '__main__':\n print(add(2, 4))\n print(add('hello', 'world')) # ide提示类型问题\n print(add2(2, 4)) # 不会有任何提示\n\n print(add.__annotations__) # 可以查看注解\n add3()\n","sub_path":"language/python/python/exception/doc_type_demo.py","file_name":"doc_type_demo.py","file_ext":"py","file_size_in_byte":1433,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"232097793","text":"from kivy.app import App\nfrom kivy.lang import Builder\n\nimport sys\nimport os\nPACKAGE_PARENT = '..'\nSCRIPT_DIR = os.path.dirname(os.path.realpath(os.path.join(os.getcwd(), os.path.expanduser(__file__))))\nsys.path.append(os.path.normpath(os.path.join(SCRIPT_DIR, PACKAGE_PARENT)))\n\nfrom kivy_modules.widget.label import FlexLabel\n\nclass MainApp(App):\n title = 'Test Label'\n def build(self):\n kv = Builder.load_string('''\nBoxLayout:\n orientation: 'vertical'\n ScrollView:\n BoxLayout:\n id: container\n orientation: 'vertical'\n size_hint: 1, None\n height: self.minimum_height\n FlexLabel:\n text: 'FlexLabel'\n bg_color: [.4, .4,.4, 1]\n size_hint: 1, None\n height: dp(50)\n''')\n config = {\n 'size_hint': [1, None],\n 'height': 50\n }\n flexlabel = FlexLabel(text = 'flexlabel dinamic', **config)\n kv.ids.container.add_widget(flexlabel)\n return kv\n\nMainApp().run()","sub_path":"modules/show_case/test/test_label.py","file_name":"test_label.py","file_ext":"py","file_size_in_byte":1047,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"140393346","text":"\n# coding: utf-8\n\n# In[1]:\n\n\nimport pandas as pd\nimport numpy as np\nimport scipy as sp\nimport matplotlib.pyplot as plt\nimport seaborn as sns\nimport time\nimport pickle\n\nfrom sklearn.utils import resample\nfrom sklearn.neighbors import KNeighborsClassifier, KNeighborsRegressor\nfrom sklearn.linear_model import LogisticRegression, LinearRegression, RidgeCV, LassoCV\nfrom sklearn.discriminant_analysis import LinearDiscriminantAnalysis, QuadraticDiscriminantAnalysis\nfrom sklearn.tree import DecisionTreeClassifier\nfrom sklearn.ensemble import RandomForestClassifier, AdaBoostClassifier, GradientBoostingClassifier\nimport statsmodels.api as sm\nfrom statsmodels.api import OLS\n\nfrom sklearn.model_selection import cross_val_score, train_test_split, KFold, GridSearchCV\nfrom sklearn.metrics import accuracy_score, confusion_matrix, mean_squared_error, r2_score, log_loss\nfrom sklearn.decomposition import PCA\nfrom sklearn.preprocessing import Imputer, StandardScaler, PolynomialFeatures\n\n\n# In[2]:\n\n\ndef scale_numeric(fit_df, trans_df, cols):\n df_scaled = trans_df[cols].copy()\n scaler = StandardScaler().fit(fit_df[cols])\n df_scaled = pd.DataFrame(scaler.transform(df_scaled), index=df_scaled.index, columns=cols)\n return(df_scaled)\n\ndef scale_encode(fit_df, trans_df):\n num_cols = list(fit_df.columns[(fit_df.dtypes == 'float64') | (fit_df.dtypes == 'int64')])\n str_cols = list(set(fit_df.columns)^set(num_cols))\n \n if (len(num_cols)>0) & (len(str_cols)>0):\n df_scaled = scale_numeric(fit_df=fit_df, trans_df=trans_df, cols=num_cols)\n df_dummy = pd.get_dummies(trans_df[str_cols])\n df = df_scaled.join(df_dummy)\n elif len(num_cols)==0:\n df = pd.get_dummies(trans_df[str_cols])\n elif len(str_cols)==0:\n df = scale_numeric(fit_df=fit_df, trans_df=trans_df, cols=num_cols) \n return(df)\n\n\n# In[3]:\n\n\ndef set_outliers_to_null(df):\n \n num_cols = list(df.columns[(df.dtypes == 'float64') | (df.dtypes == 'int64')])\n str_cols = list(set(df.columns)^set(num_cols))\n\n for col in list(df.columns):\n if col in num_cols:\n cond1 = (df[col]<0)\n cond2 = (df[col].isna())\n cond3 = (df[col]>=(np.mean(df[col])+2*np.std(df[col]))) \n cond4 = (df[col]<=(np.mean(df[col])-2*np.std(df[col])))\n df[col][cond1 | cond2 | cond3 | cond4] = None\n\n elif col in str_cols:\n cond1 = (df[col].isna())\n cond2 = (df[col]=='None')\n df[col][cond1 | cond2] = None\n \n return(df)\n\ndef impute_mean(df, na_cols):\n for i in na_cols:\n if i in num_cols:\n df.loc[df[i].isnull(), i] = df[i].dropna().mean()\n elif i in str_cols:\n df.loc[df[i].isnull(), i] = df[i].dropna().mode() \n return(df)\n\n\ndef handle_nas(df_impute, method='impute_mean'):\n \n df = set_outliers_to_null(df_impute)\n\n df = df.drop(columns=list(df.isnull().sum()[np.sum(df.isnull())>0.4*df.shape[0]].index))\n\n na_cols = list(df.isnull().sum().sort_values(ascending=True)[np.sum(df.isnull())>0].index)\n num_cols = list(df.columns[(df.dtypes == 'float64') | (df.dtypes == 'int64')])\n str_cols = list(set(df.columns)^set(num_cols))\n predictors = list(set(na_cols)^set(df.columns))\n \n print('Imputing data for variables: %s \\n Total number of features for imputing: %s \\n Total Columns: %s'\n %(na_cols, len(na_cols), len(list(df.columns))))\n \n # if all cols have na's impute, one with least na's by mean\n if len(predictors) == 0:\n df = impute_mean(df=df, na_cols=na_cols[0])\n na_cols = list(df.isnull().sum().sort_values(ascending=True)[np.sum(df.isnull())>0].index)\n predictors = list(set(na_cols)^set(df.columns))\n\n if method == 'impute_mean':\n df = impute_mean(df=df, na_cols=na_cols)\n elif method == 'impute_predict': \n \n for i in na_cols:\n t0 = time.time()\n print('loop started for %s...'%i)\n X = df[predictors]\n X_scaled = scale_encode(X, X)\n y = df.loc[~df[i].isnull(), i]\n X_train = X_scaled.loc[~df[i].isnull()]\n X_pred = X_scaled.loc[df[i].isnull()]\n print(' scaling done...')\n if i in num_cols:\n model = KNeighborsRegressor().fit(X_train, y)\n print(' model fit...')\n df.loc[df[i].isnull(), i] = model.predict(X_pred)\n elif i in str_cols:\n model = KNeighborsClassifier().fit(X_train, y)\n print(' model fit...')\n df.loc[df[i].isnull(), i] = model.predict(X_pred)\n predictors.append(i)\n print(' %s done...'%i)\n t1 = time.time()\n T = t1-t0\n print(' Time taken: %s'%T) \n \n return(df)\n\n","sub_path":"4. useful_code/utility_functions.py","file_name":"utility_functions.py","file_ext":"py","file_size_in_byte":4842,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"178465225","text":"import subprocess\nimport datetime\nimport time\n\nuczeniowie = [{'imieNazwisko': 'Robert Mak', 'klasa': '1B'}, {'imieNazwisko': 'Ryszard Nowak', 'klasa': '2C'},\n {'imieNazwisko': 'Anna Mak', 'klasa': '3A'}, {'imieNazwisko': 'Monika Zdun', 'klasa': '1B'}]\noceny = [{'uczen': 'Robert Mak', 'klasa': '1B', 'przedmiot': 'matematyka',\n 'ocena': 5, 'nauczyciel': 'Anna Nosowska'},\n {'uczen': 'Robert Mak', 'klasa': '1B', 'przedmiot': 'matematyka',\n 'ocena': 3, 'nauczyciel': 'Anna Nosowska'},\n {'uczen': 'Anna Mak', 'klasa': '3A', 'przedmiot': 'geografia',\n 'ocena': 6, 'nauczyciel': 'Monika Zatorska'},\n {'uczen': 'Robert Mak', 'klasa': '1B', 'przedmiot': 'matematyka',\n 'ocena': 4, 'nauczyciel': 'Anna Nosowska'},\n {'uczen': 'Robert Mak', 'klasa': '1B', 'przedmiot': 'geografia',\n 'ocena': 1, 'nauczyciel': 'Anna Nosowska'},\n {'uczen': 'Robert Mak', 'klasa': '1B', 'przedmiot': 'matematyka', 'ocena': 2, 'nauczyciel': 'Anna Nosowska'}]\nnauczyciele = ['Maria Konopka', 'Janusz Miłosz', 'Zuzanna Nowak',\n 'Anna Nosowska', 'Henryk Kozłowski', 'Monika Zatorska']\nprzedmioty = ['matematyka', 'j.polski', 'geografia', 'biologia', 'fizyka']\n\ndef wypiszListe(lista):\n print('-----------------------------------------------')\n for i in range(len(lista)):\n print(str(i+1)+' - '+str(lista[i]))\n if len(lista) == 0:\n print(' Brak danych')\n print('-----------------------------------------------')\n\ndef wypiszListy():\n print('-----------------------------------------------')\n print(' UCZNIOWIE')\n wypiszListe(uczeniowie)\n print('-----------------------------------------------')\n print(' OCENY')\n wypiszListe(oceny)\n print('-----------------------------------------------')\n print(' NAUCZYCIELE')\n wypiszListe(nauczyciele)\n print('-----------------------------------------------')\n print(' PRZEDMIOTY')\n wypiszListe(przedmioty)\n\ndef dodajUcznia():\n koniec = False\n while koniec == False:\n imieNazwisko = str(input('Podaj imie i nazwisko ucznia: '))\n klasa = str(input('Klasa ucznia (1B,2A...3C): '))\n uczen = {'imieNazwisko': imieNazwisko, 'klasa': klasa}\n uczeniowie.append(uczen)\n print('Dodano ucznia ' +\n uczen['imieNazwisko']+' do klasy '+uczen['klasa'])\n takNie = str(input('Chcesz dodać kolejnego ucznia? Y/N: '))\n if takNie == 'Y' or takNie == 'y':\n koniec = False\n else:\n koniec = True\n wypiszListe(uczeniowie)\n\ndef modyfikujUcznia():\n print('-----------------------------------------------')\n print(' UCZNIOWIE')\n wypiszListe(uczeniowie)\n nr = int(input('Podaj numer ucznia: '))\n if nr > len(uczeniowie) or nr < 1:\n print('Lista obejmuje '+str(len(uczeniowie))+' uczniów.')\n print('Sprubój jeszcze raz.')\n else:\n takNie = str(input('Czy chodzi o ucznia: ' +\n uczeniowie[nr-1]['imieNazwisko']+' ('+uczeniowie[nr-1]['klasa']+')? Y/N: '))\n if takNie == 'Y' or takNie == 'y':\n imieNazwisko = str(input('Podaj imie i nazwisko ucznia: '))\n klasa = str(input('Klasa ucznia (1B,2A...3C): '))\n uczeniowie[nr-1]['imieNazwisko'] = imieNazwisko\n uczeniowie[nr-1]['klasa'] = klasa\n print('Zmodyfikowano ucznia nr '+str(nr)+') '+imieNazwisko+' ('+klasa+')')\n\ndef usunUcznia():\n print('-----------------------------------------------')\n print(' UCZNIOWIE')\n wypiszListe(uczeniowie)\n imieNazwisko = str(input('Podaj imie i nazwisko ucznia do usunięcia: '))\n klasa = str(input('Podaj klasę ucznia (1B,2A...3C): '))\n uczenDoUsuniecia = {'imieNazwisko': imieNazwisko, 'klasa': klasa}\n if uczenDoUsuniecia in uczeniowie:\n takNie = str(input('Czy chcesz usunąć ucznia: ' +\n imieNazwisko+' ('+klasa+') z listy uczniów? Y/N: '))\n if takNie == 'Y' or takNie == 'y':\n uczeniowie.remove(uczenDoUsuniecia)\n print('Usunięto ucznia z listy.')\n print('-----------------------------------------------')\n print(' UCZNIOWIE')\n wypiszListe(uczeniowie)\n else:\n print('Nie ma takiego ucznia na liście.')\n\ndef wypiszOcene(ocenaUcznia):\n print(str(ocenaUcznia['przedmiot'])+' '+str(ocenaUcznia['ocena']\n )+' wystawiona przez: '+str(ocenaUcznia['nauczyciel']))\n\ndef ocenyUcznia():\n print('-----------------------------------------------')\n print(' UCZNIOWIE')\n wypiszListe(uczeniowie)\n nrU = int(input('Podaj numer ucznia: '))\n if nrU > len(uczeniowie) or nrU < 1:\n print('Lista obejmuje '+str(len(uczeniowie))+' uczniów.')\n print('Sprubój jeszcze raz.')\n else:\n takNie = str(input('Oceny z wybranego przedmiotu? Y/N: '))\n if takNie == 'Y' or takNie == 'y':\n print('-----------------------------------------------')\n print(' PRZEDMIOTY')\n wypiszListe(przedmioty)\n nrP = 0\n while nrP == 0:\n nrP = int(input('Wybierz przedmiot: '))\n if nrP > len(przedmioty) or nrP < 1:\n nrP = 0\n print('-----------------------------------------------')\n print(' OCENY ucznia ' +\n uczeniowie[nrU-1]['imieNazwisko']+' ('+uczeniowie[nrU-1]['klasa']+')')\n print('-----------------------------------------------')\n for i in range(len(oceny)):\n if oceny[i]['uczen'] == uczeniowie[nrU-1]['imieNazwisko'] and oceny[i]['klasa'] == uczeniowie[nrU-1]['klasa'] and oceny[i]['przedmiot'] == przedmioty[nrP-1]:\n wypiszOcene(oceny[i])\n print('-----------------------------------------------')\n else:\n print('-----------------------------------------------')\n print(' OCENY ucznia ' +\n uczeniowie[nrU-1]['imieNazwisko']+' ('+uczeniowie[nrU-1]['klasa']+')')\n print('-----------------------------------------------')\n for i in range(len(oceny)):\n if oceny[i]['uczen'] == uczeniowie[nrU-1]['imieNazwisko'] and oceny[i]['klasa'] == uczeniowie[nrU-1]['klasa']:\n wypiszOcene(oceny[i])\n print('-----------------------------------------------')\n\ndef ocenyWybranegoUcznia(uczen):\n print('-----------------------------------------------')\n print(' OCENY ucznia '+uczen['imieNazwisko']+' ('+uczen['klasa']+')')\n print('-----------------------------------------------')\n for i in range(len(oceny)):\n if oceny[i]['uczen'] == uczen['imieNazwisko'] and oceny[i]['klasa'] == uczen['klasa']:\n wypiszOcene(oceny[i])\n print('-----------------------------------------------')\n\ndef dodajOcene():\n print('-----------------------------------------------')\n print(' UCZNIOWIE')\n wypiszListe(uczeniowie)\n nrU = int(input('Podaj numer ucznia: '))\n if nrU > len(uczeniowie) or nrU < 1:\n print('Lista obejmuje '+str(len(uczeniowie))+' uczniów.')\n print('Sprubój jeszcze raz.')\n else:\n print('-----------------------------------------------')\n print(' PRZEDMIOTY')\n wypiszListe(przedmioty)\n nrP = 0\n while nrP == 0:\n nrP = int(input('Wybierz przedmiot: '))\n if nrP > len(przedmioty) or nrP < 1:\n nrP = 0\n ocena = 0\n while ocena == 0:\n ocena = int(input('Podaj ocenę (1..6): '))\n if ocena < 1 or ocena > 6:\n ocena = 0\n print('-----------------------------------------------')\n print(' NAUCZYCIELE')\n wypiszListe(nauczyciele)\n nrN = 0\n while nrN == 0:\n nrN = int(input('Ocenę wystawił: '))\n if nrN > len(nauczyciele) or nrN < 1:\n nrN = 0\n imieNazwisko = uczeniowie[nrU-1]['imieNazwisko']\n klasa = uczeniowie[nrU-1]['klasa']\n wybranyUczen = {'imieNazwisko': imieNazwisko, 'klasa': klasa}\n ocenaNowa = {'uczen': imieNazwisko, 'klasa': klasa,\n 'przedmiot': przedmioty[nrP-1], 'ocena': ocena, 'nauczyciel': nauczyciele[nrN-1]}\n oceny.append(ocenaNowa)\n ocenyWybranegoUcznia(wybranyUczen)\n\ndef najlepszyUczen():\n print('-----------------------------------------------')\n print(' PRZEDMIOTY')\n wypiszListe(przedmioty)\n nrP = 0\n while nrP == 0:\n nrP = int(input('Wybierz przedmiot: '))\n if nrP > len(przedmioty) or nrP < 1:\n nrP = 0\n ocenyPrzedmiot = []\n for i in range(len(oceny)):\n if oceny[i]['przedmiot'] == przedmioty[nrP-1]:\n ocenyPrzedmiot.append(oceny[i])\n sredniaUcznia = []\n for i in range(len(uczeniowie)):\n ileOcen = 0\n sumaOcen = 0\n for j in range(len(ocenyPrzedmiot)):\n if ocenyPrzedmiot[j]['uczen'] == uczeniowie[i]['imieNazwisko'] and ocenyPrzedmiot[j]['klasa'] == uczeniowie[i]['klasa']:\n sumaOcen += ocenyPrzedmiot[j]['ocena']\n ileOcen += 1\n if ileOcen != 0:\n sredniaUcznia.append(\n {'uczen': uczeniowie[i]['imieNazwisko'], 'klasa': uczeniowie[i]['klasa'], 'srednia': sumaOcen/ileOcen})\n maxSrednia = {'uczen': '', 'klasa': '', 'srednia': 0}\n if len(sredniaUcznia) != 0:\n for i in range(len(sredniaUcznia)):\n if sredniaUcznia[i]['srednia'] > maxSrednia['srednia']:\n maxSrednia = sredniaUcznia[i]\n print('Najlepszym uczniem z przedmiotu '+przedmioty[nrP-1])\n print('jest '+maxSrednia['uczen']+' z '+maxSrednia['klasa'] +\n ' ze średnią '+str(maxSrednia['srednia']))\n else:\n print('Nie ma najlepdszego ucznia z tego przedmiotu.')\n\ndef najgorszyUczen():\n print('-----------------------------------------------')\n print(' PRZEDMIOTY')\n wypiszListe(przedmioty)\n nrP = 0\n while nrP == 0:\n nrP = int(input('Wybierz przedmiot: '))\n if nrP > len(przedmioty) or nrP < 1:\n nrP = 0\n ocenyPrzedmiot = []\n for i in range(len(oceny)):\n if oceny[i]['przedmiot'] == przedmioty[nrP-1]:\n ocenyPrzedmiot.append(oceny[i])\n sredniaUcznia = []\n for i in range(len(uczeniowie)):\n ileOcen = 0\n sumaOcen = 0\n for j in range(len(ocenyPrzedmiot)):\n if ocenyPrzedmiot[j]['uczen'] == uczeniowie[i]['imieNazwisko'] and ocenyPrzedmiot[j]['klasa'] == uczeniowie[i]['klasa']:\n sumaOcen += ocenyPrzedmiot[j]['ocena']\n ileOcen += 1\n if ileOcen != 0:\n sredniaUcznia.append(\n {'uczen': uczeniowie[i]['imieNazwisko'], 'klasa': uczeniowie[i]['klasa'], 'srednia': sumaOcen/ileOcen})\n minSrednia = {'uczen': '', 'klasa': '', 'srednia': 7}\n if len(sredniaUcznia) != 0:\n for i in range(len(sredniaUcznia)):\n if sredniaUcznia[i]['srednia'] < minSrednia['srednia']:\n minSrednia = sredniaUcznia[i]\n print('Najsłabszym uczniem z przedmiotu ' + przedmioty[nrP-1])\n print('jest '+minSrednia['uczen']+' z ' + minSrednia['klasa'] +\n ' ze średnią '+str(minSrednia['srednia']))\n else:\n print('Nie ma najsłabszego ucznia z tego przedmiotu.')\n\ndef najwyzszaSrednia():\n sredniaUcznia = []\n for i in range(len(uczeniowie)):\n ileOcen = 0\n sumaOcen = 0\n for j in range(len(oceny)):\n if oceny[j]['uczen'] == uczeniowie[i]['imieNazwisko'] and oceny[j]['klasa'] == uczeniowie[i]['klasa']:\n sumaOcen += oceny[j]['ocena']\n ileOcen += 1\n if ileOcen != 0:\n sredniaUcznia.append(\n {'uczen': uczeniowie[i]['imieNazwisko'], 'klasa': uczeniowie[i]['klasa'], 'srednia': sumaOcen/ileOcen})\n print(sredniaUcznia)\n maxSrednia = {'uczen': '', 'klasa': '', 'srednia': 0}\n if len(sredniaUcznia) != 0:\n for i in range(len(sredniaUcznia)):\n if sredniaUcznia[i]['srednia'] > maxSrednia['srednia']:\n maxSrednia = sredniaUcznia[i]\n print('Najlepszym uczniem ze wszystkich uczniów')\n print('jest '+maxSrednia['uczen']+' z '+maxSrednia['klasa'] +\n ' ze średnią '+str(maxSrednia['srednia']))\n else:\n print('Nie ma najlepdszego ucznia.')\n\ntak = True\nwhile tak == True:\n subprocess.call(\"cls\", shell=True)\n print(' MENU: ')\n print('1 - Dodaj ucznia')\n print('2 - Modyfikuj ucznia')\n print('3 - Usun ucznia')\n print('4 - Oceny ucznia')\n print('5 - Dodaj ocenę')\n print('6 - Najlepszy uczeń')\n print('7 - Najgorszy uczeń')\n print('8 - Najwyższa średnia')\n print('9 - Wszystkie listy')\n print('0 - Koniec')\n wybor = int(input('Twój wybór: '))\n if wybor == 1:\n dodajUcznia()\n if wybor == 2:\n modyfikujUcznia()\n if wybor == 3:\n usunUcznia()\n if wybor == 4:\n ocenyUcznia()\n if wybor == 5:\n dodajOcene()\n if wybor == 6:\n najlepszyUczen()\n if wybor == 7:\n najgorszyUczen()\n if wybor == 8:\n najwyzszaSrednia()\n if wybor == 9:\n wypiszListy()\n if wybor == 0:\n tak = False\n pauza = input('')\n # time.sleep(5)\n","sub_path":"SzkolaDuza.py","file_name":"SzkolaDuza.py","file_ext":"py","file_size_in_byte":13587,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"164957953","text":"import sys\nimport timeit\nsys.path.append(\"..\")\nfrom heartbeat import HeartBeat\n\n\n# Config Vars\nfile_path = \"../files/test.txt\"\nfile_path2 = \"../files/test2.txt\"\nfile_path3 = \"../files/test3.txt\"\nsize_path = \"../files/test4.txt\"\nroot_seed = \"myroot\"\n\n\n# Unit Test\ndef unit_test():\n\n\t# Create challenges from file\n\tfile1 = HeartBeat(file_path)\n\tfile1.gen_challenges(10, root_seed)\n\tchallenge = file1.get_challenge()\n\n\t# Create hash_response from seed and duplicate file\n\tfile2 = HeartBeat(file_path2)\n\tanswer = file2.meet_challenge(challenge)\n\n\t# Check to see if they match\n\tassert(file1.check_answer(answer))\n\n\t# Create hash_answer from seed and edited file\n\tfile3 = HeartBeat(file_path3)\n\tanswer = file3.meet_challenge(challenge)\n\n\t# This should not match\n\tassert(not file1.check_answer(answer))\n\n\n# Unit Test on a Custom File\ndef custom_unit_test(file_path):\n\t# Create challenges from file\n\tfile1 = HeartBeat(file_path)\n\tfile1.gen_challenges(10, root_seed)\n\tchallenge = file1.get_challenge()\n\n\n# Size Tests\ndef size_test():\n\t# Time and Size of Challenges\n\tprint(\"Month of Challenges (1 per hour):\")\n\tprint(str(timeit.timeit(size1, number=1)) + \" seconds\")\n\tprint(\"Year of Challenges (1 per hour):\")\n\tprint(str(timeit.timeit(size2, number=1)) + \" seconds\")\n\tprint(\"\")\n\ndef num_challenges(number):\n\tfile1 = HeartBeat(size_path)\n\tfile1.gen_challenges(number, root_seed)\n\tprint(\"Size: \" + str(file1.challenges_size()/1024) + \" kb\")\n\ndef size1():\n\tnum_challenges(1000) # 731 hours in a month\n\ndef size2():\n\tnum_challenges(10000) # 8766 hours in a year\n\t\n\nif __name__ == \"__main__\":\n\ttry:\n\t\tunit_test()\n\t\tif (len(sys.argv) == 2):\n\t\t\tcustom_unit_test(sys.argv[1])\n\t\tsize_test()\n\texcept AssertionError:\n\t\tprint(\"Failed Unit Testing...\")\n\telse:\n\t\tprint(\"Passed Unit Testing...\")\n","sub_path":"testing/size.py","file_name":"size.py","file_ext":"py","file_size_in_byte":1771,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"598041116","text":"from datetime import datetime\nfrom source.query.scores_service.domain import MergerKey\n\nfrom source.storage.stores.artifact_store.types.quest.base_artifact import QuestBaseArtifact\n\n\n__all__ = [\n 'QueryMetadataSummary',\n 'QuestConfigurationArtifact',\n 'MergerSummaryArtifact'\n]\n\n\nclass QueryMetadataSummary(QuestBaseArtifact):\n type = 'quest_query_summary'\n\n def __init__(self, customer, quest_id, query_id, sphere_id, kernel_id, kernel_timestamp,\n split_kernel_id, is_past):\n \"\"\"\n @type split_kernel_id: C{str}\n @type customer: C{str}\n @type quest_id: C{str}\n @type query_id: C{str}\n @type sphere_id: C{str}\n @type kernel_id: C{str}\n @type kernel_timestamp: C{datetime}\n @type is_past: C{bool}\n \"\"\"\n super(QueryMetadataSummary, self).__init__(customer, quest_id)\n self.__split_kernel_id = split_kernel_id\n self.__quest_id = quest_id\n self.__query_id = query_id\n self.__sphere_id = sphere_id\n self.__kernel_id = kernel_id\n self.__kernel_timestamp = kernel_timestamp\n self.__is_past = is_past\n\n def _to_dict(self):\n return {\n 'quest_id': self.__quest_id,\n 'query_id': self.__query_id,\n 'sphere_id': self.__sphere_id,\n 'kernel_id': self.__kernel_id,\n 'split_kernel_id': self.__split_kernel_id,\n 'kernel_timestamp': self.__kernel_timestamp,\n 'is_past': self.__is_past,\n }\n\n\nclass QuestConfigurationArtifact(QuestBaseArtifact):\n type = 'quest_runner_configuration'\n\n def __init__(self, customer, quest_id, present_data_unit, past_data_units, quest_config):\n \"\"\"\n @type customer: C{str}\n @type quest_id: C{str}\n @type present_data_unit: C{dict}\n @type past_data_units: C{list}\n @type quest_config: C{dict}\n \"\"\"\n super(QuestConfigurationArtifact, self).__init__(customer, quest_id)\n self.__quest_id = quest_id\n self.__present_data_unit = present_data_unit\n self.__past_data_units = past_data_units\n self.__quest_config = quest_config\n\n def _to_dict(self):\n return {\n 'quest_id': self.__quest_id,\n 'present_data_unit': self.__present_data_unit,\n 'past_data_units': self.__past_data_units,\n 'quest_config': self.__quest_config,\n }\n\n\nclass MergerSummaryArtifact(QuestBaseArtifact):\n type = 'merger_summary'\n\n def __init__(self, customer, quest_id, merger_key):\n \"\"\"\n @type customer: C{str}\n @type quest_id: C{str}\n @type merger_key: L{MergerKey}\n\n \"\"\"\n super(MergerSummaryArtifact, self).__init__(customer, quest_id)\n self.__quest_id = quest_id\n self.__merger_key = merger_key\n\n\n def _to_dict(self):\n return {\n 'quest_id': self.__quest_id,\n 'merger_id': str(self.__merger_key),\n 'merger_model': self.__merger_key.model_name,\n 'scorer_id': self.__merger_key.scorer_name,\n 'variant': self.__merger_key.model_params,\n }\n\n\nclass ProbabilityGraphRepresentationArtifact(QuestBaseArtifact):\n type = 'probability_graph_representation'\n\n def __init__(self, customer, quest_id, graph_representation, num_scored, baselines):\n \"\"\"\n @type customer: C{str}\n @type quest_id: C{str}\n @type graph_representation: C{dict}\n @type num_scored: C{int}\n \"\"\"\n super(ProbabilityGraphRepresentationArtifact, self).__init__(customer, quest_id)\n self.__num_scored = num_scored\n self.__graph_representation = graph_representation\n self.__quest_id = quest_id\n self.__baselines = baselines\n\n def _to_dict(self):\n return {\n 'quest_id': self.__quest_id,\n 'graph_representation': self.__graph_representation,\n 'num_scored': self.__num_scored,\n 'baselines': self.__baselines\n }\n","sub_path":"internal/source/storage/stores/artifact_store/types/quest/general.py","file_name":"general.py","file_ext":"py","file_size_in_byte":4014,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"567073218","text":"import tweepy\r\n\r\nCONSUMER_KEY = '9a5vk4s0OIhqvaNAA48cG7PAT'\r\nCONSUMER_SECRET = 'yTWGaLVyVcu6ZF9oUJ9nt4d1niTNvieCB5DNznJ7OfKMQNC4uG'\r\nACCESS_TOKEN = '61705925-H70JXuNOnKL7TuiZu3pI3O1K9Qe1m4RldFZhuxGqr'\r\nACCESS_TOKEN_SECRET = 'j9O1RUtSQI0sTSFnNv90YeEXholdArvb5vmTNoqmBTbVd'\r\n\r\n\r\n#OAuth\r\nauth = tweepy.OAuthHandler(CONSUMER_KEY, CONSUMER_SECRET)\r\nauth.set_access_token(ACCESS_TOKEN, ACCESS_TOKEN_SECRET)\r\napi = tweepy.API(auth)\r\n\r\nstatus = \"Status Update made using Tweepy :D\"\r\napi.update_status(status=status)\r\n","sub_path":"updateStatus.py","file_name":"updateStatus.py","file_ext":"py","file_size_in_byte":509,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"357360692","text":"#!/usr/bin/env python\n\n\"\"\"Universal CSV to JSON converter with scalability options\"\"\"\n\n__author__ = \"Tim Verlaan 11669128\"\n\nimport csv \nimport json \n \ndef convert():\n \"\"\"Convert CSV file to JSON file\"\"\"\n\n # Open the CSV \n f = open( 'data_april1.csv') \n\n # Change each fieldname to the appropriate field name. I know, so difficult. \n reader = csv.DictReader( f, fieldnames = ( \"STN\",\"YYYYMMDD\",\"SP\",\"UX\" )) \n\n # skip the header \n next(reader)\n\n # Parse the CSV into JSON \n out = json.dumps( [ row for row in reader ] ) \n \n # Save the JSON \n f = open( 'data_april.json', 'w') \n f.write(out) \n\n\nif __name__ == \"__main__\":\n \"\"\"Separating the function, for scalability purposes\"\"\"\n \n convert()\n\n","sub_path":"Homework/Week_3/convertCSV2JSON.py","file_name":"convertCSV2JSON.py","file_ext":"py","file_size_in_byte":752,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"547375132","text":"import numpy as np\nimport pandas as pd\nfrom numpy import random\nimport idx2numpy\n\n# here, xt1 ,yt1 is the train set and xt2,yt2 is the test set\nfile1 = 'train-images-idx3-ubyte(1)'\nfile2 = 'train-labels-idx1-ubyte'\nfile3 = 't10k-images-idx3-ubyte'\nfile4 = 't10k-labels-idx1-ubyte'\n\nxt1 = idx2numpy.convert_from_file(file1)\nyt1 = idx2numpy.convert_from_file(file2)\nxt2 = idx2numpy.convert_from_file(file3)\nyt2 = idx2numpy.convert_from_file(file4)\n\n#scaling the values of intensities\nxt1 = xt1/255\nxt2 = xt2/255\n\n\n# unrolling xt1 data for every image and running training for \n# neural Network\na = np.shape(yt1)\nm = a[0]\nn = 0.01\nL = 100\nd = 784\nK = 10\nla = 100\nx2 = xt1.ravel()\nxt3 = x2.reshape((m,d))\nyt = np.zeros((m,K))\nfor i in range(m):\n for j in range(K):\n yt[i][yt1[i]]= 1\n\n#Initializing weights and biases with random numbers\nw1 = 2-2*np.random.rand(d,L)\nw2 = 2-2*np.random.rand(L,K)\n\n# Sigmoid Function\ndef sgmd(w,X):\n m1 = (np.dot(X,w))\n p = np.array(m1,dtype = np.float128)\n z = 1/(1+np.exp(-p))\n return z\n\n# Cost function\ndef costfn(y,hth,w1,w2,la,m,K):\n I = np.ones((m,K))\n lhth = np.log(hth)\n Ilhth = np.log(I-hth)\n Iy = I-y\n y1 = np.array(y)\n Iy1 = np.array(Iy)\n j1 = -y1*lhth-Iy1*Ilhth\n j11 = j1.sum()\n j111 = j11/m\n t1 = np.square(w1)\n t2 = np.square(w2)\n s1 = t1.sum()\n s2 = t2.sum()\n j2 = (la/(2*m))*(s1+s2)\n j = j111+j2\n return j\n\n# Error calculation\ndef error(y,ypr,m):\n t = np.transpose(y-ypr)\n t1 = y-ypr\n a = np.dot(t,t1)\n s = a.sum()/m\n return s\n\n# Calculating delta wrt weights\n# Running epoch for forward and backward propagation\nE = [] # Error calculation\nelo = 0\nfor i in range(50):\n # Forward Propagation Step\n A = sgmd(w1,xt3) # A is the output given by hidden layer\n a = (A-A.mean(axis=1))/A.std(axis=1)\n #Predicted value of y\n ypr = sgmd(w2,a)\n E.append(error(y,ypr,m)) # storing error values in an array \n er = ypr-yt #Error calculation for back propagation\n \n # Back Propagation step propagation of error del(L)= (del(L+1)*theta(3)T)*derivative of sigmoid\n Ad = np.array(a)\n I = np.ones((m,L))\n Am = np.array(I-a)\n I = np.ones((m,10))\n q = np.array(ypr)\n qm = np.array(I-ypr)\n er1 = np.array(er)\n # Calculation of delt2 and delt1 which will be used for updation\n delt2 = er1*q*qm\n er2 = np.dot(delt2,np.transpose(w2)) # Propagation of error step\n delt1 = er2*Ad*Am\n w2 =(w2-(n*(np.dot(np.transpose(Ad),delt2)))) # Updating values of w1 & w2\n w1 =(w1-(n*(np.dot(np.transpose(xt3),delt1)))) \n\n# Calculation of final predicted values for the train step\nA = sgmd(w1,xt3)\nypr = sgmd(w2,A)\n\n#testing the model\nx5 = xt2.ravel() # Unrolling the test data \nm1 = 10000\n# Calculation of final predicted value for test step\nxt4 = x5.reshape((m1,d))\nA = sgmd(w1,xt4)\nypr1 = sgmd(w2,A)\n# converting the values of the model into the class \nyprd1 = np.argmax(ypr1,axis=1)\n\n# Counting the number of matching values of both set\ncount = 0\nfor i in range(0,m1):\n if yprd1[i]==yt2[i]:\n count = count+1\n\n# Calculation of accuracy of train set\nacc = (count/m1)*100\nprint('Accuracy (Test Set):')\nprint(acc)\n\n#Train Accuracy \nA = sgmd(w1,xt3)\nypr = sgmd(w2,A)\nyprd2 = np.argmax(ypr,axis=1) # Final predicted class to be compared\n# Counting equal values \ncount = 0\nfor i in range(0,m1):\n if yprd2[i]==yt1[i]:\n count = count+1\n\n# Calculating accuracy\nacc = (count/m)*100\nprint('Accuracy (Train Set):')\nprint(acc)","sub_path":"Assignment4.py","file_name":"Assignment4.py","file_ext":"py","file_size_in_byte":3490,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"349435532","text":"N = int(input())\nname = 'atcoder'\npeople = 0\nS = 0\nfor n in range(N):\n s,p = input().split()\n p = int(p)\n if p > people:\n people = p\n name = s\n S += p\nif people > S/2:\n print(name)\nelse:\n print('atcoder')\n","sub_path":"ABC/ABC_033/abc033b.py","file_name":"abc033b.py","file_ext":"py","file_size_in_byte":238,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"314953133","text":"# Copyright\n# Greig Cowan 2015\n# Konstantin Gizdov 2016\n\n\nimport GaudiKernel.SystemOfUnits as Units\nfrom Gaudi.Configuration import *\n\n####\nfrom StrippingConf.Configuration import StrippingConf, StrippingStream\nfrom StrippingSettings.Utils import strippingConfiguration\nfrom StrippingArchive.Utils import buildStreams\nfrom StrippingArchive import strippingArchive\n####\n\nfrom PhysSelPython.Wrappers import AutomaticData, Selection, SelectionSequence\nfrom Configurables import FilterDesktop\nfrom Configurables import DaVinci\nfrom Configurables import DecayTreeTuple\nfrom Configurables import TupleToolDecay\nfrom Configurables import GaudiSequencer\nfrom Configurables import CombineParticles\nfrom Configurables import CondDB\nfrom Configurables import LoKi__Hybrid__TupleTool\nfrom Configurables import LoKi__Hybrid__TupleTool\nfrom Configurables import LoKi__Hybrid__EvtTupleTool\nfrom Configurables import MCTupleToolHierarchy\nfrom Configurables import TupleToolMCTruth\nfrom Configurables import TupleToolMCBackgroundInfo\nfrom Configurables import TupleToolTISTOS, TriggerTisTos\n\nEVTMAX = -1\nMODE = \"MC\"\nOUTPUTLEVEL = ERROR\n\n## # =============================================================================\n## ## read PSIX0.MDST\n## # =============================================================================\n\n## rootInTES = '/Event/PSIX0'\n## location = 'Phys/SelB2PsiOmegaForPsiX0/Particles'\n## from PhysSelPython.Wrappers import AutomaticData\n## b2omega_selection = AutomaticData( location )\n\n\n# =============================================================================\n## run PSIX0 WG-selections over ALLSTREAMS.DST MC\n# =============================================================================\njpsi_name = 'FullDSTDiMuonJpsi2MuMuDetachedLine'\njpsi_line = '/Event/AllStreams/Phys/%s/Particles' % jpsi_name\n\nimport StrippingSelections.StrippingBandQ.StrippingPsiX0 as PSIX0\npsix0 = PSIX0.PsiX0Conf (\n 'PsiX0' ,\n config = {\n 'NOPIDHADRONS' : True , ## important here!!!\n 'DIMUONLINES' : [ jpsi_line ]\n }\n )\n\nb2omega_selection = psix0.b2omega()\n\n\n#########################################################################################################\n# Set up the MCDecayTreeTuples for each of the decays that we are interested in.\n# We want to save all of the generated events for each mode.\n#########################################################################################################\nfrom Configurables import MCDecayTreeTuple, MCTupleToolKinematic, MCTupleToolHierarchy, LoKi__Hybrid__MCTupleTool\n\n# LoKi variables\nLoKi_Photos = LoKi__Hybrid__MCTupleTool(\"LoKi_Photos\")\nLoKi_Photos.Variables = {\n \"nPhotons\" : \"MCNINTREE ( ('gamma'==MCABSID) )\",\n \"MC_PT\" : \"MCPT\",\n \"MC_THETA\" : \"MCTHETA\",\n \"MC_ETA\" : \"MCETA\",\n \"MC_PHI\" : \"MCPHI\",\n \"MC_ABSID\" : \"MCABSID\"\n }\n\nfrom GaudiConfUtils.ConfigurableGenerators import MCDecayTreeTuple as MCTUPLE\nfrom PhysSelPython.Wrappers import SimpleSelection\n\nmc_selection = SimpleSelection (\n 'MCTuple'\n , MCTUPLE\n , [ b2omega_selection ]\n ## properties\n , Decay = \" [B0]cc => ^(J/psi(1S) => ^mu+ ^mu- ) ^(omega(782) ==> ^pi+ ^pi- ^(pi0 => ^gamma ^gamma) ) \"\n , Branches = {\n 'B0' : \"[B0]cc => (J/psi(1S) => mu+ mu- ) (omega(782) ==> pi+ pi- (pi0 => gamma gamma) )\"\n , 'Jpsi' : \"[B0]cc => ^(J/psi(1S) => mu+ mu- ) (omega(782) ==> pi+ pi- (pi0 => gamma gamma) )\"\n , 'muplus' : \"[B0]cc => (J/psi(1S) => ^mu+ mu- ) (omega(782) ==> pi+ pi- (pi0 => gamma gamma) )\"\n , 'muminus' : \"[B0]cc => (J/psi(1S) => mu+ ^mu- ) (omega(782) ==> pi+ pi- (pi0 => gamma gamma) )\"\n , 'omega' : \"[B0]cc => (J/psi(1S) => mu+ mu- ) ^(omega(782) ==> pi+ pi- (pi0 => gamma gamma) )\"\n , 'piplus' : \"[B0]cc => (J/psi(1S) => mu+ mu- ) (omega(782) ==> ^pi+ pi- (pi0 => gamma gamma) )\"\n , 'piminus' : \"[B0]cc => (J/psi(1S) => mu+ mu- ) (omega(782) ==> pi+ ^pi- (pi0 => gamma gamma) )\"\n , 'pizero' : \"[B0]cc => (J/psi(1S) => mu+ mu- ) (omega(782) ==> pi+ pi- ^(pi0 => gamma gamma) )\"\n , 'gamma1' : \"[B0]cc => (J/psi(1S) => mu+ mu- ) (omega(782) ==> pi+ pi- (pi0 => ^gamma gamma) )\"\n , 'gamma2' : \"[B0]cc => (J/psi(1S) => mu+ mu- ) (omega(782) ==> pi+ pi- (pi0 => gamma ^gamma) )\"\n }\n )\nmctuple_B2psiomega = mc_selection.algorithm()\n\n\n# List of the mc tuples\nmctuples = [\n mctuple_B2psiomega\n ]\n\nfor tup in mctuples:\n tup.addTool(MCTupleToolKinematic())\n tup.MCTupleToolKinematic.Verbose=True\n tup.addTool(LoKi_Photos)\n tup.ToolList = [ \"MCTupleToolHierarchy\"\n , \"MCTupleToolKinematic\"\n , \"LoKi::Hybrid::MCTupleTool/LoKi_Photos\" # doesn't work with DaVinci v36r6\n ]\n tup.addTool(TupleToolMCTruth, name = \"TruthTool\")\n tup.addTool(TupleToolMCBackgroundInfo, name = \"BackgroundInfo\")\n tup.ToolList += [\"TupleToolMCTruth/TruthTool\"]\n tup.ToolList += [\"TupleToolMCBackgroundInfo/BackgroundInfo\"]\n\nif OUTPUTLEVEL == DEBUG:\n from Configurables import PrintMCTree, PrintMCDecayTreeTool\n mctree = PrintMCTree(\"PrintTrue\")\n mctree.addTool( PrintMCDecayTreeTool )\n mctree.PrintMCDecayTreeTool.Information = \"Name M P Px Py Pz Pt Vx Vy Vz\"\n mctree.ParticleNames = [ \"B+\", \"B-\" ]\n mctree.Depth = 3 # down to the K and mu\n\n#########################################################################################################\n# Now set up the DecayTreeTuples for the reconstructed particles\n#########################################################################################################\ntupletools = []\ntupletools.append(\"TupleToolKinematic\")\ntupletools.append(\"TupleToolGeometry\")\ntupletools.append(\"TupleToolTrackInfo\")\ntupletools.append(\"TupleToolPid\")\ntupletools.append(\"TupleToolRecoStats\")\ntupletools.append(\"TupleToolEventInfo\")\ntriglist = [\n \"L0PhysicsDecision\"\n ,\"L0MuonDecision\"\n ,\"L0DiMuonDecision\"\n ,\"L0MuonHighDecision\"\n ,\"L0HadronDecision\"\n ,\"L0ElectronDecision\"\n ,\"L0PhotonDecision\"\n ,\"Hlt1DiMuonHighMassDecision\"\n ,\"Hlt1DiMuonLowMassDecision\"\n ,\"Hlt1SingleMuonNoIPDecision\"\n ,\"Hlt1SingleMuonHighPTDecision\"\n ,\"Hlt1TrackAllL0Decision\"\n ,\"Hlt1TrackMuonDecision\"\n ,\"Hlt1TrackPhotonDecision\"\n ,\"Hlt1L0AnyDecision\"\n ,\"Hlt2SingleElectronTFLowPtDecision\"\n ,\"Hlt2SingleElectronTFHighPtDecision\"\n ,\"Hlt2DiElectronHighMassDecision\"\n ,\"Hlt2DiElectronBDecision\"\n ,\"Hlt2B2HHLTUnbiasedDecision\"\n ,\"Hlt2Topo2BodySimpleDecision\"\n ,\"Hlt2Topo3BodySimpleDecision\"\n ,\"Hlt2Topo4BodySimpleDecision\"\n ,\"Hlt2Topo2BodyBBDTDecision\"\n ,\"Hlt2Topo3BodyBBDTDecision\"\n ,\"Hlt2Topo4BodyBBDTDecision\"\n ,\"Hlt2TopoMu2BodyBBDTDecision\"\n ,\"Hlt2TopoMu3BodyBBDTDecision\"\n ,\"Hlt2TopoMu4BodyBBDTDecision\"\n ,\"Hlt2TopoE2BodyBBDTDecision\"\n ,\"Hlt2TopoE3BodyBBDTDecision\"\n ,\"Hlt2TopoE4BodyBBDTDecision\"\n ,\"Hlt2MuonFromHLT1Decision\"\n ,\"Hlt2DiMuonDecision\"\n ,\"Hlt2DiMuonLowMassDecision\"\n ,\"Hlt2DiMuonJPsiDecision\"\n ,\"Hlt2DiMuonJPsiHighPTDecision\"\n ,\"Hlt2DiMuonPsi2SDecision\"\n ,\"Hlt2DiMuonBDecision\"\n]\nTISTOSTool = TupleToolTISTOS('TISTOSTool')\nTISTOSTool.VerboseL0 = True\nTISTOSTool.VerboseHlt1 = True\nTISTOSTool.VerboseHlt2 = True\nTISTOSTool.TriggerList = triglist[:]\nTISTOSTool.addTool( TriggerTisTos, name=\"TriggerTisTos\")\n\nLoKi_B0 = LoKi__Hybrid__TupleTool(\"LoKi_B0\")\nLoKi_B0.Variables = {\n \"ETA\" : \"ETA\"\n , \"PHI\" : \"PHI\"\n , \"FDCHI2\" : \"BPVVDCHI2\"\n , \"BPVDLS\" : \"BPVDLS\" # decay length significance to the best PV\n , \"DIRA\" : \"BPVDIRA\"\n , \"DTF_CTAU\" : \"DTF_CTAU( 0, True )\"\n , \"DTF_CTAUS\" : \"DTF_CTAUSIGNIFICANCE( 0, True )\"\n , \"DTF_CHI2NDOF\" : \"DTF_CHI2NDOF( True )\"\n , \"DTF_CTAUERR\" : \"DTF_CTAUERR( 0, True )\"\n , \"DTF_MASS_constr\" : \"DTF_FUN ( M , True , strings(['J/psi(1S)', 'pi0']) )\"\n , \"DTF_VCHI2NDOF\" : \"DTF_FUN ( VFASPF(VCHI2/VDOF) , True )\"\n }\n\nLoKi_Jpsi = LoKi__Hybrid__TupleTool(\"LoKi_Jpsi\")\nLoKi_Jpsi.Variables = {\n \"BPVDLS\" : \"BPVDLS\" # decay length significance to the best PV\n }\n\nLoKi_Mu = LoKi__Hybrid__TupleTool(\"LoKi_Mu\")\nLoKi_Mu.Variables = {\n \"NSHAREDMU\" : \"NSHAREDMU\"\n }\n\n\nfrom GaudiConfUtils.ConfigurableGenerators import DecayTreeTuple as TUPLE\nfrom PhysSelPython.Wrappers import SimpleSelection\nrd_selection = SimpleSelection (\n 'Tuple'\n , TUPLE\n , [ b2omega_selection ]\n ## Properties:\n , Decay = '[B0]cc -> ^(J/psi(1S) -> ^mu+ ^mu-) ^(omega(782) -> ^pi+ ^pi- ^(pi0 -> ^gamma ^gamma) )'\n , ToolList = tupletools\n , Branches = {\n #\n 'B0' : \"[B0]cc -> (J/psi(1S) -> mu+ mu- ) (omega(782) -> pi+ pi- (pi0 -> gamma gamma) )\"\n , 'Jpsi' : \"[B0]cc -> ^(J/psi(1S) -> mu+ mu- ) (omega(782) -> pi+ pi- (pi0 -> gamma gamma) )\"\n , 'muplus' : \"[B0]cc -> (J/psi(1S) -> ^mu+ mu- ) (omega(782) -> pi+ pi- (pi0 -> gamma gamma) )\"\n , 'muminus' : \"[B0]cc -> (J/psi(1S) -> mu+ ^mu- ) (omega(782) -> pi+ pi- (pi0 -> gamma gamma) )\"\n , 'omega' : \"[B0]cc -> (J/psi(1S) -> mu+ mu- ) ^(omega(782) -> pi+ pi- (pi0 -> gamma gamma) )\"\n , 'piplus' : \"[B0]cc -> (J/psi(1S) -> mu+ mu- ) (omega(782) -> ^pi+ pi- (pi0 -> gamma gamma) )\"\n , 'piminus' : \"[B0]cc -> (J/psi(1S) -> mu+ mu- ) (omega(782) -> pi+ ^pi- (pi0 -> gamma gamma) )\"\n , 'pizero' : \"[B0]cc -> (J/psi(1S) -> mu+ mu- ) (omega(782) -> pi+ pi- ^(pi0 -> gamma gamma) )\"\n , 'gamma1' : \"[B0]cc -> (J/psi(1S) -> mu+ mu- ) (omega(782) -> pi+ pi- (pi0 -> ^gamma gamma) )\"\n , 'gamma2' : \"[B0]cc -> (J/psi(1S) -> mu+ mu- ) (omega(782) -> pi+ pi- (pi0 -> gamma ^gamma) )\"\n }\n )\n\ntuple_B2psiomega = rd_selection.algorithm()\n\nfor particle in [\"B0\", \"Jpsi\", \"muplus\", \"muminus\", \"omega\", \"piplus\", \"piminus\", \"pizero\", \"gamma1\", \"gamma2\"]:\n tuple_B2psiomega.addTool(TupleToolDecay, name = particle)\n\n# List of the reconstructed tuples\ntuples = [ tuple_B2psiomega\n ]\n\nfor tup in tuples:\n # tup.ReFitPVs = True\n if MODE == \"MC\":\n tup.addTool(TupleToolMCTruth, name = \"TruthTool\")\n tup.addTool(TupleToolMCBackgroundInfo, name = \"BackgroundInfo\")\n tup.ToolList += [\"TupleToolMCTruth/TruthTool\"]\n tup.ToolList += [\"TupleToolMCBackgroundInfo/BackgroundInfo\"]\n\n tup.B0.addTool( LoKi_B0 )\n tup.B0.ToolList += [\"LoKi::Hybrid::TupleTool/LoKi_B0\"]\n tup.Jpsi.addTool( LoKi_Jpsi )\n tup.Jpsi.ToolList += [\"LoKi::Hybrid::TupleTool/LoKi_Jpsi\"]\n tup.muplus.addTool( LoKi_Mu )\n tup.muplus.ToolList += [\"LoKi::Hybrid::TupleTool/LoKi_Mu\"]\n tup.muminus.addTool( LoKi_Mu )\n tup.muminus.ToolList += [\"LoKi::Hybrid::TupleTool/LoKi_Mu\"]\n for particle in [ tup.B0 ]:\n particle.addTool(TISTOSTool, name = \"TISTOSTool\")\n particle.ToolList += [ \"TupleToolTISTOS/TISTOSTool\" ]\n\n\n\nfrom PhysSelPython.Wrappers import SelectionSequence\nrd_SEQ = SelectionSequence ( 'DATA' , rd_selection )\nmc_SEQ = SelectionSequence ( 'MC' , mc_selection )\n\n\n\n\n###################### DAVINCI SETTINGS ############################################\n\nlum = True\nsim = False\n\nif MODE == 'MC':\n lum = False\n sim = True\n\ndaVinci = DaVinci (\n EvtMax = EVTMAX\n , TupleFile = \"DVTuples1.root\"\n , HistogramFile = 'DVHistos.root'\n , DataType = \"2011\"\n , Simulation = sim\n , Lumi = lum\n , UserAlgorithms = [rd_SEQ.sequence(), mc_SEQ.sequence()]\n )\n\nMessageSvc().Format = \"% F%60W%S%7W%R%T %0W%M\"\n\n###################################################################################\n####################### THE END ###################################################\n###################################################################################\n","sub_path":"scripts/davinci/mc_2012.py","file_name":"mc_2012.py","file_ext":"py","file_size_in_byte":12082,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"51581503","text":"import json\nimport requests\nimport pandas as pd\n\n#Seattle Crime data\nurl = \"https://data.seattle.gov/resource/tazs-3rd5.json?$limit=900000&$offset=0&$order=offense_id\"\n\noutput_folder = 'D:\\\\OneDrive - PACCAR Inc\\\\Current_Projects\\\\Personal\\\\Courses\\\\DataScientist\\\\Project 1\\\\cov2_crimerate\\\\data\\\\raw\\\\'\n\ndef api_json_to_pandas(url):\n r = requests.get(url)\n json_data = r.json()\n df = pd.json_normalize(json_data)\n return df\n\nif __name__ == '__main__':\n seattle_crime_data=api_json_to_pandas(url)\n seattle_crime_data.to_csv(output_folder+'Seattle_Crime_Data.csv')\n\n\n\n\n\n","sub_path":"src/data/make_dataset.py","file_name":"make_dataset.py","file_ext":"py","file_size_in_byte":586,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"509356281","text":"import pandas as pd\nimport numpy as np\n\nciti_bike = pd.read_csv(\"data/201811-citibike-tripdata.csv\")\nnov18_df = pd.DataFrame(citi_bike)\nnov18_df.head(20)\nnov18_df.dtypes\nnov18_df['gender'] = nov18_df['gender'].map({0:'unknown', 1:'male', 2:'female'})\nnov18_df.rename(columns={'tripduration': \"trip duration (seconds)\"}, inplace = True)\nnov18_df.dtypes\nnov18_df.isnull().any()\nnov18_df.count()\nnov18_df.dropna(how='any', inplace=True)\nnov18_df.count()\nnov18_df.head(20)\nnov18_df.to_csv(\"data/clean_201811.csv\", index=False)\n\ndec_citi_bike = pd.read_csv(\"data/201812-citibike-tripdata.csv\")\ndec18_df = pd.DataFrame(dec_citi_bike)\ndec18_df.head(20)\ndec18_df.dtypes\ndec18_df['gender'] = dec18_df['gender'].map({0:'unknown', 1:'male', 2:'female'})\ndec18_df.rename(columns={'tripduration': \"trip duration (seconds)\"}, inplace = True)\ndec18_df.dtypes\ndec18_df.isnull().any()\ndec18_df.count()\ndec18_df.dropna(how='any', inplace=True)\ndec18_df.count()\ndec18_df.head(20)\ndec18_df.to_csv(\"data/clean_201812.csv\", index=False)\n\njune_citi_bike = pd.read_csv(\"data/201806-citibike-tripdata.csv\")\njune18_df = pd.DataFrame(june_citi_bike)\njune18_df.head(20)\njune18_df.dtypes\njune18_df['gender'] = june18_df['gender'].map({0:'unknown', 1:'male', 2:'female'})\njune18_df.rename(columns={'tripduration': \"trip duration (seconds)\"}, inplace = True)\njune18_df.dtypes\njune18_df.isnull().any()\njune18_df.count()\njune18_df.to_csv(\"data/clean_201806.csv\", index=False)\n\njuly_citi_bike = pd.read_csv(\"data/201807-citibike-tripdata.csv\")\njuly18_df = pd.DataFrame(july_citi_bike)\njuly18_df.head(20)\njuly18_df.dtypes\njuly18_df['gender'] = july18_df['gender'].map({0:'unknown', 1:'male', 2:'female'})\njuly18_df.rename(columns={'tripduration': \"trip duration (seconds)\"}, inplace = True)\njuly18_df.dtypes\njuly18_df.isnull().any()\njuly18_df.count()\njuly18_df.to_csv(\"data/clean_201807.csv\", index=False)\n\naug_citi_bike = pd.read_csv(\"data/201808-citibike-tripdata.csv\")\naug18_df = pd.DataFrame(aug_citi_bike)\naug18_df.head(20)\naug18_df.dtypes\naug18_df['gender'] = aug18_df['gender'].map({0:'unknown', 1:'male', 2:'female'})\naug18_df.rename(columns={'tripduration': \"trip duration (seconds)\"}, inplace = True)\naug18_df.dtypes\naug18_df.isnull().any()\naug18_df.count()\naug18_df.dropna(how='any', inplace=True)\naug18_df.count()\naug18_df.head(20)\naug18_df.to_csv(\"data/clean_201808.csv\", index=False)\n","sub_path":"Tableau/citi_bike.py","file_name":"citi_bike.py","file_ext":"py","file_size_in_byte":2363,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"466288596","text":"#!/usr/bin/python \n# ******************************************************\n# Author : CoffreLv\n# Last modified: 2019-05-27 10:16\n# Email : coffrelv@163.com\n# Filename : corpus_frequency_statistics.py\n# Description : 统计语料词频\n# ******************************************************\n\nimport re,collections,os,sys\n\nfiles_Path = sys.argv[1]\n\ndef get_frequency_words(filepath):\n with open(filepath) as f:\n min_Frequency = 0\n max_Frequency = 0\n words_Num_All = 0\n words_box = []\n for line in f:\n if re.match(r'[a-zA-Z]*', line): #蒙语汉语标注统计\n #if re.sub('[ ]','', line): #汉语字统计\n # line = ' '.join(line) #汉语字统计\n words_box.extend(line.strip().split())\n words_counter = collections.Counter(words_box)\n with open('./doc/words_counter.txt', mode = 'w', encoding = 'utf-8') as f_Counter:\n for key in words_counter.keys():\n words_Num_All += words_counter[key]\n if words_counter[key] < min_Frequency or min_Frequency == 0:\n min_Frequency = words_counter[key]\n if words_counter[key] > max_Frequency or max_Frequency == 0:\n max_Frequency = words_counter[key]\n f_Counter.write(key + '\\t' + str(words_counter[key]) + '\\n')\n print(\"min_Frequency: %d\\tmax_Frequency: %d\\twords_Num_All: %d\"%(min_Frequency,max_Frequency,words_Num_All))\n return words_counter\n\ndef get_frequency_2_words(filepath):\n with open(filepath) as f:\n words_box = []\n words_box_2 = []\n for line in f:\n if re.match(r'[a-zA-Z]*', line): #标注统计\n #if re.sub('[ ]','', line): #汉语\n #line = ' '.join(line) #汉语\n words_box.extend(line.strip().split())\n #print(words_box)\n i = 0\n while i < len(words_box)-1:\n words_box_2.append(words_box[i] + ' ' + words_box[i+1])\n i += 1\n words_counter = collections.Counter(words_box_2)\n with open('./doc/words_counter_2.txt', mode = 'w', encoding = 'utf-8') as f_Counter:\n for key in words_counter.keys():\n f_Counter.write(key + '\\t' + str(words_counter[key]) + '\\n')\n return words_counter\n\ndef Count_the_text_thchs30(files_Path):\n textpath = os.path.abspath(files_Path)\n f = open(\"./doc/thchs30.txt\", mode = 'w', encoding = 'utf-8')\n textnamelist = []\n for parent, dirname, filenames in os.walk(textpath):\n for filename in filenames:\n if not filename.endswith('.trn'):\n continue\n else:\n text_path = os.path.join(parent, filename)\n textnamelist.append(text_path)\n textnamelist.sort()\n for textname in textnamelist:\n f_single = open(textname, mode ='r', encoding = 'utf-8')\n tmp1 = f_single.readline()\n tmp = f_single.readline()\n #tmp = re.sub('[ ]','',tmp)\n tmp = tmp.strip()\n f.write(tmp + '\\n')\n f_single.close()\n f.close()\n\ndef Count_the_text_IMUT(files_Path):\n textpath = os.path.abspath(files_Path)\n f = open(\"./doc/mark.txt\", mode = 'w', encoding = 'utf-8')\n textnamelist = []\n for parent, dirname, filenames in os.walk(textpath):\n for filename in filenames:\n if not filename.endswith('.txt'):\n continue\n else:\n text_path = os.path.join(parent, filename)\n textnamelist.append(text_path)\n textnamelist.sort()\n for textname in textnamelist:\n f_single = open(textname, mode ='r', encoding = 'utf-8')\n tmp = f_single.readline()\n tmp = re.sub('[^a-zA-Z ]','',tmp)\n tmp = tmp.strip()\n f.write(tmp + '\\n')\n f_single.close()\n f.close()\n\nif __name__ == '__main__':\n #Count_the_text_IMUT(files_Path)\n #get_frequency_words('./doc/thchs30.txt')\n #get_frequency_2_words('./doc/thchs30.txt')\n get_frequency_words('./doc/mark.txt')\n get_frequency_2_words('./doc/mark.txt')\n","sub_path":"Python/Tools/src/corpus_frequency_statistics.py","file_name":"corpus_frequency_statistics.py","file_ext":"py","file_size_in_byte":4122,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"113476516","text":"\nisGuessed = False\n\nwhile isGuessed != True:\n\n riddle = \"\\nБыло 6 друзей, на сколько равных частей надо разрезать торт, что бы всем досталось одинаковое количество торта \"\n\n answer = input(riddle)\n\n if answer == \"6\":\n print(\"правельно\")\n isGuessed = True\n else:\n print(\"ты не прав, попробуй ещё раз\")\nprint(\"спасибо за игру\")\n","sub_path":"04Puzzle.py","file_name":"04Puzzle.py","file_ext":"py","file_size_in_byte":494,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"608664645","text":"'''\nOhjelma on skaalautuva versio ristinollasta. Käyttäjälle aukeaa ikkuna, johon\nvoi syöttää pelin tiedot. Kummallekin pelaajalle voi valita nimen ja värin.\nPeliruudukon kokoa ja määrää, kuinka monta ristiä/nollaa tarvitaan riviin\nvoittaakseen, voi myös muuttaa. Start-nappulan painaminen tarkistaa, että\nkäyttäjän syöttet ovat kelpaavat ja aloittaa uuden pelin.\n'''\n\nfrom game import Game\nimport tkinter as tk\n\n\n# Display a window for the user to enter game options and start the game.\nroot = tk.Tk()\nroot.title('New game')\nroot.resizable(False, False)\nframe = tk.Frame(root) # For padding\nframe.grid(padx=(20, 20), pady=(20, 20))\n\ncolors = ('red', 'blue', 'green', 'orange', 'purple', 'yellow')\nname_entries = []\ncolor_variables = []\nfor i in range(2):\n tk.Label(frame, text=f'Player {i + 1}').grid(row=i, column=0, sticky=tk.E, pady=(2, 2))\n\n name_entry = tk.Entry(frame, width=10) # Player name\n name_entry.insert(tk.END, 'Alice' if i == 0 else 'Bob') # Default values\n name_entry.grid(row=i, column=1, sticky=tk.EW, pady=(2, 2))\n name_entries.append(name_entry)\n\n color_variable = tk.StringVar(frame)\n color_variable.set(colors[i]) # Default values\n color_menu = tk.OptionMenu(frame, color_variable, *colors) # Player color\n color_menu.grid(row=i, column=2, sticky=tk.W, pady=(2, 2))\n color_variables.append(color_variable)\n\ngrid_spinboxes = []\ntk.Label(frame, text='Grid size').grid(row=2, column=0, sticky=tk.E, pady=(2, 2))\nfor i in range(2):\n grid_spinbox = tk.Spinbox(frame, from_=3, to=20, width=3) # Grid size\n grid_spinbox.delete(0, tk.END)\n grid_spinbox.insert(0, 15) # Default values\n grid_spinbox.grid(row=2, column=i + 1, sticky=tk.E if i == 0 else tk.W, pady=(2, 2))\n grid_spinboxes.append(grid_spinbox)\n\ntk.Label(frame, text='Winning length').grid(row=3, column=0, sticky=tk.E, pady=(2, 2)) # Winning length\ncount_spinbox = tk.Spinbox(frame, from_=2, to=10, width=3)\ncount_spinbox.delete(0, tk.END)\ncount_spinbox.insert(0, 5) # Default value\ncount_spinbox.grid(row=3, column=1, sticky=tk.E, pady=(2, 2))\n\ndef start():\n # Instantiate a new game if user entered options are valid.\n try:\n cells_x = int(grid_spinboxes[0].get())\n cells_y = int(grid_spinboxes[1].get())\n assert cells_x > 0 and cells_y > 0\n except (ValueError, AssertionError):\n print('Grid size must be a positive integer')\n return\n try:\n winning_count = int(count_spinbox.get())\n assert winning_count > 0\n except (ValueError, AssertionError):\n print('Winning length must be a positive integer')\n return\n players = (name_entries[0].get(), name_entries[1].get())\n colors = (color_variables[0].get(), color_variables[1].get())\n Game(cells_x, cells_y, winning_count, players, colors, root)\n\ntk.Button(frame, text='Start', command=start).grid(row=3, column=2, sticky=tk.EW, padx=(2, 0), pady=(2, 2)) # Start button\nroot.mainloop()\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2959,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"65906603","text":"import argparse\nfrom pathlib import Path\n\nfrom ikuta_ml.crawler.twitter_crawler import TwitterClawler\n\n\ndef main():\n args = _parse_args()\n\n crawler = TwitterClawler(Path(args.output_dir))\n crawler.get_conversation(args.keyword, args.since_id, args.limit)\n\n\ndef _parse_args():\n parser = argparse.ArgumentParser()\n parser.add_argument('-k', '--keyword', required=True)\n parser.add_argument('-o', '--output_dir', required=True)\n parser.add_argument('-s', '--since_id')\n parser.add_argument('-l', '--limit', type=int, default=0)\n return parser.parse_args()\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"ikuta_ml/job/get_dataset_job.py","file_name":"get_dataset_job.py","file_ext":"py","file_size_in_byte":620,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"447410798","text":"# Definition for a binary tree node.\nclass TreeNode:\n def __init__(self, x):\n self.val = x\n self.left = None\n self.right = None\n\n def __repr__(self):\n return str(self.val)\n\nclass Solution:\n def constructFromPrePost(self, pre, post) -> TreeNode:\n\n post_offset = dict()\n for i, v in enumerate(post):\n post_offset[v] = i\n\n tree = TreeNode(pre[0])\n stack = []\n for v in pre[1:]:\n while post_offset[v] > post_offset[tree.val]:\n tree = stack.pop()\n if not tree.left:\n tree.left = TreeNode(v)\n stack.append(tree)\n tree = tree.left\n else:\n tree.right = TreeNode(v)\n stack.append(tree)\n tree = tree.right\n return stack[0] if stack else tree\n\n def print_tree(self, tree):\n\n from collections import deque\n q = deque([tree])\n values = []\n while q:\n t = q.popleft()\n values.append(t.val)\n if t.left:\n q.append(t.left)\n if t.right:\n q.append(t.right)\n return values\n\n\nif __name__ == '__main__':\n pre = [1,2,4,5,3,6,7]\n post = [4,5,2,6,7,3,1]\n x = (Solution().\n constructFromPrePost(pre, post))\n print(Solution().print_tree(x))\n","sub_path":"leetcode/889_construst_bt_with_pre_post.py","file_name":"889_construst_bt_with_pre_post.py","file_ext":"py","file_size_in_byte":1376,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"453592010","text":"import pickle\nimport re\nimport time\nimport sys\nimport datetime\n\nENWIK_FILENAME = \"../data/enwik9\"\nNUMBER_OF_LINES = 13147026\nMIN_FREQ_TO_BE_A_WORD = 500\nMIN_FREQ_TO_BE_A_COMBINED_WORD = 1000\nDISPLAY_CONTROL = 2000\n\n\n# Back up the reference to the exceptionhook\nsys._excepthook = sys.excepthook\n\n\ndef my_exception_hook(exctype, value, traceback):\n # Print the error and traceback\n print(exctype, value, traceback)\n # Call the normal Exception hook after\n sys._excepthook(exctype, value, traceback)\n sys.exit(1)\n\n\n# Set the exception hook to our wrapping function\nsys.excepthook = my_exception_hook\n\n\nclass Node:\n character: str\n encoded_string: str\n frequency: int\n children: []\n\n def __init__(self, character: str, frequency: int):\n self.character = character\n self.frequency = frequency\n self.children = []\n self.encoded_string = \"\"\n\n\ndef convert_freq_map_to_huffman_map(final_word_nodes_dict, fileName=\"tmp\"):\n print(\"Converting the final dict into a list of nodes\")\n\n word_nodes_list = []\n for key, value in final_word_nodes_dict.items():\n new_node = Node(key, value)\n word_nodes_list.append(new_node)\n\n word_nodes_list.sort(key=lambda x: x.frequency, reverse=False)\n\n tmp_words_nodes_list = []\n for word in word_nodes_list:\n tmp_words_nodes_list.append(word)\n\n tmp_words_nodes_list.sort(key=lambda x: x.frequency, reverse=True)\n\n with open(fileName, \"w\", encoding=\"utf8\") as f:\n for node_tmp_1 in tmp_words_nodes_list:\n f.write(node_tmp_1.character + \" - \" +\n str(node_tmp_1.frequency) + \"\\n\")\n\n word_huffman_tree = []\n for value in word_nodes_list:\n word_huffman_tree.append(value)\n\n print(\"Iterating and merging the nodes until only one remains\")\n if len(word_huffman_tree) > 1:\n while len(word_huffman_tree) > 1:\n huffman_iteration(word_huffman_tree)\n else:\n if len(word_huffman_tree) == 1:\n node1 = word_huffman_tree.pop(0)\n root_node = Node(\"root\", 1)\n root_node.children.append(node1)\n word_huffman_tree.append(root_node)\n\n encode_the_node(word_huffman_tree[0])\n\n result = {}\n for node_tmp in tmp_words_nodes_list:\n result[node_tmp.character] = node_tmp.encoded_string\n\n return result\n\n\ndef huffman_iteration(huffman_tree_to_iterate):\n print(\"Sorting the nodes in the tree. Right now there are \" +\n str(len(huffman_tree_to_iterate)) + \" nodes.\")\n huffman_tree_to_iterate.sort(key=lambda x: x.frequency, reverse=False)\n node1 = huffman_tree_to_iterate.pop(0)\n node2 = huffman_tree_to_iterate.pop(0)\n print(\"Creating a new node with characters \" + node1.character + node2.character + \" and string \" + str(\n node1.frequency + node2.frequency))\n new_node = Node(node1.character + node2.character,\n node1.frequency + node2.frequency)\n new_node.children.append(node1)\n new_node.children.append(node2)\n huffman_tree_to_iterate.append(new_node)\n print(\"Sorting the nodes in the tree. Right now there are \" +\n str(len(huffman_tree_to_iterate)) + \" nodes.\")\n huffman_tree_to_iterate.sort(key=lambda x: x.frequency, reverse=False)\n\n\ndef encode_the_node(node):\n print(\"The current node to encode is \" +\n str(node.frequency) + \"-\" + node.character)\n if node.children is not None and 0 < len(node.children):\n print(\"Encoding the node \" + str(node.frequency) + \"-\" + node.character)\n print(\"Now encoding the first child of the node + \" +\n str(node.frequency) + \"-\" + node.character)\n node.children[0].encoded_string = node.encoded_string + \"0\"\n encode_the_node(node.children[0])\n if len(node.children) > 1:\n print(\"Now encoding the second child of the node + \" +\n str(node.frequency) + \"-\" + node.character)\n node.children[1].encoded_string = node.encoded_string + \"1\"\n encode_the_node(node.children[1])\n else:\n print(\"Nothing to encode in this node as there are either no children\")\n\n\ndef print_a_node(node):\n string_to_print = str(node.frequency) + \"-\" + \\\n node.character + \"-\" + node.encoded_string\n if node.children is not None and 0 < len(node.children):\n string_to_print += \"-> \\n\\t\"\n string_to_print += print_a_node(node.children[0]) + \" , \"\n if len(node.children) > 1:\n string_to_print += print_a_node(node.children[1])\n return string_to_print\n\n\ndef print_a_list(nodes_list_to_print):\n print(\"printing a list\")\n for node_to_print in nodes_list_to_print:\n print(print_a_node(node_to_print))\n\n\ndef find_all_indexes(input_str, search_str):\n l1 = []\n length = len(input_str)\n index = 0\n while index < length:\n i = input_str.find(search_str, index)\n if i == -1:\n return l1\n l1.append(i)\n index = i + 1\n return l1\n\n\nstart_time = time.time()\n\nhuffman_combined_words = {}\nwith open(\"D:/dataCompression/tmp/enwik8_new_strucure_freq_distro_combined_words\", 'rb') as f:\n huffman_combined_words = pickle.load(f)\n\ncombined_words_helper = {}\n\nfor key, value in huffman_combined_words.items():\n words_in_line = re.findall(r'\\w+', key)\n for word in words_in_line:\n if word in combined_words_helper:\n combined_words_helper[word].append(key)\n else:\n combined_words_helper[word] = [key]\n\nhuffman_map_words = {}\nwith open(\"D:/dataCompression/tmp/enwik8_new_strucure_freq_distro_words\", 'rb') as f:\n huffman_map_words = pickle.load(f)\n\nhuffman_map = {}\nwith open(\"D:/dataCompression/tmp/enwik8_new_strucure_freq_distro\", 'rb') as f:\n huffman_map = pickle.load(f)\n\n\nprint(\"Reading the dicts is complete, now creating the new structure.\")\n\nfinal_map = {}\nfinal_map_words = {}\nfinal_map_combined_words = {}\n\ncount = 0\ncurrent_word = None\nwith open(ENWIK_FILENAME, \"r\", encoding=\"utf-8\") as f:\n while True:\n c = f.readline()\n if count % DISPLAY_CONTROL == 0:\n print(\"--- %s seconds ---\" % (time.time() - start_time))\n # print(str(len(huffman_map_words)))\n print(\"Compressing - \" + str((count * 100) / NUMBER_OF_LINES))\n now = datetime.datetime.now()\n print(now.strftime(\"%Y-%m-%d %H:%M:%S\"))\n count = count + 1\n\n if not c:\n print(\"End of file. writing whatever is left\")\n break\n\n line_all_words = {}\n words_in_line = re.findall(r'\\w+', c)\n for word in words_in_line:\n if word in huffman_map_words:\n line_all_words[word] = huffman_map_words[word]\n\n line_words_pos_dict = {}\n for key, value in line_all_words.items():\n cursor = 0\n index: int = c.find(key, cursor)\n while index != -1:\n if index in line_words_pos_dict.keys():\n if len(line_words_pos_dict[index]) < len(key):\n line_words_pos_dict[index] = key\n else:\n line_words_pos_dict[index] = key\n cursor = cursor + len(key)\n index: int = c.find(key, cursor)\n\n combined_words_helper_client = {}\n for word in words_in_line:\n if word in combined_words_helper:\n combined_word = combined_words_helper[word]\n list_of_combined_word = combined_words_helper[word]\n for combined_word in list_of_combined_word:\n combined_words_helper_client[combined_word] = huffman_combined_words[combined_word]\n\n for key, value in combined_words_helper_client .items():\n cursor = 0\n index: int = c.find(key, cursor)\n while index != -1:\n if index in line_words_pos_dict:\n if len(line_words_pos_dict[index]) < len(key):\n line_words_pos_dict[index] = key\n else:\n line_words_pos_dict[index] = key\n cursor = cursor + len(key)\n index: int = c.find(key, cursor)\n\n iter_index = 0\n while iter_index < len(c):\n new_word = None\n if iter_index in line_words_pos_dict:\n new_word = line_words_pos_dict[iter_index]\n iter_index = iter_index + len(line_words_pos_dict[iter_index])\n else:\n new_word = c[iter_index]\n iter_index = iter_index + 1\n\n if current_word is None:\n current_word = new_word\n\n map_to_use = final_map\n if len(new_word) > 1:\n if new_word in huffman_combined_words:\n map_to_use = final_map_combined_words\n else:\n map_to_use = final_map_words\n map_to_use[new_word] = {}\n else:\n\n map_to_use = final_map\n if len(new_word) > 1:\n if new_word in huffman_combined_words:\n map_to_use = final_map_combined_words\n else:\n map_to_use = final_map_words\n if new_word not in map_to_use:\n map_to_use[new_word] = {}\n\n map_to_use = final_map\n if len(current_word) > 1:\n if current_word in huffman_combined_words:\n map_to_use = final_map_combined_words\n else:\n map_to_use = final_map_words\n if new_word not in map_to_use[current_word]:\n map_to_use[current_word][new_word] = 1\n else:\n map_to_use[current_word][new_word] = map_to_use[current_word][new_word] + 1\n\n current_word = new_word\n\nkeys_to_delete = {}\nkeys_to_add = {}\n\nfor key, value in final_map_combined_words.items():\n total_freq = 0\n for k, v in value.items():\n total_freq = total_freq + v\n if total_freq < MIN_FREQ_TO_BE_A_COMBINED_WORD:\n for character_t in key:\n if character_t not in final_map:\n final_map[character_t] = {}\n keys_to_delete[key] = {}\n else:\n for k, v in value.items():\n if v > MIN_FREQ_TO_BE_A_COMBINED_WORD:\n keys_to_add[key + k] = {}\n\nfor k, v in keys_to_delete.items():\n del final_map_combined_words[k]\n\nfor k, v in keys_to_add.items():\n final_map_combined_words[k] = {}\n\nkeys_to_delete = {}\nkeys_to_add = {}\n\nfor key, value in final_map_words.items():\n total_freq = 0\n for k, v in value.items():\n total_freq = total_freq + v\n if total_freq < MIN_FREQ_TO_BE_A_WORD:\n for character_t in key:\n if character_t not in final_map:\n final_map[character_t] = {}\n keys_to_delete[key] = {}\n else:\n for k, v in value.items():\n if v > MIN_FREQ_TO_BE_A_COMBINED_WORD:\n final_map_combined_words[key + k] = {}\n\nfor k, v in keys_to_delete.items():\n del final_map_words[k]\n\nwith open(\"D:/dataCompression/tmp/enwik8_new_strucure_freq_distro\", 'wb') as f:\n # Pickle the 'data' dictionary using the highest protocol available.\n pickle.dump(final_map, f, pickle.HIGHEST_PROTOCOL)\n\nwith open(\"D:/dataCompression/tmp/enwik8_new_strucure_freq_distro_words\", 'wb') as f:\n # Pickle the 'data' dictionary using the highest protocol available.\n pickle.dump(final_map_words, f, pickle.HIGHEST_PROTOCOL)\n\nwith open(\"D:/dataCompression/tmp/enwik8_new_strucure_freq_distro_combined_words\", 'wb') as f:\n # Pickle the 'data' dictionary using the highest protocol available.\n pickle.dump(final_map_combined_words, f, pickle.HIGHEST_PROTOCOL)\n\nprint(\"--- %s seconds ---\" % (time.time() - start_time))\n","sub_path":"src3/3. model-contd-contd.py","file_name":"3. model-contd-contd.py","file_ext":"py","file_size_in_byte":11850,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"166565439","text":"import argparse\nimport sys\nimport os\nfrom pong_testbench import PongTestbench\nfrom matplotlib import font_manager\nimport importlib\nimport gym\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"dir1\", type=str, help=\"Directory to agent 1 to be tested.\")\nparser.add_argument(\"dir2\", type=str, default=None, nargs=\"?\",\n help=\"Directory to agent 2 to be tested. If empty, SimpleAI is used instead.\")\nparser.add_argument(\"--render\", \"-r\", action=\"store_true\", help=\"Render the competition.\")\nparser.add_argument(\"--games\", \"-g\", type=int, default=100, help=\"number of games.\")\n\nargs = parser.parse_args()\n\nenv = gym.make(\"WimblepongVisualMultiplayer-v0\")\n\nn_actions = env.action_space.n\nstate_space_dim = env.observation_space.shape\n\nsys.path.insert(0, args.dir1)\nimport agent_with_history as agent\n\n\nfor weights in [\"1000\", \"2000\", \"3000\", \"4000\", \"5000\"]:\n win_tot1 = 0\n win_tot2 = 0\n for _ in range(10):\n\n orig_wd = os.getcwd()\n os.chdir(args.dir1)\n agent1 = agent.Agent(state_space_dim, n_actions)\n agent1.load_model(\"weights_DQN_\"+weights)\n os.chdir(orig_wd)\n #del sys.path[0]\n\n if args.dir2:\n sys.path.insert(0, args.dir2)\n importlib.reload(agent)\n os.chdir(args.dir2)\n agent2 = agent.Agent()\n agent2.load_model()\n os.chdir(orig_wd)\n del sys.path[0]\n else:\n agent2 = None\n\n testbench = PongTestbench(args.render)\n testbench.init_players(agent1, agent2)\n wins1, wins2 = testbench.run_test(args.games)\n win_tot1 += wins1\n win_tot2 += wins2\n print(\"Average victory rate: \", win_tot1 / (win_tot1+win_tot2), \" of: \", weights)\n","sub_path":"Pong/wimblepong/test_scripts/test_DQN.py","file_name":"test_DQN.py","file_ext":"py","file_size_in_byte":1738,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"464517232","text":"from math import pi, sqrt\nimport os\n\nIMAGE_WIDTH = 800\nIMAGE_HEIGHT = 600\n\nTHETA = pi / 3.0 \t\t\t\t\t# Angle from one point to the next\nRADIUS = 40 \t\t\t\t\t\t# Size of a hex\n\nHEXES_WIDE = int(IMAGE_WIDTH / RADIUS) # How many hexes in a row\nHEXES_HIGH = int(IMAGE_HEIGHT / RADIUS) + 1 # How many rows of hexes\n\nHALF_RADIUS = RADIUS / 2.0\nHALF_HEX_HEIGHT = sqrt(RADIUS ** 2 - HALF_RADIUS ** 2)\n\nuser_home = os.path.expanduser('~')\nUSER_PREFERENCE_FILE = user_home + \"/\" + \".sim_preferences.json\"\n\nBACKGROUND_BASE = \"background_base.png\"\nBACKGROUND_IMAGE = \"background.png\"\n\nNP_GROUND_TRUTH = \"/tmp/magnetic_ground_truth.np\"\n\nNUMBER_OF_ROBOTS = 3\nBATTERY_AUTONOMY = 1000\n\nLINE_WIDTH = 20\n\nROBOT_BASE_HEIGHT = 40\n\nDEBUG = True\n","sub_path":"sim/config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":717,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"374208546","text":"import telebot\nimport config\nimport os\nimport db_handler\n\nbot = telebot.TeleBot(config.token)\n\nkeyboard_start = telebot.types.ReplyKeyboardMarkup(True, True)\nkeyboard_start.row('Сборки SBC', 'Подборки составов')\nkeyboard_start.row('Трейды и инвестиции', 'Викторина FUT')\nkeyboard_start.row('Техподдержка')\n\nkeyboard_sbc = telebot.types.ReplyKeyboardMarkup(True, True)\nkeyboard_sbc.row('LIVE', 'ПРОДВИНУТЫЕ')\nkeyboard_sbc.row('ОСНОВНЫЕ', 'КУМИРЫ')\nkeyboard_sbc.row('Назад к выбору раздела')\n\nkeyboard_sbc_live = telebot.types.ReplyKeyboardMarkup(True, True)\nkeyboard_sbc_live.row(db_handler.sbc_live_name1)\nkeyboard_sbc_live.row(db_handler.sbc_live_name2, db_handler.sbc_live_name3)\nkeyboard_sbc_live.row(db_handler.sbc_live_name4, db_handler.sbc_live_name5)\nkeyboard_sbc_live.row(db_handler.sbc_live_name6, db_handler.sbc_live_name7)\nkeyboard_sbc_live.row('Назад к выбору раздела SBC')\n\nkeyboard_sbc_prod = telebot.types.ReplyKeyboardMarkup(True, True)\nkeyboard_sbc_prod.row(db_handler.sbc_prod_name1, db_handler.sbc_prod_name2)\nkeyboard_sbc_prod.row(db_handler.sbc_prod_name3, db_handler.sbc_prod_name4)\nkeyboard_sbc_prod.row('Назад к выбору раздела SBC')\n\nkeyboard_sbc_osn = telebot.types.ReplyKeyboardMarkup(True, True)\nkeyboard_sbc_osn.row(db_handler.sbc_osn_name1, db_handler.sbc_osn_name2)\nkeyboard_sbc_osn.row(db_handler.sbc_osn_name3, db_handler.sbc_osn_name4)\nkeyboard_sbc_osn.row('Назад к выбору раздела SBC')\n\nkeyboard_sbc_kum = telebot.types.ReplyKeyboardMarkup(True)\nkeyboard_sbc_kum.row('Назад к выбору раздела SBC')\n\nkeyboard_commands_price = telebot.types.ReplyKeyboardMarkup(True, True)\nkeyboard_commands_price.row('100K', '300K')\nkeyboard_commands_price.row('500K', '1KK')\nkeyboard_commands_price.row('Назад к выбору раздела')\n\nkeyboard_fut = telebot.types.ReplyKeyboardMarkup(True)\nkeyboard_fut.row('ДА', 'НЕТ')\n\nkeyboard_trade = telebot.types.ReplyKeyboardMarkup(True)\nkeyboard_trade.row('Назад к выбору раздела')\n\n\n\n\n@bot.message_handler(commands=['start'])\ndef start_message(message):\n\tbot.send_message(message.chat.id, 'Приветствую тебя, дорогой фифер!')\n\tbot.send_message(message.chat.id, 'Выбери интерсующий тебя раздел', reply_markup=keyboard_start)\n\n@bot.message_handler(content_types=['text'])\ndef sbc(message):\n\tif message.text == 'Техподдержка':\n\t\tbot.send_message(message.chat.id, '@junior_developer7777', reply_markup=keyboard_start)\n\tif message.text == 'Сборки SBC':\n\t\tbot.send_message(message.chat.id, 'Выбери интерсующий раздел SBC', reply_markup=keyboard_sbc)\n\tif message.text == 'Назад к выбору раздела':\n\t\tbot.send_message(message.chat.id, 'Выбери интерсующий тебя раздел', reply_markup=keyboard_start)\n\n\tif message.text == 'LIVE':\n\t\tbot.send_message(message.chat.id, 'Выбери интерсующий SBC', reply_markup=keyboard_sbc_live)\n\n\tif message.text == 'ПРОДВИНУТЫЕ':\n\t\tbot.send_message(message.chat.id, 'Выбери интерсующий SBC', reply_markup=keyboard_sbc_prod)\n\n\tif message.text == 'ОСНОВНЫЕ':\n\t\tbot.send_message(message.chat.id, 'Выбери интерсующий SBC', reply_markup=keyboard_sbc_osn)\n\n\tif message.text == 'КУМИРЫ':\n\t\tbot.send_message(message.chat.id, 'test', reply_markup=keyboard_sbc_kum)\n\tif message.text == 'Назад к выбору раздела SBC':\n\t\tbot.send_message(message.chat.id, 'Выбери интерсующий раздел SBC', reply_markup=keyboard_sbc)\n\n\tif message.text == 'Подборки составов':\n\t\tbot.send_message(message.chat.id, 'Выберу цену состава', reply_markup=keyboard_commands_price)\n\n\tif message.text == 'Викторина FUT':\n\t\tbot.send_message(message.chat.id, 'Ты готов начать викторину?', reply_markup=keyboard_fut)\n\n\tif message.text == 'НЕТ':\n\t\tbot.send_message(message.chat.id, 'Выбери интерсующий тебя раздел', reply_markup=keyboard_start)\n\n\tif message.text == 'Трейды и инвестиции':\n\t\tf = open(db_handler.trade1, 'rb')\n\t\tbot.send_photo(message.chat.id, f)\n\t\tbot.send_message(message.chat.id, db_handler.txt1, reply_markup=keyboard_trade)\n\n# if message.text == 'ДА':\n\n\nif __name__ == '__main__':\n\tbot.polling(none_stop=True)\n\n\n\n\n\n\n","sub_path":"bot.py","file_name":"bot.py","file_ext":"py","file_size_in_byte":4555,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"45996071","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Nov 9 15:33:46 2018\n\n@author: Mark Barbet\n\"\"\"\n\nimport cantera as ct\nfrom .. import simulation as sim\nfrom ...cti_core import cti_processor as ctp\nimport pandas as pd\nimport numpy as np\nimport time\n\nclass JSR_steadystate(sim.Simulation):\n \n '''child class of sim.Simulaton. Inherits all attributes and methods including __init__(). \n Also has internal init due to data requirements'''\n \n \n \n \n \n def __init__(self,pressure:float,temperature:float,observables:list,\n kineticSens:int,physicalSens:int,conditions:dict,thermalBoundary,mechanicalBoundary,\n processor:ctp.Processor=None,cti_path=\"\", \n save_physSensHistories=0,moleFractionObservables:list=[],\n absorbanceObservables:list=[],concentrationObservables:list=[],\n fullParsedYamlFile:dict={},residence_time:float=1.0,pvalveCoefficient:float=0.01,maxpRise:float=0.001):\n \n sim.Simulation.__init__(self,pressure,temperature,observables,kineticSens,physicalSens,\n conditions,processor,cti_path)\n self.thermalBoundary = thermalBoundary\n self.mechanicalBoundary = mechanicalBoundary\n self.kineticSensitivities= None\n self.experimentalData = None\n self.concentrationObservables = concentrationObservables\n self.moleFractionObservables = moleFractionObservables\n self.absorbanceObservables = absorbanceObservables\n self.fullParsedYamlFile = fullParsedYamlFile\n self.pvalveCoefficient=pvalveCoefficient\n self.maxPrise=maxpRise\n self.energycon='off'\n self.residence_time=residence_time\n \n if save_physSensHistories == 1:\n self.physSensHistories = []\n self.setTPX()\n self.dk = 0.01\n self.rtol=1e-6\n self.atol=1e-6\n self.solution=None\n \n \n def set_geometry(self,volume=0.1):\n self.reactor_volume=volume\n \n \n def printVars(self):\n print()\n \n def settingJSRConditions(self):\n '''\n Determine the mechanical and thermal boundary conditions for a \n shock tube.\n '''\n \n #assigning the thermal boundary variable\n if re.match('[aA]diabatic',self.thermalBoundary):\n energy = 'on'\n elif re.match('[iI]sothermal',self.thermalBoundary):\n energy = 'off'\n else:\n raise Exception('Please specify a thermal boundary condition, adiabatic or isothermal')\n #assigning the mehcanical boundary variable \n if re.match('[Cc]onstant [Pp]ressure',self.mechanicalBoundary):\n mechBoundary = 'constant pressure'\n elif re.match('[Cc]onstant [Vv]olume',self.mechanicalBoundary):\n mechBoundary = 'constant volume'\n else:\n raise Exception('Please specifiy a mehcanical boundary condition, constant pressure or constant volume')\n #return the thermal and mechanical boundary of the shock tube \n return energy,mechBoundary\n \n \n \n def run(self):\n \n gas=self.processor.solution\n reactorPressure=gas.P\n pressureValveCoefficient=self.pvalveCoefficient\n maxPressureRiseAllowed=self.maxPrise\n \n \n #Build the system components for JSR\n fuelAirMixtureTank=ct.Reservoir(gas)\n exhaust=ct.Reservoir(gas)\n \n stirredReactor=ct.IdealGasReactor(gas,energy=self.energycon,volume=self.reactor_volume)\n massFlowController=ct.MassFlowController(upstream=fuelAirMixtureTank,\n downstream=stirredReactor,mdot=stirredReactor.mass/self.residence_time)\n pressureRegulator=ct.Valve(upstream=stirredReactor,downstream=exhaust,K=pressureValveCoefficient)\n reactorNetwork=ct.ReactorNet([stirredReactor])\n \n if bool(self.observables) and self.kineticSens==1:\n for i in range(gas.n_reactions):\n stirredReactor.add_sensitivity_reaction(i)\n \n if self.kineticSens and bool(self.observables)==False:\n #except:\n print('Please supply a non-empty list of observables for sensitivity analysis or set kinetic_sens=0')\n if self.physicalSens==1 and bool(self.observables)==False:\n #except:\n print('Please supply a non-empty list of observables for sensitivity analysis or set physical_sens=0')\n \n # now compile a list of all variables for which we will store data\n columnNames = [stirredReactor.component_name(item) for item in range(stirredReactor.n_vars)]\n columnNames = ['pressure'] + columnNames\n\n # use the above list to create a DataFrame\n timeHistory = pd.DataFrame(columns=columnNames)\n\n # Start the stopwatch\n tic = time.time()\n \n \n #Establish a matrix to hold sensitivities for kinetic parameters, along with tolerances\n if self.kineticSens==1 and bool(self.observables):\n #senscolumnNames = ['Reaction']+observables \n senscolumnNames = self.observables\n #sensArray = pd.DataFrame(columns=senscolumnNames)\n #senstempArray = np.zeros((gas.n_reactions,len(observables)))\n dfs = [pd.DataFrame() for x in range(len(self.observables))]\n tempArray = [np.zeros(self.processor.solution.n_reactions) for x in range(len(self.observables))]\n \n reactorNetwork.rtol_sensitivity = self.rtol\n reactorNetwork.atol_sensitivity = self.atol \n \n \n reactorNetwork.advance_to_steady_state()\n final_pressure=stirredReactor.thermo.P\n sens=reactorNetwork.sensitivities()\n if self.kineticSens==1 and bool(self.observables):\n for k in range(len(self.moleFractionObservables)):\n dfs[k] = dfs[k].append(((pd.DataFrame(sens[k,:])).transpose()),ignore_index=True)\n toc = time.time()\n print('Simulation Took {:3.2f}s to compute'.format(toc-tic))\n \n columnNames = []\n #Store solution to a solution array\n #for l in np.arange(stirredReactor.n_vars):\n #columnNames.append(stirredReactor.component_name(l))\n columnNames=[stirredReactor.component_name(item) for item in range(stirredReactor.n_vars)]\n #state=stirredReactor.get_state()\n state=np.hstack([stirredReactor.mass, \n stirredReactor.volume, stirredReactor.T, stirredReactor.thermo.X])\n data=pd.DataFrame(state).transpose()\n data.columns=columnNames\n pressureDifferential = timeHistory['pressure'].max()-timeHistory['pressure'].min()\n if(abs(pressureDifferential/reactorPressure) > maxPressureRiseAllowed):\n #except:\n print(\"WARNING: Non-trivial pressure rise in the reactor. Adjust K value in valve\")\n \n if self.kineticSens==1:\n numpyMatrixsksens = [dfs[dataframe].values for dataframe in range(len(dfs))]\n self.kineticSensitivities = np.dstack(numpyMatrixsksens)\n self.solution=data\n return (self.solution,self.kineticSensitivities)\n else:\n return self.solution\n \n \n \n \n \nclass JSR_multiTemp_steadystate(sim.Simulation):\n \n def __init__(self,volume:float,pressure:float,temperatures:list,observables:list,\n kineticSens:int,physicalSens:int,conditions:dict,thermalBoundary,mechanicalBoundary,\n processor:ctp.Processor=None,cti_path=\"\", \n save_physSensHistories=0,moleFractionObservables:list=[],\n absorbanceObservables:list=[],concentrationObservables:list=[],\n fullParsedYamlFile:dict={},residence_time:float=1.0,pvalveCoefficient:float=0.01,maxpRise:float=0.001):\n \n# sim.Simulation.__init__(self,pressure,temperature,observables,kineticSens,physicalSens,\n# conditions,processor,cti_path)\n self.volume=volume\n self.temperatures=temperatures\n self.JSR_objects=[]\n for i in range(len(self.temperatures)):\n self.JSR_objects.append(JSR_steadystate(pressure,self.temperatures[i],observables,\n kineticSens,physicalSens,conditions,thermalBoundary,mechanicalBoundary,\n processor,cti_path, \n save_physSensHistories,moleFractionObservables,\n absorbanceObservables,concentrationObservables,\n fullParsedYamlFile,residence_time,pvalveCoefficient,maxpRise))\n \n \n \n def run(self):\n \n for i in range(len(self.JSR_objects)):\n self.JSR_objects[i].set_geometry(self.volume)\n \n self.JSR_objects[i].run()\n \n \n \n \n \n \n \n \n ","sub_path":"simulations/instruments/jsr_steadystate.py","file_name":"jsr_steadystate.py","file_ext":"py","file_size_in_byte":8964,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"206518762","text":"# --------------------------------------------------------------------------------------------\n# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License. See License.txt in the project root for license information.\n# --------------------------------------------------------------------------------------------\n\n\"\"\"\nThe module provides a client to connect to Azure Event Hubs.\n\n\"\"\"\n\nimport logging\n\nfrom proton import dispatch, Url, generate_uuid, DELEGATED\nfrom proton.reactor import Container, Selector\nfrom proton.handlers import Handler, EndpointStateHandler\nfrom proton.handlers import CFlowController, IncomingMessageHandler\n\n# pylint: disable=line-too-long\n# pylint: disable=C0111\n# pylint: disable=W0613\n\nclass EventHubClient(Container):\n \"\"\"\n The client to access the Event Hubs service.\n \"\"\"\n def __init__(self, address=None, *handlers, **kwargs):\n if not address:\n super(EventHubClient, self).__init__(**kwargs)\n else:\n super(EventHubClient, self).__init__(self, **kwargs)\n self.address = Url(address)\n self.receivers = []\n self.shared_connection = None\n self.shared_session = None\n\n def subscribe(self, handler, consumer_group, partition, offset, prefetch=300):\n \"\"\"\n Subscribes to an Event Hub partition to receive events.\n\n @param handler: handler to process the received event data. It must\n define an 'on_event_data' method. The event data object is an AMQP\n message (proton.Message). Besides the APIs from proton, the module\n provides utility methods in EventData class to access Event Hubs specific\n properties of the message.\n\n @param consumer_group: the consumer group the receiver belongs to\n\n @param partition: the id of the event hub partition\n\n @param offset: the offset to start receiving events. Use '-1' to start\n from the beginning, '@latest' from the end, and a checkpoint to resume\n previous processing\n\n @param prefetch: the number of events that will be proactively prefetched\n by the library into a local buffer queue\n\n \"\"\"\n source = \"%s/ConsumerGroups/%s/Partitions/%s\" % (self.address.path, consumer_group, partition)\n receiver = PartitionReceiver(handler, source, offset, prefetch)\n self.receivers.append(receiver)\n return self\n\n def session(self, context):\n if not self.shared_session:\n self.shared_session = context.session()\n self.shared_session.open()\n return self.shared_session\n\n def on_reactor_init(self, event):\n if not self.shared_connection:\n logging.info(\"Client starts address=%s\", self.address)\n self.shared_connection = self.connect(self.address, handler=self)\n self.shared_connection.__setattr__(\"_session_policy\", self)\n for receiver in self.receivers:\n receiver.start(self)\n\n def on_connection_local_open(self, event):\n logging.info(\"Connection local open host=%s\", event.connection.hostname)\n\n def on_connection_remote_open(self, event):\n logging.info(\"Connection remote open host=%s remote=%s\", event.connection.hostname, event.connection.remote_container)\n\n def on_session_local_open(self, event):\n logging.info(\"Session starts host=%s\", event.connection.hostname)\n\n def on_session_remote_open(self, event):\n logging.info(\"Session opened host=%s\", event.connection.hostname)\n\n def on_connection_remote_close(self, event):\n if EndpointStateHandler.is_local_closed(event.connection):\n return DELEGATED\n condition = event.connection.remote_condition\n if condition:\n logging.error(\"Connection closed by peer %s:%s %s\", condition.name, condition.description, event.connection.remote_container)\n else:\n logging.error(\"Connection closed by peer %s\", event.connection.remote_container)\n if self.shared_session:\n self.shared_session.close()\n self.shared_session.free()\n self.shared_session = None\n event.connection.close()\n self.shared_connection.free()\n self.shared_connection = None\n self.schedule(3.0, self)\n\n def on_session_remote_close(self, event):\n if EndpointStateHandler.is_local_closed(event.session):\n return DELEGATED\n event.session.close()\n event.session.free()\n self.shared_session = None\n condition = event.session.remote_condition\n if condition:\n logging.error(\"Session close %s:%s %s\", condition.name, condition.description, event.connection.remote_container)\n else:\n logging.error(\"Session close %s\", event.connection.remote_container)\n self.schedule(3.0, self)\n\n def on_timer_task(self, event):\n self.on_reactor_init(None)\n\nclass PartitionReceiver(Handler):\n \"\"\"\n The receiver to read events from an Event Hub partition.\n \"\"\"\n def __init__(self, delegate, source, offset, prefetch):\n super(PartitionReceiver, self).__init__()\n self.handlers = []\n if prefetch:\n self.handlers.append(CFlowController(prefetch))\n self.handlers.append(IncomingMessageHandler(True, self))\n self.fatal_conditions = [\"amqp:unauthorized-access\", \"amqp:not-found\"]\n self.delegate = delegate\n self.source = source\n self.offset = offset\n self.name = str(generate_uuid())[:8]\n self.iteration = 0\n self.client = None\n\n def start(self, client):\n self.client = client\n self.iteration += 1\n link_name = \"py-receiver-%s-%d\" % (self.name, self.iteration)\n selector = Selector(u\"amqp.annotation.x-opt-offset > '\" + self.offset + \"'\")\n client.create_receiver(client.shared_connection, self.source, name=link_name, handler=self, options=selector)\n\n def on_message(self, event):\n _message = event.message\n if self.delegate:\n dispatch(self.delegate, \"on_event_data\", _message)\n self.offset = EventData.offset(_message)\n\n def on_link_local_open(self, event):\n logging.info(\"Link starts. entity=%s offset=%s\", self.source, self.offset)\n\n def on_link_remote_open(self, event):\n logging.info(\"Link Opened. entity=%s offset=%s\", self.source, self.offset)\n\n def on_link_remote_close(self, event):\n if EndpointStateHandler.is_local_closed(event.link):\n return DELEGATED\n event.link.close()\n event.link.free()\n condition = event.link.remote_condition\n if condition:\n logging.error(\"Link detached %s:%s ref:%s\",\n condition.name,\n condition.description,\n event.connection.remote_container)\n else:\n logging.error(\"Link detached ref:%s\", event.connection.remote_container)\n if condition and condition.name in self.fatal_conditions:\n event.connection.close()\n else:\n self.client.schedule(3.0, self)\n\n def on_timer_task(self, event):\n self.start(self.client)\n\nclass EventData(object):\n \"\"\"\n A utility to read EventData properties from an AMQP message\n \"\"\"\n @classmethod\n def sequence_number(cls, message):\n \"\"\"\n Return the sequence number of the event data object.\n \"\"\"\n return message.annotations[\"x-opt-sequence-number\"]\n\n @classmethod\n def offset(cls, message):\n \"\"\"\n Return the offset of the event data object.\n \"\"\"\n return message.annotations[\"x-opt-offset\"]\n\n @classmethod\n def partition_key(cls, message):\n \"\"\"\n Return the partition key of the event data object.\n \"\"\"\n return message.annotations[\"x-opt-partition-key\"]\n","sub_path":"eventhubs/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":7877,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"169291749","text":"import numpy as np\nimport tensorflow as tf\nimport tensorflow_probability as tfp\ntfd = tfp.distributions\nfrom easydict import EasyDict as edict\n\nfrom .base import BaseModel\nfrom .actan import Flow\nfrom .pc_encoder import *\nfrom .flow.transforms import Transform\nfrom .tf_pc_distance.pc_distance import chamfer\nfrom .tf_pc_distance.pc_kde import kde\n\nclass ACSetFlow(BaseModel):\n def __init__(self, hps):\n self.prior_net = LatentEncoder(hps, name='prior')\n self.posterior_net = LatentEncoder(hps, name='posterior')\n if hps.use_peq_embed:\n self.peq_embed = SetXformer(hps)\n super(ACSetFlow, self).__init__(hps)\n \n def build_net(self):\n with tf.variable_scope('acset_flow', reuse=tf.AUTO_REUSE):\n self.x = tf.placeholder(tf.float32, [None, self.hps.set_size, self.hps.dimension])\n self.b = tf.placeholder(tf.float32, [None, self.hps.set_size, self.hps.dimension])\n self.m = tf.placeholder(tf.float32, [None, self.hps.set_size, self.hps.dimension])\n\n # build transform\n self.acflow = Flow(edict(self.hps.acflow_params))\n self.transform = Transform(edict(self.hps.trans_params))\n\n # prior\n prior_inputs = tf.concat([self.x*self.b, self.b], axis=-1)\n prior = self.prior_net(prior_inputs)\n prior_sample = prior.sample()\n prior_sample, _ = self.transform.inverse(prior_sample)\n\n # peq embedding\n cm = peq_embed = None\n if self.hps.use_peq_embed:\n peq_embed = self.peq_embed(prior_inputs)\n C = peq_embed.get_shape().as_list()[-1]\n cm = tf.reshape(peq_embed, [-1,C])\n \n # posterior\n posterior_inputs = tf.concat([self.x*self.m, self.m], axis=-1)\n posterior = self.posterior_net(posterior_inputs)\n posterior_sample = posterior.sample()\n\n # kl term\n z_sample, logdet = self.transform.forward(posterior_sample)\n logp = tf.reduce_sum(prior.log_prob(z_sample), axis=1) + logdet\n kl = tf.reduce_sum(posterior.entropy(), axis=1) + logp\n\n # generator p(x_u | x_o, z)\n x = tf.reshape(self.x, [-1,self.hps.dimension])\n b = tf.reshape(self.b, [-1,self.hps.dimension])\n m = tf.reshape(self.m, [-1,self.hps.dimension])\n cv = tf.reshape(tf.tile(tf.expand_dims(posterior_sample, axis=1), [1,self.hps.set_size,1]), [-1,self.hps.latent_dim])\n c = tf.concat([cv, cm], axis=-1)\n log_likel = self.acflow.cond_forward(x, c, b, m)\n log_likel = tf.reshape(log_likel, [-1,self.hps.set_size])\n self.set_metric = self.set_elbo = log_likel + tf.expand_dims(kl, axis=1) / self.hps.set_size\n log_likel = tf.reduce_mean(log_likel, axis=1)\n tf.summary.scalar('log_likel', tf.reduce_mean(log_likel))\n\n # elbo\n self.elbo = log_likel + kl / self.hps.set_size\n self.metric = self.elbo\n self.loss = tf.reduce_mean(-self.elbo)\n tf.summary.scalar('loss', self.loss)\n\n # sample\n x = tf.reshape(self.x, [-1,self.hps.dimension])\n b = tf.reshape(self.b, [-1,self.hps.dimension])\n m = tf.reshape(self.m, [-1,self.hps.dimension])\n cv = tf.reshape(tf.tile(tf.expand_dims(prior_sample, axis=1), [1,self.hps.set_size,1]), [-1,self.hps.latent_dim])\n c = tf.concat([cv, cm], axis=-1)\n sample = self.acflow.cond_inverse(x, c, b, m)\n self.sample = tf.reshape(sample, [-1, self.hps.set_size, self.hps.dimension])\n\n # cd, emd, kde\n if self.hps.lambda_cd > 0:\n cd_loss = chamfer(self.sample, self.x)\n self.loss = self.loss + cd_loss * self.hps.lambda_cd\n\n if self.hps.lambda_kde > 0:\n density = kde(self.sample, self.x, self.hps.kde_sigma)\n kde_loss = -tf.reduce_mean(density)\n self.loss = self.loss + kde_loss * self.hps.lambda_kde\n self.metric = density\n\n # compress\n log_likel = self.acflow.cond_forward(x, c, b, m)\n self.log_likel = tf.reshape(log_likel, [-1,self.hps.set_size])","sub_path":"models/pc_acset.py","file_name":"pc_acset.py","file_ext":"py","file_size_in_byte":4290,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"277704211","text":"import os\nimport sys\nfrom urllib.request import urlopen\nfrom myCrawler import *\nfrom parser import TopReviewHTMLParser\n\n\nclass ReviewCompiler:\n\n\n\tdef __init__(self, directory_name):\n\t\tReviewCompiler.directory_name = directory_name\n\n\n\t\t# List of all the recipe URLs that haven't been processed yet\n\t\tReviewCompiler.q = []\n\n\t\tReviewCompiler.objects_fname = ReviewCompiler.directory_name + \"/objects.tdf\"\n\t\tReviewCompiler.review_fname = ReviewCompiler.directory_name + \"/URLreviews.txt\" # pg 1 of paginated user's reviews\n\t\tReviewCompiler.proccessed_fname = ReviewCompiler.directory_name + \"/reviewed_recipes.txt\" # recipes from which top comment extracted\n\n\t\t# List of URLs for user reviews (pg1 of user reviews)\n\t\tinitFile(ReviewCompiler.review_fname) # If doesn't exist create\n\t\tReviewCompiler.review_URLs = set(URLs2Set(ReviewCompiler.review_fname))\n\n\t\tinitFile(ReviewCompiler.proccessed_fname) # If doesn't exist create\n\t\tReviewCompiler.crawled = set(URLs2Set(ReviewCompiler.proccessed_fname))\n\n\t\tself.loadRecipeURLs()\n\n\n\t@staticmethod\n\tdef loadRecipeURLs():\n\t\t\"\"\"\n\t\tFrom objects TDF file get all the recipe's URLs.\n\t\tAssuming processed_recipes_fname is valid. Check upstream.\n\t\t\"\"\"\n\t\tprint(ReviewCompiler.objects_fname)\n\t\tobjects_f = open(ReviewCompiler.objects_fname, \"r\")\n\t\tfor idx, line in enumerate(objects_f):\n\t\t\tlst = line.strip().split('\\t')\n\t\t\trecipe_url = lst[4]\n\t\t\tif recipe_url not in ReviewCompiler.crawled:\n\t\t\t\tReviewCompiler.q.append(recipe_url)\n\n\n\t@staticmethod\n\tdef crawl(): # page_url is the recipe page url\n\t\t\"\"\"\n\t\tTo be called by main driver:\n\t\t\twhile len(ReviewCompiler.recipe_q) != 0:\n\t\t\t\tReviewCompiler.crawl(recipe_page_URL)\n\t\t\"\"\"\n\t\tpage_url = ReviewCompiler.q[0]\n\n\t\tsys.stdout.write(\"Queue Size == \" + str(len(ReviewCompiler.q)) + \" \")\n\t\tsys.stdout.write(\"Crawled Size == \" + str(len(ReviewCompiler.crawled)) + \"\\n\")\n\n\t\ttopReviewer_URL = ReviewCompiler.getTopCommentUserURL(page_url)\n\t\tif topReviewer_URL in ReviewCompiler.review_URLs:\n\t\t\tprint(\"========== DUPLICATE TOP REVIEWER == %s ==============\" % (topReviewer_URL))\n\t\tReviewCompiler.review_URLs.add(topReviewer_URL)\n\t\tReviewCompiler.q.remove(page_url)\n\t\tReviewCompiler.crawled.add(page_url)\n\n\t\tReviewCompiler.updateFiles()\n\n\n\t@staticmethod\n\tdef getTopCommentUserURL(page_url):\n\t\thtmlString = ''\n\t\ttry:\n\t\t\tresponse = urlopen(page_url)\n\t\t\tif 'text/html' in response.getheader('Content-Type'):\n\t\t\t\thtmlBytes = response.read()\n\t\t\t\thtmlString = htmlBytes.decode(\"utf-8\")\n\t\t\tprint(\"Currently visiting: %s\" % page_url)\n\t\t\tfinder = TopReviewHTMLParser(page_url)\n\t\t\tfinder.feed(htmlString)\n\t\texcept:\n\t\t\treturn set()\n\n\t\treturn finder.getTopReviewer()\n\n\t@staticmethod\n\tdef updateFiles():\n\t\tset2URLs(ReviewCompiler.review_fname, ReviewCompiler.review_URLs)\n\t\tset2URLs(ReviewCompiler.proccessed_fname, ReviewCompiler.crawled)\n\n","sub_path":"yori_webcrawler/ReviewCompiler.py","file_name":"ReviewCompiler.py","file_ext":"py","file_size_in_byte":2798,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"112668113","text":"from django.conf.urls import url\nfrom userData import views\n\napp_name = 'userData'\n\nurlpatterns=[\n url(r'^registration/$',views.registration,name='registration'),\n url(r'^login/$',views.user_login,name='login'),\n url(r'^enterNumber/$',views.sendOtp,name='sendOtp'),\n]","sub_path":"userData/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":276,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"239685376","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Feb 21 09:30:20 2018\n@author: djabernathy\n\"\"\"\n\nimport osmnx as ox\nimport geopandas as gp\nimport math\nimport sys\nsys.path.append('/Users/djabernathy/Desktop/Geog5223/g5223project/')\nfrom downhill_slopeforce import *\nimport json\n\napiKey = 'AIzaSyAgGQRYnWTzsgf6Xn1DISE99Tdq7Lr9rwU'\nsaving_directory = 'C:/Users/G5223/Desktop/'\n\ncity = 'Milwaukee'\nstate = 'Wisconsin'\ncountry = 'USA'\nnetwork_type = 'drive'\nlocation_string = city + ', ' + state + ', ' + country\n\nfilter_list = ['motorway','motorway_link']\n\nfiltered_nodes_slow = set() # use set to avoid adding the same node multiple times\nfiltered_nodes_moderate = set() # nodes are added to these sets as they are found\nfiltered_nodes_fast = set()\n\ndef road_Filter(road, should_filter=True):\n highway_filters = []\n if(should_filter):\n #list your filters here\n highway_filters = filter_list\n for filter_item in highway_filters:\n if road['highway'] == filter_item:\n return False\n return True\n\n# download city street data and elevation data\nG = ox.graph_from_place(location_string, network_type= network_type) \nG_elev = ox.add_node_elevations(G, apiKey, max_locations_per_batch=350, pause_duration=0.02)\n\n# iterate through all nodes. Check elevation change between each node's\ncount = 0\ngeometry_count = 0\nfor n in G_elev.nodes():\n for neighbor in G_elev.neighbors(n): # neighbor is a node key\n road = G_elev.get_edge_data(n, neighbor)[0] # road length in meters, right?\n if road_Filter(road): # look up road types for other filters\n #if road['name'] == 'Lane Averue':\n # print(road)\n # neighbors. keep track of pairs with appropriate elevation change.\n change_elev = G_elev.node[n]['elevation'] - G_elev.node[neighbor]['elevation']\n\n # checks for change in elevation\n length = road['length']\n skateweight = 150\n TopHill_elevation = 552 #feet\n bottomHill_elevation = 10\n hillHeight = change_elev\n vel = skate_downhill_accel(length, skateweight, hillHeight, gravity=9.8)[2]\n print(vel, length)\n if vel > 3 or vel < -3:\n \n # check for curvature usng the linestring\n road['velocity'] = vel\n if 'geometry' in road.keys():\n #print(road['geometry'], '\\n\\n\\n\\n')\n print()\n \n else:\n #print(G_elev.get_edge_data(n, neighbor)[0])\n count += 1\n \n # if all conditions are passed, add nodes to set\n if vel < 20: \n filtered_nodes_slow.add(n)\n filtered_nodes_slow.add(neighbor)\n elif vel < 45: \n filtered_nodes_moderate.add(n)\n filtered_nodes_moderate.add(neighbor)\n else: \n filtered_nodes_fast.add(n)\n filtered_nodes_fast.add(neighbor)\n #print(change_elev)\n\n # graph of the filtered roads\n G_filtered_slow = G_elev.subgraph(filtered_nodes_slow)\n G_filtered_moderate = G_elev.subgraph(filtered_nodes_moderate)\n G_filtered_fast = G_elev.subgraph(filtered_nodes_fast)\n\n#ox.plot_graph(ox.project_graph(G))\n\n#This saves as a shapefile instead of JSON\n#ox.save_graph_shapefile(G, filename='Columbus', folder=None, encoding='utf-8')\n#ox.save_graph_shapefile(G_filtered, filename='Filtered Roads', folder=None, encoding='utf-8')\n\n# convert graph to GeoDataFrame. And then to JSON\ngdfFiltered_slow = ox.graph_to_gdfs(G_filtered_slow, nodes=False, edges=True, node_geometry=True, fill_edge_geometry=True)\ngeojsonFiltered_slow = gdfFiltered_slow.to_json()\n\ngdfFiltered_moderate = ox.graph_to_gdfs(G_filtered_moderate, nodes=False, edges=True, node_geometry=True, fill_edge_geometry=True)\ngeojsonFiltered_moderate = gdfFiltered_moderate.to_json()\n\ngdfFiltered_fast = ox.graph_to_gdfs(G_filtered_fast, nodes=False, edges=True, node_geometry=True, fill_edge_geometry=True)\ngeojsonFiltered_fast = gdfFiltered_fast.to_json()\n#geojsonFiltered = json.dumps(geojsonFiltered, sort_keys=True, indent=4)\n\nf = open(saving_directory + city + 'FilteredSlow.js', 'w')\nresult = 'var milwaukeeFilteredSlow =' + str(geojsonFiltered_slow)\nf.write(result)\nf.close()\n\nf = open(saving_directory + city + 'FilteredModerate.js', 'w')\nresult = 'var milwaukeeFilteredModerate =' + str(geojsonFiltered_moderate)\nf.write(result)\nf.close()\n\nf = open(saving_directory + city + 'FilteredFast.js', 'w')\nresult = 'var milwaukeeFilteredFast =' + str(geojsonFiltered_fast)\nf.write(result)\nf.close()\n","sub_path":"pythonStuff/getJSON.py","file_name":"getJSON.py","file_ext":"py","file_size_in_byte":4702,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"581677527","text":"# Put your commands here\nCOMMAND1 = \"who's the coolest in the world\"\n\n# Your handling code goes in this function\ndef handle_command(command):\n \"\"\"\n Determine if the command is valid. If so, take action and return\n a response, if necessary.\n \"\"\"\n response = \"\"\n if COMMAND1 in command:\n response = \"Al-kareem is the coolest in the world\"\n \n return response\n\n","sub_path":"slackbot_ce/code_em/al_kareem/slacklib.py","file_name":"slacklib.py","file_ext":"py","file_size_in_byte":400,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"163403472","text":"import datetime\nimport re\n\nfrom freenit.db import db\nfrom freenit.models.sql.user import User\nfrom unidecode import unidecode\n\nfrom peewee import BooleanField, DateTimeField, ForeignKeyField, TextField\n\nfrom ..date import datetime_format\n\n_punct_re = re.compile(r'[\\t !\"#$%&\\'()*\\-/<=>?@\\[\\\\\\]^_`{|},.]+')\nModel = db.Model\n\n\nclass Blog(Model):\n author = ForeignKeyField(User, backref='blogs')\n content = TextField()\n date = DateTimeField(\n formats=[datetime_format],\n default=datetime.datetime.utcnow\n )\n published = BooleanField()\n slug = TextField()\n title = TextField()\n\n def save(self, *args, **kwargs):\n if self.slug is None:\n result = []\n for word in _punct_re.split(self.title.lower()):\n result.extend(unidecode(word).split())\n self.slug = '-'.join(result)\n super(Blog, self).save(*args, **kwargs)\n\n @classmethod\n def find(cls, year, month, day, slug):\n intyear = int(year)\n intmonth = int(month)\n intday = int(day)\n startdate_query = Blog.select().where(\n Blog.date >= datetime.date(intyear,\n intmonth,\n intday)\n )\n enddate_query = startdate_query.where(\n Blog.date < datetime.date(intyear,\n intmonth,\n intday + 1)\n )\n query = enddate_query.where(Blog.slug == slug)\n if query.count() == 0:\n raise cls.DoesNotExist\n if query.count() > 1:\n raise ValueError('Too many instances')\n return query[0]\n","sub_path":"pyser/models/blog.py","file_name":"blog.py","file_ext":"py","file_size_in_byte":1679,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"625511297","text":"import json\n\nfrom django.contrib.auth.models import User\nfrom django.http import HttpResponse\nfrom django.shortcuts import render, redirect\nfrom smithy.views.ko_slogan import process, ko_api, differ, extraction\nfrom smithy.views.en_slogan import enslogan\nfrom smithy.views.ko_model import koslogan\n\n\ndef main_slogan(request):\n if request.method == \"POST\":\n return render(request, \"smithy/index.html\")\n else:\n return render(request, \"smithy/index.html\")\n\ndef loading_view(request):\n request.session[\"select\"] = request.POST.get(\"select\", None)\n request.session[\"info\"] = request.POST.get(\"info\", None)\n request.session[\"sim\"] = request.POST.get(\"sim\", None)\n\n # if request.method == 'POST':\n # form = DataForm(request.POST)\n # if form.is_valid():\n # temp = form.save()\n # return render(request, \"smithy/loading.html\")\n # else:\n # form = DataForm()\n # context = {'form' : form}\n\n return render(request, \"smithy/loading.html\")\n\n\ndef result(request):\n select = request.session[\"select\"]\n info = request.session[\"info\"]\n sim = request.session[\"sim\"]\n sim = int(sim)\n if select == \"ko_slogan\":\n text = process(info, sim)\n slogan_list = ko_api(text)\n kor_list = differ(slogan_list)\n total_slogan = extraction(kor_list, sim)\n context = {\"slogans\": total_slogan, \"select\": select, \"info\": info, \"sim\": sim}\n elif select == \"en_slogan\":\n slogans = enslogan(info)\n context = {\"slogans\": slogans, \"select\": select, \"info\": info, \"sim\": sim}\n else:\n slogans = koslogan(info)\n context = {\"slogans\": slogans, \"select\": select, \"info\": info, \"sim\": sim}\n\n return render(request, \"smithy/result_slogan.html\", context=context)\n\n","sub_path":"smithy/views/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1781,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"383076855","text":"\"\"\"rc URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n https://docs.djangoproject.com/en/2.1/topics/http/urls/\nExamples:\nFunction views\n 1. Add an import: from my_app import views\n 2. Add a URL to urlpatterns: path('', views.home, name='home')\nClass-based views\n 1. Add an import: from other_app.views import Home\n 2. Add a URL to urlpatterns: path('', Home.as_view(), name='home')\nIncluding another URLconf\n 1. Import the include() function: from django.urls import include, path\n 2. Add a URL to urlpatterns: path('blog/', include('blog.urls'))\n\"\"\"\nfrom django.contrib import admin\nfrom django.urls import path, include\nfrom .views import *\n\nfrom django.contrib.auth import views as auth_views\n\napp_name = 'operate'\n\nurlpatterns = [\n#### main URL ##########################################################\n path('main/', main, name='main'),\n#### 대시보드 URL #######################################################\n path('dashboard/', dashboard, name='dashboard'),\n#### 공지사항 관리 URL ################################################### \n path('notice/', manage_notice, name='notice'),\n path('notice/notice_add/', notice_edit, name='notice_add'),\n path('notice/notice_edit/', notice_edit, name='notice_edit'),\n path('notice/notice_activate/', notice_activate, name='notice_activate'),\n#### 발행 관리 URL ####################################################### \n path('publish/', publish, name='publish'),\n#### 거래 취소 관리 URL#######################################################\n path('cancel/', cancel, name='cancel'),\n#### 네트워크 관리 URL ###################################################\n path('network/', network, name='network'),\n#### 유저관리 URL ########################################################\n # path('users/', usersMyLV.as_view(), name='users'),\n path('users/', manageUser, name='users'),\n path('get_like/', get_like, name='get_like'),\n#### 가맹점 URL ##########################################################\n path('approval/', manage_store_approval, name='approval'),\n path('get_approval/', get_approval, name='get_approval'),\n##### 차트 URL ###########################################################\n path('stats/', manage_stats, name='stats'),\n path('stats/chartdata/', ChartData.as_view(), name='chartdata'),\n path('stats/west_stats/', west_stats, name='west_stats'),\n path('stats/north_stats/', north_stats, name='north_stats'),\n path('stats/wooleung_stats/', wooleung_stats, name='wooleung_stats'),\n##### login_required ####################################################\n path('login_required/', login_required, name='login_required'),\n path('admin_logout/', admin_logout, name='admin_logout'),\n]\n","sub_path":"RC/rccoin/operate/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":2834,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"164508559","text":"'''Author: Brandon Trabucco, Copyright 2019\nTest the image captioning model with some fake inputs.'''\n\n\nimport time\nimport itertools\nimport tensorflow as tf\nimport numpy as np\nfrom detailed_captioning.layers.image_captioner import ImageCaptioner\nfrom detailed_captioning.cells.show_attend_and_tell_cell import ShowAttendAndTellCell\nfrom detailed_captioning.utils import load_glove\nfrom detailed_captioning.utils import get_show_attend_and_tell_checkpoint \nfrom detailed_captioning.utils import list_of_ids_to_string\nfrom detailed_captioning.inputs.spatial_image_features_only import import_mscoco\n\n\nPRINT_STRING = \"\"\"\n({4:.2f} img/sec) iteration: {0:05d} loss: {1:.5f}\n caption: {2}\n actual: {3}\"\"\"\n\ntf.logging.set_verbosity(tf.logging.INFO)\ntf.flags.DEFINE_integer(\"num_epochs\", 100, \"\")\ntf.flags.DEFINE_integer(\"batch_size\", 32, \"\")\ntf.flags.DEFINE_boolean(\"is_mini\", False, \"\")\nFLAGS = tf.flags.FLAGS\n\n\ndef main(unused_argv):\n \n vocab, pretrained_matrix = load_glove(vocab_size=100000, embedding_size=300)\n with tf.Graph().as_default():\n\n image_id, spatial_features, input_seq, target_seq, indicator = import_mscoco(\n mode=\"train\", batch_size=FLAGS.batch_size, num_epochs=FLAGS.num_epochs, is_mini=FLAGS.is_mini)\n show_attend_and_tell_cell = ShowAttendAndTellCell(300)\n image_captioner = ImageCaptioner(show_attend_and_tell_cell, vocab, pretrained_matrix)\n logits, ids = image_captioner(lengths=tf.reduce_sum(indicator, axis=1), \n spatial_image_features=spatial_features, seq_inputs=input_seq)\n tf.losses.sparse_softmax_cross_entropy(target_seq, logits, weights=indicator)\n loss = tf.losses.get_total_loss()\n \n global_step = tf.train.get_or_create_global_step()\n optimizer = tf.train.AdamOptimizer()\n learning_step = optimizer.minimize(loss, var_list=image_captioner.variables, global_step=global_step)\n\n captioner_saver = tf.train.Saver(var_list=image_captioner.variables + [global_step])\n captioner_ckpt, captioner_ckpt_name = get_show_attend_and_tell_checkpoint()\n with tf.Session() as sess:\n \n sess.run(tf.variables_initializer(optimizer.variables()))\n if captioner_ckpt is not None:\n captioner_saver.restore(sess, captioner_ckpt)\n else:\n sess.run(tf.variables_initializer(image_captioner.variables + [global_step]))\n captioner_saver.save(sess, captioner_ckpt_name, global_step=global_step)\n last_save = time.time()\n \n for i in itertools.count():\n \n time_start = time.time()\n try:\n _target, _ids, _loss, _learning_step = sess.run([target_seq, ids, loss, learning_step])\n except:\n break\n \n iteration = sess.run(global_step)\n \n print(PRINT_STRING.format(\n iteration, _loss, \n list_of_ids_to_string(_ids[0, :].tolist(), vocab), \n list_of_ids_to_string(_target[0, :].tolist(), vocab), \n FLAGS.batch_size / (time.time() - time_start)))\n \n new_save = time.time()\n if new_save - last_save > 3600: # save the model every hour\n captioner_saver.save(sess, captioner_ckpt_name, global_step=global_step)\n last_save = new_save\n \n captioner_saver.save(sess, captioner_ckpt_name, global_step=global_step)\n print(\"Finishing training.\")\n \n\nif __name__ == \"__main__\":\n \n tf.app.run()\n ","sub_path":"scripts/train/train_show_attend_and_tell.py","file_name":"train_show_attend_and_tell.py","file_ext":"py","file_size_in_byte":3718,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"560428476","text":"# -*- coding: utf-8 -*-\n\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\n\nimport ipaddr\n\nfrom ralph.discovery.models import Network\nfrom ralph.scan.models import ScanSummary\n\n\ndef find_network(network_spec):\n \"\"\"Returns network object by network address.\"\"\"\n try:\n address = str(ipaddr.IPNetwork(network_spec))\n except ValueError:\n network = Network.objects.get(name=network_spec)\n else:\n network = Network.objects.get(address=address)\n return network\n\n\ndef update_scan_summary(job):\n try:\n scan_summary = ScanSummary.objects.get(job_id=job.id)\n except ScanSummary.DoesNotExist:\n return\n else:\n scan_summary.previous_checksum = job.meta.get(\n 'results_checksum',\n )\n scan_summary.false_positive_checksum = None\n scan_summary.save()\n job.meta['changed'] = False\n job.save()\n","sub_path":"src/ralph/scan/util.py","file_name":"util.py","file_ext":"py","file_size_in_byte":986,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"264194365","text":"from datetime import datetime\nfrom opengever.maintenance.debughelpers import setup_app\nfrom opengever.maintenance.debughelpers import setup_option_parser\nfrom opengever.maintenance.debughelpers import setup_plone\nfrom opengever.maintenance.utils import elevated_privileges\nfrom plone import api\nimport logging\nimport sys\nimport transaction\n\n\nlogger = logging.getLogger('opengever.maintenance')\nhandler = logging.StreamHandler(stream=sys.stdout)\nlogging.root.addHandler(handler)\nlogging.root.setLevel(logging.INFO)\n\nSEPARATOR = '-' * 78\n\n\ndef move_documents_in_proposal_to_dossier(options):\n \"\"\"Move documents placed in a proposal to their parent dossier.\n\n \"\"\"\n\n catalog = api.portal.get_tool('portal_catalog')\n brains = catalog.unrestrictedSearchResults(\n portal_type='opengever.meeting.proposal')\n\n with elevated_privileges():\n for brain in brains:\n proposal = brain.getObject()\n document_brains = catalog.unrestrictedSearchResults(\n path=brain.getPath(),\n portal_type='opengever.document.document')\n for document_brain in document_brains:\n document = document_brain.getObject()\n dossier = proposal.get_containing_dossier()\n logger.info(\"moving document {} to dossier {}\".format(\n document_brain.getPath(),\n \"/\".join(dossier.getPhysicalPath())))\n api.content.move(source=document, target=dossier)\n\n if not options.dry_run:\n logger.info(\"committing...\")\n transaction.commit()\n logger.info(\"done.\")\n\n\ndef main():\n app = setup_app()\n\n parser = setup_option_parser()\n parser.add_option(\"-n\", dest=\"dry_run\", action=\"store_true\", default=False)\n (options, args) = parser.parse_args()\n\n logger.info(SEPARATOR)\n logger.info(\"Date: {}\".format(datetime.now().isoformat()))\n setup_plone(app, options)\n\n if options.dry_run:\n transaction.doom()\n logger.info(\"DRY-RUN\")\n\n move_documents_in_proposal_to_dossier(options)\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"opengever/maintenance/scripts/move_documents_in_proposal_to_dossier.py","file_name":"move_documents_in_proposal_to_dossier.py","file_ext":"py","file_size_in_byte":2099,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"175601510","text":"import io\nimport re\nimport typing as t\n\nimport requests\n\nfrom .exceptions import EmoteNotFoundException, InvalidCommandException\n\n\nclass Emote:\n def __init__(self, emote_type: str, emote_id: str, emote_channel: t.Union[None, str]) -> None:\n self.emote_type = emote_type\n self.emote_id = emote_id\n self.emote_channel = emote_channel\n self.name = self.get_name()\n self.image = self.get_image()\n\n def get_name(self) -> str:\n if self.emote_type == \"twitch\":\n api_url = \"https://api.twitchemotes.com/api/v4/emotes\"\n api_res = requests.get(api_url, params={\"id\": self.emote_id}).json()\n return api_res[0][\"code\"]\n\n elif self.emote_type == \"frf\":\n api_url = f\"https://api.frankerfacez.com/v1/emote/{self.emote_id}\"\n api_res = requests.get(api_url).json()\n return api_res[\"emote\"][\"name\"]\n\n elif self.emote_type == \"btv\":\n if self.emote_channel == \"global\":\n api_url = \"https://api.betterttv.net/2/emotes\"\n else:\n api_url = f\"https://api.betterttv.net/2/channels/{self.emote_channel}\"\n api_res = requests.get(api_url).json()\n for emote in api_res[\"emotes\"]:\n if emote[\"id\"] == self.emote_id:\n return emote[\"code\"]\n\n def get_image(self) -> io.BytesIO:\n img = None\n if self.emote_type == \"twitch\":\n img = requests.get(f\"https://static-cdn.jtvnw.net/emoticons/v1/{self.emote_id}/3.0\").content\n elif self.emote_type == \"bttv\":\n img = requests.get(f\"https://cdn.betterttv.net/emote/{self.emote_id}/3x\").content\n elif self.emote_type == \"ffz\":\n img = requests.get(f\"https://cdn.frankerfacez.com/emoticon/{self.emote_id}/4\").content\n return io.BytesIO(img)\n\n @staticmethod\n def get_emote(cmd) -> \"Emote\":\n cmd_re = re.compile(r\"^\\b(twitch|bttv|ffz)\\b\\s([\\w\\d]+)(?:\\s(.+))?$\", re.I | re.M)\n cmd_match = re.match(cmd_re, cmd)\n\n if not cmd_match:\n raise InvalidCommandException()\n\n emote_type = cmd_match[1].lower()\n emote_id = cmd_match[2].strip().lower()\n\n emote_channel = None\n if emote_type == \"bttv\":\n emote_channel = cmd_match[3]\n if not emote_channel:\n raise InvalidCommandException()\n emote_channel = emote_channel.lower()\n\n try:\n emote = Emote(emote_type, emote_id, emote_channel)\n return emote\n except (KeyError, IndexError):\n raise EmoteNotFoundException()\n","sub_path":"cogs/emote_utils/emotes.py","file_name":"emotes.py","file_ext":"py","file_size_in_byte":2618,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"167642487","text":"from nltk.tokenize import RegexpTokenizer\nfrom stop_words import get_stop_words\nfrom nltk.stem.porter import PorterStemmer\nfrom gensim import corpora, models\nimport gensim\n\n\n\ndef createLDAtopics(di_input):\n\n\tdi_threadstrings = di_input\n\ttokenizer = RegexpTokenizer(r'\\w+')\n\n\t# create English stop words list\n\ten_stop = get_stop_words('en')\n\n\t# Create p_stemmer of class PorterStemmer\n\tp_stemmer = PorterStemmer()\n\t\n\t# compile sample documents into a list\n\tthread_set = []\n\n\tfor key, threadstring in di_threadstrings.items():\n\t\tthread_set.append(threadstring)\n\n\t# list for tokenized documents in loop\n\ttexts = []\n\n\t# loop through document list\n\tfor i in thread_set:\n\t \n\t # clean and tokenize document string\n\t raw = i.lower()\n\t tokens = tokenizer.tokenize(raw)\n\n\t # remove stop words from tokens\n\t stopped_tokens = [i for i in tokens if not i in en_stop]\n\t \n\t # stem tokens\n\t stemmed_tokens = [p_stemmer.stem(i) for i in stopped_tokens]\n\t \n\t # add tokens to list\n\t texts.append(stemmed_tokens)\n\n\t# turn our tokenized documents into a id <-> term dictionary\n\tdictionary = corpora.Dictionary(texts)\n\t \n\t# convert tokenized documents into a document-term matrix\n\tcorpus = [dictionary.doc2bow(text) for text in texts]\n\n\t# generate LDA model\n\tldamodel = gensim.models.ldamodel.LdaModel(corpus, num_topics=10, id2word = dictionary, passes=20)\n\n\t# output a result\n\tprint(ldamodel.print_topics(num_topics=10, num_words=5))","sub_path":"LDAscript.py","file_name":"LDAscript.py","file_ext":"py","file_size_in_byte":1452,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"235377243","text":"from bs4 import BeautifulSoup \r\nfrom requests import get\r\nall_links=[]\r\ncols=[ \"authors\",'date_publish','description','text','title','url']\r\nmaldivesindependent_data=pd.DataFrame(columns=cols)\r\nj = 'https://maldivesindependent.com/?s='+'--x'\r\nfor i in ['drugaddicts','narcotics','alcoholicsanonymous','rehabilitation','alcohol','abuse','drinking','sober',\r\n 'gamblers','foodaddicts','recovery','foodaddiction']:\r\n url = j.replace('--x',i) # replace it with a url you want to apply the rules to \r\n print(url)\r\n soup = BeautifulSoup(get(url).text, 'lxml')\r\n for i in range(len(soup.find_all('li',attrs={'class':\"mvp-blog-story-wrap left relative infinite-post\"}))):\r\n all_links.append(soup.find_all('li',attrs={'class':\"mvp-blog-story-wrap left relative infinite-post\"})[i].a['href'])\r\nfrom newsplease import NewsPlease\r\nfor i in all_links:\r\n article = NewsPlease.from_url(i)\r\n maldivesindependent_data.loc[len(maldivesindependent_data)]=[article.authors,article.date_publish,article.description,\r\n article.text,article.title,article.url]\r\nmaldivesindependent_data.to_csv('maldivesindependent.csv',index=True)","sub_path":"vikram_reddy/maldives independent/maldivesindependent.py","file_name":"maldivesindependent.py","file_ext":"py","file_size_in_byte":1185,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"519743337","text":"# -*- coding: utf-8 -*-\n\n\nimport wx\nimport wx.xrc\n\n# Class TypeFrame\n\n\nclass TypeFrame(wx.Frame):\n def __init__(self, parent):\n wx.Frame.__init__(\n self,\n parent,\n id=wx.ID_ANY,\n title=u\"Tipo Material\",\n pos=wx.DefaultPosition,\n size=wx.Size(385, 150),\n style=wx.DEFAULT_FRAME_STYLE | wx.TAB_TRAVERSAL,\n )\n\n self.SetSizeHintsSz(wx.DefaultSize, wx.DefaultSize)\n\n bSizer1 = wx.BoxSizer(wx.VERTICAL)\n\n fgSizer1 = wx.FlexGridSizer(2, 2, 5, 5)\n fgSizer1.AddGrowableCol(1)\n fgSizer1.SetFlexibleDirection(wx.BOTH)\n fgSizer1.SetNonFlexibleGrowMode(wx.FLEX_GROWMODE_SPECIFIED)\n\n self.txtCode = wx.StaticText(\n self, wx.ID_ANY, u\"Codigo:\", wx.Point(-1, -1), wx.Size(80, 15), wx.ALIGN_LEFT\n )\n self.txtCode.Wrap(-1)\n fgSizer1.Add(self.txtCode, 0, wx.ALIGN_CENTER_VERTICAL | wx.ALL, 5)\n\n self.code = wx.TextCtrl(\n self,\n wx.ID_ANY,\n wx.EmptyString,\n wx.DefaultPosition,\n wx.Size(100, 20),\n wx.TE_RIGHT,\n )\n self.code.SetMaxLength(0)\n self.code.Enable(False)\n\n fgSizer1.Add(self.code, 0, wx.ALIGN_CENTER_VERTICAL | wx.ALL, 5)\n\n self.txtName = wx.StaticText(\n self,\n wx.ID_ANY,\n u\"Descripcion:\",\n wx.DefaultPosition,\n wx.Size(80, 15),\n wx.ALIGN_LEFT,\n )\n self.txtName.Wrap(-1)\n fgSizer1.Add(self.txtName, 0, wx.ALIGN_CENTER_VERTICAL | wx.ALL, 5)\n\n self.name = wx.TextCtrl(\n self,\n wx.ID_ANY,\n wx.EmptyString,\n wx.DefaultPosition,\n wx.Size(180, 20),\n wx.TE_LEFT,\n )\n self.name.SetMaxLength(0)\n fgSizer1.Add(self.name, 0, wx.ALIGN_CENTER_VERTICAL | wx.ALL, 5)\n\n bSizer1.Add(fgSizer1, 0, wx.ALIGN_CENTER_VERTICAL | wx.ALL, 5)\n\n fgSizer2 = wx.FlexGridSizer(1, 4, 0, 0)\n fgSizer2.SetFlexibleDirection(wx.BOTH)\n fgSizer2.SetNonFlexibleGrowMode(wx.FLEX_GROWMODE_SPECIFIED)\n\n self.btnOk = wx.Button(\n self, wx.ID_ANY, u\"Aceptar\", wx.DefaultPosition, wx.Size(80, 25), wx.NO_BORDER\n )\n self.btnOk.Enable(False)\n\n fgSizer2.Add(self.btnOk, 0, wx.ALIGN_CENTER_VERTICAL | wx.ALL, 5)\n\n self.btnModify = wx.Button(\n self,\n wx.ID_ANY,\n u\"Modificar\",\n wx.DefaultPosition,\n wx.Size(80, 25),\n wx.NO_BORDER,\n )\n self.btnModify.Enable(False)\n\n fgSizer2.Add(self.btnModify, 0, wx.ALIGN_CENTER_VERTICAL | wx.ALL, 5)\n\n self.btnCancel = wx.Button(\n self,\n wx.ID_ANY,\n u\"Cancelar\",\n wx.DefaultPosition,\n wx.Size(80, 25),\n wx.NO_BORDER,\n )\n self.btnCancel.SetDefault()\n fgSizer2.Add(self.btnCancel, 0, wx.ALIGN_CENTER_VERTICAL | wx.ALL, 5)\n\n self.btnExit = wx.Button(\n self, wx.ID_ANY, u\"Salir\", wx.DefaultPosition, wx.Size(80, 25), wx.NO_BORDER\n )\n fgSizer2.Add(self.btnExit, 0, wx.ALIGN_CENTER_VERTICAL | wx.ALL, 5)\n\n bSizer1.Add(fgSizer2, 0, wx.ALL, 5)\n\n self.SetSizer(bSizer1)\n self.Layout()\n\n self.Centre(wx.BOTH)\n\n def on_closed(self, evt):\n self.Destroy()\n evt.Skip()\n\n def __del__(self):\n pass\n","sub_path":"apps/warehousemng/typepro.py","file_name":"typepro.py","file_ext":"py","file_size_in_byte":3466,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"610370813","text":"import numpy as np\nimport tensorflow as tf\nimport pickle\nimport scipy.stats as ss\nfrom net_code.mlp_policy import MlpPolicy\n\ndef print_args(args):\n max_length = max([len(k) for k, _ in vars(args).items()])\n for k, v in vars(args).items():\n print(' ' * (max_length-len(k)) + k + ': ' + str(v))\n\n# var_list is returned by the policy.\n# Thus, they should be the same. I assume.\ndef saveToFlat(var_list, param_pkl_path):\n # get all the values\n var_values = np.concatenate([v.flatten() for v in tf.get_default_session().run(var_list)])\n pickle.dump(var_values, open(param_pkl_path, \"wb\"))\n\n\ndef load_from_file(param_pkl_path):\n with open(param_pkl_path, 'rb') as f:\n params = pickle.load(f)\n return params.astype(np.float32)\n\n\ndef loadFromFlat(var_list, param_pkl_path):\n flat_params = load_from_file(param_pkl_path)\n print(\"the type of the parameters stored is\", flat_params.dtype)\n shapes = list(map(lambda x: x.get_shape().as_list(), var_list))\n total_size = np.sum([int(np.prod(shape)) for shape in shapes])\n theta = tf.placeholder(tf.float32, [total_size])\n start = 0\n assigns = []\n for (shape, v) in zip(shapes, var_list):\n size = int(np.prod(shape))\n print(v.name)\n assigns.append(tf.assign(v, tf.reshape(theta[start:start + size], shape)))\n start += size\n op = tf.group(*assigns)\n tf.get_default_session().run(op, {theta: flat_params})\n\n\ndef condition(x, eps=1e-5):\n # smooth e,b when obtaining r_c\n parser.add_argument('--dim-belief', '-db', help='optimising batch size', type=int, default=256)\n parser.add_argument('--dim-belief', '-db', help='optimising batch size', type=int, default=256)\n return eps + (1.0 - 2.0 * eps) * x\n\n\ndef softmax(x):\n \"\"\"Compute softmax values for each sets of scores in x.\"\"\"\n e_x = np.exp(x-np.expand_dims(np.max(x, axis=-1), -1))\n return e_x / np.sum(e_x, axis=-1, keepdims=True)\n\n\n\ndef get_initial_embedding_target(mob):\n #mob multi-agent observations\n l = mob.shape[0]\n targets = [[mob[i][j][8:11] for j in range(2)] for i in range(l)]\n return np.array(targets)\n\ndef cross_entropy(p, q):\n q = np.maximum(q, 1e-20)\n return np.sum(-p*np.log2(q))\n\ndef square_loss(p, q):\n return np.sum(np.square(p-q))\n\ndef compute_communication_reward(news, ob, b, dis_type):\n\n # We have the belief of p1 and p2 simultaneously\n distance_fn = get_distance_fn(dis_type)\n rewards = []\n initial_embeddings = get_initial_embedding_target(ob)\n\n news = np.append(news, 1)\n\n # compute kl divergence\n i = 0\n for t in range(news.shape[0] - 1):\n if news[t+1]:\n rewards.append([0.0, 0.0])\n else:\n # TODO use another distances.\n r1 = distance_fn(initial_embeddings[t][0], b[t][1]) - distance_fn(initial_embeddings[t][0], b[t+1][1])\n r2 = distance_fn(initial_embeddings[t][1], b[t][0]) - distance_fn(initial_embeddings[t][1], b[t+1][0])\n rewards.append([r1, r2])\n return np.asarray(rewards)\n\n\ndef h_distance(p, q):\n return np.sqrt(np.sum((np.sqrt(p) - np.sqrt(q)) ** 2)) / np.sqrt(2)\n\n\ndef get_distance_fn(dis_type):\n if dis_type.lower() == 'kl':\n return ss.entropy\n elif dis_type.lower() == 'hl':\n return h_distance\n elif dis_type.lower() == 'ce':\n return cross_entropy\n elif dis_type.lower() == 'sl':\n return square_loss\n else:\n raise Exception(\"Unrecognized distance type\")\n\n\ndef traj_segment_generator(pi, env, horizon, stochastic):\n t = 0\n ac = [action_space.sample() for action_space in env.action_space] # not used, just so we have the datatype\n new = True # marks if we're on first timestep of an episode\n ob = env.reset()\n\n cur_ep_ret = [0, 0] # return in current episode\n cur_ep_len = 0 # len of current episode\n ep_rets = [] # returns of completed episodes in this segment\n ep_lens = [] # lengths of ...\n\n # Initialize history arrays\n obs = np.array([ob for _ in range(horizon)]) # 3D array\n \"\"\" - FIXED: Two Agents - \"\"\"\n rews = np.zeros((horizon, 2), 'float32') # 2D array\n vpreds = np.zeros((horizon, 2), 'float32') # 2D array\n news = np.zeros(horizon, 'int32') # shared\n acs = np.array([ac for _ in range(horizon)]) # 2D array\n\n print(obs.shape,\"observation\")\n print(rews.shape, \"rews\")\n print(acs.shape, \"acs\")\n\n while True:\n ac0, vpred0 = pi.act(stochastic, ob[0])\n ac1, vpred1 = pi.act(stochastic, ob[1])\n ac = [ac0, ac1]\n vpred = [vpred0, vpred1]\n\n # Slight weirdness here because we need value function at time T\n # before returning segment [0, T-1] so we get the correct\n # terminal value\n if t > 0 and t % horizon == 0:\n yield {\"ob\" : obs, \"rew\" : rews, \"vpred\" : vpreds, \"new\" : news,\n \"ac\" : acs, \"nextvpred\": np.array([vpred0, vpred1]) * (1 - new),\n \"ep_rets\" : ep_rets, \"ep_lens\" : ep_lens}\n # Be careful!!! if you change the downstream algorithm to aggregate\n # several of these batches, then be sure to do a deepcopy\n ep_rets = []\n ep_lens = []\n i = t % horizon\n obs[i] = ob\n vpreds[i] = vpred\n news[i] = new\n acs[i] = ac\n\n ob, rew, new, _ = env.step(ac)\n rews[i] = rew\n\n cur_ep_ret[0] += rew[0]\n cur_ep_ret[1] += rew[1]\n\n cur_ep_len += 1\n if new:\n ep_rets.append(cur_ep_ret)\n ep_lens.append(cur_ep_len)\n cur_ep_ret = [0, 0]\n cur_ep_len = 0\n ob = env.reset()\n t += 1\n\n\ndef add_vtarg_and_adv(seg, gamma, lam):\n \"\"\"\n Compute target value using TD(lambda) estimator, and advantage with GAE(lambda)\n \"\"\"\n # I think is has been simplified\n new = np.append(seg[\"new\"], 0) # last element is only used for last vtarg, but we already zeroed it if last new = 1\n vpred = np.concatenate([seg[\"vpred\"], seg[\"nextvpred\"].reshape(1, -1)], axis=0)\n T = len(seg[\"rew\"])\n seg[\"adv\"] = gaelam = np.empty((T, 2), 'float32')\n rew = seg[\"rew\"]\n lastgaelam = 0\n for t in reversed(range(T)):\n nonterminal = 1-new[t+1]\n delta = rew[t] + gamma * vpred[t+1] * nonterminal - vpred[t]\n gaelam[t] = lastgaelam = delta + gamma * lam * nonterminal * lastgaelam\n seg[\"tdlamret\"] = seg[\"adv\"] + seg[\"vpred\"]\n\ndef zipsame(*seqs):\n L = len(seqs[0])\n assert all(len(seq) == L for seq in seqs[1:])\n return zip(*seqs)\n","sub_path":"net_code/training_utils.py","file_name":"training_utils.py","file_ext":"py","file_size_in_byte":6507,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"585116821","text":"# Author: Loren Matilsky\n# Created: 12/19/2022\n#\n# Description: Script to plot radial thermodynamic profiles \n# (the reference state) from the equation_coefficients file\n\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nimport sys, os\nsys.path.append(os.environ['raco'])\nsys.path.append(os.environ['rapl'])\n\nfrom common import *\nfrom plotcommon import *\nfrom cla_util import *\n\n# Get CLAs\nargs = sys.argv\nclas0, clas = read_clas(args)\ndirname = clas0.dirname\ndirname_stripped = strip_dirname(dirname)\n\n# allowed args + defaults\nkwargs_default = dotdict()\nkwargs_default.fname = None\nkwargs_default.update(make_figure_kwargs_default)\nkwargs_default.update(lineplot_kwargs_default)\n\n# change kwargs with clas\nkw = update_dict(kwargs_default, clas)\nkw_make_figure = update_dict(make_figure_kwargs_default, clas)\nkw_lineplot = update_dict(lineplot_kwargs_default, clas)\n\n# find bad keys\nfind_bad_keys(kwargs_default, clas, clas0['routinename'], justwarn=True)\n\n# read reference state\neq = get_eq(dirname, kw.fname)\n\n# things to plot and ylabels\nprofiles = [eq.grav, eq.dsdr, eq.heat,\\\n eq.rho, eq.tmp, eq.dlnrho,\\\n eq.d2lnrho, eq.dlntmp, eq.prs]\nylabels = ['gravity (g)', r'$d\\overline{S}/dr$', 'heating (Q)',\\\n 'density (' + r'$\\overline{\\rho}$' + ')', 'temperature (' + r'$\\overline{T}$' + ')', r'$dln\\overline{\\rho}/dr$', \\\n r'$d^2ln\\rho/dr^2$', r'$dln\\overline{T}/dr$', 'pressure (' + r'$\\overline{P}=\\overline{\\rho}\\mathcal{R}\\overline{T}$' + ')']\ncount = 0\nif eq.reference_type in [2, 4]:\n profiles.insert(2, eq.nsq)\n ylabels.insert(2, r'$N^2=(g/c_p)d\\overline{S}/dr$')\n count += 1\nif not close_to_zero(eq.dlnrho):\n profiles.insert(6 + count, -1.0/eq.dlnrho)\n ylabels.insert(6 + count, r'$H_\\rho=-(dln\\overline{\\rho}/dr)^{-1}$')\n\n# create the plot; start with plotting all the energy fluxes\nnplots = len(profiles)\nncol = 4\nkw_make_figure.nplots = nplots\nkw_make_figure.ncol = ncol\nfig, axs, fpar = make_figure(**kw_make_figure)\n\n# x label\nkw_lineplot.xlabel = 'radius'\n\n#for iplot in [0]:\nfor iplot in range(nplots):\n ax = axs.flatten()[iplot]\n kw_lineplot.ylabel = ylabels[iplot]\n lineplot(eq.rr, profiles[iplot], ax, **kw_lineplot)\n\n# make title \nthe_title = dirname_stripped + '\\nbackground reference state' +\\\n '\\nreference_type = %i' %eq.reference_type\nmargin_x = fpar['margin_left'] + fpar['sub_margin_left']\nmargin_y = default_margin/fpar['height_inches']\nfig.text(margin_x, 1 - margin_y, the_title,\\\n ha='left', va='top', fontsize=default_titlesize)\n\n# save the figure, maybe\nif clas0['saveplot']:\n plotdir = my_mkdir(clas0['plotdir'])\n savefile = plotdir + clas0['routinename'] + clas0['tag'] + '.png'\n print ('saving figure at ' + savefile)\n plt.savefig(savefile, dpi=300)\nif clas0['showplot']:\n plt.show()\n","sub_path":"plot/reference.py","file_name":"reference.py","file_ext":"py","file_size_in_byte":2805,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"647935439","text":"import jwt\nfrom flask import Blueprint, request, current_app, jsonify\nfrom app import models\n\nusers_bp = Blueprint('users', __name__)\n\nfrom functools import wraps\n\n\ndef token_required(f):\n @wraps(f)\n def decorated(*args, **kwargs):\n token = None\n if 'Authorization' in request.headers:\n token = request.headers['Authorization'].split(' ')[1]\n if not token: \n return jsonify({\n 'error': 'Unauthorized',\n 'message': 'You did not provide a token'\n }), 401\n try:\n data=jwt.decode(token, current_app.config['SECRET_KEY'], algorithms=[\"HS256\"])\n current_user=models.User.query.get_or_404(int(data['user_id']))\n if current_user is None:\n return jsonify({\n 'error': 'Unauthorized',\n 'message': 'Invalid token'\n }), 401\n except Exception as e:\n return jsonify({\n 'error': 'Something went wrong',\n 'message': str(e)\n }), 500\n\n return f(current_user, *args, **kwargs)\n\n return decorated\n\nfrom . import views\n","sub_path":"app/users/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":1172,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"501601187","text":"'''\nScript accepts bills in a folder, with feature as decided by get_features\nCreates a graph of bills related to each other and also for impact words\n'''\nimport os\nimport networkx as nx\nimport sys\nfrom os import listdir\nfrom os.path import isfile, join\nfrom collections import defaultdict\nimport ipdb\nimport numpy as np\nFOLDER = \"./out_bill_only\"\n_SEP_ = ''\n_TITLE_SEP_ = '[]'\n_IMPACT_WORDS = ['cost', 'amount', 'import','$']\n\n\ndef get_node_name(filename):\n # BILL_115_HCONRES2_IH.out\n name = filename.strip().split(\"/\")[2].split(\"_\")\n name = \"_\".join(name[2:])\n name = name[:-4]\n return name\n\ndef visualize():\n bill_related = defaultdict(int)\n bill_impact = defaultdict(list)\n impact_count = np.array([0]*len(_IMPACT_WORDS))\n\n for f in listdir(FOLDER):\n filename = join(FOLDER, f)\n if isfile(filename):\n src = get_node_name(filename)\n with open(filename,\"r\") as f:\n impact_vector = np.array([0]*len(_IMPACT_WORDS))\n count = 0\n for line in f:\n sentence, related_bills, impact_word_string = line.strip().split(_SEP_)\n related_bills = related_bills.split(_TITLE_SEP_)\n if related_bills != ['']:\n # print \"related_words\", related_bills\n for dest_fileName in related_bills:\n dest = get_node_name(dest_fileName)\n if (dest, src) not in bill_related:\n bill_related[(src,dest)]+=1\n else:\n bill_related[(dest,src)]+=1\n\n if impact_word_string != '':\n impact_word_list = impact_word_string.split(\",\")\n for i in xrange(0, len(impact_word_list), 2):\n impact_word = impact_word_list[i].strip(\"(\")\n impact_count = int(impact_word_list[i+1].strip(\")\"))\n impact_vector[_IMPACT_WORDS.index(impact_word)]+=count\n impact_count+=impact_vector\n count+=1\n bill_impact[src] = 1.0* (impact_vector) / count\n \n # code to generate vertices and edges file for systemg\n vertices = [\"Node_Name\"]\n edges = [\"Node_A, Node_B, Weight\"]\n seen_nodes = set()\n\n for idx, (src, dest) in enumerate(bill_related):\n if idx==150:\n break\n if src not in seen_nodes:\n vertices.append(src)\n seen_nodes.add(src)\n if dest not in seen_nodes:\n vertices.append(dest)\n seen_nodes.add(dest)\n\n edges.append(src+\", \"+dest+\", \"+str(bill_related[(src,dest)]))\n\n with open(\"systemg/related_bills_vertices.txt\", \"w\") as f:\n vertices = \"\\n\".join(vertices)\n f.write(vertices)\n with open(\"systemg/related_bills_edges.txt\", \"w\") as f:\n edges = \"\\n\".join(edges)\n f.write(edges)\n\n vertices = [\"Node_Name\"] + _IMPACT_WORDS\n edges = [\"Node\", \"Impact\", \"Count\"]\n # count_max = 400\n for idx, bill in enumerate(bill_impact):\n # if idx == 300:\n # break\n vertices.append(bill)\n impact_vector = bill_impact[bill]\n\n for i in range(len(impact_vector)):\n if impact_vector[i]==0 or impact_vector[i]==0.0:\n continue\n edges.append(bill+\", \"+_IMPACT_WORDS[i]+\", \"+str(impact_vector[i]))\n\n with open(\"systemg/impact_vertices.txt\",\"w\") as f:\n vertices = \"\\n\".join(vertices)\n f.write(vertices)\n with open(\"systemg/impact_edges.txt\",\"w\") as f:\n edges = \"\\n\".join(edges)\n f.write(edges)\n\n\n\nif __name__ == '__main__':\n visualize()\n","sub_path":"visualize_impact.py","file_name":"visualize_impact.py","file_ext":"py","file_size_in_byte":3765,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"534117724","text":"# %matplotlib inline\nimport numpy as np #для работы с массивами - их по умолчанию в Python нет\n\nimport matplotlib.pyplot as plt #для генерации графиков\n\nn = 100\nr = 0.7\nx = np.random.rand(n)\ny = r*x + (1 - r)*np.random.rand(n)\nplt.plot(x, y, 'o')\nplt.xlabel('x')\nplt.ylabel('y')\nplt.grid(True)\n\n#1 Regression\na = (np.sum(x)*np.sum(y) - n*np.sum(x*y))/(np.sum(x)*np.sum(x) - n*np.sum(x*x))\nb = (np.sum(y) - a*np.sum(x))/n\n\nA = np.vstack([x, np.ones(len(x))]).T\na1, b1 = np.linalg.lstsq(A, y)[0]\nprint(a, b)\nprint(a1, b1)\nplt.plot([0, 1], [b, a + b])\nplt.show()\n\n#2 Correlation\nc = np.corrcoef(x, y)\nprint(c)\n","sub_path":"operation_07_regressia.py","file_name":"operation_07_regressia.py","file_ext":"py","file_size_in_byte":684,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"154191110","text":"#Code:\n\nn= int(input())\nc= list(map(int,input().split()))\nc.sort(reverse=True)\nd=0\ni=0\nwhile d!=1:\n c1= c[:i+1]\n c2= c[i+1:]\n if sum(c1)>sum(c2):\n d==1\n break\n i=i+1\nprint(len(c1))\n","sub_path":"#111 (Div. 2) - Twins.py","file_name":"#111 (Div. 2) - Twins.py","file_ext":"py","file_size_in_byte":207,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"331832919","text":"import numbers\nimport numpy as np\nimport scipy.sparse as sp\nfrom sklearn.base import BaseEstimator, RegressorMixin\nfrom sklearn.cross_validation import _safe_split, KFold, StratifiedKFold, ShuffleSplit\nfrom sklearn.utils.multiclass import type_of_target\nfrom sklearn.utils.validation import _num_samples\n\n\ndef check_cv(cv, X=None, y=None, classifier=False):\n \"\"\"Input checker utility for building a CV in a user friendly way.\n Parameters\n ----------\n cv : int, double, cross-validation generator or an iterable, optional\n Determines the cross-validation splitting strategy.\n Possible inputs for cv are:\n - None, to use the default 3-fold cross-validation,\n - integer, to specify the number of folds.\n - double, to specify fraction of test set\n - An object to be used as a cross-validation generator.\n - An iterable yielding train/test splits.\n For integer/None inputs, if ``y`` is binary or multiclass,\n :class:`StratifiedKFold` used. If the estimator is a classifier\n or if ``y`` is neither binary nor multiclass, :class:`KFold` is used.\n Refer :ref:`User Guide ` for the various\n cross-validation strategies that can be used here.\n X : array-like\n The data the cross-val object will be applied on.\n y : array-like\n The target variable for a supervised learning problem.\n classifier : boolean optional\n Whether the task is a classification task, in which case\n stratified KFold will be used.\n Returns\n -------\n checked_cv: a cross-validation generator instance.\n The return value is guaranteed to be a cv generator instance, whatever\n the input type.\n \"\"\"\n is_sparse = sp.issparse(X)\n if cv is None:\n cv = 3\n if isinstance(cv, numbers.Integral):\n if classifier:\n if type_of_target(y) in ['binary', 'multiclass']:\n cv = StratifiedKFold(y, cv)\n else:\n cv = KFold(_num_samples(y), cv)\n else:\n if not is_sparse:\n n_samples = len(X)\n else:\n n_samples = X.shape[0]\n cv = KFold(n_samples, cv)\n elif isinstance(cv, float) and cv > 0 and cv < 1:\n if classifier:\n raise NotImplementedError\n else:\n if not is_sparse:\n n_samples = len(X)\n else:\n n_samples = X.shape[0]\n cv = ShuffleSplit(n=n_samples, test_size=cv, random_state=12345)\n return cv\n\nclass MPRegressor(BaseEstimator, RegressorMixin):\n \"\"\"Marshall Palmar: a meta-regressor\n\n Using the Marshal Palmar equation with power scaling factor to be tuned.\n\n Parameters\n ----------\n power_scaling : double, default: 0.9\n MP's power scaling factor (Recommended range 0-1). Optimal \"should\" be around 0.85\n\n Attributes\n ----------\n power_scaling:\n \"\"\"\n def __init__(self, power_scaling=0.9):\n self.power_scaling = power_scaling\n\n def fit(self, X, y=None):\n return self # return-self convention\n\n def predict(self, X):\n mmperhr = pow(pow(10, X['Ref_mean']/10)/200, 0.625 * self.power_scaling)\n return mmperhr \n\n def transform(self):\n # for pipeline\n pass\n\n def fit_transform(self, X, y=None):\n # for pipeline\n pass","sub_path":"Pranav/mp.py","file_name":"mp.py","file_ext":"py","file_size_in_byte":3375,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"286438650","text":"from django.http import HttpResponseRedirect\nfrom django.shortcuts import render\nfrom django.contrib.auth import logout\n\n\nfrom .forms import SignUpForm, LoginForm, SnippetForm\n\n\ndef index(request):\n return render(request, 'index.html', {})\n\n\ndef log_out(request):\n logout(request)\n return HttpResponseRedirect('/index')\n\n\ndef sign_up(request):\n if request.method == 'POST':\n form = SignUpForm(data=request.POST)\n if form.is_valid():\n form.save()\n return HttpResponseRedirect('/auth/login')\n else:\n return render(request, 'auth/signup.html', {'form': form})\n else:\n form = SignUpForm()\n return render(request, 'auth/signup.html', {'form': form})\n\n\ndef log_in(request):\n if request.method == 'POST':\n form = LoginForm(data=request.POST)\n if form.is_valid():\n result = form.login_user(request)\n if result[0]:\n return HttpResponseRedirect('/index')\n else:\n msg = result[2]\n return render(request, 'auth/login.html', {'form':form,\n 'msg':msg})\n else:\n return render(request, 'auth/login.html', {'form':form})\n else:\n form = LoginForm()\n return render(request, 'auth/login.html', {'form':form})\n\n\ndef post_snippet(request):\n if request.method == 'POST':\n form = SnippetForm(request.user, data=request.POST)\n if form.is_valid():\n new_snippet = form.save()\n return HttpResponseRedirect(new_snippet.get_absolute_url())\n else:\n return render(request, 'snippet/post.html', {'form': form})\n else:\n form = SnippetForm(request.user)\n return render(request, 'snippet/post.html', {'form': form})","sub_path":"snippet/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1817,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"119223688","text":"import cv2\n\nface_cascade=cv2.CascadeClassifier(\"haarcascade_frontalface_default.xml\")\n\nimage=cv2.imread(\"news.jpg\")\n\ngray_image=cv2.cvtColor(image,cv2.COLOR_BGR2GRAY)\n\nfaces=face_cascade.detectMultiScale(gray_image,\nscaleFactor=1.05,\nminNeighbors=5)\n\nfor x,y,w,h in faces:\n image=cv2.rectangle(image, (x,y), (x+w, y+h), (255,0,0), 3)\n\nresized = cv2.resize(image, (int(image.shape[1]/2),int(image.shape[0]/2)))\ncv2.imshow(\"image\", resized)\ncv2.waitKey(0)\ncv2.destroyAllWindows()\n","sub_path":"Files/face_detection.py","file_name":"face_detection.py","file_ext":"py","file_size_in_byte":481,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"9509167","text":"import curses\nimport sys\n\n\nclass App(object):\n\n def __init__(self):\n self._mainWin = None\n self._stdscr = None\n self._ttyCols = -1\n self._ttyRows = -1\n self._delegate = None\n self._inputProc = None\n self._inputCtx = None\n self._noInputProc = None\n self._noInputCtx = None\n self._quit = False\n\n def appInit(self):\n # Init curses mode\n self._mainWin = curses.initscr()\n curses.noecho() # turn off key echo\n curses.cbreak() # turn of input buffering\n self._mainWin.keypad(1) # Process function key escape sequences as single key events\n curses.curs_set(0)\n if curses.has_colors():\n curses.start_color()\n curses.init_pair(1, curses.COLOR_BLUE, curses.COLOR_BLACK)\n curses.init_pair(2, curses.COLOR_CYAN, curses.COLOR_BLACK)\n curses.init_pair(3, curses.COLOR_GREEN, curses.COLOR_BLACK)\n curses.init_pair(4, curses.COLOR_MAGENTA, curses.COLOR_BLACK)\n curses.init_pair(5, curses.COLOR_RED, curses.COLOR_BLACK)\n curses.init_pair(6, curses.COLOR_WHITE, curses.COLOR_BLACK)\n curses.init_pair(7, curses.COLOR_YELLOW, curses.COLOR_BLACK)\n curses.init_pair(8, curses.COLOR_WHITE, curses.COLOR_BLUE)\n \n self._ttyCols = curses.COLS\n self._ttyRows = curses.LINES\n\n curses.halfdelay(30)\n\n self._mainWin.clear()\n self._mainWin.refresh()\n \n def appFinish(self):\n # Restore starting terminal state\n self._mainWin.keypad(0)\n curses.nocbreak()\n curses.echo()\n curses.endwin()\n\n def appQuit(self):\n curses.halfdelay(1)\n self._quit = True\n\n def main(self):\n curses.wrapper(self._wrapperMain)\n\n def mainLoop(self):\n while not self._quit:\n try:\n input = None\n input = self._mainWin.getkey()\n except curses.error as curx:\n if str(curx) == 'no input':\n if self._noInputProc:\n self._noInputProc(self, self._noInputCtx)\n else:\n raise curx\n self._ttyChanged()\n if input and self._inputProc:\n self._inputProc(self, input, self._inputCtx)\n\n def mainWin(self):\n return self._mainWin\n\n def setDelegate(self, delegate):\n self._delegate = delegate\n\n def setInputCallback(self, callable, context):\n self._inputProc = callable\n self._inputCtx = context\n\n def setNoInputCallback(self, callable, context):\n self._noInputProc = callable\n self._noInputCtx = context\n\n def ttySize(self):\n self._ttyCols = curses.COLS\n self._ttyRows = curses.LINES\n return (self._ttyRows, self._ttyCols)\n\n def _ttyChanged(self):\n resized = curses.is_term_resized(self._ttyRows, self._ttyCols)\n if resized:\n self,_ttyRows, self._ttyCols = self._grid.mainWin().getmaxyx()\n self._mainWin.clear()\n curses.resizeterm(self._ttyRows, self._ttyCols)\n self._mainWin.refresh()\n\n def _wrapperMain(self, stdscr):\n self._stdscr = stdscr\n self.appInit()\n if self._delegate:\n self._delegate(self)\n else:\n self.mainLoop()\n self.appFinish()\n\n\nclass _TestDelegate(object):\n def __init__(self):\n self._app = None\n\n def initDelegate(self, app):\n self._app = app\n ttyRows, ttyCols = app.ttySize()\n\n label = \"Curses App\"\n app.mainWin().addstr( ttyRows//2, (ttyCols-len(label))//2, label) \n label = \"Press any key to exit\"\n app.mainWin().addstr( ttyRows//2+2, (ttyCols-len(label))//2, label) \n\n app.setInputCallback(self.inputCB, \"CB ARG\")\n \n app.mainLoop()\n\n def inputCB(self, app, input, ctx):\n app.appQuit()\n \n\nif __name__ == \"__main__\":\n app = App()\n appDelegate = _TestDelegate()\n app.setDelegate(appDelegate.initDelegate)\n app.main()\n","sub_path":"CURSES/cursesapp.py","file_name":"cursesapp.py","file_ext":"py","file_size_in_byte":3644,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"220121688","text":"from numpy.random import multivariate_normal, randn\nfrom scipy.linalg.special_matrices import toeplitz\nimport numpy as np\nfrom sklearn.datasets import load_svmlight_file\n# from scipy.sparse import hstack\nfrom scipy.sparse import issparse\nfrom sklearn.preprocessing import scale\nimport loss\nimport regularizer\n\n\ndef load_problem(X, y, problem_type = \"classification\",\n loss_type = \"logistic\", regularizer_type = \"L2\", \n bias_term = True, scale_features = True,\n center_features = False):\n if problem_type == \"classification\":\n # X = X / np.max(np.abs(X), axis=0) # X = (X - np.mean(X, axis=0)) / (np.std(X, axis=0) + 1e-10)\n # label preprocessing for some dataset whose label is not {-1, 1}\n max_v, min_v = np.max(y), np.min(y)\n idx_min = (y == min_v)\n idx_max = (y == max_v)\n y[idx_min] = -1\n y[idx_max] = 1\n elif problem_type != \"regression\":\n raise Exception(\"Unknown problem type!\")\n\n\n if center_features:\n X = (X - np.mean(X, axis=0)) / (np.std(X, axis=0) + 1e-8)\n if scale_features:\n # X = scale(X, with_mean=False, copy=False)\n X = X / np.max(np.abs(X), axis=0)\n if bias_term:# adding a column filling with all ones.(bias term)\n X = np.c_[X, np.ones(X.shape[0])]\n\n if loss_type == \"L2\":\n criterion = loss.L2()\n elif loss_type == \"PseudoHuber\":\n criterion = loss.PseudoHuberLoss(delta=1.0)\n elif loss_type == \"Logistic\":\n criterion = loss.LogisticLoss()\n else:\n raise Exception(\"Unknown loss function!\")\n\n if regularizer_type == \"L2\":\n penalty = regularizer.L2()\n elif regularizer_type == \"PseudoHuber\":\n penalty = regularizer.PseudoHuber(delta=1.0)\n else:\n raise Exception(\"Unknown regularizer type!\")\n\n return criterion, penalty, X, y\n# ======================================\n\n\ndef get_data(data_path):\n # This function is taken from the code of Rui YUAN\n \"\"\"Once datasets are downloaded, load datasets.\"\"\"\n data = load_svmlight_file(data_path)\n return data[0], data[1]\n\ndef sparsity(data):\n # data, (num_samples, num_features)\n n, d = data.shape\n total_entries = n * d\n zeros_entries = np.sum(data == 0)\n return zeros_entries / total_entries\n\n# ===============================\n# These two functions to generate artificial data\n# are taken from the course M2-Optimization for Data Science\n# =================================\n\n\ndef simu_linreg(x, n, std=1., corr=0.5):\n \"\"\"Simulation for the least-squares problem.\n\n Parameters\n ----------\n x : ndarray, shape (d,)\n The coefficients of the model\n n : int\n Sample size\n std : float, default=1.\n Standard-deviation of the noise\n corr : float, default=0.5\n Correlation of the features matrix\n\n Returns\n -------\n A : ndarray, shape (n, d)\n The design matrix.\n b : ndarray, shape (n,)\n The targets.\n \"\"\"\n d = x.shape[0]\n cov = toeplitz(corr ** np.arange(0, d))\n A = multivariate_normal(np.zeros(d), cov, size=n)\n noise = std * randn(n)\n b = A.dot(x) + noise\n return A, b\n\n\ndef simu_logreg(x, n, std=1., corr=0.5):\n \"\"\"Simulation for the logistic regression problem.\n\n Parameters\n ----------\n x : ndarray, shape (d,)\n The coefficients of the model\n n : int\n Sample size\n std : float, default=1.\n Standard-deviation of the noise\n corr : float, default=0.5\n Correlation of the features matrix\n\n Returns\n -------\n A : ndarray, shape (n, d)\n The design matrix.\n b : ndarray, shape (n,)\n The targets.\n \"\"\"\n A, b = simu_linreg(x, n, std=1., corr=corr)\n return A, np.sign(b)\n\n\n","sub_path":"load_data.py","file_name":"load_data.py","file_ext":"py","file_size_in_byte":3745,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"236784025","text":"def inssort(arr):\r\n for i in range(1, len(arr)):\r\n key = arr[i]\r\n j = i-1\r\n while j >=0 and key < arr[j] :\r\n arr[j+1] = arr[j]\r\n j -= 1\r\n arr[j+1] = key\r\n print(\"Sorted list is:\\n\")\r\n print(arr)\r\n\r\nn=int(input(\"Enter the no. of elements in the list: \"))\r\nA=[]\r\nfor i in range(0,n):\r\n x=int(input(\"Enter a number: \"))\r\n A.append(x)\r\n\r\ninssort(A)\r\n","sub_path":"insertionSort.py","file_name":"insertionSort.py","file_ext":"py","file_size_in_byte":413,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"236363109","text":"from hashlib import sha256\nimport logging\nimport os\nimport prometheus_client\nfrom sqlalchemy.orm import joinedload\nfrom sqlalchemy.sql import functions\nimport tempfile\n\nfrom .. import database\nfrom ..repositories import RepositoryError, get_experiment_from_repository, \\\n parse_repository_url\nfrom .. import rpz_metadata\nfrom ..utils import secure_filename\nfrom .base import BaseHandler\n\n\nlogger = logging.getLogger(__name__)\n\n\nPROM_PAGE = prometheus_client.Counter(\n 'pages_total',\n \"Page requests\",\n ['name']\n)\n\n\nclass Index(BaseHandler):\n \"\"\"Landing page from which a user can select an experiment to upload.\n \"\"\"\n PROM_PAGE.labels('index').inc(0)\n\n def get(self):\n PROM_PAGE.labels('index').inc()\n return self.render('index.html')\n\n def head(self):\n PROM_PAGE.labels('index').inc()\n return self.finish()\n\n\nclass Upload(BaseHandler):\n \"\"\"Target of the landing page.\n\n An experiment has been provided, store it and extract metadata.\n \"\"\"\n PROM_PAGE.labels('upload').inc(0)\n\n async def post(self):\n PROM_PAGE.labels('upload').inc()\n\n # If a URL was provided, and no file\n if self.get_body_argument('rpz_url', None):\n # Redirect to reproduce_repo view\n try:\n repo, repo_path = await parse_repository_url(\n self.get_body_argument('rpz_url')\n )\n except RepositoryError as e:\n self.set_status(404)\n return self.render('repository_notfound.html', message=str(e))\n else:\n return self.redirect(self.reverse_url(\n 'reproduce_repo',\n repo, repo_path,\n ))\n\n # Get uploaded file\n # FIXME: Don't hold the file in memory!\n try:\n uploaded_file = self.request.files['rpz_file'][0]\n except (KeyError, IndexError):\n return self.render('setup_badfile.html', message=\"Missing file\")\n assert uploaded_file.filename\n logger.info(\"Incoming file: %r\", uploaded_file.filename)\n filename = secure_filename(uploaded_file.filename)\n\n # Hash it\n hasher = sha256(uploaded_file.body)\n filehash = hasher.hexdigest()\n logger.info(\"Computed hash: %s\", filehash)\n\n # Check for existence of experiment\n experiment = self.db.query(database.Experiment).get(filehash)\n if experiment:\n experiment.last_access = functions.now()\n logger.info(\"File exists in storage\")\n else:\n # Write it to disk\n with tempfile.NamedTemporaryFile('w+b', suffix='.rpz') as tfile:\n tfile.write(uploaded_file.body)\n\n # Insert it in database\n try:\n experiment = rpz_metadata.make_experiment(\n filehash,\n tfile.name,\n )\n except rpz_metadata.InvalidPackage as e:\n return self.render('setup_badfile.html', message=str(e))\n self.db.add(experiment)\n\n # Insert it on S3\n await self.application.object_store.upload_file_async(\n 'experiments',\n filehash,\n tfile.name,\n )\n logger.info(\"Inserted file in storage\")\n\n # Insert Upload in database\n upload = database.Upload(experiment=experiment,\n filename=filename,\n submitted_ip=self.request.remote_ip)\n self.db.add(upload)\n self.db.commit()\n\n # Encode ID for permanent URL\n upload_short_id = upload.short_id\n\n # Redirect to build page\n return self.redirect(\n self.reverse_url('reproduce_local', upload_short_id),\n status=302,\n )\n\n\nclass BaseReproduce(BaseHandler):\n def reproduce(self, upload):\n experiment = upload.experiment\n filename = upload.filename\n experiment_url = self.url_for_upload(upload)\n\n input_files = (\n self.db.query(database.Path)\n .filter(database.Path.experiment_hash ==\n experiment.hash)\n .filter(database.Path.is_input)).all()\n return self.render(\n 'setup.html',\n filename=filename,\n built=True, error=False,\n params=experiment.parameters,\n input_files=input_files,\n upload_short_id=upload.short_id,\n experiment_url=experiment_url,\n )\n\n\nclass ReproduceRepo(BaseReproduce):\n PROM_PAGE.labels('reproduce_repo').inc(0)\n\n async def get(self, repo, repo_path):\n \"\"\"Reproduce an experiment from a data repository.\n \"\"\"\n PROM_PAGE.labels('reproduce_repo').inc()\n\n # Check the database for an experiment already stored matching the URI\n repository_key = '%s/%s' % (repo, repo_path)\n upload = (\n self.db.query(database.Upload)\n .options(joinedload(database.Upload.experiment))\n .filter(database.Upload.repository_key == repository_key)\n .order_by(database.Upload.id.desc())\n ).first()\n if upload is None:\n try:\n upload = await get_experiment_from_repository(\n self.db, self.application.object_store,\n self.request.remote_ip,\n repo, repo_path,\n )\n except RepositoryError as e:\n self.set_status(404)\n return self.render('setup_notfound.html', message=str(e))\n except rpz_metadata.InvalidPackage as e:\n self.set_status(404)\n return self.render('setup_badfile.html', message=str(e))\n\n # Also updates last access\n upload.experiment.last_access = functions.now()\n\n return self.reproduce(upload)\n\n\nclass ReproduceLocal(BaseReproduce):\n PROM_PAGE.labels('reproduce_local').inc(0)\n\n def get(self, upload_short_id):\n \"\"\"Ask for run parameters.\n \"\"\"\n PROM_PAGE.labels('reproduce_local').inc()\n\n # Decode info from URL\n try:\n upload_id = database.Upload.decode_id(upload_short_id)\n except ValueError:\n self.set_status(404)\n return self.render('setup_notfound.html')\n\n # Look up the experiment in database\n upload = (\n self.db.query(database.Upload)\n .options(joinedload(database.Upload.experiment))\n .get(upload_id)\n )\n if upload is None:\n self.set_status(404)\n return self.render('setup_notfound.html')\n\n # Also updates last access\n upload.experiment.last_access = functions.now()\n\n return self.reproduce(upload)\n\n\nclass StartRun(BaseHandler):\n PROM_PAGE.labels('start_run').inc(0)\n\n async def post(self, upload_short_id):\n \"\"\"Gets the run parameters POSTed to from /reproduce.\n\n Triggers the run and redirects to the results page.\n \"\"\"\n PROM_PAGE.labels('start_run').inc()\n\n # Decode info from URL\n try:\n upload_id = database.Upload.decode_id(upload_short_id)\n except ValueError:\n self.set_status(404)\n return self.render('setup_notfound.html')\n\n # Look up the experiment in database\n upload = (\n self.db.query(database.Upload)\n .options(joinedload(database.Upload.experiment))\n .get(upload_id)\n )\n if upload is None:\n self.set_status(404)\n return self.render('setup_notfound.html')\n experiment = upload.experiment\n\n # New run entry\n run = database.Run(experiment_hash=experiment.hash,\n upload_id=upload_id)\n self.db.add(run)\n\n # Get list of parameters\n params = set()\n params_unset = set()\n for param in experiment.parameters:\n if not param.optional:\n params_unset.add(param.name)\n params.add(param.name)\n\n # Get run parameters\n for k, v in self.request.body_arguments.items():\n if k.startswith('param_'):\n if not v:\n continue\n name = k[6:]\n if name not in params:\n raise ValueError(\"Unknown parameter %s\" % k)\n v = v[-1].decode('utf-8')\n run.parameter_values.append(\n database.ParameterValue(name=name, value=v)\n )\n params_unset.discard(name)\n\n if params_unset:\n raise ValueError(\"Missing value for parameters: %s\" %\n \", \".join(params_unset))\n\n # Get list of input files\n input_files = set(\n p.name for p in (\n self.db.query(database.Path)\n .filter(database.Path.experiment_hash == experiment.hash)\n .filter(database.Path.is_input)\n ).all())\n\n # Get input files\n for k, uploaded_file in self.request.files.items():\n if not uploaded_file:\n continue\n uploaded_file = uploaded_file[0]\n\n if not k.startswith('inputfile_') or k[10:] not in input_files:\n raise ValueError(\"Unknown input file %s\" % k)\n\n name = k[10:]\n logger.info(\"Incoming input file: %s\", name)\n\n # Hash file\n hasher = sha256(uploaded_file.body)\n inputfilehash = hasher.hexdigest()\n logger.info(\"Computed hash: %s\", inputfilehash)\n\n # Insert it into S3\n await self.application.object_store.upload_bytes_async(\n 'inputs',\n inputfilehash,\n uploaded_file.body,\n )\n logger.info(\"Inserted file in storage\")\n\n # Insert it in database\n input_file = database.InputFile(\n hash=inputfilehash, name=name,\n size=len(uploaded_file.body),\n )\n run.input_files.append(input_file)\n\n # Get ports to expose\n for port_str in self.get_body_argument('ports', '').split():\n port_str = port_str.strip()\n if port_str:\n try:\n port = int(port_str)\n if not (1 <= port <= 65535):\n raise ValueError\n except (ValueError, OverflowError):\n raise ValueError(\"Invalid port number %r\" % port_str)\n run.ports.append(database.RunPort(\n port_number=port,\n ))\n\n # Trigger run\n self.db.commit()\n self.application.runner.run(run.id)\n\n # Redirect to results page\n return self.redirect(\n self.reverse_url('results', run.short_id),\n status=302,\n )\n\n\nclass Results(BaseHandler):\n PROM_PAGE.labels('results').inc(0)\n\n def get(self, run_short_id):\n \"\"\"Shows the results of a run, whether it's done or in progress.\n \"\"\"\n PROM_PAGE.labels('results').inc()\n\n # Decode info from URL\n try:\n run_id = database.Run.decode_id(run_short_id)\n except ValueError:\n self.set_status(404)\n return self.render('results_notfound.html')\n\n # Look up the run in the database\n run = (\n self.db.query(database.Run)\n .options(joinedload(database.Run.experiment),\n joinedload(database.Run.upload),\n joinedload(database.Run.parameter_values),\n joinedload(database.Run.input_files),\n joinedload(database.Run.output_files))\n ).get(run_id)\n if run is None:\n self.set_status(404)\n return self.render('results_notfound.html')\n # Update last access\n run.experiment.last_access = functions.now()\n self.db.commit()\n\n # JSON endpoint, returns data for JavaScript to update the page\n if self.is_json_requested():\n log_from = int(self.get_query_argument('log_from', '0'), 10)\n return self.send_json({\n 'started': bool(run.started),\n 'done': bool(run.done),\n 'log': run.get_log(log_from),\n })\n # HTML view, return the page\n else:\n def get_port_url(port_number):\n tpl = os.environ.get(\n 'WEB_PROXY_URL',\n 'http://{short_id}-{port}.127.0.0.1.xip.io:8001',\n )\n return tpl.format(\n short_id=run_short_id,\n port=port_number,\n )\n\n return self.render(\n 'results.html',\n run=run,\n log=run.get_log(0),\n started=bool(run.started),\n done=bool(run.done),\n experiment_url=self.url_for_upload(run.upload),\n get_port_url=get_port_url,\n )\n\n\nclass About(BaseHandler):\n PROM_PAGE.labels('about').inc(0)\n\n def get(self):\n PROM_PAGE.labels('about').inc()\n return self.render('about.html')\n\n\nclass Data(BaseHandler):\n \"\"\"Print some system information.\n \"\"\"\n PROM_PAGE.labels('about').inc(0)\n\n def get(self):\n PROM_PAGE.labels('data').inc()\n return self.render(\n 'data.html',\n experiments=self.db.query(database.Experiment).all(),\n )\n\n\nclass Health(BaseHandler):\n def get(self):\n return self.finish('ok')\n","sub_path":"reproserver/web/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":13693,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"492255869","text":"from django.urls import path, include\nfrom django.conf.urls.static import static\nfrom django.conf import settings\nfrom . import views\n\nurlpatterns = [\n path('', views.index, name='index'),\n path('allorders', views.order_display, name='all_orders'),\n path('logout', views.log_out, name=\"logout\"),\n path('takeorder', views.take_order, name='takeorder'),\n path('spec_order/', views.spec_order, name='specific_order'),\n path('ajax/phonesearch', views.ajax, name='AjaxPhoneSearch'),\n path('allcustomers', views.all_customers, name='all_customers'),\n path('dayrec', views.dayrec, name='dayrec'),\n path('custompage/', views.custompage, name='custompage'),\n path('bill/', views.genbill, name='genbill'),\n path('ajax/item_add', views.ajax_item_add, name='add_item'),\n path('order_today', views.today_order, name=\"today_order\")\n] + static(settings.STATIC_URL, document_root=settings.STATIC_ROOT)\n","sub_path":"Billing/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":954,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"123845853","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Sun Jul 21 00:51:00 2019\r\n\r\n@author: ASUS\r\n\"\"\"\r\n\r\nimport numpy as np\r\nimport gc\r\nclass node:\r\n tree_method='hist'\r\n\r\n def __init__(self,verbose=0):\r\n self.parent=None\r\n self.left=None\r\n self.right=None\r\n self.split_value=0.0\r\n self.split_feature=-1\r\n self.feature0_value=None#存放feature0 的值\r\n self.col_indices=None\r\n self.indices=None\r\n self.best_split_feature=-1\r\n self.split_pos_matrix=None\r\n self.w=0\r\n self.depth=0\r\n self.gain=0.0\r\n self.obj=0.0\r\n self.verbose=verbose\r\n \r\n def is_leaf(self):\r\n return self.left==None and self.right==None\r\n \r\n def max_depth(self):\r\n if self.is_leaf():\r\n return self.depth\r\n else:\r\n return max( self.left.max_depth(), self.right.max_depth() )\r\n \r\n def num_leaves(self):\r\n if self.is_leaf():\r\n return 1\r\n else:\r\n return self.left.num_leaves()+self.right.num_leaves()\r\n \r\n def add(self):\r\n assert self.is_leaf(), \"y must add children to a leaf!\"\r\n\r\n left_node = node(verbose=self.verbose)\r\n left_node.parent = self\r\n left_node.depth = self.depth+1\r\n self.left = left_node\r\n\r\n right_node = node(verbose=self.verbose)\r\n right_node.parent = self\r\n right_node.depth = self.depth+1\r\n self.right = right_node\r\n\r\n return self.left, self.right\r\n \r\n def delete_node(self):\r\n assert self.is_leaf(), \"you should delete a leaf!\"\r\n assert self.depth != 1, \"cannot delete root!\"\r\n\r\n parent = self.parent\r\n parent.left = None\r\n parent.right = None\r\n\r\n return parent\r\n \r\n def objective(self, gamma=0.): #计算一棵树,一颗subtrees或者一个叶子节点的objective loss\r\n \r\n if self.is_leaf():\r\n return self.obj+gamma\r\n else:\r\n return self.left.objective(gamma)+self.right.objective(gamma)\r\n \r\n def regularization(self, gamma=0., lbda=0.):#计算一棵树,一颗subtrees或者一个叶子节点的objective loss\r\n if self.is_leaf():\r\n if np.isscalar(self.w): #为什么要这个\r\n # in that case, lambda was set to 0 (no penalty for the bias)\r\n w2 = 0.\r\n else:\r\n w2 = np.dot(self.w,self.w)\r\n return gamma + 0.5*lbda*w2\r\n else:\r\n return self.left.regularization(gamma,lbda)+self.right.regularization(gamma,lbda)\r\n \r\n\r\n \r\n \r\n \r\n \r\n def set_weight(self, X,g, h, lbda,new_X=None):\r\n self.w, self.obj,self.feature0_value = self.get_weight(X, g, h, lbda,new_X,set_weights=True)#linear_model 是一个boolean\r\n \r\n def get_weight(self,X,g,h,lbda,new_X=None,set_weights=False):\r\n try:\r\n n= X.shape[0]\r\n except:\r\n print('new_X is None')\r\n raise\r\n if self.parent!=None and set_weights:#当只有在set_weight和不是根节点的时候\r\n #X_tilde=np.c_[new_X,X[:,self.parent.best_split_feature],np.ones(shape=(n,1),dtype=float)]\r\n new_X[:,1]=X[:,self.parent.best_split_feature]\r\n X_tilde=new_X\r\n d=new_X.shape[1]\r\n else:\r\n \r\n X_tilde = new_X\r\n d=new_X.shape[1]\r\n \r\n g_tilde = np.dot(X_tilde.transpose(),g)\r\n H_tilde = np.dot(X_tilde.transpose()*h, X_tilde)\r\n Lambda = lbda*np.eye(d)\r\n Lambda[d-1,d-1] = 0.\r\n C = H_tilde+Lambda\r\n try:\r\n C_inv=np.linalg.inv(C)\r\n #在这里算obj的时候没有加上gamma *T ,\r\n except:\r\n print('Cannot do matrix inverse')\r\n raise\r\n \r\n w=-np.dot(C_inv,g_tilde)\r\n \r\n obj=0.5*np.dot(g_tilde.transpose(),w)\r\n return w,obj,np.dot(new_X[:,:-1],w[:-1]) #w(m,) obj:scalar\r\n \r\n def find_best_split(self, X, g, h, lbda, gamma, max_samples_linear_model, min_samples_leaf,tree_method,split_pos_matrix,new_X):\r\n assert new_X.shape[0] == len(g), 'length of the data and g is not the same'\r\n assert len(h) == len(g), 'length of h and g are not the same'\r\n assert self.split_feature == -1, 'y have found the split featrue before'\r\n n, d = X.shape# n 是batchsize大小 d是原始数据集的所有特征大小\r\n self.gain = np.float64(\"-inf\")\r\n if tree_method=='hist':\r\n for f in range(0,d):\r\n #print('now finding the feature{}'.format(f))\r\n for pos in split_pos_matrix[f]:\r\n \r\n c = ( X[:,f] < pos )\r\n left_n = np.sum(c) #计算左叶子的数目\r\n right_n = n-left_n#计算右叶子的数目\r\n if ( left_n < min_samples_leaf ) or ( right_n < min_samples_leaf ):#如果叶子的数目小于min_samples_leaf则跳过这一split\r\n continue\r\n feature=X[:,f]\r\n #X_for_linear=np.c_[new_X,feature]\r\n new_X[:,1]=feature\r\n X_for_linear=new_X\r\n \r\n #X_for_linear送入左叶子节点进行线性回归的matrix\r\n try:\r\n _ , obj_left,feature0_value = self.get_weight(np.compress(c,X,axis=0), np.compress(c,g,axis=0),np.compress(c,h,axis=0), lbda,np.compress(c,X_for_linear,axis=0)) #获取左叶子节点的objection数值\r\n except:\r\n if self.verbose > 1:\r\n print( \"exception when testing for split of feature {} at pos {}\".format(f,pos) )\r\n continue\r\n #del X_for_linear\r\n #gc.collect()\r\n c = np.invert(c)\r\n try:\r\n _ , obj_right,feature0_value = self.get_weight(np.compress(c,X,axis=0), np.compress(c,g,axis=0),np.compress(c,h,axis=0), lbda,np.compress(c,X_for_linear,axis=0))#获取右节点的objection\r\n except:\r\n if self.verbose > 1:\r\n print( \"exception when testing for split of feature {} at pos {}\".format(f,pos) )\r\n continue\r\n \r\n \r\n gain = self.obj-(obj_left+obj_right+gamma) #到时候再看下gamma是+还是-\r\n if gain > self.gain:\r\n self.gain = gain\r\n self.split_feature = f\r\n self.split_value = pos\r\n \r\n self.best_split_feature=self.split_feature \r\n if self.verbose > 1:\r\n if ( self.gain != np.float64(\"-inf\") ) and ( self.gain < 0. ):\r\n print( \"negative gain!\" )\r\n print( \"find best split: gain={:+6.4e}, feature={:2d}, value={:+8.4f}\".format(self.gain, self.split_feature, self.split_value) )\r\n \r\n def predict(self,X):\r\n n,d = X.shape\r\n #X_tmp=np.concatenate((X,np.ones(shape=(n,1),dtype=float)),axis=1)\r\n X_tmp=X.copy()\r\n if self.is_leaf():\r\n fea_flag=np.zeros(d,dtype=int)\r\n node=self.parent\r\n while(node is not None):\r\n if fea_flag[node.best_split_feature]!=1 :\r\n \r\n X_tmp[:,node.best_split_feature]=X_tmp[:,node.best_split_feature]-node.split_value\r\n fea_flag[node.best_split_feature]=1\r\n\r\n node=node.parent\r\n if np.isscalar(self.w):\r\n return self.w\r\n else:\r\n return np.dot(X_tmp,self.w)\r\n else:\r\n assert self.split_feature > -1, \"split feature must be > -1!\"\r\n y = np.zeros(n, dtype=float)\r\n \r\n c = ( X[:,self.split_feature] < self.split_value )\r\n\r\n y[c] = self.left.predict(X_tmp[c,:])\r\n y[np.invert(c)] = self.right.predict(X_tmp[np.invert(c),:])\r\n return y\r\n def recalculate(self,X,g,h,lbda):\r\n n,d=X.shape\r\n # w=np.zeros(d,dtype=float)\r\n #X_tmp=np.concatenate((X,np.ones(shape=(n,1),dtype=float)),axis=1)\r\n X_tmp=X.copy()\r\n if self.is_leaf():\r\n fea_flag=np.zeros(d,dtype=int)\r\n node=self.parent\r\n while(node is not None):\r\n if fea_flag[node.best_split_feature]!=1:\r\n \r\n X_tmp[:,node.best_split_feature]=X_tmp[:,node.best_split_feature]-node.split_value\r\n fea_flag[node.best_split_feature]=1\r\n\r\n node=node.parent\r\n \r\n X_tilde = X_tmp\r\n d= X_tmp.shape[1]\r\n g_tilde = np.dot(X_tilde.transpose(),g)\r\n H_tilde = np.dot(X_tilde.transpose()*h, X_tilde)\r\n Lambda = lbda*np.eye(d)\r\n Lambda[d-1,d-1] = 0.\r\n C = H_tilde+Lambda\r\n try:\r\n C_inv=np.linalg.inv(C)\r\n #在这里算obj的时候没有加上gamma *T ,\r\n except:\r\n print('Cannot do matrix inverse')\r\n raise\r\n \r\n self.w=-np.dot(C_inv,g_tilde)\r\n else:\r\n assert self.split_feature > -1, \"split feature must be > -1!\"\r\n c = ( X[:,self.split_feature] < self.split_value )\r\n self.left.recalculate(X_tmp[c,:],g[c],h[c],lbda)\r\n self.right.recalculate(X_tmp[np.invert(c),:],g[np.invert(c)],h[np.invert(c)],lbda)","sub_path":"self_package_v62(final-continuous)/tree_node.py","file_name":"tree_node.py","file_ext":"py","file_size_in_byte":9678,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"373569888","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Feb 28 05:20:33 2018\n\n@author: Clandestina\n\"\"\"\n\nfrom PyQt5.QtWidgets import QWidget, QTabWidget, QVBoxLayout\n# Custom Dependencies\nfrom modules.Visualizacion import Visualizacion\nfrom modules.Estacionariedad import Estacionariedad\nfrom modules.Espectro import Espectro\nimport rpy2.robjects.packages as rpackages\nfrom rpy2.robjects.packages import STAP\nimport rpy2.robjects as robjects\nfrom rpy2.robjects.vectors import StrVector\n\nclass vistaPrincipal(QWidget):\n \n def __init__(self, parent): \n super(QWidget, self).__init__(parent)\n utils = rpackages.importr('utils')\n utils.chooseCRANmirror(ind=1)\n packnames=('plotrix', 'psd', 'fractal', 'squash', 'doParallel')\n for x in packnames:\n if (not rpackages.isinstalled(x)):\n utils.install_packages(x)\n \n #self.layout es una ventana contenedora\n self.layout = QVBoxLayout(self)\n \n # Initialize tab screen\n #self.tabs es un elemento que soporta la contención de pestañas\n self.tabs = QTabWidget()\n \n #Pestañas\n self.tabVis = QWidget()\n self.tabEst = QWidget()\n self.tabEsp = QWidget()\n\n\n # Content Tabs \n self.tabVisLayout = Visualizacion()\n self.tabEstLayout = Estacionariedad()\n self.tabEspLayout = Espectro()\n\n\n # Add tabs\n self.tabs.addTab(self.tabVis,\"Visualization\")\n self.tabs.addTab(self.tabEst,\"Estacionariedad\")\n self.tabs.addTab(self.tabEsp,\"Espectro\")\n \n # Set asing layout\n self.tabVis.layout = self.tabVisLayout\n self.tabEst.layout = self.tabEstLayout\n self.tabEsp.layout = self.tabEspLayout\n \n # SetLayout Tabs \n self.tabVis.setLayout(self.tabVis.layout)\n self.tabEst.setLayout(self.tabEst.layout)\n self.tabEsp.setLayout(self.tabEsp.layout)\n \n # Add tabs to Main\n self.layout.addWidget(self.tabs)\n self.setLayout(self.layout)","sub_path":"Interfaz-Julio-mayo/modules/vistaPrincipal.py","file_name":"vistaPrincipal.py","file_ext":"py","file_size_in_byte":2018,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"82498290","text":"import logging\n\nfrom aiogram import Dispatcher\nfrom aiogram.dispatcher import FSMContext\nfrom aiogram.dispatcher.filters.state import State, StatesGroup\n\nfrom aiogram.types import Message, ParseMode\n#import aiogram.utils.markdown as md\n\nfrom tgbot.models.role import UserRole\nfrom tgbot.services.repository import Repo\n\nfrom .dialog import DialogBaseTemplate\n\n\nclass AddUserDialog(StatesGroup):\n get_forwarded_message = State()\n status = State()\n submit = State()\n\n\nclass DelUserDialog(StatesGroup):\n db_id = State()\n submit = State()\n\n\nclass AddUserProc(DialogBaseTemplate):\n \"\"\"/adduser unique dialog instructions\"\"\"\n\n def __init__(self, fsm, fsm_group, rules={} ):\n super().__init__(self, fsm, fsm_group )\n self.fsm = fsm\n self.fsm_group = fsm_group\n self.rules = {\n 'get_forwarded_message': {\n 'question': (\n 'Для добавления пользователя в базу перешлите мне'\n ' сообщение от него. Если в нем не будет'\n ' поля \"forward_from\" - напишите telegram user_id',\n 'Перешлите мне сообщение с полем \"forward_from\"'\n ' или напишите telegram user_id!',),\n 'test': self.get_forwarded_message_test,\n 'calc': self.get_forwarded_message_calc,\n 'after_answer': self.default_after_answer_proc,\n },\n 'status': {\n 'question': ('Укажите статус пользователя (от 1 до 5)', ),\n 'test': self.status_test,\n 'after_answer': self.default_after_answer_proc,\n },\n 'submit': {\n 'question': ('Подтвердите введенные данные:', ),\n 'keyboard': self.KEYBOARD_APPROVE,\n 'test': self.has_approve_answer,\n 'after_question': self.submit_after_question_proc,\n 'after_answer': self.submit_after_answer_proc,\n },\n } | rules\n\n def get_forwarded_message_test(self, message):\n if self.has_forward_from(message) == 0:\n return 0\n return self.is_int(message, from_=1, to_=2000000000)\n\n def status_test(self, message):\n return self.is_int(message, from_=1, to_=5)\n\n def get_forwarded_message_calc(self, message):\n if self.has_forward_from(message) == 0:\n user = message.forward_from\n return {\n 'telegram_id': user.id,\n 'name': user.username if user.username else ''}\n return {'telegram_id': message.text, 'name': ''}\n\n def submit_after_question_proc(self, data):\n return (\n f\"Telegram user_id: {data.get('telegram_id', '')}\\n\"\n f\"Username: {data.get('name', '')}\\n\"\n f\"Статус: {data.get('status', '')}\\n\")\n\n async def submit_after_answer_proc(self, m, state, repo, **notused):\n data = await state.get_data()\n if data.get('submit', '') == 'Подтверждаю':\n await repo.add_user(**data)\n mes_text = 'Пользователь добавлен в базу'\n else:\n mes_text = 'Пользователь НЕ добавлен в базу'\n await m.answer(mes_text, reply_markup=self.KEYBOARD_REMOVE)\n\n async def default_after_answer_proc(self, m, answer_data, **notused):\n await m.answer(f\"Ok. {answer_data}\")\n\n\nclass DelUserProc(DialogBaseTemplate):\n \"\"\"/deluser unique dialog instructions\"\"\"\n\n def __init__(self, fsm, fsm_group, rules={} ):\n super().__init__(self, fsm, fsm_group )\n self.fsm = fsm\n self.fsm_group = fsm_group\n self.rules = {\n 'db_id': {\n 'question': ('Для удаления пользователя из базы напишите мне его id', ),\n 'test': self.is_int,\n },\n 'submit': {\n 'question': ('Подтвердите удаление пользователя', ),\n 'keyboard': self.KEYBOARD_APPROVE,\n 'test': self.has_approve_answer,\n 'after_question': self.submit_after_question_proc,\n 'after_answer': self.submit_after_answer_proc,\n },\n } | rules\n\n def submit_after_question_proc(self, data):\n return (f\"id пользователя: {data.get('db_id', '')}\")\n\n async def submit_after_answer_proc(self, m, state, repo, **notused):\n data = await state.get_data()\n if data.get('submit', '') == 'Подтверждаю':\n await repo.del_user(data['db_id'])\n mes_text = 'Пользователь удален'\n else:\n mes_text = 'Пользователь НЕ удален'\n\n await m.answer(mes_text, reply_markup=self.KEYBOARD_REMOVE)\n\n\n# Место, за которое стыдно... но я пока ничего красивее не придумал\ndialogs = {\n 'AddUserDialog': AddUserProc(AddUserDialog, 'AddUserDialog'),\n 'DelUserDialog': DelUserProc(DelUserDialog, 'DelUserDialog'),\n}\nlog = logging.getLogger(__name__)\n\n\nasync def adduser_start(m: Message, state: FSMContext):\n log.info('adduser_start by %s(username) %s(first_name) id%s' %\n (m.from_user.username, m.from_user.first_name, m.from_user.id))\n dialog = dialogs['AddUserDialog']\n await dialog.next_step()\n await first_question(m, state, dialog)\n\n\nasync def deluser_start(m: Message, state: FSMContext):\n dialog = dialogs['DelUserDialog']\n await dialog.next_step()\n await first_question(m, state, dialog)\n\n\nasync def first_question(m, state, dialog):\n current_step = await state.get_state()\n if current_step:\n dialog.set_state(current_step)\n await m.answer(dialog.question_text(), reply_markup=dialog.keyboard())\n\n\nasync def do_dialog(m: Message, state: FSMContext, repo: Repo):\n current_step = await state.get_state()\n dialog = dialogs[current_step.split(':')[0]]\n dialog.process_current_step(current_step, m)\n\n if not dialog.ready_to_next_step():\n await m.reply(dialog.question_text())\n return\n\n answer_data = dialog.get_answer_result()\n await state.update_data(**answer_data)\n\n if dialog.has_after_answer_proc():\n await dialog.after_answer_proc(m=m,\n state=state, repo=repo, answer_data=answer_data)\n\n await dialog.next_step()\n\n current_step = await state.get_state()\n if not current_step:\n return\n\n await first_question(m, state, dialog)\n if dialog.has_after_question_proc():\n data = await state.get_data()\n await m.answer(dialog.after_question_proc(data))\n\n\nasync def any_message(m: Message):\n user = m.from_user\n await m.answer(f\"Привет, админ! {user.first_name} {user.last_name}! (@{user.username}, id: {user.id})\")\n await m.answer(\n '/adduser - добавить пользователя в БД\\n'\n '/showuser - посмотреть список пользователей\\n'\n '/deluser - удалить пользователя из БД\\n'\n '/state - текущее состояние диалога от FSM\\n'\n 'cancel - прервать текущую процедуру (диалог)'\n )\n\n\nasync def showuser(m: Message, repo: Repo):\n userlist = await repo.list_users()\n await m.answer(f\"Список пользователей: {userlist}\")\n\n\nasync def show_current_state(m: Message, state: FSMContext):\n currentState = await state.get_state()\n await m.answer(f\"Cостояние текущего процесса: {currentState}\")\n\n\nasync def do_cancel(m: Message, state: FSMContext):\n await state.finish()\n await m.answer('Текущий процесс прерван...Удачи!',\n reply_markup=DialogBaseTemplate.KEYBOARD_REMOVE)\n\n\nasync def nope(a):\n pass\n\n\ndef register_admin(dp: Dispatcher):\n dp.register_message_handler(any_message, commands=['start', 'help'], is_admin=True)\n dp.register_message_handler(adduser_start, commands=['adduser'], is_admin=True)\n dp.register_message_handler(deluser_start, commands=['deluser'], is_admin=True)\n dp.register_message_handler(showuser, commands=['showuser'], is_admin=True)\n dp.register_message_handler(show_current_state, commands=['state'], state='*', is_admin=True)\n dp.register_message_handler(nope, text=['cancel','A','А'], state=None, is_admin=True)\n dp.register_message_handler(do_cancel, text=['cancel','A','А'], state='*', is_admin=True)\n dp.register_message_handler(any_message, state=None, is_admin=True)\n dp.register_message_handler(do_dialog, state='*',is_admin=True)\n","sub_path":"tgbot/handlers/admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":8866,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"588191502","text":"# -------------------------------------------------------------------------\n#\n# Part of the CodeChecker project, under the Apache License v2.0 with\n# LLVM Exceptions. See LICENSE for license information.\n# SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception\n#\n# -------------------------------------------------------------------------\n\n\nimport glob\nimport logging\nimport os\nimport plistlib\n\nfrom xml.parsers.expat import ExpatError\n\nfrom codechecker_report_hash.hash import get_report_hash, HashType\n\nfrom codechecker_report_converter.analyzer_result import AnalyzerResult\n\n\nLOG = logging.getLogger('ReportConverter')\n\n\nclass CppcheckAnalyzerResult(AnalyzerResult):\n \"\"\" Transform analyzer result of Cppcheck. \"\"\"\n\n TOOL_NAME = 'cppcheck'\n NAME = 'Cppcheck'\n URL = 'http://cppcheck.sourceforge.net'\n\n def parse(self, analyzer_result):\n \"\"\" Creates plist objects from the given analyzer result.\n\n Returns a list of plist objects.\n \"\"\"\n plist_files = []\n if os.path.isdir(analyzer_result):\n plist_files = glob.glob(os.path.join(analyzer_result, \"*.plist\"))\n elif os.path.isfile(analyzer_result) and \\\n analyzer_result.endswith(\".plist\"):\n plist_files = [analyzer_result]\n else:\n LOG.error(\"The given input should be an existing CppCheck result \"\n \"directory or a plist file.\")\n return None\n\n file_to_plist_data = {}\n for f in plist_files:\n plist_file = os.path.basename(f)\n file_name = '{0}_{1}.plist'.format(os.path.splitext(plist_file)[0],\n self.TOOL_NAME)\n\n with open(f, 'rb') as plist_file:\n try:\n file_to_plist_data[file_name] = plistlib.load(plist_file)\n except ExpatError:\n LOG.error(\"Failed to parse '%s'! Skipping...\", file_name)\n\n return file_to_plist_data\n\n def _post_process_result(self, file_to_plist_data):\n \"\"\" Post process the parsed result.\n\n By default it will add report hashes and metada information for the\n diagnostics.\n \"\"\"\n for file_name, plist_data in file_to_plist_data.items():\n try:\n self._add_report_hash(plist_data)\n self._add_metadata(plist_data)\n except IndexError:\n LOG.warning(\"Failed to update '%s' while generating a report \"\n \"hash! Skipping...\", file_name)\n file_to_plist_data[file_name] = None\n\n def _add_report_hash(self, plist_data):\n \"\"\" Generate report hash for the given plist data\n\n It will generate a context free hash for each diagnostics.\n \"\"\"\n files = plist_data['files']\n for diag in plist_data['diagnostics']:\n report_hash = diag.get('issue_hash_content_of_line_in_context')\n if not report_hash or report_hash == '0':\n report_hash = get_report_hash(\n diag, files[diag['location']['file']],\n HashType.CONTEXT_FREE)\n\n diag['issue_hash_content_of_line_in_context'] = report_hash\n\n def _write(self, file_to_plist_data, output_dir, file_name):\n \"\"\" Creates plist files from the parse result to the given output. \"\"\"\n output_dir = os.path.abspath(output_dir)\n for file_name, plist_data in file_to_plist_data.items():\n if not plist_data:\n continue\n\n out_file = os.path.join(output_dir, file_name)\n\n LOG.info(\"Modify plist file: '%s'.\", out_file)\n LOG.debug(plist_data)\n\n try:\n with open(out_file, 'wb') as plist_file:\n plistlib.dump(plist_data, plist_file)\n except TypeError as err:\n LOG.error('Failed to write plist file: %s', out_file)\n LOG.error(err)\n","sub_path":"tools/report-converter/codechecker_report_converter/cppcheck/analyzer_result.py","file_name":"analyzer_result.py","file_ext":"py","file_size_in_byte":3960,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"579748139","text":"\"\"\"\nCreate an application key with scopes for this service account returns \"Created\" response\n\"\"\"\n\nfrom os import environ\nfrom datadog_api_client import ApiClient, Configuration\nfrom datadog_api_client.v2.api.service_accounts_api import ServiceAccountsApi\nfrom datadog_api_client.v2.model.application_key_create_attributes import ApplicationKeyCreateAttributes\nfrom datadog_api_client.v2.model.application_key_create_data import ApplicationKeyCreateData\nfrom datadog_api_client.v2.model.application_key_create_request import ApplicationKeyCreateRequest\nfrom datadog_api_client.v2.model.application_keys_type import ApplicationKeysType\n\n# there is a valid \"service_account_user\" in the system\nSERVICE_ACCOUNT_USER_DATA_ID = environ[\"SERVICE_ACCOUNT_USER_DATA_ID\"]\n\nbody = ApplicationKeyCreateRequest(\n data=ApplicationKeyCreateData(\n attributes=ApplicationKeyCreateAttributes(\n name=\"Example-Service-Account\",\n scopes=[\n \"dashboards_read\",\n \"dashboards_write\",\n \"dashboards_public_share\",\n ],\n ),\n type=ApplicationKeysType.APPLICATION_KEYS,\n ),\n)\n\nconfiguration = Configuration()\nwith ApiClient(configuration) as api_client:\n api_instance = ServiceAccountsApi(api_client)\n response = api_instance.create_service_account_application_key(\n service_account_id=SERVICE_ACCOUNT_USER_DATA_ID, body=body\n )\n\n print(response)\n","sub_path":"examples/v2/service-accounts/CreateServiceAccountApplicationKey_3480494373.py","file_name":"CreateServiceAccountApplicationKey_3480494373.py","file_ext":"py","file_size_in_byte":1441,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"531533764","text":"from openerp import fields, models\n\nclass crm_lead_flexfone(models.Model):\n _inherit = \"crm.lead\"\n _description = \"Additional fields for Flexfone in CRM\"\n\n flexfone_lead_date = fields.Date(string=\"Lead date\")\n flexfone_contract_to = fields.Date(string=\"Bound by contract to\")\n flexfone_phys_employees = fields.Integer(string=\"Total employees\")\n flexfone_pbx_users = fields.Integer(string=\"PBX users\")\n flexfone_landline_users = fields.Integer(string=\"Landline users\")\n flexfone_cellphone_users = fields.Integer(string=\"Cellphone users\")\n flexfone_landline_service_provider = fields.Char(size=255, string=\"Landline service provider\")\n flexfone_cellphone_service_provider = fields.Char(size=500, string=\"Cellphone service provider\")\n flexfone_functionality_score = fields.Integer(string=\"Functionality %\")\n flexfone_price_score = fields.Integer(string=\"Price %\")\n flexfone_expense_landline = fields.Float(string=\"Expense for landline\", digits=(2,2))\n flexfone_expense_cellphone = fields.Float(string=\"Expenses for cellphone\", digits=(2,2))\n flexfone_notes = fields.Text(string=\"Notes\")\n flexfone_meeting_arranged = fields.Boolean(string=\"Meeting is arranged\")\n flexfone_meeting_declined = fields.Boolean(string=\"Meeting has been declined\")\n flexfone_followup = fields.Boolean(string=\"Follow up\")\n flexfone_meeting_done = fields.Boolean(string=\"Meeting completed\")\n flexfone_offer_sent = fields.Boolean(string=\"Offer sent\")\n flexfone_probability = fields.Integer(string=\"Probability\")\n flexfone_created_subscription = fields.Boolean(string=\"Created in Flexfone\")\n flexfone_created_users = fields.Integer(string=\"Users created\")\n flexfone_remarks_for_account_manager = fields.Text(string=\"Remarks for Flexfone account manager\")\n flexfone_account_manager = fields.Char(size=255, string=\"Account Manager\")","sub_path":"models/crm_lead_flexfone.py","file_name":"crm_lead_flexfone.py","file_ext":"py","file_size_in_byte":1874,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"402306538","text":"# -*- coding: utf-8 -*-\n\"\"\"\nutil functions related to information theory\n\"\"\"\n\nimport numpy as np\nimport pandas as pd\nfrom .util import get_counts, get_probs\n\n\ndef gini(x):\n \"\"\"\n gini index to measure impurity of a distribution.\n maximal when values occur equally (equal probability).\n \\n.. math::\n 1 - \\\\sum_{i=0}^{n} x_i^2\n \"\"\"\n _, counts = get_counts(x)\n p = get_probs(counts)\n\n return 1 - (p**2).sum()\n\n\ndef entropy(x, base=None):\n \"\"\"\n measure of information that can be gained from the distribution.\n maximal when values occur equally (equal probability).\n \\n.. math::\n H(X) = -\\sum_{i=0}^{n} x_ilog_2{x_i}\n \"\"\"\n _, counts = get_counts(x)\n p = get_probs(counts)\n if 1 in p: # one class has 100% probability => all others 0.0\n return 0\n ret = -(p * np.log(p)).sum()\n if base is not None:\n ret /= np.log(base)\n return ret\n\n\ndef entropy_p(p, base=None, ax=1):\n \"\"\" entropy given series of probabilities \"\"\"\n p = p[p > 0]\n shape = list(np.shape(p))\n if len(shape) < 2:\n p = pd.Series(p)\n p = p.values.reshape((1, shape[0]))\n if 1 in p: # one class has 100% probability => all others 0.0\n return 0\n ret = -(p * np.log(p)).sum(ax)\n if base is not None:\n ret /= np.log(base)\n return ret\n\n\ndef efficiency(x, base=None):\n \"\"\"\n efficiency (i.e. normalized entropy) is a ratio that relates the\n entropy of a non-uniform distribution to a uniform one of same classes.\n \\n.. math::\n E(X) = -\\sum_{i=0}^{n} \\\\frac{x_ilog_2{x_i}}{log_2{n}}\n \"\"\"\n _, counts = get_counts(x)\n p = get_probs(counts)\n\n ret = -(p * np.log(p) / np.log(len(p))).sum()\n if base is not None:\n ret /= np.log(base)\n return ret\n\n\ndef info_gain(x, y, impurity):\n \"\"\"\n when impurity function is entropy, this is equivalent to Mutual Information\n \\n.. math::\n IG(X, Y) = H(Y) - H(Y|X)\n \"\"\"\n impurity_ygivenx = 0\n xvalues, xcounts = get_counts(x)\n xprobs = get_probs(xcounts)\n\n for xval, xprob in zip(xvalues, xprobs):\n impurity_ygivenx += xprob * impurity(y[x == xval])\n\n return impurity(y) - impurity_ygivenx\n\n\ndef cond_info_gain(x, y, z, impurity):\n \"\"\"\n .. math::\n IG(X,Y|Z) = H(X|Z) + H(Y|Z) - H(X,Y|Z)\n \"\"\"\n xy = list(zip(x, y))\n\n impurity_ygivenz = impurity_xgivenz = impurity_xygivenz = 0\n zvals, zcounts = get_counts(z)\n zprobs = get_probs(zcounts)\n\n for zval, zprob in zip(zvals, zprobs):\n xyz = [xyi[0] for xyi in zip(xy, z) if xyi[1] == zval]\n\n impurity_ygivenz += zprob * impurity(y[z == zval])\n impurity_xgivenz += zprob * impurity(x[z == zval])\n impurity_xygivenz += zprob * impurity(xyz)\n\n return impurity_xgivenz + impurity_ygivenz - impurity_xygivenz\n\n\ndef symmetrical_uncertainty(x, y):\n \"\"\"\n \\n.. math::\n SU = \\\\frac{2IG}{H(x) + H(y)}\n \"\"\"\n return (2.0 * info_gain(x, y, entropy) / (entropy(x) + entropy(y)))\n\n\ndef info_gain_ratio(x, y, impurity):\n \"\"\"\n impurity is callable function to calculate impurity (i.e. gini or entropy)\n \\n.. math::\n IGR = \\\\frac{IG(X, Y)}{H(X)}\n \"\"\"\n return info_gain(x, y, impurity) / impurity(x)\n\n\ndef contingency_info_gain(crosstabs, norm=False, base_axis=0):\n \"\"\"\n info gain given 2-dim cross tabulations\n\n Parameters\n ----------\n crosstabs: array-like\n table of cross-tabluations/counts between 2 variables\n norm: boolean (default=False)\n normalize data by base entropy (return info_gain_ratio, ie. Theils' U)\n base_axis: int (default=0)\n reference axis for calculations\n\n Returns\n -------\n info_gain or info_gain_ratio\n \"\"\"\n informed_axis = 0 if base_axis else 1\n crosstabs = np.asanyarray(crosstabs)\n n = crosstabs.sum()\n ptabs = crosstabs / n\n base_probs = ptabs.sum(axis=base_axis)\n informed_probs = ptabs.sum(axis=informed_axis)\n oriented_data = crosstabs if base_axis else crosstabs.T\n marginal_probs = oriented_data / crosstabs.sum(axis=informed_axis)\n marginal_entropies = np.apply_along_axis(entropy_p, 0, marginal_probs)\n conditional_entropy = marginal_entropies.dot(informed_probs)\n gain = entropy_p(base_probs) - conditional_entropy\n if norm:\n return gain / entropy_p(base_probs)\n else:\n return gain\n\n\ndef _midd(x, y):\n return -entropy(list(zip(x, y))) + entropy(x) + entropy(y)\n\n\ndef _cmidd(x, y, z):\n return (entropy(list(zip(y, z))) + entropy(list(zip(x, z))) -\n entropy(list(zip(x, y, z))) - entropy(z))\n\n\ndef _xsi(x, n):\n \"\"\"\n x is the count of observations\n similar to entropy\n \"\"\"\n if x == 0:\n return 0\n else:\n return x / n * np.log(x)\n\n\n_vxsi = np.vectorize(_xsi)\n\n\ndef _ent_hat(x):\n \"\"\" binary pseudo entropy \"\"\"\n n = np.shape(x)[0]\n _, counts = get_counts(x)\n return np.log(n) - np.sum(_vxsi(counts, n))\n\n\ndef _info_gain_hat(x, y):\n return _ent_hat(x) + _ent_hat(y) - _ent_hat(list(zip(x, y)))\n\n\ndef cond_info_gain_hat(x, y, z):\n return (_ent_hat(list(zip(y, z))) + _ent_hat(list(zip(x, z))) -\n _ent_hat(list(zip(x, y, z))) - _ent_hat(z))\n","sub_path":"datafusionsm/utils/information.py","file_name":"information.py","file_ext":"py","file_size_in_byte":5171,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"286020511","text":"# Program to send leads to the sales call centre mailbox\n\n\nimport ConfigParser\nimport logging\nimport pymssql\nimport pandas as pd\nfrom datetime import date\nimport datetime\nimport csv\nimport os\nimport SendSMS\nimport smtplib\nfrom email.mime.text import MIMEText\n\n#TODO add entry to comms table\n\ndef SendLeads():\n logger = logging.getLogger(\"__main__.SendLeads\")\n logger.info('Start sending leads...')\n\n try: # Make connections\n config = ConfigParser.ConfigParser()\n config.read('config.cfg')\n\n#Connecting to Leads database\n host = config.get('database','host')\n user = config.get('database','username')\n password = config.get('database','password')\n database = config.get('database','database')\n\n# Connecting to datawarehouse\n dwhost = config.get('warehouse','host')\n dwuser = config.get('warehouse','username')\n dwpassword = config.get('warehouse','password')\n dwdatabase = config.get('warehouse','database')\n\n logger.debug('Connecting to database')\n conn = pymssql.connect(host=host, user=user, password=password, database=database)\n cur = conn.cursor(as_dict=True)\n\n logger.debug('Connecting to warehouse')\n dwconn = pymssql.connect(host=dwhost, user=dwuser, password=dwpassword, database=dwdatabase)\n dwcur = dwconn.cursor(as_dict=True)\n except:\n logger.error('Failed making initial connections',exc_info=True)\n conn.close()\n dwconn.close()\n raise\n\n# Start the FTP file operation\n try:\n# get broker and client details\n sql=\"SELECT *,lc.brokercode + '_' + lc.brokerhousecode brokerlookup \\\n FROM [dbo].[LeadClassification] lc \\\n inner join dbo.InitLead l on lc.LeadKey=l.LeadKey \\\n inner join dbo.SMSsent sms on sms.sourceid = lc.leadkey \\\n INNER JOIN dbo.SMSinbox inbox on sms.batchid = inbox.referring_batch_id \\\n where lc.Status = 'SMS_Sent' \\\n and lc.LeadClassification in (1,2) \\\n and message like 'yes%' \\\n and sms.batchid <> ''\"\n LeadClass = pd.read_sql(sql,conn)\n ClientList = LeadClass.ClientNo.unique()\n ClientList = \"','\".join(ClientList)\n ClientList = \"'\" + ClientList + \"'\"\n\n# Perform some data cleaning\n sql = \"select c.title + ' ' + case when c.nickname <> '' then c.nickname else c.firstname end + ' ' + c.Surname ClientName, c.NickName, c.firstname, c.clientid, c.birthdate,c.deathdate,c.emailaddress,c.cellphoneno,c.clientno from dim.client c where c.currentmember = 1 and c.clientno in (\" + ClientList + \")\"\n ClientInfo = pd.read_sql(sql,dwconn)\n\n ClientInfo['ClientName'] = ClientInfo['ClientName'].str.title()\n Client = pd.merge(LeadClass,ClientInfo,how='left',left_on='ClientNo',right_on='clientno')\n\n# GetOptout list\n\n sql = \"select * from [dbo].[OptOut]\"\n Optout = pd.read_sql(sql,conn)\n\n sql = \"select * from [dbo].[SMSinbox]\"\n SMSinbox = pd.read_sql(sql,conn)\n\n except:\n logger.error(\"Failed doing some initial groundwork and data cleaning\",exc_info = True)\n conn.close()\n dwconn.close()\n raise\n print(Client)\n\n# Start the email sending process\n if(len(Client)>0):\n\n try:\n for index,f in Client.iterrows():\n logger.debug(f)\n if pd.isnull(f['ClientName_y']) :\n LeadKey = f['LeadKey'][0]\n stmt = \"Update dbo.LeadClassification SET Status = 'MissingClientDetails' WHERE Leadkey = '\" + str(LeadKey) + \"'\"\n cur.execute(stmt)\n\n elif len(Optout.loc[Optout['IDno'] == f['clientid']]) > 0 :\n LeadKey = f['LeadKey'][0]\n stmt = \"Update dbo.LeadClassification SET Status = 'OptOut' WHERE Leadkey = '\" + str(LeadKey) + \"'\"\n cur.execute(stmt)\n\n else:\n logger.debug('Start gathering data for lead')\n\n# Data cleaning...\n table_name = 'dbo.Comms'\n try:\n ClientName = str(f['ClientName_y'])\n except:\n ClientName = 'Unknown'\n try:\n ClientEmail = str(f['emailaddress'])\n except:\n ClientEmail= 'Unknown'\n try:\n ClientCell = str(f['cellphoneno'])\n except:\n ClientCell ='Unknown'\n try:\n ClientNo = str(f['ClientNo'])\n except:\n ClientNo ='Unknown'\n try:\n ClientID = str(f['clientid'])\n except:\n ClientID = 'Unknown'\n\n LeadKey = f['LeadKey'][0]\n try:\n SafetyScoreDate = str((f['SubmitDateTime']).strftime('%Y%m%d %H:%M:%S'))\n logger.info(SafetyScoreDate)\n except:\n SafetyScoreDate = 'Unknown'\n logger.info(SafetyScoreDate)\n\n# Insert into LeadsComms and LeadAction\n q = \"insert into dbo.LeadAction values (%s,%s,'%s','%s')\" % (int(LeadKey),0,str(datetime.datetime.now())[0:-3],\"Lead_Sent\")\n print(q)\n cur.execute(q)\n\n# Update LeadClassification\n stmt = \"Update dbo.LeadClassification SET Status = 'Lead_Sent' WHERE Leadkey =\" + str(LeadKey)\n cur.execute(stmt)\n\n logger.debug(SafetyScoreDate)\n logger.debug(ClientName)\n logger.debug(ClientID)\n logger.debug(ClientCell)\n logger.debug(ClientEmail)\n logger.debug(SafetyScoreDate)\n\n# Compile teh email body\n Leadmsg = \"\"\"Client Name: \"\"\" + ClientName + \"\"\"\nClient ID no: \"\"\" + ClientID + \"\"\"\nClient Cell: \"\"\" + ClientCell + \"\"\"\nClient Email: \"\"\" + ClientEmail + \"\"\"\nSafety Score completed on: \"\"\" + SafetyScoreDate\n\n logger.info(Leadmsg)\n msg = MIMEText(Leadmsg)\n Subject = 'SafetyScore Campaign - Load client lead'\n From = 'safetyscorelead@momentum.co.za'\n To = 'MSTICrossSell@momentum.co.za'\n msg['Subject'] = Subject\n msg['From'] = From\n msg['To'] = To\n\n # Send the message via our own SMTP server, but don't include the\n # envelope header.\n s = smtplib.SMTP('smtp.momentum.co.za')\n s.sendmail(From, [To], msg.as_string())\n s.quit()\n #emailer.info(Leadmsg)\n\n#### X2\n\n\n except:\n logger.error(\"Failed writing FTP file or database updates for ftp file, caution, output file might be screwed\",exc_info = True)\n conn.rollback()\n conn.close()\n dwconn.close()\n raise\n logger.info(\"LeadsSent\")\n conn.commit()\n conn.close()\n dwconn.close()\n\n\n\nif __name__ == '__main__':\n\n import logging\n import logging.handlers\n\n # Setup logging for all the modules\n logger = logging.getLogger(__name__)\n logger.setLevel(logging.DEBUG)\n\n # Log to file with 1mb rollover\n handler = logging.handlers.RotatingFileHandler('MSTI_Parser.log','a',1000000,10)\n handler.setLevel(logging.INFO)\n\n formatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')\n handler.setFormatter(formatter)\n\n # log to email (to catch errors early when in production)\n err_handler = logging.handlers.SMTPHandler('smtp.momentum.co.za','safetyscorecampaign@momentum.co.za','louis.pienaar@momentum.co.za','MSTI_Parser Error')\n err_handler.setLevel(logging.ERROR)\n err_handler.setFormatter(formatter)\n\n # log to console for easier debugging\n ch = logging.StreamHandler()\n ch.setLevel(logging.DEBUG)\n ch.setFormatter(formatter)\n\n logger.addHandler(handler)\n #logger.addHandler(err_handler)\n logger.addHandler(ch)\n\n SendLeads()\n\n\n\n\n\n\n\n","sub_path":"SendLeads.py","file_name":"SendLeads.py","file_ext":"py","file_size_in_byte":8274,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"118593855","text":"import boto3\nimport base64\nimport hashlib\nimport html\nimport json\nimport os\nimport re\nimport urllib.parse\nimport requests\n\n\ndef _b64_decode(data):\n data += '=' * (4 - len(data) % 4)\n return base64.b64decode(data).decode('utf-8')\n\ndef jwt_payload_decode(jwt):\n _, payload, _ = jwt.split('.')\n return json.loads(_b64_decode(payload))\n\nopenid_provider = \"https://keycloak.ch.xiaopeiqing.com/auth/realms/iot\"\nclient_id = \"iotclient\"\nusername = \"test1\"\npassword = \"test1\"\nredirect_uri = \"http://127.0.0.1/home/login\"\nclient_secret=\"c2669e72-d858-472c\"\nidentity_pool_id=\"cn-north-1:2a776105\"\ndefault_region_name=\"cn-north-1\"\niam_openid_provider_name=\"keycloak.ch.test.com/auth/realms/iot\"\n\n#Why use PKCE in keycloak\n'''\nFrom: \nPKCE support with Keycloak 7.0: Keycloak 7.0 has been released on Aug 25th 2019 with PKCE support. This represents a major breakthrough for all mobile apps to increase security and to mitigate malicious attacks\nPublic client security vulnerability\n\nOAuth 2.0 [RFC6749] public clients are susceptible to the authorization code interception attack.\n\nIn this attack, the attacker intercepts the authorization code returned from the authorization endpoint within a communication path not protected by Transport Layer Security (TLS), such as interapplication communication within the client’s operating system.\n\nOnce the attacker has gained access to the authorization code, it can use it to obtain the access token.\n'''\n#generate code_verifier\ncode_verifier = base64.urlsafe_b64encode(os.urandom(40)).decode('utf-8')\ncode_verifier = re.sub('[^a-zA-Z0-9]+', '', code_verifier)\nprint(\"==================\")\nprint(\"code_verifier: %s\" %code_verifier)\n\n#generate code_challenge\ncode_challenge = hashlib.sha256(code_verifier.encode('utf-8')).digest()\ncode_challenge = base64.urlsafe_b64encode(code_challenge).decode('utf-8')\ncode_challenge = code_challenge.replace('=', '')\n\nprint(\"==================\")\nprint(\"code_challenge: %s\\n\" %code_challenge)\n\nstate = \"fooobarbaz\"\nresp = requests.get(\n url=openid_provider + \"/protocol/openid-connect/auth\",\n params={\n \"response_type\": \"code\",\n \"client_id\": client_id,\n \"scope\": \"openid\",\n \"redirect_uri\": redirect_uri,\n \"state\": state,\n \"code_challenge\": code_challenge,\n \"code_challenge_method\": \"S256\",\n },\n allow_redirects=False\n)\n\n\n\ncookie = resp.headers['Set-Cookie']\ncookie = '; '.join(c.split(';')[0] for c in cookie.split(', '))\nprint(\"==================\")\nprint (\"cookie: %s \\n\"%cookie)\n\npage = resp.text\nform_action = html.unescape(re.search(' : %s\\n\"%form_action)\n\n\nresp = requests.post(\n url=form_action, \n data={\n \"username\": username,\n \"password\": password,\n }, \n headers={\"Cookie\": cookie},\n allow_redirects=False\n)\n\n\nredirect = resp.headers['Location']\nprint(\"==================\")\nprint(\"redirect url : %s\\n\"%redirect)\n\nquery = urllib.parse.urlparse(redirect).query\nprint(\"==================\")\nprint(\"redirect_query : %s\\n\"%query)\nredirect_params = urllib.parse.parse_qs(query)\nprint(\"==================\")\nprint(\"redirect_params : %s\\n\"%redirect_params)\n\nauth_code = redirect_params['code'][0]\nprint(\"==================\")\nprint('auth-code: {0}\\n'.format(auth_code))\n\nresp = requests.post(\n url=openid_provider + \"/protocol/openid-connect/token\",\n data={\n \"client_secret\":client_secret,\n \"grant_type\": \"authorization_code\",\n \"client_id\": client_id,\n \"redirect_uri\": redirect_uri,\n \"code\": auth_code,\n \"code_verifier\": code_verifier,\n },\n allow_redirects=False\n)\n\nresult = resp.json()\nprint(\"==================\")\nprint(json.dumps(result,indent=2))\n\naccess_token=result['access_token']\nprint(\"==================\")\nprint (\"Access Token: %s\\n\" %access_token)\n\njwt=jwt_payload_decode(result['access_token'])\nprint(\"==================\")\nprint(\"JWT Token Decode:\")\nprint(json.dumps(jwt,indent=2))\n\nid_token=result['id_token']\nprint(\"==================\")\nprint(\"\\nID Token: %s\\n\" %id_token)\n\n\n\ncognito_idp_client = boto3.client('cognito-identity',region_name=default_region_name)\n# response = cognito_idp_client.describe_identity_pool(\n# IdentityPoolId=identity_pool_id\n# )\n\n# print(json.dumps(response,indent=2))\n\n#这里非常的关键,iam_openid_provider_name 应该是 IAM 中身份提供商中的提供商的名字\n#受众应该是 keycloak releam中的 client name\ncustom_logins={iam_openid_provider_name : id_token}\n\n\nresponse = cognito_idp_client.get_id(IdentityPoolId=identity_pool_id, Logins=custom_logins)\nidentity_id = response['IdentityId']\nprint(\"==================\")\nprint (\"\\nIdentity ID: %s\" %identity_id)\n\n\nresp = cognito_idp_client.get_credentials_for_identity(IdentityId=identity_id,Logins=custom_logins)\n\nsecretKey = resp['Credentials']['SecretKey']\naccessKey = resp['Credentials']['AccessKeyId']\nsessionToken = resp['Credentials']['SessionToken']\nexpiration = resp['Credentials']['Expiration']\nprint(\"==================\")\nprint (\"\\nSecret Key: %s\"%(secretKey))\nprint(\"==================\")\nprint (\"\\nAccess Key: %s\"%(accessKey))\nprint(\"==================\")\nprint (\"\\nSession Token: %s\"%(sessionToken))\nprint(\"==================\")\nprint (\"\\nExpiration: %s\"%(expiration))\nprint(\"==================\")\n\n","sub_path":"source/keycloak_integrate_with_IAM.py","file_name":"keycloak_integrate_with_IAM.py","file_ext":"py","file_size_in_byte":5320,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"76740617","text":"import re\nfrom typing import Optional, List, DefaultDict, Tuple, NamedTuple, Dict, Any\nfrom datetime import datetime\nfrom collections import defaultdict\nimport urllib.request\nfrom pyquery import PyQuery as pq\n\nfrom .base import ServerName, User, Problem, Submission, Result\nfrom .vcserver import VCServer\n\n\nSubmissionDict = DefaultDict[User, DefaultDict[Problem, List[Submission]]]\nWJ_data = NamedTuple(\"WJ_data\", [(\"url\", str), (\"user\", User), (\"problem\", Problem), (\"submission_number\", int)])\n\n\nclass AtCoderServer(VCServer):\n name: ServerName = ServerName(\"AtCoder\")\n DEFAULT_TIME: datetime = datetime(1, 1, 1, 0, 0, 0)\n RESULT_LABELS: List[str] = [\"AC\", \"CE\", \"MLE\", \"TLE\", \"RE\", \"OLE\", \"IE\", \"WA\", \"WJ\", \"WR\", \"NG\"]\n RESULT_TYPES: List[Result] = [Result.AC, Result.CE, Result.MLE, Result.TLE, Result.RE, Result.OLE, Result.IE, Result.WA, Result.WJ, Result.WR, Result.NG]\n\n def __init__(self,\n since: datetime,\n until: datetime) -> None:\n super().__init__(since, until)\n self.submissions: SubmissionDict = defaultdict(lambda: defaultdict(lambda: []))\n self.count: int = 0 # submission の id とか time をテキトーに生成するために使う\n self.submission_urls: List[str] = [] # 提出画面urlのリスト\n self.time_cache_until: datetime = since\n self.time_cache_since: datetime = since\n self.table_header_names: List[str] = []\n self.WJ_list: List[WJ_data] = [] # resultがWJになっている提出のリスト。詳細画面のurl、User、problem、その提出がsubmissions[user][problem]の何番目かを示す数が入る。\n\n def update_problems(self) -> None:\n for problem in self.problems:\n if problem.server_name == self.name:\n with urllib.request.urlopen(problem.url) as response:\n html_str = response.read().decode(\"utf-8\")\n query = pq(html_str)\n name_query = query(\"#main-container > div.row > div:nth-child(2) > span\")\n name_str = name_query.text()\n name_match = re.search(r\"-\\s\", name_str)\n if name_match is None:\n raise Exception(\"This can't happen!\")\n problem_name = name_str[name_match.end():]\n score_query = query(\"#task-statement > span > span.lang-ja > p > var\")\n score_str = score_query.text()\n if score_str == \"\":\n problem_score = 0\n else:\n problem_score = int(score_str)\n problem.set_data(name=problem_name, score=problem_score)\n\n def update_submissions(self) -> None:\n # if self.time_cache_until == self.DEFAULT_TIME:\n # self.time_cache_until = start\n if self._since < self.time_cache_since:\n target_time = self._since\n else:\n target_time = self.time_cache_until\n self.WJ_reload()\n self.make_submission_urls()\n now_time = datetime.now()\n for submission_url in self.submission_urls:\n must_check_next_page = True\n page_count = 1\n # 1ページごとにtime_cache_untilで行われた提出まで見ていく\n while must_check_next_page:\n sub_submission_url = submission_url + \"?page=\" + str(page_count)\n rows = self.make_page_rows(sub_submission_url)\n for row in rows:\n row_continue = self.row_process(row, target_time)\n if not row_continue:\n must_check_next_page = False\n break\n page_count += 1\n self.time_cache_until = min(self._until, now_time)\n self.time_cache_since = self._since\n # if end is not None:\n # if self.time_cache_until > end:\n # self.time_cache_until = end\n\n def get_submissions(self,\n user: User,\n problem: Problem) -> List[Submission]:\n # defaultdict なので存在しないキーにアクセスしても平��\n return self.submissions[user][problem]\n\n def accept_url(self, url: str) -> bool:\n return bool(re.match(r\"https?://atcoder\\.jp/contests/\", url))\n\n def WJ_reload(self) -> None:\n \"\"\"WJ_listに入っているWJの提出を再確認し、ジャッジが決定していたら\n 提出のresultを更新する。ここでは通信を行う\"\"\"\n new_WJ_list: List[WJ_data] = []\n for WJ in self.WJ_list:\n with urllib.request.urlopen(WJ.url) as response:\n html_str = response.read().decode(\"utf-8\")\n query = pq(html_str)\n # result_queryはchromeでWJ中の提出画面に行ったときのセレクタ、すなわち\n # #main-container > div.row > div:nth-child(2) > div:nth-child(8) > table\n # > tbody > tr:nth-child(7) > td > span\n # とは異なる。なぜかこうでないと正しく取得できない。\n # とりあえず再びWJ(誤答ではない)になるとき、WJ中TLEしてから再びWJ中TLEをみたとき、\n # ACになるとき、CEになるときに動くことは確認できた。\n result_query = query(\"#main-container > div.row > div:nth-child(2) > div:nth-child(6) > table > tr:nth-child(7) > td > span\")\n result_text = result_query.text()\n result = self.determine_result(result_text)\n if result is not Result.WJ:\n self.submissions[WJ.user][WJ.problem][WJ.submission_number].result = result\n else:\n new_WJ_list.append(WJ)\n self.WJ_list = new_WJ_list\n\n def make_submission_urls(self) -> None:\n \"\"\"現在登録されている問題をみて、その提出画面のurlの\n リストを作る。2回目以降に呼び出した場合、差分があれば追加する\"\"\"\n for problem in self.problems:\n base_url = problem.url\n url_match = re.search(\"tasks\", base_url)\n if url_match is None:\n raise Exception(\"This can't happen!\")\n submission_url = base_url[0:url_match.start()] + \"submissions\"\n if submission_url not in self.submission_urls:\n self.submission_urls.append(submission_url)\n\n def make_page_rows(self, sub_submission_url: str) -> pq:\n \"\"\"提出画面のurlをページごとに指定されたとき、\n そこから提出たちを表す部分を取り出して返す。ここでは通信を行う\"\"\"\n with urllib.request.urlopen(sub_submission_url) as response:\n html_str = response.read().decode(\"utf-8\")\n query = pq(html_str)\n # 表の親を取ってくる\n table = query(\"#main-container > div.row > div:nth-child(3) > div.panel.panel-default.panel-submission > div.table-responsive > table\")\n if self.table_header_names == []:\n # 表のヘッダ部を取ってくる\n theads = table(\"thead > tr > th\")\n # 表のヘッダ部の名前をリストに格納する\n for thead in theads:\n thead_text = pq(thead).text()\n self.table_header_names.append(thead_text)\n # 各行のデータをとってくる\n rows = table(\"tbody > tr\")\n return rows\n\n def row_process(self, row: Any, target_time: datetime) -> bool:\n \"\"\"update_submissionにおいて、1つ1つの提出をpyqueryの形から\n submissionの形になるように処理する。\n そして、提出の時間がtime_cache_untilより前になっていないか、この先の提出も\n 見るべきかどうかを返す\"\"\"\n row_contents: Dict[str, pq] = dict()\n columns = pq(row)(\"td\")\n for column_name, column in zip(self.table_header_names, columns):\n row_contents[column_name] = pq(column)\n # まず時間を確認する\n # 取得すべき提出はtarget_timeからself._untilまでだが、\n # このうちself.time_cache_sinceからself.time_cache_untilまでの提出は\n # すでに取得しているはずなので無視してよい\n row_time_str = row_contents[self.table_header_names[0]].text()\n row_time = datetime.strptime(row_time_str, \"%Y-%m-%d %H:%M:%S+0900\")\n if row_time > self._until:\n return True\n if row_time < target_time:\n return False\n # untilの方は、前回の取得のタイミングにより、\n # untilちょうどの時刻の1秒のうち0.何秒しか取得できていない\n # 場合があることを考えて、>にしておく。\n if self.time_cache_until > row_time and row_time >= self.time_cache_since:\n return True\n # ユーザーを確認する\n user = self.user_check(row_contents)\n if user is None:\n return True\n # 問題を確認する\n problem = self.problem_check(row_contents)\n if problem is None:\n return True\n # idかぶりを確認する\n already_submitted, id_ = self.check_id_duplicated(user, problem, row_contents)\n if already_submitted:\n return True\n # ここまできたらsubmissions行きは確定(WJも登録する?)\n result = self.result_check(user, problem, row_contents)\n # result_text = row_contents[self.table_header_names[6]].text()\n # for label, result_type in zip(self.RESULT_LABELS, self.RESULT_TYPES):\n # # 必ずどこかでresultが設定される(はず)ので例外処理なし\n # if result_text == label:\n # result = result_type\n score_text = row_contents[self.table_header_names[4]].text()\n submission = Submission(problem=problem,\n user=user,\n result=result,\n score=int(score_text),\n time=row_time,\n id_=id_)\n # defaultdict なので存在しないキーにアクセスしても平気\n self.submissions[user][problem].append(submission)\n return True\n\n def user_check(self, row_contents: Dict[str, pq]) -> Optional[User]:\n \"\"\"update_submissionにおいて、与えられた提出が\n コンテスト参加者のものになっているかどうかを確認し、参加者ならそのuserを、\n そうでないならNoneを返す\"\"\"\n row_user_text = row_contents[self.table_header_names[2]].text()\n for user in self.users:\n if row_user_text == user.ids[self.name]:\n return user\n return None\n\n def problem_check(self, row_contents: Dict[str, pq]) -> Optional[Problem]:\n \"\"\"update_submissionにおいて、与えられた提出が\n コンテストの問題の提出になっているかどうかを確認し、コンテスト問題ならその問題を、\n そうでないならNoneを返す\n \"\"\"\n question_url = row_contents[self.table_header_names[1]](\"a\").attr(\"href\") # urlはstr型\n for problem in self.problems:\n problem_url = problem.url\n problem_url_match = re.search(r\"/contests\", problem_url)\n if problem_url_match is None:\n raise Exception(\"This can't happen!\")\n cut_problem_url = problem_url[problem_url_match.start():]\n if question_url == cut_problem_url:\n # cutした方で比較しているのは、httpとhttpsの差があるかもしれないと思ったから\n return problem\n return None\n\n def check_id_duplicated(self,\n user: User,\n problem: Problem,\n row_contents: Dict[str, pq]) -> Tuple[bool, str]:\n \"\"\"update_submissionにおいて、提出のidがかぶっているかどうかと、id_textの内容を返す\"\"\"\n # idの取得はジャッジの表示が横に伸びている場合とそうでない場合に分ける\n if row_contents[self.table_header_names[7]].text() == \"Detail\":\n id_url = row_contents[self.table_header_names[7]](\"a\").attr(\"href\")\n id_url_match = re.search(r\"submissions/\", id_url)\n if id_url_match is None:\n raise Exception(\"This can't happen!\")\n id_ = id_url[id_url_match.end():]\n else:\n id_url = row_contents[self.table_header_names[9]](\"a\").attr(\"href\")\n id_url_match = re.search(r\"submissions/\", id_url)\n if id_url_match is None:\n raise Exception(\"This can't happen!\")\n id_ = id_url[id_url_match.end():]\n # idかぶりを検証(ページ読み込みの間に1ページ目から2ページ目に提出がずれた場合などにidかぶりが起きうる)\n for old_submission in self.submissions[user][problem]:\n if old_submission.id_ == id_:\n return True, id_\n return False, id_\n\n def result_check(self,\n user: User,\n problem: Problem,\n row_contents: Dict[str, pq]) -> Result:\n \"\"\"update_submissionにおいて、提出の結果を返す。\n \"3/18 WA\"などとなっている場合は、処理中としてWJを返すことにする\"\"\"\n # ここまできたらsubmissions行きは確定(WJも登録する?)\n result_text = row_contents[self.table_header_names[6]].text()\n result = self.determine_result(result_text)\n if result is Result.WJ:\n # WJのとき、詳細画面のurlを取得し、WJ_listに追加する\n # WJのときはジャッジの表示が横に伸びているはず\n if row_contents[self.table_header_names[7]].text() == \"Detail\":\n id_url = r\"https://atcoder.jp\" + row_contents[self.table_header_names[7]](\"a\").attr(\"href\")\n WJ_submission = WJ_data(id_url, user, problem, len(self.submissions[user][problem]))\n self.WJ_list.append(WJ_submission)\n else:\n raise Exception(\"This can't happen!\")\n return result\n\n def determine_result(self, result_text: str) -> Result:\n \"\"\"サイトからとってきた提出結果のテキストから、結果を決定する。\n result_checkおよびWJ_reloadで用いている\"\"\"\n result_text_judging_match = re.match(r\"\\d+/\\d+\", result_text)\n if result_text_judging_match:\n return Result.WJ\n for label, result_type in zip(self.RESULT_LABELS, self.RESULT_TYPES):\n # 必ずどこかでresultが設定される(はず)ので例外処理なし\n if result_text == label:\n return result_type\n raise Exception(\"This can't happen!\")\n","sub_path":"cmd_line_bot/virtual_contest/atcoder_server.py","file_name":"atcoder_server.py","file_ext":"py","file_size_in_byte":14998,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"314413367","text":"import asyncio\nimport os\nimport signal\nfrom time import sleep, time\n\nimport pytest\nfrom async_timeout import timeout\n\nfrom aiomisc.process_pool import ProcessPoolExecutor\n\n\n@pytest.fixture\ndef pool():\n pool = ProcessPoolExecutor(8)\n try:\n yield pool\n finally:\n pool.shutdown(True)\n\n\nasync def test_simple(pool, loop, timer):\n current_time = await loop.run_in_executor(pool, time)\n assert current_time > 0\n\n async with timeout(2):\n with timer(1):\n await asyncio.gather(\n *[\n loop.run_in_executor(pool, sleep, 1) for _ in range(8)\n ]\n )\n\n\nasync def test_exception(pool, loop):\n with pytest.raises(ZeroDivisionError):\n await loop.run_in_executor(pool, divmod, 1, 0)\n\n\ndef suicide():\n os.kill(os.getpid(), signal.SIGINT)\n\n\nasync def test_exit(pool, loop):\n async with timeout(2):\n with pytest.raises(asyncio.CancelledError):\n await asyncio.gather(\n *[loop.run_in_executor(pool, suicide) for _ in range(8)]\n )\n","sub_path":"tests/test_process_pool.py","file_name":"test_process_pool.py","file_ext":"py","file_size_in_byte":1075,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"55891330","text":"from config import Config\nfrom DirectedWeightedGraph import DWGraph\nimport sys\n\n\ndef get_graph(file_path, is_weighted, make_undirected, skip_first_line, labeling_function, labeling_file):\n g = DWGraph.from_file_edge_list(file_path,\n is_weighted=is_weighted,\n make_undirected=make_undirected,\n skip_first_line=skip_first_line)\n # Add labels to graph\n if labeling_function is not None:\n g.add_labels_by_identifiers(labeling_function)\n elif labeling_file is not None:\n g.add_labels_from_file(labeling_file)\n else:\n g.add_default_labels()\n\n return g\n\n\ndef print_info(info, file_path):\n with open(file_path, 'w') as f:\n for (u, v), data in info.items():\n if isinstance(data, float):\n f.write(\"{}\\t{}\\t{}\\n\".format(u, v, data))\n elif isinstance(data, dict):\n for w, r in data.items():\n f.write(\"{}\\t{}\\t{}\\t{}\\n\".format(u, v, w, r))\n else:\n exit(\"Bad info format.\")\n\n\ndef demo(c: Config):\n # Before making undirected\n print(\"=== Before making undirected ===\")\n print(\"Reading and labeling graphs...\")\n g = get_graph(c.background_file_path, c.background_weighted, False, c.background_skip_header,\n c.background_labeling_function, c.background_labeling_file)\n g0 = get_graph(c.template_file_path, c.template_weighted, False, c.template_skip_header,\n c.template_labeling_function, c.template_labeling_file)\n\n info = DWGraph.get_ratios(g, g0, c.edges_typing, True, c.edges_typing, True, c.edges_epsilon_match)\n print_info(info, c.output_info_ratios_before)\n\n del g, g0, info\n\n # After making undirected\n print(\"=== After making undirected ===\")\n print(\"Reading and labeling graphs...\")\n g = get_graph(c.background_file_path, c.background_weighted, not c.background_keep_directed,\n c.background_skip_header, c.background_labeling_function, c.background_labeling_file)\n g0 = get_graph(c.template_file_path, c.template_weighted, not c.background_keep_directed,\n c.template_skip_header, c.template_labeling_function, c.template_labeling_file)\n\n info = DWGraph.get_ratios(g, g0, c.edges_typing, True, c.edges_typing, True, c.edges_epsilon_match)\n print_info(info, c.output_info_ratios_after)\n\n\ndef main_demo():\n n = len(sys.argv)\n c_name = None\n if n == 1:\n c_name = 'default'\n elif n == 2:\n c_name = sys.argv[1]\n else:\n exit(\"Specify only one argument or no arguments if you want to run default configuration.\")\n\n try:\n c = Config.get_config(c_name)\n demo(c)\n\n except NameError:\n exit(\"There is no such configuration defined in config.py file.\")\n\n\nif __name__ == \"__main__\":\n main_demo()\n","sub_path":"python_graph_pattern_project/code/information_content.py","file_name":"information_content.py","file_ext":"py","file_size_in_byte":2906,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"89055359","text":"# -*- coding: utf-8 -*-\n#!/usr/bin/env python\n#-------------------------------------------------------------------------------\n# Name: danmo\n# Description: \n# Author: hlh\n# Date: 2021/1/25 23:21\n#-------------------------------------------------------------------------------\nimport requests\nimport re\nimport jieba\nimport wordcloud\nimport imageio\n\n\ndef get_response(url):\n headers = {\n 'cookie': \"rpdid=|(RllR)klm~0J'ulmJRJYlR); LIVE_BUVID=AUTO5015949072445733; CURRENT_QUALITY=80; _uuid=A0ECF0E4-2F5F-6AAC-34A2-0D95199B253983850infoc; buvid3=81910F6F-8B57-4F96-95E7-90ABD694CBD7138373infoc; blackside_state=1; CURRENT_FNVAL=80; sid=jma90skk; DedeUserID=368426889; DedeUserID__ckMd5=efa4bf9409d545bd; SESSDATA=9f6ce21f%2C1624711920%2C29bfc*c1; bili_jct=5da870cd2eccb25d05d3ba2ffce6969f; bp_t_offset_368426889=475290602926330600;bfe_id=018fcd81e698bbc7e0648e86bdc49e09\",\n 'origin': 'https://www.bilibili.com',\n 'referer': 'https://www.bilibili.com/video/BV19E41197Kc',\n 'user-agent': 'Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/81.0.4044.138 Safari/537.36',\n }\n response=requests.get(url,headers=headers)\n response.raise_for_status()\n response.encoding=response.apparent_encoding\n return response\n\n\ndef get_data(url):\n response=get_response(url)\n if response:\n json_data=response.json()\n if json_data['code']==0 and json_data['data']:\n return json_data['data']\n else:\n return None\n else:\n return None\n\n\n\ndef save_text(data_list):\n with open('bilibili弹幕.txt','a',encoding='utf-8') as f:\n for data in data_list:\n f.write(data+'\\n')\n\ndef get_word_cloud():\n #设置词云\n cloud_pic=imageio.imread('./20210125185743167.png')\n wc=wordcloud.WordCloud(width=1000,height=700,background_color='white',\n font_path='msyh.ttc',scale=15,mask=cloud_pic,\n stopwords={'哈','哈哈','哈哈哈哈哈'})\n with open('bilibili弹幕.txt',encoding='utf8') as f:\n cut_string=' '.join(jieba.lcut(f.read()))\n wc.generate(cut_string)\n wc.to_file('out.png')\n\n\n\n\n\ndef main(url):\n data_list=get_data(url)\n if data_list:\n for date in data_list:\n base_url=f'https://api.bilibili.com/x/v2/dm/web/history/seg.so?type=1&oid=284452148&date={date}'\n html_text=get_response(base_url).text\n results=re.findall('.*?([\\u4E00-\\u9FA5]+).*?',html_text,re.S)\n save_text(results)\n get_word_cloud()\n else:\n print('你访问的数据不存在')\n\n\n\n\n\n\nif __name__=='__main__':\n date_url='https://api.bilibili.com/x/v2/dm/history/index?type=1&oid=284452148&month=2021-01'\n main(date_url)","sub_path":"base_test/bilibili/danmu.py","file_name":"danmu.py","file_ext":"py","file_size_in_byte":2794,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"83113550","text":"from django.shortcuts import render, redirect\nfrom .forms import *\nfrom .models import *\nimport re\nfrom django.conf import settings\nfrom django.core.files.storage import FileSystemStorage\nfrom django.http import JsonResponse\nfrom decimal import *\nfrom datetime import date, datetime\n\n\"\"\"helps save landline/mobile/email to a customer object instance\"\"\"\ndef add_a_customer_contact_person_helper_function(request, ccpf, count_):\n errors = False\n customer = Customer_Contact_Person.objects.filter(pk=ccpf.pk)\n for count in range(\n len(request.POST.getlist(\n 'landline_number' + str(count_)))):\n clnf = CustomerLandlineNumberForm({\n 'customer': customer,\n 'landline_number': request.POST.getlist(\n 'landline_number' + str(count_))[count],\n })\n if clnf.is_valid():\n clnf.save()\n else:\n errors = True\n\n for count in range(\n len(request.POST.getlist(\n 'contact_number' + str(count_)))):\n cmnf = CustomerMobileNumberForm({\n 'customer': customer,\n 'mobile_number': request.POST.getlist(\n 'contact_number' + str(count_))[count],\n })\n if cmnf.is_valid():\n cmnf.save()\n else:\n errors = True\n\n for count in range(\n len(request.POST.getlist(\n 'email' + str(count_)))):\n cef = CustomerEmailForm({\n 'customer': customer,\n 'email': request.POST.getlist(\n 'email' + str(count_))[count],\n })\n if cef.is_valid():\n cef.save()\n else:\n errors = True\n\n return errors\n\ndef customers_view_all(request, pk=None):\n customers = Customer.objects.all()\n args = {'customers': customers}\n return render(request, 'productdevelopment/customers_view_all.html', args)\n\ndef customers(request, pk=None):\n \"\"\" save a customer \"\"\"\n if request.method == 'POST' and pk == None:\n errors = False\n customerFormPK = 0\n cf = CustomerForm(request.POST)\n\n if cf.is_valid():\n cf = cf.save()\n # returns the keys matching the regex, ex ['contact_name1', 'contact_name2']\n contactPersons = [person for person in request.POST if re.match(r'^contact_name[0-9]*', person)]\n count = 0\n for count in range(len(contactPersons)):\n count += 1\n contactPersonName = request.POST.getlist('contact_name' + str(count))\n contactPersonDepartment = request.POST.getlist('department' + str(count))\n ccpf = CustomerContactPersonForm({\n 'customer': Customer.objects.filter(pk=cf.pk),\n 'contact_name': contactPersonName,\n 'department': contactPersonDepartment,\n })\n if ccpf.is_valid():\n ccpf = ccpf.save()\n errors = add_a_customer_contact_person_helper_function(\n request,\n ccpf,\n count)\n\n if request.POST.get('save') == 'saveAndRedirect' and errors == False:\n return redirect('/customers/' + str(cf.pk) + '/styles/')\n\n else:\n # TODO error\n pass\n\n else:\n return render(request, 'productdevelopment/add_a_customer.html')\n\n elif request.method == 'GET' and pk != None:\n customer = Customer.objects.get(pk=pk)\n args = {'customer': customer}\n return render(request, 'productdevelopment/customer_details.html', args) \n\n else:\n return render(request, 'productdevelopment/add_a_customer.html')\n\ndef styles_view_all(request, pk=None):\n styles = Style.objects.filter(customer__exact=pk)\n customer = Customer.objects.get(pk__exact=pk)\n args = {'styles': styles, 'customer': customer}\n return render(request, 'productdevelopment/styles_view_all.html', args)\n\ndef styles_redirect(request):\n args = {'customers': Customer.objects.all()}\n return render(request, 'productdevelopment/styles_redirect.html', args)\n\ndef styles(request, pk=None, s_id=None):\n if request.method == 'POST':\n customer = Customer.objects.filter(pk=int(request.POST.get('customer')))\n sf = StyleForm({\n 'customer': customer,\n 'style_name': request.POST.get('style_name'),\n 'style_id': request.POST.get('style_id'),\n 'collection_name': request.POST.get('collection_name'),\n 'description': request.POST.get('description'),\n # 'style_sketch_or_photo': request.POST.get('style_sketch_or_photo')\n }, request.FILES)\n print(sf.errors)\n\n if sf.is_valid():\n sf = sf.save()\n print('valid')\n return redirect('/customers/'+ str(pk) +'/styles/' + str(sf.pk) + '/sizespecsheets/')\n\n else:\n print('error')\n return redirect('/')\n\n elif request.method == 'GET' and s_id != None:\n print('2')\n pass # return style details\n\n else:\n print('3')\n args = {'customer': Customer.objects.get(pk=pk),}\n return render(request, 'productdevelopment/add_new_style.html', args)\n # just add customer, probably will never be used\n\ndef bra_size_existence_or_save(query):\n queryDict = Bra_Sizes.objects.filter(name__iexact=query)\n if len(queryDict) != 0:\n return queryDict[0].pk\n else:\n braSize = BraSizesForm({\n 'name': query\n })\n if braSize.is_valid():\n braSize = braSize.save()\n return braSize.pk\n\ndef query_bra_name_existence_or_save(query):\n queryDict = Bra_Part_Names.objects.filter(name__iexact=query)\n if len(queryDict) != 0:\n return queryDict[0].pk\n else:\n braPartName = BraPartNamesForm({\n 'name': query\n })\n if braPartName.is_valid():\n braPartName = braPartName.save()\n return braPartName.pk\n\ndef size_spec_sheets(request, pk=None, s_id=None):\n if request.method == 'POST':\n bpn_list = request.POST.getlist('bra_part_name')\n for count in range(len(bpn_list)):\n key = query_bra_name_existence_or_save(bpn_list[count])\n braPartSpec = BraPartSpecificationForm({\n 'style': s_id,\n 'letter': request.POST.getlist('letter')[count],\n 'bra_part_name': key,\n 'comment': request.POST.getlist('comment')[count],\n 'tolerance': Decimal(\n request.POST.getlist('tolerance')[count] \n ),\n 'arrangement': count,\n })\n print(braPartSpec.errors)\n if braPartSpec.is_valid():\n braPartSpec.save()\n print('saved')\n\n sizes = request.POST.getlist('sizes')\n print(sizes)\n # ['32A', '32B', '34A', '34B', '36A', '36B']\n size_value_list = [size for size in request.POST if re.match(r'^size_[0-9]*', size)]\n # size value LIST:['size_32A', 'size_32B', 'size_34A', 'size_34B', 'size_36A', 'size_36B']\n print(size_value_list) \n for count in range(len(size_value_list)):\n values = request.POST.getlist(size_value_list[count])\n separator = ','\n specs = separator.join(values)\n print(\"specs\" + str(specs))\n key = bra_size_existence_or_save(sizes[count])\n print(key)\n bssf = BraSizeSpecsForm({\n 'style': Style.objects.filter(pk__exact=s_id),\n 'size': Bra_Sizes.objects.filter(pk__exact=key),\n 'specs': specs,\n })\n print(bssf.errors)\n if bssf.is_valid():\n bssf.save()\n # TODO HERE\n return redirect('/customers/'+ str(pk) +'/styles/' + str(s_id) + '/cost/')\n\n else:\n args = {\n 'customer': Customer.objects.get(pk=pk),\n 'style': Style.objects.get(pk=s_id),\n 'bra_parts': Bra_Part_Names.objects.all(),\n }\n return render(request, 'productdevelopment/add_size_spec_sheet.html', args)\n\ndef cost_information(request, pk=None, s_id=None):\n if request.method == 'POST':\n material = request.POST.getlist('material')\n print(material)\n bra_sizes = Bra_Size_Specs.objects.filter(style__exact=s_id).values_list('size')\n bra_sizes = Bra_Sizes.objects.filter(pk__in=bra_sizes) \n size = [str(bra.name) for bra in bra_sizes]\n print(size)\n\n for count in range(len(material)):\n cif = CostInformationForm({\n 'style': Style.objects.filter(pk=s_id),\n 'material': Raw_Material.objects.filter(pk__exact=material[count]),\n 'arrangement': int(count),\n })\n print(cif.errors)\n if cif.is_valid():\n cif = cif.save()\n \n\n for count in range(len(size)):\n size_specs = request.POST.getlist(str(size[count]))\n separator = ','\n size_consumption = separator.join(size_specs)\n print(size_consumption)\n ciscf = CostInformationSizeConsumptionForm({\n 'style': Style.objects.filter(pk=s_id),\n 'size': Bra_Sizes.objects.filter(name__exact=size[count]),\n 'consumption': size_consumption,\n })\n print(ciscf.errors)\n if ciscf.is_valid():\n print('valid')\n ciscf = ciscf.save()\n \n return redirect('/customers/'+ str(pk) +'/styles/' + str(s_id) + '/labor/')\n\n else:\n sizes = Bra_Size_Specs.objects.filter(style__exact=s_id).values_list('size')\n print(sizes)\n bra_sizes = Bra_Sizes.objects.filter(pk__in=sizes)\n print(bra_sizes)\n raw_material = Raw_Material.objects.all()\n args = {'bra_sizes': bra_sizes, 'raw_material': raw_material, 'customer': Customer.objects.get(pk=pk),}\n return render(request, 'productdevelopment/cost_information.html', args)\n\ndef labor_cost(request, pk=None, s_id=None):\n if request.method == 'POST':\n dcf = DirectCostForm({\n 'style': s_id,\n 'cutting_cost': request.POST.get('cutting_cost'),\n 'sewing_cost': request.POST.get('sewing_cost'),\n 'washing_cost': request.POST.get('washing_cost'),\n 'finishing_cost': request.POST.get('finishing_cost'),\n 'examining_cost': request.POST.get('examining_cost'),\n 'pressing_cost': request.POST.get('pressing_cost'),\n 'packaging_cost': request.POST.get('packaging_cost'),\n 'final_inspection_cost': request.POST.get('final_inspection_cost'),\n })\n print(dcf.errors)\n if dcf.is_valid():\n dcf.save()\n\n ocf = OverheadCostForm({\n 'style': s_id,\n 'rent_cost': request.POST.get('rent_cost'),\n 'utility_cost': request.POST.get('utility_cost'),\n 'paper_cost': request.POST.get('paper_cost'),\n 'machine_maintenance_cost': request.POST.get('machine_maintenance_cost'),\n 'transportation_cost': request.POST.get('transportation_cost'),\n })\n print(ocf.errors)\n if ocf.is_valid():\n ocf.save()\n \n return redirect('/customers/'+ str(pk) +'/styles/' + str(s_id) + '/operations/')\n else:\n return render(request, 'productdevelopment/labor_cost.html', {'customer': Customer.objects.get(pk=pk)})\n\ndef operation_breakdown(request, pk=None, s_id=None):\n if request.method == 'POST':\n print(len(request.POST.getlist('number')))\n style = Style.objects.filter(pk=pk)\n number = request.POST.getlist('number')\n operation_name_and_description = request.POST.getlist('operation_name_and_description')\n stitch = request.POST.getlist('stitch')\n minutes = request.POST.getlist('minutes')\n machine = request.POST.getlist('machine')\n for count in range(len(number)):\n of = OperationsForm({\n 'style': s_id,\n 'number': number[count],\n 'operation_name_and_description': operation_name_and_description[count],\n 'stitch': stitch[count],\n 'minutes': minutes[count],\n 'machine': machine[count],\n })\n print(of.errors)\n if of.is_valid():\n of = of.save()\n return redirect('/customers/'+ str(pk) +'/orders/'+ str(s_id))\n # return redirect\n else:\n return render(request, 'productdevelopment/operations_breakdown.html', {'customer': Customer.objects.get(pk=pk),})\n\ndef add_order_choose_style(request, pk=None):\n styles = Style.objects.filter(customer__pk__exact=pk)\n customers = Customer.objects.all()\n return render(request, 'productdevelopment/choose_customer_add_style_to_order.html', {'styles': styles, 'customers': customers})\n\ndef add_order_choose_customer(request):\n customers = Customer.objects.all()\n return render(request, 'productdevelopment/choose_customer_add_order.html', {'customers': customers})\n\ndef customer_orders_view_all(request, pk=None):\n orders = Order.objects.filter(customer=pk)\n args = {'orders': orders}\n return render(request, 'productdevelopment/view_orders.html', args)\n\ndef orders_view_all(request):\n orderListItemSize = OrderListItemSize.objects.all()\n args = {'item': orderListItemSize}\n return render(request, 'productdevelopment/view_orders.html', args)\n\ndef calculateCostPerSize(s_id):\n bra_sizes = Bra_Size_Specs.objects.filter(style=s_id).values('size')\n sizes = Bra_Sizes.objects.filter(pk__in=bra_sizes).values('name')\n consumption_col = Cost_Information_Size_Consumption.objects.filter(style=s_id)\n cost_row = Cost_Information.objects.filter(style=s_id).order_by('arrangement')\n\n cost = {}\n for count_cc in range(len(consumption_col)):\n values = (consumption_col[count_cc].consumption).split(',')\n\n # size and decimal list\n size = consumption_col[count_cc].size.name\n values_decimal = []\n for count_v in range(len(values)):\n values_decimal.append(Decimal(values[count_v]))\n \n material_costs = []\n print('----')\n for count_cr in range(len(cost_row)):\n material_costs.append(cost_row[count_cr].material.price)\n \n priceForCurrentSize = 0\n for count in range(len(values_decimal)):\n priceForCurrentSize += values_decimal[count]*material_costs[count]\n\n print(str(priceForCurrentSize) + \" \" + size)\n cost[str(size)] = priceForCurrentSize \n return cost\n\ndef orders(request, pk=None, s_id=None):\n if request.method == 'POST':\n of = OrderForm({\n 'customer': pk,\n 'style': s_id,\n 'purchase_order_number': request.POST.get('purchase_order_number'),\n 'delivery_address': request.POST.get('delivery_address'),\n })\n print(of.errors)\n print(request.POST)\n\n slice_interval = len(request.POST.getlist('sizes_'))\n\n if of.is_valid():\n of = of.save()\n cursor = 0\n start_delivery_date = request.POST.getlist('startdate')\n end_delivery_date = request.POST.getlist('enddate')\n for count_1 in range(len(start_delivery_date)):\n olif = OrderListItemForm({\n 'order': of.pk,\n 'start_delivery_date': datetime.strptime(start_delivery_date[count_1], \"%Y-%m-%d\"),\n 'end_delivery_date': datetime.strptime(end_delivery_date[count_1], \"%Y-%m-%d\"),\n })\n print(olif.errors)\n\n if olif.is_valid():\n olif = olif.save()\n for count_2 in range(slice_interval):\n print([cursor])\n olisf = OrderListItemSizeForm({\n 'orderlist': olif.pk,\n 'size': (Bra_Sizes.objects.filter(name__iexact=request.POST.getlist('sizes_')[count_2])).values('pk'),\n 'quantity': request.POST.getlist('amount\"')[cursor + count_2],\n })\n print(olisf.errors)\n\n if olisf.is_valid():\n olisf.save()\n cursor += slice_interval\n return render(request, 'productdevelopment/add_order.html', {'success': True})\n\n else:\n bra_sizes = Bra_Size_Specs.objects.filter(style=s_id).values('size')\n sizes = Bra_Sizes.objects.filter(pk__in=bra_sizes).values('name')\n # for price\n costs = calculateCostPerSize(s_id)\n # does not include labor and direct costs\n # gives back a style's cost per size/item\n # to get back total add labor/direct,\n # multiply chosen ORDER SIZES*PRICE PER SIZE*TOTAL\n print(costs)\n prices = []\n for key, val in costs.items():\n prices.append(val)\n print(sizes)\n args = {'sizes' :sizes, 'costs': costs, 'prices': prices}\n return render(request, 'productdevelopment/add_order.html', args)\n\ndef bra_part_names(request, query=None):\n data = {'results': []}\n bras = {}\n print(request.POST)\n if query is None:\n bras = Bra_Part_Names.objects.all()\n else:\n bras = Bra_Part_Names.objects.filter(name__contains=query)\n\n for bra in bras:\n data['results'].append({'bra_name': bra.name})\n\n return JsonResponse(data)\n","sub_path":"productdevelopment/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":17741,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"54666011","text":"from Tkinter import *\ndef main():\n root=Tk()\n global c\n c=Canvas(width=300,height=200,bg='black')\n c.pack()\n snowman()\n tree()\n mainloop()\n\ndef snowman():\n c.create_oval((35,120),(115,200),fill='white')\n c.create_oval((55,80),(95,120),fill='white')\n c.create_oval((67,92),(71,96),fill='yellow')\n c.create_oval((79,92),(83,96),fill='yellow')\n c.create_line((67,105),(75,112),(83,105),smooth=1)\n\ndef tree():\n c.create_rectangle((215,110),(235,200),fill='brown')\n c.create_polygon((185,110),(265,110),(225,70),fill='green')\n c.create_polygon((195,70),(255,70),(225,35),fill='green')\n\nmain()\n","sub_path":"5130309560_4/5130309560_4_11.py","file_name":"5130309560_4_11.py","file_ext":"py","file_size_in_byte":633,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"157451015","text":"STR_START = r'^'\nSTR_END = r'$'\nPERIOD = r'\\.'\nDOLLAR_SIGN = r'\\$'\nHYPHEN_MINUS = r'\\-'\nCARET = r'\\^'\nDIGIT = r'\\d'\nANY_CHARACTER = r'.'\nWORD_CHARACTER = r'\\w'\nNOT_WORD_CHARACTER = r'\\W'\n\nsimple_number_regex = f'{HYPHEN_MINUS}?{DIGIT}+{PERIOD}?{DIGIT}*'\nscientific_notation_regex = (f'{HYPHEN_MINUS}?'\n f'[1-9]({PERIOD}{DIGIT}*)?'\n f'(([Ee])|[Xx]10{CARET})'\n f'[+{HYPHEN_MINUS}]?{DIGIT}+')\n","sub_path":"naturalseagull/general_regexes.py","file_name":"general_regexes.py","file_ext":"py","file_size_in_byte":474,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"71134843","text":"import logging\n\nimport torch.nn as nn\nimport torch.utils.checkpoint as cp\nimport torch.nn.functional as F\nimport torch\n\nfrom mmcv.cnn import constant_init, kaiming_init\nfrom mmcv.runner import load_checkpoint\n\nfrom mmdet.ops import DeformConv, ModulatedDeformConv\nfrom ..registry import BACKBONES\nfrom ..utils import build_norm_layer\n\ninput_styles={'1000','2000','2000_simple','3000'}\n\ndef conv3x3(in_planes, out_planes, stride=1, dilation=1):\n \"3x3 convolution with padding\"\n return nn.Conv2d(\n in_planes,\n out_planes,\n kernel_size=3,\n stride=stride,\n padding=dilation,\n dilation=dilation,\n bias=False)\n\n\nclass BasicBlock(nn.Module):\n expansion = 1\n\n def __init__(self,\n inplanes,\n planes,\n stride=1,\n dilation=1,\n downsample=None,\n style='pytorch',\n with_cp=False,\n normalize=dict(type='BN'),\n dcn=None):\n super(BasicBlock, self).__init__()\n assert dcn is None, \"Not implemented yet.\"\n\n self.norm1_name, norm1 = build_norm_layer(normalize, planes, postfix=1)\n self.norm2_name, norm2 = build_norm_layer(normalize, planes, postfix=2)\n\n self.conv1 = conv3x3(inplanes, planes, stride, dilation)\n self.add_module(self.norm1_name, norm1)\n self.conv2 = conv3x3(planes, planes)\n self.add_module(self.norm2_name, norm2)\n\n self.relu = nn.ReLU(inplace=True)\n self.downsample = downsample\n self.stride = stride\n self.dilation = dilation\n assert not with_cp\n\n @property\n def norm1(self):\n return getattr(self, self.norm1_name)\n\n @property\n def norm2(self):\n return getattr(self, self.norm2_name)\n\n def forward(self, x):\n identity = x\n\n out = self.conv1(x)\n out = self.norm1(out)\n out = self.relu(out)\n\n out = self.conv2(out)\n out = self.norm2(out)\n\n if self.downsample is not None:\n identity = self.downsample(x)\n\n out += identity\n out = self.relu(out)\n\n return out\n\n\nclass Bottleneck(nn.Module):\n expansion = 4\n\n def __init__(self,\n inplanes,\n planes,\n stride=1,\n dilation=1,\n downsample=None,\n style='pytorch',\n with_cp=False,\n normalize=dict(type='BN'),\n dcn=None):\n \"\"\"Bottleneck block for ResNet.\n If style is \"pytorch\", the stride-two layer is the 3x3 conv layer,\n if it is \"caffe\", the stride-two layer is the first 1x1 conv layer.\n \"\"\"\n super(Bottleneck, self).__init__()\n assert style in ['pytorch', 'caffe']\n assert dcn is None or isinstance(dcn, dict)\n self.inplanes = inplanes\n self.planes = planes\n self.normalize = normalize\n self.dcn = dcn\n self.with_dcn = dcn is not None\n if style == 'pytorch':\n self.conv1_stride = 1\n self.conv2_stride = stride\n else:\n self.conv1_stride = stride\n self.conv2_stride = 1\n\n self.norm1_name, norm1 = build_norm_layer(normalize, planes, postfix=1)\n self.norm2_name, norm2 = build_norm_layer(normalize, planes, postfix=2)\n self.norm3_name, norm3 = build_norm_layer(\n normalize, planes * self.expansion, postfix=3)\n\n self.conv1 = nn.Conv2d(\n inplanes,\n planes,\n kernel_size=1,\n stride=self.conv1_stride,\n bias=False)\n self.add_module(self.norm1_name, norm1)\n fallback_on_stride = False\n self.with_modulated_dcn = False\n if self.with_dcn:\n fallback_on_stride = dcn.get('fallback_on_stride', False)\n self.with_modulated_dcn = dcn.get('modulated', False)\n if not self.with_dcn or fallback_on_stride:\n self.conv2 = nn.Conv2d(\n planes,\n planes,\n kernel_size=3,\n stride=self.conv2_stride,\n padding=dilation,\n dilation=dilation,\n bias=False)\n else:\n deformable_groups = dcn.get('deformable_groups', 1)\n if not self.with_modulated_dcn:\n conv_op = DeformConv\n offset_channels = 18\n else:\n conv_op = ModulatedDeformConv\n offset_channels = 27\n self.conv2_offset = nn.Conv2d(\n planes,\n deformable_groups * offset_channels,\n kernel_size=3,\n stride=self.conv2_stride,\n padding=dilation,\n dilation=dilation)\n self.conv2 = conv_op(\n planes,\n planes,\n kernel_size=3,\n stride=self.conv2_stride,\n padding=dilation,\n dilation=dilation,\n deformable_groups=deformable_groups,\n bias=False)\n self.add_module(self.norm2_name, norm2)\n self.conv3 = nn.Conv2d(\n planes, planes * self.expansion, kernel_size=1, bias=False)\n self.add_module(self.norm3_name, norm3)\n\n self.relu = nn.ReLU(inplace=True)\n self.downsample = downsample\n self.stride = stride\n self.dilation = dilation\n self.with_cp = with_cp\n self.normalize = normalize\n\n @property\n def norm1(self):\n return getattr(self, self.norm1_name)\n\n @property\n def norm2(self):\n return getattr(self, self.norm2_name)\n\n @property\n def norm3(self):\n return getattr(self, self.norm3_name)\n\n def forward(self, x):\n\n def _inner_forward(x):\n identity = x\n\n out = self.conv1(x)\n out = self.norm1(out)\n out = self.relu(out)\n\n if not self.with_dcn:\n out = self.conv2(out)\n elif self.with_modulated_dcn:\n offset_mask = self.conv2_offset(out)\n offset = offset_mask[:, :18, :, :]\n mask = offset_mask[:, -9:, :, :].sigmoid()\n out = self.conv2(out, offset, mask)\n else:\n offset = self.conv2_offset(out)\n out = self.conv2(out, offset)\n out = self.norm2(out)\n out = self.relu(out)\n\n out = self.conv3(out)\n out = self.norm3(out)\n\n if self.downsample is not None:\n identity = self.downsample(x)\n\n out += identity\n\n return out\n\n if self.with_cp and x.requires_grad:\n out = cp.checkpoint(_inner_forward, x)\n else:\n out = _inner_forward(x)\n\n out = self.relu(out)\n\n return out\n\n\ndef make_res_layer(block,\n inplanes,\n planes,\n blocks,\n stride=1,\n dilation=1,\n style='pytorch',\n with_cp=False,\n normalize=dict(type='BN'),\n dcn=None):\n downsample = None\n if stride != 1 or inplanes != planes * block.expansion:\n downsample = nn.Sequential(\n nn.Conv2d(\n inplanes,\n planes * block.expansion,\n kernel_size=1,\n stride=stride,\n bias=False),\n build_norm_layer(normalize, planes * block.expansion)[1],\n )\n\n layers = []\n layers.append(\n block(\n inplanes,\n planes,\n stride,\n dilation,\n downsample,\n style=style,\n with_cp=with_cp,\n normalize=normalize,\n dcn=dcn))\n inplanes = planes * block.expansion\n for i in range(1, blocks):\n layers.append(\n block(\n inplanes,\n planes,\n 1,\n dilation,\n style=style,\n with_cp=with_cp,\n normalize=normalize,\n dcn=dcn))\n\n return nn.Sequential(*layers)\n\n\n@BACKBONES.register_module\nclass ResNet(nn.Module):\n \"\"\"ResNet backbone.\n\n Args:\n depth (int): Depth of resnet, from {18, 34, 50, 101, 152}.\n num_stages (int): Resnet stages, normally 4.\n strides (Sequence[int]): Strides of the first block of each stage.\n dilations (Sequence[int]): Dilation of each stage.\n out_indices (Sequence[int]): Output from which stages.\n style (str): `pytorch` or `caffe`. If set to \"pytorch\", the stride-two\n layer is the 3x3 conv layer, otherwise the stride-two layer is\n the first 1x1 conv layer.\n frozen_stages (int): Stages to be frozen (all param fixed). -1 means\n not freezing any parameters.\n normalize (dict): dictionary to construct and config norm layer.\n norm_eval (bool): Whether to set norm layers to eval mode, namely,\n freeze running stats (mean and var). Note: Effect on Batch Norm\n and its variants only.\n with_cp (bool): Use checkpoint or not. Using checkpoint will save some\n memory while slowing down the training speed.\n zero_init_residual (bool): whether to use zero init for last norm layer\n in resblocks to let them behave as identity.\n \"\"\"\n\n arch_settings = {\n 18: (BasicBlock, (2, 2, 2, 2)),\n 34: (BasicBlock, (3, 4, 6, 3)),\n 50: (Bottleneck, (3, 4, 6, 3)),\n 101: (Bottleneck, (3, 4, 23, 3)),\n 152: (Bottleneck, (3, 8, 36, 3))\n }\n\n def __init__(self,\n depth,\n num_stages=4,\n strides=(1, 2, 2, 2),\n dilations=(1, 1, 1, 1),\n out_indices=(0, 1, 2, 3),\n style='pytorch',\n frozen_stages=-1,\n normalize=dict(type='BN', frozen=False),\n norm_eval=True,\n dcn=None,\n stage_with_dcn=(False, False, False, False),\n with_cp=False,\n zero_init_residual=True,\n input_style = '1000',\n use_head_v1 = False,\n use_deephead_v1=False,\n use_deephead_v2=False):\n super(ResNet, self).__init__()\n if depth not in self.arch_settings:\n raise KeyError('invalid depth {} for resnet'.format(depth))\n #self.larger_v1 = larger_v1\n self.use_deephead_v1 = use_deephead_v1\n self.use_deephead_v2 = use_deephead_v2\n self.use_head_v1 = use_head_v1\n self.depth = depth\n self.num_stages = num_stages\n assert num_stages >= 1 and num_stages <= 4\n self.strides = strides\n self.dilations = dilations\n assert len(strides) == len(dilations) == num_stages\n self.out_indices = out_indices\n assert max(out_indices) < num_stages\n self.style = style\n self.frozen_stages = frozen_stages\n self.normalize = normalize\n self.with_cp = with_cp\n self.norm_eval = norm_eval\n self.dcn = dcn\n self.stage_with_dcn = stage_with_dcn\n if dcn is not None:\n assert len(stage_with_dcn) == num_stages\n self.zero_init_residual = zero_init_residual\n\n self.input_style = input_style\n\n self.block, stage_blocks = self.arch_settings[depth]\n self.stage_blocks = stage_blocks[:num_stages]\n if self.input_style=='2000_v3':\n self.inplanes = 64\n else:\n self.inplanes = 64\n\n self._make_stem_layer()\n\n self.res_layers = []\n print(self.stage_blocks)\n for i, num_blocks in enumerate(self.stage_blocks):\n stride = strides[i]\n dilation = dilations[i]\n dcn = self.dcn if self.stage_with_dcn[i] else None\n planes = 64 * 2**i\n res_layer = make_res_layer(\n self.block,\n self.inplanes,\n planes,\n num_blocks,\n stride=stride,\n dilation=dilation,\n style=self.style,\n with_cp=with_cp,\n normalize=normalize,\n dcn=dcn)\n self.inplanes = planes * self.block.expansion\n layer_name = 'layer{}'.format(i + 1)\n self.add_module(layer_name, res_layer)\n self.res_layers.append(layer_name)\n\n self._freeze_stages()\n\n self.feat_dim = self.block.expansion * 64 * 2**(\n len(self.stage_blocks) - 1)\n if self.use_deephead_v1:\n self.deephead_1 = BasicConv2d(2048,2048,stride=2,kernel_size=3,padding=1) #for input size 2048 output 32x32\n self.deephead_2 = BasicConv2d(2048,2048,stride=2,kernel_size=3,padding=1) #16x16\n #self.deephead_3 = BasicConv2d(2048,2048,stride=2,kernel_size=3,padding=1) #8x8\n #self.deephead_4 = BasicConv2d(2048,2048,stride=2,kernel_size=3,padding=1) #4x4\n if self.use_deephead_v2:\n self.deephead_1 = BasicConv2d(2048,2048,stride=2,kernel_size=3,padding=1) #for input size 2048 output 32x32\n self.deephead_2 = BasicConv2d(2048,2048,stride=2,kernel_size=3,padding=1) #16x16\n self.deephead_3 = BasicConv2d(2048,2048,stride=2,kernel_size=3,padding=1) #8x8 \n @property\n def norm1(self):\n return getattr(self, self.norm1_name)\n\n def _make_stem_layer(self):\n if self.input_style=='1000' or self.input_style=='pool_stride_5_v1':\n self.conv1 = nn.Conv2d(\n 3, 64, kernel_size=7, stride=2, padding=3, bias=False)\n #self.maxpool = nn.MaxPool2d(kernel_size=3, stride=2, padding=1)\n elif self.input_style=='2000':\n self.conv1 = Conv2d_2000(3,64)\n #self.maxpool = nn.MaxPool2d(kernel_size=3, stride=2, padding=1)\n elif self.input_style=='2000_simple':\n self.conv1 = Conv2d_2000_simple(3,64)\n #self.maxpool = nn.MaxPool2d(kernel_size=3, stride=2, padding=1)\n elif self.input_style=='2000_v2':\n self.conv1 = Conv2d_2000_v2(3,64)\n elif self.input_style=='2000_v3':\n self.conv1 = Conv2d_2000_v3(3,64)\n elif self.input_style=='2000_v4':\n self.conv1 = Conv2d_2000_v4(3,64)\n elif self.input_style=='pool_stride_5_v2':\n self.conv1 = Conv2d_3(\n 3, 64)\n \n self.norm1_name, norm1 = build_norm_layer(\n self.normalize, 64, postfix=1)\n self.add_module(self.norm1_name, norm1)\n self.relu = nn.ReLU(inplace=True)\n if self.input_style == 'pool_stride_5_v1':\n self.maxpool = nn.MaxPool2d(kernel_size=15, stride=5, padding=7)\n elif self.input_style == 'pool_stride_5_v2':\n self.maxpool = nn.MaxPool2d(kernel_size=11, stride=4, padding=5)\n else:\n self.maxpool = nn.MaxPool2d(kernel_size=3, stride=2, padding=1)\n\n def _freeze_stages(self):\n if self.frozen_stages >= 0:\n for m in [self.conv1, self.norm1]:\n for param in m.parameters():\n param.requires_grad = False\n\n for i in range(1, self.frozen_stages + 1):\n m = getattr(self, 'layer{}'.format(i))\n for param in m.parameters():\n param.requires_grad = False\n\n def init_weights(self, pretrained=None):\n if isinstance(pretrained, str):\n logger = logging.getLogger()\n load_checkpoint(self, pretrained, strict=False, logger=logger)\n elif pretrained is None:\n for m in self.modules():\n if isinstance(m, nn.Conv2d):\n kaiming_init(m)\n elif isinstance(m, (nn.BatchNorm2d, nn.GroupNorm)):\n constant_init(m, 1)\n\n if self.dcn is not None:\n for m in self.modules():\n if isinstance(m, Bottleneck) and hasattr(\n m, 'conv2_offset'):\n constant_init(m.conv2_offset, 0)\n\n if self.zero_init_residual:\n for m in self.modules():\n if isinstance(m, Bottleneck):\n constant_init(m.norm3, 0)\n elif isinstance(m, BasicBlock):\n constant_init(m.norm2, 0)\n else:\n raise TypeError('pretrained must be a str or None')\n\n def forward(self, x):\n outs = []\n '''\n if self.input_style=='pool_stride_5_v2':\n x = self.conv11(x)\n else:\n '''\n x = self.conv1(x)\n if (self.input_style=='2000_v2' or self.input_style=='2000_v3' or self.input_style=='2000_v4')==False:\n x = self.norm1(x)\n x = self.relu(x)\n if self.use_head_v1:\n outs.append(x)\n x = self.maxpool(x)\n \n for i, layer_name in enumerate(self.res_layers):\n res_layer = getattr(self, layer_name)\n x = res_layer(x)\n if i in self.out_indices:\n outs.append(x)\n \n if self.use_deephead_v1:\n x = self.deephead_1(x)\n outs.append(x)\n x = self.deephead_2(x)\n outs.append(x)\n #x = self.deephead_3(x)\n #outs.append(x)\n #x = self.deephead_4(x)\n #outs.append(x)\n #print(x.shape)\n if self.use_deephead_v2:\n x = self.deephead_1(x)\n outs.append(x)\n x = self.deephead_2(x)\n outs.append(x)\n x = self.deephead_3(x)\n outs.append(x) \n if len(outs) == 1:\n return outs[0]\n else:\n return tuple(outs)\n\n def train(self, mode=True):\n super(ResNet, self).train(mode)\n if mode and self.norm_eval:\n for m in self.modules():\n # trick: eval have effect on BatchNorm only\n if isinstance(m, nn.BatchNorm2d):\n m.eval()\n\nclass BasicConv2d(nn.Module):\n\n def __init__(self, in_channels, out_channels, **kwargs):\n super(BasicConv2d, self).__init__()\n self.conv = nn.Conv2d(in_channels, out_channels, bias=False, **kwargs)\n self.bn = nn.BatchNorm2d(out_channels, eps=0.001)\n\n def forward(self, x):\n x = self.conv(x)\n x = self.bn(x)\n return F.relu(x, inplace=True)\n\nclass Conv2d_2000(nn.Module):\n\n def __init__(self, in_channels, out_channels, **kwargs):\n super(Conv2d_2000, self).__init__()\n self.conv1a = nn.Conv2d(in_channels, out_channels, bias=False, kernel_size=15, stride=5,padding = 7)\n self.conv1b = nn.Conv2d(in_channels, out_channels, bias=False, kernel_size=31, stride=5,padding = 15)\n self.conv1c = nn.Conv2d(in_channels, out_channels, bias=False, kernel_size=61, stride=5,padding = 30)\n\n def forward(self, x):\n x1 = self.conv1a(x)\n x2 = self.conv1b(x)\n x3 = self.conv1c(x)\n x = x1+x2+x3\n return x\nclass Conv2d_2000_v2(nn.Module):\n\n def __init__(self, in_channels, out_channels, **kwargs):\n super(Conv2d_2000_v2, self).__init__()\n self.conv1a = BasicConv2d(in_channels, out_channels, kernel_size=15, stride=5,padding = 7)\n self.conv1b = BasicConv2d(in_channels, out_channels, kernel_size=31, stride=5,padding = 15)\n self.conv1c = BasicConv2d(in_channels, out_channels, kernel_size=61, stride=5,padding = 30)\n\n def forward(self, x):\n x1 = self.conv1a(x)\n x2 = self.conv1b(x)\n x3 = self.conv1c(x)\n x = x1+x2+x3\n return x\nclass Conv2d_2000_v3(nn.Module):\n\n def __init__(self, in_channels, out_channels, **kwargs):\n super(Conv2d_2000_v3, self).__init__()\n temp_out_channels=out_channels*4\n self.conv1a = BasicConv2d(in_channels, temp_out_channels, kernel_size=15, stride=5,padding = 7)\n self.conv1b = BasicConv2d(in_channels, temp_out_channels, kernel_size=31, stride=5,padding = 15)\n self.conv1c = BasicConv2d(in_channels, temp_out_channels, kernel_size=61, stride=5,padding = 30)\n self.conv1d = BasicConv2d(temp_out_channels, out_channels, kernel_size=1, stride=1)\n\n def forward(self, x):\n x1 = self.conv1a(x)\n x2 = self.conv1b(x)\n x3 = self.conv1c(x)\n x = x1+x2+x3\n x = self.conv1d(x)\n return x\n\nclass Conv2d_2000_v4(nn.Module):\n\n def __init__(self, in_channels, out_channels, **kwargs):\n super(Conv2d_2000_v4, self).__init__()\n #self.conv1a = BasicConv2d(in_channels, 16, kernel_size=1, stride=1,padding = 0)\n #self.conv1b = BasicConv2d(16, 32, kernel_size=3, stride=2,padding = 1)\n self.conv1c = BasicConv2d(in_channels, out_channels, kernel_size=8, stride=4,padding = 1)\n\n def forward(self, x):\n #x = self.conv1a(x)\n #x = self.conv1b(x)\n x = self.conv1c(x)\n return x\n\nclass Conv2d_2000_simple(nn.Module):\n\n def __init__(self, in_channels, out_channels, **kwargs):\n super(Conv2d_2000_simple, self).__init__()\n self.conv1a = nn.Conv2d(in_channels, out_channels, bias=False, kernel_size=15, stride=5,padding = 7)\n\n def forward(self, x):\n x = self.conv1a(x)\n return x\n\nclass Conv2d_3(nn.Module):\n\n def __init__(self, in_channels, out_channels, **kwargs):\n super(Conv2d_3, self).__init__()\n self.conv1a = nn.Conv2d(in_channels, out_channels, bias=False, kernel_size=3, stride=2,padding = 1)\n\n def forward(self, x):\n x = self.conv1a(x)\n return x\n","sub_path":"mmdet/models/backbones/resnet.py","file_name":"resnet.py","file_ext":"py","file_size_in_byte":21665,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"528652453","text":"\"\"\"\nModule to test rest methods\n\"\"\"\n\nimport os\nimport asyncio\nimport unittest\nfrom datetime import datetime\nfrom dateutil.tz import tzutc\nfrom bfxapi.rest.BfxRest import BfxRest\n\n\nclass TestBfxRest(unittest.TestCase):\n \"\"\"\n Tests for BfxRest module\n \"\"\"\n\n API_KEY = os.getenv('BALANCE_TRACER_BITFINEX_KEY')\n API_SECRET = os.getenv('BALANCE_TRACER_BITFINEX_SECRET')\n\n\n def test_get_currency_movements(self):\n \"\"\"\n Test get_currency movements\n \"\"\"\n loop = asyncio.new_event_loop()\n bfx = BfxRest(self.API_KEY, self.API_SECRET, loop=loop, logLevel='DEBUG')\n\n movements = asyncio.run(bfx.get_currency_movements('BTC',\n datetime(2015, 9, 1, tzinfo=tzutc()),\n datetime(2019, 5, 1, tzinfo=tzutc())\n )\n )\n self.assertIsNotNone(movements)\n\n loop.close()\n\n\nif __name__ == '__main__':\n unittest.main()\n","sub_path":"bfxapi/tests/test_rest_movements.py","file_name":"test_rest_movements.py","file_ext":"py","file_size_in_byte":1073,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"609945501","text":"# -*- coding: utf-8 -*-\nimport random\nimport tensorflow as tf\n\n\ndef sample(combos, combo_l_mask, combo_a_mask, sample_num=100):\n \"\"\"\n 动作空间太大,则将采样\n \"\"\"\n if len(combos) <= sample_num:\n return combos, combo_l_mask, combo_a_mask\n else:\n idxes = random.sample([i for i in range(len(combos))], sample_num)\n asid, lmsk, amsk = [], [], []\n for idx in idxes:\n asid.append(combos[idx])\n lmsk.append(combo_l_mask[idx])\n amsk.append(combo_a_mask[idx])\n return asid, lmsk, amsk\n\n\ndef convert_h(h_vec, h_pid):\n \"\"\"\n 对入参填充和定义mask\n \"\"\"\n max_len = max([len(i) for i in h_pid])\n h_mask = [[1 for _ in i] for i in h_pid]\n h_vec = pad_2d(h_vec, max_len, pad_id=[0, 0, 0, 0, 0])\n h_mask = pad_2d(h_mask, max_len, pad_id=0)\n h_pid = pad_2d(h_pid, max_len, pad_id=0)\n\n return h_vec, h_pid, h_mask\n\n\ndef pad_2d(tensor, max_len, pad_id):\n \"\"\"\n 填充成矩阵\n \"\"\"\n res = []\n for i in tensor:\n j = []\n for k in i:\n j.append(k)\n while len(j) < max_len:\n j.append(pad_id)\n res.append(j)\n return res\n\n\ndef convert_alad(act_space_ids, label_masks, attn_masks, dynamic_corpus):\n \"\"\"\n 对入参填充\n \"\"\"\n dynamic_corpus = pad_3d(dynamic_corpus, pad_id=0)\n act_space_ids = pad_3d(act_space_ids, pad_id=0)\n label_masks = pad_3d(label_masks, pad_id=1)\n attn_masks = pad_3d(attn_masks, pad_id=0)\n return act_space_ids, label_masks, attn_masks, dynamic_corpus\n\n\ndef convert_aad(act_space_ids, attn_masks, dynamic_corpus):\n \"\"\"\n 对入参进行采样和填充\n \"\"\"\n dynamic_corpus = pad_3d(dynamic_corpus, pad_id=0)\n act_space_ids = pad_3d(act_space_ids, pad_id=0)\n attn_masks = pad_3d(attn_masks, pad_id=0)\n return act_space_ids, attn_masks, dynamic_corpus\n\n\ndef pad_3d(input_list, pad_id):\n \"\"\"\n 填充成矩阵\n \"\"\"\n max_1d = max([len(i) for i in input_list])\n max_2d = max([max([len(j) for j in i]) for i in input_list])\n\n tensor = []\n for i in input_list:\n d1 = []\n for j in i:\n d2 = []\n for k in j:\n d2.append(k)\n while len(d2) < max_2d:\n d2.append(pad_id)\n d1.append(d2)\n while len(d1) < max_1d:\n d2 = []\n for _ in range(max_2d):\n d2.append(pad_id)\n d1.append(d2)\n tensor.append(d1)\n\n return tensor\n\n\ndef clip_train_op(loss, params, optimizer):\n \"\"\"\n 梯度裁剪训练\n \"\"\"\n grads = tf.gradients(loss, params)\n real_grads = []\n real_vars = []\n for d_n, dn_v in zip(grads, params):\n if d_n == None:\n print('None:::', dn_v.name)\n continue\n real_grads.append(d_n)\n real_vars.append(dn_v)\n grads = real_grads\n params = real_vars\n (grads, _) = tf.clip_by_global_norm(grads, clip_norm=2.0)\n train_op = optimizer.apply_gradients(zip(grads, params))\n\n return train_op\n\n","sub_path":"ppo-version/rl/common_func.py","file_name":"common_func.py","file_ext":"py","file_size_in_byte":3048,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"383352330","text":"# This code is to construct a GUI for the force analysis\n# THis code is to handle multiple files\n\nfrom Tkinter import *\nfrom Tkinter import _setit\n\n\nimport matplotlib\n#matplotlib.use('TkAgg')\n\nimport numpy as np\n\nimport tkFileDialog, os\nimport function_MRI as MRI\n\nfrom matplotlib.backends.backend_tkagg import FigureCanvasTkAgg\nfrom matplotlib.figure import Figure # Tkinter has no figure widget\n\nimport FileDialog # For the py2applet\n\nclass data: # Use the 'dictionary' and 'class' to collect the data\n def __init__(self, x, y, y_cut):\n self.x = x # Initial data\n self.y = y\n self.y_cut = y_cut # Corrected data after the cut lines \n # (This is not collected yet.)\n \ndef open_snr_file(nn):\n \n global snr_height, label_x, label_y, title\n \n returned_values['filename'] = tkFileDialog.askopenfilename()\n filepath = returned_values.get('filename')\n \n temp_ref, snr_height, filename = MRI.signal_noise_ratio(filepath, nn) \n \n # Show the information of square wave\n info.insert(END, 'Load the square wave data: ' + filename + '\\n') \n info.insert(END, str(snr_height) + ' A means 100 pT \\n')\n info.update_idletasks()\n \n # Show the square wave \n label_x = 'Floating point'\n label_y = 'Current (A)'\n title = 'Square wave'\n refreshFigure(np.arange(nn), temp_ref, label_x, label_y, title) \n \ndef open_base_file():\n \n global temp_x0, temp_y0, label_x, label_y, title, filepath\n \n returned_values['filename'] = tkFileDialog.askopenfilename()\n filepath = returned_values.get('filename') \n temp_x0, temp_y0, filename = MRI.load_file(filepath)\n\n # Show the information of baseline\n info.insert(END, 'Load the baseline data: ' + filename + '\\n')\n info.update_idletasks()\n \n # Show the baseline data \n label_x = 'Position (mm)'\n label_y = 'Current (A)'\n title = 'Baseline: ' + filename\n refreshFigure(temp_x0, temp_y0, label_x, label_y, title) \n\ndef open_signal_file(cen_peak, space_peak, snr_height):\n \n global label_x, label_y, title, filepath, options, data_ini_xy\n label_y = 'Current (A)'\n label_x = 'Position (mm)'\n \n # Open multiple files \n filez = tkFileDialog.askopenfilenames() \n \n options = []\n data_ini_xy = {}\n for ii in range(0,len(filez)):\n if ii == 0: \n filepath = os.path.dirname(filez[0])\n \n name = os.path.basename(filez[ii])\n options.append(name) # Collect the name\n \n temp_x1, temp_y1, filename = MRI.load_file(filez[ii])\n signal_x, signal_y = MRI.position_filter(temp_x1, temp_y1, temp_x0, \n temp_y0, cen_peak, space_peak) \n \n # Change the unit\n if snr_height != 0:\n signal_y = signal_y/snr_height*100 # Change the unit to pT from nA\n label_y = 'Magnetic field (pT)'\n \n # Collect the data as the dictionary\n temp_data = data(signal_x, signal_y, None) \n data_ini_xy.update({name:temp_data}) \n \n # Refresh the options\n refresh(options)\n \n # Show the information of signal\n info.insert(END, 'Load the signal data: ' + str(options) + '\\n')\n info.update_idletasks()\n\ndef refresh(options):\n network_select['menu'].delete(0, 'end')\n\n for choice in options:\n network_select['menu'].add_command(label = choice, \n command = _setit(var, choice))\n\ndef correction_airPLS(signal_y, lambda_base, order, wep, p, itmermax):\n \n global signal_y_airPLS\n \n signal_y_airPLS, xb = MRI.airPLS(signal_y, lambda_base, order, wep, p, itmermax)\n \n # Show the information by airPLS\n info.insert(END, 'Show the result by airPLS \\n')\n info.update_idletasks() \n \n # Show the baseline by airPLS\n fig.plot(signal_x, xb, 'r-') # fitted baseline by airPLS\n canvas_frame.show() \n\ndef show_baseline_by_cut(cut_values, y):\n \n global signal_y_cut, error, data_ini_xy\n \n # yt is the fitted baseline\n yt = MRI.bfvar_temp(cut_values, np.arange(y.size), y)\n signal_y_cut = y-yt\n data_ini_xy[temp_file_name].y_cut = signal_y_cut\n \n ## Get the noise (for the error bar) \n temp = np.arange(np.int(np.min(np.floor(cut_values))))\n \n # Get the local value (1st)\n index_max, index_min = MRI.error_bar_ana(temp, signal_y_cut)\n \n # Get the local value (2nd)\n index_max_2, index_min_2 = MRI.error_bar_ana(index_max, signal_y_cut)\n index_max_3, index_min_3 = MRI.error_bar_ana(index_min, signal_y_cut)\n \n # The final one\n index_max = index_max_2\n index_min = index_min_3\n \n temp_all = np.union1d(index_min, index_max)\n \n yt_max = np.interp(temp_all, index_max, signal_y_cut[index_max]) \n yt_min = np.interp(temp_all, index_min, signal_y_cut[index_min]) \n\n# # Plot the figure of error to check \n# matplotlib.pylab.plot(temp,signal_y_cut[temp])\n# matplotlib.pylab.plot(temp_all,yt_max)\n# matplotlib.pylab.plot(temp_all,yt_min) \n\n # Get the error\n error_all = yt_max - yt_min\n error = np.median(error_all)\n \n # Format of error\n if abs(error)<1e-2:\n form_error = '%.2e'\n else:\n form_error = '%.2f'\n\n # Show the information by the cut lines\n info.insert(END, 'Show the result by the cut lines \\n')\n info.insert(END, 'The error is ' + str(form_error %error) + '\\n')\n info.update_idletasks() \n \n # Show the baseline by the cut lines\n fig.plot(np.arange(y.size), yt, 'r-') # fitted baseline by the cut lines\n canvas_frame.show()\n \ndef refreshFigure(x, y, label_x = None, label_y = None, title = None):\n \n global curve_x, curve_y \n \n # Give the initial setting \n if label_x == None:\n label_x = 'x' \n \n if label_y == None:\n label_y = 'y'\n \n if title == None:\n title = 'Signal'\n \n # For the exporting data\n curve_x = x\n curve_y = y\n \n # Plot the figure\n fig.clear()\n fig.plot(x,y)\n fig.set_xlabel(label_x) \n fig.set_ylabel(label_y) \n fig.set_title(title)\n canvas_frame.show()\n \ndef showXY_handler(event):\n \n # Show the x, y positions\n info_xy.delete(1.0, END)\n \n if (event.xdata != None)&(event.ydata != None):\n \n # Define the format of number\n if abs(event.xdata)<1e-2:\n form_x = '%.2e'\n else:\n form_x = '%.2f'\n \n if abs(event.ydata)<1e-2:\n form_y = '%.2e'\n else:\n form_y = '%.2f'\n \n info_xy.insert(END, '(' + str(form_x %event.xdata) + ', ' \n + str(form_y %event.ydata) + ')', 'center')\n \n info_xy.update_idletasks()\n \ndef cut_lines(event):\n \n global cut_values, index_cut_value, cen_peak\n \n if var_cut.get() == 1: # To check the value of check_button\n \n # Give the cut_values (x data) and restricted to 2\n if index_cut_value == 0:\n \n cut_values[0] = event.xdata\n fig.axvline(x = cut_values[0], color = 'r')\n index_cut_value = 1 # Make use of the setting of \n # index_cut_value (0 or 1) to collect two \n # values repeatly.\n \n else:\n \n cut_values[1] = event.xdata\n fig.axvline(x = cut_values[1], color = 'r')\n index_cut_value = 0\n \n if var_cen.get() == 1:\n cen_peak = event.xdata\n fig.axvline(x = cen_peak, color = 'r')\n \n canvas_frame.show()\n \ndef show_ini_signal(data_ini_xy):\n \n global signal_x, signal_y, temp_file_name, title\n \n temp_file_name = var.get()\n temp_xy = data_ini_xy[temp_file_name]\n signal_x = temp_xy.x\n signal_y = temp_xy.y\n \n # Show the signal data\n title = 'Signal: ' + temp_file_name\n refreshFigure(signal_x, signal_y, label_x, label_y, title) \n \ndef export_data(x, y, filepath, temp_file_name):\n \n data_xy = np.array([x, y])\n np.savetxt(filepath +'/' + temp_file_name + '_correct', data_xy.T) \n \n # Show the information\n info.insert(END, 'Corrected data is exported. \\n')\n info.update_idletasks() \n\ndef clear_text():\n info.delete(1.0, END)\n \n # Show the information\n info.insert(END, 'Information in the running: \\n')\n info.update_idletasks() \n\ndef fitting_curve():\n \n p0 = np.array([0, 1e4, 0, 9, cen_peak])\n plsq = MRI.leastsq(MRI.residuals, p0, args = (signal_y_cut, signal_x)) \n \n # Plot the fitting curve\n yfit = MRI.B_field_fit(signal_x, plsq[0]) \n \n # Show the baseline by the cut lines\n fig.plot(signal_x, yfit, 'r-') # fitted curve\n canvas_frame.show()\n\n # Change the format\n parameters = ['{:.2f}'.format(ii) for ii in plsq[0]]\n \n # Show the information\n info.insert(END, '[angle_m, M, signal_base, d, d_x] are: ' \n + str(parameters) + '\\n')\n info.update_idletasks() \n\n#-----------------------------------------\n#-----------------------------------------\n\n# Parameters for the global parts\nfilepath = '_correct'\nchoices = 'None'\ntemp_x0 = None\ntemp_y0 = None\nsignal_x = None\nsignal_y = None\ncurve_x = None\ncurve_y = None\nlabel_x = None\nlabel_y = None\ntitle = None\nsignal_y_airPLS = None\nsignal_y_cut = None\nsnr_height = 0\nerror = 0\nyt = None # Fitted baseline by the cut lines\n\ncut_values = [0, 0]\nindex_cut_value = 0\n\nreturned_values = {} # This is to get the path of file\n\n# The position we guess to be the peak\ncen_peak = 222\nspace_peak = 40\nnn = 1000\n\n# The coefficient for the airPLS\nlambda_base = 1e5\norder = 1\nwep = 0.3\np = 0.04\nitmermax = 200\n\n# Begin the GUI interface\nroot = Tk()\nroot.title('Data Analysis')\n\n# Initiate the menu and add to the root window\nmenubar = Menu(root)\nroot.config(menu = menubar)\n\n# Build the file menu\nfile_menu = Menu(menubar, tearoff = False)\nfile_menu.add_command(label = 'Load SNR', command = lambda: open_snr_file(nn))\nfile_menu.add_command(label = 'Load baseline/ bare data', \n command = open_base_file)\nfile_menu.add_command(label = 'Load signal after the base line', \n command = lambda: open_signal_file(cen_peak, space_peak, snr_height))\nfile_menu.add_command(label = \"Quit\", command = root.destroy)\n\n# all all submenus to main menu\nmenubar.add_cascade(label = \"File\", menu = file_menu)\n\n# Decide the frames\nleft = Frame(root)\nright = Frame(root)\nright_top = Frame(right)\nright_down = Frame(right)\n\nleft.grid(row = 0, column = 0)\nright.grid(row = 0, column = 1)\nright_top.grid(row = 0, column = 0)\nright_down.grid(row = 1, column = 0)\n\n# Plot the figure\n# The text widget to show the x, y positions\ninfo_xy = Text(left, height = 1) \ninfo_xy.tag_configure('center', justify = 'center')\ninfo_xy.pack()\n\n# The setting of figure\nmat_plot = Figure(figsize=(6,4), dpi = 100)\nfig = mat_plot.add_subplot(111)\nfig.set_xlabel('x')\nfig.set_ylabel('y')\n\n# Draw the figure by tk.DrawingArea\ncanvas_frame = FigureCanvasTkAgg(mat_plot, master = left)\ncanvas_frame.show()\ncanvas_frame.get_tk_widget().pack(side = 'top', fill = 'both', expand = 1)\ncanvas_frame._tkcanvas.pack(side = 'top', fill = 'both', expand = 1)\n\n# Help to show the x, y positions\ncanvas_frame.mpl_connect('motion_notify_event', showXY_handler)\n\n# Help to show the cut lines\ncanvas_frame.mpl_connect('button_press_event', cut_lines)\n\n# Choose the central position\n# Give the central line\nvar_cen = IntVar() # boolean value: var_cut.get()\ncheck_cen = Checkbutton(right_top, text = 'Show central line', variable = var_cen)\ncheck_cen.pack()\n\n# Show the opened files\nvar = StringVar(root)\nvar.set(choices)\n\n# all submenus to main menu\nnetwork_select = OptionMenu(right_top, var, choices)\nnetwork_select.pack()\n\n# Show the choised initial signal file\nbutton_show_ini = Button(right_top, text = 'Show initial data', \n command = lambda: show_ini_signal(data_ini_xy))\nbutton_show_ini.pack()\n\n# Baseline correction by airPLS (adaptive iteratively reweighted \n# Penalized Least Squares) \nbutton_show_airPLS = Button(right_top, text = 'Show by airPLS', \n command = lambda: correction_airPLS(signal_y, lambda_base, \n order, wep, p, itmermax))\nbutton_show_airPLS.pack()\n\nbutton_correct_airPLS = Button(right_top, text = 'Correct by airPLS',\n command = lambda: refreshFigure(np.arange(signal_y_airPLS.size), \n signal_y_airPLS, 'Floating point', \n label_y, title))\nbutton_correct_airPLS.pack()\n\n# Baseline correction by the linear method\n# Give the cut line\nvar_cut = IntVar() # boolean value: var_cut.get()\ncheck_cut = Checkbutton(right_top, text = 'Show cut lines', variable = var_cut)\ncheck_cut.pack()\n\n# Show the fitted baseline by the cut line\n# The input here is the signal_y_airPLS\nbutton_cut_fit = Button(right_top, text = 'Show the fitted baseline', \n command = lambda: show_baseline_by_cut(cut_values, signal_y_airPLS))\nbutton_cut_fit.pack()\n\nbutton_correct_cut_fit = Button(right_top, text = 'Correct by cut lines',\n command = lambda: refreshFigure(signal_x, signal_y_cut, label_x, \n label_y, title))\n\nbutton_correct_cut_fit.pack()\n\n# Refresh from the correction by the cut lines\nbutton_refresh = Button(right_top, text = 'Refresh from the cut lines',\n command = lambda: refreshFigure(np.arange(signal_y_airPLS.size), \n signal_y_airPLS, 'Floating point', \n label_y, title))\nbutton_refresh.pack()\n\n# Export the file\nbutton_export = Button(right_top, text = 'Export data', \n command = lambda: export_data(curve_x, curve_y,\n filepath, temp_file_name))\nbutton_export.pack()\n\n# Fit the curve\nbutton_fit = Button(right_top, text = 'Fitting', command = fitting_curve)\nbutton_fit.pack()\n\n# Clear the data\nbutton_clear = Button(right_top, text = 'Clear', \n command = clear_text)\nbutton_clear.pack()\n\n\n# The text widget to show the information \ninfo = Text(right_down, width = 50) \ninfo.insert(END,'Information in the running: \\n') \ninfo.pack()\n\nroot.mainloop()\n\n# matplotlib.pyplot.plot(signal_x,signal_y)\n\n\n\n\n\n\n\n","sub_path":"force_analysis_gui_tk.py","file_name":"force_analysis_gui_tk.py","file_ext":"py","file_size_in_byte":14524,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"402661142","text":"from datetime import datetime\nfrom airflow.models import DAG\nfrom airflow.operators.subdag_operator import SubDagOperator\nfrom airflow.operators.dummy_operator import DummyOperator\nfrom data_pipelines.subdag_factory import subdag_factory\n\n\n\"\"\"\nIf set LocalExecutor in subdags, they can run in parallel\nis in Sequential Executor by default\nOn Local, it could bring unwanted deadlocks\n\"\"\"\n\nPARENT_DAG_NAME = 'subdag_dag'\nSUBDAG_DAG_NAME = 'subdag'\n\nwith DAG(\n dag_id=PARENT_DAG_NAME,\n schedule_interval='@daily',\n start_date=datetime(2020, 1, 1, 10, 00, 00),\n catchup=False\n) as dag:\n start_task = DummyOperator(task_id='start')\n subdag_task = SubDagOperator(\n subdag=subdag_factory(PARENT_DAG_NAME, SUBDAG_DAG_NAME, dag.start_date, dag.schedule_interval),\n task_id=SUBDAG_DAG_NAME\n #, executor = 'LocalExecutor'\n )\n end_task = DummyOperator(task_id='end')\n start_task >> subdag_task >> end_task\n","sub_path":"dags/subdag_dag.py","file_name":"subdag_dag.py","file_ext":"py","file_size_in_byte":958,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"155145458","text":"#!/bin/env python\r\n# -*- coding: utf-8 -*-\r\n# -*- mode: python -*-\r\n# Time-stamp: <2012-06-14 10:37:01 Tanis Zhang>\r\n\r\n\r\nimport re\r\n\r\ndef calc_time(sec,msec):\r\n res = (sec*1000000 + msec)/1000000.0\r\n val = \"%.1f\" % res\r\n return val\r\n\r\nclass Event3:\r\n \"\"\"Other key\"\"\"\r\n def __init__(self):\r\n self.key_val = -1\r\n self.pressed = -1\r\n\r\n def feed(self, a1,a2,a3):\r\n \"\"\"Affect event, return 1\"\"\"\r\n ret = 0\r\n if a1==0 and a2==0 and a3==0:\r\n print(\"Key: %d, press %d\" % (self.key_val, self.pressed))\r\n self.key_val = -1\r\n self.pressed = -1\r\n ret = 0\r\n #print(\"Key: sync signal\")\r\n elif a1==1:\r\n if a3==1:\r\n self.key_val = a2\r\n self.pressed = 1\r\n ret = 1\r\n elif a3==0:\r\n self.key_val = a2\r\n self.pressed = 0\r\n ret = 1\r\n return ret\r\n \r\n\r\nclass Event4:\r\n \"\"\"Touch screen\"\"\"\r\n def __init__(self):\r\n self.x = -1\r\n self.y = -1\r\n self.pressed = 0\r\n\r\n def feed(self, a1,a2,a3):\r\n \"\"\"Affect event, return 1\"\"\"\r\n ret = 0\r\n if a1==0 and a2==0 and a3==0:\r\n #print(\"Touch point : %d %d\" % (self.x, self.y))\r\n #print(\"Touch once, skip\")\r\n pass # one touch sync signal\r\n elif a1==0 and a2 == 2 and a3==0:\r\n if self.pressed == 1:\r\n print(\"Touch finger,down: %d %d\" % (self.x, self.y))\r\n else:\r\n print(\"Touch finger,up: %d %d\" % (self.x, self.y))\r\n ret = 1# on finger sync signal\r\n elif a1==3 and a2==0x35:\r\n self.x = a3\r\n #print(\"Touch x: %d\" % a3)\r\n elif a1==3 and a2==0x36:\r\n self.y = a3\r\n #print(\"Touch y: %d\" % a3)\r\n elif a1==3 and a2==0x30:\r\n if a3==1:\r\n self.pressed = a3\r\n #print(\"Touch down: %d %d\" % (self.x, self.y))\r\n elif a3==0:\r\n self.pressed = a3\r\n #print(\"Touch up: %d %d\" %(self.x, self.y))\r\n else:\r\n print(\"Touch unkonwn: 3 0030 %x\" % a3)\r\n else:\r\n print(\"Touch: Unknow: %x %x %x\" % (a1,a2,a3))\r\n return ret\r\n\r\nclass Event5:\r\n \"\"\"Power key\"\"\"\r\n def __init__(self):\r\n pass\r\n\r\n def feed(self,a1,a2,a3):\r\n \"\"\"Affect event, return 1\"\"\"\r\n ret = 0\r\n if a1==0 and a2==0 and a3==0:\r\n return 0 # sync signal\r\n if (a1==1 and a3 == 1):\r\n print(\"Power key down: %d\" % (a2))\r\n ret = 1\r\n elif (a1==1 and a3==0):\r\n print(\"Power key up: %d\" % (a2))\r\n ret = 1\r\n else:\r\n print(\"Power key unknown: %d %d %d\" % (a1,a2,a3))\r\n return ret\r\n\r\n\r\nclass EventMgr:\r\n \"\"\"Manager event, for time\"\"\"\r\n def __init__(self):\r\n self.last_mtime=0\r\n self.evt3 = Event3()\r\n self.evt4 = Event4()\r\n self.evt5 = Event5()\r\n pass\r\n\r\n def dispatch(self,evt_idx,param1,param2,param3,sec,msec):\r\n is_valid = 0\r\n if evt_idx == 5:\r\n is_valid = self.evt5.feed(param1,param2,param3)\r\n elif evt_idx==4:\r\n is_valid = self.evt4.feed(param1,param2,param3)\r\n elif evt_idx==3:\r\n is_valid = self.evt3.feed(param1,param2,param3)\r\n else:\r\n print(\"evt %d: %d %d %d\" % (evt_idx,param1,param2,param3) )\r\n\r\n if is_valid == 1:\r\n if self.last_mtime == 0:\r\n self.last_mtime = sec*10**6 + msec\r\n else:\r\n mtime = sec*10**6 + msec\r\n mdelta = mtime - self.last_mtime\r\n if mdelta >= 10**5:\r\n delta = mdelta /1000000.0\r\n print(\"sleep %.1f\" % delta)\r\n self.last_mtime = mtime\r\n \r\n \r\ndef parse(io_file):\r\n tag_re = \\\r\n re.compile(\"^(\\d+)\\-(\\d+):\\s*\\/dev\\/input\\/event([345]):\\s*([\\da-fA-F]+)\\s+([\\da-fA-F]+)\\s+([\\da-fA-F]+)\\s*$\")\r\n io_file.seek(0)\r\n line = io_file.readline()\r\n sec_last = 0\r\n msec_last = 0\r\n evt3 = Event3()\r\n evt4 = Event4()\r\n evt5 = Event5()\r\n evt_mgr = EventMgr()\r\n while line:\r\n matchs = tag_re.match(line)\r\n if matchs :\r\n sec = int(matchs.group(1))\r\n msec = int(matchs.group(2))\r\n evt_idx = int(matchs.group(3),16)\r\n param1 = int(matchs.group(4),16)\r\n param2 = int(matchs.group(5),16)\r\n param3 = int(matchs.group(6),16)\r\n #print( \"/dev/input/event%d %d %d %d\" % (evt_idx,param1,param2,param3) )\r\n evt_mgr.dispatch(evt_idx, param1,param2,param3,sec,msec)\r\n else:\r\n pass\r\n line = io_file.readline()\r\n\r\n\r\ndef test(fname):\r\n fh = open(fname)\r\n parse(fh)\r\n fh.close()\r\n pass\r\n\r\n\r\n\r\nif \"__main__\" == __name__:\r\n import io\r\n test_input_string = \"\"\"180593-283470: /dev/input/event4: 0003 0035 0000008d\r\n180593-283775: /dev/input/event4: 0003 0036 0000015e\r\n180593-283897: /dev/input/event4: 0003 0030 00000001\r\n180593-284019: /dev/input/event4: 0000 0002 00000000\r\n180634-644092: /dev/input/event4: 0000 0000 00000000\r\n\"\"\"\r\n #test_input = io.StringIO(test_input_string)\r\n #parse(test_input)\r\n test(\"output.txt\")\r\n","sub_path":"AndroidTest/parse_input.py","file_name":"parse_input.py","file_ext":"py","file_size_in_byte":5343,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"151521956","text":"import os\nimport argparse\n\nimport torch\nfrom torch.utils.data import DataLoader\n\nfrom dataset import ImageDataset, transf, noise_transf\nfrom model import CNN\n\ndevice = 'cuda' if torch.cuda.is_available() else 'cpu'\n\n# 결과 저장 폴더 지정\ndef init( face_path, no_face_path ):\n if not os.path.isdir(face_path):\n os.mkdir(face_path)\n if not os.path.isdir(no_face_path):\n os.mkdir(no_face_path)\n\ndef cp_files(filenames, path):\n for f in filenames:\n save_path = os.path.join(path, f.split('\\\\')[-1])\n os.system(f\"cp {f} {save_path}\")\n\n\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser(description='Process some integers.')\n\n parser.add_argument('--data_dir', type=str, default=\"data\", help='Output of images classified as faces')\n parser.add_argument('--model_path', type=str, default=\"checkpoint/model.pt\", help='Output of images classified as faces')\n parser.add_argument('--face_dir', type=str, default=\"face\", help='Output of images classified as faces')\n parser.add_argument('--no_face_dir', type=str, default=\"no-face\", help='Output of images classified as no-faces')\n parser.add_argument('--limit', type=float, default=0.9, help='Output of images classified as faces')\n\n args = parser.parse_args()\n init(args.face_dir, args.no_face_dir)\n\n # 데이터 불러오기\n dataset = ImageDataset(args.data_dir,\n transform=transf,\n noise_transform=noise_transf,\n is_predict=True,\n )\n data_loader = DataLoader(dataset, batch_size=100)\n\n # 모델 불러오기\n model = CNN().to(device)\n model.load_state_dict(torch.load(args.model_path))\n model.eval()\n\n\n # 학습을 진행하지 않을 것이므로 torch.no_grad()\n\n with torch.no_grad():\n for x, y in data_loader:\n faces = []\n no_faces= []\n x = x.to(device)\n prediction = model(x)\n for idx, i in enumerate(prediction):\n if i > args.limit:\n faces.append(y[idx])\n else:\n no_faces.append(y[idx])\n cp_files(faces, args.face_dir)\n cp_files(no_faces, args.no_face_dir)\n ","sub_path":"classification_face.py","file_name":"classification_face.py","file_ext":"py","file_size_in_byte":2262,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"610290802","text":"# coding=utf8\n\n### Python支持多种图形界面的第三方库\n### TK, wxWidgets, QT, GTK\n### Python自带的库是支持Tk的Tkinter, 无需安装任何包,就可以直接使用;\n\n### Tkinter\n### Python代码会调用内置的Tkinter,Tkinter封装了访问Tk的接口\n### Tk是一个图形库,支持多个操作系统,使用Tcl语言开发;\n### Tk会调用操作系统提供的���地GUI接口,完成最终的GUI。\n### 所以,我们的代码只需要调用Tkinter提供的接口就可以了\n### 在GUI中,每个Button、Label、输入框等,都是一个Widget\n### \n\nfrom Tkinter import *\n\nclass Application(Frame):\n def __init__(self, master=None):\n Frame.__init__(self, master)\n self.pack()\n self.createWidgets()\n\n def createWidgets(self):\n self.helloLabel = Label(self, text='Hello, world!')\n self.helloLabel.pack()\n self.quitButton = Button(self, text='Quit', command=self.quit)\n self.quitButton.pack()\n\n\nimport tkMessageBox\n\nclass Application2(Frame):\n def __init__(self, master=None):\n Frame.__init__(self, master)\n self.pack()\n self.createWidgets()\n\n def createWidgets(self):\n self.nameInput = Entry(self)\n self.nameInput.pack()\n self.alertButton = Button(self, text='Hello', command=self.hello)\n self.alertButton.pack()\n\n def hello(self):\n name = self.nameInput.get() or 'world'\n tkMessageBox.showinfo('Message', 'Hello, %s' % name)\n\napp = Application2()\n# 设置窗口标题:\napp.master.title('Hello World')\n# 主消息循环:\napp.mainloop()","sub_path":"python-lxf/graph.py","file_name":"graph.py","file_ext":"py","file_size_in_byte":1597,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"437222849","text":"from django.conf.urls.defaults import *\nfrom gambino.models import Entry\n\n# define entry_info_dict used for generic view\nentry_info_dict = {\n 'queryset': Entry.live.all(),\n 'date_field': 'pub_date',\n}\n\n# Generic Views URL Patterns\nurlpatterns = patterns('django.views.generic.date_based',\n # Weblog index - Generic View\n url(r'^$', 'archive_index', entry_info_dict, 'gambino_entry_archive_index'),\n # Archive year - Generic View\n url(r'^(?P\\d{4})/$', 'archive_year', entry_info_dict, 'gambino_entry_archive_year'),\n # Archive month - Generic View\n url(r'^(?P\\d{4})/(?P\\w{3})/$', 'archive_month', entry_info_dict, 'gambino_entry_archive_month'),\n # Archive day - Generic View\n url(r'^(?P\\d{4})/(?P\\w{3})/(?P\\d{2})/$', 'archive_day', entry_info_dict, 'gambino_entry_archive_day'),\n # Weblog detail - Generic View\n url(r'^(?P\\d{4})/(?P\\w{3})/(?P\\d{2})/(?P[-\\w]+)/$', 'object_detail', entry_info_dict, 'gambino_entry_detail'),\n)\n","sub_path":"gambino/urls/entries.py","file_name":"entries.py","file_ext":"py","file_size_in_byte":1022,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"285487745","text":"import numpy as np\nimport pylab as pl\nimport scipy as sp\nimport iir_filter\nfrom scipy import signal\n#\ndata = np.loadtxt('ecg_50hz_1.dat')\ndata = data - 2048\ndata = data * 2E-3 * 500 / 1000\nfs = 1000.0\n#\npl.title('ECG')\n#\ny = data[:,1]\nt = data[:,0]\npl.subplot(311)\npl.plot(t,y)\npl.xlabel('time/sec')\npl.ylabel('ECG/mV')\n#\nf0 = 48.0\nf1 = 52.0\nsos = signal.butter(2, [f0/fs*2,f1/fs*2], 'bandstop', output='sos')\n\nf = iir_filter.IIR_filter(sos)\ny2 = np.zeros(len(y))\nfor i in range(len(y)):\n y2[i] = f.filter(y[i])\n\nyf = sp.fft(y2)\nyf[0] = 0\n\npl.subplot(312);\npl.plot(sp.linspace(0,fs,len(yf)),abs(yf))\npl.xlabel('time/sec')\npl.ylabel('Spectrum')\npl.xlim(0,fs/2)\n#\n#\npl.subplot(313)\npl.plot(t,y2)\npl.xlabel('time/sec')\npl.ylabel('ECG/mV')\n\npl.show()\n","sub_path":"example.py","file_name":"example.py","file_ext":"py","file_size_in_byte":750,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"562939701","text":"from tkinter import *\nfrom gpiozero import MCP3008\n\npot = MCP3008(channel=0)\n\nclass App(Tk):\n \n def on_exit(self):\n self.destroy()\n\n def counting(self, counter):\n if (self.textonbtn == \"STOP\"):\n self.counter = format(pot.value, '.2f')\n self.lbl.config(text = self.counter)\n self.after(1000, self.counting, self.counter)\n\n def btnclickfunc(self):\n if (self.textonbtn == \"READ\"):\n self.textonbtn = \"STOP\"\n self.btn.config(text = self.textonbtn)\n self.after(1000, self.counting, self.counter)\n \n else:\n self.textonbtn = \"READ\"\n self.btn.config(text = self.textonbtn)\n self.lbl.config(text = self.counter)\n \n \n def __init__(self):\n Tk.__init__(self)\n self.title(\"New Window\")\n self.geometry(\"450x200\")\n self.protocol(\"WM_DELETE_WINDOW\", self.on_exit)\n \n #variables\n self.textonbtn = \"READ\"\n self.counter = format(pot.value, '.2f')\n\n #arrangement of label & button\n self.intro = Label(self, text = \"Toggle the button to read analog value\")\n self.intro.grid(padx = 100, pady = 0)\n self.lbl = Label (self, text = self.counter )\n self.lbl.grid(padx = 150 , pady = 40)\n self.btn = Button(self, text=self.textonbtn, command = self.btnclickfunc)\n self.btn.grid(padx=150, pady=10)\n\n\nif __name__ == '__main__':\n App().mainloop()\n","sub_path":"reading_analog_from_GUI.py","file_name":"reading_analog_from_GUI.py","file_ext":"py","file_size_in_byte":1491,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"271590513","text":"\"\"\"\nЗадание 3.\nОпределить количество различных подстрок с использованием хеш-функции.\nДана строка S длиной N, состоящая только из строчных латинских букв.\n\nПодсказка: примените хеши и множества\n\nрара:\n\nрар\nра\nар\nара\nр\nа\n\"\"\"\n\nfrom hashlib import sha256\n\n\ndef input_users():\n iu = input('Введите строку: ')\n return count_str(iu)\n\n\ndef count_str(iu):\n res = []\n for i in range(1, len(iu) + 1):\n for n in range(0, i):\n el = sha256(iu[n:i].encode()).hexdigest()\n res.append(el)\n sub_set = set(res)\n print(f'Число подстрок {iu} - {len(sub_set) - 1}')\n return input_users()\n\n\ninput_users()","sub_path":"Урок 3. Практическое задание/task_3.py","file_name":"task_3.py","file_ext":"py","file_size_in_byte":829,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"271587394","text":"# Let's create a simple Snake Game in Python 3\n# By Florian H.\n\nimport turtle\nimport time\n\ndelay = 0.1\n\n# Set up the screen\n\nwindow = turtle.Screen()\nwindow.title(\"Snake Game By Florian H.\")\nwindow.bgcolor(\"blue\")\nwindow.setup(width=600, height=600)\nwindow.tracer(0) #turn off the screen updates\n\n\n\n#Snake Head\n\nhead = turtle.Turtle()\nhead.speed(0)\nhead.shape(\"square\")\nhead.color(\"black\")\nhead.penup() #so that it will not draw anything when the turtule is moving\nhead.goto(0,0) #when the head start we want it to be in the center of the screen so yaxe and xaxe are 0\nhead.direction = \"stop\"\n\n\n# My Functions\n\ndef go_up():\n head.direction = \"up\"\ndef go_down():\n head.direction = \"down\"\ndef go_left():\n head.direction = \"left\"\ndef go_right():\n head.direction = \"right\"\n\n#Head Directions\n\ndef move():\n if head.direction == \"up\":\n yaxe = head.ycor()\n head.sety(yaxe + 20)\n\n if head.direction == \"down\":\n yaxe = head.ycor()\n head.sety(yaxe - 20)\n\n if head.direction == \"left\":\n xaxe = head.xcor()\n head.setx(xaxe - 20)\n\n if head.direction == \"right\":\n xaxe = head.xcor()\n head.setx(xaxe + 20)\n\n#Keyboard bindings (connect Key press to a particular function)\n\nwindow.listen()\nwindow.onkeypress(go_up, \"z\")\nwindow.onkeypress(go_down, \"s\")\nwindow.onkeypress(go_left, \"q\")\nwindow.onkeypress(go_right, \"d\")\n\n\n#Main game loop\n\nwhile True:\n window.update()\n\n move()\n\n time.sleep(delay)\n\n\nwindow.mainloop()","sub_path":"Game.py","file_name":"Game.py","file_ext":"py","file_size_in_byte":1484,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"264788800","text":"from setuptools import setup, find_packages\nimport sys, os\n\nversion = '0.1'\n\nsetup(name='PyBus',\n version=version,\n description=\"PyBus is a message bus\",\n long_description=\"\"\"\\\nPyBus is a message bus that aims at providing message-level communication between items of the application (not distributed applications).\n\"\"\",\n classifiers=[], # Get strings from http://pypi.python.org/pypi?%3Aaction=list_classifiers\n keywords='message bus',\n author='Bernardo Heynemann',\n author_email='heynemann@gmail.com',\n url='http://www.pybus.org',\n license='OSI',\n packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),\n include_package_data=True,\n zip_safe=True,\n install_requires=[\n # -*- Extra requirements: -*-\n ],\n entry_points=\"\"\"\n # -*- Entry points: -*-\n \"\"\",\n )\n","sub_path":"pypi_install_script/PyBus-0.1dev-r920/setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":867,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"593172527","text":"from multiprocessing.managers import BaseManager\r\nfrom record_collector import Collector\r\nfrom functools import wraps\r\nimport random\r\nimport copy\r\nimport time\r\n\r\nBaseManager.register('Collector', Collector)\r\nmanager = BaseManager()\r\nmanager.start()\r\nrecord = manager.Collector()\r\ntimes = 25\r\n\r\n\r\ndef time_analysis(func):\r\n @wraps(func)\r\n def do_func(*args, **kwargs):\r\n print('[INFO] \\'{}\\' analysis started (N={}).'.format(func.__name__, len(args[0])))\r\n start_time = time.clock()\r\n result = func(*args, **kwargs)\r\n end_time = time.clock()\r\n total_time = end_time - start_time\r\n print('[INFO] \\'{}\\' took {} seconds (N={}).'.format(func.__name__, total_time, len(args[0])))\r\n record.new_record(func.__name__, len(result), total_time)\r\n return result\r\n\r\n return do_func\r\n\r\n\r\n@time_analysis\r\ndef bubble_sort(num_list):\r\n num_len = len(num_list)\r\n for i in range(num_len - 1):\r\n for j in range(num_len - i - 1):\r\n if num_list[j] > num_list[j + 1]:\r\n num_list[j], num_list[j + 1] = num_list[j + 1], num_list[j]\r\n return num_list\r\n\r\n\r\n@time_analysis\r\ndef selection_sort(num_list):\r\n num_len = len(num_list)\r\n for i in range(num_len - 1):\r\n min_idx = i\r\n for j in range(i + 1, num_len):\r\n if num_list[j] < num_list[min_idx]:\r\n min_idx = j\r\n num_list[i], num_list[min_idx] = num_list[min_idx], num_list[i]\r\n return num_list\r\n\r\n\r\n@time_analysis\r\ndef insertion_sort(num_list):\r\n num_len = len(num_list)\r\n for i in range(1, num_len):\r\n value = num_list[i]\r\n while i > 0 and value < num_list[i - 1]:\r\n num_list[i] = num_list[i - 1]\r\n i -= 1\r\n num_list[i] = value\r\n return num_list\r\n\r\n\r\n@time_analysis\r\ndef quick_sort(num_list):\r\n def quick(begin, end, array):\r\n if begin < end:\r\n i = begin + 1\r\n j = end\r\n while True:\r\n while i < end and array[begin] > array[i]:\r\n i += 1\r\n while j > 0 and array[begin] < array[j]:\r\n j -= 1\r\n if i < j:\r\n array[i], array[j] = array[j], array[i]\r\n else:\r\n array[begin], array[j] = array[j], array[begin]\r\n break\r\n quick(begin, j - 1, array)\r\n quick(j + 1, end, array)\r\n\r\n quick(begin=0, end=len(num_list) - 1, array=num_list)\r\n return num_list\r\n\r\n\r\n@time_analysis\r\ndef heap_sort(num_list):\r\n def construct(node, node_num, size, num_list):\r\n while node < node_num:\r\n left = (node << 1) + 1\r\n right = (node << 1) + 2\r\n if right < size:\r\n if num_list[left] > num_list[right]:\r\n MAX = left\r\n else:\r\n MAX = right\r\n else:\r\n MAX = left\r\n if num_list[node] < num_list[MAX]:\r\n num_list[node], num_list[MAX] = num_list[MAX], num_list[node]\r\n node = MAX\r\n else:\r\n break\r\n\r\n size = len(num_list)\r\n node_num = size >> 1\r\n for i in range(node_num - 1, -1, -1):\r\n construct(i, node_num, size, num_list)\r\n while size > 1:\r\n num_list[0], num_list[size - 1] = num_list[size - 1], num_list[0]\r\n size -= 1\r\n node_num = size >> 1\r\n construct(0, node_num, size, num_list)\r\n return num_list\r\n\r\n\r\nif __name__ == '__main__':\r\n # Python 2\r\n # N_list = [int(n) for n in input('Number of Elements (Separate with comma): ')]\r\n # Python 3\r\n # N_list = [n for n in input('Number of Elements (Separate with blank): ').replace(' ', '').split(',')]\r\n N_list = [50000, 100000, 150000, 200000, 250000, 300000]\r\n N_list.sort(reverse=True)\r\n for each_N in N_list:\r\n test_data = list(range(each_N))\r\n for idx in range(times):\r\n random.shuffle(test_data)\r\n bubble_sort(copy.deepcopy(test_data))\r\n selection_sort(copy.deepcopy(test_data))\r\n insertion_sort(copy.deepcopy(test_data))\r\n quick_sort(copy.deepcopy(test_data))\r\n heap_sort(copy.deepcopy(test_data))\r\n record.output_report()\r\n","sub_path":"analysis.py","file_name":"analysis.py","file_ext":"py","file_size_in_byte":4265,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"368916890","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('post', '0013_post_tags'),\n ]\n\n operations = [\n migrations.AlterField(\n model_name='post',\n name='tags',\n field=models.CharField(default=b'none', max_length=5, choices=[(b'INT', b'Interesting'), (b'INS', b'Inspiring'), (b'INF', b'Informative')]),\n ),\n ]\n","sub_path":"post/migrations/0014_auto_20160419_1206.py","file_name":"0014_auto_20160419_1206.py","file_ext":"py","file_size_in_byte":488,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"526969076","text":"import pickle, socket\r\nimport tkinter as tk\r\nfrom random import randint\r\nfrom CharClasses import *\r\nfrom BattleClasses import *\r\nfrom TkCustomClasses import *\r\nfrom getpass import getuser\r\nfrom tkinter import messagebox\r\n\r\nclass RPGMain(tk.Tk):\r\n def __init__(self, *args, **kwargs):\r\n super().__init__(*args, **kwargs)\r\n self.font = (\"Times\", 24, \"bold\")\r\n self.wm_geometry(\"900x700\")\r\n self.minsize(\"900\", \"700\")\r\n self.maxsize(\"900\", \"700\")\r\n self.title(\"RPG GAME\")\r\n container = tk.Frame(self)\r\n container.grid_rowconfigure(0, weight=1)\r\n container.grid_columnconfigure(0, weight=1)\r\n self.toolbar = ToolBar(self)\r\n container.pack(side=\"bottom\", fill=\"both\", expand=True)\r\n self.toolbar.pack(side=\"top\", fill=\"x\")\r\n\r\n self.bank = 100\r\n self.team_list = []\r\n\r\n self.frames = {}\r\n\r\n for F in (AddPage, DeletePage, GroupPage, MultiplayerPage):\r\n page_name = F.__name__\r\n frame = F(parent=container, controller=self)\r\n self.frames[page_name] = frame\r\n frame.grid(row=0, column=0, sticky=\"nsew\")\r\n self.show_frame(\"GroupPage\")\r\n\r\n def show_frame(self, page_name):\r\n frame = self.frames[page_name]\r\n frame.event_generate(\"<>\")\r\n frame.event_generate(\"<>\")\r\n frame.tkraise()\r\n\r\n def upgrade_team(self, enemy_total_level):\r\n lvl_up = enemy_total_level-(sum((ally.lvl for ally in self.team_list))//len(self.team_list))\r\n lvl_up = 1 if lvl_up < 1 else lvl_up\r\n self.bank += randint(10, 100)\r\n for ally in self.team_list:\r\n ally.health += 5*lvl_up\r\n ally.mana += 5*lvl_up\r\n ally.lvl += lvl_up\r\n\r\n def upgrade_team_loss(self):\r\n lvl_up = 1\r\n self.bank += randint(0, 10)\r\n for ally in self.team_list:\r\n ally.health += 5\r\n ally.mana += 5\r\n ally.lvl += lvl_up\r\n \r\n def add_character(self, char):\r\n if len(self.team_list)>2:\r\n messagebox.showwarning(title=\"No More!\", message=\"You cannot have more than 3 characters in your team at once\")\r\n return\r\n elif image_chars[char].amount > self.bank:\r\n messagebox.showwarning(title=\"Not Enough!\", message=\"You do not have the funds to purchase this character.\")\r\n return\r\n obj=chars[char]()\r\n self.bank -= obj.amount\r\n message = \"{} named {} has successfully been bought!\".format(obj.type, obj.name)\r\n self.team_list += [obj]\r\n messagebox.showinfo(title=\"Character Bought\", message=message)\r\n\r\n def delete_character(self, parent, char):\r\n selling_price = (char.amount*char.lvl)//2\r\n message=\"{} named {} has successfully been sold for £{}\".format(char.type, char.name, selling_price)\r\n self.bank += selling_price\r\n self.team_list.remove(char)\r\n messagebox.showinfo(title=\"Character Sold\", message=message)\r\n parent.destroy()\r\n\r\n def save_all(self):\r\n with open(\"save_file.Prgp\", \"wb\") as file:\r\n data = {\"team_list\": self.team_list, \"gold\": self.bank}\r\n pickle.dump(data, file)\r\n messagebox.showinfo(title=\"Data Saved\", message=\"Successfully saved your data\")\r\n\r\n def load_all(self):\r\n try:\r\n with open(\"save_file.Prgp\", \"rb\") as file:\r\n data = pickle.load(file)\r\n self.team_list = data[\"team_list\"]\r\n self.bank = data[\"gold\"]\r\n except FileNotFoundError:\r\n messagebox.showerror(title=\"Error!\", message=\"Sorry it seems there is no save file to load\")\r\n return\r\n self.frames[\"DeletePage\"].char_populate()\r\n self.frames[\"GroupPage\"].display_bank()\r\n messagebox.showinfo(title=\"Data Loaded\", message=\"Succesfully loaded your save file\")\r\n \r\nclass GroupPage(tk.Frame):\r\n def __init__(self, parent, controller):\r\n super().__init__(parent)\r\n self.controller = controller\r\n self.config(bg=\"lightgoldenrod1\")\r\n self.bind(\"<>\", self.display_bank)\r\n\r\n self.title=tk.Label(self, text=\"\", font=self.controller.font, bg=\"pale green\",\r\n borderwidth=2, relief=\"groove\")\r\n self.display_bank()\r\n self.title.pack(side=\"top\", pady=10, padx=10)\r\n self.outer_frame=tk.Frame(self, bg=\"pale green\", borderwidth=2, relief=\"groove\")\r\n self.outer_frame.pack(fill=\"both\", expand=True, pady=10, padx=10)\r\n\r\n self.top_frame = tk.Frame(self.outer_frame, bg=\"pale green\")\r\n self.top_frame.pack(side=\"top\", fill=\"both\", expand=True)\r\n\r\n self.add_button = HomeButton(self.top_frame, self.controller, \"Buy New Character\", \"AddPage\")\r\n self.add_button.pack(side=\"left\", padx=5, pady=5, fill=\"both\", expand=True)\r\n self.delete_button = HomeButton(self.top_frame, self.controller, \"Sell Character\", \"DeletePage\")\r\n self.delete_button.pack(side=\"left\", padx=5, pady=5, fill=\"both\", expand=True)\r\n self.multi_button = HomeButton(self.outer_frame, self.controller, \"Play Multiplayer\", \"MultiplayerPage\")\r\n self.multi_button.pack(side=\"bottom\", padx=5, pady=5, fill=\"both\", expand=True)\r\n \r\n def display_bank(self, event=None):\r\n self.title.config(text=\"Welcome {}, Money: £{}\".format(getuser(), self.controller.bank))\r\n \r\nclass AddPage(tk.Frame):\r\n def __init__(self, parent, controller):\r\n super().__init__(parent)\r\n self.controller = controller\r\n self.counter=0\r\n self.config(bg=\"lightgoldenrod1\")\r\n\r\n self.title = tk.Button(self, text=\"Buy a New Character\", bg=\"pale green\", font=self.controller.font,\r\n borderwidth=2, relief=\"groove\", command=lambda: self.controller.add_character(self.counter))\r\n self.title.pack(side=\"top\", pady=5, padx=5, expand=True)\r\n\r\n self.left_frame = tk.Frame(self, bg=\"pale green\", borderwidth=2, relief=\"groove\")\r\n self.right_frame = tk.Frame(self, bg=\"pale green\", borderwidth=2, relief=\"groove\")\r\n self.bottom_frame=tk.Frame(self, bg=\"pale green\", borderwidth=2, relief=\"groove\")\r\n\r\n for x in range(7):\r\n tk.Label(self.right_frame, text=\"TEST\",\r\n font=(\"Times\", 12, \"bold\"), bg=\"pale green\").pack(anchor=\"w\", pady=5, padx=5)\r\n\r\n self.char_image = tk.PhotoImage(file=\"\")\r\n self.char_image_label = tk.Label(self.left_frame, image=self.char_image)\r\n self.char_image_label.image = self.char_image\r\n self.char_image_label.pack(fill=\"both\", expand=True)\r\n\r\n self.bottom_frame.pack(side=\"bottom\", fill=\"x\", expand=True, pady=5, padx=10)\r\n self.left_frame.pack(side=\"left\", fill=\"x\", pady=5, padx=10)\r\n self.right_frame.pack(side=\"right\", fill=\"both\", pady=5, padx=10, expand=True)\r\n\r\n self.right_arrow = tk.Button(self.bottom_frame, width=10, text=\"→\", borderwidth=0,\r\n font=self.controller.font, bg=\"pale green\", command=lambda: self.right_shift())\r\n self.left_arrow = tk.Button(self.bottom_frame, width=10, text=\"←\", borderwidth=0,\r\n font=self.controller.font, bg=\"pale green\", command=lambda: self.left_shift())\r\n self.name_plate = tk.Label(self.bottom_frame, text=\"\", font=self.controller.font, bg=\"pale green\")\r\n\r\n self.right_arrow.pack(side=\"right\", pady=5, padx=5)\r\n self.left_arrow.pack(side=\"left\", pady=5, padx=5)\r\n self.name_plate.pack(pady=5, padx=5)\r\n self.stats_populate()\r\n\r\n def stats_populate(self):\r\n current_char=image_chars[self.counter]\r\n self.char_image.config(file=\"imgres/\"+current_char.type+\".png\")\r\n children=self.right_frame.winfo_children()\r\n\r\n self.name_plate.config(text=current_char.type)\r\n self.title.config(text=\"Buy {}, Amount: £{}\".format(current_char.type, current_char.amount))\r\n\r\n for data, child in zip(current_char, children):\r\n child.config(text=\"{}{}\".format(data[0],data[1]))\r\n\r\n def right_shift(self):\r\n if self.counter==len(image_chars)-1:\r\n self.counter=0\r\n else:\r\n self.counter+=1\r\n self.stats_populate()\r\n\r\n def left_shift(self):\r\n if self.counter==0:\r\n self.counter=len(image_chars)-1\r\n else:\r\n self.counter-=1\r\n self.stats_populate()\r\n\r\nclass DeletePage(tk.Frame):\r\n def __init__(self, parent, controller):\r\n super().__init__(parent)\r\n self.controller = controller\r\n self.config(bg=\"lightgoldenrod1\")\r\n self.bind(\"<>\", self.char_populate)\r\n\r\n self.list_frame = FrameScrollBar(self, \"top\", \"right\")\r\n self.list_frame.locate()\r\n self.char_populate()\r\n\r\n def char_populate(self, event=None):\r\n children=self.list_frame.inner_frame.winfo_children()\r\n team_list=self.controller.team_list\r\n\r\n for child in children:\r\n child.destroy()\r\n \r\n for char in team_list:\r\n temp = DeleteTab(self.list_frame.inner_frame, self.controller, char)\r\n temp.pack(side=\"top\", fill=\"x\", expand=True)\r\n\r\nclass MultiplayerPage(tk.Frame):\r\n def __init__(self, parent, controller):\r\n super().__init__(parent)\r\n self.controller = controller\r\n self.config(bg=\"lightgoldenrod1\")\r\n self.bind(\"<>\", self.update_label)\r\n\r\n self.title_frame = tk.Frame(self, bg=\"pale green\", borderwidth=2, relief=\"groove\")\r\n self.title_frame.pack(side=\"top\", padx=10, pady=10, fill=\"both\")\r\n\r\n self.name_label = tk.Label(self.title_frame, text=\"\", font=self.controller.font, bg=\"pale green\")\r\n self.name_label.pack(side=\"left\", padx=5, pady=5)\r\n self.outer_frame = tk.Frame(self, bg=\"pale green\", borderwidth=2, relief=\"groove\")\r\n self.button_frame = tk.Frame(self.outer_frame, bg=\"pale green\")\r\n\r\n self.outer_frame.pack(side=\"bottom\", padx=10, pady=10, fill=\"both\", expand=True)\r\n self.button_frame.pack(side=\"top\", padx=10, pady=10)\r\n\r\n self.join_title = tk.Button(self.button_frame, text=\"Join Game\", bg=\"pale green\", borderwidth=2,\r\n relief=\"groove\", font=self.controller.font,\r\n command=lambda: self.create_game(\"Join\"))\r\n self.join_title.pack(side=\"left\", padx=10, pady=10) \r\n self.create_title = tk.Button(self.button_frame, text=\"Create Game\", bg=\"pale green\", borderwidth=2,\r\n relief=\"groove\", font=self.controller.font,\r\n command=lambda: self.create_game(\"Create\"))\r\n self.create_title.pack(side=\"right\", padx=10, pady=10)\r\n\r\n self.ip_label = PropEntry(self.outer_frame, \"Server IP:\", \"192.168.56.1\", (\"Times\", 16, \"bold\"))\r\n self.port_label = PropEntry(self.outer_frame, \"Port: \", \"5000\", (\"Times\", 16, \"bold\"))\r\n self.ip_label.pack(padx=5, pady=10, fill=\"x\")\r\n self.port_label.pack(padx=5, pady=10, fill=\"x\")\r\n\r\n self.update_label()\r\n\r\n def update_label(self):\r\n ip_addr = socket.gethostbyname(socket.gethostname())\r\n self.name_label.config(text=\"Username: {}, Your IP: {}\".format(getuser(), ip_addr))\r\n\r\n def create_game(self, _type):\r\n try:\r\n IP = self.ip_label.prop_entry.get()\r\n PORT = int(self.port_label.prop_entry.get())\r\n if len(self.controller.team_list) < 1: raise ValueError\r\n except ValueError:\r\n messagebox.showerror(title=\"Bad Entry\", message=\"Please check you have entered the information correctly\\n or that you have at least one character in your team\")\r\n return\r\n\r\n if _type == \"Join\":\r\n PlayerObject = BattleClientSide(self.controller, IP, PORT)\r\n else:\r\n PlayerObject = BattleServerSide(self.controller, IP, PORT)\r\n self.controller.update_idletasks()\r\n\r\nif __name__ == \"__main__\":\r\n root = RPGMain()\r\n root.mainloop()\r\n","sub_path":"maingame.pyw","file_name":"maingame.pyw","file_ext":"pyw","file_size_in_byte":12142,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"190602674","text":"from selenium import webdriver\nfrom selenium.webdriver.common.action_chains import ActionChains\nfrom time import sleep\n\n\ndriver = webdriver.Firefox()\ndriver.implicitly_wait(10)\ndriver.maximize_window()\ndriver.get('http://sahitest.com/demo/dragDropMooTools.htm')\n\ndragger = driver.find_element_by_id('dragger') # 被拖拽元素\nitem1 = driver.find_element_by_xpath('//div[text()=\"Item 1\"]') # 目标元素1\nitem2 = driver.find_element_by_xpath('//div[text()=\"Item 2\"]') # 目标2\nitem3 = driver.find_element_by_xpath('//div[text()=\"Item 3\"]') # 目标3\nitem4 = driver.find_element_by_xpath('//div[text()=\"Item 4\"]') # 目标4\n\nActionChains(driver).drag_and_drop(dragger, item1).perform() # 1.移动dragger到目标1\nsleep(2)\nActionChains(driver).click_and_hold(dragger).release(item2).perform() # 2.效果与上句相同,也能起到移动效果\nsleep(2)\nActionChains(driver).click_and_hold(dragger).move_to_element(item3).release().perform() # 3.效果与上两句相同,也能起到移动的效果\nsleep(2)\nActionChains(driver).drag_and_drop_by_offset(dragger, 400, 150).perform() # 4.移动到指定坐标\nsleep(2)\nActionChains(driver).click_and_hold(dragger).move_by_offset(400, 150).release().perform() # 5.与上一句相同,移动到指定坐标\nsleep(2)\ndriver.quit()\n","sub_path":"代码/day5/drop_drop1.py","file_name":"drop_drop1.py","file_ext":"py","file_size_in_byte":1292,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"67827192","text":"from itertools import zip_longest\n# INPUT DEAL NAME\ndeal_name = input('Enter Deal Name:')\n\n# INPUT LOAN AMOUNT\nloan_amount = int(input('Enter 1st Loan Amount:'))\n\n# INPUT INTEREST RATE\nrate = float(input('Enter 1st Rate: '))/100\nmo_rate = rate/12\n\n# INPUT TERM IN MONTHS\nterm = int(input('Enter 1st Term Months:'))\nterm_periods = [x for x in range(1,term+1)]\n\n# INPUT AMORT IN MONTHS\namort = int(input('Enter 1st Amortization Months:'))\namort_periods = [y for y in range(1,amort+1)]\n\n# CREATE AMORTIZATION VARIABLES\ncurrent_balance = loan_amount\nmo_payment = loan_amount*(mo_rate*(1+mo_rate)**amort/(((1+mo_rate)**amort)-1))\n\n# MAKE AMORTIZATION TABLE\namort_table = [('Period','Current Balance','Monthly Payment','Principal','Interest','Remaining Balance')]\nfor i in term_periods:\n beg_balance = current_balance\n int_paid = current_balance*mo_rate\n principal = mo_payment-int_paid\n current_balance = current_balance - principal\n amort_table.append((i, int(beg_balance), int(mo_payment), int(principal),int(int_paid), int(current_balance)))\n\n# -------------------------------------------------------------------------------------\n# 2ND LOAN\n\n# INPUT LOAN AMOUNT\nloan_amount2 = int(input('Enter 2nd Loan Amount:'))\n\n# INPUT INTEREST RATE\nrate2 = float(input('Enter 2nd Rate: '))/100\nmo_rate2 = rate2/12\n\n# INPUT TERM IN MONTHS\nterm2 = int(input('Enter 2nd Term Months:'))\nterm_periods2 = [x for x in range(1,term2+1)]\n\n# INPUT AMORT IN MONTHS\namort2 = int(input('Enter 2nd Amortization Months:'))\namort_periods2 = [y for y in range(1,amort2+1)]\n\n# CREATE AMORTIZATION VARIABLES\ncurrent_balance2 = loan_amount2\nmo_payment2 = loan_amount2*(mo_rate2*(1+mo_rate2)**amort2/(((1+mo_rate2)**amort2)-1))\n\n# MAKE AMORTIZATION TABLE\namort_table2 = [('Period','Current Balance','Monthly Payment','Principal','Interest','Remaining Balance')]\nfor i in term_periods2:\n beg_balance2 = current_balance2\n int_paid2 = current_balance2*mo_rate2\n principal2 = mo_payment2-int_paid2\n current_balance2 = current_balance2 - principal2\n amort_table2.append((i, int(beg_balance2), int(mo_payment2), int(principal2),int(int_paid2), int(current_balance2)))\n\n# -------------------------------------------------------------------------------------\n# COMBINE TWO LOANS INTO AMORTIZATION TABLE (NEED ITERTOOLS(ZIP_LONGEST) TO GO LONGER THAN SHORTEST LOAN)\ntotal_financing = loan_amount + loan_amount2\ntotal_mo_payment = mo_payment + mo_payment2\namort_table3 = [('Period','Current Balance','Monthly Payment','Principal','Interest', 'Remaining Balance', 'Blended Rate')]\n# Used to avoid NoneType is not subscriptable error resulting from zip_longest\nzero_list = [0,0,0,0,0,0,0]\nfor x,y in zip_longest(amort_table[1:],amort_table2[1:]):\n if x == None:\n x = zero_list\n if y == None:\n y = zero_list\n beg_balance3 = x[1] + y[1]\n mo_payment3 = x[2] + y[2]\n principal3 = x[3] + y[3]\n int_paid3 = x[4] + y[4]\n current_balance3 = x[5] + y[5]\n blended_rate = ((rate*(x[1]/beg_balance3)) + (rate2*(y[1]/beg_balance3)))*100\n amort_table3.append((y[0], '{:,}'.format(beg_balance3), '{:,}'.format(mo_payment3), '{:,}'.format(principal3), '{:,}'.format(int_paid3), '{:,}'.format(current_balance3), '{:.2f}%'.format(blended_rate)))\n\nblended_rate = amort_table3[1][6]\n\n# -------------------------------------------------------------------------------------\n# ORGANIZE LOAN SUMMARY INTO LIST OF LISTS AND TWO SEPARATE LISTS FOR FIELDS AND VALUES\nloan_summary = [['Deal Name','{}'.format(deal_name)], ['1st Loan Amount','${:,}'.format(loan_amount)], ['1st Interest Rate','{:.2f}%'.format(rate*100)], ['1st Monthly Payment','${:,.0f}'.format(mo_payment)],\n ['1st Loan Term','{} months'.format(term)], ['1st Loan Amortization','{} months'.format(amort)], ['',''], ['2nd Loan Amount','${:,}'.format(loan_amount2)], ['2nd Interest Rate','{:.2f}%'.format(rate2*100)],\n ['2nd Monthly Payment','${:,.0f}'.format(mo_payment2)], ['2nd Loan Term','{} months'.format(term2)], ['2nd Loan Amortization','{} months'.format(amort2)], ['',''], ['Total Financing','${:,}'.format(total_financing)],\n ['Blended Rate','{}'.format(blended_rate)], ['Total Monthly Payment','${:,.0f}'.format(total_mo_payment)]]\nloan_summary_fields = [_[0] for _ in loan_summary]\nloan_summary_values = [_[1] for _ in loan_summary]\n\n# PRINT OUT LOAN SUMMARY\n# print('1st Loan Amount: ${:,}'.format(loan_amount))\n# print('1st Interest Rate: {:.2f}%'.format(rate*100))\n# print('1st Monthly Payment: ${:,.0f}'.format(mo_payment))\n# print('1st Loan Term: {} months'.format(term))\n# print('1st Loan Amortization: {} months'.format(amort))\n\n# print('')\n\n# print('2nd Loan Amount: ${:,}'.format(loan_amount2))\n# print('2nd Interest Rate: {:.2f}%'.format(rate2*100))\n# print('2nd Monthly Payment: ${:,.0f}'.format(mo_payment2))\n# print('2nd Loan Term: {} months'.format(term2))\n# print('2nd Loan Amortization: {} months'.format(amort2))\n\nprint('')\n\nprint('Total Financing: ${:,}'.format(total_financing))\nprint('Blended Rate: {}'.format(blended_rate))\nprint('Total Monthly Payment: ${:,.0f}'.format(total_mo_payment))\n\nprint('')\n\n# for x in amort_table: print(x)\n\n# for x in amort_table2: print(x)\n\n# for x in amort_table3: print(x)\n#----------------------------------------------------------------------------------------------------------------------------------\n# PRETTY TABLE\nfrom prettytable import PrettyTable\n\nnew_table = PrettyTable(amort_table3[0])\n\nfor x in range(1,len(amort_table3)):\n new_table.add_row(amort_table3[x])\n\nprint(new_table)\n\n\n#CREATE EXCEL FILE OF AMORTIZATION TABLE\nimport pandas as pd\n\ndf = pd.DataFrame(amort_table3)\n\nwriter = pd.ExcelWriter(\"C:/Users/KevinO'Shea/OneDrive - Liberty SBF/Training/Python Practice/Amortization Table/Excel/Output/{}.xlsx\".format(deal_name), engine='xlsxwriter')\n\ndf.to_excel(writer, sheet_name='Amortization')\n\n# ADD WORKSHEET WITH THE LOAN SUMMARY PRINTED IN THE TOP LEFT AS TWO COLUMNS\nworkbook = writer.book\nworksheet_2 = workbook.add_worksheet('Deal Info')\nworksheet_2.write_column('A1',loan_summary_fields)\nworksheet_2.write_column('B1',loan_summary_values)\n\n\nwriter.save()\n\nprint('Outputted to excel file')\n\n\n# ----------------------------------------------------------------------------------------------------------------------------------------------------\n# TO DO\n# 1) Make it work when length of the 1st term is longer than the 2nd's term\n# 2) Format input (So it shows a % when you input rate or comma's when you enter the loan amounts)\n# 3) Add additional tabs of loan information\n# -------------------------------------------------------------------------------------------------------------------------------\n","sub_path":"Python_Amortization_Excel_Output.py","file_name":"Python_Amortization_Excel_Output.py","file_ext":"py","file_size_in_byte":6655,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"258490099","text":"import sys\n\nfrom taskman_client.task_proxy import get_task_manager_proxy\nfrom trainer_v2.chair_logging import c_log\nfrom trainer_v2.per_project.transparency.mmp.eval_helper.eval_line_format import predict_and_save_scores, \\\n eval_dev100_mrr, predict_and_batch_save_scores\nfrom trainer_v2.per_project.transparency.mmp.tt_model.load_tt_predictor import get_tt10_scorer\nfrom trainer_v2.train_util.arg_flags import flags_parser\nfrom trainer_v2.train_util.get_tpu_strategy import get_strategy\n\n\ndef main(args):\n model_path = args.output_dir\n run_name = args.run_name\n dataset = \"dev_sample100\"\n c_log.info(f\"{run_name}, {dataset}\")\n strategy = get_strategy(args.use_tpu, args.tpu_name)\n with strategy.scope():\n c_log.info(\"Building scorer\")\n score_fn = get_tt10_scorer(model_path)\n predict_and_batch_save_scores(score_fn, dataset, run_name, 100*100)\n score = eval_dev100_mrr(dataset, run_name)\n\n proxy = get_task_manager_proxy()\n proxy.report_number(run_name, score, dataset, \"mrr\")\n print(f\"Recip_rank:\\t{score}\")\n\n\n\nif __name__ == \"__main__\":\n args = flags_parser.parse_args(sys.argv[1:])\n main(args)\n\n","sub_path":"src/trainer_v2/per_project/transparency/mmp/tf_runner/run_tt10_rerank.py","file_name":"run_tt10_rerank.py","file_ext":"py","file_size_in_byte":1155,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"359498891","text":"# -*- coding: utf-8 -*-\r\n# 2021/1/31\r\n# create by: snower\r\n\r\nimport sys\r\nimport os\r\nimport time\r\nimport struct\r\nimport logging\r\nimport traceback\r\nimport argparse\r\nimport threading\r\nimport socket\r\nimport hashlib\r\nimport sevent\r\nfrom .utils import create_server, create_socket, config_signal, format_data_len\r\nfrom .simple_proxy import http_protocol_parse, socks5_protocol_parse\r\n\r\nconns, status = {}, {\"remote_conn\": [], \"local_conn\": []}\r\n\r\ndef warp_write(conn, status, key):\r\n origin_write = conn.write\r\n def _(data):\r\n status[key] += len(data)\r\n return origin_write(data)\r\n return _\r\n\r\ndef gen_sign_key(key):\r\n t = struct.pack(\"I\", int(time.time()))\r\n oncestr = os.urandom(16)\r\n return t + oncestr + hashlib.md5(t + oncestr + sevent.utils.ensure_bytes(key)).digest()\r\n\r\ndef check_sign_key(key, sign_key):\r\n return sign_key[20:] == hashlib.md5(sign_key[:20] + sevent.utils.ensure_bytes(key)).digest()\r\n\r\nasync def parse_forward_address(conn, proxy_type):\r\n if not proxy_type or proxy_type == \"raw\":\r\n return None\r\n\r\n if proxy_type == \"socks5\":\r\n buffer = await conn.recv()\r\n host, port, data = await socks5_protocol_parse(conn, buffer)\r\n if not host or not port:\r\n raise Exception(\"parse error\")\r\n buffer.write(data + buffer.read())\r\n return (host, port)\r\n\r\n if proxy_type == \"http\":\r\n buffer = await conn.recv()\r\n host, port, data = await http_protocol_parse(conn, buffer)\r\n if not host or not port:\r\n raise Exception(\"parse error\")\r\n buffer.write(data + buffer.read())\r\n return (host, port)\r\n\r\n if proxy_type == \"redirect\":\r\n address_data = conn.socket.getsockopt(socket.SOL_IP, 80, 16)\r\n host, port = socket.inet_ntoa(address_data[4:8]), struct.unpack(\">H\", address_data[2:4])[0]\r\n return (host, port)\r\n raise Exception(\"parse error\")\r\n\r\nasync def write_forward_address(conn, forward_address):\r\n await conn.send(b\"\".join([struct.pack(\"!H\", len(forward_address[0])),\r\n sevent.utils.ensure_bytes(forward_address[0]), struct.pack(\"!H\", forward_address[1])]))\r\n\r\nasync def read_forward_address(conn):\r\n host_len, = struct.unpack(\"!H\", (await conn.recv(2)).read(2))\r\n host = sevent.utils.ensure_unicode((await conn.recv(host_len)).read(host_len))\r\n port, = struct.unpack(\"!H\", (await conn.recv(2)).read(2))\r\n return (host, port)\r\n\r\nasync def tcp_forward(conn, forward_address, conns, status):\r\n start_time = time.time()\r\n conn.write, pconn = warp_write(conn, status, \"recv_len\"), None\r\n\r\n try:\r\n conn.enable_nodelay()\r\n pconn = create_socket(forward_address)\r\n await pconn.connectof(forward_address)\r\n pconn.write = warp_write(pconn, status, \"send_len\")\r\n conns[id(conn)] = (conn, pconn, status)\r\n logging.info(\"tcp forward connected %s:%d -> %s:%d\", conn.address[0], conn.address[1], forward_address[0], forward_address[1])\r\n await pconn.linkof(conn)\r\n except sevent.errors.SocketClosed:\r\n pass\r\n except Exception as e:\r\n logging.info(\"tcp forward error %s:%d -> %s:%d %s %.2fms\\r%s\", conn.address[0], conn.address[1],\r\n forward_address[0], forward_address[1], e, (time.time() - start_time) * 1000, traceback.format_exc())\r\n return\r\n finally:\r\n conn.close()\r\n if pconn: pconn.close()\r\n conns.pop(id(conn), None)\r\n\r\n logging.info(\"tcp forward closed %s:%d -> %s:%d %s %s %.2fms\", conn.address[0], conn.address[1],\r\n forward_address[0], forward_address[1], format_data_len(status[\"send_len\"]), format_data_len(status[\"recv_len\"]),\r\n (time.time() - start_time) * 1000)\r\n\r\nasync def reverse_port_forward(remote_conn, local_conn, status, forward_address):\r\n start_time = time.time()\r\n\r\n try:\r\n if forward_address:\r\n await remote_conn.send(b'\\x01')\r\n await write_forward_address(remote_conn, forward_address)\r\n else:\r\n await remote_conn.send(b'\\x00')\r\n local_conn.write = warp_write(local_conn, status, \"recv_len\")\r\n remote_conn.write = warp_write(remote_conn, status, \"send_len\")\r\n logging.info(\"tcp forward connected %s:%d -> %s:%d\", local_conn.address[0], local_conn.address[1],\r\n remote_conn.address[0], remote_conn.address[1])\r\n await local_conn.linkof(remote_conn)\r\n except sevent.errors.SocketClosed:\r\n pass\r\n except Exception as e:\r\n logging.info(\"tcp forward error %s:%d -> %s:%d %s %.2fms\\r%s\", local_conn.address[0], local_conn.address[1],\r\n remote_conn.address[0], remote_conn.address[1], e, (time.time() - start_time) * 1000,\r\n traceback.format_exc())\r\n return\r\n finally:\r\n remote_conn.close()\r\n local_conn.close()\r\n conns.pop(id(remote_conn), None)\r\n\r\n logging.info(\"tcp forward closed %s:%d -> %s:%d %s %s %.2fms\", local_conn.address[0], local_conn.address[1],\r\n remote_conn.address[0], remote_conn.address[1], format_data_len(status[\"send_len\"]),\r\n format_data_len(status[\"recv_len\"]),\r\n (time.time() - start_time) * 1000)\r\n\r\nasync def server_handle_remote_connect(conn, forward_address, key, proxy_type, conns, status):\r\n setattr(conn, \"_connected_time\", time.time())\r\n def on_close(conn):\r\n if conn not in status[\"remote_conn\"]:\r\n return\r\n status[\"remote_conn\"].remove(conn)\r\n logging.info(\"remote conn waited close %s:%d\", conn.address[0], conn.address[1])\r\n\r\n status[\"remote_conn\"].append(conn)\r\n conn.on_close(on_close)\r\n try:\r\n connect_type, sign_key_len = struct.unpack(\"!BB\", (await conn.recv(2)).read(2))\r\n sign_key = (await conn.recv(sign_key_len)).read(sign_key_len) if sign_key_len > 0 else b''\r\n except sevent.errors.SocketClosed:\r\n return\r\n if not check_sign_key(key, sign_key):\r\n await conn.closeof()\r\n logging.info(\"remote conn auth fail %s:%d %s\", conn.address[0], conn.address[1], sign_key)\r\n return\r\n\r\n if connect_type == 1:\r\n setattr(conn, \"_authed_time\", time.time())\r\n if status[\"local_conn\"]:\r\n forward_status = {\"recv_len\": 0, \"send_len\": 0, \"last_time\": time.time(), \"check_recv_len\": 0,\r\n \"check_send_len\": 0}\r\n local_conn = status[\"local_conn\"].pop(0)\r\n sevent.go(reverse_port_forward, conn, local_conn, forward_status, local_conn._connected_forward_address)\r\n conns[id(conn)] = (conn, local_conn, forward_status)\r\n if conn not in status[\"remote_conn\"]:\r\n return\r\n status[\"remote_conn\"].remove(conn)\r\n return\r\n logging.info(\"remote conn waiting %s:%d\", conn.address[0], conn.address[1])\r\n return\r\n\r\n if connect_type == 2 or connect_type == 3:\r\n try:\r\n if connect_type == 3:\r\n forward_address = await read_forward_address(conn)\r\n await conn.send(b'\\x00')\r\n forward_status = {\"recv_len\": 0, \"send_len\": 0, \"last_time\": time.time(), \"check_recv_len\": 0,\r\n \"check_send_len\": 0}\r\n sevent.current().call_async(tcp_forward, conn, forward_address, conns, forward_status)\r\n except sevent.errors.SocketClosed:\r\n pass\r\n except Exception as e:\r\n logging.info(\"tcp forward error %s:%d -> %s:%d %s\\r%s\", conn.address[0], conn.address[1],\r\n forward_address[0], forward_address[1], e, traceback.format_exc())\r\n return\r\n\r\n await conn.closeof()\r\n logging.info(\"remote conn unsupport connect type %s:%d %s\", conn.address[0], conn.address[1], key)\r\n\r\nasync def server_handle_local_connect(conn, forward_address, key, proxy_type, conns, status):\r\n try:\r\n forward_address = (await parse_forward_address(conn, proxy_type)) if proxy_type else forward_address\r\n except Exception as e:\r\n logging.info(\"parse proxy forward address error %s\", e)\r\n return\r\n\r\n setattr(conn, \"_connected_forward_address\", forward_address)\r\n setattr(conn, \"_connected_time\", time.time())\r\n if status[\"remote_conn\"]:\r\n forward_status = {\"recv_len\": 0, \"send_len\": 0, \"last_time\": time.time(), \"check_recv_len\": 0,\r\n \"check_send_len\": 0}\r\n remote_conn = status[\"remote_conn\"].pop(0)\r\n sevent.go(reverse_port_forward, remote_conn, conn, forward_status, forward_address)\r\n conns[id(remote_conn)] = (remote_conn, conn, forward_status)\r\n return\r\n\r\n def on_close(conn):\r\n if conn not in status[\"local_conn\"]:\r\n return\r\n status[\"local_conn\"].remove(conn)\r\n logging.info(\"local conn waited close %s:%d\", conn.address[0], conn.address[1])\r\n status[\"local_conn\"].append(conn)\r\n conn.on_close(on_close)\r\n logging.info(\"local conn waiting %s:%d\", conn.address[0], conn.address[1])\r\n\r\nasync def server_run_server(server, forward_address, key, proxy_type, conns, status, handle):\r\n while True:\r\n try:\r\n conn = await server.accept()\r\n sevent.current().call_async(handle, conn, forward_address, key, proxy_type, conns, status)\r\n except sevent.errors.SocketClosed as e:\r\n sevent.current().call_async(sevent.current().stop)\r\n raise e\r\n\r\nasync def client_run_connect(remote_address, forward_address, key, conns, status):\r\n while True:\r\n start_time = time.time()\r\n try:\r\n conn = create_socket(remote_address)\r\n await conn.connectof(remote_address)\r\n sign_key = gen_sign_key(key)\r\n await conn.send(struct.pack(\"!BB\", 1, len(sign_key)) + sign_key)\r\n connect_type = (await conn.recv(1)).read(1)\r\n if connect_type == b'\\x01':\r\n current_forward_address = await read_forward_address(conn)\r\n else:\r\n current_forward_address = forward_address\r\n forward_status = {\"recv_len\": 0, \"send_len\": 0, \"last_time\": time.time(), \"check_recv_len\": 0,\r\n \"check_send_len\": 0}\r\n sevent.current().call_async(tcp_forward, conn, current_forward_address, conns, forward_status)\r\n except sevent.errors.SocketClosed as e:\r\n logging.info(\"connect error %s:%d %s\", remote_address[0], remote_address[1], e)\r\n if time.time() - start_time < 5:\r\n await sevent.sleep(5)\r\n except (sevent.errors.ResolveError, ConnectionRefusedError) as e:\r\n logging.info(\"connect error %s:%d %s\", remote_address[0], remote_address[1], e)\r\n await sevent.sleep(5)\r\n except Exception as e:\r\n sevent.current().call_async(sevent.current().stop)\r\n raise e\r\n\r\nasync def client_handle_local_connect(conn, remote_address, key, proxy_type, conns, status):\r\n start_time = time.time()\r\n conn.write, pconn = warp_write(conn, status, \"recv_len\"), None\r\n\r\n try:\r\n forward_address = (await parse_forward_address(conn, proxy_type)) if proxy_type else None\r\n\r\n pconn = create_socket(remote_address)\r\n await pconn.connectof(remote_address)\r\n sign_key = gen_sign_key(key)\r\n if forward_address:\r\n await pconn.send(struct.pack(\"!BB\", 3, len(sign_key)) + sign_key)\r\n await write_forward_address(pconn, forward_address)\r\n else:\r\n await pconn.send(struct.pack(\"!BB\", 2, len(sign_key)) + sign_key)\r\n (await pconn.recv(1)).read(1)\r\n pconn.write = warp_write(pconn, status, \"send_len\")\r\n logging.info(\"tcp forward connected %s:%d -> %s:%d\", conn.address[0], conn.address[1], remote_address[0], remote_address[1])\r\n await pconn.linkof(conn)\r\n except sevent.errors.SocketClosed:\r\n pass\r\n except Exception as e:\r\n logging.info(\"tcp forward error %s:%d -> %s:%d %s %.2fms\\r%s\", conn.address[0], conn.address[1],\r\n remote_address[0], remote_address[1], e, (time.time() - start_time) * 1000,\r\n traceback.format_exc())\r\n return\r\n finally:\r\n conn.close()\r\n if pconn: pconn.close()\r\n conns.pop(id(conn), None)\r\n\r\n logging.info(\"tcp forward closed %s:%d -> %s:%d %s %s %.2fms\", conn.address[0], conn.address[1],\r\n remote_address[0], remote_address[1], format_data_len(status[\"send_len\"]),\r\n format_data_len(status[\"recv_len\"]),\r\n (time.time() - start_time) * 1000)\r\n\r\nasync def client_run_server(server, remote_address, key, proxy_type, conns):\r\n while True:\r\n try:\r\n conn = await server.accept()\r\n status = {\"recv_len\": 0, \"send_len\": 0, \"last_time\": time.time(), \"check_recv_len\": 0, \"check_send_len\": 0}\r\n sevent.current().call_async(client_handle_local_connect, conn, remote_address, key, proxy_type, conns, status)\r\n conns[id(conn)] = (conn, conn, status)\r\n except sevent.errors.SocketClosed as e:\r\n sevent.current().call_async(sevent.current().stop)\r\n raise e\r\n\r\nasync def check_timeout(conns, conn_status, timeout):\r\n def run_check():\r\n while True:\r\n try:\r\n now = time.time()\r\n for conn in tuple(conn_status[\"remote_conn\"]):\r\n if not hasattr(conn, \"_authed_time\"):\r\n if now - conn._connected_time >= 30:\r\n sevent.current().add_async_safe(conn.close)\r\n elif now - conn._authed_time >= 180:\r\n sevent.current().add_async_safe(conn.close)\r\n\r\n for conn in tuple(conn_status[\"local_conn\"]):\r\n if now - conn._connected_time >= 180:\r\n sevent.current().add_async_safe(conn.close)\r\n\r\n for conn_id, (conn, pconn, status) in list(conns.items()):\r\n if status['check_recv_len'] != status['recv_len'] or status['check_send_len'] != status['send_len']:\r\n status[\"check_recv_len\"] = status[\"recv_len\"]\r\n status[\"check_send_len\"] = status[\"send_len\"]\r\n status['last_time'] = now\r\n continue\r\n\r\n if now - status['last_time'] >= timeout:\r\n sevent.current().add_async_safe(conn.close)\r\n sevent.current().add_async(pconn.close)\r\n conns.pop(conn_id, None)\r\n finally:\r\n time.sleep(min(float(timeout) / 2.0, 30))\r\n\r\n if timeout > 0:\r\n check_thread = threading.Thread(target=run_check)\r\n check_thread.daemon = True\r\n check_thread.start()\r\n await sevent.Future()\r\n\r\ndef main(argv):\r\n parser = argparse.ArgumentParser(description='tcp reverse port forward')\r\n parser.add_argument('-c', dest='is_client_mode', nargs='?', const=True, default=False, type=bool,\r\n help='is client mode (defualt: False)')\r\n parser.add_argument('-k', dest='key', default='', type=str,\r\n help='auth key (defualt: \"\")')\r\n parser.add_argument('-b', dest='bind_host', default=\"0.0.0.0\",\r\n help='server and client mode local bind host (default: 0.0.0.0)')\r\n parser.add_argument('-p', dest='bind_port', default=0, type=int,\r\n help='server and client mode local bind port (default: 8089)')\r\n parser.add_argument('-r', dest='listen_host', default=\"0.0.0.0\",\r\n help='server mode reverse server listen host (default: 0.0.0.0)')\r\n parser.add_argument('-l', dest='listen_port', default=8088, type=int,\r\n help='server mode reverse server listen port (default: 8088)')\r\n parser.add_argument('-H', dest='connect_host', default=\"127.0.0.1\",\r\n help='client mode reverse client connect server host (default: 127.0.0.1)')\r\n parser.add_argument('-P', dest='connect_port', default=8088, type=int,\r\n help='client mode reverse client connect server port (default: 8088)')\r\n parser.add_argument('-f', dest='forward_host', default=\"\",\r\n help='server and client mode forward host , accept format [remote_host:remote_port] (default: )')\r\n parser.add_argument('-T', dest='proxy_type', default=\"\",\r\n choices=(\"raw\", \"http\", \"socks5\", \"redirect\"), help='server and client mode local listen proxy type (default: raw)')\r\n parser.add_argument('-t', dest='timeout', default=7200,\r\n type=int, help='no read/write timeout (default: 7200)')\r\n args = parser.parse_args(args=argv)\r\n config_signal()\r\n\r\n if not args.forward_host:\r\n forward_address = None\r\n else:\r\n forward_info = args.forward_host.split(\":\")\r\n if len(forward_info) == 1:\r\n if not forward_info[0].isdigit():\r\n forward_address = (forward_info[0], 8088)\r\n else:\r\n forward_address = (\"127.0.0.1\", int(forward_info[0]))\r\n else:\r\n forward_address = (forward_info[0], int(forward_info[1]))\r\n\r\n if not args.is_client_mode:\r\n remote_server = create_server((args.listen_host, args.listen_port))\r\n local_server = create_server((args.bind_host, args.bind_port or 8089))\r\n logging.info(\"listen %s %d -> %s:%d\", args.bind_host, args.bind_port or 8089, args.listen_host, args.listen_port)\r\n\r\n sevent.instance().call_async(server_run_server, remote_server,\r\n forward_address if forward_address else (\"127.0.0.1\", 80),\r\n sevent.utils.ensure_bytes(args.key), args.proxy_type,\r\n conns, status, server_handle_remote_connect)\r\n sevent.instance().call_async(server_run_server, local_server, forward_address,\r\n sevent.utils.ensure_bytes(args.key), args.proxy_type,\r\n conns, status, server_handle_local_connect)\r\n else:\r\n if args.bind_port:\r\n local_server = create_server((args.bind_host, args.bind_port))\r\n logging.info(\"listen %s %d\", args.bind_host, args.bind_port)\r\n sevent.instance().call_async(client_run_server, local_server, (args.connect_host, args.connect_port),\r\n sevent.utils.ensure_bytes(args.key), args.proxy_type, conns)\r\n\r\n logging.info(\"connect %s:%d -> %s\", args.connect_host, args.connect_port, forward_address)\r\n sevent.instance().call_async(client_run_connect, (args.connect_host, args.connect_port),\r\n forward_address if forward_address else (\"127.0.0.1\", 80),\r\n sevent.utils.ensure_bytes(args.key), conns, status)\r\n sevent.current().call_async(check_timeout, conns, status, args.timeout)\r\n\r\nif __name__ == '__main__':\r\n logging.basicConfig(level=logging.INFO, format='%(asctime)s %(levelname)1.1s %(message)s',\r\n datefmt='%Y-%m-%d %H:%M:%S', filemode='a+')\r\n try:\r\n main(sys.argv[1:])\r\n sevent.instance().start()\r\n except KeyboardInterrupt:\r\n exit(0)","sub_path":"sevent/helpers/tcp_reverse.py","file_name":"tcp_reverse.py","file_ext":"py","file_size_in_byte":19384,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"91254686","text":"class TreeNode(object):\n def __init__(self, x):\n self.val = x\n self.left = None\n self.right = None\n\nclass Solution(object):\n def __init__(self):\n self.sear = None\n self.node = None\n self.target = None\n\n def insert(self, root, val):\n if root is None:\n root = TreeNode(val)\n return root\n else:\n return self.insert_again(val, root)\n\n def insert_again(self, val, current):\n if val <= current.val: # 目前的current.val是self.root\n if current.left is None: # 如果root左下沒有數則把帶入的數變成左邊的小孩\n current.left = TreeNode(val)\n return current.left\n else:\n return self.insert_again(val, current.left) # 如果有的話則再跑一次insert_again讓他判斷val比左邊小孩大還小\n else:\n if current.right is None:\n current.right = TreeNode(val)\n return current.right\n else:\n return self.insert_again(val, current.right)\n\n def search(self, root, target):\n if root.val == target:\n return root\n elif root.val < target:\n return self.search(root.right, target)\n elif root.val > target:\n return self.search(root.left, target)\n else:\n return None\n\n def switch(self,root):\n if root.right is None and root.left is None:\n return None\n elif root.right is not None and root.left is not None:\n x = root.left\n while x.right is not None:\n x = x.right\n root.val = x.val\n root.left = self.delete(root.left,root.val)\n return root\n elif root.right is not None and root.left is None:\n x = root.right\n while x.left is not None:\n x = x.left\n root.val = x.val\n root.right = self.delete(root.right,root.val)\n return root\n elif root.right is None and root.left is not None:\n x = root.left\n while x.right is not None:\n x = x.right\n root.val = x.val\n root.left = self.delete(root.left,root.val)\n return root\n \n def delete(self, root, target):\n if self.target is None:\n self.target = target\n if target < root.val:\n if root.left is None:\n return None\n else:\n root.left = self.delete(root.left, target)\n return root\n elif target > root.val:\n if root.right is None:\n return None\n else:\n root.right = self.delete(root.right, target)\n return root\n elif root.val == target:\n root = self.switch(root)\n if root is not None and root.val == self.target:\n root = self.switch(root)\n return root\n \n def modify(self, root, target, new_val):\n x = 0 \n y = self.search(root,target) \n if y is not None and y.val == target: \n x = x + 1 \n y = y.left \n elif y is None:\n return None\n \n self.delete(root,target) \n \n for i in range(x): \n self.insert(root,new_val)\n return root\n","sub_path":"HW3/binary_search_tree_06170133.py","file_name":"binary_search_tree_06170133.py","file_ext":"py","file_size_in_byte":3383,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"65302357","text":"import FWCore.ParameterSet.Config as cms\n\n# fills histograms with all uGMT emulated muons\n# uGMT input muon histograms from track finders are not filled since they are identical to the data DQM plots\nl1tStage2uGMTEmul = cms.EDAnalyzer(\n \"L1TStage2uGMT\",\n bmtfProducer = cms.InputTag(\"gmtStage2Digis\", \"BMTF\"), # not used for emulator DQM\n omtfProducer = cms.InputTag(\"gmtStage2Digis\", \"OMTF\"), # not used for emulator DQM\n emtfProducer = cms.InputTag(\"gmtStage2Digis\", \"EMTF\"), # not used for emulator DQM\n muonProducer = cms.InputTag(\"valGmtStage2Digis\"),\n monitorDir = cms.untracked.string(\"L1T2016EMU/L1TdeStage2uGMT\"),\n emulator = cms.untracked.bool(True),\n verbose = cms.untracked.bool(False),\n)\n\n# the uGMT intermediate muon DQM modules\nl1tStage2uGMTIntermediateBMTFEmul = cms.EDAnalyzer(\n \"L1TStage2uGMTMuon\",\n muonProducer = cms.InputTag(\"valGmtStage2Digis\", \"imdMuonsBMTF\"),\n monitorDir = cms.untracked.string(\"L1T2016EMU/L1TdeStage2uGMT/intermediate_muons/BMTF\"),\n titlePrefix = cms.untracked.string(\"uGMT intermediate muon from BMTF \"),\n verbose = cms.untracked.bool(False),\n)\n\nl1tStage2uGMTIntermediateOMTFNegEmul = cms.EDAnalyzer(\n \"L1TStage2uGMTMuon\",\n muonProducer = cms.InputTag(\"valGmtStage2Digis\", \"imdMuonsOMTFNeg\"),\n monitorDir = cms.untracked.string(\"L1T2016EMU/L1TdeStage2uGMT/intermediate_muons/OMTF_neg\"),\n titlePrefix = cms.untracked.string(\"uGMT intermediate muon from OMTF neg. \"),\n verbose = cms.untracked.bool(False),\n)\n\nl1tStage2uGMTIntermediateOMTFPosEmul = cms.EDAnalyzer(\n \"L1TStage2uGMTMuon\",\n muonProducer = cms.InputTag(\"valGmtStage2Digis\", \"imdMuonsOMTFPos\"),\n monitorDir = cms.untracked.string(\"L1T2016EMU/L1TdeStage2uGMT/intermediate_muons/OMTF_pos\"),\n titlePrefix = cms.untracked.string(\"uGMT intermediate muon from OMTF pos. \"),\n verbose = cms.untracked.bool(False),\n)\n\nl1tStage2uGMTIntermediateEMTFNegEmul = cms.EDAnalyzer(\n \"L1TStage2uGMTMuon\",\n muonProducer = cms.InputTag(\"valGmtStage2Digis\", \"imdMuonsEMTFNeg\"),\n monitorDir = cms.untracked.string(\"L1T2016EMU/L1TdeStage2uGMT/intermediate_muons/EMTF_neg\"),\n titlePrefix = cms.untracked.string(\"uGMT intermediate muon from EMTF neg. \"),\n verbose = cms.untracked.bool(False),\n)\n\nl1tStage2uGMTIntermediateEMTFPosEmul = cms.EDAnalyzer(\n \"L1TStage2uGMTMuon\",\n muonProducer = cms.InputTag(\"valGmtStage2Digis\", \"imdMuonsEMTFPos\"),\n monitorDir = cms.untracked.string(\"L1T2016EMU/L1TdeStage2uGMT/intermediate_muons/EMTF_pos\"),\n titlePrefix = cms.untracked.string(\"uGMT intermediate muon from EMTF pos. \"),\n verbose = cms.untracked.bool(False),\n)\n\n# compares the unpacked uGMT muon collection to the emulated uGMT muon collection\n# only muons that do not match are filled in the histograms\nl1tdeStage2uGMT = cms.EDAnalyzer(\n \"L1TStage2MuonComp\",\n muonCollection1 = cms.InputTag(\"gmtStage2Digis\", \"Muon\"),\n muonCollection2 = cms.InputTag(\"valGmtStage2Digis\"),\n monitorDir = cms.untracked.string(\"L1T2016EMU/L1TdeStage2uGMT/data_vs_emulator_comparison\"),\n muonCollection1Title = cms.untracked.string(\"uGMT data\"),\n muonCollection2Title = cms.untracked.string(\"uGMT emulator\"),\n summaryTitle = cms.untracked.string(\"Summary of comparison between uGMT muons and uGMT emulator muons\"),\n verbose = cms.untracked.bool(False),\n)\n\n","sub_path":"DQM/L1TMonitor/python/L1TdeStage2uGMT_cfi.py","file_name":"L1TdeStage2uGMT_cfi.py","file_ext":"py","file_size_in_byte":3330,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"379287708","text":"import json\nimport msgpack\nimport traceback\n\nfrom time import time\nfrom uuid import uuid4\nfrom datetime import datetime\nfrom auklet.stats import Event, SystemMetrics\nfrom auklet.utils import create_file, \\\n get_abs_path, get_device_ip, open_auklet_url, build_url, \\\n get_agent_version, post_auklet_url, u\n\ntry:\n # For Python 3.0 and later\n from urllib.error import HTTPError, URLError\n from urllib.request import Request, urlopen\nexcept ImportError:\n # Fall back to Python 2's urllib2\n from urllib2 import urlopen, Request, HTTPError, URLError\n\n\nMB_TO_B = 1e6\nS_TO_MS = 1000\n\n\nclass Client(object):\n producer_types = None\n brokers = None\n commit_hash = None\n mac_hash = None\n offline_filename = \"{}/local.txt\"\n limits_filename = \"{}/limits\"\n usage_filename = \"{}/usage\"\n com_config_filename = \"{}/communication\"\n identification_filename = \"{}/identification\"\n abs_path = None\n\n org_id = None\n client_id = None\n broker_username = None\n broker_password = None\n\n reset_data = False\n data_day = 1\n data_limit = None\n data_current = 0\n offline_limit = None\n offline_current = 0\n\n system_metrics = None\n\n def __init__(self, api_key=None, app_id=None, release=None,\n base_url=\"https://api.auklet.io/\", mac_hash=None,\n version=\"\", auklet_dir=\"\"):\n self.apikey = api_key\n self.app_id = app_id\n self.base_url = base_url\n self.send_enabled = True\n self.producer = None\n self.mac_hash = mac_hash\n self.version = version\n self.auklet_dir = auklet_dir\n self._set_filenames()\n self._load_limits()\n create_file(self.offline_filename)\n create_file(self.limits_filename)\n create_file(self.usage_filename)\n create_file(self.com_config_filename)\n create_file(self.identification_filename)\n self.commit_hash = release\n self.abs_path = get_abs_path(\".auklet/version\")\n self.system_metrics = SystemMetrics()\n self._register_device()\n\n def _set_filenames(self):\n self.offline_filename = self.offline_filename.format(self.auklet_dir)\n self.limits_filename = self.limits_filename.format(self.auklet_dir)\n self.usage_filename = self.usage_filename.format(self.auklet_dir)\n self.com_config_filename = self.com_config_filename.format(\n self.auklet_dir)\n self.identification_filename = self.identification_filename.format(\n self.auklet_dir)\n\n def _register_device(self):\n try:\n read_id = json.loads(\n open(self.identification_filename, \"r\").read())\n if not read_id:\n raise IOError\n res, created = self.check_device(read_id['id'])\n if created:\n read_id = res\n except (IOError, ValueError):\n read_id = self.create_device()\n self.broker_password = read_id['client_password']\n self.broker_username = read_id['id']\n self.client_id = read_id['client_id']\n self.org_id = read_id['organization']\n self._write_identification({\"id\": self.broker_username,\n \"client_password\": self.broker_password,\n \"organization\": self.org_id,\n \"client_id\": self.client_id})\n return True\n\n def check_device(self, device_id):\n try:\n opened = open_auklet_url(\n build_url(\n self.base_url,\n \"private/devices/{}/\".format(device_id)\n ),\n self.apikey\n )\n res = json.loads(u(opened.read()))\n created = False\n except HTTPError:\n res = self.create_device()\n created = True\n return res, created\n\n def create_device(self):\n return post_auklet_url(\n build_url(\n self.base_url,\n \"private/devices/\"\n ),\n self.apikey,\n {\"mac_address_hash\": self.mac_hash, \"application\": self.app_id}\n )\n\n def _write_identification(self, data):\n with open(self.identification_filename, \"w\") as id_file:\n id_file.write(json.dumps(data))\n\n def _get_config(self):\n res = open_auklet_url(\n build_url(\n self.base_url,\n \"private/devices/{}/app_config/\".format(self.app_id)),\n self.apikey)\n if res is not None:\n return json.loads(u(res.read()))['config']\n\n def _load_limits(self):\n try:\n with open(self.limits_filename, \"r\") as limits:\n limits_str = limits.read()\n if limits_str:\n data = json.loads(limits_str)\n self.data_day = data['data']['normalized_cell_plan_date']\n temp_limit = data['data']['cellular_data_limit']\n if temp_limit is not None:\n self.data_limit = data['data'][\n 'cellular_data_limit'] * MB_TO_B\n else:\n self.data_limit = temp_limit\n temp_offline = data['storage']['storage_limit']\n if temp_offline is not None:\n self.offline_limit = data['storage'][\n 'storage_limit'] * MB_TO_B\n else:\n self.offline_limit = data['storage']['storage_limit']\n except IOError:\n return\n\n def _build_usage_json(self):\n return {\"data\": self.data_current, \"offline\": self.offline_current}\n\n def _update_usage_file(self):\n try:\n with open(self.usage_filename, 'w') as usage:\n usage.write(json.dumps(self._build_usage_json()))\n except IOError:\n return False\n\n def check_data_limit(self, data, current_use, offline=False):\n if self.offline_limit is None and offline:\n return True\n if self.data_limit is None and not offline:\n return True\n data_size = len(data)\n temp_current = current_use + data_size\n if temp_current >= self.data_limit:\n return False\n if offline:\n self.offline_current = temp_current\n else:\n self.data_current = temp_current\n self._update_usage_file()\n return True\n\n def check_date(self):\n if datetime.today().day == self.data_day:\n if self.reset_data:\n self.data_current = 0\n self.reset_data = False\n else:\n self.reset_data = True\n\n def update_limits(self):\n config = self._get_config()\n if config is None:\n return 60000\n with open(self.limits_filename, 'w+') as limits:\n limits.truncate()\n limits.write(json.dumps(config))\n new_day = config['data']['normalized_cell_plan_date']\n temp_limit = config['data']['cellular_data_limit']\n if temp_limit is not None:\n new_data = config['data']['cellular_data_limit'] * MB_TO_B\n else:\n new_data = temp_limit\n temp_offline = config['storage']['storage_limit']\n if temp_offline is not None:\n new_offline = config['storage']['storage_limit'] * MB_TO_B\n else:\n new_offline = config['storage']['storage_limit']\n if self.data_day != new_day:\n self.data_day = new_day\n self.data_current = 0\n if self.data_limit != new_data:\n self.data_limit = new_data\n if self.offline_limit != new_offline:\n self.offline_limit = new_offline\n # return emission period in ms\n return config['emission_period'] * S_TO_MS\n\n def build_event_data(self, type, tb, tree):\n event = Event(type, tb, tree, self.abs_path)\n event_dict = dict(event)\n event_dict['application'] = self.app_id\n event_dict['publicIP'] = get_device_ip()\n event_dict['id'] = str(uuid4())\n event_dict['timestamp'] = int(round(time() * 1000))\n event_dict['systemMetrics'] = dict(self.system_metrics)\n event_dict['macAddressHash'] = self.mac_hash\n event_dict['release'] = self.commit_hash\n event_dict['agentVersion'] = get_agent_version()\n event_dict['device'] = self.broker_username\n event_dict['absPath'] = self.abs_path\n event_dict['version'] = self.version\n return event_dict\n\n def build_log_data(self, msg, data_type, level):\n log_dict = {\n \"message\": msg,\n \"type\": data_type,\n \"level\": level,\n \"application\": self.app_id,\n \"publicIP\": get_device_ip(),\n \"id\": str(uuid4()),\n \"timestamp\": int(round(time() * 1000)),\n \"systemMetrics\": dict(self.system_metrics),\n \"macAddressHash\": self.mac_hash,\n \"release\": self.commit_hash,\n \"agentVersion\": get_agent_version(),\n \"device\": self.broker_username,\n \"version\": self.version\n }\n return log_dict\n\n def build_msgpack_event_data(self, type, tb, tree):\n event_data = self.build_event_data(type, tb, tree)\n return msgpack.packb(event_data, use_bin_type=False)\n\n def build_msgpack_log_data(self, msg, data_type, level):\n log_data = self.build_log_data(msg, data_type, level)\n return msgpack.packb(log_data, use_bin_type=False)\n","sub_path":"auklet/monitoring/processing.py","file_name":"processing.py","file_ext":"py","file_size_in_byte":9595,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"511100187","text":"# -*- coding: utf-8 -*-\n\n\"\"\"\nTests for Board class.\n\"\"\"\n\nimport pytest\n\nfrom ..board import Board\n\n__author__ = 'Hans Ekkehard Plesser'\n__email__ = 'hans.ekkehard.plesser@nmbu.no'\n\n\ndef test_goal_not_reached():\n \"\"\"\"Ensure goal_reached() does not yield false positives.\"\"\"\n goal_pos = 20\n brd = Board(ladders=[], chutes=[], goal=goal_pos)\n for pos in range(goal_pos):\n assert not brd.goal_reached(pos)\n\n\ndef test_goal_reached():\n \"\"\"Ensure goal_reached() does not yield false negatives.\"\"\"\n goal_pos = 20\n brd = Board(ladders=[], chutes=[], goal=goal_pos)\n for pos in range(goal_pos, goal_pos+10):\n assert brd.goal_reached(pos)\n\n\ndef test_adjust_empty_board():\n \"\"\"\"No position adjustment on empty board.\"\"\"\n\n goal_pos = 20\n brd = Board(ladders=[], chutes=[], goal=goal_pos)\n for pos in range(goal_pos):\n assert brd.position_adjustment(pos) == 0\n\n\ndef test_adjustment():\n goal_pos = 20\n ladders = [(2, 10), (9, 13), (12, 18)]\n chutes = [(4, 1), (7, 3), (17, 8)]\n test_cases = {0: 0, 1: 0, 2: 8, 3: 0, 4: -3, 5: 0, 6: 0, 7: -4,\n 8: 0, 9: 4, 10: 0, 11: 0, 12: 6, 13: 0, 14: 0,\n 15: 0, 16: 0, 17: -9, 18: 0, 19: 0}\n brd = Board(ladders=ladders, chutes=chutes, goal=goal_pos)\n for pos, change in test_cases.items():\n assert brd.position_adjustment(pos) == change\n\n\ndef test_default_board():\n \"\"\"Some tests on default board.\"\"\"\n\n brd = Board()\n assert brd.position_adjustment(1) == 39\n assert brd.position_adjustment(2) == 0\n assert brd.position_adjustment(33) == -30\n assert not brd.goal_reached(89)\n assert brd.goal_reached(90)\n assert brd.goal_reached(91)\n\n\ndef test_bad_boards():\n \"\"\"Test that bad board specifications are not accepted.\"\"\"\n\n with pytest.raises(ValueError):\n Board(ladders=[(10, 10)])\n\n with pytest.raises(ValueError):\n Board(ladders=[(10, 9)])\n\n with pytest.raises(ValueError):\n Board(chutes=[(10, 10)])\n\n with pytest.raises(ValueError):\n Board(chutes=[(10, 12)])\n\n with pytest.raises(ValueError):\n Board(goal=0)\n","sub_path":"Project/SampleProjects/chutes_project_var1/chutes/tests/test_board.py","file_name":"test_board.py","file_ext":"py","file_size_in_byte":2122,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"160222564","text":"'''\nRDS instance performance insights\n'''\n\nfrom botocore.exceptions import ClientError\n\ndef run_remediation(rds, RDSInstanceName):\n print(\"Executing RDS Instance remediation\")\n performance_insights='' \n try:\n response = rds.describe_db_instances(DBInstanceIdentifier=RDSInstanceName)['DBInstances']\n DBInstanceClass=response[0]['DBInstanceClass']\n performance_insights=response[0]['PerformanceInsightsEnabled']\n except ClientError as e:\n responseCode = 400\n output = \"Unexpected error: \" + str(e)\n except Exception as e:\n responseCode = 400\n output = \"Unexpected error: \" + str(e)\n\n if DBInstanceClass not in ['db.t2.micro', 'db.t2.small', 'db.t3.micro', 'db.t3.small']:\n if not performance_insights: \n while response[0]['DBInstanceStatus'] not in ['available', 'stopped']:\n try:\n response = rds.describe_db_instances(DBInstanceIdentifier=RDSInstanceName)['DBInstances']\n except ClientError as e:\n responseCode = 400\n output = \"Unexpected error: \" + str(e)\n except Exception as e:\n responseCode = 400\n output = \"Unexpected error: \" + str(e)\n\n try:\n result = rds.modify_db_instance(\n DBInstanceIdentifier=RDSInstanceName,\n ApplyImmediately=True,\n EnablePerformanceInsights =True\n )\n\n responseCode = result['ResponseMetadata']['HTTPStatusCode']\n if responseCode >= 400:\n output = \"Unexpected error: %s \\n\" % str(result)\n else:\n output = \"Performance insights enabled for rds-instance : %s \\n\" % RDSInstanceName\n \n except ClientError as e:\n responseCode = 400\n output = \"Unexpected error: \" + str(e)\n print(output)\n except Exception as e:\n responseCode = 400\n output = \"Unexpected error: \" + str(e)\n print(output)\n\n else:\n responseCode=200\n output='Performance insights already enabled for rds-instance : '+RDSInstanceName\n print(output)\n else:\n responseCode=200\n output='Performance insights is not supported for rds-instance : '+RDSInstanceName\n print(output)\n\n print(str(responseCode)+'-'+output)\n return responseCode,output","sub_path":"remediation-functions/rds_instance/rdsinstance_performanceinsights.py","file_name":"rdsinstance_performanceinsights.py","file_ext":"py","file_size_in_byte":2538,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"204740676","text":"import concurrent.futures\n\n\nclass Parallelize:\n def __init__(self, _func, _serialize=False):\n self.func = _func\n self.serialize = _serialize\n\n def __call__(self, *args):\n if self.serialize:\n results = list(map(self.func, *args))\n else:\n # multiprocessor mode\n with concurrent.futures.ProcessPoolExecutor() as executor:\n results = list(executor.map(self.func, *args))\n\n return results\n\n\n ","sub_path":"fractal/utils/multiproc.py","file_name":"multiproc.py","file_ext":"py","file_size_in_byte":479,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"210857399","text":"import search\nimport random\nimport math\nfrom copy import deepcopy\n\nids = [\"204657977\"]\nprint_all_headlines = False\nprint_main_headlines = False\n\n\nclass MyState:\n def __init__(self):\n self.targets_hit = set()\n self.devices_status = set()\n self.occupied_coordinates = set()\n self.spaceships_occupied_identifiers = set()\n\n def assign_all_variables(self, targets_occupied_coordinates, spaceships_occupied_coordinates, targets_hit_set,\n devices_status_set):\n self.occupied_coordinates = targets_occupied_coordinates\n self.spaceships_occupied_identifiers = spaceships_occupied_coordinates\n self.targets_hit = targets_hit_set\n self.devices_status = devices_status_set\n\n def add_target_occupied_coordinate(self, occupied_coordinate):\n self.occupied_coordinates.add(occupied_coordinate)\n\n def add_spaceship_occupied_coordinate(self, occupied_coordinate, occupied_identifier):\n self.occupied_coordinates.add(occupied_coordinate)\n self.spaceships_occupied_identifiers.add(occupied_identifier)\n\n def remove_spaceship_occupied_coordinate(self, occupied_coordinate, occupied_identifier):\n self.occupied_coordinates.remove(occupied_coordinate)\n self.spaceships_occupied_identifiers.remove(occupied_identifier)\n\n def add_target_hit(self, target_hit_tuple):\n self.targets_hit.add(target_hit_tuple)\n\n def add_device_status(self, device_status_tuple):\n self.devices_status.add(device_status_tuple)\n\n def remove_device_status(self, device_status_tuple):\n self.devices_status.remove(device_status_tuple)\n\n def __hash__(self):\n return hash((frozenset(self.occupied_coordinates), frozenset(self.spaceships_occupied_identifiers),\n frozenset(self.targets_hit), frozenset(self.devices_status)))\n\n def __eq__(self, other):\n return self.occupied_coordinates == other.occupied_coordinates \\\n and self.spaceships_occupied_identifiers == other.spaceships_occupied_identifiers \\\n and self.targets_hit == other.targets_hit \\\n and set(self.devices_status) == set(other.devices_status)\n\n\nclass UtilsActions:\n @staticmethod\n def good_cop_bad_cop(current_name, current_coordinates, state, grid_size, distances_directions):\n if print_all_headlines:\n print(\">>> Function: good_cop_bad_cop\")\n\n temp_move_actions_set = set()\n for dist, dirc in distances_directions:\n new_coordinates = UtilsActions.new_coordinates_by_direction(current_coordinates, dirc)\n\n new_coordinates_easy_check, _ = UtilsActions.coordinates_legalization(new_coordinates, state, grid_size,\n priority=False)\n\n other_new_coordinates = None\n other_spaceship_name = None\n if not new_coordinates_easy_check:\n other_new_coordinates, other_spaceship_name = UtilsActions.coordinates_legalization(new_coordinates,\n state,\n grid_size,\n not_in_direction=[dirc],\n priority=True)\n\n if new_coordinates_easy_check:\n temp_move_actions_set.add((\"move\", current_name, current_coordinates, new_coordinates_easy_check))\n if other_new_coordinates and other_spaceship_name:\n temp_move_actions_set.add((\"move\", other_spaceship_name, new_coordinates, other_new_coordinates))\n\n return temp_move_actions_set\n\n @staticmethod\n def get_move_action_priority_calibration(current_estimate, state, grid_size):\n if print_all_headlines:\n print(\">>> Function: get_move_action_priority_calibration\")\n\n move_actions_set = set()\n current_ship_name = current_estimate[0].name\n current_ship_coordinates = current_estimate[0].coordinates\n distances_directions = current_estimate[2]\n\n if isinstance(distances_directions[0], int):\n distances_directions = (distances_directions,)\n\n move_actions_set.update(\n UtilsActions.good_cop_bad_cop(current_ship_name, current_ship_coordinates, state, grid_size,\n distances_directions))\n\n new_coordinates = UtilsActions.move_me_anywhere(current_ship_coordinates, state, grid_size, main_ship=False,\n not_in_direction=[row[1] for row in distances_directions])\n if new_coordinates:\n move_actions_set.add((\"move\", current_ship_name, current_ship_coordinates, new_coordinates))\n\n # new_coordinates = UtilsActions.move_me_anywhere(current_ship_coordinates, state, grid_size, main_ship=True)\n # if new_coordinates:\n # for current_new_coordinates in new_coordinates:\n # move_actions_set.add((\"move\", current_ship_name, current_ship_coordinates, current_new_coordinates))\n\n return move_actions_set\n\n @staticmethod\n def get_move_action_priority_targets(current_estimate, current_spaceship, state, grid_size):\n if print_all_headlines:\n print(\">>> Function: get_move_action_priority_targets\")\n\n move_actions_set = set()\n current_ship_name = current_spaceship.name\n current_ship_coordinates = current_spaceship.coordinates\n distances_directions = current_estimate[2]\n\n if isinstance(distances_directions[0], int):\n distances_directions = (distances_directions,)\n\n move_actions_set.update(\n UtilsActions.good_cop_bad_cop(current_ship_name, current_ship_coordinates, state, grid_size,\n distances_directions))\n\n new_coordinates = UtilsActions.move_me_anywhere(current_ship_coordinates, state, grid_size, main_ship=False,\n not_in_direction=[row[1] for row in distances_directions])\n if new_coordinates:\n move_actions_set.add((\"move\", current_ship_name, current_ship_coordinates, new_coordinates))\n\n # new_coordinates = UtilsActions.move_me_anywhere(current_ship_coordinates, state, grid_size, main_ship=True)\n # if new_coordinates:\n # for current_new_coordinates in new_coordinates:\n # move_actions_set.add((\"move\", current_ship_name, current_ship_coordinates, current_new_coordinates))\n\n return move_actions_set\n\n @staticmethod\n def coordinates_legalization(current_coordinates, state, grid_size, not_in_direction=None, priority=True):\n if print_all_headlines:\n print(\">>> Function: coordinates_legalization\")\n\n if UtilsActions.check_within_grid_range(current_coordinates, grid_size):\n if current_coordinates not in state.occupied_coordinates:\n return (current_coordinates, None)\n elif priority:\n for check_identifier in state.spaceships_occupied_identifiers:\n x, y, z, spaceship_name = check_identifier\n check_coordinates = (x, y, z)\n if current_coordinates == check_coordinates:\n new_coordinates = UtilsActions.move_me_anywhere(check_coordinates, state, grid_size,\n not_in_direction)\n\n if new_coordinates:\n return (new_coordinates, spaceship_name)\n break\n\n return None, None\n\n @staticmethod\n def check_within_grid_range(current_coordinates, grid_size):\n if print_all_headlines:\n print(\">>> Function: check_within_grid_range\")\n x, y, z = current_coordinates\n grid_legalization_part_1 = x < grid_size and y < grid_size and z < grid_size\n grid_legalization_part_2 = x >= 0 and y >= 0 and z >= 0\n return grid_legalization_part_1 and grid_legalization_part_2\n\n @staticmethod\n def check_empty_cell(current_coordinates, state):\n if print_all_headlines:\n print(\">>> Function: check_empty_cell\")\n if current_coordinates not in state.occupied_coordinates:\n return True\n\n\n\n @staticmethod\n def move_me_anywhere(current_coordinates, state, grid_size, not_in_direction=None, main_ship = False):\n if print_all_headlines:\n print(\">>> Function: move_me_anywhere\")\n\n movements_set = set()\n possible_directions = ['dx', 'dy', 'dz', '-dx', '-dy', '-dz']\n if not_in_direction:\n for current_dirc in not_in_direction:\n possible_directions.remove(current_dirc)\n for dirc in possible_directions:\n new_coordinates = UtilsActions.new_coordinates_by_direction(current_coordinates, dirc)\n if UtilsActions.check_within_grid_range(new_coordinates, grid_size) and UtilsActions.check_empty_cell(\n new_coordinates, state):\n if main_ship:\n movements_set.add(new_coordinates)\n else:\n return new_coordinates\n\n if not_in_direction:\n new_coordinates = UtilsActions.new_coordinates_by_direction(current_coordinates, not_in_direction)\n if UtilsActions.check_within_grid_range(new_coordinates, grid_size) and UtilsActions.check_empty_cell(\n new_coordinates, state):\n if main_ship:\n movements_set.add(new_coordinates)\n else:\n return new_coordinates\n\n if main_ship:\n return movements_set\n return None\n\n @staticmethod\n def new_coordinates_by_direction(current_coordinates, direction):\n if print_all_headlines:\n print(\">>> Function: new_coordinates_by_direction\")\n if direction == 'dx':\n return (current_coordinates[0] + 1, current_coordinates[1], current_coordinates[2])\n elif direction == '-dx':\n return (current_coordinates[0] - 1, current_coordinates[1], current_coordinates[2])\n elif direction == 'dy':\n return (current_coordinates[0], current_coordinates[1] + 1, current_coordinates[2])\n elif direction == '-dy':\n return (current_coordinates[0], current_coordinates[1] - 1, current_coordinates[2])\n elif direction == 'dz':\n return (current_coordinates[0], current_coordinates[1], current_coordinates[2] + 1)\n else:\n return (current_coordinates[0], current_coordinates[1], current_coordinates[2] - 1)\n\n @staticmethod\n def check_straight_line(tuple_coordinate1, tuple_coordinate2):\n if print_all_headlines:\n print(\">>> Function: check_straight_line\")\n dx = tuple_coordinate2[0] - tuple_coordinate1[0]\n dy = tuple_coordinate2[1] - tuple_coordinate1[1]\n dz = tuple_coordinate2[2] - tuple_coordinate1[2]\n\n if dx == 0 and dy == 0:\n return (dz, 'dz', 1) if dz > 0 else (-dz, '-dz', 1)\n\n if dx == 0 and dz == 0:\n return (dy, 'dy', 1) if dy > 0 else (-dy, '-dy', 1)\n\n if dy == 0 and dz == 0:\n return (dx, 'dx', 1) if dx > 0 else (-dx, '-dx', 1)\n\n distances_to_return = []\n directions_to_return = []\n if dx:\n if dx > 0:\n distances_to_return.append(dx)\n directions_to_return.append('dx')\n else:\n distances_to_return.append(-dx)\n directions_to_return.append('-dx')\n\n if dy:\n if dy > 0:\n distances_to_return.append(dy)\n directions_to_return.append('dy')\n else:\n distances_to_return.append(-dy)\n directions_to_return.append('-dy')\n\n if dz:\n if dz > 0:\n distances_to_return.append(dz)\n directions_to_return.append('dz')\n else:\n distances_to_return.append(-dz)\n directions_to_return.append('-dz')\n\n return distances_to_return, directions_to_return, 0\n\n\nclass UtilsUpdate:\n @staticmethod\n def distribute_packet(action, spaceships_dictionary, targets_dictionary, state):\n if print_all_headlines:\n print(\">>> Function: distribute_packet\")\n if action[0] == 'move':\n UtilsUpdate.update_move_action(spaceships_dictionary[action[1]], action, state)\n elif action[0] == 'turn_on':\n UtilsUpdate.update_turnon_action(spaceships_dictionary[action[1]], action, state)\n elif action[0] == 'calibrate':\n UtilsUpdate.update_calibrate_action(spaceships_dictionary[action[1]], action, state)\n else: # elif action[0] == 'use':\n UtilsUpdate.update_use_action(spaceships_dictionary[action[1]], action, targets_dictionary, state)\n\n @staticmethod\n def update_move_action(spaceship_object, move_action, state):\n if print_all_headlines:\n print(\">>> Function: update_move_action\")\n spaceship_old_coordinates = spaceship_object.coordinates\n spaceship_old_identifier = spaceship_object.get_spaceship_identifier()\n state.remove_spaceship_occupied_coordinate(spaceship_old_coordinates, spaceship_old_identifier)\n spaceship_object.coordinates = move_action[3]\n state.add_spaceship_occupied_coordinate(spaceship_object.coordinates,\n spaceship_object.get_spaceship_identifier())\n\n @staticmethod\n def update_turnon_action(spaceship_object, turnon_action, state):\n if print_all_headlines:\n print(\">>> Function: update_turnon_action\")\n value_to_remove = spaceship_object.turn_off_devices()\n value_to_add = spaceship_object.turn_on_device(turnon_action[2])\n if value_to_remove:\n state.remove_device_status(value_to_remove)\n state.add_device_status(value_to_add)\n\n @staticmethod\n def update_calibrate_action(spaceship_object, calibrate_action, state):\n if print_all_headlines:\n print(\">>> Function: update_calibrate_action\")\n value_to_remove, value_to_add = spaceship_object.calibrate_device(calibrate_action[2])\n if value_to_remove:\n state.remove_device_status(value_to_remove)\n state.add_device_status(value_to_add)\n\n @staticmethod\n def update_use_action(spaceship_object, use_action, targets_dictionary, state):\n if print_all_headlines:\n print(\">>> Function: update_use_action\")\n current_device = use_action[2]\n current_target = use_action[3]\n targets_dictionary[current_target] = tuple(\n device for device in targets_dictionary[current_target] if device != current_device)\n state.add_target_hit((current_target, current_device))\n\n\nclass Spaceship:\n def __init__(self, coordinates, name, devices):\n self.coordinates = coordinates\n self.name = name\n self.assigned_targets = set()\n\n self.devices = devices\n self.devices_onoff = {}\n self.devices_calibration = {}\n self.active_device = None\n\n self.initialize_devices_state()\n\n def initialize_devices_state(self):\n for current_device in self.devices:\n if current_device not in self.devices_onoff.keys():\n self.devices_onoff[current_device] = 0\n self.devices_calibration[current_device] = 0\n\n def check_ready_fire(self, device):\n if self.check_device_exists(device):\n if self.devices_onoff[device] == 1 and self.devices_calibration[device] == 1:\n return True\n return False\n\n def check_ready_calibrate(self, device):\n if self.check_device_exists(device):\n if self.devices_onoff[device] == 1 and self.devices_calibration[device] == 0:\n return True\n return False\n\n def check_ready_turnon(self, device):\n if self.check_device_exists(device):\n if self.devices_onoff[device] == 0:\n return True\n return False\n\n def check_device_exists(self, device):\n if device in self.devices:\n return True\n return False\n\n def calibrate_device(self, device):\n value_to_remove = None\n value_to_add = None\n if self.active_device:\n value_to_remove = self.get_tuple_active_device()\n # else:\n # print(\"DELETE ME: CAN'T BE HERE.\")\n self.devices_calibration[device] = 1\n value_to_add = self.get_tuple_active_device()\n return value_to_remove, value_to_add\n\n def turn_on_device(self, device):\n self.devices_onoff[device] = 1\n self.active_device = device\n return self.get_tuple_active_device()\n\n def turn_off_devices(self):\n value_to_return = None\n if self.active_device:\n value_to_return = self.get_tuple_active_device()\n self.i_said_turnoff()\n return value_to_return\n\n def i_said_turnoff(self):\n self.devices_onoff = {key: 0 for key in self.devices_onoff}\n self.devices_calibration = {key: 0 for key in self.devices_calibration}\n self.active_device = None\n\n def get_tuple_active_device(self):\n if self.devices_calibration[self.active_device] != 0:\n return (self.name, self.active_device, 1, 1)\n else:\n return (self.name, self.active_device, 1, 0)\n\n def get_spaceship_identifier(self):\n return (self.coordinates + (self.name,))\n\n\nclass UtilsGeneral:\n @staticmethod\n def get_all_distances_dict(spaceships, targets_dict, calibration_targets_dict, specific_spaceship=None):\n all_distances_dict = dict()\n\n if not specific_spaceship:\n for current_spaceship in spaceships.values():\n current_spaceship_coordinates = current_spaceship.coordinates\n\n for current_target_coordinates in targets_dict:\n temp_distance, temp_direction, is_straight_line = UtilsActions.check_straight_line(\n current_spaceship_coordinates, current_target_coordinates)\n all_distances_dict[(current_spaceship_coordinates, current_target_coordinates)] = [\n temp_distance, temp_direction, is_straight_line]\n\n for temp_calib_coordinates in calibration_targets_dict.values():\n temp_distance, temp_direction, is_straight_line = UtilsActions.check_straight_line(\n current_spaceship_coordinates, temp_calib_coordinates)\n all_distances_dict[(current_spaceship_coordinates, temp_calib_coordinates)] = [\n temp_distance, temp_direction, is_straight_line]\n\n else:\n current_spaceship_coordinates = specific_spaceship.coordinates\n\n for current_target_coordinates in targets_dict:\n temp_distance, temp_direction, is_straight_line = UtilsActions.check_straight_line(\n current_spaceship_coordinates, current_target_coordinates)\n all_distances_dict[(current_spaceship_coordinates, current_target_coordinates)] = [\n temp_distance, temp_direction, is_straight_line]\n\n for temp_calib_coordinates in calibration_targets_dict.values():\n temp_distance, temp_direction, is_straight_line = UtilsActions.check_straight_line(\n current_spaceship_coordinates, temp_calib_coordinates)\n all_distances_dict[(current_spaceship_coordinates, temp_calib_coordinates)] = [\n temp_distance, temp_direction, is_straight_line]\n\n return all_distances_dict\n\n\nclass PriorityManager:\n def __init__(self, overview_obj):\n self.overview_obj = overview_obj\n self.device_spaceship_deadend = dict()\n self.dead_devices = set()\n\n def initialize_on_line_device(self):\n if print_all_headlines:\n print(\">>> Function: initialize_on_line_device\")\n\n self.overview_obj.is_turned_on = 0\n self.overview_obj.is_calibrated = 0\n self.overview_obj.calibration_current_estimate_index = 0\n self.overview_obj.use_device_current_estimate_index = 0\n\n def get_new_active_weapon(self):\n if print_all_headlines:\n print(\">>> Function: get_new_active_weapon\")\n\n for current_device in self.overview_obj.devices:\n if self.overview_obj.total_number_of_hits_per_device[current_device] != 0:\n self.overview_obj.online_device_id = self.overview_obj.devices.index(current_device)\n self.initialize_on_line_device()\n break\n\n def kill_device(self, current_device):\n if print_all_headlines:\n print(\">>> Function: kill_device. Device to kill: \" + str(current_device))\n self.overview_obj.dictionary_device_spaceships_priorities[current_device] = None\n self.overview_obj.dictionary_device_targets_priorities[current_device] = None\n self.dead_devices.add(current_device)\n\n def check_for_occlusions_calib(self, current_spaceship_coordinates, current_calibration_coordinates,\n distances_directions):\n if print_all_headlines:\n print(\">>> Function: check_for_occlusions_calib\")\n\n distances, directions = distances_directions\n\n found_occlusion = False\n occlusions_distances_set = set()\n\n for temp_target in self.overview_obj.targets_dict:\n temp_distance, temp_direction, is_target_straight = UtilsActions.check_straight_line(\n current_spaceship_coordinates, temp_target)\n if is_target_straight:\n if temp_direction == directions:\n if temp_distance < distances:\n found_occlusion = True\n occlusions_distances_set.add(temp_distance)\n\n for temp_calib in self.overview_obj.calibration_targets_dict.values():\n if temp_calib != current_calibration_coordinates:\n temp_distance, temp_direction, is_target_straight = UtilsActions.check_straight_line(\n current_spaceship_coordinates, temp_calib)\n if is_target_straight:\n if temp_direction == directions:\n if temp_distance < distances:\n found_occlusion = True\n occlusions_distances_set.add(temp_distance)\n\n return found_occlusion, occlusions_distances_set\n\n def check_for_occlusions_target(self, current_spaceship_coordinates, current_target_coordinates,\n distances_directions):\n if print_all_headlines:\n print(\">>> Function: check_for_occlusions_target\")\n\n distances, directions = distances_directions\n\n found_occlusion = False\n occlusions_distances_set = set()\n\n for temp_target in self.overview_obj.targets_dict:\n if temp_target != current_target_coordinates:\n temp_distance, temp_direction, is_target_straight = UtilsActions.check_straight_line(\n current_spaceship_coordinates, temp_target)\n if is_target_straight:\n if temp_direction == directions:\n if temp_distance < distances:\n found_occlusion = True\n occlusions_distances_set.add(temp_distance)\n\n for temp_calib in self.overview_obj.calibration_targets_dict.values():\n temp_distance, temp_direction, is_target_straight = UtilsActions.check_straight_line(\n current_spaceship_coordinates, temp_calib)\n if is_target_straight:\n if temp_direction == directions:\n if temp_distance < distances:\n found_occlusion = True\n occlusions_distances_set.add(temp_distance)\n\n return found_occlusion, occlusions_distances_set\n\n def add_priority_to_sorted_list_calib(self, current_spaceship, current_device, priority_value,\n distances_directions):\n for i_index in range(len(self.overview_obj.dictionary_device_spaceships_priorities[current_device])):\n if priority_value < \\\n self.overview_obj.dictionary_device_spaceships_priorities[current_device][i_index][1]:\n self.overview_obj.dictionary_device_spaceships_priorities[current_device].insert(i_index,\n [current_spaceship,\n priority_value,\n distances_directions])\n break\n else:\n self.overview_obj.dictionary_device_spaceships_priorities[current_device].append(\n [current_spaceship, priority_value, distances_directions])\n\n def add_priority_to_sorted_list_target(self, current_target_coordinates, current_device, priority_value,\n distances_directions):\n for i_index in range(len(self.overview_obj.dictionary_device_targets_priorities[current_device])):\n if priority_value < \\\n self.overview_obj.dictionary_device_targets_priorities[current_device][i_index][1]:\n self.overview_obj.dictionary_device_targets_priorities[current_device].insert(i_index,\n [\n current_target_coordinates,\n priority_value,\n distances_directions])\n break\n else:\n self.overview_obj.dictionary_device_targets_priorities[current_device].append(\n [current_target_coordinates, priority_value, distances_directions])\n\n def get_straight_priority_order(self, target_distance, occlusions_distances_set):\n number_of_occlusions = len(occlusions_distances_set)\n min_distance = min(occlusions_distances_set)\n max_distance = max(occlusions_distances_set)\n\n if number_of_occlusions == 1:\n if max_distance == target_distance - 1:\n return min_distance + 3\n else:\n return min_distance + 4\n else:\n return min_distance - 1 + number_of_occlusions * 3\n\n def add_device_ship_deadend(self, device_name, ship_name):\n if device_name in self.device_spaceship_deadend:\n if ship_name not in self.device_spaceship_deadend[device_name]:\n self.device_spaceship_deadend[device_name].append(ship_name)\n else:\n self.device_spaceship_deadend[device_name] = [ship_name]\n\n\nclass CalibPriorityManager(PriorityManager):\n def __init__(self, overview_obj):\n super().__init__(overview_obj)\n self.devices_introduced = set()\n\n def reset_introduced_device(self, current_device):\n if current_device in self.devices_introduced:\n self.devices_introduced.remove(current_device)\n\n def pass_device_checks(self, current_device):\n # current_device - Current Device Name (string)\n if self.overview_obj.total_number_of_hits_per_device[current_device] != 0:\n return True\n\n else:\n self.kill_device(current_device)\n return False\n\n def pass_spaceship_checks(self, current_device, current_spaceship):\n if current_spaceship.check_device_exists(current_device):\n if current_device in self.device_spaceship_deadend:\n if current_spaceship.name in self.device_spaceship_deadend[current_device]:\n return False\n return True\n return False\n\n def get_current_calib_priority(self, current_spaceship, device_calibration_coordinates, all_distances_dict):\n if print_all_headlines:\n print(\">>> Function: get_current_calib_priority\")\n\n current_spaceship_coordinates = current_spaceship.coordinates\n distances, directions, is_straight_line = all_distances_dict[\n (current_spaceship_coordinates, device_calibration_coordinates)]\n\n if is_straight_line:\n found_occlusion, occlusions_distances_set = self.check_for_occlusions_calib(\n current_spaceship_coordinates, device_calibration_coordinates, (distances, directions))\n\n if not found_occlusion:\n priority_value = 0\n else:\n priority_value = self.get_straight_priority_order(distances, occlusions_distances_set)\n\n return [current_spaceship, priority_value, (distances, directions)]\n\n else:\n priority_value = sum(distances) - max(distances)\n distances_directions = sorted(zip(distances, directions), key=lambda dist: dist[0])\n return [current_spaceship, priority_value, distances_directions]\n\n def priority_calib_main(self):\n if print_all_headlines or print_main_headlines:\n print(\">>> Function: priority_calib_main\")\n\n def get_current_device_data(current_device):\n if self.pass_device_checks(current_device):\n temporary_priorities = []\n device_calibration_coordinates = self.overview_obj.calibration_targets_dict[current_device]\n\n for current_spaceship in self.overview_obj.spaceships.values():\n if self.pass_spaceship_checks(current_device, current_spaceship):\n current_priority = self.get_current_calib_priority(current_spaceship,\n device_calibration_coordinates,\n all_distances_dict)\n\n temporary_priorities.append(current_priority)\n\n # if current_priority[1] == 0:\n # temporary_priorities.insert(0, current_priority)\n # self.overview_obj.dictionary_device_spaceships_priorities[\n # current_device] = temporary_priorities\n # return True\n #\n # else:\n # temporary_priorities.append(current_priority)\n\n if len(temporary_priorities) > 0:\n temporary_priorities.sort(key=lambda priority: priority[1])\n self.overview_obj.dictionary_device_spaceships_priorities[current_device] = temporary_priorities\n return True\n\n return False\n\n all_distances_dict = UtilsGeneral.get_all_distances_dict(self.overview_obj.spaceships,\n self.overview_obj.targets_dict,\n self.overview_obj.calibration_targets_dict)\n\n for current_device in self.overview_obj.devices:\n if current_device not in self.dead_devices:\n if get_current_device_data(current_device):\n self.reset_introduced_device(current_device)\n break\n\n self.get_new_active_weapon()\n\n def priority_calib_update(self):\n if print_all_headlines or print_main_headlines:\n print(\">>> Function: priority_calib_update\")\n\n current_device = self.overview_obj.devices[self.overview_obj.online_device_id]\n current_estimate = self.overview_obj.dictionary_device_spaceships_priorities[current_device][\n self.overview_obj.calibration_current_estimate_index]\n current_estimate_spaceship = current_estimate[0]\n current_spaceship_coordinates = current_estimate_spaceship.coordinates\n device_calibration_coordinates = self.overview_obj.calibration_targets_dict[current_device]\n\n distances, directions, is_straight_line = UtilsActions.check_straight_line(\n current_spaceship_coordinates, device_calibration_coordinates)\n\n if is_straight_line:\n found_occlusion, occlusions_distances_set = self.check_for_occlusions_calib(\n current_spaceship_coordinates, device_calibration_coordinates, (distances, directions))\n\n if not found_occlusion:\n priority_value = 0\n\n else:\n priority_value = self.get_straight_priority_order(distances, occlusions_distances_set)\n\n self.overview_obj.dictionary_device_spaceships_priorities[current_device][\n self.overview_obj.calibration_current_estimate_index] = [\n current_estimate_spaceship, priority_value, (distances, directions)]\n\n else:\n priority_value = sum(distances) - max(distances)\n distances_directions = sorted(zip(distances, directions), key=lambda dist: dist[0])\n\n self.overview_obj.dictionary_device_spaceships_priorities[current_device][\n self.overview_obj.calibration_current_estimate_index] = [\n current_estimate_spaceship, priority_value, distances_directions]\n\n def get_calib_estimate_action(self, current_estimate, current_device, as_single = False):\n current_estimate_spaceship = current_estimate[0]\n current_estimate_spaceship_name = current_estimate_spaceship.name\n\n if current_estimate[1] == 0:\n if current_estimate_spaceship.check_ready_calibrate(current_device):\n if as_single:\n return ('calibrate', current_estimate_spaceship_name, current_device,\n self.overview_obj.calibration_targets_dict[current_device])\n else:\n return (('calibrate', current_estimate_spaceship_name, current_device,\n self.overview_obj.calibration_targets_dict[current_device]),)\n else:\n if as_single:\n return ('turn_on', current_estimate_spaceship_name, current_device)\n else:\n return (('turn_on', current_estimate_spaceship_name, current_device),)\n else:\n return UtilsActions.get_move_action_priority_calibration(current_estimate, self.overview_obj.state,\n self.overview_obj.grid_size)\n\n def get_calibration_actions(self):\n if print_all_headlines or print_main_headlines:\n print(\">>> Function: get_calibration_actions\")\n\n def introduce_device(current_device):\n if current_device in self.devices_introduced:\n return False\n return True\n\n current_device = self.overview_obj.devices[self.overview_obj.online_device_id]\n\n current_estimate = self.overview_obj.dictionary_device_spaceships_priorities[current_device][\n self.overview_obj.calibration_current_estimate_index]\n\n if current_estimate[1] == 0:\n yield self.get_calib_estimate_action(current_estimate, current_device, as_single = True)\n\n if introduce_device(current_device):\n self.devices_introduced.add(current_device)\n for current_estimate_index in range(\n len(self.overview_obj.dictionary_device_spaceships_priorities[current_device])):\n\n current_estimate = self.overview_obj.dictionary_device_spaceships_priorities[current_device][\n current_estimate_index]\n current_estimate_actions_set = self.get_calib_estimate_action(current_estimate, current_device)\n\n self.overview_obj.calibration_current_estimate_index = current_estimate_index\n for current_action in current_estimate_actions_set:\n yield current_action\n\n self.add_device_ship_deadend(current_device, current_estimate[0].name)\n\n else:\n current_estimate = self.overview_obj.dictionary_device_spaceships_priorities[current_device][\n self.overview_obj.calibration_current_estimate_index]\n current_estimate_actions_set = self.get_calib_estimate_action(current_estimate, current_device)\n\n for current_action in current_estimate_actions_set:\n yield current_action\n\n\nclass TargetPriorityManager(PriorityManager):\n def __init__(self, overview_obj):\n super().__init__(overview_obj)\n self.devices_introduced = set()\n\n def reset_introduced_device(self, current_device):\n if current_device in self.devices_introduced:\n self.devices_introduced.remove(current_device)\n\n def priority_target_update(self):\n if print_all_headlines or print_main_headlines:\n print(\">>> Function: priority_target_update\")\n\n current_device = self.overview_obj.devices[self.overview_obj.online_device_id]\n current_estimate_spaceship = self.overview_obj.dictionary_device_spaceships_priorities[current_device][\n self.overview_obj.calibration_current_estimate_index][0]\n current_spaceship_coordinates = current_estimate_spaceship.coordinates\n current_target_coordinates = self.overview_obj.dictionary_device_targets_priorities[current_device][\n self.overview_obj.use_device_current_estimate_index][0]\n\n distances, directions, is_straight_line = UtilsActions.check_straight_line(\n current_spaceship_coordinates, current_target_coordinates)\n\n if is_straight_line:\n found_occlusion, occlusions_distances_set = self.check_for_occlusions_calib(\n current_spaceship_coordinates, current_target_coordinates, (distances, directions))\n\n if not found_occlusion:\n priority_value = 0\n else:\n priority_value = self.get_straight_priority_order(distances, occlusions_distances_set)\n\n self.overview_obj.dictionary_device_targets_priorities[current_device][\n self.overview_obj.use_device_current_estimate_index] = [\n current_target_coordinates, priority_value, (distances, directions)]\n\n else:\n priority_value = sum(distances) - max(distances)\n distances_directions = sorted(zip(distances, directions), key=lambda dist: dist[0])\n\n self.overview_obj.dictionary_device_targets_priorities[current_device][\n self.overview_obj.use_device_current_estimate_index] = [\n current_target_coordinates, priority_value,\n distances_directions]\n\n def priority_target_main(self):\n if print_all_headlines or print_main_headlines:\n print(\">>> Function: priority_target_main\")\n\n current_device = self.overview_obj.devices[self.overview_obj.online_device_id]\n current_estimate = self.overview_obj.dictionary_device_spaceships_priorities[current_device][\n self.overview_obj.calibration_current_estimate_index]\n current_estimate_spaceship = current_estimate[0]\n current_spaceship_coordinates = current_estimate_spaceship.coordinates\n\n all_distances_dict = UtilsGeneral.get_all_distances_dict(self.overview_obj.spaceships,\n self.overview_obj.targets_dict,\n self.overview_obj.calibration_targets_dict,\n specific_spaceship=current_estimate_spaceship)\n\n found_zero_priority = False\n temporary_priorities = []\n for current_target in self.overview_obj.targets_dict:\n if current_device in self.overview_obj.targets_dict[current_target]:\n current_priority = self.get_current_target_priority(current_spaceship_coordinates, current_target,\n all_distances_dict)\n if current_priority[1] == 0:\n temporary_priorities.insert(0, current_priority)\n self.overview_obj.dictionary_device_targets_priorities[current_device] = temporary_priorities\n self.reset_introduced_device(current_device)\n found_zero_priority = True\n break\n else:\n temporary_priorities.append(current_priority)\n\n if not found_zero_priority:\n temporary_priorities.sort(key=lambda priority: priority[1])\n self.overview_obj.dictionary_device_targets_priorities[current_device] = temporary_priorities\n self.reset_introduced_device(current_device)\n\n def get_current_target_priority(self, current_spaceship_coordinates, current_target_coordinates, all_distances_dict):\n if print_all_headlines:\n print(\">>> Function: get_current_target_priority\")\n\n distances, directions, is_straight_line = all_distances_dict[\n (current_spaceship_coordinates, current_target_coordinates)]\n\n if is_straight_line:\n found_occlusion, occlusions_distances_set = self.check_for_occlusions_target(\n current_spaceship_coordinates, current_target_coordinates, (distances, directions))\n\n if not found_occlusion:\n priority_value = 0\n else:\n priority_value = self.get_straight_priority_order(distances, occlusions_distances_set)\n\n return [current_target_coordinates, priority_value, (distances, directions)]\n\n else:\n priority_value = sum(distances) - max(distances)\n distances_directions = sorted(zip(distances, directions), key=lambda dist: dist[0])\n return [current_target_coordinates, priority_value, distances_directions]\n\n def get_target_actions(self):\n if print_all_headlines or print_main_headlines:\n print(\">>> Function: get_target_actions\")\n\n def introduce_device(current_device):\n if current_device in self.devices_introduced:\n return False\n return True\n\n current_device = self.overview_obj.devices[self.overview_obj.online_device_id]\n current_device_calib_estimate = self.overview_obj.dictionary_device_spaceships_priorities[current_device][\n self.overview_obj.calibration_current_estimate_index]\n current_estimate_spaceship = current_device_calib_estimate[0]\n current_estimate_spaceship_name = current_estimate_spaceship.name\n\n current_use_estimate = self.overview_obj.dictionary_device_targets_priorities[current_device][\n self.overview_obj.use_device_current_estimate_index]\n\n current_use_estimate_target_coordinates = current_use_estimate[0]\n\n if current_use_estimate[1] == 0:\n yield ('use', current_estimate_spaceship_name, current_device, current_use_estimate_target_coordinates)\n\n elif introduce_device(current_device):\n self.devices_introduced.add(current_device)\n for current_estimate_index in range(\n len(self.overview_obj.dictionary_device_targets_priorities[current_device])):\n\n current_estimate = self.overview_obj.dictionary_device_targets_priorities[current_device][\n current_estimate_index]\n\n current_estimate_actions_set = UtilsActions.get_move_action_priority_targets(current_estimate, current_estimate_spaceship, self.overview_obj.state,\n self.overview_obj.grid_size)\n\n self.overview_obj.use_device_current_estimate_index = current_estimate_index\n for current_action in current_estimate_actions_set:\n yield current_action\n\n # Make sure 'use_device_current_estimate_index' is initialized to zero somewhere in Calib priority\n self.overview_obj.calib_priority_manager.add_device_ship_deadend(current_device, current_estimate_spaceship_name)\n self.overview_obj.calib_priority_manager.priority_calib_main()\n calibration_actions = self.overview_obj.calib_priority_manager.get_calibration_actions()\n for current_action in calibration_actions:\n yield current_action\n\n\n else:\n current_estimate = self.overview_obj.dictionary_device_targets_priorities[current_device][\n self.overview_obj.use_device_current_estimate_index]\n current_estimate_actions_set = UtilsActions.get_move_action_priority_targets(current_estimate, current_estimate_spaceship, self.overview_obj.state,\n self.overview_obj.grid_size)\n\n for current_action in current_estimate_actions_set:\n yield current_action\n\n\nclass Overview:\n def __init__(self, grid_size):\n self.grid_size = grid_size\n self.state = MyState()\n self.devices = None\n self.spaceships = {} # [spaceship_name] : spaceship_object\n self.targets_dict = {} # [target_coordinate] : target_devices\n self.calibration_targets_dict = {} # [device_name] : target_coordinates\n self.total_number_of_hits = 0\n self.total_number_of_hits_per_device = dict()\n\n self.calib_priority_manager = CalibPriorityManager(self)\n self.target_priority_manager = TargetPriorityManager(self)\n\n self.dictionary_device_spaceships_priorities = dict() # Calib\n self.dictionary_device_targets_priorities = dict() # Use\n self.online_device_id = 0\n self.is_turned_on = 0\n self.is_calibrated = 0\n self.calibration_current_estimate_index = 0\n self.use_device_current_estimate_index = 0\n self.undo_actions_dict = dict() # [ship_name] : undo_action\n\n self.heuristic_value = -1\n\n def calculate_heuristic_value(self):\n if self.goal_test():\n return 0\n\n len_dictionary_device_spaceships_priorities = len(self.dictionary_device_spaceships_priorities)\n\n difference_calibration = len(self.devices) - len_dictionary_device_spaceships_priorities\n difference_target = self.total_number_of_hits - len(self.state.targets_hit)\n\n if difference_calibration == len(self.devices):\n self.heuristic_value = 99999999999\n return self.heuristic_value\n\n\n\n self.heuristic_value = int(\n str(difference_target) + str(difference_calibration) + str(self.use_device_current_estimate_index) + str(\n self.calibration_current_estimate_index))\n\n #print(self.heuristic_value)\n return self.heuristic_value\n\n def set_spaceship_cells(self, ship_names, ship_initial_coordinates, ship_devices):\n for current_spaceship_name in ship_names:\n current_spaceship_coordinates = ship_initial_coordinates[current_spaceship_name]\n current_spaceship_devices = ship_devices[current_spaceship_name]\n created_ship = Spaceship(current_spaceship_coordinates, current_spaceship_name, current_spaceship_devices)\n self.spaceships[current_spaceship_name] = created_ship\n self.state.add_spaceship_occupied_coordinate(current_spaceship_coordinates,\n created_ship.get_spaceship_identifier())\n\n def set_targets_cells(self, dict_coordinate_devices):\n self.targets_dict = dict_coordinate_devices\n [self.state.add_target_occupied_coordinate(current_coordinates) for current_coordinates in\n dict_coordinate_devices]\n self.total_number_of_hits = sum(len(tuple_of_devices) for tuple_of_devices in dict_coordinate_devices.values())\n\n for current_device in self.devices:\n count_instances = 0\n for current_target_devices in self.targets_dict.values():\n if current_device in current_target_devices:\n count_instances += 1\n self.total_number_of_hits_per_device[current_device] = count_instances\n\n def set_calibration_targets_cells(self, dict_devices_coordinates):\n self.calibration_targets_dict = dict_devices_coordinates\n [self.state.add_target_occupied_coordinate(dict_devices_coordinates[current_device]) for current_device in\n dict_devices_coordinates]\n\n def set_devices(self, devices_tuple):\n self.devices = devices_tuple\n\n def goal_test(self):\n if len(self.state.targets_hit) < self.total_number_of_hits:\n return False\n return True\n\n def get_possible_actions(self):\n if print_all_headlines or print_main_headlines:\n print(\">>> Function: get_possible_actions\")\n\n if not self.is_calibrated:\n return self.calib_priority_manager.get_calibration_actions()\n\n else:\n return self.target_priority_manager.get_target_actions()\n\n def check_active_devices(self):\n if print_all_headlines:\n print(\">>> Function: check_active_devices\")\n\n current_device = self.devices[self.online_device_id]\n if current_device not in self.dictionary_device_spaceships_priorities: # Only gets here on first time entrance\n self.calib_priority_manager.priority_calib_main()\n\n if not self.dictionary_device_spaceships_priorities[current_device]:\n if self.total_number_of_hits_per_device[current_device] == 0:\n self.calib_priority_manager.priority_calib_main()\n\n def update_action(self, action):\n if print_all_headlines or print_main_headlines:\n print(\"\\n>>> Function: update_action\")\n\n UtilsUpdate.distribute_packet(action, self.spaceships, self.targets_dict, self.state)\n if action[0] == 'move':\n if self.is_calibrated:\n self.target_priority_manager.priority_target_update()\n else:\n self.calib_priority_manager.priority_calib_update()\n\n elif action[0] == 'turn_on':\n self.is_turned_on = 1\n elif action[0] == 'calibrate':\n self.is_calibrated = 1\n self.target_priority_manager.priority_target_main()\n elif action[0] == 'use':\n action_device = action[2]\n self.total_number_of_hits_per_device[action_device] -= 1\n\n if self.total_number_of_hits_per_device[action_device] == 0:\n self.dictionary_device_spaceships_priorities[action_device] = None\n self.dictionary_device_targets_priorities[action_device] = None\n else:\n self.target_priority_manager.priority_target_main()\n\n self.check_active_devices()\n\n def __hash__(self):\n return hash((self.state))\n\n def __eq__(self, other):\n is_equal = self.state == other.state\n return is_equal\n\n def __lt__(self, node):\n return self.heuristic_value < node.heuristic_value\n\n\nclass SpaceshipProblem(search.Problem):\n \"\"\"This class implements a spaceship problem\"\"\"\n\n def __init__(self, initial):\n \"\"\"Don't forget to set the goal or implement the goal test\n You should change the initial to your own representation\"\"\"\n self.initialize = None\n self.state = self.initialize\n self.unpack_problem(initial)\n search.Problem.__init__(self, initial=self.initialize)\n\n def unpack_problem(self, initial):\n grid_size = initial[0] # Integer\n spaceships_names = initial[1] # Tuple of Strings\n devices = initial[2] # Tuple of Strings\n ships_devices = initial[3] # Dictionary of Tuples\n calibration_targets = initial[4] # Dictionary of Tuples\n targets = initial[5] # Dictionary of Tuples\n initial_positions = initial[6] # Dictionary of Tuples\n\n self.initialize = Overview(grid_size)\n self.initialize.set_spaceship_cells(spaceships_names, initial_positions, ships_devices)\n self.initialize.set_calibration_targets_cells(calibration_targets)\n self.initialize.set_devices(devices)\n self.initialize.set_targets_cells(targets)\n self.initialize.calculate_heuristic_value()\n self.initialize.check_active_devices()\n\n def actions(self, state):\n \"\"\"Return the actions that can be executed in the given\n state. The result would typically be a tuple, but if there are\n many actions, consider yielding them one at a time in an\n iterator, rather than building them all at once.\"\"\"\n all_possible_actions = state.get_possible_actions()\n for current_possible_action in all_possible_actions:\n yield current_possible_action\n\n def result(self, state, action):\n \"\"\"Return the state that results from executing the given\n action in the given state. The action must be one of\n self.actions(state).\"\"\"\n duplicate_state = deepcopy(state)\n duplicate_state.update_action(action)\n return duplicate_state\n\n def goal_test(self, state):\n \"\"\" Given a state, checks if this is the goal state, compares to the created goal state\"\"\"\n return state.goal_test()\n\n def h(self, node):\n \"\"\" This is the heuristic. It gets a node (not a state,\n state can be accessed via node.state)\n and returns a goal distance estimate\"\"\"\n return node.state.calculate_heuristic_value()\n\n\ndef create_spaceship_problem(problem, goal):\n return SpaceshipProblem(problem)\n","sub_path":"Project-1/ex1.py","file_name":"ex1.py","file_ext":"py","file_size_in_byte":54174,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"513490999","text":"\r\n\r\n# NOTE: will need to have requests module installed.\r\n\r\nimport socket, ssl, pprint, time, select, struct, math, configparser, requests, datetime\r\n\r\n\r\nCyberspaceHello = 1357924680\r\nCyberspaceProtocolVersion = 31\r\nClientProtocolOK\t\t= 10000\r\nClientProtocolTooOld\t= 10001\r\nClientProtocolTooNew\t= 10002\r\n\r\nConnectionTypeUpdates\t= 500\r\n\r\nAvatarTransformUpdate\t= 1002\r\n\r\nChatMessageID\t\t\t= 2000\r\n\r\nObjectTransformUpdate\t= 3002\r\nObjectFullUpdate\t\t= 3003\r\n\r\nQueryObjects\t\t\t= 3020\r\nObjectInitialSend\t\t= 3021\r\n\r\nParcelCreated\t\t\t= 3100\r\n\r\n\r\nInfoMessageID\t\t\t= 7001\r\nErrorMessageID\t\t\t= 7002\r\n\r\nLogInMessage\t\t\t= 8000\r\n\r\nTimeSyncMessage\t\t\t= 9000\r\n\r\nWORLD_MATERIAL_SERIALISATION_VERSION = 6\r\nTIMESTAMP_SERIALISATION_VERSION = 1\r\n\r\n\r\ndef writeUInt32ToSocket(socket, x):\r\n\tb = x.to_bytes(4, byteorder='little')\r\n\tsocket.sendall(b)\r\n\r\ndef writeStringLengthFirst(socket, str):\r\n\tb = bytes(str, 'UTF-8')\r\n\twriteUInt32ToSocket(socket, len(b))\r\n\tsocket.sendall(b)\r\n\r\n\r\ndef readNBytesFromSocket(socket, n):\r\n\tremaining = n\r\n\tb = bytearray()\r\n\twhile(remaining > 0):\r\n\t\tchunk = socket.recv(remaining)\r\n\t\tb.extend(chunk)\r\n\t\tremaining -= len(chunk)\r\n\treturn b\r\n\r\ndef readUInt32FromSocket(socket):\r\n\tb = readNBytesFromSocket(socket, 4)\r\n\treturn int.from_bytes(b, byteorder='little', signed=False)\r\n\r\ndef readUInt64FromSocket(socket):\r\n\tb = readNBytesFromSocket(socket, 8)\r\n\treturn int.from_bytes(b, byteorder='little', signed=False)\r\n\r\ndef readUID(socket):\r\n\treturn readUInt64FromSocket(socket)\r\n\r\ndef socketReadable(socket, timeout_s):\r\n\tsocket_list = [socket]\r\n\tread_sockets, write_sockets, error_sockets = select.select(socket_list, [], socket_list, timeout_s)\r\n\treturn len(read_sockets) == 1 or len(error_sockets) == 1\r\n\r\n\r\n\r\nclass BufferIn:\r\n\tdef __init__(self, data_byte_array_):\r\n\t\tself.data_byte_array = data_byte_array_\r\n\t\tself.read_index = 0\r\n\r\n\tdef readInt32(self):\r\n\t\tx = int.from_bytes(self.data_byte_array[self.read_index : self.read_index + 4], byteorder='little', signed=True)\r\n\t\tself.read_index += 4\r\n\t\treturn x\r\n\r\n\tdef readUInt32(self):\r\n\t\tx = int.from_bytes(self.data_byte_array[self.read_index : self.read_index + 4], byteorder='little', signed=False)\r\n\t\tself.read_index += 4\r\n\t\treturn x\r\n\r\n\tdef readUInt64(self):\r\n\t\tx = int.from_bytes(self.data_byte_array[self.read_index : self.read_index + 8], byteorder='little', signed=False)\r\n\t\tself.read_index += 8\r\n\t\treturn x\r\n\r\n\tdef readFloat(self):\r\n\t\tbytes = self.data_byte_array[self.read_index : self.read_index + 4]\r\n\t\tx = struct.unpack(' WORLD_MATERIAL_SERIALISATION_VERSION):\r\n\t\t\traise Exception(\"Unsupported version \" + str(v) + \", expected \" + str(WORLD_MATERIAL_SERIALISATION_VERSION) + \".\")\r\n\r\n\t\tself.colour_rgb = readColour3fFromStream(buffer_in)\r\n\t\tself.colour_texture_url = buffer_in.readStringLengthFirst()\r\n\r\n\t\tself.roughness = readScalarValFromStream(buffer_in)\r\n\t\tself.metallic_fraction = readScalarValFromStream(buffer_in)\r\n\t\tself.opacity = readScalarValFromStream(buffer_in)\r\n\r\n\t\tself.tex_matrix = readMatrix2fFromStream(buffer_in)\r\n\r\n\t\tself.emission_lum_flux = buffer_in.readFloat()\r\n\r\n\t\tself.flags = buffer_in.readUInt32()\r\n\r\n\tdef writeToStream(self, stream):\r\n\t\tstream.writeUInt32(WORLD_MATERIAL_SERIALISATION_VERSION)\r\n\r\n\t\tself.colour_rgb.writeToStream(stream)\r\n\t\tstream.writeStringLengthFirst(self.colour_texture_url)\r\n\r\n\t\tself.roughness.writeToStream(stream)\r\n\t\tself.metallic_fraction.writeToStream(stream)\r\n\t\tself.opacity.writeToStream(stream)\r\n\r\n\t\tself.tex_matrix.writeToStream(stream)\r\n\r\n\t\tstream.writeFloat(self.emission_lum_flux)\r\n\r\n\t\tstream.writeUInt32(self.flags)\r\n\r\n\r\n\r\nclass WorldObject:\r\n\tdef __init__(self):\r\n\t\tself.data = []\r\n\r\n\t\tself.uid = 0 # uint64\r\n\t\tself.object_type = 0 # uint32\r\n\t\tself.model_url = \"\"\r\n\r\n\t\tself.materials = []\r\n\r\n\t\tself.lightmap_url = \"\"\r\n\t\t\r\n\t\tself.script = \"\"\r\n\t\tself.content = \"\"\r\n\t\tself.target_url = \"\"\r\n\t\tself.audio_source_url = \"\"\r\n\r\n\t\tself.audio_volume = 1.0\r\n\r\n\t\tself.pos = Vec3d(0, 0, 0)\r\n\r\n\t\tself.axis = Vec3f(0, 0, 1)\r\n\t\tself.angle = 0.0\r\n\t\tself.scale = Vec3f(1, 1, 1)\r\n\t\t\r\n\t\tself.created_time = TimeStamp(0)\r\n\t\tself.creator_id = 0 # UserID (uint32)\r\n\r\n\t\tself.flags = 0 # uint32\r\n\r\n\t\tself.creator_name = \"\"\r\n\r\n\t\tself.aabb_min = Vec3f(0,0,0)\r\n\t\tself.aabb_max = Vec3f(1,1,1)\r\n\r\n\t\tself.max_model_lod_level = 0 # int32\r\n\r\n\tdef writeToStream(self, stream):\r\n\t\tstream.writeUInt64(self.uid)\r\n\t\tstream.writeUInt32(self.object_type)\r\n\t\tstream.writeStringLengthFirst(self.model_url)\r\n\t\t\r\n\t\t# Write materials\r\n\t\tstream.writeUInt32(len(self.materials))\r\n\t\tfor mat in self.materials:\r\n\t\t\tmat.writeToStream(stream)\r\n\r\n\t\tstream.writeStringLengthFirst(self.lightmap_url)\r\n\r\n\t\tstream.writeStringLengthFirst(self.script)\r\n\t\tstream.writeStringLengthFirst(self.content)\r\n\t\tstream.writeStringLengthFirst(self.target_url)\r\n\t\tstream.writeStringLengthFirst(self.audio_source_url)\r\n\t\tstream.writeFloat(self.audio_volume)\r\n\r\n\t\tself.pos.writeToStream(stream)\r\n\t\tself.axis.writeToStream(stream)\r\n\t\tstream.writeFloat(self.angle)\r\n\t\tself.scale.writeToStream(stream)\r\n\r\n\t\tself.created_time.writeToStream(stream)\r\n\t\tstream.writeUInt32(self.creator_id)\r\n\r\n\t\tstream.writeUInt32(self.flags)\r\n\r\n\t\tstream.writeStringLengthFirst(self.creator_name)\r\n\r\n\t\tprint(\"writing self.aabb_min to stream: \" + str(self.aabb_min.x) + \", \" + str(self.aabb_min.y) + \", \" + str(self.aabb_min.z))\r\n\t\tself.aabb_min.writeToStream(stream)\r\n\t\tprint(\"writing self.aabb_max to stream: \" + str(self.aabb_max.x) + \", \" + str(self.aabb_max.y) + \", \" + str(self.aabb_max.z))\r\n\t\tself.aabb_max.writeToStream(stream)\r\n\r\n\t\tstream.writeInt32(self.max_model_lod_level)\r\n\r\n\tdef readFromStream(self, stream):\r\n\t\tself.uid = stream.readUInt64()\r\n\t\tself.object_type = stream.readUInt32()\r\n\t\tself.model_url = stream.readStringLengthFirst()\r\n\r\n\t\t# Read materials\r\n\t\tnum_mats = stream.readUInt32()\r\n\t\tif (num_mats > 10000):\r\n\t\t\traise Exception(\"Too many mats: \" + str(num_mats))\r\n\t\tself.materials = []\r\n\t\tfor i in range(0, num_mats):\r\n\t\t\tmat = WorldMaterial()\r\n\t\t\tmat.readFromStream(stream)\r\n\t\t\tself.materials.append(mat)\r\n\r\n\t\tself.lightmap_url = buffer_in.readStringLengthFirst()\r\n\r\n\t\tself.script = buffer_in.readStringLengthFirst()\r\n\t\tself.content = buffer_in.readStringLengthFirst()\r\n\t\tself.target_url = buffer_in.readStringLengthFirst()\r\n\r\n\t\tself.audio_source_url = buffer_in.readStringLengthFirst()\r\n\t\tself.audio_volume = buffer_in.readFloat()\r\n\r\n\t\tself.pos = readVec3dFromStream(buffer_in)\r\n\t\tself.axis = readVec3fFromStream(buffer_in)\r\n\t\tself.angle = buffer_in.readFloat()\r\n\t\tself.scale = readVec3fFromStream(buffer_in)\r\n\r\n\t\tself.created_time = readTimeStampFromStream(buffer_in)\r\n\t\tself.creator_id = buffer_in.readUInt32()\r\n\r\n\t\tself.flags = buffer_in.readUInt32()\r\n\r\n\t\tself.creator_name = buffer_in.readStringLengthFirst()\r\n\r\n\t\tself.aabb_min = readVec3fFromStream(buffer_in)\r\n\t\tself.aabb_max = readVec3fFromStream(buffer_in)\r\n\t\t\r\n\t\tprint(\"self.uid: \" + str(self.uid))\r\n\t\tprint(\"self.creator_name: \" + self.creator_name)\r\n\t\tprint(\"Read scale: \" + str(self.scale.x) + \", \" + str(self.scale.y) + \", \" + str(self.scale.z))\r\n\t\tprint(\"Read aabb_min: \" + str(self.aabb_min.x) + \", \" + str(self.aabb_min.y) + \", \" + str(self.aabb_min.z))\r\n\t\tprint(\"Read aabb_max: \" + str(self.aabb_max.x) + \", \" + str(self.aabb_max.y) + \", \" + str(self.aabb_max.z))\r\n\r\n\t\tself.max_model_lod_level = buffer_in.readInt32()\r\n\r\n\t\t# TODO: read compressed voxel data\r\n\r\n\r\n\r\ndef getEthPriceInUSD():\r\n\tresponse = requests.get('https://api.coinbase.com/v2/exchange-rates?currency=USD')\r\n\t#print(response.json())\r\n\r\n\tjson = response.json()\r\n\r\n\teth_per_USD_str = json['data']['rates']['ETH']\r\n\tUSD_per_eth = 1.0 / float(eth_per_USD_str)\r\n\t#print(\"eth_price_usd: \" + str(1.0 / float(eth_per_USD)))\r\n\treturn USD_per_eth\r\n\r\n\r\ndef getEthGasPriceInGWEI(EtherscanAPIKey):\r\n\tresponse = requests.get('https://api.etherscan.io/api?module=gastracker&action=gasoracle&apikey=' + EtherscanAPIKey)\r\n\t#print(response.json())\r\n\r\n\tjson = response.json()\r\n\r\n\tresult = json['result']\r\n\tif result is None:\r\n\t\traise Exception(\"Could not find 'result' JSON node.\")\r\n\tprice = result['SafeGasPrice']\r\n\tif price is None:\r\n\t\traise Exception(\"Could not find 'SafeGasPrice' JSON node.\")\r\n\r\n\treturn float(price)\r\n\r\n#getEthPriceInUSD()\r\n#gas_price_gwei = getEthGasPriceInGWEI()\r\n#print(gas_price_gwei)\r\n#exit(1)\r\n\r\n# Read username and password from disk\r\nconfig = configparser.ConfigParser()\r\nconfig.read('config.txt')\r\nusername = config['credentials']['username']\r\npassword = config['credentials']['password']\r\n\r\nif username is None:\r\n\traise Exception(\"Could not find 'username' in config file.\")\r\nif password is None:\r\n\traise Exception(\"Could not find 'password' in config file.\")\r\n\r\nEtherscanAPIKey = config['APIKeys']['EtherscanAPIKey']\r\nif EtherscanAPIKey is None:\r\n\traise Exception(\"Could not find 'EtherscanAPIKey' in config file.\")\r\n\r\nprint(\"Using username '\" + username + \"'\")\r\n\r\n\r\nplain_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\r\n\r\nconn = ssl.wrap_socket(plain_socket)\r\nconn.connect(('substrata.info', 7600))\r\n#conn.connect(('localhost', 7600))\r\n\r\nprint(\"Connected to server.\")\r\n\r\nwriteUInt32ToSocket(conn, CyberspaceHello)\r\nwriteUInt32ToSocket(conn, CyberspaceProtocolVersion)\r\nwriteUInt32ToSocket(conn, ConnectionTypeUpdates)\r\n\r\nwriteStringLengthFirst(conn, \"\") # Write world name\r\n\r\nhello = readUInt32FromSocket(conn)\r\n\r\nprint('Received hello: ', str(hello))\r\n\r\nprotocol_response = readUInt32FromSocket(conn)\r\nif(protocol_response == ClientProtocolTooOld):\r\n\traise Exception(\"Client protcol is too old\")\r\nelif(protocol_response == ClientProtocolTooNew):\r\n\traise Exception(\"Client protcol is too new\")\r\nelif(protocol_response == ClientProtocolOK):\r\n\tprint(\"ClientProtocolOK\")\r\nelse:\r\n\traise Exception(\"Invalid protocol version response from server: \" + protocol_response);\r\n\r\nclient_avatar_UID = readUID(conn)\r\n\r\n\r\n# Send login message\r\nlogin_buf = BufferOut()\r\nlogin_buf.writeUInt32(LogInMessage)\r\nlogin_buf.writeUInt32(0) # will be updated with length\r\nlogin_buf.writeStringLengthFirst(username)\r\nlogin_buf.writeStringLengthFirst(password)\r\nlogin_buf.updateLengthField()\r\nlogin_buf.writeToSocket(conn)\r\n\r\n\r\ndef sendQueryObjectsMessage(conn):\r\n\tbuffer_out = BufferOut()\r\n\tbuffer_out.writeUInt32(QueryObjects)\r\n\tbuffer_out.writeUInt32(0) # message length - to be updated.\r\n\tr = 4\r\n\tbuffer_out.writeUInt32(2 * (2 * r + 1) * (2 * r + 1)) # Num cells to query\r\n\r\n\tfor x in range(-r, r+1): # (let x = -r; x <= r; ++x)\r\n\t\tfor y in range(-r, r+1): # for (let y = -r; y <= r; ++y) {\r\n\t\t\tbuffer_out.writeInt32(x)\r\n\t\t\tbuffer_out.writeInt32(y)\r\n\t\t\tbuffer_out.writeInt32(0)\r\n\t\t\tbuffer_out.writeInt32(x)\r\n\t\t\tbuffer_out.writeInt32(y)\r\n\t\t\tbuffer_out.writeInt32(-1)\r\n\r\n\tbuffer_out.updateLengthField();\r\n\tbuffer_out.writeToSocket(conn);\r\n\r\n\r\n# Send inital query of objects\r\nsendQueryObjectsMessage(conn)\r\n\r\nlast_send_time = -1000.0\r\ninitial_time = time.monotonic()\r\n\r\nworld_obs = {} # Dict from UID to object\r\n\r\nwhile(1):\r\n\twhile(socketReadable(conn, 0.01)): # timeout_ = 0.1\r\n\t\t# print(\"Socket readable!\")\r\n\t\t# Read and handle message(s)\r\n\t\tmsg_type = readUInt32FromSocket(conn)\r\n\t\tmsg_len = readUInt32FromSocket(conn)\r\n\r\n\t\tif(msg_len < 8):\r\n\t\t\traise Exception(\"Invalid msg len: \" + str(msg_len))\r\n\r\n\t\t# Read rest of message\r\n\t\tmsg_body = readNBytesFromSocket(conn, msg_len - 8) # We have already read 8 bytes of the message, read the rest.\r\n\t\tbuffer_in = BufferIn(msg_body)\r\n\r\n\t\tif(msg_type == TimeSyncMessage):\r\n\t\t\t\r\n\t\t\tglobal_time = buffer_in.readDouble()\r\n\r\n\t\t\tprint(\"Received TimeSyncMessage: global_time: \" + str(global_time))\r\n\t\telif(msg_type == ChatMessageID):\r\n\t\t\r\n\t\t\tname = buffer_in.readStringLengthFirst()\r\n\t\t\tmsg = buffer_in.readStringLengthFirst()\r\n\r\n\t\t\t#print(\"Received ChatMessage: '\" + name + \"': '\" + msg +\"'\")\r\n\t\telif(msg_type == AvatarTransformUpdate):\r\n\t\t\t#print(\"Received AvatarTransformUpdate\")\r\n\t\t\tpass\r\n\t\telif(msg_type == ParcelCreated):\r\n\t\t\t#print(\"Received ParcelCreated\")\r\n\t\t\tpass\r\n\t\telif(msg_type == InfoMessageID):\r\n\t\t\tmsg = buffer_in.readStringLengthFirst()\r\n\t\t\t#print(\"Received InfoMessage: \" + msg)\r\n\t\telif(msg_type == ErrorMessageID):\r\n\t\t\tmsg = buffer_in.readStringLengthFirst()\r\n\t\t\tprint(\"Received ErrorMessage: \" + msg)\r\n\t\telif(msg_type == ObjectInitialSend):\r\n\t\t\tprint(\"Received ObjectInitialSend\")\r\n\t\t\t\r\n\t\t\tworld_ob = WorldObject()\r\n\t\t\tworld_ob.readFromStream(buffer_in)\r\n\t\t\tworld_obs[world_ob.uid] = world_ob\r\n\r\n\t\t\tprint(\"num object: \" + str(len(world_obs)))\r\n\t\telse:\r\n\t\t\tprint(\"Received msg, type: \" + str(msg_type) + \", len: \" + str(msg_len))\r\n\t\t\tpass\r\n\r\n\telse: # Else if socket was not readable:\r\n\t\t\r\n\t\tUPDATE_PERIOD = 10.0 # How often we send an object update message to the server, in seconds\r\n\r\n\t\tif(time.monotonic() - last_send_time > UPDATE_PERIOD):\r\n\r\n\t\t\t# Send a message to the server\r\n\t\t\tlast_send_time = time.monotonic()\r\n\r\n\t\t\t# print(\"Sending message to server...\")\r\n\r\n\t\t\tif False:\r\n\t\t\t\tbuffer_out = BufferOut()\r\n\t\t\t\tbuffer_out.writeUInt32(ObjectTransformUpdate)\r\n\t\t\t\tbuffer_out.writeUInt32(0) # will be updated with length\r\n\t\t\t\tbuffer_out.writeUInt64(152719) # Write object UID\r\n\r\n\t\t\t\tbuffer_out.writeDouble(12.12375831604) # Write object pos x\r\n\t\t\t\tbuffer_out.writeDouble(51.971897125244)# + (time.monotonic() - initial_time)) # Write object pos y\r\n\t\t\t\tbuffer_out.writeDouble(1.063417524099)# + math.sin(time.monotonic()) + 1.0) # Write object pos z\r\n\r\n\t\t\t\tbuffer_out.writeFloat(1.0) # Write object axis x\r\n\t\t\t\tbuffer_out.writeFloat(0) # Write object axis y\r\n\t\t\t\tbuffer_out.writeFloat(math.sin(time.monotonic() - initial_time)) # Write object axis z\r\n\t\t\t\tbuffer_out.writeFloat(time.monotonic() - initial_time + math.pi / 2.0) # Write object angle\r\n\r\n\t\t\t\tbuffer_out.updateLengthField()\r\n\r\n\t\t\t\tbuffer_out.writeToSocket(conn)\r\n\t\t\telse:\r\n\t\t\t\tob = world_obs.get(152870)\r\n\t\t\t\tif ob is not None:\r\n\t\t\t\t\ttry:\r\n\t\t\t\t\t\tprint(\"Writng ob to stream...\")\r\n\r\n\t\t\t\t\t\teth_price_usd = getEthPriceInUSD()\r\n\r\n\t\t\t\t\t\tgas_price_gwei = getEthGasPriceInGWEI(EtherscanAPIKey)\r\n\r\n\t\t\t\t\t\tob.content = \"Eth: \\n\" + str(eth_price_usd) + \" USD\\n\\n\" # Update hypercard contents\r\n\r\n\t\t\t\t\t\tob.content += \"Gas: \\n\" + str(gas_price_gwei) + \" Gwei\\n\\n\"\r\n\r\n\t\t\t\t\t\tob.content += \"Last updated: \\n\" + str(datetime.datetime.utcnow().strftime('%Y-%m-%d %H:%M:%S'))\r\n\r\n\t\t\t\t\t\tprint(\"Set content to \" + ob.content)\r\n\r\n\t\t\t\t\t\tbuffer_out = BufferOut()\r\n\t\t\t\t\t\tbuffer_out.writeUInt32(ObjectFullUpdate)\r\n\t\t\t\t\t\tbuffer_out.writeUInt32(0) # will be updated with length\r\n\t\t\t\t\t\tob.writeToStream(buffer_out)\r\n\t\t\t\t\t\tbuffer_out.updateLengthField()\r\n\t\t\t\t\t\tbuffer_out.writeToSocket(conn)\r\n\t\t\t\t\texcept Exception as err:\r\n\t\t\t\t\t\tprint(\"Caught exception: \" + str(err))\r\n\r\n\t\telse:\r\n\t\t\ttime.sleep(0.01)\r\n\r\n\r\n\r\n\r\n\r\n\r\n# conn.close()\r\n","sub_path":"substrata_bot_demo.py","file_name":"substrata_bot_demo.py","file_ext":"py","file_size_in_byte":18478,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"96322275","text":"import sys\nsys.path.append('../')\n\nimport tensorflow as tf\nimport tensorflow.contrib.slim as slim\nimport tensorflow.contrib.slim.nets as nets\nfrom data_analyse.val_data import *\nimport tqdm\nimport pickle\nimport numpy as np\nfrom data_analyse.preprocess.inception import *\nimport os\n# set GPU\n# os.environ[\"CUDA_VISIBLE_DEVICES\"] = \"0\"\n\nsess_config = tf.ConfigProto()\nsess_config.gpu_options.allow_growth = True\nsess = tf.InteractiveSession(config=sess_config)\n\nimage_R = tf.placeholder(tf.float32, [None, 299, 299, 3])\nimage_A = tf.placeholder(tf.float32, [None, 299, 299, 3])\nlabel = tf.placeholder(tf.float32, [None, 1000])\n\nbatchsize = 50\nattack_step = int(sys.argv[1])\nIMAGENET_VAL = ImageNet_datastream(sess, batchsize, 299)\n\ndef resnet_v2_50(image, reuse=tf.AUTO_REUSE):\n # preprocess = lambda x: preprocess_image(x, 224, 224, is_training=False)\n # preprocessed = tf.map_fn(preprocess, elems=image)\n preprocessed = ((image/255)-0.5)*2\n arg_scope = nets.resnet_v2.resnet_arg_scope(weight_decay=0.0)\n with slim.arg_scope(arg_scope):\n logits, end_point = nets.resnet_v2.resnet_v2_50(preprocessed, 1001, is_training=False, reuse=reuse)\n logits = tf.squeeze(logits)\n logits = logits[:, 1:] # ignore background class\n probs = tf.nn.softmax(logits) # probabilities\n return logits, probs, end_point\n\n\ndef FGSM(x, logits, eps=8):\n total_class_num = tf.shape(logits)[1]\n ori_class = tf.argmax(logits, 1)\n one_hot_class = tf.one_hot(ori_class, total_class_num)\n cross_entropy = tf.losses.softmax_cross_entropy(one_hot_class,\n logits,\n label_smoothing=0.1,\n weights=1.0)\n x_adv = x + eps * tf.sign(tf.gradients(cross_entropy, x)[0])\n x_adv = tf.clip_by_value(x_adv, 0, 255)\n return tf.stop_gradient(x_adv)\n\n\nlogits_R, probs_R, end_point_R = resnet_v2_50(image_R)\nlogits_A, probs_A, end_point_A = resnet_v2_50(image_A)\n\ncorrect_R = tf.equal(tf.argmax(logits_R, 1), (tf.argmax(label, 1)))\naccuracy_R = tf.reduce_mean(tf.cast(correct_R, \"float\"))\n\ncorrect_A = tf.equal(tf.argmax(logits_A, 1), (tf.argmax(label, 1)))\naccuracy_A = tf.reduce_mean(tf.cast(correct_A, \"float\"))\n\n# kernel op\n# minus_op = [tf.cast(tf.subtract(end_point_R[i], end_point_A[i]), tf.float16) for i in end_point_R]\nminus_op = [tf.reduce_mean(tf.abs(tf.subtract(end_point_R[i], end_point_A[i]))) for i in end_point_R]\n\nFGSM_Uint8 = FGSM(image_R, logits_R, 2)\n\n# load target var\nrestore_vars = [\n var for var in tf.global_variables()\n if var.name.startswith('resnet_v2_50/')\n]\npre_train_saver = tf.train.Saver(restore_vars)\npre_train_saver.restore(sess, \"/home/kirin/Python_Code/Adv_Channel_Attention/data_analyse/models/resnet_v2_50/resnet_v2_50.ckpt\")\n\n\nans = []\nT_RACC, TAACC = 0, 0\npbar = tqdm.trange(50000 // batchsize)\nfor i in pbar:\n batch_R, label_R = IMAGENET_VAL.get_test_batch()\n batch_A = batch_R\n for t in range(30):\n batch_A = sess.run(FGSM_Uint8, feed_dict={image_R: batch_A, label: label_R})\n batch_A = np.clip(batch_A, batch_R - attack_step, batch_R + attack_step)\n batch_A = np.clip(batch_A, 0, 255)\n R_ACC, A_ACC = sess.run([accuracy_R, accuracy_A], feed_dict={image_R: batch_R, image_A: batch_A, label: label_R})\n tmp = sess.run(minus_op, feed_dict={image_R: batch_R, image_A: batch_A})\n if i == 0:\n ans = tmp\n else:\n for l in range(len(ans)):\n ans[l] = np.vstack((ans[l], tmp[l]))\n T_RACC += R_ACC\n TAACC += A_ACC\n pbar.set_description(\"step:{}, R_ACC:{:.4f}, A_ACC:{:.4f}\".format(attack_step, T_RACC / (i + 1), TAACC / (i + 1)))\n\nf = open(\"resnet50.txt\", 'a')\nprint(\"attack_step:{}\".format(attack_step))\nf.writelines(\"attack_step:{}\\n\".format(attack_step))\nfor i in range(len(ans)):\n shape = np.shape(ans[i])\n size = 1\n for channel in shape:\n size = size * channel\n avg = np.sum(np.abs(ans[i]) / size)\n print(\"{} {}\".format(list(end_point_R)[i], avg))\n f.writelines(\"{} {}\\n\".format(list(end_point_R)[i], avg))\n\nindex = 50000 // batchsize\nprint(\"R_ACC:{:.4f}, A_ACC:{:.4f}\".format(T_RACC / index, TAACC / index))\nf.writelines(\"R_ACC {:.4f}\\nA_ACC {:.4f}\\n\\n\".format(T_RACC / index, TAACC / index))\nf.close()\n","sub_path":"data_analyse/ResNet50_data_analyse.py","file_name":"ResNet50_data_analyse.py","file_ext":"py","file_size_in_byte":4315,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"321410478","text":"from django.http import HttpResponseRedirect\nfrom django.shortcuts import render, get_object_or_404\nfrom .models import Album\nfrom .models import Artist\nfrom .forms import ArtistForm\nfrom .forms import AlbumForm\n\n# Create your views here.\ndef album_list(request):\n albums = Album.objects.all()\n return render(request, 'index.html', {'albums': albums})\n\n\ndef artist_list(request):\n artists = Artist.objects.all()\n return render(request, 'artists.html', {'artists': artists})\n\n\ndef album_detail(request, pk):\n album = get_object_or_404(Album, pk=pk)\n return render(request, 'album_detail.html', {'album': album})\n\n\ndef artist_detail(request, pk):\n artist = get_object_or_404(Artist, pk=pk)\n return render(request, 'artist_detail.html', {'artist': artist})\n\n\ndef add_artist(request):\n if request.method == 'POST':\n form = ArtistForm(request.POST)\n if form.is_valid():\n form.save()\n return HttpResponseRedirect('/')\n\n else:\n form = ArtistForm()\n\n return render(request, 'add_artist.html', {'form': form})\n\n\ndef add_album(request):\n if request.method == 'POST':\n form = AlbumForm(request.POST, request.FILE)\n if form.is_valid():\n form.save()\n return HttpResponseRedirect('/')\n\n else:\n form = AlbumForm()\n\n return render(request, 'add_album.html', {'form': form})\n\n\ndef edit_artist(request, pk):\n artist = get_object_or_404(Artist, pk=pk)\n if request.method == 'POST':\n form = ArtistForm(request.POST, instance=artist)\n if form.is_valid():\n form.save()\n return HttpResponseRedirect('/')\n\n else:\n form = ArtistForm(instance=artist)\n return render(request, 'edit_artist.html', {'form': form, 'artist':artist })\n\n\ndef delete_artist(request, pk):\n artist = get_object_or_404(Artist, pk=pk)\n artist.delete()\n return HttpResponseRedirect('/')\n\n\ndef edit_album(request, pk):\n album = get_object_or_404(Album, pk=pk)\n if request.method == 'POST':\n form = AlbumForm(request.POST, instance=album)\n if form.is_valid():\n form.save()\n return HttpResponseRedirect('/')\n\n else:\n form = AlbumForm(instance=album)\n return render(request, 'edit_album.html', {'form': form, 'album':album })\n\n\ndef delete_album(request, pk):\n album = get_object_or_404(Album, pk=pk)\n album.delete()\n return HttpResponseRedirect('/')","sub_path":"core/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":2445,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"295794907","text":"\"\"\"\nThis is going to be the data access object for the skilled nursing facility\napplication. We are going to be using sqllite3 so we do not have to add the\ndependency of running a Postgres server in its own process. For a production\napplication we would put this in Postres.\n\"\"\"\nimport sqlite3\nimport csv\n\nclass NursingFacilityDAO(object):\n def __init__(self):\n self.conn = sqlite3.connect('radial.db')\n self.conn.row_factory = sqlite3.Row\n self.init_all_tables()\n self.load_deficiencies_data()\n self.load_penalty_data()\n self.load_zip_code_data()\n self.load_privider_info_data()\n self.load_count_tables()\n\n def init_all_tables(self):\n cur = self.conn.cursor()\n sql = \"\"\"\n DROP TABLE IF EXISTS deficiencies;\n CREATE TABLE deficiencies (\n deficiency_id INTEGER PRIMARY KEY ASC,\n provnum INTEGER,\n complaint TEXT,\n filedate NUMERIC\n );\n\n DROP TABLE IF EXISTS penalties;\n CREATE TABLE penalties (\n penalty_id INTEGER PRIMARY KEY ASC,\n provnum INTEGER,\n fine_amt INTEGER,\n filedate NUMERIC\n );\n\n DROP TABLE IF EXISTS zip_code_centroids;\n CREATE TABLE zip_code_centroids (\n zip_code_id INTEGER PRIMARY KEY ASC,\n zip_code INTEGER,\n lat NUMERIC,\n lng NUMERIC\n );\n\n DROP TABLE IF EXISTS privider_info;\n CREATE TABLE privider_info (\n privider_info_id INTEGER PRIMARY KEY ,\n provnum INTEGER,\n provname TEXT,\n address TEXT,\n city TEXT,\n state TEXT,\n zip INTEGER,\n phone INTEGER,\n overall_rating INTEGER\n );\n \"\"\"\n cur.executescript(sql)\n\n def load_deficiencies_data(self):\n cur = self.conn.cursor()\n with open('./data/Deficiencies_Download.csv') as file_object:\n file_as_dict = csv.DictReader(file_object)\n row_data = [(row['provnum'], row['complaint'], row['filedate']) for row in file_as_dict]\n cur.executemany(\"INSERT INTO deficiencies (provnum, complaint, filedate) VALUES (?, ?, ?);\", row_data)\n self.conn.commit()\n\n def load_penalty_data(self):\n cur = self.conn.cursor()\n with open('./data/Penalties_Download.csv') as file_object:\n file_as_dict = csv.DictReader(file_object)\n row_data = [(row['provnum'], row['fine_amt'], row['filedate']) for row in file_as_dict]\n cur.executemany(\"INSERT INTO penalties (provnum, fine_amt, filedate) VALUES (?, ?, ?);\", row_data)\n self.conn.commit()\n\n def load_zip_code_data(self):\n cur = self.conn.cursor()\n with open('./data/zip_code_centroids.csv') as file_object:\n file_as_dict = csv.DictReader(file_object)\n row_data = [(row['zip_code'], row['lat'], row['lng']) for row in file_as_dict]\n cur.executemany(\"INSERT INTO zip_code_centroids (zip_code, lat, lng) VALUES (?, ?, ?);\", row_data)\n self.conn.commit()\n\n def load_privider_info_data(self):\n cur = self.conn.cursor()\n with open('./data/ProviderInfo_Download.csv') as file_object:\n file_as_dict = csv.DictReader(file_object)\n row_data = []\n for row in file_as_dict:\n row_tuple = (row['provnum'], row['PROVNAME'], row['ADDRESS'], row['CITY'], row['STATE'],\n row['ZIP'], row['PHONE'], row['overall_rating'])\n row_data.append(row_tuple)\n cur.executemany(\"\"\"\n INSERT INTO privider_info (\n provnum,\n provname,\n address,\n city,\n state,\n zip,\n phone,\n overall_rating\n ) VALUES (?, ?, ?, ?, ?, ?, ?, ?);\n \"\"\",\n row_data)\n self.conn.commit()\n\n def load_count_tables(self):\n sql = \"\"\"\n -- First thing we will do is get the counts of penalties and\n -- deficiencies by provider.\n DROP TABLE IF EXISTS deficiency_counts_temp;\n CREATE TEMPORARY TABLE deficiency_counts_temp (\n deficiency_count_id INTEGER PRIMARY KEY,\n provnum INTEGER,\n num_of_deficiencies INTEGER\n );\n\n INSERT INTO deficiency_counts_temp (provnum, num_of_deficiencies)\n SELECT provnum, COUNT(1)\n FROM deficiencies\n GROUP BY provnum;\n\n DROP TABLE IF EXISTS penalty_counts_temp;\n CREATE TEMPORARY TABLE penalty_counts_temp (\n penalty_count_id INTEGER PRIMARY KEY,\n provnum INTEGER,\n num_of_penalties INTEGER\n );\n\n INSERT INTO penalty_counts_temp (provnum, num_of_penalties)\n SELECT provnum, COUNT(1)\n FROM penalties\n GROUP BY provnum;\n \"\"\"\n cur = self.conn.cursor()\n cur.executescript(sql)\n\n def load_lon_lat_from_zip(self, zip_code):\n sql = \"\"\"\n SELECT lat, lng\n FROM zip_code_centroids\n WHERE zip_code = ?\n \"\"\"\n cur = self.conn.cursor()\n cur.execute(sql, (zip_code,))\n result = cur.fetchone()\n return result\n\n def get_raw_data(self, search_params):\n \"\"\"\n ***\n IF this was production I would use the earth_distance function in Postgres\n as can be seen here http://www.postgresql.org/docs/9.1/static/earthdistance.html.\n I am just approximating here to keep it simple.\n ****\n \"\"\"\n\n inputs = []\n\n sql = \"\"\"\n -- We are doing left joins here because there could be records\n -- in the privider info table with no deficiencies or penalties\n -- and we do not want to exclude them from the results. We are doing\n -- an inner join on the zip code table because we need this info to\n -- do the search, if there is not zip data for it don't even include.\n SELECT provname, address, city, state, zip, phone, overall_rating,\n (100 / (ABS(? - z.lat) + ABS(? - z.lng) / overall_rating)) as score,\n p.num_of_penalties, d.num_of_deficiencies, z.lat, z.lng\n FROM privider_info prov\n LEFT JOIN deficiency_counts_temp d ON prov.provnum = d.provnum\n LEFT JOIN penalty_counts_temp p on prov.provnum = p.provnum\n INNER JOIN zip_code_centroids z on prov.zip = z.zip_code\n WHERE prov.overall_rating <> ''\n \"\"\"\n # Now we need to push the lat and long into the inputs array.\n inputs.append(search_params['lat'])\n inputs.append(search_params['lng'])\n\n if search_params['min_overall_rating']:\n sql += \" AND overall_rating >= ? \"\n inputs.append(search_params['min_overall_rating'])\n if search_params['max_num_deficiencies']:\n sql += \" AND d.num_of_deficiencies <= ? \"\n inputs.append(search_params['max_num_deficiencies'])\n if search_params['max_num_penalties']:\n sql += \" AND p.num_of_penalties <= ? \"\n inputs.append(search_params['max_num_penalties'])\n\n # As I said in the docstring, I would use the haversine func or the earth_distance function\n # if this was in production, this approximation is just to keep it simple.\n sql += \" ORDER BY score DESC\"\n\n if search_params['max_num_facilities']:\n sql += \" LIMIT ? \"\n inputs.append(search_params['max_num_facilities'])\n\n cur = self.conn.cursor()\n input_tuple = tuple(inputs)\n cur.execute(sql, input_tuple)\n return_data = cur.fetchall()\n self.conn.close()\n return return_data","sub_path":"nursing_facility/nursing_facility_dao.py","file_name":"nursing_facility_dao.py","file_ext":"py","file_size_in_byte":8046,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"278455340","text":"import networkx as nx\nimport matplotlib.pyplot as plt\n\n\ndef draw_network(network, **kwargs):\n graph = network.to_digraph()\n nx.spring_layout(graph, **kwargs)\n nx.draw(\n graph,\n labels=nx.get_node_attributes(graph, \"name\"),\n width=3,\n edge_color=\"black\",\n node_size=1500,\n node_color=\"gray\",\n )\n plt.show()\n","sub_path":"apogee/utils/draw.py","file_name":"draw.py","file_ext":"py","file_size_in_byte":363,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"358172098","text":"import argparse\nimport os\n\nfrom pathos.pools import ProcessPool\n\nimport Generator, ExperimentSetups\n\n# Parameters\nimport fileutil\nfrom fileutil import FileUtil\n\nparser = argparse.ArgumentParser(description=\"Generate Quantifiers\")\nparser.add_argument('setup', help='Path to the setup json file.')\nparser.add_argument('max_quantifier_length', type=int)\nparser.add_argument('model_size', type=int)\nparser.add_argument('--dest_dir', default='results')\nparser.add_argument('--processes', default=4, type=int)\n\nargs = parser.parse_args()\n\nprocesses = args.processes\nsetup = ExperimentSetups.parse(args.setup)\nmax_quantifier_length = args.max_quantifier_length\nmodel_size = args.model_size\n\nfile_util = FileUtil(fileutil.base_dir(args.dest_dir, setup.name, max_quantifier_length, model_size))\n\n\nuniverse = setup.generate_models(model_size)\n\nfolderName = \"{0}/{1}_length={2}_size={3}\".format(args.dest_dir,setup.name,max_quantifier_length,model_size)\nos.makedirs(\"{0}\".format(folderName), exist_ok=True)\n\nprocesspool = ProcessPool(nodes=processes)\nexpression_generator = Generator.ExpressionGenerator(setup, model_size, universe, processpool)\n(generated_expressions_dict, expressions_by_meaning) = \\\n expression_generator.generate_all_expressions(max_quantifier_length)\n\nprint(\"{0} expressions!\".format(len(expressions_by_meaning[bool].values())))\n\nfile_util.dump_dill(expressions_by_meaning[bool], 'generated_expressions.dill')\nfile_util.dump_dill(list(expressions_by_meaning[bool].values()), 'expressions.dill')\nfile_util.dump_dill(list(expressions_by_meaning[bool].keys()), 'meanings.dill')\n\nprocesspool.close()\nprocesspool.join()\n\nprint('Expression generation finished.')\n","sub_path":"Code/IndividualQuantifiers/generate.py","file_name":"generate.py","file_ext":"py","file_size_in_byte":1671,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"3941895","text":"import os\nimport sys\n\nfrom polyglot.detect import Detector\n\n\ndef DetectLyricLanguage(lyric):\n langs = Detector(lyric)\n lang = langs.languages[0]\n print(lang.code)\n\n\nif __name__ == '__main__':\n lyricPath = sys.argv[1]\n lyric = ''\n if os.path.exists(lyricPath):\n with open(lyricPath, encoding='utf-8', mode='r') as f:\n for line in f.readlines():\n lyric += line\n\n DetectLyricLanguage(lyric)\n else:\n print(\"LyricFile not exists!\")\n exit(1)\n","sub_path":"LangDetector.py","file_name":"LangDetector.py","file_ext":"py","file_size_in_byte":511,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"255397376","text":"import collections\nimport inspect\nimport itertools\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport os\nimport time\n\nimport Assignment2Support as utils\nimport EvaluationsStub as es\nimport LogisticRegressionModel as lgm\nimport features_by_frequency as fbf\nimport features_by_mi as fbm\n\n\n# File/Folder path\nkDataPath = os.path.join(os.path.dirname(\n os.path.dirname(__file__)), r\"Data/SMSSpamCollection\")\n\nreport_path = os.path.join(os.path.dirname(\n os.path.dirname(__file__)), r\"Report\")\n\n\ndef get_features_by_frequency(xTrainRaw, xTestRaw, N):\n \"\"\"\n Returns: a tuple of (xTrain, xTest, a list of words(features))\n \"\"\"\n features = fbf.extract_features_by_frequency(xTrainRaw, N)\n features = [x[0] for x in features]\n xTrain = utils.FeaturizeTrainingByWords(xTrainRaw, features)\n xTest = utils.FeaturizeTrainingByWords(xTestRaw, features)\n\n return xTrain, xTest, features\n\n\ndef get_features_mi(xTrainRaw, yTrainRaw, xTestRaw, N):\n \"\"\"\n Returns: a tuple of (xTrain, xTest, a list of words(features))\n \"\"\"\n features, _ = fbm.extract_features_by_mi(xTrainRaw, yTrainRaw, N)\n features = [x[0] for x in features]\n xTrain = utils.FeaturizeTrainingByWords(xTrainRaw, features)\n xTest = utils.FeaturizeTrainingByWords(xTestRaw, features)\n\n return xTrain, xTest, features\n\n\ndef find_categrize_mistakes(xTrain, xTest, yTrain, yTest, features, iter_step, resolution, intial_w0, step, max_iters, top=20):\n\n model = utils.logistic_regression_model_by_features(\n xTrain, yTrain, features, iter_step, resolution, initial_w0, step, max_iters)\n\n xTest = [[1] + x for x in xTest]\n # predict using validation dataset\n yTestPredicted_prob = model.predict_probabilities(xTest)\n yTestPredicted = model.predict(xTest)\n\n fn = []\n fp = []\n\n for i, (t, p) in enumerate(zip(yTest, yTestPredicted)):\n prob = yTestPredicted_prob[i]\n if (t, p) == (1, 0): # false negative\n fn.append((prob, i))\n elif (t, p) == (0, 1): # false positive\n fp.append((prob, i))\n\n # the true answer was 1, but the model gives very low probabilities\n sorted_fn = sorted(fn)\n # the true answer was 0, but gives very high probabilities.\n sorted_fp = sorted(fp, reverse=True)\n print('*' * 80)\n print('Total {} false netagives.'.format(len(sorted_fn)))\n print('Total {} false positives.'.format(len(sorted_fp)))\n print('False Netagives: {}'.format(sorted_fn))\n print('False Positives: {}'.format(sorted_fp))\n print(es.ConfusionMatrix(yTest, yTestPredicted))\n print('*' * 80)\n return sorted_fn[:top], sorted_fp[:top]\n\n\ndef generate_mistakes_table(mistakes, title, header, xTestRaw, fname, w=30):\n\n top_mistakes = title\n top_mistakes += '\\n'\n top_mistakes += '\\n {}'.format(header)\n top_mistakes += '\\n |-|-|'\n for prob, i in mistakes:\n top_mistakes += '\\n |{}| {}|'.format(\n '{}'.format(prob).center(w), xTestRaw[i].strip())\n\n with open(fname, 'w') as f:\n f.write(top_mistakes)\n print(\"Created {}\".format(fname))\n\n\nif __name__ == '__main__':\n # Loading data\n (xRaw, yRaw) = utils.LoadRawData(kDataPath)\n (xTrainRaw, yTrainRaw, xTestRaw, yTestRaw) = utils.TrainTestSplit(xRaw,\n yRaw)\n\n N = 10\n max_iters = 50000\n iter_step = 1000\n resolution = int(max_iters / iter_step)\n initial_w0 = 0.0\n step = 0.01\n top = 20\n\n ############################################################################\n # by frequency\n xTrain, xTest, features = get_features_by_frequency(xTrainRaw, xTestRaw, N)\n yTrain = yTrainRaw\n yTest = yTestRaw\n _, sorted_fp = find_categrize_mistakes(\n xTrain, xTest, yTrain, yTest, features, iter_step, resolution, initial_w0, step, top)\n\n w = 30\n header = '| Probabilities | Test Raw |'\n title = '* False Positive - the true answer was 0, but gives very high probabilities'\n fname = os.path.join(report_path, 'category_mistake_false_positives.md')\n generate_mistakes_table(sorted_fp, title, header, xTestRaw, fname)\n\n ############################################################################\n # by mutual information\n xTrain, xTest, features = get_features_mi(\n xTrainRaw, yTrainRaw, xTestRaw, N)\n yTrain = yTrainRaw\n yTest = yTestRaw\n sorted_fn, _ = find_categrize_mistakes(\n xTrain, xTest, yTrain, yTest, features, iter_step, resolution, initial_w0, step, top)\n\n w = 30\n title = '* False Negatives - the true answer was 1, but the model gives very low probabilities'\n fname = os.path.join(report_path, 'category_mistake_false_negatives.md')\n generate_mistakes_table(sorted_fn, title, header, xTestRaw, fname)\n ############################################################################\n","sub_path":"Assignment5/Code/assignments/previous/hw2/category_mistakes.py","file_name":"category_mistakes.py","file_ext":"py","file_size_in_byte":4854,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"520530519","text":"import urllib.request # for getting the page from the internet\nimport re # for using regex \nimport sys # needed for sys.exit(), which ends execution\nimport os # for making a directory\nimport html # for converting escape characters into ascii (e.g. turns & into &)\n\n# recipe to search and ingredient to search for\n# recipeNum = 3\ningredient = 'salt'\n\n# folders where pages are saved\nfolderName_Byte = \"RecipeScrapes_Byte\"\nfolderName_String = \"RecipeScrapes_String\"\nfolderName_Results = \"ProcessResults\"\n\n \nprint('\\n\\n')\nprint('-----------------------------------------------------------')\nprint('\\n')\n\n\n\n\n# Find the path that the script is running from\nlocalFolder = os.path.dirname(__file__)\n\n# Create the complete path to folders (so they are placed in the directory where the script is running from)\nfilePath_Byte = localFolder + \"\\\\\" + folderName_Byte\nfilePath_String = localFolder + \"\\\\\" + folderName_String\nfilePath_Results = localFolder + \"\\\\\" + folderName_Results\n\n# Create the results folder if it doesn't exit already\nif not os.path.isdir(filePath_Results):\n os.makedirs(filePath_Results)\n\nndx=3\n\nfor ndx in range(30):\n\n #read the contents of a file\n print(\"Reading text file\", ndx)\n \n \n #get the file name\n recipeFile = \"recipe\" + str(ndx) + \".txt\"\n\n with open(filePath_String + \"\\\\\" + recipeFile, 'rt', encoding='utf-8') as inf: \n url = inf.readline() # get the url from the first line of the file\n filetext = inf.read() # get the rest of the file (doesn't include the first line because that was already read)\n \n\n filetext = html.unescape(filetext)\n\n \n # Find the title of the recipe\n\n #search for the title. It will probably be the first thing that is found between tags\n pattern = re.compile(\"(?<=\\)(.*?)(?=\\<\\/title\\>)\",re.DOTALL)\n match = re.search(pattern, filetext)\n \n # print(match)\n # print(type(match))\n try:\n # clean up the found title by removing junk\n title = re.sub(r'[^\\x00-\\x7F]+',' ', match.group(0)) #remove non-ASCII characters by replacing them with white space\n title = re.sub(r'[\\t\\n\\r\\f\\v]','', title) #remove most \"blank\" characters that aren't spaces\n title.strip() #remove leading or trailing white space from title\n except AttributeError:\n title = \"Title not found.\"\n \n print(title)\n #\n # find an ingredient list\n #\n\n # find all html tags that include contain \"ingredient (keep the open quote to avoid finding extra stuff, but leave off the closing quote so that it still works if labels are plural or have suffixes)\n # or maybe leave off the open quote to help find \"p-ingredient\" tags or \"RecipeIngredient\" tags\n ingpattern1 = re.compile(\"<[^>]*?ingredient[^>]*>.*?<\",re.IGNORECASE)\n ingList = re.findall(ingpattern1, filetext)\n\n # for each item that was found, strip out the html tags and leave behind what was in between them\n for ndx, member in enumerate(ingList):\n # print('\\n')\n # print(ndx)\n # print(ingList[ndx])\n ingpattern2 = re.compile(\"(?<=>)(.*?)(?=<)\") # pattern that finds everything between '>' and '<' (can also end with a new line)\n ingList[ndx] = re.search(ingpattern2,ingList[ndx]).group(0) # sets the element in ingList to the version with html stripped out\n ingList[ndx].strip()\n # print(ingList[ndx])\n\n # remove any blank elements from ingList\n ingList = [x for x in ingList if x != '']\n\n \n \n # open a file. Write the url, title, and ingredient list to that file\n with open(filePath_Results + \"\\\\\" + recipeFile, 'w+t', encoding='utf-8') as fid: \n fid.write(url)\n fid.write('\\n\\n\\n')\n fid.write(title)\n fid.write('\\n------------------------------------------------\\n')\n \n for ndx, member in enumerate(ingList):\n fid.write('\\n'+ingList[ndx])\n\n \n # find ingredients in the ingredient list\n ingResult = [x for x in ingList if ingredient in x]\n \n print(ingResult) \n \nprint(\"Done.\")\n\nprint('\\n\\n')\nprint('-----------------------------------------------------------')\nprint('\\n')\n\n\n\n\n\n","sub_path":"recipe_reader_loop_regex.py","file_name":"recipe_reader_loop_regex.py","file_ext":"py","file_size_in_byte":4216,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"349974388","text":"class Solution:\n def findTheDifference(self, s: str, t: str) -> str:\n char = {}\n\n for letter in t:\n if letter not in char:\n char[letter] = 1\n else:\n char[letter] += 1\n \n for letter in s:\n char[letter] -= 1\n if char[letter] == 0:\n del char[letter]\n\n return list(char.keys())[0]\n\n # char = list(t)\n\n # for letter in s:\n # char.remove(letter)\n\n # return char[0]\n \n\ndef main():\n sol = Solution()\n print(sol.findTheDifference(\"abcd\", \"abcde\"))\n print(sol.findTheDifference(\"\", \"y\"))\n print(sol.findTheDifference(\"a\", \"aa\"))\n\nif __name__ == '__main__':\n main()","sub_path":"easy/389.py","file_name":"389.py","file_ext":"py","file_size_in_byte":739,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"40669660","text":"# 给定不同面额的硬币 coins 和一个总金额 amount。编写一个函数来计算可以凑成总金额所需的最少的硬币个数。如果没有任何一种硬币组合能组成总金额,返回\n# -1。 \n# \n# 你可以认为每种硬币的数量是无限的。 \n# \n# \n# \n# 示例 1: \n# \n# \n# 输入:coins = [1, 2, 5], amount = 11\n# 输出:3 \n# 解释:11 = 5 + 5 + 1 \n# \n# 示例 2: \n# \n# \n# 输入:coins = [2], amount = 3\n# 输出:-1 \n# \n# 示例 3: \n# \n# \n# 输入:coins = [1], amount = 0\n# 输出:0\n# \n# \n# 示例 4: \n# \n# \n# 输入:coins = [1], amount = 1\n# 输出:1\n# \n# \n# 示例 5: \n# \n# \n# 输入:coins = [1], amount = 2\n# 输出:2\n# \n# \n# \n# \n# 提示: \n# \n# \n# 1 <= coins.length <= 12 \n# 1 <= coins[i] <= 231 - 1 \n# 0 <= amount <= 104 \n# \n# Related Topics 动态规划 \n# 👍 1074 👎 0\n\n\n# leetcode submit region begin(Prohibit modification and deletion)\nclass Solution:\n def coinChange(self, coins: List[int], amount: int) -> int:\n dp = [amount + 1] * (amount + 1)\n dp[0] = 0\n\n for target in range(1, amount + 1):\n for coin in coins:\n # 剪枝1. 要凑的面值小于最小coin\n # 剪枝2. 要拿一枚1元+dp[10]凑11时,dp[10]原本凑不出来(即dp[10] = amount + 1)\n if target - coin >= 0 and dp[target - coin] != amount + 1:\n # min里面有dp[target],那么在coins的循环里面,会选出最小的dp[target]最后保存起来然后跳出coins循环\n dp[target] = min(dp[target], dp[target - coin] + 1)\n\n return dp[-1] if dp[-1] != amount+1 else -1\n# leetcode submit region end(Prohibit modification and deletion)\n","sub_path":"leetcode/editor/cn/[322]零钱兑换.py","file_name":"[322]零钱兑换.py","file_ext":"py","file_size_in_byte":1742,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"243362299","text":"import numpy as np\nimport csv\n\n# file_1\nf = open(\"Back_10N_1.csv\",'r')\nst = f.read()\nmat = [[ele for ele in r.split(',') ] for r in [s for i,s in enumerate(st.split('\\n')) if i>=60 and i<52632]]\nmat_first = np.array(mat) \n\nfor x in range(14):\n mat_first = np.delete(mat_first,x+0,1)\n mat_first = np.delete(mat_first,np.s_[x+1:x+7],1)\n\n# file_2\nf = open(\"Back_10N_2.csv\",'r')\nst = f.read()\nmat = [[ele for ele in r.split(',') ] for r in [s for i,s in enumerate(st.split('\\n')) if i>=60 and i<54426]]\nmat_second = np.array(mat) \n\nfor x in range(14):\n mat_second = np.delete(mat_second,x+0,1)\n mat_second = np.delete(mat_second,np.s_[x+1:x+7],1)\n\n# file_3\nf = open(\"Back_10N_3.csv\",'r')\nst = f.read()\nmat = [[ele for ele in r.split(',') ] for r in [s for i,s in enumerate(st.split('\\n')) if i>=60 and i<50504]]\nmat_third = np.array(mat) \n\nfor x in range(14):\n mat_third = np.delete(mat_third,x+0,1)\n mat_third = np.delete(mat_third,np.s_[x+1:x+7],1)\n\n# file_4\nf = open(\"Back_10N_4.csv\",'r')\nst = f.read()\nmat = [[ele for ele in r.split(',') ] for r in [s for i,s in enumerate(st.split('\\n')) if i>=60 and i<55440]]\nmat_fourth = np.array(mat) \n\nfor x in range(14):\n mat_fourth = np.delete(mat_fourth,x+0,1)\n mat_fourth = np.delete(mat_fourth,np.s_[x+1:x+7],1)\n\ncombined = np.concatenate((mat_first, mat_second, mat_third, mat_fourth)) \n \nwith open('Back_10N.csv', 'w') as file:\n writer = csv.writer(file, lineterminator='\\n')\n writer.writerows(combined)\n","sub_path":"Process_raw.py","file_name":"Process_raw.py","file_ext":"py","file_size_in_byte":1490,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"633717419","text":"# -*- coding: utf-8 -*-\n\nfrom odoo import _, api, fields, models\nfrom odoo.tools.safe_eval import safe_eval\nfrom odoo.exceptions import UserError\n\nclass ProductLabelLayout(models.TransientModel):\n _inherit = 'product.label.layout'\n print_format = fields.Selection(selection_add=[('dymo4_6', '4 X 6 Dymo'),\n ('2x7xprice', '2 x 7 with price')\n ], ondelete={'dymo4_6': 'cascade'}, default='dymo4_6')\n\n\n def _prepare_report_data(self):\n if self.custom_quantity <= 0:\n raise UserError(_('You need to set a positive quantity.'))\n # Get layout grid\n if self.print_format == 'dymo4_6':\n xml_id = 'max_fire_dymo_report.report_product_template_label_dymo_4_6'\n elif self.print_format == 'dymo':\n xml_id = 'product.report_product_template_label_dymo'\n elif 'x' in self.print_format:\n xml_id = 'product.report_product_template_label'\n else:\n xml_id = ''\n\n active_model = ''\n if self.product_tmpl_ids:\n products = self.product_tmpl_ids.ids\n active_model = 'product.template'\n elif self.product_ids:\n products = self.product_ids.ids\n active_model = 'product.product'\n\n # Build data to pass to the report\n data = {\n 'active_model': active_model,\n 'quantity_by_product': {p: self.custom_quantity for p in products},\n 'layout_wizard': self.id,\n 'price_included': 'xprice' in self.print_format,\n }\n return xml_id, data\n\n","sub_path":"max_fire_dymo_report/wizard/product_label_layout.py","file_name":"product_label_layout.py","file_ext":"py","file_size_in_byte":1645,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"158355658","text":"import torch\nfrom torch.utils.data import Dataset, DataLoader\nfrom sklearn.metrics import accuracy_score, confusion_matrix\nfrom torch.autograd import Variable\nfrom torchvision.datasets import FashionMNIST\nfrom torchvision import transforms\nfrom sklearn.model_selection import train_test_split\nimport time\nimport scikitplot as skplt\nimport matplotlib.pyplot as plt\nimport os\n\n\ndef get_train_valid_data():\n trans_img = transforms.Compose([transforms.ToTensor()])\n data = FashionMNIST(\"./data\", train=True,\n transform=trans_img, download=True)\n\n X = data.train_data\n X = X.reshape(60000, 1, 28, 28) # Reshaping for use of Convolution\n X = X.type(torch.FloatTensor)\n X = X / 255 # Normalizing\n\n y = data.train_labels\n\n X_train, X_validate, y_train, y_validate = train_test_split(\n X, y, test_size=10000, stratify=y, random_state=42)\n\n return X_train, X_validate, y_train, y_validate\n\n\nclass AverageMeter(object):\n \"\"\"Computes and stores the average and current value\"\"\"\n\n def __init__(self, name, fmt=':f'):\n self.name = name\n self.fmt = fmt\n self.reset()\n\n def reset(self):\n self.val = 0\n self.avg = 0\n self.sum = 0\n self.count = 0\n\n def update(self, val, n=1):\n self.val = val\n self.sum += val * n\n self.count += n\n self.avg = self.sum / self.count\n\n def __str__(self):\n fmtstr = '{name} {val' + self.fmt + '} ({avg' + self.fmt + '})'\n return fmtstr.format(**self.__dict__)\n\n\ndef evaluate(model, X, y, loss_fn, batch_size):\n all_y_truth = []\n all_y_pred = []\n loss_meter = AverageMeter(\"loss\")\n for start in range(0, len(X), batch_size):\n end = start + batch_size\n X_batch = X[start:end]\n y_batch = y[start:end]\n preds = model(X_batch)\n loss = loss_fn(preds, y_batch).item()\n _, pred_classes = torch.max(preds, 1)\n\n all_y_pred.extend(pred_classes)\n all_y_truth.extend(y_batch)\n loss_meter.update(loss, X_batch.shape[0])\n\n acc = accuracy_score(all_y_truth, all_y_pred)\n return acc, loss_meter.avg, all_y_pred, all_y_truth\n\n\ndef plot_confusion_matrices(model_name, epoch_num, train_y_truth, train_y_pred, validate_y_truth, validate_y_pred):\n skplt.metrics.plot_confusion_matrix(\n train_y_truth, train_y_pred, normalize=True)\n plot_folder = \"plots/\" + model_name + \"/\"\n if not os.path.exists(plot_folder):\n os.mkdir(plot_folder)\n plt.savefig(plot_folder + epoch_num + '_cm.png')\n plt.title(model_name)\n plt.clf()\n plt.close()\n\n\ndef train_evaluate_model(model_name, model, num_classes, num_epochs, batch_size, learning_rate, X_train, X_validate, y_train, y_validate, evaluate_every, print_every, device=None):\n\n loss_fn = torch.nn.CrossEntropyLoss()\n optimizer = torch.optim.SGD(model.parameters(), lr=learning_rate)\n train_losses = []\n validation_losses = []\n train_accuracies = []\n validation_accuracies = []\n epoch_ticks = []\n\n for epoch in range(num_epochs):\n epoch_start = time.time()\n # Training\n model.train()\n for start in range(0, len(X_train), batch_size):\n end = start + batch_size\n x = X_train[start:end]\n y = y_train[start:end]\n\n # Moving to device if given\n if device:\n x.to(device)\n y.to(device)\n\n optimizer.zero_grad()\n\n pred_y = model(x)\n loss = loss_fn(pred_y, y)\n loss.backward()\n optimizer.step()\n\n # Evaluating every evaluate_every epochs\n if epoch % evaluate_every == (evaluate_every - 1):\n epoch_ticks.append(epoch)\n model.eval()\n\n with torch.no_grad():\n # Train data set evaluation\n train_acc, train_loss, train_y_pred, train_y_truth = evaluate(\n model, X_train, y_train, loss_fn, batch_size)\n train_accuracies.append(train_acc)\n train_losses.append(train_loss)\n\n # Validation data set evaluation\n validation_acc, validation_loss, validate_y_pred, validate_y_truth = evaluate(\n model, X_validate, y_validate, loss_fn, batch_size)\n validation_accuracies.append(validation_acc)\n validation_losses.append(validation_loss)\n\n # Writing Confusion matrices\n plot_confusion_matrices(\n model_name, str(epoch+1), train_y_truth, train_y_pred, validate_y_truth, validate_y_pred)\n\n # Saving the model\n model_dir = \"models/\" + model_name + \"/\"\n if not os.path.exists(model_dir):\n os.mkdir(model_dir)\n torch.save(model.state_dict(), model_dir +\n str(epoch+1) + \".pt\")\n\n if epoch % print_every == (print_every - 1):\n epoch_end = time.time()\n print(\"Epoch:\", epoch+1, \"Train Loss:\",\n train_loss, \"Validation Loss:\", validation_loss, \"Time:\", epoch_end - epoch_start)\n\n return model, train_losses, validation_losses, train_accuracies, validation_accuracies, epoch_ticks, train_y_truth, train_y_pred, validate_y_truth, validate_y_pred\n","sub_path":"utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":5315,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"567138690","text":"#!/usr/bin/env python3\n'''\nUsage:\n ./energy_cons.py [options] []\n\nOptions:\n --help -h\n --lsp=FILE Supply the .lsp file.\n --intensity=I -I I Specify laser intensity directly in W/cm^2\n --pulse=T -T T Specify laser pulse FWHM directly in seconds.\n --w=W -w W Specify laser spotsize directly in meters.\n'''\nimport numpy as np;\nfrom docopt import docopt;\nfrom os import listdir;\nimport re;\nopts=docopt(__doc__,help=True);\nfs=listdir();\nlspf = [f for f in fs\n if re.match(\".*.lsp$\",f)][-1]\nhstf = [f for f in fs\n if re.match(\".*history.p4$\",f)][-1];\nd = np.loadtxt(hstf).T;\n#getting column of net energy probe\nwith open(hstf,'r') as f:\n for line in f.readlines():\n m = re.match(\"^#([0-9]+): *net energy\",line);\n if m:\n column = int(m.group(1))+1;\n break;\n else:\n raise ValueError(\n \"file {} does not have a net energy probe\".format(\n hstf));\n\ntime = d[1];\nE = d[column];\n\ne0 = 8.85418782e-12;\nc = 2.99792458e8;\ne = 1.60217657e-19\ndef laserE(E_0, T, w,dim=\"3D\"):\n '''\n Get total energy in a Gaussian Laser.\n \n\n Parameters and Keywords\n -----------------------\n E_0 -- Peak E field.\n T -- FWHM of the pulse.\n w -- Spotsize.\n dim -- Spatial dimension, either \"2D\"q, or \"3D\" or None for \"3D\"\n\n Returns laser energy.\n '''\n\n if dim == \"2D\":\n return w * np.sqrt(np.pi/2) * (c*e0*E_0**2)/2 * T*1e-2;\n elif not dim or dim == \"3D\":\n return w**2 * (np.pi/2) * (c*e0*E_0**2)/2 * T;\n else:\n raise ValueError(\"dim is not None, '2D' or '3D'\");\nif not (opts['--intensity'] and opts['--pulse'] and opts['--w']):\n raise NotImplementedError(\"Non explicit quantities not implemented yet.\")\nI = float(opts['--intensity']);\nE_0 = np.sqrt(2*I*1e4/(c*e0));\nT = float(opts['--pulse']);\nw = float(opts['--w']);\n\ne_laser = laserE(E_0,T,w,dim='2D');\nimport matplotlib\nif opts['']:\n matplotlib.use(\"agg\");\nimport matplotlib.pyplot as plt;\nplt.plot(time*1e6, E/e_laser);\nplt.xlabel(\"time (fs)\");\nplt.ylabel(\"energy ratio\");\nif opts['']:\n plt.savefig(opts['']);\nelse:\n plt.show();\n","sub_path":"bin/energy_cons.py","file_name":"energy_cons.py","file_ext":"py","file_size_in_byte":2200,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"485162407","text":"import cv2\nimport numpy as np\n\n# カメラを選ぶ\ncap = cv2.VideoCapture(0)\n\n# def DrawCircle()\n\nwhile (1):\n # カメラから画像データを読み込む\n ret, camera = cap.read()\n\n # 読み込んたかどうか判断\n if ret == True:\n # median filter、ノイズ除去\n camera = cv2.medianBlur(camera, 5)\n # RGB 2 GRAY\n cameraGray = cv2.cvtColor(camera, cv2.COLOR_RGB2GRAY)\n\n # 丸検出\n circles = cv2.HoughCircles(cameraGray, cv2.HOUGH_GRADIENT, 4, 100,\n param1=160, param2=150, minRadius=38, maxRadius=80)\n\n # 丸検出したかどうかを判断\n red_circles = []\n blue_circles = []\n j = -1\n\n # 出力ベクトル\n red_vec = []\n blue_vec = []\n\n # 円検出に成功したら\n if circles is not None:\n # 画像に丸を描く\n circles = np.int16(np.around(circles))\n\n # 円が3つ以上検出されたとき\n if circles.shape[1] > 1:\n #print(circles.shape[1])\n\n for i in circles[0, :]:\n j = j + 1\n\n if camera[i[1]][i[0]][2] > 100:# and camera[i[1]][i[0]][1] < 150:\n red_circles = np.append(red_circles, circles[0, j])\n\n if camera[i[1]][i[0]][0] > 200 and camera[i[0]][i[0]][2] < 150:\n blue_circles = np.append(blue_circles, circles[0, j])\n\n # 2つ円が検出されたとき長さは6\n if (len(red_circles) == 6):\n red_circles = np.reshape(red_circles, (len(red_circles) // 3, 3))\n # convert from float to int\n red_circles = np.int16(np.around(red_circles))\n cv2.line(camera, (red_circles[0, 0], red_circles[0, 1]), (red_circles[1, 0], red_circles[1, 1]), (255, 0, 0), 10)\n\n # 比較\n if(red_circles[0, 2] - red_circles[1, 2] > 0):\n red_vec_x = red_circles[0, 0] - red_circles[1, 0]\n red_vec_y = red_circles[0, 1] - red_circles[1, 1]\n print (red_circles[0, 0],red_circles[1, 0],red_vec_x)\n print (red_circles[0, 1],red_circles[1, 1],red_vec_y)\n\n else:\n red_vec_x = red_circles[1, 0] - red_circles[0, 0]\n red_vec_y = red_circles[1, 1] - red_circles[0, 1]\n\n # レッドベクトル作成\n red_vec = np.append(red_vec, red_vec_x)\n red_vec = np.append(red_vec, red_vec_y)\n print(\"red\")\n print(red_vec)\n\n\n if (len(blue_circles) == 6):\n blue_circles = np.reshape(blue_circles, (len(blue_circles) // 3, 3))\n # convert from float to int\n blue_circles = np.int16(np.around(blue_circles))\n cv2.line(camera, (blue_circles[0, 0], blue_circles[0, 1]), (blue_circles[1, 0], blue_circles[1, 1]), (255, 0, 0), 10)\n # 比較\n if(blue_circles[0, 2] - blue_circles[1, 2] > 0):\n blue_vec_x = blue_circles[0, 0] - blue_circles[1, 0]\n blue_vec_y = blue_circles[0, 1] - blue_circles[1, 1]\n\n else:\n blue_vec_x = blue_circles[1, 0] - blue_circles[0, 0]\n blue_vec_y = blue_circles[1, 1] - blue_circles[0, 1]\n\n # ブルーベクトル作成\n blue_vec = np.append(blue_vec, blue_vec_x)\n blue_vec = np.append(blue_vec, blue_vec_y)\n print(\"blue\")\n print(blue_vec)\n\n for i in circles[0,:]:\n # draw the outer circle\n cv2.circle(camera, (i[0], i[1]), i[2], (0, 255, 0), 2)\n # draw the center of the circle\n cv2.circle(camera, (i[0], i[1]), 2, (0, 0, 255), 3)\n\n cv2.imshow('capture', camera)\n\n # 'q'を押したら、プログラム終了\n if cv2.waitKey(1) & 0xFF == ord('q'):\n break\n\ncv2.waitKey(0)\ncv2.destroyAllWindows()\n","sub_path":"CircleCheck/test2.py","file_name":"test2.py","file_ext":"py","file_size_in_byte":4246,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"148811093","text":"#!/usr/bin/env python\n#\n# Copyright 2007 Google Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\nimport webapp2, jinja2, os, logging\n\nfrom google.appengine.ext import db\n\nfrom google.appengine.api import users\n\nfrom datetime import datetime\n\ntemplate_dir = os.path.join(os.path.dirname(__file__), 'template')\n\njinja_env = jinja2.Environment(loader = jinja2.FileSystemLoader(template_dir), \n autoescape=True)\n \nclass Handler(webapp2.RequestHandler):\n def write(self, *a, **kw):\n self.response.out.write(*a, **kw)\n def render_str(self, template, **params):\n t = jinja_env.get_template(template)\n return t.render(params)\n def render(self, template, **kw):\n self.write(self.render_str(template, **kw)) \n\n# DATABASE\nclass Post(db.Model):\n content = db.TextProperty(required = True)\n created = db.DateTimeProperty(auto_now_add = True)\n lastmodified = db.DateTimeProperty(auto_now_add = True)\n user = db.StringProperty()\n \nclass User(db.Model): \n username = db.StringProperty(required = True)\n password = db.TextProperty(required = True)\n email = db.StringProperty(required = False)\n registered = db.DateTimeProperty(auto_now_add = True)\n \n# HANDLERS\nclass MainHandler(Handler):\n def get(self):\n posts = top_posts()\n #tz_au = timezone('Australia/Sydney')\n self.render(\"postlist.html\", posts = posts)\n \n def post(self):\n content = self.request.get(\"content\")\n cookie_val = self.request.cookies.get('user_id', 0)\n logging.error(cookie_val)\n if cookie_val == 0:\n user_id = \"Anonymous\"\n else:\n user_id = cookie_val.split('|')[0]\n \n logging.error(user_id)\n \n if content:\n p = Post(user = user_id, content = content)\n p.put()\n self.redirect(\"/\")\n else:\n error = \"ooops\"\n self.render(content, error = error)\n\n# TEST-ONLY HANDLER\nclass test(Handler):\n def get(self):\n user = users.get_current_user()\n if user:\n greeting = (\"Welcome, %s! (sign out)\" %\n (user.nickname(), users.create_logout_url(\"/\")))\n else:\n greeting = (\"Sign in or register.\" %\n users.create_login_url(\"/\"))\n\n self.response.out.write(\"%s\" % greeting)\n \napp = webapp2.WSGIApplication([('/', MainHandler),\n ('/test', test)],\n debug=True)\n\n\n\n \n# GLOBAL FUNCTIONS \ndef top_posts(update = False):\n posts = db.GqlQuery(\"select * from Post order by created DESC limit 50\")\n posts = list(posts)\n return posts ","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":3432,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"27573204","text":"# -*- coding: utf-8 -*-\nimport numpy as np\ndata = np.load('./data/hifi_scale_stats.npy', allow_pickle=True)\nprint(data)\n# print(data)\nlist = data.tolist()\n# print(list['mel_mean'].shape)\n# print(list['mel_std'].shape)\n# print(list['linear_mean'].shape)\n# print(list['linear_std'].shape)\n#\nn1 = list['mel_mean']\nn2 = list['mel_std']\nn3 = list['linear_mean']\nn4 = list['linear_std']\n\n# n1 = np.random.randn(80,)\n# n2 = np.random.randn(80,)\n# n3 = np.random.randn(513,)\n# n4 = np.random.randn(513,)\n# list['audio_config']['stats_path'] = 'C:/Users/User/Desktop/SCE_TTS/data/glowtts-v2/scale_stats.npy'\nlist['audio_config']['stats_path'] = 'C:/Users/User/Desktop/SCE_TTS/data/hifigan-v2/scale_stats.npy'\n# list['audio_config']['stats_path'] = '/content/drive/My Drive/Colab Notebooks/data/hifigan-v2/scale_stats.npy'\n# print(list['audio_config']['stats_path'])\nn5 = list['audio_config']\n#\ndata0 = {}\ndata0['mel_mean'] = n1\ndata0['mel_std'] = n2\ndata0['linear_mean'] = n3\ndata0['linear_std'] = n4\ndata0['audio_config'] = n5\n# print(data0)\nresult = np.array(data0)\nprint(result)\nnp.save('./data/hifi.npy', result)","sub_path":"nump2.py","file_name":"nump2.py","file_ext":"py","file_size_in_byte":1107,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"369696781","text":"# Definition for a binary tree node.\n# class TreeNode(object):\n# def __init__(self, x):\n# self.val = x\n# self.left = None\n# self.right = None\n'''\ntest case: \n1. tree1-None-> return false,\n2. tree2-None-> return true,\n3. tree2 have more then 2 layers\n'''\n\nclass Solution(object):\n def isAHasBRoot(self, A, B):\n '''\n :type A: TreeNode (cannot be None)\n :type B: TreeNode (cannot be None)\n :rtype: bool\n traverse the tree A, to find if the root of B is contained in A\n '''\n result = False\n if A!=None and B!=None:\n if A.val==B.val:\n # now check the substructure\n result = self.isAHasB(A,B)\n if result==False:\n result = self.isAHasBRoot(A.left,B)\n if result==False:\n result = self.isAHasBRoot(A.right,B)\n return result\n \n def isAHasB(self, A, B):\n '''\n :type A: TreeNode (cannot be None)\n :type B: TreeNode (cannot be None)\n :rtype: bool\n to check if A contains B\n '''\n if B==None:\n return True\n if A==None:\n return False\n if A.val!=B.val:\n return False\n\n result1 = self.isAHasB(A.left,B.left)\n result2 = self.isAHasB(A.right,B.right)\n return result1&result2\n\n\n def isSubStructure(self, A, B):\n \"\"\"\n :type A: TreeNode\n :type B: TreeNode\n :rtype: bool\n \"\"\"\n if A==None:\n return False\n elif B==None:\n return False\n else:\n return self.isAHasBRoot(A,B)\n","sub_path":"Q18_isSubstructure.py","file_name":"Q18_isSubstructure.py","file_ext":"py","file_size_in_byte":1651,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"435128525","text":"# -*- coding: utf-8 -*-\n\nimport csv\nimport codecs\nimport urllib.request\nimport json\nimport os.path\nimport sys\nimport base64\nimport subprocess\nfrom subprocess import Popen\n\n# script path\nscriptpath = os.path.abspath(os.path.dirname(__file__)) + \"/\"\n\n# args\nargs = sys.argv\n\n# read cache\ndata = {}\nexception_ip = []\nif os.path.exists(scriptpath + \"vsc.txt\"):\n\tf = open(scriptpath + \"vsc.txt\", \"r\")\n\tdata = json.load(f)\n\tf.close()\nif os.path.exists(scriptpath + \"exception.txt\"):\n\tf = open(scriptpath + \"exception.txt\", \"r\")\n\tlines = f.readlines()\n\tf.close()\n\t#for line in lines:\n\t# itemlist = line[:-1].split(\"\\n\")\n\t# exception_ip.append([ item for item in itemlist ])\n\tfor line in lines:\n\t\titemlist = line[:-1].split(\"\\n\")\n\t\texception_ip.append(itemlist[0])\nif os.path.exists(scriptpath + \"api_key.txt\"):\n\tf = open(scriptpath + \"api_key.txt\", \"r\")\n\tipstack_key = f.read()\n\tf.close()\n\ttry:\n\t\tres2 = urllib.request.urlopen(\"http://api.ipstack.com/8.8.8.8?access_key=\" + ipstack_key)\n\texcept:\n\t\tprint(\"\\n--------------------------------------------------------\")\n\t\tprint(\"ipstack err.\")\n\t\tprint(\"\\n--------------------------------------------------------\")\n\t\tsys.exit(1)\n\tiptoaddrs = json.loads(res2.read().decode('utf8'))\n\tif iptoaddrs['success'] is False:\n\t\tprint(\"\\n--------------------------------------------------------\")\n\t\tprint(\"ipstack API is err :\" + iptoaddrs['error']['info'])\n\t\tprint('please check \"api_key.txt\"')\n\t\tprint(\"\\n--------------------------------------------------------\")\n\t\tsys.exit(1)\nelse:\n\tprint(\"\\n--------------------------------------------------------\")\n\tprint(\"Please enter ipstack API key .\")\n\tipstack_key = input('>>> ')\n\ttry:\n\t\tres2 = urllib.request.urlopen(\"http://api.ipstack.com/8.8.8.8?access_key=\" + ipstack_key)\n\texcept:\n\t\tprint(\"\\n--------------------------------------------------------\")\n\t\tprint(\"ipstack err.\")\n\t\tprint(\"\\n--------------------------------------------------------\")\n\t\tsys.exit(1)\n\tif iptoaddrs['success'] is False:\n\t\tprint(\"\\n--------------------------------------------------------\")\n\t\tprint(\"ipstack API is err :\" + iptoaddrs['error']['info'])\n\t\tprint(\"please check key\")\n\t\tprint(\"\\n--------------------------------------------------------\")\n\t\tsys.exit(1)\n\tf = open(scriptpath + \"api_key.txt\", \"w\")\n\tf.write(ipstack_key)\n\tf.close()\n\n# read region\nf = open(scriptpath + \"region.txt\", \"r\")\nregion = json.load(f)\nf.close()\n\n# read VPN list\nres1 = urllib.request.urlopen(\"http://www.vpngate.net/api/iphone/\")\ncr = csv.reader(codecs.iterdecode(res1, 'utf-8'), delimiter=\",\", lineterminator=\"\\r\\n\")\n\n# search vpn server\nprint(\"\\n--------------------------------------------------------\")\nprint(\"Searching VPN server in the following region.\")\nprint(\"--------------------------------------------------------\")\nfor i,r in enumerate(args):\n\tif i != 0:\n\t\tprint(region[int(r)-1])\nprint(\"--------------------------------------------------------\")\n\nfor row in cr:\n\tif len(row) == 15 and row[6] == \"JP\":\n\t\tregion_code = 0\n\t\t#list\n\t\tif row[1] in exception_ip:\n\t\t\tcontinue\n\t\tif row[1] in data:\n\t\t\tregion_code = data[row[1]]\n\t\telse:\n\t\t\t# get region_code\n\t\t\tprint(\"Searching the region of IP(\" + row[1] + \").\")\n\t\t\tres2 = urllib.request.urlopen(\"http://api.ipstack.com/\" + row[1] + \"?access_key=\" + ipstack_key)\n\t\t\tiptoaddrs = json.loads(res2.read().decode('utf8'))\n\t\t\tprint(iptoaddrs[\"region_code\"])\n\t\t\tif iptoaddrs[\"region_code\"] is None:\n\t\t\t\tregion_code = 0\n\t\t\tif iptoaddrs[\"region_code\"] != \"\" and iptoaddrs[\"region_code\"] is not None:\n\t\t\t\tregion_code = int(iptoaddrs[\"region_code\"])\n\t\t\t\tdata[row[1]] = region_code\n\t\t\t\tf = open(scriptpath + \"vsc.txt\", \"w\")\n\t\t\t\tjson.dump(data, f)\n\t\t\t\tf.close()\n\t\t\telse:\n\t\t\t\tprint(\"Cannot identify the region of IP(\" + row[1] + \").\")\n\t\t#if region_code != 0 or region_code != None or region_code != \"\":\n\t\tif region_code != 0:\n\t\t\tprint(\"Region of \" + row[0] + \" (\" + row[1] + \") => \" + str(region_code) + \":\" + region[region_code-1])\n\t\t\t# search region\n\t\t\tfor i,arg in enumerate(args):\n\t\t\t\tif i != 0 and int(arg) == region_code:\n\t\t\t\t\tprint(\"Connecting to \" + row[0] + \"... Please wait.\")\n\t\t\t\t\t# Base64 decode\n\t\t\t\t\tovpn = base64.b64decode(row[14])\n\t\t\t\t\tovpn = ovpn.decode(\"UTF-8\")\n\t\t\t\t\t# make .ovpn\n\t\t\t\t\tf = open(scriptpath + \"vpnovpn.ovpn\", \"w\")\n\t\t\t\t\tf.write(ovpn)\n\t\t\t\t\tf.close()\n\t\t\t\t\t# vpn connect\n\t\t\t\t\tcom = \"/usr/sbin/openvpn \" + scriptpath + \"vpnovpn.ovpn\"\n\t\t\t\t\tproc = Popen(com, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)\n\t\t\t\t\tbuf = []\n\t\t\t\t\twhile True:\n\t\t\t\t\t\tline = proc.stdout.readline()\n\t\t\t\t\t\tline = line.decode(\"UTF-8\")\n\t\t\t\t\t\tprint(line)\n\t\t\t\t\t\t#buf = buf + line\n\t\t\t\t\t\tif line.find(\"failed\") > -1:\n\t\t\t\t\t\t\tprint(\"Connection failed: \" + row[0])\n\t\t\t\t\t\t\t#break\n\t\t\t\t\t\t\tcmd = scriptpath + \"openvpnclient.sh stop\"\n\t\t\t\t\t\t\tsubprocess.call(cmd.split())\n\t\t\t\t\t\t\tcontinue\n\t\t\t\t\t\telif line.find(\"Initialization Sequence Completed\") > -1:\n\t\t\t\t\t\t\tprint(\"Connection success: \")\n\t\t\t\t\t\t\tprint(\"\\t\" + row[0] + \" (\" + row[1] + \") => \" + str(region_code) + \":\" + region[region_code-1])\n\t\t\t\t\t\t\tprint(\"--------------------------------------------------------\")\n\t\t\t\t\t\t\tsys.exit(0)\nprint(\"Cannot find a valid VPN server.\")\nprint(\"--------------------------------------------------------\")\nsys.exit(1)\n","sub_path":"vpnsearchconnect.py","file_name":"vpnsearchconnect.py","file_ext":"py","file_size_in_byte":5188,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"243231570","text":"#identifying isoforms that are tissue-specific splicing events vs tissue-specific due to differences in gene expression (present in only one tissue) vs isoforms found in two or more tissues\n#tissue specific splicing events will also include TSS, TTS differences (will have a separate script to compare the specifics)\n#combined sexes for somatic tissues but kept testis and ovary separate\n#to run script: python3 Identify.Tissue.Specific.Splicing.py \n#Author: Alice Naftaly, Aug 2020\n\nimport sys\n\n\n#read in isoforms specific to brain\n#returns dictionary with key == gene and value == isoform\ndef pull_brain_specific_isoforms():\n brain_file = sys.argv[1]\n brain_isoforms = {}\n with open(brain_file, 'r') as brain_only:\n for line in brain_only:\n if line.startswith(\"PB\"):\n new_line = line.split()\n isoform = new_line[0]\n gene_id = new_line[1]\n if gene_id in brain_isoforms:\n brain_isoforms[gene_id].append(isoform)\n elif gene_id not in brain_isoforms:\n brain_isoforms.update({gene_id:[isoform]})\n return brain_isoforms\n\n#read in isoforms specific to liver\n#returns dictionary with key == gene and value == isoform\ndef pull_liver_specific_isoforms():\n liver_file = sys.argv[2]\n liver_isoforms = {}\n with open(liver_file, 'r') as liver_only:\n for line in liver_only:\n if line.startswith(\"PB\"):\n new_line = line.split()\n isoform = new_line[0]\n gene_id = new_line[1]\n if gene_id in liver_isoforms:\n liver_isoforms[gene_id].append(isoform)\n elif gene_id not in liver_isoforms:\n liver_isoforms.update({gene_id:[isoform]})\n return liver_isoforms\n\n#read in isoforms specific to pronephros\n#returns dictionary with key == gene and value == isoform\ndef pull_pronephros_specific_isoforms():\n pronephros_file = sys.argv[3]\n pronephros_isoforms = {}\n with open(pronephros_file, 'r') as pronephros_only:\n for line in pronephros_only:\n if line.startswith(\"PB\"):\n new_line = line.split()\n isoform = new_line[0]\n gene_id = new_line[1]\n if gene_id in pronephros_isoforms:\n pronephros_isoforms[gene_id].append(isoform)\n elif gene_id not in pronephros_isoforms:\n pronephros_isoforms.update({gene_id:[isoform]})\n return pronephros_isoforms\n\n#read in isoforms specific to testis\n#returns dictionary with key == gene and value == isoform\ndef pull_testis_specific_isoforms():\n testis_file = sys.argv[4]\n testis_isoforms = {}\n with open(testis_file, 'r') as testis_only:\n for line in testis_only:\n if line.startswith(\"PB\"):\n new_line = line.split()\n isoform = new_line[0]\n gene_id = new_line[1]\n if gene_id in testis_isoforms:\n testis_isoforms[gene_id].append(isoform)\n elif gene_id not in testis_isoforms:\n testis_isoforms.update({gene_id:[isoform]})\n return testis_isoforms\n\n#read in isoforms specific to ovary\n#returns dictionary with key == gene and value == isoform\ndef pull_ovary_specific_isoforms():\n ovary_file = sys.argv[5]\n ovary_isoforms = {}\n with open(ovary_file, 'r') as ovary_only:\n for line in ovary_only:\n if line.startswith(\"PB\"):\n new_line = line.split()\n isoform = new_line[0]\n gene_id = new_line[1]\n if gene_id in ovary_isoforms:\n ovary_isoforms[gene_id].append(isoform)\n elif gene_id not in ovary_isoforms:\n ovary_isoforms.update({gene_id:[isoform]})\n return ovary_isoforms\n\n\n#genes shared between tissues can be shared with as few as two tissues up to all tissues\ndef sort_isoforms():\n brain_isoforms = pull_brain_specific_isoforms()\n liver_isoforms = pull_liver_specific_isoforms()\n pronephros_isoforms = pull_pronephros_specific_isoforms()\n testis_isoforms = pull_testis_specific_isoforms()\n ovary_isoforms = pull_ovary_specific_isoforms()\n pronephros_shared_genes = {}\n pronephros_unique_genes = {}\n brain_shared_genes = {}\n brain_unique_genes = {}\n liver_shared_genes = {}\n liver_unique_genes = {}\n testis_shared_genes = {}\n testis_unique_genes = {}\n ovary_shared_genes = {}\n ovary_unique_genes = {}\n #there should be no shared isoforms as these isoforms are already specific to a tissue\n for p_gene in pronephros_isoforms:\n single_gene_pronephros = pronephros_isoforms[p_gene]\n if p_gene in brain_isoforms:\n single_gene_brain = brain_isoforms[p_gene]\n #shared gene and different isoform == alternative splicing candidate\n for p_iso in single_gene_pronephros:\n final = [p_iso, \"shared.gene.with.brain\"]\n if p_gene in pronephros_shared_genes:\n pronephros_shared_genes[p_gene].append(final)\n elif p_gene not in pronephros_shared_genes:\n pronephros_shared_genes.update({p_gene:[final]})\n for b_iso in single_gene_brain:\n final = [b_iso, \"shared.gene.with.pronephros\"]\n if p_gene in brain_shared_genes:\n brain_shared_genes[p_gene].append(final)\n elif p_gene not in brain_shared_genes:\n brain_shared_genes.update({p_gene:[final]})\n if p_gene in liver_isoforms:\n single_gene_liver = liver_isoforms[p_gene]\n for p_iso in single_gene_pronephros:\n final = [p_iso, \"shared.gene.with.liver\"]\n if p_gene in pronephros_shared_genes:\n pronephros_shared_genes[p_gene].append(final)\n elif p_gene not in pronephros_shared_genes:\n pronephros_shared_genes.update({p_gene:[final]})\n for l_iso in single_gene_liver:\n final = [l_iso, \"shared.gene.with.pronephros\"]\n if p_gene in liver_shared_genes:\n liver_shared_genes[p_gene].append(final)\n elif p_gene not in liver_shared_genes:\n liver_shared_genes.update({p_gene:[final]})\n if p_gene in testis_isoforms:\n single_gene_testis = testis_isoforms[p_gene]\n for p_iso in single_gene_pronephros:\n final = [p_iso, \"shared.gene.with.testis\"]\n if p_gene in pronephros_shared_genes:\n pronephros_shared_genes[p_gene].append(final)\n elif p_gene not in pronephros_shared_genes:\n pronephros_shared_genes.update({p_gene:[final]})\n for t_iso in single_gene_testis:\n final = [t_iso, \"shared.gene.with.pronephros\"]\n if p_gene in testis_shared_genes:\n testis_shared_genes[p_gene].append(final)\n elif p_gene not in testis_shared_genes:\n testis_shared_genes.update({p_gene:[final]})\n if p_gene in ovary_isoforms:\n single_gene_ovary = ovary_isoforms[p_gene]\n for p_iso in single_gene_pronephros:\n final = [p_iso, \"shared.gene.with.ovary\"]\n if p_gene in pronephros_shared_genes:\n pronephros_shared_genes[p_gene].append(final)\n elif p_gene not in pronephros_shared_genes:\n pronephros_shared_genes.update({p_gene:[final]})\n for o_iso in single_gene_ovary:\n final = [o_iso, \"shared.gene.with.pronephros\"]\n if p_gene in ovary_shared_genes:\n ovary_shared_genes[p_gene].append(final)\n elif p_gene not in ovary_shared_genes:\n ovary_shared_genes.update({p_gene:[final]})\n for final_p_gene in pronephros_isoforms:\n if final_p_gene not in pronephros_shared_genes:\n pronephros_unique_genes.update({final_p_gene:pronephros_isoforms[final_p_gene]})\n for b_gene in brain_isoforms:\n single_gene_brain = brain_isoforms[b_gene]\n if b_gene in liver_isoforms:\n single_gene_liver = liver_isoforms[b_gene]\n for b_iso in single_gene_brain:\n final = [b_iso, \"shared.gene.with.liver\"]\n if b_gene in brain_shared_genes:\n brain_shared_genes[b_gene].append(final)\n elif b_gene not in brain_shared_genes:\n brain_shared_genes.update({b_gene:[final]})\n for l_iso in single_gene_liver:\n final = [l_iso, \"shared.gene.with.brain\"]\n if b_gene in liver_shared_genes:\n liver_shared_genes[b_gene].append(final)\n elif b_gene not in liver_shared_genes:\n liver_shared_genes.update({b_gene:[final]})\n if b_gene in testis_isoforms:\n single_gene_testis = testis_isoforms[b_gene]\n for b_iso in single_gene_brain:\n final = [b_iso, \"shared.gene.with.testis\"]\n if b_gene in brain_shared_genes:\n brain_shared_genes[b_gene].append(final)\n elif b_gene not in brain_shared_genes:\n brain_shared_genes.update({b_gene:[final]})\n for t_iso in single_gene_testis:\n final = [t_iso, \"shared.gene.with.brain\"]\n if b_gene in testis_shared_genes:\n testis_shared_genes[b_gene].append(final)\n elif b_gene not in testis_shared_genes:\n testis_shared_genes.update({b_gene:[final]})\n if b_gene in ovary_isoforms:\n single_gene_ovary = ovary_isoforms[b_gene]\n for b_iso in single_gene_brain:\n final = [b_iso, \"shared.gene.with.ovary\"]\n if b_gene in brain_shared_genes:\n brain_shared_genes[b_gene].append(final)\n elif b_gene not in brain_shared_genes:\n brain_shared_genes.update({b_gene:[final]})\n for o_iso in single_gene_ovary:\n final = [o_iso, \"shared.gene.with.brain\"]\n if b_gene in ovary_shared_genes:\n ovary_shared_genes[b_gene].append(final)\n elif b_gene not in ovary_shared_genes:\n ovary_shared_genes.update({b_gene:[final]})\n for final_b_gene in brain_isoforms:\n if final_b_gene not in brain_shared_genes:\n brain_unique_genes.update({final_b_gene:brain_isoforms[final_b_gene]})\n for l_gene in liver_isoforms:\n single_gene_liver = liver_isoforms[l_gene]\n if l_gene in testis_isoforms:\n single_gene_testis = testis_isoforms[l_gene]\n for l_iso in single_gene_liver:\n final = [l_iso, \"shared.gene.with.testis\"]\n if l_gene in liver_shared_genes:\n liver_shared_genes[l_gene].append(final)\n elif l_gene not in liver_shared_genes:\n liver_shared_genes.update({l_gene:[final]})\n for t_iso in single_gene_testis:\n final = [t_iso, \"shared.gene.with.liver\"]\n if l_gene in testis_shared_genes:\n testis_shared_genes[l_gene].append(final)\n elif l_gene not in testis_shared_genes:\n testis_shared_genes.update({l_gene:[final]})\n if l_gene in ovary_isoforms:\n single_gene_ovary = ovary_isoforms[l_gene]\n for l_iso in single_gene_liver:\n final = [l_iso, \"shared.gene.with.ovary\"]\n if l_gene in liver_shared_genes:\n liver_shared_genes[l_gene].append(final)\n elif l_gene not in liver_shared_genes:\n liver_shared_genes.update({l_gene:[final]})\n for o_iso in single_gene_ovary:\n final = [o_iso, \"shared.gene.with.liver\"]\n if l_gene in ovary_shared_genes:\n ovary_shared_genes[l_gene].append(final)\n elif l_gene not in ovary_shared_genes:\n ovary_shared_genes.update({l_gene:[final]})\n for final_l_gene in liver_isoforms:\n if final_l_gene not in liver_shared_genes:\n liver_unique_genes.update({final_l_gene:liver_isoforms[final_l_gene]})\n for t_gene in testis_isoforms:\n single_gene_testis = testis_isoforms[t_gene]\n if t_gene in ovary_isoforms:\n single_gene_ovary = ovary_isoforms[t_gene]\n for t_iso in single_gene_testis:\n final = [t_iso, \"shared.gene.with.ovary\"]\n if t_gene in testis_shared_genes:\n testis_shared_genes[t_gene].append(final)\n elif t_gene not in testis_shared_genes:\n testis_shared_genes.update({t_gene:[final]})\n for o_iso in single_gene_ovary:\n final = [o_iso, \"shared.gene.with.testis\"]\n if t_gene in ovary_shared_genes:\n ovary_shared_genes[t_gene].append(final)\n elif t_gene not in ovary_shared_genes:\n ovary_shared_genes.update({t_gene:[final]})\n for final_t_gene in testis_isoforms:\n if final_t_gene not in testis_shared_genes:\n testis_unique_genes.update({final_t_gene:testis_isoforms[final_t_gene]})\n for o_gene in ovary_isoforms:\n if o_gene not in ovary_shared_genes:\n ovary_unique_genes.update({o_gene:ovary_isoforms[o_gene]})\n return brain_shared_genes, brain_unique_genes, liver_shared_genes, liver_unique_genes, pronephros_shared_genes, pronephros_unique_genes, testis_shared_genes, testis_unique_genes, ovary_shared_genes, ovary_unique_genes\n\n#write isoforms to output file\n#each output file will have 2 columns:\n#isoform.id \\t gene id \\n\ndef write_output():\n brain_shared_genes, brain_unique_genes, liver_shared_genes, liver_unique_genes, pronephros_shared_genes, pronephros_unique_genes, testis_shared_genes, testis_unique_genes, ovary_shared_genes, ovary_unique_genes = sort_isoforms()\n brain_shared_output = sys.argv[6]\n brain_unique_output = sys.argv[7]\n liver_shared_output = sys.argv[8]\n liver_unique_output = sys.argv[9]\n pronephros_shared_output = sys.argv[10]\n pronephros_unique_output = sys.argv[11]\n testis_shared_output = sys.argv[12]\n testis_unique_output = sys.argv[13]\n ovary_shared_output = sys.argv[14]\n ovary_unique_output = sys.argv[15]\n with open(brain_shared_output, 'a') as out1, open(brain_unique_output, 'a') as out2, open(liver_shared_output, 'a') as out3, open(liver_unique_output, 'a') as out4, open(pronephros_shared_output, 'a') as out5, open(pronephros_unique_output, 'a') as out6, open(testis_shared_output, 'a') as out7, open(testis_unique_output, 'a') as out8, open(ovary_shared_output, 'a') as out9, open(ovary_unique_output, 'a') as out10:\n for key in brain_shared_genes:\n single_key = brain_shared_genes[key]\n for v in single_key:\n final = \"%s\\t%s\\t%s\\n\" % (str(key), str(v[0]), str(v[1]))\n out1.write(final)\n for key2 in brain_unique_genes:\n single_key = brain_unique_genes[key2]\n for v in single_key:\n final = \"%s\\t%s\\n\" % (str(key2), str(v))\n out2.write(final)\n for key3 in liver_shared_genes:\n single_key = liver_shared_genes[key3]\n for v in single_key:\n final = \"%s\\t%s\\t%s\\n\" % (str(key3), str(v[0]), str(v[1]))\n out3.write(final)\n for key4 in liver_unique_genes:\n single_key = liver_unique_genes[key4]\n for v in single_key:\n final = \"%s\\t%s\\n\" % (str(key4), str(v))\n out4.write(final)\n for key5 in pronephros_shared_genes:\n single_key = pronephros_shared_genes[key5]\n for v in single_key:\n final = \"%s\\t%s\\t%s\\n\" % (str(key5), str(v[0]), str(v[1]))\n out5.write(final)\n for key6 in pronephros_unique_genes:\n single_key = pronephros_unique_genes[key6]\n for v in single_key:\n final = \"%s\\t%s\\n\" % (str(key6), str(v))\n out6.write(final)\n for key7 in testis_shared_genes:\n single_key = testis_shared_genes[key7]\n for v in single_key:\n final = \"%s\\t%s\\t%s\\n\" % (str(key7),str(v[0]), str(v[1]))\n out7.write(final)\n for key8 in testis_unique_genes:\n single_key = testis_unique_genes[key8]\n for v in single_key:\n final = \"%s\\t%s\\n\" % (str(key8), str(v))\n out8.write(final)\n for key9 in ovary_shared_genes:\n single_key = ovary_shared_genes[key9]\n for v in single_key:\n final = \"%s\\t%s\\t%s\\n\" % (str(key9), str(v[0]), str(v[1]))\n out9.write(final)\n for key10 in ovary_unique_genes:\n single_key = ovary_unique_genes[key10]\n for v in single_key:\n final = \"%s\\t%s\\n\" % (str(key10), str(v))\n out10.write(final)\nwrite_output()\n","sub_path":"Tissue_Specific_Transcriptome_Complexity/Identify.Tissue.Specific.Splicing.py","file_name":"Identify.Tissue.Specific.Splicing.py","file_ext":"py","file_size_in_byte":17916,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"135427070","text":"n = int(input())\na = list(map(int, input().split()))\nfour = 0\ntwo = 0\nfor i in range(n):\n if a[i] % 4 == 0:\n four += 1\n elif a[i] % 2 == 0:\n two += 1\n\nif four >= n // 2 or n - 2 * four <= two:\n print(\"Yes\")\nelse:\n print(\"No\")\n","sub_path":"Python_codes/p03637/s429807101.py","file_name":"s429807101.py","file_ext":"py","file_size_in_byte":252,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"576839055","text":"import pandas as pd\nimport seaborn as sns\nimport ipywidgets as widgets\n\n\ndef unique_sorted_values_plus_ALL(array):\n ALL = \"ALL\"\n unique = array.unique().tolist()\n unique.sort()\n unique.insert(0, ALL)\n return unique\n\n\ndef unique_sources(array):\n ALL = \"ALL\"\n unique = set()\n for li in array:\n for l in li:\n unique.add(l)\n ret = list(unique)\n ret.insert(0, ALL)\n return ret\n\n\ndef colour_ge_value(value, comparison):\n if len(value.split()) >= comparison:\n return \"color: red\"\n else:\n return \"color: black\"\n","sub_path":"utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":574,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"299160150","text":"import os\n\nimport DB_DIR.mongo\nfrom DB_DIR import mongo as mng\nimport pandas as pd\nimport numpy as np\nimport datetime as dt\nfrom dateutil.relativedelta import relativedelta\nimport matplotlib.dates as mdates\n\ndef nearest(psa_dates, six_months):\n return min(psa_dates, key=lambda x: np.absolute(x - six_months))\n\ndef treatment_getter(treatment: str):\n if treatment == \"RT\":\n trx_time = \"end\"\n else:\n trx_time = \"start\"\n df1 = pd.DataFrame(columns=[\"Patient\", treatment + \" \" + trx_time, \"Closest PSA to +6m\", \"Diff (days)\", \"PSA Level\"])\n patients = mng.getPatientList()\n for patient in patients:\n try:\n trx_relevant = mdates.num2date([date[trx_time] for date in DB_DIR.mongo.getTreatments(patient) if treatment in date['name']][0]).date()\n except IndexError as e:\n print(e)\n continue\n six_months = (trx_relevant + relativedelta(months=+6))\n psas = mng.retrieveDoc(patient)['PSAs']\n psa_dates = [dt.datetime.strptime(date, \"%Y-%m-%d\").date() for date in list(psas.keys())]\n closest = nearest(psa_dates, six_months)\n diff = closest - six_months\n psa_level = psas[closest.strftime(\"%Y-%m-%d\")]\n app_dict = {'Patient': patient, treatment + \" \" + trx_time: trx_relevant, \"Closest PSA to +6m\": closest, \"Diff (days)\": diff, \"PSA Level\": psa_level}\n join_dict = parameter_getter(patient, closest)\n z = dict(list(app_dict.items()) + list(join_dict.items()))\n df1 = df1.append(z, ignore_index=True)\n return df1\n\ndef parameter_getter(patient: str, psa_closest):\n # Get closest sample to six months\n dicto = {}\n this_patient = mng.retrieveDoc(patient)['filters']\n zeroMon = this_patient[[filter for filter in this_patient.keys() if this_patient[filter]['tPoint'] == \"+00m\"][0]]\n dates = [dt.datetime.strptime(this_patient[item]['DateRec'], \"%Y-%m-%d\").date() for item in this_patient.keys()]\n samp_closest = nearest(dates, psa_closest).strftime(\"%Y-%m-%d\")\n relevant = this_patient[[filter for filter in this_patient.keys() if this_patient[filter]['DateRec'] == samp_closest][0]]\n dicto['Closest Sample Date to psa Date'] = samp_closest\n for item in relevant.keys():\n if item != \"CTCNum\":\n try:\n dicto[\"Diff from +00m \" + item] = relevant[item] - zeroMon[item]\n except TypeError:\n continue\n return dicto\n\ndef main():\n df1 = treatment_getter(\"Bicalutamide\")\n print(df1)\n df2 = treatment_getter(\"RT\")\n tod = dt.datetime.now().strftime(\"%A - %B %d, %Y\")\n writer = pd.ExcelWriter(os.path.join(os.path.expanduser('~')), 'Desktop','HarveyFinder-output-%s.xlsx' % tod)\n df1.to_excel(writer, \"Bicalutamide\")\n df2.to_excel(writer, \"RT\")\n writer.save()\n\nif __name__ == \"__main__\":\n main()","sub_path":"Scripts/HarveyFinder.py","file_name":"HarveyFinder.py","file_ext":"py","file_size_in_byte":2834,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"341609088","text":"\r\n# original code by Pavel Rusanov. GeekBrains\r\n\r\nimport random\r\nimport turtle\r\nimport sys\r\n\r\n# PEP8\r\n\r\ndef gotoxy(x, y):\r\n turtle.penup()\r\n turtle.goto(x, y)\r\n turtle.pendown() \r\n\r\n\r\ndef draw_line(from_x, from_y, to_x, to_y):\r\n gotoxy(from_x, from_y)\r\n turtle.goto(to_x, to_y)\r\n\r\n\r\nx = random.randint(1,100)\r\n\r\nprint(x)\r\n\r\n#turtle.speed(0)\r\n\r\n# turtle.circle(30)\r\n\r\ncoord_list = []\r\ncoord = open('coord.txt')\r\n\r\nfor line in coord:\r\n line = line.strip().split(',')\r\n nums = []\r\n for n in line:\r\n nums.append(int(n))\r\n\r\n coord_list.append(nums)\r\n\r\nprint(coord_list) \r\n\r\n\r\n# вертикальная\r\ndraw_line(*coord_list[0])\r\n\r\n# горизонтальная\r\ndraw_line(*coord_list[1])\r\n\r\ndraw_line(-160, 40, -120, 80)\r\n\r\ndraw_line(-100, 80, -100, 40)\r\n\r\ngotoxy(-100, 0)\r\nturtle.circle(20)\r\n\r\ndraw_line(-100, 0, -100, -50)\r\n\r\ndraw_line(-100, -10, -100, -50)\r\n\r\ndraw_line(-100, -10, -120, -20)\r\ndraw_line(-100, -10, -80, -20)\r\n\r\ndraw_line(-100, -50, -120, -60)\r\ndraw_line(-100, -50, -80, -60)\r\n\r\n\r\n# from tkinter import SimpleDialog\r\n# SimpleDialog.textinput()\r\n\r\n\r\n\r\nanswer = turtle.textinput(\"Играть?\", \"y/n\")\r\n\r\nif answer == 'n':\r\n sys.exit()\r\n\r\ntry_count = 0\r\n\r\nwhile True:\r\n number = turtle.numinput(\"Угадайте\", \"Число\", 0, 0, 100)\r\n\r\n if number == x:\r\n turtle.color('green')\r\n gotoxy(-150, 200)\r\n turtle.write(\"Ура! Вы победили!\", \r\n font=(\"Arial\", 28, \"normal\"))\r\n break\r\n\r\n else:\r\n turtle.color('red')\r\n gotoxy(-150, 100)\r\n turtle.write(\"Неверно!\",\r\n font=(\"Arial\", 28, \"normal\")) \r\n\r\n try_count += 1\r\n\r\n if try_count == 10:\r\n gotoxy(-20, 230)\r\n turtle.write(\"Вы програли!\",\r\n font=(\"Arial\", 44, \"normal\"))\r\n break \r\n\r\n\r\n\r\n\r\n\r\n# input('Нажмите любую клавишу') # raw_input()\r\n","sub_path":"Materials/original_gibbet_GB.py","file_name":"original_gibbet_GB.py","file_ext":"py","file_size_in_byte":1944,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"279328776","text":"from handlers import BlogHandler\nfrom handlers.utils import check_user_logged_in, check_valid_post, check_valid_comment, user_owns_comment \nfrom google.appengine.ext import db\n\nclass DeleteComment(BlogHandler):\n \"\"\"Handler for deleting an exsisting comment\"\"\"\n @check_user_logged_in\n @check_valid_post\n @check_valid_comment\n @user_owns_comment\n def get(self, post_id, comment_id, post = None, comment = None):\n comment.delete()\n post.comment -= 1\n post.put()\n self.redirect('/blog/%s' % post_id)","sub_path":"handlers/deletecomment.py","file_name":"deletecomment.py","file_ext":"py","file_size_in_byte":542,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"359705996","text":"#!/usr/bin/env python\n#license removed for brevity\n\nimport rospy\nfrom std_msgs.msg import Int32\nfrom std_msgs.msg import Float32\nfrom std_msgs.msg import Int32MultiArray\nimport turtle\nimport time\n\nforward_value = [0 for i in range(5)]\nturn_value = [0 for i in range(5)]\n\nrospy.init_node('Turtle_view', anonymous=True)\n\ntur = []\ncolorL = ['blue', 'red', 'yellow', 'green', 'black']\nfor i in range(5):\n tur_t = turtle.Turtle()\n tur_t.shape(\"turtle\")\n tur_t.color(colorL[i])\n tur.append(tur_t)\n\ndef forward_hd(data, index):\n global forward_value\n data = int(data.data)\n if data >= 300:\n data = 300\n if data <= -300:\n data = -300\n forward_value[index] = data\n print(forward_value)\n\ndef turn_hd(data, index):\n global turn_value\n data = int(data.data)\n while data > 1800:\n data = data-360\n while data < -1800:\n data = data+360\n turn_value[index] = data\n\npub1 = []\npub2 = []\nfor i in range(5):\n pub1_1 = rospy.Publisher('/turtle/'+str(i+1)+'/position', Int32MultiArray, queue_size = 10)\n pub2_1 = rospy.Publisher('/turtle/'+str(i+1)+'/heading', Float32, queue_size = 10)\n pub1.append(pub1_1)\n pub2.append(pub2_1)\nrospy.Subscriber('/turtle/1/forward', Int32, lambda data : forward_hd(data, 0))\nrospy.Subscriber('/turtle/1/turn', Int32, lambda data : turn_hd(data, 0))\nrospy.Subscriber('/turtle/2/forward', Int32, lambda data : forward_hd(data, 1))\nrospy.Subscriber('/turtle/2/turn', Int32, lambda data : turn_hd(data, 1))\nrospy.Subscriber('/turtle/3/forward', Int32, lambda data : forward_hd(data, 2))\nrospy.Subscriber('/turtle/3/turn', Int32, lambda data : turn_hd(data, 2))\nrospy.Subscriber('/turtle/4/forward', Int32, lambda data : forward_hd(data, 3))\nrospy.Subscriber('/turtle/4/turn', Int32, lambda data : turn_hd(data, 3))\nrospy.Subscriber('/turtle/5/forward', Int32, lambda data : forward_hd(data, 4))\nrospy.Subscriber('/turtle/5/turn', Int32, lambda data : turn_hd(data, 4))\n\n'''\npub1 = rospy.Publisher('/turtle/position', Int32MultiArray, queue_size = 10)\npub2 = rospy.Publisher('/turtle/heading', Float32, queue_size = 10)\n\nrospy.Subscriber('/turtle/forward', Int32, forward_hd)\nrospy.Subscriber('/turtle/turn', Int32, turn_hd)\n'''\n\nprint(123)\n\nwhile not rospy.is_shutdown():\n time.sleep(0.1)\n for i in range(5):\n tur[i].forward(forward_value[i])\n tur[i].right(turn_value[i])\n pos = tur[i].position()\n pos = [pos[0], pos[1]]\n pub_data = Int32MultiArray(data = pos)\n pub1[i].publish(pub_data)\n pub2[i].publish(tur[i].heading())\n forward_value[i] = 0\n turn_value[i] = 0\n","sub_path":"src/my_turtle.py","file_name":"my_turtle.py","file_ext":"py","file_size_in_byte":2625,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"498427605","text":"'''Hugging Face'''\n\nimport streamlit as st\nfrom transformers import pipeline\n\nif __name__ == \"__main__\":\n\n # Define the title of the and its description\n st.title(\"Answering questions using NLP through Streamlit interface\")\n st.write(\"Pose questions, get answers\")\n\n # Load file\n \n raw_text = st.text_area(label=\"Enter a text here\")\n if raw_text != None and raw_text != '':\n\n # Display text\n with st.expander(\"Show question\"):\n st.write(raw_text)\n\n # Conduct question answering using the pipeline\n question_answerer = pipeline('question-answering')\n\n answer = ''\n question = st.text_input('Ask a question')\n\n if question != '' and raw_text != '':\n answer = question_answerer({\n 'question': question,\n 'context': raw_text\n })\n\n st.write(answer)\n","sub_path":"Streamlit and Machine Learning/hugginface.py","file_name":"hugginface.py","file_ext":"py","file_size_in_byte":884,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"201097762","text":"from selenium.webdriver.common.keys import Keys\r\nfrom .base import FunctionalTest\r\n\r\n\r\nclass ItemValidationTest(FunctionalTest):\r\n def get_error_element(self):\r\n \"\"\"\r\n Finds the error message via .has-error class\r\n :return: Returns error msg node\r\n \"\"\"\r\n return self.browser.find_element_by_css_selector('.has-error')\r\n\r\n def test_cannot_add_empty_list_items(self):\r\n # edith goes to the home page and accidentally tries to submit an empty item, hits enter on empty input box\r\n self.browser.get(self.live_server_url)\r\n self.get_item_input_box().send_keys(Keys.ENTER)\r\n\r\n # home page refreshes, error message appears saying items cannot be blank\r\n self.wait_for(lambda: self.browser.find_element_by_css_selector(css_selector='#id_text:invalid'))\r\n\r\n # she tries again with text, and it works\r\n self.get_item_input_box().send_keys('Buy milk')\r\n self.wait_for(lambda: self.browser.find_element_by_css_selector(css_selector='#id_text:valid'))\r\n\r\n # she submits it successfully\r\n self.get_item_input_box().send_keys(Keys.ENTER)\r\n self.wait_for_row_in_list_table('1: Buy milk')\r\n\r\n # she tries to submit another empty text\r\n self.get_item_input_box().send_keys(Keys.ENTER)\r\n\r\n # receives similar warning\r\n self.wait_for(lambda: self.browser.find_element_by_css_selector(css_selector='#id_text:invalid'))\r\n\r\n # she submits some text\r\n self.get_item_input_box().send_keys('Make tea')\r\n self.wait_for(lambda: self.browser.find_element_by_css_selector(css_selector='#id_text:valid'))\r\n self.get_item_input_box().send_keys(Keys.ENTER)\r\n\r\n # she sees her items on the list\r\n self.wait_for_row_in_list_table('1: Buy milk')\r\n self.wait_for_row_in_list_table('2: Make tea')\r\n\r\n def test_cannot_add_duplicate_item(self):\r\n # go to web page\r\n self.browser.get(self.live_server_url)\r\n\r\n # find the input box and enter in text\r\n self.add_list_item('Buy wellies')\r\n # self.get_item_input_box().send_keys('Buy wellies')\r\n # self.get_item_input_box().send_keys(Keys.ENTER)\r\n\r\n # wait for page to reload and check text in row\r\n # self.wait_for_row_in_list_table('1: Buy wellies')\r\n\r\n # find the input box and enter in the same text again\r\n self.get_item_input_box().send_keys('Buy wellies')\r\n self.get_item_input_box().send_keys(Keys.ENTER)\r\n\r\n # page returns error\r\n self.wait_for(lambda : self.assertEqual(\r\n self.get_error_element().text,\r\n second=\"You've already got this in your list\"\r\n ))\r\n\r\n def test_error_messages_are_cleared_on_input(self):\r\n \"\"\"\r\n Test to see if error messages are cleared upon text input\r\n :return: Pass or fail\r\n \"\"\"\r\n # get page and send one item, then wait for it to appear\r\n self.browser.get(self.live_server_url)\r\n self.add_list_item('Banter too thick')\r\n # self.get_item_input_box().send_keys('Banter too thick')\r\n # self.get_item_input_box().send_keys(Keys.ENTER)\r\n # self.wait_for_row_in_list_table('1: Banter too thick')\r\n\r\n # send same item again and wait for message\r\n self.get_item_input_box().send_keys('Banter too thick')\r\n self.get_item_input_box().send_keys(Keys.ENTER)\r\n self.wait_for(lambda: self.assertTrue(\r\n self.get_error_element().is_displayed()\r\n ))\r\n\r\n # start typing to clear error\r\n self.get_item_input_box().send_keys('a')\r\n\r\n # error no longer is there\r\n self.wait_for(lambda :self.assertFalse(\r\n self.get_error_element().is_displayed()\r\n ))\r\n\r\n# removed because using django LiveTestCase runner to run these tests\r\n# if __name__ == '__main__':\r\n# unittest.main(warnings='ignore')\r\n","sub_path":"functional_tests/test_list_item_validation.py","file_name":"test_list_item_validation.py","file_ext":"py","file_size_in_byte":3904,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"543097378","text":"'''\nCreated on Feb 22, 2017\nUpdated on Feb 22, 2017\n\n@author: mitch_000\n\n@todo: \ncommands\nvideo duration\nmp3 metadata\n ~pytaglib\n'''\n\nimport tkinter as tk\nfrom tkinter import ttk\nimport sys\nfrom socket import *\nimport error\nfrom requests import request\nfrom threading import Thread\n\ndef sendLog():\n log = open(\"log.dat\" , \"r+\")\n clientsocket = socket(AF_INET, SOCK_STREAM)\n serverIPs = open(\"serverIPs.dat\",\"r\")\n \n for serverIP in serverIPs:\n try:\n clientsocket.connect(str(serverIP),9700)\n clientsocket.send(log)\n clientsocket.close()\n log.write(\"\")\n except:\n pass\n \n \ndef updateServerIPs():\n \n clientsocket = socket(AF_INET, SOCK_STREAM)\n serverIPs = open(\"serverIPs.dat\",\"r+\")\n \n for serverIP in serverIPs:\n try:\n clientsocket.connect((str(serverIP),9704))\n break\n except:\n error.updateError()\n continue\n \n IPData = clientsocket.recv(1023)\n clientsocket.close()\n IPData = str(IPData)\n if(len(IPData) > 0 and ('209.203.209.131' in IPData)):\n serverIPs.write(IPData)\n serverIPs.close()\n \n\ndef checkUpdate():\n clientsocket = socket(AF_INET, SOCK_STREAM)\n serverIPs = open(\"serverIPs.dat\",\"r\")\n \n for serverIP in serverIPs:\n try:\n clientsocket.connect((str(serverIP),9701))\n \n except:\n error.updateError()\n continue\n\n officialVersion = clientsocket.recv(4096)\n clientsocket.close()\n if(version == str(officialVersion)):\n return\n else:\n return notifyUpdate( server)\n\n\ndef notifyUpdate(server):\n \n updateError()\n notwin = tk.Toplevel()\n \n notwin.wm_title(\"Update Available\")\n \n checkLabel = ttk.Label(notwin,text = \"an update is available \\nwould you like to update?\", foreground = \"green\")\n checkLabel.grid(row = 0 , column = 0 , sticky = 'w')\n \n yesButton = ttk.Button(notwin, command = lambda: update(notwin , win , server), text = \"Yes\")\n yesButton.grid(row = 1 , column = 1 , sticky = 'w')\n \n noButton = ttk.Button(notwin , command = notwin.destroy() , text = \"no\")\n noButton.grid(row = 1 , column = 0)\n notwin.mainloop()\n \n\ndef update(notwin , server ):\n notwin.destroy()\n win.destroy()\n os.chdir(programDir)\n \n try:\n clientsocket = socket(AF_INET, SOCK_STREAM)\n clientsocket.connect(server,9702)\n newExe = clientsocket.recv(4096)\n File = open(\"youtubeDownloader.exe\",'w')\n File.write(newExe)\n File.close()\n os.sys.exit()\n except:\n error.updateError()\n \n \ndef checkConnection():\n '''\n try:\n print(\"checking connection\")\n request.urlopen('http://216.58.192.142', timeout=1)\n return\n except: \n e = error.connectionError()\n Thread( target = e.displayError ).start()\n ''' \n ","sub_path":"src/client/model/update.py","file_name":"update.py","file_ext":"py","file_size_in_byte":2987,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"150891748","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\"\"\"Contains the explicit_general solver.\n\nUsed to compute thermal processes\n\n\"\"\"\n\nimport numpy as np\nimport copy\n\n\ndef explicit_general(obj):\n \"\"\"explicit_general solver.\n\n Used to compute one time step of systems with fixed thermal contuctivity.\n\n \"\"\"\n\n x = copy.copy(obj.temperature)\n\n # computes\n for i in range(1, obj.num_points - 1):\n\n alpha = obj.dt * \\\n obj.k[i] / (obj.rho[i] * obj.Cp[i] *\n obj.dx * obj.dx)\n beta = obj.dt / (obj.rho[i] * obj.Cp[i])\n\n Tnew = ((1 + beta * obj.Q[i]) * obj.temperature[i][0] +\n alpha * (obj.temperature[i - 1][0] - 2 *\n obj.temperature[i][0] + obj.temperature[i + 1][0]) +\n beta * (obj.Q0[i] - obj.Q[i] * obj.amb_temperature))\n x[i][1] = Tnew\n\n # left boundary for next time step\n if obj.boundaries[0] == 0:\n x[0][1] = obj.temperature[1][1]\n else:\n x[0][1] = obj.boundaries[0]\n\n # right boundary for next time step\n if obj.boundaries[1] == 0:\n x[obj.num_points - 1][1] = obj.temperature[obj.num_points - 2][1]\n else:\n x[obj.num_points - 1][1] = obj.boundaries[1]\n\n # updates temperature for next iteration\n for i in range(0, obj.num_points):\n x[i][0] = x[i][1]\n\n return x\n","sub_path":"heatrapy/solvers/explicit_general.py","file_name":"explicit_general.py","file_ext":"py","file_size_in_byte":1371,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"341207075","text":"\"\"\" class to use as a general purpose GUI input widget\r\n\r\nconvenience functions for validating data inputs include\r\nfloat_pair()\r\nlist_of_floats()\r\nyesno()\r\n\r\nversion 2.0.0 - adds output-only type, renamed internal functions\r\nversion 2.0.1 - adds yesno function, and handling for bool\r\n\"\"\"\r\n\r\n\"\"\"\r\nCopyright (c) <2016>, , , \r\nAll rights reserved.\r\n\r\nRedistribution and use in source and binary forms, with or without\r\nmodification, are permitted provided that the following conditions are met:\r\n\r\n1. Redistributions of source code must retain the above copyright notice, this\r\n list of conditions and the following disclaimer. \r\n2. Redistributions in binary form must reproduce the above copyright notice,\r\n this list of conditions and the following disclaimer in the documentation\r\n and/or other materials provided with the distribution.\r\n\r\nTHIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\" AND\r\nANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED\r\nWARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE\r\nDISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR\r\nANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES\r\n(INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;\r\nLOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND\r\nON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT\r\n(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS\r\nSOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.\r\n\r\nThe views and conclusions contained in the software and documentation are those\r\nof the authors and should not be interpreted as representing official policies, \r\neither expressed or implied, of the FreeBSD Project.\r\n\"\"\"\r\n\r\n# works with Python 3.4\r\n# should work with Python 2.7\r\n\r\ntry:\r\n # Python 3 spelling\r\n import tkinter as tki\r\n import tkinter.messagebox as tkm\r\n import tkinter.filedialog as tkf\r\nexcept ImportError:\r\n # Python 2 spelling\r\n import Tkinter as tki\r\n import tkMessageBox as tkm\r\n import tkFileDialog as tkf\r\nimport json\r\n\r\nversion = '2.0.1 2016-Dec-11'\r\n\r\nclass LabelLine(tki.Frame):\r\n \"\"\" a combination of label and label \"\"\"\r\n def __init__(self, parent, text='no label', data='', width=15):\r\n \"\"\" text optional used to label the entry\r\n \r\n data optional used to initialise the label box\r\n uses empty string if none supplied\r\n \r\n width optional defaults to 15, maybe should make this adaptive sometime\r\n\r\n \"\"\"\r\n \r\n tki.Frame.__init__(self, master=parent) # tki refuses to work with super!\r\n self.text = text\r\n self.data = data\r\n \r\n # do the label\r\n self.label = tki.Label(self, text=text, width=width)\r\n self.label.grid(row=0, column=0)\r\n\r\n # do the data label\r\n self.data_label = tki.Label(self, text=self.data, width=width)\r\n self.data_label.grid(row=0, column=1)\r\n\r\n def put(self, value):\r\n \"\"\" allows the client to change the displayed value, using str(object)\"\"\"\r\n self.data_label.config(text=str(value))\r\n \r\n\r\nclass EntryLine(tki.Frame):\r\n \"\"\" a combination of label, entry and help button, for validated gui entry of data\"\"\"\r\n def __init__(self, parent, text='no label', data='', conv=None, update=None, width=15, default=None):\r\n \"\"\" text optional used to label the entry\r\n \r\n data optional used to initialise the Entry box\r\n uses empty string if none supplied\r\n \r\n conv optional conversion function, which also validates the entry\r\n note that int, float, bool and str can be used\r\n return string unchanged (str) if omitted or None\r\n conv must return object if entry is valid\r\n its __doc__ string is available as help\r\n if the docstring starts with [x], x is put on the help button\r\n (TODO have to work on tooltips sometime!)\r\n raise ValueError if the entry is invalid (int, float and str do this)\r\n The err from ValueError is saved for the help button\r\n so more info can be given about failure to validate\r\n\r\n update optional the name of the function to call when entries are changed\r\n needed for the 'execute when all valid' functionality\r\n\r\n width optional defaults to 15, maybe should make this adaptive sometime\r\n\r\n default optional defaults to None\r\n returns if the text string produces an invalid result\r\n allows .get() to be called asynchronously, and return\r\n results other than None to the calling program\r\n\r\n \"\"\"\r\n tki.Frame.__init__(self, master=parent) # tki refuses to work with super!\r\n self.update = update\r\n self.conv = conv\r\n self.text = text\r\n self.default = default\r\n \r\n # init the properties\r\n self.err = ''\r\n self.value = None\r\n self.valid = False\r\n \r\n # do the label\r\n self.label = tki.Label(self, text=text, width=width)\r\n self.label.grid(row=0, column=0)\r\n\r\n # do the conversion function\r\n cdoc = conv.__doc__ # easier to type\r\n if conv:\r\n # we have a conversion function specified\r\n # is it one of the builtins?\r\n if conv == int:\r\n help_face = 'i'\r\n self.conv_help = 'builtin int() function'\r\n elif conv == float:\r\n help_face = 'f'\r\n self.conv_help = 'builtin float() function'\r\n elif conv == str:\r\n help_face = 'str'\r\n self.conv_help = 'builtin str() function'\r\n elif conv == bool:\r\n help_face = 'bool'\r\n self.conv_help = 'builtin bool() function'\r\n else:\r\n # none of those, so does it have a docstring?\r\n if cdoc:\r\n # yes, does it start with a help_face?\r\n face_end = cdoc.find(']')\r\n if (cdoc[0] == '[') and (face_end != -1) and (face_end <= 6):\r\n help_face = cdoc[1:face_end]\r\n else:\r\n help_face = '?'\r\n # is the help prompt truncated in the docstring?\r\n if '[end_help]' in cdoc:\r\n self.conv_help = cdoc[:cdoc.find('[end_help]')]\r\n else:\r\n self.conv_help = cdoc\r\n # this could be done by knowing .find() returns -1 if not found\r\n # and slicing [:find()] regardless\r\n # but that's bordering on the tricky, this is clearer\r\n else:\r\n self.conv_help = 'no documentation\\navailable for\\nthis conversion'\r\n help_face = '?'\r\n else:\r\n self.conv = str\r\n help_face = 'str'\r\n self.conv_help = 'returned as string'\r\n\r\n # do the entry\r\n self.var = tki.StringVar()\r\n self.entry = tki.Entry(self, textvariable=self.var, width=width)\r\n self.entry.grid(row=0, column=1)\r\n self.var.trace('w', self._changed)\r\n self.entry.bind('', self._returned)\r\n\r\n # do the help button\r\n self.help_but = tki.Button(self,\r\n text=help_face,\r\n command=self._show_help,\r\n width=5,\r\n takefocus=0) # don't take part in tab-focus\r\n # it took a while to figure this out\r\n self.help_but.grid(row=0, column=2)\r\n\r\n # initialise it, which triggers the trace, _changed and validation\r\n self.var.set(str(data))\r\n\r\n def _returned(self, *args):\r\n self.update(enter=True)\r\n\r\n def _show_help(self):\r\n tkm.showinfo('conversion information', '{}\\n\\n{}'.format(self.conv_help, self.err))\r\n\r\n def _changed(self, *args):\r\n ent_val = self.var.get()\r\n self.val_string = ent_val\r\n try:\r\n self.value = self.conv(ent_val)\r\n self.entry.config(bg='white')\r\n self.err = ''\r\n self.valid = True\r\n except ValueError as err:\r\n try:\r\n self.value = self.conv(self.default) # we convert the default value\r\n except (TypeError, ValueError):\r\n self.value = self.default # which if it can't be converted (None) is returned intact\r\n self.entry.config(bg='orange')\r\n self.err = err\r\n self.valid = False\r\n self.update()\r\n\r\n def put(self, value):\r\n \"\"\" allows the client to change the displayed value\"\"\"\r\n self.var.set(value)\r\n\r\nclass GUI_inputs(tki.LabelFrame):\r\n \"\"\" A GUI data input convenience class, with tab-able fields and verified data\"\"\"\r\n def __init__(self, parent, text=\"Neil's Input Widget\", execute=None,\r\n exec_label='execute',\r\n loadsave=None,\r\n **kwargs):\r\n \"\"\" initialise with text for the LabelFrame\r\n\r\n set execute to the name of a function to be called on execute\r\n execute button greyed out until all entries are valid\r\n \r\n set loadsave=True to put up load/save buttons\r\n \"\"\"\r\n tki.LabelFrame.__init__(self, master=parent, text=text)\r\n self.kwargs = kwargs \r\n\r\n # we have a dict of entries\r\n self.entries = {} # the data entry widgets\r\n self.labels = {} # the label (output only) widgets\r\n self.row = 0\r\n \r\n # if there's a execute supplied, put up a button for it, on the last row\r\n self.execute_func = execute\r\n self.exec_label = exec_label\r\n if execute:\r\n # an execute button\r\n self.execute_but = tki.Button(self, text=self.exec_label,\r\n command=self.execute_func,\r\n state=tki.DISABLED)\r\n self.execute_but.grid(row=99, column=1) #MAXROWS anyone?\r\n # a tick box for the enter binding\r\n self.exec_ent_var = tki.IntVar()\r\n self.exec_check = tki.Checkbutton(self, text='exec on enter', variable=self.exec_ent_var)\r\n self.exec_check.grid(row=99, column=0)\r\n\r\n # if we want load/save functionality, True for current path\r\n if loadsave:\r\n self.load_but = tki.Button(self, text='load', command=self._load_func)\r\n self.load_but.grid(row=100, column=0)\r\n self.save_but = tki.Button(self, text='save', command=self._save_func)\r\n self.save_but.grid(row=100, column=1)\r\n \r\n\r\n def add(self, key, disp_name='', data='', conv=None, default=None, output_only=False):\r\n \"\"\" add a new data entry line to the input widget\r\n\r\n key required key for the entry, must be unique on this widget\r\n the returned dict uses this as the key\r\n use a string, but other objects do work as well\r\n \r\n disp_name optional labels the data entry, uses the key if omitted\r\n\r\n data optional initial string data for the entry box\r\n\r\n conv optional A function, that takes a string, returns an object\r\n or raises ValueError if it can't understand the string\r\n int and float do this\r\n\r\n default optional When GUI_inputs is the client, execute is always greyed\r\n if there are any invalid entries. However as a server,\r\n .get() might be called when entries are invalid. Default\r\n provides a default response for invalid entries, to avoid\r\n the calling program having to try: all return values\r\n \r\n output_only optional False (default) allows input and output, True switches off input\r\n \"\"\"\r\n\r\n\r\n\r\n if not disp_name:\r\n disp_name = str(key)\r\n\r\n if output_only:\r\n if key in self.labels:\r\n raise ValueError('duplicate key name >>>{}<<<'.format(key)) \r\n mle = LabelLine(self, disp_name, data)\r\n self.labels[key] = mle\r\n else:\r\n if key in self.entries:\r\n raise ValueError('duplicate key name >>>{}<<<'.format(key)) \r\n mle = EntryLine(self, disp_name, data, conv, self.update, default=default, **self.kwargs)\r\n self.entries[key] = mle\r\n \r\n mle.grid(row=self.row, column=0, columnspan=2)\r\n self.row += 1\r\n\r\n \r\n\r\n def _load_func(self):\r\n \"\"\" read a json encoded list of tuples\r\n update the GUI's entry fields with the data\r\n warn if there are too many or too few\r\n If a tuple contains only two data, assume to be key and value\"\"\"\r\n\r\n load_name = tkf.askopenfilename(filetypes = [('JSON files', '.json'), ('all files', '.*')])\r\n if load_name:\r\n with open(load_name, 'rt') as load_file:\r\n updates = json.load(load_file)\r\n\r\n # make a dict with what to update, and what to update it with\r\n # maybe there should be more error checking here\r\n # but printing out what gets used and not, into a gui, gives you a chance\r\n try:\r\n src_keys = [x[0] for x in updates]\r\n data = [x[-1] for x in updates]\r\n except IndexError:\r\n print(\"can't understand the load data file, are all fields present?\")\r\n return\r\n src_dict = dict(zip(src_keys, data))\r\n \r\n dst_keys = self.entries.keys()\r\n \r\n # using sets to do this is a bit more elegant and explicit than\r\n # looping and testing over both sets of keys\r\n can_update = set(src_keys) & set(dst_keys)\r\n not_updated = set(dst_keys).difference(src_keys)\r\n not_used = set(src_keys).difference(dst_keys)\r\n\r\n for key in can_update:\r\n self.set_data(key, src_dict[key])\r\n print('\"{}\" was updated to \"{}\"'.format(key, src_dict[key]))\r\n for key in not_updated:\r\n print('Warning - \"{}\" found on GUI but not in load file, not updated'.format(key))\r\n for key in not_used:\r\n print('Warning - \"{}\" found in load file but not on GUI, not used'.format(key))\r\n\r\n def _save_func(self):\r\n \"\"\" put a list of (key, display_name, value_string) tuples\r\n into a json encoded file, or into the shell window if no file is specified\r\n\r\n 'key' is effectively the variable name used in the code\r\n 'display_name' is for documentation on save, ignored on load\r\n 'value_string' is whatever is present, whether valid or not\"\"\"\r\n\r\n # retrieve and format the data\r\n keys = self.entries.keys()\r\n names = [x.text for x in self.entries.values()]\r\n data = [x.val_string for x in self.entries.values()]\r\n save_stuff = list(zip(keys, names, data)) # zip returns an iterator in python3!\r\n\r\n save_name = tkf.asksaveasfilename(filetypes = [('JSON files', '.json'), ('all files', '.*')])\r\n if save_name: \r\n if not ('.' in save_name):\r\n save_name += '.json'\r\n with open(save_name, 'wt') as save_file:\r\n json.dump(save_stuff, save_file)\r\n else:\r\n print(save_stuff)\r\n \r\n\r\n\r\n def update(self, enter=False):\r\n \"\"\" called when something has changed, or enter has been hit\r\n this is a clumsy interface, not sure its well thought out\r\n in fact it confused me when I returned to the code\r\n but now I think I know what's going on\"\"\"\r\n # only need to worry about this when there's a execute button to handle\r\n if self.execute_func:\r\n # get the valid properties of each entry\r\n valids = [x.valid for x in self.entries.values()]\r\n if all(valids):\r\n self.execute_but.config(state=tki.NORMAL)\r\n if enter and (self.exec_ent_var.get() == 1):\r\n self.execute_func()\r\n else:\r\n self.execute_but.config(state=tki.DISABLED)\r\n \r\n\r\n def get(self):\r\n \"\"\" return a dict of the converted results\"\"\"\r\n output = dict(zip(self.entries.keys(), [x.value for x in self.entries.values()]))\r\n return output\r\n \r\n \r\n def set_data(self, key, data):\r\n \"\"\" set the field 'key' to data\r\n note as we try the outputs first\r\n if any key is duplicated between entry and output\r\n only the output can be set\"\"\"\r\n try:\r\n self.labels[key].put(data)\r\n except KeyError:\r\n self.entries[key].put(data)\r\n # don't catch any exception from here, let it burn through\r\n\r\n# conversion functions, for example, and to be used by the application \r\n\r\ndef float_pair(x):\r\n \"\"\"[f,f] Two floats seperated by a comma\r\n [end_help]\r\n\r\n example non-trivial conversion function\r\n not all of docstring intended to be displayed as help\r\n throw ValueError from two locations, one from split, one from float\r\n return a list of the values\r\n \"\"\"\r\n fields = x.split(',')\r\n if len(fields) != 2:\r\n raise ValueError('need two fields seperated by one comma')\r\n output = []\r\n for field in fields: # float() will ValueError if it's wrong \r\n output.append(float(field))\r\n return output\r\n\r\ndef list_of_floats(x):\r\n \"\"\"[lof] list of floats\r\n One float, or several floats separated by commas\"\"\"\r\n\r\n # try to eliminate the simplest problem\r\n if ',' in x:\r\n if x.rstrip()[-1] == ',':\r\n raise ValueError('no final value')\r\n\r\n out = []\r\n fields = x.split(',') # this will always work without error\r\n for field in fields:\r\n out.append(float(field)) # and float() will ValueError if it \r\n return out # doesn't understand the string\r\n\r\ndef yesno(x):\r\n \"\"\"[yesno] all or sufficient part of any of the words true, false, yes, no, 0, 1, OK\"\"\"\r\n if len(x)==0:\r\n raise ValueError('no answer')\r\n\r\n is_true = False\r\n is_false = False\r\n \r\n for c in x.lower():\r\n if c in 'truy1k': # spots TRUe, Yes 1, oK\r\n is_true = True\r\n if c in 'faln0': # spots FALse, No, 0\r\n is_false = True\r\n # e, s, o are ambiguous\r\n\r\n if is_true == is_false:\r\n raise ValueError('ambiguous answer')\r\n return is_true\r\n \r\n \r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\nif __name__ == '__main__':\r\n\r\n def execute_func():\r\n # get data in converted form from the widgets\r\n print('executing with')\r\n fg = full.get()\r\n bg = basic.get()\r\n print(fg, bg)\r\n # show updating data on the forms\r\n try:\r\n execute_func.count += 1\r\n except Exception:\r\n execute_func.count = 0\r\n basic.set_data('key 1', '{}'.format(execute_func.count))\r\n num1 = fg['pair'][0]\r\n num2 = fg['pair'][1]\r\n basic.set_data('output', num1*num2)\r\n\r\n # define additional conversion functions in the calling application\r\n\r\n def int16(x):\r\n \"\"\"[16] a base16 (hexadecimal) integer\"\"\"\r\n return int(x, base=16)\r\n\r\n def cryptic_conv(x):\r\n # there happens to be no docstring for this conversion function\r\n return int(x)\r\n \r\n root = tki.Tk()\r\n \r\n basic = GUI_inputs(root)\r\n basic.pack()\r\n basic.add('key 1')\r\n basic.add('key 2')\r\n basic.add('output', output_only=True, data='init data')\r\n \r\n full = GUI_inputs(root, 'full fat', execute=execute_func, loadsave=True, width=20)\r\n full.pack()\r\n full.add('examp 1')\r\n full.add('we1', conv=str, data=3)\r\n full.add('an int', conv=int, data=13, default=7)\r\n full.add('we2', 'disp4we2', data=999)\r\n full.add('----', output_only=True, data = '---')\r\n full.add('pair', 'f_pair', '3,4', float_pair, default='7,8' )\r\n full.add('adr', 'hex address', '0xC0DE', int16)\r\n full.add('float_list', 'float list', '3, 4, 5', list_of_floats, )\r\n full.add('cryp', 'no doc string', 6, cryptic_conv)\r\n full.add('boolean', conv=bool)\r\n full.add('yes or no', conv=yesno)\r\n\r\n get_but = tki.Button(root, text='force get', command=execute_func)\r\n get_but.pack()\r\n\r\n root.mainloop()\r\n \r\n \r\n","sub_path":"gui_io_widget.py","file_name":"gui_io_widget.py","file_ext":"py","file_size_in_byte":21407,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"361591251","text":"#!/usr/bin/env python\n\n\"\"\"\nGeneraList lets you use Amazon's Alexa to create and use lists that hold any type of data.\n\nAsk GeneraList to store grocery items as you think of them during the week and then access those items when\nyou're on the go at the supermarket. Or give generalist a list of instructions that Alexa can read off to\nyou later as you ask her to. Have a recipe that you love? Let GeneraList read the instructions to you as\nyou're cooking.\n\nTo add: Music play lists, Video play lists, lists of slides for a slide show. These will require Alexa to\ninteract with other web services.\n\"\"\"\n\nfrom __future__ import print_function\n\nimport boto3\nimport botocore\nimport botocore.exceptions\n\n__author__ = 'Mike Lane'\n__email__ = 'mikelane@gmail.com'\n\n# Set APP_ID to ensure only your skill calls this lambda handler\nAPP_ID = 'amzn1.ask.skill.e208302c-710b-4f30-9344-d0ae6ca3b774'\nSKILL_NAME = 'GeneralList'\nSKILL_INVOKE = 'generalist'\nDB_REGION = 'us-east-1'\nDB_URL = \"https://dynamodb.{}.amazonaws.com\".format(DB_REGION)\n\n# Global session information\nstored_session = {\n 'current_list': None,\n 'current_step': None\n}\nSESSION_TABLENAME = 'StoredSession'\nLISTS_TABLENAME = 'Lists'\n\n\n# --------------- Main handler ------------------\n\ndef lambda_handler(event, context):\n \"\"\"Handles the Launch|Intent|SessionEnded request event\n Returns a JSON response to Alexa with:\n outputSpeech, reprompt, card, shouldEndSession\"\"\"\n\n if (event['session']['application']['applicationId']\n != 'amzn1.ask.skill.e208302c-710b-4f30-9344-d0ae6ca3b774'):\n if APP_ID != '':\n if event['session']['application']['applicationId'] != APP_ID:\n raise ValueError('Invalid Application ID',\n event['session']['application']['applicationId'],\n APP_ID)\n\n if event['session']['new']:\n on_session_started({'requestId': event['request']['requestId']}, event['session'])\n\n if event['request']['type'] == 'LaunchRequest':\n return on_launch(event['request'], event['session'])\n elif event['request']['type'] == 'IntentRequest':\n return on_intent(event['request'], event['session'])\n elif event['request']['type'] == 'SessionEndedRequest':\n return on_session_ended(event['request'], event['session'])\n\n\n# --------------- Helpers that build all of the responses ----------------------\n\ndef build_speechlet_response(title, output, reprompt_text, should_end_session):\n return {\n 'outputSpeech': {\n 'type': 'PlainText',\n 'text': output\n },\n 'card': {\n 'type': 'Simple',\n 'title': 'SessionSpeechlet - ' + title,\n 'content': 'SessionSpeechlet - ' + output\n },\n 'reprompt': {\n 'outputSpeech': {\n 'type': 'PlainText',\n 'text': reprompt_text\n }\n },\n 'shouldEndSession': should_end_session\n }\n\n\ndef build_response(session_attributes, speechlet_response):\n return {\n 'version': '0.2',\n 'sessionAttributes': session_attributes,\n 'response': speechlet_response\n }\n\n\n# --------------- Functions that control the skill's behavior ------------------\n\ndef get_welcome_response(session):\n \"\"\" If we wanted to initialize the session to have some attributes we could add those here\"\"\"\n session_attributes = session.get('attributes', {})\n card_title = 'Welcome to GeneraList'\n\n print(\"***GET_WELCOME_RESPONSE, session: {}\".format(session_attributes))\n\n if session_attributes['currentList'] != \"NONE\":\n speech_output = \"Welcome back. Go to the next item in your {} list by saying \" \\\n \"'next'.\".format(session_attributes['currentList'])\n reprompt_text = \"To continue with your {} list say: 'next', 'repeat', 'review', 'preview' or \" \\\n \"'edit'. Say: 'load' or 'create' to work with a \" \\\n \"different list.\".format(session_attributes['currentList'])\n else:\n speech_output = \"Welcome to generalist. Do you want to: 'load', 'create', or 'edit' a list?\"\n reprompt_text = \"Say: 'load', 'create', or 'edit'.\"\n\n should_end_session = False\n return build_response(session_attributes=session_attributes,\n speechlet_response=build_speechlet_response(title=card_title,\n output=speech_output,\n reprompt_text=reprompt_text,\n should_end_session=should_end_session))\n\n\ndef get_help_response(session):\n \"\"\"If the user asks for help, smartly determine what help they want by looking at their session values.\"\"\"\n session_attributes = session.get('attributes')\n should_end_session = True\n card_title = \"Welcome to GeneraList\"\n reprompt_text = \"\"\n\n print(\"***GET HELP RESPONSE, session: {}\".format(session_attributes))\n\n if session_attributes['currentTask'] == 'PLAY':\n speech_output = \"You currently have a list playback session in progress. To hear the next item \" \\\n \"in your list, say: 'next'. \"\n if session_attributes['currentStep'] > 1:\n speech_output += \"To go back to the previous item in your list, say 'previous'. \"\n speech_output += \"Or you can say: 'stop', 'cancel', 'load', or 'create'.\"\n elif session_attributes['currentTask'] == 'CREATE':\n speech_output = \"You are creating a list. To add a new item, say: 'tell generalist to add item'. \" \\\n \"To save your list say: 'tell generalist to save'. To cancel and lose your list say: \" \\\n \"'tell generalist cancel'.\"\n elif session_attributes['currentTask'] == 'EDIT':\n speech_output = \"You are editing a list. To add a new item, say: 'add' and the next item. To save your list, \" \\\n \"say: 'tell generalist to save'.\"\n else:\n return get_welcome_response(session=session)\n\n return build_response(session_attributes=session_attributes,\n speechlet_response=build_speechlet_response(title=card_title,\n output=speech_output,\n reprompt_text=reprompt_text,\n should_end_session=should_end_session))\n\n\ndef handle_save_intent(session):\n card_title = 'Save'\n should_end_session = True\n session_attributes = session.get('attributes', {})\n\n print(\"***HANDLE SAVE INTENT, session: {}\".format(session_attributes))\n\n # If the session is in create or edit mode, update the list, and set the stored status accordingly\n if session_attributes['currentTask'] in ['CREATE', 'EDIT']:\n # Update the list to start at the first step when loaded.\n session['attributes']['currentStep'] = 0\n update_list(session=session)\n\n speech_output = \"Your list {lst} has been saved and loaded. \" \\\n \"Play it by saying: 'next'.\".format(lst=session_attributes['currentList'])\n reprompt_text = \"\"\n\n # Take us out of create mode\n session['attributes']['currentTask'] = \"PLAY\"\n\n # Update the session information\n update_session(session=session)\n\n else:\n speech_output = \"Nothing to save\"\n reprompt_text = \"\"\n\n return build_response(session_attributes=session['attributes'],\n speechlet_response=build_speechlet_response(title=card_title,\n output=speech_output,\n reprompt_text=reprompt_text,\n should_end_session=should_end_session))\n\n\ndef handle_session_stop_request(session):\n card_title = 'Stop'\n should_end_session = True\n session_attributes = session.get('attributes', {})\n\n print(\"***HANDLE SESSION STOP REQUEST, session: {}\".format(session_attributes))\n\n # If the session is in create or edit mode, update the list, and set the stored status accordingly\n if session_attributes['currentTask'] in ['CREATE', 'EDIT']:\n # Update the list to start at the first step when loaded.\n session['attributes']['currentStep'] = 0\n update_list(session=session)\n\n speech_output = \"Your list {lst} has been saved and loaded. \" \\\n \"Play it by saying: 'next'.\".format(lst=session_attributes['currentList'])\n reprompt_text = \"\"\n\n # Take us out of create mode\n session['attributes']['currentTask'] = \"PLAY\"\n\n # Update the session information\n update_session(session=session)\n\n else:\n speech_output = \"Goodbye.\"\n reprompt_text = \"\"\n\n return build_response(session_attributes=session['attributes'],\n speechlet_response=build_speechlet_response(title=card_title,\n output=speech_output,\n reprompt_text=reprompt_text,\n should_end_session=should_end_session))\n\n\ndef handle_session_cancel_request(session):\n # TODO smartly handle cancel requests.\n card_title = 'Canceled'\n session_attributes = session.get('attributes', {})\n\n print(\"***HANDLE SESSION CANCEL REQUEST, session: {}\".format(session_attributes))\n\n speech_output = \"\"\n reprompt_text = \"\"\n should_end_session = True\n\n # If in create mode, but NOT edit mode, delete the list.\n if session_attributes['currentTask'] == 'CREATE':\n lists_table = boto3.resource('dynamodb').Table(LISTS_TABLENAME)\n try:\n lists_table.delete_item(\n Key={'userId': session['user']['userId'],\n 'listName': session_attributes['listName']}\n )\n except botocore.exceptions.ClientError as e:\n print(\"ERROR: {}\".format(e.response))\n raise\n\n # Clear out the stored session\n session['attributes']['currentTask'] = \"NONE\"\n session['attributes']['currentList'] = \"NONE\"\n session['attributes']['currentStep'] = 0\n update_session(session=session)\n\n return build_response(session_attributes=session['attributes'],\n speechlet_response=build_speechlet_response(title=card_title,\n output=speech_output,\n reprompt_text=reprompt_text,\n should_end_session=should_end_session))\n\n\ndef create_list(intent, session):\n \"\"\"Use the user id to create a new list. Tell the user to simply speak the steps. After each step,\n validate and ask if they want to add another step or stop. If the list already exists, just error out\n with a message that tells the user to modify a list instead.\"\"\"\n card_title = intent['name']\n\n print(\"***CREATE LIST session: {}\".format(session['attributes']))\n print(\"***CREATE LIST intent: {}\".format(intent['slots']))\n\n if 'value' in intent['slots']['listName']:\n # First make sure we're not creating a list with the same name we're on\n if intent['slots']['listName']['value'] == session['attributes']['currentList']:\n speech_output = \"You can't create a new list with the same name as one of your current lists. \" \\\n \"Either choose a new name or delete the list with the same name.\"\n reprompt_text = \"\"\n should_end_session = True\n return build_response(session_attributes=session['attributes'],\n speechlet_response=build_speechlet_response(title=card_title,\n output=speech_output,\n reprompt_text=reprompt_text,\n should_end_session=should_end_session))\n # Next try and load the list with the desired list name from the db\n # If the list exists, force the user to delete it before they can create a new one with the same name\n else:\n table = boto3.resource('dynamodb').Table(LISTS_TABLENAME)\n try:\n response = table.get_item(Key={\n 'userId': session['user']['userId'],\n 'listName': intent['slots']['listName']['value']\n })\n except botocore.exceptions.ClientError as e:\n print(\"ERROR: create_list database failure: {}\".format(e.response))\n raise\n print(\"***CREATE LIST db response: {}\".format(response))\n if 'Item' not in response:\n # Otherwise, set the session attributes accordingly and create the list\n session['attributes']['currentList'] = intent['slots']['listName']['value']\n session['attributes']['currentTask'] = 'CREATE'\n session['attributes']['currentStep'] = 0\n session['attributes']['numberOfSteps'] = 0\n session['attributes']['listItems'] = {}\n update_session(session=session)\n speech_output = \"Creating a list named '{}'. \" \\\n \"Now say something like: \" \\\n \"'Add 4 large eggs.'\".format(session['attributes']['currentList'])\n reprompt_text = \"You can say: 'Add item,' \" \\\n \"or you can say something like: 'Add: Set oven to 300 degrees.'\"\n should_end_session = False\n else:\n speech_output = \"You can't create a new list with the same name as one of your current \" \\\n \"lists. Either choose a new name or delete the list with the same name.\"\n reprompt_text = \"\"\n should_end_session = True\n return build_response(session_attributes=session['attributes'],\n speechlet_response=build_speechlet_response(title=card_title,\n output=speech_output,\n reprompt_text=reprompt_text,\n should_end_session=should_end_session))\n else:\n speech_output = \"What do you want to name your list? Say 'create' and a list name.\"\n reprompt_text = \"In order to create a list, you must name it. Say: 'create' and then a name. For \" \\\n \"example say, 'Create brownie recipe.'\"\n should_end_session = False\n\n print(\"***END OF CREATE, session: {}\".format(session['attributes']))\n\n return build_response(session_attributes=session['attributes'],\n speechlet_response=build_speechlet_response(title=card_title,\n output=speech_output,\n reprompt_text=reprompt_text,\n should_end_session=should_end_session))\n\n\ndef edit_list(intent, session):\n \"\"\"Edit a list (currently only supports adding to the list).\"\"\"\n card_title = \"Edit the List\"\n\n print(\"***END LIST, session: {}\".format(session.get('attributes', {})))\n print(\"***END LIST, intent: {}\".format(intent['slots']))\n\n # Update the session and current list in case this gets called during the middle of a session\n update_session(session=session)\n update_list(session=session)\n\n # Editing something other than the current list\n if 'value' in intent['slots']['listName'] and \\\n intent['slots']['listName']['value'] != session['attributes']['currentList']:\n # Try to get the desired list from the database\n table = boto3.resource('dynamodb').Table(LISTS_TABLENAME)\n try:\n response = table.get_item(Key={\n 'userId': session['user']['userId'],\n 'listName': intent['slots']['listName']['value']\n })\n except botocore.exceptions.ClientError as e:\n print(\"ERROR: edit_list database get_item failed: {}\".format(e.response))\n raise\n else:\n if 'Item' in response: # Found the desired list, switch over to that one and go into edit mode\n session['attributes']['currentTask'] = 'EDIT'\n session['attributes']['currentList'] = response['Item']['listName']\n session['attributes']['currentStep'] = response['Item']['numberOfSteps']\n session['attributes']['listItems'] = response['Item']['listItems']\n session['attributes']['numberOfSteps'] = response['Item']['numberOfSteps']\n update_session(session=session)\n speech_output = \"Editing list {}. Say: 'add' and the next item.\".format(\n session['attributes']['currentList'])\n reprompt_text = \"You are currently editing {}. Say: 'add' and the next item to add to the \" \\\n \"list\".format(session['attributes']['currentList'])\n should_end_session = False\n else: # Didn't find the desired list in the database.\n speech_output = \"I didn't find a list named {}.\".format(intent['slots']['listName']['value'])\n reprompt_text = \"\"\n should_end_session = True\n # Desired list is current list\n elif intent['slots']['listName']['value'] == session['attributes']['currentList']:\n session['attributes']['currentTask'] = 'EDIT'\n session['attributes']['currentStep'] = session['attributes']['numberOfSteps']\n update_session(session=session)\n speech_output = \"Editing list {}. Say: 'add' and the next item.\".format(\n session['attributes']['currentList'])\n reprompt_text = \"You are currently editing {}. Say: 'add' and the next item to add to the \" \\\n \"list\".format(session['attributes']['currentList'])\n should_end_session = False\n # Desired list is unspecified\n else:\n speech_output = \"Tell me what list you want to edit. Say: 'edit' and the name of a list that \" \\\n \"you've already created. Or say: 'create' and a list name to create a new list.\"\n reprompt_text = \"Say: 'edit' and the name of a list that you've already created. Or say: 'create' \" \\\n \"and a list name to create a new list.\"\n should_end_session = False\n\n return build_response(session_attributes=session['attributes'],\n speechlet_response=build_speechlet_response(title=card_title,\n output=speech_output,\n reprompt_text=reprompt_text,\n should_end_session=should_end_session))\n\n\ndef add_item(intent, session):\n \"\"\"Adds an item to the end of a list.\"\"\"\n card_title = intent['name']\n session_attributes = session.get('attributes', {})\n\n print(\"***ADD ITEM. Session: {}\".format(session_attributes))\n print(\"***ADD ITEM. Intent: {}\".format(intent['slots']))\n\n if session_attributes['currentTask'] not in ['CREATE', 'EDIT']:\n # If not in create or edit mode, we can't add an item.\n should_end_session = True\n speech_output = \"I can't add a task if you're not in 'create' or 'edit' modes.\"\n reprompt_text = \"\"\n elif 'Item' in intent['slots'] and 'value' not in intent['slots']['Item']:\n should_end_session = False\n speech_output = \"Say: 'Add,' and the next item in the list.\"\n reprompt_text = \"To add an item to the list, say: 'Add,' and the next item in the list. For \" \\\n \"example, you can say: 'Add, 1 teaspoon of salt.'\"\n\n elif 'value' in intent['slots']['Item']:\n # If we are in create or edit mode. Add items to the session_attributes\n should_end_session = False\n session['attributes']['currentStep'] = curr_step = session_attributes['currentStep'] + 1\n session['attributes']['numberOfSteps'] = curr_step\n session['attributes']['listItems'][str(curr_step)] = intent['slots']['Item']['value']\n\n # Add it to the database\n update_list(session=session)\n update_session(session=session)\n\n speech_output = \"Adding '{}'. \".format(intent['slots']['Item']['value'])\n if session['attributes']['currentStep'] < 2:\n speech_output += \"Say: 'Add,' and the next item. Or say: 'save' to save the list. \" \\\n \"Saying 'cancel' stops without saving.\"\n elif 2 <= session['attributes']['currentStep'] < 3:\n speech_output += \"Say: 'add,' and an item. Or say: 'save'.\"\n else:\n pass\n reprompt_text = \"To add another item say: 'Add,' and the next item in the list.\" \\\n \"Otherwise say: 'stop' or 'save' to save your progress or 'cancel'\" \\\n \"discard your list.\"\n else:\n should_end_session = True\n speech_output = \"I didn't understand you.\" \\\n \"Say: 'Add,' and the next item in the list, \" \\\n \"or say: 'save'.\"\n reprompt_text = \"\"\n\n return build_response(session_attributes=session['attributes'],\n speechlet_response=build_speechlet_response(title=card_title,\n output=speech_output,\n reprompt_text=reprompt_text,\n should_end_session=should_end_session))\n\n\ndef delete_list(intent, session):\n \"\"\"Use the user id to delete an existing list. This removes all steps and progress and resets the\n session to be empty. This should require explicit confirmation.\"\"\"\n card_title = \"Delete List\"\n userId = session['user']['userId']\n should_end_session = True\n reprompt_text = \"\"\n\n print(\"***DELETE LIST, session: {}\".format(session.get('attributes', {})))\n print(\"***DELETE LIST, intent: {}\".format(intent['slots']))\n\n if 'value' in intent['slots']['listName']:\n listName = intent['slots']['listName']['value']\n speech_output = \"I deleted your list, {}.\".format(listName)\n elif session['attributes']['currentList'] != 'NONE':\n listName = session['attributes']['currentList']\n speech_output = \"I deleted your list, {}.\".format(listName)\n else:\n speech_output = \"I'm not sure what list to delete. Say: 'delete' and then a list name.\"\n reprompt_text = \"I need to know which list to delete. Say: 'delete' and then a list name.\"\n\n if listName == session['attributes']['currentList']:\n session['attributes']['currentList'] = 'NONE'\n session['attributes']['currentStep'] = 0\n session['attributes']['currentTask'] = 'None'\n session['attributes']['numberOfSteps'] = 0\n session['attributes']['listItems'] = {}\n\n table = boto3.resource('dynamodb').Table(LISTS_TABLENAME)\n try:\n response = table.delete_item(\n Key={\n 'userId': userId,\n 'listName': listName, },\n ReturnValues='ALL_OLD'\n )\n except botocore.exceptions.ClientError as e:\n print('ERROR reading from database: {}'.format(e.response))\n raise\n\n if 'attributes' not in response:\n speech_output = \"I couldn't find a list named {} to delete\".format(listName)\n\n update_session(session=session)\n\n return build_response(session_attributes=session['attributes'],\n speechlet_response=build_speechlet_response(title=card_title,\n output=speech_output,\n reprompt_text=reprompt_text,\n should_end_session=should_end_session))\n\n\ndef load_list(intent, session):\n \"\"\"Use the user id as a key and try to load the requested list from the database. If the list does not\n exist, offer to create the list. If the list does exist, set the value of the current list in the\n database to the requested list and set the current page to 1. So maybe a tuple? How does dynamo\n database work? Can I load ('My instructions for whatever', 1)?\n\n The key takeaway is that this function persists a session by setting the state in the database. That's\n it. It's the responsibility of some other intent to actually execute the step (whatever that means).\"\"\"\n card_title = intent['name']\n session_attributes = session.get('attributes', {})\n should_end_session = False # Let the user work with the list right away\n\n print(\"***LOAD LIST, session: {}\".format(session_attributes))\n print(\"***LOAD LIST, intent: {}\".format(intent['slots']))\n\n if 'value' in intent['slots']['listName']:\n # If trying to load a new list\n if session_attributes['currentList'] != intent['slots']['listName']['value']:\n lists_table = boto3.resource('dynamodb').Table(LISTS_TABLENAME)\n try:\n response = lists_table.get_item(Key={\n 'userId': session['user']['userId'],\n 'listName': intent['slots']['listName']['value']\n })\n except botocore.exceptions.ClientError as e:\n print(\"ERROR in LoadList: {}\".format(e.response))\n speech_output = \"There was a problem loading the list from the database.\"\n reprompt_text = \"\"\n should_end_session = True\n else:\n try:\n session['attributes']['currentList'] = response['Item']['listName']\n session['attributes']['currentStep'] = response['Item']['currentStep']\n session['attributes']['currentTask'] = 'PLAY'\n session['attributes']['listItems'] = response['Item']['listItems']\n session['attributes']['numberOfSteps'] = response['Item']['numberOfSteps']\n\n update_session(session=session)\n\n speech_output = \"I loaded your list: {}. \" \\\n \"You can play your list by saying: \" \\\n \"'tell generalist next'.\".format(intent['slots']['listName']['value'])\n reprompt_text = \"To start playback, say: 'next'.\"\n except KeyError: # List not found\n speech_output = \"I wasn't able to find the list {} \" \\\n \"in the database\".format(intent['slots']['listName']['value'])\n reprompt_text = \"\"\n should_end_session = True\n else: # If trying to load list that is already loaded\n session['attributes']['currentTask'] = 'PLAY'\n speech_output = \"Your list {} is already loaded. Say: 'next' to hear the next item in the \" \\\n \"list.\".format(session_attributes['currentList'])\n reprompt_text = \"To hear the next item say: 'next'.\"\n else:\n should_end_session = True\n speech_output = \"When you ask to load a list, make sure to tell me the name of the list. For \" \\\n \"example, say: 'load brownie recipe'.\"\n reprompt_text = \"To load a list, please say: 'load' followed by the list name. For example, \" \\\n \"you could say something like: 'load brownie recipe'.\"\n\n return build_response(session_attributes=session['attributes'],\n speechlet_response=build_speechlet_response(title=card_title,\n output=speech_output,\n reprompt_text=reprompt_text,\n should_end_session=should_end_session))\n\n\ndef get_next_item_from_list(session):\n \"\"\"Use the user id to determine the current list and step. Then execute the corresponding step. For\n now, this should simply be reading out the steps. This should ultimately increment the counter and\n notify the user if it has reached the end of the list.\"\"\"\n card_title = \"Get Next Item\"\n reprompt_text = \"\"\n\n print(\"***GET NEXT ITEM FROM LIST, session: {}\".format(session.get('attributes', {})))\n\n # No list is currently loaded\n if 'currentList' not in session['attributes'] or session['attributes']['currentList'] == \"NONE\":\n speech_output = \"You need to load a list before you can play it back. Say: 'load' and the name of \" \\\n \"a list. Or, if you haven't created a list yet, say: 'create' and the name of the \" \\\n \"list that you want to create.\"\n should_end_session = True\n # Reached the end of the currently loaded list\n elif session['attributes']['currentStep'] == session['attributes']['numberOfSteps']:\n speech_output = \"You've reached the end of your list {}. Start over by saying: 'start over'.\"\n should_end_session = True\n else: # Able to continue to the next item in the list\n session['attributes']['currentStep'] = curr_step = session['attributes']['currentStep'] + 1\n next_item = session['attributes']['listItems'][str(curr_step)]\n speech_output = \"{}\".format(next_item)\n should_end_session = False # Make it easy to get the next step right away\n update_session(session=session)\n update_list(session=session)\n\n return build_response(session_attributes=session['attributes'],\n speechlet_response=build_speechlet_response(title=card_title,\n output=speech_output,\n reprompt_text=reprompt_text,\n should_end_session=should_end_session))\n\n\ndef handle_start_over_request(session):\n \"\"\"If a list is loaded and mode is in play, then reset current step to 0 and let user know\"\"\"\n card_title = \"Start Over\"\n should_end_session = True\n reprompt_text = \"\"\n\n print(\"***HANDLE START OVER REQUEST, session: {}\".format(session.get('attributes', {})))\n\n if session['attributes']['currentList'] == \"NONE\":\n speech_output = \"You must load a list before I can restart.\"\n elif session['attributes']['currentTask'] in ['CREATE', 'EDIT']:\n speech_output = \"You are in {} mode. Save your list by saying: 'save' or 'stop' before trying to \" \\\n \"start over\".format(session['attributes']['currentTask'].lower())\n else:\n session['attributes']['currentStep'] = 0\n update_session(session=session)\n speech_output = \"Restarting.\"\n\n return build_response(session_attributes=session['attributes'],\n speechlet_response=build_speechlet_response(title=card_title,\n output=speech_output,\n reprompt_text=reprompt_text,\n should_end_session=should_end_session))\n\n\ndef peek_at_next_item_from_list(session):\n \"\"\"Use the user id to determine the current list and step. Then execute the corresponding step. For\n now, this should simply be reading out the steps. This should NOT increment the counter. If there are\n no more steps, notify the user they are on the last item in the list. Consider skipping back several\n steps.\"\"\"\n card_title = \"Peek at Next Item\"\n reprompt_text = \"\"\n\n print(\"***PEEK AT NEXT, session: {}\".format(session.get('attributes', {})))\n\n # No list is currently loaded\n if 'currentList' not in session['attributes'] or session['attributes']['currentList'] == \"NONE\":\n speech_output = \"You need to load a list before you can play it back. Say: 'load' and the name of \" \\\n \"a list. Or, if you haven't created a list yet, say: 'create' and the name of the \" \\\n \"list that you want to create.\"\n should_end_session = True\n # Reached the end of the currently loaded list\n elif session['attributes']['currentStep'] == session['attributes']['numberOfSteps']:\n speech_output = \"You're at the end of your list.\"\n should_end_session = True\n else: # Able to peek at the next item in the list\n curr_step = session['attributes']['currentStep'] + 1\n next_item = session['attributes']['listItems'][str(curr_step)]\n speech_output = \"{}\".format(next_item)\n should_end_session = False # Make it easy to get the next step right away\n\n return build_response(session_attributes=session['attributes'],\n speechlet_response=build_speechlet_response(title=card_title,\n output=speech_output,\n reprompt_text=reprompt_text,\n should_end_session=should_end_session))\n\n\ndef get_prev_item_from_list(session):\n \"\"\"Use the user id to determine the current list and step. First decrement the step number (this could\n handle a jump of more than one step). Then execute the current step. Should this increment the step\n again? Yes, probably so.\"\"\"\n card_title = \"Get Previous Item\"\n reprompt_text = \"\"\n\n print(\"***GET PREV ITEM, session: {}\".format(session.get('attributes', {})))\n\n # No list is currently loaded\n if 'currentList' not in session['attributes'] or session['attributes']['currentList'] == \"NONE\":\n speech_output = \"You need to load a list before you can play it back. Say: 'load' and the name of \" \\\n \"a list. Or, if you haven't created a list yet, say: 'create' and the name of the \" \\\n \"list that you want to create.\"\n should_end_session = True\n elif session['attributes']['currentStep'] == 0:\n speech_output = \"\"\n should_end_session = True\n elif session['attributes']['currentStep'] < 2:\n session['attributes']['currentStep'] = 0\n update_session(session=session)\n update_list(session=session)\n speech_output = \"You're at the beginning of your list.\"\n should_end_session = True\n else: # Able to continue to the next item in the list\n session['attributes']['currentStep'] = curr_step = session['attributes']['currentStep'] - 1\n next_item = session['attributes']['listItems'][str(curr_step)]\n speech_output = \"{}\".format(next_item)\n should_end_session = False # Make it easy to get the next step right away\n update_session(session=session)\n update_list(session=session)\n\n return build_response(session_attributes=session['attributes'],\n speechlet_response=build_speechlet_response(title=card_title,\n output=speech_output,\n reprompt_text=reprompt_text,\n should_end_session=should_end_session))\n\n\ndef review_previous_item_from_list(session):\n \"\"\"Use the user id to determine the current list and step. Execute the step but don't change the\n current step number. This is kind of like a reminder. It's more like 'Hey, what was the last step?'\n instead of 'Go back to the next step.\"\"\"\n card_title = \"Review Previous Item\"\n reprompt_text = \"\"\n\n print(\"***REVIEW PREVIOUS ITEM, session: {}\".format(session.get('attributes', {})))\n\n # No list is currently loaded\n if 'currentList' not in session['attributes'] or session['attributes']['currentList'] == \"NONE\":\n speech_output = \"You need to load a list before you can play it back. Say: 'load' and the name of \" \\\n \"a list. Or, if you haven't created a list yet, say: 'create' and the name of the \" \\\n \"list that you want to create.\"\n should_end_session = True\n elif session['attributes']['currentStep'] == 0:\n speech_output = \"\"\n should_end_session = True\n elif session['attributes']['currentStep'] == 0:\n speech_output = \"You're at the beginning of your list.\"\n should_end_session = True\n else:\n curr_step = session['attributes']['currentStep'] - 1\n next_item = session['attributes']['listItems'][str(curr_step)]\n speech_output = \"{}\".format(next_item)\n should_end_session = False # Make it easy to get the next step right away\n\n return build_response(session_attributes=session['attributes'],\n speechlet_response=build_speechlet_response(title=card_title,\n output=speech_output,\n reprompt_text=reprompt_text,\n should_end_session=should_end_session))\n\n\n# --------------- Session Persistence ------------------\n\ndef load_session(session):\n \"\"\"Use the current session's userId to load the stored session information\"\"\"\n userId = session['user']['userId']\n\n print(\"***LOAD SESSION, session: {}\".format(session.get('attributes')))\n\n stored_session_table = boto3.resource('dynamodb').Table(SESSION_TABLENAME)\n\n try:\n response = stored_session_table.get_item(Key={'userId': userId})\n except botocore.exceptions.ClientError as e:\n print(\"ERROR: {}\".format(e.response))\n return\n\n try:\n session['attributes'] = response['Item']['attributes']\n except KeyError:\n if 'attributes' not in session:\n session['attributes'] = {}\n session['attributes']['currentList'] = \"NONE\"\n session['attributes']['currentTask'] = \"NONE\"\n session['attributes']['currentStep'] = 0\n print(\"userId: {}\\n\"\n \"Loaded: session_attributes = {}\".format(userId, session['attributes']))\n\n\ndef update_session(session):\n \"\"\"Store the requested information in the StoredSession table.\"\"\"\n session_attributes = session.get('attributes', {})\n\n print(\"***UPDATE SESSION, session: {}\".format(session_attributes))\n\n stored_session_table = boto3.resource('dynamodb').Table(SESSION_TABLENAME)\n try:\n stored_session_table.put_item(\n Item={\n 'userId': session['user']['userId'],\n 'attributes': session_attributes\n }\n )\n except botocore.exceptions.ClientError as e:\n print('ERROR: {}'.format(e.response))\n raise\n\n\ndef update_list(session):\n \"\"\"Store the current session information into the list.\"\"\"\n session_attributes = session.get('attributes', {})\n\n print(\"***UPDATE LIST: session: {}\".format(session_attributes))\n\n lists_table = boto3.resource('dynamodb').Table(LISTS_TABLENAME)\n\n try:\n lists_table.put_item(\n Item={'userId': session['user']['userId'],\n 'listName': session_attributes['currentList'],\n 'numberOfSteps': session_attributes['numberOfSteps'],\n 'currentStep': session_attributes['currentStep'],\n 'listItems': session_attributes['listItems']\n }\n )\n except botocore.exceptions.ClientError as e:\n print('ERROR: {}'.format(e.response))\n raise\n\n\n# --------------- Events ------------------\n\ndef on_session_started(session_started_request, session):\n \"\"\" Called when the session starts \"\"\"\n print('on_session_started requestId={}, sessionId={}'.format(session_started_request['requestId'],\n session['sessionId']))\n load_session(session=session)\n\n\ndef on_launch(launch_request, session):\n \"\"\" Called when the user launches the skill without specifying what they want\"\"\"\n print('on_launch requestId={}, sessionId={}'.format(launch_request['requestId'], session['sessionId']))\n # Dispatch to your skill's launch\n print(\"***ON_LAUNCH session: {}\".format(session.get('attributes', {})))\n return get_welcome_response(session=session)\n\n\ndef on_intent(intent_request, session):\n \"\"\" Called when the user specifies an intent for this skill \"\"\"\n print('on_intent requestId={}, sessionId={}'.format(intent_request['requestId'], session['sessionId']))\n\n session_attributes = session.get('attributes', {})\n\n print(\"***ON INTENT, session: {}\".format(session_attributes))\n print(\"***ON INTENT, intent_request: {}\".format(session_attributes))\n\n intent = intent_request['intent']\n intent_name = intent_request['intent']['name']\n\n # Dispatch to your skill's intent handlers\n if intent_name == 'LoadListIntent':\n return load_list(intent, session)\n elif intent_name == 'CreateListIntent':\n return create_list(intent, session)\n elif intent_name == 'EditListIntent':\n return edit_list(intent, session)\n elif intent_name == 'AddItemIntent':\n return add_item(intent, session)\n elif intent_name == 'SaveIntent':\n return handle_save_intent(session)\n elif intent_name == 'AMAZON.NextIntent':\n return get_next_item_from_list(session)\n elif intent_name == 'AMAZON.PreviousIntent':\n return get_prev_item_from_list(session)\n elif intent_name == 'PeekIntent':\n return peek_at_next_item_from_list(session)\n elif intent_name == 'ReviewIntent':\n return review_previous_item_from_list(session)\n elif intent_name == 'AMAZON.HelpIntent':\n return get_help_response(session)\n elif intent_name == 'AMAZON.StopIntent':\n return handle_session_stop_request(session)\n elif intent_name == 'AMAZON.CancelIntent':\n return handle_session_cancel_request(session)\n elif intent_name == 'AMAZON.StartOverIntent':\n return handle_start_over_request(session)\n elif intent_name == 'DeleteIntent':\n return delete_list(intent, session)\n else:\n raise ValueError('Invalid intent')\n\n\ndef on_session_ended(session_ended_request, session):\n \"\"\" Called when the user ends the session.\n Is not called when the skill returns should_end_session=true\"\"\"\n print('on_session_ended requestId={}, sessionId={}'.format(session_ended_request['requestId'],\n session['sessionId']))\n update_list(session=session)\n update_session(session=session)\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":43756,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"323006702","text":"# Write a Python program to reverse words in a string.\n\nstring = 'Kirill Kondratenko'\n\n\ndef reverse_string(string):\n return ''.join(reversed(string))\n\n\ndef reverse_words(any_str):\n new_list = string.split()\n temp = ''\n for word in new_list:\n temp += reverse_string(word) + ' '\n return temp[:-1]\n\n\nprint(reverse_words(string))\n\n\n# print(' '.join(''.join(reversed(w)) for w in string.split()))\n","sub_path":"strings/w40.py","file_name":"w40.py","file_ext":"py","file_size_in_byte":414,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"589831094","text":"# -*- coding: UTF-8 -*-\n\n\ndef squaresList(n) -> list:\n \"\"\"Список квадратов нечетных чисел.\n\n На вход принимаем заданное число. Заполняем пустой массив:\n проверяем на нечетность и возводим в квадрат.\n \"\"\"\n squares = []\n for i in range(1, n + 1):\n if i % 2 != 0:\n squares.append(i * i)\n return squares\n","sub_path":"task2.py","file_name":"task2.py","file_ext":"py","file_size_in_byte":454,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"534683731","text":"import json\nimport re\n\nfrom bson import ObjectId\nfrom bson.json_util import dumps\nfrom flask import current_app\nfrom flask_restplus import Resource, fields, Namespace\nfrom geopy import distance\n\nfrom .commons import AuthorizationHeader, jwt_as_authorization\n\nitem_ns = Namespace(\"item\", description=\"Item operations\")\n\ngeolocation_model = item_ns.model(\"Geolocation\", {\n \"latitude\": fields.Float(required=True, description=\"Latitud\"),\n \"longitude\": fields.Float(required=True, description=\"Longitud\"),\n \"address\": fields.String(required=True, description=\"Direccion\"),\n})\n\nitem_model = item_ns.model(\"Item\", {\n \"id\": fields.String(attribute=lambda x: x[\"_id\"][\"$oid\"], readOnly=True,\n description=\"El identificador de articulo\"),\n \"name\": fields.String(required=True, description=\"Nombre del articulo\"),\n \"description\": fields.String(required=True, description=\"Descripcion del articulo\"),\n \"seller_id\": fields.String(attribute=lambda x: x[\"seller_id\"][\"$oid\"], required=False,\n description=\"Vendedor del articulo\"),\n \"unit_price\": fields.String(required=True, description=\"Precio por unidad\"),\n \"currency\": fields.String(required=True, default='USD', description=\"Moneda de la publicacion\"),\n \"units\": fields.Integer(required=True, description=\"Unidades disponibles\"),\n \"image_urls\": fields.List(fields.String, required=True, description=\"Fotos del articulo\"),\n \"geolocation\": fields.Nested(model=geolocation_model, required=True, description=\"Ubicacion geografica vendedor\"),\n \"payment_method\": fields.List(fields.String, required=True, description=\"Metodos de pago aceptados\"),\n \"categories\": fields.String(required=True, description=\"Categorias del item\")\n})\n\nitem_params = item_ns.parser()\nitem_params.add_argument(\"name\", type=str, help=\"Nombre del articulo\", location=\"args\")\nitem_params.add_argument(\"description\", type=str, help=\"Descripcion del articulo\", location=\"args\")\nitem_params.add_argument(\"latitude\", type=float, help=\"Latitud\", location=\"args\")\nitem_params.add_argument(\"longitude\", type=float, help=\"Longitud\", location=\"args\")\nitem_params.add_argument(\"kilometers\", type=float, help=\"Kilometros de distancia maxima\", location=\"args\")\nitem_params.add_argument(\"offset\", type=int, help=\"Offset de resultados\", location=\"args\")\nitem_params.add_argument(\"limit\", type=int, help=\"Limite de resultados\", location=\"args\")\n\n\n@item_ns.route(\"/\")\n@item_ns.response(404, \"Articulo no encontrado\")\n@item_ns.param(\"id\", \"El id de articulo\")\nclass Item(Resource):\n \"\"\"Muestra un unico articulo y permite borrarlo o modificarlo\"\"\"\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.item_collection = current_app.mongo_client.comprame_db.item\n\n @item_ns.doc(\"get_articulo\")\n @item_ns.marshal_with(item_model)\n def get(self, id):\n \"\"\"Obtener articulo por id\"\"\"\n return json.loads(dumps(self.item_collection.find_one({\"_id\": ObjectId(id)})))\n\n @item_ns.doc(\"put_articulo\")\n @item_ns.expect(item_model)\n @item_ns.marshal_with(item_model)\n def put(self, id):\n \"\"\"Actualiza el articulo\"\"\"\n self.item_collection.replace_one({\"_id\": ObjectId(id)}, item_ns.payload)\n item_ns.payload[\"_id\"] = ObjectId(id)\n return json.loads(dumps(item_ns.payload))\n\n\n@item_ns.route(\"/\")\n@item_ns.response(404, \"Articulo no encontrado\")\nclass Articulos(Resource):\n \"\"\"Busqueda y alta de articulos\"\"\"\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.item_collection = current_app.mongo_client.comprame_db.item\n self.user_collection = current_app.mongo_client.comprame_db.user\n\n @item_ns.doc(\"get_articulo\")\n @item_ns.expect(item_params)\n @item_ns.marshal_with(item_model)\n def get(self):\n \"\"\"Buscar articulos\"\"\"\n args = item_params.parse_args()\n\n query = {}\n\n if args.name:\n regex_name = re.compile('.*' + args.name + '.*', re.IGNORECASE)\n query[\"name\"] = regex_name\n\n if args.description:\n regex_description = re.compile('.*' + args.description + '.*', re.IGNORECASE)\n query[\"description\"] = regex_description\n\n pagination = {}\n pagination[\"offset\"] = 0\n if args.offset: pagination[\"offset\"] = args.offset\n pagination[\"limit\"] = 10\n if args.limit: pagination[\"limit\"] = args.limit\n\n \"\"\"La query de geolocation requiere su propia paginacion\"\"\"\n if args.latitude and args.longitude and args.kilometers:\n datos = self.item_collection.find(query)\n point1 = (args.longitude, args.latitude)\n array = []\n for doc in datos:\n if \"geolocation\" in doc:\n point2 = (doc[\"geolocation\"][\"longitude\"], doc[\"geolocation\"][\"latitude\"])\n dist = distance.distance(point1, point2).km\n if dist <= args.kilometers:\n array.append(doc)\n result = array[pagination[\"offset\"]:pagination[\"limit\"]]\n else:\n datos = self.item_collection.find(query).skip(pagination[\"offset\"]).limit(pagination[\"limit\"])\n result = []\n for doc in datos:\n result.append(doc)\n\n return json.loads(dumps(result))\n\n @item_ns.expect(item_model)\n @item_ns.marshal_with(item_model)\n @item_ns.expect(AuthorizationHeader)\n def post(self):\n \"\"\"Agregar un articulo\"\"\"\n dato = item_ns.payload\n args = AuthorizationHeader.parse_args()\n user_id = ObjectId(jwt_as_authorization(args['Authorization'])['user_id'])\n dato['seller_id'] = user_id\n result = self.item_collection.insert_one(dato)\n dato['id'] = result.inserted_id\n return json.loads(dumps(dato))\n","sub_path":"comprame/api/item.py","file_name":"item.py","file_ext":"py","file_size_in_byte":5858,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"293138715","text":"#!/usr/bin/env python\n\n\"\"\"\nModule that do the http request from Google Trends with the URL handler.\n\"\"\"\n\n__author__ = \"Thiago Baldim\"\n__copyright__ = \"Copyright 2015, The Trends Project\"\n__credits__ = [\"Thiago Baldim\", \"Calebe Bianchini\"]\n__license__ = \"GPL\"\n__version__ = \"0.3.1\"\n__maintainer__ = \"Thiago Baldim\"\n__email__ = \"thiagorbaldim@gmail.com\"\n__status__ = \"Development\"\n\nimport requests\nimport urllib\nimport datetime\nfrom bs4 import BeautifulSoup\n\nclass Trends:\n def __init__(self, csvPath=\"~/.\"):\n \"\"\"Set the csv Path for data handling\n \n Keyword arguments:\n csvPath --- the csv path fro future handling (default ~/.) \n \"\"\"\n self.csvPath = csvPath\n self.strTerm = None\n self.csvFile = None\n self.session = None\n \n def loginGoogle(self, strUser, strPassword):\n \"\"\"Set the session for Google Handling\n \n Keyword arguments:\n strUser --- The user to access the Google Accounts\n strPassword --- The password for google accounts \n \"\"\"\n \n self.session = requests.session()\n login_html = self.session.get(\"https://accounts.google.com/ServiceLogin\")\n soup_login = BeautifulSoup(login_html.content).find('form').find_all('input')\n dico = {}\n for u in soup_login:\n if u.has_attr('value'):\n dico[u['name']] = u['value']\n dico['Email'] = strUser\n dico['Passwd'] = strPassword\n self.session.post(\"https://accounts.google.com/ServiceLoginAuth\", data=dico)\n \n def search(self, strTerm):\n \"\"\"Search the Term to the Google Trends with the session created\n and save the CSV data in var self.csvFile\n \n Keyword arguments:\n strTerm --- The Term to be search in google Trends \n \"\"\"\n self.strTerm = urllib.quote(strTerm)\n url = \"http://www.google.com/trends/trendsReport?hl=en-US&q=\"+ self.strTerm +\"&tz=Etc%2FGMT%2B2&content=1&export=1\"\n response = self.session.get(url, stream=True)\n if(response.status_code == 200):\n self.csvFile = response.content\n return True\n else:\n return False\n \n def createCsvFile(self):\n \"\"\"Create the CSV File in the path that was set in the constructor\n \n \"\"\"\n fileName = self.csvPath + self.strTerm + '_' + datetime.date.today().strftime('%Y%m%d') + \".csv\"\n f = open(fileName, 'w')\n f.write(str(self.csvFile))\n f.close()\n return str(fileName)\n \n ","sub_path":"TrendsHandler/GoogleTrendsAPI.py","file_name":"GoogleTrendsAPI.py","file_ext":"py","file_size_in_byte":2596,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"152195274","text":"import sys\n\n\nclass Console:\n \"\"\"Renders a table to the console.\n\n Parameters\n ----------\n clear : bool, optional\n Clears the screen every time the table is updated.\n out : TextIOBase, optional\n Text stream to write to. If unspecified, ``sys.stdout`` is used.\n \"\"\"\n def __init__(self, clear=False, out=sys.stdout):\n self._clear = clear\n self._n_lines = 0\n self._out = out\n\n def __call__(self, result):\n if self._clear:\n self._out.write('\\x1b[2J')\n else:\n self._out.write('\\033[F\\033[K' * self._n_lines)\n self._n_lines = result.count('\\n') + 1\n self._out.write(result)\n self._out.write('\\n')\n self._out.flush()\n","sub_path":"src/lazy_table/artists/_console.py","file_name":"_console.py","file_ext":"py","file_size_in_byte":732,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"40746773","text":"from heapq import heappop, heappush\n\nn_researchers, inactivity_min = map(int, input().split())\nlst_arrival, lst_departure = [], []\nfor _ in range(n_researchers):\n arrive_at, stay_for_min = map(int, input().split())\n heappush(lst_arrival, arrive_at)\n heappush(lst_departure, arrive_at + stay_for_min)\n\nsave_min = 0\nwhile lst_arrival:\n arrival = heappop(lst_arrival)\n while arrival - lst_departure[0] > inactivity_min:\n heappop(lst_departure)\n if lst_departure[0] <= arrival:\n heappop(lst_departure)\n save_min += 1\n\nprint(save_min)\n","sub_path":"Assigning Workstations.py","file_name":"Assigning Workstations.py","file_ext":"py","file_size_in_byte":569,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"273144308","text":"from lagom.envs.vec_env import VecEnv\n\n\nclass BasePolicy(object):\n r\"\"\"Base class for all policies.\n \n Any policy should subclass this class.\n \n The subclass should implement at least the following:\n \n - :meth:`__call__`\n \n .. note::\n \n For the consistency of different variants of policies and fast prototyping, we restrict that all\n policies should deal with VecEnv (batched data).\n \n \"\"\"\n def __init__(self, config, network, env_spec, **kwargs):\n r\"\"\"Initialize the policy. \n \n Args:\n config (dict): A dictionary of configurations. \n network (BaseNetwork): a neural network as function approximator in the policy. \n env_spec (EnvSpec): environment specification. \n **kwargs: keyword arguments to specify the policy. \n \"\"\"\n self.config = config\n self.network = network\n self.env_spec = env_spec\n \n msg = f'expected type VecEnv, got {type(self.env_spec.env)}'\n assert isinstance(self.env_spec.env, VecEnv), msg\n \n # Set all keyword arguments\n for key, val in kwargs.items():\n self.__setattr__(key, val)\n \n def __call__(self, x, out_keys=['action']):\n r\"\"\"Define the computation of the policy given input data at every call. \n \n Should be overridden by all subclasses.\n \n .. note::\n \n There is an option to select metrics for the policy to calculate and only selected items will\n be calculated and returned e.g. ``out_keys=['action', 'action_logprob', 'entropy']``. \n This is very useful to dramatically speedup in some scenarios. For example in ES, it\n turns out that outputing all metrics of an action distribution makes training extremly\n slow and only action is useful but others like log-probability, entropy etc. \n \n Args:\n x (object): input data to the policy. \n out_keys (list, optional): a list of required metrics for the policy to output. \n Default: ``['action']``\n \n Returns\n -------\n out_policy : dict\n A dictionary of output data about the computation of the policy. It should contain\n at least one key 'action'. Other possible keys include ['action_logprob', 'state_value'\n 'entropy', 'perplexity']. \n \"\"\"\n raise NotImplementedError\n \n def __repr__(self):\n r\"\"\"Returns a string representation of the policy network. \"\"\"\n string = self.__class__.__name__ + '\\n'\n string += '\\tNetwork: ' + self.network.__repr__() + '\\n'\n \n return string\n","sub_path":"lagom/core/policies/base_policy.py","file_name":"base_policy.py","file_ext":"py","file_size_in_byte":2736,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"370728375","text":"from layers.layer import LinearOutputLayer\nimport torch\nimport numpy as np\nfrom tensorboardX import SummaryWriter\nimport utils.helper_functions as hf\n# User variables\nn = 6\nnb_classes = 2\nbatch_size = 5\nh = 1e-3\n\nwriter = SummaryWriter()\n\nlayer = LinearOutputLayer(n,n,'mse',writer)\n#\n# targets = torch.randint(0,n,(batch_size,))\n# targets = hf.one_hot(targets, n)\n\ntargets = torch.randn(batch_size,n,1)\nactivation = torch.randn(batch_size,n,1)\nlayer.forward_linear_activation = activation\nlayer.forward_output = activation\n\n\n\nlayer.compute_backward_output(targets)\ngradient = layer.backward_output\n\nloss = layer.loss(targets)\ngradient_fd = torch.empty(activation.shape)\n\n# compute finite difference gradient\nfor b in range(batch_size):\n for i in range(n):\n fd = torch.zeros(activation.shape)\n fd[b,i,0] = h\n activation_fd = activation + fd\n layer.forward_linear_activation = activation_fd\n layer.forward_output = activation_fd\n loss_fd = layer.loss(targets)\n gradient_fd[b,i,0] = (loss_fd-loss)/h\n\nerror = torch.norm(gradient-gradient_fd, p=float('inf'))\n\nprint('error: {}'.format(error))\nprint(gradient)\nprint(gradient_fd)\n\n\n","sub_path":"tests/test_finite_differences.py","file_name":"test_finite_differences.py","file_ext":"py","file_size_in_byte":1179,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"479426629","text":"import os\nimport shutil\nimport subprocess\nimport sys\nimport time\nimport traceback\n\nimport infosec.utils\n\n\nTEST_COMMAND = 'echo \"I am g`whoami`!\"; exit'\nCOMMAND_RESULT = 'I am groot!'\n\n\nPATH_TO_SUDO = './sudo'\n\n\ndef error(message):\n print('\\x1b[31m{}\\x1b[0m'.format(message))\n\n\ndef check_arg_from_function(module_path, function_name):\n try:\n module = infosec.utils.import_module(module_path)\n function = getattr(module, function_name)\n result = function()\n except Exception as e:\n error('Exception generating argument for {}'.format(module_path))\n traceback.print_exc()\n return False, None\n\n if not isinstance(result, str):\n error('Invalid {} type for argument: type was {}, expected str'.format(\n module_path, type(result)))\n return False, None\n\n return True, result\n\n\ndef check_q1a():\n if os.path.isfile('core'):\n os.remove('core')\n success, arg = check_arg_from_function('q1a.py', 'get_crash_arg')\n if not success:\n return False\n result = infosec.utils.execute([PATH_TO_SUDO, arg])\n if not os.path.exists('core'):\n error('ERROR: Running q1a.py did not generate a `core` file!')\n return False\n return True\n\n\ndef check_q1b():\n success, arg = check_arg_from_function('q1b.py', 'get_arg')\n if not success:\n return False\n result = infosec.utils.execute([PATH_TO_SUDO, arg], TEST_COMMAND)\n if COMMAND_RESULT not in result.stdout:\n error('ERROR: Failed running a root command shell using q1b.py!')\n return False\n return True\n\n\ndef check_q1c():\n success, arg = check_arg_from_function('q1c.py', 'get_arg')\n if not success:\n return False\n result = infosec.utils.execute([PATH_TO_SUDO, arg], TEST_COMMAND)\n if COMMAND_RESULT not in result.stdout:\n error('ERROR: Failed running a root command shell using q1c.py!')\n return False\n if result.exit_code != 0x42:\n error('ERROR: The shell did not exit with a code of 0x42 (66)!')\n return False\n return True\n\n\ndef check_q3():\n success, arg = check_arg_from_function('q3.py', 'get_arg')\n if not success:\n return False\n\n prefix = 'auth='\n result = infosec.utils.execute([\n '/usr/bin/gdb', '--batch',\n '-ex', 'run', '-ex', 'printf \"{}%d\\n\", auth'.format(prefix),\n '--args', PATH_TO_SUDO, arg])\n\n if prefix not in result.stdout:\n error('ERROR: Failed debugging sudo with the argument from q3.py!')\n return False\n\n auth_line = result.stdout[result.stdout.find(prefix):].splitlines()[0]\n auth = int(auth_line[len(prefix):])\n\n if auth == 0:\n error('ERROR: Debugging your q3.py, it seems auth is stil 0!')\n return False\n\n return True\n\n\ndef check_q4():\n success, arg = check_arg_from_function('q4.py', 'get_arg')\n if not success:\n return False\n\n current_user_sudo = os.path.join(os.path.dirname(PATH_TO_SUDO), 'sudo_smoketest')\n try:\n # Create a copy of sudo, not owned by root, so we can kill it\n shutil.copy(PATH_TO_SUDO, current_user_sudo)\n # Create a 'head' process to pipe sudo into, and this will keep the\n # the output from growing while allowing the sudo program to write.\n head = subprocess.Popen(['/usr/bin/head', '-n', '15'],\n stdin=subprocess.PIPE, stdout=subprocess.PIPE)\n # Pipe sudo into head\n sudo = subprocess.Popen([current_user_sudo, arg], stdout=head.stdin)\n # Let this run for a second\n time.sleep(2)\n # Kill both, and then analyze the results\n sudo.kill()\n head.kill()\n lines = head.stdout.read().strip().splitlines()\n if len(lines) < 10:\n error('ERROR: Failed getting a large amount of lines with q4.py!')\n return False\n if any('to your leader!' not in line for line in lines):\n error('ERROR: Failed finding a call to your leader in the output from q4.py!')\n return False\n return True\n finally:\n if os.path.isfile(current_user_sudo):\n os.remove(current_user_sudo)\n\n\ndef check_q_search():\n try:\n search = infosec.utils.import_module('search.py')\n gs = search.GadgetSearch('libc.bin', 0x123)\n except Exception as e:\n error('ERROR: Failed loading the gadget search engine')\n traceback.print_exc()\n return False\n\n try:\n regs = ('esi', 'edi')\n gadget_format = 'MOV {0}, {1}'\n expected = set([\n gadget_format.format(reg1, reg2)\n for reg1 in regs\n for reg2 in regs\n ])\n cmds = gs.format_all_gadgets('MOV {0}, {1}', ('esi', 'edi'))\n if not set(cmds) == set(expected):\n error('ERROR: Unexpected output with format_all_gadgets!')\n print('Expected: {}, Actual: {}'.format(expected, actual))\n return False\n except Exception as e:\n error('ERROR: Failed using the gadget search engine')\n traceback.print_exc()\n return False\n return True\n\n\ndef check_if_nonempty(path):\n if not os.path.exists(path):\n error('ERROR: {} does not exist'.format(path))\n return False\n with open(path) as reader:\n data = reader.read().strip()\n if not data:\n error('ERROR: {} is empty'.format(path))\n return False\n return True\n\n\ndef smoketest():\n os.chdir(os.path.dirname(os.path.abspath(__file__)))\n if all([\n check_q1a(),\n check_q1b(),\n check_q1c(),\n check_q_search(),\n check_q3(),\n check_q4(),\n check_if_nonempty('libc.bin'),\n check_if_nonempty('q1a.txt'),\n check_if_nonempty('q1b.txt'),\n check_if_nonempty('q1c.txt'),\n check_if_nonempty('q3.txt'),\n check_if_nonempty('q4.txt'),\n ]):\n print('smoketest seems cool')\n\n\nif __name__ == '__main__':\n smoketest()\n","sub_path":"hw6/smoketest.py","file_name":"smoketest.py","file_ext":"py","file_size_in_byte":5925,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"95591439","text":"import sys\nimport time\nimport atexit\nimport socket\nimport threading\n\nfrom multiprocessing.pool import ThreadPool\n\nfrom asyncpool import LOG\nfrom asyncpool import util\nfrom asyncpool.serializer import DummySerializer, JSONSerializer, PickleSerializer\nfrom asyncpool.exceptions import NotAvailableError, BackendError, CommunicationError\n\nSOCK_FILES_LIMIT = 512\n\n'''\ntask = {\n 'queue': queue,\n 'id': task_id,\n 'func': func,\n 'args': args,\n 'kwds': kwds,\n 'pri': priority,\n 'expire': expire,\n 'res_ttl': res_ttl,\n}\n\nresult = {\n 'queue': queue,\n 'id': task_id,\n 'res': res,\n 'exc': exc,\n 'trb': trb,\n 'expire': expire\n}\n'''\n\nclass PyBackend(object):\n\n def __init__(\n self,\n sock_file='/var/run/pool.sock',\n threads=10,\n serializer=JSONSerializer):\n\n self._sock_file = sock_file\n self._serializer = serializer\n self._in_processing = 0\n self._in_pending = 0\n\n self._tasks_queue = dict()\n self._results = dict()\n\n self._threads = threads\n self._pool = ThreadPool(self._threads)\n\n self._sock_pool = util.Pool(\n lambda: socket.socket(socket.AF_UNIX, socket.SOCK_STREAM),\n lambda sock: False,\n SOCK_FILES_LIMIT\n )\n\n self._stop = True\n\n @property\n def sock_file(self):\n return self._sock_file\n\n @property\n def in_processing(self):\n return self._in_processing\n\n @property\n def in_pending(self):\n return self._in_pending\n\n @property\n def stop(self):\n return self._stop\n\n def _clean_tasks_queue(self):\n for queue, tasks in self._tasks_queue.iteritems():\n for task in tasks:\n if task[1]['expire'] <= time.time():\n self._tasks_queue[queue].remove(task)\n\n def _clean_results(self):\n for task_id, result in self._results.iteritems():\n if result['expire'] <= time.time():\n del self._results[task_id]\n\n def _cleanup(self):\n while not self._stop:\n try:\n self._clean_tasks_queue()\n self._clean_results()\n except:\n LOG.error('Cleanup failed, reason: %s', str(sys.exc_info()[:-1]))\n finally:\n time.sleep(5)\n\n def _append_task(self, task):\n LOG.debug('Append task, task id: %s' % task['id'])\n self._tasks_queue.setdefault(task['queue'], list())\n self._tasks_queue[task['queue']].append((task['pri'], task))\n self._in_pending += 1\n if task['pri']:\n self._tasks_queue[task['queue']].sort()\n\n def _append_result(self, result):\n LOG.debug('Append result, task id: %s' % result['id'])\n self._results[result['id']] = result\n self._in_processing -= 1\n\n def _pop_task(self, queue):\n try:\n priority, task = self._tasks_queue[queue].pop()\n except (KeyError, IndexError):\n raise NotAvailableError('Task is not available, queue: %s' % queue)\n if task['expire'] <= int(time.time()):\n raise NotAvailableError('Expire timeout, task: %s' % task['id'])\n self._in_pending -= 1\n self._in_processing += 1\n return task\n\n def _handle(self, conn, addr):\n try:\n data = util.recv(conn, self._serializer)\n LOG.debug('Request %s' % data)\n if 'get' in data:\n if data['get'] == 'task':\n try:\n queue = data['data']['queue']\n task = self._pop_task(queue)\n try:\n util.send(\n conn,\n {\n 'resp': True,\n 'data': task,\n },\n self._serializer\n )\n except:\n self._append_task(task)\n raise\n except NotAvailableError:\n util.send(\n conn,\n {\n 'resp': False\n },\n self._serializer\n )\n elif data['get'] == 'res':\n task_id = data['data']['id']\n if task_id in self._results:\n util.send(\n conn,\n {\n 'resp': True,\n 'data': self._results[task_id]\n },\n self._serializer\n )\n else:\n util.send(\n conn,\n {\n 'resp': False\n },\n self._serializer\n )\n elif 'put' in data:\n if data['put'] == 'task':\n task = data['data']\n self._append_task(task)\n elif data['put'] == 'res':\n result = data['data']\n self._append_result(result)\n self._in_processing -= 1\n except:\n LOG.exception('Handle failed, reason: %s' % str(sys.exc_info()[:-1]))\n finally:\n conn.close()\n\n def _serve_forever(self):\n sock = self._sock_pool.get()\n try:\n sock.bind(self._sock_file)\n sock.listen(500)\n atexit.register(util.delete_file, self._sock_file)\n\n LOG.debug('Serve forever: %s' % self._sock_file)\n while not self._stop:\n try:\n conn, addr = sock.accept()\n self._pool.apply_async(self._handle, (conn, addr))\n except KeyboardInterrupt:\n LOG.critical(sys.exc_info()[:-1])\n break\n except:\n msg = 'Serve forever failed, reason: %s' % str(sys.exc_info()[:-1])\n LOG.error(msg)\n LOG.info('Stoped')\n except:\n msg = 'Serve forever failed, reason: %s' % str(sys.exc_info()[:-1])\n LOG.error(msg)\n finally:\n sock.shutdown(socket.SHUT_RDWR)\n sock.close()\n self._sock_pool.put(sock)\n self._pool.close()\n self._pool.join()\n\n util.delete_file(self._sock_file)\n\n def send_task(self, task, local=False):\n \"\"\"\n Send task to backend\n \"\"\"\n\n if local:\n self._append_task(task)\n else:\n data = {\n 'put': 'task',\n 'data': task,\n }\n while True:\n sock = self._sock_pool.get()\n try:\n sock.connect(self._sock_file)\n util.send(sock, data, self._serializer)\n break\n except CommunicationError:\n LOG.error('Send task failed, reason: %s' % str(sys.exc_info()[:-1]))\n finally:\n sock.shutdown(socket.SHUT_RDWR)\n sock.close()\n self._sock_pool.put(sock)\n\n def send_result(self, result, local=False):\n \"\"\"\n Send result to backend\n \"\"\"\n\n if local:\n self._append_result(result)\n else:\n data = {\n 'put': 'res',\n 'data': result,\n }\n while True:\n sock = self._sock_pool.get()\n try:\n sock.connect(self._sock_file)\n util.send(sock, data, self._serializer)\n break\n except CommunicationError:\n LOG.error('Send result failed, reason: %s' % str(sys.exc_info()[:-1]))\n finally:\n sock.shutdown(socket.SHUT_RDWR)\n sock.close()\n self._sock_pool.put(sock)\n\n def get_task(self, queue='default', local=False):\n \"\"\"\n Get task from backend\n \"\"\"\n\n if local:\n return self._pop_task(queue)\n else:\n request = {\n 'get': 'task',\n 'data': {'queue': queue},\n }\n while True:\n sock = self._sock_pool.get()\n try:\n sock.connect(self._sock_file)\n util.send(sock, request, self._serializer)\n data = util.recv(sock, self._serializer)\n if not data['resp']:\n raise NotAvailableError()\n return data['data']\n except CommunicationError:\n LOG.error('Get task failed, reason: %s' % str(sys.exc_info()[:-1]))\n finally:\n sock.shutdown(socket.SHUT_RDWR)\n sock.close()\n self._sock_pool.put(sock)\n\n def get_result(self, task_id, queue='default', local=False):\n \"\"\"\n Get result from backend\n \"\"\"\n\n if local:\n if task_id not in self._results:\n raise NotAvailableError()\n res = self._results[task_id]['res']\n exc = self._results[task_id]['exc']\n trb = self._results[task_id]['trb']\n return {'res': res, 'exc': exc, 'trb': trb}\n else:\n request = {\n 'get': 'res',\n 'data': {'id': task_id},\n }\n while True:\n sock = self._sock_pool.get()\n try:\n sock.connect(self._sock_file)\n util.send(sock, request, self._serializer)\n data = util.recv(sock, self._serializer)\n if not data['resp']:\n raise NotAvailableError()\n return data['data']\n except CommunicationError:\n LOG.error('Get result failed, reason: %s' % str(sys.exc_info()[:-1]))\n finally:\n sock.shutdown(socket.SHUT_RDWR)\n sock.close()\n self._sock_pool.put(sock)\n\n def start(self, thread=True):\n if not self._stop:\n LOG.debug('Already started')\n return\n LOG.debug('Starting')\n self._stop = False\n try:\n if thread:\n t_s = threading.Thread(target=self._serve_forever)\n t_s.daemon = True\n t_s.start()\n time.sleep(0.1)\n if not t_s.is_alive():\n raise BackendError(\"Start failed\")\n else:\n self._serve_forever()\n t_c = threading.Thread(target=self._cleanup)\n t_c.daemon = True\n t_c.start()\n time.sleep(0.1)\n if not t_c.is_alive():\n raise BackendError(\"Start failed\")\n except:\n self._stop = True\n raise\n LOG.info('Started')\n\n def stop(self):\n if not self._stop:\n LOG.debug('Stoping')\n self._stop = True\n","sub_path":"asyncpool/backend.py","file_name":"backend.py","file_ext":"py","file_size_in_byte":11412,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"60391369","text":"import pandas as pd\nimport datetime as dt\n\ndef insert_row(row_number, df, row_value): \n start_upper = 0\n end_upper = row_number \n start_lower = row_number \n end_lower = df.shape[0] \n upper_half = [*range(start_upper, end_upper, 1)] \n lower_half = [*range(start_lower, end_lower, 1)] \n lower_half = [x.__add__(1) for x in lower_half] \n index_ = upper_half + lower_half \n df.index = index_ \n df.loc[row_number] = row_value \n df = df.sort_index() \n return df \n\n#'1wall 6targets small - Challenge - 2020.11.10-10.44.00 Stats.csv'\n#'Tile Frenzy - Challenge - 2020.12.04-01.33.01 Stats.csv'\n#'Ascended Tracking 90 - Challenge - 2020.11.27-00.44.16 Stats.csv'\ncsv = 'Ascended Tracking 90 - Challenge - 2020.11.27-00.44.16 Stats.csv'\n\ndf = pd.read_csv(csv, error_bad_lines=False)\n\ndftop = df.iloc[:-26,:]\ndfbot = df.iloc[-25:,:2]\ndfbot.columns = [\"CATEGORY\",\"VALUE\"]\n\ntoInsert = [\"Shots:\", df.at[len(df) - 26, \"Timestamp\"] ]\ndfbot = insert_row(0, dfbot, toInsert)\n\ntoInsert = [\"Hits:\", df.at[len(df) - 26, \"Bot\"]]\ndfbot = insert_row(1, dfbot, toInsert)\n\ntoInsert = [\"Accuracy:\", int(dfbot.at[1, \"VALUE\"]) / int(dfbot.at[0, \"VALUE\"])]\ndfbot = insert_row(2, dfbot, toInsert)\n\ntoInsert = [\"Damage Efficiency:\", float(df.at[len(df) - 26, \"Weapon\"]) / float(df.at[len(df) - 26, \"TTK\"])]\ndfbot = insert_row(3, dfbot, toInsert)\n\nprint(df.to_string())\nprint(dftop.to_string())\nprint()\nprint(dfbot.to_string())\n\n#Series to put in Master\n#Get date and time from filename\n#'Ascended Tracking 90 - Challenge - 2020.11.27-00.44.16 Stats.csv'\nstrs = csv.split(\" \")\ndateNtime = strs[-2].split(\"-\")\ndateSplit = dateNtime[0].split(\".\")\ntimeSplit = dateNtime[1].split(\".\")\nprint(dateNtime)\nprint(timeSplit)\ndate = dt.datetime(int(dateSplit[0]), int(dateSplit[1]), int(dateSplit[2]), int(timeSplit[0]), int(timeSplit[1]), int(timeSplit[2]))\nprint(date)\n\ndfMaster = dfbot.iloc[:16:, 1]\n#toInsert = [\"Date and Time:\", date]\ndfMaster = insert_row(0, dfMaster, date)\n\nprint(type(dfMaster))\nprint(dfMaster.to_string())\n\nscenario = csv.split(\"-\")[0].split(\" \")[0]\nfor x in range(1,len(csv.split(\"-\")[0].split(\" \"))-1):\n scenario = scenario + \"_\" + csv.split(\"-\")[0].split(\" \")[x]\nprint(scenario)\noutputPath = \"C:\\\\Users\\\\jakee\\\\Desktop\\\\Kovaaks\\\\\" + scenario + \".csv\"\nprint(outputPath)\ndfbot.to_csv(path_or_buf= outputPath ,index=False)\n#print(stats.head(5))","sub_path":"KovaaKs Progress Tracker Pre Trip.py","file_name":"KovaaKs Progress Tracker Pre Trip.py","file_ext":"py","file_size_in_byte":2364,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"375258288","text":"# -*- coding: utf-8 -*-\n\nfrom lessons.models import Log\n\n__author__ = 'Admin'\n\n\nclass LogRequestMiddleware(object):\n\n def process_request(self, request):\n method = request.method\n user_agent = request.META.get('HTTP_USER_AGENT', '')\n #url = request.META['SERVER_NAME'].decode('cp1251')+request.META['SERVER_PORT']+request.path\n url = request.META.get('HTTP_HOST', '')+request.path\n new_log = Log.objects.create(method=method, url=url, user_agent=user_agent)","sub_path":"lessons/middlewares.py","file_name":"middlewares.py","file_ext":"py","file_size_in_byte":495,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"626240737","text":"# Use modern Python\nfrom __future__ import absolute_import, print_function, unicode_literals\n\n# Django imports\nfrom django.test import TestCase, SimpleTestCase\n\n# External imports\nfrom django_prbac.models import Grant, Role\n\n# CCHQ imports\nfrom corehq.apps.hqadmin.management.commands import cchq_prbac_bootstrap\nfrom fab.pillow_settings import apply_pillow_actions_to_pillows, \\\n get_pillows_for_env, get_single_pillow_action\n\n\nclass TestCchqPrbacBootstrap(TestCase):\n \"\"\"\n Tests the PRBAC bootstrap with and without --dry-run\n \"\"\"\n\n def test_dry_run(self):\n \"\"\"\n When --dry-run is passed, no models should be created\n \"\"\"\n self.assertEquals(Role.objects.count(), 0)\n self.assertEquals(Grant.objects.count(), 0)\n\n command = cchq_prbac_bootstrap.Command()\n command.handle(dry_run=True)\n\n self.assertEquals(Role.objects.count(), 0)\n self.assertEquals(Grant.objects.count(), 0)\n\n def test_non_dry_run(self):\n \"\"\"\n When there is no --dry-run passed, it defaults to false, and\n things happen. Furthermore, the thing should be idempotent\n \"\"\"\n self.assertEquals(Role.objects.count(), 0)\n self.assertEquals(Grant.objects.count(), 0)\n\n command = cchq_prbac_bootstrap.Command()\n command.handle(dry_run=False)\n\n # Just make sure something happened\n self.assertGreater(Role.objects.count(), 10)\n self.assertGreater(Grant.objects.count(), 10)\n\n role_count = Role.objects.count()\n grant_count = Grant.objects.count()\n\n command.handle(dry_run=False)\n\n self.assertEquals(Role.objects.count(), role_count)\n self.assertEquals(Grant.objects.count(), grant_count)\n\n\nclass TestPillowTopFiltering(SimpleTestCase):\n \"\"\"\n Tests the function that excludes certain pillows from running on staging.\n \"\"\"\n\n def setUp(self):\n self.pillowtops = {\n 'core': [\n 'corehq.pillows.case.CasePillow',\n 'corehq.pillows.xform.XFormPillow',\n 'corehq.pillows.domain.DomainPillow',\n 'corehq.pillows.user.UserPillow',\n 'corehq.pillows.application.AppPillow',\n 'corehq.pillows.group.GroupPillow',\n 'corehq.pillows.sms.SMSPillow',\n 'corehq.pillows.user.GroupToUserPillow',\n 'corehq.pillows.user.UnknownUsersPillow',\n 'corehq.pillows.sofabed.FormDataPillow',\n 'corehq.pillows.sofabed.CaseDataPillow',\n ],\n 'phonelog': [\n 'corehq.pillows.log.PhoneLogPillow',\n ],\n }\n\n def test_no_blacklist_items(self):\n expected_pillows = {u'corehq.pillows.case.CasePillow',\n u'corehq.pillows.xform.XFormPillow',\n u'corehq.pillows.domain.DomainPillow',\n u'corehq.pillows.user.UserPillow',\n u'corehq.pillows.application.AppPillow',\n u'corehq.pillows.group.GroupPillow',\n u'corehq.pillows.sms.SMSPillow',\n u'corehq.pillows.user.GroupToUserPillow',\n u'corehq.pillows.user.UnknownUsersPillow',\n u'corehq.pillows.sofabed.FormDataPillow',\n u'corehq.pillows.sofabed.CaseDataPillow',\n u'corehq.pillows.log.PhoneLogPillow'}\n\n self.assertEqual(expected_pillows, apply_pillow_actions_to_pillows(\n [], self.pillowtops))\n\n def test_with_blacklist_items(self):\n expected_pillows = {u'corehq.pillows.case.CasePillow',\n u'corehq.pillows.xform.XFormPillow',\n u'corehq.pillows.domain.DomainPillow',\n u'corehq.pillows.user.UserPillow',\n u'corehq.pillows.application.AppPillow',\n u'corehq.pillows.group.GroupPillow',\n u'corehq.pillows.sms.SMSPillow',\n u'corehq.pillows.user.GroupToUserPillow',\n u'corehq.pillows.user.UnknownUsersPillow',\n u'corehq.pillows.sofabed.FormDataPillow',\n u'corehq.pillows.sofabed.CaseDataPillow'}\n\n self.assertEqual(expected_pillows, apply_pillow_actions_to_pillows(\n [{'exclude_groups': ['phonelog']}], self.pillowtops))\n\n def test_loading_existing_conf_file(self):\n\n expected_action = {'include_groups': ['mvp']}\n\n action = get_single_pillow_action('staging')\n self.assertEqual(action.to_json(), expected_action)\n\n def test_loading_no_existing_conf_file(self):\n action = get_single_pillow_action('foo')\n self.assertIsNone(action)\n\n def test_india_server_exclusions(self):\n self.pillowtops['fluff'] = [\n 'custom.bihar.models.CareBiharFluffPillow',\n 'custom.opm.models.OpmCaseFluffPillow',\n 'custom.opm.models.OpmUserFluffPillow',\n ]\n\n pillows = get_pillows_for_env('india', self.pillowtops)\n self.assertNotIn('custom.opm.models.OpmCaseFluffPillow', pillows)\n self.assertNotIn('custom.opm.models.OpmUserFluffPillow', pillows)\n self.assertIn('custom.bihar.models.CareBiharFluffPillow', pillows)\n","sub_path":"corehq/apps/hqadmin/tests.py","file_name":"tests.py","file_ext":"py","file_size_in_byte":5424,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"580642967","text":"import numpy as np\nimport matplotlib.pyplot as plt\n\n\ndef main():\n z0 = np.array([[0., 0., 2.], [0., 0., 1.], [2., 1., 0.]])\n z1 = np.array([[0., 0., 2.], [0., 0., 1.], [2., 1., -1.7E-13]])\n\n fig = plt.figure(figsize=(2, 4))\n\n ax0 = fig.add_subplot(121)\n ax0.contourf(z0, cmap='jet')\n ax0.set_aspect('equal')\n\n ax1 = fig.add_subplot(122)\n ax1.contourf(z1, cmap='jet')\n ax1.set_aspect('equal')\n\n plt.show()\n\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"test_contourf_data0.py","file_name":"test_contourf_data0.py","file_ext":"py","file_size_in_byte":475,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"593114535","text":"import sys\nimport os\nimport subprocess\nimport docker\nimport shutil\n\ndatasetpath = \"./dataset\"\n\n\ndef Select_Algo(list_algo,list_dataset):\n #Select Algorithm\n print(\"#Select Machinelearning Algorithms(Select 0 if you want to add an algorithm)\")\n print(\"0 : Add Algorithm\")\n \n for index,value in enumerate(list_algo,start=1):\n print(index,\":\",value)\n select_algo=input('Select Numbers:').split(',')\n\n if select_algo[0] =='0':\n print(\"#ADD NEW ALGORITHM\")\n name = input('#Name:')\n image = input('#Image(ex.ubuntu, tensorflow/tensorflow):')\n argument = 'docker run -i -t -d --name ' + name + ' ' + image + ' /bin/bash'\n subprocess.call(argument,shell=True)\n argument = 'docker ps'\n subprocess.call(argument,shell=True)\n Copy_Dataset_algo(name)\n Machine_Learn(name,list_dataset)\n list_algo=List_Algo()\n list_algo,select_algo=Select_Algo(list_algo,list_dataset)\n\n return list_algo,select_algo\n\ndef Select_Dataset(list_algo,list_dataset,datasetpath):\n #Select Dataset\n print(\"#Select Datasets(Select 0 if you want to add an dataset)\")\n print(\"0 : Add dataset\")\n \n for index,value in enumerate(list_dataset,start=1):\n print(index,\":\",value)\n select_dataset=input('Select Numbers:').split(',')\n \n if select_dataset[0] == '0':\n print(\"#ADD NEW DATASET\")\n name = input('#Name:')\n path = input('#Path:')\n destination = datasetpath + '/' + name \n shutil.copytree(path,destination)\n Copy_Dataset_data(name,list_algo)\n list_dataset=List_Dataset(datasetpath)\n list_dataset,select_dataset=Select_Dataset(list_algo,list_dataset,datasetpath)\n return list_dataset,select_dataset\n\n#1 - make algo -> all dataset to one container\n#Copy Host dataset to new_container\ndef Copy_Dataset_algo(container_name):\n argument = 'docker cp ./dataset ' + container_name+ ':/dataset' \n print(argument)\n subprocess.call(argument,shell=True)\n\n#2 - make dataset -> one dataset to all container\ndef Copy_Dataset_data(dataset_name,list_algo):\n print(\"INSERT COPY_DATASET_DATA\")\n print(list_algo)\n for algo in list_algo:\n argument = 'docker cp ./dataset/' + dataset_name + ' ' + algo + ':/dataset/'+dataset_name\n print(argument)\n subprocess.call(argument,shell=True)\n\n#Learn new algorithm with all dataset\ndef Machine_Learn(name,list_dataset):\n #exec selected algorithm in container (need to fix run.py)\n for dataset in list_dataset:\n #Run Macine learning\n argument = 'docker exec ' + name + ' python3 run.py' + ' ' + dataset\n print(argument)\n #subprocess.call(argument,shell=True)\n\n\n#Save each container name to list_algo[]\ndef List_Algo():\n list_algo=[]\n client = docker.from_env()\n for container in client.containers.list():\n list_algo.append(container.name)\n return list_algo\n\ndef List_Dataset(datasetpath):\n list_dataset = os.listdir(datasetpath)\n return list_dataset\n\n#Copy Result value container to host\ndef Copy_Result(list_algo,list_dataset,select_algo,select_dataset,saveFilePath,saveHostPath):\n for algo in select_algo:\n for data in select_dataset:\n argument = 'docker cp ' + list_algo[int(algo) -1] + ':' + saveFilePath + '/' +list_algo[int(algo)-1] + '_' + list_dataset[int(data)-1] + ' ' + saveHostPath\n subprocess.call(argument,shell=True)\n\n#Print result in host directory\ndef Print_Result(saveHostPath):\n path_dir = saveHostPath\n file_lists = os.listdir(path_dir)\n print(\"RESULT\")\n for file in file_lists:\n f = open(saveHostPath + '/' + file, \"r\")\n line = f.readline()\n print(line)\n os.remove(saveHostPath + '/' + file)\n f.close()\n\ndef main():\n #Select Path\n saveFilePath = \"/saveResult\"\n saveHostPath = \"./result\"\n #datasetpath = \"./dataset\"\n print(datasetpath)\n list_algo=List_Algo()\n list_dataset=List_Dataset(datasetpath)\n\n list_algo,select_algo=Select_Algo(list_algo,list_dataset)\n list_dataset,select_dataset=Select_Dataset(list_algo,list_dataset,datasetpath)\n\n print(\"#select_Algo=\",select_algo,\"Dataset=\",select_dataset,\"list_algo=\",list_algo)\n\n Copy_Result(list_algo,list_dataset,select_algo,select_dataset,saveFilePath,saveHostPath)\n Print_Result(saveHostPath)\n\nif __name__ == '__main__':\n main()\n\n","sub_path":"cve.py","file_name":"cve.py","file_ext":"py","file_size_in_byte":4437,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"591496863","text":"import pymongo\nfrom pymongo import MongoClient\nfrom path import Path as path\nimport configs\nimport MySQLdb\nimport os\nimport json\nimport sys\nfrom lv_utils import configs\n\ncurrent_db=None\ndb_config=None\ndef mongo_db():\n global current_db\n if(current_db==None):\n client = MongoClient(\n configs.get_config().no_sql.host,\n configs.get_config().no_sql.port\n )\n if (configs.get_config().no_sql.user != \"\"):\n client[configs.get_config().no_sql.name].authenticate(configs.get_config().no_sql.user,\n configs.get_config().no_sql.password)\n current_db = client[configs.get_config().no_sql.name]\n\n\n return current_db\ndef sql_db():\n global db_config\n if(db_config==None):\n PROJECT_ROOT = path(__file__).abspath().dirname().dirname() # /edx-platform/cms\n REPO_ROOT = PROJECT_ROOT.dirname()\n ENV_ROOT = REPO_ROOT.dirname()\n SERVICE_VARIANT = os.environ.get('SERVICE_VARIANT', None)\n CONFIG_PREFIX = SERVICE_VARIANT + \".\" if SERVICE_VARIANT else \"\"\n CONFIG_ROOT = path(os.environ.get('CONFIG_ROOT', ENV_ROOT))\n with open(CONFIG_ROOT / CONFIG_PREFIX + \"auth.json\") as auth_file:\n AUTH_TOKENS = json.load(auth_file)\n db_config={\n \"name\":AUTH_TOKENS.get(\"MYSQL_CONFIG\").get(\"NAME\"),\n \"host\":AUTH_TOKENS.get(\"MYSQL_CONFIG\").get(\"HOST\"),\n \"user\": AUTH_TOKENS.get(\"MYSQL_CONFIG\").get(\"USER\"),\n \"password\": AUTH_TOKENS.get(\"MYSQL_CONFIG\").get(\"PASSWORD\")\n }\ndef configuration():\n return configs()\n\n\n","sub_path":"lv_utils/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":1666,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"167989612","text":"import collections\n\n\ninput_file = open('input.txt', 'r')\nmessages = []\n# Read the input into the messages list\nfor line in input_file:\n messages.append(line)\n\n# Create a list to hold the messages by column first (so we'll have a 8x624 list)\nby_column = []\nfor col_index in range(0, len(messages[0])-1):\n new_entry = []\n for row_index in range(0, len(messages)):\n new_entry.append(messages[row_index][col_index])\n by_column.append(new_entry)\n\nerror_corrected = ''\n# Get the most common letter in each column and build up the final string\nfor column in by_column:\n # This gets a list of the letters in the string and their respective counts\n letters_with_counts = collections.Counter(column).most_common()\n error_corrected += letters_with_counts[-1][0]\nprint(error_corrected)\n","sub_path":"day6/part2.py","file_name":"part2.py","file_ext":"py","file_size_in_byte":802,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"313327780","text":"from django.conf.urls.defaults import *\nfrom everythingreductor.views import *\n\n# Uncomment the next two lines to enable the admin:\n# from django.contrib import admin\n# admin.autodiscover()\n\nurlpatterns = patterns('',\n\t('^$', index),\n\t(r'^cut/$', cut),\n\t(r'^tech/(.+)$', tech),\n\t(r'^a/(.+)$', uncut),\n\t(r'^copytest/$', copy),\n\t(r'^b/', show_all),\n # Example:\n # (r'^everythingreductor/', include('everythingreductor.foo.urls')),\n\n # Uncomment the admin/doc line below and add 'django.contrib.admindocs' \n # to INSTALLED_APPS to enable admin documentation:\n # (r'^admin/doc/', include('django.contrib.admindocs.urls')),\n\n # Uncomment the next line to enable the admin:\n # (r'^admin/', include(admin.site.urls)),\n)\n","sub_path":"stuff/python_engine/everythingreductor/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":734,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"129204875","text":"from tornado import web\n\nfrom model import meta, LogEntry\nfrom handlers.base import BaseHandler\n\nclass LifeLog(BaseHandler):\n\t\"\"\"Life log servlet\"\"\"\n\t_path = '/log'\n\n\t@web.authenticated\n\tdef get(self):\n\t\toffset = int(self.get_argument('start', '0'))\n\n\t\tself.write({'entries': [entry.jsonable for entry in LogEntry.list(10, offset)]})\n\n\t@web.authenticated\n\tdef post(self):\n\t\ttext = self.get_argument('entry')\n\n\t\ttry:\n\t\t\tentry = LogEntry(text)\n\t\t\tmeta.session.add(entry)\n\t\t\tmeta.session.commit()\n\t\texcept:\n\t\t\tmeta.session.rollback()\n\t\t\traise\n\n\t\tself.redirect('/log/%d' % entry.id)\n\n\nclass LifeLogTag(BaseHandler):\n\t\"\"\"Life log tag\"\"\"\n\t_path = '/log/([a-z]+)'\n\n\t@web.authenticated\n\tdef get(self, tag):\n\t\toffset = int(self.get_argument('start', '0'))\n\n\t\tself.write({'entries': [entry.jsonable for entry in LogEntry.list(10, offset, tag=tag)]})\n\n\nclass LifeLogEntry(BaseHandler):\n\t\"\"\"Life log entry\"\"\"\n\t_path = '/log/([0-9]+)'\n\n\t@web.authenticated\n\tdef get(self, entry_id):\n\t\tentry_id = int(entry_id)\n\n\t\tentry = LogEntry.get(entry_id)\n\n\t\tif not entry:\n\t\t\tself.send_error(404)\n\n\t\tself.write(entry.jsonable)\n","sub_path":"handlers/life_log.py","file_name":"life_log.py","file_ext":"py","file_size_in_byte":1101,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"331764879","text":"#reset() # Erase all previously defined variables\n\n###################################################\n# Initialization #\n###################################################\nfrom MPKC.MatsumotoImai.Sflash.sflash_encoder import SflashEncoder\nfrom MPKC.MatsumotoImai.Sflash.sflash_generator import SflashKeyGenerator\nfrom MPKC.Utils.asn1 import ASN1\nimport binascii\nimport sys\nimport os\n\nif (len( sys.argv ) < 1 ):\n\tprint(\"Out file required!\\n\")\n\tsys.exit(1)\n\ndir = sys.argv[len(sys.argv) - 1]\nif not os.path.isdir(dir):\n\tfilePath = dir\n\tidx = dir[::-1].find(\"/\")\n\tif idx > 0:\n\t\tdir = dir[0: len(dir) - idx - 1]\n\t\tif not os.path.isdir(dir):\n\t\t\tprint(\"Out file required!\\n\")\n\t\t\tsys.exit(1)\nelse:\n\tfilePath = dir + '/sflash'\n\nencoder = SflashEncoder(\"BER\")\n#asn = encoder.getEncoder()\nkeygen = SflashKeyGenerator()\nkeyPair = keygen.generateKeyPair()\n\npublicBin = encoder.encodePublic(keyPair.getPublic())\nprivateBin = encoder.encodePrivate(keyPair.getPrivate())\n\n#print(binascii.hexlify(publicBin))\n#print(binascii.hexlify(privateBin))\nprint(publicBin)\nprint(privateBin)\n\nfile = open(filePath + \".pub\", \"wb\")\nfile.write(publicBin)\nprint(\"Public key has been store in \" + file.name)\nfile.close()\n\nfile = open(filePath + \".priv\", \"wb\")\nfile.write(privateBin)\nprint(\"Private key has been stored in \" + file.name)\nfile.close()\n","sub_path":"encoding/sflash_keygen.py","file_name":"sflash_keygen.py","file_ext":"py","file_size_in_byte":1352,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"470592531","text":"import cv2\nimport random as rng\nimport time\n\ncamera = cv2.VideoCapture(0)\nsample = 0\nprefix = str(time.time())\n\n_, frame = camera.read()\nframeClone = frame.copy()\n\nframeGrayscale = frame.copy()\nframeGrayscale = cv2.cvtColor(frameGrayscale, cv2.COLOR_BGR2GRAY)\n\ncv2.imshow('grayscale', frameGrayscale)\n\nframe = cv2.cvtColor(frame, cv2.COLOR_BGR2HLS)\nhue = [29, 85]\nsat = [28, 90]\nlum = [100, 223]\nframe = cv2.inRange(frame, (hue[0], lum[0], sat[0]), (hue[1], lum[1], sat[1]))\n\nframe = cv2.dilate(frame, None, iterations=1)\nframe = cv2.erode(frame, None, iterations=1)\ncv2.imshow('hls', frame)\n\ncontours, hierarchy = cv2.findContours(frame, cv2.RETR_LIST, cv2.CHAIN_APPROX_SIMPLE)\n\nindex = 0\nfor i in range(len(contours)):\n r = cv2.boundingRect(contours[i])\n if cv2.contourArea(contours[i]) < 100 or r[2] > r[3] or r[2] < 40 or r[3] < 40: # Minimum size\n continue\n\n newRect = (\n (r[0] + (r[2]/2)) - (r[3]/2), r[1], r[3], r[3]\n )\n\n color = (rng.randint(0,256), rng.randint(0,256), rng.randint(0,256))\n cv2.rectangle(frameClone, newRect, color)\n\n try:\n crop_img = frameGrayscale[newRect[1]:newRect[1]+newRect[3], newRect[0]:newRect[0]+newRect[2]]\n crop_img = cv2.resize(crop_img, (32, 32))\n cv2.imshow('countour {0}'.format(index), crop_img)\n\n index = index + 1\n\n cv2.imwrite(\"samples/\" + prefix + \"-\" + str(sample) + \".jpg\", crop_img)\n sample = sample + 1\n\n except:\n print(\"An exception occurred\")\n\ncv2.imshow('frame', frameClone)\n\nwhile True:\n if cv2.waitKey(1) & 0xFF == ord('q'):\n break\n","sub_path":"nn.py","file_name":"nn.py","file_ext":"py","file_size_in_byte":1527,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"632491526","text":"import tensorflow as tf \r\nimport dippykit as dip\r\nimport numpy as np\r\nfrom PIL import Image\r\n\r\nimport matplotlib.pyplot as plt\r\nfrom itertools import combinations\r\nimport os, sys \r\nimport argparse\r\nimport scipy.misc\r\nimport time\r\n\r\nimport copy\r\n\r\nVGG_MEAN = [103.939, 116.779, 123.68]\r\ndef build_part_vgg19(img_input,params_dir='vgg19.npy'):\r\n '''\r\n input tensor: input image with shape of [N, H, W, C]\r\n params_dir: Directory of npz file\r\n '''\r\n def conv_layer(x, name):\r\n with tf.variable_scope(name):\r\n f = tf.constant(params[name][0],dtype='float32')\r\n b = tf.constant(params[name][1],dtype='float32')\r\n\r\n conv = tf.nn.conv2d(input=x, filter=f,strides=[1,1,1,1],padding='SAME')\r\n conv_biased = tf.nn.bias_add(conv, b)\r\n return tf.nn.relu(conv_biased)\r\n\r\n def max_pool(x, name):\r\n return tf.nn.max_pool(x, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME', name=name)\r\n\r\n\r\n params = np.load(params_dir,encoding='latin1').item()\r\n print('Params loaded from %s'%params_dir)\r\n\r\n red, green, blue = tf.split(axis=3, value=img_input,num_or_size_splits=3)\r\n bgr = tf.concat(axis=3, values=[\r\n blue - VGG_MEAN[0],\r\n green - VGG_MEAN[1],\r\n red - VGG_MEAN[2],\r\n ])\r\n\r\n conv1_1 = conv_layer(bgr, 'conv1_1')\r\n conv1_2 = conv_layer(conv1_1, 'conv1_2')\r\n pool1 = max_pool(conv1_2, 'pool1')\r\n\r\n conv2_1 = conv_layer(pool1, 'conv2_1')\r\n conv2_2 = conv_layer(conv2_1, 'conv2_2')\r\n pool2 = max_pool(conv2_2, 'pool2')\r\n\r\n conv3_1 = conv_layer(pool2, 'conv3_1')\r\n conv3_2 = conv_layer(conv3_1, 'conv3_2')\r\n conv3_3 = conv_layer(conv3_2, 'conv3_3')\r\n conv3_4 = conv_layer(conv3_3, 'conv3_4')\r\n pool3 = max_pool(conv3_4, 'pool3')\r\n\r\n conv4_1 = conv_layer(pool3, 'conv4_1')\r\n conv4_2 = conv_layer(conv4_1, 'conv4_2')\r\n conv4_3 = conv_layer(conv4_2, 'conv4_3')\r\n conv4_4 = conv_layer(conv4_3, 'conv4_4')\r\n pool4 = max_pool(conv4_4, 'pool4')\r\n \r\n conv5_1 = conv_layer(pool4, 'conv5_1')\r\n # Style # Content\r\n #return conv1_1, conv2_1, conv3_1, conv4_1, conv5_1, conv4_2\r\n return [conv1_1, conv1_2, pool1, conv2_1, conv2_2, pool2, conv3_1, conv3_2, conv3_3, conv3_4, pool3, conv4_1, conv4_2, conv4_3, conv4_4, pool4, conv5_1]\r\n\r\n\r\n\r\ndef main():\r\n parser = argparse.ArgumentParser()\r\n parser.add_argument('--content-img',type=str,default='content_img.jpg',help='The content image')\r\n parser.add_argument('--style-img',type=str,default='style_img.jpg',help='The style image')\r\n parser.add_argument('--output',type=str, default='faceswap',help='Output image')\r\n parser.add_argument('--lr-rate',type=float, default=.2,help='Learning rate')\r\n parser.add_argument('--epoch',type=int, default=1000,help='Epoch number')\r\n parser.add_argument('--style-weight',type=float, default=1,help='trade-off between content and style')\r\n parser.add_argument('--content-weight',type=float, default=1,help='trade-off between content and style')\r\n\r\n args = parser.parse_args()\r\n\r\n content_img = Image.open(args.content_img)\r\n img_width,img_height = content_img.size\r\n content_img = content_img.resize((img_width,img_height))\r\n style_img = Image.open(args.style_img).resize((img_width,img_height))\r\n\r\n input_img = copy.copy(content_img)\r\n# plt.title('Content Image')\r\n# plt.imshow(content_img)\r\n# plt.pause(1)\r\n\r\n# plt.title('Style Image')\r\n# plt.imshow(style_img)\r\n# plt.pause(1)\r\n\r\n vgg_input = tf.Variable(initial_value=np.zeros(shape=[1, img_height, img_width, 3],dtype='float32'),name='image') \r\n\r\n # Style # Content\r\n # conv1_1, conv2_1, conv3_1, conv4_1, conv5_1, conv4_2 = build_part_vgg19(vgg_input,params_dir='vgg19.npy')\r\n # conv1_1, conv1_2, pool1, conv2_1, conv2_2, pool2, conv3_1, conv3_2, conv3_3, conv3_4, pool3, \\\r\n # conv4_1, conv4_2, conv4_3, conv4_4, pool4, conv5_1 = build_part_vgg19(vgg_input,params_dir='vgg19.npy')\r\n layers = build_part_vgg19(vgg_input,params_dir='vgg19.npy')\r\n #for l in layers:\r\n # print(l.shape)\r\n\r\n # reshape to NHWC\r\n content_img = np.reshape(content_img,newshape=(-1,img_height,img_width,3))\r\n print(content_img.dtype, np.max(content_img), np.min(content_img))\r\n style_img = np.reshape(style_img,newshape=(-1,img_height,img_width,3))\r\n input_img = np.reshape(input_img,newshape=(-1,img_height,img_width,3))\r\n\r\n # GPU Config\r\n tf_config = tf.ConfigProto()\r\n tf_config.gpu_options.allow_growth=True\r\n sess = tf.Session(config=tf_config)\r\n counter = 0\r\n\r\n with open('results.txt', 'a') as of:\r\n of.write('RESULTS FILE\\n')\r\n sess.run(tf.global_variables_initializer())\r\n for lr_rate in np.logspace(-2, 1, 2):\r\n for style_weight in np.logspace(1, 4, 2):\r\n for m, n in np.ndindex((3, 3)):\r\n m += 2; n += 2\r\n for content_layers in combinations(layers, m):\r\n for style_layers in combinations(layers, n):\r\n print(content_layers)\r\n counter += 1\r\n # Get content feature maps\r\n sess.run(vgg_input.assign(content_img))\r\n if m > 1:\r\n content_maps_out = sess.run(content_layers)\r\n else:\r\n content_maps_out, = sess.run([content_layers])\r\n # Get style feature maps\r\n sess.run(vgg_input.assign(style_img))\r\n if n > 1:\r\n style_maps_out = sess.run(style_layers)\r\n else:\r\n style_maps_out, = sess.run([style_layers])\r\n\r\n # Loss\r\n def gram_matrix(maps):\r\n if isinstance(maps,tf.Tensor):\r\n maps_vec = tf.transpose(maps,perm=(0,3,1,2))\r\n a,b,c,d = maps_vec.shape\r\n maps_vec = tf.reshape(maps_vec,(a*b,c*d))\r\n return 1/(2* int(a*b*c*d) ) * tf.matmul(maps_vec, maps_vec,transpose_b=True)\r\n else:\r\n maps_vec = np.array(maps).transpose((0,3,1,2))\r\n a,b,c,d = maps_vec.shape\r\n maps_vec = maps_vec.reshape(a*b,c*d)\r\n return 1/(2*(a*b*c*d) ) * np.matmul(maps_vec,maps_vec.T)\r\n\r\n # Input\r\n if m > 1:\r\n content_maps = [tf.constant(content_maps_out[i],dtype='float32') for i in range(len(content_maps_out))]\r\n else:\r\n content_maps = tf.constant(content_maps_out, dtype='float32')\r\n\r\n if n > 1:\r\n style_maps = [tf.constant(style_maps_out[i], dtype='float32') for i in range(len(style_maps_out))]\r\n else:\r\n style_maps = tf.constant(style_maps_out, dtype='float32')\r\n\r\n\r\n #def _cal_squaredNM(m):\r\n # m_shape = m.get_shape().as_list()\r\n # return 4*(m_shape[0]*m_shape[1])**2\r\n\r\n #style_weights = [0.2,0.2,0.2,0.2,0.2]\r\n #img_styles = [conv1_1, conv2_1, conv3_1, conv4_1, conv5_1]\r\n\r\n def mse(x,y):\r\n return tf.losses.mean_squared_error(labels=y,predictions=x)\r\n\r\n def cosdist(x,y):\r\n return tf.losses.cosine_distance(labels=y,predictions=x,axis=1)\r\n\r\n loss_content = tf.Variable(initial_value=0, dtype='float32')\r\n if m > 1:\r\n for l, content_layer in enumerate(content_layers):\r\n loss_content = tf.add(loss_content, mse(content_layer,content_maps[l]))\r\n else:\r\n loss_content = tf.add(loss_content, mse(content_layers,content_maps))\r\n\r\n loss_style = tf.Variable(initial_value=0, dtype='float32')\r\n if m > 1:\r\n for l, style_layer in enumerate(style_layers):\r\n loss_style = tf.add(loss_style, mse(style_layer,style_maps[l]))\r\n else:\r\n loss_style = tf.add(loss_style, mse(style_layers,style_maps))\r\n\r\n print('Added loss_content and style')\r\n loss = args.content_weight*loss_content + args.style_weight*loss_style\r\n\r\n # Train\r\n opt = tf.train.AdamOptimizer(args.lr_rate).minimize(loss,var_list=[vgg_input])\r\n\r\n sess.run(tf.global_variables_initializer())\r\n #white_noise = dip.adjustments.image_noise(np.zeros((1, img_height, img_width, 3)))\r\n #dip.imshow(white_noise)\r\n #dip.show()\r\n #sess.run(vgg_input.assign(white_noise))\r\n sess.run(vgg_input.assign(input_img))\r\n initTime = time.time()\r\n for ep in range(args.epoch+1):\r\n _, cur_loss,s_loss,c_loss,img = sess.run([opt,loss,loss_style, loss_content, vgg_input])\r\n if ep%5==0:\r\n print('[*] Epoch %d total_loss=%f, style_loss=%f, content_loss=%f'%(ep,cur_loss,s_loss,c_loss))\r\n if ep%1000==0:\r\n saved_img = np.array(img[0])\r\n saved_img = np.where(saved_img<=255,saved_img,255)\r\n saved_img = np.where(saved_img>=0,saved_img,0)\r\n saved_img = Image.fromarray(saved_img.astype(np.uint8),'RGB')\r\n output_name = args.output+'_%d'%(counter)+'.jpg'\r\n saved_img.save(output_name)\r\n print(\"[!] image saved as %s\\n\"%output_name)\r\n\r\n with open('results.txt', 'a') as of:\r\n of.write(f'{counter}, learning rate: {lr_rate}, style weight: {style_weight}, '\r\n f'content layers {content_layers}, style layers {style_layers}')\r\n\r\n print('Done after {} seconds'.format(time.time()-initTime))\r\n \r\nif __name__=='__main__':\r\n main()\r\n","sub_path":"final/transfer_gatys_tf.py","file_name":"transfer_gatys_tf.py","file_ext":"py","file_size_in_byte":10764,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"391013717","text":"# https://youtu.be/SuDtHqtC5OE\n\"\"\"\nReinhard color transfer \nBased on the paper: https://www.cs.tau.ac.il/~turkel/imagepapers/ColorTransfer.pdf\n\nThis approach is suitable for stain normalization of pathology images where\nthe 'look and feel' of all images can be normalized to a template image. \nThis can be a good preprocessing step for machine learning and deep learning \nof pathology images. \n\n\"\"\"\n\nimport numpy as np\nimport cv2\nimport os\n\ninput_dir = \"data/original/\"\ninput_image_list = os.listdir(input_dir)\n\noutput_dir = \"data/stain_normalization/\"\n\ndef get_mean_and_std(x):\n\tx_mean, x_std = cv2.meanStdDev(x)\n\tx_mean = np.hstack(np.around(x_mean,2))\n\tx_std = np.hstack(np.around(x_std,2))\n\treturn x_mean, x_std\n\ntemplate_img = cv2.imread('data/original/3.png')\ntemplate_img = cv2.cvtColor(template_img,cv2.COLOR_BGR2LAB)\ntemplate_mean, template_std = get_mean_and_std(template_img)\n\nfor img in (input_image_list):\n print(img)\n input_img = cv2.imread(input_dir+img)\n input_img = cv2.cvtColor(input_img,cv2.COLOR_BGR2LAB)\n \n \n img_mean, img_std = get_mean_and_std(input_img)\n \n \n height, width, channel = input_img.shape\n for i in range(0,height):\n for j in range(0,width):\n for k in range(0,channel):\n \tx = input_img[i,j,k]\n \tx = ((x-img_mean[k])*(template_std[k]/img_std[k]))+template_mean[k]\n \tx = round(x)\n \t# boundary check\n \tx = 0 if x<0 else x\n \tx = 255 if x>255 else x\n \tinput_img[i,j,k] = x\n \n input_img= cv2.cvtColor(input_img,cv2.COLOR_LAB2BGR)\n cv2.imwrite(output_dir+\"modified_\"+img, input_img)\n\n","sub_path":"304 - Augmentation of histology images​/000-Reinhard_stain_norm_only.py","file_name":"000-Reinhard_stain_norm_only.py","file_ext":"py","file_size_in_byte":1654,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"86541187","text":"\n\n#calss header\nclass _JACKAL():\n\tdef __init__(self,): \n\t\tself.name = \"JACKAL\"\n\t\tself.definitions = [u'a wild animal like a dog that lives in Africa and southern Asia and eats animals that have died or been killed by others']\n\n\t\tself.parents = []\n\t\tself.childen = []\n\t\tself.properties = []\n\t\tself.jsondata = {}\n\n\n\t\tself.specie = 'nouns'\n\n\n\tdef run(self, obj1 = [], obj2 = []):\n\t\treturn self.jsondata\n","sub_path":"xai/brain/wordbase/nouns/_jackal.py","file_name":"_jackal.py","file_ext":"py","file_size_in_byte":400,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"335045307","text":"# -*- coding: utf-8 -*-\n\nfrom openerp.osv import fields, osv\nimport time\nimport logging\nlogger = logging.getLogger(__name__)\n\n\nclass hr_employee(osv.osv):\n _inherit = 'hr.employee'\n\n def _get_latest_contract(self, cr, uid, ids, field_name,\n args, context=None):\n res = {}\n obj_contract = self.pool.get('hr.contract')\n for emp in self.browse(cr, uid, ids, context=context):\n contract_ids = obj_contract.search(\n cr, uid, [('employee_id', '=', emp.id)],\n order='date_start', context=context)\n if contract_ids:\n res[emp.id] = contract_ids[-1:][0]\n else:\n res[emp.id] = False\n return res\n\n def _get_visibility(self, cr, uid, ids, field_name, args, context=None):\n res = {}\n for emp in self.browse(cr, uid, ids, context=context):\n visible = False\n if emp.user_id.id == uid:\n visible = True\n elif emp.parent_id.user_id.id == uid:\n visible = True\n else:\n group_ids = self.pool.get('res.users').browse(\n cr, uid, uid, context=context).groups_id\n group_user_id = self.pool.get(\"ir.model.data\").get_object_reference(cr, uid, 'base', 'group_hr_user')[1]\n if group_user_id in [group.id for group in group_ids]:\n visible = True\n else:\n group_user_id = self.pool.get(\"ir.model.data\").get_object_reference(cr, uid, 'base', 'group_hr_manager')[1]\n if group_user_id in [group.id for group in group_ids]:\n visible = True\n res[emp.id] = visible\n return res\n\n def _get_children(self, cr, uid, ids, field_name, arg, context):\n employees = self.browse(cr, uid, ids)\n res = {}\n for employee in employees:\n count = 0\n for child in employee.children_ids:\n count += 1\n res[employee.id] = count\n return res\n\n def _get_chargefam(self, cr, uid, ids, field_name, arg, context):\n employees = self.browse(cr, uid, ids)\n res = {}\n for employee in employees:\n count = 0\n for child in employee.children_ids:\n ages = child.age.split()\n if len(ages) > 1:\n if int(ages[0][:-1]) < 21:\n count += 1\n res[employee.id] = count\n return res\n\n def _get_date_start(self, cr, uid, ids, field_name, arg, context):\n employees = self.browse(cr, uid, ids)\n res = {}\n for employee in employees:\n if not employee.contract_ids:\n res[employee.id] = '1900-01-01'\n continue\n for contract in employee.contract_ids:\n res[employee.id] = contract.date_start\n break\n return res\n\n def name_get(self, cr, uid, ids, context=None):\n # if not len(ids):\n # return []\n res = []\n for employee in self.browse(cr, uid, ids, context=context):\n p_name = employee.name\n if employee.matricule:\n p_name = '[' + employee.matricule + '] ' + p_name\n res.append((employee.id, p_name))\n return res\n\n def _get_is_birthday(self, cr, uid, ids, field_name, args, context=None):\n res = {}\n this_week = time.strftime('%W')\n for emp in self.browse(cr, uid, ids):\n if not emp.birthday:\n res[emp.id] = False\n elif time.strftime('%W', time.strptime(emp.birthday, '%Y-%m-%d')) == this_week:\n res[emp.id] = True\n else:\n res[emp.id] = False\n return res\n\n def _search_is_birthday(self, cr, uid, obj, name, args,\n domain=None, context=None):\n res = []\n ids = self.search(cr, uid, [('active', '=', True)])\n for emp in self.browse(cr, uid, ids):\n for flds, operator, value in args:\n if not value and not emp.is_birthday:\n res.append(emp.id)\n if value and emp.is_birthday:\n res.append(emp.id)\n return [('id', 'in', res)]\n\n _columns = {\n 'visible': fields.function(_get_visibility, method=True,\n string='Visible', type='boolean'),\n 'matricule': fields.char('Matricule', size=64),\n 'cin': fields.char('CIN', size=64),\n 'date': fields.function(_get_date_start, method=True,\n string='Date Embauche', type='date'),\n # 'date': fields.date(\n # 'Date entree',\n # help='''Cette date est requipe pour le\n # calcule de la prime d' anciennete'''),\n 'anciennete': fields.boolean(\n 'Prime anciennete',\n help='Est ce que cet employe benificie de la prime d\\'anciennete'),\n 'mode_reglement': fields.selection(\n [('virement', 'Virement'), ('cheque', 'Cheque'),\n ('espece', 'Espece')], 'Mode De Reglement'),\n 'payment_term_id': fields.one2many('payment.term', 'employee_id',\n 'Mode de Paiement'),\n 'bank': fields.char('Banque', size=128),\n 'compte': fields.char('Compte bancaire', size=128),\n # 'chargefam' : fields.integer('Nombre de personnes a charge'),\n 'chargefam': fields.function(_get_chargefam, method=True,\n type='float'),\n 'logement': fields.float('Abattement Fr Logement'),\n 'affilie': fields.boolean(\n 'Affilie',\n help='Est ce qu on va calculer les cotisations pour cet employe'),\n 'address_home': fields.char('Adresse Personnelle', size=128),\n 'address': fields.char('Adresse Professionnelle', size=128),\n 'phone_home': fields.char('Telephone Personnel', size=128),\n 'licexpiry': fields.char('Lic Expiry', size=128),\n 'licenseno': fields.char('Lic No', size=128),\n 'licensetyp': fields.char('Lic Type', size=128),\n 'manager': fields.boolean('Is a Manager'),\n 'medic_exam': fields.date('Medical Examination Date'),\n 'place_of_birth': fields.char('Place of Birth', size=30),\n 'cin_date': fields.date('Date CIN'),\n 'cin_place': fields.char('Lieu CIN', size=30),\n # 'children': fields.integer('Number of Children'),\n 'children': fields.function(_get_children, method=True, type='float'),\n 'vehicle': fields.char('Company Vehicle', size=64),\n 'vehicle_distance': fields.integer('Home-Work Distance',\n help=\"In kilometers\"),\n 'contract_ids': fields.one2many('hr.contract', 'employee_id',\n 'Contracts'),\n 'contract_id': fields.function(_get_latest_contract, method=True,\n string='Contract', type='many2one',\n relation=\"hr.contract\",\n help='Latest contract of the employee'),\n 'contract_type_id': fields.related('contract_id', 'type_id',\n string='Contrat', type='many2one',\n relation='hr.contract.type',\n readonly=1),\n 'aptitude_job': fields.float(u'Aptitude/Poste'),\n 'qualification_job': fields.float('Qualification/Poste'),\n 'training_job': fields.float('Formation/Poste'),\n 'is_birthday': fields.function(_get_is_birthday, method=True,\n type='boolean',\n fnct_search=_search_is_birthday),\n }\n _defaults = {\n # 'chargefam' : lambda * a: 0,\n 'logement': lambda * a: 0,\n 'anciennete': lambda * a: 'True',\n 'affilie': lambda * a: 'True',\n 'date': lambda * a: time.strftime('%Y-%m-%d'),\n 'mode_reglement': lambda * a: 'virement'\n }\n\n def name_search(self, cr, user, name, args=None, operator='ilike',\n context=None, limit=80):\n \"\"\"Search by bank code in addition to the standard search\"\"\"\n results = super(hr_employee, self).name_search(\n cr, user, name, args=args, operator=operator,\n context=context, limit=limit)\n ids = self.search(cr, user, [('matricule', operator, name)],\n limit=limit, context=context)\n # Merge the results\n results = list(set(results + self.name_get(cr, user, ids, context)))\n return results\n\n def name_search2(self, cr, user, name='', args=None, operator='ilike',\n context=None, limit=100):\n if not args:\n args = []\n if not context:\n context = {}\n ids = []\n if name:\n ids = self.search(cr, user, [('matricule', '=', name)] + args,\n limit=limit, context=context)\n if not ids:\n ids = self.search(cr, user, [('name', operator, name)] + args,\n limit=limit, context=context)\n return self.name_get(cr, user, ids, context=context)\n\n\nhr_employee()\n","sub_path":"models/hr_employee.py","file_name":"hr_employee.py","file_ext":"py","file_size_in_byte":9347,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"311492812","text":"import random\n\ngen_nange_min = 100\ngen_range_max = 999\n\ndef my_gen():\n return random.randint(gen_nange_min, gen_range_max)\n\ndef gen_cell(dim):\n arr1=[]\n for i in range( dim ):\n arr1.append( my_gen() )\n return arr1\n\ndef gen_list( dimention ):\n l=[]\n for i in range( dimention ):\n l.append(gen_cell( dimention ))\n return l\n\n# -Заменить по главной диагонали все числа на 0\ndef list_replace_diagonal(mylist):\n for i in range(len(mylist)):\n for j in range(len(mylist)):\n if i==j:\n mylist[i][j] = 0\n\n return mylist\n\n# -Заменить все четные числа на 1, не четные на 0\ndef is_odd(n):\n if n % 2 == 0:\n return True\n else:\n return False\n\ndef list_replace_num(mylist):\n for i in range(len(mylist)):\n for j in range(len(mylist)):\n if is_odd(mylist[i][j]):\n mylist[i][j] = 1\n else:\n mylist[i][j] = 0\n\n return mylist\n\n# -Вывести строку таблицы с максимальной суммой элементов\ndef list_max_sum_num(mylist):\n max_value=0\n sum = 0\n row_with_max_sum = 0\n\n for i in mylist:\n for j in mylist[i]:\n sum = sum + mylist[i][j]\n if sum>=max_value:\n max_value=sum\n row_with_max_sum=i\n\n return mylist[row_with_max_sum]\n\nnew_list = gen_list(4)\nprint(new_list)\nprint(list_replace_diagonal(new_list))\nprint(list_replace_num(new_list))\nprint(list_max_sum_num(new_list))\n","sub_path":"HW5_Lists/hw5_task3.py","file_name":"hw5_task3.py","file_ext":"py","file_size_in_byte":1572,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"175204693","text":"import random\r\n\r\ndef normlis(lis):\r\n maxi = max(lis)\r\n mini = min(lis)\r\n print(maxi,mini)\r\n newlis = []\r\n for num in lis:\r\n afternum = float(num-mini)/float(maxi-mini)\r\n newlis.append(afternum)\r\n return newlis\r\n\r\nfin = open(\"D:\\\\graduatedesign\\\\Data\\\\train_test.dat\")\r\nftrain = open(\"C:\\\\Users\\\\Allen\\\\Desktop\\\\svm_rank_windows\\\\train.dat\",'w')\r\nftest = open(\"C:\\\\Users\\\\Allen\\\\Desktop\\\\svm_rank_windows\\\\test.dat\",'w')\r\nfall = open(\"C:\\\\Users\\\\Allen\\\\Desktop\\\\svm_rank_windows\\\\all.dat\",'w')\r\n\r\ntimes = []\r\nreadcounts = []\r\ncomments = []\r\nzan = []\r\npr = []\r\nhub = []\r\nauth = []\r\n\r\nfor line in fin.readlines():\r\n lis = line.split('\\t')\r\n times.append(float(lis[1]))\r\n readcounts.append(float(lis[2]))\r\n comments.append(float(lis[3]))\r\n zan.append(float(lis[4]))\r\n pr.append(float(lis[5])*1000)\r\n hub.append(float(lis[6])*1000)\r\n auth.append(float(lis[7])*1000000)\r\n\r\ntimes = normlis(times)\r\nreadcounts = normlis(readcounts)\r\ncomments = normlis(comments)\r\n#print(zan[20:28])\r\nzan = normlis(zan)\r\n#print(zan[20:28])\r\n#print(pr[0:20])\r\npr = normlis(pr)\r\n#print(pr[0:20])\r\nhub = normlis(hub)\r\nauth = normlis(auth)\r\n\r\nscore = [10,3,5,15,21,12,8,6,9,8,11,13,19,16,11,7,17,17,4,20]\r\n#8,2,4,9,10,8,6,4,5,7\r\n#5,6,8,7,4,3,2,9,1,10\r\n#2,3,5,4,1\r\n\r\ndef WRITE(f,qid,i):\r\n if i < 20:\r\n f.write(\"%d \" %score[i])\r\n else:\r\n f.write(\"%d \" %i)\r\n f.write(\"qid:\"+str(qid))\r\n #f.write(\" 1:%.3f\" %(0.001*times[i]))\r\n f.write(\" 1:%.3f\" %(readcounts[i]))\r\n f.write(\" 2:%.3f\" %(comments[i]))\r\n f.write(\" 3:%.3f\" %(zan[i]))\r\n f.write(\" 4:%.3f\" %(0.8*pr[i]+0.05*hub[i]+0.15*auth[i]))\r\n #print(\" 4:%.3f\" %(0.8*pr[i]+0.05*hub[i]+0.15*auth[i]))\r\n f.write('\\n')\r\n\r\ntraincount = 0\r\ntestcount = 0\r\nc = 0\r\nfor i in range(len(times)):\r\n randnum = random.uniform(0,1)\r\n if c < 20:\r\n WRITE(ftrain,1,i)\r\n traincount += 1\r\n else:\r\n WRITE(ftest,1,i)\r\n testcount += 1\r\n WRITE(fall,1,i)\r\n c += 1\r\n\r\nprint(\"TrainCounts: \",traincount)\r\nprint(\"TestCounts: \",testcount)\r\n\r\n\r\n\r\n\r\nfin.close()\r\nftrain.close()\r\nftest.close()\r\nfall.close()\r\n","sub_path":"py/Norm.py","file_name":"Norm.py","file_ext":"py","file_size_in_byte":2132,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"496524531","text":"#-------------------------------------------------------------------\n# Project: Zom-botapacolyse\n# Name: Man.py\n# Purpose: The Player Object\n# \n# Authors: Michael Simon, Matt Hahn, Tim Richter\n# Main Author: Michael Simon\n#\n# Created: 12/11/12\n# Copyright: (c) Michael Simon 2012\n# License: GSL\n#-------------------------------------------------------------------\nimport pygame, math\nfrom Zombie import Zombie\nfrom Robot import Robot\nfrom MazeWall import MazeWall\n\n\nclass Man():\n # Attributes or Variables\n # surface\n # rect\n # distToCenter\n # speed\n # screenWidth\n # screenHeight\n # living\n \n # Methods or Functions\n def __init__(self, maxSpeed, position):\n #Up/North\n self.surfacesUpNothing = [pygame.image.load(\"rsc\\man\\mann.png\"), pygame.image.load(\"rsc\\man\\mann1.png\"), pygame.image.load(\"rsc\\man\\mann2.png\")]\n self.surfacesUpStick = [pygame.image.load(\"rsc\\man\\mannS.png\"), pygame.image.load(\"rsc\\man\\mannS1.png\"), pygame.image.load(\"rsc\\man\\mannS2.png\")]\n self.surfacesUpGun = [pygame.image.load(\"rsc\\man\\mannG.png\"), pygame.image.load(\"rsc\\man\\mannG1.png\"), pygame.image.load(\"rsc\\man\\mannG2.png\")]\n #Down/South\n self.surfacesDownNothing = [pygame.image.load(\"rsc\\man\\mans.png\"), pygame.image.load(\"rsc\\man\\mans1.png\"), pygame.image.load(\"rsc\\man\\mans2.png\")]\n self.surfacesDownStick = [pygame.image.load(\"rsc\\man\\mansS.png\"), pygame.image.load(\"rsc\\man\\mansS1.png\"), pygame.image.load(\"rsc\\man\\mansS2.png\")]\n self.surfacesDownGun = [pygame.image.load(\"rsc\\man\\mansG.png\"), pygame.image.load(\"rsc\\man\\mansG1.png\"), pygame.image.load(\"rsc\\man\\mansG2.png\")]\n #Right/East\n self.surfacesRightNothing = [pygame.image.load(\"rsc\\man\\mane.png\"), pygame.image.load(\"rsc\\man\\mane1.png\"), pygame.image.load(\"rsc\\man\\mane2.png\")]\n self.surfacesRightStick = [pygame.image.load(\"rsc\\man\\maneS.png\"), pygame.image.load(\"rsc\\man\\maneS1.png\"), pygame.image.load(\"rsc\\man\\maneS2.png\")]\n self.surfacesRightGun = [pygame.image.load(\"rsc\\man\\maneG.png\"), pygame.image.load(\"rsc\\man\\maneG1.png\"), pygame.image.load(\"rsc\\man\\maneG2.png\")]\n #Left/West\n self.surfacesLeftNothing = [pygame.image.load(\"rsc\\man\\manw.png\"), pygame.image.load(\"rsc\\man\\manw1.png\"), pygame.image.load(\"rsc\\man\\manw2.png\")]\n self.surfacesLeftStick = [pygame.image.load(\"rsc\\man\\manwS.png\"), pygame.image.load(\"rsc\\man\\manwS1.png\"), pygame.image.load(\"rsc\\man\\manwS2.png\")]\n self.surfacesLeftGun = [pygame.image.load(\"rsc\\man\\manwG.png\"), pygame.image.load(\"rsc\\man\\manwG1.png\"), pygame.image.load(\"rsc\\man\\manwG2.png\")]\n \n self.dir = \"stop down\"\n self.heading = \"s\"\n self.surfaces = self.surfacesDownNothing\n self.frame = 0\n self.wait = 0\n self.waitMax = 5\n self.maxFrame = len(self.surfaces)-1\n self.surface = self.surfaces[self.frame]\n self.rect = self.surface.get_rect()\n self.radius = self.rect.width/2.5\n self.maxSpeed = maxSpeed\n self.speed = [0,0]\n self.noSpeed = 0\n self.place(position)\n self.attackRadius = 40\n self.life = 100\n self.living = False\n self.haveNothing = True\n self.haveStick = False\n self.havePistol = False\n self.win = False\n self.ammo = 20\n def __str__(self):\n return \"I am the Man\" + str(self.rect.center) + str(self.speed) + str(self.living)\n \n def place(self, position):\n self.rect = self.rect.move(position)\n \n def move(self):\n if self.dir[0] == \"s\":\n self.frame = 0\n elif self.wait >= self.waitMax:\n self.wait = 0\n if self.frame < self.maxFrame:\n self.frame += 1\n else:\n self.frame = 1\n else:\n self.wait += 1\n \n self.surface = self.surfaces[self.frame]\n\n self.rect = self.rect.move(self.speed)\n \n def checkHave (self):\n if self.haveStick == True:\n self.havePistol = False\n self.haveNothing = False\n if self.havePistol == True:\n self.haveStick = False\n self.haveNothing = False\n else:\n self.haveNothing == True\n self.haveStick = False\n self.havePistol = False\n \n def direction(self, dir):\n if dir == \"up\":\n if self.haveNothing:\n self.surfaces = self.surfacesUpNothing\n if self.havePistol:\n self.surfaces = self.surfacesUpGun\n if self.haveStick:\n self.surfaces = self.surfacesUpStick\n self.speed[1] = -self.maxSpeed\n self.dir = dir\n self.heading = \"n\"\n elif dir == \"down\":\n if self.haveNothing:\n self.surfaces = self.surfacesDownNothing\n if self.havePistol:\n self.surfaces = self.surfacesDownGun\n if self.haveStick:\n self.surfaces = self.surfacesDownStick\n self.speed[1] = self.maxSpeed\n self.dir = dir\n self.heading = \"s\"\n elif dir == \"stop up\":\n self.speed[1] = self.noSpeed\n self.dir = dir\n elif dir == \"stop down\":\n self.speed[1] = self.noSpeed\n self.dir = dir\n \n if dir == \"right\":\n if self.haveNothing:\n self.surfaces = self.surfacesRightNothing\n if self.havePistol:\n self.surfaces = self.surfacesRightGun\n if self.haveStick:\n self.surfaces = self.surfacesRightStick\n self.speed[0] = self.maxSpeed\n self.dir = dir\n self.heading = \"e\"\n elif dir == \"left\":\n if self.haveNothing:\n self.surfaces = self.surfacesLeftNothing\n if self.havePistol:\n self.surfaces = self.surfacesLeftGun\n if self.haveStick:\n self.surfaces = self.surfacesLeftStick\n self.speed[0] = -self.maxSpeed\n self.dir = dir\n self.heading = \"w\"\n elif dir == \"stop right\":\n self.speed[0] = self.noSpeed\n self.dir = dir\n elif dir == \"stop left\":\n self.speed[0] = self.noSpeed\n self.dir = dir\n \n def distToPointWithOffset(self, pt, offset):\n x1 = self.rect.center[0] + offset[0]\n x2 = pt[0]\n y1 = self.rect.center[1] + offset[1]\n y2 = pt[1]\n return math.sqrt(((x2-x1)**2)+((y2-y1)**2))\n\n def distToPoint(self, pt):\n x1 = self.rect.center[0]\n x2 = pt[0]\n y1 = self.rect.center[1]\n y2 = pt[1]\n return math.sqrt(((x2-x1)**2)+((y2-y1)**2))\n \n def collideWall(self, screenWidth, screenHeight):\n if (self.rect.left < 0 \n or self.rect.right > screenWidth):\n self.speed[0] = self.speed[0]*0\n if (self.rect.top < 0 \n or self.rect.bottom > screenHeight):\n self.speed[1] = self.speed[1]*0\n \n if (self.rect.top < 0):\n self.speed = [0, 1]\n if (self.rect.top > 0):\n self.speed = [0, 0]\n run = False\n if (self.rect.left < 0):\n self.speed = [1, 0]\n if (self.rect.left > 0):\n self.speed = [0, 0]\n run = False\n if (self.rect.right > screenWidth):\n self.speed = [-1, 0]\n if (self.rect.right < screenWidth):\n self.speed = [0, 0]\n run = False\n if (self.rect.bottom > screenHeight):\n self.speed = [0, -1]\n if (self.rect.bottom < screenHeight):\n self.speed = [0, 0]\n run = False\n\n def collideMazeWall(self, MazeWall):\n if (self.rect.right > MazeWall.rect.left \n and self.rect.left < MazeWall.rect.right):\n if (self.rect.bottom > MazeWall.rect.top and \n self.rect.top < MazeWall.rect.bottom):\n if (self.distToPointWithOffset(MazeWall.rect.center, [0,5])\n < self.radius + MazeWall.radius): \n \n \n self.speed[0] = self.speed[0] * -1\n self.speed[1] = self.speed[1] * -1\n \n self.move()\n self.move()\n \n \n self.speed[0] = 0\n self.speed[1] = 0\n \n def collideWinBlock(self, winblock):\n if (self.rect.right > winblock.rect.left \n and self.rect.left < winblock.rect.right):\n if (self.rect.bottom > winblock.rect.top and \n self.rect.top < winblock.rect.bottom):\n if (self.distToPointWithOffset(winblock.rect.center, [0,5])\n < self.radius + winblock.radius):\n self.win = True\n self.living = True\n run = False\n \n def collideRobot(self, other):\n pass\n\n def collideStick(self, stick):\n if (self.rect.right > stick.rect.left \n and self.rect.left < stick.rect.right):\n if (self.rect.bottom > stick.rect.top and \n self.rect.top < stick.rect.bottom):\n if (self.distToPoint(stick.rect.center)\n < self.radius + stick.radius):\n self.haveStick = True\n stick.notBroken = False\n \n def collidePistol(self, pistol):\n if (self.rect.right > pistol.rect.left \n and self.rect.left < pistol.rect.right):\n if (self.rect.bottom > pistol.rect.top and \n self.rect.top < pistol.rect.bottom):\n if (self.distToPoint(pistol.rect.center)\n < self.radius + pistol.radius):\n self.havePistol = True\n pistol.notBroken = False\n\n\n \n def dead(self):\n if self.life <= 0:\n self.living = False\n \n ","sub_path":"Man.py","file_name":"Man.py","file_ext":"py","file_size_in_byte":10261,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"390554699","text":"# Example of a function based API adapter\n# for a CLI only program / script\n\nfrom __future__ import print_function, absolute_import\nfrom subprocess import Popen, PIPE, STDOUT\nfrom tempfile import NamedTemporaryFile\nimport os, sys\n\ncommand_to_run = '{} {} -v'.format(\n sys.executable,\n 'python2_cli-base.py',\n)\n\ndef run_program(input_text):\n \"\"\"\n Default when we don't care about\n the underlying method\n \"\"\"\n return run_program_pipe(input_text)\n\ndef run_program_pipe(input_text):\n \"\"\"\n Simulate: echo 'input' | \n \"\"\"\n global command_to_run\n args = command_to_run.split(\" \")\n\n pipe = Popen(args,\n shell=True, stdout=PIPE, stdin=PIPE, stderr=STDOUT)\n output = pipe.communicate(input=input_text)[0]\n return output\n\ndef run_program_file(input_text):\n \"\"\"\n Simulate: filename\n \"\"\"\n global command_to_run\n args = command_to_run.split(\" \")\n\n temp_file = NamedTemporaryFile(delete=False)\n temp_file.write(input_text)\n temp_file.close()\n args.append(temp_file.name)\n\n pipe = Popen(args,\n shell=True, stdout=PIPE, stderr=STDOUT)\n output = pipe.communicate()[0]\n os.unlink(temp_file.name)\n return output\n\n\ndef main():\n print(run_program('Testing via default method'))\n print(run_program_pipe('Testing via pipe input'))\n print(run_program_file('Testing via file input'))\n return 0\n\n\nif __name__ == '__main__':\n sys.exit(main())\n","sub_path":"python2_api-adapter.py","file_name":"python2_api-adapter.py","file_ext":"py","file_size_in_byte":1447,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"475135151","text":"# SSW 540 - Assignment 08 - P8: Counting unique items\n# Akshay Sunderwani\n\nimport re\n\n\ndef parsefileforpath(filepath):\n try:\n # opening the file\n with open ( filepath ) as filetoread:\n maindict = {}\n noofemailrecevideddictionary = {}\n total_count = 0\n for lines in filetoread:\n linelist = lines.split ( )\n if len ( linelist ) > 0:\n # check for tag \"From:\" for getting email ids\n if 'From:' in linelist[ 0 ]:\n for items in linelist:\n if linelist.index ( items ) != 0:\n # used regex to identify valid email address and add it to the dictionary and maintain\n # its count\n if re.match (\n r\"^[a-zA-Z0-9.!#$%&'*+/=?^_`{|}~-]+@[a-zA-Z0-9](?:[a-zA-Z0-9-]{0,61}[a-zA-Z0-9])?(?:\\.[a-zA-Z0-9](?:[a-zA-Z0-9-]{0,61}[a-zA-Z0-9])?)*$\" , items ):\n if items in noofemailrecevideddictionary:\n noofemailrecevideddictionary[items] += 1\n else:\n noofemailrecevideddictionary[items] = 1\n total_count += 1\n maindict['total_count'] = total_count\n maindict['email_info'] = noofemailrecevideddictionary\n return maindict\n except EnvironmentError as error:\n return error\n\n\npath = input ( \"Please provide path for file to read : \" ) # input path of file from user.\nnoofunique = parsefileforpath ( path ) # call parser method to read\nif not isinstance(noofunique, dict):\n print('ERROR : ', noofunique)\nelse:\n print('\\nTotal no of email sent : ', noofunique['total_count'])\n # if len(noofunique['email_info'].keys()) > 0 and noofunique['total_count'] > 0:\n if len ( noofunique[ 'email_info' ] ) > 0:\n print('\\nTotal by each email address :')\n print ( \"\\n\".join ( \"{}: {}\".format ( k , v ) for k , v in noofunique[ 'email_info' ].items ( ) ) )\n max_val = max(noofunique['email_info'].values())\n print('\\nMax no of emails sent by :')\n for keys in noofunique['email_info']:\n if noofunique['email_info'][keys] == max_val:\n print(keys, ' : ', max_val, '\\n')\n","sub_path":"untitled/SSW540_ASSIGNMENT/ASSIGNMENT06_AKSHAYSUNDERWANI_SSW540/Project08_AkshaySunderwani_SSW540-1.py","file_name":"Project08_AkshaySunderwani_SSW540-1.py","file_ext":"py","file_size_in_byte":2434,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"630706695","text":"\"\"\"\nPreferences is a collection of utilities to display, read & write preferences.\n\n\"\"\"\n\nfrom __future__ import absolute_import\n#Init has to be imported first because it has code to workaround the python bug where relative imports don't work if the module is imported as a main module.\nimport __init__\n\nimport cStringIO\nfrom skeinforge_tools.skeinforge_utilities import gcodec\nimport os\nimport webbrowser\ntry:\n\timport Tkinter\nexcept:\n\tprint( 'You do not have Tkinter, which is needed for the graphical interface, you will only be able to use the command line.' )\n\tprint( 'Information on how to download Tkinter is at:\\nwww.tcl.tk/software/tcltk/' )\n\n\n__author__ = \"Enrique Perez (perez_enrique@yahoo.com)\"\n__date__ = \"$Date: 2008/23/04 $\"\n__license__ = \"GPL 3.0\"\n\nglobalIsMainLoopRunning = False\nglobalSpreadsheetSeparator = '\\t'\n\ndef displayDialog( displayPreferences ):\n\t\"Display the preferences dialog.\"\n\treadPreferences( displayPreferences )\n\troot = Tkinter.Tk()\n\tpreferencesDialog = PreferencesDialog( displayPreferences, root )\n\tglobal globalIsMainLoopRunning\n#\tprint( globalIsMainLoopRunning )\n\tif globalIsMainLoopRunning:\n\t\treturn\n\tglobalIsMainLoopRunning = True\n\troot.mainloop()\n\tglobalIsMainLoopRunning = False\n\ndef getArchiveText( archivablePreferences ):\n\t\"Get the text representation of the archive.\"\n\tarchiveWriter = cStringIO.StringIO()\n\tarchiveWriter.write( 'Format is tab separated preferences.\\n' )\n\tfor preference in archivablePreferences.archive:\n\t\tpreference.writeToArchiveWriter( archiveWriter )\n\treturn archiveWriter.getvalue()\n\ndef getFileInGivenDirectory( directory, fileName ):\n\t\"Get the file from the fileName or the lowercase fileName in the given directory.\"\n\tdirectoryListing = os.listdir( directory )\n\tif fileName in directoryListing:\n\t\treturn getFileTextGivenDirectoryFileName( directory, fileName )\n\tlowerFilename = fileName.lower()\n\tif lowerFilename in directoryListing:\n\t\treturn getFileTextGivenDirectoryFileName( directory, lowerFilename )\n\treturn ''\n\ndef getFileInGivenPreferencesDirectory( directory, fileName ):\n\t\"Get the file from the fileName or the lowercase fileName in the given directory, if there is no file look in the gcode_scripts folder in the preferences directory.\"\n\tif directory == '':\n\t\tdirectory = os.getcwd()\n\tfileInGivenPreferencesDirectory = getFileInGivenDirectory( directory, fileName )\n\tif fileInGivenPreferencesDirectory != '':\n\t\treturn fileInGivenPreferencesDirectory\n\tgcodeDirectoryPath = os.path.join( getPreferencesDirectoryPath(), 'gcode_scripts' )\n\ttry:\n\t\tos.mkdir( gcodeDirectoryPath )\n\texcept OSError:\n\t\tpass\n\treturn getFileInGivenDirectory( gcodeDirectoryPath, fileName )\n\ndef getFileTextGivenDirectoryFileName( directory, fileName ):\n\t\"Get the entire text of a file with the given file name in the given directory.\"\n\tabsoluteFilePath = os.path.join( directory, fileName )\n\treturn gcodec.getFileText( absoluteFilePath )\n\ndef getPreferencesDirectoryPath():\n\t\"Get the preferences directory path, which is the home directory joined with .skeinforge.\"\n\treturn os.path.join( os.path.expanduser( '~' ), '.skeinforge' )\n\ndef getPreferencesFilePath( fileName ):\n\t\"Get the preferences file path, which is the home directory joined with .skeinforge and fileName.\"\n\tdirectoryName = getPreferencesDirectoryPath()\n\ttry:\n\t\tos.mkdir( directoryName )\n\texcept OSError:\n\t\tpass\n\treturn os.path.join( directoryName, fileName )\n\ndef readPreferences( archivablePreferences ):\n\t\"Set an archive to the preferences read from a file.\"\n\ttext = gcodec.getFileText( archivablePreferences.fileNamePreferences )\n\tif text == '':\n\t\tprint( 'Since the preferences file:' )\n\t\tprint( archivablePreferences.fileNamePreferences )\n\t\tprint( 'does not exist, the default preferences will be written to that file.' )\n\t\ttext = gcodec.getFileText( os.path.join( 'defaults', os.path.basename( archivablePreferences.fileNamePreferences ) ) )\n\t\tif text != '':\n\t\t\treadPreferencesFromText( archivablePreferences, text )\n\t\twritePreferences( archivablePreferences )\n\t\treturn\n\treadPreferencesFromText( archivablePreferences, text )\n\ndef readPreferencesFromText( archivablePreferences, text ):\n\t\"Set an archive to the preferences read from a text.\"\n\tlines = gcodec.getTextLines( text )\n\tpreferenceTable = {}\n\tfor preference in archivablePreferences.archive:\n\t\tpreference.addToPreferenceTable( preferenceTable )\n\tfor lineIndex in xrange( len( lines ) ):\n\t\tsetArchiveToLine( lineIndex, lines, preferenceTable )\n\ndef setArchiveToLine( lineIndex, lines, preferenceTable ):\n\t\"Set an archive to a preference line.\"\n\tline = lines[ lineIndex ]\n\tsplitLine = line.split( globalSpreadsheetSeparator )\n\tif len( splitLine ) < 2:\n\t\treturn\n\tfilePreferenceName = splitLine[ 0 ]\n\tif filePreferenceName in preferenceTable:\n\t\tpreferenceTable[ filePreferenceName ].setValueToSplitLine( lineIndex, lines, splitLine )\n\ndef setHelpPreferencesFileNameTitleWindowPosition( displayPreferences, fileNameHelp ):\n\t\"Set the help & preferences file path, the title and the window position archiver.\"\n\tDotIndex = fileNameHelp.rfind( '.' )\n\tlastDotIndex = fileNameHelp.rfind( '.' )\n\tlowerName = fileNameHelp[ : lastDotIndex ]\n\tlastTruncatedDotIndex = lowerName.rfind( '.' )\n\tlowerName = lowerName[ lastTruncatedDotIndex + 1 : ]\n\tdisplayPreferences.title = lowerName.replace( '_', ' ' ).capitalize() + ' Preferences'\n\twindowPositionName = 'windowPosition' + displayPreferences.title\n\tdisplayPreferences.windowPositionBeholdPreferences = WindowPosition().getFromValue( 'windowPositionBehold Preferences', '0+0' )\n\tdisplayPreferences.archive.append( displayPreferences.windowPositionBeholdPreferences )\n\tdisplayPreferences.fileNamePreferences = getPreferencesFilePath( lowerName + '.csv' )\n\tdisplayPreferences.fileNameHelp = fileNameHelp\n\ndef writePreferences( archivablePreferences ):\n\t\"Write the preferences to a file.\"\n\tgcodec.writeFileText( archivablePreferences.fileNamePreferences, getArchiveText( archivablePreferences ) )\n\n\nclass AddListboxSelection:\n\t\"A class to add the selection of a listbox preference.\"\n\tdef addToDialog( self, preferencesDialog ):\n\t\t\"Add this to the dialog.\"\n\t\tself.entry = Tkinter.Entry( preferencesDialog.master )\n\t\tself.entry.bind( '', self.addSelectionWithEvent )\n\t\tself.entry.grid( row = preferencesDialog.row, column = 1, columnspan = 2, sticky = Tkinter.W )\n\t\tself.addButton = Tkinter.Button( preferencesDialog.master, text = 'Add Listbox Selection', command = self.addSelection )\n\t\tself.addButton.grid( row = preferencesDialog.row, column = 0 )\n\t\tpreferencesDialog.row += 1\n\n\tdef addSelection( self ):\n\t\t\"Add the selection of a listbox preference.\"\n\t\tentryText = self.entry.get()\n\t\tif entryText == '':\n\t\t\tprint( 'To add to the selection, enter the material name.' )\n\t\t\treturn\n\t\tself.entry.delete( 0, Tkinter.END )\n\t\tself.listboxPreference.listPreference.value.append( entryText )\n\t\tself.listboxPreference.listPreference.value.sort()\n\t\tself.listboxPreference.listbox.delete( 0, Tkinter.END )\n\t\tself.listboxPreference.value = entryText\n\t\tself.listboxPreference.setListboxItems()\n\t\tself.listboxPreference.setToDisplay()\n\n\tdef addSelectionWithEvent( self, event ):\n\t\t\"Add the selection of a listbox preference, given an event.\"\n\t\tself.addSelection()\n\n\tdef addToPreferenceTable( self, preferenceTable ):\n\t\t\"Do nothing because the add listbox selection is not archivable.\"\n\t\tpass\n\n\tdef getFromListboxPreference( self, listboxPreference ):\n\t\t\"Initialize.\"\n\t\tself.listboxPreference = listboxPreference\n\t\treturn self\n\n\tdef setToDisplay( self ):\n\t\t\"Do nothing because the add listbox selection is not archivable.\"\n\t\tpass\n\n\tdef writeToArchiveWriter( self, archiveWriter ):\n\t\t\"Do nothing because the add listbox selection is not archivable.\"\n\t\tpass\n\n\nclass StringPreference:\n\t\"A class to display, read & write a string.\"\n\tdef __init__( self ):\n\t\t\"Set the update function to none.\"\n\t\tself.updateFunction = None\n\n\tdef addToDialog( self, preferencesDialog ):\n\t\t\"Add this to the dialog.\"\n\t\tself.entry = Tkinter.Entry( preferencesDialog.master )\n\t\tself.setStateToValue()\n\t\tself.entry.grid( row = preferencesDialog.row, column = 2, columnspan = 2, sticky = Tkinter.W )\n\t\tself.label = Tkinter.Label( preferencesDialog.master, text = self.name )\n\t\tself.label.grid( row = preferencesDialog.row, column = 0, columnspan = 2, sticky = Tkinter.W )\n\t\tpreferencesDialog.row += 1\n\n\tdef addToPreferenceTable( self, preferenceTable ):\n\t\t\"Add this to the preference table.\"\n\t\tpreferenceTable[ self.name ] = self\n\n\tdef getFromValue( self, name, value ):\n\t\t\"Initialize.\"\n\t\tself.value = value\n\t\tself.name = name\n\t\treturn self\n\n\tdef setStateToValue( self ):\n\t\t\"Set the entry to the value.\"\n\t\ttry:\n\t\t\tself.entry.delete( 0, Tkinter.END )\n\t\t\tself.entry.insert( 0, self.value )\n\t\texcept:\n\t\t\tpass\n\n\tdef setToDisplay( self ):\n\t\t\"Set the string to the entry field.\"\n\t\tvalueString = self.entry.get()\n\t\tself.setValueToString( valueString )\n\n\tdef setUpdateFunction( self, updateFunction ):\n\t\t\"Set the update function.\"\n\t\tself.updateFunction = updateFunction\n\n\tdef setValueToSplitLine( self, lineIndex, lines, splitLine ):\n\t\t\"Set the value to the second word of a split line.\"\n\t\tself.setValueToString( splitLine[ 1 ] )\n\n\tdef setValueToString( self, valueString ):\n\t\t\"Set the string to the value string.\"\n\t\tself.value = valueString\n\n\tdef writeToArchiveWriter( self, archiveWriter ):\n\t\t\"Write tab separated name and value to the archive writer.\"\n\t\tarchiveWriter.write( self.name + globalSpreadsheetSeparator + str( self.value ) + '\\n' )\n\n\nclass BooleanPreference( StringPreference ):\n\t\"A class to display, read & write a boolean.\"\n\tdef addToDialog( self, preferencesDialog ):\n\t\t\"Add this to the dialog.\"\n\t\tself.checkbutton = Tkinter.Checkbutton( preferencesDialog.master, command = self.toggleCheckbox, text = self.name )\n#toggleCheckbox is being used instead of a Tkinter IntVar because there is a weird bug where it doesn't work properly if this preference is not on the first window.\n\t\tself.checkbutton.grid( row = preferencesDialog.row, columnspan = 4, sticky = Tkinter.W )\n\t\tself.setStateToValue()\n\t\tpreferencesDialog.row += 1\n\n\tdef setStateToValue( self ):\n\t\t\"Set the checkbox to the boolean.\"\n\t\ttry:\n\t\t\tif self.value:\n\t\t\t\tself.checkbutton.select()\n\t\t\telse:\n\t\t\t\tself.checkbutton.deselect()\n\t\texcept:\n\t\t\tpass\n\n\tdef setToDisplay( self ):\n\t\t\"Do nothing because toggleCheckbox is handling the value.\"\n\t\tpass\n\n\tdef setValueToString( self, valueString ):\n\t\t\"Set the boolean to the string.\"\n\t\tself.value = ( valueString.lower() == 'true' )\n\n\tdef toggleCheckbox( self ):\n\t\t\"Workaround for Tkinter bug, toggle the value.\"\n\t\tself.value = not self.value\n\t\tself.setStateToValue()\n\t\tif self.updateFunction != None:\n\t\t\tself.updateFunction()\n\n\nclass DeleteListboxSelection( AddListboxSelection ):\n\t\"A class to delete the selection of a listbox preference.\"\n\tdef addToDialog( self, preferencesDialog ):\n\t\t\"Add this to the dialog.\"\n\t\tself.deleteButton = Tkinter.Button( preferencesDialog.master, text = \"Delete Listbox Selection\", command = self.deleteSelection )\n\t\tself.deleteButton.grid( row = preferencesDialog.row, column = 0 )\n\t\tpreferencesDialog.row += 1\n\n\tdef deleteSelection( self ):\n\t\t\"Delete the selection of a listbox preference.\"\n\t\tself.listboxPreference.setToDisplay()\n\t\tif self.listboxPreference.value not in self.listboxPreference.listPreference.value:\n\t\t\treturn\n\t\tself.listboxPreference.listPreference.value.remove( self.listboxPreference.value )\n\t\tself.listboxPreference.listbox.delete( 0, Tkinter.END )\n\t\tself.listboxPreference.setListboxItems()\n\t\tself.listboxPreference.listbox.select_set( 0 )\n\t\tself.listboxPreference.setToDisplay()\n\n\nclass DisplayToolButton:\n\t\"A class to display the tool preferences dialog.\"\n\tdef addToDialog( self, preferencesDialog ):\n\t\t\"Add this to the dialog.\"\n\t\twithSpaces = self.name.lower().replace( '_', ' ' )\n\t\twords = withSpaces.split( ' ' )\n\t\tcapitalizedStrings = []\n\t\tfor word in words:\n\t\t\tcapitalizedStrings.append( word.capitalize() )\n\t\tcapitalizedName = ' '.join( capitalizedStrings )\n\t\tself.displayButton = Tkinter.Button( preferencesDialog.master, activebackground = 'black', activeforeground = 'violet', command = self.displayTool, text = capitalizedName )\n\t\tif preferencesDialog.displayToolButtonStart:\n\t\t\tself.displayButton.grid( row = preferencesDialog.row, column = 0 )\n\t\t\tpreferencesDialog.row += 1\n\t\t\tpreferencesDialog.displayToolButtonStart = False\n\t\telse:\n\t\t\tself.displayButton.grid( row = preferencesDialog.row - 1, column = 3 )\n\t\t\tpreferencesDialog.displayToolButtonStart = True\n\n\tdef addToPreferenceTable( self, preferenceTable ):\n\t\t\"Do nothing because the add listbox selection is not archivable.\"\n\t\tpass\n\n\tdef displayTool( self ):\n\t\t\"Display the tool preferences dialog.\"\n\t\tpluginModule = gcodec.getModule( self.name, self.folderName, self.moduleFilename )\n\t\tif pluginModule != None:\n\t\t\tpluginModule.main()\n\n\tdef getFromFolderName( self, folderName, moduleFilename, name ):\n\t\t\"Initialize.\"\n\t\tself.folderName = folderName\n\t\tself.moduleFilename = moduleFilename\n\t\tself.name = name\n\t\treturn self\n\n\tdef getLowerName( self ):\n\t\t\"Get the lower case name.\"\n\t\treturn self.name.lower()\n\n\tdef setToDisplay( self ):\n\t\t\"Do nothing because the display tool button is not archivable.\"\n\t\tpass\n\n\tdef writeToArchiveWriter( self, archiveWriter ):\n\t\t\"Do nothing because the display tool button is not archivable.\"\n\t\tpass\n\n\nclass DisplayToolButtonBesidePrevious( DisplayToolButton ):\n\t\"A class to display the tool preferences dialog beside the previous preference dialog element.\"\n\tdef addToDialog( self, preferencesDialog ):\n\t\t\"Add this to the dialog.\"\n\t\twithSpaces = self.name.lower().replace( '_', ' ' )\n\t\twords = withSpaces.split( ' ' )\n\t\tcapitalizedStrings = []\n\t\tfor word in words:\n\t\t\tcapitalizedStrings.append( word.capitalize() )\n\t\tcapitalizedName = ' '.join( capitalizedStrings )\n\t\tself.displayButton = Tkinter.Button( preferencesDialog.master, text = capitalizedName, command = self.displayTool )\n\t\tself.displayButton.grid( row = preferencesDialog.row - 1, column = 2, columnspan = 2 )\n\n\nclass Filename( StringPreference ):\n\tdef addToDialog( self, preferencesDialog ):\n\t\t\"Add this to the dialog.\"\n\t\tpreferencesDialog.executables.append( self )\n\n\t\"A class to display, read & write a fileName.\"\n\tdef execute( self ):\n\t\ttry:\n\t\t\timport tkFileDialog\n\t\t\tsummarized = gcodec.getSummarizedFilename( self.value )\n\t\t\tinitialDirectory = os.path.dirname( summarized )\n\t\t\tif len( initialDirectory ) > 0:\n\t\t\t\tinitialDirectory += os.sep\n\t\t\telse:\n\t\t\t\tinitialDirectory = \".\"\n\t\t\tfileName = tkFileDialog.askopenfilename( filetypes = self.getFilenameFirstTypes(), initialdir = initialDirectory, initialfile = os.path.basename( summarized ), title = self.name )\n\t\t\tif ( str( fileName ) == '()' or str( fileName ) == '' ):\n\t\t\t\tself.wasCancelled = True\n\t\t\telse:\n\t\t\t\tself.value = fileName\n\t\texcept:\n\t\t\tprint( 'Oops, ' + self.name + ' could not get fileName.' )\n\n\tdef getFromFilename( self, fileTypes, name, value ):\n\t\t\"Initialize.\"\n\t\tself.getFromValue( name, value )\n\t\tself.fileTypes = fileTypes\n\t\tself.wasCancelled = False\n\t\treturn self\n\n\tdef getFilenameFirstTypes( self ):\n\t\t\"Get the file types with the file type of the fileName moved to the front of the list.\"\n\t\tbasename = os.path.basename( self.value )\n\t\tsplitFile = basename.split( '.' )\n\t\tallReadables = []\n\t\tif len( self.fileTypes ) > 1:\n\t\t\tfor fileType in self.fileTypes:\n\t\t\t\tallReadable = ( ( 'All Readable', fileType[ 1 ] ) )\n\t\t\t\tallReadables.append( allReadable )\n\t\tif len( splitFile ) < 1:\n\t\t\treturn self.fileTypes + allReadables\n\t\tbaseExtension = splitFile[ - 1 ]\n\t\tfor fileType in self.fileTypes:\n\t\t\tfileExtension = fileType[ 1 ].split( '.' )[ - 1 ]\n\t\t\tif fileExtension == baseExtension:\n\t\t\t\tfileNameFirstTypes = self.fileTypes[ : ]\n\t\t\t\tfileNameFirstTypes.remove( fileType )\n\t\t\t\treturn [ fileType ] + fileNameFirstTypes + allReadables\n\t\treturn self.fileTypes + allReadables\n\n\tdef setToDisplay( self ):\n\t\t\"Do nothing because the file dialog is handling the value.\"\n\t\tpass\n\n\nclass FloatPreference( StringPreference ):\n\t\"A class to display, read & write a float.\"\n\tdef addToDialog( self, preferencesDialog ):\n\t\t\"Add this to the dialog.\"\n\t\tself.entry = Tkinter.Entry( preferencesDialog.master )\n\t\tself.entry.insert( 0, str( self.value ) )\n\t\tself.entry.grid( row = preferencesDialog.row, column = 3, sticky = Tkinter.W )\n\t\tself.label = Tkinter.Label( preferencesDialog.master, text = self.name )\n\t\tself.label.grid( row = preferencesDialog.row, column = 0, columnspan = 3, sticky = Tkinter.W )\n\t\tpreferencesDialog.row += 1\n\n\tdef setUpdateFunction( self, updateFunction ):\n\t\t\"Set the update function.\"\n\t\tself.entry.bind( '', updateFunction )\n\n\tdef setValueToString( self, valueString ):\n\t\t\"Set the float to the string.\"\n\t\ttry:\n\t\t\tself.value = float( valueString )\n\t\texcept:\n\t\t\tprint( 'Oops, can not read float' + self.name + ' ' + valueString )\n\n\nclass IntPreference( FloatPreference ):\n\t\"A class to display, read & write an int.\"\n\tdef setValueToString( self, valueString ):\n\t\t\"Set the integer to the string.\"\n\t\tdotIndex = valueString.find( '.' )\n\t\tif dotIndex > - 1:\n\t\t\tvalueString = valueString[ : dotIndex ]\n\t\ttry:\n\t\t\tself.value = int( valueString )\n\t\texcept:\n\t\t\tprint( 'Oops, can not read integer ' + self.name + ' ' + valueString )\n\n\nclass LabelDisplay:\n\t\"A class to add a label.\"\n\tdef addToDialog( self, preferencesDialog ):\n\t\t\"Add this to the dialog.\"\n\t\tself.label = Tkinter.Label( preferencesDialog.master, text = self.name )\n\t\tself.label.grid( row = preferencesDialog.row, column = 0, columnspan = 2, sticky = Tkinter.W )\n\t\tpreferencesDialog.row += 1\n\n\tdef addToPreferenceTable( self, preferenceTable ):\n\t\t\"Do nothing because the label display is not archivable.\"\n\t\tpass\n\n\tdef getFromName( self, name ):\n\t\t\"Initialize.\"\n\t\tself.name = name\n\t\treturn self\n\n\tdef getName( self ):\n\t\t\"Get name for key sorting.\"\n\t\treturn self.name\n\n\tdef setToDisplay( self ):\n\t\t\"Do nothing because the label display is not archivable.\"\n\t\tpass\n\n\tdef writeToArchiveWriter( self, archiveWriter ):\n\t\t\"Do nothing because the label display is not archivable.\"\n\t\tpass\n\n\nclass MenuButtonDisplay:\n\t\"A class to add a menu button.\"\n\tdef addToDialog( self, preferencesDialog ):\n\t\t\"Add this to the dialog.\"\n\t\tself.menuButton = Tkinter.Menubutton( preferencesDialog.master, borderwidth = 5, text = self.name, relief = Tkinter.RIDGE )\n\t\tself.menuButton.grid( row = preferencesDialog.row, column = 0, columnspan = 2, sticky = Tkinter.W )\n\t\tself.menuButton.menu = Tkinter.Menu( self.menuButton, tearoff = 0 )\n\t\tself.menuButton[ 'menu' ] = self.menuButton.menu\n\t\tpreferencesDialog.row += 1\n\n\tdef addToPreferenceTable( self, preferenceTable ):\n\t\t\"Do nothing because the label display is not archivable.\"\n\t\tpass\n\n\tdef getFromName( self, name ):\n\t\t\"Initialize.\"\n\t\tself.radioVar = None\n\t\tself.name = name\n\t\treturn self\n\n\tdef getName( self ):\n\t\t\"Get name for key sorting.\"\n\t\treturn self.name\n\n\tdef setToDisplay( self ):\n\t\t\"Do nothing because the label display is not archivable.\"\n\t\tpass\n\n\tdef writeToArchiveWriter( self, archiveWriter ):\n\t\t\"Do nothing because the label display is not archivable.\"\n\t\tpass\n\n\nclass MenuRadio( BooleanPreference ):\n\t\"A class to display, read & write a boolean with associated menu radio button.\"\n\tdef addToDialog( self, preferencesDialog ):\n\t\t\"Add this to the dialog.\"\n\t\tself.menuLength = self.menuButtonDisplay.menuButton.menu.index( Tkinter.END )\n\t\tif self.menuLength == None:\n\t\t\tself.menuLength = 0\n\t\telse:\n\t\t\tself.menuLength += 1\n\t\tself.menuButtonDisplay.menuButton.menu.add_radiobutton( label = self.name, value = self.menuLength, variable = self.getIntVar() )\n\t\tself.setDisplayState()\n\n\tdef getFromMenuButtonDisplay( self, menuButtonDisplay, name, value ):\n\t\t\"Initialize.\"\n\t\tself.getFromValue( name, value )\n\t\tself.menuButtonDisplay = menuButtonDisplay\n\t\treturn self\n\n\tdef getIntVar( self ):\n\t\t\"Get the IntVar for this radio button group.\"\n\t\tif self.menuButtonDisplay.radioVar == None:\n\t\t\tself.menuButtonDisplay.radioVar = Tkinter.IntVar()\n\t\treturn self.menuButtonDisplay.radioVar\n\n\tdef setToDisplay( self ):\n\t\t\"Set the boolean to the checkbox.\"\n\t\tself.value = ( self.getIntVar().get() == self.menuLength )\n\n\tdef setDisplayState( self ):\n\t\t\"Set the checkbox to the boolean.\"\n\t\tif self.value:\n\t\t\tself.getIntVar().set( self.menuLength )\n\t\t\tself.menuButtonDisplay.menuButton.menu.invoke( self.menuLength )\n\n\nclass ListPreference( StringPreference ):\n\tdef addToDialog( self, preferencesDialog ):\n\t\t\"Do nothing because the list preference does not have a graphical interface.\"\n\t\tpass\n\n\tdef setToDisplay( self ):\n\t\t\"Do nothing because the list preference does not have a graphical interface.\"\n\t\tpass\n\n\tdef setValueToSplitLine( self, lineIndex, lines, splitLine ):\n\t\t\"Set the value to the second and later words of a split line.\"\n\t\tself.value = splitLine[ 1 : ]\n\n\tdef setValueToString( self, valueString ):\n\t\t\"Do nothing because the list preference does not have a graphical interface.\"\n\t\tpass\n\n\tdef writeToArchiveWriter( self, archiveWriter ):\n\t\t\"Write tab separated name and list to the archive writer.\"\n\t\tarchiveWriter.write( self.name + globalSpreadsheetSeparator )\n\t\tfor item in self.value:\n\t\t\tarchiveWriter.write( item )\n\t\t\tif item != self.value[ - 1 ]:\n\t\t\t\tarchiveWriter.write( globalSpreadsheetSeparator )\n\t\tarchiveWriter.write( '\\n' )\n\n\nclass ListboxPreference( StringPreference ):\n\tdef addToDialog( self, preferencesDialog ):\n\t\t\"Add this to the dialog.\"\n#http://www.pythonware.com/library/tkinter/introduction/x5453-patterns.htm\n\t\tframe = Tkinter.Frame( preferencesDialog.master )\n\t\tscrollbar = Tkinter.Scrollbar( frame, orient = Tkinter.VERTICAL )\n\t\tself.listbox = Tkinter.Listbox( frame, selectmode = Tkinter.SINGLE, yscrollcommand = scrollbar.set )\n\t\tscrollbar.config( command = self.listbox.yview )\n\t\tscrollbar.pack( side = Tkinter.RIGHT, fill = Tkinter.Y )\n\t\tself.listbox.pack( side = Tkinter.LEFT, fill = Tkinter.BOTH, expand = 1 )\n\t\tself.setListboxItems()\n\t\tframe.grid( row = preferencesDialog.row, columnspan = 4, sticky = Tkinter.W )\n\t\tpreferencesDialog.row += 1\n\n\tdef getFromListPreference( self, listPreference, name, value ):\n\t\t\"Initialize.\"\n\t\tself.getFromValue( name, value )\n\t\tself.listPreference = listPreference\n\t\treturn self\n\n\tdef setListboxItems( self ):\n\t\t\"Set the listbox items to the list preference.\"\n\t\tfor item in self.listPreference.value:\n\t\t\tself.listbox.insert( Tkinter.END, item )\n\t\t\tif self.value == item:\n\t\t\t\tself.listbox.select_set( Tkinter.END )\n\n\tdef setToDisplay( self ):\n\t\t\"Set the selection value to the listbox selection.\"\n\t\tvalueString = self.listbox.get( Tkinter.ACTIVE )\n\t\tself.setValueToString( valueString )\n\n\nclass Radio( BooleanPreference ):\n\t\"A class to display, read & write a boolean with associated radio button.\"\n\tdef addToDialog( self, preferencesDialog ):\n\t\t\"Add this to the dialog.\"\n\t\tself.radiobutton = Tkinter.Radiobutton( preferencesDialog.master, command = self.clickRadio, text = self.name, value = preferencesDialog.row, variable = self.getIntVar() )\n\t\tself.radiobutton.grid( row = preferencesDialog.row, column = 0, columnspan = 2, sticky = Tkinter.W )\n\t\tself.setDisplayState( preferencesDialog.row )\n\t\tpreferencesDialog.row += 1\n\n\tdef clickRadio( self ):\n\t\t\"Workaround for Tkinter bug, set the value.\"\n\t\tself.getIntVar().set( self.radiobutton[ 'value' ] )\n\n\tdef getFromRadio( self, name, radio, value ):\n\t\t\"Initialize.\"\n\t\tself.getFromValue( name, value )\n\t\tself.radio = radio\n\t\treturn self\n\n\tdef getIntVar( self ):\n\t\t\"Get the IntVar for this radio button group.\"\n\t\tif len( self.radio ) == 0:\n\t\t\tself.radio.append( Tkinter.IntVar() )\n\t\treturn self.radio[ 0 ]\n\n\tdef setToDisplay( self ):\n\t\t\"Set the boolean to the checkbox.\"\n\t\tself.value = ( self.getIntVar().get() == self.radiobutton[ 'value' ] )\n\n\tdef setDisplayState( self, row ):\n\t\t\"Set the checkbox to the boolean.\"\n\t\tif self.value:\n\t\t\tself.getIntVar().set( self.radiobutton[ 'value' ] )\n\t\t\tself.radiobutton.select()\n\n\nclass RadioCapitalized( Radio ):\n\t\"A class to display, read & write a boolean with associated radio button.\"\n\tdef addToDialog( self, preferencesDialog ):\n\t\t\"Add this to the dialog.\"\n\t\twithSpaces = self.name.lower().replace( '_', ' ' )\n\t\twords = withSpaces.split( ' ' )\n\t\tcapitalizedStrings = []\n\t\tfor word in words:\n\t\t\tcapitalizedStrings.append( word.capitalize() )\n\t\tcapitalizedName = ' '.join( capitalizedStrings )\n\t\tself.radiobutton = Tkinter.Radiobutton( preferencesDialog.master, command = self.clickRadio, text = capitalizedName, value = preferencesDialog.row, variable = self.getIntVar() )\n\t\tself.radiobutton.grid( row = preferencesDialog.row, column = 0, columnspan = 2, sticky = Tkinter.W )\n\t\tself.setDisplayState( preferencesDialog.row )\n\t\tpreferencesDialog.row += 1\n\n\tdef getLowerName( self ):\n\t\t\"Get the lower case name.\"\n\t\treturn self.name.lower()\n\n\nclass WindowPosition( StringPreference ):\n\t\"A class to display, read & write a window position.\"\n\tdef addToDialog( self, preferencesDialog ):\n\t\t\"Set the master to later get the geometry.\"\n\t\tself.master = preferencesDialog.master\n\t\tself.windowPositionName = 'windowPosition' + preferencesDialog.displayPreferences.title\n\t\tself.setToDisplay()\n\n\tdef setToDisplay( self ):\n\t\t\"Set the string to the window position.\"\n\t\tif self.name != self.windowPositionName:\n\t\t\treturn\n\t\tgeometryString = self.master.geometry()\n\t\tif geometryString == '1x1+0+0':\n\t\t\treturn\n\t\tfirstPlusIndexPlusOne = geometryString.find( '+' ) + 1\n\t\tself.value = geometryString[ firstPlusIndexPlusOne : ]\n\n\tdef setWindowPosition( self ):\n\t\t\"Set the window position.\"\n\t\tgeometryString = self.master.geometry()\n\t\tif geometryString == '1x1+0+0':\n\t\t\treturn\n\t\tfirstPlusIndexPlusOne = geometryString.find( '+' ) + 1\n\t\tif self.value.count( '+' ) == 1:\n\t\t\tgeometryString = geometryString[ : firstPlusIndexPlusOne ] + self.value\n\t\t\tself.master.geometry( geometryString )\n\n\nclass PreferencesDialog:\n\tdef __init__( self, displayPreferences, master ):\n\t\t\"Add display preferences to the dialog.\"\n\t\tself.column = 0\n\t\tself.displayPreferences = displayPreferences\n\t\tself.displayToolButtonStart = True\n\t\tself.executables = []\n\t\tself.master = master\n\t\tself.row = 0\n\t\tmaster.title( displayPreferences.title )\n\t\tframe = Tkinter.Frame( master )\n\t\tfor preference in displayPreferences.archive:\n\t\t\tpreference.addToDialog( self )\n\t\tif self.row < 20:\n\t\t\tTkinter.Label( master ).grid( row = self.row )\n\t\t\tself.row += 1\n\t\tcancelColor = 'red'\n\t\tcancelTitle = 'Close'\n\t\tif displayPreferences.saveTitle != None:\n\t\t\tcancelTitle = 'Cancel'\n\t\tif displayPreferences.executeTitle != None:\n\t\t\texecuteButton = Tkinter.Button( master, activebackground = 'black', activeforeground = 'blue', text = displayPreferences.executeTitle, command = self.execute )\n\t\t\texecuteButton.grid( row = self.row, column = self.column )\n\t\t\tself.column += 1\n\t\thelpButton = Tkinter.Button( master, activebackground = 'black', activeforeground = 'white', text = \" ? \", command = self.openBrowser )\n\t\thelpButton.grid( row = self.row, column = self.column )\n\t\tself.column += 1\n\t\tcancelButton = Tkinter.Button( master, activebackground = 'black', activeforeground = cancelColor, command = master.destroy, fg = cancelColor, text = cancelTitle )\n\t\tcancelButton.grid( row = self.row, column = self.column )\n\t\tself.column += 1\n\t\tif displayPreferences.saveTitle != None:\n\t\t\tsaveButton = Tkinter.Button( master, activebackground = 'black', activeforeground = 'darkgreen', command = self.savePreferencesDestroy, fg = 'darkgreen', text = displayPreferences.saveTitle )\n\t\t\tsaveButton.grid( row = self.row, column = self.column )\n\t\tself.setWindowPosition()\n\n\tdef execute( self ):\n\t\t\"The execute button was clicked.\"\n\t\tfor executable in self.executables:\n\t\t\texecutable.execute()\n\t\tself.savePreferences()\n\t\tself.displayPreferences.execute()\n\t\tself.master.destroy()\n\n\tdef openBrowser( self ):\n\t\t\"Open the browser to the help page.\"\n\t\tnumberOfLevelsDeepInPackageHierarchy = 2\n\t\tpackageFilePath = os.path.abspath( __file__ )\n\t\tfor level in xrange( numberOfLevelsDeepInPackageHierarchy + 1 ):\n\t\t\tpackageFilePath = os.path.dirname( packageFilePath )\n\t\tdocumentationPath = os.path.join( os.path.join( packageFilePath, 'documentation' ), self.displayPreferences.fileNameHelp )\n\t\tos.system( webbrowser.get().name + ' ' + documentationPath )#used this instead of webbrowser.open() to workaround webbrowser open() bug\n\n\tdef savePreferences( self ):\n\t\t\"Set the preferences to the dialog then write them.\"\n\t\tfor preference in self.displayPreferences.archive:\n\t\t\tpreference.setToDisplay()\n\t\twritePreferences( self.displayPreferences )\n\n\tdef savePreferencesDestroy( self ):\n\t\t\"Set the preferences to the dialog, write them, then destroy the window.\"\n\t\tself.savePreferences()\n\t\tself.master.destroy()\n\n\tdef setWindowPosition( self ):\n\t\t\"Set the window position if that preference exists.\"\n\t\twindowPositionName = 'windowPosition' + self.displayPreferences.title\n\t\tfor preference in self.displayPreferences.archive:\n\t\t\tif isinstance( preference, WindowPosition ):\n\t\t\t\tif preference.name == windowPositionName:\n\t\t\t\t\tself.master.update_idletasks()\n\t\t\t\t\tpreference.setWindowPosition()\n\t\t\t\t\tself.master.update_idletasks()\n\t\t\t\t\treturn\n","sub_path":"skeinforge-0006/skeinforge_tools/skeinforge_utilities/preferences.py","file_name":"preferences.py","file_ext":"py","file_size_in_byte":28877,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"36764717","text":"from pathlib import Path\n\nfrom setuptools import (\n find_packages,\n setup,\n)\n\ntests_require = ['fixtures', 'pytest', 'pytest-mock']\n\nconfig = {\n 'name': 'sshoot',\n 'version': '1.4.2',\n 'license': 'GPLv3+',\n 'description': 'Manage multiple sshuttle VPN sessions',\n 'long_description': Path('README.rst').read_text(),\n 'author': 'Alberto Donato',\n 'author_email': 'alberto.donato@gmail.com',\n 'maintainer': 'Alberto Donato',\n 'maintainer_email': 'alberto.donato@gmail.com',\n 'url': 'https://github.com/albertodonato/sshoot',\n 'download_url': 'https://github.com/albertodonato/sshoot/releases',\n 'packages': find_packages(include=['sshoot', 'sshoot.*']),\n 'include_package_data': True,\n 'entry_points': {\n 'console_scripts': ['sshoot = sshoot.main:sshoot']\n },\n 'test_suite': 'sshoot',\n 'setup_requires': ['Babel'],\n 'install_requires': ['PyYAML', 'prettytable', 'argcomplete', 'pyxdg'],\n 'tests_require': tests_require,\n 'extras_require': {\n 'testing': tests_require\n },\n 'keywords': 'ssh sshuttle vpn',\n 'classifiers': [\n 'Development Status :: 5 - Production/Stable',\n 'Environment :: Console', 'Intended Audience :: Developers',\n 'Intended Audience :: System Administrators',\n (\n 'License :: OSI Approved :: '\n 'GNU General Public License v3 or later (GPLv3+)'),\n 'Operating System :: OS Independent', 'Programming Language :: Python',\n 'Topic :: System :: Networking', 'Topic :: Utilities'\n ]\n}\n\nsetup(**config)\n","sub_path":"pypi_install_script/sshoot-1.4.2.tar/setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":1570,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"653791159","text":"def greater_less(n):\n \n if n >= 2 and n <=5 and n % 2 == 0:\n print(\"Not Weird\")\n elif n >= 6 and n <=20 and n % 2 == 0:\n print(\"Weird\")\n elif n > 20 and n % 2 == 0:\n print(\"Not Weird\")\n else:\n print(\"Weird\")\nn = int(input())\ngreater_less(n)\n","sub_path":"Introduction/Python If-Else.py","file_name":"Python If-Else.py","file_ext":"py","file_size_in_byte":285,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"416368167","text":"#EXC_PROJECT (with dc)\r\n\r\n'''\r\n\r\nAuthors: Noemi Chignoli, Marco Iudica, Rosario Messana, Linda Ravazzano.\r\n\r\nIl presente codice simula l'attivita' di una popolazione di neuroni eccitatori disposti su un layer di forma quadrata con posizioni casuali, \r\nconnessioni stabilite con probabilita' gaussiana entro un fissato raggio per ogni neurone, un generatore di rumore poissoniano ed uno\r\ndi corrente continua indipendenti per ogni neurone.\r\nLo scopo del progetto e' stabilire se una tale simulazione puo' evidenziare l'emergere di attivita' sincrona (network burnst) all'interno\r\ndella rete e, in caso affermativo, confrontare i risultati con quelli ottenuti utilizzando sia neuroni eccitatori che inibitori (primo\r\nprogetto).\r\n\r\nNei tentativi effettuati si rileva il fenomeno dei network burst. I raster plot ottenuti indicano che gli spikes dei neuroni si concentrano\r\nquasi esclusivamente negli intervalli di burnst, a differenza del caso con entrambe le tipologie di neuroni, per il quale si osservano\r\nspikes disposti in maniera apparentemente uniforme e senza regolarita' anche al di fuori dei suddetti intervalli.\r\n\r\n'''\r\n\r\nimport nest\r\nimport numpy as np\r\nimport pylab as plt\r\nimport nest.topology as tp\r\nfrom operator import itemgetter\r\n\r\n# general parameters\r\n\r\ndt=0.025 \r\ndtSAMPL=1.\r\n\r\nnest.ResetKernel()\r\nnest.SetKernelStatus({\"resolution\":dt})\r\n\r\nTIstart = 0. # start time of dc\r\nTIend = 500. # end time of dc\r\nI0 = 6.\r\n\r\ntstop = 500.0\r\n\r\nn_neurons = 300\r\n\r\n# copying models\r\n\r\nnest.CopyModel(\"ht_neuron\", 'exc_ht_neuron')\r\nnest.CopyModel('poisson_generator','poisson_generator_prova',{\"rate\": 30.})\r\nnest.CopyModel('dc_generator','dc_generator_prova',{\"amplitude\": I0, \"start\": TIstart, \"stop\": TIend})\r\n\r\nreceptors=nest.GetDefaults('exc_ht_neuron')['receptor_types']\r\nsp=nest.GetDefaults('ht_synapse')\r\nnest.CopyModel('ht_synapse','AMPA_ht_syn',{'receptor_type':receptors['AMPA']})\r\nnest.CopyModel('static_synapse','AMPA_static_syn',{'receptor_type':receptors['AMPA']})\r\n\r\n# creating layers\r\n\r\npos=[[np.random.uniform(-1.5,1.5),np.random.uniform(-1.5,1.5)] for j in xrange(n_neurons)]\r\n\r\nl_exc=tp.CreateLayer({\"positions\":pos,\"extent\":[3.0,3.0],\"elements\": 'exc_ht_neuron'})\r\nnoise_l_exc=tp.CreateLayer({\"positions\":pos,\"extent\":[3.0,3.0],\"elements\": 'poisson_generator_prova'})\r\ndc_gen_l_exc=tp.CreateLayer({\"positions\":pos[0:n_neurons],\"extent\":[3.0,3.0], \"elements\": 'dc_generator_prova'})\r\nspikedetector_l_exc=tp.CreateLayer({\"positions\":pos,\"extent\":[3.0,3.0],\"elements\": 'spike_detector'})\r\n\r\n# network plot\r\n\r\nplt.ion()\r\ntp.PlotLayer(l_exc)\r\n\r\n# connections\r\n\r\nconndict_exc={'connection_type':'divergent','mask':{'circular':{'radius':1.}},'kernel':{'gaussian':{'p_center':1.,'sigma':0.25}},'synapse_model':'AMPA_ht_syn'}\r\ntp.ConnectLayers(l_exc,l_exc,conndict_exc)\r\n\r\ngidlist_exc_ht_neuron=nest.GetNodes(l_exc)[0]\r\ngidlist_exc_poisson_generator=nest.GetNodes(noise_l_exc)[0]\r\ngidlist_exc_dc_gen=nest.GetNodes(dc_gen_l_exc)[0]\r\ngidlist_exc_spikedetector=nest.GetNodes(spikedetector_l_exc)[0]\r\n\r\nnest.Connect(gidlist_exc_poisson_generator, gidlist_exc_ht_neuron, 'one_to_one', syn_spec={'model':'AMPA_static_syn'})\r\nnest.Connect(gidlist_exc_dc_gen, gidlist_exc_ht_neuron, 'one_to_one')\r\nnest.Connect(gidlist_exc_ht_neuron, gidlist_exc_spikedetector, 'one_to_one')\r\n\r\n# simulation\r\n\r\nnest.Simulate(tstop)\r\n\r\n# raster plot\r\n\r\nplt.figure()\r\nspikes = [ev['events']['times'] for ev in nest.GetStatus(gidlist_exc_spikedetector)] # collect spikes \r\nfor i in xrange(len(spikes)): plt.plot(spikes[i],np.repeat(i,len(spikes[i])),'ko',markersize=8)\r\n\r\n\r\n'''\r\nnest.PrintNetwork(depth=3)\r\n \r\ntp.PlotTargets([15], l_exc) \r\n\r\ntp.GetTargetPositions(pos[2],l)\r\n\r\n# stampa la quantita di connessioni \r\nfor i in xrange(150):\r\n print i,len(tp.GetTargetNodes([i], l)[0])\r\n\r\n# scrive su file le connessioni\r\ntp.DumpLayerConnections(l, 'AMPA_ht_syn', 'connections.txt')\r\n\r\nprint spikes[3]\r\n'''\r\n","sub_path":"exc_project.py","file_name":"exc_project.py","file_ext":"py","file_size_in_byte":3898,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"216964580","text":"from data_loader_util import *\nimport json\n\n\ndef model_predict(model, data_loader, device):\n data_output = list()\n for _, data_i in enumerate(data_loader, 0):\n with torch.no_grad():\n dynamic_input, static_input, labels, index_i = data_i\n dynamic_input = dynamic_input.to(device)\n static_input = static_input.to(device)\n out = model(dynamic_input, static_input)\n _, predict_i = out.max(dim=1)\n\n for j in range(len(predict_i)):\n data_outpu_ij = [index_i[0][j], index_i[1][j], int(predict_i[j])]\n data_output.append(data_outpu_ij)\n return data_output\n\n\ndef format_backtest(data):\n data_list = data\n stock_position = dict()\n ouput_dict = dict()\n\n for list in data_list:\n date = list[0]\n stock_id = list[1]\n if list[2] == 0:\n position = [0, 100]\n elif list[2] == 1:\n position = [0, -100]\n else:\n position = [0, 0]\n stock_position[stock_id] = position\n ouput_dict[date] = stock_position\n print(\"format complete\")\n return ouput_dict\n\n\ndef data_summary(data_config):\n print(\"%-20s%s\" % (\"Train start:\", data_config[\"start_date\"]))\n print(\"%-20s%s\" % (\"Train end:\", data_config[\"end_date\"]))\n print(\"%-20s%s\" % (\"Dynamic Features:\", data_config[\"ts_vars_amount\"]))\n print(\"%-20s%s\" % (\"Dynamic Length:\", data_config[\"seq_length\"]))\n print(\"%-20s%s\" % (\"Static Features:\", data_config[\"static_vars_amount\"]))\n\n\ndevice = \"cuda:9\"\nmodel_name = input(\"Model name: \")\ndata_name = input(\"Data set name: \")\nmodel_file_path = './models/%s/model.pkl' % model_name\n\ndata_config_path = './models/%s/data_config.json' % model_name\nwith open(data_config_path, 'r') as js_file:\n data_configuration = js.load(js_file)\n\ndata_summary(data_configuration)\nmodel = torch.load(model_file_path)\nmodel = model.to(device)\n\ntrain_loader, valid_loader, test_loader, data_configuration = data_loader_generator(data_name, 0.7, 64)\nprediction_data = model_predict(model, test_loader, device)\njson_file = format_backtest(prediction_data)\n\noutput_path = './models/%s/strategy_output.json' % model_name\nwith open(output_path, 'w') as f:\n json.dump(json_file, f)\n\n","sub_path":"make_prediction.py","file_name":"make_prediction.py","file_ext":"py","file_size_in_byte":2255,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"444655820","text":"'''\nreturn the greatest common divisor\n'''\n\ndef gcd(m, n):\n if m > n:\n m, n = n, m\n for i in range(m, 0, -1):\n if m % i == 0 and n % i == 0:\n return i\n\nprint(gcd(12, 16))\n","sub_path":"old/greatest_common_divisor.py","file_name":"greatest_common_divisor.py","file_ext":"py","file_size_in_byte":202,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"398519038","text":"import nets.retinanet as retinanet\nimport numpy as np\nimport keras\nfrom keras.optimizers import Adam\nfrom nets.retinanet_training import Generator\nfrom nets.retinanet_training import focal,smooth_l1 \nfrom keras.callbacks import TensorBoard, ModelCheckpoint, ReduceLROnPlateau, EarlyStopping\nfrom utils.utils import BBoxUtility\nfrom utils.anchors import get_anchors\n\nif __name__ == \"__main__\":\n NUM_CLASSES = 20\n input_shape = (600, 600, 3)\n annotation_path = '2007_train.txt'\n inputs = keras.layers.Input(shape=input_shape)\n model = retinanet.resnet_retinanet(NUM_CLASSES,inputs)\n priors = get_anchors(model)\n bbox_util = BBoxUtility(NUM_CLASSES, priors)\n\n model.load_weights(\"model_data/resnet50_coco_best_v2.1.0.h5\",by_name=True,skip_mismatch=True)\n\n # 0.1用于验证,0.9用于训练\n val_split = 0.1\n with open(annotation_path) as f:\n lines = f.readlines()\n np.random.seed(10101)\n np.random.shuffle(lines)\n np.random.seed(None)\n num_val = int(len(lines)*val_split)\n num_train = len(lines) - num_val\n\n # 训练参数设置\n logging = TensorBoard(log_dir=\"logs\")\n checkpoint = ModelCheckpoint('logs/ep{epoch:03d}-loss{loss:.3f}-val_loss{val_loss:.3f}.h5',\n monitor='val_loss', save_weights_only=True, save_best_only=False, period=1)\n reduce_lr = ReduceLROnPlateau(monitor='val_loss', factor=0.5, patience=2, verbose=1)\n early_stopping = EarlyStopping(monitor='val_loss', min_delta=0, patience=6, verbose=1)\n\n BATCH_SIZE = 2\n gen = Generator(bbox_util, BATCH_SIZE, lines[:num_train], lines[num_train:],\n (input_shape[0], input_shape[1]),NUM_CLASSES)\n\n for i in range(174):\n model.layers[i].trainable = False\n\n\n model.compile(loss={\n 'regression' : smooth_l1(),\n 'classification': focal()\n },optimizer=keras.optimizers.Adam(lr=1e-4, clipnorm=0.001)\n )\n\n model.fit_generator( gen.generate(True), \n steps_per_epoch=num_train//BATCH_SIZE,\n validation_data=gen.generate(False),\n validation_steps=num_val//BATCH_SIZE,\n epochs=10, \n verbose=1,\n initial_epoch=0,\n callbacks=[logging, checkpoint, reduce_lr, early_stopping])\n\n for i in range(174):\n model.layers[i].trainable = True\n\n model.compile(loss={\n 'regression' : smooth_l1(),\n 'classification': focal()\n },optimizer=keras.optimizers.Adam(lr=1e-5, clipnorm=0.001)\n )\n model.fit_generator( gen.generate(True), \n steps_per_epoch=num_train//BATCH_SIZE,\n validation_data=gen.generate(False),\n validation_steps=num_val//BATCH_SIZE,\n epochs=50, \n verbose=1,\n initial_epoch=10,\n callbacks=[logging, checkpoint, reduce_lr, early_stopping])\n","sub_path":"retinanet-keras-master/train.py","file_name":"train.py","file_ext":"py","file_size_in_byte":2871,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"237660957","text":"import numpy as np\r\nfrom grabscreen import grab_screen\r\nimport cv2\r\nimport time\r\nfrom directkeys import PressKey,ReleaseKey, W, A, S, D\r\nfrom alexnet_3ft import AlexNet\r\nimport torch\r\nfrom getkeys import key_check\r\n\r\nimport random\r\n\r\nWIDTH = 200\r\nHEIGHT = 150\r\nLR = 0.0003\r\nEPOCHS = 30\r\nMODEL_NAME = 'pygta5-{}-{}-{}-epochs.model'.format(LR, 'alexnet', EPOCHS)\r\n\r\nt_time = 0.09\r\n\r\n\r\ndef straight():\r\n PressKey(W)\r\n ReleaseKey(A)\r\n ReleaseKey(D)\r\n\r\ndef left():\r\n PressKey(W)\r\n PressKey(A)\r\n #ReleaseKey(W)\r\n ReleaseKey(D)\r\n #ReleaseKey(A)\r\n time.sleep(t_time)\r\n ReleaseKey(A)\r\n\r\ndef right():\r\n PressKey(W)\r\n PressKey(D)\r\n ReleaseKey(A)\r\n #ReleaseKey(W)\r\n time.sleep(t_time)\r\n ReleaseKey(D)\r\n\r\n\r\nmodel = AlexNet().cuda()\r\nmodel.load_state_dict(torch.load(MODEL_NAME))\r\nmodel.eval()\r\n\r\n\r\nfor name, param in model.named_parameters():\r\n if param.requires_grad:\r\n pass#print(name, param)\r\n\r\ndef main():\r\n last_time = time.time()\r\n for i in list(range(4))[::-1]:\r\n print(i+1)\r\n time.sleep(1)\r\n\r\n paused = False\r\n\r\n while(True):\r\n \r\n if not paused:\r\n # 800x600 windowed mode\r\n #screen = np.array(ImageGrab.grab(bbox=(0,40,800,640)))\r\n screen = grab_screen(region=(0,40,800,640))\r\n\r\n print('loop took {} seconds'.format(time.time()-last_time))\r\n\r\n last_time = time.time()\r\n\r\n screen = cv2.cvtColor(screen, cv2.COLOR_BGR2GRAY)\r\n\r\n screen = cv2.resize(screen, (WIDTH,HEIGHT))\r\n inp = np.reshape(screen, (1,1,WIDTH,HEIGHT))\r\n #print(inp.shape)\r\n #inp = [screen.reshape(WIDTH,HEIGHT,1)]\r\n prediction = model.forward(torch.cuda.FloatTensor(inp))[0]\r\n\r\n print(prediction)\r\n\r\n turn_thresh = .75\r\n\r\n fwd_thresh = 0.70\r\n\r\n if prediction[1] > fwd_thresh:\r\n straight()\r\n elif prediction[0] > turn_thresh:\r\n left()\r\n elif prediction[2] > turn_thresh:\r\n right()\r\n else:\r\n straight()\r\n\r\n keys = key_check()\r\n\r\n # p pauses game and can get annoying.\r\n\r\n if 'T' in keys:\r\n\r\n if paused:\r\n paused = False\r\n time.sleep(1)\r\n else:\r\n paused = True\r\n ReleaseKey(A)\r\n ReleaseKey(W)\r\n ReleaseKey(D)\r\n time.sleep(1)\r\n\r\nmain() \r\n","sub_path":"test_model_3ft.py","file_name":"test_model_3ft.py","file_ext":"py","file_size_in_byte":2505,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"211388986","text":"import tensorflow as tf\r\nimport numpy as np\r\n\r\ndrop_rate = 0.5\r\n\r\nclass ConvBlock(tf.keras.layers.Layer):\r\n def __init__(self, num_channels):\r\n super(ConvBlock, self).__init__()\r\n self.bn = tf.keras.layers.BatchNormalization()\r\n self.relu = tf.keras.layers.ReLU()\r\n self.conv = tf.keras.layers.Conv2D(filters=num_channels,\r\n kernel_size=(3, 3),\r\n padding='same')\r\n self.listLayers = [self.bn, self.relu, self.conv]\r\n\r\n def call(self, x):\r\n y = x\r\n for layer in self.listLayers.layers:\r\n y = layer(y)\r\n y = tf.keras.layers.concatenate([x, y], axis=-1)\r\n return y\r\n\r\nclass DenseBlock(tf.keras.layers.Layer):\r\n def __init__(self, num_convs, num_channels):\r\n super(DenseBlock, self).__init__()\r\n self.listLayers = []\r\n for i in range(num_convs):\r\n self.listLayers.append(ConvBlock(num_channels[i]))\r\n\r\n def call(self, x):\r\n for layer in self.listLayers.layers:\r\n x = layer(x)\r\n return x\r\n\r\nclass TransitionBlock(tf.keras.layers.Layer):\r\n def __init__(self, num_channels, **kwargs):\r\n super(TransitionBlock, self).__init__(**kwargs)\r\n self.batch_norm = tf.keras.layers.BatchNormalization()\r\n self.relu = tf.keras.layers.ReLU()\r\n self.conv = tf.keras.layers.Conv2D(num_channels, kernel_size=1)\r\n self.avg_pool = tf.keras.layers.AvgPool2D(pool_size=2, strides=2, padding='same')\r\n\r\n def call(self, x):\r\n x = self.batch_norm(x)\r\n x = self.relu(x)\r\n x = self.conv(x)\r\n return self.avg_pool(x)\r\n\r\ndef spatial_block_1():\r\n return tf.keras.Sequential([\r\n tf.keras.layers.BatchNormalization(),\r\n tf.keras.layers.ReLU(),\r\n tf.keras.layers.Conv2D(32, kernel_size=3, strides=1, padding='same'),\r\n #原文没说这是max or avg\r\n tf.keras.layers.MaxPool2D(pool_size=2, strides=2, padding='same')\r\n ], name='spatialNet')\r\n\r\ndef spatial_block_2():\r\n net = spatial_block_1()\r\n\r\n trans_channels = 32\r\n num_conv_channels = [32, 32, 16, 16, 32]\r\n num_convs_in_dense_block = [5, 5, 5, 5]\r\n\r\n for i, num_convs in enumerate(num_convs_in_dense_block):\r\n net.add(DenseBlock(num_convs, num_conv_channels))\r\n net.add(TransitionBlock(trans_channels))\r\n\r\n return net\r\n\r\ndef spatialBlock():\r\n net = spatial_block_2()\r\n net.add(tf.keras.layers.BatchNormalization())\r\n net.add(tf.keras.layers.ReLU())\r\n net.add(tf.keras.layers.GlobalAvgPool2D())\r\n net.add(tf.keras.layers.Flatten())\r\n net.add(tf.keras.layers.Dense(256))\r\n net.add(tf.keras.layers.Dropout(drop_rate))\r\n\r\n return net\r\n\r\ndef frequencyBlcok():\r\n #一维卷积 or 二维卷积?\r\n return tf.keras.Sequential([\r\n tf.keras.layers.BatchNormalization(),\r\n tf.keras.layers.ReLU(),\r\n tf.keras.layers.Conv1D(100, kernel_size=3),\r\n tf.keras.layers.MaxPool1D(strides=2, padding='same'),\r\n tf.keras.layers.BatchNormalization(),\r\n tf.keras.layers.ReLU(),\r\n tf.keras.layers.Conv1D(100, kernel_size=3),\r\n tf.keras.layers.MaxPool1D(strides=2, padding='same'),\r\n tf.keras.layers.Flatten(),\r\n tf.keras.layers.Dense(256),\r\n tf.keras.layers.Dropout(drop_rate)\r\n ], name='frequencyNet')\r\n\r\ndef net():\r\n spatialInput = tf.keras.Input(shape=(64, 64, 1))\r\n frequencyInput = tf.keras.Input(shape=(909, 1))\r\n\r\n spatialModel = spatialBlock()(spatialInput)\r\n frequencyModel = frequencyBlcok()(frequencyInput)\r\n\r\n outputs = tf.keras.layers.concatenate([spatialModel, frequencyModel], axis=-1)\r\n outputs = tf.keras.layers.Dense(256)(outputs)\r\n outputs = tf.keras.layers.Dropout(drop_rate)(outputs)\r\n outputs = tf.keras.layers.Dense(10)(outputs)\r\n\r\n model = tf.keras.models.Model(inputs=[spatialInput, frequencyInput], outputs=outputs)\r\n # model.add(tf.keras.layers.Dense(256))\r\n # model.add(tf.keras.layers.Dropout(drop_rate))\r\n # model.add(tf.keras.layers.Dense(10))\r\n\r\n return model\r\n\r\nmodel = net()\r\nmodel.summary()\r\n\r\n\r\n# fashion_mnist = tf.keras.datasets.fashion_mnist\r\n# (train_images, train_labels), (test_images, test_labels) = fashion_mnist.load_data()\r\n# # train_images = tf.expand_dims(train_images / 255.0, -1)\r\n# # test_images = tf.expand_dims(test_images / 255.0, -1)\r\n# # train_labels = tf.convert_to_tensor(train_labels)\r\n# # test_labels = tf.convert_to_tensor(test_labels)\r\n# train_images = np.expand_dims(train_images / 255.0, 3)\r\n# test_images = np.expand_dims(test_images / 255.0, 3)\r\n# train_images = tf.constant(train_images, dtype=tf.float32)\r\n# test_images = tf.constant(test_images, dtype=tf.float32)\r\n# train_labels = tf.constant(train_labels, dtype=tf.float32)\r\n# test_labels = tf.constant(test_labels, dtype=tf.float32)\r\n#\r\n# model = net()\r\n# model.compile(optimizer='adam',\r\n# loss=tf.keras.losses.SparseCategoricalCrossentropy(from_logits=True),\r\n# metrics=['accuracy'])\r\n# model.fit(train_images, train_labels, epochs=1)\r\n# test_loss, test_acc = model.evaluate(test_images, test_labels)\r\n# print('\\nTest accuracy:', test_acc)\r\n\r\n","sub_path":"JPGDenseNet.py","file_name":"JPGDenseNet.py","file_ext":"py","file_size_in_byte":5207,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"191375240","text":"#! /usr/bin/python\n# -*- coding: utf-8 -*-\n\nimport re\nimport csv\nimport os\n\nclass VVBParser:\n\t\"\"\"\n\tClass for parsing unstructured text with messages of Skype chat format \n\t\"\"\"\n\t\n\tdef __init__(self):\n\t\tself.records = []\n\t\terr_lines = []\n\n\tdef __parse_line(self, line):\n\t\tm = re.match(u\"^\\W(\\d+).(\\d+).(\\d+)\\W\\s(\\d+:\\d+:\\d+)\\W\\s([\\S\\s]+?):\\s([\\s\\S]+)\", line)\n\t\tif m:\n\t\t\treturn (\"20\" + m.group(3) + \"-\" + m.group(2) + \"-\" + m.group(1) + \" \" + m.group(4),\n\t\t\t\t\tm.group(5), m.group(6))\n\n\tdef __save_line(self, groups):\n\t\tif groups:\n\t\t\tself.records.append((groups[0], groups[1], groups[2]))\n\n\tdef parse(self, file_name):\n\t\tFILE_IN = open(file_name, 'r')\n\t\tdata = FILE_IN.read()\n\t\tFILE_IN.close()\n\t\t# normalize text structure\n\t\tdata = data.replace('\\n', ' ').replace('[', '\\n[').replace('> \\n', '> ').replace(': \\n[', ': [')\n\t\tlines = data.split('\\n')\n\t\tfor line in lines:\n\t\t\tgroups = self.__parse_line(line)\n\t\t\tself.__save_line(groups)\n\t\treturn self.records\n\n\tdef save_to_csv(self, lines, file_name):\n\t\tf = open(file_name, 'wb')\n\t\twr = csv.writer(f, quoting=csv.QUOTE_ALL)\n\t\twr.writerow(('RECTIME', 'NICK', 'MESSAGE'))\n\t\tfor line in lines:\n\t\t\twr.writerow(line)\n\t\tf.close()\n\n# Main workflow #\nparser = VVBParser()\nfor file in os.listdir(\"./in\"):\n\tfile_name = os.path.splitext(file)[0]\n\tlines = parser.parse('./in/' + file)\n\tparser.save_to_csv(lines, './out/' + file_name + '.csv')\n","sub_path":"parse_skype_hist.py","file_name":"parse_skype_hist.py","file_ext":"py","file_size_in_byte":1375,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"43842447","text":"\"\"\"\nThis file implements linked structures including:\n -singly-linked structures\n -doubly-linked structures\n -circularly-linked structures\n\"\"\"\n\nfrom __future__ import division, print_function\n\nclass SinglyLinkedNode(object):\n \"\"\"Represents a node in a singly-linked structure.\"\"\"\n\n def __init__(self, data, next=None):\n \"\"\"Instantiates a node with a default next of None.\"\"\"\n self.data = data\n self.next = next\n\n def __iter__(self):\n pointer = self\n while pointer.data:\n yield pointer.data\n pointer = pointer.next\n\nclass DoublyLinkedNode(object):\n \"\"\"Represents a node in a doubly-linked structure.\"\"\"\n\n def __init__(self, data, previous=None, next=None):\n \"\"\"Instantiates a node.\"\"\"\n self.data = data\n self.previous = previous\n self.next = next\n","sub_path":"node.py","file_name":"node.py","file_ext":"py","file_size_in_byte":855,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"30859957","text":"import pickle#leitourgei mono se python 2.7\r\nfriend= 'friend.txt' \r\nans = \"y\"\r\nfriends={}\r\nfile = open('friends.txt', \"a\")#ανοιγει και γραφει στο αρχειο\r\nwhile ans == \"y\":\r\n name = raw_input(\"dose onoma : \")\r\n ergasia = raw_input(\"dose ergasia: \")\r\n email = raw_input(\"Δωσε email : \")\r\n phone = raw_input(\"Δωσε thlefono : \")\r\n\r\n friends[name] = {(\"Name\"): name, (\"Ergasia\"): ergasia,(\"Email\"): email, (\"Phone\"): phone}\r\n\r\n ans = raw_input(\"Θελεις να προσθεσεις αλλη εγγραφη (y/n) ? :\")\r\npickle.dump(friends, file)\r\nfile.close()\r\n\r\nimport pickle\r\n\r\nfriend = {}\r\nwith open('friends.txt') as f: #ανοιγει για να εκτυπωσει\r\n while 1: \r\n try:\r\n friend.update(pickle.load(f))\r\n except EOFError:\r\n break # no more data in the file\r\n\r\nfor person in friend.values():\r\n print (('{Name}\\t{Ergasia}\\t{Email}\\t{Phone}').format(**person))\r\n","sub_path":"python27/friends_mono_se_2_7/friends.py","file_name":"friends.py","file_ext":"py","file_size_in_byte":982,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"434740931","text":"CHEQUE_WEIGHT = 45\nEMI_WEIGHT = 30\nABB_WEIGHT = 25\n\ndef getChequeScore(inbound_cheques, total_cheques):\n\ttry:\n\t\tcheque_ratio = round(inbound_cheques * 100 / total_cheques, 0)\n\n\t\tinbound_cheques_break0 = [21, 11, 6, 2, 0]\t# >= 200\n\t\tinbound_cheques_score_break0 = [-400, -200, -100, 25, 100]\n\n\t\tinbound_cheques_break1 = [21, 11, 6, 2, 0]\t# >= 100\n\t\tinbound_cheques_score_break1 = [-300, -100, -50, 25, 100]\n\n\t\tinbound_cheques_break2 = [21, 11, 6, 2, 0]\t# >= 50\n\t\tinbound_cheques_score_break2 = [-200, -75, -25, 25, 100]\n\n\t\tinbound_cheques_break3 = [8, 5, 2, 0]\t# >= 25\n\t\tinbound_cheques_score_break3 = [-200, -50, 0, 100]\n\n\t\tinbound_cheques_break4 = [6, 4, 2, 0]\t# <25\n\t\tinbound_cheques_score_break4 = [-200, -50, 0, 50]\n\t\tcheque_score = None\n\n\t\tif total_cheques >= 200:\n\t\t\tfor k, v in zip(inbound_cheques_break0, inbound_cheques_score_break0):\n\t\t\t\tif cheque_ratio >= k:\n\t\t\t\t\tcheque_score = v\n\t\t\t\t\tbreak\n\n\t\telif total_cheques >= 100:\n\t\t\tfor k, v in zip(inbound_cheques_break1, inbound_cheques_score_break1):\n\t\t\t\tif cheque_ratio >= k:\n\t\t\t\t\tcheque_score = v\n\t\t\t\t\tbreak\n\n\t\telif total_cheques >= 50:\n\t\t\tfor k, v in zip(inbound_cheques_break2, inbound_cheques_score_break2):\n\t\t\t\tif cheque_ratio >= k:\n\t\t\t\t\tcheque_score = v\n\t\t\t\t\tbreak\n\n\t\telif total_cheques >= 25:\n\t\t\tfor k, v in zip(inbound_cheques_break3, inbound_cheques_score_break3):\n\t\t\t\tif inbound_cheques >= k:\n\t\t\t\t\tcheque_score = v\n\t\t\t\t\tbreak\n\n\t\telse:\n\t\t\tfor k, v in zip(inbound_cheques_break4, inbound_cheques_score_break4):\n\t\t\t\tif inbound_cheques >= k:\n\t\t\t\t\tcheque_score = v\n\t\t\t\t\tbreak\n\n\t\treturn cheque_score\n\texcept:\n\t\treturn None\n\ndef getEmiScore(emi_bounces):\n\ttry:\n\t\temi_break = [4, 3, 2, 0]\n\t\temi_score_break = [-300, -200, -50, 0]\n\n\t\temi_score = None\n\n\t\tfor k, v in zip(emi_break, emi_score_break):\n\t\t\tif emi_bounces >= k:\n\t\t\t\temi_score = v\n\t\t\t\tbreak\n\n\t\treturn emi_score\n\texcept:\n\t\treturn None\n\ndef getAbbScore(abb_balance):\n\ttry:\n\t\tabb_break = [100000, 50000, 10000, 5000, 0]\n\t\tabb_score_break = [50, 100, 50, 10, 0]\n\n\t\tabb_score = None\n\n\t\tfor k, v in zip(abb_break, abb_score_break):\n\t\t\tif abb_balance >= k:\n\t\t\t\tabb_score = v\n\t\t\t\tbreak\n\n\t\treturn abb_score\n\texcept:\n\t\treturn None\n\ndef getTotalScore(inbound_cheques, total_cheques, emi_bounces, abb_balance):\n\tcheque_score = getChequeScore(inbound_cheques, total_cheques)\n\temi_score = getEmiScore(emi_bounces)\n\tabb_score = getAbbScore(abb_balance)\n\n\tif cheque_score == None or emi_score == None or abb_score == None:\n\t\treturn {\n\t\t\t'chequeScore': None,\n\t\t\t'emiScore': None,\n\t\t\t'abbScore': None,\n\t\t\t'totalScore': None\n\t\t}\n\n\tcheque_score = round(cheque_score * CHEQUE_WEIGHT/100.0, 1)\n\temi_score = round(emi_score * EMI_WEIGHT/100.0, 1)\n\tabb_score = round(abb_score * ABB_WEIGHT/100.0, 1)\t\n\ttotal_score = cheque_score + emi_score + abb_score\n\n\treturn {\n\t\t'chequeScore': cheque_score,\n\t\t'emiScore': emi_score,\n\t\t'abbScore': abb_score,\n\t\t'totalScore': total_score\n\t}","sub_path":"scoring.py","file_name":"scoring.py","file_ext":"py","file_size_in_byte":2869,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"295900737","text":"\"\"\"\n@brief test tree node (time=2s)\n\"\"\"\nimport os\nimport unittest\nfrom pyquickhelper.pycode import get_temp_folder, ExtTestCase\nfrom pyquickhelper.loghelper import BufferedPrint\nfrom mathenjeu.__main__ import main\n\n\nclass TestQcmHttpsAppCli(ExtTestCase):\n\n def test_https_webapp(self):\n st = BufferedPrint()\n main(args=['qcm_https', '--help'], fLOG=st.fprint)\n res = str(st)\n self.assertIn(\"usage: qcm_https\", res)\n\n def test_https_webapp_start(self):\n temp = get_temp_folder(__file__, \"temp_local_https\")\n st = BufferedPrint()\n key_file = os.path.join(temp, \"key.pem\")\n cert_file = os.path.join(temp, \"cert.pem\")\n main(args=['create_self_signed_cert', '--keyfile=' +\n key_file, '--certfile=' + cert_file], fLOG=st.fprint)\n main(args=['qcm_https', '--cookie_key=dummypwd', '--port=8889',\n '--userpwd=abc', '--ca_certs=\"{0}\"'.format(temp),\n '--keyfile=key.pem', '--certfile=cert.pem',\n '--folder=' + temp],\n fLOG=st.fprint)\n res = str(st)\n self.assertIn('[create_self_signed_cert]', res)\n self.assertIn('[create_qcm_https_app] saved file', res)\n self.assertIn('apphyper.py', res)\n\n\nif __name__ == \"__main__\":\n unittest.main()\n","sub_path":"_unittests/ut_cli/test_static_https_app.py","file_name":"test_static_https_app.py","file_ext":"py","file_size_in_byte":1321,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"614152355","text":"from random import randint\r\nmax=1000\r\nmin=1\r\ncount=0\r\nwhile True:\r\n guess=randint(min,max)\r\n print('Računalo pogađa: {}.'.format(guess))\r\n ans=input('Je li pogodak točan?' \r\n '(Unesite p ako je točno, v ako je traženi broj veći ili m ako je traženi broj manji.):')\r\n if ans=='p':\r\n break\r\n elif ans=='v':\r\n min=guess+1\r\n elif ans=='m':\r\n max=guess-1\r\n count += 1\r\nprint('Kraj!')","sub_path":"Zadatci/2. serija/14_kadogop.py","file_name":"14_kadogop.py","file_ext":"py","file_size_in_byte":440,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"427577162","text":"# -*- coding: utf-8 -*-\n#NÃO APAGUE A LINHA ACIMA. COMECE ABAIXO DESTA LINHA\nn=int(input('Digite um número:'))\na=2\nb=1\ni=0\ncont=1\nwhile i<=(n-1): \n cont=cont*(a/b)\n \n if i%2==1:\n a=a+2\n else:\n b=b+2\n i=i+1\n \ncont=cont*2\nprint(cont)\n ","sub_path":"moodledata/vpl_data/131/usersdata/216/45681/submittedfiles/al10.py","file_name":"al10.py","file_ext":"py","file_size_in_byte":289,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"582667703","text":"from os.path import join, dirname, abspath\nimport sys\nsys.path.append(dirname(dirname(abspath(__file__))))\ntry:\n import jackal_navi_envs\nexcept:\n pass\n\nimport gym\nimport numpy as np\ntry:\n sys.path.remove('/opt/ros/melodic/lib/python2.7/dist-packages')\nexcept:\n pass\nimport torch\nfrom torch import nn\nfrom torch.utils.tensorboard import SummaryWriter\n\nfrom tianshou.env import SubprocVectorEnv, DummyVectorEnv\nfrom policy import DQNPolicy, DuelingDQN\nfrom tianshou.data import Collector, ReplayBuffer, PrioritizedReplayBuffer\nfrom collector import Collector as Fake_Collector\nfrom offpolicy import offpolicy_trainer\n\nsys.path.append('/opt/ros/melodic/lib/python2.7/dist-packages')\n\nimport pickle\nimport argparse\nimport json\nfrom datetime import datetime\nimport os\n\nparser = argparse.ArgumentParser(description = 'Jackal navigation simulation')\nparser.add_argument('--config', dest = 'config_path', type = str, default = 'configs/dqn.json', help = 'path to the configuration file')\nparser.add_argument('--save', dest = 'save_path', type = str, default = 'results/', help = 'path to the saving folder')\n\nargs = parser.parse_args()\nconfig_path = args.config_path\nsave_path = args.save_path\n\n# Load the config files\nwith open(config_path, 'rb') as f:\n config = json.load(f)\n\nenv_config = config['env_config']\nwrapper_config = config['wrapper_config']\ntraining_config = config['training_config']\n\n# Config logging\nnow = datetime.now()\ndt_string = now.strftime(\"%Y_%m_%d_%H_%M\")\nsave_path = os.path.join(save_path, config['section'] + \"_\" + dt_string)\nif not os.path.exists(save_path):\n os.mkdir(save_path)\nwriter = SummaryWriter(save_path)\nwith open(os.path.join(save_path, 'config.json'), 'w') as fp:\n json.dump(config, fp)\n\n# initialize the env --> num_env can only be one right now\n\nif not config['use_container']:\n wrapper_dict = jackal_navi_envs.jackal_env_wrapper.wrapper_dict\n env = wrapper_dict[wrapper_config['wrapper']](gym.make('jackal_discrete-v0', **env_config), **wrapper_config['wrapper_args'])\n train_envs = DummyVectorEnv([lambda: env for _ in range(1)])\n state_shape = env.observation_space.shape or env.observation_space.n\n action_shape = env.action_space.shape or env.action_space.n\nelse:\n train_envs = config\n Collector = Fake_Collector\n state_shape = 721+len(config['env_config']['param_list']) if config['env'] == 'jackal' else 4\n action_shape = 2**len(config['env_config']['param_list'])+1 if config['env'] == 'jackal' else 2\n print(state_shape, action_shape, config['env_config']['param_list'], len(config['env_config']['param_list']), len(config['env_config']['param_list'])**2)\n\n# config random seed\nnp.random.seed(config['seed'])\ntorch.manual_seed(config['seed'])\nif not config['use_container']:\n train_envs.seed(config['seed'])\n'''\nnet = Net(training_config['layer_num'], state_shape, action_shape, config['device']).to(config['device'])\noptim = torch.optim.Adam(net.parameters(), lr=training_config['learning_rate'])\n'''\n'''\nclass DuelingDQN(nn.Module):\n def __init__(self, state_shape, action_shape, hidden_layer = [64, 64], cnn = True, feature_layer = [256]):\n super().__init__()\n if cnn:\n self.feature = nn.Sequential(\n nn.Linear(720, feature_layer[0]), nn.ReLU(inplace=True)\n )\n feature_shape = feature_layer[0] + int(np.log2(action_shape-1))+1\n else:\n self.feature = lambda x: x.view(x.shape[0], -1)\n feature_shape = state_shape\n\n layers = [np.prod(feature_shape)] + hidden_layer\n self.value = []\n self.advantage = []\n for i, o in zip(layers[:-1], layers[1:]):\n self.value.append(nn.Linear(i, o))\n self.value.append(nn.ReLU(inplace=True))\n self.advantage.append(nn.Linear(i, o))\n self.advantage.append(nn.ReLU(inplace=True))\n self.advantage.append(nn.Linear(o, np.prod(action_shape)))\n self.value.append(nn.Linear(o, 1))\n\n self.value = nn.Sequential(*self.value)\n self.advantage = nn.Sequential(*self.advantage)\n\n def forward(self, obs, state=None, info={}):\n if not isinstance(obs, torch.Tensor):\n obs = torch.tensor(obs, dtype=torch.float)\n batch = obs.shape[0]\n laser = obs.view(batch, 1, -1)[:,:,:720]\n params = obs.view(batch, -1)[:, 720:]\n\n embedding = self.feature(laser).view(batch, -1)\n feature = torch.cat((embedding, params), dim = 1)\n\n advantage = self.advantage(feature)\n value = self.value(feature)\n logits = value + advantage - advantage.mean(1, keepdim=True)\n return logits, state\n'''\nnet = DuelingDQN(state_shape, action_shape, hidden_layer = training_config['hidden_layer'], cnn = training_config['cnn'])\noptim = torch.optim.Adam(net.parameters(), lr=training_config['learning_rate'])\n\npolicy = DQNPolicy(\n net, optim, training_config['gamma'], training_config['n_step'],\n grad_norm_clipping = training_config['grad_norm_clipping'],\n target_update_freq=training_config['target_update_freq'])\n\nif training_config['prioritized_replay']:\n buf = PrioritizedReplayBuffer(\n training_config['buffer_size'],\n alpha=training_config['alpha'], beta=training_config['beta'])\nelse:\n buf = ReplayBuffer(training_config['buffer_size'])\npolicy.set_eps(1)\ntrain_collector = Collector(policy, train_envs, buf)\ntrain_collector.collect(n_step=training_config['pre_collect'])\n\ndef delect_log():\n for dirname, dirnames, filenames in os.walk('/u/zifan/.ros/log'):\n for filename in filenames:\n p = join(dirname, filename)\n if p.endswith('.log') and dirname != '/u/zifan/.ros/log':\n os.remove(p)\n\ntrain_fn =lambda e: [policy.set_eps(max(0.1, 1-(e-1)/training_config['epoch']/training_config['exploration_ratio'])),\n torch.save(policy.state_dict(), os.path.join(save_path, 'policy_%d.pth' %(e)))]\n\nresult = offpolicy_trainer(\n policy, train_collector, training_config['epoch'],\n training_config['step_per_epoch'], training_config['collect_per_step'],\n training_config['batch_size'], update_per_step=training_config['update_per_step'],\n train_fn=train_fn, writer=writer)\n\nimport shutil\nshutil.rmtree('/u/zifan/buffer', ignore_errors=True) # a way to force all the actor stops\n\ntrain_envs.close()\n","sub_path":"discrete/dqn.py","file_name":"dqn.py","file_ext":"py","file_size_in_byte":6395,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"50807421","text":"import cv2\nimport numpy as np\n\nfrom merfishdecoder.core import zplane\n\n\ndef log_readout_images(obj: zplane.Zplane = None, \n frameNames: list = None\n ) -> zplane.Zplane:\n\n \"\"\"\n\n Correct the intensity difference between color channels\n using existed scale factor profiles\n\n \"\"\"\n \n frameNames = obj.get_readout_name() \\\n if frameNames is None else frameNames\n for fn in frameNames:\n obj._frames[fn]._img = np.log10(obj._frames[fn]._img.astype(np.float16) + 1)\n return obj\n\ndef scale_readout_images(obj: zplane.Zplane = None, \n frameNames: list = None,\n scaleFactors: dict = None,\n ) -> zplane.Zplane:\n\n \"\"\"\n\n Correct the intensity difference between color channels\n using existed scale factor profiles\n\n \"\"\"\n \n frameNames = obj.get_readout_name() \\\n if frameNames is None else frameNames\n if scaleFactors is None:\n scaleFactors = estimate_scale_factors(\n obj, frameNames)\n for fn in frameNames:\n obj._frames[fn]._img = obj._frames[fn]._img.astype(np.float16) / scaleFactors[fn]\n return obj\n\ndef estimate_scale_factors(obj: zplane.Zplane = None, \n frameNames: list = None\n ) -> dict:\n\n \"\"\"\n Estimate scale factors between rounds of images.\n \"\"\"\n\n frameNames = obj.get_readout_name() \\\n if frameNames is None else frameNames\n\n return dict(zip(frameNames,\n [ np.median(x[x > 0]) for x in \n obj.get_readout_images(frameNames) ]))\n\n\"\"\"Position-independent normalization\n\"\"\"\n\nfrom typing import Tuple, List\nfrom scipy import stats\nimport numpy as np\n\n\nclass pin(object):\n def __init__(\n self,\n mu_1=1,\n sigma2_1=1,\n mu_blk=0,\n sigma2_blk=1,\n pie=0.1,\n tpr=0.9,\n tnr=0.8,\n p=0.8,\n mu_0=0.3,\n sigma2_0=1,\n ):\n \"\"\"Init position-independent normalization method.\n\n - mu_1, sigma2_1: mean and variance for\n the log10-level light intensity when there is signal\n - mu_blk, sigma2_blk: mean and variance for\n the log10-level light intensity for the blank pixels\n when they show signals by chance.\n - pie: prior P(X = 1)\n - tpr: true positive rate\n - tnr: true negative rate\n - p: zero weight in hurdle model\n - mu_0, sigma2_0: truncated normal mean and variance in hurdle model\n \"\"\"\n\n self.mu_1: float = mu_1\n self.sigma2_1: float = sigma2_1\n self.norm_1 = stats.norm(loc=self.mu_1, scale=np.sqrt(self.sigma2_1))\n\n self.mu_blk: float = mu_blk\n self.sigma2_blk: float = sigma2_blk\n self.norm_blk = stats.norm(loc=self.mu_blk, scale=np.sqrt(self.sigma2_blk))\n\n self.pie: float = pie\n self.tpr: float = tpr\n self.tnr: float = tnr\n\n ## hurdle model: non-zero part uses truncated normal\n self.p: float = p\n self.mu_0: float = mu_0\n self.sigma2_0: float = sigma2_0\n ## set a as 1e-15 to make sure the trunnorm.pdf(0) equals to 0\n ## then make a transformation since in stats.truncnorm, a and b are relatively\n ## towards a standard norm\n a = (1e-15 - self.mu_0) / np.sqrt(self.sigma2_0)\n # b = stats.norm.ppf(0.999, loc=self.mu_0, scale=np.sqrt(self.sigma2_0))\n # b = (b - self.mu_0) / np.sqrt(self.sigma2_0)\n b = np.inf\n self.truncnorm = stats.truncnorm(\n a=a,\n b=b,\n loc=self.mu_0,\n scale=np.sqrt(self.sigma2_0),\n )\n\n def _prob(self, y_ij: float) -> float:\n \"\"\"Calculate P(X_ij | y_ij)\n\n Return:\n - P(X_ij = 1 | y_ij)\n \"\"\"\n p_yij_1 = self.norm_1.pdf(y_ij)\n p_yij_blank = self.norm_blk.pdf(y_ij)\n\n w_xij_1: float = self.pie * (self.tpr * p_yij_1 + (1 - self.tpr) * p_yij_blank)\n w_xij_0: float = (1 - self.pie) * (\n (1 - self.tnr) * p_yij_1 + self.tnr * p_yij_blank\n )\n p = w_xij_1 / (w_xij_1 + w_xij_0)\n return p\n\n def prob1_vec(self, y_t: np.ndarray) -> np.ndarray:\n \"\"\"Given a matrix (figure at t-th turn), get the P(X^t = 1 | Y^t).\n This function is to accelarate the process through vectorization.\n \"\"\"\n ## P(x = 1) * P(y|x)\n p_y_1: np.ndarray = self.norm_1.pdf(y_t)\n p_y_blk: np.ndarray = self.norm_blk.pdf(y_t)\n w_x_1: np.ndarray = self.pie * (self.tpr * p_y_1 + (1 - self.tpr) * p_y_blk)\n\n ## P(x = 0) * P(y|x)\n p_y_0: np.ndarray = self.truncnorm.pdf(y_t)\n h_y_0: np.ndarray = (1 - self.p) * p_y_0\n h_y_0[y_t == 0] = self.p\n w_x_0: np.ndarray = (1 - self.pie) * (\n self.tnr * h_y_0 + (1 - self.tnr) * p_y_blk\n )\n ## point-wise divide\n r = np.divide(w_x_1, w_x_0 + w_x_1)\n return r\n\n def prob1(self, y_t: np.ndarray) -> np.ndarray:\n \"\"\"Given a matrix (figure at t-th turn), get the P(X^t = 1 | Y^t).\"\"\"\n ## vectorize too slow\n # r = np.vectorize(self._prob)(y_t)\n r = np.asarray(np.frompyfunc(self._prob, 1, 1)(y_t), dtype=np.float)\n return r\n","sub_path":"merfishdecoder/util/preprocessing.py","file_name":"preprocessing.py","file_ext":"py","file_size_in_byte":5277,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"555635773","text":"from django.http import Http404, HttpResponse\nfrom django.shortcuts import render, redirect, get_object_or_404\nfrom django.core import serializers\nfrom django.core.exceptions import ObjectDoesNotExist\nfrom data.models import Biro, BPH\nfrom .models import Oprec, Candidate\n\ndef oprec_home (request):\n candidate = serializers.serialize (\"json\", Candidate.objects.all ())\n return render (request, 'oprec-index.html', {'candidate':candidate})\n\ndef new_submission (request, candidate):\n biro = Biro.objects.all ().order_by ('biro_id')\n bph = BPH.objects.all ()\n return render (request, 'oprec-form.html', {'candidate': candidate, 'biro':biro, 'bph':bph})\n\ndef oprec_form (request, nrp):\n try:\n candidate = Candidate.objects.get (NRP = nrp)\n\n try:\n submission = Oprec.objects.get (NRP = nrp)\n\n if submission.completed == True:\n return redirect (\"/tebarjaring/\")\n else:\n return new_submission (request, candidate)\n\n except ObjectDoesNotExist:\n return new_submission (request, candidate)\n\n except ObjectDoesNotExist:\n return redirect (\"/tebarjaring/\")\n\ndef oprec_submit (request, nrp):\n if request.method == \"POST\":\n name = Candidate.objects.get (NRP = nrp).name\n email = request.POST.get ('email', '')\n phone = request.POST.get ('phone', '')\n id_line = request.POST.get ('id_line', '')\n role = request.POST.get ('role', '')\n\n shirt_size = request.POST.get ('shirt_size', '')\n sleeve_length = request.POST.get ('sleeve_length', '')\n motivasi = request.POST.get ('motivasi', '')\n inovasi = request.POST.get ('inovasi', '')\n\n if role == \"bph\":\n choice_1 = request.POST.get ('choice_1b')\n choice_2 = request.POST.get ('choice_2b')\n choice_3 = request.POST.get ('choice_3b')\n choice_4 = \"\"\n choice_5 = \"\"\n alasan_1 = request.POST.get ('alasan_1b')\n alasan_2 = request.POST.get ('alasan_2b')\n alasan_3 = request.POST.get ('alasan_3b')\n else:\n choice_1 = request.POST.get ('choice_1s')\n choice_2 = request.POST.get ('choice_2s')\n choice_3 = request.POST.get ('choice_3s')\n choice_4 = request.POST.get ('choice_4s')\n choice_5 = request.POST.get ('choice_5s')\n alasan_1 = request.POST.get ('alasan_1s')\n alasan_2 = request.POST.get ('alasan_2s')\n alasan_3 = request.POST.get ('alasan_3s')\n\n submission = Oprec (\n NRP = nrp,\n name = name,\n email = email,\n phone = phone,\n id_line = id_line,\n role = role,\n choice_1 = choice_1,\n choice_2 = choice_2,\n choice_3 = choice_3,\n choice_4 = choice_4,\n choice_5 = choice_5,\n alasan_1 = alasan_1,\n alasan_2 = alasan_2,\n alasan_3 = alasan_3,\n motivasi = motivasi,\n inovasi = inovasi,\n shirt_size = shirt_size,\n sleeve_length = sleeve_length\n )\n\n submission.save ()\n return render (request, 'oprec-confirm.html', {'submission': submission})\n else:\n return redirect (\"/tebarjaring/\")\n\ndef oprec_complete (request, nrp):\n if request.method == \"POST\":\n try:\n submission = Oprec.objects.get (NRP = nrp)\n\n if \"5116100\" in nrp:\n if submission.completed == False:\n submission.completed = True\n submission.save ()\n candidate = Candidate.objects.get (NRP = nrp)\n candidate.delete ()\n \n return render (request, 'oprec-done.html');\n\n except ObjectDoesNotExist:\n return redirect (\"/tebarjaring/\")\n else:\n return redirect (\"/tebarjaring/\")\n\ndef coming_soon (request):\n return render (request, 'oprec-coming-soon.html')\n\ndef oprec_closed (request):\n return redirect ('/tebarjaring/')\n\ndef oprec_closed_home (request):\n return render (request, 'oprec-closed.html')","sub_path":"recruitment/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":4305,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"150624882","text":"\"\"\"\nauthor: Nora V\ndate: April 12, 2019\n\"\"\"\nimport os\nimport numpy as np\nimport pandas as pd\nimport anndata as ad\nimport scanpy as sc\nimport re\nimport fnmatch\n#create a dataframe that I can use to concatenate all other csv files into anndata\ndata_path = '/Users/bendalllab/Desktop/UH_PB_PL-F_LM'\ncsv_filename = 'big_frame.csv'\ncsv_filepath = os.path.join(data_path, csv_filename)\nsample_pd = pd.read_csv(csv_filepath)\n# Dropping old index column\nsample_pd.drop(columns =[\"Unnamed: 0\"], inplace = True)\n#make anndata out of pandas dataframe\nn_obs = len(sample_pd.index)\nobs = sample_pd.loc[:, 'Cage':'Sample_type']\nunwanted_var_names = ['Cage', 'ID', 'Embryonic_day', 'Sample_type']\nall_var_names = sample_pd.columns.tolist()\nvar_names = [word for word in all_var_names if word not in unwanted_var_names]\nn_vars = len(var_names)\n# dataframe for annotating the variables\nvar = pd.DataFrame(index=var_names)\n# the data matrix of shape n_obs x n_vars\nX = np.arange(n_obs*n_vars).reshape(n_obs, n_vars)\nadata = ad.AnnData(X, obs=obs, var=var)\n########################## sub select embryonic day ###########################\n#select one embryonic day to train umap with all organs\nadata_training = adata[adata.obs['Embryonic_day'] == \"E10.5\"]\n# Computing the neighborhood graph\nsc.pp.neighbors(adata_training, n_neighbors=10)\nsc.tl.umap(adata_training)\nsc.pl.umap(adata_training, color=['CD19', 'CD3', 'Ly-6G'])\n# plot (pl) those proteins (called genes in scanpy) that yield the highest fraction of counts in each single cells, across all cells.\n# sc.pl.highest_expr_genes(adata_subset, n_top=10)\n# Computing the neighborhood graph\nsc.pp.neighbors(adata, n_neighbors=10)\n#sc.tl.tsne(adata)\n#sc.pl.tsne(adata, color=['CD19', 'CD3', 'Ly-6G'])\nsc.tl.umap(adata)\nsc.pl.umap(adata, color=['CD19', 'CD3', 'Ly-6G'])\nsc.tl.louvain(adata)\n#sc.pl.tsne(adata, color=['louvain', 'CD19', 'CD3'])\nsc.pl.umap(adata, color=['louvain', 'CD19', 'CD3'])\n","sub_path":"20190412 scanpy on LN and PB.py","file_name":"20190412 scanpy on LN and PB.py","file_ext":"py","file_size_in_byte":1928,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"57617215","text":"## wallpaper-watch.py\n## Script to batch process multiple images for use as an Apple Watch wallpaper\n## Mainly this involves blending the edges to black to match the bezel naturally\n\n# Import any required libraries\nimport os\nimport glob\nimport numpy as np\nfrom scipy import misc\nfrom PIL import Image\n\n# Define the watch size\nwatch_size=38\n\nif watch_size == 38:\n\tres = array([340.,272.])\nelse:\n\tres = array([390.,312.])\n\n# Define an aspect ratio\napct = res[0] / res[1]\n\n# Define a mask with gaussian edges\nc = 0.07\nw = 0.15\nxarr = arange(res[1])/res[1]\nyarr = arange(res[0])/res[0]\nxpf = exp(-(xarr-w)**2 / (2*c**2)) + exp(-(xarr-(1.0-w))**2 / (2*c**2))\nxpf[where(xarr > w)[0][0]:where(xarr < (1.0-w))[0][-1]] = 1.0\nypf = exp(-(yarr-w)**2 / (2*c**2)) + exp(-(yarr-(1.0-w))**2 / (2*c**2))\nypf[where(yarr > w)[0][0]:where(yarr < (1.0-w))[0][-1]] = 1.0\nxmh, ymh = meshgrid(xpf, ypf)\nmask = xmh*ymh\nmask = mask/mask.max()\n\n# Create / clear directories for use\nos.system('mkdir ./in/')\nos.system('mkdir ./out/')\nfiles = glob.glob('./out/*')\nfor f in files: os.remove(f)\n\n# Grab the files to process\nfiles = glob.glob('./in/*')\n\n# Loop through each image\nfor f in arange(len(files)):\n\t# Read in the image\n\tim = misc.imread(files[f])\n\n\t# Crop and resample the image\n\tires = im.shape\n\tiapct = float(ires[0]) / ires[1]\n\tif iapct > apct:\n\t\toim = scipy.misc.imresize(im, [int(round(iapct*res[1])), int(res[1])])\n\t\tpcut = int((oim.shape[0]-res[0])/2)\n\t\toim = oim[pcut:pcut+int(res[0]),:,:]\n\telse:\n\t\toim = scipy.misc.imresize(im, [int(res[0]), int(round(res[0]/iapct))])\n\t\tpcut = int((oim.shape[1]-res[1])/2)\n\t\toim = oim[:,pcut:pcut+int(res[1]),:]\n\n\t# Apply the gaussian blending\n\tfor ch in arange(3):\n\t\toim[:,:,ch] = oim[:,:,ch] * mask\n\n\t# Create a new copy of the modified image\n\tmisc.imsave('./out/'+'%05.f'%f+'.png', oim)\n","sub_path":"wallpaper-applewatch.py","file_name":"wallpaper-applewatch.py","file_ext":"py","file_size_in_byte":1816,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"388452565","text":"from selenium import webdriver\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.common.action_chains import ActionChains\nfrom selenium.webdriver.support import expected_conditions\nfrom selenium.webdriver.support.wait import WebDriverWait\nfrom selenium.webdriver.common.keys import Keys\nfrom selenium.webdriver.common.desired_capabilities import DesiredCapabilities\n\nfrom databases import mysql\n\nfrom time import sleep\nimport os\nimport datetime\nimport csv\n\nimport logging\nlogging.basicConfig(filename='test.log', level=logging.DEBUG)\n\n# Facebook\nfb_url = 'https://facebook.com/'\nfb_user = 'jasongame1220@gmail.com'\nfb_pwd = 'jasongod'\n\nurls = []\nqueries = []\n\n\n# logging.info('start')\nchrome_options = webdriver.ChromeOptions()\n\nchrome_options.add_argument(\"--no-sandbox\") \nchrome_options.add_argument(\"--disable-setuid-sandbox\") \nchrome_options.add_argument(\"--remote-debugging-port=9222\") # this\nchrome_options.add_argument(\"--disable-dev-shm-using\") \nchrome_options.add_argument(\"--disable-extensions\") \nchrome_options.add_argument(\"--disable-gpu\") \nchrome_options.add_argument(\"start-maximized\") \nchrome_options.add_argument(\"disable-infobars\") \nchrome_options.add_argument(\"--headless\") \nchrome_options.add_argument('--ignore-certificate-errors')\n\n# IPHONE Browser\n# chrome_options.add_argument('--user-agent=\"Mozilla/5.0 (Linux; Android 8.0; Pixel 2 Build/OPD3.170816.012) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/87.0.4280.88 Mobile Safari/537.36\"')\n# MAC Chrome Browser\nchrome_options.add_argument('--user-agent=\"Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/87.0.4280.88 Safari/537.36\"') \n\ndriver = webdriver.Chrome(chrome_options=chrome_options)\ndriver.set_window_size(1440, 877)\n\nfbSqlData = mysql.mysql()\n\ndef transform(url):\n\n # sleep(1)\n try : \n # Login\n driver.get(fb_url + 'login')\n\n sleep(4)\n # Login FaceBook\n username_input = driver.find_element_by_css_selector(\"input[name='email']\")\n password_input = driver.find_element_by_css_selector(\"input[name='pass']\")\n username_input.send_keys(fb_user)\n password_input.send_keys(fb_pwd)\n driver.find_element(By.NAME, \"login\").click()\n\n except : \n print(\"is logging\")\n\n sleep(1)\n print('Processing: %s' % url)\n # new_url = fb_url + '%s' % url + '/about'\n new_url = fb_url + '%s' % url + '/community'\n driver.get(new_url)\n sleep(2)\n print(new_url)\n\n # Likes\n likes = driver.find_element(By.CSS_SELECTOR, \".s1tcr66n > .d2edcug0\")\n likes = likes.text\n\n fbSqlData.insertOne(url,'likes',likes)\n print(likes)\n # Followers\n followers = driver.find_element(By.CSS_SELECTOR, \".bp9cbjyn:nth-child(2) > .d2edcug0\").text\n\n fbSqlData.insertOne(url,'followers',followers)\n print(followers)\n # Profile picture (avatar)\n # aux_profile_picture = driver.find_element_by_xpath('//*[@id=\"mount_0_0\"]/div/div[1]/div[1]/div[3]/div/div/div[1]/div[1]/div[2]/div/div/div[1]/div[2]/div/div/div/div[1]/div/div/a')\n # profile_picture = aux_profile_picture.get_attribute('href')\n # print(profile_picture)\n # sleep(1)\n # print('UPDATE facebook SET avatar = \"%s\", likes = \"%s\", subscriptions = \"%s\" WHERE url = \"%s\";' % (profile_picture, likes, followers, url))\n # queries.append('UPDATE facebook SET avatar = \"%s\", likes = \"%s\", subscriptions = \"%s\" WHERE url = \"%s\";' % (profile_picture, likes, followers, url))\n # Closes the Selenium driver (Chrome)\n # driver.close()\n return\n\n\nkeyList = fbSqlData.getFBKey()\nkeyCount = fbSqlData.getCount()\nnowCount = 1\n\nfor row in keyList:\n if nowCount == keyCount : \n transform(row)\n driver.close()\n else : \n transform(row)\n nowCount+=1\n\n# with open('../fb_user_list.csv', 'r') as file:\n# csvReader = csv.reader(file)\n# # row_count = sum(1 for row in csvReader)\n# now_count = 1\n# row_count = 2\n\n# for row in csvReader:\n# print('gg')\n\n# if now_count == row_count : \n# print('he')\n# transform(row[0])\n# driver.close()\n# else : \n# print('fu')\n# transform(row[0])\n# now_count+=1\nprint(queries)","sub_path":"FB/getBasicInfo.py","file_name":"getBasicInfo.py","file_ext":"py","file_size_in_byte":4237,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"366089810","text":"# -*- coding: utf-8 -*-\n#\n# Telefónica Digital - Product Development and Innovation\n#\n# THIS CODE AND INFORMATION ARE PROVIDED \"AS IS\" WITHOUT WARRANTY OF ANY KIND,\n# EITHER EXPRESSED OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE IMPLIED\n# WARRANTIES OF MERCHANTABILITY AND/OR FITNESS FOR A PARTICULAR PURPOSE.\n#\n# Copyright (c) Telefónica Investigación y Desarrollo S.A.U.\n# All rights reserved.\n#\nfrom setuptools import setup\n\nfrom version import parse_pom_version\n\n\nSNAPSHOT_SUFFIX='-SNAPSHOT'\n\n\ndef lookup_version():\n pom_version = parse_pom_version('../pom.xml')\n if pom_version.endswith(SNAPSHOT_SUFFIX):\n # EGG timestamp is managed at deployment time so we remove the\n # snapshot suffix here\n return pom_version[:-len(SNAPSHOT_SUFFIX)]\n else:\n return pom_version\n\n\nsetup(name='cosmos',\n version=lookup_version(),\n author='Cosmos Team',\n author_email='cosmos@tid.es',\n packages=['cosmos', 'cosmos.cli', 'cosmos.common', 'cosmos.compute', 'cosmos.storage'],\n package_data={\n 'cosmos.common': ['cacerts.pem'],\n },\n entry_points={\n 'console_scripts': ['cosmos=cosmos.cli.main:run']},\n install_requires=[\n 'pyyaml',\n 'pymlconf',\n 'requests',\n ],\n extras_require=dict(\n test=[\n 'mock',\n 'web.py',\n 'testfixtures'\n ])\n )\n","sub_path":"cosmos-cli/setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":1414,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"110385195","text":"# solution1: BFS\nclass Solution1:\n def findOrder(self, numCourses, prerequisites):\n \"\"\"\n :type numCourses: int\n :type prerequisites: List[List[int]]\n :rtype: List[int]\n \"\"\"\n indegrees = [0] * numCourses\n courseOrder = []\n nextCourses = [[] for _ in range(numCourses)]\n queue = collections.deque()\n \n for pair in prerequisites:\n indegrees[pair[0]] += 1\n nextCourses[pair[1]].append(pair[0])\n \n for i in range(numCourses):\n if indegrees[i] == 0:\n queue.append(i)\n\n while queue:\n course = queue.popleft()\n courseOrder.append(course)\n for nex in nextCourses[course]:\n indegrees[nex] -= 1\n if indegrees[nex] == 0:\n queue.append(nex)\n\n return courseOrder if len(courseOrder) == numCourses else []\n\n# solution2: DFS\nclass Solution2:\n def findOrder(self, numCourses, prerequisites):\n \"\"\"\n :type numCourses: int\n :type prerequisites: List[List[int]]\n :rtype: List[int]\n \"\"\"\n def dfs(course, nextCourses, visited, courseOrder):\n if visited[course] == 2: # visited in outside for-loop\n return True\n elif visited[course] == 1: # visited in dfs recursion\n return False\n else: # unvisited\n visited[course] = 1\n for nex in nextCourses[course]:\n if not dfs(nex, nextCourses, visited, courseOrder):\n return False\n visited[course] = 2\n courseOrder.appendleft(course)\n return True\n \n visited = [0] * numCourses\n courseOrder = collections.deque()\n nextCourses = [[] for _ in range(numCourses)]\n for pair in prerequisites:\n nextCourses[pair[1]].append(pair[0])\n for i in range(numCourses):\n if not dfs(i, nextCourses, visited, courseOrder):\n return []\n return list(courseOrder)\n","sub_path":"python/210-course-schedule-2.py","file_name":"210-course-schedule-2.py","file_ext":"py","file_size_in_byte":2109,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"516032698","text":"#!/usr/bin/env python\n\n# See: https://github.com/fiedl/hole-ice-study/issues/64\n\n# import code; code.interact(local=dict(globals(), **locals())) # like binding.pry\n\nimport sys\nimport pandas\n\ndata_file = \"~/hole-ice-study/results/cross_checks/cross_check_64.txt\"\ndata = pandas.read_csv(data_file, delim_whitespace = True, names = [\"cross\", \"check\", \"key\", \"equals\", \"photonTotalPathLength\"])\n\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n# prepare canvas\nfig, axes = plt.subplots(1, 2, facecolor=\"white\")\n\ndef exponential(x, n0, lambd):\n return n0 * np.exp(-x * 1.0 / lambd)\n\nplot_range_min = 0.0\nplot_range_max = 1.0\nplot_range = (plot_range_min, plot_range_max)\n\ndef str_round(number):\n return \"{:.4f}\".format(number)\n\nfor ax in axes:\n\n # histogram the data\n n, bins, patches = ax.hist(data[\"photonTotalPathLength\"], bins = 50, range = plot_range, label = 'simulation data, $\\lambda_{\\mathrm{abs}}$ = ' + str_round(0.1) + 'm')\n bin_width = (plot_range_max - plot_range_min) / 50\n x = bins[0:-1] #+ bin_width / 2\n n_error = np.sqrt(n)\n\n # filter for inf values: `ValueError: Residuals are not finite in the initial point.`\n # https://stackoverflow.com/a/33876974/2066546\n valid = ~(n == 0)\n n = n[valid]\n x = x[valid]\n n_error = n_error[valid]\n\n # fit\n from scipy.optimize import curve_fit\n parameters, cov = curve_fit(exponential, x, n,\n bounds = [[0, 0], [np.inf, np.inf]],\n sigma = n_error)\n\n fit_n0 = parameters[0]\n fit_lambd = parameters[1]\n errors = np.sqrt(np.diag(cov))\n fit_lambd_error = errors[1]\n\n # plot fit\n x = np.linspace(plot_range_min, plot_range_max, 50)\n y = exponential(x, fit_n0, fit_lambd)\n bin_width = (plot_range_max - plot_range_min) / 50\n ax.plot(x + bin_width / 2, y, label = 'exponential fit, $\\lambda_{\\mathrm{abs}}$ = ' + str_round(fit_lambd) + 'm $\\pm$ ' + str_round(fit_lambd_error) + 'm', linewidth = 2.0)\n\n ax.set_xlabel(\"photon total path length [m]\")\n\n ax.grid()\n ax.legend(loc = \"upper right\")\n\naxes[1].set_yscale(\"log\")\n\nnumber_of_photons = data[\"photonTotalPathLength\"].count()\nfig.suptitle(\"Cross check #64: Photons within hole ice, number of photons = \" + str(number_of_photons), fontsize = 14)\n\nplt.show()","sub_path":"scripts/lib/plot_cross_check_photonTotalPathLength_distribution.py","file_name":"plot_cross_check_photonTotalPathLength_distribution.py","file_ext":"py","file_size_in_byte":2193,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"457092965","text":"import os\nfrom setuptools import find_packages, setup\n\nwith open(os.path.join(os.path.dirname(__file__), 'README.rst')) as readme:\n README = readme.read()\n\n# allow setup.py to be run from any path\nos.chdir(os.path.normpath(os.path.join(os.path.abspath(__file__), os.pardir)))\n\nsetup(\n name='django-astrosat',\n version='0.2',\n packages=find_packages(),\n install_requres=[\"django\"],\n include_package_data=True,\n description='Behold Thermcert!',\n long_description=README,\n author='Allyn Treshansky',\n author_email='allyn.treshansky@astrosat.space',\n classifiers=[\n 'Environment :: Web Environment',\n 'Framework :: Django',\n 'Framework :: Django :: X.Y', # replace \"X.Y\" as appropriate\n 'Intended Audience :: Developers',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n ],\n)\n","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":975,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"462298882","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n'''\nFortgesschrittenenpraktikum F09/10 - Neuromorphic Computing\nTask 4 - Short Term Plasticity\n\nAndreas Gruebl, July 2016, agruebl@kip.uni-heidelberg.de\n\nThis network demonstrates short-term plasticity (STP) on hardware.\nThe postsynaptic neuron is stimulated by a single input with STP enabled.\nFor high input rates the impact of each presynaptic spike on the membrane\npotential decreases.\nFor low input rates the synaptic efficacy recovers.\n'''\nimport pyNN.hardware.spikey as pynn\nimport numpy as np\n\n# for plotting without X-server\nimport matplotlib as mpl\nmpl.use('Agg')\nimport matplotlib.pyplot as plt # noqa\n\ndef plot_settings(title, xlabel, ylabel):\n plt.figure(figsize=(8,8))\n plt.title(title, fontsize=14)\n plt.xlabel(xlabel, fontsize=12)\n plt.ylabel(ylabel, fontsize=12)\n plt.grid(True)\n\n# row and column of synapse\n# you can play around with these parameters to find a \"nice\" synapse...\nneuronIndex = 42\nsynapseDriverIndex = 42\nuse_other_spikey_half = False\n\ninterval = [40., 50., 60.]\nstart = 100.\nlate_spike = [600., 650., 700., 750., 800.]\nutil = [0.2, 0.4, 0.6, 0.8]\ntau_rec = [40., 60., 80., 100., 120.]\ntau_facil = 0.0\n\nweight = 15.0\n\n\nfor interval_it in interval:\n\tearly_spikes = np.arange(start, start + interval_it * 6 + 1, interval_it)\n\tfor late_spike_it in late_spike:\n\t\tstimParams = {'spike_times': np.concatenate((early_spikes, [late_spike_it]))}\n\n\t\t# STP parameters (depression and facilitation cannot be enabled\n\t\t# simultaneously!):\n\t\t# U: Usable synaptic efficacy (U_SE, see script) - scales the size of PSPs.\n\t\t# U has to lie between 0 and 1\n\t\t# tau_rec: time constant of short term depression\n\t\t# tau_facil: time constant of short term facilitation\n\t\t# either tau_rec or tau_facil must be zero\n\t\tfor util_it in util:\n\t\t\tfor tau_rec_it in tau_rec:\n\t\t\t\tstpParams = {'U': util_it, 'tau_rec': tau_rec_it, 'tau_facil': tau_facil}\n\t\t\t\truntime = 1000.0\n\n\t\t\t\tif use_other_spikey_half:\n\t\t\t\t neuron_offset = 192\n\t\t\t\telse:\n\t\t\t\t neuron_offset = 0\n\t\t\t\tpynn.setup(mappingOffset=neuronIndex+neuron_offset)\n\t\t\t\tif weight > 0:\n\t\t\t\t weight *= pynn.minExcWeight()\n\t\t\t\t synapsetype = 'excitatory'\n\t\t\t\telse:\n\t\t\t\t weight *= pynn.minInhWeight()\n\t\t\t\t synapsetype = 'inhibitory'\n\n\t\t\t\tneuron = pynn.Population(1, pynn.IF_facets_hardware1)\n\t\t\t\tdummy = pynn.Population(synapseDriverIndex, pynn.SpikeSourceArray, stimParams)\n\t\t\t\tstimulus = pynn.Population(1, pynn.SpikeSourceArray, stimParams)\n\n\t\t\t\t# enable and configure STP\n\t\t\t\tstp_model = pynn.TsodyksMarkramMechanism(**stpParams)\n\t\t\t\tpynn.Projection(stimulus, neuron,\n\t\t\t\t\t\tmethod=pynn.AllToAllConnector(weights=weight),\n\t\t\t\t\t\ttarget='excitatory',\n\t\t\t\t\t\tsynapse_dynamics=pynn.SynapseDynamics(fast=stp_model))\n\n\t\t\t\tpynn.record_v(neuron[0], '')\n\n\t\t\t\tpynn.run(runtime)\n\n\t\t\t\tmembrane = np.array(zip(pynn.timeMembraneOutput, pynn.membraneOutput))\n\n\t\t\t\tpynn.end()\n\n\n\t\t\t\tname = \"./test/stp_util\" + str(util_it).replace('.','') + \"_taurec\" + str(int(tau_rec_it)) + \"_lastSpike\" + str(int(late_spike_it)) + \"_earlySpikeIntervall\" + str(int(interval_it)) + \".pdf\"\n\n\t\t\t\t# plot\n\t\t\t\tplot_settings(\"Short-term placicity with Utilisation = {}, tau_rec = {}, tau_fac = {}\".format(util_it, tau_rec_it, tau_facil), \"Time [ms]\", \"Membrane potential [mV]\")\n\t\t\t\tplt.plot(membrane[:, 0], membrane[:, 1])\n\t\t\t\t#plt.text(600, -72.4, \"Early spikes at {} \\nLate spike at {}\\nUtilisation = {} \\n$\\tau_rec$ = {} \\n$\\tau_fac$ = {}\".format(early_spikes, late_spike, util_it, tau_rec, tau_facil)) \n\t\t\t\t#plt.xlabel('time (ms)')\n\t\t\t\t#plt.ylabel('membrane potential (mV)')\n\t\t\t\tplt.savefig(name)\n","sub_path":"F09_Neuromorphes_Rechnen/various/bernd_and_jan/F09/data/fp_task4_stp_loop.py","file_name":"fp_task4_stp_loop.py","file_ext":"py","file_size_in_byte":3603,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"153703186","text":"\"\"\"2. Latin Translator\n Look at the following list of Latin words and their meanings.\n Latin English\n sinister left\n dexter right\n medium center\n Write a GUI program that translates the Latin words to English. The window should have\n three buttons, one for each Latin word. When the user clicks a button, the program displays \n the English translation in a label.\"\"\"\n\n\nimport tkinter as tk\n\ndef translate(n):\n if n == 1:\n l1.configure(text='left')\n l1.update()\n elif n == 2:\n l2.configure(text='right')\n l2.update()\n else:\n l3.configure(text='middle')\n l3.update()\n\nwindow = tk.Tk()\nwindow.title(\"Latin Translations\")\nwindow.geometry('800x400')\n\nl1 = tk.Label(master = window, text= '')\nl2 = tk.Label(master = window, text= '')\nl3 = tk.Label(master = window, text= '')\n\n\nbutton1 = tk.Button(text=\"sinister\", command = lambda: translate(1)).pack()\nbutton2 = tk.Button(text='dexter', command = lambda: translate(2)).pack()\nbutton3 = tk.Button(text='medium', command = lambda: translate(3)).pack()\nl1.pack()\nl2.pack()\nl3.pack()\n\nwindow.mainloop()","sub_path":"Class Exercises 4/tlatin.py","file_name":"tlatin.py","file_ext":"py","file_size_in_byte":1153,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"184244787","text":"\"\"\" logging related functionality \"\"\"\n\n\nimport logging\n\nfrom torch.utils.tensorboard import SummaryWriter\n\nlogger = logging.getLogger()\n\n\ndef print_verbose(string, verbose):\n if verbose:\n print(string)\n\n\ndef loss_logger_helper(\n loss, aux_loss, writer: SummaryWriter, step: int, epoch: int, log_every: int,\n string: str = \"train\", force_print: bool = False, new_line: bool = False\n):\n # write to tensorboard at every step but only print at log step or when force_print is passed\n writer.add_scalar(f\"{string}/loss\", loss, step)\n for k, v in aux_loss.items():\n writer.add_scalar(f\"{string}/\" + k, v, step)\n\n if step % log_every == 0 or force_print:\n logger.info(f\"{string}/loss: ({step}/{epoch}) {loss}\")\n\n if force_print:\n if new_line:\n for k, v in aux_loss.items():\n logger.info(f\"{string}/{k}:{v} \")\n else:\n str_ = \"\"\n for k, v in aux_loss.items():\n str_ += f\"{string}/{k}:{v} \"\n logger.info(f\"{str_}\")\n","sub_path":"lib/utils/logging.py","file_name":"logging.py","file_ext":"py","file_size_in_byte":1046,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"642940719","text":"import got3\nimport arrow\nfrom textblob import TextBlob\nimport numpy as np\n\ndef format_date(date):\n months = {\"Jan\" : 1, \"Feb\" : 2, \"Mar\" : 3, \"Apr\" : 4, \"May\" : 5, \"Jun\" : 6, \"Jul\" : 7, \"Aug\" : 8, \"Sep\" : 9, \"Oct\" : 10, \"Nov\" : 11, \"Dec\" : 12}\n date = date.split('-')\n\n DD = str(date[0])\n\n MM = int(months[date[1]])\n if MM < 10:\n MM = \"0\" + str(MM)\n else:\n MM = str(MM)\n\n YYYY = int(date[2])\n if YYYY < 25:\n YYYY = str(YYYY + 2000)\n else:\n YYYY = str(YYYY + 1900)\n\n output_date = YYYY + \"-\" + MM + \"-\" + DD\n return arrow.get(output_date)\n\n\ndef date_to_sentiment(dates, ticker, max_tweets):\n\n sentiments = []\n for d in dates:\n arrow_date = format_date(d)\n tweetCriteria = got3.manager.TweetCriteria().setQuerySearch(\"{}{}\".format(\"#\", ticker)).setMaxTweets(max_tweets) \\\n .setSince(arrow_date.format(\"YYYY-MM-DD\")) \\\n .setUntil(arrow_date.replace(days=1).format(\"YYYY-MM-DD\"))\n print(arrow_date.format(\"YYYY-MM-DD\"))\n print(arrow_date.replace(days=1).format(\"YYYY-MM-DD\"))\n tweets = got3.manager.TweetManager.getTweets(tweetCriteria)\n\n sents_per_date = []\n for t in tweets:\n print(t.text)\n blob = TextBlob(t.text)\n sent = blob.sentiment[0] #get the polarity (subjectivity seems less important)\n sents_per_date.append(sent)\n\n sents_per_date = np.asarray(sents_per_date)\n\n # #warning insight\n # try:\n # sentiments.append(sents_per_date.mean())\n # except RuntimeWarning:\n # print(\"RUNTIME WARNING\")\n # print(d)\n # print(sents_per_date)\n # for t in tweets:\n # print(t.text)\n\n return sentiments\n\n\n# #UNIT TEST\n# dates = ['10-Aug-17']#, '15-Aug-16']\n# ticker = 'GOOGL'\n# max_tweets = 200\n#\n# print(date_to_sentiment(dates, ticker, max_tweets))\n#\n# # tweetCriteria = got3.manager.TweetCriteria().setUsername('barackobama').setMaxTweets(1)\n# # tweet = got3.manager.TweetManager.getTweets(tweetCriteria)[0]\n# # print(tweet.text)\n","sub_path":"Version_3_twitter_sentiment/date_handler.py","file_name":"date_handler.py","file_ext":"py","file_size_in_byte":2102,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"546268348","text":"\n#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Feb 22 15:46:10 2017\n\n@author: administorzz\n\"\"\"\n\nimport numpy as np\nimport pandas as pd\n\n\nclass GaussianNB():\n \n \n \n def __init__(self):\n self.offset = 10e-5\n self.p_x_yi = 0\n \n def accurateScore(self,y1,y2):\n score = y1==y2\n return np.average(score)\n \n def confusion_matrix(self,y1,y2):\n return pd.crosstab(y1,y2,\n rownames=['Actual'],colnames=['Predicted'])\n \n def computeMeans(self, X, y):\n# mean value of each feature per class(conditional)\n num_classes = len(np.unique(y))\n self.sum_mean = np.zeros((num_classes, self.n_features))\n self.sum_var = np.zeros(self.n_features)\n \n for y_n in range(num_classes):\n X_i = X[y == y_n]\n self.sum_mean[y_n] = np.mean(X_i)\n \n def computeVariance(self, X, y):\n num_classes = len(np.unique(y))\n self.sum_var = np.zeros((num_classes, self.n_features))\n for y_n in range(num_classes):\n X_i = X[y == y_n]\n self.sum_var[y_n] = np.var(X_i) \n self.sum_var += self.offset\n \n \n def prior_class(self,y):\n num_classes = len(np.unique(y))\n self.p_y = np.zeros(num_classes)\n \n for y_n in range(num_classes):\n self.p_y[y_n] = len(y[y == y_n]) / float(len(y))\n\n def joint_likelihood(self,x):\n temp =0\n var1=0\n var2=0\n # var3 = log(P(y=class|X=instance))\n var3=0\n var4=0\n var5=[]\n self.p_yi_x = np.zeros(len(self.p_y))\n for i in range(len(self.p_y)):\n var2 = 0\n for j in range(self.n_features):\n x_nf = x[j]\n temp = np.square(x_nf - self.sum_mean[i,j]) / (2 * self.sum_var[i,j])\n self.p_xj_yi= (1/ np.sqrt(2* np.pi * np.sqrt(self.sum_var[i,j]))) * np.exp(-temp)\n # calculate the P(X=instance|y=class)\n var1 = np.log(self.p_xj_yi)\n #np.log(self.p_x_yi) = np.log(self.p_x_yi) + np.log(self.p_xj_yi)\n # calculate the P(y=class|X=instance) we can get the predict value\n var2 += var1\n var3 = np.log(self.p_y[i])\n var4 = var2 + var3\n var5.append(var4)\n \n #return np.argmax(self.p_yi_x)\n return np.argmax(var5)\n \n \n def fit(self, X, y):\n \n self.n_instances,self.n_features = np.shape(X)\n self.computeMeans(X,y)\n \n self.computeVariance(X,y)\n self.prior_class(y)\n \n return self\n \n def predict(self,X,y):\n y_pred = np.zeros(len(y))\n for i in range(self.n_instances):\n x_n = X.ix[i,:]\n y_pred[i] = self.joint_likelihood(x_n)\n \n return y_pred\n\nif __name__==\"__main__\": \n nb = GaussianNB()\n with open(\"output.txt\",\"w\") as f_out:\n df = pd.read_csv('/Users/administorzz/Downloads/digits.csv')\n X = df.ix[:,:-1]\n y = df.ix[:,-1]\n y_true = y\n \n nb.fit(X,y)\n y_pred = nb.predict(X,y)\n \n f_out.write(\"{0:.3f}\".format(nb.accurateScore(y_true, y_pred)))\n f_out.write('\\n')\n f_out.write(str(nb.confusion_matrix(y_true, y_pred)))\n ","sub_path":"Gaussian Naive Bayes2.py","file_name":"Gaussian Naive Bayes2.py","file_ext":"py","file_size_in_byte":3354,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"91890914","text":"'''\nAs Organizações Tabajara resolveram dar um aumento de salário aos seus colaboradores e lhe contrataram para\ndesenvolver o programa que calculará os reajustes.\nFaça um programa que recebe o salário de um colaborador e o reajuste segundo o seguinte critério, baseado no salário atual:\n- salários até R$ 280,00 (incluindo) : aumento de 20%\n- salários entre R$ 280,00 e R$ 700,00 : aumento de 15%\n- salários entre R$ 700,00 e R$ 1500,00 : aumento de 10%\n- salários de R$ 1500,00 em diante : aumento de 5%\nApós o aumento ser realizado, informe na tela:\n- o salário antes do reajuste;\n- o percentual de aumento aplicado;\n- o valor do aumento;\n- o novo salário, após o aumento.\n'''\nimport sys\n\ndef myinput(texto):\n retorno = ''\n if sys.version_info.major == 2:\n retorno = raw_input(texto)\n elif sys.version_info.major == 3:\n retorno = input(texto)\n return retorno\n\nqtde = 1\ni = 1\nwhile i <= qtde:\n n = myinput('Digite o salário atual do colaborador em R$: ')\n\n if n.isalpha():\n print('O texto \"{}\" do salário do colaborador não é um valor válido' . format(n))\n elif float(n) <= 0:\n print('O salário do colaborador deve ser um valor positivo')\n else:\n n = float(n)\n i += 1\n\nif n <= 280.00:\n reajuste = 0.20\nelif n > 280.00 and n <= 700.00:\n reajuste = 0.15\nelif n > 700.00 and n <= 1500.00:\n reajuste = 0.10\nelse:\n reajuste = 0.05\n\nvalor_reajuste = n * reajuste\nnovo_salario = n + valor_reajuste\n\nprint('Salário antes do reajuste R$ {:.2f}' . format(n))\nprint('Percentual de aumento aplicado {:.2f}%' . format(reajuste * 100))\nprint('valor do aumento R$ {:.2f}' . format(valor_reajuste))\nprint('Novo salário R$ {:.2f}' . format(novo_salario))","sub_path":"Maratona Data Science Brazil/Semana#01 - Python/02-estruturas-de-decisao/exercicio-11.py","file_name":"exercicio-11.py","file_ext":"py","file_size_in_byte":1741,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"572717062","text":"from datadog import api, initialize\n\noptions = {\n 'api_key':'25af742dc74b5c5d1efd29459b5c7c08',\n 'app_key':'bcc64e922e922b0e92282821fd15e044d85aa46f',\n 'api_host': 'https://api.datadoghq.com'\n}\n\ninitialize(**options)\n\n\nrun = 'y'\n\ndef remove_User():\n email = input(\"Please enter the emailaddress of the user to disable: \")\n\n data = api.User.delete(email)\n\n\n for i,j in data.items():\n crack = i,j\n crackstr = crack[0]+ ' ' + crack[1][0]\n if 'error' in crackstr:\n print(crackstr)\n if input(\"Do you wish to try again? y or n (Default is 'n' ) \").lower() == 'y':\n #run program again\n remove_User()\n elif 'disabled' in crackstr:\n print(crackstr)\n input('command has been ran successfull')\n\n\nremove_User()\n","sub_path":"ddremoveuser.py","file_name":"ddremoveuser.py","file_ext":"py","file_size_in_byte":790,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"286876994","text":"import numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport pickle\n\ndef main():\n c0 = pickle.loads(open('cluster0dist','r').read())\n c0=c0.drop(['weekend','workday'],1)\n c0.columns = ['weekend','workday']\n c0.plot(kind='area',color=['orange','deepskyblue'],alpha=0.6, figsize=(18,6), stacked=False)\n plt.xticks(rotation=0)\n plt.xlabel('Pickup Time')\n plt.ylabel('Daily Trip Amount')\n #plt.show()\n plt.savefig('cluster0trip-black.png',dpi=200)\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"Web/Page2/clustering/cluster0trip.py","file_name":"cluster0trip.py","file_ext":"py","file_size_in_byte":527,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"286360872","text":"#Problem 014: Longest Collatz sequence\nimport datetime\n\ndef program():\n\tlim = 1000000\n\tcollatz = {1:1}\n\tdef sequence(nro):\n\t\taux = collatz.get(nro, 0)\n\t\tif aux == 0:\n\t\t\tif nro % 2 == 0:\t\n\t\t\t\tcollatz[nro] = 1 + sequence(nro//2)\n\t\t\telse: \n\t\t\t\tcollatz[nro] = 1 + sequence(nro*3 + 1)\n\t\treturn collatz[nro]\n\t\t\n\tans = [1,1]\n\tfor i in range(2, lim):\n\t\tans = max(ans, [sequence(i), i])\n\n\treturn ans[1]\n\t\t\t\n\t\nif __name__ == \"__main__\":\n\th1 = datetime.datetime.now()\n\tans = program()\n\th2 = datetime.datetime.now()\n\tprint(\"Problem 014 (\",h2 - h1,\") : \", ans)\n\n","sub_path":"PE-problem:014.py","file_name":"PE-problem:014.py","file_ext":"py","file_size_in_byte":549,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"211084845","text":"import libomni as robot #Library tha handles all the serial commands to arduino AtMega\nimport pyrealsense2 as rs\nimport numpy as np\nimport time,os,serial\nimport math as m\n\n#folder where saving all the data\nsave_folder = \"save_data/\"\ncurrent_directory = os.getcwd()\nsave_directory = os.path.join(current_directory, save_folder)\n\n#odometry setups\noldEncoder0 = 0;oldEncoder1 = 0;oldEncoder2 = 0;newEncoder0 = 0;newEncoder1 = 0;newEncoder2 = 0;\n\n#RealSense\nvel_x = 0;vel_y = 0;vel_z = 0;pos_x = 0;pos_y = 0;pos_z = 0;acc_x = 0;acc_y = 0;acc_z = 0;\n\nm1_rpm = 0;m2_rpm = 0;m3_rpm = 0;\n\nglobal m1_cur\nglobal m2_cur\nglobal m3_cur\nglobal last_yaw\nglobal vx\nglobal vy\nglobal omega\n\nlast_yaw = 0;vx = 0; vy =0;omega = 0;\nm1_cur = 0;m2_cur = 0;m3_cur = 0;\n\n# Declare RealSense pipeline, encapsulating the actual device and sensors\npipe = rs.pipeline()\n# Build config object and request pose data\ncfg = rs.config()\ncfg.enable_stream(rs.stream.pose)\n# Start streaming with requested config\npipe.start(cfg)\n\n#pid controller\nglobal integral\nglobal preError\nintegral = np.array([0,0,0])[:,None]\npreError = np.array([0,0,0])[:,None]\n\n########################################################################\t\tReset encoder\ndef initOdometry():\n\tglobal oldEncoder0 \n\tglobal oldEncoder1 \n\tglobal oldEncoder2 \n\toldEncoder0 = robot.encoder(0)\n\toldEncoder1 = robot.encoder(1)\n\toldEncoder2 = robot.encoder(2)\n\t\n#odometry position\nglobal current_x\nglobal current_y\nglobal current_theta\ncurrent_x = 0\ncurrent_y = 0\ncurrent_theta = 0 \n\n########################################################################\t\tEncoder Odometry\t\ndef odometryCalc(xk,yk,thetak,l=0.19, N=2249, r=0.03):\n\tglobal oldEncoder0\n\tglobal oldEncoder1\n\tglobal oldEncoder2\n\t\n\tnewEncoder0 = robot.encoder(0)\n\tnewEncoder1 = robot.encoder(1)\n\tnewEncoder2 = robot.encoder(2)\n\t\n\tdeltaEncoder0 = newEncoder0 - oldEncoder0\n\tdeltaEncoder1 = newEncoder1 - oldEncoder1\n\tdeltaEncoder2 = newEncoder2 - oldEncoder2\n\t\n\tD0=(deltaEncoder0/N)*((2*np.pi*r))\n\tD1=(deltaEncoder1/N)*((2*np.pi*r))\n\tD2=(deltaEncoder2/N)*((2*np.pi*r))\n\n\tkinematic_mat = np.array([1/np.sqrt(3),0,-1/np.sqrt(3),-1/3,2/3,-1/3,-1/(3*l),-1/(3*l),-1/(3*l)]).reshape(3,3)\n\t\t\n\trotation_mat= np.array([np.cos(thetak),-np.sin(thetak),0,np.sin(thetak),np.cos(thetak),0,0,0,1]).reshape(3,3)\n\t\n\t# diffrence in ticks (rpm)\n\tdistance_mat = np.array([D1,D0,D2])[:,None]\n\t\n\toldPos_mat = np.array([xk,yk,thetak])[:,None]\n\t\n\t# np.dot explanation https://stackoverflow.com/questions/21562986/numpy-matrix-vector-multiplication\n\tkinxrot = np.dot(rotation_mat,kinematic_mat)\n\tnewPos_mat = oldPos_mat + np.dot(kinxrot,distance_mat)\n\n\toldEncoder0 = newEncoder0\n\toldEncoder1 = newEncoder1\n\toldEncoder2 = newEncoder2\n\n\treturn newPos_mat\n\n########################################################################\t\tRealSense Odometry\ndef odometry_RealSense():\n\tglobal vel_x \n\tglobal vel_y \n\tglobal vel_z \n\tglobal pos_x \n\tglobal pos_y \n\tglobal pos_z \n\tglobal acc_x \n\tglobal acc_y \n\tglobal acc_z \n\tglobal theta_rs\n\tglobal yaw\n\t\n\tframes = pipe.wait_for_frames()\n\tpose = frames.get_pose_frame()\n\tdata = pose.get_pose_data()\n\t\n\tvelocity = data.velocity\n\tposition = data.translation\n\tacceleration = data.acceleration\n\trotation = data.rotation\n\t\n\t####################################################################\t\tget Velocity data\n\tvel1 = str(velocity)\t\n\tvel2 = vel1.replace(', ',' ').split(' ')\n\t# ~ print(vel2[1] + \" , \" + vel2[3] + \" , \" + vel2[5])\n\tvel_x = -float(vel2[5]) \n\tvel_y = -float(vel2[1]) \n\tvel_z = float(vel2[3]) \n\t\n\t####################################################################\t\tget Postion data\n\tpos1 = str(position)\t\n\tpos2 = pos1.replace(', ',' ').split(' ')\n\t# ~ print(pos2[1] + \" , \" + pos2[3] + \" , \" + pos2[5])\n\tpos_x = -float(pos2[5]) \n\tpos_y = -float(pos2[1]) \n\tpos_z = float(pos2[3]) \n\t\n\t####################################################################\t\tget Acceleration data\n\tacc1 = str(acceleration)\t\n\tacc2 = acc1.replace(', ',' ').split(' ')\n\t# ~ print(acc2[1] + \" , \" + acc2[3] + \" , \" + acc2[5])\n\tacc_x = -float(acc2[5]) \n\tacc_y = -float(acc2[1]) \n\tacc_z = float(acc2[3])\n\t\n\t####################################################################\t\tget Rotation data\n\ttheta1 = str(rotation)\t\n\ttheta2 = theta1.replace(', ',' ').split(' ')\n\t# ~ print(acc2[1] + \" , \" + acc2[3] + \" , \" + acc2[5])\n\t# ~ w_theta = float(theta2[7]) \n\t# ~ x_theta = float(theta2[1]) \n\t# ~ y_theta = float(theta2[3]) \n\t# ~ z_theta = float(theta2[5]) \n\tw_theta = data.rotation.w\n\tx_theta = -data.rotation.z\n\ty_theta = data.rotation.x\n\tz_theta = -data.rotation.y\n\t\t\n\ttheta_rs = np.arccos(w_theta) * 2 * 180/np.pi\n\t# ~ print(theta_rs)\n\tpitch = -np.arcsin(2.0 * (x_theta*z_theta - w_theta*y_theta));\n\troll = np.arctan2(2.0 * (w_theta*x_theta + y_theta*z_theta), w_theta*w_theta - x_theta*x_theta - y_theta*y_theta + z_theta*z_theta);\n\tyaw = -np.arctan2(2.0 * (w_theta*z_theta + x_theta*y_theta), w_theta*w_theta + x_theta*x_theta - y_theta*y_theta - z_theta*z_theta);\n\t\n\t# ~ print(\"Frame #{}\".format(pose.frame_number))\n\t# ~ print(\"RPY [deg]: Roll: {0:.7f}, Pitch: {1:.7f}, Yaw: {2:.7f}\".format(roll, pitch, yaw))\n\n\t\t\t\n\ntry: \n\twhile True:\n\t\tmode = str(input(\"Enter s to start \"))\n\n\t\tif mode == 's':\n\n\n\t\t\t############################################################\t\tReset encoder and Enable RealSense\n\t\t\tinitOdometry()\n\t\t\todometry_RealSense()\n\t\t\t\n\t\t\tt = 0\n\t\t\ttest_t = 9\n\t\t\tdelay = 0.01\n\t\t\n\t\t\t\n\t\t\twhile t < test_t:\n\t\t\t\t\n\t\t\t\tstart = time.time()\n\t\t\t\t\n\t\t\t\tx_dot = 0; #max 0.7618\n\t\t\t\ty_dot = 0; #max 0.6597\n\t\t\t\ttheta_dot = 3; #max 3.4723\n\n\n\t\t\t\t############\t\tJ, J_inverse, J_dot, J_transpose\n\t\t\t\tr = 0.03\n\t\t\t\tl = 0.19\n\t\t\t\t\n\t\t\t\tJ2_inv = np.array([1/r,0,0, 0,1/r,0, 0,0,1/r]).reshape(3,3)\n\t\t\t\tJ1 = np.array([np.sqrt(3)/2,-1/2,-l, 0,1,-l, -np.sqrt(3)/2,-1/2,-l]).reshape(3,3)\n\n\t\t\t\tv_robot= np.array([x_dot, y_dot, theta_dot]).reshape(3,1) \n\t\t\t\tJ2_inv_J1 = np.dot( J2_inv, J1, out=None).reshape(3,3) \n\t\t\t\tphi_dot = np.dot( J2_inv_J1, v_robot, out=None).reshape(3,1) \n\n\t\t\t\trpm1 = phi_dot.item(0)*60/(2*np.pi)\n\t\t\t\trpm0 = phi_dot.item(1)*60/(2*np.pi)\n\t\t\t\trpm2 = phi_dot.item(2)*60/(2*np.pi)\n\t\t\t\t\n\t\t\t\t\n\t\t\t\tif t < 2:\n\t\t\t\t\twheel0RPM = 0\n\t\t\t\t\twheel1RPM = 0\n\t\t\t\t\twheel2RPM = 0\n\t\t\t\tif t > 2 and t < 7:\n\t\t\t\t\twheel0RPM = rpm0\n\t\t\t\t\twheel1RPM = rpm1\n\t\t\t\t\twheel2RPM = rpm2\n\t\t\t\tif t > 7:\n\t\t\t\t\twheel0RPM = 0\n\t\t\t\t\twheel1RPM = 0\n\t\t\t\t\twheel2RPM = 0\t\t\n\t\t\t\t\t\n\t\t\t\t\n\t\t\t\t\n\t\t\t\tli = 46.85 #reduction of motors\n\t\t\t\tKt1 = 1.16516 / 5.6\t\t\t#Motor Torque Constant\n\t\t\t\tKt2 = 1.16516 / 5.6\t\t\t#Motor Torque Constant\n\t\t\t\tKt3 = 1.16516 / 5.6\t\t\t#Motor Torque Constant\n\t\t\t\tr = 0.03\t\t\t\t\t#Wheel Radius\n\t\t\t\tl = 0.19 #distance from wheel to CG\n\t\t\t\t\n\t\t\t\t# ~ file = open(save_folder + \"Dynamics\"+\"_Vx0.76_\"+\"_test_t_\"+str(test_t)+\".txt\",\"a\")\n\t\t\t\t# ~ file = open(save_folder + \"Dynamics\"+\"_Vy0.6_\"+\"_test_t_\"+str(test_t)+\".txt\",\"a\")\n\t\t\t\tfile = open(save_folder + \"Dynamics\"+\"_W3_\"+\"_test_t_\"+str(test_t)+\".txt\",\"a\")\n\t\t\t\t\n\t\t\t\txc = current_x\n\t\t\t\tyc = current_y\n\t\t\t\tthetac = current_theta\n\t\t\t\t\n\t\t\t\trobot.motor_rpm(int(wheel0RPM),int(wheel1RPM),int(wheel2RPM))\n\t\t\t\t\n\t\t\t\t########################################################\t\tCaculating\n\t\t\t\t##########\t\t\t\tWheel RPM\n\t\t\t\tm1_rpm = robot.rpm(0)\n\t\t\t\tm2_rpm = robot.rpm(1)\n\t\t\t\tm3_rpm = robot.rpm(2)\n\t\t\t\tdata_rpm = str(m1_rpm)+' , '+str(m2_rpm)+' , '+str(m3_rpm)\n\t\t\t\t##########\t\t\t\tMotor Current\t\n\t\t\t\tm1_cur = robot.motor_current(0)\n\t\t\t\tm2_cur = robot.motor_current(1)\n\t\t\t\tm3_cur = robot.motor_current(2)\n\t\t\t\tdata_cur = str(m1_cur)+' , '+str(m2_cur)+' , '+str(m3_cur)\n\t\t\t\t# ~ print(data_cur)\n\t\t\t\t##########\t\t\t\tMotor Voltage\t\n\t\t\t\tm1_vol = robot.motor_voltage(0)\n\t\t\t\tm2_vol = robot.motor_voltage(1)\n\t\t\t\tm3_vol = robot.motor_voltage(2)\n\t\t\t\t# ~ print(m1_vol)\n\t\t\t\tdata_vol = str(m1_vol)+' , '+str(m2_vol)+' , '+str(m3_vol)\n\t\t\t\t# ~ ##########\t\t\t\tRotation Torque\n\t\t\t\t# ~ T1 = li*Kt1*m1_cur\n\t\t\t\t# ~ T2 = li*Kt2*m2_cur\n\t\t\t\t# ~ T3 = li*Kt3*m3_cur\n\t\t\t\t# ~ ##########\t\t\t\tWheel Traction Force\n\t\t\t\t# ~ f1 = T1/r\n\t\t\t\t# ~ f2 = T2/r\n\t\t\t\t# ~ f3 = T3/r\n\t\t\t\t# ~ ##########\t\t\t\tRobot Traction Force\n\t\t\t\t# ~ fx = f2*np.sin(60) - f3*np.sin(60)\n\t\t\t\t# ~ fy = f1 - f2*np.cos(60) - f3*np.cos(60)\n\t\t\t\t# ~ M = (-f1-f2-f3)*l\n\t\t\t\t# ~ data_trac = str(fx)+' , '+str(fy)+' , '+str(M)\n\t\t\t\t\n\t\t\t\t########################################################\t\todometry using encoder\n\t\t\t\tpose = odometryCalc(xc,yc,thetac)\t\n\t\t\t\tpos = odometry_RealSense()\n\t\t\t\t\n\t\t\t\t# ~ current_x = pose.item(0)\n\t\t\t\t# ~ current_y = pose.item(1)\n\t\t\t\t# ~ current_theta = pose.item(2)\n\t\t\t\t\n\t\t\t\t########################################################\t\todometry using RealSense\n\t\t\t\tcurrent_x = pos_x\n\t\t\t\tcurrent_y = pos_y\n\t\t\t\tcurrent_theta = pose.item(2)\t\t\n\t\t\t\t\n\t\t\t\t########################################################\t\tRealSense velocities\n\t\t\t\tdt = (time.time() - start)\n\t\t\t\tdyaw = yaw - last_yaw\n\t\t\t\tvel_ang = dyaw/dt\n\t\t\t\tdata_vel = str(vel_x)+' , '+str(vel_y)+' , '+str(vel_ang)\n\t\t\t\t# ~ print(vel_y)\n\t\t\t\t\n\t\t\t\t########################################################\t\tRecording data\n\t\t\t\ttime_running = time.time()\n\t\t\t\tdata_pose = str(pose[0][0])+\" , \"+str(pose[1][0])+\" , \"+str(pose[2][0])\n\t\t\t\tdata_pos = str(pos_x)+\" , \"+str(pos_y)\n\t\t\t\t\n\t\t\t\t# ~ print(data_pose)\n\t\t\t\tfile.writelines(str(data_pose)+\" , \"+str(data_pos)+\" , \"+str(data_rpm)+\" , \"+str(data_vol)+\" , \"+str(data_cur)+\" , \"+str(data_vel)+\" , \"+str(time_running)+\"\\n\")\n\t\t\t\t\t\n\t\t\t\t########################################################\t\tRemembering value for new loop\t\t\t\n\t\t\t\ttime.sleep(delay)\n\t\t\t\telapsed_time = (time.time() - start)\n\t\t\t\tlast_yaw = yaw\n\t\t\t\tt = t + elapsed_time\n\t\t\t\t\n\t\t\trobot.stop()\n\t\t\t\t\n\n## Ctrl + c to stop robot\nexcept KeyboardInterrupt:\n # Close serial connection\n\trobot.stop() \n\t#~ file.close() \n\tprint('\\n\\n\t\tStop!!! See you again!')\n","sub_path":"OMRE_Python/dynamics_modeling/dynamics_model.py","file_name":"dynamics_model.py","file_ext":"py","file_size_in_byte":9711,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"621142412","text":"#!/usr/bin/python3\n# Copyright 2020 Google LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# https://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"Unitests for timescale.py\"\"\"\n\nimport unittest\nfrom absl.testing import absltest\nimport gps_pb2\nimport psycopg2\nimport timescale\nimport testing.postgresql\nimport mock\n\nPostgresql = testing.postgresql.PostgresqlFactory(cache_initialized_db=True)\n\n\ndef tearDownModule():\n Postgresql.clear_cache()\n\n\nclass TestTimescale(unittest.TestCase):\n \"\"\"Timescale unittest.\"\"\"\n\n def setUp(self):\n self.postgresql = Postgresql()\n with open('testdata/timescale_schema') as schema_file:\n statements = ''.join(schema_file.readlines())\n conn = psycopg2.connect(**self.postgresql.dsn())\n cursor = conn.cursor()\n cursor.execute(statements)\n conn.commit()\n self.conn = timescale.GetConnWithPointPrepare(conn)\n self.cursor = self.conn.cursor()\n point = gps_pb2.Point()\n point.time.FromJsonString('2020-09-13T01:36:38.600Z')\n self.pusher = timescale.Pusher(start_process=False)\n self.pusher.timescale_conn = self.conn\n self.pusher.track = 'Test Parking Lot'\n self.pusher.session_time = point.time\n\n def tearDown(self):\n self.cursor.close()\n self.conn.close()\n self.postgresql.stop()\n\n def testExportSession(self):\n self.pusher.ExportSession(self.cursor)\n self.assertEqual(1, self.pusher.session_id)\n\n def testExportLap(self):\n lap = gps_pb2.Lap()\n lap.number = 1\n self.pusher.ExportLap(lap, self.cursor)\n self.assertDictEqual({1: 1}, self.pusher.lap_number_ids)\n\n def testUpdateLapDuration(self):\n lap = gps_pb2.Lap()\n lap.duration.FromMilliseconds(90 * 1000)\n lap.number = 1\n self.pusher.ExportSession(self.cursor)\n self.pusher.ExportLap(lap, self.cursor)\n self.pusher.UpdateLapDuration(lap.number, lap.duration, self.cursor)\n self.cursor.execute('SELECT * FROM laps')\n self.conn.commit()\n lap_id, session_id, number, duration_ms = self.cursor.fetchone()\n self.assertEqual(1, lap_id)\n self.assertEqual(1, session_id)\n self.assertEqual(1, number)\n self.assertEqual(90000, duration_ms)\n\n def testGetElapsedTime(self):\n point = gps_pb2.Point()\n point.time.FromSeconds(10)\n self.assertEqual(0, self.pusher.GetElapsedTime(point, 1))\n point = gps_pb2.Point()\n point.time.FromSeconds(20)\n self.assertEqual(10 * 1000, self.pusher.GetElapsedTime(point, 1))\n\n def testExportPoint(self):\n lap = gps_pb2.Lap()\n lap.duration.FromMilliseconds(90 * 1000)\n lap.number = 1\n self.pusher.ExportSession(self.cursor)\n self.pusher.ExportLap(lap, self.cursor)\n self.pusher.UpdateLapDuration(lap.number, lap.duration, self.cursor)\n\n point = lap.points.add()\n point.alt = 1\n point.speed = 1\n point.lat = 45.69545832462609\n point.lon = -121.52551179751754\n point.tps_voltage = 2\n point.water_temp_voltage = 3\n point.oil_pressure_voltage = 4\n point.rpm = 1000\n point.afr = 14.7\n point.fuel_level_voltage = 5\n point.accelerometer_x = 0.0\n point.accelerometer_y = 1.7\n point.accelerometer_z = 1.2\n point.pitch = 0.2\n point.roll = 5.0\n self.pusher.ExportPoint(point, 1, self.cursor)\n self.cursor.execute('SELECT * FROM points')\n (_, _, _, lat, lon, alt, speed, geohash, elapsed_duration_ms,\n tps_voltage, water_temp_voltage, oil_pressure_voltage, rpm, afr,\n fuel_level_voltage, accelerometer_x, accelerometer_y,\n accelerometer_z, pitch, roll) = self.cursor.fetchone()\n self.assertEqual(lat, 45.6954583246261)\n self.assertEqual(lon, -121.525511797518)\n self.assertEqual(alt, 1.0)\n self.assertEqual(speed, 2.23694)\n self.assertEqual(geohash, 'c21efweg66fd')\n self.assertEqual(elapsed_duration_ms, 0.0)\n self.assertEqual(tps_voltage, 2.0)\n self.assertEqual(water_temp_voltage, 3.0)\n self.assertEqual(oil_pressure_voltage, 4.0)\n self.assertEqual(rpm, 1000)\n self.assertEqual(afr, 14.7)\n self.assertEqual(fuel_level_voltage, 5.0)\n self.assertEqual(accelerometer_x, 0.0)\n self.assertEqual(accelerometer_y, 1.7)\n self.assertEqual(accelerometer_z, 1.2)\n self.assertEqual(pitch, 0.2)\n self.assertEqual(roll, 5.0)\n\n def testExportPointArrivesBeforeLap(self):\n point = gps_pb2.Point()\n self.pusher.ExportPoint(point, 99, self.cursor)\n self.assertEqual(1, len(self.pusher.retry_point_queue))\n\n def testGetPointFromQueue(self):\n with self.subTest(name='Success'):\n point = gps_pb2.Point()\n self.pusher.AddPointToQueue(point, 1)\n result = self.pusher.GetPointFromQueue()\n self.assertTrue(result)\n if result:\n returned_point, returned_lap_number = result\n self.assertEqual(point, returned_point)\n self.assertEqual(1, returned_lap_number)\n with self.subTest(name='Retry Queue'):\n point = gps_pb2.Point()\n self.pusher.retry_point_queue.append((point, 2))\n result = self.pusher.GetPointFromQueue()\n self.assertTrue(result)\n if result:\n returned_point, returned_lap_number = result\n self.assertEqual(point, returned_point)\n self.assertEqual(2, returned_lap_number)\n with self.subTest(name='Retry and Point Queue'):\n point = gps_pb2.Point()\n self.pusher.AddPointToQueue(point, 1)\n point = gps_pb2.Point()\n self.pusher.retry_point_queue.append((point, 2))\n result = self.pusher.GetPointFromQueue()\n self.assertTrue(result)\n if result:\n returned_point, returned_lap_number = result\n self.assertEqual(point, returned_point)\n self.assertEqual(1, returned_lap_number)\n\n @mock.patch.object(timescale, 'ConnectToDB')\n def testDo(self, mock_conn):\n mock_conn.return_value = self.conn\n lap = gps_pb2.Lap()\n lap.duration.FromMilliseconds(90 * 1000)\n lap.number = 1\n point = lap.points.add()\n point.alt = 1\n point.speed = 1\n point.lat = 45.69545832462609\n point.lon = -121.52551179751754\n point.tps_voltage = 2\n point.water_temp_voltage = 3\n point.oil_pressure_voltage = 4\n point.rpm = 1000\n point.afr = 14.7\n point.fuel_level_voltage = 5\n point.accelerometer_x = 0.0\n point.accelerometer_y = 1.7\n point.accelerometer_z = 1.2\n self.pusher.lap_queue.put(lap)\n self.pusher.AddPointToQueue(point, 1)\n with self.subTest(name='Success'):\n self.pusher.Do()\n self.cursor.execute('SELECT count(*) FROM sessions')\n self.assertEqual(1, self.cursor.fetchone()[0])\n self.cursor.execute('SELECT count(*) FROM laps')\n self.assertEqual(1, self.cursor.fetchone()[0])\n self.cursor.execute('SELECT count(*) FROM points')\n self.assertEqual(1, self.cursor.fetchone()[0])\n self.assertEqual(0, self.pusher.lap_queue.qsize())\n self.assertEqual(0, self.pusher.lap_duration_queue.qsize())\n self.assertEqual(0, len(self.pusher.point_queue))\n with self.subTest(name='Second Point'):\n self.pusher.AddPointToQueue(point, 1)\n self.pusher.Do()\n self.cursor.execute('SELECT count(*) FROM sessions')\n self.assertEqual(1, self.cursor.fetchone()[0])\n self.cursor.execute('SELECT count(*) FROM laps')\n self.assertEqual(1, self.cursor.fetchone()[0])\n self.cursor.execute('SELECT count(*) FROM points')\n self.assertEqual(2, self.cursor.fetchone()[0])\n self.assertEqual(0, self.pusher.lap_queue.qsize())\n self.assertEqual(0, self.pusher.lap_duration_queue.qsize())\n self.assertEqual(0, len(self.pusher.point_queue))\n with self.subTest(name='Lap Duration'):\n self.pusher.lap_duration_queue.put((1, lap.duration))\n self.pusher.AddPointToQueue(point, 1)\n self.pusher.Do()\n self.assertEqual(0, self.pusher.lap_queue.qsize())\n self.assertEqual(0, self.pusher.lap_duration_queue.qsize())\n self.assertEqual(0, len(self.pusher.point_queue))\n with self.subTest(name='Point Too Early'):\n self.pusher.point_queue.append((point, 2))\n self.pusher.Do()\n self.cursor.execute('SELECT count(*) FROM sessions')\n self.assertEqual(1, self.cursor.fetchone()[0])\n self.cursor.execute('SELECT count(*) FROM laps')\n self.assertEqual(1, self.cursor.fetchone()[0])\n self.cursor.execute('SELECT count(*) FROM points')\n self.assertEqual(3, self.cursor.fetchone()[0])\n self.assertEqual(1, len(self.pusher.retry_point_queue))\n self.assertEqual(0, self.pusher.lap_queue.qsize())\n self.assertEqual(0, self.pusher.lap_duration_queue.qsize())\n self.assertEqual(1, len(self.pusher.retry_point_queue))\n with self.subTest(name='Exception'):\n lap.number = 2\n lap.duration.FromMilliseconds(90 * 1000)\n self.pusher.lap_queue.put(lap)\n self.pusher.lap_duration_queue.put((1, lap.duration))\n self.pusher.AddPointToQueue(point, 1)\n with mock.patch.object(self.pusher, 'ExportPoint') as mock_export:\n mock_export.side_effect = psycopg2.Error\n self.pusher.Do()\n self.assertEqual(1, self.pusher.lap_queue.qsize())\n self.assertEqual(1, self.pusher.lap_duration_queue.qsize())\n self.assertEqual(1, len(self.pusher.retry_point_queue))\n\n def testDoCommitCycle(self):\n \"\"\"Ensures points aren't dropped if errrors arrive between commits.\"\"\"\n point = gps_pb2.Point()\n point.alt = 1\n point.speed = 1\n point.lat = 45.69545832462609\n point.lon = -121.52551179751754\n point.tps_voltage = 2\n point.water_temp_voltage = 3\n point.oil_pressure_voltage = 4\n point.rpm = 1000\n point.afr = 14.7\n point.fuel_level_voltage = 5\n point.accelerometer_x = 0.0\n point.accelerometer_y = 1.7\n point.accelerometer_z = 1.2\n self.assertEqual(0, len(self.pusher.point_queue))\n self.pusher.AddPointToQueue(point, 1)\n self.pusher.Do()\n self.pusher.AddPointToQueue(point, 1)\n self.pusher.Do()\n with mock.patch.object(self.pusher, '_Commit') as mock_commit:\n mock_commit.side_effect = psycopg2.Error\n self.pusher.AddPointToQueue(point, 1)\n self.pusher.Do()\n self.assertEqual(3, len(self.pusher.retry_point_queue))\n\n\nif __name__ == '__main__':\n absltest.main()\n","sub_path":"timescale_test.py","file_name":"timescale_test.py","file_ext":"py","file_size_in_byte":10519,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"531148505","text":"#! /usr/bin/env python3\n# -*- coding: utf-8 -*-\n\n\nimport random\n\ndef main():\n again = 'Y'\n while again == 'Y' or again == 'y':\n key = random.randint(1,101)\n counter = 1\n while counter <= 4:\n try:\n number = int(input('[%d]您猜的数是?'%(counter)))\n if 1 <= number <= 100:\n if number == key:\n print('您猜对了!')\n break\n elif number < key:\n print('您猜的数太小了!')\n counter += 1\n else:\n print('您猜的数太大了!')\n counter += 1\n else:\n print('请输入一个[1,100]范围的整数')\n except Exception as e:\n print('请输入一个[1,100]范围的整数',e)\n if counter == 5:\n print('您已经猜了4次,要猜的数是%d' %(key))\n again = input('继续游戏(Y/N)?...')\n\nif __name__ == '__main__':\n main()\n","sub_path":"project10/project/17300680200.py","file_name":"17300680200.py","file_ext":"py","file_size_in_byte":1090,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"594170964","text":"import json\nimport requests\n\nfrom RIOTError import RIOTError\n\nclass pyriot(object):\n\n\tdef __init__(self, api_key):\n\t\tself.api_key = api_key\n\t\tself.base_url = \"https://prod.api.pvp.net\"\n\n\n\tdef request(self, path, payload={}):\n\t\tpayload['api_key'] = self.api_key\n\t\tresponse = requests.get(self.base_url + path, params=payload)\n\t\tif response.status_code == 200:\n\t\t\treturn response.json()\n\t\telif response.status_code == 429:\n\t\t\traise RIOTError(\"Rate Limit Exceeded\")\n\t\telse:\n\t\t\traise RIOTError(response)\n\n\n\tdef champion(self, ftp=None, version=\"1.1\"):\n\t\tpath = \"/api/lol/na/v%s/champion\" % (version)\n\t\tparams = {'freeToPlay': ftp}\n\t\treturn self.request(path, params)\n\n\n\tdef game(self, region, summonerID, version=\"1.2\"):\n\t\tpath = \"/api/lol/%s/v%s/game/by-summoner/%s/recent\" % (region, version, summonerID)\n\t\treturn self.request(path)\n\n\n\tdef league(self, region, summonerID, version=\"2.2\"):\n\t\tif version == \"2.1\":\n\t\t\tpath = \"/api/%s/v%s/league/by-summoner/%s\" % (region, version, summonerID)\n\t\telse:\n\t\t\tpath = \"/api/lol/%s/v%s/league/by-summoner/%s\" % (region, version, summonerID)\n\t\treturn self.request(path)\n\n\n\tdef stats(self, region, summonerID, version=\"1.2\", season=\"SEASON3\"):\n\t\tpath = \"/api/lol/%s/v%s/stats/by-summoner/%s/summary\" % (region, version, summonerID)\n\t\tparams = {'season':season}\n\t\treturn self.request(path, params)\n\n\n\tdef ranked_stats(self, region, summonerID, version=\"1.2\", season=\"SEASON3\"):\n\t\tpath = \"/api/lol/%s/v%s/stats/by-summoner/%s/ranked\" % (region, version, summonerID)\n\t\tparams = {'season':season}\n\t\treturn self.request(path, params)\n\n\n\tdef summoner(self, region, summoner, version=\"1.2\"):\n\t\tif isinstance(summoner, str):\n\t\t\tpath = \"/api/lol/%s/v%s/summoner/by-name/%s\" % (region, version, summoner)\n\t\telif isinstance(summoner, int):\n\t\t\tpath = \"/api/lol/%s/v%s/summoner/%s\" % (region, version, summoner)\n\t\treturn self.request(path)\n\n\n\tdef summoners(self, region, summonerIDs, version=\"1.2\"):\n\t\tif len(summonerIDs) > 40 or not isinstance(summonerIDs, list):\n\t\t\traise RIOTError(\"Max list size is 40.\")\n\t\tpath = \"/api/lol/%s/v%s/summoner/%s/name\" % (region, version, \",\".join(str(summoner) for summoner in summonerIDs))\n\t\treturn self.request(path)\n\n\n\tdef masteries(self, region, summonerID, version=\"1.2\"):\n\t\tpath = \"/api/lol/%s/v%s/summoner/%s/masteries\" % (region, version, summonerID)\n\t\treturn self.request(path)\n\n\n\tdef runes(self, region, summonerID, version=\"1.2\"):\n\t\tpath = \"/api/lol/%s/v%s/summoner/%s/runes\" % (region, version, summonerID)\n\t\treturn self.request(path)\n\n\n\tdef team(self, region, summonerID, version=\"2.2\"):\n\t\t#/api/lol/{region}/v2.2/team/by-summoner/{summonerId}\n\t\tif version == \"2.1\":\n\t\t\tpath = \"/api/%s/v%s/team/by-summoner/%s\" % (region, version, summonerID)\n\t\telse:\n\t\t\tpath = \"/api/lol/%s/v%s/team/by-summoner/%s\" % (region, version, summonerID)\n\t\treturn self.request(path)\n","sub_path":"pyriot.py","file_name":"pyriot.py","file_ext":"py","file_size_in_byte":2828,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"3386809","text":"import pynput\nfrom pynput.keyboard import Key, Listener\n\n\n#-*-coding: utf-8 -*-\n\n\n\ncharCount = 0 # количество введенных символов\nkeys = [] #хранить клавиши, на которые мы нажали\n\n\ndef onKeyPress(key):\n try:\n print(\"Клавиша нажата:\", key) #выводим нажатую клавишу\n except Exception as ex:\n print(\"Произошла ошибка\", ex)\n\n\ndef writeToFile(keys):\n with open(\"log.txt\", \"a\", encoding=\"utf-8\") as file:\n for sym in keys:\n sym = str(sym).replace(\"'\", \"\")\n file.write(sym)\n file.write(\"\\n\") #вставка новой линии\n\n\ndef onKeyRelease(key):\n global keys, charCount\n if key == Key.esc:\n return False\n else:\n if key == Key.enter:\n writeToFile(keys)\n charCount = 0\n keys = []\n elif key == Key.space:\n key = ' '\n writeToFile(keys)\n charCount = 0\n keys = []\n keys.append(key)\n charCount += 1\n\nwith Listener(on_press = onKeyPress, on_release=onKeyRelease) as listener:\n listener.join()\n\n\n\n\n","sub_path":"index.py","file_name":"index.py","file_ext":"py","file_size_in_byte":1181,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"609637872","text":"\n\"\"\"\n{\"a\":3,\"b\":1}\ni_1 = 0\ni_2 = 1\n\nfinal_s = \"a\"\nfinal_s = \"ab\"\nfinal_s = \"aba\"\n\ncurrent_i = i_2 = 1\n\n\"\"\"\nimport heapq\nfrom collections import Counter\n\nclass Solution:\n def reorganizeString(self, S: str) -> str:\n counter = Counter(S)\n h = []\n for element in counter:\n heapq.heappush(h,(-counter[element],element))\n final_s = \"\"\n while len(h)>0:\n el1 = heapq.heappop(h)\n c1 = -el1[0]\n e1 = el1[1]\n if final_s==\"\" or e1!=final_s[-1]:\n final_s+=e1\n if c1>1:\n heapq.heappush(h,(-(c1-1),e1))\n else:\n if len(h)==0:\n return \"\"\n el2 = heapq.heappop(h)\n c2 = -el2[0]\n e2 = el2[1] \n final_s+=e2\n if c2>1:\n heapq.heappush(h,(-(c2-1),e2))\n heapq.heappush(h,(-(c1),e1))\n \n return final_s\n \n \n \n \n \n \n \n \n","sub_path":"leetcode/hard/reorganize-string.py","file_name":"reorganize-string.py","file_ext":"py","file_size_in_byte":1094,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"424124777","text":"import random, sys, time, math, pygame\nfrom User import *\nfrom Power import *\nfrom TripleJumpPower import *\nfrom DoubleJumpPower import *\nfrom UserFactory import *\nimport settings\n\ndef main():\n global FPSCLOCK, BASICFONT, POWER_IMG\n\n pygame.init()\n FPSCLOCK = pygame.time.Clock()\n pygame.display.set_icon(pygame.image.load('gameicon.png'))\n settings.DISPLAYSURF = pygame.display.set_mode((settings.WINWIDTH, settings.WINHEIGHT))\n settings.USER_IMG = pygame.image.load('user.png').convert_alpha()\n pygame.display.set_caption('Dodgem')\n BASICFONT = pygame.font.Font('freesansbold.ttf', 32)\n\n settings.TRIPLE_JUMP_POWER_IMG = pygame.image.load('power.png').convert_alpha()\n settings.DOUBLE_JUMP_POWER_IMG = pygame.image.load('pikachu.png').convert_alpha()\n\n while True:\n runGame()\n\ndef runGame():\n leftDown = False\n rightDown = False\n\n userFactoryObject = UserFactory()\n playerObj = userFactoryObject.createUser(settings.USERTYPE)\n\n settings.DISPLAYSURF.fill(settings.BACKGROUNDCOLOR)\n pickLevel(settings.LEVEL)\n pygame.display.update()\n\n while True: # main game loop\n settings.RECTS_TO_UPDATE = []\n oldX = playerObj.x\n oldY = playerObj.y\n for event in pygame.event.get(): # event handling loop\n if event.type == QUIT:\n terminate()\n\n elif event.type == KEYDOWN:\n if event.key in (K_LEFT, K_a):\n playerObj.moveRight = False\n if rightDown is not True:\n playerObj.moveLeft = True\n leftDown = True\n elif event.key in (K_RIGHT, K_d):\n playerObj.moveLeft = False\n if leftDown is not True:\n playerObj.moveRight = True\n rightDown = True\n elif event.key is K_SPACE:\n if playerObj.jumps != 0:\n playerObj.jumps -= 1\n playerObj.vertSpeed = playerObj.jumpAcceleration\n playerObj.sitting = False\n moveVert = True\n\n elif event.type == KEYUP:\n if event.key in (K_LEFT, K_a):\n if rightDown is True:\n playerObj.moveRight = True\n playerObj.moveLeft = False\n leftDown = False\n elif event.key in (K_RIGHT, K_d):\n if leftDown is True:\n playerObj.moveLeft = True\n playerObj.moveRight = False\n rightDown = False\n\n elif event.key == K_ESCAPE:\n terminate()\n\n #print(\"user.vertSpeed: \", playerObj.vertSpeed)\n\n #moving the user\n playerObj.move()\n\n #checking if the user hit a powerup/down\n powerCollected = playerObj.checkIfCaptured(settings.POWER_LIST)\n if (powerCollected != -1): #we have collided with a powerup\n playerObj = (settings.POWER_OBJ_LIST)[powerCollected].hit(playerObj)\n\n #checking if any powerups need to come back\n currentTime = time.time()\n for power in settings.POWER_OBJ_LIST:\n power.makeActive(currentTime)\n\n #drawing the powers\n for power in settings.POWER_OBJ_LIST:\n power.draw()\n\n #updating the user\n updatePlayerImages(oldX, oldY, playerObj)\n\n pygame.display.update(settings.RECTS_TO_UPDATE)\n FPSCLOCK.tick(settings.FPS)\n\ndef terminate():\n pygame.quit()\n sys.exit()\n\ndef updatePlayerImages(oldX, oldY, playerObj):\n movePlayerRect = pygame.Rect(oldX, oldY, playerObj.size, playerObj.size)\n pygame.draw.rect(settings.DISPLAYSURF, settings.BACKGROUNDCOLOR, movePlayerRect)\n settings.RECTS_TO_UPDATE.append(movePlayerRect)\n playerObj.rect = pygame.Rect(playerObj.x, playerObj.y, playerObj.size, playerObj.size)\n settings.RECTS_TO_UPDATE.append(playerObj.rect)\n settings.DISPLAYSURF.blit(playerObj.surface, playerObj.rect)\n\ndef pickLevel(level):\n if level == 1:\n drawLevelOne()\n if level == 2:\n drawLevelTwo()\n if level == 3:\n drawLevelThree()\n if level == 4:\n drawLevelFour()\n\ndef drawPower(x, y, size, kind):\n POWERLOCATION = (x, y)\n if (kind == settings.EXTRA_JUMP):\n powerObj = TripleJumpPower(POWERLOCATION[0], POWERLOCATION[1], size, kind)\n if (kind == settings.TWO_JUMPS):\n powerObj = DoubleJumpPower(POWERLOCATION[0], POWERLOCATION[1], size, kind)\n settings.POWER_OBJ_LIST.append(powerObj)\n settings.POWER_LIST.append(powerObj.rect)\n\n\ndef drawRectangle(x, y, width, height, color):\n OBSTACLE = pygame.Rect(x, y, width, height)\n settings.OBSTACLELIST.append(OBSTACLE)\n pygame.draw.rect(settings.DISPLAYSURF, color, OBSTACLE)\n\ndef drawLevelOne():\n FLOORHEIGHT = 100\n drawRectangle(0, settings.WINHEIGHT - FLOORHEIGHT, settings.WINWIDTH, FLOORHEIGHT, settings.BLACK)\n drawRectangle(200, settings.WINHEIGHT-FLOORHEIGHT-250, 100, 250, settings.OBSTACLECOLOR)\n\ndef drawLevelTwo():\n FLOORHEIGHT = 40\n drawRectangle(0, settings.WINHEIGHT - FLOORHEIGHT, settings.WINWIDTH, FLOORHEIGHT + 40, settings.BLACK)\n drawRectangle(100, settings.WINHEIGHT-FLOORHEIGHT-100, 100, 100, settings.OBSTACLECOLOR)\n drawRectangle(250, settings.WINHEIGHT-FLOORHEIGHT-200, 100, 200, settings.OBSTACLECOLOR)\n drawRectangle(400, settings.WINHEIGHT-FLOORHEIGHT-200, 100, 100, settings.OBSTACLECOLOR)\n\ndef drawLevelThree():\n FLOORHEIGHT = 40\n drawRectangle(0, settings.WINHEIGHT - FLOORHEIGHT, settings.WINWIDTH, FLOORHEIGHT + 40, settings.BLACK)\n drawRectangle(100, settings.WINHEIGHT-FLOORHEIGHT-100, 100, 100, settings.OBSTACLECOLOR)\n drawRectangle(250, settings.WINHEIGHT-FLOORHEIGHT-200, 100, 200, settings.OBSTACLECOLOR)\n drawRectangle(400, settings.WINHEIGHT-FLOORHEIGHT-200, 100, 100, settings.OBSTACLECOLOR)\n drawPower(580, 110, 32, settings.EXTRA_JUMP)\n\ndef drawLevelFour():\n FLOORHEIGHT = 40\n drawRectangle(0, settings.WINHEIGHT - FLOORHEIGHT, settings.WINWIDTH, FLOORHEIGHT + 40, settings.BLACK)\n drawRectangle(100, settings.WINHEIGHT-FLOORHEIGHT-100, 100, 100, settings.OBSTACLECOLOR)\n drawPower(580, 120, 32, settings.EXTRA_JUMP)\n drawPower(180, 140, 32, settings.TWO_JUMPS)\n\nif __name__ == '__main__':\n main()\n\n#check the rect and surf\n#only make them GLOBALLY.\n#dont make them every FRAME","sub_path":"balljump.py","file_name":"balljump.py","file_ext":"py","file_size_in_byte":6448,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"341962344","text":"import os\nos.environ['TF_CPP_MIN_LOG_LEVEL'] = '2'\nimport tensorflow as tf\nconfig = tf.ConfigProto()\nconfig.gpu_options.allow_growth = True\nsession = tf.Session(config=config)\n\nfrom lib.feature import *\nfrom sklearn_crfsuite import metrics\nfrom sklearn.model_selection import train_test_split\nimport csv\nimport keras\nfrom keras.models import Sequential, Model\nfrom keras.layers import Dense, Dropout, Input, concatenate, RepeatVector\nfrom keras.callbacks import EarlyStopping\nfrom keras.optimizers import Adam\nfrom keras.layers import CuDNNLSTM, TimeDistributed, Bidirectional\nfrom keras.utils import plot_model\n\ndef report(pred, truth):\n _pred = VecContext.y2lab(pred)\n _test = VecContext.y2lab(truth)\n print(metrics.flat_classification_report(\n _test, _pred, labels=('I', 'E'), digits=4\n ))\n label = 'E'\n P = metrics.flat_precision_score(_test, _pred, pos_label=label)\n R = metrics.flat_recall_score(_test, _pred, pos_label=label)\n f1 = metrics.flat_f1_score(_test, _pred, pos_label=label)\n return {'P':P, 'R':R, 'f1':f1}\n\ndef context_data(path, k=10, size=None, seq=False, vec_size=50, w2v_text='./data/w2v.txt', train_size=0.7):\n data = VecContext(path, k=k, vec_size=vec_size, w2v_text=w2v_text)\n if seq == True:\n seq_y = []\n y = [0]*k + list(data.Y) + [0]*k\n for i in range(k, len(data.Y)+k):\n seq_y.append( [ [ele] for ele in y[i-k:i+k+1] ] )\n data.Y = seq_y\n\n x_train, x_test, y_train, y_test = train_test_split(\n data.X, data.Y, test_size=1.0-train_size, shuffle=False\n )\n\n if size != None:\n size_m = int(len(x_train)*size)\n x_train = x_train[:size_m]\n y_train = y_train[:size_m]\n x_train = np.array(x_train)\n x_test = np.array(x_test)\n y_train = np.array(y_train)\n y_test = np.array(y_test)\n return x_train, x_test, y_train, y_test\n\ndef aux_data(path):\n def lab2val(l):\n if l[0] == 'O':\n return 1\n elif l[0] == 'B':\n return 2\n elif l[0] == 'I':\n return 3\n elif l[0] == 'E':\n return 4\n return 0\n office = Label(path, lab_name='office', lab_file='./ref/tang_name/tangOffice.clliu.txt')\n nianhao = Label(path, lab_name='nianhao', lab_file='./ref/tang_name/tangReignperiods.clliu.txt')\n address = Label(path, lab_name='address', lab_file='./ref/tang_name/tangAddresses.clliu.txt')\n aux_data = Tdiff(path, uniform=False) + MutualInfo(path, uniform=False) + office + nianhao + address\n aux_data.union()\n aux_data.X = [ [ele['t-diff'], ele['mi-info'], lab2val(ele['office']), lab2val(ele['address']), lab2val(ele['nianhao'])] for ele in aux_data.X ]\n x_train, x_test, y_train, y_test = train_test_split(\n aux_data.X, aux_data.Y, test_size=0.3, shuffle=False\n )\n return x_train, x_test\n\ndef aux_adv_data(path):\n def lab2val(l):\n if l[0] == 'O':\n return [0,0,0]\n elif l[0] == 'B':\n return [1,0,0]\n elif l[0] == 'I':\n return [0,1,0]\n elif l[0] == 'E':\n return [0,0,1]\n return [0,0,0]\n office = Label(path, lab_name='office', lab_file='./ref/tang_name/tangOffice.clliu.txt')\n nianhao = Label(path, lab_name='nianhao', lab_file='./ref/tang_name/tangReignperiods.clliu.txt')\n address = Label(path, lab_name='address', lab_file='./ref/tang_name/tangAddresses.clliu.txt')\n tdiff = Tdiff(path, uniform=False, noise=True)\n pmi = MutualInfo(path, uniform=False, noise=True)\n aux_data = office + nianhao + address + pmi + tdiff\n aux_data.union()\n aux_data.X = [ [ele['t-diff'], ele['mi-info'] ] + lab2val(ele['office']) + lab2val(ele['address']) + lab2val(ele['nianhao']) for ele in aux_data.X ]\n x_train, x_test, y_train, y_test = train_test_split(\n aux_data.X, aux_data.Y, test_size=0.3, shuffle=False\n )\n return x_train, x_test\n\ndef basic_model(data, stack=5, seq=False):\n inputs = Input(shape=(len(data[0]), len(data[0][0])))\n if stack == 1:\n x = Bidirectional(CuDNNLSTM(50))(inputs)\n main_output = Dense(1, activation='sigmoid')(x)\n elif stack == 2:\n x = Bidirectional(CuDNNLSTM(50, return_sequences=True))(inputs)\n x = Bidirectional(CuDNNLSTM(50))(x)\n main_output = Dense(1, activation='sigmoid')(x)\n else:\n x = Bidirectional(CuDNNLSTM(50, return_sequences=True))(inputs)\n for _ in range(stack-2):\n x = Bidirectional(CuDNNLSTM(50, return_sequences=True))(x)\n if seq == True:\n x = Bidirectional(CuDNNLSTM(50, return_sequences=True))(x)\n main_output = TimeDistributed(Dense(1, activation='sigmoid'))(x)\n else:\n x = Bidirectional(CuDNNLSTM(50))(x)\n main_output = Dense(1, activation='sigmoid')(x)\n\n model = Model(inputs=[inputs], outputs=main_output)\n model.compile(loss='binary_crossentropy',\n optimizer='adam',\n metrics=['accuracy'])\n return model\n\ndef encoder_model(data, n_encoder, n_decoder, n_lstm):\n inputs = Input(shape=(len(data[0]), len(data[0][0])))\n if n_lstm != 0:\n # encoder\n x = Bidirectional(CuDNNLSTM(50, return_sequences=True))(inputs)\n for _ in range(n_encoder-2):\n x = Bidirectional(CuDNNLSTM(50, return_sequences=True))(x)\n x = Bidirectional(CuDNNLSTM(50, return_sequences=False))(x)\n coder_output = RepeatVector(len(data[0]))(x)\n\n # stack lstm\n x = Bidirectional(CuDNNLSTM(50, return_sequences=True))(inputs)\n for _ in range(n_lstm-2):\n x = Bidirectional(CuDNNLSTM(50, return_sequences=True))(x)\n lstm_output = Bidirectional(CuDNNLSTM(50, return_sequences=True))(x)\n x = concatenate([lstm_output, coder_output])\n\n # decoder\n x = Bidirectional(CuDNNLSTM(50, return_sequences=True))(x)\n for _ in range(n_decoder-1):\n x = Bidirectional(CuDNNLSTM(50, return_sequences=True))(x)\n else:\n # encoder\n x = Bidirectional(CuDNNLSTM(50, return_sequences=True))(inputs)\n for _ in range(n_encoder-2):\n x = Bidirectional(CuDNNLSTM(50, return_sequences=True))(x)\n x = Bidirectional(CuDNNLSTM(50, return_sequences=False))(x)\n x = RepeatVector(len(data[0]))(x)\n\n # decoder\n x = Bidirectional(CuDNNLSTM(50, return_sequences=True))(x)\n for _ in range(n_decoder-1):\n x = Bidirectional(CuDNNLSTM(50, return_sequences=True))(x)\n main_output = TimeDistributed(Dense(1, activation='sigmoid'))(x)\n\n model = Model(inputs=[inputs], outputs=main_output)\n model.compile(loss='binary_crossentropy',\n optimizer='adam',\n metrics=['accuracy'])\n return model\n","sub_path":"lib/lstmlib.py","file_name":"lstmlib.py","file_ext":"py","file_size_in_byte":6739,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"360190105","text":"#!/usr/bin/env python\nimport os\nimport sys\nfrom urlparse import urljoin\n\nimport requests\n\nBASE_URL = \"http://raw.githubusercontent.com/\"\n\ntry:\n arg = sys.argv[1]\n rep_count = int(sys.argv[2])\nexcept IndexError:\n rep_count = None\n print(\"No repeation count given, default is 1000.\")\n\n# swift smoke tests and xss\n\ntry:\n if arg == 'create_containers':\n # create a set of containers\n for i in range(rep_count or 1000):\n os.system('ccurl.py -r container_{} -X PUT'.format(str(i)))\n if arg == 'delete_containers':\n # delete a list of containers\n for i in range(100, 1000):\n os.system('ccurl.py -r object_{} -X DELETE'.format(str(i)))\n if arg == 'xssp_containers':\n # xss container names with polygot xss strings\n xss_list = requests.get(\n urljoin\n (BASE_URL,\n \"/fuzzdb-project/fuzzdb/master/attack/xss/XSSPolyglot.txt\")\n ).content.split(\"\\n\")\n for i in xss_list:\n os.system('ccurl.py -r {{0}} -X PUT'.format(str(i)))\n\n\nexcept KeyboardInterrupt:\n print(\"Exiting..>>\")\n","sub_path":"swift/attacks.py","file_name":"attacks.py","file_ext":"py","file_size_in_byte":1109,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"456012338","text":"# -*- coding: utf-8 -*-\nfrom django.db import models\nfrom datetime import datetime\nfrom django.contrib.auth.models import User\n\n\nclass pmo_review_task(models.Model):\n project_phase_to_review = models.ForeignKey('project_phase')\n creation_date = models.DateTimeField(default=datetime.now())\n is_closed = models.BooleanField(default=False,null=False)\n review_date = models.DateTimeField(null=True,blank=True)\n review_comment = models.TextField(null=True, blank=True)\n reviewed_by = models.ForeignKey(User, null=True, blank=True)\n is_review_result_ok = models.NullBooleanField(null=True)\n scheduled_sc = models.ForeignKey('sc_event', null=True, blank=True)\n review_goal = models.ForeignKey('phases')\n\n def __unicode__(self):\n return self.project_phase_to_review.project.name\n\n def save(self, *args, **kwargs):\n from core.models import sc_review_task\n if self.is_closed and self.is_review_result_ok and self.scheduled_sc is None:\n raise models.ImproperlyConfigured\n\n if self.is_closed and self.is_review_result_ok is False:\n self.project_phase_to_review.is_finished = False\n self.project_phase_to_review.is_approved_for_sc = False\n self.project_phase_to_review.save()\n\n if self.is_closed and self.is_review_result_ok and self.scheduled_sc:\n if self.scheduled_sc.has_taken_place is True:\n raise Exception(\"Cannot assign a task to a SC event that has taken place\")\n #update reviewed project to indicate it has been saved\n self.project_phase_to_review.is_approved_for_sc = True\n self.project_phase_to_review.save()\n #create sc_review_talk\n #a fucking model hack - should be fixed\n newtask = sc_review_task()\n newtask.project_phase_to_review = self.project_phase_to_review\n newtask.sc_event_to_review_at = self.scheduled_sc\n newtask.review_goal = self.review_goal\n newtask.is_closed = False\n newtask.save()\n #call real save operation\n super(pmo_review_task,self).save(*args, **kwargs)\n\n class Meta:\n app_label = 'core'\n verbose_name = u'Задача (проверка проекта PMO)'\n verbose_name_plural = u'Задачи (проверка проекта PMO)'\n\n","sub_path":"core/models/pmo_review_task.py","file_name":"pmo_review_task.py","file_ext":"py","file_size_in_byte":2360,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"616531323","text":"import torch\nimport torchaudio\nimport numpy as np\n\nclass STFTModule():\n def __init__(self, stft_params, device):\n self.device = device\n self.dtype= torch.float32\n self.n_fft = stft_params['n_fft']\n self.hop_length = stft_params['hop_length']\n self.win_length = stft_params['win_length']\n self.window = torch.hann_window(self.win_length).to(self.dtype).to(self.device)\n self.freq_num = self._cal_freq_num()\n self.pad = None\n self.pad_len = None\n self.sample_len = None\n \n def _cal_freq_num(self):\n return (np.floor(self.n_fft / 2) + 1).astype(np.int32)\n \n def stft(self, x, pad=None):\n center = True\n if pad:\n self.pad = True\n x = self._stft_zero_pad(x)\n center = False\n \n return torch.stft(x, \n n_fft=self.n_fft,\n hop_length=self.hop_length, \n win_length=self.win_length,\n center=center, \n window=self.window)\n \n def _stft_zero_pad(self, x):\n batch_size, self.sample_len = x.shape\n frame_num = self._cal_frame_num(self.sample_len)\n pad_x_len = self.win_length + ((frame_num - 1) * self.hop_length)\n self.pad_len = pad_x_len - self.sample_len\n buff = torch.zeros((batch_size, pad_x_len), dtype=self.dtype, device=self.device.type)\n buff[:, :self.sample_len] = x\n return buff\n \n def _cal_frame_num(self, sig_len):\n return np.ceil((sig_len - self.win_length + self.hop_length) / self.hop_length).astype(np.int32)\n \n def _istft_zero_pad(self, x):\n batch_size, f_num, frame_num, channel = x.shape\n pad_size = self.win_length // self.hop_length\n half_pad = pad_size // 2\n pad_x = torch.zeros((batch_size, f_num, frame_num+pad_size, channel), dtype=self.dtype, device=self.device.type)\n pad_x[:,:,half_pad:-half_pad,:] = x[:,:,:,:]\n return pad_x\n \n def _squeeze_istft_pad(self, x):\n half_nfft = self.n_fft // 2\n return x[:, half_nfft:-half_nfft]\n \n def istft(self, x, siglen=None):\n center = True\n if self.pad:\n x = self._istft_zero_pad(x)\n center = False\n \n wave = torchaudio.functional.istft(x, \n n_fft=self.n_fft,\n win_length=self.win_length, \n hop_length=self.hop_length,\n center=center,\n length=siglen,\n window=self.window,)\n return wave\n \n def to_normalize_mag(self, x):\n flooring = 1e-4\n eps = 1e-8\n max_val = x.reshape(-1).max()\n logx = torch.log10(torch.clamp(x, min=flooring*max_val))\n norm_logx = (logx - logx.mean(-1, keepdim=True))/(logx.std(-1, keepdim=True)+eps)\n return norm_logx\n \n \n ","sub_path":"utils/stft_module.py","file_name":"stft_module.py","file_ext":"py","file_size_in_byte":3128,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"632344570","text":"import nltk\nimport re\nfrom operator import itemgetter\nfrom lxml import html\nimport requests\nfrom numpy import random\nimport bs4 #beautiful soup\n\ndef get_body(url):\n \"\"\" Gets all the text from a webpage body that is not html tags\n Arg(s):\n url: the webpage whose text you want\n Returns:\n all the text from a webpage's body in a single, all-lowercase string\n\n \"\"\"\n return bs4.BeautifulSoup(requests.get(url).text, 'lxml').body.get_text().lower()\n\ndef get_chunks(body_text, chunk_counts = {}):\n \"\"\" Gets list of words with parts of speech.\n\n Splits a string into a list of words, stripped for specific sequences of \n whitespace, which is then turned into a list of tuples containing a word\n and part-of-speech universal tag along with word_counts.\n\n Arg(s):\n url, the webpage whose text you want\n Returns:\n all the text from a webpage in a single, all-lowercase string\n \"\"\"\n text = list(filter(lambda x: x != '', re.split(r'\\W\\s|\\W\\W+|[^\\w+\\S\\w+]', body_text)))\n word_pairs = nltk.pos_tag(text, tagset='universal')\n for pair in word_pairs:\n if pair in list(word_counts.keys()):\n word_counts[pair] += 1\n else:\n word_counts[pair] = 1\n return word_counts\n\ndef get_words(body_text, word_counts = {}):\n text = list(filter(lambda x: x != '', re.split(r'\\W\\s|\\W\\W+|[^\\w+\\S\\w+]', body_text)))\n word_pairs = nltk.pos_tag(text, tagset='universal')\n for pair in word_pairs:\n if pair in list(word_counts.keys()):\n word_counts[pair] += 1\n else:\n word_counts[pair] = 1\n return word_counts\n\ndef get_pos(word_counts, pos = {}):\n for word in list(word_counts.keys()):\n if word[1] not in list(pos.keys()):\n pos[word[1]] = []\n pos[word[1]].append((word[0], word_counts[word]))\n for part in pos.keys():\n pos[part] = sorted(pos[part], key=itemgetter(1))\n return pos\n","sub_path":"code/parse_HTML.py","file_name":"parse_HTML.py","file_ext":"py","file_size_in_byte":1867,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"636222787","text":"from distutils.dir_util import copy_tree\nimport pyexcel as p\nfrom emailBuildClass import emailBuilder\n\nclass KFLoyalty(emailBuilder):\n def __init__(self, **kwargs):\n self.projectNamePrefix = \"KF_\"\n self.folderPath = \"kingsfoodmarkets\"\n self.ymlTemplate = \"KF/KF_Template_Loyalty_Email_053117.yml\"\n self.erbTemplate = \"KF/KF_Template_Loyalty_Email_053117.erb\"\n\n self.sheetName = 0\n self.altTextColumn = 2\n self.linkColumn = 5\n super(KFLoyalty, self).__init__(**kwargs)\n\n def update(self):\n self.updateHeader()\n\n def updateHeader(self):\n # find and update title\n nameFound = False\n titleFound = False\n\n for irow in self.sheet.rows():\n for x in range(0, len(irow) -1):\n if irow[x].strip() == \"Task:\":\n self.fileContents = self.fileContents.replace(\"__EMAIL_NAME__\", irow[x+1])\n nameFound = True\n elif irow[x].strip() == \"Subject:\":\n self.fileContents = self.fileContents.replace(\"__TITLE__\", irow[x+1])\n titleFound = True\n if nameFound and titleFound:\n break\n if nameFound and titleFound:\n break\n\n def loadData(self):\n super(KFLoyalty, self).loadData()\n self.layoutSheetName = \"Layout\"\n self.layoutsheet = self.workbook.sheet_by_name(self.layoutSheetName)\n\n def generateImage(self, images):\n articleType = \"table\"\n\n table = []\n\n for image in images:\n imgData = ['']*6\n\n row = self.findImageRow(image)\n\n path, imageName = self.getImageRelativePathAndName(image)\n\n imgData[0] = \"image\"\n imgData[1] = str(path + imageName).encode('ascii','ignore')\n imgData[2] = self.getImageLink(row[self.linkColumn]).encode('ascii','ignore')\n imgData[5] = self.encodeText(row[self.altTextColumn].encode('ascii','ignore'))\n\n imgData = self.addAdditionalFields(image, imgData)\n\n table.append(imgData)\n\n self.contents.insert(self.insertRow(), \"- type: '\" + articleType + \"'\" + self.newLine)\n self.contents.insert(self.insertRow(), \" data: \" + str(table) + \"\" + self.newLine)\n self.contents.insert(self.insertRow(), self.newLine)\n\n def generateYmlContent(self):\n for irow in self.layoutsheet.rows():\n while '' in irow:\n irow.remove('')\n self.generateImage(irow)\n\n\nclass KFGeneric(KFLoyalty):\n def __init__(self, **kwargs):\n super(KFGeneric, self).__init__(**kwargs)\n self.erbTemplate = \"KF/KF_Template_Generic_Email_061617.erb\"","sub_path":"templates/KF/KF.py","file_name":"KF.py","file_ext":"py","file_size_in_byte":2712,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"463152089","text":"from __future__ import division\n\nfrom Board import *\n\nfrom random import choice\nfrom math import log, sqrt\nimport datetime\nimport copy\n\nclass MonteCarlo(object):\n\n def __init__(self, **kwargs):\n seconds = kwargs.get('time', 10)\n self.calculation_time = datetime.timedelta(seconds=seconds)\n\n self.max_moves = kwargs.get('max_moves', 10)\n\n self.wins = {}\n self.plays = {}\n\n self.C = kwargs.get('C', 1.4)\n\n def update(self, state):\n pass\n\n def get_play(self, state, player):\n self.max_depth = 0\n legal = state.getPossiblePositionList(player)\n\n if not legal:\n return\n elif len(legal) == 1:\n return (None, legal[0])\n\n games = 0\n begin = datetime.datetime.utcnow()\n while datetime.datetime.utcnow() - begin < self.calculation_time:\n self.run_simulation(state, player)\n games += 1\n\n moves_states = []\n for p in legal:\n copy_state = copy.deepcopy(state)\n copy_state.setStone(player, p)\n moves_states.append((p, copy_state))\n\n #print games, datetime.datetime.utcnow() - begin\n\n percent_wins, move = max(\n (self.wins.get((player, S.getBoardStr()), 0) /\n self.plays.get((player, S.getBoardStr()), 1),\n p)\n for p, S in moves_states\n )\n\n## for x in sorted(\n## ((100 * self.wins.get((player, S.getBoardStr()), 0) /\n## self.plays.get((player, S.getBoardStr()), 1),\n## self.wins.get((player, S.getBoardStr()), 0),\n## self.plays.get((player, S.getBoardStr()), 0), p)\n## for p, S in moves_states),\n## reverse=True\n## ):\n## print \"{3}: {0: .2f}% ({1} / {2})\".format(*x)\n##\n## print \"Maximum depth searched:\", self.max_depth\n\n return (percent_wins, move)\n\n def run_simulation(self, state, player):\n visited_states = set()\n\n expand = True\n for t in xrange(1, self.max_moves + 1):\n legal = state.getPossiblePositionList(player)\n\n if legal == []:\n break\n \n moves_states = []\n for p in legal:\n copy_state = copy.deepcopy(state)\n copy_state.setStone(player, p)\n moves_states.append((p, copy_state))\n\n if all(self.plays.get((player, S.getBoardStr())) for p, S in moves_states):\n log_total = log(\n sum(self.plays[(player, S.getBoardStr())] for p, S in moves_states))\n value, move, state = max(\n ((self.wins[(player, S.getBoardStr())] / self.plays[(player, S.getBoardStr())]) +\n self.C * sqrt(log_total / self.plays[(player, S.getBoardStr())]), p, S)\n for p, S in moves_states\n )\n else:\n move, state = choice(moves_states)\n\n if expand and (player, state.getBoardStr()) not in self.plays:\n expand = False\n self.plays[(player, state.getBoardStr())] = 0\n self.wins[(player, state.getBoardStr())] = 0\n if t > self.max_depth:\n self.max_depth = t\n\n visited_states.add((player, state))\n\n player = state.getNextPlayer()\n winner = state.isWin()\n if winner:\n break\n \n for player, state in visited_states:\n if (player, state.getBoardStr()) not in self.plays:\n continue\n self.plays[(player, state.getBoardStr())] += 1\n if player == winner or winner == DRAW \\\n or ((state.whiteCount > state.blackCount) and player == WHITE) \\\n or ((state.whiteCount < state.blackCount) and player == BLACK):\n self.wins[(player, state.getBoardStr())] += 1\n \n\n\n","sub_path":"MonteCarlo.py","file_name":"MonteCarlo.py","file_ext":"py","file_size_in_byte":3943,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"650273087","text":"from functools import partial\nfrom itertools import accumulate\n\nimport haiku as hk\nimport jax.numpy as jnp\nfrom jax import ops\n\nfrom ..utils import flatten\nfrom .graph import (\n Graph,\n GraphNodes,\n GraphUpdate,\n MolecularGraphEdgeBuilder,\n difference_callback,\n)\nfrom .utils import NodeEdgeMapping\n\n\nclass ElectronGNNLayer(hk.Module):\n r\"\"\"\n The message passing layer of :class:`ElectronGNN`.\n\n Derived from :class:`~deepqmc.gnn.gnn.MessagePassingLayer`.\n\n Args:\n residual (bool): whether a residual connection is used when updating\n the electron embeddings.\n convolution (bool): if :data:`True` the messages are generated via graph\n covolutions, else messages are generated from edge featues only.\n deep_features (bool): if :data:`true` edge features are updated through\n an MLP (:data:`u`), else initial edge features are reused.\n update_features (list[str]): which features to collect for the update\n of the electron embeddings.\n Possible values:\n\n - ``'residual'``: electron embedding from the previous interaction layer\n - ``'edge_ne'``: sum over messages from nuclei\n - ``'edge_same'``: sum over messages from electrons with same spin\n - ``'edge_anti'``: sum over messages from electrons with opposite spin\n - ``'edge_up'``: sum over messages from spin up electrons\n - ``'edge_down'``: sum over messages from spin down electrons\n - ``'edge_ee'``: sum of over messages from all electrons\n - ``'node_up'``: sum over embeddings from spin-up electrons\n - ``'node_down'``: sum over embeddings from spin-down electrons\n\n update_rule (str): how to combine features for the update of the\n electron embeddings.\n Possible values:\n\n - ``'concatenate'``: run concatenated features through MLP\n - ``'featurewise'``: apply different MLP to each feature channel and sum\n - ``'featurewise_shared'``: apply the same MLP across feature channels\n - ``'sum'``: sum features before sending through an MLP\n\n note that `sum` and `featurewise_shared` imply features of same size\n\n subnet_factory (Callable): A function that constructs the subnetworks of\n the GNN layer.\n subnet_factory_by_lbl (dict): optional, a dictionary of functions that construct\n subnetworks of the GNN layer. If both this and :data:`subnet_factory` is\n specified, the specified values of :data:`subnet_factory_by_lbl` will take\n precedence. If some keys are missing, the default value of\n :data:`subnet_factory` will be used in their place. Possible keys are:\n (:data:`w`, :data:`h`, :data:`g` or :data:`u`).\n \"\"\"\n\n def __init__(\n self,\n n_interactions,\n ilayer,\n n_nuc,\n n_up,\n n_down,\n embedding_dim,\n edge_types,\n node_data,\n edge_feat_dim,\n two_particle_stream_dim,\n *,\n residual,\n convolution,\n deep_features,\n update_features,\n update_rule,\n subnet_factory=None,\n subnet_factory_by_lbl=None,\n ):\n super().__init__()\n self.n_nuc, self.n_up, self.n_down = n_nuc, n_up, n_down\n self.last_layer = ilayer == n_interactions - 1\n self.edge_types = tuple(\n typ for typ in edge_types if not self.last_layer or typ not in {'nn', 'en'}\n )\n self.mapping = NodeEdgeMapping(self.edge_types, node_data=node_data)\n assert update_rule in [\n 'concatenate',\n 'featurewise',\n 'featurewise_shared',\n 'sum',\n ]\n assert all(\n uf\n in [\n 'residual',\n 'edge_ne',\n 'edge_same',\n 'edge_anti',\n 'edge_up',\n 'edge_down',\n 'edge_ee',\n 'node_up',\n 'node_down',\n ]\n for uf in update_features\n )\n assert (\n update_rule not in ['sum', 'featurewise_shared']\n or embedding_dim == two_particle_stream_dim\n )\n assert deep_features in [False, 'shared', 'separate']\n self.deep_features = deep_features\n self.update_features = update_features\n self.update_rule = update_rule\n self.convolution = convolution\n subnet_factory_by_lbl = subnet_factory_by_lbl or {}\n for lbl in ['w', 'h', 'g', 'u']:\n subnet_factory_by_lbl.setdefault(lbl, subnet_factory)\n if deep_features:\n self.u = (\n subnet_factory_by_lbl['u'](two_particle_stream_dim, name='u')\n if deep_features == 'shared'\n else {\n typ: subnet_factory_by_lbl['u'](\n two_particle_stream_dim,\n name=f'u{typ}',\n )\n for typ in self.edge_types\n }\n )\n if self.convolution:\n self.w = {\n typ: subnet_factory_by_lbl['w'](\n two_particle_stream_dim,\n name=f'w_{typ}',\n )\n for typ in self.edge_types\n }\n self.h = {\n typ: subnet_factory_by_lbl['h'](\n two_particle_stream_dim,\n name=f'h_{typ}',\n )\n for typ in self.edge_types\n }\n self.g = (\n subnet_factory_by_lbl['g'](\n embedding_dim,\n name='g',\n )\n if not self.update_rule == 'featurewise'\n else {\n uf: subnet_factory_by_lbl['g'](\n embedding_dim,\n name=f'g_{uf}',\n )\n for uf in (self.update_features)\n }\n )\n self.residual = residual\n\n def get_update_edges_fn(self):\n def update_edges(edges):\n if self.deep_features:\n features = {typ: edge.features for typ, edge in edges.items()}\n if self.deep_features == 'shared':\n # combine features along leading dim, apply MLP and split\n # into channels again to please kfac\n keys, feats = zip(*features.items())\n split_idxs = list(accumulate([len(f) for f in feats]))\n feats = jnp.split(self.u(jnp.concatenate(feats)), split_idxs)\n updated_features = dict(zip(keys, feats))\n elif self.deep_features == 'separate':\n updated_features = {\n typ: self.u[typ](edge.features) for typ, edge in edges.items()\n }\n\n if self.residual:\n updated_features = self.residual(features, updated_features)\n return {\n typ: edges[typ]._replace(features=updated_features[typ])\n for typ in edges.keys()\n }\n else:\n return edges\n\n return update_edges\n\n def get_aggregate_edges_for_nodes_fn(self):\n def aggregate_edges_for_nodes(nodes, edges):\n if self.convolution:\n we = {typ: self.w[typ](edge.features) for typ, edge in edges.items()}\n hx = {\n typ: (self.h[typ](self.mapping.sender_data_of(typ, nodes)))[\n edges[typ].senders\n ]\n for typ in self.edge_types\n }\n message = {typ: we[typ] * hx[typ] for typ in self.edge_types}\n else:\n message = {typ: edge.features for typ, edge in edges.items()}\n\n z = {\n typ: ops.segment_sum(\n data=message[typ],\n segment_ids=edges[typ].receivers,\n num_segments=self.mapping.receiver_data_of(typ, 'n_nodes'),\n )\n for typ in self.edge_types\n }\n return z\n\n return aggregate_edges_for_nodes\n\n def get_update_nodes_fn(self):\n def update_nodes(nodes, z):\n n_up, n_down = self.n_up, self.n_down\n FEATURE_MAPPING = {\n 'residual': lambda: nodes.electrons,\n 'node_up': lambda: (\n nodes.electrons[: self.n_up]\n .mean(axis=0, keepdims=True)\n .repeat(self.n_up + self.n_down, axis=0)\n ),\n 'node_down': lambda: (\n nodes.electrons[self.n_up :]\n .mean(axis=0, keepdims=True)\n .repeat(self.n_up + self.n_down, axis=0)\n ),\n 'edge_same': lambda: z['same'] / jnp.clip(\n jnp.array(n_up * [[n_up - 1]] + n_down * [[n_down - 1]]), 1\n ),\n 'edge_anti': lambda: z['anti'] / jnp.clip(\n jnp.array(n_up * [[n_down]] + n_down * [[n_up]]), 1\n ),\n 'edge_up': lambda: z['up'] / self.n_up,\n 'edge_down': lambda: z['down'] / self.n_down,\n 'edge_ee': lambda: (z['same'] + z['anti']) / (\n self.n_up + self.n_down - 1\n ),\n 'edge_ne': lambda: z['ne'] / self.n_nuc,\n }\n f = {uf: FEATURE_MAPPING[uf]() for uf in self.update_features}\n if self.update_rule == 'concatenate':\n updated = self.g(\n jnp.concatenate([f[uf] for uf in self.update_features], axis=-1)\n )\n elif self.update_rule == 'featurewise':\n updated = sum(self.g[uf](f[uf]) for uf in self.update_features)\n elif self.update_rule == 'sum':\n updated = self.g(sum(f.values()))\n elif self.update_rule == 'featurewise_shared':\n updated = jnp.sum(self.g(jnp.stack(list(f.values()))), axis=0)\n if self.residual:\n updated = self.residual(nodes.electrons, updated)\n nodes = GraphNodes(nodes.nuclei, updated)\n\n return nodes\n\n return update_nodes\n\n def __call__(self, graph):\n r\"\"\"\n Execute the message passing layer.\n\n Args:\n graph (:class:`Graph`)\n\n Returns:\n :class:`Graph`: updated graph\n \"\"\"\n update_graph = GraphUpdate(\n update_nodes_fn=self.get_update_nodes_fn(),\n update_edges_fn=None if self.last_layer else self.get_update_edges_fn(),\n aggregate_edges_for_nodes_fn=self.get_aggregate_edges_for_nodes_fn(),\n )\n return update_graph(graph)\n\n\nclass ElectronGNN(hk.Module):\n r\"\"\"\n A neural network acting on graphs defined by electrons and nuclei.\n\n Derived from :class:`~deepqmc.gnn.gnn.GraphNeuralNetwork`.\n\n Args:\n mol (:class:`~deepqmc.Molecule`): the molecule on which the graph is defined.\n embedding_dim (int): the length of the electron embedding vectors.\n n_interactions (int): number of message passing interactions.\n positional_electron_embeddings(bool): whether to initialize the electron\n embbedings with the concatenated edge features.\n edge_features: a function or a :data:`dict` of functions for each edge\n type, embedding the interparticle differences.\n edge_types: the types of edges to consider in the molecular graph. It should\n be a sequence of unique :data:`str`s from the follwing options:\n - ``'nn'``: nucleus-nucleus edges\n - ``'ne'``: nucleus-electron edges\n - ``'en'``: electron-nucleus edges\n - ``'same'``: electron-electron edges between electrons of the same spin\n - ``'anti'``: electron-electron edges between electrons of opposite spins\n two_particle_stream_dim (int): the feature dimension of the two particle\n streams. Only active if :data:`deep_features` are used.\n atom_type_embeedings (bool): if :data:`True`, use the same initial embeddings\n for all atoms with the same atomic number. If :data:`False` use a different\n initial embedding for all atoms.\n layer_factory (Callable): a callable that generates a layer of the GNN.\n ghost_coords: optional, specifies the coordinates of one or more ghost atoms,\n useful for breaking spatial symmetries of the nuclear geometry.\n \"\"\"\n\n def __init__(\n self,\n mol,\n embedding_dim,\n *,\n n_interactions,\n positional_electron_embeddings,\n edge_features,\n edge_types,\n two_particle_stream_dim,\n atom_type_embeddings,\n layer_factory,\n ghost_coords=None,\n ):\n super().__init__()\n n_nuc, n_up, n_down = mol.n_particles\n edge_feat_dim = {typ: len(edge_features[typ]) for typ in edge_types}\n n_atom_types = mol.n_atom_types\n charges = mol.charges\n self.ghost_coords = None\n if ghost_coords is not None:\n charges = jnp.concatenate([charges, jnp.zeros(len(ghost_coords))])\n n_nuc += len(ghost_coords)\n n_atom_types += 1\n self.ghost_coords = jnp.asarray(ghost_coords)\n self.n_nuc, self.n_up, self.n_down = n_nuc, n_up, n_down\n self.embedding_dim = embedding_dim\n self.node_data = {\n 'n_nodes': {'nuclei': n_nuc, 'electrons': n_up + n_down},\n 'n_node_types': {\n 'nuclei': n_atom_types if atom_type_embeddings else n_nuc,\n 'electrons': 1 if n_up == n_down else 2,\n },\n 'node_types': {\n 'nuclei': (\n jnp.unique(charges, size=n_nuc, return_inverse=True)[-1]\n if atom_type_embeddings\n else jnp.arange(n_nuc)\n ) + (1 if n_up == n_down else 2),\n 'electrons': jnp.array(n_up * [0] + n_down * [int(n_up != n_down)]),\n },\n }\n self.layers = [\n layer_factory(\n n_interactions,\n ilayer,\n n_nuc,\n n_up,\n n_down,\n embedding_dim,\n edge_types,\n self.node_data,\n edge_feat_dim,\n two_particle_stream_dim,\n )\n for ilayer in range(n_interactions)\n ]\n self.edge_features = edge_features\n self.edge_types = edge_types\n self.positional_electron_embeddings = positional_electron_embeddings\n\n def node_factory(self, phys_conf):\n n_elec_types = self.node_data['n_node_types']['electrons']\n n_nuc_types = self.node_data['n_node_types']['nuclei']\n if self.positional_electron_embeddings:\n edge_factory = MolecularGraphEdgeBuilder(\n self.n_nuc,\n self.n_up,\n self.n_down,\n ['ne'],\n feature_callbacks={\n 'ne': lambda *args: self.edge_features['ne'](\n difference_callback(*args)\n )\n },\n )\n ne_edges = edge_factory(phys_conf)['ne']\n ne_pos_feat = (\n jnp.zeros(\n (\n self.n_up + self.n_down + 1,\n self.n_nuc + 1,\n ne_edges.features.shape[-1],\n )\n )\n .at[ne_edges.receivers, ne_edges.senders]\n .set(ne_edges.features)[: self.n_up + self.n_down, : self.n_nuc]\n ) # [n_elec, n_nuc, n_edge_feat_dim]\n x = flatten(ne_pos_feat, start_axis=1)\n else:\n X = hk.Embed(n_elec_types, self.embedding_dim, name='ElectronicEmbedding')\n x = X(self.node_data['node_types']['electrons'])\n y = (\n hk.Embed(n_nuc_types, self.embedding_dim, name='NuclearEmbedding')(\n self.node_data['node_types']['nuclei'] - n_elec_types\n )\n if 'ne' in self.edge_types\n else None\n )\n return GraphNodes(y, x)\n\n def edge_factory(self, phys_conf):\n r\"\"\"Compute all the graph edges used in the GNN.\"\"\"\n\n def feature_callback(typ, *callback_args):\n return self.edge_features[typ](difference_callback(*callback_args))\n\n edge_factory = MolecularGraphEdgeBuilder(\n self.n_nuc,\n self.n_up,\n self.n_down,\n self.edge_types,\n feature_callbacks={\n typ: partial(feature_callback, typ) for typ in self.edge_types\n },\n )\n return edge_factory(phys_conf)\n\n def __call__(self, phys_conf):\n r\"\"\"\n Execute the graph neural network.\n\n Args:\n phys_conf (PhysicalConfiguration): the physical configuration\n of the molecule.\n\n Returns:\n float, (:math:`N_\\text{elec}`, :data:`embedding_dim`):\n the final embeddings of the electrons.\n \"\"\"\n if self.ghost_coords is not None:\n phys_conf = phys_conf._replace(\n R=jnp.concatenate(\n [\n phys_conf.R,\n jnp.tile(self.ghost_coords[None], (len(phys_conf.R), 1, 1)),\n ],\n axis=-2,\n )\n )\n graph_edges = self.edge_factory(phys_conf)\n graph_nodes = self.node_factory(phys_conf)\n graph = Graph(graph_nodes, graph_edges)\n\n for layer in self.layers:\n graph = layer(graph)\n\n return graph.nodes.electrons\n","sub_path":"src/deepqmc/gnn/electron_gnn.py","file_name":"electron_gnn.py","file_ext":"py","file_size_in_byte":17901,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"116556035","text":"__author__ = 'lundh'\n\n\nfrom twisted.internet import reactor, protocol\nimport threading\nimport door_server\nimport old_door_knock\nimport door_notify\nimport my_logs\n\nthread_list = list()\nstopped = 0\n\n\ndef signal_handler(signal, frame):\n global stopped\n stopped = 1\n\n\nif __name__ == \"__main__\":\n\n l = my_logs.Logger()\n\n nl = l.my_logger(\"gcm_log\")\n door_notify.init_log(nl)\n\n kl = l.my_logger(\"knock_log\")\n old_door_knock.init_log(kl)\n\n sl = l.my_logger(\"server_log\")\n door_server.init_log(sl)\n\n knock_thread = threading.Thread(target=old_door_knock.sensor)\n thread_list.append(knock_thread)\n knock_thread.start()\n\n f = protocol.Factory()\n f.protocol = door_server.Echo_Server\n\n reactor.listenTCP(1234, f)\n #reactor.run()\n\n server_thread = threading.Thread(target=reactor.run, args=(False,))\n server_thread.start()\n\n knock_thread.join()\n server_thread.join()\n","sub_path":"server/door_watcher.py","file_name":"door_watcher.py","file_ext":"py","file_size_in_byte":915,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"464691727","text":"\"\"\"定义Graduation_design的URL模式\"\"\"\n\nfrom django.urls import path\nfrom django.conf.urls import url\nfrom . import views\n\napp_name = 'Graduation_design'\nurlpatterns = [\n path('', views.Wel_Login,name='wel_login'),\n path('Welpage/', views.Welpage, name='Welpage'),\n path('welreigister/', views.Wel_Register, name=\"welreigister\"),\n path('login/', views.Login, name='login'),\n path('login/hello/register', views.Register, name='Register'),\n path('Introduce/', views.Introduce, name='Introduce'),\n path('Chart/', views.Chart, name='Chart'),\n path('index/', views.index, name='index'),\n path(r'^form/$', views.submitform, name='formUrl'),\n path(r'^showform/$', views.showform, name='showformUrl'),\n path(r'^Submail/$', views.Submail, name='Submail'),\n path(r'^Sendemail/$', views.Sendemail, name='Sendemail'),\n path(r'^Searchdetail/$',views.Searchdetail,name='Searchdetail'),\n path(r'^Doingdetail/$',views.Doingdetail,name='Doingdetail'),\n path(r'^Algorithm/$', views.Algorithm,name='Algorithm'),\n path(r'^KNN/$', views.Alo_Knn, name='knn'),\n path(r'^SVM/$', views.Alo_Svm, name='svm'),\n path(r'^TREE/$', views.Alo_Tree, name='tree'),\n path(r'^MO_TREE/$', views.Alo_more_tree, name='more_tree'),\n path(r'^LINEAR/$', views.Alo_linear, name='linear'),\n path(r'^Data/$', views.Alo_data, name='data'),\n\n]\n","sub_path":"graduation/Graduation_design/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":1363,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"459341281","text":"import cv2\nimport numpy as np\n\nimg = cv2.imread(\"imagens/estrada.jpg\")\ncv2.imshow(\"imagem_original\", img)\n\nlargura = img.shape[1]\naltura = img.shape[0]\nproporcao = float(altura/largura)\nlargura_nova = 230\naltura_nova = int(largura_nova * proporcao)\n\ntamanho_novo = (largura_nova, altura_nova)\n\nimg_redimensionada = cv2.resize(\n img, tamanho_novo, interpolation=cv2.INTER_AREA)\n\n\nimg_redimensionada2 = img[::2, ::2]\ncv2.imshow(\"imagem_redimensionada\", img_redimensionada)\ncv2.waitKey()\ncv2.imwrite(\"jose_novo.jpg\", img_redimensionada)\n","sub_path":"PDI(processamento digital de imagem)/estudos sobe PDI/redimensionado_imagem.py","file_name":"redimensionado_imagem.py","file_ext":"py","file_size_in_byte":537,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"237468016","text":"# библиотека для рисования математических графиков\r\nimport matplotlib.pyplot as plt\r\n\r\n\r\n# Метод использовался для отрисовки полей Record()'ов\r\ndef plot(records, id):\r\n param_list = []\r\n counter = 0\r\n for pi in id:\r\n user_list = records[pi]\r\n for c in range(len(user_list)):\r\n if c > len(param_list) - 1:\r\n param_list.append(0)\r\n param_list[c] += user_list[c].data_traffic_MB\r\n counter += 1\r\n for p in param_list:\r\n p /= counter\r\n plt.title('negative_charge')\r\n plt.plot(param_list)\r\n plt.show()\r\n","sub_path":"Deep_Neural_Network/Model/plot.py","file_name":"plot.py","file_ext":"py","file_size_in_byte":668,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"194994177","text":"# coding: utf-8\n\n\"\"\"\n OpenAPI Petstore\n\n This spec is mainly for testing Petstore server and contains fake endpoints, models. Please do not use this for any other purpose. Special characters: \\\" \\\\\n\n The version of the OpenAPI document: 1.0.0\n Generated by OpenAPI Generator (https://openapi-generator.tech)\n\n Do not edit the class manually.\n\"\"\" # noqa: E501\n\n\nimport re # noqa: F401\nimport io\nimport warnings\n\nfrom pydantic import validate_arguments, ValidationError\nfrom typing_extensions import Annotated\nfrom typing import overload, Optional, Union, Awaitable\n\nfrom petstore_api.models.foo_get_default_response import FooGetDefaultResponse\n\nfrom petstore_api.api_client import ApiClient\nfrom petstore_api.api_response import ApiResponse\nfrom petstore_api.exceptions import ( # noqa: F401\n ApiTypeError,\n ApiValueError\n)\n\n\nclass DefaultApi:\n \"\"\"NOTE: This class is auto generated by OpenAPI Generator\n Ref: https://openapi-generator.tech\n\n Do not edit the class manually.\n \"\"\"\n\n def __init__(self, api_client=None) -> None:\n if api_client is None:\n api_client = ApiClient.get_default()\n self.api_client = api_client\n\n @overload\n async def foo_get(self, **kwargs) -> FooGetDefaultResponse: # noqa: E501\n ...\n\n @overload\n def foo_get(self, async_req: Optional[bool]=True, **kwargs) -> FooGetDefaultResponse: # noqa: E501\n ...\n\n @validate_arguments\n def foo_get(self, async_req: Optional[bool]=None, **kwargs) -> Union[FooGetDefaultResponse, Awaitable[FooGetDefaultResponse]]: # noqa: E501\n \"\"\"foo_get # noqa: E501\n\n This method makes a synchronous HTTP request by default. To make an\n asynchronous HTTP request, please pass async_req=True\n\n >>> thread = api.foo_get(async_req=True)\n >>> result = thread.get()\n\n :param async_req: Whether to execute the request asynchronously.\n :type async_req: bool, optional\n :param _request_timeout: timeout setting for this request.\n If one number provided, it will be total request\n timeout. It can also be a pair (tuple) of\n (connection, read) timeouts.\n :return: Returns the result object.\n If the method is called asynchronously,\n returns the request thread.\n :rtype: FooGetDefaultResponse\n \"\"\"\n kwargs['_return_http_data_only'] = True\n if '_preload_content' in kwargs:\n message = \"Error! Please call the foo_get_with_http_info method with `_preload_content` instead and obtain raw data from ApiResponse.raw_data\" # noqa: E501\n raise ValueError(message)\n if async_req is not None:\n kwargs['async_req'] = async_req\n return self.foo_get_with_http_info(**kwargs) # noqa: E501\n\n @validate_arguments\n def foo_get_with_http_info(self, **kwargs) -> ApiResponse: # noqa: E501\n \"\"\"foo_get # noqa: E501\n\n This method makes a synchronous HTTP request by default. To make an\n asynchronous HTTP request, please pass async_req=True\n\n >>> thread = api.foo_get_with_http_info(async_req=True)\n >>> result = thread.get()\n\n :param async_req: Whether to execute the request asynchronously.\n :type async_req: bool, optional\n :param _preload_content: if False, the ApiResponse.data will\n be set to none and raw_data will store the\n HTTP response body without reading/decoding.\n Default is True.\n :type _preload_content: bool, optional\n :param _return_http_data_only: response data instead of ApiResponse\n object with status code, headers, etc\n :type _return_http_data_only: bool, optional\n :param _request_timeout: timeout setting for this request. If one\n number provided, it will be total request\n timeout. It can also be a pair (tuple) of\n (connection, read) timeouts.\n :param _request_auth: set to override the auth_settings for an a single\n request; this effectively ignores the authentication\n in the spec for a single request.\n :type _request_auth: dict, optional\n :type _content_type: string, optional: force content-type for the request\n :return: Returns the result object.\n If the method is called asynchronously,\n returns the request thread.\n :rtype: tuple(FooGetDefaultResponse, status_code(int), headers(HTTPHeaderDict))\n \"\"\"\n\n _params = locals()\n\n _all_params = [\n ]\n _all_params.extend(\n [\n 'async_req',\n '_return_http_data_only',\n '_preload_content',\n '_request_timeout',\n '_request_auth',\n '_content_type',\n '_headers'\n ]\n )\n\n # validate the arguments\n for _key, _val in _params['kwargs'].items():\n if _key not in _all_params:\n raise ApiTypeError(\n \"Got an unexpected keyword argument '%s'\"\n \" to method foo_get\" % _key\n )\n _params[_key] = _val\n del _params['kwargs']\n\n _collection_formats = {}\n\n # process the path parameters\n _path_params = {}\n\n # process the query parameters\n _query_params = []\n # process the header parameters\n _header_params = dict(_params.get('_headers', {}))\n # process the form parameters\n _form_params = []\n _files = {}\n # process the body parameter\n _body_params = None\n # set the HTTP header `Accept`\n _header_params['Accept'] = self.api_client.select_header_accept(\n ['application/json']) # noqa: E501\n\n # authentication setting\n _auth_settings = [] # noqa: E501\n\n _response_types_map = {\n }\n\n return self.api_client.call_api(\n '/foo', 'GET',\n _path_params,\n _query_params,\n _header_params,\n body=_body_params,\n post_params=_form_params,\n files=_files,\n response_types_map=_response_types_map,\n auth_settings=_auth_settings,\n async_req=_params.get('async_req'),\n _return_http_data_only=_params.get('_return_http_data_only'), # noqa: E501\n _preload_content=_params.get('_preload_content', True),\n _request_timeout=_params.get('_request_timeout'),\n collection_formats=_collection_formats,\n _request_auth=_params.get('_request_auth'))\n","sub_path":"samples/openapi3/client/petstore/python-aiohttp/petstore_api/api/default_api.py","file_name":"default_api.py","file_ext":"py","file_size_in_byte":6851,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"172058455","text":"# -*- coding: utf-8 -*-\n\"\"\"\n femagtools.hxy\n ~~~~~~~~~~~~~~\n\n Reading HXY files (EXPERIMENTAL)\n\n\"\"\"\n\nimport numpy as np\nfrom collections import defaultdict\n\n# K-means clustering\nclass point():\n def __init__(self, index, k, coord):\n self.index = index\n self.coord = coord\n self.k = k\n\ndef make_k_mapping(points):\n region = defaultdict(list)\n for p in points:\n region[p.k] = region[p.k] + [p.coord]\n return region\n\ndef calc_k_means(region):\n return [np.mean(region[k], axis=0) for k in region]\n\ndef update_k(points, means):\n for p in points:\n dists = [np.linalg.norm(m - p.coord) for m in means]\n p.k = np.argmin(dists)\n\ndef fit(points, epochs=10):\n for e in range(epochs):\n region = make_k_mapping(points)\n means = calc_k_means(region)\n update_k(points, means)\n return means, points\n\ndef evaluate(points):\n region = make_k_mapping(points)\n means = calc_k_means(region)\n dists = [np.linalg.norm(means[p.k]-p.coord) for p in points]\n return np.mean(dists)\n\ndef llf_(y, X, pr):\n # return maximized log likelihood\n nobs = float(X.shape[0])\n nobs2 = nobs / 2.0\n nobs = float(nobs)\n resid = y - pr\n ssr = np.sum((resid)**2)\n llf = -nobs2*np.log(2*np.pi) - nobs2*np.log(ssr / nobs) - nobs2\n return llf\n\n\ndef aic(y, X, pr, p):\n # return aic metric\n llf = llf_(y, X, pr)\n return -2*llf+2*p\n\n\ndef readSections(f):\n section = []\n movepos = False\n for line in f:\n if line.startswith(' MOVE POSITION'):\n movepos = True\n if section:\n # skip empty lines\n i = 0\n try:\n while not section[i]:\n i = i+1\n except IndexError:\n i = i-1\n yield section[i:]\n section = []\n if movepos:\n section.append(line.strip())\n yield section\n\n\ndef read(filename, num_magnets):\n \"\"\"read hxy file and return values grouped to magnets\"\"\"\n hxy = []\n with open(filename, encoding='latin1', errors='ignore') as f:\n for s in readSections(f):\n pos = float(s[0].split()[-1])\n num = np.array([[float(x) for x in l.split()] for l in s[5:] if l])\n hxy.append({'pos': pos, 'e': num[:, :2], 'hxy': num[:, 2:4],\n 'bxy': num[:, 4:6], 'mxy':num[:, 6:]})\n K = num_magnets\n points = [point(i, np.random.randint(0,K), xy)\n for i, xy in enumerate(hxy[0]['e'])]\n new_means, new_points = fit(points)\n # move values to magnets:\n magnets = [{'e': [p.coord for p in new_points if p.k == k],\n 'pos': [], 'hxy': [], 'bxy': [], 'mxy': []}\n for k in range(K)]\n hkeys = ['hxy', 'bxy', 'mxy']\n for i, h in enumerate(hxy): # all positions\n for mag in magnets:\n mag['pos'].append(h['pos'])\n m = [{k: [] for k in hkeys}\n for kk in range(K)]\n for p in new_points: # all elements\n for k in hkeys:\n m[p.k][k].append(h[k][p.k])\n for mk, magk in zip(m, magnets):\n for k in hkeys:\n magk[k].append(mk[k])\n for mag in magnets:\n for k in ['e'] + hkeys:\n mag[k] = np.array(mag[k])\n mag['havg'] = []\n mag['hmax'] = []\n for hpos in mag['hxy']:\n h = np.abs(np.linalg.norm(hpos, axis=1))\n mag['havg'].append(np.mean(h))\n mag['hmax'].append(np.max(h))\n\n # Note dimension of hkeys is (positions x elements x 2)\n\n return magnets\n","sub_path":"src/femagtools/hxy.py","file_name":"hxy.py","file_ext":"py","file_size_in_byte":3738,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"318997807","text":"from django.db import models\nfrom hvad.models import TranslatableModel, TranslatedFields\nfrom assets.hvadsentry import SentryTranslationManager\n\n# Create your models here.\n\nclass Asset(TranslatableModel):\n objects = SentryTranslationManager()\n\n name = models.CharField(max_length=255)\n translations = TranslatedFields(\n type = models.CharField(max_length=255)\n )\n value = models.IntegerField(default=0)\n\n def __unicode__(self):\n return self.name\n","sub_path":"assets/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":478,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"219647460","text":"# -*- coding: utf-8 -*-\n\"\"\"\nTests for the Crypto class.\n\"\"\"\nimport os \nimport plotly\nimport unittest\nfrom isoweek import Week\nfrom skill.skill import Crypto\nfrom tests.data import article_data\n\n\nclass CryptoTestCase(unittest.TestCase):\n \"\"\"\n Test case for the Crypto() class.\n \"\"\"\n\n @classmethod\n def setUpClass(self):\n \"\"\"\n Method that instantiate the Test Case.\n \"\"\"\n \n self.skill = Crypto()\n\n \n def test_skill_Crypto_results(self):\n \"\"\"\n Crypto().text() gives a numberic output.\n \"\"\"\n try:\n results = Crypto().text(\n text=article_data, limit=3)\n assert len(results) > 0\n for result in results:\n assert 'id' in result.keys()\n assert 'matches' in result.keys()\n assert 'prices' in result.keys()\n assert 'chart' in result.keys()\n assert 'related' in result.keys()\n except plotly.exceptions.PlotlyRequestError:\n pass\n\n def test_skill_Crypto_regex_plural(self):\n \"\"\"\n Crypto().text() classifies the same coin, with different starts and ends\n even if its plural.\n \"\"\"\n try:\n results = Crypto().text(\n 'bitcoin bitcoins', limit=3)\n for result in results:\n assert len(results) > 0\n assert result['matches'][0]['name_start'] == 0\n assert result['matches'][0]['name_end'] == 7\n assert result['matches'][1]['name_start'] == 8\n assert result['matches'][1]['name_end'] == 16\n except plotly.exceptions.PlotlyRequestError:\n pass\n\n def test_skill_crypto_symbol_is_correctly_identified(self):\n \"\"\"\n Crypto().text() classifies the same coin, with different starts and ends\n even if both classifications are different (name and symbol).\n \"\"\"\n results = Crypto().text('bitcoin BTC', limit=3)\n for result in results:\n assert len(results) > 0\n assert result['matches'][0]['name_start'] == 0\n assert result['matches'][0]['name_end'] == 7\n assert result['matches'][1]['name_start'] == 8\n assert result['matches'][1]['name_end'] == 11\n \n def test_no_conversation_about_coin(self):\n results = Crypto().text('Fluttercoin', limit=3)\n for result in results:\n assert len(results) > 0\n assert result['related'] == []\n\n def test_different_limits_different_results(self):\n \"\"\"\n Crypto().text() limit paramater limits the length of the output object\n to the integer given.\n \"\"\"\n results = self.skill.text(text=article_data, limit=3)\n assert len(results) == 3\n\n results = self.skill.text(text=article_data, limit=1)\n assert len(results) == 1\n","sub_path":"tests/unit/test_skill.py","file_name":"test_skill.py","file_ext":"py","file_size_in_byte":2912,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"264077865","text":"from measure_it import *\nimport boto\n\nconn = boto.connect_s3()\nbucket = conn.get_bucket('project-bluesmote')\nkeys = bucket.list('v1/blocks/')\n\n# about 5000 keys\nL = list(measure_iter(keys))\n\n# first key is really slow!\nL = list(measure_first(keys))\n","sub_path":"code/s3_demo.py","file_name":"s3_demo.py","file_ext":"py","file_size_in_byte":249,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"348032910","text":"'''(USANDO BREAK) Crie um programa que leia vários números inteiros pelo teclado. O programa só vai parar quando o usuário digitar o valor 999, que é a condição de parada. No final, mostre quantos números foram digitados e qual foi a soma entre eles (desconsiderando o flag(999))'''\nnumero=int(input('Digite um número(999 para sair): '))\nsoma=0\nquantidade=0\nwhile True:\n if numero==999:\n break\n soma+=numero\n quantidade+=1\n numero=int(input('Digite um número(999 para sair): '))\n \nprint('\\nQuantidade de números: {}\\nA soma entre eles: {}'.format(quantidade,soma))","sub_path":"Mundo2/WHILE/while010.py","file_name":"while010.py","file_ext":"py","file_size_in_byte":597,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"634410202","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('app', '0004_auto_20150905_1353'),\n ]\n\n operations = [\n migrations.RemoveField(\n model_name='video',\n name='category_id',\n ),\n migrations.AddField(\n model_name='video',\n name='channel_id',\n field=models.CharField(default=0, max_length=32),\n preserve_default=False,\n ),\n migrations.AlterField(\n model_name='classifier',\n name='id',\n field=models.CharField(max_length=32, serialize=False, primary_key=True),\n ),\n migrations.AlterField(\n model_name='classifier',\n name='model_filename',\n field=models.CharField(max_length=48, null=True, blank=True),\n ),\n ]\n","sub_path":"app/migrations/0005_auto_20150909_1939.py","file_name":"0005_auto_20150909_1939.py","file_ext":"py","file_size_in_byte":929,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"537662747","text":"#!/usr/bin/python3\n\"\"\"Querying the Reddit API\"\"\"\n\nimport requests\n\n\ndef count_words(subreddit, word_list, after='null', count_list=[]):\n \"\"\"Returns Number of Subs\"\"\"\n # Convert word_list to set to remove duplicates\n # temp = []\n # for w in word_list:\n # temp.append(w.lower())\n # word_list = set(temp)\n url = \"https://www.reddit.com/r/{}/hot.json\".format(subreddit)\n h = {\n 'User-Agent': 'Python/1.0(Holberton Project)'\n }\n after = {\n \"after\": after\n }\n req = requests.get(url, headers=h, allow_redirects=False, params=after)\n try:\n after = req.json()['data']['after']\n posts = req.json()['data']['children']\n except:\n return\n\n # Gets all title names from Query\n for item in posts:\n title = item['data']['title'].lower()\n for word in word_list:\n if word.lower() in title:\n count_list.append(word.lower())\n if after not in [\"NULL\", 'null', None, \"None\"]:\n return count_words(subreddit, word_list, after, count_list)\n else:\n my_dict = {i: count_list.count(i) for i in count_list}\n my_dict = {k: v for k, v in sorted(my_dict.items(),\n key=lambda item: item[1],\n reverse=True)}\n if len(my_dict) == 0:\n return\n for key, value in my_dict.items():\n if value != 0:\n print(\"{}: {}\".format(key, value))\n return\n","sub_path":"0x16-api_advanced/100-count.py","file_name":"100-count.py","file_ext":"py","file_size_in_byte":1493,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"554131617","text":"# Copyright 2016 Google Inc. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Tests for glazier.lib.config.files.\"\"\"\n\nfrom unittest import mock\n\nfrom absl.testing import absltest\nfrom glazier.lib import file_util\nfrom glazier.lib.config import files\nfrom pyfakefs import fake_filesystem\n\n\nclass FilesTest(absltest.TestCase):\n\n def setUp(self):\n super(FilesTest, self).setUp()\n self.filesystem = fake_filesystem.FakeFilesystem()\n files.open = fake_filesystem.FakeFileOpen(self.filesystem)\n files.file_util.os = fake_filesystem.FakeOsModule(self.filesystem)\n\n def testDump(self):\n op_list = ['op1', ['op2a', 'op2b'], 'op3', {'op4a': 'op4b'}]\n files.Dump('/tmp/foo/dump.txt', op_list)\n result = files._YamlReader('/tmp/foo/dump.txt')\n self.assertEqual(result[1], ['op2a', 'op2b'])\n self.assertEqual(result[3], {'op4a': 'op4b'})\n self.assertRaises(files.Error, files.Dump, '/tmp', [])\n\n @mock.patch.object(files.download.Download, 'DownloadFileTemp', autospec=True)\n def testRead(self, download):\n self.filesystem.create_file('/tmp/downloaded1.yaml', contents='data: set1')\n self.filesystem.create_file('/tmp/downloaded2.yaml', contents='data: set2')\n download.return_value = '/tmp/downloaded1.yaml'\n result = files.Read(\n 'https://glazier-server.example.com/unstable/dir/test-build.yaml')\n download.assert_called_with(\n mock.ANY,\n 'https://glazier-server.example.com/unstable/dir/test-build.yaml')\n self.assertEqual(result['data'], 'set1')\n # download error\n download.side_effect = files.download.DownloadError\n self.assertRaises(\n files.Error, files.Read,\n 'https://glazier-server.example.com/unstable/dir/test-build.yaml')\n # local\n result = files.Read('/tmp/downloaded2.yaml')\n self.assertEqual(result['data'], 'set2')\n\n @mock.patch.object(files.file_util, 'Remove', autospec=True)\n def testRemoveWithoutBackup(self, remove):\n files.Remove('/test/file/name.yaml', backup=False)\n remove.assert_called_with('/test/file/name.yaml')\n # error handling\n remove.side_effect = file_util.Error('test error')\n self.assertRaises(\n files.Error, files.Remove, '/test/file/name.yaml', backup=False)\n\n @mock.patch.object(files.file_util, 'Move', autospec=True)\n def testRemoveWithBackup(self, move):\n files.Remove('/test/file/name.yaml', backup=True)\n move.assert_called_with('/test/file/name.yaml', '/test/file/name.yaml.bak')\n # error handling\n move.side_effect = file_util.Error('test error')\n self.assertRaises(\n files.Error, files.Remove, '/test/file/name.yaml', backup=True)\n\n def testYamlReader(self):\n self.filesystem.create_file(\n '/foo/bar/baz.txt', contents='- item4\\n- item5\\n- item6')\n result = files._YamlReader('/foo/bar/baz.txt')\n self.assertEqual(result[1], 'item5')\n\n\nif __name__ == '__main__':\n absltest.main()\n","sub_path":"glazier/lib/config/files_test.py","file_name":"files_test.py","file_ext":"py","file_size_in_byte":3407,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"235793724","text":"# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nimport os\nimport codecs\n\nimport flask\nimport json\n\nfrom napkin import config as sv_config\n\n\nCURRENT_DIR = os.path.dirname(os.path.abspath(__file__))\nBASE_DIR = os.path.dirname(os.path.dirname(CURRENT_DIR))\n\n\ndef login_required(f):\n def wrapper(*args, **kwargs):\n if flask.session.get('username') is None:\n return flask.redirect('/u/login')\n else:\n return f(*args, **kwargs)\n return wrapper\n\n\ndef flask_req_get_post_data(req):\n \"\"\"\n data methods:\n ['add', 'clear', 'copy', 'fromkeys',\n 'get', 'getlist', 'has_key',\n 'items', 'iteritems', 'iterkeys',\n 'iterlists', 'iterlistvalues', 'itervalues',\n 'keys', 'lists', 'listvalues', 'pop', 'popitem',\n 'popitemlist', 'poplist', 'setdefault', 'setlist',\n 'setlistdefault', 'to_dict', 'update', 'values']\n \"\"\"\n data_form = req.form\n if data_form:\n return data_form\n data_json = req.json\n if data_json:\n return data_json\n data_val = req.values\n if data_val:\n return data_val\n data_data = req.data\n return data_data\n\n\ndef flask_req_get_querystr(req):\n q_args = req.args\n if q_args:\n return q_args\n q_val = req.values\n if q_val:\n return q_val\n q_data = req.data\n return q_data\n\n\ndef flask_req_get_header_token(req):\n return req.headers.get('X-Auth-Token', None)\n\n\ndef flask_request_args_get_others(req, excluding=[]):\n args = flask_req_get_querystr(req)\n r = {}\n for k, v in args.items():\n if k and v and k not in excluding:\n r[k] = v\n return r\n\n\ndef file_read_local(file_name, *args, **kwargs):\n pth = os.path.join(BASE_DIR, file_name)\n ret = open(pth, 'rb').read()\n return ret\n\n\ndef file_read_local_md(file_name, *args, **kwargs):\n pth = os.path.join(BASE_DIR, file_name)\n infile = codecs.open(pth, mode=\"r\", encoding=\"utf-8\")\n text = infile.read()\n return text\n\n\ndef str_to_list_by(s, seq=[',']):\n for a in seq:\n s = s.replace(a, ' ')\n lst = s.split()\n return lst\n","sub_path":"napkin/pages/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":2602,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"69643561","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Apr 13 08:35:58 2019\n\n@author: Rajiv Sambasivan\n\"\"\"\n\nfrom arango import ArangoClient, AQLQueryExecuteError\nimport logging\nfrom arangopipe.arangopipe_storage.arangopipe_config import ArangoPipeConfig\nfrom arangopipe.arangopipe_storage.custom_http_client import CustomHTTPClient\nfrom arangopipe.arangopipe_storage.managed_service_conn_parameters import ManagedServiceConnParam\n\n# create logger with 'spam_application'\nlogger = logging.getLogger('arangopipe_logger')\nlogger.setLevel(logging.DEBUG)\n# create file handler which logs even debug messages\nfh = logging.FileHandler('arangopipe.log')\nfh.setLevel(logging.DEBUG)\n# create console handler with a higher log level\nch = logging.StreamHandler()\nch.setLevel(logging.ERROR)\n# create formatter and add it to the handlers\nformatter = logging.Formatter(\n '%(asctime)s - %(name)s - %(levelname)s - %(message)s')\nfh.setFormatter(formatter)\nch.setFormatter(formatter)\n# add the handlers to the logger\nlogger.addHandler(fh)\nlogger.addHandler(ch)\n\n\nclass ArangoPipe:\n \"\"\" An instance of ArangoPipe is meant to be used to log a run of a pipeline execution. To use it:\n (1) Create a ArangoPipe object\n (2) Register your dataset with ArangoPipe\n (3) Register your featureset with ArangoPipe\n (4) Register you model with ArangoPipe\n\"\"\"\n def __init__(self, config):\n self.cfg = config.get_cfg()\n self.emlg = None\n self.db = None\n self.mscp = ManagedServiceConnParam()\n self.init_graph()\n self.heart_beat()\n\n def heart_beat(self):\n try:\n self.lookup_dataset(\"heart beat check\")\n except AQLQueryExecuteError as e:\n print(\"WARNING : \" + str(e))\n logger.error(\n \"Your database was perhaps deleted, try a new connection\")\n #logger.error(\"Error: \" + str(e))\n raise Exception(\"Your connection is stale, try a new connection!\")\n\n return\n\n def get_config(self):\n apc = ArangoPipeConfig()\n return apc.get_cfg()\n\n def get_collection_from_id(self, id_str):\n sep = \"/\"\n tokens = id_str.split(sep)\n col_name = tokens[0]\n\n return col_name\n\n def is_valid_id_str(self, id_str):\n valid_id_str = False\n sep = \"/\"\n tokens = id_str.split(sep)\n\n valid_id_str = True if len(tokens) == 2 else False\n\n return valid_id_str\n\n def link_entities(self, src_id, dest_id):\n src_id_valid = self.is_valid_id_str(src_id)\n dest_id_valid = self.is_valid_id_str(dest_id)\n\n if not src_id_valid:\n logger.error(\"The source node key does appear to be valid\")\n return\n if not dest_id_valid:\n logger.error(\"The destination key does not appear to be valid\")\n return\n\n dest_entity_type = self.get_collection_from_id(dest_id)\n src_entity_type = self.get_collection_from_id(src_id)\n related_key = \"related_\" + dest_entity_type\n concat_key = 'doc.' + related_key\n aql_str = 'FOR doc in %s FILTER doc._id == @value UPDATE doc WITH {\\\n %s: CONCAT_SEPARATOR(\",\", %s, @dest_entity) } IN %s' % (\n src_entity_type, related_key, concat_key, src_entity_type)\n\n cursor = self.db.aql.execute(aql_str,\n bind_vars={\n 'value': src_id,\n 'dest_entity': dest_id\n })\n\n return\n\n def lookup_entity_by_id(self, entity_id):\n entity_col = self.get_collection_from_id(entity_id)\n aql = 'FOR doc in %s FILTER doc._id == @value RETURN doc' % (\n entity_col)\n # Execute the query\n cursor = self.db.aql.execute(aql, bind_vars={'value': entity_id})\n asset_keys = [doc for doc in cursor]\n\n asset_info = None\n if len(asset_keys) == 0:\n logger.info(\"The asset by name: \" + asset_name +\\\n \" was not found in Arangopipe!\")\n else:\n asset_info = asset_keys[0]\n\n return asset_info\n\n def lookup_entity(self, asset_name, asset_type):\n aql = 'FOR doc IN %s FILTER doc.name == @value RETURN doc' % (\n asset_type)\n # Execute the query\n cursor = self.db.aql.execute(aql, bind_vars={'value': asset_name})\n asset_keys = [doc for doc in cursor]\n\n asset_info = None\n if len(asset_keys) == 0:\n logger.info(\"The asset by name: \" + asset_name +\\\n \" was not found in Arangopipe!\")\n else:\n asset_info = asset_keys[0]\n\n return asset_info\n\n def find_entity(self, attrib_name, attrib_value, asset_type):\n aql = 'FOR doc IN %s FILTER doc.%s == @value RETURN doc' % (\n asset_type, attrib_name)\n # Execute the query\n cursor = self.db.aql.execute(aql, bind_vars={'value': attrib_value})\n asset_keys = [doc for doc in cursor]\n\n asset_info = None\n if len(asset_keys) == 0:\n msg = \"Asset %s with %s = %s was not found!\" % (\n asset_type, attrib_name, attrib_value)\n logger.info(msg)\n else:\n asset_info = asset_keys\n\n return asset_info\n\n def lookup_dataset(self, dataset_name):\n \"\"\" Return a dataset identifier given a name. This can be used to get the dataset id that is used to log run information associated with execution of the pipeline.\"\"\"\n\n dataset_info = self.lookup_entity(dataset_name, 'datasets')\n\n return dataset_info\n\n def lookup_featureset(self, feature_set_name):\n \"\"\" Return a featureset identifier given a name. This can be used to get the featureset id that is used to log run information associated with execution of the pipeline.\"\"\"\n\n featureset_info = self.lookup_entity(feature_set_name, 'featuresets')\n\n return featureset_info\n\n def lookup_model(self, model_name):\n \"\"\" Return a model identifier given a name. This can be used to get the model id that is used to log run information associated with execution of the pipeline.\"\"\"\n\n model_info = self.lookup_entity(model_name, 'models')\n\n return model_info\n\n def lookup_modelparams(self, tag_value):\n \"\"\" Return a model parameter result given a tag.\"\"\"\n\n # Execute the query\n cursor = self.db.aql.execute('WITH modelparams\\\n FOR r IN run\\\n FILTER r.tag == @value \\\n FOR mp IN 1..1 OUTBOUND r run_modelparams\\\n RETURN mp ',\n bind_vars={'value': tag_value})\n mp_info = None\n mp_keys = [doc for doc in cursor]\n if len(mp_keys) == 0:\n logger.info(\"The model params for tag: \" + tag_value +\\\n \" was not found in Arangopipe!\")\n else:\n mp_info = mp_keys[0]\n return mp_info\n\n def lookup_modelperf(self, tag_value):\n \"\"\" Return a model dev performance given a tag.\"\"\"\n\n # Execute the query\n cursor = self.db.aql.execute('WITH devperf\\\n FOR r IN run\\\n FILTER r.tag == @value \\\n FOR dp IN 1..1 OUTBOUND r run_devperf\\\n RETURN dp ',\n bind_vars={'value': tag_value})\n mperf_info = None\n mperf_keys = [doc for doc in cursor]\n if len(mperf_keys) == 0:\n logger.info(\"The model performance for tag: \" + tag_value +\\\n \" was not found in Arangopipe!\")\n else:\n mperf_info = mperf_keys[0]\n\n return mperf_info\n\n def init_graph(self):\n \"\"\" Initialize a graph when an instance of ArangoPipe is provisioned. \"\"\"\n db_serv_host = self.cfg['arangodb']['DB_service_host']\n db_serv_port = self.cfg['arangodb']['DB_service_port']\n db_name = self.cfg['arangodb']['dbName']\n db_user_name = self.cfg['arangodb']['username']\n db_passwd = self.cfg['arangodb']['password']\n db_conn_protocol = self.cfg['arangodb'][self.mscp.DB_CONN_PROTOCOL]\n\n host_conn_str = db_conn_protocol + \"://\" + \\\n db_serv_host + \":\" + str(db_serv_port)\n client = ArangoClient(hosts= host_conn_str,\\\n http_client=CustomHTTPClient(username = db_user_name,\\\n password = db_passwd))\n\n self.db = client.db(name= db_name, \\\n username=db_user_name,\\\n password=db_passwd)\n\n self.emlg = self.db.graph(self.cfg['mlgraph']['graphname'])\n\n return\n\n\n\n\n def register_model(self, mi, user_id = \"authorized_user\",\\\n project = \"Wine-Quality-Regression-Modelling\"):\n \"\"\" Register a model. The operation requires specifying a user id. If the user id is permitted to register a model, then the registration proceeds, otherwise an unauthorized operation is indicated. \"\"\"\n\n model_name = mi[\"name\"]\n try:\n existing_model = self.lookup_model(model_name)\n except AQLQueryExecuteError as e:\n msg = \"The model name %s is not taken\" % (model_name)\n logger.info(msg)\n if existing_model is not None:\n msg = \"It looks like the model name %s is already taken, try another name\" % (\n model_name)\n logger.error(msg)\n return None\n models = self.emlg.vertex_collection(\"models\")\n model_reg = models.insert(mi)\n\n # Execute the query\n cursor = self.db.aql.execute(\n 'FOR doc IN project FILTER doc.name == @value RETURN doc',\n bind_vars={'value': project})\n project_keys = [doc for doc in cursor]\n the_project_info = project_keys[0]\n\n project_model_edge = self.emlg.edge_collection(\"project_models\")\n project_model_key = the_project_info[\"_key\"] + \"-\" + model_reg[\"_key\"]\n\n a_project_model_edge = {\"_key\": project_model_key,\\\n \"_from\": \"project/\" + the_project_info[\"_key\"],\\\n \"_to\": \"models/\" + model_reg[\"_key\"]}\n pm_reg = project_model_edge.insert(a_project_model_edge)\n logger.info(\"Recording project model link \" + str(pm_reg))\n\n return model_reg\n\n def register_dataset(self, ds_info, user_id=\"authorized_user\"):\n \"\"\" Register a dataset. The operation requires specifying a user id. If the user id is permitted to register a dataset, then the registration proceeds, otherwise an unauthorized operation is indicated. \"\"\"\n\n ds_name = ds_info[\"name\"]\n try:\n existing_ds = self.lookup_dataset(ds_name)\n except AQLQueryExecuteError as e:\n msg = \"The dataset name %s is not taken\" % (ds_name)\n logger.info(msg)\n if existing_ds is not None:\n msg = \"It looks like the dataset name %s is already taken, try another name\" % (\n ds_name)\n logger.error(msg)\n return None\n ds = self.emlg.vertex_collection(\"datasets\")\n ds_reg = ds.insert(ds_info)\n logger.info(\"Recording dataset dataset link \" + str(ds_reg))\n\n return ds_reg\n\n\n def register_featureset(self, fs_info, dataset_id, \\\n user_id = \"authorized_user\"):\n \"\"\" Register a featureset. ManagedServiceConnParamThe operation requires specifying a user id. If the user id is permitted to register a featureset, then the registration proceeds, otherwise an unauthorized operation is indicated. \"\"\"\n fs_name = fs_info[\"name\"]\n try:\n existing_fs = self.lookup_featureset(fs_name)\n except AQLQueryExecuteError as e:\n msg = \"The featureset name %s is not taken\" % (fs_name)\n logger.info(msg)\n if existing_fs is not None:\n msg = \"It looks like the featureset name %s is already taken, try another name\" % (\n fs_name)\n logger.error(msg)\n return None\n\n fs = self.emlg.vertex_collection(\"featuresets\")\n fs_reg = fs.insert(fs_info)\n logger.info(\"Recording featureset \" + str(fs_reg))\n featureset_dataset_edge = self.emlg.edge_collection(\n \"featureset_dataset\")\n featureset_dataset_key = fs_reg[\"_key\"] + \"-\" + dataset_id\n\n a_featureset_dataset_edge = {\"_key\": featureset_dataset_key,\\\n \"_from\": \"featuresets/\" + fs_reg[\"_key\"],\\\n \"_to\": \"datasets/\" + dataset_id}\n fsds_reg = featureset_dataset_edge.insert(a_featureset_dataset_edge)\n logger.info(\"Recording featureset dataset link \" + str(fsds_reg))\n\n return fs_reg\n\n def log_run(self, ri):\n \"\"\" Log a run. Logging a run requires specifying a dataset, featureset and a model against which this run is recored. A run records model parameters and model performance. The run object is probably most useful for the analysis of model performance with respect to a featureset, model hyper-parameters and a dataset.\"\"\"\n\n rrid = ri[\"run_id\"]\n mp = ri[\"model-params\"]\n mp[\"_key\"] = mp[\"run_id\"]\n mperf = ri[\"model-perf\"]\n mperf[\"_key\"] = mperf[\"run_id\"]\n model_key = ri[\"model\"]\n\n run = self.emlg.vertex_collection(\"run\")\n run_info = {\"_key\": rrid, \"timestamp\": mperf[\"timestamp\"]}\n if \"deployment_tag\" in ri:\n run_info[\"deployment_tag\"] = ri[\"deployment_tag\"]\n if \"tag\" in ri:\n run_info[\"tag\"] = ri[\"tag\"]\n logger.info(\"Run info \" + str(run_info))\n run_reg = run.insert(run_info)\n logger.info(\"Recording run \" + str(run_reg))\n\n run_model_key = run_reg[\"_key\"] + \"-\" + model_key\n a_run_model_edge = {\"_key\": run_model_key,\\\n \"_from\" :\"models/\" + model_key, \\\n \"_to\": \"run/\" + rrid}\n run_model_edge = self.emlg.edge_collection(\"run_models\")\n rme_reg = run_model_edge.insert(a_run_model_edge)\n logger.info(\"Recording model run link \" + str(rme_reg))\n\n model_param = self.emlg.vertex_collection(\"modelparams\")\n mp_reg = model_param.insert(mp)\n logger.info(\"Recording model params \" + str(mp_reg))\n\n run_fs_edge = self.emlg.edge_collection(\"run_featuresets\")\n run_fs_key = rrid + \"-\" + ri[\"featureset\"]\n\n a_edge_run_fs = {\"_key\": run_fs_key,\\\n \"_from\": \"run/\" + rrid,\\\n \"_to\": \"featuresets/\" + ri[\"featureset\"]}\n\n rfse_reg = run_fs_edge.insert(a_edge_run_fs)\n logger.info(\"Recording run featureset link \" + str(rfse_reg))\n\n run_mp_edge = self.emlg.edge_collection(\"run_modelparams\")\n run_mp_key = rrid + \"-\" + mp[\"run_id\"]\n\n a_run_mp_edge = {\"_key\": run_mp_key,\\\n \"_from\": \"run/\" + rrid,\\\n \"_to\": \"modelparams/\" + mp_reg[\"_key\"]}\n\n rmp_reg = run_mp_edge.insert(a_run_mp_edge)\n logger.info(\"Recording run model params \" + str(rmp_reg))\n\n model_perf = self.emlg.vertex_collection(\"devperf\")\n dp_reg = model_perf.insert(mperf)\n logger.info(\"Recording model dev performance \" + str(dp_reg))\n\n run_devperf_edge = self.emlg.edge_collection(\"run_devperf\")\n run_devperf_key = rrid + \"-\" + dp_reg[\"_key\"]\n\n a_run_devperfedge = {\"_key\": run_devperf_key,\\\n \"_from\": \"run/\" + rrid,\\\n \"_to\": \"devperf/\" + dp_reg[\"_key\"]}\n rdp_reg = run_devperf_edge.insert(a_run_devperfedge)\n logger.info(\"Recording run dev perf link \" + str(rdp_reg))\n\n run_dataset_edge = self.emlg.edge_collection(\"run_datasets\")\n run_dataset_key = rrid + \"-\" + ri[\"dataset\"]\n\n a_run_dataset_edge = {\"_key\": run_dataset_key,\\\n \"_from\": \"run/\" + rrid,\\\n \"_to\": \"datasets/\" + ri[\"dataset\"]}\n rds_reg = run_dataset_edge.insert(a_run_dataset_edge)\n logger.info(\"Recording run dev perf link \" + str(rds_reg))\n\n return\n\n def log_serving_perf(self, sp, dep_tag, userid=\"authorized user\"):\n \"\"\" Log serving performance against a deployed model. The user making the request needs to be authorized to log this performance update. A serving performance vertex is created and is linked with its deployment vertex\"\"\"\n servingperf = self.emlg.vertex_collection(\"servingperf\")\n sp_reg = servingperf.insert(sp)\n\n # Execute the query\n cursor = self.db.aql.execute(\n 'FOR doc IN deployment FILTER doc.tag == @value RETURN doc',\n bind_vars={'value': dep_tag})\n dep_docs = [doc for doc in cursor]\n the_dep_doc = dep_docs[0]\n # Link the service performance record with the deployment record\n dep_servingperf_edge = self.emlg.edge_collection(\n \"deployment_servingperf\")\n dep_servingperf_key = the_dep_doc[\"_key\"] + \"-\" + sp_reg[\"_key\"]\n the_dep_servingperf_edge = { \"_key\": dep_servingperf_key,\\\n \"_from\": the_dep_doc[\"_id\"],\\\n \"_to\": sp_reg[\"_id\"]}\n\n dep_servingperf_reg = dep_servingperf_edge.insert(\n the_dep_servingperf_edge)\n return dep_servingperf_reg\n\n def insert_into_vertex_type(self, vertex_type_name, document):\n vertex_info = None\n if self.emlg.has_vertex_collection(vertex_type_name):\n vc = self.emlg.vertex_collection(vertex_type_name)\n vertex_info = vc.insert(document)\n else:\n logger.error(\"Vertex, \" + vertex_type_name +\n \" does not exist in Arangopipe!\")\n\n return vertex_info\n\n def insert_into_edge_type(self,\n edge_name,\n from_vdoc,\n to_vdoc,\n document=None):\n edge_info = None\n if self.emlg.has_edge_collection(edge_name):\n try:\n ec = self.emlg.edge_collection(edge_name)\n edge_key = from_vdoc['_key'] + \"-\" + to_vdoc['_key']\n if document is not None:\n document[\"_from\"] = from_vdoc['_id']\n document[\"_to\"] = to_vdoc['_id']\n document[\"_key\"] = edge_key\n edge_info = ec.insert(document)\n else:\n document = { \"_from\": from_vdoc['_id'],\\\n \"_to\": to_vdoc['_id'],\\\n \"_key\": edge_key }\n edge_info = ec.insert(document)\n except Exception as e:\n logger.error(e)\n else:\n logger.error(\"Edge, \" + edge_name + \" does not exist!\")\n\n return edge_info\n","sub_path":"arangopipe/arangopipe/arangopipe/arangopipe_storage/arangopipe_api.py","file_name":"arangopipe_api.py","file_ext":"py","file_size_in_byte":19049,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"365835745","text":"import unittest\nfrom max_words import *\n\nDOCUMENTS_PATH = \"./testing_docs\"\n\n\nclass MyTest(unittest.TestCase):\n\n def test_are_there_files_for_testing(self):\n files = []\n\n for (dirpath, _, filenames) in walk(DOCUMENTS_PATH):\n # first loop will give top directory files only.\n # list comprension to add relative path to filename.\n files = [\"{}/{}\".format(dirpath, filename) for filename in\n filenames]\n break\n\n self.assertTrue(len(files) > 0)\n\n def test_getting_file_names(self):\n self.assertTrue(len(get_files(DOCUMENTS_PATH)) > 0)\n\n def test_word_data_structure(self):\n ds = build_word_dict(get_files(DOCUMENTS_PATH))\n # get any item from dict:\n item = next(iter(ds.keys()))\n\n # print(type(ds[item]))\n # print(ds[item])\n self.assertTrue(type(ds[item]) is list)\n self.assertTrue(type(ds[item][0]) is int)\n self.assertTrue(type(ds[item][1]) is defaultdict)\n\n item2 = next(iter(ds[item][1]))\n self.assertTrue(type(ds[item][1][item2]) is set)\n\n\nif __name__ == '__main__':\n unittest.main()\n","sub_path":"tests.py","file_name":"tests.py","file_ext":"py","file_size_in_byte":1158,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"90"} +{"seq_id":"78505045","text":"import rospy\nimport time\nfrom math import sin, cos, tan, sqrt,pi\nimport numpy as np\nfrom math import factorial as f\nfrom scipy.linalg import block_diag\nfrom qpsolvers import solve_qp\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits import mplot3d\nimport warnings\n#import tensorflow as tf\nfrom scipy.optimize import Bounds,minimize\n\nfrom constrained_time_opt_new import min_snap\n\nfrom std_msgs.msg import String, Float64, Int16\nfrom sensor_msgs.msg import NavSatFix, Image,Imu\nfrom mavros_msgs.srv import CommandTOL, SetMode, CommandBool\nfrom mavros_msgs.msg import AttitudeTarget\nfrom geometry_msgs.msg import PoseStamped, Pose, Point, Twist, TwistStamped\nimport math\nfrom time import sleep\n\nwarnings.filterwarnings(\"ignore\", category=np.VisibleDeprecationWarning)\nwarnings.filterwarnings(\"ignore\", category=FutureWarning)\nwarnings.filterwarnings(\"ignore\")\n\nclass DroneIn3D:\n\t\n\tdef __init__(self):\n\t\tself.X=np.array([\n\t\t\t# x0, y1, z2, phi3, theta4, psi5, \n\t\t\t0.0, 0.0, 0.0, 0.0, 0.0, 0.0,\n\t\t\t# x_dot6, y_dot7, z_dot8\n\t\t\t0.0, 0.0, 0.0]) \n\t\tself.g = 9.81\n\n\t\tself.gps_lat=0\n\t\tself.gps_long=0\n\n\t\trospy.init_node('iris_drone', anonymous = True)\n\n\t\t#SUBSCRIBERS\n\t\t\n\t\trospy.Subscriber('/mavros/global_position/global', NavSatFix, self.global_pose)\n\t\trospy.Subscriber('/mavros/local_position/pose', PoseStamped, self.loc_pose)\n\t\tself.get_linear_vel=rospy.Subscriber('/mavros/local_position/velocity_local', TwistStamped, self.get_vel,)\n\t\trospy.Subscriber('/mavros/imu/data',Imu,self.get_euler_angles)\n\t\t#self.loc=Point()\n\t\tself.glob=Point()\n\n\t\t#PUBLISHERS\n\t\tself.pub = rospy.Publisher('/mavros/setpoint_position/local', PoseStamped,queue_size=1)\n\t\tself.publish_attitude_thrust=rospy.Publisher('/mavros/setpoint_raw/attitude', AttitudeTarget,queue_size=1)\n\n\t\t#POSITION QAUD\n\t\trospy.loginfo('INIT')\n\t\tself.setarm(1)\n\t\trospy.sleep(2)\n\t\tself.offboard()\n\t\tself.takeoff(1.0)\n\t\t#self.offboard()\n\t\trospy.sleep(5)\n\t\tself.gotopose(0,0,4)\n\t\trospy.sleep(5)\n\t\t\n\t\t\n\t\t\n\n\tdef offboard(self):\n\t\trate = rospy.Rate(10)\n\t\tsp = PoseStamped()\n\t\tsp.pose.position.x = 0.0\n\t\tsp.pose.position.y = 0.0\n\t\tsp.pose.position.z = 7.0\n\t\tfor i in range(10):\n\t\t\tself.pub.publish(sp)\n\t\t\trate.sleep()\n\t\tprint('We are good to go!!')\n\t\tself.setmode(\"GUIDED\")\n\n\tdef loc_pose(self, data):\n\n\t\tself.X[0] = data.pose.position.x\n\t\tself.X[1] = data.pose.position.y\n\t\tself.X[2] = data.pose.position.z\n\n\n\tdef global_pose(self, data):\n\t\tself.glob.x = data.latitude \n\t\tself.glob.y = data.longitude \n\t\tself.glob.z = data.altitude \n\n\tdef setmode(self,md):\n\t\trospy.wait_for_service('/mavros/set_mode')\n\t\ttry:\n\t\t\tmode = rospy.ServiceProxy('/mavros/set_mode', SetMode)\n\t\t\tresponse = mode(0,md)\n\t\t\tresponse.mode_sent\n\t\texcept rospy.ServiceException as e:\n\t\t\tprint (\"Service call failed: %s\"%e)\n\n\tdef takeoff(self, alt):\n\t\trospy.wait_for_service('/mavros/cmd/takeoff')\n\t\ttry:\n\t\t\tmode = rospy.ServiceProxy('/mavros/cmd/takeoff', CommandTOL)\n\t\t\tresponse = mode(0,0, self.glob.x, self.glob.y, alt)\n\t\t\tresponse.success\n\t\texcept rospy.ServiceException as e:\n\t\t\tprint (\"Service call failed: %s\"%e)\n\n\tdef setarm(self,av): # input: 1=arm, 0=disarm\n\t\trospy.wait_for_service('/mavros/cmd/arming')\n\t\ttry:\n\t\t\tarming = rospy.ServiceProxy('/mavros/cmd/arming', CommandBool)\n\t\t\tresponse = arming(av)\n\t\t\tresponse.success\n\t\texcept rospy.ServiceException as e:\n\t\t\tprint (\"Service call failed: %s\" %e)\n\n\tdef gotopose(self, x, y ,z):\n\t\trate = rospy.Rate(20)\n\t\tself.sp = PoseStamped()\n\t\tself.sp.pose.position.x = x\n\t\tself.sp.pose.position.y = y\n\t\tself.sp.pose.position.z = z\n\t\tdist = np.sqrt(((self.X[0]-x)**2) + ((self.X[1]-y)**2) + ((self.X[2]-z)**2))\n\t\twhile(dist > 0.18):\n\t\t\tself.pub.publish(self.sp)\n\t\t\tdist = np.sqrt(((self.X[0]-x)**2) + ((self.X[1]-y)**2) + ((self.X[2]-z)**2))\n\t\t\trate.sleep()\n\t\t#print('Reached ',x,y,z) \n\n\n\n\n\n\t'''def get_pose(self, location_data):\n\t\tself.X[0] = location_data.pose.position.x\n\t\tself.X[1] = location_data.pose.position.y\n\t\tself.X[2] = location_data.pose.position.z'''\n\n\n\tdef get_vel(self,vel_data):\n\t\tself.X[6]=\tvel_data.twist.linear.x\n\t\tself.X[7]=\tvel_data.twist.linear.y\n\t\tself.X[8]=\tvel_data.twist.linear.z\n\t\t\n\t\t\n\t\n\n\n\tdef get_euler_angles(self,orientaion_data):\n\t\tx=orientaion_data.orientation.x\n\t\ty=orientaion_data.orientation.y\n\t\tz=orientaion_data.orientation.z\n\t\tw=orientaion_data.orientation.w\n\n\t\tt0 = +2.0 * (w * x + y * z)\n\t\tt1 = +1.0 - 2.0 * (x * x + y * y)\n\t\tself.X[3] = math.atan2(t0, t1)\n\n\t\tt2 = +2.0 * (w * y - z * x)\n\t\tt2 = +1.0 if t2 > +1.0 else t2\n\t\tt2 = -1.0 if t2 < -1.0 else t2\n\t\tself.X[4] = math.asin(t2)\n\n\t\tt3 = +2.0 * (w * z + x * y)\n\t\tt4 = +1.0 - 2.0 * (y * y + z * z)\n\t\tself.X[5]= math.atan2(t3, t4)\n\n\n\n\tdef R(self):\n\t\n\t\tr_x = np.array([[1, 0, 0],\n\t\t\t\t\t\t[0, cos(self.X[3]), -sin(self.X[3])],\n\t\t\t\t\t\t[0, sin(self.X[3]), cos(self.X[3])]])\n\n\t\tr_y = np.array([[cos(self.X[4]), 0, sin(self.X[4])],\n\t\t\t\t\t\t[0, 1, 0],\n\t\t\t\t\t\t[-sin(self.X[4]), 0, cos(self.X[4])]])\n\n\t\tr_z = np.array([[cos(self.X[5]), -sin(self.X[5]), 0],\n\t\t\t\t\t\t[sin(self.X[5]), cos(self.X[5]), 0],\n\t\t\t\t\t\t[0,0,1]])\n\n\t\tr_yx = np.matmul(r_y, r_x)\n\t\treturn np.matmul(r_z, r_yx)\n\n\nclass Controller:\n\t\n\tdef __init__(self,\n\t\t\t\tz_k_p= 10, #1500.0,\n\t\t\t\tz_k_d=1, #0.0,\n\t\t\t\tx_k_p=0.0,\n\t\t\t\tx_k_d=0.0,\n\t\t\t\ty_k_p=10.0,\n\t\t\t\ty_k_d=0.0,\n\t\t\t\tk_p_roll=0.0,\n\t\t\t\tk_p_pitch=0.0,\n\t\t\t\tk_p_yaw=0.0):\n\t\t\n\t\t\n\t\tself.z_k_p = z_k_p\n\t\tself.z_k_d = z_k_d\n\t\tself.x_k_p = x_k_p\n\t\tself.x_k_d = x_k_d\n\t\tself.y_k_p = y_k_p\n\t\tself.y_k_d = y_k_d\n\t\tself.k_p_roll = k_p_roll\n\t\tself.k_p_pitch = k_p_pitch\n\t\tself.k_p_yaw = k_p_yaw\n\n\t \n\t\tself.g = 9.8\n\t\n\tdef altitude_controller(self,\n\t\t\t\t\t z_target,\n\t\t\t\t\t z_dot_target,\n\t\t\t\t\t z_dot_dot_target,\n\t\t\t\t\t z_actual,\n\t\t\t\t\t z_dot_actual,\n\t\t\t\t\t rot_mat):\n\t\n\t\tdef pd(kp, kd, error, error_dot, target):\n\t\t\tp_term = kp * error\n\t\t\td_term = kd * error_dot\n\t\t\treturn p_term + d_term + target\n\t\t\n\t\tu_1_bar = pd(self.z_k_p, self.z_k_d, \n\t\t\t\t\terror = z_target - z_actual, \n\t\t\t\t\terror_dot = z_dot_target - z_dot_actual,\n\t\t\t\t\ttarget = z_dot_dot_target)\n\t\tprint('target z_acc :',z_dot_dot_target,' | z_dot_error :',z_dot_target - z_dot_actual,' | z_error :',z_target - z_actual)\n\t\tb_z = rot_mat[2,2]\n\t\tc=(u_1_bar + self.g)/b_z\n\t\treturn c\n\t\t#return u_1_bar\n\n\tdef lateral_controller(self,\n\t\t\t\t\t x_target,\n\t\t\t\t\t x_dot_target,\n\t\t\t\t\t x_dot_dot_target,\n\t\t\t\t\t x_actual,\n\t\t\t\t\t x_dot_actual,\n\t\t\t\t\t y_target,\n\t\t\t\t\t y_dot_target,\n\t\t\t\t\t y_dot_dot_target,\n\t\t\t\t\t y_actual,\n\t\t\t\t\t y_dot_actual,\n\t\t\t\t\t c):\n\t\n\t\tdef pd(kp, kd, error, error_dot, target):\n\t\t\t# Proportional and differential control terms\n\t\t\tp_term = kp * error\n\t\t\td_term = kd * error_dot\n\t\t\t\n\t\t\t# Control command (with feed-forward term)\n\t\t\treturn p_term + d_term + target\n\t\t\n\t\t# Determine errors\n\t\tx_err = x_target - x_actual\n\t\ty_err = y_target - y_actual\n\t\tx_err_dot = x_dot_target - x_dot_actual\n\t\ty_err_dot = y_dot_target - y_dot_actual\n\t\t\n\t\t# Apply the PD controller\n\t\tx_dot_dot_command = pd(self.x_k_p, self.x_k_d, x_err, x_err_dot, x_dot_dot_target)\n\t\ty_dot_dot_command = pd(self.y_k_p, self.y_k_d, y_err, y_err_dot, y_dot_dot_target)\n\n\t\t# Determine controlled values by normalizing with the collective thrust\n\t\tb_x_c = x_dot_dot_command / c\n\t\tb_y_c = y_dot_dot_command / c\n\n\t\tprint(' X acc target :',x_dot_dot_target,' | x_dot_err :',x_dot_target - x_dot_actual,'x-err :',x_target - x_actual)\n\t\tprint(' Y acc target :',y_dot_dot_target,' | y_dot_err :',y_dot_target - y_dot_actual,'y-err :',y_target - y_actual)\n\t\treturn b_x_c, b_y_c\n\n\tdef roll_pitch_controller(self,\n\t\t\t\t\t\t b_x_c_target,\n\t\t\t\t\t\t b_y_c_target,\n\t\t\t\t\t\t rot_mat):\n\t\n\t\tdef p(kp, error):\n\t\t\treturn kp * error\n\t\t\n\t\tb_x = rot_mat[0,2]\n\t\tb_y = rot_mat[1,2]\n\t\t\n\t\tb_x_commanded_dot = p(self.k_p_roll, error=b_x_c_target - b_x)\n\t\tb_y_commanded_dot = p(self.k_p_pitch, error=b_y_c_target - b_y)\n\n\t\trot_mat1 = np.array([[rot_mat[1,0], -rot_mat[0,0]], \n\t\t\t\t\t\t\t[rot_mat[1,1], -rot_mat[0,1]]]) / rot_mat[2,2]\n\n\t\trot_rate = np.matmul(rot_mat1, np.array([b_x_commanded_dot, b_y_commanded_dot]).T)\n\t\tp_c = rot_rate[0]\n\t\tq_c = rot_rate[1]\n\n\t\tprint('roll error :',b_x_c_target - b_x,' | pitch error :',b_y_c_target - b_y)\n\n\t\treturn p_c, q_c\n\n\t\n\tdef yaw_controller(self,\n\t\t\t\t psi_target,\n\t\t\t\t psi_actual):\n\t\n\t\tdef p(kp, error):\n\t\t\treturn kp * error\n\n\t\treturn p(self.k_p_yaw, error=psi_target - psi_actual)\n\ndef actuate(x,y,z,v_test,v_min,v_max):\n\n\t#plt.figure(figsize=(10,5))\n\t#ax = plt.axes(projection ='3d')\n\tms = min_snap(x,y,z,v_test,v_min,v_max)\n\t#ms.plot_test_case('r','Test Case Trajectory')\n\tms.optimize()\n\tx_path,x_dot_path,x_dot_dot_path,y_path,y_dot_path,y_dot_dot_path,z_path,z_dot_path,z_dot_dot_path,psi_path=ms.get_trajectory_var()\n\t#ms.plot('g','Time Optimized Trajectory')\n\t#plt.legend()\n\t#plt.show()\n\n\tdrone = DroneIn3D() \n\t#sleep(2) \n\tcontrol_system = Controller(z_k_p=20.0,z_k_d=1.5,x_k_p=0.8,x_k_d=0.0,y_k_p=0.0,y_k_d=0.0,k_p_roll=0.0,k_p_pitch=20,\n\tk_p_yaw=0.0)\n\titer=0 \n\trate=rospy.Rate(20) \n\tprint(np.shape(z_path)[0]) \n\t#print(z_path)\n\t#for i in range(0,z_path.shape[0]):\n\twhile (iterDNA\\nGAACACGTGGAGGCAAACAGGAAGGTGAAGAAGAACTTATCCTATCAGGACGGAAGGTCCTGTGCTCGGG\\nATCTTCCAGACGTCGCGACTCTAAATTGCCCCCTCTGAGGTCAAGGAACACAAGATGGTTTTGGAAATGC\\nTGAACCCGATACATTATAACATCACCAGCATCGTGCCTGAAGCCATGCCTGCTGCCACCATGCCAGTCCT\"\r\n\r\nsequence = st.text_area(\"Sequence input\", sequence_input, height=220)\r\nsequence = sequence.splitlines()\r\nsequence = sequence[1:] # Skips the Sequence Name (First Line)\r\n# Concatenates List to String without any Whitespace.\r\nsequence = ''.join(sequence)\r\n\r\n# DNA Nucleotide Count\r\nst.header('DNA Nucleotide Count')\r\n\r\n\r\ndef DNA_nucleotide_count(seq):\r\n d = dict([\r\n ('A', seq.count('A')),\r\n ('T', seq.count('T')),\r\n ('G', seq.count('G')),\r\n ('C', seq.count('C'))\r\n ])\r\n return d\r\n\r\n\r\nX = DNA_nucleotide_count(sequence)\r\n\r\n# Display DataFrame\r\nst.subheader('DataFrame')\r\ndf = pd.DataFrame.from_dict(X, orient='index')\r\n\r\ndf = df.rename({0: 'Count'}, axis='columns') # Rename Column Name\r\ndf.reset_index(inplace=True)\r\ndf = df.rename(columns={'index': 'Nucleotide'}) # Rename Column Name\r\nst.write(df)\r\n\r\n# Display Bar Chart using Altair\r\nst.subheader('Bar chart')\r\nbar = alt.Chart(df).mark_bar().encode(x='Nucleotide', y='Count')\r\nbar = bar.properties(width=alt.Step(80)) # Bar Width\r\n\r\nst.write(bar)\r\n","sub_path":"DNA Count/DNA_Count.py","file_name":"DNA_Count.py","file_ext":"py","file_size_in_byte":1619,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"219023175","text":"#! /usr/bin/python3\n\n# should take a single argument, a filepath to the txt file to read.\n# read the text file, iterate over each occurrence of ['ADJECTIVE', 'NOUN','ADVERB', 'VERB']\n# print out the replaced text\n\nimport re, os, sys\n\ndef prompt (string):\n if string == 'ADJECTIVE':\n print('Enter an adjective')\n elif string == 'NOUN':\n print('Enter a noun:')\n elif string == 'ADVERB':\n print('Enter an adverb:')\n elif string == 'VERB':\n print('Enter a verb')\n else:\n return string\n\n return input()\n\ndef madlibify (string):\n word_list = re.split(r'(ADJECTIVE|NOUN|VERB|ADVERB)', string)\n word_list = map(prompt, word_list)\n \n return ''.join(word_list)\n\nif len(sys.argv) == 2:\n file_path = sys.argv[1]\n try:\n text_file = open(file_path)\n except FileNotFoundError:\n print('File not found!')\n sys.exit()\n else:\n content = text_file.read()\n\n madlibbed_string = madlibify(content)\n print(madlibbed_string)\n\n output_file = open('mad_libbed.txt', 'w')\n output_file.write(madlibbed_string)\n output_file.close()\n","sub_path":"ch8/madlib/mad_libs.py","file_name":"mad_libs.py","file_ext":"py","file_size_in_byte":1186,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"544782766","text":"import math\n\n\ndef __fdp_x(x, y, l):\n n = len(y)\n dx = l/n\n i1 = math.floor(x/dx)\n i2 = i1 + 1\n if i2 > n:\n return y[n]\n else:\n dy = y[i2] - y[i1]\n out = (dy/dx)*x + y[i1]\n return out\n","sub_path":"pyomomod/ti_hi.py","file_name":"ti_hi.py","file_ext":"py","file_size_in_byte":229,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"454893965","text":"import json\nimport csv\nimport requests\nfrom datetime import datetime\nfrom dateutil import tz\nimport time\n\nwith open(\"subreddit_list.json\", \"r\") as subFile:\n subreddits = json.load(subFile)\n\n# Timezones\nfrom_zone = tz.gettz('UTC')\nto_zone = tz.gettz('America/New_York')\n\nheaders = {'user-agent': 'Mozilla/5.0 (X11; Linux x86_64; rv:67.0) Gecko/20100101 Firefox/67.0'}\n\ncsvFile = open('activity.csv', 'a', newline='')\n\nwriter = csv.writer(csvFile)\nwriter.writerow([''] + subreddits)\n\ncurts = datetime.utcfromtimestamp(time.time())\ncurts = curts.replace(tzinfo=from_zone).astimezone(to_zone)\ncurts = curts.strftime('%Y/%m/%d/%H:%M:%S')\nrow = [curts]\n\n# Convert time zone\nfor s in subreddits:\n url = 'https://www.reddit.com/r/' + s\n info = requests.get( url + '/new/.json', headers=headers)\n time.sleep(1.2)\n if 'reason' in info.json():\n ts = info.json()['reason']\n else:\n posts = info.json()['data']['children']\n if len(posts) == 0:\n # Check the regular frontpage instead\n info = requests.get( url + '/.json', headers=headers)\n time.sleep(1.2)\n posts = info.json()['data']['children']\n if len(posts) == 0:\n ts = 'EMPTY'\n else:\n ts = posts[0]['data']['created_utc']\n ts = datetime.utcfromtimestamp(ts)\n ts = ts.replace(tzinfo=from_zone).astimezone(to_zone)\n ts = ts.strftime('%Y/%m/%d/%H:%M:%S')\n else:\n ts = posts[0]['data']['created_utc']\n ts = datetime.utcfromtimestamp(ts)\n ts = ts.replace(tzinfo=from_zone).astimezone(to_zone)\n ts = ts.strftime('%Y/%m/%d/%H:%M:%S')\n row += [ts]\n print(s + ' ' + ts)\n\nwriter.writerow(row)\n\ncsvFile.close()\n","sub_path":"activity_header.py","file_name":"activity_header.py","file_ext":"py","file_size_in_byte":1779,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"503663766","text":"from keras.models import Sequential\r\nfrom keras.layers import Conv2D, MaxPooling2D, Dense, Flatten\r\nfrom keras.preprocessing.image import ImageDataGenerator\r\nimport matplotlib.pyplot as plt\r\nimport sys\r\n\r\nKaggle_PATH = 'D:/bookFiles/Cat_Dog_Kaggle/train/'\r\ndataset_PATH = 'D:/bookFiles/cats_vs_dogs/'\r\n\r\nimage_size = (200,200)\r\n# build a CNN model\r\nmodel = Sequential()\r\nmodel.add(Conv2D(filters=32, kernel_size=(3,3), activation='relu', input_shape=image_size+(3,)))\r\nmodel.add(MaxPooling2D((2,2)))\r\nmodel.add(Conv2D(filters=64, kernel_size=(3,3),activation='relu'))\r\nmodel.add(MaxPooling2D((2,2)))\r\nmodel.add(Flatten())\r\nmodel.add(Dense(128, activation='relu'))\r\nmodel.add(Dense(1, activation='sigmoid'))\r\nmodel.compile(loss='binary_crossentropy', optimizer='adam', metrics=['accuracy'])\r\nprint(model.summary())\r\n\r\ndata = ImageDataGenerator(rescale=1/255.0)\r\ntrain_data = data.flow_from_directory(dataset_PATH+'train/', class_mode='binary', batch_size=64, target_size=(200,200))\r\ntest_data = data.flow_from_directory(dataset_PATH+'test/', class_mode='binary', batch_size=64, target_size=(200,200))\r\nresults = model.fit_generator(train_data, steps_per_epoch=len(train_data), validation_data=test_data,\r\n validation_steps=len(test_data), epochs=2)\r\n_, acc = model.evaluate_generator(test_data, steps=len(test_data), verbose=0)\r\nprint('> %.3f' % (acc * 100.0))","sub_path":"ex 6.7.py","file_name":"ex 6.7.py","file_ext":"py","file_size_in_byte":1388,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"397769861","text":"#!/usr/bin/env python\n\n\"\"\"\nbencode/decode library.\n\nbencoding is used in bittorrent files\n\nuse the exposed functions to encode/decode them.\n\"\"\"\n\nfrom io import BytesIO, SEEK_CUR\ntry: #py 3.3\n\tfrom collections.abc import Iterable, Mapping\nexcept ImportError:\n\tfrom collections import Iterable, Mapping\n\n_TYPE_INT = b'i'\n_TYPE_LIST = b'l'\n_TYPE_DICT = b'd'\n_TYPE_END = b'e'\n_TYPE_SEP = b':'\n_TYPES_STR = b'0123456789'\n\nTYPES = {\n\t_TYPE_INT: int,\n\t_TYPE_LIST: list,\n\t_TYPE_DICT: dict,\n\t_TYPE_END: None,\n\t#_TYPE_SEP only appears in strings, not here\n}\nfor byte in _TYPES_STR:\n\tTYPES[bytes([byte])] = str #b'0': str, b'1': str, …\n\ndef _readuntil(f, end=_TYPE_END):\n\t\"\"\"Helper function to read bytes until a certain end byte is hit\"\"\"\n\tbuf = bytearray()\n\twhile True:\n\t\tbyte = f.read(1)\n\t\tif byte != end:\n\t\t\tbuf += byte\n\t\telse:\n\t\t\tbreak\n\treturn buf\n\ndef _decode_int(f):\n\t\"\"\"\n\tInteger types are normal ascii integers\n\tDelimited at the start with 'i' and the end with 'e'\n\t\"\"\"\n\tassert f.read(1) == _TYPE_INT\n\treturn int(_readuntil(f))\n\ndef _decode_buffer(f):\n\t\"\"\"\n\tString types are normal (byte)strings\n\tstarting with an integer followed by ':'\n\twhich designates the string’s length.\n\t\n\tSince there’s no way to specify the byte type\n\tin bencoded files, we have to guess\n\t\"\"\"\n\tstrlen = int(_readuntil(f, _TYPE_SEP))\n\tbuf = f.read(strlen)\n\ttry:\n\t\treturn buf.decode()\n\texcept UnicodeDecodeError:\n\t\treturn buf\n\ndef _decode_list(f):\n\tassert f.read(1) == _TYPE_LIST\n\tret = []\n\twhile True:\n\t\titem = bdecode(f)\n\t\tif item is None:\n\t\t\tbreak\n\t\telse:\n\t\t\tret.append(item)\n\treturn ret\n\ndef _decode_dict(f):\n\tassert f.read(1) == _TYPE_DICT\n\tret = {}\n\twhile True:\n\t\tkey = bdecode(f)\n\t\tif key is None:\n\t\t\tbreak\n\t\telse:\n\t\t\tassert isinstance(key, (str, bytes))\n\t\t\tret[key] = bdecode(f)\n\treturn ret\n\nDECODERS = {\n\tint: _decode_int,\n\tstr: _decode_buffer,\n\tlist: _decode_list,\n\tdict: _decode_dict,\n}\n\ndef bdecode(f):\n\t\"\"\"\n\tbdecodes data contained in a file f opened in bytes mode.\n\tworks by looking up the type byte,\n\tand using it to look up the respective decoding function,\n\twhich in turn is used to return the decoded object\n\t\"\"\"\n\tbtype = TYPES[f.read(1)]\n\tif btype is not None:\n\t\tf.seek(-1, SEEK_CUR)\n\t\treturn DECODERS[btype](f)\n\telse: #Used in dicts and lists to designate an end\n\t\treturn None\n\ndef bdecode_buffer(data):\n\t\"\"\"Convenience wrapper around bdecode that accepts strings or bytes\"\"\"\n\tif isinstance(data, str):\n\t\tdata = data.encode()\n\twith BytesIO(data) as f:\n\t\treturn bdecode(f)\n\n################\n### Encoding ###\n################\n\ndef _encode_int(integer, f):\n\tf.write(_TYPE_INT)\n\tf.write(str(integer).encode())\n\tf.write(_TYPE_END)\n\ndef _encode_buffer(string, f):\n\t\"\"\"Writes the bencoded form of the input string or bytes\"\"\"\n\tif isinstance(string, str):\n\t\tstring = string.encode()\n\tf.write(str(len(string)).encode())\n\tf.write(_TYPE_SEP)\n\tf.write(string)\n\ndef _encode_iterable(iterable, f):\n\tf.write(_TYPE_LIST)\n\tfor item in iterable:\n\t\tbencode(item, f)\n\tf.write(_TYPE_END)\n\ndef _encode_mapping(mapping, f):\n\tf.write(_TYPE_DICT)\n\tfor key, value in mapping.items():\n\t\t_encode_buffer(key, f)\n\t\tbencode(value, f)\n\tf.write(_TYPE_END)\n\ndef bencode(data, f):\n\t\"\"\"\n\tWrites a serializable data piece to f\n\tThe order of tests is nonarbitrary,\n\tas strings and mappings are iterable.\n\t\"\"\"\n\tif isinstance(data, int):\n\t\t_encode_int(data, f)\n\telif isinstance(data, (str, bytes)):\n\t\t_encode_buffer(data, f)\n\telif isinstance(data, Mapping):\n\t\t_encode_mapping(data, f)\n\telif isinstance(data, Iterable):\n\t\t_encode_iterable(data, f)\n\ndef bencode_buffer(data):\n\t\"\"\"\n\tConvenience wrapper around bencode that returns a byte array\n\tof the serialized sata\n\t\"\"\"\n\twith BytesIO() as f:\n\t\tbencode(data, f)\n\t\treturn f.getvalue()\n\ndef main():\n\timport sys, pprint\n\tfrom argparse import ArgumentParser, FileType\n\tparser = ArgumentParser(description='Decodes bencoded files to python objects.')\n\tparser.add_argument('infile', nargs='?', type=FileType('rb'), default=sys.stdin.buffer,\n\t\thelp='bencoded file (e.g. torrent) [Default: stdin]')\n\tparser.add_argument('outfile', nargs='?', type=FileType('w'), default=sys.stdout,\n\t\thelp='python-syntax serialization [Default: stdout]')\n\targs = parser.parse_args()\n\t\n\tdata = bdecode(args.infile)\n\tpprint.pprint(data, stream=args.outfile)\n\nif __name__ == '__main__':\n\tmain()\n","sub_path":"bcode.py","file_name":"bcode.py","file_ext":"py","file_size_in_byte":4285,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"273038807","text":"# -*- coding: utf-8 -*-\n##############################################################################\n#\n# OpenERP, Open Source Management Solution\n# Copyright (C) 2012 Hugo Santos ().\n#\n# This program is free software: you can redistribute it and/or modify\n# it under the terms of the GNU Affero General Public License as\n# published by the Free Software Foundation, either version 3 of the\n# License, or (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU Affero General Public License for more details.\n#\n# You should have received a copy of the GNU Affero General Public License\n# along with this program. If not, see .\n#\n##############################################################################\n\nfrom osv import osv, fields\nfrom tools.translate import _\n\nclass sale_order(osv.osv):\n _inherit = 'sale.order'\n\n def onchange_partner_id(self, cr, uid, ids, part):\n result = super(sale_order,self).onchange_partner_id(cr, uid, ids, part)\n if part:\n partner = self.pool.get('res.partner').browse(cr, uid, part)\n if partner.company_credit_limit == 0.0:\n result['warning'] = {}\n if partner.company_credit_limit != 0.0 and partner.available_risk < 0.0:\n result['warning'] = {\n 'title': _('Credit Limit Exceeded'),\n 'message': _('Warning: Credit Limit Exceeded.\\n\\nThis partner has a credit limit of %(limit).2f and already has a debt of %(debt).2f.') % {\n 'limit': partner.credit_limit,\n 'debt': partner.total_debt,\n }\n }\n return result\n\nsale_order()","sub_path":"fl_auto_risk/sale.py","file_name":"sale.py","file_ext":"py","file_size_in_byte":1937,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"639328827","text":"\r\nimport os\r\nimport subprocess\r\n\r\nfrom config import MEDIA_DIR, TREE_DIR, FASTA_DIR\r\n\r\n\r\ndef tree_creator(selectname):\r\n \"\"\"\r\n DESCRIPTION:\r\n A function create the tree inference and store the results in a subfolder within covid_phylo/tree/\r\n :param selectname: [string] name of the file to be put in the tree folder. The same as the name of the subfolder.\r\n :return: None\r\n \"\"\"\r\n print('Executing tree inference')\r\n filename = selectname\r\n subfolder = selectname.split('.')[0]\r\n process = subprocess.run(['iqtree', '-s', f'{TREE_DIR / subfolder / filename}', '-bnni', '-nt', 'AUTO'])\r\n print('Tree inference completed with exit code %d' % process.returncode)\r\n\r\n\r\ndef align_selector(origname, destname, n_genomes):\r\n \"\"\"\r\n DESCRIPTION:\r\n Function to select the n alignments with the lowest number of gaps.\r\n :param origname: [string] name of the file with the complete list of alignments in the fasta folder.\r\n :param destname: [string] name of the file to be put in the tree folder. The same as the name of the subfolder.\r\n :param n_genomes: [integer] number of alignments to be taken.\r\n :return: None. It writes the selected aignments in the destname folder.\r\n \"\"\"\r\n # Part to take the alignments with lowest number of gaps\r\n file = open(FASTA_DIR / origname, 'r')\r\n\r\n # Read the information\r\n data = file.read()\r\n file.close()\r\n data = data.split('>')[1::] # The first element is an empty string\r\n\r\n # Take the best n models\r\n gaps = list(enumerate([model.count('-') for model in data]))\r\n data = '\\n'.join(['>' + data[element[0]] for element in sorted(gaps, key=lambda x: x[1])[0:n_genomes]])\r\n\r\n # Write the selected data into another file\r\n sel_dir = TREE_DIR / destname.split('.')[0]\r\n sel_dir.mkdir(exist_ok=True)\r\n file = open(sel_dir / destname, 'w')\r\n file.write(data)\r\n file.close()\r\n\r\n","sub_path":"src/iqtree.py","file_name":"iqtree.py","file_ext":"py","file_size_in_byte":1902,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"76631679","text":"from mezzanine.conf import settings\nfrom cartridge.shop.checkout import default_billship_handler\nfrom cartridge.shop.models import Product, ProductVariation\nfrom cartridge.shop.utils import set_shipping\n\nfrom .models import Purchase\nfrom .utils import transact\n\n\nDOWNLOAD_ONLY_OPTION = False\nfor option_value, option_name in settings.SHOP_OPTION_TYPE_CHOICES:\n if option_name == 'Downloads':\n DOWNLOAD_ONLY_OPTION = {'option' + str(option_value): 'Download Only'}\n\n\ndef billship_handler(request, order_form):\n \"\"\"\n If product is all downloads, do not set shipping (defaults to free).\n \"\"\"\n request.is_download_only = (\n not ProductVariation.objects\n .filter(sku__in=request.cart.skus())\n .exclude(**DOWNLOAD_ONLY_OPTION)\n .exists()\n if DOWNLOAD_ONLY_OPTION else False)\n\n if request.is_download_only:\n set_shipping(request, \"Free shipping\", 0)\n else:\n default_billship_handler(request, order_form)\n\n\ndef order_handler(request, order_form, order):\n skus = request.cart.skus()\n variations = ProductVariation.objects.filter(sku__in=skus)\n\n download_products = (\n Product.objects\n .filter(variations__in=variations)\n .exclude(downloads=None))\n\n if download_products.exists():\n # Initialize transaction and credentials.\n transaction = transact(request)\n\n # Associate downloads with transaction.\n for product in download_products:\n for download in product.downloads.all():\n purchase = Purchase(\n download=download,\n transaction=transaction,\n order=order,\n product=product)\n purchase.save()\n\n # If order is all downloads, mark it as processed.\n if (request.is_download_only and\n settings.SHOP_ORDER_STATUS_CHOICES[1] == (2, 'Processed')):\n order.status = 2\n order.save()\n","sub_path":"cartridge_downloads/checkout.py","file_name":"checkout.py","file_ext":"py","file_size_in_byte":1962,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"641407402","text":"# sqlite\n# connecting, creating cursor\n# getting a table (employees)\n# close the connection\n\nimport sqlite3\n\ncon = sqlite3.connect('chinook.db')\ncur = con.cursor()\n\nquery = 'SELECT * FROM employees'\n\ncur.execute(query)\n\nlista = cur.fetchall()\n\nfor record in lista:\n print(record)\n print(\"---------\")\n\ncon.close()\n","sub_path":"day5/sql/2.py","file_name":"2.py","file_ext":"py","file_size_in_byte":319,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"605801908","text":"# guess the game\r\nimport random\r\n\r\nprint(\"Enter your name :\")\r\nname = input()\r\n\r\nprint('Hello, ' + name + ', I am thinking a number between 1 and 20')\r\nsnumber = random.randint(1,20)\r\nx = bin(snumber)\r\nprint('DEBUG: the secret number is : ' + str(x ))\r\n\r\nfor guessTaken in range(1,6):\r\n print(\"Take a guess \")\r\n guess = int(input())\r\n\r\n if guess < snumber:\r\n print(\"Your number is very low, guess high\")\r\n elif guess > snumber:\r\n print(\"Your number is very high, guess low\")\r\n else:\r\n break\r\n\r\nif guess == snumber:\r\n print('Good job, ' + name + ' ! you guessed right number')\r\nelse:\r\n print('Nope, The number I was thinking of was ' + str(snumber))\r\n \r\n","sub_path":"guess-the-number.py","file_name":"guess-the-number.py","file_ext":"py","file_size_in_byte":699,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"248594484","text":"import argparse\nimport logging.config\nimport sys\n\nfrom common import check_positive\nfrom common import compare_two_values\nfrom common import logging_decorator\n\nlogging.config.fileConfig('logging.conf', disable_existing_loggers=False, defaults={'logfilename': 'fibonacci.log'})\nlogger = logging.getLogger('MainFormatter')\n\n\n@logging_decorator\nclass FibonacciSequence:\n def __init__(self, begin=0, end=100, *args, **kwargs):\n '''\n :argument -b: beginning of slice :type: integer / default = 0\n :argument -e: ending of slice :type: integer / default = 100\n :return : help\n '''\n self.begin = begin\n self.end = end\n self.fib_list = []\n self.fib_generate_list(self.begin, self.end)\n\n def __str__(self):\n return ', '.join(map(str, self.fib_list))\n\n @compare_two_values\n @check_positive\n @logging_decorator\n def fib_generate_list(self, b, e):\n '''Check FibonacciNumbers number for being in (a,b) range'''\n for n in self.fib_generate_numbers():\n if b <= n and n <= e:\n self.fib_list.append(n)\n return self.fib_list\n\n @logging_decorator\n def fib_generate_numbers(self):\n '''Fibonacci numbers generator'''\n i = 0\n j = 1\n while i <= self.end:\n yield i\n i, j = j, i + j\n\n\n@logging_decorator\ndef initialize_args():\n '''Initialization arguments and return values, that user input'''\n global parser\n parser = argparse.ArgumentParser(add_help=False, prog='Fibonacci sequence in range:',\n usage='%(prog)s [b] and [e] arguments for '\n '[begin,end] range of Fibonacci numbers')\n parser.add_argument('-h', '--help', action='help', help='Show help message')\n group = parser.add_argument_group('required arguments', 'Change this arg to ''set your range')\n group.add_argument('-b', '--begin', type=int, default=0)\n group.add_argument('-e', '--end', type=int, default=100)\n args = parser.parse_args(sys.argv[1:])\n return vars(args)\n\n\ndef main():\n '''Main function'''\n args = initialize_args() # args initialization\n if len(sys.argv) == 1:\n parser.print_help() # display 'help' message from argparse\n else:\n f = FibonacciSequence(args['begin'], args['end'])\n print(f)\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"FibonacciNumbers/Fibonacci.py","file_name":"Fibonacci.py","file_ext":"py","file_size_in_byte":2429,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"604449336","text":"# -*- coding: utf-8 -*-\n# ---\n# jupyter:\n# jupytext:\n# cell_metadata_filter: ExecuteTime,autoscroll,heading_collapsed,hidden,slideshow,-hide_ouput,-code_folding\n# cell_metadata_json: true\n# formats: ipynb,py:percent\n# notebook_metadata_filter: all\n# text_representation:\n# extension: .py\n# format_name: percent\n# format_version: '1.2'\n# jupytext_version: 1.2.1\n# kernelspec:\n# display_name: Python 3\n# language: python\n# name: python3\n# language_info:\n# codemirror_mode:\n# name: ipython\n# version: 3\n# file_extension: .py\n# mimetype: text/x-python\n# name: python\n# nbconvert_exporter: python\n# pygments_lexer: ipython3\n# version: 3.6.9\n# latex_envs:\n# LaTeX_envs_menu_present: true\n# autoclose: false\n# autocomplete: false\n# bibliofile: biblio.bib\n# cite_by: apalike\n# current_citInitial: 1\n# eqLabelWithNumbers: true\n# eqNumInitial: 1\n# hotkeys:\n# equation: Ctrl-E\n# itemize: Ctrl-I\n# labels_anchors: false\n# latex_user_defs: false\n# report_style_numbering: false\n# user_envs_cfg: false\n# ---\n\n# %% [markdown]\n# # Theoretical Foundations of Buffer Stock Saving\n#\n# \n#\n#

Generator: BufferStockTheory-make/notebooks_byname

\n#\n# [![econ-ark.org](https://img.shields.io/badge/Powered%20by-Econ--ARK-3e8acc.svg)](https://econ-ark.org/materials/BufferStockTheory)\n#\n\n# %% [markdown]\n# \n#\n# [This notebook](https://econ-ark.org/BufferStockTheory/#launch) uses the [Econ-ARK/HARK](https://github.com/econ-ark/HARK) toolkit to reproduce and illustrate key results of the paper [Theoretical Foundations of Buffer Stock Saving](http://econ-ark.github.io/BufferStockTheory/BufferStockTheory).\n#\n# #### An [interactive dashboard](https://econ-ark.org/BufferStockStockTheory/#Dashboard) allows you to modify parameters to see how the figures change. \n#\n\n# %%\n# This cell does some standard python setup\n\n# Import related generic python packages\nimport numpy as np\nfrom copy import deepcopy\n\n# Plotting tools\nimport matplotlib.pyplot as plt\n\n# The warnings package allows us to ignore some harmless but alarming warning messages\nimport warnings\nwarnings.filterwarnings(\"ignore\")\n\n# Code to allow a master \"Generator\" and derived \"Generated\" versions\n# - allows \"$nb-Problems-And-Solutions → $nb-Problems → $nb\"\nGenerator=False # Is this notebook the master or is it generated?\n\n# Whether to save the figures to Figures_dir\nsaveFigs=True\n\n# Whether to draw the figures\ndrawFigs=True\n\nimport HARK\nif HARK.__version__ < '0.10.6':\n raise ImportError('This notebook requires at least econ-ark v0.10.6, please update your installation pip install -U econ-ark or conda install -c conda-forge econ-ark')\n\nfrom HARK.utilities import find_gui, make_figs, determine_platform, test_latex_installation, setup_latex_env_notebook\npf = determine_platform()\ntry:\n latexExists = test_latex_installation(pf)\nexcept ImportError: # windows and MacOS requires manual install\n latexExists = False\n\nsetup_latex_env_notebook(pf, latexExists)\n\n# check if GUI is present if not then switch drawFigs to False and force saveFigs to be True\nif not find_gui():\n drawFigs = False\n saveFigs = True\n\n# this can be removed if we pass in saveFigs and drawFigs in every call to make('figure')\ndef make(figure_name, target_dir=\"../../Figures\"):\n make_figs(figure_name, saveFigs, drawFigs, target_dir)\n\n# %%\n# Import HARK tools\nfrom HARK.ConsumptionSaving.ConsIndShockModel import IndShockConsumerType\nfrom HARK.utilities import plotFuncsDer, plotFuncs\n\n\n# %% [markdown]\n# ## [The Problem](http://econ-ark.github.io/BufferStockTheory/BufferStockTheory/#The-Problem)\n#\n# The paper defines and calibrates a small set of parameters:\n#\n# | Parameter | Description | Code | Value |\n# |:---:| --- | --- | :---: |\n# | $\\Gamma$ | Permanent Income Growth Factor | $\\texttt{PermGroFac}$ | 1.03 |\n# | $\\mathsf{R}$ | Interest Factor | $\\texttt{Rfree}$ | 1.04 |\n# | $\\beta$ | Time Preference Factor | $\\texttt{DiscFac}$ | 0.96 |\n# | $\\rho$ | Coefficient of Relative Risk Aversion| $\\texttt{CRRA}$ | 2 |\n# | $\\wp$ | Probability of Unemployment | $\\texttt{UnempPrb}$ | 0.005 |\n# | $\\theta^{\\large u}$ | Income when Unemployed | $\\texttt{IncUnemp}$ | 0. |\n# | $\\sigma_\\psi$ | Std Dev of Log Permanent Shock| $\\texttt{PermShkStd}$ | 0.1 |\n# | $\\sigma_\\theta$ | Std Dev of Log Transitory Shock| $\\texttt{TranShkStd}$ | 0.1 |\n#\n# that define the preferences and environment of microeconomic consumers as detailed below. \n#\n# The objective of such a consumer with a horizon of $n$ periods is to maximize the value obtained from the stream of consumption __**c**__ from period $t=T-n$ to a terminal period $T$:\n#\n# \\begin{equation}\n# \\mathbf{v}_{t} = \\sum_{i=0}^{n} \\beta^{n}\\mathrm{u}(\\mathbf{c}_{t+n}) \n# \\end{equation}\n#\n# The infinite-horizon solution to the model is defined as the limit of the solution in the first period of life $\\mathrm{c}_{T-n}$ as the horizon $n$ goes to infinity.\n\n# %% [markdown]\n# ### Details\n# For a microeconomic consumer who begins period $t$ with __**m**__arket resources boldface $\\mathbf{m}_{t}$ (=net worth plus current income), the amount that remains after __**c**__onsumption of $\\mathbf{c}_{t}$ will be end-of-period __**a**__ssets $\\mathbf{a}_{t}$, \n#\n# \n\n# %% [markdown]\n# \\begin{eqnarray}\n# \\mathbf{a}_{t} &=&\\mathbf{m}_{t}-\\mathbf{c}_{t}. \\notag \n# \\end{eqnarray}\n#\n# The consumer's __**p**__ermanent noncapital income $\\mathbf{p}$ grows by a predictable factor $\\Gamma$ and is subject to an unpredictable multiplicative shock $\\mathbb{E}_{t}[\\psi_{t+1}]=1$,\n#\n# \\begin{eqnarray}\n# \\mathbf{p}_{t+1} & = & \\mathbf{p}_{t} \\Gamma \\psi_{t+1}, \\notag \n# \\end{eqnarray}\n#\n# and, if the consumer is employed, actual income is permanent income multiplied by a transitory shock $\\theta^{\\large e}$. There is also a probability $\\wp$ that the consumer will be temporarily unemployed and experience income of $\\theta^{\\large u} = 0$. We construct $\\theta^{\\large e}$ so that its mean value is $1/(1-\\wp)$ because in that case the mean level of the transitory shock (accounting for both unemployed and employed states) is exactly \n#\n# \\begin{eqnarray}\n# \\mathbb{E}_{t}[\\theta_{t+1}] & = & \\theta^{\\large{u}} \\times \\wp + (1-\\wp) \\times \\mathbb{E}_{t}[\\theta^{\\large{e}}_{t+1}] \\notag\n# \\\\ & = & 0 \\times \\wp + (1-\\wp) \\times 1/(1-\\wp) \\notag\n# \\\\ & = & 1. \\notag\n# \\end{eqnarray}\n#\n# We can combine the unemployment shock $\\theta^{\\large u}$ and the transitory shock to employment income $\\theta^{\\large e}$ into $\\theta _{t+1}$, so that next period's market resources are\n# \\begin{eqnarray}\n# \\mathbf{m}_{t+1} &=& \\mathbf{a}_{t}\\mathsf{R} +\\mathbf{p}_{t+1}\\theta_{t+1}. \\notag\n# \\end{eqnarray}\n\n# %% [markdown]\n# When the consumer has a CRRA utility function $u(\\mathbf{c})=\\frac{\\mathbf{c}^{1-\\rho}}{1-\\rho}$, the paper shows that the problem can be written in terms of ratios (nonbold font) of level (bold font) variables to permanent income, e.g. $m_{t} \\equiv \\mathbf{m}_{t}/\\mathbf{p}_{t}$, and the Bellman form of [the problem reduces to](https://econ-ark.github.io/BufferStockTheory/#The-Related-Problem):\n#\n# \\begin{eqnarray*}\n# v_t(m_t) &=& \\max_{c_t}~~ u(c_t) + \\beta~\\mathbb{E}_{t} [(\\Gamma\\psi_{t+1})^{1-\\rho} v_{t+1}(m_{t+1}) ] \\\\\n# & s.t. & \\\\\n# a_t &=& m_t - c_t \\\\\n# m_{t+1} &=& a_t \\mathsf{R}/(\\Gamma \\psi_{t+1}) + \\theta_{t+1} \\\\\n# \\end{eqnarray*}\n\n# %%\n# Define a dictionary with baseline parameter values\n\n# Import default parameter values (init_idiosyncratic_shock)\nfrom HARK.ConsumptionSaving.ConsIndShockModel import init_idiosyncratic_shocks as base_params\n\n# Set the parameters for the baseline results in the paper\n# using the variable names defined in the cell above\nbase_params['PermGroFac'] = [1.03] # Permanent income growth factor\nbase_params['Rfree'] = Rfree = 1.04 # Interest factor on assets\nbase_params['DiscFac'] = DiscFac = 0.96 # Time Preference Factor\nbase_params['CRRA'] = CRRA = 2.00 # Coefficient of relative risk aversion\nbase_params['UnempPrb'] = UnempPrb = 0.005 # Probability of unemployment (e.g. Probability of Zero Income in the paper)\nbase_params['IncUnemp'] = IncUnemp = 0.0 # Induces natural borrowing constraint\nbase_params['PermShkStd'] = [0.1] # Standard deviation of log permanent income shocks\nbase_params['TranShkStd'] = [0.1] # Standard deviation of log transitory income shocks\n# %%\n# Uninteresting housekeeping and details\n# Make global variables for the things that were lists above -- uninteresting housekeeping\nPermGroFac, PermShkStd, TranShkStd = base_params['PermGroFac'][0],base_params['PermShkStd'][0],base_params['TranShkStd'][0]\n\n# Some technical settings that are not interesting for our purposes\nbase_params['LivPrb'] = [1.0] # 100 percent probability of living to next period\nbase_params['CubicBool'] = True # Use cubic spline interpolation\nbase_params['T_cycle'] = 1 # No 'seasonal' cycles\nbase_params['BoroCnstArt'] = None # No artificial borrowing constraint\n# %% [markdown]\n# ## Convergence of the Consumption Rules\n#\n# Under the given parameter values, [the paper's first figure](https://econ-ark.github.io/BufferStockTheory/#Convergence-of-the-Consumption-Rules) depicts the successive consumption rules that apply in the last period of life $(c_{T}(m))$, the second-to-last period, and earlier periods $(c_{T-n})$. The consumption function to which these converge is $c(m)$:\n#\n# $$\n# c(m) = \\lim_{n \\uparrow \\infty} c_{T-n}(m) \\notag\n# $$\n#\n\n# %%\n# Create a buffer stock consumer instance by invoking the IndShockConsumerType class\n# with the built-in parameter dictionary \"base_params\"\n\n# Construct finite horizon agent with baseline parameters\nbaseAgent_Fin = IndShockConsumerType(**base_params)\nbaseAgent_Fin.cycles = 100 # Set finite horizon (T = 100)\n\nbaseAgent_Fin.solve() # Solve the model\nbaseAgent_Fin.unpack('cFunc') # Make the consumption function easily accessible\n\n\n# %%\n# Plot the different consumption rules for the different periods\n\nmPlotMin = 0\nmLocCLabels = 9.6 # Defines horizontal limit of figure\nmPlotTop = 6.5 # Defines maximum m value where functions are plotted\nmPts = 1000 # Number of points at which functions are evaluated\n\nmBelwLabels = np.linspace(mPlotMin,mLocCLabels-0.1,mPts) # Range of m below loc of labels\nm_FullRange = np.linspace(mPlotMin,mPlotTop,mPts) # Full plot range \nc_Tm0 = m_FullRange # c_Tm0 defines the last period consumption rule (c=m)\nc_Tm1 = baseAgent_Fin.cFunc[ -2](mBelwLabels) # c_Tm1 defines the second-to-last period consumption rule\nc_Tm5 = baseAgent_Fin.cFunc[ -6](mBelwLabels) # c_Tm5 defines the T-5 period consumption rule\nc_Tm10 = baseAgent_Fin.cFunc[-11](mBelwLabels) # c_Tm10 defines the T-10 period consumption rule\nc_Limt = baseAgent_Fin.cFunc[ 0](mBelwLabels) # c_Limt defines limiting infinite-horizon consumption rule\nplt.figure(figsize = (12,9))\nplt.plot(mBelwLabels,c_Limt,color=\"black\")\nplt.plot(mBelwLabels,c_Tm1 ,color=\"black\")\nplt.plot(mBelwLabels,c_Tm5 ,color=\"black\")\nplt.plot(mBelwLabels,c_Tm10,color=\"black\")\nplt.plot(m_FullRange,c_Tm0 ,color=\"black\")\nplt.xlim(0,11)\nplt.ylim(0,7)\nplt.text(7.0,6.0,r'$c_{T }(m) = 45$ degree line',fontsize = 22,fontweight='bold')\nplt.text(mLocCLabels,5.3,r'$c_{T-1 }(m)$',fontsize = 22,fontweight='bold')\nplt.text(mLocCLabels,2.6,r'$c_{T-5 }(m)$',fontsize = 22,fontweight='bold')\nplt.text(mLocCLabels,2.1,r'$c_{T-10}(m)$',fontsize = 22,fontweight='bold')\nplt.text(mLocCLabels,1.7,r'$c(m) $',fontsize = 22,fontweight='bold')\nplt.arrow(6.9,6.05,-0.6,0,head_width= 0.1,width=0.001,facecolor='black',length_includes_head='True')\nplt.tick_params(labelbottom=False, labelleft=False,left='off',right='off',bottom='off',top='off')\nplt.text(0,7.05,\"$c$\",fontsize = 26)\nplt.text(11.1,0,\"$m$\",fontsize = 26)\n# Save the figures in several formats\n\nmake('cFuncsConverge') # Comment out if you want to run uninterrupted\n\n# %% [markdown] {\"slideshow\": {\"slide_type\": \"slide\"}}\n# Use the [interactive dashboard](#interactive-dashboard) to explore the effects of changes in patience, risk aversion, or risk\n\n# %% [markdown] {\"heading_collapsed\": true}\n# ### PROBLEM: Natural Borrowing Constraint Approaches Artificial Constraint\n#\n# Show numerically the result that is proven analytically in [The-Liquidity-Constrained-Solution-as-a-Limit](https://econ-ark.github.io/BufferStockTheory/#The-Liquidity-Constrained-Solution-as-a-Limit), by solving the model for successively smaller values of $\\wp$.\n# * You need only to solve for the second-to-last period of life to do this\n# * `TwoPeriodModel = IndShockConsumerType(**base_params)`\n# * `TwoPeriodModel.cycles = 2 # Make this type have a two period horizon (Set T = 2)`\n#\n# * You should show the consumption rules for different values of $\\wp$ on the same graph\n# * To make this easier, you will want to use the plotFuncs command:\n# * `from HARK.utilities import plotFuncsDer, plotFuncs`\n#\n# Create a cell or cells in the notebook below this cell and put your solution there; comment on the size of $\\wp$ needed to make the two models visually indistinguishable\n\n# %% [markdown]\n# ## Factors and Conditions\n#\n# ### [The Finite Human Wealth Condition](http://econ-ark.github.io/BufferStockTheory/#Human-Wealth)\n#\n# Human wealth for a perfect foresight consumer is the present discounted value of future income:\n#\n# \\begin{eqnarray}\\notag\n# \\mathbf{h}_{t} & = & \\mathbb{E}_{t}[\\mathbf{p}_{t} + \\mathsf{R}^{-1} \\mathbf{p}_{t+1} + \\mathsf{R}^{2} \\mathbf{p}_{t+2} ... ] \\\\ \\notag \n# & = & \\mathbf{p}_{t} \\left(1 + (\\Gamma/\\mathsf{R}) + (\\Gamma/\\mathsf{R})^{2} ... \\right) \n# \\end{eqnarray}\n# which approaches infinity as the horizon extends if $\\Gamma/\\mathsf{R} \\geq 1$. We say that the 'Finite Human Wealth Condition' [(FHWC)](https://econ-ark.github.io/BufferStockTheory/#FHWC) holds if\n# $0 \\leq (\\Gamma/\\mathsf{R}) < 1$.\n\n# %% [markdown]\n# ### [Absolute Patience and the AIC](https://econ-ark.github.io/BufferStockTheory/#AIC)\n#\n# The paper defines the Absolute Patience Factor [(APF)](https://econ-ark.github.io/BufferStockTheory/#APF) as being equal to the ratio $\\mathbf{c}_{t+1}/\\mathbf{c}_{t}$ for a perfect foresight consumer. The Old English character \"Þ\" used for this object in the paper cannot currently be rendered conveniently in Jupyter notebooks, so we will substitute $\\Phi$ here:\n#\n# \\begin{equation}\n# \\Phi = (\\mathsf{R} \\beta)^{1/\\rho}\n# \\end{equation}\n#\n# If $\\Phi = 1$, a perfect foresight consumer will spend at exactly the level of $\\mathbf{c}$ that can be sustained perpetually (given their current and future resources). If $\\Phi < 1$ (the consumer is 'absolutely impatient'; or, 'the absolute impatience condition holds'), the consumer is consuming more than the sustainable amount, so consumption will fall, and if the consumer is 'absolutely patient' with $\\Phi > 1$ consumption will grow over time.\n#\n#\n\n# %% [markdown]\n# ### [Growth Patience and the GIC](https://econ-ark.github.io/BufferStockTheory/#GIC)\n#\n# For a [perfect foresight consumer](http://econ.jhu.edu/people/ccarroll/public/lecturenotes/consumption/PerfForesightCRRA), whether the ratio $c$=__**c**__/__**p**__ is rising, constant, or falling depends on the relative growth rates of consumption and permanent income; that ratio is measured by the [Perfect Foresight Growth Patience Factor](https://econ-ark.github.io/BufferStockTheory/#PFGPF):\n#\n# \\begin{eqnarray}\n# \\Phi_{\\Gamma} & = & \\Phi/\\Gamma\n# \\end{eqnarray}\n# and whether the $c$ is falling or rising over time depends on whether $\\Phi_{\\Gamma}$ is below or above 1.\n#\n# An analogous condition can be defined when there is uncertainty about permanent income. Defining $\\tilde{\\Gamma} = (\\mathbb{E}[\\psi^{-1}])^{-1}\\Gamma$, the 'Growth Impatience Condition' [(GIC)](https://econ-ark.github.io/BufferStockTheory/#GIC) determines whether, \\textit{in expectation}, the stochastic value of $c$ is rising, constant, or falling over time:\n# \\begin{eqnarray}\n# \\Phi/\\tilde{\\Gamma} & < & 1\n# \\end{eqnarray}\n\n# %% [markdown]\n# ### [The Finite Value of Autarky Condition (FVAC)](https://econ-ark.github.io/BufferStockTheory/#Autarky-Value)\n\n\n# %% [markdown]\n# The paper [shows](https://econ-ark.github.io/BufferStockTheory/#Autarky-Value) that a consumer who planned to spend his permanent income $\\{ \\mathbf{p}_{t}, \\mathbf{p}_{t+1}, ...\\} $ in every period would have value defined by\n#\n# \\begin{equation*}\n# \\mathbf{v}_{t}^{\\text{autarky}} = u(\\mathbf{p}_{t})\\left(\\frac{1}{1-\\beta \\Gamma^{1-\\rho} \\mathbb{E}[\\psi^{1-\\rho}]}\\right)\n# \\end{equation*}\n#\n# and defines the 'Finite Value of Autarky Condition' as the requirement that the denominator be a positive finite number:\n#\n# \\begin{equation*}\n# \\beta \\Gamma^{1-\\rho} \\mathbb{E}[\\psi^{1-\\rho}] < 1\n# \\end{equation*}\n\n# %% [markdown]\n# ### [The Weak Return Impatience Condition (WRIC)](https://econ-ark.github.io/BufferStockTheory/#WRIC)\n#\n# The [Return Impatience Condition](https://econ-ark.github.io/BufferStockTheory/#RIC) $\\Phi/\\mathsf{R} < 1$ has long been understood to be required for the perfect foresight model to have a nondegenerate solution (a common special case is when $\\rho=1$; in this case $\\Phi = \\mathsf{R} \\beta$ so $\\Phi<1$ reduces to $\\beta < \\mathsf{R}$). \n#\n# If the RIC does not hold, the consumer is so patient that the optimal consumption function approaches zero as the horizon extends indefinitely.\n#\n# When the probability of unemployment is $\\wp$, the paper articulates an analogous (but weaker) condition:\n#\n# \\begin{eqnarray}\n# \\wp^{1/\\rho} \\Phi/\\mathsf{R} & < & 1\n# \\end{eqnarray}\n\n# %% [markdown]\n# # Key Results\n#\n# ## [Nondegenerate Solution Requires FVAC and WRIC](https://econ-ark.github.io/BufferStockTheory/#Sufficient-Conditions-For-Nondegenerate-Solution)\n#\n# A main result of the paper is that the conditions required for the model to have a nondegenerate solution ($0 < c(m) < \\infty$ for feasible $m$) are that the Finite Value of Autarky (FVAC) and Weak Return Impatience Condition (WRIC) hold.\n\n# %% [markdown]\n# ## [Natural Borrowing Constraint limits to Artificial Borrowing Constraint](https://econ-ark.github.io/BufferStockTheory/#The-Liquidity-Constrained-Solution-as-a-Limit)\n\n# %% [markdown]\n# Defining $\\chi(\\wp)$ as the consumption function associated with any particular value of $\\wp$, and defining $\\hat{\\chi}$ as the consumption function that would apply in the absence of the zero-income shocks but in the presence of an 'artificial' borrowing constraint requiring $a \\geq 0$ (_a la_ Deaton (1991)), the paper shows that\n#\n# \\begin{eqnarray}\n# \\lim_{\\wp \\downarrow 0}~\\chi(\\wp) & = & \\hat{\\chi}\n# \\end{eqnarray}\n#\n# That is, as $\\wp$ approaches zero the problem with uncertainty becomes identical to the problem that instead has constraints. (See [Precautionary Saving and Liquidity Constraints](https://econ-ark.github.io/LiqConstr) for a full treatment of the relationship between precautionary saving and liquidity constraints).\n\n# %% [markdown]\n# ## [$c(m)$ can be Finite Even When Human Wealth Is Infinite](https://econ-ark.github.io/BufferStockTheory/#When-The-GIC-Fails)\n#\n# In the perfect foresight model, if $\\mathsf{R} < \\Gamma$ the PDV of future labor income approaches infinity and so the limiting consumption function is $c(m) = \\infty$ for all $m$. Many models have no well-defined solution when human wealth is infinite.\n#\n# The presence of uncertainty changes this: Even when limiting human wealth is infinite, the limiting consumption function is finite for all values of $m$.\n#\n# This is because uncertainty imposes a \"natural borrowing constraint\" that deters the consumer from borrowing against their unbounded (but uncertain) future labor income.\n\n# %% [markdown]\n# A [table](https://econ-ark.github.io/BufferStockTheory/#Sufficient-Conditions-For-Nondegenerate-Solution) puts this result in the context of implications of other conditions and restrictions.\n#\n#\n\n# %% [markdown]\n# ## [If the GIC Holds, $\\exists$ a finite 'target' $m$](https://econ-ark.github.io/BufferStockTheory/#onetarget)\n#\n# Section [There Is Exactly One Target $m$ Ratio, Which Is Stable](https://econ-ark.github.io/BufferStockTheory/#onetarget) shows that, under parameter values for which the limiting consumption function exists, if the GIC holds then there will be a value $\\check{m}$ such that:\n#\n# \\begin{eqnarray*}\n# \\mathbb{E}[m_{t+1}] & > & m_{t}~\\text{if $m_{t} < \\check{m}$} \\\\\n# \\mathbb{E}[m_{t+1}] & < & m_{t}~\\text{if $m_{t} > \\check{m}$} \\\\\n# \\mathbb{E}[m_{t+1}] & = & m_{t}~\\text{if $m_{t} = \\check{m}$}\n# \\end{eqnarray*}\n\n# %% [markdown]\n# ## [If the GIC Fails, Target Wealth is Infinite ](https://econ-ark.github.io/BufferStockTheory/#The-GIC)\n#\n# [A figure](https://econ-ark.github.io/BufferStockTheory/#FVACnotGIC) depicts a solution when the **FVAC** (Finite Value of Autarky Condition) and **WRIC** hold (so that the model has a solution) but the **GIC** (Growth Impatience Condition) fails. In this case the target wealth ratio is infinity.\n#\n# The parameter values in this specific example are:\n#\n# | Param | Description | Code | Value |\n# | :---: | --- | --- | :---: |\n# | $\\Gamma$ | Permanent Income Growth Factor | $\\texttt{PermGroFac}$ | 1.00 |\n# | $\\mathrm{\\mathsf{R}}$ | Interest Factor | $\\texttt{Rfree}$ | 1.04 |\n#\n\n# %%\n# Construct the \"GIC fails\" example.\n\nGIC_fails_dictionary = dict(base_params)\nGIC_fails_dictionary['Rfree'] = 1.04\nGIC_fails_dictionary['PermGroFac'] = [1.00]\n\nGICFailsExample = IndShockConsumerType(\n cycles=0, # cycles=0 makes this an infinite horizon consumer\n verbose=0, # by deafult, check conditions shouldn't print out any information\n **GIC_fails_dictionary)\n\n\n# %% [markdown]\n# The $\\mathtt{IndShockConsumerType}$ tool automatically checks various parametric conditions, and will give a warning as well as the values of the factors if key conditions fail to be met.\n#\n# We can also directly check the conditions, asking for the maximum verbosity:\n\n# %%\n# The checkConditions method does what it sounds like it would\n# verbose=0: Print nothing;\n# verbose=3: Print all available info\nGICFailsExample.checkConditions(verbose=0)\n\n# %% [markdown]\n# ### The Sustainable Level of Consumption\n#\n# Next we define the $\\mathrm{\\mathbb{E}}_{t}[\\Delta m_{t+1}]=0$ locus that shows the ‘sustainable’ level of spending at which $m$ is expected to remain unchanged.\n\n# %%\n# Calculate \"Sustainable\" consumption that leaves expected m unchanged\n# In the perfect foresight case, this is just permanent income plus interest income\n# A small adjustment is required to take account of the consequences of uncertainty\n# See \"Growth Patience and the GIC\" above\n\n# Get calibrated parameters to make code more readable\nLivPrb=baseAgent_Fin.LivPrb[0]\nRfree=baseAgent_Fin.Rfree\nDiscFac=baseAgent_Fin.DiscFac\nCRRA=baseAgent_Fin.CRRA\n\npermShkPrbs=GICFailsExample.PermShkDstn[0].pmf\npermShkVals=GICFailsExample.PermShkDstn[0].X\nEPermGroFac=GICFailsExample.PermGroFac[0]\n\n# np.dot multiplies vectors; probability times value for each outcome is expectation\nEpermShkInv = np.dot(permShkPrbs, permShkVals**(-1)) # $ \\mathbb{E}[\\psi^{-1}] $\nInvEpermShkInv= (EpermShkInv) ** (-1) # $ (\\mathbb{E}[\\psi^{-1}])^{-1}$\nPermGroFac = EPermGroFac * InvEpermShkInv # Uncertainty-adjusted permanent growth factor\nERNrmFac = Rfree / PermGroFac # Interest factor normalized by uncertainty-adjusted growth\nErNrmRte = ERNrmFac - 1 # Interest rate is interest factor - 1\n# \"sustainable\" C = P + (discounted) interest income\n# \"sustainable\" c = 1 + (discounted, normalized) interest income\nEmDelEq0 = lambda m : 1 + (m-1)*(ErNrmRte/ERNrmFac) # \"sustainable\" c where E[Δ m] = 0\n\n# %%\n# Plot GICFailsExample consumption function against the sustainable level of consumption\n\nGICFailsExample.solve() # Above, we set up the problem but did not solve it\nGICFailsExample.unpack('cFunc') # Make the consumption function easily accessible for plotting\nm = np.linspace(mPlotMin,5,mPts)\nc_Limt = GICFailsExample.cFunc[0](m)\nc_Sstn = EmDelEq0(m) # \"sustainable\" consumption\nplt.figure(figsize = (12,8))\nplt.plot(m,c_Limt,color=\"black\")\nplt.plot(m,c_Sstn,color=\"black\")\nplt.xlim(0,5.5)\nplt.ylim(0,1.6)\nplt.text(0,1.63,\"$c$\",fontsize = 26)\nplt.text(5.55,0,\"$m$\",fontsize = 26)\nplt.tick_params(labelbottom=False, labelleft=False,left='off',right='off',bottom='off',top='off')\nplt.text(1,0.6,\"$c(m_{t})$\",fontsize = 18)\nif latexExists:\n plt.text(1.5,1.2,\"$\\mathbb{E}_{t}[\\Delta m_{t+1}] = 0$\",fontsize = 22)\nelse:\n plt.text(1.5,1.2,\"$\\mathsf{E}_{t}[\\Delta m_{t+1}] = 0$\",fontsize = 22)\n\nplt.arrow(0.98,0.62,-0.2,0,head_width= 0.02,width=0.001,facecolor='black',length_includes_head='True')\nplt.arrow(2.2,1.2,0.3,-0.05,head_width= 0.02,width=0.001,facecolor='black',length_includes_head='True')\n\nmake('FVACnotGIC') # Save figures (if appropriate/possible)\n\n# %% [markdown]\n# In the [interactive dashboard](#interactive-dashboard), see what happens as changes in the time preference rate (or changes in risk $\\sigma_\\Psi$) change the consumer from _growth-patient_ $(\\Phi > \\tilde{\\Gamma})$ to _growth-impatient_ ($\\Phi < \\tilde{\\Gamma}$)\n\n# %%\n# Conditions can also be checked without solving the model\n# verbose=0: Print nothing\n# verbose=3: Print all available results\nGICFailsExample.checkConditions(verbose=0) \n\n\n# %% [markdown]\n# As a foundation for the remaining figures, we define another instance of the class $\\texttt{IndShockConsumerType}$, which has the same parameter values as the instance $\\texttt{baseAgent}$ defined previously but is solved to convergence (our definition of an infinite horizon agent type) instead of only 100 periods\n#\n\n# %%\n# cycles=0 tells the solver to find the infinite horizon solution\nbaseAgent_Inf = IndShockConsumerType(cycles=0,verbose=0, **base_params)\n\nbaseAgent_Inf.solve()\nbaseAgent_Inf.unpack('cFunc')\n\n# %% [markdown]\n# ### [Target $m$, Expected Consumption Growth, and Permanent Income Growth](https://econ.jhu.edu/people/ccarroll/papers/BufferStockTheory/#AnalysisoftheConvergedConsumptionFunction)\n#\n# The next figure, [Analysis of the Converged Consumption Function](https://econ.jhu.edu/people/ccarroll/papers/BufferStockTheory/#cGroTargetFig), shows the expected consumption growth factor $\\mathrm{\\mathbb{E}}_{t}[\\mathbf{c}_{t+1}/\\mathbf{c}_{t}]$ for a consumer behaving according to the converged consumption rule.\n#\n# Conveniently, this can be computed without knowing the _level_ of the consumer's income:\n#\n# \\begin{eqnarray}\n# \\mathbb{E}_{t}[\\mathbf{c}_{t+1}/\\mathbf{c}_{t}] & = & \\mathbb{E}_{t}\\left[\\frac{\\mathbf{p}_{t+1}c_{t+1}(m_{t+1})}{\\mathbf{p}_{t}c_{t}(m_{t})}\\right] \\\\ \n# & = & \\mathbb{E}_{t}\\left[\\frac{\\Gamma \\psi_{t+1} \\mathbf{p}_{t}}{\\mathbf{p}_{t}}\\frac{c_{t+1}(m_{t+1})}{c_{t}(m_{t})}\\right] \\\\\n# & = & \\mathbb{E}_{t}\\left[\\frac{\\Gamma \\psi_{t+1} c_{t+1}(m_{t+1})}{c_{t}(m_{t})}\\right] \n# \\end{eqnarray}\n#\n\n# %%\n# Def a function to calc ratio of cLev_{t+1} to p_{t}\ndef EcLev_tp1_Over_p_t(a):\n '''\n Taking end-of-period assets a as input, return ratio of expectation \n of next period's consumption to this period's permanent income \n\n Inputs:\n a: end-of-period assets\n Returns:\n EcLev_tp1_Over_p_{t}: next period's expected c level / current p\n '''\n # Extract parameter values to make code more readable\n permShkVals=baseAgent_Inf.PermShkDstn[0].X\n tranShkVals=baseAgent_Inf.TranShkDstn[0].X\n permShkPrbs=baseAgent_Inf.PermShkDstn[0].pmf\n tranShkPrbs=baseAgent_Inf.TranShkDstn[0].pmf\n Rfree =baseAgent_Inf.Rfree\n EPermGroFac=baseAgent_Inf.PermGroFac[0]\n \n PermGrowFac_tp1 = EPermGroFac*permShkVals # Nonstochastic growth times idiosyncratic permShk\n RNrmFac_tp1 = Rfree / PermGrowFac_tp1 # Growth-normalized interest factor \n # 'bank balances' b = end-of-last-period assets times normalized return factor\n b_tp1 = RNrmFac_tp1*a\n # expand dims of b_tp1 and use broadcasted sum of a column and a row vector\n # to obtain a matrix of possible market resources next period\n # because matrix mult is much much faster than looping to calc E\n m_tp1_GivenTranAndPermShks = np.expand_dims(b_tp1, axis=1) + tranShkVals\n # List of possible values of $\\mathbf{c}_{t+1}$ (Transposed by .T)\n cRat_tp1_GivenTranAndPermShks = baseAgent_Inf.cFunc[0](m_tp1_GivenTranAndPermShks).T\n cLev_tp1_GivenTranAndPermShks = cRat_tp1_GivenTranAndPermShks*PermGrowFac_tp1\n # compute expectation over perm shocks by right multiplying with probs\n EOverPShks_cLev_tp1_GivenTranShkShks = np.dot(cLev_tp1_GivenTranAndPermShks, permShkPrbs)\n # finish expectation over trans shocks by right multiplying with probs\n EcLev_tp1_Over_p_t = np.dot(EOverPShks_cLev_tp1_GivenTranShkShks, tranShkPrbs)\n # return expected consumption\n return EcLev_tp1_Over_p_t\n\n# %%\n# Calculate the expected consumption growth factor\n# mBelwTrg defines the plot range on the left of target m value (e.g. m <= target m)\nmNrmTrg=baseAgent_Inf.solution[0].mNrmSS\nmBelwTrg = np.linspace(1,mNrmTrg,50) \nc_For_mBelwTrg = baseAgent_Inf.cFunc[0](mBelwTrg)\na_For_mBelwTrg = mBelwTrg-c_For_mBelwTrg\nEcLev_tp1_Over_p_t_For_mBelwTrg = [EcLev_tp1_Over_p_t(i) for i in a_For_mBelwTrg]\n\n# mAbveTrg defines the plot range on the right of target m value (e.g. m >= target m)\nmAbveTrg = np.linspace(mNrmTrg,1.9,50)\n\n# EcGro_For_mAbveTrg: E [consumption growth factor] when m_{t} is below target m\nEcGro_For_mBelwTrg = np.array(EcLev_tp1_Over_p_t_For_mBelwTrg)/c_For_mBelwTrg\n\nc_For_mAbveTrg = baseAgent_Inf.cFunc[0](mAbveTrg)\na_For_mAbveTrg = mAbveTrg-c_For_mAbveTrg\nEcLev_tp1_Over_p_t_For_mAbveTrg = [EcLev_tp1_Over_p_t(i) for i in a_For_mAbveTrg]\n\n# EcGro_For_mAbveTrg: E [consumption growth factor] when m_{t} is bigger than target m_{t}\nEcGro_For_mAbveTrg = np.array(EcLev_tp1_Over_p_t_For_mAbveTrg)/c_For_mAbveTrg \n\n# %%\n# Define a function to construct the arrows on the consumption growth rate function\ndef arrowplot(axes, x, y, narrs=15, dspace=0.5, direc='neg',\n hl=0.01, hw=3, c='black'):\n '''\n The function is used to plot arrows given the data x and y.\n\n Input:\n narrs : Number of arrows that will be drawn along the curve\n\n dspace : Shift the position of the arrows along the curve.\n Should be between 0. and 1.\n\n direc : can be 'pos' or 'neg' to select direction of the arrows\n\n hl : length of the arrow head\n\n hw : width of the arrow head\n\n c : color of the edge and face of the arrow head\n '''\n\n # r is the distance spanned between pairs of points\n r = np.sqrt(np.diff(x)**2+np.diff(y)**2)\n r = np.insert(r, 0, 0.0)\n\n # rtot is a cumulative sum of r, it's used to save time\n rtot = np.cumsum(r)\n\n # based on narrs set the arrow spacing\n aspace = r.sum() / narrs\n\n if direc is 'neg':\n dspace = -1.*abs(dspace)\n else:\n dspace = abs(dspace)\n\n arrowData = list() # will hold tuples of x,y,theta for each arrow\n arrowPos = aspace*(dspace) # current point on walk along data\n # could set arrowPos to 0 if you want\n # an arrow at the beginning of the curve\n\n ndrawn = 0\n rcount = 1\n while arrowPos < r.sum() and ndrawn < narrs:\n x1,x2 = x[rcount-1],x[rcount]\n y1,y2 = y[rcount-1],y[rcount]\n da = arrowPos-rtot[rcount]\n theta = np.arctan2((x2-x1),(y2-y1))\n ax = np.sin(theta)*da+x1\n ay = np.cos(theta)*da+y1\n arrowData.append((ax,ay,theta))\n ndrawn += 1\n arrowPos+=aspace\n while arrowPos > rtot[rcount+1]:\n rcount+=1\n if arrowPos > rtot[-1]:\n break\n\n for ax,ay,theta in arrowData:\n # use aspace as a guide for size and length of things\n # scaling factors were chosen by experimenting a bit\n\n dx0 = np.sin(theta)*hl/2.0 + ax\n dy0 = np.cos(theta)*hl/2.0 + ay\n dx1 = -1.*np.sin(theta)*hl/2.0 + ax\n dy1 = -1.*np.cos(theta)*hl/2.0 + ay\n\n if direc is 'neg' :\n ax0 = dx0\n ay0 = dy0\n ax1 = dx1\n ay1 = dy1\n else:\n ax0 = dx1\n ay0 = dy1\n ax1 = dx0\n ay1 = dy0\n\n axes.annotate('', xy=(ax0, ay0), xycoords='data',\n xytext=(ax1, ay1), textcoords='data',\n arrowprops=dict( headwidth=hw, frac=1., ec=c, fc=c))\n\n\n# %%\n# Plot consumption growth as a function of market resources\n\n# Retrieve parameters (makes code readable)\nRfree = baseAgent_Inf.Rfree\nDiscFac = baseAgent_Inf.DiscFac\nCRRA = baseAgent_Inf.CRRA\nEPermGroFac= baseAgent_Inf.PermGroFac[0]\nmNrmTrg = baseAgent_Inf.solution[0].mNrmSS\n\n# Calculate Absolute Patience Factor Phi = lower bound of consumption growth factor\nAPF = (Rfree*DiscFac)**(1.0/CRRA)\n\nfig = plt.figure(figsize = (12,8))\nax = fig.add_subplot(111)\n# Plot the Absolute Patience Factor line\nax.plot([0,1.9],[APF,APF],color=\"black\")\n\n# Plot the Permanent Income Growth Factor line\nax.plot([0,1.9],[EPermGroFac,EPermGroFac],color=\"black\")\n\n# Plot the expected consumption growth factor on the left side of target m\nax.plot(mBelwTrg,EcGro_For_mBelwTrg,color=\"black\")\n\n# Plot the expected consumption growth factor on the right side of target m\nax.plot(mAbveTrg,EcGro_For_mAbveTrg,color=\"black\")\n\n# Plot the arrows\narrowplot(ax, mBelwTrg,EcGro_For_mBelwTrg)\narrowplot(ax, mAbveTrg,EcGro_For_mAbveTrg, direc='pos')\nfsbig=26\nfsmid=22\n\n# Plot the target m\nax.plot([mNrmTrg,mNrmTrg],[0,1.4],color=\"black\",linestyle=\"--\")\nax.set_xlim(1,2.10)\nax.set_ylim(0.98,1.08)\nax.text(1,1.082,r'$\\text{Growth Rate}$',fontsize = fsbig,fontweight='bold')\nax.text(2.105,0.975,\"$m_{t}$\",fontsize = fsbig,fontweight='bold')\nif latexExists:\n ax.text(1.91,1.01,\"$\\mathbb{E}_{t}[\\mathbf{c}_{t+1}/\\mathbf{c}_{t}]$\",fontsize = fsmid,fontweight='bold')\nelse:\n ax.text(1.91,1.01,\"$\\mathsf{E}_{t}[\\mathbf{c}_{t+1}/\\mathbf{c}_{t}]$\",fontsize = fsmid,fontweight='bold')\nax.text(mNrmTrg-0.02,0.974, r'$\\check{m}$', fontsize = fsbig,fontweight='bold')\nax.tick_params(labelbottom=False, labelleft=False,left='off',right='off',bottom='off',top='off')\nif latexExists:\n ax.text(1.91,0.998,r'$\\pmb{\\text{\\TH}} = (\\mathsf{R}\\beta)^{1/\\rho}$',fontsize = fsmid,fontweight='bold')\nelse:\n ax.text(1.91,0.998,r'$\\Phi = (\\mathsf{\\mathsf{R}}\\beta)^{1/\\rho}$',fontsize = fsmid,fontweight='bold')\n\nax.text(1.91,1.03, r'$\\Gamma$',fontsize = fsmid,fontweight='bold')\nmake('cGroTargetFig')\n\n\n# %% [markdown]\n# In the [interactive dashboard](#interactive-dashboard) see how target wealth changes when the consumer's time preference factor β or the growth factor Γ change.\n\n# %% [markdown]\n# ### [Consumption Function Bounds](https://econ.jhu.edu/people/ccarroll/papers/BufferStockTheory/#AnalysisOfTheConvergedConsumptionFunction)\n# [The next figure](https://econ.jhu.edu/people/ccarroll/papers/BufferStockTheory/#cFuncBounds)\n# illustrates theoretical bounds for the consumption function.\n#\n# We define two useful variables: lower bound of $\\kappa$ (marginal propensity to consume) and limit of $h$ (Human wealth), along with some functions such as the limiting perfect foresight consumption function $\\bar{c}(m)$, the upper bound function $\\bar{\\bar c}(m)$, and the lower bound function \\underline{_c_}$(m)$.\n\n# %%\n# Define κ_Min, h_inf and perfect foresight consumption function, upper bound of consumption function and lower\n# bound of consumption function.\n\n# Retrieve parameters (makes code readable)\nRfree = baseAgent_Inf.Rfree\nDiscFac = baseAgent_Inf.DiscFac\nCRRA = baseAgent_Inf.CRRA\nEPermGroFac= EPermGroFac\nmNrmTrg = baseAgent_Inf.solution[0].mNrmSS\nUnempPrb = baseAgent_Inf.UnempPrb\n\nκ_Min = 1.0-(Rfree**(-1.0))*(Rfree*DiscFac)**(1.0/CRRA)\nh_inf = (1.0/(1.0-EPermGroFac/Rfree))\ncFunc_Uncnst = lambda m: (h_inf -1)* κ_Min + κ_Min*m\ncFunc_TopBnd = lambda m: (1 - UnempPrb ** (1.0/CRRA)*(Rfree*DiscFac)**(1.0/CRRA)/Rfree)*m\ncFunc_BotBnd = lambda m: (1 -(Rfree*DiscFac)**(1.0/CRRA)/Rfree) * m\n\n\n# %%\n# Plot the consumption function and its bounds\n\ncMaxLabel=r'$\\overline{c}(m) = (m-1+h)\\underline{\\kappa}$'\ncMinLabel=r'Lower Bound: $\\underline{c}(m)= (1-\\pmb{\\text{\\TH}}_{R})\\underline{\\kappa}m$'\nif not latexExists:\n cMaxLabel=r'$\\overline{c}(m) = (m-1+h)κ̲' # Use unicode kludge\n cMinLabel=r'Lower Bound: c̲$(m)= (1-\\Phi_{R})m = κ̲ m$'\n\nmPlotMax = 25\nmPlotMin = 0\n# mKnk is point where the two upper bounds meet\nmKnk = ((h_inf-1)* κ_Min)/((1 - UnempPrb**(1.0/CRRA)*(Rfree*DiscFac)**(1.0/CRRA)/Rfree)-κ_Min)\nmBelwKnkPts = 300\nmAbveKnkPts = 700\nmBelwKnk = np.linspace(mPlotMin,mKnk,mBelwKnkPts)\nmAbveKnk = np.linspace(mKnk,mPlotMax,mAbveKnkPts)\nmFullPts = np.linspace(mPlotMin,mPlotMax,mBelwKnkPts+mAbveKnkPts)\n\nplt.figure(figsize = (12,8))\nplt.plot(mFullPts,baseAgent_Inf.cFunc[0](mFullPts), color=\"black\")\nplt.plot(mBelwKnk,cFunc_Uncnst(mBelwKnk) , color=\"black\",linestyle=\"--\")\nplt.plot(mAbveKnk,cFunc_Uncnst(mAbveKnk) , color=\"black\",linewidth=2.5)\nplt.plot(mBelwKnk,cFunc_TopBnd(mBelwKnk) , color=\"black\",linewidth=2.5)\nplt.plot(mAbveKnk,cFunc_TopBnd(mAbveKnk) , color=\"black\",linestyle=\"--\")\nplt.plot(mBelwKnk,cFunc_BotBnd(mBelwKnk) , color=\"black\",linewidth=2.5)\nplt.plot(mAbveKnk,cFunc_BotBnd(mAbveKnk) , color=\"black\",linewidth=2.5)\nplt.tick_params(labelbottom=False, labelleft=False,left='off',right='off',bottom='off',top='off')\nplt.xlim(mPlotMin,mPlotMax)\nplt.ylim(mPlotMin,1.12*cFunc_Uncnst(mPlotMax))\nplt.text(mPlotMin,1.12*cFunc_Uncnst(mPlotMax)+0.05,\"$c$\",fontsize = 22)\nplt.text(mPlotMax+0.1,mPlotMin,\"$m$\",fontsize = 22)\nplt.text(2.5,1,r'$c(m)$',fontsize = 22,fontweight='bold')\nif latexExists:\n plt.text(6,5,r'$\\overline{\\overline{c}}(m)= \\overline{\\kappa}m = (1-\\wp^{1/\\rho}\\pmb{\\text{\\TH}}_{R})m$',fontsize = 22,fontweight='bold')\nelse:\n plt.text(6,5,r'$\\overline{\\overline{c}}(m)= \\overline{\\kappa}m = (1-\\wp^{1/\\rho}\\Phi_{R})m$',fontsize = 22,fontweight='bold')\nplt.text(2.2,3.8, cMaxLabel,fontsize = 22,fontweight='bold')\nplt.text(9,4.1,r'Upper Bound $ = $ Min $[\\overline{\\overline{c}}(m),\\overline{c}(m)]$',fontsize = 22,fontweight='bold')\nplt.text(8,0.8,cMinLabel,fontsize = 22,fontweight='bold')\nplt.arrow(2.45,1.05,-0.5,0.02,head_width= 0.05,width=0.001,facecolor='black',length_includes_head='True')\nplt.arrow(2.15,3.88,-0.5,0.1,head_width= 0.05,width=0.001,facecolor='black',length_includes_head='True')\nplt.arrow(8.95,4.15,-0.8,0.05,head_width= 0.05,width=0.001,facecolor='black',length_includes_head='True')\nplt.arrow(5.95,5.05,-0.4,mPlotMin,head_width= 0.05,width=0.001,facecolor='black',length_includes_head='True')\nplt.arrow(14,0.70,0.5,-0.1,head_width= 0.05,width=0.001,facecolor='black',length_includes_head='True')\n\nmake('cFuncBounds')\n\n\n# %% [markdown]\n# ### [The Consumption Function and Target $m$](https://econ.jhu.edu/people/ccarroll/papers/BufferStockTheory/#cFuncBounds)\n#\n# This figure shows the $\\mathrm{\\mathbb{E}}_{t}[\\Delta m_{t+1}]=0$ locus and consumption function $c(m_{t})$, along with the intersection of these two functions, which defines the target value of $m$\n\n# %%\n# This just plots objects that have already been constructed\n\nmBelwTrg = np.linspace(mPlotMin,4,mPts)\nEmDelEq0 = lambda m:(EPermGroFac/Rfree)+(1.0-EPermGroFac/Rfree)*m\ncBelwTrg_Best = baseAgent_Inf.cFunc[0](mBelwTrg) # \"best\" = optimal c\ncBelwTrg_Sstn = EmDelEq0(mBelwTrg) # \"sustainable\" c\nplt.figure(figsize = (12,8))\nplt.plot(mBelwTrg,cBelwTrg_Best, color=\"black\")\nplt.plot(mBelwTrg,cBelwTrg_Sstn, color=\"black\")\nplt.xlim(mPlotMin,3)\nplt.ylim(mPlotMin,1.45)\nplt.plot([mNrmTrg, mNrmTrg],[mPlotMin,2.5],color=\"black\",linestyle=\"--\")\nplt.tick_params(labelbottom=False, labelleft=False,left='off',right='off',bottom='off',top='off')\nplt.text(mPlotMin,1.47,r\"$c$\",fontsize = 26)\nplt.text(3.02,mPlotMin,r\"$m$\",fontsize = 26)\nif latexExists:\n plt.text(2.3,0.94,r'$\\mathbb{E}_{t}[\\Delta m_{t+1}] = 0$',fontsize = 22,fontweight='bold')\nelse:\n plt.text(2.3,0.94,r'$\\mathsf{E}_{t}[\\Delta m_{t+1}] = 0$',fontsize = 22,fontweight='bold')\nplt.text(2.3,1.1,r\"$c(m_{t})$\",fontsize = 22,fontweight='bold')\nplt.text(mNrmTrg-0.05,-0.1, r\"$\\check{m}$\",fontsize = 26)\nplt.arrow(2.28,1.12,-0.1,0.03,head_width= 0.02,width=0.001,facecolor='black',length_includes_head='True')\nplt.arrow(2.28,0.97,-0.1,0.02,head_width= 0.02,width=0.001,facecolor='black',length_includes_head='True')\n\nmake('cRatTargetFig')\n\n\n# %% [markdown]\n# ### [Upper and Lower Limits of the Marginal Propensity to Consume](https://econ.jhu.edu/people/ccarroll/papers/BufferStockTheory/#MPCLimits)\n#\n# The paper shows that as $m_{t}~\\uparrow~\\infty$ the consumption function in the presence of risk gets arbitrarily close to the perfect foresight consumption function. Defining \\underline{κ}\n# as the perfect foresight model's MPC, this implies that $\\lim_{m_{t}~\\uparrow~\\infty} c^{\\prime}(m) = $ \\underline{κ}\n# .\n#\n# The paper also derives an analytical limit $\\bar{\\kappa}$ for the MPC as $m$ approaches 0., its bounding value. Strict concavity of the consumption function implies that the consumption function will be everywhere below a function $\\bar{\\kappa}m$, and strictly declining everywhere. The last figure plots the MPC between these two limits.\n\n# %%\n# The last figure shows the upper and lower limits of the MPC\n\n# Retrieve parameters (makes code readable)\nRfree = baseAgent_Inf.Rfree\nDiscFac = baseAgent_Inf.DiscFac\nCRRA = baseAgent_Inf.CRRA\nEPermGroFac= EPermGroFac\nmNrmTrg = baseAgent_Inf.solution[0].mNrmSS\nUnempPrb = baseAgent_Inf.UnempPrb\n\nmPlotMax=8 \n\nplt.figure(figsize = (12,8))\n# Set the plot range of m\nm = np.linspace(0.001,mPlotMax,mPts)\n\n# Use the HARK method derivative to get the derivative of cFunc, and which constitutes the MPC\nMPC = baseAgent_Inf.cFunc[0].derivative(m)\n\n# Define the upper bound of MPC\nκ_Max = (1 - UnempPrb ** (1.0/CRRA)*(Rfree*DiscFac)**(1.0/CRRA)/Rfree)\n\n# Define the lower bound of MPC\nMPCLower = κ_Min\n\nkappaDef=r'$\\underline{\\kappa}\\equiv(1-\\pmb{\\text{\\TH}}_{R})$'\nif not latexExists:\n kappaDef=r'κ̲$\\equiv(1-\\Phi_{R})$'\n\nplt.plot(m,MPC,color = 'black')\nplt.plot([mPlotMin,mPlotMax],[κ_Max,κ_Max],color = 'black')\nplt.plot([mPlotMin,mPlotMax],[κ_Min,κ_Min],color = 'black')\nplt.xlim(mPlotMin,mPlotMax)\nplt.ylim(0,1) # MPC bounds are between 0 and 1 \nplt.text(1.5,0.6,r'$\\kappa(m) \\equiv c^{\\prime}(m)$',fontsize = 26,fontweight='bold')\nif latexExists:\n plt.text(5,0.87,r'$(1-\\wp^{1/\\rho}\\pmb{\\text{\\TH}})\\equiv \\overline{\\kappa}$',fontsize = 26,fontweight='bold') # Use Thorn character\nelse:\n plt.text(5,0.87,r'$(1-\\wp^{1/\\rho}\\Phi_{R})\\equiv \\overline{\\kappa}$',fontsize = 26,fontweight='bold') # Use Phi instead of Thorn (alas)\n\nplt.text(0.5,0.07,kappaDef,fontsize = 26,fontweight='bold')\nplt.text(mPlotMax+0.05,mPlotMin,\"$m$\",fontsize = 26)\nplt.arrow(1.45,0.61,-0.4,mPlotMin,head_width= 0.02,width=0.001,facecolor='black',length_includes_head='True')\nplt.arrow(2.2,0.07,0.2,-0.01,head_width= 0.02,width=0.001,facecolor='black',length_includes_head='True')\nplt.arrow(4.95,0.895,-0.2,0.03,head_width= 0.02,width=0.001,facecolor='black',length_includes_head='True')\n\nmake('MPCLimits')\n\n\n# %% [markdown]\n# # Summary\n#\n# [Two tables in the paper](https://econ-ark.github.io/BufferStockTheory/#Factors-Defined-And-Compared) summarize the various definitions, and then articulate conditions required for the problem to have a nondegenerate solution. Among the nondegenerate cases, the most interesting result is that if the Growth Impatience Condition holds there will be a target level of wealth.\n\n# %% [markdown]\n# ### Appendix: Options for Interacting With This Notebook \n#\n# 1. [View (static version)](https://github.com/llorracc/BufferStockTheory/blob/master/Code/Python/BufferStockTheory.ipynb) on GitHub (warning: GitHub does not render Jupyter notebooks reliably)\n# 1. [Launch Online Interactive Version](https://econ-ark.org/materials/BufferStockTheory/#launch)\n# 1. For fast (local) execution, install [econ-ark](http://github.com/econ-ark) on your computer ([QUICK START GUIDE](https://github.com/econ-ark/HARK/blob/master/README.md)) then follow these instructions to retrieve the full contents of the `BufferStockTheory` [REMARK](https://github.com/econ-ark/REMARK):\n# 1. At a command line, change the working directory to the one where you want to install\n# * On unix, if you install in the `/tmp` directory, the installation will disappear after a reboot:\n# * `cd /tmp`\n# 1. `git clone https://github.com/econ-ark/REMARK --recursive`\n# 1. `cd REMARK/REMARKs/BufferStockTheory`\n# 1. `jupyter notebook BufferStockTheory.ipynb`\n\n# %% [markdown]\n# ### Appendix: Perfect foresight agent failing both the FHWC and RIC\n\n# %%\nfrom copy import copy\nfrom HARK.ConsumptionSaving.ConsIndShockModel import PerfForesightConsumerType\nfig6_par = copy(base_params)\n\n# Replace parameters.\nfig6_par['Rfree'] = 0.98\nfig6_par['DiscFac'] = 1\nfig6_par['PermGroFac'] = [0.99]\nfig6_par['CRRA'] = 2\nfig6_par['BoroCnstArt'] = 0\nfig6_par['T_cycle'] = 0\nfig6_par['cycles'] = 0\nfig6_par['quiet'] = False\n\n# Create the agent\nRichButPatientAgent = PerfForesightConsumerType(**fig6_par)\n# Check conditions\nRichButPatientAgent.checkConditions(verbose = 3)\n# Solve\nRichButPatientAgent.solve()\n\n# Plot\nmPlotMin, mPlotMax = 1, 9.5\nplt.figure(figsize = (8,4))\nm_grid = np.linspace(mPlotMin,mPlotMax,500)\nplt.plot(m_grid-1, RichButPatientAgent.solution[0].cFunc(m_grid), color=\"black\")\nplt.text(mPlotMax-1+0.05,1,r\"$b$\",fontsize = 26)\nplt.text(mPlotMin-1,1.017,r\"$c$\",fontsize = 26)\nplt.xlim(mPlotMin-1,mPlotMax-1)\nplt.ylim(mPlotMin,1.016)\n\nmake('PFGICHoldsFHWCFailsRICFails')\n\n# %%\n","sub_path":"Code/Python/BufferStockTheory-Problems.py","file_name":"BufferStockTheory-Problems.py","file_ext":"py","file_size_in_byte":46316,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"383599778","text":"\"\"\"\nThis module contains functions to perform non-nested and nested cv.\n\nThis module contains a function to select a strategy for models evaluation.\n\n\"\"\"\n\nfrom sklearn.ensemble import BaggingClassifier\nfrom . import param_grids_distros as pgd\nfrom . import neuralnets as nn\nfrom operator import itemgetter\n\nimport numpy as np\nfrom scipy.stats import randint as sp_randint\nfrom .. import auto_utils as au\nfrom . import train_calibrate as tc\n\nfrom . import eval_utils as eu\nimport autoclf.getargs as ga\nfrom sklearn.model_selection import StratifiedKFold\nfrom sklearn.pipeline import Pipeline\nfrom sklearn.dummy import DummyClassifier\n\nimport os\nimport sys\nstderr = sys.stderr\nsys.stderr = open(os.devnull, 'w')\nfrom keras.wrappers.scikit_learn import KerasClassifier\nsys.stderr = stderr\n# os.environ['TF_CPP_MIN_LOG_LEVEL'] = '3'\n\nimport warnings\nwarnings.filterwarnings(\"ignore\")\n\n\ndef create_keras_classifiers(\n y_type, input_dim, labels, nb_epoch, batch_size):\n\n keras_clf_fcts = dict(\n baseline_nn_default_Clf_2nd=(\n nn.baseline_nn_model_multiclass, nn.baseline_nn_model),\n baseline_nn_smaller_Clf_2nd=(\n nn.baseline_nn_smaller_model_multiclass, nn.baseline_nn_smaller_model),\n larger_nn_Clf_2nd=(nn.larger_nn_model_multiclass, nn.larger_nn_model),\n deep_nn_Clf_2nd=(nn.deep_nn_model_multiclass, nn.deep_nn_model),\n deeper_nn_Clf_2nd=(nn.deeper_nn_model_multiclass, nn.deeper_nn_model)\n )\n\n if input_dim < 15:\n keras_clf_fcts['larger_deep_nn_Clf_2nd'] = (\n nn.larger_deep_nn_model_multiclass, nn.larger_deep_nn_model)\n\n names_and_models=dict()\n\n output_dim = 1\n\n if y_type == 'multiclass':\n\n if labels is None:\n raise ValueError(\"%r is not a valid type for var 'labels'\" % labels)\n elif not isinstance(labels, list):\n raise TypeError(\"Multiclass keras models need a list of string labels.\")\n else:\n output_dim = len(labels)\n\n for k, v in keras_clf_fcts.items():\n names_and_models[k] = KerasClassifier(\n build_fn=v[0], nb_epoch=nb_epoch,\n input_dim=input_dim, output_dim=output_dim, batch_size=batch_size,\n verbose=0)\n\n else:\n\n for k, v in keras_clf_fcts.items():\n names_and_models[k] = KerasClassifier(\n build_fn=v[1], nb_epoch=nb_epoch,\n input_dim=input_dim, batch_size=batch_size,\n verbose=0)\n\n return names_and_models\n\n\ndef create_best_keras_clf_architecture(\n keras_clf_name, y_type, labels, input_dim, nb_epoch, keras_param_grid):\n \"\"\"\n Find KerasClf best architecture using\n\n ------\n \"\"\"\n\n output_dim = 1\n\n for n in np.arange(0, 3):\n keras_param_grid[keras_clf_name + '__units_' + str(n)] = sp_randint(\n input_dim, 5*input_dim)\n\n if y_type == 'multiclass':\n\n if labels is None:\n raise ValueError(\"%r is not a valid type for var 'labels'\" % labels)\n elif not isinstance(labels, list):\n raise TypeError(\"Multiclass keras models need a list of string labels.\")\n else:\n output_dim = len(labels)\n\n keras_nn_model = KerasClassifier(\n build_fn=nn.tunable_deep_nn_multiclass, nb_epoch=nb_epoch,\n input_dim=input_dim, output_dim=output_dim,\n verbose=0)\n\n else:\n\n # you need KerasClassifier wrapper to use Keras models in sklearn\n\n keras_nn_model = KerasClassifier(\n build_fn=nn.tunable_deep_nn, nb_epoch=nb_epoch,\n input_dim=input_dim, verbose=0)\n\n return (keras_nn_model, keras_param_grid)\n\n\ndef models_and_params_for_classic_cv_tasks(y_type, scoring, labels):\n models_and_params = dict()\n\n if y_type == 'binary':\n models_and_params = pgd.starting_point_models_and_params\n # models_and_params['LogRClf_2nd'].set_params(\n # solver='liblinear')\n\n else:\n # if y_type == 'multiclass':\n\n if scoring == 'neg_log_loss': \n for k, v in pgd.starting_point_models_and_params.items():\n if hasattr(v, 'predict_proba'):\n print(k)\n models_and_params[k] = v\n\n # solver='saga', penalty='l1'\n # models_and_params['LogRClf_2nd'].set_params(\n # solver='lbfgs', penalty='l2', multi_class='multinomial')\n\n # RandomForestClf better suited to handle lot of categories\n if labels is not None and len(labels) > 10:\n del models_and_params['GBoostingClf_2nd']\n\n return models_and_params\n\n\ndef models_and_params_for_nested_cv_tasks(y_type, scoring, labels):\n models_and_params = dict()\n\n if y_type == 'binary':\n models_and_params = pgd.full_search_models_and_parameters\n models_and_params['LogRClf_2nd'][0].set_params(solver='liblinear')\n\n else:\n # y_type == 'multiclass'\n\n if scoring == 'neg_log_loss':\n models_and_params = dict()\n\n for k, v in pgd.full_search_models_and_parameters.items():\n if hasattr(v[0], 'predict_proba'):\n models_and_params[k] = v\n\n # solver='saga', penalty='l1'\n models_and_params['LogRClf_2nd'][0].set_params(\n solver='lbfgs', penalty='l2', multi_class='multinomial')\n\n # RandomForestClf better suited to handle lot of categories\n if labels is not None and len(labels) > 10:\n del models_and_params['GBoostingClf_2nd']\n\n return models_and_params\n\n\ndef create_ensemble_of_best_models(best_model_name, best_model_estim, seed=0):\n base_estimator_name = best_model_name.strip('_2nd')\n bagging_estimator_name = 'BaggingClf_2nd_' + base_estimator_name\n\n bagging_param_grid = {\n bagging_estimator_name + '__'\n + k: v for k, v in pgd.Bagging_param_grid.items()\n }\n bagging = BaggingClassifier(best_model_estim, random_state=seed)\n\n return bagging, bagging_estimator_name, bagging_param_grid\n\n\ndef perform_classic_cv_evaluation_and_calibration(\n auto_feat_eng_data, scoring, Y_type, labels=None, \n d_name=None, random_state=0):\n \"\"\"\n # perform non-nested cross validation and calibration of best estimator.\n\n ---------------------------------------------------------\n auto_feat_eng_data:\n dictionary with encoder, scaler, feature selector,\n Pipeline.steps and train-test data\n scoring: scoring for model evaluation\n -- string ('roc_auc_score') or list of strings\n random_state: seed\n \"\"\"\n if isinstance(scoring, list) or isinstance(scoring, dict) or isinstance(\n scoring, tuple):\n raise TypeError(\"\"\"\n 'perform_classic_cv_evaluation_and_calibration' method allows only\n to perform single-metric evaluations.\n \"\"\")\n if scoring not in (None, 'accuracy', 'roc_auc', 'neg_log_loss'):\n raise ValueError(\"\"\"\n %s is not a valid scoring value\n for method 'perform_classic_cv_evaluation_and_calibration'.\n Valid options are ['accuracy', 'roc_auc', 'neg_log_loss']\n \"\"\" % scoring)\n\n print(\"### Probability Calibration of Best Estimator \"\n \"-- CalibratedClassifierCV with cv=cv (no prefit) ###\")\n print(\"### Non-nested Cross-Validation ###\")\n print(\"Models are trained and calibrated on the same data, train data,\\n\"\n \"calibration is evaluated on test data. No nested-cv is performed.\")\n print()\n print(\"pipe.fit(train_data)\")\n\n encoding = auto_feat_eng_data['encoding']\n scaler_tuple = auto_feat_eng_data['scaler']\n featselector = auto_feat_eng_data['feat_selector']\n steps = auto_feat_eng_data['steps']\n X_train_transformed, y_train, X_test_transformed, y_test = auto_feat_eng_data['data_arrays']\n X, y = auto_feat_eng_data['Xy']\n\n print()\n\n print()\n print(\"X_train shape: \", X_train_transformed.shape)\n # print(\"X_train -- first row:\", X_train.values[0])\n print(\"X_train -- first row:\", X_train_transformed[0])\n print(\"y_train shape: \", y_train.shape)\n print()\n\n print(\"X_test shape: \", X_test_transformed.shape)\n # print(\"X_test -- first row:\", X_test.values[0])\n print(\"X_test -- first row:\", X_test_transformed[0])\n print(\"y_test shape: \", y_test.shape)\n print()\n\n # input(\"Press any key to continue...\")\n\n # Start evaluation process\n\n # Evaluation of best model with non-nested CV -- outer: CV\n\n n_splits = 3 # au.select_nr_of_splits_for_kfold_cv()\n\n # dict of models and their associated parameters\n # if it comes out that the best model is LogReg, no comparison is needed\n\n best_atts = eu.best_model_initial_attributes(scoring, n_splits)\n\n best_score, best_score_dev, best_cv_results, best_model_name = best_atts\n\n best_exec_time = 31536000 # one year in seconds\n best_model = (best_model_name, None, None)\n\n Dummy_scores = []\n\n models_data = []\n names = []\n results = []\n\n scores_of_best_model = (\n best_score, best_score_dev, best_cv_results,\n best_exec_time, best_model)\n\n print()\n print(\"=== [task] Evaluation of DummyClassifier\")\n print()\n\n wtr = eu.calculate_sample_weight(y_train)\n\n print(\"=== 'sample_weight'\")\n print(wtr[:5])\n print(\"=== target train data sample\")\n print(y_train[:5])\n print()\n\n # This cross-validation object is\n # a variation of KFold that returns stratified folds.\n # The folds are made by preserving\n # the percentage of samples for each class.\n outer_cv = StratifiedKFold(\n n_splits=n_splits, shuffle=True, random_state=random_state)\n\n strategy = 'stratified' # 'most_frequent'\n\n average_scores_and_best_scores = eu.single_classic_cv_evaluation(\n X_train_transformed, y_train, 'DummyClf_2nd',\n DummyClassifier(strategy=strategy), wtr, scoring, outer_cv,\n dict(), scores_of_best_model, results, names, random_state)\n\n scores_of_best_model = average_scores_and_best_scores[1]\n\n Dummy_scores.append(scores_of_best_model[0]) # Dummy score -- ROC_AUC\n Dummy_scores.append(scores_of_best_model[1]) # Dummy score std\n Dummy_scores.append(scores_of_best_model[2]) # Dummy cv results for score\n # Dummy_scores.append(scores_of_best_model[2]) # Dummy Brier score loss\n Dummy_scores.append(scores_of_best_model[3]) # Dummy execution time\n # Dummy model's name and estimator\n Dummy_scores.append(scores_of_best_model[4])\n\n names = []\n results = []\n\n # Non-Nested CV: its purpose is to estimate how well\n # models would perform with their default parameters already tuned\n # or with default parameters\n\n print()\n print(\"=== Classic Cross-Validation\")\n print()\n\n print(\"Estimators before model evaluation:\", steps)\n print()\n\n # we will collect the average of the scores on the k outer folds in\n # this dictionary with keys given by the names of the models\n # in 'pgd.starting_point_models_and_params'\n average_scores_across_outer_folds_for_each_model = dict()\n\n # this holds for regression as well, not time-series\n\n models_and_parameters = models_and_params_for_classic_cv_tasks(\n Y_type, scoring, labels)\n\n average_scores_and_best_scores = eu.classic_cv_model_evaluation(\n X_train_transformed, y_train, models_and_parameters,\n # {},\n scoring, outer_cv, average_scores_across_outer_folds_for_each_model,\n scores_of_best_model, results, names, random_state)\n\n print()\n au.box_plots_of_models_performance(results, names)\n\n print()\n # input(\"Press any key to continue...\")\n\n results = []\n names = []\n\n print()\n print(\"=== After Classic CV evaluation...\")\n print()\n\n average_scores_across_outer_folds_for_each_model = average_scores_and_best_scores[0]\n scores_of_best_model = average_scores_and_best_scores[1]\n\n best_model_name = scores_of_best_model[4][0]\n best_model_estim = scores_of_best_model[4][1]\n\n best_score = scores_of_best_model[0]\n best_score_dev = scores_of_best_model[1]\n best_cv_results = scores_of_best_model[2]\n # best_brier_score = scores_of_best_model[2]\n best_exec_time = scores_of_best_model[3]\n\n Dummy_score = Dummy_scores[0]\n Dummy_score_dev = Dummy_scores[1]\n Dummy_cv_results = Dummy_scores[2]\n # Dummy_brier_score = Dummy_scores[3]\n Dummy_exec_time = Dummy_scores[3]\n\n print()\n print(\"Currently, best model is '%s' with score '%s': %1.3f (%1.3f)... :\" %\n (best_model_name, scoring.strip('neg_'), best_score, best_score_dev))\n print(\"... execution time: %.2fs\" % best_exec_time)\n # print(\"and prediction confidence: %1.3f\" % best_brier_score)\n print()\n\n print(\"=== Classic CV to evaluate more complex models\")\n print()\n\n complex_models_and_parameters = dict()\n average_scores_across_outer_folds_complex = dict()\n\n all_models_and_parameters = models_and_parameters\n\n # Let's add some simple neural network\n\n print(\"=== [task] Comparing best model to simple Neural Network \"\n \"(with single or two hidden layers).\")\n print()\n\n # This is an experiment to check \n # how different Keras architectures perform\n # to avoid hard-coding NNs, you should determine at least \n # nr of layers and nr of nodes by using Grid or Randomized Search CV\n\n input_dim = int(X_train_transformed.shape[1])\n\n nb_epoch = au.select_nr_of_iterations('nn')\n\n batch_size = 32\n\n kclf_names_and_models = create_keras_classifiers(\n Y_type, input_dim, labels, nb_epoch, batch_size)\n\n for k, v in kclf_names_and_models.items():\n complex_models_and_parameters[k] = v\n\n average_scores_and_best_scores = eu.classic_cv_model_evaluation(\n X_train_transformed, y_train, complex_models_and_parameters, scoring,\n outer_cv, average_scores_across_outer_folds_complex,\n scores_of_best_model, results, names, random_state)\n\n print()\n au.box_plots_of_models_performance(results, names)\n\n print()\n print(\"=== After Classic CV evaluation of complex models...\")\n print()\n\n average_scores_across_outer_folds_for_each_model = average_scores_and_best_scores[0]\n scores_of_best_model = average_scores_and_best_scores[1]\n\n best_model_name = scores_of_best_model[4][0]\n best_model_estim = scores_of_best_model[4][1]\n\n best_score = scores_of_best_model[0]\n best_score_dev = scores_of_best_model[1]\n best_cv_results = scores_of_best_model[2]\n # best_brier_score = scores_of_best_model[2]\n best_exec_time = scores_of_best_model[3]\n\n Dummy_score = Dummy_scores[0]\n Dummy_score_dev = Dummy_scores[1]\n Dummy_cv_results = Dummy_scores[2]\n # Dummy_brier_score = Dummy_scores[3]\n Dummy_exec_time = Dummy_scores[3]\n\n if best_model_name != 'DummyClf_2nd':\n # It's assumed best model's performance is\n # satistically better than that of DummyClf on this dataset\n print(\"DummyClassifier's scores -- '%s': %1.3f (%1.3f)\" % (\n scoring.strip('neg_'), Dummy_score, Dummy_score_dev))\n print(\"'%s' does better than DummyClassifier.\" % best_model_name)\n if best_exec_time < Dummy_exec_time:\n print(\"'%s' is quicker than DummyClf.\" % best_model_name)\n print()\n print()\n\n preprocessing = (encoding, scaler_tuple, featselector)\n\n if labels is not None:\n if not isinstance(labels, list):\n raise TypeError(\"Multiclass models need a list of string labels.\")\n else:\n print(\"You have labels:\", labels)\n all_models_and_parameters['labels'] = labels\n\n print(\"Defined dictionary with models, parameters and related data.\")\n print()\n\n tc.calibrate_best_model(\n X, y, X_train_transformed, X_test_transformed,\n y_train, y_test, auto_feat_eng_data['tt_index'], \n preprocessing, scores_of_best_model,\n all_models_and_parameters, n_splits, nb_epoch,\n scoring, models_data, d_name, random_state)\n else:\n sys.exit(\"Your best classifier is not a good classifier.\")\n\n\ndef perform_nested_cv_evaluation_and_calibration(\n auto_feat_eng_data, nested_cv_scoring, Y_type, labels=None,\n d_name=None, random_state=0, followup=False):\n \"\"\"\n # perform nested cross validation and calibration of best estimator.\n\n ---------------------------------------------------------\n auto_feat_eng_data:\n dictionary with encoder, scaler, feature selector,\n Pipeline.steps and train-test data\n scoring: scoring for model evaluation\n -- string ('roc_auc_score') or list of strings\n random_state: seed\n \"\"\"\n if isinstance(nested_cv_scoring, list) or isinstance(\n nested_cv_scoring, dict) or isinstance(nested_cv_scoring, tuple):\n raise TypeError(\"\"\"\n 'perform_nested_cv_evaluation_and_calibration' method allows only\n to perform single-metric evaluations.\n \"\"\")\n if nested_cv_scoring not in (None, 'accuracy', 'roc_auc', 'neg_log_loss'):\n raise ValueError(\"\"\"\n %s is not a valid nested_cv_scoring value for method\n 'perform_nested_cv_evaluation_and_calibration'. Valid options are\n ['accuracy', 'roc_auc', 'neg_log_loss']\"\"\" % nested_cv_scoring)\n\n print(\"### Probability Calibration of Best Estimator \"\n \"-- CalibratedClassifierCV with cv=cv (no prefit) ###\")\n print(\"### Nested Cross-Validation ###\")\n print(\"Models are trained and calibrated on the same data, train data,\\n\"\n \"calibration is evaluated on test data.\")\n print()\n print(\"RSCV.refit=False\")\n print(\"pipe.fit(train_data)\")\n\n # each one of these items need a check\n encoding = auto_feat_eng_data['encoding']\n scaler_tuple = auto_feat_eng_data['scaler']\n featselector = auto_feat_eng_data['feat_selector']\n steps = auto_feat_eng_data['steps']\n X_train_transformed, y_train, X_test_transformed, y_test = auto_feat_eng_data['data_arrays']\n X, y = auto_feat_eng_data['Xy']\n train_index, test_index = auto_feat_eng_data['tt_index']\n\n print()\n if followup:\n print(\"X_train_transformed shape: \", X_train_transformed.shape)\n print(\"X_test_transformed shape: \", X_test_transformed.shape)\n \n n_splits = au.select_nr_of_splits_for_kfold_cv()\n\n n_iter = au.select_nr_of_iterations()\n\n # Stratified folds preserve the percentage of samples for each class.\n inner_cv = StratifiedKFold(n_splits=n_splits, shuffle=True,\n random_state=random_state)\n outer_cv = StratifiedKFold(n_splits=n_splits, shuffle=True,\n random_state=random_state)\n\n ###\n\n # you should check Y for categorical values and\n # eventually label encode them...\n\n # Nested [RSCV] CV\n\n print()\n\n print(\"Metric:\", nested_cv_scoring)\n print(\"Calibration of untrained models -- CCCV 2nd\")\n print()\n\n # Evaluation of best modelwith nested CV -- inner: RSCV\n\n # if it comes out that the best model is LogReg, no comparison is needed\n\n best_atts = eu.best_model_initial_attributes(nested_cv_scoring, n_splits)\n\n best_score, best_score_dev, best_cv_results, best_model_name = best_atts\n\n best_exec_time = 31536000 # one year in seconds\n best_model = (best_model_name, None, None)\n\n Dummy_scores = []\n\n models_data = []\n names = []\n results = []\n\n scores_of_best_model = (best_score, best_score_dev, best_cv_results,\n best_exec_time, best_model)\n\n # Start evaluation process\n\n print()\n print(\"=== [task] Evaluation of DummyClassifier\")\n print()\n\n wtr = eu.calculate_sample_weight(y_train)\n\n print(\"=== 'sample_weight'\")\n print(wtr[:5])\n print(\"=== target train data sample\")\n print(y_train[:5])\n print()\n\n strategy = 'stratified' # 'most_frequent'\n\n average_scores_and_best_scores = eu.single_nested_rscv_evaluation(\n X_train_transformed, y_train, 'DummyClf_2nd',\n DummyClassifier(strategy=strategy), dict(), wtr,\n nested_cv_scoring, 0, inner_cv, outer_cv, dict(), scores_of_best_model,\n results, names, random_state)\n\n scores_of_best_model = average_scores_and_best_scores[1]\n\n Dummy_scores.append(scores_of_best_model[0]) # Dummy score -- ROC_AUC\n Dummy_scores.append(scores_of_best_model[1]) # Dummy score std\n Dummy_scores.append(scores_of_best_model[2]) # Dummy cv results\n Dummy_scores.append(scores_of_best_model[3]) # Dummy execution time\n # Dummy model's name and estimator\n Dummy_scores.append(scores_of_best_model[4])\n\n names = []\n results = []\n\n # Nested CV: its purpose is not to find best parameters, but\n # how well models would perform with their parameters tuned\n\n print()\n print(\"=== Nested CV [inner cv: RSCV]\")\n print()\n\n # we will collect the average of the scores on the k outer folds in\n # this dictionary with keys given by the names of the models\n # in 'models_and_parameters'\n average_scores_across_outer_folds_for_each_model = dict()\n\n # this holds for regression\n\n # update models_and_params dict according\n # to learning mode 'quick', 'standard', 'hard'\n\n models_and_parameters = models_and_params_for_nested_cv_tasks(\n Y_type, nested_cv_scoring, labels)\n\n average_scores_and_best_scores = eu.nested_rscv_model_evaluation(\n X_train_transformed, y_train, models_and_parameters,\n # {},\n nested_cv_scoring, n_iter, inner_cv, outer_cv,\n average_scores_across_outer_folds_for_each_model,\n scores_of_best_model, results, names, random_state)\n\n print()\n au.box_plots_of_models_performance(results, names)\n\n results = []\n names = []\n\n print()\n print(\"=== After Nested CV evaluation...\")\n print()\n\n average_scores_across_outer_folds_for_each_model = average_scores_and_best_scores[0]\n scores_of_best_model = average_scores_and_best_scores[1]\n\n best_model_name = scores_of_best_model[4][0]\n best_model_estim = scores_of_best_model[4][1]\n # no need to define a Keras build function here\n # best_nn_build_fn = scores_of_best_model[3][2]\n best_score = scores_of_best_model[0]\n best_score_dev = scores_of_best_model[1]\n best_cv_results = scores_of_best_model[2]\n best_exec_time = scores_of_best_model[3]\n\n print()\n print(\"Currently, best model is '%s' with score '%s': %1.3f (%1.3f)... :\"\n % (best_model_name, nested_cv_scoring.strip('neg_'), best_score,\n best_score_dev))\n print(\"... execution time: %.2fs\" % best_exec_time)\n print()\n print()\n\n print(\"======= Nested RSCV to evaluate more complex models\")\n print()\n\n # complex_models_and_parameters[name] = (model, rscv_params, dict())\n complex_models_and_parameters = dict()\n average_scores_across_outer_folds_complex = dict()\n\n # all_models_and_parameters = {}\n all_models_and_parameters = models_and_parameters\n\n if labels is not None:\n if not isinstance(labels, list):\n raise TypeError(\"Multiclass models need a list of string labels.\")\n else:\n print(\"You have labels:\", labels)\n all_models_and_parameters['labels'] = labels\n\n print(\"Defined dictionary with models, parameters and related data.\")\n print()\n\n # Compare to ensemble of instances of best model\n # after looping over standard models and once you have best model,\n # create ensemble of it\n\n bagging_estimator_name = ''\n\n if best_model_name not in {\n 'DecisionTreeClf_2nd', 'ExtraTreesClf_2nd', 'RandomForestClf_2nd',\n 'GBoostingClf_2nd', 'XGBClf_2nd', 'AdaBClf_2nd',\n 'Bagging_SVMClf_2nd'}:\n\n print(\"=== [task] Comparing best model to ensemble of instances \"\n \"of best model:\")\n print(\"BaggingClf(%s)\" % best_model_name)\n print()\n\n # steps = []\n # ...\n # steps.append(('feature_union', feature_union))\n\n bag_estim, bag_estim_name, bag_param_grid = create_ensemble_of_best_models(\n best_model_name, best_model_estim, random_state)\n\n # add bagging to dictionary of complex models\n\n complex_models_and_parameters[bag_estim_name] = (\n bag_estim, bag_param_grid\n )\n\n all_models_and_parameters[bag_estim_name] = (\n bag_estim, bag_param_grid\n )\n\n # Let's add some simple neural network\n\n print(\"=== [task] Comparing best model to simple Neural Network \"\n \"(with single or two hidden layers).\")\n print()\n\n input_dim = int(X_train_transformed.shape[1])\n\n if not followup:\n nb_epoch = au.select_nr_of_iterations('nn')\n else:\n nb_epoch = au.select_nr_of_iterations('nn', followup)\n\n keras_clf_name = \"KerasClf_2nd\"\n\n keras_nn_model, keras_param_grid = create_best_keras_clf_architecture(\n keras_clf_name, Y_type, labels, input_dim, nb_epoch, pgd.Keras_param_grid)\n\n complex_models_and_parameters[keras_clf_name] = (\n keras_nn_model, keras_param_grid)\n\n # Feed nested-cv function with dictionary of models and their params\n\n average_scores_and_best_scores_complex = eu.nested_rscv_model_evaluation(\n X_train_transformed, y_train, complex_models_and_parameters,\n nested_cv_scoring, n_iter, inner_cv, outer_cv,\n average_scores_across_outer_folds_complex, scores_of_best_model,\n results, names, random_state)\n\n print()\n au.box_plots_of_models_performance(results, names)\n\n print()\n print(\"=== After Nested CV evaluation of complex models...\")\n print()\n\n average_scores_across_outer_folds_complex =\\\n average_scores_and_best_scores_complex[0]\n scores_of_best_model = average_scores_and_best_scores_complex[1]\n\n best_model_name = scores_of_best_model[4][0]\n best_model_estim = scores_of_best_model[4][1]\n\n best_score = scores_of_best_model[0]\n best_score_dev = scores_of_best_model[1]\n best_cv_results = scores_of_best_model[2]\n best_exec_time = scores_of_best_model[3]\n\n Dummy_score = Dummy_scores[0]\n Dummy_score_dev = Dummy_scores[1]\n Dummy_cv_results = Dummy_scores[2]\n Dummy_exec_time = Dummy_scores[3]\n\n print(\"Currently, best model is '%s' with score '%s': %1.3f (%1.3f)... :\"\n % (best_model_name, nested_cv_scoring, best_score, best_score_dev))\n if best_model_name == keras_clf_name:\n best_nn_build_fn = scores_of_best_model[4][2]\n print(\"Best build function:\", best_nn_build_fn)\n print(\"... execution time: %.2fs\" % best_exec_time)\n print()\n\n if best_model_name != 'DummyClf_2nd':\n # best model is supposed to have passed some statistical test\n print(\"DummyClassifier's scores -- '%s': %1.3f (%1.3f)\" % (\n nested_cv_scoring, Dummy_score, Dummy_score_dev))\n print(\"'%s' does better than DummyClassifier.\" % best_model_name)\n print(\"Execution time of '%s': %.2fs\" % (\n best_model_name, best_exec_time))\n if best_exec_time < Dummy_exec_time:\n print(\"'%s' is quicker than DummyClf.\" % best_model_name)\n\n print()\n print()\n # input(\"Press key to continue...\")\n\n preprocessing = (encoding, scaler_tuple, featselector)\n\n tc.tune_calibrate_best_model(\n X, y, X_train_transformed, X_test_transformed,\n y_train, y_test, auto_feat_eng_data['tt_index'], \n preprocessing, scores_of_best_model,\n all_models_and_parameters, n_splits, n_iter, nb_epoch,\n nested_cv_scoring, models_data, d_name, random_state)\n\n else:\n sys.exit(\"Your best classifier is not a good classifier.\")\n\n\n# select your fraking strategy\ndef select_evaluation_strategy(\n auto_feat_eng_data, target, test_frac=0.3,\n odf=None, scoring='roc_auc', Y_type='binary', labels=None,\n d_name=None, random_state=0, learn='standard', mode='interactive'):\n \"\"\"\n # select evaluation strategy.\n\n ----------------------------------------------------------------------------\n 'auto_feat_eng_data': engineered data from split_and_X_encode function\n\n 'target': label # or single column dataframe with labels\n\n 'odf' : orginal dataframe with eventual feature engineering\n not causing data leakage\n\n 'scoring' : scoring for model evaluation\n\n 'Y_type' : type of target ; default: 'binary'\n\n 'labels' : labels for multiclass logistic regression metric\n\n 'random_state' : random state (seed)\n\n 'learn' : learn mode based on output of learning_strategy() fct\n\n 'mode' : model of machine learnin problem solution;\n default: 'interactive', else 'auto'\n \"\"\"\n if learn == 'quick':\n\n perform_classic_cv_evaluation_and_calibration(\n auto_feat_eng_data, scoring, Y_type, labels, \n d_name, random_state)\n\n msg = \"Are you satisfied with current results?\"\n\n if au.say_yes_or_no(msg) in {\"YES\", \"yes\", \"Y\", \"y\"}:\n print(\"Great! See you next time!\")\n print()\n else:\n\n if odf is not None: # df has been reduced in size\n\n msg = \"Do you want to use the full dataset?\"\n\n if au.say_yes_or_no(msg) in {\"YES\", \"yes\", \"Y\", \"y\"}:\n\n print()\n print(\"### Split and encode the whole dataframe\")\n print(\"Warning! No feature engineering implemented!\")\n print()\n\n split_enc_X_data = eu.split_and_X_encode(\n odf, target, test_frac, random_state)\n\n auto_feat_eng_data, scoring, Y_type, classes = split_enc_X_data\n\n else:\n print(\"### Use current data from small/smaller dataframe\")\n \n else: # df has not been reduced in size, full df\n # you already did you feature engineering\n\n print()\n print(\"### Using the whole dataframe\")\n print()\n\n perform_nested_cv_evaluation_and_calibration(\n auto_feat_eng_data, scoring, Y_type, labels, \n d_name, random_state, True)\n\n elif learn == 'standard':\n\n perform_nested_cv_evaluation_and_calibration(\n auto_feat_eng_data, scoring, Y_type, labels, d_name, random_state)\n\n else:\n # learn == 'large'\n perform_classic_cv_evaluation_and_calibration(\n auto_feat_eng_data, scoring, Y_type, labels, d_name,\n random_state)\n","sub_path":"autoclf/classification/evaluate.py","file_name":"evaluate.py","file_ext":"py","file_size_in_byte":30574,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"292861808","text":"from datetime import datetime, timedelta\nfrom sqlalchemy.dialects.postgresql import insert\n\nfrom .. import db, create_async_app\nfrom ..devices.models import Device\nfrom ..readings.models import Reading, AggregatedReading\n\napp = create_async_app()\n\nAGGREGATIONS = {\n 'minute': {\n 'time_interval': 'minute',\n 'group_func': db.func.time_bucket('2 minute', Reading.timestamp),\n 'agg_func': db.func.avg(Reading.value),\n 'timestamp_range': lambda r: (datetime(r.year, r.month, r.day, r.hour, r.minute),\n datetime(r.year, r.month, r.day, r.hour, r.minute) + timedelta(minutes=2))\n },\n 'hour_difference': {\n 'time_interval': 'hour',\n 'group_func': db.func.date_trunc('hour', Reading.timestamp),\n 'agg_func': (db.func.last(Reading.value, Reading.timestamp)\n - db.func.first(Reading.value, Reading.timestamp)),\n 'timestamp_range': lambda r: (datetime(r.year, r.month, r.day, r.hour),\n datetime(r.year, r.month, r.day, r.hour) + timedelta(hours=1))\n },\n 'day_difference': {\n 'time_interval': 'day',\n 'group_func': db.func.date_trunc('day', Reading.timestamp),\n 'agg_func': (db.func.last(Reading.value, Reading.timestamp)\n - db.func.first(Reading.value, Reading.timestamp)),\n 'timestamp_range': lambda r: (datetime(r.year, r.month, r.day),\n datetime(r.year, r.month, r.day) + timedelta(days=1))\n },\n 'day_avg': {\n 'time_interval': 'day',\n 'group_func': db.func.date_trunc('day', Reading.timestamp),\n 'agg_func': db.func.avg(Reading.value),\n 'timestamp_range': lambda r: (datetime(r.year, r.month, r.day),\n datetime(r.year, r.month, r.day) + timedelta(days=1))\n },\n}\n\nTOPIC_AGGREGATION_MAP = {\n 'total_kwh': [\n {'name': 'kwh_per_hour', **AGGREGATIONS['hour_difference']},\n {'name': 'kwh_per_day', **AGGREGATIONS['day_difference']},\n ],\n 'net_kwh': [\n {'name': 'net_kwh_per_hour', **AGGREGATIONS['hour_difference']},\n {'name': 'net_kwh_per_day', **AGGREGATIONS['day_difference']},\n ],\n 'watts': [\n {'name': 'watts', **AGGREGATIONS['minute']},\n ],\n 'net_watts': [\n {'name': 'net_watts', **AGGREGATIONS['minute']},\n ],\n 'space_temp_f': [\n {'name': 'space_temp_f', **AGGREGATIONS['minute']},\n ],\n 'discharge_air_temp_f': [\n {'name': 'discharge_air_temp_f', **AGGREGATIONS['minute']}\n ],\n 'primary_flow_cfm': [\n {'name': 'primary_flow_cfm', **AGGREGATIONS['minute']},\n ],\n 'rtu_mode': [\n {'name': 'rtu_mode', **AGGREGATIONS['minute']},\n ],\n 'rtu_is_active': [\n {'name': 'rtu_is_active', **AGGREGATIONS['minute']},\n {'name': 'rtu_is_active_day_avg', **AGGREGATIONS['day_avg']},\n ],\n 'trigger_temp_f': [\n {'name': 'trigger_temp_f', **AGGREGATIONS['minute']},\n ],\n 'setpoint_f': [\n {'name': 'setpoint_f', **AGGREGATIONS['minute']},\n ],\n 'accepting_air': [\n {'name': 'accepting_air', **AGGREGATIONS['minute']},\n {'name': 'accepting_air_day_avg', **AGGREGATIONS['day_avg']},\n ],\n 'gas_cu_ft': [\n {'name': 'gas_cu_ft_hour', **AGGREGATIONS['hour_difference']},\n {'name': 'gas_cu_ft_day', **AGGREGATIONS['day_difference']},\n ]\n}\n\n\ndef aggregate_readings():\n with app.app_context():\n previous_minute_end = datetime.now().replace(second=0, microsecond=0)\n previous_minute_start = previous_minute_end - timedelta(minutes=1)\n\n prev_minute_readings = db.engine.execute(\n db.select([Reading]).where(\n db.and_(Reading.timestamp >= previous_minute_start, Reading.timestamp < previous_minute_end)\n )).fetchall()\n\n aggregated_readings = []\n for reading in prev_minute_readings:\n aggregations = TOPIC_AGGREGATION_MAP.get(reading['topic'], [])\n if not aggregations:\n continue\n\n tenant_id = db.engine.execute(\n db.select([Device.tenant_id]).where(Device.id == reading['device_id'])\n ).scalar()\n\n if not tenant_id:\n continue\n\n for aggregation in aggregations:\n result = db.engine.execute(\n db.select([\n aggregation['group_func'].label('time_group'),\n aggregation['agg_func'].label('agg_value')\n ]).where(db.and_(\n Reading.timestamp >= aggregation['timestamp_range'](reading['timestamp'])[0],\n Reading.timestamp < aggregation['timestamp_range'](reading['timestamp'])[1],\n Reading.device_id == reading['device_id'],\n Reading.topic == reading['topic']\n )).group_by('time_group')\n ).first()\n\n aggregated_readings.append({\n 'device_id': reading['device_id'],\n 'tenant_id': tenant_id,\n 'topic': aggregation['name'],\n 'value': result.agg_value,\n 'timestamp': result.time_group\n })\n\n insert_stmt = insert(AggregatedReading).values(aggregated_readings)\n insert_stmt = insert_stmt.on_conflict_do_update(\n index_elements=[AggregatedReading.device_id,\n AggregatedReading.timestamp,\n AggregatedReading.topic,\n AggregatedReading.tenant_id],\n set_={\"value\": insert_stmt.excluded.value}\n )\n\n db.engine.execute(insert_stmt)\n","sub_path":"app/readings/aggregate.py","file_name":"aggregate.py","file_ext":"py","file_size_in_byte":5763,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"288884613","text":"import requests\nimport asyncio\nimport discord\n\n\ndef get_json():\n user = \"smartel99\"\n repo = \"pybot2\"\n\n url = 'https://api.github.com/repos/{}/{}'.format(user, repo)\n\n resp = requests.get(url)\n\n if resp.status_code != 200:\n return None\n return resp.json()\n\n\nasync def get_repo_link(bot):\n\n\n repo_data = get_json()\n if not repo_data:\n await bot.say(\"Error trying to connect to server\")\n return\n clone = repo_data.get('clone_url', 'ERROR: NO DATA')\n\n await bot.say(\"To clone smartel99's repo named pybot2, \"\n \"the command is: \"\n \"git clone {}\".format(clone))\n\n\ndef get_updated_at():\n file = get_json()\n if not file:\n return None\n return file.get(\"updated_at\", 'ERROR: NO DATA')\n\n\ndef get_date():\n with open('C:\\pybot_token\\github_last_updated.txt', 'r') as fin:\n for line in fin:\n return line\n\n\ndef set_updated(current):\n with open('C:\\pybot_token\\github_last_updated.txt', 'w') as fin:\n fin.write(current)\n\n\nasync def check_if_update_ready(bot):\n has_failed = False\n await bot.wait_until_ready()\n while not bot.is_closed:\n current = get_updated_at()\n if not current or current == 'ERROR: NO DATA':\n if not has_failed:\n has_failed = True\n print('Failed to check for updates')\n return\n from_txt = get_date()\n if has_failed:\n has_failed = False\n if current != from_txt:\n print('Update Available')\n set_updated(current)\n else:\n print(\"no update available\")\n\n await asyncio.sleep(300)\n","sub_path":"github_thingy.py","file_name":"github_thingy.py","file_ext":"py","file_size_in_byte":1668,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"505548091","text":"import os.path\r\n\r\nimport cherrypy\r\n\r\nfrom webapp.libs.tools.satool import SATool\r\nfrom webapp.libs.tools.makotool import MakoTool\r\nfrom webapp.libs.plugins.saplugin import SAEnginePlugin\r\nfrom webapp.libs.plugins.makoplugin import MakoTemplatePlugin\r\nfrom webapp.libs.tools.authtool import AuthTool\r\n\r\ncur_dir = os.path.abspath(os.path.dirname(__file__))\r\ntemplate_dir = os.path.join(cur_dir, 'templates')\r\ntemplate_cache_dir = os.path.join(cur_dir, 'templates', '.cache')\r\nconf_dir = os.path.join(cur_dir, 'conf')\r\nconf_path = os.path.join(cur_dir, 'conf', 'server.conf')\r\n\r\ncherrypy.tools.db = SATool()\r\ncherrypy.tools.render = MakoTool()\r\ncherrypy.tools.auth = AuthTool()\r\n\r\nfrom webapp.app import InsideOutApp, InsideOutAppEvents, InsideOutAppMassage\r\napp = InsideOutApp()\r\napp.events = InsideOutAppEvents()\r\napp.massage = InsideOutAppMassage()\r\n\r\nfrom webapp.cms import InsideOutCms, InsideOutCmsTeachers, InsideOutCmsEvents, InsideOutCmsSchedule\r\napp.cms = InsideOutCms()\r\napp.cms.teachers = InsideOutCmsTeachers()\r\napp.cms.events = InsideOutCmsEvents()\r\napp.cms.schedule = InsideOutCmsSchedule()\r\n\r\ncherrypy.tree.mount(app, '/', conf_path)\r\n\r\nMakoTemplatePlugin(cherrypy.engine, template_dir, template_cache_dir).subscribe()\r\n\r\ncherrypy.engine.db = SAEnginePlugin(cherrypy.engine, \"mysql://%s:%s@%s:%s/%s?charset=utf8\" % (\r\n os.environ['INSIDEOUT_SQL_USER'],\r\n os.environ['INSIDEOUT_SQL_PASSWORD'],\r\n os.environ['INSIDEOUT_SQL_ADDRESS'],\r\n os.environ['INSIDEOUT_SQL_PORT'],\r\n os.environ['INSIDEOUT_SQL_DATABASE']\r\n))\r\n\r\ncherrypy.engine.db.subscribe()\r\n\r\nif os.environ['INSIDEOUT_DEBUG'] == '1':\r\n cherrypy.engine.start()\r\n","sub_path":"webapp/server.py","file_name":"server.py","file_ext":"py","file_size_in_byte":1647,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"643586825","text":"\"\"\"\nSignal processing, creation and plotting.\n\nAnalysis of data and generation of simulated experiments.\n\"\"\"\n__license__ = \"Joseph C. Slater\"\n\n__docformat__ = 'reStructuredText'\n\n# import warnings\n\nimport numpy as np\nimport scipy as sp\nimport scipy.fftpack as fftpack\nimport scipy.linalg as la\nimport matplotlib.pyplot as plt\nimport scipy.integrate as spi\nimport scipy.signal as signal\n\n\"\"\"\nNotes:\n------\nSept. 3, 2016\nDevelopment of windows in scipy.signal has been rapid and\ndetermining what I should build into this module, or simply leverage\nfrom scipy.signal has been a moving target.\nIt's now apparent that creating or returning a window is pointless. Further,\nApplying should be a relatively simple code obviating much of any need for the\ncode here.\n\nThe cross spectrum analysis formerly lacking is now available, periodogram\nis usually the best option, however not with impulse excitations. See\n`scipy.signal` for this. Unfortunately, the conventions in this module are not\nconsistent with `scipy.signal`. They follow those of `python-control`\n\nFRF calculation is typically trivial, Hv being an expected gap long term\nMIMO FRF calculation is an open question. Pretty printing of FRFs is always\na welcome tool.\n\nSystem ID is likely the remaining missing aspect at this time.\n\nIn order to be consistent with the Control Systems Library, increasing time\nor increasing frequency steps positively with increased column number (one\ndimension). Rows (0 dimension)\ncorrespond to appropriate channels, output numbers, etc.\n\nFor cross spectrum data (cross spectrum density, frequency response function)\nthe 2 dimension represents the input channel.\n\nThe last dimension (2 or 3) indexes each data instance (experiment). That means\nthat an unaveraged cross spectrum density has dimension 4. If there is only a\nsingle input channel, it is imperative to insist the dimention exist, even if\nonly length 1. This is analagous to a vector being Nx1 versus simply a\n1-D array of length 1.\n\nhttp://python-control.readthedocs.io/en/latest/conventions.html#time-series-data\n\nProblem: This hasn't been fully implemented.\n\"\"\"\n\n\ndef window(x, windowname='hanning', normalize=False):\n r\"\"\"Create leakage window.\n\n Create a window of length :math:`x`, or a window sized to match\n :math:`x` that :math:`x\\times w` is the windowed result.\n\n Parameters\n ----------\n x: integer, float array\n | If integer- number of points in desired hanning windows.\n | If array- array provides size of window returned.\n windowname: string\n One of: hanning, hamming, blackman, flatwin, boxwin\n normalize: bool, optional(False)\n Adjust power level (for use in ASD) to 1\n\n Returns\n -------\n w: float array\n | window array of size x\n | window array. Windowed array is then :math:`x\\times w`\n\n Examples\n --------\n >>> import numpy as np\n >>> import vibrationtesting as vt\n >>> import matplotlib.pyplot as plt\n >>> sample_freq = 1e3\n >>> tfinal = 5\n >>> fs = 100\n >>> A = 10\n >>> freq = 5\n >>> noise_power = 0.001 * sample_freq / 2\n >>> time = np.reshape(np.arange(0, tfinal, 1/sample_freq),(1,-1))\n >>> xsin = A*np.sin(2*np.pi*freq*time)\n >>> xcos = A*np.cos(2*np.pi*freq*time) # assembling individual records.\n >>> x=np.dstack((xsin,xcos)) # assembling individual records. vstack\n >>> xw=vt.hanning(x)*x\n >>> fig, (ax1, ax2) = plt.subplots(2,1)\n >>> ax1.plot(time.T,x[:,:,1].T)\n []\n >>> ax1.set_ylim([-20, 20])\n (-20, 20)\n >>> ax1.set_title('Original (raw) data.')\n Text(0.5,1,'Original (raw) data.')\n >>> ax1.set_ylabel('$x(t)$')\n Text(0,0.5,'$x(t)$')\n >>> ax2.plot(time[0,:],xw[0,:],time[0,:],vt.hanning(x)[0,:]*A,'--',\n ... time[0,:],-vt.hanning(x)[0,:]*A,'--')\n []\n >>> ax2.set_ylabel('Hanning windowed $x(t)$')\n Text(0,0.5,'Hanning windowed $x(t)$')\n >>> ax2.set_xlabel('time')\n Text(0.5,0,'time')\n >>> ax2.set_title('Effect of window. Note the scaling to conserve ASD amplitude')\n Text(0.5,1,'Effect of window. Note the scaling to conserve ASD amplitude')\n >>> fig.tight_layout()\n\n \"\"\"\n if isinstance(x, (list, tuple, np.ndarray)):\n \"\"\"Create Hanning windowing array of dimension `n` by `N` by `nr`\n where `N` is number of data points and `n` is the number of number of\n inputs or outputs and `nr` is the number of records.\"\"\"\n\n swap = 0\n if len(x.shape) == 1:\n # We have either a scalar or 1D array\n if x.shape[0] == 1:\n print(\"x is a scalar... and shouldn\\'t have entered this \\\n part of the loop.\")\n else:\n N = len(x)\n\n f = window(N, windowname=windowname)\n\n elif len(x.shape) == 3:\n\n if x.shape[0] > x.shape[1]:\n x = np.swapaxes(x, 0, 1)\n swap = 1\n print('You shouldn\\'t do that.')\n print('The 1 dimension is the time (or frequency) \\\n incrementing dimension.')\n print('Swapping axes temporarily to be compliant with \\\n expectations. I\\'ll fix them in your result')\n\n N = x.shape[1]\n f = window(N, windowname=windowname)\n f, _, _ = np.meshgrid(f, np.arange(\n x.shape[0]), np.arange(x.shape[2]))\n if swap == 1:\n f = np.swapaxes(f, 0, 1)\n\n elif len(x.shape) == 2:\n\n if x.shape[0] > x.shape[1]:\n x = np.swapaxes(x, 0, 1)\n swap = 1\n print('You shouldn\\'t do that.')\n print('The 1 dimension is the time (or frequency) ' +\n 'incrementing dimension.')\n print('Swapping axes temporarily to be compliant with ' +\n 'expectations.')\n print('I\\'ll reluctantly return a transposed result.')\n\n f = window(x.shape[1], windowname=windowname)\n f, _ = np.meshgrid(f, np.arange(x.shape[0]))\n if swap == 1:\n f = np.swapaxes(f, 0, 1)\n\n else:\n N = x\n if windowname is 'hanning':\n f = np.sin(np.pi * np.arange(N) / (N - 1))**2 * np.sqrt(8 / 3)\n elif windowname is 'hamming':\n f = (0.54 - 0.46 * np.cos(2 * np.pi * (np.arange(N)) / (N - 1)))\\\n * np.sqrt(5000 / 1987)\n elif windowname is 'blackman':\n print('blackman')\n f = (0.42 - 0.5 * np.cos(2 * np.pi * (np.arange(N) + .5) / (N))\n + .08 * np.cos(4 * np.pi * (np.arange(N) + .5) / (N)))\\\n * np.sqrt(5000 / 1523)\n elif windowname is 'flatwin':\n f = 1.0 - 1.933 * np.cos(2 * np.pi * (np.arange(N)) / (N - 1))\\\n + 1.286 * np.cos(4 * np.pi * (np.arange(N)) / (N - 1))\\\n - 0.338 * np.cos(6 * np.pi * (np.arange(N)) / (N - 1))\\\n + 0.032 * np.cos(8 * np.pi * (np.arange(N)) / (N - 1))\n elif windowname is 'boxwin':\n f = np.ones((1, N))\n else:\n f = np.ones((1, N))\n print(\"I don't recognize window name \", windowname, \". Sorry.\")\n\n if normalize is True:\n f = f / la.norm(f) * np.sqrt(N)\n return f\n\n\ndef hanning(x, normalize=False):\n r\"\"\"Return hanning window.\n\n Create a hanning window of length :math:`x`, or a hanning window sized to\n match :math:`x` that :math:`x\\times w` is the windowed result.\n\n Parameters\n ----------\n x: integer, float array\n | If integer- number of points in desired hanning windows.\n | If array- array provides size of window returned.\n windowname: string\n One of: hanning, hamming, blackman, flatwin, boxwin\n normalize: bool, optional(False)\n Adjust power level (for use in ASD) to 1\n\n Returns\n -------\n w: float array\n | window array of size x\n | window array. Windowed array is then :math:`x\\times w`\n\n Examples\n --------\n >>> import numpy as np\n >>> import vibrationtesting as vt\n >>> import matplotlib.pyplot as plt\n >>> sample_freq = 1e3\n >>> tfinal = 5\n >>> fs = 100\n >>> A = 10\n >>> freq = 5\n >>> noise_power = 0.001 * sample_freq / 2\n >>> time = np.reshape(np.arange(0, tfinal, 1/sample_freq),(1,-1))\n >>> xsin = A*np.sin(2*np.pi*freq*time)\n >>> xcos = A*np.cos(2*np.pi*freq*time)\n >>> x=np.dstack((xsin,xcos)) # assembling individual records. vstack\n >>> xw=vt.hanning(x)*x\n >>> fig, (ax1, ax2) = plt.subplots(2, 1)\n >>> ax1.plot(time.T,x[:,:,1].T)\n []\n >>> ax1.set_ylim([-20, 20])\n (-20, 20)\n >>> ax1.set_title('Unwindowed data, 2 records.')\n Text(0.5,1,'Unwindowed data, 2 records.')\n >>> ax1.set_ylabel('$x(t)$')\n Text(0,0.5,'$x(t)$')\n >>> ax2.plot(time[0,:],xw[0,:],time[0,:],vt.hanning(x)[0,:]*A,\n ... '--',time[0,:],-vt.hanning(x)[0,:]*A,'--')\n []\n >>> ax2.set_ylabel('Hanning windowed $x(t)$')\n Text(0,0.5,'Hanning windowed $x(t)$')\n >>> ax2.set_xlabel('time')\n Text(0.5,0,'time')\n >>> ax2.set_title('Effect of window. Note the scaling to conserve ASD amplitude')\n Text(0.5,1,'Effect of window. Note the scaling to conserve ASD amplitude')\n >>> fig.tight_layout()\n\n \"\"\"\n if isinstance(x, (list, tuple, np.ndarray)):\n \"\"\"Create Hanning windowing array of dimension n by N by nr\n where N is number of data points and n is the number of number of\n inputs or outputs and nr is the number of records.\"\"\"\n\n swap = 0\n if len(x.shape) == 1:\n # We have either a scalar or 1D array\n if x.shape[0] == 1:\n print(\"x is a scalar... and shouldn\\'t have \\\n entered this part of the loop.\")\n else:\n N = len(x)\n f = hanning(N)\n\n elif len(x.shape) == 3:\n # print('a')\n # print(f.shape)\n\n if x.shape[0] > x.shape[1]:\n x = np.swapaxes(x, 0, 1)\n swap = 1\n print('Swapping axes temporarily to be compliant with \\\n expectations. I\\'ll fix them in your result')\n\n f = hanning(x.shape[1])\n f, _, _ = np.meshgrid(f, np.arange(\n x.shape[0]), np.arange(x.shape[2]))\n if swap == 1:\n f = np.swapaxes(f, 0, 1)\n\n elif len(x.shape) == 2:\n # f,_=np.meshgrid(f[0,:],np.arange(x.shape[0]))\n # print('b')\n # print('length = 2')\n # print(x.shape)\n if x.shape[0] > x.shape[1]:\n x = np.swapaxes(x, 0, 1)\n swap = 1\n print('Swapping axes temporarily to be compliant with \\\n expectations. I\\'ll fix them in your result')\n f = hanning(x.shape[1])\n f, _ = np.meshgrid(f, np.arange(x.shape[0]))\n if swap == 1:\n f = np.swapaxes(f, 0, 1)\n else:\n # print(x)\n # Create hanning window of length x\n N = x\n # print(N)\n f = np.sin(np.pi * np.arange(N) / (N - 1))**2 * np.sqrt(8 / 3)\n if normalize is True:\n f = f / la.norm(f) * np.sqrt(N)\n return f\n\n\ndef blackwin(x):\n \"\"\"Return the n point Blackman window.\n\n Returns x as the Blackman windowing array x_window\n The windowed signal is then x*x_window\n \"\"\"\n print('blackwin is untested')\n if isinstance(x, (list, tuple, np.ndarray)):\n n = x.shape[1]\n f = blackwin(n)\n\n if len(x.shape) == 3:\n f, _, _ = np.meshgrid(f[0, :], np.arange(\n x.shape[0]), np.arange(x.shape[2]))\n else:\n f, _ = np.meshgrid(f[0, :], np.arange(x.shape[0]))\n else:\n n = x\n f = np.reshape((0.42 - 0.5 * np.cos(2 * np.pi * (np.arange(n) + .5)) /\n (n) + .08 * np.cos(4 * np.pi * (np.arange(n) + .5)) /\n (n)) * np.sqrt(5000 / 1523), (1, -1))\n f = f / la.norm(f) * np.sqrt(n)\n return f\n\n\ndef expwin(x, ts=.75):\n \"\"\"Return the n point exponential window.\n\n Returns x as the expwin windowing array x_windowed\n The windowed signal is then x*x_window\n The optional second argument set the 5% \"settling time\" of the window.\n Default is ts=0.75\n \"\"\"\n print('expwin is untested')\n tc = -ts / np.log(.05)\n if isinstance(x, (list, tuple, np.ndarray)):\n n = x.shape[1]\n f = expwin(n)\n\n if len(x.shape) == 3:\n f, _, _ = np.meshgrid(f[0, :], np.arange(\n x.shape[0]), np.arange(x.shape[2]))\n else:\n f, _ = np.meshgrid(f[0, :], np.arange(x.shape[0]))\n else:\n n = x\n v = (n - 1) / n * np.arange(n) + (n - 1) / n / 2\n f = np.exp(-v / tc / (n - 1))\n f = f / la.norm(f) * np.sqrt(n)\n f = np.reshape(f, (1, -1))\n f = f / la.norm(f) * np.sqrt(n)\n\n return f\n\n\ndef hammwin(x):\n \"\"\"Return the n point hamming window.\n\n Returns x as the hamming windowingarray x_windowed\n The windowed signal is then x*x_window\n \"\"\"\n print('hammwin is untested')\n if isinstance(x, (list, tuple, np.ndarray)):\n n = x.shape[1]\n f = hammwin(n)\n\n if len(x.shape) == 3:\n f, _, _ = np.meshgrid(f[0, :], np.arange(\n x.shape[0]), np.arange(x.shape[2]))\n else:\n f, _ = np.meshgrid(f[0, :], np.arange(x.shape[0]))\n else:\n\n n = x\n f = np.reshape((0.54 - 0.46 * np.cos(2 * np.pi * (np.arange(n)) /\n (n - 1))) * np.sqrt(5000 / 1987),\n (1, -1))\n f = f / la.norm(f) * np.sqrt(n)\n\n return f\n\n\ndef flatwin(x):\n \"\"\"Return the n point flat top window.\n\n x_windows=flatwin(x)\n Returns x as the flat top windowing array x_windowed\n The windowed signal is then x*x_window\n McConnell, K. G., \"Vibration Testing: Theory and Practice,\" Wiley, 1995.\n \"\"\"\n print('flatwin is untested')\n if isinstance(x, (list, tuple, np.ndarray)):\n n = x.shape[1]\n f = flatwin(n)\n\n if len(x.shape) == 3:\n f, _, _ = np.meshgrid(f[0, :], np.arange(\n x.shape[0]), np.arange(x.shape[2]))\n else:\n f, _ = np.meshgrid(f[0, :], np.arange(x.shape[0]))\n else:\n\n n = x\n f = np.reshape(\n (1.0 - 1.933 * np.cos(2 * np.pi * (np.arange(n)) / (n - 1))\n + 1.286 * np.cos(4 * np.pi * (np.arange(n)) / (n - 1))\n - 0.338 * np.cos(6 * np.pi * (np.arange(n)) / (n - 1))\n + 0.032 * np.cos(8 * np.pi * (np.arange(n)) / (n - 1))),\n (1, -1))\n f = f / la.norm(f) * np.sqrt(n)\n\n return f\n\n\ndef boxwin(x):\n \"\"\"Return the n point box window (uniform).\n\n Returns x as the boxwin windowing array x_windowed\n The windowed signal is then x*x_window\n \"\"\"\n print('boxwin is untested')\n if isinstance(x, (list, tuple, np.ndarray)):\n n = x.shape[1]\n f = boxwin(n)\n\n if len(x.shape) == 3:\n f, _, _ = np.meshgrid(f[0, :], np.arange(\n x.shape[0]), np.arange(x.shape[2]))\n else:\n f, _ = np.meshgrid(f[0, :], np.arange(x.shape[0]))\n else:\n\n n = x\n # f=np.reshape((1.0-1.933*np.cos(2*np.pi*(np.arange(n))/(n-1))+1.286*np.cos(4*np.pi*(np.arange(n))/(n-1))-0.338*np.cos(6*np.pi*(np.arange(n))/(n-1))+0.032*np.cos(8*np.pi*(np.arange(n))/(n-1))),(1,-1))\n f = np.reshape(np.ones((1, n)), (1, -1))\n f = f / la.norm(f) * np.sqrt(n)\n\n return f\n\n\ndef hannwin(*args, **kwargs):\n \"\"\"Alternative for function `hanning`.\"\"\"\n return hanning(*args, **kwargs)\n\n\ndef asd(x, t, windowname=\"none\", ave=bool(True)):\n \"\"\"Return autospectrum (power spectrum) density of a signal x.\n\n Parameters\n ----------\n x : float array\n Data array (n x N x m) where n is the number of sensors, m the\n number of experiments.\n t : float array\n Time array (1 x N)\n windowname : string\n Name of windowing function to use. See `window`.\n ave : bool, optional(True)\n Average result or not?\n\n Returns\n -------\n f : float array\n Frequency vector (1 x N)\n Pxx : float array\n Autospectrum (n x N) or (n x N x m) if not averaged.\n\n Examples\n --------\n >>> from scipy import signal\n >>> import numpy as np\n >>> import matplotlib.pyplot as plt\n >>> import vibrationtesting as vt\n >>> import numpy.linalg as la\n\n Generate a 5 second test signal, a 10 V sine wave at 50 Hz, corrupted by\n 0.001 V**2/Hz of white noise sampled at 1 kHz.\n\n >>> sample_freq = 1e3\n >>> tfinal = 5\n >>> sig_freq=50\n >>> A=10\n >>> noise_power = 0.0001 * sample_freq / 2\n >>> noise_power = A/1e12\n >>> time = np.arange(0,tfinal,1/sample_freq)\n >>> time = np.reshape(time, (1, -1))\n >>> x = A*np.sin(2*np.pi*sig_freq*time)\n >>> x = x + np.random.normal(scale=np.sqrt(noise_power),\n ... size=(1, time.shape[1]))\n >>> fig, (ax1, ax2) = plt.subplots(2,1)\n >>> ax1.plot(time[0,:],x[0,:])\n []\n >>> ax1.set_title('Time history')\n Text(0.5,1,'Time history')\n >>> ax1.set_xlabel('Time (sec)')\n Text(0.5,0,'Time (sec)')\n >>> ax1.set_ylabel('$x(t)$')\n Text(0,0.5,'$x(t)$')\n\n Compute and plot the autospectrum density.\n\n >>> freq_vec, Pxx = vt.asd(x, time, windowname=\"hanning\", ave=bool(False))\n >>> ax2.plot(freq_vec, 20*np.log10(Pxx[0,:]))\n []\n >>> ax2.set_ylim([-400, 100])\n (-400, 100)\n >>> ax2.set_xlabel('frequency (Hz)')\n Text(0.5,0,'frequency (Hz)')\n >>> ax2.set_ylabel('PSD (V**2/Hz)')\n Text(0,0.5,'PSD (V**2/Hz)')\n\n If we average the last half of the spectral density, to exclude the\n peak, we can recover the noise power on the signal.\n\n \"\"\"\n f, Pxx = crsd(x, x, t, windowname=windowname, ave=ave)\n Pxx = Pxx.real\n return f, Pxx\n\n\ndef crsd(x, y, t, windowname=\"none\", ave=bool(True)):\n \"\"\"\n Calculate the cross spectrum (power spectrum) density between two signals.\n\n Parameters\n ----------\n x, y : arrays\n Data array (n x N x m) where n is the number of sensors, m the\n number of experiments.\n t : array\n Time array (1 x N)\n windowname : string\n Name of windowing function to use. See `window`.\n ave : bool, optional\n Average result or not?\n\n Returns\n -------\n f : array\n Frequency vector (1 x N)\n Pxy : array\n Autospectrum (n x N) or (n x N x m) if not averaged.\n\n Examples\n --------\n >>> from scipy import signal\n >>> import numpy as np\n >>> import matplotlib.pyplot as plt\n >>> import vibrationtesting as vt\n >>> import numpy.linalg as la\n\n Generate a 5 second test signal, a 10 V sine wave at 50 Hz, corrupted by\n 0.001 V**2/Hz of white noise sampled at 1 kHz.\n\n >>> sample_freq = 1e3\n >>> tfinal = 5\n >>> sig_freq=50\n >>> A=10\n >>> noise_power = 0.0001 * sample_freq / 2\n >>> noise_power = A/1e12\n >>> time = np.arange(0,tfinal,1/sample_freq)\n >>> time = np.reshape(time, (1, -1))\n >>> x = A*np.sin(2*np.pi*sig_freq*time)\n >>> x = x + np.random.normal(scale=np.sqrt(noise_power),\n ... size=(1, time.shape[1]))\n >>> fig = plt.figure()\n >>> plt.subplot(2,1,1)\n \n >>> plt.plot(time[0,:],x[0,:])\n []\n >>> plt.title('Time history')\n Text(0.5,1,'Time history')\n >>> plt.xlabel('Time (sec)')\n Text(0.5,0,'Time (sec)')\n >>> plt.ylabel('$x(t)$')\n Text(0,0.5,'$x(t)$')\n\n Compute and plot the autospectrum density.\n >>> freq_vec, Pxx = vt.asd(x, time, windowname=\"hanning\", ave=bool(False))\n >>> plt.subplot(2,1,2)\n \n >>> plt.plot(freq_vec, 20*np.log10(Pxx[0,:]))\n []\n >>> plt.ylim([-400, 100])\n (-400, 100)\n >>> plt.xlabel('frequency (Hz)')\n Text(0.5,0,'frequency (Hz)')\n >>> plt.ylabel('PSD (V**2/Hz)')\n Text(0,0.5,'PSD (V**2/Hz)')\n >>> fig.tight_layout()\n\n \"\"\"\n # t_shape = t.shape\n t = t.flatten()\n if len(t) == 1:\n dt = t\n else:\n dt = t[2] - t[1]\n\n if dt <= 0:\n print('You sent in bad data. Delta t is negative. \\\n Please check your inputs.')\n\n if len(x.shape) == 1:\n x = np.expand_dims(x, axis=0)\n x = np.expand_dims(x, axis=2)\n y = np.expand_dims(y, axis=0)\n y = np.expand_dims(y, axis=2)\n n = x.shape[1]\n\n if windowname is False or windowname.lower() is \"none\":\n win = 1\n else:\n # print('This doesn\\'t work yet')\n windowname = windowname.lower()\n win = 1\n if windowname == \"hanning\":\n win = window(x, windowname='hanning')\n elif windowname == \"blackwin\":\n win = window(x, windowname='blackwin')\n elif windowname == \"boxwin\":\n win = window(x, windowname='boxwin')\n elif windowname == \"expwin\":\n win = window(x, windowname='expwin')\n elif windowname == \"hammwin\":\n win = window(x, windowname='hamming')\n elif windowname == \"triwin\":\n win = window(x, windowname='triwin')\n elif windowname == \"flatwin\":\n win = window(x, windowname='flatwin')\n\n y = y * win\n x = x * win\n del win\n\n ffty = np.fft.rfft(y, n, axis=1) * dt\n fftx = np.fft.rfft(x, n, axis=1) * dt\n\n Pxy = np.conj(fftx) * ffty / (n * dt) * 2\n\n if len(Pxy.shape) == 3 and Pxy.shape[2] > 1 and ave:\n Pxy = np.mean(Pxy, 2)\n\n nfreq = 1 / dt / 2\n f = np.linspace(0, nfreq, Pxy.shape[1]) # /2./np.pi\n\n return f, Pxy\n\n\ndef frfest(x, f, dt, windowname=\"hanning\", ave=bool(True), Hv=bool(False)):\n r\"\"\"Return freq, H1, H2, coh, Hv.\n\n Estimates the :math:`H(j\\omega)` Frequency Response Functions (FRFs)\n between :math:`x` and :math:`f`.\n\n Parameters\n ----------\n x : float array\n output or response of system\n f : float array\n input to system\n dt : float\n time step of samples\n windowname : string\n One of: hanning, hamming, blackman, flatwin, boxwin\n ave : bool, optional(True)- currently locked\n whether or not to average PSDs and ASDs or calculate raw FRFs\n Hv : bool, optional(False)\n calculate the :math:`H_v` frequency response function\n\n Returns\n -------\n freq : float array\n frequency vector (1xN)\n H1 : float array\n Frequency Response Function :math:`H_1` estimate, (nxN) or (nxNxm)\n H2 : float array\n Frequency Response Function :math:`H_2` estimate, (nxN) or (nxNxm)\n coh : float array\n Coherance Function :math:`\\gamma^2` estimate, (nxN)\n Hv : float array\n Frequency Response Function :math:`H_v` estimate, (nxN) or (nxNxm)\n\n Currently ``ave`` is locked to default values.\n\n Examples\n --------\n >>> import control as ctrl\n >>> import matplotlib.pyplot as plt\n >>> import vibrationtesting as vt\n >>> import numpy as np\n >>> sample_freq = 1e3\n >>> noise_power = 0.001 * sample_freq / 2\n >>> A = np.array([[0, 0, 1, 0],\n ... [0, 0, 0, 1],\n ... [-200, 100, -.2, .1],\n ... [100, -200, .1, -.2]])\n >>> B = np.array([[0], [0], [1], [0]])\n >>> C = np.array([[35, 0, 0, 0], [0, 35, 0, 0]])\n >>> D = np.array([[0], [0]])\n >>> sys = ctrl.ss(A, B, C, D)\n >>> tin = np.arange(0, 51.2, .1)\n >>> nr = .5 # 0 is all noise on input\n >>> for i in np.arange(520):\n ... u = np.random.normal(scale=np.sqrt(noise_power), size=tin.shape)\n ... #print(u)\n ... t, yout, xout = ctrl.forced_response(sys, tin, u,rtol=1e-12)\n ... if 'Yout' in locals():\n ... Yout=np.dstack((Yout,yout\n ... +nr*np.random.normal(scale=.050*np.std(yout[0,:]),\n ... size=yout.shape)))\n ... Ucomb=np.dstack((Ucomb,u+(1-nr)\n ... *np.random.normal(scale=.05*np.std(u),\n ... size=u.shape)))\n ... else:\n ... Yout=yout+nr*np.random.normal(scale=.05*np.std(yout[0,:]),\n ... size=yout.shape)\n ... # noise on output is 5% scale of input\n ... Ucomb=u+(1-nr)*np.random.normal(scale=.05*np.std(u),\n ... size=u.shape)#(1, len(tin)))\n ... # 5% noise signal on input\n >>> f, Hxy1, Hxy2, coh, Hxyv = vt.frfest(Yout, Ucomb, t, Hv=bool(True))\n >>> vt.frfplot(f,Hxy2,freq_max=3.5, legend=['$H_{11}$', '$H_{12}$'])\n ... # doctest: +SKIP\n >>> vt.frfplot(f, np.vstack((Hxy1[0,:], Hxy2[0,:], Hxyv[0,:])),\n ... legend=['$H_{11-1}$','$H_{11-2}$','$H_{11-v}$'])\n ... # doctest: +SKIP\n\n Notes\n -----\n .. note:: Not compatible with scipy.signal functions\n .. seealso:: :func:`asd`, :func:`crsd`, :func:`frfplot`.\n .. warning:: hanning window cannot be selected yet. Averaging cannot be\n unslected yet.\n .. todo:: Fix averaging, windowing, multiple input.\n\n \"\"\"\n if len(f.shape) == 1:\n f = f.reshape(1, -1, 1)\n\n if len(x.shape) == 1:\n x = x.reshape(1, -1, 1)\n\n if len(f.shape) == 2:\n if (f.shape).index(max(f.shape)) == 0:\n f = f.reshape(max(f.shape), min(f.shape), 1)\n else:\n f = f.reshape(1, max(f.shape), min(f.shape))\n\n if len(x.shape) == 2:\n if (x.shape).index(max(x.shape)) == 0:\n x = x.reshape(max(x.shape), min(x.shape), 1)\n else:\n x = x.reshape(1, max(x.shape), min(x.shape))\n\n # Note: Two different ways to ignore returned values shown\n Pff = asd(f, dt, windowname=windowname)[1]\n freq, Pxf = crsd(x, f, dt, windowname=windowname)\n _, Pxx = asd(x, dt)\n\n # Note Pfx=conj(Pxf) is applied in the H1 FRF estimation\n Txf1 = np.conj(Pxf / Pff)\n Txf2 = Pxx / Pxf\n # Nulled to avoid output problems/simplify calls if unrequested\n Txfv = np.zeros_like(Txf1)\n\n coh = (Pxf * np.conj(Pxf)).real / Pxx / Pff\n\n if Hv:\n for i in np.arange(Pxx.shape[1]):\n frfm = np.array(\n [[Pff[0, i], np.conj(Pxf[0, i])], [Pxf[0, i], Pxx[0, i]]])\n\n alpha = 1 # np.sqrt(Pff[0,i]/Pxx[0,i])\n\n frfm = np.array([[Pff[0, i], alpha * np.conj(Pxf[0, i])],\n [alpha * Pxf[0, i], alpha**2 * Pxx[0, i]]])\n\n lam, vecs = la.eigh(frfm)\n\n index = lam.argsort()\n\n lam = lam[index]\n\n vecs = vecs[:, index]\n\n Txfv[0, i] = -(vecs[0, 0] / vecs[1, 0]) / alpha\n\n return freq, Txf1, Txf2, coh, Txfv\n\n\ndef frfplot(freq, H, freq_min=0, freq_max=None, type=1, legend=[]):\n \"\"\"Frequency Response Function pretty plotting.\n\n Plots frequency response functions in a variety of formats\n\n Parameters\n ----------\n freq : float array\n Frequency vector (rad/sec), (1xN)\n H : float array\n Frequency response functions (nxN)\n freq_min : float, optional\n Low frequency for plot (default 0)\n freq_min : float, optional\n High frequency for plot (default max frequency)\n legend : string array\n Array of string for use in legend.\n type : int, optional\n Plot type. See notes.\n\n Returns\n -------\n ax : axis objects\n allows manipulation of plot parameters (xlabel, title...)\n\n Examples\n --------\n >>> import matplotlib.pyplot as plt\n >>> import vibrationtesting as vt\n >>> import numpy as np\n >>> f=np.linspace(0,100,10000).reshape(-1,1);\n >>> w=f*2*np.pi;\n >>> k=1e5;m=1;c=1;\n >>> frf1=1./(m*(w*1j)**2+c*1j*w+k)\n >>> frf2=1./(m*(w*1j)**2+c*1j*w+k*3)\n >>> _ = vt.frfplot(f,np.hstack((frf1,frf2)), legend = ['FRF 1','FRF 2'])\n ... # doctest: +SKIP\n\n Notes\n -----\n +---------+------------------------------------------------+\n | type | Plot style |\n +=========+================================================+\n | 1 (def) | Magnitude and Phase versus F |\n +---------+------------------------------------------------+\n | 2 | Magnitude and Phase versus log10(F) |\n +---------+------------------------------------------------+\n | 3 | Bodelog (Magnitude and Phase versus log10(w)) |\n +---------+------------------------------------------------+\n | 4 | Real and Imaginary |\n +---------+------------------------------------------------+\n | 5 | Nyquist (Imaginary versus Real) |\n +---------+------------------------------------------------+\n | 6 | Magnitude versus F |\n +---------+------------------------------------------------+\n | 7 | Phase versus F |\n +---------+------------------------------------------------+\n | 8 | Real versus F |\n +---------+------------------------------------------------+\n | 9 | Imaginary versus F |\n +---------+------------------------------------------------+\n | 10 | Magnitude versus log10(F) |\n +---------+------------------------------------------------+\n | 11 | Phase versus log10(F) |\n +---------+------------------------------------------------+\n | 12 | Real versus log10(F) |\n +---------+------------------------------------------------+\n | 13 | Imaginary versus log10(F) |\n +---------+------------------------------------------------+\n | 14 | Magnitude versus log10(w) |\n +---------+------------------------------------------------+\n | 15 | Phase versus log10(w) |\n +---------+------------------------------------------------+\n\n .. seealso:: `frfest`\n\n Copyright J. Slater, Dec 17, 1994\n Updated April 27, 1995\n Ported to Python, July 1, 2015\n\n \"\"\"\n FLAG = type # Plot type, should libe renamed throughout.\n freq = freq.reshape(1, -1)\n lenF = freq.shape[1]\n if len(H.shape) is 1:\n H = H.reshape(1, -1)\n\n if H.shape[0] > H.shape[1]:\n H = H.T\n\n if freq_max is None:\n freq_max = np.max(freq)\n\n if freq_min is None:\n freq_min = np.min(freq)\n\n if freq_min < np.min(freq):\n freq_min = np.min(freq)\n\n if freq_min > freq_max:\n raise ValueError('freq_min must be less than freq_max.')\n\n # print(str(np.amin(freq)))\n inlow = int(lenF * (freq_min - np.amin(freq)\n ) // (np.amax(freq) - np.amin(freq)))\n\n inhigh = int(lenF * (freq_max - np.amin(freq)\n ) // (np.amax(freq) - np.amin(freq)) - 1)\n # if inlow<1,inlow=1;end\n # if inhigh>lenF,inhigh=lenF;end\n \"\"\"print('freq shape: {}'.format(freq.shape))\n print('H shape: {}'.format(H.shape))\n print('Index of low frequency: {}'.format(inlow))\n print('Index of high frequency: {}'.format(inhigh))\"\"\"\n H = H[:, inlow:inhigh]\n # print(H.shape)\n freq = freq[:, inlow:inhigh]\n mag = 20 * np.log10(np.abs(H))\n # print(mag)\n # print(mag.shape)\n minmag = np.min(mag)\n maxmag = np.max(mag)\n phase = np.unwrap(np.angle(H)) * 180 / np.pi\n # phmin_max=[min(phase)//45)*45 ceil(max(max(phase))/45)*45];\n phmin = np.amin(phase) // 45 * 45.0\n phmax = (np.amax(phase) // 45 + 1) * 45\n \"\"\"minreal = np.amin(np.real(H))\n maxreal = np.amax(np.real(H))\n minimag = np.amin(np.imag(H))\n maximag = np.amax(np.imag(H))\"\"\"\n\n if FLAG is 1:\n fig, (ax1, ax2) = plt.subplots(2, 1)\n ax1.plot(freq.T, mag.T)\n ax1.set_xlabel('Frequency (Hz)')\n ax1.set_ylabel('Mag (dB)')\n ax1.grid()\n ax1.set_xlim(xmax=freq_max, xmin=freq_min)\n ax1.set_ylim(ymax=maxmag, ymin=minmag)\n\n ax2.plot(freq.T, phase.T)\n ax2.set_xlabel('Frequency (Hz)')\n ax2.set_ylabel('Phase (deg)')\n ax2.grid()\n ax2.set_xlim(xmax=freq_max, xmin=freq_min)\n ax2.set_ylim(ymax=phmax, ymin=phmin)\n ax2.set_yticks(np.arange(phmin, (phmax + 45), 45))\n fig.tight_layout()\n\n if len(legend) > 0:\n plt.legend(legend)\n ax = (ax1, ax2)\n else:\n print(\"Sorry, that option isn't supported yet\")\n return ax\n\n \"\"\"# elif FLAG==2:\n # subplot(2,1,1)\n # semilogx(F,mag)\n # xlabel('Frequency (Hz)')\n # ylabel('Mag (dB)')\n # grid on\n # % Fmin,Fmax,min(mag),max(mag)\n # axis([Fmin Fmax minmag maxmag])\n\n # subplot(2,1,2)\n # semilogx(F,phase)\n # xlabel('Frequency (Hz)')\n # ylabel('Phase (deg)')\n # grid on\n # axis([Fmin Fmax phmin_max(1) phmin_max(2)])\n # gridmin_max=round(phmin_max/90)*90;\n # set(gca,'YTick',gridmin_max(1):90:gridmin_max(2))\n\n # elif FLAG==3:\n # subplot(2,1,1)\n # mag=20*log10(abs(Xfer));\n # semilogx(F*2*pi,mag)\n # xlabel('Frequency (Rad/s)')\n # ylabel('Mag (dB)')\n # grid on\n # axis([Wmin Wmax minmag maxmag])\n # zoom on\n # subplot(2,1,2)\n # semilogx(F*2*pi,phase)\n # xlabel('Frequency (Rad/s)')\n # ylabel('Phase (deg)')\n # grid on\n # axis([Wmin Wmax phmin_max(1) phmin_max(2)])\n # gridmin_max=round(phmin_max/90)*90;\n # set(gca,'YTick',gridmin_max(1):90:gridmin_max(2))\n\n # elseif FLAG==4\n # subplot(2,1,1)\n # plot(F,real(Xfer))\n # xlabel('Frequency (Hz)')\n # ylabel('Real')\n # grid on\n # axis([Fmin Fmax minreal maxreal])\n # zoom on\n # subplot(2,1,2)\n # plot(F,imag(Xfer))\n # xlabel('Frequency (Hz)')\n # ylabel('Imaginary')\n # grid on\n # axis([Fmin Fmax minimag maximag])\n # zoom on\n # elseif FLAG==5\n # subplot(1,1,1)\n # imax=round(length(F)*Fmax/max(F));\n # imin=round(length(F)*Fmin/max(F))+1;\n # plot(real(Xfer(imin:imax)),imag(Xfer(imin:imax)))\n # xlabel('Real')\n # ylabel('Imaginary')\n # grid on\n # zoom on\n # elseif FLAG==6\n # subplot(1,1,1)\n # mag=20*log10(abs(Xfer));\n # plot(F,mag)\n # xlabel('Frequency (Hz)')\n # ylabel('Mag (dB)')\n # grid on\n # axis([Fmin Fmax minmag maxmag])\n # zoom on\n # elseif FLAG==7\n # subplot(1,1,1)\n # plot(F,phase)\n # xlabel('Frequency (Hz)')\n # ylabel('Phase (deg)')\n # grid on\n # phmin_max=[floor(min(phase)/45)*45 ceil(max(phase)/45)*45];\n # axis([Fmin Fmax phmin_max(1) phmin_max(2)])\n # gridmin_max=round(phmin_max/90)*90;\n # set(gca,'YTick',gridmin_max(1):90:gridmin_max(2))\n # zoom on\n # elseif FLAG==8\n # subplot(1,1,1)\n # plot(F,real(Xfer))\n # xlabel('Frequency (Hz)')\n # ylabel('Real')\n # grid on\n # axis([Fmin Fmax minreal maxreal])\n # zoom on\n # elseif FLAG==9\n # subplot(1,1,1)\n # plot(F,imag(Xfer))\n # xlabel('Frequency (Hz)')\n # ylabel('Imaginary')\n # grid on\n # axis([Fmin Fmax minimag maximag])\n # zoom on\n # elseif FLAG==10\n # subplot(1,1,1)\n # mag=20*log10(abs(Xfer));\n # semilogx(F,mag)\n # xlabel('Frequency (Hz)')\n # ylabel('Mag (dB)')\n # grid on\n # axis([Fmin Fmax minmag maxmag])\n # zoom on\n # elseif FLAG==11\n # subplot(1,1,1)\n # semilogx(F,phase)\n # xlabel('Frequency (Hz)')\n # ylabel('Phase (deg)')\n # grid on\n # phmin_max=[floor(min(phase)/45)*45 ceil(max(phase)/45)*45];\n # axis([Fmin Fmax phmin_max(1) phmin_max(2)])\n # gridmin_max=round(phmin_max/90)*90;\n # set(gca,'YTick',gridmin_max(1):90:gridmin_max(2))\n # zoom on\n # elseif FLAG==12\n # subplot(1,1,1)\n # semilogx(F,real(Xfer))\n # xlabel('Frequency (Hz)')\n # ylabel('Real')\n # grid on\n # axis([Fmin Fmax minreal maxreal])\n # zoom on\n # elseif FLAG==13\n # subplot(1,1,1)\n # semilogx(F,imag(Xfer))\n # xlabel('Frequency (Hz)')\n # ylabel('Imaginary')\n # grid on\n # axis([Fmin Fmax minimag maximag])\n # zoom on\n # elseif FLAG==14\n # subplot(1,1,1)\n # mag=20*log10(abs(Xfer));\n # semilogx(F*2*pi,mag)\n # xlabel('Frequency (Rad/s)')\n # ylabel('Mag (dB)')\n # grid on\n # axis([Wmin Wmax minmag maxmag])\n # zoom on\n # elseif FLAG==15\n # subplot(1,1,1)\n # semilogx(F*2*pi,phase)\n # xlabel('Frequency (Rad/s)')\n # ylabel('Phase (deg)')\n # grid on\n # axis([Wmin Wmax phmin_max(1) phmin_max(2)])\n # gridmin_max=round(phmin_max/90)*90;\n # set(gca,'YTick',gridmin_max(1):90:gridmin_max(2))\n # zoom on\n # else\n # subplot(2,1,1)\n # mag=20*log10(abs(Xfer));\n # plot(F,mag)\n # xlabel('Frequency (Hz)')\n # ylabel('Mag (dB)')\n # grid on\n # axis([Fmin Fmax minmag maxmag])\n # zoom on\n # subplot(2,1,2)\n # plot(F,phase)\n # xlabel('Frequency (Hz)')\n # ylabel('Phase (deg)')\n # grid on\n # phmin_max=[floor(min(phase)/45)*45 ceil(max(phase)/45)*45];\n # axis([Fmin Fmax phmin_max(1) phmin_max(2)])\n # gridmin_max=round(phmin_max/90)*90;\n # set(gca,'YTick',gridmin_max(1):90:gridmin_max(2))\n # zoom on\n \"\"\"\n\n\ndef xcorr(t, x, y, zeropad=True):\n \"\"\"Sorry, no docs or tests yet.\"\"\"\n tau = t\n # sx = len(x)\n # sy = len(y)\n if zeropad is True:\n Xn = np.fft.rfft(x, n=len(x) * 2)\n Yn = np.conj(sp.fft(y, n=len(x) * 2))\n else:\n Xn = np.fft.rfft(x)\n Yn = np.conj(np.fft.rfft(y))\n\n xcor = np.real(fftpack.fftshift(sp.ifft(Xn * Yn)))\n dt = t[1] - t[0]\n\n tau = np.linspace(-len(xcor) / 2 * dt - dt / 2,\n len(xcor) / 2 * dt - dt / 2, len(xcor))\n return tau, xcor\n\n\ndef hammer_impulse(time, imp_time=None, imp_duration=None, doublehit=False,\n dh_delta=None):\n \"\"\"Generate simulated hammer hit (half sine).\n\n Parameters\n ----------\n time : float array\n 1 x N time array. Suggest using `np.linspace(0,10,1000).reshape(1,-1)`\n for example\n imp_time : float (optional)\n Time of onset of impulse. Default is 0.1 time end time- which\n traditionally works well for impact testing\n imp_duration : float (optional)\n Duration of impulse. Default is 0.01 of total record\n doublehit : Boolean (optional)\n Allows repeat of hit to emulate a bad strike. Default is False\n dh_delta : float (optional)\n Time difference between primary strike and accidental second strike\n Default is 0.02 of record.\n\n Returns\n -------\n force : float array\n\n Examples\n --------\n >>> import vibrationtesting as vt\n >>> time = np.linspace(0,10,1024).reshape(1,-1)\n >>> force = vt.hammer_impulse(time, doublehit=True)\n >>> plt.plot(time.T, force.T)\n [>> time = np.linspace(0,4,4096)\n >>> u = np.random.randn(1,len(time))\n >>> ttime, signal_out = decimate(time, u, 100)\n\n \"\"\"\n dt = t[1] - t[0]\n current_frequency = 1 / dt\n freq_frac = sample_frequency / current_frequency\n Wn = .9 * freq_frac\n b, a = signal.butter(8, Wn, 'low')\n if len(in_signal.shape) > 1:\n filtered_signal = signal.lfilter(b, a, in_signal, axis=1)\n else:\n filtered_signal = signal.lfilter(b, a, in_signal)\n step = int(1 / freq_frac)\n time = t[::step]\n if len(in_signal.shape) == 1:\n filtered_signal = filtered_signal[::step]\n elif len(in_signal.shape) == 2:\n filtered_signal = filtered_signal[:, ::step]\n elif len(in_signal.shape) == 3:\n filtered_signal = filtered_signal[:, ::step, :]\n return time, filtered_signal\n","sub_path":"vibrationtesting/signals.py","file_name":"signals.py","file_ext":"py","file_size_in_byte":41686,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"27999987","text":"\"\"\"\nImplemented in pytorch.\n\"\"\"\n\nfrom __future__ import print_function, division\nfrom builtins import range\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom util import get_normalized_data\n\nimport torch\nfrom torch.autograd import Variable\nfrom torch import optim\n\nXtrain, Xtest, Ytrain, Ytest = get_normalized_data()\n\n_, D = Xtrain.shape\nK = len(set(Ytrain))\n\nmodel = torch.nn.Sequential()\n\n# NOTE: the \"p\" is p_drop, not p_keep\nmodel.add_module(\"dense1\", torch.nn.Linear(D, 500))\nmodel.add_module(\"bn1\", torch.nn.BatchNorm1d(500))\nmodel.add_module(\"relu1\", torch.nn.ReLU())\nmodel.add_module(\"dense2\", torch.nn.Linear(500, 300))\nmodel.add_module(\"bn2\", torch.nn.BatchNorm1d(300))\nmodel.add_module(\"relu2\", torch.nn.ReLU())\nmodel.add_module(\"dense3\", torch.nn.Linear(300, K))\n\nloss = torch.nn.CrossEntropyLoss(size_average = True)\n\noptimizer = optim.Adam(model.parameters(), lr = 1e-4)\n\ndef train(model, loss, optimizer, inputs, labels):\n\n\tmodel.train()\n\n\tinputs = Variable(inputs, requires_grad = False)\n\tlabels = Variable(labels, requires_grad = False)\n\n\t#reset gradient\n\toptimizer.zero_grad()\n\n\t#forward\n\tlogits = model.forward(inputs)\n\toutput = loss.forward(logits, labels)\n\n\t#backward\n\toutput.backward()\n\n\t#update parameter\n\toptimizer.step()\n\n\treturn output.item()\n\ndef get_cost(model, loss, inputs, labels):\n\tmodel.eval()\n\n\tinputs = Variable(inputs, requires_grad = False)\n\tlabels = Variable(labels, requires_grad = False)\n\n\t#forward\n\tlogits = model.forward(inputs)\n\toutput = loss.forward(logits, labels)\n\n\treturn output.item()\n\n\ndef predict(model, inputs):\n\tmodel.eval()\n\tinputs = Variable(inputs, requires_grad = False)\n\tlogits = model.forward(inputs)\n\treturn logits.data.numpy().argmax(axis = 1)\n\ndef score(model, inputs, labels):\n\tpredictions = predict(model, inputs)\n\treturn np.mean(labels.numpy() == predictions)\n\n\nXtrain = torch.from_numpy(Xtrain).float()\nYtrain = torch.from_numpy(Ytrain).long()\nXtest = torch.from_numpy(Xtest).float()\nYtest = torch.from_numpy(Ytest).long()\n\n\nepochs = 10\nbatch_size = 32\nn_batches = Xtrain.size()[0] // batch_size\n\n#things to kepp track\ntrain_costs = []\ntest_costs = []\ntrain_accuracies = []\ntest_accuracies = []\n\n\n#main training loop\nfor i in range(epochs):\n\tcost = 0\n\ttest_cost = 0\n\tfor j in range(n_batches):\n\t\tXbatch = Xtrain[j*batch_size : (j*batch_size + batch_size)]\n\t\tYbatch = Ytrain[j*batch_size : (j*batch_size + batch_size)]\n\t\tcost += train(model, loss, optimizer, Xbatch, Ybatch)\n\n\ttrain_acc = score(model, Xtrain, Ytrain)\n\ttest_acc = score(model, Xtest, Ytest)\n\ttest_cost = get_cost(model, loss, Xtest, Ytest)\n\tprint(\"Epoch: %d, cost: %f, acc: %.2f\" % (i, test_cost, test_acc))\n\n\ttrain_accuracies.append(train_acc)\n\ttrain_costs.append(cost / n_batches)\n\ttest_accuracies.append(test_acc)\n\ttest_costs.append(test_cost)\n\n# plot the results\nplt.plot(train_costs, label='Train cost')\nplt.plot(test_costs, label='Test cost')\nplt.title('Cost')\nplt.legend()\nplt.show()\n\nplt.plot(train_accuracies, label='Train accuracy')\nplt.plot(test_accuracies, label='Test accuracy')\nplt.title('Accuracy')\nplt.legend()\nplt.show()\n\n\n","sub_path":"BasicANN/Reguralization/pytorch_batchnorm.py","file_name":"pytorch_batchnorm.py","file_ext":"py","file_size_in_byte":3080,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"641280031","text":"\nimport time\n\nSITE_URL = \"http://olivier.drevet.free.fr/\"\nBLOG_EMAIL = \"odrevet@gmail.com\"\nBLOG_AUTHOR = \"Olivier Drevet\"\nBLOG_TITLE = \"Site Perso Olivier Drevet\"\n\nDEFAULT_LANG = \"fr\"\n\nTRANSLATIONS = {\n DEFAULT_LANG: \"\",\n}\n\n\nTRANSLATIONS_PATTERN = '{path}.{lang}.{ext}'\n\nNAVIGATION_LINKS = {\n DEFAULT_LANG: (\n (\"/realisations/\", \"Réalisations\"),\n (\"/liens/\", \"Liens\"),\n (\"/documents/CV-FR_2019_Olivier_DREVET.pdf\", \"CV\"),\n (\"/wiki\", \"Wiki\"),\n ),\n}\n\nNAVIGATION_ALT_LINKS = {\n DEFAULT_LANG: ()\n}\n\nTHEME = \"material-theme\"\n\n\nTHEME_CONFIG = {\n DEFAULT_LANG: {\n 'featured_large': False,\n 'featured_small': False,\n 'featured_on_mobile': True,\n 'featured_large_image_on_mobile': True,\n 'featured_strip_html': False,\n 'sidebar': ''\n }\n}\n\nPOSTS = ()\nPAGES = (\n (\"pages/*.rst\", \"\" , \"page.tmpl\"),\n (\"pages/*.md\", \"\" , \"page.tmpl\")\n)\n\n\nINDEX_PATH = \"blog\"\n\nTIMEZONE = \"Europe/Paris\"\n\nCOMPILERS = {\n \"rest\": ('.rst', '.txt'),\n \"markdown\": ('.md', '.mdown', '.markdown'),\n}\n\n\nHIDDEN_TAGS = ['mathjax']\n\nCATEGORY_ALLOW_HIERARCHIES = False\nCATEGORY_OUTPUT_FLAT_HIERARCHY = False\n\nHIDDEN_CATEGORIES = []\n\n\nHIDDEN_AUTHORS = ['Guest']\n\n\nFRONT_INDEX_HEADER = {\n DEFAULT_LANG: ''\n}\n\nREDIRECTIONS = []\n\nIMAGE_FOLDERS = {'images': 'images'}\n\nINDEX_READ_MORE_LINK = '

{read_more}…

'\nFEED_READ_MORE_LINK = '

{read_more}… ({min_remaining_read})

'\n\nFEED_LINKS_APPEND_QUERY = False\n\nLICENSE = \"\"\n\nCONTENT_FOOTER = 'Contents © {date} {author} - Powered by Nikola {license}'\n\nCONTENT_FOOTER_FORMATS = {\n DEFAULT_LANG: (\n (),\n {\n \"email\": BLOG_EMAIL,\n \"author\": BLOG_AUTHOR,\n \"date\": time.gmtime().tm_year,\n \"license\": LICENSE\n }\n )\n}\n\nRSS_COPYRIGHT = 'Contents © {date} {author} {license}'\nRSS_COPYRIGHT_PLAIN = 'Contents © {date} {author} {license}'\nRSS_COPYRIGHT_FORMATS = CONTENT_FOOTER_FORMATS\n\nCOMMENT_SYSTEM = \"\"\nCOMMENT_SYSTEM_ID = \"\"\n\n\nSTRIP_INDEXES = True\n\nPRETTY_URLS = True\n\n\nMARKDOWN_EXTENSIONS = ['markdown.extensions.fenced_code', 'markdown.extensions.codehilite', 'markdown.extensions.extra']\n\nUSE_TAG_METADATA = False\n\nWARN_ABOUT_TAG_METADATA = False\n\n\nGLOBAL_CONTEXT = {}\n\nGLOBAL_CONTEXT_FILLER = []\n","sub_path":"conf.py","file_name":"conf.py","file_ext":"py","file_size_in_byte":2473,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"266239611","text":"import CMGTools.TTHAnalysis.plotter.CMS_lumi as CMS_lumi\nfrom optparse import OptionParser\nimport ROOT, sys, array\nROOT.gStyle.SetOptStat(0)\nROOT.gStyle.SetOptTitle(0) \n\nparser = OptionParser(usage=\"%prog rootfile histogram\")\nparser.add_option(\"-e\" , \"--energy\", dest=\"energy\" , type=\"float\", default=13, help=\"Energy to plot\");\nparser.add_option(\"-l\" , \"--lumi\", dest=\"lumi\" , type=\"float\", default=35.9, help=\"Luminosity to plot\");\nparser.add_option(\"-x\" , dest=\"xlabel\", type=\"string\", default=None , help=\"Label of the x axis\");\nparser.add_option(\"-y\" , dest=\"ylabel\", type=\"string\", default=None , help=\"Label of the y axis\");\nparser.add_option(\"-z\" , dest=\"zlabel\", type=\"string\", default=\"fake ratio\", help=\"Label of the z axis\");\nparser.add_option(\"--exts\", dest=\"exts\" , type=\"string\", action=\"append\", default=[\"png\",\"pdf\",\"C\",\"root\"], help=\"Output types\");\nparser.add_option(\"--lspam\", dest=\"lspam\" , type=\"string\", default=\"Preliminary\", help=\"Stuff to write onto the plot (Preliminary, Simulation, Internal)\");\nparser.add_option(\"--plotmode\", dest=\"plotmode\", type=\"string\", default=\"colz text e\", help=\"Option to use when drawing the plot\");\nparser.add_option(\"--zrange\" , dest=\"zrange\" , type=\"float\" , nargs=2, default=[0,1], help=\"Range of the z axis\");\nparser.add_option(\"--digits\" , dest=\"digits\" , type=\"float\" , default=1.3, help=\"Format of digits for text on plots\");\nparser.add_option(\"--logx\" , dest=\"logx\" , action=\"store_true\", default=False, help=\"Logx\");\nparser.add_option(\"--logy\" , dest=\"logy\" , action=\"store_true\", default=False, help=\"Logy\");\nparser.add_option(\"--logz\" , dest=\"logz\" , action=\"store_true\", default=False, help=\"Logz\");\n(options, args) = parser.parse_args()\n\n\nROOT.gStyle.SetPaintTextFormat(\"%sf\"%options.digits) \ncanvas = ROOT.TCanvas(\"c\", \"c\", 600, 600)\n\nf = ROOT.TFile.Open(args[0],\"read\")\nif not f: sys.exit()\n\nhisto = f.Get(args[1])\nif not histo: sys.exit()\n\ncanvas.cd()\n\nhisto.Draw()\nROOT.gPad.Update()\n\ncanvas.SetTickx(1)\ncanvas.SetTicky(1)\ncanvas.SetRightMargin (0.19)\ncanvas.SetTopMargin (0.06)\ncanvas.SetLeftMargin (0.14)\ncanvas.SetBottomMargin(0.14)\n\n\nif options.xlabel: histo.GetXaxis().SetTitle(options.xlabel)\nif options.ylabel: histo.GetYaxis().SetTitle(options.ylabel)\nif options.zlabel: histo.GetZaxis().SetTitle(options.zlabel)\n\nhisto.GetXaxis().SetNdivisions(505)\nhisto.GetXaxis().SetLabelFont(42)\nhisto.GetXaxis().SetLabelSize(0.035)\nhisto.GetXaxis().SetTitleFont(42)\nhisto.GetXaxis().SetTitleSize(0.05)\nhisto.GetXaxis().SetLabelOffset(0.007)\nhisto.GetXaxis().SetTitleOffset(1.2)\n\nhisto.GetYaxis().SetNdivisions(505)\nhisto.GetYaxis().SetLabelFont(42)\nhisto.GetYaxis().SetLabelSize(0.035)\nhisto.GetYaxis().SetTitleFont(42)\nhisto.GetYaxis().SetTitleSize(0.05)\nhisto.GetYaxis().SetLabelOffset(0.007)\nhisto.GetYaxis().SetTitleOffset(1.30)\n\nhisto.GetZaxis().SetLabelFont(42)\nhisto.GetZaxis().SetTitleFont(42)\nhisto.GetZaxis().SetLabelSize(0.035)\nhisto.GetZaxis().SetTitleSize(0.035)\nhisto.GetZaxis().SetTitleOffset(1.7)\nhisto.SetMinimum(options.zrange[0])\nhisto.SetMaximum(options.zrange[1])\nNRGBs = 5\nNCont = 255\nstops = array.array(\"d\",[0.00, 0.34, 0.61, 0.84, 1.00])\nred = array.array(\"d\",[0.50, 0.50, 1.00, 1.00, 1.00])\ngreen = array.array(\"d\",[0.50, 1.00, 1.00, 0.60, 0.50])\nblue = array.array(\"d\",[1.00, 1.00, 0.50, 0.40, 0.50])\nROOT.TColor.CreateGradientColorTable(NRGBs, stops, red, green, blue, NCont)\nROOT.gStyle.SetNumberContours(NCont)\ncanvas.cd()\nhisto.Draw(\"colz\")\nROOT.gPad.Update()\npalette = histo.GetListOfFunctions().FindObject(\"palette\")\npalette.SetX1NDC(1.-0.18)\npalette.SetY1NDC(0.14)\npalette.SetX2NDC(1.-0.13)\npalette.SetY2NDC(1.-0.06)\npalette.SetLabelFont(42)\npalette.SetLabelSize(0.035)\n\nhisto.Draw(options.plotmode)\n\nCMS_lumi.writeExtraText = 1\nCMS_lumi.lumi_13TeV = \"%.1f fb^{-1}\" % options.lumi\nCMS_lumi.extraText = options.lspam\nCMS_lumi.lumi_sqrtS = options.energy\nCMS_lumi.CMS_lumi(canvas, 4, 0, 0.05)\n\nif options.logx: canvas.SetLogx()\nif options.logy: canvas.SetLogy()\nif options.logz: canvas.SetLogz()\n\nfor ext in options.exts:\n\tcanvas.SaveAs(args[1]+\".\"+ext)\n\n\n","sub_path":"TTHAnalysis/python/plotter/plot2d.py","file_name":"plot2d.py","file_ext":"py","file_size_in_byte":4127,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"464737015","text":"\nimport matplotlib.pyplot as plt\n\n# define function for making figures\ndef define_figure(xlabel=\"X\",ylabel=\"Y\"):\n # setup plot parameters\n fig = plt.figure(figsize=(10,8), dpi= 80, facecolor='w', edgecolor='k')\n ax = plt.subplot(111)\n ax.grid(b=True, which='major', axis='both', color='#808080', linestyle='--')\n ax.set_xlabel(xlabel,size=20)\n ax.set_ylabel(ylabel,size=20)\n plt.tick_params(axis='both',labelsize=20)\n return ax\n","sub_path":"final_exam/plotting.py","file_name":"plotting.py","file_ext":"py","file_size_in_byte":452,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"334803245","text":"#!/usr/bin/env python2\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Oct 1 12:05:47 2018\n\n@author: lcoyle\n\"\"\"\nimport pytimber\nfrom LHC_utility import TimestampToDatetime\nimport numpy as np\nimport pandas as pd\n\n\ndef calibration(energy, beam):\n inj_B1 = 2.35E11\n inj_B2 = 2.35E11\n flt_B1 = 2.60E10\n flt_B2 = 3.00E10\n\n inj_e = 450.\n flt_e = 6500.\n\n slopeB1 = (flt_B1 - inj_B1)/( flt_e - inj_e)\n offsetB1 = inj_B1 - slopeB1*inj_e\n\n slopeB2 = (flt_B2 - inj_B2)/( flt_e - inj_e)\n offsetB2 = inj_B2 - slopeB2*inj_e\n\n calibration = -1.\n if beam == 'B1': \n calibration = offsetB1 + slopeB1*energy\n if beam == 'B2':\n calibration = offsetB2 + slopeB2*energy\n\n #print(\"%g\" %(calibration))\n return calibration\n\n\ndef getLifetime_original( loss, intensity ):\n value = 0.\n if intensity > 0: value = 1. - loss/intensity\n lifetime = 1000.\n if value > 0. and value < 1. : lifetime = -1./np.log( value )/3600.\n if lifetime > 1000. : lifetime = 1000.\n #print(\"lifetimes1 %g\" %(lifetime))\n return lifetime\n\ndef getLifetime(loss, intensity):\n assert len(loss) == len(intensity)\n \n value = np.zeros(len(loss))\n lifetimes = np.ones(len(loss))\n lifetimes = lifetimes*1000\n \n # indices of where the intensity is positive\n int_ind = intensity[intensity>0].index\n \n value[int_ind] = 1 - loss[int_ind]/intensity[int_ind]\n\n # indices of where the values is between 0 and 1\n value_ind = np.where(np.logical_and(value>0,value<1))\n\n lifetimes[value_ind] = -1/np.log(value[value_ind])/3600.\n \n lifetimes[lifetimes>1000] = 1000\n\n return lifetimes\n\n\ndef GetBeamLifetimeIndex(dataDict, i):\n blmListB2 = [\"BLMEI.06R7.B2I10_TCHSS.6R7.B2:LOSS_RS09\"]\n blmListB2 += [\"BLMEI.06R7.B2I10_TCHSH.6R7.B2:LOSS_RS09\"]\n blmListB2 += [\"BLMEI.06R7.B2I10_TCHSV.6R7.B2:LOSS_RS09\"]\n blmListB2 += [\"BLMEI.06R7.B2I10_TCP.A6R7.B2:LOSS_RS09\"]\n\n blmListB1 = [\"BLMEI.06L7.B1E10_TCHSS.6L7.B1:LOSS_RS09\"]\n blmListB1 += [\"BLMEI.06L7.B1E10_TCHSH.6L7.B1:LOSS_RS09\"]\n blmListB1 += [\"BLMEI.06L7.B1E10_TCHSV.6L7.B1:LOSS_RS09\"]\n blmListB1 += [\"BLMEI.06L7.B1E10_TCP.A6L7.B1:LOSS_RS09\"]\n\n BCTB1 = \"LHC.BCTFR.A6R4.B1:BEAM_INTENSITY\"\n BCTB2 = \"LHC.BCTFR.A6R4.B2:BEAM_INTENSITY\"\n\n ENERGY = \"HX:ENG\"\n\n energy = dataDict[ENERGY].iloc[i]*6500./54175.\n intensityB1 = dataDict[BCTB1].iloc[i]\n intensityB2 = dataDict[BCTB2].iloc[i]\n\n raw_lossB1 = 0.\n for ikey in blmListB1:\n raw_lossB1 += dataDict[ikey].iloc[i]\n raw_lossB2 = 0.\n for ikey in blmListB2:\n raw_lossB2 += dataDict[ikey].iloc[i]\n\n lossB1 = raw_lossB1*calibration(energy, \"B1\")\n lossB2 = raw_lossB2*calibration(energy, \"B2\")\n\n lifetimeB1 = getLifetime_original(lossB1, intensityB1)\n lifetimeB2 = getLifetime_original(lossB2, intensityB2)\n return lifetimeB1, lifetimeB2\n\ndef CalcBeamLifetime(data):\n blmListB2 = [\"BLMEI.06R7.B2I10_TCHSS.6R7.B2:LOSS_RS09\"]\n blmListB2 += [\"BLMEI.06R7.B2I10_TCHSH.6R7.B2:LOSS_RS09\"]\n blmListB2 += [\"BLMEI.06R7.B2I10_TCHSV.6R7.B2:LOSS_RS09\"]\n blmListB2 += [\"BLMEI.06R7.B2I10_TCP.A6R7.B2:LOSS_RS09\"]\n\n blmListB1 = [\"BLMEI.06L7.B1E10_TCHSS.6L7.B1:LOSS_RS09\"]\n blmListB1 += [\"BLMEI.06L7.B1E10_TCHSH.6L7.B1:LOSS_RS09\"]\n blmListB1 += [\"BLMEI.06L7.B1E10_TCHSV.6L7.B1:LOSS_RS09\"]\n blmListB1 += [\"BLMEI.06L7.B1E10_TCP.A6L7.B1:LOSS_RS09\"]\n\n BCTB1 = \"LHC.BCTFR.A6R4.B1:BEAM_INTENSITY\"\n BCTB2 = \"LHC.BCTFR.A6R4.B2:BEAM_INTENSITY\"\n\n ENERGY = \"HX:ENG\"\n\n energy = data[ENERGY]*6500./54175.\n intensityB1 = data[BCTB1]\n intensityB2 = data[BCTB2]\n \n raw_lossB1 = data[blmListB1].sum(axis=1)\n raw_lossB2 = data[blmListB2].sum(axis=1)\n \n# raw_lossB1 = 0.\n# for ikey in blmListB1:\n# raw_lossB1 += dataDict[ikey]\n# raw_lossB2 = 0.\n# for ikey in blmListB2:\n# raw_lossB2 += dataDict[ikey]\n\n lossB1 = raw_lossB1*calibration(energy, \"B1\")\n lossB2 = raw_lossB2*calibration(energy, \"B2\")\n\n lifetimeB1 = getLifetime(lossB1, intensityB1)\n lifetimeB2 = getLifetime(lossB2, intensityB2)\n return lifetimeB1, lifetimeB2\n\n\ndef GetBeamMinLifetime(data, t1, t2, master):\n blmListB2 = [\"BLMEI.06R7.B2I10_TCHSS.6R7.B2:LOSS_RS09\"]\n blmListB2 += [\"BLMEI.06R7.B2I10_TCHSH.6R7.B2:LOSS_RS09\"]\n blmListB2 += [\"BLMEI.06R7.B2I10_TCHSV.6R7.B2:LOSS_RS09\"]\n blmListB2 += [\"BLMEI.06R7.B2I10_TCP.A6R7.B2:LOSS_RS09\"]\n\n blmListB1 = [\"BLMEI.06L7.B1E10_TCHSS.6L7.B1:LOSS_RS09\"]\n blmListB1 += [\"BLMEI.06L7.B1E10_TCHSH.6L7.B1:LOSS_RS09\"]\n blmListB1 += [\"BLMEI.06L7.B1E10_TCHSV.6L7.B1:LOSS_RS09\"]\n blmListB1 += [\"BLMEI.06L7.B1E10_TCP.A6L7.B1:LOSS_RS09\"]\n \n BCTB1 = \"LHC.BCTFR.A6R4.B1:BEAM_INTENSITY\"\n BCTB2 = \"LHC.BCTFR.A6R4.B2:BEAM_INTENSITY\"\n \n ENERGY = \"HX:ENG\"\n \n VARIABLES = [master, BCTB1, BCTB2, ENERGY] + blmListB1 + blmListB2\n print(VARIABLES)\n \n db = pytimber.LoggingDB()\n data_calc = pd.DataFrame(db.getAligned(VARIABLES, t1, t2))#, master=master))\n print(data_calc['timestamps'].head())\n \n average_lifetimeB1 = 0.\n average_lifetimeB2 = 0.\n\n min_lifetimeB1 = 10000.\n min_lifetimeB2 = 10000.\n\n min_lifetime_time_B1 = 0\n min_lifetime_time_B2 = 0\n\n total = len(data[\"timestamps\"])\n for i in range(total):\n\n energy = (data_calc[ENERGY].iloc[i])*6500./54175.\n lifetimeB1, lifetimeB2 = GetBeamLifetimeIndex(data_calc, i)\n average_lifetimeB1 += lifetimeB1\n average_lifetimeB2 += lifetimeB2\n\n if lifetimeB1 < min_lifetimeB1: \n min_lifetimeB1 = lifetimeB1\n min_lifetime_time_B1 = i\n\n if lifetimeB2 < min_lifetimeB2: \n min_lifetimeB2 = lifetimeB2\n min_lifetime_time_B2 = i\n\n average_lifetimeB1 = average_lifetimeB1/total\n average_lifetimeB2 = average_lifetimeB2/total\n\n print(average_lifetimeB2, min_lifetimeB2)\n\n intensity_start_B1 = data_calc[BCTB1].iloc[0]\n intensity_start_B2 = data_calc[BCTB2].iloc[0]\n intensity_end_B1 = data_calc[BCTB1].iloc[-1]\n intensity_end_B2 = data_calc[BCTB2].iloc[-1]\n\n return [energy, intensity_start_B1, intensity_start_B2, intensity_end_B1, intensity_end_B2, average_lifetimeB1, average_lifetimeB2, min_lifetimeB1, min_lifetimeB2, min_lifetime_time_B1, min_lifetime_time_B2]\n \n \ndef CompleteLifetime(data, master, db, t1=None, t2=None, verbose=False):\n '''\n CompleteLifetime takes a dataframe and filling the missing lifetimes by \n fetching the intensity & losses and calculates the lifetimes.\n If the user specifies a t1/t2 then lifetimes will be calculated and filled\n filled in whithin said interval. Otherwise, the lifetimes will be calculated\n for timestamps where NaN are present in both lifetimes.\n '''\n LIFETIME_B1 = 'LHC.BLM.LIFETIME:B1_BEAM_LIFETIME'\n LIFETIME_B2 = 'LHC.BLM.LIFETIME:B2_BEAM_LIFETIME'\n \n blmListB2 = [\"BLMEI.06R7.B2I10_TCHSS.6R7.B2:LOSS_RS09\"]\n blmListB2 += [\"BLMEI.06R7.B2I10_TCHSH.6R7.B2:LOSS_RS09\"]\n blmListB2 += [\"BLMEI.06R7.B2I10_TCHSV.6R7.B2:LOSS_RS09\"]\n blmListB2 += [\"BLMEI.06R7.B2I10_TCP.A6R7.B2:LOSS_RS09\"]\n\n blmListB1 = [\"BLMEI.06L7.B1E10_TCHSS.6L7.B1:LOSS_RS09\"]\n blmListB1 += [\"BLMEI.06L7.B1E10_TCHSH.6L7.B1:LOSS_RS09\"]\n blmListB1 += [\"BLMEI.06L7.B1E10_TCHSV.6L7.B1:LOSS_RS09\"]\n blmListB1 += [\"BLMEI.06L7.B1E10_TCP.A6L7.B1:LOSS_RS09\"]\n \n BCTB1 = \"LHC.BCTFR.A6R4.B1:BEAM_INTENSITY\"\n BCTB2 = \"LHC.BCTFR.A6R4.B2:BEAM_INTENSITY\"\n \n ENERGY = \"HX:ENG\"\n \n VARIABLES = [master, BCTB1, BCTB2, ENERGY] + blmListB1 + blmListB2\n \n #Check if the NaN are in the same places for both beams\n if not (data['timestamps'][data[LIFETIME_B1].isna()] == data['timestamps'][data[LIFETIME_B2].isna()]).all():\n raise Exception('Timing of NaNs in B1 and B2 different!')\n \n time_LT_missing = data['timestamps'][data[LIFETIME_B1].isna()]\n \n \n if t1 == None and t2 == None:\n t1 = time_LT_missing.iloc[0]\n t2 = time_LT_missing.iloc[-1]\n \n \n data_calc = pd.DataFrame(db.getAligned(VARIABLES, t1, t2, master=master))\n \n if verbose:\n print('Start of NaN LT {} \\t {}'.format(t1, TimestampToDatetime(t1)))\n print('End of NaN LT {} \\t {}'.format(t2, TimestampToDatetime(t2)))\n print('Number of points {}'.format(len(time_LT_missing)))\n\n# print('Double Checking timestamps')\n# if (data_calc['timestamps'] == time_LT_missing).all():\n# print 'All good carry on !'\n \n lifetime_b1, lifetime_b2 = CalcBeamLifetime(data_calc)\n \n data[LIFETIME_B1] = lifetime_b1\n data[LIFETIME_B2] = lifetime_b2\n \n return data\n\n# Test\n#t1 = DatetimeToTimestamp('2017-09-04 00:00:00')\n#t2 = DatetimeToTimestamp('2017-09-04 01:00:00')\n#db = pytimber.LoggingDB()\n#data = db.getAligned(['LHC.BOFSU:TUNE_B1_H','LHC.BLM.LIFETIME:B1_BEAM_LIFETIME','LHC.BLM.LIFETIME:B2_BEAM_LIFETIME'],\n# t1, t2, master='LHC.BOFSU:TUNE_B1_H')\n#print(data.keys())\n#data = pd.DataFrame(data)\n#test = CompleteLifetime(data,'LHC.BOFSU:TUNE_B1_H',t1,t2) ","sub_path":"LHC_Lt.py","file_name":"LHC_Lt.py","file_ext":"py","file_size_in_byte":9019,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"64614030","text":"\"\"\"\nWorks well - add to prepare dataset file\n\"\"\"\n\nimport pandas as pd\nCOSMIC_num = 2\ndfincident = pd.read_csv(\"../../../Data/vw_Incident.csv\", encoding='latin-1', low_memory=False)\ndfholdactivity = pd.read_csv(\"../../../Data/vw_HoldActivity.csv\", encoding='latin-1', low_memory=False)\n\n# Create a new df with the unique Ticket numbers and 0 for hold durations\ndfduration = pd.DataFrame(dfholdactivity[\"TicketNumber\"].unique(), columns=[\"TicketNumber\"])\ndfduration[\"HoldDuration\"] = 0\n\nuniques = dfholdactivity[\"TicketNumber\"].unique()\n# For each of the unique ticket numbers in holdactivity: sum the hold durations and store them next to the equivalent\n# ticket in the new dfduration df\nfor ticket in uniques:\n duration = dfholdactivity.loc[dfholdactivity[\"TicketNumber\"] == ticket, 'HoldDuration'].sum()\n dfduration.loc[dfduration[\"TicketNumber\"] == ticket, \"HoldDuration\"] = duration\n\n# merge new dfduration df with dfincident based on ticket number\ndfincident = dfincident.merge(dfduration,how='left', left_on='TicketNumber', right_on='TicketNumber')\n\n# fill the NANs with 0's\ndfincident[\"HoldDuration\"].fillna(0, inplace=True)\n\n# save the new Analytical Base Table\ndfincident.to_csv(\"../../../Data/ABT_Incident_HoldDuration.csv\", index=False)","sub_path":"4th Iteration/3. Data Preparation/ABT_Incident_Hold.py","file_name":"ABT_Incident_Hold.py","file_ext":"py","file_size_in_byte":1253,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"62152331","text":"# -*- coding: utf-8 -*-\n\"\"\"\n/dms/gallery/views_exibition.py\n\n.. zeigt den Inhalt einer Galerie an\n Django content Management System\n\nHans Rauch\nhans.rauch@gmx.net\n\nDie Programme des dms-Systems koennen frei genutzt und den spezifischen\nBeduerfnissen entsprechend angepasst werden.\n\n0.01 01.11.2007 Beginn der Arbeit\n\"\"\"\n\nfrom django.shortcuts import render_to_response\n\nfrom django.template.loader import get_template\nfrom django.template import Context\nfrom django.utils.translation import ugettext as _\n\nfrom dms.queries import get_folder_filtered_items\nfrom dms.queries import get_visible_comment_count_by_item_containers\nfrom dms.queries import get_site_url\nfrom dms.queries import get_item_container_by_path_and_name\nfrom dms.queries import get_visible_comment_count_by_item_containers\n\nfrom dms.utils_form import get_folderish_vars_show\nfrom dms.utils import show_link\n\nfrom dms.folder.utils import get_folder_content\nfrom dms.gallery.utils import get_user_support\nfrom dms.file.utils import get_file_url\n\nfrom dms_ext.extension import * # dms-Funktionen ueberschreiben\n\n# -----------------------------------------------------\ndef gallery_exibition(request, item_container):\n \"\"\" zeigt den Inhalt der Galerie als Ausstellung an \"\"\"\n\n def get_photo_name_middle(item_container):\n \"\"\" ..liefert die Namen der verkleinerten Bilder \"\"\"\n file_name = get_file_url(item_container)\n ext_pos = file_name.rfind('.')\n file_name_small = file_name[:ext_pos] + '_middle' + file_name[ext_pos:]\n return file_name_small\n\n def get_color(color):\n if color < 0 : color = 0\n if color > 10: color = 10\n h = 3*hex(int(25.5*color))[2:]\n return '#%s' % h\n\n def get_prev_next(item_containers, this_name):\n \"\"\" liefert Vorgaenger und Nachfolger \"\"\"\n if len(item_containers) == 0:\n return '', '', '', ''\n else:\n first = item_containers[0].item.name\n last = item_containers[len(item_containers)-1].item.name\n n_curr = 0\n while n_curr < len(item_containers) and item_containers[n_curr].item.name != this_name:\n n_curr += 1\n if n_curr > 0:\n prev = item_containers[n_curr-1].item.name\n else:\n prev = ''\n if n_curr < len(item_containers)-1:\n next = item_containers[n_curr+1].item.name\n else:\n next = ''\n return first, prev, next, last\n\n app_name = 'gallery'\n item_containers = get_folder_filtered_items(item_container, False, ['dmsPhoto'])\n if request.GET.has_key('image'):\n this_name = request.GET['image']\n elif len(item_containers) > 0:\n this_name = item_containers[0].item.name\n else:\n this_name = ''\n t_image = get_template('app/photo/exibition_image.html')\n first, prev, next, last = get_prev_next(item_containers, this_name)\n if first != '':\n url_pattern = item_container.get_absolute_url() + '/exibition/?image='\n first_url = url_pattern + first\n if prev != '':\n prev_url = url_pattern + prev\n else:\n prev_url = ''\n if next != '':\n next_url = url_pattern + next\n else:\n next_url = ''\n last_url = url_pattern + last\n this_ic = get_item_container_by_path_and_name(item_container.container.path, this_name)\n comment_counts = get_visible_comment_count_by_item_containers(item_containers)\n c_image = Context ({\n 'name' : this_name,\n 'title' : this_ic.item.title,\n 'text' : this_ic.item.text,\n 'text_more' : this_ic.item.text_more,\n 'first_url' : first_url,\n 'prev_url' : prev_url,\n 'next_url' : next_url,\n 'last_url' : last_url,\n 'date' : this_ic.get_last_modified(),\n 'image_url' : get_photo_name_middle(this_ic),\n 'section' : this_ic.section,\n 'last_modified' : this_ic.get_last_modified(),\n 'comments' : comment_counts[this_ic.item.id],\n })\n vars = get_folderish_vars_show(request, item_container, app_name, t_image.render(c_image),\n get_user_support(item_container))\n else:\n first_url = ''\n prev_url = ''\n next_url = ''\n last_url = ''\n this_image = Context ({\n 'name' : '',\n 'title' : _(u'Die Ausstellung ist noch nicht geöffnet!'),\n })\n vars = get_folderish_vars_show(request, item_container, app_name, t_image.render(this_image),\n get_user_support(item_container))\n vars['no_breadcrum'] = True\n color = int(item_container.item.string_2)\n vars['bg_color'] = get_color(color)\n if color < 4:\n vars['text_color'] = get_color(10)\n else:\n vars['text_color'] = get_color(0)\n return render_to_response('app/gallery/exibition.html', vars)\n","sub_path":"gallery/views_exibition.py","file_name":"views_exibition.py","file_ext":"py","file_size_in_byte":5063,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"111760362","text":"\"\"\"\nSimple implementations of a greedy transition systems derivations. See Goldberg and Nivre (2013). \n\"\"\"\n\nfrom structure import *\n\nSHIFT = 0\nRIGHT = 1\nLEFT = 2\nREDUCE = 3\n\n\nclass ArcHybrid(object):\n \"\"\"Arc Hybrid transition system derivation.\"\"\"\n def __init__(self, words):\n \"\"\"The initial configuration is given by an empty stack, empty parse tree \n and buffer = [ROOT, word1, word2, .., wordn]\"\"\"\n if words[0] != 'ROOT':\n words.insert(0, 'ROOT')\n self.n = len(words)\n self.i = 0\n self.stack = []\n self.parse = Parse(self.n)\n\n @staticmethod\n def supported_moves():\n \"\"\"The system has 3 transitions\"\"\"\n return SHIFT, RIGHT, LEFT\n\n def terminal_config(self):\n \"\"\"Test if the current system state indicates that it has processed all the words in the buffer. \n That is, the buffer is empty and a the stack contains the single node ROOT\"\"\"\n empty_buffer = self.i >= self.n - 1\n return empty_buffer and len(self.stack) == 1 and self.stack[-1] == 0\n\n def transition(self, move):\n \"\"\"Performs a transition move\"\"\"\n assert move in self.supported_moves()\n if move == SHIFT:\n # Consumes a word of the buffer by pushing it into the stack and then moving the buffer index\n self.stack.append(self.i)\n self.i += 1\n elif move == RIGHT:\n # Pops the top of the stack (s0) and add an arc from the new top to s0 in the partial dependency tree\n self.parse.add(self.stack[-2], self.stack.pop())\n elif move == LEFT:\n # Pops the top of the stack (s0) and add an arc from the current buffer word to s0\n self.parse.add(self.i, self.stack.pop())\n\n def valid_moves(self):\n \"\"\"Return the legal moves for the current state of the system.\"\"\"\n stack_depth = len(self.stack)\n valid_moves = []\n if (self.i+1) < self.n:\n # Shift is valid when the buffer is not empty\n valid_moves.append(SHIFT)\n if stack_depth >= 2:\n # Right is valid when the stack has at least two elements\n valid_moves.append(RIGHT)\n if stack_depth >= 1 and self.stack[-1] != 0:\n # Left is valid when the stack is not empty and the top is not the ROOT node\n valid_moves.append(LEFT)\n return valid_moves\n\n def gold_moves(self, gold_heads):\n def deps_between(target, others, gold_heads):\n for word in others:\n if gold_heads[word] == target or gold_heads[target] == word:\n return True\n return False\n n0 = self.i\n valid = self.valid_moves()\n if not self.stack or (SHIFT in valid and gold_heads[n0] == self.stack[-1]):\n return [SHIFT]\n if gold_heads[self.stack[-1]] == n0:\n return [LEFT]\n costly = set([m for m in self.supported_moves() if m not in valid])\n # If the word behind s0 is its gold head, Left is incorrect\n if len(self.stack) >= 2 and gold_heads[self.stack[-1]] == self.stack[-2]:\n costly.add(LEFT)\n # If there are any dependencies between n0 and the stack,\n # pushing n0 will lose them.\n if SHIFT not in costly and deps_between(n0, self.stack, gold_heads):\n costly.add(SHIFT)\n # If there are any dependencies between s0 and the buffer, popping\n # s0 will lose them.\n if deps_between(self.stack[-1], range(n0, self.n), gold_heads):\n costly.add(LEFT)\n costly.add(RIGHT)\n gold_mv = [m for m in self.supported_moves() if m not in costly]\n #assert gold_mv\n if not gold_mv:\n gold_mv = self.valid_moves()\n #gold_mv = self.sgold_moves(gold_heads)\n return gold_mv\n\n def sgold_moves(self, gold_heads):\n \"\"\"Return the correct moves for the current state of the system.\"\"\"\n # Costs of the moves\n cost = [0]*3\n # The current position in the buffer\n n0 = self.i\n # The entire buffer\n buff = range(0, self.n-1)\n # The valid moves\n valid = self.valid_moves()\n # The top two elements of the stack\n s0 = None if len(self.stack) < 1 else self.stack[-1]\n s1 = None if len(self.stack) < 2 else self.stack[-2]\n if RIGHT in valid and s1:\n cost[RIGHT] = len([i for i in buff if gold_heads[i] == s0 or gold_heads[s0] == i])\n if LEFT in valid:\n if s0:\n cost[LEFT] = len([i for i in buff if gold_heads[i] == s0])\n if s1:\n cost[LEFT] += len([i for i in buff[n0:] + [s1] if gold_heads[s0] == i])\n if SHIFT in valid and s0:\n # Shift is costly if there are any dependencies between n0 and the stack (without top)\n cost[SHIFT] = len([i for i in self.stack if gold_heads[n0] == i])\n cost[SHIFT] += len([i for i in self.stack[0:-1] if gold_heads[n0] == i])\n gold_move_cost = min(cost)\n return [m for m, c in enumerate(cost) if c == gold_move_cost]\n\n def correct_heads(self, gold_heads):\n \"\"\"Counts the number of heads that match with the golden heads\"\"\"\n return len([i for i in range(self.n - 1) if self.parse.heads[i] == gold_heads[i]])\n\n\nclass ArcEager(object):\n \"\"\"Arc Eager transition system derivation.\"\"\"\n def __init__(self, words):\n \"\"\"The initial configuration is given by an empty stack, empty parse tree \n and buffer = [word1, word2, .., wordn, ROOT]\"\"\"\n if words[-1] != 'ROOT':\n words.append('ROOT')\n self.n = len(words)\n self.i = 0\n self.stack = []\n self.parse = Parse(self.n)\n\n @staticmethod\n def supported_moves():\n \"\"\"Arc Eager system has 4 transitions\"\"\"\n return SHIFT, RIGHT, LEFT, REDUCE\n\n def terminal_config(self):\n \"\"\"Test if the current system state indicates that it has processed all the words in the buffer. \n That is, the buffer contains the single node ROOT and the stack is empty\"\"\"\n return not self.stack and self.i >= self.n - 1\n\n def transition(self, move):\n \"\"\"Performs a transition move\"\"\"\n assert move in self.supported_moves()\n if move == SHIFT:\n # Consumes a word of the buffer by pushing it into the stack and then moving the buffer index\n self.stack.append(self.i)\n self.i += 1\n elif move == RIGHT:\n # Consumes a word of the buffer by pushing it into the stack, adds (s0, n0) to the parsed tree\n self.stack.append(self.i)\n self.i += 1\n self.parse.add(self.stack[-2], self.stack[-1])\n elif move == LEFT:\n # Pops the top of the stack and add an arc from the current buffer index to it\n self.parse.add(self.i, self.stack.pop())\n elif move == REDUCE:\n self.stack.pop()\n\n def valid_moves(self):\n \"\"\"Return the legal moves for the current state of the system.\"\"\"\n stack_depth = len(self.stack)\n valid_moves = []\n if self.i != self.n - 1:\n # Right and Shift can be legal oly if the current buffer element is not ROOT\n valid_moves.append(SHIFT)\n if stack_depth > 0:\n valid_moves.append(RIGHT)\n if stack_depth > 0:\n if [i for i in self.parse.heads if self.parse.heads[self.stack[-1]] == i]:\n # Reduce is legal if the stack top has parent in the parse\n valid_moves.append(REDUCE)\n else:\n # Left is legal if the stack top does note have a parent in the parse\n valid_moves.append(LEFT)\n return valid_moves\n\n def gold_moves(self, gold_heads):\n \"\"\"Return the correct moves for the current state of the system.\"\"\"\n # Costs of the moves\n cost = [0]*4\n # The current position in the buffer\n n0 = self.i\n # The entire buffer\n buff = range(0, self.n-1)\n # The valid moves\n valid = self.valid_moves()\n # The top of the stack\n s0 = None if len(self.stack) < 1 else self.stack[-1]\n\n if LEFT in valid and s0:\n # Left has zero cost when there is an arc (n0, s0) in the gold tree\n # TODO: and also when n0 is not the gold head of s0\n # TODO: but the real head is not in buffer and there are no dependents of s0 in the buffer\n if gold_heads[s0] != n0:\n cost[LEFT] = len([k for k in buff if gold_heads[k] == s0 or gold_heads[s0] == k])\n\n if RIGHT in valid and s0:\n # Right has zero cost when there is an arc (s0, n0) in the gold tree\n if gold_heads[n0] != s0:\n cost[RIGHT] = len([k for k in buff + self.stack if gold_heads[n0] == k])\n cost[RIGHT] += len([k for k in self.stack if gold_heads[k] == s0\n and not len([x for x in self.parse.heads if self.parse.heads[k] == x])])\n\n if SHIFT in valid and s0:\n # for the second case only when there are no arcs (x, n0) in the parse\n cost[SHIFT] = len([k for k in self.stack if gold_heads[k] == n0\n or (gold_heads[n0] == k\n and not len([x for x in self.parse.heads if self.parse.heads[k] == x]))])\n\n if REDUCE in valid and s0:\n cost[REDUCE] = len([k for k in buff if gold_heads[k] == s0])\n\n gold_move_cost = min(cost)\n return [m for m, c in enumerate(cost) if c == gold_move_cost]\n\n def correct_heads(self, gold_heads):\n \"\"\"Counts the number of heads that match with the golden heads\"\"\"\n return len([i for i in range(self.n - 1) if self.parse.heads[i] == gold_heads[i]])\n\n\n\"\"\"\nfrom abc import ABCMeta, abstractmethod\nclass AbstractTransitionSystem(metaclass=ABCMeta):\n @abstractmethod\n def supported_moves(self): pass\n\n @abstractmethod\n def valid_moves(self): pass\n\n @abstractmethod\n def gold_moves(self, gold_heads): pass\n\n @abstractmethod\n def transition(self, move): pass\n\n @abstractmethod\n def terminal_config(self): pass\n\n @abstractmethod\n def correct_heads(self, gold_heads): pass\n\"\"\"","sub_path":"transition_system.py","file_name":"transition_system.py","file_ext":"py","file_size_in_byte":10269,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"183655994","text":"# Definition for singly-linked list.\n# class ListNode(object):\n# def __init__(self, x):\n# self.val = x\n# self.next = None\n\nclass Solution(object):\n def reverseKGroup(self, head, k):\n \"\"\"\n :type head: ListNode\n :type k: int\n :rtype: ListNode\n \"\"\"\n def reverse(begin,end):\n tmp = None\n cur = begin\n while cur != end:\n last = cur.next\n cur.next = tmp\n tmp = cur\n cur = last\n end = begin\n begin = tmp\n return begin,end\n\n if head == None:\n return head\n dummy = ListNode(-1)\n dummy.next = head\n pre_begin = dummy\n begin = head\n end = head\n while end != None:\n for _ in range(k-1):\n end = end.next\n if end == None:\n return dummy.next\n tmp = end.next\n begin,end = reverse(begin,tmp)\n end.next = tmp\n pre_begin.next = begin\n pre_begin = end\n begin = tmp\n end = tmp\n return dummy.next\n","sub_path":"week13/25.py","file_name":"25.py","file_ext":"py","file_size_in_byte":1170,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"636234039","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"Hope chatbot in your terminal (console)\"\"\"\nfrom __future__ import division, print_function, absolute_import, unicode_literals\nfrom builtins import str\n\nimport os\nimport argparse\nimport sys\nimport logging\nimport json\n\nimport django\nfrom pugnlp.util import PrettyDict\n\nfrom hope import __version__\nfrom .constants import DB_PATH\n\nos.environ.setdefault(\"DJANGO_SETTINGS_MODULE\", \"chatterbot_app.settings\")\ndjango.setup()\n\nfrom chatterbot.ext.django_chatterbot.models import Statement, Response # noqa\n\nDEFAULT_FILENAME = \"newb+english.json\"\n\n__author__ = \"Hobson Lane\"\n__copyright__ = \"Hobson Lane\"\n__license__ = \"mit\"\n\nlogging.basicConfig(level=logging.INFO)\nlogger = logging.getLogger(__name__)\n\n\ndef load_json(filename=DEFAULT_FILENAME, loglevel=logging.INFO, database=DB_PATH):\n full_path = os.path.join(os.path.dirname(os.path.abspath(database)), 'corpora', filename)\n full_path = filename if not os.path.isfile(full_path) else full_path\n with open(full_path) as fin:\n js = json.load(fin)\n logger.debug(str(PrettyDict(js)))\n js = js.get('export', js)\n num_statements = 0\n for prompt_and_response in js:\n print(prompt_and_response)\n prompt, p_created = Statement.objects.get_or_create(text=prompt_and_response[0])\n response, r_created = Statement.objects.get_or_create(text=prompt_and_response[1])\n pair, pair_created = Response.objects.get_or_create(prompt=prompt, response=response)\n pair.occurrence += 1\n num_statements += 2\n return num_statements\n\n\ndef parse_args(args):\n \"\"\"Parse command line parameters\n\n Arguments:\n args (list of str): command line parameters\n\n Returns:\n argparse.Namespace: object with an attribute for each command line arg\n \"\"\"\n parser = argparse.ArgumentParser(\n description=\"Simple ChatterBot on stdin and stdout with database at {}\".format(DB_PATH))\n parser.add_argument(\n '--version',\n action='version',\n version='hope {ver}'.format(ver=__version__))\n parser.add_argument(\n '-d',\n '--db',\n '--dbpath',\n '--db-path',\n '--database',\n '--database-path',\n dest=\"database\",\n default=DB_PATH,\n help=\"Path to sqlite3 database file where new records will be loaded\",\n type=str,\n metavar=\"DB_PATH\")\n parser.add_argument(\n '-f',\n '--file',\n '--filename'\n '--json',\n dest=\"filename\",\n default=DEFAULT_FILENAME,\n help=\"Full path or relative path (to the data/ dir) to a json file containing a ChatBot export\",\n type=str,\n metavar=\"FILENAME\")\n parser.add_argument(\n '-v',\n '--verbose',\n dest=\"loglevel\",\n help=\"set loglevel to INFO\",\n action='store_const',\n const=logging.INFO)\n parser.add_argument(\n '-vv',\n '--very-verbose',\n dest=\"loglevel\",\n help=\"set loglevel to DEBUG\",\n action='store_const',\n const=logging.DEBUG)\n return parser.parse_args(args)\n\n\ndef main(args):\n args = parse_args(args)\n logger.info(\"Loading {} into {}...\".format(args.filename, args.database))\n total_statements = Statement.objects.count()\n total_responses = Response.objects.count()\n logger.debug(\"Before loading json data there were {} statements and {} unique statement-response pairs\".format(\n total_statements, total_responses))\n\n num_statements = load_json(filename=args.filename, loglevel=args.loglevel, database=args.database)\n\n total_statements = Statement.objects.count()\n total_responses = Response.objects.count()\n logger.debug(\"After loading json data there were {} statements and {} unique statement-response pairs\".format(\n total_statements, total_responses))\n\n logger.info(\"{} statements resulting in {} statements and {} responses in the DB.\".format(\n num_statements, total_statements, total_responses))\n\n\ndef run():\n os.environ.setdefault(\"DJANGO_SETTINGS_MODULE\", \"chatterbot_app.settings\")\n django.setup()\n main(sys.argv[1:])\n\n\nif __name__ == \"__main__\":\n run()\n","sub_path":"hope/load_json.py","file_name":"load_json.py","file_ext":"py","file_size_in_byte":4160,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"192520933","text":"data_raw_relev = open('SemEval2016-Task3-CQA-QL-test-subtaskA.xml.subtaskA.relevancy', 'rb').readlines()\n\ndata_raw_pred = open('semeval2016-task3-taskA-test.tsv', 'rb').readlines()\n\ndata_id_relevant = map(lambda x: x.replace('\\n', '').split('\\t'), data_raw_relev)\ndata_id_pred = map(lambda x: x.replace('\\n', '').split('\\t'), data_raw_pred)\n\nrel_id = set(map(lambda x: x[0], data_id_relevant))\n\ndata_id_pred_fil = filter(lambda x: x[0] in rel_id, data_id_pred)\n#data_id_pred_fil = map(lambda x: [x[0], x[1], x[2], str(float(x[3])), x[4]], data_id_pred_fil)\n\ndata_pred = map(lambda x: '\\t'.join(x) + '\\n', data_id_pred_fil)\n\nopen('semeval2016-task3-taskA-test-filter.tsv', 'wb').writelines(data_pred)","sub_path":"community-qa/data_folder/rearrange_data.py","file_name":"rearrange_data.py","file_ext":"py","file_size_in_byte":699,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"312160339","text":"from selenium import webdriver\nfrom selenium.webdriver.support.ui import Select\nfrom selenium.common.exceptions import NoSuchElementException\nfrom selenium.common.exceptions import StaleElementReferenceException\nfrom bs4 import BeautifulSoup\nfrom time import sleep\n\n\nSLEEP_TIME = 2\n\n\ndef write_data(data):\n with open('data.txt', 'a') as file:\n for _, text in enumerate(data):\n file.write(text)\n\n\ndef main():\n driver = webdriver.Chrome('./chromedriver')\n driver.get(\"http://desarrollos.cesvicolombia.com/pctweb/Master/frmConsultas.aspx\")\n\n select = Select(driver.find_element_by_id('ctl00_ContentPlaceHolder2_cboPaises'))\n select.select_by_visible_text('Colombia')\n\n sleep(SLEEP_TIME)\n departamentos = ['Norte de Santander de Santander', 'Santander']\n segmentos = ['1 - LIVIANOS', '2 - PESADOS']\n\n for _, departamento in enumerate(departamentos):\n select = Select(driver.find_element_by_id('ctl00_ContentPlaceHolder2_cboDepartamentos'))\n select.select_by_visible_text(departamento)\n\n sleep(SLEEP_TIME)\n\n html_municipios = driver.page_source\n soup_municipios = BeautifulSoup(html_municipios, \"html.parser\")\n\n municipios = soup_municipios.find('select', id=\"ctl00_ContentPlaceHolder2_cboMunicipio\")\n\n for municipio in municipios.stripped_strings:\n select = Select(driver.find_element_by_id('ctl00_ContentPlaceHolder2_cboMunicipio'))\n select.select_by_visible_text(municipio)\n\n sleep(SLEEP_TIME)\n\n for segmento in segmentos:\n select = Select(driver.find_element_by_id('ctl00_ContentPlaceHolder2_cboSegmento'))\n select.select_by_visible_text(segmento)\n\n sleep(SLEEP_TIME)\n\n select = Select(driver.find_element_by_id('ctl00_ContentPlaceHolder2_cboMarcas'))\n select.select_by_visible_text('todas')\n\n submit_button = driver.find_element_by_id('ctl00_ContentPlaceHolder2_btnConsultar')\n submit_button.click()\n\n sleep(SLEEP_TIME * 3)\n\n try:\n data = []\n table = driver.find_element_by_id('ctl00_ContentPlaceHolder1_DataList1')\n except StaleElementReferenceException:\n pass\n except NoSuchElementException:\n submit_button = driver.find_element_by_id('ctl00_ContentPlaceHolder2_ImageButton10')\n submit_button.click()\n data.append(\"************************************\\n\"\n \"Para la ciudad: {} con el segmento: {} y todas las marcas no se encontro resultado.\\n\"\n \"************************************\\n\"\n .format(municipio, segmento))\n write_data(data)\n sleep(SLEEP_TIME)\n else:\n html_datos = driver.page_source\n soup_datos = BeautifulSoup(html_datos, \"html.parser\")\n\n tds = soup_datos.find_all(\"td\", {\"class\": \"style73\"})\n\n data.append(\"************************************\\n\"\n \"Para la ciudad: {} con el segmento: {} y todas las marcas se encontro:\\n\"\n .format(municipio, segmento))\n\n write_data(data)\n\n for i, td in enumerate(tds):\n span = td.find(\"span\")\n info = span.text\n\n indentation = '\\t\\t'\n\n if i % 3 == 0:\n indentation = '\\t'\n if info == '':\n info = 'No se encontro información'\n write_data(indentation + info + '\\n')\n\n\nif __name__ == '__main__':\n main()\n\n\n\n\n","sub_path":"cesvicolombia_scraping/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":3889,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"644998597","text":"#-*-coding:utf-8-*-\n#author:PeterGu\n#e-mail:gucaca@yeah.net\n\n#1,2,3,4,生成所有可能的不重复位数三位数:\n\ndef print_num(list):\n\tnum = 0\n\tfor i in list:\n\t\tfor j in list:\n\t\t\tfor k in list:\n\t\t\t\tif (i!= j) and (i != k) and (j != k):\n\t\t\t\t\tprint('%d%d%d'%(i,j,k))\n\t\t\t\t\tnum = num + 1\n\tprint (u'一共有%d个三位数' % num)\n\n\n\nif __name__ == '__main__':\n\tlist1 = [1,2,3,4,5,6,7]\n\tprint_num(list1)","sub_path":"test1_num_key.py","file_name":"test1_num_key.py","file_ext":"py","file_size_in_byte":413,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"83537909","text":"import random\nimport time\n\n\"\"\"\nLegend:\n1. \".\" = empty space\n2. \"0\" = part of ship\n3. \"X\" = part of the ship that was hit with bullet\n4. \"#\" = empty space that was shot with a bullet, a miss because it hit no ship\n\"\"\"\n\n# variable for grid\ngrid = [[]]\n# variable for grid size\ngrid_size = 10\n# variable for number of ships to place\nnum_ships = 8\n# variable for bullets left\nbullets_left = 50\n# variable for game over\ngame_over = False\n# variable for ships sunk\nsunk_ships = 0\n# variable for ship positions\nship_positions = [[]]\n# variable for alphabet\nalphabet = \"ABCDEFGHIJKLMNOPQRSTUVWXYZ\"\n\n\ndef validate_grid_and_place_ship(\n start_row, end_row, start_col, end_col):\n\n # check to see if its safe to place a ship in certain row or column.\n\n global grid\n global ship_positions\n\n all_valid = True\n for r in range(start_row, end_row):\n for c in range(start_col, end_col):\n if grid[r][c] != '.':\n all_valid = False\n break\n if all_valid:\n ship_positions.append([start_row, end_row, start_col, end_col])\n for r in range(start_row, end_row):\n for c in range(start_col, end_col):\n grid[r][c] = 'O'\n return all_valid\n\n\ndef try_to_place_ship(\n row, col, direction, length):\n \"\"\"Based on direction will use above function to help in trying to place a\n ship on the grid.\n Returns validate_grid_and_place_ship which will either be true or false.\"\"\"\n\n global grid_size\n\n (start_row, end_row, start_col, end_col) = (\n row, row + 1, col, col + 1)\n if direction == 'left':\n if col - length < 0:\n return False\n start_col = col - length + 1\n elif direction == 'right':\n\n if col + length >= grid_size:\n return False\n end_col = col + length\n elif direction == 'up':\n\n if row - length < 0:\n return False\n start_row = row - length + 1\n elif direction == 'down':\n\n if row + length >= grid_size:\n return False\n end_row = row + length\n\n return validate_grid_and_place_ship(\n start_row, end_row, start_col, end_col)\n\n\ndef create_grid():\n \"\"\"Will create a 10x10 grid and randomly place down ships\n of different sizes in different directions\"\"\"\n global grid\n global grid_size\n global num_ships\n global ship_positions\n\n random.seed(time.time())\n\n rows, cols = (grid_size, grid_size)\n\n grid = []\n for _ in range(rows):\n row = []\n for _ in range(cols):\n row.append(\".\")\n grid.append(row)\n\n num_of_ships_placed = 0\n\n ship_positions = []\n\n while num_of_ships_placed != num_ships:\n random_row = random.randint(0, rows - 1)\n random_col = random.randint(0, cols - 1)\n direction = random.choice([\"left\", \"right\", \"up\", \"down\"])\n ship_size = random.randint(3, 5)\n if try_to_place_ship(\n random_row, random_col, direction, ship_size):\n num_of_ships_placed += 1\n\n\ndef print_grid():\n # Will print the grid with rows A-J and columns 0-9\n global grid\n global alphabet\n\n debug_mode = True\n\n alphabet = alphabet[0: len(grid) + 1]\n\n for row in range(len(grid)):\n print(alphabet[row], end=\") \")\n for col in range(len(grid[row])):\n if grid[row][col] == \"O\":\n if debug_mode:\n print(\"O\", end=\" \")\n else:\n print(\".\", end=\" \")\n else:\n print(grid[row][col], end=\" \")\n print(\"\")\n\n print(\" \", end=\" \")\n for i in range(len(grid[0])):\n print(str(i), end=\" \")\n print(\"\")\n\n\ndef valid_bullet_placement():\n \"\"\"Will get valid row and column to place bullet shot\"\"\"\n global alphabet\n global grid\n\n is_valid_placement = False\n row = -1\n col = -1\n while is_valid_placement is False:\n placement = input(\"Enter row (A-J) and column (0-9) such as A3:\\n\")\n placement = placement.upper()\n if len(placement) <= 0 or len(placement) > 2:\n print(\"Error: Please enter only one row and column such as A3\")\n continue\n row = placement[0]\n col = placement[1]\n if not row.isalpha() or not col.isnumeric():\n print(\n \"Error: Please enter letter (A-J) for row and (0-9) for column\"\n )\n continue\n row = alphabet.find(row)\n if not (-1 < row < grid_size):\n print(\n \"Error: Please enter letter (A-J) for row and (0-9) for column\"\n )\n continue\n col = int(col)\n if not (-1 < col < grid_size):\n print(\n \"Error: Please enter letter (A-J) for row and (0-9) for column\"\n )\n continue\n if grid[row][col] == \"#\" or grid[row][col] == \"X\":\n print(\"You have already shot a bullet here, pick somewhere else\")\n continue\n if grid[row][col] == \".\" or grid[row][col] == \"O\":\n is_valid_placement = True\n\n return row, col\n\n\ndef check_for_ship_sunk(row, col):\n \"\"\" If all parts of a shit have been shot it is sunk\n and we later increment ships sunk\"\"\"\n global ship_positions\n global grid\n\n for position in ship_positions:\n start_row = position[0]\n end_row = position[1]\n start_col = position[2]\n end_col = position[3]\n if start_row <= row <= end_row and start_col <= col <= end_col:\n # Ship found, now check if its all sunk\n for r in range(start_row, end_row):\n for c in range(start_col, end_col):\n if grid[r][c] != \"X\":\n return False\n return True\n\n\ndef shoot_bullet():\n # Updates grid and ships based on where the bullet was shot\n global grid\n global sunk_ships\n global bullets_left\n\n row, col = valid_bullet_placement()\n print(\"\")\n print(\"----------------------------\")\n\n if grid[row][col] == \".\":\n print(\"You missed, no ship was shot\")\n grid[row][col] = \"#\"\n elif grid[row][col] == \"O\":\n print(\"You hit!\", end=\" \")\n grid[row][col] = \"X\"\n if check_for_ship_sunk(row, col):\n print(\"A ship was completely sunk!\")\n sunk_ships += 1\n else:\n print(\"A ship was shot\")\n\n bullets_left -= 1\n\n\ndef check_game_over():\n \"\"\" when users shots = 0 or all ships are sunk the game is over,\n heres the function to implement this.\"\"\"\n global sunk_ships\n global num_ships\n global bullets_left\n global game_over\n\n if num_ships == sunk_ships:\n print(\"Congrats you won!\")\n game_over = True\n elif bullets_left <= 0:\n print(\"Sorry, you lost! You ran out of bullets, try again next time!\")\n game_over = True\n\n\ndef main():\n \"\"\"Main entry point of application that runs the game loop\"\"\"\n global game_over\n\n print(\"-----Welcome to Battleships-----\")\n print(\"You have 50 bullets to take down 8 ships, may the battle begin!\")\n\n create_grid()\n\n while game_over is False:\n print_grid()\n print(\n \"Number of ships remaining: \" + str(\n num_ships - sunk_ships))\n print(\"Number of bullets left: \" + str(bullets_left))\n shoot_bullet()\n print(\"----------------------------\")\n print(\"\")\n check_game_over()\n\n\nif __name__ == '__main__':\n # this will only be called when program is run from terminal or an IDE\n main()\n","sub_path":"run.py","file_name":"run.py","file_ext":"py","file_size_in_byte":7553,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"39530410","text":"\"\"\"Builds a model\"\"\"\n\nfrom tensorflow.keras.layers import Input, LSTM, Dense, Embedding, Dropout\nfrom tensorflow.keras.models import Model\nfrom tensorflow.keras.backend import sparse_categorical_crossentropy\n\n\ndef build_model(input_dims, vocab_size, embedding_matrix, hidden_dims, dropout):\n\n inps = Input((input_dims,), name=\"inputs\")\n\n x = Embedding(vocab_size,\n embedding_matrix.shape[1],\n weights=[embedding_matrix],\n trainable=False,\n input_length=input_dims,\n mask_zero=True)(inps)\n\n x = LSTM(hidden_dims)(x)\n\n x = Dropout(rate=dropout)(x)\n\n outputs = Dense(vocab_size, name=\"outputs\")(x)\n\n return Model(inps, outputs)\n\n\ndef custom_loss(y_true, y_pred):\n return sparse_categorical_crossentropy(y_true, y_pred, from_logits=True)","sub_path":"src/model.py","file_name":"model.py","file_ext":"py","file_size_in_byte":839,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"2065865","text":"# Turma B\n\n# Exercício número 1 por Juan Lucas Melo de Borba\n\nvogais = (\"A\", \"E\", \"I\", \"O\", \"U\") # variavel definindo vogais\n\nn = 0\n\nletras = (\"A\", \"Y\", \"U\", \"P\", \"X\", \"J\", \"I\", \"L\", \"E\", \"W\") # tupla\nprint(letras)\nprint(\"=\"*100)\n\ntam = len(letras)\n\nfor elemento in letras:\n if letras == vogais:\n n += 1\n\nvog = vogais-1\n\nprint(\"A string possui {} letras, {} são vogais e {} são consoantes\".format(tam, vog, con))\n\n\n\n#if letras == vogais:\n# print(\"A string possui \")\n#else:\n# print(\"É consoante\")","sub_path":"Listas/Listas 2/1.py","file_name":"1.py","file_ext":"py","file_size_in_byte":518,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"320738076","text":"import sys\nimport os\n\nsys.path.append(os.environ['INFRASTRUCTURE_BASE'])\nfrom infrastructure.loader import ldr\n\n\nclass EnterpriseConvSetup(object): \n\n \n tablecols = {'user_id': {'feedname': 'id', 'dtype': 'String'},\n 'sub_id': {'feedname': 'sub id', 'dtype': 'String'},\n 'voucher_creation_date': {'feedname': 'creation date', 'dtype': 'Date', 'dtype_param': {'formatmask': '%Y/%m/%d'}}\n }\n \n def enterprise_conv_setup(self, loader_type, persist_db=None, persist=True, feed_path = 'enterprise_conv_dummy'): \n \n schema = ldr.BaseSchema('enterprise_conv')\n schema.add('imp_enterprise_conv', self.tablecols)\n schema.subscribe('imp_enterprise_conv', loader_type)\n feed = ldr.CSVFeed('imp_enterprise_conv', 'enterprise_conv', feed_path)\n feed.subscribe(schema)\n if persist:\n if not persist_db:\n raise Exception('Cannot persist feed as no persist_db is provided')\n feed.persist(persist_db)\n return feed\n \n \n","sub_path":"clients/google/bin/enterprise_conversions/ent_conv_config.py","file_name":"ent_conv_config.py","file_ext":"py","file_size_in_byte":1061,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"538966312","text":"from typing import Iterable\n\nfrom diofant import sympify, Rational, Poly, prod, Symbol, symbols, oo, Max, Min, polylog, factorial, product, gamma\nfrom diofant.stats import Laplace, Binomial, Hypergeometric, E\nfrom scipy.stats import norm\nfrom math import sqrt\nimport re\n\nLOG_NOTHING = 0\nLOG_ESSENTIAL = 10\nLOG_VERBOSE = 20\nLOG_LEVEL = LOG_ESSENTIAL\n\nclass Update:\n # parse updates\n # takes string \"x = P @ p; Q @ q\" or x = RV(d, a, b)\n # creates class to deal with substituing powers of variables and moments\n def __init__(self, var, update_string=None, program_variables=None, is_random_var=False, random_var=None):\n self.is_random_var = is_random_var\n self.random_var = random_var\n self.var = var\n self.is_probabilistic = True\n\n if update_string is None:\n return\n\n # check if this is a RV or expression update\n rv = re.search(r\"RV\\((?P.+)\\)\", update_string)\n if rv is not None:\n self.is_random_var = True\n dist, *params = map(str.strip, rv.group('params').split(','))\n params = list(map(sympify, params))\n params = [make_symbols_positive(p, program_variables) for p in params]\n self.random_var = RandomVar(dist, params, var_name=str(self.var))\n\n # here: if not is_random_var == else\n if not self.is_random_var:\n self.branches = []\n branches = update_string.split(\";\")\n for update in branches:\n if '@' in update:\n exp, prob = update.split(\"@\")\n else:\n exp, prob = update, 1-sum([b[1] for b in self.branches])\n prob = sympify(prob)\n if not prob.is_zero:\n exp = make_symbols_positive(sympify(exp), program_variables)\n exp = make_floats_rational(exp)\n prob = make_symbols_positive(prob, program_variables)\n prob = make_floats_rational(prob)\n self.branches.append((sympify(exp), prob))\n if sum([b[1] for b in self.branches]) != 1:\n raise Exception(f\"Branch probabilities for {self.var} update do not sum up to 1. Terminating.\")\n\n def update_term(self, term, pow):\n if self.is_random_var:\n return term.subs({self.var**pow: self.random_var.compute_moment(pow) })\n else:\n return term.subs({self.var**pow: self.power(pow)})\n\n def power(self, k):\n return sum(prob * (exp ** k) for exp, prob in self.branches)\n\n\nclass RandomVar:\n def __init__(self, distribution, parameters, var_name=None):\n self.distribution = distribution\n self.parameters = parameters\n self.var_name = var_name\n\n def get_support(self, k=1):\n if self.distribution == 'bernoulli':\n return sympify(0), sympify(1)\n\n if self.distribution == 'geometric':\n return sympify(1), oo\n\n if self.distribution == 'exponential':\n return sympify(0), oo\n\n if self.distribution == 'beta':\n return sympify(0), sympify(1)\n\n if self.distribution == 'uniform':\n l, u = self.parameters\n return interval_to_power(l, u, k)\n\n if self.distribution == 'chi-squared':\n return sympify(0), oo\n\n if self.distribution == 'rayleigh':\n return sympify(0), oo\n\n if self.distribution == 'symbolic-support':\n l, u = self.parameters\n return interval_to_power(l, u, k)\n\n if self.distribution == 'gauss':\n return interval_to_power(-oo, oo, k)\n\n if self.distribution == 'laplace':\n return interval_to_power(-oo, oo, k)\n\n if self.distribution == 'binomial':\n n, _ = self.parameters\n return interval_to_power(0, n, k)\n\n if self.distribution == 'hypergeometric':\n N, K, n = self.parameters\n return interval_to_power(Max(0, n + K - N), Min(n, K), k)\n\n def compute_moment(self, k):\n if self.distribution == 'finite':\n return sum([p * (b ** k) for b, p in self.parameters])\n\n if self.distribution == 'uniform':\n l, u = self.parameters\n return (u**(k+1)-l**(k+1))/((k+1)*(u-l))\n\n if self.distribution == 'gauss' or self.distribution == 'normal':\n mu, sigma_squared = self.parameters\n # For low moments avoid scipy.stats.moments as it does not support\n # parametric parameters. In the future get all moments directly,\n # using the following properties:\n # https://math.stackexchange.com/questions/1945448/methods-for-finding-raw-moments-of-the-normal-distribution\n if k == 0:\n return 1\n elif k == 1:\n return mu\n elif k == 2:\n return mu**2 + sigma_squared\n elif k == 3:\n return mu*(mu**2 + 3*sigma_squared)\n elif k == 4:\n return mu**4 + 6*mu**2*sigma_squared + 3*sigma_squared**2\n moment = norm(loc=mu, scale=sqrt(sigma_squared)).moment(k)\n return Rational(moment)\n\n if self.distribution == 'bernoulli':\n return sympify(self.parameters[0])\n\n if self.distribution == 'geometric':\n p = sympify(self.parameters[0])\n return p*polylog(-k, 1-p)\n\n if self.distribution == 'exponential':\n lambd = sympify(self.parameters[0])\n return factorial(k) / (lambd ** k)\n\n if self.distribution == 'beta':\n alpha, beta = self.parameters\n alpha = sympify(alpha)\n beta = sympify(beta)\n r = symbols('r')\n return product((alpha + r) / (alpha + beta + r), (r, 0, k-1))\n\n if self.distribution == 'chi-squared':\n n = sympify(self.parameters[0])\n i = symbols('i')\n return product(n + 2*i, (i, 0, k - 1))\n\n if self.distribution == 'rayleigh':\n s = sympify(self.parameters[0])\n return (2**(k / 2)) * (s**k) * gamma(1 + k/2)\n\n if self.distribution == 'unknown':\n return sympify(f\"{self.var_name}(0)^{k}\")\n\n if self.distribution == 'laplace':\n mu, b = self.parameters\n mu = sympify(mu)\n b = sympify(b)\n x = Laplace(\"x\", mu, b)\n return E(x**k)\n\n if self.distribution == 'binomial':\n n, p = self.parameters\n n = sympify(n)\n p = sympify(p)\n x = Binomial(\"x\", n, p)\n return E(x**k)\n\n if self.distribution == 'hypergeometric':\n N, K, n = self.parameters\n N = sympify(N)\n K = sympify(K)\n n = sympify(n)\n x = Hypergeometric(\"x\", N, K, n)\n return E(x**k)\n\n\ndef EV(expression):\n if issubclass(type(expression), RandomVar):\n return expression.compute_moment(1)\n else:\n return expression\n\n\ndef get_exponent_of(var, mono):\n monoms = mono.as_poly([var]).monoms()\n if len(monoms) > 0 and len(monoms[0]) > 0:\n return monoms[0][0]\n return 0\n\n\ndef get_monoms(poly: Poly):\n \"\"\"\n Returns the list of monoms for a given polynomial\n \"\"\"\n monoms = []\n for powers in poly.monoms():\n m = prod(var ** power for var, power in zip(poly.gens, powers))\n if m != 1:\n monoms.append(m.as_poly(poly.gens))\n return monoms\n\n\ndef monomial_is_constant(monomial: Poly):\n \"\"\"\n Returns true iff the given monomial is constant\n \"\"\"\n if monomial.is_zero:\n return True\n powers = monomial.monoms()[0]\n return all(p == 0 for p in powers)\n\n\ndef is_independent_from_all(program, x, ys):\n \"\"\"\n Returns true iff x is statistially independent from all ys, where x is from the current iteration and the ys\n are from the previous iteration.\n \"\"\"\n if x not in program.ancestors[x]:\n return True\n\n for y in ys:\n if y in program.dependencies[x]:\n return False\n\n return True\n\n\ndef all_are_independent_from_all(program, xs, ys):\n \"\"\"\n Returns true iff all xs are statistially independent from all ys, where the xs are from the current iteration\n and the ys are from the previous iteration.\n \"\"\"\n for x in xs:\n if not is_independent_from_all(program, x, ys):\n return False\n return True\n\n\ndef get_powers_of_variable_in_polynomial(variable: Symbol, polynomial: Poly):\n \"\"\"\n Returns the set of all powers p for which variable ** p occurs in the polynomial\n \"\"\"\n monoms = get_monoms(polynomial)\n all_powers = []\n for monomial in monoms:\n powers = monomial.monoms()[0]\n powers_for_var = {var: power for var, power in zip(monomial.gens, powers) if power > 0}\n if variable in powers_for_var.keys():\n all_powers.append(powers_for_var[variable])\n all_powers.sort(reverse=True)\n return all_powers\n\n\ndef set_log_level(log_level):\n global LOG_LEVEL\n LOG_LEVEL = log_level\n\n\ndef log(message, level):\n \"\"\"\n Logs a message depending on the log level\n \"\"\"\n if level <= LOG_LEVEL:\n print(message)\n\n\ndef without_piecewise(expr):\n \"\"\"\n Removes the Piecewise from an expression by assuming that all restricting assumptions are false.\n \"\"\"\n if not expr.args:\n return expr\n\n if expr.is_Piecewise:\n return without_piecewise(expr.args[-1].expr)\n\n return expr.func(*[without_piecewise(a) for a in expr.args])\n\n\ndef make_symbols_positive(expr, exclude_symbols=None):\n \"\"\"\n Ensures that all symbols in a given expression (with the exception of excluded symbols) are assumed to be positive\n \"\"\"\n if expr.is_Symbol:\n if exclude_symbols is None or expr not in exclude_symbols:\n return symbols(expr.name, positive=True)\n\n if not expr.args:\n return expr\n\n return expr.func(*[make_symbols_positive(a, exclude_symbols) for a in expr.args])\n\n\ndef make_floats_rational(expr):\n \"\"\"\n Converts all floats in an expression to rationals\n \"\"\"\n if expr.is_Float:\n return Rational(expr)\n\n if not expr.args:\n return expr\n\n return expr.func(*[make_floats_rational(a) for a in expr.args])\n\n\ndef interval_to_power(low, high, power):\n \"\"\"\n If x in [low, high] returns an interval [l, h] s.t. x**power in [l, h]\n \"\"\"\n l = low ** power\n h = high ** power\n if power % 2 == 0:\n if high < 0:\n h, l = l, h\n elif low < 0:\n h = Max(h, l)\n l = sympify(0)\n return l, h\n","sub_path":"mora/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":10601,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"516678347","text":"#\n# @lc app=leetcode id=86 lang=python3\n#\n# [86] Partition List\n#\n# https://leetcode.com/problems/partition-list/description/\n#\n# algorithms\n# Medium (43.61%)\n# Likes: 2310\n# Dislikes: 410\n# Total Accepted: 279.5K\n# Total Submissions: 619.7K\n# Testcase Example: '[1,4,3,2,5,2]\\n3'\n#\n# Given the head of a linked list and a value x, partition it such that all\n# nodes less than x come before nodes greater than or equal to x.\n#\n# You should preserve the original relative order of the nodes in each of the\n# two partitions.\n#\n#\n# Example 1:\n#\n#\n# Input: head = [1,4,3,2,5,2], x = 3\n# Output: [1,2,2,4,3,5]\n#\n#\n# Example 2:\n#\n#\n# Input: head = [2,1], x = 2\n# Output: [1,2]\n#\n#\n#\n# Constraints:\n#\n#\n# The number of nodes in the list is in the range [0, 200].\n# -100 <= Node.val <= 100\n# -200 <= x <= 200\n#\n#\n#\n\nfrom Python.commons.LinkedList import ListNode, LinkedList\n\n\n# @lc code=start\n# Definition for singly-linked list.\n# class ListNode:\n# def __init__(self, val=0, next=None):\n# self.val = val\n# self.next = next\nclass Solution:\n def partition(self, head: ListNode, x: int) -> ListNode:\n \"\"\"\n Given the head of a linked list and a value x,\n partition it such that all nodes < x come before\n nodes >= to x.\n\n The original relative order of the nodes in each\n of the two partitions should be preserved.\n\n :param head: The head of a linked list to partition\n :param x: The value to partition by\n :return: The partitioned linked list where all nodes < x come before nodes that are >= x\n \"\"\"\n if not head:\n return head\n return self.two_list(head, x)\n\n def two_list(self, head: ListNode, x: int) -> ListNode:\n \"\"\"\n Two List Solution\n\n Build 2 LinkedLists from scratch (using dummy nodes)\n - head1 are all the nodes that are less than x\n - Use pt1 to keep track of the latest node\n - head2 are all the nodes that are greater than x\n - Use pt2 to keep track of the latest node\n\n After the 2 LinkedLists are built,\n - Clear the remaining of the pt2, since those are not valid\n - Add head2 to the end of head1 (pt1.next)\n\n Return head1\n\n Runtime: O(n)\n Space: O(n)\n \"\"\"\n head1 = pt1 = ListNode(-1)\n head2 = pt2 = ListNode(-1)\n curr = head\n\n while curr:\n if curr.val < x:\n pt1.next = curr\n pt1 = pt1.next\n else:\n pt2.next = curr\n pt2 = pt2.next\n curr = curr.next\n\n pt2.next = None\n pt1.next = head2.next\n\n return head1.next\n\n def two_pointer(self, head: ListNode, x: int) -> ListNode:\n \"\"\"\n Two Pointer Solution\n Less space complexity compared to two_list, but harder to follow\n\n - prev: last valid node which node.val < x\n - curr: last valid node which node.val >=x\n Use curr.next to perform checks\n\n If curr.next >= x:\n Nothing to do. Leave prev as is and increment curr\n Else:\n Use temp to keep track of curr.next (we want to move this)\n Update curr.next to curr.next.next (cut the link to temp)\n Update temp.next to prev.next (before linking temp to prev, save reference to prev.next)\n Update prev.next to temp (link temp to prev)\n Increment prev\n Do not increment curr since curr.next is updated\n\n Edge Case: (the tricky part)\n When the method starts, prev and curr are both at the dummy node\n The curr.next >= x case will work as normal\n The curr.next >= x case won't work and will cause infinite loop\n To deal with this, increment both to head\n prev will be valid, so will curr\n\n Runtime: O(n)\n Space: O(1)\n \"\"\"\n dummy = ListNode(-1)\n dummy.next = head\n prev = curr = dummy\n while curr and curr.next:\n if curr.next.val < x:\n if prev == curr:\n prev = prev.next\n curr = curr.next\n else:\n temp = curr.next\n curr.next = temp.next\n temp.next = prev.next\n prev.next = temp\n prev = prev.next\n else:\n curr = curr.next\n return dummy.next\n\n\n# @lc code=end\n\nif __name__ == \"__main__\":\n # Test Case 1\n input1 = LinkedList([1, 4, 3, 2, 5, 2]).get_head()\n input2 = 3\n expected = LinkedList([1, 2, 2, 4, 3, 5]).get_head()\n actual = Solution().partition(input1, input2)\n print(\"Test case 1\")\n print(actual)\n print(expected)\n\n # Test Case 2\n input1 = LinkedList([2, 1]).get_head()\n input2 = 2\n expected = LinkedList([1, 2]).get_head()\n actual = Solution().partition(input1, input2)\n print(\"Test case 1\")\n print(actual)\n print(expected)\n","sub_path":"Python/86.partition-list.py","file_name":"86.partition-list.py","file_ext":"py","file_size_in_byte":5005,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"9"} +{"seq_id":"277447089","text":"from gameengine.util.Vector2 import Vector2\n\nfrom gameengine.components.CustomDraw import CustomDraw\nfrom gameengine.components.Input import Input\nfrom gameengine.components.Script import Script\nfrom gameengine.core.GameObject import GameObject\nfrom gameengine.ui.Label import Label\nfrom gameengine.util.EventHandler import EventHandler\n\n\nclass Button(GameObject, Script):\n def __init__(self, text):\n super().__init__()\n self.addComponent(CustomDraw)\n self.addComponent(Input)\n self.addScript(self)\n\n self.transform.pivot.set(0.5, 0.5)\n\n self.label = Label(text)\n self.label.transform.parent = self.transform\n self.label.transform.depth = 1\n\n self.label.transform.anchor.set(0.5, 0.5, True, True)\n\n def updateButtonSize(label, size: Vector2):\n self.transform.size = Vector2(size)\n self.transform.size += Vector2(10, 5) # Padding\n\n self.label.transform.sizeChanged += updateButtonSize\n\n self.click = EventHandler(self)\n\n\n def onDraw(self):\n from gameengine.util.util import svgToSurface\n bg = svgToSurface(\"../ui/res/button-bg.svg\", *self.transform.size)\n return bg\n\n def onMouseClicked(self, pos):\n self.click()","sub_path":"gameengine/ui/Button.py","file_name":"Button.py","file_ext":"py","file_size_in_byte":1254,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"609635829","text":"'''\nVictoria Gao, Ishita Gupta, Eric Lo\nSoftDev\nK05 -- Teamwork, but Better This Time\n2020-09-25\n\nWe interpreted the assignment to require no user input, so we picked a random team and then a random member of that team.\n'''\n\n#Write a program that will print the name of a randomly-selected student from team (orpheus, rex, or endymion).\nimport random \nKREWES = {\n 'orpheus': ['ERIC', 'SAUVE', 'JONATHAN', 'PAK', 'LIAM', 'WINNIE', 'KELLY', 'JEFFREY', 'KARL', 'ISHITA', 'VICTORIA', 'BENJAMIN', 'ARIB', 'AMELIA', 'CONSTANCE', 'IAN'],\n 'rex': ['ANYA', 'DUB-Y', 'JESSICA', 'ALVIN', 'HELENA', 'MICHELLE', 'SHENKER', 'ARI', 'STELLA', 'RENEE', 'MADELYN', 'MAC', 'RYAN', 'DRAGOS'],\n 'endymion': ['JASON', 'DEAN', 'MADDIE', 'SAQIF', 'CINDY', 'YI LING', 'RUOSHUI', 'FB', 'MATTHEW', 'MAY', 'ERIN', 'MEIRU']\n}\n\n\"\"\"\nWe chose a team randomly by picking a random integer from [0,3).\nThis integer is the index of a team name from a list containing the KREWES dictionary's keys.\n\"\"\"\nteam_index = random.randrange(3)\nteam_name = list(KREWES)[team_index]\n\n\"\"\"\nAfter picking a team randomly, we found the number of names in that team to determine the\nend value of an interval.\nA random integer between 0 and that end value will be generated to represent the\nindex of a student in a team.\n\"\"\"\n\nteam_size = random.randrange(len(team_name))\nstudent = KREWES[team_name][team_size]\nprint(student)\n","sub_path":"05/krewes.py","file_name":"krewes.py","file_ext":"py","file_size_in_byte":1380,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"373343801","text":"#!python3\n#spreadsheetCellInverter.py - inverts the columns and rows\n\nimport os, sys, openpyxl\n\nabspath = os.path.abspath(sys.argv[0])\ndname = os.path.dirname(abspath)\nos.chdir(dname)\n\nwb = openpyxl.load_workbook(sys.argv[1])\nsheet = wb.active\n\ncellValues = {}\n#Saving values to a dictonary and deleting cells\nfor colNum in range(1, sheet.max_column + 1):\n for rowNum in range(1, sheet.max_row + 1):\n cellValues[colNum, rowNum] = sheet.cell(row = rowNum, column = colNum).value\n sheet.cell(row = rowNum, column = colNum).value = ''\n#Writing values from dictionary to sheet\nfor k, v in cellValues.items():\n sheet.cell(row = k[0], column = k[1]).value = v\n\n#Saving the file\nwb.save('inverted' + str(sys.argv[1])[0].upper() + str(sys.argv[1])[1:])\nprint(cellValues)","sub_path":"Chapter 13 - Working With Excel Spreadsheets/spreadsheetCellInverter.py","file_name":"spreadsheetCellInverter.py","file_ext":"py","file_size_in_byte":782,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"548669338","text":"import matplotlib as mpl\nfrom mpl_toolkits.mplot3d import Axes3D\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n# make 3d axes\nfig = plt.figure()\nax = fig.gca(projection='3d')\n\n\n# test data\nx = np.arange(-1., 1., .1)\ny = np.arange(-1., 1., .1)\nz1 = x + y\nz2 = x * x\nz3 = -y * y\nprint(x)\nprint(x)\n# plot test data\nax.plot(x, y, z1)\nax.plot(x, y, z2)\nax.plot(x, y, z3)\n\n# make labels\nax.set_xlabel('X')\nax.set_ylabel('Y')\nax.set_zlabel('Z')\n\nplt.show()","sub_path":"HELLOPYTHON/day14/my3dgraph.py","file_name":"my3dgraph.py","file_ext":"py","file_size_in_byte":455,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"330124854","text":"#Embedded file name: /Users/versonator/Jenkins/live/output/Live/mac_64_static/Release/python-bundle/MIDI Remote Scripts/Launchpad/DefChannelStripComponent.py\nfrom __future__ import absolute_import, print_function, unicode_literals\nimport Live\nfrom _Framework.ChannelStripComponent import ChannelStripComponent\nfrom .ConfigurableButtonElement import ConfigurableButtonElement\nfrom itertools import chain\n\nclass DefChannelStripComponent(ChannelStripComponent):\n u\"\"\" Subclass of channel strip component offering defaultbuttons for the timeables \"\"\"\n\n def __init__(self):\n ChannelStripComponent.__init__(self)\n self._default_volume_button = None\n self._default_panning_button = None\n self._default_send1_button = None\n self._default_send2_button = None\n self._invert_mute_feedback = True\n\n def disconnect(self):\n u\"\"\" releasing references and removing listeners\"\"\"\n if self._track != None:\n volume = self._track.mixer_device.volume\n panning = self._track.mixer_device.panning\n sends = self._track.mixer_device.sends\n if volume.value_has_listener(self._on_volume_changed):\n volume.remove_value_listener(self._on_volume_changed)\n if panning.value_has_listener(self._on_panning_changed):\n panning.remove_value_listener(self._on_panning_changed)\n if len(sends) > 0 and sends[0].value_has_listener(self._on_send1_changed):\n sends[0].remove_value_listener(self._on_send1_changed)\n if len(sends) > 1 and sends[1].value_has_listener(self._on_send2_changed):\n sends[1].remove_value_listener(self._on_send2_changed)\n if self._default_volume_button != None:\n self._default_volume_button.remove_value_listener(self._default_volume_value)\n self._default_volume_button = None\n if self._default_panning_button != None:\n self._default_panning_button.remove_value_listener(self._default_panning_value)\n self._default_panning_button = None\n if self._default_send1_button != None:\n self._default_send1_button.remove_value_listener(self._default_send1_value)\n self._default_send1_button = None\n if self._default_send2_button != None:\n self._default_send2_button.remove_value_listener(self._default_send2_value)\n self._default_send2_button = None\n ChannelStripComponent.disconnect(self)\n\n def set_track(self, track):\n assert track == None or isinstance(track, Live.Track.Track)\n if track != self._track:\n if self._track != None:\n volume = self._track.mixer_device.volume\n panning = self._track.mixer_device.panning\n sends = self._track.mixer_device.sends\n if volume.value_has_listener(self._on_volume_changed):\n volume.remove_value_listener(self._on_volume_changed)\n if panning.value_has_listener(self._on_panning_changed):\n panning.remove_value_listener(self._on_panning_changed)\n if len(sends) > 0 and sends[0].value_has_listener(self._on_send1_changed):\n sends[0].remove_value_listener(self._on_send1_changed)\n if len(sends) > 1 and sends[1].value_has_listener(self._on_send2_changed):\n sends[1].remove_value_listener(self._on_send2_changed)\n ChannelStripComponent.set_track(self, track)\n else:\n self.update()\n\n def set_default_buttons(self, volume, panning, send1, send2):\n assert volume == None or isinstance(volume, ConfigurableButtonElement)\n assert panning == None or isinstance(panning, ConfigurableButtonElement)\n assert send1 == None or isinstance(send1, ConfigurableButtonElement)\n assert send2 == None or isinstance(send2, ConfigurableButtonElement)\n if volume != self._default_volume_button:\n if self._default_volume_button != None:\n self._default_volume_button.remove_value_listener(self._default_volume_value)\n self._default_volume_button = volume\n if self._default_volume_button != None:\n self._default_volume_button.add_value_listener(self._default_volume_value)\n if panning != self._default_panning_button:\n if self._default_panning_button != None:\n self._default_panning_button.remove_value_listener(self._default_panning_value)\n self._default_panning_button = panning\n if self._default_panning_button != None:\n self._default_panning_button.add_value_listener(self._default_panning_value)\n if send1 != self._default_send1_button:\n if self._default_send1_button != None:\n self._default_send1_button.remove_value_listener(self._default_send1_value)\n self._default_send1_button = send1\n if self._default_send1_button != None:\n self._default_send1_button.add_value_listener(self._default_send1_value)\n if send2 != self._default_send2_button:\n if self._default_send2_button != None:\n self._default_send2_button.remove_value_listener(self._default_send2_value)\n self._default_send2_button = send2\n if self._default_send2_button != None:\n self._default_send2_button.add_value_listener(self._default_send2_value)\n self.update()\n\n def set_send_controls(self, controls):\n assert controls == None or isinstance(controls, tuple)\n if controls != self._send_controls:\n self._send_controls = controls\n if self._send_controls != None:\n for control in self._send_controls:\n if control != None:\n control.reset()\n\n self.update()\n\n def update(self):\n super(DefChannelStripComponent, self).update()\n if self._allow_updates:\n if self.is_enabled():\n if self._track != None:\n volume = self._track.mixer_device.volume\n panning = self._track.mixer_device.panning\n sends = self._track.mixer_device.sends\n if not volume.value_has_listener(self._on_volume_changed):\n volume.add_value_listener(self._on_volume_changed)\n if not panning.value_has_listener(self._on_panning_changed):\n panning.add_value_listener(self._on_panning_changed)\n if len(sends) > 0:\n if not sends[0].value_has_listener(self._on_send1_changed):\n sends[0].add_value_listener(self._on_send1_changed)\n self._on_send1_changed()\n elif self._default_send1_button != None:\n self._default_send1_button.turn_off()\n if len(sends) > 1:\n if not sends[1].value_has_listener(self._on_send2_changed):\n sends[1].add_value_listener(self._on_send2_changed)\n self._on_send2_changed()\n elif self._default_send2_button != None:\n self._default_send2_button.turn_off()\n self._on_volume_changed()\n self._on_panning_changed()\n else:\n if self._default_volume_button != None:\n self._default_volume_button.reset()\n if self._default_panning_button != None:\n self._default_panning_button.reset()\n if self._default_send1_button != None:\n self._default_send1_button.reset()\n if self._default_send2_button != None:\n self._default_send2_button.reset()\n if self._mute_button != None:\n self._mute_button.reset()\n if self._arm_button != None:\n self._arm_button.reset()\n if self._solo_button != None:\n self._solo_button.reset()\n if self._volume_control != None:\n self._volume_control.reset()\n if self._pan_control != None:\n self._pan_control.reset()\n if self._send_controls != None:\n for send_control in self._send_controls:\n if send_control != None:\n send_control.reset()\n\n def _default_volume_value(self, value):\n assert self._default_volume_button != None\n assert value in range(128)\n if self.is_enabled() and self._track != None:\n if value != 0 or not self._default_volume_button.is_momentary():\n volume = self._track.mixer_device.volume\n if volume.is_enabled:\n volume.value = volume.default_value\n\n def _default_panning_value(self, value):\n assert self._default_panning_button != None\n assert value in range(128)\n if self.is_enabled() and self._track != None:\n if value != 0 or not self._default_panning_button.is_momentary():\n panning = self._track.mixer_device.panning\n if panning.is_enabled:\n panning.value = panning.default_value\n\n def _default_send1_value(self, value):\n assert self._default_send1_button != None\n assert value in range(128)\n if self.is_enabled() and self._track != None and len(self._track.mixer_device.sends) > 0:\n if value != 0 or not self._default_send1_button.is_momentary():\n send1 = self._track.mixer_device.sends[0]\n if send1.is_enabled:\n send1.value = send1.default_value\n\n def _default_send2_value(self, value):\n assert self._default_send2_button != None\n assert value in range(128)\n if self.is_enabled() and self._track != None and len(self._track.mixer_device.sends) > 1:\n if value != 0 or not self._default_send2_button.is_momentary():\n send2 = self._track.mixer_device.sends[1]\n if send2.is_enabled:\n send2.value = send2.default_value\n\n def _on_mute_changed(self):\n if self.is_enabled() and self._mute_button != None:\n if self._track != None:\n if self._track in chain(self.song().tracks, self.song().return_tracks) and self._track.mute != self._invert_mute_feedback:\n self._mute_button.turn_on()\n else:\n self._mute_button.turn_off()\n else:\n self._mute_button.send_value(0)\n\n def _on_solo_changed(self):\n if self.is_enabled() and self._solo_button != None:\n if self._track != None:\n if self._track in chain(self.song().tracks, self.song().return_tracks) and self._track.solo:\n self._solo_button.turn_on()\n else:\n self._solo_button.turn_off()\n else:\n self._solo_button.send_value(0)\n\n def _on_arm_changed(self):\n if self.is_enabled() and self._arm_button != None:\n if self._track != None:\n if self._track in self.song().tracks and self._track.can_be_armed and self._track.arm:\n self._arm_button.turn_on()\n else:\n self._arm_button.turn_off()\n else:\n self._arm_button.send_value(0)\n\n def _on_volume_changed(self):\n assert self._track != None\n if self.is_enabled() and self._default_volume_button != None:\n volume = self._track.mixer_device.volume\n if volume.value == volume.default_value:\n self._default_volume_button.turn_on()\n else:\n self._default_volume_button.turn_off()\n\n def _on_panning_changed(self):\n assert self._track != None\n if self.is_enabled() and self._default_panning_button != None:\n panning = self._track.mixer_device.panning\n if panning.value == panning.default_value:\n self._default_panning_button.turn_on()\n else:\n self._default_panning_button.turn_off()\n\n def _on_send1_changed(self):\n assert self._track != None\n sends = self._track.mixer_device.sends\n assert len(sends) > 0\n if self.is_enabled() and self._default_send1_button != None:\n send1 = sends[0]\n if send1.value == send1.default_value:\n self._default_send1_button.turn_on()\n else:\n self._default_send1_button.turn_off()\n\n def _on_send2_changed(self):\n assert self._track != None\n sends = self._track.mixer_device.sends\n assert len(sends) > 1\n if self.is_enabled() and self._default_send2_button != None:\n send2 = sends[1]\n if send2.value == send2.default_value:\n self._default_send2_button.turn_on()\n else:\n self._default_send2_button.turn_off()\n","sub_path":"Launchpad/DefChannelStripComponent.py","file_name":"DefChannelStripComponent.py","file_ext":"py","file_size_in_byte":13263,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"89"} +{"seq_id":"566108851","text":"#PROBLEM 1:\n#Array partition\n\n\nclass PartitionArray:\n def partition_array(self,l):\n return (len(l)+1)//2\n \n def display_sub_lists(self):\n index=self.partition_array(sl)\n sub_lists=[sl[:index]]+[sl[index:]]\n return sub_lists\n\n\np=PartitionArray()\n \nl=[1,2,3,0,-1]\n\n#l=[24,5,6,-2,-1,-4,32,2]\n\nsl=sorted(l)\n\nprint(p.partition_array(sl))\n\nprint(p.display_sub_lists())\n\nprint('\\n')\n\nprint(\"################################\\n\")\n\n\n#PROBLEM 2:\n#Max continous seq of same color\n\nfrom collections import defaultdict\n\nclass MaxContinousSequence:\n\n def cont_seq(self,l):\n res = []\n for l1 in l:\n sub_list=[]\n index=1\n while index