diff --git "a/5095.jsonl" "b/5095.jsonl" new file mode 100644--- /dev/null +++ "b/5095.jsonl" @@ -0,0 +1,668 @@ +{"seq_id":"567033939","text":"\nprint(\"Welcome to the distance calculator. Follow the prompts to calculate the distance between two coordinates!\")\n\nx1 , y1 = eval(input(\"Please enter an x and y coordinate seperated by a comma: \"))\nx2 , y2 = eval(input(\"Please enter another x and y coordinate seperated by a comma: \"))\n\nxTravel = max(x1, x2) - min(x1, x2)\nyTravel = max(y1, y2) - min(y1, y2)\ntotalDistance = ((xTravel ** 2) + (yTravel) ** 2) ** 0.5\n\nprint(\"The total distance between the two points is:\", totalDistance)","sub_path":"Exercise2.14.py","file_name":"Exercise2.14.py","file_ext":"py","file_size_in_byte":489,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"} +{"seq_id":"265282053","text":"from flask import Flask, render_template, request, redirect\r\nimport queueManager, inputCheck, flags\r\nclass ServerManager:\r\n app = Flask(__name__)\r\n\r\n # def __init__(self):\r\n # self.app = Flask(__name__)\r\n # self.qManager = queueManager.QueueManager(queueManager.QueueManager.dbPath)\r\n\r\n def start(self):\r\n self.app.run(host=\"0.0.0.0\")\r\n\r\n @app.route(\"/\") \r\n def index():\r\n return render_template(\"index.html\")\r\n \r\n @app.route(\"/admin\")\r\n def admin():\r\n return render_template(\"admin.html\")\r\n\r\n @app.route('/submit', methods=[\"POST\"])\r\n def onSubmit():\r\n spotify_uri = ''\r\n youtube_url = str(request.form.get('youtube_url'))\r\n ctx_media = str(request.form.get('ctx_media'))\r\n ctx_button = str(request.form.get('ctx_button'))\r\n\r\n print(ctx_media + '-Link wurde per ' + ctx_button + ' submitted')\r\n\r\n if not spotify_uri == '':\r\n url = str(spotify_uri)\r\n ctx = flags.ctx_spotify\r\n else:\r\n url = str(youtube_url)\r\n ctx = flags.ctx_youtube\r\n \r\n if ctx_button == 'submit_normal':\r\n print('Url wird der Queue hinzugefügt')\r\n queueManager.QueueManager(queueManager.QueueManager.dbPath).add(url, ctx)\r\n elif ctx_button == 'submit_firstQ':\r\n print('Url wird an der ersten Stelle der Queue hinzugefügt')\r\n queueManager.QueueManager(queueManager.QueueManager.dbPath).addFirst(url, ctx)\r\n flags.setSkip(True)\r\n \r\n return redirect(\"/\")\r\n \r\n # Checken, ob es sich um einen Youtube oder Spotify Link handelt\r\n @app.route('/checkInput', methods=['POST'])\r\n def checkInput():\r\n spotify_uri = ''\r\n youtube_url = str(request.form.get('youtube_url'))\r\n _return = {'success': '', 'error': '', 'ctx_media': ''}\r\n\r\n if (not spotify_uri == '' and youtube_url == '') or (spotify_uri == '' and not youtube_url == ''):\r\n if inputCheck.checkSpotifyYoutube(spotify_uri) or inputCheck.checkSpotifyYoutube(youtube_url):\r\n _return['success'] = True\r\n\r\n if spotify_uri == '':\r\n _return['ctx_media'] = flags.ctx_youtube\r\n else:\r\n _return['ctx_media'] = flags.ctx_spotify\r\n else:\r\n _return['success'] = False\r\n _return['error'] = 'Eingabe war nicht korrekt! Bitte Eingabe prüfen'\r\n elif spotify_uri == '' and youtube_url == '':\r\n _return['success'] = False\r\n _return['error'] = 'Es muss mindestens ein Feld gefüllt werden'\r\n elif not spotify_uri == '' and not youtube_url == '':\r\n _return['success'] = False\r\n _return['error'] = 'Es darf nur ein Feld gefüllt werden'\r\n else:\r\n _return['success'] = False\r\n _return['error'] = 'Unbekannter Fehler bitte dem Benutzerservice bescheid geben ;)'\r\n \r\n if queueManager.QueueManager(queueManager.QueueManager.dbPath).getQueueLength() >= flags.max_queue_length:\r\n _return['success'] = False\r\n _return['error'] += '\\nDie Maximale Queue-Länge von ' + flags.max_queue_length + ' darf nicht überschritten werden'\r\n\r\n return _return\r\n \r\n # Admin Optionen übernehmen\r\n @app.route('/adminOptions', methods=['POST'])\r\n def setAdminOptions():\r\n ctx_button = str(request.form.get('ctx_button'))\r\n \r\n if ctx_button == 'skip':\r\n flags.doSkip = True\r\n elif ctx_button == 'playpause':\r\n if flags.doPause != True:\r\n flags.doPause = True\r\n else:\r\n flags.doPause = False","sub_path":"MediaPlayServer_noSpotify/serverManager.py","file_name":"serverManager.py","file_ext":"py","file_size_in_byte":3727,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"} +{"seq_id":"384585380","text":"from flask import Flask, request, jsonify\nfrom flask_cors import CORS\nimport json, os, data\n\napp = Flask(__name__)\nCORS(app)\n\n\n@app.route('/', methods=['GET'])\ndef hello():\n return \"
{}
'.format(render_wiki_file(\n wiki_file.id,\n wiki_file.name,\n file_type,\n tostring=True\n ))\n\n if upload_from_upload_page:\n wiki_page = get_object_or_404(\n WikiPage.select(\n WikiPage.id,\n WikiPage.markdown,\n WikiPage.current_version,\n WikiPage.modified_on),\n WikiPage.id==wiki_page_id\n )\n\n diff = make_patch(xstr(wiki_page.markdown), xstr(wiki_page.markdown)+file_markdown)\n WikiPageVersion.create(\n wiki_page=wiki_page,\n diff=diff,\n version=wiki_page.current_version,\n modified_on=wiki_page.modified_on\n )\n\n (WikiPageIndex\n .update(markdown=wiki_page.markdown+file_markdown)\n .where(WikiPageIndex.docid==wiki_page.id)\n .execute())\n\n (WikiPage\n .update(\n markdown=WikiPage.markdown+file_markdown,\n html=WikiPage.html+file_html,\n current_version=WikiPage.current_version+1,\n modified_on=datetime.utcnow())\n .where(WikiPage.id==wiki_page.id)\n .execute())\n\n return ''\n return file_markdown\n\n\n@blueprint.route('/reference/\\n')\n\t\tfp.write(\"==语音资料==\\n\")\n\t\tfp.write(\"'''''注:改造舰娘的语音只列出不重复的台词。'''''\\n\\n\")\n\t\tfp.write(\"\\n\".join(output))\n\t\tfp.write(\"\")\n\n\n","sub_path":"generater.py","file_name":"generater.py","file_ext":"py","file_size_in_byte":3363,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"}
+{"seq_id":"563218058","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\nimport os\nimport functools\nimport asyncio\n\n\nclass PingProtocol(asyncio.SubprocessProtocol):\n\n FD_NAMES = [\"stdin\", \"stdout\", \"stderr\"]\n\n def __init__(self, done_future: asyncio.Future):\n super().__init__()\n self.done = done_future\n self.buffer = bytearray()\n\n def connection_made(self, transport):\n print(\"process started {}, {}\".format(\n transport.get_pid(), type(transport)))\n self.transport = transport\n\n def pipe_data_received(self, fd, data):\n print(\"read {} bytes from {}\".format(len(data), self.FD_NAMES[fd]))\n if fd == 1:\n self.buffer.extend(data)\n\n def process_exited(self):\n print(\"process exited\")\n return_code = self.transport.get_returncode()\n print(\"return code {}\".format(return_code))\n\n cmd_output = bytes(self.buffer).decode()\n self.done.set_result((return_code, cmd_output))\n\n\n@asyncio.coroutine\ndef run_ping(loop: asyncio.BaseEventLoop):\n print(\"in run_ping\")\n\n cmd_done = asyncio.Future(loop=loop)\n args = [\"www.baidu.com\"]\n if os.name == \"nt\":\n args.extend([\"-n\", \"1\"])\n elif os.name == \"posix\":\n args.extend([\"-c\", \"1\"])\n factory = functools.partial(PingProtocol, cmd_done)\n proc = loop.subprocess_exec(\n factory,\n \"ping\", *args,\n stdin=None,\n stderr=None)\n try:\n print(\"launching process\")\n transport, protocol = yield from proc\n print(\"waiting for process to complete\")\n yield from cmd_done\n except Exception as ex:\n print(ex)\n raise\n else:\n transport.close()\n\n return cmd_done.result()\n\n\ndef main():\n event_loop = asyncio.get_event_loop()\n try:\n return_code, results = event_loop.run_until_complete(\n run_ping(event_loop))\n print(results)\n finally:\n event_loop.close()\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"standard/050.asyncio/use_subprocess/asyncio_subprocess_protocol.py","file_name":"asyncio_subprocess_protocol.py","file_ext":"py","file_size_in_byte":1966,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"}
+{"seq_id":"73055560","text":"#!/usr/bin/env python3\nimport sys, pathlib, unittest\nsys.path.append('../src')\nfrom directed_graph import DirectedGraph\nfrom vertex import Vertex\n\nclass TestVertex(unittest.TestCase):\n def test_illegal_edge(self):\n g = DirectedGraph()\n with self.assertRaises(ValueError):\n g.addEdge(2, Vertex('b'))\n g.addEdge(Vertex('b'), 2)\n g.adjacencies(2)\n\n def test_legal_edge(self):\n g = DirectedGraph()\n g.addEdge(Vertex('b'), Vertex('c'))\n\n def test_iteration(self):\n g = DirectedGraph()\n g.addEdge(Vertex('b'), Vertex('c'))\n for v in g:\n self.assertEqual(v, Vertex('b'))\n\n def test_adjacencies(self):\n g = DirectedGraph()\n g.addEdge(Vertex('b'), Vertex('c'))\n g.addEdge(Vertex('b'), Vertex('d'))\n self.assertEqual(len(g.adjacencies(Vertex('b'))), 2)\n self.assertEqual(len(g.adjacencies(Vertex('c'))), 0)\n\n def test_amountOfVertices(self):\n g = DirectedGraph()\n self.assertEqual(g.amountOfVertices(), 0)\n g.addEdge(Vertex('b'), Vertex('c'))\n self.assertEqual(g.amountOfVertices(), 2)\n g.addEdge(Vertex('b'), Vertex('d'))\n self.assertEqual(g.amountOfVertices(), 3)\n\nif __name__ == '__main__':\n unittest.main()\n","sub_path":"problem_016_find_mother_vertex_dfs/tests/directed_graph_tests.py","file_name":"directed_graph_tests.py","file_ext":"py","file_size_in_byte":1181,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"}
+{"seq_id":"578490749","text":"#coding=utf-8\n'''\n计算器手机自动化\n'''\nfrom appium import webdriver\n#1.获取手机信息\nimport time\n\ndesired_cpas={\n 'platformName' : 'Android',\n 'platformVersion' : '4.4.4',\n 'deviceName': '192.168.56.101:5555',\n 'appPackage' : 'com.android.calculator2',\n 'appActivity' : '.Calculator'\n}\n'''\n#android setting 中查看 一下具体内容\n#a。平台名称\n#desired_caps['platformName']='Android'\ndesired_cpas['platformName']='Android'\n#b.android 版本\ndesired_cpas['platformVersion']='4.4.4'\n#c.设备名称 win+r dos窗口 输入命令 adb devices\ndesired_cpas['deviceName']='192.168.56.101:5555'\n#d.存放的包名 adb shell--进入手机内部 进来如果不是root用户 我们先切换到root用户,通过 su root进行切换\n# cd /data/data -----ls 查看要找的包名;如计算器的包名为:com.android.calculator2\ndesired_cpas['appPackage']='com.android.calculator2'\n#e.存放Activity名称 .Calculator\ndesired_cpas['appActivity']='.Calculator'\n'''\n#2>启动appium, 将手机信息导入 ip+端口号+appium固定的路径\ndriver=webdriver.Remote('http://127.0.0.1:4723/wd/hub',desired_cpas)\ntime.sleep(3)\n#计算 2+5=7\n#借助SDK中的具体路径 D:\\adt-bundle-windows-sdk\\sdk\\tools\n#com.android.calculator2:id/digit2 content_desc /text 都可以用by_name\n#清除结果\ndriver.find_element_by_id(\"com.android.calculator2:id/clear\").click()\n#计算\ndriver.find_element_by_id('com.android.calculator2:id/digit2').click()\ndriver.find_element_by_id('com.android.calculator2:id/plus').click()\ndriver.find_element_by_id('com.android.calculator2:id/digit5').click()\ndriver.find_element_by_id('com.android.calculator2:id/equal').click()\n#结果 class class_name\nresult=driver.find_element_by_class_name(\"android.widget.EditText\").text\nprint(result)\nif int(result)==7:\n print(\"测试通过\")\nelse:\n print(\"测试不通过\")\n#清除结果\ndriver.find_element_by_id(\"com.android.calculator2:id/clear\").click()\n#关闭app\ndriver.quit()\n\n","sub_path":"weekend01/calc2.py","file_name":"calc2.py","file_ext":"py","file_size_in_byte":2006,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"}
+{"seq_id":"304641990","text":"from mongoengine import *\n\nfrom .audit import Audit\n\nclass Echange(Audit, EmbeddedDocument):\n accord = ReferenceField(\"Accord\", required = True)\n departements = ListField(ReferenceField(\"Departement\"), required = True)\n places = StringField()\n\n\n def get_departments_str(self):\n return \", \".join(d.nom for d in self.departements)\n\n\n def get_summary_str(self):\n dpts = self.get_departments_str()\n return \"{} ({}): {} places\".format(self.accord.nom, dpts, self.places)\n","sub_path":"dao/echange.py","file_name":"echange.py","file_ext":"py","file_size_in_byte":502,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"}
+{"seq_id":"390676670","text":"#!/usr/bin/env python\n\nimport threading, Irc, IrcBuilder\n\ndef join_greeter(bot, network, message):\n if message['command'] == 'JOIN':\n print(message)\n name = message['prefix']['name']\n if name != bot.nickname:\n channel = message['params']['channel'][1:]\n text = \"Hello \" + name + \"!\"\n bot.send(network, IrcBuilder.privmsg([channel], text))\n\nclass MessageListener(threading.Thread):\n def __init__(self, bot):\n threading.Thread.__init__(self)\n self.bot = bot\n def run(self):\n while True:\n message = self.bot.messages.get()\n if not message:\n break\n for handler in self.bot.message_handlers:\n handler(bot, message[0], message[1])\n if message[1]['command'] == 'PRIVMSG':\n pass\n\n\n\nclass Bot(Irc.Client):\n def __init__(self, nickname, userinfo = None, prefix = '.'):\n Irc.Client.__init__(self, nickname, userinfo)\n self.prefix = prefix\n self.message_handlers = []\n self.command_handlers = {}\n self.message_listener = MessageListener(self)\n def start(self):\n self.message_listener.start()\n def stop(self):\n self.messages.put(False)\n self.message_listener.join()\n print('Listener Thread died')\n def add_message_hander(self, handler):\n self.message_handlers.append(handler)\n def add_command_handler(self, command, handler):\n self.command_handlers[command] = handler\n\n\ndef main():\n pass\n\nif __name__ == '__main__':\n main()\n\nbot = Bot('Faroosh')\nbot.add_network('IRCHighway', 'irc.irchighway.net')\nbot.connect('IRCHighway')\nbot.send('IRCHighway', IrcBuilder.join(['#kiss']))\nbot.send('IRCHighway', IrcBuilder.privmsg(['#kiss'], 'Alright!'))\nbot.add_message_hander(join_greeter)\nbot.start()\n","sub_path":"Bot.py","file_name":"Bot.py","file_ext":"py","file_size_in_byte":1848,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"}
+{"seq_id":"91434553","text":"import torch\nimport torch.utils.data\nimport copy\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torch.optim as optim\nimport torch.autograd as autograd\nimport torch.backends.cudnn as cudnn\nimport numpy as np\nimport os\nimport csv\nimport time\nimport gzip\nimport scipy.optimize\nimport curve_fit\nfrom Logger import SummaryPrint\nfrom misc import bcolors\nimport misc\nimport matplotlib.pyplot as plt\nfrom DataLoaders import PFPSampler\n\nclass Wrapper(object):\n def __init__(self, args, network_class, data_loader, device,auto,num_net = 1):\n super(Wrapper, self).__init__()\n self.args = args\n self.device = device\n self.window_size = args.window_size\n self.ctr = 0\n self.nfp = 0\n self.epoch = 0\n self.num_norm = 0\n self.num_ana = 0\n self.norm_limit = 2000\n self.ana_limit = 2000\n self.norm_error = []\n self.ana_error = []\n self.auto=auto\n self.fit_curve = args.fit_curve\n self.gen_train = args.ryu_datagen_train\n self.class_conv = args.class_conv\n self.num_net = args.multi_net\n self.network_list = []\n self.sumPrint_list = []\n self.validSumPrint_list = []\n self.networks_trained = []\n self.ryu_test = False\n print(auto,auto,auto,auto)\n self.ckpt_dir = 'ckpts/{0}'.format(args.run_name)\n if not os.path.isdir('ckpts'): #this is making the ckpts folder\n os.mkdir('ckpts')\n\n if not os.path.isdir(self.ckpt_dir): #this is making the run ckpt folder\n os.makedirs(self.ckpt_dir)\n\n print('Creating Network')\n self.data_loader = data_loader\n print('data loader: ', data_loader)\n x,y = next(iter(self.data_loader))\n x_size = x.size()\n if self.num_net > 1: #if parameter is defined, will now generate a list of networks\n\n self.network = network_class(args,x_size)\n self.network_list.append(self.network)\n self.sumPrint = SummaryPrint(args,self.network.loss_names(), self.ckpt_dir, 'train' if not args.test else 'test')\n self.sumPrint_list.append(self.sumPrint)\n self.validSumPrint_list.append(SummaryPrint(args,self.network.loss_names(),self.ckpt_dir, 'valid', color=bcolors.OKBLUE))\n self.validSumPrint = self.validSumPrint_list[0]\n i = 1\n while i < self.num_net:\n tempnet = network_class(args,x_size)\n self.network_list.append(tempnet) #not sure if i can prematurely send all of them to device\n self.sumPrint_list.append(SummaryPrint(args,self.network_list[i].loss_names(),self.ckpt_dir,'train' if not args.test else 'test'))\n self.validSumPrint_list.append(SummaryPrint(args,self.network_list[i].loss_names(), self.ckpt_dir, 'valid', color=bcolors.OKBLUE))\n i += 1\n #send all the networks to the device\n for i in range(len(self.network_list)):\n self.network_list[i] = self.network_list[i].to(device)\n else:\n self.network = network_class(args, x_size)\n if args.print_network:\n print(self.network)\n #will need to make this into a list\n self.sumPrint = SummaryPrint(args,self.network.loss_names(), self.ckpt_dir, 'train' if not args.test else 'test')\n #temporarily disable validation for multi-network network\n if not args.test:\n self.validSumPrint = SummaryPrint(args,self.network.loss_names(), self.ckpt_dir, 'valid', color=bcolors.OKBLUE)\n\n print(bcolors.OKBLUE + 'Moving to specified device' + bcolors.ENDC)\n self.network = self.network.to(device)\n cudnn.benchmark = True\n if self.args.rmsprop:\n self.optimizer = optim.RMSprop(self.network.parameters(), lr=args.lr, weight_decay=self.args.l2_reg)\n elif self.args.sgd:\n self.optimizer = optim.SGD(self.network.parameters(), lr=args.lr, momentum=0.9, weight_decay=self.args.l2_reg)\n else:\n self.optimizer = optim.Adam(self.network.parameters(), lr=args.lr, weight_decay=self.args.l2_reg)\n \n\n def load(self, resume=False):\n if self.args.checkpoint is None:\n print(bcolors.OKBLUE + 'Loading Checkpoint: ' + self.args.run_name + bcolors.ENDC)\n checkpoint = torch.load(self.ckpt_dir+\"/ckpt.pth\")\n else:\n print(bcolors.OKBLUE + 'Loading Checkpoint: ' + self.args.checkpoint + bcolors.ENDC)\n checkpoint = torch.load(\"ckpts/%s/ckpt.pth\" % self.args.checkpoint)\n self.network.load_state_dict(checkpoint['network'])\n if resume:\n self.optimizer.load_state_dict(checkpoint['opt'])\n self.epoch = checkpoint['epoch']\n print(bcolors.OKBLUE + 'Finished Loading Checkpoint ' + bcolors.ENDC)\n\n def save(self):\n print(bcolors.OKBLUE + 'Saving Checkpoint: ' + self.args.run_name + bcolors.ENDC)\n torch.save({\n 'network':self.network.state_dict(),\n 'opt':self.optimizer.state_dict(),\n 'epoch':self.epoch+1,\n 'args':self.args,\n }, self.ckpt_dir+\"/ckpt.pth\")\n\n def _iter(self, x, y, sumPrint, backwards=True):\n x = x[:,0,:]\n x = x.view(x.shape[0],1,x.shape[1])\n x, y = x.to(self.device), y.to(self.device)\n y_bar = self.network(x)\n loss_l = self.network.loss(x,y,y_bar)\n if backwards == False and self.args.ryu_testing == True:\n batch = 0\n batch_test = []\n while batch < 512:\n label = y[batch].item()\n\n if label == 2 and self.num_norm > self.norm_limit:\n return [l.data.item() for l in loss_l]\n if label == 1 and self.num_ana > self.ana_limit:\n return [l.data.item() for l in loss_l]\n truth = x[batch,:,:].cpu().numpy()\n batch_test.append(truth)\n\n guess = y_bar[batch,:,:].detach().cpu().numpy()\n l2 = np.linalg.norm(truth-guess)\n\n if label == 2:\n #print('**************', label)\n #print('**************', l2)\n self.norm_error.append(l2)\n self.num_norm += 1 \n elif label == 1:\n print(\"-------------\", l2)\n self.ana_error.append(l2)\n self.num_ana += 1\n batch += 1\n print(\"\")\n \"\"\"for i in range(len(batch_test)):\n j = i\n while j < len(batch_test):\n #print(batch_test[i])\n print(\"Norm error between: \", i, j, \" is\", np.linalg.norm(batch_test[i]-batch_test[j]))\n j+= 1\n exit(1)\"\"\"\n if backwards:\n loss_l = self.network.loss(x,y,y_bar)\n self.optimizer.zero_grad()\n loss_l[0].backward()\n self.optimizer.step()\n return [l.data.item() for l in loss_l]\n\n def gen_training_testing(self, x, y,training=True): #LEFT OFF HERES\n x_copy = torch.Tensor.numpy(x)\n x_copy = x_copy.tolist()\n i = 0\n while(i < len(x_copy)):\n sample = x_copy[i]\n truth = y[i].item()\n \n strii = ','.join(str(i) for i in sample) + ',' + str(truth) + '\\n'\n #print(truth)\n wrote = False\n if training:\n if (self.num_norm < self.norm_limit and truth == 0) or (truth != 0 and self.num_ana < self.ana_limit):\n with open('./trainingData.txt.gz','ab') as f:\n f.write(strii)\n wrote = True\n else:\n if (self.num_norm < self.norm_limit and truth == 0) or (truth != 0 and self.num_ana < self.ana_limit):\n with open('./testingData.txt.gz','ab') as f:\n f.write(strii)\n wrote = True\n i += 1\n if wrote == True:\n if truth == 0:\n self.num_norm += 1\n else:\n self.num_ana += 1\n #print(self.num_ana, self.num_norm)\n if self.num_norm > self.norm_limit and self.num_ana > self.ana_limit:\n break\n return 0\n\n # def _ryu_iter(self, x, y, sumPrint, backwards=True):\n # x_copy = torch.Tensor.numpy(x)\n # x_copy = x_copy.tolist()\n # #print(\"---------------\", len(x_copy))\n #\n # x,y = x.to(self.device), y.to(self.device)\n # y_bar = self.network(x)\n # loss_l = self.network.loss(x, y, y_bar)\n # if backwards:\n # self.optimizer.zero_grad()\n # loss_l[0].backward()\n # self.optimizer.step()\n # with open('./trainingData.txt.gz','ab') as f:\n # f.write(','.join(str(i) for i in x_copy)+','+str(y)+'\\n')\n # else:\n # with open('./testingData.txt.gz','ab') as f:\n # f.write(','.join(str(i) for i in x_copy)+','+str(y)+'\\n')\n # return [l.data.item() for l in loss_l]\n\n def run_epoch(self, data_loader, test=False, ryu_test=False):\n data_loader=self.data_loader\n #data_loader.switch_train((not test) and (self.auto))\n self.sumPrint.start_epoch(self.epoch, len(data_loader))\n for j, (data, target) in enumerate(data_loader):\n self.sumPrint.start_iter(j)\n res = self._iter(data, target, self.sumPrint, backwards=not test)\n self.sumPrint.end_iter(j, res)\n \n rets = self.sumPrint.end_epoch()\n self.ctr+=1\n #print(data_loader.dataset.list_of_training_files)\n #exit(1)\n if not test: \n data_loader.switch_train(test)\n self.network.eval()\n self.validSumPrint.start_epoch(self.epoch, len(data_loader))\n for j, (data, target) in enumerate(data_loader):\n self.validSumPrint.start_iter(j)\n res = self._iter(data, target, self.validSumPrint, backwards=False)\n self.validSumPrint.end_iter(j, res,weights=[1.0,1.0,res[1]/100.0,(100.0-res[1])/100])\n self.network.train()\n\n val_rets = self.validSumPrint.end_epoch()\n else:\n val_rets = None\n\n return rets, val_rets\n\n\n # def ryu_testing(self,val=False,test=True):\n # #print(\"in ryu_testing\")\n # #data_loader.switch_train((not test) and (self.auto))\n # #print(data_loader.dataset.train)\n # #exit(1)\n # self.sumPrint.start_epoch(self.epoch, len(self.data_loader))\n # for j, (data, target) in enumerate(self.data_loader):\n #\n # self.sumPrint.start_iter(j)\n # res = 0\n # if val == False:\n # #print(\"calling iter\")\n # res = self._iter(data, target, self.sumPrint, backwards=False)\n # self.sumPrint.end_iter(j, res)\n #\n # rets = self.sumPrint.end_epoch()\n # val_rets = None\n # return rets, val_rets\n\n def test(self, load=True):\n #testing\n print(bcolors.OKBLUE+'*******TESTING********'+bcolors.ENDC)\n data_loader = PFPSampler(self.args, train=False)\n #load checkpoint\n if load:\n self.load()\n #set no gradients\n self.network.eval()\n #run epoch\n \n rets, _ = self.run_epoch(data_loader, True)\n rets = [self.args.run_name] + rets #run name\n return rets\n\n # def ryu_test_procedure(self, load = True):\n # print(bcolors.OKBLUE+'*******TESTING********'+bcolors.ENDC)\n # self.load()\n # self.network.eval()\n # while(self.num_norm < self.norm_limit or self.num_ana < self.ana_limit):\n # rets, _ = self.ryu_testing(False,not self.data_loader.dataset.train)\n # print(self.num_norm, \" || \", self.num_ana)\n #\n # print(len(self.norm_error))\n # print(len(self.ana_error))\n #\n # #fit gaussian curve\n # if self.fit_curve == True:\n # p0 = [1., -1., 1., 1., -1., 1.]\n # bin_centres = (min(self.norm_error) + max(self.norm_error))/2\n # coeff, var_matrix = curve_fit(gaussx, bin_centres, self.norm_error, p0=p0)\n # hist_fit = gauss\n #\n #\n # rets = [self.args.run_name] + rets #run name\n # g_min = min(min(self.ana_error),min(self.norm_error))\n # g_max = max(max(self.ana_error),max(self.norm_error))\n # bins = np.linspace(g_min,g_max,100)\n # plt.hist(self.norm_error,bins,alpha=0.5,label=\"Norm\")\n # plt.hist(self.ana_error,bins,alpha=0.5,label=\"Ano\")\n # plt.legend(loc='upper right')\n # plt.savefig((str)(self.args.run_name)+\"-ANO vs NORM errors.png\")\n # return rets\n\n\n def train(self):\n x = 100\n for epoch in range(x): # loop over the dataset multiple times\n\n running_loss = 0.0\n for i, data in enumerate(PFPSampler, 0):\n # get the inputs; data is a list of [inputs, labels]\n inputs, labels = data\n\n # zero the parameter gradients\n self.optimizer.zero_grad()\n\n # forward + backward + optimize\n outputs = self.network(inputs)\n loss = criterion(outputs, labels)\n loss.backward()\n self.optimizer.step()\n\n # print statistics\n running_loss += loss.item()\n if i % 2000 == 1999: # print every 2000 mini-batches\n print('[%d, %5d] loss: %.3f' %\n (epoch + 1, i + 1, running_loss / 2000))\n running_loss = 0.0\n\nprint('Finished Training')\n\ndef gaussx(x, *p):\n A1, mu1, sigma1= p\n return A1*np.exp(-(x-mu1)**2/(2.*sigma1**2))\n\n def train(self):\n if self.args.load_checkpoint or self.args.resume or (self.args.checkpoint is not None):\n self.load(self.args.resume)\n data_loader = self.data_loader\n\n print(bcolors.OKBLUE+'*******TRAINING********'+bcolors.ENDC)\n self.network.train()\n while self.epoch < self.args.epochs:\n _, val_ret = self.run_epoch(data_loader, False)\n self.epoch += 1\n if self.epoch % self.args.checkpoint_every == 0:\n print(\"Saving...\")\n self.save()\n self.save()\n\n # def ryu_data_gen(self):\n # if self.args.load_checkpoint or self.args.resume or (self.args.checkpoint is not None):\n # self.load(self.args.resume)\n # data_loader = PFPSampler(self.args,train=True)\n # #print(bcolors.OKBLUE+'*******TRAINING*******'+bccolors.ENDC)\n # while self.num_norm < self.norm_limit and self.num_ana < self.ana_limit:\n # _, val_ret = self.ryu_run_epoch(data_loader, False)\n # #print('Normal: ', self.num_norm)\n # #print(\"Ana: \", self.num_ana)\n","sub_path":"apriori_characterization/Wrapper.py","file_name":"Wrapper.py","file_ext":"py","file_size_in_byte":14999,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"}
+{"seq_id":"192239363","text":"'''3. Write a program that accepts sequence of lines as input and prints the\r\n lines after making all characters in the sentence capitalized. Suppose the\r\n following input is supplied to the program:\r\n Hello world\r\n Practice makes perfect\r\n Then, the output should be:\r\n HELLO WORLD\r\n PRACTICE MAKES PERFECT [Marks: 3]'''\r\n\r\n\r\nlines = []\r\nwhile True:\r\n x = input()\r\n if x:\r\n lines.append(x.upper())\r\n else:\r\n break;\r\nfor x in lines:\r\n print(x)\r\n\r\n \r\n'''WHILE TRUE MEANS LOOP FOREEVER'''\r\n","sub_path":"Assignment 04 28th may/3.capitalized.py","file_name":"3.capitalized.py","file_ext":"py","file_size_in_byte":556,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"}
+{"seq_id":"60203261","text":"#Module 6 Week 4 Sprint\n\nimport json\nimport pymongo as pm\nimport pandas as pd \n\nclient = pm.MongoClient(\"mongodb://localhost:27017/\")\nmyDb = client['Data_Tracker']\ncollection = myDb[\"DataVault\"]\nprint(collection)\n\nstockData = pd.read_csv('Inventory.csv')\nprint(stockData)\n\nmongoData = []\nn=1\ncol =stockData.columns\nfor i in range(len(stockData)):\n row = stockData.iloc[i:n,:]\n template = {}\n for c in col:\n value = row.get_value(index=i,col=c)\n try:\n value = float(value)\n except:\n x=0\n template[c]= value\n mongoData.append(template)\n n+=1\n\nprint(len(mongoData),mongoData)\n\nfor item in mongoData:\n #insert = collection.insert_one(item)\n print(item, insert,'\\n')\n\n#Clear documents in collection \ndef clearCollection(col):\n x = col.delete_many({})\n print(x.deleted_count,'documents deleted')\n\nfor info in collection.find().sort('Amount',-1):\n print(info)\n\n#Filtering to display a specific category\nsearchQuery = {'Category':{'$regex':'CHOCOLATES'}}\nchocData = collection.find(searchQuery).sort('Amount',-1)\nfor choc in chocData:\n print(choc)\n\n#Create collection of top 3 categories\ntopCollec = myDb['Top 3']\nprint(topCollec)\nfor stuff in mongoData:\n if stuff['Category'] in top3cats:\n some=0\n #insert = topCollec.insert_one(stuff)\n\ntop3data = topCollec.find().sort('Amount',1)\nfor d in top3data:\n print(d)\n\n#Deleting entries from top 3\nitemsToDel = ['Mutton_Curry','Squash', 'Twista']\nfor x in itemsToDel:\n del_query = {'Product':{'$regex': x}}\n deletedItems = topCollec.delete_many(del_query)\n print(deletedItems.deleted_count,'items delted')\n\n#Product to find an update\n#object id of item to update :{'_id': ObjectId('5e957c42492b0732c5bd3644')}\nresult = collection.update_many({'Product': 'Fritos'},\n {'$inc': {'Amount': 3}}, upsert=True)\n\n# boolean confirmation that the API call went through\nprint (\"acknowledged:\", result.acknowledged)\n\n# integer of the number of docs modified\nprint (\"number of docs updated:\", result.modified_count)\n\n# dict object with more info on API call\nprint (\"raw_result:\", result.raw_result)\n\n#Search and filter for the least 5 items \n\nbottom3 = collection.find()\ncount=0\nwhile count<3:\n for b in bottom3:\n if (b['Amount']<10):\n print(b)\n count+=1\n else:\n continue\n","sub_path":"Sprint.py","file_name":"Sprint.py","file_ext":"py","file_size_in_byte":2389,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"}
+{"seq_id":"384825916","text":"import pandas as pd\nfrom rdkit import Chem\nfrom rdkit.Chem.SaltRemover import SaltRemover\nremover = SaltRemover()\nfrom rdkit import RDLogger\nlg = RDLogger.logger()\nlg.setLevel(RDLogger.ERROR)\n\ndef convert_to_rdkit(smi):\n try:\n mol = Chem.MolFromSmiles(smi)\n mol = remove_salt(mol)\n new_smi = Chem.MolToSmiles(mol)\n return new_smi\n except:\n print(f\"{smi} not accepted by rdkit\")\n return None\n\ndef remove_salt(mol):\n try:\n mol = remover.StripMol(mol, dontRemoveEverything=True)\n except:\n pass\n return mol\n\ndef remove_duplicate_smiles(df):\n df = df.drop_duplicates(subset=['SMILES'])\n return df\n\ndef create_mol_col(smi):\n try:\n return Chem.MolFromSmiles(smi)\n except:\n print(f'error with smi {smi}')\n return None\n\ndef get_inchi_key(mol):\n try:\n return Chem.MolToInchiKey(mol)\n except:\n return None\n\ndef load_mol_columns(df):\n df['mol'] = df['SMILES'].apply(create_mol_col)\n df = df.dropna(subset=['mol'])\n df['rdmol'] = df['mol'].apply(mol_binary)\n df = df.dropna(subset=['rdmol'])\n df['inchi_key'] = df['mol'].apply(get_inchi_key)\n df = df.dropna(subset=['inchi_key'])\n\n df = df.where(pd.notnull(df), None)\n\n return df\n\ndef nan_to_none(df):\n df = df.where(pd.notnull(df), None)\n return df\n\ndef create_merged_df(paths):\n list_dfs = load_dfs(paths)\n merged_df = merge_dfs(list_dfs)\n merged_df = nan_to_none(merged_df)\n return merged_df\n\ndef load_dfs(paths):\n print(f'loading {len(paths)} dataframes')\n list_dfs = []\n for path in paths:\n df = pd.read_csv(path, index_col=0)\n list_dfs.append(df)\n return list_dfs\n\ndef merge_dfs(list_dfs):\n print('merging dataframes')\n all_df = list_dfs.pop(0)\n for df in list_dfs:\n all_df = all_df.merge(df, on='SMILES', how='outer')\n return all_df","sub_path":"buyable_molecules/create_csvs/funcs/dataframe_functions.py","file_name":"dataframe_functions.py","file_ext":"py","file_size_in_byte":1887,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"}
+{"seq_id":"509172764","text":"#celsius to farenhieh coversion table\nc=list(range(0,101,10))\nf=list()\nprint(\"celsius list: \" ,c )\nfor i in range(len(c)):\n far=9/5*c[i]+32\n f.append(far)\nprint(\"\\tcelsius\\tfahrenheit\")\nprint(\"------------------------\")\nfor i in range(len(c)):\n print(\"\\t\",c[i],\"\\t\",f[i])\n \n","sub_path":"pythonprog/celsiustofahrenheit.py","file_name":"celsiustofahrenheit.py","file_ext":"py","file_size_in_byte":286,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"}
+{"seq_id":"499504417","text":"# ====================================================================\n# This code will load the results/outputs from 01_1.py and produce the\n# plots for the paper\n# ====================================================================\n\nimport numpy as np\nimport bbi\nfrom matplotlib import pyplot as plt\nimport pickle\nplt.rcParams.update({'font.size': 14})\n\ndef harmonic_mean(error, axis=0):\n return np.exp(np.mean( np.log(error), axis=axis))\n\n# general: load reference likelihood and data needed in all plots\ncontent = np.load('input/01_model.npz')\ngrid = content['grid']\n\nll_true = np.load('output/01_reference_ll.npy')\n\n# load result of computation step 1)\ncontent = np.load('output/01_main_ll.npz', allow_pickle = True)\nn_eval = content['n_eval']\nll_m = content['ll_m']\nll_l = content['ll_l']\nll_a = content['ll_a']\n\nnodes_m = content['nodes_m'].item()\nnodes_l = content['nodes_l'].item()\nnodes_a = content['nodes_a'].item()\n\ncontent = pickle.load( open('output/01_pre_nodes_f.pkl', 'rb')) \nnodes_lhs = content[2][0] # first index = 2 means third run which is initial sample = 15\n\ncontent = np.load('output/01_miracle.npz', allow_pickle = True)\nll_1 = content['ll_1']\n\n# compute errors from loglikelihoods\nerrors_m = bbi.compute_errors(ll_true, ll_m)\nerrors_l = bbi.compute_errors(ll_true, ll_l)\nerrors_a = bbi.compute_errors(ll_true, ll_a)\nerrors_1 = bbi.compute_errors(ll_true, ll_1)\n\nnp.savetxt('output/exp1_nodes_m.txt', grid[nodes_m.idx], fmt= '%g', delimiter = ' ')\nnp.savetxt('output/exp1_nodes_l.txt', grid[nodes_l.idx], fmt= '%g', delimiter = ' ')\nnp.savetxt('output/exp1_nodes_a.txt', grid[nodes_a.idx], fmt= '%g', delimiter = ' ')\nnp.savetxt('output/exp1_nodes_lhs15.txt', grid[nodes_lhs.idx], fmt= '%g', delimiter = ' ')\n\nf = plt.figure(figsize=(16,16))\nx = np.linspace(0,1,51)\nimg = plt.contourf(x,x,np.exp(ll_true).reshape(51,51))\n\nimg.set_cmap('Blues')\n\n#f.savefig(\"figures/01_fig_0_solution.pdf\", bbox_inches='tight')\n\nplt.axis('off')\n#plt.savefig(\"figures/01_fig_0_solution.png\", bbox_inches='tight')\n\n# %% Plot 1: Comparison of new methods with random picking\n\nplt.rcParams.update({'font.size': 14})\n# load results of 20 random gpes\ncontent = np.load('output/01_20_random_matern.npz', allow_pickle = True)\nnodes_gpe = content['nodes_gpe']\n#ll_gpe = content['ll_gpe']\nerrors_gpe = content['errors_gpe']\n\nerror_average = np.exp(np.mean( np.log(errors_gpe), axis=0))\ne_mean_random = harmonic_mean(errors_gpe)\ne_0_random = errors_gpe.min(axis=0)\ne_100_random = errors_gpe.max(axis=0)\n\n#error_sorted = np.sort(errors_gpe, axis = 0)\n#error_0 = error_sorted[0,:]\n#error_5 = error_sorted[1,:]\n#error_95 = error_sorted[19,:]\n#error_100 = error_sorted[-1,:]\n\n#error_50 = error_sorted[10,:]\n \n# plot\nf = plt.figure(figsize=(12,8))\nplt.xlim(0,30)\nplt.ylim(1e-6,1e2)\nplt.xlabel('Number of model evaluations')\nplt.ylabel('Error (KL-divergence)')\n\nplt.semilogy(n_eval, errors_m, label = 'dynamic MAP estimate',linewidth = 2.5)\nplt.semilogy(n_eval, errors_a, label = 'average criterion',linewidth = 2.5)\nplt.semilogy(n_eval, errors_l, label = 'linearization',linewidth = 2.5)\nplt.semilogy(n_eval, errors_1, label = 'miracle',linewidth = 2.5)\n\nplt.fill_between(n_eval, e_0_random, e_100_random, label = '5% to 95% percentiles', color = 'lightgray')\nplt.semilogy(n_eval, e_mean_random, label = 'mean of random hyper parameters', color = 'black',linewidth = 2.5)\n\nplt.legend(loc=3, frameon = False).set_zorder(-1)\n\nplt.show()\n\n#f.savefig(\"figures/01_fig_1_random.pdf\", bbox_inches='tight')\n\nplot_data = np.full((31,8), np.nan)\nplot_data[:,0] = n_eval\n\nplot_data[:,1] = errors_m\nplot_data[:,2] = errors_a\nplot_data[:,3] = errors_l\nplot_data[:,4] = errors_1\nplot_data[:,5] = e_mean_random\nplot_data[:,6] = e_0_random\nplot_data[:,7] = e_100_random\n\nnp.savetxt('output/exp1_fig1.data', plot_data, '%2i %1.3e %1.3e %1.3e %1.3e %1.3e %1.3e %1.3e')\n\n#%% Plot 2: Comparison with separate sampling phase approach\n\n\nplt.rcParams.update({'font.size': 14})\n# load data\n\nn_eval_pre = pickle.load( open('output/01_pre50_lhs_n_eval.pkl','rb'))\nerrors_raw_fix = pickle.load( open('output/01_pre_errors_f.pkl','rb'))\nerrors_raw_re = pickle.load( open('output/01_pre_errors_r.pkl','rb'))\n\ne_mean_fix = []\ne_0_fix = []\ne_100_fix = []\ne_mean_re = []\ne_0_re = []\ne_100_re = []\n \nfor e in errors_raw_fix:\n e_mean_fix.append(harmonic_mean(e))\n e_0_fix.append(e.min(axis=0))\n e_100_fix.append(e.max(axis=0))\n \nfor e in errors_raw_re:\n e_mean_re.append(harmonic_mean(e))\n e_0_re.append(e.min(axis=0))\n e_100_re.append(e.max(axis=0))\n\nf = plt.figure(figsize=(12,8))\nplt.xlabel('Number of model evaluations')\nplt.ylabel('Error (KL-divergence)')\n\nplt.semilogy(n_eval, errors_m, label = 'dynamic MAP estimate',linewidth = 1)\nplt.semilogy(n_eval, errors_a, label = 'average criterion',linewidth = 1)\nplt.semilogy(n_eval, errors_l, label = 'linearization',linewidth = 1)\n\nfor i, (e, n) in enumerate(zip(e_mean_fix,n_eval_pre)):\n if i == 0:\n plt.semilogy(n, e, 'k', label = 'exploratory phase, fixed hyper parameters',linewidth = 2.5) \n else:\n plt.semilogy(n, e, 'k',linewidth = 2.5) \n\nfor i, (e, n) in enumerate(zip(e_mean_re,n_eval_pre)):\n if i == 0:\n plt.semilogy(n, e, 'C3', label = 'exploratory phase, re-estimate',linewidth = 2.5)\n else:\n plt.semilogy(n, e, 'C3',linewidth = 2.5)\n \n\nerror_new = np.array([errors_m, errors_l, errors_a])\nerror_new = np.sort(error_new, axis = 0)\nerror_upper = error_new[0,:]\nerror_lower = error_new[2,:]\n\nplt.xlim(0,30)\nplt.legend(loc=3, frameon = False).set_zorder(-1)\nplt.show()\n\n#f.savefig(\"figures/01_fig_2_presampled.pdf\", bbox_inches='tight')\n\n#for (e1,e2,n) in zip(e_0_fix, e_mean_re, n_eval_pre):\nfor i,n in enumerate(n_eval_pre):\n prefix = n[0]\n filename = \"output/exp1_fig2_{}.data\".format(prefix)\n plot_data = np.full((n.size, 7), np.nan)\n plot_data[:,0] = n\n plot_data[:,1] = e_mean_fix[i]\n plot_data[:,2] = e_0_fix[i]\n plot_data[:,3] = e_100_fix[i]\n plot_data[:,4] = e_mean_re[i]\n plot_data[:,5] = e_0_re[i]\n plot_data[:,6] = e_100_re[i]\n np.savetxt(filename, plot_data, '%2i %1.3e %1.3e %1.3e %1.3e %1.3e %1.3e')\n\n#%% Plot 3: Investigate sensitivity to field parameter prior\n\nll_true = np.load('output/01_reference_ll.npy', allow_pickle = True)\n\ncontent = np.load('output/01_main_ll.npz', allow_pickle = True)\nn_eval = content['n_eval']\nll_m = content['ll_m']\n\ncontent = np.load('output/01_prior_sensitivity_ll.npz', allow_pickle = True)\nll_p1 = content['ll_p1']\nll_p2 = content['ll_p2']\nll_p3 = content['ll_p3']\n\n# compute errors from loglikelihoods\nerrors_m = bbi.compute_errors(ll_true, ll_m)\nerrors_p1 = bbi.compute_errors(ll_true, ll_p1)\nerrors_p2 = bbi.compute_errors(ll_true, ll_p2)\nerrors_p3 = bbi.compute_errors(ll_true, ll_p3)\n\n# plot\nf = plt.figure(figsize=(12,8))\nplt.xlabel('Number of model evaluations')\nplt.ylabel('Error (KL-divergence)')\nplt.xlim(0,30)\nplt.semilogy(n_eval, errors_m, label = 'default',linewidth = 2.5)\n\nplt.semilogy(n_eval, errors_p1, label = 'wide upper and lower',linewidth = 2.5)\nplt.semilogy(n_eval, errors_p2, label = 'wide upper',linewidth = 2.5)\nplt.semilogy(n_eval, errors_p3, label = 'narrow',linewidth = 2.5)\n\nplt.legend()\nplt.show()\n\n#f.savefig(\"figures/01_fig_3_sensitivity.pdf\", bbox_inches='tight')\n\nplot_data = np.full((31,5), np.nan)\nplot_data[:,0] = n_eval\n\nplot_data[:,1] = errors_m\nplot_data[:,2] = errors_p1\nplot_data[:,3] = errors_p2\nplot_data[:,4] = errors_p3\n\nnp.savetxt('output/exp1_fig3.data', plot_data, '%2i %1.3e %1.3e %1.3e %1.3e')\n","sub_path":"experiment 1/01_2_plot.py","file_name":"01_2_plot.py","file_ext":"py","file_size_in_byte":7544,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"}
+{"seq_id":"285743627","text":"# Given a path to data file. parse it into a map\nimport csv\nimport math\n\n\ndef parse_txt(path):\n with open(path) as csv_file:\n csv_reader = csv.reader(csv_file, delimiter='\\t')\n index = -1\n station_headers = []\n stations = dict()\n for row in csv_reader:\n index = index + 1\n # Ignore first line\n if index is 0:\n continue\n\n # Parse the locations\n if index is 1:\n for i in range(4, len(row), 4):\n station_headers.append(row[i])\n stations[row[i]] = {\n 'long': float(row[i + 1]),\n 'lat': float(row[i + 2]),\n 'site_id': row[i + 3] if i + 3 < len(row) else '',\n 'data': {}\n }\n if index > 3:\n header_index = 0\n for i in range(4, len(row), 4):\n stations[station_headers[header_index]]['data'][row[0]] = float(math.nan if row[i] == \"\" else row[i])\n\n header_index = header_index + 1\n return stations\n","sub_path":"Interpolation/csaparser.py","file_name":"csaparser.py","file_ext":"py","file_size_in_byte":1146,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"}
+{"seq_id":"403120956","text":"import matplotlib.pyplot as plt\nimport numpy as np\nfrom math import pi,sqrt\n\n\n#essai1 = \"Essai_01/0_90.Txt\"\nessai2 = \"Essai_02/0_90.Txt\"\nmodele1 = \"\"\n\nfid = open(essai2,\"r\")\nfid.readline()\nfid.readline()\nfichier = []\nfor ligne in fid :\n ligne = ligne.replace(\" \",\" \")\n ligne = ligne.replace(\" \",\" \")\n ligne = ligne.replace(\" \",\" \")\n ligne = ligne.replace(\" \",\" \")\n ligne = ligne.replace(\" \",\" \")\n ligne = ligne.replace(\" \",\" \")\n ligne = ligne.replace(\" \",\" \")\n ligne = ligne.replace(\"\\n\",\"\")\n ligne = ligne.split(\" \")\n if len(ligne)>1:\n for i in range (len(ligne)-1,-1,-1):\n if ligne[i] == \"\":\n del ligne[0]\n else : \n ligne[i] = float(ligne[i])\n fichier.append(ligne)\n\nfid.close()\n\ntemps =[] # temps en ms dans le fichier, indice 0\npos_bras=[] # position du bras en degrés, indice 2\ntx_bras=[] # taux de rotation du bras en rad/s, indice 5\ntx_mot=[] # taux de rotation du moteur en rad/s, indice 6\ngamma = []\nfor ligne in fichier:\n temps.append(ligne[0]/1000.)\n pos_bras.append(ligne[2]*180/pi)\n tx_bras.append(ligne[5])\n tx_mot.append(ligne[6])\n \ngamma=[0]\nfor i in range(len(tx_mot)-1):\n deltaT = (temps[i+1]-temps[i])\n gammatmp = gamma[i]+deltaT*tx_mot[i]\n gamma.append(gammatmp)\n\n\n\ngamma = np.array(gamma)\ndgamma = np.array(tx_mot)\n\n## Loi Entrée Sortie Théorique\na,b,c,d = 106.3, 59, 70, 80\np = 4\nlambda0 = sqrt(d*d+(b+c)**2)\n\ntps = np.linspace(0,4,1000)\ndgam = np.ones(1000)*40 # 40 rad/s\ngam = 40*tps\n\ngamma = gam\ndgamma = dgam\n\ndtheta = - ((lambda0-p*gamma/(2*pi))*((-p*dgamma)/(2*pi))/(a*b)) / ( np.sqrt(1- (((lambda0-(p*gamma)/(2*pi))**2-a*a-b*b)/(2*a*b))**2))\n\n\ntheta = np.arccos((((lambda0-p*gamma/(2*pi))**2)-a*a-b*b)/(2*a*b))-np.arctan(d/c)\ntheta = theta*360/(2*pi)\n\n\n\"\"\"\nfor i in range(len(gamma)):\n gam = gamma[i]\n dgam = tx_bras[i]\n pp = -(((np.sqrt(d*d+c*c+b*b+2*b*c)-p*gamma/(2*pi))*(-p*dgamma/(2*pi)))/(a*b))/(np.sqrt(1-((((np.sqrt(d*d+c*c+b*b+2*b*c)-p*gamma/(2*pi))**2)-a*a-b*b)/(2*a*b))**2))\n print(\n\"\"\"\n \n\n\n\n\n\n#plt.plot(temps,tx_mot,label = \"Moteur\")\n#plt.plot(temps,tx_bras,label = \"Bras\")\nplt.plot(tps,theta,label = \"Vitesse bras simulee\")\n\nplt.legend()\nplt.show()","sub_path":"11_Maxpid/EtudeMaxpid/images/LoiES_Essai_Modele/exploitationEssais.py","file_name":"exploitationEssais.py","file_ext":"py","file_size_in_byte":2255,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"}
+{"seq_id":"514215965","text":"import pandas as pd\nimport numpy as np\nimport csv\n\ndef getXY(filename):\n data = pd.read_csv(filename, sep=',',header=None)\n X = data[data.columns[0:29]].values[1:].astype(np.float)\n y = np.transpose([data[data.columns[30]].values[1:].astype(np.float)])\n return X, y\n\nif __name__ == \"__main__\":\n getXY('../dataset/250.csv')","sub_path":"scripts/getData.py","file_name":"getData.py","file_ext":"py","file_size_in_byte":337,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"}
+{"seq_id":"574224461","text":"#!/usr/bin/env python\nimport sys, os\nimport wx\n\nclass main_window(wx.Frame):\n def __init__(self, parent, id, title):\n wxFrame.__init__(self, parent, -1, title, size = (200, 100), style=wxDEFAULT_FRAME_STYLE|wxNO_FULL_REPAINT_ON_RESIZE)\n self.control = wxTextCtrl(self, -1, style=wxTE_MULTILINE)\n self.Show(true)\n\nclass App(wx.App):\n def OnInit(self):\n frame = main_window(None, -1, \"wxPython: (A Demonstration)\")\n self.SetTopWindow(frame)\n return true\n\napp = App(0)\napp.MainLoop()\n\n","sub_path":"wxpython/ex02.py","file_name":"ex02.py","file_ext":"py","file_size_in_byte":531,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"}
+{"seq_id":"530723415","text":"import re\n\nimport requests\n\nurl = \"https://es.wiktionary.org/wiki/Ap%C3%A9ndice:C%C3%B3digos_de_idioma\"\nwith requests.get(url) as req:\n req.raise_for_status()\n content = req.text\n\npattern = r\"':'',\r\n '
':'
{rawcode}
'\r\n # Print the code\r\n print(code)\r\n\r\n\r\n\r\n\r\n'''\r\n /‾/ /‾/‾‾‾‾‾‾‾‾/‾| /‾| /‾/\r\n / /___/ /‾‾‾/ /‾‾/ |/ | / /\r\n / ___ / / / / | | / |/ /\r\n / / / / / / / /|__/| | /____\r\n/_/ /_/ /_/ /_/_____|_|_____/\r\n /‾‾‾‾‾‾/‾‾‾‾‾\\|‾‾‾‾‾\\ | _____|\r\n| /‾‾‾| /‾\\ | |‾\\ \\| |___\r\n| | | | | | | | | ___|\r\n| \\___| \\_/ | |_/ /| |_____\r\n \\______\\_____/|_____/ |_______|\r\n'''\r\n\r\n\r\n# Head the function htmlCode\r\ndef html_code(code,inline = False, Return = False, color = False):\r\n if color != False:\r\n setColorScheme(color)\r\n global ColorToType\r\n ColorToType = {\r\n \"Border\" : Border,\r\n \"Background\" : Background,\r\n \"Foreground\" : Foreground,\r\n \"Comment\" : Comment,\r\n \"String\" : String,\r\n \"Keywords\" : Keywords,\r\n \"Builtins\" : Builtins,\r\n \"Definitions\" : Definitions\r\n }\r\n # Create a copy of code\r\n rawcode = code\r\n # Loop through the replacements\r\n for item in Replacers:\r\n # and replace\r\n rawcode = rawcode.replace(item,Replacers[item])\r\n # Create an empty highlighted code string\r\n HighlightedCode = ''\r\n\r\n # If the code does not end in a new line, make it\r\n if rawcode[-6:]!='
{HighlightedCode}
'\r\n # Print the code\r\n \r\n print(code)\r\n if Return == True:\r\n return code\r\n \r\n\r\n\r\n#htmlCode(input('Code Here:'))\r\n\r\nprint('\\n\\nInitiate using the \"htmlCode\" function\\n>>> htmlCode(\\'\\'\\'CODE GOES HERE\\'\\'\\') or >>> htmlCode(\\'\\'\\'CODE GOES HERE\\'\\'\\', inline = True)\\nPlease use triple quotes for multiline code\\n'+'-'*30+'\\nThe htmlCode(\\'\\'\\'CODE GOES HERE\\'\\'\\') function returns an object that can be previewed using the OBJECT.preview() command.\\n An immediate preview can be achieved by using the command htmlCode(\\'\\'\\'CODE GOES HERE\\'\\'\\').preview()\\n')\r\n\r\n\r\nclass htmlCode():\r\n \"\"\" ../‾‾‾‾‾‾/‾‾‾‾‾\\|‾‾‾���‾\\ | _____|.\r\n .| /‾‾‾| /‾\\ | |‾\\ \\| |___....\r\n .| | | | | | | | | ___|...\r\n .| \\___| \\_/ | |_/ /| |_____..\r\n ..\\______\\_____/|_____/ |_______|.\r\n \"\"\"\r\n def __init__(self, code, inline=False, color = False):\r\n print('⩋'*40+'\\n')\r\n if type(code)==tuple or type(code)==list:\r\n self.code = ''\r\n for section in code:\r\n if type(color)==tuple or type(color)==list:\r\n clr = color[code.index(section)]\r\n else:\r\n clr = color\r\n self.code += html_code(section,inline = inline, Return = True,color=clr)[:-(6+7*(inline==False))]\r\n else:\r\n self.code = html_code(code,inline = inline, Return = True)\r\n print('\\n'+'⩊'*40)\r\n \r\n def preview(self):\r\n Preview(self.code)\r\n\r\n\r\ndef HighlightCode(): \r\n try:\r\n inline = False\r\n code = Input(f'Code (end code with \"{DEFAULT_INPUT_TERMINATION_CODE}\" on an empty line after the code):',ml=True)\r\n if code.find('\\n') ==-1: inline = bool(input('Inline? (True/False)'))\r\n preview = bool(input('Preview? (True/False)'))\r\n except ValueError:\r\n print('Please make sure your input for the \"Inline?\" and \"Preview?\" prompts were booleans')\r\n return\r\n print()\r\n code = html_code(code,inline = inline, Return = True)\r\n if preview:Preview(code)\r\n\r\ndef ThemeDisplay():\r\n code = []\r\n colors = []\r\n for Theme in Themes:\r\n colors.append(Theme)\r\n code.append( f'''# {Theme} Theme Test\r\ndef Function():\r\n print('Hello')''')\r\n htmlCode(code,color = colors).preview()\r\n \r\n","sub_path":"Html_Code_Formatter_v1.1.0.py","file_name":"Html_Code_Formatter_v1.1.0.py","file_ext":"py","file_size_in_byte":30672,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"} +{"seq_id":"389308329","text":"import urllib.request\nimport json\nimport pathlib\nimport time\n\n\n# ----------------------------------------------\n# Config\n# ----------------------------------------------\n\n# TAGS SEARCH\n\napi_url = 'https://derpibooru.org/api/v1/json/search/images?page={0}&sf=score&q={1}'\n\nban_list = ' -eqg, -anthro, -sketch, -screencap, -3d, -traditional art, -animated, -pixel art, -id card, -pony town, -fat, -meme, -derped, -tumblr'\nocs = 'oc, solo, pony,' + ban_list #-eqg, -anthro, -sketch, -screencap, -traditional art, -animated, -pixel art, -id card, -pony town, -fat, -meme, -derped'\npony_test = 'pony' #history is name of dataset\n\n# configure this\ntwi_tags = 'twilight sparkle, upvotes.lt:400, -humanized, -3d, -anthro, solo, -eqg, -animated, -plushie, -meme, -sketch'\n\n# filter things like captions and screenshots\nfilter_bad = ', -edit, -sketch, -caption, -fat, -derped, -tumblr, -animated, -pixel art, -traditional art'\n\n# pony / starlight glimmer\ntwi_and_star_2_tags = 'starlight glimmer || twilight sparkle, -humanized, -3d, -anthro, solo, -eqg, -plushie, -meme, -sketch'\n#starlight glimmer, twilight sparkle, -humanized, -3d, -anthro, solo, -eqg, -plushie, -meme, -sketch, -edit, -sketch, -caption, -fat, -derped, -tumblr, -animated, -pixel art, -traditional art\n\n# CONFIG ------------------------\n\n# set tags here\ntags = twi_and_star_2_tags + filter_bad\n\nsave_folder = 'glimmy_twi_3'\n\n# download img to hdd\ndownload = True\n# 0.5 is ok (sometimes web error), .77 now \nsleep_time = 1\n# start from page X \nstart_page = 1 \n\n# images to download (converts to pages) you need 10-100k images\nimage_count = 100000 \n\n#derpi api sizes https://derpibooru.org/api/v1/json/search/images?page=1&sf=score&q=pony\nsize = 'medium' \nfile_formats = [\"image/jpeg\", \"image/png\"]\n\n\n# ----------------------------------------------\n# Code below\n# ----------------------------------------------\n\npages = image_count // 15\n\nest_time = (pages * sleep_time * 15) / 60\n\nprint('estimated time:', est_time , 'min')\n\n\noutput_dir = pathlib.Path.cwd() / save_folder\n\n#process = True\n\nstart_time = time.time()\nctr = 0\nfor page_idx in range(start_page, pages+1):\n print('request page:', page_idx)\n time.sleep(sleep_time) # sleep\n\n # get page request\n req = urllib.request.urlopen(api_url.format(str(page_idx), tags).replace(' ', '%20'))\n\n data = req.read()\n json_object = json.loads(data.decode('utf8'))\n\n for image_idx, image in enumerate(json_object['images']):\n end_time = time.time()\n print('page>', str(page_idx) + '/' + str(pages), 'index>', str(image_idx)+'/15', 'dl>', ctr, '|| time left:', est_time-((end_time - start_time) / 60), 'min') # metrics\n \n if not image['mime_type'] in file_formats: continue\n image_url = image['representations'][size]\n image_url_parts = pathlib.PurePath(image_url).parts # https, derpicdn.net, img, 2020 ...\n\n # save with this file name\n image_filename_id = image_url_parts[6] + '.' + image_url_parts[7].split('.')[1] # id + extension 4013041350 + .png\n print('| source', image_url, '\\n| writing to:', output_dir / pathlib.Path(image_filename_id))\n if download:\n # download image request\n urllib.request.urlretrieve(image_url, output_dir / pathlib.Path(image_filename_id))\n time.sleep(sleep_time) # sleep\n ctr += 1\n\n","sub_path":"pbooru_downloader.py","file_name":"pbooru_downloader.py","file_ext":"py","file_size_in_byte":3384,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"} +{"seq_id":"276169475","text":"from xml.etree import ElementTree\nfrom requests.exceptions import HTTPError, ConnectionError\n\nfrom periscope.controllers.helpers import retry_on\nfrom periscope.controllers.cache import cache_manager as cm\nfrom periscope.services.uniprot import uniprot\n\n\nns_map = {'u': 'http://uniprot.org/uniprot'}\n\n\n@retry_on([HTTPError, ConnectionError], 10)\ndef get_protein_name(uniprot_id):\n xml_string = uniprot.get_xml(uniprot_id)\n root = ElementTree.fromstring(xml_string)\n fullname_elem = root.find('u:entry/u:protein/*/u:fullName', namespaces=ns_map)\n\n if fullname_elem is not None:\n return fullname_elem.text\n else:\n return None\n\n\n@retry_on([HTTPError, ConnectionError], 10)\ndef get_protein_ids(uniprot_id):\n xml_string = uniprot.get_xml(uniprot_id)\n root = ElementTree.fromstring(xml_string)\n\n ids = []\n for shortname_elem in root.findall('u:entry/u:protein/*/u:shortName', namespaces=ns_map):\n ids.append(shortname_elem.text)\n\n return ids\n","sub_path":"periscope/controllers/uniprot.py","file_name":"uniprot.py","file_ext":"py","file_size_in_byte":986,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"} +{"seq_id":"150668315","text":"#coding: utf-8\n\n## dummy\nimport os, sys\ncurrent_path = os.path.abspath(os.path.dirname(__file__))\ncurrent_path = current_path + \"/..\"\nsys.path.append(current_path)\nsys.path.append(current_path + \"/lib\")\nconfig_name = \"trendfollow_dummy\"\n\nimport traceback\nfrom compute_indicator import ComputeIndicator\nfrom datetime import datetime, timedelta\nimport time\nimport logging\n\nnow = datetime.now()\nnow = now.strftime(\"%Y%m%d%H%M%S\")\nlogfilename = \"%s/log/indicator_%s.log\" %(current_path, now)\nlogging.basicConfig(filename=logfilename, level=logging.INFO)\n\n\nif __name__ == \"__main__\":\n# instrument = \"GBP_JPY\"\n args = sys.argv\n instrument = args[1]\n base_time = datetime.now()\n# base_time = base_time.strftime(\"%Y-%m-%d %H:00:00\")\n base_time = base_time.strftime(\"%Y-%m-%d 00:00:00\")\n base_time = datetime.strptime(base_time, \"%Y-%m-%d %H:%M:%S\")\n time_width = 60 * 200\n compute_indicator = ComputeIndicator(instrument, time_width, base_time)\n\n while True:\n try:\n now = datetime.now()\n base_time = base_time + timedelta(minutes=1)\n while now < base_time:\n time.sleep(1)\n now = datetime.now()\n \n span = \"1m\"\n compute_indicator.computeInsertIndicator(base_time, span)\n\n except Exception as e:\n logging.info(e.args)\n logging.info(traceback.format_exc())\n\n\n","sub_path":"utility/insert_indicator_master_1m.py","file_name":"insert_indicator_master_1m.py","file_ext":"py","file_size_in_byte":1409,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"} +{"seq_id":"230249776","text":"\n\"\"\"\n\nComputing features about accelerometer orientations\n\nAuthor: Binod Thapa Chhetry\n\nDate: Jul 10, 2018\n\"\"\"\nimport numpy as np\nfrom numpy.linalg import norm\nfrom SWaN_accel.utils import *\n\nclass EnergyFeature:\n def __init__(self, X, subwins=30):\n self._X = X\n self._subwins = subwins\n\n @staticmethod\n def energies(X):\n X = as_float64(X)\n if not has_enough_samples(X):\n print(\n '''One of sub windows do not have enough samples, will ignore in\n feature computation''')\n energies = np.array([np.nan])\n else:\n energies = np.array([np.sum(np.square(X))/(X.shape[0])])\n return vec2rowarr(energies)\n \n \n def get_energies(self):\n result = apply_over_subwins(\n self._X, EnergyFeature.energies, subwins=self._subwins)\n\n self._energies = np.concatenate(result, axis=0)\n return self\n\n def smv_energy_sum(self):\n smv_energy_sum = np.nansum(self._energies, axis=0)\n result = vec2rowarr(smv_energy_sum)\n result = add_name(result, self.smv_energy_sum.__name__)\n return result\n\n def smv_energy_var(self):\n smv_energy_var = np.nanvar(self._energies, axis=0)\n result = vec2rowarr(smv_energy_var)\n result = add_name(result, self.smv_energy_var.__name__)\n return result\n\n ","sub_path":"build/lib/SWaN_accel/energy.py","file_name":"energy.py","file_ext":"py","file_size_in_byte":1373,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"68"} +{"seq_id":"213158694","text":"#!/usr/bin/python3\n\"\"\"amenities\"\"\"\n\nfrom flask import abort, jsonify, make_response, request\nfrom models import storage\nfrom api.v1.views import app_views\nfrom models.amenity import Amenity\n\n\n@app_views.route('/amenities', methods=['GET'],\n strict_slashes=False)\ndef get_amenities():\n \"\"\" get amenities\"\"\"\n get_amenitys = []\n for amenity in storage.all(Amenity).values():\n get_amenitys.append(amenity.to_dict())\n return jsonify(get_amenitys)\n\n\n@app_views.route('/amenities/