diff --git "a/4550.jsonl" "b/4550.jsonl"
new file mode 100644--- /dev/null
+++ "b/4550.jsonl"
@@ -0,0 +1,43 @@
+{"seq_id":"24550494422","text":"'''\nCode generation routines that provide a more direct\nconversion to Python bytecode.\n'''\n\nfrom codegen import codegens\nfrom code import CodeGen, CmpOp, Flags\nimport ast\n\n@codegens(ast.Program)\ndef codegen(self, c):\n mainclass, *classes = self.children\n \n c.setLine(1)\n \n #codegen classes\n for cls in classes:\n cls.codegen(c)\n\n #make main function\n main = CodeGen(c.filename, 'main')\n main.setFlags(Flags.NEWLOCALS | Flags.OPTIMIZED)\n mainclass.codegen(main)\n c.LOAD_CONST(main)\n c.MAKE_FUNCTION()\n c.STORE_NAME('main')\n\n #ifmain\n c.LOAD_NAME('__name__')\n c.LOAD_CONST('__main__')\n c.COMPARE_OP(CmpOp.EQUAL)\n dest = c.POP_JUMP_IF_FALSE()\n c.LOAD_NAME('main')\n c.CALL_FUNCTION()\n c.POP_TOP()\n\n #module return\n dest()\n c.LOAD_CONST(None)\n c.RETURN_VALUE()\n\n@codegens(ast.MainClassDecl)\ndef codegen(self, c):\n c.setLine(1)\n stmts = self.children\n for stmt in stmts:\n stmt.codegen(c)\n \n #return null;\n c.LOAD_CONST(None)\n c.RETURN_VALUE()\n\n@codegens(ast.ClassDecl)\ndef codegen(self, c):\n c.LOAD_BUILD_CLASS()\n\n cls = CodeGen(c.filename, self.name)\n cls.setFlags(Flags.NEWLOCALS)\n cls.argcount = 1\n cls.setLine(1)\n cls.LOAD_FAST('__locals__')\n cls.STORE_LOCALS()\n cls.LOAD_NAME('__name__')\n cls.STORE_NAME('__module__')\n\n #define constructor to initialize class variables\n init = CodeGen(cls.filename, '__init__')\n init.setFlags(Flags.NEWLOCALS | Flags.OPTIMIZED)\n init.argcount = 1\n init.varnames = ['self']\n if self.parent:\n init.LOAD_GLOBAL(self.parent)\n init.LOAD_ATTR('__init__')\n init.LOAD_FAST('self')\n init.CALL_FUNCTION(1)\n init.POP_TOP()\n\n for var in self.classvars:\n vartype = var.typename\n if vartype == ast.IntType:\n init.LOAD_CONST(0)\n elif vartype == ast.BoolType:\n init.LOAD_CONST(False)\n else:\n init.LOAD_CONST(None)\n init.LOAD_FAST('self')\n init.STORE_ATTR('_' + var.ID)\n init.LOAD_CONST(None)\n init.RETURN_VALUE()\n cls.LOAD_CONST(init)\n cls.MAKE_FUNCTION()\n cls.STORE_NAME('__init__')\n\n #generate methods\n def genMethod(method, methodname):\n func = CodeGen(cls.filename, methodname)\n func.setFlags(Flags.NEWLOCALS | Flags.OPTIMIZED)\n func.argcount = len(method.formallist) + 1\n func.varnames = ['self'] + list(map(lambda formal: formal.ID, method.formallist))\n method.codegen(func)\n cls.LOAD_CONST(func)\n cls.MAKE_FUNCTION()\n cls.STORE_NAME(methodname)\n for method in self.children:\n genMethod(method, method.ID)\n if method.ID == 'toString':\n genMethod(method, '__str__')\n\n cls.LOAD_CONST(None)\n cls.RETURN_VALUE()\n\n c.LOAD_CONST(cls)\n c.MAKE_FUNCTION()\n\n c.LOAD_CONST(self.name)\n if self.parent:\n c.LOAD_GLOBAL(self.parent)\n c.CALL_FUNCTION(3)\n else:\n c.CALL_FUNCTION(2)\n c.STORE_NAME(self.name)\n\n@codegens(ast.MethodDecl)\ndef codegen(self, c):\n c.setLine(1)\n *stmts, expr = self.children\n\n if self.isGenerator():\n c.setFlags(Flags.GENERATOR)\n\n for stmt in stmts:\n stmt.codegen(c)\n\n if expr:\n expr.codegen(c)\n else:\n c.LOAD_CONST(None)\n c.RETURN_VALUE()\n\n@codegens(ast.StmtList)\ndef codegen(self, c):\n stmts = self.children\n for s in stmts:\n s.codegen(c)\n\n@codegens(ast.Decl)\ndef codegen(self, c):\n (expr,) = self.children\n expr.codegen(c)\n \n context = self.context\n typename = context.localVarType(self.name)\n c.STORE_FAST(self.name)\n\n@codegens(ast.Assignment)\ndef codegen(self, c):\n (expr,) = self.children\n expr.codegen(c)\n \n context = self.context\n typename = context.localVarType(self.name)\n if typename:\n c.STORE_FAST(self.name)\n else:\n c.LOAD_FAST('self')\n c.STORE_ATTR('_' + self.name)\n\n@codegens(ast.Yield)\ndef codegen(self, c):\n (expr,) = self.children\n expr.codegen(c)\n\n c.YIELD_VALUE()\n c.POP_TOP()\n\n@codegens(ast.If)\ndef codegen(self, c):\n cond, ifstmt = self.children\n cond.codegen(c)\n\n # binary array size coincides with binary location of instructions\n jumpLoc = c.POP_JUMP_IF_FALSE()\n\n # Codegen the ifstmt\n ifstmt.codegen(c)\n\n # Mark the jump to after the if block\n jumpLoc()\n\n@codegens(ast.IfElse)\ndef codegen(self, c):\n cond, ifstmt, elsestmt = self.children\n cond.codegen(c)\n\n # binary array size coincides with binary location of instructions\n jumpLoc = c.POP_JUMP_IF_FALSE()\n\n # Codegen the ifstmt\n ifstmt.codegen(c)\n\n # Skip the else block\n endOfIf = c.JUMP_FORWARD()\n\n # Mark the jump to the else block\n jumpLoc()\n\n # Codegen the else stmt\n elsestmt.codegen(c)\n\n # Mark the end of the if statement\n endOfIf()\n\n\n@codegens(ast.While)\ndef codegen(self, c):\n cond, stmt = self.children\n\n #Start the loop\n loop = c.SETUP_LOOP()\n\n #Condition\n loopStart = c.marker()\n cond.codegen(c)\n jumpEnd = c.POP_JUMP_IF_FALSE()\n\n #Body\n stmt.codegen(c)\n c.JUMP_ABSOLUTE(loopStart)\n\n #Loop end\n jumpEnd()\n c.POP_BLOCK()\n\n #Instruction position after loop\n loop()\n\n@codegens(ast.ForEach)\ndef codegen(self, c):\n expr, stmt = self.children\n\n #Start the loop\n loop = c.SETUP_LOOP()\n\n #Expression\n expr.codegen(c)\n c.GET_ITER()\n loopStart = c.marker()\n jumpEnd = c.FOR_ITER()\n c.STORE_FAST(self.name)\n\n #Body\n stmt.codegen(c)\n c.JUMP_ABSOLUTE(loopStart)\n\n #Loop end\n jumpEnd()\n c.POP_BLOCK()\n\n #Instruction position after loop\n loop()\n c.popStack(1)\n\n@codegens(ast.Break)\ndef codegen(self, c):\n c.BREAK_LOOP()\n\n@codegens(ast.NewInstance)\ndef codegen(self, c):\n c.LOAD_GLOBAL(self.name)\n c.CALL_FUNCTION()\n\n@codegens(ast.This)\ndef codegen(self, c):\n c.LOAD_FAST('self')\n\n@codegens(ast.ID)\ndef codegen(self, c):\n context = self.context\n typename = context.localVarType(self.name)\n if typename:\n c.LOAD_FAST(self.name)\n else:\n c.LOAD_FAST('self')\n c.LOAD_ATTR('_' + self.name)\n\n@codegens(ast.Call)\ndef codegen(self, c):\n obj, *args = self.children\n func = self.func\n\n obj.codegen(c)\n c.LOAD_ATTR(func)\n\n for arg in args:\n arg.codegen(c)\n\n c.CALL_FUNCTION(len(args))\n\n","repo_name":"Javapyc/Javapyc","sub_path":"src/fastgen.py","file_name":"fastgen.py","file_ext":"py","file_size_in_byte":6368,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"84"}
+{"seq_id":"5262895747","text":"from PyQt5 import QtGui, QtCore\r\nfrom PyQt5.QtCore import Qt, pyqtSignal, QObject, pyqtSlot, QSize\r\nfrom PyQt5.QtGui import QImage, QPixmap, QIcon, QPainter, QColor, QFont, QBrush, QPen, QPolygon\r\nfrom PyQt5.QtWidgets import (QApplication, QVBoxLayout, QGroupBox, QMainWindow, QFrame, QGridLayout,\r\n QPushButton, QHBoxLayout, QTabWidget, QWidget, QLabel, QDialog,\r\n QPlainTextEdit, QLineEdit, QMenu,\r\n QScrollArea, QToolButton, QSizePolicy, QComboBox,\r\n QFileDialog, QSplitter, QListWidget, QListWidgetItem, QGraphicsView, QGraphicsScene, QMenuBar, QMessageBox)\r\nfrom PyQt5 import QtWidgets, QtGui, QtCore\r\n\r\nimport os\r\nimport json\r\nimport classifier\r\nimport utils\r\nimport h5py\r\nimport time\r\nimport re\r\nimport cv2\r\nimport control\r\nimport tab\r\nimport edit\r\n\r\n\r\nclass mainWindow(QMainWindow):\r\n _parse_projects = pyqtSignal()\r\n _open_project = pyqtSignal(str)\r\n _show_all = pyqtSignal(int)\r\n _edit_task = pyqtSignal(int)\r\n _edit_description = pyqtSignal(str)\r\n _refresh_tree = pyqtSignal()\r\n def __init__(self, screen):\r\n QMainWindow.__init__(self, flags=QtCore.Qt.Window)\r\n self.screen = screen\r\n self.file = None\r\n self.task_count = 0\r\n self.codenamecolor = []\r\n self.init_ui()\r\n \r\n def init_ui(self):\r\n self.adjust_window()\r\n self.init_widgets()\r\n self.place_blocks()\r\n self.connect_ui()\r\n\r\n def adjust_window(self):\r\n self.main_frame = QFrame()\r\n self.setCentralWidget(self.main_frame)\r\n self.setWindowTitle(\"Segmentation app. 1.1::Release\")\r\n self.screen.setHeight(self.screen.height() - 50)\r\n #self.screen.setWidth(self.screen.width() - 10)\r\n #print(self.screen)\r\n #size = QSize(1366, 768)\r\n #self.setMaximumSize(self.screen.size())\r\n #self.setMinimumSize(self.screen.size())\r\n self.main_layout = QGridLayout()\r\n self.main_frame.setLayout(self.main_layout)\r\n\r\n def init_widgets(self):\r\n self.tab = tab.tab(self)\r\n self.projectControl = control.projectControl(self)\r\n self.taskDescription = control.taskDescription(self)\r\n self.viewTree = control.polygonTree(parent=self, main=self)\r\n self.viewToolbar = control.viewToolbar()\r\n\r\n def place_blocks(self):\r\n self.main_layout.addWidget(self.tab, 1, 0, 4, 1)\r\n self.main_layout.addWidget(self.projectControl, 1, 1)\r\n self.main_layout.addWidget(self.taskDescription, 1, 1)\r\n self.main_layout.addWidget(self.viewTree, 1, 1)\r\n #self.main_layout.addWidget(self.viewToolbar, 0, 0)\r\n self.tab.view_layout.setMenuBar(self.viewToolbar)\r\n self.show_tab1()\r\n\r\n def connect_ui(self):\r\n self._parse_projects.connect(self.tab.parse_projects)\r\n self._open_project.connect(self.on_open_project)\r\n self.tab.currentChanged.connect(self.show_tab)\r\n #self.taskDescription.btn_addtask.clicked.connect(self.add_task)\r\n #self.taskDescription.btn_edit_task.clicked.connect(self.on_edit_task)\r\n #self.viewTree.btn_previous.clicked.connect(self.previous_view)\r\n #self.viewTree.btn_next.clicked.connect(self.next_view)\r\n #self.viewTree.btn_edit_task.clicked.connect(self.on_edit_task)\r\n self.viewToolbar.first.triggered.connect(self.previous_view_10)\r\n self.viewToolbar.first.triggered.connect(self.previous_polygons_10)\r\n self.viewToolbar.previous.triggered.connect(self.previous_view)\r\n self.viewToolbar.previous.triggered.connect(self.previous_polygons)\r\n self.viewToolbar.next.triggered.connect(self.next_view)\r\n self.viewToolbar.next.triggered.connect(self.next_polygons)\r\n self.viewToolbar.last.triggered.connect(self.next_view_10)\r\n self.viewToolbar.last.triggered.connect(self.next_polygons_10)\r\n self.viewToolbar.hideall.triggered.connect(self.hideall)\r\n self.viewToolbar.add.triggered.connect(self.add_task)\r\n self.viewToolbar.reseg.triggered.connect(self.reseg)\r\n self.viewToolbar.delete.triggered.connect(self.delete)\r\n\r\n self._edit_task.connect(self.on_edit_task)\r\n self._refresh_tree.connect(self.viewTree.fill)\r\n self.viewTree.itemSelectionChanged.connect(self.send_selected)\r\n self.viewToolbar.showall.triggered.connect(self.send_all)\r\n #self._edit_description.connect(self.on_edit_description)\r\n \r\n def show_tab(self): \r\n if self.tab.currentWidget() == self.tab.split:\r\n self.show_tab2()\r\n elif self.tab.currentWidget() == self.tab.projects:\r\n self.show_tab1()\r\n elif self.tab.currentWidget() == self.tab.view:\r\n self.show_tab3()\r\n\r\n def show_tab2(self):\r\n self.projectControl.setVisible(False)\r\n self.viewTree.setVisible(False)\r\n self.viewToolbar.setVisible(False)\r\n self.taskDescription.setVisible(True)\r\n\r\n def show_tab1(self):\r\n self.projectControl.setVisible(True)\r\n self.viewTree.setVisible(False)\r\n self.viewToolbar.setVisible(False)\r\n self.taskDescription.setVisible(False)\r\n\r\n def show_tab3(self):\r\n self.projectControl.setVisible(False)\r\n self.taskDescription.setVisible(False)\r\n self.viewTree.setVisible(True)\r\n self.viewToolbar.setVisible(True)\r\n\r\n def send_selected(self):\r\n items = []\r\n for item in self.viewTree.selectedItems():\r\n if self.viewTree.indexOfTopLevelItem(item) == -1:\r\n items.append(item)\r\n self.tab.view._selectedItems.emit(items)\r\n\r\n def send_all(self):\r\n items = []\r\n for item in self.viewTree.findItems('', Qt.MatchRecursive, 3):\r\n if self.viewTree.indexOfTopLevelItem(item) != -1:\r\n #childcount = item.childCount\r\n for index in range(item.childCount()):\r\n items.append(item.child(index))\r\n #print(item.text(0))\r\n #items.append(item)\r\n #print(items)\r\n self.tab.view_w.show_shapes(items)\r\n\r\n\r\n @pyqtSlot(str)\r\n def on_open_project(self, project_path):\r\n if self.file:\r\n self.file.close()\r\n self.file = h5py.File(project_path, 'r+')\r\n self.task_count = self.file.attrs[classifier.hdfs.TASK_COUNT.value]\r\n self.update_codenamecolor()\r\n self.tab.parse_tasks()\r\n #self.taskDescription.parse_description(self.file)\r\n self.tab.parse_view()\r\n #self.viewTree.adjust_pallete(self.file)\r\n self.viewTree.fill()\r\n self.projectControl.description.updateWidgetDescription()\r\n #self.projectControl.description.updateitem(self.file.attrs[classifier.hdfs.DESCRIPTION.value])\r\n #self.task_count = self.file.attrs[classifier.hdfs.TASK_COUNT.value]\r\n #self.tab.update_info()\r\n #print(self.task_count)\r\n\r\n def adjust_opened_project(self):\r\n self.task_count = self.file.attrs[classifier.hdfs.TASK_COUNT.value]\r\n self.tab.parse_tasks()\r\n #self.taskDescription.parse_description(self.file)\r\n self.tab.parse_view()\r\n self.tab.parse_projects()\r\n\r\n def add_task(self):\r\n if self.file:\r\n hdf = self.file\r\n image = QFileDialog.getOpenFileName()[0]\r\n if image:\r\n tasks_count = hdf.attrs[classifier.hdfs.TASK_COUNT.value]\r\n task = hdf.create_dataset(str(tasks_count), data=cv2.imread(image))\r\n task.attrs[classifier.tasks.COUNT.value] = 0\r\n task.attrs[classifier.tasks.STATUS.value] = classifier.tasks.TO_DO.value\r\n task.attrs[classifier.aerial.SOURCE.value] = 'не задано'\r\n task.attrs[classifier.aerial.ALTITUDE.value] = 'не задано'\r\n task.attrs[classifier.aerial.LATITUDE.value] = 'не задано'\r\n task.attrs[classifier.aerial.LONGITUDE.value] = 'не задано'\r\n task.attrs[classifier.aerial.SUN.value] = 'не задано'\r\n task.attrs[classifier.aerial.SPATIAL.value] = 'не задано'\r\n task.attrs[classifier.aerial.SIZE.value] = 'не задано'\r\n task.attrs[classifier.aerial.DATE.value] = 'не задано'\r\n task.attrs[classifier.aerial.TIME.value] = 'не задано'\r\n hdf.attrs[classifier.hdfs.TASK_COUNT.value] += 1\r\n self.adjust_opened_project()\r\n\r\n def reseg(self):\r\n if self.file:\r\n #print(self.file[str(self.viewTree.index)].attrs[classifier.tasks.STATUS.value])\r\n if self.file[str(self.viewTree.index)].attrs[classifier.tasks.STATUS.value] == classifier.tasks.DONE.value:\r\n self.file[str(self.viewTree.index)].attrs[classifier.tasks.STATUS.value] = classifier.tasks.IN_PROGRESS.value\r\n self.adjust_opened_project()\r\n #print(self.file[str(self.viewTree.index)].attrs[classifier.tasks.STATUS.value])\r\n\r\n def delete(self):\r\n if self.file:\r\n msg = QMessageBox()\r\n msg.setWindowTitle(\"Удаление задачи\")\r\n msg.setInformativeText(\"Задача и данные разметки будут удалены. \\nУдалить?\")\r\n #msg.setStandardButtons(QMessageBox.Cancel | QMessageBox.OK)\r\n msg.addButton('Удалить', QMessageBox.AcceptRole)\r\n msg.addButton('Отмена', QMessageBox.RejectRole)\r\n rtn = msg.exec_()\r\n #print(rtn)\r\n if rtn == 0:\r\n current_index = self.viewTree.index\r\n \r\n #del self.file[str(current_index)]\r\n for index in range(current_index, self.task_count - 1):\r\n del self.file[str(index)]\r\n #if index < self.task_count:\r\n self.file[str(index)] = self.file[str(index + 1)]\r\n del self.file[str(self.task_count - 1)]\r\n\r\n self.file.attrs[classifier.hdfs.TASK_COUNT.value] -= 1\r\n self.adjust_opened_project()\r\n\r\n\r\n\r\n\r\n def deleteFileAttr(self, name):\r\n self.main.file[str(self.index)].attrs.__delitem__(name)\r\n self.main.file[str(self.index)].attrs[classifier.tasks.COUNT.value] -= 1\r\n self.updateFileNames(name)\r\n\r\n def updateFileNames(self, deleted_name):\r\n for name, value in self.main.file[str(self.index)].attrs.items():\r\n if utils.ispoints(name):\r\n #if name != classifier.tasks.COUNT.value and name != classifier.tasks.STATUS.value:\r\n if int(name) > int(deleted_name):\r\n self.main.file[str(self.index)].attrs.__delitem__(name)\r\n name = int(name)\r\n name -= 1\r\n self.main.file[str(self.index)].attrs[str(name)] = value\r\n self.main._refresh_tree.emit()\r\n\r\n @pyqtSlot(int)\r\n def on_edit_task(self, index=-1):\r\n current_task = self.tab.view_w.current_task()\r\n if index != -1:\r\n current_task = index\r\n #self.edit = edit.edit_widget(index=current_task, main=self, hdf=self.file)\r\n self.edit = edit.editWidget(parent=self, main=self, index=current_task)\r\n self.edit.exec_()\r\n\r\n @pyqtSlot(str)\r\n def on_edit_description(self, newdescription):\r\n if self.file: #корректно проверяю открыт ли hdf????\r\n self.file.attrs[classifier.hdfs.DESCRIPTION.value] = newdescription\r\n\r\n def update_codenamecolor(self):\r\n self.codenamecolor.clear()\r\n codes = classifier.classes.code()\r\n names = classifier.classes.name()\r\n colors= classifier.classes.color()\r\n for code, name, color in zip(codes, names, colors):\r\n if str(code) in self.file.attrs[classifier.hdfs.CLASSES.value]:\r\n self.codenamecolor.append((code, name, color))\r\n\r\n def on_show_all(self):\r\n #print(\"showall\")\r\n self._show_all.emit(1)\r\n\r\n def previous_view_10(self):\r\n if self.file:\r\n self.tab.change_view(index=-10)\r\n\r\n def previous_view(self):\r\n if self.file:\r\n self.tab.change_view(index=-1)\r\n\r\n def next_view(self):\r\n if self.file:\r\n self.tab.change_view(index=+1)\r\n\r\n def next_view_10(self):\r\n if self.file:\r\n self.tab.change_view(index=+10)\r\n\r\n def previous_polygons(self):\r\n if self.file:\r\n self.viewTree.update(index=-1)\r\n\r\n def previous_polygons_10(self):\r\n if self.file:\r\n self.viewTree.update(index=-10)\r\n\r\n def next_polygons(self):\r\n if self.file:\r\n self.viewTree.update(index=+1)\r\n\r\n def next_polygons_10(self):\r\n if self.file:\r\n self.viewTree.update(index=+10)\r\n\r\n def hideall(self):\r\n if self.file:\r\n self.tab.view_w.discard()\r\n\r\n def get_name(self, code):\r\n for triple in self.codenamecolor:\r\n if triple[0] == code:\r\n return triple[1]\r\n\r\n def get_code(self, name):\r\n for triple in self.codenamecolor:\r\n if triple[1] == name:\r\n return triple[0]\r\n\r\n def get_color(self, code):\r\n for triple in self.codenamecolor:\r\n if triple[0] == code:\r\n return triple[2]\r\n #if triple[0] == '000':\r\n if code == 0:\r\n return 5\r\n\r\n def adjust_code(self, taskindex, itemindex, newcode):\r\n attr = self.file[str(taskindex)].attrs[str(itemindex)]\r\n attr = re.sub(r';[0-9][0-9][0-9];', ';' + newcode + ';', attr)\r\n self.file[str(taskindex)].attrs[str(itemindex)] = attr\r\n\r\n\r\n def send_task_index(self, index):\r\n self.taskDescription.update_aerial(index)\r\n\r\n \r\n","repo_name":"AMKalinin/labeling3","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":13865,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"34760556201","text":"\"\"\"CRCache module: contains the CRCache class\"\"\"\n\nfrom direct.directnotify import DirectNotifyGlobal\nimport DistributedObject\n\nclass CRCache:\n notify = DirectNotifyGlobal.directNotify.newCategory(\"CRCache\")\n\n def __init__(self, maxCacheItems=10):\n self.maxCacheItems = maxCacheItems\n self.dict = {}\n self.fifo = []\n return None\n\n def flush(self):\n \"\"\"\n Delete each item in the cache then clear all references to them\n \"\"\"\n assert(self.checkCache())\n CRCache.notify.debug(\"Flushing the cache\")\n for distObj in self.dict.values():\n distObj.deleteOrDelay()\n # Null out all references to the objects so they will get gcd\n self.dict = {}\n self.fifo = []\n\n def cache(self, distObj):\n # Only distributed objects are allowed in the cache\n assert(isinstance(distObj, DistributedObject.DistributedObject))\n assert(self.checkCache())\n # Get the doId\n doId = distObj.getDoId()\n # Error check\n if self.dict.has_key(doId):\n CRCache.notify.warning(\"Double cache attempted for distObj \"\n + str(doId))\n else:\n # Call disable on the distObj\n distObj.disableAndAnnounce()\n \n # Put the distObj in the fifo and the dict\n self.fifo.append(distObj)\n self.dict[doId] = distObj\n \n if len(self.fifo) > self.maxCacheItems:\n # if the cache is full, pop the oldest item\n oldestDistObj = self.fifo.pop(0)\n # and remove it from the dictionary\n del(self.dict[oldestDistObj.getDoId()])\n # and delete it\n oldestDistObj.deleteOrDelay()\n \n # Make sure that the fifo and the dictionary are sane\n assert(len(self.dict) == len(self.fifo))\n return None\n\n def retrieve(self, doId):\n assert(self.checkCache())\n if self.dict.has_key(doId):\n # Find the object\n distObj = self.dict[doId]\n # Remove it from the dictionary\n del(self.dict[doId])\n # Remove it from the fifo\n self.fifo.remove(distObj)\n # return the distObj\n return distObj\n else:\n # If you can't find it, return None\n return None\n\n def contains(self, doId):\n return self.dict.has_key(doId)\n \n def delete(self, doId):\n assert(self.checkCache())\n assert(self.dict.has_key(doId))\n # Look it up\n distObj = self.dict[doId]\n # Remove it from the dict and fifo\n del(self.dict[doId])\n self.fifo.remove(distObj)\n # and delete it\n distObj.deleteOrDelay()\n \n def checkCache(self):\n # For debugging; this verifies that the cache is sensible and\n # returns true if so.\n from pandac.PandaModules import NodePath\n for obj in self.dict.values():\n if isinstance(obj, NodePath):\n assert(not obj.isEmpty() and obj.getTopNode() != render.node())\n return 1\n","repo_name":"PlumpMath/Panda3d-1.0.5","sub_path":"direct/src/distributed/CRCache.py","file_name":"CRCache.py","file_ext":"py","file_size_in_byte":3166,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"32084417609","text":"import torch\nfrom torch.utils.data import Dataset\nimport os\nimport numpy as np\n\nclass LoadDataset_from_numpy(Dataset):\n # Initialize your data, download, etc.\n def __init__(self, np_dataset):\n super(LoadDataset_from_numpy, self).__init__()\n\n # load files\n X_train = np.load(np_dataset[0])[\"x\"]\n y_train = np.load(np_dataset[0])[\"y\"]\n\n for np_file in np_dataset[1:]:\n X_train = np.vstack((X_train, np.load(np_file)[\"x\"]))\n y_train = np.append(y_train, np.load(np_file)[\"y\"])\n\n self.len = X_train.shape[0]\n self.x_data = torch.from_numpy(X_train)\n self.y_data = torch.from_numpy(y_train).long()\n\n # Correcting the shape of input to be (Batch_size, #channels, seq_len) where #channels=1\n if len(self.x_data.shape) == 3:\n if self.x_data.shape[1] != 1:\n self.x_data = self.x_data.permute(0, 2, 1)\n else:\n self.x_data = self.x_data.unsqueeze(1)\n\n def __getitem__(self, index):\n return self.x_data[index], self.y_data[index]\n\n def __len__(self):\n return self.len\n\n\ndef data_generator_np(training_files, subject_files, batch_size):\n train_dataset = LoadDataset_from_numpy(training_files)\n test_dataset = LoadDataset_from_numpy(subject_files)\n\n # to calculate the ratio for the CAL\n all_ys = np.concatenate((train_dataset.y_data, test_dataset.y_data))\n all_ys = all_ys.tolist()\n num_classes = len(np.unique(all_ys))\n counts = [all_ys.count(i) for i in range(num_classes)]\n\n train_loader = torch.utils.data.DataLoader(dataset=train_dataset,\n batch_size=batch_size,\n shuffle=True,\n drop_last=False,\n num_workers=0)\n\n test_loader = torch.utils.data.DataLoader(dataset=test_dataset,\n batch_size=batch_size,\n shuffle=False,\n drop_last=False,\n num_workers=0)\n\n return train_loader, test_loader, counts\n","repo_name":"emadeldeen24/AttnSleep","sub_path":"data_loader/data_loaders.py","file_name":"data_loaders.py","file_ext":"py","file_size_in_byte":2243,"program_lang":"python","lang":"en","doc_type":"code","stars":121,"dataset":"github-code","pt":"84"}
+{"seq_id":"35906824225","text":"# -*- coding: utf-8 -*-\nfrom openprocurement.auctions.core.adapters import AuctionConfigurator\nfrom openprocurement.auctions.dgf.models import (\n DGFOtherAssets,\n DGFFinancialAssets\n)\nfrom openprocurement.auctions.core.plugins.awarding.v2_1.adapters import (\n AwardingV2_1ConfiguratorMixin\n)\n\n\nclass AuctionDGFOtherAssetsConfigurator(AuctionConfigurator,\n AwardingV2_1ConfiguratorMixin):\n name = 'Auction Dgf Configurator'\n model = DGFOtherAssets\n\n\nclass AuctionDGFFinancialAssetsConfigurator(AuctionConfigurator,\n AwardingV2_1ConfiguratorMixin):\n name = 'Auction Dgf Configurator'\n model = DGFFinancialAssets","repo_name":"korobkoff/openprocurement.auctions.dgf","sub_path":"openprocurement/auctions/dgf/adapters.py","file_name":"adapters.py","file_ext":"py","file_size_in_byte":712,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"72972874195","text":"import cv2\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport sys\n\nif len(sys.argv) < 2:\n raise RuntimeError(\"Need one argument\")\n\nimg = cv2.imread(sys.argv[1], 0)\nf = np.fft.fft2(img)\nfshift = np.fft.fftshift(f)\n\nmagnitude_spectrum = 20*np.log(np.abs(fshift))\n\nplt.clf()\n\nplt.subplot(121)\nplt.imshow(img, cmap=\"gray\")\nplt.title(\"Input image\")\nplt.xticks([])\nplt.yticks([])\n\nplt.subplot(122)\nplt.imshow(magnitude_spectrum, cmap=\"gray\")\nplt.title(\"Magnitude spectrum\")\nplt.xticks([])\nplt.yticks([])\n\nplt.savefig(\"TEMP_test_fft.pdf\")\n\n","repo_name":"f-fathurrahman/ffr-python-stuffs","sub_path":"cv2/test_fft2_01.py","file_name":"test_fft2_01.py","file_ext":"py","file_size_in_byte":543,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"29809113465","text":"''' \nBinary search uses a technique of divide and conquer technique. \nThis type of technique is used in the case of sorted lists. \nInstead of searching an element one by one in the list, it directly goes to the middle element of the list, \ndivides the array into 2 parts, and decides element lies in which sub-array the element exists.\n\nBinary Search is a high-speed searching algorithm having the complexity of (logn)\n\n'''\n\ndef BinarySearch(Num_List, target, low=None,high=None):\n \n #lowest possible index you can check\n if low is None:\n low = 0\n \n #highest possible index\n if high is None:\n high = len(Num_List) - 1\n \n #Target is Not in the list \n if high < low:\n return -1\n \n mid = (low + high ) // 2 ## 2\n \n if Num_List[mid] == target:\n return mid \n \n elif target < Num_List[mid]:\n #using recursion to go through the list\n return BinarySearch(Num_List, target, low, mid - 1 )\n \n else:\n return BinarySearch(Num_List, target, mid + 1, high)\n\nNum_List = [1, 2, 3, 6, 9, 10, 11, 23, 34, 45] #list must be sorted for this to work properly\n\ntarget = 10\n\nprint(\"An item is found at index: \" + str(BinarySearch(Num_List, target))) #returns the position of the number you searched for\n","repo_name":"InnocentSuta/PythonChallenges","sub_path":"01_BinarySearch.py","file_name":"01_BinarySearch.py","file_ext":"py","file_size_in_byte":1296,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"905136099","text":"#!/usr/bin/env python3\nfrom http.server import BaseHTTPRequestHandler, HTTPServer\n\n\nclass WebServerHandler(BaseHTTPRequestHandler):\n def do_GET(self):\n if self.path.endswith(\"/hello\"):\n self.send_response(200)\n self.send_header('Content-type', 'text/html; charset=utf-8')\n self.end_headers()\n message = \"\"\n message += \"
Hello!\"\n self.wfile.write(message.encode())\n print(message)\n return\n else:\n self.send_error(404, 'File Not Found: %s' % self.path)\n\n\ndef main():\n try:\n port = 8000\n httpd = HTTPServer(('', port), WebServerHandler)\n print(\"Web Server running on port %s\" % port)\n httpd.serve_forever()\n except KeyboardInterrupt:\n print(\"^C entered, stopping web server....\")\n httpd.socket.close()\n\n\nif __name__ == '__main__':\n main()\n","repo_name":"S-Dey/udacity-full-stack-nanodegree","sub_path":"Term II/Unit03-The-Backend--Applications/Lesson01-02/webserver.py","file_name":"webserver.py","file_ext":"py","file_size_in_byte":928,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"84"}
+{"seq_id":"9729616541","text":"import sys\nfrom PyQt5.QtWidgets import QApplication, QWidget, QPushButton\nfrom PyQt5.QtCore import Qt, QMimeData\nfrom PyQt5.QtGui import QDrag\n\n\n# 左键可以点击,右键可以拖拽。\nclass Button(QPushButton):\n\n def __init__(self, title, parent):\n super().__init__(title, parent)\n\n def mouseMoveEvent(self, e):\n if e.buttons() != Qt.RightButton:\n return\n\n # QMimeData类为记录自身MIME信息的数据提供了一个容器\n mimeData = QMimeData()\n\n # 定义鼠标拖动\n drag = QDrag(self)\n drag.setMimeData(mimeData)\n drag.setHotSpot(e.pos() - self.rect().topLeft())\n\n dropAction = drag.exec_(Qt.MoveAction)\n\n def mousePressEvent(self, e):\n super().mousePressEvent(e)\n if e.button() == Qt.LeftButton:\n print('press')\n\n\nclass Example(QWidget):\n\n def __init__(self):\n super().__init__()\n self.init_ui()\n\n def init_ui(self):\n\n self.setAcceptDrops(True)\n\n self.button = Button('Button', self)\n self.button.move(100, 65)\n\n self.setGeometry(300, 300, 300, 150)\n self.setWindowTitle('Click and Move')\n\n def dragEnterEvent(self, e):\n e.accept()\n\n def dropEvent(self, e):\n position = e.pos()\n self.button.move(position)\n\n e.setDropAction(Qt.MoveAction)\n e.accept()\n\nif __name__ == '__main__':\n app = QApplication(sys.argv)\n ex = Example()\n ex.show()\n sys.exit(app.exec_())\n","repo_name":"slimboye/AT_Examples_Tutoral_PyQt5","sub_path":"8.Drag/drog2.py","file_name":"drog2.py","file_ext":"py","file_size_in_byte":1491,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"12866390492","text":"from base import Coverage, Ranks, ArgumentumParser\n\n\n# running calc. - main.py -f coverage_example.csv -b m2\n\nparam_dict = ArgumentumParser.arg_parser()\nmetrics = ArgumentumParser.get_metrics()\nprint(\";\"+\";\".join(metrics))\n\ncov = Coverage.Coverage(param_dict[\"file\"], param_dict[\"buggy-methods\"])\ncov.set_coverage_data(param_dict[\"file\"])\nrankContainer = Ranks.RankContainer(cov, metrics)\nrankContainer.add_ranks()\nrankContainer.printMinRanks()\n\n","repo_name":"bvancsics/FL_example","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":446,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"84"}
+{"seq_id":"1303245263","text":"import random\n\nfrom django.core.management.base import BaseCommand\n\nfrom myslack import factories, models\n\nUSER = 200\nWORKSPACES = 5\nCHANNELS = 10\nTHREADS_MIN = 5\nTHREADS_MAX = 15\nCOMMENTS_MIN = 1\nCOMMENTS_MAX = 20\nREACTIONS_MIN = 1\nREACTIONS_MAX = 25\n\n\nclass Command(BaseCommand):\n \"\"\"Custom `filldb` command.\n Django commands docs:\n https://docs.djangoproject.com/en/3.1/howto/custom-management-commands/\n \"\"\"\n help = 'Fill DB with sample data'\n\n def handle(self, *args, **options):\n print('Creating Users...')\n users = factories.UserFactory.create_batch(size=USER)\n print('Creating Workspaces...')\n workspaces = factories.WorkspaceFactory.create_batch(size=WORKSPACES)\n\n # Every user has profile in at least 1 Workspace\n print('Adding Users to Workspaces...')\n for user in users:\n workspaces = random.sample(workspaces, random.randint(1, WORKSPACES))\n for workspace in workspaces:\n factories.ProfileFactory.create(user=user, workspace=workspace)\n\n print('Adding content to Workspaces...')\n for workspace in workspaces:\n\n print(f'Filling {workspace}...')\n print('\\tAdding channels...')\n channels = factories.ChannelFactory.create_batch(size=CHANNELS, workspace=workspace)\n all_profiles = list(models.Profile.objects.filter(workspace_id=workspace))\n profiles_number = len(all_profiles)\n\n for channel in channels:\n\n print(f'\\tFilling {channel}...')\n print('\\t\\tAdding Channel members...')\n profiles = random.sample(all_profiles, random.randint(profiles_number // 2, profiles_number))\n\n for profile in profiles:\n factories.ChannelMembershipFactory.create(profile=profile, channel=channel)\n\n print('\\t\\tAdding Threads...')\n threads = [\n factories.ThreadFactory(author=random.choice(profiles), channel=channel)\n for _ in range(random.randint(THREADS_MIN, THREADS_MAX))\n ]\n\n print('\\t\\tAdding Comments to Threads...')\n comments = []\n for thread in threads:\n comments.extend([\n factories.CommentFactory(thread=thread, author=random.choice(profiles))\n for _ in range(random.randint(COMMENTS_MIN, COMMENTS_MAX))\n ])\n\n print('\\t\\tAdding Reactions to Threads...')\n for thread in threads:\n factories.ThreadReactionFactory(thread=thread, profile=random.choice(profiles))\n\n print('\\t\\tAdding Reactions to Comments...')\n for comment in comments:\n factories.CommentReactionFactory(comment=comment, profile=random.choice(profiles))\n","repo_name":"fr0mhell/myslack-django","sub_path":"myslack/management/commands/fill_db.py","file_name":"fill_db.py","file_ext":"py","file_size_in_byte":2890,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"37894451770","text":"import requests\nimport json\n\n#Task 1\nurl = 'https://api.github.com'\nuser ='AlexanderKolenko123'\n\nres = requests.get(f'{url}/users/{user}/repos')\nprint(res.status_code)\nprint(res)\n\nall_reps = []\nfor repos in res.json():\n all_reps.append(repos['name'])\n\n with open('_1.json', 'w') as f:\n to_json = json.dump(all_reps, f)\n\n#Task 2\n\nendpoint = \"https://api.nasa.gov/mars-photos/api/v1/rovers/curiosity/photos\"\napi_key = \"gZkkDU0fvcsLSzQYuLh2z9pjGQbIksAaMgccSZbH\"\nquery_params = {\"api_key\": api_key, \"earth_date\": \"2020-10-21\"}\nresponse = requests.get(endpoint, params=query_params)\nprint(response)\nphotos = response.json()[\"photos\"]\nprint(f\"Found {len(photos)} photos\")\n\nwith open('_2.json', 'w') as f:\n to_json = json.dump(photos, f)","repo_name":"AlexanderKolenko123/CollectingProcessingData","sub_path":"Homework_1/HomeWork1.py","file_name":"HomeWork1.py","file_ext":"py","file_size_in_byte":749,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"22131756628","text":"import os\r\nimport time\r\nimport shutil\r\n\r\ndef obter_data_atual():\r\n return time.time()\r\n\r\ndef calcular_diferenca_tres_dias():\r\n return 3 * 24 * 60 * 60 # 3 dias em segundos\r\n\r\ndef listar_arquivos(caminho_backup, arquivo_log):\r\n for root, dirs, files in os.walk(caminho_backup):\r\n for nome_arquivo in files:\r\n caminho_completo = os.path.join(root, nome_arquivo)\r\n \r\n tamanho = os.path.getsize(caminho_completo)\r\n data_criacao = os.path.getctime(caminho_completo)\r\n data_modificacao = os.path.getmtime(caminho_completo)\r\n \r\n arquivo_log.write(f\"Nome: {nome_arquivo}\\n\")\r\n arquivo_log.write(f\"Tamanho: {tamanho} bytes\\n\")\r\n arquivo_log.write(f\"Data de Criação: {time.ctime(data_criacao)}\\n\")\r\n arquivo_log.write(f\"Data da Última Modificação: {time.ctime(data_modificacao)}\\n\\n\")\r\n \r\n yield caminho_completo, data_criacao\r\n\r\ndef copiar_arquivos_recentes(caminho_destino, arquivo_log_destino, data_atual, diferenca_tres_dias):\r\n with open(arquivo_log_destino, 'w') as arquivo_log_destino:\r\n for caminho_completo, data_criacao in listar_arquivos(caminho_backup, arquivo_log_destino):\r\n if data_atual - data_criacao <= diferenca_tres_dias:\r\n destino = os.path.join(caminho_destino, os.path.basename(caminho_completo))\r\n shutil.copy2(caminho_completo, destino)\r\n arquivo_log_destino.write(f\"Arquivo copiado para {destino}\\n\")\r\n\r\ndef remover_arquivos_antigos(caminho_backup, data_atual, diferenca_tres_dias):\r\n for caminho_completo, data_criacao in listar_arquivos(caminho_backup, open(arquivo_saida, 'w')):\r\n if data_atual - data_criacao > diferenca_tres_dias:\r\n os.remove(caminho_completo)\r\n arquivo_saida.write(f\"Arquivo removido: {caminho_completo}\\n\")\r\n\r\nif __name__ == \"__main__\":\r\n caminho_backup = '/home/valcann/backupsFrom'\r\n arquivo_saida = '/home/valcann/backupsFrom.log'\r\n caminho_destino = '/home/valcann/backupsTo'\r\n arquivo_log_destino = '/home/valcann/backupsTo.log'\r\n\r\n data_atual = obter_data_atual()\r\n diferenca_tres_dias = calcular_diferenca_tres_dias()\r\n\r\n copiar_arquivos_recentes(caminho_destino, arquivo_log_destino, data_atual, diferenca_tres_dias)\r\n remover_arquivos_antigos(caminho_backup, data_atual, diferenca_tres_dias)\r\n","repo_name":"gcosta2107/backup-script","sub_path":"Automation.py","file_name":"Automation.py","file_ext":"py","file_size_in_byte":2418,"program_lang":"python","lang":"pt","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"1694121250","text":"# -*- coding: utf-8 -*-\nimport scrapy\nfrom ..items import SearchItem\n\n# https://www.biduo.cc/search.php?keyword=\n\n\nclass SearchSpider(scrapy.Spider):\n name = 'Search'\n allowed_domains = ['www.biduo.cc']\n start_urls = [\n 'https://www.biduo.cc/search.php?keyword={keyword}&page=8'.format(keyword='大道')]\n # 当前页数\n current_page = 1\n # 结果页总数\n page_count = 0\n\n def parse(self, response):\n result_list = response.xpath(\"//div[@class='result-list']/div\")\n if not result_list:\n return\n pages = response.xpath(\n \"//div[@class='search-result-page']/div/a[@title='末页']/@href\").re_first(r\"/search.php\\?keyword=.+&page=(\\d+)\")\n if pages:\n self.page_count = int(pages)\n for item in result_list:\n book_info = SearchItem()\n # 封面图片地址\n book_info['image_url'] = item.xpath(\n './div[1]/a/img/@src').extract_first()\n # 书籍地址\n book_info['source_url'] = item.xpath(\n './div[1]/a/@href').extract_first()\n # 书名\n book_info['name'] = item.xpath(\n './div[2]/h3/a/span/text()').extract_first()\n # 描述\n book_info['description'] = item.xpath(\n './div[2]/p/text()').extract_first()\n # 作者\n book_info['author'] = item.xpath(\n './div[2]/div/p[1]/span[2]/text()').extract_first().strip()\n # 书籍类型\n book_info['book_type'] = item.xpath(\n './div[2]/div/p[2]/span[2]/text()').extract_first().strip()\n # 更新时间\n book_info['update_time'] = item.xpath(\n './div[2]/div/p[3]/span[2]/text()').extract_first().strip()\n # 最新章节\n book_info['latest_chapters'] = item.xpath(\n './div[2]/div/p[4]/a/text()').extract_first().strip()\n book_info['latest_chapter_url'] = item.xpath(\n './div[2]/div/p[4]/a/@href').extract_first().strip()\n book_info['page_count'] = self.page_count\n yield book_info\n","repo_name":"Sean-Zheng/ReaderClub","sub_path":"BookScrapy/BookScrapy/spiders/Search.py","file_name":"Search.py","file_ext":"py","file_size_in_byte":2168,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"84"}
+{"seq_id":"28149356566","text":"# -*- mode: python -*-\n\nblock_cipher = None\n\n__version__ = \"0.1.1\"\n\nAPP_NAME = \"PDFx\"\n\ndatas = [\n (\"qml/*\", \"qml\"),\n (\"images/*.png\", \"images\")\n]\n\na = Analysis(['pdfxgui.py'],\n pathex=['/Users/chris/Projects/chris/pdf-link-extractor/pdfx-gui/source'],\n binaries=None,\n datas=datas,\n hiddenimports=[],\n hookspath=[],\n runtime_hooks=[],\n excludes=[],\n win_no_prefer_redirects=False,\n win_private_assemblies=False,\n cipher=block_cipher)\npyz = PYZ(a.pure, a.zipped_data,\n cipher=block_cipher)\nexe = EXE(pyz,\n a.scripts,\n a.binaries,\n a.zipfiles,\n a.datas,\n name=APP_NAME,\n debug=False,\n strip=False,\n upx=True,\n console=False,\n icon='images/icon.icns')\n\napp = BUNDLE(exe,\n name='PDFx.app',\n icon='images/icon.icns',\n bundle_identifier='com.metachris.pdfx',\n info_plist={\n 'CFBundleName': APP_NAME,\n 'CFBundleDisplayName': APP_NAME,\n 'CFBundleGetInfoString': \"Extract references from PDF documents\",\n 'CFBundleIdentifier': \"com.metachris.osx.sandwich\",\n 'CFBundleVersion': __version__,\n 'CFBundleShortVersionString': __version__,\n 'NSHumanReadableCopyright': u\"Copyright © 2016, Chris Hager, All Rights Reserved\"\n })\n","repo_name":"metachris/pdfx-gui","sub_path":"source/pdfxgui-onefile.spec","file_name":"pdfxgui-onefile.spec","file_ext":"spec","file_size_in_byte":1505,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"84"}
+{"seq_id":"26289730062","text":"import pytest\nimport os\nimport importlib\n\nimport atom3d.datasets as da\n\n\n# -- Dataset Loaders\n\n\ndef test_load_dataset_lmdb():\n dataset = da.load_dataset('tests/test_data/lmdb', 'lmdb')\n assert len(dataset) == 4\n for df in dataset:\n print(df)\n print(df['atoms'].x.dtype)\n assert df['atoms'].x.dtype in ['float', 'float32', 'float64']\n assert df['atoms'].y.dtype in ['float', 'float32', 'float64']\n assert df['atoms'].z.dtype in ['float', 'float32', 'float64']\n\n\ndef test_load_dataset_list():\n dataset = da.load_dataset('tests/test_data/list/pdbs.txt', 'pdb')\n assert len(dataset) == 4\n for df in dataset:\n print(df)\n print(df['atoms'].x.dtype)\n assert df['atoms'].x.dtype in ['float', 'float32', 'float64']\n assert df['atoms'].y.dtype in ['float', 'float32', 'float64']\n assert df['atoms'].z.dtype in ['float', 'float32', 'float64']\n\ndef test_load_dataset_list_nonexistent():\n dataset = da.load_dataset('tests/test_data/list/nonexistent.txt', 'pdb')\n assert len(dataset) == 1\n with pytest.raises(FileNotFoundError):\n df = dataset[0]\n\n#def test_load_dataset_sharded():\n# dataset = da.load_dataset('tests/test_data/sharded', 'sharded')\n# assert len(dataset) == 4\n\n\ndef test_load_dataset_pdb():\n dataset = da.load_dataset('tests/test_data/pdb', 'pdb')\n assert len(dataset) == 4\n for df in dataset:\n print(df)\n print(df['atoms'].x.dtype)\n assert df['atoms'].x.dtype in ['float', 'float32', 'float64']\n assert df['atoms'].y.dtype in ['float', 'float32', 'float64']\n assert df['atoms'].z.dtype in ['float', 'float32', 'float64']\n\n\n@pytest.mark.skipif(not importlib.util.find_spec(\"rdkit\") is not None,\n reason=\"Reading SDF files requires RDKit!\")\ndef test_load_dataset_sdf():\n dataset = da.load_dataset('tests/test_data/sdf', 'sdf')\n assert len(dataset) == 4\n for df in dataset:\n print(df)\n print(df['atoms'].x.dtype)\n assert df['atoms'].x.dtype in ['float', 'float32', 'float64']\n assert df['atoms'].y.dtype in ['float', 'float32', 'float64']\n assert df['atoms'].z.dtype in ['float', 'float32', 'float64']\n\n\n@pytest.mark.skipif(not importlib.util.find_spec(\"rosetta\") is not None,\n reason=\"Reading silent files requires pyrosetta!\")\ndef test_load_dataset_silent():\n dataset = da.load_dataset('tests/test_data/silent', 'silent')\n for x in dataset:\n print(x)\n assert len(dataset) == 8\n\n\ndef test_load_dataset_xyz():\n dataset = da.load_dataset('tests/test_data/xyz', 'xyz')\n assert len(dataset) == 3\n for df in dataset:\n print(df)\n print(df['atoms'].x.dtype)\n assert df['atoms'].x.dtype in ['float', 'float32', 'float64']\n assert df['atoms'].y.dtype in ['float', 'float32', 'float64']\n assert df['atoms'].z.dtype in ['float', 'float32', 'float64']\n\n\ndef test_load_dataset_xyzgdb():\n file_list = ['tests/test_data/xyz-gdb/dsgdb9nsd_000005.xyz', \n 'tests/test_data/xyz-gdb/dsgdb9nsd_000212.xyz', \n 'tests/test_data/xyz-gdb/dsgdb9nsd_001458.xyz']\n dataset = da.load_dataset(file_list, 'xyz-gdb')\n assert len(dataset) == 3\n for df in dataset:\n print(df)\n print(df['atoms'].x.dtype)\n assert df['atoms'].x.dtype in ['float', 'float32', 'float64']\n assert df['atoms'].y.dtype in ['float', 'float32', 'float64']\n assert df['atoms'].z.dtype in ['float', 'float32', 'float64']\n\n\n# -- Creator for LMDB dataset\n\n\ndef test_make_lmdb_dataset():\n # Load PDB dataset\n dataset = da.load_dataset('tests/test_data/pdb', 'pdb')\n assert len(dataset) == 4\n # Create LMDB dataset from PDB dataset\n da.make_lmdb_dataset(dataset, 'tests/test_data/_output_lmdb',\n filter_fn=None, serialization_format='json',\n include_bonds=False)\n # Try to load generated dataset\n new_dataset = da.load_dataset('tests/test_data/_output_lmdb', 'lmdb')\n assert len(new_dataset) == 4\n # Remove temporary files\n os.remove('tests/test_data/_output_lmdb/data.mdb')\n os.remove('tests/test_data/_output_lmdb/lock.mdb')\n os.rmdir('tests/test_data/_output_lmdb')\n\n","repo_name":"drorlab/atom3d","sub_path":"tests/datasets/test_datasets.py","file_name":"test_datasets.py","file_ext":"py","file_size_in_byte":4264,"program_lang":"python","lang":"en","doc_type":"code","stars":278,"dataset":"github-code","pt":"84"}
+{"seq_id":"21451208147","text":"from __future__ import (absolute_import, division, print_function,\n unicode_literals)\nfrom copy import deepcopy\n\nimport pandas as pd\n\n\nclass facet_null(object):\n\n def __init__(self, shrink=True):\n self.shrink = shrink\n\n def __radd__(self, gg):\n gg = deepcopy(gg)\n gg.facet = self\n return gg\n\n def train_layout(self, data):\n layout = pd.DataFrame({'PANEL': 1, 'ROW': 1, 'COL': 1,\n 'SCALE_X': 1, 'SCALE_Y': 1},\n index=[0])\n self.nrow = 1\n self.ncol = 1\n return layout\n\n def map_layout(self, data, layout):\n \"\"\"\n Assign a data points to panels\n\n Parameters\n ----------\n data : DataFrame\n dataframe for a layer\n layout : DataFrame\n As returned by self.train_layout\n \"\"\"\n data['PANEL'] = 1\n return data\n","repo_name":"unif2/ggplot","sub_path":"ggplot/facets/facet_null.py","file_name":"facet_null.py","file_ext":"py","file_size_in_byte":933,"program_lang":"python","lang":"en","doc_type":"code","dataset":"github-code","pt":"84"}
+{"seq_id":"34818614472","text":"from chinacapi.cloud_phone_region import CloudPhoneRegion\nfrom chinacapi.entity.cloud_phone import CloudPhone\nfrom chinacapi.chinac_api import ChinacApi\n\n\n# 创建 ChinacApi 实例\nca = ChinacApi('your_access_key', 'your_access_secret')\ncpa = ca.cloud_phone_api # 云手机API\n\n# 列出所有云手机\ncpl = cpa.list_cloud_phone()\nfor cp in cpl:\n print(cp)\n\n# 通过ID构造云手机对象\ncp = CloudPhone({\n 'id': 'cp-xxxxxxxxxxxxxx',\n 'region': CloudPhoneRegion.SZ,\n})\n\n# 获取云手机详情\nprint(cpa.describe_cloud_phone(cp.region, cp.id))\n\n# 列出云手机ADB白名单\nips = cpa.list_cloud_phone_adb_white_ip(CloudPhoneRegion.SZ)\nprint(ips)\n\n# 设置云手机ADB白名单\nprint(cpa.set_cloud_phone_adb_white_ip(CloudPhoneRegion.SZ, list(set(ips + [\n '1.1.1.1', '1.1.1.2', '1.1.1.3',\n '1.1.1.4', '1.1.1.5', '1.1.1.6',\n]))))\nips = cpa.list_cloud_phone_adb_white_ip(CloudPhoneRegion.SZ)\nprint(ips)\n\n\nprint(cpa.reboot_cloud_phone(cp)) # 重启云手机\nprint(cpa.start_cloud_phone(cp)) # 开启云手机\nprint(cpa.stop_cloud_phone(cp)) # 关闭云手机\n","repo_name":"AkagiYui/ChinacApi","sub_path":"example/cloud_phone.py","file_name":"cloud_phone.py","file_ext":"py","file_size_in_byte":1077,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"84"}
+{"seq_id":"41557268746","text":"import re\nimport json\nimport os\nimport logging\nfrom io import BytesIO\nimport zipfile\n\nfrom pywhale.lib.utils import Utils\n\n\n# Module for parsing email attachments\nclass AttachmentParser:\n\n attachment = None\n\n indicators = []\n directory = None\n merged_links = [\n\n ]\n logger = logging.getLogger(\"Attachments\")\n files = []\n\n def __init__(self, attachment, indicator_directory=None):\n self.attachment = attachment\n self.directory = indicator_directory\n if not self.directory:\n self.directory = os.path.join(os.path.abspath(os.path.dirname(__file__)), \"../\", \"attachments\")\n self.load_indicators()\n\n def load_indicators(self):\n self.indicators = []\n for i_file in os.listdir(self.directory):\n full_path = os.path.join(self.directory, i_file)\n try:\n with open(full_path, 'r') as f:\n indicator = json.load(f)\n if indicator not in self.indicators:\n self.indicators.append(indicator)\n except Exception as e:\n self.logger.error(\"Error loading script %s: %s\" % (full_path, str(e)))\n self.logger.debug(\"Loaded %d attachment scripts\" % len(self.indicators))\n\n def unzip(self, data):\n if type(data) == str:\n data = data.encode()\n f = BytesIO(data)\n try:\n zf = zipfile.ZipFile(f, 'r')\n out_stream = []\n for zfile in zf.infolist():\n out_stream.append({\"name\": zfile.filename, \"body\": zf.read(zfile)})\n except Exception as e:\n self.logger.error(\"Error reading zipped file data: %s\" % str(e))\n return []\n return out_stream\n\n def run_indicators(self, indicators, body):\n results = []\n body_bytes = type(body) is not str\n for i in indicators:\n if 'type' in i:\n match = Utils.contains(needle=i['needles'], haystack=body, search_type=i['type'])\n if match:\n results.append({\n \"match\": match,\n \"indicator\": i['indicator'],\n \"output\": i['output']\n })\n\n return results\n\n def run_file_indicators(self, indicators, filename, parent):\n results = []\n for i in indicators:\n match = Utils.contains(needle=i['needles'], haystack=filename, search_type=i['type'])\n if match:\n results.append({\n \"location\": filename,\n \"match\": match,\n \"name\": parent['name'],\n \"indicator\": i['indicator'],\n \"output\": i['output']\n })\n\n return results\n\n def run_script(self, script):\n output = []\n body = [{\"name\": self.attachment['filename'], \"body\": self.attachment['body']}]\n self.logger.info(\"Processing attachment %s with script %s\" % (self.attachment['filename'], script['name']))\n\n # unzip with BytesIO\n if script['decoder'] == \"zip\":\n zipdata = self.unzip(self.attachment['body'])\n for f in zipdata:\n body.append(f)\n self.logger.debug(\"Found %d files in zipfile: %s\" % (len(body), self.attachment['filename']))\n\n # scan one or multiple file bodies for strings\n for f in body:\n self.files.append({\"name\": f['name'], \"length\": len(f['body']), \"parent\": self.attachment['filename']})\n results = self.run_indicators(script['indicators'], f['body'])\n if results:\n for r in results:\n output.append({\n \"attachment\": self.attachment['filename'],\n \"location\": f['name'],\n \"match\": r\n })\n # for scanning zipped files\n if 'files' in script:\n results = self.run_file_indicators(script['files'], f['name'], parent=script)\n if results:\n for r in results:\n output.append({\n \"key\": self.attachment['filename'],\n \"location\": f['name'],\n \"match\": r\n })\n return output\n\n def process(self):\n full_output = []\n for script in self.indicators:\n matches = [script['file_type']] if type(script['file_type']) == str else script['file_type']\n for m_type in matches:\n if self.attachment['filename'].lower().endswith(m_type):\n result = self.run_script(script)\n if result:\n full_output.extend(result)\n return full_output\n","repo_name":"stefan2200/pywhale","sub_path":"pywhale/lib/attachments.py","file_name":"attachments.py","file_ext":"py","file_size_in_byte":4790,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"84"}
+{"seq_id":"5186763453","text":"from SpiffWorkflow.bpmn.specs.data_spec import TaskDataReference\nfrom SpiffWorkflow.bpmn.parser.util import one\nfrom SpiffWorkflow.bpmn.parser.ValidationException import ValidationException\nfrom SpiffWorkflow.bpmn.parser.TaskParser import TaskParser\nfrom SpiffWorkflow.bpmn.parser.task_parsers import SubprocessParser\n\nfrom SpiffWorkflow.camunda.specs.business_rule_task import BusinessRuleTask\nfrom SpiffWorkflow.camunda.specs.multiinstance_task import SequentialMultiInstanceTask, ParallelMultiInstanceTask\nfrom SpiffWorkflow.camunda.specs.user_task import Form, FormField, EnumFormField\n\nCAMUNDA_MODEL_NS = 'http://camunda.org/schema/1.0/bpmn'\n\n\nclass CamundaTaskParser(TaskParser):\n\n def parse_extensions(self, node=None):\n extensions = {}\n extension_nodes = self.xpath('.//bpmn:extensionElements/camunda:properties/camunda:property')\n for ex_node in extension_nodes:\n extensions[ex_node.get('name')] = ex_node.get('value')\n return extensions\n\n def _add_multiinstance_task(self, loop_characteristics):\n\n sequential = loop_characteristics.get('isSequential') == 'true'\n prefix = 'bpmn:multiInstanceLoopCharacteristics'\n\n cardinality = self.xpath(f'./{prefix}/bpmn:loopCardinality')\n cardinality = cardinality[0].text if len(cardinality) > 0 else None\n collection = self.attribute('collection', 'camunda', loop_characteristics)\n if cardinality is None and collection is None:\n self.raise_validation_exception('A multiinstance task must specify a cardinality or a collection')\n\n element_var = self.attribute('elementVariable', 'camunda', loop_characteristics)\n condition = self.xpath(f'./{prefix}/bpmn:completionCondition')\n condition = condition[0].text if len(condition) > 0 else None\n\n original = self.spec.task_specs.pop(self.task.name)\n\n # We won't include the data input, because sometimes it is the collection, and other times it\n # is the cardinality. The old MI task evaluated the cardinality at run time and treated it like\n # a cardinality if it evaluated to an int, and as the data input if if evaluated to a collection\n # I highly doubt that this is the way Camunda worked then, and I know that's not how it works\n # now, and I think we should ultimately replace this with something that corresponds to how\n # Camunda actually handles things; however, for the time being, I am just going to try to \n # replicate the old behavior as closely as possible.\n # In our subclassed MI task, we'll update the BPMN multiinstance attributes when the task starts.\n params = {\n 'task_spec': '',\n 'cardinality': cardinality, \n 'data_output': TaskDataReference(collection) if collection is not None else None,\n 'output_item': TaskDataReference(element_var) if element_var is not None else None,\n 'condition': condition,\n }\n if sequential:\n self.task = SequentialMultiInstanceTask(self.spec, original.name, **params)\n else:\n self.task = ParallelMultiInstanceTask(self.spec, original.name, **params)\n self._copy_task_attrs(original, loop_characteristics)\n\n\nclass BusinessRuleTaskParser(CamundaTaskParser):\n dmn_debug = None\n\n def create_task(self):\n decision_ref = self.get_decision_ref(self.node)\n return BusinessRuleTask(self.spec, self.bpmn_id, \n dmnEngine=self.process_parser.parser.get_engine(decision_ref, self.node),\n **self.bpmn_attributes)\n\n @staticmethod\n def get_decision_ref(node):\n return node.attrib['{' + CAMUNDA_MODEL_NS + '}decisionRef']\n\n\nclass UserTaskParser(CamundaTaskParser):\n \"\"\"Base class for parsing User Tasks\"\"\"\n\n def create_task(self):\n form = self.get_form()\n return self.spec_class(self.spec, self.bpmn_id, form=form, **self.bpmn_attributes)\n\n def get_form(self):\n \"\"\"Camunda provides a simple form builder, this will extract the\n details from that form and construct a form model from it. \"\"\"\n form = Form()\n try:\n form.key = self.attribute('formKey', 'camunda')\n except KeyError:\n return form\n for xml_field in self.xpath('.//camunda:formData/camunda:formField'):\n if xml_field.get('type') == 'enum':\n field = self.get_enum_field(xml_field)\n else:\n field = FormField()\n\n field.id = xml_field.get('id')\n field.type = xml_field.get('type')\n field.label = xml_field.get('label')\n field.default_value = xml_field.get('defaultValue')\n\n prefix = '{' + self.nsmap.get('camunda') + '}'\n for child in xml_field:\n if child.tag == f'{prefix}properties':\n for p in child:\n field.add_property(p.get('id'), p.get('value'))\n\n if child.tag == f'{prefix}validation':\n for v in child:\n field.add_validation(v.get('name'), v.get('config'))\n\n form.add_field(field)\n return form\n\n def get_enum_field(self, xml_field):\n field = EnumFormField()\n\n for child in xml_field:\n if child.tag == '{' + self.nsmap.get('camunda') + '}value':\n field.add_option(child.get('id'), child.get('name'))\n return field\n\n\n# These classes need to be able to use the overriden _add_multiinstance_task method\n# so they have to inherit from CamundaTaskParser. Therefore, the parsers have to just\n# be copied, because both they and the CamundaTaskParser inherit from the base task\n# parser. I am looking forward to the day when I can replaced all of this with\n# something sane and sensible.\n\nclass SubWorkflowParser(CamundaTaskParser):\n\n def create_task(self):\n subworkflow_spec = SubprocessParser.get_subprocess_spec(self)\n return self.spec_class(self.spec, self.bpmn_id, subworkflow_spec=subworkflow_spec, **self.bpmn_attributes)\n\n\nclass CallActivityParser(CamundaTaskParser):\n \"\"\"Parses a CallActivity node.\"\"\"\n\n def create_task(self):\n subworkflow_spec = SubprocessParser.get_call_activity_spec(self)\n return self.spec_class(self.spec, self.bpmn_id, subworkflow_spec=subworkflow_spec, **self.bpmn_attributes)\n\n\nclass ScriptTaskParser(TaskParser):\n \"\"\"\n Parses a script task\n \"\"\"\n\n def create_task(self):\n script = self.get_script()\n return self.spec_class(self.spec, self.bpmn_id, script=script, **self.bpmn_attributes)\n\n def get_script(self):\n \"\"\"\n Gets the script content from the node. A subclass can override this\n method, if the script needs to be pre-parsed. The result of this call\n will be passed to the Script Engine for execution.\n \"\"\"\n try:\n return one(self.xpath('.//bpmn:script')).text\n except AssertionError as ae:\n raise ValidationException(\n \"Invalid Script Task. No Script Provided. \" + str(ae),\n node=self.node, file_name=self.filename)\n","repo_name":"sartography/SpiffWorkflow","sub_path":"SpiffWorkflow/camunda/parser/task_spec.py","file_name":"task_spec.py","file_ext":"py","file_size_in_byte":7187,"program_lang":"python","lang":"en","doc_type":"code","stars":1532,"dataset":"github-code","pt":"84"}
+{"seq_id":"23429181284","text":"\nimport time\nfrom ustruct import unpack, unpack_from\nfrom array import array\n\n# BME280 default address.\nBME280_I2CADDR = 0x76\n\n# Operating Modes\nBME280_OSAMPLE_1 = 1\nBME280_OSAMPLE_2 = 2\nBME280_OSAMPLE_4 = 3\nBME280_OSAMPLE_8 = 4\nBME280_OSAMPLE_16 = 5\n\nBME280_REGISTER_CONTROL_HUM = 0xF2\nBME280_REGISTER_CONTROL = 0xF4\n\n\nclass BME280:\n\n def __init__(self,\n mode=BME280_OSAMPLE_1,\n address=BME280_I2CADDR,\n i2c=None,\n **kwargs):\n # Check that mode is valid.\n if mode not in [BME280_OSAMPLE_1, BME280_OSAMPLE_2, BME280_OSAMPLE_4,\n BME280_OSAMPLE_8, BME280_OSAMPLE_16]:\n raise ValueError(\n 'Unexpected mode value {0}. Set mode to one of '\n 'BME280_ULTRALOWPOWER, BME280_STANDARD, BME280_HIGHRES, or '\n 'BME280_ULTRAHIGHRES'.format(mode))\n self._mode = mode\n self.address = address\n if i2c is None:\n raise ValueError('An I2C object is required.')\n self.i2c = i2c\n\n # load calibration data\n dig_88_a1 = self.i2c.readfrom_mem(self.address, 0x88, 26)\n dig_e1_e7 = self.i2c.readfrom_mem(self.address, 0xE1, 7)\n self.dig_T1, self.dig_T2, self.dig_T3, self.dig_P1, \\\n self.dig_P2, self.dig_P3, self.dig_P4, self.dig_P5, \\\n self.dig_P6, self.dig_P7, self.dig_P8, self.dig_P9, \\\n _, self.dig_H1 = unpack(\"> 4)\n\n self.dig_H6 = unpack_from(\"> 4\n raw_press = ((readout[0] << 16) | (readout[1] << 8) | readout[2]) >> 4\n # temperature(0xFA): ((msb << 16) | (lsb << 8) | xlsb) >> 4\n raw_temp = ((readout[3] << 16) | (readout[4] << 8) | readout[5]) >> 4\n # humidity(0xFD): (msb << 8) | lsb\n raw_hum = (readout[6] << 8) | readout[7]\n\n result[0] = raw_temp\n result[1] = raw_press\n result[2] = raw_hum\n\n def read_compensated_data(self, result=None):\n \"\"\" Reads the data from the sensor and returns the compensated data.\n\n Args:\n result: array of length 3 or alike where the result will be\n stored, in temperature, pressure, humidity order. You may use\n this to read out the sensor without allocating heap memory\n\n Returns:\n array with temperature, pressure, humidity. Will be the one from\n the result parameter if not None\n \"\"\"\n self.read_raw_data(self._l3_resultarray)\n raw_temp, raw_press, raw_hum = self._l3_resultarray\n # temperature\n var1 = ((raw_temp >> 3) - (self.dig_T1 << 1)) * (self.dig_T2 >> 11)\n var2 = (((((raw_temp >> 4) - self.dig_T1) *\n ((raw_temp >> 4) - self.dig_T1)) >> 12) * self.dig_T3) >> 14\n self.t_fine = var1 + var2\n temp = (self.t_fine * 5 + 128) >> 8\n\n # pressure\n var1 = self.t_fine - 128000\n var2 = var1 * var1 * self.dig_P6\n var2 = var2 + ((var1 * self.dig_P5) << 17)\n var2 = var2 + (self.dig_P4 << 35)\n var1 = (((var1 * var1 * self.dig_P3) >> 8) +\n ((var1 * self.dig_P2) << 12))\n var1 = (((1 << 47) + var1) * self.dig_P1) >> 33\n if var1 == 0:\n pressure = 0\n else:\n p = 1048576 - raw_press\n p = (((p << 31) - var2) * 3125) // var1\n var1 = (self.dig_P9 * (p >> 13) * (p >> 13)) >> 25\n var2 = (self.dig_P8 * p) >> 19\n pressure = ((p + var1 + var2) >> 8) + (self.dig_P7 << 4)\n\n # humidity\n h = self.t_fine - 76800\n h = (((((raw_hum << 14) - (self.dig_H4 << 20) -\n (self.dig_H5 * h)) + 16384)\n >> 15) * (((((((h * self.dig_H6) >> 10) *\n (((h * self.dig_H3) >> 11) + 32768)) >> 10) +\n 2097152) * self.dig_H2 + 8192) >> 14))\n h = h - (((((h >> 15) * (h >> 15)) >> 7) * self.dig_H1) >> 4)\n h = 0 if h < 0 else h\n h = 419430400 if h > 419430400 else h\n humidity = h >> 12\n\n if result:\n result[0] = temp\n result[1] = pressure\n result[2] = humidity\n return result\n\n return array(\"i\", (temp, pressure, humidity))\n\n @property\n def values(self):\n \"\"\" human readable values \"\"\"\n\n t, p, h = self.read_compensated_data()\n\n p = p // 256\n pi = p // 100\n pd = p - pi * 100\n\n hi = h // 1024\n hd = h * 100 // 1024 - hi * 100\n return (\"{}C\".format(t / 100), \"{}.{:02d}hPa\".format(pi, pd),\n \"{}.{:02d}%\".format(hi, hd))\n\n\nimport esp\nimport time\nimport json\nimport machine\nimport network\nfrom umqtt.robust import MQTTClient\n\np2 = machine.Pin(4, machine.Pin.PULL_UP)\n\ni2c_1 = machine.I2C(scl=machine.Pin(0), sda=machine.Pin(2))\ni2c_2 = machine.I2C(scl=machine.Pin(14), sda=machine.Pin(12))\nbme1 = BME280(i2c=i2c_1)\nbme2 = BME280(i2c=i2c_2)\n\nBLINK_DELAY=0.1\nMEASR_DELAY=60\n\ndef blink():\n p2.on()\n time.sleep(BLINK_DELAY)\n p2.off()\n\n\ndef get_rssi():\n nets = sta.scan()\n for n in nets:\n if n[0].decode() == settings[\"network\"]:\n return n[3]\n return None\n\n\ndef ctof(t):\n return t * (9 / 5) + 32\n\n\nf = open(\"settings.json\")\nraw_settings = \"\".join(f.readlines())\nf.close()\n\nsettings = json.loads(raw_settings)\n\ncredentials = bytes(\"channels/{:s}/publish/{:s}\".format(settings[\"channel\"], settings[\"apikey\"]), 'utf-8')\n\nsta = network.WLAN(network.STA_IF)\nsta.active(True)\nsta.connect(settings[\"network\"], settings[\"password\"])\n\n#randomNum = int.from_bytes(uos.urandom(3), 'little')\n#myMqttClient = bytes(\"client_\"+str(randomNum), 'utf-8')\n\nc = MQTTClient(\n client_id = \"client_1\",\n server = settings[\"broker\"],\n user = settings[\"userid\"],\n password = settings[\"mqttapikey\"],\n port = 1883\n )\nwhile True:\n print(\"connecting to wifi\")\n if sta.isconnected():\n print(\"connected!\")\n c.connect()\n print(\"mqtt connected!\")\n break\n time.sleep(1)\n\n\ntime.sleep(5)\n\ncounter = 0\nwhile True:\n blink()\n #print(\"BME1\", bme1.values)\n #print(\"BME2\", bme2.values)\n rv1 = bme1.values\n rv2 = bme2.values\n t1 = ctof(float(rv1[0][:-1]))\n t2 = ctof(float(rv2[0][:-1]))\n h1 = float(rv1[2][:-1])\n h2 = float(rv2[2][:-1])\n payload = bytes(\"field1={:.1f}&field2={:.1f}&field3={:.1f}&field4={:.1f}&field5={:d}&field6={:d}\\n\".format(t1, t2, h1, h2, get_rssi(), esp.freemem()), 'utf-8')\n c.publish(credentials, payload)\n counter += 1\n time.sleep(MEASR_DELAY)\n","repo_name":"stephanellis/nodemcu-mpy-fedora","sub_path":"main_freezer.py","file_name":"main_freezer.py","file_ext":"py","file_size_in_byte":8343,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"23617871423","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Oct 15 11:38:36 2018\n\n@author: megan wojciechowicz, maayan lab \n\n\"\"\"\nimport pronto\nfrom pronto.relationship import Relationship\nimport eutils_query as EQ\nimport copy\nimport numpy as np\nimport pandas as pd\n\n# This function takes a list of parent terms and finds corresponding children\n# Terms with no children are kept in the returned list \n# Input: ontology object, list of parent terms \n# Returns: list of terms \ndef find_children_terms(ont, parents):\n new_terms = []\n for parent in parents:\n children = []\n for term in ont:\n ID = term.id \n try:\n isa = ont[ID].relations[Relationship('is_a')]\n for x in isa:\n if x.name == parent:\n children.append(term.name)\n new_terms.append(term.name)\n except:\n pass\n if len(children)==0:\n new_terms.append(parent) \n return(new_terms)\n\n# This function loops through each term in a list of terms and checks if there are also parent terms in the list\n# If a parent term exists, all pubMed ids from child term are added to the parent term \n# Input: ontology object, list of terms, 2D list of pubMed ids associated with each term \n# Returns: 2D list of pubMed ids \ndef back_propogation(ont,terms, ids):\n new_ids = copy.deepcopy(ids)\n for idx_term,term in enumerate(terms):\n for o in ont:\n if term == o.name:\n parents = [x.name for x in o.rparents()]\n for parent in parents:\n if parent in terms:\n print('parent:'+parent+'\\t'+'child:'+term)\n idx_parent = terms.index(parent)\n for x in ids[idx_term]:\n new_ids[idx_parent].append(x)\n return(new_ids) \n\n\n\n \n\n############################# FOR GO ontology only #############################\n\n# This function subsets data in association files from http://geneontology.org/gene-associations/\n# Code copied from GITHUB: https://github.com/MaayanLab/EnrichrPythonScripts/blob/master/WormEnrichr/GO/process_GO_wrombase.ipynb by Zachary Flamholz\n# Input: dataframe , GO class type ('P'->biological process,'C'->cellular component,'F'->molecular function)\n# Returns: subsetted dataframe \ndef clean_data (df, go_type): \n matrix = np.chararray((df.shape[0], 17), itemsize=150, unicode=True)\n for i, row in enumerate(df.itertuples()):\n lst = row[1].split('\\t')\n matrix[i, :] = lst\n df_clean = pd.DataFrame(data=matrix)\n df_clean.columns = [\"DB\", \"DB gene ID\", \"Gene symbol\", \"Qualifier\", \"GO ID\", \"Reference\", \n \"Evidence code\", \"Evidence from\", \"GO class\", \"attribute\", \"Locus tag\",\n \"gene/protein\", \"tax id\", \"date\", \"Assigned by\", \"additional information\", \"empty\"]\n df_clean= df_clean[df_clean[\"GO class\"] == go_type]\n df_clean= df_clean[df_clean[\"Evidence code\"] != 'IEA']# remove any annotation assigned by electronic matching and with the NOT qualifier which is used to specify a gene is not associated with a term\n df_clean= df_clean[df_clean[\"Qualifier\"] != 'NOT']\n return(df_clean)\n\n# This function takes subsetted gene association dataframe and extracts all GO ids\n# Maps GO ids to GO terms \n# Input: dataframe , dictionary with GO ids --> GO terms \n# Returns list of unique GO terms with GO ids e.g. ['GO term1 (GO:id1)', 'GO term2 (GO:id2)']\ndef get_GO_terms(df, lookup_table): \n process = []\n for x in df['GO ID']:\n try:\n name = lookup_table[x]\n process.append(name + ' ' +'('+x+')')\n except: \n pass\n unique_terms = np.unique(process)\n return(unique_terms)\n\n############################# ^ FOR GO ontology only ^ #############################\n\n\n\n\n\n############################## Upload ontology files ###############################\n\n# fly anatomy ontology file: http://www.obofoundry.org/ontology/fbbt.html \nfly_anatomy_ont = pronto.Ontology('/Users/maayanlab/Downloads/fbbt.obo.txt')\n# fly phenotype ontology: http://www.obofoundry.org/ontology/dpo.html\nfly_phenotype_ont = pronto.Ontology('/Users/maayanlab/Downloads/fbcv.obo.txt')\n\n# worm anatomy ontology file: http://www.obofoundry.org/ontology/wbbt.html\nworm_anatomy_ont = pronto.Ontology('/Users/maayanlab/Downloads/wbbt.owl')\n# worm phenptype file: http://www.obofoundry.org/ontology/wbphenotype.html\nworm_phenotype_ont = pronto.Ontology('/Users/maayanlab/Downloads/wbphenotype.obo.txt')\n\n# zebrafish anatomy ontology file: http://www.obofoundry.org/ontology/zfa.html\nzebrafish_anatomy_ont = pronto.Ontology('/Users/maayanlab/Downloads/zfa.obo.txt')\n# zebrafish phenotype ontology file: https://zfin.org/downloads/gene_expression_phenotype.txt\nzebrafish_phenotype_ont = pd.DataFrame.from_csv('/Users/maayanlab/Downloads/phenotype_fish_2018.10.19.txt',sep= '\\t', header=1).reset_index()\n\n# yeast cellular component ontology : http://www.geneontology.org/ontology/subsets/goslim_yeast.obo\nyeast_anatomy_ont = pronto.Ontology('/Users/maayanlab/Downloads/goslim_yeast.obo.txt')\n# yeast phenotype ontology file: https://www.yeastgenome.org/search?category=download&page=0&topic=Genotype+and+phenotype&year=2017\nyeast_phenotype_ont = pd.DataFrame.from_csv('phenotype_data.20170114.tab',sep= '\\t', header=None).reset_index()\n\n# GO ontology: http://snapshot.geneontology.org/ontology/go.obo\ngo_ont = pronto.Ontology('/Users/maayanlab/Downloads/go-basic.obo')\ngo_lookup = {}\nfor term in go_ont:\n go_lookup[term.id]=term.name\n \n############################# ^ Upload ontology files ^ ##############################\n \n\n\n\n\n\n############################## Parent terms of interest ##############################\n# parent terms of interest\nfly_anatomy_parents=[\n 'tagma',\n 'egg',\n 'embryo',\n 'organism',\n 'appendage',\n 'organ system',\n 'adipose system',\n 'circulatory system',\n 'digestive system',\n 'excretory system',\n 'integumentary system',\n 'muscle system',\n 'nervous system',\n 'reproductive system',\n 'tracheal system',\n 'portion of tissue',\n 'multi-tissue structure',\n 'sense organ',\n 'cell', \n 'acellular anatomical structure',\n 'developing material anatomical entity'\n ] \nworm_anatomy_parents=[\n 'axis',\n 'body region',\n 'cell component',\n 'dorsal nerve cord',\n 'extracellular component',\n 'ganglion',\n 'gubernaculum',\n 'lateral nerve cord',\n 'nerve ring',\n 'organ',\n 'organism',\n 'pharyngeal nerve',\n 'pharyngeal nerve process',\n 'pharyngeal segment',\n 'pharyngeal-intestinal valve',\n 'somatic gonad ',\n 'stoma',\n 'vas deferens',\n 'vas deferens valve region',\n 'ventral nerve cord'\n ]\nfly_phenotype_parents=[\n 'fertile',\n 'phenotype',\n 'viable',\n 'wild-type']\nworm_phenotype_parents=[\n 'body region pigmentation variant',\n 'cell development variant',\n 'cell morphology variant',\n 'cell physiology variant',\n 'cell pigmentation variant',\n 'electrophysiology variant',\n 'organ system development variant',\n 'organ system morphology variant',\n 'organ system physiology variant',\n 'organ system pigmentation variant',\n 'organism behavior variant',\n 'organism development variant',\n 'organism morphology variant',\n 'organism physiology variant',\n 'organism pigmentation variant',\n 'pericellular component development variant',\n 'pericellular component morphology variant',\n 'pericellular component physiology variant',\n 'population fitness variant'\n ]\n\n############################# ^ Parent terms of interest ^ ############################\n\n\n\n\n\n\n\n######################## Get terms to search PubMed ####################################\n \n# fly terms\nfly_anatomy_terms = [x for x in np.unique(find_children_terms(fly_anatomy_ont, fly_anatomy_parents))]\nfly_phenotype_terms = [x for x in np.unique(find_children_terms(fly_phenotype_ont, fly_phenotype_parents))]\n# fly gene and GO:terms --> GO Term processes names associated with fly\n# file: http://geneontology.org/gene-associations/gene_association.fb.gz\nfly_gene_associtation_biological_process = clean_data(pd.read_csv('/Users/maayanlab/Downloads/fb.gaf', sep = '`', skiprows=16, header=None),'P')[['Gene symbol', 'GO ID']]\nfly_biological_process_terms = get_GO_terms(fly_gene_associtation_biological_process, go_lookup)\n\n\n# worm terms\nworm_anatomy_terms = find_children_terms(worm_anatomy_ont, worm_anatomy_parents)\nworm_phenotype_terms = [x for x in np.unique(find_children_terms(worm_phenotype_ont, worm_phenotype_parents))]\n# worm gene and GO:terms --> GO Term processes names associated with worm\n# file: http://geneontology.org/gene-associations/gene_association.wb.gz\nworm_gene_associtation_biological_process = clean_data(pd.read_csv('/Users/maayanlab/Downloads/wb.gaf', sep = '~', skiprows=14, header=None),'P')[['Gene symbol', 'GO ID']]\nworm_biological_process_terms = get_GO_terms(worm_gene_associtation_biological_process, go_lookup)\n\n\n# zebrafish terms\nzebrafish_anatomy_terms=[]\nfor z in zebrafish_anatomy_ont:\n if z.name == 'zebrafish anatomical entity':\n zebrafish_anatomy_terms = [x.name for x in z.rchildren(level=3)]\nzebrafish_phenotype_terms = []\nfor i,x in enumerate(zebrafish_phenotype_ont['Affected Structure or Process 1 superterm Name']):\n zebrafish_phenotype_terms.append(x)\nzebrafish_phenotype_terms = np.unique(zebrafish_phenotype_terms)\n# zebrafish gene and GO:terms --> GO Term processes names associated with zebrafish\n# file: http://geneontology.org/gene-associations/gene_association.zfin.gz\nzebrafish_gene_associtation_biological_process = clean_data(pd.read_csv('/Users/maayanlab/Downloads/zfin.gaf', sep = '~', skiprows=16, header=None),'P')[['Gene symbol', 'GO ID']]\nzebrafish_biological_process_terms = get_GO_terms(zebrafish_gene_associtation_biological_process, go_lookup)\n\n\n# yeast terms \nyeast_anatomy_terms=[]\nfor y in yeast_anatomy_ont:\n if y.name == 'cellular_component':\n yeast_anatomy_terms = [x.name for x in y.rchildren(level=-1)]\nyeast_phenotype_terms = np.unique([x.split(':')[0] for x in yeast_phenotype_ont.loc[:,9]])\n# yeast gene and GO:terms --> GO Term processes names associated with yeast\n# file: http://geneontology.org/gene-associations/gene_association.sgd.gz\nyeast_gene_associtation_biological_process = clean_data(pd.read_csv('/Users/maayanlab/Downloads/sgd.gaf', sep = '~', skiprows=18, header=None),'P')[['Gene symbol', 'GO ID']]\nyeast_biological_process_terms = get_GO_terms(yeast_gene_associtation_biological_process, go_lookup)\n\n####################### ^ Get terms to search PubMed ^ ###################################\n\n\n\n\n\n############################## Get PubMed Results #########################################\n\n\n# get pubmed results, perform back propogation, and write to file \ndef pubmed_results_to_file(organism,terms, type_terms,ontology_file, backpropogate):\n ids = EQ.get_pubmed_ids(organism,[x.split(' (')[0] for x in terms])# GO Terms in parentheses must be removed for term search \n if backpropogate =='yes':\n ids = back_propogation(ontology_file,terms, ids)\n writefile = open(organism.replace(' ', '_')+'_'+type_terms+'_pubMed_Ids.txt',\"w\")\n for i,x in enumerate (terms):\n if len(ids[i])>0:\n ids2 = np.unique(ids[i])\n writefile.write(str(x).replace(' ','_')+'\\t'+'\\t'.join([str(k) for k in ids2])+'\\n')\n \n\n#pubmed_results_to_file('Drosophila melanogaster',fly_anatomy_terms,'anatomy',fly_anatomy_ont,'yes')\n#pubmed_results_to_file('Drosophila melanogaster',fly_phenotype_terms,'phenotype',fly_phenotype_ont,'yes')\n#pubmed_results_to_file('Drosophila melanogaster',fly_biological_process_terms,'GO_biological_process',go_ont,'yes')\n \n#pubmed_results_to_file('Caenorhabditis elegans',worm_anatomy_terms,'anatomy',worm_anatomy_ont,'yes')\n#pubmed_results_to_file('Caenorhabditis elegans',worm_phenotype_terms,'phenotype',worm_phenotype_ont,'yes') \n#pubmed_results_to_file('Caenorhabditis elegans',worm_biological_process_terms,'GO_biological_process',go_ont,'yes') \n \n#pubmed_results_to_file('Danio_rerio',zebrafish_anatomy_terms,'anatomy',zebrafish_anatomy_ont,'yes')\n#pubmed_results_to_file('Danio_rerio',zebrafish_phenotype_terms,'phenotype',zebrafish_phenotype_ont,'no') # no back propogation with a simple list of terms (no ontology tree)\n#pubmed_results_to_file('Danio_rerio',zebrafish_biological_process_terms,'GO_biological_process',go_ont,'yes') \n \n#pubmed_results_to_file('Saccharomyces_cerevisiae',yeast_anatomy_terms,'cellular_component',yeast_anatomy_ont,'yes')\n#pubmed_results_to_file('Saccharomyces_cerevisiae',yeast_phenotype_terms,'phenotype',yeast_phenotype_ont,'no') # no back propogation with a simple list of terms (no ontology tree)\n#pubmed_results_to_file('Saccharomyces_cerevisiae',yeast_biological_process_terms,'GO_biological_process',go_ont,'yes')\n\n############################## ^ Get PubMed Results ^ ######################################### \n","repo_name":"MaayanLab/GeneRif_lists","sub_path":"terms_to_pubMed_ids.py","file_name":"terms_to_pubMed_ids.py","file_ext":"py","file_size_in_byte":14419,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"84"}
+{"seq_id":"39478479953","text":"from django import forms\nfrom django.forms import widgets\nfrom .models import Topic, Entry\n\nclass TopicForm(forms.ModelForm):\n \"\"\" Build forms to enter the information by the user.\"\"\"\n class Meta:\n model = Topic\n fields = ['text']\n labels = {'text': ''}\n\nclass EntryForm(forms.ModelForm):\n class Meta:\n model = Entry\n fields = ['text']\n labels = {'text': 'Entry:'}\n widgets = {'text': forms.Textarea(attrs={'cols': 80})}","repo_name":"das88768/Learning_log","sub_path":"learning_logs/forms.py","file_name":"forms.py","file_ext":"py","file_size_in_byte":478,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"84"}
+{"seq_id":"8897925113","text":"import os\nimport random\n\nfrom services.keyboard_service import KeyboardService\nfrom services.video_service import VideoService\nfrom directing.director import Director\nfrom shared.point import Point\nfrom casting.cast import Cast\nfrom casting.actor import Actor\nfrom casting.artifact import Artifact\nfrom shared.color import Color\n\n\nFRAME_RATE = 12\nMAX_X = 900\nMAX_Y = 600\nCELL_SIZE = 15\nFONT_SIZE = 15\nCOLS = 60\nROWS = 40\nCAPTION = \"Greed\"\nDATA_PATH = os.path.dirname(os.path.abspath(__file__)) + \"/data/messages.txt\"\nWHITE = Color(255, 255, 255)\nDEFAULT_ARTIFACTS = 40\n\n\ndef main():\n \n # create the cast\n cast = Cast()\n \n # create the banner\n banner = Actor()\n banner.set_text(\"\")\n banner.set_font_size(FONT_SIZE)\n banner.set_color(WHITE)\n banner.set_position(Point(CELL_SIZE, 0))\n cast.add_actor(\"banners\", banner)\n \n # create the robot\n x = int(MAX_X / 2)\n y = int(MAX_Y / 2)\n position = Point(x, y)\n\n robot = Actor()\n robot.set_text(\"#\")\n robot.set_font_size(FONT_SIZE)\n robot.set_color(WHITE)\n robot.set_position(position)\n cast.add_actor(\"robots\", robot)\n\n \n\n for n in range(ROWS):\n\n x = random.randint(0, MAX_X)\n y = MAX_Y\n position = Point(x, y)\n position = position.scale(CELL_SIZE)\n\n r = random.randint(0, 255)\n g = random.randint(0, 255)\n b = random.randint(0, 255)\n color = Color(r, g, b)\n \n if n % 2 == 0 :\n gems = Artifact()\n gems.set_font_size(FONT_SIZE)\n gems.set_color(color)\n gems.set_position(position)\n cast.add_actor(\"gems\", gems)\n \n else :\n rocks = Artifact()\n rocks.set_font_size(FONT_SIZE)\n rocks.set_color(color)\n rocks.set_position(position)\n cast.add_actor(\"rocks\", rocks)\n \n # start the game\n keyboard_service = KeyboardService(CELL_SIZE)\n video_service = VideoService(CAPTION, MAX_X, MAX_Y, CELL_SIZE, FRAME_RATE)\n director = Director(keyboard_service, video_service)\n director.start_game(cast)\n\n\nif __name__ == \"__main__\":\n main()","repo_name":"JakeZalesny/CSE210-04","sub_path":"__main__.py","file_name":"__main__.py","file_ext":"py","file_size_in_byte":2150,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"73631490833","text":"import sqlite3\n\nfrom create_folders import Folder\n\n\nclass Database:\n def __init__(self):\n pass\n\n def create_connection(self, db_file):\n conn = None\n conn = sqlite3.connect(db_file)\n return conn\n\n def create_table(self, db_file):\n sql_create_crypto_table = \"\"\"CREATE TABLE IF NOT EXISTS crypto (\n id real PRIMARY KEY,\n DATE text,\n CRYPTO text,\n TREND text,\n ROC real,\n PURCHASE real,\n UNITS real,\n EXCHANGE real,\n PRED_LABEL text,\n STAT_LABEL text,\n ACC_BALANCE real,\n PROFIT real,\n SPOT_PRICE real,\n NATIVE_BALANCE real\n\n ); \"\"\"\n file_path = Folder().create_folder_link(\"SQLDatabase\", db_file)\n conn = self.create_connection(file_path)\n if conn is not None:\n conn.execute(sql_create_crypto_table)\n conn.commit()\n conn.close()\n\n else:\n print(\"Error! cannot create the database connection.\")\n\n def create_entry_crypto(self, db_file, data):\n sql = \"\"\" INSERT INTO crypto (id,DATE,CRYPTO,TREND, ROC, PURCHASE,UNITS, EXCHANGE, PRED_LABEL, STAT_LABEL, ACC_BALANCE, PROFIT, SPOT_PRICE, NATIVE_BALANCE)\n VALUES(?,?,?,?,?,?,?,?,?,?,?,?,?,?) \"\"\"\n file_path = Folder().create_folder_link(\"SQLDatabase\", db_file)\n conn = self.create_connection(file_path)\n if conn is not None:\n c = conn.execute(sql, data)\n conn.commit()\n conn.close()\n return c.lastrowid\n\n else:\n print(\"Error! cannot create the database connection.\")\n\n def main(self):\n # Code for creating database connection and table\n print(\"Creating database connection and table...\")\n","repo_name":"x4rMa/Crypto-Trading-Bot","sub_path":"database.py","file_name":"database.py","file_ext":"py","file_size_in_byte":2248,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"69945191635","text":"import json\nfrom typing import Union\nclass SettingsNotLocked(Exception):\n pass\n\nclass Settings:\n def __init__(self, js: dict) -> None:\n self._js = js\n print(self._js)\n \n def set_sys_channel(self, guild_id: str, chan_id: int) -> None:\n if guild_id in self._js:\n self._js[guild_id][\"systemchannel\"] = chan_id\n else:\n self._js[guild_id] = {\"systemchannel\": chan_id}\n \n def get_sys_channel(self, guild_id: str) -> Union[int, None]:\n try:\n return self._js[guild_id][\"systemchannel\"]\n except (IndexError, KeyError):\n return None\n def set_ann_channel(self, guild_id: str, chan_id: int) -> None:\n if guild_id in self._js:\n self._js[guild_id][\"announcementchannel\"] = chan_id\n else:\n self._js[guild_id] = {\"announcementchannel\": chan_id}\n \n def get_greet(self, guild_id: str) -> Union[bool, None]:\n try:\n return self._js[guild_id][\"greet\"]\n except (IndexError, KeyError):\n return None\n def set_greet(self, guild_id: str, greet: bool) -> None:\n if guild_id in self._js:\n self._js[guild_id][\"greet\"] = greet\n else:\n self._js[guild_id] = {\"greet\": greet}\n \n def get_mute_everyone(self, guild_id: str) -> Union[bool, None]:\n try:\n return self._js[guild_id][\"mute\"]\n except (IndexError, KeyError):\n return None\n def set_mute_everyone(self, guild_id: str, greet: bool) -> None:\n if guild_id in self._js:\n self._js[guild_id][\"mute\"] = greet\n else:\n self._js[guild_id] = {\"mute\": greet}\n \n def add_reactor_channel(self, guild_id: str, msg_id: str, reactions: dict) -> None:\n if guild_id in self._js:\n try:\n self._js[guild_id][\"reactors\"][msg_id] = reactions\n except KeyError:\n self._js[guild_id][\"reactors\"] = {msg_id: reactions}\n else:\n self._js[guild_id] = {\"reactors\": {msg_id: reactions}}\n \n def get_reaction_role(self, guild_id: str, msg_id: str, emoji: str) -> Union[str, None]:\n try:\n return self._js[guild_id][msg_id][emoji]\n except (IndexError, KeyError):\n print(\"8h\", self._js, guild_id, msg_id, emoji)\n return None\n \n def get_ann_channel(self, guild_id: str) -> Union[int, None]:\n try:\n return self._js[guild_id][\"announcementchannel\"]\n except (IndexError, KeyError):\n return None\n def get(self) -> dict:\n ret = self._js \n del self\n return ret\nclass SettingsManager:\n def __init__(self) -> None:\n self.filename = \"settings.json\"\n self.lock = None\n \n def __enter__(self) -> Settings:\n while self.lock is not None:\n pass\n self.lock = Settings(json.load(open(self.filename))[\"settings\"])\n return self.lock\n \n def __exit__(self, *_args) -> None:\n json.dump({\"settings\": self.lock.get()}, open(self.filename, \"w\"))\n self.lock = None","repo_name":"ayushashi11/balak","sub_path":"settings_manager.py","file_name":"settings_manager.py","file_ext":"py","file_size_in_byte":3103,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"70432226836","text":"import sys\ninput = sys.stdin.readline\n\n\ndef bin(s, e):\n if s == e:\n if a[s] > -a[i]:\n return s\n else:\n return s+1\n m = (s + e)>>1\n if a[m] > -a[i]:\n return bin(s, m)\n else:\n return bin(m+1, e)\n\n\nn = int(input())\na = list(map(int, input().split()))\nb = list(map(int, input().split()))\nfor i in range(n):\n a[i] -= b[i]\ncnt = 0\na.sort()\nidx1 = 0\nidx2 = 1\nfor i in range(n-1):\n cnt += n - bin(i+1, n-1)\nprint(cnt)\n","repo_name":"PyeongGang-Kim/TIL","sub_path":"algorithm/cf/3.12/d.py","file_name":"d.py","file_ext":"py","file_size_in_byte":478,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"3086198779","text":"#!/usr/bin/python\n# -*- coding: UTF-8 -*-\n\ninversion_num = 0\n\n\ndef merge_sort_counting(nums, start, end):\n if start >= end:\n return\n\n mid = (start + end)//2\n merge_sort_counting(nums, start, mid)\n merge_sort_counting(nums, mid+1, end)\n merge(nums, start, mid, end)\n\n\ndef merge(nums, start, mid, end):\n global inversion_num\n i = start\n j = mid+1\n tmp = []\n while i <= mid and j <= end:\n if nums[i] <= nums[j]:\n inversion_num += j - mid - 1\n tmp.append(nums[i])\n i += 1\n else:\n tmp.append(nums[j])\n j += 1\n\n while i <= mid:\n # 这时nums[i]的逆序数是整个nums[mid+1: end+1]的长度\n inversion_num += end - mid\n tmp.append(nums[i])\n i += 1\n\n while j <= end:\n tmp.append(nums[j])\n j += 1\n\n nums[start: end+1] = tmp\n\n\nif __name__ == '__main__':\n print('--- count inversion number using merge sort ---')\n # nums = [5, 0, 4, 2, 3, 1, 6, 8, 7]\n nums = [5, 0, 4, 2, 3, 1, 3, 3, 3, 6, 8, 7]\n print('nums : {}'.format(nums))\n merge_sort_counting(nums, 0, len(nums)-1)\n print('sorted: {}'.format(nums))\n print('inversion number: {}'.format(inversion_num))\n","repo_name":"wangzheng0822/algo","sub_path":"python/38_divide_and_conquer/merge_sort_counting.py","file_name":"merge_sort_counting.py","file_ext":"py","file_size_in_byte":1230,"program_lang":"python","lang":"en","doc_type":"code","stars":22332,"dataset":"github-code","pt":"84"}
+{"seq_id":"70598072276","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations\n\n\ndef run(apps, schema_editor):\n DashboardBookmark = apps.get_model(\"geosight_data\", \"DashboardBookmark\")\n DashboardIndicatorLayerIndicator = apps.get_model(\n \"geosight_data\", \"DashboardIndicatorLayerIndicator\")\n\n for bookmark in DashboardBookmark.objects.all():\n dashboard = bookmark.dashboard\n selected_indicators = bookmark.selected_indicators.all()\n for selected_indicator in selected_indicators:\n layers = dashboard.dashboardindicatorlayer_set.all()\n for layer in layers:\n layer_indicators = layer.dashboardindicatorlayerindicator_set\n try:\n selected_layer = layer_indicators.all().get(\n indicator__id=selected_indicator.id\n )\n bookmark.selected_indicator_layer = selected_layer.object\n bookmark.save()\n except DashboardIndicatorLayerIndicator.DoesNotExist:\n pass\n\n\nclass Migration(migrations.Migration):\n dependencies = [\n ('geosight_data', '0033_dashboardbookmark_selected_indicator_layer')\n ]\n\n operations = [\n migrations.RunPython(run, migrations.RunPython.noop),\n ]\n","repo_name":"dimasciput/GeoSight","sub_path":"django_project/geosight/data/migrations/0034_dashboard_bookmark_migrations.py","file_name":"0034_dashboard_bookmark_migrations.py","file_ext":"py","file_size_in_byte":1322,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"75080543634","text":"def determine(maxNumber, numbers):\n numbersSpoken = {}\n prevNumbersSpoken = {}\n\n for number in numbers:\n numbersSpoken[int(number)] = numbers.index(number) + 1\n numberSpoken = numbers[-1]\n\n for i in range(len(numbers), maxNumber):\n if numberSpoken in prevNumbersSpoken.keys():\n numberSpoken = i - prevNumbersSpoken[numberSpoken]\n else:\n numberSpoken = 0\n\n if numberSpoken in numbersSpoken.keys():\n prevNumbersSpoken[numberSpoken] = numbersSpoken[numberSpoken]\n numbersSpoken[numberSpoken] = i + 1\n\nnumbers = []\ninput = open(\"input.txt\")\n\nfor line in input:\n for char in line.strip().split(','):\n numbers.append(int(char))\n\ndetermine(2020, numbers)\ndetermine(30000000, numbers)\n","repo_name":"Jorissen9/AoC","sub_path":"Day 15/rambunctiousRecitation.py","file_name":"rambunctiousRecitation.py","file_ext":"py","file_size_in_byte":770,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"13442849418","text":"def pattern(n):\r\n for i in range(1,n+1):\r\n if i%2!=0:\r\n k=i+1\r\n else:\r\n k=i\r\n for j in range(k,n):\r\n print(end=\" \")\r\n for g in range(k):\r\n print(\"*\",end=\" \")\r\n print(\"\\n\")\r\n\r\n\r\nn=int(input(\"Enter the number of rows: \"))\r\npattern(n)\r\n \r\n \r\n","repo_name":"sushrao1996/CompetitiveProgramming_Beginner","sub_path":"Day2_Challenge.py","file_name":"Day2_Challenge.py","file_ext":"py","file_size_in_byte":335,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"2237877855","text":"'''\nCreated on 21 Sep 2015\n\n@author: sm6110\n\nAdapted from BeamLib/src/optimiser/wrapper/pysrc/lib.read\n\n'''\n\nimport h5py\nfrom warnings import warn\n\n\ndef h5file(filename,ReadList=None):\n '''\n Read entries of a HDF5 file and attaches them to class.\n Groups are saved as sub-classes, while dataset values are saved as class \n attributes\n \n Important: though this method allows to read all the input/output classes \n required to run the aeroelastic solutions, ctypes variable are not saved as\n such!\n \n If a ReadList=None all groups/attributes are read. \n If ReadList is given, the following format is used:\n \n ReadList = ['name']\n \n if name is an attribute, this is read as such.\n if name refers to a group, all entries of the group are read\n if name = 'grp/attr' only the attribute attr of the group grp is read\n \n '''\n \n class H: pass\n Hinst=H()\n \n # read and scan file\n hdfile=h5py.File(filename,'r')\n NamesList=[] # dataset names\n hdfile.visit(NamesList.append)\n \n # find SubList (required in case a class has to be read)\n if ReadList is None:\n SubList=NamesList\n else:\n SubList=[]\n for name in NamesList:\n for readname in ReadList:\n if readname in name:\n SubList.append(name)\n\n \n for name in SubList:\n #print('Found %s...'%name)\n if type(hdfile[name]) is h5py._hl.group.Group: # @UndefinedVariable\n #print(' %s is a group!' %name)\n if hasattr(Hinst,'name') is False:\n setattr(Hinst,name,H())\n else:\n if '/' in name:\n subnames=name.split('/')\n if hasattr(Hinst,subnames[0]) is False:\n setattr(Hinst,subnames[0],H())\n #print(' Extracting class %s'%subnames[0])\n subclass=getattr(Hinst,subnames[0])\n #print(' Allocating attribute %s'%subnames[1])\n setattr(subclass,subnames[1],hdfile[name].value)\n #print(' copying subclass %s back'%subnames[0])\n setattr(Hinst,subnames[0],subclass)\n else:\n setattr(Hinst,name,hdfile[name].value)\n\n '''\n #setattr(XBinst,'FollowerForce',hdfile['FollowerForce'].value)\n\n # parameters for constant beam span reconstruction \n XBinst = conditional_reading(hdfile,XBinst,'cs_l2')\n\n setattr(XBinst,'',hdfile[''].value)\n XBinst = conditional_reading(hdfile,XBinst,'')\n '''\n \n # close and return\n hdfile.close() \n \n a=1\n \n return Hinst \n \n\n \ndef h5series(rootname,ReadList=None,N0=0):\n ''' \n Given a list of datasets, creates a list of lists for all the solutions\n run for a DOE or optimisation. the output is in a list of classes, each\n containing all the attributes/groups read.\n \n See h5file for ReadList format\n '''\n \n outlist=[]\n \n cc=N0\n go_on=True\n \n while go_on is True:\n cc_str = '%.3d' % (cc)\n filename = rootname + cc_str + '.h5'\n try:\n print( 'Reading: %s' %(filename) )\n Hinst = h5file(filename,ReadList)\n outlist.append(Hinst)\n cc=cc+1\n except OSError:\n print( '%s not found. %s files read in total!' %(filename,cc_str) )\n go_on=False\n return outlist\n \n \n\n\ndef h5list(fileslist,ReadList=None):\n ''' \n Equivalent to h5series but reads files from an user defined list (fileslist)\n \n All the attributes in 'attrlist' are read and stored in outlist.\n '''\n \n outlist=[]\n \n for filename in fileslist:\n \n print( 'Reading: %s' %(filename) )\n Hinst = h5file(filename,ReadList)\n outlist.append(Hinst)\n\n return outlist\n\n\n\n\n\ndef conditional_reading(hdfile,obj,fieldname): \n ''' \n Given a hdf5 file 'hdfile' and the object obj, the routine:\n a. if the field field-name is found and has a value, assigns it to the\n attribute obj.field-name. \n b. does nothing otherwise \n '''\n \n try:\n val = hdfile[fieldname].value\n if val!='not found' and val!='no value':\n setattr(obj,fieldname,val)\n except:\n warn('Attribute \"%s\" not found!!!' %fieldname)\n \n return obj \n\n\n\n\ndef collect(Hlist,ReadList):\n '''\n Given a list of classes Hlist given in output by h5series or h5list methods,\n the function extracts specific attributes from each element of Hlist and\n assigns the related values to a bunch of output variables.\n \n If ReadList contains subclasses, these will be saved as such.\n \n @warning: method not working with nested subclasses\n '''\n \n \n Nlist = len(Hlist) \n Nattr = len(ReadList)\n \n ValList = []\n \n for ii in range(Nattr):\n \n name = ReadList[ii]\n if '/' in name:\n subnames=name.split('/')\n Nlevels = len(subnames)\n subclass_list = [getattr(Hlist[hh],subnames[0]) for hh in range(Nlist)] \n for cc in range(1,Nlevels):\n subclass_list = [ getattr(subclass_list[hh],subnames[cc]) for hh in range(Nlist) ] \n ValList.append( subclass_list ) \n else:\n ValList.append( getattr(Hlist[hh],name) for hh in range(Nlist) ) \n \n \n return tuple(ValList)\n \n\n\n\n\nif __name__=='__main__':\n \n # read one file \n filename='/home/sm6110/git/SHARPy_studies/aerocomp/hale/opt/res_rig/opt_hale_noyaw_sig100_Umag0_fi4_Umagflow25_alpha0_010.h5'\n hd=h5file(filename,['cost','gdis','geq/val']) \n \n # read a series\n #ReadList=['Design','gdis/val','gdis/jac']\n ReadList=None\n fileroot='/home/sm6110/git/SHARPy_studies/aerocomp/hale/opt/res_rig/opt_hale_noyaw_sig100_Umag0_fi4_Umagflow25_alpha0_'\n Hseries=h5series(fileroot,ReadList)\n \n # or a list of files\n FileList=['/home/sm6110/git/SHARPy_studies/aerocomp/hale/opt/res_rig/opt_hale_noyaw_sig100_Umag0_fi4_Umagflow25_alpha0_001.h5',\n '/home/sm6110/git/SHARPy_studies/aerocomp/hale/opt/res_rig/opt_hale_noyaw_sig100_Umag0_fi4_Umagflow25_alpha0_008.h5',\n ]\n \n # and collect them\n ValList = collect( Hseries, ['Design/x', 'cost/val', 'cost/jac'] )\n \n a=1\n\n \n\n\n\n","repo_name":"SalvatoreMaraniello/SHARPy","sub_path":"src/PyLibs/io/read.py","file_name":"read.py","file_ext":"py","file_size_in_byte":6356,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"84"}
+{"seq_id":"23700306468","text":"from flask import Flask, flash, render_template, request, redirect, url_for\nfrom flask_sqlalchemy import SQLAlchemy\nfrom flask_cors import CORS, cross_origin\nfrom flask_limiter import Limiter\nfrom flask_limiter.util import get_remote_address\nfrom pathlib import Path\n# from werkzeug.utils import secure_filename\nimport psycopg2\nimport os\n\napp = Flask(__name__)\n\nlimiter = Limiter(\n app,\n key_func=get_remote_address,\n default_limits=[\"2000 per day\", \"1000 per hour\"],\n storage_uri=\"memory://\",\n)\n\nCORS(app, support_credentials=True)\n\nDEVELOPMENT = 'dev'\nPRODUCTION = 'prod'\n\nENV = DEVELOPMENT\n\nconn = psycopg2.connect(database=\"share_db\", user=\"postgres\",\n password=\"12345\", host=\"localhost\")\n# print(\"connected to the database\")\n# print(\"unable to connect to the database\")\nmycursor = conn.cursor()\n\nif ENV == DEVELOPMENT:\n app.debug = True\n app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://postgres:12345@localhost/share_db'\nelse:\n # remote appdatabase uri\n app.config['SQLALCHEMY_DATABASE_URI'] = ''\n app.debug = False\n\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False\n\napp.config['UPLOAD_FOLDER'] = 'uploaded_files'\nif not os.path.exists('uploaded_files'):\n os.mkdir('uploaded_files')\n\nALLOWED_EXTENSIONS = {'txt', 'pdf', 'png', 'jpg', 'jpeg', 'gif'}\napp.config['MAX_CONTENT_PATH'] = 30 * 1000 * 1000 # 30mb\n\n\ndef allowed_file(filename):\n return '.' in filename and filename.rsplit('.', 1)[1].lower() in ALLOWED_EXTENSIONS\n\n\ndef file_upload(filerequest, name: str):\n if 'file' not in filerequest.files:\n flash('No file part')\n return redirect(filerequest.url)\n\n print(filerequest.files)\n file = filerequest.files['file']\n # If the user does not select a file, the browser submits an\n # empty file without a filename.\n if file.filename == '':\n flash('No selected file')\n return redirect(filerequest.url)\n if file and allowed_file(file.filename):\n # filename = secure_filename(file.filename)\n filename = name + \".\" + file.filename.rsplit('.', 1)[1].lower()\n file.save(os.path.join(app.config['UPLOAD_FOLDER'], filename))\n return redirect(url_for('file', name=filename))\n\n\ndb = SQLAlchemy(app)\n\n\nclass RideShare(db.Model):\n __tablename__ = 'rideshare'\n id = db.Column(db.Integer, primary_key=True)\n name = db.Column(db.String(200), unique=True)\n start = db.Column(db.String(200))\n end = db.Column(db.String(200))\n date = db.Column(db.String(12))\n time = db.Column(db.String(12))\n request = db.Column(db.Integer) # 0 for request 1 for offer\n phone_number = db.Column(db.String(200))\n no_of_people = db.Column(db.Integer)\n\n def __init__(self, name, start, end, date, time, request, phone_number, no_of_people):\n self.name = name\n self.start = start\n self.end = end\n self.date = date\n self.time = time\n self.request = request\n self.phone_number = phone_number\n self.no_of_people = no_of_people\n\n\nclass nearby_events(db.Model):\n __tablename__ = 'rideshare'\n id = db.Column(db.Integer, primary_key=True)\n name = db.Column(db.String(200), unique=True)\n start = db.Column(db.String(200))\n end = db.Column(db.String(200))\n date = db.Column(db.String(12))\n time = db.Column(db.String(12))\n\n def __init__(self, name, start, end, date, time, request, phone_number, no_of_people):\n self.name = name\n self.start = start\n self.end = end\n self.date = date\n self.time = time\n\n\nclass BookShare(db.Model):\n __tablename__ = 'bookshare'\n id = db.Column(db.Integer, primary_key=True)\n name = db.Column(db.String(200), unique=True)\n author = db.Column(db.String(200))\n contact_details = db.Column(db.String(200))\n description = db.Column(db.String(200))\n\n def __init__(self, name, author, contact_details, description):\n self.name = name\n self.author = author\n self.contact_details = contact_details\n self.description = description\n\n\nclass Feed(db.Model):\n __tablename__ = 'posts'\n\n id = db.Column(db.Integer, primary_key=True)\n text = db.Column(db.String(200))\n name = db.Column(db.String(200))\n\n def __init__(self, text, name):\n self.text = text\n self.name = name\n\n\nclass NotesShare(db.Model):\n __tablename__ = 'notes'\n id = db.Column(db.Integer, primary_key=True)\n name = db.Column(db.String(200))\n user_name = db.Column(db.String(200))\n description = db.Column(db.String(200))\n course = db.Column(db.String(200))\n downloads = db.Column(db.Integer)\n\n def __init__(self, name, user_name, description, course, downloads):\n self.name = name\n self.user_name = user_name\n self.description = description\n self.course = course\n self.downloads = downloads\n\n\n@app.route('/')\n@cross_origin(supports_credentials=True)\ndef index():\n mycursor.execute(\"SELECT * FROM posts\")\n posts = mycursor.fetchall()\n if request.args.get(\"show_webpage\") == \"1\":\n return render_template('home.html', posts=posts)\n\n allowed_types = [\"png\", \"jpg\", \"jpeg\", \"gif\", \"heif\", \"heic\"]\n for i in posts:\n for type in allowed_types:\n uploaded_images = Path(f\"/uploaded_files/feed_{i[0]}.{type}\")\n if uploaded_images.exists():\n file_path = f\"/uploaded_files/feed_{i[0]}.{type}\"\n posts.append(file_path)\n break\n\n return posts\n\n\n@app.route('/submit_text', methods=['POST'])\n@cross_origin(supports_credentials=True)\ndef submit_text():\n if request.method == 'POST':\n text = request.json['text']\n name = request.json['name']\n # print(text)\n # text = request.form['text']\n if \"file\" in request.files:\n file_upload(request, \"feed_\" + str(db.session.query(Feed.id > 0).count()))\n\n if text == \"\":\n return {\"error\": \"Empty Text field\"}\n if name == \"\":\n return {\"error\": \"Empty name field\"}\n else:\n data = Feed(text, name)\n db.session.add(data)\n db.session.commit()\n return {\"message\": \"Success\"}\n\n\n############rideshare#################################################\n\n\n@app.route('/rideshare_form')\n@cross_origin(supports_credentials=True)\ndef request_ride():\n requests = int(request.args.get('req_type'))\n return render_template('ride.html', request_=requests)\n\n\n# returns available rides from database\n@app.route('/display_rides')\n@cross_origin(supports_credentials=True)\ndef rides():\n mycursor.execute(\"SELECT * FROM rideshare\")\n rides = mycursor.fetchall()\n requests = int(request.args.get('req_type'))\n return {\"rides\": rides, \"request_type\": requests}\n\n\n# stores submitted files from the json to db\n@app.route('/submit_ride', methods=['POST'])\n@cross_origin(supports_credentials=True)\ndef submit_ride():\n if request.method == 'POST':\n name = request.json[\"name\"]\n start = request.json[\"start\"]\n end = request.json[\"end\"]\n date = request.json[\"date\"]\n # print(date, type(date))\n time = request.json[\"time\"]\n # print(time, type(time))\n\n no_of_people = int(request.json[\"no_of_people\"]) if request.json[\"no_of_people\"] != \"\" else 0\n try:\n request_ = request.json[\"request\"]\n request_ = 1\n except:\n request_ = request.json[\"offer\"]\n request_ = 0\n\n phone_number = (request.json[\"phone_number\"]).replace(\"-\", \"\")\n # print(customer,dealer,rating,comment)\n if name == \"\" or start == \"\" or end == \"\" or date == \"\" or time == \"\" or no_of_people == 0 or phone_number == \"\":\n return {\"message\": \"One or more empty fields\"}\n if db.session.query(RideShare).filter(RideShare.phone_number == phone_number).count() == 0:\n data = RideShare(name, start, end, date, time, request_, phone_number, no_of_people)\n db.session.add(data)\n db.session.commit()\n return {\"message\": \"success\"}\n return {\"message\": \"Already submitted\"}\n\n\n#######################################rideshare####################################\n#######################################request_book################################\n\n\n@app.route('/share_book')\n@cross_origin(supports_credentials=True)\ndef share_book():\n return render_template('book.html', request_=False)\n\n\n@app.route('/display_books')\n@cross_origin(supports_credentials=True)\ndef books():\n mycursor.execute(\"SELECT * FROM bookshare\")\n all_books = mycursor.fetchall()\n books = []\n for i in all_books:\n books.append({\"name\": i[1], \"author\": i[2], \"contact\": i[3], \"description\": i[4]})\n return books\n\n\n@app.route('/submit_book', methods=['POST'])\n@cross_origin(supports_credentials=True)\ndef submit_book():\n if request.method == 'POST':\n name = request.json[\"name\"]\n author = request.json[\"author\"]\n contact_details = request.json[\"contact\"]\n description = request.json[\"description\"]\n # print(date, type(date))\n # print(time, type(time))\n try:\n request_ = request.json[\"request\"]\n request_ = 1\n except:\n request_ = request.json[\"offer\"]\n request_ = 0\n # print(customer,dealer,rating,comment)\n if name == \"\" or author == \"\" or contact_details == \"\":\n return {\"message\": \"one or more empty fields\"}\n else:\n data = BookShare(name, author, contact_details, description)\n db.session.add(data)\n db.session.commit()\n return {\"message\": \"success\"}\n\n\n##################################books-end#########################################################\n##################################noteshare#########################################################\n@app.route('/share_note', methods=['POST', 'GET'])\n@cross_origin(supports_credentials=True)\n@limiter.limit('50 per day')\ndef share_notes():\n if request.method == 'POST':\n name = request.form[\"name\"]\n user_name = request.form[\"username\"]\n description = request.form[\"description\"]\n course = request.form[\"course\"]\n downloads = request.form[\"downloads\"]\n check = [name, user_name, course, description, downloads]\n if \"\" in check:\n return {\"message\": \"one or more empty fields: name,username,course,description,downloads,file\"}\n else:\n data = NotesShare(name, user_name, description, course, downloads)\n db.session.add(data)\n db.session.commit()\n file_upload(request, str(db.session.query(NotesShare.id > 0).count()))\n return {\"message\": \"success\"}\n\n\n@app.route('/display_notes')\n@cross_origin(supports_credentials=True)\ndef notes():\n mycursor.execute(\"SELECT * FROM notes\")\n all_notes = mycursor.fetchall()\n\n return all_notes\n\n\n@app.route('/upload_note', methods=['POST', 'GET'])\n@cross_origin(supports_credentials=True)\ndef file():\n return render_template(\"success.html\")\n\n\n#################################noteshare-end###################################################\n@app.route('/submit_file', methods=['POST', 'GET'])\n@cross_origin(supports_credentials=True)\n@limiter.limit('50 per day')\ndef file_submit():\n if request.method == 'POST':\n # check if the post request has the file part\n return file_upload(request, \"cat02\")\n\n\nif __name__ == \"__main__\":\n try:\n app.run(host=\"150.243.207.105\", port=8000)\n except:\n app.run()\n","repo_name":"shibampokerail/trushare_backend","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":11588,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"20220081229","text":"from timeit import Timer\n\ndef divideBy2(num):\n num_lst = []\n while num > 0:\n num, rem = divmod(num, 2)\n return num_lst\n\nx = 0\nt1 = Timer(\"divideBy2(x)\", \"from __main__ import divideBy2, x\")\nprint(t1.timeit(number=1000))\n","repo_name":"jddelia/algos_and_ds","sub_path":"Section3/test.py","file_name":"test.py","file_ext":"py","file_size_in_byte":236,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"31619120404","text":"#!/usr/bin/python3\n#\nimport logging\nimport sys\nfrom libSensor import getBattery_th1\nfrom libSQLite import SQL\n\n#settings\nsys.path.append('/boot')\nfrom SAST_settings import SENS, AMBIENT, IFTTT, GAS, SAST_NAME, MC_TEMP\n\n###### Logging\nlogging.basicConfig(format='%(asctime)s %(message)s',level=logging.INFO)\n\nif __name__ == '__main__':\n ## --- バッテリーステータス更新\n macs = []\n S=SQL()\n for mac in SENS:\n macs.append(mac['MAC'])\n logging.info(f\"get Battery STASUS MAC={macs}\") \n\n batts = getBattery_th1( macs )\n for batinfo in batts:\n logging.info(f\" {batinfo['MAC']} {batinfo['battery']}% {batinfo['rssi']}dBm\") \n S.updateBattery( batinfo )\n\n","repo_name":"semit-takahashi/semit_ast","sub_path":"sastV1/SAST_machineSTAT.py","file_name":"SAST_machineSTAT.py","file_ext":"py","file_size_in_byte":690,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"13061794941","text":"from flask import jsonify\r\nfrom functools import wraps\r\n\r\nfrom app.utils.token import decode_token\r\n\r\n'''\r\n Decorator for checking the authentication in the requests\r\n'''\r\n\r\ndef isAuth(request):\r\n def decorator(f):\r\n @wraps(f)\r\n def wrapped(*args, **kwargs):\r\n\r\n # Authorization token from headers\r\n token = request.headers.get('Authorization')\r\n\r\n # Check conditions...\r\n if token is None:\r\n return jsonify({\r\n \"success\": False,\r\n \"message\": \"Authentication is required\"\r\n })\r\n\r\n # Decoding\r\n status, msg, data = decode_token(token)\r\n \r\n if status is True:\r\n # Continue...\r\n return f(data, *args, **kwargs)\r\n else:\r\n return jsonify({\r\n \"success\": False,\r\n \"message\": msg\r\n })\r\n\r\n return wrapped\r\n return decorator\r\n\r\n\r\n'''\r\n Simple and funny admin decorator/middleware\r\n'''\r\n\r\ndef isAdmin(request):\r\n def decorator(f):\r\n @wraps(f)\r\n def wrapped(*args, **kwargs):\r\n\r\n isAdmin = request.headers.get('Admin')\r\n\r\n if isAdmin:\r\n # Continue...\r\n return f(*args, **kwargs)\r\n else:\r\n return jsonify({\r\n \"success\": False,\r\n \"message\": \"You are not an admin\"\r\n })\r\n\r\n return wrapped\r\n return decorator\r\n","repo_name":"fuadaghazada/Que-Base-SDP","sub_path":"flask_api/app/helpers/isAuth.py","file_name":"isAuth.py","file_ext":"py","file_size_in_byte":1550,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"84"}
+{"seq_id":"18151667524","text":"import json\nimport logging\nfrom typing import Any, Dict, List, Optional, cast\n\nfrom aws_orbit.utils import boto3_client\n\n_logger: logging.Logger = logging.getLogger(__name__)\n\n\ndef get_role(role_name: str) -> Optional[Dict[str, Any]]:\n _logger.debug(f\"Getting Role: {role_name}\")\n\n iam_client = boto3_client(\"iam\")\n try:\n return cast(Dict[str, Any], iam_client.get_role(RoleName=role_name))\n except iam_client.exceptions.NoSuchEntityException:\n return None\n\n\ndef get_open_id_connect_provider(account_id: str, open_id_connect_provider_id: str) -> Optional[Dict[str, Any]]:\n open_id_connect_provider_arn = f\"arn:aws:iam::{account_id}:oidc-provider/{open_id_connect_provider_id}\"\n _logger.debug(f\"Getting OpenIDConnectProvider: {open_id_connect_provider_arn}\")\n\n iam_client = boto3_client(\"iam\")\n try:\n return cast(\n Dict[str, Any],\n iam_client.get_open_id_connect_provider(OpenIDConnectProviderArn=open_id_connect_provider_arn),\n )\n except iam_client.exceptions.NoSuchEntityException:\n return None\n\n\ndef update_assume_role_roles(\n account_id: str,\n role_name: str,\n roles_to_add: Optional[List[str]] = None,\n roles_to_remove: Optional[List[str]] = None,\n) -> None:\n if not roles_to_add and not roles_to_remove:\n raise Exception(\"One of roles_to_add or roles_to_remove is required\")\n\n _logger.debug(f\"Updating AssumeRolePolicy for {role_name}, Adding: {roles_to_add}, Removing: {roles_to_remove}\")\n\n iam_client = boto3_client(\"iam\")\n assume_role_policy = iam_client.get_role(RoleName=role_name)[\"Role\"][\"AssumeRolePolicyDocument\"]\n\n statements = []\n roles_to_add_set = (\n set()\n if roles_to_add is None\n else {f\"arn:aws:iam::{account_id}:role/{role}\" for role in roles_to_add if get_role(role)}\n )\n roles_to_remove_set = (\n set() if roles_to_remove is None else {f\"arn:aws:iam::{account_id}:role/{role}\" for role in roles_to_remove}\n )\n\n _logger.debug(\"current_policies: %s\", assume_role_policy[\"Statement\"])\n for statement in assume_role_policy[\"Statement\"]:\n arn = statement.get(\"Principal\", {}).get(\"AWS\", None)\n if arn in roles_to_remove_set:\n _logger.debug(\"Removing %s from AssumeRolePolicy\", arn)\n continue\n elif arn in roles_to_add_set:\n _logger.debug(\"AssumeRolePolicy Statement (%s) found containing %s\", statement, arn)\n roles_to_add_set.remove(arn)\n statements.append(statement)\n else:\n _logger.debug(\"Keeping %s in AssumeRolePolicy\", statement)\n statements.append(statement)\n\n for arn in roles_to_add_set:\n _logger.debug(\"Adding %s to AssumeRolePolicy\", arn)\n statements.append({\"Effect\": \"Allow\", \"Action\": \"sts:AssumeRole\", \"Principal\": {\"AWS\": arn}})\n\n assume_role_policy[\"Statement\"] = statements\n policy_body = json.dumps(assume_role_policy)\n _logger.debug(\"policy_body: %s\", policy_body)\n iam_client.update_assume_role_policy(RoleName=role_name, PolicyDocument=policy_body)\n\n\ndef add_assume_role_statement(role_name: str, statement: Dict[str, Any]) -> None:\n _logger.debug(f\"Adding AssumeRolePolicy for {role_name}, Adding: {statement}\")\n\n iam_client = boto3_client(\"iam\")\n assume_role_policy = iam_client.get_role(RoleName=role_name)[\"Role\"][\"AssumeRolePolicyDocument\"]\n statements = assume_role_policy[\"Statement\"]\n\n if statement in statements:\n _logger.debug(\"Skipping Statement already contained by the AssumeRolePolicy\")\n else:\n statements.append(statement)\n assume_role_policy[\"Statement\"] = statements\n policy_body = json.dumps(assume_role_policy)\n _logger.debug(\"policy_body: %s\", policy_body)\n iam_client.update_assume_role_policy(RoleName=role_name, PolicyDocument=policy_body)\n\n\ndef remove_assume_role_statement(role_name: str, statement: Dict[str, Any]) -> None:\n _logger.debug(f\"Removing AssumeRolePolicy for {role_name}, Removing: {statement}\")\n\n iam_client = boto3_client(\"iam\")\n assume_role_policy = iam_client.get_role(RoleName=role_name)[\"Role\"][\"AssumeRolePolicyDocument\"]\n statements = assume_role_policy[\"Statement\"]\n\n if statement in statements:\n try:\n _logger.debug(\"Statement found in AssumeRolePolicy .. removing\")\n statements.remove(statement)\n assume_role_policy[\"Statement\"] = statements\n policy_body = json.dumps(assume_role_policy)\n _logger.debug(\"policy_body: %s\", policy_body)\n iam_client.update_assume_role_policy(RoleName=role_name, PolicyDocument=policy_body)\n except iam_client.exceptions.NoSuchEntityException:\n _logger.error(\"Issues with finding the role....moving on\")\n else:\n _logger.debug(\"Statement not found in policy, moving on\")\n","repo_name":"awslabs/aws-orbit-workbench","sub_path":"cli/aws_orbit/services/iam.py","file_name":"iam.py","file_ext":"py","file_size_in_byte":4864,"program_lang":"python","lang":"en","doc_type":"code","stars":127,"dataset":"github-code","pt":"84"}
+{"seq_id":"42566333482","text":"from json import dumps\nfrom flask import Blueprint, jsonify, request\nfrom app.init import redis\nfrom . import logics as lg\nfrom flask_login import login_user, current_user\nfrom app.packages.short import logics as sh_lg\n\nbp = Blueprint(\"api\", __name__, url_prefix=\"/api\")\n\n\n@bp.route(\"/login\", methods=[\"GET\", \"POST\"])\n@lg.error_check\ndef login():\n if request.method == \"POST\" and request.is_json:\n user = request.get_json()\n if user is not None:\n user = lg.check_to_login(user)\n if user:\n login_user(user=user, remember=False, )\n token = current_user.generate_signature().strip().decode()\n urls = sh_lg.get_user_urls(current_user)\n [redis.set(url.name, url.url) for url in urls]\n return jsonify(status=200, message={\"msg\": \"token generated.\", \"token\": token}), 200\n return jsonify(status=200, message=\"wrong request.\", ), 200\n\n\n@bp.route(\"/signup\", methods=[\"GET\", \"POST\"])\n@lg.error_check\ndef signup():\n if request.method == \"POST\" and request.is_json:\n user = request.get_json()\n if user is not None:\n user = lg.check_to_sign_up(user)\n if user:\n return jsonify(status=200, message={\"msg\": \"signup.\", }), 200\n return jsonify(status=200, message=\"wrong request.\", ), 200\n\n\n@bp.route(\"/get-urls\", methods=[\"GET\", \"POST\"])\n@lg.token_required\n@lg.error_check\ndef get_user_urls():\n urls = sh_lg.get_user_urls(current_user)\n urls = lg.dump_urls(urls)\n return jsonify(status=200, message={\"msg\": \"user urls.\", \"urls\": urls}), 200\n\n\n@bp.route(\"/add-url\", methods=[\"GET\", \"POST\"])\n@lg.token_required\n@lg.error_check\ndef add_url():\n if request.is_json:\n url = request.json.get(\"url\")\n if url is not None:\n url = lg.add_url(url)\n redis.set(url.name, url.url)\n return jsonify(status=200, message={\"msg\": \"url added\", \"url\": redis.get(url.name).decode(\"utf-8\")}), 200\n return jsonify(status=200, message=\"wrong request.\", ), 200\n\n\n@bp.route(\"/get-url\", methods=[\"GET\", \"POST\"])\n@lg.token_required\n@lg.error_check\ndef get_user_url():\n if request.is_json:\n name = request.json.get(\"name\")\n if name is not None:\n url = redis.get(name).decode(\"utf-8\")\n return jsonify(status=200, message={\"msg\": \"url\", \"url\": url}), 200\n return jsonify(status=200, message=\"wrong request.\", ), 200","repo_name":"pd-Shah/ShortMe","sub_path":"app/packages/api/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":2443,"program_lang":"python","lang":"en","doc_type":"code","stars":2,"dataset":"github-code","pt":"84"}
+{"seq_id":"10911158615","text":"'''\nAuthor: lmio 2091319361@qq.com\nDate: 2023-08-30 16:52:31\nLastEditors: lmio 2091319361@qq.com\nDescription: 1654. 到家的最少跳跃次数\n'''\n\nfrom collections import deque\nfrom typing import List\n\n\n\nclass Solution:\n def minimumJumps(self, forbidden: List[int], a: int, b: int, x: int) -> int:\n lower,upper = 0, max(x,max(forbidden)+a+b)\n visit = set(forbidden)\n q = deque([(0,0)])\n while q:\n pos, step = q.popleft()\n if pos == x:\n return step\n if pos in visit:\n continue\n visit.add(pos)\n for diff in (a,-b):\n step += 1\n pos += diff\n if pos in visit:\n break\n if lower <= pos <= upper:\n q.append((pos, step))\n return -1","repo_name":"Clear-Love/leetcode","sub_path":"code/1654.py","file_name":"1654.py","file_ext":"py","file_size_in_byte":845,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"23526002458","text":"\"\"\"\nTests for time series api client aka QuantumLeap\n\"\"\"\nimport logging\nimport unittest\nfrom random import random\nimport requests\nimport time\nfrom typing import List\nfrom filip.clients.ngsi_v2 import \\\n ContextBrokerClient, \\\n QuantumLeapClient\nfrom filip.models.base import FiwareHeader\nfrom filip.models.ngsi_v2.context import ContextEntity\nfrom filip.models.ngsi_v2.subscriptions import Message\nfrom filip.utils.cleanup import clean_test, clear_all\nfrom tests.config import settings\n\n\nlogger = logging.getLogger(__name__)\n\n\ndef create_entities() -> List[ContextEntity]:\n \"\"\"\n Create entities with random values\n Returns:\n\n \"\"\"\n def create_attr():\n return {'temperature': {'value': random(),\n 'type': 'Number'},\n 'humidity': {'value': random(),\n 'type': 'Number'},\n 'co2': {'value': random(),\n 'type': 'Number'}}\n\n return [ContextEntity(id='Kitchen', type='Room', **create_attr()),\n ContextEntity(id='LivingRoom', type='Room', **create_attr())]\n\n\ndef create_time_series_data(num_records: int = 50000):\n \"\"\"\n creates large testing data sets that should remain on the server.\n This is mainly to reduce time for testings\n \"\"\"\n fiware_header = FiwareHeader(service=settings.FIWARE_SERVICE,\n service_path=\"/static\")\n\n with QuantumLeapClient(url=settings.QL_URL, fiware_header=fiware_header) \\\n as client:\n\n for i in range(num_records):\n notification_message = Message(data=create_entities(),\n subscriptionId=\"test\")\n client.post_notification(notification_message)\n\n\nclass TestTimeSeries(unittest.TestCase):\n \"\"\"\n Test class for time series api client\n \"\"\"\n\n def setUp(self) -> None:\n \"\"\"\n Setup test data\n Returns:\n None\n \"\"\"\n self.fiware_header = FiwareHeader(\n service=settings.FIWARE_SERVICE,\n service_path=settings.FIWARE_SERVICEPATH)\n self.ql_client = QuantumLeapClient(\n url=settings.QL_URL,\n fiware_header=self.fiware_header)\n\n self.cb_client = ContextBrokerClient(\n url=settings.CB_URL,\n fiware_header=self.fiware_header)\n\n def test_meta_endpoints(self) -> None:\n \"\"\"\n Test meta data endpoints\n Returns:\n None\n \"\"\"\n with QuantumLeapClient(\n url=settings.QL_URL,\n fiware_header=self.fiware_header) \\\n as client:\n self.assertIsNotNone(client.get_version())\n self.assertIsNotNone(client.get_health())\n\n @clean_test(fiware_service=settings.FIWARE_SERVICE,\n fiware_servicepath=settings.FIWARE_SERVICEPATH,\n cb_url=settings.CB_URL,\n ql_url=settings.QL_URL)\n def test_input_endpoints(self) -> None:\n \"\"\"\n Test input endpoint\n Returns:\n None\n \"\"\"\n entities = create_entities()\n for entity in entities:\n self.cb_client.post_entity(entity)\n\n with QuantumLeapClient(\n url=settings.QL_URL,\n fiware_header=self.fiware_header) \\\n as client:\n notification_message = Message(data=entities,\n subscriptionId=\"test\")\n client.post_subscription(cb_url=settings.CB_URL,\n ql_url=settings.QL_URL,\n entity_id=entities[0].id)\n client.post_notification(notification_message)\n time.sleep(1)\n\n @clean_test(fiware_service=settings.FIWARE_SERVICE,\n fiware_servicepath=settings.FIWARE_SERVICEPATH,\n cb_url=settings.CB_URL,\n ql_url=settings.QL_URL)\n def test_entity_context(self) -> None:\n \"\"\"\n Test entities endpoint\n Returns:\n None\n \"\"\"\n entities = create_entities()\n with QuantumLeapClient(\n url=settings.QL_URL,\n fiware_header=self.fiware_header) \\\n as client:\n notification_message = Message(data=entities,\n subscriptionId=\"test\")\n client.post_notification(notification_message)\n\n time.sleep(1)\n entities = client.get_entities(entity_type=entities[0].type)\n for entity in entities:\n logger.debug(entity.json(indent=2))\n\n def test_query_endpoints_by_id(self) -> None:\n \"\"\"\n Test queries with default values\n\n Returns:\n None\n \"\"\"\n with QuantumLeapClient(\n url=settings.QL_URL,\n fiware_header=self.fiware_header.copy(\n update={'service_path': '/static'})) \\\n as client:\n\n entities = create_entities()\n\n with self.assertRaises(requests.RequestException):\n client.get_entity_by_id(entity_id=entities[0].id,\n entity_type='MyType')\n for entity in entities:\n # get by id\n attrs_id = client.get_entity_by_id(entity_id=entity.id,\n aggr_period='minute',\n aggr_method='avg',\n attrs='temperature,co2')\n logger.debug(attrs_id.json(indent=2))\n logger.debug(attrs_id.to_pandas())\n\n attrs_values_id = client.get_entity_values_by_id(\n entity_id=entity.id)\n logger.debug(attrs_values_id.to_pandas())\n self.assertEqual(len(attrs_values_id.index), 10000)\n\n attr_id = client.get_entity_attr_by_id(\n entity_id=entity.id, attr_name=\"temperature\")\n logger.debug(attr_id.to_pandas())\n self.assertEqual(len(attr_id.index), 10000)\n\n attr_values_id = client.get_entity_attr_values_by_id(\n entity_id=entity.id, attr_name=\"temperature\")\n logger.debug(attr_values_id.to_pandas())\n self.assertEqual(len(attrs_values_id.index), 10000)\n\n def test_query_endpoints_by_type(self) -> None:\n \"\"\"\n Test queries by type with default values\n\n Returns:\n None\n \"\"\"\n with QuantumLeapClient(\n url=settings.QL_URL,\n fiware_header=self.fiware_header.copy(\n update={'service_path': '/static'})) \\\n as client:\n\n entities = create_entities()\n\n for entity in entities:\n # get by type\n attrs_type = client.get_entity_by_type(\n entity_type=entity.type)\n for entity_id in attrs_type:\n logger.debug(entity_id.to_pandas())\n\n self.assertEqual(sum([len(entity_id.index) for\n entity_id in attrs_type]),\n 10000)\n\n attrs_values_type = client.get_entity_values_by_type(\n entity_type=entity.type, )\n for entity_id in attrs_values_type:\n logger.debug(entity_id.to_pandas())\n self.assertEqual(sum([len(entity_id.index) for\n entity_id in attrs_values_type]),\n 10000)\n\n attr_type = client.get_entity_attr_by_type(\n entity_type=entity.type, attr_name=\"temperature\")\n for entity_id in attr_type:\n logger.debug(entity_id.to_pandas())\n self.assertEqual(sum([len(entity_id.index) for\n entity_id in attr_type]),\n 10000)\n\n attr_values_type = client.get_entity_attr_values_by_type(\n entity_type=entity.type, attr_name=\"temperature\")\n for entity_id in attr_values_type:\n logger.debug(entity_id.to_pandas())\n self.assertEqual(sum([len(entity_id.index) for\n entity_id in attr_values_type]),\n 10000)\n\n def test_test_query_endpoints_with_args(self) -> None:\n \"\"\"\n Test arguments for queries\n\n Returns:\n None\n \"\"\"\n with QuantumLeapClient(\n url=settings.QL_URL,\n fiware_header=self.fiware_header.copy(\n update={'service_path': '/static'})) \\\n as client:\n\n for entity in create_entities():\n # test limit\n for limit in range(5000, 25000, 5000):\n records = client.get_entity_by_id(\n entity_id=entity.id,\n attrs='temperature,co2',\n limit=limit)\n\n logger.debug(records.json(indent=2))\n logger.debug(records.to_pandas())\n self.assertEqual(len(records.index), limit)\n\n # test last_n\n for last_n in range(5000, 25000, 5000):\n limit = 15000\n last_n_records = client.get_entity_by_id(\n entity_id=entity.id,\n attrs='temperature,co2',\n limit=limit,\n last_n=last_n)\n self.assertGreater(last_n_records.index[0],\n records.index[0])\n self.assertEqual(len(last_n_records.index),\n min(last_n, limit))\n\n # test offset\n old_records = None\n for offset in range(5000, 25000, 5000):\n # with limit\n records = client.get_entity_by_id(\n entity_id=entity.id,\n attrs='temperature,co2',\n offset=offset)\n\n if old_records:\n self.assertLess(old_records.index[0],\n records.index[0])\n old_records = records\n\n old_records = None\n for offset in range(5000, 25000, 5000):\n # test with last_n\n records = client.get_entity_by_id(\n entity_id=entity.id,\n attrs='temperature,co2',\n offset=offset,\n last_n=5)\n if old_records:\n self.assertGreater(old_records.index[0],\n records.index[0])\n old_records = records\n\n def tearDown(self) -> None:\n \"\"\"\n Clean up server\n Returns:\n None\n \"\"\"\n clear_all(fiware_header=self.fiware_header,\n cb_url=settings.CB_URL,\n ql_url=settings.QL_URL)\n\n self.ql_client.close()\n self.cb_client.close()\n","repo_name":"RWTH-EBC/FiLiP","sub_path":"tests/clients/test_ngsi_v2_timeseries.py","file_name":"test_ngsi_v2_timeseries.py","file_ext":"py","file_size_in_byte":11330,"program_lang":"python","lang":"en","doc_type":"code","stars":16,"dataset":"github-code","pt":"84"}
+{"seq_id":"39848263326","text":"import pyglet\nfrom pyglet.window import mouse\nfrom pyglet.gl import gl\nfrom pyglet import resource\nfrom pyglet import sprite\nfrom pyglet import clock\nfrom random import randint\n\nresource.path = ['./resources/']\nresource.reindex()\npyglet.options['debug_gl'] = False\n\nSCREENW = 800\nSCREENH = 600\nFULLSCREEN = False\nwindow = pyglet.window.Window(SCREENW, SCREENH, caption=\"Bomb Your Opponents\",\n fullscreen=FULLSCREEN)\n\nicon16 = resource.image('icon16.png')\nicon32 = resource.image('icon32.png')\nwindow.set_icon(icon16, icon32)\n\ndefault_cur = window.get_system_mouse_cursor(window.CURSOR_DEFAULT)\nchoose_cur = window.get_system_mouse_cursor(window.CURSOR_HAND)\n\n\ndef center_image(image):\n image.anchor_x = image.width // 2\n image.anchor_y = image.height // 2\n\n\n# Load images\ngame_BG = resource.image('background.png')\nplatform = resource.image('platform2.png')\nship1 = resource.image('ship1.png')\nship2 = resource.image('ship2.png')\nstats_box = resource.image('statsBox.png')\nbomb = resource.image('bomb.png')\ntrophy = resource.image('trophy.png')\nmain_BG = resource.image('mainBG.png')\ngame_logo = resource.image('logo.png')\nstart_unselected = resource.image('startUnselected.png')\nstart_selected = resource.image('startSelected.png')\ninfo = resource.image('info.png')\ngo_back = resource.image('credits_back.png')\npage_left = resource.image('pageL.png')\npage_right = resource.image('pageR.png')\n\n# Anchoring certain images\ncenter_image(bomb)\ncenter_image(game_logo)\ncenter_image(start_unselected)\ncenter_image(start_selected)\n\n# Allow transparency for blits\ngl.glEnable(gl.GL_BLEND)\ngl.glBlendFunc(gl.GL_SRC_ALPHA, gl.GL_ONE_MINUS_SRC_ALPHA)\n\n# Animations\nexplosion_frames = [\n resource.image('explosion1.png'),\n resource.image('explosion2.png'),\n resource.image('explosion3.png'),\n resource.image('explosion4.png')\n]\n\nfor img in explosion_frames:\n img.anchor_x = img.width / 2\n img.anchor_y = img.height / 2\n\n\nclass Player():\n def __init__(self, name, hp, status, pId):\n self.name = name\n self.hp = hp\n self.status = status\n self.pId = pId\n\n\nclass Engine():\n def __init__(self, currentScreen):\n self.mouse_x = 0\n self.mouse_y = 0\n self.currentScreen = currentScreen\n self.paused = True\n\n def on_click(self, x, y, button):\n self.currentScreen.on_click(x, y, button)\n\n def mouseXY(self, x, y, dx, dy):\n self.mouse_x = x\n self.mouse_y = y\n\n def draw(self):\n self.currentScreen.draw()\n\n def update(self, dt):\n window.set_mouse_cursor(default_cur)\n self.currentScreen.update(dt)\n\n def setCurrentScreen(self, currentScreen):\n self.currentScreen = currentScreen\n\n\nclass Screen():\n def __init__(self):\n self.paused = True\n\n def draw(self):\n pass\n\n def on_click(self, x, y, button):\n pass\n\n def update(self, dt):\n pass\n\n\nclass MainMenu(Screen):\n\n logo = sprite.Sprite(game_logo, x=SCREENW//2, y=480)\n bg = sprite.Sprite(main_BG, x=0, y=0)\n selected = sprite.Sprite(start_selected, x=400, y=200)\n unselected = sprite.Sprite(start_unselected, x=selected.x, y=selected.y)\n info_button = sprite.Sprite(info, x=5, y=5)\n\n def __init__(self):\n self.mouse_overButton = False\n self.p1 = Player(\"Red\", 100, \"alive\", ship1)\n self.p2 = Player(\"Blue\", 100, \"alive\", ship2)\n\n # Texts\n self.copyrights = pyglet.text.Label(\"Made by webguy88 in 2020\",\n x=655, y=15, anchor_x='center',\n anchor_y='center', font_size=16,\n color=(0, 0, 0, 255),\n bold=True)\n\n self.version = pyglet.text.Label(\"v 1.1\",\n x=768, y=40, anchor_x='center',\n anchor_y='center', font_size=16,\n color=(0, 0, 0, 255),\n bold=True)\n\n # Regions\n self.start = Region(250, 150, 300, 100)\n self.information = Region(5, 5, 71, 64)\n\n def draw(self):\n self.bg.draw()\n self.logo.draw()\n self.copyrights.draw()\n self.version.draw()\n self.info_button.draw()\n\n if self.mouse_overButton:\n self.selected.draw()\n else:\n self.unselected.draw()\n\n def on_click(self, x, y, button):\n if self.start.contain(x, y):\n engine.setCurrentScreen(game)\n\n elif self.information.contain(x, y):\n engine.setCurrentScreen(credit)\n\n def update(self, dt):\n if self.start.contain(engine.mouse_x, engine.mouse_y):\n self.mouse_overButton = True\n window.set_mouse_cursor(choose_cur)\n else:\n self.mouse_overButton = False\n\n if self.information.contain(engine.mouse_x, engine.mouse_y):\n window.set_mouse_cursor(choose_cur)\n\n\nclass Credit(Screen):\n\n credits_back = sprite.Sprite(go_back, x=10, y=500)\n pageL = sprite.Sprite(page_left, x=350, y=5)\n pageR = sprite.Sprite(page_right, x=420, y=5)\n gameplay = resource.image('gameplay.png')\n\n def __init__(self):\n self.page = 1\n self.back = Region(10, 500, 135, 86)\n self.goRight = Region(420, 5, 64, 64)\n self.goLeft = Region(350, 5, 64, 64)\n\n # Text\n self.license = pyglet.text.Label(\n \"\"\"\n Copyright (c) 2006-2008 Alex Holkner\n Copyright (c) 2008-2020 pyglet contributors\n All rights reserved.\n\n Redistribution and use in source and binary forms, with or without\n modification, are permitted provided that the following conditions are met:\n\n * Redistributions of source code must retain the above copyright\n notice, this list of conditions and the following disclaimer.\n * Redistributions in binary form must reproduce the above copyright\n notice, this list of conditions and the following disclaimer in\n the documentation and/or other materials provided with the\n distribution.\n * Neither the name of pyglet nor the names of its\n contributors may be used to endorse or promote products\n derived from this software without specific prior written\n permission.\n\n THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS\n \"AS IS\" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT\n LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS\n FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE\n COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,\n INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,\n BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;\n LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER\n CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT\n LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN\n ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE\n POSSIBILITY OF SUCH DAMAGE.\n \"\"\",\n x=550, y=230, anchor_x='center', anchor_y='center', font_size=10,\n bold=True, color=(0, 0, 0, 255),\n multiline=True, width=800, height=600\n )\n\n self.license_header = pyglet.text.Label(\"License\", x=400, y=550,\n anchor_x='center',\n anchor_y='center',\n font_size=24, bold=True,\n color=(0, 0, 0, 255))\n\n def draw(self):\n\n if self.page == 1:\n main_menu.bg.draw()\n self.license_header.draw()\n self.pageR.draw()\n self.license.draw()\n\n if self.page == 2:\n self.gameplay.blit(0, 0)\n self.pageL.draw()\n\n self.credits_back.draw()\n\n def on_click(self, x, y, button):\n if self.back.contain(x, y):\n engine.setCurrentScreen(main_menu)\n\n # Go to page 2\n if self.page == 1 and self.goRight.contain(x, y):\n self.page = 2\n\n # Go to page 1\n elif self.page == 2 and self.goLeft.contain(x, y):\n self.page = 1\n\n def update(self, dt):\n if self.back.contain(engine.mouse_x, engine.mouse_y):\n window.set_mouse_cursor(choose_cur)\n\n if self.page == 1 and self.goRight.contain(engine.mouse_x,\n engine.mouse_y):\n window.set_mouse_cursor(choose_cur)\n\n elif self.page == 2 and self.goLeft.contain(engine.mouse_x,\n engine.mouse_y):\n window.set_mouse_cursor(choose_cur)\n\n\nclass Game(Screen):\n\n # Turning preloaded images into sprite\n bg = sprite.Sprite(game_BG, x=0, y=0)\n player1 = sprite.Sprite(ship1, x=83, y=130)\n player2 = sprite.Sprite(ship2, x=593, y=250)\n explosion_animation = pyglet.image.Animation.from_image_sequence(\n explosion_frames, duration=0.1, loop=True)\n explosion_sound = resource.media('explode.wav', streaming=False)\n bomb_flying = resource.media('bombfly.wav', streaming=False)\n\n # Media player stuff\n exp_player = pyglet.media.Player()\n exp_player.queue(explosion_sound)\n bmbf = pyglet.media.Player()\n bmbf.queue(bomb_flying)\n\n def __init__(self):\n # Firing system\n self.fired = False\n self.who_firedBomb = 0\n self.bomb_x = -10\n self.bomb_y = -10\n self.explosion_x = -10\n self.explosion_y = -10\n\n # Rest of instances\n self.audio_player = None\n self.explosion = sprite.Sprite(self.explosion_animation,\n x=self.explosion_x,\n y=self.explosion_y)\n self.bombSpr = sprite.Sprite(bomb, x=self.bomb_x, y=self.bomb_y)\n self.p1 = main_menu.p1\n self.p2 = main_menu.p2\n self.turn = 2\n self.mouse_overPlayer = 0\n self.player_hit = 0\n self.p1_hitbox = Region(self.player1.x, self.player1.y, 139, 71)\n self.p2_hitbox = Region(self.player2.x, self.player2.y, 139, 71)\n self.bomb_hitbox = Region(self.bombSpr.x, self.bombSpr.y, 48, 55)\n\n # Text\n\n self.p1name = pyglet.text.Label(f\"Name: {self.p1.name}\",\n x=(self.player1.x + 53),\n y=(self.player1.y - 42),\n anchor_x='center', anchor_y='center',\n font_size=16, bold=True,\n color=(0, 0, 0, 255))\n\n self.p2name = pyglet.text.Label(f\"Name: {self.p2.name}\",\n x=(self.player2.x + 53),\n y=(self.player2.y - 42),\n anchor_x='center', anchor_y='center',\n font_size=16, bold=True,\n color=(0, 0, 0, 255))\n\n self.p1HP = pyglet.text.Label(f\"HP: {self.p1.hp}\",\n x=(self.player1.x + 37),\n y=(self.player1.y - 69),\n anchor_x='center', anchor_y='center',\n font_size=16, bold=True,\n color=(0, 0, 0, 255))\n\n self.p2HP = pyglet.text.Label(f\"HP: {self.p2.hp}\",\n x=(self.player2.x + 37),\n y=(self.player2.y - 69),\n anchor_x='center', anchor_y='center',\n font_size=16, bold=True,\n color=(0, 0, 0, 255))\n\n self.p1_status = pyglet.text.Label(f\"Status: {self.p1.status}\",\n x=(self.player1.x + 60),\n y=(self.player1.y - 95),\n anchor_x='center',\n anchor_y='center',\n font_size=16, bold=True,\n color=(0, 0, 0, 255))\n\n self.p2_status = pyglet.text.Label(f\"Status: {self.p2.status}\",\n x=(self.player2.x + 60),\n y=(self.player2.y - 95),\n anchor_x='center',\n anchor_y='center',\n font_size=16, bold=True,\n color=(0, 0, 0, 255))\n\n # Turn texts\n self.turn_text = None\n\n self.turn1 = pyglet.text.Label(f\"It's {self.p1.name}'s turn!\",\n x=SCREENW/2, y=570,\n anchor_x='center', anchor_y='center',\n font_size=28, bold=True,\n color=(255, 255, 255, 255))\n\n self.turn2 = pyglet.text.Label(f\"It's {self.p2.name}'s turn!\",\n x=SCREENW/2, y=570,\n anchor_x='center', anchor_y='center',\n font_size=28, bold=True,\n color=(255, 255, 255, 255))\n\n # Choosing system\n self.choose = randint(0, 1)\n\n # Debugging goes here\n ...\n\n def play_explosion(self):\n self.audio_player = pyglet.media.Player()\n self.audio_player.queue(self.explosion_sound)\n self.audio_player.play()\n pass\n\n def play_fly(self):\n self.audio_player = pyglet.media.Player()\n self.audio_player.queue(self.bomb_flying)\n self.audio_player.play()\n pass\n\n def pause_explosion(self):\n self.audio_player.pause()\n pass\n\n def pause_fly(self):\n self.audio_player.pause()\n\n def select_player(self):\n if self.choose == 0:\n if self.p1.status == \"alive\":\n self.turn = 1\n\n elif self.p1.hp <= 0:\n self.turn = -1\n\n elif self.choose == 1:\n if self.p2.status == \"alive\":\n self.turn = 2\n\n elif self.p2.hp <= 0:\n self.turn = -1\n\n # Reset all data\n def restart(self):\n self.p1.name = \"Red\"\n self.p2.name = \"Blue\"\n self.p1.hp = 100\n self.p2.hp = 100\n self.p1.status = \"alive\"\n self.p2.status = \"alive\"\n self.player_hit = 0\n self.fired = False\n self.who_firedBomb = 0\n\n # Drawing stuff\n def draw(self):\n self.bg.draw()\n platform.blit(70, 100) # In this case we use 1 image to draw multiple\n platform.blit(580, 220)\n self.player1.draw()\n self.player2.draw()\n\n # Stats\n if self.mouse_overPlayer == 1:\n stats_box.blit((self.player1.x - 30), (self.player1.y - 120))\n self.p1name.draw()\n self.p1HP.draw()\n self.p1_status.draw()\n\n if self.mouse_overPlayer == 2:\n stats_box.blit((self.player2.x - 30), (self.player2.y - 120))\n self.p2name.draw()\n self.p2HP.draw()\n self.p2_status.draw()\n\n if self.turn == 1 and not self.fired:\n self.turn_text = self.turn1\n self.bombSpr.x = (self.player1.x + 75)\n self.bombSpr.y = (self.player1.y + 40)\n\n elif self.turn == 2 and not self.fired:\n self.turn_text = self.turn2\n self.bombSpr.x = (self.player2.x + 75)\n self.bombSpr.y = (self.player2.y + 40)\n\n elif self.turn == 0:\n self.player_hit = 0\n self.bombSpr.x = -10\n self.bombSpr.y = -10\n\n self.turn_text.draw()\n\n # Bomb stuff\n if self.fired:\n self.bombSpr.draw()\n\n if self.player_hit == 1:\n self.explosion.x = (self.player1.x + 75)\n self.explosion.y = (self.player1.y + 40)\n draw_explosion(1)\n\n elif self.player_hit == 2:\n self.explosion.x = (self.player2.x + 75)\n self.explosion.y = (self.player2.y + 40)\n draw_explosion(1)\n\n # Debugging goes here\n ...\n\n def on_click(self, x, y, button):\n\n # Can only click whenever it's the player's turn and\n # they are not under the explosion animation\n\n # Player 1\n if self.p2_hitbox.contain(x, y) and self.turn == 1 \\\n and self.player_hit == 0 and not self.fired:\n self.fired = True\n self.play_fly()\n self.who_firedBomb = 1\n\n # Player 2\n if self.p1_hitbox.contain(x, y) and self.turn == 2 \\\n and self.player_hit == 0 and not self.fired:\n self.fired = True\n self.play_fly()\n self.who_firedBomb = 2\n\n def update(self, dt):\n self.mouse_overPlayer = 0\n\n # Choose player 1\n if self.p1_hitbox.contain(engine.mouse_x, engine.mouse_y):\n window.set_mouse_cursor(choose_cur)\n self.mouse_overPlayer = 1\n self.bomb_x = self.player1.x\n self.bomb_y = self.player1.y\n\n # Choose player 2\n elif self.p2_hitbox.contain(engine.mouse_x, engine.mouse_y):\n window.set_mouse_cursor(choose_cur)\n self.mouse_overPlayer = 2\n self.bomb_x = self.player2.x\n self.bomb_y = self.player2.y\n\n # Player 1 fired bomb\n if self.fired and self.who_firedBomb == 1:\n self.bombSpr.x += 10\n self.bombSpr.y += 2.5\n self.bombSpr.rotation += 10\n\n # Player 2 fired bomb\n if self.fired and self.who_firedBomb == 2:\n self.bombSpr.x -= 10\n self.bombSpr.y -= 2.5\n self.bombSpr.rotation -= 10\n\n # Bomb hits player 1\n if self.p1_hitbox.contain(self.bombSpr.x, self.bombSpr.y) \\\n and self.turn == 2:\n self.player_hit = 1\n self.fired = False\n self.pause_fly()\n self.play_explosion()\n clock.schedule_once(stop_explosion, 1)\n\n # Remove HP and give next turn\n if self.player_hit == 1:\n self.p1.hp -= randint(5, 20)\n self.choose = randint(0, 1)\n self.select_player()\n\n # Bomb hits player 2\n if self.p2_hitbox.contain(self.bombSpr.x, self.bombSpr.y) \\\n and self.turn == 1:\n self.player_hit = 2\n self.fired = False\n self.pause_explosion()\n self.play_explosion()\n clock.schedule_once(stop_explosion, 1)\n\n # Remove HP and give next turn\n if self.player_hit == 2:\n print(self.p2.hp)\n self.p2.hp -= randint(5, 20)\n self.choose = randint(0, 1)\n self.select_player()\n\n self.set_stats_text()\n\n def set_stats_text(self):\n self.p1HP = pyglet.text.Label(f\"HP: {self.p1.hp}\",\n x=(self.player1.x + 37),\n y=(self.player1.y - 69),\n anchor_x='center', anchor_y='center',\n font_size=16, bold=True,\n color=(0, 0, 0, 255))\n\n self.p2HP = pyglet.text.Label(f\"HP: {self.p2.hp}\",\n x=(self.player2.x + 37),\n y=(self.player2.y - 69),\n anchor_x='center', anchor_y='center',\n font_size=16, bold=True,\n color=(0, 0, 0, 255))\n\n self.p1_status = pyglet.text.Label(f\"Status: {self.p1.status}\",\n x=(self.player1.x + 60),\n y=(self.player1.y - 95),\n anchor_x='center',\n anchor_y='center',\n font_size=16, bold=True,\n color=(0, 0, 0, 255))\n\n self.p2_status = pyglet.text.Label(f\"Status: {self.p2.status}\",\n x=(self.player2.x + 60),\n y=(self.player2.y - 95),\n anchor_x='center',\n anchor_y='center',\n font_size=16, bold=True,\n color=(0, 0, 0, 255))\n\n # Check HP and status\n if self.p1.hp <= 0:\n self.p1.status = \"dead\"\n engine.setCurrentScreen(WinnerScreen(self.p2.name))\n\n elif self.p2.hp <= 0:\n self.p2.status = \"dead\"\n engine.setCurrentScreen(WinnerScreen(self.p1.name))\n\n\nclass WinnerScreen(Screen):\n\n prize = sprite.Sprite(trophy, x=0, y=0)\n victory_sound = resource.media('victory.wav', streaming=False)\n\n victory_player = pyglet.media.Player()\n victory_player.queue(victory_sound)\n\n def __init__(self, winner):\n self.winner = winner\n self.leave = Region(0, 0, SCREENW, SCREENH)\n self.victory_player.play()\n\n self.winner_text = pyglet.text.Label(f\"\"\"Congratulations, {self.winner}!\nYou have won!\"\"\",\n x=400, y=300, anchor_x='center',\n anchor_y='center',\n font_size=24, bold=True)\n\n def draw(self):\n self.winner_text.draw()\n self.prize.opacity = 50\n self.prize.draw()\n\n def on_click(self, x, y, button):\n if self.leave.contain(x, y):\n engine.setCurrentScreen(main_menu)\n game.restart()\n\n def update(self, dt):\n pass\n\n\n# Gives life to the region\nclass Rect:\n\n def __init__(self, x, y, w, h):\n self.set(x, y, w, h)\n\n def draw(self):\n pyglet.graphics.draw(4, gl.GL_QUADS, self._quad)\n\n def set(self, x=None, y=None, w=None, h=None):\n self._x = self._x if x is None else x\n self._y = self._y if y is None else y\n self._w = self._w if w is None else w\n self._h = self._h if h is None else h\n self._quad = ('v2f', (self._x, self._y,\n self._x + self._w, self._y,\n self._x + self._w, self._y + self._h,\n self._x, self._y + self._h))\n\n def __repr__(self):\n return f\"Rect(x={self._x}, y={self._y}, w={self._w}, h={self._h})\"\n\n\n# Allows clicks in certain areas\nclass Region(object):\n\n def __init__(self, x, y, width, height):\n self.x = x\n self.y = y\n self.width = width\n self.height = height\n\n def contain(self, x, y):\n inside_x = False\n inside_y = False\n\n if x >= self.x and x <= (self.x + self.width):\n inside_x = True\n\n if y >= self.y and y <= (self.y + self.height):\n inside_y = True\n\n if inside_x and inside_y:\n return True\n else:\n return False\n\n def draw(self):\n r = Rect(self.x, self.y, self.width, self.height)\n r.draw()\n\n\n# Class instances\nmain_menu = MainMenu()\ncredit = Credit()\ngame = Game()\nengine = Engine(main_menu)\n\n\n# Window events\n@window.event\ndef on_draw():\n window.clear()\n engine.draw()\n\n\n@window.event\ndef on_mouse_press(x, y, button, modifiers):\n if button & mouse.LEFT:\n engine.on_click(x, y, button)\n\n\n@window.event\ndef on_mouse_motion(x, y, dx, dy):\n engine.mouseXY(x, y, dx, dy)\n pass\n\n\n@window.event\ndef update(dt):\n engine.update(dt)\n pass\n\n\n@window.event\ndef stop_explosion(dt):\n print(game.player_hit)\n game.player_hit = 0\n pass\n\n\n@window.event\ndef draw_explosion(dt):\n game.explosion.draw()\n\n\nclock.schedule_interval(update, 1/30)\nclock.schedule_once(draw_explosion, 1)\n\npyglet.app.run()\n","repo_name":"webguy88/byo","sub_path":"bombG.py","file_name":"bombG.py","file_ext":"py","file_size_in_byte":24447,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"84"}
+{"seq_id":"26207125542","text":"import numpy as np\ndef rotate(matrix):\n \"\"\"\n :type matrix: List[List[int]]\n :rtype: None Do not return anything, modify matrix in-place instead.\n https://leetcode-cn.com/problems/rotate-image/\n \"\"\"\n # Solution 1: 原地旋转\n N = len(matrix)\n for row in range(N // 2):\n # 要全部遍历该行所有元素\n for col in range(N):\n matrix[row][col],matrix[N -1-row][col] = matrix[N -1-row][col],matrix[row][col]\n\n for row in range(N):\n # 沿对角线交换只用遍历到row\n for col in range(row):\n # 沿对角线交换\n matrix[row][col], matrix[col][row] = matrix[col][row], matrix[row][col]\n return matrix\n\n # Solution 2: 引入另一个数组\n # N = len(matrix)\n # mat = [[0]*N for _ in range(N)]\n # print(mat)\n # for i in range(N):\n # for j in range(N):\n # mat[j][N-1-i] = matrix[i][j] # 倒数第i个元素索引: N-1-i\n # # 不能写成 matrix = matrix_new\n # matrix[:] = mat\n # return matrix\n\n # Solution 3: 原地旋转\n # N = len(matrix)\n # for row in range(N//2):\n # if N % 2 == 0:\n # col_end_itr = N/2\n # else:\n # col_end_itr = (N+1)//2 # // 返回整数, / 返回小数 2.0\n #\n # for col in range(col_end_itr):\n # temp = matrix[row][col]\n # matrix[row][col] = matrix[N-col-1][row]\n # matrix[N-col-1][row] = matrix[N-row-1][N-col-1]\n # matrix[N - row - 1][N - col - 1] = matrix[col][N-row-1]\n # matrix[col][N - row - 1] = temp\n # return matrix\n\n\n\nmatrix = [[1,2,3],[4,5,6],[7,8,9]]\nprint(rotate(matrix))\n","repo_name":"yaxi333/Leetcode_Execises","sub_path":"Array/48.rotate array.py","file_name":"48.rotate array.py","file_ext":"py","file_size_in_byte":1666,"program_lang":"python","lang":"en","doc_type":"code","stars":0,"dataset":"github-code","pt":"84"}
+{"seq_id":"20678562061","text":"# coding=utf-8\n\"\"\"\n@Time : 2020/8/6 20:31 \n@Author : Haojun Gao (github.com/VincentGaoHJ)\n@Email : vincentgaohj@gmail.com haojun.gao@u.nus.edu\n@Sketch : 剑指 Offer 21. 调整数组顺序使奇数位于偶数前面\n https://leetcode-cn.com/problems/diao-zheng-shu-zu-shun-xu-shi-qi-shu-wei-yu-ou-shu-qian-mian-lcof/\n\"\"\"\n\n\nclass Solution(object):\n def exchange(self, nums):\n \"\"\"\n :type nums: List[int]\n :rtype: List[int]\n \"\"\"\n odd_idx = 0\n for idx, num in enumerate(nums):\n if num % 2 == 1:\n nums[idx], nums[odd_idx] = nums[odd_idx], nums[idx]\n odd_idx += 1\n return nums\n\n\ndef offer_21(nums):\n \"\"\"\n 输入一个整数数组,实现一个函数来调整该数组中数字的顺序,使得所有奇数位于数组的前半部分,所有偶数位于数组的后半部分。\n 输入:nums = [1,2,3,4]\n 输出:[1,3,2,4]\n 注:[3,1,2,4] 也是正确的答案之一。\n 提示:\n 1 <= nums.length <= 50000\n 1 <= nums[i] <= 10000\n :param nums: [list]\n :return:\n \"\"\"\n solution = Solution()\n output = solution.exchange(nums)\n\n print(output)\n\n\nif __name__ == '__main__':\n # 剑指 Offer 21. 调整数组顺序使奇数位于偶数前面\n offer_21([1, 2, 3, 4])\n","repo_name":"VincentGaoHJ/Sword-For-Offer","sub_path":"rsc/1_easy/offer_21.py","file_name":"offer_21.py","file_ext":"py","file_size_in_byte":1314,"program_lang":"python","lang":"en","doc_type":"code","stars":1,"dataset":"github-code","pt":"84"}