diff --git "a/6309.jsonl" "b/6309.jsonl" new file mode 100644--- /dev/null +++ "b/6309.jsonl" @@ -0,0 +1,693 @@ +{"seq_id":"631630416","text":"\"\"\"This program allows the user to input an integer representing the size of a list, it then allows the user to insert that many elements into the list\"\"\"\n\ndef populate_list(int_object):\n new_list = []\n while len(new_list) != int_object:\n list_element = input(\"Enter an element to put into the list: \")\n new_list.append(list_element)\n return new_list\n\ndef main_func():\n size_num = int(input(\"Enter the size of the list: \"))\n value_list = populate_list(size_num)\n\n print(value_list)\n\nmain_func()","sub_path":"midterm_list_exercise1.py","file_name":"midterm_list_exercise1.py","file_ext":"py","file_size_in_byte":526,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"265521713","text":"#!/usr/bin/python3\nimport re\n\ndef main():\n regexPline()\n regexPgroup()\n\ndef regexPline():\n fh = open('text.txt')\n for line in fh:\n if re.search('year', line):\n print(line, end ='')\n print()\n\ndef regexPgroup():\n fh = open('text.txt')\n for line in fh:\n match = re.search('year', line)\n if match:\n print(match.group())\n print()\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"Python/Essentials_Python3/Chapter6/regex_search.py","file_name":"regex_search.py","file_ext":"py","file_size_in_byte":432,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"229298846","text":"from django.contrib.auth.decorators import login_required, permission_required\nfrom django.http import Http404, HttpResponseRedirect\nfrom django.shortcuts import render\nfrom .models import PayPeriod, Timecard, Task\nfrom django.contrib.auth.models import User\nfrom hq import models as hq_models\nfrom atom.models import LaborGroup, LaborItem, LaborClass\nfrom django.db.models import Sum\nimport datetime\n\n\n@login_required(login_url=\"/login/\")\ndef overview(request):\n \"\"\"/kronos/\n Return overview render.\n \"\"\"\n timecards = Timecard.objects.all().filter(employee=request.user, complete=False)\n return render(request, 'kronos/overview.html', {\n 'timecards': timecards\n })\n\n@login_required(login_url=\"/login/\")\ndef timecard_add(request):\n \"\"\"/kronos/timecard/add/\n If POST, process data and create timecard.\n \"\"\"\n if request.method == \"POST\":\n p = request.POST[\"pay_period\"]\n pay_period = PayPeriod.objects.get(pk=p)\n timecard = Timecard.objects.create(\n employee=request.user,\n pay_period=pay_period\n )\n timecard.save()\n\n return HttpResponseRedirect('/kronos/%s' % (timecard.id))\n\n else:\n pay_periods = PayPeriod.objects.order_by('-start')[:3]\n return render(request, 'kronos/timecard_add.html', {'pay_periods': pay_periods})\n\n@login_required(login_url=\"/login/\")\ndef timecard_detail(request, timecard_id):\n \"\"\"/kronos/1/\n Return timecard detail render.\n \"\"\"\n timecard = Timecard.objects.get(pk=timecard_id)\n tasks = Task.objects.all().filter(timecard=timecard).order_by('started')\n\n return render(request, 'kronos/timecard_detail.html', {\n 'timecard': timecard,\n 'tasks': tasks\n })\n\n@login_required(login_url=\"/login/\")\ndef timecard_complete(request, timecard_id):\n \"\"\"/kronos/1/complete/\n Have to be a manager.\n If POST process data.\n \"\"\"\n if request.method == \"POST\":\n timecard = Timecard.objects.get(pk=timecard_id)\n timecard.submission_notes = request.POST[\"submission_notes\"]\n timecard.complete = True\n timecard.save()\n\n return HttpResponseRedirect('/kronos/')\n else:\n try:\n timecard = Timecard.objects.get(pk=timecard_id)\n tasks = Task.objects.all().filter(timecard=timecard).order_by('started')\n except Timecard.DoesNotExist:\n raise Http404(\"Timecard does not exist.\")\n return render(request, 'kronos/timecard_review.html', {\n 'timecard': timecard,\n 'tasks': tasks\n })\n\n@permission_required('kronos.can_review_timecards')\n@login_required(login_url=\"/login/\")\ndef timecard_review(request, timecard_id):\n \"\"\"/kronos/1/review/\n Have to be a manager.\n If POST, timecard_id, mark timecard as reviewed.\n \"\"\"\n if request.method == \"POST\":\n timecard = Timecard.objects.get(pk=timecard_id)\n timecard.reviewed = True\n timecard.save()\n\n return HttpResponseRedirect('/kronos/timecards/')\n else:\n try:\n timecard = Timecard.objects.get(pk=timecard_id)\n tasks = Task.objects.all().filter(timecard=timecard).order_by('started')\n except Timecard.DoesNotExist:\n raise Http404(\"Timecard does not exist.\")\n return render(request, 'kronos/timecard_review.html', {\n 'timecard': timecard,\n 'tasks': tasks\n })\n\n@login_required(login_url=\"/login/\")\ndef timecard_complete_index(request):\n \"\"\"/kronos/complete/\n Return a list of completed timecards.\n \"\"\"\n try:\n timecard = Timecard.objects.order_by('-pay_period').filter(complete=True, employee=request.user)\n except Timecard.DoesNotExist:\n raise Http404(\"No timecards does not exist.\")\n return render(request, 'kronos/timecard_complete.html', {'timecards': timecard})\n\n@permission_required('kronos.can_review_timecards')\n@login_required(login_url=\"/login/\")\ndef timecard_index(request):\n \"\"\"/kronos/timecards/\n Return a list of all timecards.\n \"\"\"\n try:\n pay_period= PayPeriod.objects.order_by('-start')\n except PayPeriod.DoesNotExist:\n raise Http404(\"Timecards do not exist.\")\n return render(request, 'kronos/timecard_index.html', {'pay_periods': pay_period})\n\n@login_required(login_url=\"/login/\")\ndef timecard_update(request, timecard_id):\n pass\n\n@login_required(login_url=\"/login/\")\ndef task_add(request, timecard_id):\n \"\"\"/kronos/task/add/\n If POST process data and add task to timecard.\n \"\"\"\n if request.method == \"POST\":\n employee = request.POST[\"employee\"]\n date = request.POST[\"date\"]\n project_id = request.POST[\"project\"]\n description = request.POST[\"description\"]\n start_time = request.POST[\"start_time\"]\n end_time = request.POST[\"end_time\"]\n labor_item_id = request.POST[\"labor_item_number\"]\n li_class_id = request.POST[\"li_class\"]\n\n started = \"%s %s\" % (date, start_time)\n finished = \"%s %s\" % (date, end_time)\n\n project = hq_models.Project.objects.get(id=project_id)\n user = User.objects.get(id=employee)\n timecard = Timecard.objects.get(id=timecard_id)\n labor_item = LaborItem.objects.get(id=labor_item_id)\n li_class = LaborClass(id=li_class_id)\n\n t = Task.objects.create(\n employee=user,\n timecard=timecard,\n project = project,\n description = description,\n started = started,\n finished = finished,\n labor_item_number = labor_item,\n li_class = li_class\n )\n t.save()\n\n return HttpResponseRedirect('/kronos/%s' % (timecard_id))\n\n else:\n li_classes = LaborClass.objects.all()\n labor_groups = LaborGroup.objects.all()\n projects = hq_models.Project.objects.all().filter(archived=False)\n return render(request, 'kronos/task_add.html', {\n 'projects': projects,\n 'timecard_id': timecard_id,\n 'labor_groups': labor_groups,\n 'li_classes': li_classes\n })\n\n@login_required(login_url=\"/login/\")\ndef timecard_delete(request, timecard_id):\n \"\"\"/kronos/1/delete/\n Deletes the timecard with id.\n \"\"\"\n t = Timecard.objects.get(pk=timecard_id)\n t.delete()\n return HttpResponseRedirect('/kronos/') \n\n@login_required(login_url=\"/login/\")\ndef task_detail(request, timecard_id, task_id):\n try:\n timecard = Timecard.objects.get(pk=timecard_id)\n task = Task.objects.get(pk=task_id)\n except Timecard.DoesNotExist:\n raise Http404(\"Task or Timecard do not exist.\")\n return render(request, 'kronos/task_detail.html', {\n 'timecard': timecard,\n 'task': task\n })\n\n\n@login_required(login_url=\"/login/\")\ndef task_update(request, timecard_id, task_id):\n \"\"\"/kronos/1/task/1/update/\n Updates task with id.\n \"\"\"\n if request.method == \"POST\":\n\n date = request.POST[\"date\"]\n project_id = request.POST[\"project\"]\n description = request.POST[\"description\"]\n start_time = request.POST[\"start_time\"]\n end_time = request.POST[\"end_time\"]\n labor_item_id = request.POST[\"labor_item_number\"]\n li_class_id = request.POST[\"li_class\"]\n\n started = \"%s %s\" % (date, start_time)\n finished = \"%s %s\" % (date, end_time)\n\n project = hq_models.Project.objects.get(pk=project_id)\n labor_item = LaborItem.objects.get(pk=labor_item_id)\n li_class = LaborClass(pk=li_class_id)\n\n task = Task.objects.get(pk=task_id)\n \n task.project = project\n task.description = description\n task.started = started\n task.finished = finished\n task.labor_item_number = labor_item\n task.li_class = li_class\n \n task.save()\n\n return HttpResponseRedirect('/kronos/%s/task/%s' % (timecard_id, task_id))\n else:\n timecard = Timecard.objects.get(pk=timecard_id)\n task = Task.objects.get(pk=task_id)\n li_classes = LaborClass.objects.all()\n labor_groups = LaborGroup.objects.all()\n projects = hq_models.Project.objects.all().filter(archived=False)\n return render(request, 'kronos/task_update.html', {\n 'timecard': timecard,\n 'task': task,\n 'projects': projects,\n 'labor_groups': labor_groups,\n 'li_classes': li_classes\n })\n\n@login_required(login_url=\"/login/\")\ndef task_copy(request, timecard_id, task_id):\n \"\"\"/kronos/1/task/1/delete/\n Copy task with id, replace times with today\n \"\"\"\n t = Task.objects.get(pk=task_id)\n t.pk = None\n\n today = datetime.datetime.now()\n \n t.started = t.started.replace(year=today.year, month=today.month, day=today.day)\n t.finished = t.finished.replace(year=today.year, month=today.month, day=today.day)\n t.save()\n return HttpResponseRedirect('/kronos/%s/' % (timecard_id))\n\n\n@login_required(login_url=\"/login/\")\ndef task_delete(request, timecard_id, task_id):\n \"\"\"/kronos/1/task/1/delete/\n Delete the task with id.\n \"\"\"\n t = Task.objects.get(pk=task_id)\n t.delete()\n return HttpResponseRedirect('/kronos/%s' % (timecard_id))\n","sub_path":"kronos/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":9240,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"476943599","text":"import hypothesis.strategies as st\nimport pytest\nfrom hypothesis import given, settings\n\nfrom slyther.parser import LParen, Quote, RParen, parse\nfrom slyther.types import NIL, Quoted, SExpression\nfrom slyther.types import String as sl\nfrom slyther.types import Symbol as s\n\nlp, rp, q = LParen(), RParen(), Quote()\n\n\ndef s_expressions(*args, **kwargs):\n return st.builds(SExpression.from_iterable, st.lists(*args, **kwargs))\n\n\ndef quoteds(strategy):\n return st.builds(Quoted, strategy)\n\n\nsymbols = st.builds(\n s, st.from_regex(r'''[^0-9\"'();.\\s][^\"'();\\s]*''', fullmatch=True))\nstrings = st.builds(sl, st.text())\nast_objects = st.deferred(lambda: (symbols\n | strings\n | st.integers()\n | st.floats(\n allow_nan=False, allow_infinity=False)\n | quoteds(ast_objects)\n | s_expressions(ast_objects)))\nast_lists = st.lists(ast_objects)\n\n\ndef deparse(exprs):\n for expr in exprs:\n if type(expr) in {int, float, s, sl}:\n yield expr\n elif isinstance(expr, Quoted):\n yield q\n yield from deparse([expr.elem])\n elif expr is NIL:\n yield lp\n yield rp\n elif isinstance(expr, SExpression):\n yield lp\n yield from deparse(expr)\n yield rp\n else:\n raise ValueError(\"bad deparse\")\n\n\n@settings(deadline=None)\n@given(ast_lists)\ndef test_valid_parses(ast):\n assert list(parse(deparse(ast))) == ast\n\n\ndef test_missing_rp():\n parser = parse(iter([lp, rp, lp, lp, rp]))\n assert next(parser) is NIL\n with pytest.raises(SyntaxError):\n next(parser)\n\n\ndef test_missing_lp():\n parser = parse(iter([10, rp]))\n assert next(parser) == 10\n with pytest.raises(SyntaxError):\n next(parser)\n\n\ndef test_bad_quotation_end():\n parser = parse(iter([q, -15, q, q]))\n assert next(parser) == Quoted(-15)\n with pytest.raises(SyntaxError):\n next(parser)\n\n\ndef test_bad_quotation_rp():\n parser = parse(iter([q, s('a'), lp, q, q, q, rp]))\n assert next(parser) == Quoted(s('a'))\n with pytest.raises(SyntaxError):\n next(parser)\n\n\n@settings(deadline=None)\n@given(st.integers(min_value=1, max_value=75))\ndef test_many_quotes(quotes):\n parser = parse(iter([q] * quotes + [lp, rp]))\n r = next(parser)\n for _ in range(quotes):\n assert isinstance(r, Quoted)\n r = r.elem\n assert r is NIL\n with pytest.raises(StopIteration):\n next(parser)\n\n\n@settings(deadline=None)\n@given(st.integers(min_value=1, max_value=10), s_expressions(ast_objects))\ndef test_many_quote_nested(quotes, inner):\n parser = parse(iter([q] * quotes + [lp, inner, rp]))\n r = next(parser)\n for _ in range(quotes):\n assert isinstance(r, Quoted)\n r = r.elem\n assert r == SExpression(inner)\n with pytest.raises(StopIteration):\n next(parser)\n","sub_path":"Mines Courses/CSCI_400/SlytherLisp/tests/d2/test_parser - Copy.py","file_name":"test_parser - Copy.py","file_ext":"py","file_size_in_byte":3033,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"197318482","text":"\n\n#calss header\nclass _NEPOTISM():\n\tdef __init__(self,): \n\t\tself.name = \"NEPOTISM\"\n\t\tself.definitions = [u'the act of using your power or influence to get good jobs or unfair advantages for members of your own family: ']\n\n\t\tself.parents = []\n\t\tself.childen = []\n\t\tself.properties = []\n\t\tself.jsondata = {}\n\n\n\t\tself.specie = 'nouns'\n\n\n\tdef run(self, obj1 = [], obj2 = []):\n\t\treturn self.jsondata\n","sub_path":"xai/brain/wordbase/nouns/_nepotism.py","file_name":"_nepotism.py","file_ext":"py","file_size_in_byte":395,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"598961530","text":"'''\n\n This file is part of MGEAR.\n\n MGEAR is free software: you can redistribute it and/or modify\n it under the terms of the GNU Lesser General Public License as published by\n the Free Software Foundation, either version 3 of the License, or\n (at your option) any later version.\n\n This program is distributed in the hope that it will be useful,\n but WITHOUT ANY WARRANTY; without even the implied warranty of\n MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n GNU Lesser General Public License for more details.\n\n You should have received a copy of the GNU Lesser General Public License\n along with this program. If not, see .\n\n Author: Jeremie Passerin geerem@hotmail.com\n Url: http://www.jeremiepasserin.com\n Date: 2011 / 07 / 13\n\n'''\n\n## @package mgear.maya.rig.component.arm_2jnt_01\n# @author Jeremie Passerin\n#\n#############################################\n# GLOBAL\n#############################################\n# Maya\nfrom pymel.core.general import *\nfrom pymel.core.animation import *\nfrom pymel.util import *\nimport pymel.core.datatypes as dt\n\nimport maya.OpenMaya as om\n\n# mgear\nfrom mgear.maya.rig.component import MainComponent\n\nimport mgear.maya.primitive as pri\nimport mgear.maya.transform as tra\nimport mgear.maya.attribute as att\nimport mgear.maya.node as nod\nimport mgear.maya.icon as ico\nimport mgear.maya.vector as vec\nimport mgear.maya.applyop as aop\nimport mgear.maya.fcurve as fcu\n\n#############################################\n# COMPONENT\n#############################################\nclass Component(MainComponent):\n\n def addObjects(self):\n\n self.normal = self.getNormalFromPos(self.guide.apos)\n\n self.length0 = vec.getDistance(self.guide.apos[0], self.guide.apos[1])\n self.length1 = vec.getDistance(self.guide.apos[1], self.guide.apos[2])\n self.length2 = vec.getDistance(self.guide.apos[2], self.guide.apos[3])\n\n # FK Controlers -----------------------------------\n t = tra.getTransformLookingAt(self.guide.apos[0], self.guide.apos[1], self.normal, \"xz\", self.negate)\n \n ### FK NEUTRAL POSE IS DIFFERENT\n self.fk0_npo = pri.addTransform(self.root, self.getName(\"fk0_npo\"), t)\n \n self.fk0_ctl = self.addCtl(self.fk0_npo, \"fk0_ctl\", t, self.color_fk, \"cube\", w=self.length0, h=self.size*.1, d=self.size*.1, po=dt.Vector(.5*self.length0*self.n_factor,0,0))\n\n t = tra.getTransformLookingAt(self.guide.apos[1], self.guide.apos[2], self.normal, \"xz\", self.negate)\n self.fk1_ctl = self.addCtl(self.fk0_ctl, \"fk1_ctl\", t, self.color_fk, \"cube\", w=self.length1, h=self.size*.1, d=self.size*.1, po=dt.Vector(.5*self.length1*self.n_factor,0,0))\n\n t = tra.getTransformLookingAt(self.guide.apos[2], self.guide.apos[3], self.normal, \"xz\", self.negate)\n self.fk2_ctl = self.addCtl(self.fk1_ctl, \"fk2_ctl\", t, self.color_fk, \"cube\", w=self.length2, h=self.size*.1, d=self.size*.1, po=dt.Vector(.5*self.length2*self.n_factor,0,0))\n self.fk_ctl = [self.fk0_ctl, self.fk1_ctl, self.fk2_ctl]\n\n # IK Controlers -----------------------------------\n\n self.ik_cns = pri.addTransformFromPos(self.root, self.getName(\"ik_cns\"), self.guide.pos[\"wrist\"])\n\n self.ikcns_ctl = self.addCtl(self.ik_cns, \"ikcns_ctl\", tra.getTransformFromPos(self.guide.pos[\"wrist\"]), self.color_ik, \"null\", w=self.size*.12)\n\n ### IK CONTROLER POSE IS DIFFERENT\n m = tra.getTransformLookingAt(self.guide.pos[\"wrist\"], self.guide.pos[\"eff\"], self.normal, \"xz\", False)\n self.ik_ctl = self.addCtl(self.ikcns_ctl, \"ik_ctl\", m, self.color_ik, \"cube\", w=self.size*.12, h=self.size*.12, d=self.size*.12)\n\n # upv\n v = self.guide.apos[2] - self.guide.apos[0]\n v = self.normal ^ v\n v.normalize()\n v *= self.size*.5\n v += self.guide.apos[1]\n\n self.upv_cns = pri.addTransformFromPos(self.root, self.getName(\"upv_cns\"), v)\n\n self.upv_ctl = self.addCtl(self.upv_cns, \"upv_ctl\", tra.getTransform(self.upv_cns), self.color_ik, \"diamond\", w=self.size*.12)\n att.setKeyableAttributes(self.upv_ctl, self.t_params)\n\n # References --------------------------------------\n self.ik_ref = pri.addTransform(self.ik_ctl, self.getName(\"ik_ref\"), tra.getTransform(self.ik_ctl))\n self.fk_ref = pri.addTransform(self.fk_ctl[2], self.getName(\"fk_ref\"), tra.getTransform(self.ik_ctl))\n\n # Chain --------------------------------------------\n # The outputs of the ikfk2bone solver\n self.bone0 = pri.addLocator(self.root, self.getName(\"0_jnt\"), tra.getTransform(self.fk_ctl[0]))\n self.bone0_shp = self.bone0.getShape()\n self.bone0_shp.setAttr(\"localPositionX\", self.n_factor*.5)\n self.bone0_shp.setAttr(\"localScale\", .5, 0, 0)\n self.bone0.setAttr(\"sx\", self.length0)\n self.bone0.setAttr(\"visibility\", False)\n\n self.bone1 = pri.addLocator(self.root, self.getName(\"1_jnt\"), tra.getTransform(self.fk_ctl[1]))\n self.bone1_shp = self.bone1.getShape()\n self.bone1_shp.setAttr(\"localPositionX\", self.n_factor*.5)\n self.bone1_shp.setAttr(\"localScale\", .5, 0, 0)\n self.bone1.setAttr(\"sx\", self.length1)\n self.bone1.setAttr(\"visibility\", False)\n\n self.ctrn_loc = pri.addTransformFromPos(self.root, self.getName(\"ctrn_loc\"), self.guide.apos[1])\n self.eff_loc = pri.addTransformFromPos(self.root, self.getName(\"eff_loc\"), self.guide.apos[2])\n\n # Mid Controler ------------------------------------\n self.mid_ctl = self.addCtl(self.ctrn_loc, \"mid_ctl\", tra.getTransform(self.ctrn_loc), self.color_ik, \"sphere\", w=self.size*.2)\n\n # Twist references ---------------------------------\n x = dt.Vector(0,-1,0)\n x = x * tra.getTransform(self.eff_loc)\n z = dt.Vector(self.normal.x,self.normal.y,self.normal.z)\n z = z * tra.getTransform(self.eff_loc)\n\n m = tra.getRotationFromAxis(x, z, \"xz\", self.negate)\n m = tra.setMatrixPosition(m, tra.getTranslation(self.ik_ctl))\n\n self.tws0_loc = pri.addTransform(self.root, self.getName(\"tws0_loc\"), tra.getTransform(self.fk_ctl[0]))\n self.tws0_rot = pri.addTransform(self.tws0_loc, self.getName(\"tws0_rot\"), tra.getTransform(self.fk_ctl[0]))\n\n self.tws1_loc = pri.addTransform(self.ctrn_loc, self.getName(\"tws1_loc\"), tra.getTransform(self.ctrn_loc))\n self.tws1_rot = pri.addTransform(self.tws1_loc, self.getName(\"tws1_rot\"), tra.getTransform(self.ctrn_loc))\n\n self.tws2_loc = pri.addTransform(self.root, self.getName(\"tws2_loc\"), tra.getTransform(self.fk_ctl[2]))\n self.tws2_rot = pri.addTransform(self.tws2_loc, self.getName(\"tws2_rot\"), tra.getTransform(self.fk_ctl[2]))\n self.tws2_rot.setAttr(\"sx\", .001)\n\n # Divisions ----------------------------------------\n # We have at least one division at the start, the end and one for the elbow.\n self.divisions = self.settings[\"div0\"] + self.settings[\"div1\"] + 3\n\n self.div_cns = []\n for i in range(self.divisions):\n\n div_cns = pri.addTransform(self.root, self.getName(\"div%s_loc\" % i))\n\n self.div_cns.append(div_cns)\n\n self.addShadow(div_cns, i)\n\n # End reference ------------------------------------\n # To help the deformation on the wrist\n self.end_ref = pri.addTransform(self.tws2_rot, self.getName(\"end_ref\"), m)\n self.addShadow(self.end_ref, \"end\")\n\n def addAttributes(self):\n\n # Anim -------------------------------------------\n self.blend_att = self.addAnimParam(\"blend\", \"Fk/Ik Blend\", \"double\", self.settings[\"blend\"], 0, 1)\n self.roll_att = self.addAnimParam(\"roll\", \"Roll\", \"double\", 0, -180, 180)\n\n self.scale_att = self.addAnimParam(\"ikscale\", \"Scale\", \"double\", 1, .001, 99)\n self.maxstretch_att = self.addAnimParam(\"maxstretch\", \"Max Stretch\", \"double\", 1.5, 1, 99)\n self.slide_att = self.addAnimParam(\"slide\", \"Slide\", \"double\", .5, 0, 1)\n self.softness_att = self.addAnimParam(\"softness\", \"Softness\", \"double\", 0, 0, 1)\n self.reverse_att = self.addAnimParam(\"reverse\", \"Reverse\", \"double\", 0, 0, 1)\n self.roundness_att = self.addAnimParam(\"roundness\", \"Roundness\", \"double\", 0, 0, 1)\n self.volume_att = self.addAnimParam(\"volume\", \"Volume\", \"double\", 1, 0, 1)\n \n # Ref\n if self.settings[\"ikrefarray\"]:\n ref_names = self.settings[\"ikrefarray\"].split(\",\")\n if len(ref_names) > 1:\n self.ikref_att = self.addAnimEnumParam(\"ikref\", \"Ik Ref\", 0, self.settings[\"ikrefarray\"].split(\",\"))\n \n if self.settings[\"upvrefarray\"]:\n ref_names = self.settings[\"upvrefarray\"].split(\",\")\n if len(ref_names) > 1:\n self.upvref_att = self.addAnimEnumParam(\"upvref\", \"UpV Ref\", 0, self.settings[\"upvrefarray\"].split(\",\"))\n\n # Setup ------------------------------------------\n # Eval Fcurve\n self.st_value = fcu.getFCurveValues(self.settings[\"st_profile\"], self.divisions)\n self.sq_value = fcu.getFCurveValues(self.settings[\"sq_profile\"], self.divisions)\n \n self.st_att = [ self.addSetupParam(\"stretch_%s\"%i, \"Stretch %s\"%i, \"double\", self.st_value[i], -1, 0) for i in range(self.divisions) ]\n self.sq_att = [ self.addSetupParam(\"squash_%s\"%i, \"Squash %s\"%i, \"double\", self.sq_value[i], 0, 1) for i in range(self.divisions) ]\n\n self.resample_att = self.addSetupParam(\"resample\", \"Resample\", \"bool\", True)\n self.absolute_att = self.addSetupParam(\"absolute\", \"Absolute\", \"bool\", False)\n\n def addOperators(self):\n\n # Visibilities -------------------------------------\n # fk\n fkvis_node = nod.createReverseNode(self.blend_att)\n \n for shp in self.fk0_ctl.getShapes():\n connectAttr(fkvis_node+\".outputX\", shp.attr(\"visibility\"))\n for shp in self.fk1_ctl.getShapes():\n connectAttr(fkvis_node+\".outputX\", shp.attr(\"visibility\"))\n for shp in self.fk2_ctl.getShapes():\n connectAttr(fkvis_node+\".outputX\", shp.attr(\"visibility\"))\n\n # ik\n for shp in self.upv_ctl.getShapes():\n connectAttr(self.blend_att, shp.attr(\"visibility\"))\n for shp in self.ikcns_ctl.getShapes():\n connectAttr(self.blend_att, shp.attr(\"visibility\"))\n for shp in self.ik_ctl.getShapes():\n connectAttr(self.blend_att, shp.attr(\"visibility\"))\n\n # IK Solver -----------------------------------------\n out = [self.bone0, self.bone1, self.ctrn_loc, self.eff_loc]\n node = aop.gear_ikfk2bone_op(out, self.root, self.ik_ref, self.upv_ctl, self.fk_ctl[0], self.fk_ctl[1], self.fk_ref, self.length0, self.length1, self.negate)\n\n connectAttr(self.blend_att, node+\".blend\")\n connectAttr(self.roll_att, node+\".roll\")\n connectAttr(self.scale_att, node+\".scaleA\")\n connectAttr(self.scale_att, node+\".scaleB\")\n connectAttr(self.maxstretch_att, node+\".maxstretch\")\n connectAttr(self.slide_att, node+\".slide\")\n connectAttr(self.softness_att, node+\".softness\")\n connectAttr(self.reverse_att, node+\".reverse\")\n\n # Twist references ---------------------------------\n pointConstraint(self.root, self.tws0_loc, maintainOffset=True)\n aop.aimCns(self.tws0_loc, self.mid_ctl, self.n_sign+\"xz\", 2, [0,1,0], self.root, False)\n\n pointConstraint(self.mid_ctl, self.tws1_loc, maintainOffset=False)\n scaleConstraint(self.mid_ctl, self.tws1_loc, maintainOffset=False)\n orientConstraint(self.mid_ctl, self.tws1_rot, maintainOffset=False)\n\n pointConstraint(self.eff_loc, self.tws2_loc, maintainOffset=False)\n scaleConstraint(self.eff_loc, self.tws2_loc, maintainOffset=False)\n orientConstraint(self.bone1, self.tws2_loc, maintainOffset=False) \n # orientConstraint(self.eff_loc, self.tws2_rot, maintainOffset=False)\n node = aop.gear_mulmatrix_op(self.eff_loc.attr(\"worldMatrix\"), self.tws2_rot.attr(\"parentInverseMatrix\"))\n dm_node = createNode(\"decomposeMatrix\")\n connectAttr(node+\".output\", dm_node+\".inputMatrix\")\n connectAttr(dm_node+\".outputRotate\", self.tws2_rot+\".rotate\")\n # att.setRotOrder(self.tws2_rot, \"YZX\")\n\n self.tws0_loc.setAttr(\"sx\", .001)\n self.tws2_loc.setAttr(\"sx\", .001)\n\n add_node = nod.createAddNode(self.roundness_att, .001)\n connectAttr(add_node+\".output\", self.tws1_rot.attr(\"sx\"))\n\n # Volume -------------------------------------------\n distA_node = nod.createDistNode(self.tws0_loc, self.tws1_loc)\n distB_node = nod.createDistNode(self.tws1_loc, self.tws2_loc)\n add_node = nod.createAddNode(distA_node+\".distance\", distB_node+\".distance\")\n div_node = nod.createDivNode(add_node+\".output\", self.root.attr(\"sx\"))\n self.volDriver_att = div_node+\".outputX\"\n\n # Divisions ----------------------------------------\n # at 0 or 1 the division will follow exactly the rotation of the controler.. and we wont have this nice tangent + roll\n for i, div_cns in enumerate(self.div_cns):\n\n if i < (self.settings[\"div0\"]+1):\n perc = i*.5 / (self.settings[\"div0\"]+1.0)\n else:\n perc = .5 + (i-self.settings[\"div0\"]-1.0)*.5 / (self.settings[\"div1\"]+1.0)\n\n perc = max(.001, min(.999, perc))\n\n # Roll\n if self.negate:\n node = aop.gear_rollsplinekine_op(div_cns, [self.tws2_rot, self.tws1_rot, self.tws0_rot], 1-perc)\n else:\n node = aop.gear_rollsplinekine_op(div_cns, [self.tws0_rot, self.tws1_rot, self.tws2_rot], perc)\n\n connectAttr(self.resample_att, node+\".resample\")\n connectAttr(self.absolute_att, node+\".absolute\")\n\n # Squash n Stretch\n node = aop.gear_squashstretch2_op(div_cns, None, getAttr(self.volDriver_att), \"x\")\n connectAttr(self.volume_att, node+\".blend\")\n connectAttr(self.volDriver_att, node+\".driver\")\n connectAttr(self.st_att[i], node+\".stretch\")\n connectAttr(self.sq_att[i], node+\".squash\")\n\n return\n\n # =====================================================\n # CONNECTOR\n # =====================================================\n ## Set the relation beetween object from guide to rig.\\n\n # @param self\n def setRelation(self):\n self.relatives[\"root\"] = self.bone0\n self.relatives[\"elbow\"] = self.bone1\n self.relatives[\"wrist\"] = self.eff_loc ## NOT GOOD TEMP !!!\n self.relatives[\"eff\"] = self.eff_loc ## NOT GOOD TEMP !!!\n\n ## standard connection definition.\n # @param self\n def connect_standard(self):\n self.connect_standardWithIkRef()","sub_path":"pythonlibs/mgear/maya/rig/component/arm_2jnt_01/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":14914,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"444580116","text":"import re\nimport os\nimport sys\nimport numpy as np\nfrom pathlib import Path\nfrom glob import iglob\nimport tensorflow as tf\nimport sentencepiece as spm\nimport csv\nimport pandas as pd\nimport argparse\n\nsys.path.append(os.path.dirname(os.path.dirname(__file__)))\n\nfrom model.transformer import transformer, CustomSchedule\nfrom module.dirHandler import mkdir_p, del_folder\nfrom module.encoder import IntegerEncoder\nfrom module.decoder import Decoder\nfrom module.parse import ParseBoolean\nfrom sklearn.model_selection import train_test_split\n\nBASE_DIR = os.getcwd()\nDATA_BASE_DIR = os.path.join(BASE_DIR, 'articles')\nSRC_BASE_DIR = os.path.join(BASE_DIR, 'src')\n\nVAL_PREPROCESSED_PATH = os.path.join(DATA_BASE_DIR,\"Valid-Preprocessed-Data\")\nVAL_SUMMARY_PREPROCESSED_PATH = os.path.join(DATA_BASE_DIR,\"Valid-Summary-Preprocessed-Data\")\nTITLE_PREPROCESSED_PATH= os.path.join(DATA_BASE_DIR,\"Title-Preprocessed-Data\")\n\nPREPROCESSED_PATH = os.path.join(DATA_BASE_DIR,\"Preprocessed-Data\")\nSUMMARY_PREPROCESSED_PATH = os.path.join(DATA_BASE_DIR,\"Summary-Preprocessed-Data\")\n\nTRANSFORMER_PREDICT_PATH = os.path.join(DATA_BASE_DIR,\"Transformer-Predict-Data\")\n\nWORD_ENCODING_DIR = os.path.join(SRC_BASE_DIR, 'Word-Encoding-Model')\nMODEL_DIR = os.path.join(SRC_BASE_DIR, 'trained-model')\nTRANS_MODEL_DIR = os.path.join(MODEL_DIR, 'Transformer')\n\n# Get argument to determine the process\nparser = argparse.ArgumentParser(description=\"Description\")\nparser.add_argument('--headline', required=True, type=ParseBoolean, help=\"If True, Generating Headline else Generating Summary\")\nargs = parser.parse_args()\n\n# Load Sentencepiece word encoding model\nsp = spm.SentencePieceProcessor()\nmodel_num = len(list(iglob(os.path.join(WORD_ENCODING_DIR, 'spm-input-*.vocab'), recursive=False))) -1\nwith open(os.path.join(WORD_ENCODING_DIR, 'spm-input-{}.vocab'.format(model_num)), encoding='utf-8') as f:\n Vo = [doc.strip().split(\"\\t\") for doc in f]\n\nsp.Load(os.path.join(WORD_ENCODING_DIR, 'spm-input-{}.model').format(model_num))\n\nD_MODEL = 128\nVOCAB_SIZE = len(Vo)\nLAYER_NUM = 6\nNUM_HEADS = 8\nDFF = 512\n\nBATCH_SIZE = 64\nBUFFER_SIZE = 5000\n\nWARMUP_STEPS = 50\nEPOCHS = 30\nSTART_TOKEN = [sp.bos_id()]\nEND_TOKEN = [sp.eos_id()]\n\nget_max_length = lambda x : np.max([len(line) for line in x])\n\ndef loss_function(y_true, y_pred):\n y_true = tf.reshape(y_true, shape=(-1, MAX_LEN-1))\n\n loss = tf.keras.losses.SparseCategoricalCrossentropy(\n from_logits=True, reduction='none')(y_true, y_pred)\n\n mask = tf.cast(tf.not_equal(y_true, 0), tf.float32)\n loss = tf.multiply(loss, mask)\n\n return tf.reduce_mean(loss)\n\n\nif __name__ == '__main__':\n\n options = {\n 'model-type' : 'Sentence-Piece',\n 'inv_wv' : None,\n 'corpus' : None,\n 'spm' : sp\n }\n\n # src & target path depend on the process\n src_data_path = PREPROCESSED_PATH\n if args.headline:\n target_data_path = TITLE_PREPROCESSED_PATH\n else :\n target_data_path = SUMMARY_PREPROCESSED_PATH\n\n # Load src & target data for training model\n # src & target data integer encoding \n input_encoded_list = IntegerEncoder(options=options, filepaths=list(iglob(os.path.join(src_data_path, '**.csv'), recursive=False))).encoder()\n output_encoded_list = IntegerEncoder(options=options, filepaths=list(iglob(os.path.join(target_data_path, '**.csv'), recursive=False))).encoder()\n \n MAX_LEN = get_max_length(input_encoded_list) + 2\n SUMMARY_MAX_LEN = get_max_length(output_encoded_list) + 2\n\n # add SOS & EOS Token (Start of Sentence, End of Sentence)\n input_encoded_list = list(map(lambda list_ : START_TOKEN + list_ + END_TOKEN, input_encoded_list))\n output_encoded_list = list(map(lambda list_ : START_TOKEN + list_ + END_TOKEN, output_encoded_list))\n\n # Divide into Train dataset & Validation dataset\n input_train, input_test, output_train, output_test = train_test_split(\n input_encoded_list, output_encoded_list, test_size=0.2, random_state=42)\n\n # Padding\n train_input_encoded_matrix = tf.keras.preprocessing.sequence.pad_sequences(\n input_train, maxlen=MAX_LEN, padding='post')\n train_summary_encoded_matrix = tf.keras.preprocessing.sequence.pad_sequences(\n output_train, maxlen=MAX_LEN, padding='post')\n test_input_encoded_matrix = tf.keras.preprocessing.sequence.pad_sequences(\n input_test, maxlen=MAX_LEN, padding='post')\n test_summary_encoded_matrix = tf.keras.preprocessing.sequence.pad_sequences(\n output_test, maxlen=MAX_LEN, padding='post')\n\n print('Train Contents Shape : {}'.format(train_input_encoded_matrix.shape))\n print('Train Summaries Shape : {}'.format(train_summary_encoded_matrix.shape))\n print('Test Contents Shape : {}'.format(test_input_encoded_matrix.shape))\n print('Test Summaries Shape : {}'.format(test_summary_encoded_matrix.shape))\n \n dataset = tf.data.Dataset.from_tensor_slices((\n {\n 'inputs': train_input_encoded_matrix, # Encoder Input\n 'dec_inputs': train_summary_encoded_matrix[:, :-1] # Decoder Input\n },\n {\n # Decoder Output, Remove \n 'outputs': train_summary_encoded_matrix[:, 1:] \n },\n ))\n dataset = dataset.cache()\n dataset = dataset.shuffle(BUFFER_SIZE)\n dataset = dataset.batch(BATCH_SIZE)\n dataset = dataset.prefetch(tf.data.experimental.AUTOTUNE)\n\n val_dataset = tf.data.Dataset.from_tensor_slices((\n {\n 'inputs': test_input_encoded_matrix, # Encoder Input\n 'dec_inputs': test_summary_encoded_matrix[:, :-1] # Decoder Input\n },\n {\n # Decoder Output, Remove \n 'outputs': test_summary_encoded_matrix[:, 1:] \n },\n ))\n val_dataset = val_dataset.cache()\n val_dataset = val_dataset.shuffle(BUFFER_SIZE)\n val_dataset = val_dataset.batch(BATCH_SIZE)\n val_dataset = val_dataset.prefetch(tf.data.experimental.AUTOTUNE)\n\n # Declaring optimizer\n lrate_scheduler = CustomSchedule(d_model=D_MODEL)\n beta_1 = 0.9 \n beta_2 = 0.98\n epsilon = 10 ** -9\n\n optimizer = tf.keras.optimizers.Adam(lrate_scheduler, beta_1=0.9, beta_2=0.98, epsilon=1e-9)\n\n # Initialize Transformer\n model = transformer(\n vocab_size=VOCAB_SIZE,\n num_layers=LAYER_NUM,\n dff=DFF,\n d_model=D_MODEL,\n num_heads=NUM_HEADS,\n dropout = 0.3)\n\n\n # Initialize model train checkpoint\n mkdir_p(checkpoint_dirpath)\n checkpoint_filepath = os.path.join(TRANS_MODEL_DIR, \"checkpoint.ckpt\")\n model_checkpoint_callback = tf.keras.callbacks.ModelCheckpoint(\n filepath=checkpoint_filepath,\n save_weights_only=True,\n monitor='loss',\n mode='max',\n save_best_only=True)\n \n # Training Model\n model.compile(optimizer=optimizer, loss=loss_function)\n model.summary()\n\n model.fit(dataset, epochs=30, verbose=2, validation_data=val_dataset, shuffle=True, callbacks=[model_checkpoint_callback])\n","sub_path":"src/train/train_transformer.py","file_name":"train_transformer.py","file_ext":"py","file_size_in_byte":6965,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"579518782","text":"from django.db import models\nfrom django.conf import settings\nfrom ckeditor.fields import RichTextField\n\nSeverity = (\n (1,'Select'),\n (2, 'Critical'),\n (3,'Major'),\n (4,'Moderate'),\n (5,'Minor'),\n (6,'Cosmetic'),\n)\nPriority=(\n (1,'Low'),\n (2, 'Medium'),\n (3,'High'),\n)\nReproducibility=(\n (1,'10%'),\n (2,'25%'),\n (3,'50%'),\n (4,'75%'),\n (5,'100%'),\n)\nclass Report(models.Model):\n '''Model for managing discussions'''\n owner = models.ForeignKey(settings.AUTH_USER_MODEL, on_delete=models.CASCADE)\n resourcetitle = models.CharField(blank=False,max_length=500, )\n severity = models.IntegerField(choices=Severity,blank=\"false\", default=1, verbose_name='Severity')\n Priority = models.IntegerField(choices=Priority,blank=\"false\", default=1, verbose_name='Priority')\n Reproducibility = models.IntegerField(choices=Reproducibility,blank=\"false\", default=1, verbose_name='Reproducibility')\n media =models.ImageField(upload_to='images/', blank=True, verbose_name='Provide Picture/Video')\n Description= RichTextField(blank=False,default=\"\", verbose_name='Description of the Issue')\n created_time = models.DateTimeField(auto_now_add=True,)\n def get_absolute_url(self):\n '''Returns the URL of the discussion'''\n return \"/report/%i/\" % self.id\n\n class Meta:\n ordering = ['-created_time']\n\nclass ReportParticipant(models.Model):\n '''Model for managing a discussion's participant'''\n participant = models.ForeignKey(settings.AUTH_USER_MODEL, on_delete=models.CASCADE)\n report = models.ForeignKey(Report, on_delete=models.CASCADE)","sub_path":"collabplatform/report/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":1617,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"118510011","text":"# coding: utf-8\nimport time\nfrom MashiroSpider import Middleware\nimport logging\n\n\nclass SleepController(Middleware):\n\n def __init__(self, sleep):\n self._last_time = time.time()\n self._sleep = sleep\n self.logger = logging.getLogger('SleepController')\n\n def before_download(self, request):\n release = self._sleep - time.time() + self._last_time\n\n if release > 0:\n self.logger.debug('Sleep %d second...', release)\n # time.sleep(release)\n time.sleep(10)\n self.logger.debug('sleep done')\n self._last_time = time.time()\n return request\n\n @classmethod\n def from_settings(cls, settings):\n return cls(settings.SLEEP)\n","sub_path":"request_middlewares/SleepController.py","file_name":"SleepController.py","file_ext":"py","file_size_in_byte":720,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"519594735","text":"import os\nimport pygame\nimport time\n\nclass Controller(object):\n\n\tos.environ['SDL_VIDEODRIVER']=\"dummy\"\n\tcontroller = None\n\n\tdef init(self):\n\t\tpygame.init()\n\t\tpygame.joystick.init()\n\t\tself.controller = pygame.joystick.Joystick(0)\n\t\tself.controller.init()\n\n\tdef listen(self):\n\t\twhile True:\n\t\t\tfor event in pygame.event.get():\n\t\t\t\tif event.type == pygame.JOYBUTTONDOWN:\n\t\t\t\t\treturn event.button\n\n\tdef game_two_listen(self,interval_time,bool):\n\t\tstart_time=time.time()\n\t\tcheck_time=time.time()\n\t\toutput=None\n\t\tbutton_pressed=False\n\n\t\twhile check_time - start_time < interval_time:\n\t\t\tfor event in pygame.event.get():\n\t\t\t\tif event.type == pygame.JOYBUTTONDOWN:\n\t\t\t\t\tbutton_pressed=True\n\n\t\t\tcheck_time = time.time()\n\n\t\t\tif button_pressed==True and bool==True:\n\t\t\t\toutput=True\n\t\t\telif button_pressed==True and bool==False:\n\t\t\t\toutput=False\n\t\t\telif button_pressed==False:\n\t\t\t\toutput=None\n\t\treturn output\n","sub_path":"controller_test.py","file_name":"controller_test.py","file_ext":"py","file_size_in_byte":896,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"582205336","text":"from datetime import datetime\nfrom datetime import timedelta\nfrom datetime import date\nimport logging\nimport requests\nfrom bs4 import BeautifulSoup as Bs\nfrom requests import HTTPError\n\nfrom application.models.model import WeatherNow, Forecast\n\nANGOLA_PROVINCES = ['Bengo', 'Benguela', 'Kuito', 'Cabinda', 'Menongue', \"N'dalatando\", 'Sumbe', 'Ondjiva',\n 'Huambo', 'Lubango', 'Luanda', 'Dundo', 'Saurimo', 'Malanje', 'Luena', 'Namibe', 'Uíge']\n\nWEEK = [\"Monday\", \"Tuesday\", \"Wednesday\", \"Thursday\", \"Friday\", \"Saturday\", \"Sunday\"]\n\nUSER_AGENT = \"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.77 \" \\\n \"Safari/537.36\"\nLANGUAGE = \"en-gb;q=0.8, en;q=0.7\"\n\nsession = requests.Session()\nsession.headers['User-Agent'] = USER_AGENT\nsession.headers['Accept-Language'] = LANGUAGE\nsession.headers['Content-Language'] = LANGUAGE\n\n\nclass WeatherUtilities(object):\n\n @classmethod\n def update_progress(cls, progress):\n print(\"\\rProgress: [{0:50s}] {1:.1f}%\".format('#' * int(progress * 50), progress * 100), end=\"\", flush=True)\n\n @classmethod\n def get_weather_soup(cls, url):\n try:\n html = session.get(url)\n soup = Bs(html.text, \"html.parser\")\n except HTTPError as error:\n # print(f\"get weather soup error with code {HTTPError}\")\n logging.debug(\"weather soup http error code : %s\", error)\n except ConnectionError as error:\n logging.debug(\"weather soup connection error code : %s\", error)\n finally:\n return soup\n\n @classmethod\n def get_weather_now(cls, url):\n data_weather = {}\n progress = 0\n try:\n for region in ANGOLA_PROVINCES:\n\n progress += 1\n working = progress / len(ANGOLA_PROVINCES)\n cls.update_progress(working)\n\n temp_url = url + f\"+{region}\"\n soup = cls.get_weather_soup(temp_url)\n result = dict()\n # extract data\n result['city_name'] = soup.find(\"div\", attrs={\"id\": \"wob_loc\"}).text\n result['temperature'] = soup.find(\"span\", attrs={\"id\": \"wob_tm\"}).text\n time_tmp = soup.find(\"div\", attrs={\"id\": \"wob_dts\"}).text\n result['time_of_day'] = cls.time_weatherdata (time_tmp)\n result['description'] = soup.find(\"span\", attrs={\"id\": \"wob_dc\"}).text\n result['preciptation'] = soup.find(\"span\", attrs={\"id\": \"wob_pp\"}).text\n result['humidity'] = soup.find(\"span\", attrs={\"id\": \"wob_hm\"}).text\n result['wind'] = soup.find(\"span\", attrs={\"id\": \"wob_ws\"}).text\n\n next_days = []\n date_initial = 0\n days = soup.find(\"div\", attrs={\"id\": \"wob_dp\"})\n for day in days.findAll(\"div\", attrs={\"class\": \"wob_df\"}):\n # extract the name of the day\n day_name = day.find(\"div\", attrs={\"class\": \"QrNVmd Z1VzSb\"}).attrs['aria-label']\n # get weather status for that day\n description = day.find(\"img\").attrs[\"alt\"]\n temp = day.findAll(\"span\", {\"class\": \"wob_t\"})\n # maximum temparature in Celsius, use temp[1].text if you want fahrenheit\n max_temp = temp[0].text\n # minimum temparature in Celsius, use temp[3].text if you want fahrenheit\n min_temp = temp[2].text\n date = cls.get_next_date(date_initial)\n date_initial = date_initial + 1\n\n n_day = Forecast(day_name, date, description, min_temp, max_temp)\n next_days.append(n_day)\n\n result['week_weather'] = next_days\n\n _weather = WeatherNow(result['city_name'], result['time_of_day'], result['temperature'],\n result['description'], result['preciptation'], result['humidity'],\n result['wind'], result['week_weather'])\n\n data_weather.update({_weather.city_name: _weather})\n except HTTPError:\n logging.debug(\"weather data error code :%s\", HTTPError)\n except ConnectionError:\n logging.debug(\"weather data connection error code :%s\", ConnectionError)\n except Exception as error:\n logging.debug(\"main error getting weather data :%s\", error)\n finally:\n return data_weather\n\n @classmethod\n def add_weather_db(cls, weather_data, db):\n weather_db = WeatherNow.query.all()\n print(weather_db)\n status = False\n try:\n if (list(weather_data.values()) != 0) and (len(weather_db) == 0):\n db.session.add_all(list(weather_data.values()))\n db.session.commit()\n status = True\n except Exception:\n logging.debug(\"error adding the weather to db: %s\", Exception)\n finally:\n db.session.close()\n return status\n\n\n\n @classmethod\n def update_weather(cls,weather_database , db):\n try:\n db.session.query(WeatherNow).filter(WeatherNow.city_name ==\n weather_database.city_name) \\\n .update({'city_name': weather_database.city_name,\n 'time_of_day': weather_database.time_of_day,\n 'temperature': weather_database.temperature,\n 'description': weather_database.description,\n 'preciptation': weather_database.preciptation,\n 'humidity': weather_database.humidity,\n 'wind': weather_database.wind})\n\n for week_day in weather_database.forecast_week:\n db.session.query(Forecast).filter(Forecast.weather_id == weather_database.id,\n Forecast.day == week_day.day).update(\n {'day': week_day.day,\n 'date': week_day.date,\n 'description': week_day.description,\n 'min_temperature': week_day.min_temperature,\n 'max_temperature': week_day.max_temperature})\n\n db.session.commit()\n except Exception:\n logging.debug(\"error updating a weather data:%s\", Exception)\n\n @classmethod\n def update_weather_db(cls, weather_data, db):\n weather_db = WeatherNow.query.all()\n try:\n if len(weather_data.values()) != 0 and len(weather_db) != 0:\n \"\"\"filter database by date to check if passed 1hr to update weather database .\"\"\"\n checked_date_weather = filter(lambda weather : cls.check_date(weather.time_of_day) , weather_db)\n \"\"\"update the filtered ones\"\"\"\n updated_weather = [cls.update_weather( weather_data[weather.city_name] , db ) for weather in checked_date_weather]\n\n if not updated_weather :\n logging.info(\"No need to update Weather database\")\n except Exception:\n logging.debug(\"error updating the weather database :%s\", Exception)\n finally:\n db.session.close()\n\n @classmethod\n def time_weatherdata (cls, t_stamp):\n\n time_system = datetime.now()\n time_weather =t_stamp.split(' ')[1] .split(':')\n w_hour = time_weather[0]\n w_minutes = time_weather[1]\n time_of_day = time_system.replace(hour=int(w_hour), minute=int(w_minutes), second=0)\n\n return time_of_day.strftime('%Y-%m-%d %H:%M:%S.%f')\n\n @classmethod\n def check_date(cls, date_db):\n\n condition = False\n try:\n time_db = datetime.strptime(date_db, '%Y-%m-%d %H:%M:%S.%f')\n time_now = datetime.now()\n delta_time = time_now - time_db\n diff_time_hr = delta_time.total_seconds() / 3600\n\n if (diff_time_hr > 1):\n condition = True\n except Exception:\n logging.debug(\"Unable to check date :%s\", Exception)\n finally:\n return condition\n\n @classmethod\n def get_next_date(cls, next_day):\n next_day = date.today() + timedelta(days=next_day)\n return next_day\n","sub_path":"application/utilities/weather_utilities.py","file_name":"weather_utilities.py","file_ext":"py","file_size_in_byte":8312,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"373843810","text":"# coding=utf-8\n\"\"\"Log handler that send emails\n============================\n\nOverrides :class:`django.utils.log.AdminEmailHandler` to avoid flooding admin with many e-mails.\n\nIf an e-mail with the same object has already been sent in the last 10 minutes, then nothing is done.\nThis duration can be configured (argument `min_interval`).\n\n\"\"\"\nfrom __future__ import unicode_literals\nimport datetime\n\nfrom django.core import mail\nfrom django.utils.log import AdminEmailHandler\nimport io\nimport sys\n\n\n__author__ = 'Matthieu Gallet'\n\nLAST_SENDS = {}\n\n\nclass FloorAdminEmailHandler(AdminEmailHandler):\n \"\"\"An exception log handler that emails log entries to site admins.\n\n If the request is passed as the first argument to the log record,\n request data will be provided in the email report.\n \"\"\"\n def __init__(self, include_html=False, email_backend=None, min_interval=600):\n super(FloorAdminEmailHandler, self).__init__(include_html=include_html, email_backend=email_backend)\n self.min_interval = min_interval\n\n def send_mail(self, subject, message, *args, **kwargs):\n interval = datetime.datetime.now() - LAST_SENDS.get(subject, datetime.datetime(1970, 1, 1))\n if interval < datetime.timedelta(0, self.min_interval):\n return\n LAST_SENDS[subject] = datetime.datetime.now()\n original_std = sys.stdout, sys.stderr\n # sys.stdout = io.StringIO()\n sys.stderr = io.StringIO()\n try:\n pass\n mail.mail_admins(subject, message, *args, connection=self.connection(), **kwargs)\n except Exception as e:\n print('<==================================================================================================')\n print('Unable to send the mail: %s' % e)\n print(subject)\n print('===================================================================================================')\n print('/!\\\\ settings.DEBUG = False BUT STILL UNABLE TO SEND MAIL TO ADMIN /!\\\\')\n print('==================================================================================================>')\n sys.stdout, sys.stdout = original_std","sub_path":"djangofloor/log.py","file_name":"log.py","file_ext":"py","file_size_in_byte":2186,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"99799366","text":"import cv2\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport math\n\n# Recibe dos puntos (P,Q) y los centroides de cada uno para calcular la distancia\n# y realizar la comparación para filtrar la lista de puntos con los que se realizará\n# el matching.\ndef validarPuntos(p,q,c_1,c_2,umbral):\n p_x,p_y,p_i = p\n q_x,q_y,q_i = q\n c_x_p,c_y_p,c_i_p = c_1[0]\n c_x_q,c_y_q,c_i_q = c_2[0]\n if abs(math.dist([p_x,p_y],[c_x_p,c_y_p]) - math.dist([q_x,q_y],[c_x_q,c_y_q])) < umbral:\n return True\n else:\n return False\n\nimg1 = cv2.imread(\"img/picadef1.png\")\nimg2 = cv2.imread(\"img/picadef4.png\")\n\n#Utilizando FAST para obtener los KP\nfast = cv2.FastFeatureDetector_create()\nkp1 = fast.detect(img1, None)\nkp2 = fast.detect(img2, None)\n\n#Obteniendo los descriptores con BRIEF\nbrief = cv2.xfeatures2d.BriefDescriptorExtractor_create()\nkpx, des1 = brief.compute(img1,kp1)\nkpy, des2 = brief.compute(img2,kp2)\n\n#Brute Force Matching\nbf = cv2.BFMatcher(cv2.NORM_HAMMING, crossCheck = True)\nmatches = bf.match(des1,des2)\n\n#Obtener las coordenadas del brute force matching\nlist_kp1 = []\nlist_kp2 = []\nfor m in matches:\n img1_index = m.queryIdx\n img2_index = m.trainIdx\n #Coordenadas\n (x1,y1) = kp1[img1_index].pt\n (x2,y2) = kp2[img2_index].pt\n list_kp1.append((x1,y1,img1_index))\n list_kp2.append((x2,y2,img2_index))\n\n\n\n#K-Means\ncriteria = (cv2.TERM_CRITERIA_EPS + cv2.TERM_CRITERIA_MAX_ITER, 20, 1.0)\n\n#Primera imagen\nlist_kp1 = np.float32(np.vstack(list_kp1))\nret_1, label_1, center_1 = cv2.kmeans(list_kp1, 1, None, criteria, 10, cv2.KMEANS_RANDOM_CENTERS)\n\n#Seguna imagen\nlist_kp2 = np.float32(np.vstack(list_kp2))\nret_2, label_2, center_2 = cv2.kmeans(list_kp2, 1, None, criteria, 10, cv2.KMEANS_RANDOM_CENTERS)\n\n#Calcular las distancia\nlist_validos = []\nfor i in range(len(list_kp1)):\n if validarPuntos(list_kp1[i],list_kp2[i],center_1,center_2,20):\n list_validos.append(matches[i])\n \nprint(len(matches))\nprint(len(list_validos))\n\nresult = cv2.drawMatches(img1, kp1, img2, kp2, list_validos, None, flags=2)\n\ncv2.imshow(\"Detector BRIEF & FAST + KMeans\",result)\ncv2.waitKey(0)\ncv2.destroyAllWindows()","sub_path":"src/kmeans/match_brief_fast_kmeans.py","file_name":"match_brief_fast_kmeans.py","file_ext":"py","file_size_in_byte":2155,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"642758023","text":"import serial\nfrom flask import Flask, render_template, request, flash, session, redirect, url_for\nfrom models import db, User\nimport datetime\nimport time\n\nclass Kitchen():\n\n\tserial = \" \"\n\n\tdef __init__(self, serial):\n\t\tself.serial = serial\n\n\tdef renderPage(self):\n now = datetime.datetime.now()\n timeString = now.strftime(\"%h %d, %Y %H:%M\")\n state = self.getStateOfLight()\n\n templateData = {\n 'title' : 'Kitchen',\n 'time' : timeString,\n 'state' : state\n }\n return render_template(\"kitchen.html\", **templateData)\n\n\tdef toggleLed(self, action):\n \tif action == 'on':\n \tself.serial.write('1')\n \telse:\n \tself.serial.write('0')\n\n\tdef getStateOfLight(self):\n self.serial.write('2')\n state = self.serial.read()\n state = str(state[0])\n\n return state\n\n\tdef validate(self):\n if 'username' not in session:\n return redirect(url_for('signin'))\n\n user = User.query.filter_by(username = session['username']).first()\n\n if user is None:\n return redirect(url_for('signin'))\n else:\n return self.renderPage()\n","sub_path":"jarvisapp/application/kitchen.py","file_name":"kitchen.py","file_ext":"py","file_size_in_byte":1381,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"205941419","text":"\nfrom Parser import Parser\n\nsyntax = (\n{ 'root' : ['*any'\n#\t\t, 'IF'\n#\t\t, 'CMP'\n#\t\t, 'ASSIGN'\n\t\t, 'EXPR']\n, 'IF' : ['<>if', 'CMP', ['*any', 'ASSIGN', 'EXPR']]\n, 'ASSIGN' : ['', '<>=', 'CMP']\n, 'EXPR': 'CMP'\n, 'CMP': ['ADD', ['*opt', '<>==', 'ADD']]\n, 'ADD' : ['VAL', ['*rep', '<>+', 'VAL']]\n, 'VAL' : ['*any', '', ''\n#\t\t, ['<>(', 'EXPR', '<>)']\n\t\t]\n})\n\nparser = Parser(syntax=syntax)\n\nfor ky in parser.syntax.keys():\n\tprint(ky)\n\tprint(\" \", (parser.syntax[ky]))\n\n\nsrc = input()\n\nsuccess, src2, tree = parser.parse(src)\n\nprint(\"success:\", success)\nprint(\"rest:\", src2)\nprint(\"tree:\", tree)\n","sub_path":"tes.py","file_name":"tes.py","file_ext":"py","file_size_in_byte":600,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"643729954","text":"#给你一个整数 n,请你帮忙计算并返回该整数「各位数字之积」与「各位数字之和」的差。\n\n#示例 1:\n#输入:n = 234\n#输出:15 \n#解释:\n#各位数之积 = 2 * 3 * 4 = 24 \n#各位数之和 = 2 + 3 + 4 = 9 \n#结果 = 24 - 9 = 15\n\n#示例 2:\n#输入:n = 4421\n#输出:21\n#解释: \n#各位数之积 = 4 * 4 * 2 * 1 = 32 \n#各位数之和 = 4 + 4 + 2 + 1 = 11 \n#结果 = 32 - 11 = 21\n\n#提示:\n#1 <= n <= 10^5\n\n# enconding: utf-8\ndef test(n):\n Product = 1\n Sum = 0\n for i in str(n):\n i = int(i)\n Sum += i \n Product *= i\n\n return Product - Sum\n\nn = 4421\nprint(test(n))\n","sub_path":"practice/prac_lecode/three_day/sum_ji.py","file_name":"sum_ji.py","file_ext":"py","file_size_in_byte":648,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"587255229","text":"import sys\n\ndef getInputcheckISBN():\n\tx = getInput();\n\tcheckISBN(x);\n\n\ndef getInput():\n\tx = raw_input(\"Type Number to check or exit to exit\\n\");\n\twhile len(x) != 10 and x != \"exit\":\n\t\tprint(\"Invalid String. Type Number to check or exit to exit\");\n\t\tx = raw_input();\n\t\n\tif(len(x) == 10):\n\t\treturn x;\n\telse:\n\t\tsys.exit();\n\n\ndef checkISBN(isbn):\n\n\tchecksum = 0;\n\tfor a in range(0,10):\n\t\tmulti = 10 - a;\n\t\tif(multi == 1 and (isbn[a] == 'X' or isbn[a] == 'x')):\n\t\t\tchecksum += 10;\n\t\telse:\n\t\t\tchecksum += int(isbn[a]) * multi;\n\n\tprint(checksum);\n\tif checksum % 11 != 0:\n\t\tprint(\"Invalid\");\n\telse:\n\t\tprint(\"Valid\");\n\ndef main():\n\tgetInputcheckISBN();\n\nmain();","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":652,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"107415565","text":"from django.conf.urls import url, include\nfrom apps.Inicio.views import index\nfrom apps.RegUser.views import Ruser_view,Verific,Validar,Iniciar,recuperar\n\nurlpatterns = [\n url(r'^nuevo$', Ruser_view, name='Usuario_crear'),\n url(r'^verif',Verific,name='Verific'),\n url(r'^validar/$',Validar,name='validar'),\n url(r'^iniciar$',Iniciar,name='iniciar'),\n url(r'^recu$',recuperar,name='recu'),\n # url(r'^',include(\"apps.inicio.urls\")),\n\n]","sub_path":"Archivos/USACMUSIC/apps/RegUser/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":451,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"362746023","text":"SOURCE_DIR = \"C:/Users/Joel/dev/python/BackupTool/testSource\"\r\nTARGET_DIR = \"C:/Users/Joel/dev/python/BackupTool/testTarget\"\r\n\r\nEXCLUDE_PATHS = [\r\n\t\"ignoredir/*\",\r\n\t\"*.png\"\r\n]\r\n\r\n# save, mirror, hardlinks\r\nMODE = \"hardlink\"\r\n\r\n# In hardlink mode this is True automatically\r\nVERSIONED = True\r\n\r\n# Uses time.strftime\r\nVERSION_NAME = \"%Y_%m_%d\"\r\n\r\n# only relevant when VERSIONED = True, will not use the directory writing to\r\n# In hardlink mode this is True automatically\r\nCOMPARE_WITH_LAST_BACKUP = True\r\n\r\nSAVE_ACTIONFILE = True\r\n\r\n# Opens the action file. Only performed if SAVE_ACTIONFILE = True.\r\nOPEN_ACTIONFILE = False\r\n\r\nAPPLY_ACTIONS = True\r\n\r\n# ordered list of possible elements \"moddate\", \"size\", \"bytes\", \"md5\" (not yet implemented)\r\nCOMPARE_METHOD = [\"moddate\", \"size\", \"bytes\"]\r\n\r\n# Caching not yet implemented\r\nCACHE_TARGET_DATA = False\r\n\r\n# Move detection not yet implemented. And I probably never will, since the gain is very small and sometimes even non-existent if the moved files are big enough\r\nMOVE_DETECTION = False\r\n\r\n# Log level, possible options: \"ERROR\", \"WARNING\", \"INFO\", \"DEBUG\"\r\nLOG_LEVEL = \"DEBUG\"\r\n\r\nSAVE_ACTIONHTML = True\r\nOPEN_ACTIONHTML = True\r\n\r\n","sub_path":"config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":1180,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"469571517","text":"#!/usr/bin/env python3\n\n\"\"\"Main.\"\"\"\n\nimport sys\nfrom cpu import *\n\n# program = [\n# # From print8.ls8\n# 0b10000010, # LDI R0,8\n# 0b00000000,\n# 0b00001000,\n# 0b01000111, # PRN R0\n# 0b00000000,\n# 0b00000001, # HLT\n# ]\nprogram = []\nfilename = sys.argv[1]\nwith open(filename) as f:\n for line in f:\n # print(line)\n # split line before and after comment symbol\n comment_split = line.split(\"#\")\n\n # # extract our number\n num = comment_split[0].strip() # trim whitespace\n\n if num == '':\n continue # ignore blank lines\n\n # convert our binary string to a number\n val = int(num, 2)\n\n # store val at address in memory\n program.append(val)\n\ncpu = CPU()\n\ncpu.load(program)\ncpu.run()\n","sub_path":"ls8/ls8.py","file_name":"ls8.py","file_ext":"py","file_size_in_byte":860,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"645173916","text":"#!/usr/bin/python3\n\"\"\"\nScript that sends a POST request to the-\npassed URL with the email as a parameter.\n\"\"\"\nimport urllib.parse\nimport urllib.request\nimport sys\n\nif __name__ == '__main__':\n url = sys.argv[1]\n value = {'email': sys.argv[2]}\n data = urllib.parse.urlencode(value)\n data = data.encode('utf-8')\n req = urllib.request.Request(url, data)\n with urllib.request.urlopen(req) as response:\n body = response.read()\n print(body.decode('utf-8'))\n","sub_path":"0x11-python-network_1/2-post_email.py","file_name":"2-post_email.py","file_ext":"py","file_size_in_byte":482,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"187744056","text":"shopping_list = []\n\nprint(\"What are we going to pick up at the store? \")\nprint(\"Enter 'FINISH' to stop adding\")\n\nwhile True:\n item = input(\"> \")\n if item == 'FINISH':\n break;\n else:\n shopping_list.append(item)\n print(\"Here is the list of items: \")\n for item in shopping_list:\n print(item)","sub_path":"ShoppingList.py","file_name":"ShoppingList.py","file_ext":"py","file_size_in_byte":336,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"189959823","text":"#! /usr/bin/env python3\nfrom airflow.models import BaseOperator\nfrom airflow.utils.decorators import apply_defaults\n\nfrom cwl_airflow.utilities.cwl import (\n execute_workflow_step,\n collect_reports\n)\nfrom cwl_airflow.utilities.report import post_status\n\n\nclass CWLStepOperator(BaseOperator):\n\n\n @apply_defaults # in case someone decided to overwrite default_args from the DAG\n def __init__(\n self,\n task_id,\n *args, **kwargs\n ):\n super().__init__(task_id=task_id, *args, **kwargs)\n\n\n def execute(self, context):\n \"\"\"\n Creates job from collected reports of all finished tasks in a DAG.\n Then executes a workflow constructed from the workflow step. Writes\n report file location to X-Com.\n \"\"\"\n\n post_status(context)\n\n _, step_report = execute_workflow_step(\n workflow=context[\"dag\"].workflow,\n task_id=self.task_id,\n job_data=collect_reports(context),\n cwl_args=context[\"dag\"].default_args[\"cwl\"]\n )\n\n return step_report\n\n\n # def on_kill(self):\n # _logger.info(\"Stop docker containers\")\n # for cidfile in glob.glob(os.path.join(self.dag.default_args[\"cidfile_dir\"], self.task_id + \"*.cid\")): # make this better, doesn't look good to read from self.dag.default_args\n # try:\n # with open(cidfile, \"r\") as inp_stream:\n # _logger.debug(f\"\"\"Read container id from {cidfile}\"\"\")\n # command = [\"docker\", \"kill\", inp_stream.read()]\n # _logger.debug(f\"\"\"Call {\" \".join(command)}\"\"\")\n # p = subprocess.Popen(command, shell=False)\n # try:\n # p.wait(timeout=10)\n # except subprocess.TimeoutExpired:\n # p.kill()\n # except Exception as ex:\n # _logger.error(f\"\"\"Failed to stop docker container with ID from {cidfile}\\n {ex}\"\"\")\n\n # # _logger.info(f\"\"\"Delete temporary output directory {self.outdir}\"\"\")\n # # try:\n # # shutil.rmtree(self.outdir)\n # # except Exception as ex:\n # # _logger.error(f\"\"\"Failed to delete temporary output directory {self.outdir}\\n {ex}\"\"\")\n","sub_path":"cwl_airflow/extensions/operators/cwlstepoperator.py","file_name":"cwlstepoperator.py","file_ext":"py","file_size_in_byte":2276,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"1"} +{"seq_id":"84267033","text":"# -*- coding: utf-8 -*-\n##############################################################################\n#\n# OpenERP, Open Source Management Solution\n# Copyright (C) 2004-2010 Tiny SPRL ().\n#\n# This program is free software: you can redistribute it and/or modify\n# it under the terms of the GNU Affero General Public License as\n# published by the Free Software Foundation, either version 3 of the\n# License, or (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU Affero General Public License for more details.\n#\n# You should have received a copy of the GNU Affero General Public License\n# along with this program. If not, see .\n#\n##############################################################################\n\nimport netsvc\nfrom osv import osv,fields\nfrom tools.translate import _\nimport time\n\nclass pos_return(osv.osv_memory):\n _name = 'pos.return'\n _description = 'Point of sale return'\n\n def default_get(self, cr, uid, fields, context=None):\n \"\"\"\n To get default values for the object.\n\n @param self: The object pointer.\n @param cr: A database cursor\n @param uid: ID of the user currently logged in\n @param fields: List of fields for which we want default values\n @param context: A standard dictionary\n\n @return: A dictionary which of fields with values.\n\n \"\"\"\n\n res = super(pos_return, self).default_get(cr, uid, fields, context=context)\n order_obj = self.pool.get('pos.order')\n if context is None:\n context={}\n active_ids = context.get('active_ids')\n for order in order_obj.browse(cr, uid, active_ids, context=context):\n for line in order.lines:\n if 'return%s'%(line.id) in fields:\n res['return%s'%(line.id)] = line.qty\n return res\n\n def view_init(self, cr, uid, fields_list, context=None):\n \"\"\"\n Creates view dynamically and adding fields at runtime.\n @param self: The object pointer.\n @param cr: A database cursor\n @param uid: ID of the user currently logged in\n @param context: A standard dictionary\n @return: New arch of view with new columns.\n \"\"\"\n res = super(pos_return, self).view_init(cr, uid, fields_list, context=context)\n order_obj=self.pool.get('pos.order')\n if context is None:\n context={}\n\n active_ids=context.get('active_ids')\n for order in order_obj.browse(cr, uid, active_ids, context=context):\n for line in order.lines:\n if 'return%s'%(line.id) not in self._columns:\n self._columns['return%s'%(line.id)] = fields.float(\"Quantity\")\n\n return res\n\n def fields_view_get(self, cr, uid, view_id=None, view_type='form', context=None, toolbar=False,submenu=False):\n\n \"\"\"\n Changes the view dynamically\n\n @param self: The object pointer.\n @param cr: A database cursor\n @param uid: ID of the user currently logged in\n @param context: A standard dictionary\n\n @return: New arch of view.\n\n \"\"\"\n _nombre = ''\n result = super(pos_return, self).fields_view_get(cr, uid, view_id, view_type, context, toolbar,submenu)\n if context is None:\n context={}\n active_model = context.get('active_model')\n if not active_model and active_model != 'pos.order':\n return result\n order_obj = self.pool.get('pos.order')\n active_id = context.get('active_id', False)\n if active_id:\n _moves_arch_lst=\"\"\"\n
\n