diff --git "a/5277.jsonl" "b/5277.jsonl" new file mode 100644--- /dev/null +++ "b/5277.jsonl" @@ -0,0 +1,680 @@ +{"seq_id":"32252796","text":"import managers\nimport json\n\nfrom models import *\nfrom unittests_resource import TestResourceApi\n\n\nargs = request.arguments\n\nappid = application.id\ncontainer = \"5073ff75-da99-44fb-a5d7-e44e5ab28598\"\n#action = args.get(\"api_name\")\nxml_param = \"\"\n#xml_data = args[\"api_data\"]\n\n# auto login\ndef login():\n\treturn managers.dispatcher.dispatch_action(appid, container, \"login\", xml_param,\n\t\t'{\"login\": \"root\", \"password\": \"root\"}'\n\t)\n\ndef apicall(action, xml_data):\n\treturn managers.dispatcher.dispatch_action(appid, container, action, \"\", xml_data)\n\ndef res_test():\n\tw = Workspace(name=\"Res test workspace\")\n\tw.save()\n\tlogs = []\n\n\ttry:\n\t\ta = Application(name=\"Res test app\", workspace_id=w.guid)\n\t\ta.save()\n\n\t\tw_objects = json.loads(apicall(\"get_objects\", \"{}\"))[1]\n\t\tw_search = [ws for ws in w_objects[\"workspaces\"] if ws[\"guid\"] == w.guid][0]\n\t\tlogs.append(\"{} workspace created \".format(w_search[\"name\"]))\n\n\t\tret = apicall(\"eapp_resources\", {})[1]\n\n\n\tfinally:\n\t\tw.delete()\n\n\treturn logs\n\ntest_case = args.get(\"formlist1\", \"res_test\")\n\nresult = \"\"\nlogin()\n\nif test_case == \"res_test\":\n\tresult = res_test()\nelif test_case == \"res_unittest\":\n\tTestResourceApi.main()\n\n\n\n#\tret = managers.dispatcher.dispatch_action(appid, container, action, xml_param, xml_data)\n#if isinstance(ret, unicode):\n#\tret = ret.encode(\"utf8\",\"ignore\")\n\nself.hypertext1.htmlcode = json.dumps(result, indent=4)\n","sub_path":"Pages/test_api/Actions-test_api/api_test.py","file_name":"api_test.py","file_ext":"py","file_size_in_byte":1385,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"466573032","text":"from summer import Multiply, Stratification\n\nfrom autumn.models.covid_19.constants import COMPARTMENTS, Clinical\nfrom autumn.models.covid_19.parameters import Parameters\nfrom autumn.models.covid_19.preprocess.vaccination import add_clinical_adjustments_to_strat\nfrom autumn.models.covid_19.preprocess.vaccination import add_vaccine_infection_and_severity\n\nCLINICAL_STRATA = [\n Clinical.NON_SYMPT,\n Clinical.SYMPT_NON_HOSPITAL,\n Clinical.SYMPT_ISOLATE,\n Clinical.HOSPITAL_NON_ICU,\n Clinical.ICU,\n]\n\nVACCINATION_STRATA = [\n \"unvaccinated\",\n \"vaccinated\",\n]\n\n\ndef get_vaccination_strat(params: Parameters) -> Stratification:\n immunity_strat = Stratification(\"vaccination\", VACCINATION_STRATA, COMPARTMENTS)\n\n # Everyone starts out unvaccinated.\n immunity_strat.set_population_split({\"unvaccinated\": 1.0, \"vaccinated\": 0.0})\n\n # Sort out the parameters to be applied.\n\n infection_efficacy, severity_efficacy = add_vaccine_infection_and_severity(\n params.vaccination.vacc_prop_prevent_infection, params.vaccination.overall_efficacy\n )\n symptomatic_adjuster, hospital_adjuster, ifr_adjuster = (1.0 - severity_efficacy,) * 3\n\n # Apply the calibration adjustment parameters.\n symptomatic_adjuster *= params.clinical_stratification.props.symptomatic.multiplier\n ifr_adjuster *= params.infection_fatality.multiplier\n\n # Add the clinical adjustments parameters as overwrites in the same way as for vaccination.\n immunity_strat = add_clinical_adjustments_to_strat(\n immunity_strat,\n VACCINATION_STRATA[0],\n VACCINATION_STRATA[1],\n params,\n symptomatic_adjuster,\n hospital_adjuster,\n ifr_adjuster,\n params.infection_fatality.top_bracket_overwrite,\n )\n\n # Apply vaccination effect against infection/transmission\n immunity_strat.add_flow_adjustments(\n \"infection\",\n {\n \"vaccinated\": Multiply(1.0 - infection_efficacy),\n \"unvaccinated\": None,\n },\n )\n\n return immunity_strat\n","sub_path":"autumn/models/covid_19/stratifications/vaccination.py","file_name":"vaccination.py","file_ext":"py","file_size_in_byte":2036,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"433792523","text":"\"\"\"\nTaboo game.\n\nPlayers take turns describing a word or phrase on a drawn card to their partner\nwithout using five common additional words.\n\"\"\"\n\nfrom collections import OrderedDict\n\nfrom kochira.auth import requires_permission\nfrom kochira.service import Service, requires_context\nfrom kochira.db import Model\n\nimport peewee\n\nservice = Service(__name__, __doc__)\n\n\n@service.setup\ndef setup_contexts(ctx):\n ctx.storage.games = {}\n\n\n@service.model\nclass Taboo(Model):\n title = peewee.CharField(255)\n taboo1 = peewee.CharField(255)\n taboo2 = peewee.CharField(255)\n taboo3 = peewee.CharField(255)\n taboo4 = peewee.CharField(255)\n taboo5 = peewee.CharField(255)\n\n @property\n def taboos(self):\n return {self.title, self.taboo1, self.taboo2, self.taboo3, self.taboo4, self.taboo5}\n\n class Meta:\n indexes = (\n ((\"title\",), True),\n )\n\n\nclass TabooStateError(Exception):\n def __init__(self, code):\n self.code = code\n\n NO_MORE_CARDS = 0\n\n\nclass Game:\n TURN_DURATION = 60\n\n def __init__(self):\n q = Taboo.select().order_by(peewee.fn.Random())\n\n if not q.exists():\n raise TabooStateError(TabooStateError.NO_MORE_CARDS)\n\n self.cards = iter(q)\n self.started = False\n\n self.players = []\n self.teams = [0, 0]\n\n self._turn_index = 0\n\n def draw(self):\n try:\n self.card = next(self.cards)\n except StopIteration:\n raise TabooStateError(TabooStateError.NO_MORE_CARDS)\n\n @property\n def team(self):\n return self._turn_index % 2\n\n @property\n def turn(self):\n return self.players[self._turn_index]\n\n @property\n def guessers(self):\n return [p for p in self.players[self.team::2]\n if p != self.turn]\n\n def _next_turn_index(self):\n return (self._turn_index + 1) % len(self.players)\n\n def advance(self):\n self._turn_index = self._next_turn_index()\n\n def submit_clue(self, sentence):\n for taboo in self.card.taboos:\n if taboo in sentence:\n return taboo\n return None\n\n def submit_guess(self, sentence):\n if self.card.title in sentence:\n self.teams[self.team] += 1\n return True\n return False\n\n def join(self, player):\n if player in self.players:\n raise ValueError(\"player is already playing\")\n\n self.players.append(player)\n\n def leave(self, player):\n self.players.remove(player)\n if self.players:\n self._turn_index %= len(self.players)\n return not self.players or (self.started and len(self.players) < 4)\n\n def start(self):\n if len(self.players) < 4:\n raise ValueError(\"not enough players\")\n self.started = True\n\n def stop(self):\n self.started = False\n\n\n@service.command(r\"!taboo add (?P[^:]+): (?P<taboo1>[^,]+), (?P<taboo2>[^,]+), (?P<taboo3>[^,]+), (?P<taboo4>[^,]+), (?P<taboo5>[^,]+)\")\n@service.command(r\"add taboo (?P<title>[^:]+): (?P<taboo1>[^,]+), (?P<taboo2>[^,]+), (?P<taboo3>[^,]+), (?P<taboo4>[^,]+), (?P<taboo5>[^,]+)\", mention=True)\n@requires_permission(\"taboo\")\ndef add_taboo(ctx, title, taboo1, taboo2, taboo3, taboo4, taboo5):\n \"\"\"\n Add a Taboo card.\n\n (It has five parameters because the regex can strictly validate the\n command. YES, I KNOW WHAT ``str.split`` IS.)\n \"\"\"\n if Taboo.select().where(Taboo.title == title).exists():\n ctx.respond(ctx._(\"That Taboo card already exists.\"))\n return\n\n taboo = Taboo.create(title=title.strip().lower(),\n taboo1=taboo1.strip().lower(),\n taboo2=taboo2.strip().lower(),\n taboo3=taboo3.strip().lower(),\n taboo4=taboo4.strip().lower(),\n taboo5=taboo5.strip().lower())\n\n taboo.save()\n ctx.respond(ctx._(\"Added Taboo card \\\"{title}\\\", with taboos: {taboos}.\").format(\n title=taboo.title,\n taboos=\", \".join(taboo.taboos)\n ))\n\n\n@service.command(r\"!taboo del (?P<title>.+)\")\n@service.command(r\"(?:remove|delete) taboo card (?P<title>.+)\", mention=True)\n@requires_permission(\"taboo\")\ndef remove_taboo(ctx, title):\n \"\"\"\n Remove a Taboo card.\n\n Delete a Taboo card from the database.\n \"\"\"\n title = title.lower()\n\n if Taboo.delete().where(Taboo.title == title).execute() == 0:\n ctx.respond(ctx._(\"Can't find that Taboo card.\"))\n else:\n ctx.respond(ctx._(\"Deleted Taboo card \\\"{title}\\\".\").format(\n title=title\n ))\n\n\n@service.command(r\"!taboo\")\n@service.command(r\"taboo\", mention=True)\ndef request_taboo(ctx):\n \"\"\"\n Request a game of Taboo.\n\n Initiate a game of Taboo.\n \"\"\"\n k = (ctx.client.name, ctx.target)\n\n if k in ctx.storage.games:\n ctx.respond(ctx._(\"A game is already in progress.\"))\n return\n\n try:\n g = Game()\n except TabooStateError as e:\n if e.code != TabooStateError.NO_MORE_CARDS:\n raise\n\n ctx.respond(ctx._(\"There are no Taboo cards.\"))\n return\n\n g.period = None\n g.join(ctx.origin)\n ctx.storage.games[k] = g\n\n ctx.message(ctx._(\"{origin} has started a game of Taboo! Send !join to join, and !start when ready!\").format(\n origin=ctx.origin\n ))\n\n ctx.add_context(\"taboo\")\n\n\n@service.command(r\"!join\")\n@requires_context(\"taboo\")\ndef join_taboo(ctx):\n \"\"\"\n Join game.\n\n Join a Taboo game in progress.\n \"\"\"\n game = ctx.storage.games[ctx.client.name, ctx.target]\n\n if ctx.origin in game.players:\n ctx.respond(ctx._(\"You're already in the game.\"))\n return\n\n game.join(ctx.origin)\n\n ctx.message(ctx._(\"{origin} has joined the game!\").format(origin=ctx.origin))\n\n\n@service.command(r\"!leave\")\n@requires_context(\"taboo\")\ndef leave(ctx):\n \"\"\"\n Leave game.\n\n Leave the game, if you're participating.\n \"\"\"\n game = ctx.storage.games[ctx.client.name, ctx.target]\n\n if ctx.origin not in game.players:\n ctx.respond(ctx._(\"You're not in this game.\"))\n return\n\n game_over = game.leave(ctx.origin)\n\n if game_over:\n do_game_over(ctx)\n return\n\n ctx.message(ctx._(\"{origin} left the game.\").format(origin=ctx.origin))\n\n if game.started:\n send_summary(ctx)\n\n\ndef show_scores(game):\n team1 = \", \".join(list(game.players)[::2])\n team2 = \", \".join(list(game.players)[1::2])\n\n scores = [(team1, game.teams[0]), (team2, game.teams[1])]\n scores.sort(key=lambda x: -x[1])\n\n return \"; \".join(\"{}: {}\".format(k, v) for k, v in scores)\n\n\ndef send_summary(ctx):\n game = ctx.storage.games[ctx.client.name, ctx.target]\n\n ctx.message(ctx.ngettext(\"{turn}: It's your turn -- explain your word but don't say any of the taboos! {guessers} is guessing. You have {time} seconds.\",\n \"{turn}: It's your turn -- explain your word but don't say any of the taboos! {guessers} are guessing. You have {time} seconds.\",\n len(game.guessers)).format(\n turn=game.turn,\n guessers=\", \".join(game.guessers),\n time=Game.TURN_DURATION,\n isare=\"is\" if len(game.guessers) == 1 else \"are\"\n ))\n\n\ndef do_game_over(ctx, prefix=\"\"):\n game = ctx.storage.games[ctx.client.name, ctx.target]\n game.stop()\n\n if game.period is not None:\n ctx.bot.scheduler.unschedule_period(game.period)\n\n ctx.message(prefix + ctx._(\"Game over! Final results: {results}\").format(\n results=show_scores(game)\n ))\n del ctx.storage.games[ctx.client.name, ctx.target]\n ctx.remove_context(\"taboo\")\n\n\n@service.command(r\"!stop\")\n@requires_context(\"taboo\")\ndef stop_taboo(ctx):\n \"\"\"\n Stop Taboo.\n\n Stop the Taboo game in progress.\n \"\"\"\n do_game_over(ctx)\n\n\n@service.command(r\"!start\")\n@requires_context(\"taboo\")\ndef start_taboo(ctx):\n \"\"\"\n Start Taboo.\n\n Start the Taboo game.\n \"\"\"\n game = ctx.storage.games[ctx.client.name, ctx.target]\n\n if ctx.origin not in game.players:\n ctx.respond(ctx._(\"You're not in this game.\"))\n return\n\n if game.started:\n ctx.respond(ctx._(\"This game is already in progress.\"))\n return\n\n if len(game.players) < 4:\n ctx.respond(ctx._(\"There aren't enough players to play yet.\"))\n return\n\n game.start()\n\n send_summary(ctx)\n do_draw(ctx)\n\n game.period = client.bot.scheduler.schedule_every(Game.TURN_DURATION, do_advance, ctx, game)\n\n\n@service.task\ndef do_advance(ctx, game):\n ctx.message(ctx._(\"{turn}: Time is up! The word was \\\"{word}\\\".\").format(\n turn=game.turn,\n word=game.card.title\n ))\n\n game.advance()\n if do_draw(ctx):\n return\n\n send_summary(ctx)\n\n\n@service.command(r\"!pass\")\n@requires_context(\"taboo\")\ndef pass_taboo(ctx):\n \"\"\"\n Pass.\n\n Pass on this card.\n \"\"\"\n game = ctx.storage.games[client.name, target]\n\n if origin not in game.players:\n ctx.respond(ctx._(\"You're not in this game.\"))\n return\n\n if origin != game.turn:\n ctx.respond(ctx._(\"It's not your turn.\"))\n return\n\n do_draw(client, target)\n\n\ndef do_draw(ctx):\n game = ctx.storage.games[ctx.client.name, ctx.target]\n\n try:\n game.draw()\n except TabooStateError as e:\n if e.code != TabooStateError.NO_MORE_CARDS:\n raise\n do_game_over(ctx, ctx._(\"Looks like we ran out of cards! \"))\n return True\n\n ctx.client.notice(game.turn, ctx._(\"Title: {title}; Taboos: {taboos}\").format(\n title=game.card.title,\n taboos=\", \".join(game.card.taboos)\n ))\n return False\n\n\n@service.hook(\"channel_message\")\ndef do_guess(ctx, target, origin, message):\n if not service.has_context(ctx.client, \"taboo\", ctx.target):\n # nobody is playing taboo.\n return\n\n game = ctx.storage.games[ctx.client.name, ctx.target]\n\n if not game.started:\n # taboo hasn't started yet.\n return\n\n card = game.card\n\n if origin == game.turn:\n maybe_taboo = game.submit_clue(message)\n if maybe_taboo is not None:\n ctx.respond(ctx._(\"BZZT! You said \\\"{taboo}\\\". The word was \\\"{title}\\\". Next word!\").format(\n taboo=maybe_taboo,\n title=card.title\n ))\n do_draw(ctx)\n elif origin in game.guessers:\n if game.submit_guess(message):\n ctx.respond(ctx._(\"Ding-ding! The word was \\\"{title}\\\". Point for team {n}. Next word!\").format(\n title=card.title,\n n=game.team + 1\n ))\n do_draw(ctx)\n","sub_path":"kochira/services/games/taboo.py","file_name":"taboo.py","file_ext":"py","file_size_in_byte":10617,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"131430519","text":"#!/home/kjeong23/softwares/bin/python3.4\n# program to calculate total MSD of any atom/molecules\n# -> display MSD(t), count(t).\n# ** ascii trajectory file: requires pbc -whole treatment!!\n#inputs: grofile, traj ASCII file \n#output: file of {t, tot_MSD(t), count(t)}\n\nimport math\nimport sys\nimport numpy\nimport timeit\n\n#atomic mass dictionary\namass={'C1':15.035, 'C2':14.027, 'C3': 14.027, 'C4': 14.027, 'C5': 14.027, \\\n'C6': 14.027, 'C7': 14.027, 'C8': 14.027, 'C9': 14.027, 'CA': 14.027, \\\n'P1': 30.9738, 'OD1': 15.9994, 'OD2': 15.9994, 'OD3': 15.9994, \\\n'N+': 14.0067, 'Me1': 15.035, 'Me2': 15.035, 'Me3': 15.035, 'Me4': 15.035, \\\n'OW': 15.9994, 'HW1': 1.008, 'HW2': 1.008}\n\ndef gro_minorsplit(str): #own needed function for splitting of .gro format. V05.\n splitstr=[str[0:10],str[10:15]] #molecule index&name, atomname\n for i in range(len(splitstr)):\n splitstr[i]=splitstr[i].replace(\" \",\"\")\n return splitstr\n\ndef ascii_trajsplit(str): #reading coord from traj ASCII. box size cannot be read by this.\n splitstr=[str[16:28],str[29:42],str[43:56]]\n for i in range(len(splitstr)):\n splitstr[i]=splitstr[i].replace(\" \",\"\")\n posdata=numpy.array([float(x) for x in splitstr])\n return posdata\n\ndef comcalc(alist,crd): #calculation of COM of a molecule\n compos=numpy.zeros(3)\n totmass,i=0.0,0\n for x in alist: #alist : list of atomnames. #crd: n*3 array of coords\n m1=amass[x]\n totmass+=m1\n compos+=m1*crd[i]\n i+=1\n compos/=totmass\n return compos\n\ndef pbcdr(ri,rf,box): #displacement vector with considering pbc\n dr=rf-ri\n for i in range(3):\n if dr[i]>(box[i]/2.0):\n dr[i]-=box[i]\n elif dr[i]<(-box[i]/2.0):\n dr[i]+=box[i]\n return dr\n\n#main fxn\ndef main():\n #Part to load coordinate file, morphology info file\n grofile = open(sys.argv[1],'r')\n trjfile = open(sys.argv[2],'r')\n outfile = open(sys.argv[3],'w')\n\n initt=float(input(\"Initial time of this trajectory (in ns)?\\n\"))\n fint=float(input(\"Final time of the whole trajectory in ns? ex) 499.96 \\n\"))\n ssize=float(input(\"Timestep between snapshots (in ns)? ex) 0.04\\n\"))\n surfstr=input(\"Name of the molecule you are tracking? ex) SOL \\n\")\n delim=input(\"delimiter: name of the final atom of 1 molecule? ex) OW \\n\")\n nsurf,lindex,entry=0,0,[]\n start_time=timeit.default_timer()\n\n #reading grofile to match molecule,atom index\n for line in grofile:\n if lindex !=0:\n if lindex==1:\n totnatom=int(line)\n elif lindex>=2 and lindex<=1+totnatom:\n split=gro_minorsplit(line)\n entry.append(split)\n if split[1]==delim:\n nsurf+=1\n lindex+=1\n\n #prepare bins\n nstep=int((fint-initt)/ssize)\n sdbin,count=numpy.zeros(nstep),numpy.zeros(nstep) #bin & stat weight for dr^2(t) ave\n rcom,box=numpy.zeros((nstep+1,nsurf,3)),numpy.zeros((nstep+1,3))\n\n #loop of reading coord for rcom storing\n lindex,sindex=0,0\n totcrd=numpy.empty((0,3),float)\n molst=int(entry[0][0].replace(surfstr,\"\"))-1 #beginning index of species\n for line in trjfile:\n if lindex==3: #box vector x\n splitstr=[line[18:30]]\n for i in range(len(splitstr)):\n splitstr[i]=splitstr[i].replace(\" \",\"\")\n box[sindex][0]=float(splitstr[0])\n elif lindex==4: #box vector y\n splitstr=[line[31:44]]\n for i in range(len(splitstr)):\n splitstr[i]=splitstr[i].replace(\" \",\"\")\n box[sindex][1]=float(splitstr[0])\n elif lindex==5: #box vector z\n splitstr=[line[45:58]]\n for i in range(len(splitstr)):\n splitstr[i]=splitstr[i].replace(\" \",\"\")\n box[sindex][2]=float(splitstr[0])\n elif lindex>=7 and lindex<=6+totnatom:\n split=ascii_trajsplit(line)\n totcrd=numpy.vstack((totcrd,split))\n if lindex==6+totnatom: #conclusion for 1 step, initialize\n tempcrd=numpy.empty((0,3),float) #coord set for 1molecule\n alist=[]\n aindex=0\n #calc of COMs\n for row in entry: #search molecule index & atom names\n molnum=int(row[0].replace(surfstr,\"\"))-molst\n alist.append(row[1])\n tempcrd=numpy.vstack((tempcrd,totcrd[aindex]))\n if row[1]==delim: #final atom of a molecule\n rcom[sindex][molnum-1]=comcalc(alist,tempcrd)\n #initialize single-molecule variables\n alist=[]\n tempcrd=numpy.empty((0,3),float)\n aindex+=1\n totcrd=numpy.empty((0,3),float)\n sindex+=1\n lindex=-1\n# elapsed=timeit.default_timer() - start_time\n# print('reading,rcom calc step {} time {:8.4f}'.format(sindex,elapsed))\n lindex+=1\n\n #after all rcom is clarified, calculate MSD with considering PBC\n for molnum in range(nsurf):\n for i in range(nstep):\n vec=numpy.zeros(3)\n for j in range(i+1,nstep+1):\n dr=pbcdr(rcom[j][molnum],rcom[j-1][molnum],box[j]) #1step displacement vector\n vec+=dr\n rmag2=numpy.dot(vec,vec)\n sdbin[j-i-1]+=rmag2\n count[j-i-1]+=1\n elapsed=timeit.default_timer() - start_time\n print('collecting data for molecule# {} time {:8.4f}'.format(molnum+1,elapsed))\n \n #printing section\n sdbin/=count\n outfile.write('{:8.4f} {:8.4f} {:8.4f}\\n'.format(0,0,0))\n for i in range(1,nstep+1):\n outfile.write('{:8.4f} {:8.4f} {:8.4f}\\n'.format(ssize*i,sdbin[i-1],count[i-1]))\n\n grofile.close()\n trjfile.close()\n outfile.close()\n\nif __name__ == \"__main__\": main()\n\n","sub_path":"py_development/data_process/micelles/totmsd_v02.py","file_name":"totmsd_v02.py","file_ext":"py","file_size_in_byte":5303,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"167630473","text":"import random\n\n# игра поле чудес, выводятся скрывающие буквы символы (кол-во символов в заданном слове)\n# рандомно выбранного слова из списка\n# предлагается ввести символ или слово целиком\n\nbase_words = [\"книга\", \"месяц\", \"ручка\", \"шарик\", \"олень\", \"носок\", \"кукарача\"]\nunknown_symb = \"\\u25a0\"\nheart_symb = \"\\u2764\"\nlife = 5\ngame_repeat = True\nwhile game_repeat:\n hidden_word = str(base_words[random.randrange(len(base_words))]) # рандомно выбранное слово из списка\n output_word = list(len(hidden_word) * unknown_symb) # скрытое выводимое слово\n while life > 0:\n if list(output_word) == list(hidden_word): # проверка отгаданного слова и вывод результата\n print(\" \".join(output_word))\n print(\"Поздравляю!\\nВы выиграли воздух!\")\n life = -1\n break\n\n print(\" \".join(output_word), \"|x\", heart_symb, life, sep=\" \")\n inputWord_Letter = input(\"Назовите букву или слово целиком: \")\n\n if inputWord_Letter in hidden_word: # если введённые символы в нашем слове то\n if list(inputWord_Letter) == list(hidden_word): # если введённые символы равны нашему слову\n for i in range(len(inputWord_Letter)): # проход по каждому введённому символу\n output_word[i] = hidden_word[i] # изменение скрытого слова\n elif len(inputWord_Letter) == 1: # в случае ввода 1ого символа\n for i in range(len(hidden_word)): # проход по заданному слову\n if hidden_word[i] == inputWord_Letter: # проверка каждого символа\n output_word[i] = hidden_word[i] # замена совпадения в скрытом слове\n else:\n print(\"Неправильно. Вы теряете жизнь\")\n life -= 1\n\n if life == 0:\n print(\"Вы проиграли, увидимся никогда!\")\n life = -1\n elif life == -1: # предлагаем сыграть ещё раз\n more_game = input(\"Сыграть ещё раз? Y/N: \")\n if more_game == \"N\" or more_game == \"n\":\n game_repeat = False\n elif more_game == \"Y\" or more_game == \"y\":\n life = 5\n","sub_path":"wheel_of_fortune_5.py","file_name":"wheel_of_fortune_5.py","file_ext":"py","file_size_in_byte":2920,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"97609963","text":"class TreeNode:\n def __init__(self, x):\n self.val = x\n self.left, self.right = None, None\n\nclass Solution:\n\n def maxDepth(self, root):\n if not root:\n return 0\n depth = 0\n stack = [(root, 1)]\n while stack:\n node, d = stack.pop()\n if depth < d:\n depth = d\n if node.right:\n stack.append((node.right, d+1))\n if node.left:\n stack.append((node.left, d+1))\n return depth\n\n\n\nif __name__ == \"__main__\":\n n0 = TreeNode(3)\n n1 = TreeNode(9)\n n2 = TreeNode(20)\n n0.left = n1\n n0.right = n2\n n3 = TreeNode(15)\n n4 = TreeNode(7)\n n2.left = n3\n n2.right = n4\n s = Solution()\n r = s.maxDepth(n0)\n print(r)\n","sub_path":"leetcode/tree/max-depth.py","file_name":"max-depth.py","file_ext":"py","file_size_in_byte":781,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"152400263","text":"def hangman(word):\n rletters = list(word)\n stage = [\n '__________',\n '| | ',\n '| o ',\n '| /|\\ ',\n '| / \\ ',\n '| ',\n ]\n\n board = ['_'] * len(word)\n\n wrong = 0\n win = False\n challenge = 1\n\n print('ハングマンにようこそ')\n \n while wrong < len(stage):\n print('\\n')\n msg = '一文字を入力してください'\n user_input = input(msg)\n\n if user_input in rletters:\n cind = rletters.index(user_input)\n board[cind] = rletters[cind]\n rletters[cind] = '$'\n print('あたり')\n e = len(rletters)\n for i in board:\n print(i, end=\"\")\n \n if '_' not in board:\n win = True\n break\n else:\n print('はずれ')\n for i in range(wrong + 1):\n print(stage[i])\n wrong += 1\n \n if win:\n print('あなたの勝ち')\n else:\n print('あなたの負け')\n\n\nhangman('test')\n\n\n\n","sub_path":"hangman.py","file_name":"hangman.py","file_ext":"py","file_size_in_byte":1098,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"542018583","text":"import sys\nfrom PyQt5.QtWidgets import QMainWindow, QApplication, QWidget, QAction, QTableWidget, QTableWidgetItem, QVBoxLayout, QDialog\nfrom PyQt5.QtGui import QIcon\nfrom PyQt5.QtCore import pyqtSlot\nfrom matplotlib.backends.qt_compat import QtCore, QtWidgets, is_pyqt5\nif is_pyqt5():\n from matplotlib.backends.backend_qt5agg import (\n FigureCanvas, NavigationToolbar2QT as NavigationToolbar)\nelse:\n from matplotlib.backends.backend_qt4agg import (\n FigureCanvas, NavigationToolbar2QT as NavigationToolbar)\nfrom matplotlib.figure import Figure\n\n# import matplotlib.pyplot as plt\n\n\nclass App(QDialog):\n\n def __init__(self):\n super().__init__()\n self.title = '1D look up table input dialog'\n self.left = 0\n self.top = 0\n self.width = 1200\n self.height = 500\n self.initUI()\n\n def initUI(self):\n self.setWindowTitle(self.title)\n self.setGeometry(self.left, self.top, self.width, self.height)\n\n self.create_table()\n self.create_plot()\n\n # Add box layout, add table to box layout and add box layout to widget\n self.layout = QVBoxLayout()\n self.layout.addWidget(self.tableWidget)\n self.layout.addWidget(self.canvas)\n self.setLayout(self.layout)\n\n # Show widget\n self.show()\n\n self.update_plot()\n\n def create_table(self):\n # Create table\n self.tableWidget = QTableWidget()\n self.tableWidget.setRowCount(2)\n self.tableWidget.setColumnCount(10)\n self.tableWidget.setItem(0, 0, QTableWidgetItem(\"Cell (1,1)\"))\n self.tableWidget.setItem(0, 1, QTableWidgetItem(\"Cell (1,2)\"))\n self.tableWidget.setItem(1, 0, QTableWidgetItem(\"Cell (2,1)\"))\n self.tableWidget.setItem(1, 1, QTableWidgetItem(\"Cell (2,2)\"))\n self.tableWidget.move(0, 0)\n\n # table selection change\n self.tableWidget.doubleClicked.connect(self.on_click)\n\n def create_plot(self):\n self.canvas = FigureCanvas(Figure(figsize = (5, 5)))\n self.axis = self.canvas.figure.subplots()\n pass\n\n @pyqtSlot()\n def on_click(self):\n print(\"\\n\")\n for currentQTableWidgetItem in self.tableWidget.selectedItems():\n print(currentQTableWidgetItem.row(), currentQTableWidgetItem.column(), currentQTableWidgetItem.text())\n\n def update_plot(self):\n self.axis.plot([0, 10], [10, 0], 'r')\n self.axis.figure.canvas.draw()\n\n\nif __name__ == '__main__':\n app = QApplication(sys.argv)\n ex = App()\n sys.exit(app.exec_())\n","sub_path":"rosi_benchmarking/soil/python/test_gui.py","file_name":"test_gui.py","file_ext":"py","file_size_in_byte":2558,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"611556064","text":"import glob\nimport re\n\nfiles = glob.glob('H:/FS3100/*.xls')\n\nnitrateData = []\nnitriteData = []\ncalcsData = []\n \nfor file in files:\n fname = file.split('\\\\')[1].split('.')[0]\n if len(fname) == 7: \n if fname[4:6] == 'N2':\n with open(file, 'r') as f:\n for line in f:\n line = line.split('\\t')\n if ( line[0].isdigit() ) and ( re.match('^S', line[2]) or re.match('^W', line[2]) or re.match('^AW', line[2]) ):\n if line[-1] == '\\n':\n nitriteTuple = (file,line[2].lstrip().rstrip(),line[-2].lstrip().rstrip())\n nitriteData.append(nitriteTuple)\n else:\n nitriteTuple = (file,line[2].lstrip().rstrip(),line[-1].lstrip().rstrip())\n nitriteData.append(nitriteTuple) \n elif fname[4:6] == 'N3':\n with open(file, 'r') as f:\n for line in f:\n line = line.split('\\t')\n if ( line[0].isdigit() ) and ( re.match('^S', line[2]) or re.match('^W', line[2]) or re.match('^AW', line[2]) ):\n if line[-1] == '\\n':\n nitrateTuple = (file,line[2].lstrip().rstrip(),line[-2].lstrip().rstrip())\n nitrateData.append(nitrateTuple)\n else:\n nitrateTuple = (file,line[2].lstrip().rstrip(),line[-1].lstrip().rstrip())\n nitrateData.append(nitrateTuple)\n\nwith open('calcs.txt', 'r') as calcsFile:\n for line in calcsFile:\n line = line.rstrip().split(' ')\n calcsTuple = (line[0],line[5])\n calcsData.append(calcsTuple)\n\nfor x in calcsData:\n for y in nitrateData:\n if x[0] in y[1]:\n for z in nitriteData:\n if x[0] in z[1]:\n with open('NitrateCalcs.csv', 'a') as output:\n output.write(x[0]+','+y[2]+','+z[2]+','+str(float(y[2])-float(z[2]))+'\\n')\n \n \n\n \n \n","sub_path":"NitrateCalcsNew.py","file_name":"NitrateCalcsNew.py","file_ext":"py","file_size_in_byte":2109,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"259849211","text":"'''\n@author: chenxi\n@file: FFM.py\n@time: 2019/9/10 19:07\n@desc: Field-aware Factorization Machine\n'''\n\nimport tensorflow as tf\nfrom Utils.Data4PyRec import Data4PyRec\nfrom tensorflow.losses import logloss\n\nclass FFM():\n def __init__(self, data, label, feature_field, embedding_size=8, lr_reg_l1=0, lr_reg_l2=0, fm_reg_l1=0, fm_reg_l2=0, loss=\"logloss\", metric=\"logloss\", opt=\"adam\", learning_rate=0.1, epochs=10, batch_size=256, verbos=1, random_seed=2018):\n # 数据参数\n self.data = data # 训练特征集\n self.label = label # 训练标签集\n self.feature_num = data.shape[1] # 特征的个数\n self.feature_field = feature_field # feature_field是一个每个feature所属的field列表\n self.field_num = len(set(feature_field)) # 特征所属的field数量\n\n # 算法特性参数\n self.embedding_size = embedding_size\n\n # 算法训练参数\n self.lr_reg_l1 = lr_reg_l1 # LR部分L1正则化系数\n self.lr_reg_l2 = lr_reg_l2 # LR部分L2正则化系数\n self.fm_reg_l1 = fm_reg_l1 # FM部分L1正则化系数\n self.fm_reg_l2 = fm_reg_l2 # FM部分L2正则化系数\n self.learning_rate = learning_rate # 学习率大小\n self.epochs = epochs # 训练迭代次数\n self.batch_size = batch_size # 一个batch数据大小\n self.verbos = verbos # 打印输出间隔,默认是1个batch一次打印,小于1为不打印输出\n\n # 其他参数\n self.random_seed = random_seed\n\n # 初始化计算图\n self.init_graph()\n\n def __del__(self):\n print(\"FFM task over\")\n self.sess.close() # 对象销毁时,停止会话,防止内存泄露\n\n def init_graph(self):\n # 构建tf计算图\n self.graph = tf.Graph()\n with self.graph.as_default():\n tf.set_random_seed(self.random_seed) # 设置随机种子大小\n # 设置输入输出\n self.X = tf.placeholder('float', [None, self.feature_num], name='X') # 输入矩阵维度: m*n,m为数据量大小,n是特征个数\n self.Y = tf.placeholder('float', [None, 1], name='Y') # 目标值维度: m*1\n\n # part1: LR部分\n self.w_0 = tf.Variable(tf.truncated_normal([1]), name=\"w_0\") # w_0\n self.w = tf.Variable(tf.truncated_normal([self.feature_num]), name=\"w\") # w\n self.lr_output = tf.add(self.w_0, tf.reduce_sum(tf.multiply(self.w, self.X), axis=1, keep_dims=True), name=\"LR_part\") # W0+WX\n\n # part2: FFM特征和filed及其他特征交叉项部分\n self.V = tf.Variable(tf.truncated_normal([self.feature_num, self.field_num, self.embedding_size])) # 大小为kfn的嵌入矩阵\n self.ffm_output = tf.Variable(0.0, tf.float32)\n # 外层遍历所有的特征\n for feature_index1 in range(self.feature_num):\n field_index_1 = int(self.feature_field[feature_index1])\n # 内层遍历feature_index1之后的特征\n for feature_index2 in range(feature_index1+1, self.feature_num):\n field_index_2 = int(self.feature_field[feature_index2])\n # 左向量,对应V_{i,fj}\n # VectorSize对应每个隐向量的长度\n vectorLeft = tf.convert_to_tensor([[feature_index1, feature_index1, i] for i in range(self.embedding_size)])\n # 在多维上进行索引去除对应的值\n weightLeft = tf.gather_nd(self.V, vectorLeft)\n weightLeftAfterCut = tf.squeeze(weightLeft)\n\n # 右向量,对应V_{j,fi}\n vectorRight = tf.convert_to_tensor([[feature_index2, feature_index1, i] for i in range(self.embedding_size)])\n weightRight = tf.gather_nd(self.V, vectorRight)\n weightRightAfterCut = tf.squeeze(weightRight)\n\n tempValue = tf.reduce_sum(tf.multiply(weightLeftAfterCut, weightRightAfterCut))\n\n indices2 = [feature_index1]\n indices3 = [feature_index2]\n\n x_i = tf.squeeze(tf.gather_nd(self.data, indices2))\n x_j = tf.squeeze(tf.gather_nd(self.data, indices3))\n\n product = tf.reduce_sum(tf.multiply(x_i, x_j))\n\n secondItemVal = tf.multiply(tempValue, product)\n\n tf.assign(self.ffm_output, tf.add(self.V, secondItemVal))\n\n self.output = tf.add(self.lr_output, self.ffm_output)\n\n # 定义目标损失\n self.obj_loss = logloss(self.Y, tf.nn.sigmoid(self.output))\n\n # 定义正则化损失\n self.lr_l1 = tf.constant(self.lr_reg_l1, name=\"lr_l1\")\n self.lr_l2 = tf.constant(self.lr_reg_l2, name=\"lr_l2\")\n self.fm_l1 = tf.constant(self.fm_reg_l1, name=\"fm_l1\")\n self.fm_l2 = tf.constant(self.fm_reg_l2, name=\"fm_l2\")\n\n self.l1_norm = tf.reduce_sum(tf.add(tf.multiply(self.lr_l1, tf.abs(self.w)), tf.multiply(self.fm_l1, tf.abs(self.V)))) # L1正则化损失函数\n self.l2_norm = tf.reduce_sum(tf.add(tf.multiply(self.lr_l2, tf.pow(self.w, 2)), tf.multiply(self.fm_l2, tf.pow(self.V, 2)))) # L2正则化损失函数\n self.norm_loss = tf.add(self.l1_norm, self.l2_norm) # 合并正则化损失\n\n # 整体损失函数\n self.loss_fun = tf.add(self.obj_loss, self.norm_loss)\n\n # 选择优化器\n self.optimizer = tf.train.AdamOptimizer(learning_rate=self.learning_rate).minimize(self.loss_fun)\n\n # 初始化\n self.saver = tf.train.Saver() # 模型保存器\n init = tf.global_variables_initializer() # 初始化变量\n self.sess = tf.Session() # 初始化tf会话\n self.sess.run(init) # 执行初始化变量\n\n # 在一个batch上训练数据\n def train_in_one_batch(self, batch_X, batch_Y):\n feed_dict = {\n self.X: batch_X,\n self.Y: batch_Y,\n }\n loss, _ = self.sess.run([self.loss, self.optimizer], feed_dict=feed_dict)\n return loss\n\n # 训练模型\n def train(self):\n # 加载数据到Data4PyRec类型\n pyRecData = Data4PyRec(self.data, self.label, batch_size=self.batch_size, is_shuffle=True,\n random_seed=self.random_seed)\n # 迭代epochs次\n for epoch in range(self.epochs):\n # 数据类型会计算是否还有剩余batch\n while pyRecData.has_next():\n # 取出下一个batch数据\n (batch_data, batch_label) = pyRecData.next()\n # 计算这个batch数据的loss\n cur_loss = self.train_in_one_batch(batch_data, batch_label)\n # 打印\n if self.verbos > 0:\n if (pyRecData.get_idx() % self.verbos == 0):\n print(\"current \" + str(self.loss) + \" is : \" + str(cur_loss) + \"!\")\n pyRecData.reset()\n\n # 使用指定的评价函数进行评估,评价函数可以和之前的目标函数不相同\n def evaluate(self, eva_X, eva_Y):\n pred = self.predict(eva_X, eva_Y)\n metric_fun = metric_select.select(self.metric)\n\n return metric_fun(eva_Y, pred)\n\n # 对指定的数据集进行预测\n def predict(self, pre_X, pre_Y):\n # 下过程同训练\n pyRecData_for_pre = Data4PyRec(pre_X, pre_Y, batch_size=self.batch_size)\n while pyRecData_for_pre.has_next():\n (batch_data_for_pre, batch_label_for_pre) = pyRecData_for_pre.next()\n feed_dict = {\n self.X: batch_data_for_pre,\n self.Y: batch_label_for_pre,\n }\n predict_part = self.sess.run(self.output, feed_dict=feed_dict)\n # 对每个batch的预测结果进行合并\n predict = None\n if pyRecData_for_pre.get_idx() == 1:\n predict = predict_part\n else:\n predict = tf.concat(0, [predict, predict_part])\n return predict\n\n # 保存模型到指定的path路径\n def save_model(self, path):\n self.saver.save(self.sess, path)\n\n\n\n\n","sub_path":"models/FFM.py","file_name":"FFM.py","file_ext":"py","file_size_in_byte":8241,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"550419500","text":"# from django.core.management.base import BaseCommand, CommandError\r\n\r\n# from core.models import State, StatePoll\r\n# from django.utils import timezone\r\n\r\nimport csv\r\n\r\n# def date_convert(str_date):\r\n# \"\"\"\r\n# Converts a string in the form YYYY-MM-DD HH:MM:SS to a datetime object\r\n# \"\"\"\r\n# str_date = str_date.split(\" \")[0]\r\n# y_m_d = str_date.split('-')\r\n# y = int(y_m_d[0])\r\n# m = int(y_m_d[1])\r\n# d = int(y_m_d[2])\r\n# date = timezone.datetime(y, m, d)\r\n\r\n# return date\r\n\r\n# class Command(BaseCommand):\r\n# help = 'adds the polls from csv to model'\r\n\r\n \r\n\r\n# def add_arguments(self, parser):\r\n# # Correct for PATH TO CSV FILE\r\n# parser.add_argument('file_name', type=str)\r\n\r\n# def handle(self, *args, **options):\r\n# file_name = options['file_name']\r\n\r\n# with open(file_name, 'r') as file:\r\n\r\n# rows = csv.reader(file, delimiter=',')\r\n \r\n# # remove header\r\n\r\n# next(rows)\r\n\r\n# # 0 1 2 3 4 5 6 7 8 9 \r\n# # Structure: State, Polling Company, URL, Start Date, End Date, Sample Size, Sample Type, MOE, Biden, Trump \r\n# for row in rows:\r\n# state_name = row[0]\r\n \r\n# pollster = row[1]\r\n# url = row[2]\r\n \r\n# # print(row)\r\n# start_date_str = row[3]\r\n# start_date = date_convert(start_date_str)\r\n \r\n\r\n# end_date_str = row[4]\r\n# end_date = date_convert(end_date_str)\r\n\r\n# if row[5] != '':\r\n# n = int(row[5])\r\n# else:\r\n# n = None\r\n# pollType = row[6]\r\n# if row[7] != '':\r\n# moe = float(row[7])\r\n# else:\r\n# moe = None\r\n# percent_biden = float(row[8])\r\n# percent_trump = float(row[9])\r\n\r\n# # id = f'{url}{state_name}{percent_biden}{percent_trump}'\r\n\r\n# # Will create new StatePoll if the call returns Does Not Exist error\r\n# try:\r\n# # get State Object\r\n# state = State.objects.get(name=state_name)\r\n \r\n \r\n# try:\r\n# exist_poll = StatePoll.objects.get(state=state, url=url, percent_biden=percent_biden, percent_trump=percent_trump)\r\n# except:\r\n# state_poll = StatePoll(state=state, start_date=start_date, end_date=end_date, percent_trump=percent_trump, percent_biden=percent_biden, n=n, pollType=pollType, pollster=pollster, moe=moe, url=url)\r\n# state_poll.save()\r\n \r\n# except:\r\n# print(state_name)\r\n\r\n\r\nwith open('polls.csv', 'r') as file:\r\n\r\n rows = csv.reader(file, delimiter=',')\r\n \r\n # remove header\r\n\r\n next(rows)\r\n count = 0\r\n for row in rows:\r\n state = row[0]\r\n if 'Maine' not in state and 'Nebraska' not in state and 'National' not in state:\r\n count+=1\r\nprint(count)","sub_path":"core/management/commands/addpoll.py","file_name":"addpoll.py","file_ext":"py","file_size_in_byte":3385,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"378756782","text":"'''\n문제 2.\n자연수 N이 주어졌을 때, 1부터 N까지 한 줄에 하나씩 출력하는 프로그램을 작성하시오.\n'''\n\nnumbers = int(input('숫자를 입력하세요: '))\n\n# 아래에 코드를 작성해 주세요.\n\nRanges = range(1,numbers+1)\nfor Range in Ranges:\n print(Ranges)\n","sub_path":"00_startcamp/03_day/list/02_int.py","file_name":"02_int.py","file_ext":"py","file_size_in_byte":299,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"164967520","text":"from __future__ import absolute_import, print_function, division\nimport os\nimport logging\nfrom six import integer_types\nfrom six.moves import StringIO, reduce\nimport theano\nfrom theano import Apply\nfrom theano import tensor\nfrom theano.sandbox.cuda.type import CudaNdarrayType\nfrom theano.sandbox.cuda import GpuOp\nfrom theano.sandbox.cuda.basic_ops import (as_cuda_ndarray_variable,\n gpu_contiguous)\nfrom theano.tensor import as_tensor_variable\n_logger = logging.getLogger(__name__)\n\n\nclass BaseGpuCorrMM(GpuOp):\n \"\"\"\n Base class for `GpuCorrMM`, `GpuCorrMM_gradWeights` and\n `GpuCorrMM_gradInputs`. Cannot be used directly.\n\n Parameters\n ----------\n border_mode : {'valid', 'full', 'half'}\n Additionally, the padding size could be directly specified by an integer\n or a pair of integers\n subsample\n Perform subsampling of the output (default: (1, 1)).\n filter_dilation\n Perform subsampling of the input, also known as dilation (default: (1, 1)).\n pad\n *deprecated*, now you should always use border_mode.\n \"\"\"\n\n check_broadcast = False\n __props__ = ('border_mode', 'subsample', 'filter_dilation')\n\n def __init__(self, border_mode=\"valid\", subsample=(1, 1),\n filter_dilation=(1, 1), pad=None, binary=False):\n if pad is not None:\n _logger.warning(\n 'do not use pad for BaseGpuCorrMM; please set padding in '\n 'border_mode parameter, see the docstring for more details')\n if border_mode != \"valid\":\n raise ValueError(\"border_mode must be 'valid' if pad is given\")\n border_mode = pad\n if isinstance(border_mode, integer_types):\n border_mode = (border_mode, border_mode)\n if isinstance(border_mode, tuple):\n pad_h, pad_w = map(int, border_mode)\n border_mode = (pad_h, pad_w)\n if not ((isinstance(border_mode, tuple) and min(border_mode) >= 0) or\n border_mode in ('valid', 'full', 'half')):\n raise ValueError(\n 'invalid border_mode {}, which must be either '\n '\"valid\", \"full\", \"half\", an integer or a pair of'\n ' integers'.format(border_mode))\n self.border_mode = border_mode\n if len(subsample) != 2:\n raise ValueError(\"subsample must have two elements\")\n if len(filter_dilation) != 2:\n raise ValueError(\"filter_dilation must have two elements\")\n self.subsample = tuple(subsample)\n self.filter_dilation = tuple(filter_dilation)\n self.binary = binary\n\n @property\n def pad(self):\n if self.border_mode != 'valid':\n return self.border_mode\n return (0, 0)\n\n def __str__(self):\n return '%s{%s, %s, %s}' % (\n self.__class__.__name__,\n self.border_mode,\n str(self.subsample),\n str(self.filter_dilation))\n\n def flops(self, inp, outp):\n \"\"\"\n Useful with the hack in profiling to print the MFlops.\n\n \"\"\"\n # if the output shape is correct, then this gives the correct\n # flops for any direction, sampling, padding, and border mode\n inputs, filters = inp\n outputs, = outp\n assert inputs[1] == filters[1]\n # nb mul and add by output pixel\n flops = filters[2] * filters[3] * 2\n # nb flops by output image\n flops *= outputs[2] * outputs[3]\n # nb patch multiplied\n flops *= inputs[1] * filters[0] * inputs[0]\n return flops\n\n def c_headers(self):\n return ['cuda_ndarray.cuh', '<stdio.h>']\n\n def c_code_cache_version(self):\n # raise this whenever modifying any of the support_code_files\n return (1, 30)\n\n def c_support_code_apply(self, node, nodename):\n # REMEMBER TO RAISE c_code_cache_version when changing any of\n # these files\n files = ['corr_gemm.cu']\n codes = [open(os.path.join(os.path.split(__file__)[0], f)).read()\n for f in files]\n return reduce(str.__add__, codes)\n\n def c_code_helper(self, bottom, weights, top, direction, sub, height=None, width=None):\n \"\"\"\n This generates the C code for GpuCorrMM (direction=\"forward\"),\n GpuCorrMM_gradWeights (direction=\"backprop weights\"), and\n GpuCorrMM_gradInputs (direction=\"backprop inputs\").\n Depending on the direction, one of bottom, weights, top will\n receive the output, while the other two serve as inputs.\n\n Parameters\n ----------\n bottom\n Variable name of the input images in the forward pass,\n or the gradient of the input images in backprop wrt. inputs\n weights\n Variable name of the filters in the forward pass,\n or the gradient of the filters in backprop wrt. weights\n top\n Variable name of the output images / feature maps in the\n forward pass, or the gradient of the outputs in the backprop passes\n direction : {'forward', 'backprop weights', 'backprop inputs'}\n \"forward\" to correlate bottom with weights and store results in top,\n \"backprop weights\" to do a valid convolution of bottom with top\n (swapping the first two dimensions) and store results in weights,\n and \"backprop inputs\" to do a full convolution of top with weights\n (swapping the first two dimensions) and store results in bottom.\n sub\n Dictionary of substitutions useable to help generating the C code.\n height\n Required if self.subsample[0] != 1, a variable giving the height of\n the filters for direction=\"backprop weights\" or the height of the\n input images for direction=\"backprop inputs\".\n Required if self.border_mode == 'half', a variable giving the height\n of the filters for direction=\"backprop weights\".\n Not required otherwise, but if a value is given this will be checked.\n width\n Required if self.subsample[1] != 1, a variable giving the width of\n the filters for direction=\"backprop weights\" or the width of the\n input images for direction=\"backprop inputs\".\n Required if self.border_mode == 'half', a variable giving the width\n of the filters for direction=\"backprop weights\".\n Not required otherwise, but if a value is given this will be checked.\n\n \"\"\"\n callBinary = 0\n if self.binary == True:\n callBinary = 1\n\n print(\"callbinary = \" + str(callBinary))\n\n dH, dW = self.subsample\n dilH, dilW = self.filter_dilation\n if self.border_mode == \"half\":\n padH = padW = -1\n elif self.border_mode == \"full\":\n padH = padW = -2\n elif isinstance(self.border_mode, tuple):\n padH, padW = self.border_mode\n else:\n assert self.border_mode == \"valid\"\n padH = padW = 0\n if direction == \"forward\":\n direction = 0\n out = top\n elif direction == \"backprop weights\":\n direction = 1\n out = weights\n elif direction == \"backprop inputs\":\n direction = 2\n out = bottom\n else:\n raise ValueError(\"direction must be one of 'forward', \"\n \"'backprop weights', 'backprop inputs'\")\n # When subsampling, we cannot unambiguously infer the height and width\n # of bottom and weights from top, so we require them to be given.\n # Similarly, when pad=\"half\", we cannot infer the weight size.\n if height:\n height = '(*(npy_int*)(PyArray_DATA(%s)))' % height\n else:\n if ((direction != 0) and (dH != 1)) or ((direction == 1) and (padH == -1)):\n raise ValueError(\"height must be given for backprop with vertical sampling or pad='half'\")\n height = '-1'\n if width:\n width = '(*(npy_int*)(PyArray_DATA(%s)))' % width\n else:\n if ((direction != 0) and (dW != 1)) or ((direction == 1) and (padW == -1)):\n raise ValueError(\"width must be given for backprop with horizontal sampling or pad='half'\")\n width = '-1'\n sub = sub.copy()\n sub.update(locals())\n\n return \"\"\"\n // Mandatory args\n int direction = %(direction)s; // forward, bprop weights, bprop inputs\n\n // Optional args\n int dH = %(dH)s;\n int dW = %(dW)s;\n int dilH = %(dilH)s;\n int dilW = %(dilW)s;\n int padH = %(padH)s;\n int padW = %(padW)s;\n int callBinary = %(callBinary)s;\n\n CudaNdarray * bottom = %(bottom)s;\n CudaNdarray * weights = %(weights)s;\n CudaNdarray * top = %(top)s;\n CudaNdarray * out2 = NULL;\n\n // Obtain or infer kernel width and height\n // (we need to know it early to be able to handle auto-padding)\n int kH, kW, dil_kH, dil_kW;\n if (direction != 1) {\n // weight is an input variable, we can just read its shape\n kH = CudaNdarray_HOST_DIMS(weights)[2];\n kW = CudaNdarray_HOST_DIMS(weights)[3];\n }\n else {\n if (%(height)s != -1) {\n // kernel height is specified (perhaps vertical subsampling or half padding)\n kH = %(height)s;\n }\n else if (padH == -2) {\n // vertical full padding, we can infer the kernel height\n kH = (2 - CudaNdarray_HOST_DIMS(bottom)[2] + (CudaNdarray_HOST_DIMS(top)[2] - 1)*dH - 1) / dilH + 1;\n }\n else {\n // explicit padding, we can infer the kernel height\n kH = (CudaNdarray_HOST_DIMS(bottom)[2] + 2*padH - (CudaNdarray_HOST_DIMS(top)[2] - 1)*dH - 1) / dilH + 1 ;\n }\n if (%(width)s != -1) {\n kW = %(width)s;\n }\n else if (padW == -2) {\n kW = (2 - CudaNdarray_HOST_DIMS(bottom)[3] + (CudaNdarray_HOST_DIMS(top)[3] - 1) * dW - 1) / dilW + 1;\n }\n else {\n kW = (CudaNdarray_HOST_DIMS(bottom)[3] + 2*padW - (CudaNdarray_HOST_DIMS(top)[3] - 1) * dW - 1) / dilW + 1;\n }\n }\n\n // Implicit dilated kernel size\n dil_kH = (kH - 1) * dilH + 1;\n dil_kW = (kW - 1) * dilW + 1;\n\n // Auto-padding if requested\n if (padH == -1) { // vertical half padding\n padH = dil_kH / 2;\n }\n else if (padH == -2) { // vertical full padding\n padH = dil_kH - 1;\n }\n else if (padH < 0) {\n PyErr_SetString(PyExc_ValueError, \"BaseGpuCorrMM: padH must be >= -2\");\n %(fail)s\n }\n if (padW == -1) { // horizontal half padding\n padW = dil_kW / 2;\n }\n else if (padW == -2) { // horizontal full padding\n padW = dil_kW - 1;\n }\n else if (padW < 0) {\n PyErr_SetString(PyExc_ValueError, \"BaseGpuCorrMM: padW must be >= -2\");\n %(fail)s\n }\n\n // Infer output shape\n int out_dim[4];\n switch(direction) {\n case 0: // forward pass\n // output is top: (batchsize, num_filters, height, width)\n // height and width: top = (bottom + 2*pad - ((weight-1)*dil + 1)) / sample + 1\n out_dim[0] = CudaNdarray_HOST_DIMS(bottom)[0];\n out_dim[1] = CudaNdarray_HOST_DIMS(weights)[0];\n out_dim[2] = (CudaNdarray_HOST_DIMS(bottom)[2] + 2*padH - ((CudaNdarray_HOST_DIMS(weights)[2]-1)*dilH + 1)) / dH + 1;\n out_dim[3] = (CudaNdarray_HOST_DIMS(bottom)[3] + 2*padW - ((CudaNdarray_HOST_DIMS(weights)[3]-1)*dilW + 1)) / dW + 1;\n if (out_dim[0] < 0 || out_dim[1] < 0 || out_dim[2] <= 0 || out_dim[3] <= 0)\n {\n PyErr_Format(PyExc_ValueError,\n \"GpuCorrMM: impossible output shape\\\\n\"\n \" bottom shape: %%ld x %%ld x %%ld x %%ld\\\\n\"\n \" weights shape: %%ld x %%ld x %%ld x %%ld\\\\n\"\n \" top shape: %%ld x %%ld x %%ld x %%ld\\\\n\",\n CudaNdarray_HOST_DIMS(bottom)[0], CudaNdarray_HOST_DIMS(bottom)[1],\n CudaNdarray_HOST_DIMS(bottom)[2], CudaNdarray_HOST_DIMS(bottom)[3],\n CudaNdarray_HOST_DIMS(weights)[0], CudaNdarray_HOST_DIMS(weights)[1],\n CudaNdarray_HOST_DIMS(weights)[2], CudaNdarray_HOST_DIMS(weights)[3],\n out_dim[0], out_dim[1], out_dim[2], out_dim[3]);\n %(fail)s\n }\n break;\n case 1: // backprop wrt. weights\n // output is weights: (num_filters, num_channels, height, width)\n // height and width: weights = (bottom + 2*pad - (top - 1) * sample - 1) / dil + 1\n out_dim[0] = CudaNdarray_HOST_DIMS(top)[1];\n out_dim[1] = CudaNdarray_HOST_DIMS(bottom)[1];\n out_dim[2] = kH; // already inferred further above\n out_dim[3] = kW; // how convenient\n if (out_dim[0] < 0 || out_dim[1] < 0 || out_dim[2] <= 0 || out_dim[3] <= 0)\n {\n PyErr_Format(PyExc_ValueError,\n \"GpuCorrMM backprop wrt. weights: impossible output shape\\\\n\"\n \" bottom shape: %%ld x %%ld x %%ld x %%ld\\\\n\"\n \" weights shape: %%ld x %%ld x %%ld x %%ld\\\\n\"\n \" top shape: %%ld x %%ld x %%ld x %%ld\\\\n\",\n CudaNdarray_HOST_DIMS(bottom)[0], CudaNdarray_HOST_DIMS(bottom)[1],\n CudaNdarray_HOST_DIMS(bottom)[2], CudaNdarray_HOST_DIMS(bottom)[3],\n out_dim[0], out_dim[1], out_dim[2], out_dim[3],\n CudaNdarray_HOST_DIMS(top)[0], CudaNdarray_HOST_DIMS(top)[1],\n CudaNdarray_HOST_DIMS(top)[2], CudaNdarray_HOST_DIMS(top)[3]);\n %(fail)s\n }\n break;\n case 2: // backprop wrt. inputs\n // output is bottom: (batchsize, num_channels, height, width)\n // height and width: bottom = (top - 1) * sample + (weights-1)*dil + 1 - 2*pad\n out_dim[0] = CudaNdarray_HOST_DIMS(top)[0];\n out_dim[1] = CudaNdarray_HOST_DIMS(weights)[1];\n out_dim[2] = (%(height)s != -1) ? %(height)s : (CudaNdarray_HOST_DIMS(top)[2] - 1) * dH + (CudaNdarray_HOST_DIMS(weights)[2]-1)*dilH + 1 - 2*padH;\n out_dim[3] = (%(width)s != -1) ? %(width)s : (CudaNdarray_HOST_DIMS(top)[3] - 1) * dW + (CudaNdarray_HOST_DIMS(weights)[3]-1)*dilW + 1 - 2*padW;\n if (out_dim[0] < 0 || out_dim[1] < 0 || out_dim[2] <= 0 || out_dim[3] <= 0)\n {\n PyErr_Format(PyExc_ValueError,\n \"GpuCorrMM backprop wrt. inputs: impossible output shape\\\\n\"\n \" bottom shape: %%ld x %%ld x %%ld x %%ld\\\\n\"\n \" weight shape: %%ld x %%ld x %%ld x %%ld\\\\n\"\n \" top shape: %%ld x %%ld x %%ld x %%ld\\\\n\",\n out_dim[0], out_dim[1], out_dim[2], out_dim[3],\n CudaNdarray_HOST_DIMS(weights)[0], CudaNdarray_HOST_DIMS(weights)[1],\n CudaNdarray_HOST_DIMS(weights)[2], CudaNdarray_HOST_DIMS(weights)[3],\n CudaNdarray_HOST_DIMS(top)[0], CudaNdarray_HOST_DIMS(top)[1],\n CudaNdarray_HOST_DIMS(top)[2], CudaNdarray_HOST_DIMS(top)[3]);\n %(fail)s\n }\n break;\n default:\n PyErr_SetString(PyExc_ValueError, \"BaseGpuCorrMM: direction must be 0, 1, or 2\\\\n\");\n %(fail)s\n }\n\n // Prepare output array\n if ( !(%(out)s\n && %(out)s->nd==4\n && CudaNdarray_is_c_contiguous(%(out)s)\n && CudaNdarray_HOST_DIMS(%(out)s)[0]==out_dim[0]\n && CudaNdarray_HOST_DIMS(%(out)s)[1]==out_dim[1]\n && CudaNdarray_HOST_DIMS(%(out)s)[2]==out_dim[2]\n && CudaNdarray_HOST_DIMS(%(out)s)[3]==out_dim[3]))\n {\n Py_XDECREF(%(out)s);\n %(out)s = (CudaNdarray*)CudaNdarray_NewDims(4,out_dim);\n if (NULL == %(out)s)\n {\n PyErr_Format(PyExc_RuntimeError,\n \"BaseGpuCorrMM: Failed to allocate output of %%d x %%d x %%d x %%d\",\n out_dim[0], out_dim[1], out_dim[2], out_dim[3]);\n %(fail)s\n }\n }\n\n // Call CUDA code\n out2 = corrMMWrapper(%(bottom)s, %(weights)s, %(top)s, direction, dH, dW, dilH, dilW, padH, padW, callBinary);\n if (out2==NULL)\n {\n %(fail)s\n }\n assert (out2 == %(out)s);\n\n\"\"\" % sub\n\n\nclass GpuCorrMM(BaseGpuCorrMM):\n \"\"\"\n GPU correlation implementation using Matrix Multiplication.\n\n Parameters\n ----------\n border_mode\n The width of a border of implicit zeros to pad the\n input with. Must be a tuple with 2 elements giving the numbers of rows\n and columns to pad on each side, or a single integer to pad the same\n on all sides, or a string shortcut setting the padding at runtime:\n ``'valid'`` for ``(0, 0)`` (valid convolution, no padding), ``'full'``\n for ``(kernel_rows - 1, kernel_columns - 1)`` (full convolution),\n ``'half'`` for ``(kernel_rows // 2, kernel_columns // 2)`` (same\n convolution for odd-sized kernels). Note that the two widths are each\n applied twice, once per side (left and right, top and bottom).\n subsample\n The subsample operation applied to each output image.\n Should be a tuple with 2 elements.\n `(sv, sh)` is equivalent to `GpuCorrMM(...)(...)[:,:,::sv, ::sh]`,\n but faster.\n Set to `(1, 1)` to disable subsampling.\n filter_dilation\n The filter dilation operation applied to each input image.\n Should be a tuple with 2 elements.\n Set to `(1, 1)` to disable filter dilation.\n pad\n Deprecated alias for `border_mode`.\n\n Notes\n -----\n Currently, the Op requires the inputs, filters and outputs to be\n C-contiguous. Use :func:`gpu_contiguous\n <theano.sandbox.cuda.basic_ops.gpu_contiguous>` on these arguments\n if needed.\n\n You can either enable the Theano flag `optimizer_including=conv_gemm`\n to automatically replace all convolution operations with `GpuCorrMM`\n or one of its gradients, or you can use it as a replacement for\n :func:`conv2d <theano.tensor.nnet.conv.conv2d>`, called as\n `GpuCorrMM(subsample=...)(image, filters)`. The latter is currently\n faster, but note that it computes a correlation -- if you need to\n compute a convolution, flip the filters as `filters[:,:,::-1,::-1]`.\n\n ..warning:: For 700 series Nvidia GPUs of compute capability 3.5 and CUDA 5.0\n to 6.0, there is a bug in CUBLAS' matrix multiplication function that\n can make GpuCorrMM or its gradients crash for some input and filter\n shapes. So if you have a Tesla K20, Tesla K40, Quadro K6000, GeForce GT\n 640 (DDR5), GeForce GTX 780 (or Ti), GeForce GTX TITAN (or Black or Z)\n and experience a crash, switching to CUDA 6.5 or CUDA 4.2 should fix it.\n If this is not possible, changing the input or filter shapes (e.g., the\n batchsize or number of filters) may also work around the CUBLAS bug.\n\n \"\"\"\n def __init__(self, border_mode=\"valid\",\n subsample=(1, 1),\n filter_dilation=(1, 1),\n pad=None,\n binary=False):\n super(GpuCorrMM, self).__init__(border_mode, subsample,\n filter_dilation, pad, binary)\n\n def make_node(self, img, kern):\n img = as_cuda_ndarray_variable(img)\n kern = as_cuda_ndarray_variable(kern)\n if img.type.ndim != 4:\n raise TypeError('img must be 4D tensor')\n if kern.type.ndim != 4:\n raise TypeError('kern must be 4D tensor')\n\n broadcastable = [img.type.broadcastable[0], kern.type.broadcastable[0],\n False, False]\n return Apply(self, [img, kern], [CudaNdarrayType(broadcastable)()])\n\n def c_code(self, node, nodename, inp, out_, sub):\n bottom, weights = inp\n top, = out_\n direction = \"forward\"\n return super(GpuCorrMM, self).c_code_helper(bottom, weights, top, direction, sub)\n\n def grad(self, inp, grads):\n bottom, weights = inp\n top, = grads\n top = gpu_contiguous(top)\n d_bottom = GpuCorrMM_gradInputs(self.border_mode,\n self.subsample,\n self.filter_dilation)(\n weights, top, bottom.shape[-2:])\n d_weights = GpuCorrMM_gradWeights(self.border_mode,\n self.subsample,\n self.filter_dilation)(\n bottom, top, weights.shape[-2:])\n return d_bottom, d_weights\n\n\nclass GpuCorrMM_gradWeights(BaseGpuCorrMM):\n \"\"\"\n Gradient wrt. filters for `GpuCorrMM`.\n\n Notes\n -----\n You will not want to use this directly, but rely on Theano's automatic\n differentiation or graph optimization to use it as needed.\n\n \"\"\"\n\n def __init__(self, border_mode=\"valid\",\n subsample=(1, 1),\n filter_dilation=(1, 1),\n pad=None):\n super(GpuCorrMM_gradWeights, self).__init__(border_mode,\n subsample,\n filter_dilation,\n pad)\n\n def make_node(self, img, topgrad, shape=None):\n img = as_cuda_ndarray_variable(img)\n topgrad = as_cuda_ndarray_variable(topgrad)\n if img.type.ndim != 4:\n raise TypeError('img must be 4D tensor')\n if topgrad.type.ndim != 4:\n raise TypeError('topgrad must be 4D tensor')\n if shape is None:\n if self.subsample != (1, 1) or self.border_mode == \"half\":\n raise ValueError('shape must be given if subsample != (1, 1)'\n ' or border_mode == \"half\"')\n height_width = []\n else:\n height_width = [shape[0], shape[1]]\n assert shape[0].ndim == 0\n assert shape[1].ndim == 0\n\n broadcastable = [topgrad.type.broadcastable[1], img.type.broadcastable[1],\n False, False]\n return Apply(self, [img, topgrad] + height_width, [CudaNdarrayType(broadcastable)()])\n\n def c_code(self, node, nodename, inp, out_, sub):\n bottom, top = inp[:2]\n height, width = inp[2:] or (None, None)\n weights, = out_\n direction = \"backprop weights\"\n return super(GpuCorrMM_gradWeights, self).c_code_helper(bottom, weights, top, direction, sub, height, width)\n\n def grad(self, inp, grads):\n bottom, top = inp[:2]\n weights, = grads\n weights = gpu_contiguous(weights)\n d_bottom = GpuCorrMM_gradInputs(self.border_mode,\n self.subsample,\n self.filter_dilation)(weights,\n top,\n bottom.shape[-2:])\n d_top = GpuCorrMM(\n self.border_mode, self.subsample, self.filter_dilation)(bottom, weights)\n d_height_width = (\n theano.gradient.DisconnectedType()(),\n ) * 2 if len(inp) == 4 else ()\n return (d_bottom, d_top) + d_height_width\n\n def connection_pattern(self, node):\n if node.nin == 2:\n return [[1], [1]]\n else:\n return [[1], [1], [0], [0]] # no connection to height, width\n\n\nclass GpuCorrMM_gradInputs(BaseGpuCorrMM):\n \"\"\"\n Gradient wrt. inputs for `GpuCorrMM`.\n\n Notes\n -----\n You will not want to use this directly, but rely on Theano's automatic\n differentiation or graph optimization to use it as needed.\n\n \"\"\"\n\n def __init__(self, border_mode=\"valid\",\n subsample=(1, 1),\n filter_dilation=(1, 1),\n pad=None):\n super(GpuCorrMM_gradInputs, self).__init__(border_mode, subsample,\n filter_dilation, pad)\n\n def make_node(self, kern, topgrad, shape=None):\n kern = as_cuda_ndarray_variable(kern)\n topgrad = as_cuda_ndarray_variable(topgrad)\n if kern.type.ndim != 4:\n raise TypeError('kern must be 4D tensor')\n if topgrad.type.ndim != 4:\n raise TypeError('topgrad must be 4D tensor')\n if shape is None:\n if self.subsample != (1, 1):\n raise ValueError('shape must be given if subsample != (1, 1)')\n height_width = []\n else:\n height_width = [shape[0], shape[1]]\n assert shape[0].ndim == 0\n assert shape[1].ndim == 0\n\n broadcastable = [topgrad.type.broadcastable[0], kern.type.broadcastable[1],\n False, False]\n return Apply(self, [kern, topgrad] + height_width, [CudaNdarrayType(broadcastable)()])\n\n def c_code(self, node, nodename, inp, out_, sub):\n weights, top = inp[:2]\n height, width = inp[2:] or (None, None)\n bottom, = out_\n direction = \"backprop inputs\"\n return super(GpuCorrMM_gradInputs, self).c_code_helper(bottom, weights, top, direction, sub, height, width)\n\n def grad(self, inp, grads):\n weights, top = inp[:2]\n bottom, = grads\n bottom = gpu_contiguous(bottom)\n d_weights = GpuCorrMM_gradWeights(self.border_mode,\n self.subsample,\n self.filter_dilation)(bottom,\n top,\n weights.shape[-2:])\n d_top = GpuCorrMM(self.border_mode,\n self.subsample,\n self.filter_dilation)(bottom, weights)\n d_height_width = (\n theano.gradient.DisconnectedType()(),\n ) * 2 if len(inp) == 4 else ()\n return (d_weights, d_top) + d_height_width\n\n def connection_pattern(self, node):\n if node.nin == 2:\n return [[1], [1]]\n else:\n return [[1], [1], [0], [0]] # no connection to height, width\n\n\n","sub_path":"Run-time/myblas.py","file_name":"myblas.py","file_ext":"py","file_size_in_byte":26299,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"586024891","text":"import math\nimport vtk\nfrom PythonMetricsCalculator import PerkEvaluatorMetric\n\nclass TraceTrajectory( PerkEvaluatorMetric ):\n\n # Static methods\n @staticmethod\n def GetMetricName():\n return \"Trace Trajectory\"\n \n @staticmethod \n def GetMetricUnit():\n return \"\"\n \n @staticmethod\n def GetAnatomyRoles():\n return { \"OutputModel\": \"vtkMRMLModelNode\" }\n \n \n # Instance methods\n def __init__( self ):\n PerkEvaluatorMetric.__init__( self )\n \n self.curvePoints = vtk.vtkPoints()\n self.curveLines = vtk.vtkCellArray()\n self.curvePolyData = vtk.vtkPolyData()\n self.counter = 0\n \n self.curvePolyData.SetPoints( self.curvePoints )\n self.curvePolyData.SetLines( self.curveLines )\n \n def SetAnatomy( self, role, node ): \n if ( role == \"OutputModel\" ):\n node.SetAndObservePolyData( self.curvePolyData )\n if ( node.GetModelDisplayNode() is None ):\n node.CreateDefaultDisplayNodes()\n modelDisplayNode = node.GetModelDisplayNode()\n return True\n \n return False\n \n def AddTimestamp( self, time, matrix, point, role ): \n # Some initialization for the first point\n if ( self.curveLines.GetNumberOfCells() == 0 ):\n self.curvePoints.InsertNextPoint( point[ 0 ], point[ 1 ], point[ 2 ] )\n self.curveLines.InsertNextCell( 1 )\n self.curveLines.InsertCellPoint( 0 )\n \n self.curvePoints.InsertPoint( self.counter + 1, point[ 0 ], point[ 1 ], point[ 2 ] )\n \n self.curveLines.InsertNextCell( 2 ) # Because there are two points in the cell\n self.curveLines.InsertCellPoint( self.counter )\n self.curveLines.InsertCellPoint( self.counter + 1 )\n self.counter += 1","sub_path":"TraceTrajectory.py","file_name":"TraceTrajectory.py","file_ext":"py","file_size_in_byte":1673,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"640644622","text":"\nimport numpy as np\nimport tensorflow as tf\n\nclass OneToOneMultiFeature:\n \n def __init__(self):\n self.X = list()\n self.Y = list()\n self.epochs = 4000\n self.timestep = 1\n self.features = 2\n self.size = 25\n self.create_dataset()\n self.model = self.create_model()\n self.run()\n self.test()\n \n def create_dataset(self):\n # Y is the multiplication of X1 and X2. For instance, first element of X1 is 2 and X2 is 3. So Y is the product of X1 and X2 i.e 6 \n # We train the lSTM to predict the cross product of two features in an input data\n\n X1 = list()\n X2 = list()\n X1 = [(x+1)*2 for x in range(self.size)]\n X2 = [(x+1)*3 for x in range(self.size)]\n self.X = np.column_stack((X1, X2))\n print(self.X)\n self.Y = [x1*x2 for x1,x2 in zip(X1,X2)]\n\n # The expected dimension to LSTM/RNN is in 3D shape i.e. (samples, time-steps, features). \n # Original shape is (25, 2)\n # We are converting to 25 as batch, 1 as timestep(remeber we are trying to implement one to one, so timestep is always one because the one input), \n # and 2 as feature or sequence length, i.e. (25, 1, 2)\n self.X = np.array(self.X).reshape(self.size, self.timestep, self.features)\n print(self.X.shape)\n\n def create_model(self):\n return tf.keras.Sequential([\n tf.keras.layers.LSTM(80, activation='relu',\n input_shape=(self.timestep, self.features)),\n tf.keras.layers.Dense(10, activation='relu'),\n tf.keras.layers.Dense(1)\n ])\n\n def run(self):\n self.model.compile(optimizer='adam', loss='mse')\n print(self.model.summary())\n self.model.fit(self.X, self.Y, epochs=self.epochs, validation_split=0.2, batch_size=5)\n\n def test(self):\n print(\"-------------------------------------------\")\n print(\"Test result: \")\n test_input = np.array([55,80])\n test_input = test_input.reshape((1, self.timestep, self.features))\n test_output = self.model.predict(test_input, verbose=0)\n print(test_output)\n\nif __name__== \"__main__\":\n OneToOneMultiFeature()","sub_path":"one_to_one/multi_feature.py","file_name":"multi_feature.py","file_ext":"py","file_size_in_byte":2232,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"645959200","text":"import json\n\nimport time\n\nMAX_PACKET_SIZE = 65536\n\ncurrent_milli_time = lambda: int(round(time.time() * 1000))\n\n\"\"\"\nSplit message into chunks and send to client\n\"\"\"\n\n\nclass PacketSender:\n @staticmethod\n async def send_packet(sock_out, response_type, data, room):\n data_size = len(data)\n\n packet_id = current_milli_time()\n\n info = {\n \"type\": response_type,\n \"id\": packet_id,\n \"size\": data_size\n }\n\n await sock_out.emit(\"response_header\", str.encode(json.dumps(info)))\n\n byte_packet_id = packet_id.to_bytes(8, byteorder='big')\n\n for i in range(0, data_size, MAX_PACKET_SIZE):\n chunk = byte_packet_id + data[i:i + MAX_PACKET_SIZE]\n await sock_out.emit(\"response\", chunk, room=room)\n","sub_path":"server/src/network/packet_sender.py","file_name":"packet_sender.py","file_ext":"py","file_size_in_byte":789,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"123910759","text":"# -*- coding: utf-8 -*-\n\nfrom django.conf.urls.defaults import patterns, include, url\n\nurlpatterns = patterns('mensajesInstantaneos.views',\n\turl(r'^index/$', 'index'),\n\turl(r'^login/$', 'login_user'),\n\turl(r'^home/$', 'home'),\n\turl(r'^crearCuenta/$', 'crearCuenta'),\n\turl(r'^logout/$', 'salir'),\n\turl(r'^perfil/(?P<usuario>\\w+)/$', 'enviarMensaje'),\n\turl(r'^find_friends/$', 'encontrarAmigos'),\n\turl(r'^agregar_amigo/(?P<amigo>\\w+)/$', 'agregarAmigo'),\n)\n","sub_path":"mensajesInstantaneos/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":455,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"452780957","text":"#!/usr/bin/env python\n\nfrom __future__ import print_function\nimport os\nfrom smugpy import SmugMug\nfrom smugphoto.helper.fileutil import FileUtil\n\n\nclass AuthUtil(object):\n def __init__(self, api_key=None, oauth_secret=None, apiKeyDir=None, accessKeyDir=None):\n self.api_key = api_key\n self.oauth_secret = oauth_secret\n self.app_name = 'MySmug'\n self._apiKeyDir = apiKeyDir\n self._accessKeyDir = accessKeyDir\n self.access_token = None\n\n def _smugmugOauthRequestToken(self, access=\"Public\", perm=\"Read\"):\n smugmug = SmugMug(api_key=self.api_key, oauth_secret=self.oauth_secret, app_name=self.app_name)\n\n # Get a token that is short-lived (probably about 5 minutes) and can be used\n # only to setup authorization at SmugMug\n response = smugmug.auth_getRequestToken()\n\n # Get the URL that the user must visit to authorize this app (implicilty includes the request token in the URL)\n url = smugmug.authorize(access=access, perm=perm)\n return url, response['Auth'] # (should contain a 'Token')\n\n @staticmethod\n def _userAuthorizeAtSmugmug(url):\n input(\"Authorize app at %s\\n\\nPress Enter when complete.\\n\" % (url))\n\n def _smugmugOauthGetAccessToken(self, requestToken):\n # Use the request token to log in (which should be authorized now)\n smugmug = SmugMug(api_key=self.api_key, oauth_secret=self.oauth_secret,\n oauth_token=requestToken['Token']['id'],\n oauth_token_secret=requestToken['Token']['Secret'],\n app_name=self.app_name)\n\n # The request token is good for 1 operation: to get an access token.\n response = smugmug.auth_getAccessToken()\n\n # The access token should be good until the user explicitly\n # disables it at smugmug.com in their settings panel.\n return response['Auth']\n\n # Log into smugmug.com with an authorized accessToken. The accessToken includes\n # the user's identity and, effectively, a password to get this application into\n # the account.\n def _smugmugOauthUseAccessToken(self, accessToken):\n # Use the access token to log in\n smugmug = SmugMug(api_key=self.api_key, oauth_secret=self.oauth_secret,\n oauth_token=accessToken['Token']['id'],\n oauth_token_secret=accessToken['Token']['Secret'],\n app_name=self.app_name)\n return smugmug\n\n def getSmug(self):\n # try to read\n self._loadAPIKey()\n\n try:\n access_token = FileUtil.readYamlToDict(self.accessKeyDir)\n if access_token is None:\n access_token = self._tryToGetAccessKeyFromAPIKey()\n except FileExistsError:\n access_token = self._tryToGetAccessKeyFromAPIKey()\n\n self.access_token = access_token\n\n return self._smugmugOauthUseAccessToken(access_token)\n\n\n def _tryToGetAccessKeyFromAPIKey(self):\n (url, requestToken) = self._smugmugOauthRequestToken()\n self._userAuthorizeAtSmugmug(url)\n access_token = self._smugmugOauthGetAccessToken(requestToken)\n FileUtil.writeDictToYaml(access_token, self.accessKeyDir)\n\n return access_token\n\n def _loadAPIKey(self):\n keys = FileUtil.readYamlToDict(self.apiKeyDir)\n self.api_key = keys[\"API Key\"]\n self.oauth_secret = keys['Oauth Secret']\n self.app_name = keys[\"App Name\"]\n\n @property\n def accessKeyDir(self):\n if self._apiKeyDir is None:\n return os.path.abspath(os.path.join(os.getcwd(), r'..\\..\\tests\\auth\\oauth_access.yaml'))\n else:\n return os.path.join(os.path.abspath(self._accessKeyDir), r'..\\..\\tests\\auth\\oauth_access.yaml')\n\n @property\n def apiKeyDir(self):\n if self._apiKeyDir is None:\n return os.path.abspath(os.path.join(os.getcwd(), r'..\\..\\tests\\auth\\keys.yaml'))\n else:\n return os.path.join(os.path.abspath(self._apiKeyDir), r'..\\..\\tests\\auth\\'keys.yaml')\n\n # def saveAccessToken(self, accessDict, accessKeyDir=None):\n # if accessKeyDir is None:\n # filepath = os.path.abspath(os.path.join(os.getcwd(), 'access_key.yaml'))\n # else:\n # filepath = os.path.join(os.path.abspath(accessKeyDir), 'access_key.yaml')\n #\n # FileUtil.writeDictToYaml(accessDict, filepath)\n #\n # def readAccessToken(self, accessKeyDir=None):\n # if accessKeyDir is None:\n # filepath = os.path.abspath(os.path.join(os.getcwd(), 'access_key.yaml'))\n # else:\n # filepath = os.path.join(os.path.abspath(accessKeyDir), 'access_key.yaml')\n #\n # return FileUtil.readYamlToDict(filepath)\n\n\nif __name__ == '__main__':\n # ###\n # ### Main\n # ###\n # API_KEY = \"Ai1WhX5ErNtHYR5YFg4qFAiww6PGZs1d\"\n # OAUTH_SECRET = \"968d0e37c50b47a2ca04b28da556a8f0\" # From SmugMug Settings -> Discovery -> API Keys\n # APP_NAME = \"mySmugTest\"\n # myAccessToken = {'User': {'Name': 'Huy Le',\n # 'URL': 'https://hizzle.smugmug.com',\n # 'id': 1813307, 'AccountType': 'Portfolio',\n # 'SmugVault': False,\n # 'FileSizeLimit': 157286400,\n # 'AccountStatus': 'Active',\n # 'NickName': 'Hizzle'},\n # 'Token': {'Secret': 'e0d8224bcb517f481ebd7a7aca751fe3d34fd657f82e998f2816f1d76424a0b3',\n # 'id': '5635a8a6c820780ddf1a5abf13a6a07b',\n # 'Access': 'Public',\n # 'Permissions': 'Read'}}\n\n mySmug = AuthUtil().getSmug()\n\n albums = mySmug.albums_get()\n for album in albums[\"Albums\"]:\n print(\"{}, {}\".format(album[\"id\"], album[\"Title\"]))\n","sub_path":"smugphoto/helper/auth.py","file_name":"auth.py","file_ext":"py","file_size_in_byte":6018,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"522473212","text":"import datetime\nimport json\nimport os\nimport sys\n\nfrom django.contrib.auth import logout as auth_logout\nfrom django.contrib.auth.decorators import login_required\nfrom django.core.urlresolvers import reverse\nfrom django.http import HttpResponse, HttpResponseRedirect, Http404\nfrom django.shortcuts import render, get_list_or_404, redirect\nfrom django.views import generic\nfrom django.views.decorators import csrf\nfrom oauth2client import client\nfrom oauth2client.service_account import ServiceAccountCredentials\nfrom distutils.util import strtobool\n\nfrom mysite.settings import PROJECT_ROOT\nfrom .fetch_data import update_database\nfrom .google_manipulation import read_bgColor, sheet_authorization, get_sheet\nfrom .google_manipulation import write_sheet, write_cell\nfrom .models import Customer, Orders\n\n# Create your views here.\n\nINFO_SHIPPED = {\n 'action': 'changeBackgroundColor',\n 'sheetId': 309611827,\n 'startRowIndex': 1835,\n 'endRowIndex': 1836,\n 'startColumnIndex': 10,\n 'endColumnIndex': 11,\n 'red': 0.98823529,\n 'green': 0.89803922,\n 'blue': 0.80392158\n}\n\nINFO_PARTIAL = {\n 'action': 'changeBackgroundColor',\n 'sheetId': 309611827,\n 'startRowIndex': 1835,\n 'endRowIndex': 1836,\n 'startColumnIndex': 10,\n 'endColumnIndex': 11,\n 'red': 0.81568629,\n 'green': 0.87843138,\n 'blue': 0.89019608\n}\n\n\nclass Index(generic.TemplateView):\n template_name = \"kelly/old/index.html\"\n\n\ndef search(request):\n if not request.method == 'POST':\n return render(request, 'kelly/index.html', {\n 'error_message': '你並未輸入任何關鍵字。'\n })\n\n keyword_type = request.POST['keyword_type']\n keyword = request.POST['keyword']\n\n if not keyword or keyword == \"\" or keyword is None:\n return render(request, 'kelly/index.html', {\n 'error_message': '你並未輸入任何關鍵字。'\n })\n\n switch = {\n 'name': select_by_name(keyword),\n 'Facebook': select_by_facebook(keyword),\n 'bank': select_by_bank(keyword),\n 'amount': select_by_amount(keyword),\n 'transfer_date': select_by_transfer_date(keyword),\n 'order': select_by_order(keyword)\n }\n\n result = switch.get(keyword_type, None)\n if not result:\n return render(request, 'kelly/index.html', {\n 'error_message': \"找不到紀錄。\"\n })\n elif result is None:\n return render(request, \"kelly/index.html\", {\n 'error_message': \"找不到紀錄。\"\n })\n else:\n if type(result[0]) == Customer:\n customer = result\n # return HttpResponseRedirect(reverse('kelly:results', kwargs={'customer': customer}))\n return HttpResponseRedirect(reverse('kelly:cus_detail', args=(customer[0].id,)))\n # return render(request, 'kelly/results.html', {'customer': customer})\n else:\n orders = result\n # return HttpResponseRedirect(reverse('kelly:results', kwargs={'customer': customer}))\n return render(request, 'kelly/results.html', {'orders': orders})\n\n\ndef select_by_name(_name):\n try:\n result = get_list_or_404(Customer, customer_name__contains=_name)\n except Http404:\n return False\n return result\n\n\ndef select_by_facebook(_facebook):\n try:\n result = get_list_or_404(Customer, customer_Facebook__contains=_facebook)\n except Http404:\n return False\n return result\n\n\ndef select_by_bank(_bank):\n try:\n result = get_list_or_404(Orders, bank__contains=_bank)\n except Http404:\n return False\n return result\n\n\ndef select_by_amount(_p_amount):\n try:\n _amount = int(_p_amount)\n result = get_list_or_404(Orders, amount=_amount)\n except (Http404, ValueError):\n return False\n return result\n\n\ndef select_by_transfer_date(_p_transfer_date):\n try:\n _transfer_date = datetime.datetime.strptime(_p_transfer_date, '%m/%d/%Y')\n result = get_list_or_404(Orders, transfer_date=_transfer_date)\n except (Http404, ValueError):\n return False\n return result\n\n\ndef select_by_order(_order):\n try:\n result = get_list_or_404(Orders, order_content__contains=_order)\n except Http404:\n try:\n result = get_list_or_404(Orders, order_model__contains=_order)\n except Http404:\n return False\n return result\n\n\nclass CustomerDetailView(generic.DetailView):\n model = Customer\n template_name = 'kelly/old/cus_detail.html'\n\n\ndef show_angular(request):\n return render(request, 'kelly/old/angular/index.html')\n\n\ndef js_index(request):\n return render(request, 'kelly/index.html', {'user': request.user})\n\n\ndef js_welcome(request):\n return render(request, 'kelly/welcome.html')\n\n\ndef js_search(request):\n\n if not request.session.get('dataFetched'):\n fetch(request)\n request.session['dataFetched'] = True\n request.session.set_expiry(600)\n\n return render(request, 'kelly/search.html')\n\n\ndef js_control(request):\n return render(request, 'kelly/control.html')\n\n\n@csrf.csrf_protect\ndef ajax_search(request):\n if request.method == 'POST':\n try:\n csrf_str = request.POST.get('csrfmiddlewaretoken', False)\n keyword_type = request.POST.get('keyword_type', False)\n keyword = request.POST.get('keyword', False)\n return render(request, 'kelly/old/ajax_test.html', {'msg': request})\n except:\n e = sys.exc_info()\n return render(request, 'kelly/old/ajax_test.html', {'msg': str(e)})\n else:\n return render(request, 'kelly/old/ajax_test.html')\n\n\n@csrf.csrf_exempt\ndef test_post(request):\n if request.method == 'GET':\n\n order = Orders.objects.all().values()\n\n myDict = dict()\n myList = list()\n\n for _var in order:\n myDict = _var\n realName, faceBook, phone = get_customer_data(_var['customer_id'])\n myDict['customer_name'] = realName\n myDict['customer_Facebook'] = faceBook\n myDict['customer_phone'] = phone\n\n myStr = datetime.datetime.strftime(myDict['transfer_date'], '%Y. %m. %d')\n intYear = int(myStr[0:4]) - 1911\n remains = myStr[4:]\n myFinalDate = str(intYear) + remains\n\n myDict['transfer_date'] = myFinalDate\n # myDict['bank'] = str(myDict['bank']).replace('-', ' | ')\n\n myList.append(myDict)\n\n jsonText = json.dumps(myList)\n\n return HttpResponse(jsonText, {'user': request.user})\n\n elif request.method == 'POST':\n\n string = request.body\n string = string.decode('utf-8')\n myDict = json.loads(string)\n\n if myDict['action'] and myDict['action'] == 'shipping':\n\n service_account_access(request)\n\n spreadsheet_service = sheet_authorization(request, use_service=True)\n\n # if not spreadsheet_service:\n # return HttpResponseRedirect(reverse('kelly:oauth2callback'))\n\n order = Orders.objects.get(pk=myDict['id'])\n if myDict.get('isShipped') == 'partial':\n order.isShipped = False\n order.isPartialShipped = True\n else:\n order.isShipped = myDict['isShipped']\n order.isPartialShipped = False\n order.save()\n\n info = None\n values = list()\n _content = list()\n\n if not order.isShipped and order.isPartialShipped:\n info = INFO_PARTIAL\n _content.append('Partial')\n values.append(_content)\n else:\n info = INFO_SHIPPED\n _content.append(str(order.isShipped).upper())\n values.append(_content)\n\n info['startRowIndex'] = order.id + 1834\n info['endRowIndex'] = order.id + 1835\n\n if not order.isShipped and not order.isPartialShipped:\n info['red'], info['green'], info['blue'] = 1.0, 1.0, 1.0\n\n result = write_sheet(spreadsheet_service, 'kelly', info)\n result2 = write_cell(spreadsheet_service, 'kelly',\n range='Form Responses 2!M%i:M%i' % (order.id + 1835, order.id + 1835),\n majorDimension='ROWS',\n values=values)\n\n return HttpResponse('%s<br>%s' % (str(result), str(result2)))\n\n elif myDict['action'] and myDict['action'] == 'update':\n\n service_account_access(request)\n\n spreadsheet_service = sheet_authorization(request, use_service=True)\n\n response = ''\n\n for single_order in myDict['allData']:\n order = Orders.objects.get(pk=single_order['id'])\n affected = 0\n for col in single_order.keys():\n\n try:\n\n if isinstance(getattr(order, col), int) and not isinstance(getattr(order, col), bool):\n\n if getattr(order, col) != int(single_order[col]):\n setattr(order, col, int(single_order[col]))\n affected += 1\n else:\n continue\n\n elif isinstance(getattr(order, col), bool):\n\n if str(single_order[col]) == 'partial' and not order.isPartialShipped:\n order.isPartialShipped = True\n order.isShipped = False\n affected += 1\n else:\n current_status = strtobool(single_order[col]) \\\n if not isinstance(single_order[col], bool) \\\n else single_order[col]\n\n if getattr(order, col) != single_order[col]:\n setattr(order, col, current_status)\n order.isPartialShipped = False\n affected += 1\n else:\n continue\n\n elif isinstance(getattr(order, col), datetime.date):\n\n _strDate = str(int(single_order[col][:3]) + 1911) + single_order[col][3:]\n current_date = datetime.datetime.strptime(_strDate, '%Y. %m. %d')\n current_date = datetime.date(current_date.year, current_date.month, current_date.day)\n\n if getattr(order, col) != current_date:\n setattr(order, col, current_date)\n affected += 1\n else:\n continue\n\n else:\n\n if getattr(order, col) != single_order[col]:\n setattr(order, col, single_order[col])\n affected += 1\n else:\n continue\n\n except AttributeError:\n\n customer = order.customer\n\n if isinstance(getattr(customer, col), int) and not isinstance(getattr(customer, col), bool):\n if getattr(customer, col) != int(single_order[col]):\n setattr(customer, col, int(single_order[col]))\n customer.save()\n affected += 1\n else:\n continue\n else:\n if getattr(customer, col) != single_order[col]:\n setattr(customer, col, single_order[col])\n customer.save()\n affected += 1\n else:\n continue\n\n order.save()\n\n if affected > 0:\n\n info = None\n values = list()\n\n _content = [\n order.customer.customer_name,\n order.customer.customer_Facebook,\n order.bank,\n order.last_five,\n order.amount,\n datetime.datetime.strftime(order.transfer_date, '%m/%d/%Y'),\n order.zip_code,\n order.address,\n order.customer.customer_phone,\n order.order_content,\n order.order_model,\n order.isShipped\n ]\n\n if not order.isShipped and order.isPartialShipped:\n info = INFO_PARTIAL\n _content[-1] = 'Partial'\n values.append(_content)\n else:\n info = INFO_SHIPPED\n _content[-1] = (str(order.isShipped).upper())\n values.append(_content)\n\n info['startRowIndex'] = order.id + 1834\n info['endRowIndex'] = order.id + 1835\n\n if not order.isShipped and not order.isPartialShipped:\n info['red'], info['green'], info['blue'] = 1.0, 1.0, 1.0\n\n result = write_sheet(spreadsheet_service, 'kelly', info)\n result2 = write_cell(spreadsheet_service, 'kelly',\n range='Form Responses 2!B%i:M%i' % (order.id + 1835, order.id + 1835),\n majorDimension='ROWS',\n values=values)\n\n response += '%s<br>%s' % (str(result), str(result2))\n\n return HttpResponse(response + '<br>Success.')\n\n else:\n\n var = ''\n\n for _str in myDict.keys():\n var += '%s: %s' % (_str, string[_str])\n\n return HttpResponse(var)\n\n\ndef get_customer_data(customer_id):\n realName = Customer.objects.get(id=customer_id).customer_name\n faceBook = Customer.objects.get(id=customer_id).customer_Facebook\n phone = Customer.objects.get(id=customer_id).customer_phone\n\n return realName, faceBook, phone\n\n\n@csrf.csrf_exempt\ndef oauth2callback(request):\n scope = 'https://www.googleapis.com/auth/spreadsheets'\n flow = client.flow_from_clientsecrets(os.path.join(PROJECT_ROOT, 'client_secret.json'),\n scope=scope,\n redirect_uri='http://kelly8118.xyz:8080/kelly/oauth2callback/')\n\n if not request.GET.get('code', False):\n auth_uri = flow.step1_get_authorize_url()\n return HttpResponseRedirect(auth_uri)\n else:\n auth_code = request.GET['code']\n credentials = flow.step2_exchange(auth_code)\n request.session['credentials'] = credentials.to_json()\n return HttpResponseRedirect(reverse('kelly:show_data'))\n\n\n@csrf.csrf_exempt\ndef service_account_access(request):\n scopes = 'https://www.googleapis.com/auth/spreadsheets'\n\n credentials = ServiceAccountCredentials.from_json_keyfile_name(\n os.path.join(PROJECT_ROOT, 'My Project-d78da7467a28.json'),\n scopes=scopes)\n request.session['credentials'] = credentials.to_json()\n\n return HttpResponse(request.session['credentials'])\n\n\n@csrf.csrf_exempt\ndef write_data(request):\n spreadsheet_service = sheet_authorization(request)\n\n if not spreadsheet_service:\n return HttpResponseRedirect(reverse('kelly:oauth2callback'))\n\n result = get_sheet(spreadsheet_service, 'kelly',\n 'Form Responses 2!K1911:K1915', getValue=False)\n\n sheets = result.get('sheets', [])\n\n if not sheets:\n return HttpResponse('No data found.')\n else:\n output = read_bgColor(sheets, as_string=False, as_255=False)\n return HttpResponse(output)\n\n # if not request.session.get('credentials', False):\n # return HttpResponseRedirect(reverse('kelly:oauth2callback'))\n #\n # credentials = client.OAuth2Credentials.from_json(request.session['credentials'])\n #\n # if credentials.access_token_expired:\n # return HttpResponseRedirect(reverse('kelly:oauth2callback'))\n # else:\n # http_auth = credentials.authorize(httplib2.Http())\n # discoveryUrl = ('https://sheets.googleapis.com/$discovery/rest?'\n # 'version=v4')\n # spreadsheet_service = discovery.build('sheets', 'v4', http=http_auth,\n # discoveryServiceUrl=discoveryUrl)\n # spreadsheetId = '1EKRGi8EmPIO1yUNGuh7L7m9VrDP3YLDYcX6UspM1Lns'\n # rangeName = 'Form Responses 2!K1884:L1885'\n #\n # result = spreadsheet_service.spreadsheets().get(\n # spreadsheetId=spreadsheetId,\n # ranges=rangeName,\n # includeGridData=True).execute()\n # sheets = result.get('sheets', [])\n # output = None\n # if not sheets:\n # return HttpResponse('No data found.')\n # else:\n # # for _sheet in sheets:\n # # _data = _sheet['data'] # A list\n # # for i in _data:\n # # _rowData = i['rowData']\n # # for j in _rowData:\n # # _values = j['values']\n # # for _cellData in _values:\n # # _effectiveFormat = _cellData['effectiveFormat']\n # # _backgroundColor = _effectiveFormat['backgroundColor']\n # # for _key in _backgroundColor.keys():\n # # output += '%s: %s' % (_key, _backgroundColor[_key]) + '<br>'\n # # output += \"<p>\"\n # output = read_bgColor(sheets, as_string=True, as_255=True)\n # return HttpResponse(output)\n\n\n@csrf.csrf_exempt\ndef write_color(request):\n spreadsheet_service = sheet_authorization(request)\n\n if not spreadsheet_service:\n return HttpResponseRedirect(reverse('kelly:oauth2callback'))\n\n info = {\n 'action': 'changeBackgroundColor',\n 'sheetId': 309611827,\n 'startRowIndex': 1836,\n 'endRowIndex': 1837,\n 'startColumnIndex': 10,\n 'endColumnIndex': 11,\n 'red': 0.81568629,\n 'green': 0.87843138,\n 'blue': 0.89019608\n }\n\n result = write_sheet(spreadsheet_service, 'kelly', info)\n\n return HttpResponse(str(result))\n\n # if not request.session.get('credentials', False):\n # return HttpResponseRedirect(reverse('kelly:oauth2callback'))\n #\n # credentials = client.OAuth2Credentials.from_json(request.session['credentials'])\n #\n # if credentials.access_token_expired:\n # return HttpResponseRedirect(reverse('kelly:oauth2callback'))\n # else:\n # http_auth = credentials.authorize(httplib2.Http())\n # discoveryUrl = ('https://sheets.googleapis.com/$discovery/rest?'\n # 'version=v4')\n # spreadsheet_service = discovery.build('sheets', 'v4', http=http_auth,\n # discoveryServiceUrl=discoveryUrl)\n # spreadsheetId = '1EKRGi8EmPIO1yUNGuh7L7m9VrDP3YLDYcX6UspM1Lns'\n # rangeName = 'Form Responses 2!K1836:K1836'\n # body = {\n # 'requests': [\n # {\n # 'addConditionalFormatRule': {\n # 'rule': {\n # 'ranges': [\n # {\n # 'sheetId': 309611827,\n # 'startRowIndex': 1836,\n # 'endRowIndex': 1837,\n # 'startColumnIndex': 10,\n # 'endColumnIndex': 11\n # }\n # ],\n # 'booleanRule': {\n # 'condition': {\n # 'type': 'NOT_BLANK',\n # 'values': []\n # },\n # 'format': {\n # 'backgroundColor': {\n # 'red': 0.81568629,\n # 'green': 0.87843138,\n # 'blue': 0.89019608\n # }\n # }\n # }\n # },\n # 'index': 0\n # }\n # }\n # ]\n # }\n #\n # result = spreadsheet_service.spreadsheets()\\\n # .batchUpdate(spreadsheetId=spreadsheetId, body=body).execute()\n # return HttpResponse(str(result))\n\n\n@csrf.csrf_exempt\ndef show_data(request):\n if not check_oauth2(request):\n return redirect(reverse('kelly:oauth2callback'))\n\n spreadsheet_service = sheet_authorization(request)\n\n if not spreadsheet_service:\n return HttpResponseRedirect(reverse('kelly:oauth2callback'))\n\n result = get_sheet(spreadsheet_service, 'kelly', 'Form Responses 2!B1836:M3000')\n\n if not result:\n return HttpResponse('No data found.')\n else:\n output = update_database(result, spreadsheet_service)\n affected, partial, created, error_count = output.get('affected'), \\\n output.get('partial'), \\\n output.get('created'), \\\n output.get('error_count')\n myStr = 'Affected: %r, Created: %r, Partially shipped: %r, Error: %r' % \\\n (affected, created, partial, error_count)\n return HttpResponse(myStr)\n\n # if credentials.access_token_expired:\n # return HttpResponseRedirect(reverse('kelly:oauth2callback'))\n # else:\n # http_auth = credentials.authorize(httplib2.Http())\n # discoveryUrl = ('https://sheets.googleapis.com/$discovery/rest?'\n # 'version=v4')\n # spreadsheet_service = discovery.build('sheets', 'v4', http=http_auth,\n # discoveryServiceUrl=discoveryUrl)\n # spreadsheetId = '1EKRGi8EmPIO1yUNGuh7L7m9VrDP3YLDYcX6UspM1Lns'\n # rangeName = 'Form Responses 2!B1836:M3000'\n #\n # result = spreadsheet_service.spreadsheets().values().get(\n # spreadsheetId=spreadsheetId, range=rangeName).execute()\n # values = result.get('values', [])\n #\n # if not values:\n # return HttpResponse('No data found.')\n # else:\n # orders = Orders.objects.all()\n # index = 0\n # affected = 0\n # error_count = 0\n # created = 0\n # partial = 0\n # myOrder = None\n #\n # for row in values:\n # try:\n # myOrder = orders[index]\n # myOrder.bank = row[2]\n # myOrder.last_five = row[3]\n # myOrder.amount = row[4]\n # myOrder.transfer_date = datetime.datetime.strptime(row[5], '%m/%d/%Y')\n # myOrder.zip_code = row[6]\n # myOrder.address = row[7]\n # myOrder.order_content = row[9]\n # myOrder.order_model = row[10]\n # myOrder.order_belong = row[0]\n # if not find_customer(myOrder.order_belong):\n # customer = Customer()\n # customer.customer_name = myOrder.order_belong\n # customer.customer_Facebook = row[1]\n # customer.customer_phone = row[8]\n # customer.save()\n # myOrder.customer_id = Customer.objects.get(customer_name=myOrder.order_belong).id\n # else:\n # myOrder.customer_id = Customer.objects.get(customer_name=myOrder.order_belong).id\n # try:\n # myOrder.isShipped = strtobool(row[11])\n # except ValueError:\n # if row[11] == 'Partial':\n # myOrder.isShipped = False\n # myOrder.isPartialShipped = True\n # myOrder.save()\n # print('Partially shipped order: %s\\n' % (str(row)))\n # partial += 1\n # affected += 1\n # index += 1\n # continue\n # else:\n # print('Cannot fetch the data: IsShipped. Row: %s\\n' % (str(row)))\n # error_count += 1\n # index += 1\n # continue\n # myOrder.save()\n # affected += 1\n # index += 1\n # except IndexError:\n # if not row:\n # break\n # myDate = datetime.datetime.strptime(row[5], '%m/%d/%Y')\n # order = Orders()\n # order.bank = row[2]\n # order.last_five = row[3]\n # order.amount = row[4]\n # order.transfer_date = myDate\n # order.zip_code = row[6]\n # order.address = row[7]\n # order.order_content = row[9]\n # order.order_model = row[10]\n # order.order_belong = row[0]\n # if not find_customer(order.order_belong):\n # customer = Customer()\n # customer.customer_name = order.order_belong\n # customer.customer_Facebook = row[1]\n # customer.customer_phone = row[8]\n # customer.save()\n # order.customer_id = Customer.objects.get(customer_name=order.order_belong).id\n # else:\n # order.customer_id = Customer.objects.get(customer_name=order.order_belong).id\n # try:\n # order.isShipped = strtobool(row[11])\n # except ValueError:\n # if row[11] == 'Partial':\n # order.isShipped = False\n # order.isPartialShipped = True\n # order.save()\n # print('Partially shipped order: %s\\n' % (str(row)))\n # partial += 1\n # created += 1\n # index += 1\n # continue\n # else:\n # print('Cannot fetch the data: IsShipped. Row: %s\\n' % (str(row)))\n # error_count += 1\n # index += 1\n # continue\n # order.save()\n # created += 1\n # index += 1\n # continue\n #\n # myStr = 'Affected: %i, Created: %i, Partially shipped: %i, Error: %i' %\\\n # (affected, created, partial, error_count)\n # return HttpResponse(myStr)\n\n\n@csrf.csrf_exempt\ndef fetch(request):\n service_account_access(request)\n\n spreadsheet_service = sheet_authorization(request, use_service=True)\n\n result = get_sheet(spreadsheet_service, 'kelly', 'Form Responses 2!B1836:M3000')\n\n if not result:\n return HttpResponse('No data found.')\n else:\n output = update_database(result, spreadsheet_service)\n affected, partial, created, error_count = output.get('affected'), \\\n output.get('partial'), \\\n output.get('created'), \\\n output.get('error_count')\n myStr = 'Affected: %r, Created: %r, Partially shipped: %r, Error: %r' % \\\n (affected, created, partial, error_count)\n return HttpResponse(myStr)\n\n\ndef logout(request):\n auth_logout(request)\n return redirect(reverse('kelly:index'))\n\n\ndef check_oauth2(request):\n try:\n check = isinstance(client.OAuth2Credentials.from_json(request.session.get('credentials')),\n client.OAuth2Credentials)\n except ValueError:\n return False\n\n return check\n","sub_path":"kelly/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":29494,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"488336604","text":"import os\nimport queue\nfrom concurrent.futures import Future, Executor as FutureExecutor\n\nfrom pympipool import cancel_items_in_queue, RaisingThread\nfrom pysqa.executor.helper import (\n reload_previous_futures,\n find_executed_tasks,\n serialize_funct,\n write_to_file,\n)\n\n\nclass Executor(FutureExecutor):\n def __init__(self, cwd=None, queue_adapter=None, queue_adapter_kwargs=None):\n self._task_queue = queue.Queue()\n self._memory_dict = {}\n self._cache_directory = os.path.abspath(os.path.expanduser(cwd))\n self._queue_adapter = queue_adapter\n reload_previous_futures(\n future_queue=self._task_queue,\n future_dict=self._memory_dict,\n cache_directory=self._cache_directory,\n )\n command = (\n \"python -m pysqa.executor --cores \"\n + str(queue_adapter_kwargs[\"cores\"])\n + \" --path \"\n + str(self._cache_directory)\n )\n self._queue_id = self._queue_adapter.submit_job(\n working_directory=self._cache_directory,\n command=command,\n **queue_adapter_kwargs\n )\n self._process = RaisingThread(\n target=find_executed_tasks,\n kwargs={\n \"future_queue\": self._task_queue,\n \"cache_directory\": self._cache_directory,\n },\n )\n self._process.start()\n\n def submit(self, fn, *args, **kwargs):\n funct_dict = serialize_funct(fn, *args, **kwargs)\n key = list(funct_dict.keys())[0]\n if key not in self._memory_dict.keys():\n self._memory_dict[key] = Future()\n _ = write_to_file(\n funct_dict=funct_dict, state=\"in\", cache_directory=self._cache_directory\n )[0]\n self._task_queue.put({key: self._memory_dict[key]})\n return self._memory_dict[key]\n\n def shutdown(self, wait=True, *, cancel_futures=False):\n if cancel_futures:\n cancel_items_in_queue(que=self._task_queue)\n self._task_queue.put({\"shutdown\": True, \"wait\": wait})\n self._queue_adapter.delete_job(process_id=self._queue_id)\n self._process.join()\n","sub_path":"pysqa/executor/executor.py","file_name":"executor.py","file_ext":"py","file_size_in_byte":2183,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"368858235","text":"from __future__ import print_function\n\nimport tensorflow as tf\nfrom PhotoHandler import get_training_set_and_test_set\nfrom PhotoHandler import get_CNN_training_set_and_test_set\nfrom PhotoHandler import CNN_jpg_width\nfrom PhotoHandler import CNN_jpg_height\n\nimport sys\n\ndp1 = float(sys.argv[1])\ndp2 = float(sys.argv[2])\nprint(\"For Dropout_1=%.1f Dropout_2=%.1f\" % (dp1, dp2))\n\n# (x_train, y_train),(x_test, y_test) = get_training_set_and_test_set()\n(x_train, y_train),(x_test, y_test) = get_CNN_training_set_and_test_set()\nx_train, x_test = x_train / 255.0, x_test / 255.0\n\n# model = tf.keras.models.Sequential([\n# tf.keras.layers.Flatten(input_shape=(30000, )),\n# tf.keras.layers.Dense(512, activation=tf.nn.relu),\n# tf.keras.layers.Dropout(0.2),\n# tf.keras.layers.Dense(5, activation=tf.nn.softmax)\n# ])\n\n# CNN Model\nmodel = tf.keras.models.Sequential([\n tf.keras.layers.Conv2D(128, (3, 3), input_shape=(CNN_jpg_width, CNN_jpg_height, 3), activation=tf.nn.relu),\n tf.keras.layers.Conv2D(128, (3, 3), activation=tf.nn.relu),\n tf.keras.layers.MaxPool2D(),\n tf.keras.layers.Dropout(dp1),\n tf.keras.layers.Flatten(),\n tf.keras.layers.Dense(512, activation=tf.nn.relu),\n tf.keras.layers.Dropout(dp2),\n tf.keras.layers.Dense(5, activation=tf.nn.softmax)\n])\n\nmodel.compile(optimizer='adam',\n loss='sparse_categorical_crossentropy',\n metrics=['accuracy'])\n\nmodel.fit(x_train, y_train, epochs=1)\nmodel.evaluate(x_test, y_test)\nmodel.fit(x_train, y_train, epochs=1)\nmodel.evaluate(x_test, y_test)\nmodel.fit(x_train, y_train, epochs=1)\nmodel.evaluate(x_test, y_test)\nmodel.fit(x_train, y_train, epochs=1)\nmodel.evaluate(x_test, y_test) # ------------------------ BEST: dp1=0.6, dp2=0.6 and acc=0.6081 loss=1.0303\nmodel.fit(x_train, y_train, epochs=1)\nmodel.evaluate(x_test, y_test)","sub_path":"Version/3.0/KerasNetwork.py","file_name":"KerasNetwork.py","file_ext":"py","file_size_in_byte":1833,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"42258061","text":"from django import template\nfrom newItem.models import Image\n\nregister = template.Library()\n\n\ndef imageMini(productid):\n\tallImage = Image.objects.all()\n\tallImage = allImage.filter(numberProduct = productid)\n\timage = allImage[0]\n\treturn image.link\n\t\nregister.filter('imageMini', imageMini)","sub_path":"newtags/templatetags/imgTags.py","file_name":"imgTags.py","file_ext":"py","file_size_in_byte":288,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"163748789","text":"import tensorflow as tf\nfrom tensorflow import keras\nimport pandas as pd\nfrom time import time\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.model_selection import GridSearchCV\nfrom tensorflow.python.keras.wrappers.scikit_learn import KerasClassifier\nfrom tensorflow.python.keras.callbacks import TensorBoard\nfrom sklearn.metrics import classification_report\n\n# PARAMETERIZED DL NN\ndef parameterized_deep_net(loss=None,\n optimizer=None,\n metrics=None,\n batch_size=None,\n epochs=None,\n validation_split=None,\n steps_per_epoch=None,\n use_multiprocessing=True):\n model = keras.Sequential()\n model.add(keras.layers.Dense(4, activation=tf.nn.relu))\n model.add(keras.layers.Dense(4, activation=tf.nn.relu))\n model.add(keras.layers.Dense(2, activation=tf.nn.softmax))\n model.compile(optimizer=optimizer, loss=loss, metrics=metrics)\n return model\n\n# GATHER DATA\ntrain, val = train_test_split(\n pd.read_csv(\"titanic_imputed_1_hot_encd_norml.csv\", sep=\",\").\n query(expr='Survived > -1'),\n test_size=0.35)\n\nX_train = train[['Pclass', 'Age', 'SibSp', 'Parch', 'Fare', 'is_male', 'embark_C', 'embark_Q']].values\ny_train = train[['Survived']].values\n\nX_val = val[['Pclass', 'Age', 'SibSp', 'Parch', 'Fare', 'is_male', 'embark_C', 'embark_Q']].values\ny_val = val[['Survived']].values\ndel val, train\n\n\n# TEST MODEL METHOD\ndef test_parameterized_tf_nn():\n batch_size = 10\n epochs = 10\n validation_split = 0.3\n use_multiprocessing = True\n steps_per_epoch = 10\n tensorboard_log = 'tb_logs'\n tensorboard_name = '6. Grid Search NB'\n\n model = parameterized_deep_net(X_train, y_train, X_val, y_val,\n batch_size=batch_size,\n epochs=epochs,\n validation_split=validation_split,\n steps_per_epoch=steps_per_epoch,\n use_multiprocessing=use_multiprocessing,\n tensorboard_log=tensorboard_log,\n tensorboard_name=tensorboard_name)\n# test_parameterized_tf_nn()\n\n# GRID SEARCHING\nbatch_size = [5, 7]\nepochs = [5, 7]\nuse_multiprocessing = [True]\nsteps_per_epoch = [5, 7]\nmetrics = [['accuracy']]\noptimizer = ['adam']\nloss = ['sparse_categorical_crossentropy']\n\nparam_grid = dict(batch_size=batch_size,\n epochs=epochs,\n use_multiprocessing=use_multiprocessing,\n steps_per_epoch=steps_per_epoch,\n metrics=metrics,\n loss=loss,\n optimizer=optimizer)\n\nscoring = {'AUC': 'roc_auc', 'Accuracy': 'accuracy', 'Balanced Accuracy': 'balanced_accuracy'}\n\ntf_nn = KerasClassifier(build_fn=parameterized_deep_net, verbose=1)\nclf = GridSearchCV(estimator=tf_nn,\n param_grid=param_grid,\n n_jobs=10,\n cv=3,\n scoring=scoring,\n refit='AUC',\n return_train_score=True)\n# https://scikit-learn.org/stable//modules/model_evaluation.html\ngrid_result = clf.fit(X_train, y_train)\n\nprint(\"Best parameters set found on development set:\")\nprint()\nprint(clf.best_params_)\nprint()\nprint(\"Grid scores on development set:\")\nprint()\nmeans = clf.cv_results_['mean_test_Accuracy']\nstds = clf.cv_results_['std_test_Accuracy']\nfor mean, std, params in zip(means, stds, clf.cv_results_['params']):\n print(\"%0.3f (+/-%0.03f) for %r\"\n % (mean, std * 2, params))\nprint()\n\nprint(\"Detailed classification report:\")\nprint()\nprint(\"The model is trained on the full development set.\")\nprint(\"The scores are computed on the full evaluation set.\")\nprint()\ny_true, y_pred = y_val, clf.predict(X_val)\nprint(classification_report(y_true, y_pred))\nprint()\n\n# ######################################################################################################################\n# ######################################################################################################################\n# ######################################################################################################################\n\nMODEL_NAME = \"TITANIC-TF-MODEL-{}\".format(int(time.time()))\n\nbatch_size = 10\nepochs = 10\nvalidation_split = 0.3\nuse_multiprocessing = True\nsteps_per_epoch = 10\n\nmodel = parameterized_deep_net(batch_size=batch_size,\n epochs=epochs,\n validation_split=validation_split,\n steps_per_epoch=steps_per_epoch,\n use_multiprocessing=use_multiprocessing)\n\n# GRID SEARCH\nmodel = KerasClassifier(build_fn=parameterized_deep_net, verbose=0)\n\nparam_grid = dict(epochs=[1, 3, 5, 7])\ngrid = GridSearchCV(estimator=model, param_grid=param_grid, n_jobs=-1, cv=3, scoring='accuracy')\ngrid_result = grid.fit(X_train, y_train)\n\nexit(0)\n","sub_path":"grid_search_deep_net.py","file_name":"grid_search_deep_net.py","file_ext":"py","file_size_in_byte":5081,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"262036058","text":"import pandas\n\nfrom sklearn.cluster import KMeans\nfrom super_dash.signals import register_jsonschema\nfrom mine.algorithm.models import Scatter\n\n\nconfig_schema = {\n \"properties\": {\n \"n_clusters\": {\n \"type\": \"number\",\n \"minimum\": 1,\n },\n \"axis\": {\n \"type\": \"array\",\n \"items\": {\n \"type\": \"string\",\n }\n }\n },\n \"required\": [\"axis\"]\n}\n\nregister_jsonschema.send(sender=None, schema=config_schema,\n import_path='mine.algorithm.kmeans')\n\n\ndef entry(ds, cfg):\n ds = pandas.read_csv(ds)\n n_clusters = cfg.get('n_clusters')\n if n_clusters:\n kmeans = KMeans(n_clusters=n_clusters)\n else:\n kmeans = KMeans()\n labels = kmeans.fit(ds[cfg['axis']]).labels_\n\n models = []\n for i in range(kmeans.n_clusters):\n scatter = Scatter(cfg.get('name'))\n scatter.label = labels[i]\n models.append(scatter)\n for i_loc, series in ds[cfg['axis']].iterrows():\n models[labels[i_loc]].add(series.tolist())\n return models, kmeans.predict\n","sub_path":"mine/algorithm/kmeans.py","file_name":"kmeans.py","file_ext":"py","file_size_in_byte":1102,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"461592895","text":"# uncompyle6 version 3.7.4\n# Python bytecode 2.4 (62061)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build\\bdist.win32\\egg\\wsgiserialize\\xmlrpc.py\n# Compiled at: 2006-12-06 22:56:10\nfrom xmlrpclib import dumps\n__all__ = [\n 'WSGIXmlRpc', 'xmlrpc']\n\ndef xmlrpc(application, **kw):\n \"\"\"Decorator for XML-RPC serialization.\"\"\"\n return WsgiXmlRpc(application, **kw)\n\n\nclass WsgiXmlRpc(object):\n \"\"\"WSGI middleware for serializing simple Python objects to XML-RPC.\"\"\"\n __module__ = __name__\n\n def __init__(self, application, **kw):\n self.application = application\n self.response = kw.get('methodresponse')\n self.name = kw.get('methodname')\n self.encoding = kw.get('encoding')\n self.allownone = kw.get('allow_none', 0)\n\n def __call__(self, environ, start_response):\n return [\n dumps(tuple([self.application(environ, start_response)]), self.response, self.name, self.encoding, self.allownone)]","sub_path":"pycfiles/wsgiserialize-0.3-py2.4/xmlrpc.py","file_name":"xmlrpc.py","file_ext":"py","file_size_in_byte":1011,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"429195290","text":"\"\"\"Download treadmill app metrics given a pattern or exact app name.\n\nThe files are downloaded to the directory specified by the --outdir command\nline option.\n\"\"\"\n\nimport functools\nimport logging\nimport os\nimport urllib.request\nimport urllib.parse\nimport urllib.error\n\nimport click\n\nfrom treadmill import cli\nfrom treadmill import context\nfrom treadmill import fs\nfrom treadmill import restclient\nfrom treadmill import rrdutils\nfrom treadmill.websocket import client as wsc\n\n_LOGGER = logging.getLogger(__name__)\n\n# TODO: this list should be discoverable from the server rather than\n# hardcoded. GET /metrics/core should return this list.\n_SYSTEM_SERVICES = [\n # Total metrics for non-treadmill (system), core services and all apps.\n 'treadmill.system',\n 'treadmill.core',\n 'treadmill.apps',\n]\n\n\ndef _find_nodeinfo_endpoints(api=None):\n \"\"\"Return all the nodeinfo endpoints in the cell.\n\n The return value is a dict with host-endpoint assigments as key-value\n pairs.\n \"\"\"\n endpoints = _get_endpoints(api)\n return _endpoints_by_hosts(endpoints)\n\n\ndef _get_endpoints(api=None):\n \"\"\"Return all the nodeinfo endpoints for the given cell.\"\"\"\n apis = context.GLOBAL.state_api(api)\n\n url = '/endpoint/{}/tcp/nodeinfo'.format(urllib.parse.quote('root.*'))\n response = restclient.get(apis, url)\n\n endpoints = [{\n 'name': end['name'],\n 'proto': end['proto'],\n 'endpoint': end['endpoint'],\n 'hostport': '{0}:{1}'.format(end['host'], end['port'])\n } for end in response.json()]\n\n if not endpoints:\n cli.bad_exit(\"Nodeinfo API couldn't be found\")\n\n return endpoints\n\n\ndef _endpoints_by_hosts(endpoints):\n \"\"\"Return a dict consisting of the host-endpoint pairs as key-values.\"\"\"\n rv = {}\n for ep in endpoints:\n host, _ = ep['hostport'].split(':')\n rv[host] = ep\n\n return rv\n\n\ndef _get_endpoint_for_host(endpoints, host):\n \"\"\"Return the nodeinfo endpoint running on the host in parameter.\"\"\"\n try:\n rv = endpoints[host]\n except KeyError:\n cli.bad_exit('Nodeinfo endpoint not found on %s', host)\n\n return rv\n\n\ndef _instance_to_host_uniq(in_=None, out_=None, uniq=None):\n \"\"\"Update out_ so it contains 'instance: {host, uniq}' as key: value pairs.\n \"\"\"\n if 'event' not in in_ or not in_['event']:\n return True\n\n if 'uniqueid' not in in_['event'] or in_['event']['uniqueid'] != uniq:\n return True\n\n out_[in_['event']['instanceid']] = in_['event']['source']\n return True\n\n\ndef _find_uniq_instance(instance, uniq, ws_api=None):\n \"\"\"Find out where the given instance/uniq is/has been running.\"\"\"\n rv = {}\n message = {'topic': '/trace', 'filter': instance, 'snapshot': True}\n on_message = functools.partial(_instance_to_host_uniq, out_=rv, uniq=uniq)\n\n wsc.ws_loop(ws_api, message, True, on_message)\n\n return rv\n\n\ndef _instance_to_host(in_=None, out_=None):\n \"\"\"Update out_ so it contains 'instance: host' as key: value pairs.\"\"\"\n if 'host' not in in_:\n return True\n\n out_[in_['name']] = in_['host']\n return True\n\n\ndef _find_running_instance(app, ws_api=None):\n \"\"\"Find the instance name(s) and host(s) corresponding to the application.\n \"\"\"\n rv = {}\n message = {'topic': '/endpoints',\n 'filter': app,\n 'proto': 'tcp',\n 'endpoint': 'ssh',\n 'snapshot': True}\n\n on_message = functools.partial(_instance_to_host, out_=rv)\n\n wsc.ws_loop(ws_api, message, True, on_message)\n\n return rv\n\n\ndef _metrics_url(*name_parts):\n \"\"\"Return the url with which the application metrics can be retrieved.\"\"\"\n return '/metrics/{}'.format(urllib.parse.quote('/'.join(name_parts)))\n\n\ndef _rrdfile(outdir, *fname_parts):\n \"\"\"Return the full path of the rrd file where the metrics will be saved.\n \"\"\"\n return os.path.join(outdir, '-'.join(fname_parts) + '.rrd')\n\n\ndef _get_app_rsrc(instance, admin_api=None, cell_api=None):\n \"\"\"Return the application's reserved resources from the manifest.\"\"\"\n try:\n mf = restclient.get(context.GLOBAL.cell_api(cell_api),\n '/instance/%s' % urllib.quote(instance)).json()\n except restclient.NotFoundError:\n mf = restclient.get(context.GLOBAL.admin_api(admin_api),\n '/app/%s' % instance).json()\n\n return {rsrc: mf[rsrc] for rsrc in ('cpu', 'disk', 'memory')\n if rsrc in mf}\n\n\ndef _get_app_metrics(endpoint, instance, timeframe, uniq='running',\n outdir=None, cell_api=None):\n \"\"\"Retreives app metrics.\"\"\"\n fs.mkdir_safe(outdir)\n reserved_rsrc = _get_app_rsrc(instance, cell_api)\n\n api = 'http://{}'.format(endpoint['hostport'])\n _download_rrd(api, _metrics_url(instance, uniq),\n _rrdfile(outdir, instance, uniq), timeframe, reserved_rsrc)\n\n\ndef _get_server_metrics(endpoint, server, timeframe, services=None,\n outdir=None):\n \"\"\"Get core services metrics.\"\"\"\n fs.mkdir_safe(outdir)\n\n api = 'http://{}'.format(endpoint['hostport'])\n\n if not services:\n services = _SYSTEM_SERVICES\n\n # FIXME: give a default value of system limit\n # otherwise the command will crash\n for svc in services:\n _download_rrd(api, _metrics_url(svc), _rrdfile(outdir, server, svc),\n timeframe, {'cpu': '0%', 'disk': '0M', 'memory': '0M'})\n\n\ndef _download_rrd(nodeinfo_url, metrics_url, rrdfile, timeframe,\n reserved_rsrc=None):\n \"\"\"Get rrd file and store in output directory.\"\"\"\n _LOGGER.info('Download metrics from %s/%s', nodeinfo_url, metrics_url)\n try:\n resp = restclient.get(nodeinfo_url, metrics_url, stream=True)\n with open(rrdfile, 'w+b') as f:\n for chunk in resp.iter_content(chunk_size=128):\n f.write(chunk)\n\n rrdutils.gen_graph(rrdfile, timeframe, rrdutils.RRDTOOL,\n reserved_rsrc=reserved_rsrc)\n except restclient.NotFoundError as err:\n _LOGGER.error('%s', err)\n cli.bad_exit('Metrics not found: %s', err)\n except rrdutils.RRDToolNotFoundError:\n cli.bad_exit('The rrdtool utility cannot be found in the PATH')\n\n\n# Disable warning about redefined-builtin 'long' in the options\n# pylint: disable=W0622\ndef init():\n \"\"\"Top level command handler.\"\"\"\n\n ctx = {}\n\n @click.group()\n @click.option('--cell-api',\n envvar='TREADMILL_CELLAPI',\n help='Cell API url to use.',\n required=False)\n @click.option('--api',\n envvar='TREADMILL_STATEAPI',\n help='State API url to use.',\n required=False)\n @click.option('--cell',\n callback=cli.handle_context_opt,\n envvar='TREADMILL_CELL',\n expose_value=False,\n required=True)\n @click.option('--outdir',\n '-o',\n help='Output directory.',\n required=True,\n type=click.Path(exists=True))\n @click.option('--ws-api', help='Websocket API url to use.', required=False)\n def metrics(cell_api, api, outdir, ws_api):\n \"\"\"Retrieve node / app metrics.\"\"\"\n ctx['cell_api'] = cell_api\n ctx['nodeinf_eps'] = _find_nodeinfo_endpoints(api)\n ctx['outdir'] = outdir\n ctx['ws_api'] = ws_api\n\n @metrics.command()\n @cli.ON_REST_EXCEPTIONS\n @click.argument('app_pattern')\n @click.option('--long', is_flag=True, default=False,\n help='Metrics for longer timeframe.')\n def running(app_pattern, long):\n \"\"\"Get the metrics of running instances.\"\"\"\n instances = _find_running_instance(app_pattern, ctx['ws_api'])\n if not instances:\n cli.bad_exit('No running instance matched the pattern.')\n\n _LOGGER.debug('Found instance(s): %s', instances)\n\n timeframe = 'long' if long else 'short'\n for inst, host in instances.items():\n endpoint = _get_endpoint_for_host(ctx['nodeinf_eps'], host)\n _LOGGER.debug(\"getting metrics from endpoint %r\", endpoint)\n\n _get_app_metrics(endpoint, inst, timeframe, outdir=ctx['outdir'],\n cell_api=ctx['cell_api'])\n\n @metrics.command()\n @cli.ON_REST_EXCEPTIONS\n @click.argument('app')\n @click.option('--long', is_flag=True, default=False,\n help='Metrics for longer timeframe.')\n def app(app, long):\n \"\"\"Get the metrics of the application in params.\"\"\"\n instance, uniq = app.split('/')\n if uniq == 'running':\n instances = _find_running_instance(instance, ctx['ws_api'])\n else:\n instances = _find_uniq_instance(instance, uniq, ctx['ws_api'])\n\n if not instances:\n cli.bad_exit('No instance found with the application name.')\n\n _LOGGER.debug('Found instance(s): %s', instances)\n\n timeframe = 'long' if long else 'short'\n for inst, host in instances.items():\n endpoint = _get_endpoint_for_host(ctx['nodeinf_eps'], host)\n _LOGGER.debug(\"getting metrics from endpoint %r\", endpoint)\n\n _get_app_metrics(endpoint, inst, timeframe, uniq,\n outdir=ctx['outdir'], cell_api=ctx['cell_api'])\n\n @metrics.command()\n @cli.ON_REST_EXCEPTIONS\n @click.argument('servers', nargs=-1)\n @click.option('--services', type=cli.LIST, help='Subset of core services.')\n @click.option('--long', is_flag=True, default=False,\n help='Metrics for longer timeframe.')\n def sys(servers, services, long):\n \"\"\"Get the metrics of the server(s) in params.\"\"\"\n timeframe = 'long' if long else 'short'\n for server in servers:\n endpoint = _get_endpoint_for_host(ctx['nodeinf_eps'], server)\n _LOGGER.debug(\"getting metrics from endpoint %r\", endpoint)\n\n _get_server_metrics(endpoint, server, timeframe, services,\n ctx['outdir'])\n\n del running\n del app\n del sys\n\n return metrics\n","sub_path":"treadmill/cli/metrics.py","file_name":"metrics.py","file_ext":"py","file_size_in_byte":10146,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"401067928","text":"import os\nimport mysql.connector\nfrom dotenv import load_dotenv # for python-dotenv method\nload_dotenv() # for python-dotenv method\n\n\ndef connect_to_database():\n # os environments are defined in .env file\n return mysql.connector.connect(host=os.environ.get('HOST'),\n user=os.environ.get('USER'),\n password=os.environ.get('PASSWORD'),\n database=os.environ.get('DATABASE'))\n\n\ndef test_sql_create_color():\n test_val = (\"blue\", \"#123458\")\n\n database = connect_to_database()\n database_cursor = database.cursor()\n\n database_cursor.execute(\"INSERT INTO colors (color_name, hex_value) \\\n VALUES (%s, %s)\", test_val)\n\n assert database_cursor.rowcount == 1\n\n\ndef test_sql_get_all_colors():\n database1 = connect_to_database()\n database_cursor1 = database1.cursor()\n database_cursor1.execute(\"SELECT * FROM colors\")\n expected_rows = len(database_cursor1.fetchall())\n\n database2 = connect_to_database()\n database_cursor2 = database2.cursor()\n database_cursor2.execute(\"SELECT COUNT(*) FROM colors\")\n # comma after actual_rows to unpack tuple of database response\n actual_rows, = database_cursor2.fetchone()\n\n assert expected_rows == actual_rows\n","sub_path":"ai/digimad_backend_ai_io_controller/tests/test_colors.py","file_name":"test_colors.py","file_ext":"py","file_size_in_byte":1334,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"246104299","text":"import os\nfrom django.test import SimpleTestCase\nfrom django.test.utils import override_settings\nfrom corehq.apps.app_manager.tests import TestFileMixin\nfrom corehq.apps.userreports.models import CustomDataSourceConfiguration, CustomReportConfiguration\n\n\nclass TestCustomReportConfig(SimpleTestCase, TestFileMixin):\n\n file_path = ('data', 'custom_reports')\n root = os.path.dirname(__file__)\n\n def test_wrap(self):\n wrapped = CustomReportConfiguration.wrap(self.get_json('custom_report_config'))\n self.assertEqual([\"example\", \"dimagi\"], wrapped.domains)\n\n def test_get_all(self):\n with override_settings(CUSTOM_UCR_REPORTS=[self.get_path('custom_report_config', 'json')]):\n all = list(CustomReportConfiguration.all())\n self.assertEqual(2, len(all))\n example, dimagi = all\n self.assertEqual('example', example.domain)\n self.assertEqual('dimagi', dimagi.domain)\n for config in all:\n self.assertEqual('Custom Title', config.title)\n\n def test_production_config(self):\n for data_source in CustomDataSourceConfiguration.all():\n data_source.validate()\n","sub_path":"corehq/apps/userreports/tests/test_custom_reports.py","file_name":"test_custom_reports.py","file_ext":"py","file_size_in_byte":1179,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"182966520","text":"#opening the python3 shell\npython3\n\nimport subprocess\nsrc = \"<source-path>\" # replace <source-path> with the source directory\ndest = \"<destination-path>\" # replace <destination-path> with the destination directory\n\nsubprocess.call([\"rsync\", \"-arq\", src, dest])\n\n# Exit from the Python shell using exit().","sub_path":"debugging/part2/final_lab/subprocess_rsync_command_line.py","file_name":"subprocess_rsync_command_line.py","file_ext":"py","file_size_in_byte":304,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"469013488","text":"import random\r\n\r\ndef power(a,b):\r\n if b==1:\r\n return a;\r\n elif b==0:\r\n return 1;\r\n elif b%2==0:\r\n return (power(a,b/2)**2);\r\n else:\r\n return (a*power(a,b//2)**2);\r\n\r\nn=int(input(\"Δώσε τον όρο της ακολουθίας Fibonacci: \"));\r\n\r\nif n<=0:\r\n print(\"Σφάλμα! Ο αριθμός που έδωσες είναι μικρότερος του 0\");\r\nelif n==1:\r\n p=0;\r\n print(\"Ο αριθμός είναι πρώτος!\");\r\n\r\nelse:\r\n a=0;\r\n b=1;\r\n c=0;\r\n p=0;\r\n while c<n:\r\n p=a+b;\r\n a=b;\r\n b=p;\r\n c=c+1;\r\n print(p);\r\n \r\n flag=False;\r\n for i in range(20):\r\n x=random.choice(range(2,n));\r\n if (power(x,n))%n==x:\r\n flag=True;\r\n \r\n\r\n if flag==False:\r\n print(\"Ο αριθμός είναι πρώτος!\");\r\n else:\r\n print(\"Ο αριθμός δεν είναι πρώτος\");\r\n\r\n ","sub_path":"2.py","file_name":"2.py","file_ext":"py","file_size_in_byte":956,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"334211138","text":"__author__ = 'haussbrandt'\nimport random\n\nprint('Choose your weapon! You can choose rock, paper or scissors. You can type \"r\", \"p\" or \"s\" if you want.')\n\n\ndef game():\n players_choice = input('Rock, paper or scissors?') # Takes player's input.\n players_choice = players_choice.lower()\n if players_choice == 'r':\n players_choice = 'rock'\n elif players_choice == 's':\n players_choice = 'scissors'\n elif players_choice == 'p':\n players_choice = 'paper'\n else:\n print('You have chosen nothing.')\n print(\"You've chosen %s.\" % players_choice)\n\n computers_choice= random.choice(['rock', 'paper', 'scissors'])\n print(\"Computer's choice is %s.\" % computers_choice)\n\n if players_choice == 'rock':\n if computers_choice == 'rock':\n print('Draw.')\n if computers_choice == 'paper':\n print('You lose.')\n if computers_choice == 'scissors':\n print('You win.')\n\n if players_choice == 'paper':\n if computers_choice == 'rock':\n print('You win.')\n if computers_choice == 'paper':\n print('Draw.')\n if computers_choice == 'scissors':\n print('You lose.')\n\n if players_choice == 'scissors':\n if computers_choice == 'rock':\n print('You lose.')\n if computers_choice == 'paper':\n print('You win.')\n if computers_choice == 'scissors':\n print('Draw.')\n\n play_again = input('Do you want to play again?')\n play_again = play_again.lower()\n if play_again == 'yes' or play_again == 'y':\n game()\ngame()","sub_path":"RockPaperScissors.py","file_name":"RockPaperScissors.py","file_ext":"py","file_size_in_byte":1608,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"55934977","text":"#!/usr/bin/python2 \n# -*- coding: utf-8 -*-\nfrom pyspark import SparkContext \nfrom pyspark.sql import SparkSession\nimport pyspark.sql.functions as fn\nfrom pyspark.ml.feature import HashingTF,IDF,StringIndexer,RegexTokenizer,VectorAssembler,StopWordsRemover,IndexToString\nfrom pyspark.ml import Pipeline\nfrom pyspark.ml.classification import NaiveBayes\nfrom pyspark.ml.evaluation import MulticlassClassificationEvaluator\n\n\ndata_enh=\"hdfs://10.190.2.112/data/data_dump.txt\"\ntrain_set=\"hdfs://10.190.2.112/data/train_set.txt\"\nval_set=\"hdfs://10.190.2.112/data/val_set.txt\"\ntest_set=\"hdfs://10.190.2.112/data/test_set.txt\"\n\nspark = SparkSession.builder.master(\"spark://10.190.2.112:7077\").appName(\"15let_me_pass\").getOrCreate()\n\ntrain_setdf = spark.read.csv(train_set,header=None,encoding=\"utf-8\",inferSchema=True,sep=\"\\t\").drop_duplicates()#.replace(['E','K'],['1','0'],\"_c6\") \nval_setdf= spark.read.csv(val_set,header=None,encoding=\"utf-8\",inferSchema=True,sep=\"\\t\").drop_duplicates()#.replace(['E','K'],['1','0'],\"_c6\") \ntest_setdf= spark.read.csv(test_set,header=None,encoding=\"utf-8\",inferSchema=True,sep=\"\\t\").drop_duplicates()#.replace(['E','K'],['1','0'],\"_c6\") \n\n\n#预处理\ndef pre(dfx):\n #选取2个特征以及一列标签\n df=dfx.select(dfx[\"_c2\"].alias(\"FN\"),dfx['_c3'].alias(\"LN\"),dfx[\"_c6\"].alias(\"gender\"))\n #分词,分词结果应为纯粹的单��\n regexTokenizer1 = RegexTokenizer(inputCol=\"FN\", outputCol=\"FNgexr\", pattern=\"\\\\W+?\")\n regexTokenizer2= RegexTokenizer(inputCol=\"LN\", outputCol=\"LNgexr\", pattern=\"\\\\W+?\")\n #性别和city的相关性几乎为无\n # regexTokenizer3= RegexTokenizer(inputCol=\"Bcity\", outputCol=\"Bcitygex\", pattern=\"\\\\W\") \n #去除个别词\n wlist=[\"a\",\"b\",\"c\",\"d\",\"e\",\"f\",\"g\",\"h\",\"i\",\"j\",\"k\",\"l\",\"m\",\"n\",\"o\",\"p\",\"q\",\"r\",\"s\",\"t\",\"u\",\"v\",\"w\",\"x\",\"y\",\"z\"]\n sr1=StopWordsRemover(inputCol=\"FNgexr\",outputCol=\"FNgex\",stopWords=wlist)\n sr2=StopWordsRemover(inputCol=\"LNgexr\",outputCol=\"LNgex\",stopWords=wlist)\n # 将分词结果转换为向量,outputCol为feature1,feature2\n hashingTF1= HashingTF(inputCol=\"FNgex\", outputCol=\"rawFeature1\", numFeatures=10240)\n idf1 = IDF(inputCol=\"rawFeature1\", outputCol=\"feature1\")\n hashingTF2= HashingTF(inputCol=\"LNgex\", outputCol=\"rawFeature2\", numFeatures=10240)\n idf2 = IDF(inputCol=\"rawFeature2\", outputCol=\"feature2\")\n # cv3 = CountVectorizer(inputCol=\"Bcitygex\", outputCol=\"feature3\", vocabSize=100000, minDF=10) \n #拼接。outputCol为features\n VA=VectorAssembler(inputCols=[\"feature1\",\"feature2\"],outputCol=\"features\")\n #pipeline \n pipeline = Pipeline(stages=[regexTokenizer1,sr1,hashingTF1,idf1,regexTokenizer2,sr2,hashingTF2,idf2,VA])\n\n model = pipeline.fit(df)\n dataset = model.transform(df)\n return dataset\n\ntrainingData=pre(train_setdf)\nval_setData=pre(val_setdf)\ntest_setData=pre(test_setdf)\n\n# trainingData.show()\n# val_setData.show()\nfrom pyspark.ml.classification import NaiveBayes\n\n#训练\nindexer = StringIndexer(inputCol = \"gender\", outputCol = \"label\").setHandleInvalid(\"skip\")\nnb = NaiveBayes(smoothing=1,modelType=\"multinomial\",featuresCol=\"features\",labelCol=\"label\")\nconverter=IndexToString(inputCol=\"label\", outputCol=\"originlabel\")\n\n#pipeline \npipeline = Pipeline(stages=[indexer,nb,converter])\nmodel = pipeline.fit(trainingData)\npredictions = model.transform(test_setData)\n\nevaluator = MulticlassClassificationEvaluator().setLabelCol(\"label\").setPredictionCol(\"prediction\")\nlrAccuracy = evaluator.evaluate(predictions)\nprint(\"Test Accuracy = \" + str(lrAccuracy))\n\n","sub_path":"sparkh/spark14.py","file_name":"spark14.py","file_ext":"py","file_size_in_byte":3555,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"373571895","text":"import pandas as pd\nimport numpy as np\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.cross_validation import train_test_split\nfrom sklearn.cross_validation import cross_val_score\nfrom sklearn.metrics import confusion_matrix\n\n\ntraindata = pd.read_csv(\"train.csv\")\ntestdata = pd.read_csv(\"test.csv\")\nid_columns = testdata[\"Id\"]\ntraindata = traindata.drop([\"Id\"], axis=1)\ntestdata = testdata.drop([\"Id\"], axis=1)\ntestdata_array = testdata.as_matrix()\n\ntraindata = traindata.loc[traindata[\"Wilderness_Area1\"] == 0]\ntraindata = traindata.loc[traindata[\"Wilderness_Area4\"] == 0]\ntraindata_array = traindata.as_matrix()\n\n\ncolumns = traindata.columns\nsoil_columns = columns[columns.str.contains(\"Soil\")]\nwild_columns = columns[columns.str.contains(\"Wild\")]\nhill_columns = columns[columns.str.contains(\"Hill\")]\nhoriz_columns = columns[columns.str.contains(\"Horiz\")]\nverti_columns = columns[columns.str.contains(\"Verti\")]\nquant_columns = [\"Slope\"]\ncont_columns = np.array(columns.difference(soil_columns))\n\ntestcolumns = testdata.columns\nzero_columns = testcolumns[np.sum(testdata, axis=0) < 50]\nzero_columns_soil = zero_columns[zero_columns.str.contains(\"Soil\")]\n\nunbalanced_columns = columns[np.sum(traindata, axis=0) < 50]\n\ncolumns_drop = hill_columns | quant_columns | zero_columns_soil \\\n | wild_columns\n\ntrain, validation = train_test_split(traindata, test_size=0.2)\n\nx = train.drop([\"Cover_Type\"], axis=1)\nx = x.drop(columns_drop, axis=1)\n\ny = train[\"Cover_Type\"]\nx_array = x.as_matrix()\ny_array = y.as_matrix()\n\ntest_x = validation.drop([\"Cover_Type\"], axis=1)\ntest_x = test_x.drop(columns_drop, axis=1)\n\ntest_x = test_x.as_matrix()\ntest_y = np.array(validation[\"Cover_Type\"])\n\ntestdata_array = testdata.drop(columns_drop, axis=1).as_matrix()\n\ntemp = pd.DataFrame(train[\"Cover_Type\"], columns=[\"Cover_Type\"])\nsample_weight = np.zeros(len(train))\nsample_weight = temp\nn1 = len(train[train[\"Cover_Type\"] == 1])\nn2 = len(train[train[\"Cover_Type\"] == 2])\nn3 = len(train[train[\"Cover_Type\"] == 3])\nn5 = len(train[train[\"Cover_Type\"] == 5])\nn6 = len(train[train[\"Cover_Type\"] == 6])\nn7 = len(train[train[\"Cover_Type\"] == 7])\nn_min = min(n1, n2, n3, n5, n6, n7)\n\nsample_weight[sample_weight == 1] = n1 / n_min\nsample_weight[sample_weight == 2] = n2 / n_min\nsample_weight[sample_weight == 3] = n_min\nsample_weight[sample_weight == 5] = n5 / n_min\nsample_weight[sample_weight == 6] = n6 / n_min\nsample_weight[sample_weight == 7] = n7 / n_min\n\n# Random forest\nprint(\"Random forest\")\nrandomforest = RandomForestClassifier(n_estimators=20)\nrandomforest.fit(x_array, y_array, sample_weight=np.array(sample_weight).reshape(sample_weight.shape[0]))\nscores = cross_val_score(randomforest, test_x, test_y, cv=10)\nprint(\"score : \", round(np.mean(scores), 4) * 100)\nconf_matrix = confusion_matrix(test_y, randomforest.predict(test_x))\nprint(conf_matrix)\n\nprint(\"\\n\")\n\nresults = randomforest.predict(testdata_array)\ndf_result = pd.DataFrame(results, index=id_columns, columns=[\"Cover_Type\"])\ndf_result.to_csv(\"result.csv\")\nprint(x.columns)\nprint(\"\\n\")\n\n","sub_path":"datamining/kaggle/_random_focus.py","file_name":"_random_focus.py","file_ext":"py","file_size_in_byte":3064,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"486092772","text":"import json\nimport os\nfrom flask import Flask, render_template, jsonify, request, send_from_directory\nfrom pydub import AudioSegment\nimport uuid\n\napp = Flask(__name__, static_url_path='/static/')\n\n@app.route(\"/\")\ndef index():\n data = {\n 'graph': [\n {\n 'id': 0\n , 'audio': 'drum_kick'\n , 'speed': 30\n , 'beats': [0, 90, 180, 270] \n }\n , {\n 'id': 1\n , 'audio': 'snare'\n , 'speed': 30\n , 'beats': [45, 120, 150, 225, 300, 330] \n }\n , {\n 'id': 2\n , 'audio': 'budha'\n , 'speed': 30\n , 'beats': [270] \n }\n , {\n 'id': 3\n , 'audio': 'sound'\n , 'speed': 30\n , 'beats': [] \n }\n ]\n }\n\n return render_template(\"index.html\", data=data)\n\n\n@app.route(\"/create\", methods = ['POST'])\ndef create():\n graph = request.json\n path = './static/audio/{0}.mp3'\n \n duration = 5 * 1000 # 5 secs\n mashup = AudioSegment.silent(duration=duration)\n\n for n in graph:\n sound = AudioSegment.from_mp3(path.format(n['audio']))\n for beat in n['beats']:\n beat = (beat + 90) % 360\n\n if 0 <= beat < 5:\n beat = 360\n\n start = (360-beat) / 360 * duration\n mashup = mashup.overlay(sound, position=start)\n\n hash = uuid.uuid4().hex\n out_f = open(path.format('/tmp/' + hash), 'wb')\n mashup.export(out_f, format='mp3')\n\n return jsonify({'url': 'static/audio/tmp/' + hash + '.mp3'})\n\n@app.route('/static/<path:path>')\ndef serve_static(path):\n return send_from_directory('static/', path)\n\nif __name__ == \"__main__\":\n app.run(debug=True)\n","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":1947,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"119592974","text":"# -*- coding: utf-8 -*-\r\n'''\r\nCreated on 20.09.2019\r\n\r\n@author: yu03\r\n'''\r\nfrom pipython import GCSDevice, datarectools, pitools, gcscommands\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\nimport time\r\nimport datetime\r\nfrom File_name_define import PI_name, name\r\n\r\nnow = datetime.datetime.now()\r\n\r\ndef Export_Data(file_name, header, out_str):\r\n print('Writing Data')\r\n with open(file_name,'w') as fid: ######################################################################################\r\n fid.writelines(header)\r\n fid.writelines(out_str)\r\n print('Finish Writing')\r\n return\r\n\r\nCONTROLLERNAME = 'E-712'\r\nSTAGES = None # connect stages to axes\r\nREFMODE = None # reference the connected stages\r\n\r\n'''\r\n Parameters for data recorder\r\n'''\r\nNUMVALUES = 500 # number of data sets to record as integer\r\n# NUMVALUES = 600\r\nRECRATE = 250 # number of recordings per second, i.e. in Hz\r\n\r\n'''\r\n Parameters for Wave Generator\r\n'''\r\nNUMPOINTS = 30000 # number of points for one sine period as integer\r\nSTARTPOS = 0.0 # start position of the circular motion as float for both axes\r\nAMPLITUDE = 1 # amplitude of the circular motion as float for both axes\r\n# AMPLITUDE = 200\r\n# if AMPLITUDE >= 50:\r\n# print('AMPLITUDE TOO LARGE')\r\n# exit()\r\nOFFSET = 0\r\nNUMCYLES = 1 # number of cycles for wave generator output\r\n# TABLERATE = 50 # duration of a wave table point in multiples of servo cycle times as integer\r\nTABLERATE = 1\r\n\r\n'''\r\n Moving Axis define\r\n'''\r\n### 1=x,2=y,3=z_rot,4=z,5=x_rot,6=y_rot\r\n\r\nmoving_axis = 2\r\n\r\nwavegens = (1, 2, 3, 4, 5, 6)\r\nwavetables = (1, 2, 3, 4, 5, 6)\r\n\r\nwith GCSDevice(CONTROLLERNAME) as pidevice:\r\n \r\n '''\r\n Initialization\r\n '''\r\n pidevice.InterfaceSetupDlg()\r\n print('connected: {}'.format(pidevice.qIDN().strip()))\r\n print('initialize connected stages...')\r\n pitools.startup(pidevice, STAGES, REFMODE)\r\n IDN = pidevice.qIDN()\r\n print('IDN: ', IDN)\r\n print('Servo Status: ', pidevice.qSVO())\r\n pidevice.WGO(wavegens, mode=[0]*len(wavegens))\r\n '''\r\n Auto-Zero\r\n '''\r\n# pidevice.ATZ({1:0, 2:0, 4:0})\r\n# time.sleep(5)\r\n '''\r\n Turn on control loop\r\n '''\r\n pidevice.SVO({'1':1,'2':1,'3':1,'4':1,'5':1,'6':1})\r\n# print('Servo Status: ', pidevice.qSVO())\r\n '''\r\n Data Recording Configuration\r\n '''\r\n drec = datarectools.Datarecorder(pidevice)\r\n drec.numvalues = NUMVALUES\r\n drec.samplefreq = RECRATE\r\n print('data recorder rate: {:.2f} Hz'.format(drec.samplefreq))\r\n drec.options = (datarectools.RecordOptions.ACTUAL_POSITION_2)\r\n# drec.sources = ('2', '3', '5') ### 2=y=lenth 3=rot_z=hor_angle, 5=rot_x=ver_angle\r\n# drec.trigsources = datarectools.TriggerSources.POSITION_CHANGING_COMMAND_1\r\n drec.trigsources = datarectools.TriggerSources.TRIGGER_IMMEDIATELY_4\r\n# drec.arm()\r\n print('Data recorder TriggerSource: ', pidevice.qDRT())\r\n \r\n print('Sampling Freq. = ', drec.samplefreq)\r\n# pidevice.RTR(60)\r\n# print('Record Table Rate: ', pidevice.qRTR())\r\n \r\n pidevice.DRC(tables=1, sources='2', options=2)\r\n pidevice.DRC(tables=2, sources='3', options=2)\r\n pidevice.DRC(tables=3, sources='5', options=2)\r\n print('Data recorder configuration: ', pidevice.qDRC())\r\n \r\n \r\n \r\n# pidevice.DRT(tables=0, sources=2, values='0')\r\n# print('Data recorder TriggerSource: ', pidevice.qDRT())\r\n \r\n '''\r\n Wave Generator Configuration\r\n '''\r\n# print('Wave Generator Num. ', pidevice.qTWG())\r\n# Servo_update_time = pidevice.qSPA(items=1, params=0x0E000200)[1][234881536]### 0x0E000200\r\n# print('Servo update time: /s', Servo_update_time)\r\n \r\n# pidevice.WAV_LIN(table=wavetables[moving_axis-1], firstpoint=1, numpoints=NUMPOINTS, append='X',\r\n# speedupdown=NUMPOINTS//10, amplitude=AMPLITUDE, offset=OFFSET, seglength=NUMPOINTS)\r\n pidevice.WAV_SIN_P(table=wavetables[1], firstpoint=1, numpoints=NUMPOINTS, append='X',\r\n center=NUMPOINTS/2, amplitude=AMPLITUDE, offset=STARTPOS, seglength=NUMPOINTS)\r\n# pidevice.WAV`_RAMP(table=wavetables[moving_axis-1], firstpoint=1, numpoints=NUMPOINTS, append='X', center=NUMPOINTS/2, \r\n# speedupdown=NUMPOINTS//10, amplitude=45, offset=0, seglength=NUMPOINTS)\r\n pidevice.WSL(wavegens, wavetables)\r\n pidevice.WGC(wavegens, [NUMCYLES]*len(wavegens))\r\n pidevice.WTR(0, tablerates=TABLERATE, interpol=1)\r\n \r\n '''\r\n Trigger Configuration\r\n '''\r\n pidevice.TWC()\r\n# for i in range(NUMPOINTS//6+1): ### 50Hz~12 for TABLERATE 25\r\n for i in range(NUMPOINTS//300+1): ### 50Hz~60 for TABLERATE 5\r\n# pidevice.TWS(lines=2, points=1+6*i, switches=1) ### 50Hz~12 for TABLERATE 25\r\n pidevice.TWS(lines=2, points=1+300*i, switches=1) ### 50Hz~60 for TABLERATE 5\r\n \r\n pidevice.CTO(lines=2, params=1, values=0.1)\r\n# pidevice.CTO(lines=2, params=2, values=2)\r\n pidevice.CTO(lines=2, params=3, values=9)\r\n# Trig_conf = pidevice.qCTO()[2] ### Y_axis=2\r\n# Trig_step = Trig_conf[1]\r\n# Trig_line = Trig_conf[2]\r\n# Trig_mode = Trig_conf[3]\r\n# # print(Trig_conf)\r\n# print('Trigger step = ', float(Trig_step))\r\n# print('Trigger line = ', Trig_line)\r\n# print('Trigger mode = ', Trig_mode)\r\n \r\n# print('Data recorder options: ', pidevice.qHDR())\r\n \r\n\r\n# pitools.waitonready(pidevice)\r\n \r\n# Table_rate = pidevice.qSPA(items=1, params=0x13000109)[1][318767369] ###0x13000109\r\n# print(Table_rate)\r\n# print(pidevice.qWTR(wavegens=1))\r\n pitools.waitontarget(pidevice, '%s'%moving_axis)\r\n print('Servo Status: ', pidevice.qSVO())\r\n \r\n '''\r\n Notice the Axis No.2!!!!\r\n '''\r\n pidevice.WGO(wavegens[moving_axis-1], mode=[1])\r\n while any(list(pidevice.IsGeneratorRunning(wavegens[moving_axis-1]).values())):\r\n print ('.')\r\n time.sleep(1.0)\r\n print('done')\r\n pidevice.WGO(wavegens[moving_axis-1], mode=[0])\r\n\r\n# time.sleep(2.0)\r\n# pidevice.WGO(wavegens=1, mode=0)\r\n \r\n '''\r\n Set Target Relative To Current Position\r\n '''\r\n# pidevice.MVR('2', 0.1) ### y = 2\r\n \r\n '''\r\n Get Target Position\r\n '''\r\n# target_position = pidevice.qMOV() \r\n# target_x, target_y, target_z_r, target_z, target_x_r, target_y_r = pidevice.qMOV()['1'],pidevice.qMOV()['2'],pidevice.qMOV()['3'],pidevice.qMOV()['4'],pidevice.qMOV()['5'],pidevice.qMOV()['6']\r\n# print(target_x, target_y, target_z_r, target_z, target_x_r, target_y_r)\r\n \r\n '''\r\n Get Real Position\r\n '''\r\n# pos = pidevice.qPOS()\r\n# pos_x, pos_y, pos_z_r, pos_z, pos_x_r, pos_y_r = pidevice.qMOV()['1'],pidevice.qMOV()['2'],pidevice.qMOV()['3'],pidevice.qMOV()['4'],pidevice.qMOV()['5'],pidevice.qMOV()['6']\r\n# print(pos_x, pos_y, pos_z_r, pos_z, pos_x_r, pos_y_r)\r\n \r\n \r\n \r\n\r\n '''\r\n Data Recording\r\n '''\r\n \r\n# header, data = drec.getdata()\r\n# y_pos, z_rot, x_rot = data[0], data[1], data[2]\r\n \r\n# samp_time = NUMVALUES/RECRATE\r\n# n_data = NUMVALUES\r\n# print('Sampling Rate = ', RECRATE)\r\n# print('Data length = ', n_data)\r\n# print('Time = ', samp_time)\r\n \r\n\r\n header, data = pidevice.qDRR(tables=[1,2,3], offset=1, numvalues=NUMVALUES)\r\n y_pos, z_rot, x_rot = data[0], data[1], data[2]\r\n# header, data = datarectools.Datarecorder(pidevice).read(offset=1, numvalues=NUMVALUES)\r\n# print('Num. of recorded points: ', pidevice.qDRL())\r\n samp_time = NUMVALUES/RECRATE\r\n n_data = NUMVALUES\r\n \r\n# print(header)\r\n# print((data))\r\n# # y_pos, z_rot, x_rot = data[0], data[1], data[2]\r\n# y_pos = data[0]\r\n \r\n# print('Sampling Rate = ', RECRATE)\r\n# print('Data length = ', n_data)\r\n# print('Time = ', samp_time)\r\n# print(len(y_pos))\r\n\r\n\r\n\r\n''' \r\n 保存文件 Exporting TXT\r\n'''\r\nheader = ['%s\\n' %(name+'_PI'),\r\n 'Local current time : %s\\n' %now.strftime(\"%Y-%m-%d %H:%M:%S\"),\r\n 'Fs = %e (Hz)\\n' %RECRATE,##########################################################################################################\r\n 'Record time: %e (s)\\n' %samp_time,############################################################################################\r\n 'Frame Number = %i\\n' %NUMVALUES,############################################################################################\r\n 'Channel_1: Position /um \\n',############################################################################################\r\n 'Channel_2: Hor_Angle /urad \\n',############################################################################################\r\n 'Channel_3: ver_Angle /urad \\n',############################################################################################\r\n 'Channel_4: xxx \\n',############################################################################################\r\n '-------------------------------------------------\\n',\r\n ]\r\nout_str = ['%f, %f, %f\\n' %(y_pos[i], z_rot[i], x_rot[i]) for i in range(NUMVALUES)]\r\n\r\n'''\r\n Output\r\n'''\r\nExport_Data(PI_name, header, out_str)\r\nprint('TXT file saved')\r\n\r\nt = np.linspace(0, samp_time, num=n_data)\r\n\r\nplt.figure(1)\r\nplt.gcf().set_size_inches(18,9)\r\n\r\nplt.subplot(3,1,1)\r\nplt.plot(y_pos, color='blue', label='length')\r\nplt.title('Length')\r\nplt.xlabel('Time /s')\r\nplt.ylabel('Position /um')\r\nplt.grid(which='both', axis='both')\r\n\r\nplt.subplot(3,1,2)\r\nplt.plot(t, z_rot, color='red', label='Hor_Angle')\r\nplt.grid(which='both', axis='both')\r\nplt.xlabel('Time /s')\r\nplt.ylabel('Angle /urad')\r\nplt.title('Hor_Angle')\r\n\r\nplt.subplot(3,1,3)\r\nplt.plot(t, x_rot, color='black', label='Ver_Angle')\r\nplt.grid(which='both', axis='both')\r\nplt.title('Ver_Angle')\r\nplt.xlabel('Time /s')\r\nplt.ylabel('Angle /urad')\r\n\r\nfigManager = plt.get_current_fig_manager()\r\nfigManager.window.showMaximized()\r\nplt.tight_layout()\r\n\r\nplt.show()\r\n","sub_path":"Test_trigger_record.py","file_name":"Test_trigger_record.py","file_ext":"py","file_size_in_byte":9937,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"430125985","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Feb 13 15:42:24 2021\n\n@author: Henao\n\"\"\"\n\ndef calcular_BMI(peso_en_libras: float, estatura_en_pulgadas: float)-> float:\n '''\n Algorithm to calculate the body mass from your weight in pounds and your height in inches.\n Parameters\n ----------\n peso_en_libras : float\n Weight in pounds\n estatura_en_pulgadas : float\n Height in inches\n\n\n Returns\n ----------\n float\n \n BMI: your body mass index \n \n '''\n peso = peso_en_libras * 0.45\n altura = estatura_en_pulgadas * 0.025\n BMI = round(float(peso / (altura**2)),2)\n return BMI\n \n \npeso_en_libras = float(input(\"Enter your weight in pounds : \"))\nestatura_en_pulgadas = float(input(\"Enter your height in inches: \"))\nindice = calcular_BMI(peso_en_libras , estatura_en_pulgadas)\nprint(\"su imc es: \", indice)\n#print(\"Your body mass index is: \" , calcular_BMI(peso_en_libras, estatura_en_pulgadas))","sub_path":"calcular_BMI.py","file_name":"calcular_BMI.py","file_ext":"py","file_size_in_byte":951,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"231250675","text":"# 80%\n\n# 0. 사용할 패키지 불러오기\nimport numpy as np\nfrom keras.models import Sequential\nfrom keras.layers import Dense, Flatten, Dropout, Activation\nfrom keras.layers.convolutional import Conv2D\nfrom keras.layers.convolutional import MaxPooling2D\nfrom keras.preprocessing.image import ImageDataGenerator, array_to_img, img_to_array, load_img\nfrom keras import backend as K\n\nimg_w, img_h = 150, 150\ntrain_data_dir = './dataset/training_set'\ntest_data_dir = './dataset/test_set'\nnb_train_samples = 100\nnb_test_samples = 50\nepochs = 25\nbatch_size = 16\n\nif K.image_data_format() == 'channels_first':\n input_shape = (3, img_w, img_h)\nelse:\n input_shape = (img_w, img_h, 3)\n\n# 1. dataset 생성\n# 변화를 줘서 부풀리기.\ntrain_datagen = ImageDataGenerator(rescale=1./255,\n # rotation_range=15,\n # width_shift_range=0.1,\n # height_shift_range=0.1,\n shear_range=0.5,\n zoom_range=[0.8, 2.0],\n horizontal_flip=True,\n # vertical_flip=True,\n fill_mode='nearest')\n\n# 훈련용 generator 생성\ntrain_generator = train_datagen.flow_from_directory(\n train_data_dir, # img 경로\n target_size=(img_w, img_h), # 패치 이미지 크기\n batch_size=batch_size, # 배치 크기\n class_mode='categorical') # categorical/binary/sparse/None\n\ntest_datagen = ImageDataGenerator(rescale=1./255)\n# 검증용 generator 생성\ntest_generator = test_datagen.flow_from_directory(\n test_data_dir,\n target_size=(img_w, img_h),\n batch_size=batch_size,\n class_mode='categorical')\n\n\n\n# 2. 모델 구성하기\nmodel = Sequential()\nmodel.add(Conv2D(32, kernel_size=(3, 3), input_shape=input_shape, activation='relu'))\n# model.add(Conv2D(64, (3, 3), activation='relu'))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\n\nmodel.add(Conv2D(32, (3, 3), activation='relu'))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\n\nmodel.add(Conv2D(64, (3, 3), activation='relu'))\nmodel.add(MaxPooling2D(pool_size=(2, 2)))\n\nmodel.add(Flatten())\nmodel.add(Dense(64, activation='relu'))\n# model.add(Dense(256, activation='relu'))\nmodel.add(Dropout(0.5))\nmodel.add(Dense(3, activation='sigmoid'))\n\n# 3. 모델 학습과정 설정하기\nmodel.compile(loss='categorical_crossentropy', optimizer='adam', metrics=['accuracy'])\n# model.compile(loss='binary_crossentropy', optimizer='rmsprop', metrics=['accuracy'])\n\n\n\n\n\n# 4. 모델 학습시키기\nmodel.fit_generator(train_generator, steps_per_epoch=nb_train_samples, epochs=epochs, validation_data=test_generator, validation_steps=nb_test_samples)\nmodel.save_weights('test_model.h5')\n\n# 5. 모델 평가하기\nprint(\"-- Evaluate(정확도) --\")\nscores = model.evaluate_generator(test_generator, steps=5)\nprint(\"%s: %.2f%%\" %(model.metrics_names[1], scores[1]*100))\n#\n# # 6. 모델 저장하기\n# from keras.models import load_model\n# model.save('testModel.h5')\n","sub_path":"cnn/makeModel.py","file_name":"makeModel.py","file_ext":"py","file_size_in_byte":3085,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"41700144","text":"#!/usr/bin/env python3\nimport argparse\nimport os\nimport shutil\nimport signal\nimport subprocess\n\n# Absolute path to Gluon JavaFX directory\nSYSTEM_INIT_BIN = \"/usr/sbin/init\"\nGLUON_JAVAFX_PATH = \"/opt/javafx-sdk\"\n\n\n# Helper method to split JVM properties specified as -Dkey=value\ndef jvm_property(data):\n parts = tuple(str(data).split('=', 1))\n return parts if len(parts) == 2 else (parts[0], '')\n\n\n# Parse known arguments and preserve others\nparser = argparse.ArgumentParser(description='Gluon JavaFX Kiosk Launcher', allow_abbrev=False)\nparser.add_argument('--add-modules', default='')\nparser.add_argument('-p', '--module-path', default='')\nparser.add_argument('-D', default=[], action='append', type=jvm_property, dest='properties')\nargs, unknown_args = parser.parse_known_args()\n\n# Patch '--module-path' option\nmodule_path = list(filter(None, args.module_path.split(':')))\nmodule_path.insert(0, GLUON_JAVAFX_PATH + '/lib')\n\n# Patch '--add-modules' option\nadd_modules = list(filter(None, args.add_modules.split(',')))\nadd_modules.insert(0, 'javafx.controls')\n\n# Patch generic properties\nproperties = dict(filter(None, args.properties))\nproperties.setdefault('glass.platform', 'Monocle')\nproperties.setdefault('monocle.platform', 'EGL')\nproperties.setdefault('monocle.platform.traceConfig', 'false')\nproperties.setdefault('monocle.egl.lib', GLUON_JAVAFX_PATH + '/lib/libgluon_drm.so')\nproperties.setdefault('egl.displayid', '/dev/dri/card0')\nproperties.setdefault('javafx.verbose', 'false')\nproperties.setdefault('prism.verbose', 'false')\n\n# Patch 'java.library.path' property\njava_library_path = list(filter(None, properties.get('java.library.path', '').split(':')))\njava_library_path.insert(0, GLUON_JAVAFX_PATH + '/lib')\nproperties['java.library.path'] = ':'.join(java_library_path)\n\n# Patch environment variables\njvm_env = os.environ.copy()\njvm_env['ENABLE_GLUON_COMMERCIAL_EXTENSIONS'] = 'true'\n\n# Build final list of JVM arguments\njvm_args = [\n '--module-path', ':'.join(module_path),\n '--add-modules', ','.join(add_modules),\n]\njvm_args.extend(['-D' + key + '=' + value for key, value in properties.items()])\njvm_args.extend(unknown_args)\n\n# Search for absolute path of JVM\njvm_path = shutil.which('java')\nif jvm_path is None:\n parser.error(\"Unable to find 'java' binary in current PATH\")\n\n# Ensure we are running as root\nif os.geteuid() != 0:\n parser.error(\"Unable to execute 'java-kiosk' without running as root\")\n\n# Run application in kiosk mode\ntry:\n # Ignore Ctrl+C for python process to ensure completion\n signal.signal(signal.SIGINT, lambda signum, frame: None)\n\n # Switch to runlevel 3 to stop X11\n subprocess.run([SYSTEM_INIT_BIN, '3'])\n\n # Execute JVM with patched options\n subprocess.run([jvm_path] + jvm_args, env=jvm_env)\nexcept KeyboardInterrupt:\n # Silently ignore KeyboardInterrupt, we expect the user to sometimes abort the script\n pass\nfinally:\n # Switch back to runlevel 5 to start X11\n subprocess.run([SYSTEM_INIT_BIN, '5'])\n","sub_path":"image/resources/java/java-kiosk.py","file_name":"java-kiosk.py","file_ext":"py","file_size_in_byte":2999,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"549269187","text":"import nltk\nfrom nltk.corpus import ConllChunkCorpusReader\n\nfrom nltk.corpus.reader.tagged import TaggedCorpusReader\nroot = '/usr/local/share/nltk_data/corpora/MASC-for-NE/'\nmasc_for_ne = TaggedCorpusReader(root,'.*', '_')\n\nsents = masc_for_ne.tagged_sents()\nne_sents = [nltk.ne_chunk(sent) for sent in sents]\n\nroot = \"/usr/local/share/nltk_data/corpora/masc_conll/\"\ngold_corpus = ConllChunkCorpusReader(root,r\".*\\.conll\", chunk_types=(\"DATE\",\"PERSON\",\"ORGANIZATION\",\"LOCATION\"))\ngold_sents = gold_corpus.chunked_sents()\n\n","sub_path":"classes/vassar/Linguistics_HW_4/Prob3/init_masc_ne.py","file_name":"init_masc_ne.py","file_ext":"py","file_size_in_byte":522,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"353570048","text":"from django.shortcuts import render\n\nfrom website.models.order_model import Order\nfrom website.models.product_order_model import ProductOrder\nfrom website.models.product_model import Product\nfrom django.contrib.auth.models import User\nfrom django.http import HttpResponseForbidden\nfrom django.http import HttpResponseRedirect\nfrom django.urls import reverse\n\n\ndef order_detail(request, order_id):\n\t\"\"\"\n This function is invoked to display the details of a user's order.\n\n ---Arguments---\n request: the full HTTP request object\n order_id(integer): the id of the order\n\n ---GET---\n Renders order_detail.html\n\n ---Context---\n 'order'(instance): the order instance\n 'orderproducts'(list): a list of the products on the order \n 'total'(integer): the total cost of an order\n\n Author: Blaise Roberts\n \"\"\"\n\n\ttemplate_name = 'order_detail.html'\n\torder = Order.objects.get(pk=order_id)\n\n\tif request.user == order.user:\n\t\t# Get seller object\n\t\tline_items = ProductOrder.objects.filter(order=order_id).values_list(\n\t\t\t'product_id').distinct()\n\t\tproduct_list = list()\n\t\ttotal = int()\n\t\tfor x in line_items:\n\t\t\tproduct = Product.objects.filter(pk=x[0])\n\t\t\tproduct_count = ProductOrder.objects.filter(product_id=x[0], \n\t\t\t\torder=order_id).count()\n\t\t\tsubtotal = product[0].price * product_count\n\t\t\ttotal += subtotal\n\t\t\tproduct_list.append((product, product_count, subtotal))\n\t\treturn render(request, template_name, {'order': order, \"orderproducts\":\n\t\t\tproduct_list, \"total\":total})\n\telse:\n\t\treturn HttpResponseForbidden('''<h1>Not your order, bruh!</h1>\n\t\t\t<img src=\"/website/static/other.jpg\">''')\n\ndef delete_product_from_order(request, product_id, order_id):\n\n\t\"\"\"\n This function is invoked to delete a product from a user's order.\n\n ---Arguments---\n request: the full HTTP request object\n product_id(integer): the id of the product\n order_id(integer): the id of the order\n\n ---Return---\n Returns HttpResponseRedirect to order_detail\n\n Author: Jeremy Bakker and Jessica Younker\n \"\"\"\n\n\torder = Order.objects.get(pk=order_id, user=request.user)\n\tif request.user == order.user:\n\t\tProductOrder.objects.filter(product_id=product_id, \n\t\t\torder_id=order_id).delete()\n\t\treturn HttpResponseRedirect(reverse('website:order_detail', \n\t\t\targs=[order.id]))\n\telse:\n\t\treturn HttpResponseForbidden('''<h1>Not your order, bruh!</h1>\n\t\t\t<img src=\"/website/static/other.jpg\">''')","sub_path":"website/views/order_detail_view.py","file_name":"order_detail_view.py","file_ext":"py","file_size_in_byte":2426,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"549207238","text":"import matplotlib.pyplot as plt\r\n\r\ndef func1():\r\n\tfd = open(\"build_tvp_time.log\", \"r\")\r\n\t# wfd = open(\"build_tvp_time.ana.log\", \"w\")\r\n\tn_lines = 0;\r\n\tx_r = []\r\n\ty_r = []\r\n\tfor line in fd:\r\n\t\tn_lines += 1\r\n\t\tif(n_lines > 1 and n_lines < 10246 and (n_lines - 2) % 3 == 0):\r\n\t\t\ta = line.find('size: [') + len('size: [')\r\n\t\t\tb = line.find(']', a)\r\n\t\t\tsub1 = (line[a:b])\r\n\t\t\tc = line.find('time: [') + len('time: [')\r\n\t\t\td = line.find(' s', c)\r\n\t\t\tsub2 = (line[c:d])\r\n\t\t\t# buf = \"%5d %f\\n\" % (int(sub1), float(sub2))\r\n\t\t\t# buf = str(sub1) + \" \" + str(sub2) + '\\n'\r\n\t\t\t# wfd.write(buf);\r\n\r\n\t\t\tx_r.append(int(sub1))\r\n\t\t\ty_r.append(float(sub2))\r\n\r\n\r\n\tfd.close()\r\n\t# wfd.flush()\r\n\t# wfd.close()\r\n\r\n\tfig = plt.figure()\r\n\tax1 = fig.add_subplot(111)\r\n\tax1.set_title('tvp build time (NBA dataset)')\r\n\tplt.xlabel('number of nodes associated with keyword')\r\n\tplt.ylabel(\"build time (second)\")\r\n\tss1 = ax1.scatter(x_r, y_r, c = 'r', marker = 'x')\r\n\tplt.legend([ss1], ['one keyword'], loc='upper left')\r\n\tplt.show()\r\n\r\n\r\n\r\nif __name__ == '__main__':\r\n\tfunc1()\r\n","sub_path":"nba_demo/build_tvp_time.py","file_name":"build_tvp_time.py","file_ext":"py","file_size_in_byte":1050,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"348849518","text":"import serial\nfrom struct import *\nfrom misoclib.tools.litescope.host.driver.reg import *\n\ndef write_b(uart, data):\n\tuart.write(pack('B',data))\n\nclass LiteScopeUARTDriver:\n\tcmds = {\n\t\t\"write\"\t: 0x01,\n\t\t\"read\"\t: 0x02\n\t}\n\tdef __init__(self, port, baudrate=115200, addrmap=None, busword=8, debug=False):\n\t\tself.port = port\n\t\tself.baudrate = str(baudrate)\n\t\tself.debug = debug\n\t\tself.uart = serial.Serial(port, baudrate, timeout=0.25)\n\t\tself.regs = build_map(addrmap, busword, self.read, self.write)\n\n\tdef open(self):\n\t\tself.uart.flushOutput()\n\t\tself.uart.close()\n\t\tself.uart.open()\n\t\tself.uart.flushInput()\n\t\ttry:\n\t\t\tself.regs.uart2wb_sel.write(1)\n\t\texcept:\n\t\t\tpass\n\n\tdef close(self):\n\t\ttry:\n\t\t\tself.regs.uart2wb_sel.write(0)\n\t\texcept:\n\t\t\tpass\n\t\tself.uart.flushOutput()\n\t\tself.uart.close()\n\n\tdef read(self, addr, burst_length=None, repeats=None):\n\t\tdatas = []\n\t\tdef to_int(v):\n\t\t\treturn 1 if v is None else v\n\t\tfor i in range(to_int(repeats)):\n\t\t\tself.uart.flushInput()\n\t\t\twrite_b(self.uart, self.cmds[\"read\"])\n\t\t\twrite_b(self.uart, burst_length)\n\t\t\twrite_b(self.uart, (addr//4 & 0xff000000) >> 24)\n\t\t\twrite_b(self.uart, (addr//4 & 0x00ff0000) >> 16)\n\t\t\twrite_b(self.uart, (addr//4 & 0x0000ff00) >> 8)\n\t\t\twrite_b(self.uart, (addr//4 & 0x000000ff))\n\t\t\tfor j in range(to_int(burst_length)):\n\t\t\t\tdata = 0\n\t\t\t\tfor k in range(4):\n\t\t\t\t\tdata = data << 8\n\t\t\t\t\tdata |= ord(self.uart.read())\n\t\t\t\tif self.debug:\n\t\t\t\t\tprint(\"RD %08X @ %08X\" %(data, (addr+j)*4))\n\t\t\t\tdatas.append(data)\n\t\treturn datas\n\n\tdef write(self, addr, data):\n\t\tif isinstance(data, list):\n\t\t\tburst_length = len(data)\n\t\telse:\n\t\t\tburst_length = 1\n\t\twrite_b(self.uart, self.cmds[\"write\"])\n\t\twrite_b(self.uart, burst_length)\n\t\twrite_b(self.uart, (addr//4 & 0xff000000) >> 24)\n\t\twrite_b(self.uart, (addr//4 & 0x00ff0000) >> 16)\n\t\twrite_b(self.uart, (addr//4 & 0x0000ff00) >> 8)\n\t\twrite_b(self.uart, (addr//4 & 0x000000ff))\n\t\tif isinstance(data, list):\n\t\t\tfor i in range(len(data)):\n\t\t\t\tdat = data[i]\n\t\t\t\tfor j in range(4):\n\t\t\t\t\twrite_b(self.uart, (dat & 0xff000000) >> 24)\n\t\t\t\t\tdat = dat << 8\n\t\t\t\tif self.debug:\n\t\t\t\t\tprint(\"WR %08X @ %08X\" %(data[i], (addr + i)*4))\n\t\telse:\n\t\t\tdat = data\n\t\t\tfor j in range(4):\n\t\t\t\twrite_b(self.uart, (dat & 0xff000000) >> 24)\n\t\t\t\tdat = dat << 8\n\t\t\tif self.debug:\n\t\t\t\tprint(\"WR %08X @ %08X\" %(data, (addr * 4)))\n","sub_path":"misoclib/tools/litescope/host/driver/uart.py","file_name":"uart.py","file_ext":"py","file_size_in_byte":2295,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"206291049","text":"import ctypes\nimport numpy as np\nimport numpy.ctypeslib as npct\n\n\"\"\"\n The mapping between the C and Python type interfaces for Tipsy.\n This is the lowest level of the Tipsy-Python interface. It is\n not intended for general use.' \n\"\"\"\n\n_native_float32_dtype = np.dtype('=f')\n_array_1d_float32 = npct.ndpointer(dtype=_native_float32_dtype, ndim=1, flags=('C','O','W','A'))\n_array_2d_float32 = npct.ndpointer(dtype=_native_float32_dtype, ndim=2, flags=('C','O','W','A'))\n\ndef _convert_array(x, name):\n if x is None:\n raise ValueError(name + ' cannot be None')\n if not isinstance(x, np.ndarray):\n raise TypeError(name + ' is not a numpy array')\n return np.require(x, dtype=_native_float32_dtype, requirements=['C_CONTIGUOUS', 'ALIGNED', 'WRITEABLE', 'OWNDATA', 'ENSUREARRAY'])\n\ndef _make_array(size, ndims=1, zero=False):\n if ndims == 1:\n return np.empty(size, dtype=_native_float32_dtype) if not zero else np.zeros(size, dtype=_native_float32_dtype)\n if ndims == 2:\n return np.empty((size, 3), dtype=_native_float32_dtype) if not zero else np.zeros((size, 3), dtype=_native_float32_dtype)\n raise ValueError(\"tipsy only supports 1d and 2d arrays\")\n\ndef _pretty_print(obj, attrs):\n repr = []\n for k in attrs:\n a = getattr(obj, k)\n repr.append('{0:10s}: {1:s}'.format(k, str(a.shape) if isinstance(a, np.ndarray) else str(a)))\n return '\\n'.join(repr)\n\nclass header():\n class struct(ctypes.Structure):\n _fields_ = [\n ('time' , ctypes.c_double),\n ('nbodies', ctypes.c_uint),\n ('ndim' , ctypes.c_int),\n ('ngas' , ctypes.c_uint),\n ('ndark' , ctypes.c_uint),\n ('nstar' , ctypes.c_uint)\n ]\n\n @classmethod\n def from_external(cls, hdr):\n self = cls()\n self.time = hdr.time\n self.nbodies = hdr.nbodies\n self.ndim = hdr.ndim\n self.ngas = hdr.ngas\n self.ndark = hdr.ndark\n self.nstar = hdr.nstar\n return self\n\n def __init__(self):\n self.c_data = header.struct()\n\n def __str__(self):\n return _pretty_print(self, ['time','nbodies','ngas','ndark','nstar'])\n \n @property\n def time(self):\n return self.c_data.time\n @time.setter\n def time(self, rhs):\n self.c_data.time = rhs\n @property\n def nbodies(self):\n return self.c_data.nbodies\n @nbodies.setter\n def nbodies(self, rhs):\n self.c_data.nbodies = rhs\n @property\n def ngas(self):\n return self.c_data.ngas\n @ngas.setter\n def ngas(self, rhs):\n self.c_data.ngas = rhs\n @property\n def ndark(self):\n return self.c_data.ndark\n @ndark.setter\n def ndark(self, rhs):\n self.c_data.ndark = rhs\n @property\n def nstar(self):\n return self.c_data.nstar\n @nstar.setter\n def nstar(self, rhs):\n self.c_data.nstar = rhs\n\n @classmethod\n def from_external(cls, time, ngas, ndark, nstar):\n self = cls()\n self.time = float(time)\n self.nbodies = int(ngas) + int(ndark) + int(nstar)\n self.ndim = 3\n self.ngas = int(ngas)\n self.ndark = int(ndark)\n self.nstar = int(nstar)\n self.c_data = header.struct.from_external(self)\n return self\n\nclass gas_data():\n class struct(ctypes.Structure):\n _fields_ = [\n ('mass' , _array_1d_float32),\n ('pos' , _array_2d_float32),\n ('vel' , _array_2d_float32),\n ('rho' , _array_1d_float32),\n ('temp' , _array_1d_float32),\n ('hsmooth', _array_1d_float32),\n ('metals' , _array_1d_float32),\n ('phi' , _array_1d_float32),\n ('size' , ctypes.c_size_t)\n ]\n \n def __init__(self):\n super().__init__()\n \n @classmethod\n def from_external(cls, other):\n self = cls()\n self.mass = other.mass.ctypes.data_as(_array_1d_float32)\n self.pos = other.pos.ctypes.data_as(_array_2d_float32)\n self.vel = other.vel.ctypes.data_as(_array_2d_float32)\n self.rho = other.rho.ctypes.data_as(_array_1d_float32)\n self.temp = other.temp.ctypes.data_as(_array_1d_float32)\n self.metals = other.metals.ctypes.data_as(_array_1d_float32)\n self.hsmooth = other.hsmooth.ctypes.data_as(_array_1d_float32)\n self.phi = other.phi.ctypes.data_as(_array_1d_float32)\n self.size = other.size\n return self\n\n def __init__(self):\n self.size = 0\n self.c_data = None\n \n def __str__(self):\n if self.c_data is not None:\n return _pretty_print(self, [k[0] for k in self.c_data._fields_])\n\n @classmethod\n def from_size(cls, size):\n self = cls()\n self.mass = _make_array(size)\n self.pos = _make_array(size, ndims=2)\n self.vel = _make_array(size, ndims=2)\n self.rho = _make_array(size)\n self.temp = _make_array(size)\n self.metals = _make_array(size)\n self.hsmooth = _make_array(size)\n self.phi = _make_array(size)\n self.size = size\n self.c_data = gas_data.struct.from_external(self)\n return self\n\n @classmethod\n def from_external(cls, mass, pos, vel, rho, temp, hsmooth, metals, phi, size):\n self = cls()\n self.mass = _convert_array(mass, 'mass')\n self.pos = _convert_array(pos, 'pos')\n self.vel = _convert_array(vel, 'vel')\n self.rho = _convert_array(rho, 'rho')\n self.temp = _convert_array(temp, 'temp')\n self.hsmooth = _convert_array(hsmooth, 'hsmooth')\n self.metals = _convert_array(metals, 'metals')\n self.phi = _convert_array(phi, 'phi')\n self.size = size\n self.c_data = gas_data.struct.from_external(self)\n return self\n\nclass dark_data():\n class struct(ctypes.Structure):\n _fields_ = [\n ('mass', _array_1d_float32),\n ('pos' , _array_2d_float32),\n ('vel' , _array_2d_float32),\n ('soft', _array_1d_float32),\n ('phi' , _array_1d_float32),\n ('size', ctypes.c_size_t)\n ]\n \n def __init__(self):\n super().__init__()\n \n @classmethod\n def from_external(cls, other):\n self = cls()\n self.mass = other.mass.ctypes.data_as(_array_1d_float32)\n self.pos = other.pos.ctypes.data_as(_array_2d_float32)\n self.vel = other.vel.ctypes.data_as(_array_2d_float32)\n self.phi = other.phi.ctypes.data_as(_array_1d_float32)\n self.soft = other.soft.ctypes.data_as(_array_1d_float32)\n self.size = other.size\n return self\n\n def __init__(self):\n self.size = 0\n self.c_data = None\n \n def __str__(self):\n if self.c_data is not None:\n return _pretty_print(self, [k[0] for k in self.c_data._fields_])\n \n @classmethod\n def from_size(cls, size):\n self = cls()\n self.mass = _make_array(size)\n self.pos = _make_array(size, ndims=2)\n self.vel = _make_array(size, ndims=2)\n self.phi = _make_array(size)\n self.soft = _make_array(size)\n self.size = size\n self.c_data = dark_data.struct.from_external(self)\n return self\n \n @classmethod\n def from_external(cls, mass, pos, vel, soft, phi, size):\n self = cls()\n self.mass = _convert_array(mass, 'mass')\n self.pos = _convert_array(pos, 'pos')\n self.vel = _convert_array(vel, 'vel')\n self.phi = _convert_array(phi, 'phi')\n \n if soft is not None and np.isscalar(soft):\n self.soft = _make_array(size, zero=True)\n self.soft += np.asscalar(np.array(soft, dtype=_native_float32_dtype))\n else:\n self.soft = _convert_array(soft, 'soft')\n \n self.size = size\n self.c_data = dark_data.struct.from_external(self)\n return self\n\nclass star_data():\n class struct(ctypes.Structure):\n _fields_ = [\n ('mass' , _array_1d_float32),\n ('pos' , _array_2d_float32),\n ('vel' , _array_2d_float32),\n ('metals', _array_1d_float32),\n ('tform' , _array_1d_float32),\n ('soft' , _array_1d_float32),\n ('phi' , _array_1d_float32),\n ('size' , ctypes.c_size_t)\n ]\n \n def __init__(self):\n super().__init__()\n\n @classmethod\n def from_external(cls, other):\n self = cls()\n self.mass = other.mass.ctypes.data_as(_array_1d_float32)\n self.pos = other.pos.ctypes.data_as(_array_2d_float32)\n self.vel = other.vel.ctypes.data_as(_array_2d_float32)\n self.metals = other.metals.ctypes.data_as(_array_1d_float32)\n self.tform = other.tform.ctypes.data_as(_array_1d_float32)\n self.phi = other.phi.ctypes.data_as(_array_1d_float32)\n self.soft = other.soft.ctypes.data_as(_array_1d_float32)\n self.size = other.size\n return self\n\n def __init__(self):\n self.size = 0\n self.c_data = None\n \n def __str__(self):\n if self.c_data is not None:\n return _pretty_print(self, [k[0] for k in self.c_data._fields_])\n \n @classmethod\n def from_size(cls, size):\n self = cls()\n self.mass = _make_array(size)\n self.pos = _make_array(size, ndims=2)\n self.vel = _make_array(size, ndims=2)\n self.metals = _make_array(size)\n self.tform = _make_array(size)\n self.phi = _make_array(size)\n self.soft = _make_array(size)\n self.size = size\n self.c_data = star_data.struct.from_external(self)\n return self\n\n @classmethod\n def from_external(cls, mass, pos, vel, metals, tform, soft, phi, size):\n self = cls()\n self.mass = _convert_array(mass, 'mass')\n self.pos = _convert_array(pos, 'pos')\n self.vel = _convert_array(vel, 'vel')\n self.metals = _convert_array(metals, 'metals')\n self.tform = _convert_array(tform, 'tform')\n self.phi = _convert_array(phi, 'phi')\n \n if soft is not None and np.isscalar(soft):\n self.soft = _make_array(size, zero=True)\n self.soft += np.asscalar(np.array(soft, dtype=_native_float32_dtype))\n else:\n self.soft = _convert_array(soft, 'soft')\n \n self.size = size\n self.c_data = star_data.struct.from_external(self)\n return self\n\n","sub_path":"tipsy_c.py","file_name":"tipsy_c.py","file_ext":"py","file_size_in_byte":10639,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"456153486","text":"import logging\n\n\nclass numList:\n \"\"\"This is a numList class.\n\n Attributes:\n :maxMin (tuple): tuple of the Max and Min values in the list\n\n :max_diff (list): list of the highest diff between 2 adj values in list\n\n :list_add (int): sum of all the values in the list\n\n \"\"\"\n\n def __init__(self, myList=[]):\n self.list = myList\n self.maxMin = None\n self.max_diff = None\n self.list_add = None\n self.max_Min()\n self.find_diff()\n self.find_sum()\n\n def max_Min(self):\n \"\"\"\n Finds the max and min in a list of positive values and returns a tuple\n\n :param inputList: Is a list of positive values\n :returns: Tuple of the max and min values\n :raises ImportError: If numpy is not installed in the env\n :raises ValueError: If there are values less than 0\n :raises TypeError: If the inputList is not an actual list\n \"\"\"\n\n inputList = self.list\n logging.basicConfig(filename='log.txt', level=logging.DEBUG)\n\n try:\n import numpy\n except ImportError:\n logging.error(\"missing a module!\")\n raise ImportError(\"missing a module!\")\n for i in inputList:\n if i < 0:\n logging.warning(\"Negative value detected\")\n raise ValueError('Negative value detected')\n if not isinstance(inputList, list):\n logging.warning('Input is not a list')\n raise TypeError('Input is not a list')\n myMin = min(inputList)\n myMax = max(inputList)\n logging.debug(inputList)\n logging.debug('Min value: %s', myMin)\n logging.debug('Max value: %s', myMax)\n maxMinTuple = (myMin, myMax)\n logging.info(maxMinTuple)\n self.maxMin = maxMinTuple\n\n def find_diff(self):\n \"\"\"\n Finds maximum difference between two adjacent numbers in a list\n\n :param my_list: Is a list of numbers\n :returns: Largest difference between two adjacent numbers\n :raises ValueError: If my_list has 0 or 1 elements\n :raises ImportError: If numpy is not installed in environment\n :raises TypeError: If element in my_list is not an int, float, complex\n \"\"\"\n\n my_list = self.list\n logging.basicConfig(filename='log.txt', level=logging.DEBUG)\n\n logging.info('Finding max difference between adjacent values in list')\n logging.debug('Printing %s', str(my_list))\n n = 0\n if len(my_list) < 2:\n logging.warning('Not enough values to calculate difference')\n raise ValueError('List too small, no difference to compare!')\n for i in range(len(my_list)-1):\n if(isinstance(my_list[i], (int, float, complex)) and\n isinstance(my_list[i+1], (int, float, complex))):\n diff = abs(my_list[i+1] - my_list[i])\n if diff > n:\n n = diff\n else:\n raise TypeError('List elements must be int, float, complex!')\n logging.debug('Returns %s', str(n))\n self.max_diff = n\n\n def find_sum(self):\n \"\"\"\n Adds a lenist of numbers\n\n :param list_var: Is a list of numbers (int, float, complex)\n :returns: Addition of values in list\n :raises ValueError: If list_var is empty\n :raises ImportError: If numpy or numbers not installed in environment\n :raises TypeError: If element in list_var is not an int, float,complex\n \"\"\"\n list_var = self.list\n try:\n import logging\n except ImportError:\n logging.warning('ImportError Logging')\n raise ImportError('Module Logging not found.')\n logging.basicConfig(filename='log.txt', level=logging.DEBUG)\n try:\n import numpy as np\n except ImportError:\n logging.warning('ImportError Numpy')\n raise ImportError('Module Numpy not found.')\n if len(list_var) == 0:\n raise ValueError('Input list is empty')\n try:\n import numbers\n except ImportError:\n logging.warning('ImportError Numbers')\n raise ImportError('Module Numbers not found.')\n if not isinstance(list_var, list):\n logging.warning('Input is not a list')\n for x in list_var:\n if isinstance(x, (int, float, complex)):\n continue\n else:\n logging.warning('List elements must be int, float or complex')\n raise TypeError('List elements must be int, float, or complex')\n logging.debug(list_var)\n value = np.sum(list_var)\n logging.info(value)\n self.list_add = value\n","sub_path":"numList.py","file_name":"numList.py","file_ext":"py","file_size_in_byte":4742,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"44463210","text":"\"\"\"This module includes the specification of the model.\"\"\"\nimport numpy as np\n\nfrom trempy.shared.shared_auxiliary import dist_class_attributes\nfrom trempy.shared.shared_auxiliary import print_init_dict\nfrom trempy.config_trempy import PREFERENCE_PARAMETERS\nfrom trempy.paras.clsParas import ParasCls\nfrom trempy.shared.clsBase import BaseCls\nfrom trempy.read.read import read\n\n\nclass ModelCls(BaseCls):\n \"\"\"This class manages all issues about the model specification.\"\"\"\n def __init__(self, fname):\n\n init_dict = read(fname)\n\n # We first tackle the more complex issue of parameter management.\n paras_obj = ParasCls(init_dict)\n\n self.attr = dict()\n\n # Parameters\n self.attr['paras_obj'] = paras_obj\n\n # Information\n upper = []\n upper += [init_dict['UNIATTRIBUTE SELF']['max']]\n upper += [init_dict['UNIATTRIBUTE OTHER']['max']]\n self.attr['upper'] = upper\n\n # Marginal utility functions\n marginals = []\n marginals += [init_dict['UNIATTRIBUTE SELF']['marginal']]\n marginals += [init_dict['UNIATTRIBUTE OTHER']['marginal']]\n self.attr['marginals'] = marginals\n\n # Cutoffs\n self.attr['cutoffs'] = init_dict['CUTOFFS']\n\n # Simulation\n self.attr['sim_agents'] = init_dict['SIMULATION']['agents']\n self.attr['sim_seed'] = init_dict['SIMULATION']['seed']\n self.attr['sim_file'] = init_dict['SIMULATION']['file']\n\n # Estimation\n self.attr['est_detailed'] = init_dict['ESTIMATION']['detailed']\n self.attr['optimizer'] = init_dict['ESTIMATION']['optimizer']\n\n self.attr['est_agents'] = init_dict['ESTIMATION']['agents']\n self.attr['num_skip'] = init_dict['ESTIMATION']['skip']\n self.attr['est_file'] = init_dict['ESTIMATION']['file']\n self.attr['maxfun'] = init_dict['ESTIMATION']['maxfun']\n self.attr['start'] = init_dict['ESTIMATION']['start']\n\n # Optimizer options\n self.attr['opt_options'] = dict()\n\n self.attr['opt_options']['SCIPY-BFGS'] = dict()\n self.attr['opt_options']['SCIPY-BFGS']['gtol'] = init_dict['SCIPY-BFGS']['gtol']\n self.attr['opt_options']['SCIPY-BFGS']['eps'] = init_dict['SCIPY-BFGS']['eps']\n\n self.attr['opt_options']['SCIPY-POWELL'] = dict()\n self.attr['opt_options']['SCIPY-POWELL']['xtol'] = init_dict['SCIPY-POWELL']['xtol']\n self.attr['opt_options']['SCIPY-POWELL']['ftol'] = init_dict['SCIPY-POWELL']['ftol']\n\n para_objs = paras_obj.get_attr('para_objs')\n\n questions = []\n for para_obj in para_objs:\n label = para_obj.get_attr('label')\n if label in PREFERENCE_PARAMETERS:\n continue\n\n questions += [label]\n\n self.attr['questions'] = sorted(questions)\n self.attr['num_questions'] = len(questions)\n\n # We now need to check the integrity of the class instance.\n self._check_integrity()\n\n def update(self, perspective, which, values):\n \"\"\"This method updates the estimation parameters.\"\"\"\n # Distribute class attributes\n paras_obj = self.attr['paras_obj']\n\n paras_obj.set_values(perspective, which, values)\n\n def write_out(self, fname):\n \"\"\"This method creates a initialization dictionary of the current class instance.\"\"\"\n init_dict = dict()\n\n labels = []\n labels += ['UNIATTRIBUTE SELF', 'UNIATTRIBUTE OTHER', 'MULTIATTRIBUTE COPULA']\n labels += ['QUESTIONS', 'CUTOFFS', 'ESTIMATION', 'SIMULATION']\n for label in labels:\n init_dict[label] = dict()\n\n paras_obj = self.attr['paras_obj']\n questions = self.attr['questions']\n\n # Preferences\n init_dict['UNIATTRIBUTE SELF']['marginal'] = self.attr['marginals'][0]\n init_dict['UNIATTRIBUTE SELF']['r'] = paras_obj.get_para('r_self')\n init_dict['UNIATTRIBUTE SELF']['max'] = self.attr['upper'][0]\n\n init_dict['UNIATTRIBUTE OTHER']['marginal'] = self.attr['marginals'][1]\n init_dict['UNIATTRIBUTE OTHER']['r'] = paras_obj.get_para('r_other')\n init_dict['UNIATTRIBUTE OTHER']['max'] = self.attr['upper'][1]\n\n init_dict['MULTIATTRIBUTE COPULA']['delta'] = paras_obj.get_para('delta')\n init_dict['MULTIATTRIBUTE COPULA']['self'] = paras_obj.get_para('self')\n init_dict['MULTIATTRIBUTE COPULA']['other'] = paras_obj.get_para('other')\n\n # Questions\n for q in questions:\n init_dict['QUESTIONS'][q] = paras_obj.get_para(q)\n\n # Cutoffs\n init_dict['CUTOFFS'] = self.attr['cutoffs']\n\n # Estimation\n init_dict['ESTIMATION']['detailed'] = self.attr['est_detailed']\n init_dict['ESTIMATION']['optimizer'] = self.attr['optimizer']\n init_dict['ESTIMATION']['agents'] = self.attr['est_agents']\n init_dict['ESTIMATION']['skip'] = self.attr['num_skip']\n init_dict['ESTIMATION']['file'] = self.attr['est_file']\n init_dict['ESTIMATION']['maxfun'] = self.attr['maxfun']\n init_dict['ESTIMATION']['start'] = self.attr['start']\n\n # Simulation\n init_dict['SIMULATION']['agents'] = self.attr['sim_agents']\n init_dict['SIMULATION']['seed'] = self.attr['sim_seed']\n init_dict['SIMULATION']['file'] = self.attr['sim_file']\n\n # Optimizer options\n init_dict['SCIPY-BFGS'] = dict()\n init_dict['SCIPY-BFGS']['gtol'] = self.attr['opt_options']['SCIPY-BFGS']['gtol']\n init_dict['SCIPY-BFGS']['eps'] = self.attr['opt_options']['SCIPY-BFGS']['eps']\n\n init_dict['SCIPY-POWELL'] = dict()\n init_dict['SCIPY-POWELL']['xtol'] = self.attr['opt_options']['SCIPY-POWELL']['xtol']\n init_dict['SCIPY-POWELL']['ftol'] = self.attr['opt_options']['SCIPY-POWELL']['ftol']\n\n print_init_dict(init_dict, fname)\n\n def _check_integrity(self):\n \"\"\"This method checks the integrity of the class instance.\"\"\"\n # Distribute class attributes for further processing.\n args = []\n args += ['paras_obj', 'sim_seed', 'sim_agents', 'sim_file', 'est_agents', 'maxfun']\n args += ['est_file', 'questions', 'start', 'num_skip']\n\n paras_obj, sim_seed, sim_agents, sim_file, est_agents, maxfun, est_file, questions, \\\n start, num_skip = dist_class_attributes(self, *args)\n\n # We restrict the identifiers for the questions between 1 and 16\n np.testing.assert_equal(12 < min(questions) <= max(questions) < 46, True)\n\n # The number of skipped individuals has to be non-negative.\n np.testing.assert_equal(0 <= num_skip, True)\n\n # We have to alternative how to start the estimation.\n np.testing.assert_equal(start in ['init', 'auto'], True)\n","sub_path":"trempy/clsModel.py","file_name":"clsModel.py","file_ext":"py","file_size_in_byte":6738,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"379963352","text":"import recommend\r\nimport pageRankRecommender as PRR\r\nimport recommenderSurprise as RS\r\nimport app\r\nimport json\r\nimport collections\r\nfrom collections import Counter\r\nimport userInfo\r\n\r\nclass RecommendationCombination():\r\n def __init__(self, numEntities = 1000): \r\n self.userFavs = {}\r\n self.topStories = {}\r\n self.stories = []\r\n self.users = []\r\n self.reviews = [] \r\n with open('resultCleanup.jl') as f: \r\n cnt = 0\r\n for line in f:\r\n if cnt > numEntities:\r\n break\r\n cnt += 1\r\n j = json.loads(line)\r\n if j[\"pT\"] == \"user\":\r\n self.users.append(\r\n {\r\n 'name':j['name'], \r\n 'stories':j['stories'],\r\n 'favorites':j['favorites']\r\n })\r\n favAuthors = []\r\n favs = j[\"favorites\"]\r\n for elem in favs:\r\n favAuthors.append(elem[\"A\"])\r\n self.userFavs[j[\"name\"]] = set(favAuthors)\r\n\r\n if j[\"pT\"] == \"story\":\r\n favs = int(j[\"otherInfo\"][\"favorites\"])\r\n author = j[\"author\"]\r\n link = j[\"storyLink\"]\r\n \r\n self.stories.append({'storyLink':j[\"storyLink\"]})\r\n\r\n if author not in self.topStories:\r\n self.topStories[author] = (link, int(favs))\r\n else:\r\n #if the current top story for the author has less favorites than the new story then make the new story the top story. else don't change anything.\r\n if int(self.topStories[author][1]) < int(favs):\r\n self.topStories[author] = (link, int(favs)) \r\n if j[\"pT\"] == \"review\":\r\n item = {}\r\n item['rO'] = j['rO']\r\n item['r'] = j['r']\r\n item['sS'] = j['sS']\r\n self.reviews.append(item)\r\n\r\n self.prRecommender = PRR.pageRankRecommender(self.users, self.stories)\r\n self.sRecommender = RS.surpriseRecommender(self.stories, self.reviews, self.users)\r\n self.sRecommender.train()\r\n\r\n def getTopAuthors(self, link):\r\n\r\n favoriteAuthors = userInfo.getFavoriteAuthors(link)\r\n\r\n basicRecommendations = recommend.recommender(favoriteAuthors, self.userFavs, self.topStories)\r\n basicRecommendations = Counter({ x : y[0] for x, y in basicRecommendations.items()})\r\n pageRankRecommendations = Counter({link: score * .1 for link, score in self.prRecommender.predictBestAuthors().items()})\r\n\r\n combinedResults = dict(basicRecommendations + pageRankRecommendations)\r\n #print(len(basicRecommendations),len(pageRankRecommendations),len(combinedResults))\r\n combinedResults = [(link, score) for link, score in combinedResults.items()]\r\n combinedResults = sorted(combinedResults, key=lambda tup: tup[1], reverse=True)[:10]\r\n #print(combinedResults[:10])\r\n\r\n\r\n return [ x for x, y in combinedResults]\r\n\r\n def getTopStories(self, link):\r\n def scaleCounter(counter, scaler):\r\n return Counter({ link : score * scaler for link, score in counter.items()})\r\n\r\n pageRankRecommendations = Counter({link: score for link, score in self.prRecommender.predictBestStories().items()})\r\n pageRankRecommendations = scaleCounter(pageRankRecommendations, .1)\r\n surpriseRecommendations = Counter(self.sRecommender.predict(link, self.stories))\r\n surpriseRecommendations = scaleCounter(surpriseRecommendations, .9)\r\n\r\n combinedResults = pageRankRecommendations + surpriseRecommendations\r\n combinedResults = [(link, score) for link, score in combinedResults.items()]\r\n combinedResults = sorted(combinedResults, key=lambda tup: tup[1], reverse=True)[:10]\r\n print(combinedResults)\r\n return [ x for x, y in combinedResults]\r\n","sub_path":"RecommendationCombination.py","file_name":"RecommendationCombination.py","file_ext":"py","file_size_in_byte":4121,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"393710694","text":"from google.appengine.ext import db\n\n\n\nclass Feed(db.Model):\n owner = db.UserProperty(required=True)\n last_fetch = db.DateTimeProperty()\n url = db.StringProperty(required=True)\n title = db.StringProperty()\n is_valid = db.BooleanProperty(default=True)\n\n\nclass ReadyData(db.Model):\n DATA_TYPES = ('feed', 'page')\n\n content = db.TextProperty(default='')\n owner = db.UserProperty(required=True)\n created = db.DateTimeProperty(auto_now_add=True)\n merged = db.IntegerProperty(default=1)\n data_type = db.StringProperty(choices=DATA_TYPES, required=True)\n\n def as_html_page(self):\n return '''\n <!DOCTYPE html>\n <html>\n <head>\n <meta http-equiv=\"Content-Type\" content=\"text/html; charset=UTF-8\">\n <title>Kindledump articles\n \n %s\n \n ''' % self.content\n","sub_path":"src/fetcher/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":912,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"181380966","text":"from torch.utils.data import dataset\nfrom torchvision import transforms\nimport numpy as np\nimport os\nimport matplotlib.pyplot as plt\nfrom torch.utils.data import DataLoader\nimport torchvision\nfrom PIL import Image\n\nsize = 32\ntrans = {\n 'train':\n transforms.Compose([\n transforms.RandomHorizontalFlip(),\n transforms.RandomCrop((size, size), padding=4),\n transforms.ToTensor(),\n transforms.Normalize(mean=[0.4914, 0.4822, 0.4465], std=[0.2023, 0.1994, 0.2010])\n ]),\n 'test':\n transforms.Compose([\n transforms.ToTensor(),\n transforms.Normalize(mean=[0.4914, 0.4822, 0.4465], std=[0.2023, 0.1994, 0.2010])\n ])\n}\n\n\ndef unpickle(file):\n import pickle\n with open(file, 'rb') as fo:\n dict = pickle.load(fo, encoding='bytes')\n return dict\n\n\nclass CIFAR10(dataset.Dataset):\n def __init__(self, mode, root='./data/cifar-10-batches-py/'):\n assert mode in ['train', 'test'], print('mode must be \"train\" or \"test\"')\n data_root = root\n data_files = {'train': ['data_batch_%d' % x for x in range(1, 6)],\n 'test': ['test_batch']}\n self.imgs = None\n self.labels = []\n # self.class_names = self._unpickle(os.path.join(data_root, 'batches.meta'))[b'label_names]\n for f in data_files[mode]:\n data_dict = unpickle(os.path.join(data_root, f))\n data = data_dict[b'data'].reshape(-1, 3, 32, 32).transpose(0, 2, 3, 1)\n if self.imgs is None:\n self.imgs = data\n else:\n self.imgs = np.vstack((self.imgs, data))\n self.labels += data_dict[b'labels']\n\n self.trans = trans[mode]\n\n def __getitem__(self, index):\n img = Image.fromarray(self.imgs[index])\n label = self.labels[index]\n img = self.trans(img)\n\n return img, label\n\n def __len__(self):\n return len(self.labels)\n\n\nclass CIFAR100(dataset.Dataset):\n def __init__(self, mode, root='./data/cifar-100-py'):\n super().__init__()\n self.data_root = root\n\n data = unpickle(os.path.join(self.data_root, mode))\n print(data.keys())\n self.fnames = data[b'filenames']\n self.imgs = data[b'data'].reshape(-1, 3, 32, 32).transpose(0, 2, 3, 1)\n self.labels = data[b'fine_labels']\n self.corse = data[b'coarse_labels']\n self.trans = trans[mode]\n\n def __getitem__(self, index):\n img = Image.fromarray(self.imgs[index])\n label = self.labels[index]\n img = self.trans(img)\n\n return img, label\n\n def __len__(self):\n return len(self.labels)\n\n\ndef get_data(num_classes=10, root='./data/cifar-10-batches-py'):\n if num_classes == 10:\n Dataset = CIFAR10\n else:\n Dataset = CIFAR100\n\n trainset = Dataset(mode='train', root=root)\n testset = Dataset(mode='test', root=root)\n\n return trainset, testset\n\n\ndef imshow(inp, title=None):\n \"\"\"Imshow for Tensor.\"\"\"\n inp = torchvision.utils.make_grid(inp)\n inp = inp.numpy().transpose((1, 2, 0))\n mean = np.array([0.485, 0.456, 0.406])\n std = np.array([0.229, 0.224, 0.225])\n inp = std * inp + mean\n inp = np.clip(inp, 0, 1)\n plt.imshow(inp)\n if title is not None:\n plt.title(title)\n plt.show()\n\n\nif __name__ == '__main__':\n data = CIFAR10(mode='train', root='./data/cifar-10-batches-py')\n loader = DataLoader(data, batch_size=16, shuffle=True, num_workers=0)\n imgs, label = iter(loader).__next__()\n imshow(imgs)\n print(label)\n","sub_path":"cifar_data.py","file_name":"cifar_data.py","file_ext":"py","file_size_in_byte":3568,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"15730038","text":"from tkinter import *\n\n\ndef main():\n root = Tk()\n App(root)\n print(\"Main loop.\")\n root.mainloop()\n\nclass App:\n def __init__(self, master):\n self.master = master \n master.protocol(\"WM_DELETE_WINDOW\", self.handler) #Exit when x pressed, notice that its the name of the function 'self.handler' and not a method call self.handler()\n\n def handler(self):\n self.master.destroy()\n print(\"Destoy root window.\")\n self.master.quit()\n print(\"Quit main loop.\")\n\nif __name__ == \"__main__\":\n main()\n\n","sub_path":"tk_test2.py","file_name":"tk_test2.py","file_ext":"py","file_size_in_byte":549,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"513157010","text":"import os\nimport datetime\nimport requests\n\nfrom flask import jsonify, request\nfrom dateutil.parser import parse\n\nfrom server import app, db, sqldb\nfrom penn.base import APIError\nfrom .models import StudySpacesBooking, User\nfrom .penndata import studyspaces, wharton\nfrom .base import cached_route\n\n\ndef get_wharton_sessionid(public=False):\n \"\"\" Try to get a GSR session id. \"\"\"\n sessionid = request.args.get('sessionid')\n cache_key = 'studyspaces:gsr:sessionid'\n\n if sessionid:\n return sessionid\n\n if public:\n if db.exists(cache_key):\n return db.get(cache_key).decode('utf8')\n\n return os.environ.get('GSR_SESSIONID')\n\n return None\n\n\ndef save_wharton_sessionid():\n sessionid = request.args.get('sessionid')\n cache_key = 'studyspaces:gsr:sessionid'\n\n if sessionid:\n db.set(cache_key, sessionid, ex=604800)\n\n\n@app.route('/studyspaces/gsr', methods=['GET'])\ndef get_wharton_gsrs_temp_route():\n \"\"\" Temporary endpoint to allow non-authenticated users to access the list of GSRs. \"\"\"\n date = request.args.get('date')\n try:\n data = wharton.get_wharton_gsrs(get_wharton_sessionid(public=True), date)\n save_wharton_sessionid()\n return jsonify(data)\n except APIError as error:\n return jsonify({'error': str(error)}), 400\n\n\n@app.route('/studyspaces/gsr/reservations', methods=['GET'])\ndef get_wharton_gsr_reservations():\n \"\"\"\n Returns JSON containing a list of Wharton GSR reservations.\n \"\"\"\n\n sessionid = get_wharton_sessionid()\n\n if not sessionid:\n return jsonify({'error': 'No Session ID provided.'})\n\n try:\n reservations = wharton.get_reservations(sessionid)\n save_wharton_sessionid()\n return jsonify({'reservations': reservations})\n except APIError as e:\n return jsonify({\"error\": str(e)}), 400\n\n\n@app.route('/studyspaces/gsr/delete', methods=['POST'])\ndef delete_wharton_gsr_reservation():\n \"\"\"\n Deletes a Wharton GSR reservation\n \"\"\"\n booking = request.form.get('booking')\n sessionid = request.form.get('sessionid')\n if not booking:\n return jsonify({\"error\": \"No booking sent to server.\"})\n if not sessionid:\n return jsonify({\"error\": \"No session id sent to server.\"})\n\n try:\n result = wharton.delete_booking(sessionid, booking)\n save_wharton_sessionid()\n return jsonify({'result': result})\n except APIError as e:\n return jsonify({\"error\": str(e)}), 400\n\n\n@app.route('/studyspaces/availability/', methods=['GET'])\ndef parse_times(building):\n \"\"\"\n Returns JSON containing all rooms for a given building.\n\n Usage:\n /studyspaces/availability/ gives all rooms for the next 24 hours\n /studyspaces/availability/?start=2018-25-01 gives all rooms in the start date\n /studyspaces/availability/?start=...&end=... gives all rooms between the two days\n \"\"\"\n if 'date' in request.args:\n start = request.args.get('date')\n end = request.args.get('date')\n else:\n start = request.args.get('start')\n end = request.args.get('end')\n\n if building == 1:\n sessionid = get_wharton_sessionid(public=True)\n try:\n rooms = wharton.get_wharton_gsrs(sessionid, date=start)\n rooms = wharton.switch_format(rooms)\n save_wharton_sessionid()\n except APIError as e:\n return jsonify({\"error\": str(e)}), 400\n else:\n try:\n rooms = studyspaces.get_rooms(building, start, end)\n rooms[\"location_id\"] = rooms[\"id\"]\n rooms[\"rooms\"] = []\n for room_list in rooms[\"categories\"]:\n for room in room_list[\"rooms\"]:\n room[\"thumbnail\"] = room[\"image\"]\n del room[\"image\"]\n room[\"room_id\"] = room[\"id\"]\n del room[\"id\"]\n room[\"gid\"] = room_list[\"cid\"]\n room[\"lid\"] = building\n room[\"times\"] = room[\"availability\"]\n del room[\"availability\"]\n for time in room[\"times\"]:\n time[\"available\"] = True\n time[\"start\"] = time[\"from\"]\n time[\"end\"] = time[\"to\"]\n del time[\"from\"]\n del time[\"to\"]\n rooms[\"rooms\"].append(room)\n except APIError as e:\n return jsonify({\"error\": str(e)}), 400\n return jsonify(rooms)\n\n\n@app.route('/studyspaces/locations', methods=['GET'])\ndef display_id_pairs():\n \"\"\"\n Returns JSON containing a list of buildings with their ids.\n \"\"\"\n def get_data():\n return {\"locations\": studyspaces.get_buildings() + [{\"lid\": 1, \"name\": \"Huntsman Hall\", \"service\": \"wharton\"}]}\n\n return cached_route('studyspaces:locations', datetime.timedelta(days=1), get_data)\n\n\n@app.route('/studyspaces/cancel', methods=['POST'])\ndef cancel_room():\n \"\"\"\n Cancels a booked room.\n \"\"\"\n try:\n user = User.get_user()\n except ValueError as err:\n return jsonify({\"error\": str(err)})\n\n booking_id = request.form.get(\"booking_id\")\n if not booking_id:\n return jsonify({\"error\": \"No booking id sent to server!\"})\n if \",\" in booking_id:\n return jsonify({\"error\": \"Only one booking may be cancelled at a time.\"})\n\n booking = StudySpacesBooking.query.filter_by(booking_id=booking_id).first()\n if booking:\n if (booking.user is not None) and (booking.user != user.id):\n return jsonify({\"error\": \"Unauthorized: This reservation was booked by someone else.\"}), 400\n if booking.is_cancelled:\n return jsonify({\"error\": \"This reservation has already been cancelled.\"}), 400\n\n if booking_id.isdigit():\n sessionid = request.form.get(\"sessionid\")\n if not sessionid:\n return jsonify({\"error\": \"No session id sent to server.\"}), 400\n try:\n wharton.delete_booking(sessionid, booking_id)\n save_wharton_sessionid()\n if booking:\n booking.is_cancelled = True\n sqldb.session.commit()\n else:\n save_booking(\n lid=1,\n email=user.email,\n booking_id=booking_id,\n is_cancelled=True,\n user=user.id\n )\n return jsonify({'result': [{\"booking_id\": booking_id, \"cancelled\": True}]})\n except APIError as e:\n return jsonify({\"error\": str(e)}), 400\n else:\n resp = studyspaces.cancel_room(booking_id)\n if \"error\" not in resp:\n if booking:\n booking.is_cancelled = True\n sqldb.session.commit()\n else:\n save_booking(\n email=user.email,\n booking_id=booking_id,\n is_cancelled=True,\n user=user.id\n )\n return jsonify({'result': resp})\n\n\n@app.route('/studyspaces/book', methods=['POST'])\ndef book_room():\n \"\"\"\n Books a room.\n \"\"\"\n try:\n room = int(request.form[\"room\"])\n except (KeyError, ValueError):\n return jsonify({\"results\": False, \"error\": \"Please specify a correct room id!\"}), 400\n\n try:\n start = parse(request.form[\"start\"])\n end = parse(request.form[\"end\"])\n except KeyError:\n return jsonify({\"results\": False, \"error\": \"No start and end parameters passed to server!\"}), 400\n\n try:\n lid = int(request.form[\"lid\"])\n except (KeyError, ValueError):\n lid = None\n\n email = None\n\n if lid == 1:\n sessionid = request.form.get(\"sessionid\")\n if not sessionid:\n return jsonify({\"results\": False, \"error\": \"You must pass a sessionid when booking a Wharton GSR!\"}), 400\n resp = wharton.book_reservation(sessionid, room, start, end)\n resp[\"results\"] = resp[\"success\"]\n room_booked = resp[\"success\"]\n del resp[\"success\"]\n if room_booked:\n save_wharton_sessionid()\n booking_id = None\n\n # Look up the reservation to get the booking id\n reservations = get_reservations(None, sessionid, 0)\n startStr = request.form[\"start\"].split(\"-\")[0]\n endStr = request.form[\"end\"].split(\"-\")[0]\n for reservation in reservations:\n resStartStr = reservation[\"fromDate\"].split(\"-\")[0]\n resEndStr = reservation[\"toDate\"].split(\"-\")[0]\n if startStr == resStartStr and endStr == resEndStr:\n booking_id = reservation[\"booking_id\"]\n break\n else:\n contact = {}\n for arg, field in [(\"fname\", \"firstname\"), (\"lname\", \"lastname\"), (\"email\", \"email\"), (\"nickname\", \"groupname\")]:\n try:\n contact[arg] = request.form[field]\n except KeyError:\n return jsonify({\"results\": False, \"error\": \"'{}' is a required parameter!\".format(field)})\n\n email = contact.get(\"email\")\n contact[\"custom\"] = {}\n contact[\"custom\"][\"q3699\"] = get_affiliation(email)\n for arg, field in [(\"q2533\", \"phone\"), (\"q2555\", \"size\"), (\"q2537\", \"size\"), (\"q3699\", \"affiliation\")]:\n try:\n contact[\"custom\"][arg] = request.form[field]\n except KeyError:\n pass\n\n resp = studyspaces.book_room(room, start.isoformat(), end.isoformat(), **contact)\n room_booked = resp.get(\"results\")\n booking_id = resp.get(\"booking_id\")\n\n try:\n user = User.get_user()\n user_id = user.id\n if email and user.email != email:\n user.email = email\n sqldb.session.commit()\n else:\n email = user.email\n except ValueError:\n user_id = None\n\n if room_booked:\n save_booking(\n lid=lid,\n rid=room,\n email=email,\n start=start.replace(tzinfo=None),\n end=end.replace(tzinfo=None),\n booking_id=booking_id,\n user=user_id\n )\n return jsonify(resp)\n\n\ndef get_affiliation(email):\n if \"wharton\" in email:\n return \"Wharton\"\n elif \"seas\" in email:\n return \"SEAS\"\n elif \"sas\" in email:\n return \"SAS\"\n else:\n return \"Other\"\n\n\n@app.route('/studyspaces/reservations', methods=['GET'])\ndef get_reservations_endpoint():\n \"\"\"\n Gets a users reservations.\n \"\"\"\n\n email = request.args.get('email')\n sessionid = request.args.get('sessionid')\n if not email and not sessionid:\n return jsonify({\"error\": \"A session id or email must be sent to server.\"}), 400\n\n libcal_search_span = request.args.get(\"libcal_search_span\")\n if libcal_search_span:\n try:\n libcal_search_span = int(libcal_search_span)\n except ValueError:\n return jsonify({\"error\": \"Search span must be an integer.\"}), 400\n else:\n libcal_search_span = 3\n\n try:\n reservations = get_reservations(email, sessionid, libcal_search_span)\n return jsonify({'reservations': reservations})\n except APIError as e:\n return jsonify({\"error\": str(e)}), 400\n\n\ndef get_reservations(email, sessionid, libcal_search_span, timeout=20):\n reservations = []\n if sessionid:\n try:\n gsr_reservations = wharton.get_reservations(sessionid, timeout)\n timezone = wharton.get_dst_gmt_timezone()\n\n for res in gsr_reservations:\n res[\"service\"] = \"wharton\"\n res[\"booking_id\"] = str(res[\"booking_id\"])\n res[\"name\"] = res[\"location\"]\n res[\"gid\"] = 1\n res[\"lid\"] = 1\n res[\"info\"] = None\n del res[\"location\"]\n\n date = datetime.datetime.strptime(res[\"date\"], \"%b %d, %Y\")\n date_str = datetime.datetime.strftime(date, \"%Y-%m-%d\")\n\n if res[\"startTime\"] == \"midnight\":\n res[\"fromDate\"] = date_str + \"T00:00:00-{}\".format(timezone)\n elif res[\"startTime\"] == \"noon\":\n res[\"fromDate\"] = date_str + \"T12:00:00-{}\".format(timezone)\n else:\n start_str = res[\"startTime\"].replace(\".\", \"\").upper()\n try:\n start_time = datetime.datetime.strptime(start_str, \"%I:%M %p\")\n except ValueError:\n start_time = datetime.datetime.strptime(start_str, \"%I %p\")\n start_str = datetime.datetime.strftime(start_time, \"%H:%M:%S\")\n res[\"fromDate\"] = \"{}T{}-{}\".format(date_str, start_str, timezone)\n\n if res[\"endTime\"] == \"midnight\":\n date += datetime.timedelta(days=1)\n date_str = datetime.datetime.strftime(date, \"%Y-%m-%d\")\n res[\"toDate\"] = date_str + \"T00:00:00-{}\".format(timezone)\n elif res[\"endTime\"] == \"noon\":\n res[\"toDate\"] = date_str + \"T12:00:00-{}\".format(timezone)\n else:\n end_str = res[\"endTime\"].replace(\".\", \"\").upper()\n try:\n end_time = datetime.datetime.strptime(end_str, \"%I:%M %p\")\n except ValueError:\n end_time = datetime.datetime.strptime(end_str, \"%I %p\")\n end_str = datetime.datetime.strftime(end_time, \"%H:%M:%S\")\n res[\"toDate\"] = \"{}T{}-{}\".format(date_str, end_str, timezone)\n\n del res[\"date\"]\n del res[\"startTime\"]\n del res[\"endTime\"]\n\n reservations.extend(gsr_reservations)\n\n except APIError:\n pass\n\n if email:\n confirmed_reservations = []\n try:\n def is_not_cancelled_in_db(booking_id):\n booking = StudySpacesBooking.query.filter_by(booking_id=booking_id).first()\n return not (booking and booking.is_cancelled)\n\n now = datetime.datetime.now()\n dateFormat = \"%Y-%m-%d\"\n i = 0\n while len(confirmed_reservations) == 0 and i < libcal_search_span:\n date = now + datetime.timedelta(days=i)\n dateStr = datetime.datetime.strftime(date, dateFormat)\n libcal_reservations = studyspaces.get_reservations(email, dateStr, timeout)\n confirmed_reservations = [res for res in libcal_reservations if (type(res) == dict and res[\"status\"] == \"Confirmed\"\n and datetime.datetime.strptime(res[\"toDate\"][:-6], \"%Y-%m-%dT%H:%M:%S\") >= now)]\n confirmed_reservations = [res for res in confirmed_reservations if is_not_cancelled_in_db(res[\"bookId\"])]\n i += 1\n\n except APIError:\n pass\n\n # Fetch reservations in database that are not being returned by API\n db_bookings = StudySpacesBooking.query.filter_by(email=email)\n db_booking_ids = [str(x.booking_id) for x in db_bookings if x.end\n and x.end > now\n and not str(x.booking_id).isdigit()\n and not x.is_cancelled]\n reservation_ids = [x[\"bookId\"] for x in confirmed_reservations]\n missing_booking_ids = list(set(db_booking_ids) - set(reservation_ids))\n if missing_booking_ids:\n missing_bookings_str = \",\".join(missing_booking_ids)\n missing_reservations = studyspaces.get_reservations_for_booking_ids(missing_bookings_str)\n confirmed_missing_reservations = [res for res in missing_reservations if res[\"status\"] == \"Confirmed\"]\n confirmed_reservations.extend(confirmed_missing_reservations)\n\n for res in confirmed_reservations:\n res[\"service\"] = \"libcal\"\n res[\"booking_id\"] = res[\"bookId\"]\n res[\"room_id\"] = res[\"eid\"]\n res[\"gid\"] = res[\"cid\"]\n del res[\"bookId\"]\n del res[\"eid\"]\n del res[\"cid\"]\n del res[\"status\"]\n del res[\"email\"]\n del res[\"firstName\"]\n del res[\"lastName\"]\n\n room_ids = \",\".join(list(set([str(x[\"room_id\"]) for x in confirmed_reservations])))\n if room_ids:\n rooms = studyspaces.get_room_info(room_ids)\n for room in rooms:\n room[\"thumbnail\"] = room[\"image\"]\n del room[\"image\"]\n del room[\"formid\"]\n\n for res in confirmed_reservations:\n room = [x for x in rooms if x[\"id\"] == res[\"room_id\"]][0]\n res[\"name\"] = room[\"name\"]\n res[\"info\"] = room\n del res[\"room_id\"]\n reservations.extend(confirmed_reservations)\n\n return reservations\n\n\ndef save_booking(**info):\n item = StudySpacesBooking(**info)\n\n sqldb.session.add(item)\n sqldb.session.commit()\n","sub_path":"server/studyspaces.py","file_name":"studyspaces.py","file_ext":"py","file_size_in_byte":17047,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"225569048","text":"import picar_4wd as fc\nimport part2\nimport time\nimport math\n\nFINE_TUNE_STEP_SIZE = 7.5\nTHRESHOLD = 100 #centimeter\nFINE_TUNE_PWR = part2.TURN_PWR\nFINE_TUNE_TIME = part2.TURN_TIME * (FINE_TUNE_STEP_SIZE/45) #make it turn FINE_TUNE_STEP_SIZE degree each time\n\nclass fineTune():\n def __init__(self, del_dir):\n self.ref_points = []\n self.curr_points = []\n self.del_dir = del_dir\n self.tune_range_degree = abs(del_dir) * 15 # turn 45 degree -> fine tune 15 degree; turn 90 degree -> fine tune 30 degree\n self.tune_range = round(self.tune_range_degree / FINE_TUNE_STEP_SIZE)\n\n def set_ref_points(self):\n ref_angle_dist = part2.get_distances( round(180/FINE_TUNE_STEP_SIZE), get_median = False)\n #ref_angle_dist = [(90, 200), (75,200), (60, 100), (45,110), (30,90), (15, 300), (0, 400), (-15, 400), (-30, 20), (-45, 10), (-60, 100), (-75, 400), (-90, 400)] ############ test #############\n self.ref_points = self._LT_threshold(ref_angle_dist)\n\n def fine_tune(self):\n # each elements in list represents FINE_TUNE_STEP_SIZE degree\n # fine tune +- self.tune_range_degree degree\n is_turn_left = (self.del_dir >= 0)\n curr_angle_dist = part2.get_distances( round(180/FINE_TUNE_STEP_SIZE), get_median = False)\n #curr_angle_dist = [(90, 90), (75,300), (60, 400), (45,400), (30,20), (15, 10), (0, 100), (-15, 400), (-30, 400), (-45, 400), (-60, 20), (-75, 10), (-90, 100)] ############ test ############\n self.curr_points = self._LT_threshold(curr_angle_dist)\n print(self.ref_points)\n print(self.curr_points)\n\n if is_turn_left == True:\n lhs = self.curr_points\n rhs = self.ref_points\n else:\n lhs = self.ref_points\n rhs = self.curr_points\n\n ideal_del_i = (abs(self.del_dir) * round(45/FINE_TUNE_STEP_SIZE)) #del_dir = 1 -> 45 degree, displacement = 45/15 = 3\n relevance_list = []\n for del_i in range( ideal_del_i - self.tune_range, ideal_del_i + self.tune_range + 1):\n relevance = self._cal_relevance(lhs, rhs, del_i)\n #relevance -= math.sqrt(abs(del_i - ideal_del_i)) ##### test #####\n relevance_list.append(relevance)\n print(relevance_list)\n fine_tune_displacement = self._get_fine_tune_displacement(relevance_list)\n \n if is_turn_left:\n fine_tune_degree = (-1)*fine_tune_displacement*FINE_TUNE_STEP_SIZE\n else:\n fine_tune_degree = fine_tune_displacement*FINE_TUNE_STEP_SIZE\n print(\"fine tune \" + str(fine_tune_degree) + \" degree\")\n \n # fine_tune_displacement > 0 means overturned, need turn back\n if (is_turn_left and fine_tune_displacement > 0) or (not is_turn_left and fine_tune_displacement < 0):\n fc.turn_right(FINE_TUNE_PWR)\n else:\n fc.turn_left(FINE_TUNE_PWR)\n time.sleep(FINE_TUNE_TIME * abs(fine_tune_displacement))\n fc.stop() \n\n def _LT_threshold(self, target_list):\n res = []\n for angle, dist in target_list:\n if dist >= THRESHOLD:\n res.append(0)\n else:\n res.append(1)\n return res\n\n def _cal_relevance(self, lhs, rhs, del_i):\n rel = 0\n list_len = len(lhs)\n for i in range(list_len - del_i):\n if lhs[del_i + i] == rhs[i]:\n rel += 1\n else:\n rel -= 0.5\n\n return rel\n\n def _get_fine_tune_displacement(self, relevance_list):\n max_relevance_index = self.tune_range\n max_relevance = relevance_list[self.tune_range]\n for i in range(2*self.tune_range + 1):\n if relevance_list[i] > max_relevance:\n max_relevance_index = i\n max_relevance = relevance_list[i]\n return max_relevance_index - self.tune_range\n","sub_path":"lab1_code/fineTuneOrient.py","file_name":"fineTuneOrient.py","file_ext":"py","file_size_in_byte":3890,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"8366962","text":"#!/usr/bin/env python3\n\"\"\"\nLaunch a python shell\n\nTry to launch ipthon or bpython. Fall back to the\nstandard python interpreter.\n\"\"\"\n\nimport services\n\n\ndef start_ipython(user_ns={}):\n \"\"\"Start the ipython shell\"\"\"\n from IPython import start_ipython\n start_ipython(argv=[], user_ns=user_ns)\n\n\ndef start_bpython(locals_={}):\n \"\"\"Start the bpython shell\"\"\"\n from bpython import embed\n embed(locals_=locals_)\n\n\ndef start_fallback(local={}):\n \"\"\"Start the fallback interpreter\"\"\"\n from code import interact\n interact(local=local)\n\n\ndef start_shell(local={}):\n \"\"\"\n Start a python shell\n \"\"\"\n shells = [start_ipython, start_bpython, start_fallback]\n for shell in shells:\n try:\n shell(local)\n except ImportError:\n pass # try next\n else:\n return\n\n\ndef console(args):\n \"\"\"\n Start the API console\n \"\"\"\n host = \"localhost:2344\"\n\n services.init(host)\n start_shell(services.__dict__)\n\n\nif __name__ == \"__main__\":\n console(None)\n","sub_path":"console/console.py","file_name":"console.py","file_ext":"py","file_size_in_byte":1032,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"336197813","text":"from functools import lru_cache\r\n\r\ndef korita(n, m, l):\r\n if n == 0:\r\n return 0\r\n if l > n: \r\n return 0\r\n counter = 0\r\n if m * (l + 1) <= n:\r\n counter += 1 + korita(n - m, m - 1, l)\r\n return counter\r\n\r\nprint(korita(9, 3, 2))","sub_path":"vaje/korita.py","file_name":"korita.py","file_ext":"py","file_size_in_byte":271,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"359784935","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\nfrom __future__ import print_function\nimport itertools\n\ndef consists_of(n, digits):\n return sorted(n) == map(str, digits)\n\ndef sum_of_fifths_pow_str(digits):\n res = str(sum(map(lambda x:x**5, digits)))\n res = (6-len(res))*'0' + res\n return res\n\ndef main():\n result = 0\n digits = range(10)*6\n for d in itertools.combinations_with_replacement(range(10), r=6):\n val = sum_of_fifths(d)\n if consists_of(val, d):\n print(d, val)\n result += int(val)\n print(result-1)\n\nif __name__ == '__main__':\n main()\n","sub_path":"pe_30.py","file_name":"pe_30.py","file_ext":"py","file_size_in_byte":605,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"114584656","text":"#coding=utf-8\n\n\"\"\"\nvalid_keys 表示 json keys 映射到 t_shixin_valid 相应的columns\n\nvalid_columns 对应 t_shixin_valid 的columns\ninvalid_columns 对应 t_shixin_invalid 的columns\n\"\"\"\n\nvalid_keys = {\n 'id': 'sys_id',\n 'iname': 'name',\n 'age': 'age',\n 'sexy': 'sex',\n 'cardNum':'card_num',\n 'businessEntity': 'business_entity',\n 'areaName': 'area_name',\n 'caseCode': 'case_code',\n 'regDate': 'reg_date',\n 'publishDate': 'publish_date',\n 'gistId': 'gist_id',\n 'courtName': 'court_name',\n 'gistUnit': 'gist_unit',\n 'duty': 'duty',\n 'performance': 'performance',\n 'disruptTypeName': 'disrupt_type_name',\n 'partyTypeName': 'party_type_name'\n}\n\nvalid_columns = valid_keys.values()\nvalid_columns.append('flag')\n\ninvalid_columns = ('sys_id', 'err_type')\n\n\n","sub_path":"current/shixin_spider/configuration/columns_cfg.py","file_name":"columns_cfg.py","file_ext":"py","file_size_in_byte":807,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"39962972","text":"def mymax(a, b):\n if a > b:\n return a\n else:\n return b\n\n\ndef test_mymax(test_data, test_result, test_id):\n if mymax(test_data[0], test_data[1]) == test_result:\n print(\"%s is correct\" % test_id)\n else:\n print(\"%s failed!\" % test_id)\n\n\nif __name__ == '__main__':\n lista1 = [1, 5]\n lista2 = [6, 1]\n lista3 = [10, 10]\n\n test_mymax(lista1, 5, \"Test 1\")\n test_mymax(lista2, 6, \"Test 2\")\n test_mymax(lista3, 10, \"Test 3\")\n","sub_path":"simple46exercises/simple/zadanie1.py","file_name":"zadanie1.py","file_ext":"py","file_size_in_byte":473,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"350265479","text":"from Messages.Message import Message\n\n\nclass BOSCO:\n\n name = \"BOSCO Protocol\"\n\n def __init__(self, **kargs):\n self.env = kargs[\"env\"]\n self.pki = kargs[\"pki\"]\n self.pki.register(self)\n self.input = None\n\n def run_node(self):\n round = self.env.get_round()\n myid = self.env.get_id(self)\n flag = 0\n if round == 0:\n self.input = self.env.get_input(myid)\n self.env.put_broadcast(self, self.pki.sign(\n self, Message(myid, self.input)))\n else:\n if flag:\n self.env.get_input_msgs(self)\n self.env.put_broadcast(self, self.pki.sign(\n self, Message(myid, self.input)))\n else:\n msgs = self.env.get_input_msgs(self)\n d = {}\n for msg in msgs:\n if(not self.pki.verify(msg)):\n raise RuntimeError\n key = msg.get_extraction()\n if key not in d:\n d[key] = 0\n d[key] = d[key]+1\n if not d:\n raise RuntimeError\n d_sorted = sorted(\n d.items(), key=lambda kv: kv[1], reverse=True)\n if(d_sorted[0][1] >= (self.env.get_n()-self.env.get_f())):\n self.env.put_output(self, d_sorted[0][0])\n self.input = d_sorted[0][0]\n self.env.put_broadcast(self, self.pki.sign(\n self, Message(myid, self.input)))\n elif (d_sorted[0][1] > (self.env.get_n()-self.env.get_f())/2):\n if len(d_sorted) > 1 and d_sorted[1][1] > (self.env.get_n()-self.env.get_f())/2:\n self.env.put_broadcast(self, self.pki.sign(\n self, Message(myid, self.input)))\n else:\n self.input = d_sorted[0][0]\n self.env.put_broadcast(self, self.pki.sign(\n self, Message(myid, self.input)))\n","sub_path":"src/Protocols/BOSCO.py","file_name":"BOSCO.py","file_ext":"py","file_size_in_byte":2096,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"119161242","text":"\n# disclaimer: this code was written for testing and self-learning, do\n# not use in any serious application, do not expect any serious security!\nimport socket, sys, logging, hashlib, random, json, sympy, pickle\n\n# global settings - make sure those are consistent\nNBITS=32 # increase this is comparing worth with Jeff Bezos or Mark Zuckerberg\n\n# we work modulo this prime\nP = (1<<510)+15 # yes it happens to be prime\n\nSECURITY=512\n\n# setup logging\nlogging.basicConfig(level=logging.INFO, format='%(asctime)s: %(message)s')\nLOG = logging.getLogger(__name__)\n\n# utility - wait for connection on port, or connect to host:port\ndef interact(info):\n info = info.split(':')\n\n if len(info) == 1:\n port = int(info[0])\n LOG.info('listening on port %d', port)\n sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n sock.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)\n\n sock.bind( ('localhost', port) )\n sock.listen(1)\n connection, client_address = sock.accept()\n LOG.info('Client connected from %s', client_address)\n\n return connection\n\n else:\n host, port = info[0], int(info[1])\n sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n LOG.info('connecting to %s:%d', host, port)\n sock.connect( (host,port) )\n LOG.info('connected!')\n return sock\n\n\n# compute 'ups', they are values that are bigger than a, and will intercept a bigger value of the other party\n# for each i, if the i-th bit is 0, set it to 1 and clear lower bits\ndef s_up(a, n=NBITS):\n return { (a - (a & ((1<>i) & 1 == 0 }\n\n# compute 'downs', they are values that are smaller than a, and will intercept a smaller value of the other party\n# for each i, if the i-th bit is 1, clear lower bits\ndef s_down(a, n=NBITS):\n return { (a - (a & ((1<>i) & 1 == 1 }\n\n# random integer of 512 bits\ndef rand512():\n return random.getrandbits(512)\n\n# return the sha512, as an integer\ndef sha512(s):\n return int.from_bytes(hashlib.sha512(s.encode()).digest(), byteorder='big')\n\ndef hashes(s, n=NBITS):\n return [*(sha512(str(x)) for x in s), *(rand512() for i in range(n-len(s)))]\n\n\n# compute x^e, b y repeated squaring, modulo p\ndef modpow(x, e, p=P):\n r, t = ((x % p) if (e & 1) else 1), x\n e >>= 1\n while e:\n t = (t * t) % p\n if e & 1:\n r = (r * t) % p\n e >>= 1\n return r\n\n# compute a number relatively prime to b, that is,\n# a number r so that r, b have no common factor\ndef rand_coprime(b):\n while True:\n r = random.randint(b // 4, b)\n if sympy.gcd(r, b) == 1:\n return r\n\nif len(sys.argv) != 3:\n print('Usage: prog.py number_to_compare [host:]port (if host is specified connect, otherwise listen and wait)')\n sys.exit()\n \nval = int(sys.argv[1])\nif (val >> NBITS) != 0:\n print('Congratulations, you own many billions, and', NBITS, 'bits are not enough for you!')\n print('You will have to up the NBITS parameter, and the other party too for this to work.')\n sys.exit()\n\nLOG.info('will compare the provided value with a private remote number')\n\nLOG.info('computing my downs/ups, plus padding to hide their sizes')\nmy_downs = hashes(s_down(val))\nmy_ups = hashes(s_up(val))\n\nLOG.info('generating a private key')\nmy_key = rand_coprime(P-1)\n\nLOG.info('encrypting and shuffling my downs/ups')\nM_my_downs = [modpow(x, my_key) for x in my_downs]\nM_my_ups = [modpow(x, my_key) for x in my_ups]\nrandom.shuffle(M_my_downs)\nrandom.shuffle(M_my_ups)\n\n# listen or connect to host:port\nc = interact(sys.argv[2])\n\nLOG.info('sending my downs/ups encrypted with my key...')\nc.sendall(pickle.dumps([M_my_downs, M_my_ups]))\nH_his_downs, H_his_ups = pickle.loads(c.recv(2*SECURITY*NBITS//8*3))\nLOG.info('...received his downs/ups encrypted with his key')\n\nLOG.info('bi-encrypting (with my key) and shuffling his downs/ups')\nHM_his_downs = [modpow(x, my_key) for x in H_his_downs]\nHM_his_ups = [modpow(x, my_key) for x in H_his_ups]\nrandom.shuffle(HM_his_downs)\nrandom.shuffle(HM_his_ups)\n\nLOG.info('sending his bi-encrypted downs/ups...')\nc.sendall(pickle.dumps([HM_his_downs, HM_his_ups]))\nHM_my_downs, HM_my_ups = pickle.loads(c.recv(2*SECURITY*NBITS//8*3))\nLOG.info('...received my bi-encrypted downs/ups')\n\nLOG.info('n. insections of my_downs and his_ups: %d (is my_value > his_value?)',\n len(set(HM_my_downs) & set(HM_his_ups)))\nLOG.info('n. insections of my_ups and his_downs: %d (is his_value > my_value?)',\n len(set(HM_my_ups) & set(HM_his_downs)))\n","sub_path":"millionaires_old.py","file_name":"millionaires_old.py","file_ext":"py","file_size_in_byte":4601,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"446337392","text":"import pytest\r\nimport agentpy as ap\r\nimport numpy as np\r\n\r\n\r\ndef test_repr():\r\n model = ap.Model()\r\n model.add_agents()\r\n model.add_env()\r\n assert model.agents.__repr__() == \"AgentList [1 agent]\"\r\n assert model.envs.__repr__() == \"EnvList [1 environment]\"\r\n assert model.objects.__repr__() == \"ObjList [2 objects]\"\r\n l1 = model.agents.id\r\n l2 = l1 + 1\r\n assert l1.__repr__() == \"AttrList of 'id': [1]\"\r\n assert l2.__repr__() == \"AttrList: [2]\"\r\n\r\n\r\ndef test_call():\r\n class MyAgent(ap.Agent):\r\n def method(self):\r\n if self.id == 2:\r\n self.model.agents[2].delete()\r\n self.model.called.append(self.id)\r\n\r\n model = ap.Model()\r\n model.called = []\r\n model.add_agents(4, MyAgent)\r\n model.agents.call('method', check_alive=True)\r\n assert model.called == [1, 2, 4]\r\n\r\n model = ap.Model()\r\n model.called = []\r\n model.add_agents(4, MyAgent)\r\n model.agents.method()\r\n assert model.called == [1, 2, 3, 4]\r\n\r\n\r\ndef test_attr_calls():\r\n model = ap.Model()\r\n model.add_agents(2)\r\n model.agents.x = 1\r\n model.agents.f = lambda: 2\r\n assert list(model.agents.x) == [1, 1]\r\n assert list(model.agents.f()) == [2, 2]\r\n with pytest.raises(AttributeError):\r\n assert list(model.agents.y) # Convert to list to call attribute\r\n with pytest.raises(TypeError):\r\n assert model.agents.x() # noqa\r\n\r\n\r\ndef test_select():\r\n \"\"\" Select subsets with boolean operators. \"\"\"\r\n model = ap.Model()\r\n model.add_agents(3)\r\n selection1 = model.agents.id == 2\r\n selection2 = model.agents.id != 2\r\n selection3 = model.agents.id < 2\r\n selection4 = model.agents.id > 2\r\n selection5 = model.agents.id <= 2\r\n selection6 = model.agents.id >= 2\r\n assert selection1 == [False, True, False]\r\n assert selection2 == [True, False, True]\r\n assert selection3 == [True, False, False]\r\n assert selection4 == [False, False, True]\r\n assert selection5 == [True, True, False]\r\n assert selection6 == [False, True, True]\r\n assert list(model.agents.select(selection1).id) == [2]\r\n\r\n\r\ndef test_random():\r\n \"\"\" Test random shuffle and selection. \"\"\"\r\n model = ap.Model()\r\n model.add_agents(2)\r\n assert len(model.agents) == len(model.agents.shuffle())\r\n assert len(model.agents.random()) == 1\r\n\r\n # Custom generator with seperate seed\r\n model = ap.Model()\r\n model.add_agents(5)\r\n generator = np.random.default_rng(1)\r\n assert len(model.agents.random(generator=generator)) == 1\r\n assert model.agents.random(generator=generator).id[0] == 3\r\n assert list(model.agents.shuffle(generator=generator).id) == [5, 1, 3, 2, 4]\r\n\r\n # Test with single agent\r\n model = ap.Model()\r\n agents = model.add_agents(1)\r\n assert model.agents.random()[0] is agents[0]\r\n assert model.agents.shuffle()[0] is agents [0]\r\n\r\n # Agentlist with no model defined directly\r\n model = ap.Model()\r\n agents = model.add_agents(3)\r\n agents = ap.AgentList(agents)\r\n model.run(steps=0, seed=1, display=False)\r\n assert agents.random()[0].id == 2\r\n\r\n # Agentlist with no model defined\r\n # (no seed control without model, test can only check if no errors)\r\n agents1 = ap.AgentList([1, 2, 3])\r\n agents1.random()\r\n\r\n\r\ndef test_sort():\r\n \"\"\" Test sorting method. \"\"\"\r\n model = ap.Model()\r\n model.add_agents(2)\r\n model.agents[0].x = 1\r\n model.agents[1].x = 0\r\n model.agents.sort('x')\r\n assert list(model.agents.x) == [0, 1]\r\n assert list(model.agents.id) == [2, 1]\r\n\r\n\r\ndef test_arithmetics():\r\n \"\"\" Test arithmetic operators \"\"\"\r\n\r\n model = ap.Model()\r\n model.add_agents(3)\r\n agents = model.agents\r\n\r\n agents.x = 1\r\n assert agents.x.attr == \"x\"\r\n assert list(agents.x) == [1, 1, 1]\r\n\r\n agents.y = ap.AttrList([1, 2, 3])\r\n assert list(agents.y) == [1, 2, 3]\r\n\r\n agents.x = agents.x + agents.y\r\n assert list(agents.x) == [2, 3, 4]\r\n\r\n agents.x = agents.x - ap.AttrList([1, 1, 1])\r\n assert list(agents.x) == [1, 2, 3]\r\n\r\n agents.x += 1\r\n assert list(agents.x) == [2, 3, 4]\r\n\r\n agents.x -= 1\r\n assert list(agents.x) == [1, 2, 3]\r\n\r\n agents.x *= 2\r\n assert list(agents.x) == [2, 4, 6]\r\n\r\n agents.x = agents.x * agents.x\r\n assert list(agents.x) == [4, 16, 36]\r\n\r\n agents.x = agents.x / agents.x\r\n assert list(agents.x)[0] == pytest.approx(1.)\r\n\r\n agents.x /= 2\r\n assert list(agents.x)[0] == pytest.approx(0.5)\r\n","sub_path":"tests/test_lists.py","file_name":"test_lists.py","file_ext":"py","file_size_in_byte":4466,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"56057105","text":"# A positive fraction whose numerator is less than its denominator is\n# called a proper fraction.\n# For any denominator, d, there will be d−1 proper fractions; for example,\n# with d = 12:\n# 1/12, 2/12, 3/12, 4/12, 5/12, 6/12, 7/12, 8/12, 9/12, 10/12, 11/12.\n# \n# We shall call a fraction that cannot be cancelled down a resilient fraction.\n# Furthermore we shall define the resilience of a denominator, R(d), to be the\n# ratio of its proper fractions that are resilient; for example, R(12) = 4/11.\n# In fact, d = 12 is the smallest denominator having a resilience R(d) < 4/10.\n# \n# Find the smallest denominator d, having a resilience R(d) < 15499/94744.\n\n# THEORY:\n# \n# If phi(n) is Euler's totient function, then R(d) = phi(d) / (d - 1).\n# If d = p1 * p2 * p3 * ... * pk, with all primes p distinct,\n# then R(d) = ((p1 - 1) * ... * (pk - 1)) / ((p1 * ... * pk) - 1).\n# Multiplying d by a prime p that it's already divisible by results in the\n# numerator and the left half of the denominator both being multiplied by p.\n# \n# This means that to minimize R(d), d should first be set equal to a product\n# of distinct primes D, and then be set equal to multiples of D until the -1\n# in the denominator ceases to be significant.\n\nfrom time import time\nimport sys\nsys.path.append(\"../Library\")\nfrom peresult import peresult\nfrom primefns import primesbelow\n\ndef solve(cap = 15499/94744):\n start = time()\n primes = primesbelow(100) # Safe overestimate\n numerator = 1\n n = 1\n for p in primes:\n if (numerator * (p - 1)) / (n * p - 1) > cap:\n numerator *= p - 1\n n *= p\n else:\n for mult in range(1, p):\n if (numerator * mult) / (n * mult - 1) < cap:\n result = n * mult\n break\n break\n else: # Loop fell through. Primes list wasn't long enough\n raise RuntimeError(\"Primes list in code too short. Edit and extend\")\n peresult(243, result, time() - start)\n\nif __name__ == \"__main__\":\n solve()\n","sub_path":"Problems 201-300/pe243Resilience.py","file_name":"pe243Resilience.py","file_ext":"py","file_size_in_byte":2033,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"46368736","text":"from splinter import Browser\nfrom bs4 import BeautifulSoup as bs\nfrom webdriver_manager.chrome import ChromeDriverManager\nimport pandas as pd\n\n\ndef init_browser():\n executable_path = {'executable_path': ChromeDriverManager().install()}\n return Browser('chrome', **executable_path, headless=False)\n\n\ndef scrape():\n # ## NASA Mars News\n url = 'https://redplanetscience.com'\n browser = init_browser()\n browser.visit(url)\n html = browser.html\n soup = bs(html, 'html.parser')\n\n results = soup.find_all(\"div\", class_=\"content_title\")\n content_titles = []\n for result in results:\n content_titles.append(result.text)\n\n results = soup.find_all(\"div\", class_=\"article_teaser_body\")\n content_para = []\n for result in results:\n content_para.append(result.text)\n\n content = []\n for i in range(len(results)):\n content.append({'title': content_titles[i], 'para': content_para[i]})\n mongo_collection = {'contents': content,\n 'featured_image_url': '',\n 'hemisphere': ''\n }\n\n# ## JPL Mars Space Images - Featured Image\n\n url = 'https://spaceimages-mars.com/'\n browser.visit(url)\n\n links_found = browser.links.find_by_partial_text('FULL IMAGE')\n for link in links_found:\n print(link[\"href\"])\n featured_image_url = link[\"href\"]\n mongo_collection['featured_image_url'] = featured_image_url\n\n url = 'https://galaxyfacts-mars.com'\n tables = pd.read_html(url)\n\n df = tables[1]\n df.to_html('MarsFacts.html')\n\n # ## Mars Hemispheres\n\n url = 'https://marshemispheres.com/'\n browser.visit(url)\n\n links = browser.links.find_by_partial_text('Hemisphere Enhanced')\n for link in links:\n print(link['href'])\n\n hemisphere_image_urls = []\n for i in range(len(links)):\n browser.links.find_by_partial_text('Hemisphere Enhanced')[i].click()\n link_img = browser.links.find_by_partial_text('Original')\n soup = bs(browser.html, 'html.parser')\n title = soup.find('h2', class_='title').text.replace(\" Enhanced\", \"\")\n hemisphere_image_urls.append(\n {\"title\": title, \"img_url\": link_img[\"href\"]})\n browser.links.find_by_partial_text('Back').click()\n mongo_collection['hemisphere'] = hemisphere_image_urls\n\n return mongo_collection\n","sub_path":"scrape_mars.py","file_name":"scrape_mars.py","file_ext":"py","file_size_in_byte":2355,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"609909755","text":"\"\"\"\nconway.py\nAuthor: Christopher Lee\nCredit: \nAssignment:\nWrite and submit a program that plays Conway's Game of Life, per \nhttps://github.com/HHS-IntroProgramming/Conway-Life\n\"\"\"\n#==================================IMPORTS======================================\nfrom ggame import *\nfrom math import floor\n#==============================COLORS_AND_LINES=================================\nblack = Color(0, 1)\nwhite = Color(0xffffff, 1)\npink = Color(0xFF0097, 1)\nblue = Color(0x00B6FF, 1)\nline = LineStyle(1, white)\nblackline = LineStyle(0.1, black)\nc = {}\ncc = []\n#================================IMPORTANT======================================\n'''gridnumber is the number of cells that there are for each row\nRecommended is 20. Max is 30 before program starts to slow'''\n#gridnumber = 20\ngridnumber = int(input('''How many cells would you like each row to have?\nRecommended is 20 cells.\nMax is 30 cells before program starts to slow.\n'''))\n#Scales screen based on gridnumber\nScreenWidth = gridnumber * 100\nScreenHeight = gridnumber * 100\n#=================================CLASSES=======================================\nclass grid(Sprite):\n g = RectangleAsset(50, 50, blackline, white)\n def __init__(self, position):\n super().__init__(grid.g, position)\n self.visible = True\nclass cell(Sprite):\n cc = RectangleAsset(50, 50, blackline, blue)\n def __init__(self, position):\n super().__init__(cell.cc, position)\n self.visible = False\nclass deadcell(Sprite):\n dc = RectangleAsset(50, 50, blackline, pink)\n def __init__(self, position):\n super().__init__(deadcell.dc, position)\n self.visible = False\n#================================CREATES GRID===================================\ndef row(x):\n xx = x\n y = 0\n for i in range(gridnumber):\n grid((xx, y))\n cell((xx, y))\n deadcell((xx, y))\n y += 50\n#----------------------------------RULES----------------------------------------\ndef rules():\n print('''\nRULES: \n1. Any live cell with fewer than two live neighbors dies, as if by underpopulation.\n\n2. Any live cell with two or three live neighbors lives on to the next generation.\n\n3. Any live cell with more than three live neighbors dies, as if by overpopulation.\n\n4. Any dead cell with exactly three live neighbors becomes a live cell, as if by reproduction.\n\nHOW TO PLAY:\n- Click where you want to add a cell to the grid\n\n- Press R to move to the next generation or start or stop the program\n\n- Press C to reset / clear the grid (not fully working)\n\n- Press S to print the cell's status\n\n- Comment out line 149 and line 151 if you want the steps to be automatic\n''')\n#==============================RUNNING_PROGRAM==================================\nclass map(App):\n def __init__(self, width, height):\n super().__init__(width, height)\n self.go = False\n rules()\n x = 0\n for i in range(gridnumber):\n row(x)\n x += 50\n map.listenKeyEvent('keydown', 'r', self.r)\n map.listenKeyEvent('keydown', 'c', self.c)\n map.listenKeyEvent('keydown', 's', self.s)\n map.listenMouseEvent('click', self.mouse)\n#----------------------------------STEP_FUNC------------------------------------\n def step(self):\n if self.go == True:\n age = 0\n coordlist = []\n for (xc, yc) in cc:\n coordlist.append((xc, yc))\n check = []\n for (xc, yc) in coordlist:\n for x in range(xc - 50, xc + 100, 50):\n if x <= ScreenWidth and x >= 0:\n for y in range(yc - 50, yc + 100, 50):\n if y <= ScreenHeight and y >= 0 and (x, y) not in check:\n check.append((x, y))\n for (xc, yc) in check:\n exist = 0\n neighbor = []\n for x in range(xc - 50, xc + 100, 50):\n if x <= ScreenWidth and x >= 0:\n for y in range(yc - 50, yc + 100, 50):\n if y <= ScreenHeight and y >= 0:\n neighbor.append((x, y))\n neighbor.remove((xc, yc))\n for (xcoord, ycoord) in neighbor:\n if (xcoord, ycoord) in coordlist:\n exist += 1\n if exist == 3 and (xc, yc) not in coordlist:\n c[(xc, yc)] = 'a'\n cell((xc, yc)).visible = True\n cc.append((xc, yc))\n elif (xc, yc) in coordlist:\n if exist == 2 or exist == 3:\n if age % 2 != 0:\n c[(xc, yc)] = 'a'\n cell((xc, yc)).visible = True\n if age % 2 == 0:\n c[(xc, yc)] = 'a'\n deadcell((xc, yc)).visible = True\n else:\n c[(xc, yc)] = 'd'\n grid((xc, yc)).visible = True\n del c[(xc, yc)]\n cc.remove((xc, yc))\n age += 1\n '''\n for coord in c:\n if c[(x, y)] == 'a':\n cell(coord).visible = True\n elif c[(x, y)] == 'd':\n grid(coord).visible = True\n elif c[(x, y)] == 'da':\n c[(x, y)] = 'a'\n cell(coord).visible = True\n '''\n#-------------------------------------vvvv--------------------------------------\n #self.go = False\n#-------------------------------------^^^^--------------------------------------\n #print('Stopping...')\n#-------------------------------MOUSE_CLICK-------------------------------------\n def mouse(self, event):\n if self.go == False:\n x = floor(event.x / 50) * 50\n y = floor(event.y / 50) * 50\n coord = (x, y)\n if x >= 0 and y >= 0 and x < gridnumber * 50 and y < gridnumber * 50:\n c[coord] = 'p'\n if c[coord] == 'a':\n print('test')\n c[coord] = 'd'\n grid(coord).visible = True\n else:\n c[coord] = 'a'\n cc.append(coord)\n cell(coord).visible = True\n#-----------------------------MOVE_TO_NEXT_GEN----------------------------------\n def r(self, event):\n self.go = not self.go\n if self.go == True:\n print('Running...')\n else:\n self.go = False\n print('Stopping...')\n#----------------------------------CLEAR----------------------------------------\n def c(self, event):\n print('Clearing...')\n x = 0\n for i in range(gridnumber):\n row(x)\n x += 50\n c.clear()\n exist = 0\n cc = []\n coordlist = []\n check = []\n neighbor = []\n#-----------------------------------UPDATE--------------------------------------\n def s(self, event):\n if c == {}:\n print('There are no alive cells')\n else:\n print(\"Printing status of cells\")\n print(c)\n print('Done')\n#====================================RUN=======================================\nmyapp = map(ScreenWidth, ScreenHeight)\nmyapp.run()\n","sub_path":"conway.py","file_name":"conway.py","file_ext":"py","file_size_in_byte":7424,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"311974393","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n\"\"\"\nAIM utility functions.\n\"\"\"\n\n\n# ----------------------------------------------------------------------------\n# Imports\n# ----------------------------------------------------------------------------\n\n# Standard library modules\nimport base64\nimport pathlib\nfrom io import BytesIO\n\n# Third-party modules\nfrom PIL import Image\n\n# First-party modules\nfrom deepux1_metrics.ux1.src.core.constants import IMAGE_QUALITY_JPEG\n\n# ----------------------------------------------------------------------------\n# Metadata\n# ----------------------------------------------------------------------------\n\n__author__ = \"Markku Laine\"\n__date__ = \"2020-08-21\"\n__email__ = \"markku.laine@aalto.fi\"\n__version__ = \"1.0\"\n\n\n# ----------------------------------------------------------------------------\n# Utility functions\n# ----------------------------------------------------------------------------\n\n\ndef read_image(filepath: pathlib.Path) -> str:\n \"\"\"\n Read an image from a file.\n\n Args:\n filepath: Input image file path\n\n Returns:\n Image encoded in Base64\n \"\"\"\n with open(filepath, \"rb\") as f:\n image_base64: str = base64.b64encode(f.read()).decode(\"utf-8\")\n\n return image_base64\n\n\ndef write_image(image_base64: str, filepath: pathlib.Path):\n \"\"\"\n Write an image to a file.\n\n Args:\n image_base64: Image encoded in Base64\n filepath: Output image file path\n \"\"\"\n with open(filepath, \"wb\") as f:\n f.write(base64.b64decode(image_base64))\n\n\ndef convert_image(\n png_image: str, jpeg_image_quality: int = IMAGE_QUALITY_JPEG\n) -> str:\n \"\"\"\n Convert an image from PNG to JPEG, encoded in Base64.\n\n (Semi-)transparent pixels are replaced with (semi-)white pixels in\n the output JPEG image.\n\n Args:\n png_image: PNG image encoded in Base64\n\n Kwargs:\n jpeg_image_quality: JPEG image quality (defaults to 70)\n\n Returns:\n JPEG image encoded in Base64\n \"\"\"\n img_rgb: Image.Image = Image.open(\n BytesIO(base64.b64decode(png_image))\n ).convert(\"RGB\")\n buffered: BytesIO = BytesIO()\n img_rgb.save(buffered, format=\"JPEG\", quality=jpeg_image_quality)\n jpeg_image_base64: str = base64.b64encode(buffered.getvalue()).decode(\n \"utf-8\"\n )\n\n return jpeg_image_base64\n","sub_path":"deepux1_metrics/ux1/src/core/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":2331,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"391064093","text":"import numpy as np\nimport bitarray\nimport sys\nimport re\nimport math\nimport argparse\nimport csv\nfrom utils import get_model, encode_context, dfs\n\nfrom arithmetic import encode_arithmetic, decode_arithmetic\nfrom block_baseline import get_bins, encode_block, decode_block\nfrom huffman_baseline import encode_huffman, decode_huffman\nfrom sample import sample\nfrom saac import encode_saac, decode_saac\n# from base64 import *\nimport re\nimport pandas as pd\n\n# -----------------------------------------------------\n# | Harvard NLP project edited by Kieran |\n# | Feature SAAC, Arithmetic, Bins, Huffman implemen- |\n# | tations on linguistic Steganography based on Text |\n# | Generation Language Model. The Basic openAI GPT-2 |\n# | language model have been included in the directo- |\n# | ry pretrained_model. Usage mentioned below. |\n# -----------------------------------------------------\n\n# Usage:\n# python run_single.py [-mode] [-unicode_enc] [-block_size] [-temp] [-precision] [-topk] [-nucleus] [-device] [-finish_sent] [-delta] [-language_model]\n\n# Simply Usage:\n# python run_single.py\n# python run_single.py -mode \"huffman\"\n# python run_single.py -mode \"saac\" -nucleus 0.98\n\n# API likely:\n# message_str: string to be hidden. 需要被隐写的人名,比如 'Kieran'\n# context: the context related to the text generation procedure. 上下文CONTEXT,此处更改为使用同目录中其他文件\n# message: Binary stream Based on message_str. text --arithmetic encode--> binary stream 根据隐写信息(人名)编码得到的二进制流\n# text: covertext. generated text that contains secret information. 生成的含有隐写信息的文本 COVERTEXT\n# message_rec: binary stream extracted from stego_text. 对隐写文本进行隐写提取得到的二进制流\n# reconst: Decoded text. message_rec --arithmetic decode--> reconst 将隐写提取得到的二进制流进行解码得到的结果,合法输入应该也为人名\n# covertext_list: 将所有人名变化得到的covertext保存到的一个list中,可供调用。\n\n# env: Windows 10, python 3.6.12, torch 1.0.1, pytorch_transformers 1.1.0,\n# bitarray 1.0.1, CUDA 10, GTX1050.\n\n\ndef main(args):\n # Initial process\n args = vars(args)\n unicode_enc = args['unicode_enc'] # 选择编码方式 \n mode = args['mode'] # 选择隐写算法\n block_size = args['block_size'] # 隐写参数batch_size\n temp = args['temp'] # 隐写参数TEMPERATURE,注意下文中最好不要新建temp变量\n precision = args['precision'] # 隐写参数\n topk = args['topk'] # 文本生成相关参数\n device = args['device'] # device,文本生成相关参数,选择GPU/CPU,默认'cuda'\n finish_sent = args['finish_sent'] # 隐写参数\n nucleus = args['nucleus'] # saac相关隐写参数\n delta = args['delta'] # saac相关隐写参数\n model_name = args['language_model'] # 文本生成模型\n context_file = args['context_file'] # 上下文文件的位置\n message_str = args['name']\n # sample_tokens = 100 # 测试用变量\n\n # PARAMETERS 默认第一次的隐写信息(人名)\n # message_str = \"Chhenl\" # string to be hidden.\n\n # VALIDATE PARAMETERS 验证隐写算法\n if mode not in ['arithmetic', 'huffman', 'bins', 'saac']:\n raise NotImplementedError\n \n # 打印隐写信息(人名)\n print(\"Default plain_text is \", message_str)\n \n # 读取上下文\n f = open(context_file, 'r', encoding='utf-8')\n context = f.read()\n f.close()\n print(\"sample context is \", context) # related to the text generation procedure.\n\n # 加载文本生成模型\n print(\"loading GPT-2 LM to GPU\")\n enc, model = get_model(model_name=model_name)\n print(\"finish loading !\")\n\n print(\"implication of {}\".format(mode))\n \n # bins隐写算法的处理\n if mode == 'bins':\n bin2words, words2bin = get_bins(len(enc.encoder), block_size)\n\n # saac隐写算法的处理\n if delta and mode == \"saac\":\n nucleus = 2 ** (-1.0 * delta)\n\n\n\n # 以下注释都为旧调试过程中的注释\n # fix situation: directly encode the text.\n # print(\"directly encode the plain txt:\\n\", enc.encode(message_str))\n # print(\"Decode back:\\n\", enc.decode(enc.encode(message_str)))\n\n # can ensure the problem arise in the arithmetic_decode as well as the arithmetic_encode function.\n\n # ----------------------start test----------------------------\n # test_str = \"hello world.\"\n # print(\"test_str = \", test_str)\n # out = enc.encode(test_str)\n # print(\"out = \", out)\n # decode_str = enc.decode(out)\n # print(\"decode_str = \", decode_str)\n # print(\"enc.encode(decode_str) = \", enc.encode(decode_str))\n # ----------------------stop test-----------------------------\n\n # Archive Basic Initialization----------------------------------\n # print(\"plain_text is {}\".format(message_str))\n # unicode_enc = False\n # mode = 'huffman'\n # block_size = 3 # for huffman and bins\n # temp = 0.9 # for arithmetic\n # precision = 26 # for arithmetic\n # sample_tokens = 100 # for sample, delete sample\n # topk = 300\n # device = 'cuda'\n # finish_sent=False # whether or not to force finish sent. If so, stats displayed will be for non-finished sentence\n # nucleus = 0.95\n # Archive Basic Initialization----------------------------------\n\n\n\n\n\n\n first_flag = 1 # 对下文中默认处理的标志\n context_tokens = encode_context(context, enc) # 对context进行语言模型相关的编码\n\n while(1):\n # ---此处在循环中,则会不断等待输入隐写信息(人名)--------------------------------------\n # ------------------------------------------------------------------------------------\n # list_for_bpw = [] # 用于计算Bits/word参数\n # list_for_DKL = [] # 用于计算KL参数\n # list_for_seq = [] # 用于标记\n \n if first_flag == 0:\n message_str = input(\"Please reenter a new plaintext:\")\n # output_amount = len(message_str)\n \n # 得到对隐写信息(人名)的大小写集合\n message_str = message_str.upper()\n arr=list(message_str)\n generated_array = dfs(arr,0,[])\n \n first_flag = 0\n covertext_list = []\n \n for temp_count in range(0, len(generated_array)):\n # First encode message to uniform bits, without any context\n # (not essential this is arithmetic vs ascii, but it's more efficient when the message is natural language)\n \n # if temp_count > 10:\n # break # 测试时最好完成修正,此处限制输出10个COVERTEXT\n \n print(\"=\"*80)\n print(\"Altering the #{} msg_str:\".format(temp_count), message_str)\n message_str = generated_array[temp_count] # 选择一个隐写信息(比如 KiErAn)\n\n\n\n # 得到message。即上文所述的字节流\n if unicode_enc:\n ba = bitarray.bitarray()\n ba.frombytes(message_str.encode('utf-8'))\n message = ba.tolist()\n else:\n message_ctx = [enc.encoder['<|endoftext|>']]\n message_str += ''\n message = decode_arithmetic(model, enc, message_str, message_ctx, precision=40, topk=60000)\n\n\n # print(\"First encode the text to a bit sequence!\")\n # print(message) # the binary stream. text--arithmetic-->binary stream\n # print(\"the length is {}\".format(len(message)))\n\n # Next encode bits into cover text, using arbitrary context\n \n\n # 下方完成隐写算法,使用不同隐写算法将字节流嵌入进生成文本中,得到out经过GPT2的解码器得到COVERTEXT\n Hq = 0\n if mode == 'arithmetic':\n out, nll, kl, words_per_bit, Hq = encode_arithmetic(model, enc, message, context_tokens, temp=temp, finish_sent=finish_sent, precision=precision, topk=topk)\n elif mode == 'huffman':\n out, nll, kl, words_per_bit = encode_huffman(model, enc, message, context_tokens, block_size, finish_sent=finish_sent)\n elif mode == 'bins':\n out, nll, kl, words_per_bit = encode_block(model, enc, message, context_tokens, block_size, bin2words, words2bin, finish_sent=finish_sent)\n elif mode == 'saac':\n out, nll, kl, words_per_bit, Hq, topk_list, case_studies = encode_saac(model, enc, message, context_tokens, device=device, temp=temp, precision=precision, topk=topk, nucleus=nucleus)\n # add thing contains device='cuda', temp=1.0, precision=26, topk=50, nucleus=0.95.\n covertext = enc.decode(out)\n covertext_list.append(covertext) # 将所有COVERTEXT保存到一个结构中,可供调用\n\n\n\n # list_for_bpw.append(1/words_per_bit) # 用于计算参数\n # list_for_DKL.append(kl) # 用于计算参数\n # list_for_seq.append(temp_count) \n # print(\"=\"*40 + \" Encoding \" + \"=\"*40)\n\n # 打印结果,COVERTEXT,此处可以将covertext进行提取。\n print('#{} generated covertext:\\n'.format(temp_count), covertext) # covertext. generated covertext that contains secret information.\n print('ppl: %0.2f, kl: %0.3f, words/bit: %0.2f, bits/word: %0.2f, entropy: %.2f' % (math.exp(nll), kl, words_per_bit, 1/words_per_bit, Hq/0.69315))\n \n\n\n\n # -----------------------------------------------------------------------------------\n # 以下为隐写提取过程, 选择不同的隐写算法对covertext进行提取,得到字节流 MESSAGE_REC\n # Decode binary message from bits using the same arbitrary context\n \n # 下方在编写时可能会使用到,这里先注释掉,接收人将自己的名字和covertext输入进行判定。\n # input_name = input(\"Please input ur name:\")\n # input_covertext = input(\"Please input the covertext:\")\n # covertext = input_covertext\n\n\n if mode == 'arithmetic':\n message_rec = decode_arithmetic(model, enc, covertext, context_tokens, temp=temp, precision=precision, topk=topk)\n elif mode == 'huffman':\n message_rec = decode_huffman(model, enc, covertext, context_tokens, block_size)\n elif mode == 'bins':\n message_rec = decode_block(model, enc, covertext, context_tokens, block_size, bin2words, words2bin)\n elif mode == 'saac':\n message_rec = decode_saac(model, enc, covertext, context_tokens, device=device, temp=temp, precision=precision, topk=topk, nucleus=nucleus)\n\n # print(\"=\"*40 + \" Recovered Message \" + \"=\"*40)\n # print(message_rec) # binary stream extracted from stego_text.\n # print(\"=\" * 80)\n # Finally map message bits back to original text\n \n # 对字节流进行解码操作,最终得到的reconst变量即为最终隐写提取所得,正常使用应为人名。\n if unicode_enc:\n message_rec = [bool(item) for item in message_rec]\n ba = bitarray.bitarray(message_rec)\n reconst = ba.tobytes().decode('utf-8', 'ignore')\n else:\n reconst = encode_arithmetic(model, enc, message_rec, message_ctx, precision=40, topk=60000)\n # reconst = encode_arithmetic(model, enc, message_rec, message_ctx, temp=temp, precision=precision, topk=topk)\n # print(\"reconst[0] is\", format(reconst[0]))\n reconst = enc.decode(reconst[0])\n print(\"The decode text is \")\n print(reconst[0:-5]) # Decoded text. message_rec --arithmetic decode--> reconst\n \n # 这里完成基本的判断,判断此时的covertext是否指向此人名,这里对应输入设置。\n # extracted_name = reconst.upper()[0:-5]\n # if extracted_name is input_name.upper():\n # print(\"YOU ARE THE ONE! (^..^)\")\n # else:\n # print(\"PITY. ('..') \")\n\n\n\n\n\n # dataframe = pd.DataFrame({'Times':list_for_seq, 'Dkl':list_for_DKL, 'Bits/Word':list_for_bpw})\n # dataframe.to_csv(\"test_{}_temp_{}_topk_{}_prec_{}_nucleus_{:.3}.csv\".format(mode, temp, topk, precision, nucleus), index=False, sep=',')\n\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument(\"-unicode_enc\", type=bool, default=False, help=\"Whether open unicode encoding method.\")\n parser.add_argument(\"-mode\", type=str, default=\"saac\", help=\"Steganography Method.\")\n parser.add_argument(\"-block_size\", type=int, default=3, help=\"Block_size is for Huffman and Bins.\")\n parser.add_argument(\"-temp\", type=float, default=0.9, help=\"Temperature, for arithmetic and saac.\")\n parser.add_argument(\"-precision\", type=int, default=26, help=\"Precision is for arithmetic and saac.\")\n parser.add_argument(\"-topk\", type=int, default=300, help=\"top K Token, for arithmetic and saac.\")\n parser.add_argument(\"-nucleus\", type=float, default=0.95, help=\"Nucleus is for saac.\")\n parser.add_argument(\"-device\", type=str, default=\"cuda\", help=\"The basic calculator when applying model.\")\n parser.add_argument(\"-finish_sent\", type=bool, default=False, help=\"\")\n parser.add_argument(\"-delta\", type=float, default=0.01, help=\"delta for adaptive arithemtic encoding method.\")\n parser.add_argument(\"-language_model\", type=str, default=\"gpt2\", help=\"Basic Languages to generate text.\")\n parser.add_argument(\"-context_file\", type=str, default=\"./context.txt\", help=\"the basic context file\")\n parser.add_argument(\"-name\", type=str, default=\"Gogo\", help=\"Name, plz.\")\n args = parser.parse_args()\n # main()\n main(args)\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n# parser = argparse.ArgumentParser()\n# parser.add_argument(\"-plaintext\", type=str, default=\"\", help=\"your secret plaintext, use a double-quotes if necessary\")\n# parser.add_argument(\"-context\", type=str, default=\"\", help=\"context used for steganography, use a double-quotes if necessary\")\n# parser.add_argument(\"-encrypt\", type=str, default=\"arithmetic\", choices=[\"arithmetic\", \"utf8\"])\n# parser.add_argument(\"-encode\", type=str, default=\"bins\", choices=[\"bins\", \"huffman\", \"arithmetic\", \"saac\"])\n# parser.add_argument(\"-lm\", type=str, default=\"gpt2\")\n# parser.add_argument(\"-device\", type=str, default=\"0\", help=\"your gpu device id\")\n# parser.add_argument(\"-block_size\", type=int, default=4, help=\"block_size for bin/huffman encoding method\")\n# parser.add_argument(\"-precision\", type=int, default=26, help=\"precision for arithmetic encoding method\")\n# parser.add_argument(\"-temp\", type=float, default=1.0, help=\"temperature for arithemtic/huffman encoding method\")\n# parser.add_argument(\"-topK\", type=int, default=50, help=\"topK for arithemtic encoding method\")\n# parser.add_argument(\"-nucleus\", type=float, default=0.95, help=\"neclues for adaptive arithemtic encoding method\")\n# parser.add_argument(\"-delta\", type=float, default=0.01, help=\"delta for adaptive arithemtic encoding method\")\n# args = parser.parse_args()\n# main(args)\n\n# basic parameters include unicode_enc, mode, block_size, temp, precision, sample_tokens, topk, device, finish_sent, nucleus\n\n\n# 12.30, fulfil the basic function api for further implementation.\n\n\n\n\n\n\n\n\n\n# result:\n# bins:\n# [0, 1, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1, 1, 1, 1, 1, 0, 1, 0, 1, 0, 1, 1, 1, 1, 0, 0, 1, 0, 0, 0, 1, 1, 1, 0, 1, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1, 0, 1, 1, 1, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0]\n# [0, 1, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1, 1, 1, 1, 1, 0, 1, 0, 1, 0, 1, 1, 1, 1, 0, 0, 1, 0, 0, 0, 1, 1, 1, 0, 1, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1, 0, 1, 1, 1, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0]\n\n# arithmetic:\n# [0, 1, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1, 1, 1, 1, 1, 0, 1, 0, 1, 0, 1, 1, 1, 1, 0, 0, 1, 0, 0, 0, 1, 1, 1, 0, 1, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1, 0, 1, 1, 1, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0]\n# [0, 1, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1, 1, 1, 1, 1, 0, 1, 0, 1, 0, 1, 1, 1, 1, 0, 0, 1, 0, 0, 0, 1, 1, 1, 0, 1, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1, 0, 1, 1, 1, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]\n\n# huffman:\n# [0, 1, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1, 1, 1, 1, 1, 0, 1, 0, 1, 0, 1, 1, 1, 1, 0, 0, 1, 0, 0, 0, 1, 1, 1, 0, 1, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1, 0, 1, 1, 1, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0]\n# [0, 1, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1, 1, 1, 1, 1, 0, 1, 0, 1, 0, 1, 1, 1, 1, 0, 0, 1, 0, 0, 0, 1, 1, 1, 0, 1, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1, 0, 1, 1, 1, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0]\n\n# 第一处:message = decode_arithmetic(model, enc, message_str, message_ctx, precision=40, topk=60000)\n# 第二处:out, nll, kl, words_per_bit, Hq = encode_arithmetic(model, enc, message, context_tokens, temp=temp, finish_sent=finish_sent, precision=precision, topk=topk)\n# 前一个:message_rec = decode_arithmetic(model, enc, text, context_tokens, temp=temp, precision=precision, topk=topk)\n# 后一个:reconst = encode_arithmetic(model, enc, message_rec, message_ctx, precision=40, topk=60000)\n\n\n# [1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 0, 0, 1, 1, 1, 1, 0, 0, 0, 1, 1, 1, 1, 0, 0, 0, 1, 1, 0, 0, 0, 0, 1, 1, 0, 0, 0, 1, 0, 0, 1, 0, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 0, 0, 0, 0, 1, 0, 0, 1, 1, 1, 1, 0, 1, 0, 0, 1, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]\n# [1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 0, 0, 1, 1, 1, 1, 0, 0, 0, 1, 1, 1, 1, 0, 0, 0, 1, 1, 0, 0, 0, 0, 1, 1, 0, 0, 0, 1, 0, 0, 1, 0, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 0, 0, 0, 0, 1, 0, 0, 1, 1, 1, 1, 0, 1, 0, 0, 1, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]","sub_path":"NeuralSteganography-master1/run_single_bak.py","file_name":"run_single_bak.py","file_ext":"py","file_size_in_byte":19211,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"642361478","text":"import os\nimport json\nimport platform\nimport datetime\nimport binascii\n\nfrom threading import RLock\nfrom geopy import geocoders\n\nimport logging\nlogger = logging.getLogger()\n\nclass WeatherUtils:\n __lock = RLock()\n __zip_mapping = {}\n\n def get_direction(bearing):\n coords = {\n 'N': [0, 22.5],\n 'NE': [22.5, 67.5],\n 'E': [67.5, 112.5],\n 'SE': [112.5, 157.5],\n 'S': [157.5, 202.5],\n 'SW': [202.5, 247.5],\n 'W': [247.5, 292.5],\n 'NW': [292.5, 337.5],\n 'N': [337.5, 360]\n }\n for k,v in coords.items():\n if bearing >= v[0] and bearing < v[1]:\n return k\n return \"\"\n\n def get_am_pm_hour_str(timestamp):\n if platform.system() == 'Windows':\n return timestamp.strftime('%#I %p')\n else:\n return timestamp.strftime('%-I %p')\n\n def load_api_dump(url):\n if 'DEBUG' in os.environ:\n hash = binascii.crc32(url.encode('utf8'))\n debug_json = f\"/tmp/{hash}.json\"\n if os.path.exists(debug_json):\n with open(debug_json) as r:\n return json.load(r)\n\n def save_api_dump(url, r):\n if 'DEBUG' in os.environ or os.path.exists('/tmp/dump-api.flag'):\n hash = binascii.crc32(url.encode('utf8'))\n debug_json = f\"/tmp/{hash}.json\"\n with open(debug_json, \"w\") as w:\n w.write(f\"URL: {url}\\r\\n\")\n w.write(r.text)\n\n def get_gps_coordinates(zip_code):\n try:\n WeatherUtils.__lock.acquire()\n if zip_code in WeatherUtils.__zip_mapping:\n return WeatherUtils.__zip_mapping[zip_code]\n\n # geopy cannot specify zip code explicitly, so not accurate\n geolocator = geocoders.Nominatim(user_agent=\"Nook-Weather\")\n location = geolocator.geocode({\"country\":\"us\", \"postalcode\":zip_code})\n coordinates = f\"{location.latitude},{location.longitude}\"\n WeatherUtils.__zip_mapping[zip_code] = coordinates\n return coordinates\n except Exception as e:\n logger.error(f\"Failed to get gps coordinates from zip {zip_code}: {e}\")\n return None\n finally:\n WeatherUtils.__lock.release()\n","sub_path":"nook-weather/weather/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":2048,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"45391548","text":"from binance.client import Client\nfrom binance.enums import *\nfrom utils.get_mac_id import get_mac\nfrom datetime import datetime\nimport api.index as api\nfrom binance.exceptions import BinanceAPIException, BinanceOrderUnknownSymbolException\nimport scripts.telegram as tel\nfrom utils.extract_coin import extract\nfrom termcolor import colored\nfrom telethon import TelegramClient, events\nimport time\nimport re\nimport math\nimport webbrowser\n\n\nclass Bot:\n def __init__(self, data, isTrial):\n self.api_key = data[\"apiKey\"]\n self.api_secret = data[\"secret\"]\n self.useTelegramCapture = str(data[\"useTelegramCapture\"]).lower()\n self.quoteOrderQty = data[\"BtcToSpend\"]\n self.take_profit = str(data[\"takeProfit\"]).lower()\n self.takeProfitAt = data[\"takeProfitAt\"]\n self.takeProfitLimit = data[\"takeProfitLimit\"]\n self.stopLoss = str(data[\"stopLoss\"]).lower()\n self.stopLossAt = data[\"stopLossAt\"]\n self.stopLossLimit = data[\"stopLossLimit\"]\n self.timeout = data[\"Timeout\"]\n self.client = Client(self.api_key, self.api_secret)\n self.isTrial = isTrial\n self.tp_order_id, self.sl_order_id, self.oco_order = False, False, False\n self.tp_triggered, self.sl_triggered = False, False\n if self.useTelegramCapture == \"true\":\n self.initialise_telegram_client(data)\n\n def initialise_telegram_client(self, data):\n self.telegram_App_id = data[\"telegram_App_id\"]\n self.telegram_api_hash = data[\"telegram_api_hash\"]\n self.channel_id = data[\"channel_id\"]\n self.Test_channel_id = data[\"Test_channel_id\"]\n self.coin_extract_status = False\n self.coin_symbol = \"\"\n self.telegram_client = TelegramClient(\n \"anon\", self.telegram_App_id, self.telegram_api_hash\n )\n\n def create_market_order(self):\n try:\n order = self.client.order_market_buy(\n symbol=self.coin_symbol,\n quoteOrderQty=self.quoteOrderQty,\n )\n return order\n except BinanceAPIException as e:\n print(colored(f\"Market Buy Error - {e}\", \"red\"))\n\n def create_market_order_sell(self):\n try:\n order = self.client.order_market_sell(\n symbol=self.coin_symbol,\n quantity=self.quantity_brought,\n )\n return order\n except BinanceAPIException as e:\n print(colored(f\"Market Sell Error - {e}\", \"red\"))\n\n def float_precision(self, f, n):\n n = int(math.log10(1 / float(n)))\n f = math.floor(float(f) * 10 ** n) / 10 ** n\n f = \"{:0.0{}f}\".format(float(f), n)\n return str(int(f)) if int(n) == 0 else f\n\n def get_price(self):\n price = None\n tickers = self.client.get_all_tickers()\n for ticker in tickers:\n if ticker[\"symbol\"] == self.coin_symbol:\n price = float(ticker[\"price\"])\n return price\n\n def get_info(self):\n try:\n tick_size = None\n step_size = None\n symbol_info = self.client.get_symbol_info(self.coin_symbol)\n for filt in symbol_info[\"filters\"]:\n if filt[\"filterType\"] == \"PRICE_FILTER\":\n tick_size = float(filt[\"tickSize\"])\n elif filt[\"filterType\"] == \"LOT_SIZE\":\n step_size = float(filt[\"stepSize\"])\n return tick_size, step_size\n except TypeError as e:\n print(colored(f\"Wrong Coin Name Entered - {e}\", \"red\"))\n\n def get_asset_info(self):\n try:\n self.tick_size, self.step_size= self.get_info()\n except TypeError as e:\n print(colored(f\"Wrong Coin Name Entered - {e}\", \"red\"))\n\n def Average(self, lst):\n return sum(lst) / len(lst)\n\n def calculate_target_price(self, price, percent_change, loss):\n if loss:\n final_price = price - ((percent_change / 100) * price)\n else:\n final_price = price + ((percent_change / 100) * price)\n return final_price\n\n def set_take_profit(self):\n # PERCENT_PRICE Filter Check\n if(self.takeProfitAt>=399):\n print(colored('WARNING : Your Take Profit value is too high, Binance allows only 400% TP to be set','yellow'))\n self.takeProfitAt=398\n if(self.takeProfitLimit>=399):\n print(colored('WARNING : Your Take Profit Limit value is too high, Binance allows only 400% TP to be set','yellow'))\n self.takeProfitLimit=398\n take_profit_price = self.calculate_target_price(\n self.price_brought, self.takeProfitLimit, False\n )\n stop_price = self.calculate_target_price(\n self.price_brought, self.takeProfitAt, False\n )\n price_formatted = self.float_precision(take_profit_price, self.tick_size)\n stop_price_formatted = self.float_precision(stop_price, self.tick_size)\n try:\n order_take_ptf = self.client.create_order(\n symbol=self.coin_symbol,\n side=SIDE_SELL,\n type=ORDER_TYPE_TAKE_PROFIT_LIMIT,\n quantity=self.quantity_brought,\n price=price_formatted,\n stopPrice=stop_price_formatted,\n timeInForce=TIME_IN_FORCE_GTC,\n )\n return order_take_ptf\n except BinanceAPIException as e:\n print(colored(f\"Take Profit Error - {e}\", \"red\"))\n\n def set_stop_loss(self):\n # PERCENT_PRICE Filter Check\n if(self.stopLossAt>=80):\n print(colored('WARNING : Your Stop Loss value is too high, Binance allows only 80% stopLoss to be set','yellow'))\n self.stopLossAt=79\n if(self.stopLossLimit>=80):\n print(colored('WARNING : Your Stop Loss Limit value is too high, Binance allows only 80% stopLoss to be set','yellow'))\n self.stopLossLimit=79\n stop_loss_price = self.calculate_target_price(\n self.price_brought, self.stopLossLimit, True\n )\n stop_price = self.calculate_target_price(\n self.price_brought, self.stopLossAt, True\n )\n price_formatted = self.float_precision(stop_loss_price, self.tick_size)\n stop_price_formatted = self.float_precision(stop_price, self.tick_size)\n try:\n order_stop_loss = self.client.create_order(\n symbol=self.coin_symbol,\n side=SIDE_SELL,\n type=ORDER_TYPE_STOP_LOSS_LIMIT,\n quantity=self.quantity_brought,\n price=price_formatted,\n stopPrice=stop_price_formatted,\n timeInForce=TIME_IN_FORCE_GTC,\n )\n return order_stop_loss\n except BinanceAPIException as e:\n print(colored(f\"Stop Loss Error - {e}\", \"red\"))\n\n def set_oco_order(self):\n take_profit_price = self.float_precision(\n self.calculate_target_price(\n self.price_brought, self.takeProfitLimit, False\n ),\n self.tick_size,\n )\n stop_loss_price = self.float_precision(\n self.calculate_target_price(self.price_brought, self.stopLossLimit, True),\n self.tick_size,\n )\n stop_price = self.float_precision(\n self.calculate_target_price(self.price_brought, self.stopLossAt, True),\n self.tick_size,\n )\n try:\n oco_order = self.client.create_oco_order(\n symbol=self.coin_symbol,\n side=SIDE_SELL,\n stopLimitTimeInForce=TIME_IN_FORCE_GTC,\n quantity=self.quantity_brought,\n stopPrice=stop_price,\n stopLimitPrice=stop_loss_price,\n price=take_profit_price,\n )\n return oco_order\n except BinanceAPIException as e:\n print(colored(f\"OCO order Error - {e}\", \"red\"))\n\n def openCurrencyChart(self):\n curreny_url_binance = (\n f\"https://www.binance.com/en/trade/{self.coin_currency}_BTC?layout=basic\"\n )\n webbrowser.open(curreny_url_binance)\n\n def start(self):\n self.coin_currency = self.coin_symbol[0 : len(self.coin_symbol) - 3]\n self.get_asset_info()\n market_order = self.create_market_order()\n price_brought_list, commission = [], 0\n if market_order:\n for i in market_order[\"fills\"]:\n price_brought_list.append(float(i[\"price\"]))\n if i[\"commissionAsset\"] == self.coin_currency:\n commission += float(i[\"commission\"])\n self.quantity_brought = float(\n self.float_precision(\n float(market_order[\"executedQty\"]) - commission, self.step_size\n )\n )\n self.price_brought = self.Average(price_brought_list)\n\n print(\n f\"Market Buy Successful at price {colored(self.price_brought,'green')} and quantity brought - {colored(self.quantity_brought,'green')}\"\n )\n if self.take_profit == \"true\" and self.stopLoss != \"true\":\n take_profit_order = self.set_take_profit()\n if take_profit_order:\n print(colored(f\"Take Profit Successfully set!\", \"green\"))\n self.tp_order_id = take_profit_order[\"orderId\"]\n if self.stopLoss == \"true\" and self.take_profit != \"true\":\n stop_loss_order = self.set_stop_loss()\n if stop_loss_order:\n print(colored(f\"Stop Loss Successfully set!\", \"green\"))\n self.sl_order_id = stop_loss_order[\"orderId\"]\n if self.take_profit == \"true\" and self.stopLoss == \"true\":\n self.oco_order = self.set_oco_order()\n if self.oco_order:\n print(\n colored(f\"Take Profit and Stop Loss Successfully Set!\", \"green\")\n )\n self.openCurrencyChart()\n if self.timeout > 0:\n t = self.timeout\n while t:\n mins, secs = divmod(t, 60)\n timer = \"{:02d}:{:02d}\".format(mins, secs)\n print(colored(timer, \"yellow\"), end=\"\\r\")\n time.sleep(1)\n t -= 1\n if self.tp_order_id:\n try:\n order_status_tp = self.client.get_order(\n symbol=self.coin_symbol, orderId=self.tp_order_id\n )\n if order_status_tp[\"status\"] == \"FILLED\":\n self.tp_triggered = True\n except BinanceAPIException as e:\n print(colored(f\"Failed to get TP Order - {e}\", \"red\"))\n elif self.sl_order_id:\n try:\n order_status_sl = self.client.get_order(\n symbol=self.coin_symbol, orderId=self.sl_order_id\n )\n if order_status_sl[\"status\"] == \"FILLED\":\n self.sl_triggered = True\n except BinanceAPIException as e:\n print(colored(f\"Failed to get SL Order - {e}\", \"red\"))\n elif self.oco_order:\n try:\n order_status_oco = self.client.get_order(\n symbol=self.coin_symbol,\n orderId=self.oco_order[\"orders\"][1][\"orderId\"],\n )\n if order_status_oco[\"status\"] == \"FILLED\":\n self.tp_triggered = True\n elif order_status_oco[\"status\"] == \"EXPIRED\":\n self.sl_triggered = True\n except BinanceAPIException as e:\n print(colored(f\"Failed to get OCO Order - {e}\", \"red\"))\n if not self.tp_triggered and not self.sl_triggered:\n if self.tp_order_id:\n try:\n self.client.cancel_order(\n symbol=self.coin_symbol, orderId=self.tp_order_id\n )\n except BinanceAPIException as e:\n print(colored(f\"Failed to Cancel TP order - {e}\", \"red\"))\n if self.sl_order_id:\n try:\n self.client.cancel_order(\n symbol=self.coin_symbol, orderId=self.sl_order_id\n )\n except BinanceAPIException as e:\n print(colored(f\"Failed to Cancel SL order - {e}\", \"red\"))\n if self.oco_order:\n try:\n self.client.cancel_order(\n symbol=self.coin_symbol, orderId=self.oco_order[\"orders\"][1][\"orderId\"]\n )\n except BinanceAPIException as e:\n print(colored(f\"Failed to Cancel OCO order - {e}\", \"red\"))\n print(\n colored(\n \"None of Take Profit or Stop Loss were triggered, Selling ASAP !\",\n \"yellow\",\n )\n )\n\n market_sell_order = self.create_market_order_sell()\n if market_sell_order and market_sell_order[\"status\"] == \"FILLED\":\n print(colored(\"Market Sell Successful !\", \"green\"))\n if market_sell_order and market_sell_order[\"status\"] != \"FILLED\":\n print(\n colored(\n \"Market Sell Partially Successfull or Failed , Please check on Binance.com\",\n \"yellow\",\n )\n )\n else:\n if self.tp_triggered:\n print(\n colored(\n \"Congrats, Your Take Profit Order was triggered !\",\n \"green\",\n )\n )\n if self.sl_triggered:\n print(colored(\"Your Stop Loss Order was triggered !\", \"green\"))\n\n\n def initialise(self):\n if self.useTelegramCapture == \"true\":\n\n @self.telegram_client.on(events.NewMessage)\n async def my_event_handler(event):\n if event.chat_id == self.channel_id:\n self.coin_symbol = (extract(event.raw_text) + \"BTC\").upper()\n if re.match(\"^[A-Z0-9-_.]{1,20}$\", self.coin_symbol):\n print(\n f\"Coin to pump detected - {colored(self.coin_symbol,'green')}\"\n )\n self.coin_extract_status = True\n\n await self.telegram_client.disconnect()\n\n self.telegram_client.start()\n print(\"Listening For messages (press Ctrl+c to exit) !\")\n self.telegram_client.run_until_disconnected()\n if not self.coin_extract_status:\n print(colored(\"Failed to detect coin, enter manually !\", \"yellow\"))\n self.coin_symbol = (input() + \"BTC\").upper()\n if self.coin_symbol != \"\":\n self.start()\n else:\n print(\"Enter Coin name\")\n self.coin_symbol = input().upper() + \"BTC\"\n print(f\"Coin Entered - {colored(self.coin_symbol,'green')}\")\n self.start()\n\n\ndef main(data, isTrial):\n client = Bot(data, isTrial)\n client.initialise()\n","sub_path":"client/scripts/index.py","file_name":"index.py","file_ext":"py","file_size_in_byte":15846,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"72453942","text":"import cv2 as cv\n\nheight = 0\nwidth = 0\n\ndef last(x):\n\t\"\"\"Return an integer value, the least significant bit of the argument.\"\"\"\n\tx = bin(x)\n\treturn x[len(x)-1]\n\ndef get_ascii(arr_list):\n\t\"\"\"Return a character, ascii equivalent of a number\"\"\"\n\tdeci = 0\n\tfor i in range(len(arr_list)):\n\t\tif arr_list[i] == '1':\n\t\t\tdeci = deci + pow(2,len(arr_list)-i)\n\n\treturn str(deci)\n\t\t\ndef getnext(list1,pos):\n\t\"\"\"Returns next character in message\"\"\"\n\tsum = 0\n\tk = 7\n\tfor i in range(pos,pos+8):\n\t\tif list1[i]=='1':\n\t\t\tsum = sum+pow(2,k)\n\t\tk = k-1\n\n\treturn chr(sum)\n\ndef get_message_size():\n\t\"\"\"Determines the size of message\"\"\"\n\tfile_obj = open(\"Res/user_input.txt\",\"r\")\n\tstring = file_obj.readline()\n\treturn len(string)\n\ndef write_To_File(message):\n\t\"\"\"Writes the decoded message to a file named Decoded.txt\"\"\"\n\tfile_obj = open(\"Res/decoded.txt\",\"w\")\n\tfile_obj.write(message)\n\n\ndef decode():\n\t\"\"\"main decoding logic\"\"\"\n\t\n\t''' Reading the steganographed Imgae file and calculating height and width of the image'''\n\tsteganoImage = cv.imread(\"Res/steganographed_image.png\",1)\n\theight,width = steganoImage.shape[:2]\n\n\t''' Initializes the message as null'''\n\tmessage = \"\"\n\n\t''' Initializes the starting conditions'''\n\tcolumn=0\n\trow=0\n\tcolor=0\n\n\t''' Last bits is an array that holds last bits of pixel values.'''\n\tlast_bits = []\n\n\tfor row in range(height):\n\t\tfor column in range(width):\n\t\t\tlast_bits.append(last(steganoImage[row][column][color]))\n\t\t\tif (column+1)%width == 0:\n\t\t\t\trow = row+1\n\t\t\t\tif row == height:\n\t\t\t\t\trow = 0\n\t\t\t\t\tcolor = color+1\n\t\t\tcolumn = (column+1)%width\n\n\tmessage_size = get_message_size()\n\n\ti=0\n\twhile True:\n\t\tnextChar = getnext(last_bits,i)\n\t\tif(nextChar == '`'):\n\t\t\ttemp = getnext(last_bits,i)\n\t\t\ti = i+8\n\t\t\ttemp2 = getnext(last_bits,i)\n\t\t\tif(temp == '`'):\n\t\t\t\tif(temp2 == '`'):\n\t\t\t\t\tbreak\n\t\t\telse:\n\t\t\t\tmessage+=temp\n\t\t\t\tmessage+=temp2\n\t\tmessage+=nextChar\n\t\ti = i+8\n\t\t\n\twrite_To_File(message)\n\ndef main():\n\tdecode()\n\t\nif __name__ == '__main__':\n\tmain()","sub_path":"LSB_Linear_Ret.py","file_name":"LSB_Linear_Ret.py","file_ext":"py","file_size_in_byte":1958,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"322692822","text":"import logging\nimport os\nfrom abc import ABC, abstractmethod\nfrom typing import Dict, Tuple\n\nimport numpy as np\nfrom keras import Model\nfrom keras import backend as keras_backend\nfrom keras import layers\nfrom keras import losses\nfrom keras import optimizers\n\nfrom modules.generators import generators\nfrom modules.image import image, image_batch, image_utils\nfrom modules.model import denseblock, callbacks, model_utils\n\nDEFAULTS = {\n 'lr': 0.001,\n 'input_shape': [256, 256, 3],\n 'batch_size': 16,\n 'image_op': 'derez',\n 'training_directory': os.path.join('data', 'training'),\n 'validation_directory': os.path.join('data', 'validation'),\n 'evaluation_directory': os.path.join('data', 'evaluation'),\n 'steps': 1000,\n}\n\nMODULE_VERSION = '0'\n\nSUPPORTED_IMAGE_OPS_VERSION_TABLE = {\n 'blur': '0',\n 'derez': '1',\n 'blur_derez': '2',\n}\n\nLOG = logging.getLogger(__name__)\n\n\ndef psnr_loss(y_true, y_pred):\n \"\"\" PSNR is Peak Signal to Noise Ratio, defined below\n PSNR = 20 * log10(MAXp) - 10 * log10(MSE)\n MAXp = maximum pixel value.\n Our framework scales to [0,1] range, so MAXp = 1.\n The 20 * log10(MAXp) reduces to 0\n\n https://en.wikipedia.org/wiki/Peak_signal-to-noise_ratio\n \"\"\"\n\n return -10.0 * keras_backend.log(\n losses.mean_squared_error(y_true, y_pred)\n ) / keras_backend.log(10.0)\n\n\nclass SRModel(ABC):\n\n def __init__(self, config: Dict, model_version: str='0'):\n self._config = self._add_defaults(\n config=config,\n model_version=model_version\n )\n\n constructed_model = self._create_model(config=self._config)\n\n self._model = self._load_or_save_model_weights(\n model=constructed_model,\n config=self._config\n )\n\n model_utils.save_config(self._config)\n # only initialized before training\n self._image_generator = None\n\n def _add_defaults(self, config: Dict, model_version: str) -> Dict:\n \"\"\"Adds essential default values to a config dictionary via Default\n \"\"\"\n for key in DEFAULTS.keys():\n config[key] = config.get(key, DEFAULTS[key])\n\n config['name'] = config.get('name', self.__class__.__name__)\n\n image_op_version = SUPPORTED_IMAGE_OPS_VERSION_TABLE[config['image_op']]\n\n config['version'] = \\\n f'v{MODULE_VERSION}.{model_version}.{image_op_version}'\n\n # handles JSON not allowing tuples\n config['input_shape'] = tuple(config['input_shape'])\n\n # Handles the fpaths is any are missing\n path_keys = {\n 'model_fpath': 'model.hdf5',\n 'config_fpath': 'config.json',\n 'log_fpath': 'log.csv',\n }\n # Maps the file extensions for each key, so the path can be constructed easily\n for path_key in path_keys.keys():\n if config.get(path_key) is None:\n config[path_key] = os.path.join(\n self._save_name(config['version']),\n path_keys[path_key]\n )\n\n return config\n\n @property\n def version(self):\n return self._config['version']\n\n @staticmethod\n def _save_name(version: str) -> str:\n \"\"\"Name for saving and reading stored weights\n\n :param version: Version of model\n :return: An initialized directory path\n \"\"\"\n dir_path = os.path.join('configs', version)\n if not os.path.isdir(dir_path):\n os.mkdir(dir_path)\n return dir_path\n\n @staticmethod\n def _load_or_save_model_weights(model: Model, config: Dict) -> Model:\n \"\"\"Loads model weight if they exist, else saves them to model_fpath\n\n :param model: Constructed Model architecture\n :param config: configuration file\n :return: Model with loaded weight (if any)\n \"\"\"\n if os.path.isfile(config['model_fpath']):\n model.load_weights(config['model_fpath'])\n LOG.info(\n 'Loaded model weights from {}'.format(\n config['model_fpath'])\n )\n else:\n model.save_weights(config['model_fpath'])\n LOG.info(\n 'Initialized new model weights to {}'.format(\n config['model_fpath'])\n )\n return model\n\n @staticmethod\n def _initialize_image_generator_with_data(\n config: Dict) -> generators.SRImageGenerator:\n \"\"\"Initializes an SRImageGenerator from a configuration\n\n :param config: configuration dictionary\n :return: SRImageGenerator with data specified by configuration data paths\n \"\"\"\n return generators.SRImageGenerator(\n training_data=image_batch.PILImageBatch.open_from_list(\n image_utils.get_image_paths_from_dir(\n dir_path=config['training_directory'],\n image_file_extension='.png'\n )\n ),\n validation_data=image_batch.PILImageBatch.open_from_list(\n image_utils.get_image_paths_from_dir(\n dir_path=config['validation_directory'],\n image_file_extension='.png'\n )\n ),\n config=config\n )\n\n @abstractmethod\n def _create_model(self, config: Dict) -> Model:\n \"\"\"Creates the model\n\n :param config: Configuration file\n :return: ImageSR model\n \"\"\"\n pass\n\n def train(self, epochs: int):\n \"\"\"Trains the model for a given number of epochs.\n\n :param epochs: Number of epochs to train the model\n \"\"\"\n # initialize the image generator if not already initialized\n if self._image_generator is None:\n self._image_generator = \\\n self._initialize_image_generator_with_data(\n config=self._config\n )\n\n initial_epoch = self._config.get('epoch', 0)\n LOG.info(\n f'Training model for {epochs} epochs, starting at {initial_epoch}'\n )\n\n self._model.compile(\n optimizer=optimizers.Adam(\n lr=self._config['lr']\n ),\n loss='mse',\n metrics=[psnr_loss]\n )\n\n model_callbacks = callbacks.callback_list(self._config)\n steps = self._config['steps']\n\n self._model.fit_generator(\n self._image_generator.training_generator(),\n steps_per_epoch=steps,\n epochs=epochs+initial_epoch,\n verbose=LOG.getEffectiveLevel() <= logging.INFO,\n callbacks=model_callbacks,\n validation_data=self._image_generator.validation_generator(),\n validation_steps=steps // 10,\n initial_epoch=initial_epoch)\n\n def enhance(self,\n img: image.PILImage,\n output_size: Tuple[int, int]) -> image.PILImage:\n \"\"\"Performs the enhancement algorithm on a given image\n\n The basic algorithm is to increase the size of the image by a factor\n of at most 2 until it reaches the correct size\n\n :param img: Original PIL image\n :param output_size: desired output shape\n :return: Enhanced PILImage at the new output_size\n \"\"\"\n output_img = img\n while output_img.size != output_size:\n new_size: Tuple[int, int] = (\n min(2*output_img.size[0], output_size[0]),\n min(2*output_img.size[1], output_size[1]),\n )\n output_img_at_new_size = output_img.resize(size=new_size)\n\n # apply a blur if a blur was used in training\n if 'blur' in self._config['image_op']:\n output_img_at_new_size = output_img_at_new_size.blur()\n\n # numpyize the image for the model\n np_image = output_img_at_new_size.numpyize()\n\n # image needs to be a rank 4 input (1,) + image_shape\n model_input = np.expand_dims(np_image, axis=0)\n\n # use the model to predict the residual image\n residual_image = self._model.predict(model_input, batch_size=1)[0]\n\n # add the residual to the original\n enhanced_np_image = np_image + residual_image\n\n # clip the values to the expected [0, 1] range\n enhanced_np_image = enhanced_np_image.clip(min=0.0, max=1.0)\n\n # set back to PIL image\n output_img = image.PILImage.from_numpy_array(enhanced_np_image)\n\n return output_img\n","sub_path":"modules/model/model.py","file_name":"model.py","file_ext":"py","file_size_in_byte":8483,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"429224458","text":"# Copyright 2020 Mechanics of Microstructures Group\n# at The University of Manchester\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport numpy as np\n\nfrom sklearn.cluster import MeanShift\nfrom scipy.stats import linregress\nimport pandas as pd\n\nfrom defdap.plotting import Plot, GrainPlot, LineSlice\n\n\nclass GrainInspector:\n \"\"\"\n Class containing the interactive grain inspector tool for slip trace analysis\n and relative displacement ratio analysis.\n\n \"\"\"\n def __init__(self, currMap, vmax=0.1):\n # Initialise some values\n self.grainID = 0\n self.currMap = currMap\n self.currEBSDMap = self.currMap.ebsdMap\n self.currDICGrain = self.currMap[self.grainID]\n self.currEBSDGrain = self.currDICGrain.ebsdGrain\n self.vmax = vmax\n \n # Draw the figure\n self.draw()\n\n def draw(self):\n \"\"\" Draw the main window, buttons, text boxes and axes.\n\n \"\"\"\n # Plot window\n self.plot = Plot(ax=None, makeInteractive=True, figsize=(14,8), title='Grain Inspector')\n \n # Buttons\n self.plot.addButton('Save\\nLine', self.saveLine, (0.73, 0.48, 0.05, 0.04))\n self.plot.addButton('Previous\\nGrain', lambda e, p: self.gotoGrain(self.grainID-1, p), (0.73, 0.94, 0.05, 0.04))\n self.plot.addButton('Next\\nGrain', lambda e, p: self.gotoGrain(self.grainID+1, p), (0.79, 0.94, 0.05, 0.04))\n self.plot.addButton('Run\\nAll STA', self.batchRunSTA, (0.81, 0.02, 0.1, 0.04))\n self.plot.addButton('Clear\\nAll Lines', self.clearAllLines, (0.89, 0.48, 0.05, 0.04))\n\n # Text boxes\n self.plot.addTextBox(label='Go to \\ngrain ID:', loc=(0.9, 0.94, 0.05, 0.04), submitHandler=self.gotoGrain)\n self.plot.addTextBox(label='Remove\\nID:', loc=(0.83, 0.48, 0.05, 0.04), submitHandler=self.removeLine)\n self.RDRGroupBox = self.plot.addTextBox(label='Run RDR\\non group:',\n loc=(0.78, 0.07, 0.05, 0.04), submitHandler=self.runRDRGroup)\n\n # Axes\n self.maxShearAx = self.plot.addAxes((0.05, 0.4, 0.65, 0.55))\n self.slipTraceAx = self.plot.addAxes((0.2, 0.05, 0.6, 0.3))\n self.unitCellAx = self.plot.addAxes((0.05, 0.055, 0.15, 0.3), proj='3d')\n self.grainInfoAx = self.plot.addAxes((0.73, 0.86, 0.25, 0.06))\n self.lineInfoAx = self.plot.addAxes((0.73, 0.55, 0.25, 0.3))\n self.groupsInfoAx = self.plot.addAxes((0.73, 0.15, 0.25, 0.3))\n self.grainPlot = self.currMap[self.grainID].plotMaxShear(fig=self.plot.fig, ax=self.maxShearAx, \n vmax=self.vmax, plotScaleBar=True, plotColourBar=True)\n self.plot.ax.axis('off')\n \n # Draw the stuff that will need to be redrawn often in a seperate function\n self.redraw()\n\n def gotoGrain(self, event, plot):\n \"\"\" Go to a specified grain ID.\n\n Parameters\n ----------\n event: int\n Grain ID to go to.\n\n \"\"\"\n ## Go to grain ID specified in event\n self.grainID=int(event)\n self.grainPlot.arrow=None\n self.currDICGrain = self.currMap[self.grainID]\n self.currEBSDGrain = self.currDICGrain.ebsdGrain\n self.redraw()\n\n def saveLine(self, event, plot):\n \"\"\" Save the start point, end point and angle of drawn line into the grain.\n\n Parameters\n ----------\n event: numpy.ndarray\n Start x, start y, end x, end y point of line passed from drawn line.\n\n \"\"\"\n # Get angle of lines\n lineAngle = 90-np.rad2deg(np.arctan2(self.drawnLine.points[3]-self.drawnLine.points[1], \n self.drawnLine.points[2]-self.drawnLine.points[0]))\n if lineAngle > 180: lineAngle -= 180\n elif lineAngle < 0: lineAngle += 180\n #lineAngle += self.currMap.ebsdTransform.rotation*-180/np.pi\n \n # Save drawn line to the DIC grain\n self.currDICGrain.pointsList.append([self.drawnLine.points, lineAngle, -1])\n \n # Group lines and redraw\n self.groupLines()\n self.redraw()\n \n def groupLines(self):\n \"\"\"\n Group the lines drawn in the current grain item using a mean shift algorithm,\n save the average angle and detect the active slip planes.\n\n \"\"\"\n angles = [x[1] for x in self.currDICGrain.pointsList]\n # For single line, don't group\n if len(angles) == 1:\n self.currDICGrain.pointsList[0][2]=0\n self.currDICGrain.groupsList = [[0, angles[0], 0, 0, 0]]\n else:\n # Run clustering algorithm for >1 line\n ms = MeanShift(bandwidth=10).fit(np.matrix([range(len(angles)), angles]).transpose())\n \n # Add group ID for each line to the points list\n for i, label in enumerate(ms.labels_):\n self.currDICGrain.pointsList[i][2] = label\n \n # Make array of groups with mean angle\n self.currDICGrain.groupsList = []\n for i in range(np.max(ms.labels_+1)):\n self.currDICGrain.groupsList.append([i, ms.cluster_centers_[i][1], 0, 0, 0])\n \n # Detect active slip systems in each group\n for group in self.currDICGrain.groupsList:\n activePlanes = []\n deviation = []\n experimentalAngle = group[1]\n for idx, theoreticalAngle in enumerate(np.rad2deg(self.currEBSDGrain.slipTraceAngles)):\n if theoreticalAngle-5 < experimentalAngle < theoreticalAngle+5:\n activePlanes.append(idx)\n deviation.append(experimentalAngle-theoreticalAngle)\n group[2] = activePlanes\n group[3] = deviation\n \n def clearAllLines(self, event, plot):\n \"\"\" Clear all lines in a given grain.\n\n \"\"\"\n\n self.currDICGrain.pointsList = []\n self.currDICGrain.groupsList = []\n self.redraw()\n\n def removeLine(self, event, plot):\n \"\"\" Remove single line [runs after submitting a text box].\n\n Parameters\n ----------\n event: int\n Line ID to remove.\n\n \"\"\"\n ## Remove single line\n del self.currDICGrain.pointsList[int(event)]\n self.redraw()\n\n def redraw(self):\n \"\"\"\n Draw items which need to be redrawn often (i.e. when changing grain ID).\n\n \"\"\"\n\n # Plot max shear for grain\n self.maxShearAx.clear()\n grainPlot = self.currMap[self.grainID].plotMaxShear(\n fig=self.plot.fig,ax=self.maxShearAx, vmax=self.vmax, plotColourBar=False, plotScaleBar=True)\n\n # Draw slip traces\n self.slipTraceAx.clear()\n self.slipTraceAx.set_aspect('equal', 'box')\n slipPlot = GrainPlot(fig=self.plot.fig, callingGrain=self.currMap[self.grainID], ax=self.slipTraceAx)\n traces = slipPlot.addSlipTraces(topOnly=True)\n self.slipTraceAx.axis('off')\n \n # Draw slip bands\n bands = [elem[1] for elem in self.currDICGrain.groupsList]\n if self.currDICGrain.groupsList != None:\n slipPlot.addSlipBands(topOnly=True, angles=list(np.deg2rad(bands)))\n \n # Draw unit cell\n self.unitCellAx.clear()\n self.currEBSDGrain.plotUnitCell(fig=self.plot.fig, ax=self.unitCellAx)\n \n # Write grain info text\n self.grainInfoAx.clear()\n self.grainInfoAx.axis('off')\n grainInfoText = 'Grain ID: {0} / {1}\\n'.format(self.grainID, len(self.currMap.grainList))\n grainInfoText += 'Min: {0:.1f} % Mean:{1:.1f} % Max: {2:.1f} %'.format(\n np.min(self.currDICGrain.maxShearList)*100,\n np.mean(self.currDICGrain.maxShearList)*100,\n np.max(self.currDICGrain.maxShearList)*100)\n self.plot.addText(self.grainInfoAx, 0, 1, grainInfoText, va='top', ha='left', fontsize=10)\n \n # Detect lines\n self.drawnLine = LineSlice(ax=self.maxShearAx, fig=self.plot.fig, action=self.grainPlot.addArrow)\n\n # Write lines text and draw lines\n linesTxt = 'List of lines\\n\\nLineID x0 y0 x1 y1 Angle Group\\n'\n\n if self.currDICGrain.pointsList != []:\n for idx, points in enumerate(self.currDICGrain.pointsList):\n linesTxt += '{0} {1:.1f} {2:.1f} {3:.1f} {4:.1f} {5:.1f} {6}\\n'.format(idx,\n points[0][0], points[0][1], points[0][2], points[0][3], points[1], points[2])\n self.grainPlot.addArrow(startEnd=points[0], clearPrev=False, persistent=True, label=idx)\n \n self.lineInfoAx.clear()\n self.lineInfoAx.axis('off')\n self.plot.addText(self.lineInfoAx, 0, 1, linesTxt, va='top', fontsize=10)\n \n # Write groups info text\n groupsTxt = 'List of groups\\n\\nGroupID Angle System Dev RDR\\n'\n if self.currDICGrain.groupsList != []:\n for idx, group in enumerate(self.currDICGrain.groupsList):\n groupsTxt += '{0} {1:.1f} {2} {3} {4:.2f}\\n'.format(\n idx, group[1], group[2], np.round(group[3],3), group[4])\n\n self.groupsInfoAx.clear()\n self.groupsInfoAx.axis('off')\n self.plot.addText(self.groupsInfoAx, 0, 1, groupsTxt, va='top', fontsize=10)\n\n def runRDRGroup(self, event, plot):\n \"\"\" Run RDR on a specified group, upon submitting a text box.\n\n Parameters\n ----------\n event: int\n Group ID specified from text box.\n\n \"\"\"\n ## Run RDR for group of lines\n if event != '':\n self.calcRDR(grain = self.currDICGrain, group=int(event))\n self.RDRGroupBox.set_val('')\n \n def batchRunSTA(self, event, plot):\n \"\"\" Run slip trace analysis on all grains which hve slip trace lines drawn.\n\n \"\"\"\n\n # Print header\n print(\"Grain\\tEul1\\tEul2\\tEul3\\tMaxSF\\tGroup\\tAngle\\tSystem\\tDev\\RDR\")\n \n # Print information for each grain\n for idx, grain in enumerate(self.currMap):\n if grain.pointsList != []:\n for group in grain.groupsList:\n maxSF = np.max([item for sublist in grain.ebsdGrain.averageSchmidFactors for item in sublist])\n eulers = self.currEBSDGrain.refOri.eulerAngles()*180/np.pi\n text = '{0}\\t{1:.1f}\\t{2:.1f}\\t{3:.1f}\\t{4:.3f}\\t'.format(\n idx, eulers[0], eulers[1], eulers[2], maxSF)\n text += '{0}\\t{1:.1f}\\t{2}\\t{3}\\t{4:.2f}'.format(\n group[0], group[1], group[2], np.round(group[3],3), group[4])\n print(text)\n\n def calcRDR(self, grain, group, showPlot=True, length=2.5):\n \"\"\" Calculates the relative displacement ratio for a given grain and group.\n\n Parameters\n ----------\n grain: int\n DIC grain ID to run RDR on.\n group: int\n group ID to run RDR on.\n showPlot: bool\n if True, show plot window.\n length: int\n length of perpendicular lines used for RDR.\n\n \"\"\"\n \n ulist=[]; vlist=[]; allxlist = []; allylist = []; \n\n # Get all lines belonging to group\n points = []\n for point in grain.pointsList:\n if point[2] == group:\n points.append(point[0])\n\n for point in points:\n x0=point[0]; y0=point[1]; x1=point[2]; y1=point[3];\n grad = (y1-y0)/(x1-x0)\n invgrad = -1/grad\n profile_length = np.sqrt((y1-y0)**2+(x1-x0)**2)\n num = np.round(profile_length*2)\n \n ### Calculate positions for each point along slip trace line (x,y)\n x, y = np.round(np.linspace(x0, x1, int(num))), np.round(np.linspace(y0, y1, int(num)))\n df = pd.DataFrame({'x':x, 'y':y}).drop_duplicates()\n x,y = df['x'].values.tolist(),df['y'].values.tolist()\n\n ## Calculate deviation from (0,0) for points along line with angle perpendicular to slip line (xnew,ynew)\n x0new = np.sqrt(length/(invgrad**2+1))*np.sign(grad)\n y0new = -np.sqrt(length/(1/invgrad**2+1))\n x1new = -np.sqrt(length/(invgrad**2+1))*np.sign(grad)\n y1new = np.sqrt(length/(1/invgrad**2+1))\n profile_length=np.sqrt((y1new-y0new)**2+(x1new-x0new)**2)\n num = np.round(profile_length)\n xnew, ynew = np.linspace(x0new, x1new, int(num)), np.linspace(y0new, y1new, int(num))\n xnew, ynew = np.around(xnew).astype(int), np.around(ynew).astype(int)\n df = pd.DataFrame({'x':xnew, 'y':ynew}).drop_duplicates()\n xnew,ynew = df['x'].values.tolist(), df['y'].values.tolist()\n \n for x,y in zip(x,y):\n xperp = []; yperp = [];\n for xdiff, ydiff in zip(xnew, ynew):\n xperp.append(int(x+xdiff))\n yperp.append(int(y+ydiff))\n allxlist.append(xperp)\n allylist.append(yperp)\n\n xmap = self.currDICGrain.extremeCoords[0] + xperp\n ymap = self.currDICGrain.extremeCoords[1] + yperp\n \n ### For all points, append u and v to list\n u = []; v = [];\n for xmap, ymap in zip(xmap,ymap):\n u.append((self.currMap.crop(self.currMap.x_map))[ymap, xmap])\n v.append((self.currMap.crop(self.currMap.y_map))[ymap, xmap])\n\n ### Take away mean\n u = u-np.mean(u); v = v-np.mean(v)\n\n ### Append to main lists (ulist,vlist)\n ulist.extend(u)\n vlist.extend(v)\n\n ### Linear regression of ucentered against vcentered\n linRegResults = linregress(x=vlist,y=ulist)\n \n # Save measured RDR\n grain.groupsList[group][4] = linRegResults.slope\n \n\n if showPlot: self.plotRDR(grain, group, ulist, vlist, allxlist, allylist, linRegResults)\n\n def plotRDR(self, grain, group, ulist, vlist, allxlist, allylist, linRegResults):\n \"\"\"\n Plot RDR figure, including location of perpendicular lines and scatter plot of ucentered vs vcentered.\n \n Parameters\n ----------\n grain: int\n DIC grain to plot.\n group: int\n Group ID to plot.\n ulist: list\n List of ucentered values.\n vlist: list\n List of vcentered values.\n allxlist: list\n List of all x values.\n allylist: list\n List of all y values.\n linRegResults: numpy.ndarray, {slope, intercept, rvalue, pvalue, stderr}\n Results from linear regression of ucentered vs vcentered.\n\n \"\"\"\n\n # Draw window and axes\n self.rdrPlot = Plot(ax=None, makeInteractive=True, title='RDR Calculation', figsize=(21, 7))\n self.rdrPlot.ax.axis('off')\n self.rdrPlot.grainAx = self.rdrPlot.addAxes((0.05, 0.07, 0.20, 0.85))\n self.rdrPlot.textAx = self.rdrPlot.addAxes((0.27, 0.07, 0.20, 0.85))\n self.rdrPlot.textAx.axis('off')\n self.rdrPlot.numLineAx = self.rdrPlot.addAxes((0.48, 0.07, 0.2, 0.85))\n self.rdrPlot.numLineAx.axis('off')\n self.rdrPlot.plotAx = self.rdrPlot.addAxes((0.75, 0.07, 0.2, 0.85))\n\n ## Draw grain plot\n self.rdrPlot.grainPlot = self.currDICGrain.plotMaxShear(fig=self.rdrPlot.fig, ax=self.rdrPlot.grainAx, \n plotColourBar=False, plotScaleBar = True) \n self.rdrPlot.grainPlot.addColourBar(label='Effective Shear Strain', fraction=0.046, pad=0.04)\n\n ## Draw all points\n self.rdrPlot.grainAx.plot(allxlist, allylist, 'rx',lw=0.5)\n for xlist, ylist in zip(allxlist, allylist):\n self.rdrPlot.grainAx.plot(xlist, ylist, '-',lw=1)\n\n ## Generate scatter plot\n slope = linRegResults.slope\n r_value = linRegResults.rvalue\n intercept = linRegResults.intercept\n std_err = linRegResults.stderr\n \n self.rdrPlot.plotAx.scatter(x=vlist,y=ulist,marker='x', lw=1)\n self.rdrPlot.plotAx.plot(\n [np.min(vlist), np.max(vlist)],[slope*np.min(vlist)+intercept,slope*np.max(vlist)+intercept], '-')\n self.rdrPlot.plotAx.set_xlabel('v-centered')\n self.rdrPlot.plotAx.set_ylabel('u-centered')\n self.rdrPlot.addText(self.rdrPlot.plotAx, 0.95, 0.01, 'Slope = {0:.3f} ± {1:.3f}\\nR-squared = {2:.3f}\\nn={3}'\n .format(slope,std_err,r_value**2,len(ulist)), va='bottom', ha='right',\n transform=self.rdrPlot.plotAx.transAxes, fontsize=10);\n\n ## Write grain info\n ebsdGrain = grain.ebsdGrain\n ebsdGrain.calcSlipTraces()\n\n if ebsdGrain.averageSchmidFactors is None:\n raise Exception(\"Run 'calcAverageGrainSchmidFactors' first\")\n\n eulers = np.rad2deg(ebsdGrain.refOri.eulerAngles())\n\n text = 'Average angle: {0:.2f}\\n'.format(grain.groupsList[group][1])\n text += 'Eulers: {0:.1f} {1:.1f} {2:.1f}\\n\\n'.format(eulers[0], eulers[1], eulers[2])\n\n self.rdrPlot.addText(self.rdrPlot.textAx, 0.15, 1, text, fontsize=10, va='top')\n\n ## Write slip system info\n RDRs = []; offset = 0; \n for idx, (ssGroup, sfGroup, slipTraceAngle) in enumerate(\n zip(grain.ebsdMap.slipSystems, ebsdGrain.averageSchmidFactors, np.rad2deg(ebsdGrain.slipTraceAngles))):\n text = \"{0:s} {1:.1f}\\n\".format(ssGroup[0].slipPlaneLabel, slipTraceAngle)\n tempRDRs = [];\n for ss, sf in zip(ssGroup, sfGroup):\n slipDirSample = ebsdGrain.refOri.conjugate.transformVector(ss.slipDir)\n text = text + \" {0:s} SF: {1:.3f} RDR: {2:.3f}\\n\".format\\\n (ss.slipDirLabel, sf,-slipDirSample[0]/slipDirSample[1])\n RDR = -slipDirSample[0]/slipDirSample[1]\n tempRDRs.append(RDR)\n RDRs.append(tempRDRs) \n\n if idx in grain.groupsList[group][2]:\n self.rdrPlot.addText(self.rdrPlot.textAx, 0.15, 0.9-offset, text, weight='bold', fontsize=10, va='top')\n else:\n self.rdrPlot.addText(self.rdrPlot.textAx, 0.15, 0.9-offset, text, fontsize=10, va='top')\n\n offset += 0.0275 * text.count('\\n')\n\n # Plot RDR values on number line\n uniqueRDRs = set()\n for x in [item for sublist in RDRs for item in sublist]: uniqueRDRs.add(x)\n self.rdrPlot.numLineAx.axvline(x=0, ymin=-20, ymax=20, c='k')\n self.rdrPlot.numLineAx.plot(np.zeros(len(uniqueRDRs)), list(uniqueRDRs), 'bo', label='Theroretical RDR values')\n self.rdrPlot.numLineAx.plot([0], slope, 'ro', label='Measured RDR value')\n self.rdrPlot.addText(self.rdrPlot.numLineAx, -0.009, slope-0.01, '{0:.3f}'.format(float(slope)))\n self.rdrPlot.numLineAx.legend(bbox_to_anchor=(1.15, 1.05))\n \n # Label RDRs by slip system on number line \n for RDR in list(uniqueRDRs):\n self.rdrPlot.addText(self.rdrPlot.numLineAx, -0.009, RDR-0.01, '{0:.3f}'.format(float(RDR)))\n txt = ''\n for idx, ssGroup in enumerate(RDRs):\n for idx2, rdr in enumerate(ssGroup):\n if rdr == RDR:\n txt += str('{0} {1} '.format(self.currEBSDMap.slipSystems[idx][idx2].slipPlaneLabel, \n self.currEBSDMap.slipSystems[idx][idx2].slipDirLabel))\n self.rdrPlot.addText(self.rdrPlot.numLineAx,0.002, RDR-0.01, txt)\n\n self.rdrPlot.numLineAx.set_ylim(slope-1, slope+1)\n self.rdrPlot.numLineAx.set_xlim(-0.01, 0.05)\n","sub_path":"defdap/inspector.py","file_name":"inspector.py","file_ext":"py","file_size_in_byte":20495,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"291243467","text":"# def server_command(cmd):\r\n# process.stdin.write(cmd+\"\\n\") #just write the command to the input stream\r\n# process = None\r\n# executable = '\"C:/Program Files/Java/jre1.8.0_191/bin/java.exe\" -Xms4G -Xmx4G -jar \"D:/gd/minecraft_1.15.2/server (6).jar\" nogui java'\r\n# while True:\r\n# command=input()\r\n# command=command.lower()\r\n# if process is not None:\r\n# if command==(\"start\"):\r\n# os.chdir(minecraft_dir)\r\n# process = subprocess.Popen(executable, stdin=subprocess.PIPE)\r\n# print(\"Server started.\")\r\n# else:\r\n# server_command(command)\r\n\r\nimport socket\r\nimport os\r\nimport time\r\nimport pickle\r\nimport numpy as np\r\nimport copy\r\nimport wexpect\r\nimport json\r\nimport random\r\n# import subprocess\r\n# import time\r\n\r\n# mc_server=subprocess.Popen('\"C:/Program Files/Java/jre1.8.0_191/bin/java.exe\" -Xms4G -Xmx4G -jar \"D:/gd/minecraft_1.15.2/server (6).jar\" nogui java',shell=False,stdout=subprocess.PIPE,stdin=subprocess.PIPE)\r\n# mc_server.stdin.write(b'/time set midnight \\n')\r\n# #mc_server.stdin.close()\r\n# mc_server.stdin.flush()\r\n\r\n#this function takes a wexpect connection and returns the player list\r\n#need to adda try catch to this\r\ndef capture_player_list(mc_server):\r\n #print(player_list)\r\n #the start and end of the substring we will be taking\r\n while(True):\r\n #print(mc_server)\r\n #this first line basically just flushed the buffer becaause we want only the output of \"/list\". It will be overwritten\r\n throwRA=mc_server.read_nonblocking()\r\n mc_server.sendline('/list')\r\n time.sleep(0.2)\r\n player_list=mc_server.read_nonblocking()\r\n if ('20 players online:' in player_list) and (not('lost connection: Disconnected' in player_list)) and (not('the game' in player_list)) and player_list.count(\"[Server thread/INFO]\")==1:\r\n print(player_list)\r\n start=player_list.find('20 players online:')+len('20 players online:')+1\r\n end=player_list.rfind('\\r\\n')\r\n players=player_list[start:end].replace(\",\",\"\").split(\" \")\r\n break\r\n return(players)\r\n\r\n#stock server \r\ndef stop_server(mc_server):\r\n mc_server.sendline('/stop')\r\n\r\n#create team block\r\ndef create_teams():\r\n landing=[\"\"]*10\r\n team1=[\"\"]*10\r\n team2=[\"\"]*10\r\n team3=[\"\"]*10\r\n team4=[\"\"]*10\r\n return([landing,team1,team2,team3,team4])\r\n\r\n#for soem reason this function is working on global varibales.\r\ndef update_position(team_data,x,y,new_value):\r\n data=copy.deepcopy(team_data)\r\n data[x][y]=new_value\r\n return(data)\r\n # data=team_data\r\n # data[x][y]=new_value\r\n #return(data)\r\n\r\n\r\n#team_to_add_to is an integer with the group number landing(0), team1(1), team2(2), team3(3), team4(4)\r\ndef add_to_team(team_data,team_to_add_to,name):\r\n #for the team which will have a member added we should find where the next empty slot is\r\n # team_data[team_to_add_to]\r\n filled_already=[]\r\n for poss in team_data[team_to_add_to]:\r\n filled_already.append(not(poss==\"\"))\r\n if sum(filled_already)>9:\r\n #to prevent an error out if over 10 players in one team just return unchanged\r\n return(team_data)\r\n else:\r\n new_free_position=sum(filled_already)\r\n #print(new_free_position)\r\n return(update_position(team_data,x=team_to_add_to,y=new_free_position,new_value=name))\r\n\r\ndef clean_slots(the_list):\r\n if the_list[the_list.index(\"\")+1]==\"\":\r\n return(the_list)\r\n else:\r\n new_list=the_list[0:the_list.index(\"\")] + the_list[the_list.index(\"\")+1:10] + [\"\"]\r\n # new_list.append(the_list[0:the_list.index(\"\")])\r\n # new_list.append(the_list[the_list.index(\"\")+1:9])\r\n # new_list.append(\"\")\r\n return(new_list)\r\n\r\n#remove player from a team\r\ndef remove_from_team(team_data,team_to_remove_from,name):\r\n team_data_copy=copy.deepcopy(team_data)\r\n temp_team=team_data_copy[team_to_remove_from]\r\n removed_team=[]\r\n for i in temp_team:\r\n if name==i:\r\n removed_team.append(\"\")\r\n else:\r\n removed_team.append(i)\r\n #the slot is now empty. players below should be bumped up\r\n team_data_copy[team_to_remove_from]=clean_slots(removed_team)\r\n return(team_data_copy)\r\n\r\n#find the position of a player in the teams\r\ndef find_member_on_team(mylist,char):\r\n for sub_list in mylist:\r\n if char in sub_list:\r\n return (mylist.index(sub_list), sub_list.index(char))\r\n raise ValueError(\"'{char}' is not in list\".format(char = char))\r\n\r\n#update the team data and account for any logins or disconnects\r\ndef check_current_login(mc_server,accounted,team_data):\r\n active_players=capture_player_list(mc_server)\r\n #find any players that are new logins\r\n player_unaccounted=[]\r\n for i in active_players:\r\n player_unaccounted.append(not(i in accounted))\r\n print(i)\r\n player_unaccounted=np.array(active_players)[np.array(player_unaccounted)]\r\n #if there are new players, we should add them to the landing page\r\n if not(len(player_unaccounted)==0):\r\n for player in player_unaccounted:\r\n team_data=add_to_team(team_data,team_to_add_to=0,name=player)\r\n accounted.append(player)\r\n #now check for player which have logged out\r\n player_left=[]\r\n for i in accounted:\r\n player_left.append(not(i in active_players))\r\n #get names of players that logged out\r\n player_logged=np.array(accounted)[np.array(player_left)]\r\n print(len(player_logged))\r\n if not(len(player_logged)==0):\r\n print(\"player left!\")\r\n for player in player_logged:\r\n #need add a function to find where the person is if they moved a team already\r\n a,b=find_member_on_team(team_data,player)\r\n team_data=remove_from_team(team_data,a,name=player)\r\n print(team_data)\r\n print(player)\r\n #remove these players from accounted\r\n accounted=np.array(accounted)[np.array([not i for i in player_left])]\r\n accounted=accounted.tolist()\r\n #return all varibales that change\r\n return(team_data,accounted)\r\n\r\ndef move_player_to_team(team_data,player_to_move,new_team):\r\n for sub_list in team_data:\r\n if player_to_move in sub_list:\r\n x,y=team_data.index(sub_list), sub_list.index(player_to_move)\r\n team_data=remove_from_team(team_data=team_data,team_to_remove_from=x,name=player_to_move)\r\n team_data=add_to_team(team_data=team_data,team_to_add_to=new_team,name=player_to_move)\r\n return(team_data)\r\n print(\"Could not find the selected player. Maybe they logged out\")\r\n print(player_to_move)\r\n return(team_data)\r\n\r\ndef change_player_gamemode(mc_server,player_name,mode):\r\n mc_server.sendline('/gamemode '+mode+\" \"+player_name)\r\n\r\ndef teleport_player(mc_server,player_name,x,y,z):\r\n mc_server.sendline('/teleport '+str(player_name)+\" \"+str(x)+\" \"+str(y)+\" \"+str(z))\r\n\r\ndef create_team(mc_server,team_name):\r\n mc_server.sendline('/team add '+team_name)\r\n\r\ndef add_member_to_team(mc_server,team_name,member_to_add):\r\n mc_server.sendline('/team join '+team_name+\" \"+member_to_add)\r\n\r\ndef kill_all_players(mc_server):\r\n mc_server.sendline('/kill @a')\r\n\r\ndef set_world_border(mc_server,worldborder_size,time=\"\"):\r\n if time==\"\":\r\n mc_server.sendline('/worldborder set '+str(worldborder_size))\r\n else:\r\n mc_server.sendline('/worldborder set '+str(worldborder_size)+\" \"+str(time))\r\n\r\ndef calculate_teams_and_spawns(team_data,worldborder_center,worldborder_start_size,worldborder_end_size,worldborder_collpase_time):\r\n #first lets find the number of teams which contain at least one player\r\n #we will fill active_team with the indexes of [team_data] which are the teams which will play\r\n active_team_indexes=[]\r\n for sublist in team_data:\r\n for underlist in sublist:\r\n if underlist!=\"\":\r\n active_team_indexes.append(team_data.index(sublist))\r\n break\r\n print(active_team_indexes)\r\n if len(active_team_indexes)==0:\r\n print(\"Error. no active teams found.\")\r\n else:\r\n print(\"teams found.\")\r\n #now lets calculate the possible spawn locations. \r\n xhigh=worldborder_center[0]+(worldborder_start_size/2)\r\n xlow=worldborder_center[0]-(worldborder_start_size/2)\r\n yhigh=worldborder_center[1]+(worldborder_start_size/2)\r\n ylow=worldborder_center[1]-(worldborder_start_size/2)\r\n loc_1=[xhigh,ylow]\r\n loc_2=[xlow,yhigh]\r\n loc_3=[xlow,ylow]\r\n loc_4=[xhigh,yhigh]\r\n possible_spawns=[loc_1,loc_2,loc_3,loc_4]\r\n random.shuffle(possible_spawns)\r\n return(active_team_indexes,possible_spawns[0:len(active_team_indexes)])\r\n\r\n# calculate_teams_and_spawns(test.teams,test.worldborder_center_location,test.worldborder_start_size,test.worldborder_end_size,3600)\r\n\r\n\r\n\r\n\r\n\r\n\r\n\r\nclass player:\r\n def __init__(self):\r\n self.teams=create_teams()\r\n self.mc_server=wexpect.spawn('java -Xms4G -Xmx4G -jar \"server (6).jar\" nogui java',cwd=os.path.abspath(\"D:\\\\gd\\\\minecraft_1.15.2\"))\r\n self.accounted=[]\r\n self.worldborder_center_location=[0,0]\r\n self.worldborder_start_size=1000\r\n self.worldborder_end_size=100\r\n self.worldborder_collapse_time=3600\r\n self.initiate_match_teams_indexes=[]\r\n self.initiate_match_teams_spawn_locations=[]\r\n self.total_players=[]\r\n self.check_players_time=0\r\n def check_players(self):\r\n if (time.time()-self.check_players_time)>1:\r\n try:\r\n self.teams,self.accounted=check_current_login(mc_server=self.mc_server,accounted=self.accounted,team_data=self.teams)\r\n self.check_players_time=time.time()\r\n except:\r\n print(\"check_players_error_rebound\")\r\n def stop_pserver(self):\r\n stop_server(self.mc_server)\r\n def move_player(self,player_to_move,new_team):\r\n self.teams=move_player_to_team(team_data=self.teams,player_to_move=player_to_move,new_team=new_team)\r\n def pre_match_calc(self):\r\n self.initiate_match_teams_indexes,self.initiate_match_teams_spawn_locations=calculate_teams_and_spawns(team_data=self.teams,worldborder_center=self.worldborder_center_location,worldborder_start_size=self.worldborder_start_size,worldborder_end_size=self.worldborder_end_size,worldborder_collpase_time=self.worldborder_collapse_time)\r\n def make_teams(self):\r\n #for each team that will play create a team\r\n for i in self.initiate_match_teams_indexes:\r\n #make a clean list of all players. make empty to fill\r\n #create the team name\r\n team_name_assign=str(\"team_\"+str(i))\r\n print(team_name_assign)\r\n create_team(mc_server=self.mc_server,team_name=team_name_assign)\r\n print(\"good\")\r\n #for each team add the members\r\n players_in_team=[]\r\n print(\"good1\")\r\n for player_in_list in self.teams[i]:\r\n if player_in_list!=\"\":\r\n players_in_team.append(player_in_list)\r\n print(players_in_team)\r\n self.total_players.append(players_in_team)\r\n print(self.total_players)\r\n #now we have the list of players, per team add them to the team\r\n print(\"good2\")\r\n for member in players_in_team:\r\n add_member_to_team(mc_server=self.mc_server,team_name=team_name_assign,member_to_add=member)\r\n players_in_team=[]\r\n #delete first element which is the empty list\r\n print(self.total_players)\r\n # del self.total_players[0]\r\n def start_match(self):\r\n self.mc_server.sendline('/say Command to Initiate Match Received.')\r\n self.mc_server.sendline('/say Calculating teams and spawn locations...')\r\n self.mc_server.sendline('/say Assigning players to their teams...')\r\n self.pre_match_calc()\r\n self.make_teams()\r\n self.mc_server.sendline('/say Killing All Players in 10 seconds... Please Respawn to be Full Health')\r\n time.sleep(10)\r\n kill_all_players(mc_server=self.mc_server)\r\n self.mc_server.sendline('/say Teleporting teams to their respective start locations in 5 seconds...')\r\n time.sleep(5)\r\n set_world_border(mc_server=self.mc_server,worldborder_size=self.worldborder_start_size,time=\"\")\r\n #\r\n #make all players creative before the jump\r\n for sublist in self.total_players:\r\n for indiv_player in sublist:\r\n change_player_gamemode(mc_server=self.mc_server,player_name=indiv_player,mode=\"creative\")\r\n print(\"creative\")\r\n #\r\n #teleport teams to staarting\r\n x=0\r\n for sublist in self.total_players:\r\n for indiv_player in sublist:\r\n teleport_player(mc_server=self.mc_server,player_name=indiv_player,x=self.initiate_match_teams_spawn_locations[int(x)][0],y=200,z=self.initiate_match_teams_spawn_locations[int(x)][1])\r\n print(indiv_player)\r\n print(self.initiate_match_teams_spawn_locations[int(x)][0],self.initiate_match_teams_spawn_locations[int(x)][0])\r\n x=x+1\r\n time.sleep(10)\r\n print(\"teleport\")\r\n #\r\n #after players survive the fall turn them back to survival\r\n #make all players creative before the jump\r\n x=0\r\n for sublist in self.total_players:\r\n for indiv_player in sublist:\r\n change_player_gamemode(mc_server=self.mc_server,player_name=indiv_player,mode=\"survival\")\r\n print(indiv_player)\r\n x=x+1\r\n set_world_border(mc_server=self.mc_server,worldborder_size=self.worldborder_end_size,time=self.worldborder_collapse_time)\r\n self.initiate_match_teams_indexes=[]\r\n self.initiate_match_teams_spawn_locations=[]\r\n self.total_players=[]\r\n\r\n\r\n\r\n# test.stop_pserver()\r\n# test=player()\r\n# test.teams\r\n# test.teams[1][0]=\"aaron\"\r\n# test.teams[2][0]=\"aaronddd\"\r\n# test.pre_match_calc()\r\n# test.make_teams()\r\n# test.total_players\r\n\r\n\r\n\r\n# test=player()\r\n# time.sleep(5)\r\n# test.check_players()\r\n# player.teams\r\n# player.accounted\r\n\r\n#aaro=capture_player_list(mc_server)\r\n\r\n\r\n#mc_server=wexpect.spawn('java -Xms4G -Xmx4G -jar \"server (6).jar\" nogui java')\r\n# mc_server=wexpect.spawn('java -Xms4G -Xmx4G -jar \"server (6).jar\" nogui java',cwd=os.path.abspath(\"D:\\\\gd\\\\minecraft_1.15.2\"))\r\n# mc_server.sendline('/time set midnight')\r\n#printout the buffer holding the console output\r\n\r\n\r\n\r\n\r\n# print(mc_server.read_nonblocking())\r\n\r\n#players=capture_player_list(mc_server)\r\n\r\n#list for players already accounted for\r\n\r\n\r\n\r\nHOST = '' # Symbolic name meaning all available interfaces\r\nPORT = 50008 # Arbitrary non-privileged port\r\ns = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\r\ns.bind((HOST, PORT))\r\ns.listen(1)\r\ntest=player()\r\ntime.sleep(5)\r\nwhile True:\r\n clientsocket, addr = s.accept()\r\n print(\"computer\",{addr},\"coonected.\")\r\n msg=clientsocket.recv(1024)\r\n msg=json.loads(msg.decode(\"utf-8\"))\r\n if msg[0]==\"ch_team\":\r\n #xtract list of player to be moved\r\n player_move=msg[2][:-1]\r\n player_move=player_move.split(\" \")\r\n for player in player_move:\r\n test.move_player(player_to_move=player,new_team=int(msg[1]))\r\n elif msg[0]==\"update_teams\": \r\n test.check_players()\r\n elif msg[0]==\"worldborder_start\":\r\n test.worldborder_start_size=int(msg[1])\r\n elif msg[0]==\"worldborder_end\":\r\n test.worldborder_end_size=int(msg[1])\r\n elif msg[0]==\"worldborder_time_move\":\r\n test.worldborder_collapse_time=int(msg[1])\r\n elif msg[0]==\"start_game\":\r\n test.start_match()\r\n temp_teams=copy.deepcopy(test.teams)\r\n temp_teams.append([test.worldborder_start_size,test.worldborder_end_size,test.worldborder_collapse_time])\r\n to_send=temp_teams\r\n # to_send=test.teams\r\n clientsocket.sendall(bytes(json.dumps(to_send),encoding=\"utf-8\"))\r\n \r\n","sub_path":"minecraft_1.15.2/hunger_games.py","file_name":"hunger_games.py","file_ext":"py","file_size_in_byte":16139,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"442518766","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Oct 12 20:57:19 2020\n\n@author: bnebe\n\"\"\"\n\nimport pandas as pd\nimport numpy as np\n\nimport sklearn.preprocessing as pp\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.pipeline import Pipeline\n\nfrom sklearn.neural_network import MLPRegressor\n\nimport time\nstart_time = time.time()\n\ndf_raw = pd.read_csv(\"final_df.csv\", index_col=0, parse_dates=True)\n\n# Copy raw file, filter down years and columns\ndf = df_raw.copy()\ndf = df.loc['2018-01-01':'2019-12-31']\ndf = df[['dewpoint', 'rel_humidity', 'temperature', 'wind_direction', 'wind_speed',\n 'Fuel_Price', 'Wind_MW', 'Solar_MW', 'Demand_DLAP_MW', 'Demand_MW', 'Year', \n 'Month', 'Day', 'Hour', 'Weekday', 'Weekend', 'LMP_Price_Per_MWh']]\n\n\n# Feature creation\n# Create wind vectors\ndf['wind_x'] = df['wind_speed'] * np.cos(np.deg2rad(df['wind_direction']))\ndf['wind_y'] = df['wind_speed'] * np.sin(np.deg2rad(df['wind_direction']))\n\n# Create two weeks worth of lagged variables 168\nfor i in range (1, (24*7)):\n df['LMP_Price_Per_MWh -' + str(i) + 'h'] = df['LMP_Price_Per_MWh'].shift(i)\n\n# Create 48 hour lagged CAISO data 48\nfor i in range (1, 24):\n for item in ['Fuel_Price', 'Wind_MW', 'Solar_MW', 'Demand_DLAP_MW', 'Demand_MW']:\n df[item + ' -' + str(i) + 'h'] = df[item].shift(i)\n\n# Create rolling averages\nfor item in ['Wind_MW', 'Solar_MW', 'Demand_DLAP_MW', 'Demand_MW', 'LMP_Price_Per_MWh', 'temperature']:\n if item in ['Solar_MW', 'Wind_MW']:\n df[item + ' 12 roll avg'] = df[item].rolling(window=12).mean()\n else:\n df[item + ' 4 roll avg'] = df[item].rolling(window=4).mean()\n\n# Target DataFrame\n# Create future y's for \ntarget_df = pd.DataFrame(data = df['LMP_Price_Per_MWh'], index = df.index)\nfor i in range (1, 24):\n target_df['LMP_Price_Per_MWh +' + str(i) + 'h'] = target_df['LMP_Price_Per_MWh'].shift(-i)\ntarget_df = target_df.drop(labels = ['LMP_Price_Per_MWh'], axis = 1)\n\n\n# Random split\nX_train, X_test, y_train, y_test = train_test_split(df, target_df, test_size=0.25, random_state=1)\n\n\n# Pre-process Train\nss = pp.StandardScaler()\nX_train = pd.DataFrame(ss.fit_transform(X_train), index=y_train.index)\n\n# Drop NANs after to avoid lagged variables losing values and skewing the scaling\nX_train = X_train.dropna()\ny_train = y_train.loc[X_train.index]\n\ny_train = y_train.dropna()\nX_train = X_train.loc[y_train.index]\n\n\n# MLP Regressor\nregr = MLPRegressor(hidden_layer_sizes = (30, 40, 50, 60, 70), activation = 'relu', solver = 'adam', alpha = 0.05, max_iter=300)\nregr.fit(X_train, y_train)\nprint('R^2:', regr.score(X_train, y_train))\n\nprint('Time to finish: ', (time.time()-start_time)/60)\n\n","sub_path":"Development/model_training.py","file_name":"model_training.py","file_ext":"py","file_size_in_byte":2671,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"415975959","text":"#!/usr/bin/env python3\n# lexicon.py jcj 2019-02-20, 2020-01-23\n\n'''A class to implement a lexicon with methods for prefixes and suffixes,\nregular expressions, anagrams, etc'''\n\nimport bisect\nimport re\nimport unicodedata as ud\nfrom collections import defaultdict\n\nINTERVAL = 5 # How often in % is progress notified?\nPASSES = 4 # How many times is each entry processed?\n\nclass LexiconError(Exception):\n\tpass\n\nclass Lexicon:\n\n\tdef __init__(self, fileName, caseBlind=False, diacBlind=False, busyWait=None):\n\t\t'''Initialize an object representing a lexicon from a disk file\n\t\tconsisting of one entry per line (unique but not necessarily sorted)\n\t\twith possible comment lines beginning with a hash sign.\n\t\tMethods are provided for lookup, with possible disregard of\n\t\tcase and/or diacritics, of whole words, prefixes, suffixes,\n\t\tregular expressions, and anagrams. Lookup of whole words and anagrams\n\t\tis optimized. Provision is made for callback to an optional\n\t\tfunction or method to display progress information.'''\n\t\tself.fileName = fileName\n\t\tself.caseBlind = caseBlind\n\t\tself.diacBlind = diacBlind\n\t\tif not busyWait:\n\t\t\tbusyWait = lambda percent, message: None\n\t\tself.words = [] # list of sorted, normalized words\n\t\tself.refs = defaultdict(list) # dict from normalized words\n\t\t # to lists of reference forms\n\t\t\t\t\t\t\t\t\t\t # eg POLISH -> [Polish, polish]\n\t\tself.anags = defaultdict(list) # dict from normalized anagrams\n\t\t # to lists of anagrammatic forms\n\t\t\t\t\t\t\t\t\t\t # eg abeert -> [beater, berate, rebate]\n\t\ttry:\n\t\t\tf = open(fileName)\n\t\texcept Exception as err:\n\t\t\traise LexiconError(err)\n\t\tbusyWait(0, 'Reading...')\n\t\t# Read in a list of lines, without showing any progress.\n\t\t# But then use the number of words read in to time progress updates\n\t\tlines = [ line.rstrip('\\n') for line in f.readlines()\n\t\t\t\t\tif not line.startswith('#') ]\n\t\tself.numLines = len(lines)\n\t\tpcPerLine = 100 / (self.numLines * PASSES) # NB result is a real\n\t\tpcDone = pcPerLine * self.numLines # for the read itself\n\t\t# build dict of normalized forms in self.refs\n\t\tbusyWait(round(pcDone), 'Normalizing...')\n\t\tfor line in lines:\n\t\t\tpcDone += pcPerLine\n\t\t\tif pcDone % INTERVAL < pcPerLine:\n\t\t\t\tbusyWait(round(pcDone), 'Normalizing...')\n\t\t\tself.refs[self.normalized(line, caseBlind, diacBlind)].append(line)\n\t\t# build sorted list of normalized word forms for lookup\n\t\tbusyWait(round(pcDone), 'Sorting...')\n\t\tself.words = sorted(self.refs) # includes only the keys\n\t\tpcDone += pcPerLine * self.numLines\n\t\t# build a separate dictionary of normalized anagram forms\n\t\tbusyWait(round(pcDone), 'Hashing...')\n\t\tfor line in lines: \n\t\t\tpcDone += pcPerLine\n\t\t\tif pcDone % INTERVAL < pcPerLine:\n\t\t\t\tbusyWait(round(pcDone), 'Hashing...')\n\t\t\thash = self.anagramHash(line)\n\t\t\tself.anags[hash].append(line)\n\t\tbusyWait(100, 'Done.')\n\t\tf.close()\n\t\n\tdef normalized(self, s, caseBlind, diacBlind):\n\t\t'''Apply needed transformations to ignore case and/or accents'''\n\t\tif caseBlind:\n\t\t\ts = s.upper()\n\t\tif diacBlind:\n\t\t\ts = ud.normalize('NFKD', s)\n\t\t\ts = ''.join([c for c in s if not ud.combining(c)])\n\t\treturn s\n\t\t\n\tdef length(self):\n\t\t'''Return the number of entries'''\n\t\treturn self.numLines\n\n\tdef anagramHash(self, s):\n\t\t'''Create a unique hash for anagram purposes, ignoring\n\t\torder, letter-case, and all punctuation. Diacritics\n\t\tare ignored only if self.diacBlind is set'''\n\t\t# Unfortunately the \\w class (Unicode 'word' characters)\n\t\t# includes the underscore, so _ must be special-cased\n\t\ts = s.replace('_', '')\n\t\t# Upper-case the string, remove all characters which are not \\w,\n\t\t# and sort the result: thus all anagrammatic strings get the same hash\n\t\treturn ''.join(sorted(re.sub(r'\\W', '',\n\t\t\t\t\t\tself.normalized(s, True, self.diacBlind))))\n\n\tdef contains(self, word):\n\t\t'''Return a list of matching words'''\n\t\tword = self.normalized(word, self.caseBlind, self.diacBlind)\n\t\ti = bisect.bisect_left(self.words, word)\n\t\tif i != len(self.words) and self.words[i] == word:\n\t\t\treturn self.refs[word]\n\t\telse:\n\t\t\treturn []\n\n\tdef regex(self, pattern):\n\t\t'''Return list of matching words'''\n\t\t# we can normalize case but we can't do anything about diacritics\n\t\tflags = re.IGNORECASE if self.caseBlind else 0\n\t\tpattern = re.compile(pattern, flags)\n\t\tmatches = []\n\t\tfor word in self.words:\n\t\t\tm = pattern.search(word)\n\t\t\tif m:\n\t\t\t\tmatches.extend(self.refs[word])\n\t\treturn matches\n\n\tdef hasPrefix(self, prefix):\n\t\t'''Return whether there is any word that begins with prefix'''\n\t\tprefix = self.normalized(prefix, self.caseBlind, self.diacBlind) \n\t\tfor word in self.words:\n\t\t\tif word.startswith(prefix):\n\t\t\t\treturn True\n\t\treturn False\n\n\tdef withPrefix(self, prefix):\n\t\t'''Return a list of words with prefix'''\n\t\tprefix = self.normalized(prefix, self.caseBlind, self.diacBlind)\n\t\treturn [ ' '.join(self.refs[w]) for w in self.words\n\t\t\t\t\tif w.startswith(prefix) ]\n\n\tdef hasSuffix(self, suffix):\n\t\t'''Return whether there is any word that ends with suffix'''\n\t\tsuffix = self.normalized(suffix, self.caseBlind, self.diacBlind)\n\t\tif self.caseBlind:\n\t\t\tsuffix = suffix.upper()\n\t\tfor word in self.words:\n\t\t\tif word.endswith(suffix):\n\t\t\t\treturn True\n\t\treturn False\n\n\tdef withSuffix(self, suffix):\n\t\t'''Return a list of words with suffix'''\n\t\tsuffix = self.normalized(suffix, self.caseBlind, self.diacBlind)\n\t\treturn [ ' '.join(self.refs[w]) for w in self.words\n\t\t\t\t\tif w.endswith(suffix) ]\n\n\n\tdef anagrams(self, word):\n\t\t'''Return a list of anagrams of word, ignoring case and punctuation''' \n\t\treturn self.anags.get(self.anagramHash(word), [])\n\ndef main():\n\tprint('This module is intended to be imported rather than run standalone.')\n\tprint('Use as \"import lexicon\" or \"from lexicon import Lexicon, LexiconError\".')\n\nif __name__ == '__main__':\n\tmain()\n\n","sub_path":"archive/non-i18n/lexicon.py","file_name":"lexicon.py","file_ext":"py","file_size_in_byte":5789,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"302319963","text":"import argparse\n\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom ridge_map import RidgeMap\n\nfrom heightmap import get_image_dims, read_image\n\n\nHEIGHTMAP_FILE = \"heightmaps/los_santos.png\" # input file\nOUTPUT_FILE = \"output/los_santos_desktop.png\" # output file\n\nNUM_LINES = 80 # ideal number of lines to include in ridge map\nX_RESOLUTION = 2 # \"resolution\" in x direction (i.e. for resolution 2, 1 in 2 data points are included)\n\n# (Y_DIM, X_DIM) = get_image_dims(HEIGHTMAP_FILE) # use dims from original file\n(Y_DIM, X_DIM) = (824, 824) # custom dims\n\nDPI = 96 # DPI of my monitor, use link to find out: https://www.infobyip.com/detectmonitordpi.php\nSCALING_FACTOR = 1 # Factor to scale output image by\n\nfig, ax = plt.subplots(figsize=(X_DIM/DPI, Y_DIM/DPI), dpi=DPI)\n\nrm = RidgeMap()\n\nvalues = read_image(HEIGHTMAP_FILE, NUM_LINES, X_RESOLUTION)\n\nvalues = rm.preprocess(values=values,\n water_ntile=40,\n lake_flatness=2,\n vertical_ratio=40)\n\nrm.plot_map(values=values,\n label='',\n label_y=0.2,\n label_x=0.2,\n label_size=20,\n linewidth=2,\n line_color=plt.get_cmap('cool'),\n kind='gradient',\n background_color=np.array([65, 74, 76])/255,\n ax=ax)\n\n# Remove margins around image\n# Solution found from discussions here: https://stackoverflow.com/questions/11837979/removing-white-space-around-a-saved-image-in-matplotlib\nplt.subplots_adjust(top=1, bottom=0, right=1, left=0, hspace=0, wspace=0)\nplt.margins(0, 0)\n\nplt.savefig(OUTPUT_FILE,\n bbox_inches='tight',\n pad_inches=0,\n dpi=DPI*SCALING_FACTOR)\n\nplt.show()","sub_path":"los_santos.py","file_name":"los_santos.py","file_ext":"py","file_size_in_byte":1720,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"47673064","text":"from collections import OrderedDict\n\n\nclass TaggedTableToDataStore:\n\n def __init__(self, row_tag_re, column_tag_re, data_store, spacer=\" \"):\n self.row_tag_re = row_tag_re\n self.column_tag_re = column_tag_re\n self.data_store = data_store\n self.spacer = spacer\n\n def get_name(self):\n return \"Tagged Table to Data Store\"\n\n def process(self, document, context):\n\n rows = self.process_node(document.get_root())\n\n for row in rows:\n context.get_store(self.data_store).add(row)\n return document\n\n def process_node(self, node):\n all_rows = []\n\n for line in node.findall(tag_name_re=self.row_tag_re):\n row = OrderedDict()\n for col in line.findall(tag_name_re=self.column_tag_re):\n if col.get_tags()[0] not in row:\n row[col.get_tags()[0]] = []\n row[col.get_tags()[0]].append(col.get_all_content())\n\n all_rows.append(row)\n\n max_cols = 0\n\n for row in all_rows:\n if len(row) > max_cols:\n max_cols = len(row)\n\n final_rows = []\n\n for row in all_rows:\n final_row = []\n for key, value in row.items():\n final_row.append(self.spacer.join(value))\n for n in range(len(final_row), max_cols):\n final_row.append(None)\n final_rows.append(final_row)\n\n return final_rows\n","sub_path":"kodexa/extractors/extractors.py","file_name":"extractors.py","file_ext":"py","file_size_in_byte":1453,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"504247205","text":"# !/usr/bin/env python3\n# -*- coding: utf-8 -*-\nimport json,requests\nfrom bs4 import BeautifulSoup\n\nclass Crawler:\n url=''\n headers={\n 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/63.0.3239.26 Safari/537.36 Core/1.63.6821.400 QQBrowser/10.3.3040.400'\n }\n data={}\n target_info={}\n #初始化\n def __init__(self,u,d):\n self.url=u\n self.data=d\n\n #返回session对象\n def getSession(self):\n s=requests.session()\n s.post(url=self.url, headers=self.headers, data=self.data)\n return s\n\n #返回get请求结果\n def getGetResponse(self,s,url):\n info=s.get(url)\n return info\n\n #返回初步处理后的结果\n def getHandleInfo(self,res):\n info = BeautifulSoup(res.content, 'html.parser')\n return info\n\n #返回目标信息\n def crawlInfo(self,info):\n target_info=info\n if (len(target_info) > 0):\n return json.dumps(target_info, ensure_ascii=False), 201\n else:\n return \"false\"\n","sub_path":"crawler.py","file_name":"crawler.py","file_ext":"py","file_size_in_byte":1080,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"388976345","text":"from external import *\nfrom TaskAssignment import *\nfrom display import *\n\nv = Visualization(\"Map.jpg\", 60)\nnum_t = 50\nt = np.concatenate((np.random.rand(num_t,1) * 1078, np.random.rand(num_t,1)*730), axis=1)\nt_list = []\n\nfor i in range(len(t)):\n t_list.append(Trolley(t[i, 0], t[i, 1]))\n\nassign = TaskAssignment(t_list, [])\ngroups = assign.grouping()\n\ngo_match = True\nwhile True:\n t_group = TrolleyGroup(0,0,t_list)\n if v.read_click_flag():\n go_match = False\n t_group.elimanate_trolleys_nearby(v.get_mouse_pos())\n t_list = t_group.get_trolleys()\n\n new_groups = []\n for g in groups:\n g.elimanate_trolleys_nearby(v.get_mouse_pos())\n if len(g.get_trolleys()) != 0:\n new_groups.append(TrolleyGroup(*g.get_position(), g.get_trolleys()))\n else:\n go_match = True\n groups = new_groups\n\n w_list = [Worker(*v.get_mouse_pos()), ]\n assign.update(worker=w_list)\n\n if go_match:\n match = assign.assign_workers_to_groups(groups)\n paths = []\n for i, w in enumerate(w_list):\n order = assign.calculate_picking_order(w, groups[match[i]])\n paths.append(order)\n\n v.update(worker=w_list, trolley=t_list, groups=groups, match=match, paths=paths)\n v.draw()\n k = cv2.waitKey(1)\n if k == ord('q'):\n break\n elif k == ord('g'):\n v.on_off_show(groups=False)\n elif k == ord('a'):\n t = np.concatenate((np.random.rand(3,1) * 1078, np.random.rand(3,1)*730), axis=1)\n for i in range(len(t)):\n t_list.append(Trolley(t[i, 0], t[i, 1]))\n assign.update(trolley=t_list)\n groups = assign.grouping()","sub_path":"one_worker_add_trolleys.py","file_name":"one_worker_add_trolleys.py","file_ext":"py","file_size_in_byte":1679,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"400775493","text":"#!/usr/bin/python\r\n# -*- coding: utf-8 -*-\r\n\r\nimport os, sys\r\nimport string\r\nimport glob\r\n\r\n\r\nwork_dir = os.getcwd();\r\nscript_dir = os.path.dirname(os.path.realpath(__file__))\r\nos.chdir(script_dir)\r\nos.chdir(os.path.join('..'));\r\n\r\nproject_dir = os.getcwd();\r\nproto_dir = os.path.join(script_dir, 'proto_v2');\r\n\r\nproto_file = []\r\nfor item in glob.glob(os.path.join(proto_dir, '*.proto')):\r\n proto_file.append('\"' + item + '\"');\r\n\r\nos.chdir(work_dir);\r\nos.system('python \"{0}\"'.format(os.path.join(project_dir, 'loader-binding', 'cxx', 'gen_protocol.py')))\r\ncpp_out_dir = os.path.join(script_dir, 'cxx');\r\n\r\nproto_src_dir = '{0}/v2'.format(cpp_out_dir)\r\nif not os.path.exists(proto_src_dir):\r\n os.mkdir(proto_src_dir)\r\nparams = ['protoc', '-I', proto_dir, '-o', os.path.join(proto_dir, 'kind.pb'), '--cpp_out={0}'.format(proto_src_dir)]\r\nparams.extend(proto_file)\r\ncmd = ' '.join(params)\r\nprint(cmd)\r\nos.system(cmd)\r\n\r\n","sub_path":"sample/gen_protocol.py","file_name":"gen_protocol.py","file_ext":"py","file_size_in_byte":924,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"390200850","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Oct 6 16:17:38 2018\n\n@author: owen\n\"\"\"\n\n# Given an array of integers, find how many pairs in the array such that \n# their sum is less than or equal to a specific target number. \n\n#class Solution:\n# \"\"\"\n# @param nums: an array of integer\n# @param target: an integer\n# @return: an integer\n# \"\"\"\n# def twoSum5(self, nums, target):\n# # write your code here\n# # sort + two pointers, time O(n log n + n^2)\n# nums.sort()\n# n = len(nums)\n# res = 0\n# for right in range(n - 1, 0, -1):\n# left = right - 1 # from right to left, find the first left that makes left + right <= target\n# while left >=0 and nums[left] + nums[right] > target:\n# left -= 1\n# res += left + 1\n# \n# return res\n\nclass Solution:\n \"\"\"\n @param nums: an array of integer\n @param target: an integer\n @return: an integer\n \"\"\"\n def twoSum5(self, nums, target):\n # write your code here\n # sort + two pointers, time O(nlogn + n)\n nums.sort()\n n = len(nums)\n res = 0\n left, right = 0, n - 1\n while left < right:\n sums = nums[left] + nums[right]\n if sums > target:\n right -= 1\n else: # for each left, find the last right that makes left + right <= target\n res += right - left\n left += 1\n \n return res","sub_path":"Two Sum - Less than or equal to target.py","file_name":"Two Sum - Less than or equal to target.py","file_ext":"py","file_size_in_byte":1517,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"513729088","text":"\"\"\"\nCopyright 2021 Tsinghua University\nApache 2.0.\nAuthor: Zheng Huahuan (zhh20@mails.tsinghua.edu.cn)\n\nThis script uses DistributedDataParallel (DDP) to train model within framework of CAT.\nDiffered from `train_dist.py`, this one supports read configurations from json file\nand is more non-hard-coding style.\n\"\"\"\n\nimport utils\nimport os\nimport argparse\nimport numpy as np\nimport model as model_zoo\nimport dataset as DataSet\nfrom _specaug import SpecAug\nfrom collections import OrderedDict\n\nimport torch\nimport torch.nn as nn\nimport torch.distributed as dist\nimport torch.multiprocessing as mp\nfrom torch.utils.data.distributed import DistributedSampler\nfrom torch.utils.data import DataLoader\n\nimport ctc_crf_base\n\n\ndef main(args):\n if not torch.cuda.is_available():\n utils.highlight_msg(\"CPU only training is unsupported.\")\n return None\n\n os.makedirs(args.dir+'/ckpt', exist_ok=True)\n setattr(args, 'ckptpath', args.dir+'/ckpt')\n if os.listdir(args.ckptpath) != [] and not args.debug and args.resume is None:\n utils.highlight_msg(\n f\"ERROR:\\nCheckpoint path `{args.ckptpath}` is not empty!\\nRefuse to run the experiment, otherwise previous files would be overwritten.\")\n raise AssertionError\n\n ngpus_per_node = torch.cuda.device_count()\n args.world_size = ngpus_per_node * args.world_size\n print(f\"Global number of GPUs: {args.world_size}\")\n mp.spawn(main_worker, nprocs=ngpus_per_node, args=(ngpus_per_node, args))\n\n\ndef main_worker(gpu, ngpus_per_node, args):\n args.gpu = gpu\n\n args.rank = args.rank * ngpus_per_node + gpu\n print(f\"Use GPU: local[{args.gpu}] | global[{args.rank}]\")\n\n dist.init_process_group(\n backend=args.dist_backend, init_method=args.dist_url,\n world_size=args.world_size, rank=args.rank)\n\n args.batch_size = args.batch_size // ngpus_per_node\n\n print(\"> Data prepare\")\n if args.h5py:\n data_format = \"hdf5\"\n utils.highlight_msg(\"H5py reading might cause error with Multi-GPUs.\")\n Dataset = DataSet.SpeechDataset\n else:\n data_format = \"pickle\"\n Dataset = DataSet.SpeechDatasetPickle\n\n tr_set = Dataset(\n f\"{args.data}/{data_format}/tr.{data_format}\")\n test_set = Dataset(\n f\"{args.data}/{data_format}/cv.{data_format}\")\n print(\"Data prepared.\")\n\n train_sampler = DistributedSampler(tr_set)\n test_sampler = DistributedSampler(test_set)\n test_sampler.set_epoch(1)\n\n trainloader = DataLoader(\n tr_set, batch_size=args.batch_size, shuffle=(train_sampler is None),\n num_workers=args.workers, pin_memory=True,\n sampler=train_sampler, collate_fn=DataSet.sortedPadCollate())\n\n testloader = DataLoader(\n test_set, batch_size=args.batch_size, shuffle=(test_sampler is None),\n num_workers=args.workers, pin_memory=True,\n sampler=test_sampler, collate_fn=DataSet.sortedPadCollate())\n\n logger = OrderedDict({\n 'log_train': ['epoch,loss,loss_real,net_lr,time'],\n 'log_eval': ['loss_real,time']\n })\n manager = utils.Manager(logger, build_model, args)\n\n # get GPU info\n gpu_info = utils.gather_all_gpu_info(args.gpu)\n\n if args.rank == 0:\n print(\"> Model built.\")\n print(\"Model size:{:.2f}M\".format(\n utils.count_parameters(manager.model)/1e6))\n\n utils.gen_readme(args.dir+'/readme.md',\n model=manager.model, gpu_info=gpu_info)\n\n # init ctc-crf, args.iscrf is set in build_model\n if args.iscrf:\n gpus = torch.IntTensor([args.gpu])\n ctc_crf_base.init_env(f\"{args.data}/den_meta/den_lm.fst\", gpus)\n\n # training\n manager.run(train_sampler, trainloader, testloader, args)\n\n if args.iscrf:\n ctc_crf_base.release_env(gpus)\n\n\nclass CAT_Model(nn.Module):\n def __init__(self, NET=None, fn_loss='crf', lamb: float = 0.1, net_kwargs: dict = None, sepcaug: nn.Module = None):\n super().__init__()\n if NET is None:\n return None\n\n self.infer = NET(**net_kwargs)\n self.specaug = sepcaug\n\n if fn_loss == \"ctc\":\n self.loss_fn = utils.CTCLoss()\n elif fn_loss == \"crf\":\n self.loss_fn = utils.CRFLoss(lamb=lamb)\n else:\n raise ValueError(f\"Unknown loss function: {fn_loss}\")\n\n def forward(self, logits, labels, input_lengths, label_lengths):\n labels = labels.cpu()\n input_lengths = input_lengths.cpu()\n label_lengths = label_lengths.cpu()\n\n netout, lens_o = self.infer(logits, input_lengths)\n netout = torch.log_softmax(netout, dim=-1)\n\n loss = self.loss_fn(netout, labels, lens_o.to(\n torch.int32).cpu(), label_lengths)\n\n return loss\n\n\ndef build_model(args, configuration, train=True) -> nn.Module:\n\n netconfigs = configuration['net']\n net_kwargs = netconfigs['kwargs']\n net = getattr(model_zoo, netconfigs['type'])\n\n if not train:\n infer_model = net(**net_kwargs)\n return infer_model\n\n if 'lossfn' not in netconfigs:\n lossfn = 'crf'\n utils.highlight_msg(\n \"Warning: not specified \\'lossfn\\' in configuration.\\nDefaultly set to \\'crf\\'\")\n else:\n lossfn = netconfigs['lossfn']\n\n if 'lamb' not in netconfigs:\n lamb = 0.01\n if lossfn == 'crf':\n utils.highlight_msg(\n \"Warning: not specified \\'lamb\\' in configuration.\\nDefaultly set to 0.01\")\n else:\n lamb = netconfigs['lamb']\n\n if 'specaug' not in netconfigs:\n specaug = None\n if args.rank == 0:\n utils.highlight_msg(\"Disable SpecAug.\")\n else:\n specaug = SpecAug(**netconfigs['specaug'])\n\n setattr(args, 'iscrf', lossfn == 'crf')\n model = CAT_Model(net, lossfn, lamb, net_kwargs, specaug)\n\n torch.cuda.set_device(args.gpu)\n model.cuda(args.gpu)\n model = torch.nn.parallel.DistributedDataParallel(\n model, device_ids=[args.gpu])\n return model\n\n\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser(description=\"recognition argument\")\n\n parser.add_argument('--batch_size', default=256, type=int, metavar='N',\n help='mini-batch size (default: 256), this is the total '\n 'batch size of all GPUs on the current node when '\n 'using Distributed Data Parallel')\n\n parser.add_argument(\"--seed\", type=int, default=0,\n help=\"Manual seed.\")\n\n parser.add_argument(\"--resume\", type=str, default=None,\n help=\"Path to location of checkpoint.\")\n\n parser.add_argument(\"--debug\", action=\"store_true\",\n help=\"Configure to debug settings, would overwrite most of the options.\")\n parser.add_argument(\"--h5py\", action=\"store_true\",\n help=\"Load data with H5py, defaultly use pickle (recommended).\")\n\n parser.add_argument(\"--config\", type=str, default=None, metavar='PATH',\n help=\"Path to configuration file of training procedure.\")\n\n parser.add_argument(\"--data\", type=str, default=None,\n help=\"Location of training/testing data.\")\n parser.add_argument(\"--dir\", type=str, default=None, metavar='PATH',\n help=\"Directory to save the log and model files.\")\n\n parser.add_argument('-p', '--print-freq', default=10, type=int,\n metavar='N', help='print frequency (default: 10)')\n\n parser.add_argument('-j', '--workers', default=4, type=int, metavar='N',\n help='number of data loading workers (default: 4)')\n parser.add_argument('--rank', default=-1, type=int,\n help='node rank for distributed training')\n parser.add_argument('--dist-url', default='tcp://127.0.0.1:13943', type=str,\n help='url used to set up distributed training')\n parser.add_argument('--dist-backend', default='nccl', type=str,\n help='distributed backend')\n parser.add_argument('--world-size', default=-1, type=int,\n help='number of nodes for distributed training')\n parser.add_argument('--gpu', default=None, type=int,\n help='GPU id to use.')\n\n args = parser.parse_args()\n\n SEED = args.seed\n torch.manual_seed(SEED)\n torch.cuda.manual_seed_all(SEED)\n np.random.seed(SEED)\n torch.backends.cudnn.deterministic = True\n\n if args.debug:\n utils.highlight_msg(\"Debugging.\")\n\n main(args)\n","sub_path":"scripts/ctc-crf/train.py","file_name":"train.py","file_ext":"py","file_size_in_byte":8508,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"267652629","text":"#! /usr/bin/python\r\n# encoding=UTF-8\r\n# version 3.x\r\n\r\n\"\"\"备份或根据备份信息比较目录的变化\"\"\"\r\n\r\nimport os, datetime, time, hashlib, re\r\nimport optparse\r\nfrom string import Template\r\n\r\n\r\ndef file_md5(file_path):\r\n md5 = None\r\n if os.path.isfile(file_path):\r\n f = open(file_path, 'rb')\r\n md5_obj = hashlib.md5()\r\n md5_obj.update(f.read())\r\n hash_code = md5_obj.hexdigest()\r\n f.close()\r\n md5 = str(hash_code).lower()\r\n return md5\r\n\r\n\r\ndef time_str(t):\r\n tt = datetime.datetime.fromtimestamp(t)\r\n return tt.strftime('%Y%m%d%H%M%S')\r\n\r\n\r\ndef file_info(f, path):\r\n \"\"\"获取文件信息\"\"\"\r\n path_len = len(path)\r\n info = Template(\"\"\"{'name':'${name}','dir':${dir},'md5':'${md5}'}\"\"\")\r\n name = re.sub('\\\\\\\\', '/', f[path_len:])\r\n # ctime = time_str(os.path.getctime(f))\r\n # mtime = time_str(os.path.getmtime(f))\r\n # atime = time(os.path.getatime(f))\r\n isdir = os.path.isdir(f)\r\n md5 = file_md5(f)\r\n return re.sub('\\s+', '', info.substitute(name=name, dir=isdir, md5=md5))\r\n\r\n\r\ndef dir_info(path):\r\n \"\"\"获取指定目录下的各个文件的信息\"\"\"\r\n da = []\r\n for root, dirs, files in os.walk(path):\r\n for dir in dirs:\r\n d = os.path.join(root, dir)\r\n da.append(file_info(d, path))\r\n for f in files:\r\n f = os.path.join(root, f)\r\n da.append(file_info(f, path))\r\n return da\r\n\r\n\r\ndef load_file_content(file):\r\n \"\"\"将文件内容已字符串的形式返回\"\"\"\r\n with open(file, 'r') as ff:\r\n return ff.read()\r\n\r\n\r\ndef usage():\r\n \"\"\"命令行帮助\"\"\"\r\n print(\"\"\"\r\n hash3compare.py \r\n usage:\r\n hash3compare.py -b -d dir 备份dir目录的信息\r\n hash3compare.py -c -f hash.txt -p dir 根据hash.txt的备份信息比较dir目录的变化\r\n hash3compare.py -c -s dir1 -p dir2 根据dir1为基础比较dir目录的变化\r\n params:\r\n -b back dir 备份目录信息\r\n -d dir 目录信息\r\n -c compare 对比变化\r\n -f hash file 从文件对比中对比\r\n -s resource path 比较的目录\r\n -p compare path 比较的目录 \r\n \"\"\")\r\n\r\n\r\nif __name__ == \"__main__\":\r\n parser = optparse.OptionParser()\r\n parser.add_option('-b', '--back', action=\"store_true\", default=False, dest=\"back\", help=\"back path info\")\r\n parser.add_option('-c', '--compare', action=\"store_true\", default=False, dest=\"compare\", help=\"compare path modify\")\r\n parser.add_option('-f', '--file', action=\"store\", dest=\"file\", help=\"compare modify with hash.txt\")\r\n parser.add_option('-s', '--source', action=\"store\", dest=\"source\", help=\"compare modify with this path\")\r\n parser.add_option('-p', '--path', action=\"store\", dest=\"path\", help='back or compare path')\r\n\r\n (options, args) = parser.parse_args()\r\n\r\n # print(options)\r\n if options.back and options.compare:\r\n usage()\r\n exit()\r\n elif options.back:\r\n if options.path and os.path.exists(os.path.abspath(options.path)):\r\n path = os.path.abspath(options.path)\r\n print(\"bak -> \" + path)\r\n dir_result = dir_info(path)\r\n cur_time = time.strftime('%Y%m%d%H%M%S', time.localtime())\r\n with open(\"hash_\" + cur_time + \".txt\", 'w') as ff:\r\n ff.write(str(dir_result))\r\n else:\r\n print(\"file not exists:\" + options.path)\r\n\r\n elif options.compare:\r\n if options.file and options.source:\r\n usage()\r\n else:\r\n base_path = os.path.abspath(options.file or options.source)\r\n path = os.path.abspath(options.path)\r\n print(\"\"\"compare modify:\\n\\tleft :%s \\n\\tright: %s\"\"\" % (base_path, path))\r\n dir_result = dir_info(path)\r\n if options.file:\r\n bak_result = eval(load_file_content(options.file))\r\n else:\r\n bak_result = dir_info(base_path)\r\n\r\n af, afmd5, bf, bfmd5 = set(), set(), set(), set()\r\n for a in dir_result:\r\n t = eval(a)\r\n af.add(t['name'])\r\n afmd5.add(t['name'] + \":\" + t['md5'])\r\n for b in bak_result:\r\n u = eval(b)\r\n bf.add(u['name'])\r\n bfmd5.add(u['name'] + \":\" + u['md5'])\r\n\r\n add_file_list = af.difference(bf)\r\n if len(add_file_list):\r\n print(\"===add file to right:\")\r\n for f in add_file_list:\r\n print(\"\\t %s\" % f)\r\n\r\n delete_file_list = bf.difference(af)\r\n if len(delete_file_list):\r\n print(\"===delete file form left:\")\r\n for f in delete_file_list:\r\n print(\"\\t %s\" % f)\r\n\r\n modify_file_list = afmd5.difference(bfmd5)\r\n if len(modify_file_list):\r\n print(\"===modify file:\")\r\n for f in modify_file_list:\r\n fn = f.split(':')[0]\r\n if fn not in add_file_list and fn not in delete_file_list:\r\n print(\"\\t\" + f)\r\n\r\n else:\r\n usage()\r\n","sub_path":"src/hash3compare.py","file_name":"hash3compare.py","file_ext":"py","file_size_in_byte":5291,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"195506064","text":"from sklearn import tree\n#from sklearn.metrics import accuracy_score\nimport pandas as pd\nfrom sklearn.model_selection import train_test_split\n\ndata = pd.read_csv('indians-diabetes.data', header = None, delimiter=' *, *', engine='python')\ndata.columns = ['num_pregent', 'glucose', 'bp', 'triceps', 'serum', 'bmi', 'dpf', 'age', 'class']\n\n#extract features and targets from the data\nfeatures = data.values[:,:8]\ntarget = data.values[:,8]\n\n#Split arrays or matrices into random train and test subsets\nfeatures_train, features_test, target_train, target_test = train_test_split(features, target, test_size = 0.33, random_state = 10)\n\n#Create a Tree Classifier\nclf = tree.DecisionTreeClassifier(min_samples_split=40)\n\n#Fit Tree classifier according to features_train, target_train\nclf.fit(features_train, target_train)\n\n#Perform classification on an array of test vectors \ntarget_pred = clf.predict([1,89,66,23,94,28.1,0.167,21])\n\nif target_pred == [1]:\n print(\"Diabetes positive\")\nelse:\n print(\"Diabetes negative\")\n\n#Returns the mean accuracy on the given test data and labels.\nprint (clf.score(features_test, target_test)*100)\n\n#print (accuracy_score(target_test, pred))","sub_path":"Diabetes_Predction/tree.py","file_name":"tree.py","file_ext":"py","file_size_in_byte":1173,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"88529778","text":"'''\nimport numpy as np \nimport matplotlib.pyplot as plt \nimport tensorflow.compat.v1 as tf \ntf.compat.v1.disable_eager_execution()\n\nACTIVATION = tf.nn.relu\t\t#每一层都使用relu\nN_LAYERS = 7\t\t\t\t#一共7层隐藏层\nN_HIDDEN_UNITS = 30\t\t\t#每个隐藏层有30个神经元\n\n#重复观看\ndef fix_seed(seed=1):\n\t#reporducible\n\tnp.random.seed(seed)\n\ttf.set_random_seed(seed)\n\n\n#打印图片\ndef plot_his(inputs, inputs_norm):\n\t#plot histogram for the inputs of every layer\n\tfor j, all_inputs in enumerate([inputs, inputs_norm]):\n\t\tfor i, input in enumerate(all_inputs):\n\t\t\tplt.subplot(2, len(all_inputs), j*len(all_inputs)+(i+1))\n\t\t\tplt.cla()\n\t\t\tif i == 0:\n\t\t\t\tthe_range = (-7,10)\n\t\t\telse:\n\t\t\t\tthe_range = (-1,1)\n\t\t\tplt.hist(input.ravel(), bins=15, range=the_range, color='#FF5733')\n\t\t\tplt.yticks(())\n\t\t\tif j == 1:\n\t\t\t\tplt.xticks(the_range)\n\t\t\telse:\n\t\t\t\tplt.xticks(())\n\t\t\tax = plt.gca()\n\t\t\tax.spines['right'].set_color('none')\n\t\t\tax.spines['top'].set_color('none')\n\t\tplt.title(\"%s normalizing\" % (\"Without\" if j == 0 else \"With\"))\n\tplt.draw()\n\tplt.pause(0.01)\n\n\n#建立神经网络\ndef built_net(xs, ys, norm):\n\tdef add_layer(inputs, in_size, out_size, activation_function=None, norm=False):\n\t\tWeights = tf.Variable(tf.random_normal([in_size, out_size], mean=0., stddev=1.))\n\t\tbiases = tf.Variable(tf.zeros([1, out_size]) + 0.1)\n\t\tWx_plus_b = tf.matmul(inputs, Weights) + biases\n\t\t\n\t\tif norm:\n\t\t\tfc_mean, fc_var = tf.nn.moments(Wx_plus_b, axes=[0])\n\t\t\tscale = tf.Variable(tf.ones([out_size]))\n\t\t\tshift = tf.Variable(tf.zeros([out_size]))\n\t\t\tepsilon = 0.001\n\n\t\t\tema = tf.train.ExponentialMovingAverage(decay=0.5)\n\t\t\tdef mean_var_with_update():\n\t\t\t\tema_apply_op = ema.apply([fc_mean, fc_var])\n\t\t\t\twith tf.control_dependencies([ema_apply_op]):\n\t\t\t\t\treturn tf.identity(fc_mean), tf.identity(fc_var)\n\t\t\tmean, var = mean_var_with_update()\n\n\n\t\t\tWx_plus_b = tf.nn.batch_normalization(Wx_plus_b, mean, var, shift, scale, epsilon)\n\n\n\n\t\tif activation_function is None:\n\t\t\toutputs = Wx_plus_b\n\t\telse:\n\t\t\toutputs = activation_function(Wx_plus_b)\n\n\t\treturn outputs\n\n\tfix_seed(1)\n\n\tif norm:\n\t\t# BN for the first input\n\t\tfc_mean, fc_var = tf.nn.moments(xs, axes=[0],)\n\t\tscale = tf.Variable(tf.ones([1]))\n\t\tshift = tf.Variable(tf.zeros([1]))\n\t\tepsilon = 0.001\n\t\t# apply moving average for mean and var when train on batch\n\t\tema = tf.train.ExponentialMovingAverage(decay=0.5)\n\t\tdef mean_var_with_update():\n\t\t\tema_apply_op = ema.apply([fc_mean, fc_var])\n\t\t\twith tf.control_dependencies([ema_apply_op]):\n\t\t\t\treturn tf.identity(fc_mean), tf.identity(fc_var)\n\t\tmean, var = mean_var_with_update()\n\t\txs = tf.nn.batch_normalization(xs, mean, var, shift, scale, epsilon)\n\n\n\t#recorde inputs for every layer \n\tlayers_inputs = [xs]\n\n\t#build hidden layers\n\tfor l_n in range(N_LAYERS):\n\t\tlayer_input = layers_inputs[l_n]\n\t\tin_size = layers_inputs[l_n].get_shape()[1].value\n\n\t\toutput = add_layer(layer_input, in_size, N_HIDDEN_UNITS, ACTIVATION, norm,)\n\t\tlayers_inputs.append(output)\n\n\n\tprediction = add_layer(layers_inputs[-1], 30, 1, activation_function=None)\n\n\tcost = tf.reduce_mean(tf.reduce_sum(tf.square(ys - prediction), reduction_indices=[1]))\n\ttrain_op = tf.train.GradientDescentOptimizer(0.001).minimize(cost)\n\treturn [train_op, cost, layers_inputs]\n\n\n#make up data\nfix_seed(1)\nx_data = np.linspace(-7, 10, 2500)[:, np.newaxis]\nnp.random.shuffle(x_data)\nnoise = np.random.normal(0, 8, x_data.shape)\ny_data = np.square(x_data) - 5 + noise\n\n\n#plot input data\nplt.scatter(x_data, y_data)\nplt.show()\n\nxs = tf.placeholder(tf.float32, [None, 1])\t\t#[num_samples, num_features]\nys = tf.placeholder(tf.float32, [None, 1])\n\ntrain_op, cost, layers_inputs = built_net(xs, ys, norm=False)\t\t\t\t\t#without BN\ntrain_op_norm, cost_norm, layers_inputs_norm = built_net(xs, ys, norm=True)\t\t#with BN\n\nsess = tf.Session()\ninit = tf.global_variables_initializer\nsess.run(init)\n\n\n#record cost\ncost_his = []\ncost_his_norm = []\nrecord_step = 5\n\nplt.ion()\nplt.figure(figsize=(7, 3))\nfor i in range(250):\n\tif i % 50 == 0:\n\t\tall_inputs, all_inputs_norm = sess.run([layers_inputs, layers_inputs_norm], feed_dict={xs: x_data, ys: y_data})\n\t\tplot_his(all_inputs, all_inputs_norm)\n\n\t# train on batch\n\tsess.run([train_op, train_op_norm], feed_dict={xs: x_data[i*10:i*10+10], ys: y_data[i*10:i*10+10]})\n\n\tif i % record_step == 0:\n\t\t#record cost\n\t\tcost_his.append(sess.run(cost, feed_dict={xs:x_data, ys:y_data}))\n\t\tcost_his_norm.append(sess.run(cost_norm, feed_dict={xs:x_data, ys:y_data}))\n\n\n\nplt.ioff()\nplt.figure()\nplt.plot(np.arange(len(cost_his))*record_step, np.array(cost_his), label='no BN') # no norm\nplt.plot(np.arange(len(cost_his))*record_step, np.array(cost_his_norm), label='BN') # norm\nplt.legend()\nplt.show()\n'''\n\n\nimport tensorflow.compat.v1 as tf \ntf.compat.v1.disable_eager_execution()\nimport numpy as np\nimport matplotlib.pyplot as plt\n\ntf.set_random_seed(1)\nnp.random.seed(1)\n\n# Hyper parameters\nN_SAMPLES = 2000\nBATCH_SIZE = 64\nEPOCH = 12\nLR = 0.03\nN_HIDDEN = 8\nACTIVATION = tf.nn.tanh\nB_INIT = tf.constant_initializer(-0.2) # use a bad bias initialization\n\n# training data\nx = np.linspace(-7, 10, N_SAMPLES)[:, np.newaxis]\nnp.random.shuffle(x)\nnoise = np.random.normal(0, 2, x.shape)\ny = np.square(x) - 5 + noise\ntrain_data = np.hstack((x, y))\n\n# test data\ntest_x = np.linspace(-7, 10, 200)[:, np.newaxis]\nnoise = np.random.normal(0, 2, test_x.shape)\ntest_y = np.square(test_x) - 5 + noise\n\n# plot input data\nplt.scatter(x, y, c='#FF9359', s=50, alpha=0.5, label='train')\nplt.legend(loc='upper left')\n\n# tensorflow placeholder\ntf_x = tf.placeholder(tf.float32, [None, 1])\ntf_y = tf.placeholder(tf.float32, [None, 1])\ntf_is_train = tf.placeholder(tf.bool, None) # flag for using BN on training or testing\n\n\nclass NN(object):\n def __init__(self, batch_normalization=False):\n self.is_bn = batch_normalization\n\n self.w_init = tf.random_normal_initializer(0., .1) # weights initialization\n self.pre_activation = [tf_x]\n if self.is_bn:\n self.layer_input = [tf.layers.batch_normalization(tf_x, training=tf_is_train)] # for input data\n else:\n self.layer_input = [tf_x]\n for i in range(N_HIDDEN): # adding hidden layers\n self.layer_input.append(self.add_layer(self.layer_input[-1], 10, ac=ACTIVATION))\n self.out = tf.layers.dense(self.layer_input[-1], 1, kernel_initializer=self.w_init, bias_initializer=B_INIT)\n self.loss = tf.losses.mean_squared_error(tf_y, self.out)\n\n # !! IMPORTANT !! the moving_mean and moving_variance need to be updated,\n # pass the update_ops with control_dependencies to the train_op\n update_ops = tf.get_collection(tf.GraphKeys.UPDATE_OPS)\n with tf.control_dependencies(update_ops):\n self.train = tf.train.AdamOptimizer(LR).minimize(self.loss)\n\n def add_layer(self, x, out_size, ac=None):\n x = tf.layers.dense(x, out_size, kernel_initializer=self.w_init, bias_initializer=B_INIT)\n self.pre_activation.append(x)\n # the momentum plays important rule. the default 0.99 is too high in this case!\n if self.is_bn: x = tf.layers.batch_normalization(x, momentum=0.4, training=tf_is_train) # when have BN\n out = x if ac is None else ac(x)\n return out\n\nnets = [NN(batch_normalization=False), NN(batch_normalization=True)] # two nets, with and without BN\n\nsess = tf.Session()\nsess.run(tf.global_variables_initializer())\n\n# plot layer input distribution\nf, axs = plt.subplots(4, N_HIDDEN+1, figsize=(10, 5))\nplt.ion() # something about plotting\n\ndef plot_histogram(l_in, l_in_bn, pre_ac, pre_ac_bn):\n for i, (ax_pa, ax_pa_bn, ax, ax_bn) in enumerate(zip(axs[0, :], axs[1, :], axs[2, :], axs[3, :])):\n [a.clear() for a in [ax_pa, ax_pa_bn, ax, ax_bn]]\n if i == 0: p_range = (-7, 10); the_range = (-7, 10)\n else: p_range = (-4, 4); the_range = (-1, 1)\n ax_pa.set_title('L' + str(i))\n ax_pa.hist(pre_ac[i].ravel(), bins=10, range=p_range, color='#FF9359', alpha=0.5)\n ax_pa_bn.hist(pre_ac_bn[i].ravel(), bins=10, range=p_range, color='#74BCFF', alpha=0.5)\n ax.hist(l_in[i].ravel(), bins=10, range=the_range, color='#FF9359')\n ax_bn.hist(l_in_bn[i].ravel(), bins=10, range=the_range, color='#74BCFF')\n for a in [ax_pa, ax, ax_pa_bn, ax_bn]:\n a.set_yticks(()); a.set_xticks(())\n ax_pa_bn.set_xticks(p_range); ax_bn.set_xticks(the_range); axs[2, 0].set_ylabel('Act'); axs[3, 0].set_ylabel('BN Act')\n plt.pause(0.01)\n\nlosses = [[], []] # record test loss\nfor epoch in range(EPOCH):\n print('Epoch: ', epoch)\n np.random.shuffle(train_data)\n step = 0\n in_epoch = True\n while in_epoch:\n b_s, b_f = (step*BATCH_SIZE) % len(train_data), ((step+1)*BATCH_SIZE) % len(train_data) # batch index\n step += 1\n if b_f < b_s:\n b_f = len(train_data)\n in_epoch = False\n b_x, b_y = train_data[b_s: b_f, 0:1], train_data[b_s: b_f, 1:2] # batch training data\n sess.run([nets[0].train, nets[1].train], {tf_x: b_x, tf_y: b_y, tf_is_train: True}) # train\n\n if step == 1:\n l0, l1, l_in, l_in_bn, pa, pa_bn = sess.run(\n [nets[0].loss, nets[1].loss, nets[0].layer_input, nets[1].layer_input,\n nets[0].pre_activation, nets[1].pre_activation],\n {tf_x: test_x, tf_y: test_y, tf_is_train: False})\n [loss.append(l) for loss, l in zip(losses, [l0, l1])] # recode test loss\n plot_histogram(l_in, l_in_bn, pa, pa_bn) # plot histogram\n\nplt.ioff()\n\n# plot test loss\nplt.figure(2)\nplt.plot(losses[0], c='#FF9359', lw=3, label='Original')\nplt.plot(losses[1], c='#74BCFF', lw=3, label='Batch Normalization')\nplt.ylabel('test loss'); plt.ylim((0, 2000)); plt.legend(loc='best')\n\n# plot prediction line\npred, pred_bn = sess.run([nets[0].out, nets[1].out], {tf_x: test_x, tf_is_train: False})\nplt.figure(3)\nplt.plot(test_x, pred, c='#FF9359', lw=4, label='Original')\nplt.plot(test_x, pred_bn, c='#74BCFF', lw=4, label='Batch Normalization')\nplt.scatter(x[:200], y[:200], c='r', s=50, alpha=0.2, label='train')\nplt.legend(loc='best'); plt.show()","sub_path":"TensorFlow/batch_n.py","file_name":"batch_n.py","file_ext":"py","file_size_in_byte":10134,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"351619735","text":"from sklearn.feature_extraction.text import CountVectorizer\nfrom math import floor\nimport numpy as np\nimport pandas as pd\nimport random\nimport pickle\nimport codecs\nimport csv\nimport os\n\n# CLEAN TEXT\n'''\n\npath = os.getcwd()+\"/gutenberg\"\n\ninitial_text2cut_1 = \"with this eBook or online at www.gutenberg.\"\ninitial_text2cut_2 = \"*END*THE SMALL PRINT! FOR PUBLIC DOMAIN ETEXTS*\"\n\nfinal_text2cut_1 = \"END OF THIS PROJECT GUTENBERG EBOOK\"\nfinal_text2cut_2 = \"END OF THE PROJECT GUTENBERG EBOOK\"\n\nwith open(\"gut_index.csv\", \"r\") as csv_file:\n reader = csv.DictReader(csv_file, delimiter=',')\n for book in reader:\n path_book = path[:]\n if len(book[\"id\"]) == 1:\n path_book += '/0/'+book[\"id\"]+'/'+book[\"id\"]+\".txt\"\n else:\n for i in range(len(book[\"id\"])-1):\n path_book += '/'+book[\"id\"][i]\n path_book += '/'+book[\"id\"]\n if path_book[-1] == 'C':\n path_book = path_book[:-1]\n try:\n possible_txts = os.listdir(path_book)\n if path_book.split(\"/\")[-1]+\".txt\" in possible_txts:\n path_book += \"/\"+book[\"id\"]+\".txt\"\n else:\n for possible_txt in possible_txts:\n if \".txt\" in possible_txt and not \"readme.txt\" in possible_txt:\n path_book += \"/\"+possible_txt\n break;\n except FileNotFoundError:\n pass\n\n try:\n with codecs.open(path_book, \"r\",encoding='utf-8', errors='ignore') as book_file:\n cv = CountVectorizer(min_df=2, stop_words='english',ngram_range=(5, 5), analyzer='word')\n analyzer = cv.build_tokenizer()\n\n #print(path_book)\n\n text = book_file.read()\n if initial_text2cut_1 in text:\n index_start = text.index(initial_text2cut_1)+len(initial_text2cut_1)+3\n elif initial_text2cut_2 in text:\n index_start = text.index(initial_text2cut_2)+len(initial_text2cut_2)+3\n else:\n index_start = 250\n \n if final_text2cut_1 in text:\n index_end = text.index(final_text2cut_1)\n elif final_text2cut_2 in text:\n index_end = text.index(final_text2cut_2)\n else:\n index_end = len(text)-200\n\n text = text[index_start:index_end]\n\n #print(index_start, index_end)\n \n tokens = analyzer(text)\n except FileNotFoundError as error:\n try:\n path_book = \"/\".join(path_book.split(\"/\")[:-2])+\"/\"+path_book.split(\"/\")[-1]+\"/\"+path_book.split(\"/\")[-1]+\".txt\"\n with codecs.open(path_book, \"r\",encoding='utf-8', errors='ignore') as book_file:\n cv = CountVectorizer(min_df=2, stop_words='english',ngram_range=(5, 5), analyzer='word')\n analyzer = cv.build_tokenizer()\n\n #print(path_book)\n\n text = book_file.read()\n if initial_text2cut_1 in text:\n index_start = text.index(initial_text2cut_1)+len(initial_text2cut_1)+3\n elif initial_text2cut_2 in text:\n index_start = text.index(initial_text2cut_2)+len(initial_text2cut_2)+3\n else:\n index_start = 250\n\n if final_text2cut_1 in text:\n index_end = text.index(final_text2cut_1)\n elif final_text2cut_2 in text:\n index_end = text.index(final_text2cut_2)\n else:\n index_end = len(text)-200\n \n text = text[index_start:index_end]\n #print(index_start, index_end)\n tokens = analyzer(text)\n except (IsADirectoryError, FileNotFoundError) as error:\n print(path_book+\" ERROR\")\n continue;\n except IsADirectoryError:\n continue;\n \n print(path_book+\" OK\")\n with open(os.getcwd()+\"/ngrams/\"+path_book.split(\"/\")[-1][:-4]+\".pkl\", 'wb') as f:\n pickle.dump(tokens, f)\n\n\n# CALCULATE SIMILARITY AND RANK BOOKS BY AUTHOR\n\n\ndf = pd.read_csv(os.getcwd()+\"/gut_index.csv\")\n\nbooks_by_author = df.groupby('author')\n\ncoincident_books = []\nless_coincident_books = []\ncoincident_books_list = []\n\nfor author, ids in books_by_author:\n ids = list(ids['id'])\n \n for i in range(len(ids)):\n for j in range(len(ids)):\n if i != j:\n directory_book_ref = os.getcwd()+\"/ngrams/\"+ids[i]\n directory_book_delete = os.getcwd()+\"/ngrams/\"+ids[j]\n\n if directory_book_ref[-1] == 'C':\n directory_book_ref = directory_book_ref[:-1]\n\n if directory_book_delete[-1] == 'C':\n directory_book_delete = directory_book_delete[:-1]\n try:\n with open(directory_book_ref+\".pkl\", 'rb') as book_ref:\n with open(directory_book_delete+\".pkl\", 'rb') as book_delete:\n book_ngrams = pickle.load(book_ref)\n book_possible_delete = pickle.load(book_delete)\n\n if(len(book_possible_delete)>0 and len(book_ngrams)>0):\n #max-min\n coincident_ngrams = len(set(book_ngrams).intersection(set(book_possible_delete)))/min(len(book_possible_delete), len(book_ngrams))\n coincident_books.append((author,ids[i],ids[j],coincident_ngrams))\n \n except FileNotFoundError as error:\n pass\n\n \n coincident_books.sort(key=lambda tup: tup[3])\n for i in range(len(coincident_books)):\n if(coincident_books[i][3] < 0.1):\n less_coincident_books.append(list(coincident_books[i]))\n\n if(len(less_coincident_books)>0):\n coincident_books_list += less_coincident_books\n print(ids)\n coincident_books = []\n less_coincident_books = []\n\nheaders = [\"author\",\"book1\",\"book2\",\"score\"]\ndf = pd.DataFrame(coincident_books_list, columns=headers)\n\nsum_scores = pd.pivot_table(df,index=[\"author\",\"book1\"],values=[\"score\"], aggfunc=np.sum).sort_values('score')\n\ndf.to_csv(\"book_scores_ordered.csv\", sep=',', encoding='utf-8')\n\n\n#os.remove()\n\n'''\n\ndef get_directory(book_id):\n \n path_book = os.getcwd()+\"/gutenberg\"\n\n if len(book_id) == 1:\n path_book += '/0/'+book_id+'/'+book_id+\".txt\"\n else:\n for i in range(len(book_id)-1):\n path_book += '/'+book_id[i]\n path_book += '/'+book_id\n if path_book[-1] == 'C':\n path_book = path_book[:-1]\n try:\n possible_txts = os.listdir(path_book)\n if path_book.split(\"/\")[-1]+\".txt\" in possible_txts:\n path_book += \"/\"+book_id+\".txt\"\n elif path_book.split(\"/\")[-1]+\"-0.txt\" in possible_txts:\n path_book += \"/\"+book_id+\"-0.txt\"\n elif path_book.split(\"/\")[-1]+\"-8.txt\" in possible_txts:\n path_book += \"/\"+book_id+\"-8.txt\"\n except FileNotFoundError:\n pass\n\n return path_book\n\nrandom.seed(9001)\n\n# Read scores by author\ndf = pd.read_csv(\"book_scores_ordered.csv\", sep=',')\n\n# Drop useless column\ndf.drop(['Unnamed: 0'], inplace=True, axis=1)\n\n#df_f = df.groupby(['author']).agg(['count'])\n\n# Filter authors with al least 3, 6 and 11 authors\ndf3 = df.groupby(['author']).filter(lambda x: x['book1'].nunique() > 2)\ndf6 = df.groupby(['author']).filter(lambda x: x['book1'].nunique() > 5)\ndf11 = df.groupby(['author']).filter(lambda x: x['book1'].nunique() > 10)\n\n# Sum of scores by book\ndf3_sum = pd.pivot_table(df3,index=[\"author\",\"book1\"],values=[\"score\"], aggfunc=np.sum)\ndf6_sum = pd.pivot_table(df6,index=[\"author\",\"book1\"],values=[\"score\"], aggfunc=np.sum)\ndf11_sum = pd.pivot_table(df11,index=[\"author\",\"book1\"],values=[\"score\"], aggfunc=np.sum)\n\n# Get list of all authors\ndistinct_authors_3 = np.array(df3.author.unique())\ntotal_size_3 = len(distinct_authors_3)\n\ndistinct_authors_6 = np.array(df6.author.unique())\ntotal_size_6 = len(distinct_authors_6)\n\ndistinct_authors_11 = np.array(df11.author.unique())\ntotal_size_11 = len(distinct_authors_11)\n\n\n# From the N remaining authors we take 100 randomly and iteratively until we consume the N authors. From those 100 authors,\n# we take once again 10 authors randomly which will be used to form the matrix of the model\n\n# Number of author from which 10 (selection_f) authors will be finally selected to be part of the matrix of the model\nselection_r = 100\n\n# Number of authors selected to be part of the matrix of the model\nselection_f = 10\n\n# Remaining books to select once the max/min score books are selected (2 for 3 books, 5 for 6 books and 10 for 11 books)\nremaining_f = 5\n\n# Choose size\ntotal_size = total_size_6\n\n# Choose authors size\ndistinct_authors = distinct_authors_6[:]\n\n# Number of iterations\nnum_iterations = floor(total_size/selection_r)\n\n# Lists containing the final 10 authors of all the iterations\nauthor_names_list = []\n\nprint(\"Total of {} authors\".format(total_size))\n\n# Generate 10 random indexes to select 'selection_f = 10' authors\nfor i in range(num_iterations):\n\n # take randomly selection_r indexes\n selected_authors_indexes = random.sample(range(total_size), selection_r)\n\n # get the author names associated to the indexes generated\n selected_authors_names = distinct_authors[selected_authors_indexes]\n\n # delete the author names selected from the list which contains all the author names\n distinct_authors = np.delete(distinct_authors, (selected_authors_indexes), axis=0)\n\n # reduce total size after deleting the selected names\n total_size = len(distinct_authors)\n\n # take randomly selection_f indexes out of selection_r\n selected_authors_indexes = random.sample(range(selection_r), selection_f)\n\n # get the author names associated to the indexes generated\n selected_authors_names = selected_authors_names[selected_authors_indexes]\n\n # save the 10 names selected in current iteration\n author_names_list.append(selected_authors_names)\n\n# Get the 3, 6 or 11 books from each author\n# When 3: 1-Min Known text and 2-Random for matrix, 1-Max Known text and 2-Random for matrix, 1-Random Known text and 2-Random for matrix\n# When 6: 1-Min Known text and 5-Random for matrix, 1-Max Known text and 5-Random for matrix, 1-Random Known text and 5-Random for matrix\n# When 11: 1-Min Known text and 10-Random for matrix, 1-Max Known text and 10-Random for matrix, 1-Random Known text and 10-Random for matrix\n\n'''\n# Min\nprint(\"\\n\\nMIM\\n~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\")\nfor i in range(len(author_names_list)):\n print(\"iteration: \"+str(i+1))\n print(\"\\n\")\n with open(\"gut_min_max_rand/size3/min/size3_min_\" +str(i+1)+ \".csv\", \"w\") as csv_file:\n writer = csv.writer(csv_file, delimiter=',')\n writer.writerow([\"author\", \"book\", \"special\", \"directory\"])\n for author in author_names_list[i]:\n print(\"________________________________________________________________________\")\n print(author+\"\\n\")\n author_df = pd.DataFrame(df3_sum.loc[author].to_records()).sort_values('score')\n\n print(\"\\nMin\")\n print(author_df.iloc[0]['book1'])\n writer.writerow([author, author_df.iloc[0]['book1'], \"True\", get_directory(author_df.iloc[0]['book1'])])\n\n author_df = author_df.drop(author_df.index[0])\n remaining_indexes = random.sample(range(len(author_df)), remaining_f)\n remaining_books = author_df.iloc[remaining_indexes]\n print(\"\\nRemaining\")\n\n for row in remaining_books.to_records():\n print(row[1])\n writer.writerow([author, row[1], \"False\", get_directory(row[1])])\n\n\n# Max\nprint(\"\\n\\nMAX\\n~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\")\nfor i in range(len(author_names_list)):\n print(\"iteration: \"+str(i+1))\n print(\"\\n\")\n with open(\"gut_min_max_rand/size3/max/size3_max_\" +str(i+1)+ \".csv\", \"w\") as csv_file:\n writer = csv.writer(csv_file, delimiter=',')\n writer.writerow([\"author\", \"book\", \"special\", \"directory\"])\n for author in author_names_list[i]:\n print(\"________________________________________________________________________\")\n print(author+\"\\n\")\n author_df = pd.DataFrame(df3_sum.loc[author].to_records()).sort_values('score')\n\n print(\"\\nMax\")\n print(author_df.iloc[-1]['book1'])\n writer.writerow([author, author_df.iloc[-1]['book1'], \"True\", get_directory(author_df.iloc[-1]['book1'])])\n\n author_df = author_df.drop(author_df.index[-1])\n remaining_indexes = random.sample(range(len(author_df)), remaining_f)\n remaining_books = author_df.iloc[remaining_indexes]\n print(\"\\nRemaining\")\n\n for row in remaining_books.to_records():\n print(row[1])\n writer.writerow([author, row[1], \"False\", get_directory(row[1])])\n\n'''\n# Random\nprint(\"\\n\\nRANDOM\\n~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\")\nfor i in range(len(author_names_list)):\n print(\"iteration: \"+str(i+1))\n print(\"\\n\")\n with open(\"gut_min_max_rand/size6/rand/size6_rand_\" +str(i+1)+ \".csv\", \"w\") as csv_file:\n writer = csv.writer(csv_file, delimiter=',')\n writer.writerow([\"author\", \"book\", \"special\", \"directory\"])\n for author in author_names_list[i]:\n print(\"________________________________________________________________________\")\n print(author+\"\\n\")\n author_df = pd.DataFrame(df6_sum.loc[author].to_records())\n\n print(author_df)\n print(\"len: \", len(author_df), \"selection_factor: \", remaining_f+1)\n remaining_indexes = random.sample(range(len(author_df)), remaining_f+1)\n remaining_books = author_df.iloc[remaining_indexes]\n\n for row in remaining_books.to_records():\n writer.writerow([author, row[1], \"False\", get_directory(row[1])])\n\n","sub_path":"gut_ngram.py","file_name":"gut_ngram.py","file_ext":"py","file_size_in_byte":14429,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"294001799","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Tue Apr 14 13:43:36 2020\r\n\r\n@author: bmsri\r\n\"\"\"\r\n\r\nimport pandas as pd\r\nimport numpy as np\r\nimport os\r\nimport cv2\r\nimport imageio\r\n\r\nfrom sklearn.utils import shuffle\r\nfrom sklearn.model_selection import train_test_split\r\nimport shutil\r\nimport matplotlib.pyplot as plt\r\nfrom keras.preprocessing.image import ImageDataGenerator\r\n\r\nNUM_AUG_IMAGES_WANTED = 10000\r\n\r\nIMAGE_HEIGHT = 200\r\nIMAGE_WIDTH = 200\r\n\r\ntrain_image_list = os.listdir('D:\\\\PR_term_proj\\\\chest-xray-pneumonia\\\\chest_xray\\\\train')\r\n\r\ndf_train = pd.DataFrame(train_image_list, columns=['image_path'])\r\n\r\ndf_train.reset_index(inplace=True, drop=True)\r\n\r\n\r\n \r\ntrain_img_path = 'D:\\\\PR_term_proj\\\\chest-xray-pneumonia\\\\chest_xray\\\\train' \r\n\r\n\r\nbase_dir = 'base_dir'\r\nos.mkdir(base_dir)\r\n\r\ntrain_dir = os.path.join(base_dir, 'train_dir')\r\nos.mkdir(train_dir)\r\n\r\n\r\nNormal = os.path.join(train_dir, 'NORMAL')\r\nos.mkdir(Normal)\r\nPneumonia = os.path.join(train_dir, 'PNEUMONIA')\r\nos.mkdir(Pneumonia)\r\n\r\n\r\nfolder_1 = os.listdir(train_img_path)\r\n\r\ntrain_path_norm = os.path.join(train_img_path, 'NORMAL')\r\ntrain_path_pne = os.path.join(train_img_path, 'PNEUMONIA')\r\n\r\ntrain_list_norm = sorted(os.listdir(train_path_norm))\r\ntrain_list_pne = sorted(os.listdir(train_path_pne))\r\n\r\n\r\ncomp_path = 'D:\\\\PR_term_proj\\\\'\r\n\r\n#df_data.set_index('image_path', inplace=True)\r\nfrom PIL import Image\r\n\r\nfor image in train_list_norm: \r\n fname = image\r\n #label = df_data.loc[image, 'target']\r\n \r\n #if fname in folder_1:\r\n \r\n src = train_path_norm + '\\\\' + fname\r\n dst = comp_path + train_dir + '\\\\' + \"NORMAL\" + '\\\\' + fname\r\n \r\n image = Image.open(src)\r\n image=image.resize([200,200])\r\n print(np.shape(image))\r\n if (len(np.shape(image)) == 2):\r\n image.save(dst)\r\n \r\n\r\nfor image in train_list_pne: \r\n fname = image\r\n \r\n src = train_path_pne + '\\\\' + fname\r\n dst = comp_path + train_dir + '\\\\' + \"PNEUMONIA\" + '\\\\' + fname\r\n \r\n image = Image.open(src)\r\n# print(np.shape(image))\r\n image=image.resize([200,200])\r\n print(np.shape(image))\r\n if (len(np.shape(image)) == 2):\r\n image.save(dst)\r\n\r\n\r\n'''\r\nData Augmentation\r\n''' \r\naug_dir = 'aug_dir'\r\nos.mkdir(aug_dir)\r\n \r\nclass_list = ['NORMAL','PNEUMONIA']\r\n\r\nfor item in class_list:\r\n \r\n \r\n img_class = item\r\n\r\n img_list = sorted(os.listdir(comp_path + 'base_dir\\\\train_dir\\\\' + img_class))\r\n\r\n os.mkdir(aug_dir+ \"\\\\\" + item)\r\n##\r\n for fname in img_list:\r\n src = os.path.join(comp_path + 'base_dir\\\\train_dir\\\\' + img_class, fname)\r\n dst = os.path.join(aug_dir+ \"\\\\\" + item, fname)\r\n shutil.copyfile(src, dst)\r\n\r\n path = aug_dir+ \"\\\\\"# + item + \"\\\\\"\r\n save_path = comp_path + 'base_dir\\\\train_dir\\\\' + img_class\r\n\r\n datagen = ImageDataGenerator(rotation_range=40, width_shift_range=0.2, height_shift_range=0.2, shear_range=0.2, zoom_range=0.2, horizontal_flip=True)\r\n#rescale=1/255, \r\n batch_size = 50\r\n \r\n aug_datagen = datagen.flow_from_directory(path, save_to_dir=aug_dir+ \"\\\\\" + item + \"\\\\\", save_format='jpg', target_size=(IMAGE_HEIGHT,IMAGE_WIDTH), batch_size=batch_size)\r\n \r\n num_files = len(os.listdir(aug_dir+ \"\\\\\" + item))\r\n \r\n num_batches = int(np.ceil((NUM_AUG_IMAGES_WANTED-num_files)/batch_size))\r\n\r\n for i in range(0,num_batches):\r\n #imgs = Image.open(aug_datagen)\r\n imgs, labels = next(aug_datagen)\r\n print(np.shape(imgs))\r\n\r\n \r\n","sub_path":"data_gen_for_aug.py","file_name":"data_gen_for_aug.py","file_ext":"py","file_size_in_byte":3486,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"275384058","text":"import scipy.misc as misc\nimport numpy as np\nimport time\nimport numpy as np\nfrom glob import glob\nimport os\nimport cv2\n\n\n\ndef resize(image,image_width):\n # image = np.int(255*image)\n image = cv2.resize(image, (image_width, image_width), interpolation=cv2.INTER_LINEAR)\n return image\n\n\ndef images_selection(file_name, image_width, image_channel, batch_size, num_generations, support_number):\n filenames = glob(os.path.join(file_name, '*.*'))\n fake_categories = len(filenames) * batch_size\n fake_images = np.zeros([fake_categories * num_generations, image_width, image_width, image_channel])\n for i,image_path in enumerate(filenames):\n store_name = file_name + '_split/'\n if not os.path.exists(store_name):\n os.mkdir(store_name)\n current_x = misc.imread(image_path)\n image_size = int(np.shape(current_x)[0]/ batch_size)\n for j in range(batch_size):\n for k in range(support_number+num_generations):\n current_iamge = current_x[image_size*j:image_size*(j+1),image_size*(k):image_size*(k+1)]\n current_iamge = resize(current_iamge,128)\n # if len(np.shape(current_iamge))<3:\n # current_iamge = np.expand_dims(current_iamge,axis=-1)\n current_name = store_name + image_path.split('/')[-1].split('png')[0] + 'batch{}_sample{}.png'.format(j,k) \n misc.imsave(current_name, current_iamge)\n\n\n\nfile_name = '/media/user/05e85ab6-e43e-4f2a-bc7b-fad887cfe312/meta_gan/MatchingGAN-SelfAttention-XS/VISUALIZATION/vggface/1shot/visual_outputs/'\nimage_width = 128\nimage_channel =3\nbatch_size = 20\nnum_generations = 60\nsupport_number = 3\nimages_selection(file_name, image_width, image_channel, batch_size, num_generations, support_number)","sub_path":"visluazation_images_selection.py","file_name":"visluazation_images_selection.py","file_ext":"py","file_size_in_byte":1790,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"619215589","text":"import sys\nimport os\nimport cv2\n\nfrom pysvso.py_pose_graph.mapblock import RuntimeBlock\nfrom pysvso.system_tracker.tracker import SVSOTracker, Tracker\nfrom pysvso.graphics.viewer import PCViewer\nfrom pysvso.validation_toolkit.tum import Program as GTReader\nfrom pysvso.py_parallel_tasking_sched.threadingpool import ThreadingPool\n\nfrom pysvso.lib.log import init_logging\nimport logging\n\ninit_logging()\n\nfrom pysvso.config import Settings\n\nsettings = Settings()\nprint(settings)\n\nProject_base = settings.PROJECT_ROOT\nCamera_device = settings.CAMERA_DEVICE\n\nclass System:\n\n def __init__(self):\n # construct a map block, where typically raw map data should be read from this point\n block = RuntimeBlock()\n self._block = block\n block.load_device(Camera_device)\n\n # initialize a tracker\n tracker = SVSOTracker().set_FromMap(block)\n self._tracker = tracker\n\n # reading ground truth\n # NOTE: modify \"global_settings.py\" to decide whether you want to use ground truth\n trajectories, timestamps, depth_images = GTReader()\n self._trajectories = trajectories\n self._timestamps = timestamps\n self._depth_images = depth_images\n\n # set ground truth we loaded before\n tracker.trajectories_gt = trajectories\n\n # set depth images\n tracker.depth_images = depth_images\n\n # viewer\n self._legacy_viewer = None\n\n self._viewer = PCViewer()\n viewer = self._viewer\n\n viewer.set_FromTracker(tracker)\n viewer.set_FromMap(block)\n # viewer.Init()\n # Pycharm has some problems to run this snippet of codes\n viewer.Start()\n print(\"[Main Thread] viewer: \", viewer)\n print(\"\\n\\n\")\n pass\n\n def run(self):\n tracker = self._tracker\n timestamps = self._timestamps\n viewer = self._viewer\n\n DRAW_ONCE = False\n\n # ORBSLAM2 first successfully triangulated frame no: #91\n\n SAVER = settings.SAVER\n if not os.path.isdir(SAVER):\n os.makedirs(SAVER)\n\n # allocating system resources\n pool = ThreadingPool(5)\n\n # pool.add_task(viewer.Start)\n\n # experiments controls\n cnt = 0 # total frames seq\n cnt0 = 0 # tracked frames seq\n\n START_FRAMES = 0\n STOP_FRAMES = 500\n TRACKED_FRAMES = 400\n\n STEP = 5\n\n P1 = START_FRAMES\n P2 = START_FRAMES + 3 * STEP # 3\n P3 = START_FRAMES + 50 * STEP # 50\n\n capture = cv2.VideoCapture(os.path.join(settings.VIDEO_DIR, settings.VIDEO_NAME))\n\n while True:\n ret, frame = capture.read()\n if not ret:\n break\n\n timestamp = float(timestamps[cnt][0])\n file_name = timestamps[cnt][1]\n\n cnt += 1\n if cnt % STEP != 1:\n # comment this line if you want track images continuously\n pass\n # continue\n\n if cnt < START_FRAMES:\n continue\n\n # Relocalization control\n if cnt > P2 and cnt < P3:\n # creating missing gaps\n logging.info(\"skiping seqences from %d to %d, cur %d\" % (\n P2, P3, cnt\n ))\n continue\n\n # Trigger relocalization mode\n if cnt == P3:\n logging.info(\"manually switch state\")\n tracker.state = Tracker.State.LOSTED\n DRAW_ONCE = False\n # store the legacy\n self._legacy_viewer = viewer\n # create a new viewer\n self._viewer = PCViewer()\n viewer = self._viewer\n viewer.set_FromTracker(tracker)\n viewer.set_FromMap(tracker._map)\n # viewer.Init()\n # Pycharm has some problems to run this snippet of codes\n viewer.Start()\n\n # the tracker is in relocalization mode and not initialized\n if tracker._is_relocalization_mode and not tracker.isInitialized():\n # I don't wan to use a new map here\n viewer.set_FromMap(tracker._map)\n pass\n\n # @todo : TODO fix encoding error\n logging.info(\"exec tracker to track motions: %d\" % cnt)\n # Hybrid of OpticalFlow and Kalman Filter predictor & deep features based Hungarian algorithm implementation, Updated on Feb 26 2020 by Author Lei Wang\n tracker.track(frame, timestamp=timestamp)\n cnt0 += 1\n\n if tracker.cur is None: # relocalization triggered!\n continue\n\n ### Deprecated codes, in favor of new implementation of WebImagaRenderer ###\n # offline task\n if tracker.cur is not None and (\n not hasattr(tracker.cur, \"rendered_img\") or tracker.cur.rendered_img is None):\n logging.info(\"skipping rendered_img at Frame#%d ...\" % tracker.cur.seq)\n continue\n\n if not tracker.cur.is_First and tracker.cur.isKeyFrame:\n if tracker.cur.depth_img is not None and not DRAW_ONCE:\n viewer.drawDepthImage(tracker.cur)\n DRAW_ONCE = True\n logging.info(\"Scheduling task of updating point cloud viewer at KeyFrame %s\" % tracker.cur)\n\n def update_map(cv, cur=None, last=None, flow_mask=None, active_frames=None):\n with cv:\n logging.info(\"Waiting for completion of updating map ...\")\n cv.wait_for(\n lambda: tracker._map.complete_updating) # awaken by local mapper, and check complete_updating variable\n logging.info(\"Updating ...\")\n viewer.Update(cur=cur, last=last, flow_mask=flow_mask, active_frames=active_frames)\n logging.info(\"Point cloud viewer updated at KeyFrame %s\" % tracker.cur)\n tracker._map.complete_updating = False\n\n if not pool.tasks.full():\n pool.add_task(update_map, tracker._map.update_map_condition,\n cur=tracker.cur, last=tracker.last_frame, flow_mask=tracker.flow_mask,\n active_frames=tracker._map.get_active_frames())\n else:\n logging.info(\"Tasks are full. Cannot push tasks into the pool.\")\n\n if not tracker.cur.is_First and not tracker.cur.isKeyFrame:\n # update pose\n viewer.Update()\n pass\n\n # rets = model.detect([tracker.last_frame.img], verbose=1)\n # r = rets[0]\n # visualize.display_instances(frame, r['rois'], r['masks'], r['class_ids'], class_names, r['scores'])\n # out_frame = save_instances(tracker.last_frame.img, r['rois'], r['masks'], r['class_ids'], class_names, r['scores'])\n\n name = os.path.join(settings.SAVER, \"images/{}.jpg\".format(cnt))\n if tracker.last_frame is None or not hasattr(tracker.last_frame,\n \"rendered_img\") or tracker.last_frame.rendered_img is None:\n out_frame = tracker.cur.rendered_img\n else:\n out_frame = cv2.add(tracker.last_frame.rendered_img, tracker.flow_mask)\n cv2.imwrite(name, out_frame)\n\n if cnt >= STOP_FRAMES:\n logging.info(\"break after %d frames\" % STOP_FRAMES)\n break\n\n if cnt0 >= TRACKED_FRAMES:\n logging.info(\"break after %d frames tracked\" % TRACKED_FRAMES)\n break\n\n if tracker.isInitialized():\n continue\n\n logging.info(\"complete reading video.\")\n capture.release()\n # viewer.Stop()\n pass\n\n\nif __name__ == \"__main__\":\n system = System()\n system.run()","sub_path":"python/pysvso/system_tracker/system.py","file_name":"system.py","file_ext":"py","file_size_in_byte":7968,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"154882531","text":"from collections import deque\n\n# 변수 선언 및 입력\nn, L, R = tuple(map(int, input().split()))\n\negg = [\n list(map(int, input().split()))\n for _ in range(n)\n]\n\nbfs_q = deque()\negg_group = []\nvisited = [\n [False for _ in range(n)]\n for _ in range(n)\n]\n\n\ndef in_range(x, y):\n return 0 <= x and x < n and 0 <= y and y < n\n\n\ndef can_go(x, y, curr_egg):\n if not in_range(x, y):\n return False\n\n egg_diff = abs(egg[x][y] - curr_egg)\n return not visited[x][y] \\\n and L <= egg_diff and egg_diff <= R\n\n\n# visited 배열을 초기화 해줍니다.\ndef initialize_visited():\n for i in range(n):\n for j in range(n):\n visited[i][j] = False\n\n\ndef bfs():\n dxs, dys = [0, 1, 0, -1], [1, 0, -1, 0]\n\n # BFS 탐색을 수행합니다.\n while bfs_q:\n curr_x, curr_y = bfs_q.popleft()\n\n for dx, dy in zip(dxs, dys):\n new_x, new_y = curr_x + dx, curr_y + dy\n\n # L, R 사이인 경우에만 합쳐질 수 있습니다.\n if can_go(new_x, new_y, egg[curr_x][curr_y]):\n bfs_q.append((new_x, new_y))\n egg_group.append((new_x, new_y))\n visited[new_x][new_y] = True\n\n\n# 계란들을 합칩니다.\ndef merge_eggs():\n sum_of_eggs = sum([\n egg[x][y]\n for x, y in egg_group\n ])\n\n for x, y in egg_group:\n egg[x][y] = sum_of_eggs // len(egg_group)\n\n\n# 조건에 맞게 계란의 양을 바꿔줍니다.\ndef move_eggs():\n global egg_group\n\n # BFS 탐색을 위한 초기화 작업을 수행합니다.\n initialize_visited()\n\n is_changed = False\n\n # 아직 방문하지 못한 칸에 대해\n # BFS 탐색을 통해 합쳐질 계란들을 찾아냅니다.\n for i in range(n):\n for j in range(n):\n if not visited[i][j]:\n # 합쳐질 계란 목록을 담을 곳을 초기화합니다.\n egg_group = []\n\n bfs_q.append((i, j))\n egg_group.append((i, j))\n visited[i][j] = True\n\n bfs()\n\n # 계란의 이동이 한번이라도 일어났는지를 확인합니다.\n if len(egg_group) > 1:\n is_changed = True\n\n # (i, j)와 관련이 있는 계란들을 합칩니다.\n merge_eggs()\n\n return is_changed\n\n\nmove_cnt = 0\n\n# 이동이 더 이상 필요 없을 때까지\n# 계란의 이동을 반복합니다.\nwhile True:\n is_changed = move_eggs()\n if not is_changed:\n break\n\n move_cnt += 1\n\nprint(move_cnt)\n","sub_path":"samsung/11/17 토스트 계란틀/33.py","file_name":"33.py","file_ext":"py","file_size_in_byte":2574,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"640850017","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Jan 29 23:50:57 2021\n\n@author: Aditya Mishra\n\"\"\"\n\n# Necessary Libraries\nimport pandas as pd\nimport matplotlib.pyplot as plt\nfrom sklearn.datasets import load_diabetes\nfrom sklearn.model_selection import train_test_split\nfrom sklearn.linear_model import LinearRegression as Sk_linear_reg\nfrom sklearn.metrics import mean_squared_error\nfrom my_linear_model import LinearRegression\n\n# Dataset\ndata = load_diabetes()\ndf = pd.DataFrame(data.data, columns=data.feature_names)\ndf['target'] = data.target\nprint('Describe Dataset\\n', df.describe())\n\n# Model\nepoch = 10_000\nmy_clf = LinearRegression(max_iter=epoch, optimizer='bgd')\nreg_clf = Sk_linear_reg()\n\nx_train, x_test, y_train, y_test = train_test_split(df.drop(['target'], axis=1), df['target'], test_size=0.3)\nmy_clf.fit(x_train, y_train)\nreg_clf.fit(x_train, y_train)\n\nmy_pred = my_clf.predict(x_test)\nreg_pred = reg_clf.predict(x_test)\n\n# Model Test Result\nprint(f\"My Model's MSE: {my_clf.mse(y_test, my_pred):.3f}\")\nprint(f\"Sklearn Model's MSE: {mean_squared_error(y_test, reg_pred):.3f}\")\n\n# Error Plot\nplt.title(\"Error in K-epochs\")\nplt.plot(range(epoch), my_clf.error_, 'r-')\nplt.xlabel('Epochs')\nplt.ylabel('Error')\nplt.show()\n","sub_path":"driver.py","file_name":"driver.py","file_ext":"py","file_size_in_byte":1227,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"148001241","text":"import hlt\nimport logging\nfrom time import time\n\n#SETUP\n\nclass Bot:\n def __init__(self, entity, my_id, myShips):\n self.ship = entity\n \n self.playerID = my_id\n self.teamShips = myShips\n \n def findNearbyEntities(self, gameMap):\n self.entitiesDistance = gameMap.nearby_entities_by_distance(self.ship)\n \n def sortNearbyEntities(self):\n self.planetsOfInterest = []\n self.closestShip = None\n tempSortedEntities = sorted(self.entitiesDistance)\n found = False\n for distance in tempSortedEntities:\n entity = self.entitiesDistance[distance][0]\n if self.isPlanet(entity, self.playerID):\n \n self.planetsOfInterest.append((entity, distance)) \n if self.isShip(entity, self.playerID, self.teamShips) and not found:\n found = True\n self.closestShip = (entity, distance)\n\n def getPriorities(self, gameMap):\n self.findNearbyEntities(gameMap)\n self.sortNearbyEntities()\n self.priorities = self.planetsOfInterest\n self.priorities.append(self.closestShip)\n if self.closestShip[1] < 25:\n self.priorities.reverse()\n return self.priorities\n\n\n def isPlanet(self, entity, myID):\n if isinstance(entity, hlt.entity.Planet):\n return (not entity.is_owned() or (entity.owner.id == myID and not entity.is_full()))\n return False\n \n def isShip(self,entity, myID, teamShips):\n return isinstance(entity, hlt.entity.Ship) and not entity in teamShips\n\nclass BotController:\n def __init__(self):\n self.game = hlt.Game(\"Settler\")\n\n self.command_queue = []\n \n def update(self):\n self.start = time()\n self.gameMap = self.game.update_map()\n self.myID = self.gameMap.my_id\n self.myBots = []\n self.command_queue = []\n \n def getShips(self):\n myShips = []\n \n for ship in self.gameMap.get_me().all_ships():\n myShips.append(ship)\n\n return myShips\n\n def createBots(self):\n ships = self.getShips()\n self.myBots = []\n for bot in ships:\n self.myBots.append(Bot(bot,self.myID,ships))\n\n def getBotPriorities(self):\n botPriorities = {}\n for bot in self.myBots:\n botPriorities[bot.ship] = bot.getPriorities(self.gameMap)\n return self.sortBotPriorities(botPriorities)\n\n def sortBotPriorities(self, botPriorities):\n tempBotPriorities = sorted(botPriorities, key = lambda bot:botPriorities[bot][0][1])\n finalBotPriorities = {}\n for bot in tempBotPriorities:\n finalBotPriorities[bot] = botPriorities[bot]\n return finalBotPriorities\n\n def getPlanetSpaces(self):\n planetSpaces = {}\n for planet in self.gameMap.all_planets():\n planetSpaces[planet] = planet.num_docking_spots-len(planet._docked_ship_ids)\n return planetSpaces\n\n def getCommands(self): \n priorities = self.getBotPriorities()\n planetSpaces = self.getPlanetSpaces()\n for bot in priorities:\n if bot.docking_status != bot.DockingStatus.UNDOCKED:\n continue\n navigateCommand = None\n for preference in priorities[bot]:\n target = preference[0]\n if self.isPlanet(target):\n if bot.can_dock(target):\n planetSpaces[target] -= 1\n navigateCommand = bot.dock(target)\n break\n if planetSpaces[target] > 0:\n planetSpaces[target] -= 1\n navigateCommand = bot.navigate(\n bot.closest_point_to(target),\n self.gameMap,\n speed=int(hlt.constants.MAX_SPEED),\n ignore_ships=False)\n break\n else:\n #Attack Ship Code\n navigateCommand = bot.navigate(\n bot.closest_point_to(target),\n self.gameMap,\n speed=int(hlt.constants.MAX_SPEED),\n ignore_ships=False)\n break\n if navigateCommand:\n self.command_queue.append(navigateCommand)\n \n def sendCommands(self):\n logging.info(self.start-time())\n self.game.send_command_queue(self.command_queue)\n \n\n def isPlanet(self,entity):\n return isinstance(entity, hlt.entity.Planet)\n\n#START GAME\nController = BotController()\n\nlogging.info(\"Initialising Robbot...\")\n\n\nwhile True:\n Controller.update()\n Controller.createBots()\n Controller.getCommands()\n Controller.sendCommands()\n \n \n \n \n","sub_path":"MyBot-v8.py","file_name":"MyBot-v8.py","file_ext":"py","file_size_in_byte":4928,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"552093519","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n# pylint: disable=wrong-import-position,import-error\n\nimport re\nimport time\nimport threading\nimport subprocess\nimport pkg_resources\nimport gi # isort:skip\ngi.require_version('Gtk', '3.0') # isort:skip\nfrom gi.repository import Gtk, Gdk, GObject, Gio # isort:skip\nfrom qubesadmin import Qubes\nfrom qubesadmin import exc\n\n# using locale.gettext is necessary for Gtk.Builder translation support to work\n# in most cases gettext is better, but it cannot handle Gtk.Builder/glade files\nimport locale\nfrom locale import gettext as _\nlocale.bindtextdomain(\"desktop-linux-manager\", \"/usr/locales/\")\nlocale.textdomain('desktop-linux-manager')\n\nclass QubesUpdater(Gtk.Application):\n # pylint: disable=too-many-instance-attributes\n\n def __init__(self, qapp):\n super().__init__(\n application_id=\"org.gnome.example\",\n flags=Gio.ApplicationFlags.FLAGS_NONE)\n\n self.qapp = qapp\n\n self.primary = False\n self.connect(\"activate\", self.do_activate)\n\n def perform_setup(self, *_args, **_kwargs):\n # pylint: disable=attribute-defined-outside-init\n self.builder = Gtk.Builder()\n self.builder.set_translation_domain(\"desktop-linux-manager\")\n self.builder.add_from_file(pkg_resources.resource_filename(\n __name__, 'updater.glade'))\n\n self.main_window = self.builder.get_object(\"main_window\")\n\n self.vm_list = self.builder.get_object(\"vm_list\")\n\n self.updates_available = self.populate_vm_list()\n\n self.no_updates_available_label = \\\n self.builder.get_object(\"no_updates_available\")\n self.no_updates_available_label.set_visible(not self.updates_available)\n\n self.allow_update_unavailable_check = \\\n self.builder.get_object(\"allow_update_unavailable\")\n self.allow_update_unavailable_check.connect(\"clicked\",\n self.set_update_available)\n\n self.next_button = self.builder.get_object(\"button_next\")\n self.next_button.connect(\"clicked\", self.next_clicked)\n\n self.cancel_button = self.builder.get_object(\"button_cancel\")\n self.cancel_button.connect(\"clicked\", self.cancel_updates)\n self.main_window.connect(\"delete-event\", self.window_close)\n self.main_window.connect(\"key-press-event\", self.check_escape)\n\n self.stack = self.builder.get_object(\"main_stack\")\n self.list_page = self.builder.get_object(\"list_page\")\n self.progress_page = self.builder.get_object(\"progress_page\")\n self.finish_page = self.builder.get_object(\"finish_page\")\n self.progress_textview = self.builder.get_object(\"progress_textview\")\n self.progress_scrolled_window = self.builder.get_object(\n \"progress_scrolled_window\")\n self.progress_listview = self.builder.get_object(\"progress_listview\")\n\n self.details_visible = True\n self.details_icon = self.builder.get_object(\"details_icon\")\n self.builder.get_object(\"details_icon_events\").connect(\n \"button-press-event\", self.toggle_details)\n self.builder.get_object(\"details_label\").connect(\n \"clicked\", self.toggle_details)\n\n self.load_css()\n\n self.main_window.show_all()\n self.toggle_details()\n\n self.update_thread = None\n self.exit_triggered = False\n\n def do_activate(self, *_args, **_kwargs):\n if not self.primary:\n self.perform_setup()\n self.primary = True\n self.hold()\n else:\n self.main_window.present()\n\n @staticmethod\n def load_css():\n style_provider = Gtk.CssProvider()\n css = b'''\n .black-border { \n border-width: 1px; \n border-color: #c6c6c6; \n border-style: solid;\n }\n '''\n style_provider.load_from_data(css)\n\n Gtk.StyleContext.add_provider_for_screen(\n Gdk.Screen.get_default(),\n style_provider,\n Gtk.STYLE_PROVIDER_PRIORITY_APPLICATION)\n\n def populate_vm_list(self):\n result = False # whether at least one VM has updates available\n for vm in self.qapp.domains:\n if vm.klass == 'AdminVM':\n try:\n state = vm.features.get('updates-available', False)\n except exc.QubesDaemonCommunicationError:\n state = False\n result = result or state\n self.vm_list.add(VMListBoxRow(vm, state))\n\n for vm in self.qapp.domains:\n if getattr(vm, 'updateable', False) and vm.klass != 'AdminVM':\n try:\n state = vm.features.get('updates-available', False)\n except exc.QubesDaemonCommunicationError:\n state = False\n result = result or state\n vmrow = VMListBoxRow(vm, state)\n self.vm_list.add(vmrow)\n vmrow.checkbox.connect('toggled', self.checkbox_checked)\n\n self.vm_list.connect(\"row-activated\", self.toggle_row_selection)\n return result\n\n def checkbox_checked(self, _emitter, *_args):\n for vm_row in self.vm_list:\n if vm_row.checkbox.get_active():\n self.next_button.set_sensitive(True)\n return\n self.next_button.set_sensitive(False)\n\n @staticmethod\n def toggle_row_selection(_emitter, row):\n if row:\n row.checkbox.set_active(not row.checkbox.get_active())\n row.set_label_text()\n\n def set_update_available(self, _emitter):\n for vm_row in self.vm_list:\n if not vm_row.updates_available:\n vm_row.set_sensitive(\n self.allow_update_unavailable_check.get_active())\n if not vm_row.get_sensitive():\n vm_row.checkbox.set_active(False)\n\n def next_clicked(self, _emitter):\n if self.stack.get_visible_child() == self.list_page:\n self.stack.set_visible_child(self.progress_page)\n\n for row in self.vm_list:\n if row.checkbox.get_active():\n self.progress_listview.add(ProgressListBoxRow(row.vm))\n\n self.progress_listview.show_all()\n\n self.next_button.set_sensitive(False)\n self.next_button.set_label(_(\"_Finish\"))\n\n # pylint: disable=attribute-defined-outside-init\n self.update_thread = threading.Thread(target=self.perform_update)\n self.update_thread.start()\n\n elif self.stack.get_visible_child() == self.progress_page:\n self.cancel_updates()\n return\n\n def toggle_details(self, *_args, **_kwargs):\n # pylint: disable=attribute-defined-outside-init\n self.details_visible = not self.details_visible\n self.progress_textview.set_visible(self.details_visible)\n\n if self.details_visible:\n self.progress_textview.show()\n self.progress_scrolled_window.show()\n else:\n self.progress_textview.hide()\n self.progress_scrolled_window.hide()\n\n if self.details_visible:\n self.details_icon.set_from_icon_name(\"pan-down-symbolic\",\n Gtk.IconSize.BUTTON)\n else:\n self.details_icon.set_from_icon_name(\"pan-end-symbolic\",\n Gtk.IconSize.BUTTON)\n\n def append_text_view(self, text):\n buffer = self.progress_textview.get_buffer()\n buffer.insert(buffer.get_end_iter(), text + '\\n')\n\n def perform_update(self):\n for row in self.progress_listview:\n if self.exit_triggered:\n GObject.idle_add(row.set_status, 'failure')\n GObject.idle_add(\n self.append_text_view,\n _(\"Canceled update for {}\\n\").format(row.vm.name))\n continue\n\n GObject.idle_add(\n self.append_text_view, _(\"Updating {}\\n\").format(row.vm.name))\n GObject.idle_add(row.set_status, 'in-progress')\n\n try:\n if row.vm.klass == 'AdminVM':\n output = subprocess.check_output(\n ['sudo', 'qubesctl', '--dom0-only', '--no-color',\n 'pkg.upgrade', 'refresh=True'],\n stderr=subprocess.STDOUT).decode()\n ansi_escape = re.compile(r'(\\x9B|\\x1B\\[)[0-?]*[ -/]*[@-~]')\n output = ansi_escape.sub('', output)\n else:\n output = subprocess.check_output(\n ['sudo', 'qubesctl', '--skip-dom0',\n '--targets=' + row.vm.name, '--show-output',\n 'state.sls', 'update.qubes-vm'],\n stderr=subprocess.STDOUT).decode()\n\n GObject.idle_add(self.append_text_view, output)\n GObject.idle_add(row.set_status, 'success')\n\n except subprocess.CalledProcessError as ex:\n GObject.idle_add(\n self.append_text_view,\n _(\"Error on updating {}: {}\\n{}\").format(\n row.vm.name, str(ex), ex.output.decode()))\n GObject.idle_add(row.set_status, 'failure')\n\n GObject.idle_add(self.next_button.set_sensitive, True)\n GObject.idle_add(self.cancel_button.set_visible, False)\n\n def cancel_updates(self, *_args, **_kwargs):\n # pylint: disable=attribute-defined-outside-init\n if self.update_thread and self.update_thread.is_alive():\n self.exit_triggered = True\n dialog = Gtk.MessageDialog(\n self.main_window, Gtk.DialogFlags.MODAL, Gtk.MessageType.OTHER,\n Gtk.ButtonsType.NONE, _(\n \"Waiting for current qube to finish updating.\"\n \" Updates for remaining qubes have been cancelled.\"))\n dialog.show()\n while self.update_thread.is_alive():\n while Gtk.events_pending():\n Gtk.main_iteration()\n time.sleep(1)\n dialog.hide()\n else:\n self.exit_updater()\n\n def check_escape(self, _widget, event, _data=None):\n if event.keyval == Gdk.KEY_Escape:\n self.cancel_updates()\n\n def window_close(self, *_args, **_kwargs):\n if self.stack.get_visible_child() == self.progress_page:\n self.cancel_updates()\n self.exit_updater()\n\n def exit_updater(self, _emitter=None):\n if self.primary:\n self.release()\n\n\ndef get_domain_icon(vm):\n icon_vm = Gtk.IconTheme.get_default().load_icon(vm.label.icon, 16, 0)\n icon_img = Gtk.Image.new_from_pixbuf(icon_vm)\n return icon_img\n\n\nclass VMListBoxRow(Gtk.ListBoxRow):\n def __init__(self, vm, updates_available, **properties):\n super().__init__(**properties)\n self.vm = vm\n\n hbox = Gtk.HBox(orientation=Gtk.Orientation.HORIZONTAL)\n\n self.label_text = vm.name\n self.updates_available = updates_available\n if self.updates_available:\n self.label_text = _(\"{vm} (updates available)\").format(\n vm=self.label_text)\n self.label = Gtk.Label()\n self.icon = get_domain_icon(self.vm)\n\n self.checkbox = Gtk.CheckButton()\n self.checkbox.set_active(self.updates_available)\n self.checkbox.set_margin_right(10)\n\n self.checkbox.connect(\"clicked\", self.set_label_text)\n self.set_sensitive(self.updates_available)\n\n self.set_label_text()\n\n hbox.pack_start(self.checkbox, False, False, 0)\n hbox.pack_start(self.icon, False, False, 0)\n hbox.pack_start(self.label, False, False, 0)\n\n # check for VMs that may be restored from older Qubes versions\n # and not support updating; this is a heuristic and may not always work\n try:\n if vm.features.get('qrexec', False) and \\\n vm.features.get('gui', False) and \\\n not vm.features.get('os', False):\n warn_icon = Gtk.Image.new_from_pixbuf(\n Gtk.IconTheme.get_default().load_icon(\n 'dialog-warning', 12, 0))\n warn_icon.set_tooltip_text(\n 'This qube may have been restored from an older version of '\n 'Qubes OS and may not be able to update itself correctly. '\n 'Please check the documentation if problems occur.')\n hbox.pack_start(warn_icon, False, False, 0)\n except exc.QubesDaemonCommunicationError:\n # we have no permission to access the vm's features, there's no\n # point in guessing original Qubes version\n pass\n\n self.add(hbox)\n\n def set_label_text(self, _=None):\n if self.checkbox.get_active():\n self.label.set_markup(\"{}\".format(self.label_text))\n else:\n self.label.set_markup(self.label_text)\n\n\nclass ProgressListBoxRow(Gtk.ListBoxRow):\n def __init__(self, vm):\n super().__init__()\n\n self.vm = vm\n\n hbox = Gtk.HBox(orientation=Gtk.Orientation.HORIZONTAL)\n\n self.icon = get_domain_icon(self.vm)\n self.icon.set_margin_right(10)\n\n self.label = Gtk.Label(vm.name)\n self.label.set_margin_right(10)\n\n self.progress_box = Gtk.HBox(orientation=Gtk.Orientation.HORIZONTAL)\n\n hbox.pack_start(self.icon, False, False, 0)\n hbox.pack_start(self.label, False, False, 0)\n hbox.pack_start(self.progress_box, False, False, 0)\n\n self.set_status('not-started')\n self.add(hbox)\n\n def set_status(self, status):\n\n if status == 'not-started':\n widget = Gtk.Spinner()\n elif status == 'in-progress':\n widget = Gtk.Spinner()\n widget.start()\n elif status == 'success':\n widget = Gtk.Image.new_from_icon_name(\"gtk-apply\",\n Gtk.IconSize.BUTTON)\n elif status == 'failure':\n widget = Gtk.Image.new_from_icon_name(\"gtk-cancel\",\n Gtk.IconSize.BUTTON)\n else:\n raise ValueError(_(\"Unknown status {}\").format(status))\n\n for child in self.progress_box.get_children():\n self.progress_box.remove(child)\n\n self.progress_box.pack_start(widget, False, False, 0)\n\n widget.show()\n\n\ndef main():\n qapp = Qubes()\n app = QubesUpdater(qapp)\n app.run()\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"qui/updater.py","file_name":"updater.py","file_ext":"py","file_size_in_byte":14654,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"242501617","text":"import tensorflow as tf\n\nhparams = tf.contrib.training.HParams(\n num_mels=80,\n frame_length_ms=50,\n frame_shift_ms=12.5,\n hop_length=int(16000 * 0.0125), # samples.\n win_length=int(16000 * 0.05), # samples.\n max_db=100,\n ref_db=20,\n preemphasis=0.97,\n max_abs_value=4.0,\n symmetric_mel=True,\n sr=16000,\n n_fft=2048,\n\n n_iter=60,\n power=1.5,\n max_generation_frames=1300,\n max_eval_batches=20,\n max_eval_sample_length=1300,\n eval_sample_per_speaker=4,\n\n vocab_size=40000,\n embed_size=512,\n encoder_hidden=512,\n decoder_hidden=640,\n n_encoder_layer=5,\n n_decoder_layer=4,\n n_attention_head=8,\n transformer_dropout_rate=0.1,\n decoder_dropout_rate=0.5,\n prenet_hidden=256,\n postnet_hidden=512,\n n_postnet_layer=5,\n\n use_knowledge_attention=True,\n knowledge_value_size=2048,\n knowledge_key_size=1024,\n knowledge_start_layer=1,\n knowledge_end_layer=5,\n knowledge_attention_head=4,\n use_key_encoder=False,\n key_encode_layers=1,\n use_identical_key_context=False,\n\n token_dropout_rate=0.0,\n\n data_format=\"nltpi\",\n input_method=\"char\",\n use_sos=True,\n remove_space=False,\n bucket_size=512,\n shuffle_training_data=True,\n batch_frame_limit=8000,\n batch_frame_quad_limit=8000000,\n max_batch_size=32,\n balanced_training=False,\n lg_prob_scale=0.2,\n adapt_start_step=30000,\n adapt_end_step=30000,\n final_adapt_rate=0.25,\n data_warmup_steps=30000,\n target_length_lower_bound=240,\n target_length_upper_bound=800,\n\n reg_weight=5e-9,\n\n multi_speaker=True,\n max_num_speaker=1000,\n speaker_embedding_size=128,\n\n multi_lingual=False,\n max_num_language=100,\n language_net_hidden=128,\n language_embedding_size=128,\n front_lang_embed=True,\n\n warmup_steps=50000,\n max_lr=1e-3,\n min_lr=1e-5,\n lr_decay_step=550000,\n lr_decay_rate=1e-2,\n adam_eps=5e-8,\n\n external_embed_dim=1024,\n use_external_embed=True,\n)\n","sub_path":"hyperparams.py","file_name":"hyperparams.py","file_ext":"py","file_size_in_byte":2002,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"580994407","text":"import numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\nimport matplotlib.cm as cm\n\n\ndef load_file(filename, names):\n return pd.read_csv(filename, header=None, names=names)\n\n\ndf = load_file('ex2data1.txt', ['first_exam', 'second_exam', 'accepted'])\nX_train, y_train = df.filter(['first_exam', 'second_exam']), df['accepted']\n\n\nz_true = df[df['accepted'] == 1]\nz_false = df[df['accepted'] == 0]\nfig, ax = plt.subplots()\nax.scatter(z_true['first_exam'], z_true['second_exam'], marker='o', c='g', label='Accepted', s=20)\nax.scatter(z_false['first_exam'], z_false['second_exam'], marker='x', c='r', label='Not accepted', s=20)\nax.legend(loc='upper right');\nax.set_xlabel('First exam')\nax.set_ylabel('Second exam')\nplt.show()\n\n\nfrom utils import sigmoid\n\nclass LogisticRegression:\n THRESHOLD = 1e-6\n\n def __init__(self, fit_method='gradient_descent', max_steps=100000,\n learning_rate=0.01, regularized=False, reg_L=0.5, log=False):\n self.weights = []\n self.max_steps = max_steps\n self.learning_rate = learning_rate\n self.regularized = regularized\n self.reg_L = reg_L\n self.cost_func = self.cost_func_regularized if regularized else self.cost_func_non_regularized\n self.cost_der = self.cost_der_regularized if regularized else self.cost_der_non_regularized\n self.fit_method = getattr(self, fit_method)\n self.log = log\n \n def fit(self, X, y):\n if hasattr(X, 'values'):\n X = X.values\n if hasattr(y, 'values'):\n y = y.values\n\n X = X.astype('float64') \n y = y.astype('float64')\n \n if not self.regularized:\n X = np.column_stack((np.ones(X.shape[0]), X))\n \n self.fit_method(X, y)\n \n def predict(self, X):\n if self.weights is None:\n raise Exception(\"Model is not trained. Call `fit` method.\")\n\n X = np.array(X)\n if not self.regularized:\n X = np.insert(X, 0, 1)\n h = self.calculate_hypotesis(X)\n return 1 if h >= 0.5 else 0\n \n def gradient_descent(self, X, y):\n self.cost_history = []\n self.weights = np.zeros(X.shape[1])\n cur_loss = self.cost_func(X, y)\n\n cur_step = 0\n while cur_step < self.max_steps:\n cur_step += 1\n self.gradient_descent_step(X, y)\n new_loss = self.cost_func(X, y)\n self.cost_history.append(new_loss)\n if abs(new_loss - cur_loss) < self.THRESHOLD:\n break\n\n cur_loss = new_loss\n \n def gradient_descent_step(self, X, y):\n gradient = self.cost_der(X, y, self.weights)\n gradient *= self.learning_rate\n self.weights -= gradient\n \n def cost_func_non_regularized(self, X, y, weights=None):\n if weights is None:\n weights = self.weights\n \n predictions = self.calculate_hypotesis(X, weights)\n cost_trues = y * np.log(predictions)\n cost_falses = (1 - y) * np.log(1 - predictions)\n total_cost = -np.mean(cost_trues + cost_falses)\n return total_cost\n \n def cost_func_regularized(self, X, y, weights=None):\n if weights is None:\n weights = self.weights\n \n cost = self.cost_func_non_regularized(X, y, weights)\n weights_R = weights[1:]\n total_cost = cost + (self.reg_L / 2 / X.shape[0]) * np.dot(weights_R.T, weights_R)\n return total_cost\n \n def calculate_hypotesis(self, X, weights=None):\n if weights is None:\n weights = self.weights\n\n return sigmoid(X.dot(weights))\n \n def cost_der_non_regularized(self, X, y, theta):\n predictions = self.calculate_hypotesis(X, weights=theta)\n sq_error = predictions - y\n gradient = np.dot(X.T, sq_error)\n gradient /= X.shape[0]\n return gradient\n\n def cost_der_regularized(self, X, y, theta):\n predictions = self.calculate_hypotesis(X, weights=theta)\n sq_error = predictions - y\n gradient_first = np.dot(X.T[:1], sq_error)\n gradient_full = np.dot(X.T[1:], sq_error) + self.reg_L * theta[1:]\n gradient = np.insert(gradient_full, 0, gradient_first)\n gradient /= X.shape[0]\n return gradient\n \n def nelder_mead_algo(self, X, y):\n from scipy.optimize import fmin\n\n N = X.shape[0]\n\n def func(theta):\n return self.cost_func(X, y, theta)\n \n init_theta = np.zeros(X.shape[1])\n self.weights = fmin(func, init_theta, xtol=self.THRESHOLD, maxfun=100000)\n \n def bfgs_algo(self, X, y):\n from scipy.optimize import fmin_bfgs\n\n N = X.shape[0]\n\n def func(theta):\n return self.cost_func(X, y, theta)\n \n def func_der(theta):\n return self.cost_der(X, y, theta)\n\n init_theta = np.zeros(X.shape[1])\n self.weights = fmin_bfgs(func, init_theta, fprime=func_der, gtol=self.THRESHOLD, disp=self.log)\n \n\ncls_grad = LogisticRegression(fit_method='gradient_descent', max_steps=300000, learning_rate=0.004)\ncls_grad.fit(X_train, y_train)\nprint(f'Minimum cost function value: {cls_grad.cost_history[-1]}')\nprint(f'Iterations: {len(cls_grad.cost_history)}')\nprint(f'Weights: {cls_grad.weights}')\n\ncls_nm = LogisticRegression(fit_method='nelder_mead_algo')\ncls_nm.fit(X_train, y_train)\nprint(f'Weights: {cls_nm.weights}')\n\ncls_bfgs = LogisticRegression(fit_method='bfgs_algo', log=True)\ncls_bfgs.fit(X_train, y_train)\nprint(f'Weights: {cls_bfgs.weights}')\n\n\nz_true = df[df['accepted'] == 1]\nz_false = df[df['accepted'] == 0]\n\ndef decision_boundary(x, weights):\n return -(weights[0] + weights[1] * x) / weights[2]\n\nfig, ax = plt.subplots()\nax.scatter(z_true['first_exam'], z_true['second_exam'], marker='o', c='g', label='Accepted', s=20)\nax.scatter(z_false['first_exam'], z_false['second_exam'], marker='x', c='r', label='Not accepted', s=20)\nax.plot(z_false['first_exam'],\n [decision_boundary(i, cls_grad.weights) for i in z_false['first_exam']],\n c='b', label='Decision boundary')\nax.legend(loc='upper right');\nax.set_xlabel('First exam')\nax.set_ylabel('Second exam')\nplt.show()\n\n\ndf = load_file('ex2data2.txt', names=['first_test', 'second_test', 'passed'])\nX_train, y_train = df.filter(['first_test', 'second_test']), df['passed']\n\n\nz_true = df[df['passed'] == 1]\nz_false = df[df['passed'] == 0]\nfig, ax_reg = plt.subplots()\nax_reg.scatter(z_true['first_test'], z_true['second_test'], marker='o', c='g', label='Passed', s=20)\nax_reg.scatter(z_false['first_test'], z_false['second_test'], marker='x', c='r', label='Not passed', s=20)\nax_reg.legend(loc='upper right');\nax_reg.set_xlabel('First test')\nax_reg.set_ylabel('Second test')\nplt.show()\n\n\ndef build_poly_features(x1, x2, log=False):\n degree = 6\n res = []\n str_res = []\n\n for i in range(degree + 1):\n for j in range(i, degree + 1):\n res.append(x1**(j - i) * x2**i)\n first = '' if j - i == 0 else 'x1' if j - i == 1 else f'x1^{j - i}'\n second = '' if i == 0 else 'x2' if i == 1 else f'x2^{i}'\n if not first and not second:\n str_append = '1'\n elif first and not second:\n str_append = first\n elif second and not first:\n str_append = second\n else:\n str_append = f\"{first}*{second}\"\n str_res.append(str_append)\n\n str_res = ' + '.join(str_res)\n if log:\n print(str_res)\n assert len(res) == 28\n return np.array(res).T\n\n\nX_poly = build_poly_features(X_train['first_test'], X_train['second_test'], log=True)\n\ncls_grad_reg = LogisticRegression(fit_method='gradient_descent', regularized=True,\n max_steps=300000, learning_rate=0.5, reg_L=0.5)\ncls_grad_reg.fit(X_poly, y_train)\nprint(f'Minimum cost function value: {cls_grad_reg.cost_history[-1]}')\nprint(f'Iterations: {len(cls_grad_reg.cost_history)}')\nprint(f'Weights: {cls_grad_reg.weights}')\n\ncls_nm_reg = LogisticRegression(fit_method='nelder_mead_algo', regularized=True)\ncls_nm_reg.fit(X_poly, y_train)\nprint(f'Weights: {cls_nm_reg.weights}')\n\ncls_bfgs_reg = LogisticRegression(fit_method='bfgs_algo', regularized=True, log=True)\ncls_bfgs_reg.fit(X_poly, y_train)\nprint(f'Weights: {cls_bfgs_reg.weights}')\n\n\nprint(f\"Predicted class: {cls_grad_reg.predict(X_poly[0])}, actual class: {y_train[0]}\")\nprint(f\"Predicted class: {cls_nm_reg.predict(X_poly[0])}, actual class: {y_train[0]}\")\nprint(f\"Predicted class: {cls_bfgs_reg.predict(X_poly[0])}, actual class: {y_train[0]}\")\n\n\ndef decision_boundary_contour(theta1, theta2, theta3):\n u = np.linspace(-1, 1.2, 50)\n v = np.linspace(-1, 1.3, 50)\n z1 = np.zeros(shape=(len(u), len(v)))\n z2 = np.zeros(shape=(len(u), len(v)))\n z3 = np.zeros(shape=(len(u), len(v)))\n for i in range(len(u)):\n for j in range(len(v)):\n z1[i, j] = build_poly_features(np.array(u[i]), np.array(v[j])).dot(theta1)\n z2[i, j] = build_poly_features(np.array(u[i]), np.array(v[j])).dot(theta2)\n z3[i, j] = build_poly_features(np.array(u[i]), np.array(v[j])).dot(theta3)\n\n z1 = z1.T\n z2 = z2.T\n z3 = z3.T\n fig, ax_reg = plt.subplots()\n ax_reg.contour(u, v, z1, levels=0, colors='b')\n ax_reg.contour(u, v, z2, levels=0, colors='g')\n ax_reg.contour(u, v, z3, levels=0, colors='y')\n z_true = df[df['passed'] == 1]\n z_false = df[df['passed'] == 0]\n ax_reg.scatter(z_true['first_test'], z_true['second_test'], marker='o', c='g', label='Passed', s=20)\n ax_reg.scatter(z_false['first_test'], z_false['second_test'], marker='x', c='r', label='Not passed', s=20)\n ax_reg.legend(loc='upper right');\n ax_reg.set_xlabel('First test')\n ax_reg.set_ylabel('Second test')\n ax_reg.set_title('Decision boundary, lambda = %f' % cls_grad_reg.reg_L)\n plt.show()\n \ndecision_boundary_contour(cls_grad_reg.weights, cls_nm_reg.weights, cls_bfgs_reg.weights)\n\n\ncls1 = LogisticRegression(fit_method='gradient_descent', max_steps=300000, learning_rate=0.5,\n regularized=True, reg_L=0.5)\ncls1.fit(X_poly, y_train)\n\ncls2 = LogisticRegression(fit_method='gradient_descent', max_steps=300000, learning_rate=0.5,\n regularized=True, reg_L=0.05)\ncls2.fit(X_poly, y_train)\n\ncls3 = LogisticRegression(fit_method='gradient_descent', max_steps=300000, learning_rate=0.5,\n regularized=True, reg_L=0.005)\ncls3.fit(X_poly, y_train)\n\ndecision_boundary_contour(cls1.weights, cls2.weights, cls3.weights)\n\n\nfrom scipy.io import loadmat\n\nmat = loadmat('ex2data3.mat')\nX_train, y_train = mat['X'], mat['y']\ny_train = y_train.reshape(y_train.shape[0])\ny_train = np.where(y_train != 10, y_train, 0)\n\n\ndef vector_to_matrix(x):\n len_vec = len(x)\n step = int(np.sqrt(len_vec))\n assert step ** 2 == len_vec, 'Matrix should be squared' \n matrix = [x[left:left+step] for left in range(0, len_vec, step)]\n np_matrix = np.array(matrix).T\n reversed_matrix = np.flip(np_matrix, axis=0)\n return reversed_matrix\n\nnums = list(range(150, 5000, 500))\npictures = [vector_to_matrix(X_train[i]) for i in nums]\n\nfig, axs = plt.subplots(2, 5, figsize=(20, 8))\nfor i, ax in enumerate(axs.flatten()):\n ax.pcolor(pictures[i], cmap=cm.gray)\n res = y_train[nums[i]]\n if res == 10:\n res = 0\n ax.set_title(f'Number {res}')\n\nplt.show()\n\n\nclass MulticlassLogisticRegression:\n classifier = LogisticRegression\n\n def __init__(self, num_classes=10):\n self.num_classes = num_classes\n self.classifiers = [\n self.classifier(fit_method='gradient_descent', learning_rate=0.5, regularized=True, reg_L=0.1)\n for i in range(num_classes)\n ]\n \n def fit(self, X, y):\n for i in range(self.num_classes):\n y_train = (y == i).astype(int)\n self.classifiers[i].fit(X, y_train)\n \n def predict(self, X):\n h = []\n for cls in self.classifiers:\n h.append(cls.calculate_hypotesis(X))\n \n return np.argmax(np.array(h), axis=0)\n \n \ncls_mult = MulticlassLogisticRegression()\ncls_mult.fit(X_train, y_train)\npred_value = cls_mult.predict(X_train[-1])\nprint(f\"Predicted class: {pred_value}, actual class: {y_train[-1]}\")\n\n\ndef accuracy(cls, X, y):\n error = cls.predict(X) - y\n return 1.0 - (float(np.count_nonzero(error)) / len(error))\n\nacc = accuracy(cls_mult, X_train, y_train)\nprint(f\"Accuracy: {acc}\")\n","sub_path":"lab2/logistic_regression.py","file_name":"logistic_regression.py","file_ext":"py","file_size_in_byte":12513,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"454372841","text":"import sys\r\ndef check(egg,count):\r\n if egg==0:return 0\r\n else:\r\n total=0\r\n for i in range(count):total+=check(egg-1,i)+1\r\n return total\r\ndef crack_egg(egg,floor,count=0):\r\n while check(egg,count) 0:\n t = 0\n unique = set()\n for c in contacts_j:\n t += c.duration\n unique.add(c.indiv_i)\n\n counts.append(len(contacts_j))\n counts_unique.append(len(unique))\n total_contact_time.append(t)\n ave_contact_time.append(t / len(contacts_j))\n ave_contact_time_unique.append(t / len(unique))\n\n else:\n empty += 1\n\n print('empty = ', empty)\n\n return dict(\n counts=counts,\n counts_unique=counts_unique,\n total_contact_time=total_contact_time,\n ave_contact_time=ave_contact_time,\n ave_contact_time_unique=ave_contact_time_unique,\n )\n\n\ndef comp_stats(arr0, arr1):\n return {\n 'mean': np.mean(arr0) / np.mean(arr1),\n 'median': np.median(arr0) / np.median(arr1),\n 'max': np.max(arr0) / np.max(arr1),\n }\n\n\ndef compute_mob_statistics(loc_tup, days, max_people, verbose=False):\n '''Computes all MobilitySimulator statistics for given `country` and `area` '''\n\n country, area = loc_tup\n\n if verbose:\n print(country, area)\n\n # get mobility simulator settings\n statistics = dict()\n mob_settings_downsampled, mob_settings_full = calibration_mob_paths[country][area]\n\n # downsampled\n with open(mob_settings_downsampled, 'rb') as fp:\n obj = pickle.load(fp)\n mob_downsampled = MobilitySimulator(**obj)\n mob_downsampled.verbose = verbose\n mob_downsampled.simulate(max_time=days * TO_HOURS, lazy_contacts=True)\n\n # full\n with open(mob_settings_full, 'rb') as fp:\n obj = pickle.load(fp)\n mob_full = MobilitySimulator(**obj)\n mob_full.verbose = verbose\n mob_full.simulate(max_time=days * TO_HOURS, lazy_contacts=True)\n\n # compute contact information\n contact_info_downsampled = get_stats(\n mob_downsampled, max_people, verbose=verbose)\n del mob_downsampled\n contact_info_full = get_stats(mob_full, max_people, verbose=verbose)\n del mob_full\n\n # summarize\n for s in contact_info_downsampled.keys():\n\n fig = plt.figure(figsize=(4, 7))\n ax0 = fig.add_subplot(211)\n ax0.hist(contact_info_downsampled[s])\n ax0.set_title('downsampled')\n xlim0 = ax0.get_xlim()\n ax1 = fig.add_subplot(212)\n ax1.hist(contact_info_full[s])\n ax1.set_title('full')\n xlim1 = ax1.get_xlim()\n\n ax0.set_xlim((min(xlim0[0], xlim1[0]), max(xlim0[1], xlim1[1])))\n ax1.set_xlim((min(xlim0[0], xlim1[0]), max(xlim0[1], xlim1[1])))\n fig.suptitle(s)\n plt.savefig('plots/betaScaling-' + loc_tup[0] + '-' + loc_tup[1] + '-' + s + '.png', format='png', facecolor=None,\n dpi=200, bbox_inches='tight')\n plt.close('all')\n\n d = comp_stats(\n contact_info_downsampled[s],\n contact_info_full[s])\n for k, v in d.items():\n statistics['ratio-' + k + '-' + s] = v\n\n # print always\n print(country, area)\n pprint(statistics)\n\n return statistics\n\n\nif __name__ == '__main__':\n\n days = 7.0\n max_people = 5000\n parallel = False\n cpu_count = 2\n\n locs = [\n ('GER', 'TU'), ('GER', 'KL'), ('GER', 'RH'), ('GER', 'TR'),\n ('CH', 'VD'), ('CH', 'BE'), ('CH', 'TI'), ('CH', 'JU'),\n ]\n\n # run in parallel for all locs\n if parallel:\n with ProcessPoolExecutor(cpu_count) as ex:\n res = ex.map(\n compute_mob_statistics,\n locs,\n [days for _ in locs],\n [max_people for _ in locs]\n )\n else:\n res = [compute_mob_statistics(\n tup, days, max_people, verbose=True) for tup in locs]\n\n # print all statistics\n all_statistics_unordered = dict(zip(locs, res))\n\n pprint(all_statistics_unordered)\n\n all_statistics = dict()\n\n for s in res[0].keys():\n all_statistics[s] = dict()\n for loc_tup in locs:\n all_statistics[s][loc_tup] = all_statistics_unordered[loc_tup][s]\n\n print('\\nStatistics by type:')\n pprint(all_statistics)\n","sub_path":"sim/betaScaling.py","file_name":"betaScaling.py","file_ext":"py","file_size_in_byte":5436,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"256013927","text":"import pyaudio\r\nimport wave\r\nimport sys\r\nCHUNK=1024\r\ndef talk(op=''):\r\n p = pyaudio.PyAudio()\r\n\r\n stream = p.open(format=p.get_format_from_width(op.getsampwidth()),\r\n channels=op.getnchannels(),\r\n rate=op.getframerate(),\r\n output=True)\r\n\r\n data = op.readframes(CHUNK)\r\n\r\n while len(data) > 0:\r\n stream.write(data)\r\n data = op.readframes(CHUNK)\r\n\r\n stream.stop_stream()\r\n stream.close()\r\n\r\n p.terminate()\r\ndef numeros():\r\n import math\r\n unidad=[\" \",\"n1\",\"n2\",\"n3\",\"n4\",\"n5\",\"n6\",\"n7\",\"n8\",\"n9\",\"n10\"]\r\n esp=[\" \",\"n11\",\"n12\",\"n13\",\"n14\",\"n15\",\"n16\",\"n17\",\"n18\",\"n19\"]\r\n decenas = [\"\",\"n10\",\"n20\", \"n30\",\"n40\",\"n50\", \"n60\",\"n70\", \"n80\", \"n90\"]\r\n centenas = [\"n100\",\"n100.2\",\"n200\",\"n300\",\"n400\",\"n500\",\"n600\",\"n700\",\"n800\",\"n900\"]\r\n \r\n local=\"./nu/\"\r\n print(\"**Suma de numeros Menor de 1000**\")\r\n su1=int(input(\"Digite el primer numero:\"))\r\n su2=int(input(\"Digite el segundo numero:\"))\r\n num=su1+su2\r\n \r\n if (num < 9):\r\n op=wave.open(local+unidad[num]+\".wav\")\r\n talk(op)\r\n elif (num==10):\r\n op=wave.open(local+\"n10.wav\")\r\n talk(op)\r\n elif (num>11 and num<20):\r\n num=num-10\r\n op=wave.open(local+esp[num]+\".wav\")\r\n talk(op)\r\n elif (num >= 20 and num < 100):\r\n u= (num%10)\r\n d=int(num/10)\r\n op=wave.open(local+decenas[d]+\".wav\")\r\n if (u==0):\r\n talk(op)\r\n else:\r\n op1=wave.open(local+\"y.wav\")\r\n op2=wave.open(local+unidad[u]+\".wav\")\r\n [talk(op),talk(op1),talk(op2)]\r\n elif (num >=100 and num <=1000):\r\n if(num==1000):\r\n op=wave.open(local+\"n1000.wav\")\r\n talk(op)\r\n elif(num<1000):\r\n c=int(num/100)\r\n d=int((num-(c*100))/10)\r\n u=int(num-(c*100+d*10))\r\n if(u==0):\r\n if(d==0):\r\n op=wave.open(local+centenas[0]+\".wav\")\r\n talk(op)\r\n else:\r\n op1=wave.open(local+centenas[c]+\".wav\")\r\n op2=wave.open(local+decenas[d]+\".wav\")\r\n [talk(op1),talk(op2)]\r\n else:\r\n if(d == 1) and (u == 1,2,3,4,5,6,7,8,9):\r\n\r\n op=wave.open(local+centenas[c]+\".wav\")\r\n op1=wave.open(local+esp[u]+\".wav\")\r\n [talk(op),talk(op1)]\r\n else:\r\n op=wave.open(local+centenas[c]+\".wav\")\r\n op1=wave.open(local+decenas[d]+\".wav\")\r\n op2=wave.open(local+\"y.wav\")\r\n op3=wave.open(local+unidad[u]+\".wav\")\r\n [talk(op),talk(op1),talk(op2),talk(op3)]\r\n\r\nnumeros()\r\n\r\n\r\n\r\n \r\n","sub_path":"facial recognition/talk.py","file_name":"talk.py","file_ext":"py","file_size_in_byte":2919,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"141520706","text":"from pychrom.modeling.cadet_modeler import CadetModeler\nfrom pychrom.core import *\nfrom pychrom.opt.casadi.build_utils import *\nimport matplotlib.animation as animation\nimport matplotlib.pyplot as plt\n\ncomps = ['A',\n 'B',\n 'C',\n 'D']\n\nGRM = GRModel(components=comps)\n\n# create sections\nGRM.load = Section(components=comps)\nfor cname in comps:\n GRM.load.set_a0(cname, 1.0)\n\nGRM.load.set_a0('A', 50.0)\nGRM.load.set_a1('A', 0.0)\nGRM.load.start_time_sec = 0.0\n\nGRM.wash = Section(components=comps)\nGRM.wash.set_a0('A', 50.0)\nGRM.wash.start_time_sec = 10.0\n\nGRM.elute = Section(components=comps)\nGRM.elute.set_a0('A', 100.0)\nGRM.elute.set_a1('A', 0.2)\nGRM.elute.start_time_sec = 90.0\n\n# create inlet\nGRM.inlet = Inlet(components=comps)\nGRM.inlet.add_section('load')\nGRM.inlet.add_section('wash')\nGRM.inlet.add_section('elute')\n\n# create binding\nGRM.salt = 'A'\nGRM.binding = SMABinding(data=\"sma.yml\")\nGRM.binding.is_kinetic = True\n\n# create column\nGRM.column = Column(data=\"column.yml\")\n\n# create outlet\nGRM.outlet = Outlet(components=comps)\n\n# connect units\nGRM.connect_unit_operations('inlet', 'column')\nGRM.connect_unit_operations('column', 'outlet')\n\nfor name in GRM.list_components():\n nu = 1.0\n GRM.column.binding_model.set_nu(name, nu)\n\ncwrapper = CasadiColumn(GRM.column)\nlspan = np.linspace(0, GRM.column.length, 50)\ncwrapper.build_model(lspan, nominal_c={'A': 50}, nominal_q={'A': 1200})\n\n# defines grid of times\ntspan = np.linspace(0, 1500, 1500)\nresults = cwrapper.solve(tspan)\n\nfor cname in results.components:\n if cname !='A':\n to_plot = results.C.sel(component=cname)\n plot2d = to_plot.sel(col_loc=GRM.column.length)\n plt.plot(plot2d.time, plot2d)\nplt.show()\n\n","sub_path":"pychrom/opt/casadi/simulating_sma.py","file_name":"simulating_sma.py","file_ext":"py","file_size_in_byte":1727,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"303688508","text":"from PyQt4.QtGui import *\nfrom PyQt4.QtCore import *\nfrom PyQt4.QtSql import *\n\nfrom view_product_details import *\n\nclass BrowseProductSearchResults(QWidget):\n cancel_button_signal = pyqtSignal()\n\n def __init__(self, search_type, search_name):\n super().__init__()\n \n self.table_view = QTableView()\n self.table_view.setSelectionBehavior(1)\n self.table_view.setEditTriggers(self.table_view.EditTrigger(0))\n self.cancel_button = QPushButton(\"Close\")\n self.information_label = QLabel('To view information about a product, click on the product and click \"View iformation\"')\n self.view_button = QPushButton(\"View information\")\n\n self.information_label.setAlignment(Qt.Alignment(4))\n\n self.layout = QVBoxLayout()\n self.layout.addWidget(self.information_label)\n self.layout.addWidget(self.table_view)\n self.layout.addWidget(self.cancel_button)\n self.layout.addWidget(self.view_button)\n\n self.setLayout(self.layout)\n\n self.create_table_model(search_type, search_name)\n\n #Connection\n self.cancel_button.clicked.connect(self.cancel_button_clicked)\n self.view_button.clicked.connect(self.view_button_clicked)\n\n def create_table_model(self, search_type, search_name):\n self.model = QSqlTableModel()\n query = QSqlQuery()\n if search_type == \"product_type_ID\":\n query.prepare(\"\"\"SELECT * FROM Product WHERE ProductTypeID = ?\"\"\")\n query.addBindValue(search_name)\n query.exec_()\n elif search_type == \"kit_name\" or \"blank_name\":\n query.prepare(\"\"\"SELECT PartID FROM Part WHERE PartName = ?\"\"\")\n query.addBindValue(search_name)\n query.exec_()\n while query.next():\n id_number = query.value(0)\n query.prepare(\"\"\"SELECT * FROM ProductParts WHERE PartID = ?\"\"\")\n query.addBindValue(id_number)\n query.exec_()\n self.model.setQuery(query)\n self.model.setEditStrategy(QSqlTableModel.OnManualSubmit)\n self.table_view.setModel(self.model)\n self.table_view.model().select()\n\n def product(self):\n self.index = self.table_view.selectedIndexes()\n self.product_id = self.table_view.model().data(self.index[0])\n query = QSqlQuery()\n query.prepare(\"\"\"SELECT Price, ProductTypeID, Quantity, ProductStatus FROM Product WHERE ProductID = ?\"\"\")\n query.addBindValue(self.product_id)\n query.exec_()\n while query.next():\n self.product_price = query.value(0)\n self.product_type_id = query.value(1)\n self.product_quantity = query.value(2)\n self.product_status = query.value(3)\n self.part_id = []\n query.prepare(\"\"\"SELECT PartID FROM ProductParts WHERE ProductID = ?\"\"\")\n query.addBindValue(self.product_id)\n query.exec_()\n while query.next():\n self.part_id.append(query.value(0))\n self.kit_id = self.part_id[0]\n self.blank_id = self.part_id[1]\n query.prepare(\"\"\"SELECT ProductType FROM ProductType WHERE ProductTypeID = ?\"\"\")\n query.addBindValue(self.product_type_id)\n query.exec_()\n while query.next():\n self.product_type = query.value(0)\n query.prepare(\"\"\"SELECT PartName FROM Part WHERE PartID = ?\"\"\")\n query.addBindValue(self.kit_id)\n query.exec_()\n while query.next():\n self.kit_name = query.value(0)\n query.prepare(\"\"\"SELECT PartName FROM Part WHERE PartID = ?\"\"\")\n query.addBindValue(self.blank_id)\n query.exec_()\n while query.next():\n self.blank_name = query.value(0)\n details = {\"product id\":self.product_id,\n \"product type\":self.product_type,\n \"price\":self.product_price,\n \"kit_name\":self.kit_name,\n \"blank_name\":self.blank_name,\n \"quantity\":self.product_quantity,\n \"status\":self.product_status}\n return details\n\n def view_button_clicked(self):\n details = self.product()\n dialog = ViewProductDetails(details)\n dialog.exec_()\n\n def cancel_button_clicked(self):\n self.cancel_button_signal.emit()\n","sub_path":"browse_product_search_results.py","file_name":"browse_product_search_results.py","file_ext":"py","file_size_in_byte":4306,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"65339408","text":"def process_scores():\n scores_txt = open(\"scores.txt\", \"r\")\n scores = {}\n for line in scores_txt:\n line = line.rstrip()\n restaurant, score = line.split(\":\")\n scores[restaurant] = int(score)\n\n return scores\n\n\ndef add_restaurant(scores):\n \n\n print(\"Please add a rating for your favorite restaurant!\")\n restaurant = input(\"Restaurant name> \")\n rating = int(input(\"Rating> \"))\n\n scores[restaurant] = rating\n\n\ndef print_sorted_scores(scores):\n \n\n for restaurant, rating in sorted(scores.items()):\n print(\"{restaurant} is rated at {rating}.\")\nscores = process_scores()\nadd_restaurant(scores)\nprint_sorted_scores(scores)\n","sub_path":"ratings.py","file_name":"ratings.py","file_ext":"py","file_size_in_byte":666,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"530766637","text":"\"\"\"Tilesets command line interface\"\"\"\nimport os\nimport json\nimport requests\nimport tempfile\n\nimport click\nimport cligj\nfrom requests_toolbelt import MultipartEncoder\n\nimport mapbox_tilesets\nfrom mapbox_tilesets import utils, errors\n\n\ndef _get_token(token=None):\n \"\"\"Get Mapbox access token from arg or environment\"\"\"\n if token is not None:\n return token\n else:\n return os.environ.get(\"MAPBOX_ACCESS_TOKEN\") or os.environ.get(\n \"MapboxAccessToken\"\n )\n\n\ndef _get_api():\n \"\"\"Get Mapbox tileset API base URL from environment\"\"\"\n return os.environ.get(\"MAPBOX_API\", \"https://api.mapbox.com\")\n\n\n@click.version_option(version=mapbox_tilesets.__version__, message=\"%(version)s\")\n@click.group()\ndef cli():\n \"\"\"This is the command line interface for the Mapbox Tilesets API.\n Thanks for joining us.\n\n This CLI requires a Mapbox access token. You can either set it in your environment as\n \"MAPBOX_ACCESS_TOKEN\" or \"MapboxAccessToken\" or pass it to each command with the --token flag.\n \"\"\"\n\n\n@cli.command(\"create\")\n@click.argument(\"tileset\", required=True, type=str)\n@click.option(\n \"--recipe\",\n \"-r\",\n required=True,\n type=click.Path(exists=True),\n help=\"path to a Recipe JSON document\",\n)\n@click.option(\"--name\", \"-n\", required=True, type=str, help=\"name of the tileset\")\n@click.option(\n \"--description\", \"-d\", required=False, type=str, help=\"description of the tileset\"\n)\n@click.option(\n \"--privacy\",\n \"-p\",\n required=False,\n type=click.Choice([\"public\", \"private\"]),\n help=\"set the tileset privacy options\",\n)\n@click.option(\"--token\", \"-t\", required=False, type=str, help=\"Mapbox access token\")\n@click.option(\"--indent\", type=int, default=None, help=\"Indent for JSON output\")\ndef create(\n tileset, recipe, name=None, description=None, privacy=None, token=None, indent=None\n):\n \"\"\"Create a new tileset with a recipe.\n\n $ tilesets create \n\n is in the form of username.handle - for example \"mapbox.neat-tileset\".\n The handle may only include \"-\" or \"_\" special characters.\n \"\"\"\n mapbox_api = _get_api()\n mapbox_token = _get_token(token)\n url = \"{0}/tilesets/v1/{1}?access_token={2}\".format(\n mapbox_api, tileset, mapbox_token\n )\n body = {}\n body[\"name\"] = name or \"\"\n body[\"description\"] = description or \"\"\n if privacy:\n body[\"private\"] = True if privacy == \"private\" else False\n\n if not utils.validate_tileset_id(tileset):\n raise errors.TilesetNameError\n\n if recipe:\n with open(recipe) as json_recipe:\n body[\"recipe\"] = json.load(json_recipe)\n\n r = requests.post(url, json=body)\n\n click.echo(json.dumps(r.json(), indent=indent))\n\n\n@cli.command(\"publish\")\n@click.argument(\"tileset\", required=True, type=str)\n@click.option(\"--token\", \"-t\", required=False, type=str, help=\"Mapbox access token\")\n@click.option(\"--indent\", type=int, default=None, help=\"Indent for JSON output\")\ndef publish(tileset, token=None, indent=None):\n \"\"\"Publish your tileset.\n\n tilesets publish \n \"\"\"\n mapbox_api = _get_api()\n mapbox_token = _get_token(token)\n url = \"{0}/tilesets/v1/{1}/publish?access_token={2}\".format(\n mapbox_api, tileset, mapbox_token\n )\n r = requests.post(url)\n if r.status_code == 200:\n click.echo(json.dumps(r.json(), indent=indent))\n click.echo(\n f\"You can view the status of your tileset with the `tilesets status {tileset}` command.\",\n err=True,\n )\n else:\n raise errors.TilesetsError(f\"{r.text}\")\n\n\n@cli.command(\"status\")\n@click.argument(\"tileset\", required=True, type=str)\n@click.option(\"--token\", \"-t\", required=False, type=str, help=\"Mapbox access token\")\n@click.option(\"--indent\", type=int, default=None, help=\"Indent for JSON output\")\ndef status(tileset, token=None, indent=None):\n \"\"\"View the current queue/processing/complete status of your tileset.\n\n tilesets status \n \"\"\"\n mapbox_api = _get_api()\n mapbox_token = _get_token(token)\n url = \"{0}/tilesets/v1/{1}/status?access_token={2}\".format(\n mapbox_api, tileset, mapbox_token\n )\n r = requests.get(url)\n\n click.echo(json.dumps(r.json(), indent=indent))\n\n\n@cli.command(\"jobs\")\n@click.argument(\"tileset\", required=True, type=str)\n@click.option(\"--stage\", \"-s\", required=False, type=str, help=\"job stage\")\n@click.option(\"--token\", \"-t\", required=False, type=str, help=\"Mapbox access token\")\n@click.option(\"--indent\", type=int, default=None, help=\"Indent for JSON output\")\ndef jobs(tileset, stage, token=None, indent=None):\n \"\"\"View all jobs for a particular tileset.\n\n tilesets jobs \n \"\"\"\n mapbox_api = _get_api()\n mapbox_token = _get_token(token)\n url = \"{0}/tilesets/v1/{1}/jobs?access_token={2}\".format(\n mapbox_api, tileset, mapbox_token\n )\n if stage:\n url = \"{0}/tilesets/v1/{1}/jobs?stage={2}&access_token={3}\".format(\n mapbox_api, tileset, stage, mapbox_token\n )\n\n r = requests.get(url)\n\n click.echo(json.dumps(r.json(), indent=indent))\n\n\n@cli.command(\"job\")\n@click.argument(\"tileset\", required=True, type=str)\n@click.argument(\"job_id\", required=True, type=str)\n@click.option(\"--token\", \"-t\", required=False, type=str, help=\"Mapbox access token\")\n@click.option(\"--indent\", type=int, default=None, help=\"Indent for JSON output\")\ndef job(tileset, job_id, token=None, indent=None):\n \"\"\"View a single job for a particular tileset.\n\n tilesets job \n \"\"\"\n mapbox_api = _get_api()\n mapbox_token = _get_token(token)\n url = \"{0}/tilesets/v1/{1}/jobs/{2}?access_token={3}\".format(\n mapbox_api, tileset, job_id, mapbox_token\n )\n r = requests.get(url)\n\n click.echo(json.dumps(r.json(), indent=indent))\n\n\n@cli.command(\"list\")\n@click.argument(\"username\", required=True, type=str)\n@click.option(\n \"--verbose\",\n \"-v\",\n required=False,\n is_flag=True,\n help=\"Will print all tileset information\",\n)\n@click.option(\"--token\", \"-t\", required=False, type=str, help=\"Mapbox access token\")\n@click.option(\"--indent\", type=int, default=None, help=\"Indent for JSON output\")\ndef list(username, verbose, token=None, indent=None):\n \"\"\"List all tilesets for an account.\n By default the response is a simple list of tileset IDs.\n If you would like an array of all tileset's information,\n use the --versbose flag.\n\n tilests list \n \"\"\"\n mapbox_api = _get_api()\n mapbox_token = _get_token(token)\n url = \"{0}/tilesets/v1/{1}?access_token={2}\".format(\n mapbox_api, username, mapbox_token\n )\n r = requests.get(url)\n if r.status_code == 200:\n if verbose:\n for tileset in r.json():\n click.echo(json.dumps(tileset, indent=indent))\n else:\n for tileset in r.json():\n click.echo(tileset[\"id\"])\n else:\n raise errors.TilesetsError(r.text)\n\n\n@cli.command(\"validate-recipe\")\n@click.argument(\"recipe\", required=True, type=click.Path(exists=True))\n@click.option(\"--token\", \"-t\", required=False, type=str, help=\"Mapbox access token\")\n@click.option(\"--indent\", type=int, default=None, help=\"Indent for JSON output\")\ndef validate_recipe(recipe, token=None, indent=None):\n \"\"\"Validate a Recipe JSON document\n\n tilesets validate-recipe \n \"\"\"\n mapbox_api = _get_api()\n mapbox_token = _get_token(token)\n url = \"{0}/tilesets/v1/validateRecipe?access_token={1}\".format(\n mapbox_api, mapbox_token\n )\n with open(recipe) as json_recipe:\n recipe_json = json.load(json_recipe)\n\n r = requests.put(url, json=recipe_json)\n click.echo(json.dumps(r.json(), indent=indent))\n\n\n@cli.command(\"view-recipe\")\n@click.argument(\"tileset\", required=True, type=str)\n@click.option(\"--token\", \"-t\", required=False, type=str, help=\"Mapbox access token\")\n@click.option(\"--indent\", type=int, default=None, help=\"Indent for JSON output\")\ndef view_recipe(tileset, token=None, indent=None):\n \"\"\"View a tileset's recipe JSON\n\n tilesets view-recipe \n \"\"\"\n mapbox_api = _get_api()\n mapbox_token = _get_token(token)\n url = \"{0}/tilesets/v1/{1}/recipe?access_token={2}\".format(\n mapbox_api, tileset, mapbox_token\n )\n r = requests.get(url)\n if r.status_code == 200:\n click.echo(json.dumps(r.json(), indent=indent))\n else:\n raise errors.TilesetsError(r.text)\n\n\n@cli.command(\"update-recipe\")\n@click.argument(\"tileset\", required=True, type=str)\n@click.argument(\"recipe\", required=True, type=click.Path(exists=True))\n@click.option(\"--token\", \"-t\", required=False, type=str, help=\"Mapbox access token\")\n@click.option(\"--indent\", type=int, default=None, help=\"Indent for JSON output\")\ndef update_recipe(tileset, recipe, token=None, indent=None):\n \"\"\"Update a Recipe JSON document for a particular tileset\n\n tilesets update-recipe \n \"\"\"\n mapbox_api = _get_api()\n mapbox_token = _get_token(token)\n url = \"{0}/tilesets/v1/{1}/recipe?access_token={2}\".format(\n mapbox_api, tileset, mapbox_token\n )\n with open(recipe) as json_recipe:\n recipe_json = json.load(json_recipe)\n\n r = requests.patch(url, json=recipe_json)\n if r.status_code == 201:\n click.echo(\"Updated recipe.\", err=True)\n click.echo(r.text)\n else:\n raise errors.TilesetsError(r.text)\n\n\n@cli.command(\"validate-source\")\n@cligj.features_in_arg\ndef validate_source(features):\n \"\"\"Validate your source file.\n $ tilesets validate-source \n \"\"\"\n click.echo(f\"Validating features\", err=True)\n\n for feature in features:\n utils.validate_geojson(feature)\n\n click.echo(\"✔ valid\")\n\n\n@cli.command(\"add-source\")\n@click.argument(\"username\", required=True, type=str)\n@click.argument(\"id\", required=True, type=str)\n@cligj.features_in_arg\n@click.option(\"--no-validation\", is_flag=True, help=\"Bypass source file validation\")\n@click.option(\"--token\", \"-t\", required=False, type=str, help=\"Mapbox access token\")\n@click.option(\"--indent\", type=int, default=None, help=\"Indent for JSON output\")\n@click.pass_context\ndef add_source(ctx, username, id, features, no_validation, token=None, indent=None):\n \"\"\"Create/add a tileset source\n\n tilesets add-source \n \"\"\"\n mapbox_api = _get_api()\n mapbox_token = _get_token(token)\n url = (\n f\"{mapbox_api}/tilesets/v1/sources/{username}/{id}?access_token={mapbox_token}\"\n )\n\n with tempfile.TemporaryFile() as file:\n for feature in features:\n if not no_validation:\n utils.validate_geojson(feature)\n file.write((json.dumps(feature) + \"\\n\").encode(\"utf-8\"))\n\n file.seek(0)\n m = MultipartEncoder(fields={\"file\": (\"file\", file)})\n resp = requests.post(\n url,\n data=m,\n headers={\n \"Content-Disposition\": \"multipart/form-data\",\n \"Content-type\": m.content_type,\n },\n )\n\n if resp.status_code == 200:\n click.echo(json.dumps(resp.json(), indent=indent))\n else:\n raise errors.TilesetsError(resp.text)\n\n\n@cli.command(\"view-source\")\n@click.argument(\"username\", required=True, type=str)\n@click.argument(\"id\", required=True, type=str)\n@click.option(\"--token\", \"-t\", required=False, type=str, help=\"Mapbox access token\")\n@click.option(\"--indent\", type=int, default=None, help=\"Indent for JSON output\")\ndef view_source(username, id, token=None, indent=None):\n \"\"\"View a Tileset Source's information\n\n tilesets view-source \n \"\"\"\n mapbox_api = _get_api()\n mapbox_token = _get_token(token)\n url = \"{0}/tilesets/v1/sources/{1}/{2}?access_token={3}\".format(\n mapbox_api, username, id, mapbox_token\n )\n r = requests.get(url)\n if r.status_code == 200:\n click.echo(json.dumps(r.json(), indent=indent))\n else:\n raise errors.TilesetsError(r.text)\n\n\n@cli.command(\"delete-source\")\n@click.argument(\"username\", required=True, type=str)\n@click.argument(\"id\", required=True, type=str)\n@click.option(\"--force\", \"-f\", is_flag=True, help=\"Circumvents confirmation prompt\")\n@click.option(\"--token\", \"-t\", required=False, type=str, help=\"Mapbox access token\")\ndef delete_source(username, id, force, token=None):\n \"\"\"Delete a Tileset Source + all of its files.\n\n tilesets delete-source \n \"\"\"\n if not force:\n click.confirm(\n \"Are you sure you want to delete {0} {1}?\".format(username, id), abort=True\n )\n mapbox_api = _get_api()\n mapbox_token = _get_token(token)\n url = \"{0}/tilesets/v1/sources/{1}/{2}?access_token={3}\".format(\n mapbox_api, username, id, mapbox_token\n )\n r = requests.delete(url)\n if r.status_code == 204:\n click.echo(\"Source deleted.\")\n else:\n raise errors.TilesetsError(r.text)\n\n\n@cli.command(\"list-sources\")\n@click.argument(\"username\", required=True, type=str)\n@click.option(\"--token\", \"-t\", required=False, type=str, help=\"Mapbox access token\")\ndef list_sources(username, token=None):\n \"\"\"List all Tileset Sources for an account. Response is an un-ordered array of sources.\n\n tilesets list-sources \n \"\"\"\n mapbox_api = _get_api()\n mapbox_token = _get_token(token)\n url = \"{0}/tilesets/v1/sources/{1}?access_token={2}\".format(\n mapbox_api, username, mapbox_token\n )\n r = requests.get(url)\n if r.status_code == 200:\n for source in r.json():\n click.echo(source[\"id\"])\n else:\n raise errors.TilesetsError(r.text)\n","sub_path":"mapbox_tilesets/scripts/cli.py","file_name":"cli.py","file_ext":"py","file_size_in_byte":13707,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"93968242","text":"import matplotlib.pyplot as plt\n\nfrom AWGN_Channel_Transmission.AWGN_Discrete_Density_Evolution import \\\n AWGN_Discrete_Density_Evolution_class_irregular as DDE_irregular\nfrom Discrete_LDPC_decoding.Information_Matching import *\n\n__author__ = \"Maximilian Stark\"\n__copyright__ = \"2016, Institute of Communications, University of Technology Hamburg\"\n__credits__ = [\"Maximilian Stark\"]\n__version__ = \"1.0\"\n__email__ = \"maximilian.stark@tuhh.de\"\n__status__ = \"Production\"\n__name__ = \"Decoder Generation\"\n__doc__ = \"\"\"This script generates a discrete decoder for the desired design-Eb/N0.\"\"\"\n\n\n\n# set noise level for DE\nEbN0_dB_mapping_gen = 0.7\nfor EbN0_dB_mapping_gen in np.array([0.6,0.7,0.8,0.9,1.0]):\n # set quantizer limits\n AD_Max_abs = 3\n plt.figure()\n\n cardinality_Y_channel = 2000\n cardinality_T_channel = 16\n cardinality_T_decoder_ops = 16\n i_max = 50\n nror = 10\n\n # 1 2 3 4 5 6 7\n d_c_dist = np.array([0,0,0,0,0,1,32399]) / 32400\n # 1 2 3 4 5 6 7 8\n d_v_dist = np.array([1,32399,19440,0,0,0,0,12960])/64800\n\n\n lambda_vec = convert_node_to_edge_degree(d_v_dist)\n rho_vec = convert_node_to_edge_degree(d_c_dist)\n\n #R_c = 1-d_v/d_c # code rate\n R_c = 1 - (d_v_dist*(np.arange(d_v_dist.shape[0])+1)).sum() / (d_c_dist*(np.arange(d_c_dist.shape[0])+1)).sum() # code rate\n\n sigma_n2 = 10**(-EbN0_dB_mapping_gen/10) / (2*R_c)\n steps = 5\n\n config = 'cas'\n # generate decoder config\n DDE_inst = DDE_irregular(sigma_n2, AD_Max_abs, cardinality_Y_channel, cardinality_T_channel,\n cardinality_T_decoder_ops, lambda_vec, rho_vec, i_max, nror , match = True)\n\n DDE_inst.run_discrete_density_evolution()\n DDE_inst.save_config(config)\n plt.plot(DDE_inst.DDE_inst_data['MI_T_dvm1_v_X_dvm1_v'],label='match')\n\n\n # DDE_inst = DDE_irregular(sigma_n2, AD_Max_abs, cardinality_Y_channel, cardinality_T_channel,\n # cardinality_T_decoder_ops, lambda_vec, rho_vec, i_max, nror , match = False)\n #\n # DDE_inst.run_discrete_density_evolution()\n # DDE_inst.save_config('adapt_no_match')\n # plt.plot(DDE_inst.DDE_inst_data['MI_T_dvm1_v_X_dvm1_v'],label='no match')\n # plt.legend(loc=4)\n ","sub_path":"Irregular_LDPC_Decoding/DVB-S2/decoder_config_generation.py","file_name":"decoder_config_generation.py","file_ext":"py","file_size_in_byte":2236,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"256562645","text":"import csv\nimport sqlite3\n\n\nconnection = sqlite3.connect('/home/haletod/PycharmProjects/Ligase/students_and_alcohol.db')\ncursor = connection.cursor()\n\n\n# create table with students info\n\n(cursor.execute(\n \"CREATE TABLE IF NOT EXISTS students_info (student_id INTEGER, sex TEXT, age INTEGER, famsize TEXT, Pstatus TEXT, \"\n \"failures INTEGER, health INTEGER, guardian TEXT, PRIMARY KEY(student_id));\"))\n\nwith open('students.csv') as csv_data:\n data = csv.DictReader(csv_data)\n to_db = [\n (i['student_id'], i['sex'], i['age'], i['famsize'], i['Pstatus'], i['failures'], i['health'], i['guardian'])\n for i in data]\n\n(cursor.executemany(\n \"INSERT INTO students_info (student_id, sex, age, famsize, Pstatus, failures, health, guardian) \"\n \"VALUES (?, ?, ?, ?, ?, ?, ?, ?);\",\n to_db))\n\n\n# create table with students alcohol consumption\n\n(cursor.execute(\n \"CREATE TABLE IF NOT EXISTS alcohol_consumption (student_id INTEGER, Daily INTEGER, Weekly INTEGER, \"\n \"PRIMARY KEY(student_id) FOREIGN KEY(student_id) REFERENCES students_info(student_id) ON DELETE CASCADE \"\n \"ON UPDATE CASCADE);\"))\n\nwith open('students.csv') as csv_data:\n data = csv.DictReader(csv_data)\n to_db = [(i['student_id'], i['Dalc'], i['Walc']) for i in data]\n\ncursor.executemany(\"INSERT INTO alcohol_consumption (student_id, Daily, Weekly) VALUES (?, ?, ?);\", to_db)\n\nconnection.commit()\nconnection.close()\n","sub_path":"SQL/SQL_DB_creation.py","file_name":"SQL_DB_creation.py","file_ext":"py","file_size_in_byte":1418,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"174232316","text":"import numpy as np\nimport os\nfrom search import * #for search engines\nfrom sokoban import SokobanState, Direction, sokoban_goal_state #for Sokoban specific classes and problems\nfrom test_problems import PROBLEMS\n\n#SOKOBAN HEURISTICS\ndef heur_displaced(state):\n '''trivial admissible sokoban heuristic'''\n '''INPUT: a sokoban state'''\n '''OUTPUT: a numeric value that serves as an estimate of the distance of the state to the goal.''' \n count = 0\n for box in state.boxes:\n if box not in state.storage:\n count += 1\n return count\n\ndef heur_manhattan_distance(state):\n#IMPLEMENT\n '''admissible sokoban heuristic: manhattan distance'''\n '''INPUT: a sokoban state'''\n '''OUTPUT: a numeric value that serves as an estimate of the distance of the state to the goal.''' \n #We want an admissible heuristic, which is an optimistic heuristic. \n #It must always underestimate the cost to get from the current state to the goal.\n #The sum Manhattan distance of the boxes to their closest storage spaces is such a heuristic. \n #When calculating distances, assume there are no obstacles on the grid and that several boxes can fit in one storage bin.\n #You should implement this heuristic function exactly, even if it is tempting to improve it.\n #Your function should return a numeric value; this is the estimate of the distance to the goal.\n manhattan_distance = 0\n current_state = []\n goal_state = []\n for box in state.boxes:\n current_state.append(box)\n for storage in state.storage:\n goal_state.append(storage)\n for i in range(len(goal_state)):\n manhattan_distance+= abs(current_state[i][0]-goal_state[i][0])\n manhattan_distance+= abs(current_state[i][1]-goal_state[i][1])\n return manhattan_distance\n\ndef heur_alternate(state):\n#IMPLEMENT\n '''a better sokoban heuristic'''\n '''INPUT: a sokoban state'''\n '''OUTPUT: a numeric value that serves as an estimate of the distance of the state to the goal.''' \n #heur_manhattan_distance has flaws. \n #Write a heuristic function that improves upon heur_manhattan_distance to estimate distance between the current state and the goal.\n #Your function should return a numeric value for the estimate of the distance to the goal.\n cost = 0\n if check_corners(state): return float(\"inf\")\n cost += robot_beside_nothing(state)\n cost += distance(state)\n return cost \n\ndef distance(state):\n final_cost = 0\n robot_distance = float(\"inf\")\n robot_position = state.robot\n for box in state.boxes:\n possible_storage = get_possible_storage(box, state)\n tempcost = []\n old_cost = float(\"inf\")\n for possible in possible_storage:\n if box == possible:\n old_cost = 0\n break\n else:\n new_cost = calculate_simple_distance(box, possible, state)\n if new_cost <= old_cost:\n old_cost = new_cost\n final_cost +=old_cost\n if box not in possible_storage:\n final_cost += get_closeness(box,state)\n new_robot_distance = calculate_simple_distance(robot_position, box, state)\n if new_robot_distance=state.width: return True\n if box[1] >=state.height: return True\n return False\n\ndef get_top(box):\n return (box[0],box[1]+1)\ndef get_bottom(box):\n return (box[0],box[1]-1)\ndef get_left(box):\n return (box[0]-1,box[1])\ndef get_right(box):\n return (box[0]+1,box[1])\n\ndef robot_beside_nothing(state):\n robot_position = state.robot\n cost = 0\n if (robot_position[0]+1, robot_position[1]) in state.boxes:\n test = (robot_position[0]+2, robot_position[1]) in state.boxes\n if test in state.boxes or test in state.obstacles:\n cost+= 2\n else:\n return cost\n if (robot_position[0]-1, robot_position[1]) in state.boxes:\n test = (robot_position[0]-2, robot_position[1]) in state.boxes\n if test in state.boxes or test in state.obstacles:\n cost+= 2\n else:\n return cost\n if (robot_position[0], robot_position[1]+1) in state.boxes:\n test = (robot_position[0], robot_position[1]+2) in state.boxes\n if test in state.boxes or test in state.obstacles:\n cost+= 2\n else:\n return cost\n if (robot_position[0], robot_position[1]-1) in state.boxes:\n test = (robot_position[0], robot_position[1]-2) in state.boxes\n if test in state.boxes or test in state.obstacles:\n cost+= 2\n else:\n return cost\n cost+=1\n if (robot_position[0]+1, robot_position[1]+1) in state.boxes: return cost\n if (robot_position[0]-1, robot_position[1]-1) in state.boxes: return cost\n if (robot_position[0]-1, robot_position[1]+1) in state.boxes: return cost\n if (robot_position[0]+1, robot_position[1]-1) in state.boxes: return cost\n return cost+2\n\ndef calculate_simple_distance(box, possible,state):\n return abs(box[0]-possible[0])+ abs(box[1]-possible[1])\n\ndef is_cornered(position, state):\n if position[0] == 0:\n if position[1] == 0: return True\n if position[1] == state.height-1: return True\n if (position[0], position[1]-1) in state.obstacles: return True\n if (position[0], position[1]+1) in state.obstacles: return True\n return False \n if position[0] == state.width-1:\n if position[1] == 0: return True\n if position[1] == state.height-1: return True\n if (position[0]-1, position[1]) in state.obstacles: return True\n if (position[0]+1, position[1]) in state.obstacles: return True \n return False \n testabove = (position[0]-1, position[1])\n testbelow = (position[0]+1, position[1])\n testleft = (position[0], position[1]-1)\n testright = (position[0], position[1]+1)\n if testabove in state.obstacles:\n if testleft in state.obstacles: return True\n if testright in state.obstacles: return True\n if testbelow in state.obstacles:\n if testleft in state.obstacles: return True\n if testright in state.obstacles: return True\n return False\n\ndef check_corners(state):\n for box in state.boxes:\n possible_storage = get_possible_storage(box, state)\n if box not in possible_storage:\n if is_cornered(box, state): return True\n # if is_edge(box, possible_storage,state): return True\n return False\n\ndef get_possible_storage(box,state):\n if state.restrictions != None:\n possible = state.restrictions[state.boxes[box]]\n if box in possible:\n return [box]\n for other_boxes in state.boxes:\n if box != other_boxes:\n if other_boxes in possible and other_boxes in state.restrictions[state.boxes[other_boxes]]:\n possible = possible.difference(other_boxes)\n return possible\n else:\n possible = []\n for place in state.storage:\n possible.append(place)\n if box in possible:\n return [box]\n for other_boxes in state.boxes: \n if box != other_boxes:\n if other_boxes in possible:\n possible.remove(other_boxes) \n return possible\n\n\ndef fval_function(sN, weight):\n#IMPLEMENT\n \"\"\"\n Provide a custom formula for f-value computation for Anytime Weighted A star.\n Returns the fval of the state contained in the sNode.\n\n @param sNode sN: A search node (containing a SokobanState)\n @param float weight: Weight given by Anytime Weighted A star\n @rtype: float\n \"\"\"\n \n #Many searches will explore nodes (or states) that are ordered by their f-value.\n #For UCS, the fvalue is the same as the gval of the state. For best-first search, the fvalue is the hval of the state.\n #You can use this function to create an alternate f-value for states; this must be a function of the state and the weight.\n #The function must return a numeric f-value.\n #The value will determine your state's position on the Frontier list during a 'custom' search.\n #You must initialize your search engine object as a 'custom' search engine if you supply a custom fval function.\n fval = sN.gval + weight*sN.hval\n return fval\n\n\ndef anytime_gbfs(initial_state, heur_fn, timebound = 10):\n#IMPLEMENT\n '''Provides an implementation of anytime greedy best-first search, as described in the HW1 handout'''\n '''INPUT: a sokoban state that represents the start state and a timebound (number of seconds)'''\n '''OUTPUT: A goal state (if a goal is found), else False''' \n time_end = os.times()[0]+timebound\n search = SearchEngine('best_first')\n search.init_search(initial_state, sokoban_goal_state, heur_fn)\n time_left = time_end - os.times()[0]\n output = False\n prev_cost = float(\"inf\")\n while time_left > 0:\n goal = search.search(time_left)\n if goal != False:\n if goal.gval 0:\n goal = search.search(time_left)\n if goal != False:\n if goal.gval 0:\n nums.append(re.findall('([0-9]+)', line))\n\t\nfor num in nums:\n for n in num:\n total = total + int(n)\n\n\nprint(total)","sub_path":"regex.py","file_name":"regex.py","file_ext":"py","file_size_in_byte":268,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"292896724","text":"\"\"\"/pidor command.\"\"\"\n\nfrom telegram import Update, InlineKeyboardButton, InlineKeyboardMarkup\nfrom telegram.ext import CallbackContext, run_async\nfrom telegram.error import BadRequest\n\nfrom main import randomizer\nfrom main.database import *\nfrom main.constants import DEVS\nfrom main.helpers import antispam_passed, check_if_group_chat, ResetError\n\n\n@run_async\n@antispam_passed\n@check_if_group_chat\n@db_session\ndef pidor(update: Update, context: CallbackContext):\n \"\"\"Get the pidor of the day from all users stored for the chat.\"\"\"\n # Check if there is already a pidor of the day\n pidor_today = select(q.user_id.full_name for q in Pidors\n if q.chat_id == Chats[update.message.chat.id]\n and q.day == date.today())[:][:]\n if pidor_today:\n update.message.reply_text(\n text=f'Пидором дня является {pidor_today[0]}!')\n return\n keyboard = InlineKeyboardMarkup.from_button(\n InlineKeyboardButton('Реролл #1 (только админы)', callback_data='Reroll.1'))\n update.message.reply_text(text=f'Пидором дня является {getnew(update).result()}!',\n parse_mode='Markdown',\n reply_markup=keyboard)\n\n\n@run_async\n@db_session\ndef getnew(update: Update) -> str:\n \"\"\"Look for new pidor.\"\"\"\n chat_users = select(q.user_id for q in User_Stats\n if q.chat_id == Chats[update.message.chat.id])[:][:]\n # Find a pidor that's still in the chat and delete those that are gone.\n while chat_users:\n pidor = randomizer.choice(chat_users)\n try:\n pidor_data = update.message.chat.get_member(user_id=pidor.id)\n if pidor_data.status not in ['restricted', 'left', 'kicked'] and \\\n not pidor_data.user.is_bot:\n break\n else:\n delete(u for u in User_Stats\n if u.user_id == pidor\n and u.chat_id == Chats[update.message.chat.id])\n chat_users.remove(pidor)\n except BadRequest:\n delete(u for u in User_Stats\n if u.user_id == pidor\n and u.chat_id == Chats[update.message.chat.id])\n chat_users.remove(pidor)\n else:\n update.message.reply_text('Нужно больше данных!')\n raise ResetError\n # Assign a tag\n Users[pidor.id].full_name = pidor_data.user.full_name\n pidor_tag = f'[{pidor.full_name}](tg://user?id={pidor.id})'\n if not Pidors.exists(chat_id=Chats[update.message.chat.id]):\n Pidors(chat_id=Chats[update.message.chat.id],\n user_id=pidor,\n day=date.today())\n else:\n Pidors[Chats[update.message.chat.id]].user_id = pidor\n Pidors[Chats[update.message.chat.id]].day = date.today()\n # Record and return\n User_Stats[Users[pidor.id],\n Chats[update.message.chat.id]].times_pidor += 1\n return pidor_tag\n\n\n@run_async\n@db_session\ndef reroll(update: Update, context: CallbackContext):\n \"\"\"Reroll pidor of the day.\"\"\"\n admins = [u.user for u in context.bot.get_chat_administrators(\n update.callback_query.message.chat.id)]\n if update.callback_query.from_user in admins or \\\n update.callback_query.from_user.id in DEVS:\n rolln = int(\n update.callback_query.message.reply_markup.inline_keyboard[0][0].callback_data.split('.')[-1]) + 1\n keyboard = InlineKeyboardMarkup.from_button(\n InlineKeyboardButton(f'Реролл #{rolln} (только админы)',\n callback_data=f'Reroll.{rolln}'))\n update.callback_query.message.edit_text(\n text=f'Пидором дня является {getnew(update.callback_query).result()}!',\n parse_mode='Markdown',\n reply_markup=keyboard)\n","sub_path":"main/commands/pidor.py","file_name":"pidor.py","file_ext":"py","file_size_in_byte":3930,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"632487797","text":"import mysql.connector\n\n#global variable allows us to update page\ntimeToUpdate = 0\n#\ndef createTables(mycursor):\n try:\n mycursor.execute(\"CREATE TABLE IF NOT EXISTS StudentInfo(BaylorID CHAR (9),\"\n \"lastName VARCHAR (30),\"\n \"firstName VARCHAR (30),\"\n \"emailAddress VARCHAR (30),\"\n \"ADV_PR_semester VARCHAR (30),\"\n \"class VARCHAR (20),\"\n \"major_minor VARCHAR(5),\"\n \"ADV_PR_grade CHAR(1),\"\n \"ADV_PR_year CHAR(4),\"\n \"PRIMARY KEY (BaylorID))\")\n except mysql.connector.Error as err:\n print(err.msg)\n\n try:\n mycursor.execute(\"CREATE INDEX assign_ibfk_1 ON StudentInfo(BaylorID)\")\n except mysql.connector.errors.ProgrammingError as err:\n index_created = True\n\n try:\n mycursor.execute(\"CREATE TABLE IF NOT EXISTS Internship(company VARCHAR (50),\"\n \"startMonth VARCHAR (15),\"\n \"startYear CHAR (4),\"\n \"endMonth VARCHAR (15),\"\n \"endYear CHAR (4),\"\n \"address VARCHAR(80),\"\n \"phoneNumber CHAR(11),\"\n \"totalHours INT,\"\n \"BaylorID CHAR(9),\"\n \"supervisorName VARCHAR(50),\"\n \"PRIMARY KEY (BaylorID, company, supervisorName),\"\n \"FOREIGN KEY (BaylorID) REFERENCES StudentInfo(BaylorID))\")\n except mysql.connector.Error as err:\n print(err.msg)\n\n\n\ndef insertIntoStudentInfo(idEntry, lastnameEntry, firstnameEntry, emailEntry, semesterEntry, classyr, major_minor, grade, year, mycursor, mydb, middleFrame, topFrame):\n try:\n sqlFormula = \"INSERT INTO StudentInfo (BaylorID, lastName, firstName, emailAddress, ADV_PR_semester, class, major_minor, ADV_PR_grade, ADV_PR_year) \" \\\n \"VALUES (%s, %s, %s, %s, %s, %s, %s, %s, %s, %s)\"\n mycursor.execute(sqlFormula, (idEntry, lastnameEntry, firstnameEntry, emailEntry, semesterEntry, classyr, major_minor, grade, year))\n mydb.commit()\n except mysql.connector.Error as error:\n print(\"could not be inserted\")\n\ndef insertIntoInternship(companyEntry, startmoEntry, startyrEntry, endmoEntry, endyrEntry, addressEntry, numberEntry, totHoursEntry, idEntry, supNameEntry, mycursor, mydb, middleFrame, topFrame):\n # print (\"The company is\", companyEntry)\n # print (\"The start month is\", startmoEntry)\n # print (\"The start year is\", startyrEntry)\n # print (\"The end month is\", endmoEntry)\n # print (\"The end year is\", endyrEntry)\n # print (\"The address is\", addressEntry)\n # print (\"The number is\", numberEntry)\n # print (\"The total hours is\", totHoursEntry)\n # print (\"The id is\", idEntry)\n # print (\"The supervisor is\", supNameEntry)\n\n try:\n sqlFormula = \"INSERT INTO Internship (company, startMonth, startYear, endMonth, endYear, address, phoneNumber, totalHours, BaylorID, supervisorName) \" \\\n \"VALUES (%s, %s, %s, %s, %s, %s, %s, %s, %s, %s)\"\n mycursor.execute(sqlFormula, (companyEntry, startmoEntry, startyrEntry, endmoEntry, endyrEntry, addressEntry, numberEntry, totHoursEntry, idEntry, supNameEntry))\n mydb.commit()\n except mysql.connector.Error as error:\n print(\"could not be inserted\")\n","sub_path":"Marquise_Working_Flask/User_Inputed_Data.py","file_name":"User_Inputed_Data.py","file_ext":"py","file_size_in_byte":3486,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"438353296","text":"from GameObject import GameObject\r\nfrom EnemyComponents import *\r\nfrom PlayerComponents import *\r\nfrom Room import Room\r\n\r\n\r\nclass EntityFactory:\r\n @classmethod\r\n def create_entity(cls, entity: dict):\r\n if entity['type'] == 'Level':\r\n collision = pygame.image.load(entity[\"collision\"])\r\n return Room(collision=pygame.image.load(entity[\"collision\"]),\r\n image=pygame.image.load(entity[\"graphics\"]),)\r\n\r\n elif entity['type'] == 'Player':\r\n collision = pygame.image.load(entity[\"collision\"])\r\n mask = pygame.mask.from_surface(collision)\r\n return GameObject(PlayerStateComponent(), GraphicsComponent(player_frames), collision=mask)\r\n\r\n elif entity['type'] == 'Slime':\r\n return GameObject(EnemyStateComponent(), GraphicsComponent(slime_frames))\r\n\r\n elif entity['type'] == 'Sword':\r\n collision = pygame.image.load(entity[\"collision\"])\r\n mask = pygame.mask.from_surface(collision)\r\n return GameObject(PlayerStateComponent(), GraphicsComponent(sword_all_frames), collision = mask)\r\n\r\n elif entity['type'] == 'Shield':\r\n collision = pygame.image.load(entity[\"collision\"])\r\n mask = pygame.mask.from_surface(collision)\r\n return GameObject(PlayerStateComponent(), GraphicsComponent(shield_all_frames), collision = mask)\r\n","sub_path":"EntityFactory.py","file_name":"EntityFactory.py","file_ext":"py","file_size_in_byte":1404,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"130961111","text":"# Create the empty lists\nflightID = []\ntrackID = []\nx = []\ny = []\nmodec = []\ncallsign = []\nicao = []\ndest = []\nadep = []\nflighttype = []\nradar = []\ntakeofftime = []\ntime = []\n\n\nxp = []\nyp = []\n\n\n# This part opens the file to be read, it's the Schiphol outbound data file\nf = open(\"outbound.txt\", \"r\")\nlines = f.readlines()\n\n# First the file is made readable by replacing ',' with ' ',\n# then the data lists are made, flightID1 and trackID1 were assumed redundant\n# It runs from [1:421] to get rid of the first line and to have\n# only two flights for testing purposes. For only one flight, use [1:216] instead.\n\n \ntxt = []\n\nfor line in lines[1:]:\n txt = line.split(',')\n flightID.append(txt[0])\n trackID.append(txt[2])\n x.append(txt[4])\n y.append(txt[5])\n modec.append(txt[6])\n callsign.append(txt[7])\n icao.append(txt[8])\n dest.append(txt[9])\n adep.append(txt[10])\n flighttype.append(txt[11])\n radar.append(txt[12])\n takeofftime.append(txt[13])\n time.append(txt[14])\n\n# Now all the numbers have to be changed to floats\nxp = map(float,x)\nyp = map(float,y)\n\n# The time has to be made usable\n# First all the lists\ntime0 = []\ntime1 = []\ntime2 = []\ntime3 = []\ntime4 = []\ntime5 = []\ntime6 = []\n\ntime0l = []\ntime1l = []\ntime2l = []\ntime3l = []\ntime4l = []\ntime5l = []\nt = []\n\nt = []\ntl = []\nh = []\nm = []\ns = []\n\n# Now we only want the actual time in h:m:s, date is removed\nfor t0 in time:\n time0 = t0.split('-')\n time0l.append(time0)\n\nfor i in range(len(time0l)):\n time1 = time0l[i][2]\n time1l.append(time1)\n\nfor t1 in time1l:\n time2 = t1.split(' ')\n time2l.append(time2)\n\nfor j in range(len(time2l)):\n time3 = time2l[j][1]\n time3l.append(time3)\n\nfor t3 in time3l:\n time4 = t3.split(':')\n time4l.append(time4)\n\n# Now the time is converted to floats\nfor k in range(len(time4l)):\n time5 = map(float,time4l[k])\n time5l.append(time5)\n\n# And the time is calculated in seconds, which is put in the final list called t\nfor l in range(len(time5l)):\n time6 = 3600*time5l[l][0] + 60*time5l[l][1] + time5l[l][2]\n t.append(time6)\n\n# If you're interested in a specific flight, use the code below to figure out when it occurs.\n# I believe the interval should then be [first point +1:last point +2]\n\n##for p in range(len(callsign)):\n## if callsign[p] == 'TRA883':\n## print p\n\n","sub_path":"Datareading.py","file_name":"Datareading.py","file_ext":"py","file_size_in_byte":2352,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"245869280","text":"# função calcular se o voto é obrigatorio\n\ndef idade(a):\n from datetime import date\n return date.today().year - a\n\n\ndef voto(i):\n if i < 16:\n return 'Negado'\n elif i > 18 and i < 65:\n return 'Obrigatório'\n elif i >= 16 or i >= 65:\n return 'Opcional'\n\n\nano = int(input('Em que ano você nasceu? '))\nidade(ano) # Calcula apenas a idade\nsituacao = voto(idade(ano)) # situacao recebe o resultado da funcao voto, q tem como paramentro\n# a idade da pessoa, calculada por outra função.\nprint(f'Você tem {idade(ano)} anos: Voto {situacao}.')\n","sub_path":"CursoEmVideo/pythonProject/ex101.py","file_name":"ex101.py","file_ext":"py","file_size_in_byte":577,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"302790781","text":"t = int(input())\nfor i in range(t):\n l = input()\n arr = input()\n arr = [int(k) for k in arr.split(\" \")]\n arr.sort()\n if arr[-1] % arr[0] == 0:\n print(arr[-1])\n else:\n if arr[-1] % 2 == 0 and arr[0] % 2 == 0:\n arr[0] /= 2\n print(arr[0] * arr[-1])\n","sub_path":"Code/CodeRecords/2779/49361/295192.py","file_name":"295192.py","file_ext":"py","file_size_in_byte":296,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"19197086","text":"from google.appengine.api import users\r\nfrom google.appengine.ext import ndb\r\n\r\nimport webapp2\r\n\r\nfrom mydicts import *\r\nfrom myschemas import *\r\nfrom myadmin import *\r\n\r\nfrom wordtemplates import *\r\nfrom utils import *\r\nfrom modelutils import *\r\n\r\ndef deleteword(request,wordid):\r\n dict_name = request.request.get('dict_name', WORDDICT)\r\n word_key = ndb.Key(urlsafe=wordid)\r\n word = word_key.get()\r\n word.key.delete()\r\n\r\ndef clearwords(request):\r\n words_query = Word.query()\r\n words = words_query.fetch()\r\n \r\n for word in words:\r\n deleteword(request,word.key.urlsafe())\r\n\r\n# [START ListWords]\r\nclass ListWords(webapp2.RequestHandler):\r\n def get(self):\r\n self.response.write('')\r\n\r\n sdict_name = self.request.get('dict_name',WORDDICT)\r\n cdict_name = self.request.get('dict_name',CHICHARDICT)\r\n\r\n words_query = Word.query(ancestor=dict_key(sdict_name)).order(-Word.date)\r\n words = words_query.fetch()\r\n\r\n wordlist = \"\"\r\n wordlist = wordlist + \"\"\r\n for word in words:\r\n wordlist = wordlist + \"\\n\"\r\n for ochichar in getwordchichars(self,word):\r\n wordlist = wordlist + \"\"\r\n viewwordform = \"
\"\r\n wordlist = wordlist + \"\"\r\n wordlist = wordlist + \"\"\r\n wordlist = wordlist + \"
\" + word.chichar +\" \" + word.translation + \"\" + ochichar.chichar + \"
\" + viewwordform + \"
\"\r\n\r\n self.response.write(LIST_WORD_TEMPLATE % wordlist)\r\n\r\n self.response.write('')\r\n# [END ListWords]\r\n\r\n\r\n# [END LoadWords]\r\n\r\ndef clearwords(request):\r\n word = Word()\r\n words_query = Word.query()\r\n words = words_query.fetch()\r\n \r\n for word in words:\r\n deleteword(request,word.key.urlsafe())\r\n \r\n\r\n# [START ClearWords]\r\nclass ClearWords(webapp2.RequestHandler):\r\n def post(self):\r\n clearwords(self)\r\n self.redirect('/')\r\n# [END ClearWords]\r\n\r\n# [START ViewWord]\r\nclass ViewWord(webapp2.RequestHandler):\r\n def get(self,wordid):\r\n self.response.write('')\r\n\r\n #dict_name = self.request.get('dict_name', WORDDICT)\r\n #word = Word(parent=dict_key(dict_name));\r\n\r\n\r\n dict_name = self.request.get('dict_name',WORDDICT)\r\n word_key = ndb.Key(urlsafe=wordid)\r\n #sandy = sandy_key.get()\r\n #key = ndb.Key(Word, wordid)\r\n #words_query = Word.query(Word.key == key)\r\n #word = words_query.fetch(1)[0]\r\n word = word_key.get()\r\n\r\n user = users.get_current_user()\r\n\r\n if user:\r\n udict_name = self.request.get('dict_name', USERDICT)\r\n viewstat = ViewStat(parent=dict_key(udict_name))\r\n viewstat.email = user.email()\r\n viewstat.word = word\r\n viewstat.put()\r\n\r\n \r\n wordchars = \"\"\r\n for chichar10 in lsplit(getwordchichars(self,word),10):\r\n wordchars = wordchars + \"\"\r\n for lchichar in chichar10:\r\n chichars_query = Chichar.query(Chichar.chichar == lchichar.chichar)\r\n qresult = chichars_query.fetch(1)\r\n chichar = qresult[0]\r\n wordchars = wordchars + \"\"\r\n wordchars = wordchars + \"\"\r\n wordchars = wordchars + \"
\"\r\n\r\n if not user == None and user.email() == ADMIN_ID:\r\n self.response.write(VIEW_WORD_ADMIN_TEMPLATE % ( word.chichar, word.pronunciation, word.translation, wordchars, word.key.urlsafe(),word.key.urlsafe()))\r\n else:\r\n self.response.write(VIEW_WORD_USER_TEMPLATE % ( word.chichar, word.pronunciation, word.translation, wordchars))\r\n \r\n\r\n self.response.write('')\r\n# [END ViewWord]\r\n\r\n# [START AddWord]\r\nclass AddWord(webapp2.RequestHandler):\r\n def get(self):\r\n user = users.get_current_user()\r\n if not user == None and user.email() == ADMIN_ID:\r\n self.response.write('')\r\n self.response.write(ADD_WORD_TEMPLATE)\r\n self.response.write('')\r\n else:\r\n self.response.write('Sorry, you must be ADMIN to access this page')\r\n# [END AddWord]\r\n\r\n# [START DoAddWord]\r\nclass DoAddWord(webapp2.RequestHandler):\r\n def post(self):\r\n dict_name = self.request.get('dict_name', WORDDICT)\r\n word = Word(parent=dict_key(dict_name));\r\n word.chichar = self.request.get('wordchichar')\r\n word.translation = self.request.get('wordtranslation')\r\n word.pronunciation = self.request.get('wordpronunciation')\r\n word.put()\r\n\r\n self.redirect(\"/viewword/\" + word.key.urlsafe())\r\n# [END DoAddWord]\r\n\r\n\r\n# [START EditWord]\r\nclass EditWord(webapp2.RequestHandler):\r\n def get(self,wordid):\r\n self.response.write('')\r\n\r\n dict_name = self.request.get('dict_name', WORDDICT)\r\n word_key = ndb.Key(urlsafe=wordid)\r\n # word = Word(parent=dict_key(dict_name));\r\n word = word_key.get()\r\n\r\n # Write the submission form and the footer of the page\r\n self.response.write(EDIT_WORD_TEMPLATE % ( word.key.urlsafe(), word.chichar, word.translation, word.pronunciation,word.key.urlsafe()))\r\n\r\n self.response.write('')\r\n\r\n# [END EditWord]\r\n\r\n# [START SaveWord]\r\nclass SaveWord(webapp2.RequestHandler):\r\n def post(self,wordid):\r\n save = self.request.get('save')\r\n cancel = self.request.get('cancel')\r\n dict_name = self.request.get('dict_name', WORDDICT)\r\n word_key = ndb.Key(urlsafe=wordid)\r\n # word = Word(parent=dict_key(dict_name));\r\n word = word_key.get()\r\n \r\n if save:\r\n word.chichar = self.request.get('word')\r\n word.translation = self.request.get('translation')\r\n word.pronunciation = self.request.get('pronunciation')\r\n word.put()\r\n\r\n self.redirect(\"/viewword/\" + word.key.urlsafe())\r\n# [END SaveWord]\r\n \r\ndef deleteword(request,wordid):\r\n dict_name = request.request.get('dict_name', WORDDICT)\r\n word_key = ndb.Key(urlsafe=wordid)\r\n word = word_key.get()\r\n word.key.delete()\r\n\r\n\r\n# [START DeleteWord]\r\nclass DeleteWord(webapp2.RequestHandler):\r\n def post(self,wordid):\r\n deleteword(self,wordid)\r\n self.redirect(\"/listwords\")\r\n# [END DeleteWord]\r\n\r\n# [START StatWords]\r\nclass StatWords(webapp2.RequestHandler):\r\n def get(self):\r\n self.response.write('')\r\n \r\n dict_name = self.request.get('dict_name',WORDDICT)\r\n words_query = Word.query(ancestor=dict_key(dict_name)).order(-Word.date)\r\n words = words_query.fetch()\r\n\r\n # Write the submission form and the footer of the page\r\n self.response.write(STAT_WORD_TEMPLATE % ( len(words) ))\r\n\r\n self.response.write('')\r\n# [END StatChiChars]\r\n\r\n","sub_path":"word.py","file_name":"word.py","file_ext":"py","file_size_in_byte":7436,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"43820261","text":"import torch\nimport torch.nn as nn\nimport torch.nn.parallel\nimport torch.backends.cudnn as cudnn\nimport torch.optim\nimport torch.utils.data\nimport torchvision\nimport torchvision.transforms as T\nimport torchvision.datasets as datasets\nimport torchvision.models as models\n\n\nprint (torch.__version__)\n\n# model in pytorch repo with weights \nmodel = models.resnet50(pretrained=True)\nmodel.cuda() # load in GPU\ncudnn.benchmark = True #? needed for profiler? \n\n# pre-process images\ntransform = T.Compose([T.Resize(256), T.CenterCrop(224), T.ToTensor()])\n# Dataset load \ntrainset = torchvision.datasets.CIFAR10(root='./data', train=True, \n download=True, transform=transform)\n# Loading ( parallel workers processes - GIL problem global lock - not running in threads) \ntrainloader = torch.utils.data.DataLoader(trainset, batch_size=8,\n shuffle=True)\n# calc loss (target and training) - and minimize it\ncriterion = nn.CrossEntropyLoss().cuda()\n# back propagation \noptimizer = torch.optim.SGD(model.parameters(), lr=0.001, momentum=0.9)\n\ndevice = torch.device(\"cuda:0\")\n# switch to training mode\nmodel.train()\n\n\nimport torch.profiler\n\ndef output_fn(p):\n p.export_chrome_trace(\"./trace/worker0-batch8/worker0.pt.trace.json\")\n \n# add context manager around training loop\nwith torch.profiler.profile(\n activities=[\n torch.profiler.ProfilerActivity.CPU,\n torch.profiler.ProfilerActivity.CUDA],\n schedule=torch.profiler.schedule(\n wait=2, # skip first 2 training steps\n warmup=2, # reach steady and skip few layers, profiling happens ignores results\n active=6), # only profile 6 steps - allows to focus and skip some layers for reducing overhead(even in prod)\n on_trace_ready=output_fn,\n record_shapes=True\n) as p:\n for step, data in enumerate(trainloader, 0):\n print(\"step:{}\".format(step))\n inputs, labels = data[0].to(device=device), data[1].to(device=device)\n\n outputs = model(inputs)\n loss = criterion(outputs, labels)\n \n optimizer.zero_grad()\n loss.backward()\n optimizer.step()\n # next training step (metadata)\n p.step() \n if step + 1 >= 10:\n break\n","sub_path":"resnet50basic.py","file_name":"resnet50basic.py","file_ext":"py","file_size_in_byte":2301,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"174360570","text":"from nose.tools import assert_equal\n\nfrom pyecharts.commons.utils import remove_key_with_none_value\nfrom pyecharts.options.global_options import (\n AnimationOpts,\n InitOpts,\n ToolBoxFeatureBrushOpts,\n ToolBoxFeatureDataViewOpts,\n ToolBoxFeatureDataZoomOpts,\n ToolBoxFeatureMagicTypeOpts,\n ToolBoxFeatureOpts,\n ToolBoxFeatureRestoreOpts,\n ToolBoxFeatureSaveAsImageOpts,\n ToolboxOpts,\n BrushOpts,\n DataZoomOpts,\n LegendOpts,\n VisualMapOpts,\n TooltipOpts,\n)\n\n\ndef test_animation_options_remove_none():\n option = AnimationOpts()\n expected = {\n \"animation\": True,\n \"animationDelay\": 0,\n \"animationDelayUpdate\": 0,\n \"animationDuration\": 1000,\n \"animationDurationUpdate\": 300,\n \"animationEasing\": \"cubicOut\",\n \"animationEasingUpdate\": \"cubicOut\",\n \"animationThreshold\": 2000,\n }\n assert_equal(expected, remove_key_with_none_value(option.opts))\n\n\ndef test_init_options_remove_none():\n option = InitOpts(animation_opts={})\n expected = {\n \"animationOpts\": {},\n \"height\": \"500px\",\n \"page_title\": \"Awesome-pyecharts\",\n \"renderer\": \"canvas\",\n \"theme\": \"white\",\n \"width\": \"900px\",\n }\n assert_equal(expected, remove_key_with_none_value(option.opts))\n\n\ndef test_toolbox_feature_options_remove_none():\n save_as_image = ToolBoxFeatureSaveAsImageOpts()\n restore = ToolBoxFeatureRestoreOpts()\n data_view = ToolBoxFeatureDataViewOpts()\n data_zoom = ToolBoxFeatureDataZoomOpts()\n magic_type = ToolBoxFeatureMagicTypeOpts()\n brush = ToolBoxFeatureBrushOpts()\n\n option = ToolBoxFeatureOpts(\n save_as_image=save_as_image,\n restore=restore,\n data_view=data_view,\n data_zoom=data_zoom,\n magic_type=magic_type,\n brush=brush,\n )\n expected = {\n \"saveAsImage\": save_as_image,\n \"restore\": restore,\n \"dataView\": data_view,\n \"dataZoom\": data_zoom,\n \"magicType\": magic_type,\n \"brush\": brush,\n }\n assert_equal(expected, remove_key_with_none_value(option.opts))\n\n\ndef test_toolbox_options_remove_none():\n option = ToolboxOpts(feature={})\n expected = {\n \"show\": True,\n \"orient\": \"horizontal\",\n \"itemSize\": 15,\n \"itemGap\": 10,\n \"left\": \"80%\",\n \"feature\": {},\n }\n assert_equal(expected, remove_key_with_none_value(option.opts))\n\n\ndef test_brush_options_remove_none():\n option = BrushOpts()\n expected = {\n \"brushMode\": \"single\",\n \"brushStyle\": {\n \"borderColor\": \"rgba(120,140,180,0.8)\",\n \"borderWidth\": 1,\n \"color\": \"rgba(120,140,180,0.3)\",\n },\n \"brushType\": \"rect\",\n \"removeOnClick\": True,\n \"throttleDelay\": 0,\n \"throttleType\": \"fixRate\",\n \"toolbox\": [\"rect\", \"polygon\", \"keep\", \"clear\"],\n \"transformable\": True,\n }\n assert_equal(expected, remove_key_with_none_value(option.opts))\n\n\ndef test_data_zoom_options_remove_none():\n option = DataZoomOpts()\n expected = {\n \"end\": 80,\n \"filterMode\": \"filter\",\n \"orient\": \"horizontal\",\n \"realtime\": True,\n \"show\": True,\n \"start\": 20,\n \"type\": \"slider\",\n \"zoomLock\": False,\n }\n assert_equal(expected, remove_key_with_none_value(option.opts))\n\n\ndef test_legend_options_remove_none():\n option = LegendOpts()\n expected = {\n \"show\": True,\n \"padding\": 5,\n \"itemGap\": 10,\n \"itemWidth\": 25,\n \"itemHeight\": 14,\n }\n assert_equal(expected, remove_key_with_none_value(option.opts))\n\n\ndef test_visual_map_options_remove_none():\n option = VisualMapOpts()\n expected = {\n \"calculable\": True,\n \"inRange\": {\"color\": [\"#50a3ba\", \"#eac763\", \"#d94e5d\"]},\n \"itemHeight\": 140,\n \"itemWidth\": 20,\n \"max\": 100,\n \"min\": 0,\n \"orient\": \"vertical\",\n \"show\": True,\n \"showLabel\": True,\n \"inverse\": False,\n \"splitNumber\": 5,\n \"type\": \"continuous\",\n \"borderWidth\": 0,\n }\n assert_equal(expected, remove_key_with_none_value(option.opts))\n\n\ndef test_tool_tip_options_remove_none():\n option = TooltipOpts(textstyle_opts=None)\n expected = {\n \"alwaysShowContent\": False,\n \"axisPointer\": {\"type\": \"line\"},\n \"borderWidth\": 0,\n \"hideDelay\": 100,\n \"padding\": 5,\n \"show\": True,\n \"showContent\": True,\n \"showDelay\": 0,\n \"trigger\": \"item\",\n \"triggerOn\": \"mousemove|click\",\n }\n assert_equal(expected, remove_key_with_none_value(option.opts))\n","sub_path":"test/test_global_options.py","file_name":"test_global_options.py","file_ext":"py","file_size_in_byte":4639,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"474658209","text":"import os\nfrom Configs import ConfigBase\n\n# define different classes per environment\n\n\nclass DEFAULT(ConfigBase):\n # commit every n files\n COMMIT_INTERVAL = 100\n\n SCRIPT_DIR = os.path.dirname(\n os.path.realpath(os.path.join(os.getcwd(), os.path.expanduser(__file__)))\n )\n\n DATA_BASE_DIR = os.path.join(SCRIPT_DIR, \"dataset/\")\n DATA_DIRS = [\n \"biorxiv_medrxiv\",\n \"comm_use_subset\",\n \"noncomm_use_subset\",\n \"custom_license\",\n ]\n METADATA_FILE = os.path.join(DATA_BASE_DIR, \"metadata.csv\")\n\n # Override label names\n JSON2GRAPH_LABELOVERRIDE = {\n \"authors\": \"Author\",\n }\n\n JSON2GRAPH_PROPOVERRIDE = {\n \"Doi\": {\"DOI\": \"id\"},\n \"Arxiv\": {\"arXiv\": \"id\"},\n \"Pmcid\": {\"PMCID\": \"id\"},\n \"Pmid\": {\"PMID\": \"id\"},\n }\n\n JSON2GRAPH_GENERATED_HASH_ID_ATTR_NAME = \"_hash_id\"\n # Define for which labels and how a hash id attr should be generated\n JSON2GRAPH_GENERATED_HASH_IDS = {\n \"Abstract\": [\"text\"], # Generate an id based on the property \"text\"\n \"Affiliation\": \"AllAttributes\", # Generate an id based all properties\n \"Author\": \"AllAttributes\",\n \"Back_matter\": \"AllAttributes\",\n \"Bibref\": \"AllAttributes\",\n \"Bib_entries\": \"AllInnerContent\", # Generate an id based all attr and childrens attr\n \"Body_text\": \"AllAttributes\",\n \"Cite_spans\": \"AllInnerContent\",\n \"Figref\": \"AllAttributes\",\n \"Location\": \"AllAttributes\",\n \"Metadata\": \"AllInnerContent\",\n \"Other_ids\": \"AllInnerContent\",\n \"Ref_entries\": \"AllInnerContent\",\n \"Ref_spans\": \"AllInnerContent\",\n \"Tabref\": \"AllAttributes\",\n }\n\n # Define which properties can be taken as primary key for specific labels\n # {\"label\":\"attribute-that-works-as-id\"}\n JSON2GRAPH_ID_ATTR = {\n \"Arxiv\": \"id\",\n \"Doi\": \"id\",\n \"Paper\": \"paper_id\",\n \"Pmcid\": \"id\",\n \"Pmid\": \"id\",\n }\n JSON2GRAPH_CONCAT_LIST_ATTR = {\"middle\": \" \"}\n JSON2GRAPH_COLLECTION_NODE_LABEL = \"CollectionHub\"\n\n\n# All following config classes inherit from DEFAULT\nclass PRODUCTION(DEFAULT):\n pass\n\n\nclass DEVELOPMENT(DEFAULT):\n COMMIT_INTERVAL = 2\n DATA_BASE_DIR = os.path.join(DEFAULT.SCRIPT_DIR, \"testdataset/\")\n DATA_DIRS = [\n \"test\",\n ]\n METADATA_FILE = os.path.join(DATA_BASE_DIR, \"metadata.csv\")\n","sub_path":"dataloader/config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":2389,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"36115775","text":"# Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserve.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport cv2\nimport paddle\nimport os.path\nfrom .base_dataset import BaseDataset, get_params, get_transform\nfrom .image_folder import make_dataset\n\nfrom .builder import DATASETS\nfrom .transforms.builder import build_transforms\n\n\n@DATASETS.register()\nclass PairedDataset(BaseDataset):\n \"\"\"A dataset class for paired image dataset.\n \"\"\"\n def __init__(self, cfg):\n \"\"\"Initialize this dataset class.\n\n Args:\n cfg (dict): configs of datasets.\n \"\"\"\n BaseDataset.__init__(self, cfg)\n self.dir_AB = os.path.join(cfg.dataroot,\n cfg.phase) # get the image directory\n self.AB_paths = sorted(make_dataset(\n self.dir_AB, cfg.max_dataset_size)) # get image paths\n\n self.input_nc = self.cfg.output_nc if self.cfg.direction == 'BtoA' else self.cfg.input_nc\n self.output_nc = self.cfg.input_nc if self.cfg.direction == 'BtoA' else self.cfg.output_nc\n self.transforms = build_transforms(cfg.transforms)\n\n def __getitem__(self, index):\n \"\"\"Return a data point and its metadata information.\n\n Parameters:\n index - - a random integer for data indexing\n\n Returns a dictionary that contains A, B, A_paths and B_paths\n A (tensor) - - an image in the input domain\n B (tensor) - - its corresponding image in the target domain\n A_paths (str) - - image paths\n B_paths (str) - - image paths (same as A_paths)\n \"\"\"\n # read a image given a random integer index\n AB_path = self.AB_paths[index]\n AB = cv2.cvtColor(cv2.imread(AB_path), cv2.COLOR_BGR2RGB)\n\n # split AB image into A and B\n h, w = AB.shape[:2]\n # w, h = AB.size\n w2 = int(w / 2)\n\n A = AB[:h, :w2, :]\n B = AB[:h, w2:, :]\n\n # apply the same transform to both A and B\n A, B = self.transforms((A, B))\n\n return {'A': A, 'B': B, 'A_paths': AB_path, 'B_paths': AB_path}\n\n def __len__(self):\n \"\"\"Return the total number of images in the dataset.\"\"\"\n return len(self.AB_paths)\n","sub_path":"ppgan/datasets/paired_dataset.py","file_name":"paired_dataset.py","file_ext":"py","file_size_in_byte":2724,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"590106962","text":"#!/usr/bin/env python\n\"\"\"\n% Create cell width for this mesh on a regular latitude-longitude grid.\n% Outputs:\n% cellWidth - m x n array, entries are desired cell width in km\n% lon - longitude, vector of length m, with entries between -180 and 180, degrees\n% lat - latitude, vector of length n, with entries between -90 and 90, degrees\n\"\"\"\nimport numpy as np\nimport jigsaw_to_MPAS.mesh_definition_tools as mdt\n\n\ndef cellWidthVsLatLon():\n\n ddeg = 0.1\n\n lat = np.arange(-90, 90.01, ddeg)\n lon = np.arange(-180, 180.01, ddeg)\n\n cellWidthSouth = 15 * np.ones(lat.size)\n cellWidthNorth = 60 * np.ones(lat.size)\n latTransition = -30\n latWidthTransition = 5\n\n cellWidthVsLat = mdt.mergeCellWidthVsLat(\n lat,\n cellWidthSouth,\n cellWidthNorth,\n latTransition,\n latWidthTransition)\n\n cellWidth = np.ones((lat.size, lon.size))\n for i in range(lon.size):\n cellWidth[:, i] = cellWidthVsLat\n\n #print 'cellWidthSouth', cellWidthSouth\n #print 'cellWidthNorth', cellWidthNorth\n #print 'cellWidthVsLat', cellWidthVsLat\n\n return cellWidth, lon, lat\n","sub_path":"testing_and_setup/compass/ocean/global_ocean/SOQU60to15/init/define_base_mesh.py","file_name":"define_base_mesh.py","file_ext":"py","file_size_in_byte":1121,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"185522591","text":"import numpy as np\nimport typing\nfrom typing import Any, Tuple\nimport tensorflow as tf\nfrom tensorflow.keras.layers.experimental import preprocessing\nimport tensorflow_text as tf_text\nfrom src.interpreter.attention_auto_encoder.shape_checker import *\n\n\n\n\nclass Encoder(tf.keras.layers.Layer):\n def __init__(self, input_vocab_size, embedding_dim, enc_units):\n super(Encoder, self).__init__()\n self.enc_units = enc_units\n self.input_vocab_size = input_vocab_size\n\n # The embedding layer converts tokens to vectors\n self.embedding = tf.keras.layers.Embedding(self.input_vocab_size,\n embedding_dim)\n\n # The GRU RNN layer processes those vectors sequentially.\n self.gru = tf.keras.layers.GRU(self.enc_units,\n # Return the sequence and state\n return_sequences=True,\n return_state=True,\n recurrent_initializer='glorot_uniform')\n\n def call(self, tokens, state=None):\n shape_checker = ShapeChecker()\n shape_checker(tokens, ('batch', 's'))\n\n # 2. The embedding layer looks up the embedding for each token.\n vectors = self.embedding(tokens)\n shape_checker(vectors, ('batch', 's', 'embed_dim'))\n\n # 3. The GRU processes the embedding sequence.\n # output shape: (batch, s, enc_units)\n # state shape: (batch, enc_units)\n output, state = self.gru(vectors, initial_state=state)\n shape_checker(output, ('batch', 's', 'enc_units'))\n shape_checker(state, ('batch', 'enc_units'))\n\n # 4. Returns the new sequence and its state.\n return output, state\n\nclass BahdanauAttention(tf.keras.layers.Layer):\n def __init__(self, units):\n super().__init__()\n # For Eqn. (4), the Bahdanau attention\n self.W1 = tf.keras.layers.Dense(units, use_bias=False)\n self.W2 = tf.keras.layers.Dense(units, use_bias=False)\n\n self.attention = tf.keras.layers.AdditiveAttention()\n\n def call(self, query, value, mask):\n shape_checker = ShapeChecker()\n shape_checker(query, ('batch', 't', 'query_units'))\n shape_checker(value, ('batch', 's', 'value_units'))\n shape_checker(mask, ('batch', 's'))\n\n # From Eqn. (4), `W1@ht`.\n w1_query = self.W1(query)\n shape_checker(w1_query, ('batch', 't', 'attn_units'))\n\n # From Eqn. (4), `W2@hs`.\n w2_key = self.W2(value)\n shape_checker(w2_key, ('batch', 's', 'attn_units'))\n\n query_mask = tf.ones(tf.shape(query)[:-1], dtype=bool)\n value_mask = mask\n\n context_vector, attention_weights = self.attention(\n inputs = [w1_query, value, w2_key],\n mask=[query_mask, value_mask],\n return_attention_scores = True,\n )\n shape_checker(context_vector, ('batch', 't', 'value_units'))\n shape_checker(attention_weights, ('batch', 't', 's'))\n\n return context_vector, attention_weights\n\nclass Decoder(tf.keras.layers.Layer):\n def __init__(self, output_vocab_size, embedding_dim, dec_units):\n super(Decoder, self).__init__()\n self.dec_units = dec_units\n self.output_vocab_size = output_vocab_size\n self.embedding_dim = embedding_dim\n\n # For Step 1. The embedding layer convets token IDs to vectors\n self.embedding = tf.keras.layers.Embedding(self.output_vocab_size,\n embedding_dim)\n\n # For Step 2. The RNN keeps track of what's been generated so far.\n self.gru = tf.keras.layers.GRU(self.dec_units,\n return_sequences=True,\n return_state=True,\n recurrent_initializer='glorot_uniform')\n\n # For step 3. The RNN output will be the query for the attention layer.\n self.attention = BahdanauAttention(self.dec_units)\n\n # For step 4. Eqn. (3): converting `ct` to `at`\n self.Wc = tf.keras.layers.Dense(dec_units, activation=tf.math.tanh,\n use_bias=False)\n\n # For step 5. This fully connected layer produces the logits for each\n # output token.\n self.fc = tf.keras.layers.Dense(self.output_vocab_size)\n\n\n\n\n def call(self,\n inputs,\n state=None):\n shape_checker = ShapeChecker()\n shape_checker(inputs.new_tokens, ('batch', 't'))\n shape_checker(inputs.enc_output, ('batch', 's', 'enc_units'))\n shape_checker(inputs.mask, ('batch', 's'))\n\n if state is not None:\n shape_checker(state, ('batch', 'dec_units'))\n\n # Step 1. Lookup the embeddings\n vectors = self.embedding(inputs.new_tokens)\n shape_checker(vectors, ('batch', 't', 'embedding_dim'))\n\n # Step 2. Process one step with the RNN\n rnn_output, state = self.gru(vectors, initial_state=state)\n\n shape_checker(rnn_output, ('batch', 't', 'dec_units'))\n shape_checker(state, ('batch', 'dec_units'))\n\n # Step 3. Use the RNN output as the query for the attention over the\n # encoder output.\n context_vector, attention_weights = self.attention(\n query=rnn_output, value=inputs.enc_output, mask=inputs.mask)\n shape_checker(context_vector, ('batch', 't', 'dec_units'))\n shape_checker(attention_weights, ('batch', 't', 's'))\n\n # Step 4. Eqn. (3): Join the context_vector and rnn_output\n # [ct; ht] shape: (batch t, value_units + query_units)\n context_and_rnn_output = tf.concat([context_vector, rnn_output], axis=-1)\n\n # Step 4. Eqn. (3): `at = tanh(Wc@[ct; ht])`\n attention_vector = self.Wc(context_and_rnn_output)\n shape_checker(attention_vector, ('batch', 't', 'dec_units'))\n\n # Step 5. Generate logit predictions:\n logits = self.fc(attention_vector)\n shape_checker(logits, ('batch', 't', 'output_vocab_size'))\n\n return DecoderOutput(logits, attention_weights), state\n\n\nclass DecoderInput(typing.NamedTuple):\n new_tokens: Any\n enc_output: Any\n mask: Any\n\nclass DecoderOutput(typing.NamedTuple):\n logits: Any\n attention_weights: Any\n\n","sub_path":"src/interpreter/attention_auto_encoder/autoencoder.py","file_name":"autoencoder.py","file_ext":"py","file_size_in_byte":5926,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"115093817","text":"import unittest\nfrom cataloger import reverse_geocode_wgs84_boundingbox\n\n\nclass TestGeocoder(unittest.TestCase):\n \"\"\"\n unittests for the geocoder function\n \"\"\"\n def setUp(self):\n self.pg_conn_str = \"postgres://james:MopMetal3@localhost:5432/mapcatalogue\"\n self.us_bbox = (-74.66163264559283, 39.650417182507944, -72.0006154054558, 41.612140592278074)\n\n def test_geocoder_returns_list(self):\n self.assertIsInstance(\n reverse_geocode_wgs84_boundingbox(self.pg_conn_str, self.us_bbox),\n list,\n 'Should be a list'\n )\n\n def test_geocoder_contains_dict(self):\n self.assertIsInstance(\n reverse_geocode_wgs84_boundingbox(self.pg_conn_str, self.us_bbox)[0],\n dict,\n 'List should contain dictionaries'\n )\n\n def test_geocoder_equals(self):\n bboxes = {\n 1: [\n (-84.7, 28.5, -66.8, 42.6),\n [{'country': 'Canada', 'continent': 'North America'},\n {'country': 'United States', 'continent': 'North America'}]\n ],\n 2: [\n (-4.080, 55.572, -2.228, 57.250),\n [{'country': 'England', 'continent': 'Europe'},\n {'country': 'Scotland', 'continent': 'Europe'}]\n ],\n }\n\n for bb in bboxes:\n bbox_coords = bboxes[bb][0]\n geographies = bboxes[bb][1]\n\n self.assertEqual(\n reverse_geocode_wgs84_boundingbox(self.pg_conn_str, bboxes[bb][0]),\n bboxes[bb][1],\n 'Geographies for BBox should equal: {0}'.format(bboxes[bb][1])\n )\n\n\nif __name__ == \"__main__\":\n unittest.main()\n\n","sub_path":"tests.py","file_name":"tests.py","file_ext":"py","file_size_in_byte":1721,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"546862590","text":"import importlib\nimport os\nimport re\nimport sys\nfrom io import TextIOWrapper\nfrom logging import getLogger\nfrom pathlib import Path\n\nimport pkg_resources # pylint: disable=C041\nimport pytest\n\nimport scrapli\nfrom scrapli.exceptions import TransportPluginError\nfrom scrapli.helper import (\n _find_transport_plugin,\n _textfsm_get_template,\n attach_duplicate_log_filter,\n genie_parse,\n get_prompt_pattern,\n resolve_file,\n strip_ansi,\n textfsm_parse,\n ttp_parse,\n)\n\nTEST_DATA_DIR = f\"{Path(scrapli.__file__).parents[1]}/tests/test_data\"\n\nIOS_ARP = \"\"\"Protocol Address Age (min) Hardware Addr Type Interface\nInternet 172.31.254.1 - 0000.0c07.acfe ARPA Vlan254\nInternet 172.31.254.2 - c800.84b2.e9c2 ARPA Vlan254\n\"\"\"\n\n\ndef test_get_prompt_pattern_class_pattern():\n class_pattern = \"^averygoodpattern$\"\n result = get_prompt_pattern(\"\", class_pattern)\n assert result == re.compile(b\"^averygoodpattern$\", re.IGNORECASE | re.MULTILINE)\n\n\ndef test_get_prompt_pattern_class_pattern_no_line_start_end_markers():\n class_pattern = \"averygoodpattern\"\n result = get_prompt_pattern(class_pattern, \"\")\n assert result == re.compile(b\"averygoodpattern\")\n\n\ndef test_get_prompt_pattern_arg_pattern():\n class_pattern = \"averygoodpattern\"\n result = get_prompt_pattern(\"^awesomepattern$\", class_pattern)\n assert result == re.compile(b\"^awesomepattern$\", re.IGNORECASE | re.MULTILINE)\n\n\ndef test_get_prompt_pattern_arg_string():\n class_pattern = \"averygoodpattern\"\n result = get_prompt_pattern(\"awesomepattern\", class_pattern)\n assert result == re.compile(b\"awesomepattern\")\n\n\ndef test_get_prompt_pattern_arg_bytes():\n class_pattern = \"averygoodpattern\"\n result = get_prompt_pattern(b\"awesomepattern\", class_pattern)\n assert result == re.compile(b\"awesomepattern\")\n\n\ndef test__strip_ansi():\n output = b\"[admin@CoolDevice.Sea1: \\x1b[1m/\\x1b[0;0m]$\"\n output = strip_ansi(output)\n assert output == b\"[admin@CoolDevice.Sea1: /]$\"\n\n\n@pytest.mark.skipif(sys.platform.startswith(\"win\"), reason=\"not supporting textfsm on windows\")\ndef test__textfsm_get_template_valid_template():\n template = _textfsm_get_template(\"cisco_nxos\", \"show ip arp\")\n template_dir = pkg_resources.resource_filename(\"ntc_templates\", \"templates\")\n assert isinstance(template, TextIOWrapper)\n assert template.name == f\"{template_dir}/cisco_nxos_show_ip_arp.textfsm\"\n\n\n@pytest.mark.skipif(sys.platform.startswith(\"win\"), reason=\"not supporting textfsm on windows\")\ndef test__textfsm_get_template_invalid_template():\n template = _textfsm_get_template(\"cisco_nxos\", \"show racecar\")\n assert not template\n\n\n@pytest.mark.skipif(sys.platform.startswith(\"win\"), reason=\"not supporting textfsm on windows\")\n@pytest.mark.parametrize(\n \"parse_type\",\n [\n (\n False,\n [\"Internet\", \"172.31.254.1\", \"-\", \"0000.0c07.acfe\", \"ARPA\", \"Vlan254\"],\n ),\n (\n True,\n {\n \"protocol\": \"Internet\",\n \"address\": \"172.31.254.1\",\n \"age\": \"-\",\n \"mac\": \"0000.0c07.acfe\",\n \"type\": \"ARPA\",\n \"interface\": \"Vlan254\",\n },\n ),\n ],\n ids=[\"to_dict_false\", \"to_dict_true\"],\n)\ndef test_textfsm_parse_success(parse_type):\n to_dict = parse_type[0]\n expected_result = parse_type[1]\n template = _textfsm_get_template(\"cisco_ios\", \"show ip arp\")\n result = textfsm_parse(template, IOS_ARP, to_dict=to_dict)\n assert isinstance(result, list)\n assert result[0] == expected_result\n\n\n@pytest.mark.skipif(sys.platform.startswith(\"win\"), reason=\"not supporting textfsm on windows\")\n@pytest.mark.parametrize(\n \"parse_type\",\n [\n (\n False,\n [\"Internet\", \"172.31.254.1\", \"-\", \"0000.0c07.acfe\", \"ARPA\", \"Vlan254\"],\n ),\n (\n True,\n {\n \"protocol\": \"Internet\",\n \"address\": \"172.31.254.1\",\n \"age\": \"-\",\n \"mac\": \"0000.0c07.acfe\",\n \"type\": \"ARPA\",\n \"interface\": \"Vlan254\",\n },\n ),\n ],\n ids=[\"to_dict_false\", \"to_dict_true\"],\n)\ndef test_textfsm_parse_success_string_path(parse_type):\n to_dict = parse_type[0]\n expected_result = parse_type[1]\n template = _textfsm_get_template(\"cisco_ios\", \"show ip arp\")\n result = textfsm_parse(template.name, IOS_ARP, to_dict=to_dict)\n assert isinstance(result, list)\n assert result[0] == expected_result\n\n\n@pytest.mark.skipif(sys.platform.startswith(\"win\"), reason=\"not supporting textfsm on windows\")\ndef test_textfsm_parse_failure():\n template = _textfsm_get_template(\"cisco_ios\", \"show ip arp\")\n result = textfsm_parse(template, \"not really arp data\")\n assert result == []\n\n\n@pytest.mark.skipif(\n sys.version_info.minor > 8, reason=\"genie not currently available for python 3.9\"\n)\n@pytest.mark.skipif(sys.platform.startswith(\"win\"), reason=\"not supporting genie on windows\")\ndef test_genie_parse_success():\n result = genie_parse(\"iosxe\", \"show ip arp\", IOS_ARP)\n assert isinstance(result, dict)\n assert (\n result[\"interfaces\"][\"Vlan254\"][\"ipv4\"][\"neighbors\"][\"172.31.254.1\"][\"ip\"] == \"172.31.254.1\"\n )\n\n\n@pytest.mark.skipif(\n sys.version_info.minor > 8, reason=\"genie not currently available for python 3.9\"\n)\n@pytest.mark.skipif(sys.platform.startswith(\"win\"), reason=\"not supporting genie on windows\")\ndef test_genie_parse_failure():\n result = genie_parse(\"iosxe\", \"show ip arp\", \"not really arp data\")\n assert result == []\n # w/out killing this module pyfakefs explodes. dont remember why/how i found that out...\n del sys.modules[\"pyats.configuration\"]\n\n\ndef test_ttp_parse():\n # example data lifted straight out of ttp docs\n data_to_parse = \"\"\"\n interface Loopback0\n description Router-id-loopback\n ip address 192.168.0.113/24\n !\n interface Vlan778\n description CPE_Acces_Vlan\n ip address 2002::fd37/124\n ip vrf CPE1\n !\n \"\"\"\n\n ttp_template = \"\"\"\n interface {{ interface }}\n ip address {{ ip }}/{{ mask }}\n description {{ description }}\n ip vrf {{ vrf }}\n \"\"\"\n\n expected = [\n [\n {\n \"ip\": \"192.168.0.113\",\n \"mask\": \"24\",\n \"description\": \"Router-id-loopback\",\n \"interface\": \"Loopback0\",\n },\n {\n \"vrf\": \"CPE1\",\n \"ip\": \"2002::fd37\",\n \"mask\": \"124\",\n \"description\": \"CPE_Acces_Vlan\",\n \"interface\": \"Vlan778\",\n },\n ]\n ]\n result = ttp_parse(template=ttp_template, output=data_to_parse)\n assert result == expected\n\n\ndef test_ttp_parse_invalid_template():\n result = ttp_parse(template=None, output=\"blah\")\n assert result == []\n\n\ndef test_ttp_parse_failed_to_parse():\n result = ttp_parse(template=\"mytemplateisneat\", output=\"blah\")\n assert result == [{}]\n\n\n@pytest.mark.skipif(\n sys.platform.startswith(\"win\"), reason=\"not dealing with windows path things in tests\"\n)\ndef test_resolve_file():\n resolved_file = resolve_file(file=f\"{TEST_DATA_DIR}/files/_ssh_config\")\n assert resolved_file == f\"{TEST_DATA_DIR}/files/_ssh_config\"\n\n\n@pytest.mark.skipif(\n sys.platform.startswith(\"win\"), reason=\"not dealing with windows path things in tests\"\n)\ndef test_resolve_file_expanduser(fs):\n fs.add_real_file(\n source_path=f\"{TEST_DATA_DIR}/files/_ssh_config\",\n target_path=f\"{os.path.expanduser('~')}/myneatfile\",\n )\n resolved_file = resolve_file(file=f\"~/myneatfile\")\n assert resolved_file == f\"{os.path.expanduser('~')}/myneatfile\"\n\n\n@pytest.mark.skipif(\n sys.platform.startswith(\"win\"), reason=\"not dealing with windows path things in tests\"\n)\ndef test_resolve_file_failure():\n with pytest.raises(ValueError) as exc:\n resolve_file(file=f\"~/myneatfile\")\n assert str(exc.value) == \"File path `~/myneatfile` could not be resolved\"\n\n\ndef test_attach_duplicate_log_filter():\n dummy_logger = getLogger(\"this_is_a_dumb_test_log\")\n assert dummy_logger.filters == []\n attach_duplicate_log_filter(logger=dummy_logger)\n # simple assert to confirm that we got the dup filter attached to the new logger\n assert dummy_logger.filters[0].__class__.__name__ == \"DuplicateFilter\"\n\n\ndef test__find_transport_plugin_failure():\n with pytest.raises(ModuleNotFoundError) as exc:\n _find_transport_plugin(transport=\"blah\")\n assert (\n str(exc.value)\n == \"\\n***** Module 'scrapli_blah' not found! ************************************************\\nTo resolve this issue, ensure you are referencing a valid transport plugin. Transport plugins should be named similar to `scrapli_paramiko` or `scrapli_ssh2`, and can be selected by passing simply `paramiko` or `ssh2` into the scrapli driver. You can install most plugins with pip: `pip install scrapli-ssh2` for example.\\n***** Module 'scrapli_blah' not found! ************************************************\"\n )\n\n\ndef test___find_transport_plugin_module_failed_to_load(monkeypatch):\n from scrapli_ssh2 import transport\n\n monkeypatch.setattr(transport, \"Transport\", None)\n with pytest.raises(TransportPluginError) as exc:\n _find_transport_plugin(transport=\"ssh2\")\n assert (\n str(exc.value)\n == \"Failed to load transport plugin `ssh2` transport class or required arguments\"\n )\n\n\ndef test_textfsm_get_template_no_textfsm(monkeypatch):\n def mock_import_module(name, package):\n raise ModuleNotFoundError\n\n monkeypatch.setattr(importlib, \"import_module\", mock_import_module)\n\n with pytest.warns(UserWarning) as warning_msg:\n _textfsm_get_template(platform=\"blah\", command=\"blah\")\n assert (\n str(warning_msg._list[0].message)\n == \"\\n***** Module 'None' not installed! ****************************************************\\nTo resolve this issue, install 'None'. You can do this in one of the following ways:\\n1: 'pip install -r requirements-textfsm.txt'\\n2: 'pip install scrapli[textfsm]'\\n***** Module 'None' not installed! ****************************************************\"\n )\n\n\ndef test_genie_parse_no_genie(monkeypatch):\n def mock_import_module(name, package):\n raise ModuleNotFoundError\n\n monkeypatch.setattr(importlib, \"import_module\", mock_import_module)\n\n with pytest.warns(UserWarning) as warning_msg:\n genie_parse(platform=\"blah\", command=\"blah\", output=\"blah\")\n assert (\n str(warning_msg._list[0].message)\n == \"\\n***** Module 'None' not installed! ****************************************************\\nTo resolve this issue, install 'None'. You can do this in one of the following ways:\\n1: 'pip install -r requirements-genie.txt'\\n2: 'pip install scrapli[genie]'\\n***** Module 'None' not installed! ****************************************************\"\n )\n\n\ndef test_ttp_parse_no_ttp(monkeypatch):\n def mock_import_module(name):\n raise ModuleNotFoundError\n\n monkeypatch.setattr(importlib, \"import_module\", mock_import_module)\n\n with pytest.warns(UserWarning) as warning_msg:\n ttp_parse(template=\"blah\", output=\"blah\")\n assert (\n str(warning_msg._list[0].message)\n == \"\\n***** Module 'None' not installed! ****************************************************\\nTo resolve this issue, install 'None'. You can do this in one of the following ways:\\n1: 'pip install -r requirements-ttp.txt'\\n2: 'pip install scrapli[ttp]'\\n***** Module 'None' not installed! ****************************************************\"\n )\n","sub_path":"tests/unit/test_helper.py","file_name":"test_helper.py","file_ext":"py","file_size_in_byte":11673,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"503169630","text":"# encoding: utf-8\n\"\"\"\nDefine logging level names\n--------------------------\n\n.. rst-class:: html-toggle\n\nIdentification\n--------------\n\nCreated on Sep 28, 2014\n\nDefine names for the logging levels\n\n@author: Jonathan Gossage\n\n@copyright: © 2015 Jonathan Gossage. All rights reserved.\n\n@license: Apache 2\n\n@contact: jgossage at gmail.com\n@deffield updated: Updated\n\n\"\"\"\n\nimport logging\n\n# Define our symbolic names for the Python logging levels\nCRITICAL = 'critical'\n\"\"\"Our name for critical logging level\"\"\"\nERROR = 'error'\n\"\"\"Our name for error logging level\"\"\"\nWARNING = 'warning'\n\"\"\"Our name for warning logging level\"\"\"\nINFO = 'info'\n\"\"\"Our name for info logging level\"\"\"\nDEBUG = 'debug'\n\"\"\"Our name for debug logging level\"\"\"\n\nLLEVELS = {CRITICAL: logging.CRITICAL,\n ERROR: logging.ERROR,\n WARNING: logging.WARNING,\n INFO: logging.INFO,\n DEBUG: logging.DEBUG}\n\"\"\"\nMap from our symbolic names to the Python logging system levels\n\"\"\"\n\nRLEVELS = {logging.CRITICAL: CRITICAL,\n logging.ERROR: ERROR,\n logging.WARNING: WARNING,\n logging.INFO: INFO,\n logging.DEBUG: DEBUG}\n\"\"\"\nMap from the Python logging system log level codes to our symbolic names\n\"\"\"\n","sub_path":"Logging/src/logutils/loglevels.py","file_name":"loglevels.py","file_ext":"py","file_size_in_byte":1246,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"647599113","text":"# Copyright (C) 2007 Matthew Neeley, Isaac Storch\n#\n# This program is free software: you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation, either version 2 of the License, or\n# (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with this program. If not, see .\n\n\"\"\"\n### BEGIN NODE INFO\n[info]\nname = XY Attenuator Server\nversion = 1.0\ndescription = \n\n[startup]\ncmdline = %PYTHON% %FILE%\ntimeout = 20\n\n[shutdown]\nmessage = 987654321\ntimeout = 20\n### END NODE INFO\n\"\"\"\n\nfrom labrad.gpib import GPIBManagedServer\nfrom labrad.server import setting\nfrom labrad import types as T\nfrom twisted.internet.defer import inlineCallbacks, returnValue\nfrom numpy import floor\n\nclass XYAttenuatorServer(GPIBManagedServer):\n name = 'XY Attenuator Server'\n deviceName = 'Hewlett-Packard 11713A'\n deviceIdentFunc = 'identify_device'\n\n @inlineCallbacks\n def setAtten(self, c, val, commands):\n \"\"\"Helper method to set either the X or Y attenuation.\n\n This method looks up the desired attenuation and gpib\n command in the provided dictionary.\n \"\"\"\n dev = self.selectedDevice(c)\n if val not in commands.keys():\n raise Exception('Invalid attenuation value.')\n\n yield dev.write(commands[val])\n returnValue(T.Value(val, 'dB'))\n\n # settings\n\n @setting(1000, server='s', address='s', idn='s')\n def identify_device(self, c, server, address, idn=None):\n devices = [('ADR GPIB Bus', 'GPIB0::28'),\n ('DR GPIB Bus', 'GPIB0::28'),\n ('Twins IBCL GPIB Bus', 'ProbeStation GPIB-422CT::28'),\n ('T1000 IBCL GPIB Bus', 'T1000 GPIB-422CT::28')]\n if (server, address) in devices:\n return self.deviceName\n\n @setting(10000, \"X Atten\", data=['v[dB]'], returns=['v[dB]'])\n def x_atten(self, c, data):\n \"\"\"Set the X attenuation.\n\n Allowed values of are 0, 1, 2, ... 11 dB.\n \"\"\"\n val = int(data.value)\n return self.setAtten(c, val, XattnDict)\n\n @setting(10001, \"Y Atten\", data=['v[dB]'], returns=['v[dB]'])\n def y_atten(self, c, data):\n \"\"\"Set the Y attenuation.\n\n Allowed values are 0, 10, 20, ... 70 dB.\n \"\"\"\n val = int(data.value)\n return self.setAtten(c, val, YattnDict)\n\n @setting(10002, \"Total Atten\", data=['v[dB]'], returns=['v[dB]v[dB]'])\n def total_atten(self, c, data):\n \"\"\"Set the total attenuation on X and Y channels (connected in series).\n\n Allowed values of are 0, 1, 2, ... 79 dB.\n Note: use x_atten and y_atten to go to 80 and 81 dB\n \"\"\"\n val = int(data.value)\n x = yield self.setAtten(c, val%10, XattnDict)\n y = yield self.setAtten(c, floor(val/10)*10, YattnDict)\n returnValue((x,y))\n\n# commands for X attenuation\nXattnDict = {\n 0: 'B1234',\n 1: 'A1B234',\n 2: 'A2B134',\n 3: 'A12B34',\n 4: 'A3B124',\n 5: 'A13B24',\n 6: 'A23B14',\n 7: 'A123B4',\n 8: 'A34B12',\n 9: 'A134B2',\n 10: 'A234B1',\n 11: 'A1234'\n}\n\n# commands for Y attenuation\nYattnDict = {\n 0: 'B5678',\n 10: 'A5B678',\n 20: 'A6B578',\n 30: 'A56B78',\n 40: 'A7B568',\n 50: 'A57B68',\n 60: 'A67B58',\n 70: 'A567B8'\n}\n\n__server__ = XYAttenuatorServer()\n\nif __name__ == '__main__':\n from labrad import util\n util.runServer(__server__)\n","sub_path":"xyattenuators.py","file_name":"xyattenuators.py","file_ext":"py","file_size_in_byte":3765,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"431622620","text":"from pandas import *\nfrom batch_fitting_class import *\nimport matplotlib.backends.backend_pdf\n\n\n####################################################################\n# start with different substrates\n####################################################################\n\n# read from excel into a pandas dataframe\nss = read_excel(\"DMSP_dosage.xlsx\", \"substrate_forpy\")\n\n# times\ndtimes = array(ss['T'])\n\n# convert to numpy arrays\na1 = array(ss['A_2090'])\na1D = array(ss['A_2090_DMSP'])\na1G = array(ss['A_2090_Glycerol'])\na1P = array(ss['A_2090_Proprionate'])\na1b = array(ss['A_2090_D7'])\na1bD = array(ss['A_2090_D7_DMSP'])\na1bG = array(ss['A_2090_D7_Glycerol'])\na1bP = array(ss['A_2090_D7_Proprionate'])\na2 = array(ss['A_379'])\na2D = array(ss['A_379_DMSP'])\na2G = array(ss['A_379_Glycerol'])\na2P = array(ss['A_379_Proprionate'])\na2b = array(ss['A_379_D7'])\na2bD = array(ss['A_379_D7_DMSP'])\na2bG = array(ss['A_379_D7_Glycerol'])\na2bP = array(ss['A_379_D7_Proprionate'])\nb1 = array(ss['B_2090'])\nb1D = array(ss['B_2090_DMSP'])\nb1G = array(ss['B_2090_Glycerol'])\nb1P = array(ss['B_2090_Proprionate'])\nb1a = array(ss['B_2090_D7'])\nb1aD = array(ss['B_2090_D7_DMSP'])\nb1aG = array(ss['B_2090_D7_Glycerol'])\nb1aP = array(ss['B_2090_D7_Proprionate'])\nb2 = array(ss['B_379'])\nb2D = array(ss['B_379_DMSP'])\nb2G = array(ss['B_379_Glycerol'])\nb2P = array(ss['B_379_Proprionate'])\nb2a = array(ss['B_379_D7'])\nb2aD = array(ss['B_379_D7_DMSP'])\nb2aG = array(ss['B_379_D7_Glycerol'])\nb2aP = array(ss['B_379_D7_Proprionate'])\n\na1sd = array(ss['A_2090_sd'])\na1Dsd = array(ss['A_2090_DMSP_sd'])\na1Gsd = array(ss['A_2090_Glycerol_sd'])\na1Psd = array(ss['A_2090_Proprionate_sd'])\na1bsd = array(ss['A_2090_D7_sd'])\na1bDsd = array(ss['A_2090_D7_DMSP_sd'])\na1bGsd = array(ss['A_2090_D7_Glycerol_sd'])\na1bPsd = array(ss['A_2090_D7_Proprionate_sd'])\na2sd = array(ss['A_379_sd'])\na2Dsd = array(ss['A_379_DMSP_sd'])\na2Gsd = array(ss['A_379_Glycerol_sd'])\na2Psd = array(ss['A_379_Proprionate_sd'])\na2bsd = array(ss['A_379_D7_sd'])\na2bDsd = array(ss['A_379_D7_DMSP_sd'])\na2bGsd = array(ss['A_379_D7_Glycerol_sd'])\na2bPsd = array(ss['A_379_D7_Proprionate_sd'])\nb1sd = array(ss['B_2090_sd'])\nb1Dsd = array(ss['B_2090_DMSP_sd'])\nb1Gsd = array(ss['B_2090_Glycerol_sd'])\nb1Psd = array(ss['B_2090_Proprionate_sd'])\nb1asd = array(ss['B_2090_D7_sd'])\nb1aDsd = array(ss['B_2090_D7_DMSP_sd'])\nb1aGsd = array(ss['B_2090_D7_Glycerol_sd'])\nb1aPsd = array(ss['B_2090_D7_Proprionate_sd'])\nb2sd = array(ss['B_379_sd'])\nb2Dsd = array(ss['B_379_DMSP_sd'])\nb2Gsd = array(ss['B_379_Glycerol_sd'])\nb2Psd = array(ss['B_379_Proprionate_sd'])\nb2asd = array(ss['B_379_D7_sd'])\nb2aDsd = array(ss['B_379_D7_DMSP_sd'])\nb2aGsd = array(ss['B_379_D7_Glycerol_sd'])\nb2aPsd = array(ss['B_379_D7_Proprionate_sd'])\n\n# put in dictionaries\ncont_a2090 = {'htimes': dtimes, 'hms': a1, 'hss': a1sd}\ncontD_a2090 = {'htimes': dtimes, 'hms': a1D, 'hss': a1Dsd}\ncontG_a2090 = {'htimes': dtimes, 'hms': a1G, 'hss': a1Gsd}\ncontP_a2090 = {'htimes': dtimes, 'hms': a1P, 'hss': a1Psd}\n\ncont_b2090 = {'htimes': dtimes, 'hms': b1, 'hss': b1sd}\ncontD_b2090 = {'htimes': dtimes, 'hms': b1D, 'hss': b1Dsd}\ncontG_b2090 = {'htimes': dtimes, 'hms': b1G, 'hss': b1Gsd}\ncontP_b2090 = {'htimes': dtimes, 'hms': b1P, 'hss': b1Psd}\n\ninf_2090 = {'htimes': dtimes, 'vtimes': dtimes,\n 'hms': a1b, 'vms': b1a, 'hss': a1bsd, 'vss': b1asd}\ninfD_2090 = {'htimes': dtimes, 'vtimes': dtimes,\n 'hms': a1bD, 'vms': b1aD, 'hss': a1bDsd, 'vss': b1aDsd}\ninfG_2090 = {'htimes': dtimes, 'vtimes': dtimes,\n 'hms': a1bG, 'vms': b1aG, 'hss': a1bGsd, 'vss': b1aGsd}\ninfP_2090 = {'htimes': dtimes, 'vtimes': dtimes,\n 'hms': a1bP, 'vms': b1aP, 'hss': a1bPsd, 'vss': b1aPsd}\n\n# put in dictionaries\ncont_a379 = {'htimes': dtimes, 'hms': a2, 'hss': a2sd}\ncontD_a379 = {'htimes': dtimes, 'hms': a2D, 'hss': a2Dsd}\ncontG_a379 = {'htimes': dtimes, 'hms': a2G, 'hss': a2Gsd}\ncontP_a379 = {'htimes': dtimes, 'hms': a2P, 'hss': a2Psd}\n\ncont_b379 = {'htimes': dtimes, 'hms': b2, 'hss': b2sd}\ncontD_b379 = {'htimes': dtimes, 'hms': b2D, 'hss': b2Dsd}\ncontG_b379 = {'htimes': dtimes, 'hms': b2G, 'hss': b2Gsd}\ncontP_b379 = {'htimes': dtimes, 'hms': b2P, 'hss': b2Psd}\n\ninf_379 = {'htimes': dtimes, 'vtimes': dtimes,\n 'hms': a2b, 'vms': b2a, 'hss': a2bsd, 'vss': b2asd}\ninfD_379 = {'htimes': dtimes, 'vtimes': dtimes,\n 'hms': a2bD, 'vms': b2aD, 'hss': a2bDsd, 'vss': b2aDsd}\ninfG_379 = {'htimes': dtimes, 'vtimes': dtimes,\n 'hms': a2bG, 'vms': b2aG, 'hss': a2bGsd, 'vss': b2aGsd}\ninfP_379 = {'htimes': dtimes, 'vtimes': dtimes,\n 'hms': a2bP, 'vms': b2aP, 'hss': a2bPsd, 'vss': b2aPsd}\n\n####################################################################\n# now do DMSP dosage\n####################################################################\n\n# read excel sheet into a pandas dataframe\nss = read_excel(\"DMSP_dosage.xlsx\", \"doses_forpy\")\n\n# times\ndtimes = array(ss['T'])\n\n# convert data from pandas data frame to arrays\na0 = array(ss['A_379_0'])\na10 = array(ss['A_379_10'])\na100 = array(ss['A_379_100'])\na500 = array(ss['A_379_500'])\nab0 = array(ss['A_379_D7_0'])\nab10 = array(ss['A_379_D7_10'])\nab100 = array(ss['A_379_D7_100'])\nab500 = array(ss['A_379_D7_500'])\nb0 = array(ss['B_379_0'])\nb10 = array(ss['B_379_10'])\nb100 = array(ss['B_379_100'])\nb500 = array(ss['B_379_500'])\nba0 = array(ss['B_379_D7_0'])\nba10 = array(ss['B_379_D7_10'])\nba100 = array(ss['B_379_D7_100'])\nba500 = array(ss['B_379_D7_500'])\n\na0sd = array(ss['A_379_0_sd'])\na10sd = array(ss['A_379_10_sd'])\na100sd = array(ss['A_379_100_sd'])\na500sd = array(ss['A_379_500_sd'])\nab0sd = array(ss['A_379_D7_0_sd'])\nab10sd = array(ss['A_379_D7_10_sd'])\nab100sd = array(ss['A_379_D7_100_sd'])\nab500sd = array(ss['A_379_D7_500_sd'])\nb0sd = array(ss['B_379_0_sd'])\nb10sd = array(ss['B_379_10_sd'])\nb100sd = array(ss['B_379_100_sd'])\nb500sd = array(ss['B_379_500_sd'])\nba0sd = array(ss['B_379_D7_0_sd'])\nba10sd = array(ss['B_379_D7_10_sd'])\nba100sd = array(ss['B_379_D7_100_sd'])\nba500sd = array(ss['B_379_D7_500_sd'])\n\n# put in dictionaries to call the function\ncont_a0 = {'htimes': dtimes, 'hms': a0, 'hss': a0sd}\ncont_a10 = {'htimes': dtimes, 'hms': a10, 'hss': a10sd}\ncont_a100 = {'htimes': dtimes, 'hms': a100, 'hss': a100sd}\ncont_a500 = {'htimes': dtimes, 'hms': a500, 'hss': a500sd}\n\ncont_b0 = {'htimes': dtimes, 'hms': b0, 'hss': b0sd}\ncont_b10 = {'htimes': dtimes, 'hms': b10, 'hss': b10sd}\ncont_b100 = {'htimes': dtimes, 'hms': b100, 'hss': b100sd}\ncont_b500 = {'htimes': dtimes, 'hms': b500, 'hss': b500sd}\n\ninf_0 = {'htimes': dtimes, 'vtimes': dtimes,\n 'hms': ab0, 'vms': ba0, 'hss': ab0sd, 'vss': ba0sd}\ninf_10 = {'htimes': dtimes, 'vtimes': dtimes, 'hms': ab10,\n 'vms': ba10, 'hss': ab10sd, 'vss': ba10sd}\ninf_100 = {'htimes': dtimes, 'vtimes': dtimes, 'hms': ab100,\n 'vms': ba100, 'hss': ab100sd, 'vss': ba100sd}\ninf_500 = {'htimes': dtimes, 'vtimes': dtimes, 'hms': ab500,\n 'vms': ba500, 'hss': ab500sd, 'vss': ba500sd}\n\n####################################################################\n# now do data from mixed experiments\n####################################################################\n\n# read excel sheet into a pandas dataframe\nsv = read_excel(\"DMSP_dosage.xlsx\", \"vir_forpy\")\nsb = read_excel(\"DMSP_dosage.xlsx\", \"bac_forpy\")\noh = read_excel(\"DMSP_dosage.xlsx\", \"onehost_forpy\")\nth = read_excel(\"DMSP_dosage.xlsx\", \"twohost_forpy\")\n\n# convert to numpy arrays\n# times\ndtimes = array(sv['T'])\n\n# algae\na2090 = array(sv['A_2090'])\na2090v = array(sv['A_2090_V'])\na2090b = array(sb['A_2090_D7'])\na2090vb = array(oh['A_2090_D7_V'])\na379 = array(sb['A_379'])\na379b = array(sb['A_379_D7'])\na379vb = array(oh['A_379_D7_V'])\na2090a379 = array(th['A_2090_A_379'])\na2090a379v = array(th['A_2090_A_379_V'])\na2090a379b = array(th['A_2090_A_379_D7'])\na2090a379vb = array(th['A_2090_A_379_D7_V'])\n\n# bacteria\nb2090b = array(sb['B_A_2090_D7'])\nb2090vb = array(oh['B_A_2090_D7_V'])\nb = array(sb['B'])\nb379b = array(sb['B_A_379_D7'])\nb379vb = array(oh['B_A_379_D7_V'])\nba2090a379b = array(th['B_A_2090_A_379_D7'])\nba2090a379vb = array(th['B_A_2090_A_379_D7_V'])\n\n# bacteria\nv2090v = array(sv['V_A_2090_V'])\nv2090vb = array(oh['V_A_2090_D7_V'])\nv379vb = array(oh['V_A_379_D7_V'])\nva2090a379b = array(th['V_A_2090_A_379_D7'])\nva2090a379vb = array(th['V_A_2090_A_379_D7_V'])\n\n# algae\na2090sd = array(sv['A_2090_sd'])\na2090vsd = array(sv['A_2090_V_sd'])\na2090bsd = array(sb['A_2090_D7_sd'])\na2090vbsd = array(oh['A_2090_D7_V_sd'])\na379sd = array(sb['A_379_sd'])\na379bsd = array(sb['A_379_D7_sd'])\na379vbsd = array(oh['A_379_D7_V_sd'])\na2090a379sd = array(th['A_2090_A_379_sd'])\na2090a379vsd = array(th['A_2090_A_379_V_sd'])\na2090a379bsd = array(th['A_2090_A_379_D7_sd'])\na2090a379vbsd = array(th['A_2090_A_379_D7_V_sd'])\n\n# bacteria\nb2090bsd = array(sb['B_A_2090_D7_sd'])\nb2090vbsd = array(oh['B_A_2090_D7_V_sd'])\nbsd = array(sb['B_sd'])\nb379bsd = array(sb['B_A_379_D7_sd'])\nb379vbsd = array(oh['B_A_379_D7_V_sd'])\nba2090a379bsd = array(th['B_A_2090_A_379_D7_sd'])\nba2090a379vbsd = array(th['B_A_2090_A_379_D7_V_sd'])\n\n# bacteria\nv2090vsd = array(sv['V_A_2090_V_sd'])\nv2090vbsd = array(oh['V_A_2090_D7_V_sd'])\nv379vbsd = array(oh['V_A_379_D7_V_sd'])\nva2090a379bsd = array(th['V_A_2090_A_379_D7_sd'])\nva2090a379vbsd = array(th['V_A_2090_A_379_D7_V_sd'])\n\n# put in dictionaries for fitting\ncont_2090 = {'htimes': dtimes, 'hms': a2090, 'hss': a2090sd}\ncont_379 = {'htimes': dtimes, 'hms': a379, 'hss': a379sd}\ncont_D7 = {'htimes': dtimes, 'hms': b, 'hss': bsd}\n\ninf_379bac = {'htimes': dtimes, 'vtimes': dtimes,\n 'hms': a379b, 'vms': b379b, 'hss': a379bsd, 'vss': b379bsd}\ninf_2090bac = {'htimes': dtimes, 'vtimes': dtimes,\n 'hms': a2090b, 'vms': b2090b, 'hss': a2090bsd, 'vss': b2090bsd}\ninf_2090vir = {'htimes': dtimes[:6], 'vtimes': dtimes[:6], 'hms': a2090v[:6],\n 'vms': v2090v[:6], 'hss': a2090vsd[:6], 'vss': v2090vsd[:6]}\n","sub_path":"mikrodyno_depHeroku/read_all_data.py","file_name":"read_all_data.py","file_ext":"py","file_size_in_byte":9933,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"152187255","text":"from maix import nn\nfrom PIL import Image, ImageDraw\nimport numpy as np\nimport time\nfrom maix import display, camera\n\ncamera.config(size=(224, 224))\n\nmodel = {\n \"param\": \"/root/models/sobel_int8.param\",\n \"bin\": \"/root/models/sobel_int8.bin\"\n}\n\ninput_size = (224, 224, 3)\noutput_size = (222, 222, 3)\n\noptions = {\n \"model_type\": \"awnn\",\n \"inputs\": {\n \"input0\": input_size\n },\n \"outputs\": {\n \"output0\": output_size\n },\n \"mean\": [127.5, 127.5, 127.5],\n \"norm\": [0.0078125, 0.0078125, 0.0078125],\n}\nprint(\"-- load model:\", model)\nm = nn.load(model, opt=options)\nprint(\"-- load ok\")\n\nwhile 1:\n img = camera.capture()\n if not img:\n time.sleep(0.01)\n continue\n print(\"-- read image ok\")\n print(\"-- forward model with image as input\")\n out = m.forward(img, quantize=True, layout=\"hwc\")\n # print(\"-- read image ok\")\n # out = out.reshape(222, 222, 3)\n print(\"-- out:\", out.shape, out.dtype)\n out = out.astype(np.float32).reshape(output_size)\n out = (np.abs(out) * 255 / out.max()).astype(np.uint8)\n img2 = Image.fromarray(out, mode=\"RGB\")\n\n display.show(img2)\n","sub_path":"ext_modules/_maix_nn/example/load_forward_sobel_edge_camera.py","file_name":"load_forward_sobel_edge_camera.py","file_ext":"py","file_size_in_byte":1142,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"476168010","text":"from sys import argv,stderr\nfrom gzip import open as gzopen\n\nmatrixIn=argv[1]\nvalidSpecimensIn=argv[2]\nvalidProbes=set()\nnumValidProbes=0\nif len(argv)==4:\n validProbeFile=argv[3]\n with open(validProbeFile) as f:\n for line in f:\n validProbes.add(line.rstrip())\n numValidProbes+=1 \n\nvalidSpecimens=set()\nvalidDonors=set()\nwith open(validSpecimensIn) as f:\n for line in f:\n validSpecimens.add(line.rstrip())\n validDonors.add('-'.join(line.rstrip().split('-')[:-1]))\n\nwith gzopen(matrixIn,'rt') as f:\n processedSpecimens=set()\n headerChunks=next(f).rstrip().split('\\t')\n validColumns=[0]\n for i in range(1,len(headerChunks)):\n currentSpecimen=headerChunks[i]\n if not currentSpecimen[-1].isdigit():\n currentSpecimen=currentSpecimen[:-1]\n specimenChunks=currentSpecimen.split('-')\n healthySpecimen=int(specimenChunks[-1])>10\n currentDonor='-'.join(specimenChunks[:-1])\n validDonor=False\n if currentSpecimen in validSpecimens or (healthySpecimen and currentDonor in validDonors):\n if currentSpecimen not in processedSpecimens:\n processedSpecimens.add(currentSpecimen)\n validColumns.append(i)\n filteredChunks=[]\n for i in validColumns:\n if i>0:\n if not headerChunks[i][-1].isdigit():\n headerChunks[i]=headerChunks[i][:-1]\n filteredChunks.append(headerChunks[i])\n print(*filteredChunks,sep='\\t')\n expectedColumns=len(headerChunks)\n for line in f:\n lineChunks=line.rstrip().split('\\t')\n if lineChunks[0].startswith(\"ENSGR\"):\n continue\n if lineChunks[0].startswith(\"ENSG\"):\n lineChunks[0]=lineChunks[0].split('.')[0]\n currentColumns=len(lineChunks)\n if currentColumns==1:\n continue\n if numValidProbes>0:\n if lineChunks[0] not in validProbes:\n continue\n if currentColumns 1:\n main(args[1])\n else:\n main()","sub_path":"test/run.py","file_name":"run.py","file_ext":"py","file_size_in_byte":1043,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"29979169","text":"'''\nCreated on 25th/June/2017\n\n@author: aubrey.wang\n'''\n\nimport email.mime.multipart\nimport email.mime.text\nimport smtplib\nimport time, datetime\nfrom urllib import error\nimport urllib.request\n\n\ndef get_raw_url(req_name, list_data):\n website = 'http://hq.sinajs.cn/list=s_' + req_name;\n \n try:\n resp = urllib.request.urlopen(website)\n html = resp.read() \n\n except error.URLError as e:\n print('URLError reason: ', e.reason)\n \n # resp = urllib.request.urlopen(website)\n # html = resp.read()\n\n str = html.decode('gbk', 'ignore');\n\n # print(time.strftime(\"%Y-%m-%d %H:%M:%S\"),str)\n \n flag_colon_appear = 0\n data_order = 0\n str_data = \"\"\n piece_data = []\n\n for char in str:\n if (char == ',') and (flag_colon_appear == 0):\n piece_data.append(req_name)\n flag_colon_appear = 1\n continue\n \n if (char == ',') and (flag_colon_appear == 1):\n flag_colon_appear = 1 \n piece_data.append( (float) (str_data) )\n \n str_data = ''\n continue\n \n if (char == '\"') and (flag_colon_appear == 1):\n flag_colon_appear = 1\n \n piece_data.append( (float) (str_data) )\n # piece_data.insert(0, time.strftime(\"%Y-%m-%d %H:%M:%S\") )\n piece_data.insert(0, time.time())\n \n print('piece_data:', piece_data)\n \n list_data.append(piece_data)\n \n piece_data = []\n str_data = ''\n \n data_order += 1\n continue\n \n if flag_colon_appear == 1:\n str_data += char \n return list_data\n\ndef get_newest_data(stock_raw_data_list):\n for get_newest_data_i in reversed(stock_raw_data_list):\n # print('get_newest_data_name:',get_newest_data_i)\n # print('data length:',len(stock_raw_data_list))\n return get_newest_data_i\n \ndef is_trade_time():\n is_trade_time = ( (datetime.datetime.now().hour == 9) and (datetime.datetime.now().minute > 29) ) \\\n or (datetime.datetime.now().hour == 10) \\\n or ( (datetime.datetime.now().hour == 11) and (datetime.datetime.now().minute < 31) ) \\\n or (datetime.datetime.now().hour == 13) \\\n or (datetime.datetime.now().hour == 14) \\\n or ( (datetime.datetime.now().hour == 15) and (datetime.datetime.now().minute < 1) )\n \n return is_trade_time\n\ndef list_write2file(file_name, send_list):\n file_to_write = open(file_name,'a')\n\n for item in send_list:\n file_to_write.write(str(item)[1:-1]+'\\n')\n \n file_to_write.close()\n \ndef send_mail(flag, strdata):\n msg = email.mime.multipart.MIMEMultipart()\n msg['from'] = '18566260586m@sina.cn'\n msg['to'] = '18566260586@163.com'\n msg['subject'] = flag\n \n content = strdata\n \n txt = email.mime.text.MIMEText(content)\n \n msg.attach(txt)\n \n smtp=smtplib.SMTP()\n \n smtp.connect('smtp.sina.cn', '25')\n smtp.login('18566260586m@sina.cn', 'wk19910415')\n \n smtp.sendmail(\"18566260586m@sina.cn\",\"18566260586@163.com\",str(msg))\n smtp.quit()\n \n print(time.strftime(\"%Y-%m-%d %H:%M:%S\"), ' send_data:', flag)","sub_path":"app_main/app_data_io.py","file_name":"app_data_io.py","file_ext":"py","file_size_in_byte":3274,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"191899194","text":"#!/usr/bin/env python3\n# -*- coding:UTF-8 -*-\n\nfrom __future__ import print_function\n\nfrom __future__ import absolute_import\n\nimport subprocess\nfrom sys import stdout\n\n\ndef main():\n print('start....')\n #TerminalCheck.py -a 192.168.0.109 -p 0 0\n# proc = subprocess.Popen('python3 TerminalCheck.py -a 192.168.0.109 -p 0 0'\n# ,stdin=subprocess.PIPE, stdout=subprocess.PIPE,\n# stderr=subprocess.PIPE,shell=True)\n \n# proc = subprocess.Popen('python3 TerminalCheck.py -a 192.168.0.109 -p 0 0'\n# ,shell=True)\n# print('haha')\n# while True:\n# output, usrerror = proc.communicate(timeout=2)\n# print(output)\n \n \n proc1 = subprocess.Popen('python3 call_sub.py'\n ,stdin=subprocess.PIPE, stdout=subprocess.PIPE,\n stderr=subprocess.PIPE,shell=True)\n exe_result1 = proc1.stdout.read()\n print(\"mdsb:\" + exe_result1.decode())\n \n now_md5sum = subprocess.check_output(\"md5sum config.txt\",shell=True)\n print(\"now_md5sum:\" + now_md5sum.decode())\n \nif __name__ == '__main__':\n main()","sub_path":"step1/configobj_test/subprocess_test.py","file_name":"subprocess_test.py","file_ext":"py","file_size_in_byte":1179,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"241551054","text":"\"\"\" Tests for `yatsm.gis.tilespec`\n\"\"\"\nimport pytest\n\nfrom yatsm.gis import tilespec\n\n\n@pytest.fixture\ndef example_spec(request):\n for k in tilespec.TILESPECS:\n return tilespec.TILESPECS[k]\n\n\nEX_CRS_GEOG = (\n (0., 0.),\n 'epsg:4326',\n (0.00025, 0.0025),\n (1., 1.),\n 'geographic'\n)\nEX_CRS_ALBERS = (\n (-2565600., 3314800.),\n 'epsg:5070',\n (30, 30),\n (250, 250),\n 'albers_conus'\n)\nEX_CRS_UTM = (\n (653385., 4828815.),\n 'epsg:32619',\n (30, 30),\n (5000, 5000),\n 'utm19n'\n)\n\n\ntilespec_params = pytest.mark.parametrize(\n ('ul', 'crs', 'res', 'size', 'desc'),\n [EX_CRS_GEOG,\n EX_CRS_ALBERS,\n EX_CRS_UTM]\n)\n\n\n# tilezilla.tilespec.TileSpec\n@tilespec_params\ndef test_tilespec(ul, crs, res, size, desc):\n tilespec.TileSpec(ul, crs, res, size, desc=desc)\n\n\n# FAILURE: CRS PARSING PROBLEMS\n@tilespec_params\ndef test_tilespec_fail_crs_1(ul, crs, res, size, desc):\n with pytest.raises(ValueError):\n tilespec.TileSpec(ul, 'not a crs', res, size, desc)\n\n\n# FAILURE: INDEXING PROBLEMS\ndef test_tilespec_fail_1(example_spec):\n with pytest.raises(IndexError):\n example_spec[-1]\n\n\ndef test_tilespec_fail_2(example_spec):\n with pytest.raises(TypeError):\n example_spec[([0, 1], [0, 1])]\n","sub_path":"tests/gis/test_gis_tilespec.py","file_name":"test_gis_tilespec.py","file_ext":"py","file_size_in_byte":1265,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"612322572","text":"# Prepare training and test data.\nimport os\nimport pyspark\nimport unittest\nimport xmlrunner\nfrom mmlspark.RankingAdapter import RankingAdapter\nfrom mmlspark.RankingEvaluator import RankingEvaluator\nfrom mmlspark.RankingTrainValidationSplit import RankingTrainValidationSplit, RankingTrainValidationSplitModel\nfrom pyspark.ml import Pipeline\nfrom pyspark.ml.feature import StringIndexer\nfrom pyspark.ml.tuning import *\nfrom pyspark.ml.tuning import *\nfrom pyspark.sql.types import *\nfrom pyspark.ml.recommendation import ALS\n\n\nclass RankingSpec(unittest.TestCase):\n\n @staticmethod\n def getRatings():\n cSchema = StructType([StructField(\"originalCustomerID\", IntegerType()),\n StructField(\"newCategoryID\", IntegerType()),\n StructField(\"rating\", IntegerType()),\n StructField(\"notTime\", IntegerType())])\n\n ratings = pyspark.sql.SparkSession.builder.getOrCreate().createDataFrame([\n (0, 1, 4, 4),\n (0, 3, 1, 1),\n (0, 4, 5, 5),\n (0, 5, 3, 3),\n (0, 7, 3, 3),\n (0, 9, 3, 3),\n (0, 10, 3, 3),\n (1, 1, 4, 4),\n (1, 2, 5, 5),\n (1, 3, 1, 1),\n (1, 6, 4, 4),\n (1, 7, 5, 5),\n (1, 8, 1, 1),\n (1, 10, 3, 3),\n (2, 1, 4, 4),\n (2, 2, 1, 1),\n (2, 3, 1, 1),\n (2, 4, 5, 5),\n (2, 5, 3, 3),\n (2, 6, 4, 4),\n (2, 8, 1, 1),\n (2, 9, 5, 5),\n (2, 10, 3, 3),\n (3, 2, 5, 5),\n (3, 3, 1, 1),\n (3, 4, 5, 5),\n (3, 5, 3, 3),\n (3, 6, 4, 4),\n (3, 7, 5, 5),\n (3, 8, 1, 1),\n (3, 9, 5, 5),\n (3, 10, 3, 3)], cSchema)\n return ratings\n\n @staticmethod\n def get_pyspark():\n return pyspark.sql.SparkSession.builder.master(\"local[*]\").config('spark.driver.extraClassPath',\n \"../../../../../BuildArtifacts/packages/m2/com/microsoft/ml/spark/mmlspark_2.11/0.0/mmlspark_2.11-0.0.jar\").getOrCreate()\n\n def test_adapter_evaluator(self):\n self.get_pyspark()\n\n ratings = self.getRatings()\n\n user_id = \"originalCustomerID\"\n item_id = \"newCategoryID\"\n rating_id = 'rating'\n\n user_id_index = \"customerID\"\n item_id_index = \"itemID\"\n\n customer_indexer = StringIndexer(inputCol=user_id, outputCol=user_id_index).fit(ratings)\n items_indexer = StringIndexer(inputCol=item_id, outputCol=item_id_index).fit(ratings)\n\n als = ALS(userCol=user_id_index, itemCol=item_id_index, ratingCol=rating_id)\n\n adapter = RankingAdapter(mode='allUsers', k=5, recommender=als)\n\n pipeline = Pipeline(stages=[customer_indexer, items_indexer, adapter])\n output = pipeline.fit(ratings).transform(ratings)\n print(str(output.take(1)) + \"\\n\")\n\n metrics = ['ndcgAt', 'fcp', 'mrr']\n for metric in metrics:\n print(metric + \": \" + str(RankingEvaluator(k=3, metricName=metric).evaluate(output)))\n\n\nif __name__ == \"__main__\":\n result = unittest.main(testRunner=xmlrunner.XMLTestRunner(output=os.getenv(\"TEST_RESULTS\", \"TestResults\")), \\\n failfast=False, buffer=False, catchbreak=False)\n","sub_path":"src/recommendation/src/test/python/RankingSpec.py","file_name":"RankingSpec.py","file_ext":"py","file_size_in_byte":3393,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"164705123","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n\nclass Solution(object):\n def maxSubArray(self, nums):\n \"\"\"\n :type nums: List[int]\n :rtype: int\n \"\"\"\n if not nums:\n return 0\n length = len(nums)\n if length == 1:\n return nums[0]\n last_max_sum = nums[0]\n max_subarray_sum = nums[0]\n for i in range(1, length):\n last_max_sum = max(last_max_sum+nums[i], nums[i])\n max_subarray_sum = max(max_subarray_sum, last_max_sum)\n return max_subarray_sum\n","sub_path":"leetcode/053.py","file_name":"053.py","file_ext":"py","file_size_in_byte":561,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"174131965","text":"from room import Room\nfrom player import Player\nfrom world import World\n\nimport random\nfrom ast import literal_eval\n\n# Load world\nworld = World()\n\n\n# You may uncomment the smaller graphs for development and testing purposes.\n# map_file = \"maps/test_line.txt\"\n# map_file = \"maps/test_cross.txt\"\n# map_file = \"maps/test_loop.txt\"\n# map_file = \"maps/test_loop_fork.txt\"\nmap_file = \"maps/main_maze.txt\"\n\n# Loads the map into a dictionary\nroom_graph = literal_eval(open(map_file, \"r\").read())\nworld.load_graph(room_graph)\n\n# Print an ASCII map\nworld.print_rooms()\nplayer = Player(world.starting_room)\n\ntraversal_path = []\n\n\n# move player helper function\ndef move_player(direction):\n player.travel(direction)\n traversal_path.append(direction)\n\n# * Recursive DFT\n\n\ndef find_path_rec(visited=None, previous=None, move=None):\n # visited dict starts as none, previous room starts as none, move input starts as none\n curr_id = player.current_room.id\n connected_rooms = player.current_room.get_exits()\n # * enable easily determining what room the player came from -- previous room is the room opposite Move\n reverse_dirs = {\n 'n': 's',\n 's': 'n',\n 'e': 'w',\n 'w': 'e'\n }\n\n # * instantiates visited set at first\n if visited == None:\n visited = {}\n\n # * create empty set in visited when curr_id not in visited\n if curr_id not in visited:\n visited[curr_id] = {}\n\n # * handles movement command -- if the function recieves a move, the curr_id is assigned to the move applied to the previous room\n if move is not None:\n visited[previous][move] = curr_id\n\n # * the previous room is the room at the opposite-provided direction\n # ! handle after checking move\n if previous is not None:\n visited[curr_id][reverse_dirs[move]] = previous\n\n # * determines if there are remaining neighbors to be visited and recursively visits\n if len(visited[curr_id]) < len(connected_rooms):\n for direction in connected_rooms:\n if direction not in visited[curr_id]:\n move_player(direction)\n find_path_rec(visited, previous=curr_id, move=direction)\n\n # * If the player has visited fewer than total rooms move the player backwards\n if len(visited) < len(room_graph):\n move_player(reverse_dirs[move])\n\n\n# ! executes recursive function to determine path\nfind_path_rec()\n\nvisited_rooms = set()\nplayer.current_room = world.starting_room\n\n\n# * Allows player to be controlled by dft function\n# Moves player according to traversal_path\nfor direction in traversal_path:\n player.travel(direction)\n current_room = player.current_room\n visited_rooms.add(current_room)\n\nprint('traversal_path: ', traversal_path)\n\nif len(visited_rooms) == len(room_graph):\n print(\n f\"TESTS PASSED: {len(traversal_path)} moves, {len(visited_rooms)} rooms visited\")\nelse:\n print(\"TESTS FAILED: INCOMPLETE TRAVERSAL\")\n print(f\"{len(room_graph) - len(visited_rooms)} unvisited rooms\")\n\n\n#######\n# UNCOMMENT TO WALK AROUND\n#######\nplayer.current_room.print_room_description(player)\nwhile True:\n cmds = input(\"-> \").lower().split(\" \")\n if cmds[0] in [\"n\", \"s\", \"e\", \"w\"]:\n player.travel(cmds[0], True)\n elif cmds[0] == \"q\":\n break\n else:\n print(\"I did not understand that command.\")\n","sub_path":"adv.py","file_name":"adv.py","file_ext":"py","file_size_in_byte":3336,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"298437830","text":"from __future__ import print_function\nimport numpy as np\nfrom spt3g import core\n\nclass PlotG3MapHistograms(object):\n def __init__(self, g3_map_key, frame_type, save_folder, save_tag, n_bins, \n log_y = False, log_x= False):\n self.g3_map_key = g3_map_key\n self.frame_type = frame_type\n self.save_folder = save_folder\n self.save_tag = save_tag\n self.bins = n_bins\n self.log = log_y\n \n self.log_x = log_x\n \n self.scan_num = 0\n\n def __call__(self, frame):\n import pylab as pl\n if self.frame_type != frame.type:\n return\n mp = frame[self.g3_map_key]\n vals = np.nan_to_num(np.asarray(mp.values()))\n pl.clf()\n\n if self.log_x:\n print(min(vals), max(vals))\n start = np.log(min(vals[np.where(vals > 0)]))/np.log(10)\n stop = np.log(max(vals))/np.log(10)\n bins = np.logspace(start - 1, stop + 1, num = self.bins)\n else:\n bins = self.bins\n #import pdb; pdb.set_trace()\n pl.hist(vals, bins = bins, log=self.log)\n\n if self.log_x:\n pl.semilogx()\n\n pl.savefig('%s/%s_%d.png' % (self.save_folder, self.save_tag, self.scan_num ))\n self.scan_num += 1\n\n\nclass ReportStatistics(object):\n def __init__(self, type, labels, funcs, also_plot = True, plot_dir = '.'):\n assert(len(labels) == len(funcs))\n\n self.type = type\n self.labels = labels\n self.funcs = funcs\n self.also_plot = also_plot\n self.vals = [ [] for l in labels]\n self.plot_dir = plot_dir\n def __call__(self, frame):\n if frame.type == self.type:\n for i in range(len(self.labels)):\n self.vals[i].append( self.funcs[i](frame))\n if frame.type == core.G3FrameType.EndProcessing:\n for i in range(len(self.labels)):\n print(self.labels[i],\":\")\n print(self.vals[i])\n print('\\n\\n')\n if self.also_plot:\n import pylab as pl\n pl.clf()\n pl.ylabel(self.labels[i])\n pl.xlabel('Frame')\n pl.plot(self.vals[i])\n pl.savefig(self.plot_dir + '/' + self.labels[i] + '.png')\n","sub_path":"util/python/datavis.py","file_name":"datavis.py","file_ext":"py","file_size_in_byte":2320,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"135019823","text":"from itertools import cycle\n\ndef byte_xor(A, B):\n\tzip_list = zip(A, cycle(B)) if len(A) > len(B) else zip(cycle(A), B)\n\treturn bytes([_a ^ _b for _a, _b in zip_list])\n\ndef get_character_score(character):\n\tkeys = {\n\t'a': 0.08167,\n 'b': 0.01492,\n 'c': 0.02782,\n 'd': 0.04253,\n 'e': 0.12702,\n 'f': 0.02228,\n 'g': 0.02015,\n 'h': 0.06094,\n 'i': 0.06094,\n 'j': 0.00153,\n 'k': 0.00772,\n 'l': 0.04025,\n 'm': 0.02406,\n 'n': 0.06749,\n 'o': 0.07507,\n 'p': 0.01929,\n 'q': 0.00095,\n 'r': 0.05987,\n 's': 0.06327,\n 't': 0.09056,\n 'u': 0.02758,\n 'v': 0.00978,\n 'w': 0.02360,\n 'x': 0.00150,\n 'y': 0.01974,\n 'z': 0.00074,\n ' ': 0.13000}\n\tif(character in keys):\n\t\treturn keys[character]\n\treturn 0\n\n\ndef get_best_english_string(_bytes):\n\tresult = {\"key\":' ', \"score\":0, \"text\":' '}\n\tfor key in range(32,127):\n\t\txor_bytes = byte_xor(_bytes, (key).to_bytes(2, byteorder='big'))\n\t\ttext = xor_bytes.decode(\"utf-8\") \n\t\ttext_score = get_english_score(text)\n\t\tif text_score > result[\"score\"]:\n\t\t\tresult[\"key\"] = key\n\t\t\tresult[\"score\"] = text_score\n\t\t\tresult[\"text\"] = text\n\treturn result\n\ndef get_english_score(text):\n\ttext = text.lower()\n\ttext_scores = []\n\tfor character in text:\n\t\ttext_scores.append(get_character_score(character))\n\treturn sum(text_scores)\n\n\ndef main():\n print(get_character_score('a') + get_character_score('b') + get_character_score('c'))\n print(get_english_score(\"abc\"))\n _bytes = bytes.fromhex(\"1b37373331363f78151b7f2b783431333d78397828372d363c78373e783a393b3736\")\n print(get_english_score(byte_xor(_bytes, b'X').decode(\"utf-8\")))\n print(get_english_score(byte_xor(_bytes, bytes(b'z')).decode(\"utf-8\")))\n\nif __name__ == '__main__':\n main()\n","sub_path":"05.Security/tasks/cryptopals-challanges/Python/crypto.py","file_name":"crypto.py","file_ext":"py","file_size_in_byte":1733,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"449140343","text":"def solve(str_number):\n num = int(str_number)\n \n if num == 0:\n return \"INSOMNIA\" \n \n result = num\n \n digits = [\"0\", \"1\", \"2\", \"3\", \"4\", \"5\", \"6\", \"7\", \"8\", \"9\"]\n \n \n while True:\n str_result = str(result)\n digits = [d for d in digits if d not in str_result]\n if len(digits) == 0:\n return str_result\n else:\n result += num\n\n\n# Read file\nwith open('input.txt') as f:\n content = f.readlines()\n\nf_output = open(\"output.txt\", \"wb\")\nfor i in range(1, int(content[0]) +1):\n f_output.write(\"Case #\" + str(i) + \": \" +solve(content[i]) + \"\\r\\n\");\nf_output.close()\n\n\n","sub_path":"codes/CodeJamCrawler/16_0_1_neat/16_0_1_Kostadinov_ProblemACountingSheep.py","file_name":"16_0_1_Kostadinov_ProblemACountingSheep.py","file_ext":"py","file_size_in_byte":647,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"333691293","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nSpyder Editor\r\n\r\nThis is a temporary script file.\r\n\"\"\"\r\nimport MDAnalysis\r\nimport matplotlib.pyplot as plt\r\nimport numpy as np\r\nimport csv\r\n\r\nparams = {'legend.fontsize': 8,\r\n 'legend.handlelength': 2}\r\nplt.rcParams.update(params)\r\n\r\nz,h1,h2,h3,h4,h5,h6,h7,h8,h9,h10,h11,h12,h13,h14,h15,h16,h17,h18,h19,h20,h21 = [],[],[],[],[],[],[],[],[],[],[],[],[],[],[],[],[],[],[],[],[],[]\r\n\r\nwith open(r\"C:\\Users\\emfla\\Desktop\\CANES_2019\\research-project-assignment\\umbrella-sampling-data\\hairpin1_pulling_umbrella\\b=20\\histo_data.xvg\") as f:\r\n reader = csv.reader(f,delimiter = \"\\t\")\r\n for line in reader:\r\n a = np.transpose(line)\r\n z = np.append(z,a[0])\r\n h1 = np.append(h1,a[1])\r\n h2 = np.append(h2, a[2])\r\n h3 = np.append(h3, a[3])\r\n h4 = np.append(h4, a[4])\r\n h5 = np.append(h5, a[5])\r\n h6 = np.append(h6, a[6])\r\n h7 = np.append(h7, a[7])\r\n h8 = np.append(h8, a[8])\r\n h9 = np.append(h9, a[9])\r\n h10 = np.append(h10, a[10])\r\n h11 = np.append(h11, a[11])\r\n h12 = np.append(h12, a[12])\r\n h13 = np.append(h13, a[13])\r\n h14 = np.append(h14, a[14])\r\n h15 = np.append(h15, a[15])\r\n h16 = np.append(h16, a[16])\r\n h17 = np.append(h17, a[17])\r\n h18 = np.append(h18, a[18])\r\n h19 = np.append(h19, a[19])\r\n h20 = np.append(h20, a[20])\r\n h21 = np.append(h21, a[21])\r\n \r\n\r\nif h1[0]!=float: \r\n z = np.array(z,dtype=float)\r\n h1 = np.array(h1,dtype=float)\r\n h2 = np.array(h2,dtype=float)\r\n h3 = np.array(h3,dtype=float)\r\n h4 = np.array(h4,dtype=float)\r\n h5 = np.array(h5,dtype=float)\r\n h6 = np.array(h6,dtype=float)\r\n h7 = np.array(h7,dtype=float)\r\n h8 = np.array(h8,dtype=float)\r\n h9 = np.array(h9,dtype=float)\r\n h10 = np.array(h10,dtype=float)\r\n h11 = np.array(h11,dtype=float)\r\n h12 = np.array(h12,dtype=float)\r\n h13 = np.array(h13,dtype=float)\r\n h14 = np.array(h14,dtype=float)\r\n h15 = np.array(h15,dtype=float)\r\n h16 = np.array(h16,dtype=float)\r\n h17 = np.array(h17,dtype=float)\r\n h18 = np.array(h18,dtype=float)\r\n h19 = np.array(h19,dtype=float)\r\n h20 = np.array(h20,dtype=float)\r\n h21 = np.array(h21,dtype=float)\r\n\r\n\r\nhistos = [h1,h2,h3,h4,h5,h6,h7,h8,h9,h10,h11,h12,h13,h14,h15,h16,h17,h18,h19,h20,h21]\r\n\r\nplt.figure()\r\nfor i in range(len(histos)):\r\n plt.plot(z,histos[i],lw=2)\r\nplt.xlabel('$z$ ($nm$)')\r\n#plt.xlim(1.2,4.2)\r\nplt.ylim(0)\r\nplt.ylabel('Count')\r\nplt.savefig('hairpin_histo.png')\r\nplt.show()\r\n\r\ndef return_intersection(hist_1,hist_2):\r\n minima = np.minimum(hist_1,hist_2)\r\n intersection = np.true_divide(np.sum(minima),np.sum(hist_2))\r\n return intersection\r\n\r\nfraction = []\r\nfor i in range(len(histos)-1):\r\n fraction.append(return_intersection(histos[i],histos[i+1]))\r\n \r\nfraction_2 = []\r\nfor i in range(len(histos)-2):\r\n fraction_2.append(return_intersection(histos[i],histos[i+2]))\r\n\r\nplt.clf()\r\nplt.figure()\r\nplt.plot(fraction,label='Fractional overlap')\r\nplt.plot(np.arange(0,35,1),np.repeat(0.3173,35),'k--',label=r'$1\\sigma$ overlap')\r\nplt.plot(np.arange(0,35,1),np.repeat(0.0455,35),'r--',label=r'$2\\sigma$ overlap')\r\nplt.plot(np.arange(0,35,1),np.repeat(0.0027,35),'g--',label=r'$3\\sigma$ overlap')\r\nplt.xlim(0,20)\r\nplt.ylim(0,1)\r\nplt.xlabel(r'Histogram pairs ($x$, $x+1$)')\r\nplt.ylabel(r'Fractional first neighbour overlap')\r\nplt.savefig('hairpin_first_overlap.png')\r\nplt.show()\r\n \r\nplt.clf()\r\nplt.figure()\r\nplt.plot(fraction_2,label='Fractional overlap')\r\nplt.plot(np.arange(0,35,1),np.repeat(0.3173,35),'k--',label=r'$1\\sigma$ overlap')\r\nplt.plot(np.arange(0,35,1),np.repeat(0.0455,35),'r--',label=r'$2\\sigma$ overlap')\r\nplt.plot(np.arange(0,35,1),np.repeat(0.0027,35),'g--',label=r'$3\\sigma$ overlap')\r\nplt.xlim(0,20)\r\nplt.ylim(0,1)\r\nplt.xlabel(r'Histogram pairs ($x$, $x+2$)')\r\nplt.ylabel(r'Fractional second neighbour overlap')\r\nplt.savefig('hairpin_second_overlap.png')\r\nplt.show()\r\n \r\n'''\r\ndef histogram_overlap(h1,h2):\r\n sm = 0.0\r\n for i in range(len(z)):\r\n sm += min(h1[i],h2[i])\r\n return sm\r\ndef total_area(h1,h2):\r\n a = 0.0\r\n for i in range(len(z)):\r\n a += max(h1[i],h2[i])\r\n return a\r\n\r\nfraction = []\r\nfor i in range(len(histos)-1):\r\n fraction.append(histogram_overlap(histos[i],histos[i+1])/total_area(histos[i],histos[i+1]))\r\n \r\nfraction_2 = []\r\nfor i in range(len(histos)-2):\r\n fraction_2.append(histogram_overlap(histos[i],histos[i+2])/total_area(histos[i],histos[i+2])) \r\n''' ","sub_path":"backup/overlap_hairpin.py","file_name":"overlap_hairpin.py","file_ext":"py","file_size_in_byte":4564,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"236619258","text":"\"\"\"\nRegular expressions are terrible, tedious, terrible things.\n\"\"\"\nimport re \n\n\ndef like_scanf(s, sep=\"\\s*\"):\n \"\"\" Convert a 'scanf()' style string to a regex \"\"\"\n mapping = \\\n {\n r\"%f\" : r\"[-+]?(\\d+(\\.\\d*)?|\\.\\d+)([eE][-+]?\\d+)?\",\n r\"%d\" : r\"[-+]?\\d+\",\n }\n valid = \"|\".join(mapping.keys())\n fmtstr = re.findall(valid, s)\n fmtlst = [mapping[x] for x in fmtstr]\n return sep.join(fmtlst)\n","sub_path":"realtime-graph-follow/webserver/lineparse.py","file_name":"lineparse.py","file_ext":"py","file_size_in_byte":426,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"127169621","text":"from django.shortcuts import render\nfrom .form import ContactForm\nfrom .models import Contact\n# Create your views here.\ndef contact(request):\n form = ContactForm(request.POST or None)\n if form.is_valid():\n obj = Contact.objects.create(**form.cleaned_data)\n print(form.cleaned_data)\n form = ContactForm()\n context = {\n \"title\": \"Contact Us\",\n \"form\": form\n }\n return render(request, \"form.html\",context)","sub_path":"contact/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":452,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"61593064","text":"import pandas as pd\nimport requests\nfrom datetime import datetime\nfrom datetime import timedelta\nimport easygui\nimport time\nfrom binance.client import Client\nimport array\nimport tulipy as ti\n\n\n\n\nclient = Client('1XGCXOq8RCHuKA0O322OHahi0Kg0KsSHsG4ai4Gbp7MmaLFwVEOxGoZ2G1KSjEAS','207ia9nrYf8OF3LDXjMPUShYxEDAQWUwJBxv1wzHUDmswHWlU1udgCHc7xxwyTiK')\ncryptoCount = 1\n\narr = array.array('i',[])\nsmall = 60\nmedium = 540\nlarge = 960\ndef find(total, smallavgPrice = 0, mediumavgPrice = 0, largeavgPrice = 0):\n BTCUSDTPrice = requests.get(\"https://api.binance.com/api/v1/ticker/price?symbol=ALGOBTC\")\n arr.insert(total, int(float(BTCUSDTPrice.json()['price'])))\n count = len(arr)\n if (count > small):\n for i in range(small):\n smallavgPrice = smallavgPrice + arr[count - 1 - i]\n\n #print(\"small: \",smallavgPrice/15)\n if (count > medium):\n for i in range(medium):\n mediumavgPrice = mediumavgPrice + arr[count - 1 - i]\n\n #print(\"medium :\",mediumavgPrice / 30)\n if (count > large):\n for i in range(large):\n largeavgPrice = largeavgPrice + arr[count - 1 - i]\n\n #print(\"large: \",largeavgPrice / 60)\n\n if (smallavgPrice > largeavgPrice and smallavgPrice > mediumavgPrice and count >= medium):\n\n print(\"Buy\")\n elif (cryptoCount > 0 and smallavgPrice <= mediumavgPrice and count >= medium):\n\n print(\"Sell\")\n time.sleep(900)\n find(total + 1)\n\n\nfind(0)\n","sub_path":"Simple Moving Average Method Conintues Data/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":1498,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"601086852","text":"'''\nGiven an unsorted array return whether an increasing subsequence of length 3 exists or not in the array.\n\nFormally the function should:\nReturn true if there exists i, j, k\nsuch that arr[i] < arr[j] < arr[k] given 0 ≤ i < j < k ≤ n-1 else return false.\nYour algorithm should run in O(n) time complexity and O(1) space complexity.\n\nExample\nGiven [1, 2, 3, 4, 5],\nreturn true.\n\nGiven [5, 4, 3, 2, 1],\nreturn false.\n'''\nclass Solution:\n \"\"\"\n @param nums: a list of integers\n @return: return a boolean\n \"\"\"\n #time: O(n), space O(n)\n # def increasingTriplet(self, nums):\n # if len(nums) < 3: return False\n \n # left_min = [0] * len(nums)\n # lmin = nums[0]\n # for i, n in enumerate(nums):\n # lmin = min(lmin, n)\n # left_min[i] = lmin\n \n # right_max = [0] * len(nums)\n # rmax = nums[-1]\n # for i in range(len(nums)-1, -1, -1):\n # rmax = max(nums[i], rmax)\n # right_max[i] = rmax\n \n # for i in range(1, len(nums)-1):\n # if left_min[i-1] < nums[i] < right_max[i+1]: return True\n # return False\n def increasingTriplet(self, nums):\n small = big = float('inf')\n for n in nums:\n if n <= small: small = n\n elif n <= big: big = n\n else: return True\n return False\n","sub_path":"leetcode/increasingTriplet.py","file_name":"increasingTriplet.py","file_ext":"py","file_size_in_byte":1372,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"373150875","text":"import sys\n\n\ntest_case = open(sys.argv[1], 'r')\nfor test in test_case:\n line = test.strip().split(' ')\n new_line = ''\n x = int(line[0])\n y = int(line[1])\n n = int(line[2])\n\n for i in range(1, n+1):\n to_append = ''\n if i % x == 0 and i % y == 0:\n to_append = 'FB'\n elif i % x == 0:\n to_append = 'F'\n elif i % y == 0:\n to_append = 'B'\n else:\n to_append = str(i)\n\n new_line = new_line + to_append + ' '\n\n print(new_line.strip())\ntest_case.close()\n\n","sub_path":"solved/fizz_buzz.py","file_name":"fizz_buzz.py","file_ext":"py","file_size_in_byte":552,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"316960344","text":"import utils\nimport classifiers\nimport numpy as np\n\nfilename = '../data/PolySVM_Normalized.txt'\n\ndef get_poly_function(c, d):\n def poly_function(x1, x2):\n value = x1.T @ x2\n value = value + c\n value = value ** d\n return value\n return poly_function\n\nif __name__ == '__main__':\n dataset = utils.load_train_data()\n dataset = utils.normalize(dataset)\n _, folds = utils.kfold(dataset, n=3)\n outfile = open(filename, 'w')\n constants = [.5, 1, 3, 5]\n powers = [1, 2, 3]\n bounds = [.1, .5, 1]\n npca = [11, 10, 9, 8]\n w, v = utils.PCA(dataset)\n\n for power in powers:\n for constant in constants:\n for bound in bounds:\n for n in npca:\n vt = v[:, :n]\n poly_function = get_poly_function(power, constant)\n scores, labels = [], []\n for fold in folds:\n train, test = fold[0], fold[1]\n train, test = np.vstack((vt.T @ train[:-1, :], train[-1])), np.vstack((vt.T @ test[:-1, :], test[-1]))\n fold_labels = test[-1, :]\n labels.append(fold_labels)\n\n alphas = classifiers.DualSVM_Train(train, poly_function, bound=bound)\n train, alphas = utils.support_vectors(train, alphas)\n fold_scores = classifiers.DualSVM_Score(train, poly_function, alphas, test)\n scores.append(fold_scores)\n\n scores = np.concatenate(scores)\n labels = np.concatenate(labels)\n mindcf, optimal_threshold = utils.minDCF(scores, labels, prior_t=.5)\n # Ignore the first field, is just handy for sorting\n print(f\"{mindcf} |.| MinDCF: {mindcf:.4f} - PCA: {n} - Opt. Thr.: {optimal_threshold:.4f} - Power: {power:.4f} - Constant: {constant:.4f} - C: {bound:.4f}\", file=outfile)\n np.save(f'../data/PolySVM-Normalized-PCA{n}-C{constant}-POW{power}Scores.npy', scores)\n\n outfile.close()\n","sub_path":"code/PolySVM_Normalized.py","file_name":"PolySVM_Normalized.py","file_ext":"py","file_size_in_byte":2110,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"510805653","text":"#!/usr/bin/env python\n\nimport epitome as epi\n\ndef run(input_name):\n output = 'detrend'\n\n print('\\nAdding detrend module.')\n\n try:\n print('\\nSet detrend order:')\n polort = epi.utilities.selector_int()\n\n # if we messed any of these up, we return None\n except ValueError as ve:\n return '', None\n\n # otherwise we print the command and return it\n line = ('. ${{DIR_PIPE}}/epitome/modules/pre/detrend {input_name} {polort}').format(\n input_name=str(input_name),\n polort=str(polort))\n\n return line, output\n","sub_path":"assets/epitome/151012-spins/epitome/commands/detrend.py","file_name":"detrend.py","file_ext":"py","file_size_in_byte":602,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"33711390","text":"from decimal import Decimal\nimport uuid\n\nfrom django.db import models\nfrom django.db.models import Sum, F\nfrom django.db.models.signals import post_save, post_delete\nfrom django.dispatch import receiver\nfrom django.conf import settings\nfrom django.core.validators import MinValueValidator, MaxValueValidator\nfrom django.core.urlresolvers import reverse\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom mptt.models import MPTTModel, TreeForeignKey\nfrom autoslug import AutoSlugField\nfrom filebrowser.fields import FileBrowseField\nfrom polymorphic.models import PolymorphicModel\n\nfrom . import settings as shop_settings\n\n\nPAYMENT_GATEWAYS_CHOICES = tuple(\n (g, shop_settings.SHOP_PAYMENT_GATEWAYS[g]['verbose_name'])\n for g in shop_settings.SHOP_PAYMENT_GATEWAYS\n)\n\n\nPRODUCT_ATTRIBUTE_VALUE_TYPE_CHOICES = (\n ('options', _('Options')),\n ('text', _('Text'))\n)\n\n\nclass Category(MPTTModel):\n title = models.CharField(_('title'), max_length=255)\n slug = AutoSlugField(\n _('slug'), populate_from='title',\n blank=True, editable=True, unique=True\n )\n parent = TreeForeignKey(\n 'self', null=True, blank=True, related_name='children', db_index=True\n )\n order = models.PositiveSmallIntegerField(_('order'), default=0)\n published = models.BooleanField(_('published'), default=True)\n description = models.TextField(_('description'), default='', blank=True)\n\n class Meta:\n app_label = 'shop'\n ordering = ('order',)\n verbose_name = _('category')\n verbose_name_plural = _('categories')\n\n class MPTTMeta:\n order_insertion_by = ('order',)\n\n @staticmethod\n def autocomplete_search_fields():\n return ('id__iexact', 'title__icontains') # nocov\n\n def save(self, *args, **kwargs):\n super(Category, self).save(*args, **kwargs)\n unpublished = Category.objects.filter(published=False)\n unpublished_descendants_pks = []\n for c in unpublished:\n unpublished_descendants_pks += c.get_descendants() \\\n .values_list('pk', flat=True)\n Category.objects.filter(pk__in=unpublished_descendants_pks) \\\n .distinct().update(published=False)\n self.__class__.objects.rebuild()\n\n def __str__(self):\n return ' > '.join(\n a.title for a in self.get_ancestors(include_self=True)\n )\n\n def get_subcategories(self):\n return self.get_children().filter(published=True)\n\n def get_products(self):\n return self.products.filter(published=True)\n\n def get_product_attributes(self):\n return self.product_attributes.filter(published=True)\n\n\nclass TaxRate(models.Model):\n title = models.CharField(_('title'), max_length=255)\n rate = models.DecimalField(\n _('rate [%]'), max_digits=16, decimal_places=4,\n validators=[\n MinValueValidator(Decimal(0)),\n MaxValueValidator(Decimal(100))\n ]\n )\n\n class Meta:\n app_label = 'shop'\n ordering = ('title',)\n verbose_name = _('tax rate')\n verbose_name_plural = _('tax rates')\n\n def __str__(self):\n return '{} ({}%)'.format(self.title, self.rate)\n\n\nclass Product(models.Model):\n title = models.CharField(_('title'), max_length=255)\n slug = AutoSlugField(\n _('slug'), populate_from='title',\n blank=True, editable=True, unique=True\n )\n order = models.PositiveSmallIntegerField(_('order'), default=0)\n published = models.BooleanField(_('published'), default=True)\n featured = models.BooleanField(_('featured'), default=False)\n sale = models.BooleanField(_('on sale'), default=False)\n category = models.ForeignKey(\n Category, verbose_name=_('category'), related_name='products'\n )\n\n description = models.TextField(\n _('description'), default='', blank=True\n )\n\n attributes = models.ManyToManyField(\n 'ProductAttribute', verbose_name=_('attributes'), blank=True,\n through='ProductAttributeValue'\n )\n\n thumbnail = FileBrowseField(\n _('thumbnail'), max_length=255, blank=True, null=True,\n directory='shop/products', extensions=['.jpg', '.jpeg', '.gif', '.png']\n )\n\n net_price = models.DecimalField(\n _('net price'), max_digits=14, decimal_places=2, blank=True,\n validators=[MinValueValidator(Decimal(0))]\n )\n tax_rate = models.ForeignKey(\n TaxRate, verbose_name=_('tax rate'),\n help_text=_('Leaving this field empty means no tax is applied to net '\n 'price.'),\n null=True, blank=True\n )\n gross_price = models.DecimalField(\n _('gross price'), max_digits=14, decimal_places=2, blank=True,\n validators=[MinValueValidator(Decimal(0))]\n )\n\n stock = models.PositiveIntegerField(_('stock'), default=0, blank=True)\n infinite_stock = models.BooleanField(\n _('infinite stock'),\n help_text=_('Setting this option will make product available to '\n 'purchase regardless of the stock.'),\n default=False\n )\n\n sales_count = models.PositiveIntegerField(\n _('sales count'), default=0, blank=True\n )\n\n class Meta:\n app_label = 'shop'\n ordering = ('order',)\n verbose_name = _('product')\n verbose_name_plural = _('products')\n\n def __str__(self):\n return self.title\n\n def as_dict(self):\n try:\n image = self.get_first_image_or_thumbnail() \\\n .version_generate('shop_thumbnail').url\n except (AttributeError, FileNotFoundError):\n image = ''\n data = {\n 'pk': self.pk,\n 'url': self.get_url(),\n 'title': self.title,\n 'image': image,\n 'net_price': float(self.net_price),\n 'tax_rate': float(self.get_tax_rate()),\n 'gross_price': float(self.gross_price)\n }\n return data\n\n def get_url(self):\n return reverse('shop:product', kwargs={\n 'category_slug': self.category.slug,\n 'product_slug': self.slug\n })\n\n def get_tax_rate(self):\n try:\n return self.tax_rate.rate\n except AttributeError:\n return 0\n\n def get_images(self):\n return self.images.filter(published=True)\n\n def get_first_image_or_thumbnail(self):\n try:\n return self.thumbnail or self.get_images()[0].image\n except IndexError:\n return None\n\n def is_available(self):\n return bool(self.stock) or self.infinite_stock\n\n def get_categories_breadcrumbs(self):\n return self.category.get_ancestors(include_self=True)\n\n def get_attributes(self):\n return ProductAttributeValue.objects.filter(\n product=self,\n attribute__published=True,\n attribute__categories=self.category\n )\n\n\nclass ProductImage(models.Model):\n product = models.ForeignKey(Product, related_name='images')\n image = FileBrowseField(\n _('image'), max_length=255, directory='shop/products',\n extensions=['.jpg', '.jpeg', '.gif', '.png']\n )\n published = models.BooleanField(_('published'), default=True)\n title = models.CharField(\n _('title'), max_length=255, default='', blank=True\n )\n order = models.PositiveSmallIntegerField(_('order'), default=0)\n\n class Meta:\n app_label = 'shop'\n ordering = ('order',)\n verbose_name = _('image')\n verbose_name_plural = _('images')\n\n\nclass ProductAttribute(models.Model):\n name = models.CharField(_('name'), max_length=255)\n slug = AutoSlugField(\n _('slug'), populate_from='name',\n blank=True, editable=True, unique=True\n )\n published = models.BooleanField(_('published'), default=True)\n order = models.PositiveSmallIntegerField(_('order'), default=0)\n value_type = models.CharField(\n _('value type'), max_length=255,\n choices=PRODUCT_ATTRIBUTE_VALUE_TYPE_CHOICES, default='options'\n )\n categories = models.ManyToManyField(\n Category, verbose_name=_('categories'), blank=True,\n related_name='product_attributes'\n )\n\n class Meta:\n app_label = 'shop'\n ordering = ('order',)\n verbose_name = _('product attribute')\n verbose_name_plural = _('product attributes')\n\n def __str__(self):\n return self.name\n\n def get_options(self):\n return self.options.all()\n\n\nclass ProductAttributeOption(models.Model):\n attribute = models.ForeignKey(\n ProductAttribute, verbose_name=_('attribute'), related_name='options'\n )\n value = models.CharField(_('value'), max_length=255)\n order = models.PositiveSmallIntegerField(_('order'), default=0)\n\n class Meta:\n app_label = 'shop'\n ordering = ('order',)\n verbose_name = _('option')\n verbose_name_plural = _('options')\n\n def __str__(self):\n return self.value\n\n\nclass ProductAttributeValue(models.Model):\n product = models.ForeignKey(Product)\n attribute = models.ForeignKey(ProductAttribute)\n option = models.ForeignKey(ProductAttributeOption, null=True, blank=True)\n text_value = models.CharField(\n _('text value'), max_length=255, null=True, blank=True\n )\n order = models.PositiveSmallIntegerField(_('order'), default=0)\n\n class Meta:\n app_label = 'shop'\n ordering = ('order',)\n verbose_name = _('attribute')\n verbose_name_plural = _('attributes')\n\n def __str__(self):\n return self.attribute.name\n\n def save(self, *args, **kwargs):\n if self.attribute.value_type == 'options':\n self.text_value = None\n elif self.attribute.value_type == 'text':\n self.option = None\n super(ProductAttributeValue, self).save(*args, **kwargs)\n\n def get_value(self):\n if self.attribute.value_type == 'options':\n return self.option.value\n elif self.attribute.value_type == 'text':\n return self.text_value\n\n\nclass DeliveryMethod(models.Model):\n title = models.CharField(\n _('title'), max_length=255, default='', blank=True\n )\n published = models.BooleanField(_('published'), default=True)\n order = models.PositiveSmallIntegerField(_('order'), default=0)\n price = models.DecimalField(\n _('price'), max_digits=14, decimal_places=2,\n validators=[MinValueValidator(Decimal(0))]\n )\n involves_shipping = models.BooleanField(\n _('involves shipping'), default=True\n )\n\n class Meta:\n app_label = 'shop'\n ordering = ('order',)\n verbose_name = _('delivery method')\n verbose_name_plural = _('delivery methods')\n\n def __str__(self):\n return self.title\n\n def get_payment_methods(self):\n return self.payment_methods.filter(published=True)\n\n def as_dict(self):\n data = {\n 'pk': self.pk,\n 'title': self.title,\n 'price': float(self.price),\n 'involves_shipping': self.involves_shipping,\n 'payment_methods': [\n p.as_dict() for p in self.get_payment_methods()\n ]\n }\n return data\n\n\nclass PaymentMethod(models.Model):\n title = models.CharField(\n _('title'), max_length=255, default='', blank=True\n )\n published = models.BooleanField(_('published'), default=True)\n order = models.PositiveSmallIntegerField(_('shop order'), default=0)\n price = models.DecimalField(\n _('price'), max_digits=14, decimal_places=2,\n validators=[MinValueValidator(Decimal(0))]\n )\n delivery_methods = models.ManyToManyField(\n DeliveryMethod, verbose_name=_('delivery methods'),\n related_name='payment_methods'\n )\n gateway = models.CharField(\n _('gateway'), max_length=255,\n choices=PAYMENT_GATEWAYS_CHOICES, null=True, blank=True\n )\n wait_for_payment = models.BooleanField(_('wait for payment'), default=True)\n\n class Meta:\n app_label = 'shop'\n ordering = ('order',)\n verbose_name = _('payment method')\n verbose_name_plural = _('payment methods')\n\n def __str__(self):\n return self.title\n\n def get_shipment_methods(self):\n return self.shipment_methods.filter(published=True)\n\n def as_dict(self):\n data = {\n 'pk': self.pk,\n 'title': self.title,\n 'price': float(self.price),\n 'gateway': self.gateway\n }\n return data\n\n def get_gateway_api_url(self):\n try:\n return shop_settings.SHOP_PAYMENT_GATEWAYS[self.gateway]['api_url']\n except AttributeError:\n return None\n\n\nclass Customer(models.Model):\n user = models.OneToOneField(\n settings.AUTH_USER_MODEL, verbose_name=_('user')\n )\n\n wishlist = models.ManyToManyField(\n Product, verbose_name=_('wishlist'), related_name='wishlists'\n )\n\n class Meta:\n app_label = 'shop'\n verbose_name = _('customer')\n verbose_name_plural = _('customers')\n\n def __str__(self):\n return self.user.username\n\n def get_wishlist(self):\n return self.wishlist.filter(\n published=True,\n category__published=True\n )\n\n def get_orders(self):\n return self.orders.all()\n\n\n@receiver(post_save, sender=settings.AUTH_USER_MODEL)\ndef create_user_hook(sender, instance, created, **kwargs):\n if created:\n Customer.objects.get_or_create(user=instance)\n\n\n@receiver(post_delete, sender=Customer)\ndef delete_user_hook(sender, instance, **kwargs):\n instance.user.delete()\n\n\nclass ShopOrder(models.Model):\n order_id = models.UUIDField(\n _('order id'), default=uuid.uuid4, editable=False\n )\n created = models.DateTimeField(\n _('creation date'), auto_now_add=True\n )\n customer = models.ForeignKey(\n Customer, verbose_name=_('customer'), related_name='orders'\n )\n delivery_method = models.ForeignKey(\n DeliveryMethod, verbose_name=_('delivery method')\n )\n payment_method = models.ForeignKey(\n PaymentMethod, verbose_name=_('payment method')\n )\n delivery_address = models.TextField(\n _('delivery address'), default='', blank=True\n )\n invoice_tax_id = models.CharField(\n _('invoice - tax ID'), max_length=255, blank=True, default=''\n )\n invoice_company = models.TextField(\n _('invoice - company name and address'),\n blank=True, default=''\n )\n additional_info = models.TextField(\n _('additional info'), default='', blank=True\n )\n\n status = models.CharField(\n _('status'), max_length=255,\n choices=shop_settings.SHOP_ORDER_STATUSES,\n )\n\n class Meta:\n app_label = 'shop'\n ordering = ('-created',)\n verbose_name = _('shop order')\n verbose_name_plural = _('shop orders')\n\n def __str__(self):\n return str(self.order_id)\n\n def get_products(self):\n return self.products.all()\n\n def get_prices(self):\n products = self.get_products().aggregate(\n total=Sum(\n F('gross_price') * F('quantity'),\n output_field=models.DecimalField()\n )\n )\n delivery_price = self.delivery_method.price\n payment_price = self.payment_method.price\n total = (products['total'] or Decimal(0) +\n delivery_price +\n payment_price)\n return {\n 'products': products['total'],\n 'delivery': delivery_price,\n 'payment': payment_price,\n 'total': total\n }\n\n @property\n def invoice(self):\n return bool(self.invoice_tax_id or self.invoice_company)\n\n\nclass ShopOrderProduct(models.Model):\n shop_order = models.ForeignKey(\n ShopOrder, verbose_name=_('shop order'), related_name='products'\n )\n product = models.ForeignKey(Product, verbose_name=_('product'))\n net_price = models.DecimalField(\n _('net price'), max_digits=14, decimal_places=2,\n validators=[MinValueValidator(Decimal(0))]\n )\n tax_rate = models.ForeignKey(\n TaxRate, verbose_name=_('tax rate'), null=True, blank=True\n )\n gross_price = models.DecimalField(\n _('gross price'), max_digits=14, decimal_places=2,\n validators=[MinValueValidator(Decimal(0))]\n )\n quantity = models.PositiveIntegerField(_('quantity'))\n\n class Meta:\n app_label = 'shop'\n verbose_name = _('product')\n verbose_name_plural = _('products')\n\n def __str__(self):\n return self.product.title\n\n def get_subtotal(self):\n return self.gross_price * self.quantity\n\n\nclass ShopOrderPayment(PolymorphicModel):\n shop_order = models.ForeignKey(\n ShopOrder, verbose_name=_('shop order'), related_name='payments'\n )\n\n class Meta:\n app_label = 'shop'\n verbose_name = _('payment')\n verbose_name_plural = _('payments')\n\n\nclass GenericPayment(ShopOrderPayment):\n amount = models.DecimalField(\n _('amount'), max_digits=14, decimal_places=2,\n validators=[MinValueValidator(Decimal(0))]\n )\n date = models.DateTimeField(_('date'))\n\n class Meta:\n app_label = 'shop'\n verbose_name = _('payment')\n verbose_name_plural = _('payments')\n\n def get_amount(self):\n return self.amount\n\n def get_date(self):\n return self.date\n","sub_path":"shop/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":17413,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"465671458","text":"import asyncio\nimport websockets\nimport serial\nimport time\nfrom pathlib import Path\n\nTIME_FORMAT = \"[%Y-%m-%d %H:%M:%S]\"\n\nclass WSServer:\n def __init__(self, queue, base_port):\n self.message_queue = queue\n self.base_port = base_port\n\n async def on_receive(self, websocket, path):\n server_port = websocket.local_address[1]\n try:\n async for message in websocket:\n print(time.strftime(TIME_FORMAT), server_port, \"received from\", websocket.remote_address[0], \"0x\" + message.hex())\n\n # Append the received message to the queue\n await self.message_queue.put((server_port, message))\n\n await websocket.send(message)\n except Exception as e:\n print(time.strftime(TIME_FORMAT), server_port, \"Unexpected connection close\", )\n print(time.strftime(TIME_FORMAT), server_port, \"Issuing disconnect\", )\n await self.message_queue.put((server_port, (server_port - self.base_port).to_bytes(1, 'little') + b\"\\x01\\x00\\x00\"))\n\n def start(self, num_controllers):\n for i in range(num_controllers):\n print(time.strftime(TIME_FORMAT), \"Starting server on port\", i + self.base_port)\n asyncio.get_event_loop().run_until_complete(websockets.serve(self.on_receive, \"\", i + self.base_port))\n print(time.strftime(TIME_FORMAT), \"Server started\")\n\n# Consumes arriving messages queue. This serializes message handling independent\n# of the server threads, to avoid races over the serial output\nasync def queue_handler(queue, serial_port, baud=115200):\n print(time.strftime(TIME_FORMAT), \"queue_handler started\")\n with serial.Serial(serial_port, 115200, timeout=1) as ser:\n while True:\n index, message = await queue.get()\n # print(\"Dequeued message\", message, \"for index\", index)\n ser.write(message)\n queue.task_done()\n\nasync def dummy_queue_handler(queue):\n print(time.strftime(TIME_FORMAT), \"queue_handler started\")\n while True:\n index, message = await queue.get()\n print(time.strftime(TIME_FORMAT), \"Dequeued message\", \"0x\" + message.hex(), \"for index\", index)\n queue.task_done()\n\nif __name__ == '__main__':\n import argparse\n\n parser = argparse.ArgumentParser(description='Run server for Remote USB Gamepad bridge')\n parser.add_argument('--num-controllers', type=int, default=2, help='Number of controllers to listen for')\n parser.add_argument('--base-port', '-p', type=int, default=8000, help='Base port number for 0th controller. Other controllers will have servers attached to port base+id')\n parser.add_argument('--baud', type=int, default=115200, help='Serial port baud rate')\n parser.add_argument('--dummy-serial', action=\"store_true\", help='Use dummy handler instead of real serial port')\n parser.add_argument('serial_port', help=\"Serial port to output commands\")\n\n args = parser.parse_args()\n\n queue = asyncio.Queue()\n \n server = WSServer(queue, args.base_port)\n server.start(args.num_controllers)\n\n if args.dummy_serial:\n asyncio.get_event_loop().run_until_complete(dummy_queue_handler(queue))\n else:\n asyncio.get_event_loop().run_until_complete(queue_handler(queue, args.serial_port, args.baud))\n asyncio.get_event_loop().run_forever()\n","sub_path":"server/gamepad_bridge_server.py","file_name":"gamepad_bridge_server.py","file_ext":"py","file_size_in_byte":3333,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"636634538","text":"# -*- coding: utf-8 -*-\r\n\r\nimport pandas as pd\r\n \r\n \r\n# 需要读取广告操作数据集中的广告ID并将其转化成list\r\nAd_operation = pd.read_csv('/cos_person/tencent/train/Ad_Operation_Data.csv')\r\nAd_op_id = Ad_operation['ad_id'].drop_duplicates(keep='first', inplace=False)\r\nlist_Ad_op_id = list(Ad_op_id)\r\n\r\nuser_feature = pd.read_csv('/cos_person/tencent/train/userFeature.csv')\r\nuser_id = user_feature['user_id'].drop_duplicates(keep='first', inplace=False)\r\nlist_user_id = list(user_id)\r\n \r\n# 定义曝光日志中的相关列\r\nExposure_Log_Data = []\r\n\r\nfor j in range(10,23):\r\n with open('/cos_public/cephfs/tesla_common/deeplearning/dataset/AI_Race/track_log/track_log_201904' + str(j) + '.out', 'r') as f:\r\n for i, line in enumerate(f):\r\n line = line.strip().split('\\t')\r\n \r\n flag_line = line\r\n \r\n if (i % 5000000) == 0:\r\n print(\"*******run \", i)\r\n \r\n if line[0] == '0' or line[1] == '0' or line[2] == '0' or line[3] == '0' or line[4] == '0':\r\n continue\r\n \r\n if ',' in line[2]:\r\n continue\r\n \r\n tmp_user_id = int(line[2]) ##不在用户特征集中的曝光\r\n if tmp_user_id not in list_user_id:\r\n continue\r\n \r\n if '.' in line[0]:\r\n continue\r\n if '.' in line[3]:\r\n continue\r\n \r\n ad_list = []\r\n ad_list_columns = ['ad_id', 'bid', 'pctr', 'quality_ecpm', 'totalEcpm', \r\n 'filter', 'label']\r\n ad_list.append(ad_list_columns)\r\n \r\n tmp_line = line[4].strip().split(';')\r\n \r\n for each in tmp_line:\r\n save_line = []\r\n each_list = each.split(',')\r\n if each_list[6] != '1':\r\n continue\r\n else:\r\n if int(each_list[0]) not in list_Ad_op_id:\r\n continue;\r\n else: \r\n line.append(int(each_list[0]))\r\n line.append(int(each_list[1]))\r\n line.append(float(each_list[2]))\r\n line.append(float(each_list[3]))\r\n line.append(float(each_list[4]))\r\n line.append(int(each_list[5]))\r\n line.append(int(each_list[6]))\r\n save_line.append(line[5])\r\n if save_line:\r\n Exposure_Log_Data.append(line)\r\n line = flag_line\r\n \r\nExposure_Log_Data = pd.DataFrame(Exposure_Log_Data) \r\nExposure_Log_Data_columns = ['Ad_Request_id', 'Ad_Request_Time','user_id',\r\n 'Ad_pos_id', 'ad_list',\r\n 'ad_id','ad_bid','pctr',\r\n 'quality_ecpm', 'totalEcpm',\r\n 'filter', 'label']\r\nExposure_Log_Data.columns = Exposure_Log_Data_columns \r\nExposure_Log_Data.to_csv('/cos_person/tencent/train/Total_Exposure_Log_Data_with_AD_list.csv', index=False,header=None)\r\n\r\nExposure_Log_Data.drop(Exposure_Log_Data.columns[['ad_list']], axis=1,inplace=True)\r\nExposure_Log_Data.to_csv('/cos_person/tencent/train/Train_Log_Data.csv', index=False,header=None)","sub_path":"(修改)处理曝光文件.py","file_name":"(修改)处理曝光文件.py","file_ext":"py","file_size_in_byte":3385,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"511286154","text":"\nclass ECSystemParameters:\n \"\"\"\n This class governs all the parameters needed to run our EC System.\n Initially set to invalid values to prevent system for running.\n\n Generation size => How many Individuals do we want in each generation?\n Genome size => We can cap the length for how long we want our initial\n population's expressions to be\n X-Training Data => The x value we will plug into our random expressions\n Y-Training Data => The y values we will match our output to to determine fitness\n Fitness Threshold => What percentage of the population will be selected to\n go on to the next generation\n Stagnation Threshold => If our fitness is not improving overall over this set\n number of generations, we reboot the system and start over\n Mutation Percentage => Of the Individuals selected for the next generation, what\n percentage will we mutate instead of crossover\n Success Threshold => Determines when we have found an equivalent expression. Their\n fitness is at or below this value.\n \"\"\"\n\n def __init__(self):\n self.generation_size = 0\n self.genome_size = 0\n self.x_training_data = []\n self.y_training_data = []\n self.fitness_threshold = -1.0\n self.stagnation_threshold = -1\n self.mutation_percentage = -1.0\n self.success_threshold = -1.0\n\n def all_parameters_set(self):\n \"\"\"\n Make sure that all values are appropriately set\n :return: Boolean true if system is ready, false otherwise\n \"\"\"\n return self.generation_size > 49 and self.genome_size > 4 and self._valid_training_data() \\\n and self.fitness_threshold > 0.0 and self.stagnation_threshold > 0 \\\n and self.mutation_percentage >= 0.0 and self.success_threshold >= 0.0\n\n def _valid_training_data(self):\n \"\"\"\n Make sure our training data exists and that we have a x for every y\n :return: Boolean true if valid data, false othewise\n \"\"\"\n return len(self.x_training_data) > 0 and len(self.x_training_data) == len(self.y_training_data)\n","sub_path":"src/ecSystem/ECSystemParameters.py","file_name":"ECSystemParameters.py","file_ext":"py","file_size_in_byte":2212,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"119699365","text":"def pokerChips2(chips):\n memory = {}\n total = len(chips)\n return count_poker_moves(chips, 0, sum(chips) / len(chips), total, memory)\n\ndef count_poker_moves(chips, moves, average, total, memory):\n str_value = ''\n max_value = max(chips)\n position = chips.index(max_value)\n new_chips = chips[position:position+1] + chips[position+1:] + chips[:position]\n\n for value in map(lambda x: str(x), new_chips):\n str_value += value + ','\n\n if str_value in memory:\n return memory[str_value]\n\n if max_value == average or moves == total:\n return moves\n\n new_array = new_chips[1:]\n new_array2 = new_array[:]\n new_array[0] += max_value - average\n new_array2[len(new_array2)-1] += max_value - average\n count1 = count_poker_moves(new_array, moves, average, total, memory)\n count2 = count_poker_moves(new_array2, moves, average, total, memory)\n\n if count1 < count2:\n total_moves = moves + count1 + 1\n else:\n total_moves = moves + count2 + 1\n memory[str_value] = total_moves\n return total_moves\n\nprint(pokerChips2([18, 22, 30, 21, 2, 20, 22, 8, 30, 30, 7, 23, 1, 22, 8, 23, 7, 22, 25, 26, 17, 30, 27, 6, 25, 29, 20, 9, 3, 25, 16, 16, 30, 30, 8, 15, 27, 25, 6, 22, 16, 10, 24, 14, 26, 0, 13, 28, 11, 5]))\n","sub_path":"old/poker_chips2b.py","file_name":"poker_chips2b.py","file_ext":"py","file_size_in_byte":1277,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"526592022","text":"import sys, itertools, operator, inspect\nimport numpy as np\nimport gym\n\nfrom gym_minigrid.minigrid import Grid, MiniGridEnv\nfrom gym_minigrid.roomgrid import RoomGrid\nimport gym_minigrid.entities as entities\nfrom gym_minigrid.entities import Goal, Wall, Door, Key, Ball, Box, Lava, COLORS, OBJECTS\n\n\nclass Empty(MiniGridEnv):\n \"\"\"\n This environment is an empty room, and the goal of the agent is to reach the green goal square, which provides a sparse reward. A small penalty is subtracted for the number of steps to reach the goal. This environment is useful, with small rooms, to validate that your RL algorithm works correctly, and with large rooms to experiment with sparse rewards and exploration. The random variants of the environment have the agent starting at a random position for each episode, while the regular variants have the agent always starting in the corner opposite to the goal.\n \"\"\"\n\n def __init__(\n self,\n size=8,\n agent_start_pos=(1,1),\n agent_start_state='right',\n max_steps=None,\n **kwargs\n ):\n self.agent_start_pos = agent_start_pos\n self.agent_start_state = agent_start_state\n\n super().__init__(\n height=size,\n width=size,\n max_steps=4 * size**2 if max_steps is None else max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n # Create an empty grid\n self.grid = Grid(height, width)\n\n # Generate the surrounding walls\n self.wall_rect(0, 0, height, width)\n\n # Place a goal square in the bottom-right corner\n self[height - 2, width - 2] = Goal()\n\n # Place the agent\n self.agent.pos = self.agent_start_pos\n self.agent.state = self.agent_start_state\n\n self.mission = 'get to the green goal square'\n\n\nclass FourRooms(MiniGridEnv):\n \"\"\"\n Classic four room reinforcement learning environment. The agent must navigate in a maze composed of four rooms interconnected by 4 gaps in the walls. To obtain a reward, the agent must reach the green goal square. Both the agent and the goal square are randomly placed in any of the four rooms.\n \"\"\"\n\n def __init__(self, agent_pos=None, goal_pos=None, max_steps=None, **kwargs):\n self._agent_default_pos = agent_pos\n self._goal_default_pos = goal_pos\n super().__init__(\n height=19, \n width=19, \n max_steps=100 if max_steps is None else max_steps, \n **kwargs)\n\n def _gen_grid(self, height, width):\n # Create the grid\n self.grid = Grid(height, width)\n\n # Generate the surrounding walls\n self.horz_wall(0, 0)\n self.horz_wall(height - 1, 0)\n self.vert_wall(0, 0)\n self.vert_wall(0, width - 1)\n\n room_w = width // 2\n room_h = height // 2\n\n # For each row of rooms\n for i in range(0, 2):\n\n # For each column\n for j in range(0, 2):\n i_top = i * room_h\n j_left = j * room_w\n i_bottom = i_top + room_h\n j_right = j_left + room_w\n\n # Right wall and door\n if j + 1 < 2:\n self.vert_wall(i_top, j_right, room_h)\n pos = (self.rng.randint(i_top + 1, i_bottom), j_right)\n self[pos].clear()\n\n # Bottom wall and door\n if i + 1 < 2:\n self.horz_wall(i_bottom, j_left, room_w)\n pos = (i_bottom, self.rng.randint(j_left + 1, j_right))\n self[pos].clear()\n\n # Randomize the player start position and orientation\n if self._agent_default_pos is not None:\n self.agent.pos = self._agent_default_pos\n self.agent.state = self.rng.choice(self.agent.STATES) # assuming random start direction\n else:\n self.place_agent()\n\n if self._goal_default_pos is not None:\n self[self._goal_default_pos] = Goal()\n else:\n self.place_obj(Goal())\n\n self.mission = 'Reach the goal'\n\n\nclass DoorKey(MiniGridEnv):\n \"\"\"\n This environment has a key that the agent must pick up in order to unlock a goal and then get to the green goal square. This environment is difficult, because of the sparse reward, to solve using classical RL algorithms. It is useful to experiment with curiosity or curriculum learning.\n \"\"\"\n\n def __init__(self, size=8, max_steps=None, **kwargs):\n super().__init__(\n height=size,\n width=size,\n max_steps=10 * size * size if max_steps is None else max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n # Create an empty grid\n self.grid = Grid(height, width)\n\n # Generate the surrounding walls\n self.wall_rect(0, 0, height, width)\n\n # Place a goal in the bottom-right corner\n self[height - 2, width - 2] = Goal()\n\n # Create a vertical splitting wall\n split_idx = self.rng.randint(2, width - 2)\n self.vert_wall(0, split_idx)\n\n # Place the agent at a random position and orientation\n # on the left side of the splitting wall\n self.place_agent(size=(height, split_idx))\n\n # Place a door in the wall\n door_idx = self.rng.randint(1, height - 2)\n self[door_idx, split_idx] = Door('yellow', state='locked')\n\n # Place a yellow key on the left side\n self.place_obj(Key('yellow'), top=(0, 0), size=(height, split_idx))\n\n self.mission = 'use the key to open the door and then get to the goal'\n\n\nclass _MultiRoom(object):\n\n def __init__(self,\n top,\n size,\n entry_door_pos,\n exit_door_pos\n ):\n self.top = top\n self.size = size\n self.entry_door_pos = entry_door_pos\n self.exit_door_pos = exit_door_pos\n\n\nclass MultiRoom(MiniGridEnv):\n \"\"\"\n This environment has a series of connected rooms with doors that must be opened in order to get to the next room. The final room has the green goal square the agent must get to. This environment is extremely difficult to solve using RL alone. However, by gradually increasing the number of rooms and building a curriculum, the environment can be solved.\n \"\"\"\n\n def __init__(self,\n min_num_rooms=6,\n max_num_rooms=6,\n max_room_size=10,\n max_steps=None,\n **kwargs\n ):\n assert min_num_rooms > 0\n assert max_num_rooms >= min_num_rooms\n assert max_room_size >= 4\n\n self.min_num_rooms = min_num_rooms\n self.max_num_rooms = max_num_rooms\n self.max_room_size = max_room_size\n\n self.rooms = []\n\n super().__init__(\n height=25,\n width=25,\n max_steps=self.max_num_rooms * 20 if max_steps is None else max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n room_list = []\n\n # Choose a random number of rooms to generate\n num_rooms = self.rng.randint(self.min_num_rooms, self.max_num_rooms + 1)\n\n while len(room_list) < num_rooms:\n cur_room_list = []\n\n entry_door_pos = (\n self.rng.randint(0, height - 2),\n self.rng.randint(0, width - 2)\n )\n\n # Recursively place the rooms\n self._place_room(\n num_rooms,\n room_list=cur_room_list,\n min_sz=4,\n max_sz=self.max_room_size,\n entry_door_wall=2,\n entry_door_pos=entry_door_pos\n )\n\n if len(cur_room_list) > len(room_list):\n room_list = cur_room_list\n\n # Store the list of rooms in this environment\n assert len(room_list) > 0\n self.rooms = room_list\n\n # Create the grid\n self.grid = Grid(height, width)\n\n prev_door_color = None\n\n # For each room\n for idx, room in enumerate(room_list):\n\n top_i, top_j = room.top\n room_height, room_width = room.size\n\n # Generate the surrounding walls\n self.horz_wall(top_i, top_j, width=room_width)\n self.horz_wall(top_i + room_height - 1, top_j, width=room_width)\n self.vert_wall(top_i, top_j, height=room_height)\n self.vert_wall(top_i, top_j + room_width - 1, height=room_height)\n\n # If this isn't the first room, place the entry door\n if idx > 0:\n # Pick a door color different from the previous one\n door_colors = set(COLORS)\n if prev_door_color:\n door_colors.remove(prev_door_color)\n # Note: the use of sorting here guarantees determinism,\n # This is needed because Python's set is not deterministic\n door_colors = self.rng.choice(sorted(door_colors))\n\n self[room.entry_door_pos] = Door(door_colors)\n prev_door_color = door_colors\n\n prev_room = room_list[idx - 1]\n prev_room.exit_door_pos = room.entry_door_pos\n\n # Randomize the starting agent position and direction\n self.place_agent(room_list[0].top, room_list[0].size)\n\n # Place the final goal in the last room\n self.goal_pos = self.place_obj(Goal(), room_list[-1].top, room_list[-1].size)\n\n self.mission = 'traverse the rooms to get to the goal'\n\n def _place_room(\n self,\n num_left,\n room_list,\n min_sz,\n max_sz,\n entry_door_wall,\n entry_door_pos\n ):\n # Choose the room size randomly\n size_i = self.rng.randint(min_sz, max_sz + 1)\n size_j = self.rng.randint(min_sz, max_sz + 1)\n\n # The first room will be at the door position\n if len(room_list) == 0:\n top_i, top_j = entry_door_pos\n # Entry on the right\n elif entry_door_wall == 0:\n i = entry_door_pos[0]\n top_i = self.rng.randint(i - size_i + 2, i)\n top_j = entry_door_pos[1] - size_j + 1\n # Entry wall on the bottom\n elif entry_door_wall == 1:\n top_i = entry_door_pos[0] - size_i + 1\n j = entry_door_pos[1]\n top_j = self.rng.randint(j - size_j + 2, j)\n # Entry wall on the left\n elif entry_door_wall == 2:\n i = entry_door_pos[0]\n top_i = self.rng.randint(i - size_i + 2, i)\n top_j = entry_door_pos[1]\n # Entry wall on the top\n elif entry_door_wall == 3:\n top_i = entry_door_pos[0]\n j = entry_door_pos[1]\n top_j = self.rng.randint(j - size_j + 2, j)\n else:\n raise ValueError(f'Entry door wall index wrong: {entry_door_wall}')\n\n # If the room is out of the grid, can't place a room here\n if top_i < 0 or top_j < 0:\n return False\n if top_i + size_i >= self.height or top_j + size_j > self.width:\n return False\n\n # If the room intersects with previous rooms, can't place it here\n for room in room_list[:-1]:\n non_overlap = \\\n top_i + size_i < room.top[0] or \\\n room.top[0] + room.size[0] <= top_i or \\\n top_j + size_j < room.top[1] or \\\n room.top[1] + room.size[1] <= top_j\n\n if not non_overlap:\n return False\n\n # Add this room to the list\n room_list.append(_MultiRoom(\n (top_i, top_j),\n (size_i, size_j),\n entry_door_pos,\n None\n ))\n\n # If this was the last room, stop\n if num_left == 1:\n return True\n\n # Try placing the next room\n for i in range(0, 8):\n\n # Pick which wall to place the out door on\n wall_set = set((0, 1, 2, 3))\n wall_set.remove(entry_door_wall)\n exit_door_wall = self.rng.choice(sorted(wall_set))\n next_entry_wall = (exit_door_wall + 2) % 4\n\n # Pick the exit door position\n # Exit on right wall\n if exit_door_wall == 0:\n exit_door_pos = (\n top_i + self.rng.randint(1, size_i - 1),\n top_j + size_j - 1\n )\n # Exit on bottom wall\n elif exit_door_wall == 1:\n exit_door_pos = (\n top_i + size_i - 1,\n top_j + self.rng.randint(1, size_j - 1)\n )\n # Exit on left wall\n elif exit_door_wall == 2:\n exit_door_pos = (\n top_i + self.rng.randint(1, size_i - 1),\n top_j\n )\n # Exit on top wall\n elif exit_door_wall == 3:\n exit_door_pos = (\n top_i,\n top_j + self.rng.randint(1, size_j - 1)\n )\n else:\n raise ValueError\n\n # Recursively create the other rooms\n success = self._place_room(\n num_left - 1,\n room_list=room_list,\n min_sz=min_sz,\n max_sz=max_sz,\n entry_door_wall=next_entry_wall,\n entry_door_pos=exit_door_pos\n )\n\n if success:\n break\n\n return True\n\n\nclass Fetch(MiniGridEnv):\n \"\"\"\n This environment has multiple objects of assorted types and colors. The agent receives a textual string as part of its observation telling it which object to pick up. Picking up the wrong object produces a negative reward.\n \"\"\"\n\n def __init__(\n self,\n size=8,\n num_objs=3,\n max_steps=None,\n **kwargs\n ):\n self.num_objs = num_objs\n\n super().__init__(\n height=size,\n width=size,\n max_steps=5 * size**2 if max_steps is None else max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n self.grid = Grid(height, width)\n\n # Generate the surrounding walls\n self.horz_wall(0, 0)\n self.horz_wall(height - 1, 0)\n self.vert_wall(0, 0)\n self.vert_wall(0, width - 1)\n\n types = ['key', 'ball']\n\n objs = []\n\n # For each object to be generated\n while len(objs) < self.num_objs:\n obj_type = self.rng.choice(types)\n obj_color = self.rng.choice(COLORS)\n\n if obj_type == 'key':\n obj = Key(obj_color)\n elif obj_type == 'ball':\n obj = Ball(obj_color)\n\n self.place_obj(obj)\n objs.append(obj)\n\n # Randomize the player start position and orientation\n self.place_agent()\n\n # Choose a random object to be picked up\n target = objs[self.rng.randint(0, len(objs))]\n self.target_type = target.type\n self.target_color = target.color\n\n # Generate the mission string\n missions = ['get a', 'go get a', 'fetch a', 'go fetch a', 'you must fetch a']\n self.mission = self.rng.choice(missions) + f' {self.target_color} {self.target_type}'\n\n def step(self, action):\n obs, reward, done, info = super().step(action)\n\n if self.agent.is_carrying:\n if self.agent.carrying.color == self.target_color and \\\n self.agent.carrying.type == self.target_type:\n reward = self._win_reward\n else:\n reward = self._lose_reward\n done = True\n\n return obs, reward, done, info\n\n\nclass GoToObject(MiniGridEnv):\n \"\"\"\n This environment is a room with four doors, one on each wall. The agent receives a textual (mission) string as input, telling it which door to go to, (eg: \"go to the red door\"). It receives a positive reward for performing the `done` action next to the correct door, as indicated in the mission string. (BUG: doesn't look like the mission had that indicated)\n \"\"\"\n\n def __init__(self, size=6, num_objs=2, max_steps=None, **kwargs):\n self.num_objs = num_objs\n\n super().__init__(\n height=size,\n width=size,\n max_steps=5 * size**2 if max_steps is None else max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n self.grid = Grid(height, width)\n\n # Generate the surrounding walls\n self.wall_rect(0, 0, height, width)\n\n # Types and colors of objects we can generate\n types = ['key', 'ball', 'box']\n\n objs = []\n # Until we have generated all the objects\n while len(objs) < self.num_objs:\n obj_type = self.rng.choice(types)\n obj_color = self.rng.choice(COLORS)\n\n # If this object already exists, try again\n if (obj_type, obj_color) in objs:\n continue\n\n if obj_type == 'key':\n obj = Key(obj_color)\n elif obj_type == 'ball':\n obj = Ball(obj_color)\n elif obj_type == 'box':\n obj = Box(obj_color)\n\n self.place_obj(obj)\n objs.append(obj)\n\n # Randomize the agent start position and orientation\n self.place_agent()\n\n # Choose a random object to be picked up\n self.target = self.rng.choice(objs)\n\n self.mission = f'go to the {self.target.color} {self.target.type}'\n\n def step(self, action):\n obs, reward, done, info = super().step(action)\n\n # Toggle/pickup action terminates the episode\n if self.actions[action] == 'toggle':\n reward = self._lose_reward\n done = True\n\n # Reward performing the done action next to the target object\n ai, aj = self.agent.pos\n ti, tj = self.target.pos\n if self.actions[action] == 'done':\n reward = self._lose_reward\n if abs(ai - ti) <= 1 and abs(aj - tj) <= 1:\n reward = self._win_reward\n done = True\n\n return obs, reward, done, info\n\n\nclass GoToDoor(MiniGridEnv):\n \"\"\"\n This environment is a room with four doors, one on each wall. The agent receives a textual (mission) string as input, telling it which door to go to, (eg: \"go to the red door\"). It receives a positive reward for performing the `done` action next to the correct door, as indicated in the mission string.\n \"\"\"\n\n def __init__(self, size=5, max_steps=None, **kwargs):\n assert size >= 5\n\n super().__init__(\n height=size,\n width=size,\n max_steps=5 * size**2 if max_steps is None else max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n # Create the grid\n self.grid = Grid(height, width)\n\n # Randomly vary the room width and height\n height = self.rng.randint(5, height + 1)\n width = self.rng.randint(5, width + 1)\n\n # Generate the surrounding walls\n self.wall_rect(0, 0, height, width)\n\n # Generate the 4 doors at random positions\n door_pos = [(0, self.rng.randint(2, width - 2)),\n (height - 1, self.rng.randint(2, width - 2)),\n (self.rng.randint(2, height - 2), 0),\n (self.rng.randint(2, height - 2), width - 1)]\n\n # Generate the door colors\n door_colors = self.rng.choice(COLORS, size=len(door_pos), replace=False)\n\n # Place the doors in the grid\n for idx, pos in enumerate(door_pos):\n color = door_colors[idx]\n self[pos] = Door(color)\n\n # Randomize the agent start position and orientation\n self.place_agent()\n\n # Select a random target door\n door_idx = self.rng.randint(0, len(door_pos))\n self.target_pos = door_pos[door_idx]\n self.target_color = door_colors[door_idx]\n\n # Generate the mission string\n self.mission = f'go to the {self.target_color} door'\n\n def step(self, action):\n obs, reward, done, info = super().step(action)\n\n ai, aj = self.agent.pos\n ti, tj = self.target_pos\n\n # Don't let the agent open any of the doors\n if self.actions[action] == 'toggle':\n reward = self._lose_reward\n done = True\n\n # Reward performing done action in front of the target door\n if self.actions[action] == 'done':\n reward = self._lose_reward\n if (ai == ti and abs(aj - tj) == 1) or (aj == tj and abs(ai - ti) == 1):\n reward = self._win_reward\n done = True\n\n return obs, reward, done, info\n\n\nclass PutNear(MiniGridEnv):\n \"\"\"\n The agent is instructed through a textual string to pick up an object and place it next to another object. This environment is easy to solve with two objects, but difficult to solve with more, as it involves both textual understanding and spatial reasoning involving multiple objects.\n \"\"\"\n\n def __init__(\n self,\n size=6,\n num_objs=2,\n max_steps=None,\n **kwargs\n ):\n self.num_objs = num_objs\n\n super().__init__(\n height=size,\n width=size,\n max_steps=5 * size if max_steps is None else max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n self.grid = Grid(height, width)\n\n # Generate the surrounding walls\n self.horz_wall(0, 0)\n self.horz_wall(height - 1, 0)\n self.vert_wall(0, 0)\n self.vert_wall(0, width - 1)\n\n # Types and colors of objects we can generate\n types = ['key', 'ball', 'box']\n\n objs = []\n obj_pos = []\n\n def near_obj(env, p1):\n for p2 in obj_pos:\n di = p1[0] - p2[0]\n dj = p1[1] - p2[1]\n if abs(di) <= 1 and abs(dj) <= 1:\n return True\n return False\n\n # Until we have generated all the objects\n while len(objs) < self.num_objs:\n obj_type = self.rng.choice(types)\n obj_color = self.rng.choice(COLORS)\n\n # If this object already exists, try again\n if (obj_type, obj_color) in objs:\n continue\n\n if obj_type == 'key':\n obj = Key(obj_color)\n elif obj_type == 'ball':\n obj = Ball(obj_color)\n elif obj_type == 'box':\n obj = Box(obj_color)\n\n self.place_obj(obj, reject_fn=near_obj)\n\n objs.append((obj_type, obj_color))\n obj_pos.append(obj.pos)\n\n # Randomize the agent start position and orientation\n self.place_agent()\n\n # Choose a random object to be moved\n obj_idx = self.rng.randint(0, len(objs))\n self.move_type, self.move_color = objs[obj_idx]\n # self.move_pos = obj_pos[obj_idx]\n\n # Choose a target object (to put the first object next to)\n while True:\n targetIdx = self.rng.randint(0, len(objs))\n if targetIdx != obj_idx:\n break\n self.target_type, self.target_color = objs[targetIdx]\n self.target_pos = obj_pos[targetIdx]\n\n self.mission = (f'put the {self.move_color} {self.move_type} near '\n f'the {self.target_color} {self.target_type}')\n\n def step(self, action):\n pre_carrying = self.agent.carrying\n\n obs, reward, done, info = super().step(action)\n\n oi, oj = self.agent.front_pos\n ti, tj = self.target_pos\n\n # If we picked up the wrong object, terminate the episode\n if self.actions[action] == 'pickup' and self.agent.is_carrying:\n if self.agent.carrying.type != self.move_type or self.agent.carrying.color != self.move_color:\n reward = self._lose_reward\n done = True\n\n # If successfully dropping an object near the target\n if self.actions[action] == 'drop' and pre_carrying:\n reward = self._lose_reward\n if self[oj, oi] is pre_carrying:\n if abs(oi - ti) <= 1 and abs(oj - tj) <= 1:\n reward = self._win_reward\n done = True\n\n return obs, reward, done, info\n\n\nclass RedBlueDoor(MiniGridEnv):\n \"\"\"\n The purpose of this environment is to test memory. The agent is randomly placed within a room with one red and one blue door facing opposite directions. The agent has to open the red door and then open the blue door, in that order. The agent, when facing one door, cannot see the door behind him. Hence, the agent needs to remember whether or not he has previously opened the other door in order to reliably succeed at completing the task.\n \"\"\"\n\n def __init__(self, size=8, max_steps=None, **kwargs):\n self.size = size\n\n super().__init__(\n height=size,\n width=2 * size,\n max_steps=20 * size * size if max_steps is None else max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n # Create an empty grid\n self.grid = Grid(height, width)\n\n # Generate the grid walls\n self.wall_rect(0, 0, self.size, 2 * self.size)\n self.wall_rect(0, self.size // 2, self.size, self.size)\n\n # Place the agent in the top-left corner\n self.place_agent(top=(0, self.size // 2), size=(self.size, self.size))\n\n # Add a red door at a random position in the left wall\n pos = self.rng.randint(1, self.size - 1)\n self.red_door = Door('red')\n self[pos, self.size // 2] = self.red_door\n\n # Add a blue door at a random position in the right wall\n pos = self.rng.randint(1, self.size - 1)\n self.blue_door = Door('blue')\n self[pos, self.size // 2 + self.size - 1] = self.blue_door\n\n # Generate the mission string\n self.mission = 'open the red door then the blue door'\n\n def step(self, action):\n red_door_opened_before = self.red_door.is_open\n blue_door_opened_before = self.blue_door.is_open\n\n obs, reward, done, info = super().step(action)\n\n red_door_opened_after = self.red_door.is_open\n blue_door_opened_after = self.blue_door.is_open\n\n if blue_door_opened_after:\n if red_door_opened_before:\n reward = self._win_reward\n done = True\n else:\n reward = self._lose_reward\n done = True\n\n elif red_door_opened_after:\n if blue_door_opened_before:\n reward = self._lose_reward\n done = True\n\n return obs, reward, done, info\n\n\nclass Memory(MiniGridEnv):\n \"\"\"\n This environment is a memory test. The agent starts in a small room\n where it sees an object. It then has to go through a narrow hallway\n which ends in a split. At each end of the split there is an object,\n one of which is the same as the object in the starting room. The\n agent has to remember the initial object, and go to the matching\n object at split.\n \"\"\"\n\n def __init__(\n self,\n size=13,\n random_length=False,\n max_steps=None,\n **kwargs\n ):\n self.random_length = random_length\n super().__init__(\n height=size,\n width=size,\n max_steps=5 * size**2 if max_steps is None else max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n self.grid = Grid(height, width)\n\n # Generate the surrounding walls\n self.horz_wall(0, 0)\n self.horz_wall(height - 1, 0)\n self.vert_wall(0, 0)\n self.vert_wall(0, width - 1)\n\n assert height % 2 == 1\n upper_room_wall = height // 2 - 2\n lower_room_wall = height // 2 + 2\n if self.random_length:\n hallway_end = self.rng.randint(4, width - 2)\n else:\n hallway_end = width - 3\n\n # Start room\n self.horz_wall(upper_room_wall, 1, width=4)\n self.horz_wall(upper_room_wall + 1, 4, width=1)\n self.horz_wall(lower_room_wall, 1, width=4)\n self.horz_wall(lower_room_wall - 1, 4, width=1)\n\n # Horizontal hallway\n self.horz_wall(upper_room_wall + 1, 5, width=hallway_end - 5)\n self.horz_wall(lower_room_wall - 1, 5, width=hallway_end - 5)\n\n # Vertical hallway\n self.vert_wall(0, hallway_end, height=height)\n self.vert_wall(0, hallway_end + 2, height=height)\n self[height // 2, hallway_end].clear()\n\n # Fix the player's start position and orientation\n self.agent.pos = (height // 2,\n self.rng.randint(1, hallway_end + 1))\n self.agent.state = 'right'\n\n # Place objects\n start_room_obj = self.rng.choice([Key, Ball])\n self[height // 2 - 1, 1] = start_room_obj('green')\n\n other_objs = self.rng.permutation([Ball, Key])\n pos0 = (height // 2 - 2, hallway_end + 1)\n pos1 = (height // 2 + 2, hallway_end + 1)\n self[pos0] = other_objs[0]('green')\n self[pos1] = other_objs[1]('green')\n\n # Choose the target objects\n if start_room_obj == other_objs[0]:\n self.success_pos = (pos0[0], pos0[1] + 1)\n self.failure_pos = (pos1[0], pos1[1] - 1)\n else:\n self.success_pos = (pos1[0], pos1[1] - 1)\n self.failure_pos = (pos0[0], pos0[1] + 1)\n\n self.mission = 'go to the matching object at the end of the hallway'\n\n def step(self, action):\n if self.actions[action] == 'pickup':\n action = self.actions.index('toggle')\n obs, reward, done, info = super().step(action)\n\n if tuple(self.agent.pos) == self.success_pos:\n reward = self._win_reward\n done = True\n if tuple(self.agent.pos) == self.failure_pos:\n reward = self._lose_reward\n done = True\n\n return obs, reward, done, info\n\n\nclass _LockedRoom(object):\n\n def __init__(self,\n top,\n size,\n door_pos\n ):\n self.top = top\n self.size = size\n self.door_pos = door_pos\n self.color = None\n self.locked = False\n\n def rand_pos(self, env):\n top_i, top_j = self.top\n size_i, size_j = self.size\n return (env.rng.randint(top_i + 1, top_i + size_i - 1),\n env.rng.randint(top_j + 1, top_j + size_j - 1))\n\n\nclass LockedRoom(MiniGridEnv):\n \"\"\"\n The environment has six rooms, one of which is locked. The agent receives a textual mission string as input, telling it which room to go to in order to get the key that opens the locked room. It then has to go into the locked room in order to reach the final goal. This environment is extremely difficult to solve with vanilla reinforcement learning alone.\n \"\"\"\n\n def __init__(self, size=19, max_steps=None, **kwargs):\n super().__init__(\n height=size, \n width=size, \n max_steps=10 * size if max_steps is None else max_steps, \n **kwargs)\n\n def _gen_grid(self, height, width):\n # Create the grid\n self.grid = Grid(height, width)\n\n # Generate the surrounding walls\n self.horz_wall(0, 0)\n self.horz_wall(height - 1, 0)\n self.vert_wall(0, 0)\n self.vert_wall(0, width - 1)\n\n # Hallway walls\n\n left_wall = width // 2 - 2\n right_wall = width // 2 + 2\n\n self.vert_wall(0, left_wall, height=height)\n self.vert_wall(0, right_wall, height=height)\n\n self.rooms = []\n\n # Room splitting walls\n for n in range(0, 3):\n i = n * (height // 3)\n self.horz_wall(i, 0, width=left_wall)\n self.horz_wall(i, right_wall, width=width - right_wall)\n\n room_height = height // 3 + 1\n room_width = left_wall + 1\n self.rooms.append(_LockedRoom(\n (i, 0),\n (room_height, room_width),\n (i + 3, left_wall)\n ))\n self.rooms.append(_LockedRoom(\n (i, right_wall),\n (room_height, room_width),\n (i + 3, right_wall)\n ))\n\n # Choose one random room to be locked\n locked_room = self.rng.choice(self.rooms)\n locked_room.locked = True\n goal_i = self.rng.randint(locked_room.top[0] + 1, locked_room.top[0] + locked_room.size[0] - 1)\n goal_j = self.rng.randint(locked_room.top[1] + 1, locked_room.top[1] + locked_room.size[1] - 1)\n self[goal_i, goal_j] = Goal()\n\n # Assign the door colors\n colors = self.rng.choice(COLORS, size=len(self.rooms))\n for room, color in zip(self.rooms, colors):\n room.color = color\n if room.locked:\n self[room.door_pos] = Door(color, state='locked')\n else:\n self[room.door_pos] = Door(color)\n\n # Select a random room to contain the key\n while True:\n key_room = self.rng.choice(self.rooms)\n if key_room != locked_room:\n break\n key_pos = key_room.rand_pos(self)\n self[key_pos] = Key(locked_room.color)\n\n # Randomize the player start position and orientation\n self.place_agent(\n top=(0, left_wall),\n size=(height, right_wall - left_wall)\n )\n\n # Generate the mission string\n self.mission = (\n f'get the {locked_room.color} key from the {key_room.color} room, '\n f'unlock the {locked_room.color} door and go to the goal'\n )\n\n\nclass KeyCorridor(RoomGrid):\n \"\"\"\n A ball is behind a locked door, the key is placed in a\n random room.\n\n This environment is similar to the locked room environment, but there are multiple registered environment configurations of increasing size, making it easier to use curriculum learning to train an agent to solve it. The agent has to pick up an object which is behind a locked door. The key is hidden in another room, and the agent has to explore the environment to find it. The mission string does not give the agent any clues as to where the key is placed.\n \"\"\"\n _requires_language = False\n\n def __init__(\n self,\n num_rows=3,\n obj_type='ball',\n room_size=6,\n max_steps=None,\n **kwargs\n ):\n self.obj_type = obj_type\n\n super().__init__(\n room_size=room_size,\n num_rows=num_rows,\n max_steps=30 * room_size**2 if max_steps is None else max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n super()._gen_grid(height, width)\n\n # Connect the middle column rooms into a hallway\n for i in range(1, self.num_rows):\n self.remove_wall(i, 1, 'up')\n\n # Add a locked door on the top left\n # Add an object behind the locked door\n room_idx = self.rng.randint(self.num_rows)\n door = self.add_door(room_idx, 2, door_idx='left', locked=True)\n self.obj = self.add_object(room_idx, 2, kind=self.obj_type)\n\n # Add a key in a random room on the left side\n self.add_object(self.rng.randint(self.num_rows), 0, 'key', door.color)\n\n # Place the agent in the middle\n self.place_agent(i=self.num_rows // 2, j=1)\n\n # Make sure all rooms are accessible\n self.connect_all()\n\n self.mission = f'pick up the {self.obj.color} {self.obj.type}'\n\n def step(self, action):\n obs, reward, done, info = super().step(action)\n\n if self.actions[action] == 'pickup':\n if self.agent.is_carrying:\n if self.agent.carrying is self.obj:\n reward = self._win_reward\n done = True\n\n return obs, reward, done, info\n\n\nclass Unlock(RoomGrid):\n \"\"\"\n The agent has to open a locked door.\n \"\"\"\n _requires_language = False\n\n def __init__(self, max_steps=None, **kwargs):\n room_size = 6\n super().__init__(\n num_rows=1,\n num_cols=2,\n room_size=room_size,\n max_steps=8 * room_size**2 if max_steps is None else max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n super()._gen_grid(height, width)\n\n # Make sure the two rooms are directly connected by a locked door\n self.door = self.add_door(0, 0, door_idx='right', locked=True)\n # Add a key to unlock the door\n self.add_object(0, 0, 'key', self.door.color)\n\n self.place_agent(i=0, j=0)\n\n self.mission = 'open the door'\n\n def step(self, action):\n obs, reward, done, info = super().step(action)\n\n if self.actions[action] == 'toggle':\n if self.door.is_open:\n reward = self._win_reward\n done = True\n\n return obs, reward, done, info\n\n\nclass UnlockPickup(RoomGrid):\n \"\"\"\n The agent has to pick up a box which is placed in another room, behind a locked door.\n \"\"\"\n _requires_language = False\n\n def __init__(self, max_steps=None, **kwargs):\n room_size = 6\n super().__init__(\n num_rows=1,\n num_cols=2,\n room_size=room_size,\n max_steps=8 * room_size**2 if max_steps is None else max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n super()._gen_grid(height, width)\n\n # Add a box to the room on the right\n self.obj = self.add_object(0, 1, kind='box')\n # Make sure the two rooms are directly connected by a locked door\n door = self.add_door(0, 0, door_idx='right', locked=True)\n # Add a key to unlock the door\n self.add_object(0, 0, 'key', door.color)\n\n self.place_agent(i=0, j=0)\n\n self.mission = f'pick up the {self.obj.color} {self.obj.type}'\n\n def step(self, action):\n obs, reward, done, info = super().step(action)\n\n if self.actions[action] == 'pickup':\n if self.agent.is_carrying:\n if self.agent.carrying is self.obj:\n reward = self._win_reward\n done = True\n\n return obs, reward, done, info\n\n\nclass BlockedUnlockPickup(RoomGrid):\n \"\"\"\n The agent has to pick up a box which is placed in another room, behind a locked door. The door is also blocked by a ball which the agent has to move before it can unlock the door. Hence, the agent has to learn to move the ball, pick up the key, open the door and pick up the object in the other room.\n \"\"\"\n _requires_language = False\n\n def __init__(self, max_steps=None, **kwargs):\n room_size = 6\n super().__init__(\n num_rows=1,\n num_cols=2,\n room_size=room_size,\n max_steps=16 * room_size**2 if max_steps is None else max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n super()._gen_grid(height, width)\n\n # Add a box to the room on the right\n self.obj = self.add_object(0, 1, kind='box')\n # Make sure the two rooms are directly connected by a locked door\n door = self.add_door(0, 0, door_idx='right', locked=True)\n # Block the door with a ball\n color = self.rng.choice(COLORS)\n self[door.pos[0], door.pos[1] - 1] = Ball(color)\n # Add a key to unlock the door\n self.add_object(0, 0, 'key', door.color)\n self.place_agent(i=0, j=0)\n\n self.mission = f'pick up the {self.obj.color} {self.obj.type}'\n\n def step(self, action):\n obs, reward, done, info = super().step(action)\n\n if self.actions[action] == 'pickup':\n if self.agent.is_carrying:\n if self.agent.carrying is self.obj:\n reward = self._win_reward\n done = True\n\n return obs, reward, done, info\n\n\nclass _ObstructedMaze(RoomGrid):\n \"\"\"\n The agent has to pick up a box which is placed in a corner of a 3x3 maze. The doors are locked, the keys are hidden in boxes and doors are obstructed by balls.\n \"\"\"\n _requires_language = False\n\n def __init__(self,\n num_rows,\n num_cols,\n num_rooms_visited,\n max_steps=None,\n **kwargs\n ):\n room_size = 6\n\n super().__init__(\n room_size=room_size,\n num_rows=num_rows,\n num_cols=num_cols,\n max_steps=4 * num_rooms_visited * room_size**2 if max_steps is None else max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n super()._gen_grid(height, width)\n\n # Define the color of the ball to pick up\n self.ball_to_find_color = COLORS[0]\n # Define the color of the balls that obstruct doors\n self.blocking_ball_color = COLORS[1]\n # Define the color of boxes in which keys are hidden\n self.box_color = COLORS[2]\n\n self.mission = f'pick up the {self.ball_to_find_color} ball'\n\n def step(self, action):\n obs, reward, done, info = super().step(action)\n\n if self.actions[action] == 'pickup':\n if self.agent.is_carrying:\n if self.agent.carrying is self.obj:\n reward = self._win_reward\n done = True\n\n return obs, reward, done, info\n\n def add_door(self, i, j, door_idx='right', color=None, locked=False, key_in_box=False, blocked=False):\n \"\"\"\n Add a door. If the door must be locked, it also adds the key.\n If the key must be hidden, it is put in a box. If the door must\n be obstructed, it adds a ball in front of the door.\n \"\"\"\n\n door = super().add_door(i, j, door_idx=door_idx, color=color, locked=locked)\n\n if blocked: # place a ball in front of the door\n if door_idx == 'right':\n offset = (0, -1)\n elif door_idx == 'down':\n offset = (-1, 0)\n elif door_idx == 'left':\n offset = (0, 1)\n elif door_idx == 'up':\n offset = (1, 0)\n pos = (door.pos[0] + offset[0], door.pos[1] + offset[1])\n self[pos] = Ball(self.blocking_ball_color)\n\n if locked:\n obj = Key(door.color)\n if key_in_box:\n box = Box(self.box_color) if key_in_box else None\n box.contains = obj\n obj = box\n self.place_in_room(i, j, obj)\n\n return door\n\n\nclass ObstructedMaze_1Dlhb(_ObstructedMaze):\n \"\"\"\n A blue ball is hidden in a 2x1 maze. A locked door separates\n rooms. Doors are obstructed by a ball and keys are hidden in boxes.\n \"\"\"\n _requires_language = False\n\n def __init__(self, key_in_box=True, blocked=True, max_steps=None, **kwargs):\n self.key_in_box = key_in_box\n self.blocked = blocked\n\n super().__init__(\n num_rows=1,\n num_cols=2,\n num_rooms_visited=2,\n max_steps=max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n super()._gen_grid(height, width)\n\n self.add_door(0, 0, door_idx='right',\n color=self.rng.choice(COLORS),\n locked=True,\n key_in_box=self.key_in_box,\n blocked=self.blocked)\n\n self.obj = self.add_object(0, 1, 'ball', color=self.ball_to_find_color)\n self.place_agent(i=0, j=0)\n\n\nclass ObstructedMaze_Full(_ObstructedMaze):\n \"\"\"\n A blue ball is hidden in one of the 4 corners of a 3x3 maze. Doors\n are locked, doors are obstructed by a ball and keys are hidden in\n boxes.\n \"\"\"\n _requires_language = False\n\n def __init__(self, agent_room=(1, 1), key_in_box=True, blocked=True,\n num_quarters=4, num_rooms_visited=25, max_steps=None, **kwargs):\n self.agent_room = agent_room\n self.key_in_box = key_in_box\n self.blocked = blocked\n self.num_quarters = num_quarters\n\n super().__init__(\n num_rows=3,\n num_cols=3,\n num_rooms_visited=num_rooms_visited,\n max_steps=max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n super()._gen_grid(height, width)\n\n middle_room = (1, 1)\n # Define positions of \"side rooms\" i.e. rooms that are neither\n # corners nor the center.\n side_rooms = {\n 'right': (1, 2),\n 'down': (2, 1),\n 'left': (1, 0),\n 'up': (0, 1)\n }\n # Define all possible colors for doors\n door_colors = self.rng.choice(COLORS, size=len(side_rooms), replace=False)\n door_colors = {\n 'right': door_colors[0],\n 'down': door_colors[1],\n 'left': door_colors[2],\n 'up': door_colors[3]\n }\n\n side_rooms = list(side_rooms.items())[:self.num_quarters]\n for door_idx, side_room in side_rooms:\n # Add a door between the center room and the side room\n door_color = door_colors[door_idx]\n self.add_door(*middle_room, door_idx=door_idx, color=door_color, locked=False)\n\n for k in [-1, 1]:\n # Add a door to each side of the side room\n side_door_idx = self._door_idx(door_idx, k)\n side_door_color = door_colors[side_door_idx]\n self.add_door(*side_room, locked=True,\n door_idx=side_door_idx,\n color=side_door_color,\n key_in_box=self.key_in_box,\n blocked=self.blocked)\n\n corners = [(0, 2), (2, 2), (2, 0), (0, 0)]\n ball_room = corners[self.rng.randint(self.num_quarters)]\n\n self.obj = self.add_object(*ball_room, 'ball', color=self.ball_to_find_color)\n self.place_agent(i=self.agent_room[0], j=self.agent_room[1])\n\n\nclass DistShift(MiniGridEnv):\n \"\"\"\n Distributional shift environment\n\n This environment is based on one of the DeepMind [AI safety gridworlds](https://github.com/deepmind/ai-safety-gridworlds). The agent starts in the top-left corner and must reach the goal which is in the top-right corner, but has to avoid stepping into lava on its way. The aim of this environment is to test an agent's ability to generalize. There are two slightly different variants of the environment, so that the agent can be trained on one variant and tested on the other.\n \"\"\"\n\n def __init__(\n self,\n width=9,\n height=7,\n agent_start_pos=(1,1),\n agent_start_state='right',\n strip2_row=2,\n max_steps=None,\n **kwargs\n ):\n self.agent_start_pos = agent_start_pos\n self.agent_start_state = agent_start_state\n self.goal_pos = (1, width - 2)\n self.strip2_row = strip2_row\n\n super().__init__(\n width=width,\n height=height,\n max_steps=4 * width * height if max_steps is None else max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n # Create an empty grid\n self.grid = Grid(height, width)\n\n # Generate the surrounding walls\n self.wall_rect(0, 0, height, width)\n\n # Place a goal square in the bottom-right corner\n self[self.goal_pos] = Goal()\n\n # Place the lava rows\n for j in range(self.width - 6):\n self[1, 3 + j] = Lava()\n self[self.strip2_row, 3 + j] = Lava()\n\n # Place the agent\n if self.agent_start_pos is not None:\n self.agent.pos = self.agent_start_pos\n self.agent.state = self.agent_start_state\n else:\n self.place_agent()\n\n self.mission = 'get to the green goal square'\n\n\nclass LavaGap(MiniGridEnv):\n \"\"\"\n The agent has to reach the green goal square at the opposite corner of the room, and must pass through a narrow gap in a vertical strip of deadly lava. Touching the lava terminate the episode with a zero reward. This environment is useful for studying safety and safe exploration.\n \"\"\"\n\n def __init__(self, size=7, obstacle_type=Lava, max_steps=None, **kwargs):\n self.obstacle_type = obstacle_type\n super().__init__(\n height=size,\n width=size,\n max_steps=4 * size**2 if max_steps is None else max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n assert width >= 5 and height >= 5\n\n # Create an empty grid\n self.grid = Grid(height, width)\n\n # Generate the surrounding walls\n self.wall_rect(0, 0, height, width)\n\n # Place the agent in the top-left corner\n self.agent.pos = (1, 1)\n self.agent.state = 'right'\n\n # Place a goal square in the bottom-right corner\n self[height - 2, width - 2] = Goal()\n\n # Generate and store random gap position\n gap_pos = (self.rng.randint(1, height - 1),\n self.rng.randint(2, width - 2))\n\n # Place the obstacle wall\n self.vert_wall(1, gap_pos[1], height=height - 2,\n obj=self.obstacle_type)\n\n # Put a hole in the wall\n self[gap_pos].clear()\n\n if type(self.obstacle_type) is Lava:\n self.mission = 'avoid the lava and get to the green goal square'\n else:\n self.mission = 'find the opening and get to the green goal square'\n\n\nclass _Crossing(MiniGridEnv):\n \"\"\"\n Environment with wall or lava obstacles, sparse reward.\n \"\"\"\n\n def __init__(self, size=9, num_crossings=1, obstacle_type=Lava, max_steps=None, **kwargs):\n self.num_crossings = num_crossings\n self.obstacle_type = obstacle_type\n super().__init__(\n height=size,\n width=size,\n max_steps=4 * size * size if max_steps is None else max_steps,\n **kwargs\n )\n\n def _gen_grid(self, height, width):\n assert width % 2 == 1 and height % 2 == 1 # odd size\n\n # Create an empty grid\n self.grid = Grid(height, width)\n\n # Generate the surrounding walls\n self.wall_rect(0, 0, height, width)\n\n # Place the agent in the top-left corner\n self.agent.pos = (1, 1)\n self.agent.state = 'right'\n\n # Place a goal square in the bottom-right corner\n self[height - 2, width - 2] = Goal()\n\n # Place obstacles (lava or walls)\n\n # Lava rivers or walls specified by direction and position in grid\n rivers = [('v', i) for i in range(2, width - 2, 2)]\n rivers += [('h', j) for j in range(2, height - 2, 2)]\n self.rng.shuffle(rivers)\n rivers = rivers[:self.num_crossings] # sample random rivers\n rivers_v = sorted([pos for d, pos in rivers if d == 'v'])\n rivers_h = sorted([pos for d, pos in rivers if d == 'h'])\n obstacle_pos = itertools.chain(\n itertools.product(range(1, height - 1), rivers_v),\n itertools.product(rivers_h, range(1, width - 1)),\n )\n for pos in obstacle_pos:\n self[pos] = self.obstacle_type()\n\n # Sample path to goal\n path = ['v'] * len(rivers_v) + ['h'] * len(rivers_h)\n self.rng.shuffle(path)\n\n # Create openings\n limits_h = [0] + rivers_h + [height - 1]\n limits_v = [0] + rivers_v + [width - 1]\n room_i, room_j = 0, 0\n for direction in path:\n if direction == 'h':\n i = limits_h[room_i + 1]\n j = self.rng.choice(\n range(limits_v[room_j] + 1, limits_v[room_j + 1]))\n room_i += 1\n elif direction == 'v':\n i = self.rng.choice(\n range(limits_h[room_i] + 1, limits_h[room_i + 1]))\n j = limits_v[room_j + 1]\n room_j += 1\n self[i, j].clear()\n\n if type(self.obstacle_type) is Lava:\n self.mission = 'avoid the lava and get to the green goal square'\n else:\n self.mission = 'find the opening and get to the green goal square'\n\n\nclass LavaCrossing(_Crossing):\n \"\"\"\n The agent has to reach the green goal square on the other corner of the room while avoiding rivers of deadly lava which terminate the episode in failure. Each lava stream runs across the room either horizontally or vertically, and has a single crossing point which can be safely used; Luckily, a path to the goal is guaranteed to exist. This environment is useful for studying safety and safe exploration.\n \"\"\"\n\n def __init__(self, size=9, num_crossings=1, max_steps=None, **kwargs):\n super().__init__(size=size, num_crossings=num_crossings, obstacle_type=Lava, max_steps=max_steps, **kwargs)\n\n\nclass SimpleCrossing(_Crossing):\n \"\"\"\n Similar to the LavaCrossing environment, the agent has to reach the green goal square on the other corner of the room, however lava is replaced by walls. This MDP is therefore much easier and and maybe useful for quickly testing your algorithms.\n \"\"\"\n\n def __init__(self, size=11, num_crossings=5, max_steps=None, **kwargs):\n super().__init__(size=size, num_crossings=num_crossings,\n obstacle_type=Wall, max_steps=max_steps, **kwargs)\n\n\nclass DynamicObstacles(MiniGridEnv):\n \"\"\"\n This environment is an empty room with moving obstacles. The goal of the agent is to reach the green goal square without colliding with any obstacle. A large penalty is subtracted if the agent collides with an obstacle and the episode finishes. This environment is useful to test Dynamic Obstacle Avoidance for mobile robots with Reinforcement Learning in Partial Observability.\n \"\"\"\n\n def __init__(\n self,\n size=8,\n agent_start_pos=(1, 1),\n agent_start_state='right',\n n_obstacles=4,\n max_steps=None,\n **kwargs\n ):\n self.agent_start_pos = agent_start_pos\n self.agent_start_state = agent_start_state\n\n # Reduce obstacles if there are too many\n if n_obstacles <= size / 2 + 1:\n self.n_obstacles = int(n_obstacles)\n else:\n self.n_obstacles = int(size / 2)\n super().__init__(\n height=size,\n width=size,\n max_steps=4 * size * size if max_steps is None else max_steps,\n **kwargs\n )\n # Allow only 3 actions permitted: left, right, forward\n self.action_space = gym.spaces.Discrete(3)\n\n def _gen_grid(self, height, width):\n # Create an empty grid\n self.grid = Grid(height, width)\n\n # Generate the surrounding walls\n self.wall_rect(0, 0, height, width)\n\n # Place a goal square in the bottom-right corner\n self[height - 2, width - 2] = Goal()\n\n # Place the agent\n if self.agent_start_pos is not None:\n self.agent.pos = self.agent_start_pos\n self.agent.state = self.agent_start_state\n else:\n self.place_agent()\n\n # Place obstacles\n self.obstacles = []\n for i_obst in range(self.n_obstacles):\n self.obstacles.append(Ball())\n self.place_obj(self.obstacles[i_obst], max_tries=100)\n\n self.mission = 'get to the green goal square'\n\n def step(self, action):\n # Invalid action\n if action >= self.action_space.n:\n action = 0\n\n # Check if there is an obstacle in front of the agent\n front_cell = self[self.agent.front_pos]\n not_clear = front_cell.entity is not None and front_cell.entity.type != 'goal'\n\n # Update obstacle positions\n for i_obst in range(len(self.obstacles)):\n old_pos = self.obstacles[i_obst].pos\n top = tuple(map(operator.add, old_pos, (-1, -1)))\n\n try:\n self.place_obj(self.obstacles[i_obst], top=top, size=(3,3), max_tries=100)\n self[old_pos].clear()\n except RecursionError:\n pass\n\n obs, reward, done, info = super().step(action)\n\n # If the agent tries to walk over an obstacle\n if self.actions[action] == 'forward' and not_clear:\n reward = self._lose_reward\n done = True\n return obs, reward, done, info\n\n return obs, reward, done, info\n\n\nclass Playground(MiniGridEnv):\n \"\"\"\n Environment with multiple rooms and random objects.\n This environment has no specific goals or rewards.\n \"\"\"\n\n def __init__(self, size=19, max_steps=100, **kwargs):\n super().__init__(height=size, width=size, max_steps=max_steps, **kwargs)\n\n def _gen_grid(self, height, width):\n # Create the grid\n self.grid = Grid(height, width)\n\n # Generate the surrounding walls\n self.horz_wall(0, 0)\n self.horz_wall(height - 1, 0)\n self.vert_wall(0, 0)\n self.vert_wall(0, width - 1)\n\n room_width = width // 3\n room_height = height // 3\n\n # For each row of rooms\n for i in range(3):\n # For each column\n for j in range(3):\n\n i_top = i * room_height\n j_left = j * room_width\n i_bottom = i_top + room_height\n j_right = j_left + room_width\n\n # Right wall and door\n if j + 1 < 3:\n self.vert_wall(i_top, j_right, height=room_height)\n pos = (self.rng.randint(i_top + 1, i_bottom - 1), j_right)\n color = self.rng.choice(COLORS)\n self[pos] = Door(color)\n\n # Bottom wall and door\n if i + 1 < 3:\n self.horz_wall(i_bottom, j_left, width=room_width)\n pos = (i_bottom, self.rng.randint(j_left + 1, j_right - 1))\n color = self.rng.choice(COLORS)\n self[pos] = Door(color)\n\n # Place random objects in the world\n types = ['key', 'ball', 'box']\n for i in range(0, 12):\n obj_type = self.rng.choice(types)\n obj_color = self.rng.choice(COLORS)\n if obj_type == 'key':\n obj = Key(obj_color)\n elif obj_type == 'ball':\n obj = Ball(obj_color)\n elif obj_type == 'box':\n obj = Box(obj_color)\n self.place_obj(obj)\n\n # Randomize the player start position and orientation\n self.place_agent()\n\n # No explicit mission in this environment\n self.mission = ''\n\n\nclass RandomObjects(MiniGridEnv):\n \"\"\"\n This environment is a blank grid filled with randomly placed objects (including wall elements). Useful for curriculum learning as the first learning stage.\n \"\"\"\n\n def __init__(self,\n size=16,\n density=.2,\n objects=OBJECTS,\n colors=COLORS,\n max_steps=100,\n surround_walls=True,\n **kwargs):\n self.density = density\n self.objects = objects\n self.colors = colors\n self.surround_walls = surround_walls\n super().__init__(height=size, width=size, max_steps=max_steps, **kwargs)\n\n def _gen_grid(self, height, width):\n # Create an empty grid\n self.grid = Grid(height, width)\n\n if self.surround_walls:\n self.horz_wall(0, 0)\n self.horz_wall(height - 1, 0)\n self.vert_wall(0, 0)\n self.vert_wall(0, width - 1)\n\n # Place a goal square at a random location\n self.place_obj(Goal())\n\n # Place random objects in the world\n if self.surround_walls:\n mean_n_objs = int((height - 2) * (width - 2) * self.density)\n else:\n mean_n_objs = int(height * width * self.density)\n n_objs = mean_n_objs\n for i in range(n_objs):\n obj = self.make_obj()\n self.place_obj(obj)\n\n # Randomize the player start position and orientation\n agent_pos = (height // 2, width // 2)\n self[agent_pos].clear()\n self.agent.pos = agent_pos\n self.agent.state = self.rng.choice(self.agent.STATES)\n\n self.mission = 'get to a green goal square'\n\n def make_obj(self):\n type_ = self.rng.choice(self.objects)\n color = self.rng.choice(self.colors)\n\n if type_ in ['wall', 'goal', 'lava']:\n obj = entities.make(type_)\n elif type_ == 'door':\n state = self.rng.choice(entities.Door.STATES)\n obj = entities.make(type_, color=color, state=state)\n elif type_ == 'box':\n if self.rng.random() < .5:\n contains = None\n else:\n contains = self.make_obj()\n obj = entities.make(type_, color=color, contains=contains)\n else:\n obj = entities.make(type_, color=color)\n\n return obj\n\n\n# Register all environments with OpenAI gym\nfor name, obj in inspect.getmembers(sys.modules[__name__]):\n if inspect.isclass(obj) and obj.__module__ == __name__ and not name.startswith('_'):\n\n gym.envs.registration.register(\n id=f'MiniGrid-{name}-v1',\n entry_point=f'gym_minigrid.envs:{name}',\n reward_threshold=.95,\n )\n","sub_path":"gym_minigrid/envs.py","file_name":"envs.py","file_ext":"py","file_size_in_byte":61429,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"468745519","text":"from datetime import date\n\nfrom django.views.generic.base import TemplateView\n\nfrom regulations.generator.api_reader import ApiReader\nfrom .utils import get_structure\n\nclient = ApiReader()\n\n\nclass SearchView(TemplateView):\n\n template_name = 'regulations/search.html'\n\n def get_context_data(self, **kwargs):\n context = super().get_context_data(**kwargs)\n results = get_data(self.request.GET.get(\"q\"))\n today = date.today()\n parts = client.effective_parts(today)\n structure = get_structure(parts)\n c = {\n 'parts': parts,\n 'toc': structure,\n 'results': results,\n }\n return {**context, **c, **self.request.GET.dict()}\n\n\ndef get_data(query):\n return client.search(query)\n","sub_path":"regulations/views/search.py","file_name":"search.py","file_ext":"py","file_size_in_byte":764,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"534989027","text":"def duplicate(a_list):\n n = []\n for i in range(1, len(a_list)):\n k = i - 1\n count = 1\n while k >= 0 and count <= 2:\n if a_list[i] == a_list[k]:\n count += 1\n k -= 1\n if count >= 2:\n if a_list[i] not in n:\n n.append(a_list[i])\n return n\n\n\nprint(duplicate([1, 2, 3, 3, 5, 6, 7, 7, 7, 7]))\n","sub_path":"menu.py","file_name":"menu.py","file_ext":"py","file_size_in_byte":386,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"261962789","text":"class BTnode:\n '''\n class to create an object of binary tree node\n '''\n\n def __init__(self, data):\n '''\n constructor to initialize fields\n such as data\n left pointer\n and right pointer\n '''\n self.data = data\n self.left = None\n self.right = None\n\n\ndef is_leaf(root):\n '''\n utility function to check\n if node is leaf node\n '''\n\n if root.left is None and root.right is None:\n return True\n\n return False\n\n\ndef same_level(root, level, store):\n '''\n Utility function to check\n if current root is a leaf node\n and store the level for it\n then call recursively for left and right\n store is a set that stores the level\n of leaf nodes found\n '''\n\n if root is None:\n return\n\n # check if leaf node\n # if a leaf node then it will not have\n # any child nodes\n if(is_leaf(root)):\n store.add(level)\n return\n\n # call for left and right\n same_level(root.left, level+1, store)\n same_level(root.right, level+1, store)\n\n return\n\n\ndef check(node):\n '''\n main function which calls other \n '''\n\n # initialze a set to store values\n store={}\n\n # call function same_level to fill values in store\n same_level(node,0,store)\n\n # if store has only one elemnt\n # implies that all the leaf nodes\n # are at same level\n if len(store) == 1:\n return True\n \n return False\n","sub_path":"leaf_same_level.py","file_name":"leaf_same_level.py","file_ext":"py","file_size_in_byte":1452,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"184353684","text":"# -*- coding: utf-8 -*-\nfrom aiida.orm.data.upf import UpfData, get_pseudos_from_structure\n\n\ndef get_pseudos_of_calc(calc):\n \"\"\"\n Return a dictionary of pseudos used by a given (pw.x, cp.x) calculation.\n\n This returns a dictionary ``pseudos`` that can be set in a builder as ``builder.pseudo = pseudos``.\n\n :param calc: a pw.x or cp.x calculation.\n :return: a dictionary where the key is the kind name and the value is the UpfData object.\n \"\"\"\n from aiida.common.links import LinkType\n\n pseudos = {}\n # I create here a dictionary that associates each kind name to a pseudo\n inputs = calc.get_inputs_dict(link_type=LinkType.INPUT)\n for linkname in inputs.keys():\n if linkname.startswith(calc._get_linkname_pseudo_prefix()):\n # Note that this string might be a sequence of kind names\n # concatenated by an underscore, see implementation in the\n # input plugin implementation.\n multiplekindstring = linkname[len(calc._get_linkname_pseudo_prefix()):]\n pseudos[multiplekindstring] = inputs[linkname]\n return pseudos\n\n\ndef validate_and_prepare_pseudos_inputs(structure, pseudos=None, pseudo_family=None):\n \"\"\"\n Use the explicitly passed pseudos dictionary or use the pseudo_family in combination with\n the structure to obtain that dictionary.\n\n The pseudos dictionary should now be a dictionary of UPF nodes with the kind as linkname\n As such, if there are multiple kinds with the same element, there will be duplicate UPF nodes\n but multiple links for the same input node are not allowed. Moreover, to couple the UPF nodes\n to the Calculation instance, we have to go through the use_pseudo method, which takes the kind\n name as an additional parameter. When creating a Calculation through a Process instance, one\n cannot call the use methods directly but rather should pass them as keyword arguments. However,\n we can pass the additional parameters by using them as the keys of a dictionary\n\n :param structure: StructureData node\n :param pseudos: a dictionary where keys are the kind names and value are UpfData nodes\n :param pseudo_family: string name of the pseudopotential family to use\n :raises: ValueError if neither pseudos or pseudo_family is specified or if no UpfData is found for\n every element in the structure\n :returns: a dictionary of UpfData nodes where the key is a tuple with the kind name\n \"\"\"\n from aiida.orm.data.base import Str\n result_pseudos = {}\n unique_pseudos = {}\n\n if pseudos and pseudo_family:\n raise ValueError('You cannot specify both \"pseudos\" and \"pseudo_family\"')\n elif pseudos is None and pseudo_family is None:\n raise ValueError('Neither an explicit pseudos dictionary nor a pseudo_family was specified')\n elif pseudo_family:\n # This will already raise some exceptions, potentially, like the ones below\n pseudos = get_pseudos_from_structure(structure, str(pseudo_family))\n\n if isinstance(pseudos, (str, unicode, Str)):\n raise TypeError('You passed \"pseudos\" as a string - maybe you wanted to pass it as \"pseudo_family\" instead?')\n\n for kind in structure.get_kind_names():\n if kind not in pseudos:\n raise ValueError('no pseudo available for element {}'.format(kind))\n elif not isinstance(pseudos[kind], UpfData):\n raise ValueError('pseudo for element {} is not of type UpfData'.format(kind))\n\n for kind, pseudo in pseudos.iteritems():\n unique_pseudos.setdefault(pseudo, []).append(kind)\n\n for pseudo, kinds in unique_pseudos.iteritems():\n result_pseudos[tuple(kinds)] = pseudo\n\n return result_pseudos\n","sub_path":"aiida_quantumespresso/utils/pseudopotential.py","file_name":"pseudopotential.py","file_ext":"py","file_size_in_byte":3702,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"267836325","text":"import pygame\nimport cards\nimport view\nimport random\nimport copy\nimport time\nimport pprint\nfrom signalslot import Signal\nfrom ai_comp import ai\nfrom enum import Enum\n\nNUM_OF_PLAYERS = 4\nSTARTING_HAND = 13\nHIGHEST_CARD = 414\nLOWEST_CARD = 102\nVIEW_TRANSPARENT = False # Make the text box not transparent\n\n\nclass GameState(Enum):\n DEALING = 0\n POINT_CHECK = 1\n BIDDING = 2\n PLAYING = 3\n ENDING = 4\n\n\nclass PlayerRole(Enum):\n UNKNOWN = 0\n ATTACKER = 1\n DEFENDER = 2\n\n\nclass Table:\n \"\"\"\n A Table is the place where all actions takes place. It is essentially a FSM, doing different\n routines at each state. It needs to keep track of the score, roles, the rules, etc. It needs\n to ask each player for decisions and respond to them accordingly. The table will also need\n to inform any decision to the Main Screen so that it can update the screen to reflect that\n change through the use of callbacks (Signal and Slot). This call should be minimised by making\n all the changes before calling to update the screen in one go.\n\n FSM cycles\n ---\n Preloop - Prepare the cards once\n - Initiate Players and connect them to the Table\n 1. Shuffle and Deal out cards to Players.\n 2a. Detect weak hands and ask for reshuffle.\n 2b. Return to (1) if any reshuffle occurs, otherwise proceed.\n 3. Bidding round. Randomly pick a starting player, in clockwise manner\n ask for a bid until it is valid.\n 3b. Proceed only if 3 consecutive skips are detected.\n 3c. Ask the winner of the bid a card not in their hand.\n 3d. Set up the player roles, trump suit, rounds to win for both side\n 3e. Play the game. Start with bid winner if NO TRUMP, otherwise\n Starting next to the bid winner.\n 4a. With the first player, ask for any card, excluding trump suits if trump\n is not broken\n 4b. With subsequent players, ask for cards that follow the suit of the first player\n , include trump suit if trump is broken. Ask for any card if the player cannot\n follow suit.\n 4c. Once all 4 players has made valid plays, announce results, update scoring. Announce\n player roles if the partner card is played. Break trump if trump is played.\n 4d. Repeat 4 until 13 rounds are made. Maybe add early win if confirmed one side wins\n 5. Ask for a new game. Go back to 1 if true.\n\n All played cards go into a hidden discard pile.\n\n \"\"\"\n update_table = Signal()\n\n def __init__(self, x, y, width, height, clear_colour, autoplay=False, view_all_cards=False):\n # TODO: Reduce the amount of update_table call\n self.x = x\n self.y = y\n self.width = width\n self.height = height\n\n self.table_font = pygame.font.SysFont(\"None\", 25)\n self.player_font = pygame.font.SysFont(\"None\", 25)\n\n # For gameplay\n self.game_state = GameState.DEALING\n self.current_round = 0\n self.passes = 0\n self.current_player = 0\n self.first_player = False # This is for bidding purposes\n self.players = []\n self.players_playzone = []\n # Table status will be made known to the player by reference\n self.table_status = {'played cards': [0, 0, 0, 0], 'leading player': 0, 'trump suit': 1,\n 'trump broken': False, 'round history': [], 'bid': 0, 'partner': 0,\n 'partner reveal': False, 'defender': {'target': 0, 'wins': 0},\n 'attacker': {'target': 0, 'wins': 0}}\n\n # Prepare the surfaces for displaying\n self.background = pygame.Surface((self.width, self.height))\n self.background.fill(clear_colour)\n self.background = self.background.convert()\n\n # TODO: Update the drawing of the table?\n # Prepare the card with dimensions\n w_deck = min(self.height, self.width) * 0.18\n l_deck = min(self.width, self.height) * 0.7\n # This is not a deck as it will never be drawn\n self.discard_deck = cards.prepare_playing_cards(int(w_deck*0.7), int(w_deck*0.9))\n game_margins = 5\n\n # Players' deck positioning\n playerx = ((self.width - l_deck)//2,\n game_margins,\n (self.width - l_deck)//2,\n self.width - w_deck - game_margins)\n playery = (self.height - w_deck - game_margins,\n (self.height - l_deck)//2,\n game_margins,\n (self.height - l_deck)//2)\n h_spacing = 20\n v_spacing = 25\n\n # Middle playfield for announcer and player playing deck positioning\n playfield_margins = 5\n margins_with_w_deck = w_deck + playfield_margins + game_margins\n playfield_x = margins_with_w_deck\n playfield_y = margins_with_w_deck\n playfield_width = self.width - margins_with_w_deck * 2\n playfield_height = self.height - margins_with_w_deck * 2\n\n playdeckx = (playfield_x + (playfield_width - w_deck) / 2,\n playfield_x,\n playfield_x + (playfield_width - w_deck) / 2,\n playfield_x + playfield_width - w_deck)\n playdecky = (playfield_y + playfield_height - w_deck,\n playfield_y + (playfield_height - w_deck) / 2,\n playfield_y,\n playfield_y + (playfield_height - w_deck) / 2)\n\n # Player stats positioning\n stats_width = 100\n self.stats_height = 100\n stats_spacing = 10\n self.player_stats_x = (playdeckx[0] - stats_width - stats_spacing,\n playdeckx[1],\n playdeckx[2] + w_deck + stats_spacing,\n playdeckx[3])\n self.player_stats_y = (playdecky[0] + w_deck - self.stats_height,\n playdecky[1] - self.stats_height - stats_spacing,\n playdecky[2],\n playdecky[3] + w_deck + stats_spacing)\n\n self.player_stats = [[], [], [], []]\n\n # TODO: change surface to use colorkey, maybe, if the performance is tanked\n # Prepare all the player surfaces\n for i in range(4):\n vert = i % 2 == 1\n spacing = h_spacing\n if vert:\n spacing = v_spacing\n\n reveal_mode = cards.DeckReveal.HIDE_ALL\n if i == 0 or view_all_cards:\n reveal_mode = cards.DeckReveal.SHOW_ALL\n self.players.append(Player(playerx[i], playery[i],\n l_deck, w_deck,\n spacing, vert_orientation=vert,\n deck_reveal=reveal_mode))\n self.players[i].connect_to_table(self.table_status)\n if i > 0:\n self.players[i].add_ai(ai.RandomAI(self.table_status))\n\n self.players_playzone.append(cards.Deck(playdeckx[i], playdecky[i],\n w_deck, w_deck, 0))\n for j in range(3):\n surf = pygame.Surface((stats_width, self.stats_height / 3), pygame.SRCALPHA)\n rendered_text = self.player_font.render(\"Player {0:d}\".format(i), True,\n (255, 0, 255)).convert_alpha()\n self.center_text_on_surface(surf, rendered_text,\n (255, 255, 255, 255 * VIEW_TRANSPARENT))\n self.player_stats[i].append(surf)\n\n if autoplay:\n self.players[0].add_ai(ai.RandomAI(self.table_status))\n\n # Announcer positioning and surface creation\n announcer_margins = 5\n announcer_spacing = announcer_margins + w_deck\n self.announcer_x = playfield_x + announcer_spacing\n self.announcer_y = playfield_y + announcer_spacing\n self.announcer_width = playfield_width - 2 * announcer_spacing\n self.announcer_height = playfield_height - 2 * announcer_spacing\n self.announcer_line = []\n for i in range(3):\n surf = pygame.Surface((self.announcer_width, self.announcer_height/3), pygame.SRCALPHA)\n self.announcer_line.append(surf)\n\n self.update_all_players(role=True, wins=True)\n\n self.write_message(\"Press P to play!\")\n\n self.ongoing = False\n\n\n\n def center_text_on_surface(self, surf, rendered_text, clear_colour):\n line_center = surf.get_rect().center\n text_rect = rendered_text.get_rect(center=line_center)\n surf.fill(clear_colour)\n surf.blit(rendered_text, text_rect)\n\n def write_message(self, text, delay_time=0.5, line=0, update_now=True):\n \"\"\"\n Write a message into the center board surface (announcer)\n :param text: String to be displayed on the center board\n :param delay_time: How much delay to put once the string is display\n :param line: Which line of the announcer to write to\n :return: None\n \"\"\"\n if 0 <= line < len(self.announcer_line):\n print(text)\n text = text.strip('\\n')\n rendered_text = self.table_font.render(text, True, (255, 255, 255)).convert_alpha()\n self.center_text_on_surface(self.announcer_line[line], rendered_text,\n (255, 255, 255, 255*VIEW_TRANSPARENT))\n if update_now:\n self.update_table.emit()\n time.sleep(delay_time)\n\n def update_players_role(self, player_num, update_now=True):\n self.player_stats[player_num][1].fill((255, 255, 255, 255*VIEW_TRANSPARENT))\n if self.players[player_num].role == PlayerRole.DEFENDER:\n rendered_text = self.player_font.render(\"Defender\", True, (0, 64, 192)).convert_alpha()\n self.center_text_on_surface(self.player_stats[player_num][1], rendered_text,\n (255, 255, 255, 255 * VIEW_TRANSPARENT))\n elif self.players[player_num].role == PlayerRole.ATTACKER:\n rendered_text = self.player_font.render(\"Attacker\", True, (192, 0, 0)).convert_alpha()\n self.center_text_on_surface(self.player_stats[player_num][1], rendered_text,\n (255, 255, 255, 255 * VIEW_TRANSPARENT))\n if update_now:\n self.update_table.emit()\n\n def update_player_wins(self, player_num, update_now=True):\n self.player_stats[player_num][2].fill((255, 255, 255, 255*VIEW_TRANSPARENT))\n if self.players[player_num].score > 1:\n rendered_text = self.player_font.render(\"Wins: {0:d}\".format(self.players[player_num].score), True,\n (255, 255, 255)).convert_alpha()\n else:\n rendered_text = self.player_font.render(\"Win: {0:d}\".format(self.players[player_num].score), True,\n (255, 255, 255)).convert_alpha()\n self.center_text_on_surface(self.player_stats[player_num][2], rendered_text,\n (255, 255, 255, 255 * VIEW_TRANSPARENT))\n if update_now:\n self.update_table.emit()\n\n def update_all_players(self, role=False, wins=True):\n for i in range(4):\n if wins:\n self.update_player_wins(i, update_now=False)\n if role:\n self.update_players_role(i, update_now=False)\n self.update_table.emit()\n\n def display_current_player(self, current=-1):\n if current >= 0:\n print(\"Player {0:d}\\n\".format(current))\n for i in range(4):\n rendered_text = self.player_font.render(\"Player {0:d}\".format(i), True,\n (255, 0, 255)).convert_alpha()\n if i == current:\n self.center_text_on_surface(self.player_stats[i][0], rendered_text,\n (0, 64, 0, 255))\n else:\n self.center_text_on_surface(self.player_stats[i][0], rendered_text,\n (255, 255, 255, 255 * VIEW_TRANSPARENT))\n\n self.update_table.emit()\n\n def update_team_scores(self):\n if self.table_status['partner reveal']:\n msg = \"Defender: {0:d}/{2:d}, Attacker: {1:d}/{3:d}\\n\".format(self.table_status['defender']['wins'],\n self.table_status['attacker']['wins'],\n self.table_status['defender']['target'],\n self.table_status['attacker']['target'])\n self.write_message(msg, line=2)\n else:\n msg = \"Defender: {0:d}?/{1:d}, Attacker: ?/{2:d}\\n\".format(self.table_status['defender']['wins'],\n self.table_status['defender']['target'],\n self.table_status['attacker']['target'])\n self.write_message(msg, line=2)\n\n def get_pos(self):\n return self.x, self.y\n\n def continue_game(self):\n \"\"\"\n This is where the FSM is. State transition should occur here.\n What takes place in the state should be in a function.\n :return: None\n \"\"\"\n # TODO: Adjust the timing of sleep\n if self.game_state == GameState.DEALING:\n self.shuffle_and_deal()\n self.write_message(\"Shuffle Complete!\")\n self.game_state = GameState.POINT_CHECK\n\n elif self.game_state == GameState.POINT_CHECK:\n if self.check_reshuffle():\n self.write_message('Reshuffle Initiated!', line=1)\n self.game_state = GameState.ENDING\n else:\n self.write_message('No Reshuffle needed!')\n self.game_state = GameState.BIDDING\n self.write_message(\"Start to Bid\")\n self.prepare_bidding()\n elif self.game_state == GameState.BIDDING:\n bid_complete = self.start_bidding()\n if bid_complete:\n self.game_state = GameState.PLAYING\n self.update_all_players(role=True, wins=True)\n self.update_team_scores()\n\n elif self.game_state == GameState.PLAYING:\n self.play_a_round()\n if self.current_round == 13:\n self.write_message(\"Game Set! Press P to play again!\")\n self.ongoing = False\n self.game_state = GameState.ENDING\n else:\n self.reset_game()\n self.game_state = GameState.DEALING\n\n def shuffle_and_deal(self):\n \"\"\"\n Shuffle and deal the discard deck to the players, which should have 52 cards.\n :return: None\n \"\"\"\n if self.discard_deck:\n for i in range(10):\n random.shuffle(self.discard_deck)\n for player in self.players:\n for i in range(STARTING_HAND):\n player.add_card(self.discard_deck.pop())\n self.update_table.emit()\n\n def check_reshuffle(self):\n \"\"\"\n Detect any possible reshuffle request within the players\n :return: True if reshuffle requested, else False\n \"\"\"\n print(\"Player Point Count\")\n for i, player in enumerate(self.players):\n print(\"Player {0:d}: {1:d}\".format(i, player.get_card_points()))\n if player.get_card_points() < 4:\n self.write_message(\"Low points detected in Player {0:d}! \".format(i))\n return player.make_decision(self.game_state, 0)\n\n def prepare_bidding(self):\n # Randomly pick a starting player, whom also is the current bid winner\n self.current_player = random.randint(1, NUM_OF_PLAYERS) - 1\n print(\"Starting Player: {0:d}\".format(self.current_player))\n self.passes = 0\n self.table_status[\"bid\"] = 11 # Lowest Bid: 1 Club by default\n self.first_player = True # Starting bidder \"privilege\" to raise the starting bid\n msg = \"Current Bid: {0:d} {1:s}\".format(self.table_status[\"bid\"] // 10,\n cards.get_suit_string(self.table_status[\"bid\"] % 10))\n self.write_message(msg, line=1, delay_time=0)\n self.display_current_player(self.current_player)\n msg = 'Bid Leader: Player {0:d}'.format((self.current_player - self.passes - 1 * (not self.first_player)) % 4)\n self.write_message(msg, line=2, delay_time=1)\n\n def start_bidding(self):\n \"\"\"\n The bidding procedure.\n :return: Whether bidding is completed\n \"\"\"\n # Highest bid: 7 NoTrump. No further check required\n if self.passes < NUM_OF_PLAYERS - 1 and self.table_status[\"bid\"] < 75:\n player_bid = self.players[self.current_player].make_decision(self.game_state, 0)\n if not player_bid:\n if not self.first_player: # Starting bidder pass do not count at the start\n self.passes += 1\n else:\n self.table_status[\"bid\"] = player_bid\n self.passes = 0\n\n if self.table_status[\"bid\"] < 75:\n self.current_player += 1\n self.current_player %= 4\n msg = \"Current Bid: {0:d} {1:s}\".format(self.table_status[\"bid\"] // 10,\n cards.get_suit_string(self.table_status[\"bid\"] % 10))\n self.write_message(msg, line=1, update_now=False)\n msg = 'Bid Leader: Player {0:d}'.format((self.current_player - self.passes\n - 1 * (not self.first_player)) % 4)\n self.write_message(msg, line=2, update_now=False)\n self.display_current_player(self.current_player)\n if self.first_player:\n self.first_player = False\n time.sleep(0.5)\n return False\n else:\n self.write_message(\"Player {0:d} is the bid winner!\".format(self.current_player), delay_time=1)\n msg = \"Player {0:d} is calling a partner...\".format(self.current_player)\n self.write_message(msg, delay_time=1)\n self.display_current_player(self.current_player)\n # Ask for the partner card\n self.table_status[\"partner\"] = self.players[self.current_player].make_decision(self.game_state, 1)\n\n # Setup the table status before the play starts\n self.table_status['partner reveal'] = False\n self.table_status[\"trump suit\"] = self.table_status[\"bid\"] % 10\n self.table_status[\"trump broken\"] = False\n self.table_status['played cards'] = [0, 0, 0, 0]\n if self.table_status['trump suit'] == 5:\n self.table_status[\"leading player\"] = self.current_player\n else:\n self.table_status[\"leading player\"] = (self.current_player + 1) % 4\n self.table_status['defender']['target'] = self.table_status[\"bid\"] // 10 + 6\n self.table_status['attacker']['target'] = 14 - self.table_status['defender']['target']\n\n # Set the roles of the players\n self.players[self.current_player].role = PlayerRole.DEFENDER\n\n self.write_message('Bidding Complete', delay_time=0)\n msg = 'Trump: {1:s}, Partner: {0:s}'.format(cards.get_card_string(self.table_status[\"partner\"]),\n cards.get_suit_string(self.table_status['trump suit']))\n self.write_message(msg, line=1, delay_time=1)\n return True\n\n def play_a_round(self):\n \"\"\"\n Ask each player to play a valid card and determine the winner of the round\n :return: None\n \"\"\"\n if not any(self.table_status[\"played cards\"]):\n # Leading player starts with the leading card, which determines the leading suit\n self.current_player = self.table_status['leading player']\n self.display_current_player(self.current_player)\n card = self.players[self.current_player].make_decision(self.game_state, 0)\n self.table_status[\"played cards\"][self.current_player] = card\n self.players_playzone[self.current_player].add_card(card)\n elif not all(self.table_status[\"played cards\"]):\n # Subsequent player make their plays, following suit if possible\n self.display_current_player(self.current_player)\n print(\"Player {0:d}\\n\".format(self.current_player))\n card = self.players[self.current_player].make_decision(self.game_state, 1)\n self.players_playzone[self.current_player].add_card(card)\n self.table_status[\"played cards\"][self.current_player] = card\n else:\n # Once all player played, find out who wins\n leading_card = self.table_status[\"played cards\"][self.table_status['leading player']]\n card_suits = [card.suit() for card in self.table_status[\"played cards\"]]\n card_nums = [card.number() for card in self.table_status[\"played cards\"]]\n follow_suits = [suit == leading_card.suit() for suit in card_suits]\n trumps = [suit == self.table_status['trump suit'] for suit in card_suits]\n trump_played = any(trumps)\n\n # Break trump if the trump suit is played\n if not self.table_status['trump broken']:\n if trump_played:\n self.table_status['trump broken'] = True\n self.write_message(\"Trump Broken!\", delay_time=1)\n\n # Determine which players to check for winner, and determine winner\n valid_nums = [card_nums[i] * ((follow_suits[i] and not trump_played) or trumps[i]) for i in range(4)]\n winning_player = valid_nums.index(max(valid_nums))\n self.write_message(\"Player {0:d} wins!\\n\".format(winning_player), delay_time=1)\n self.players[winning_player].score += 1\n self.update_player_wins(winning_player)\n\n # Clean up the cards, update score, set the next leading player, update round history\n for deck in self.players_playzone:\n self.discard_deck.append(deck.remove_card())\n\n if self.players[winning_player].role == PlayerRole.DEFENDER:\n self.table_status['defender']['wins'] += 1\n elif self.players[winning_player].role == PlayerRole.ATTACKER:\n self.table_status['attacker']['wins'] += 1\n\n self.table_status['leading player'] = winning_player\n self.table_status['round history'].append(copy.copy(self.table_status[\"played cards\"]))\n self.update_team_scores()\n self.table_status[\"played cards\"] = [0]*4\n self.current_round += 1\n self.update_table.emit()\n return\n\n if not self.table_status['partner reveal']:\n if card.value == self.table_status['partner']:\n self.table_status['partner reveal'] = True\n self.write_message(\"Partner Revealed!\", delay_time=1)\n self.reveal_all_roles(self.current_player)\n self.update_all_players(role=True, wins=False)\n\n self.current_player += 1\n self.current_player %= 4\n self.update_table.emit()\n time.sleep(0.5)\n\n def reveal_all_roles(self, partner):\n self.players[partner].role = PlayerRole.DEFENDER\n self.table_status['defender']['wins'] += self.players[partner].score\n for i in range(4):\n if self.players[i].role == PlayerRole.UNKNOWN:\n self.players[i].role = PlayerRole.ATTACKER\n self.table_status['attacker']['wins'] += self.players[i].score\n\n def reset_game(self):\n for player in self.players:\n while not player.is_empty():\n self.discard_deck.append(player.remove_card())\n player.score = 0\n player.role = PlayerRole.UNKNOWN\n\n for i in range(4):\n self.update_players_role(i)\n self.update_player_wins(i)\n self.table_status['defender']['wins'] = 0\n self.table_status['attacker']['wins'] = 0\n self.table_status[\"played cards\"] = [0]*4\n self.table_status['round history'] = []\n self.current_round = 0\n self.write_message(\"\", line=1, update_now=False)\n self.write_message(\"\", line=2)\n self.display_current_player()\n print(len(self.discard_deck))\n self.update_table.emit()\n\n\nclass Player(cards.Deck):\n \"\"\"\n A player is essentially a Deck with decision making function or AI component if it is a bot\n that returns a valid action for the Table/Board.\n\n The player has the knowledge of Table status in the form of a dictionary (as it is mutable, thus passed by ref)\n so all validation is done by the player\n\n Possible decisions, each decision has to be enum maybe:\n - Query the board status (i.e. current round, player status), AI most likely need a lot more\n - Query the last round\n - Attempt to play a card\n - Play the validate move\n\n The player also implements method to play from the terminal\n if it is not a bot.\n\n \"\"\"\n def __init__(self, *args, ai_component=None, **kwargs):\n super().__init__(*args, **kwargs)\n\n self.role = PlayerRole.UNKNOWN\n self.AI = ai_component\n self._table_status = None # This is found in Table and updated through Table\n self.score = 0\n\n def connect_to_table(self, table):\n self._table_status = table\n\n def add_ai(self, ai_comp):\n self.AI = ai_comp\n ai_comp.connect_to_player(self)\n\n def make_decision(self, game_state, sub_state):\n \"\"\"\n The player will need to make a decision depending on the game state and sub-state\n :param game_state: Current game state\n :param sub_state: Sub-state which affects the output for the current game state\n :return: For Bidding: Either a bid or a partner call, int\n For Playing: A Card\n For Reshuffle: bool, True to reshuffle, False otherwise\n \"\"\"\n if game_state == GameState.POINT_CHECK:\n if self.AI:\n return self.AI.request_reshuffle()\n if input(\"Reshuffle? (y/n)\").lower() == 'y':\n return self.request_reshuffle()\n if game_state == GameState.BIDDING:\n if sub_state == 0:\n if self.AI:\n return self.AI.make_a_bid()\n return self.make_a_bid()\n else:\n if self.AI:\n return self.AI.call_partner()\n return self.call_partner()\n if game_state == GameState.PLAYING:\n if self.AI:\n play = self.AI.make_a_play(sub_state)\n [_, pos] = self.check_card_in(play)\n return self.remove_card(pos)\n return self.make_a_play(sub_state)\n\n def make_a_bid(self):\n \"\"\"\n The procedure to make a bid\n :return: A valid bid number\n \"\"\"\n while True:\n bid = input(\"Please input a bid in the format 'number' + 'suit' \\n\"\n \"To pass, enter nothing. \\n\"\n \"e.g 4d is 4 Diamond, 6n is 6 No Trump \\n\")\n\n if not bid:\n return 0\n\n bid = cards.convert_bid_string(bid)\n if bid < 0:\n print(\"Error in processing bid\")\n continue\n\n if self._table_status[\"bid\"] < bid:\n return bid\n else:\n if bid > 75:\n print(\"You cannot bid beyond 7 No Trump\")\n else:\n print(\"You might need to bid higher\")\n\n def call_partner(self):\n \"\"\"\n The procedure to call a partner\n :return: A valid card value\n \"\"\"\n current_card_values = self.get_deck_values()\n while True:\n partner = input(\"Please call your partner card. Enter card number + suit number \\n\"\n \"e.g. qs is Queen Spade, 8c is 8 Clubs, ah is Ace Hearts\\n\")\n\n partner = cards.convert_input_string(partner)\n if partner in current_card_values:\n print(\"Please call a card outside of your hand\")\n elif cards.card_check(partner):\n return partner\n else:\n print(\"Invalid card call\")\n\n def make_a_play(self, substate):\n \"\"\"\n The procedure to make a play in a round\n :return: A valid Card\n \"\"\"\n while True:\n play = input(\"Please play a card.Enter card number + suit number \\n\"\n \"e.g. qs is Queen Spade, 8c is 8 Clubs, ah is Ace Hearts\\n\")\n if play == \"v\":\n pprint.pprint(self._table_status)\n else:\n play = cards.convert_input_string(play)\n if play > 0:\n if substate == 0:\n valid = self.check_for_valid_plays(play, True)\n else:\n valid = self.check_for_valid_plays(play, False)\n\n if valid:\n [_, pos] = self.check_card_in(play)\n return self.remove_card(pos)\n\n print(\"Invalid play\")\n\n def view_last_round(self):\n pass\n\n def check_for_valid_plays(self, card, leading):\n \"\"\"\n Check if the card played is valid\n :param card: int\n :param leading: bool\n :return:\n \"\"\"\n if not self.check_card_in(card):\n return False\n card_suit = cards.get_card_suit(card)\n if leading:\n if not self._table_status['trump broken'] and \\\n card_suit == self._table_status['trump suit']:\n if any([not cards.get_card_suit(crd) == self._table_status['trump suit'] for crd in self.get_deck_values()]):\n return False\n else:\n leading_card_suit = self._table_status['played cards'][self._table_status[\"leading player\"]].suit()\n if not card_suit == leading_card_suit and \\\n any([cards.get_card_suit(crd) == leading_card_suit for crd in\n self.get_deck_values()]):\n return False\n\n return True\n\n def get_card_points(self):\n suit_points = 0\n card_points = []\n current_suit = 1\n card_position = 0\n for (i, card) in enumerate(self.cards):\n if card.suit() != current_suit:\n suit_points += (i-card_position) // 5\n card_position = i\n current_suit = card.suit()\n card_points.append(max(0, card.number() - 10))\n suit_points += (STARTING_HAND-card_position) // 5\n return suit_points + sum(card_points)\n\n def request_reshuffle(self):\n # Players can choose NOT to reshuffle\n # But always reshuffle for simplicity\n return True\n\n\nclass MainPlayer(cards.PlayerDeck):\n def __init__(self, *args, ai_component=None, **kwargs):\n super().__init__(*args, **kwargs)\n\n self.AI = ai_component\n self.table_status = None # This is found in Table and updated through Table\n\n def connect_to_table(self, table):\n self.table_status = table\n\n def make_a_bid(self):\n pass\n\n def make_a_play(self):\n pass\n\n def view_last_round(self):\n pass\n\n def check_for_valid_moves(self):\n pass\n\n\nclass TestView(view.PygView):\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.table = Table(0, 0, self.width, self.height, (0, 0, 255))\n self.table.update_table.connect(self.draw_table)\n self.draw_table()\n\n def draw_table(self, **kwargs):\n self.screen.blit(self.background, (0, 0))\n self.screen.blit(self.table.background, self.table.get_pos())\n for player in self.table.players:\n self.screen.blit(player.deck_surface, player.get_pos())\n for playerzone in self.table.players_playzone:\n self.screen.blit(playerzone.deck_surface, playerzone.get_pos())\n pygame.display.flip()\n\n def run(self):\n running = True\n while running:\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n running = False\n elif event.type == pygame.KEYDOWN:\n if event.key == pygame.K_ESCAPE:\n running = False\n if event.key == pygame.K_p:\n print('add cards')\n pass\n\n # milliseconds = self.clock.tick(self.fps)\n # self.playtime += milliseconds / 1000.0\n\n # self.draw_function()\n\n pygame.quit()\n\n\nif __name__ == '__main__':\n test_view = TestView(900, 600, clear_colour=(0, 0, 0))\n test_view.run()\n","sub_path":"players.py","file_name":"players.py","file_ext":"py","file_size_in_byte":33164,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"348462650","text":"import numpy as np\nimport pandas as pd\nfrom cea.constants import HOURS_IN_YEAR\n\n\ndef calculate_incident_radiation(radiation, radiation_csv):\n \"\"\"\n Calculate the output file \"radiation_csv\" based on the radiation dataframe.\n :param radiation:\n :param radiation_csv:\n :return:\n \"\"\"\n\n # Import Radiation table and compute the Irradiation in W in every building's surface\n column_names = ['T%i' % (i + 1) for i in range(HOURS_IN_YEAR)]\n for column in column_names:\n # transform all the points of solar radiation into Wh\n radiation[column] = radiation[column] * radiation['Awall_all']\n\n # sum up radiation load per building\n # NOTE: this looks like an ugly hack because it is: in order to work around a pandas MemoryError, we group/sum the\n # columns individually...\n grouped_data_frames = {}\n for column in column_names:\n df = pd.DataFrame(data={'Name': radiation['Name'],\n column: radiation[column]})\n grouped_data_frames[column] = df.groupby(by='Name').sum()\n radiation_load = pd.DataFrame(index=grouped_data_frames.values()[0].index)\n for column in column_names:\n radiation_load[column] = grouped_data_frames[column][column]\n\n incident_radiation = np.round(radiation_load[column_names], 2)\n incident_radiation.to_csv(radiation_csv)\n\n return # total solar radiation in areas exposed to radiation in Watts\n\n\nif __name__ == '__main__':\n import argparse\n\n parser = argparse.ArgumentParser()\n parser.add_argument('--radiation-pickle', help='path to a pickle of the radiation dataframe')\n parser.add_argument('--radiation-csv', help='path to a pickle of the radiation dataframe')\n args = parser.parse_args()\n\n radiation = pd.read_pickle(args.radiation_pickle)\n calculate_incident_radiation(radiation, args.radiation_csv)\n","sub_path":"legacy/radiation_arcgis/calculate_incident_radiation.py","file_name":"calculate_incident_radiation.py","file_ext":"py","file_size_in_byte":1859,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"465625228","text":"# -*- coding:utf-8 -*-\r\n# Author: washing\r\n# DateTime: 2022/08/30 15:11\r\n# File: 0998.py\r\n# Desc: \r\n\r\n\r\n# Definition for a binary tree node.\r\n# class TreeNode:\r\n# def __init__(self, val=0, left=None, right=None):\r\n# self.val = val\r\n# self.left = left\r\n# self.right = right\r\nclass Solution:\r\n def insertIntoMaxTree(self, root: Optional[TreeNode], val: int) -> Optional[TreeNode]:\r\n t = TreeNode(val)\r\n if root.val < val:\r\n t.left = root\r\n return t\r\n cur = root\r\n while cur:\r\n if not cur.right:\r\n cur.right = t\r\n break\r\n elif cur.right.val < val:\r\n t.left = cur.right\r\n cur.right = t\r\n break\r\n cur = cur.right\r\n return root\r\n","sub_path":"Solutions/0998/0998.py","file_name":"0998.py","file_ext":"py","file_size_in_byte":819,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"586201108","text":"import time\nfrom neopixel import *\nimport argparse\n\n# LED strip configuration:\nLED_COUNT = 350 # Number of LED pixels.\nLED_PIN = 18 # GPIO pin connected to the pixels (18 uses PWM!).\n#LED_PIN = 10 # GPIO pin connected to the pixels (10 uses SPI /dev/spidev0.0).\nLED_FREQ_HZ = 800000 # LED signal frequency in hertz (usually 800khz)\nLED_DMA = 10 # DMA channel to use for generating signal (try 10)\nLED_BRIGHTNESS = 180 # Set to 0 for darkest and 255 for brightest\nLED_INVERT = False # True to invert the signal (when using NPN transistor level shift)\nLED_CHANNEL = 0 # set to '1' for GPIOs 13, 19, 41, 45 or 53\nSTEP_SIZE = 28 # The size of a step in the unit of LED lights which will be lit up\n\n\ndef clearStrip(strip):\n #Clear all the lights on the strip\n for i in range(0, strip.numPixels(), 1):\n strip.setPixelColor(i, 0)\n strip.show()\n\ndef lightStep(strip, stepNum):\n clearStrip(strip)\n for i in range(stepNum * STEP_SIZE, stepNum * STEP_SIZE + STEP_SIZE, 1):\n if (i==0):\n strip.setPixelColor(i+8, Color(0, 128,0))\n else:\n strip.setPixelColor(i, Color(0, 128, 0))\n strip.show()\n\ndef lightIndv(strip, stepNum):\n clearStrip(strip)\n # For the first step light LEDs 0-35\n if (stepNum == 0):\n for i in range(0, 35, 1):\n strip.setPixelColor(i, Color(0, 128, 0))\n strip.show()\n \n else:\n for i in range(stepNum*STEP_SIZE+7, stepNum*STEP_SIZE+7 + STEP_SIZE, 1):\n strip.setPixelColor(i, Color(0, 128, 0))\n strip.show()\n\n\ndef flashVirginiaTech(strip, secs, wait_ms=500):\n #Flash every other light maroon and orage, then switch\n p = 0\n while p < secs:\n for i in range(0, strip.numPixels(), 1):\n if i % 2 == 0:\n strip.setPixelColor(i, Color(0, 128, 0))\n if i % 2 == 1:\n strip.setPixelColor(i, Color(128, 255, 0))\n strip.show()\n time.sleep(wait_ms/1000.0)\n for i in range(0, strip.numPixels(), 1):\n if i % 2 == 0:\n strip.setPixelColor(i, Color(128, 255, 0))\n if i % 2 == 1:\n strip.setPixelColor(i, Color(0, 128, 0))\n strip.show()\n time.sleep(wait_ms/1000.0)\n p += 1\n\n\n# Main program logic follows:\nif __name__ == '__main__':\n # Process arguments\n parser = argparse.ArgumentParser()\n parser.add_argument('-c', '--clear', action='store_true', help='clear the display on exit')\n args = parser.parse_args()\n\n # Create NeoPixel object with appropriate configuration.\n strip = Adafruit_NeoPixel(LED_COUNT, LED_PIN, LED_FREQ_HZ, LED_DMA, LED_INVERT, LED_BRIGHTNESS, LED_CHANNEL)\n # Intialize the library (must be called once before other functions).\n strip.begin()\n\n print ('Press Ctrl-C to quit.')\n if not args.clear:\n print('Use \"-c\" argument to clear LEDs on exit')\n try:\n print(\"Let's try lighting up steps 1-10\\n\")\n for i in range(0, 10, 1):\n for x in range(0,10,1):\n print(\"Lighting step %d\", x)\n lightIndv(strip, x)\n time.sleep(500/1000.0)\n clearStrip(strip)\n\n #flashVirginiaTech(strip, 2)\n #clearStrip(strip)\n\n #x =2 \n #print(\"Turn on step {}\".format(x))\n #lightIndv(strip, x)\n #time.sleep(10)\n #clearStrip(strip)\n\n except KeyboardInterrupt:\n if args.clear == False:\n clearStrip(strip)\n\n\n","sub_path":"testLights.py","file_name":"testLights.py","file_ext":"py","file_size_in_byte":3531,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"295278790","text":"#Node Representation\nclass Node:\n\n #Constructor\n def __init__(self, state, parent_node, operator, depth, path_cost):\n self.state = state\n self.parent_node = parent_node\n self.operator = operator\n self.depth = depth\n self.path_cost = path_cost\n\n \n\n#Problem Formulation\nclass General_problem:\n\n #Constructor\n def __init__(self, initial_state, operators, goal_test):\n self.initial_state = initial_state\n self.operator = operators #List of available operators\n self.goal_test = goal_test\n self.path_cost = 0 #initial path cost always 0\n\n #Define General Successor Function that takes a node and returns (state, action) pairs possible resulting\n def successor_function(self, node):\n successors = [];\n return successors\n \n #Define General Cost Function that whenever we expand a node it updates its cost by adding the edge cost\n def path_cost_function(self, node):\n extra_cost = 0\n #Given a file name such us pathcosts\n text_name = input(\"Please Enter your file name: \")\n text_name = text_name +\".txt\"\n f = open(text_name, 'r')\n for i in f:\n #for each line in the file\n line = i.split()\n #read the distance between the node and its parent node line[2]\n if (line[0] == node.parent_node.state and line[1] == node.state) or (line[1] == node.parent_node.state and line[0] == node.state):\n extra_cost = int(line[2])\n break\n #add the cost of getting to that state from the previous state to the total path cost leading to the parent node\n return extra_cost + node.parent_node.path_cost\n","sub_path":"General.py","file_name":"General.py","file_ext":"py","file_size_in_byte":1576,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"99172945","text":"\"\"\"\nThis spider is a ApprenticeshipVacancy spider created on top of the ATSSpider\nscrapy crawl apprenticeshipvacancy -a mining_job_id=9999 -a iteration=1 -a extract=1 -a url=\"https://apprenticeshipvacancymatchingservice.lsc.gov.uk/navms/Forms/Vacancy/SearchVacancy.aspx\"\n\nsample url:\n https://apprenticeshipvacancymatchingservice.lsc.gov.uk/navms/Forms/Vacancy/SearchVacancy.aspx\n\"\"\"\n\nfrom urlparse import urljoin\nfrom scrapy.http import Request, FormRequest\nfrom scrapy.selector import Selector\nfrom urlparse import urlparse\n\nfrom brightcorp.base.atsspiders import ATSSpider\nfrom brightcorp.items import BrightcorpItemLoader\nfrom brightcorp.processors import Prefix, NormalizedJoin, ConvertDateString\nfrom brightcorp.lib.utils import get_hidden_inputs\n\n\nclass ApprenticeshipVacancy(ATSSpider):\n\n name = \"apprenticeshipvacancy\"\n ref_re = r'(\\d+)$'\n page = 0\n download_delay = 0.3\n tot_pages = ''\n is_first_page = True\n\n def parse(self, response):\n formdata = get_hidden_inputs(response)\n formdata['ctl00$ContentBody$SearchByRadioButtonList'] = 'Occupation'\n formdata['ctl00$ContentBody$ApprenticeshipOccupationDropDownList'] = '0'\n formdata['ctl00$ContentBody$ApprenticeshipFrameworkListBox'] = '0'\n formdata['ctl00$ContentBody$ApprenticeshiptypeOldValue'] = '999'\n formdata['ctl00$ContentBody$ApprenticeshipCategoryRadioButtonList'] = '0'\n formdata['ctl00$ContentBody$ApprenticeshiptypeDropDownList'] = '999'\n formdata['ctl00$ContentBody$SearchButton'] = 'Search'\n req = FormRequest(\n response.url, formdata=formdata, callback=self.parse_jobs_list\n )\n yield req\n\n def parse_jobs_list(self, response):\n sel = Selector(response)\n if not self.tot_pages:\n tot_pages = sel.xpath(\n '//select[contains(@name, \"PagesDropDownList\")]/option[1]/text()'\n ).re(self.ref_re)\n if tot_pages:\n self.tot_pages = tot_pages[0]\n\n jobs = sel.xpath('//table[@class=\"results\"]//tr[td]')\n for job in jobs:\n job_url = job.xpath('./td[1]/a/@href').extract()\n if job_url:\n job_url = urljoin(response.url, job_url[0])\n meta = {\n 'title': job.xpath('./td[1]/a//text()').extract(),\n 'loc': job.xpath('./td[3]/span/text()').extract(),\n 'cat': job.xpath('./td[4]/text()').extract(),\n 'date': job.xpath('./td[5]/text()').extract(),\n }\n yield Request(\n job_url, callback=self.parse_job_callback(), meta=meta\n )\n\n if self.is_first_page and self.tot_pages:\n self.is_first_page = False\n for i in xrange(1, int(self.tot_pages) + 1, 1):\n formdata = get_hidden_inputs(response)\n formdata['ctl00$ContentBody$ResultsPager$PagesDropDownList'] = str(i)\n yield FormRequest(\n response.url, callback=self.parse_jobs_list, formdata=formdata\n )\n\n def parse_job(self, response):\n loader = BrightcorpItemLoader(response=response)\n loader.add_value('url', response.url)\n loader.add_value('title', response.meta.get('title'))\n loader.add_value('location', response.meta.get('loc'))\n loader.add_value('jobcategory', response.meta.get('cat'))\n loader.add_value(\n 'expiration_date', response.meta.get('date'), ConvertDateString('%d/%m/%Y')\n )\n loader.add_value(\n 'referencenumber', response.url,\n Prefix('%s-' % self.name), re=self.ref_re\n )\n\n loader.add_xpath('baseSalary', '//p[@id=\"vacancy-wage\"]/text()')\n loader.add_xpath('workhours', '//p[@itemprop=\"workHours\"]/text()')\n loader.add_xpath('requirements', '//section[@id=\"course-info\"]/node()')\n loader.add_xpath('duration', '//p[@id=\"vacancy-expected-duration\"]/text()')\n loader.add_xpath(\n 'description',\n [\n '//h2[text()=\"Apprenticeship summary\"]|//div[@itemprop=\"responsibilities\"]',\n '//h2[text()=\"Traineeship details\"]|//h2[text()=\"Traineeship details\"]/following-sibling::div[1]'\n ]\n )\n loader.add_xpath(\n 'responsibilities',\n '//h2[text()=\"Training provider\"]|//h2[text()=\"Training provider\"]/following-sibling::div[1]'\n )\n loader.add_xpath(\n 'company',\n [\n '//p[@id=\"vacancy-employer-name\"]/text()',\n '//h3[text()=\"Training provider\"]/following-sibling::p[1]/text()'\n ]\n )\n loader.add_xpath(\n 'company_description',\n '//h2[text()=\"About the employer\"]|//h2[text()=\"About the employer\"]/following-sibling::div[1]',\n NormalizedJoin()\n )\n\n yield loader.load_item()\n","sub_path":"brightcorp/brightcorp/spiders/apprenticeshipvacancy.py","file_name":"apprenticeshipvacancy.py","file_ext":"py","file_size_in_byte":4987,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"635090837","text":"from django import forms\nfrom django.urls import reverse\nfrom django.forms import Textarea,CheckboxSelectMultiple\n\nfrom .models import Post, Comment, Profile\nfrom django.contrib.auth import get_user_model\n\nfrom crispy_forms.helper import FormHelper\nfrom crispy_forms.layout import Div, Layout, Submit\nfrom crispy_forms.bootstrap import InlineField\n\nfrom django.contrib.auth.forms import PasswordChangeForm\nfrom django_registration.forms import RegistrationFormCaseInsensitive, RegistrationFormTermsOfService\n\nclass CrispyModelForm(forms.ModelForm):\n\tdef __init__(self, *args, **kwargs):\n\t\tsuper(CrispyModelForm, self).__init__(*args, **kwargs)\n\t\tself.helper = FormHelper(self)\n\nclass CrispyForm(forms.Form):\n\tdef __init__(self, *args, **kwargs):\n\t\tsuper(CrispyForm, self).__init__(*args, **kwargs)\n\t\tself.helper = FormHelper(self)\n\nclass PostForm(CrispyModelForm):\n\tclass Meta:\n\t\tmodel = Post\n\t\tfields = ['title','entry','private','show_recent']\n\t\tlabels = {\n\t\t\t'show_recent': 'Show on Front Page',\n\t\t\t'private': 'Visibility'\n\t\t}\n\nclass CommentForm(CrispyModelForm):\n\tclass Meta:\n\t\tmodel = Comment\n\t\tfields = ['entry']\n\t\tlabels = {\n\t\t\t'entry': ''\n\t\t}\n\t\twidgets = {\n\t\t\t'entry': Textarea(attrs={'cols': 40, 'rows': 4})\n\t\t}\n\nclass ConfirmDeleteForm(CrispyForm):\n\tdef __init__(self, *args, **kwargs):\n\t\tsuper(ConfirmDeleteForm, self).__init__(*args, **kwargs)\n\t\tself.helper.add_input(Submit('submit', 'Cancel', css_class='btn-secondary'))\n\t\tself.helper.add_input(Submit('submit', 'Delete', css_class='btn-primary'))\n\nclass UserForm(CrispyModelForm):\n\tdef __init__(self, *args, **kwargs):\n\t\tsuper(UserForm, self).__init__(*args, **kwargs)\n\t\tself.helper.form_tag = False\n\n\tclass Meta:\n\t\tmodel = get_user_model()\n\t\tfields = ['last_name','email']\n\t\tlabels = {\n\t\t\t'last_name': 'Name',\n\t\t\t'email': 'Email Address'\n\t\t}\n\nclass UserProfileForm(CrispyModelForm):\n\tdef __init__(self, *args, **kwargs):\n\t\tsuper(UserProfileForm, self).__init__(*args, **kwargs)\n\t\tself.helper.form_tag = False\n\t\tself.helper.add_input(Submit('submit', 'Save', css_class='btn-primary'))\n\n\tclass Meta:\n\t\tmodel = Profile\n\t\tfields = ['profile','avatar','banner']\n\nclass PasswordChangeForm(PasswordChangeForm):\n\tdef __init__(self, *args, **kwargs):\n\t\tsuper().__init__(*args, **kwargs)\n\t\tself.helper = FormHelper(self)\n\t\tself.helper.add_input(Submit('submit', 'Change Password', css_class='btn-primary'))\n\nclass UploadFilesForm(CrispyForm):\n\tfiles = forms.FileField(widget=forms.ClearableFileInput(attrs={'multiple': True}))\n\n\tdef __init__(self, *args, **kwargs):\n\t\tsuper(UploadFilesForm, self).__init__(*args, **kwargs)\n\t\tself.helper.form_tag = False\n\t\tself.helper.form_show_labels = False\n\nclass CreateFolderForm(CrispyForm):\n\tnew_folder = forms.CharField(strip=True)\n\n\tdef __init__(self, *args, **kwargs):\n\t\tsuper(CreateFolderForm, self).__init__(*args, **kwargs)\n\t\tself.helper.form_tag = False\n\t\tself.helper.form_show_labels = False\n\nclass FileRenameForm(forms.Form):\n\told_name = forms.CharField(strip=True)\n\tnew_name = forms.CharField(strip=True)\n\nclass FileDeleteForm(forms.Form):\n\tdelete = forms.CharField(strip=True)\n\nclass RegistrationForm(RegistrationFormCaseInsensitive, RegistrationFormTermsOfService):\n\tpass\n\nbulk_actions = [\n\t('visible_only_me','Set Visible to Only Me'),\n\t('visible_registered','Set Visible to Registered Users'),\n\t('visible_regular','Set Visible to Regular Users'),\n\t('visible_public','Set Visible to Public'),\n\t('visible_staff','Set Visible to Staff'),\n\t('delete','Delete Posts (NO UNDO)'),\n]\n\nclass UserManagePostsForm(CrispyForm):\n\tposts = forms.ModelMultipleChoiceField(queryset=None,widget=CheckboxSelectMultiple)\n\taction = forms.ChoiceField(choices=bulk_actions)\n\tdef __init__(self, user, *args, **kwargs):\n\t\tsuper().__init__(*args, **kwargs)\n\t\tself.fields['posts'].queryset = Post.posts_visible(user).filter(user=user)\n\t\tself.helper.form_tag = False\n\t\tself.helper.field_template = 'bootstrap4/layout/inline_field.html'\n\t\tself.helper.layout = Div(Layout(\n\t\t\t\tInlineField('action', css_class='flex-grow-1'),\n\t\t\t\tSubmit('submit', 'Update Posts', css_class='btn-primary ml-3'),\n\t\t),css_class='form-inline justify-content-end')\n\nfrom django.utils import timezone\nclass UserManagePostsByDateForm(CrispyForm):\n\tolder_than = forms.DateTimeField(initial=timezone.now)\n\taction = forms.ChoiceField(choices=bulk_actions)\n\tconfirm = forms.BooleanField(initial=False, required=True)\n\tdef __init__(self, *args, **kwargs):\n\t\tsuper().__init__(*args, **kwargs)\n\t\tself.helper.add_input(Submit('submit', 'Update Posts', css_class='btn-primary'))\n","sub_path":"user/forms.py","file_name":"forms.py","file_ext":"py","file_size_in_byte":4523,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"645055301","text":"#!/usr/bin/env python\n\n\"\"\"\n#The differential controller script is responsible sending velocity commands\n#to the Arduino that controls de DC Motors. Here the Twist message coming from \n#the teleop device is converted into MotorCommand message used by the Arduino.\n\"\"\"\n\nimport rospy\nfrom deepfind_package.msg import MotorCommand\nfrom geometry_msgs.msg import Twist\n\n# Motor global variables \nglobal dict\ndict = {'MOTOR_CAP': 40, 'FORWARD': 0, 'BACKWARD': 1}\n\n# Converts incoming Twist message to Motorommand message and\n# sends the command to the Arduino.\ndef diff_control_callback(twistData):\n command = MotorCommand()\n command.leftMotorPower = int((twistData.linear.x + twistData.angular.z) * dict['MOTOR_CAP'])\n command.rightMotorPower = int((twistData.linear.x - twistData.angular.z) * dict['MOTOR_CAP'])\n\n if command.leftMotorPower > 0:\n \tcommand.leftMotorDirection = dict['FORWARD']\n else:\n command.leftMotorDirection = dict['BACKWARD']\n \n if command.rightMotorPower > 0:\n command.rightMotorDirection = dict['FORWARD']\n else:\n command.rightMotorDirection = dict['BACKWARD']\n\n command.leftMotorPower = abs(command.leftMotorPower)\n command.rightMotorPower = abs(command.rightMotorPower)\n \t\n pub = rospy.Publisher('motor_speed', MotorCommand, queue_size = 10)\n pub.publish(command)\n\n# Main function, waits for new Twist messages to arrive\n# Creates node and subscriber\ndef diff_controller_listener():\n rospy.init_node('differential_controller')\n rospy.Subscriber('cmd_vel', Twist, diff_control_callback)\n rospy.spin()\n\nif __name__ == '__main__':\n diff_controller_listener()\n\n\n","sub_path":"catkin_ws/src/deepfind_package/scripts/diff_motor_controller.py","file_name":"diff_motor_controller.py","file_ext":"py","file_size_in_byte":1605,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"636744835","text":"import numpy as np\nimport matplotlib.pyplot as plt\ndef model(x, a, b, c):\n return a*np.square(x) + b*x + c\n\ndef loglikelihood(x_obs, y_obs, sigma_y_obs, a, b, c):\n d = y_obs - model(x_obs, a, b, c)\n d = d/sigma_y_obs\n d = -0.5 * np.sum(d**2)\n return d\n\ndef logprior(a, b, c):\n p = -np.inf\n if -3 < a < -0.5 and -3 < b < 3 and -1 < c < 10 :\n p = 0.0\n return p\n\nx_obs = np.array([-2.0,1.3,0.4,5.0,0.1, -4.7, 3.0, -3.5,-1.1])\ny_obs = np.array([ -1.931, 2.38, 1.88, -24.22, 3.31, -21.9, -5.18, -12.23, 0.822])\nsigma_y_obs = np.array([ 2.63, 6.23, -1.461, 1.376, -4.72, 1.313, -4.886, -1.091, 0.8054]) \n\nN = 50000\nlista_a = np.zeros(N)\nlista_b = np.zeros(N)\nlista_c = np.zeros(N)\nlogposterior = np.zeros(N)\n\nlista_a[0] = np.random.uniform(-3, -0.5)\nlista_b[0] = np.random.uniform(-3, 3)\nlista_c[0] = np.random.uniform(-1, 10)\ni = 0\nlogposterior[0] = loglikelihood(x_obs, y_obs, sigma_y_obs, lista_a[i-1], lista_b[i-1], lista_c[i-1]) + logprior(lista_a[i-1], lista_b[i-1], lista_c[i-1])\n\nsigma_delta_c = 0.1\nsigma_delta_b = 0.08\nsigma_delta_a = 0.05\n\n#### DATOS\nfor i in range(1,N):\n propuesta_a = np.random.normal(loc=lista_a[i-1], scale=sigma_delta_a)\n propuesta_b = np.random.normal(loc=lista_b[i-1], scale=sigma_delta_b)\n propuesta_c = np.random.normal(loc=lista_c[i-1], scale=sigma_delta_c)\n\n logposterior_viejo = loglikelihood(x_obs, y_obs, sigma_y_obs, lista_a[i-1], lista_b[i-1], lista_c[i-1]) + logprior(lista_a[i-1], lista_b[i-1], lista_c[i-1])\n logposterior_nuevo = loglikelihood(x_obs, y_obs, sigma_y_obs, propuesta_a, propuesta_b, propuesta_c) + logprior(propuesta_a, propuesta_b, propuesta_c)\n\n r = min(1,np.exp(logposterior_nuevo-logposterior_viejo))\n alpha = np.random.random()\n if(alpha={a_model:.4f}$\")\n\nplt.subplot(1, 3, 2)\nplt.hist(lista_b, bins=40, density = True)\nplt.xlabel(\"$b$\")\nplt.title(f\"$={b_model:.4f}$\")\n\nplt.subplot(1, 3, 3)\nplt.hist(lista_c, bins=40, density = True)\nplt.xlabel(\"$c$\")\nplt.title(f\"$={c_model:.4f}$\")\nplt.tight_layout()\n#plt.savefig(\"individual_histograms.pdf\", dpi = 200)\n\nplt.figure(figsize=(12, 7))\nxplot = np.linspace(x_obs.min(), x_obs.max(), 100)\ny_model = model(xplot, a_model , b_model, c_model)\n\nplt.errorbar(x_obs,y_obs, yerr=sigma_y_obs, fmt='o', label = \"Data\")\nplt.plot(xplot, y_model, label = f\"Model: $y = {a_model:.3f}x^2 + {b_model:.3f}x + {c_model:.3f}$\")\nplt.xlabel(\"x\")\nplt.ylabel(\"y\")\nplt.legend(loc = 'best')\nplt.savefig(\"results_plot.pdf\", dpi = 200)\nplt.show()","sub_path":"Exercises/Ejercicio 6/DanielOchoa_Ejercicio6.py","file_name":"DanielOchoa_Ejercicio6.py","file_ext":"py","file_size_in_byte":3025,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"24174591","text":"from zope.pagetemplate import pagetemplatefile, engine\nfrom zope.app.pagetemplate import viewpagetemplatefile\n\nclass Context(engine.ZopeContextBase):\n def translate(self, msgid, domain=None, mapping=None, default=None):\n return i18n.translate(\n msgid, domain, mapping, context=self.principal, default=default)\n\nclass ZopeEngine(engine.ZopeEngine):\n _create_context = Context\n def getContext(self, __namespace=None, **namespace):\n if __namespace:\n if namespace:\n namespace.update(__namespace)\n else:\n namespace = __namespace\n\n context = self._create_context(self, namespace)\n\n # Put principal into context so path traversal can find it\n if 'principal' in namespace:\n context.principal = namespace['principal']\n\n # Put context into context so path traversal can find it\n if 'context' in namespace:\n context.context = namespace['context']\n\n return context\n\nEngine = engine._TrustedEngine(ZopeEngine())\n\nclass AppPT(object):\n def pt_getEngine(self):\n return Engine\n\nclass PageTemplateFile(AppPT, pagetemplatefile.PageTemplateFile):\n\n def __init__(self, filename, _prefix=None):\n _prefix = self.get_path_from_prefix(_prefix)\n super(PageTemplateFile, self).__init__(filename, _prefix)\n\n def pt_getContext(self, instance, **_kw):\n # instance is object with 'context' and 'principal' atttributes.\n namespace = super(PageTemplateFile, self).pt_getContext(**_kw)\n namespace['view'] = instance\n namespace['context'] = context = instance.context\n return namespace\n\n def __call__(self, instance, *args, **keywords):\n namespace = self.pt_getContext(\n instance=instance, args=args, options=keywords)\n s = self.pt_render(\n namespace,\n showtal=getattr(instance, 'showTAL', 0),\n sourceAnnotations=getattr(instance, 'sourceAnnotations', 0),\n )\n return s\n\n def __get__(self, instance, type):\n return viewpagetemplatefile.BoundPageTemplate(self, instance)\n","sub_path":"zc.notification/trunk/src/zc/notification/requestless.py","file_name":"requestless.py","file_ext":"py","file_size_in_byte":2139,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"425344626","text":"destinations = [\"Paris, France\", \"Shanghai, China\", \"Los Angeles, USA\", \"Sao Paulo, Brazil\", \"Cairo, Egypt\"]\ntest_traveler = ['Erin Wilkes', 'Shanghai, China', ['historical site','art']]\n#To obtain the index of a particular destination in the destinations list \ndef get_destination_index(destination):\n for i in range(len(destinations)):\n if( destinations[i] == destination):\n destination_index = i\n return destination_index\n\n#print(get_destination_index(\"Los Angeles, USA\"))\n#print(get_destination_index(\"Hyderabad, India\")) \n#print(destinations.index(\"Hyderabad, India\"))\n\n#to get the index of the location of the traveller(current location)\ndef get_traveler_location(traveler):\n traveler_destination = traveler[1] #as traveller is a list containing name of the traveller, location and interests.\n traveler_destination_index = get_destination_index(traveler_destination)\n return traveler_destination_index\n#example of how to use get_traveler_location with traveler -'test_traveler' which is defined above\ntest_destination_index = get_traveler_location(test_traveler)\n#print(test_destination_index ) #location of test_traveler is printed on terminal\n\n#defining a list of attractions. This list will contain sublists containing each of the attractions present in a destination, hence the range is len()+1\nattractions = [[] for i in range(len(destinations)+1)]\n#print(attractions)- will print a list containing 5 empty sublists as 5 is the no of destinations in total\n\n#The following function adds attraction(s) to a particular destination\ndef add_attraction(destination, attraction):\n destination_index = get_destination_index(destination) \n \n attractions_for_destinations = attractions[destination_index]\n attractions_for_destinations.append(attraction)\n return \n\nadd_attraction(\"Los Angeles, USA\", [\"Venice Beach\", [\"beach\"]])\n#print(attractions)\nadd_attraction(\"Paris, France\", [\"the Louvre\", [\"art\", \"museum\"]])\nadd_attraction(\"Paris, France\", [\"Arc de Triomphe\", [\"historical site\", \"monument\"]])\nadd_attraction(\"Shanghai, China\", [\"Yu Garden\", [\"garden\", \"historcical site\"]])\nadd_attraction(\"Shanghai, China\", [\"Yuz Museum\", [\"art\", \"museum\"]])\nadd_attraction(\"Shanghai, China\", [\"Oriental Pearl Tower\", [\"skyscraper\", \"viewing deck\"]])\nadd_attraction(\"Los Angeles, USA\", [\"LACMA\", [\"art\", \"museum\"]])\nadd_attraction(\"Sao Paulo, Brazil\", [\"So Paulo Zoo\", [\"zoo\"]])\nadd_attraction(\"Sao Paulo, Brazil\", [\"Ptio do Colgio\", [\"historical site\"]])\nadd_attraction(\"Cairo, Egypt\", [\"Pyramids of Giza\", [\"monument\", \"historical site\"]])\nadd_attraction(\"Cairo, Egypt\", [\"Egyptian Museum\", [\"museum\"]])\n#print(attractions)\n\ndef find_attractions(destination, interests): #this function attractions in a particular destination based on your interests\n destination_index = get_destination_index(destination)\n attractions_in_city = attractions[destination_index]\n attraction_with_interests = []\n possible_attraction = []\n attraction_tags = []\n #return interests\n #return attractions_in_city\n for attraction in attractions_in_city:\n possible_attraction=attraction\n attraction_tags=attraction[1]\n #return possible_attraction, attraction_tags, interests\n for interest in interests:\n for tag in attraction_tags:\n if(tag == interest):\n attraction_with_interests.append(possible_attraction[0])\n return attraction_with_interests[0]\n#print(find_attractions('Shanghai, China', ['art', 'museum']))\n\ndef get_attractions_for_traveler(traveler): #main engine which would give you your recommendations according to your desired destinantion and interests\n interests_string = 'a'\n traveler_destination = traveler[1]\n traveler_interests = traveler[2]\n traveler_attractions = find_attractions(traveler_destination, traveler_interests)\n interests_string = 'Hi '+ traveler[0] +', we think you will like these places around '+ traveler_destination + ': '\n for attraction in traveler_attractions:\n interests_string = interests_string + attraction \n interests_string = interests_string \n \n \n return interests_string + \".\"\n\n\n#Enter your information here. I know you need a GUI coz u a rich kid but I a noob so I built this, will add GUI in the future.\nHarsimrat_kohli = get_attractions_for_traveler(['Harsimrat Kohli', 'Paris, France', ['monument']])\nprint(Harsimrat_kohli)\n\n \n \n \n \n \n \n \n \n \n \n\n\n\n\n\n\n\n \n \n\n\n\n\n \n \n\n \n\n","sub_path":"Desktop/Python codes/Boredless_tourist.py","file_name":"Boredless_tourist.py","file_ext":"py","file_size_in_byte":4392,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"603812169","text":"def bubble(ls):\n for i in range(len(ls)-1):\n j=0\n while jls[j+1][0] or(ls[j][0]==ls[j+1][0] and ls[j][1]>ls[j+1][1]):\n ls[j],ls[j+1]=ls[j+1],ls[j]\n j=j+1\n for i in range(len(ls)):\n for j in range(len(ls[i])):\n ls[i][j]=ls[i][j]+1\n return ls\n\ndef canmove(ls):\n n=ls[0]\n m=ls[1]\n canleft=True\n canright=True\n canup=True\n candown=True\n if m==0:\n canleft=False\n if canleft and not barrier.__contains__([n,m-1]):\n return True\n if m==M-1:\n canright=False\n if canright and not barrier.__contains__([n,m+1]):\n return True\n if n==0:\n canup=False\n if canup and not barrier.__contains__([n-1,m]):\n return True\n if n==N-1:\n candown=False\n if candown and not barrier.__contains__([n+1,m]):\n return True\n return False\n\n\nnums=input().split(\" \")\nN=int(nums[0])\nM=int(nums[1])\nbarrier=[]\nfor i in range(N):\n s=input()\n for j in range(M):\n if s[j]=='#':\n barrier.append([i,j])\n#i+j偶数为黑点,奇数为白点\neven=[]\nodd=[]\ncant=[]#不能走的点\nprint(barrier)\nfor i in range(N):\n for j in range(M):\n if not barrier.__contains__([i,j]):\n if not canmove([i,j]):\n cant.append([i,j])\n else:\n if (i+j)%2==0:\n even.append([i,j])\n else:\n odd.append([i,j])\nresult=[]+cant\nif len(even)>len(odd):\n result=result+even\nelif len(odd)>len(even):\n result=result+odd\nprint(result)\nif len(result)==0:\n print(0)\nelse:\n result=bubble(result)\n print(len(result))\n for i in range(len(result)):\n s=str(result[i][0])\n for j in range(1,len(result[i])):\n s=s+\" \"+str(result[i][j])\n print(s)\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n","sub_path":"Code/CodeRecords/2348/60796/292993.py","file_name":"292993.py","file_ext":"py","file_size_in_byte":1868,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"310376222","text":"'''\nCreated on 23 May 2020\n\n@author: snake91\n'''\n#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# vispy: gallery 60\n\n\"\"\"\nDynamic planar graph layout.\n\"\"\"\n\n#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# vispy: gallery 2\n# Copyright (c) Vispy Development Team. All Rights Reserved.\n# Distributed under the (new) BSD License. See LICENSE.txt for more info.\n\n\"\"\"\nMultiple real-time digital signals with GLSL-based clipping.\n\"\"\"\n\nfrom vispy import gloo\nfrom vispy import app\nimport numpy as np\nimport math\n\n# Number of cols and rows in the table.\nnrows = 1\nncols = 1\n\n# Number of signals.\nm = nrows*ncols\n\n# Number of samples per signal.\nn = 1000\n\n# Various signal amplitudes.\namplitudes = .1 + .2 * np.random.rand(m, 1).astype(np.float32)\n\n# Generate the signals as a (m, n) array.\ny = amplitudes * np.random.randn(m, n).astype(np.float32)\n\n# Color of each vertex (TODO: make it more efficient by using a GLSL-based\n# color map and the index).\ncolor = np.repeat(np.random.uniform(size=(m, 3), low=.5, high=.9),\n n, axis=0).astype(np.float32)\n\n# Signal 2D index of each vertex (row and col) and x-index (sample index\n# within each signal).\nindex = np.c_[np.repeat(np.repeat(np.arange(ncols), nrows), n),\n np.repeat(np.tile(np.arange(nrows), ncols), n),\n np.tile(np.arange(n), m)].astype(np.float32)\n\nVERT_SHADER = open(\"/home/snake91/git/ShareCode/stats/cpxnetw/vispyplot/vertexshader.cpp\").read() \nFRAG_SHADER = open(\"/home/snake91/git/ShareCode/stats/cpxnetw/vispyplot/fragshader.cpp\").read()\n\n\nclass Canvas(app.Canvas):\n def __init__(self):\n app.Canvas.__init__(self, title='Use your wheel to zoom!', size = (1280,720),\n keys='interactive')\n self.program = gloo.Program(VERT_SHADER, FRAG_SHADER)\n self.program['a_position'] = y.reshape(-1, 1)\n self.program['a_color'] = color\n self.program['a_index'] = index\n self.program['u_scale'] = (1., 1.)\n self.program['u_size'] = (nrows, ncols)\n self.program['u_n'] = n\n\n gloo.set_viewport(0, 0, *self.physical_size)\n\n self._timer = app.Timer('auto', connect=self.on_timer, start=True)\n\n gloo.set_state(clear_color='black', blend=True,\n blend_func=('src_alpha', 'one_minus_src_alpha'))\n\n self.show()\n\n def on_resize(self, event):\n gloo.set_viewport(0, 0, *event.physical_size)\n\n def on_mouse_wheel(self, event):\n dx = np.sign(event.delta[1]) * .05\n scale_x, scale_y = self.program['u_scale']\n scale_x_new, scale_y_new = (scale_x * math.exp(2.5*dx),\n scale_y * math.exp(0.0*dx))\n self.program['u_scale'] = (max(1, scale_x_new), max(1, scale_y_new))\n self.update()\n\n def on_timer(self, event):\n \"\"\"Add some data at the end of each signal (real-time signals).\"\"\"\n k = 100\n y[:, :-k] = y[:, k:] # here it deletes part of the numbers\n y[:, -k:] = amplitudes * np.random.randn(m, k) # here it adds more numbers\n\n self.program['a_position'].set_data(y.ravel().astype(np.float32))\n self.update()\n\n def on_draw(self, event):\n gloo.clear()\n self.program.draw('line_strip')\n\nif __name__ == '__main__':\n c = Canvas()\n app.run()\n \n \n \n ","sub_path":"stats/cpxnetw/vispyplot/vispyplot.py","file_name":"vispyplot.py","file_ext":"py","file_size_in_byte":3304,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"534042429","text":"import argparse\n\ndef init_parser():\n parser = argparse.ArgumentParser(\n prog='ott controller load',\n formatter_class=argparse.ArgumentDefaultsHelpFormatter\n )\n parser.add_argument(\n 'file',\n default='http://developer.trimet.org/schedule/gtfs.zip',\n help='URL or local directory path to GTFS zip FILE',\n )\n parser.add_argument(\n '-c',\n '--create',\n action='store_true',\n default=False,\n help='create new db...'\n )\n parser.add_argument(\n '-d',\n '--database_url',\n default='sqlite:///gtfsdb.db',\n help='DATABASE URL with appropriate privileges'\n )\n parser.add_argument(\n '-l',\n '--local',\n action='store_true',\n default=False,\n help='local ott table updates to the db (e.g., fast / no reload of gtfsdb tables)...'\n )\n parser.add_argument(\n '-g',\n '--is_geospatial',\n action='store_true',\n default=False,\n help='Database IS GEOSPATIAL',\n )\n parser.add_argument(\n '-s',\n '--schema',\n default=None,\n help='Database SCHEMA name',\n )\n args = parser.parse_args()\n return args\n\n","sub_path":"ott/controller/util/argparse_db_loader.py","file_name":"argparse_db_loader.py","file_ext":"py","file_size_in_byte":1212,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"381099603","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n \nimport socket\nimport time\nimport router\nimport view\nimport index\nfrom optparse import OptionParser\n\n# Опции запуска\nopts = index.opts\n\n# Проверка файла БД\ndef init():\n try:\n open(opts.db_file, \"r\", encoding=opts.encode)\n except FileNotFoundError:\n import db\n try:\n with open(opts.base_dir + \"db.sql\", \"r\", encoding=opts.encode) as sql:\n for line in sql:\n db.execute(line)\n except:\n pass\n\n# Запуск сервера\ndef new_server():\n global opts\n\n # Экземпляр роутера\n r = router.RequestHandler()\n\n # Вызов функции настройки роутинга\n index.set_handlers(r)\n\n # Инициализация сокета\n try:\n serversocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n serversocket.bind((opts.server, opts.port))\n serversocket.listen(opts.listen)\n except socket.error:\n print('Не удалось создать сокет')\n exit()\n\n print('Ожидание запроса...')\n # Ожидание и обработка коннектов\n try:\n while True:\n conn, addr = serversocket.accept()\n print(\"Новый запрос от \" + addr[0])\n try:\n data = r.Route(conn)\n # Прочие ошибки\n except router.RouteException as err:\n html = view.HTMLObject(\"%s - %s\" % (err.code, err.message))\n html.Style(func=view.main_css, param=dict(filename=opts.css_file))\n html.Body(func=view.error_body, content=err.__dict__)\n r.do_post(conn, err.code, tp=\"text/html; charset=%s\" % opts.encode, data=str(html))\n # Ошибка 500\n except:\n err = router.RouteException(\"500\", \"Internal Server Error\")\n html = view.HTMLObject(\"%s - %s\" % (err.code, err.message))\n html.Style(func=view.main_css, param=dict(filename=opts.css_file))\n html.Body(func=view.error_body, content=err.__dict__)\n r.do_post(conn, err.code, tp=\"text/html; charset=%s\" % opts.encode, data=str(html))\n else:\n try:\n r.do_post(conn, tp=\"text/html; charset=%s\" % opts.encode, data=data)\n except:\n r.do_get(conn, tp=\"text/html; charset=%s\" % opts.encode)\n finally:\n conn.close()\n finally: \n serversocket.close()\n\ndef main():\n # Запуск сервера\n try:\n new_server()\n except ConnectionAbortedError as e:\n print(e)\n main()\n\nif __name__ == \"__main__\":\n try:\n init()\n main()\n except KeyboardInterrupt:\n print(\"Завершение работы...\")\n exit()","sub_path":"server.py","file_name":"server.py","file_ext":"py","file_size_in_byte":2924,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"596665602","text":"\"\"\"\nCyclic Support\n==============\n\"\"\"\n\nfrom ansys import dpf\nfrom ansys.dpf.core.scoping import Scoping\nfrom ansys.grpc.dpf import cyclic_support_pb2, cyclic_support_pb2_grpc\n\n\nclass CyclicSupport:\n \"\"\"Represents a cyclic support, which describes a model with cyclic symmetry.\n\n The model has the necessary data for cyclic (and multistage) expansion.\n\n Parameters\n ----------\n cyclic_support : ansys.grpc.dpf.cyclic_support_pb2.CyclicSupport message\n Cyclic support.\n server : DPFServer , optional\n Server with the channel connected to the remote or local instance. The default is\n ``None``, in which case an attempt is made to use the the global server.\n\n Examples\n --------\n Get a cyclic support from a model.\n\n >>> from ansys.dpf import core as dpf\n >>> from ansys.dpf.core import examples\n >>> multi_stage = examples.download_multi_stage_cyclic_result()\n >>> model = dpf.Model(multi_stage)\n >>> result_info = model.metadata.result_info\n >>> cyc_support = result_info.cyclic_support\n >>> cyc_support.num_sectors()\n 6\n >>> cyc_support.num_stages\n 2\n\n \"\"\"\n\n def __init__(self, cyclic_support, server=None):\n \"\"\"Initialize time frequency support with its `TimeFreqSupport` message (if possible).\"\"\"\n if server is None:\n server = dpf.core._global_server()\n\n self._server = server\n self._stub = self._connect()\n self._message = cyclic_support\n\n def __str__(self):\n \"\"\"Describe the entity.\n\n Returns\n -------\n str\n Description of the entity.\n \"\"\"\n from ansys.dpf.core.core import _description\n\n return _description(self._message, self._server)\n\n @property\n def num_stages(self) -> int:\n \"\"\"Number of cyclic stages in the model\n\n Examples\n --------\n >>> from ansys.dpf.core import Model\n >>> from ansys.dpf.core import examples\n >>> multi_stage = examples.download_multi_stage_cyclic_result()\n >>> cyc_support = Model(multi_stage).metadata.result_info.cyclic_support\n >>> cyc_support.num_stages\n 2\n\n Returns\n -------\n int\n Number of cyclic stages in the model.\n \"\"\"\n return self._stub.List(self._message).num_stages\n\n def num_sectors(self, stage_num=0) -> int:\n \"\"\"Number of sectors to expand on 360 degrees.\n\n Parameters\n ----------\n stage_num : int , optional\n Number of the stages required (from 0 to num_stages).\n\n Returns\n -------\n int\n Number of sectors to expand on 360 degrees.\n\n Examples\n --------\n >>> from ansys.dpf.core import Model\n >>> from ansys.dpf.core import examples\n >>> multi_stage = examples.download_multi_stage_cyclic_result()\n >>> cyc_support = Model(multi_stage).metadata.result_info.cyclic_support\n >>> cyc_support.num_sectors(0)\n 6\n >>> cyc_support.num_sectors(1)\n 12\n\n \"\"\"\n return self._stub.List(self._message).stage_infos[stage_num].num_sectors\n\n def base_nodes_scoping(self, stage_num=0) -> int:\n \"\"\"Retrieve a nodal scoping containing node IDs in the\n base sector of the given stage.\n\n Parameters\n ----------\n stage_num : int, optional\n Number of the stage required (from 0 to num_stages).\n\n Returns\n -------\n base_nodes_scoping : Scoping\n Nodes IDs in the base sector of the given stage.\n\n Examples\n --------\n >>> from ansys.dpf.core import Model\n >>> from ansys.dpf.core import examples\n >>> multi_stage = examples.download_multi_stage_cyclic_result()\n >>> cyc_support = Model(multi_stage).metadata.result_info.cyclic_support\n >>> base = cyc_support.base_nodes_scoping(0)\n\n \"\"\"\n return Scoping(\n scoping=self._stub.List(self._message)\n .stage_infos[stage_num]\n .base_nodes_scoping,\n server=self._server,\n )\n\n def base_elements_scoping(self, stage_num=0) -> int:\n \"\"\"Retrieve an elemental scoping containing elements IDs in the\n base sector of the given stage.\n\n Parameters\n ----------\n stage_num : int, optional\n Number of the stage required (from 0 to num_stages).\n\n Returns\n -------\n base_elements_scoping : Scoping\n Elements ids in the base sector of the given stage.\n\n Examples\n --------\n >>> from ansys.dpf.core import Model\n >>> from ansys.dpf.core import examples\n >>> multi_stage = examples.download_multi_stage_cyclic_result()\n >>> cyc_support = Model(multi_stage).metadata.result_info.cyclic_support\n >>> base = cyc_support.base_elements_scoping(stage_num=1)\n\n \"\"\"\n return Scoping(\n scoping=self._stub.List(self._message)\n .stage_infos[stage_num]\n .base_elements_scoping,\n server=self._server,\n )\n\n def sectors_set_for_expansion(self, stage_num=0) -> int:\n \"\"\"Retrieve a sector's scoping of the already expanded results\n and mesh or the list of sectors that will be expanded by default.\n\n A sector's scoping starts from 0, with the maximum equal to num_sectors-1.\n\n Parameters\n ----------\n stage_num : int, optional\n Number of the stage required (from 0 to num_stages).\n\n Returns\n -------\n sectors_set_for_expansion : Scoping\n List of sectors (starting from 0 to max = num_sectors-1).\n\n Examples\n --------\n >>> from ansys.dpf.core import Model\n >>> from ansys.dpf.core import examples\n >>> multi_stage = examples.download_multi_stage_cyclic_result()\n >>> cyc_support = Model(multi_stage).metadata.result_info.cyclic_support\n >>> print(cyc_support.sectors_set_for_expansion(stage_num=1).ids)\n [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11]\n\n \"\"\"\n return Scoping(\n scoping=self._stub.List(self._message)\n .stage_infos[stage_num]\n .sectors_for_expansion,\n server=self._server,\n )\n\n def expand_node_id(self, node_id, sectors=None, stage_num=0):\n \"\"\"Retrieve the node IDs corresponding to the base sector node ID given in the input\n after expansion.\n\n Parameters\n ----------\n node_id : int\n Base sector's node ID to expand.\n sectors : Scoping , list of int, optional\n List of sectors to expand (from 0 to ``num_sectors - 1``).\n The default is ``None``, in which case all sectors are expanded.\n stage_num : int, optional\n Number of the stage required (from 0 to ``num_stages``).\n\n Returns\n -------\n sectors_set_for_expansion : Scoping\n List of sectors (starting from 0 to ``num_sectors - 1``).\n\n Examples\n --------\n >>> from ansys.dpf.core import Model\n >>> from ansys.dpf.core import examples\n >>> multi_stage = examples.download_multi_stage_cyclic_result()\n >>> cyc_support = Model(multi_stage).metadata.result_info.cyclic_support\n >>> print(cyc_support.expand_node_id(1,stage_num=0).ids)\n [1, 3596, 5816, 8036, 10256, 12476]\n\n \"\"\"\n if isinstance(sectors, list):\n sectors = Scoping(ids=sectors, location=\"sectors\", server=self._server)\n\n request = cyclic_support_pb2.GetExpandedIdsRequest()\n request.support.CopyFrom(self._message)\n request.node_id = node_id\n request.stage_num = stage_num\n if sectors:\n request.sectors_to_expand.CopyFrom(sectors._message)\n return Scoping(\n scoping=self._stub.GetExpandedIds(request).expanded_ids, server=self._server\n )\n\n def expand_element_id(self, element_id, sectors=None, stage_num=0):\n \"\"\"Retrieves the element IDs corresponding to the base sector element ID given in the input\n after expansion.\n\n Parameters\n ----------\n element_id : int\n Base sector's element ID to expand.\n sectors : Scoping or list of int, optional\n List of sectors to expand (from 0 to ``num_sectors - 1``).\n The default is ``None``, in which case all sectors are expanded.\n stage_num : int, optional\n Number of the stage required (from 0 to ``num_stages``).\n\n Returns\n -------\n sectors_set_for_expansion : Scoping\n List of sectors (starting from 0 to ``num_sectors - 1``).\n\n Examples\n --------\n >>> from ansys.dpf.core import Model\n >>> from ansys.dpf.core import examples\n >>> multi_stage = examples.download_multi_stage_cyclic_result()\n >>> cyc_support = Model(multi_stage).metadata.result_info.cyclic_support\n >>> print(cyc_support.expand_element_id(1,stage_num=0).ids)\n [1, 1558, 2533, 3508, 4483, 5458]\n\n \"\"\"\n if isinstance(sectors, list):\n sectors = Scoping(ids=sectors, location=\"sectors\", server=self._server)\n\n request = cyclic_support_pb2.GetExpandedIdsRequest()\n request.support.CopyFrom(self._message)\n request.element_id = element_id\n request.stage_num = stage_num\n if sectors:\n request.sectors_to_expand.CopyFrom(sectors._message)\n return Scoping(\n scoping=self._stub.GetExpandedIds(request).expanded_ids, server=self._server\n )\n\n def _connect(self):\n \"\"\"Connect to the grpc service\"\"\"\n return cyclic_support_pb2_grpc.CyclicSupportServiceStub(self._server.channel)\n\n def __del__(self):\n try:\n self._stub.Delete(self._message)\n except:\n pass\n","sub_path":"ansys/dpf/core/cyclic_support.py","file_name":"cyclic_support.py","file_ext":"py","file_size_in_byte":9878,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"57769082","text":"import sys\nimport collections\n\nif __name__ == \"__main__\":\n # 读取第一行的n\n n = sys.stdin.readline().strip()\n n = int(n)\n\n def convertToDecimal(number, base):\n return int(str(number), base=base)\n\n def trans_map(cint):\n if cint < 0:\n print(\"不合法\")\n return\n elif cint < 10:\n return cint\n\n elif cint >= 10:\n return chr(cint - 10 + 65)\n\n def decimalConvertToBase(origin, n):\n # 10进制转换为任意进制的数\n list = []\n while True:\n # 取商\n s = origin // n\n # 取余数\n tmp = origin % n\n list.append(trans_map(tmp))\n if s == 0:\n break\n origin = s\n list.reverse()\n list = [str(each) for each in list]\n return ''.join(list)\n\n\n for _ in range(n):\n base = int(sys.stdin.readline().strip())\n line2 = sys.stdin.readline().strip().split()\n num1, num2, operation = line2\n\n\n ans = float('-inf')\n\n","sub_path":"others/2019Autumn/2tecentQ4.py","file_name":"2tecentQ4.py","file_ext":"py","file_size_in_byte":1054,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"158279363","text":"def merge(left, right):\n result = []\n i ,j = 0, 0\n while i < len(left) and j < len(right):\n if left[i] <= right[j]:\n result.append(left[i])\n i += 1\n else:\n result.append(right[j])\n j += 1\n result += left[i:]\n result += right[j:]\n return result\n\n\n\ndef mergeSort(alist):\n if len(alist)==1:\n return alist\n print(\"Splitting \",alist)\n mid = len(alist)//2\n lefthalf = alist[:mid]\n righthalf = alist[mid:]\n\n left = mergeSort(lefthalf)\n right = mergeSort(righthalf)\n return merge(left, right)\n\nlist_values=[2,6,3,4,7,9,3,4,6]\nprint(mergeSort(list_values))","sub_path":"Sorts/merge.py","file_name":"merge.py","file_ext":"py","file_size_in_byte":654,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"121182602","text":"#written by kenan arica. github.com/kenanarica if you have internships for me\n\nfrom prettytable import PrettyTable\nimport json\nimport requests\nimport os\nimport sys\nfile = open(\"nodeConfigs.json\", \"a\")\nnodes = []\njsonNodes = []\naccess_token = \"\"\nbikeWaggleConfig = [ #sensorname, sensorID, enabled/disabled, sensing freq(seconds)\n ['MetMAC', 0x00, False, 30], ['TMP112', 0x01, False, 30], ['HTU21D', 0x02, False, 30], ['HIH4030', 0x03, False, 30], \n ['BMP180', 0x04, False, 30], ['PR103J2', 0x05, False, 30], ['TSL250RDMS', 0x06, False, 30], ['MMA8452Q', 0x07, False, 30],\n ['SPV1840LR5H-B', 0x08, False, 30], ['TSYS01', 0x09, False, 30], ['HMC5883L', 0x0A, False, 30], ['HIH6130', 0x0B, False, 30],\n ['APDS_9006_020', 0x0C, False, 30], ['TSL260', 0x0D, False, 30], ['TSL250RDLS', 0x0E, False, 30], ['MLX75305', 0x0F, False, 30],\n ['ML8511', 0x10, False, 30], ['TMP421', 0x13, False, 30], ['Chemsense', 0x2A, False, 30], ['AlphaHisto', 0x28, False, 30]\n]\n\ndef getToken(args):\n tokenToReturn = args\n\n for node in jsonNodes:\n \n if args == node[\"nodeID\"]:\n print(\"This node's token was matched with token of node \" + node[\"nodeID\"])\n tokenToReturn = node[\"token\"]\n return tokenToReturn\n\n\ndef sendConfig(node, args, function): \n access_token = node[\"token\"]\n \n payload = {'params': args, 'access_token':access_token}\n requests.post(\"https://api.particle.io/v1/devices/{0}/{1}/\".format(node[\"deviceID\"], function), payload)\n \n\n\"\"\" TODO:\n\n* do -h for --configure\n* make disAll and enAll, dis and en multi-node function\n* do the removeNode function\n* make -e and -d disable/enable all nodes on given node, and make it multi-node if it isn't-d\n\n\n\n\"\"\"\n\n\n\ndef fixBool(arg):\n if arg.lower() == 'yes' or arg.lower() == 'y' or arg.lower() == 'true' or arg.lower() == 'en':\n return True\n if arg.lower() == 'no' or arg.lower() == 'n' or arg.lower() == 'false' or arg.lower == 'dis':\n return False\n else:\n rerun = input(\"[X] That's not a valid option, please type True, False, yes or no: \")\n return fixBool(rerun)\n\n# changeSensorConfig(newargs = ID, enabled, senseFreq)\n# changeNodeConfig()\n# changeSensorConfig(newArgs = \"001;en;30\")\n#print(fixBool(input(\"test\")))\n\n#seeing if the config file exists. if it doesn't, create it.\n\n#seeing if the nodeConfigs.json file exists. if it doesn't, create it.\ntry:\n \n file = open(\"nodeConfigs.json\", 'r')\nexcept IOError:\n\n file = open(\"nodeConfigs.json\", 'w')\n print(\"[WARNING] No previous nodes found, and no nodeConfigs.json file found. If it's your first time using the controller, don't worry. A file will be created.\")\n\n\n\n\n\n\n\"\"\"\nStuff to do: A lot!\n\nadd a function to change enabling on different sensors and changing the frequency of sensing\n\n\n\n\"\"\"\n#change this to a request called \"enableAllSensors\" in the particle cloud\ndef enableAll():\n for node in jsonNodes:\n sendConfig(node, \"enableall\", \"nodeConfig\")\n \n\n \n#change this to a request called \"enableAllSensors\" in the particle cloud\ndef disableAll():\n for node in jsonNodes:\n sendConfig(node, \"disableall\", \"nodeConfig\")\n \ndef createCustomConfig(): \n numOfSensors = int(input(\"How many sensors would you like to have on your node?\"))\n configuration = []\n for num in range(0, numOfSensors):\n tempConfig = []\n tempConfig.append(input(\"Sensor name? \"))\n tempConfig.append(input(\"Sensor ID? \"))\n tempConfig.append(fixBool(input(\"Enabled? [True/False]\")))\n tempConfig.append(input(\"How often, in seconds, should the sensor collect data? \"))\n configuration.append(tempConfig)\n print(tempConfig)\n print(configuration)\n return configuration\n \n\n\n\ndef loadNodes():\n global jsonNodes, nodes\n jsonNodes = []\n nodes = []\n \n configFile = open(\"nodeConfigs.json\", \"r\")\n for line in configFile:\n jsonData = json.loads(line)\n print(\"[✓] Node with ID \" + jsonData[\"nodeID\"] + \" loaded\")\n jsonNodes.append(jsonData)\n #nodes.append(tempNodeObject)\n #print(jsonNodes)\n #print(nodes)\n\ndef configure(unparsedargs):\n #make sure nodes are loaded, we're going to overwrite the configs file every time this function is called.\n #print(unparsedargs)\n nodeToConfigure = unparsedargs[0] #name of the node that the user wants to configure\n multipleNodes = False\n \n if ',' in nodeToConfigure:\n nodeToConfigure = nodeToConfigure.split(\",\")\n multipleNodes = True\n\n if '-rn' in unparsedargs:\n newName = unparsedargs[unparsedargs.index(\"-rn\") + 1]\n \n if not multipleNodes:\n \n for node in jsonNodes:\n if node[\"nodeID\"] == nodeToConfigure:\n print(\"[✓] Node previously named {0} is now named {1}.\".format(node[\"name\"], newName))\n node[\"name\"] = newName\n else:\n for tempNode in nodeToConfigure:\n for node in jsonNodes:\n if node[\"nodeID\"] == tempNode:\n print(\"[✓] Node previously named {0} is now named {1}.\".format(node[\"name\"], newName))\n node[\"name\"] = newName\n \n \n\n if '-id' in unparsedargs:\n newID = unparsedargs[unparsedargs.index(\"-id\") + 1]\n \n if not multipleNodes:\n \n for node in jsonNodes:\n if node[\"nodeID\"] == nodeToConfigure:\n print(\"[✓] Node previously with the ID {0} now has the ID {1}.\".format(node[\"nodeID\"], newID))\n node[\"nodeID\"] = newID\n \n if '-d' in unparsedargs:\n newDisabledSetting = unparsedargs[unparsedargs.index(\"-d\") + 1]\n \n if not multipleNodes:\n \n for node in jsonNodes:\n if node[\"nodeID\"] == nodeToConfigure:\n print(\"[✓] Node with the ID \" + node[\"nodeID\"] + \" is now disabled.\" )\n node[\"enabled\"] = False\n sendConfig(node, \"disableall\", \"nodeConfig\")\n \n else:\n newDisabledSetting = unparsedargs[unparsedargs.index(\"-d\") + 1]\n print(nodeToConfigure)\n for nodeLabelToModify in nodeToConfigure:\n #please help me what is going on here\n for node in jsonNodes:\n \n if node[\"name\"] == nodeLabelToModify:\n \n print(\"[✓] Node with the ID \" + node[\"nodeID\"] + \" is now disabled.\" )\n node[\"enabled\"] = False\n\n sendConfig(node, \"disableall\", \"nodeConfig\")\n \n\n\n if '-e' in unparsedargs:\n newEnabledSetting = unparsedargs[unparsedargs.index(\"-e\") + 1]\n \n if not multipleNodes:\n \n for node in jsonNodes:\n if node[\"nodeID\"] == nodeToConfigure:\n print(\"[✓] Node with the ID \" + node[\"nodeID\"] + \" is now enabled.\" )\n node[\"enabled\"] = True\n sendConfig(node, \"enableall\", \"nodeConfig\")\n \n else:\n newEnabledSetting = unparsedargs[unparsedargs.index(\"-e\") + 1]\n print(nodeToConfigure)\n for nodeLabelToModify in nodeToConfigure:\n #please help me what is going on here\n for node in jsonNodes:\n \n if node[\"nodeID\"] == nodeLabelToModify:\n print(\"[✓] Node with the ID \" + node[\"nodeID\"] + \" is now enabled.\" )\n node[\"enabled\"] = True\n sendConfig(node, \"enableall\", \"nodeConfig\")\n \n \n\n\n if '-sdf' in unparsedargs:\n newSDFSetting = unparsedargs[unparsedargs.index(\"-sdf\") + 1]\n \n if not multipleNodes:\n \n for node in jsonNodes:\n if node[\"nodeID\"] == nodeToConfigure:\n print(\"[✓] Node previously with the sending frequency setting {0} now has the setting {1} seconds.\".format(node[\"sendingFrequency\"], newSDFSetting))\n node[\"sendingFrequency\"] = newSDFSetting\n params = \"freqreport-\" + newSDFSetting\n sendConfig(node, params, \"nodeConfig\")\n\n else:\n newSDFSetting = unparsedargs[unparsedargs.index(\"-sdf\") + 1]\n print(nodeToConfigure)\n for nodeLabelToModify in nodeToConfigure:\n #please help me what is going on here\n for node in jsonNodes:\n \n if node[\"nodeID\"] == nodeLabelToModify:\n \n print(\"[✓] Node previously with the sending frequency setting {0} now has the setting {1} seconds.\".format(node[\"sendingFrequency\"], newSDFSetting))\n node[\"sendingFrequency\"] = newSDFSetting\n params = \"freqreport-\" + newSDFSetting\n sendConfig(node, params, \"nodeConfig\")\n \n\n\n\n if '-ssf' in unparsedargs:\n newSSFSetting = unparsedargs[unparsedargs.index(\"-ssf\") + 1]\n if not multipleNodes:\n \n for node in jsonNodes:\n if node[\"nodeID\"] == nodeToConfigure:\n print(\"[✓] Node previously with the sensing frequency setting {0} now has the setting {1} seconds.\".format(node[\"sensingFrequency\"], newSSFSetting))\n node[\"sensingFrequency\"] = newSSFSetting\n\n for sensor in node[\"config\"]:\n sensor[3] = newSSFSetting\n print(\"Setting sensor with ID {0} to a sensing frequency of {1} seconds\".format(sensor[3], newSSFSetting))\n params = \"{0};_;{1}\".format(sensor[1], newSSFSetting)\n sendConfig(node, params, \"sensorConfig\")\n\n\n else:\n newSSFSetting = unparsedargs[unparsedargs.index(\"-ssf\") + 1]\n print(nodeToConfigure)\n for nodeLabelToModify in nodeToConfigure:\n #please help me what is going on here\n for node in jsonNodes:\n \n if node[\"nodeID\"] == nodeLabelToModify:\n print(\"[✓] Node previously with the sensing frequency setting {0} now has the setting {1} seconds.\".format(node[\"sensingFrequency\"], newSSFSetting))\n node[\"sensingFrequency\"] = newSSFSetting\n \n\n for sensor in node[\"config\"]:\n sensor[3] = newSSFSetting\n print(\"Setting sensor with ID {0} to a sensing frequency of {1} seconds\".format(sensor[3], newSSFSetting))\n params = \"{0};_;{1}\".format(sensor[1], newSSFSetting)\n sendConfig(node, params, \"sensorConfig\")\n\n \n if '-sm' in unparsedargs:\n newStatusSetting = fixBool(unparsedargs[unparsedargs.index(\"-sm\") + 1])\n command = \"\"\n\n if newStatusSetting:\n command = \"Status\" #STATUS MESSAGE NOT IMPLEMENTED\n else: \n command = \"notStatus\"\n\n if not multipleNodes:\n \n for node in jsonNodes:\n if node[\"nodeID\"] == nodeToConfigure:\n print(\"[✓] Node previously with the status message setting {0} now has the setting {1}.\".format(node[\"statusMessage\"], newStatusSetting))\n node[\"statusMessage\"] = newStatusSetting\n else:\n newStatusSetting = unparsedargs[unparsedargs.index(\"-sm\") + 1]\n print(nodeToConfigure)\n for nodeLabelToModify in nodeToConfigure:\n #please help me what is going on here\n for node in jsonNodes:\n \n if node[\"nodeID\"] == nodeLabelToModify:\n print(\"[✓] Node previously with the status message setting {0} now has the setting {1}.\".format(node[\"statusMessage\"], newStatusSetting))\n node[\"statusMessage\"] = newStatusSetting\n \n #params = command\n #payload = {'params': params, 'access_token': access_token}\n #requests.post(\"https://api.particle.io/v1/devices/{0}/{1}/\".format(node[\"deviceID\"], \"nodeConfig\"), payload)\n\n \n\n\n if '-sd' in unparsedargs:\n newSDSetting = fixBool(unparsedargs[unparsedargs.index(\"-sd\") + 1])\n command = \"\"\n\n if newSDSetting:\n command = \"enableSD\"\n else:\n command = \"disableSD\"\n\n if not multipleNodes:\n \n for node in jsonNodes:\n if node[\"nodeID\"] == nodeToConfigure:\n print(\"[✓] Node previously with the save to SD setting {0} now has the setting {1}.\".format(node[\"saveToSD\"], newSDSetting))\n node[\"saveToSD\"] = newSDSetting\n params = command\n sendConfig(node, params, \"nodeConfig\")\n else:\n print(nodeToConfigure)\n for nodeLabelToModify in nodeToConfigure:\n\n for node in jsonNodes:\n \n if node[\"nodeID\"] == nodeLabelToModify:\n print(\"[✓] Node previously with the save to SD setting {0} now has the setting {1}.\".format(node[\"saveToSD\"], newSDSetting))\n node[\"saveToSD\"] = newSDSetting\n print(\"Sending command...\")\n params = command\n sendConfig(node, params, \"nodeConfig\")\n\n\n #sen functions:\n # add/remove sensor\n # redo entire sensor config\n # change individual sensor configs \n\n###########SEN BLOCK\n\n\n if '-sen' in unparsedargs: \n function = unparsedargs[unparsedargs.index(\"-sen\") + 1]\n \n ###adding or removing a sensor. --configure [nodeName(s)] -sen [sensorID(s)] [add/remove/config] [-e [T/F], -id [newID], -ssf [seconds] ] ]\n\n # Ex usage: --configure [nodeID] -sen add\n\n if function == 'add':\n\n tempConfig = []\n tempConfig.append(input(\"Sensor name? \"))\n tempConfig.append(input(\"Sensor ID? \"))\n tempConfig.append(fixBool(input(\"Enabled? [True/False]\")))\n tempConfig.append(input(\"How often, in seconds, should the sensor collect data? \"))\n\n \n if not multipleNodes:\n\n for node in jsonNodes:\n \n if node[\"nodeID\"] == nodeToConfigure:\n node[\"config\"].append(tempConfig)\n print(\"[✓] Sensor with ID {0} has been added to Node {1}\".format(tempConfig[1], node[\"nodeID\"]))\n else:\n for nodeLabelToModify in nodeToConfigure:\n\n for node in jsonNodes:\n\n if node[\"nodeID\"] == nodeLabelToModify: \n node[\"config\"].append(tempConfig)\n print(\"[✓] Sensor with ID {0} has been added to Node {1}\".format(tempConfig[1], node[\"nodeID\"]))\n\n # Ex. Usage: --configure [nodeID] -sen rm [sensorID] \n\n elif function == 'rm':\n if multipleNodes == False:\n \n sensorToRemove = unparsedargs[unparsedargs.index(\"rm\") + 1]\n if ',' in sensorToRemove:\n sensorToRemove = sensorToRemove.split(\",\")\n\n for node in jsonNodes:\n if node[\"nodeID\"] == nodeToConfigure:\n for individualSensor in sensorToRemove:\n for sensor in node[\"config\"]:\n if sensor[1] == individualSensor:\n print(\"[✓] removing sensor with ID \" + individualSensor + \" from node with ID \" + node[\"nodeID\"])\n node[\"config\"].remove(sensor)\n\n else:\n \n for node in jsonNodes:\n if node[\"nodeID\"] == nodeToConfigure:\n\n for sensor in node[\"config\"]:\n if sensor[1] == sensorToRemove:\n print(\"[✓] removing sensor with ID \" + sensorToRemove)\n node[\"config\"].remove(sensor)\n else: \n \n for individualNode in nodeToConfigure:\n\n sensorToRemove = unparsedargs[unparsedargs.index(\"rm\") + 1]\n if ',' in sensorToRemove:\n sensorToRemove = sensorToRemove.split(\",\")\n\n for node in jsonNodes:\n if node[\"nodeID\"] == individualNode:\n for individualSensor in sensorToRemove:\n for sensor in node[\"config\"]:\n if sensor[1] == individualSensor:\n print(\"[✓] removing sensor with ID \" + individualSensor + \" from node with ID \" + node[\"nodeID\"])\n node[\"config\"].remove(sensor)\n\n else:\n \n for node in jsonNodes:\n if node[\"nodeID\"] == individualNode:\n\n for sensor in node[\"config\"]:\n if sensor[1] == sensorToRemove:\n print(\"[✓] removing sensor with ID \" + sensorToRemove + \" from node with ID \" + node[\"nodeID\"])\n node[\"config\"].remove(sensor)\n \n # ex. Usage: --configure [nodeID] -sen dis [sensorID]\n \n elif function.lower() == 'dis':\n\n sensorToDisable = unparsedargs[unparsedargs.index(\"dis\") + 1]\n\n if not multipleNodes:\n \n if ',' in sensorToDisable:\n sensorToDisable = sensorToDisable.split(',')\n\n for individualSensor in sensorToDisable:\n \n\n for node in jsonNodes:\n if node[\"nodeID\"] == nodeToConfigure:\n\n for sensor in node[\"config\"]:\n if sensor[1] == individualSensor:\n sensor[2] = False\n print(\"Setting sensor with ID \" + individualSensor + \" to disabled.\")\n \n params = \"{0};dis;_\".format(sensor[1])\n sendConfig(node, params, \"sensorConfig\")\n else:\n\n for node in jsonNodes:\n if node[\"nodeID\"] == nodeToConfigure:\n\n for sensor in node[\"config\"]:\n if sensor[1] == sensorToDisable:\n sensor[2] = False\n print(\"Setting sensor with ID \" + sensorToDisable + \" to disabled.\")\n \n params = \"{0};dis;_\".format(sensor[1])\n sendConfig(node, params, \"sensorConfig\")\n\n # ex. Usage: --configure [nodeID] -sen en [sensorID]\n\n\n elif function.lower() == 'en':\n\n sensorToEnable = unparsedargs[unparsedargs.index(\"en\") + 1]\n\n if not multipleNodes:\n \n if ',' in sensorToEnable:\n sensorToEnable = sensorToEnable.split(',')\n\n for individualSensor in sensorToEnable:\n \n\n for node in jsonNodes:\n if node[\"nodeID\"] == nodeToConfigure:\n\n for sensor in node[\"config\"]:\n if sensor[1] == individualSensor:\n sensor[2] = True\n print(\"Setting sensor with ID \" + individualSensor + \" to Enabled.\")\n \n params = \"{0};en;_\".format(sensor[1])\n sendConfig(node, params, \"sensorConfig\")\n else:\n\n for node in jsonNodes:\n if node[\"nodeID\"] == nodeToConfigure:\n\n for sensor in node[\"config\"]:\n if sensor[1] == sensorToEnable:\n sensor[2] = True\n print(\"Setting sensor with ID \" + sensorToEnable + \" to Enabled.\")\n\n params = \"{0};en;_\".format(sensor[1])\n sendConfig(node, params, \"sensorConfig\")\n\n # ex. Usage: --configure [nodeID] -sen freq [newSetting] [sensorID]\n\n elif function.lower() == 'freq':\n\n sensorToFreq = unparsedargs[unparsedargs.index(\"freq\") + 2]\n newFreqSetting = unparsedargs[unparsedargs.index(\"freq\") + 1]\n\n if not multipleNodes:\n \n if ',' in sensorToFreq:\n sensorToFreq = sensorToFreq.split(',')\n\n for individualSensor in sensorToFreq:\n \n\n for node in jsonNodes:\n if node[\"nodeID\"] == nodeToConfigure:\n\n for sensor in node[\"config\"]:\n if sensor[1] == individualSensor:\n sensor[2] = True\n print(\"Setting sensor with ID \" + individualSensor + \" to a sensing freq of \" + newFreqSetting + \" seconds.\")\n params = \"{0};_;{1}\".format(sensor[1], newFreqSetting)\n sendConfig(node, params, \"sensorConfig\") \n \n else:\n\n for node in jsonNodes:\n if node[\"nodeID\"] == nodeToConfigure:\n\n for sensor in node[\"config\"]:\n if sensor[1] == sensorToFreq:\n sensor[2] = True\n print(\"Setting sensor with ID \" + individualSensor + \" to a sensing freq of \" + newFreqSetting + \" seconds.\")\n\n params = \"{0};_;{1}\".format(sensor[1], newFreqSetting)\n sendConfig(node, params, \"sensorConfig\")\n\n with open('nodeConfigs.json', 'w') as outfile:\n for node in jsonNodes:\n json.dump(node, outfile)\n outfile.write(\"\\n\")\n #at the end we're going to save this to our objects and overwrite the nodeConfigs.json file.\n\n\ndef listNodes(args):\n\n # --list -sen [nodeID]\n\n\n if '-sen' in args:\n t = PrettyTable(['Sensor name', 'Sensor ID', 'Enabled/Disabled', 'Reporting Frequency'])\n nodeToSelect = args[args.index(\"-sen\") + 1]\n for node in jsonNodes:\n if node[\"nodeID\"] == nodeToSelect:\n for sensor in node[\"config\"]:\n t.add_row([sensor[0], sensor[1], sensor[2], sensor[3]])\n else:\n \n t = PrettyTable(['Name', 'nodeID', 'deviceID', 'enabled', 'Sending Freq', 'Sensing Freq', 'Status Message', 'Save to SD', 'Access token'])\n for node in jsonNodes:\n \n t.add_row([node[\"name\"], node[\"nodeID\"], node[\"deviceID\"], node[\"enabled\"], node[\"sendingFrequency\"], node[\"sensingFrequency\"], node[\"statusMessage\"], node[\"saveToSD\"], node[\"token\"]])\n print(t)\n\n\n\ndef addNode():\n nodeName = input(\"What would you like to name your node? \")\n nodeID = input(\"What is the node ID? \")\n deviceID = input(\"What is the deviceID? \")\n sendingFrequency = input(\"What would you like the SENDING frequency to be? \")\n sensingFrequency = input(\"What would you like the SENSING frequency to be? \")\n statusMessage = input(\"Would you like the node to send a status message? [True/False] \")\n saveToSD = input(\"Would you like the node to save to SD during connection loss? [True/False] \")\n defaultConfigYN = input(\"Use default config for bike waggle? [True/False]\")\n tempToken = getToken(input(\"What is this node's access token? If you want to use the same token as an existing node, just input the ID.\"))\n \n sensorConfig = []\n if defaultConfigYN.lower().startswith(\"t\"):\n sensorConfig = bikeWaggleConfig\n elif defaultConfigYN.lower().startswith(\"f\"):\n sensorConfig = createCustomConfig()\n\n\n jsonToAppend = {\n \"name\" : nodeName, # -rn \n \"nodeID\" : nodeID, # -id\n \"deviceID\" : deviceID, # -d\n \"enabled\" : False, # -e\n \"sendingFrequency\" : sendingFrequency, # -sdf\n \"sensingFrequency\" : sensingFrequency, # -ssf\n \"statusMessage\" : statusMessage, # -sm\n \"saveToSD\" : saveToSD, # - sd\n \"config\" : sensorConfig, # - sen\n \"token\" : tempToken\n }\n \n jsonNodes.append(jsonToAppend)\n\n \n with open('nodeConfigs.json', 'a') as outfile:\n json.dump(jsonToAppend, outfile)\n outfile.write(\"\\n\")\n print(jsonNodes)\n #configString = \"{{ \\\"name\\\" : \\\"{}\\\", }}\"\n\nloadNodes()\n#list of node names or node ID's. compare this to a node-lookup table if it's in name form.\n\n#import OS package and get args\nargs = sys.argv[1:] #get args here, removing the first \"controller.py entry\"\n#first, check if args are equal to non-function commands such as list nodes, help, disableAll, enableAll, addNode, removeNode\nif len(args) > 0:\n \n if args[0] == \"--help\" or args[0] == \"help\":\n\n print(\"\\nThis is a tool to configure your micro-waggle modules! Here's a list of commands: \\n\\n --help : gives you all the help you need! \\n --list : lists all nodes and their configurations \\n --enAll : enables all nodes! By default, all nodes are off out of the box. \\n --disAll : disables all nodes \\n --add: adds a new node, the parameters come after you type it! \\n --rm : lists your nodes and allows you to remove one \\n \")\n\n elif args[0] == \"--list\":\n\n listNodes(args)\n\n elif args[0] == \"--enAll\":\n\n enableAll()\n\n elif args[0] == \"--disAll\":\n\n disableAll(args[1:])\n\n elif args[0] == \"--add\":\n\n addNode()\n\n elif args[0] == \"--rm\":\n\n removeNode()\n\n elif args[0] == \"--configure\":\n\n configure(args[1:])\n\n # --config [node name] -sensorID []\nelse: \n print(\"\\nThis is a tool to configure your micro-waggle modules! Here's a list of commands: \\n\\n --help : gives you all the help you need! \\n --list : lists all nodes and their configurations \\n --enAll : enables all nodes! By default, all nodes are off out of the box. \\n --disAll : disables all nodes \\n --add: adds a new node, the parameters come after you type it! \\n --rm : lists your nodes and allows you to remove one \\n \")\n\n# I have to fix the --help interface. I'll get to it one day.\n\n","sub_path":"integrated/software/devicecontroller/controller.py","file_name":"controller.py","file_ext":"py","file_size_in_byte":27274,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"110658006","text":"import gym\nimport tensorflow as tf\nfrom Agent import Agent\n\n# Hyper parameters\nENV_NAME = \"CartPole-v0\"\nDISPLAY_EPISODE = 50\nMAX_STEP = 1000\nHIDDEN_UNITS = 20\n\n# Initial OpenAI Gym env and DQN agent\nenv = gym.make(ENV_NAME)\nagent = Agent(env)\n# Load model\nsaver = tf.train.Saver()\nsaver.restore(agent.sess, './model/CartPole.ckpt')\n\n# Display\nfor i in xrange(DISPLAY_EPISODE):\n state = env.reset()\n for j in xrange(MAX_STEP):\n env.render()\n action = agent.action(state, policy='greedy')\n state, reward, done, _ = env.step(action)\n if done:\n break\n\n print(\"Episode: %02d\" % (i))\n","sub_path":"result.py","file_name":"result.py","file_ext":"py","file_size_in_byte":626,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"89656126","text":"#! /usr/bin/env python\n# -*- coding: utf-8\n# @author \nimport logging\nfrom sys import argv, exit\nfrom pronto import Pronto, ProntoHttp404\n\nlogger = logging.getLogger('pronto_logger')\nlogger.setLevel(logging.CRITICAL)\np = Pronto(strict=True)\ntry:\n pr = p.problem_report(argv[1])\nexcept ProntoHttp404:\n print('NOT FOUND')\n exit(1)\nelse:\n print(pr.xml())\n","sub_path":"server/pronto/getpr.py","file_name":"getpr.py","file_ext":"py","file_size_in_byte":388,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"489298847","text":"broker_ip=\"34.123.208.229\"\nbroker_port=1883\ndb_name=\"test_database\"\npath_data_heating=\"temperature_in\"\ntopic_grzalka=\"grzalka_test\"\ntopic_grzalka2=\"grzalka_test2\"\ndic={\"on\":\"1\",\"off\":\"0\"}\npath_data_temperature=\"temperature_in\"\npath_data_wiatr_sila=\"./../../Data/WindS.csv\"\npath_data_wiatr_kierunek=\"./../../Data/WindD.csv\"\npath_data_entrance=\"./../../Data/Entrance.csv\"\ntopic ={ \"harmonogram_new\":\"harmonogram_new\",\n \"light_salon\":\"light_salon\",\n \"heating_switch\":\"heating_switch\",\n \"grzalka\":\"grzalka\",\n \"temperatura\":\"temperatura\"}\n\ncollections={\n \"temperature_in\":\"symulated_temp\"\n \n}","sub_path":"src/Configs/config_test_termostat_con.py","file_name":"config_test_termostat_con.py","file_ext":"py","file_size_in_byte":618,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"525163051","text":"import numpy as np\nfrom PIL import Image\nfrom tqdm import tqdm\n\n\ndef chip_image(img, chip_size=(300, 300)):\n \"\"\"\n Segment an image into NxWxH chips\n\n Args:\n img : Array of image to be chipped\n chip_size : A list of (width,height) dimensions for chips\n\n Outputs:\n An ndarray of shape (N,W,H,3) where N is the number of chips,\n W is the width per chip, and H is the height per chip.\n\n \"\"\"\n width, height, _ = img.shape\n wn, hn = chip_size\n images = np.zeros((int(width / wn) * int(height / hn), wn, hn, 3))\n k = 0\n for i in tqdm(range(int(width / wn))):\n for j in range(int(height / hn)):\n chip = img[wn * i:wn * (i + 1), hn * j:hn * (j + 1), :3]\n images[k] = chip\n\n k = k + 1\n\n return images.astype(np.uint8)\n\n\nif __name__ == \"__main__\":\n arr = np.array(Image.open(\"./Images/CF013540.jpg\"))\n chip_size = (300, 300)\n img = chip_image(arr, chip_size)\n print(img.shape)\n\n chipresult = \"./chipresult/\"\n for index in range(img.shape[0]):\n a = img[index]\n r = Image.fromarray(a[:, :, 0]).convert('L')\n g = Image.fromarray(a[:, :, 1]).convert('L')\n b = Image.fromarray(a[:, :, 2]).convert('L')\n image = Image.merge(\"RGB\", (r, g, b))\n image.save(chipresult + str(index) + \".jpg\", 'jpg')\n\n\n","sub_path":"src/semantic-segmentation/test/chip.py","file_name":"chip.py","file_ext":"py","file_size_in_byte":1345,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"557390362","text":"from django.shortcuts import render, redirect\nfrom django.core.paginator import Paginator, EmptyPage, PageNotAnInteger\nfrom multiprocessing import Process,Manager\nfrom operator import itemgetter\n\nimport pickle\n\n# homepage\ndef index(request):\n\ttry:\n\t\ttopics = pickle.load(open('topics.pickle', 'rb'))\n\texcept (IOError, OSError) as e:\n\t\ttopics = []\n\n\t# display top 20 with descending upvotes\n\t# use django's pagination to display the rest of the topics\n\tsorted_topics = sorted(topics, key=itemgetter('upvote'), reverse=True)\n\tpaginator = Paginator(sorted_topics, 20)\n\tpage = request.GET.get('page')\n\n\ttry:\n\t\tpg_topic = paginator.page(page)\n\texcept PageNotAnInteger:\n\t\t# If page is not an integer, deliver first page.\n\t\tpg_topic = paginator.page(1)\n\texcept EmptyPage:\n\t\t# If page is out of range (e.g. 9999), deliver last page of results.\n\t\tpg_topic = paginator.page(paginator.num_pages)\n\n\t# only dict is allowed to be passed to the template\n\treturn render(request, 'thread/index.html', {'topics' : pg_topic})\n\n# method for posting topic\ndef post(request):\n\n\ttry:\n\t\ttopics = pickle.load(open('topics.pickle', 'rb'))\n\texcept (IOError, OSError) as e:\n\t\ttopics = []\n\n\t# structure of topics consists of a list of dictionaries\n\t# each dict store the details of each topic, such as upvote, downvote, content, id\n\tif request.method == 'POST':\n\t\ttopic = {}\n\t\ttopic['id'] = len(topics)\n\t\ttopic['content'] = request.POST['post']\n\t\ttopic['upvote'] = 0\n\t\ttopic['downvote'] = 0\n\t\ttopics.append(topic)\n\n\t\t# save changes as pickled python list\n\t\tpickle.dump(topics, open('topics.pickle', 'wb'))\n\n\t# redirect back to homepage\n\treturn redirect('/thread/')\n\n# method for voting\ndef vote(request):\n\t# check if topics exist\n\ttry:\n\t\ttopics = pickle.load(open('topics.pickle', 'rb'))\n\texcept (IOError, OSError) as e:\n\t\t# just redirect to homepage if thread does not exist\n\t\treturn redirect('/thread/')\n\n\tif request.method == 'POST':\n\t\t# update topic based on id\n\t\ttopic_id = int(request.POST['id'])\n\n\t\t# distinguish between upvoting and downvoting\n\t\tif 'upvote' in request.POST:\n\t\t\ttopics[topic_id]['upvote'] += 1\n\t\telif 'downvote' in request.POST:\n\t\t\ttopics[topic_id]['downvote'] += 1\n\n\t\t# save changes as pickled python list\n\t\tpickle.dump(topics, open('topics.pickle', 'wb'))\n\n\t# redirect back to homepage\n\treturn redirect('/thread/')\n\n# not required for main functionalities\n# implemented for convenience to remove all the posts\ndef clear(request):\n\tpickle.dump([], open('topics.pickle', 'wb'))\n\treturn redirect('/thread/')","sub_path":"thread/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":2501,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"536347279","text":"from django.shortcuts import render, HttpResponse\nfrom apps.hello.models import Contact, HttpRequestLog\nfrom apps.hello.forms import ContactForm\nfrom django.contrib.auth.decorators import login_required\nimport json\n\nimport datetime\n\nlogin_url = '/login/'\n\n\ndef hello(request):\n contact = Contact.objects.all()[0]\n age = int((datetime.date.today() - contact.date_of_birth).days / 365.25)\n return render(request, 'hello/index.html',\n {'contact': contact, 'age': age})\n\n\ndef http_requests(request):\n requests = HttpRequestLog.objects.all().order_by('-date')[:10]\n request.session['viewed_nmb'] = HttpRequestLog.objects.count()\n return render(request, 'hello/requests.html',\n {'requests': requests})\n\n\ndef ajax_request(request):\n response_data = {'total': HttpRequestLog.objects.count()}\n if 'viewed_nmb' in request.session:\n response_data['total'] -= request.session['viewed_nmb']\n return HttpResponse(json.dumps(response_data),\n content_type='application/json')\n\n\n@login_required(login_url=login_url)\ndef edit_form(request):\n current_entry = Contact.objects.all()[0]\n if request.method == 'POST' and request.is_ajax():\n form = ContactForm(request.POST, request.FILES, instance=current_entry)\n if form.is_valid():\n form.save()\n return HttpResponse(json.dumps({'success': 'success'}),\n content_type='application/json')\n else:\n return HttpResponse(json.dumps({'error': form.errors}))\n else:\n form = ContactForm(instance=current_entry)\n return render(request, 'hello/edit_form.html',\n {'form': form, 'entry': current_entry})\n","sub_path":"apps/hello/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1731,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"93343017","text":"from random import randint\n\ndef main():\n \"\"\"Guess the number!!\"\"\"\n\n #The correct answer cannot change every loop so it has to stay outside the loop\n correctNumber = randint(0, 100)\n guessNumber = input(\"What's your guess from 0 to 100, man?\\n\")\n \n #Gotta make sure we're giving an int alright\n try:\n GuessVal = int(guessNumber)\n except ValueError:\n print(\"That's not an int\")\n\n while(1):\n\n #So what'cha say? The main game loop\n if(GuessVal == correctNumber):\n print(\"Yay! , how'd you do that?\\n\") \n break\n \n elif(GuessVal > correctNumber):\n print(\"Not quite there, you wanna try something lesser?\\n\")\n GuessVal = int(input(\"What's your guess from 0 to 100, man?\\n\"))\n continue\n \n else:\n print(\"Try something bigger\\n\")\n GuessVal = int(input(\"What's your guess from 0 to 100, man?\\n\"))\n continue\n \nif __name__ == '__main__':\n main()","sub_path":"Knight Lab Projects/2.Guess Number.py","file_name":"2.Guess Number.py","file_ext":"py","file_size_in_byte":1009,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"244493853","text":"import subprocess\nimport key_store\nimport pdb\nimport os\n\n\ndef perform_encryption(messages: list, keystore_data, is_text_encryption: bool):\n\n messages = messages if isinstance(messages, list) else [messages]\n key = key_store.get_password(\n keystore_data.path, keystore_data.password, keystore_data.key_identifier\n )\n if is_text_encryption:\n return encrypt_text(messages, keystore_data.encryption_mode, key)\n else:\n return encrypt_files(messages, keystore_data.encryption_mode, key)\n\n\n# TODO: add -iv from devrandom\n# NEVER USE URANDOM\ndef encrypt_text(plaintexts: list, encryption_mode: str, key: str):\n\n ciphertexts = []\n\n for message in plaintexts:\n\n open_ssl_commands = [\"openssl\", \"enc\", f\"-{encryption_mode}\", \"-nosalt\"]\n\n if \"cbc\" in encryption_mode:\n with open(\"/dev/random\", \"rb\") as f:\n iv = f.read(16).hex()\n open_ssl_commands += [\"-K\", f\"{key}\", \"-iv\", f\"{iv}\"]\n else:\n open_ssl_commands += [\"-k\", f\"{key}\"]\n\n ciphertext = subprocess.check_output(open_ssl_commands, input=message)\n ciphertexts.append(ciphertext)\n\n return ciphertexts\n\n\ndef encrypt_files(\n file_paths: list, encryption_mode: str, key: str, challenge: bool = False\n):\n encrypted_paths = []\n\n if challenge:\n for path in file_paths:\n open_ssl_commands = [\n \"openssl\",\n \"enc\",\n f\"-{encryption_mode}\",\n \"-nosalt\",\n \"-k\",\n f\"{key}\",\n \"-iv\",\n f\"{os.urandom(16).hex()}\",\n \"-in\",\n f\"{path}\",\n \"-out\",\n f\"challenge.enc\",\n ]\n subprocess.check_output(open_ssl_commands)\n\n else:\n for path in file_paths:\n\n enc_path = f\"{path}.enc\"\n encrypted_paths.append(enc_path)\n\n open_ssl_commands = [\n \"openssl\",\n \"enc\",\n f\"-{encryption_mode}\",\n \"-nosalt\",\n \"-k\",\n f\"{key}\",\n \"-iv\",\n f\"{os.urandom(16).hex()}\",\n \"-in\",\n f\"{path}\",\n \"-out\",\n enc_path,\n ]\n subprocess.check_output(open_ssl_commands)\n return encrypted_paths\n\n\n##### DECRYPTION #####\ndef perform_decryption(\n ciphertexts: list, keystore_data, encryption_mode: str, is_text_decryption: bool\n):\n\n ciphertexts = ciphertexts if isinstance(ciphertexts, list) else [ciphertexts]\n\n key = key_store.get_password(\n keystore_data.path, keystore_data.password, keystore_data.key_identifier\n )\n\n if is_text_decryption:\n encrypt_text(ciphertexts, encryption_mode, key)\n else:\n decrypt_files(ciphertexts, encryption_mode, key)\n\n\ndef decrypt_text(ciphertexts: list, encryption_mode: str, key: str) -> list:\n plaintexts = []\n\n for ciphertext in ciphertexts:\n open_ssl_commands = [\n \"openssl\",\n \"enc\",\n \"-d\",\n f\"{encryption_mode}\",\n \"-nosalt\",\n \"-k\",\n f\"{key}\",\n ]\n plaintext = subprocess.check_output(open_ssl_commands, input=ciphertext)\n plaintexts.append(plaintext)\n\n return plaintexts\n\n\ndef decrypt_files(file_paths, encryption_mode, key):\n\n for path in file_paths:\n open_ssl_commands = [\n \"openssl\",\n \"enc\",\n \"-d\",\n f\"{encryption_mode}\",\n \"-nosalt\",\n \"-k\",\n f\"{key}\",\n \"-in\",\n f\"{path}\",\n \"-out\",\n f\"{path}.dec\",\n ]\n output = subprocess.check_output(open_ssl_commands)\n","sub_path":"List_3/cipher.py","file_name":"cipher.py","file_ext":"py","file_size_in_byte":3774,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"640331184","text":"import numpy as np\nfrom data_utils import load_vocab\nimport constants\n\n\n# NLPLAB_W2V = 'data/w2v_model/wikipedia-pubmed-and-PMC-w2v.bin'\n# NLPLAB_W2V = 'data/w2v_model/BioWordVec_PubMed_MIMICIII_d200.vec.bin'\nNLPLAB_W2V = 'data/w2v_model/w2v_retrain.bin'\n\n\ndef export_trimmed_nlplab_vectors(vocab, trimmed_filename, dim=200, bin=NLPLAB_W2V):\n \"\"\"\n Saves glove vectors in numpy array\n\n Args:\n vocab: dictionary vocab[word] = index\n trimmed_filename: a path where to store a matrix in npy\n dim: (int) dimension of embeddings\n :param bin:\n \"\"\"\n # embeddings contains embedding for the pad_tok as well\n embeddings = np.zeros([len(vocab) + 1, dim])\n with open(bin, 'rb') as f:\n header = f.readline()\n vocab_size, layer1_size = map(int, header.split())\n print('nlplab vocab size', vocab_size)\n binary_len = np.dtype('float32').itemsize * layer1_size\n\n count = 0\n m_size = len(vocab)\n for line in range(vocab_size):\n word = []\n while True:\n ch = f.read(1)\n if ch == b' ':\n word = b''.join(word)\n break\n if ch != b'\\n':\n word.append(ch)\n word = word.decode(\"utf-8\")\n\n if word in vocab:\n count += 1\n embedding = np.fromstring(f.read(binary_len), dtype='float32')\n word_idx = vocab[word]\n embeddings[word_idx] = embedding\n else:\n f.read(binary_len)\n\n print('Missing rate {}'.format(1.0 * (m_size - count)/m_size))\n np.savez_compressed(trimmed_filename, embeddings=embeddings)\n\n\nvocab_words = load_vocab(constants.ALL_WORDS)\nexport_trimmed_nlplab_vectors(vocab_words, 'w2v_retrain_nlplab.npz')\n","sub_path":"data/w2v_model/trim_w2v.py","file_name":"trim_w2v.py","file_ext":"py","file_size_in_byte":1819,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"173865345","text":"'''\nCreated on Jun 12, 2014\n\n@author: CKenley\n'''\nimport logging\nfrom django.http import Http404\nfrom django.shortcuts import render\n\nlogging.basicConfig(\n format=\"%(asctime)s : %(levelname)s : %(message)s\",\n filename=\"user.log\",\n level=logging.DEBUG\n )\n\n\ndef ProfileView(request):\n if 'user' in request.session:\n user = request.session['user']\n return render(request, \"user/profile.html\", {\"user\": user})\n errmsg = \"User could not be found in session!\"\n logging.error(errmsg)\n raise Http404(errmsg)\n","sub_path":"BTC/btc_wizard/user/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":538,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"126213086","text":"'''Crie um programa que tenha uma tupla totalmente preenchida com uma contagem por extenso, de zero até vinte. Seu\r\nprograma deverá ler um número pelo teclado (entre 0 e 20) e mostrá-lo por extenso.'''\r\n\r\nnums = ('zero', 'um', 'dois', 'tres', 'quatro', 'cinco', 'seis', 'sete', 'oito', 'nove', 'dez', 'onze', 'doze', 'treze',\r\n'quatorze', 'quinze', 'dezesseis', 'dezessete', 'dezoito', 'dezenove', 'vinte')\r\n\r\nn = int(input('Digite um número de 0 a 20: '))\r\nwhile True:\r\n if n < 0 or n > 20:\r\n n = int(input('Valor digitado inválido. Digite um número de 0 a 20:'))\r\n else:\r\n break\r\nprint(nums[n])\r\n\r\nresp = input('Você quer que continue a contagem? [S/N] ').upper().strip()[0]\r\n\r\nwhile True:\r\n if resp != 'S' and resp != 'N':\r\n resp = input('Resposta inválida. Você quer que continue a contagem? [S/N] ').upper().strip()[0]\r\n\r\n if resp == 'S':\r\n if n != 20:\r\n n += 1\r\n print(nums[n])\r\n\r\n elif resp == 'N':\r\n break\r\n\r\nprint('Sequencia terminada.')\r\n","sub_path":"Ex072.py","file_name":"Ex072.py","file_ext":"py","file_size_in_byte":1030,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"489876372","text":"\nfrom __future__ import division\nimport os\nimport math\nimport json\nimport base64\nimport tempfile\nimport requests\nimport argparse\nimport matplotlib\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.mplot3d import Axes3D, axes3d\n\n\ndef stabilityclass_day(u,cloud,UV):\n if (cloud > 0.5): return 'D'\n if UV < 300:\n if (u<2): return 'B'\n if (2<=u<=5): return 'C'\n if (u>5): return 'D'\n elif 300 <= UV <= 600:\n if (u<2): return 'A'\n if 2<=u<5: return 'B'\n if 5<=u<=6: return 'C'\n if u>6: return 'D'\n elif UV > 600:\n if (u<3): return 'A'\n if (3<=u<=5): return 'B'\n if (u>5): return 'C'\n\ndef stabilityclass_night(u,cloud):\n if (cloud > 0.5): return 'D'\n if 0.355: return 'D'\n elif cloud<=0.35:\n if u<=3: return 'E'\n if u>3: return 'D'\n\ndef stabilityclass_latlon(lat,lon):\n ''' Calls an API to get weather data from location\n Returns stability class at day and night based on weather'''\n import json\n import requests\n # Gather weather parameters time='day'\n try:\n apikey='a6a267d35d8c445bbc4f74dca9543661'\n url='https://api.weatherbit.io/v2.0/current?lat={}&lon={}&key={}'.format(lat,lon,apikey)\n json_response = requests.get(url).json()\n except:\n raise Exception('API call was not possible')\n RH=json_response['data'][0]['rh']\n Irradiance=json_response['data'][0]['solar_rad']\n rain=json_response['data'][0]['precip']\n clouds=(json_response['data'][0]['clouds'])/100\n u=json_response['data'][0]['wind_spd']\n UV=json_response['data'][0]['uv']\n city=json_response['data'][0]['city_name']\n country=json_response['data'][0]['country_code']\n\n stabilityclasses={}\n stabilityclasses['Day']= stabilityclass_day(u,clouds,Irradiance)\n stabilityclasses['Night']= stabilityclass_night(u,clouds)\n\n return (stabilityclasses,u,RH,Irradiance,rain,clouds,UV,city, country)\n\ndef stabilityclass_input(u,cloud,UV):\n u=float(u) ; cloud=float(cloud)\n stabilityclasses={}\n stabilityclasses['Day']= stabilityclass_day(u,cloud,UV)\n stabilityclasses['Night']= stabilityclass_night(u,cloud)\n return stabilityclasses\n\n\n## Functions for printing the graph\n\ndef graph_2D(allXs,allYs,allCs,stabilityclass,u,time):\n matplotlib.use('agg')\n plt.scatter(allYs,allXs,c=allCs,cmap='nipy_spectral_r')\n cbar = plt.colorbar()\n cbar.set_label('Number of aeciospores deposited')\n plt.clim(0,5000)\n plt.xlabel('Horizontal plane (m)')\n plt.ylabel('Downwind of source distance (m)')\n plt.title('Stability class %s. Wind speed: %s m/s' % (stabilityclass,u))\n\n with tempfile.TemporaryFile(suffix=\".png\") as tmpfile:\n plt.savefig(tmpfile,format=\"png\")\n plt.clf()\n tmpfile.seek(0)\n return base64.b64encode(tmpfile.read())\n\n\ndef graph_3D(allXs,allYs,allZs,allCs, stability_class,u,time):\n fig = plt.figure()\n ax = fig.gca(projection='3d')\n p = ax.scatter(allXs, allYs, allZs, zdir='z', s=20, c=allCs, cmap='nipy_spectral_r', depthshade=True)\n ax.legend()\n fig.colorbar(p)\n plt.ylabel('Cross-wind (m)')\n plt.xlabel('Distance (m)')\n plt.yticks([0,1,2])\n plt.title('Stability Class: %s. Wind: %s m/s' %(stability_class,u))\n plt.savefig('3D_%s_%s'% (city,time))\n\n\n\ndef calculateCs(stability_class,x,y,z,H,Q0,u,I,R, time):\n Vs=0.0113 # m/s 11.3 mm/s # Vd=1.27 # Vs=1.13 # urediniospores = 11.5 mm/s\n z=0\n stabilities={ #a b 10P q\n 'A': [0.28,0.9,0.527,0.865],\n 'B': [0.23,0.85,0.371,0.866],\n 'C': [0.22,0.8,0.209,0.897],\n 'D': [0.2,0.76,0.128,0.905],\n 'E': [0.15,0.73,0.098,0.902],\n 'F': [0.12,0.67,0.065,0.902],\n }\n # h=0.7\n hd=0.2\n z0=0.13*hd #0.029\n kz0= (10*z0)**(0.53*(x**-0.22))\n d=0.56*hd\n a = stabilities[stability_class][0]\n b = stabilities[stability_class][1]\n p = stabilities[stability_class][2]\n q = stabilities[stability_class][3]\n sigy=kz0*p*(x**q)\n sigz=kz0*a*(x**b)\n sig2y=sigy**2\n sig2z=sigz**2\n secondpart=math.exp(-(((H-z)**2)/(2*sig2z)))+math.exp(-(((H+z-2*d)**2)/(2*sig2z)))\n\n if time=='Day': Fs=math.exp(-(I*x)/5555*u) #5555*u) #18.01\n if time=='Night': Fs=1\n\n Yw=0.000272*(R**0.7873)\n Yd1=math.sqrt(2/math.pi)*(Vs/x)\n Y2a=((10*z0)**(0.53*(x**(-0.22))))*(x**(0.22-b+1))\n Y2b=math.log(10*z0)*(0.53*0.22)\n Yd2=Y2a*Y2b/a\n Yd=Yd1*Yd2\n\n Fd=math.exp((-(Yw+abs(Yd))*x)/u)\n Q=Q0*Fd*Fs\n\n C=(Q/u)*(math.exp((-y**2)/(2*sig2y))/(2*math.pi*sigz*sigy))*secondpart\n return C\n\n\ndef runmodel(graph,H,Q,u,I,R,clouds,stabilityclasses):\n xmax=100.02\n Xlist= np.arange(0.1,xmax,0.1) #Xlist= np.arange(0.001,20,0.001)\n Ylist=np.arange(-5,5,0.1) # Zlist=np.arange(0,H*2,0.1)\n z=0\n times=['Day','Night']\n maxdistances={}\n for time in times:\n allCs=[]\n allXs=[]\n allYs=[]\n for x in Xlist:\n for y in Ylist:\n stabilityclass=stabilityclasses[time]\n C=calculateCs(stabilityclass,x,y,z,H,Q,u,I,R,time)\n allCs.append(C)\n allYs.append(y)\n allXs.append(x)\n\n\n str_img=graph_2D(allXs,allYs,allCs,stabilityclass,u,time)\n\n Ccum = np.cumsum(allCs)\n max99=max(Ccum)*0.999\n max95=max(Ccum)*0.95\n max75=max(Ccum)*0.75\n max50=max(Ccum)*0.50\n\n X99 = round(allXs[[n for n,i in enumerate(Ccum) if i> max99][0]],1)\n X95 = round(allXs[[n for n,i in enumerate(Ccum) if i> max95][0]],1)\n X75 = round(allXs[[n for n,i in enumerate(Ccum) if i> max75][0]],1)\n X50 = round(allXs[[n for n,i in enumerate(Ccum) if i> max50][0]],1)\n\n valuesaty0=[]\n Xmax=\"more than 100\"\n for i,x in enumerate(allXs):\n if str(round(allYs[i],2))==\"-0.0\" or str(round(allYs[i],2))==\"0.0\":\n if allCs[i]<1 and x>X99:\n Xmax=round(x,2)\n break\n\n maxdistances[time]=[X95,X75,X50,X99,Xmax,str_img]\n return maxdistances\n","sub_path":"dispersal/GPModel/GPM_django_PF.py","file_name":"GPM_django_PF.py","file_ext":"py","file_size_in_byte":6151,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"401997026","text":"# 人工神经网络第四次作业1-2,异联想DHNN\nimport random\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport os\nimport copy\nfrom numpy import nan\nfrom typing import Optional\ndef data_loader(path): # 数据生成:获得全部数据\n with open(path,'r',encoding='UTF-8') as f:\n data = []\n for line in f:\n d = [int(j) for j in line.strip()]\n data.append(d)\n return np.array(data)\ndef visualize(W,title,xlabel,ylabel,path): # 可视化\n plt.clf()\n for i,w in enumerate(W):\n cols,rows = np.where(w.reshape(7,5) != nan)\n plt.subplot(2,4,i+1)\n ax = plt.gca() # 获取到当前坐标轴信息\n ax.xaxis.set_ticks_position('top') # 将X坐标轴移到上面\n ax.invert_yaxis() # 翻转y轴\n plt.margins(0.1)\n plt.scatter(rows,cols,s=w*100,alpha=0.8,c=np.random.RandomState(0).rand(len(rows)),cmap='Dark2') # 绘制字符点\n plt.grid(visible=False) # 不显示网格线\n plt.draw()\n plt.tight_layout() # 自适应调整子图大小\n plt.savefig(path) # 保存图像\ndef bam(W,x,y_last): # X状态更新\n \"\"\"\n input : W: connnection weight; x: flattrend letter vector, \n output: y: predection for x; x_new: updated x; ee: energy\n \"\"\"\n y = np.sign(x @ W) # 1x3\n y[np.where(y==0)] = y_last[np.where(y==0)]\n x_new = np.sign(y @ W.T) # 1x35\n x_new[np.where(x_new==0)] = x[np.where(x_new==0)]\n return x_new,y\ndef cal_energy(W,x,y): # 计算当前的能量函数值\n return -((x @ W) @ y)\ndef addnoise(c,noise_ratio = 0.1): # 按一定比例增加噪声\n noisenum = int(len(c) * noise_ratio)\n noisepos = [1]*len(c)\n noisepos[:noisenum] = [-1]*noisenum\n np.random.shuffle(noisepos)\n cc = np.array([x*y for x,y in zip(c,noisepos)])\n return cc\ndef showChar(c,offsetx,offsety,height,weight): # 单个字符绘制函数\n cc = list(zip(*([iter(c)]*weight)))\n x = [] # 非字符点\n y = []\n X = [] # 字符点\n Y = []\n for id,a in enumerate(cc):\n YY = offsety + height - id\n for iidd,b in enumerate(a):\n XX = offsetx + iidd\n if b == 0:\n x.append(XX)\n y.append(YY)\n else:\n X.append(XX)\n Y.append(YY)\n plt.scatter(x,y,s=50,alpha=1,marker='*') # 绘制非字符点\n plt.scatter(X,Y,s=500,alpha=0.8,c=np.random.RandomState(0).rand(len(X)),cmap='Dark2') # 绘制字符点\ndef calCharXY(id,height,weight): # 计算字符绘制坐标函数\n offsetx = id*height\n offsety = weight + 5\n if id >= 4:\n offsetx = (id-4)*height\n offsety = 0\n return offsetx,offsety\ndef savePltPic(title,xlabel,ylabel,path): # 我的标准保存图像函数\n font = {'family':'serif','style':'italic','weight':'bold','color':'black','size':20} # 设置标签字体\n plt.title(title,fontdict=font,fontsize=18) # 显示传递函数类型\n plt.xlabel(xlabel,fontdict=font,fontsize=15) # 设置x轴标签\n plt.ylabel(ylabel,fontdict=font,fontsize=15) # 设置y轴标签\n plt.grid(visible=False) # 不显示网格线\n plt.tight_layout() # 自适应调整子图大小\n plt.savefig(path) # 保存图像\n plt.close() # 关闭图像\nnp.random.seed(1) # 设置随机数种子\nplt.rcParams['figure.figsize'] = (12,8)\ndata_path = './DATA/char.txt' # 八字母数据集路径\nnumber_data = data_loader(data_path) # 读取八字母数据集\nCharStr = np.array(['A','B','C','D','E','F','G','H','I','J','K','L','M','N','O','P','Q','R','S','T','U','V','W','X','Y','Z']) # 字母列表初始化,选择ZHUOQING一共8个字母\ntarget = np.array(['G','N','I','Q','O','U','H','Z'])\n# 生成训练数据和标签\nX = [] # 8x35\nY = [] # 8x3\nfor index,letter in enumerate(target):\n id_ = np.where(CharStr == letter)[0][0]\n X.append(number_data[id_])\n bi = bin(index).replace('0b','')\n la = np.array([int(i) for i in f'{bi:0>3}'])\n la[la == 0] = -1\n Y.append(la)\nif False: # 绘制八字母数据集\n for id,xx in enumerate(X):\n offsetx,offsety = calCharXY(id=id,height=7,weight=5) # 计算字符绘制坐标\n showChar(c=xx,offsetx=offsetx,offsety=offsety,height=7,weight=5) # 绘制单个字符\n savePltPic(title='Alphabet dataset',xlabel='x',ylabel='y',path='./OUTPUT/8number_char.jpg')\nX = np.array(X).astype('float32')*2-1 # 二进制(0,1)转换为双极性(-1,1)\nY = np.array(Y).astype('float32')\nw_matrix = X.T @ Y # 计算权系数矩阵W\nif False: # 绘制权系数矩阵\n plt.imshow(w_matrix)\n savePltPic(title='Weight coefficient matrix',xlabel='x',ylabel='y',path='./OUTPUT/8number_w.jpg')\nenergy = np.zeros(8) # 计算能量函数\nfor i in range(8):\n energy[i] = cal_energy(w_matrix,X[i],Y[i])\n# print(f'Energy: {energy}')\n# # 从噪声数据中还原到对应的标签\n# timeNum = 2\n# # plt.draw()\n# # plt.pause(0.2)\n# for noi in range(4): # 4种加噪声的结果\n# # 生成加噪声数据X_noise\n# X_noise = np.zeros(X.shape)\n# for i,x in enumerate(X):\n# X_noise[i] = addnoise(x,0.2) \n# # 保留原始数据,在X_copy上更新状\n# X_copy = X_noise.copy()\n# # visualize(X_copy,f'./OUTPUT/8number{noi}.png')\n# # 还原到对应的标签\n# for j,x in enumerate(X_noise): # 由于更新的是状态,权值不变,那么<对所有样本迭代一遍再重复time次>和<逐个对单个样本迭代time次> 一样\n# y = np.sign(x @ w_matrix)\n# for _ in range(timeNum): # 迭代次数\n# ee = cal_energy(w_matrix,x,y)\n# x,y = bam(w_matrix,x,y)\n# X_copy[j] = x\n# print(y,Y[j],(y==Y[j]).all()) \n# print('----------------------------------------')\n# # visualize(X_copy,f'./OUTPUT/v{noi}.png')","sub_path":"NeuralModel/ANN27.py","file_name":"ANN27.py","file_ext":"py","file_size_in_byte":5856,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"21979046","text":"from dictionaries.exe_1 import *\ndicSum=create_dic()\nprint(dicSum)\nkey = int(input(\"enter wanted key\"))\nfor i in dicSum.keys():\n print(i)\n if i==key:\n print(\"yes\")\n break\n else:\n print(\"no\")\ncreate_dic()","sub_path":"dictionaries/exe_2.py","file_name":"exe_2.py","file_ext":"py","file_size_in_byte":233,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"362350392","text":"\"\"\"\n For models: model_v1_x.py\n\"\"\"\n\nimport os\nimport time\nimport sys\nimport shutil\nimport random\nfrom time import strftime\nfrom argparse import ArgumentParser\nimport numpy as np\nimport torch\nimport torch.utils.data\nimport torch.nn.functional as F\ntorch.multiprocessing.set_sharing_strategy('file_system')\nfrom PIL import Image\nfrom subprocess import call\nfrom sapien_data import PartNetSapienDataset\nimport utils\nfrom geometry_utils import render_pts\nimport sapien.core as sapien\nBASE_DIR = os.path.dirname(os.path.abspath(__file__))\n\nimport logging\nlogger = logging.getLogger(\"trimesh\")\nlogger.setLevel(logging.ERROR)\n\n\ndef train(conf):\n # create training and validation datasets and data loaders\n data_features = ['pc', 'shape_id']\n\n train_dataset = PartNetSapienDataset(train=True)\n utils.printout(conf.flog, str(train_dataset))\n train_dataloader = torch.utils.data.DataLoader(train_dataset, batch_size=conf.batch_size, shuffle=True, pin_memory=True, \\\n num_workers=conf.num_workers, drop_last=True, collate_fn=utils.collate_feats, worker_init_fn=utils.worker_init_fn)\n\n val_dataset = PartNetSapienDataset(train=False)\n utils.printout(conf.flog, str(val_dataset))\n val_dataloader = torch.utils.data.DataLoader(val_dataset, batch_size=conf.batch_size, shuffle=False, pin_memory=True, \\\n num_workers=0, drop_last=True, collate_fn=utils.collate_feats, worker_init_fn=utils.worker_init_fn)\n\n # load network model\n model_def = utils.get_model_module(conf.model_version)\n\n # create models\n network = model_def.Network(conf)\n utils.printout(conf.flog, '\\n' + str(network) + '\\n')\n\n models = [network]\n model_names = ['network']\n\n # create optimizers\n network_opt = torch.optim.Adam(network.parameters(), lr=conf.lr, weight_decay=conf.weight_decay)\n optimizers = [network_opt]\n optimizer_names = ['network_opt']\n\n # learning rate scheduler\n network_lr_scheduler = torch.optim.lr_scheduler.StepLR(network_opt, step_size=conf.lr_decay_every, gamma=conf.lr_decay_by)\n\n # create logs\n if not conf.no_console_log:\n header = ' Time Epoch Dataset Iteration Progress(%) LR ReconLoss KLDivLoss TotalLoss'\n if not conf.no_tb_log:\n # https://github.com/lanpa/tensorboard-pytorch\n from tensorboardX import SummaryWriter\n train_writer = SummaryWriter(os.path.join(conf.exp_dir, 'train'))\n val_writer = SummaryWriter(os.path.join(conf.exp_dir, 'val'))\n\n # send parameters to device\n for m in models:\n m.to(conf.device)\n for o in optimizers:\n utils.optimizer_to_device(o, conf.device)\n\n # start training\n start_time = time.time()\n\n last_checkpoint_step = None\n last_train_console_log_step, last_val_console_log_step = None, None\n train_num_batch = len(train_dataloader)\n val_num_batch = len(val_dataloader)\n\n # train for every epoch\n for epoch in range(conf.epochs):\n if not conf.no_console_log:\n utils.printout(conf.flog, f'training run {conf.exp_name}')\n utils.printout(conf.flog, header)\n\n train_batches = enumerate(train_dataloader, 0)\n val_batches = enumerate(val_dataloader, 0)\n train_fraction_done = 0.0\n val_fraction_done = 0.0\n val_batch_ind = -1\n\n # train for every batch\n for train_batch_ind, batch in train_batches:\n train_fraction_done = (train_batch_ind + 1) / train_num_batch\n train_step = epoch * train_num_batch + train_batch_ind\n\n log_console = not conf.no_console_log and (last_train_console_log_step is None or \\\n train_step - last_train_console_log_step >= conf.console_log_interval)\n if log_console:\n last_train_console_log_step = train_step\n\n # set models to training mode\n for m in models:\n m.train()\n\n # forward pass (including logging)\n total_loss = forward(batch=batch, data_features=data_features, network=network, conf=conf, is_val=False, \\\n step=train_step, epoch=epoch, batch_ind=train_batch_ind, num_batch=train_num_batch, start_time=start_time, \\\n log_console=log_console, log_tb=not conf.no_tb_log, tb_writer=train_writer, lr=network_opt.param_groups[0]['lr'])\n\n # optimize one step\n network_opt.zero_grad()\n total_loss.backward()\n network_opt.step()\n network_lr_scheduler.step()\n\n # save checkpoint\n with torch.no_grad():\n if last_checkpoint_step is None or train_step - last_checkpoint_step >= conf.checkpoint_interval:\n utils.printout(conf.flog, 'Saving checkpoint ...... ')\n utils.save_checkpoint(models=models, model_names=model_names, dirname=os.path.join(conf.exp_dir, 'ckpts'), \\\n epoch=epoch, prepend_epoch=True, optimizers=optimizers, optimizer_names=model_names)\n utils.printout(conf.flog, 'DONE')\n last_checkpoint_step = train_step\n\n # validate one batch\n while val_fraction_done <= train_fraction_done and val_batch_ind+1 < val_num_batch:\n val_batch_ind, val_batch = next(val_batches)\n\n val_fraction_done = (val_batch_ind + 1) / val_num_batch\n val_step = (epoch + val_fraction_done) * train_num_batch - 1\n\n log_console = not conf.no_console_log and (last_val_console_log_step is None or \\\n val_step - last_val_console_log_step >= conf.console_log_interval)\n if log_console:\n last_val_console_log_step = val_step\n\n # set models to evaluation mode\n for m in models:\n m.eval()\n\n with torch.no_grad():\n # forward pass (including logging)\n __ = forward(batch=val_batch, data_features=data_features, network=network, conf=conf, is_val=True, \\\n step=val_step, epoch=epoch, batch_ind=val_batch_ind, num_batch=val_num_batch, start_time=start_time, \\\n log_console=log_console, log_tb=not conf.no_tb_log, tb_writer=val_writer, lr=network_opt.param_groups[0]['lr'])\n\n # save the final models\n utils.printout(conf.flog, 'Saving final checkpoint ...... ')\n utils.save_checkpoint(models=models, model_names=model_names, dirname=os.path.join(conf.exp_dir, 'ckpts'), \\\n epoch=epoch, prepend_epoch=False, optimizers=optimizers, optimizer_names=optimizer_names)\n utils.printout(conf.flog, 'DONE')\n\n\ndef forward(batch, data_features, network, conf, \\\n is_val=False, step=None, epoch=None, batch_ind=0, num_batch=1, start_time=0, \\\n log_console=False, log_tb=False, tb_writer=None, lr=None):\n # prepare input\n input_pcs = torch.cat(batch[data_features.index('pc')], dim=0).to(conf.device) # B x N x 3\n batch_size = input_pcs.shape[0]\n\n # forward through the network\n output_pcs, pc_feats, ret_list = network(input_pcs) # B x N x 3, B x P\n\n # for each type of loss, compute losses per data\n recon_loss_per_data = network.get_loss(input_pcs, output_pcs)\n\n kldiv_loss_per_data = torch.zeros_like(recon_loss_per_data)\n if conf.probabilistic:\n kldiv_loss_per_data = ret_list['kldiv_loss']\n\n # for each type of loss, compute avg loss per batch\n recon_loss = recon_loss_per_data.mean()\n kldiv_loss = kldiv_loss_per_data.mean()\n\n # compute total loss\n total_loss = recon_loss + conf.kldiv_loss_weight * kldiv_loss\n\n # display information\n data_split = 'train'\n if is_val:\n data_split = 'val'\n\n with torch.no_grad():\n # log to console\n if log_console:\n utils.printout(conf.flog, \\\n f'''{strftime(\"%H:%M:%S\", time.gmtime(time.time()-start_time)):>9s} '''\n f'''{epoch:>5.0f}/{conf.epochs:<5.0f} '''\n f'''{data_split:^10s} '''\n f'''{batch_ind:>5.0f}/{num_batch:<5.0f} '''\n f'''{100. * (1+batch_ind+num_batch*epoch) / (num_batch*conf.epochs):>9.1f}% '''\n f'''{lr:>5.2E} '''\n f'''{recon_loss.item():>10.5f}'''\n f'''{kldiv_loss.item():>10.5f}'''\n f'''{total_loss.item():>10.5f}''')\n conf.flog.flush()\n\n # log to tensorboard\n if log_tb and tb_writer is not None:\n tb_writer.add_scalar('recon_loss', recon_loss.item(), step)\n tb_writer.add_scalar('kldiv_loss', kldiv_loss.item(), step)\n tb_writer.add_scalar('total_loss', total_loss.item(), step)\n tb_writer.add_scalar('lr', lr, step)\n\n # gen visu\n if is_val and (not conf.no_visu) and epoch % conf.num_epoch_every_visu == 0:\n visu_dir = os.path.join(conf.exp_dir, 'val_visu')\n out_dir = os.path.join(visu_dir, 'epoch-%04d' % epoch)\n input_pcs_dir = os.path.join(out_dir, 'input_pcs')\n output_pcs_dir = os.path.join(out_dir, 'output_pcs')\n info_dir = os.path.join(out_dir, 'info')\n\n if batch_ind == 0:\n # create folders\n os.mkdir(out_dir)\n os.mkdir(input_pcs_dir)\n os.mkdir(output_pcs_dir)\n os.mkdir(info_dir)\n\n if batch_ind < conf.num_batch_every_visu:\n utils.printout(conf.flog, 'Visualizing ...')\n\n for i in range(batch_size):\n fn = 'data-%03d.png' % (batch_ind * batch_size + i)\n\n # render_pts(os.path.join(input_pcs_dir, fn), input_pcs[i].cpu().numpy())\n # render_pts(os.path.join(output_pcs_dir, fn), output_pcs[i].cpu().numpy())\n # or to render using matplotlib\n utils.render_pc(os.path.join(input_pcs_dir, fn), input_pcs[i].cpu().numpy())\n utils.render_pc(os.path.join(output_pcs_dir, fn), output_pcs[i].cpu().numpy())\n\n with open(os.path.join(info_dir, fn.replace('.png', '.txt')), 'w') as fout:\n fout.write('shape_id: %s\\n' % batch[data_features.index('shape_id')][i])\n fout.write('recon_loss: %f\\n' % recon_loss_per_data[i].item())\n fout.write('kldiv_loss: %f\\n' % kldiv_loss_per_data[i].item())\n\n if batch_ind == conf.num_batch_every_visu - 1:\n # visu html\n utils.printout(conf.flog, 'Generating html visualization ...')\n sublist = 'input_pcs,output_pcs,info'\n cmd = 'cd %s && python %s . 10 htmls %s %s > /dev/null' % (out_dir, os.path.join(BASE_DIR, '../utils/gen_html_hierarchy_local.py'), sublist, sublist)\n call(cmd, shell=True)\n utils.printout(conf.flog, 'DONE')\n\n return total_loss\n\n\nif __name__ == '__main__':\n\n ### get parameters\n parser = ArgumentParser()\n\n # main parameters (required)\n parser.add_argument('--exp_suffix', type=str, help='exp suffix')\n parser.add_argument('--model_version', type=str, help='model def file')\n\n # main parameters (optional)\n parser.add_argument('--device', type=str, default='cuda:0', help='cpu or cuda:x for using cuda on GPU number x')\n parser.add_argument('--seed', type=int, default=3124256514, help='random seed (for reproducibility) [specify -1 means to generate a random one]')\n #parser.add_argument('--seed', type=int, default=-1, help='random seed (for reproducibility) [specify -1 means to generate a random one]')\n parser.add_argument('--log_dir', type=str, default='logs', help='exp logs directory')\n parser.add_argument('--data_dir', type=str, help='data directory')\n parser.add_argument('--val_data_dir', type=str, help='data directory')\n parser.add_argument('--overwrite', action='store_true', default=False, help='overwrite if exp_dir exists [default: False]')\n\n # network settings\n parser.add_argument('--num_point', type=int, default=2048)\n parser.add_argument('--decoder_type', type=str, default='fc')\n parser.add_argument('--loss_type', type=str, default='cd')\n parser.add_argument('--kldiv_loss_weight', type=float, default=1e-4)\n parser.add_argument('--probabilistic', action='store_true', default=False, help='probabilistic [default: False]')\n\n # training parameters\n parser.add_argument('--epochs', type=int, default=1000)\n parser.add_argument('--batch_size', type=int, default=16)\n parser.add_argument('--num_workers', type=int, default=5)\n parser.add_argument('--lr', type=float, default=.001)\n parser.add_argument('--weight_decay', type=float, default=1e-5)\n parser.add_argument('--lr_decay_by', type=float, default=0.9)\n parser.add_argument('--lr_decay_every', type=float, default=5000)\n\n # loss weights\n\n # logging\n parser.add_argument('--no_tb_log', action='store_true', default=False)\n parser.add_argument('--no_console_log', action='store_true', default=False)\n parser.add_argument('--console_log_interval', type=int, default=10, help='number of optimization steps beween console log prints')\n parser.add_argument('--checkpoint_interval', type=int, default=10000, help='number of optimization steps beween checkpoints')\n\n # visu\n parser.add_argument('--num_batch_every_visu', type=int, default=1, help='num batch every visu')\n parser.add_argument('--num_epoch_every_visu', type=int, default=10, help='num epoch every visu')\n parser.add_argument('--no_visu', action='store_true', default=False, help='no visu? [default: False]')\n\n # parse args\n conf = parser.parse_args()\n\n\n ### prepare before training\n # make exp_name\n conf.exp_name = f'exp-{conf.model_version}-{conf.exp_suffix}'\n\n # mkdir exp_dir; ask for overwrite if necessary\n conf.exp_dir = os.path.join(conf.log_dir, conf.exp_name)\n if os.path.exists(conf.exp_dir):\n if not conf.overwrite:\n response = input('A training run named \"%s\" already exists, overwrite? (y/n) ' % conf.exp_name)\n if response != 'y':\n exit(1)\n shutil.rmtree(conf.exp_dir)\n os.mkdir(conf.exp_dir)\n os.mkdir(os.path.join(conf.exp_dir, 'ckpts'))\n if not conf.no_visu:\n os.mkdir(os.path.join(conf.exp_dir, 'val_visu'))\n\n # control randomness\n if conf.seed < 0:\n conf.seed = random.randint(1, 10000)\n random.seed(conf.seed)\n np.random.seed(conf.seed)\n torch.manual_seed(conf.seed)\n\n # save config\n torch.save(conf, os.path.join(conf.exp_dir, 'conf.pth'))\n\n # file log\n flog = open(os.path.join(conf.exp_dir, 'train_log.txt'), 'w')\n conf.flog = flog\n\n # backup command running\n utils.printout(flog, ' '.join(sys.argv) + '\\n')\n utils.printout(flog, f'Random Seed: {conf.seed}')\n\n # backup python files used for this training\n os.system('cp data.py models/%s.py %s %s' % (conf.model_version, __file__, conf.exp_dir))\n\n # set training device\n device = torch.device(conf.device)\n utils.printout(flog, f'Using device: {conf.device}\\n')\n conf.device = device\n\n ### start training\n train(conf)\n\n\n ### before quit\n # close file log\n flog.close()\n\n","sub_path":"exps/exp_baseline1/train_v1.py","file_name":"train_v1.py","file_ext":"py","file_size_in_byte":15280,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"43277799","text":"from django.shortcuts import render,HttpResponse\nfrom layui import models\nfrom django.core import serializers\n\n# Create your views here.\ndef test(request):\n if request.method==\"GET\":\n print(\"用户已经创建\")\n # for i in range(0,100):\n # obj=models.Cuetomer.objects.create(name='我是%s号'%i,age=i)\n return render(request, 'test1.html')\n if request.method==\"POST\":\n print(request.POST.get(\"time\"),'time')\n\n return render(request, 'test1.html')\n\nimport json\n\ndef getinfo(request):\n if request.method==\"GET\":\n print(\"get_info geT\")\n return HttpResponse(\"GET\")\n if request.method==\"POST\":\n print('page',request.POST.get(\"start\"))\n page=request.POST.get(\"start\")\n\n print(\"来取数据了\")\n data=models.Cuetomer.objects.filter(id=page).values('name')[0]['name']\n print(data)\n ret = {'status': False, 'mydata': ''}\n ret['mydata'] = data\n # return render(request,'test1.html',{mydata:'123'})\n return HttpResponse(json.dumps(ret))\n\n\n\n\n\n","sub_path":"mylayui/layui/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1065,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"364144198","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nimport datetime\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('name', '0015_insert_review_name_data'),\n ('group', '0008_auto_20160505_0523'),\n ('person', '0014_auto_20160613_0751'),\n ('doc', '0012_auto_20160207_0537'),\n ]\n\n operations = [\n migrations.CreateModel(\n name='NextReviewerInTeam',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('next_reviewer', models.ForeignKey(to='person.Person')),\n ('team', models.ForeignKey(to='group.Group')),\n ],\n options={\n 'verbose_name': 'next reviewer in team setting',\n 'verbose_name_plural': 'next reviewer in team settings',\n },\n bases=(models.Model,),\n ),\n migrations.CreateModel(\n name='ResultUsedInReviewTeam',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('result', models.ForeignKey(to='name.ReviewResultName')),\n ('team', models.ForeignKey(to='group.Group')),\n ],\n options={\n 'verbose_name': 'review result used in team setting',\n 'verbose_name_plural': 'review result used in team settings',\n },\n bases=(models.Model,),\n ),\n migrations.CreateModel(\n name='ReviewerSettings',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('min_interval', models.IntegerField(default=30, verbose_name=b'Can review at most', choices=[(7, b'Once per week'), (14, b'Once per fortnight'), (30, b'Once per month'), (61, b'Once per two months'), (91, b'Once per quarter')])),\n ('filter_re', models.CharField(help_text=b'Draft names matching regular expression should not be assigned', max_length=255, verbose_name=b'Filter regexp', blank=True)),\n ('skip_next', models.IntegerField(default=0, verbose_name=b'Skip next assignments')),\n ('remind_days_before_deadline', models.IntegerField(help_text=b\"To get an email reminder in case you forget to do an assigned review, enter the number of days before a review deadline you want to receive it. Clear the field if you don't want a reminder.\", null=True, blank=True)),\n ('person', models.ForeignKey(to='person.Person')),\n ('team', models.ForeignKey(to='group.Group')),\n ],\n options={\n 'verbose_name_plural': 'reviewer settings',\n },\n bases=(models.Model,),\n ),\n migrations.CreateModel(\n name='ReviewRequest',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('old_id', models.IntegerField(help_text=b'ID in previous review system', null=True, blank=True)),\n ('time', models.DateTimeField(default=datetime.datetime.now)),\n ('deadline', models.DateField()),\n ('requested_rev', models.CharField(help_text=b'Fill in if a specific revision is to be reviewed, e.g. 02', max_length=16, verbose_name=b'requested revision', blank=True)),\n ('reviewed_rev', models.CharField(max_length=16, verbose_name=b'reviewed revision', blank=True)),\n ('doc', models.ForeignKey(related_name='reviewrequest_set', to='doc.Document')),\n ('requested_by', models.ForeignKey(to='person.Person')),\n ('result', models.ForeignKey(blank=True, to='name.ReviewResultName', null=True)),\n ('review', models.OneToOneField(null=True, blank=True, to='doc.Document')),\n ('reviewer', models.ForeignKey(blank=True, to='person.Email', null=True)),\n ('state', models.ForeignKey(to='name.ReviewRequestStateName')),\n ('team', models.ForeignKey(to='group.Group')),\n ('type', models.ForeignKey(to='name.ReviewTypeName')),\n ],\n options={\n },\n bases=(models.Model,),\n ),\n migrations.CreateModel(\n name='ReviewWish',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('time', models.DateTimeField(default=datetime.datetime.now)),\n ('doc', models.ForeignKey(to='doc.Document')),\n ('person', models.ForeignKey(to='person.Person')),\n ('team', models.ForeignKey(to='group.Group')),\n ],\n options={\n 'verbose_name_plural': 'review wishes',\n },\n bases=(models.Model,),\n ),\n migrations.CreateModel(\n name='TypeUsedInReviewTeam',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('team', models.ForeignKey(to='group.Group')),\n ('type', models.ForeignKey(to='name.ReviewTypeName')),\n ],\n options={\n 'verbose_name': 'review type used in team setting',\n 'verbose_name_plural': 'review type used in team settings',\n },\n bases=(models.Model,),\n ),\n migrations.CreateModel(\n name='UnavailablePeriod',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('start_date', models.DateField(default=datetime.date.today, help_text=b\"Choose the start date so that you can still do a review if it's assigned just before the start date - this usually means you should mark yourself unavailable for assignment some time before you are actually away.\")),\n ('end_date', models.DateField(help_text=b'Leaving the end date blank means that the period continues indefinitely. You can end it later.', null=True, blank=True)),\n ('availability', models.CharField(max_length=30, choices=[(b'canfinish', b'Can do follow-ups'), (b'unavailable', b'Completely unavailable')])),\n ('person', models.ForeignKey(to='person.Person')),\n ('team', models.ForeignKey(to='group.Group')),\n ],\n options={\n },\n bases=(models.Model,),\n ),\n ]\n","sub_path":"ietf/review/migrations/0001_initial.py","file_name":"0001_initial.py","file_ext":"py","file_size_in_byte":6704,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"216315239","text":"#!/usr/bin/python\nfrom struct import *\nimport os\n\n\nlibc_binsh = pack(\"= int(nSE * sub)], size=nselect)\n filteredactn = np.vstack((filteredactn1, filteredactn2))\n\n return filteredactn\n\n\ndef get_CPs(rates, pref_msk, actn, dt, smoothwin=100e-3, step=5):\n \"\"\"\n pref_msk is tailored to this population rates and it's a bool!\n get choice probability obtaining the distribution of each timepoint\n for active neurons of a population\n\n :param: rates: ntrls, nSE, timepoints\n :param: pref_msk:\n :param: actn:\n :param: dt:\n :param: smoothwin:\n :param: step:\n :return: aucs # CPs\n \"\"\"\n # params\n timepoints = rates.shape[2]\n this_time = np.linspace(0, timepoints, int(timepoints / step), dtype=int)[:-1]\n nselect = actn.shape[0]\n newdt = dt * step\n kernel = np.ones((int(smoothwin / newdt)))\n prefrates = rates[pref_msk == True, :, :]\n nprefrates = rates[pref_msk == False, :, :]\n\n # allocate variable to save CP\n aucs = np.zeros((nselect, this_time.shape[0]))\n smoothauc = aucs.copy()\n\n # for each neuron that is active\n for i, n in tqdm(enumerate(actn)):\n\n # define max rate of neuron\n maxrate = max(2, rates[:, n, :].max() + 1)\n\n # for each timepoint\n for j, t in enumerate(this_time):\n # get this rate across all trials\n pref = prefrates[:, n, t:t + step]\n npref = nprefrates[:, n, t:t + step]\n\n # hist\n x1, e1 = np.histogram(pref, bins=np.arange(maxrate), density=True)\n x2, e2 = np.histogram(npref, bins=np.arange(maxrate), density=True)\n\n # cumulative distribution\n cx1 = np.concatenate(([0], np.cumsum(x1)))\n cx2 = np.concatenate(([0], np.cumsum(x2)))\n\n # auc\n aucs[i, j] = mtr.auc(cx1, cx2) # reversed because pref > npref\n\n smoothauc[i] = np.convolve(aucs[i], kernel, mode='same') / (smoothwin/newdt)\n\n return smoothauc\n\n\ndef get_corr(rates, actn, dt, smoothwin=250e-3, step=10):\n \"\"\"\n Pearson correlations at each time point between two neurons\n bins are size step.\n\n :param: rates:\n :param: actn:\n :param: dt:\n :param: smoothwin:\n :param: step:\n :return: corrall, corrii, corrij\n \"\"\"\n # params\n timepoints = rates.shape[2]\n this_time = np.linspace(0, timepoints, int(timepoints / step), dtype=int)[:-1]\n actn1 = actn[0]\n actn2 = actn[1]\n nselect = actn.shape[1]\n\n # allocate variables\n corrs = np.zeros((nselect ** 2 * 2, this_time.shape[0]))\n corrs1 = np.zeros((nselect ** 2, this_time.shape[0]))\n corrs2 = corrs1.copy()\n\n # massive for loop through both subpopulations\n for i1, n1 in tqdm(enumerate(actn1)):\n for i2, n2 in enumerate(actn2):\n for j, t in enumerate(this_time):\n\n # get rates for each case\n x11 = rates[:, n1, t:t + step].mean(axis=1)\n x12 = rates[:, actn1[i2], t:t + step].mean(axis=1)\n\n x21 = rates[:, actn2[i1], t:t + step].mean(axis=1)\n x22 = rates[:, n2, t:t + step].mean(axis=1)\n\n x1 = rates[:, n1, t:t + step].mean(axis=1)\n x2 = rates[:, n2, t:t + step].mean(axis=1)\n\n # check if we have info aside from zero\n if not (np.nonzero(x11)[0].size == False) or not (np.nonzero(x12)[0].size == False):\n # correlations between pop1\n corrs1[int(nselect * i1) + i2, j] = np.corrcoef(x11, x12)[0, 1]\n\n if not (np.nonzero(x21)[0].size == False) or not (np.nonzero(x22)[0].size == False):\n # correlations between pop2\n corrs2[int(nselect * i2) + i1, j] = np.corrcoef(x21, x22)[0, 1]\n\n if not (np.nonzero(x1)[0].size == False) or not (np.nonzero(x2)[0].size == False):\n # correlations across pops)\n k = np.corrcoef(x1, x2)\n corrs[int(nselect * i1) + i2, j] = k[0, 1]\n corrs[-int(nselect * i1) + i2, j] = k[1, 0]\n\n # return as corrsall, corrsii, corrsij\n corrsall = np.concatenate((corrs, corrs1, corrs2), axis=0)\n corrsii = np.concatenate((corrs1, corrs2), axis=0)\n\n # allocate variables for smoothing\n newdt = dt * step\n kernel = np.ones((int(smoothwin / newdt)))\n smoothcorrsall = np.zeros(corrsall.shape)\n smoothcorrsii = np.zeros(corrsii.shape)\n smoothcorrsij = np.zeros(corrs.shape)\n\n for n in np.arange(corrsall.shape[0]):\n smoothcorrsall[n] = np.convolve(corrsall[n], kernel, mode='same') / (smoothwin/newdt)\n\n if n < corrs.shape[0]:\n smoothcorrsii[n] = np.convolve(corrsii[n], kernel, mode='same') / (smoothwin/newdt)\n smoothcorrsij[n] = np.convolve(corrs[n], kernel, mode='same') / (smoothwin/newdt)\n\n return smoothcorrsall, smoothcorrsii, smoothcorrsij\n\n\n# decorator for experiment\nthisexperiment = '2018-12-11-09h04m25s'\n\n\n@experiment_opener({'test0': thisexperiment}, load_path, show=plt_show)\ndef plot_fig2(tables_task_ids):\n \"\"\"\n Using the experiment_opener decorator automates some of the tedious aspects of handling experiment\n files, including opening and closing the file, plus it also calls plt.show() if you ask it to.\n And finally, it fixes a problem with SVG files so that they don't explode Inkscape if you import them.\n\n :param tables_task_ids: dict mapping from user supplied name to a tuple of (tables, task_ids)\n :return:\n \"\"\"\n from snep.tables.experiment import ExperimentTables\n\n tables, task_ids = tables_task_ids['test0']\n assert isinstance(tables, ExperimentTables) # This allows PyCharm to autocomplete method names for tables\n params = tables.get_general_params(True)\n param_ranges = tables.read_param_ranges()\n\n # filter tasks to only the ones that reach the targets\n targets = [{('c',): 0, ('bfb',): 0}]\n target_ids = filter_tasks(task_ids, targets)\n\n # -------------------------------------\n # Get experiment results and params\n # -------------------------------------\n # Simulation times\n ntrls = len(target_ids)\n sub = params['sen']['populations']['sub']\n settletime = params['simulation']['settletime'] / second\n runtime = params['simulation']['runtime'] / second - settletime\n stimon = params['simulation']['stimon'] / second - settletime\n stimoff = params['simulation']['stimoff'] / second - settletime\n pops, timepoints = tables.get_raw_data(task_ids[0])['poprates_dec'].shape\n dt = runtime / timepoints\n nSE, downsampltimepoints = tables.get_computed(task_ids[0])['spikes'].shape\n time = np.linspace(0, runtime, timepoints)\n downsampltime = np.linspace(0, runtime, downsampltimepoints)\n downsampldt = runtime / downsampltimepoints\n\n # allocate variables\n rateDE = np.empty((ntrls, pops, timepoints), dtype='float32')\n rateSE = np.empty((ntrls, pops, timepoints), dtype='float32')\n spksSE = np.empty((ntrls, nSE, downsampltimepoints), dtype='float32')\n # evntSE = np.empty((ntrls, nSE, downsampltimepoints), dtype='float32')\n # brstSE = np.empty((ntrls, nSE, downsampltimepoints), dtype='float32')\n # snglSE = np.empty((ntrls, nSE, downsampltimepoints), dtype='float32')\n pref_msk = np.empty((ntrls, 1), dtype='int')\n\n # loop through trials and retrieve results\n for trl, tid in tqdm(enumerate(target_ids)):\n # get neurometric info of all neurons\n computed = tables.get_computed(tid)\n spksSE[trl] = computed['spikes']\n # evntSE[trl] = computed['events']\n # brstSE[trl] = computed['bursts']\n # snglSE[trl] = computed['singles']\n\n # population rates\n raw_data = tables.get_raw_data(tid)\n rateDE[trl] = raw_data['poprates_dec'] # 0: pref, 1: npref\n rateSE[trl] = raw_data['poprates_sen'] # 0: pref, 1: npref\n pref_msk[trl] = raw_data['pref_msk']\n\n # -------------------------------------\n # Choice probability and correlations\n # -------------------------------------\n # accuracy\n acc = pref_msk.sum() / ntrls\n\n # get active neurons, 100 per subpopulation\n actn = get_actn(spksSE, sub)\n\n # a calculation every 1, 5 or 10 ms?\n stepCP = 10\n auc1 = get_CPs(spksSE, np.logical_not(pref_msk), actn[0], downsampldt, step=stepCP)\n auc2 = get_CPs(spksSE, pref_msk.astype(bool), actn[1], downsampldt, step=stepCP)\n auc12 = np.concatenate((auc1, auc2), axis=0)\n\n stepCorr = 50\n corrsall, corrsii, corrsij = get_corr(spksSE, actn, downsampldt, step=stepCorr)\n\n # -------------------------------------\n # Plot figure 2\n # -------------------------------------\n fig, axs = plt.subplots(4, 1, figsize=(8, 12), sharex=True)\n\n fig.add_axes(axs[0])\n plt.plot(time, rateDE[:, 0, :].mean(axis=0), c='C3', lw=2, label='preferred')\n plt.plot(time, rateDE[:, 1, :].mean(axis=0), c='C0', lw=2, label='non-preferred')\n plt.axvline(x=stimon, color='gray', ls='dashed', lw=1.5)\n plt.axvline(x=stimoff, color='gray', ls='dashed', lw=1.5)\n plt.title('Integration circuit')\n plt.ylabel('Population rate (sp/s)') # , {'horizontalalignment': 'right'}\n plt.ylim(0, 50)\n # plt.legend(loc='center right', bbox_to_anchor=(1.22, 0.82))\n\n # sensory circuit\n fig.add_axes(axs[1])\n plt.plot(time, rateSE[:, 0, :].mean(axis=0), c='C3', lw=2, label='preferred')\n plt.plot(time, rateSE[:, 1, :].mean(axis=0), c='C0', lw=2, label='pon-preferred')\n plt.axvline(x=stimon, color='gray', ls='dashed', lw=2)\n plt.axvline(x=stimoff, color='gray', ls='dashed', lw=2)\n plt.title('Sensory circuit')\n plt.ylabel('Population rate (sp/s)')\n plt.ylim(0, 20) # 0, 15\n plt.legend(loc='center', bbox_to_anchor=(0.76, 0.91), ncol=2, fontsize='x-small')\n\n # CPs\n # clean to plot\n aucm = auc12.mean(axis=0)\n ymin = 0.45\n cleanaucm = np.ones(aucm.shape) * np.nan\n cleanaucm[aucm > ymin] = aucm[aucm > ymin]\n\n fig.add_axes(axs[2])\n plt.plot(downsampltime[::stepCP][1:], cleanaucm, 'k', lw=2)\n plt.axvline(x=stimon, color='gray', ls='dashed', lw=2)\n plt.axvline(x=stimoff, color='gray', ls='dashed', lw=2)\n plt.ylabel('Choice prob.')\n plt.ylim(ymin, ymin + 0.2) # ymin+0.2\n\n # correlations\n fig.add_axes(axs[3])\n plt.plot(downsampltime[::stepCorr][1:], np.nanmean(corrsall, axis=0), c='k', lw=2, label='EE')\n plt.plot(downsampltime[::stepCorr][1:], np.nanmean(corrsii, axis=0), c='C4', lw=2, label='EiEi')\n plt.plot(downsampltime[::stepCorr][1:], np.nanmean(corrsij, axis=0), c='C2', lw=2, label='EjEj')\n plt.axvline(x=stimon, color='gray', ls='dashed', lw=2)\n plt.axvline(x=stimoff, color='gray', ls='dashed', lw=2)\n plt.xlim(stimon - 0.5, stimoff + 0.5)\n plt.ylim(-0.2, 0.2) # -0.25, 0.25\n plt.xlabel('Time (s)')\n plt.ylabel('Noise correlations')\n plt.legend(loc='center', bbox_to_anchor=(0.77, 0.95), ncol=3, fontsize='x-small')\n\n # save figure\n #savepath = '/Users/PSR/Documents/WS19/MasterThesis/Experiments/run_hierarchical/'\n fig.savefig(load_path + '/' + thisexperiment + '/figure2.png')\n plt.close(fig)\n\n # -------------------------------------\n # Save analysis\n # -------------------------------------\n thisanalysisname = '/CPs-' + str(ntrls) + 'trls-' + str(targets) + '.pkl'\n\n # save variables\n with open(savepath + thisexperiment + thisanalysisname, 'wb') as f:\n pickle.dump([pref_msk,\n actn,\n auc12,\n [corrsall, corrsii, corrsij]], f)\n\n # TODO: plot burst probability and coherence levels\n # TODO: plot accuracy!\n\nif __name__ == '__main__':\n plot_fig2()\n # explore_data()\n","sub_path":"choiceProb.py","file_name":"choiceProb.py","file_ext":"py","file_size_in_byte":12568,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"509193079","text":"# Register your models here.\n#from django.contrib import admin\n#Activer l une ou l autre des 2 instruction ci dessous JPN 13/10/2015\nfrom django.contrib.gis import admin\n#from leaflet.admin import LeafletGeoAdmin\nfrom cnls.osmgeo_inline import OSMGeoTabularInline\n\n\nfrom .models import Organisme, Utilisateur, Action, Typeintervention, Cible, ActionLocalisation, ActionCible, ActionTypeintervention#, Status\n\n## PERMISSIONS (ajout nvx element de la liste) ##\n\nclass CibleAdmin(admin.ModelAdmin):\n\n def has_add_permission(self, request, obj=None):\n# return False\n return True\n\nclass TypeinterventionAdmin(admin.ModelAdmin):\n\n def has_add_permission(self, request, obj=None):\n return True\n\n\n## SECTIONS ##\n\n#class ActionCibleAdmin(admin.TabularInline):\nclass ActionCibleInline(admin.TabularInline):\n model = ActionCible\n extra = 2\n max_num = 3 # TODO Augmenter en production\n\n#class ActionTypeinterventionAdmin(admin.TabularInline):\nclass ActionTypeinterventionInline(admin.TabularInline):\n model = ActionTypeintervention \n extra = 2\n max_num = 3 # TODO Augmenter en production\n\n#class ActionLocalisationAdmin(admin.TabularInline):\n#class ActionLocalisationInline(admin.TabularInline):\nclass ActionLocalisationInline(OSMGeoTabularInline):\n model = ActionLocalisation \n extra = 1\n max_num = 2\n scale_text = False\n openlayers_url = '/static/OpenLayers.js'\n layerswitcher = False\n default_zoom = 3\n #'map_width': 200, 'map_height': 200, 'default_lon': -22, 'default_lat': 43, 'default_zoom': 10, 'layerswitcher': False, 'max_zoom': 15, 'min_zoom': 5, 'scale_text': False, 'debug' = True, }\n # cf. liste des paramètres modifiables https://github.com/django/django/blob/master/django/contrib/gis/admin/options.py\n\"\"\" \nclass ActionLocalisationAdmin(admin.OSMGeoAdmin):\n model = ActionLocalisation\n scale_text = False\n default_zoom = 3\n layerswitcher = False\n openlayers_url = '/static/OpenLayers.js'\n# map_width = 100\n# map_height = 100\n default_lon = -22\n default_lat = 43\n\"\"\"\n\n \n## L'Admin Principal compose des SECTIONS ##\n\nclass ActionAdmin(admin.ModelAdmin):\n#class ActionAdmin(admin.OSMGeoAdmin):\n model = Action\n radio_fields = {\"echelle_localisation\": admin.HORIZONTAL, \"devise\": admin.HORIZONTAL, \"avancement\": admin.HORIZONTAL}\n# inlines = [ActionLocalisationInline]#, ActionCibleInline, ActionTypeinterventionInline] # On a agrege les sections\n fieldsets = (\n (u'Informations générales', {\n 'fields': ('titre', 'organisme', 'typeintervention', 'cible', 'objectif', 'operateur',),\n 'classes': ('wide',),\n# 'description': 'texte',\n }),\n (u'Localisation', {\n 'fields': ('echelle_localisation',),\n 'classes': ('wide',),\n# 'description': 'texte',\n }),\n (u'Période', {\n 'fields': ('date_debut', 'date_fin', 'duree', 'avancement'),\n 'classes': ('wide',),\n# 'description': 'texte',\n }),\n (u'Objectifs', {\n 'fields': ('objectif', 'priorite_psn', 'resultat_cf_annee_ant',),\n 'classes': ('wide',),\n# 'description': 'texte',\n }),\n (u'Fonds', {\n 'fields': (('montant_prevu', 'montant_disponible',), 'devise', 'bailleurfond'),\n 'classes': ('wide',),\n# 'description': 'texte',\n }),\n (u'Contact', {\n 'fields': ('createur', 'contact', 'origine'),\n 'classes': ('wide',),\n# 'description': 'texte',\n }),\n\n (u'Informations avancées', {\n 'classes': ('wide',), #'collapse',),\n 'fields': ('description', 'commentaire'),\n }), \n )\n filter_horizontal = ('cible', 'typeintervention')\n \n\n# On enregistre les classes que l'on veut pouvoir modifier depuis l'interface d'administration, suivies éventuellement des modifications de l'interface par défaut\n\n#admin.site.register(mdgRegion, admin.OSMGeoAdmin)\nadmin.site.register(Organisme)\nadmin.site.register(Utilisateur)\nadmin.site.register(Action,ActionAdmin)\n#admin.site.register(ActionLocalisation, ActionLocalisationAdmin) #admin.OSMGeoAdmin) #, LeafletGeoAdmin)\nadmin.site.register(Typeintervention)\n\n#admin.site.register(Status)\nadmin.site.register(Cible)\n","sub_path":"cnls/admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":4393,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"654260841","text":"def process_ls(s):\n lines = []\n lines = s.splitlines()\n List = []\n ans = []\n for line in lines:\n if(line[0] != 'd'):\n List.append(line.split())\n for i in sorted(List, key = lambda x: (-int(x[4]), x[8:])): \n ans.append(' '.join(i[8:]))\n return ans\n","sub_path":"önn3/Python/ProcessIs.py","file_name":"ProcessIs.py","file_ext":"py","file_size_in_byte":293,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"57424174","text":"from django.urls import path\nfrom . import views\n\nurlpatterns = [\n path('', views.home, name='home'),\n path('make_data', views.make_data, name='make_data'),\n\n path('post/', views.PostRUDView.as_view(), name='post_rud'),\n path('post/new', views.PostCreateView.as_view(), name='post_new'),\n]\n","sub_path":"frameworks/python/django/08_rest_crud/myapp/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":310,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"601387493","text":"# -*- coding: utf-8 -*-\n\n# Define here the models for your spider middleware\n#\n# See documentation in:\n# https://docs.scrapy.org/en/latest/topics/spider-middleware.html\n\nfrom scrapy import signals\nfrom random import choice\nfrom scrapy.http import HtmlResponse\nfrom selenium import webdriver\nfrom scrapy.exceptions import NotConfigured\nfrom scrapy.exceptions import CloseSpider\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.chrome.options import Options\nfrom selenium.webdriver.support.ui import WebDriverWait\nfrom selenium.webdriver.support import expected_conditions as EC\n\nimport importlib\nimport os\nfrom pathlib import Path\n\nclass BabyscrapeSpiderMiddleware(object):\n # Not all methods need to be defined. If a method is not defined,\n # scrapy acts as if the spider middleware does not modify the\n # passed objects.\n\n @classmethod\n def from_crawler(cls, crawler):\n # This method is used by Scrapy to create your spiders.\n s = cls()\n crawler.signals.connect(s.spider_opened, signal=signals.spider_opened)\n return s\n\n def process_spider_input(self, response, spider):\n # Called for each response that goes through the spider\n # middleware and into the spider.\n\n # Should return None or raise an exception.\n return None\n\n def process_spider_output(self, response, result, spider):\n # Called with the results returned from the Spider, after\n # it has processed the response.\n\n # Must return an iterable of Request, dict or Item objects.\n for i in result:\n yield i\n\n def process_spider_exception(self, response, exception, spider):\n # Called when a spider or process_spider_input() method\n # (from other spider middleware) raises an exception.\n\n # Should return either None or an iterable of Request, dict\n # or Item objects.\n pass\n\n def process_start_requests(self, start_requests, spider):\n # Called with the start requests of the spider, and works\n # similarly to the process_spider_output() method, except\n # that it doesn’t have a response associated.\n\n # Must return only requests (not items).\n for r in start_requests:\n yield r\n\n def spider_opened(self, spider):\n spider.logger.info('Spider opened: %s' % spider.name)\n\n\nclass BabyscrapeDownloaderMiddleware(object):\n visited_pages =[]\n @classmethod\n def from_crawler(cls, crawler):\n # This method is used by Scrapy to create your spiders.\n s = cls()\n crawler.signals.connect(s.spider_opened, signal=signals.spider_opened)\n return s\n\n def process_request(self, request, spider):\n if 'robots.txt' not in request.url:\n if spider.name == 'hotel' and not spider.readmore_clicked:\n return self.readmore_click_response(request, spider)\n else:\n pass\n else:\n return None\n\n def readmore_click_response(self, request, spider):\n options = webdriver.ChromeOptions()\n options.add_argument('--headless')\n options.add_argument(\"enable-automation\")\n options.add_argument(\"--no-sandbox\")\n options.add_argument(\"--disable-infobars\")\n options.add_argument(\"--disable-dev-shm-usage\")\n options.add_argument(\"--disable-browser-side-navigation\")\n options.add_argument(\"--disable-gpu\")\n driver = webdriver.Chrome(options=options)\n #driver = webdriver.Chrome('chromedriver_win.exe', options=options)\n driver.get(request.url)\n readmore_css = 'span._3maEfNCR:nth-of-type(1)'\n attempts = 0\n\n while attempts < 2:\n try:\n readmore_present = EC.presence_of_element_located((By.CSS_SELECTOR, readmore_css))\n element = WebDriverWait(driver, 3).until(readmore_present)\n element.click()\n break\n except:\n attempts += 1\n print('Did not locate the \"Read more\" element, retrying: {}/2 '.format(attempts))\n if attempts == 2:\n driver.close()\n driver.quit()\n return HtmlResponse(url=request.url, body=\"Emergency\", encoding='utf-8', request=request)\n #raise CloseSpider(reason='Readmore Element not Found')\n\n body = driver.page_source\n drive_url = driver.current_url\n driver.close()\n driver.quit()\n spider.readmore_clicked = True\n return HtmlResponse(url=drive_url, body=body, encoding='utf-8', request=request)\n\n def process_response(self, request, response, spider):\n # Called with the response returned from the downloader.\n # Must either;\n # - return a Response object\n # - return a Request object\n # - or raise IgnoreRequest\n return response\n\n def process_exception(self, request, exception, spider):\n # Called when a download handler or a process_request()\n # (from other downloader middleware) raises an exception.\n\n # Must either:\n # - return None: continue processing this exception\n # - return a Response object: stops process_exception() chain\n # - return a Request object: stops process_exception() chain\n pass\n\n def spider_opened(self, spider):\n spider.logger.info('Spider opened: %s' % spider.name)\n\n\nclass RotateUserAgentMiddleware(object):\n \"\"\"Rotate user-agent for each request.\"\"\"\n def __init__(self, user_agents):\n self.enabled = False\n self.user_agents = user_agents\n\n @classmethod\n def from_crawler(cls, crawler):\n user_agents = crawler.settings.get('USER_AGENT_CHOICES', [])\n\n if not user_agents:\n raise NotConfigured(\"USER_AGENT_CHOICES not set or empty\")\n\n o = cls(user_agents)\n crawler.signals.connect(o.spider_opened, signal=signals.spider_opened)\n return o\n\n def spider_opened(self, spider):\n self.enabled = getattr(spider, 'rotate_user_agent', self.enabled)\n\n def process_request(self, request, spider):\n if not self.enabled or not self.user_agents:\n return\n request.headers['user-agent'] = choice(self.user_agents)\n print('User agent switched to : ' + request.headers['user-agent'].decode(\"utf-8\"))\n\n","sub_path":"babyscrape/middlewares.py","file_name":"middlewares.py","file_ext":"py","file_size_in_byte":6323,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"472186575","text":"from tensorflow.keras import optimizers\nfrom tensorflow.python.keras.layers import LSTM, RepeatVector, TimeDistributed, Dense\n\nfrom utils.algos.base_algos import SequentialMl\nfrom utils.dataset import Dataset, SequenceDataset\nfrom utils.preprocess import AnomalyDetection\nimport numpy as np\nimport pandas as pd\nimport os\nimport glob\nimport re\n\nclass Sequential(SequentialMl):\n\n def __init__(self, dataset, dataset_name, epochs=10, batch=64, lr=0.1):\n super().__init__(dataset, dataset_name=dataset_name, epochs=epochs, batch=batch, lr=lr)\n\n\n def compile_sequential(self):\n adam = optimizers.Adam(self.lr)\n self.sequential.compile(loss=\"mse\",\n optimizer=adam, metrics=['accuracy'])\n\n\nclass LSTMAutoencoder(Sequential):\n\n def __init__(self, dataset: Dataset, dataset_name, epochs=10, batch=64, lr=0.1):\n super().__init__(dataset=dataset, dataset_name=dataset_name, epochs=epochs, batch=batch, lr=lr)\n\n def init(self):\n super().init()\n timesteps, n_features = self.dataset.input_shape()\n\n self.sequential.add(\n LSTM(32, activation='relu', input_shape=(timesteps, n_features), return_sequences=True))\n self.sequential.add(LSTM(16, activation='relu', return_sequences=False))\n self.sequential.add(RepeatVector(timesteps))\n # Decoder\n self.sequential.add(LSTM(16, activation='relu', return_sequences=True))\n self.sequential.add(LSTM(32, activation='relu', return_sequences=True))\n self.sequential.add(TimeDistributed(Dense(n_features)))\n self.compile_sequential()\n\n\n\n\n def name(self):\n return \"LSTM Autoencoder\"\n\ndef clear_output_file():\n csvs = glob.glob('./../../results/**/*.csv', recursive=True)\n pngs = glob.glob('./../../results/**/*.png', recursive=True)\n for file in csvs + pngs:\n os.remove(file)\n\nif __name__ == '__main__':\n clear_output_file()\n names = []\n full_names = []\n regex = r\"Anomalies\\\\((.*?)\\.csv)\"\n for name in glob.glob('./../../data/Anomalies/*'):\n result = re.search(regex, name)\n if result:\n names.append(result.group(2))\n full_names.append(result.group(1))\n\n for i in range(len(full_names)):\n path = \"\"\n for step in range(3, 5):\n # Load and preprocess data\n preprocess_data = AnomalyDetection(data_path=\"./../../data/Anomalies/\" + full_names[i], seq_size=step, threshold_label=50)\n preprocess_data.preprocess()\n\n #Create an instant of SequenceDataset which already have some helpfull method. Example: load_test()..\n dataset = SequenceDataset(preprocess_data)\n\n #Create an instant of model which include method like: train(), predict()..\n model = LSTMAutoencoder(dataset=dataset, dataset_name=names[i]+\"/step\" + str(step), epochs=20, batch=64, lr=0.1 )\n preprocess_data.plot_data(data=preprocess_data.raw_data, title=names[i], path=model.pwd + model.image_path(names[i]))\n # model.train()\n model.load_model()\n history = model.load_history()\n model.plot_hist(history)\n print(\"Model predict: \\n\", model.predict())\n\n error_df = model.evaluation_metric(dataset.X_test, dataset.y_test)\n\n model.ROC(error_df=error_df)\n threshold_rt = model.precision_recal_curve(error_df)\n\n max_accuracy, thres = model.best_accuracy(error_df=error_df, threshold_rt=threshold_rt)\n\n model.confusion_matric(thres, error_df)\n model.reconstruction_error_for_2_class(thres, error_df)\n\n d = {\"DatasetName\": names[i], \"Step\": step, \"Threshold\": thres, \"Accuracy\": max_accuracy}\n df = pd.DataFrame([d])\n path = \"./../../results/\" + names[i] + \"/accuracies.csv\"\n if os.path.exists(path):\n df.to_csv(path, mode=\"a\", header=False)\n else:\n df.to_csv(path)\n df = pd.read_csv(path)\n df = df.sort_values(by=['Accuracy'], ascending=False)\n df.to_csv(path)\n break\n\n\n\n\n","sub_path":"utils/algos/sequential.py","file_name":"sequential.py","file_ext":"py","file_size_in_byte":4114,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"51"} +{"seq_id":"151362409","text":"\n\n#calss header\nclass _COT():\n\tdef __init__(self,): \n\t\tself.name = \"COT\"\n\t\tself.definitions = [u'a small bed for a baby or young child with high bars around the sides so that the child cannot fall out', u'a light bed that can be folded so that it can be easily carried and stored', u'a narrow bed']\n\n\t\tself.parents = []\n\t\tself.childen = []\n\t\tself.properties = []\n\t\tself.jsondata = {}\n\n\n\t\tself.specie = 'nouns'\n\n\n\tdef run(self, obj1 = [], obj2 = []):\n\t\treturn self.jsondata\n","sub_path":"xai/brain/wordbase/nouns/_cot.py","file_name":"_cot.py","file_ext":"py","file_size_in_byte":473,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"178835946","text":"import unittest\nimport webnotes\nimport copy\n\nfrom webnotes.model.doclist import DocList\nfrom webnotes.model.doc import Document\nfrom webnotes.model.code import get_obj\nfrom webnotes.utils import flt\n\nsql = webnotes.conn.sql\n\n\nclass TestAH(unittest.TestCase):\n\tdef setUp(self):\n\t\twebnotes.conn.begin()\n\n\tdef tearDown(self):\n\t\twebnotes.conn.rollback()\n\n\tdef testInsert(self):\n#\t\ttd = TestData()\n\t\td = DocList()\n\n\t\tcount_before = flt(sql(\"select count(*) from tab\"+_doctype)[0][0])\n\t\tif docok:\n\t\t\tfor i in docok:\n\t\t\t\td.doc = i\n\t\t\t\td.children = None\n\t\t\t\td.doc.fields['__islocal']=1\n\t\t\t\td.save(1)\n\t\tcount_after = flt(sql(\"select count(*) from tab\"+_doctype)[0][0])\n\t\tself.assertTrue(count_before+len(docok)==count_after)\n\n\tdef testFailAssert(self):\n#\t\ttd = TestData()\n\t\tif docnotok:\n\t\t\twith self.assertRaises(Exception) as context:\n\t\t\t\td = DocList()\n\t\t\t\td.doc = docnotok[0]\n\t\t\t\td.children = None\n\t\t\t\td.doc.fields['__islocal']=1\n\t\t\t\td.save(1)\n\n# Test Data\n\n#class TestData():\n\ntabOK =\t[\n\t\t{'account_name': 'acc1', 'parent_account': 'Indirect Expenses - TC', 'group_or_ledger': 'Ledger', 'is_pl_account': 'Yes', 'debit_or_credit': 'Debit', 'company': 'Test Company'},\n\t\t{'account_name': 'acc2', 'parent_account': 'Indirect Expenses - TC', 'group_or_ledger': 'Ledger', 'is_pl_account': 'Yes', 'debit_or_credit': 'Debit', 'company': 'Test Company'},\n\t\t{'account_name': 'acc3', 'parent_account': 'Indirect Expenses - TC', 'group_or_ledger': 'Ledger', 'is_pl_account': 'Yes', 'debit_or_credit': 'Debit', 'company': 'Test Company'}\n\t]\n\ntabNotOK = \t[\n\t\t]\n\n_doctype = 'Account'\n\nfor i in tabOK: i['doctype']=_doctype\nfor i in tabNotOK: i['doctype']=_doctype\n\ndocok = [Document(fielddata=r) for r in tabOK]\ndocnotok = [Document(fielddata=r) for r in tabNotOK]\n\n\n","sub_path":"erpnext/accounts/doctype/account/test_account.py","file_name":"test_account.py","file_ext":"py","file_size_in_byte":1748,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"347101115","text":"\nfrom . import plot_expr_graph as peg\nfrom .data_graph import FactorExprNode\n\ndef ggplot(data=None, aes=None):\n p = peg.BokehPlot(data)\n if aes is not None:\n return p + aes\n else:\n return p\n\ndef aes(x=None, y=None, **kw):\n return peg.Aes(**kw)\n\ndef geom_point(position=None, aes=None):\n \"\"\" **position** is one of the position_* adjustment\n functions like jitter, dodge, etc.\n \"\"\"\n g = peg.GeomPoint()\n if aes:\n g.aes = aes\n if position is not None:\n g.position = position\n return g\n \ndef geom_line(aes=None):\n if aes:\n g = peg.GeomLine(aes)\n else:\n g = peg.GeomLine()\n return g\n\ndef facet_grid(factor_expr):\n node = peg.FacetGrid()\n node.factor_expr = FactorExprNode.from_string_expr(factor_expr)\n return node\n\ndef facet_wrap(factor_expr):\n node = peg.FacetWrap()\n node.factor_expr = FactorExprNode.from_string_expr(factor_expr)\n return node\n\n\n\n\n\n","sub_path":"bokeh/attic/bokeh_ggplot.py","file_name":"bokeh_ggplot.py","file_ext":"py","file_size_in_byte":953,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"199398783","text":"#!/usr/bin/python\n\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport csv\nfrom mpl_toolkits.mplot3d import Axes3D\nimport time\nimport linecache\n\ndef train_svm():\n\t'''\n\tPicking Veg(1004) and Facade (1400) as the two classes\n\tLet Veg have label 1\n\tLet Facade have label -1\n\t'''\n\n\tlam = 0.495\n\tnum_lines = open('training_set.node_features').read().count('\\n')\n\twts = 2 * np.random.uniform(0, 1, 10) - 1\n\tc = 0\n\tsse = []\n\titr = []\n\tfor i in range(num_lines):\n\t\tline = linecache.getline('training_set.node_features',i+1)\n\t\tvals = [float(j) for j in line.split()]\n\t\tfeature = vals[5:15]\n\t\tf_vec = np.asarray(feature,dtype=np.float32)\n\t\tnode_id = int(vals[4])\n\t\tif node_id == 1004 or node_id == 1400:\n\t\t\tc = c + 1\n\t\t\talpha_t = 0.01/np.sqrt(float(c))\n\t\t\tmodel_value = np.dot(wts, f_vec)\n\t\t\tif node_id == 1004: y = 1\n\t\t\telse : y = -1\n\t\t\teps = 0.0\n\t\t\tif model_value >= eps and node_id == 1004: # true positive\n\t\t\t\twts = wts - 2*alpha_t*lam*wts\n\t\t\telif model_value >= eps and node_id == 1400: # false positive\n\t\t\t\twts = wts - 2*alpha_t*lam*wts + alpha_t*y*f_vec\n\t\t\telif model_value < eps and node_id == 1004: # false negative\n\t\t\t\twts = wts - 2*alpha_t*lam*wts + alpha_t*y*f_vec\n\t\t\telif model_value < eps and node_id == 1400: # true negative\n\t\t\t\twts = wts - 2*alpha_t*lam*wts\n\n\t\t\tif c == 1:\n\t\t\t\tdiff = 0\n\t\t\t\titr.append(c)\n\t\t\t\tnum_lines = open('test_set.node_features').read().count('\\n')\n\t\t\t\tfor i in range(num_lines):\n\t\t\t\t\tline = linecache.getline('test_set.node_features',i+1)\n\t\t\t\t\tvals = [float(j) for j in line.split()]\n\t\t\t\t\tfeature = vals[5:15]\n\t\t\t\t\tf_vec = np.asarray(feature,dtype=np.float32)\n\t\t\t\t\tnode_id = int(vals[4])\n\t\t\t\t\tif node_id == 1004 or node_id == 1400:\n\t\t\t\t\t\tresult = np.dot(wts,f_vec)\n\t\t\t\t\t\tif node_id == 1004: y = 1\n\t\t\t\t\t\telse: y = -1\n\t\t\t\t\t\tdiff = diff + (y - result)**2\n\t\t\t\tsse.append(diff)\n\n\t\t\tif c%1000 == 0:\n\t\t\t\tdiff = 0\n\t\t\t\titr.append(c)\n\t\t\t\tnum_lines = open('test_set.node_features').read().count('\\n')\n\t\t\t\tfor i in range(num_lines):\n\t\t\t\t\tline = linecache.getline('test_set.node_features',i+1)\n\t\t\t\t\tvals = [float(j) for j in line.split()]\n\t\t\t\t\tfeature = vals[5:15]\n\t\t\t\t\tf_vec = np.asarray(feature,dtype=np.float32)\n\t\t\t\t\tnode_id = int(vals[4])\n\t\t\t\t\tif node_id == 1004 or node_id == 1400:\n\t\t\t\t\t\tresult = np.dot(wts,f_vec)\n\t\t\t\t\t\tif node_id == 1004: y = 1\n\t\t\t\t\t\telse: y = -1\n\t\t\t\t\t\tdiff = diff + (y - result)**2\n\t\t\t\tsse.append(diff)\n\n\tplt.plot(itr, sse, 'r--')\n\tplt.xlabel('Iteration Number')\n\tplt.ylabel('Sum Squared Error')\n\tplt.title('Error Convergence plot')\n\tplt.grid(True)\n\tplt.show()\n\n\nif __name__ == '__main__':\n\ttrain_svm()\n","sub_path":"lab2_svm_convergence_plots.py","file_name":"lab2_svm_convergence_plots.py","file_ext":"py","file_size_in_byte":2549,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"385780532","text":"__author__ = 'wbk3zd'\n\nimport os\nimport shutil\nimport threading\nimport time\nfrom copy import deepcopy\n\nimport zmq\n\ntry:\n from Helpers.Encodings import *\n from Helpers.Logging.OneDirLogger import EventLogger\nexcept ImportError:\n from FinalOneDir.Helpers.Encodings import *\n from FinalOneDir.Logging.OneDirLogger import EventLogger\n\n\nclass SyncResponder():\n def __init__(self, msg_identifier):\n #Components\n self.context = zmq.Context()\n self.logger = EventLogger()\n\n #Attributes\n self.msg_identifier = msg_identifier\n self.config = None\n self.listen_flag = threading.Event()\n self.listen_flag.clear()\n\n #Networking\n self.internal_request_lock = threading.RLock()\n self.internal_request_socket = self.context.socket(zmq.PUSH)\n self.server_sync_throw_lock = threading.RLock()\n self.server_sync_throw_socket = self.context.socket(zmq.SUB)\n\n \"\"\"\n Public methods\n \"\"\"\n def initialize(self, config):\n \"\"\"\n Sets up configuration values and connects sockets\n \"\"\"\n self.config = config\n\n #Initialize components\n logfile = \".\" + SLASH + \"responder.log\"\n self.logger.init_session(logfile)\n\n #Socket connections\n self.internal_request_socket.connect(\"tcp://localhost:\" + self.config[\"INTERNAL_REQUEST_PORT\"])\n self.logger.log(\"INFO\",\"Connecting responder to internal client controller over tcp port \" + self.config[\"INTERNAL_REQUEST_PORT\"] + \"...\")\n\n #Subscribe to sync throws for configured username\n self.server_sync_throw_socket.setsockopt(zmq.SUBSCRIBE, self.config[\"USERNAME\"].encode('ascii', 'replace'))\n self.server_sync_throw_socket.connect(\"tcp://\" + self.config[\"SERVER_ADDR\"] + \":\" + self.config[\"SERVER_SYNC_THROW_PORT\"])\n self.logger.log(\"INFO\",\"Subscribed to sync directives at tcp://\" + self.config[\"SERVER_ADDR\"] + \":\" + self.config[\"SERVER_SYNC_THROW_PORT\"] + \" for user \" + self.config[\"USERNAME\"] + \"...\")\n\n def start(self):\n \"\"\"\n Spawns a new thread with target _listen_ to listen for sync\n directives published by server.\n \"\"\"\n if self.listen_flag.is_set():\n return\n else:\n self.listen_flag.set()\n threading.Thread(target=self._listen_).start()\n self.logger.log(\"INFO\",\"Responder is listening for sync directives at tcp://\" + self.config[\"SERVER_ADDR\"] + \":\" + self.config[\"SERVER_SYNC_THROW_PORT\"] + \" for user \" + self.config[\"USERNAME\"] + \"...\")\n\n def pause(self):\n \"\"\"\n Causes any thread in _listen_ to exit gracefully\n \"\"\"\n self.logger.log(\"INFO\",\"Responder has paused. No longer listening for sync directives\")\n self.listen_flag.clear()\n\n def stop(self):\n \"\"\"\n Like pause, but allows for additional cleanup.\n Causes any thread in _listen_ to exit gracefully\n \"\"\"\n self.logger.log(\"INFO\",\"Responder is being killed. Going down permanently.\")\n self.listen_flag.clear()\n\n \"\"\"\n Protected methods\n \"\"\"\n def _listen_(self):\n \"\"\"\n Run in a separate thread. Listens for sync directives published by server\n for the subscribed username. Dispatches caught directives to a new thread\n for processing.\n \"\"\"\n while(self.listen_flag.is_set()):\n #Receive and dispatch until the end of time (or until listen_flag is cleared)\n msg = self.server_sync_throw_socket.recv_multipart()\n msg = decode(msg)\n\n #Remove topic from message where topic is username\n msg.remove(msg[0])\n\n #Dispatch command\n threading.Thread(target=self._dispatch_, args=(msg,)).start()\n\n #Strip away file contents before logging message\n msg_clone = deepcopy(msg)\n if msg_clone[0] == self.msg_identifier[\"FILESYNC\"]:\n msg_clone[-1] = \"\"\n\n #Log\n self.logger.log(\"INFO\",\"Sync Directive received: \" + str(msg_clone))\n\n\n def _dispatch_(self, msg):\n \"\"\"\n Entry point for all threads spawned from a message received in\n _listen_. Identifies the sync directive type and calls the appropriate\n internal method to handle.\n \"\"\"\n #Check to see if message was empty\n if not msg[0]:\n self.logger.log(\"ERROR\",\"Empty message received from server\")\n return\n\n #Send internal request to controller to stop daemon monitoring of directory, we are about to write\n out = [self.msg_identifier[\"STOP_MONITORING\"], str(threading.current_thread().ident)]\n with self.internal_request_lock:\n self.internal_request_socket.send_multipart(encode(out))\n\n #Give controller and daemon a moment to get their affairs in order\n time.sleep(1)\n\n #Dispatch\n if msg[0] == self.msg_identifier[\"FILESYNC\"]:\n self._on_sync_(msg)\n elif msg[0] == self.msg_identifier[\"MKDIR\"]:\n self._on_mkdir_(msg)\n elif msg[0] == self.msg_identifier[\"DELETE\"]:\n self._on_remove_(msg)\n elif msg[0] == self.msg_identifier[\"MOVE\"]:\n self._on_move_(msg)\n elif msg[0] == self.msg_identifier[\"KILL\"]:\n msg = [self.msg_identifier[\"KILL\"]]\n self.internal_request_socket.send_multipart(encode(msg)) #Notify controller of impending doom\n else:\n self.logger.log(\"ERROR\",\"Unrecognized message. Closing without handle: \" + str(msg))\n\n #Cleanup (resumes daemon)\n self._on_finish_()\n\n def _on_sync_(self, msg):\n \"\"\"\n Handles file sync events by writing sent contents\n to disk.\n \"\"\"\n #Get absolute path by appending path base\n dest_path = self.config[\"PATH_BASE\"] + msg[1]\n\n #Create the target directory if it does not exist\n if not os.path.exists(os.path.dirname(dest_path)):\n os.makedirs(os.path.dirname(dest_path))\n\n #Log and write\n self.logger.log(\"INFO\",\"Updating file at \" + dest_path)\n with open(dest_path, 'wb') as user_file:\n user_file.write(msg[2])\n\n def _on_mkdir_(self, msg):\n \"\"\"\n Creates a directory at the specified relative path if it does not exist\n \"\"\"\n #Create final destination by appending path base\n dest_path = self.config[\"PATH_BASE\"] + msg[1]\n\n #Create a directory, or not, the choice is yours\n if(os.path.isdir(dest_path)):\n self.logger.log(\"INFO\",\"Directory already exists, ignoring make command: \" + str(msg))\n else:\n self.logger.log(\"INFO\",\"Creating directory at \" + dest_path)\n os.makedirs(dest_path)\n\n def _on_remove_(self, msg):\n \"\"\"\n Deletes the filesystem object at the specified target, recursively if relevant\n \"\"\"\n dest_path = self.config[\"PATH_BASE\"] + msg[1]\n\n #If object does not exist, all done\n if not os.path.exists(dest_path):\n self.logger.log(\"WARNING\", dest_path + \" does not exist. Can't remove: \" + str(msg))\n #Otherwise, remove as appropriate\n elif(os.path.isdir(dest_path)):\n self.logger.log(\"INFO\",\"Removing entire file tree at \" + dest_path)\n shutil.rmtree(dest_path)\n else:\n self.logger.log(\"INFO\",\"Removing file at\" + dest_path)\n os.remove(dest_path)\n\n def _on_move_(self, msg):\n \"\"\"\n Called anytime a file system object is moved\n \"\"\"\n\n #Get absolute paths\n src_path = self.config[\"PATH_BASE\"] + msg[1]\n dest_path = self.config[\"PATH_BASE\"] + msg[2]\n\n #If source doesn't exist, throw an error\n if not os.path.exists(src_path):\n self.logger.log(\"ERROR\",\"File system object at \" + dest_path + \" does not exist. Cannot move: \" + str(msg))\n #Otherwise, handle as appropriate\n elif(os.path.isdir(src_path)):\n self.logger.log(\"INFO\",\"Moving directory at \" + src_path + \" to \" + dest_path)\n shutil.copytree(src_path, dest_path)\n shutil.rmtree(src_path)\n else:\n self.logger.log(\"INFO\",\"Moving file at\" + src_path + \"to\" + dest_path)\n shutil.copy2(src_path, dest_path)\n os.remove(src_path)\n\n def _on_finish_(self):\n \"\"\"\n Called by every thread after it has finished its dispatch task.\n Asks controller to restore daemon operation. Other cleanup can go here\n as well.\n \"\"\"\n with self.server_sync_throw_lock:\n msg = [self.msg_identifier[\"START_MONITORING\"],str(threading.current_thread().ident)]\n self.internal_request_socket.send_multipart(encode(msg))","sub_path":"FinalOneDir/Controllers/Client/ClientSyncResponder.py","file_name":"ClientSyncResponder.py","file_ext":"py","file_size_in_byte":8888,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"491957033","text":"from flask import Flask\nfrom .isite import Site_Main_Flask_Obj\nfrom logging import getLogger\nimport sys\n\nlogger = getLogger(__name__)\nlogger.setLevel('DEBUG')\n\nprint('hello', file=sys.stderr)\n\napp = Flask(__name__)\napp.register_blueprint(Site_Main_Flask_Obj)\n\n\nport=5000\nprint('hello2', file=sys.stderr)\nlogger.debug('starting iserver on port %d' % port)\nprint('starting iserver on port %d' % port, file=sys.stderr)\napp.run(host='0.0.0.0', port=port)\n","sub_path":"iserver/iserver.py","file_name":"iserver.py","file_ext":"py","file_size_in_byte":451,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"52"} +{"seq_id":"487927849","text":"from django_datatables_view.base_datatable_view import BaseDatatableView\nfrom django.utils.html import escape\nfrom register_disease_data.models import MedicineType\nfrom django.db.models import Q\nfrom django.http import HttpResponse\nfrom django.shortcuts import get_object_or_404\nimport traceback\n\n\nclass medicineType_datatable(BaseDatatableView):\n order_columns = ['name']\n columns = ['name','id']\n def get_initial_queryset(self):\n return MedicineType.objects.order_by(\"id\")\n def filter_queryset(self, qs):\n search = self.request.GET.get('search[value]', None)\n if search:\n qs = qs.filter(name__contains=search)\n filter_customer = self.request.GET.get('MedicineType', None)\n\n if filter_customer:\n customer_parts = filter_customer.split(' ')\n qs_params = None\n for part in customer_parts:\n q = Q(customer_firstname__contains=part)|Q(customer_lastname__contains=part)\n qs_params = qs_params | q if qs_params else q\n qs = qs.filter(qs_params)\n return qs\n\n def prepare_results(self, qs):\n\n json_data = []\n for item in qs:\n json_data.append([\n escape(item.name),\n item.id\n ])\n return json_data\n\n\n\ndef browse_medicineType_name(request):\n if request.method == 'POST':\n key = request.POST.get('key')\n if not key:\n return HttpResponse(\"\")\n filtered_data=MedicineType.objects.filter(name__contains=key)[0:10]\n html='';\n for data in filtered_data:\n html=html.__add__('