code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
from database.db0 import db0, ConstDB from database.db3 import db3, ConstDB3 from utils.errors import KeyDuplicateError, ReadOnlyDeny from utils.utils import eui_64_to_48, eui_48_to_64 from binascii import hexlify from enum import Enum import enum from userver.frequency_plan import FrequencyPlan from userver.object.as...
[ "utils.utils.eui_64_to_48", "database.db0.db0.hgetall", "database.db_sql.db_sql.session.delete", "database.db0.db0.exists", "binascii.hexlify", "utils.errors.KeyDuplicateError", "database.db0.db0.keys", "database.db0.db0.pipeline", "sqlalchemy.ForeignKey", "sqlalchemy.BINARY", "eviltransform.wgs...
[((2130, 2171), 'eviltransform.wgs2gcj', 'eviltransform.wgs2gcj', (['self.lat', 'self.lng'], {}), '(self.lat, self.lng)\n', (2151, 2171), False, 'import eviltransform\n'), ((3764, 3773), 'sqlalchemy.BINARY', 'BINARY', (['(8)'], {}), '(8)\n', (3770, 3773), False, 'from sqlalchemy import Column, String, BINARY\n'), ((381...
import logging import pytz from urllib.parse import urlencode, urljoin, urlparse, parse_qs import requests from django.conf import settings from ol_openedx_canvas_integration.constants import DEFAULT_ASSIGNMENT_POINTS log = logging.getLogger(__name__) class CanvasClient: def __init__(self, canvas_course_id): ...
[ "urllib.parse.urlencode", "requests.Session", "requests.utils.parse_header_links", "logging.getLogger", "urllib.parse.parse_qs", "urllib.parse.urlparse" ]
[((228, 255), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (245, 255), False, 'import logging\n'), ((564, 582), 'requests.Session', 'requests.Session', ([], {}), '()\n', (580, 582), False, 'import requests\n'), ((1064, 1077), 'urllib.parse.urlparse', 'urlparse', (['url'], {}), '(url)\n'...
import builtins import collections import contextlib import glob import io import os import string from typing import Dict, Any, Tuple, List, Collection, Optional import attrs import pandas as pd from databutler.pat import astlib from databutler.pat.analysis.type_analysis.mypy_types import SerializedMypyType DF_TYPE...
[ "databutler.pat.astlib.with_deep_replacements", "io.StringIO", "os.path.dirname", "databutler.pat.astlib.walk", "collections.defaultdict", "databutler.pat.astlib.iter_body_stmts", "databutler.pat.astlib.is_constant", "databutler.pat.astlib.get_constant_value", "databutler.pat.astlib.parse_expr", "...
[((887, 921), 'attrs.define', 'attrs.define', ([], {'eq': '(False)', 'repr': '(False)'}), '(eq=False, repr=False)\n', (899, 921), False, 'import attrs\n'), ((1248, 1273), 'attrs.field', 'attrs.field', ([], {'factory': 'dict'}), '(factory=dict)\n', (1259, 1273), False, 'import attrs\n'), ((2735, 2780), 'databutler.pat.a...
## @package bot # Module that sets up the Geometrize Twitter bot. # # Invoke this script to run the bot i.e. "python bot.py". import sys import config import dependency_locator import geometrize import geometrize_bot import launch_text import on_status_event import tweepy # Print welcome text. launch_text.print_la...
[ "geometrize.test_geometrize", "tweepy.API", "config.validate_credentials", "on_status_event.on_on_demand_status_event", "dependency_locator.get_geometrize_executable_path", "launch_text.print_launch_text", "sys.exit", "on_status_event.on_account_watcher_status_event", "dependency_locator.geometrize_...
[((300, 331), 'launch_text.print_launch_text', 'launch_text.print_launch_text', ([], {}), '()\n', (329, 331), False, 'import launch_text\n'), ((1034, 1062), 'geometrize.test_geometrize', 'geometrize.test_geometrize', ([], {}), '()\n', (1060, 1062), False, 'import geometrize\n'), ((1306, 1382), 'tweepy.OAuthHandler', 't...
from django.views.generic.base import View from django.template.response import HttpResponse from share.oaipmh.repository import OAIRepository class OAIPMHView(View): CONTENT_TYPE = 'text/xml' def get(self, request): return self.oai_response(**request.GET) def post(self, request): retur...
[ "share.oaipmh.repository.OAIRepository", "django.template.response.HttpResponse" ]
[((416, 431), 'share.oaipmh.repository.OAIRepository', 'OAIRepository', ([], {}), '()\n', (429, 431), False, 'from share.oaipmh.repository import OAIRepository\n'), ((509, 558), 'django.template.response.HttpResponse', 'HttpResponse', (['xml'], {'content_type': 'self.CONTENT_TYPE'}), '(xml, content_type=self.CONTENT_TY...
import pandas as pd import psycopg2 #from sqlalchemy import create_engine psql_credeintal = { 'database': 'wode', 'user': 'wode', 'password': '***', 'host': '192.168.3.11', 'port': '5432' } con = psycopg2.connect(**psql_credeintal) def get_winrate(user_id): query = "SEL...
[ "pandas.read_sql_query", "psycopg2.connect" ]
[((241, 276), 'psycopg2.connect', 'psycopg2.connect', ([], {}), '(**psql_credeintal)\n', (257, 276), False, 'import psycopg2\n'), ((425, 454), 'pandas.read_sql_query', 'pd.read_sql_query', (['query', 'con'], {}), '(query, con)\n', (442, 454), True, 'import pandas as pd\n')]
from django.core.files.base import ContentFile import base64 import uuid def get_report_image(data): _, image_binary = data.split(';base64') decoded_image = base64.b64decode(image_binary) img_name = str(uuid.uuid4())[:10] + '.png' data = ContentFile(decoded_image, name=img_name) return data
[ "uuid.uuid4", "base64.b64decode", "django.core.files.base.ContentFile" ]
[((168, 198), 'base64.b64decode', 'base64.b64decode', (['image_binary'], {}), '(image_binary)\n', (184, 198), False, 'import base64\n'), ((257, 298), 'django.core.files.base.ContentFile', 'ContentFile', (['decoded_image'], {'name': 'img_name'}), '(decoded_image, name=img_name)\n', (268, 298), False, 'from django.core.f...
#!/usr/bin/python # ########################################################################## # # Social-Engineer Toolkit Persistence Service # # Right now this is a pretty lame attempt at a service but will grow over time. The text file it reads in from isn't # really a good idea, but it...
[ "win32serviceutil.HandleCommandLine", "subprocess.Popen", "os.path.abspath", "sys.getwindowsversion", "servicemanager.StartServiceCtrlDispatcher", "win32event.SetEvent", "time.sleep", "win32event.WaitForSingleObject", "servicemanager.PrepareToHostSingle", "win32event.CreateEvent", "servicemanage...
[((1068, 1122), 'win32serviceutil.ServiceFramework.__init__', 'win32serviceutil.ServiceFramework.__init__', (['self', 'args'], {}), '(self, args)\n', (1110, 1122), False, 'import win32serviceutil\n'), ((1148, 1188), 'win32event.CreateEvent', 'win32event.CreateEvent', (['None', '(0)', '(0)', 'None'], {}), '(None, 0, 0, ...
import requests from model.json_check import * from model.input_data import * # Запрос на получение настроек всех объектов CAM def test_GetV1AllCamerasCode200(): data = "success" response = requests.get(url="http://"+slave_ip+":"+restPort+"/api/v1/cameras/", auth=auth) user_resp_code = "200" assert st...
[ "requests.get" ]
[((200, 291), 'requests.get', 'requests.get', ([], {'url': "('http://' + slave_ip + ':' + restPort + '/api/v1/cameras/')", 'auth': 'auth'}), "(url='http://' + slave_ip + ':' + restPort + '/api/v1/cameras/',\n auth=auth)\n", (212, 291), False, 'import requests\n'), ((528, 623), 'requests.get', 'requests.get', ([], {'...
import random from impiccato_disegno import d_impiccato, logo from parole_impiccato import lista_parole scelta_parola = random.choice(lista_parole) print(logo) game_over = False energia = len(d_impiccato)-1 campo_gioco = [] for i in scelta_parola: campo_gioco += '_' while not game_over: indovina = input('In...
[ "random.choice" ]
[((121, 148), 'random.choice', 'random.choice', (['lista_parole'], {}), '(lista_parole)\n', (134, 148), False, 'import random\n')]
import unittest from pyoxigraph import * XSD_STRING = NamedNode("http://www.w3.org/2001/XMLSchema#string") XSD_INTEGER = NamedNode("http://www.w3.org/2001/XMLSchema#integer") RDF_LANG_STRING = NamedNode("http://www.w3.org/1999/02/22-rdf-syntax-ns#langString") class TestNamedNode(unittest.TestCase): def test_cons...
[ "unittest.main" ]
[((7347, 7362), 'unittest.main', 'unittest.main', ([], {}), '()\n', (7360, 7362), False, 'import unittest\n')]
import getpass # importing getpass dpass = getpass.getpass(prompt = "Enter the password: ") # by default it shows 'password' print(f"The entered password is {dpass}")
[ "getpass.getpass" ]
[((46, 92), 'getpass.getpass', 'getpass.getpass', ([], {'prompt': '"""Enter the password: """'}), "(prompt='Enter the password: ')\n", (61, 92), False, 'import getpass\n')]
from django.db import models import datetime class Region(models.Model): name = models.CharField(max_length=200) class University(models.Model): address = models.CharField(max_length=255) affilation_name = models.CharField(max_length=255) author_count = models.IntegerField(default=0) city = mode...
[ "django.db.models.TextField", "django.db.models.CharField", "django.db.models.ForeignKey", "datetime.date", "django.db.models.FloatField", "django.db.models.BooleanField", "django.db.models.IntegerField", "django.db.models.DateField" ]
[((86, 118), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(200)'}), '(max_length=200)\n', (102, 118), False, 'from django.db import models\n'), ((167, 199), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(255)'}), '(max_length=255)\n', (183, 199), False, 'from django.db...
import enum import datetime from chat_bots.models import Sender from slack_bot.bot import SlackBot from telegram_bot.bot import TelegramBot class MessengerType(enum.Enum): Telegram = 'Telegram' Slack = 'Slack' class SenderBots: new_employee_channel_id = None new_employee_chat_bot = None access_r...
[ "chat_bots.models.Sender.objects.first", "datetime.date.today", "datetime.timedelta", "slack_bot.bot.SlackBot", "chat_bots.models.Sender.objects.filter", "telegram_bot.bot.TelegramBot", "datetime.datetime.combine", "datetime.datetime.now" ]
[((2521, 2543), 'chat_bots.models.Sender.objects.first', 'Sender.objects.first', ([], {}), '()\n', (2541, 2543), False, 'from chat_bots.models import Sender\n'), ((438, 460), 'chat_bots.models.Sender.objects.first', 'Sender.objects.first', ([], {}), '()\n', (458, 460), False, 'from chat_bots.models import Sender\n'), (...
import math import random __BASE_SEED = 1 def constant_thresholds(graph, value): """Sets a constant threshold for every node of the graph.""" # Store threshold assignment in a dictionary thresholds = dict() # Add a constant attribute to each node for node in graph.Nodes(): thresholds[n...
[ "random.seed", "math.floor" ]
[((1132, 1163), 'random.seed', 'random.seed', (['(__BASE_SEED + seed)'], {}), '(__BASE_SEED + seed)\n', (1143, 1163), False, 'import random\n'), ((750, 779), 'math.floor', 'math.floor', (['(degree * fraction)'], {}), '(degree * fraction)\n', (760, 779), False, 'import math\n')]
""" Minimal character-level Vanilla RNN model. Written b_y <NAME> (@karpathy) BSD License """ import numpy as np import unicodedata import string import codecs # data I/O data = codecs.open('data/potter.txt', 'r', encoding='utf8', errors='ignore').read() fake = codecs.open('data/output.txt', 'w', encoding='utf8') char...
[ "unicodedata.normalize", "numpy.zeros_like", "codecs.open", "numpy.copy", "numpy.random.randn", "numpy.log", "unicodedata.category", "numpy.zeros", "numpy.clip", "numpy.exp", "numpy.dot", "numpy.sqrt" ]
[((263, 315), 'codecs.open', 'codecs.open', (['"""data/output.txt"""', '"""w"""'], {'encoding': '"""utf8"""'}), "('data/output.txt', 'w', encoding='utf8')\n", (274, 315), False, 'import codecs\n'), ((1301, 1327), 'numpy.zeros', 'np.zeros', (['(hidden_size, 1)'], {}), '((hidden_size, 1))\n', (1309, 1327), True, 'import ...
import sys input = sys.stdin.readline import heapq as hq # input t = int(input()) for _ in range(t): chapter = int(input()) pages = list(map(int, input().split())) # process ''' 각 장이 섞여도 됨. >> 힙에서 제일 작은 거 두 개 합치고 다시 넣음. ''' sol = 0 hq.heapify(pages) for _ in range(chapter - 1): cost = hq.heappop(pages) + hq....
[ "heapq.heappush", "heapq.heapify", "heapq.heappop" ]
[((240, 257), 'heapq.heapify', 'hq.heapify', (['pages'], {}), '(pages)\n', (250, 257), True, 'import heapq as hq\n'), ((351, 375), 'heapq.heappush', 'hq.heappush', (['pages', 'cost'], {}), '(pages, cost)\n', (362, 375), True, 'import heapq as hq\n'), ((297, 314), 'heapq.heappop', 'hq.heappop', (['pages'], {}), '(pages)...
"""Optimization result.""" import warnings from collections import Counter from copy import deepcopy from typing import Sequence, Union import numpy as np import pandas as pd from ..objective import History from ..problem import Problem from ..util import assign_clusters, delete_nan_inf OptimizationResult = Union['...
[ "pandas.DataFrame", "copy.deepcopy", "numpy.argmax", "collections.Counter", "numpy.isnan", "numpy.argmin", "numpy.max", "numpy.mean", "numpy.array", "numpy.min", "warnings.warn" ]
[((5219, 5238), 'copy.deepcopy', 'deepcopy', (['self.list'], {}), '(self.list)\n', (5227, 5238), False, 'from copy import deepcopy\n'), ((9657, 9674), 'pandas.DataFrame', 'pd.DataFrame', (['lst'], {}), '(lst)\n', (9669, 9674), True, 'import pandas as pd\n'), ((10229, 10359), 'warnings.warn', 'warnings.warn', (['"""get_...
import datetime import pytest from unittest.mock import patch from django.core.exceptions import ValidationError from freezegun import freeze_time from phonenumber_field.phonenumber import PhoneNumber from wagtail.core.models import Page from wagtail_extensions.blocks import ( DepartmentBlock, ImagesBlock, LinkBl...
[ "wagtail.core.models.Page", "wagtail_extensions.blocks.OpeningTimeBlock", "wagtail_extensions.blocks.OpeningTimesBlock.time_keyfunc", "wagtail_extensions.blocks.OpeningTimesBlock", "datetime.time", "wagtail_extensions.blocks.OpeningTimesBlock.group_times", "wagtail.core.models.Page.objects.get", "unit...
[((3823, 3848), 'freezegun.freeze_time', 'freeze_time', (['"""2017-01-01"""'], {}), "('2017-01-01')\n", (3834, 3848), False, 'from freezegun import freeze_time\n'), ((4437, 4462), 'freezegun.freeze_time', 'freeze_time', (['"""2017-01-01"""'], {}), "('2017-01-01')\n", (4448, 4462), False, 'from freezegun import freeze_t...
""" BlueSky simulation client class """ # TODO: Need to re-add the tests for string parsing/units from the old API tests import os from typing import List from semver import VersionInfo from .bluesky_aircraft_controls import BlueSkyAircraftControls from .bluesky_simulator_controls import BlueSkySimulatorControls from...
[ "bluebird.sim_client.bluesky.bluesky_client.BlueSkyClient", "os.getenv", "semver.VersionInfo.parse" ]
[((548, 575), 'os.getenv', 'os.getenv', (['"""BS_MIN_VERSION"""'], {}), "('BS_MIN_VERSION')\n", (557, 575), False, 'import os\n'), ((695, 729), 'semver.VersionInfo.parse', 'VersionInfo.parse', (['_BS_MIN_VERSION'], {}), '(_BS_MIN_VERSION)\n', (712, 729), False, 'from semver import VersionInfo\n'), ((1623, 1638), 'blueb...
# -*- coding: utf-8 -*- """ Created on Mon Jul 27 08:49:21 2020 @author: rafae """ import pandas as pd import numpy as np def gerarbanco(): banco = pd.read_stata("Microcefalia MS analysis 20160609.dta") circ = pd.read_csv("circumference.csv",sep=";",index_col='sem') banco.index = range(...
[ "pandas.read_csv", "pandas.DataFrame", "pandas.read_stata" ]
[((170, 224), 'pandas.read_stata', 'pd.read_stata', (['"""Microcefalia MS analysis 20160609.dta"""'], {}), "('Microcefalia MS analysis 20160609.dta')\n", (183, 224), True, 'import pandas as pd\n'), ((237, 295), 'pandas.read_csv', 'pd.read_csv', (['"""circumference.csv"""'], {'sep': '""";"""', 'index_col': '"""sem"""'})...
from redis import StrictRedis class SentinelClient(object): def __init__(self, url, db=None, name='default', health_check_interval=30): """create a redis client. Args: url: redis server url. db: redis database, default 0. name: client name, default 'default'. ...
[ "redis.StrictRedis.from_url" ]
[((452, 578), 'redis.StrictRedis.from_url', 'StrictRedis.from_url', ([], {'url': 'url', 'db': 'db', 'client_name': 'name', 'health_check_interval': 'health_check_interval', 'decode_responses': '(True)'}), '(url=url, db=db, client_name=name,\n health_check_interval=health_check_interval, decode_responses=True)\n', (4...
from django.utils.translation import ugettext_lazy as _ MALE = "m" FEMALE = "f" UNSPECIFIED = "-" GENDERS = {(MALE, _("male")), (FEMALE, _("female")), (UNSPECIFIED, _("don't want to answer"))}
[ "django.utils.translation.ugettext_lazy" ]
[((117, 126), 'django.utils.translation.ugettext_lazy', '_', (['"""male"""'], {}), "('male')\n", (118, 126), True, 'from django.utils.translation import ugettext_lazy as _\n'), ((149, 160), 'django.utils.translation.ugettext_lazy', '_', (['"""female"""'], {}), "('female')\n", (150, 160), True, 'from django.utils.transl...
import os import logging import unittest from Uefi.Capsule.CatGenerator import * #must run from build env or set PYTHONPATH env variable to point to the PythonLibrary folder class CatGeneratorTest(unittest.TestCase): def test_win10_OS(self): o = CatGenerator("x64", "win10") self.assertEqual(o.OperatingSyst...
[ "os.path.join" ]
[((1463, 1515), 'os.path.join', 'os.path.join', (['"""c:"""', '"""test"""', '"""badpath"""', '"""inf2cat.exe"""'], {}), "('c:', 'test', 'badpath', 'inf2cat.exe')\n", (1475, 1515), False, 'import os\n')]
import numpy as np import matplotlib.pyplot as plt from typing import Tuple, Union, TypeVar, Iterable, Dict from goa import problems T = TypeVar("T") def plot_population( problem: problems.BaseProblem, X: Union[T, Iterable[T]], ax: plt.Axes = None, c: str = "darkblue", linestyle: str = ":", ...
[ "numpy.subtract", "numpy.asarray", "matplotlib.pyplot.figure", "numpy.cos", "typing.TypeVar" ]
[((141, 153), 'typing.TypeVar', 'TypeVar', (['"""T"""'], {}), "('T')\n", (148, 153), False, 'from typing import Tuple, Union, TypeVar, Iterable, Dict\n'), ((608, 635), 'matplotlib.pyplot.figure', 'plt.figure', ([], {'figsize': 'figsize'}), '(figsize=figsize)\n', (618, 635), True, 'import matplotlib.pyplot as plt\n'), (...
# Standard library import os from pathlib import Path # Third party from cryptography.fernet import Fernet from dotenv import load_dotenv # Constants CRYPTO_KEY_ENV_VAR = 'CRYPTO_KEY' ENV_VAR_EXIST = f'The {CRYPTO_KEY_ENV_VAR} environment variable already exist. Cannot continue as the crypto key may still be in use....
[ "dotenv.load_dotenv", "os.environ.get", "cryptography.fernet.Fernet.generate_key", "pathlib.Path.home" ]
[((408, 445), 'dotenv.load_dotenv', 'load_dotenv', ([], {'dotenv_path': 'ENV_VAR_PATH'}), '(dotenv_path=ENV_VAR_PATH)\n', (419, 445), False, 'from dotenv import load_dotenv\n'), ((337, 348), 'pathlib.Path.home', 'Path.home', ([], {}), '()\n', (346, 348), False, 'from pathlib import Path\n'), ((487, 521), 'os.environ.ge...
import csv import config import awb infilename = config.datafolder+'wikidata/wdlid_ElhId.csv' resourceitem = "Q19" #Q19: Elhuyar #positem = "Q8" # Q7: substantibo, Q8: aditza with open(infilename, encoding="utf-8") as csvfile: sourcedict = csv.DictReader(csvfile) lex_elh = {} for row in sourcedict: lex_elh[row['...
[ "csv.DictReader", "awb.updateclaim", "awb.setqualifier" ]
[((243, 266), 'csv.DictReader', 'csv.DictReader', (['csvfile'], {}), '(csvfile)\n', (257, 266), False, 'import csv\n'), ((516, 560), 'awb.updateclaim', 'awb.updateclaim', (['awbid', '"""P1"""', 'wdid', '"""string"""'], {}), "(awbid, 'P1', wdid, 'string')\n", (531, 560), False, 'import awb\n'), ((571, 644), 'awb.setqual...
import numpy as np from casim.calculations import word_entropy def test_word_entropy(): test_arr = np.array([1, 0, 0, 1, 1, 0, 1, 0]) assert np.round(word_entropy(test_arr, 3), decimals=1) == 2.5
[ "casim.calculations.word_entropy", "numpy.array" ]
[((105, 139), 'numpy.array', 'np.array', (['[1, 0, 0, 1, 1, 0, 1, 0]'], {}), '([1, 0, 0, 1, 1, 0, 1, 0])\n', (113, 139), True, 'import numpy as np\n'), ((161, 186), 'casim.calculations.word_entropy', 'word_entropy', (['test_arr', '(3)'], {}), '(test_arr, 3)\n', (173, 186), False, 'from casim.calculations import word_en...
import numpy as np import matplotlib.pyplot as plt import glob from sys import argv from os.path import exists as file_exists methods = ['drude', 'c36'] mol1, mol2 = str(argv[1]), str(argv[2]) sysname = mol1+'_'+mol2 def blockavg(x,nblocks=30): lblock = int(len(x)/nblocks) m = [] for i in range(nblocks): ...
[ "matplotlib.pyplot.savefig", "numpy.concatenate", "numpy.std", "matplotlib.pyplot.close", "matplotlib.pyplot.legend", "os.path.exists", "matplotlib.pyplot.figure", "numpy.mean", "numpy.array", "numpy.loadtxt", "glob.glob", "matplotlib.pyplot.ylabel", "matplotlib.pyplot.xlabel", "matplotlib...
[((420, 431), 'numpy.array', 'np.array', (['m'], {}), '(m)\n', (428, 431), True, 'import numpy as np\n'), ((1330, 1344), 'numpy.array', 'np.array', (['osmp'], {}), '(osmp)\n', (1338, 1344), True, 'import numpy as np\n'), ((1374, 1386), 'matplotlib.pyplot.figure', 'plt.figure', ([], {}), '()\n', (1384, 1386), True, 'imp...
# Copyright 2022 Google LLC # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # https://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, sof...
[ "tensorflow.image.random_crop", "tensorflow.data.TFRecordDataset", "tensorflow.io.decode_image", "tensorflow.io.parse_single_example", "tensorflow.concat", "tensorflow.constant", "absl.logging.warning", "tensorflow.data.experimental.sample_from_datasets", "tensorflow.io.FixedLenFeature", "tensorfl...
[((2435, 2482), 'tensorflow.io.parse_single_example', 'tf.io.parse_single_example', (['sample', 'feature_map'], {}), '(sample, feature_map)\n', (2461, 2482), True, 'import tensorflow as tf\n'), ((4124, 4157), 'tensorflow.concat', 'tf.concat', (['image_to_crop'], {'axis': '(-1)'}), '(image_to_crop, axis=-1)\n', (4133, 4...
import matplotlib.pyplot as plt import numpy as np #returns the binding energy predicted by nuclear liquid drop model def BE_liquidDrop(N,Z): #N=num of neutrons, Z=num of protons #num of nucleons A = N+Z #physical constants (from Alex's notes, in MeV) a1 = 15.49 a2 = 17.23 a3 = 0.697 a4 = 22.6 #nuclear li...
[ "numpy.zeros", "matplotlib.pyplot.imshow", "matplotlib.pyplot.show" ]
[((1336, 1392), 'matplotlib.pyplot.imshow', 'plt.imshow', (['mat'], {'interpolation': '"""nearest"""', 'origin': '"""lower"""'}), "(mat, interpolation='nearest', origin='lower')\n", (1346, 1392), True, 'import matplotlib.pyplot as plt\n'), ((1412, 1422), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (1420, 14...
import gym from gym.spaces import Box import numpy as np import torch as T import rl.environments from rl.data.buffer import TrajBuffer, ReplayBuffer # from rl.data.buffer import TrajBuffer, ReplayBufferNP # def make_env(env_id, seed, idx, capture_video, run_name): # def thunk(): # print('in thunk') # ...
[ "gym.make", "rl.data.buffer.ReplayBuffer", "torch.Tensor", "rl.data.buffer.TrajBuffer", "torch.no_grad" ]
[((1634, 1648), 'gym.make', 'gym.make', (['name'], {}), '(name)\n', (1642, 1648), False, 'import gym\n'), ((1890, 1904), 'gym.make', 'gym.make', (['name'], {}), '(name)\n', (1898, 1904), False, 'import gym\n'), ((2726, 2817), 'rl.data.buffer.TrajBuffer', 'TrajBuffer', (['self.obs_dim', 'self.act_dim', 'horizon', 'num_t...
#!/usr/bin/env python #Update your .twconfig file on this same directory #with your own api keys and secrets #Get them signing up at https://apps.twitter.com #Install required modules with #'pip install -r requirements.txt' import configparser import os import sys import json import twitter def jdefault(o): r...
[ "sys.stdout.write", "configparser.RawConfigParser", "json.dumps" ]
[((477, 520), 'sys.stdout.write', 'sys.stdout.write', (['"""reading config file... """'], {}), "('reading config file... ')\n", (493, 520), False, 'import sys\n'), ((532, 562), 'configparser.RawConfigParser', 'configparser.RawConfigParser', ([], {}), '()\n', (560, 562), False, 'import configparser\n'), ((675, 716), 'sy...
import tensorflow as tf def neural_mf(user_input, item_input, y_, num_users, num_items, embed_partitioner=None): embed_dim = 8 layers = [64, 32, 16, 8] learning_rate = 0.01 with tf.compat.v1.variable_scope('nmf', dtype=tf.float32): with tf.device('/cpu:0'): User_Embedding = tf.comp...
[ "tensorflow.nn.relu", "tensorflow.compat.v1.variable_scope", "tensorflow.nn.embedding_lookup", "tensorflow.device", "tensorflow.reduce_mean", "tensorflow.concat", "tensorflow.nn.sigmoid_cross_entropy_with_logits", "tensorflow.compat.v1.train.GradientDescentOptimizer", "tensorflow.multiply", "tenso...
[((196, 248), 'tensorflow.compat.v1.variable_scope', 'tf.compat.v1.variable_scope', (['"""nmf"""'], {'dtype': 'tf.float32'}), "('nmf', dtype=tf.float32)\n", (223, 248), True, 'import tensorflow as tf\n'), ((263, 282), 'tensorflow.device', 'tf.device', (['"""/cpu:0"""'], {}), "('/cpu:0')\n", (272, 282), True, 'import te...
import pytest from anndata import AnnData from pandas.testing import assert_frame_equal import numpy as np from squidpy.gr import moran, ripley_k, co_occurrence MORAN_K = "moranI" def test_ripley_k(adata: AnnData): """Check ripley score and shape.""" ripley_k(adata, cluster_key="leiden") # assert rip...
[ "pandas.testing.assert_frame_equal", "squidpy.gr.co_occurrence", "numpy.testing.assert_allclose", "pytest.raises", "squidpy.gr.ripley_k", "numpy.random.choice", "numpy.array_equal", "pytest.mark.parametrize", "squidpy.gr.moran" ]
[((1825, 1866), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""n_jobs"""', '[1, 2]'], {}), "('n_jobs', [1, 2])\n", (1848, 1866), False, 'import pytest\n'), ((3614, 3679), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (["('n_jobs', 'n_splits')", '[(1, 2), (2, 2)]'], {}), "(('n_jobs', 'n_splits'), [(...
""" Tests for pyhap.characteristic """ import uuid from unittest import mock import pytest import pyhap.characteristic as characteristic from pyhap.characteristic import Characteristic PROPERTIES = { "Format": characteristic.HAP_FORMAT.INT, "Permissions": [characteristic.HAP_PERMISSIONS.READ] } def get_char...
[ "uuid.uuid1", "pytest.raises", "unittest.mock.Mock" ]
[((1509, 1520), 'unittest.mock.Mock', 'mock.Mock', ([], {}), '()\n', (1518, 1520), False, 'from unittest import mock\n'), ((2733, 2744), 'unittest.mock.Mock', 'mock.Mock', ([], {}), '()\n', (2742, 2744), False, 'from unittest import mock\n'), ((1351, 1376), 'pytest.raises', 'pytest.raises', (['ValueError'], {}), '(Valu...
from random_survival_forest import RandomSurvivalForest, concordance_index from lifelines import datasets from sklearn.model_selection import train_test_split import time rossi = datasets.load_rossi() # Attention: duration column must be index 0, event column index 1 in y y = rossi.loc[:, ["arrest", "week"]] X = rossi...
[ "random_survival_forest.concordance_index", "sklearn.model_selection.train_test_split", "time.time", "random_survival_forest.RandomSurvivalForest", "lifelines.datasets.load_rossi" ]
[((180, 201), 'lifelines.datasets.load_rossi', 'datasets.load_rossi', ([], {}), '()\n', (199, 201), False, 'from lifelines import datasets\n'), ((377, 432), 'sklearn.model_selection.train_test_split', 'train_test_split', (['X', 'y'], {'test_size': '(0.25)', 'random_state': '(10)'}), '(X, y, test_size=0.25, random_state...
import gym import numpy as np import torch import torch.optim as optim from utils_main import make_env, save_files from neural_network import ActorCritic from ppo_method import ppo from common.multiprocessing_env import SubprocVecEnv from itertools import count use_cuda = torch.cuda.is_available() device = torch.devic...
[ "numpy.stack", "ppo_method.ppo", "utils_main.save_files", "gym.make", "common.multiprocessing_env.SubprocVecEnv", "numpy.save", "neural_network.ActorCritic", "torch.FloatTensor", "torch.cat", "itertools.count", "numpy.hstack", "torch.cuda.is_available", "torch.device", "utils_main.make_env...
[((274, 299), 'torch.cuda.is_available', 'torch.cuda.is_available', ([], {}), '()\n', (297, 299), False, 'import torch\n'), ((309, 352), 'torch.device', 'torch.device', (["('cuda' if use_cuda else 'cpu')"], {}), "('cuda' if use_cuda else 'cpu')\n", (321, 352), False, 'import torch\n'), ((456, 475), 'common.multiprocess...
import tensorflow as tf from .config import cfg class YoloHead(tf.keras.layers.Layer): def __init__(self, grid_size, classes, strides, anchors, xyscale, i): super().__init__() self.grid_size = grid_size self.classes = classes self.strides = strides self.anchors = anchors ...
[ "tensorflow.range", "tensorflow.math.reduce_max", "tensorflow.reshape", "tensorflow.concat", "tensorflow.constant", "tensorflow.stack", "tensorflow.cast", "tensorflow.shape", "tensorflow.exp", "tensorflow.boolean_mask", "tensorflow.split", "tensorflow.sigmoid", "tensorflow.expand_dims" ]
[((470, 564), 'tensorflow.reshape', 'tf.reshape', (['feature_map', '(batch_size, self.grid_size, self.grid_size, 3, 5 + self.classes)'], {}), '(feature_map, (batch_size, self.grid_size, self.grid_size, 3, 5 +\n self.classes))\n', (480, 564), True, 'import tensorflow as tf\n'), ((654, 709), 'tensorflow.split', 'tf.sp...
""" Author: <NAME> File: menu.py A general class for creating menus Parameters: pos - (x,y) position for the top-left corner of the menu dims - (width, height) pixels of the menu commands - list of dictionaries specifying the button attributes padding - (horizontal, vertical) padding between border an...
[ "pygame.Surface", "polybius.graphics.components.Button", "pygame.font.SysFont", "polybius.graphics.basics.drawable.Drawable.__init__", "polybius.graphics.utils.window.Window.__init__" ]
[((1042, 1092), 'polybius.graphics.basics.drawable.Drawable.__init__', 'Drawable.__init__', (['self', '""""""', 'pos'], {'worldBound': '(False)'}), "(self, '', pos, worldBound=False)\n", (1059, 1092), False, 'from polybius.graphics.basics.drawable import Drawable\n'), ((1101, 1122), 'polybius.graphics.utils.window.Wind...
from flask import flash, redirect, render_template, request, url_for from flask_login import login_required from sqlalchemy import and_ from . import unitOfMeasurements from . forms import UnitOfMeasurementForm from .. import db from .. decorators import adminRequired from .. models import UnitOfMeasurement modelName ...
[ "sqlalchemy.and_", "flask.url_for", "flask.flash", "flask.render_template" ]
[((521, 625), 'flask.render_template', 'render_template', (['"""unitOfMeasurements/unitOfMeasurements.html"""'], {'unitOfMeasurements': 'unitOfMeasurements'}), "('unitOfMeasurements/unitOfMeasurements.html',\n unitOfMeasurements=unitOfMeasurements)\n", (536, 625), False, 'from flask import flash, redirect, render_te...
from unittest import mock import pytest from bs4 import BeautifulSoup from conf import wsgi @pytest.mark.django_db @pytest.mark.parametrize('script_name,prefix', (('/sso', '/sso/accounts/'), ('', '/accounts/'))) def test_set_script_name(rf, script_name, prefix): environ = rf._base_environ( PATH_INFO='/a...
[ "conf.wsgi.application", "pytest.mark.parametrize", "bs4.BeautifulSoup" ]
[((120, 219), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""script_name,prefix"""', "(('/sso', '/sso/accounts/'), ('', '/accounts/'))"], {}), "('script_name,prefix', (('/sso', '/sso/accounts/'),\n ('', '/accounts/')))\n", (143, 219), False, 'import pytest\n'), ((487, 546), 'conf.wsgi.application', 'wsg...
# Generated by Django 2.1.2 on 2018-11-04 16:16 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('game_catalog', '0010_auto_20181104_1036'), ] operations = [ migrations.RemoveField( model_name=...
[ "django.db.migrations.RemoveField" ]
[((273, 339), 'django.db.migrations.RemoveField', 'migrations.RemoveField', ([], {'model_name': '"""publishedgame"""', 'name': '"""edition"""'}), "(model_name='publishedgame', name='edition')\n", (295, 339), False, 'from django.db import migrations, models\n'), ((384, 454), 'django.db.migrations.RemoveField', 'migratio...
import asyncio from bashbot.commands import Command from bashbot.commands.syntax import SyntaxBuilder from bashbot.session_manager import SessionManager class RenameCommand(Command): def __init__(self): super().__init__() self.name = "Rename terminal" self.aliases = [".rename"] s...
[ "bashbot.session_manager.SessionManager.get_session", "asyncio.get_event_loop", "bashbot.commands.syntax.SyntaxBuilder" ]
[((627, 670), 'bashbot.session_manager.SessionManager.get_session', 'SessionManager.get_session', (['message.channel'], {}), '(message.channel)\n', (653, 670), False, 'from bashbot.session_manager import SessionManager\n'), ((1027, 1051), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (1049, 1051...
import os, sys if __name__ == "__main__": negative_identity_weight = [0, 0, 0] positive_identity_weight = [0, 0, 0] negative_classes_weight = [0, 0, 0] positive_classes_weight = [0, 0, 0] negative_policy_weight = [0, 0, 0] positive_policy_weight = [0, 0, 0] for i in range(len(negative_iden...
[ "os.system" ]
[((421, 451), 'os.system', 'os.system', (['"""rm -rf ./resource"""'], {}), "('rm -rf ./resource')\n", (430, 451), False, 'import os, sys\n'), ((460, 529), 'os.system', 'os.system', (['f"""python ./SST/image_classification_policy.py --log {log}"""'], {}), "(f'python ./SST/image_classification_policy.py --log {log}')\n",...
#!/usr/bin/env python3 """ Limits: - can't read brackets in brackets - avoid using more then one bracket per line. Normally it should work, but you should use it carefull - problems by differncing ttl/domaine, if domain[:-1].isdigit()""" VERSION = 1.1 DOMAIN_MAX_LENGHT = 255 CLASSES = ["IN", "HS", "CH",...
[ "os.path.isfile", "sqlite3.connect", "time.strftime", "os.path.basename" ]
[((9084, 9108), 'os.path.basename', 'path.basename', (['self.file'], {}), '(self.file)\n', (9097, 9108), True, 'import os.path as path\n'), ((12963, 12981), 'time.strftime', 'strftime', (['"""%Y%m%d"""'], {}), "('%Y%m%d')\n", (12971, 12981), False, 'from time import strftime\n'), ((13868, 13885), 'sqlite3.connect', 'sq...
import pytorch_lightning as pl import torch import typing as th import torchmetrics from .. import utils as ebad_utils class AnoTrainer(pl.LightningModule): def __init__( self, model_cls: th.Union[str, torch.nn.Module], input_shape: th.Union[th.Tuple[int], th.List[int]], ...
[ "torchmetrics.AUROC", "torch.rand_like", "torch.cat", "torch.randn_like" ]
[((646, 692), 'torchmetrics.AUROC', 'torchmetrics.AUROC', ([], {'num_classes': '(2)', 'pos_label': '(1)'}), '(num_classes=2, pos_label=1)\n', (664, 692), False, 'import torchmetrics\n'), ((719, 765), 'torchmetrics.AUROC', 'torchmetrics.AUROC', ([], {'num_classes': '(2)', 'pos_label': '(1)'}), '(num_classes=2, pos_label...
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved """ Given an URL in string, make a request, fetch its content, and parse using BeautifulSoup. """ import time import requests import logging import urllib.parse as urlparse from bs4 import BeautifulSoup class URLContentFetc...
[ "requests.get", "bs4.BeautifulSoup", "urllib.parse.urlparse", "time.time" ]
[((1719, 1746), 'urllib.parse.urlparse', 'urlparse.urlparse', (['self.url'], {}), '(self.url)\n', (1736, 1746), True, 'import urllib.parse as urlparse\n'), ((2022, 2033), 'time.time', 'time.time', ([], {}), '()\n', (2031, 2033), False, 'import time\n'), ((2050, 2150), 'requests.get', 'requests.get', (['self.url'], {'he...
from utils.models import base, product # # Bronze Age # BronzePottery = base.ManufactureBuilding( name="🏺🏠 Гончарня", products=[product.dish, product.jug, product.amphora], create_price=[340, 490], create_time_sec=1800, manpower=108 ) BronzePlantation = base.ManufactureBuilding( name="🍇🏠 ...
[ "utils.models.base.ManufactureBuilding" ]
[((74, 243), 'utils.models.base.ManufactureBuilding', 'base.ManufactureBuilding', ([], {'name': '"""🏺🏠 Гончарня"""', 'products': '[product.dish, product.jug, product.amphora]', 'create_price': '[340, 490]', 'create_time_sec': '(1800)', 'manpower': '(108)'}), "(name='🏺🏠 Гончарня', products=[product.dish,\n produc...
# Generated by Django 2.1.3 on 2018-11-09 00:49 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('workertasks', '0002_auto_20181109_0046'), ] operations = [ migrations.AlterField( model_name='assignment', name='beg...
[ "django.db.models.DateTimeField", "django.db.models.IntegerField" ]
[((347, 378), 'django.db.models.DateTimeField', 'models.DateTimeField', ([], {'null': '(True)'}), '(null=True)\n', (367, 378), False, 'from django.db import migrations, models\n'), ((507, 538), 'django.db.models.DateTimeField', 'models.DateTimeField', ([], {'null': '(True)'}), '(null=True)\n', (527, 538), False, 'from ...
import cotk from cotk._utils.file_utils import get_resource_file_path from cotk.dataloader.dataloader import * from collections import Counter import numpy as np from itertools import chain class Score(DataField): def get_next(self, dataset): r"""read text and returns the next label(integer). Note that it...
[ "itertools.chain.from_iterable", "collections.Counter", "numpy.max", "numpy.array" ]
[((15676, 15731), 'numpy.array', 'np.array', (["[self.data[key]['score'][i] for i in indexes]"], {}), "([self.data[key]['score'][i] for i in indexes])\n", (15684, 15731), True, 'import numpy as np\n'), ((15048, 15074), 'numpy.max', 'np.max', (["res['post_length']"], {}), "(res['post_length'])\n", (15054, 15074), True, ...
import string import numpy as np import pandas as pd import pytest from plotnine import (ggplot, aes, geom_point, geom_jitter, geom_bar, geom_col, geom_boxplot, geom_text, geom_rect, after_stat, position_dodge, position_dodge2, position_jitter, positio...
[ "plotnine.geom_boxplot", "plotnine.position_dodge2", "numpy.ones", "numpy.arange", "plotnine.position_dodge", "plotnine.aes", "plotnine.position_stack", "pandas.DataFrame", "plotnine.position_jitter", "plotnine.after_stat", "numpy.random.RandomState", "plotnine.position_nudge", "pytest.raise...
[((520, 553), 'numpy.random.RandomState', 'np.random.RandomState', (['(1234567890)'], {}), '(1234567890)\n', (541, 553), True, 'import numpy as np\n'), ((560, 612), 'pandas.DataFrame', 'pd.DataFrame', (["{'x': [1, 2, 1, 2], 'y': [1, 1, 2, 2]}"], {}), "({'x': [1, 2, 1, 2], 'y': [1, 1, 2, 2]})\n", (572, 612), True, 'impo...
from setuptools import setup, find_packages NAME = 'golr-schema-generator' DESCRIPTION = 'GOlr Schema Generator' URL = 'https://github.com/deepakunni3/golr-schema-generator' AUTHOR = '<NAME>' EMAIL = '<EMAIL>' REQUIRES_PYTHON = '>=3.7.0' VERSION = '0.0.1' LICENSE = 'BSD3' REQUIRED = [ 'PyYAML>=5.3' ] EXTRAS = { ...
[ "setuptools.find_packages" ]
[((593, 608), 'setuptools.find_packages', 'find_packages', ([], {}), '()\n', (606, 608), False, 'from setuptools import setup, find_packages\n')]
import json import click from tokenizer import get_grouped_tokens, TokenName NULL = "null" #from click documentation to support alias command class AliasedGroup(click.Group): def get_command(self, ctx, cmd_name): rv = click.Group.get_command(self, ctx, cmd_name) if rv is not None: ret...
[ "click.Group.get_command", "json.load", "click.option", "click.echo", "json.dumps", "click.command", "click.File", "click.ClickException", "tokenizer.get_grouped_tokens" ]
[((645, 660), 'click.command', 'click.command', ([], {}), '()\n', (658, 660), False, 'import click\n'), ((662, 785), 'click.option', 'click.option', (['"""--expression"""', '"""-e"""'], {'type': 'click.STRING', 'help': '"""jq style expression to search in the json"""', 'required': '(True)'}), "('--expression', '-e', ty...
import os BASE_DIR = os.path.abspath(os.path.dirname(__file__)) MAIN_DB_URL = os.path.join(BASE_DIR, 'bucketlist.sqlite') TEST_DB_URL = os.path.join(BASE_DIR, 'test.sqlite') class BaseConfig(object): ''' The class holds base config for each environment ''' SECRET_KEY = os.getenv('SECRET_KEY', 'This s...
[ "os.path.dirname", "os.path.join", "os.getenv" ]
[((79, 122), 'os.path.join', 'os.path.join', (['BASE_DIR', '"""bucketlist.sqlite"""'], {}), "(BASE_DIR, 'bucketlist.sqlite')\n", (91, 122), False, 'import os\n'), ((137, 174), 'os.path.join', 'os.path.join', (['BASE_DIR', '"""test.sqlite"""'], {}), "(BASE_DIR, 'test.sqlite')\n", (149, 174), False, 'import os\n'), ((38,...
import datetime import os from typing import List, Optional import unittest from parentopticon.db import test_utilities from parentopticon.db.model import ColumnInteger, ColumnText, Model class ModelTests(test_utilities.DBTestCase): "Test all of our logic around the model class." class MyTable(Model): COLUMNS = {...
[ "parentopticon.db.model.ColumnText", "parentopticon.db.model.ColumnInteger" ]
[((330, 381), 'parentopticon.db.model.ColumnInteger', 'ColumnInteger', ([], {'autoincrement': '(True)', 'primary_key': '(True)'}), '(autoincrement=True, primary_key=True)\n', (343, 381), False, 'from parentopticon.db.model import ColumnInteger, ColumnText, Model\n'), ((395, 410), 'parentopticon.db.model.ColumnInteger',...
# -*- coding: utf-8 -*- """The UTMPX binary file event formatter.""" from plaso.formatters import interface from plaso.formatters import manager from plaso.lib import errors class UtmpxSessionFormatter(interface.ConditionalEventFormatter): """Formatter for an UTMPX session event.""" DATA_TYPE = u'mac:utmpx:even...
[ "plaso.formatters.manager.FormattersManager.RegisterFormatter" ]
[((1969, 2035), 'plaso.formatters.manager.FormattersManager.RegisterFormatter', 'manager.FormattersManager.RegisterFormatter', (['UtmpxSessionFormatter'], {}), '(UtmpxSessionFormatter)\n', (2012, 2035), False, 'from plaso.formatters import manager\n')]
import pytest import numpy as np from sklearn.datasets import load_iris from numpy.testing import assert_array_equal from numpy.testing import assert_allclose from .._kdclassifier import KDClassifierRF from .._RBFSamplerORF import RBFSamplerORF from .._RBFSamplerSORF import RBFSamplerSORF @pytest.fixture def data()...
[ "sklearn.datasets.load_iris" ]
[((333, 359), 'sklearn.datasets.load_iris', 'load_iris', ([], {'return_X_y': '(True)'}), '(return_X_y=True)\n', (342, 359), False, 'from sklearn.datasets import load_iris\n')]
from itertools import cycle, islice import re distances = {} for line in open('input.txt'): m = re.match(r'(\w+) can fly (\d+) km/s for (\d+) seconds, ' r'but then must rest for (\d+) seconds.', line) reindeer, speed, time, rest = m.groups() distances[reindeer] = cycle([int(speed)] * int(...
[ "re.match", "itertools.islice" ]
[((102, 215), 're.match', 're.match', (['"""(\\\\w+) can fly (\\\\d+) km/s for (\\\\d+) seconds, but then must rest for (\\\\d+) seconds."""', 'line'], {}), "(\n '(\\\\w+) can fly (\\\\d+) km/s for (\\\\d+) seconds, but then must rest for (\\\\d+) seconds.'\n , line)\n", (110, 215), False, 'import re\n'), ((360, ...
import datetime as dt import pandas import random import smtplib from astroid import Pass MY_EMAIL = "your_email" PASSWORD = "<PASSWORD>" today = dt.datetime.now() today_tuple = (today.month, today.day) data = pandas.read_csv("birthdays.csv") birthday_dict = {(data_row.month, data_row.day): data_row for (index, da...
[ "pandas.read_csv", "smtplib.SMTP_SSL", "datetime.datetime.now", "random.randint" ]
[((149, 166), 'datetime.datetime.now', 'dt.datetime.now', ([], {}), '()\n', (164, 166), True, 'import datetime as dt\n'), ((214, 246), 'pandas.read_csv', 'pandas.read_csv', (['"""birthdays.csv"""'], {}), "('birthdays.csv')\n", (229, 246), False, 'import pandas\n'), ((651, 685), 'smtplib.SMTP_SSL', 'smtplib.SMTP_SSL', (...
# -*- coding: utf-8 -*- from __future__ import (absolute_import, division, print_function) import warnings from math import exp import numpy as np def fit_factory(discard=1): def fit(x, y): p = np.polyfit(x, y, 1) v = np.polyval(p, x) e = np.abs(y - v) drop_idxs = np.argsort(e)[-...
[ "numpy.abs", "numpy.log", "numpy.polyfit", "numpy.polyval", "numpy.argsort", "numpy.diff", "numpy.array", "warnings.warn", "numpy.delete", "numpy.all" ]
[((210, 229), 'numpy.polyfit', 'np.polyfit', (['x', 'y', '(1)'], {}), '(x, y, 1)\n', (220, 229), True, 'import numpy as np\n'), ((242, 258), 'numpy.polyval', 'np.polyval', (['p', 'x'], {}), '(p, x)\n', (252, 258), True, 'import numpy as np\n'), ((271, 284), 'numpy.abs', 'np.abs', (['(y - v)'], {}), '(y - v)\n', (277, 2...
# -*- coding: utf-8 -*- # Learn more: https://github.com/kennethreitz/setup.py import os, sys from setuptools import setup, find_packages def read_requirements() -> List: """Parse requirements from requirements.txt.""" reqs_path = os.path.join('.', 'requirements.txt') with open(reqs_path, 'r') as f: ...
[ "os.path.join", "setuptools.find_packages" ]
[((242, 279), 'os.path.join', 'os.path.join', (['"""."""', '"""requirements.txt"""'], {}), "('.', 'requirements.txt')\n", (254, 279), False, 'import os, sys\n'), ((667, 707), 'setuptools.find_packages', 'find_packages', ([], {'exclude': "('tests', 'docs')"}), "(exclude=('tests', 'docs'))\n", (680, 707), False, 'from se...
from flatpak_indexer.models import FlatpakBuildModel, ImageModel, ImageBuildModel, RegistryModel IMAGE1 = { "Annotations": {"key1": "value1"}, "Architecture": "amd64", "Digest": "sha256:baabaa", "Labels": {"key2": "value2"}, "MediaType": "application/vnd.docker.distribution.manifest.v2+json", ...
[ "flatpak_indexer.models.ImageBuildModel.from_json", "flatpak_indexer.models.RegistryModel.from_json", "flatpak_indexer.models.ImageModel.from_json" ]
[((1756, 1789), 'flatpak_indexer.models.RegistryModel.from_json', 'RegistryModel.from_json', (['REGISTRY'], {}), '(REGISTRY)\n', (1779, 1789), False, 'from flatpak_indexer.models import FlatpakBuildModel, ImageModel, ImageBuildModel, RegistryModel\n'), ((1897, 1930), 'flatpak_indexer.models.RegistryModel.from_json', 'R...
import time import random from MT19937 import seed_mt from MT19937 import extract_number def delay( seconds ): #delay thì Ctrl + C được, còn time.sleep() thì không start = time.time() while time.time() - start < seconds: pass def main(): start_time = time.time() print('Pending....
[ "MT19937.extract_number", "random.randint", "MT19937.seed_mt", "time.time" ]
[((185, 196), 'time.time', 'time.time', ([], {}), '()\n', (194, 196), False, 'import time\n'), ((286, 297), 'time.time', 'time.time', ([], {}), '()\n', (295, 297), False, 'import time\n'), ((400, 418), 'MT19937.seed_mt', 'seed_mt', (['timestamp'], {}), '(timestamp)\n', (407, 418), False, 'from MT19937 import seed_mt\n'...
from praline.client.project.pipeline.stage_resources import StageResources from praline.client.project.pipeline.stages.stage import stage from praline.client.repository.remote_proxy import RemoteProxy from praline.common.file_system import FileSystem, join from typing import Any, Dict clang_format_style_file_contents...
[ "praline.common.file_system.join", "praline.client.project.pipeline.stages.stage.stage" ]
[((1044, 1176), 'praline.client.project.pipeline.stages.stage.stage', 'stage', ([], {'requirements': "[['project_directory']]", 'output': "['clang_format_style_file', 'clang_format_executable']", 'predicate': 'predicate'}), "(requirements=[['project_directory']], output=[\n 'clang_format_style_file', 'clang_format_e...
import unittest from kalliope.core.NeuronModule import MissingParameterException from kalliope.neurons.slack.slack import Slack class TestSlack(unittest.TestCase): def setUp(self): self.slack_token="<PASSWORD>" self.channel = "kalliochannel" self.message = "kalliomessage" def testPa...
[ "kalliope.neurons.slack.slack.Slack" ]
[((457, 484), 'kalliope.neurons.slack.slack.Slack', 'Slack', ([], {}), '(**parameters_to_test)\n', (462, 484), False, 'from kalliope.neurons.slack.slack import Slack\n')]
# -*- coding: utf-8 -*- # # Copyright 2016 Google LLC. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless requir...
[ "six.iteritems", "googlecloudsdk.command_lib.util.args.labels_util.ProcessUpdateArgsLazy", "googlecloudsdk.api_lib.util.apis.GetClientInstance", "googlecloudsdk.core.properties.VALUES.core.project.GetOrFail", "googlecloudsdk.core.resources.REGISTRY.Parse", "apitools.base.py.list_pager.YieldFromList", "g...
[((1140, 1179), 'googlecloudsdk.api_lib.util.apis.GetClientInstance', 'apis.GetClientInstance', (['"""spanner"""', '"""v1"""'], {}), "('spanner', 'v1')\n", (1162, 1179), False, 'from googlecloudsdk.api_lib.util import apis\n'), ((1189, 1228), 'googlecloudsdk.api_lib.util.apis.GetMessagesModule', 'apis.GetMessagesModule...
from setuptools import setup import pkg_resources import autocomplete def get_metadata_version(): """ Tries to get the version from the django_autocomplete.egg-info directory. """ try: pkg = list(pkg_resources.find_distributions('.', True))[0] except IndexError: return autocomplete...
[ "pkg_resources.find_distributions", "autocomplete.get_mercurial_version", "setuptools.setup" ]
[((431, 1075), 'setuptools.setup', 'setup', ([], {'name': '"""django-autocomplete"""', 'version': 'version', 'description': '"""autocomplete utilities for django"""', 'author': '"""<NAME>"""', 'author_email': '"""<EMAIL>"""', 'url': '"""http://bitbucket.org/tyrion/django-autocomplete"""', 'download_url': '"""http://bit...
# -*- coding: utf-8 -*- """Testing the functions in typhon.plots. """ import os from typhon import plots class TestPlots: """Testing the plot functions.""" def test_figsize(self): """Test golden ratio for figures sizes.""" ret = plots.figsize(10) assert ret == (10, 6.1803398874989481...
[ "typhon.plots.get_subplot_arrangement", "typhon.plots.styles", "os.path.isfile", "typhon.plots.get_available_styles", "typhon.plots.figsize" ]
[((256, 273), 'typhon.plots.figsize', 'plots.figsize', (['(10)'], {}), '(10)\n', (269, 273), False, 'from typhon import plots\n'), ((445, 477), 'typhon.plots.get_subplot_arrangement', 'plots.get_subplot_arrangement', (['(8)'], {}), '(8)\n', (474, 477), False, 'from typhon import plots\n'), ((758, 773), 'typhon.plots.st...
# encoding: utf-8 import ckan.plugins as plugins import ckan.plugins.toolkit as toolkit import logging try: from ckan.common import config # CKAN 2.7 and later except ImportError: from pylons import config # CKAN 2.7 and later log = logging.getLogger(__name__) def dataset_count(): """Return a count of a...
[ "ckan.plugins.implements", "pylons.config.get", "ckan.plugins.toolkit.add_template_directory", "ckan.plugins.toolkit.get_action", "ckan.plugins.toolkit.add_public_directory", "ckan.plugins.toolkit.add_resource", "logging.getLogger" ]
[((243, 270), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (260, 270), False, 'import logging\n'), ((1074, 1113), 'ckan.plugins.implements', 'plugins.implements', (['plugins.IConfigurer'], {}), '(plugins.IConfigurer)\n', (1092, 1113), True, 'import ckan.plugins as plugins\n'), ((1118, 1...
''' This file is used for changing icons color from black to white, so icon color could be changed in code. Icon images are from https://github.com/iconic/open-iconic/tree/master/png ''' import os from PIL import Image # https://pillow.readthedocs.io/en/stable/ def find_files(): """ Finds all files from icons fo...
[ "os.fsdecode", "os.fsencode", "os.listdir", "PIL.Image.open" ]
[((370, 391), 'os.fsencode', 'os.fsencode', (['"""icons/"""'], {}), "('icons/')\n", (381, 391), False, 'import os\n'), ((449, 470), 'os.listdir', 'os.listdir', (['directory'], {}), '(directory)\n', (459, 470), False, 'import os\n'), ((832, 849), 'PIL.Image.open', 'Image.open', (['image'], {}), '(image)\n', (842, 849), ...
from django.core.management.base import BaseCommand from reader.importer.PerseusBatchImporter import PerseusBatchImporter from reader.importer.batch_import import JSONImportPolicy import os import sys class Command(BaseCommand): help = "Imports all Perseus XML documents from a directory that match the i...
[ "reader.importer.batch_import.JSONImportPolicy", "os.path.split", "reader.importer.PerseusBatchImporter.PerseusBatchImporter" ]
[((2171, 2189), 'reader.importer.batch_import.JSONImportPolicy', 'JSONImportPolicy', ([], {}), '()\n', (2187, 2189), False, 'from reader.importer.batch_import import JSONImportPolicy\n'), ((2294, 2453), 'reader.importer.PerseusBatchImporter.PerseusBatchImporter', 'PerseusBatchImporter', ([], {'perseus_directory': 'dire...
"""An "optimizer" that draws random samples. Scientific Machine Learning Benchmark A benchmark of regression models in chem- and materials informatics. 2019-2020, Citrine Informatics. """ from typing import Optional, Any from smlb import ( params, Random, RandomVectorSampler, VectorSpaceData, Opt...
[ "smlb.RandomVectorSampler", "smlb.params.integer" ]
[((903, 939), 'smlb.params.integer', 'params.integer', (['num_samples'], {'above': '(0)'}), '(num_samples, above=0)\n', (917, 939), False, 'from smlb import params, Random, RandomVectorSampler, VectorSpaceData, Optimizer, TrackedTransformation\n'), ((964, 1031), 'smlb.RandomVectorSampler', 'RandomVectorSampler', ([], {...
try: import orjson as json except ImportError: import json import toml import yaml def format_requirements(data: dict) -> str: result = [] for group, values in data.items(): result.append(f"### {group.upper()}\n") for extras, version in values.items(): result.append(f"{ext...
[ "toml.dumps", "yaml.dump", "json.dumps" ]
[((457, 473), 'json.dumps', 'json.dumps', (['data'], {}), '(data)\n', (467, 473), False, 'import json\n'), ((609, 625), 'toml.dumps', 'toml.dumps', (['data'], {}), '(data)\n', (619, 625), False, 'import toml\n'), ((668, 683), 'yaml.dump', 'yaml.dump', (['data'], {}), '(data)\n', (677, 683), False, 'import yaml\n')]
#!/usr/bin/env python3 import os, numpy as np, argparse def relFit(nu, eps): return 7.33972668 * np.power(eps, 1/6.0) / np.sqrt(nu) def etaFit(nu, eps): return np.power(eps, -0.25) * np.power(nu, 0.75) def lambdaFit(nu, eps): return 5.35507603 * np.power(eps,-1/6.0) * np.sqrt(nu); def runspec(nu, eps, run, cs...
[ "argparse.ArgumentParser", "numpy.power", "os.system", "numpy.linspace", "numpy.log10", "os.getenv", "numpy.sqrt" ]
[((762, 779), 'numpy.sqrt', 'np.sqrt', (['(nu / eps)'], {}), '(nu / eps)\n', (769, 779), True, 'import os, numpy as np, argparse\n'), ((1307, 1324), 'numpy.sqrt', 'np.sqrt', (['(nu / eps)'], {}), '(nu / eps)\n', (1314, 1324), True, 'import os, numpy as np, argparse\n'), ((1329, 1496), 'os.system', 'os.system', (['("""e...
from __future__ import absolute_import from __future__ import division from __future__ import print_function from collections import defaultdict import numpy as np def find_smallest_positive(alist): # find first positive value minpos = -1 for x in alist: if x > 0: minpos = x ...
[ "collections.defaultdict", "numpy.mean" ]
[((3311, 3328), 'collections.defaultdict', 'defaultdict', (['list'], {}), '(list)\n', (3322, 3328), False, 'from collections import defaultdict\n'), ((4362, 4382), 'numpy.mean', 'np.mean', (['importances'], {}), '(importances)\n', (4369, 4382), True, 'import numpy as np\n')]
import torch.nn as nn from ..registry import HEADS from ..utils import ConvModule from .bbox_head import BBoxHead import torch import torch.nn.functional as F import mmcv from mmdet.core import mask_target, mask_bg_target, force_fp32, bbox_target, bbox_overlaps from ..losses import accuracy from ..builder import build...
[ "torch.nn.ReLU", "torch.nn.ModuleList", "torch.nn.init.xavier_uniform_", "mmdet.core.mask_target", "torch.cat", "torch.nn.init.constant_", "torch.nn.Linear", "mmdet.core.mask_bg_target" ]
[((4509, 4530), 'torch.nn.ReLU', 'nn.ReLU', ([], {'inplace': '(True)'}), '(inplace=True)\n', (4516, 4530), True, 'import torch.nn as nn\n'), ((5421, 5436), 'torch.nn.ModuleList', 'nn.ModuleList', ([], {}), '()\n', (5434, 5436), True, 'import torch.nn as nn\n'), ((6058, 6073), 'torch.nn.ModuleList', 'nn.ModuleList', ([]...
#!/usr/bin/python3 from services.loggerServices.loggerService import LoggerService logger = LoggerService().logger def create_match(self, data): return self.client.FMT["matches"].insert_one(data).inserted_id def update_match(self, data): return self.client.FMT["matches"].update_one(data, upsert=True).inser...
[ "services.loggerServices.loggerService.LoggerService" ]
[((93, 108), 'services.loggerServices.loggerService.LoggerService', 'LoggerService', ([], {}), '()\n', (106, 108), False, 'from services.loggerServices.loggerService import LoggerService\n')]
#!/usr/bin/env python3 import json import logging import os import re import subprocess # nosec import sys from typing import Dict, Optional, Tuple, cast SRC_VERSION_RE = re.compile(r"^.*\(([^=]*)===?([^=]*)\)$") VERSION_RE = re.compile(r"^([^=]*)==([^=]*)$") LOG = logging.getLogger(__name__) def _get_package_versi...
[ "json.dump", "subprocess.check_output", "logging.getLogger", "re.compile" ]
[((173, 215), 're.compile', 're.compile', (['"""^.*\\\\(([^=]*)===?([^=]*)\\\\)$"""'], {}), "('^.*\\\\(([^=]*)===?([^=]*)\\\\)$')\n", (183, 215), False, 'import re\n'), ((228, 260), 're.compile', 're.compile', (['"""^([^=]*)==([^=]*)$"""'], {}), "('^([^=]*)==([^=]*)$')\n", (238, 260), False, 'import re\n'), ((268, 295)...
# -*- coding: utf-8 -*- ############################################################################## # # OpenERP, Open Source Management Solution # Copyright (C) 2004-2010 Tiny SPRL (<http://tiny.be>). # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU...
[ "subprocess.Popen", "os.unlink", "tempfile.mkstemp", "os.close", "os.path.splitext", "os.write", "logging.getLogger" ]
[((1064, 1091), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1081, 1091), False, 'import logging\n'), ((2877, 2905), 'tempfile.mkstemp', 'tempfile.mkstemp', ([], {'suffix': 'ext'}), '(suffix=ext)\n', (2893, 2905), False, 'import tempfile\n'), ((2922, 2943), 'os.write', 'os.write', (['f...
#!/usr/bin/env python3 from __future__ import print_function import morfessor import sys import logging import lzma import os import math def main(allowed_chars_file, model): allowed_chars = {line.strip() for line in open(allowed_chars_file, encoding='utf-8') if len(line.strip()) == 1} model = morfessor.M...
[ "morfessor.MorfessorIO", "sys.exit", "logging.basicConfig" ]
[((575, 614), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.INFO'}), '(level=logging.INFO)\n', (594, 614), False, 'import logging\n'), ((1005, 1017), 'sys.exit', 'sys.exit', (['(-1)'], {}), '(-1)\n', (1013, 1017), False, 'import sys\n'), ((309, 332), 'morfessor.MorfessorIO', 'morfessor.Morfessor...
from __future__ import print_function from create_tree import * import numpy as np import random DATA_DIR = "../data/" def curriculum_depth(i, num_examples, max_depth): curriculum_max_depth= int((max_depth*i)/num_examples) #print(i, curriculum_max_depth,) if curriculum_max_depth > 0: random_depth ...
[ "numpy.random.randint" ]
[((326, 365), 'numpy.random.randint', 'np.random.randint', (['curriculum_max_depth'], {}), '(curriculum_max_depth)\n', (343, 365), True, 'import numpy as np\n'), ((857, 877), 'numpy.random.randint', 'np.random.randint', (['(2)'], {}), '(2)\n', (874, 877), True, 'import numpy as np\n')]
# Copyright 2015 F5 Networks Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writi...
[ "io.StringIO", "pytest.config.getoption", "distutils.version.LooseVersion", "pytest.fixture", "pytest.raises" ]
[((943, 975), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""function"""'}), "(scope='function')\n", (957, 975), False, 'import pytest\n'), ((1343, 1375), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""function"""'}), "(scope='function')\n", (1357, 1375), False, 'import pytest\n'), ((1649, 1681), 'pytes...
""" 功能:分形树 版本:1.0 日期:2018/08/19 """ import turtle def draw_branch(branch_length, pen_size): if(branch_length > 0): turtle.forward(branch_length) turtle.right(20) draw_branch(branch_length-10, pen_size) turtle.left(40) draw_branch(branch_length-10, pen_size) ...
[ "turtle.backward", "turtle.forward", "turtle.exitonclick", "turtle.right", "turtle.penup", "turtle.left", "turtle.pendown" ]
[((415, 431), 'turtle.right', 'turtle.right', (['(90)'], {}), '(90)\n', (427, 431), False, 'import turtle\n'), ((436, 450), 'turtle.penup', 'turtle.penup', ([], {}), '()\n', (448, 450), False, 'import turtle\n'), ((455, 474), 'turtle.forward', 'turtle.forward', (['(300)'], {}), '(300)\n', (469, 474), False, 'import tur...
# Generated by Django 2.1.4 on 2019-04-26 07:24 from django.db import migrations class Migration(migrations.Migration): dependencies = [ ('core', '0014_auto_20190417_1639'), ] operations = [ migrations.AlterUniqueTogether( name='documentpart', unique_together={('...
[ "django.db.migrations.AlterUniqueTogether" ]
[((224, 321), 'django.db.migrations.AlterUniqueTogether', 'migrations.AlterUniqueTogether', ([], {'name': '"""documentpart"""', 'unique_together': "{('order', 'document')}"}), "(name='documentpart', unique_together={(\n 'order', 'document')})\n", (254, 321), False, 'from django.db import migrations\n')]
# Generated by Django 2.2.13 on 2021-07-22 14:49 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('des', '0032_auto_20210713_2127'), ] operations = [ migrations.AlterField( model_name='astrometryjob', name='status'...
[ "django.db.models.IntegerField" ]
[((340, 522), 'django.db.models.IntegerField', 'models.IntegerField', ([], {'choices': "[(1, 'Idle'), (2, 'Running'), (3, 'Completed'), (4, 'Failed'), (5,\n 'Aborted'), (6, 'Warning'), (7, 'Launched')]", 'default': '(1)', 'verbose_name': '"""Status"""'}), "(choices=[(1, 'Idle'), (2, 'Running'), (3, 'Completed'),\n ...
import pytest from pyformlang.finite_automaton import NondeterministicFiniteAutomaton from project import BooleanMatrices @pytest.fixture def nfa(): nfa = NondeterministicFiniteAutomaton() nfa.add_transitions( [ (0, "X", 1), (0, "X", 2), (1, "Y", 2), (1...
[ "pyformlang.finite_automaton.NondeterministicFiniteAutomaton", "pytest.mark.parametrize", "project.BooleanMatrices" ]
[((441, 542), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""label,expected_nnz"""', "[('X', 2), ('Y', 1), ('Z', 1), ('S', 1), ('W', 2)]"], {}), "('label,expected_nnz', [('X', 2), ('Y', 1), ('Z', 1),\n ('S', 1), ('W', 2)])\n", (464, 542), False, 'import pytest\n'), ((888, 1033), 'pytest.mark.parametrize...
# -*- coding: utf-8 -*- from scripts import tabledef from flask import session from sqlalchemy.orm import sessionmaker from contextlib import contextmanager import bcrypt import sys, subprocess, ipaddress, time, datetime, json, os, csv, copy from watson_developer_cloud import DiscoveryV1 EnvID="5aec3469-82f9-49cb-9...
[ "sqlalchemy.orm.sessionmaker", "watson_developer_cloud.DiscoveryV1" ]
[((911, 1021), 'watson_developer_cloud.DiscoveryV1', 'DiscoveryV1', ([], {'version': '"""2018-03-05"""', 'username': '"""9e523dc4-1206-4898-a30f-faf75cd8526b"""', 'password': '"""<PASSWORD>"""'}), "(version='2018-03-05', username=\n '9e523dc4-1206-4898-a30f-faf75cd8526b', password='<PASSWORD>')\n", (922, 1021), Fals...
# -*- coding: utf-8 -*- from celery.task import task @task def publish_recipe(recipe): from recipes.models import Recipe try: recipe = Recipe.objects.get(id=recipe.id) recipe.published=True recipe.save() except Recipe.DoesNotExist: pass
[ "recipes.models.Recipe.objects.get" ]
[((141, 173), 'recipes.models.Recipe.objects.get', 'Recipe.objects.get', ([], {'id': 'recipe.id'}), '(id=recipe.id)\n', (159, 173), False, 'from recipes.models import Recipe\n')]
# Copyright (c) 2018-2021 Trim21 <<EMAIL>> # Copyright (c) 2008-2014 <NAME> <<EMAIL>> # Licensed under the MIT license. from typing import TYPE_CHECKING, Any, Dict, Tuple, Union, Generator from typing_extensions import Literal from transmission_rpc.lib_types import Field if TYPE_CHECKING: from transmission_rpc.c...
[ "transmission_rpc.lib_types.Field" ]
[((1972, 1990), 'transmission_rpc.lib_types.Field', 'Field', (['value', '(True)'], {}), '(value, True)\n', (1977, 1990), False, 'from transmission_rpc.lib_types import Field\n'), ((2086, 2104), 'transmission_rpc.lib_types.Field', 'Field', (['value', '(True)'], {}), '(value, True)\n', (2091, 2104), False, 'from transmis...
from decimal import Decimal as D from django.test import TestCase from nose.plugins.attrib import attr import mock from oscar.apps.shipping import methods from oscar.apps.shipping.models import OrderAndItemCharges @attr('shipping') class TestStandardMethods(TestCase): def setUp(self): self.non_discount...
[ "decimal.Decimal", "oscar.apps.shipping.methods.TaxExclusiveOfferDiscount", "mock.Mock", "oscar.apps.shipping.methods.Free", "nose.plugins.attrib.attr" ]
[((219, 235), 'nose.plugins.attrib.attr', 'attr', (['"""shipping"""'], {}), "('shipping')\n", (223, 235), False, 'from nose.plugins.attrib import attr\n'), ((832, 843), 'mock.Mock', 'mock.Mock', ([], {}), '()\n', (841, 843), False, 'import mock\n'), ((947, 1005), 'oscar.apps.shipping.methods.TaxExclusiveOfferDiscount',...
# Network import numpy as np import pandas as pd import simulator import random from igraph import * import matplotlib.pyplot as plt class Network(): """docstring for Network""" def __init__(self, simulator): # Genero un grafo random self.g = Graph.Erdos_Renyi(simulator.num_nodi,simulator.p_link) # Ini...
[ "pandas.DataFrame", "numpy.count_nonzero", "numpy.concatenate", "numpy.copy", "matplotlib.pyplot.clf", "numpy.zeros", "numpy.nonzero", "numpy.random.random", "matplotlib.pyplot.gca", "matplotlib.pyplot.ylabel", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.savefig" ]
[((408, 441), 'numpy.zeros', 'np.zeros', (['(simulator.num_nodi, 1)'], {}), '((simulator.num_nodi, 1))\n', (416, 441), True, 'import numpy as np\n'), ((458, 506), 'numpy.zeros', 'np.zeros', (['(simulator.num_nodi, 1)'], {'dtype': 'np.int8'}), '((simulator.num_nodi, 1), dtype=np.int8)\n', (466, 506), True, 'import numpy...
from django.contrib import admin # Register your models here. from .models import shortenedUrl admin.site.register(shortenedUrl)
[ "django.contrib.admin.site.register" ]
[((97, 130), 'django.contrib.admin.site.register', 'admin.site.register', (['shortenedUrl'], {}), '(shortenedUrl)\n', (116, 130), False, 'from django.contrib import admin\n')]
from __future__ import print_function from __future__ import unicode_literals from __future__ import division from __future__ import absolute_import from future import standard_library import twitter_scraper as twitter_scraper import json import codecs standard_library.install_aliases() def main(): def build_list...
[ "codecs.open", "twitter_scraper.scraper.Scraper.get_tweets", "future.standard_library.install_aliases", "json.dumps", "twitter_scraper.scraper.SearchParams" ]
[((253, 287), 'future.standard_library.install_aliases', 'standard_library.install_aliases', ([], {}), '()\n', (285, 287), False, 'from future import standard_library\n'), ((1052, 1109), 'twitter_scraper.scraper.Scraper.get_tweets', 'twitter_scraper.scraper.Scraper.get_tweets', (['search_params'], {}), '(search_params)...
import yaml import click from accelbyte_py_sdk.api.iam import admin_get_banned_users_v3 from ._utils import login_as as login_as_internal @click.command() @click.argument("active_only", type=bool) @click.argument("ban_type") @click.argument("offset", type=int) @click.argument("limit", type=int) @click.option("--na...
[ "click.argument", "click.echo", "click.option", "click.command", "click.Choice", "accelbyte_py_sdk.api.iam.admin_get_banned_users_v3" ]
[((144, 159), 'click.command', 'click.command', ([], {}), '()\n', (157, 159), False, 'import click\n'), ((161, 201), 'click.argument', 'click.argument', (['"""active_only"""'], {'type': 'bool'}), "('active_only', type=bool)\n", (175, 201), False, 'import click\n'), ((203, 229), 'click.argument', 'click.argument', (['""...
from aocd import get_data def part1(a): return sum(len(l) - len(l.encode('utf-8').decode('unicode_escape')) + 2 for l in a) def part2(a): return sum(len(l.encode('unicode_escape').decode('utf-8').replace('"', '\\"')) - len(l) + 2 for l in a) if __name__ == '__main__': data = get_data(day=8, year=2015)...
[ "aocd.get_data" ]
[((294, 320), 'aocd.get_data', 'get_data', ([], {'day': '(8)', 'year': '(2015)'}), '(day=8, year=2015)\n', (302, 320), False, 'from aocd import get_data\n')]
from adafruit_led_animation.animation import Animation import adafruit_framebuf class TextScroll(Animation): def __init__(self, grid_object, speed, text, color, font_name='font5x8.bin', name=None): self._text = text self._font_name = font_name self._frame = 0 # We're only using th...
[ "adafruit_framebuf.FrameBuffer" ]
[((461, 584), 'adafruit_framebuf.FrameBuffer', 'adafruit_framebuf.FrameBuffer', (['self._buffer', 'grid_object.width', 'grid_object.height'], {'buf_format': 'adafruit_framebuf.MVLSB'}), '(self._buffer, grid_object.width, grid_object.\n height, buf_format=adafruit_framebuf.MVLSB)\n', (490, 584), False, 'import adafru...
import numpy as np import tensorflow as tf from tensorflow import keras from tensorflow.keras import layers, models, activations from models.cnn_gru.cnn import CNN, Encoder class CnnGru(tf.keras.Model): def __init__(self, seq_len): super().__init__() self.seq_len = seq_len self.cnn = CNN()...
[ "models.cnn_gru.cnn.Encoder", "tensorflow.nn.relu", "tensorflow.keras.layers.Dropout", "tensorflow.keras.layers.Dense", "models.cnn_gru.cnn.CNN", "tensorflow.keras.layers.Flatten" ]
[((315, 320), 'models.cnn_gru.cnn.CNN', 'CNN', ([], {}), '()\n', (318, 320), False, 'from models.cnn_gru.cnn import CNN, Encoder\n'), ((416, 442), 'models.cnn_gru.cnn.Encoder', 'Encoder', (['self.cnn', 'ip_dims'], {}), '(self.cnn, ip_dims)\n', (423, 442), False, 'from models.cnn_gru.cnn import CNN, Encoder\n'), ((465, ...
'''Create an omim mapping based on information from Face2Gene using the Face2Gene library.''' from pprint import pprint import json from lib.api.face2gene import Face2Gene from lib.model.config import ConfigManager config_data = ConfigManager() f2g_session = Face2Gene(config=config_data) s_list = f2g_session.brows...
[ "lib.api.face2gene.Face2Gene", "lib.model.config.ConfigManager", "json.dump" ]
[((232, 247), 'lib.model.config.ConfigManager', 'ConfigManager', ([], {}), '()\n', (245, 247), False, 'from lib.model.config import ConfigManager\n'), ((263, 292), 'lib.api.face2gene.Face2Gene', 'Face2Gene', ([], {'config': 'config_data'}), '(config=config_data)\n', (272, 292), False, 'from lib.api.face2gene import Fac...
from googletrans import LANGUAGES from googletrans import Translator translator = Translator() async def ajuda(message, comandos :dict): msg = "```\n" for c in comandos.keys(): msg += comandos[c][1]+'\n' msg += "```" await message.channel.send(msg) async def traduz(message, _): msg = message.conten...
[ "googletrans.LANGUAGES.values", "googletrans.Translator", "googletrans.LANGUAGES.keys" ]
[((83, 95), 'googletrans.Translator', 'Translator', ([], {}), '()\n', (93, 95), False, 'from googletrans import Translator\n'), ((1063, 1079), 'googletrans.LANGUAGES.keys', 'LANGUAGES.keys', ([], {}), '()\n', (1077, 1079), False, 'from googletrans import LANGUAGES\n'), ((495, 511), 'googletrans.LANGUAGES.keys', 'LANGUA...