code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
"""SentencePiece Tokenization for Wiki Dataset Example: * python scripts/wiki_sp_tokenize_json.py --word --unigram """ import gzip import json import subprocess from pathlib import Path import sentencepiece as spm import joblib import numpy as np import click from tqdm import tqdm from opencc import OpenCC from wi...
[ "subprocess.run", "gzip.open", "sentencepiece.SentencePieceProcessor", "json.loads", "wiki_tokenize_json.clean_text", "click.option", "click.command", "opencc.OpenCC", "pathlib.Path", "numpy.array", "wiki_tokenize_json.filter_texts" ]
[((564, 577), 'opencc.OpenCC', 'OpenCC', (['"""t2s"""'], {}), "('t2s')\n", (570, 577), False, 'from opencc import OpenCC\n'), ((2890, 2905), 'click.command', 'click.command', ([], {}), '()\n', (2903, 2905), False, 'import click\n'), ((2907, 2943), 'click.option', 'click.option', (['"""--word"""'], {'is_flag': '(True)'}...
from django.conf.urls import url, include from django.contrib import admin from django.views.generic import TemplateView urlpatterns = [ url(r'^admin/', admin.site.urls), url(r'^', include('server.urls'), name='server'), ]
[ "django.conf.urls.include", "django.conf.urls.url" ]
[((143, 174), 'django.conf.urls.url', 'url', (['"""^admin/"""', 'admin.site.urls'], {}), "('^admin/', admin.site.urls)\n", (146, 174), False, 'from django.conf.urls import url, include\n'), ((191, 213), 'django.conf.urls.include', 'include', (['"""server.urls"""'], {}), "('server.urls')\n", (198, 213), False, 'from dja...
# Adapted for numpy/ma/cdms2 by convertcdms.py # Import the modules needed for the tuturial import vcs, cdms2 as cdms, cdutil, time, os, sys # Open data file: filepath = os.path.join(vcs.sample_data, 'clt.nc') cdmsfile = cdms.open( filepath ) # Extract a 3 dimensional data set and get a subset of the time dimension d...
[ "cdms2.open", "os.path.join", "vcs.init" ]
[((171, 210), 'os.path.join', 'os.path.join', (['vcs.sample_data', '"""clt.nc"""'], {}), "(vcs.sample_data, 'clt.nc')\n", (183, 210), False, 'import vcs, cdms2 as cdms, cdutil, time, os, sys\n'), ((222, 241), 'cdms2.open', 'cdms.open', (['filepath'], {}), '(filepath)\n', (231, 241), True, 'import vcs, cdms2 as cdms, cd...
"""Install or upgrade a bcbio-nextgen installation. """ from __future__ import print_function import os import subprocess import sys import yaml from bcbiovm.docker import manage, mounts DEFAULT_IMAGE = "chapmanb/bcbio-nextgen-devel" def full(args, dockerconf): """Full installaction of docker image and data. ...
[ "yaml.load", "os.makedirs", "os.path.getsize", "bcbiovm.docker.manage.run_bcbio_cmd", "yaml.dump", "os.path.exists", "os.path.realpath", "subprocess.check_output", "bcbiovm.docker.mounts.prepare_system", "sys.exit", "os.path.join", "subprocess.check_call" ]
[((486, 548), 'bcbiovm.docker.mounts.prepare_system', 'mounts.prepare_system', (['args.datadir', "dockerconf['biodata_dir']"], {}), "(args.datadir, dockerconf['biodata_dir'])\n", (507, 548), False, 'from bcbiovm.docker import manage, mounts\n'), ((2802, 2887), 'subprocess.check_call', 'subprocess.check_call', (["['dock...
import enum from pandas.io.pytables import DuplicateWarning from py2neo import Node, Relationship, Graph, NodeMatcher import pandas as pd from operator import itemgetter from typing import List, Dict import random graph = Graph("http://localhost:7474", username="neo4j", password='<PASSWORD>') main_ingr = set(['apple'...
[ "py2neo.Graph", "pandas.DataFrame" ]
[((223, 294), 'py2neo.Graph', 'Graph', (['"""http://localhost:7474"""'], {'username': '"""neo4j"""', 'password': '"""<PASSWORD>"""'}), "('http://localhost:7474', username='neo4j', password='<PASSWORD>')\n", (228, 294), False, 'from py2neo import Node, Relationship, Graph, NodeMatcher\n'), ((1890, 1907), 'pandas.DataFra...
import cwgen import os import sys import PySimpleGUI as sg class CwGenUI: # General # GUI - window config WINDOW_DESCRIPTION = 'CW training material generator by SP6HFE' # GUI - text config E2CW_VER_LOCAL_KEY = '-E2CW VER LOCAL-' E2CW_VER_ONLINE_KEY = '-E2CW VER ONLINE-' ...
[ "PySimpleGUI.Button", "PySimpleGUI.Input", "PySimpleGUI.theme", "cwgen.CwGen", "PySimpleGUI.Frame", "PySimpleGUI.Slider", "PySimpleGUI.FileBrowse", "PySimpleGUI.Table", "PySimpleGUI.Text", "PySimpleGUI.Window", "os.path.normpath", "os.path.isfile", "PySimpleGUI.Column", "PySimpleGUI.VSepar...
[((21381, 21401), 'PySimpleGUI.theme', 'sg.theme', (['"""Default1"""'], {}), "('Default1')\n", (21389, 21401), True, 'import PySimpleGUI as sg\n'), ((2063, 2076), 'cwgen.CwGen', 'cwgen.CwGen', ([], {}), '()\n', (2074, 2076), False, 'import cwgen\n'), ((10656, 10698), 'PySimpleGUI.Window', 'sg.Window', (['self.WINDOW_DE...
# TODO: use sliding windows instead of chunking. # the chunking method fails when part of a dupe string # crosses the border between chunks import colorama from colorama import Fore from tqdm import trange, tqdm import os colorama.init(autoreset=True) DEFAULT_MIN = 5 DEFAULT_MAX = 10 def striplist(l): # clean ...
[ "colorama.init", "tqdm.trange" ]
[((224, 253), 'colorama.init', 'colorama.init', ([], {'autoreset': '(True)'}), '(autoreset=True)\n', (237, 253), False, 'import colorama\n'), ((1951, 1971), 'tqdm.trange', 'trange', (['max', 'min', '(-1)'], {}), '(max, min, -1)\n', (1957, 1971), False, 'from tqdm import trange, tqdm\n')]
import boto3 from botocore.exceptions import ClientError import json import os import time import datetime from dateutil import tz from lib.account import * from lib.common import * import logging logger = logging.getLogger() logger.setLevel(logging.DEBUG) logging.getLogger('botocore').setLevel(logging.WARNING) log...
[ "dateutil.tz.gettz", "json.loads", "logging.getLogger", "json.dumps" ]
[((210, 229), 'logging.getLogger', 'logging.getLogger', ([], {}), '()\n', (227, 229), False, 'import logging\n'), ((523, 572), 'json.loads', 'json.loads', (["event['Records'][0]['Sns']['Message']"], {}), "(event['Records'][0]['Sns']['Message'])\n", (533, 572), False, 'import json\n'), ((261, 290), 'logging.getLogger', ...
#!/usr/bin/python3 from PyQt5.QtGui import QBrush, QColor, QPainter from PyQt5.QtChart import QChartView, QChart, QPieSeries from portfolio.utils import confighandler from portfolio.db.fdbhandler import balances from portfolio.db.cdbhandler import cbalances from portfolio.gui.ui_components.fonts import ChartTitleFont...
[ "PyQt5.QtGui.QColor", "portfolio.db.fdbhandler.balances.get_total_balance_all_accounts", "PyQt5.QtChart.QPieSeries", "portfolio.db.cdbhandler.cbalances.get_all_accounts_with_amount_fiat", "portfolio.gui.ui_components.fonts.ChartTitleFont", "portfolio.db.cdbhandler.cbalances.get_total_balance_all_accounts_...
[((590, 598), 'PyQt5.QtChart.QChart', 'QChart', ([], {}), '()\n', (596, 598), False, 'from PyQt5.QtChart import QChartView, QChart, QPieSeries\n'), ((1618, 1630), 'PyQt5.QtChart.QPieSeries', 'QPieSeries', ([], {}), '()\n', (1628, 1630), False, 'from PyQt5.QtChart import QChartView, QChart, QPieSeries\n'), ((2903, 2919)...
#coding=utf8 import re import itchat from itchat.content import * ''' 0.0.1版本 功能: 1.匹配群聊关键字 说 ,然后回复接受到的消息 ''' # 群聊监控 @itchat.msg_register(TEXT, isGroupChat = True) def groupchat_reply(msg): room_name = itchat.search_chatrooms(userName=msg[u'FromUserName']) print(u"来自-%s-群聊消息|%s:%s"%(room_name[ u'NickName'],ms...
[ "itchat.msg_register", "re.match", "itchat.search_chatrooms", "itchat.send_msg", "itchat.auto_login", "itchat.run" ]
[((120, 163), 'itchat.msg_register', 'itchat.msg_register', (['TEXT'], {'isGroupChat': '(True)'}), '(TEXT, isGroupChat=True)\n', (139, 163), False, 'import itchat\n'), ((578, 629), 'itchat.auto_login', 'itchat.auto_login', ([], {'hotReload': '(True)', 'enableCmdQR': '(True)'}), '(hotReload=True, enableCmdQR=True)\n', (...
#! /usr/bin/python3 # -*- coding: utf-8 -*- #################################################################################################### import io #################################################################################################### def TailFile(p_FileName, p_BufferSize=io.DEFAULT_BUFFER_SIZE,...
[ "os.remove", "sys.exit", "time.time" ]
[((2399, 2410), 'sys.exit', 'sys.exit', (['(0)'], {}), '(0)\n', (2407, 2410), False, 'import sys\n'), ((2461, 2472), 'time.time', 'time.time', ([], {}), '()\n', (2470, 2472), False, 'import time\n'), ((2631, 2642), 'time.time', 'time.time', ([], {}), '()\n', (2640, 2642), False, 'import time\n'), ((2796, 2807), 'time.t...
# Copyright (c) Meta Platforms, Inc. and affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import argparse import textwrap from typing import Dict, Optional import libcst as cst from libcst.codemod import CodemodContext, VisitorBa...
[ "textwrap.dedent", "libcst.Comma" ]
[((634, 1244), 'textwrap.dedent', 'textwrap.dedent', (['"""\n Codemod that adds trailing commas to arguments in function\n headers and function calls.\n\n The idea is that both the black and yapf autoformatters will\n tend to split headers and function calls so that there\n is one par...
""" Tests with the Izhikevich neuron model. """ import numpy as np import matplotlib.pyplot as plt import pyNN.nest as sim from pyNN.utility.plotting import Figure, Panel # === Configure the simulator ================================================ duration = 100 dt = 0.01 sim.setup(timestep=dt, min_delay=0.1) ...
[ "pyNN.nest.run", "numpy.count_nonzero", "pyNN.nest.setup", "pyNN.nest.end", "matplotlib.pyplot.scatter", "matplotlib.pyplot.legend", "numpy.logspace", "numpy.array", "pyNN.utility.plotting.Panel", "matplotlib.pyplot.savefig", "pyNN.nest.Izhikevich" ]
[((281, 318), 'pyNN.nest.setup', 'sim.setup', ([], {'timestep': 'dt', 'min_delay': '(0.1)'}), '(timestep=dt, min_delay=0.1)\n', (290, 318), True, 'import pyNN.nest as sim\n'), ((857, 874), 'pyNN.nest.run', 'sim.run', (['duration'], {}), '(duration)\n', (864, 874), True, 'import pyNN.nest as sim\n'), ((1345, 1402), 'mat...
import sys from argparse import ArgumentParser parser = ArgumentParser() parser.add_argument('--indexes', dest='inFile', required=True, help='Path to indexes file, one per line.') parser.add_argument('--config', dest='config', required=True) o=parser.parse_args() outFile = open(f"{o.config}.summa...
[ "sys.stderr.write", "argparse.ArgumentParser" ]
[((57, 73), 'argparse.ArgumentParser', 'ArgumentParser', ([], {}), '()\n', (71, 73), False, 'from argparse import ArgumentParser\n'), ((4423, 4533), 'sys.stderr.write', 'sys.stderr.write', (['f"""{index}/{runID}.dcs.filt.no_overlap.region.c{c}-{C}.d{d}.unique.countmuts.txt\n"""'], {}), "(\n f'{index}/{runID}.dcs.fil...
from pathlib import Path from setuptools import find_packages, setup long_description: str = (Path(__file__).parent.resolve() / "README.md").read_text( encoding="utf-8" ) setup( name="crawler", version="0.0.0", description="A Web Crawler", long_description=long_description, long_description_co...
[ "pathlib.Path", "setuptools.find_packages" ]
[((831, 857), 'setuptools.find_packages', 'find_packages', ([], {'where': '"""src"""'}), "(where='src')\n", (844, 857), False, 'from setuptools import find_packages, setup\n'), ((96, 110), 'pathlib.Path', 'Path', (['__file__'], {}), '(__file__)\n', (100, 110), False, 'from pathlib import Path\n')]
import json import os import requests data_file = './description.json' with open(data_file) as f: data = json.load(f) home_page = data['homepage'] seqs = data['sequences'] for v in seqs: link = '%s%s' % (home_page,v['annotations']['url']) print('download %s' % link) os.system('wget %s -O %s_ann.zip' %...
[ "json.load", "os.system" ]
[((109, 121), 'json.load', 'json.load', (['f'], {}), '(f)\n', (118, 121), False, 'import json\n'), ((285, 339), 'os.system', 'os.system', (["('wget %s -O %s_ann.zip' % (link, v['name']))"], {}), "('wget %s -O %s_ann.zip' % (link, v['name']))\n", (294, 339), False, 'import os\n'), ((438, 492), 'os.system', 'os.system', ...
import hyperion import time import colorsys # Get the parameters rotationTime = float(hyperion.args.get('rotation-time', 2.0)) colorOne = hyperion.args.get('color_one', (255,0,0)) colorTwo = hyperion.args.get('color_two', (0,0,255)) colorsCount = hyperion.args.get('colors_count', hyperion.ledCount/2) reverse = bool(hy...
[ "colorsys.rgb_to_hsv", "colorsys.hsv_to_rgb", "hyperion.abort", "hyperion.setColor", "time.sleep", "hyperion.args.get" ]
[((139, 182), 'hyperion.args.get', 'hyperion.args.get', (['"""color_one"""', '(255, 0, 0)'], {}), "('color_one', (255, 0, 0))\n", (156, 182), False, 'import hyperion\n'), ((192, 235), 'hyperion.args.get', 'hyperion.args.get', (['"""color_two"""', '(0, 0, 255)'], {}), "('color_two', (0, 0, 255))\n", (209, 235), False, '...
""" --- title: Attention with Linear Biases (ALiBi) Experiment summary: This experiment trains an Attention with Linear Biases (ALiBi) based model on Tiny Shakespeare dataset. --- # [Attention with Linear Biases (ALiBi)](index.html) Experiment This is an annotated PyTorch experiment to train a [ALiBi model](index.htm...
[ "labml.configs.option", "labml_nn.transformers.TransformerConfigs", "labml_nn.transformers.alibi.AlibiMultiHeadAttention", "labml.experiment.create", "labml.experiment.configs", "labml.experiment.add_pytorch_models", "labml_helpers.datasets.text.SequentialUnBatchedDataset", "labml.experiment.start", ...
[((2260, 2327), 'labml.configs.calculate', 'calculate', (['TransformerConfigs.encoder_attn', '"""alibi_mha"""', '_alibi_mha'], {}), "(TransformerConfigs.encoder_attn, 'alibi_mha', _alibi_mha)\n", (2269, 2327), False, 'from labml.configs import option, calculate\n'), ((2328, 2395), 'labml.configs.calculate', 'calculate'...
#coding=utf-8 ''' Created on 2015-9-24 @author: Devuser ''' from doraemon.ci.pagefactory.ci_pageworker import CIPageWorker from doraemon.ci.viewmodels.ci_left_nav_bar import CIServiceLeftNavBar from doraemon.ci.viewmodels.ci_sub_nav_bar import CIServiceSubNavBar from doraemon.ci.viewmodels.vm_ci_deploy_service import ...
[ "business.ci.ci_service.CIService.get_products_include_me", "doraemon.ci.pagefactory.ci_pageworker.CIPageWorker.__init__", "doraemon.project.pagefactory.project_common_pageworker.ProjectCommonControllPageWorker.get_myproject_dropdown_list", "business.ci.ci_service.CIService.get_product_ci_services", "doraem...
[((754, 790), 'doraemon.ci.pagefactory.ci_pageworker.CIPageWorker.__init__', 'CIPageWorker.__init__', (['self', 'request'], {}), '(self, request)\n', (775, 790), False, 'from doraemon.ci.pagefactory.ci_pageworker import CIPageWorker\n'), ((974, 1016), 'business.ci.ci_service.CIService.get_products_include_me', 'CIServi...
# Code generated by lark_sdk_gen. DO NOT EDIT. import unittest import pylark import pytest from tests.test_conf import app_all_permission, app_no_permission from tests.test_helper import mock_get_tenant_access_token_failed def mock(*args, **kwargs): raise pylark.PyLarkError(scope="scope", func="func", code=1, ms...
[ "pylark.GetBitableViewListReq", "pylark.PyLarkError", "pylark.CreateBitableViewReq", "pylark.GetBitableRecordReq", "pylark.DeleteBitableRecordReq", "pylark.GetBitableTableListReq", "pytest.raises", "pylark.UpdateBitableFieldReq", "pylark.DeleteBitableFieldReq", "pylark.GetBitableMetaReq", "pylar...
[((263, 336), 'pylark.PyLarkError', 'pylark.PyLarkError', ([], {'scope': '"""scope"""', 'func': '"""func"""', 'code': '(1)', 'msg': '"""mock-failed"""'}), "(scope='scope', func='func', code=1, msg='mock-failed')\n", (281, 336), False, 'import pylark\n'), ((388, 478), 'pylark.PyLarkError', 'pylark.PyLarkError', ([], {'s...
import angr from taint import is_tainted, taintedUnconstrainedBits import logging l = logging.getLogger(__name__) class IROpHook(angr.SimStatePlugin): """ Allows hooking the computation of operations performed in the symbolic execution. (requires our fork of angr to actually respect the hook) """ ...
[ "taint.taintedUnconstrainedBits", "taint.is_tainted", "logging.getLogger" ]
[((87, 114), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (104, 114), False, 'import logging\n'), ((843, 908), 'taint.taintedUnconstrainedBits', 'taintedUnconstrainedBits', (['state', '"""secret"""', 'irop._output_size_bits'], {}), "(state, 'secret', irop._output_size_bits)\n", (867, 90...
import pandas as pd import numpy as np from sklearn.feature_extraction.text import CountVectorizer from sklearn.metrics.pairwise import cosine_similarity def get_title_from_index(index): return df[df.index == index]["title"].values[0] def get_index_from_title(title): return df[df.title.str.lower() == title.lower()...
[ "pandas.read_csv", "sklearn.feature_extraction.text.CountVectorizer", "sklearn.metrics.pairwise.cosine_similarity" ]
[((990, 1029), 'pandas.read_csv', 'pd.read_csv', (['"""./data/movie_dataset.csv"""'], {}), "('./data/movie_dataset.csv')\n", (1001, 1029), True, 'import pandas as pd\n'), ((1264, 1281), 'sklearn.feature_extraction.text.CountVectorizer', 'CountVectorizer', ([], {}), '()\n', (1279, 1281), False, 'from sklearn.feature_ext...
import django_filters from django_filters import CharFilter from .models import * class FoodFilter(django_filters.FilterSet): food_name = CharFilter(field_name = 'name' , lookup_expr = 'icontains',label='search food items') class Meta: model = Food fields = ['food_name']
[ "django_filters.CharFilter" ]
[((140, 226), 'django_filters.CharFilter', 'CharFilter', ([], {'field_name': '"""name"""', 'lookup_expr': '"""icontains"""', 'label': '"""search food items"""'}), "(field_name='name', lookup_expr='icontains', label=\n 'search food items')\n", (150, 226), False, 'from django_filters import CharFilter\n')]
# -*- coding: utf-8 -*- import os from config import Config from flask import Flask, send_from_directory from werkzeug.contrib.fixers import ProxyFix import logging from gtts import gTTS from pydub import AudioSegment import hashlib try: from urllib.parse import unquote_plus except: from urllib import unquote_p...
[ "werkzeug.contrib.fixers.ProxyFix", "config.Config", "gtts.gTTS", "flask.Flask", "os.environ.get", "os.path.isfile", "pydub.AudioSegment.from_file", "flask.send_from_directory", "urllib.unquote_plus", "logging.getLogger" ]
[((334, 342), 'config.Config', 'Config', ([], {}), '()\n', (340, 342), False, 'from config import Config\n'), ((350, 365), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (355, 365), False, 'from flask import Flask, send_from_directory\n'), ((563, 581), 'urllib.unquote_plus', 'unquote_plus', (['text'], {}),...
import ctypes import sys from pathlib import Path from .space import SpaceClient _known_tokens = 'ap cons nil neg c b s isnil car eq mul add lt div i t f cdr SCAN number FUN DEF galaxy GG' _Tokens = {s:i for i, s in enumerate(_known_tokens.split(), 1)} class AlienProxy: def __init__(self): pass class ...
[ "pathlib.Path", "ctypes.cdll.LoadLibrary", "ctypes.POINTER" ]
[((4013, 4040), 'ctypes.cdll.LoadLibrary', 'ctypes.cdll.LoadLibrary', (['fn'], {}), '(fn)\n', (4036, 4040), False, 'import ctypes\n'), ((4055, 4085), 'ctypes.POINTER', 'ctypes.POINTER', (['ctypes.c_int64'], {}), '(ctypes.c_int64)\n', (4069, 4085), False, 'import ctypes\n'), ((3890, 3904), 'pathlib.Path', 'Path', (['__f...
# Augur: A Step Towards Realistic Drift Detection in Production MLSystems - Code # Copyright 2022 Carnegie Mellon University. # # NO WARRANTY. THIS CARNEGIE MELLON UNIVERSITY AND SOFTWARE ENGINEERING INSTITUTE MATERIAL IS FURNISHED ON AN "AS-IS" BASIS. CARNEGIE MELLON UNIVERSITY MAKES NO WARRANTIES OF ANY KIND, EITH...
[ "sklearn.model_selection.KFold", "utils.logging.print_and_log", "tensorflow.keras.callbacks.EarlyStopping", "tensorflow.keras.callbacks.ModelCheckpoint" ]
[((2726, 2787), 'tensorflow.keras.callbacks.EarlyStopping', 'tfcb.EarlyStopping', (['"""val_loss"""'], {'patience': 'patience', 'mode': '"""min"""'}), "('val_loss', patience=patience, mode='min')\n", (2744, 2787), True, 'import tensorflow.keras.callbacks as tfcb\n'), ((2804, 2856), 'tensorflow.keras.callbacks.ModelChec...
''' utility functions ''' __author__ = '<NAME>' import os from os.path import join from os.path import abspath import json import pandas as pd import numpy as np from configs import config as cf def is_available(filename): ''' [filename] : str ''' return os.path.isfile(filename) def chunks(lst,...
[ "json.load", "pandas.read_csv", "os.path.isfile", "os.path.join", "pandas.concat", "numpy.vstack" ]
[((278, 302), 'os.path.isfile', 'os.path.isfile', (['filename'], {}), '(filename)\n', (292, 302), False, 'import os\n'), ((965, 980), 'pandas.concat', 'pd.concat', (['data'], {}), '(data)\n', (974, 980), True, 'import pandas as pd\n'), ((1598, 1669), 'pandas.read_csv', 'pd.read_csv', (['data_path'], {'delimiter': '""",...
from django.contrib import admin from .models import Order # Register your models here. class OrderAdmin(admin.ModelAdmin): list_display = ('id', 'name', 'phone_number', 'submitted') list_filter = ('name', 'submitted') readonly_fields = ('submitted',) fieldsets = ( (None, {'fields': ('name', ...
[ "django.contrib.admin.site.register" ]
[((810, 848), 'django.contrib.admin.site.register', 'admin.site.register', (['Order', 'OrderAdmin'], {}), '(Order, OrderAdmin)\n', (829, 848), False, 'from django.contrib import admin\n')]
import click from rapt.connection import get_vr from rapt.models import query, models from rapt.util import edit_yaml, dump_yaml from pprint import pformat @click.command() def buildpack(): tmpl = { 'repo_url': '', 'repo_type': ['git', 'hg'], 'description': '', 'order': 0, } ...
[ "pprint.pformat", "rapt.models.models.Buildpack", "click.confirm", "rapt.connection.get_vr", "click.echo", "click.command", "rapt.util.dump_yaml", "rapt.models.query" ]
[((161, 176), 'click.command', 'click.command', ([], {}), '()\n', (174, 176), False, 'import click\n'), ((330, 338), 'rapt.connection.get_vr', 'get_vr', ([], {}), '()\n', (336, 338), False, 'from rapt.connection import get_vr\n'), ((546, 603), 'click.echo', 'click.echo', (['"""Creating buildpack with following config:\...
# -*- coding: utf-8 -*- from django.shortcuts import render from django.contrib.auth.decorators import login_required from geoprisma import views as geoprisma_views @login_required def maprender(request, *args, **kwargs): wsName = kwargs.get('wsName') viewId = kwargs.get('viewId') if not viewId: ...
[ "django.shortcuts.render", "geoprisma.views.maprender" ]
[((353, 403), 'geoprisma.views.maprender', 'geoprisma_views.maprender', (['request', 'wsName', 'viewId'], {}), '(request, wsName, viewId)\n', (378, 403), True, 'from geoprisma import views as geoprisma_views\n'), ((516, 581), 'django.shortcuts.render', 'render', (['request', "('example_project/' + templateName)", 'rend...
""" @Fire https://github.com/fire717 """ from lib import init, Data, MoveNet, Task from config import cfg from lib.utils.utils import arg_parser # Script to create and save as images all the various outputs of the model def main(cfg): init(cfg) model = MoveNet(num_classes=cfg["num_classes"], ...
[ "lib.MoveNet", "lib.utils.utils.arg_parser", "lib.Task", "lib.Data", "lib.init" ]
[((246, 255), 'lib.init', 'init', (['cfg'], {}), '(cfg)\n', (250, 255), False, 'from lib import init, Data, MoveNet, Task\n'), ((270, 358), 'lib.MoveNet', 'MoveNet', ([], {'num_classes': "cfg['num_classes']", 'width_mult': "cfg['width_mult']", 'mode': '"""train"""'}), "(num_classes=cfg['num_classes'], width_mult=cfg['w...
#!/usr/bin/env python3 # [rights] Copyright 2020 brianddk at github https://github.com/brianddk # [license] Apache 2.0 License https://www.apache.org/licenses/LICENSE-2.0 # [repo] github.com/brianddk/reddit/blob/master/python/elec-p2sh-hodl.py # [btc] BTC-b32: bc1qwc2203uym96u0nmq04pcgqfs9ldqz9l3mz8fpj # [tipja...
[ "electrum.constants.set_testnet", "electrum.bitcoin.deserialize_privkey", "electrum.transaction.PartialTxInput", "electrum.transaction.PartialTxOutput.from_address_and_value", "electrum.ecc.ECPrivkey", "electrum.bitcoin.push_script", "electrum.transaction.TxOutpoint", "electrum.transaction.PartialTran...
[((1640, 1653), 'electrum.constants.set_testnet', 'set_testnet', ([], {}), '()\n', (1651, 1653), False, 'from electrum.constants import set_testnet\n'), ((2262, 2286), 'electrum.bitcoin.deserialize_privkey', 'deserialize_privkey', (['wif'], {}), '(wif)\n', (2281, 2286), False, 'from electrum.bitcoin import deserialize_...
from django.contrib import admin from app.models import * class TrackAdmin(admin.ModelAdmin): list_display=('title','description',) class SessionAdmin(admin.ModelAdmin): list_display = ('title','status',) search_fields = ['title','abstract'] list_filter = ('track','speaker',) actions =...
[ "django.contrib.admin.site.register" ]
[((1079, 1121), 'django.contrib.admin.site.register', 'admin.site.register', (['Speaker', 'SpeakerAdmin'], {}), '(Speaker, SpeakerAdmin)\n', (1098, 1121), False, 'from django.contrib import admin\n'), ((1122, 1160), 'django.contrib.admin.site.register', 'admin.site.register', (['Track', 'TrackAdmin'], {}), '(Track, Tra...
#!/usr/bin/env python """ Defines unit-tests for the module at hms_core.data_objects. """ ####################################### # Any needed from __future__ imports # # Create an "__all__" list to support # # "from module import member" use # ####################################### __all__ = [ # Test-case...
[ "uuid.uuid4", "unittest.TestSuite", "time.time", "datetime.datetime.strptime", "uuid.UUID", "unittest.TestLoader", "datetime.datetime.fromtimestamp", "datetime.datetime.now", "unittest.TestResult" ]
[((1371, 1391), 'unittest.TestSuite', 'unittest.TestSuite', ([], {}), '()\n', (1389, 1391), False, 'import unittest\n'), ((1891, 1905), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (1903, 1905), False, 'from datetime import datetime\n'), ((1907, 1941), 'datetime.datetime.fromtimestamp', 'datetime.fromtime...
from __future__ import annotations import argparse import asyncio import getpass import inspect import json import logging import logging.config import os import threading import time import traceback import warnings import webbrowser from dataclasses import dataclass, field from datetime import datetime, timedelta fr...
[ "os.remove", "argparse.ArgumentParser", "typing.cast", "rich.traceback.install", "logging.getLevelName", "os.path.join", "rich.logging.RichHandler", "traceback.print_exc", "warnings.simplefilter", "shell.Command", "bakalariapi.BakalariAPI", "os.path.dirname", "prompt_toolkit.input.create_inp...
[((1762, 1790), 'rich.traceback.install', 'tb_install', ([], {'show_locals': '(True)'}), '(show_locals=True)\n', (1772, 1790), True, 'from rich.traceback import install as tb_install\n'), ((1872, 1965), 'platformdirs.PlatformDirs', 'platformdirs.PlatformDirs', ([], {'appauthor': '"""BakalariAPI"""', 'appname': '"""baka...
import json import codecs import pandas as pd import boto3 csv_path = "./fixed_tweets.csv" save_path = "./fixed_tweets.json" df = pd.read_csv(csv_path, header=None, encoding = "utf-8") df.columns =["tweet"] df_json = df.to_dict(orient='records') resource = boto3.resource('dynamodb', region_name='ap-northeast-1') #...
[ "pandas.read_csv", "boto3.resource" ]
[((133, 185), 'pandas.read_csv', 'pd.read_csv', (['csv_path'], {'header': 'None', 'encoding': '"""utf-8"""'}), "(csv_path, header=None, encoding='utf-8')\n", (144, 185), True, 'import pandas as pd\n'), ((261, 317), 'boto3.resource', 'boto3.resource', (['"""dynamodb"""'], {'region_name': '"""ap-northeast-1"""'}), "('dyn...
"""Authentication routes.""" from __future__ import annotations from flask import g from flask_accept import accept_fallback from keeper.api import api from keeper.auth import password_auth from keeper.logutils import log_route from ._models import AuthTokenResponse @api.route("/token") @accept_fallback @log_rout...
[ "keeper.logutils.log_route", "keeper.api.api.route", "flask.g.user.generate_auth_token" ]
[((274, 293), 'keeper.api.api.route', 'api.route', (['"""/token"""'], {}), "('/token')\n", (283, 293), False, 'from keeper.api import api\n'), ((312, 323), 'keeper.logutils.log_route', 'log_route', ([], {}), '()\n', (321, 323), False, 'from keeper.logutils import log_route\n'), ((1250, 1278), 'flask.g.user.generate_aut...
""" Test princomp extraction from CLI """ import argparse import os import numpy as np from demo_utils import get_random_data from hebbnets.networks import MultilayerHahNetwork np.set_printoptions(suppress=True) def _argparse(): parser = argparse.ArgumentParser( prog="Testing HebbNet principal c...
[ "numpy.set_printoptions", "demo_utils.get_random_data", "argparse.ArgumentParser", "hebbnets.networks.MultilayerHahNetwork", "numpy.array_str", "numpy.argsort", "numpy.linalg.norm", "numpy.array", "numpy.squeeze" ]
[((186, 220), 'numpy.set_printoptions', 'np.set_printoptions', ([], {'suppress': '(True)'}), '(suppress=True)\n', (205, 220), True, 'import numpy as np\n'), ((254, 410), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'prog': '"""Testing HebbNet principal components"""', 'description': '"""Testing HebbNet p...
# -*- coding: utf-8 -*- """ ## Author: <NAME> ## Copyright: Copyright 2018-2019, Packt Publishing Limited ## Version: 0.0.1 ## Maintainer: <NAME> ## Email: <EMAIL> ## Linkedin: https://www.linkedin.com/in/linus1/ ## Contributor : {if you debug, append your name here} ## Contributor Email : {if you debug, appen...
[ "matplotlib.pyplot.xlim", "sklearn.pipeline.Pipeline", "numpy.random.seed", "matplotlib.pyplot.show", "matplotlib.pyplot.ylim", "numpy.random.randn", "sklearn.model_selection.cross_val_score", "matplotlib.pyplot.scatter", "matplotlib.pyplot.legend", "matplotlib.pyplot.setp", "sklearn.linear_mode...
[((618, 635), 'numpy.random.seed', 'np.random.seed', (['(0)'], {}), '(0)\n', (632, 635), True, 'import numpy as np\n'), ((804, 827), 'numpy.cos', 'np.cos', (['(1.5 * np.pi * X)'], {}), '(1.5 * np.pi * X)\n', (810, 827), True, 'import numpy as np\n'), ((1018, 1045), 'matplotlib.pyplot.figure', 'plt.figure', ([], {'figsi...
import numpy as np import argparse from utils import Audio def sample_wav_audio(path): audio = Audio() mel = audio.audio_to_mel(path) samples = audio.mel_sample(mel, width=128, k=5) return samples def save_embeddings(name, samples): audio = Audio() avg_embed = np.zeros(256, dtype=np.float32) ...
[ "utils.Audio", "numpy.save", "numpy.zeros", "argparse.ArgumentParser" ]
[((100, 107), 'utils.Audio', 'Audio', ([], {}), '()\n', (105, 107), False, 'from utils import Audio\n'), ((263, 270), 'utils.Audio', 'Audio', ([], {}), '()\n', (268, 270), False, 'from utils import Audio\n'), ((287, 318), 'numpy.zeros', 'np.zeros', (['(256)'], {'dtype': 'np.float32'}), '(256, dtype=np.float32)\n', (295...
import argparse import logging import os import pickle import random import ujson import sys import math from ctypes import c_ulong from multiprocessing import Array, Queue from multiprocessing.sharedctypes import RawArray from queue import Empty from time import time import numpy as np import resource from scipy.spa...
[ "argparse.ArgumentParser", "experiments.duplicate_bug_detection_deep_learning.getDataHandlerLexiconEmb", "sklearn.metrics.accuracy_score", "numpy.around", "multiprocessing.Queue", "data.bug_report_database.BugReportDatabase.fromJson", "experiments.duplicate_bug_detection_deep_learning.getModel", "skle...
[((2982, 3001), 'logging.getLogger', 'logging.getLogger', ([], {}), '()\n', (2999, 3001), False, 'import logging\n'), ((3010, 3016), 'time.time', 'time', ([], {}), '()\n', (3014, 3016), False, 'from time import time\n'), ((6412, 6437), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (6435, 6437)...
"""Tests for registering custom formatters.""" from pyramid.testing import DummyRequest from pyramid.testing import testConfig def test_add_formatter() -> None: """Test registration of a custom formatter.""" with testConfig() as config: request = DummyRequest() config.include("pyramid_opena...
[ "pyramid.testing.testConfig", "pyramid.testing.DummyRequest" ]
[((225, 237), 'pyramid.testing.testConfig', 'testConfig', ([], {}), '()\n', (235, 237), False, 'from pyramid.testing import testConfig\n'), ((267, 281), 'pyramid.testing.DummyRequest', 'DummyRequest', ([], {}), '()\n', (279, 281), False, 'from pyramid.testing import DummyRequest\n')]
#!/usr/bin/env python """Test USCG specific 8:367:22 area notice message Version 23 samples.""" import datetime import unittest from ais_area_notice import m366_22 # from m366_22 import AreaNotice # from m366_22 import AreaNoticeCircle # from m366_22 import AreaNoticeRectangle # from m366_22 import AreaNoticeSector ...
[ "unittest.main", "unittest.skip", "ais_area_notice.m366_22.AreaNotice", "datetime.datetime.utcnow" ]
[((1203, 1252), 'unittest.skip', 'unittest.skip', (['"""TODO(schwehr): Fix this failure."""'], {}), "('TODO(schwehr): Fix this failure.')\n", (1216, 1252), False, 'import unittest\n'), ((1593, 1608), 'unittest.main', 'unittest.main', ([], {}), '()\n', (1606, 1608), False, 'import unittest\n'), ((614, 640), 'datetime.da...
#!/usr/bin/env python # -*- coding: utf-8 -*- ######################################################################## # # Copyright 2015 Baidu, Inc. # ######################################################################## """ File: test_create_thumbnail.py Date: 2015/06/10 15:15:40 """ import os import sys impo...
[ "mediaBase.MediaBase.__init__", "os.path.abspath", "nose.tools.assert_true", "nose.tools.assert_raises_regexp", "sys.path.insert", "baidubce.services.media.media_client.MediaClient", "time.sleep", "nose.tools.assert_is_not_none" ]
[((504, 536), 'sys.path.insert', 'sys.path.insert', (['(0)', '_COMMON_PATH'], {}), '(0, _COMMON_PATH)\n', (519, 536), False, 'import sys\n'), ((432, 457), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (447, 457), False, 'import os\n'), ((1010, 1044), 'mediaBase.MediaBase.__init__', 'mediaBas...
#!/usr/bin/env python3 """ This module specifies class Monitored designated for processing of cryptocurrencies blocks """ import threading import logging from timeit import default_timer as timer from datetime import datetime, timedelta from .coin import BTC, BCH, DASH, ZEC, LTC, ETH from .database import Database from...
[ "threading.Thread", "timeit.default_timer", "threading.Event", "datetime.datetime.now", "logging.getLogger" ]
[((357, 384), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (374, 384), False, 'import logging\n'), ((492, 509), 'threading.Event', 'threading.Event', ([], {}), '()\n', (507, 509), False, 'import threading\n'), ((2015, 2079), 'threading.Thread', 'threading.Thread', ([], {'target': 'self....
import threading class MultiThreading(object): def __init__(self, scrapers): self.scrapers = scrapers def run(self): threads = [] for i in range(len(self.scrapers)): t = threading.Thread(target=self.scrapers[i].start) t.start() threads.append(t) ...
[ "threading.Thread" ]
[((218, 265), 'threading.Thread', 'threading.Thread', ([], {'target': 'self.scrapers[i].start'}), '(target=self.scrapers[i].start)\n', (234, 265), False, 'import threading\n')]
import glob import numpy as np from matplotlib import pyplot as plt for filename in glob.glob("*.dat"): print(filename) name = filename.split(".")[0] data = np.loadtxt(filename, delimiter=",") size = int(np.sqrt(len(data))) data = data.reshape((size, size)) fig, ax = plt.subplots(figsize=(5.12,...
[ "matplotlib.pyplot.close", "matplotlib.pyplot.subplots", "numpy.loadtxt", "glob.glob", "matplotlib.pyplot.tick_params", "matplotlib.pyplot.tight_layout", "matplotlib.pyplot.savefig" ]
[((85, 103), 'glob.glob', 'glob.glob', (['"""*.dat"""'], {}), "('*.dat')\n", (94, 103), False, 'import glob\n'), ((170, 205), 'numpy.loadtxt', 'np.loadtxt', (['filename'], {'delimiter': '""","""'}), "(filename, delimiter=',')\n", (180, 205), True, 'import numpy as np\n'), ((293, 327), 'matplotlib.pyplot.subplots', 'plt...
# -*- coding: utf-8 -*- """ Created on Sun Aug 6 00:25:27 2017 @author: Wayne """ import pandas as pd import xgboost as xgb import numpy as np from sklearn.model_selection import train_test_split import pickle #%% mydf1= mydf[outliers.outliers==False] z = np.log(data.trip_duration+1) X = mydf1 Xtest ...
[ "pandas.DataFrame", "sklearn.cluster.MiniBatchKMeans", "numpy.log", "xgboost.train", "sklearn.model_selection.train_test_split", "pandas.get_dummies", "pickle.load", "numpy.mean", "numpy.exp", "numpy.intersect1d", "pandas.concat", "xgboost.DMatrix", "numpy.vstack" ]
[((273, 303), 'numpy.log', 'np.log', (['(data.trip_duration + 1)'], {}), '(data.trip_duration + 1)\n', (279, 303), True, 'import numpy as np\n'), ((342, 360), 'xgboost.DMatrix', 'xgb.DMatrix', (['Xtest'], {}), '(Xtest)\n', (353, 360), True, 'import xgboost as xgb\n'), ((899, 952), 'sklearn.model_selection.train_test_sp...
# -*- coding: utf-8 -*- import tensorflow as tf import numpy as np import cPickle import ipdb class Detector(): def __init__(self,weight_file_path,n_labels): self.image_mean=[103.939,116.779,123.68] self.n_labels=n_labels with open(weight_file_path)as f: self.pretrained_weights=cPickle.load(f) def get_weight...
[ "tensorflow.nn.relu", "tensorflow.constant_initializer", "tensorflow.reshape", "tensorflow.reduce_mean", "tensorflow.concat", "cPickle.load", "tensorflow.variable_scope", "tensorflow.nn.bias_add", "tensorflow.nn.max_pool", "tensorflow.matmul", "tensorflow.get_variable", "tensorflow.random_norm...
[((1521, 1539), 'numpy.prod', 'np.prod', (['shape[1:]'], {}), '(shape[1:])\n', (1528, 1539), True, 'import numpy as np\n'), ((1544, 1573), 'tensorflow.reshape', 'tf.reshape', (['bottom', '[-1, dim]'], {}), '(bottom, [-1, dim])\n', (1554, 1573), True, 'import tensorflow as tf\n'), ((2136, 2154), 'numpy.prod', 'np.prod',...
# -*- coding: utf-8 -*- import os from unittest import TestCase import tempfile from server import db def use_temp_db(filename): def _use_temp_db(fn): def wrapper(obj): with tempfile.TemporaryDirectory() as dirname: fn(obj, os.path.join(dirname, filename)) return wrapp...
[ "tempfile.NamedTemporaryFile", "tempfile.TemporaryDirectory", "server.db.ItemLabelDBController", "server.db.MatchResultDBController", "os.path.join", "server.db.RatedMatchResultDBController" ]
[((695, 731), 'server.db.MatchResultDBController', 'db.MatchResultDBController', (['filename'], {}), '(filename)\n', (721, 731), False, 'from server import db\n'), ((1176, 1212), 'server.db.MatchResultDBController', 'db.MatchResultDBController', (['filename'], {}), '(filename)\n', (1202, 1212), False, 'from server impo...
"""Quantitative Proteomic Service Usage: pyqp api pyqp cli <proteomicTSV> <proteomeXML> [--field=<quantity_column>] [--adress=<apiAdress>] [--port=<apiPort>] [--verbose] [--topScore=<pathway_number>] Options: -h --help Show this screen. --field=<quantity column> csv column header featuring signal ...
[ "unigo.uloads", "pyproteinsExt.uniprot.EntrySet", "requests.get", "docopt.docopt" ]
[((1037, 1052), 'docopt.docopt', 'docopt', (['__doc__'], {}), '(__doc__)\n', (1043, 1052), False, 'from docopt import docopt\n'), ((1406, 1471), 'pyproteinsExt.uniprot.EntrySet', 'createUniprotCollection', ([], {'collectionXML': "arguments['<proteomeXML>']"}), "(collectionXML=arguments['<proteomeXML>'])\n", (1429, 1471...
from kafka import KafkaConsumer, TopicPartition from json import loads from sqlalchemy import create_engine, Table, Column, Integer, String from sqlalchemy.ext.declarative import declarative_base import os user = os.getenv('MYSQL_user') pw = os.getenv('MYSQL') str_sql = 'mysql+mysqlconnector://' + user + ':' + pw + '@...
[ "sqlalchemy.String", "sqlalchemy.ext.declarative.declarative_base", "sqlalchemy.Column", "sqlalchemy.create_engine", "os.getenv" ]
[((214, 237), 'os.getenv', 'os.getenv', (['"""MYSQL_user"""'], {}), "('MYSQL_user')\n", (223, 237), False, 'import os\n'), ((243, 261), 'os.getenv', 'os.getenv', (['"""MYSQL"""'], {}), "('MYSQL')\n", (252, 261), False, 'import os\n'), ((348, 370), 'sqlalchemy.create_engine', 'create_engine', (['str_sql'], {}), '(str_sq...
import os from google.cloud import speech_v1p1beta1 as speech import io #Set env variable, because it resets every shell session os.environ["GOOGLE_APPLICATION_CREDENTIALS"] = "/home/robin_jf_andersson/mbox_speaker_diarization/mbox1-28508a73fde1.json" def speaker_diarization(audio_file, channels, sample_rate, nbr_of_...
[ "google.cloud.speech_v1p1beta1.RecognitionConfig", "google.cloud.speech_v1p1beta1.RecognitionAudio", "google.cloud.speech_v1p1beta1.SpeechClient" ]
[((343, 364), 'google.cloud.speech_v1p1beta1.SpeechClient', 'speech.SpeechClient', ([], {}), '()\n', (362, 364), True, 'from google.cloud import speech_v1p1beta1 as speech\n'), ((492, 532), 'google.cloud.speech_v1p1beta1.RecognitionAudio', 'speech.RecognitionAudio', ([], {'content': 'content'}), '(content=content)\n', ...
from manimlib.imports import * from ManimProjects.utils.Parabola import Parabola from ManimProjects.utils.geometry import CText class Prob1(Parabola): CONFIG = { 'x_min' : -5 } def construct(self): self.adjust_x_range() graph = self.get_graph(color=LIGHT_BROWN) directrix = s...
[ "ManimProjects.utils.geometry.CText" ]
[((1736, 1748), 'ManimProjects.utils.geometry.CText', 'CText', (['"""焦点弦"""'], {}), "('焦点弦')\n", (1741, 1748), False, 'from ManimProjects.utils.geometry import CText\n'), ((8883, 8900), 'ManimProjects.utils.geometry.CText', 'CText', (['"""这些交点在准线上"""'], {}), "('这些交点在准线上')\n", (8888, 8900), False, 'from ManimProjects.ut...
'''faça um programa que mostre a tabuada de vários números, um de cada vez, para cada valor digitado pelo usuário. O programa será interrompido quando o valor solicitado for negativo.''' from time import sleep n = 0 cont = 0 while n >= 0: print('--' * 15) print('\033[33mPara cancelar, digite um número negativo....
[ "time.sleep" ]
[((489, 497), 'time.sleep', 'sleep', (['(1)'], {}), '(1)\n', (494, 497), False, 'from time import sleep\n')]
#!/usr/bin/env python # -*- coding: utf-8 -*- import json import os from contextlib import contextmanager import click import crayons def open_url(url): click.echo("Opening {}.".format(crayons.white(url, bold=True))) click.launch(url) def get_config_file_path(): home = os.path.expanduser("~") retur...
[ "json.dump", "crayons.white", "os.path.exists", "click.launch", "os.path.expanduser" ]
[((228, 245), 'click.launch', 'click.launch', (['url'], {}), '(url)\n', (240, 245), False, 'import click\n'), ((287, 310), 'os.path.expanduser', 'os.path.expanduser', (['"""~"""'], {}), "('~')\n", (305, 310), False, 'import os\n'), ((640, 660), 'os.path.exists', 'os.path.exists', (['path'], {}), '(path)\n', (654, 660),...
import json import requests import ccxt import time import os import pandas as pd from datetime import datetime, timedelta import operator import csv import cfg liquid = ccxt.liquid(cfg.liquid_misc_credential) exchange = liquid since = exchange.milliseconds() - 86400000 # -1 day from now def save_and_get_str(): ...
[ "pandas.DataFrame", "pandas.read_csv", "time.sleep", "ccxt.liquid", "os.path.isfile" ]
[((172, 211), 'ccxt.liquid', 'ccxt.liquid', (['cfg.liquid_misc_credential'], {}), '(cfg.liquid_misc_credential)\n', (183, 211), False, 'import ccxt\n'), ((784, 875), 'pandas.DataFrame', 'pd.DataFrame', ([], {'columns': "['utc', 'time', 'type', 'amount', 'price', 'fee', 'takerOrMaker']"}), "(columns=['utc', 'time', 'typ...
import re from datetime import datetime from datetime import timedelta import dateutil.parser import pytz import tzlocal # datetime objct for beginning of epoch T_EPOCH = datetime(1970, 1, 1, tzinfo=pytz.utc) DEFAULT = object() # singleton, for args with default values class DateTimeError(Exception): """ custo...
[ "tzlocal.get_localzone", "datetime.datetime", "datetime.datetime.strptime", "datetime.timedelta", "datetime.datetime.now", "re.compile" ]
[((173, 210), 'datetime.datetime', 'datetime', (['(1970)', '(1)', '(1)'], {'tzinfo': 'pytz.utc'}), '(1970, 1, 1, tzinfo=pytz.utc)\n', (181, 210), False, 'from datetime import datetime\n'), ((634, 670), 're.compile', 're.compile', (["('^' + single_delta + '$')"], {}), "('^' + single_delta + '$')\n", (644, 670), False, '...
""" This module implements connections for CUBRIDdb. Presently there is only one class: Connection. Others are unlikely. However, you might want to make your own subclasses. In most cases, you will probably override Connection.default_cursor with a non-standard Cursor class. """ from CUBRIDdb.cursors import * import t...
[ "_cubrid.connect" ]
[((634, 667), '_cubrid.connect', '_cubrid.connect', (['*args'], {}), '(*args, **kwargs2)\n', (649, 667), False, 'import types, _cubrid\n')]
from setuptools import setup setup(name='tutorials', version='POC', url='https://github.com/pyomeca/tutorials.git', author='pyomeca', packages=['src'], zip_safe=False)
[ "setuptools.setup" ]
[((30, 178), 'setuptools.setup', 'setup', ([], {'name': '"""tutorials"""', 'version': '"""POC"""', 'url': '"""https://github.com/pyomeca/tutorials.git"""', 'author': '"""pyomeca"""', 'packages': "['src']", 'zip_safe': '(False)'}), "(name='tutorials', version='POC', url=\n 'https://github.com/pyomeca/tutorials.git', ...
import random import json import math async def battle_attack(x, y, u, e, call): if x == y: await call.answer("❗ Противник увернулся от удара", show_alert=True) return e.health, e.defence else: if e.defence <= 0: e.health -= u.damage return e.health, e.defence ...
[ "math.floor", "random.randint", "random.uniform" ]
[((2701, 2723), 'random.randint', 'random.randint', (['(1)', '(100)'], {}), '(1, 100)\n', (2715, 2723), False, 'import random\n'), ((3000, 3026), 'math.floor', 'math.floor', (['(n * multiplier)'], {}), '(n * multiplier)\n', (3010, 3026), False, 'import math\n'), ((3170, 3194), 'random.uniform', 'random.uniform', (['(0....
import pandas as pd from news_classifier.database import db def load_data(projection: dict) -> pd.DataFrame: """ Load the data from the Mongo collection and transform into a pandas dataframe :projection: A dictionary with the fields to load from database :return: A pandas dataframe with the data ...
[ "pandas.DataFrame", "news_classifier.database.db.read_articles" ]
[((343, 382), 'news_classifier.database.db.read_articles', 'db.read_articles', ([], {'projection': 'projection'}), '(projection=projection)\n', (359, 382), False, 'from news_classifier.database import db\n'), ((408, 430), 'pandas.DataFrame', 'pd.DataFrame', (['articles'], {}), '(articles)\n', (420, 430), True, 'import ...
#!/usr/bin/env python import argparse import sys from collections import defaultdict DEFAULT_OUT = "stackcollapse-merged.txt" def merge(files, dst): data = defaultdict(lambda: 0) for file in files: with open(file, "r") as fp: for line in fp.readlines(): stack, hits = lin...
[ "collections.defaultdict", "argparse.ArgumentParser" ]
[((164, 187), 'collections.defaultdict', 'defaultdict', (['(lambda : 0)'], {}), '(lambda : 0)\n', (175, 187), False, 'from collections import defaultdict\n'), ((546, 582), 'argparse.ArgumentParser', 'argparse.ArgumentParser', (['sys.argv[0]'], {}), '(sys.argv[0])\n', (569, 582), False, 'import argparse\n'), ((596, 686)...
import os from fastapi import Depends, FastAPI, Response, status from fastapi.middleware.cors import CORSMiddleware from fastapi_users import FastAPIUsers from fastapi_users.authentication import JWTAuthentication from sqlalchemy.orm import Session from .database.session import database, user_db from .deps import db_...
[ "fastapi_users.authentication.JWTAuthentication", "fastapi_users.FastAPIUsers", "fastapi.Depends", "os.getenv", "fastapi.FastAPI" ]
[((1057, 1080), 'os.getenv', 'os.getenv', (['"""APP_SECRET"""'], {}), "('APP_SECRET')\n", (1066, 1080), False, 'import os\n'), ((1103, 1195), 'fastapi_users.authentication.JWTAuthentication', 'JWTAuthentication', ([], {'secret': 'APP_SECRET', 'lifetime_seconds': '(3600)', 'tokenUrl': '"""/auth/jwt/login"""'}), "(secret...
"""Utility functions.""" import subprocess import logging import os import shutil import stat import itertools from collections import OrderedDict from pkg_resources import resource_string import pandas as pd from genometools.expression import ExpGeneTable from genometools import gtf import singlecell _LOGGER = lo...
[ "subprocess.Popen", "os.chmod", "os.stat", "pandas.read_csv", "os.walk", "pkg_resources.resource_string", "itertools.product", "collections.OrderedDict", "genometools.gtf.parse_attributes", "shutil.copyfileobj", "os.path.join", "logging.getLogger" ]
[((318, 345), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (335, 345), False, 'import logging\n'), ((2162, 2175), 'os.stat', 'os.stat', (['path'], {}), '(path)\n', (2169, 2175), False, 'import os\n'), ((2180, 2221), 'os.chmod', 'os.chmod', (['path', '(st.st_mode | stat.S_IEXEC)'], {}), ...
import os import frida from flask import Flask, jsonify, request from hook import start_hook REMOTE_DEVICE = os.getenv('REMOTE_DEVICE', '') app = Flask(__name__) api = start_hook(REMOTE_DEVICE) @app.route('/sign') def sign(): global api url = request.args.get('url', '') headers = dict(request.headers)...
[ "flask.request.args.get", "flask.Flask", "flask.jsonify", "os.getenv", "hook.start_hook" ]
[((111, 141), 'os.getenv', 'os.getenv', (['"""REMOTE_DEVICE"""', '""""""'], {}), "('REMOTE_DEVICE', '')\n", (120, 141), False, 'import os\n'), ((150, 165), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (155, 165), False, 'from flask import Flask, jsonify, request\n'), ((172, 197), 'hook.start_hook', 'star...
import PySimpleGUI as sg from ..model.objectdata import ObjectData from ..model.search import SearchableList from ..view import newproduction from . import get_string_unit, RACES, filter_listbox from myconfigparser import Section def open_window(data): options = SearchableList() for u in data: unit = ...
[ "myconfigparser.Section", "PySimpleGUI.popup" ]
[((320, 336), 'myconfigparser.Section', 'Section', (['data[u]'], {}), '(data[u])\n', (327, 336), False, 'from myconfigparser import Section\n'), ((1030, 1049), 'PySimpleGUI.popup', 'sg.popup', (['"""Success"""'], {}), "('Success')\n", (1038, 1049), True, 'import PySimpleGUI as sg\n')]
# Generated by Django 3.2.3 on 2021-08-10 02:38 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('artists', '0002_artist_bio'), ('beats', '0002_instrumental_img_file'), ] operations = [ ...
[ "django.db.models.ForeignKey", "django.db.models.ManyToManyField" ]
[((435, 567), 'django.db.models.ForeignKey', 'models.ForeignKey', ([], {'blank': '(True)', 'null': '(True)', 'on_delete': 'django.db.models.deletion.DO_NOTHING', 'related_name': '"""+"""', 'to': '"""artists.artist"""'}), "(blank=True, null=True, on_delete=django.db.models.\n deletion.DO_NOTHING, related_name='+', to...
import os import sys sys.path.append(os.path.join(sys.path[0], '../')) from smart_pipeline import Pipeline data = [1,2,3,4,5] # Define a data function def onlyOdd(item): return False if item%2==0 else True pl = Pipeline() # Adds function into pipeline pl.addDataPipe(onlyOdd) res = pl(data) for item in res: ...
[ "os.path.join", "smart_pipeline.Pipeline" ]
[((220, 230), 'smart_pipeline.Pipeline', 'Pipeline', ([], {}), '()\n', (228, 230), False, 'from smart_pipeline import Pipeline\n'), ((38, 70), 'os.path.join', 'os.path.join', (['sys.path[0]', '"""../"""'], {}), "(sys.path[0], '../')\n", (50, 70), False, 'import os\n')]
import tensorflow as tf batch_size = 4 feature_num = 3 csv1 = [ "harden|james|curry", "wrestbrook|harden|durant", "paul|towns", ] csv2 = [ "curry", "wrestbrook|harden|durant", "paul|towns", ] csv3 = [ "harden|james|curry", "durant", "paul|towns", ] csv4 = [ "wrestbrook|harden|durant", "wrestbrook|harden|...
[ "tensorflow.contrib.layers.sparse_column_with_hash_bucket", "tensorflow.global_variables_initializer", "tensorflow.contrib.layers.embedding_column", "tensorflow.Session", "tensorflow.string_split", "tensorflow.feature_column.input_layer", "tensorflow.placeholder", "tensorflow.shape", "pdb.set_trace"...
[((392, 450), 'tensorflow.placeholder', 'tf.placeholder', ([], {'shape': '[None, feature_num]', 'dtype': 'tf.string'}), '(shape=[None, feature_num], dtype=tf.string)\n', (406, 450), True, 'import tensorflow as tf\n'), ((464, 601), 'tensorflow.contrib.layers.sparse_column_with_hash_bucket', 'tf.contrib.layers.sparse_col...
""" Commands for index operations """ import os import re import sys import time from typing import List from zipfile import ZipFile, ZIP_DEFLATED from datetime import datetime as DateTime import click from mnotes.environment import MnoteEnvironment, pass_env, echo_line, save_global_index_data from mnotes.notes.in...
[ "mnotes.environment.save_global_index_data", "mnotes.environment.echo_line", "zipfile.ZipFile", "click.argument", "click.echo", "click.style", "datetime.datetime.now", "time.time", "os.path.relpath", "click.group", "os.path.join", "sys.exit", "re.compile" ]
[((413, 439), 're.compile', 're.compile', (['"""[^a-z0-9\\\\-]"""'], {}), "('[^a-z0-9\\\\-]')\n", (423, 439), False, 'import re\n'), ((443, 497), 'click.group', 'click.group', ([], {'name': '"""index"""', 'invoke_without_command': '(True)'}), "(name='index', invoke_without_command=True)\n", (454, 497), False, 'import c...
import numpy as np import os import torch import torch.nn as nn import pytorch_lightning as pl from data.VOCdevkit.vocdata import VOCDataset from torch.utils.data import DataLoader from torchvision.transforms import Compose, Resize, ToTensor, Normalize, GaussianBlur from torchvision.transforms.functional import Interp...
[ "numpy.delete", "torch.gather", "torch.utils.data.DataLoader", "os.path.join", "numpy.sum", "torch.argsort", "torchvision.transforms.ToTensor", "torch.cumsum", "numpy.max", "torch.nn.functional.interpolate", "torchvision.transforms.GaussianBlur", "torchvision.transforms.Normalize", "torch.no...
[((1354, 1483), 'torch.utils.data.DataLoader', 'DataLoader', (['self.dataset'], {'batch_size': 'attn_batch_size', 'shuffle': '(False)', 'num_workers': 'num_workers', 'drop_last': '(True)', 'pin_memory': '(True)'}), '(self.dataset, batch_size=attn_batch_size, shuffle=False,\n num_workers=num_workers, drop_last=True, ...
import socket, threading server = socket.socket(socket.AF_INET, socket.SOCK_STREAM) host = "127.0.0.1" port = 9090 server.bind((host, port)) server.listen(5) clients = list() end = list() def get(): while True: client, addr = server.accept() clients.append(client) print(f'сервер подключе...
[ "threading.Thread", "socket.socket" ]
[((35, 84), 'socket.socket', 'socket.socket', (['socket.AF_INET', 'socket.SOCK_STREAM'], {}), '(socket.AF_INET, socket.SOCK_STREAM)\n', (48, 84), False, 'import socket, threading\n'), ((1215, 1262), 'threading.Thread', 'threading.Thread', ([], {'target': 'send_mes', 'name': '"""input"""'}), "(target=send_mes, name='inp...
# Autoencoder using convolutional layers # Dataset : MNIST # Requires : PIL, matplotlib # Inspired by https://pytorch.org/tutorials/beginner/blitz/cifar10_tutorial.html # To compress data : net.encode(data) # To decompress data : net.decode(data) # To mutate data : net(data) import os import numpy as np import matplot...
[ "torch.nn.MSELoss", "matplotlib.pyplot.show", "torch.nn.ConvTranspose2d", "torch.utils.data.DataLoader", "torch.load", "torch.nn.Conv2d", "os.path.exists", "numpy.transpose", "torchvision.transforms.ToPILImage", "torchvision.utils.make_grid", "torch.cuda.is_available", "torch.nn.Linear", "to...
[((2274, 2295), 'torchvision.transforms.ToTensor', 'transforms.ToTensor', ([], {}), '()\n', (2293, 2295), False, 'from torchvision import transforms\n'), ((2306, 2373), 'torchvision.datasets.MNIST', 'MNIST', ([], {'root': 'dataset_dir', 'train': '(True)', 'download': '(True)', 'transform': 'trans'}), '(root=dataset_dir...
# Monocyte - Search and Destroy unwanted AWS Resources relentlessly. # Copyright 2015 Immobilien Scout GmbH # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licens...
[ "monocyte.handler.Resource", "boto.ec2.regions", "boto.ec2.connect_to_region" ]
[((2322, 2360), 'boto.ec2.connect_to_region', 'ec2.connect_to_region', (['resource.region'], {}), '(resource.region)\n', (2343, 2360), False, 'from boto import ec2\n'), ((4121, 4159), 'boto.ec2.connect_to_region', 'ec2.connect_to_region', (['resource.region'], {}), '(resource.region)\n', (4142, 4159), False, 'from boto...
from pathlib import Path import requests import string import urllib import random import json import sys def acapyla(quote, voicename="willfromafar"): try: voiceid = "enu_" + voicename + "_22k_ns.bvcu" except IndexError: voiceid = "enu_willfromafar_22k_ns.bvcu" letters = string.ascii_lower...
[ "requests.post", "random.choice", "pathlib.Path.home" ]
[((545, 587), 'requests.post', 'requests.post', ([], {'url': 'noncerl', 'data': 'noncedata'}), '(url=noncerl, data=noncedata)\n', (558, 587), False, 'import requests\n'), ((1074, 1133), 'requests.post', 'requests.post', ([], {'url': 'synthrl', 'data': 'synthdata', 'headers': 'headers'}), '(url=synthrl, data=synthdata, ...
# -*- coding: utf-8 -*- """ This module """ import attr import typing from ..core.model import ( Property, Resource, Tag, GetAtt, TypeHint, TypeCheck, ) from ..core.constant import AttrMeta #--- Property declaration --- @attr.s class PropJobDefinitionAuthorizationConfig(Property): """ AWS Object Type =...
[ "attr.validators.instance_of" ]
[((6867, 6924), 'attr.validators.instance_of', 'attr.validators.instance_of', (['TypeCheck.intrinsic_str_type'], {}), '(TypeCheck.intrinsic_str_type)\n', (6894, 6924), False, 'import attr\n'), ((7264, 7296), 'attr.validators.instance_of', 'attr.validators.instance_of', (['int'], {}), '(int)\n', (7291, 7296), False, 'im...
#!/usr/bin/env python # coding: utf-8 # __buildin__ modules import smtplib from email.mime.text import MIMEText from monitor.utils.settings import EMAIL_SERVER from monitor.utils.settings import EMAIL_PORT from monitor.utils.settings import EMAIL_USER from monitor.utils.settings import EMAIL_PASS from monitor.utils.s...
[ "email.mime.text.MIMEText", "smtplib.SMTP_SSL" ]
[((562, 597), 'email.mime.text.MIMEText', 'MIMEText', (['content', '"""plain"""', '"""utf-8"""'], {}), "(content, 'plain', 'utf-8')\n", (570, 597), False, 'from email.mime.text import MIMEText\n'), ((708, 726), 'smtplib.SMTP_SSL', 'smtplib.SMTP_SSL', ([], {}), '()\n', (724, 726), False, 'import smtplib\n')]
# Copyright (c) 2018 <NAME> # Copyright (c) 2018 <NAME> # # Distributed under the Boost Software License, Version 1.0. (See accompanying # file LICENSE_1_0.txt or copy at http://www.boost.org/LICENSE_1_0.txt) from phylanx import Phylanx import numpy as np @Phylanx def foo(): local_a = np.array((2, 1)) lo...
[ "numpy.array" ]
[((297, 313), 'numpy.array', 'np.array', (['(2, 1)'], {}), '((2, 1))\n', (305, 313), True, 'import numpy as np\n'), ((364, 381), 'numpy.array', 'np.array', (['(57, 1)'], {}), '((57, 1))\n', (372, 381), True, 'import numpy as np\n')]
import importlib import json import os import sys from pprint import pprint as pp import pytest import utils as util from ixload import IxLoadTestSettings as TestSettings from ixload import IxLoadUtils as IxLoadUtils from ixload import IxRestUtils as IxRestUtils from ixnetwork_restpy import SessionAssistant from ixnet...
[ "ixload.IxRestUtils.getConnection", "ixload.IxLoadTestSettings.IxLoadTestSettings", "importlib.import_module", "os.path.dirname", "pytest.fixture", "sys.path.insert" ]
[((467, 498), 'sys.path.insert', 'sys.path.insert', (['(0)', 'targets_dir'], {}), '(0, targets_dir)\n', (482, 498), False, 'import sys\n'), ((502, 533), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""session"""'}), "(scope='session')\n", (516, 533), False, 'import pytest\n'), ((722, 771), 'pytest.fixture', 'pyt...
import matplotlib.pyplot as plt import matplotlib import numpy as np import datetime from dateutil.tz import tzutc def plot_water_levels(station, dates, levels): """Task 2E: Plots water level against time""" #Assign variables range_high = [station.typical_range[1]]*len(dates) range_low = [station.ty...
[ "matplotlib.pyplot.title", "numpy.poly1d", "matplotlib.pyplot.show", "matplotlib.pyplot.plot", "numpy.polyfit", "matplotlib.pyplot.legend", "matplotlib.pyplot.xticks", "numpy.linspace", "matplotlib.dates.date2num", "matplotlib.pyplot.ylabel", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.tigh...
[((363, 407), 'matplotlib.pyplot.plot', 'plt.plot', (['dates', 'levels'], {'label': '"""Water Level"""'}), "(dates, levels, label='Water Level')\n", (371, 407), True, 'import matplotlib.pyplot as plt\n'), ((412, 461), 'matplotlib.pyplot.plot', 'plt.plot', (['dates', 'range_high'], {'label': '"""Typical High"""'}), "(da...
from app import database def get_single_json_entity(entity_query): query_result_proxy = database.session.execute(entity_query) database.session.commit() row_proxies = [r for r in query_result_proxy] if len(row_proxies) == 1: json_entity = {k: v for k, v in row_proxies[0].items()} else: ...
[ "app.database.session.commit", "app.database.session.execute" ]
[((94, 132), 'app.database.session.execute', 'database.session.execute', (['entity_query'], {}), '(entity_query)\n', (118, 132), False, 'from app import database\n'), ((137, 162), 'app.database.session.commit', 'database.session.commit', ([], {}), '()\n', (160, 162), False, 'from app import database\n')]
#!/usr/bin/env python # -*- coding: utf-8 -*- # Copyright (c) 2020 Huawei Device Co., Ltd. # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # #...
[ "os.path.abspath", "os.path.split", "os.path.join", "argparse.ArgumentParser" ]
[((1913, 1938), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (1936, 1938), False, 'import argparse\n'), ((2244, 2269), 'os.path.split', 'os.path.split', (['args.input'], {}), '(args.input)\n', (2257, 2269), False, 'import os\n'), ((2288, 2316), 'os.path.abspath', 'os.path.abspath', (['args.ou...
from basicts.utils.registry import SCALER_REGISTRY """ data normalization and re-normalization """ # ====================================== re-normalizations ====================================== # @SCALER_REGISTRY.register() def re_max_min_normalization(x, **kwargs): _min, _max = kwargs['min'], kwargs['max'] ...
[ "basicts.utils.registry.SCALER_REGISTRY.register" ]
[((202, 228), 'basicts.utils.registry.SCALER_REGISTRY.register', 'SCALER_REGISTRY.register', ([], {}), '()\n', (226, 228), False, 'from basicts.utils.registry import SCALER_REGISTRY\n'), ((393, 419), 'basicts.utils.registry.SCALER_REGISTRY.register', 'SCALER_REGISTRY.register', ([], {}), '()\n', (417, 419), False, 'fro...
from jinja2 import Environment, PackageLoader, select_autoescape import yaml import json import pkg_resources import os env = Environment( loader=PackageLoader('kite_metrics', 'schemas'), ) cache = {} def _schema_exists(filename): return pkg_resources.resource_exists('kite_metrics', 'schemas/{}'.format(file...
[ "jinja2.PackageLoader", "json.loads" ]
[((152, 192), 'jinja2.PackageLoader', 'PackageLoader', (['"""kite_metrics"""', '"""schemas"""'], {}), "('kite_metrics', 'schemas')\n", (165, 192), False, 'from jinja2 import Environment, PackageLoader, select_autoescape\n'), ((1501, 1521), 'json.loads', 'json.loads', (['rendered'], {}), '(rendered)\n', (1511, 1521), Fa...
############################################################################### # Copyright (c) 2017-2020 <NAME> (Cisco Systems), # # <NAME> (Cisco Systems), <NAME> (Cisco Systems) and others # # # # All rights r...
[ "base.utils.configuration.Configuration" ]
[((1360, 1375), 'base.utils.configuration.Configuration', 'Configuration', ([], {}), '()\n', (1373, 1375), False, 'from base.utils.configuration import Configuration\n')]
# -*- coding: utf-8 -*- # Generated by Django 1.11.17 on 2020-11-02 11:30 from __future__ import unicode_literals from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('activities', '0027_contributionvalue'), ('time_bas...
[ "django.db.migrations.RemoveField", "django.db.models.OneToOneField", "django.db.models.DateField" ]
[((2833, 2904), 'django.db.migrations.RemoveField', 'migrations.RemoveField', ([], {'model_name': '"""application"""', 'name': '"""current_period"""'}), "(model_name='application', name='current_period')\n", (2855, 2904), False, 'from django.db import migrations, models\n'), ((508, 685), 'django.db.models.OneToOneField...
from __future__ import annotations from amulet.world_interface.chunk.interfaces.leveldb.leveldb_14.interface import ( LevelDB14Interface, ) class LevelDB15Interface(LevelDB14Interface): def __init__(self): LevelDB14Interface.__init__(self) self.features["chunk_version"] = 15 INTERFACE_CLAS...
[ "amulet.world_interface.chunk.interfaces.leveldb.leveldb_14.interface.LevelDB14Interface.__init__" ]
[((225, 258), 'amulet.world_interface.chunk.interfaces.leveldb.leveldb_14.interface.LevelDB14Interface.__init__', 'LevelDB14Interface.__init__', (['self'], {}), '(self)\n', (252, 258), False, 'from amulet.world_interface.chunk.interfaces.leveldb.leveldb_14.interface import LevelDB14Interface\n')]
#!/usr/bin/env python3 # -*- coding: utf-8 -*- # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed t...
[ "jobson_systemtests.run", "argparse.ArgumentParser" ]
[((727, 789), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Run jobson system tests"""'}), "(description='Run jobson system tests')\n", (750, 789), False, 'import argparse\n'), ((1446, 1573), 'jobson_systemtests.run', 'jobson_systemtests.run', ([], {'specs_dir': 'args.specs_dir', 'host'...
from discord.ext import commands import discord import yaml import os class Bot(commands.Bot): async def get_prefix(self, bot, message): """Fetches the current prefix in the guild from the database.""" return commands.when_mentioned_or("onyx ") def __init__(self): """Initialize the bot...
[ "discord.ext.commands.when_mentioned_or", "discord.Intents.default", "yaml.safe_load", "os.getenv" ]
[((728, 746), 'os.getenv', 'os.getenv', (['"""TOKEN"""'], {}), "('TOKEN')\n", (737, 746), False, 'import os\n'), ((230, 265), 'discord.ext.commands.when_mentioned_or', 'commands.when_mentioned_or', (['"""onyx """'], {}), "('onyx ')\n", (256, 265), False, 'from discord.ext import commands\n'), ((442, 459), 'yaml.safe_lo...
import urllib.parse from django.shortcuts import render #from ckan_model import stub as ckan from ckan_model import production as ckan #from ..ckan_model import production as ckan import django.http def index(request: django.http.HttpRequest): payload = {} google = ckan.Search() payload['top_tags'] = ...
[ "django.shortcuts.render", "ckan_model.production.Search" ]
[((280, 293), 'ckan_model.production.Search', 'ckan.Search', ([], {}), '()\n', (291, 293), True, 'from ckan_model import production as ckan\n'), ((349, 395), 'django.shortcuts.render', 'render', (['request', '"""index.html"""'], {'context': 'payload'}), "(request, 'index.html', context=payload)\n", (355, 395), False, '...
import numpy as np from scipy.integrate import quad import matplotlib.pyplot as plt def Redshift(n0, n1, n2, z1=3.6, z=np.linspace(0,10,num=1001)): Rlow = np.power((1.0 + z), n1) Rhigh = np.power((1.0 + z), n2) rbrk = np.power((1.0 + z1), n1 - n2) R = Rlow * (z <= z1) + rbrk * Rhigh * (z > z1) R *= n0 / R[0] ret...
[ "matplotlib.pyplot.show", "matplotlib.pyplot.plot", "scipy.integrate.quad", "numpy.power", "numpy.savetxt", "numpy.zeros", "numpy.linspace", "matplotlib.pyplot.ylabel", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.grid", "numpy.concatenate" ]
[((365, 385), 'matplotlib.pyplot.plot', 'plt.plot', (['z', 'R', '"""-k"""'], {}), "(z, R, '-k')\n", (373, 385), True, 'import matplotlib.pyplot as plt\n'), ((384, 401), 'matplotlib.pyplot.xlabel', 'plt.xlabel', (['"""$z$"""'], {}), "('$z$')\n", (394, 401), True, 'import matplotlib.pyplot as plt\n'), ((403, 434), 'matpl...
from waio.keyboard.reply import QuickReplyContentText, QuickReply, KeyboardButton, QuickReplyContentImage from callback import callback_reply_keyboard def generate_keyboard_place(): kb_content = QuickReplyContentText( header="Куда вы сегодня хотите сходить?", text="Выберите из предложенного списка...
[ "waio.keyboard.reply.QuickReplyContentText", "waio.keyboard.reply.KeyboardButton", "callback.callback_reply_keyboard.new", "waio.keyboard.reply.QuickReplyContentImage" ]
[((201, 322), 'waio.keyboard.reply.QuickReplyContentText', 'QuickReplyContentText', ([], {'header': '"""Куда вы сегодня хотите сходить?"""', 'text': '"""Выберите из предложенного списка"""', 'caption': '""""""'}), "(header='Куда вы сегодня хотите сходить?', text=\n 'Выберите из предложенного списка', caption='')\n",...
#!/usr/bin/env python # Python script for fast text file searching using keyword index on disk. # # Author: <NAME> <<EMAIL>> # Last Change: November 1, 2015 # URL: http://peterodding.com/code/vim/notes/ # License: MIT # # This Python script can be used by the notes.vim plug-in to perform fast # keyword searches in the...
[ "codecs.encode", "pickle.dump", "getopt.getopt", "logging.basicConfig", "os.path.join", "codecs.decode", "os.path.isdir", "Levenshtein.distance", "os.walk", "time.time", "re.findall", "pickle.load", "os.path.getmtime", "sys.exit", "os.path.expanduser", "fnmatch.fnmatch", "logging.get...
[((2271, 2313), 'os.path.expanduser', 'os.path.expanduser', (['"""~/notes-index.pickle"""'], {}), "('~/notes-index.pickle')\n", (2289, 2313), False, 'import os\n'), ((2315, 2355), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.DEBUG'}), '(level=logging.DEBUG)\n', (2334, 2355), False, 'import logg...
"""Implementation of the FunctionInjectionWalker class.""" from inspect import ( signature as inspect_signature) from types import ( CodeType, FunctionType) from typing import ( Iterator, Optional) from .abstract_injection_walker import ( AbstractInjectionWalker) from .code_object_injection_w...
[ "types.FunctionType", "inspect.signature" ]
[((6470, 6496), 'types.FunctionType', 'FunctionType', (['code_obj', '{}'], {}), '(code_obj, {})\n', (6482, 6496), False, 'from types import CodeType, FunctionType\n'), ((6524, 6547), 'inspect.signature', 'inspect_signature', (['func'], {}), '(func)\n', (6541, 6547), True, 'from inspect import signature as inspect_signa...
# # Copyright (c) nexB Inc. and others. All rights reserved. # http://nexb.com and https://github.com/nexB/tracecode-toolkit/ # The TraceCode software is licensed under the Apache License version 2.0. # Data generated with TraceCode require an acknowledgment. # TraceCode is a trademark of nexB Inc. # # You may not use ...
[ "tracecode.pathutils.common_path_suffix", "tracecode.pathutils.common_path_prefix" ]
[((1673, 1721), 'tracecode.pathutils.common_path_prefix', 'pathutils.common_path_prefix', (['"""/a/b/c"""', '"""/a/b/c"""'], {}), "('/a/b/c', '/a/b/c')\n", (1701, 1721), False, 'from tracecode import pathutils\n'), ((1814, 1860), 'tracecode.pathutils.common_path_prefix', 'pathutils.common_path_prefix', (['"""/a/b/c"""'...
# Generated by Django 3.2.8 on 2021-10-24 01:32 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('dim_sum', '0001_initial'), ] operations = [ migrations.AddField( model_name='dimsum', name='history', fi...
[ "django.db.models.TextField" ]
[((324, 371), 'django.db.models.TextField', 'models.TextField', ([], {'default': '"""Write summary here."""'}), "(default='Write summary here.')\n", (340, 371), False, 'from django.db import migrations, models\n')]
# Copyright (c) 2020-2021 CRS4 # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribu...
[ "tests.utils.pick_and_register_workflow", "logging.getLogger" ]
[((1196, 1215), 'logging.getLogger', 'logging.getLogger', ([], {}), '()\n', (1213, 1215), False, 'import logging\n'), ((1302, 1357), 'tests.utils.pick_and_register_workflow', 'utils.pick_and_register_workflow', (['user1', 'valid_workflow'], {}), '(user1, valid_workflow)\n', (1334, 1357), False, 'from tests import utils...
# ============================================= # -*- coding: utf-8 -*- # @Time : 2020/5/14 上午10:50 # @Author : xiao9616 # @Email : <EMAIL> # @File : BaseModel.py # @Software: PyCharm # ============================================ import logging import tensorflow as tf import os f...
[ "tensorflow.math.ceil", "logging.basicConfig", "tensorflow.keras.metrics.Mean", "tensorflow.print", "tensorflow.keras.optimizers.schedules.ExponentialDecay", "tensorflow.config.experimental.set_memory_growth", "logging.info", "tensorflow.keras.optimizers.Adam", "tensorflow.debugging.set_log_device_p...
[((454, 635), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.DEBUG', 'format': '"""%(asctime)s %(levelname)s %(message)s"""', 'datefmt': '"""%a, %d %b %Y %H:%M:%S"""', 'filename': '"""./yolo4/logs/train.log"""', 'filemode': '"""w+"""'}), "(level=logging.DEBUG, format=\n '%(asctime)s %(levelnam...
import networkx as nx import sys from collections import defaultdict sys.setrecursionlimit(100000000) def direct_paths(lines): world = nx.Graph() for row,line in enumerate(lines): for col,obj in enumerate(line): if obj != '.': continue if line[col - 1] == '.': ...
[ "networkx.dijkstra_path_length", "networkx.dijkstra_path", "collections.defaultdict", "networkx.Graph", "sys.setrecursionlimit" ]
[((70, 102), 'sys.setrecursionlimit', 'sys.setrecursionlimit', (['(100000000)'], {}), '(100000000)\n', (91, 102), False, 'import sys\n'), ((1461, 1538), 'networkx.dijkstra_path', 'nx.dijkstra_path', (['W', "portal_connections['AA'][0]", "portal_connections['ZZ'][0]"], {}), "(W, portal_connections['AA'][0], portal_conne...