code
stringlengths
22
1.05M
apis
listlengths
1
3.31k
extract_api
stringlengths
75
3.25M
from django.contrib import admin from apps.iotdb_cloud_core.models import IoTDBRelease admin.site.register(IoTDBRelease)
[ "django.contrib.admin.site.register" ]
[((89, 122), 'django.contrib.admin.site.register', 'admin.site.register', (['IoTDBRelease'], {}), '(IoTDBRelease)\n', (108, 122), False, 'from django.contrib import admin\n')]
import copy import numpy as np PXL2CM = 0.035277778 def print_formatted_stats(stats): """ Print formatted results for result tables """ print("& {:.2f} & {:.2f} & {:.2f} & {:.2f} \\\\" .format(np.mean(stats['tracked_until_end_ratio']), np...
[ "copy.deepcopy", "numpy.abs", "numpy.sum", "numpy.median", "numpy.asarray", "numpy.mean" ]
[((1224, 1252), 'numpy.asarray', 'np.asarray', (['alignment_errors'], {}), '(alignment_errors)\n', (1234, 1252), True, 'import numpy as np\n'), ((1280, 1304), 'numpy.abs', 'np.abs', (['alignment_errors'], {}), '(alignment_errors)\n', (1286, 1304), True, 'import numpy as np\n'), ((1327, 1354), 'numpy.asarray', 'np.asarr...
from pathlib import Path from torchvision.datasets import VisionDataset import numpy as np from PIL import Image class MHPv1(VisionDataset): """ MHP dataset : Multi-Human Parsing V1은 human parsing 만 있고, v2는 pose 포함 https://github.com/ZhaoJ9014/Multi-Human-Parsing or https://lv-mhp.github.io/ ...
[ "pathlib.Path", "PIL.Image.open" ]
[((2751, 2761), 'pathlib.Path', 'Path', (['root'], {}), '(root)\n', (2755, 2761), False, 'from pathlib import Path\n'), ((3223, 3240), 'PIL.Image.open', 'Image.open', (['fname'], {}), '(fname)\n', (3233, 3240), False, 'from PIL import Image\n'), ((3306, 3319), 'PIL.Image.open', 'Image.open', (['f'], {}), '(f)\n', (3316...
from sys import stdin def count(S, m, n): tabla = [[0 for x in range(m)] for x in range(n+1)] for i in range(m): tabla[0][i] = 1 for i in range(1, n+1): for j in range(m): x = tabla[i - S[j]][j] if i-S[j] >= 0 else 0 y = tabla[i][j-1] if j >= 1 else 0 ...
[ "sys.stdin.readline" ]
[((404, 420), 'sys.stdin.readline', 'stdin.readline', ([], {}), '()\n', (418, 420), False, 'from sys import stdin\n'), ((453, 469), 'sys.stdin.readline', 'stdin.readline', ([], {}), '()\n', (467, 469), False, 'from sys import stdin\n')]
from django.db import models from django.utils import timezone from django.forms import ModelForm from django.contrib.auth.models import User from django.db.models.signals import post_save from django.dispatch import receiver from django.core.exceptions import ValidationError import secrets import os class Profile(mod...
[ "django.db.models.FileField", "django.db.models.OneToOneField", "django.db.models.TextField", "django.core.exceptions.ValidationError", "django.db.models.ForeignKey", "django.db.models.CharField", "django.dispatch.receiver", "secrets.token_urlsafe", "django.db.models.BooleanField", "django.db.mode...
[((516, 548), 'django.dispatch.receiver', 'receiver', (['post_save'], {'sender': 'User'}), '(post_save, sender=User)\n', (524, 548), False, 'from django.dispatch import receiver\n'), ((343, 395), 'django.db.models.OneToOneField', 'models.OneToOneField', (['User'], {'on_delete': 'models.CASCADE'}), '(User, on_delete=mod...
__source__ = 'https://leetcode.com/problems/k-empty-slots/' # Time: O() # Space: O() # # Description: Leetcode # 683. K Empty Slots # #There is a garden with N slots. In each slot, there is a flower. The N flowers will bloom one by one in N days. # In each day, there will be exactly one flower blooming and it will be ...
[ "unittest.main" ]
[((1422, 1437), 'unittest.main', 'unittest.main', ([], {}), '()\n', (1435, 1437), False, 'import unittest\n')]
from flask import render_template,request,redirect,url_for from . import main from ..request import get_news from ..request import get_news, get_news_articles,search_article # Views @main.route('/') def index(): ''' function that returns the index page and its data ''' #Get popular news general_ne...
[ "flask.render_template", "flask.url_for", "flask.request.args.get" ]
[((677, 707), 'flask.request.args.get', 'request.args.get', (['"""news_query"""'], {}), "('news_query')\n", (693, 707), False, 'from flask import render_template, request, redirect, url_for\n'), ((1227, 1285), 'flask.render_template', 'render_template', (['"""articles.html"""'], {'id': 'id', 'articles': 'articles'}), "...
#!/usr/bin/env python """ hostlists plugin to recursively query plugins based on type. This makes it possible to obtain lists of hosts by recursively querying multiple backends. For example: * Query dns for www.foo.com * Get a list of two hostnames back haproxy1.ny.foo.com and haproxy1.lax.foo.com. * Qu...
[ "hostlists.plugin_manager.get_plugins" ]
[((2060, 2073), 'hostlists.plugin_manager.get_plugins', 'get_plugins', ([], {}), '()\n', (2071, 2073), False, 'from hostlists.plugin_manager import get_plugins\n')]
import asyncio from PIL import Image from ..Base import Base class BiasGame(Base): def __init__(self, *args): super().__init__(*args) async def create_bias_game_image(self, first_idol_id, second_idol_id): """Uses thread pool to create bias game image to prevent IO blocking.""" # (sel...
[ "PIL.Image.open" ]
[((2858, 2918), 'PIL.Image.open', 'Image.open', (['f"""{self.ex.keys.bias_game_location}bracket8.png"""'], {}), "(f'{self.ex.keys.bias_game_location}bracket8.png')\n", (2868, 2918), False, 'from PIL import Image\n'), ((910, 968), 'PIL.Image.open', 'Image.open', (['f"""{self.ex.keys.bias_game_location}versus.png"""'], {...
# -*-mode: python; encoding: utf-8; test-case-name: tests.test_app-*- # ======================================================================== """ Copyright |(c)| 2017 `Dropbox, Inc.`_ .. |(c)| unicode:: u+a9 .. _`Dropbox, Inc.`: https://www.dropbox.com/ Please see the accompanying ``LICENSE`` and ``CREDIT...
[ "flask.request.form.get", "future.utils.bytes_to_native_str", "hmac.compare_digest", "flask.url_for", "os.path.join", "flask.redirect", "flask.request.args.get", "flask.request.headers.get", "flask.abort", "hashlib.sha256", "flask.render_template", "os.urandom", "hmac.new", "sqlite3.connec...
[((1619, 1640), 'flask.Flask', 'flask.Flask', (['__name__'], {}), '(__name__)\n', (1630, 1640), False, 'import flask\n'), ((1883, 1918), 'flask.session.get', 'flask.session.get', (['_SESSION_USER_ID'], {}), '(_SESSION_USER_ID)\n', (1900, 1918), False, 'import flask\n'), ((5429, 5477), 'flask.session.pop', 'flask.sessio...
from PIL import ImageFont class Letter: """ letter class- each letter is one of these objects, and is rendered in order. """ def __init__(self,char,size,font,color = (255,255,255,255),b=False,i=False,u=False): """ char: character. size: size of letter. font: PIL truetype font obj...
[ "PIL.ImageFont.truetype" ]
[((572, 602), 'PIL.ImageFont.truetype', 'ImageFont.truetype', (['font', 'size'], {}), '(font, size)\n', (590, 602), False, 'from PIL import ImageFont\n')]
''' ############################################################################# # # M A I N L O O P & P R I M A R Y F U N C T I O N S # ############################################################################# ''' import getopt import os import random import signal import sys try: import numpy...
[ "getopt.getopt", "os.path.join", "rpn.debug.typename", "os.path.isfile", "random.seed", "signal.signal", "sys.setrecursionlimit", "os.path.expanduser", "sys.exit" ]
[((1384, 1396), 'sys.exit', 'sys.exit', (['(64)'], {}), '(64)\n', (1392, 1396), False, 'import sys\n'), ((1646, 1673), 'sys.setrecursionlimit', 'sys.setrecursionlimit', (['(2000)'], {}), '(2000)\n', (1667, 1673), False, 'import sys\n'), ((1697, 1710), 'random.seed', 'random.seed', ([], {}), '()\n', (1708, 1710), False,...
"""Main application initilization.""" import os.path from flask import Flask from flask_sqlalchemy import SQLAlchemy from flask_login import LoginManager BOOK_PATH = os.path.normpath(os.path.join(os.path.dirname(__file__), '..', 'books')) # Make sure to add your own secret key in config.py SECRET_KEY = "<KEY>"...
[ "flask_sqlalchemy.SQLAlchemy", "flask_login.LoginManager", "flask.Flask" ]
[((530, 545), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (535, 545), False, 'from flask import Flask\n'), ((668, 683), 'flask_sqlalchemy.SQLAlchemy', 'SQLAlchemy', (['app'], {}), '(app)\n', (678, 683), False, 'from flask_sqlalchemy import SQLAlchemy\n'), ((733, 747), 'flask_login.LoginManager', 'LoginM...
# Copyright 2021 Arm Inc. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agree...
[ "tensorflow.keras.layers.Conv2D", "tensorflow.pad", "tensorflow.eye", "numpy.zeros", "models.quantize_utils.compute_ranges", "tensorflow.transpose", "tensorflow.constant", "tensorflow.Variable", "tensorflow.nn.conv2d", "tensorflow.keras.Sequential", "tensorflow.keras.layers.Layer.__init__", "t...
[((2255, 2282), 'tensorflow.keras.Sequential', 'tf.keras.Sequential', (['layers'], {}), '(layers)\n', (2274, 2282), True, 'import tensorflow as tf\n'), ((2789, 2825), 'tensorflow.keras.layers.Layer.__init__', 'tf.keras.layers.Layer.__init__', (['self'], {}), '(self)\n', (2819, 2825), True, 'import tensorflow as tf\n'),...
import random from LogicClass.MonteCarloClass.ArborescenteTree import ArborescenteTree class MonteCarloMove: def __init__(self, isInspector, isPhantom, numberOfRoom): self.tree = ArborescenteTree() self.isPhantom = isPhantom self.isInspector = isInspector self.numberOfRoom = number...
[ "random.randint", "LogicClass.MonteCarloClass.ArborescenteTree.ArborescenteTree" ]
[((193, 211), 'LogicClass.MonteCarloClass.ArborescenteTree.ArborescenteTree', 'ArborescenteTree', ([], {}), '()\n', (209, 211), False, 'from LogicClass.MonteCarloClass.ArborescenteTree import ArborescenteTree\n'), ((2441, 2481), 'random.randint', 'random.randint', (['(0)', '(self.numberOfRoom - 1)'], {}), '(0, self.num...
""" Tests for the Paver commands for updating test databases and its utility methods """ import os import shutil import tarfile from tempfile import mkdtemp from unittest import TestCase from unittest.mock import call, patch, Mock import boto from pavelib import database from pavelib.utils import db_utils from pave...
[ "unittest.mock.patch.object", "pavelib.utils.db_utils.extract_files_from_zip", "unittest.mock.Mock", "pavelib.database.update_local_bokchoy_db_from_s3", "unittest.mock.patch", "unittest.mock.call", "os.path.isfile", "tempfile.mkdtemp", "tarfile.open", "os.path.join", "boto.s3.key.Key" ]
[((530, 580), 'unittest.mock.patch', 'patch', (['"""pavelib.utils.db_utils.verify_files_exist"""'], {}), "('pavelib.utils.db_utils.verify_files_exist')\n", (535, 580), False, 'from unittest.mock import call, patch, Mock\n'), ((2404, 2432), 'unittest.mock.patch.object', 'patch.object', (['db_utils', '"""sh"""'], {}), "(...
from functools import lru_cache from typing import Optional from pygame.freetype import get_default_font, SysFont font_cache = {} @lru_cache(100) def get_font(fontname: Optional[str] = None, size: int = 12, bold: bool = False, italic: bool = False): if fontname is None: fontname = get_default_font() ...
[ "pygame.freetype.get_default_font", "functools.lru_cache", "pygame.freetype.SysFont" ]
[((136, 150), 'functools.lru_cache', 'lru_cache', (['(100)'], {}), '(100)\n', (145, 150), False, 'from functools import lru_cache\n'), ((329, 378), 'pygame.freetype.SysFont', 'SysFont', (['fontname', 'size'], {'bold': 'bold', 'italic': 'italic'}), '(fontname, size, bold=bold, italic=italic)\n', (336, 378), False, 'from...
import requests def get_event(user_key, latitude, longitude): url = "http://api.eventful.com/json/events/search?" url += "&app_key=" + user_key url += "&date=Future" #+ date url += "&page_size=100" url += "&sort_order=popularity" url += "&sort_direction=descending" url += "&q=music" u...
[ "requests.get" ]
[((417, 434), 'requests.get', 'requests.get', (['url'], {}), '(url)\n', (429, 434), False, 'import requests\n')]
from typing import List, Text, Tuple import logging import re import lxml from .. import utils MAIN_PAGE = "Wikiquote:Accueil" logger = logging.getLogger(__name__) logger.addHandler(logging.NullHandler()) def extract_quotes(tree: lxml.html.HtmlElement, max_quotes: int) -> List[Text]: # French wiki uses a "ci...
[ "re.search", "logging.getLogger", "logging.NullHandler" ]
[((141, 168), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (158, 168), False, 'import logging\n'), ((187, 208), 'logging.NullHandler', 'logging.NullHandler', ([], {}), '()\n', (206, 208), False, 'import logging\n'), ((1088, 1118), 're.search', 're.search', (['"""«(.+?)»(.+)"""', 'line']...
# Generated by Django 3.2.6 on 2022-02-14 13:25 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ("database", "0060_set_ordering_on_tablewebhook_models"), ] operations = [ migrations.AlterField( model_name="numberfield", ...
[ "django.db.models.IntegerField" ]
[((374, 563), 'django.db.models.IntegerField', 'models.IntegerField', ([], {'choices': "[(0, '1'), (1, '1.0'), (2, '1.00'), (3, '1.000'), (4, '1.0000'), (5, '1.00000')\n ]", 'default': '(0)', 'help_text': '"""The amount of digits allowed after the point."""'}), "(choices=[(0, '1'), (1, '1.0'), (2, '1.00'), (3, '1.00...
from IPython import display from torch.utils.data import DataLoader from torchvision import transforms, datasets import os import tensorflow as tf from tensorflow import nn, layers from tensorflow.contrib import layers as clayers import numpy as np import errno import torchvision.utils as vutils from tensorboardX impo...
[ "tensorflow.trainable_variables", "tensorflow.contrib.layers.flatten", "tensorflow.reshape", "tensorflow.zeros_like", "torchvision.datasets.CIFAR10", "matplotlib.pyplot.figure", "numpy.random.normal", "tensorflow.layers.conv2d_transpose", "torchvision.transforms.Normalize", "tensorflow.nn.leaky_re...
[((989, 1045), 'torch.utils.data.DataLoader', 'DataLoader', (['dataset'], {'batch_size': 'batch_size', 'shuffle': '(True)'}), '(dataset, batch_size=batch_size, shuffle=True)\n', (999, 1045), False, 'from torch.utils.data import DataLoader\n'), ((3937, 3993), 'tensorflow.placeholder', 'tf.placeholder', (['tf.float32'], ...
# -*- coding: UTF-8 -*- # Copyright 2013-2017 <NAME> # License: BSD (see file COPYING for details) """See :doc:`/specs/vat`. .. autosummary:: :toctree: utils .. fixtures.novat fixtures.euvatrates """ from django.utils.translation import ugettext_lazy as _ from lino.api import ad import six class Plugin(a...
[ "django.utils.translation.ugettext_lazy" ]
[((440, 448), 'django.utils.translation.ugettext_lazy', '_', (['"""VAT"""'], {}), "('VAT')\n", (441, 448), True, 'from django.utils.translation import ugettext_lazy as _\n')]
from flask import Blueprint admin = Blueprint('admin', __name__, url_prefix='/admin', template_folder='templates')
[ "flask.Blueprint" ]
[((38, 116), 'flask.Blueprint', 'Blueprint', (['"""admin"""', '__name__'], {'url_prefix': '"""/admin"""', 'template_folder': '"""templates"""'}), "('admin', __name__, url_prefix='/admin', template_folder='templates')\n", (47, 116), False, 'from flask import Blueprint\n')]
# Generated by Django 2.1.5 on 2019-01-29 23:48 from django.db import migrations, models class Migration(migrations.Migration): initial = True dependencies = [ ] operations = [ migrations.CreateModel( name='Call', fields=[ ('id', models.AutoField(aut...
[ "django.db.models.CharField", "django.db.models.IntegerField", "django.db.models.AutoField" ]
[((300, 393), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (316, 393), False, 'from django.db import migrations, models\...
import csv import logging import numpy as np import datajoint as dj import pathlib import scipy.io as scio from tifffile import imread from . import InsertBuffer from .reference import ccf_ontology from . import get_schema_name schema = dj.schema(get_schema_name('ccf')) log = logging.getLogger(__name__) @schem...
[ "scipy.io.loadmat", "pathlib.Path", "numpy.where", "tifffile.imread", "datajoint.conn", "logging.getLogger" ]
[((284, 311), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (301, 311), False, 'import logging\n'), ((1964, 1982), 'tifffile.imread', 'imread', (['stack_path'], {}), '(stack_path)\n', (1970, 1982), False, 'from tifffile import imread\n'), ((3995, 4057), 'scipy.io.loadmat', 'scio.loadmat'...
import os import discord import requests import json from dotenv import load_dotenv from discord.ext import commands from datetime import datetime load_dotenv() TOKEN = os.getenv('DISCORD_TOKEN') API_KEY = os.getenv('API_KEY') HEADERS = { "x-api-key" : API_KEY } bot = commands.Bot(command_prefix = "-") ROOT_URL...
[ "discord.Color.blurple", "dotenv.load_dotenv", "requests.get", "discord.ext.commands.Bot", "datetime.datetime.now", "os.getenv" ]
[((148, 161), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (159, 161), False, 'from dotenv import load_dotenv\n'), ((170, 196), 'os.getenv', 'os.getenv', (['"""DISCORD_TOKEN"""'], {}), "('DISCORD_TOKEN')\n", (179, 196), False, 'import os\n'), ((207, 227), 'os.getenv', 'os.getenv', (['"""API_KEY"""'], {}), "('...
import logging from common import settings from elasticapm.contrib.flask import ElasticAPM log = logging.getLogger(__name__) def configure_app(flask_app): flask_app.config.SWAGGER_UI_DOC_EXPANSION = settings.RESTPLUS_SWAGGER_UI_DOC_EXPANSION flask_app.config.RESTPLUS_VALIDATE = settings.RESTPLUS_VALIDATE ...
[ "elasticapm.contrib.flask.ElasticAPM", "logging.getLogger" ]
[((99, 126), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (116, 126), False, 'import logging\n'), ((1040, 1093), 'elasticapm.contrib.flask.ElasticAPM', 'ElasticAPM', (['flask_app'], {'logging': 'settings.APM_LOG_LEVEL'}), '(flask_app, logging=settings.APM_LOG_LEVEL)\n', (1050, 1093), Fa...
"""CLI for data preparation and processing.""" import argparse from utils import data_prep from utils import read_one_row from utils import save_input parser = argparse.ArgumentParser() parser.add_argument( "--save_row", type=int, default="0", help="Saves a single row to a file defaults to row 0", ) p...
[ "utils.save_input", "utils.data_prep", "argparse.ArgumentParser", "utils.read_one_row" ]
[((162, 187), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (185, 187), False, 'import argparse\n'), ((837, 858), 'utils.data_prep', 'data_prep', (['input_file'], {}), '(input_file)\n', (846, 858), False, 'from utils import data_prep\n'), ((865, 899), 'utils.read_one_row', 'read_one_row', (['s...
import os import sys import unittest from io import StringIO import bpy import tests.test_utils as tutils from blendernc.preferences import get_addon_preference @tutils.refresh_state def create_nodes(file, var): node_groups = bpy.data.node_groups if tutils.is_blendernc_in_nodetree(node_groups): node...
[ "io.StringIO", "bpy.ops.image.new", "os.path.abspath", "tests.test_utils.is_blendernc_in_nodetree", "unittest.TextTestRunner", "tests.test_utils.build_dict_blendernc_prop", "bpy.context.scene.frame_set", "bpy.data.node_groups.new", "unittest.defaultTestLoader.loadTestsFromTestCase", "tests.test_ut...
[((4711, 4774), 'unittest.defaultTestLoader.loadTestsFromTestCase', 'unittest.defaultTestLoader.loadTestsFromTestCase', (['Test_settings'], {}), '(Test_settings)\n', (4759, 4774), False, 'import unittest\n'), ((4851, 4864), 'sys.exit', 'sys.exit', (['ret'], {}), '(ret)\n', (4859, 4864), False, 'import sys\n'), ((262, 3...
from __future__ import print_function import os import warnings import pytest from graphene_django.utils.testing import graphql_query from graphql_jwt.settings import jwt_settings from graphql_jwt.shortcuts import get_token from neo4j.exceptions import ClientError as CypherError from neobolt.exceptions import ClientE...
[ "warnings.simplefilter", "graphene_django.utils.testing.graphql_query", "neomodel.db.set_connection", "graphql_jwt.shortcuts.get_token", "pytest.fixture", "kaffepause.users.test.factories.UserFactory", "neomodel.clear_neo4j_database", "os.environ.get", "warnings.warn", "kaffepause.accounts.test.fa...
[((612, 640), 'pytest.fixture', 'pytest.fixture', ([], {'autouse': '(True)'}), '(autouse=True)\n', (626, 640), False, 'import pytest\n'), ((3998, 4026), 'pytest.fixture', 'pytest.fixture', ([], {'autouse': '(True)'}), '(autouse=True)\n', (4012, 4026), False, 'import pytest\n'), ((4167, 4195), 'pytest.fixture', 'pytest....
# Copyright (C) 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writi...
[ "proto.Field" ]
[((749, 783), 'proto.Field', 'proto.Field', (['proto.INT32'], {'number': '(1)'}), '(proto.INT32, number=1)\n', (760, 783), False, 'import proto\n'), ((1009, 1043), 'proto.Field', 'proto.Field', (['proto.INT32'], {'number': '(1)'}), '(proto.INT32, number=1)\n', (1020, 1043), False, 'import proto\n'), ((1251, 1285), 'pro...
import numpy as np import pandas as pd import sys import os from utils import DATA_DIR class Dataset(object): def __init__(self, DATA_NAME): self.DATA_NAME = DATA_NAME print("Initializing dataset:", DATA_NAME) sys.stdout.flush() data = pd.read_csv(os.path.join(DATA_DIR, "...
[ "numpy.sum", "numpy.ones", "numpy.array", "sys.stdout.flush", "numpy.random.choice", "pandas.factorize", "os.path.join" ]
[((249, 267), 'sys.stdout.flush', 'sys.stdout.flush', ([], {}), '()\n', (265, 267), False, 'import sys\n'), ((500, 536), 'pandas.factorize', 'pd.factorize', (["data['item_id'].values"], {}), "(data['item_id'].values)\n", (512, 536), True, 'import pandas as pd\n'), ((570, 606), 'pandas.factorize', 'pd.factorize', (["dat...
import setuptools with open("README.md", "r") as fh: long_description = fh.read() setuptools.setup( name='HPexome', version='1.2.1', author="<NAME>", author_email="<EMAIL>", description="An automated tool for processing whole-exome sequencing data", long_description=long_description, l...
[ "setuptools.find_packages" ]
[((416, 442), 'setuptools.find_packages', 'setuptools.find_packages', ([], {}), '()\n', (440, 442), False, 'import setuptools\n')]
import os import random import syft as sy import pandas as pd import numpy as np from PIL import Image from tqdm import tqdm from torch import ( # pylint:disable=no-name-in-module manual_seed, stack, cat, std_mean, save, is_tensor, from_numpy, randperm, default_generator, ) from tor...
[ "numpy.isin", "albumentations.Lambda", "albumentations.RandomSunFlare", "albumentations.GaussNoise", "albumentations.Resize", "pandas.read_csv", "albumentations.RandomShadow", "torch.cat", "albumentations.RandomFog", "numpy.mean", "albumentations.Normalize", "torch.std_mean", "os.path.join",...
[((4313, 4474), 'torchvision.transforms.RandomAffine', 'transforms.RandomAffine', ([], {'degrees': 'args.rotation', 'translate': '(args.translate, args.translate)', 'scale': '(1.0 - args.scale, 1.0 + args.scale)', 'shear': 'args.shear'}), '(degrees=args.rotation, translate=(args.translate,\n args.translate), scale=(...
from math import trunc def two_decimal_places(number): factor = float(10 ** 2) return trunc(number * factor) / factor
[ "math.trunc" ]
[((96, 118), 'math.trunc', 'trunc', (['(number * factor)'], {}), '(number * factor)\n', (101, 118), False, 'from math import trunc\n')]
import warnings warnings.simplefilter("ignore", UserWarning) warnings.simplefilter("ignore", FutureWarning) import argparse import os import pandas as pd import numpy as np from torch import nn from torch.utils.data import DataLoader from tqdm import tqdm from collections import defaultdict from catalyst.utils impo...
[ "numpy.random.seed", "pandas.DataFrame.from_dict", "warnings.simplefilter", "argparse.ArgumentParser", "torch.utils.data.DataLoader", "catalyst.utils.any2device", "os.path.exists", "collections.defaultdict", "pytorch_toolbelt.utils.fs.change_extension", "os.environ.get", "pytorch_toolbelt.utils....
[((18, 62), 'warnings.simplefilter', 'warnings.simplefilter', (['"""ignore"""', 'UserWarning'], {}), "('ignore', UserWarning)\n", (39, 62), False, 'import warnings\n'), ((63, 109), 'warnings.simplefilter', 'warnings.simplefilter', (['"""ignore"""', 'FutureWarning'], {}), "('ignore', FutureWarning)\n", (84, 109), False,...
from django.db import models class Course(models.Model): created = models.DateTimeField(auto_now_add=True) title = models.CharField(max_length=100, blank=True, default='') content = models.TextField() owner = models.ForeignKey('auth.User', related_name='Course', on_delete=models.CASCADE) class M...
[ "django.db.models.CharField", "django.db.models.DateTimeField", "django.db.models.TextField", "django.db.models.ForeignKey" ]
[((73, 112), 'django.db.models.DateTimeField', 'models.DateTimeField', ([], {'auto_now_add': '(True)'}), '(auto_now_add=True)\n', (93, 112), False, 'from django.db import models\n'), ((125, 181), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(100)', 'blank': '(True)', 'default': '""""""'}), "(m...
#! /usr/bin/env python3 import sys import csv import argparse import numpy as np import pandas as pd from mll_calc.all_jobs import parent_jobs, kid_jobs def row_calcs(ext_test): if 'no' in ext_test: #db_rows = 450240 #max_jobs = 9750 db_rows = 90048 * 4 max_jobs = 978 * 4 else:...
[ "numpy.arange", "csv.writer", "sys.exit" ]
[((993, 1059), 'sys.exit', 'sys.exit', (['"""total expected jobs does not equal one of db_row lists"""'], {}), "('total expected jobs does not equal one of db_row lists')\n", (1001, 1059), False, 'import sys\n'), ((414, 443), 'numpy.arange', 'np.arange', (['(0)', 'db_rows', 'n_rows'], {}), '(0, db_rows, n_rows)\n', (42...
#!/usr/bin/env python3 import os, sys import setuptools # Get text from README.txt with open("README.md", "r") as fp: readme_text = fp.read() # Get __version__ without importing with open(os.path.join(os.path.dirname(__file__),"ndsb", "__init__.py"), "r") as f: for line in f: if line.startswith("__ver...
[ "os.path.dirname", "setuptools.setup" ]
[((383, 1072), 'setuptools.setup', 'setuptools.setup', ([], {'name': '"""ndsb"""', 'version': '__version__', 'description': '"""Collect data, turn it into static artifacts and beam it to a vault."""', 'license': '"""MIT"""', 'author': '"""<NAME>"""', 'author_email': '"""<EMAIL>"""', 'url': '"""https://github.com/dbbs-l...
# import the needed packages import pickle from sklearn import preprocessing import time from os import listdir from os.path import isfile, join from random import randint, uniform import numpy as np from matplotlib import pyplot as plt import cv2 as cv from scipy import ndimage from skimage import morphol...
[ "cv2.GaussianBlur", "numpy.random.seed", "cv2.bitwise_and", "numpy.argmax", "scipy.ndimage.binary_fill_holes", "numpy.ones", "numpy.exp", "cv2.imshow", "cv2.inRange", "numpy.unique", "pandas.DataFrame", "cv2.contourArea", "numpy.zeros_like", "random.randint", "cv2.cvtColor", "scipy.ndi...
[((622, 640), 'numpy.random.seed', 'np.random.seed', (['(26)'], {}), '(26)\n', (636, 640), True, 'import numpy as np\n'), ((1152, 1169), 'numpy.unique', 'np.unique', (['target'], {}), '(target)\n', (1161, 1169), True, 'import numpy as np\n'), ((3827, 3865), 'numpy.exp', 'np.exp', (['(-((x - mean) ** 2 / (2 * var)))'], ...
############################################################ # -*- coding: utf-8 -*- # # # # # # # #### # ## ## # ## # # # # # # # # # # # ### # # ## # ## ## # # # # # # # #### # # Python-based Tool for interaction with the 10micron mounts # GUI with PyQT5 ...
[ "os.path.basename", "os.getcwd", "copy.copy", "numpy.clip", "astrometry.transform.Transform", "numpy.interp", "operator.itemgetter", "logging.getLogger" ]
[((621, 648), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (638, 648), False, 'import logging\n'), ((727, 756), 'astrometry.transform.Transform', 'transform.Transform', (['self.app'], {}), '(self.app)\n', (746, 756), False, 'from astrometry import transform\n'), ((13679, 13807), 'numpy....
# Project: hardInfo # Author: <NAME> # Date Started: March 18, 2022 # Copyright: (c) Copyright 2022 <NAME> # Module: model/LsBlk.py # Date Started: March 23, 2022 # Purpose: Store and provide API for Linux lsblk command. # Development: # Arguments to include ...
[ "subprocess.Popen", "json.loads", "view.Components.JsonTreeView", "tkinter.messagebox.askyesno", "tkinter.LabelFrame", "tkinter.Tk" ]
[((1875, 1954), 'tkinter.messagebox.askyesno', 'messagebox.askyesno', (['"""Exit program """', "('Exit the ' + PROGRAM_TITLE + ' program?')"], {}), "('Exit program ', 'Exit the ' + PROGRAM_TITLE + ' program?')\n", (1894, 1954), False, 'from tkinter import Tk, messagebox, LabelFrame, BOTH, RAISED\n'), ((2041, 2045), 'tk...
import argparse def encrypt(message: str, key: dict[int, int]) -> str: encrypted = map(lambda char: key[char], message) return encrypted def decrypt(message: str, key: dict[int, int]) -> str: decrypted = map(lambda char: key[char], message) return decrypted def main(): parser = argparse.ArgumentParser() ...
[ "argparse.ArgumentParser" ]
[((293, 318), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (316, 318), False, 'import argparse\n')]
""" Obit Plotting class Create a plot object using newOPlot which allows specifying the output and background color. If no output is specified this information will be prompted. Next, the plotting region must be specified using either PSetPlot, one of the XY plotting routines (PXYPlot, PXYOver, or PXYErr) PGrayScal...
[ "Obit.PlotInitPlot", "Obit.PlotXYPlot", "Obit.CreateOPlot", "Obit.PlotSetPage", "Obit.PlotXYErr", "Obit.PlotDrawSymbol", "Obit.PlotSetColor", "Obit.PlotText", "InfoList.InfoList", "Obit.PlotSetLineStyle", "Obit.PlotContour", "math.cos", "Obit.OPlot_Get_me", "Obit.PlotGetList", "Obit.OPlo...
[((5153, 5211), 'Obit.PlotInitPlot', 'Obit.PlotInitPlot', (['out.me', 'output', 'bgcolor', 'nx', 'ny', 'err.me'], {}), '(out.me, output, bgcolor, nx, ny, err.me)\n', (5170, 5211), False, 'import Obit, _Obit, InfoList, Image\n'), ((7952, 8001), 'Obit.PlotXYPlot', 'Obit.PlotXYPlot', (['plot.me', 'symbol', 'n', 'x', 'y', ...
# -------------- #Code starts here import sys def palindrome(num): numstr = str(num) for i in range(num+1,sys.maxsize): if str(i)== str(i)[::-1]: return i palindrome(123) # -------------- #Code starts here from collections import Counter def a_scramble(str_1,str_2): list_str1 = Counter(st...
[ "math.sqrt" ]
[((630, 642), 'math.sqrt', 'math.sqrt', (['x'], {}), '(x)\n', (639, 642), False, 'import math\n')]
from utils import open_image, open_greyscale_bmp from workspace_calc import WorkspaceCalculator from workspace_view import WorkspaceView class Workspace: def __init__(self, app_page, room_name, robot_name): room_bmp = open_greyscale_bmp(room_name) robot_bmp = open_greyscale_bmp(robot_name) ...
[ "utils.open_greyscale_bmp", "workspace_view.WorkspaceView", "workspace_calc.WorkspaceCalculator", "utils.open_image" ]
[((232, 261), 'utils.open_greyscale_bmp', 'open_greyscale_bmp', (['room_name'], {}), '(room_name)\n', (250, 261), False, 'from utils import open_image, open_greyscale_bmp\n'), ((282, 312), 'utils.open_greyscale_bmp', 'open_greyscale_bmp', (['robot_name'], {}), '(robot_name)\n', (300, 312), False, 'from utils import ope...
# Generated by Django 2.2.13 on 2021-06-08 10:08 import os from django.db import migrations def create_premier_tenant(apps, schema_editor): # We can't import the Person model directly as it may be a newer # version than this migration expects. We use the historical version. Client = apps.get_model('Custom...
[ "django.db.migrations.RunPython", "os.getenv" ]
[((397, 416), 'os.getenv', 'os.getenv', (['"""DOMAIN"""'], {}), "('DOMAIN')\n", (406, 416), False, 'import os\n'), ((1645, 1697), 'django.db.migrations.RunPython', 'migrations.RunPython', (['create_premier_tenant', 'reverse'], {}), '(create_premier_tenant, reverse)\n', (1665, 1697), False, 'from django.db import migrat...
from common.input_validation import ( extract_phone_number, ) def test_extract_phone_number(): assert extract_phone_number('510501622') == None assert extract_phone_number('5105016227') == '15105016227' assert extract_phone_number('15105016227') == '15105016227' assert extract_phone_number('+15105...
[ "common.input_validation.extract_phone_number" ]
[((112, 145), 'common.input_validation.extract_phone_number', 'extract_phone_number', (['"""510501622"""'], {}), "('510501622')\n", (132, 145), False, 'from common.input_validation import extract_phone_number\n'), ((165, 199), 'common.input_validation.extract_phone_number', 'extract_phone_number', (['"""5105016227"""']...
import unittest import math import datasets from pdffigures_utils import get_num_pages_in_pdf class TestDataset(unittest.TestCase): def test_pages_annotated_consistency(self): for dataset in datasets.DATASETS.values(): dataset = dataset() pages_annotated = dataset.get_annotated_pa...
[ "unittest.main", "datasets.DATASETS.values", "pdffigures_utils.get_num_pages_in_pdf", "math.ceil", "datasets.DatasetPartition" ]
[((2637, 2652), 'unittest.main', 'unittest.main', ([], {}), '()\n', (2650, 2652), False, 'import unittest\n'), ((206, 232), 'datasets.DATASETS.values', 'datasets.DATASETS.values', ([], {}), '()\n', (230, 232), False, 'import datasets\n'), ((1464, 1490), 'datasets.DATASETS.values', 'datasets.DATASETS.values', ([], {}), ...
# -*- coding: utf-8 -*- import botocore import boto3 import io from datetime import datetime import s3Uploader # Refs : https://boto3.readthedocs.io/en/latest/reference/services/s3.html s3 = boto3.client('s3') def main(): # [追加する時] # バケットがなければ作成 # あればそれを使う。 # ファイルの重複チェック # 重複していれば、削除し更新 # 重...
[ "s3Uploader.s3Uploader", "datetime.datetime.now", "boto3.client" ]
[((194, 212), 'boto3.client', 'boto3.client', (['"""s3"""'], {}), "('s3')\n", (206, 212), False, 'import boto3\n'), ((1821, 1881), 's3Uploader.s3Uploader', 's3Uploader.s3Uploader', (['bucketName', 'objectName', '"""./image.jpg"""'], {}), "(bucketName, objectName, './image.jpg')\n", (1842, 1881), False, 'import s3Upload...
from django.test import TestCase from majority_judgment.tools import get_ranking, get_ratings, majority_grade class MajorityJudgmentTestCase(TestCase): fixtures = ['election.json'] # def setUp(self): def test_ranking(self): election_id = 2 ranking = get_ranking(election_id) ranki...
[ "majority_judgment.tools.majority_grade", "majority_judgment.tools.get_ranking" ]
[((282, 306), 'majority_judgment.tools.get_ranking', 'get_ranking', (['election_id'], {}), '(election_id)\n', (293, 306), False, 'from majority_judgment.tools import get_ranking, get_ratings, majority_grade\n'), ((584, 608), 'majority_judgment.tools.get_ranking', 'get_ranking', (['election_id'], {}), '(election_id)\n',...
from typing import List, Union import pandas as pd from zvdata import IntervalLevel from zvt.api.common import get_kdata_schema from zvt.factors.algorithm import MacdTransformer, MaTransformer from zvt.factors.factor import Factor, Transformer, Accumulator class TechnicalFactor(Factor): def __init__(self, ...
[ "zvt.factors.algorithm.MacdTransformer", "zvt.api.common.get_kdata_schema" ]
[((1309, 1326), 'zvt.factors.algorithm.MacdTransformer', 'MacdTransformer', ([], {}), '()\n', (1324, 1326), False, 'from zvt.factors.algorithm import MacdTransformer, MaTransformer\n'), ((1500, 1542), 'zvt.api.common.get_kdata_schema', 'get_kdata_schema', (['entity_type'], {'level': 'level'}), '(entity_type, level=leve...
# Generated by Django 2.2.12 on 2020-06-16 13:14 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('memes', '0001_initial'), ] operations = [ migrations.RemoveField( model_name='task', ...
[ "django.db.migrations.RemoveField", "django.db.models.URLField", "django.db.models.IntegerField", "django.db.models.ForeignKey" ]
[((256, 310), 'django.db.migrations.RemoveField', 'migrations.RemoveField', ([], {'model_name': '"""task"""', 'name': '"""name"""'}), "(model_name='task', name='name')\n", (278, 310), False, 'from django.db import migrations, models\n'), ((456, 482), 'django.db.models.URLField', 'models.URLField', ([], {'null': '(True)...
from entities.remote_radio_head import RemoteRadioHead from entities.hypervisor import Hypervisor from entities.baseband_unit import BasebandUnit from entities.switch import Switch from forwarding.forwarding import Forwarding class StatHistory(object): history = {} def get(key, current): if (key in St...
[ "forwarding.forwarding.Forwarding", "entities.remote_radio_head.RemoteRadioHead", "entities.baseband_unit.BasebandUnit", "entities.switch.Switch", "entities.hypervisor.Hypervisor" ]
[((651, 677), 'forwarding.forwarding.Forwarding', 'Forwarding', (['self.env', 'self'], {}), '(self.env, self)\n', (661, 677), False, 'from forwarding.forwarding import Forwarding\n'), ((6337, 6377), 'entities.switch.Switch', 'Switch', (['self.env', '"""physical"""', '"""external"""'], {}), "(self.env, 'physical', 'exte...
# coding=utf-8 # Author: <NAME> & <NAME> # Date: Jan 06, 2021 # # Description: Utility functions # import os import re import functools import pickle import numpy as np # # Functions to handle Twitter text # re_all_after_retweet = re.compile(r"rt @[a-zA-Z0-9_]+.+", re.IGNORECASE | re.UNICODE) def removeAllAfterRetwe...
[ "pickle.dump", "os.makedirs", "os.path.dirname", "os.path.exists", "os.path.isfile", "pickle.load", "functools.wraps", "re.compile" ]
[((232, 293), 're.compile', 're.compile', (['"""rt @[a-zA-Z0-9_]+.+"""', '(re.IGNORECASE | re.UNICODE)'], {}), "('rt @[a-zA-Z0-9_]+.+', re.IGNORECASE | re.UNICODE)\n", (242, 293), False, 'import re\n'), ((499, 613), 're.compile', 're.compile', (['"""(#Repost @\\\\w+ with @repostapp)|(#EzRepost @\\\\w+ with @ezrepostapp...
import hashlib import sys if sys.version_info[0] < 3: import urllib else: import urllib.parse as urllib class Gravtr(object): GRAVATAR_URL = 'https://www.gravatar.com/avatar/' GRAVATAR_URL_UNSECURE = 'http://www.gravatar.com/avatar/' class ratingType(object): G = 'g' PG = 'pg' ...
[ "hashlib.md5", "urllib.parse.urlencode" ]
[((1050, 1074), 'urllib.parse.urlencode', 'urllib.urlencode', (['params'], {}), '(params)\n', (1066, 1074), True, 'import urllib.parse as urllib\n'), ((664, 687), 'hashlib.md5', 'hashlib.md5', (['self.email'], {}), '(self.email)\n', (675, 687), False, 'import hashlib\n')]
import os import unittest from MuseParse.tests.testUsingXML.xmlSet import xmlSet, parsePiece from MuseParse.classes.ObjectHierarchy.TreeClasses.BaseTree import Search, FindByIndex from MuseParse.classes.ObjectHierarchy.TreeClasses.NoteNode import NoteNode from MuseParse.classes.ObjectHierarchy.TreeClasses.MeasureNode ...
[ "MuseParse.classes.ObjectHierarchy.TreeClasses.BaseTree.Search", "os.path.join", "MuseParse.tests.testUsingXML.xmlSet.xmlSet.setUp" ]
[((550, 583), 'os.path.join', 'os.path.join', (['directory', 'partname'], {}), '(directory, partname)\n', (562, 583), False, 'import os\n'), ((642, 660), 'MuseParse.tests.testUsingXML.xmlSet.xmlSet.setUp', 'xmlSet.setUp', (['self'], {}), '(self)\n', (654, 660), False, 'from MuseParse.tests.testUsingXML.xmlSet import xm...
import os import pathlib from ctypes import * from ctypes import _SimpleCData from ctypes import _Pointer from .common import CEnum class SNResult(CEnum): SN_S_OK = (0) SN_S_PENDING = (1) SN_S_NO_MSG = (3) SN_S_TM_VERSION = (4) SN_S_REPLACED = (5) SN_S_NO_ACTION = (6) SN_S_CONNECTED = SN_S...
[ "os.getcwd", "os.getenv" ]
[((2153, 2164), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (2162, 2164), False, 'import os\n'), ((2208, 2232), 'os.getenv', 'os.getenv', (['"""SN_PS3_PATH"""'], {}), "('SN_PS3_PATH')\n", (2217, 2232), False, 'import os\n')]
import sys import unittest import os script_dir = os.path.dirname(os.path.realpath(__file__)) sys.path.insert(1, os.path.abspath( os.path.join(script_dir, os.path.join('..', '..')))) from pake import process import pake.program import pake class ProcessTest(unittest.TestCase): def test_call(self): ...
[ "pake.de_init", "pake.process.check_output", "os.path.realpath", "pake.process.call", "pake.init", "pake.process.check_call", "os.path.join" ]
[((68, 94), 'os.path.realpath', 'os.path.realpath', (['__file__'], {}), '(__file__)\n', (84, 94), False, 'import os\n'), ((1840, 1870), 'pake.de_init', 'pake.de_init', ([], {'clear_conf': '(False)'}), '(clear_conf=False)\n', (1852, 1870), False, 'import pake\n'), ((1885, 1896), 'pake.init', 'pake.init', ([], {}), '()\n...
# Copyright 2021 Huawei Technologies Co., Ltd # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to...
[ "src.loss.Quadrupletloss", "argparse.ArgumentParser", "mindspore.train.callback.ModelCheckpoint", "src.dataset.create_dataset1", "moxing.file.copy_parallel", "numpy.mean", "mindspore.train.serialization.load_checkpoint", "os.path.join", "mindspore.train.serialization.load_param_into_net", "mindspo...
[((1586, 1597), 'mindspore.common.set_seed', 'set_seed', (['(1)'], {}), '(1)\n', (1594, 1597), False, 'from mindspore.common import set_seed\n'), ((1608, 1667), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Image classification"""'}), "(description='Image classification')\n", (1631, 166...
# -*- coding: utf-8 -*- """testing script""" import os import sys from functools import reduce import numpy as np import pandas as pd import nltk # Natural Language Tool Kit from fuzzywuzzy import fuzz, process # Fuzzy String Matching import jellyfish # Distance metrics from sklearn.feature_extraction.text import T...
[ "functools.reduce", "sklearn.feature_extraction.text.TfidfVectorizer", "sklearn.metrics.pairwise.linear_kernel" ]
[((3797, 3848), 'functools.reduce', 'reduce', (["(lambda x, y: f'{x} {y}')", '[x[0] for x in l1]'], {}), "(lambda x, y: f'{x} {y}', [x[0] for x in l1])\n", (3803, 3848), False, 'from functools import reduce\n'), ((3882, 3933), 'functools.reduce', 'reduce', (["(lambda x, y: f'{x} {y}')", '[x[0] for x in l1]'], {}), "(la...
# Generated by Django 2.2 on 2019-05-23 12:31 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('dashboard', '0001_initial'), ] operations = [ migrations.AlterField( model_name='entrylist', name='awards', ...
[ "django.db.models.CharField" ]
[((328, 398), 'django.db.models.CharField', 'models.CharField', ([], {'blank': '(True)', 'default': '"""未设置"""', 'max_length': '(200)', 'null': '(True)'}), "(blank=True, default='未设置', max_length=200, null=True)\n", (344, 398), False, 'from django.db import migrations, models\n'), ((527, 597), 'django.db.models.CharFie...
from aparse import click from viewformer.utils.click import LazyGroup @click.group(cls=LazyGroup) def main(): pass @main.group(cls=LazyGroup) def dataset(): pass @main.group(cls=LazyGroup) def visualize(): pass @main.group(cls=LazyGroup) def model(): pass @main.group(cls=LazyGroup) def evaluat...
[ "aparse.click.group" ]
[((73, 99), 'aparse.click.group', 'click.group', ([], {'cls': 'LazyGroup'}), '(cls=LazyGroup)\n', (84, 99), False, 'from aparse import click\n')]
""" Copyright (c) 2022 Huawei Technologies Co.,Ltd. openGauss is licensed under Mulan PSL v2. You can use this software according to the terms and conditions of the Mulan PSL v2. You may obtain a copy of Mulan PSL v2 at: http://license.coscl.org.cn/MulanPSL2 THIS SOFTWARE IS PROVIDED ON AN "AS IS" BASIS, W...
[ "yat.test.Node", "testcase.utils.CommonSH.CommonSH", "testcase.utils.Logger.Logger" ]
[((956, 974), 'testcase.utils.CommonSH.CommonSH', 'CommonSH', (['"""dbuser"""'], {}), "('dbuser')\n", (964, 974), False, 'from testcase.utils.CommonSH import CommonSH\n'), ((1000, 1014), 'yat.test.Node', 'Node', (['"""dbuser"""'], {}), "('dbuser')\n", (1004, 1014), False, 'from yat.test import Node\n'), ((1034, 1042), ...
from GitLabApi import objects from core.tests.test_view import LoginMethods from core.tests.test_view import SimpleUrlsTestsCases from django.db.models import QuerySet from django.urls import reverse from groups import models from groups.sidebar import GroupSidebar, FutureGroupSidebar from groups.tests import test_form...
[ "groups.models.AddSubgroup.objects.get", "groups.tests.test_forms.SubgroupAndMembersFromFileFormTests.valid_file_data.items", "groups.tests.test_forms.TaskGroupFormTests.valid_form_data.items", "groups.tests.models.AddProjectCreateMethods", "groups.models.AddMember.objects.get", "groups.tests.test_forms.A...
[((7695, 7748), 'groups.tests.test_forms.TaskGroupFormTests.valid_form_data.items', 'test_forms.TaskGroupFormTests.valid_form_data.items', ([], {}), '()\n', (7746, 7748), False, 'from groups.tests import test_forms\n'), ((9189, 9247), 'groups.models.TaskGroup.objects.get', 'models.TaskGroup.objects.get', ([], {'parent_...
#!/usr/bin/env python3 from api import DB from api.models.listing import Listing def get_all_queries(): """ Returns all stored listing queries. """ return list(Listing.query.all())
[ "api.models.listing.Listing.query.all" ]
[((183, 202), 'api.models.listing.Listing.query.all', 'Listing.query.all', ([], {}), '()\n', (200, 202), False, 'from api.models.listing import Listing\n')]
import asyncio import json from datetime import date from decimal import Decimal from typing import Dict from aiohttp import ClientResponse from dynaconf import settings _PRICE_HISTORY_API = f"{settings.BENZAK_API_URL}/price-history/" async def load_price(logger, session, price: Dict): logger.debug( f"c...
[ "json.dumps" ]
[((411, 454), 'json.dumps', 'json.dumps', (['price'], {'indent': '(2)', 'sort_keys': '(True)'}), '(price, indent=2, sort_keys=True)\n', (421, 454), False, 'import json\n')]
"""playnetmano_rm common internal object model""" from oslo_utils import versionutils from oslo_versionedobjects import base from playnetmano_rm import objects VersionedObjectDictCompat = base.VersionedObjectDictCompat class Playnetmano_rmObject(base.VersionedObject): """Base class for playnetmano_rm objects. ...
[ "oslo_utils.versionutils.convert_version_to_tuple" ]
[((1417, 1467), 'oslo_utils.versionutils.convert_version_to_tuple', 'versionutils.convert_version_to_tuple', (['cls.VERSION'], {}), '(cls.VERSION)\n', (1454, 1467), False, 'from oslo_utils import versionutils\n')]
import random import datetime import dateparser from faker import Faker from .base import Filth class DateOfBirthFilth(Filth): type = 'date_of_birth' min_age_years = 18 max_age_years = 100 @staticmethod def generate(faker: Faker) -> str: """Generates an example of this ``Filth`` type, us...
[ "dateparser.parse", "datetime.date.today", "random.choice" ]
[((1071, 1098), 'dateparser.parse', 'dateparser.parse', (['self.text'], {}), '(self.text)\n', (1087, 1098), False, 'import dateparser\n'), ((937, 959), 'random.choice', 'random.choice', (['formats'], {}), '(formats)\n', (950, 959), False, 'import random\n'), ((1193, 1214), 'datetime.date.today', 'datetime.date.today', ...
import psycopg2 from psycopg2.extensions import connection, cursor from psycopg2.extras import DictCursor from typing import Dict from src.log.logger import logger from contextlib import contextmanager @contextmanager def get_connection(params: Dict[str, str]) -> connection: """ Get a connection using a cont...
[ "src.log.logger.logger.debug", "psycopg2.connect" ]
[((467, 493), 'psycopg2.connect', 'psycopg2.connect', ([], {}), '(**params)\n', (483, 493), False, 'import psycopg2\n'), ((696, 739), 'src.log.logger.logger.debug', 'logger.debug', (['"""Closing database connection"""'], {}), "('Closing database connection')\n", (708, 739), False, 'from src.log.logger import logger\n')...
import unittest from test.test01 import soma class TesteSoma(unittest.TestCase): def test_retorno_soma_10_10(self): self .assertEqual(soma(10, 10), 20)
[ "test.test01.soma" ]
[((147, 159), 'test.test01.soma', 'soma', (['(10)', '(10)'], {}), '(10, 10)\n', (151, 159), False, 'from test.test01 import soma\n')]
import pytest from data_structures.heap import Heap @pytest.fixture def base_heap(): heap = Heap() heap.push(1) heap.push(2) heap.push(3) heap.push(4) heap.push(5) return heap def test_heap_init(): basic_heap = Heap() init_list_heap = Heap([9, 8, 7, 5, 1, 2]) assert isinstanc...
[ "data_structures.heap.Heap" ]
[((98, 104), 'data_structures.heap.Heap', 'Heap', ([], {}), '()\n', (102, 104), False, 'from data_structures.heap import Heap\n'), ((247, 253), 'data_structures.heap.Heap', 'Heap', ([], {}), '()\n', (251, 253), False, 'from data_structures.heap import Heap\n'), ((275, 299), 'data_structures.heap.Heap', 'Heap', (['[9, 8...
import os import json import boto3 from botocore.exceptions import ClientError from cryptography.fernet import Fernet dynamodb = boto3.resource('dynamodb') s3 = boto3.resource('s3') vault_table = dynamodb.Table(os.environ.get('VAULT_TABLE_NAME')) vault_table_partition_key = os.environ.get('VAULT_TABLE_KEY') vault_tabl...
[ "os.environ.get", "cryptography.fernet.Fernet", "boto3.resource", "json.dumps" ]
[((130, 156), 'boto3.resource', 'boto3.resource', (['"""dynamodb"""'], {}), "('dynamodb')\n", (144, 156), False, 'import boto3\n'), ((162, 182), 'boto3.resource', 'boto3.resource', (['"""s3"""'], {}), "('s3')\n", (176, 182), False, 'import boto3\n'), ((276, 309), 'os.environ.get', 'os.environ.get', (['"""VAULT_TABLE_KE...
import errno import glob import os import re import shutil from pypadre.core.model.code.code_mixin import CodeMixin, PythonPackage, PythonFile, GenericCall, \ GitIdentifier, RepositoryIdentifier, PipIdentifier, Function from pypadre.pod.backend.i_padre_backend import IPadreBackend from pypadre.pod.repository.i_rep...
[ "re.escape", "pypadre.core.model.code.code_mixin.GitIdentifier", "pypadre.pod.repository.local.file.generic.i_file_repository.File", "shutil.copy", "shutil.copytree", "pypadre.core.model.code.code_mixin.Function", "os.path.join", "pypadre.core.model.code.code_mixin.PipIdentifier" ]
[((1002, 1039), 'pypadre.pod.repository.local.file.generic.i_file_repository.File', 'File', (['"""metadata.json"""', 'JSonSerializer'], {}), "('metadata.json', JSonSerializer)\n", (1006, 1039), False, 'from pypadre.pod.repository.local.file.generic.i_file_repository import File\n'), ((706, 732), 'shutil.copytree', 'shu...
#!/usr/bin/python3 # -*- coding: utf-8 -*- """ Created on Mon Oct 8 20:54:26 2018 @author: andrea """ import sys import json from xml.sax import make_parser from urllib.request import urlretrieve from smallsmilhandler import SmallSMILHandler class KaraokeLocal(SmallSMILHandler): def __init__(self, fichero): ...
[ "json.dumps", "urllib.request.urlretrieve", "smallsmilhandler.SmallSMILHandler", "sys.exit", "xml.sax.make_parser" ]
[((377, 390), 'xml.sax.make_parser', 'make_parser', ([], {}), '()\n', (388, 390), False, 'from xml.sax import make_parser\n'), ((425, 443), 'smallsmilhandler.SmallSMILHandler', 'SmallSMILHandler', ([], {}), '()\n', (441, 443), False, 'from smallsmilhandler import SmallSMILHandler\n'), ((1112, 1134), 'json.dumps', 'json...
import os import sys import tempfile from datetime import datetime from pprint import pprint import ray from ray import tune from ray.rllib.agents import Trainer from ray.tune.logger import UnifiedLogger from ray.tune.result import DEFAULT_RESULTS_DIR # os.environ['TF_CPP_MIN_LOG_LEVEL'] = '2' from command_line_tools...
[ "ray.init", "command_line_tools.run_tools.setup_run", "ray.nodes", "ray.tune.run", "ray.is_initialized", "ray.shutdown", "ray.cluster_resources" ]
[((2877, 2902), 'command_line_tools.run_tools.setup_run', 'setup_run', (['default_config'], {}), '(default_config)\n', (2886, 2902), False, 'from command_line_tools.run_tools import setup_run\n'), ((3195, 3305), 'ray.tune.run', 'tune.run', (['train'], {'name': "config['name']", 'trial_name_creator': "(lambda trial: con...
"""Protocol-related functions.""" # Copyright 2020-2021 Blue Brain Project / EPFL # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless r...
[ "bluepyopt.ephys.locations.NrnSomaDistanceCompLocation", "bluepyopt.ephys.protocols.SweepProtocol", "bluepyopt.ephys.protocols.SequenceProtocol", "emodelrunner.features.define_efeatures", "emodelrunner.recordings.RecordingCustom", "emodelrunner.protocols.StepProtocol", "emodelrunner.protocols.SweepProto...
[((1156, 1183), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1173, 1183), False, 'import logging\n'), ((1196, 1300), 'bluepyopt.ephys.locations.NrnSeclistCompLocation', 'ephys.locations.NrnSeclistCompLocation', ([], {'name': '"""soma"""', 'seclist_name': '"""somatic"""', 'sec_index': '...
import os from flask import Flask, redirect from flask import request from flask import jsonify import hashlib app = Flask(__name__) c = 0 clients = [] chat = [] #[from, to, status[0sent, 1accepted, 2rejected]] requests = {} requests_sent = {} version = 5 additive = 0 def getUID(ip): return hashlib.sha256(str(...
[ "os.environ.get", "flask.jsonify", "flask.Flask", "flask.redirect" ]
[((118, 133), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (123, 133), False, 'from flask import Flask, redirect\n'), ((2333, 2356), 'flask.redirect', 'redirect', (['"""/"""'], {'code': '(302)'}), "('/', code=302)\n", (2341, 2356), False, 'from flask import Flask, redirect\n'), ((2676, 2697), 'flask.json...
# Copyright (c) 2020, Huawei Technologies.All rights reserved. # # Licensed under the BSD 3-Clause License (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # https://opensource.org/licenses/BSD-3-Clause # # Unless required by applicable law...
[ "numpy.random.uniform", "torch.ne", "copy.deepcopy", "common_utils.run_tests", "util_test.create_common_tensor" ]
[((8284, 8295), 'common_utils.run_tests', 'run_tests', ([], {}), '()\n', (8293, 8295), False, 'from common_utils import TestCase, run_tests\n'), ((894, 918), 'torch.ne', 'torch.ne', (['input1', 'input2'], {}), '(input1, input2)\n', (902, 918), False, 'import torch\n'), ((1051, 1075), 'torch.ne', 'torch.ne', (['input1',...
from elasticsearch.helpers import scan import utils.helpers as hp valueField = { 'ps_packetloss': 'packet_loss', 'ps_owd': 'delay_mean', 'ps_retransmits': 'retransmits', 'ps_throughput': 'throughput' } def query4Avg(idx, dateFrom, dateTo): val_fld = val...
[ "utils.helpers.es.search" ]
[((2956, 2991), 'utils.helpers.es.search', 'hp.es.search', ([], {'index': 'idx', 'body': 'query'}), '(index=idx, body=query)\n', (2968, 2991), True, 'import utils.helpers as hp\n')]
from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import pytest from mock import patch from night_scheduler.framework.sun.sun import Sun class TestSun(object): FAKE_LATITUDE = "00" FAKE_LONGITUDE = "11" F...
[ "mock.patch", "night_scheduler.framework.sun.sun.Sun" ]
[((1033, 1054), 'mock.patch', 'patch', (['"""requests.get"""'], {}), "('requests.get')\n", (1038, 1054), False, 'from mock import patch\n'), ((1229, 1327), 'night_scheduler.framework.sun.sun.Sun', 'Sun', ([], {'latitude': 'TestSun.FAKE_LATITUDE', 'longitude': 'TestSun.FAKE_LONGITUDE', 'date': 'TestSun.FAKE_DATE'}), '(l...
#!/usr/bin/env python3 import argparse import sys import psutil from wrap_scriptlet import wrap_scriptlet def run(): parser = argparse.ArgumentParser() parser.add_argument('pid') args = parser.parse_args(sys.argv[1:]) process = psutil.Process(int(args.pid)) return process.cmdline() sys.exit(wra...
[ "wrap_scriptlet.wrap_scriptlet", "argparse.ArgumentParser" ]
[((133, 158), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (156, 158), False, 'import argparse\n'), ((317, 336), 'wrap_scriptlet.wrap_scriptlet', 'wrap_scriptlet', (['run'], {}), '(run)\n', (331, 336), False, 'from wrap_scriptlet import wrap_scriptlet\n')]
''' By Zhenghang(<NAME> ############################################################################################ It's a light server based on FLASK micro framework, 1.Requirements: Python 3, Flask and relevant packages 2. How does this work? (1) Firstly, modify the host IP address of your own environment. (2)...
[ "csv.reader", "csv.writer", "os.path.getsize", "flask.Flask", "app.models.User.query.all" ]
[((1817, 1832), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (1822, 1832), False, 'from flask import Flask, request\n'), ((2366, 2389), 'app.models.User.query.all', 'models.User.query.all', ([], {}), '()\n', (2387, 2389), False, 'from app import db, models\n'), ((3430, 3488), 'csv.writer', 'csv.writer', ...
from rx.core import Observable from rx.internal import extensionmethod @extensionmethod(Observable) def do_while(self, condition): """Repeats source as long as condition holds emulating a do while loop. Keyword arguments: condition -- {Function} The condition which determines if the source will b...
[ "rx.internal.extensionmethod", "rx.core.Observable.while_do" ]
[((74, 101), 'rx.internal.extensionmethod', 'extensionmethod', (['Observable'], {}), '(Observable)\n', (89, 101), False, 'from rx.internal import extensionmethod\n'), ((480, 516), 'rx.core.Observable.while_do', 'Observable.while_do', (['condition', 'self'], {}), '(condition, self)\n', (499, 516), False, 'from rx.core i...
import os import networkx as nx import scipy.sparse as sp def search_cycle(dir_adj): dir_adj = nx.from_scipy_sparse_matrix(A=dir_adj, create_using=nx.DiGraph) cycles = list(nx.algorithms.cycles.simple_cycles(dir_adj)) num_cycle = len(cycles) q = [] for i in range(num_cycle): q.append(len(cy...
[ "networkx.from_scipy_sparse_matrix", "networkx.algorithms.cycles.simple_cycles", "os.path.join" ]
[((424, 464), 'os.path.join', 'os.path.join', (['dataset_path', 'dataset_name'], {}), '(dataset_path, dataset_name)\n', (436, 464), False, 'import os\n'), ((100, 163), 'networkx.from_scipy_sparse_matrix', 'nx.from_scipy_sparse_matrix', ([], {'A': 'dir_adj', 'create_using': 'nx.DiGraph'}), '(A=dir_adj, create_using=nx.D...
from asitiger.status import ( AxisEnabledStatus, AxisStatus, JoystickStatus, LimitStatus, MotorStatus, RampingDirection, RampingStatus, Status, status_from_decimal, statuses_for_rdstat, ) RDSTAT_RESPONSE = ":A 10N 138" def test_status_from_decimal_types(): axis = status_f...
[ "asitiger.status.Status.from_flag", "asitiger.status.status_from_decimal", "asitiger.status.statuses_for_rdstat" ]
[((312, 336), 'asitiger.status.status_from_decimal', 'status_from_decimal', (['(210)'], {}), '(210)\n', (331, 336), False, 'from asitiger.status import AxisEnabledStatus, AxisStatus, JoystickStatus, LimitStatus, MotorStatus, RampingDirection, RampingStatus, Status, status_from_decimal, statuses_for_rdstat\n'), ((809, 8...
from django.contrib.auth.models import User from rest_framework import serializers from .models import Film, ExtraInfo, Recenzja, Aktor class UserSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = User fields = ['id', 'username', 'email','password'] extra_kwargs = {'pa...
[ "django.contrib.auth.models.User.objects.create_user" ]
[((423, 465), 'django.contrib.auth.models.User.objects.create_user', 'User.objects.create_user', ([], {}), '(**validated_data)\n', (447, 465), False, 'from django.contrib.auth.models import User\n')]
""" Module containing the definitions and methods to compute a variety of indices used to study ENSO """ from typing import List, Optional, Tuple import numpy as np import xarray as xr from eofs.xarray import Eof from .core import compute_anomaly, compute_climatology, xconvolve class ECindex: """ Computes ...
[ "numpy.deg2rad", "numpy.zeros", "xarray.merge", "numpy.array", "numpy.sqrt" ]
[((2338, 2349), 'numpy.zeros', 'np.zeros', (['(2)'], {}), '(2)\n', (2346, 2349), True, 'import numpy as np\n'), ((3047, 3073), 'xarray.merge', 'xr.merge', (['[eindex, cindex]'], {}), '([eindex, cindex])\n', (3055, 3073), True, 'import xarray as xr\n'), ((3931, 3954), 'numpy.array', 'np.array', (['smooth_kernel'], {}), ...
# -*- coding: utf-8 -*- """ Copyright (c) 2020, University of Southampton All rights reserved. Licensed under the BSD 3-Clause License. See LICENSE.md file in the project root for full license information. """ import pynmea2 from auv_nav.sensors import Category, Usbl from auv_nav.tools.time_conversions import date_ti...
[ "oplab.get_raw_folder", "pynmea2.parse", "auv_nav.tools.time_conversions.date_time_to_epoch", "auv_nav.tools.time_conversions.read_timezone", "oplab.get_file_list", "auv_nav.sensors.Usbl" ]
[((791, 814), 'auv_nav.tools.time_conversions.read_timezone', 'read_timezone', (['timezone'], {}), '(timezone)\n', (804, 814), False, 'from auv_nav.tools.time_conversions import date_time_to_epoch, read_timezone\n'), ((939, 1038), 'auv_nav.sensors.Usbl', 'Usbl', (['mission.usbl.std_factor', 'mission.usbl.std_offset', '...
import unittest from hupun.page.hupun_goods.goods_information import GoodsInformation from hupun.page.hupun_goods.goods_information_sku import GoodsInformationsku from hupun.page.in_sale_store_table.export_file_download_req import ExportFileDownloadReq from hupun.page.in_sale_store_table.export_task_query import Expor...
[ "hupun.page.order_goods.OrderGoods", "hupun.page.purchase_order.PurchaseOrder", "hupun.page.sync_module.choose_purchase_bill_sku.ChoosePurBillSku", "pyspider.helper.date.Date.now", "hupun_slow_crawl.model.es.store_house.StoreHouse", "hupun.page.sync_module.confirm_purchase_stock.ConfirmPurBillStock", "h...
[((7478, 7493), 'unittest.main', 'unittest.main', ([], {}), '()\n', (7491, 7493), False, 'import unittest\n'), ((3246, 3256), 'pyspider.helper.date.Date.now', 'Date.now', ([], {}), '()\n', (3254, 3256), False, 'from pyspider.helper.date import Date\n'), ((1814, 1952), 'hupun.page.order_goods.OrderGoods', 'OrderGoods', ...
# Generated by Django 3.0.8 on 2020-07-29 00:27 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('log_api', '0008_auto_20200728_2126'), ] operations = [ migrations.AlterField( model_name='execution', name='archived...
[ "django.db.models.BooleanField" ]
[((341, 400), 'django.db.models.BooleanField', 'models.BooleanField', ([], {'default': '(False)', 'verbose_name': '"""Archived"""'}), "(default=False, verbose_name='Archived')\n", (360, 400), False, 'from django.db import migrations, models\n')]
import asyncio, random import nacre class EightBallSession: answers = [ "It is certain", "It is decidedly so", "Without a doubt", "Yes definitely", "You may rely on it", "As I see it, yes", "Most likely", "Outlook good", "Yes", "Signs point to yes", "Reply hazy try again", "Ask again later",...
[ "nacre.handle.isMessageEvent", "random.choice" ]
[((1081, 1108), 'random.choice', 'random.choice', (['self.answers'], {}), '(self.answers)\n', (1094, 1108), False, 'import asyncio, random\n'), ((849, 884), 'nacre.handle.isMessageEvent', 'nacre.handle.isMessageEvent', (['update'], {}), '(update)\n', (876, 884), False, 'import nacre\n')]
# Copyright (c) 2013, GreyCube Technologies and contributors # For license information, please see license.txt from __future__ import unicode_literals import frappe from frappe import _ from frappe.utils import cint import shutil, os from frappe.modules import scrub, get_module_path def copy_report( module="NPro...
[ "frappe.modules.scrub", "frappe.get_doc", "frappe.db.commit", "shutil.copyfile", "frappe.modules.get_module_path" ]
[((533, 551), 'frappe.db.commit', 'frappe.db.commit', ([], {}), '()\n', (549, 551), False, 'import frappe\n'), ((622, 645), 'frappe.modules.get_module_path', 'get_module_path', (['module'], {}), '(module)\n', (637, 645), False, 'from frappe.modules import scrub, get_module_path\n'), ((1098, 1133), 'shutil.copyfile', 's...
from mazikeen.MakedirsBlock import MakedirsBlock from mazikeen.GeneratorException import GeneratorException def generateMakedirs(data): if not isinstance(data, str): raise GeneratorException("'makedirs' block not recognized") return MakedirsBlock(data)
[ "mazikeen.GeneratorException.GeneratorException", "mazikeen.MakedirsBlock.MakedirsBlock" ]
[((250, 269), 'mazikeen.MakedirsBlock.MakedirsBlock', 'MakedirsBlock', (['data'], {}), '(data)\n', (263, 269), False, 'from mazikeen.MakedirsBlock import MakedirsBlock\n'), ((185, 238), 'mazikeen.GeneratorException.GeneratorException', 'GeneratorException', (['"""\'makedirs\' block not recognized"""'], {}), '("\'makedi...
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import models, migrations from django.conf import settings class Migration(migrations.Migration): dependencies = [ ('idc', '0001_initial'), migrations.swappable_dependency(settings.AUTH_USER_MODEL), ] operati...
[ "django.db.migrations.swappable_dependency", "django.db.models.CharField", "django.db.models.ForeignKey", "django.db.models.BooleanField", "django.db.models.AutoField", "django.db.models.IntegerField", "django.db.models.DateTimeField" ]
[((243, 300), 'django.db.migrations.swappable_dependency', 'migrations.swappable_dependency', (['settings.AUTH_USER_MODEL'], {}), '(settings.AUTH_USER_MODEL)\n', (274, 300), False, 'from django.db import models, migrations\n'), ((433, 484), 'django.db.models.AutoField', 'models.AutoField', ([], {'serialize': '(False)',...
import unittest from solver import buddy_strings class TestSolver(unittest.TestCase): def test_buddy_strings(self): self.assertEqual(buddy_strings("ab" , "ba" ), True ) self.assertEqual(buddy_strings("ab" , "ab" ), False) self.assertEqual(buddy_strings("aa" , "aa" ), T...
[ "unittest.main", "solver.buddy_strings" ]
[((496, 511), 'unittest.main', 'unittest.main', ([], {}), '()\n', (509, 511), False, 'import unittest\n'), ((140, 165), 'solver.buddy_strings', 'buddy_strings', (['"""ab"""', '"""ba"""'], {}), "('ab', 'ba')\n", (153, 165), False, 'from solver import buddy_strings\n'), ((209, 234), 'solver.buddy_strings', 'buddy_strings...
#!/usr/bin/env python3 import unittest from src.executor.Printer import Printer from src.data.VideoItem import VideoItem class TestIExecutor(unittest.TestCase): def test_compiles(self): self.assertEqual(True, True) def test_printer(self): printer = Printer() printer.run(VideoItem(met...
[ "unittest.main", "src.executor.Printer.Printer", "src.data.VideoItem.VideoItem" ]
[((381, 396), 'unittest.main', 'unittest.main', ([], {}), '()\n', (394, 396), False, 'import unittest\n'), ((277, 286), 'src.executor.Printer.Printer', 'Printer', ([], {}), '()\n', (284, 286), False, 'from src.executor.Printer import Printer\n'), ((307, 346), 'src.data.VideoItem.VideoItem', 'VideoItem', ([], {'metadata...
"""Main execution body for program. Contains GUI interface and exporting class that creates files instead of generating HTML Reports Author: <NAME> Last Updated: 28/02/2017 """ import argparse import webbrowser import textwrap import xlrd from tkinter import * from tkinter import filedialog, ttk from threading import...
[ "textwrap.dedent", "threading.Thread", "webbrowser.open", "webbrowser.open_new", "xlrd.open_workbook", "tkinter.ttk.Progressbar", "tkinter.filedialog.askdirectory", "tkinter.filedialog.askopenfiles", "tkinter.filedialog.asksaveasfile", "tkinter.filedialog.askopenfile" ]
[((3301, 3369), 'tkinter.ttk.Progressbar', 'ttk.Progressbar', (['mainwindow'], {'orient': '"""horizontal"""', 'mode': '"""determinate"""'}), "(mainwindow, orient='horizontal', mode='determinate')\n", (3316, 3369), False, 'from tkinter import filedialog, ttk\n'), ((4123, 4315), 'tkinter.filedialog.askopenfiles', 'filedi...
from django.test import TestCase, override_settings from social_django.compat import reverse @override_settings(SOCIAL_AUTH_GITHUB_KEY = '1', SOCIAL_AUTH_GITHUB_SECRET='2') class AuthTestcase(TestCase): def setUp(self): session = self.client.session session["github_status"] = "1" session.s...
[ "social_django.compat.reverse", "django.test.override_settings" ]
[((95, 171), 'django.test.override_settings', 'override_settings', ([], {'SOCIAL_AUTH_GITHUB_KEY': '"""1"""', 'SOCIAL_AUTH_GITHUB_SECRET': '"""2"""'}), "(SOCIAL_AUTH_GITHUB_KEY='1', SOCIAL_AUTH_GITHUB_SECRET='2')\n", (112, 171), False, 'from django.test import TestCase, override_settings\n'), ((393, 446), 'social_djang...
import datetime import json import logging import os import threading import time from abc import ABC, abstractmethod import pika from tools.mongo_dao import MongoDB class StopCondition(ABC): def __init__(self, stop_condition_parameters: dict, experiment_description: dict, experiment_id: str): self.even...
[ "threading.Thread", "pika.ConnectionParameters", "json.dumps", "time.sleep", "datetime.timedelta", "os.getenv", "logging.getLogger" ]
[((329, 366), 'os.getenv', 'os.getenv', (['"""BRISE_EVENT_SERVICE_HOST"""'], {}), "('BRISE_EVENT_SERVICE_HOST')\n", (338, 366), False, 'import os\n'), ((393, 435), 'os.getenv', 'os.getenv', (['"""BRISE_EVENT_SERVICE_AMQP_PORT"""'], {}), "('BRISE_EVENT_SERVICE_AMQP_PORT')\n", (402, 435), False, 'import os\n'), ((931, 98...