code stringlengths 22 1.05M | apis listlengths 1 3.31k | extract_api stringlengths 75 3.25M |
|---|---|---|
from django.contrib import admin
from apps.iotdb_cloud_core.models import IoTDBRelease
admin.site.register(IoTDBRelease)
| [
"django.contrib.admin.site.register"
] | [((89, 122), 'django.contrib.admin.site.register', 'admin.site.register', (['IoTDBRelease'], {}), '(IoTDBRelease)\n', (108, 122), False, 'from django.contrib import admin\n')] |
import copy
import numpy as np
PXL2CM = 0.035277778
def print_formatted_stats(stats):
"""
Print formatted results for result tables
"""
print("& {:.2f} & {:.2f} & {:.2f} & {:.2f} \\\\" .format(np.mean(stats['tracked_until_end_ratio']),
np... | [
"copy.deepcopy",
"numpy.abs",
"numpy.sum",
"numpy.median",
"numpy.asarray",
"numpy.mean"
] | [((1224, 1252), 'numpy.asarray', 'np.asarray', (['alignment_errors'], {}), '(alignment_errors)\n', (1234, 1252), True, 'import numpy as np\n'), ((1280, 1304), 'numpy.abs', 'np.abs', (['alignment_errors'], {}), '(alignment_errors)\n', (1286, 1304), True, 'import numpy as np\n'), ((1327, 1354), 'numpy.asarray', 'np.asarr... |
from pathlib import Path
from torchvision.datasets import VisionDataset
import numpy as np
from PIL import Image
class MHPv1(VisionDataset):
"""
MHP dataset : Multi-Human Parsing
V1은 human parsing 만 있고, v2는 pose 포함
https://github.com/ZhaoJ9014/Multi-Human-Parsing
or https://lv-mhp.github.io/
... | [
"pathlib.Path",
"PIL.Image.open"
] | [((2751, 2761), 'pathlib.Path', 'Path', (['root'], {}), '(root)\n', (2755, 2761), False, 'from pathlib import Path\n'), ((3223, 3240), 'PIL.Image.open', 'Image.open', (['fname'], {}), '(fname)\n', (3233, 3240), False, 'from PIL import Image\n'), ((3306, 3319), 'PIL.Image.open', 'Image.open', (['f'], {}), '(f)\n', (3316... |
from sys import stdin
def count(S, m, n):
tabla = [[0 for x in range(m)] for x in range(n+1)]
for i in range(m):
tabla[0][i] = 1
for i in range(1, n+1):
for j in range(m):
x = tabla[i - S[j]][j] if i-S[j] >= 0 else 0
y = tabla[i][j-1] if j >= 1 else 0
... | [
"sys.stdin.readline"
] | [((404, 420), 'sys.stdin.readline', 'stdin.readline', ([], {}), '()\n', (418, 420), False, 'from sys import stdin\n'), ((453, 469), 'sys.stdin.readline', 'stdin.readline', ([], {}), '()\n', (467, 469), False, 'from sys import stdin\n')] |
from django.db import models
from django.utils import timezone
from django.forms import ModelForm
from django.contrib.auth.models import User
from django.db.models.signals import post_save
from django.dispatch import receiver
from django.core.exceptions import ValidationError
import secrets
import os
class Profile(mod... | [
"django.db.models.FileField",
"django.db.models.OneToOneField",
"django.db.models.TextField",
"django.core.exceptions.ValidationError",
"django.db.models.ForeignKey",
"django.db.models.CharField",
"django.dispatch.receiver",
"secrets.token_urlsafe",
"django.db.models.BooleanField",
"django.db.mode... | [((516, 548), 'django.dispatch.receiver', 'receiver', (['post_save'], {'sender': 'User'}), '(post_save, sender=User)\n', (524, 548), False, 'from django.dispatch import receiver\n'), ((343, 395), 'django.db.models.OneToOneField', 'models.OneToOneField', (['User'], {'on_delete': 'models.CASCADE'}), '(User, on_delete=mod... |
__source__ = 'https://leetcode.com/problems/k-empty-slots/'
# Time: O()
# Space: O()
#
# Description: Leetcode # 683. K Empty Slots
#
#There is a garden with N slots. In each slot, there is a flower. The N flowers will bloom one by one in N days.
# In each day, there will be exactly one flower blooming and it will be ... | [
"unittest.main"
] | [((1422, 1437), 'unittest.main', 'unittest.main', ([], {}), '()\n', (1435, 1437), False, 'import unittest\n')] |
from flask import render_template,request,redirect,url_for
from . import main
from ..request import get_news
from ..request import get_news, get_news_articles,search_article
# Views
@main.route('/')
def index():
'''
function that returns the index page and its data
'''
#Get popular news
general_ne... | [
"flask.render_template",
"flask.url_for",
"flask.request.args.get"
] | [((677, 707), 'flask.request.args.get', 'request.args.get', (['"""news_query"""'], {}), "('news_query')\n", (693, 707), False, 'from flask import render_template, request, redirect, url_for\n'), ((1227, 1285), 'flask.render_template', 'render_template', (['"""articles.html"""'], {'id': 'id', 'articles': 'articles'}), "... |
#!/usr/bin/env python
"""
hostlists plugin to recursively query plugins based on type.
This makes it possible to obtain lists of hosts by recursively
querying multiple backends.
For example:
* Query dns for www.foo.com
* Get a list of two hostnames back haproxy1.ny.foo.com and
haproxy1.lax.foo.com.
* Qu... | [
"hostlists.plugin_manager.get_plugins"
] | [((2060, 2073), 'hostlists.plugin_manager.get_plugins', 'get_plugins', ([], {}), '()\n', (2071, 2073), False, 'from hostlists.plugin_manager import get_plugins\n')] |
import asyncio
from PIL import Image
from ..Base import Base
class BiasGame(Base):
def __init__(self, *args):
super().__init__(*args)
async def create_bias_game_image(self, first_idol_id, second_idol_id):
"""Uses thread pool to create bias game image to prevent IO blocking."""
# (sel... | [
"PIL.Image.open"
] | [((2858, 2918), 'PIL.Image.open', 'Image.open', (['f"""{self.ex.keys.bias_game_location}bracket8.png"""'], {}), "(f'{self.ex.keys.bias_game_location}bracket8.png')\n", (2868, 2918), False, 'from PIL import Image\n'), ((910, 968), 'PIL.Image.open', 'Image.open', (['f"""{self.ex.keys.bias_game_location}versus.png"""'], {... |
# -*-mode: python; encoding: utf-8; test-case-name: tests.test_app-*-
# ========================================================================
"""
Copyright |(c)| 2017 `Dropbox, Inc.`_
.. |(c)| unicode:: u+a9
.. _`Dropbox, Inc.`: https://www.dropbox.com/
Please see the accompanying ``LICENSE`` and ``CREDIT... | [
"flask.request.form.get",
"future.utils.bytes_to_native_str",
"hmac.compare_digest",
"flask.url_for",
"os.path.join",
"flask.redirect",
"flask.request.args.get",
"flask.request.headers.get",
"flask.abort",
"hashlib.sha256",
"flask.render_template",
"os.urandom",
"hmac.new",
"sqlite3.connec... | [((1619, 1640), 'flask.Flask', 'flask.Flask', (['__name__'], {}), '(__name__)\n', (1630, 1640), False, 'import flask\n'), ((1883, 1918), 'flask.session.get', 'flask.session.get', (['_SESSION_USER_ID'], {}), '(_SESSION_USER_ID)\n', (1900, 1918), False, 'import flask\n'), ((5429, 5477), 'flask.session.pop', 'flask.sessio... |
from PIL import ImageFont
class Letter:
""" letter class- each letter is one of these objects, and is rendered in order. """
def __init__(self,char,size,font,color = (255,255,255,255),b=False,i=False,u=False):
"""
char: character.
size: size of letter.
font: PIL truetype font obj... | [
"PIL.ImageFont.truetype"
] | [((572, 602), 'PIL.ImageFont.truetype', 'ImageFont.truetype', (['font', 'size'], {}), '(font, size)\n', (590, 602), False, 'from PIL import ImageFont\n')] |
'''
#############################################################################
#
# M A I N L O O P & P R I M A R Y F U N C T I O N S
#
#############################################################################
'''
import getopt
import os
import random
import signal
import sys
try:
import numpy... | [
"getopt.getopt",
"os.path.join",
"rpn.debug.typename",
"os.path.isfile",
"random.seed",
"signal.signal",
"sys.setrecursionlimit",
"os.path.expanduser",
"sys.exit"
] | [((1384, 1396), 'sys.exit', 'sys.exit', (['(64)'], {}), '(64)\n', (1392, 1396), False, 'import sys\n'), ((1646, 1673), 'sys.setrecursionlimit', 'sys.setrecursionlimit', (['(2000)'], {}), '(2000)\n', (1667, 1673), False, 'import sys\n'), ((1697, 1710), 'random.seed', 'random.seed', ([], {}), '()\n', (1708, 1710), False,... |
"""Main application initilization."""
import os.path
from flask import Flask
from flask_sqlalchemy import SQLAlchemy
from flask_login import LoginManager
BOOK_PATH = os.path.normpath(os.path.join(os.path.dirname(__file__), '..', 'books'))
# Make sure to add your own secret key in config.py
SECRET_KEY = "<KEY>"... | [
"flask_sqlalchemy.SQLAlchemy",
"flask_login.LoginManager",
"flask.Flask"
] | [((530, 545), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (535, 545), False, 'from flask import Flask\n'), ((668, 683), 'flask_sqlalchemy.SQLAlchemy', 'SQLAlchemy', (['app'], {}), '(app)\n', (678, 683), False, 'from flask_sqlalchemy import SQLAlchemy\n'), ((733, 747), 'flask_login.LoginManager', 'LoginM... |
# Copyright 2021 Arm Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agree... | [
"tensorflow.keras.layers.Conv2D",
"tensorflow.pad",
"tensorflow.eye",
"numpy.zeros",
"models.quantize_utils.compute_ranges",
"tensorflow.transpose",
"tensorflow.constant",
"tensorflow.Variable",
"tensorflow.nn.conv2d",
"tensorflow.keras.Sequential",
"tensorflow.keras.layers.Layer.__init__",
"t... | [((2255, 2282), 'tensorflow.keras.Sequential', 'tf.keras.Sequential', (['layers'], {}), '(layers)\n', (2274, 2282), True, 'import tensorflow as tf\n'), ((2789, 2825), 'tensorflow.keras.layers.Layer.__init__', 'tf.keras.layers.Layer.__init__', (['self'], {}), '(self)\n', (2819, 2825), True, 'import tensorflow as tf\n'),... |
import random
from LogicClass.MonteCarloClass.ArborescenteTree import ArborescenteTree
class MonteCarloMove:
def __init__(self, isInspector, isPhantom, numberOfRoom):
self.tree = ArborescenteTree()
self.isPhantom = isPhantom
self.isInspector = isInspector
self.numberOfRoom = number... | [
"random.randint",
"LogicClass.MonteCarloClass.ArborescenteTree.ArborescenteTree"
] | [((193, 211), 'LogicClass.MonteCarloClass.ArborescenteTree.ArborescenteTree', 'ArborescenteTree', ([], {}), '()\n', (209, 211), False, 'from LogicClass.MonteCarloClass.ArborescenteTree import ArborescenteTree\n'), ((2441, 2481), 'random.randint', 'random.randint', (['(0)', '(self.numberOfRoom - 1)'], {}), '(0, self.num... |
"""
Tests for the Paver commands for updating test databases and its utility methods
"""
import os
import shutil
import tarfile
from tempfile import mkdtemp
from unittest import TestCase
from unittest.mock import call, patch, Mock
import boto
from pavelib import database
from pavelib.utils import db_utils
from pave... | [
"unittest.mock.patch.object",
"pavelib.utils.db_utils.extract_files_from_zip",
"unittest.mock.Mock",
"pavelib.database.update_local_bokchoy_db_from_s3",
"unittest.mock.patch",
"unittest.mock.call",
"os.path.isfile",
"tempfile.mkdtemp",
"tarfile.open",
"os.path.join",
"boto.s3.key.Key"
] | [((530, 580), 'unittest.mock.patch', 'patch', (['"""pavelib.utils.db_utils.verify_files_exist"""'], {}), "('pavelib.utils.db_utils.verify_files_exist')\n", (535, 580), False, 'from unittest.mock import call, patch, Mock\n'), ((2404, 2432), 'unittest.mock.patch.object', 'patch.object', (['db_utils', '"""sh"""'], {}), "(... |
from functools import lru_cache
from typing import Optional
from pygame.freetype import get_default_font, SysFont
font_cache = {}
@lru_cache(100)
def get_font(fontname: Optional[str] = None, size: int = 12, bold: bool = False, italic: bool = False):
if fontname is None:
fontname = get_default_font()
... | [
"pygame.freetype.get_default_font",
"functools.lru_cache",
"pygame.freetype.SysFont"
] | [((136, 150), 'functools.lru_cache', 'lru_cache', (['(100)'], {}), '(100)\n', (145, 150), False, 'from functools import lru_cache\n'), ((329, 378), 'pygame.freetype.SysFont', 'SysFont', (['fontname', 'size'], {'bold': 'bold', 'italic': 'italic'}), '(fontname, size, bold=bold, italic=italic)\n', (336, 378), False, 'from... |
import requests
def get_event(user_key, latitude, longitude):
url = "http://api.eventful.com/json/events/search?"
url += "&app_key=" + user_key
url += "&date=Future" #+ date
url += "&page_size=100"
url += "&sort_order=popularity"
url += "&sort_direction=descending"
url += "&q=music"
u... | [
"requests.get"
] | [((417, 434), 'requests.get', 'requests.get', (['url'], {}), '(url)\n', (429, 434), False, 'import requests\n')] |
from typing import List, Text, Tuple
import logging
import re
import lxml
from .. import utils
MAIN_PAGE = "Wikiquote:Accueil"
logger = logging.getLogger(__name__)
logger.addHandler(logging.NullHandler())
def extract_quotes(tree: lxml.html.HtmlElement, max_quotes: int) -> List[Text]:
# French wiki uses a "ci... | [
"re.search",
"logging.getLogger",
"logging.NullHandler"
] | [((141, 168), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (158, 168), False, 'import logging\n'), ((187, 208), 'logging.NullHandler', 'logging.NullHandler', ([], {}), '()\n', (206, 208), False, 'import logging\n'), ((1088, 1118), 're.search', 're.search', (['"""«(.+?)»(.+)"""', 'line']... |
# Generated by Django 3.2.6 on 2022-02-14 13:25
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
("database", "0060_set_ordering_on_tablewebhook_models"),
]
operations = [
migrations.AlterField(
model_name="numberfield",
... | [
"django.db.models.IntegerField"
] | [((374, 563), 'django.db.models.IntegerField', 'models.IntegerField', ([], {'choices': "[(0, '1'), (1, '1.0'), (2, '1.00'), (3, '1.000'), (4, '1.0000'), (5, '1.00000')\n ]", 'default': '(0)', 'help_text': '"""The amount of digits allowed after the point."""'}), "(choices=[(0, '1'), (1, '1.0'), (2, '1.00'), (3, '1.00... |
from IPython import display
from torch.utils.data import DataLoader
from torchvision import transforms, datasets
import os
import tensorflow as tf
from tensorflow import nn, layers
from tensorflow.contrib import layers as clayers
import numpy as np
import errno
import torchvision.utils as vutils
from tensorboardX impo... | [
"tensorflow.trainable_variables",
"tensorflow.contrib.layers.flatten",
"tensorflow.reshape",
"tensorflow.zeros_like",
"torchvision.datasets.CIFAR10",
"matplotlib.pyplot.figure",
"numpy.random.normal",
"tensorflow.layers.conv2d_transpose",
"torchvision.transforms.Normalize",
"tensorflow.nn.leaky_re... | [((989, 1045), 'torch.utils.data.DataLoader', 'DataLoader', (['dataset'], {'batch_size': 'batch_size', 'shuffle': '(True)'}), '(dataset, batch_size=batch_size, shuffle=True)\n', (999, 1045), False, 'from torch.utils.data import DataLoader\n'), ((3937, 3993), 'tensorflow.placeholder', 'tf.placeholder', (['tf.float32'], ... |
# -*- coding: UTF-8 -*-
# Copyright 2013-2017 <NAME>
# License: BSD (see file COPYING for details)
"""See :doc:`/specs/vat`.
.. autosummary::
:toctree:
utils
.. fixtures.novat fixtures.euvatrates
"""
from django.utils.translation import ugettext_lazy as _
from lino.api import ad
import six
class Plugin(a... | [
"django.utils.translation.ugettext_lazy"
] | [((440, 448), 'django.utils.translation.ugettext_lazy', '_', (['"""VAT"""'], {}), "('VAT')\n", (441, 448), True, 'from django.utils.translation import ugettext_lazy as _\n')] |
from flask import Blueprint
admin = Blueprint('admin', __name__, url_prefix='/admin',
template_folder='templates')
| [
"flask.Blueprint"
] | [((38, 116), 'flask.Blueprint', 'Blueprint', (['"""admin"""', '__name__'], {'url_prefix': '"""/admin"""', 'template_folder': '"""templates"""'}), "('admin', __name__, url_prefix='/admin', template_folder='templates')\n", (47, 116), False, 'from flask import Blueprint\n')] |
# Generated by Django 2.1.5 on 2019-01-29 23:48
from django.db import migrations, models
class Migration(migrations.Migration):
initial = True
dependencies = [
]
operations = [
migrations.CreateModel(
name='Call',
fields=[
('id', models.AutoField(aut... | [
"django.db.models.CharField",
"django.db.models.IntegerField",
"django.db.models.AutoField"
] | [((300, 393), 'django.db.models.AutoField', 'models.AutoField', ([], {'auto_created': '(True)', 'primary_key': '(True)', 'serialize': '(False)', 'verbose_name': '"""ID"""'}), "(auto_created=True, primary_key=True, serialize=False,\n verbose_name='ID')\n", (316, 393), False, 'from django.db import migrations, models\... |
import csv
import logging
import numpy as np
import datajoint as dj
import pathlib
import scipy.io as scio
from tifffile import imread
from . import InsertBuffer
from .reference import ccf_ontology
from . import get_schema_name
schema = dj.schema(get_schema_name('ccf'))
log = logging.getLogger(__name__)
@schem... | [
"scipy.io.loadmat",
"pathlib.Path",
"numpy.where",
"tifffile.imread",
"datajoint.conn",
"logging.getLogger"
] | [((284, 311), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (301, 311), False, 'import logging\n'), ((1964, 1982), 'tifffile.imread', 'imread', (['stack_path'], {}), '(stack_path)\n', (1970, 1982), False, 'from tifffile import imread\n'), ((3995, 4057), 'scipy.io.loadmat', 'scio.loadmat'... |
import os
import discord
import requests
import json
from dotenv import load_dotenv
from discord.ext import commands
from datetime import datetime
load_dotenv()
TOKEN = os.getenv('DISCORD_TOKEN')
API_KEY = os.getenv('API_KEY')
HEADERS = {
"x-api-key" : API_KEY
}
bot = commands.Bot(command_prefix = "-")
ROOT_URL... | [
"discord.Color.blurple",
"dotenv.load_dotenv",
"requests.get",
"discord.ext.commands.Bot",
"datetime.datetime.now",
"os.getenv"
] | [((148, 161), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (159, 161), False, 'from dotenv import load_dotenv\n'), ((170, 196), 'os.getenv', 'os.getenv', (['"""DISCORD_TOKEN"""'], {}), "('DISCORD_TOKEN')\n", (179, 196), False, 'import os\n'), ((207, 227), 'os.getenv', 'os.getenv', (['"""API_KEY"""'], {}), "('... |
import logging
from common import settings
from elasticapm.contrib.flask import ElasticAPM
log = logging.getLogger(__name__)
def configure_app(flask_app):
flask_app.config.SWAGGER_UI_DOC_EXPANSION = settings.RESTPLUS_SWAGGER_UI_DOC_EXPANSION
flask_app.config.RESTPLUS_VALIDATE = settings.RESTPLUS_VALIDATE
... | [
"elasticapm.contrib.flask.ElasticAPM",
"logging.getLogger"
] | [((99, 126), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (116, 126), False, 'import logging\n'), ((1040, 1093), 'elasticapm.contrib.flask.ElasticAPM', 'ElasticAPM', (['flask_app'], {'logging': 'settings.APM_LOG_LEVEL'}), '(flask_app, logging=settings.APM_LOG_LEVEL)\n', (1050, 1093), Fa... |
"""CLI for data preparation and processing."""
import argparse
from utils import data_prep
from utils import read_one_row
from utils import save_input
parser = argparse.ArgumentParser()
parser.add_argument(
"--save_row",
type=int,
default="0",
help="Saves a single row to a file defaults to row 0",
)
p... | [
"utils.save_input",
"utils.data_prep",
"argparse.ArgumentParser",
"utils.read_one_row"
] | [((162, 187), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (185, 187), False, 'import argparse\n'), ((837, 858), 'utils.data_prep', 'data_prep', (['input_file'], {}), '(input_file)\n', (846, 858), False, 'from utils import data_prep\n'), ((865, 899), 'utils.read_one_row', 'read_one_row', (['s... |
import os
import sys
import unittest
from io import StringIO
import bpy
import tests.test_utils as tutils
from blendernc.preferences import get_addon_preference
@tutils.refresh_state
def create_nodes(file, var):
node_groups = bpy.data.node_groups
if tutils.is_blendernc_in_nodetree(node_groups):
node... | [
"io.StringIO",
"bpy.ops.image.new",
"os.path.abspath",
"tests.test_utils.is_blendernc_in_nodetree",
"unittest.TextTestRunner",
"tests.test_utils.build_dict_blendernc_prop",
"bpy.context.scene.frame_set",
"bpy.data.node_groups.new",
"unittest.defaultTestLoader.loadTestsFromTestCase",
"tests.test_ut... | [((4711, 4774), 'unittest.defaultTestLoader.loadTestsFromTestCase', 'unittest.defaultTestLoader.loadTestsFromTestCase', (['Test_settings'], {}), '(Test_settings)\n', (4759, 4774), False, 'import unittest\n'), ((4851, 4864), 'sys.exit', 'sys.exit', (['ret'], {}), '(ret)\n', (4859, 4864), False, 'import sys\n'), ((262, 3... |
from __future__ import print_function
import os
import warnings
import pytest
from graphene_django.utils.testing import graphql_query
from graphql_jwt.settings import jwt_settings
from graphql_jwt.shortcuts import get_token
from neo4j.exceptions import ClientError as CypherError
from neobolt.exceptions import ClientE... | [
"warnings.simplefilter",
"graphene_django.utils.testing.graphql_query",
"neomodel.db.set_connection",
"graphql_jwt.shortcuts.get_token",
"pytest.fixture",
"kaffepause.users.test.factories.UserFactory",
"neomodel.clear_neo4j_database",
"os.environ.get",
"warnings.warn",
"kaffepause.accounts.test.fa... | [((612, 640), 'pytest.fixture', 'pytest.fixture', ([], {'autouse': '(True)'}), '(autouse=True)\n', (626, 640), False, 'import pytest\n'), ((3998, 4026), 'pytest.fixture', 'pytest.fixture', ([], {'autouse': '(True)'}), '(autouse=True)\n', (4012, 4026), False, 'import pytest\n'), ((4167, 4195), 'pytest.fixture', 'pytest.... |
# Copyright (C) 2020 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writi... | [
"proto.Field"
] | [((749, 783), 'proto.Field', 'proto.Field', (['proto.INT32'], {'number': '(1)'}), '(proto.INT32, number=1)\n', (760, 783), False, 'import proto\n'), ((1009, 1043), 'proto.Field', 'proto.Field', (['proto.INT32'], {'number': '(1)'}), '(proto.INT32, number=1)\n', (1020, 1043), False, 'import proto\n'), ((1251, 1285), 'pro... |
import numpy as np
import pandas as pd
import sys
import os
from utils import DATA_DIR
class Dataset(object):
def __init__(self, DATA_NAME):
self.DATA_NAME = DATA_NAME
print("Initializing dataset:", DATA_NAME)
sys.stdout.flush()
data = pd.read_csv(os.path.join(DATA_DIR, "... | [
"numpy.sum",
"numpy.ones",
"numpy.array",
"sys.stdout.flush",
"numpy.random.choice",
"pandas.factorize",
"os.path.join"
] | [((249, 267), 'sys.stdout.flush', 'sys.stdout.flush', ([], {}), '()\n', (265, 267), False, 'import sys\n'), ((500, 536), 'pandas.factorize', 'pd.factorize', (["data['item_id'].values"], {}), "(data['item_id'].values)\n", (512, 536), True, 'import pandas as pd\n'), ((570, 606), 'pandas.factorize', 'pd.factorize', (["dat... |
import setuptools
with open("README.md", "r") as fh:
long_description = fh.read()
setuptools.setup(
name='HPexome',
version='1.2.1',
author="<NAME>",
author_email="<EMAIL>",
description="An automated tool for processing whole-exome sequencing data",
long_description=long_description,
l... | [
"setuptools.find_packages"
] | [((416, 442), 'setuptools.find_packages', 'setuptools.find_packages', ([], {}), '()\n', (440, 442), False, 'import setuptools\n')] |
import os
import random
import syft as sy
import pandas as pd
import numpy as np
from PIL import Image
from tqdm import tqdm
from torch import ( # pylint:disable=no-name-in-module
manual_seed,
stack,
cat,
std_mean,
save,
is_tensor,
from_numpy,
randperm,
default_generator,
)
from tor... | [
"numpy.isin",
"albumentations.Lambda",
"albumentations.RandomSunFlare",
"albumentations.GaussNoise",
"albumentations.Resize",
"pandas.read_csv",
"albumentations.RandomShadow",
"torch.cat",
"albumentations.RandomFog",
"numpy.mean",
"albumentations.Normalize",
"torch.std_mean",
"os.path.join",... | [((4313, 4474), 'torchvision.transforms.RandomAffine', 'transforms.RandomAffine', ([], {'degrees': 'args.rotation', 'translate': '(args.translate, args.translate)', 'scale': '(1.0 - args.scale, 1.0 + args.scale)', 'shear': 'args.shear'}), '(degrees=args.rotation, translate=(args.translate,\n args.translate), scale=(... |
from math import trunc
def two_decimal_places(number):
factor = float(10 ** 2)
return trunc(number * factor) / factor
| [
"math.trunc"
] | [((96, 118), 'math.trunc', 'trunc', (['(number * factor)'], {}), '(number * factor)\n', (101, 118), False, 'from math import trunc\n')] |
import warnings
warnings.simplefilter("ignore", UserWarning)
warnings.simplefilter("ignore", FutureWarning)
import argparse
import os
import pandas as pd
import numpy as np
from torch import nn
from torch.utils.data import DataLoader
from tqdm import tqdm
from collections import defaultdict
from catalyst.utils impo... | [
"numpy.random.seed",
"pandas.DataFrame.from_dict",
"warnings.simplefilter",
"argparse.ArgumentParser",
"torch.utils.data.DataLoader",
"catalyst.utils.any2device",
"os.path.exists",
"collections.defaultdict",
"pytorch_toolbelt.utils.fs.change_extension",
"os.environ.get",
"pytorch_toolbelt.utils.... | [((18, 62), 'warnings.simplefilter', 'warnings.simplefilter', (['"""ignore"""', 'UserWarning'], {}), "('ignore', UserWarning)\n", (39, 62), False, 'import warnings\n'), ((63, 109), 'warnings.simplefilter', 'warnings.simplefilter', (['"""ignore"""', 'FutureWarning'], {}), "('ignore', FutureWarning)\n", (84, 109), False,... |
from django.db import models
class Course(models.Model):
created = models.DateTimeField(auto_now_add=True)
title = models.CharField(max_length=100, blank=True, default='')
content = models.TextField()
owner = models.ForeignKey('auth.User', related_name='Course', on_delete=models.CASCADE)
class M... | [
"django.db.models.CharField",
"django.db.models.DateTimeField",
"django.db.models.TextField",
"django.db.models.ForeignKey"
] | [((73, 112), 'django.db.models.DateTimeField', 'models.DateTimeField', ([], {'auto_now_add': '(True)'}), '(auto_now_add=True)\n', (93, 112), False, 'from django.db import models\n'), ((125, 181), 'django.db.models.CharField', 'models.CharField', ([], {'max_length': '(100)', 'blank': '(True)', 'default': '""""""'}), "(m... |
#! /usr/bin/env python3
import sys
import csv
import argparse
import numpy as np
import pandas as pd
from mll_calc.all_jobs import parent_jobs, kid_jobs
def row_calcs(ext_test):
if 'no' in ext_test:
#db_rows = 450240
#max_jobs = 9750
db_rows = 90048 * 4
max_jobs = 978 * 4
else:... | [
"numpy.arange",
"csv.writer",
"sys.exit"
] | [((993, 1059), 'sys.exit', 'sys.exit', (['"""total expected jobs does not equal one of db_row lists"""'], {}), "('total expected jobs does not equal one of db_row lists')\n", (1001, 1059), False, 'import sys\n'), ((414, 443), 'numpy.arange', 'np.arange', (['(0)', 'db_rows', 'n_rows'], {}), '(0, db_rows, n_rows)\n', (42... |
#!/usr/bin/env python3
import os, sys
import setuptools
# Get text from README.txt
with open("README.md", "r") as fp:
readme_text = fp.read()
# Get __version__ without importing
with open(os.path.join(os.path.dirname(__file__),"ndsb", "__init__.py"), "r") as f:
for line in f:
if line.startswith("__ver... | [
"os.path.dirname",
"setuptools.setup"
] | [((383, 1072), 'setuptools.setup', 'setuptools.setup', ([], {'name': '"""ndsb"""', 'version': '__version__', 'description': '"""Collect data, turn it into static artifacts and beam it to a vault."""', 'license': '"""MIT"""', 'author': '"""<NAME>"""', 'author_email': '"""<EMAIL>"""', 'url': '"""https://github.com/dbbs-l... |
# import the needed packages
import pickle
from sklearn import preprocessing
import time
from os import listdir
from os.path import isfile, join
from random import randint, uniform
import numpy as np
from matplotlib import pyplot as plt
import cv2 as cv
from scipy import ndimage
from skimage import morphol... | [
"cv2.GaussianBlur",
"numpy.random.seed",
"cv2.bitwise_and",
"numpy.argmax",
"scipy.ndimage.binary_fill_holes",
"numpy.ones",
"numpy.exp",
"cv2.imshow",
"cv2.inRange",
"numpy.unique",
"pandas.DataFrame",
"cv2.contourArea",
"numpy.zeros_like",
"random.randint",
"cv2.cvtColor",
"scipy.ndi... | [((622, 640), 'numpy.random.seed', 'np.random.seed', (['(26)'], {}), '(26)\n', (636, 640), True, 'import numpy as np\n'), ((1152, 1169), 'numpy.unique', 'np.unique', (['target'], {}), '(target)\n', (1161, 1169), True, 'import numpy as np\n'), ((3827, 3865), 'numpy.exp', 'np.exp', (['(-((x - mean) ** 2 / (2 * var)))'], ... |
############################################################
# -*- coding: utf-8 -*-
#
# # # # # # ####
# ## ## # ## # #
# # # # # # # # # ###
# # ## # ## ## #
# # # # # # ####
#
# Python-based Tool for interaction with the 10micron mounts
# GUI with PyQT5 ... | [
"os.path.basename",
"os.getcwd",
"copy.copy",
"numpy.clip",
"astrometry.transform.Transform",
"numpy.interp",
"operator.itemgetter",
"logging.getLogger"
] | [((621, 648), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (638, 648), False, 'import logging\n'), ((727, 756), 'astrometry.transform.Transform', 'transform.Transform', (['self.app'], {}), '(self.app)\n', (746, 756), False, 'from astrometry import transform\n'), ((13679, 13807), 'numpy.... |
# Project: hardInfo
# Author: <NAME>
# Date Started: March 18, 2022
# Copyright: (c) Copyright 2022 <NAME>
# Module: model/LsBlk.py
# Date Started: March 23, 2022
# Purpose: Store and provide API for Linux lsblk command.
# Development:
# Arguments to include ... | [
"subprocess.Popen",
"json.loads",
"view.Components.JsonTreeView",
"tkinter.messagebox.askyesno",
"tkinter.LabelFrame",
"tkinter.Tk"
] | [((1875, 1954), 'tkinter.messagebox.askyesno', 'messagebox.askyesno', (['"""Exit program """', "('Exit the ' + PROGRAM_TITLE + ' program?')"], {}), "('Exit program ', 'Exit the ' + PROGRAM_TITLE + ' program?')\n", (1894, 1954), False, 'from tkinter import Tk, messagebox, LabelFrame, BOTH, RAISED\n'), ((2041, 2045), 'tk... |
import argparse
def encrypt(message: str, key: dict[int, int]) -> str:
encrypted = map(lambda char: key[char], message)
return encrypted
def decrypt(message: str, key: dict[int, int]) -> str:
decrypted = map(lambda char: key[char], message)
return decrypted
def main():
parser = argparse.ArgumentParser()
... | [
"argparse.ArgumentParser"
] | [((293, 318), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (316, 318), False, 'import argparse\n')] |
"""
Obit Plotting class
Create a plot object using newOPlot which allows specifying the output
and background color. If no output is specified this information
will be prompted.
Next, the plotting region must be specified using either PSetPlot,
one of the XY plotting routines (PXYPlot, PXYOver, or PXYErr)
PGrayScal... | [
"Obit.PlotInitPlot",
"Obit.PlotXYPlot",
"Obit.CreateOPlot",
"Obit.PlotSetPage",
"Obit.PlotXYErr",
"Obit.PlotDrawSymbol",
"Obit.PlotSetColor",
"Obit.PlotText",
"InfoList.InfoList",
"Obit.PlotSetLineStyle",
"Obit.PlotContour",
"math.cos",
"Obit.OPlot_Get_me",
"Obit.PlotGetList",
"Obit.OPlo... | [((5153, 5211), 'Obit.PlotInitPlot', 'Obit.PlotInitPlot', (['out.me', 'output', 'bgcolor', 'nx', 'ny', 'err.me'], {}), '(out.me, output, bgcolor, nx, ny, err.me)\n', (5170, 5211), False, 'import Obit, _Obit, InfoList, Image\n'), ((7952, 8001), 'Obit.PlotXYPlot', 'Obit.PlotXYPlot', (['plot.me', 'symbol', 'n', 'x', 'y', ... |
# --------------
#Code starts here
import sys
def palindrome(num):
numstr = str(num)
for i in range(num+1,sys.maxsize):
if str(i)== str(i)[::-1]:
return i
palindrome(123)
# --------------
#Code starts here
from collections import Counter
def a_scramble(str_1,str_2):
list_str1 = Counter(st... | [
"math.sqrt"
] | [((630, 642), 'math.sqrt', 'math.sqrt', (['x'], {}), '(x)\n', (639, 642), False, 'import math\n')] |
from utils import open_image, open_greyscale_bmp
from workspace_calc import WorkspaceCalculator
from workspace_view import WorkspaceView
class Workspace:
def __init__(self, app_page, room_name, robot_name):
room_bmp = open_greyscale_bmp(room_name)
robot_bmp = open_greyscale_bmp(robot_name)
... | [
"utils.open_greyscale_bmp",
"workspace_view.WorkspaceView",
"workspace_calc.WorkspaceCalculator",
"utils.open_image"
] | [((232, 261), 'utils.open_greyscale_bmp', 'open_greyscale_bmp', (['room_name'], {}), '(room_name)\n', (250, 261), False, 'from utils import open_image, open_greyscale_bmp\n'), ((282, 312), 'utils.open_greyscale_bmp', 'open_greyscale_bmp', (['robot_name'], {}), '(robot_name)\n', (300, 312), False, 'from utils import ope... |
# Generated by Django 2.2.13 on 2021-06-08 10:08
import os
from django.db import migrations
def create_premier_tenant(apps, schema_editor):
# We can't import the Person model directly as it may be a newer
# version than this migration expects. We use the historical version.
Client = apps.get_model('Custom... | [
"django.db.migrations.RunPython",
"os.getenv"
] | [((397, 416), 'os.getenv', 'os.getenv', (['"""DOMAIN"""'], {}), "('DOMAIN')\n", (406, 416), False, 'import os\n'), ((1645, 1697), 'django.db.migrations.RunPython', 'migrations.RunPython', (['create_premier_tenant', 'reverse'], {}), '(create_premier_tenant, reverse)\n', (1665, 1697), False, 'from django.db import migrat... |
from common.input_validation import (
extract_phone_number,
)
def test_extract_phone_number():
assert extract_phone_number('510501622') == None
assert extract_phone_number('5105016227') == '15105016227'
assert extract_phone_number('15105016227') == '15105016227'
assert extract_phone_number('+15105... | [
"common.input_validation.extract_phone_number"
] | [((112, 145), 'common.input_validation.extract_phone_number', 'extract_phone_number', (['"""510501622"""'], {}), "('510501622')\n", (132, 145), False, 'from common.input_validation import extract_phone_number\n'), ((165, 199), 'common.input_validation.extract_phone_number', 'extract_phone_number', (['"""5105016227"""']... |
import unittest
import math
import datasets
from pdffigures_utils import get_num_pages_in_pdf
class TestDataset(unittest.TestCase):
def test_pages_annotated_consistency(self):
for dataset in datasets.DATASETS.values():
dataset = dataset()
pages_annotated = dataset.get_annotated_pa... | [
"unittest.main",
"datasets.DATASETS.values",
"pdffigures_utils.get_num_pages_in_pdf",
"math.ceil",
"datasets.DatasetPartition"
] | [((2637, 2652), 'unittest.main', 'unittest.main', ([], {}), '()\n', (2650, 2652), False, 'import unittest\n'), ((206, 232), 'datasets.DATASETS.values', 'datasets.DATASETS.values', ([], {}), '()\n', (230, 232), False, 'import datasets\n'), ((1464, 1490), 'datasets.DATASETS.values', 'datasets.DATASETS.values', ([], {}), ... |
# -*- coding: utf-8 -*-
import botocore
import boto3
import io
from datetime import datetime
import s3Uploader
# Refs : https://boto3.readthedocs.io/en/latest/reference/services/s3.html
s3 = boto3.client('s3')
def main():
# [追加する時]
# バケットがなければ作成
# あればそれを使う。
# ファイルの重複チェック
# 重複していれば、削除し更新
# 重... | [
"s3Uploader.s3Uploader",
"datetime.datetime.now",
"boto3.client"
] | [((194, 212), 'boto3.client', 'boto3.client', (['"""s3"""'], {}), "('s3')\n", (206, 212), False, 'import boto3\n'), ((1821, 1881), 's3Uploader.s3Uploader', 's3Uploader.s3Uploader', (['bucketName', 'objectName', '"""./image.jpg"""'], {}), "(bucketName, objectName, './image.jpg')\n", (1842, 1881), False, 'import s3Upload... |
from django.test import TestCase
from majority_judgment.tools import get_ranking, get_ratings, majority_grade
class MajorityJudgmentTestCase(TestCase):
fixtures = ['election.json']
# def setUp(self):
def test_ranking(self):
election_id = 2
ranking = get_ranking(election_id)
ranki... | [
"majority_judgment.tools.majority_grade",
"majority_judgment.tools.get_ranking"
] | [((282, 306), 'majority_judgment.tools.get_ranking', 'get_ranking', (['election_id'], {}), '(election_id)\n', (293, 306), False, 'from majority_judgment.tools import get_ranking, get_ratings, majority_grade\n'), ((584, 608), 'majority_judgment.tools.get_ranking', 'get_ranking', (['election_id'], {}), '(election_id)\n',... |
from typing import List, Union
import pandas as pd
from zvdata import IntervalLevel
from zvt.api.common import get_kdata_schema
from zvt.factors.algorithm import MacdTransformer, MaTransformer
from zvt.factors.factor import Factor, Transformer, Accumulator
class TechnicalFactor(Factor):
def __init__(self,
... | [
"zvt.factors.algorithm.MacdTransformer",
"zvt.api.common.get_kdata_schema"
] | [((1309, 1326), 'zvt.factors.algorithm.MacdTransformer', 'MacdTransformer', ([], {}), '()\n', (1324, 1326), False, 'from zvt.factors.algorithm import MacdTransformer, MaTransformer\n'), ((1500, 1542), 'zvt.api.common.get_kdata_schema', 'get_kdata_schema', (['entity_type'], {'level': 'level'}), '(entity_type, level=leve... |
# Generated by Django 2.2.12 on 2020-06-16 13:14
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('memes', '0001_initial'),
]
operations = [
migrations.RemoveField(
model_name='task',
... | [
"django.db.migrations.RemoveField",
"django.db.models.URLField",
"django.db.models.IntegerField",
"django.db.models.ForeignKey"
] | [((256, 310), 'django.db.migrations.RemoveField', 'migrations.RemoveField', ([], {'model_name': '"""task"""', 'name': '"""name"""'}), "(model_name='task', name='name')\n", (278, 310), False, 'from django.db import migrations, models\n'), ((456, 482), 'django.db.models.URLField', 'models.URLField', ([], {'null': '(True)... |
from entities.remote_radio_head import RemoteRadioHead
from entities.hypervisor import Hypervisor
from entities.baseband_unit import BasebandUnit
from entities.switch import Switch
from forwarding.forwarding import Forwarding
class StatHistory(object):
history = {}
def get(key, current):
if (key in St... | [
"forwarding.forwarding.Forwarding",
"entities.remote_radio_head.RemoteRadioHead",
"entities.baseband_unit.BasebandUnit",
"entities.switch.Switch",
"entities.hypervisor.Hypervisor"
] | [((651, 677), 'forwarding.forwarding.Forwarding', 'Forwarding', (['self.env', 'self'], {}), '(self.env, self)\n', (661, 677), False, 'from forwarding.forwarding import Forwarding\n'), ((6337, 6377), 'entities.switch.Switch', 'Switch', (['self.env', '"""physical"""', '"""external"""'], {}), "(self.env, 'physical', 'exte... |
# coding=utf-8
# Author: <NAME> & <NAME>
# Date: Jan 06, 2021
#
# Description: Utility functions
#
import os
import re
import functools
import pickle
import numpy as np
#
# Functions to handle Twitter text
#
re_all_after_retweet = re.compile(r"rt @[a-zA-Z0-9_]+.+", re.IGNORECASE | re.UNICODE)
def removeAllAfterRetwe... | [
"pickle.dump",
"os.makedirs",
"os.path.dirname",
"os.path.exists",
"os.path.isfile",
"pickle.load",
"functools.wraps",
"re.compile"
] | [((232, 293), 're.compile', 're.compile', (['"""rt @[a-zA-Z0-9_]+.+"""', '(re.IGNORECASE | re.UNICODE)'], {}), "('rt @[a-zA-Z0-9_]+.+', re.IGNORECASE | re.UNICODE)\n", (242, 293), False, 'import re\n'), ((499, 613), 're.compile', 're.compile', (['"""(#Repost @\\\\w+ with @repostapp)|(#EzRepost @\\\\w+ with @ezrepostapp... |
import hashlib
import sys
if sys.version_info[0] < 3:
import urllib
else:
import urllib.parse as urllib
class Gravtr(object):
GRAVATAR_URL = 'https://www.gravatar.com/avatar/'
GRAVATAR_URL_UNSECURE = 'http://www.gravatar.com/avatar/'
class ratingType(object):
G = 'g'
PG = 'pg'
... | [
"hashlib.md5",
"urllib.parse.urlencode"
] | [((1050, 1074), 'urllib.parse.urlencode', 'urllib.urlencode', (['params'], {}), '(params)\n', (1066, 1074), True, 'import urllib.parse as urllib\n'), ((664, 687), 'hashlib.md5', 'hashlib.md5', (['self.email'], {}), '(self.email)\n', (675, 687), False, 'import hashlib\n')] |
import os
import unittest
from MuseParse.tests.testUsingXML.xmlSet import xmlSet, parsePiece
from MuseParse.classes.ObjectHierarchy.TreeClasses.BaseTree import Search, FindByIndex
from MuseParse.classes.ObjectHierarchy.TreeClasses.NoteNode import NoteNode
from MuseParse.classes.ObjectHierarchy.TreeClasses.MeasureNode ... | [
"MuseParse.classes.ObjectHierarchy.TreeClasses.BaseTree.Search",
"os.path.join",
"MuseParse.tests.testUsingXML.xmlSet.xmlSet.setUp"
] | [((550, 583), 'os.path.join', 'os.path.join', (['directory', 'partname'], {}), '(directory, partname)\n', (562, 583), False, 'import os\n'), ((642, 660), 'MuseParse.tests.testUsingXML.xmlSet.xmlSet.setUp', 'xmlSet.setUp', (['self'], {}), '(self)\n', (654, 660), False, 'from MuseParse.tests.testUsingXML.xmlSet import xm... |
import os
import pathlib
from ctypes import *
from ctypes import _SimpleCData
from ctypes import _Pointer
from .common import CEnum
class SNResult(CEnum):
SN_S_OK = (0)
SN_S_PENDING = (1)
SN_S_NO_MSG = (3)
SN_S_TM_VERSION = (4)
SN_S_REPLACED = (5)
SN_S_NO_ACTION = (6)
SN_S_CONNECTED = SN_S... | [
"os.getcwd",
"os.getenv"
] | [((2153, 2164), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (2162, 2164), False, 'import os\n'), ((2208, 2232), 'os.getenv', 'os.getenv', (['"""SN_PS3_PATH"""'], {}), "('SN_PS3_PATH')\n", (2217, 2232), False, 'import os\n')] |
import sys
import unittest
import os
script_dir = os.path.dirname(os.path.realpath(__file__))
sys.path.insert(1, os.path.abspath(
os.path.join(script_dir, os.path.join('..', '..'))))
from pake import process
import pake.program
import pake
class ProcessTest(unittest.TestCase):
def test_call(self):
... | [
"pake.de_init",
"pake.process.check_output",
"os.path.realpath",
"pake.process.call",
"pake.init",
"pake.process.check_call",
"os.path.join"
] | [((68, 94), 'os.path.realpath', 'os.path.realpath', (['__file__'], {}), '(__file__)\n', (84, 94), False, 'import os\n'), ((1840, 1870), 'pake.de_init', 'pake.de_init', ([], {'clear_conf': '(False)'}), '(clear_conf=False)\n', (1852, 1870), False, 'import pake\n'), ((1885, 1896), 'pake.init', 'pake.init', ([], {}), '()\n... |
# Copyright 2021 Huawei Technologies Co., Ltd
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to... | [
"src.loss.Quadrupletloss",
"argparse.ArgumentParser",
"mindspore.train.callback.ModelCheckpoint",
"src.dataset.create_dataset1",
"moxing.file.copy_parallel",
"numpy.mean",
"mindspore.train.serialization.load_checkpoint",
"os.path.join",
"mindspore.train.serialization.load_param_into_net",
"mindspo... | [((1586, 1597), 'mindspore.common.set_seed', 'set_seed', (['(1)'], {}), '(1)\n', (1594, 1597), False, 'from mindspore.common import set_seed\n'), ((1608, 1667), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Image classification"""'}), "(description='Image classification')\n", (1631, 166... |
# -*- coding: utf-8 -*-
"""testing script"""
import os
import sys
from functools import reduce
import numpy as np
import pandas as pd
import nltk # Natural Language Tool Kit
from fuzzywuzzy import fuzz, process # Fuzzy String Matching
import jellyfish # Distance metrics
from sklearn.feature_extraction.text import T... | [
"functools.reduce",
"sklearn.feature_extraction.text.TfidfVectorizer",
"sklearn.metrics.pairwise.linear_kernel"
] | [((3797, 3848), 'functools.reduce', 'reduce', (["(lambda x, y: f'{x} {y}')", '[x[0] for x in l1]'], {}), "(lambda x, y: f'{x} {y}', [x[0] for x in l1])\n", (3803, 3848), False, 'from functools import reduce\n'), ((3882, 3933), 'functools.reduce', 'reduce', (["(lambda x, y: f'{x} {y}')", '[x[0] for x in l1]'], {}), "(la... |
# Generated by Django 2.2 on 2019-05-23 12:31
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('dashboard', '0001_initial'),
]
operations = [
migrations.AlterField(
model_name='entrylist',
name='awards',
... | [
"django.db.models.CharField"
] | [((328, 398), 'django.db.models.CharField', 'models.CharField', ([], {'blank': '(True)', 'default': '"""未设置"""', 'max_length': '(200)', 'null': '(True)'}), "(blank=True, default='未设置', max_length=200, null=True)\n", (344, 398), False, 'from django.db import migrations, models\n'), ((527, 597), 'django.db.models.CharFie... |
from aparse import click
from viewformer.utils.click import LazyGroup
@click.group(cls=LazyGroup)
def main():
pass
@main.group(cls=LazyGroup)
def dataset():
pass
@main.group(cls=LazyGroup)
def visualize():
pass
@main.group(cls=LazyGroup)
def model():
pass
@main.group(cls=LazyGroup)
def evaluat... | [
"aparse.click.group"
] | [((73, 99), 'aparse.click.group', 'click.group', ([], {'cls': 'LazyGroup'}), '(cls=LazyGroup)\n', (84, 99), False, 'from aparse import click\n')] |
"""
Copyright (c) 2022 Huawei Technologies Co.,Ltd.
openGauss is licensed under Mulan PSL v2.
You can use this software according to the terms and conditions of the Mulan PSL v2.
You may obtain a copy of Mulan PSL v2 at:
http://license.coscl.org.cn/MulanPSL2
THIS SOFTWARE IS PROVIDED ON AN "AS IS" BASIS, W... | [
"yat.test.Node",
"testcase.utils.CommonSH.CommonSH",
"testcase.utils.Logger.Logger"
] | [((956, 974), 'testcase.utils.CommonSH.CommonSH', 'CommonSH', (['"""dbuser"""'], {}), "('dbuser')\n", (964, 974), False, 'from testcase.utils.CommonSH import CommonSH\n'), ((1000, 1014), 'yat.test.Node', 'Node', (['"""dbuser"""'], {}), "('dbuser')\n", (1004, 1014), False, 'from yat.test import Node\n'), ((1034, 1042), ... |
from GitLabApi import objects
from core.tests.test_view import LoginMethods
from core.tests.test_view import SimpleUrlsTestsCases
from django.db.models import QuerySet
from django.urls import reverse
from groups import models
from groups.sidebar import GroupSidebar, FutureGroupSidebar
from groups.tests import test_form... | [
"groups.models.AddSubgroup.objects.get",
"groups.tests.test_forms.SubgroupAndMembersFromFileFormTests.valid_file_data.items",
"groups.tests.test_forms.TaskGroupFormTests.valid_form_data.items",
"groups.tests.models.AddProjectCreateMethods",
"groups.models.AddMember.objects.get",
"groups.tests.test_forms.A... | [((7695, 7748), 'groups.tests.test_forms.TaskGroupFormTests.valid_form_data.items', 'test_forms.TaskGroupFormTests.valid_form_data.items', ([], {}), '()\n', (7746, 7748), False, 'from groups.tests import test_forms\n'), ((9189, 9247), 'groups.models.TaskGroup.objects.get', 'models.TaskGroup.objects.get', ([], {'parent_... |
#!/usr/bin/env python3
from api import DB
from api.models.listing import Listing
def get_all_queries():
"""
Returns all stored listing queries.
"""
return list(Listing.query.all())
| [
"api.models.listing.Listing.query.all"
] | [((183, 202), 'api.models.listing.Listing.query.all', 'Listing.query.all', ([], {}), '()\n', (200, 202), False, 'from api.models.listing import Listing\n')] |
import asyncio
import json
from datetime import date
from decimal import Decimal
from typing import Dict
from aiohttp import ClientResponse
from dynaconf import settings
_PRICE_HISTORY_API = f"{settings.BENZAK_API_URL}/price-history/"
async def load_price(logger, session, price: Dict):
logger.debug(
f"c... | [
"json.dumps"
] | [((411, 454), 'json.dumps', 'json.dumps', (['price'], {'indent': '(2)', 'sort_keys': '(True)'}), '(price, indent=2, sort_keys=True)\n', (421, 454), False, 'import json\n')] |
"""playnetmano_rm common internal object model"""
from oslo_utils import versionutils
from oslo_versionedobjects import base
from playnetmano_rm import objects
VersionedObjectDictCompat = base.VersionedObjectDictCompat
class Playnetmano_rmObject(base.VersionedObject):
"""Base class for playnetmano_rm objects.
... | [
"oslo_utils.versionutils.convert_version_to_tuple"
] | [((1417, 1467), 'oslo_utils.versionutils.convert_version_to_tuple', 'versionutils.convert_version_to_tuple', (['cls.VERSION'], {}), '(cls.VERSION)\n', (1454, 1467), False, 'from oslo_utils import versionutils\n')] |
import random
import datetime
import dateparser
from faker import Faker
from .base import Filth
class DateOfBirthFilth(Filth):
type = 'date_of_birth'
min_age_years = 18
max_age_years = 100
@staticmethod
def generate(faker: Faker) -> str:
"""Generates an example of this ``Filth`` type, us... | [
"dateparser.parse",
"datetime.date.today",
"random.choice"
] | [((1071, 1098), 'dateparser.parse', 'dateparser.parse', (['self.text'], {}), '(self.text)\n', (1087, 1098), False, 'import dateparser\n'), ((937, 959), 'random.choice', 'random.choice', (['formats'], {}), '(formats)\n', (950, 959), False, 'import random\n'), ((1193, 1214), 'datetime.date.today', 'datetime.date.today', ... |
import psycopg2
from psycopg2.extensions import connection, cursor
from psycopg2.extras import DictCursor
from typing import Dict
from src.log.logger import logger
from contextlib import contextmanager
@contextmanager
def get_connection(params: Dict[str, str]) -> connection:
"""
Get a connection using a cont... | [
"src.log.logger.logger.debug",
"psycopg2.connect"
] | [((467, 493), 'psycopg2.connect', 'psycopg2.connect', ([], {}), '(**params)\n', (483, 493), False, 'import psycopg2\n'), ((696, 739), 'src.log.logger.logger.debug', 'logger.debug', (['"""Closing database connection"""'], {}), "('Closing database connection')\n", (708, 739), False, 'from src.log.logger import logger\n')... |
import unittest
from test.test01 import soma
class TesteSoma(unittest.TestCase):
def test_retorno_soma_10_10(self):
self .assertEqual(soma(10, 10), 20)
| [
"test.test01.soma"
] | [((147, 159), 'test.test01.soma', 'soma', (['(10)', '(10)'], {}), '(10, 10)\n', (151, 159), False, 'from test.test01 import soma\n')] |
import pytest
from data_structures.heap import Heap
@pytest.fixture
def base_heap():
heap = Heap()
heap.push(1)
heap.push(2)
heap.push(3)
heap.push(4)
heap.push(5)
return heap
def test_heap_init():
basic_heap = Heap()
init_list_heap = Heap([9, 8, 7, 5, 1, 2])
assert isinstanc... | [
"data_structures.heap.Heap"
] | [((98, 104), 'data_structures.heap.Heap', 'Heap', ([], {}), '()\n', (102, 104), False, 'from data_structures.heap import Heap\n'), ((247, 253), 'data_structures.heap.Heap', 'Heap', ([], {}), '()\n', (251, 253), False, 'from data_structures.heap import Heap\n'), ((275, 299), 'data_structures.heap.Heap', 'Heap', (['[9, 8... |
import os
import json
import boto3
from botocore.exceptions import ClientError
from cryptography.fernet import Fernet
dynamodb = boto3.resource('dynamodb')
s3 = boto3.resource('s3')
vault_table = dynamodb.Table(os.environ.get('VAULT_TABLE_NAME'))
vault_table_partition_key = os.environ.get('VAULT_TABLE_KEY')
vault_tabl... | [
"os.environ.get",
"cryptography.fernet.Fernet",
"boto3.resource",
"json.dumps"
] | [((130, 156), 'boto3.resource', 'boto3.resource', (['"""dynamodb"""'], {}), "('dynamodb')\n", (144, 156), False, 'import boto3\n'), ((162, 182), 'boto3.resource', 'boto3.resource', (['"""s3"""'], {}), "('s3')\n", (176, 182), False, 'import boto3\n'), ((276, 309), 'os.environ.get', 'os.environ.get', (['"""VAULT_TABLE_KE... |
import errno
import glob
import os
import re
import shutil
from pypadre.core.model.code.code_mixin import CodeMixin, PythonPackage, PythonFile, GenericCall, \
GitIdentifier, RepositoryIdentifier, PipIdentifier, Function
from pypadre.pod.backend.i_padre_backend import IPadreBackend
from pypadre.pod.repository.i_rep... | [
"re.escape",
"pypadre.core.model.code.code_mixin.GitIdentifier",
"pypadre.pod.repository.local.file.generic.i_file_repository.File",
"shutil.copy",
"shutil.copytree",
"pypadre.core.model.code.code_mixin.Function",
"os.path.join",
"pypadre.core.model.code.code_mixin.PipIdentifier"
] | [((1002, 1039), 'pypadre.pod.repository.local.file.generic.i_file_repository.File', 'File', (['"""metadata.json"""', 'JSonSerializer'], {}), "('metadata.json', JSonSerializer)\n", (1006, 1039), False, 'from pypadre.pod.repository.local.file.generic.i_file_repository import File\n'), ((706, 732), 'shutil.copytree', 'shu... |
#!/usr/bin/python3
# -*- coding: utf-8 -*-
"""
Created on Mon Oct 8 20:54:26 2018
@author: andrea
"""
import sys
import json
from xml.sax import make_parser
from urllib.request import urlretrieve
from smallsmilhandler import SmallSMILHandler
class KaraokeLocal(SmallSMILHandler):
def __init__(self, fichero):
... | [
"json.dumps",
"urllib.request.urlretrieve",
"smallsmilhandler.SmallSMILHandler",
"sys.exit",
"xml.sax.make_parser"
] | [((377, 390), 'xml.sax.make_parser', 'make_parser', ([], {}), '()\n', (388, 390), False, 'from xml.sax import make_parser\n'), ((425, 443), 'smallsmilhandler.SmallSMILHandler', 'SmallSMILHandler', ([], {}), '()\n', (441, 443), False, 'from smallsmilhandler import SmallSMILHandler\n'), ((1112, 1134), 'json.dumps', 'json... |
import os
import sys
import tempfile
from datetime import datetime
from pprint import pprint
import ray
from ray import tune
from ray.rllib.agents import Trainer
from ray.tune.logger import UnifiedLogger
from ray.tune.result import DEFAULT_RESULTS_DIR
# os.environ['TF_CPP_MIN_LOG_LEVEL'] = '2'
from command_line_tools... | [
"ray.init",
"command_line_tools.run_tools.setup_run",
"ray.nodes",
"ray.tune.run",
"ray.is_initialized",
"ray.shutdown",
"ray.cluster_resources"
] | [((2877, 2902), 'command_line_tools.run_tools.setup_run', 'setup_run', (['default_config'], {}), '(default_config)\n', (2886, 2902), False, 'from command_line_tools.run_tools import setup_run\n'), ((3195, 3305), 'ray.tune.run', 'tune.run', (['train'], {'name': "config['name']", 'trial_name_creator': "(lambda trial: con... |
"""Protocol-related functions."""
# Copyright 2020-2021 Blue Brain Project / EPFL
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
# http://www.apache.org/licenses/LICENSE-2.0
# Unless r... | [
"bluepyopt.ephys.locations.NrnSomaDistanceCompLocation",
"bluepyopt.ephys.protocols.SweepProtocol",
"bluepyopt.ephys.protocols.SequenceProtocol",
"emodelrunner.features.define_efeatures",
"emodelrunner.recordings.RecordingCustom",
"emodelrunner.protocols.StepProtocol",
"emodelrunner.protocols.SweepProto... | [((1156, 1183), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1173, 1183), False, 'import logging\n'), ((1196, 1300), 'bluepyopt.ephys.locations.NrnSeclistCompLocation', 'ephys.locations.NrnSeclistCompLocation', ([], {'name': '"""soma"""', 'seclist_name': '"""somatic"""', 'sec_index': '... |
import os
from flask import Flask, redirect
from flask import request
from flask import jsonify
import hashlib
app = Flask(__name__)
c = 0
clients = []
chat = []
#[from, to, status[0sent, 1accepted, 2rejected]]
requests = {}
requests_sent = {}
version = 5
additive = 0
def getUID(ip):
return hashlib.sha256(str(... | [
"os.environ.get",
"flask.jsonify",
"flask.Flask",
"flask.redirect"
] | [((118, 133), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (123, 133), False, 'from flask import Flask, redirect\n'), ((2333, 2356), 'flask.redirect', 'redirect', (['"""/"""'], {'code': '(302)'}), "('/', code=302)\n", (2341, 2356), False, 'from flask import Flask, redirect\n'), ((2676, 2697), 'flask.json... |
# Copyright (c) 2020, Huawei Technologies.All rights reserved.
#
# Licensed under the BSD 3-Clause License (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://opensource.org/licenses/BSD-3-Clause
#
# Unless required by applicable law... | [
"numpy.random.uniform",
"torch.ne",
"copy.deepcopy",
"common_utils.run_tests",
"util_test.create_common_tensor"
] | [((8284, 8295), 'common_utils.run_tests', 'run_tests', ([], {}), '()\n', (8293, 8295), False, 'from common_utils import TestCase, run_tests\n'), ((894, 918), 'torch.ne', 'torch.ne', (['input1', 'input2'], {}), '(input1, input2)\n', (902, 918), False, 'import torch\n'), ((1051, 1075), 'torch.ne', 'torch.ne', (['input1',... |
from elasticsearch.helpers import scan
import utils.helpers as hp
valueField = {
'ps_packetloss': 'packet_loss',
'ps_owd': 'delay_mean',
'ps_retransmits': 'retransmits',
'ps_throughput': 'throughput'
}
def query4Avg(idx, dateFrom, dateTo):
val_fld = val... | [
"utils.helpers.es.search"
] | [((2956, 2991), 'utils.helpers.es.search', 'hp.es.search', ([], {'index': 'idx', 'body': 'query'}), '(index=idx, body=query)\n', (2968, 2991), True, 'import utils.helpers as hp\n')] |
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import pytest
from mock import patch
from night_scheduler.framework.sun.sun import Sun
class TestSun(object):
FAKE_LATITUDE = "00"
FAKE_LONGITUDE = "11"
F... | [
"mock.patch",
"night_scheduler.framework.sun.sun.Sun"
] | [((1033, 1054), 'mock.patch', 'patch', (['"""requests.get"""'], {}), "('requests.get')\n", (1038, 1054), False, 'from mock import patch\n'), ((1229, 1327), 'night_scheduler.framework.sun.sun.Sun', 'Sun', ([], {'latitude': 'TestSun.FAKE_LATITUDE', 'longitude': 'TestSun.FAKE_LONGITUDE', 'date': 'TestSun.FAKE_DATE'}), '(l... |
#!/usr/bin/env python3
import argparse
import sys
import psutil
from wrap_scriptlet import wrap_scriptlet
def run():
parser = argparse.ArgumentParser()
parser.add_argument('pid')
args = parser.parse_args(sys.argv[1:])
process = psutil.Process(int(args.pid))
return process.cmdline()
sys.exit(wra... | [
"wrap_scriptlet.wrap_scriptlet",
"argparse.ArgumentParser"
] | [((133, 158), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (156, 158), False, 'import argparse\n'), ((317, 336), 'wrap_scriptlet.wrap_scriptlet', 'wrap_scriptlet', (['run'], {}), '(run)\n', (331, 336), False, 'from wrap_scriptlet import wrap_scriptlet\n')] |
'''
By Zhenghang(<NAME>
############################################################################################
It's a light server based on FLASK micro framework,
1.Requirements: Python 3, Flask and relevant packages
2. How does this work?
(1) Firstly, modify the host IP address of your own environment.
(2)... | [
"csv.reader",
"csv.writer",
"os.path.getsize",
"flask.Flask",
"app.models.User.query.all"
] | [((1817, 1832), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (1822, 1832), False, 'from flask import Flask, request\n'), ((2366, 2389), 'app.models.User.query.all', 'models.User.query.all', ([], {}), '()\n', (2387, 2389), False, 'from app import db, models\n'), ((3430, 3488), 'csv.writer', 'csv.writer', ... |
from rx.core import Observable
from rx.internal import extensionmethod
@extensionmethod(Observable)
def do_while(self, condition):
"""Repeats source as long as condition holds emulating a do while loop.
Keyword arguments:
condition -- {Function} The condition which determines if the source
will b... | [
"rx.internal.extensionmethod",
"rx.core.Observable.while_do"
] | [((74, 101), 'rx.internal.extensionmethod', 'extensionmethod', (['Observable'], {}), '(Observable)\n', (89, 101), False, 'from rx.internal import extensionmethod\n'), ((480, 516), 'rx.core.Observable.while_do', 'Observable.while_do', (['condition', 'self'], {}), '(condition, self)\n', (499, 516), False, 'from rx.core i... |
import os
import networkx as nx
import scipy.sparse as sp
def search_cycle(dir_adj):
dir_adj = nx.from_scipy_sparse_matrix(A=dir_adj, create_using=nx.DiGraph)
cycles = list(nx.algorithms.cycles.simple_cycles(dir_adj))
num_cycle = len(cycles)
q = []
for i in range(num_cycle):
q.append(len(cy... | [
"networkx.from_scipy_sparse_matrix",
"networkx.algorithms.cycles.simple_cycles",
"os.path.join"
] | [((424, 464), 'os.path.join', 'os.path.join', (['dataset_path', 'dataset_name'], {}), '(dataset_path, dataset_name)\n', (436, 464), False, 'import os\n'), ((100, 163), 'networkx.from_scipy_sparse_matrix', 'nx.from_scipy_sparse_matrix', ([], {'A': 'dir_adj', 'create_using': 'nx.DiGraph'}), '(A=dir_adj, create_using=nx.D... |
from asitiger.status import (
AxisEnabledStatus,
AxisStatus,
JoystickStatus,
LimitStatus,
MotorStatus,
RampingDirection,
RampingStatus,
Status,
status_from_decimal,
statuses_for_rdstat,
)
RDSTAT_RESPONSE = ":A 10N 138"
def test_status_from_decimal_types():
axis = status_f... | [
"asitiger.status.Status.from_flag",
"asitiger.status.status_from_decimal",
"asitiger.status.statuses_for_rdstat"
] | [((312, 336), 'asitiger.status.status_from_decimal', 'status_from_decimal', (['(210)'], {}), '(210)\n', (331, 336), False, 'from asitiger.status import AxisEnabledStatus, AxisStatus, JoystickStatus, LimitStatus, MotorStatus, RampingDirection, RampingStatus, Status, status_from_decimal, statuses_for_rdstat\n'), ((809, 8... |
from django.contrib.auth.models import User
from rest_framework import serializers
from .models import Film, ExtraInfo, Recenzja, Aktor
class UserSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = User
fields = ['id', 'username', 'email','password']
extra_kwargs = {'pa... | [
"django.contrib.auth.models.User.objects.create_user"
] | [((423, 465), 'django.contrib.auth.models.User.objects.create_user', 'User.objects.create_user', ([], {}), '(**validated_data)\n', (447, 465), False, 'from django.contrib.auth.models import User\n')] |
"""
Module containing the definitions and methods to compute
a variety of indices used to study ENSO
"""
from typing import List, Optional, Tuple
import numpy as np
import xarray as xr
from eofs.xarray import Eof
from .core import compute_anomaly, compute_climatology, xconvolve
class ECindex:
"""
Computes ... | [
"numpy.deg2rad",
"numpy.zeros",
"xarray.merge",
"numpy.array",
"numpy.sqrt"
] | [((2338, 2349), 'numpy.zeros', 'np.zeros', (['(2)'], {}), '(2)\n', (2346, 2349), True, 'import numpy as np\n'), ((3047, 3073), 'xarray.merge', 'xr.merge', (['[eindex, cindex]'], {}), '([eindex, cindex])\n', (3055, 3073), True, 'import xarray as xr\n'), ((3931, 3954), 'numpy.array', 'np.array', (['smooth_kernel'], {}), ... |
# -*- coding: utf-8 -*-
"""
Copyright (c) 2020, University of Southampton
All rights reserved.
Licensed under the BSD 3-Clause License.
See LICENSE.md file in the project root for full license information.
"""
import pynmea2
from auv_nav.sensors import Category, Usbl
from auv_nav.tools.time_conversions import date_ti... | [
"oplab.get_raw_folder",
"pynmea2.parse",
"auv_nav.tools.time_conversions.date_time_to_epoch",
"auv_nav.tools.time_conversions.read_timezone",
"oplab.get_file_list",
"auv_nav.sensors.Usbl"
] | [((791, 814), 'auv_nav.tools.time_conversions.read_timezone', 'read_timezone', (['timezone'], {}), '(timezone)\n', (804, 814), False, 'from auv_nav.tools.time_conversions import date_time_to_epoch, read_timezone\n'), ((939, 1038), 'auv_nav.sensors.Usbl', 'Usbl', (['mission.usbl.std_factor', 'mission.usbl.std_offset', '... |
import unittest
from hupun.page.hupun_goods.goods_information import GoodsInformation
from hupun.page.hupun_goods.goods_information_sku import GoodsInformationsku
from hupun.page.in_sale_store_table.export_file_download_req import ExportFileDownloadReq
from hupun.page.in_sale_store_table.export_task_query import Expor... | [
"hupun.page.order_goods.OrderGoods",
"hupun.page.purchase_order.PurchaseOrder",
"hupun.page.sync_module.choose_purchase_bill_sku.ChoosePurBillSku",
"pyspider.helper.date.Date.now",
"hupun_slow_crawl.model.es.store_house.StoreHouse",
"hupun.page.sync_module.confirm_purchase_stock.ConfirmPurBillStock",
"h... | [((7478, 7493), 'unittest.main', 'unittest.main', ([], {}), '()\n', (7491, 7493), False, 'import unittest\n'), ((3246, 3256), 'pyspider.helper.date.Date.now', 'Date.now', ([], {}), '()\n', (3254, 3256), False, 'from pyspider.helper.date import Date\n'), ((1814, 1952), 'hupun.page.order_goods.OrderGoods', 'OrderGoods', ... |
# Generated by Django 3.0.8 on 2020-07-29 00:27
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('log_api', '0008_auto_20200728_2126'),
]
operations = [
migrations.AlterField(
model_name='execution',
name='archived... | [
"django.db.models.BooleanField"
] | [((341, 400), 'django.db.models.BooleanField', 'models.BooleanField', ([], {'default': '(False)', 'verbose_name': '"""Archived"""'}), "(default=False, verbose_name='Archived')\n", (360, 400), False, 'from django.db import migrations, models\n')] |
import asyncio, random
import nacre
class EightBallSession:
answers = [
"It is certain",
"It is decidedly so",
"Without a doubt",
"Yes definitely",
"You may rely on it",
"As I see it, yes",
"Most likely",
"Outlook good",
"Yes",
"Signs point to yes",
"Reply hazy try again",
"Ask again later",... | [
"nacre.handle.isMessageEvent",
"random.choice"
] | [((1081, 1108), 'random.choice', 'random.choice', (['self.answers'], {}), '(self.answers)\n', (1094, 1108), False, 'import asyncio, random\n'), ((849, 884), 'nacre.handle.isMessageEvent', 'nacre.handle.isMessageEvent', (['update'], {}), '(update)\n', (876, 884), False, 'import nacre\n')] |
# Copyright (c) 2013, GreyCube Technologies and contributors
# For license information, please see license.txt
from __future__ import unicode_literals
import frappe
from frappe import _
from frappe.utils import cint
import shutil, os
from frappe.modules import scrub, get_module_path
def copy_report(
module="NPro... | [
"frappe.modules.scrub",
"frappe.get_doc",
"frappe.db.commit",
"shutil.copyfile",
"frappe.modules.get_module_path"
] | [((533, 551), 'frappe.db.commit', 'frappe.db.commit', ([], {}), '()\n', (549, 551), False, 'import frappe\n'), ((622, 645), 'frappe.modules.get_module_path', 'get_module_path', (['module'], {}), '(module)\n', (637, 645), False, 'from frappe.modules import scrub, get_module_path\n'), ((1098, 1133), 'shutil.copyfile', 's... |
from mazikeen.MakedirsBlock import MakedirsBlock
from mazikeen.GeneratorException import GeneratorException
def generateMakedirs(data):
if not isinstance(data, str):
raise GeneratorException("'makedirs' block not recognized")
return MakedirsBlock(data) | [
"mazikeen.GeneratorException.GeneratorException",
"mazikeen.MakedirsBlock.MakedirsBlock"
] | [((250, 269), 'mazikeen.MakedirsBlock.MakedirsBlock', 'MakedirsBlock', (['data'], {}), '(data)\n', (263, 269), False, 'from mazikeen.MakedirsBlock import MakedirsBlock\n'), ((185, 238), 'mazikeen.GeneratorException.GeneratorException', 'GeneratorException', (['"""\'makedirs\' block not recognized"""'], {}), '("\'makedi... |
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import models, migrations
from django.conf import settings
class Migration(migrations.Migration):
dependencies = [
('idc', '0001_initial'),
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
]
operati... | [
"django.db.migrations.swappable_dependency",
"django.db.models.CharField",
"django.db.models.ForeignKey",
"django.db.models.BooleanField",
"django.db.models.AutoField",
"django.db.models.IntegerField",
"django.db.models.DateTimeField"
] | [((243, 300), 'django.db.migrations.swappable_dependency', 'migrations.swappable_dependency', (['settings.AUTH_USER_MODEL'], {}), '(settings.AUTH_USER_MODEL)\n', (274, 300), False, 'from django.db import models, migrations\n'), ((433, 484), 'django.db.models.AutoField', 'models.AutoField', ([], {'serialize': '(False)',... |
import unittest
from solver import buddy_strings
class TestSolver(unittest.TestCase):
def test_buddy_strings(self):
self.assertEqual(buddy_strings("ab" , "ba" ), True )
self.assertEqual(buddy_strings("ab" , "ab" ), False)
self.assertEqual(buddy_strings("aa" , "aa" ), T... | [
"unittest.main",
"solver.buddy_strings"
] | [((496, 511), 'unittest.main', 'unittest.main', ([], {}), '()\n', (509, 511), False, 'import unittest\n'), ((140, 165), 'solver.buddy_strings', 'buddy_strings', (['"""ab"""', '"""ba"""'], {}), "('ab', 'ba')\n", (153, 165), False, 'from solver import buddy_strings\n'), ((209, 234), 'solver.buddy_strings', 'buddy_strings... |
#!/usr/bin/env python3
import unittest
from src.executor.Printer import Printer
from src.data.VideoItem import VideoItem
class TestIExecutor(unittest.TestCase):
def test_compiles(self):
self.assertEqual(True, True)
def test_printer(self):
printer = Printer()
printer.run(VideoItem(met... | [
"unittest.main",
"src.executor.Printer.Printer",
"src.data.VideoItem.VideoItem"
] | [((381, 396), 'unittest.main', 'unittest.main', ([], {}), '()\n', (394, 396), False, 'import unittest\n'), ((277, 286), 'src.executor.Printer.Printer', 'Printer', ([], {}), '()\n', (284, 286), False, 'from src.executor.Printer import Printer\n'), ((307, 346), 'src.data.VideoItem.VideoItem', 'VideoItem', ([], {'metadata... |
"""Main execution body for program. Contains GUI interface and exporting class that creates files instead
of generating HTML Reports
Author: <NAME>
Last Updated: 28/02/2017
"""
import argparse
import webbrowser
import textwrap
import xlrd
from tkinter import *
from tkinter import filedialog, ttk
from threading import... | [
"textwrap.dedent",
"threading.Thread",
"webbrowser.open",
"webbrowser.open_new",
"xlrd.open_workbook",
"tkinter.ttk.Progressbar",
"tkinter.filedialog.askdirectory",
"tkinter.filedialog.askopenfiles",
"tkinter.filedialog.asksaveasfile",
"tkinter.filedialog.askopenfile"
] | [((3301, 3369), 'tkinter.ttk.Progressbar', 'ttk.Progressbar', (['mainwindow'], {'orient': '"""horizontal"""', 'mode': '"""determinate"""'}), "(mainwindow, orient='horizontal', mode='determinate')\n", (3316, 3369), False, 'from tkinter import filedialog, ttk\n'), ((4123, 4315), 'tkinter.filedialog.askopenfiles', 'filedi... |
from django.test import TestCase, override_settings
from social_django.compat import reverse
@override_settings(SOCIAL_AUTH_GITHUB_KEY = '1', SOCIAL_AUTH_GITHUB_SECRET='2')
class AuthTestcase(TestCase):
def setUp(self):
session = self.client.session
session["github_status"] = "1"
session.s... | [
"social_django.compat.reverse",
"django.test.override_settings"
] | [((95, 171), 'django.test.override_settings', 'override_settings', ([], {'SOCIAL_AUTH_GITHUB_KEY': '"""1"""', 'SOCIAL_AUTH_GITHUB_SECRET': '"""2"""'}), "(SOCIAL_AUTH_GITHUB_KEY='1', SOCIAL_AUTH_GITHUB_SECRET='2')\n", (112, 171), False, 'from django.test import TestCase, override_settings\n'), ((393, 446), 'social_djang... |
import datetime
import json
import logging
import os
import threading
import time
from abc import ABC, abstractmethod
import pika
from tools.mongo_dao import MongoDB
class StopCondition(ABC):
def __init__(self, stop_condition_parameters: dict, experiment_description: dict, experiment_id: str):
self.even... | [
"threading.Thread",
"pika.ConnectionParameters",
"json.dumps",
"time.sleep",
"datetime.timedelta",
"os.getenv",
"logging.getLogger"
] | [((329, 366), 'os.getenv', 'os.getenv', (['"""BRISE_EVENT_SERVICE_HOST"""'], {}), "('BRISE_EVENT_SERVICE_HOST')\n", (338, 366), False, 'import os\n'), ((393, 435), 'os.getenv', 'os.getenv', (['"""BRISE_EVENT_SERVICE_AMQP_PORT"""'], {}), "('BRISE_EVENT_SERVICE_AMQP_PORT')\n", (402, 435), False, 'import os\n'), ((931, 98... |