code stringlengths 20 1.05M | apis list | extract_api stringlengths 75 5.24M |
|---|---|---|
# Copyright (c) 2010-2014 openpyxl
import pytest
from openpyxl.styles.borders import Border, Side
from openpyxl.styles.fills import GradientFill
from openpyxl.styles.colors import Color
from openpyxl.writer.styles import StyleWriter
from openpyxl.tests.helper import get_xml, compare_xml
class DummyWorkbook:
st... | [
"openpyxl.styles.colors.Color",
"openpyxl.tests.helper.get_xml",
"openpyxl.tests.helper.compare_xml",
"openpyxl.styles.borders.Border"
] | [((552, 573), 'openpyxl.tests.helper.get_xml', 'get_xml', (['writer._root'], {}), '(writer._root)\n', (559, 573), False, 'from openpyxl.tests.helper import get_xml, compare_xml\n'), ((910, 936), 'openpyxl.tests.helper.compare_xml', 'compare_xml', (['xml', 'expected'], {}), '(xml, expected)\n', (921, 936), False, 'from ... |
from unittest import TestCase
from io import StringIO
import json
class TestDump(TestCase):
def test_dump(self):
sio = StringIO()
json.dump({}, sio)
self.assertEquals(sio.getvalue(), '{}')
def test_dumps(self):
self.assertEquals(json.dumps({}), '{}')
def test_encode_truef... | [
"io.StringIO",
"json.dumps",
"json.dump"
] | [((133, 143), 'io.StringIO', 'StringIO', ([], {}), '()\n', (141, 143), False, 'from io import StringIO\n'), ((152, 170), 'json.dump', 'json.dump', (['{}', 'sio'], {}), '({}, sio)\n', (161, 170), False, 'import json\n'), ((272, 286), 'json.dumps', 'json.dumps', (['{}'], {}), '({})\n', (282, 286), False, 'import json\n')... |
# -*- coding: utf-8 -*-
#
# Tencent is pleased to support the open source community by making QT4C available.
# Copyright (C) 2020 THL A29 Limited, a Tencent company. All rights reserved.
# QT4C is licensed under the BSD 3-Clause License, except for the third-party components listed below.
# A copy of the BSD 3-Cla... | [
"os.path.abspath",
"os.path.dirname",
"sys.argv.count",
"unittest.TestLoader"
] | [((439, 464), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (454, 464), False, 'import os\n'), ((485, 510), 'os.path.dirname', 'os.path.dirname', (['test_dir'], {}), '(test_dir)\n', (500, 510), False, 'import os\n'), ((611, 632), 'unittest.TestLoader', 'unittest.TestLoader', ([], {}), '()\n'... |
from cffi import FFI
ffibuilder = FFI()
ffibuilder.cdef("""
int test(int t);
""")
ffibuilder.set_source("_pi_cffi",
"""
#include "brute.h"
""",
sources=['brute.c'])
if __name__ == "__main__":
ffibuilder.compile(verbose = Tru... | [
"cffi.FFI"
] | [((35, 40), 'cffi.FFI', 'FFI', ([], {}), '()\n', (38, 40), False, 'from cffi import FFI\n')] |
"""Board Module"""
import copy
from typing import Tuple, List
from src.coordinate import Coordinate
from src.snake import Snake
class Board:
"""Track the cooardinates for all snakes and food in the game."""
def __init__(self, data):
self._data = data
self._snakes = None
self._foods = No... | [
"src.snake.Snake",
"src.coordinate.Coordinate",
"copy.deepcopy"
] | [((1708, 1727), 'copy.deepcopy', 'copy.deepcopy', (['self'], {}), '(self)\n', (1721, 1727), False, 'import copy\n'), ((493, 510), 'src.snake.Snake', 'Snake', (['snake_data'], {}), '(snake_data)\n', (498, 510), False, 'from src.snake import Snake\n'), ((789, 810), 'src.coordinate.Coordinate', 'Coordinate', (['food_data'... |
import os
import warnings
from django.conf import settings
CAPTCHA_FONT_PATH = getattr(settings, 'CAPTCHA_FONT_PATH', os.path.normpath(os.path.join(os.path.dirname(__file__), '..', 'fonts/Vera.ttf')))
CAPTCHA_FONT_SIZE = getattr(settings, 'CAPTCHA_FONT_SIZE', 22)
CAPTCHA_LETTER_ROTATION = getattr(settings, 'CAPTCHA_L... | [
"warnings.warn",
"os.path.dirname"
] | [((2103, 2141), 'warnings.warn', 'warnings.warn', (['msg', 'DeprecationWarning'], {}), '(msg, DeprecationWarning)\n', (2116, 2141), False, 'import warnings\n'), ((2368, 2406), 'warnings.warn', 'warnings.warn', (['msg', 'DeprecationWarning'], {}), '(msg, DeprecationWarning)\n', (2381, 2406), False, 'import warnings\n'),... |
# Generated by Django 2.2.21 on 2021-06-23 12:43
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('resources', '0125_add_timmi_payload_model'),
]
operations = [
migrations.AddField(
model_name=... | [
"django.db.models.BooleanField"
] | [((409, 521), 'django.db.models.BooleanField', 'models.BooleanField', ([], {'default': '(False)', 'verbose_name': '"""Disallow overlapping reservations in this unit per user."""'}), "(default=False, verbose_name=\n 'Disallow overlapping reservations in this unit per user.')\n", (428, 521), False, 'from django.db imp... |
import logging
from lora_multihop import serial_connection, variables
def config_module(configuration=variables.MODULE_CONFIG):
if serial_connection.execute_command(configuration, [variables.STATUS_OK]):
serial_connection.execute_command('AT+SEND=1', [variables.STATUS_OK])
serial_connection.execu... | [
"logging.debug",
"lora_multihop.serial_connection.execute_command",
"logging.warning",
"lora_multihop.serial_connection.response_q.get",
"lora_multihop.serial_connection.str_to_bytes",
"lora_multihop.serial_connection.bytes_to_str"
] | [((138, 209), 'lora_multihop.serial_connection.execute_command', 'serial_connection.execute_command', (['configuration', '[variables.STATUS_OK]'], {}), '(configuration, [variables.STATUS_OK])\n', (171, 209), False, 'from lora_multihop import serial_connection, variables\n'), ((445, 491), 'logging.warning', 'logging.war... |
import abc
from typing import Dict, Callable
import tensorflow as tf
from flink_ml_framework.context import Context
from flink_ml_framework.java_file import *
from ..runner import tf_helper, io_helper
from ..runner.output_writer import DirectOutputWriter
try:
from flink_ml_tensorflow.tensorflow_context import TF... | [
"flink_ml_tensorflow2.tensorflow_context.TFContext",
"tensorflow.data.TFRecordDataset",
"tensorflow.config.threading.set_intra_op_parallelism_threads",
"importlib.import_module"
] | [((1728, 1746), 'flink_ml_tensorflow2.tensorflow_context.TFContext', 'TFContext', (['context'], {}), '(context)\n', (1737, 1746), False, 'from flink_ml_tensorflow2.tensorflow_context import TFContext\n'), ((1311, 1347), 'importlib.import_module', 'importlib.import_module', (['module_name'], {}), '(module_name)\n', (133... |
import pytest
ENCODING = 'utf-8'
@pytest.fixture(scope='function', autouse=True)
def setup_case(request):
def destroy_case():
from corm import annihilate_keyspace_tables, SESSIONS
annihilate_keyspace_tables('mykeyspace')
for keyspace_name, session in SESSIONS.copy().items():
if... | [
"corm.register_table",
"corm.select",
"corm.keyspace_exists",
"datetime.datetime.utcnow",
"corm.annihilate_keyspace_tables",
"corm.sync_schema",
"corm.insert",
"corm.keyspace_destroy",
"uuid.uuid4",
"corm.cp",
"corm.obtain_session",
"pytest.fixture",
"corm.SESSIONS.copy",
"random.randint",... | [((36, 82), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""function"""', 'autouse': '(True)'}), "(scope='function', autouse=True)\n", (50, 82), False, 'import pytest\n'), ((735, 760), 'corm.register_table', 'register_table', (['TestModel'], {}), '(TestModel)\n', (749, 760), False, 'from corm import register_tab... |
import json
import os
from utilities.SaveLoadJson import SaveLoadJson as SLJ
from utilities.LineCount import LineCount as LC
import subprocess
from geolite2 import geolite2
class getData:
#Get Data Functions ------------------------------------------------------
@staticmethod
def getDATA():
resu... | [
"utilities.SaveLoadJson.SaveLoadJson.load",
"subprocess.Popen",
"json.dumps",
"geolite2.geolite2.reader"
] | [((789, 807), 'json.dumps', 'json.dumps', (['result'], {}), '(result)\n', (799, 807), False, 'import json\n'), ((865, 890), 'utilities.SaveLoadJson.SaveLoadJson.load', 'SLJ.load', (['"""dataStore.txt"""'], {}), "('dataStore.txt')\n", (873, 890), True, 'from utilities.SaveLoadJson import SaveLoadJson as SLJ\n'), ((1124,... |
# All credit to https://stackoverflow.com/questions/46571448/tkinter-and-a-html-file - thanks DELICA - https://stackoverflow.com/users/7027346/delica
from cefpython3 import cefpython as cef
import ctypes
try:
import tkinter as tk
from tkinter import messagebox
except ImportError:
import Tkinter as tk
impo... | [
"logging.getLogger",
"Tkinter.Grid.rowconfigure",
"logging.StreamHandler",
"tkinter.messagebox.askokcancel",
"Tkinter.Frame.__init__",
"logging.Formatter",
"Tkinter.Tk",
"Tkinter.Tcl",
"cefpython3.cefpython.MessageLoopWork",
"cefpython3.cefpython.Initialize",
"platform.system",
"platform.archi... | [((418, 435), 'cefpython3.cefpython.WindowUtils', 'cef.WindowUtils', ([], {}), '()\n', (433, 435), True, 'from cefpython3 import cefpython as cef\n'), ((589, 622), 'logging.getLogger', '_logging.getLogger', (['"""tkinter_.py"""'], {}), "('tkinter_.py')\n", (607, 622), True, 'import logging as _logging\n'), ((4152, 4176... |
#!/usr/bin/env python3
# encoding: utf-8
import sys
import urllib.parse
import selenium.webdriver
def exit():
driver.quit()
sys.exit(0)
driver = selenium.webdriver.Firefox()
# for some reason, detectportal.firefox.com and connectivitycheck.gstatic.com are not blocked
# therefore, they cannot be used to detect con... | [
"sys.exit"
] | [((129, 140), 'sys.exit', 'sys.exit', (['(0)'], {}), '(0)\n', (137, 140), False, 'import sys\n')] |
from django.shortcuts import render
from django.contrib.auth.decorators import login_required
from django.http import JsonResponse
from django.views.decorators.csrf import csrf_exempt
from . import helpers
# Create your views here.
@csrf_exempt
def convert_video(request, version):
# Get video
video = reques... | [
"django.http.JsonResponse"
] | [((532, 565), 'django.http.JsonResponse', 'JsonResponse', (['context'], {'safe': '(False)'}), '(context, safe=False)\n', (544, 565), False, 'from django.http import JsonResponse\n')] |
from itertools import count
import numpy as np
class Particle(object):
"""Object containing all the properties for a single particle"""
_ids = count(0)
def __init__(self, main_data=None, x=np.zeros(2)):
self.id = next(self._ids)
self.main_data = main_data
self.x = np.array(x)
... | [
"numpy.array",
"numpy.zeros",
"itertools.count"
] | [((154, 162), 'itertools.count', 'count', (['(0)'], {}), '(0)\n', (159, 162), False, 'from itertools import count\n'), ((205, 216), 'numpy.zeros', 'np.zeros', (['(2)'], {}), '(2)\n', (213, 216), True, 'import numpy as np\n'), ((305, 316), 'numpy.array', 'np.array', (['x'], {}), '(x)\n', (313, 316), True, 'import numpy ... |
from unittest import TestCase
from pyRdfa import pyRdfa
class NonXhtmlTest(TestCase):
"""
RDFa that is in not well-formed XHTML is passed through html5lib.
These tests make sure that this RDFa can be processed both from
a file, and from a URL.
"""
target1 = '<og:isbn>9780596516499</og:isbn>'... | [
"pyRdfa.pyRdfa"
] | [((459, 467), 'pyRdfa.pyRdfa', 'pyRdfa', ([], {}), '()\n', (465, 467), False, 'from pyRdfa import pyRdfa\n'), ((623, 631), 'pyRdfa.pyRdfa', 'pyRdfa', ([], {}), '()\n', (629, 631), False, 'from pyRdfa import pyRdfa\n')] |
from des109 import moeda
preco = float(input('Digite o preço pretendido: €'))
print(f'''A metade do preço é {(moeda.metade(preco))}
O dobro do preço é {(moeda.dobra(preco))}
Aumentando o preço 10% temos {(moeda.aumentar(preco, 10))}
Diminuindo o preço 13% temos {(moeda.aumentar(preco, 13))}''')
| [
"des109.moeda.metade",
"des109.moeda.dobra",
"des109.moeda.aumentar"
] | [((113, 132), 'des109.moeda.metade', 'moeda.metade', (['preco'], {}), '(preco)\n', (125, 132), False, 'from des109 import moeda\n'), ((163, 181), 'des109.moeda.dobra', 'moeda.dobra', (['preco'], {}), '(preco)\n', (174, 181), False, 'from des109 import moeda\n'), ((214, 239), 'des109.moeda.aumentar', 'moeda.aumentar', (... |
# coding=utf-8
# *** WARNING: this file was generated by the Pulumi SDK Generator. ***
# *** Do not edit by hand unless you're certain you know what you are doing! ***
import warnings
import pulumi
import pulumi.runtime
from typing import Any, Mapping, Optional, Sequence, Union, overload
from .. import _utilities
from... | [
"pulumi.get",
"pulumi.Alias",
"pulumi.getter",
"pulumi.set",
"pulumi.ResourceOptions",
"pulumi.ResourceOptions.merge"
] | [((6655, 6694), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourceGroupName"""'}), "(name='resourceGroupName')\n", (6668, 6694), False, 'import pulumi\n'), ((7070, 7122), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""authorizedAzureSubscriptionIds"""'}), "(name='authorizedAzureSubscriptionIds')\n", (708... |
import os
import glob
import cv2
import numpy as np
import torch
from torchvision.transforms import transforms
from natsort import natsorted
from models import resmasking_dropout1
from utils.datasets.fer2013dataset import EMOTION_DICT
from barez import show
transform = transforms.Compose(
[
transforms.ToPI... | [
"torchvision.transforms.transforms.ToPILImage",
"models.resmasking_dropout1",
"numpy.uint8",
"cv2.resize",
"torch.mean",
"torch.unsqueeze",
"torch.load",
"numpy.min",
"torch.flatten",
"numpy.max",
"torchvision.transforms.transforms.ToTensor",
"os.path.basename",
"numpy.concatenate",
"torch... | [((774, 799), 'models.resmasking_dropout1', 'resmasking_dropout1', (['(3)', '(7)'], {}), '(3, 7)\n', (793, 799), False, 'from models import resmasking_dropout1\n'), ((894, 971), 'torch.load', 'torch.load', (['"""./saved/checkpoints/Z_resmasking_dropout1_rot30_2019Nov30_13.32"""'], {}), "('./saved/checkpoints/Z_resmaski... |
# coding=utf-8
# *** WARNING: this file was generated by the Pulumi SDK Generator. ***
# *** Do not edit by hand unless you're certain you know what you are doing! ***
import warnings
import pulumi
import pulumi.runtime
from typing import Any, Mapping, Optional, Sequence, Union
from ... import _utilities, _tables
from... | [
"pulumi.get",
"pulumi.getter",
"pulumi.set",
"warnings.warn",
"pulumi.log.warn",
"pulumi.runtime.invoke",
"pulumi.InvokeOptions"
] | [((436, 620), 'warnings.warn', 'warnings.warn', (['"""The \'latest\' version is deprecated. Please migrate to the function in the top-level module: \'azure-native:notificationhubs:getNamespace\'."""', 'DeprecationWarning'], {}), '(\n "The \'latest\' version is deprecated. Please migrate to the function in the top-le... |
import json
from pygments import highlight
from pygments.lexers import JsonLexer
from pygments.formatters import TerminalFormatter
def print_json_obj(json_object):
json_str = json.dumps(json_object, indent=4, sort_keys=True)
print(highlight(json_str, JsonLexer(), TerminalFormatter()))
def print_json_str(jso... | [
"pygments.lexers.JsonLexer",
"json.dumps",
"pygments.formatters.TerminalFormatter"
] | [((181, 230), 'json.dumps', 'json.dumps', (['json_object'], {'indent': '(4)', 'sort_keys': '(True)'}), '(json_object, indent=4, sort_keys=True)\n', (191, 230), False, 'import json\n'), ((261, 272), 'pygments.lexers.JsonLexer', 'JsonLexer', ([], {}), '()\n', (270, 272), False, 'from pygments.lexers import JsonLexer\n'),... |
#!/usr/bin/env python3
import os
from opendbc.can.parser import CANParser
from cereal import car
from selfdrive.car.interfaces import RadarInterfaceBase
RADAR_MSGS_C = list(range(0x2c2, 0x2d4+2, 2)) # c_ messages 706,...,724
RADAR_MSGS_D = list(range(0x2a2, 0x2b4+2, 2)) # d_ messages
LAST_MSG = max(RADAR_MSGS_C + RA... | [
"cereal.car.RadarData.new_message",
"os.path.splitext",
"cereal.car.RadarData.RadarPoint.new_message"
] | [((2329, 2356), 'cereal.car.RadarData.new_message', 'car.RadarData.new_message', ([], {}), '()\n', (2354, 2356), False, 'from cereal import car\n'), ((1589, 1612), 'os.path.splitext', 'os.path.splitext', (['dbc_f'], {}), '(dbc_f)\n', (1605, 1612), False, 'import os\n'), ((2670, 2708), 'cereal.car.RadarData.RadarPoint.n... |
import os
import numpy as np
import tensorflow as tf
from image_quality.utils import utils
class TrainDataGenerator(tf.keras.utils.Sequence):
'''inherits from Keras Sequence base object, allows to use multiprocessing in .fit_generator'''
def __init__(self, samples, img_dir, batch_size, n_classes, basenet_preproc... | [
"image_quality.utils.utils.random_crop",
"image_quality.utils.utils.load_image",
"image_quality.utils.utils.random_horizontal_flip",
"image_quality.utils.utils.normalize_labels",
"numpy.random.shuffle"
] | [((1440, 1471), 'numpy.random.shuffle', 'np.random.shuffle', (['self.indexes'], {}), '(self.indexes)\n', (1457, 1471), True, 'import numpy as np\n'), ((1878, 1924), 'image_quality.utils.utils.load_image', 'utils.load_image', (['img_file', 'self.img_load_dims'], {}), '(img_file, self.img_load_dims)\n', (1894, 1924), Fal... |
# -*- coding: utf-8 -*-
"""
Licensed to the Apache Software Foundation (ASF) under one
or more contributor license agreements. See the NOTICE file
distributed with this work for additional information
regarding copyright ownership. The ASF licenses this file
to you under the Apache License, Versi... | [
"qiskit.quantum_info.operators.channel.Chi",
"qiskit.quantum_info.operators.channel.Kraus",
"qiskit.quantum_info.operators.channel.Choi",
"qiskit.quantum_info.operators.channel.PTM",
"qat.comm.quops.ttypes.QuantumChannel",
"numpy.array",
"qiskit.quantum_info.operators.channel.SuperOp",
"numpy.real",
... | [((1512, 1556), 'qat.comm.datamodel.ttypes.Matrix', 'Matrix', (['array.shape[0]', 'array.shape[1]', 'data'], {}), '(array.shape[0], array.shape[1], data)\n', (1518, 1556), False, 'from qat.comm.datamodel.ttypes import Matrix, ComplexNumber\n'), ((2444, 2558), 'qat.comm.quops.ttypes.QuantumChannel', 'QuantumChannel', ([... |
""" util.auth2: Authentication tools
This module is based off of util.auth, except with the action
paradigm removed.
"""
from flask import session
from app.models import Account
from app.util import course as course_util
# Session keys
SESSION_EMAIL = 'email'
def create_account(email: str, password: str, f... | [
"app.models.Account.objects.get_or_404",
"app.util.course.set_courses",
"app.models.Account"
] | [((795, 893), 'app.models.Account', 'Account', ([], {'email': 'email', 'first_name': 'first_name', 'last_name': 'last_name', 'fsuid': 'fsuid', 'is_admin': '(False)'}), '(email=email, first_name=first_name, last_name=last_name, fsuid=\n fsuid, is_admin=False)\n', (802, 893), False, 'from app.models import Account\n')... |
from PyQt5.QtWidgets import *
from matplotlib.backends.backend_qt5agg import FigureCanvas
from matplotlib.figure import Figure
from matplotlib.backends.backend_qt5agg import NavigationToolbar2QT as NavigationToolbar
class PstaticWidget(QWidget):
def __init__(self, parent=None):
QWidget.__init__(self... | [
"matplotlib.figure.Figure",
"matplotlib.backends.backend_qt5agg.NavigationToolbar2QT",
"matplotlib.backends.backend_qt5agg.FigureCanvas"
] | [((360, 368), 'matplotlib.figure.Figure', 'Figure', ([], {}), '()\n', (366, 368), False, 'from matplotlib.figure import Figure\n'), ((451, 481), 'matplotlib.backends.backend_qt5agg.FigureCanvas', 'FigureCanvas', (['self.fig_pstatic'], {}), '(self.fig_pstatic)\n', (463, 481), False, 'from matplotlib.backends.backend_qt5... |
import os
class Config:
CSRF_ENABLED = True
SECRET_KEY = 'your-very-very-secret-key'
SQLALCHEMY_DATABASE_URI = 'postgresql:///flask_template_dev'
SQLALCHEMY_TRACK_MODIFICATIONS = False
SQLALCHEMY_ECHO = True
class Development(Config):
ENV = 'development'
DEBUG = True
TESTING = False
... | [
"os.getenv"
] | [((419, 610), 'os.getenv', 'os.getenv', (['"""DATABASE_URL"""', '"""postgres://firhokdcdnfygz:93231d3f2ae1156cabfc40f7e4ba08587a77f68a5e2072fbcbbdb30150ba4bcb@ec2-107-22-253-158.compute-1.amazonaws.com:5432/df9c5vvl0s21da"""'], {}), "('DATABASE_URL',\n 'postgres://firhokdcdnfygz:93231d3f2ae1156cabfc40f7e4ba08587a77f... |
import itertools
import numpy as np
import pandas as pd
def find_intersections(formula_lists,group_labels,exclusive = True):
"""
Docstring for function pyKrev.find_intersections
====================
This function compares n lists of molecular formula and outputs a dictionary containing the intersections... | [
"pandas.DataFrame",
"itertools.combinations"
] | [((1483, 1515), 'pandas.DataFrame', 'pd.DataFrame', ([], {'data': 'formula_lists'}), '(data=formula_lists)\n', (1495, 1515), True, 'import pandas as pd\n'), ((1176, 1215), 'itertools.combinations', 'itertools.combinations', (['group_labels', 'i'], {}), '(group_labels, i)\n', (1198, 1215), False, 'import itertools\n')] |
import os
import glob
import shutil
from tinytag import TinyTag
""" root = 'C:/'
copy_to = '/copy to/folder'
tag = TinyTag.get('C:/Users/jchap/OneDrive/Pictures/(VERYRAREBOYZ) (feat. $ki Mask The Slump God and Drugz).mp3')
print(tag.artist)
print('song duration: '+str(tag.duration))
"""
f = []
f=glob.gl... | [
"os.path.join",
"os.path.dirname",
"tinytag.TinyTag.get",
"glob.glob",
"os.walk"
] | [((313, 355), 'glob.glob', 'glob.glob', (['"""C:/Users/jchap/OneDrive/*.mp3"""'], {}), "('C:/Users/jchap/OneDrive/*.mp3')\n", (322, 355), False, 'import glob\n'), ((544, 558), 'os.walk', 'os.walk', (['"""C:/"""'], {}), "('C:/')\n", (551, 558), False, 'import os\n'), ((807, 837), 'tinytag.TinyTag.get', 'TinyTag.get', ([... |
#import modules
import os
import csv
#input
csvpath = os.path.join('Resources', 'budget_data.csv')
#output
outfile = os.path.join('Analysis', 'pybankstatements.txt')
#declare variables
months = []; total_m = 1; net_total = 0; total_change = 0; monthly_changes = []; greatest_inc = ['', 0]; greatest_dec = ['', 0]
#open... | [
"os.path.join",
"csv.reader"
] | [((54, 98), 'os.path.join', 'os.path.join', (['"""Resources"""', '"""budget_data.csv"""'], {}), "('Resources', 'budget_data.csv')\n", (66, 98), False, 'import os\n'), ((117, 165), 'os.path.join', 'os.path.join', (['"""Analysis"""', '"""pybankstatements.txt"""'], {}), "('Analysis', 'pybankstatements.txt')\n", (129, 165)... |
from rest_framework import serializers
from cms.api.serializers import UniCMSContentTypeClass, UniCMSCreateUpdateSerializer
from cms.medias.serializers import MediaSerializer
from . models import Carousel, CarouselItem, CarouselItemLink, CarouselItemLinkLocalization, CarouselItemLocalization
class CarouselForeignK... | [
"cms.medias.serializers.MediaSerializer"
] | [((2384, 2415), 'cms.medias.serializers.MediaSerializer', 'MediaSerializer', (['instance.image'], {}), '(instance.image)\n', (2399, 2415), False, 'from cms.medias.serializers import MediaSerializer\n')] |
#!/usr/bin/env python3
"""
Copyright (c) 2018-2021 Intel Corporation
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applic... | [
"cv2.vconcat",
"model_api.performance_metrics.PerformanceMetrics",
"images_capture.open_images_capture",
"cv2.imshow",
"logging.info",
"argparse.ArgumentParser",
"pathlib.Path",
"time.perf_counter",
"cv2.VideoWriter",
"numpy.concatenate",
"cv2.VideoWriter_fourcc",
"cv2.waitKey",
"numpy.squee... | [((1125, 1220), 'logging.basicConfig', 'log.basicConfig', ([], {'format': '"""[ %(levelname)s ] %(message)s"""', 'level': 'log.DEBUG', 'stream': 'sys.stdout'}), "(format='[ %(levelname)s ] %(message)s', level=log.DEBUG,\n stream=sys.stdout)\n", (1140, 1220), True, 'import logging as log\n'), ((1249, 1279), 'argparse... |
# coding: utf-8
"""
[AHOI cookbook](/ahoi/docs/cookbook/index.html) [Data Privacy](/sandboxmanager/#/privacy) [Terms of Service](/sandboxmanager/#/terms) [Imprint](https://sparkassen-hub.com/impressum/) © 2016‐2017 Starfinanz - Ein Unternehmen der Finanz Informatik # noqa: E501
OpenAPI sp... | [
"six.iteritems"
] | [((6926, 6959), 'six.iteritems', 'six.iteritems', (['self.swagger_types'], {}), '(self.swagger_types)\n', (6939, 6959), False, 'import six\n')] |
# Copyright 2015-2017 ARM Limited, Google and contributors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applica... | [
"os.listdir",
"trappy.register_dynamic_ftrace",
"os.path.join",
"shutil.copytree",
"os.path.dirname",
"json.load",
"os.path.basename",
"os.path.abspath",
"trappy.FTrace",
"trappy.unregister_dynamic_ftrace",
"json.dump",
"trappy.SysTrace"
] | [((2584, 2599), 'trappy.FTrace', 'trappy.FTrace', ([], {}), '()\n', (2597, 2599), False, 'import trappy\n'), ((2754, 2794), 'trappy.FTrace', 'trappy.FTrace', (['uncached_trace.trace_path'], {}), '(uncached_trace.trace_path)\n', (2767, 2794), False, 'import trappy\n'), ((4683, 4749), 'os.path.join', 'os.path.join', (['u... |
from django.contrib.auth.decorators import login_required
from django.contrib.auth.models import User
from django.shortcuts import render
from django.urls import reverse
from django.http import HttpResponseRedirect, HttpResponse
from django.utils import timezone
from olaf.models import *
from olaf.forms import *
from... | [
"django.shortcuts.render",
"django.utils.timezone.now",
"django.contrib.auth.models.User.objects.filter",
"django.urls.reverse",
"olaf.chess.controller.proccess_move",
"olaf.utility.usertools.get_translated_game_board",
"olaf.utility.usertools.new_game",
"olaf.utility.usertools.logout_user"
] | [((3350, 3391), 'django.shortcuts.render', 'render', (['request', 'fail_template', 'fail_args'], {}), '(request, fail_template, fail_args)\n', (3356, 3391), False, 'from django.shortcuts import render\n'), ((5689, 5719), 'olaf.utility.usertools.logout_user', 'usertools.logout_user', (['request'], {}), '(request)\n', (5... |
#!/usr/bin/python
# -*- coding: utf-8 -*-
# [Import start]
from flask import Blueprint, jsonify
# [Import end]
app = Blueprint(
'hoge',
__name__,
url_prefix='/hoge'
)
@app.route('/test')
def hoge():
return "\nhogehoge"
| [
"flask.Blueprint"
] | [((119, 166), 'flask.Blueprint', 'Blueprint', (['"""hoge"""', '__name__'], {'url_prefix': '"""/hoge"""'}), "('hoge', __name__, url_prefix='/hoge')\n", (128, 166), False, 'from flask import Blueprint, jsonify\n')] |
#!/usr/bin/env python3
import apt_pkg
import sys
from apt_pkg import CURSTATE_INSTALLED, version_compare
from operator import lt, le, eq, ge, gt
# Function mappings for relationship operators.
relation_operators = {"<<": lt, "<=": le, "=": eq, ">=": ge, ">>": gt}
# Set up APT cache.
apt_pkg.init()
cache = apt_pkg.Ca... | [
"apt_pkg.version_compare",
"apt_pkg.Cache",
"apt_pkg.init"
] | [((287, 301), 'apt_pkg.init', 'apt_pkg.init', ([], {}), '()\n', (299, 301), False, 'import apt_pkg\n'), ((310, 329), 'apt_pkg.Cache', 'apt_pkg.Cache', (['None'], {}), '(None)\n', (323, 329), False, 'import apt_pkg\n'), ((2105, 2147), 'apt_pkg.version_compare', 'version_compare', (['installed_version', 'pkgver'], {}), '... |
# Generated by Django 3.0.7 on 2020-08-24 06:17
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('datasets', '0008_auto_20200821_1427'),
]
operations = [
migrations.AddField(
model_name='rawdar',
name='AsB',
... | [
"django.db.models.FloatField",
"django.db.models.CharField"
] | [((332, 372), 'django.db.models.FloatField', 'models.FloatField', ([], {'blank': '(True)', 'null': '(True)'}), '(blank=True, null=True)\n', (349, 372), False, 'from django.db import migrations, models\n'), ((493, 632), 'django.db.models.CharField', 'models.CharField', ([], {'choices': "[('1', 'below detection level'), ... |
import os
import math
import time
import geohash
import geojson
from geojson import MultiLineString
from shapely import geometry
import shapefile
import numpy
import datetime as dt
import pandas as pd
import logging
logger = logging.getLogger(__name__)
source_shape_file_path = "C:/temp/2018/"
threshold = 60*60
cols = ... | [
"logging.getLogger",
"shapefile.Reader",
"datetime.datetime.strptime",
"os.path.join",
"pandas.DataFrame",
"os.walk"
] | [((226, 253), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (243, 253), False, 'import logging\n'), ((454, 485), 'os.walk', 'os.walk', (['source_shape_file_path'], {}), '(source_shape_file_path)\n', (461, 485), False, 'import os\n'), ((2012, 2045), 'pandas.DataFrame', 'pd.DataFrame', (['... |
import pygame_sdl2
pygame_sdl2.import_as_pygame()
import pygame
import os
import random
import math
from Ball import Ball
def save_state(balls):
"""
Saves the game state.
"""
stateString = ""
with open("state.txt", "w") as f:
for ball in balls:
stateString += "{} {} {} {} {}".f... | [
"os.path.exists",
"pygame.init",
"pygame.event.get",
"pygame.display.set_mode",
"pygame.display.flip",
"pygame.display.Info",
"pygame_sdl2.import_as_pygame",
"os.unlink",
"pygame.time.Clock",
"pygame.font.Font",
"random.randint",
"Ball.Ball"
] | [((19, 49), 'pygame_sdl2.import_as_pygame', 'pygame_sdl2.import_as_pygame', ([], {}), '()\n', (47, 49), False, 'import pygame_sdl2\n'), ((999, 1026), 'os.path.exists', 'os.path.exists', (['"""state.txt"""'], {}), "('state.txt')\n", (1013, 1026), False, 'import os\n'), ((1076, 1089), 'pygame.init', 'pygame.init', ([], {... |
from __future__ import division
import math, copy
import argparse
from brownian import Brownian
import scipy
import LLRcalc
class sprt:
def __init__(self, alpha=0.05, beta=0.05, elo0=0, elo1=5, elo_model="logistic"):
assert elo_model in ("logistic", "normalized")
self.elo_model = elo_model
... | [
"LLRcalc.results_to_pdf",
"argparse.ArgumentParser",
"math.sqrt",
"math.log",
"LLRcalc.stats",
"LLRcalc.L_",
"brownian.Brownian"
] | [((4157, 4182), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (4180, 4182), False, 'import argparse\n'), ((330, 358), 'math.log', 'math.log', (['(beta / (1 - alpha))'], {}), '(beta / (1 - alpha))\n', (338, 358), False, 'import math, copy\n'), ((376, 404), 'math.log', 'math.log', (['((1 - beta)... |
"""Simple Hail query example."""
import click
import hail as hl
from bokeh.io.export import get_screenshot_as_png
from analysis_runner import output_path
GNOMAD_HGDP_1KG_MT = (
'gs://gcp-public-data--gnomad/release/3.1/mt/genomes/'
'gnomad.genomes.v3.1.hgdp_1kg_subset_dense.mt'
)
@click.command()
@click.op... | [
"hail.hadoop_open",
"click.option",
"hail.sample_qc",
"hail.hadoop_exists",
"hail.read_matrix_table",
"hail.init",
"bokeh.io.export.get_screenshot_as_png",
"click.command",
"analysis_runner.output_path",
"hail.plot.histogram"
] | [((295, 310), 'click.command', 'click.command', ([], {}), '()\n', (308, 310), False, 'import click\n'), ((312, 397), 'click.option', 'click.option', (['"""--rerun"""'], {'help': '"""Whether to overwrite cached files"""', 'default': '(False)'}), "('--rerun', help='Whether to overwrite cached files', default=False\n )... |
# Copyright 2020 The TensorFlow Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applica... | [
"tensorflow.python.framework.convert_to_constants.convert_variables_to_constants_from_session_graph",
"collections.namedtuple",
"tensorflow.python.ops.random_ops.random_uniform",
"tensorflow.python.saved_model.load.load",
"tensorflow.python.client.session.Session",
"tensorflow.core.protobuf.config_pb2.Con... | [((3602, 3623), 'functools.lru_cache', 'functools.lru_cache', ([], {}), '()\n', (3621, 3623), False, 'import functools\n'), ((4355, 4376), 'functools.lru_cache', 'functools.lru_cache', ([], {}), '()\n', (4374, 4376), False, 'import functools\n'), ((4806, 4892), 'collections.namedtuple', 'collections.namedtuple', (['"""... |
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT License.
from typing import List, Pattern
from recognizers_text.utilities import RegExpUtility
from recognizers_number.number import BaseNumberParser
from recognizers_number.number.italian.extractors import ItalianIntegerExtractor,... | [
"recognizers_number.ItalianCardinalExtractor",
"recognizers_number.number.italian.extractors.ItalianIntegerExtractor",
"recognizers_text.utilities.RegExpUtility.get_safe_reg_exp",
"recognizers_number.ItalianOrdinalExtractor",
"recognizers_number.number.italian.parsers.ItalianNumberParserConfiguration"
] | [((5851, 5915), 'recognizers_text.utilities.RegExpUtility.get_safe_reg_exp', 'RegExpUtility.get_safe_reg_exp', (['ItalianDateTime.AllHalfYearRegex'], {}), '(ItalianDateTime.AllHalfYearRegex)\n', (5881, 5915), False, 'from recognizers_text.utilities import RegExpUtility\n'), ((5956, 6023), 'recognizers_text.utilities.Re... |
def test_setupcall():
"""
Test the call of the setup function
"""
import jupyter_libertem_proxy as jx
print("\nRunning test_setupcall...")
print(jx.setup_libertem())
| [
"jupyter_libertem_proxy.setup_libertem"
] | [((170, 189), 'jupyter_libertem_proxy.setup_libertem', 'jx.setup_libertem', ([], {}), '()\n', (187, 189), True, 'import jupyter_libertem_proxy as jx\n')] |
import math
import imageio
import cv2 as cv
import numpy as np
import transformer
def fix_rotation(img):
img_copy = img.copy()
img = cv.cvtColor(img, cv.COLOR_BGR2GRAY)
rows, cols = img.shape
img = cv.adaptiveThreshold(img, 255, cv.ADAPTIVE_THRESH_MEAN_C, cv.THRESH_BINARY_INV, 15, 9)
kernel = cv.g... | [
"cv2.imshow",
"cv2.warpPerspective",
"cv2.destroyAllWindows",
"cv2.approxPolyDP",
"math.hypot",
"imageio.get_writer",
"imageio.get_reader",
"cv2.arcLength",
"cv2.medianBlur",
"cv2.contourArea",
"cv2.waitKey",
"cv2.getPerspectiveTransform",
"cv2.minEnclosingCircle",
"cv2.morphologyEx",
"c... | [((142, 177), 'cv2.cvtColor', 'cv.cvtColor', (['img', 'cv.COLOR_BGR2GRAY'], {}), '(img, cv.COLOR_BGR2GRAY)\n', (153, 177), True, 'import cv2 as cv\n'), ((215, 306), 'cv2.adaptiveThreshold', 'cv.adaptiveThreshold', (['img', '(255)', 'cv.ADAPTIVE_THRESH_MEAN_C', 'cv.THRESH_BINARY_INV', '(15)', '(9)'], {}), '(img, 255, cv... |
# Altere o Programa 8.20 de forma que o usuário tenha três chances de acertar o número
# O programa termina se o usuário acertar ou errar três vezes
# Programa 8.20 do livro, página 184
# Programa 8.20 - Adivinhando o número
#
# import random
#
# n = random.randint(1, 10)
# x = int(input('Escolha um número entre 1 e 1... | [
"random.randint"
] | [((435, 456), 'random.randint', 'random.randint', (['(1)', '(10)'], {}), '(1, 10)\n', (449, 456), False, 'import random\n')] |
import setuptools
# To use a consistent encoding
from codecs import open
from os import path
here = path.abspath(path.dirname(__file__))
with open(path.join(here, 'README.md'), encoding='utf-8') as f:
long_description = f.read()
setuptools.setup(
name="atm76",
version="0.1.0",
author="<NAME>",
a... | [
"os.path.dirname",
"setuptools.find_packages",
"os.path.join"
] | [((115, 137), 'os.path.dirname', 'path.dirname', (['__file__'], {}), '(__file__)\n', (127, 137), False, 'from os import path\n'), ((150, 178), 'os.path.join', 'path.join', (['here', '"""README.md"""'], {}), "(here, 'README.md')\n", (159, 178), False, 'from os import path\n'), ((542, 568), 'setuptools.find_packages', 's... |
import logging
import asyncio
from agent.check_plugins import AbstractCheckPlugin
# Do khong biet dung thu vien asyncio ntn ca nen em dung thu vien request
# python
import requests
import sys
import time
from datetime import datetime
logger = logging.getLogger(__name__)
class Download(AbstractCheckPlugin):
@as... | [
"logging.getLogger",
"datetime.datetime.now",
"time.clock"
] | [((245, 272), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (262, 272), False, 'import logging\n'), ((448, 460), 'time.clock', 'time.clock', ([], {}), '()\n', (458, 460), False, 'import time\n'), ((728, 740), 'time.clock', 'time.clock', ([], {}), '()\n', (738, 740), False, 'import time\n... |
import weakref
import os
import requests
import ssl
from ssl import SSLContext
import logging
from ssl_context_builder.builder.builder import SslContextBuilder
from ssl_context_builder.http_impl.requests_wrapper.ssl_adapter import SslAdapter
class RequestsSecureSession:
def __init__(self, ssl_context: SSLContex... | [
"os.path.exists",
"requests.Session",
"logging.warning",
"ssl.DER_cert_to_PEM_cert",
"ssl_context_builder.http_impl.requests_wrapper.ssl_adapter.SslAdapter",
"os.remove"
] | [((838, 856), 'requests.Session', 'requests.Session', ([], {}), '()\n', (854, 856), False, 'import requests\n'), ((2087, 2116), 'logging.warning', 'logging.warning', (['warn_message'], {}), '(warn_message)\n', (2102, 2116), False, 'import logging\n'), ((2660, 2680), 'os.path.exists', 'os.path.exists', (['path'], {}), '... |
# Ghetto Fixtures
from codebox import app
from codebox.apps.auth.models import User
from codebox.apps.snippets.models import Snippet
from codebox.apps.organizations.models import Organization, OrganizationMember
from flask import g
client = app.test_client()
_ctx = app.test_request_context()
_ctx.push()
app.preproces... | [
"codebox.app.test_client",
"codebox.apps.snippets.models.Snippet.objects.create",
"codebox.apps.organizations.models.Organization.objects.create",
"codebox.app.preprocess_request",
"codebox.apps.organizations.models.OrganizationMember.objects.create",
"flask.g.redis.flushdb",
"codebox.app.test_request_c... | [((243, 260), 'codebox.app.test_client', 'app.test_client', ([], {}), '()\n', (258, 260), False, 'from codebox import app\n'), ((268, 294), 'codebox.app.test_request_context', 'app.test_request_context', ([], {}), '()\n', (292, 294), False, 'from codebox import app\n'), ((307, 331), 'codebox.app.preprocess_request', 'a... |
import function_exercise_01 as st
st.sandwich_toppings('meatballs', 'salad')
| [
"function_exercise_01.sandwich_toppings"
] | [((35, 77), 'function_exercise_01.sandwich_toppings', 'st.sandwich_toppings', (['"""meatballs"""', '"""salad"""'], {}), "('meatballs', 'salad')\n", (55, 77), True, 'import function_exercise_01 as st\n')] |
import json
import time
from functools import lru_cache
from multiprocessing import Pool, Process
from threading import Thread, Timer
from typing import Any, Dict, List
from datetime import datetime
import hashlib
import inspect
import requests
import waitress
from bottle import BaseTemplate, Bottle, request, response,... | [
"requests.post",
"utils.utils.decompress",
"bottle.Bottle",
"multiprocessing.Process",
"utils.logger.logger.info",
"time.sleep",
"core.Block.from_json",
"utils.storage.read_header_list_from_db",
"core.BlockChain",
"bottle.error",
"bottle.template",
"utils.utils.compress",
"json.dumps",
"co... | [((723, 731), 'bottle.Bottle', 'Bottle', ([], {}), '()\n', (729, 731), False, 'from bottle import BaseTemplate, Bottle, request, response, static_file, template, error\n'), ((817, 829), 'core.BlockChain', 'BlockChain', ([], {}), '()\n', (827, 829), False, 'from core import Block, BlockChain, SingleOutput, Transaction, ... |
import io
import os
from setuptools import setup
def read(file_name):
"""Read a text file and return the content as a string."""
with io.open(os.path.join(os.path.dirname(__file__), file_name),
encoding='utf-8') as f:
return f.read()
setup(
name='recmetrics',
url='https://gi... | [
"os.path.dirname"
] | [((166, 191), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (181, 191), False, 'import os\n')] |
import pygame
class Texto:
def __init__(self, screen, text, x, y, text_size = 20, fuente = 'Calibri', italic = False, bold= False, subrayado= False, color = (250, 240, 230), bg = [] ):
self.screen = screen
fg = color
self.coord = x, y
#load font, prepare values
f... | [
"pygame.Color",
"pygame.font.Font",
"pygame.font.SysFont"
] | [((326, 352), 'pygame.font.Font', 'pygame.font.Font', (['None', '(80)'], {}), '(None, 80)\n', (342, 352), False, 'import pygame\n'), ((421, 459), 'pygame.font.SysFont', 'pygame.font.SysFont', (['fuente', 'text_size'], {}), '(fuente, text_size)\n', (440, 459), False, 'import pygame\n'), ((2049, 2070), 'pygame.Color', 'p... |
# This is the code to train the xgboost model with cross-validation for each unique room in the dataset.
# Models are dumped into ./models and results are dumped into two csv files in the current work directory.
import argparse
import json
import math
import os
import pickle
import warnings
from typing import Tuple
i... | [
"pandas.read_csv",
"numpy.array",
"xgboost.DMatrix",
"sklearn.metrics.r2_score",
"numpy.random.RandomState",
"argparse.ArgumentParser",
"xgboost.train",
"json.dumps",
"pandas.set_option",
"xgboost.cv",
"pandas.DataFrame",
"sklearn.model_selection.train_test_split",
"hyperopt.hp.quniform",
... | [((792, 817), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (815, 817), False, 'import argparse\n'), ((1688, 1721), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {}), "('ignore')\n", (1711, 1721), False, 'import warnings\n'), ((1722, 1764), 'pandas.set_option', 'pd.s... |
import os
from setuptools import setup
# Read the version
g = {}
with open(os.path.join("editorconfig", "version.py"), "rt") as fp:
exec(fp.read(), g)
v = g['VERSION']
version = ".".join(str(x) for x in v[:3])
if v[3] != "final":
version += "-" + v[3]
setup(
name='EditorConfig',
versio... | [
"os.path.join"
] | [((76, 118), 'os.path.join', 'os.path.join', (['"""editorconfig"""', '"""version.py"""'], {}), "('editorconfig', 'version.py')\n", (88, 118), False, 'import os\n')] |
# -*- coding: utf-8 -*-
"""Tests for sktime annotators."""
import pandas as pd
import pytest
from sktime.registry import all_estimators
from sktime.utils._testing.estimator_checks import _make_args
ALL_ANNOTATORS = all_estimators(estimator_types="series-annotator", return_names=False)
@pytest.mark.parametrize("Est... | [
"pytest.mark.parametrize",
"sktime.utils._testing.estimator_checks._make_args",
"sktime.registry.all_estimators"
] | [((218, 288), 'sktime.registry.all_estimators', 'all_estimators', ([], {'estimator_types': '"""series-annotator"""', 'return_names': '(False)'}), "(estimator_types='series-annotator', return_names=False)\n", (232, 288), False, 'from sktime.registry import all_estimators\n'), ((292, 344), 'pytest.mark.parametrize', 'pyt... |
#!/usr/bin/env python
from __future__ import print_function
from kaldi.segmentation import NnetSAD, SegmentationProcessor
from kaldi.nnet3 import NnetSimpleComputationOptions
from kaldi.util.table import SequentialMatrixReader
# Construct SAD
model = NnetSAD.read_model("final.raw")
post = NnetSAD.read_average_poster... | [
"kaldi.nnet3.NnetSimpleComputationOptions",
"kaldi.segmentation.NnetSAD.make_sad_transform",
"kaldi.util.table.SequentialMatrixReader",
"kaldi.segmentation.NnetSAD.read_average_posteriors",
"kaldi.segmentation.NnetSAD",
"kaldi.segmentation.SegmentationProcessor",
"kaldi.segmentation.NnetSAD.read_model",... | [((254, 285), 'kaldi.segmentation.NnetSAD.read_model', 'NnetSAD.read_model', (['"""final.raw"""'], {}), "('final.raw')\n", (272, 285), False, 'from kaldi.segmentation import NnetSAD, SegmentationProcessor\n'), ((293, 343), 'kaldi.segmentation.NnetSAD.read_average_posteriors', 'NnetSAD.read_average_posteriors', (['"""po... |
"""
Comparison between the efficiency of the Boyer-Moore algorithm and the naive substring search algorithm.
The runtimes for both algorithms are plotted on the same axes.
"""
import matplotlib.pyplot as plt
import numpy as np
import string
import time
import random
from bm_alg import boyer_moore_match, naive_match
#... | [
"bm_alg.naive_match",
"random.choice",
"matplotlib.pyplot.ylabel",
"matplotlib.pyplot.xlabel",
"matplotlib.pyplot.plot",
"matplotlib.pyplot.title",
"bm_alg.boyer_moore_match",
"time.time",
"matplotlib.pyplot.legend",
"matplotlib.pyplot.show"
] | [((2371, 2416), 'matplotlib.pyplot.plot', 'plt.plot', (['x', 'y_naive'], {'label': '"""Naive Algorithm"""'}), "(x, y_naive, label='Naive Algorithm')\n", (2379, 2416), True, 'import matplotlib.pyplot as plt\n'), ((2421, 2469), 'matplotlib.pyplot.plot', 'plt.plot', (['x', 'y_bm'], {'label': '"""Boyer-Moore Algorithm"""'}... |
import json
from typing import Type, TYPE_CHECKING
from django.core.exceptions import ObjectDoesNotExist
from django.utils.decorators import method_decorator
from django.views.decorators.cache import cache_page
from rest_framework import viewsets, filters
from rest_framework.exceptions import NotFound
from rest_framew... | [
"json.dumps",
"rest_framework.response.Response",
"django.views.decorators.cache.cache_page",
"indicators.views.GeoJSONRenderer",
"maps.models.DataLayer.objects.get",
"maps.models.DataLayer.objects.all"
] | [((1032, 1055), 'maps.models.DataLayer.objects.all', 'DataLayer.objects.all', ([], {}), '()\n', (1053, 1055), False, 'from maps.models import DataLayer\n'), ((1501, 1517), 'json.dumps', 'json.dumps', (['data'], {}), '(data)\n', (1511, 1517), False, 'import json\n'), ((3018, 3035), 'rest_framework.response.Response', 'R... |
# -*- coding: utf-8 -*-
"""
Provide download function by request
"""
from datetime import datetime
import logging
import time
import urllib.parse
import requests
from bs4 import BeautifulSoup
class Throttle(object):
"""Throttle downloading by sleeping between requests to same domain."""
de... | [
"logging.warn",
"requests.Session",
"time.sleep",
"requests.Request",
"datetime.datetime.now",
"logging.info",
"logging.error"
] | [((891, 905), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (903, 905), False, 'from datetime import datetime\n'), ((1334, 1352), 'requests.Session', 'requests.Session', ([], {}), '()\n', (1350, 1352), False, 'import requests\n'), ((2118, 2161), 'requests.Request', 'requests.Request', (['"""GET"""', 'url']... |
# -*- coding: utf-8 -*-
#
# The MIT License (MIT)
#
# Copyright (C) 2017 <NAME> <<EMAIL>>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy of
# this software and associated documentation files (the "Software"), to deal in
# the Software without restriction, including without limitation t... | [
"urlparse.urljoin",
"requests.adapters.HTTPAdapter",
"requests.Session"
] | [((1981, 1999), 'requests.Session', 'requests.Session', ([], {}), '()\n', (1997, 1999), False, 'import requests\n'), ((2311, 2337), 'requests.adapters.HTTPAdapter', 'HTTPAdapter', ([], {'max_retries': '(1)'}), '(max_retries=1)\n', (2322, 2337), False, 'from requests.adapters import HTTPAdapter\n'), ((4972, 5000), 'urlp... |
import yaml
from ruamel.yaml import YAML
from ruamel.yaml.error import YAMLError
try:
from yaml import CSafeLoader as SafeLoader
except ImportError:
from yaml import SafeLoader
from dvc.exceptions import StageFileCorruptedError
from dvc.utils.compat import open
def load_stage_file(path):
with open(path,... | [
"dvc.utils.compat.open",
"dvc.exceptions.StageFileCorruptedError",
"yaml.dump",
"yaml.load",
"ruamel.yaml.YAML"
] | [((310, 343), 'dvc.utils.compat.open', 'open', (['path', '"""r"""'], {'encoding': '"""utf-8"""'}), "(path, 'r', encoding='utf-8')\n", (314, 343), False, 'from dvc.utils.compat import open\n'), ((960, 966), 'ruamel.yaml.YAML', 'YAML', ([], {}), '()\n', (964, 966), False, 'from ruamel.yaml import YAML\n'), ((1132, 1165),... |
import numpy as np
from pyad.nn import NeuralNet
from sklearn.datasets import load_breast_cancer
from sklearn.model_selection import train_test_split
np.random.seed(0)
data = load_breast_cancer()
X_train, X_test, y_train, y_test = train_test_split(
data.data, data.target, train_size=0.8, random_state=0
)
nn = Ne... | [
"sklearn.model_selection.train_test_split",
"sklearn.datasets.load_breast_cancer",
"numpy.max",
"pyad.nn.NeuralNet",
"numpy.random.seed"
] | [((151, 168), 'numpy.random.seed', 'np.random.seed', (['(0)'], {}), '(0)\n', (165, 168), True, 'import numpy as np\n'), ((176, 196), 'sklearn.datasets.load_breast_cancer', 'load_breast_cancer', ([], {}), '()\n', (194, 196), False, 'from sklearn.datasets import load_breast_cancer\n'), ((233, 305), 'sklearn.model_selecti... |
import ssl
import nltk
from textblob import TextBlob
from nltk.corpus import stopwords
# set SSL
try:
_create_unverified_https_context = ssl._create_unverified_context
except AttributeError:
pass
else:
ssl._create_default_https_context = _create_unverified_https_context
# download noun data (if required... | [
"textblob.TextBlob",
"nltk.corpus.stopwords.words",
"nltk.download"
] | [((322, 344), 'nltk.download', 'nltk.download', (['"""brown"""'], {}), "('brown')\n", (335, 344), False, 'import nltk\n'), ((345, 367), 'nltk.download', 'nltk.download', (['"""punkt"""'], {}), "('punkt')\n", (358, 367), False, 'import nltk\n'), ((368, 394), 'nltk.download', 'nltk.download', (['"""stopwords"""'], {}), "... |
from django_celery_beat.models import PeriodicTask, IntervalSchedule
from django.core.management.base import BaseCommand
from django.db import IntegrityError
class Command(BaseCommand):
def handle(self, *args, **options):
try:
schedule_channel, created = IntervalSchedule.objects.get_or_create... | [
"django_celery_beat.models.PeriodicTask.objects.create",
"django_celery_beat.models.IntervalSchedule.objects.get_or_create"
] | [((282, 360), 'django_celery_beat.models.IntervalSchedule.objects.get_or_create', 'IntervalSchedule.objects.get_or_create', ([], {'every': '(4)', 'period': 'IntervalSchedule.HOURS'}), '(every=4, period=IntervalSchedule.HOURS)\n', (320, 360), False, 'from django_celery_beat.models import PeriodicTask, IntervalSchedule\n... |
# required modules
import numpy as np
import matplotlib.pyplot as plt
import matplotlib.gridspec as gridspec
from matplotlib import cm
from matplotlib.colors import Normalize
from mpl_toolkits.mplot3d import Axes3D
from matplotlib.animation import FuncAnimation
# two-dimesional version
def plot_mse_loss_surface_2d(fi... | [
"numpy.sqrt",
"matplotlib.pyplot.ylabel",
"numpy.array",
"numpy.gradient",
"numpy.arange",
"matplotlib.pyplot.xlabel",
"matplotlib.pyplot.plot",
"numpy.diff",
"numpy.linspace",
"matplotlib.gridspec.GridSpec",
"numpy.min",
"numpy.meshgrid",
"matplotlib.pyplot.cm.ScalarMappable",
"matplotlib... | [((434, 480), 'numpy.linspace', 'np.linspace', (['w1_range[0]', 'w1_range[1]'], {'num': 'n_w'}), '(w1_range[0], w1_range[1], num=n_w)\n', (445, 480), True, 'import numpy as np\n'), ((502, 548), 'numpy.linspace', 'np.linspace', (['w2_range[0]', 'w2_range[1]'], {'num': 'n_w'}), '(w2_range[0], w2_range[1], num=n_w)\n', (5... |
# Copyright 2019 Google LLC
#
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,... | [
"qkeras.quantized_bits",
"numpy.sqrt",
"numpy.random.rand",
"qkeras.utils.quantized_model_from_json",
"qkeras.QActivation",
"numpy.array",
"tensorflow.keras.backend.clear_session",
"qkeras.extract_model_operations",
"os.remove",
"tensorflow.keras.layers.Input",
"qkeras.binary",
"numpy.testing.... | [((1761, 1793), 'tensorflow.keras.layers.Input', 'Input', (['(28, 28, 1)'], {'name': '"""input"""'}), "((28, 28, 1), name='input')\n", (1766, 1793), False, 'from tensorflow.keras.layers import Input\n'), ((2913, 2946), 'tensorflow.keras.models.Model', 'Model', ([], {'inputs': '[x_in]', 'outputs': '[x]'}), '(inputs=[x_i... |
from typing import Optional, List, TypeVar, Generic, Callable
import discord.ui
from .item import Item
from .select_option import SelectOption
from .custom import CustomSelect
def _default_check(_: discord.Interaction) -> bool:
return True
C = TypeVar("C", bound=discord.ui.Select)
class Select(Item, Generic... | [
"typing.TypeVar"
] | [((254, 291), 'typing.TypeVar', 'TypeVar', (['"""C"""'], {'bound': 'discord.ui.Select'}), "('C', bound=discord.ui.Select)\n", (261, 291), False, 'from typing import Optional, List, TypeVar, Generic, Callable\n')] |
#!/usr/bin/env python3
import os
import sys
import time
sys.path.append(os.getcwd()+'/lib')
import random
from dataclasses import dataclass, field
from ObsInfo import ObsInfo
def generate_random_obs(num_obs: int, size_list: list, config_data):
"""
config_file_name = "config.json"
json_file = open(config_f... | [
"ObsInfo.ObsInfo",
"random.uniform",
"os.getcwd"
] | [((72, 83), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (81, 83), False, 'import os\n'), ((548, 659), 'random.uniform', 'random.uniform', (["config_data['LAB_SPACE_LIMIT']['LIMIT_X'][0]", "config_data['LAB_SPACE_LIMIT']['LIMIT_X'][1]"], {}), "(config_data['LAB_SPACE_LIMIT']['LIMIT_X'][0], config_data[\n 'LAB_SPACE_L... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
# License: BSD-3 (https://tldrlegal.com/license/bsd-3-clause-license-(revised))
# Copyright (c) 2016-2021, <NAME>; Luczywo, Nadia
# All rights reserved.
# =============================================================================
# DOCS
# ===============================... | [
"numpy.asarray",
"numpy.min"
] | [((2605, 2620), 'numpy.asarray', 'np.asarray', (['arr'], {}), '(arr)\n', (2615, 2620), True, 'import numpy as np\n'), ((2632, 2669), 'numpy.min', 'np.min', (['arr'], {'axis': 'axis', 'keepdims': '(True)'}), '(arr, axis=axis, keepdims=True)\n', (2638, 2669), True, 'import numpy as np\n')] |
# -*- coding: utf-8 -*-
#
# This class was auto-generated from the API references found at
# https://support.direct.ingenico.com/documentation/api/reference/
#
from ingenico.direct.sdk.data_object import DataObject
from ingenico.direct.sdk.domain.address import Address
from ingenico.direct.sdk.domain.company_informatio... | [
"ingenico.direct.sdk.domain.address.Address",
"ingenico.direct.sdk.domain.company_information.CompanyInformation",
"ingenico.direct.sdk.domain.personal_information_token.PersonalInformationToken"
] | [((2597, 2606), 'ingenico.direct.sdk.domain.address.Address', 'Address', ([], {}), '()\n', (2604, 2606), False, 'from ingenico.direct.sdk.domain.address import Address\n'), ((2941, 2961), 'ingenico.direct.sdk.domain.company_information.CompanyInformation', 'CompanyInformation', ([], {}), '()\n', (2959, 2961), False, 'f... |
import time
import pykeyboard
# TODO: Replace following two lines with the code that activate the application.
print('Activate the application 3 seconds.')
time.sleep(3)
k = pykeyboard.PyKeyboard()
k.press_key(k.left_key)
time.sleep(1) # Hold down left key for 1 second.
k.release_key(k.left_key)
| [
"pykeyboard.PyKeyboard",
"time.sleep"
] | [((159, 172), 'time.sleep', 'time.sleep', (['(3)'], {}), '(3)\n', (169, 172), False, 'import time\n'), ((178, 201), 'pykeyboard.PyKeyboard', 'pykeyboard.PyKeyboard', ([], {}), '()\n', (199, 201), False, 'import pykeyboard\n'), ((226, 239), 'time.sleep', 'time.sleep', (['(1)'], {}), '(1)\n', (236, 239), False, 'import t... |
import os
import tarfile
from abc import ABC, abstractmethod
from glob import glob
import shutil
import random
import zstandard
"""
This registry is for automatically downloading and extracting datasets.
To register a class you need to inherit the DataDownloader class, provide name, filetype and url attributes, and
(... | [
"tarfile.open",
"os.makedirs",
"os.path.join",
"os.environ.get",
"os.path.isfile",
"os.path.isdir",
"os.path.basename",
"os.system",
"zstandard.ZstdDecompressor",
"os.remove"
] | [((648, 684), 'os.environ.get', 'os.environ.get', (['"""DATA_DIR"""', '"""./data"""'], {}), "('DATA_DIR', './data')\n", (662, 684), False, 'import os\n'), ((4552, 4588), 'os.makedirs', 'os.makedirs', (['DATA_DIR'], {'exist_ok': '(True)'}), '(DATA_DIR, exist_ok=True)\n', (4563, 4588), False, 'import os\n'), ((1524, 1562... |
# -*- coding: utf-8 -*-
import os
from django.db import models
from django.db.models.signals import post_delete
from django.dispatch import receiver
from .base import Pessoa
from djangosige.apps.login.models import Usuario
from djangosige.configs.settings import MEDIA_ROOT
def logo_directory_path(inst... | [
"django.db.models.ForeignKey",
"os.path.splitext",
"os.path.join",
"django.db.models.ImageField",
"django.dispatch.receiver",
"django.db.models.CharField"
] | [((1620, 1657), 'django.dispatch.receiver', 'receiver', (['post_delete'], {'sender': 'Empresa'}), '(post_delete, sender=Empresa)\n', (1628, 1657), False, 'from django.dispatch import receiver\n'), ((535, 638), 'django.db.models.ImageField', 'models.ImageField', ([], {'upload_to': 'logo_directory_path', 'default': '"""i... |
import os
from nltk.translate.bleu_score import corpus_bleu
from nltk.translate.bleu_score import SmoothingFunction
import json
from tqdm import tqdm, trange
from random import sample
import numpy as np
import pickle
import argparse
import bert_eval_acc
import svm_eval_acc
smooth = SmoothingFunction()
def eval_bleu... | [
"nltk.translate.bleu_score.corpus_bleu",
"nltk.translate.bleu_score.SmoothingFunction",
"json.loads",
"argparse.ArgumentParser",
"svm_eval_acc.main",
"bert_eval_acc.main"
] | [((285, 304), 'nltk.translate.bleu_score.SmoothingFunction', 'SmoothingFunction', ([], {}), '()\n', (302, 304), False, 'from nltk.translate.bleu_score import SmoothingFunction\n'), ((5559, 5584), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (5582, 5584), False, 'import argparse\n'), ((5799, 5... |
"""Constants for the UniFi component."""
import logging
LOGGER = logging.getLogger(__package__)
DOMAIN = "unifi"
CONTROLLER_ID = "{host}-{site}"
CONF_CONTROLLER = "controller"
CONF_SITE_ID = "site"
UNIFI_WIRELESS_CLIENTS = "unifi_wireless_clients"
CONF_ALLOW_BANDWIDTH_SENSORS = "allow_bandwidth_sensors"
CONF_BLOCK... | [
"logging.getLogger"
] | [((66, 96), 'logging.getLogger', 'logging.getLogger', (['__package__'], {}), '(__package__)\n', (83, 96), False, 'import logging\n')] |
import numpy as np
from treelas import post_order, TreeInstance
def test_demo_3x7_postord():
parent = np.array([0, 4, 5, 0, 3, 4, 7, 8, 5, 6, 7, 8,
9, 14, 17, 12, 15, 16, 19, 16, 17])
po = post_order(parent, include_root=True)
expect = np.array([12, 11, 19, 20, 21, 14, 15, 18, 17, 1... | [
"numpy.abs",
"numpy.unique",
"treelas.post_order",
"treelas.TreeInstance",
"numpy.array",
"numpy.fromstring"
] | [((108, 193), 'numpy.array', 'np.array', (['[0, 4, 5, 0, 3, 4, 7, 8, 5, 6, 7, 8, 9, 14, 17, 12, 15, 16, 19, 16, 17]'], {}), '([0, 4, 5, 0, 3, 4, 7, 8, 5, 6, 7, 8, 9, 14, 17, 12, 15, 16, 19, 16,\n 17])\n', (116, 193), True, 'import numpy as np\n'), ((222, 259), 'treelas.post_order', 'post_order', (['parent'], {'inclu... |
import argparse
import os
import torch
import yaml
DEFAULT_DEVICE = 'cuda:0'
def load_config():
parser = argparse.ArgumentParser(description='UNet3D training')
parser.add_argument('--config', type=str, help='Path to the YAML config file', required=True)
args = parser.parse_args()
config = _load_conf... | [
"torch.cuda.is_available",
"argparse.ArgumentParser"
] | [((113, 167), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""UNet3D training"""'}), "(description='UNet3D training')\n", (136, 167), False, 'import argparse\n'), ((473, 498), 'torch.cuda.is_available', 'torch.cuda.is_available', ([], {}), '()\n', (496, 498), False, 'import torch\n')] |
import tensorflow as tf
from tensorflow import keras
from tensorflow.keras import layers
from load_cora import load_cora
from baseline_model import create_ffn
from utils import run_experiment
from utils import display_learning_curves
# Graph convolution layer
class GraphConvLayer(layers.Layer):
def __init__(
... | [
"tensorflow.expand_dims",
"load_cora.load_cora",
"tensorflow.unstack",
"tensorflow.ones",
"tensorflow.math.unsorted_segment_mean",
"tensorflow.nn.l2_normalize",
"utils.display_learning_curves",
"utils.run_experiment",
"tensorflow.concat",
"tensorflow.math.unsorted_segment_sum",
"tensorflow.gathe... | [((6709, 6729), 'load_cora.load_cora', 'load_cora', ([], {'verbose': '(1)'}), '(verbose=1)\n', (6718, 6729), False, 'from load_cora import load_cora\n'), ((7127, 7156), 'tensorflow.ones', 'tf.ones', ([], {'shape': 'edges.shape[1]'}), '(shape=edges.shape[1])\n', (7134, 7156), True, 'import tensorflow as tf\n'), ((7921, ... |
from db import db
class RisklayerPrognosis(db.Model):
__tablename__ = 'risklayer_prognosis'
datenbestand = db.Column(db.TIMESTAMP, primary_key=True, nullable=False)
prognosis = db.Column(db.Float, nullable=False)
# class RisklayerPrognosisSchema(SQLAlchemyAutoSchema):
# class Meta:
# strict ... | [
"db.db.Column"
] | [((118, 175), 'db.db.Column', 'db.Column', (['db.TIMESTAMP'], {'primary_key': '(True)', 'nullable': '(False)'}), '(db.TIMESTAMP, primary_key=True, nullable=False)\n', (127, 175), False, 'from db import db\n'), ((192, 227), 'db.db.Column', 'db.Column', (['db.Float'], {'nullable': '(False)'}), '(db.Float, nullable=False)... |
# SPDX-FileCopyrightText: 2014 MicroPython & CircuitPython contributors (https://github.com/adafruit/circuitpython/graphs/contributors)
#
# SPDX-License-Identifier: MIT
import argparse
import os
import sys
sys.path.append("../../tools/usb_descriptor")
from adafruit_usb_descriptor import audio, audio10, cdc, hid, mi... | [
"adafruit_usb_descriptor.cdc.CallManagement",
"adafruit_usb_descriptor.midi.InJackDescriptor",
"argparse.FileType",
"adafruit_usb_descriptor.midi.OutJackDescriptor",
"adafruit_usb_descriptor.midi.Header",
"argparse.ArgumentParser",
"adafruit_usb_descriptor.cdc.Header",
"adafruit_usb_descriptor.cdc.Uni... | [((209, 254), 'sys.path.append', 'sys.path.append', (['"""../../tools/usb_descriptor"""'], {}), "('../../tools/usb_descriptor')\n", (224, 254), False, 'import sys\n'), ((821, 885), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Generate USB descriptors."""'}), "(description='Generate USB... |
from tqdm import tqdm
import pandas as pd
import numpy as np, argparse, time, pickle, random, os, datetime
import torch
import torch.optim as optim
from model import MaskedNLLLoss, BC_LSTM
from dataloader import MELDDataLoader
from sklearn.metrics import f1_score, confusion_matrix, accuracy_score, classification_re... | [
"sklearn.metrics.classification_report",
"numpy.array",
"torch.cuda.is_available",
"argparse.ArgumentParser",
"model.BC_LSTM",
"dataloader.MELDDataLoader",
"numpy.random.seed",
"numpy.concatenate",
"numpy.argmin",
"torch.argmax",
"numpy.argmax",
"time.time",
"sklearn.metrics.accuracy_score",... | [((427, 450), 'torch.manual_seed', 'torch.manual_seed', (['seed'], {}), '(seed)\n', (444, 450), False, 'import torch\n'), ((455, 483), 'torch.cuda.manual_seed', 'torch.cuda.manual_seed', (['seed'], {}), '(seed)\n', (477, 483), False, 'import torch\n'), ((488, 520), 'torch.cuda.manual_seed_all', 'torch.cuda.manual_seed_... |
# noinspection PyUnresolvedReferences
import os
import re
# TODO I'm going to need to make a dictionary for my big list of stuff i care about and what's needed for
# every file type....
RAF = ['EXIF:LensModel', 'MakerNotes:RawImageHeight', 'MakerNotes:RawImageWidth', 'EXIF:CreateDate', 'EXIF:ModifyDate',
'EXI... | [
"re.split",
"os.popen",
"os.path.splitext"
] | [((3281, 3305), 'os.path.splitext', 'os.path.splitext', (['filein'], {}), '(filein)\n', (3297, 3305), False, 'import os\n'), ((1764, 1788), 'os.path.splitext', 'os.path.splitext', (['filein'], {}), '(filein)\n', (1780, 1788), False, 'import os\n'), ((2026, 2053), 're.split', 're.split', (['"""\\\\s+:\\\\s+"""', 'each']... |
import logging
from typing import List, Callable
import numpy as np
from pyquaternion import Quaternion
from pyrep import PyRep
from pyrep.errors import IKError
from pyrep.objects import Dummy, Object
from rlbench import utils
from rlbench.action_modes import ArmActionMode, ActionMode
from rlbench.backend.exceptions ... | [
"numpy.flip",
"numpy.abs",
"numpy.allclose",
"numpy.random.get_state",
"numpy.linalg.pinv",
"numpy.minimum",
"pyquaternion.Quaternion",
"numpy.square",
"pyrep.objects.Dummy.create",
"numpy.array",
"numpy.append",
"numpy.dot",
"numpy.matmul",
"numpy.linalg.norm",
"numpy.shape",
"numpy.t... | [((1721, 1735), 'pyrep.objects.Dummy.create', 'Dummy.create', ([], {}), '()\n', (1733, 1735), False, 'from pyrep.objects import Dummy, Object\n'), ((8536, 8557), 'numpy.array', 'np.array', (['action[:-1]'], {}), '(action[:-1])\n', (8544, 8557), True, 'import numpy as np\n'), ((14997, 15012), 'numpy.transpose', 'np.tran... |
from django import forms
from django.contrib.contenttypes.forms import generic_inlineformset_factory
from django.contrib.contenttypes.models import ContentType
from django.db import models
from django.test import TestCase
from django.test.utils import isolate_apps
from .models import (
Animal, ForProxyMode... | [
"django.db.models.PositiveIntegerField",
"django.contrib.contenttypes.models.ContentType.objects.get_for_model",
"django.test.utils.isolate_apps",
"django.contrib.contenttypes.forms.generic_inlineformset_factory"
] | [((7814, 7847), 'django.test.utils.isolate_apps', 'isolate_apps', (['"""generic_relations"""'], {}), "('generic_relations')\n", (7826, 7847), False, 'from django.test.utils import isolate_apps\n'), ((692, 742), 'django.contrib.contenttypes.forms.generic_inlineformset_factory', 'generic_inlineformset_factory', (['Tagged... |
#!/usr/bin/python
# -*- coding: utf-8 -*-
# Copyright (C) 2017, <NAME> <<EMAIL>>
# vim: set ts=4 sts=4 sw=4 expandtab smartindent:
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without ... | [
"os.path.exists",
"os.mkdir"
] | [((2249, 2272), 'os.path.exists', 'path.exists', (['"""jug-demo"""'], {}), "('jug-demo')\n", (2260, 2272), False, 'from os import path\n'), ((2350, 2370), 'os.mkdir', 'os.mkdir', (['"""jug-demo"""'], {}), "('jug-demo')\n", (2358, 2370), False, 'import os\n')] |
#!/usr/bin/env python
# ROS Libraries
import actionlib
from actionlib_msgs.msg import GoalStatus
from control_msgs.msg import JointTrajectoryControllerState, FollowJointTrajectoryAction, FollowJointTrajectoryGoal
from kuri_wandering_robot.msg import Power
from wandering_behavior.msg import WanderAction, WanderGoal
impo... | [
"rospy.logwarn",
"rospy.init_node",
"rospy.Rate",
"threading.Lock",
"wandering_behavior.msg.WanderGoal",
"rospy.Duration.from_sec",
"rospy.spin",
"rospy.Subscriber",
"sent_messages_database.SentMessagesDatabase.load",
"actionlib.SimpleActionClient",
"rospy.get_param",
"rospy.Time.now",
"rosp... | [((19163, 19202), 'rospy.init_node', 'rospy.init_node', (['"""kuri_wandering_robot"""'], {}), "('kuri_wandering_robot')\n", (19178, 19202), False, 'import rospy\n'), ((19257, 19269), 'rospy.spin', 'rospy.spin', ([], {}), '()\n', (19267, 19269), False, 'import rospy\n'), ((2419, 2451), 'rospy.get_param', 'rospy.get_para... |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
#
# Copyright (C) 2010 <NAME> <<EMAIL>>
# Licensed under the GNU LGPL v2.1 - http://www.gnu.org/licenses/lgpl.html
"""
Automated tests for checking transformation algorithms (the models package).
"""
import logging
import unittest
import numpy as np
from gensim.corpora... | [
"logging.basicConfig",
"gensim.models.rpmodel.RpModel.load",
"gensim.matutils.sparse2full",
"numpy.allclose",
"gensim.test.utils.get_tmpfile",
"gensim.models.rpmodel.RpModel",
"numpy.array",
"numpy.random.seed",
"gensim.test.utils.datapath",
"unittest.main"
] | [((2218, 2314), 'logging.basicConfig', 'logging.basicConfig', ([], {'format': '"""%(asctime)s : %(levelname)s : %(message)s"""', 'level': 'logging.DEBUG'}), "(format='%(asctime)s : %(levelname)s : %(message)s',\n level=logging.DEBUG)\n", (2237, 2314), False, 'import logging\n'), ((2315, 2330), 'unittest.main', 'unit... |
import tensorflow as tf
from tensorflow import keras
class CondGeneratorModel(keras.Model):
def __init__(self):
super(CondGeneratorModel, self).__init__()
# Expand 7*7*128 features into a (7,7,128) tensor
self.dense_1 = keras.layers.Dense(7*7*256)
self.reshape_1 = keras.layers.Resh... | [
"tensorflow.keras.layers.Reshape",
"tensorflow.keras.layers.Conv2DTranspose",
"tensorflow.keras.layers.Conv2D",
"tensorflow.keras.layers.Dropout",
"tensorflow.keras.layers.LeakyReLU",
"tensorflow.keras.layers.BatchNormalization",
"tensorflow.keras.layers.Embedding",
"tensorflow.math.multiply",
"tens... | [((250, 281), 'tensorflow.keras.layers.Dense', 'keras.layers.Dense', (['(7 * 7 * 256)'], {}), '(7 * 7 * 256)\n', (268, 281), False, 'from tensorflow import keras\n'), ((303, 336), 'tensorflow.keras.layers.Reshape', 'keras.layers.Reshape', (['(7, 7, 256)'], {}), '((7, 7, 256))\n', (323, 336), False, 'from tensorflow imp... |
# -*- coding: utf-8 -*-
from selectable.decorators import login_required
from maestros.models import TiposMedidasActuacion, TiposLimitesCriticos, TiposMedidasVigilancia, TiposTemperaturas, TiposFrecuencias, Zonas, Terceros, CatalogoEquipos, Personal, Consumibles, ParametrosAnalisis, Actividades, Etapas, Peligros, Tipos... | [
"selectable.registry.registry.register",
"maestros_generales.models.Empresas.objects.filter"
] | [((1100, 1140), 'selectable.registry.registry.register', 'registry.register', (['TPActuacionPrevLookup'], {}), '(TPActuacionPrevLookup)\n', (1117, 1140), False, 'from selectable.registry import registry\n'), ((1686, 1726), 'selectable.registry.registry.register', 'registry.register', (['TPActuacionCorrLookup'], {}), '(... |
"""Define commands for Python 2.7"""
import argparse
import traceback
from . import util
from .cmd import run
from .cmd import extractpipenv
def main():
"""Main function"""
print("This version is not supported! It has limitted analysis features")
parser = argparse.ArgumentParser(description='Analyze Jupyt... | [
"traceback.print_exc",
"argparse.ArgumentParser"
] | [((270, 334), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Analyze Jupyter Notebooks"""'}), "(description='Analyze Jupyter Notebooks')\n", (293, 334), False, 'import argparse\n'), ((778, 799), 'traceback.print_exc', 'traceback.print_exc', ([], {}), '()\n', (797, 799), False, 'import tr... |
# -*- coding: utf-8 -*-
"""This python module aims to manage
`DokuWiki <https://www.dokuwiki.org/dokuwiki>`_ wikis by using the
provided `XML-RPC API <https://www.dokuwiki.org/devel:xmlrpc>`_. It is
compatible with python2.7 and python3+.
Installation
------------
It is on `PyPi <https://pypi.python.org/pypi/dokuwik... | [
"os.path.exists",
"collections.OrderedDict",
"xmlrpclib.Binary",
"os.makedirs",
"datetime.datetime.utcnow",
"datetime.datetime.strptime",
"re.compile",
"base64.b64encode",
"os.path.join",
"xmlrpclib.Transport.parse_response",
"base64.b64decode",
"datetime.datetime.now",
"xmlrpclib.Transport.... | [((945, 1007), 're.compile', 're.compile', (['"""(?P<proto>https?)://(?P<host>[^/]*)(?P<uri>/.*)?"""'], {}), "('(?P<proto>https?)://(?P<host>[^/]*)(?P<uri>/.*)?')\n", (955, 1007), False, 'import re\n'), ((1261, 1310), 'datetime.datetime.strptime', 'datetime.strptime', (['date[:-5]', '"""%Y-%m-%dT%H:%M:%S"""'], {}), "(d... |
import setuptools
import re
with open("README.md", "r") as fh:
long_description = fh.read()
# get version from _version.py file, from below
# https://stackoverflow.com/questions/458550/standard-way-to-embed-version-into-python-package
VERSION_FILE = "test_aide/_version.py"
version_file_str = open(VERSION_FILE, "r... | [
"setuptools.find_packages",
"re.search"
] | [((391, 440), 're.search', 're.search', (['VERSION_STR_RE', 'version_file_str', 're.M'], {}), '(VERSION_STR_RE, version_file_str, re.M)\n', (400, 440), False, 'import re\n'), ((977, 1003), 'setuptools.find_packages', 'setuptools.find_packages', ([], {}), '()\n', (1001, 1003), False, 'import setuptools\n')] |
#! /usr/bin/env python3
import importlib
import logging
import os
import subprocess
from setuptools import setup
from setuptools.command.install import install as install
from setuptools.command.develop import develop as develop
logger = logging.getLogger(__name__)
stan_model_files = [
os.path.join("nonperiod... | [
"logging.getLogger",
"setuptools.command.install.install.run",
"setuptools.command.install.install.finalize_options",
"logging.info",
"pbio.misc.shell_utils.check_programs_exist",
"os.path.exists",
"subprocess.call",
"shlex.quote",
"setuptools.setup",
"logging.warning",
"os.path.dirname",
"pbi... | [((242, 269), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (259, 269), False, 'import logging\n'), ((4350, 4416), 'setuptools.setup', 'setup', ([], {'cmdclass': "{'install': SetupInstall, 'develop': SetupDevelop}"}), "(cmdclass={'install': SetupInstall, 'develop': SetupDevelop})\n", (43... |
import os
import h5py
import nibabel as nb
import numpy as np
import torch
import torch.utils.data as data
from torchvision import transforms
import utils.preprocessor as preprocessor
# transform_train = transforms.Compose([
# transforms.RandomCrop(200, padding=56),
# transforms.ToTensor(),
# ])
class Imdb... | [
"os.listdir",
"nibabel.load",
"os.path.join",
"torch.from_numpy",
"utils.preprocessor.estimate_weights_mfb",
"utils.preprocessor.remap_labels",
"numpy.max",
"utils.preprocessor.rotate_orientation",
"numpy.min",
"utils.preprocessor.reduce_slices",
"utils.preprocessor.remove_black",
"numpy.round... | [((4062, 4083), 'nibabel.load', 'nb.load', (['file_path[0]'], {}), '(file_path[0])\n', (4069, 4083), True, 'import nibabel as nb\n'), ((4766, 4828), 'utils.preprocessor.rotate_orientation', 'preprocessor.rotate_orientation', (['volume', 'labelmap', 'orientation'], {}), '(volume, labelmap, orientation)\n', (4797, 4828),... |
# -*- coding: utf-8 -*-
# Form implementation generated from reading ui file 'design.ui'
#
# Created by: PyQt5 UI code generator 5.15.4
#
# WARNING: Any manual changes made to this file will be lost when pyuic5 is
# run again. Do not edit this file unless you know what you are doing.
from PyQt5 import QtC... | [
"PyQt5.QtWidgets.QWidget",
"PyQt5.QtWidgets.QPlainTextEdit",
"PyQt5.QtCore.QMetaObject.connectSlotsByName",
"PyQt5.QtCore.QRect",
"PyQt5.QtWidgets.QLabel",
"PyQt5.QtWidgets.QPushButton",
"PyQt5.QtWidgets.QCheckBox",
"PyQt5.QtWidgets.QLineEdit"
] | [((527, 556), 'PyQt5.QtWidgets.QWidget', 'QtWidgets.QWidget', (['MainWindow'], {}), '(MainWindow)\n', (544, 556), False, 'from PyQt5 import QtCore, QtGui, QtWidgets\n'), ((638, 674), 'PyQt5.QtWidgets.QLabel', 'QtWidgets.QLabel', (['self.centralwidget'], {}), '(self.centralwidget)\n', (654, 674), False, 'from PyQt5 impo... |
from simulation.car import spawn_drivers
from simulation.passenger import spawn_passengers
from simulation.core import World, Clock
conf = {
"x": 100,
"y": 100,
"drivers": 200,
"users": 1000,
"start": "2019-07-08T00:00:00",
"end": "2019-07-08T00:01:00"
}
clock = Clock(conf["start"], conf["end"... | [
"simulation.passenger.spawn_passengers",
"simulation.car.spawn_drivers",
"simulation.core.World",
"simulation.core.Clock"
] | [((289, 322), 'simulation.core.Clock', 'Clock', (["conf['start']", "conf['end']"], {}), "(conf['start'], conf['end'])\n", (294, 322), False, 'from simulation.core import World, Clock\n'), ((363, 405), 'simulation.core.World', 'World', (["[conf['x'], conf['y']]"], {'clock': 'clock'}), "([conf['x'], conf['y']], clock=clo... |
import json
d1 = {}
with open("/home/qinyuan/zs/out/bart-large-with-description-grouped-1e-5-outerbsz4-innerbsz32-adapterdim4-unfreeze-dec29/test_predictions.jsonl") as fin:
for line in fin:
d = json.loads(line)
d1[d["id"]] = d["output"][0]["answer"]
d2 = {}
dq = {}
with open("/home/qinyuan/zs/out... | [
"json.loads"
] | [((208, 224), 'json.loads', 'json.loads', (['line'], {}), '(line)\n', (218, 224), False, 'import json\n'), ((450, 466), 'json.loads', 'json.loads', (['line'], {}), '(line)\n', (460, 466), False, 'import json\n'), ((663, 679), 'json.loads', 'json.loads', (['line'], {}), '(line)\n', (673, 679), False, 'import json\n')] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.