code
stringlengths
20
1.05M
apis
list
extract_api
stringlengths
75
5.24M
# Copyright (c) 2010-2014 openpyxl import pytest from openpyxl.styles.borders import Border, Side from openpyxl.styles.fills import GradientFill from openpyxl.styles.colors import Color from openpyxl.writer.styles import StyleWriter from openpyxl.tests.helper import get_xml, compare_xml class DummyWorkbook: st...
[ "openpyxl.styles.colors.Color", "openpyxl.tests.helper.get_xml", "openpyxl.tests.helper.compare_xml", "openpyxl.styles.borders.Border" ]
[((552, 573), 'openpyxl.tests.helper.get_xml', 'get_xml', (['writer._root'], {}), '(writer._root)\n', (559, 573), False, 'from openpyxl.tests.helper import get_xml, compare_xml\n'), ((910, 936), 'openpyxl.tests.helper.compare_xml', 'compare_xml', (['xml', 'expected'], {}), '(xml, expected)\n', (921, 936), False, 'from ...
from unittest import TestCase from io import StringIO import json class TestDump(TestCase): def test_dump(self): sio = StringIO() json.dump({}, sio) self.assertEquals(sio.getvalue(), '{}') def test_dumps(self): self.assertEquals(json.dumps({}), '{}') def test_encode_truef...
[ "io.StringIO", "json.dumps", "json.dump" ]
[((133, 143), 'io.StringIO', 'StringIO', ([], {}), '()\n', (141, 143), False, 'from io import StringIO\n'), ((152, 170), 'json.dump', 'json.dump', (['{}', 'sio'], {}), '({}, sio)\n', (161, 170), False, 'import json\n'), ((272, 286), 'json.dumps', 'json.dumps', (['{}'], {}), '({})\n', (282, 286), False, 'import json\n')...
# -*- coding: utf-8 -*- # # Tencent is pleased to support the open source community by making QT4C available. # Copyright (C) 2020 THL A29 Limited, a Tencent company. All rights reserved. # QT4C is licensed under the BSD 3-Clause License, except for the third-party components listed below. # A copy of the BSD 3-Cla...
[ "os.path.abspath", "os.path.dirname", "sys.argv.count", "unittest.TestLoader" ]
[((439, 464), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (454, 464), False, 'import os\n'), ((485, 510), 'os.path.dirname', 'os.path.dirname', (['test_dir'], {}), '(test_dir)\n', (500, 510), False, 'import os\n'), ((611, 632), 'unittest.TestLoader', 'unittest.TestLoader', ([], {}), '()\n'...
from cffi import FFI ffibuilder = FFI() ffibuilder.cdef(""" int test(int t); """) ffibuilder.set_source("_pi_cffi", """ #include "brute.h" """, sources=['brute.c']) if __name__ == "__main__": ffibuilder.compile(verbose = Tru...
[ "cffi.FFI" ]
[((35, 40), 'cffi.FFI', 'FFI', ([], {}), '()\n', (38, 40), False, 'from cffi import FFI\n')]
"""Board Module""" import copy from typing import Tuple, List from src.coordinate import Coordinate from src.snake import Snake class Board: """Track the cooardinates for all snakes and food in the game.""" def __init__(self, data): self._data = data self._snakes = None self._foods = No...
[ "src.snake.Snake", "src.coordinate.Coordinate", "copy.deepcopy" ]
[((1708, 1727), 'copy.deepcopy', 'copy.deepcopy', (['self'], {}), '(self)\n', (1721, 1727), False, 'import copy\n'), ((493, 510), 'src.snake.Snake', 'Snake', (['snake_data'], {}), '(snake_data)\n', (498, 510), False, 'from src.snake import Snake\n'), ((789, 810), 'src.coordinate.Coordinate', 'Coordinate', (['food_data'...
import os import warnings from django.conf import settings CAPTCHA_FONT_PATH = getattr(settings, 'CAPTCHA_FONT_PATH', os.path.normpath(os.path.join(os.path.dirname(__file__), '..', 'fonts/Vera.ttf'))) CAPTCHA_FONT_SIZE = getattr(settings, 'CAPTCHA_FONT_SIZE', 22) CAPTCHA_LETTER_ROTATION = getattr(settings, 'CAPTCHA_L...
[ "warnings.warn", "os.path.dirname" ]
[((2103, 2141), 'warnings.warn', 'warnings.warn', (['msg', 'DeprecationWarning'], {}), '(msg, DeprecationWarning)\n', (2116, 2141), False, 'import warnings\n'), ((2368, 2406), 'warnings.warn', 'warnings.warn', (['msg', 'DeprecationWarning'], {}), '(msg, DeprecationWarning)\n', (2381, 2406), False, 'import warnings\n'),...
# Generated by Django 2.2.21 on 2021-06-23 12:43 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('resources', '0125_add_timmi_payload_model'), ] operations = [ migrations.AddField( model_name=...
[ "django.db.models.BooleanField" ]
[((409, 521), 'django.db.models.BooleanField', 'models.BooleanField', ([], {'default': '(False)', 'verbose_name': '"""Disallow overlapping reservations in this unit per user."""'}), "(default=False, verbose_name=\n 'Disallow overlapping reservations in this unit per user.')\n", (428, 521), False, 'from django.db imp...
import logging from lora_multihop import serial_connection, variables def config_module(configuration=variables.MODULE_CONFIG): if serial_connection.execute_command(configuration, [variables.STATUS_OK]): serial_connection.execute_command('AT+SEND=1', [variables.STATUS_OK]) serial_connection.execu...
[ "logging.debug", "lora_multihop.serial_connection.execute_command", "logging.warning", "lora_multihop.serial_connection.response_q.get", "lora_multihop.serial_connection.str_to_bytes", "lora_multihop.serial_connection.bytes_to_str" ]
[((138, 209), 'lora_multihop.serial_connection.execute_command', 'serial_connection.execute_command', (['configuration', '[variables.STATUS_OK]'], {}), '(configuration, [variables.STATUS_OK])\n', (171, 209), False, 'from lora_multihop import serial_connection, variables\n'), ((445, 491), 'logging.warning', 'logging.war...
import abc from typing import Dict, Callable import tensorflow as tf from flink_ml_framework.context import Context from flink_ml_framework.java_file import * from ..runner import tf_helper, io_helper from ..runner.output_writer import DirectOutputWriter try: from flink_ml_tensorflow.tensorflow_context import TF...
[ "flink_ml_tensorflow2.tensorflow_context.TFContext", "tensorflow.data.TFRecordDataset", "tensorflow.config.threading.set_intra_op_parallelism_threads", "importlib.import_module" ]
[((1728, 1746), 'flink_ml_tensorflow2.tensorflow_context.TFContext', 'TFContext', (['context'], {}), '(context)\n', (1737, 1746), False, 'from flink_ml_tensorflow2.tensorflow_context import TFContext\n'), ((1311, 1347), 'importlib.import_module', 'importlib.import_module', (['module_name'], {}), '(module_name)\n', (133...
import pytest ENCODING = 'utf-8' @pytest.fixture(scope='function', autouse=True) def setup_case(request): def destroy_case(): from corm import annihilate_keyspace_tables, SESSIONS annihilate_keyspace_tables('mykeyspace') for keyspace_name, session in SESSIONS.copy().items(): if...
[ "corm.register_table", "corm.select", "corm.keyspace_exists", "datetime.datetime.utcnow", "corm.annihilate_keyspace_tables", "corm.sync_schema", "corm.insert", "corm.keyspace_destroy", "uuid.uuid4", "corm.cp", "corm.obtain_session", "pytest.fixture", "corm.SESSIONS.copy", "random.randint",...
[((36, 82), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""function"""', 'autouse': '(True)'}), "(scope='function', autouse=True)\n", (50, 82), False, 'import pytest\n'), ((735, 760), 'corm.register_table', 'register_table', (['TestModel'], {}), '(TestModel)\n', (749, 760), False, 'from corm import register_tab...
import json import os from utilities.SaveLoadJson import SaveLoadJson as SLJ from utilities.LineCount import LineCount as LC import subprocess from geolite2 import geolite2 class getData: #Get Data Functions ------------------------------------------------------ @staticmethod def getDATA(): resu...
[ "utilities.SaveLoadJson.SaveLoadJson.load", "subprocess.Popen", "json.dumps", "geolite2.geolite2.reader" ]
[((789, 807), 'json.dumps', 'json.dumps', (['result'], {}), '(result)\n', (799, 807), False, 'import json\n'), ((865, 890), 'utilities.SaveLoadJson.SaveLoadJson.load', 'SLJ.load', (['"""dataStore.txt"""'], {}), "('dataStore.txt')\n", (873, 890), True, 'from utilities.SaveLoadJson import SaveLoadJson as SLJ\n'), ((1124,...
# All credit to https://stackoverflow.com/questions/46571448/tkinter-and-a-html-file - thanks DELICA - https://stackoverflow.com/users/7027346/delica from cefpython3 import cefpython as cef import ctypes try: import tkinter as tk from tkinter import messagebox except ImportError: import Tkinter as tk impo...
[ "logging.getLogger", "Tkinter.Grid.rowconfigure", "logging.StreamHandler", "tkinter.messagebox.askokcancel", "Tkinter.Frame.__init__", "logging.Formatter", "Tkinter.Tk", "Tkinter.Tcl", "cefpython3.cefpython.MessageLoopWork", "cefpython3.cefpython.Initialize", "platform.system", "platform.archi...
[((418, 435), 'cefpython3.cefpython.WindowUtils', 'cef.WindowUtils', ([], {}), '()\n', (433, 435), True, 'from cefpython3 import cefpython as cef\n'), ((589, 622), 'logging.getLogger', '_logging.getLogger', (['"""tkinter_.py"""'], {}), "('tkinter_.py')\n", (607, 622), True, 'import logging as _logging\n'), ((4152, 4176...
#!/usr/bin/env python3 # encoding: utf-8 import sys import urllib.parse import selenium.webdriver def exit(): driver.quit() sys.exit(0) driver = selenium.webdriver.Firefox() # for some reason, detectportal.firefox.com and connectivitycheck.gstatic.com are not blocked # therefore, they cannot be used to detect con...
[ "sys.exit" ]
[((129, 140), 'sys.exit', 'sys.exit', (['(0)'], {}), '(0)\n', (137, 140), False, 'import sys\n')]
from django.shortcuts import render from django.contrib.auth.decorators import login_required from django.http import JsonResponse from django.views.decorators.csrf import csrf_exempt from . import helpers # Create your views here. @csrf_exempt def convert_video(request, version): # Get video video = reques...
[ "django.http.JsonResponse" ]
[((532, 565), 'django.http.JsonResponse', 'JsonResponse', (['context'], {'safe': '(False)'}), '(context, safe=False)\n', (544, 565), False, 'from django.http import JsonResponse\n')]
from itertools import count import numpy as np class Particle(object): """Object containing all the properties for a single particle""" _ids = count(0) def __init__(self, main_data=None, x=np.zeros(2)): self.id = next(self._ids) self.main_data = main_data self.x = np.array(x) ...
[ "numpy.array", "numpy.zeros", "itertools.count" ]
[((154, 162), 'itertools.count', 'count', (['(0)'], {}), '(0)\n', (159, 162), False, 'from itertools import count\n'), ((205, 216), 'numpy.zeros', 'np.zeros', (['(2)'], {}), '(2)\n', (213, 216), True, 'import numpy as np\n'), ((305, 316), 'numpy.array', 'np.array', (['x'], {}), '(x)\n', (313, 316), True, 'import numpy ...
from unittest import TestCase from pyRdfa import pyRdfa class NonXhtmlTest(TestCase): """ RDFa that is in not well-formed XHTML is passed through html5lib. These tests make sure that this RDFa can be processed both from a file, and from a URL. """ target1 = '<og:isbn>9780596516499</og:isbn>'...
[ "pyRdfa.pyRdfa" ]
[((459, 467), 'pyRdfa.pyRdfa', 'pyRdfa', ([], {}), '()\n', (465, 467), False, 'from pyRdfa import pyRdfa\n'), ((623, 631), 'pyRdfa.pyRdfa', 'pyRdfa', ([], {}), '()\n', (629, 631), False, 'from pyRdfa import pyRdfa\n')]
from des109 import moeda preco = float(input('Digite o preço pretendido: €')) print(f'''A metade do preço é {(moeda.metade(preco))} O dobro do preço é {(moeda.dobra(preco))} Aumentando o preço 10% temos {(moeda.aumentar(preco, 10))} Diminuindo o preço 13% temos {(moeda.aumentar(preco, 13))}''')
[ "des109.moeda.metade", "des109.moeda.dobra", "des109.moeda.aumentar" ]
[((113, 132), 'des109.moeda.metade', 'moeda.metade', (['preco'], {}), '(preco)\n', (125, 132), False, 'from des109 import moeda\n'), ((163, 181), 'des109.moeda.dobra', 'moeda.dobra', (['preco'], {}), '(preco)\n', (174, 181), False, 'from des109 import moeda\n'), ((214, 239), 'des109.moeda.aumentar', 'moeda.aumentar', (...
# coding=utf-8 # *** WARNING: this file was generated by the Pulumi SDK Generator. *** # *** Do not edit by hand unless you're certain you know what you are doing! *** import warnings import pulumi import pulumi.runtime from typing import Any, Mapping, Optional, Sequence, Union, overload from .. import _utilities from...
[ "pulumi.get", "pulumi.Alias", "pulumi.getter", "pulumi.set", "pulumi.ResourceOptions", "pulumi.ResourceOptions.merge" ]
[((6655, 6694), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""resourceGroupName"""'}), "(name='resourceGroupName')\n", (6668, 6694), False, 'import pulumi\n'), ((7070, 7122), 'pulumi.getter', 'pulumi.getter', ([], {'name': '"""authorizedAzureSubscriptionIds"""'}), "(name='authorizedAzureSubscriptionIds')\n", (708...
import os import glob import cv2 import numpy as np import torch from torchvision.transforms import transforms from natsort import natsorted from models import resmasking_dropout1 from utils.datasets.fer2013dataset import EMOTION_DICT from barez import show transform = transforms.Compose( [ transforms.ToPI...
[ "torchvision.transforms.transforms.ToPILImage", "models.resmasking_dropout1", "numpy.uint8", "cv2.resize", "torch.mean", "torch.unsqueeze", "torch.load", "numpy.min", "torch.flatten", "numpy.max", "torchvision.transforms.transforms.ToTensor", "os.path.basename", "numpy.concatenate", "torch...
[((774, 799), 'models.resmasking_dropout1', 'resmasking_dropout1', (['(3)', '(7)'], {}), '(3, 7)\n', (793, 799), False, 'from models import resmasking_dropout1\n'), ((894, 971), 'torch.load', 'torch.load', (['"""./saved/checkpoints/Z_resmasking_dropout1_rot30_2019Nov30_13.32"""'], {}), "('./saved/checkpoints/Z_resmaski...
# coding=utf-8 # *** WARNING: this file was generated by the Pulumi SDK Generator. *** # *** Do not edit by hand unless you're certain you know what you are doing! *** import warnings import pulumi import pulumi.runtime from typing import Any, Mapping, Optional, Sequence, Union from ... import _utilities, _tables from...
[ "pulumi.get", "pulumi.getter", "pulumi.set", "warnings.warn", "pulumi.log.warn", "pulumi.runtime.invoke", "pulumi.InvokeOptions" ]
[((436, 620), 'warnings.warn', 'warnings.warn', (['"""The \'latest\' version is deprecated. Please migrate to the function in the top-level module: \'azure-native:notificationhubs:getNamespace\'."""', 'DeprecationWarning'], {}), '(\n "The \'latest\' version is deprecated. Please migrate to the function in the top-le...
import json from pygments import highlight from pygments.lexers import JsonLexer from pygments.formatters import TerminalFormatter def print_json_obj(json_object): json_str = json.dumps(json_object, indent=4, sort_keys=True) print(highlight(json_str, JsonLexer(), TerminalFormatter())) def print_json_str(jso...
[ "pygments.lexers.JsonLexer", "json.dumps", "pygments.formatters.TerminalFormatter" ]
[((181, 230), 'json.dumps', 'json.dumps', (['json_object'], {'indent': '(4)', 'sort_keys': '(True)'}), '(json_object, indent=4, sort_keys=True)\n', (191, 230), False, 'import json\n'), ((261, 272), 'pygments.lexers.JsonLexer', 'JsonLexer', ([], {}), '()\n', (270, 272), False, 'from pygments.lexers import JsonLexer\n'),...
#!/usr/bin/env python3 import os from opendbc.can.parser import CANParser from cereal import car from selfdrive.car.interfaces import RadarInterfaceBase RADAR_MSGS_C = list(range(0x2c2, 0x2d4+2, 2)) # c_ messages 706,...,724 RADAR_MSGS_D = list(range(0x2a2, 0x2b4+2, 2)) # d_ messages LAST_MSG = max(RADAR_MSGS_C + RA...
[ "cereal.car.RadarData.new_message", "os.path.splitext", "cereal.car.RadarData.RadarPoint.new_message" ]
[((2329, 2356), 'cereal.car.RadarData.new_message', 'car.RadarData.new_message', ([], {}), '()\n', (2354, 2356), False, 'from cereal import car\n'), ((1589, 1612), 'os.path.splitext', 'os.path.splitext', (['dbc_f'], {}), '(dbc_f)\n', (1605, 1612), False, 'import os\n'), ((2670, 2708), 'cereal.car.RadarData.RadarPoint.n...
import os import numpy as np import tensorflow as tf from image_quality.utils import utils class TrainDataGenerator(tf.keras.utils.Sequence): '''inherits from Keras Sequence base object, allows to use multiprocessing in .fit_generator''' def __init__(self, samples, img_dir, batch_size, n_classes, basenet_preproc...
[ "image_quality.utils.utils.random_crop", "image_quality.utils.utils.load_image", "image_quality.utils.utils.random_horizontal_flip", "image_quality.utils.utils.normalize_labels", "numpy.random.shuffle" ]
[((1440, 1471), 'numpy.random.shuffle', 'np.random.shuffle', (['self.indexes'], {}), '(self.indexes)\n', (1457, 1471), True, 'import numpy as np\n'), ((1878, 1924), 'image_quality.utils.utils.load_image', 'utils.load_image', (['img_file', 'self.img_load_dims'], {}), '(img_file, self.img_load_dims)\n', (1894, 1924), Fal...
# -*- coding: utf-8 -*- """ Licensed to the Apache Software Foundation (ASF) under one or more contributor license agreements. See the NOTICE file distributed with this work for additional information regarding copyright ownership. The ASF licenses this file to you under the Apache License, Versi...
[ "qiskit.quantum_info.operators.channel.Chi", "qiskit.quantum_info.operators.channel.Kraus", "qiskit.quantum_info.operators.channel.Choi", "qiskit.quantum_info.operators.channel.PTM", "qat.comm.quops.ttypes.QuantumChannel", "numpy.array", "qiskit.quantum_info.operators.channel.SuperOp", "numpy.real", ...
[((1512, 1556), 'qat.comm.datamodel.ttypes.Matrix', 'Matrix', (['array.shape[0]', 'array.shape[1]', 'data'], {}), '(array.shape[0], array.shape[1], data)\n', (1518, 1556), False, 'from qat.comm.datamodel.ttypes import Matrix, ComplexNumber\n'), ((2444, 2558), 'qat.comm.quops.ttypes.QuantumChannel', 'QuantumChannel', ([...
""" util.auth2: Authentication tools This module is based off of util.auth, except with the action paradigm removed. """ from flask import session from app.models import Account from app.util import course as course_util # Session keys SESSION_EMAIL = 'email' def create_account(email: str, password: str, f...
[ "app.models.Account.objects.get_or_404", "app.util.course.set_courses", "app.models.Account" ]
[((795, 893), 'app.models.Account', 'Account', ([], {'email': 'email', 'first_name': 'first_name', 'last_name': 'last_name', 'fsuid': 'fsuid', 'is_admin': '(False)'}), '(email=email, first_name=first_name, last_name=last_name, fsuid=\n fsuid, is_admin=False)\n', (802, 893), False, 'from app.models import Account\n')...
from PyQt5.QtWidgets import * from matplotlib.backends.backend_qt5agg import FigureCanvas from matplotlib.figure import Figure from matplotlib.backends.backend_qt5agg import NavigationToolbar2QT as NavigationToolbar class PstaticWidget(QWidget): def __init__(self, parent=None): QWidget.__init__(self...
[ "matplotlib.figure.Figure", "matplotlib.backends.backend_qt5agg.NavigationToolbar2QT", "matplotlib.backends.backend_qt5agg.FigureCanvas" ]
[((360, 368), 'matplotlib.figure.Figure', 'Figure', ([], {}), '()\n', (366, 368), False, 'from matplotlib.figure import Figure\n'), ((451, 481), 'matplotlib.backends.backend_qt5agg.FigureCanvas', 'FigureCanvas', (['self.fig_pstatic'], {}), '(self.fig_pstatic)\n', (463, 481), False, 'from matplotlib.backends.backend_qt5...
import os class Config: CSRF_ENABLED = True SECRET_KEY = 'your-very-very-secret-key' SQLALCHEMY_DATABASE_URI = 'postgresql:///flask_template_dev' SQLALCHEMY_TRACK_MODIFICATIONS = False SQLALCHEMY_ECHO = True class Development(Config): ENV = 'development' DEBUG = True TESTING = False ...
[ "os.getenv" ]
[((419, 610), 'os.getenv', 'os.getenv', (['"""DATABASE_URL"""', '"""postgres://firhokdcdnfygz:93231d3f2ae1156cabfc40f7e4ba08587a77f68a5e2072fbcbbdb30150ba4bcb@ec2-107-22-253-158.compute-1.amazonaws.com:5432/df9c5vvl0s21da"""'], {}), "('DATABASE_URL',\n 'postgres://firhokdcdnfygz:93231d3f2ae1156cabfc40f7e4ba08587a77f...
import itertools import numpy as np import pandas as pd def find_intersections(formula_lists,group_labels,exclusive = True): """ Docstring for function pyKrev.find_intersections ==================== This function compares n lists of molecular formula and outputs a dictionary containing the intersections...
[ "pandas.DataFrame", "itertools.combinations" ]
[((1483, 1515), 'pandas.DataFrame', 'pd.DataFrame', ([], {'data': 'formula_lists'}), '(data=formula_lists)\n', (1495, 1515), True, 'import pandas as pd\n'), ((1176, 1215), 'itertools.combinations', 'itertools.combinations', (['group_labels', 'i'], {}), '(group_labels, i)\n', (1198, 1215), False, 'import itertools\n')]
import os import glob import shutil from tinytag import TinyTag """ root = 'C:/' copy_to = '/copy to/folder' tag = TinyTag.get('C:/Users/jchap/OneDrive/Pictures/(VERYRAREBOYZ) (feat. $ki Mask The Slump God and Drugz).mp3') print(tag.artist) print('song duration: '+str(tag.duration)) """ f = [] f=glob.gl...
[ "os.path.join", "os.path.dirname", "tinytag.TinyTag.get", "glob.glob", "os.walk" ]
[((313, 355), 'glob.glob', 'glob.glob', (['"""C:/Users/jchap/OneDrive/*.mp3"""'], {}), "('C:/Users/jchap/OneDrive/*.mp3')\n", (322, 355), False, 'import glob\n'), ((544, 558), 'os.walk', 'os.walk', (['"""C:/"""'], {}), "('C:/')\n", (551, 558), False, 'import os\n'), ((807, 837), 'tinytag.TinyTag.get', 'TinyTag.get', ([...
#import modules import os import csv #input csvpath = os.path.join('Resources', 'budget_data.csv') #output outfile = os.path.join('Analysis', 'pybankstatements.txt') #declare variables months = []; total_m = 1; net_total = 0; total_change = 0; monthly_changes = []; greatest_inc = ['', 0]; greatest_dec = ['', 0] #open...
[ "os.path.join", "csv.reader" ]
[((54, 98), 'os.path.join', 'os.path.join', (['"""Resources"""', '"""budget_data.csv"""'], {}), "('Resources', 'budget_data.csv')\n", (66, 98), False, 'import os\n'), ((117, 165), 'os.path.join', 'os.path.join', (['"""Analysis"""', '"""pybankstatements.txt"""'], {}), "('Analysis', 'pybankstatements.txt')\n", (129, 165)...
from rest_framework import serializers from cms.api.serializers import UniCMSContentTypeClass, UniCMSCreateUpdateSerializer from cms.medias.serializers import MediaSerializer from . models import Carousel, CarouselItem, CarouselItemLink, CarouselItemLinkLocalization, CarouselItemLocalization class CarouselForeignK...
[ "cms.medias.serializers.MediaSerializer" ]
[((2384, 2415), 'cms.medias.serializers.MediaSerializer', 'MediaSerializer', (['instance.image'], {}), '(instance.image)\n', (2399, 2415), False, 'from cms.medias.serializers import MediaSerializer\n')]
#!/usr/bin/env python3 """ Copyright (c) 2018-2021 Intel Corporation Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applic...
[ "cv2.vconcat", "model_api.performance_metrics.PerformanceMetrics", "images_capture.open_images_capture", "cv2.imshow", "logging.info", "argparse.ArgumentParser", "pathlib.Path", "time.perf_counter", "cv2.VideoWriter", "numpy.concatenate", "cv2.VideoWriter_fourcc", "cv2.waitKey", "numpy.squee...
[((1125, 1220), 'logging.basicConfig', 'log.basicConfig', ([], {'format': '"""[ %(levelname)s ] %(message)s"""', 'level': 'log.DEBUG', 'stream': 'sys.stdout'}), "(format='[ %(levelname)s ] %(message)s', level=log.DEBUG,\n stream=sys.stdout)\n", (1140, 1220), True, 'import logging as log\n'), ((1249, 1279), 'argparse...
# coding: utf-8 """ [AHOI cookbook](/ahoi/docs/cookbook/index.html) [Data Privacy](/sandboxmanager/#/privacy) [Terms of Service](/sandboxmanager/#/terms) [Imprint](https://sparkassen-hub.com/impressum/) &copy; 2016&dash;2017 Starfinanz - Ein Unternehmen der Finanz Informatik # noqa: E501 OpenAPI sp...
[ "six.iteritems" ]
[((6926, 6959), 'six.iteritems', 'six.iteritems', (['self.swagger_types'], {}), '(self.swagger_types)\n', (6939, 6959), False, 'import six\n')]
# Copyright 2015-2017 ARM Limited, Google and contributors # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applica...
[ "os.listdir", "trappy.register_dynamic_ftrace", "os.path.join", "shutil.copytree", "os.path.dirname", "json.load", "os.path.basename", "os.path.abspath", "trappy.FTrace", "trappy.unregister_dynamic_ftrace", "json.dump", "trappy.SysTrace" ]
[((2584, 2599), 'trappy.FTrace', 'trappy.FTrace', ([], {}), '()\n', (2597, 2599), False, 'import trappy\n'), ((2754, 2794), 'trappy.FTrace', 'trappy.FTrace', (['uncached_trace.trace_path'], {}), '(uncached_trace.trace_path)\n', (2767, 2794), False, 'import trappy\n'), ((4683, 4749), 'os.path.join', 'os.path.join', (['u...
from django.contrib.auth.decorators import login_required from django.contrib.auth.models import User from django.shortcuts import render from django.urls import reverse from django.http import HttpResponseRedirect, HttpResponse from django.utils import timezone from olaf.models import * from olaf.forms import * from...
[ "django.shortcuts.render", "django.utils.timezone.now", "django.contrib.auth.models.User.objects.filter", "django.urls.reverse", "olaf.chess.controller.proccess_move", "olaf.utility.usertools.get_translated_game_board", "olaf.utility.usertools.new_game", "olaf.utility.usertools.logout_user" ]
[((3350, 3391), 'django.shortcuts.render', 'render', (['request', 'fail_template', 'fail_args'], {}), '(request, fail_template, fail_args)\n', (3356, 3391), False, 'from django.shortcuts import render\n'), ((5689, 5719), 'olaf.utility.usertools.logout_user', 'usertools.logout_user', (['request'], {}), '(request)\n', (5...
#!/usr/bin/python # -*- coding: utf-8 -*- # [Import start] from flask import Blueprint, jsonify # [Import end] app = Blueprint( 'hoge', __name__, url_prefix='/hoge' ) @app.route('/test') def hoge(): return "\nhogehoge"
[ "flask.Blueprint" ]
[((119, 166), 'flask.Blueprint', 'Blueprint', (['"""hoge"""', '__name__'], {'url_prefix': '"""/hoge"""'}), "('hoge', __name__, url_prefix='/hoge')\n", (128, 166), False, 'from flask import Blueprint, jsonify\n')]
#!/usr/bin/env python3 import apt_pkg import sys from apt_pkg import CURSTATE_INSTALLED, version_compare from operator import lt, le, eq, ge, gt # Function mappings for relationship operators. relation_operators = {"<<": lt, "<=": le, "=": eq, ">=": ge, ">>": gt} # Set up APT cache. apt_pkg.init() cache = apt_pkg.Ca...
[ "apt_pkg.version_compare", "apt_pkg.Cache", "apt_pkg.init" ]
[((287, 301), 'apt_pkg.init', 'apt_pkg.init', ([], {}), '()\n', (299, 301), False, 'import apt_pkg\n'), ((310, 329), 'apt_pkg.Cache', 'apt_pkg.Cache', (['None'], {}), '(None)\n', (323, 329), False, 'import apt_pkg\n'), ((2105, 2147), 'apt_pkg.version_compare', 'version_compare', (['installed_version', 'pkgver'], {}), '...
# Generated by Django 3.0.7 on 2020-08-24 06:17 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('datasets', '0008_auto_20200821_1427'), ] operations = [ migrations.AddField( model_name='rawdar', name='AsB', ...
[ "django.db.models.FloatField", "django.db.models.CharField" ]
[((332, 372), 'django.db.models.FloatField', 'models.FloatField', ([], {'blank': '(True)', 'null': '(True)'}), '(blank=True, null=True)\n', (349, 372), False, 'from django.db import migrations, models\n'), ((493, 632), 'django.db.models.CharField', 'models.CharField', ([], {'choices': "[('1', 'below detection level'), ...
import os import math import time import geohash import geojson from geojson import MultiLineString from shapely import geometry import shapefile import numpy import datetime as dt import pandas as pd import logging logger = logging.getLogger(__name__) source_shape_file_path = "C:/temp/2018/" threshold = 60*60 cols = ...
[ "logging.getLogger", "shapefile.Reader", "datetime.datetime.strptime", "os.path.join", "pandas.DataFrame", "os.walk" ]
[((226, 253), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (243, 253), False, 'import logging\n'), ((454, 485), 'os.walk', 'os.walk', (['source_shape_file_path'], {}), '(source_shape_file_path)\n', (461, 485), False, 'import os\n'), ((2012, 2045), 'pandas.DataFrame', 'pd.DataFrame', (['...
import pygame_sdl2 pygame_sdl2.import_as_pygame() import pygame import os import random import math from Ball import Ball def save_state(balls): """ Saves the game state. """ stateString = "" with open("state.txt", "w") as f: for ball in balls: stateString += "{} {} {} {} {}".f...
[ "os.path.exists", "pygame.init", "pygame.event.get", "pygame.display.set_mode", "pygame.display.flip", "pygame.display.Info", "pygame_sdl2.import_as_pygame", "os.unlink", "pygame.time.Clock", "pygame.font.Font", "random.randint", "Ball.Ball" ]
[((19, 49), 'pygame_sdl2.import_as_pygame', 'pygame_sdl2.import_as_pygame', ([], {}), '()\n', (47, 49), False, 'import pygame_sdl2\n'), ((999, 1026), 'os.path.exists', 'os.path.exists', (['"""state.txt"""'], {}), "('state.txt')\n", (1013, 1026), False, 'import os\n'), ((1076, 1089), 'pygame.init', 'pygame.init', ([], {...
from __future__ import division import math, copy import argparse from brownian import Brownian import scipy import LLRcalc class sprt: def __init__(self, alpha=0.05, beta=0.05, elo0=0, elo1=5, elo_model="logistic"): assert elo_model in ("logistic", "normalized") self.elo_model = elo_model ...
[ "LLRcalc.results_to_pdf", "argparse.ArgumentParser", "math.sqrt", "math.log", "LLRcalc.stats", "LLRcalc.L_", "brownian.Brownian" ]
[((4157, 4182), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (4180, 4182), False, 'import argparse\n'), ((330, 358), 'math.log', 'math.log', (['(beta / (1 - alpha))'], {}), '(beta / (1 - alpha))\n', (338, 358), False, 'import math, copy\n'), ((376, 404), 'math.log', 'math.log', (['((1 - beta)...
"""Simple Hail query example.""" import click import hail as hl from bokeh.io.export import get_screenshot_as_png from analysis_runner import output_path GNOMAD_HGDP_1KG_MT = ( 'gs://gcp-public-data--gnomad/release/3.1/mt/genomes/' 'gnomad.genomes.v3.1.hgdp_1kg_subset_dense.mt' ) @click.command() @click.op...
[ "hail.hadoop_open", "click.option", "hail.sample_qc", "hail.hadoop_exists", "hail.read_matrix_table", "hail.init", "bokeh.io.export.get_screenshot_as_png", "click.command", "analysis_runner.output_path", "hail.plot.histogram" ]
[((295, 310), 'click.command', 'click.command', ([], {}), '()\n', (308, 310), False, 'import click\n'), ((312, 397), 'click.option', 'click.option', (['"""--rerun"""'], {'help': '"""Whether to overwrite cached files"""', 'default': '(False)'}), "('--rerun', help='Whether to overwrite cached files', default=False\n )...
# Copyright 2020 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applica...
[ "tensorflow.python.framework.convert_to_constants.convert_variables_to_constants_from_session_graph", "collections.namedtuple", "tensorflow.python.ops.random_ops.random_uniform", "tensorflow.python.saved_model.load.load", "tensorflow.python.client.session.Session", "tensorflow.core.protobuf.config_pb2.Con...
[((3602, 3623), 'functools.lru_cache', 'functools.lru_cache', ([], {}), '()\n', (3621, 3623), False, 'import functools\n'), ((4355, 4376), 'functools.lru_cache', 'functools.lru_cache', ([], {}), '()\n', (4374, 4376), False, 'import functools\n'), ((4806, 4892), 'collections.namedtuple', 'collections.namedtuple', (['"""...
# Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. from typing import List, Pattern from recognizers_text.utilities import RegExpUtility from recognizers_number.number import BaseNumberParser from recognizers_number.number.italian.extractors import ItalianIntegerExtractor,...
[ "recognizers_number.ItalianCardinalExtractor", "recognizers_number.number.italian.extractors.ItalianIntegerExtractor", "recognizers_text.utilities.RegExpUtility.get_safe_reg_exp", "recognizers_number.ItalianOrdinalExtractor", "recognizers_number.number.italian.parsers.ItalianNumberParserConfiguration" ]
[((5851, 5915), 'recognizers_text.utilities.RegExpUtility.get_safe_reg_exp', 'RegExpUtility.get_safe_reg_exp', (['ItalianDateTime.AllHalfYearRegex'], {}), '(ItalianDateTime.AllHalfYearRegex)\n', (5881, 5915), False, 'from recognizers_text.utilities import RegExpUtility\n'), ((5956, 6023), 'recognizers_text.utilities.Re...
def test_setupcall(): """ Test the call of the setup function """ import jupyter_libertem_proxy as jx print("\nRunning test_setupcall...") print(jx.setup_libertem())
[ "jupyter_libertem_proxy.setup_libertem" ]
[((170, 189), 'jupyter_libertem_proxy.setup_libertem', 'jx.setup_libertem', ([], {}), '()\n', (187, 189), True, 'import jupyter_libertem_proxy as jx\n')]
import math import imageio import cv2 as cv import numpy as np import transformer def fix_rotation(img): img_copy = img.copy() img = cv.cvtColor(img, cv.COLOR_BGR2GRAY) rows, cols = img.shape img = cv.adaptiveThreshold(img, 255, cv.ADAPTIVE_THRESH_MEAN_C, cv.THRESH_BINARY_INV, 15, 9) kernel = cv.g...
[ "cv2.imshow", "cv2.warpPerspective", "cv2.destroyAllWindows", "cv2.approxPolyDP", "math.hypot", "imageio.get_writer", "imageio.get_reader", "cv2.arcLength", "cv2.medianBlur", "cv2.contourArea", "cv2.waitKey", "cv2.getPerspectiveTransform", "cv2.minEnclosingCircle", "cv2.morphologyEx", "c...
[((142, 177), 'cv2.cvtColor', 'cv.cvtColor', (['img', 'cv.COLOR_BGR2GRAY'], {}), '(img, cv.COLOR_BGR2GRAY)\n', (153, 177), True, 'import cv2 as cv\n'), ((215, 306), 'cv2.adaptiveThreshold', 'cv.adaptiveThreshold', (['img', '(255)', 'cv.ADAPTIVE_THRESH_MEAN_C', 'cv.THRESH_BINARY_INV', '(15)', '(9)'], {}), '(img, 255, cv...
# Altere o Programa 8.20 de forma que o usuário tenha três chances de acertar o número # O programa termina se o usuário acertar ou errar três vezes # Programa 8.20 do livro, página 184 # Programa 8.20 - Adivinhando o número # # import random # # n = random.randint(1, 10) # x = int(input('Escolha um número entre 1 e 1...
[ "random.randint" ]
[((435, 456), 'random.randint', 'random.randint', (['(1)', '(10)'], {}), '(1, 10)\n', (449, 456), False, 'import random\n')]
import setuptools # To use a consistent encoding from codecs import open from os import path here = path.abspath(path.dirname(__file__)) with open(path.join(here, 'README.md'), encoding='utf-8') as f: long_description = f.read() setuptools.setup( name="atm76", version="0.1.0", author="<NAME>", a...
[ "os.path.dirname", "setuptools.find_packages", "os.path.join" ]
[((115, 137), 'os.path.dirname', 'path.dirname', (['__file__'], {}), '(__file__)\n', (127, 137), False, 'from os import path\n'), ((150, 178), 'os.path.join', 'path.join', (['here', '"""README.md"""'], {}), "(here, 'README.md')\n", (159, 178), False, 'from os import path\n'), ((542, 568), 'setuptools.find_packages', 's...
import logging import asyncio from agent.check_plugins import AbstractCheckPlugin # Do khong biet dung thu vien asyncio ntn ca nen em dung thu vien request # python import requests import sys import time from datetime import datetime logger = logging.getLogger(__name__) class Download(AbstractCheckPlugin): @as...
[ "logging.getLogger", "datetime.datetime.now", "time.clock" ]
[((245, 272), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (262, 272), False, 'import logging\n'), ((448, 460), 'time.clock', 'time.clock', ([], {}), '()\n', (458, 460), False, 'import time\n'), ((728, 740), 'time.clock', 'time.clock', ([], {}), '()\n', (738, 740), False, 'import time\n...
import weakref import os import requests import ssl from ssl import SSLContext import logging from ssl_context_builder.builder.builder import SslContextBuilder from ssl_context_builder.http_impl.requests_wrapper.ssl_adapter import SslAdapter class RequestsSecureSession: def __init__(self, ssl_context: SSLContex...
[ "os.path.exists", "requests.Session", "logging.warning", "ssl.DER_cert_to_PEM_cert", "ssl_context_builder.http_impl.requests_wrapper.ssl_adapter.SslAdapter", "os.remove" ]
[((838, 856), 'requests.Session', 'requests.Session', ([], {}), '()\n', (854, 856), False, 'import requests\n'), ((2087, 2116), 'logging.warning', 'logging.warning', (['warn_message'], {}), '(warn_message)\n', (2102, 2116), False, 'import logging\n'), ((2660, 2680), 'os.path.exists', 'os.path.exists', (['path'], {}), '...
# Ghetto Fixtures from codebox import app from codebox.apps.auth.models import User from codebox.apps.snippets.models import Snippet from codebox.apps.organizations.models import Organization, OrganizationMember from flask import g client = app.test_client() _ctx = app.test_request_context() _ctx.push() app.preproces...
[ "codebox.app.test_client", "codebox.apps.snippets.models.Snippet.objects.create", "codebox.apps.organizations.models.Organization.objects.create", "codebox.app.preprocess_request", "codebox.apps.organizations.models.OrganizationMember.objects.create", "flask.g.redis.flushdb", "codebox.app.test_request_c...
[((243, 260), 'codebox.app.test_client', 'app.test_client', ([], {}), '()\n', (258, 260), False, 'from codebox import app\n'), ((268, 294), 'codebox.app.test_request_context', 'app.test_request_context', ([], {}), '()\n', (292, 294), False, 'from codebox import app\n'), ((307, 331), 'codebox.app.preprocess_request', 'a...
import function_exercise_01 as st st.sandwich_toppings('meatballs', 'salad')
[ "function_exercise_01.sandwich_toppings" ]
[((35, 77), 'function_exercise_01.sandwich_toppings', 'st.sandwich_toppings', (['"""meatballs"""', '"""salad"""'], {}), "('meatballs', 'salad')\n", (55, 77), True, 'import function_exercise_01 as st\n')]
import json import time from functools import lru_cache from multiprocessing import Pool, Process from threading import Thread, Timer from typing import Any, Dict, List from datetime import datetime import hashlib import inspect import requests import waitress from bottle import BaseTemplate, Bottle, request, response,...
[ "requests.post", "utils.utils.decompress", "bottle.Bottle", "multiprocessing.Process", "utils.logger.logger.info", "time.sleep", "core.Block.from_json", "utils.storage.read_header_list_from_db", "core.BlockChain", "bottle.error", "bottle.template", "utils.utils.compress", "json.dumps", "co...
[((723, 731), 'bottle.Bottle', 'Bottle', ([], {}), '()\n', (729, 731), False, 'from bottle import BaseTemplate, Bottle, request, response, static_file, template, error\n'), ((817, 829), 'core.BlockChain', 'BlockChain', ([], {}), '()\n', (827, 829), False, 'from core import Block, BlockChain, SingleOutput, Transaction, ...
import io import os from setuptools import setup def read(file_name): """Read a text file and return the content as a string.""" with io.open(os.path.join(os.path.dirname(__file__), file_name), encoding='utf-8') as f: return f.read() setup( name='recmetrics', url='https://gi...
[ "os.path.dirname" ]
[((166, 191), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (181, 191), False, 'import os\n')]
import pygame class Texto: def __init__(self, screen, text, x, y, text_size = 20, fuente = 'Calibri', italic = False, bold= False, subrayado= False, color = (250, 240, 230), bg = [] ): self.screen = screen fg = color self.coord = x, y #load font, prepare values f...
[ "pygame.Color", "pygame.font.Font", "pygame.font.SysFont" ]
[((326, 352), 'pygame.font.Font', 'pygame.font.Font', (['None', '(80)'], {}), '(None, 80)\n', (342, 352), False, 'import pygame\n'), ((421, 459), 'pygame.font.SysFont', 'pygame.font.SysFont', (['fuente', 'text_size'], {}), '(fuente, text_size)\n', (440, 459), False, 'import pygame\n'), ((2049, 2070), 'pygame.Color', 'p...
# This is the code to train the xgboost model with cross-validation for each unique room in the dataset. # Models are dumped into ./models and results are dumped into two csv files in the current work directory. import argparse import json import math import os import pickle import warnings from typing import Tuple i...
[ "pandas.read_csv", "numpy.array", "xgboost.DMatrix", "sklearn.metrics.r2_score", "numpy.random.RandomState", "argparse.ArgumentParser", "xgboost.train", "json.dumps", "pandas.set_option", "xgboost.cv", "pandas.DataFrame", "sklearn.model_selection.train_test_split", "hyperopt.hp.quniform", ...
[((792, 817), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (815, 817), False, 'import argparse\n'), ((1688, 1721), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {}), "('ignore')\n", (1711, 1721), False, 'import warnings\n'), ((1722, 1764), 'pandas.set_option', 'pd.s...
import os from setuptools import setup # Read the version g = {} with open(os.path.join("editorconfig", "version.py"), "rt") as fp: exec(fp.read(), g) v = g['VERSION'] version = ".".join(str(x) for x in v[:3]) if v[3] != "final": version += "-" + v[3] setup( name='EditorConfig', versio...
[ "os.path.join" ]
[((76, 118), 'os.path.join', 'os.path.join', (['"""editorconfig"""', '"""version.py"""'], {}), "('editorconfig', 'version.py')\n", (88, 118), False, 'import os\n')]
# -*- coding: utf-8 -*- """Tests for sktime annotators.""" import pandas as pd import pytest from sktime.registry import all_estimators from sktime.utils._testing.estimator_checks import _make_args ALL_ANNOTATORS = all_estimators(estimator_types="series-annotator", return_names=False) @pytest.mark.parametrize("Est...
[ "pytest.mark.parametrize", "sktime.utils._testing.estimator_checks._make_args", "sktime.registry.all_estimators" ]
[((218, 288), 'sktime.registry.all_estimators', 'all_estimators', ([], {'estimator_types': '"""series-annotator"""', 'return_names': '(False)'}), "(estimator_types='series-annotator', return_names=False)\n", (232, 288), False, 'from sktime.registry import all_estimators\n'), ((292, 344), 'pytest.mark.parametrize', 'pyt...
#!/usr/bin/env python from __future__ import print_function from kaldi.segmentation import NnetSAD, SegmentationProcessor from kaldi.nnet3 import NnetSimpleComputationOptions from kaldi.util.table import SequentialMatrixReader # Construct SAD model = NnetSAD.read_model("final.raw") post = NnetSAD.read_average_poster...
[ "kaldi.nnet3.NnetSimpleComputationOptions", "kaldi.segmentation.NnetSAD.make_sad_transform", "kaldi.util.table.SequentialMatrixReader", "kaldi.segmentation.NnetSAD.read_average_posteriors", "kaldi.segmentation.NnetSAD", "kaldi.segmentation.SegmentationProcessor", "kaldi.segmentation.NnetSAD.read_model",...
[((254, 285), 'kaldi.segmentation.NnetSAD.read_model', 'NnetSAD.read_model', (['"""final.raw"""'], {}), "('final.raw')\n", (272, 285), False, 'from kaldi.segmentation import NnetSAD, SegmentationProcessor\n'), ((293, 343), 'kaldi.segmentation.NnetSAD.read_average_posteriors', 'NnetSAD.read_average_posteriors', (['"""po...
""" Comparison between the efficiency of the Boyer-Moore algorithm and the naive substring search algorithm. The runtimes for both algorithms are plotted on the same axes. """ import matplotlib.pyplot as plt import numpy as np import string import time import random from bm_alg import boyer_moore_match, naive_match #...
[ "bm_alg.naive_match", "random.choice", "matplotlib.pyplot.ylabel", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.plot", "matplotlib.pyplot.title", "bm_alg.boyer_moore_match", "time.time", "matplotlib.pyplot.legend", "matplotlib.pyplot.show" ]
[((2371, 2416), 'matplotlib.pyplot.plot', 'plt.plot', (['x', 'y_naive'], {'label': '"""Naive Algorithm"""'}), "(x, y_naive, label='Naive Algorithm')\n", (2379, 2416), True, 'import matplotlib.pyplot as plt\n'), ((2421, 2469), 'matplotlib.pyplot.plot', 'plt.plot', (['x', 'y_bm'], {'label': '"""Boyer-Moore Algorithm"""'}...
import json from typing import Type, TYPE_CHECKING from django.core.exceptions import ObjectDoesNotExist from django.utils.decorators import method_decorator from django.views.decorators.cache import cache_page from rest_framework import viewsets, filters from rest_framework.exceptions import NotFound from rest_framew...
[ "json.dumps", "rest_framework.response.Response", "django.views.decorators.cache.cache_page", "indicators.views.GeoJSONRenderer", "maps.models.DataLayer.objects.get", "maps.models.DataLayer.objects.all" ]
[((1032, 1055), 'maps.models.DataLayer.objects.all', 'DataLayer.objects.all', ([], {}), '()\n', (1053, 1055), False, 'from maps.models import DataLayer\n'), ((1501, 1517), 'json.dumps', 'json.dumps', (['data'], {}), '(data)\n', (1511, 1517), False, 'import json\n'), ((3018, 3035), 'rest_framework.response.Response', 'R...
# -*- coding: utf-8 -*- """ Provide download function by request """ from datetime import datetime import logging import time import urllib.parse import requests from bs4 import BeautifulSoup class Throttle(object): """Throttle downloading by sleeping between requests to same domain.""" de...
[ "logging.warn", "requests.Session", "time.sleep", "requests.Request", "datetime.datetime.now", "logging.info", "logging.error" ]
[((891, 905), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (903, 905), False, 'from datetime import datetime\n'), ((1334, 1352), 'requests.Session', 'requests.Session', ([], {}), '()\n', (1350, 1352), False, 'import requests\n'), ((2118, 2161), 'requests.Request', 'requests.Request', (['"""GET"""', 'url']...
# -*- coding: utf-8 -*- # # The MIT License (MIT) # # Copyright (C) 2017 <NAME> <<EMAIL>> # # Permission is hereby granted, free of charge, to any person obtaining a copy of # this software and associated documentation files (the "Software"), to deal in # the Software without restriction, including without limitation t...
[ "urlparse.urljoin", "requests.adapters.HTTPAdapter", "requests.Session" ]
[((1981, 1999), 'requests.Session', 'requests.Session', ([], {}), '()\n', (1997, 1999), False, 'import requests\n'), ((2311, 2337), 'requests.adapters.HTTPAdapter', 'HTTPAdapter', ([], {'max_retries': '(1)'}), '(max_retries=1)\n', (2322, 2337), False, 'from requests.adapters import HTTPAdapter\n'), ((4972, 5000), 'urlp...
import yaml from ruamel.yaml import YAML from ruamel.yaml.error import YAMLError try: from yaml import CSafeLoader as SafeLoader except ImportError: from yaml import SafeLoader from dvc.exceptions import StageFileCorruptedError from dvc.utils.compat import open def load_stage_file(path): with open(path,...
[ "dvc.utils.compat.open", "dvc.exceptions.StageFileCorruptedError", "yaml.dump", "yaml.load", "ruamel.yaml.YAML" ]
[((310, 343), 'dvc.utils.compat.open', 'open', (['path', '"""r"""'], {'encoding': '"""utf-8"""'}), "(path, 'r', encoding='utf-8')\n", (314, 343), False, 'from dvc.utils.compat import open\n'), ((960, 966), 'ruamel.yaml.YAML', 'YAML', ([], {}), '()\n', (964, 966), False, 'from ruamel.yaml import YAML\n'), ((1132, 1165),...
import numpy as np from pyad.nn import NeuralNet from sklearn.datasets import load_breast_cancer from sklearn.model_selection import train_test_split np.random.seed(0) data = load_breast_cancer() X_train, X_test, y_train, y_test = train_test_split( data.data, data.target, train_size=0.8, random_state=0 ) nn = Ne...
[ "sklearn.model_selection.train_test_split", "sklearn.datasets.load_breast_cancer", "numpy.max", "pyad.nn.NeuralNet", "numpy.random.seed" ]
[((151, 168), 'numpy.random.seed', 'np.random.seed', (['(0)'], {}), '(0)\n', (165, 168), True, 'import numpy as np\n'), ((176, 196), 'sklearn.datasets.load_breast_cancer', 'load_breast_cancer', ([], {}), '()\n', (194, 196), False, 'from sklearn.datasets import load_breast_cancer\n'), ((233, 305), 'sklearn.model_selecti...
import ssl import nltk from textblob import TextBlob from nltk.corpus import stopwords # set SSL try: _create_unverified_https_context = ssl._create_unverified_context except AttributeError: pass else: ssl._create_default_https_context = _create_unverified_https_context # download noun data (if required...
[ "textblob.TextBlob", "nltk.corpus.stopwords.words", "nltk.download" ]
[((322, 344), 'nltk.download', 'nltk.download', (['"""brown"""'], {}), "('brown')\n", (335, 344), False, 'import nltk\n'), ((345, 367), 'nltk.download', 'nltk.download', (['"""punkt"""'], {}), "('punkt')\n", (358, 367), False, 'import nltk\n'), ((368, 394), 'nltk.download', 'nltk.download', (['"""stopwords"""'], {}), "...
from django_celery_beat.models import PeriodicTask, IntervalSchedule from django.core.management.base import BaseCommand from django.db import IntegrityError class Command(BaseCommand): def handle(self, *args, **options): try: schedule_channel, created = IntervalSchedule.objects.get_or_create...
[ "django_celery_beat.models.PeriodicTask.objects.create", "django_celery_beat.models.IntervalSchedule.objects.get_or_create" ]
[((282, 360), 'django_celery_beat.models.IntervalSchedule.objects.get_or_create', 'IntervalSchedule.objects.get_or_create', ([], {'every': '(4)', 'period': 'IntervalSchedule.HOURS'}), '(every=4, period=IntervalSchedule.HOURS)\n', (320, 360), False, 'from django_celery_beat.models import PeriodicTask, IntervalSchedule\n...
# required modules import numpy as np import matplotlib.pyplot as plt import matplotlib.gridspec as gridspec from matplotlib import cm from matplotlib.colors import Normalize from mpl_toolkits.mplot3d import Axes3D from matplotlib.animation import FuncAnimation # two-dimesional version def plot_mse_loss_surface_2d(fi...
[ "numpy.sqrt", "matplotlib.pyplot.ylabel", "numpy.array", "numpy.gradient", "numpy.arange", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.plot", "numpy.diff", "numpy.linspace", "matplotlib.gridspec.GridSpec", "numpy.min", "numpy.meshgrid", "matplotlib.pyplot.cm.ScalarMappable", "matplotlib...
[((434, 480), 'numpy.linspace', 'np.linspace', (['w1_range[0]', 'w1_range[1]'], {'num': 'n_w'}), '(w1_range[0], w1_range[1], num=n_w)\n', (445, 480), True, 'import numpy as np\n'), ((502, 548), 'numpy.linspace', 'np.linspace', (['w2_range[0]', 'w2_range[1]'], {'num': 'n_w'}), '(w2_range[0], w2_range[1], num=n_w)\n', (5...
# Copyright 2019 Google LLC # # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing,...
[ "qkeras.quantized_bits", "numpy.sqrt", "numpy.random.rand", "qkeras.utils.quantized_model_from_json", "qkeras.QActivation", "numpy.array", "tensorflow.keras.backend.clear_session", "qkeras.extract_model_operations", "os.remove", "tensorflow.keras.layers.Input", "qkeras.binary", "numpy.testing....
[((1761, 1793), 'tensorflow.keras.layers.Input', 'Input', (['(28, 28, 1)'], {'name': '"""input"""'}), "((28, 28, 1), name='input')\n", (1766, 1793), False, 'from tensorflow.keras.layers import Input\n'), ((2913, 2946), 'tensorflow.keras.models.Model', 'Model', ([], {'inputs': '[x_in]', 'outputs': '[x]'}), '(inputs=[x_i...
from typing import Optional, List, TypeVar, Generic, Callable import discord.ui from .item import Item from .select_option import SelectOption from .custom import CustomSelect def _default_check(_: discord.Interaction) -> bool: return True C = TypeVar("C", bound=discord.ui.Select) class Select(Item, Generic...
[ "typing.TypeVar" ]
[((254, 291), 'typing.TypeVar', 'TypeVar', (['"""C"""'], {'bound': 'discord.ui.Select'}), "('C', bound=discord.ui.Select)\n", (261, 291), False, 'from typing import Optional, List, TypeVar, Generic, Callable\n')]
#!/usr/bin/env python3 import os import sys import time sys.path.append(os.getcwd()+'/lib') import random from dataclasses import dataclass, field from ObsInfo import ObsInfo def generate_random_obs(num_obs: int, size_list: list, config_data): """ config_file_name = "config.json" json_file = open(config_f...
[ "ObsInfo.ObsInfo", "random.uniform", "os.getcwd" ]
[((72, 83), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (81, 83), False, 'import os\n'), ((548, 659), 'random.uniform', 'random.uniform', (["config_data['LAB_SPACE_LIMIT']['LIMIT_X'][0]", "config_data['LAB_SPACE_LIMIT']['LIMIT_X'][1]"], {}), "(config_data['LAB_SPACE_LIMIT']['LIMIT_X'][0], config_data[\n 'LAB_SPACE_L...
#!/usr/bin/env python # -*- coding: utf-8 -*- # License: BSD-3 (https://tldrlegal.com/license/bsd-3-clause-license-(revised)) # Copyright (c) 2016-2021, <NAME>; Luczywo, Nadia # All rights reserved. # ============================================================================= # DOCS # ===============================...
[ "numpy.asarray", "numpy.min" ]
[((2605, 2620), 'numpy.asarray', 'np.asarray', (['arr'], {}), '(arr)\n', (2615, 2620), True, 'import numpy as np\n'), ((2632, 2669), 'numpy.min', 'np.min', (['arr'], {'axis': 'axis', 'keepdims': '(True)'}), '(arr, axis=axis, keepdims=True)\n', (2638, 2669), True, 'import numpy as np\n')]
# -*- coding: utf-8 -*- # # This class was auto-generated from the API references found at # https://support.direct.ingenico.com/documentation/api/reference/ # from ingenico.direct.sdk.data_object import DataObject from ingenico.direct.sdk.domain.address import Address from ingenico.direct.sdk.domain.company_informatio...
[ "ingenico.direct.sdk.domain.address.Address", "ingenico.direct.sdk.domain.company_information.CompanyInformation", "ingenico.direct.sdk.domain.personal_information_token.PersonalInformationToken" ]
[((2597, 2606), 'ingenico.direct.sdk.domain.address.Address', 'Address', ([], {}), '()\n', (2604, 2606), False, 'from ingenico.direct.sdk.domain.address import Address\n'), ((2941, 2961), 'ingenico.direct.sdk.domain.company_information.CompanyInformation', 'CompanyInformation', ([], {}), '()\n', (2959, 2961), False, 'f...
import time import pykeyboard # TODO: Replace following two lines with the code that activate the application. print('Activate the application 3 seconds.') time.sleep(3) k = pykeyboard.PyKeyboard() k.press_key(k.left_key) time.sleep(1) # Hold down left key for 1 second. k.release_key(k.left_key)
[ "pykeyboard.PyKeyboard", "time.sleep" ]
[((159, 172), 'time.sleep', 'time.sleep', (['(3)'], {}), '(3)\n', (169, 172), False, 'import time\n'), ((178, 201), 'pykeyboard.PyKeyboard', 'pykeyboard.PyKeyboard', ([], {}), '()\n', (199, 201), False, 'import pykeyboard\n'), ((226, 239), 'time.sleep', 'time.sleep', (['(1)'], {}), '(1)\n', (236, 239), False, 'import t...
import os import tarfile from abc import ABC, abstractmethod from glob import glob import shutil import random import zstandard """ This registry is for automatically downloading and extracting datasets. To register a class you need to inherit the DataDownloader class, provide name, filetype and url attributes, and (...
[ "tarfile.open", "os.makedirs", "os.path.join", "os.environ.get", "os.path.isfile", "os.path.isdir", "os.path.basename", "os.system", "zstandard.ZstdDecompressor", "os.remove" ]
[((648, 684), 'os.environ.get', 'os.environ.get', (['"""DATA_DIR"""', '"""./data"""'], {}), "('DATA_DIR', './data')\n", (662, 684), False, 'import os\n'), ((4552, 4588), 'os.makedirs', 'os.makedirs', (['DATA_DIR'], {'exist_ok': '(True)'}), '(DATA_DIR, exist_ok=True)\n', (4563, 4588), False, 'import os\n'), ((1524, 1562...
# -*- coding: utf-8 -*- import os from django.db import models from django.db.models.signals import post_delete from django.dispatch import receiver from .base import Pessoa from djangosige.apps.login.models import Usuario from djangosige.configs.settings import MEDIA_ROOT def logo_directory_path(inst...
[ "django.db.models.ForeignKey", "os.path.splitext", "os.path.join", "django.db.models.ImageField", "django.dispatch.receiver", "django.db.models.CharField" ]
[((1620, 1657), 'django.dispatch.receiver', 'receiver', (['post_delete'], {'sender': 'Empresa'}), '(post_delete, sender=Empresa)\n', (1628, 1657), False, 'from django.dispatch import receiver\n'), ((535, 638), 'django.db.models.ImageField', 'models.ImageField', ([], {'upload_to': 'logo_directory_path', 'default': '"""i...
import os from nltk.translate.bleu_score import corpus_bleu from nltk.translate.bleu_score import SmoothingFunction import json from tqdm import tqdm, trange from random import sample import numpy as np import pickle import argparse import bert_eval_acc import svm_eval_acc smooth = SmoothingFunction() def eval_bleu...
[ "nltk.translate.bleu_score.corpus_bleu", "nltk.translate.bleu_score.SmoothingFunction", "json.loads", "argparse.ArgumentParser", "svm_eval_acc.main", "bert_eval_acc.main" ]
[((285, 304), 'nltk.translate.bleu_score.SmoothingFunction', 'SmoothingFunction', ([], {}), '()\n', (302, 304), False, 'from nltk.translate.bleu_score import SmoothingFunction\n'), ((5559, 5584), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (5582, 5584), False, 'import argparse\n'), ((5799, 5...
"""Constants for the UniFi component.""" import logging LOGGER = logging.getLogger(__package__) DOMAIN = "unifi" CONTROLLER_ID = "{host}-{site}" CONF_CONTROLLER = "controller" CONF_SITE_ID = "site" UNIFI_WIRELESS_CLIENTS = "unifi_wireless_clients" CONF_ALLOW_BANDWIDTH_SENSORS = "allow_bandwidth_sensors" CONF_BLOCK...
[ "logging.getLogger" ]
[((66, 96), 'logging.getLogger', 'logging.getLogger', (['__package__'], {}), '(__package__)\n', (83, 96), False, 'import logging\n')]
import numpy as np from treelas import post_order, TreeInstance def test_demo_3x7_postord(): parent = np.array([0, 4, 5, 0, 3, 4, 7, 8, 5, 6, 7, 8, 9, 14, 17, 12, 15, 16, 19, 16, 17]) po = post_order(parent, include_root=True) expect = np.array([12, 11, 19, 20, 21, 14, 15, 18, 17, 1...
[ "numpy.abs", "numpy.unique", "treelas.post_order", "treelas.TreeInstance", "numpy.array", "numpy.fromstring" ]
[((108, 193), 'numpy.array', 'np.array', (['[0, 4, 5, 0, 3, 4, 7, 8, 5, 6, 7, 8, 9, 14, 17, 12, 15, 16, 19, 16, 17]'], {}), '([0, 4, 5, 0, 3, 4, 7, 8, 5, 6, 7, 8, 9, 14, 17, 12, 15, 16, 19, 16,\n 17])\n', (116, 193), True, 'import numpy as np\n'), ((222, 259), 'treelas.post_order', 'post_order', (['parent'], {'inclu...
import argparse import os import torch import yaml DEFAULT_DEVICE = 'cuda:0' def load_config(): parser = argparse.ArgumentParser(description='UNet3D training') parser.add_argument('--config', type=str, help='Path to the YAML config file', required=True) args = parser.parse_args() config = _load_conf...
[ "torch.cuda.is_available", "argparse.ArgumentParser" ]
[((113, 167), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""UNet3D training"""'}), "(description='UNet3D training')\n", (136, 167), False, 'import argparse\n'), ((473, 498), 'torch.cuda.is_available', 'torch.cuda.is_available', ([], {}), '()\n', (496, 498), False, 'import torch\n')]
import tensorflow as tf from tensorflow import keras from tensorflow.keras import layers from load_cora import load_cora from baseline_model import create_ffn from utils import run_experiment from utils import display_learning_curves # Graph convolution layer class GraphConvLayer(layers.Layer): def __init__( ...
[ "tensorflow.expand_dims", "load_cora.load_cora", "tensorflow.unstack", "tensorflow.ones", "tensorflow.math.unsorted_segment_mean", "tensorflow.nn.l2_normalize", "utils.display_learning_curves", "utils.run_experiment", "tensorflow.concat", "tensorflow.math.unsorted_segment_sum", "tensorflow.gathe...
[((6709, 6729), 'load_cora.load_cora', 'load_cora', ([], {'verbose': '(1)'}), '(verbose=1)\n', (6718, 6729), False, 'from load_cora import load_cora\n'), ((7127, 7156), 'tensorflow.ones', 'tf.ones', ([], {'shape': 'edges.shape[1]'}), '(shape=edges.shape[1])\n', (7134, 7156), True, 'import tensorflow as tf\n'), ((7921, ...
from db import db class RisklayerPrognosis(db.Model): __tablename__ = 'risklayer_prognosis' datenbestand = db.Column(db.TIMESTAMP, primary_key=True, nullable=False) prognosis = db.Column(db.Float, nullable=False) # class RisklayerPrognosisSchema(SQLAlchemyAutoSchema): # class Meta: # strict ...
[ "db.db.Column" ]
[((118, 175), 'db.db.Column', 'db.Column', (['db.TIMESTAMP'], {'primary_key': '(True)', 'nullable': '(False)'}), '(db.TIMESTAMP, primary_key=True, nullable=False)\n', (127, 175), False, 'from db import db\n'), ((192, 227), 'db.db.Column', 'db.Column', (['db.Float'], {'nullable': '(False)'}), '(db.Float, nullable=False)...
# SPDX-FileCopyrightText: 2014 MicroPython & CircuitPython contributors (https://github.com/adafruit/circuitpython/graphs/contributors) # # SPDX-License-Identifier: MIT import argparse import os import sys sys.path.append("../../tools/usb_descriptor") from adafruit_usb_descriptor import audio, audio10, cdc, hid, mi...
[ "adafruit_usb_descriptor.cdc.CallManagement", "adafruit_usb_descriptor.midi.InJackDescriptor", "argparse.FileType", "adafruit_usb_descriptor.midi.OutJackDescriptor", "adafruit_usb_descriptor.midi.Header", "argparse.ArgumentParser", "adafruit_usb_descriptor.cdc.Header", "adafruit_usb_descriptor.cdc.Uni...
[((209, 254), 'sys.path.append', 'sys.path.append', (['"""../../tools/usb_descriptor"""'], {}), "('../../tools/usb_descriptor')\n", (224, 254), False, 'import sys\n'), ((821, 885), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Generate USB descriptors."""'}), "(description='Generate USB...
from tqdm import tqdm import pandas as pd import numpy as np, argparse, time, pickle, random, os, datetime import torch import torch.optim as optim from model import MaskedNLLLoss, BC_LSTM from dataloader import MELDDataLoader from sklearn.metrics import f1_score, confusion_matrix, accuracy_score, classification_re...
[ "sklearn.metrics.classification_report", "numpy.array", "torch.cuda.is_available", "argparse.ArgumentParser", "model.BC_LSTM", "dataloader.MELDDataLoader", "numpy.random.seed", "numpy.concatenate", "numpy.argmin", "torch.argmax", "numpy.argmax", "time.time", "sklearn.metrics.accuracy_score",...
[((427, 450), 'torch.manual_seed', 'torch.manual_seed', (['seed'], {}), '(seed)\n', (444, 450), False, 'import torch\n'), ((455, 483), 'torch.cuda.manual_seed', 'torch.cuda.manual_seed', (['seed'], {}), '(seed)\n', (477, 483), False, 'import torch\n'), ((488, 520), 'torch.cuda.manual_seed_all', 'torch.cuda.manual_seed_...
# noinspection PyUnresolvedReferences import os import re # TODO I'm going to need to make a dictionary for my big list of stuff i care about and what's needed for # every file type.... RAF = ['EXIF:LensModel', 'MakerNotes:RawImageHeight', 'MakerNotes:RawImageWidth', 'EXIF:CreateDate', 'EXIF:ModifyDate', 'EXI...
[ "re.split", "os.popen", "os.path.splitext" ]
[((3281, 3305), 'os.path.splitext', 'os.path.splitext', (['filein'], {}), '(filein)\n', (3297, 3305), False, 'import os\n'), ((1764, 1788), 'os.path.splitext', 'os.path.splitext', (['filein'], {}), '(filein)\n', (1780, 1788), False, 'import os\n'), ((2026, 2053), 're.split', 're.split', (['"""\\\\s+:\\\\s+"""', 'each']...
import logging from typing import List, Callable import numpy as np from pyquaternion import Quaternion from pyrep import PyRep from pyrep.errors import IKError from pyrep.objects import Dummy, Object from rlbench import utils from rlbench.action_modes import ArmActionMode, ActionMode from rlbench.backend.exceptions ...
[ "numpy.flip", "numpy.abs", "numpy.allclose", "numpy.random.get_state", "numpy.linalg.pinv", "numpy.minimum", "pyquaternion.Quaternion", "numpy.square", "pyrep.objects.Dummy.create", "numpy.array", "numpy.append", "numpy.dot", "numpy.matmul", "numpy.linalg.norm", "numpy.shape", "numpy.t...
[((1721, 1735), 'pyrep.objects.Dummy.create', 'Dummy.create', ([], {}), '()\n', (1733, 1735), False, 'from pyrep.objects import Dummy, Object\n'), ((8536, 8557), 'numpy.array', 'np.array', (['action[:-1]'], {}), '(action[:-1])\n', (8544, 8557), True, 'import numpy as np\n'), ((14997, 15012), 'numpy.transpose', 'np.tran...
from django import forms from django.contrib.contenttypes.forms import generic_inlineformset_factory from django.contrib.contenttypes.models import ContentType from django.db import models from django.test import TestCase from django.test.utils import isolate_apps from .models import ( Animal, ForProxyMode...
[ "django.db.models.PositiveIntegerField", "django.contrib.contenttypes.models.ContentType.objects.get_for_model", "django.test.utils.isolate_apps", "django.contrib.contenttypes.forms.generic_inlineformset_factory" ]
[((7814, 7847), 'django.test.utils.isolate_apps', 'isolate_apps', (['"""generic_relations"""'], {}), "('generic_relations')\n", (7826, 7847), False, 'from django.test.utils import isolate_apps\n'), ((692, 742), 'django.contrib.contenttypes.forms.generic_inlineformset_factory', 'generic_inlineformset_factory', (['Tagged...
#!/usr/bin/python # -*- coding: utf-8 -*- # Copyright (C) 2017, <NAME> <<EMAIL>> # vim: set ts=4 sts=4 sw=4 expandtab smartindent: # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without ...
[ "os.path.exists", "os.mkdir" ]
[((2249, 2272), 'os.path.exists', 'path.exists', (['"""jug-demo"""'], {}), "('jug-demo')\n", (2260, 2272), False, 'from os import path\n'), ((2350, 2370), 'os.mkdir', 'os.mkdir', (['"""jug-demo"""'], {}), "('jug-demo')\n", (2358, 2370), False, 'import os\n')]
#!/usr/bin/env python # ROS Libraries import actionlib from actionlib_msgs.msg import GoalStatus from control_msgs.msg import JointTrajectoryControllerState, FollowJointTrajectoryAction, FollowJointTrajectoryGoal from kuri_wandering_robot.msg import Power from wandering_behavior.msg import WanderAction, WanderGoal impo...
[ "rospy.logwarn", "rospy.init_node", "rospy.Rate", "threading.Lock", "wandering_behavior.msg.WanderGoal", "rospy.Duration.from_sec", "rospy.spin", "rospy.Subscriber", "sent_messages_database.SentMessagesDatabase.load", "actionlib.SimpleActionClient", "rospy.get_param", "rospy.Time.now", "rosp...
[((19163, 19202), 'rospy.init_node', 'rospy.init_node', (['"""kuri_wandering_robot"""'], {}), "('kuri_wandering_robot')\n", (19178, 19202), False, 'import rospy\n'), ((19257, 19269), 'rospy.spin', 'rospy.spin', ([], {}), '()\n', (19267, 19269), False, 'import rospy\n'), ((2419, 2451), 'rospy.get_param', 'rospy.get_para...
#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright (C) 2010 <NAME> <<EMAIL>> # Licensed under the GNU LGPL v2.1 - http://www.gnu.org/licenses/lgpl.html """ Automated tests for checking transformation algorithms (the models package). """ import logging import unittest import numpy as np from gensim.corpora...
[ "logging.basicConfig", "gensim.models.rpmodel.RpModel.load", "gensim.matutils.sparse2full", "numpy.allclose", "gensim.test.utils.get_tmpfile", "gensim.models.rpmodel.RpModel", "numpy.array", "numpy.random.seed", "gensim.test.utils.datapath", "unittest.main" ]
[((2218, 2314), 'logging.basicConfig', 'logging.basicConfig', ([], {'format': '"""%(asctime)s : %(levelname)s : %(message)s"""', 'level': 'logging.DEBUG'}), "(format='%(asctime)s : %(levelname)s : %(message)s',\n level=logging.DEBUG)\n", (2237, 2314), False, 'import logging\n'), ((2315, 2330), 'unittest.main', 'unit...
import tensorflow as tf from tensorflow import keras class CondGeneratorModel(keras.Model): def __init__(self): super(CondGeneratorModel, self).__init__() # Expand 7*7*128 features into a (7,7,128) tensor self.dense_1 = keras.layers.Dense(7*7*256) self.reshape_1 = keras.layers.Resh...
[ "tensorflow.keras.layers.Reshape", "tensorflow.keras.layers.Conv2DTranspose", "tensorflow.keras.layers.Conv2D", "tensorflow.keras.layers.Dropout", "tensorflow.keras.layers.LeakyReLU", "tensorflow.keras.layers.BatchNormalization", "tensorflow.keras.layers.Embedding", "tensorflow.math.multiply", "tens...
[((250, 281), 'tensorflow.keras.layers.Dense', 'keras.layers.Dense', (['(7 * 7 * 256)'], {}), '(7 * 7 * 256)\n', (268, 281), False, 'from tensorflow import keras\n'), ((303, 336), 'tensorflow.keras.layers.Reshape', 'keras.layers.Reshape', (['(7, 7, 256)'], {}), '((7, 7, 256))\n', (323, 336), False, 'from tensorflow imp...
# -*- coding: utf-8 -*- from selectable.decorators import login_required from maestros.models import TiposMedidasActuacion, TiposLimitesCriticos, TiposMedidasVigilancia, TiposTemperaturas, TiposFrecuencias, Zonas, Terceros, CatalogoEquipos, Personal, Consumibles, ParametrosAnalisis, Actividades, Etapas, Peligros, Tipos...
[ "selectable.registry.registry.register", "maestros_generales.models.Empresas.objects.filter" ]
[((1100, 1140), 'selectable.registry.registry.register', 'registry.register', (['TPActuacionPrevLookup'], {}), '(TPActuacionPrevLookup)\n', (1117, 1140), False, 'from selectable.registry import registry\n'), ((1686, 1726), 'selectable.registry.registry.register', 'registry.register', (['TPActuacionCorrLookup'], {}), '(...
"""Define commands for Python 2.7""" import argparse import traceback from . import util from .cmd import run from .cmd import extractpipenv def main(): """Main function""" print("This version is not supported! It has limitted analysis features") parser = argparse.ArgumentParser(description='Analyze Jupyt...
[ "traceback.print_exc", "argparse.ArgumentParser" ]
[((270, 334), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Analyze Jupyter Notebooks"""'}), "(description='Analyze Jupyter Notebooks')\n", (293, 334), False, 'import argparse\n'), ((778, 799), 'traceback.print_exc', 'traceback.print_exc', ([], {}), '()\n', (797, 799), False, 'import tr...
# -*- coding: utf-8 -*- """This python module aims to manage `DokuWiki <https://www.dokuwiki.org/dokuwiki>`_ wikis by using the provided `XML-RPC API <https://www.dokuwiki.org/devel:xmlrpc>`_. It is compatible with python2.7 and python3+. Installation ------------ It is on `PyPi <https://pypi.python.org/pypi/dokuwik...
[ "os.path.exists", "collections.OrderedDict", "xmlrpclib.Binary", "os.makedirs", "datetime.datetime.utcnow", "datetime.datetime.strptime", "re.compile", "base64.b64encode", "os.path.join", "xmlrpclib.Transport.parse_response", "base64.b64decode", "datetime.datetime.now", "xmlrpclib.Transport....
[((945, 1007), 're.compile', 're.compile', (['"""(?P<proto>https?)://(?P<host>[^/]*)(?P<uri>/.*)?"""'], {}), "('(?P<proto>https?)://(?P<host>[^/]*)(?P<uri>/.*)?')\n", (955, 1007), False, 'import re\n'), ((1261, 1310), 'datetime.datetime.strptime', 'datetime.strptime', (['date[:-5]', '"""%Y-%m-%dT%H:%M:%S"""'], {}), "(d...
import setuptools import re with open("README.md", "r") as fh: long_description = fh.read() # get version from _version.py file, from below # https://stackoverflow.com/questions/458550/standard-way-to-embed-version-into-python-package VERSION_FILE = "test_aide/_version.py" version_file_str = open(VERSION_FILE, "r...
[ "setuptools.find_packages", "re.search" ]
[((391, 440), 're.search', 're.search', (['VERSION_STR_RE', 'version_file_str', 're.M'], {}), '(VERSION_STR_RE, version_file_str, re.M)\n', (400, 440), False, 'import re\n'), ((977, 1003), 'setuptools.find_packages', 'setuptools.find_packages', ([], {}), '()\n', (1001, 1003), False, 'import setuptools\n')]
#! /usr/bin/env python3 import importlib import logging import os import subprocess from setuptools import setup from setuptools.command.install import install as install from setuptools.command.develop import develop as develop logger = logging.getLogger(__name__) stan_model_files = [ os.path.join("nonperiod...
[ "logging.getLogger", "setuptools.command.install.install.run", "setuptools.command.install.install.finalize_options", "logging.info", "pbio.misc.shell_utils.check_programs_exist", "os.path.exists", "subprocess.call", "shlex.quote", "setuptools.setup", "logging.warning", "os.path.dirname", "pbi...
[((242, 269), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (259, 269), False, 'import logging\n'), ((4350, 4416), 'setuptools.setup', 'setup', ([], {'cmdclass': "{'install': SetupInstall, 'develop': SetupDevelop}"}), "(cmdclass={'install': SetupInstall, 'develop': SetupDevelop})\n", (43...
import os import h5py import nibabel as nb import numpy as np import torch import torch.utils.data as data from torchvision import transforms import utils.preprocessor as preprocessor # transform_train = transforms.Compose([ # transforms.RandomCrop(200, padding=56), # transforms.ToTensor(), # ]) class Imdb...
[ "os.listdir", "nibabel.load", "os.path.join", "torch.from_numpy", "utils.preprocessor.estimate_weights_mfb", "utils.preprocessor.remap_labels", "numpy.max", "utils.preprocessor.rotate_orientation", "numpy.min", "utils.preprocessor.reduce_slices", "utils.preprocessor.remove_black", "numpy.round...
[((4062, 4083), 'nibabel.load', 'nb.load', (['file_path[0]'], {}), '(file_path[0])\n', (4069, 4083), True, 'import nibabel as nb\n'), ((4766, 4828), 'utils.preprocessor.rotate_orientation', 'preprocessor.rotate_orientation', (['volume', 'labelmap', 'orientation'], {}), '(volume, labelmap, orientation)\n', (4797, 4828),...
# -*- coding: utf-8 -*- # Form implementation generated from reading ui file 'design.ui' # # Created by: PyQt5 UI code generator 5.15.4 # # WARNING: Any manual changes made to this file will be lost when pyuic5 is # run again. Do not edit this file unless you know what you are doing. from PyQt5 import QtC...
[ "PyQt5.QtWidgets.QWidget", "PyQt5.QtWidgets.QPlainTextEdit", "PyQt5.QtCore.QMetaObject.connectSlotsByName", "PyQt5.QtCore.QRect", "PyQt5.QtWidgets.QLabel", "PyQt5.QtWidgets.QPushButton", "PyQt5.QtWidgets.QCheckBox", "PyQt5.QtWidgets.QLineEdit" ]
[((527, 556), 'PyQt5.QtWidgets.QWidget', 'QtWidgets.QWidget', (['MainWindow'], {}), '(MainWindow)\n', (544, 556), False, 'from PyQt5 import QtCore, QtGui, QtWidgets\n'), ((638, 674), 'PyQt5.QtWidgets.QLabel', 'QtWidgets.QLabel', (['self.centralwidget'], {}), '(self.centralwidget)\n', (654, 674), False, 'from PyQt5 impo...
from simulation.car import spawn_drivers from simulation.passenger import spawn_passengers from simulation.core import World, Clock conf = { "x": 100, "y": 100, "drivers": 200, "users": 1000, "start": "2019-07-08T00:00:00", "end": "2019-07-08T00:01:00" } clock = Clock(conf["start"], conf["end"...
[ "simulation.passenger.spawn_passengers", "simulation.car.spawn_drivers", "simulation.core.World", "simulation.core.Clock" ]
[((289, 322), 'simulation.core.Clock', 'Clock', (["conf['start']", "conf['end']"], {}), "(conf['start'], conf['end'])\n", (294, 322), False, 'from simulation.core import World, Clock\n'), ((363, 405), 'simulation.core.World', 'World', (["[conf['x'], conf['y']]"], {'clock': 'clock'}), "([conf['x'], conf['y']], clock=clo...
import json d1 = {} with open("/home/qinyuan/zs/out/bart-large-with-description-grouped-1e-5-outerbsz4-innerbsz32-adapterdim4-unfreeze-dec29/test_predictions.jsonl") as fin: for line in fin: d = json.loads(line) d1[d["id"]] = d["output"][0]["answer"] d2 = {} dq = {} with open("/home/qinyuan/zs/out...
[ "json.loads" ]
[((208, 224), 'json.loads', 'json.loads', (['line'], {}), '(line)\n', (218, 224), False, 'import json\n'), ((450, 466), 'json.loads', 'json.loads', (['line'], {}), '(line)\n', (460, 466), False, 'import json\n'), ((663, 679), 'json.loads', 'json.loads', (['line'], {}), '(line)\n', (673, 679), False, 'import json\n')]