code
stringlengths
31
1.05M
apis
list
extract_api
stringlengths
97
1.91M
import numpy as np X = 2 * np.random.randn(100, 5) y = 2.5382 * np.cos(X[:, 3]) + X[:, 0] ** 2 - 0.5 from pysr import PySRRegressor model = PySRRegressor( niterations=40, binary_operators=["+", "*"], unary_operators=[ "cos", "exp", "sin", "inv(x) = 1/x", # Custom operator...
[ "pysr.PySRRegressor", "numpy.random.randn", "numpy.cos" ]
[((143, 321), 'pysr.PySRRegressor', 'PySRRegressor', ([], {'niterations': '(40)', 'binary_operators': "['+', '*']", 'unary_operators': "['cos', 'exp', 'sin', 'inv(x) = 1/x']", 'model_selection': '"""best"""', 'loss': '"""loss(x, y) = (x - y)^2"""'}), "(niterations=40, binary_operators=['+', '*'], unary_operators=\n ...
import json import os import cv2 import numpy as np from dgp.datasets.synchronized_dataset import SynchronizedScene from dgp.utils.visualization_engine import visualize_dataset_3d, visualize_dataset_2d, visualize_dataset_sample_3d, visualize_dataset_sample_2d from tests import TEST_DATA_DIR def dummy_caption(datase...
[ "os.path.exists", "numpy.allclose", "dgp.utils.visualization_engine.visualize_dataset_sample_3d", "dgp.utils.visualization_engine.visualize_dataset_2d", "os.path.join", "dgp.utils.visualization_engine.visualize_dataset_3d", "dgp.datasets.synchronized_dataset.SynchronizedScene", "cv2.imread", "os.rem...
[((498, 611), 'os.path.join', 'os.path.join', (['TEST_DATA_DIR', '"""dgp"""', '"""test_scene/scene_01/scene_a8dc5ed1da0923563f85ea129f0e0a83e7fe1867.json"""'], {}), "(TEST_DATA_DIR, 'dgp',\n 'test_scene/scene_01/scene_a8dc5ed1da0923563f85ea129f0e0a83e7fe1867.json')\n", (510, 611), False, 'import os\n'), ((671, 871),...
import numpy as np import serial import struct import threading import time from array import array from datetime import datetime class ImuData: def __init__(self, t=0.0, freq=0, ypr=np.zeros(3), a=np.zeros(3), \ W=np.zeros(3)): self.t = t self.freq = freq self.ypr = ypr ...
[ "threading.Thread.__init__", "array.array", "threading.Lock", "time.sleep", "numpy.array", "numpy.zeros", "struct.unpack", "serial.Serial", "datetime.datetime.now" ]
[((191, 202), 'numpy.zeros', 'np.zeros', (['(3)'], {}), '(3)\n', (199, 202), True, 'import numpy as np\n'), ((206, 217), 'numpy.zeros', 'np.zeros', (['(3)'], {}), '(3)\n', (214, 217), True, 'import numpy as np\n'), ((235, 246), 'numpy.zeros', 'np.zeros', (['(3)'], {}), '(3)\n', (243, 246), True, 'import numpy as np\n')...
''' objective :- ------------ detect and classify shapes and their location in an image with low latency and high accuracy. it must account for false positives and empty images. modules used :- --------------- 1 - open cv for image processing tasks. 2 - easyocr for text-recognition tasks. 3 - threading for running tex...
[ "cv2.cv2.getRotationMatrix2D", "cv2.cv2.warpAffine", "cv2.cv2.imread", "time.perf_counter", "AlphanumericCharacterDetection.recogniser.Recognize", "numpy.array" ]
[((2094, 2143), 'cv2.cv2.getRotationMatrix2D', 'cv2.getRotationMatrix2D', (['image_center', 'angle', '(1.0)'], {}), '(image_center, angle, 1.0)\n', (2117, 2143), False, 'from cv2 import cv2\n'), ((2154, 2228), 'cv2.cv2.warpAffine', 'cv2.warpAffine', (['image', 'rot_mat', 'image.shape[1::-1]'], {'flags': 'cv2.INTER_LINE...
""" Custom added maze tasks with dense rewards and progressively farther goals For creating expert demonstrations """ from typing import Dict, List, Type, Tuple import numpy as np from mujoco_maze.custom_maze_task import ( GoalRewardLargeUMaze, GoalRewardRoom3x5, GoalRewardRoom3x10, ) from mujoco_maze.t...
[ "mujoco_maze.task_common.euc_dist", "mujoco_maze.task_common.RewardThresholdList", "numpy.argmax", "numpy.sum", "numpy.array" ]
[((4318, 4357), 'mujoco_maze.task_common.RewardThresholdList', 'RewardThresholdList', (['[-70]', '[-70]', 'None'], {}), '([-70], [-70], None)\n', (4337, 4357), False, 'from mujoco_maze.task_common import MazeGoal, MazeTask, GREEN, euc_dist, RewardThresholdList\n'), ((4397, 4455), 'mujoco_maze.task_common.RewardThreshol...
from functools import lru_cache import math import logging from enum import Enum from typing import Optional, List, Tuple, Any, Union, Dict, Callable from concurrent.futures import ThreadPoolExecutor from io import BytesIO import requests import numpy as np from google.api_core import retry from PIL import Image from ...
[ "logging.getLogger", "pygeotile.point.Point.from_latitude_longitude", "pygeotile.point.Point.from_pixel", "math.tan", "pydantic.validator", "numpy.hstack", "concurrent.futures.ThreadPoolExecutor", "pygeotile.point.Point.from_meters", "io.BytesIO", "math.radians", "numpy.max", "numpy.array", ...
[((765, 792), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (782, 792), False, 'import logging\n'), ((1532, 1551), 'pydantic.validator', 'validator', (['"""bounds"""'], {}), "('bounds')\n", (1541, 1551), False, 'from pydantic import BaseModel, validator\n'), ((3019, 3035), 'pydantic.vali...
import numpy import time import threading import matplotlib.pyplot from matplotlib.animation import FuncAnimation class VisualizationWindow: def __init__(self, signal_collector): self.figure, self.axes = matplotlib.pyplot.subplots(7, 1, sharex=True) self.figure.subplots_adjust(hspace=0) ...
[ "matplotlib.animation.FuncAnimation", "numpy.array", "time.time" ]
[((3421, 3432), 'time.time', 'time.time', ([], {}), '()\n', (3430, 3432), False, 'import time\n'), ((3562, 3634), 'matplotlib.animation.FuncAnimation', 'FuncAnimation', (['self.figure', 'self.update_plots'], {'interval': '(1000)', 'blit': '(False)'}), '(self.figure, self.update_plots, interval=1000, blit=False)\n', (35...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- import numpy as np import matplotlib.pyplot as plt import h5py as h5 import os.path as fs import keras from keras.models import Sequential from keras.utils import np_utils from keras.layers.core import Dense, Activation, Dropout from sklearn.model_selection import trai...
[ "sklearn.metrics.f1_score", "keras.layers.core.Activation", "sklearn.decomposition.PCA", "os.path.join", "numpy.heaviside", "keras.models.Sequential", "keras.optimizers.SGD", "keras.utils.np_utils.to_categorical", "numpy.empty", "umap.UMAP", "sklearn.preprocessing.scale", "keras.layers.core.De...
[((1697, 1729), 'numpy.empty', 'np.empty', (['(n, 1)'], {'dtype': 'np.int32'}), '((n, 1), dtype=np.int32)\n', (1705, 1729), True, 'import numpy as np\n'), ((1871, 1902), 'keras.utils.np_utils.to_categorical', 'np_utils.to_categorical', (['labels'], {}), '(labels)\n', (1894, 1902), False, 'from keras.utils import np_uti...
import os import cv2 import numpy as np import tensorflow as tf import tensorflow.compat.v1 as tf from PIL import Image from PIL import ImageFont from PIL import ImageDraw from flask_ngrok import run_with_ngrok from flask import Flask,request,send_from_directory,render_template # GLOBAL ACCESS os.environ['TF_CPP_MIN_L...
[ "flask.render_template", "flask.Flask", "PIL.Image.new", "numpy.array", "PIL.ImageDraw.Draw", "tensorflow.compat.v1.get_collection", "tensorflow.compat.v1.get_default_graph", "flask.send_from_directory", "tensorflow.compat.v1.nn.ctc_greedy_decoder", "numpy.asarray", "PIL.ImageFont.truetype", "...
[((414, 442), 'tensorflow.compat.v1.disable_eager_execution', 'tf.disable_eager_execution', ([], {}), '()\n', (440, 442), True, 'import tensorflow.compat.v1 as tf\n'), ((449, 484), 'flask.Flask', 'Flask', (['__name__'], {'static_url_path': '""""""'}), "(__name__, static_url_path='')\n", (454, 484), False, 'from flask i...
import time,os,math,inspect,re,sys,random,argparse from env import SenseEnv from torch.autograd import Variable import numpy as np from itertools import count from collections import namedtuple from tensorboardX import SummaryWriter import torch import torch.nn as nn import torch.nn.functional as F import torch.optim a...
[ "torch.nn.ReLU", "torch.nn.CrossEntropyLoss", "numpy.random.rand", "torch.max", "torch.from_numpy", "torch.nn.MSELoss", "torch.cuda.is_available", "torch.nn.functional.softmax", "os.path.exists", "torch.nn.BatchNorm2d", "tensorboardX.SummaryWriter", "argparse.ArgumentParser", "numpy.asarray"...
[((408, 423), 'tensorboardX.SummaryWriter', 'SummaryWriter', ([], {}), '()\n', (421, 423), False, 'from tensorboardX import SummaryWriter\n'), ((438, 484), 'collections.namedtuple', 'namedtuple', (['"""SavedAction"""', "['action', 'value']"], {}), "('SavedAction', ['action', 'value'])\n", (448, 484), False, 'from colle...
import numpy as np from sklearn import svm class character: def __init__(self, raw_character): self.identity = None def recognize(characters, classifier): for char in characters: data = np.reshape(char.image_centered, np.prod(char.image_centered.shape)).reshape(1, -1) char.ide...
[ "numpy.prod" ]
[((253, 287), 'numpy.prod', 'np.prod', (['char.image_centered.shape'], {}), '(char.image_centered.shape)\n', (260, 287), True, 'import numpy as np\n')]
# -*- coding: utf-8 -*- from skimage.viewer import utils from skimage.viewer.utils import dialogs from skimage.viewer.qt import QtCore, QtGui, has_qt from numpy.testing.decorators import skipif @skipif(not has_qt) def test_event_loop(): utils.init_qtapp() timer = QtCore.QTimer() timer.singleShot(10, QtGui...
[ "skimage.viewer.utils.init_qtapp", "numpy.testing.decorators.skipif", "skimage.viewer.utils.dialogs._format_filename", "skimage.viewer.utils.dialogs.open_file_dialog", "skimage.viewer.qt.QtCore.QTimer", "skimage.viewer.utils.dialogs.save_file_dialog", "skimage.viewer.utils.start_qtapp", "skimage.viewe...
[((197, 215), 'numpy.testing.decorators.skipif', 'skipif', (['(not has_qt)'], {}), '(not has_qt)\n', (203, 215), False, 'from numpy.testing.decorators import skipif\n'), ((367, 385), 'numpy.testing.decorators.skipif', 'skipif', (['(not has_qt)'], {}), '(not has_qt)\n', (373, 385), False, 'from numpy.testing.decorators ...
from aoi_envs.MultiAgent import MultiAgentEnv import numpy as np class MobileEnv(MultiAgentEnv): def __init__(self, agent_velocity=1.0, initialization='Random', biased_velocities=False, flocking=False, random_acceleration=True, aoi_reward=True, flocking_position_control=False, num_agents=40): ...
[ "numpy.clip", "numpy.copy", "numpy.random.normal", "numpy.multiply", "numpy.where", "numpy.nanmean", "numpy.sum", "numpy.cos", "numpy.random.uniform", "numpy.sin", "numpy.nansum", "numpy.divide" ]
[((1140, 1226), 'numpy.random.uniform', 'np.random.uniform', (['(-self.max_velocity)', 'self.max_velocity'], {'size': '(self.n_agents, 2)'}), '(-self.max_velocity, self.max_velocity, size=(self.\n n_agents, 2))\n', (1157, 1226), True, 'import numpy as np\n'), ((3705, 3773), 'numpy.clip', 'np.clip', (['acceleration',...
from torch.autograd import Variable import torch.nn.functional as F import scripts.utils as utils import torch.nn as nn import numpy as np import torch class CrossEntropy2d(nn.Module): def __init__(self, size_average=True, ignore_label=255): super(CrossEntropy2d, self).__init__() self.size_average...
[ "torch.sort", "torch.nn.functional.nll_loss", "numpy.put", "torch.eye", "torch.unsqueeze", "torch.LongTensor", "scripts.utils.mean", "torch.pow", "torch.from_numpy", "torch.randn", "numpy.array", "numpy.zeros", "torch.nn.functional.cross_entropy", "torch.nn.functional.log_softmax", "torc...
[((1750, 1777), 'torch.nn.functional.log_softmax', 'F.log_softmax', (['input'], {'dim': '(1)'}), '(input, dim=1)\n', (1763, 1777), True, 'import torch.nn.functional as F\n'), ((2124, 2202), 'torch.nn.functional.nll_loss', 'F.nll_loss', (['log_p', 'target'], {'ignore_index': '(250)', 'weight': 'weight', 'size_average': ...
import logging import tflite import numpy as np from tflite2onnx import mapping from tflite2onnx.op.common import Operator from tflite2onnx.op.binary import PowerWrapper logger = logging.getLogger('tflite2onnx') class Rsqrt(Operator): # use square root as input operator and propagate output to power TypeMap...
[ "logging.getLogger", "tflite2onnx.op.binary.PowerWrapper", "numpy.full" ]
[((181, 213), 'logging.getLogger', 'logging.getLogger', (['"""tflite2onnx"""'], {}), "('tflite2onnx')\n", (198, 213), False, 'import logging\n'), ((1405, 1436), 'tflite2onnx.op.binary.PowerWrapper', 'PowerWrapper', (['self.TFactory', '(-1)'], {}), '(self.TFactory, -1)\n', (1417, 1436), False, 'from tflite2onnx.op.binar...
""" ******************************************************************************** main file to execute ******************************************************************************** """ import time import numpy as np import tensorflow as tf import matplotlib.pyplot as plt from pinn import PINN from config_gpu im...
[ "matplotlib.pyplot.grid", "numpy.sqrt", "matplotlib.pyplot.ylabel", "numpy.array", "numpy.linalg.norm", "fdm.FDM", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.plot", "numpy.linspace", "numpy.empty", "params.params", "numpy.meshgrid", "matplotlib.pyplot.yscale", "tensorflow.device", "p...
[((473, 494), 'config_gpu.config_gpu', 'config_gpu', ([], {'gpu_flg': '(1)'}), '(gpu_flg=1)\n', (483, 494), False, 'from config_gpu import config_gpu\n'), ((689, 697), 'params.params', 'params', ([], {}), '()\n', (695, 697), False, 'from params import params\n'), ((856, 883), 'numpy.linspace', 'np.linspace', (['tmin', ...
import os, sys import numpy as np import pandas as pd import matplotlib.pyplot as plt import skimage.io from skimage.transform import resize from imgaug import augmenters as iaa from random import randint import PIL from PIL import Image import cv2 from sklearn.utils import class_weight, shuffle import keras import wa...
[ "keras.layers.Conv2D", "tensorflow.equal", "pandas.read_csv", "imgaug.augmenters.GaussianBlur", "keras.backend.floatx", "numpy.array", "tensorflow.is_nan", "keras.layers.Dense", "tensorflow.ones_like", "imgaug.augmenters.Fliplr", "numpy.arange", "numpy.divide", "imgaug.augmenters.Flipud", ...
[((384, 417), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {}), "('ignore')\n", (407, 417), False, 'import warnings\n'), ((8892, 8931), 'numpy.arange', 'np.arange', (['paths.shape[0]'], {'dtype': 'np.int'}), '(paths.shape[0], dtype=np.int)\n', (8901, 8931), True, 'import numpy as np\n'), ((8...
import numpy as np from . import utils def tile_position(x0, y0, x1=None, y1=None): """Need doc string...""" if x1 is None and y1 is None: x1 = x0 y1 = y0 if (x0.size != y0.size) or (x1.size != y1.size): raise ValueError("x0 and y0 or x1 and y1 size do not match.") x0g = np...
[ "numpy.sqrt" ]
[((1225, 1251), 'numpy.sqrt', 'np.sqrt', (['(dx ** 2 + dy ** 2)'], {}), '(dx ** 2 + dy ** 2)\n', (1232, 1251), True, 'import numpy as np\n')]
#!/usr/bin/env python # coding: utf-8 import sys sys.path.append("../") import pandas as pd import numpy as np import pathlib import pickle import os import itertools import argparse import logging import helpers.feature_helpers as fh from collections import Counter OUTPUT_DF_TR = 'df_steps_tr.csv' OUTPUT_DF_VAL =...
[ "logging.getLogger", "logging.basicConfig", "numpy.mean", "pickle.dump", "argparse.ArgumentParser", "pandas.read_csv", "numpy.max", "itertools.chain.from_iterable", "numpy.array", "pandas.concat", "numpy.std", "pandas.notnull", "sys.path.append" ]
[((50, 72), 'sys.path.append', 'sys.path.append', (['"""../"""'], {}), "('../')\n", (65, 72), False, 'import sys\n'), ((785, 842), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Create cv features"""'}), "(description='Create cv features')\n", (808, 842), False, 'import argparse\n'), ((1...
import copy import importlib import os import numpy as np import tensorflow as tf import logging tf.get_logger().setLevel(logging.ERROR) from client import Client from server import Server from model import ServerModel from baseline_constants import MAIN_PARAMS, MODEL_PARAMS from fedbayes_helper import * from fedbaye...
[ "os.path.exists", "tensorflow.reset_default_graph", "importlib.import_module", "numpy.ones", "numpy.average", "tensorflow.logging.set_verbosity", "metrics.writer.get_metrics_names", "server.Server", "numpy.array", "numpy.zeros", "utils.matching.cnn_pfnm.layerwise_sampler", "utils.matching.cnn_...
[((836, 877), 'metrics.writer.get_metrics_names', 'metrics_writer.get_metrics_names', (['metrics'], {}), '(metrics)\n', (868, 877), True, 'import metrics.writer as metrics_writer\n'), ((98, 113), 'tensorflow.get_logger', 'tf.get_logger', ([], {}), '()\n', (111, 113), True, 'import tensorflow as tf\n'), ((3894, 3929), '...
import numpy as np import pyverilator import os from . import to_float, to_fix_point_int import taichi as ti from .all_python_functions import calc_next_pos_and_velocity, rectify_positions_and_velocities, \ rectify_positions_in_collision, calc_after_collision_velocity, two_ball_collides, normalize_vector def rect...
[ "numpy.allclose", "pyverilator.PyVerilator.build", "numpy.random.rand", "numpy.ones", "taichi.init", "os.chdir", "numpy.array", "numpy.zeros", "taichi.GUI" ]
[((978, 1001), 'os.chdir', 'os.chdir', (['"""./pyverilog"""'], {}), "('./pyverilog')\n", (986, 1001), False, 'import os\n'), ((1012, 1069), 'pyverilator.PyVerilator.build', 'pyverilator.PyVerilator.build', (['"""rectify_p_in_collision.v"""'], {}), "('rectify_p_in_collision.v')\n", (1041, 1069), False, 'import pyverilat...
import unittest from operator import attrgetter from typing import Dict import numpy as np from PIL import Image from _pytest._code import ExceptionInfo from lunavl.sdk.errors.errors import ErrorInfo from lunavl.sdk.faceengine.engine import VLFaceEngine from lunavl.sdk.image_utils.geometry import Rect from lunavl.sdk...
[ "operator.attrgetter", "PIL.Image.open", "lunavl.sdk.faceengine.engine.VLFaceEngine", "numpy.array", "numpy.ndarray", "lunavl.sdk.image_utils.image.VLImage.fromNumpyArray" ]
[((463, 477), 'lunavl.sdk.faceengine.engine.VLFaceEngine', 'VLFaceEngine', ([], {}), '()\n', (475, 477), False, 'from lunavl.sdk.faceengine.engine import VLFaceEngine\n'), ((2576, 2596), 'PIL.Image.open', 'Image.open', (['ONE_FACE'], {}), '(ONE_FACE)\n', (2586, 2596), False, 'from PIL import Image\n'), ((2645, 2680), '...
from pygame import init, display, time, event, draw, QUIT from numpy import arange def grid(janela, comprimento, tamanho_linha, tamanho_quadrado): def draw_grid(v): draw.line(janela, (255, 255, 255), (v * tamanho_quadrado, 0), (v * tamanho_quadrado, comprimento)) ...
[ "pygame.display.set_caption", "pygame.init", "pygame.draw.line", "pygame.event.get", "pygame.display.set_mode", "pygame.display.flip", "pygame.time.Clock", "numpy.arange" ]
[((1554, 1560), 'pygame.init', 'init', ([], {}), '()\n', (1558, 1560), False, 'from pygame import init, display, time, event, draw, QUIT\n'), ((1594, 1622), 'pygame.display.set_mode', 'display.set_mode', (['tela_cheia'], {}), '(tela_cheia)\n', (1610, 1622), False, 'from pygame import init, display, time, event, draw, Q...
from operator import truediv import cv2 from time import sleep import HandTrackingModule as htm import os import autopy import numpy as np import math import mediapipe as mp #import modules #variables frameR=20 #frame rduction frameR_x=800 frameR_y=110 wCam,hCam=1300 ,400 pTime=0 smoothening = 5 #need to tune plocX, p...
[ "cv2.rectangle", "autopy.mouse.click", "autopy.screen.size", "HandTrackingModule.handDetector", "time.sleep", "cv2.imshow", "cv2.circle", "cv2.VideoCapture", "numpy.interp", "autopy.mouse.move", "cv2.waitKey" ]
[((360, 379), 'cv2.VideoCapture', 'cv2.VideoCapture', (['(0)'], {}), '(0)\n', (376, 379), False, 'import cv2\n'), ((423, 451), 'HandTrackingModule.handDetector', 'htm.handDetector', ([], {'maxHands': '(1)'}), '(maxHands=1)\n', (439, 451), True, 'import HandTrackingModule as htm\n'), ((463, 483), 'autopy.screen.size', '...
from abc import ABC, abstractmethod import numpy as np import matplotlib.pyplot as plt from heatlib.units import Time from heatlib.boundary_conditions import Boundary_Condition from heatlib.domains import Domain_Constant_1D, Domain_Variable_1D from heatlib.solvers import Solver_1D ####################################...
[ "heatlib.units.Time", "matplotlib.pyplot.subplots", "numpy.interp", "matplotlib.pyplot.show" ]
[((2137, 2171), 'matplotlib.pyplot.subplots', 'plt.subplots', ([], {'figsize': 'self.figsize'}), '(figsize=self.figsize)\n', (2149, 2171), True, 'import matplotlib.pyplot as plt\n'), ((2840, 2850), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (2848, 2850), True, 'import matplotlib.pyplot as plt\n'), ((3671, ...
# BSD 3-Clause License # # Copyright (c) 2016-21, University of Liverpool # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # # * Redistributions of source code must retain the above copyright notic...
[ "numpy.load", "conkit.core.distancefile.DistanceFile", "conkit.core.distogram.Distogram" ]
[((2857, 2875), 'conkit.core.distancefile.DistanceFile', 'DistanceFile', (['f_id'], {}), '(f_id)\n', (2869, 2875), False, 'from conkit.core.distancefile import DistanceFile\n'), ((2946, 2970), 'conkit.core.distogram.Distogram', 'Distogram', (['"""distogram_1"""'], {}), "('distogram_1')\n", (2955, 2970), False, 'from co...
from cv2 import cv2 from collections import Counter from PIL import Image, ImageDraw, ImageFont from scipy.fftpack import dct from sklearn.cluster import KMeans import matplotlib.pyplot as plt import gmpy2 import numpy as np import time import os """ Transparency If putting the pixel with RGBA = (Ra, Ga, Ba, Aa) over...
[ "sklearn.cluster.KMeans", "numpy.mean", "PIL.Image.fromarray", "numpy.median", "os.listdir", "numpy.copy", "PIL.Image.new", "os.path.join", "PIL.ImageFont.truetype", "numpy.asarray", "numpy.array", "numpy.zeros", "PIL.ImageDraw.Draw", "scipy.fftpack.dct", "os.path.dirname", "numpy.argm...
[((540, 574), 'os.path.join', 'os.path.join', (['"""icons"""', '"""generated"""'], {}), "('icons', 'generated')\n", (552, 574), False, 'import os\n'), ((626, 655), 'numpy.array', 'np.array', (['img'], {'dtype': 'np.uint8'}), '(img, dtype=np.uint8)\n', (634, 655), True, 'import numpy as np\n'), ((666, 686), 'PIL.Image.f...
# ====================================================================== # Copyright CERFACS (October 2018) # Contributor: <NAME> (<EMAIL>) # # This software is governed by the CeCILL-B license under French law and # abiding by the rules of distribution of free software. You can use, # modify and/or redistribute ...
[ "numpy.random.choice", "qtoolkit.data_structures.quantum_circuit.quantum_circuit.QuantumCircuit", "qtoolkit.maths.matrix.distances.gloa_objective_function", "numpy.zeros", "numpy.random.randint", "numpy.argmin", "qtoolkit.maths.matrix.generation.quantum_circuit.generate_random_quantum_circuit" ]
[((5397, 5433), 'numpy.zeros', 'numpy.zeros', (['(p,)'], {'dtype': 'numpy.float'}), '((p,), dtype=numpy.float)\n', (5408, 5433), False, 'import numpy\n'), ((6224, 6249), 'numpy.argmin', 'numpy.argmin', (['self._costs'], {}), '(self._costs)\n', (6236, 6249), False, 'import numpy\n'), ((8461, 8529), 'numpy.random.choice'...
from optparse import OptionParser import os import sys import time import numpy as np import pandas as pd import tensorflow as tf import utils import get_site_features import tf_utils np.set_printoptions(threshold=np.inf, linewidth=200) pd.options.mode.chained_assignment = None if __name__ == '__main__': pars...
[ "tf_utils._float_feature", "pandas.read_csv", "tf_utils._int64_feature", "optparse.OptionParser", "get_site_features.get_sites_from_utr", "utils.rev_comp", "tensorflow.train.FeatureList", "utils.one_hot_encode", "tensorflow.python_io.TFRecordWriter", "tensorflow.train.SequenceExample", "numpy.se...
[((187, 239), 'numpy.set_printoptions', 'np.set_printoptions', ([], {'threshold': 'np.inf', 'linewidth': '(200)'}), '(threshold=np.inf, linewidth=200)\n', (206, 239), True, 'import numpy as np\n'), ((325, 339), 'optparse.OptionParser', 'OptionParser', ([], {}), '()\n', (337, 339), False, 'from optparse import OptionPar...
""" Feature Selection Test 3 Random Forest, heatmap """ import matplotlib.pyplot as plt from mlxtend.plotting import scatterplotmatrix from sklearn.ensemble import RandomForestClassifier from sklearn.feature_selection import SelectFromModel from sklearn.model_selection import train_test_split from mlxtend.plotting im...
[ "numpy.mean", "matplotlib.pyplot.savefig", "pandas.read_csv", "sklearn.feature_selection.SelectFromModel", "numpy.corrcoef", "sklearn.ensemble.RandomForestClassifier", "sklearn.preprocessing.StandardScaler", "numpy.argsort", "matplotlib.pyplot.tight_layout", "matplotlib.pyplot.title", "matplotli...
[((442, 475), 'pandas.read_csv', 'pd.read_csv', (['"""NewBioDegWCols.csv"""'], {}), "('NewBioDegWCols.csv')\n", (453, 475), True, 'import pandas as pd\n'), ((1183, 1199), 'sklearn.preprocessing.StandardScaler', 'StandardScaler', ([], {}), '()\n', (1197, 1199), False, 'from sklearn.preprocessing import StandardScaler\n'...
import pandas as pd import matplotlib.pyplot as plt import matplotlib.style as style import numpy as np import os style.use('ggplot') grid_list = ['grid.168010.e', 'grid.1032.0', 'grid.7177.6', 'grid.194645.b', 'grid.6571.5'] dirname = os.getcwd() dirname = dirname + '/Data/' df_ARWU2018 = pd.read_csv(dirname + 'AR...
[ "pandas.read_csv", "numpy.absolute", "os.getcwd", "matplotlib.style.use", "matplotlib.pyplot.subplots", "matplotlib.pyplot.show" ]
[((115, 134), 'matplotlib.style.use', 'style.use', (['"""ggplot"""'], {}), "('ggplot')\n", (124, 134), True, 'import matplotlib.style as style\n'), ((239, 250), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (248, 250), False, 'import os\n'), ((295, 350), 'pandas.read_csv', 'pd.read_csv', (["(dirname + 'ARWU/ARWURanking_2...
#!/bin/python3 # encoding: utf-8 import sys import numpy as np from time import time ''' x [0, 2] => idx start 0, end 3 [3, 5] => idx start 3, end 6 [6, 8] => idx start 6, end 9 ((0 + (r_idx // 3 * 3)): (3 + (r_idx // 3 * 3)), (0 + (c_idx // 3 * 3)): (3 + (c_idx // 3 * 3))) np.random.randint(1, 10) ''' sys.setrecu...
[ "sys.setrecursionlimit", "numpy.array", "time.time" ]
[((309, 339), 'sys.setrecursionlimit', 'sys.setrecursionlimit', (['(10 ** 7)'], {}), '(10 ** 7)\n', (330, 339), False, 'import sys\n'), ((2394, 2678), 'numpy.array', 'np.array', (['[[5, 3, 0, 0, 7, 0, 0, 0, 0], [6, 0, 0, 1, 9, 5, 0, 0, 0], [0, 9, 8, 0, 0, \n 0, 0, 6, 0], [8, 0, 0, 0, 6, 0, 0, 0, 3], [4, 0, 0, 8, 0, ...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ **Project Name:** MakeHuman **Product Home Page:** http://www.makehumancommunity.org/ **Github Code Home Page:** https://github.com/makehumancommunity/ **Authors:** <NAME>, <NAME> **Copyright(c):** MakeHuman Team 2001-2019 **Licensing:** ...
[ "getpath.thoroughFindFile", "transformations.affine_matrix_from_points", "material.Material", "io.open", "numpy.array", "makehuman.getAssetLicense", "getpath.formatPath", "animation.VertexBoneWeights.fromFile", "os.remove", "log.notice", "numpy.asarray", "time.perf_counter", "os.path.split",...
[((1684, 1705), 'numpy.identity', 'np.identity', (['(3)', 'float'], {}), '(3, float)\n', (1695, 1705), True, 'import numpy as np\n'), ((20405, 20424), 'io.open', 'io.open', (['path', '"""wb"""'], {}), "(path, 'wb')\n", (20412, 20424), False, 'import io\n'), ((20649, 20670), 'os.path.dirname', 'os.path.dirname', (['path...
import numpy as np import unittest import pytest from pysph.base.particle_array import ParticleArray import pysph.tools.mesh_tools as G from pysph.base.utils import get_particle_array # Data of a unit length cube def cube_data(): points = np.array([[0., 0., 0.], [0., 1., 0.], ...
[ "pysph.tools.mesh_tools.get_points_from_mgrid", "numpy.ones", "pysph.tools.mesh_tools.prism", "pysph.tools.mesh_tools.surface_points", "pysph.base.particle_array.ParticleArray", "pysph.tools.mesh_tools.surf_points_uniform", "pysph.tools.mesh_tools._get_surface_mesh", "pysph.tools.mesh_tools._in_triang...
[((245, 396), 'numpy.array', 'np.array', (['[[0.0, 0.0, 0.0], [0.0, 1.0, 0.0], [1.0, 1.0, 0.0], [1.0, 0.0, 0.0], [0.0, \n 0.0, 1.0], [0.0, 1.0, 1.0], [1.0, 0.0, 1.0], [1.0, 1.0, 1.0]]'], {}), '([[0.0, 0.0, 0.0], [0.0, 1.0, 0.0], [1.0, 1.0, 0.0], [1.0, 0.0, 0.0\n ], [0.0, 0.0, 1.0], [0.0, 1.0, 1.0], [1.0, 0.0, 1.0...
import xarray as xr import pandas as pd import cartopy import cartopy.crs as ccrs import matplotlib as mpl import matplotlib.pyplot as plt from matplotlib.cm import get_cmap import numpy as np from cartopy.mpl.gridliner import LONGITUDE_FORMATTER, LATITUDE_FORMATTER import shapely.geometry as sgeom import cartopy.featu...
[ "copy.copy", "numpy.arange", "numpy.multiply", "numpy.linspace", "matplotlib.cm.get_cmap", "matplotlib.pyplot.savefig", "numpy.amin", "cartopy.crs.PlateCarree", "shapely.geometry.LineString", "matplotlib.colors.Normalize", "matplotlib.pyplot.title", "xarray.open_dataset", "cartopy.crs.Geodet...
[((22322, 22367), 'xarray.open_dataset', 'xr.open_dataset', (["(path + 'pressure_anomaly.nc')"], {}), "(path + 'pressure_anomaly.nc')\n", (22337, 22367), True, 'import xarray as xr\n'), ((22465, 22526), 'xarray.open_dataset', 'xr.open_dataset', (['"""atms597_proj3/data/pressure_anomaly_new.nc"""'], {}), "('atms597_proj...
#!/usr/bin/python # coding: utf-8 import numpy as np from .facet import Facet class PosRegion(): """ Implement the convex polytope """ def __init__(self, pos_samples): """ Params: pos_samples (np.array): dim+1 positive samples to create the (dim)-polytope...
[ "numpy.delete", "numpy.vstack" ]
[((4221, 4259), 'numpy.delete', 'np.delete', (['self.vertices', 'to_remove', '(0)'], {}), '(self.vertices, to_remove, 0)\n', (4230, 4259), True, 'import numpy as np\n'), ((862, 905), 'numpy.delete', 'np.delete', (['self.vertices', 'sample_id'], {'axis': '(0)'}), '(self.vertices, sample_id, axis=0)\n', (871, 905), True,...
import gym import gym_sokoban import torch import numpy as np import random import time from utilities.channelConverter import hwc2chw from experts.utils import get_distance from external_actions import get_astar_action import warnings warnings.simplefilter("ignore", UserWarning) def test_the_agent(agent, data_path...
[ "external_actions.get_astar_action", "experts.utils.get_distance", "time.sleep", "numpy.sum", "utilities.channelConverter.hwc2chw", "warnings.simplefilter", "torch.no_grad", "gym.make" ]
[((239, 283), 'warnings.simplefilter', 'warnings.simplefilter', (['"""ignore"""', 'UserWarning'], {}), "('ignore', UserWarning)\n", (260, 283), False, 'import warnings\n'), ((660, 714), 'gym.make', 'gym.make', (['"""Curriculum-Sokoban-v2"""'], {'data_path': 'data_path'}), "('Curriculum-Sokoban-v2', data_path=data_path)...
""" Test `sinethesizer.effects.equalizer` module. Author: <NAME> """ from typing import Any, Dict, List import numpy as np import pytest from scipy.signal import spectrogram from sinethesizer.effects.equalizer import apply_equalizer from sinethesizer.synth.core import Event from sinethesizer.oscillators import gen...
[ "sinethesizer.effects.equalizer.apply_equalizer", "numpy.ones", "scipy.signal.spectrogram", "numpy.testing.assert_almost_equal", "numpy.array", "numpy.vstack" ]
[((4672, 4697), 'numpy.vstack', 'np.vstack', (['(sound, sound)'], {}), '((sound, sound))\n', (4681, 4697), True, 'import numpy as np\n'), ((4920, 4965), 'sinethesizer.effects.equalizer.apply_equalizer', 'apply_equalizer', (['sound', 'event', 'kind'], {}), '(sound, event, kind, **kwargs)\n', (4935, 4965), False, 'from s...
from __future__ import annotations import datetime import json from abc import ABC, abstractmethod from collections import defaultdict, deque, namedtuple from typing import ( Any, Deque, Dict, Iterator, List, Mapping, Optional, Set, TextIO, Tuple, ) import numpy as np import sc...
[ "collections.namedtuple", "collections.deque", "numpy.array", "collections.defaultdict", "datetime.date", "json.load" ]
[((396, 442), 'collections.namedtuple', 'namedtuple', (['"""ColumnValue"""', "['column', 'value']"], {}), "('ColumnValue', ['column', 'value'])\n", (406, 442), False, 'from collections import defaultdict, deque, namedtuple\n'), ((6342, 6374), 'numpy.array', 'np.array', (['data'], {'dtype': 'np.float32'}), '(data, dtype...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- import numpy as np alpha = np.random.rand(7) alpha /= np.linalg.norm(alpha, 1) n = 40 def index_to_position(index): p = 0 a, b, c, d, e, f = index index = [a, d, b, e, c, f] for i in index: p = p * n + i return p if __name__ == "__main__": ...
[ "numpy.random.rand", "numpy.linalg.norm" ]
[((76, 93), 'numpy.random.rand', 'np.random.rand', (['(7)'], {}), '(7)\n', (90, 93), True, 'import numpy as np\n'), ((103, 127), 'numpy.linalg.norm', 'np.linalg.norm', (['alpha', '(1)'], {}), '(alpha, 1)\n', (117, 127), True, 'import numpy as np\n'), ((476, 493), 'numpy.random.rand', 'np.random.rand', (['(7)'], {}), '(...
import torch import torch.optim as optim from torch import autograd import numpy as np from tqdm import trange import trimesh from skimage import measure import warnings import time from pipelines.utils.point_utils import sample_points_from_ray, np_get_occupied_idx, occupancy_sparse_to_dense from pipelines.utils.postpr...
[ "skimage.measure.marching_cubes_lewiner", "numpy.isin", "torch.from_numpy", "numpy.array", "numpy.arange", "pipelines.utils.point_utils.sample_points_from_ray", "pipelines.utils.postprocess_utils.remove_backface", "numpy.max", "numpy.stack", "numpy.linspace", "numpy.concatenate", "numpy.min", ...
[((2985, 3015), 'numpy.concatenate', 'np.concatenate', (['[v, n]'], {'axis': '(1)'}), '([v, n], axis=1)\n', (2999, 3015), True, 'import numpy as np\n'), ((3032, 3049), 'numpy.min', 'np.min', (['v'], {'axis': '(0)'}), '(v, axis=0)\n', (3038, 3049), True, 'import numpy as np\n'), ((3065, 3082), 'numpy.max', 'np.max', (['...
import numpy as np from time import time from typing import List, Tuple from tsp_heuristics.heuristics.utils import get_tour_distance def nn_algo( dist_matrix: np.array, start: int = 0 ) -> Tuple[List, float]: """ From a start city index, get an Tour according to the Nearest Neighbor algorithm from...
[ "numpy.argmin", "time.time", "tsp_heuristics.heuristics.utils.get_tour_distance" ]
[((666, 672), 'time.time', 'time', ([], {}), '()\n', (670, 672), False, 'from time import time\n'), ((994, 1026), 'numpy.argmin', 'np.argmin', (['dist_matrix[Tour[-1]]'], {}), '(dist_matrix[Tour[-1]])\n', (1003, 1026), True, 'import numpy as np\n'), ((1266, 1302), 'tsp_heuristics.heuristics.utils.get_tour_distance', 'g...
import os import numpy as np from sys import platform, path if platform == "linux" or platform == "linux2": path.insert(1, os.path.dirname(os.getcwd()) + "/src") FILE_NAME = os.path.dirname(os.getcwd()) + "/data" + "/xAPI-Edu-Data-Edited.csv" elif platform == "win32": path.insert(1, os.path.dirname(os.getc...
[ "numpy.argpartition", "DataPreprocessing.FeaturePreprocess", "os.getcwd", "DataProcessing.save_file", "DataProcessing.load_file", "DataProcessing.ModelValidating", "DataPreprocessing.Preprocess", "DataProcessing.ModelTuning" ]
[((2478, 2504), 'DataPreprocessing.Preprocess', 'Preprocess', ([], {'data': 'FILE_NAME'}), '(data=FILE_NAME)\n', (2488, 2504), False, 'from DataPreprocessing import Preprocess, FeaturePreprocess\n'), ((4600, 4677), 'DataPreprocessing.FeaturePreprocess', 'FeaturePreprocess', (['X_data'], {'n_components': 'n_components',...
import gym from baselines import deepq from baselines.common.atari_wrappers_deprecated import wrap_dqn, ScaledFloatFrame from cloud_environment import CloudEnvironment import numpy as np import collections import os import csv import pandas as pd #Logging def logger_callback(locals,globals): done = locals[...
[ "numpy.mean", "cloud_environment.CloudEnvironment", "csv.writer", "baselines.deepq.test", "numpy.sum", "numpy.zeros", "collections.Counter", "baselines.deepq.models.cnn_to_mlp", "numpy.concatenate", "os.stat", "pandas.concat" ]
[((2638, 2670), 'pandas.concat', 'pd.concat', (['[e_df, ci_df]'], {'axis': '(1)'}), '([e_df, ci_df], axis=1)\n', (2647, 2670), True, 'import pandas as pd\n'), ((3001, 3216), 'cloud_environment.CloudEnvironment', 'CloudEnvironment', ([], {'img_size': 'img_size', 'radius': '[12, 13]', 'sequence_stride': '(1)', 'channels'...
#!/usr/bin/env python from argparse import ArgumentParser, ArgumentDefaultsHelpFormatter import numpy as np import scipy.io import glob import os import csv import random import tensorflow as tf import transition_model_common as tm # import sys # sys.path.append('./tensorflow_hmm') # import tensorflow_hmm.hmm as hmm ...
[ "tensorflow.nn.softmax", "transition_model_common.create_model", "tensorflow.cast", "tensorflow.variables_initializer", "transition_model_common.RobotDataLoader", "os.path.exists", "numpy.reshape", "argparse.ArgumentParser", "tensorflow.squared_difference", "tensorflow.Session", "tensorflow.plac...
[((350, 365), 'transition_model_common.DataLoader', 'tm.DataLoader', ([], {}), '()\n', (363, 365), True, 'import transition_model_common as tm\n'), ((583, 630), 'transition_model_common.create_model', 'tm.create_model', (['n_input', 'n_classes'], {'train': '(True)'}), '(n_input, n_classes, train=True)\n', (598, 630), T...
import matplotlib.pyplot as plt import seaborn as sns import pandas as pd import numpy as np sns.set() path = r'C:\Users\HP\PycharmProjects\Operaciones_Calor\Data.xlsx' df = pd.read_excel("Data.xlsx") df2 = pd.read_excel("time.xlsx") df3 = np.transpose(df) ax = sns.heatmap(data=df3) plt.show()
[ "seaborn.set", "seaborn.heatmap", "pandas.read_excel", "numpy.transpose", "matplotlib.pyplot.show" ]
[((94, 103), 'seaborn.set', 'sns.set', ([], {}), '()\n', (101, 103), True, 'import seaborn as sns\n'), ((175, 201), 'pandas.read_excel', 'pd.read_excel', (['"""Data.xlsx"""'], {}), "('Data.xlsx')\n", (188, 201), True, 'import pandas as pd\n'), ((208, 234), 'pandas.read_excel', 'pd.read_excel', (['"""time.xlsx"""'], {})...
""" This module deals with querying and downloading LAT FITS files. """ # Scientific Library import numpy as np import pandas as pd # Requests Urls and Manupilate Files from astropy.utils.data import download_files_in_parallel, download_file from astroquery import fermi from tqdm import tqdm import requests import ...
[ "logging.basicConfig", "os.makedirs", "functools.reduce", "tqdm.tqdm", "logging.warning", "numpy.any", "numpy.sum", "shutil.copyfile", "os.path.dirname", "signal.alarm", "retry.api.retry_call", "pandas.read_html", "pandas.isna", "logging.info" ]
[((457, 499), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.WARNING'}), '(level=logging.WARNING)\n', (476, 499), False, 'import logging\n'), ((2439, 2470), 'pandas.isna', 'pd.isna', (['self.grbs.at[row, col]'], {}), '(self.grbs.at[row, col])\n', (2446, 2470), True, 'import pandas as pd\n'), ((24...
# Copyright (c) 2020, <NAME>, Honda Research Institute Europe GmbH, and # Technical University of Darmstadt. # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # 1. Redistributions of source code mus...
[ "numpy.clip", "pyrado.policies.recurrent.rnn.LSTMPolicy", "pyrado.environment_wrappers.base.EnvWrapper.__init__", "torch.cuda.is_available", "pyrado.algorithms.step_based.svpg.SVPGBuilder", "pyrado.save", "pyrado.sampling.parallel_evaluation.eval_domain_params", "numpy.mean", "pyrado.sampling.sample...
[((21755, 21781), 'torch.cat', 'to.cat', (['(state, action)', '(1)'], {}), '((state, action), 1)\n', (21761, 21781), True, 'import torch as to\n'), ((6471, 6608), 'numpy.random.random_sample', 'np.random.random_sample', (["(svpg_hp['algo']['num_particles'], svpg_hp['algo']['horizon'], adr_hp[\n 'evaluation_steps'], ...
""" Stats stuff! """ import textwrap import numpy as np import pandas as pd from scipy import stats from IPython.display import display from .boxes import * from .table_display import * __DEBUG__ = False def debug(*args, **kwargs): if __DEBUG__: print(*args, **kwargs) class Chi2Result(object): "...
[ "textwrap.dedent", "IPython.display.display", "numpy.linalg.solve", "scipy.stats.chi2.cdf", "scipy.stats.chi2_contingency", "pandas.crosstab", "numpy.diag", "numpy.kron", "numpy.zeros", "numpy.outer", "pandas.concat" ]
[((2089, 2112), 'pandas.crosstab', 'pd.crosstab', (['col1', 'col2'], {}), '(col1, col2)\n', (2100, 2112), True, 'import pandas as pd\n'), ((2139, 2165), 'scipy.stats.chi2_contingency', 'stats.chi2_contingency', (['xs'], {}), '(xs)\n', (2161, 2165), False, 'from scipy import stats\n'), ((4582, 4616), 'numpy.zeros', 'np....
import os import numpy as np import matplotlib.pyplot as plt from .kshell_utilities import atomic_numbers, loadtxt from .general_utilities import create_spin_parity_list, gamma_strength_function_average class LEE: def __init__(self, directory): self.bin_width = 0.2 self.E_max = 30 self.Ex_m...
[ "numpy.ceil", "os.listdir", "numpy.sum", "numpy.linspace", "os.path.isdir", "matplotlib.pyplot.subplots", "matplotlib.pyplot.show" ]
[((563, 605), 'numpy.linspace', 'np.linspace', (['(0)', 'E_max_adjusted', '(n_bins + 1)'], {}), '(0, E_max_adjusted, n_bins + 1)\n', (574, 605), True, 'import numpy as np\n'), ((7272, 7286), 'matplotlib.pyplot.subplots', 'plt.subplots', ([], {}), '()\n', (7284, 7286), True, 'import matplotlib.pyplot as plt\n'), ((7579,...
"""Tests for Pipeline class""" import pytest from cognigraph.nodes.pipeline import Pipeline import numpy as np from numpy.testing import assert_array_equal from cognigraph.tests.prepare_pipeline_tests import ( create_dummy_info, ConcreteSource, ConcreteProcessor, ConcreteOutput, ) @pytest.fixture(scop...
[ "cognigraph.tests.prepare_pipeline_tests.ConcreteSource", "cognigraph.tests.prepare_pipeline_tests.create_dummy_info", "numpy.ones", "cognigraph.tests.prepare_pipeline_tests.ConcreteProcessor", "cognigraph.tests.prepare_pipeline_tests.ConcreteOutput", "numpy.zeros", "pytest.fixture", "numpy.all", "c...
[((301, 333), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""function"""'}), "(scope='function')\n", (315, 333), False, 'import pytest\n'), ((363, 379), 'cognigraph.tests.prepare_pipeline_tests.ConcreteSource', 'ConcreteSource', ([], {}), '()\n', (377, 379), False, 'from cognigraph.tests.prepare_pipeline_tests ...
#!/usr/bin/env python '''Testing for read_rockstar.py @author: <NAME> @contact: <EMAIL> @status: Development ''' import glob from mock import call, patch import numpy as np import numpy.testing as npt import os import pdb import pytest import unittest import galaxy_dive.read_data.rockstar as read_rockstar import gal...
[ "numpy.testing.assert_allclose", "galaxy_dive.read_data.rockstar.RockstarReader" ]
[((603, 660), 'galaxy_dive.read_data.rockstar.RockstarReader', 'read_rockstar.RockstarReader', (['"""./tests/data/rockstar_dir"""'], {}), "('./tests/data/rockstar_dir')\n", (631, 660), True, 'import galaxy_dive.read_data.rockstar as read_rockstar\n'), ((901, 938), 'numpy.testing.assert_allclose', 'npt.assert_allclose',...
from typing import Union, Optional, Sequence, Any, Mapping, List, Tuple, Callable from collections.abc import Iterable import operator import numpy as np import pandas as pd import matplotlib.pyplot as plt from matplotlib.axes import Axes def pie_marker( ratios: Sequence[float], res: int = 50, ...
[ "numpy.abs", "numpy.column_stack", "numpy.sum", "numpy.linspace", "matplotlib.pyplot.scatter", "pandas.DataFrame", "matplotlib.pyplot.subplots", "matplotlib.pyplot.get_cmap" ]
[((3013, 3061), 'pandas.DataFrame', 'pd.DataFrame', (["{'x': x, 'y': y, 'ratios': ratios}"], {}), "({'x': x, 'y': y, 'ratios': ratios})\n", (3025, 3061), True, 'import pandas as pd\n'), ((3646, 3664), 'matplotlib.pyplot.get_cmap', 'plt.get_cmap', (['cmap'], {}), '(cmap)\n', (3658, 3664), True, 'import matplotlib.pyplot...
""" Test script for data.py classes. """ import os import numpy as np import pytest from bilby.core.prior import PriorDict, Uniform from cwinpy import HeterodynedData, MultiHeterodynedData, TargetedPulsarLikelihood class TestTargetedPulsarLikelhood(object): """ Tests for the TargetedPulsarLikelihood class. ...
[ "numpy.random.normal", "cwinpy.HeterodynedData", "numpy.ones", "cwinpy.MultiHeterodynedData", "cwinpy.TargetedPulsarLikelihood", "numpy.linspace", "pytest.raises", "bilby.core.prior.PriorDict", "bilby.core.prior.Uniform", "os.remove" ]
[((372, 417), 'numpy.linspace', 'np.linspace', (['(1000000000.0)', '(1000086340.0)', '(1440)'], {}), '(1000000000.0, 1000086340.0, 1440)\n', (383, 417), True, 'import numpy as np\n'), ((429, 473), 'numpy.random.normal', 'np.random.normal', (['(0.0)', '(1e-25)'], {'size': '(1440, 2)'}), '(0.0, 1e-25, size=(1440, 2))\n',...
############################################################################### # mockDensData.py: generate mock data following a given density ############################################################################### import os, os.path import pickle import multiprocessing from optparse import OptionParser import...
[ "mwdust.Drimmel03", "define_rcsample.get_rcsample", "densprofiles.logit", "numpy.log", "multiprocessing.cpu_count", "numpy.array", "mwdust.Green15", "os.path.exists", "numpy.linspace", "numpy.nanmax", "numpy.amax", "numpy.amin", "mwdust.Marshall06", "pickle.load", "galpy.util.bovy_coords...
[((1685, 1714), 'fitDens._setup_densfunc', 'fitDens._setup_densfunc', (['type'], {}), '(type)\n', (1708, 1714), False, 'import fitDens\n'), ((2767, 2793), 'os.path.exists', 'os.path.exists', (['selectFile'], {}), '(selectFile)\n', (2781, 2793), False, 'import os, os.path\n'), ((2999, 3029), 'numpy.linspace', 'numpy.lin...
import os import numpy as np import pandas as pd import yaml from . import model as model_lib from . import training, tensorize, io_local def main(): #Turn off warnings: os.environ["TF_CPP_MIN_LOG_LEVEL"] = "3" ###Load training data - Put the path to your own data here training_data_path = "/root/trai...
[ "numpy.stack", "numpy.sqrt", "pandas.read_csv", "yaml.dump" ]
[((368, 399), 'pandas.read_csv', 'pd.read_csv', (['training_data_path'], {}), '(training_data_path)\n', (379, 399), True, 'import pandas as pd\n'), ((2552, 2594), 'numpy.stack', 'np.stack', (['training_df.relative_intensities'], {}), '(training_df.relative_intensities)\n', (2560, 2594), True, 'import numpy as np\n'), (...
from time import perf_counter import numpy as np from sklearn.model_selection import RepeatedStratifiedKFold from sklearn.metrics import * from sklearn.neighbors import KNeighborsClassifier from sklearn.svm import SVC from sklearn.decomposition import PCA from sklearn.manifold import TSNE import matplotlib.pyplot as pl...
[ "sklearn.metrics.f1_score", "numpy.load", "numpy.random.seed", "hpsklearn.random_forest" ]
[((545, 563), 'numpy.random.seed', 'np.random.seed', (['(42)'], {}), '(42)\n', (559, 563), True, 'import numpy as np\n'), ((578, 605), 'numpy.load', 'np.load', (['"""data/train_X.npy"""'], {}), "('data/train_X.npy')\n", (585, 605), True, 'import numpy as np\n'), ((619, 645), 'numpy.load', 'np.load', (['"""data/test_X.n...
import numpy as np from scipy.spatial.distance import pdist, squareform from scipy import exp from scipy.linalg import eigh def rbf_kernel_pca(X, gamma, n_components): """ RBF kernel PCA implementation. Parameters ------------ X: {NumPy ndarray}, shape = [n_samples, n_features] gamma: float ...
[ "scipy.linalg.eigh", "scipy.spatial.distance.squareform", "numpy.sqrt", "numpy.ones", "scipy.exp", "scipy.spatial.distance.pdist", "numpy.dot" ]
[((660, 683), 'scipy.spatial.distance.pdist', 'pdist', (['X', '"""sqeuclidean"""'], {}), "(X, 'sqeuclidean')\n", (665, 683), False, 'from scipy.spatial.distance import pdist, squareform\n'), ((759, 779), 'scipy.spatial.distance.squareform', 'squareform', (['sq_dists'], {}), '(sq_dists)\n', (769, 779), False, 'from scip...
# %% import timeit import tqdm from os import path import inspect import numpy as np import dill import init import fastg3.crisp as g3crisp from plot_utils import plot_bench from constants import N_REPEATS, N_STEPS, DILL_FOLDER from number_utils import format_number from dataset_utils import AVAILABLE_DATASETS, load_d...
[ "init.gen_sampling_benchmark", "number_utils.format_number", "inspect.stack", "tqdm.tqdm", "os.path.join", "os.path.isfile", "dataset_utils.load_dataset", "timeit.timeit", "plot_utils.plot_bench", "numpy.arange", "init.gen_time_benchmark" ]
[((718, 743), 'init.gen_time_benchmark', 'init.gen_time_benchmark', ([], {}), '()\n', (741, 743), False, 'import init\n'), ((757, 780), 'tqdm.tqdm', 'tqdm.tqdm', (['frac_samples'], {}), '(frac_samples)\n', (766, 780), False, 'import tqdm\n'), ((1367, 1396), 'init.gen_sampling_benchmark', 'init.gen_sampling_benchmark', ...
# -*- coding: utf-8 -*- # Copyright 2018 <NAME> & <NAME>. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # #...
[ "numpy.array", "numpy.log" ]
[((888, 901), 'numpy.array', 'np.array', (['out'], {}), '(out)\n', (896, 901), True, 'import numpy as np\n'), ((903, 918), 'numpy.array', 'np.array', (['label'], {}), '(label)\n', (911, 918), True, 'import numpy as np\n'), ((1551, 1564), 'numpy.array', 'np.array', (['out'], {}), '(out)\n', (1559, 1564), True, 'import n...
#!/usr/bin/env python from __future__ import division, print_function try: range = xrange except NameError: pass import os import sys import h5py import json import time import numpy import ctypes import signal import logging import argparse import threading from functools import reduce from datetime impo...
[ "logging.getLogger", "datetime.datetime.utcfromtimestamp", "logging.StreamHandler", "numpy.sqrt", "bifrost.packet_capture.PacketCaptureCallback", "numpy.log10", "bifrost.affinity.get_core", "ctypes.create_string_buffer", "time.sleep", "numpy.isfinite", "datetime.timedelta", "mnc.mcs.Client", ...
[((1240, 1261), 'operations.FileOperationsQueue', 'FileOperationsQueue', ([], {}), '()\n', (1259, 1261), False, 'from operations import FileOperationsQueue\n'), ((1203, 1228), 'os.path.abspath', 'os.path.abspath', (['__file__'], {}), '(__file__)\n', (1218, 1228), False, 'import os\n'), ((30595, 30682), 'argparse.Argume...
import copy import numpy as np import pandas as pd import os import contextlib from sklearn.metrics import f1_score, accuracy_score from sklearn.model_selection import StratifiedKFold from sklearn.linear_model import LogisticRegression from sklearn.pipeline import make_pipeline from sklearn.preprocessing import Standa...
[ "contextlib.redirect_stdout", "sklearn.metrics.f1_score", "numpy.where", "sklearn.linear_model.LogisticRegression", "sklearn.model_selection.StratifiedKFold", "sklearn.preprocessing.StandardScaler", "numpy.zeros", "numpy.empty", "copy.deepcopy", "pandas.DataFrame", "sklearn.metrics.accuracy_scor...
[((358, 423), 'sklearn.model_selection.StratifiedKFold', 'StratifiedKFold', ([], {'n_splits': 'NFOLDS', 'shuffle': '(True)', 'random_state': 'SEED'}), '(n_splits=NFOLDS, shuffle=True, random_state=SEED)\n', (373, 423), False, 'from sklearn.model_selection import StratifiedKFold\n'), ((559, 586), 'numpy.where', 'np.wher...
import numpy as np import yt from matplotlib import rc fsize = 17 rc('text', usetex=False) rc('font', size=fsize)#, ftype=42) line_width = 3 point_size = 30 import matplotlib as mpl mpl.use('Agg') import matplotlib.pyplot as plt from galaxy_analysis.particle_analysis import particle_types as pdef def plot_dtd(ds): ...
[ "matplotlib.use", "galaxy_analysis.particle_analysis.particle_types.snIa", "matplotlib.pyplot.minorticks_on", "matplotlib.pyplot.close", "numpy.array", "galaxy_analysis.particle_analysis.particle_types.white_dwarfs", "yt.load", "matplotlib.rc", "matplotlib.pyplot.tight_layout", "matplotlib.pyplot....
[((66, 90), 'matplotlib.rc', 'rc', (['"""text"""'], {'usetex': '(False)'}), "('text', usetex=False)\n", (68, 90), False, 'from matplotlib import rc\n'), ((91, 113), 'matplotlib.rc', 'rc', (['"""font"""'], {'size': 'fsize'}), "('font', size=fsize)\n", (93, 113), False, 'from matplotlib import rc\n'), ((182, 196), 'matpl...
"""Tests cac.models.classification.ClassificationModel""" import os from os.path import dirname, join, exists from copy import deepcopy import torch import wandb import unittest from tqdm import tqdm import numpy as np from torch import optim from cac.config import Config from cac.utils.logger import set_logger, color ...
[ "numpy.unique", "numpy.sort", "os.path.join", "numpy.array", "torch.cuda.is_available", "copy.deepcopy", "unittest.main", "cac.models.classification.ClassificationModel", "cac.config.Config" ]
[((3251, 3266), 'unittest.main', 'unittest.main', ([], {}), '()\n', (3264, 3266), False, 'import unittest\n'), ((587, 602), 'cac.config.Config', 'Config', (['version'], {}), '(version)\n', (593, 602), False, 'from cac.config import Config\n'), ((1202, 1220), 'copy.deepcopy', 'deepcopy', (['self.cfg'], {}), '(self.cfg)\...
import numpy as np from numba import jit from numba.core import types from numba.tests.support import TestCase, tag import unittest # Array overlaps involving a displacement def array_overlap1(src, dest, k=1): assert src.shape == dest.shape dest[k:] = src[:-k] def array_overlap2(src, dest, k=1): assert...
[ "unittest.main", "numba.jit", "numpy.arange" ]
[((3823, 3838), 'unittest.main', 'unittest.main', ([], {}), '()\n', (3836, 3838), False, 'import unittest\n'), ((1959, 1977), 'numba.jit', 'jit', ([], {'nopython': '(True)'}), '(nopython=True)\n', (1962, 1977), False, 'from numba import jit\n'), ((2113, 2136), 'numpy.arange', 'np.arange', (['(0)', '(N ** ndim)'], {}), ...
import pandas as pd import numpy as np from copy import deepcopy import warnings from sklearn.base import BaseEstimator, TransformerMixin from sklearn.model_selection import cross_val_predict from sklearn.model_selection import KFold, StratifiedKFold from sklearn.externals.joblib import Parallel, delayed from gravity_l...
[ "copy.deepcopy", "sklearn.externals.joblib.delayed", "pandas.DataFrame", "numpy.hstack", "gravity_learn.utils.force_array", "sklearn.model_selection.StratifiedKFold", "gravity_learn.utils.check_is_fitted", "sklearn.externals.joblib.Parallel", "sklearn.model_selection.cross_val_predict", "gravity_l...
[((2660, 2780), 'warnings.warn', 'warnings.warn', (['"""EnsemblerClassifier is deprecated, please use FullStackClassifier instead"""', 'DeprecationWarning'], {}), "(\n 'EnsemblerClassifier is deprecated, please use FullStackClassifier instead'\n , DeprecationWarning)\n", (2673, 2780), False, 'import warnings\n'),...
import numpy as np from sklearn.exceptions import NotFittedError from sklearn.linear_model import SGDClassifier from sklearn.linear_model.base import LinearClassifierMixin from sklearn.utils import check_array import faiss def _default_index(d): index = faiss.index_factory(d, "IVF2048,Flat", faiss.METRIC_INNER_P...
[ "faiss.index_factory", "sklearn.utils.check_array", "numpy.ascontiguousarray" ]
[((261, 327), 'faiss.index_factory', 'faiss.index_factory', (['d', '"""IVF2048,Flat"""', 'faiss.METRIC_INNER_PRODUCT'], {}), "(d, 'IVF2048,Flat', faiss.METRIC_INNER_PRODUCT)\n", (280, 327), False, 'import faiss\n'), ((713, 748), 'sklearn.utils.check_array', 'check_array', (['X'], {'accept_sparse': '(False)'}), '(X, acc...
""" Created on Thursday Mar 26 2020 <NAME> based on https://www.kaggle.com/bardor/covid-19-growing-rate https://github.com/CSSEGISandData/COVID-19 https://github.com/imdevskp https://www.kaggle.com/yamqwe/covid-19-status-israel """ import datetime import numpy as np import pandas as pd import seaborn as...
[ "matplotlib.pyplot.grid", "matplotlib.pyplot.ylabel", "numpy.log", "seaborn.scatterplot", "datetime.timedelta", "folium.CircleMarker", "numpy.arange", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.plot", "matplotlib.pyplot.style.use", "numpy.diff", "folium.Map", "numpy.max", "numpy.linspa...
[((554, 586), 'matplotlib.pyplot.style.use', 'plt.style.use', (['"""dark_background"""'], {}), "('dark_background')\n", (567, 586), True, 'import matplotlib.pyplot as plt\n'), ((1813, 1839), 'matplotlib.pyplot.figure', 'plt.figure', ([], {'figsize': '(9, 7)'}), '(figsize=(9, 7))\n', (1823, 1839), True, 'import matplotl...
#!/usr/bin/env python # coding: utf8 # # Copyright (c) 2021 Centre National d'Etudes Spatiales (CNES). # # This file is part of PANDORA_MCCNN # # https://github.com/CNES/Pandora_MCCNN # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the Licens...
[ "argparse.ArgumentParser", "numpy.arange", "numpy.where", "rasterio.open", "numba.njit", "os.path.join", "numpy.squeeze", "numpy.stack", "numpy.sum", "numpy.random.seed", "numpy.zeros_like", "glob.glob", "numpy.random.shuffle" ]
[((987, 993), 'numba.njit', 'njit', ([], {}), '()\n', (991, 993), False, 'from numba import njit\n'), ((4961, 5011), 'glob.glob', 'glob.glob', (["(input_dir + '/*/left_epipolar_disp.tif')"], {}), "(input_dir + '/*/left_epipolar_disp.tif')\n", (4970, 5011), False, 'import glob\n'), ((5076, 5093), 'numpy.arange', 'np.ara...
import numpy as np # NOTE: these all assume a sample rate of 1000Hz and 0-centered(ish) BUTTER2_45_55_NOTCH = [[0.95654323, -1.82035157, 0.95654323, 1., -1.84458768, 0.9536256 ], [1. , -1.90305207, 1. , 1., -1.87701816, 0.95947072]] BUTTER4_45_55_NOTCH = [[0.92117099, -1.75303637, ...
[ "numpy.array" ]
[((2752, 2768), 'numpy.array', 'np.array', (['coeffs'], {}), '(coeffs)\n', (2760, 2768), True, 'import numpy as np\n'), ((2786, 2820), 'numpy.array', 'np.array', (['([[0.0] * 2] * self.order)'], {}), '([[0.0] * 2] * self.order)\n', (2794, 2820), True, 'import numpy as np\n')]
# -*- coding: utf-8 -*- """ Created on Tue May 12 08:23:58 2020 @author: sumanth """ import numpy as np import cv2 from tensorflow.keras.applications.mobilenet_v2 import preprocess_input from tensorflow.keras.preprocessing.image import img_to_array def pre_dect(frame,faceNet,model): (h, w) = frame.s...
[ "cv2.dnn.blobFromImage", "numpy.array", "tensorflow.keras.applications.mobilenet_v2.preprocess_input", "cv2.cvtColor", "numpy.expand_dims", "tensorflow.keras.preprocessing.image.img_to_array", "cv2.resize" ]
[((341, 409), 'cv2.dnn.blobFromImage', 'cv2.dnn.blobFromImage', (['frame', '(1.0)', '(300, 300)', '(104.0, 177.0, 123.0)'], {}), '(frame, 1.0, (300, 300), (104.0, 177.0, 123.0))\n', (362, 409), False, 'import cv2\n'), ((976, 1013), 'cv2.cvtColor', 'cv2.cvtColor', (['face', 'cv2.COLOR_BGR2RGB'], {}), '(face, cv2.COLOR_B...
import shutil import subprocess # nosec # have to use subprocess import warnings from collections import Counter from copy import deepcopy from os import listdir, makedirs from os.path import abspath, basename, dirname, exists, isfile, join from subprocess import PIPE # nosec # have to use subprocess from tempfile im...
[ "f90nml.write", "dateutil.relativedelta.relativedelta", "f90nml.Namelist", "copy.deepcopy", "numpy.arange", "os.path.exists", "os.listdir", "subprocess.run", "pandas.DataFrame", "warnings.warn", "numpy.ones", "openscm_units.unit_registry", "os.path.isfile", "os.path.dirname", "scmdata.ru...
[((1232, 1247), 'os.listdir', 'listdir', (['source'], {}), '(source)\n', (1239, 1247), False, 'from os import listdir, makedirs\n'), ((1164, 1195), 'shutil.copytree', 'shutil.copytree', (['source', 'target'], {}), '(source, target)\n', (1179, 1195), False, 'import shutil\n'), ((1260, 1274), 'os.path.exists', 'exists', ...
import json import json import boto3 import re import json import collections import os import pandas as pd import csv from csv import writer # boto3 S3 initialization s3_client = boto3.client("s3") import numpy as np def lambda_handler(event, context): # TODO implement bucketname = 'sourcedatab00870639' ...
[ "boto3.client", "os.path.join", "numpy.zeros", "boto3.resource", "pandas.DataFrame", "csv.reader" ]
[((181, 199), 'boto3.client', 'boto3.client', (['"""s3"""'], {}), "('s3')\n", (193, 199), False, 'import boto3\n'), ((625, 666), 'os.path.join', 'os.path.join', (['sourcebucket', 'file_key_name'], {}), '(sourcebucket, file_key_name)\n', (637, 666), False, 'import os\n'), ((2414, 2440), 'numpy.zeros', 'np.zeros', (['(si...
""" Friends-of-Friends (FOF) for N-body simulations <NAME> - Oct 2016 """ from __future__ import absolute_import, print_function from lizard.periodic import pad_unitcube from scipy.spatial import Delaunay from scipy.sparse import csr_matrix, csgraph from numpy import square, flatnonzero, ones, zeros_like, cumsum, con...
[ "numpy.argsort", "numpy.random.RandomState", "pylab.ylim", "lizard.periodic.pad_unitcube", "pylab.plot", "numpy.sort", "numpy.flatnonzero", "numpy.diff", "lizard.log.VerboseTimingLog", "pylab.xlim", "numpy.concatenate", "numpy.square", "numpy.bincount", "pylab.show", "scipy.sparse.csgrap...
[((1200, 1220), 'lizard.periodic.pad_unitcube', 'pad_unitcube', (['pos', 'b'], {}), '(pos, b)\n', (1212, 1220), False, 'from lizard.periodic import pad_unitcube\n'), ((1523, 1540), 'scipy.spatial.Delaunay', 'Delaunay', (['all_pos'], {}), '(all_pos)\n', (1531, 1540), False, 'from scipy.spatial import Delaunay\n'), ((163...
import numpy as np from numpy.core.fromnumeric import size class BrownionPathGen: def __init__(self, NumPaths, Maturity): self.NumPaths = NumPaths self.Maturity = Maturity # this is in days # this is not optimal lets make a matrix of the std normal and then perform the operation to # ch...
[ "numpy.random.standard_normal", "numpy.zeros" ]
[((453, 503), 'numpy.random.standard_normal', 'np.random.standard_normal', ([], {'size': '[self.NumPaths, 1]'}), '(size=[self.NumPaths, 1])\n', (478, 503), True, 'import numpy as np\n'), ((1037, 1065), 'numpy.zeros', 'np.zeros', (['[self.NumPaths, 1]'], {}), '([self.NumPaths, 1])\n', (1045, 1065), True, 'import numpy a...
""" Module defining transfer functions """ from typing import List, Optional, Dict, Any, Union from pydantic import validator, constr import numpy as np import plotly.graph_objects as go from plotly.subplots import make_subplots from resistics.common import Metadata class Component(Metadata): """ Data class ...
[ "plotly.subplots.make_subplots", "pydantic.validator", "numpy.reciprocal", "numpy.power", "pydantic.constr", "numpy.unwrap", "numpy.array", "warnings.warn", "numpy.mod", "numpy.arctan" ]
[((3479, 3500), 'pydantic.constr', 'constr', ([], {'max_length': '(16)'}), '(max_length=16)\n', (3485, 3500), False, 'from pydantic import validator, constr\n'), ((8972, 9002), 'pydantic.validator', 'validator', (['"""name"""'], {'always': '(True)'}), "('name', always=True)\n", (8981, 9002), False, 'from pydantic impor...
import numpy import pytest import orthopy import quadpy from helpers import check_degree_ortho schemes = [ quadpy.e2r2.haegemans_piessens_a(), quadpy.e2r2.haegemans_piessens_b(), quadpy.e2r2.rabinowitz_richter_1(), quadpy.e2r2.rabinowitz_richter_2(), quadpy.e2r2.rabinowitz_richter_3(), quadpy....
[ "quadpy.e2r2.rabinowitz_richter_3", "quadpy.e2r2.rabinowitz_richter_4", "quadpy.e2r2.stroud_15_1", "numpy.sqrt", "quadpy.e2r2.haegemans_piessens_b", "quadpy.e2r2.rabinowitz_richter_1", "quadpy.e2r2.stroud_4_1", "quadpy.e2r2.rabinowitz_richter_2", "quadpy.e2r2.rabinowitz_richter_5", "quadpy.e2r2.ha...
[((770, 812), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""scheme"""', 'schemes'], {}), "('scheme', schemes)\n", (793, 812), False, 'import pytest\n'), ((113, 147), 'quadpy.e2r2.haegemans_piessens_a', 'quadpy.e2r2.haegemans_piessens_a', ([], {}), '()\n', (145, 147), False, 'import quadpy\n'), ((153, 187)...
""" Displays FISH data, raw and deconvolved, with spots detected using starFISH """ from skimage.io import imread import numpy as np from napari import Viewer, gui_qt raw = imread('data-njs/smFISH/raw.tif') deconvolved = imread('data-njs/smFISH/deconvolved.tif') spots = np.loadtxt('data-njs/smFISH/spots.csv', delimit...
[ "napari.Viewer", "numpy.roll", "napari.gui_qt", "skimage.io.imread", "numpy.loadtxt" ]
[((175, 208), 'skimage.io.imread', 'imread', (['"""data-njs/smFISH/raw.tif"""'], {}), "('data-njs/smFISH/raw.tif')\n", (181, 208), False, 'from skimage.io import imread\n'), ((223, 264), 'skimage.io.imread', 'imread', (['"""data-njs/smFISH/deconvolved.tif"""'], {}), "('data-njs/smFISH/deconvolved.tif')\n", (229, 264), ...
import numpy as np from scipy.spatial.distance import pdist, squareform import scipy.cluster.hierarchy as hy import matplotlib.pyplot as plt # Creating a cluster of clusters function def clusters(number=20, cnumber=5, csize=10): # Note that the way the clusters are positioned is Gaussian randomness. rnum = np...
[ "scipy.spatial.distance.squareform", "scipy.cluster.hierarchy.dendrogram", "numpy.random.rand", "numpy.where", "scipy.spatial.distance.pdist", "numpy.column_stack", "numpy.max", "matplotlib.pyplot.figure", "scipy.cluster.hierarchy.linkage", "numpy.vstack", "numpy.random.randn" ]
[((1070, 1088), 'scipy.spatial.distance.pdist', 'pdist', (['cls[:, 0:2]'], {}), '(cls[:, 0:2])\n', (1075, 1088), False, 'from scipy.spatial.distance import pdist, squareform\n'), ((1093, 1106), 'scipy.spatial.distance.squareform', 'squareform', (['D'], {}), '(D)\n', (1103, 1106), False, 'from scipy.spatial.distance imp...
from unittest import mock import gym import numpy as np from tests.fixtures.envs.dummy import DummyEnv class DummyDiscretePixelEnv(DummyEnv): """ A dummy discrete pixel environment. It follows Atari game convention, where actions are 'NOOP', 'FIRE', ... It also contains self.unwrapped....
[ "unittest.mock.Mock", "numpy.ones", "gym.spaces.Discrete", "gym.spaces.Box", "numpy.random.uniform", "numpy.full" ]
[((1029, 1040), 'unittest.mock.Mock', 'mock.Mock', ([], {}), '()\n', (1038, 1040), False, 'from unittest import mock\n'), ((1127, 1195), 'gym.spaces.Box', 'gym.spaces.Box', ([], {'low': '(0)', 'high': '(255)', 'shape': 'self._obs_dim', 'dtype': 'np.uint8'}), '(low=0, high=255, shape=self._obs_dim, dtype=np.uint8)\n', (...
import torch import numpy as np from ctc_decoders import Scorer, ctc_beam_search_decoder_batch """ # 安装语言模型 sudo apt-get install build-essential libboost-all-dev cmake zlib1g-dev libbz2-dev liblzma-dev git clone https://github.com/NVIDIA/OpenSeq2Seq -b ctc-decoders mv OpenSeq2Seq/decoders . rm -rf OpenSeq2Seq cd decod...
[ "torch.log_softmax", "ctc_decoders.ctc_beam_search_decoder_batch", "numpy.max", "ctc_decoders.Scorer", "torch.no_grad", "numpy.zeros_like", "torch.randn", "torch.IntTensor" ]
[((859, 874), 'torch.no_grad', 'torch.no_grad', ([], {}), '()\n', (872, 874), False, 'import torch\n'), ((2354, 2403), 'torch.randn', 'torch.randn', (['(2, 1000, 4334)'], {'dtype': 'torch.float32'}), '((2, 1000, 4334), dtype=torch.float32)\n', (2365, 2403), False, 'import torch\n'), ((1119, 1329), 'ctc_decoders.ctc_bea...
from datetime import datetime, date, timedelta import pandas as pd import networkx as nx from itertools import combinations import numpy as np class TeamworkStudyRunner: def __init__(self, notes, window_in_days, step_in_days): notes.sort_values('date', inplace=True) self.notes = notes sel...
[ "numpy.timedelta64", "itertools.combinations", "networkx.Graph", "numpy.arange" ]
[((330, 365), 'numpy.timedelta64', 'np.timedelta64', (['window_in_days', '"""D"""'], {}), "(window_in_days, 'D')\n", (344, 365), True, 'import numpy as np\n'), ((386, 419), 'numpy.timedelta64', 'np.timedelta64', (['step_in_days', '"""D"""'], {}), "(step_in_days, 'D')\n", (400, 419), True, 'import numpy as np\n'), ((533...
# nodenet/utilities/commons.py # Description: # "commons.py" provide commons utilities that can be use widely. # Copyright 2018 NOOXY. All Rights Reserved. from nodenet.imports.commons import * import numpy as np2 # np2 for cupy compabable def cut_dataset_by_ratio_ramdom(datasets, cut_ratio = 0.1): dimension = le...
[ "numpy.delete" ]
[((1010, 1047), 'numpy.delete', 'np2.delete', (['input_data', 'index'], {'axis': '(0)'}), '(input_data, index, axis=0)\n', (1020, 1047), True, 'import numpy as np2\n'), ((1070, 1108), 'numpy.delete', 'np2.delete', (['output_data', 'index'], {'axis': '(0)'}), '(output_data, index, axis=0)\n', (1080, 1108), True, 'import...
import tensorflow as tf import csv import time from datetime import timedelta import sys import numpy as np from tensorflow.python.training import training_util from tensorflow.contrib import slim from tensorflow.python.ops import variables as tf_variables from ..configuration import * from .. import trainer, evaluator...
[ "tensorflow.cast", "logging.getLogger", "tensorflow.python.training.training_util.get_or_create_global_step", "tensorflow.check_numerics", "tensorflow.Variable", "tensorflow.train.Saver", "numpy.sum", "tensorflow.control_dependencies", "tensorflow.assign_add", "time.time", "tensorflow.python.ops...
[((769, 842), 'csv.reader', 'csv.reader', (['file'], {'delimiter': '""","""', 'quotechar': '"""\\""""', 'quoting': 'csv.QUOTE_MINIMAL'}), '(file, delimiter=\',\', quotechar=\'"\', quoting=csv.QUOTE_MINIMAL)\n', (779, 842), False, 'import csv\n'), ((1423, 1439), 'tensorflow.ConfigProto', 'tf.ConfigProto', ([], {}), '()\...
# -*- coding: utf-8 -*- """ Created on Fri Aug 27 15:16:34 2021 @author: ag """ import numpy as np import matplotlib.pyplot as plt from matplotlib.patches import Ellipse import matplotlib.transforms as transforms import re def confidence_ellipse(x, y, n_std=1.0, weights=None, ax=None, facecolor='none', **kwargs): ...
[ "numpy.mean", "numpy.all", "numpy.sqrt", "matplotlib.pyplot.gca", "numpy.argsort", "numpy.array", "numpy.sum", "numpy.cov", "matplotlib.transforms.Affine2D", "numpy.cumsum", "numpy.interp", "re.sub", "matplotlib.patches.Ellipse" ]
[((1466, 1486), 'numpy.sqrt', 'np.sqrt', (['(1 + pearson)'], {}), '(1 + pearson)\n', (1473, 1486), True, 'import numpy as np\n'), ((1506, 1526), 'numpy.sqrt', 'np.sqrt', (['(1 - pearson)'], {}), '(1 - pearson)\n', (1513, 1526), True, 'import numpy as np\n'), ((1541, 1674), 'matplotlib.patches.Ellipse', 'Ellipse', (['(0...
import sys import numpy as np rng = np.random.default_rng() # dt = np.dtype('i,i,i,i,i,i,i,i,i,i,U16,U16,U16,U16,f,f,f,f,f,f,f,f') nrows = 2000000 filename = 'data/bigmixed.csv' print("Generating {}".format(filename)) with open(filename, 'w') as f: for k in range(nrows): values1 = rng.integers(1, 1000...
[ "sys.stdout.flush", "numpy.random.default_rng" ]
[((39, 62), 'numpy.random.default_rng', 'np.random.default_rng', ([], {}), '()\n', (60, 62), True, 'import numpy as np\n'), ((723, 741), 'sys.stdout.flush', 'sys.stdout.flush', ([], {}), '()\n', (739, 741), False, 'import sys\n')]
#!/usr/bin/env python3 import argparse import re import sys import zipfile import numpy as np import bert_wrapper if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("input_conllu", type=str, help="Input CoNLL-U file") parser.add_argument("output_npz", type=str, help="Output...
[ "zipfile.ZipFile", "argparse.ArgumentParser", "bert_wrapper.BertWrapper", "re.match", "numpy.save" ]
[((157, 182), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {}), '()\n', (180, 182), False, 'import argparse\n'), ((1946, 2147), 'bert_wrapper.BertWrapper', 'bert_wrapper.BertWrapper', ([], {'language': 'args.language', 'size': 'args.size', 'casing': 'args.casing', 'layer_indices': 'args.layer_indices', 'w...
import sys, math, numpy, struct import matplotlib.pyplot as plt class readBinaryModels(object): '''Class for reading binary models''' def __init__(self, fil): '''Initialize''' super(readBinaryModels, self).__init__() self.fread = open(fil, "rb") self.head = None self.m...
[ "matplotlib.pyplot.ylabel", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.plot", "numpy.array", "struct.unpack", "matplotlib.pyplot.ylim", "math.log10", "matplotlib.pyplot.yscale", "matplotlib.pyplot.show" ]
[((7333, 7343), 'matplotlib.pyplot.show', 'plt.show', ([], {}), '()\n', (7341, 7343), True, 'import matplotlib.pyplot as plt\n'), ((6152, 6172), 'numpy.array', 'numpy.array', (['evolEps'], {}), '(evolEps)\n', (6163, 6172), False, 'import sys, math, numpy, struct\n'), ((6204, 6224), 'numpy.array', 'numpy.array', (['evol...
from sklearn.linear_model import LinearRegression from sklearn.preprocessing import PolynomialFeatures from sklearn.metrics import mean_squared_error, r2_score from sklearn.pipeline import make_pipeline import matplotlib.pyplot as plt import numpy as np import random #=================================================...
[ "sklearn.preprocessing.PolynomialFeatures", "numpy.asarray", "matplotlib.pyplot.scatter", "sklearn.linear_model.LinearRegression", "matplotlib.pyplot.show" ]
[((1680, 1696), 'numpy.asarray', 'np.asarray', (['days'], {}), '(days)\n', (1690, 1696), True, 'import numpy as np\n'), ((1705, 1722), 'numpy.asarray', 'np.asarray', (['cases'], {}), '(cases)\n', (1715, 1722), True, 'import numpy as np\n'), ((1782, 1806), 'matplotlib.pyplot.scatter', 'plt.scatter', (['days', 'cases'], ...
# Copyright 2020 Huawei Technologies Co., Ltd # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to...
[ "mindspore.context.set_context", "numpy.array", "scipy.stats.poisson", "mindspore.nn.probability.distribution.Poisson", "mindspore.Tensor" ]
[((924, 992), 'mindspore.context.set_context', 'context.set_context', ([], {'mode': 'context.GRAPH_MODE', 'device_target': '"""Ascend"""'}), "(mode=context.GRAPH_MODE, device_target='Ascend')\n", (943, 992), True, 'import mindspore.context as context\n'), ((1334, 1355), 'scipy.stats.poisson', 'stats.poisson', ([], {'mu...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- from nltk.tokenize import TweetTokenizer from nltk.stem.snowball import SnowballStemmer import numpy as np from collections import defaultdict,Counter import logging as logger from nortok.stopwords import get_norwegian_stopwords import pickle import gzip def dd_def(): ...
[ "nltk.tokenize.TweetTokenizer", "pickle.dump", "gzip.open", "pickle.load", "collections.Counter", "nltk.stem.snowball.SnowballStemmer", "numpy.zeros", "collections.defaultdict", "nortok.stopwords.get_norwegian_stopwords" ]
[((494, 503), 'collections.Counter', 'Counter', ([], {}), '()\n', (501, 503), False, 'from collections import defaultdict, Counter\n'), ((533, 552), 'collections.defaultdict', 'defaultdict', (['dd_def'], {}), '(dd_def)\n', (544, 552), False, 'from collections import defaultdict, Counter\n'), ((1159, 1178), 'collections...
import numpy.testing as test import numpy as np from unittest import TestCase from PyFVCOM.ocean import * class OceanToolsTest(TestCase): def setUp(self): """ Make a set of data for the various ocean tools functions """ self.lat = 30 self.z = np.array(9712.02) self.t = np.array(4...
[ "numpy.array", "numpy.testing.assert_equal", "numpy.arange", "numpy.testing.assert_almost_equal" ]
[((275, 292), 'numpy.array', 'np.array', (['(9712.02)'], {}), '(9712.02)\n', (283, 292), True, 'import numpy as np\n'), ((310, 322), 'numpy.array', 'np.array', (['(40)'], {}), '(40)\n', (318, 322), True, 'import numpy as np\n'), ((340, 352), 'numpy.array', 'np.array', (['(40)'], {}), '(40)\n', (348, 352), True, 'import...
from torch.utils.data import Sampler import numpy as np def get_chunks(l, n): for i in range(0, len(l), n): yield l[i:i + n] def flatten(l): return [item for sublist in l for item in sublist] class LengthSortSampler(Sampler): def __init__(self, data_source, bs): super().__init__(data_s...
[ "numpy.argsort", "numpy.random.shuffle" ]
[((704, 733), 'numpy.random.shuffle', 'np.random.shuffle', (['chunk_inds'], {}), '(chunk_inds)\n', (721, 733), True, 'import numpy as np\n'), ((576, 595), 'numpy.argsort', 'np.argsort', (['lengths'], {}), '(lengths)\n', (586, 595), True, 'import numpy as np\n')]
import math import numpy as np class Strategy: """Options strategy class. Takes in a number of `StrategyLeg`'s (option contracts), and filters that determine entry and exit conditions. """ def __init__(self, schema): self.schema = schema self.legs = [] self.conditions = []...
[ "numpy.sign" ]
[((2077, 2096), 'numpy.sign', 'np.sign', (['entry_cost'], {}), '(entry_cost)\n', (2084, 2096), True, 'import numpy as np\n')]
import os, math import numpy as np import pandas as pd import matplotlib.pyplot as plt #from matplotlib.collections import PatchCollection from sklearn import linear_model from pandas.plotting import register_matplotlib_converters register_matplotlib_converters() from importlib import reload # Constants #files = ['tim...
[ "pandas.read_csv", "numpy.log", "math.sqrt", "math.log", "numpy.array", "pandas.plotting.register_matplotlib_converters", "math.exp", "math.log10", "pandas.to_datetime", "numpy.arange", "matplotlib.pyplot.close", "pandas.DataFrame", "matplotlib.pyplot.savefig", "matplotlib.pyplot.gcf", "...
[((231, 263), 'pandas.plotting.register_matplotlib_converters', 'register_matplotlib_converters', ([], {}), '()\n', (261, 263), False, 'from pandas.plotting import register_matplotlib_converters\n'), ((13594, 13643), 'sklearn.linear_model.LinearRegression', 'linear_model.LinearRegression', ([], {'fit_intercept': '(True...
#importing some useful packages import matplotlib.pyplot as plt import matplotlib.image as mpimg import numpy as np import cv2 import os import pickle import glob import sys class ParameterFinder: def __init__(self, image, _h_channel_low=0, _h_channel_high=255, _l_channel_low=0, _l_channel_high=255, ...
[ "cv2.imshow", "cv2.destroyAllWindows", "numpy.arctan2", "numpy.max", "cv2.waitKey", "cv2.blur", "glob.glob", "pickle.load", "cv2.cvtColor", "cv2.createTrackbar", "cv2.imread", "cv2.namedWindow", "numpy.set_printoptions", "cv2.imwrite", "pickle.dump", "numpy.power", "cv2.destroyWindow...
[((16370, 16391), 'os.chdir', 'os.chdir', (['WORKING_DIR'], {}), '(WORKING_DIR)\n', (16378, 16391), False, 'import os\n'), ((16442, 16462), 'cv2.imread', 'cv2.imread', (['FILENAME'], {}), '(FILENAME)\n', (16452, 16462), False, 'import cv2\n'), ((16743, 16767), 'cv2.imshow', 'cv2.imshow', (['"""input"""', 'IMG'], {}), "...
#!/usr/bin/env python3 # Copyright (c) 2021 <NAME> # # This software is released under the MIT License. # https://opensource.org/licenses/MIT """This module does comparison of two images""" import argparse import os from io import BytesIO from typing import Union, List, Tuple from pathlib import Path import numpy a...
[ "PIL.Image.fromarray", "PIL.Image.open", "emrtd_face_access.print_to_sg.SetInterval", "argparse.ArgumentParser", "pathlib.Path", "io.BytesIO", "os.path.join", "os.path.isfile", "numpy.array", "face_recognition.face_distance", "face_recognition.face_encodings", "cv2.cv2.dnn.blobFromImage", "c...
[((455, 468), 'emrtd_face_access.print_to_sg.SetInterval', 'SetInterval', ([], {}), '()\n', (466, 468), False, 'from emrtd_face_access.print_to_sg import SetInterval\n'), ((1153, 1202), 'cv2.cv2.dnn.readNetFromCaffe', 'cv2.dnn.readNetFromCaffe', (['config_file', 'model_file'], {}), '(config_file, model_file)\n', (1177,...
from typing import NoReturn from ...base import BaseEstimator import numpy as np from numpy.linalg import det, inv from scipy.stats import multivariate_normal class LDA(BaseEstimator): """ Linear Discriminant Analysis (LDA) classifier Attributes ---------- self.classes_ : np.ndarray of shape (n_c...
[ "numpy.unique", "scipy.stats.multivariate_normal.pdf", "numpy.log", "numpy.array", "numpy.zeros", "numpy.linalg.inv", "numpy.outer" ]
[((1651, 1683), 'numpy.unique', 'np.unique', (['y'], {'return_counts': '(True)'}), '(y, return_counts=True)\n', (1660, 1683), True, 'import numpy as np\n'), ((2212, 2246), 'numpy.zeros', 'np.zeros', (['(X.shape[1], X.shape[1])'], {}), '((X.shape[1], X.shape[1]))\n', (2220, 2246), True, 'import numpy as np\n'), ((2491, ...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Oct 1 08:04:50 2019 @author: alexandradarmon """ import random import numpy as np import seaborn as sns import matplotlib.pyplot as plt from punctuation.config import options from punctuation.visualisation.heatmap_functions import heatmap, annotate_he...
[ "matplotlib.pyplot.hist", "matplotlib.pyplot.ylabel", "numpy.array", "matplotlib.pyplot.axvline", "numpy.arange", "numpy.histogram", "matplotlib.pyplot.xlabel", "matplotlib.pyplot.plot", "webcolors.hex_to_rgb", "numpy.linspace", "matplotlib.pyplot.yticks", "matplotlib.pyplot.ylim", "matplotl...
[((1343, 1356), 'webcolors.hex_to_rgb', 'hex_to_rgb', (['i'], {}), '(i)\n', (1353, 1356), False, 'from webcolors import hex_to_rgb\n'), ((2515, 2536), 'numpy.arange', 'np.arange', (['(0)', '(1)', '(0.01)'], {}), '(0, 1, 0.01)\n', (2524, 2536), True, 'import numpy as np\n'), ((3412, 3437), 'numpy.arange', 'np.arange', (...
import argparse import os import numpy as np import torch import torch.optim as optim from torch.utils.data import DataLoader, TensorDataset from torchvision import transforms from torchvision import datasets from utils.lib import * from utils.pgd_attack import * from models.resnet import ResNet def test(model, data...
[ "os.path.exists", "argparse.ArgumentParser", "os.makedirs", "torch.load", "os.path.join", "torch.Tensor", "torch.utils.data.TensorDataset", "torchvision.transforms.RandomHorizontalFlip", "torchvision.transforms.RandomCrop", "numpy.array", "torchvision.datasets.CIFAR10", "torch.sum", "torch.u...
[((813, 913), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""Generate augmented training dataset and extract features"""'}), "(description=\n 'Generate augmented training dataset and extract features')\n", (836, 913), False, 'import argparse\n'), ((1467, 1506), 'os.path.join', 'os.pat...