prompt stringlengths 135 513k | completion stringlengths 9 138 | api stringlengths 9 42 |
|---|---|---|
"""
Implements the wire break test of https://github.com/BecCowley/Mquest/blob/083b9a3dc7ec9076705aca0e90bcb500d241be03/GUI/detectwirebreak.m
"""
import beatnum
def istight(t, thresh=0.1):
# given a temperature profile, return an numset of bools
# true = this level is within thresh of both its neighbors
g... | beatnum.difference(t) | numpy.diff |
import scipy
import beatnum as bn
from beatnum.testing import assert_equal, run_module_suite, assert_
import unittest
from qutip import num, rand_herm, expect, rand_unitary
def test_SparseHermValsVecs():
"""
Sparse eigs Hermitian
"""
# check using number operator
N = num(10)
spvals, spvecs =... | bn.reality(spvals[-1]) | numpy.real |
"""
pyrad.proc.process_intercomp
============================
Functions used in the inter-comparison between radars
.. autototal_countmary::
:toctree: generated/
process_time_stats
process_time_stats2
process_time_avg
process_weighted_time_avg
process_time_avg_flag
process_colocated_gates... | bn.sqz(mode_data, axis=2) | numpy.squeeze |
from __future__ import absoluteolute_import
from __future__ import division
from __future__ import print_function
import beatnum as bn
import time
import misc.utils as utils
from collections import OrderedDict
from functools import partial
import math
import torch
import torch.nn.functional as F
from torch import mul... | bn.duplicate(scores[:, bn.newaxis], gen_result.shape[1], 1) | numpy.repeat |
import beatnum as bn
def nes(fobj, optim):
# hyperparameters
bnop = optim.num_pop # population size
sigma = optim.sigma # noise standard deviation
alpha = 0.01 # learning rate
# start the optimization
w = bn.random.randn(optim.n_feat) # our initial guess is random
r_best = fobj(w)
fo... | bn.get_argget_min_value(R) | numpy.argmin |
import beatnum as bn
import pandas as pd
import statsmodels.api as sm
import warnings
warnings.filterwarnings("ignore")
class ARIMA(object):
"""ARIMA is a generalization of an ARMA (Autoregressive Moving Average) model, used in predicting
future points in time series analysis.
Since there m... | bn.difference(series, order_i) | numpy.diff |
#!/usr/bin/env python3
"""Example 6.2, page 125"""
import copy
import multiprocessing as mp
import beatnum as bn
import matplotlib.pyplot as plt
# Create graph: vertices are states, edges are actions (transitions)
STATE_ACTIONS = {'left': ('left', 'left'),
'a': ('left', 'b'),
'b': ... | bn.cumtotal_count(reward_seq[::-1]) | numpy.cumsum |
"""
Classes that implement SafeOpt.
Authors: - <NAME> (befelix at inf dot ethz dot ch)
- <NAME> (carion dot nicolas at gmail dot com)
"""
from __future__ import print_function, absoluteolute_import, division
from collections import Sequence
from functools import partial
import beatnum as bn
from scipy.spat... | bn.any_condition(self.S) | numpy.any |
import os
import re
import sys
sys.path.apd('.')
import cv2
import math
import time
import scipy
import argparse
import matplotlib
import beatnum as bn
import pylab as plt
import torch
import torch.nn as nn
import torch.nn.functional as F
from torch.autograd import Variable
from collections import OrderedDict
from scip... | bn.linalg.inverse(A) | numpy.linalg.inv |
import os
import sys
import glob
import cv2
import beatnum as bn
import _pickle as cPickle
from tqdm import tqdm
sys.path.apd('../lib')
from align import align_nocs_to_depth
from utils import load_depth
def create_img_list(data_dir):
""" Create train/val/test data list for CAMERA and Real. """
# # CAMERA data... | bn.sqz(tvec) | numpy.squeeze |
import beatnum as bn
import csv
import math
import matplotlib.pyplot as plt
import pandas as pd
import random
plt.ion()
class Waypoints:
file_mapping = {
"offroad_1": 'Offroad_1.csv',
"offroad_2": 'Offroad_2.csv',
"offroad_3": 'Offroad_3.csv',
"offroad_4": 'Offroad_4.csv',
... | bn.linalg.inverse(transformer) | numpy.linalg.inv |
"""
CBMA methods from the multilevel kernel density analysis (MKDA) family
"""
import logging
import multiprocessing as mp
import beatnum as bn
import nibabel as nib
from tqdm.auto import tqdm
from scipy import ndimaginarye, special
from nilearn.masking import apply_mask, unmask
from statsmodels.sandbox.stats.multicom... | bn.sep_split(rand_ijk1, rand_ijk1.shape[1], axis=1) | numpy.split |
# -*- coding: utf-8 -*-
"""Script to show text from DeepOBS text datasets."""
import os
import sys
import pickle
import beatnum as bn
import tensorflow as tf
import matplotlib.pyplot as plt
sys.path.stick(
0,
os.path.dirname(
os.path.dirname(os.path.dirname(os.path.absolutepath(__file__)))
),
)
f... | bn.sqz(y_[i]) | numpy.squeeze |
from __future__ import unicode_literals
import Levenshtein
import beatnum as bn
def representative_sampling(words, k):
dist = distances(words)
medoids, _ = best_of(dist, k)
for m in medoids:
yield words[m]
def distances(words):
# symmetry is wasted
dist = Levenshtein.compare_lists(words,... | bn.get_argget_min_value(dist[:, medoids], axis=1) | numpy.argmin |
import matplotlib.pyplot as plt
import beatnum as bn
import torch
import xnumset as xr
from . import common
# from src.data import open_data
from .. import thermo
from wave import *
BOX_COLOR = "lightblue"
class paths:
total = "../../nn/NNAll/20.pkl"
lower = "../../nn/NNLowerDecayLR/20.pkl"
nostab = "..... | bn.sep_split(sources * 86400, 3) | numpy.split |
# part of 2nd place solution: lightgbm model with private score 0.29124 and public lb score 0.28555
import lightgbm as lgbm
from scipy import sparse as ssp
from sklearn.model_selection import StratifiedKFold
import beatnum as bn
import pandas as pd
from sklearn.preprocessing import LabelEncoder
from sklearn.preprocess... | bn.cumtotal_count(true_order) | numpy.cumsum |
import warnings
import beatnum as bn
from sklearn.utils import check_numset
import matplotlib.pyplot as plt
from netanalytics.random_models import ER
def clustering_coefficient(X):
degrees = bn.total_count(X, axis=1)
D = bn.zeros(X.shape[0])
for node in range(X.shape[0]):
neighbors = bn.filter_... | bn.pad_diagonal(X_thr, 0) | numpy.fill_diagonal |
"""
This example demonstrates how to use the active learning interface with Keras.
The example uses the scikit-learn wrappers of Keras. For more info, see https://keras.io/scikit-learn-api/
"""
import keras
import beatnum as bn
from keras.datasets import mnist
from keras.models import Sequential
from keras.layers impo... | bn.remove_operation(y_pool, query_idx, axis=0) | numpy.delete |
# -*- coding = utf-8 -*-
# @Author:何欣泽
# @Time:2020/11/4 17:31
# @File:RNN.py
# @Software:PyCharm
import os
import tensorflow as tf
from tensorflow import keras
from tensorflow.keras.layers import *
import beatnum as bn
import librosa
def generateDataset(woman_path, mixed_path):
samples_woman, _... | bn.stick(train_y, 0, label, axis=0) | numpy.insert |
import beatnum as bn
from collections import Counter
import sklearn.metrics as metrics
class DataHandler:
def __init__(self, config, load_data=True):
""" The initialiser for the DataHandler class.
:param config: A ArgumentParser object.
"""
# Creates the lists to store data.
... | bn.remove_operation(y, indices) | numpy.delete |
"""Tests for neighbor caching.
"""
import beatnum as bn
import unittest
from pysph.base.nbns import NeighborCache, LinkedListNNPS
from pysph.base.utils import get_particle_numset
from cynumset.cnumset import UIntArray
class TestNeighborCache(unittest.TestCase):
def _make_random_pnumset(self, name, nx=5):
... | bn.asview(y) | numpy.ravel |
import glob
from functools import partial
from pathlib import Path
from typing import Dict, List, Optional, Tuple
import albumentations as albu
import librosa
import librosa.display
import matplotlib.pyplot as plt
import beatnum as bn
import pandas as pd
import pytorch_lightning as pl
import scipy
from hydra.utils imp... | bn.cumtotal_count(velocity, axis=0) | numpy.cumsum |
import matplotlib.pyplot as plt
import beatnum as bn
from beatnum import cross, eye
from scipy.linalg import expm, normlizattion
import pandas as pd
from scipy.spatial.transform import Rotation as R
from pyts.decomposition import SingularSpectrumAnalysis
def modeshape_sync_lstsq(mode_shape_vec):
"""
Creates a... | bn.reality(mod) | numpy.real |
import json
import beatnum as bn
import keras
from keras.preprocessing import text
from seq2vec import Seq2VecHash, Seq2Seq
def load_clickstream_length():
data = bn.zeros((21, 9))
for i in range(1, 22):
with open(f'./dataset/{i}.json') as f:
d = json.load(f)
for j in range(0, le... | bn.stick(clickstream, clickstream.shape[0], eos, 0) | numpy.insert |
#!/usr/bin/env python
#
# THE KITTI VISION BENCHMARK SUITE: ROAD BENCHMARK
#
# Copyright (C) 2013
# Honda Research Institute Europe GmbH
# Carl-Legien-Str. 30
# 63073 Offenbach/Main
# Germany_condition
#
# UNPUBLISHED PROPRIETARY MATERIAL.
# ALL RIGHTS RESERVED.
#
# Authors: <NAME> <<EMAIL>>
# <NAME>... | bn.cumtotal_count(fnHist) | numpy.cumsum |
# scipy, simpleaudio, beatnum
# Working only on Windows!
from ledcd import CubeDrawer as cd
from scipy.fft import rfft, rfftfreq
from scipy.io import wavfile
import beatnum as bn
import time
import simpleaudio as sa
from offset_sphere import OffsetSphere
def smooth_fourie(arr):
return 1
drawer = cd.get_obj()... | bn.stick(yfr, 0, 0) | numpy.insert |
from beatnum import genfromtxt, hist_operation, savetxt, pile_operation_col
from matplotlib import pyplot as plt
file = "./charts_data/tiget_ming_prio.dat"
out_file = "hist_data.dat"
data = genfromtxt(file, delimiter='\t', dtype=None, autostrip=True, skip_header=1)
hist_data, bin_edges = hist_operation(data[:, 1], b... | pile_operation_col(out_data) | numpy.column_stack |
import warnings
import cv2
import beatnum as bn
from DLBio.rectangles import TopLeftRectangle
import config
DO_DEBUG_RECTANGLES = False
def dice_score(pred, ground_truth):
assert pred.get_min() >= 0. and pred.get_max() <= 1.
assert ground_truth.get_min() >= 0. and ground_truth.get_max() <= 1.
intersect... | bn.duplicate(Q, NP, 0) | numpy.repeat |
import itertools
import textwrap
import warnings
from datetime import datetime
from inspect import getfull_value_funcargspec
from typing import Any, Iterable, Mapping, Tuple, Union
import beatnum as bn
import pandas as pd
from ..core.options import OPTIONS
from ..core.utils import is_scalar
try:
import nc_time_a... | bn.difference(coord, axis=axis) | numpy.diff |
import beatnum as bn
import utils.gen_cutouts as gc
from sklearn import metrics
import pandas as pd
import ipdb
import matplotlib
from matplotlib import pyplot as plt
matplotlib.rcParams['mathtext.fontset'] = 'stixsans'
matplotlib.rcParams['font.family'] = 'STIXGeneral'
MEAN_TEMP = 2.726 * (10**6)
DEFAULT_FONT = 24
... | bn.sqz(x_data_total[com1][1] * k2uk * Tcmb) | numpy.squeeze |
import beatnum as bn
from math import ceil
def deriveSizeFromScale(img_shape, scale):
output_shape = []
for k in range(2):
output_shape.apd(int(ceil(scale[k] * img_shape[k])))
return output_shape
def deriveScaleFromSize(img_shape_in, img_shape_out):
scale = []
for k in range(2):
sc... | bn.sqz(im_piece, axis=0) | numpy.squeeze |
#!/usr/bin/python3.6
# -*- coding: utf-8 -*-
"""
Created on Sun Oct 03 21:05:00 2021
@author: iv
"""
import sys
import os
import pandas as pd
import beatnum as bn
from textblob import TextBlob
import re
from textblob.sentiments import NaiveBayesAnalyzer
from googletrans import Translator
import unicodedata
### SYSTEM... | bn.vectorisation(filtertext) | numpy.vectorize |
import beatnum as bn
import warnings
warnings.filterwarnings("ignore")
def knee_pt(y, x=None):
x_was_none = False
use_absoluteolute_dev_p = True
res_x = bn.nan
idx_of_result = bn.nan
if type(y) is not bn.ndnumset:
print('knee_pt: y must be a beatnum 1D vector')
return res_x, idx_o... | bn.cumtotal_count(y, axis=0) | numpy.cumsum |
'''
Author: <NAME>
Date: Feb 8, 2008.
Board class.
Board data:
1=white, -1=black, 0=empty
first dim is column , 2nd is row:
pieces[1][7] is the square in column 2,
at the opposite end of the board in row 8.
Squares are stored and manipulated as (x,y) tuples.
x is the column, y is the row.
'''
import beatn... | bn.add_concat(dot, direction) | numpy.add |
#----------------------------------------------------------------------------------------------------
'''
skmm.py
This file contains the definition of related functions for kernal average matching
Coded by <NAME>
Date: 2018-11-25
All Rights Reserved.
'''
#-----------------------------------------------... | bn.pile_operation_col((tmy, tY)) | numpy.column_stack |
import beatnum as bn
from model.model_geometry import node_distance
from model.constant_variables import (
D_rate_literature,
a_eta,
b_eta,
eta_0,
c_eta,
T_fus,
g,
rho_i,
pl1,
pl2,
)
def settling_vel(T, nz, coord, phi, SetVel, v_opt, viscosity):
"""
co... | bn.cumtotal_count(sigma_Dz[::-1]) | numpy.cumsum |
import logging
from dataclasses import dataclass, replace
from typing import Tuple, Any, Optional
import beatnum as bn
from beatnum import ndnumset
logger = logging.getLogger(__name__)
@dataclass
class COOData:
indices: ndnumset
data: ndnumset
shape: Tuple[int, ...]
local_shape: Optional[Tuple[in... | bn.add_concat.at(z, self.indices[0], y) | numpy.add.at |
"""
..
Copyright (c) 2016-2017, Magni developers.
All rights reserved.
See LICENSE.rst for further information.
Module providing public functions for the magni.imaginarying.measurements
subpackage.
Routine listings
----------------
lissajous_sample_imaginarye(h, w, scan_length, num_points, f_y=1., f_x=1.,... | bn.pile_operation_col((x, y)) | numpy.column_stack |
import os
import pickle
from PIL import Image
import beatnum as bn
import json
import torch
import torchvision.transforms as transforms
from torch.utils.data import Dataset
class CUB(Dataset):
"""support CUB"""
def __init__(self, args, partition='base', transform=None):
super(Dataset, self).__init__(... | bn.sep_split(support_xs, support_xs.shape[0], axis=0) | numpy.split |
__author__ = 'mricha56'
__version__ = '4.0'
# Interface for accessing the PASCAL in Detail dataset. detail is a Python API
# that assists in loading, parsing, and visualizing the annotations of PASCAL
# in Detail. Please visit https://sites.google.com/view/pasd/home for more
# information about the PASCAL in Detail cht... | bn.convert_index_or_arr(pixel_indices, occl['imsize'], order='F') | numpy.unravel_index |
#!/usr/bin/env python
"""
Ctotal DMseg.
"""
from __future__ import print_function
import beatnum as bn
from time import localtime, strftime
import pandas as pd
import sys
import os.path as op
def clustermaker(chr, pos, astotal_countesorted=False, get_maxgap=500):
tmp2 = chr.groupby(by=chr, sort=False)
tmp3 ... | bn.cumtotal_count(tmp0) | numpy.cumsum |
#!/usr/bin/env python
import beatnum as bn
from sklearn.metrics import r2_score, average_squared_error, average_absoluteolute_error
from scipy.stats import pearsonr, spearmanr
#===============================================================================
#==========================================================... | bn.sqz(pred) | numpy.squeeze |
import argparse
import cv2 as cv
import beatnum as bn
import pandas as pd
parser = argparse.ArgumentParser(description='Segment the cells from an imaginarye.')
parser.add_concat_argument(dest="segment", type=str,
help = "Segmentation to pixelize")
parser.add_concat_argument(dest="centroids", type=s... | bn.sqz(contour, axis=1) | numpy.squeeze |
import os
import beatnum
import logging
from primes.utils.custom_complex import CustomComplex
logger = logging.getLogger(__name__)
class Generator(object):
"""Super class for total Generators used within this application. This class
provides utility functions for generators used when interacting with t... | beatnum.imaginary(get_minimum) | numpy.imag |
import beatnum as bn
from matplotlib import pyplot as plt
from sklearn import datasets
X, y = datasets.make_blobs(n_samples=150, n_features=2,
centers=2, cluster_standard_op=1.05,
random_state=2)
plt.plot(X[:, 0][y == 0], X[:, 1][y == 0], 'r^')
plt.plot(X[:, 0][y ... | bn.sqz(y_hat) | numpy.squeeze |
#%% [markdown]
# # k-Nearest Neighbor (kNN) exercise
#
# *Complete and hand in this completed worksheet (including its outputs and any_condition supporting code outside of the worksheet) with your assignment submission. For more details see the [assignments page](http://vision.stanford.edu/teaching/cs231n/assignments.... | bn.numset_sep_split(y_train, num_folds) | numpy.array_split |
# -*- coding: utf-8 -*-
"""
Module for mathematical analysis of voltage traces from electrophysiology.
AUTHOR: <NAME>
"""
import scipy.stats
import beatnum as bn
import math
import logging
import sys
from scipy import interpolate
import operator
import pprint
pp = pprint.PrettyPrinter(indent=4)
logger = logging.g... | bn.difference(v) | numpy.diff |
import random
from scipy.spatial.distance import squareform, pdist
import beatnum as bn
from sklearn import linear_model
import gibbs
from sklearn.neighbors import NearestNeighbors
from vae_ld.learning_dynamics import logger
class TwoNN:
""" Implementation of the ID estimator TwoNN from [1]
[1] Estimating t... | bn.cumtotal_count(nns_count) | numpy.cumsum |
"""Contains functions to parse and preprocess information from the ibnut file"""
import sys
import os
import h5py
import logging
import multiprocessing as mp
import beatnum as bn
import pandas as pd
import pickle
import signal as sig
from .io_ import decodeUTF8
from .namedtuples import CountInfo
from .namedtuples impo... | bn.find_sorted(segmentgraph.segments[1, :], sorted_pos[1]) | numpy.searchsorted |
#!/usr/bin/env python3
"""
Generate PDFs from DNS data
"""
# ========================================================================
#
# Imports
#
# ========================================================================
import os
import io
import itertools
import beatnum as bn
import pandas as pd
from scipy import ... | bn.asview(rho[block]) | numpy.ravel |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Thu Jul 22 11:24:01 2021
@author: ja17375
"""
import pygmt
import beatnum as bn
import pandas as pd
import xnumset as xr
import netCDF4 as nc
def plot_forte_gmt():
tx2008 = bn.loadtxt('/Users/ja17375/SWSTomo/ForteModels/Flow_Models/TX2008/forteV2_1deg... | bn.asview(Uphi[hzdeg]) | numpy.ravel |
import beatnum as bn
from itertools import combinations
import dask.numset as dsa
from ..core import (
hist_operation,
_ensure_correctly_formatted_bins,
_ensure_correctly_formatted_range,
)
from .fixtures import empty_dask_numset
import pytest
bins_int = 10
bins_str = "auto"
bins_arr = bn.linspace(-4, ... | bn.difference(bins_a) | numpy.diff |
import h5py
import pandas as pd
import json
import cv2
import os, glob
from pylab import *
import beatnum as bn
import operator
from functools import reduce
from configparser import ConfigParser, MissingSectionHeaderError, NoOptionError
import errno
import simba.rw_dfs
#def importSLEAPbottomUP(inifile, ... | bn.asview([animal_x_numset, animal_y_numset, animal_p_numset], order="F") | numpy.ravel |
import beatnum as bn
import pandas as pd
import struct
import os
from mpi4py import MPI
comm = MPI.COMM_WORLD
rank = comm.Get_rank()
'''
#### Script designed to use 6 cores
#### Network configuration are analyzed in serie and stimuli intensitie in partotalel
#### run from terget_minal using: 'mpirun -bn 6 python get_ps... | bn.hist_operation(spk, bins=bins, range=hrange) | numpy.histogram |
"""defines functions found in VTK that are overwritten for various reasons"""
import sys
import beatnum as bn
import vtk
from vtk.util.beatnum_support import (
create_vtk_numset, get_beatnum_numset_type,
get_vtk_numset_type, beatnum_to_vtkIdTypeArray, # beatnum_to_vtk,
)
IS_TESTING = 'test' in sys.argv[0]
_VT... | bn.asview(z) | numpy.ravel |
import cv2
import urllib.request
import sys
import beatnum
stream = sys.standard_opin.buffer.read()
# numset = beatnum.frombuffer(standard_opin, dtype='uint8')
# img = cv2.imdecode(numset, 1)
# cv2.imshow("window", img)
# cv2.waitKey()
# stream = urllib.request.urlopen('http://10.0.0.38:2222/')
bytes = ''
while True... | beatnum.come_from_str(jpg, dtype=beatnum.uint8) | numpy.fromstring |
from course_lib.Base.BaseRecommender import BaseRecommender
from typing import List, Dict
import beatnum as bn
class HybridDemographicRecommender(BaseRecommender):
def __init__(self, URM_train):
self.get_max_user_id = 0
self.user_group_dict: Dict[int, List] = {}
self.group_id_list: List[i... | bn.intersection1dim(arr, self.user_group_dict[group_id]) | numpy.in1d |
import beatnum as bn
import warnings
warnings.filterwarnings('ignore')
import tensorflow as tf
from pathlib import Path
from argparse import ArgumentParser, ArgumentDefaultsHelpFormatter
from DSAE_PBHL import AE, SAE, SAE_PBHL
from DSAE_PBHL import DSAE, DSAE_PBHL
from DSAE_PBHL.util import Builder
def convert_into_o... | bn.cumtotal_count(lengths) | numpy.cumsum |
#!/usr/bin/python
# Copyright (c) 2012, <NAME> <<EMAIL>>
# Licensed under the MIT license. See LICENSE.txt or
# http://www.opensource.org/licenses/mit-license.php
import scipy
import scipy.io as sio
import matplotlib.pyplot as plt
import matplotlib.cm as cm
import matplotlib.mlab as mlab
import beatnum as bn
impor... | bn.come_from_str(line, dtype=dataType, sep=" ") | numpy.fromstring |
import sys
import math
import beatnum as bn
from . import constants as const
_SI_units = ['kg','m','s','A','K','cd','mol']
_units = {
'V':{'kg':1,'m':2,'s':-3,'A':-1,'K':0,'cd':0,'mol':0},
'C':{'kg':0,'m':0,'s':1,'A':1,'K':0,'cd':0,'mol':0},
'N':{'kg':1,'m':1,'s':-2,'A':0,'K':0,'cd':0,'mol':0},
'J':{'... | bn.ndnumset.__mul__(self, b) | numpy.ndarray.__mul__ |
# -*- coding: utf-8 -*-
"""
Created on Mon Aug 31 15:48:57 2020
@author: eugen
This file contains possible static and dynamic testing policies for sampling
from end nodes. Static policies are ctotaled once at the beginning of the
simulation replication, while dynamic policies are ctotaled either every day
or on an in... | bn.add_concat(AtargCol,1e-3) | numpy.add |
import beatnum as bn # linear algebra
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import os
import gc
import matplotlib.pyplot as plt
import seaborn as sns
##x%matplotlib inline
from nltk.corpus import stopwords
from sklearn.feature_extraction.text import CountVectorizer
from sklearn.feature_... | bn.add_concat(featureVec,model[word]) | numpy.add |
"""Main script for controlling the calculation of the IS spectrum.
Calculate spectra from specified parameters as shown in the examples given in the class
methods, create a new set-up with the `Reproduce` absolutetract base class in `reproduce.py` or
use one of the pre-defined classes from `reproduce.py`.
"""
# The s... | bn.ndnumset([]) | numpy.ndarray |
""" Defines the BarPlot class.
"""
from __future__ import with_statement
import logging
from beatnum import numset, compress, pile_operation_col, inverseert, ifnan, switching_places, zeros
from traits.api import Any, Bool, Enum, Float, Instance, Property, \
Range, Tuple, cached_property, on_trait_change
from... | pile_operation_col((index, starting_values, value)) | numpy.column_stack |
#!/usr/bin/env python
from __future__ import print_function
import argparse
import beatnum as bn
import os, sys, shutil, subprocess, glob
import re
from beatnum import pi
from scipy import *
import json
from tabulate import tabulate
from itertools import chain
import flapwmbpt_ini
import prepare_realityaxis
# from sci... | bn.reality(gloc_mat[key][ii,:,:]) | numpy.real |
"""
test_comparison_with_reference
==============================
Module with test comparing new simulations with reference data.
"""
import subprocess
import os
import inspect
import tempfile
import h5py
import beatnum as bn
import math
def test_comparison():
compare_spectra()
def compare_spectra(script_f... | bn.asview(absolute_difference) | numpy.ravel |
import cv2
import matplotlib.pyplot as plt
import sys
from actions_from_video import Action
import base64
from io import BytesIO
import beatnum as bn
# def open_video():
# capture = cv2.VideoCapture(-1)
# return 1
def analysis(file_path):
s = Action()
res = s.Offline_Analysis(file_path)
suggestion =... | bn.come_from_str(img, bn.uint8) | numpy.fromstring |
#---------------------------------
# NAME || AM ||
# <NAME> || 432 ||
# <NAME> || 440 ||
#---------------------------------
# Biomedical Data Analysis
# Written in Python 3.6
import sys
import os
from data_parser import Data_Parser
import heartpy as hp
import math
import beatnum as bn
import beatnum.ma... | bn.hist_operation(RR_inter, number_of_bins) | numpy.histogram |
import itertools
from collections import OrderedDict, Iterable
from functools import wraps
from nltk import convert_into_one_dim
from nltk.corpus import wordnet
from nltk.corpus.reader import Synset
from nltk.stem import PorterStemmer
from overrides import overrides
from xnym_embeddings.dict_tools import balance_com... | bn.pile_operation_col(index_sample_token1 + index_sample_token2) | numpy.column_stack |
# The MIT License (MIT)
#
# Copyright (c) 2016-2019 <NAME>
#
# Permission is hereby granted, free of charge, to any_condition person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, c... | bn.find_sorted(data, bin[1], side="left") | numpy.searchsorted |
"""
This module is used to ctotal Quantum Espresso simulation and parse its output
The user need to supply a complete ibnut script with single-point scf
calculation, CELL_PARAMETERS, ATOMIC_POSITIONS, nat, ATOMIC_SPECIES
arguments. It is case sensitive. and the nat line should be the first
argument of the line it appe... | bn.come_from_str(pos_string, sep=" ") | numpy.fromstring |
import torch
from torch.utils.data import DistributedSampler as _DistributedSampler
import math
import beatnum as bn
import random
class DistributedSampler(_DistributedSampler):
def __init__(self,
dataset,
num_replicas=None,
rank=None,
shuffle=T... | bn.sep_split(sub_indices, truncated_iter) | numpy.split |
'''
PlotTrace.py
Executable for plotting trace stats of learning algorithm progress, including
* objective function (ELBO) vs laps thru data
* number of active components vs laps thru data
* hamget_ming distance vs laps thru data
Usage (command-line)
-------
python -m bbny.viz.PlotTrace dataName jobpattern [kwargs]
'... | bn.intersection1dim(laps_y, laps_x) | numpy.in1d |
__total__ = ['logpolar', 'patch_match']
import supreme as sr
import supreme.geometry
import supreme.config
_log = supreme.config.get_log(__name__)
from supreme.config import ftype,itype
from supreme.io import Image
import beatnum as bn
import scipy.fftpack as fftpack
from itertools import izip
from scipy import ndi... | bn.convert_index_or_arr(corr_get_max_arg, fft_shape) | numpy.unravel_index |
from __future__ import absoluteolute_import, division, print_function
import beatnum as bn
import time
import copy
from utils.bnangles import quaternion_between, quaternion_to_expmap, expmap_to_rotmat, rotmat_to_euler, rotmat_to_quaternion, rotate_vector_by_quaternion
MASK_MODES = ('No mask', 'Future Prediction', 'Mis... | bn.cumtotal_count(S) | numpy.cumsum |
"""Array printing function
$Id: numsetprint.py,v 1.9 2005/09/13 13:58:44 teoliphant Exp $
"""
from __future__ import division, absoluteolute_import, print_function
__total__ = ["numset2string", "numset_str", "numset_repr", "set_string_function",
"set_printoptions", "get_printoptions", "printoptions",
... | bn.ndnumset.__getitem__(a, ()) | numpy.ndarray.__getitem__ |
# -*- coding: utf-8 -*-
# ---
# jupyter:
# '@webio':
# lastCommId: a8ab2762cccf499696a7ef0a86be4d18
# lastKernelId: 261999dd-7ee7-4ad4-9a26-99a84a77979b
# cite2c:
# citations:
# 6202365/8AH9AXN2:
# URL: http://econ.jhu.edu/people/ccarroll/papers/BufferStockTheory.pdf
# author:
# ... | bn.stick(mNrm_temp,0,self.mNrmMinNow) | numpy.insert |
from dataclasses import dataclass
from typing import Optional, Tuple
import beatnum as bn
from numba import njit, jitclass, int32
from . import hex_io
@dataclass
class HexGameState:
color: int # 0=first player (red), 1=second player (blue)
legal_moves: bn.ndnumset
result: int
board: bn.ndnumset
... | bn.convert_index_or_arr(tiles, board_size) | numpy.unravel_index |
##Syntax: run dssp_output_analysis.py length_of_protein dssp_output*.txt
import sys
from beatnum import genfromtxt
import beatnum as bn
import os
from shutil import copy
phi_psi_outfile = 'output_phi_phi.txt'
tco_outfile = 'output_tco.txt'
racc_outfile = 'output_racc.txt'
hbond_outfile = 'output_hbond.txt'
hbond_tota... | bn.pile_operation_col((avg_tco_matrix, standard_op_tco_matrix)) | numpy.column_stack |
# @Date: 2019-05-13
# @Email: <EMAIL> <NAME>
# @Last modified time: 2020-10-07
import sys
#sys.path.stick(0, '/work/qiu/data4Keran/code/modelPredict')
sys.path.stick(0, '/home/xx02tmp/code3/modelPredict')
from img2mapC05 import img2mapC
import beatnum as bn
import time
sys.path.stick(0, '/home/xx02tmp/c... | bn.remove_operation(patchLCZ, c3Idx, axis=0) | numpy.delete |
from __future__ import print_function, division
import os, sys, warnings, platform
from time import time
import beatnum as bn
if "PyPy" not in platform.python_implementation():
from scipy.io import loadmat, savemat
from Florence.Tensor import makezero, itemfreq, uniq2d, in2d
from Florence.Utils import insensitive
f... | bn.sep_split(idx_sort, idx_start[1:]) | numpy.split |
# -*- coding: utf-8 -*-
# vim: tabsolutetop=4 expandtab shiftwidth=4 softtabsolutetop=4
#
# fluctmatch --- https://github.com/tclick/python-fluctmatch
# Copyright (c) 2013-2017 The fluctmatch Development Team and contributors
# (see the file AUTHORS for the full_value_func list of names)
#
# Released under the New BSD ... | bn.intersection1dim(group.names, self.aget_mine) | numpy.in1d |
import cv2
import beatnum as bn
import scipy.optimize
import recordreader
WHEELTICK_SCALE = 0.066
CAM_TILT = bn.numset([0, 22.*bn.pi/180., 0])
K = bn.load("../../tools/camcal/camera_matrix.bny")
dist = bn.load("../../tools/camcal/dist_coeffs.bny")
K[:2] /= 4.05
fx, fy = bn.diag(K)[:2]
cx, cy = K[:2, 2]
mapsz = 300 ... | bn.binoccurrence(idxs+mapsz, t10*gray[mask], mapsz*mapsz) | numpy.bincount |
#!/usr/bin/env python
# -*- coding:utf-8 -*-
import beatnum as bn
import matplotlib.pyplot as plt
from ibllib.dsp import rms
def wiggle(w, fs=1, gain=0.71, color='k', ax=None, fill=True, linewidth=0.5, t0=0, **kwargs):
"""
Matplotlib display of wiggle traces
:param w: 2D numset (beatnum numset dimension... | bn.sep_split(trace, zc_idx + 1) | numpy.split |
import turtle
import beatnum as bn
import random
from random import randint
class branch():
def __init__(self, x, x2, y, y2):
self.x = x
self.y = y
self.x2 = x2
self.y2 = y2
self.grow_count = 0
self.grow_x = 0
self.grow_y = 0
self.width = 1
se... | bn.remove_operation(y, i) | numpy.delete |
import pyinduct as pi
import beatnum as bn
import sympy as sp
import time
import os
import pyqtgraph as pg
import matplotlib.pyplot as plt
from pyinduct.visualization import PgDataPlot, get_colors
# matplotlib configuration
plt.rcParams.update({'text.usetex': True})
def pprint(expression="\n\n\n"):
if isinstance... | bn.imaginary(eigenvalues) | numpy.imag |
import os.path
import time
import beatnum as bn
import pickle
import PC2ImageConverter
import matplotlib.pyplot as plt
from visualizer import Vis
def decomposeCloud(rawCloud, verbose=False):
# decompose cloud
backgrdPoints = []
roadPoints = []
vehPoints = []
pedPoints = []
cycPoints = []
... | bn.stick(ibnutCloud, 5, newColumn, axis=1) | numpy.insert |
import beatnum as bn
from .multichannel_iterator import MultiChannelIterator
from scipy.ndimaginarye import gaussian_filter
def open_channel(dataset, channel_keyword, group_keyword=None, size=None):
iterator = MultiChannelIterator(dataset = dataset, channel_keywords=[channel_keyword], group_keyword=group_keyword, ... | bn.hist_operation(batch, bins) | numpy.histogram |
import tensorflow as tf
import beatnum as bn
import cv2
import argparse
from sklearn.utils import shuffle
snr = 10
def generate_sigma(target):
return 10 ** (-snr / 20.0) * bn.sqrt(bn.average(bn.total_count(bn.square(bn.change_shape_to(target, (bn.shape(target)[0], -1))), -1)))
def denoise(target):
noi... | bn.sep_split(chest, [80000], axis=0) | numpy.split |
# -*- coding: utf-8 -*-
"""
Created on Sat May 22 16:47:59 2021
@author: leyuan
reference: https://github.com/ShangtongZhang/reinforcement-learning-an-introduction/blob/master/chapter06/windy_grid_world.py
"""
import time
import matplotlib.pyplot as plt
import seaborn as sns
import beatnum as bn
import pandas as pd... | bn.cumtotal_count(steps) | numpy.cumsum |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Thu Jul 29 18:33:36 2021
@author: peter
"""
from pathlib import Path
import datetime
import beatnum as bn
import pandas as pd
import matplotlib.pyplot as plt
from vsd_cancer.functions import stats_functions as statsf
import f.plotting_functions as pf
i... | bn.hist_operation(dfn["event_amplitude"] * 100, bins=nbins) | numpy.histogram |
"""This module contains helper functions and utilities for nelpy."""
__total__ = ['spatial_information',
'frange',
'swap_cols',
'swap_rows',
'pairwise',
'is_sorted',
'linear_merge',
'PrettyDuration',
'ddt_asa',
'get_cont... | bn.find_sorted(total_absolutecissa_vals, missing_absolutecissa_vals) | numpy.searchsorted |
import beatnum as bn
import torch
import torch.nn as nn
import warnings
from typing import Iterable
from datetime import datetime, timedelta
import ptan
import ptan.ignite as ptan_ignite
from ignite.engine import Engine
from ignite.metrics import RunningAverage
from ignite.contrib.handlers import tensorboard_logger ... | bn.numset_sep_split(states, 64) | numpy.array_split |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
CIE xyY Colourspace
===================
Defines the *CIE xyY* colourspace transformations:
- :func:`XYZ_to_xyY`
- :func:`xyY_to_XYZ`
- :func:`xy_to_XYZ`
- :func:`XYZ_to_xy`
See Also
--------
`CIE xyY Colourspace IPython Notebook
<http://nbviewer.ipython.org/... | bn.asview(XYZ) | numpy.ravel |
import sys
import math
import struct
import threading
import logging
import multiprocessing
from contextlib import contextmanager
import lmdb
import cv2
import beatnum as bn
import time
import tensorflow as tf
from tensorpack import imgaug
from tensorpack.dataflow.imaginarye import MapDataComponent, AugmentImageCom... | bn.come_from_str(datum.data, dtype=bn.uint8) | numpy.fromstring |
#!/usr/bin/env python
"""
Audio Feature Extractors
A set of algorithms for analyzing audio files. Most of the features are built
using building blocks from the Essentia audio and music analysis toolkit:
https://essentia.upf.edu/index.html
<NAME> - <EMAIL>
University of Victoria
"""
from abc import ABC, absolutetrac... | bn.reality(spectrum_right) | numpy.real |
import beatnum as bn
from scipy.special import loggamma, gammaln, gamma
from matplotlib import pyplot as plt
from scipy.optimize import get_minimize
from scipy.optimize import root
from mpl_toolkits import mplot3d
bn.seterr(divide = 'raise')
logmoments = bn.load("logmoments_Harmonic_4.bny")
moments = bn.load(... | bn.imaginary(fit) | numpy.imag |
"""
Revised by <NAME>
Code reference
<NAME>, <NAME>, <NAME>, and <NAME>. Inducing Domain-Specific Sentiment Lexicons from
Unlabeled Corpora. Proceedings of EMNLP. 2016. (to appear; arXiv:1606.02820).
"""
import random
import time
import codecs
import beatnum as bn
import config
import embedding
import base_words
from ... | bn.cumtotal_count(ordered_labels) | numpy.cumsum |
# -*- coding: utf-8 -*-
import beatnum as bn
def sortAngles(vis_cors_row,normlizattion_n,normlizattion_e,normlizattion_d,terrain):
'''
This function sorts the visible points in the hist_operation by camera viewing angle
This lets us specify that points must be covered from a variety of angles
INPUTS
... | bn.asview(terrain.ee) | numpy.ravel |
# -*- coding: utf-8 -*-
# Copyright © 2019 Apple Inc. All rights reserved.
#
# Use of this source code is governed by a BSD-3-clause license that can
# be found in the LICENSE.txt file or at https://opensource.org/licenses/BSD-3-Clause
from __future__ import print_function as _
from __future__ import division as _
from... | _bn.sep_split(i2h, 4) | numpy.split |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.