prompt stringlengths 19 1.03M | completion stringlengths 4 2.12k | api stringlengths 8 90 |
|---|---|---|
from typing import Union, Dict, Tuple, List
import os
import tempfile
import zipfile as zf
import logging
import patsy
import pandas as pd
import numpy as np
import scipy.sparse
import xarray as xr
import dask
import dask.array
try:
import anndata
except ImportError:
anndata = None
logger = logging.getLogge... | pd.read_csv(fpath, header=None, sep=delim) | pandas.read_csv |
# Copyright (c) 2020 Civic Knowledge. This file is licensed under the terms of the
# MIT license included in this distribution as LICENSE
import logging
import re
from collections import defaultdict, deque
from pathlib import Path
from time import time
import pandas as pd
from synpums.util import *
''
_logger = logg... | pd.DataFrame(series) | pandas.DataFrame |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
## Copyright 2015-2021 PyPSA Developers
## You can find the list of PyPSA Developers at
## https://pypsa.readthedocs.io/en/latest/developers.html
## PyPSA is released under the open source MIT License, see
## https://github.com/PyPSA/PyPSA/blob/master/LICENSE.txt
"""
T... | pd.to_numeric(duals['Marginal'], 'coerce') | pandas.to_numeric |
import math
import pandas as pd
import os
from abc import ABC, abstractclassmethod
class Base(ABC):
'''Abstract class'''
def __init__(self, formula, iter = 8, dec_places=4):
self.formula = lambda x: eval(formula)
self.a = int(input('Value of A = '))
self.b = int(input('Value of B = ')... | pd.DataFrame() | pandas.DataFrame |
import pandas as pd
import matplotlib.pyplot as plt
from datetime import datetime, timedelta
plt.style.use('fivethirtyeight')
import warnings
warnings.filterwarnings("ignore")
import os
from stockscraperalpha import getStockInfo
from functions import SMA, EMA,DEMA,MACD,RSI
def show_menu():
return """[1]change stock
[... | pd.DataFrame() | pandas.DataFrame |
"""
This tests whether the Study object was created correctly.
No computation or visualization tests yet.
"""
from __future__ import absolute_import, division, print_function
from __future__ import unicode_literals
from six import iteritems
from collections import Iterable
import itertools
import json
import matplotl... | pdt.assert_equal(study.default_sample_subset, 'all_samples') | pandas.util.testing.assert_equal |
import pandas as pd
import numpy as np
import jinja2
import math
import re
class Plate_Desc:
def __init__(self, title, descrip, serialnum, date_time, datafname):
self.title = title
self.descrip = descrip
self.serialnum = serialnum
self.date_time = date_time
self.datafname =... | pd.Index([]) | pandas.Index |
import numpy as np
import pandas as pd
import os, errno
import datetime
import uuid
import itertools
import yaml
import subprocess
import scipy.sparse as sp
from scipy.spatial.distance import squareform
from sklearn.decomposition.nmf import non_negative_factorization
from sklearn.cluster import KMeans
from sklearn.me... | pd.DataFrame(index=tpm.index) | pandas.DataFrame |
import pandas as pd
import datetime
import numpy as np
from tpau_gtfsutilities.gtfs.gtfssingleton import gtfs as gtfs_singleton
from tpau_gtfsutilities.helpers.datetimehelpers import seconds_since_zero
from tpau_gtfsutilities.helpers.datetimehelpers import seconds_to_military
def get_trip_duration_seconds(gtfs_overr... | pd.concat([min_arrival_times, max_arrival_times], axis=1) | pandas.concat |
# -*- coding: utf-8 -*-
from __future__ import absolute_import, division, print_function
from multiprocess import Pool
import simplejson as json
import six
import sys
from cytoolz import compose
import numpy as np
import pandas as pd
import h5py
from ._util import parse_bins, parse_kv_list_param, parse_field_param
fr... | pd.__version__.split('.') | pandas.__version__.split |
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not u... | pandas.DataFrame() | pandas.DataFrame |
# Author: <NAME> <<EMAIL>>
#
# License: Apache Software License 2.0
"""Tests for Drift package."""
import numpy as np
import pandas as pd
import plotly.graph_objects
import pytest
from sklearn.impute import SimpleImputer
from nannyml.chunk import CountBasedChunker, DefaultChunker, PeriodBasedChunker, SizeBasedC... | pd.testing.assert_frame_equal(expected_drift, drift.data[['key', 'reconstruction_error']]) | pandas.testing.assert_frame_equal |
# pylint: disable=E1101
from datetime import time, datetime
from datetime import timedelta
import numpy as np
from pandas.core.index import Index, Int64Index
from pandas.tseries.frequencies import infer_freq, to_offset
from pandas.tseries.offsets import DateOffset, generate_range, Tick
from pandas.tseries.tools impo... | tools.to_datetime(data) | pandas.tseries.tools.to_datetime |
import sys
import time
import math
import warnings
import numpy as np
import pandas as pd
from os import path
sys.path.append(path.dirname(path.dirname(path.abspath(__file__))))
from fmlc.triggering import triggering
from fmlc.baseclasses import eFMU
from fmlc.stackedclasses import controller_stack
class testcontroll... | pd.isna(df3['b'][0]) | pandas.isna |
# Arithmetic tests for DataFrame/Series/Index/Array classes that should
# behave identically.
# Specifically for datetime64 and datetime64tz dtypes
from datetime import (
datetime,
time,
timedelta,
)
from itertools import (
product,
starmap,
)
import operator
import warnings
import numpy as np
impo... | pd.array(expected, dtype="bool") | pandas.array |
"""
This creates Figure S1 - Full Cytokine plots
"""
import numpy as np
import pandas as pd
import seaborn as sns
from scipy.stats import pearsonr
from tfac.dataImport import form_tensor, import_cytokines
from .common import getSetup
def fig_S1_setup():
tensor, _, patInfo = form_tensor()
plasma, _ = import_cy... | pd.DataFrame(plasma_slice) | pandas.DataFrame |
#!/usr/bin/env python
from gensim.models.doc2vec import Doc2Vec, TaggedDocument
from scipy.spatial.distance import pdist, squareform
from sklearn import datasets, preprocessing, manifold
from nltk.corpus import stopwords
from nltk.stem import PorterStemmer
from nltk.tokenize import word_tokenize
from glob import glob... | pandas.DataFrame(embedding, index=distance.index, columns=["x", "y"]) | pandas.DataFrame |
import pandas as pd
from pandas import DataFrame
from sklearn.ensemble import RandomForestRegressor
from sklearn.feature_selection import f_regression
from sklearn.linear_model import LinearRegression
from sklearn.preprocessing import StandardScaler
from sklearn.svm import SVR, LinearSVR
from metalfi.src.data.dataset ... | pd.concat(data) | pandas.concat |
# %%
import os
import sys
os.chdir(os.path.dirname(os.getcwd())) # make directory one step up the current directory
from pymaid_creds import url, name, password, token
import pymaid
rm = pymaid.CatmaidInstance(url, token, name, password)
import navis
import numpy as np
import pandas as pd
import seaborn as sns
import ... | pd.DataFrame(inputs.values, index = inputs.index, columns = ['axon_input', 'dendrite_input']) | pandas.DataFrame |
import numpy as np
import pandas as pd;
# python list
data = [1,2,3,4,5];
# numpy
ndata = np.array(data);
# pandas
pdata = pd.Series(data);
print(pdata[0]);
print(pdata.values);
print(pdata.index);
pdata2 = pd.Series(data, index=['A','B','C','D','E']);
print(pdata2);
print(pdata2['C']);
# dic
data2 = {'name':'kim'... | pd.Series(data2) | pandas.Series |
"""Functions related to SCOP classes of structures."""
import os
import pandas as pd
SCOP_CLA_LATEST_FILE = 'atom3d/data/metadata/scop-cla-latest.txt'
PDB_CHAIN_SCOP2_UNIPROT_FILE = 'atom3d/data/metadata/pdb_chain_scop2_uniprot.csv'
PDB_CHAIN_SCOP2B_SF_UNIPROT_FILE = 'atom3d/data/metadata/pdb_chain_scop2b_sf_uniprot... | pd.to_numeric(scop2_uniprot['sf-domid']) | pandas.to_numeric |
"""Excel Model"""
__docformat__ = "numpy"
# pylint: disable=abstract-class-instantiated
import logging
import pandas as pd
from openbb_terminal.decorators import log_start_end
from openbb_terminal.rich_config import console
logger = logging.getLogger(__name__)
@log_start_end(log=logger)
def load_configuration(ex... | pd.ExcelWriter(file) | pandas.ExcelWriter |
from typing import List
import numpy as np
import itertools
import sklearn.neighbors
import functools
import joblib
import psutil
import warnings
import pandas as pd
from numpy.typing import NDArray
from ..base import BaseExplainer
from ._cadex_parallel import compute_criterion # type: ignore
class ECE(BaseExplain... | pd.DataFrame(k_subset, columns=self._col_names) | pandas.DataFrame |
#!/usr/bin/env python
# coding=utf-8
"""
@version: 0.1
@author: li
@file: factor_solvency.py
@time: 2019-01-28 11:33
"""
import gc, six
import json
import numpy as np
import pandas as pd
from pandas.io.json import json_normalize
from utilities.calc_tools import CalcTools
from utilities.singleton import Singleton
# fr... | pd.merge(factor_solvency, cash_flow, how='outer', on="security_code") | pandas.merge |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
Description : This code do basic statistical tests (i.e., student t-test, fold change,
Benjamini-Hochberg false discovery rate adjustment) for peak table generated
by MZmine-2.53
Copyright : (c) LemasLab, 02/23/2020
Author : <NAME>
Lic... | pd.isnull(row["row identity (main ID)"]) | pandas.isnull |
import numpy as np
from datetime import timedelta
from distutils.version import LooseVersion
import pandas as pd
import pandas.util.testing as tm
from pandas import to_timedelta
from pandas.util.testing import assert_series_equal, assert_frame_equal
from pandas import (Series, Timedelta, DataFrame, Timestamp, Timedelt... | Timedelta('0 days 01:00:00') | pandas.Timedelta |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
qualify donor data
"""
# %% REQUIRED LIBRARIES
import os
import argparse
import json
import ast
import pandas as pd
import datetime as dt
import numpy as np
# %% USER INPUTS (choices to be made in order to run the code)
codeDescription = "qualify donor data"
parser... | pd.to_datetime(basalData.time) | pandas.to_datetime |
# -*- coding: utf-8 -*-
"""
Created on Wed Mar 25 14:10:00 2020
@author: asweet
"""
import pandas as pd
import numpy as np
from sqlalchemy.types import Integer, Numeric, String, DateTime
from sqlalchemy import create_engine
import urllib.parse
from sys import platform
from abc import ABC, abstractmethod
... | pd.read_csv(url_confirmed) | pandas.read_csv |
from collections import OrderedDict
import datetime
from datetime import timedelta
from io import StringIO
import json
import os
import numpy as np
import pytest
from pandas.compat import is_platform_32bit, is_platform_windows
import pandas.util._test_decorators as td
import pandas as pd
from pandas import DataFrame... | pd.Timestamp("20130101") | pandas.Timestamp |
# -*- coding: utf-8 -*-
"""
Created on Thu Sep 23 08:06:31 2021
@author: bcamc
"""
#%% Import Packages
import matplotlib as mpl
import matplotlib.pyplot as plt
from mpl_toolkits.axes_grid1.inset_locator import InsetPosition, inset_axes
from matplotlib.lines import Line2D
import pandas as pd
import numpy ... | pd.Series(ind, name='ind') | pandas.Series |
# pylint: disable-msg=E1101,W0613,W0603
import os
import copy
from collections import defaultdict
import numpy as np
import pandas.json as _json
from pandas.tslib import iNaT
from pandas.compat import StringIO, long, u
from pandas import compat, isnull
from pandas import Series, DataFrame, to_datetime
from pandas.io.... | compat.iteritems(meta_vals) | pandas.compat.iteritems |
"""
Makes Data Available in Standard Formats. Creates the following data:
df_gene_description DF
cn.GENE_ID (index), cn.GENE_NAME, cn.LENGTH, cn.PRODUCT, cn.START, cn.END, cn.STRAND
Dataframes in dfs_centered_adjusted_read_count
cn.GENE_ID (index), columns: time indices
hypoxia curve DF
cn.SAMPLE, cn.... | pd.read_csv(path) | pandas.read_csv |
import os
import json
import sqlite3 as sql
import pandas as pd
from datetime import datetime, timedelta
import numpy as np
from catboost import CatBoostRegressor
from typing import List
import logging
from constants import (
UTILS_PATH, STREETS_FOR_PREDICTIONS_FILE,
CURRENT_FILENAMES_FILE, SELECT_DATETIME_STR... | pd.to_datetime(df[DATETIME]) | pandas.to_datetime |
import json
from datetime import datetime
import pandas as pd
from collections import namedtuple
from FinanceTools import *
import numpy as np
import sys
class StatusInvest:
def __init__(self, inFile):
self.orders = pd.read_csv(inFile)
self.orders['Category'] = self.orders['Category'].appl... | pd.to_datetime(self.orders['Date']) | pandas.to_datetime |
'''
Importing pandasTools enables several features that allow for using RDKit molecules as columns of a Pandas dataframe.
If the dataframe is containing a molecule format in a column (e.g. smiles), like in this example:
>>> from rdkit.Chem import PandasTools
>>> import pandas as pd
>>> import os
>>> from rdkit import R... | pd.version.version.split('.') | pandas.version.version.split |
from incrementalSearch import incrementalSearch
from bisection import bisection
from newton import newton
from falseRule import falseRule
from fixedPoint import fixedPoint
from multipleRoots import multipleRoots
from gaussPartialPivot import partialPivot
from gaussSimple import gaussSimple
from gaussTotal import gaussT... | pd.DataFrame(table) | pandas.DataFrame |
# Copyright 2020 <NAME>, <NAME>, <NAME>, <NAME>, <NAME>
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
# http://www.apache.org/licenses/LICENSE-2.0
# Unless required b... | DataFrame(lsParsed[0], columns=["WORD", "TAG", "CHUNK", "ROLE", "PNP"]) | pandas.DataFrame |
from sqlalchemy import func
import pandas as pd
import numpy as np
from cswd.sql.base import session_scope
from cswd.sql.models import (Issue, StockDaily, Adjustment, DealDetail,
SpecialTreatment, SpecialTreatmentType)
DAILY_COLS = ['symbol', 'date',
'open', 'high', 'low', 'c... | pd.DataFrame(columns=ADJUSTMENT_COLS) | pandas.DataFrame |
# -*- coding: utf-8 -*-
# pylint: disable=W0612,E1101
from datetime import datetime
import operator
import nose
from functools import wraps
import numpy as np
import pandas as pd
from pandas import Series, DataFrame, Index, isnull, notnull, pivot, MultiIndex
from pandas.core.datetools import bday
from pandas.core.n... | assert_panel_equal(sorted_panel, self.panel) | pandas.util.testing.assert_panel_equal |
import numpy as np
from fconcrete import config as c
import matplotlib.pyplot as plt
import time
import ezdxf
import pandas as pd
_Q = c._Q
def cond(x, singular=False, order=0):
"""
If It is singular, return 1 if x>0 else 0.
If It is not singular, return x**order if x>0 else 0
"""
if... | pd.DataFrame(array_table) | pandas.DataFrame |
""" Prep a series of graphs for the algorithm diagram figure.
"""
from collections import defaultdict, Counter
from copy import deepcopy
from itertools import combinations
from math import sqrt, factorial # , comb
import numpy
import pandas
def binom_coeff(n, k):
"""
apparently `from math import comb` isn't ... | pandas.DataFrame.from_dict(edges) | pandas.DataFrame.from_dict |
from PhiRelevance.PhiUtils1 import phiControl,phi
import pandas as pd
import matplotlib.pyplot as plt
import numpy as np
import math
from random import seed
from random import randint
from random import random
class SmoteRRegression:
"""
Class SmoteRRegression takes arguments as follows:
data - Pandas... | pd.concat(frames) | pandas.concat |
import os
import pandas as pd
import numpy as np
import untangle
import requests
import scripts.manipulation as manipulation
mais_path = "../../bd+/mais_projects/data/alesp"
def parse_deputados(download=True): # sourcery no-metrics
if download:
r = requests.get(
"https://www.al.sp.gov.br... | pd.read_csv("../data/servidores/deputados_alesp_aux.csv") | pandas.read_csv |
"""
Tests for zipline/utils/pandas_utils.py
"""
from unittest import skipIf
import pandas as pd
from zipline.testing import parameter_space, ZiplineTestCase
from zipline.testing.predicates import assert_equal
from zipline.utils.pandas_utils import (
categorical_df_concat,
nearest_unequal_elements,
new_pan... | pd.Series([100, 102, 103], dtype='int64') | pandas.Series |
# This file is generated from image_classification/dataset.md automatically through:
# d2lbook build lib
# Don't edit it directly
#@save_all
#@hide_all
import pathlib
import pandas as pd
from matplotlib import pyplot as plt
from typing import Union, Sequence, Callable, Optional
import fnmatch
import numpy as np
imp... | pd.DataFrame(entries) | pandas.DataFrame |
import asyncio
import concurrent.futures
import itertools
import logging
import math
import random
import sys
import dask.array as da
import dask.dataframe as dd
import numpy as np
import pandas as pd
import pytest
import scipy
import toolz
from dask.distributed import Future
from distributed.utils_test import ( # no... | pd.DataFrame(search.history_) | pandas.DataFrame |
#!/usr/bin/env python
# coding: utf-8
# In[79]:
import numpy as np
import pandas as pd
import datetime
import csv
# In[45]:
#modifing the text file to make the split of date and temperture pairs.
def textModify():
with open('temperature.txt', 'r') as file :
filedata = file.read()
filedata = filed... | pd.read_csv('temperature.txt',sep="$", header=None) | pandas.read_csv |
from __future__ import division
import pytest
import numpy as np
from datetime import timedelta
from pandas import (
Interval, IntervalIndex, Index, isna, notna, interval_range, Timestamp,
Timedelta, compat, date_range, timedelta_range, DateOffset)
from pandas.compat import lzip
from pandas.tseries.offsets imp... | tm.assert_index_equal(result, expected) | pandas.util.testing.assert_index_equal |
# -*- coding: utf-8 -*-
"""
Created on 12/20/2019
@author: Azhu
"""
import glob
import pdfquery
import os
import tabula as tb
import cv2
import numpy as np
import pandas as pd
import PyPDF2
from PyPDF2 import PdfFileReader, PdfFileWriter
###############################################################################... | pd.DataFrame() | pandas.DataFrame |
import scipy.io as sio
import numpy as np
import pandas as pd
import tables
import pickle
from scipy.interpolate import interp1d
import os
from ismore import settings
from utils.constants import *
pkl_name = os.path.expandvars('$BMI3D/riglib/ismore/traj_reference_interp.pkl')
mat_name = os.path.expandvars('$HOME/Des... | pd.DataFrame(kin, columns=columns) | pandas.DataFrame |
import pytest
from datetime import datetime
import pandas as pd
from tadpole_algorithms.transformations import convert_to_year_month, \
convert_to_year_month_day, map_string_diagnosis
def test_forecastDf_date_conversion():
forecastDf = pd.DataFrame([{'Forecast Date': '2019-07'}])
assert pd.api.types.is... | pd.api.types.is_datetime64_ns_dtype(d4Df_new2['ScanDate']) | pandas.api.types.is_datetime64_ns_dtype |
import pandas as pd
ds = | pd.Series([2, 4, 6, 8, 10, 12, 14, 16, 18, 20]) | pandas.Series |
"""
Copyright (C) 2018 <NAME> (<EMAIL>)
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, sof... | pd.merge(self.eg_edges, events_meta, left_on='source', right_index=True) | pandas.merge |
# -*- coding:utf-8 -*-
# /usr/bin/env python
"""
Date: 2021/7/27 15:14
Desc: 东方财富-经济数据-英国
http://data.eastmoney.com/cjsj/foreign_4_0.html
"""
import pandas as pd
import requests
import demjson
# Halifax房价指数月率
def macro_uk_halifax_monthly():
"""
东方财富-经济数据-英国-Halifax 房价指数月率
http://data.eastmoney.com/cjsj/fo... | o_numeric(temp_df["现值"]) | pandas.to_numeric |
import os
from re import X
import sys
import logging
import argparse
import shutil
import numpy as np
from pandas import read_csv, DataFrame
from Bio.SeqIO import parse
from itertools import product
from subprocess import call
from scipy.spatial.distance import pdist, squareform
import plotly.express as px
#from pand... | read_csv('./data/CONCAT_base.tsv', sep='\t', index_col=[0]) | pandas.read_csv |
import os
import shutil
import dash
import base64
import io
import pandas as pd
import geopandas as gpd
from shapely.geometry import Polygon
import numpy as np
import dash_table
import dash_bootstrap_components as dbc
import dash_html_components as html
import dash_core_components as dcc
from dash.dependencies import I... | pd.DataFrame(datatable_load_profile) | pandas.DataFrame |
import pandas
from functools import reduce
from .. import SystemClass
from ..decorators import actions, mode
from typing import Iterable, Tuple, Callable
@mode
def run_static_pf(
distSys: SystemClass,
actions: Iterable[Callable] = (lambda distSys: None,),
tools: Iterable[Callable] = (lambda distSys: None,... | pandas.concat([df1, df2], ignore_index=True) | pandas.concat |
# -*- coding: utf-8 -*-
"""
Created on Thu Jun 4 16:22:57 2020
@author: Natalie
"""
import os
import sys
import click
import pickle
import pandas as pd
import numpy as np
import geopandas as gpd
import imageio
from shapely.geometry import Point
import json
from bokeh.io import output_file
from bokeh.plotting import... | pd.merge(tmp_df, age_count_temp, how='left', left_index=True,right_index=True) | pandas.merge |
# Copyright (c) 2019-2021 - for information on the respective copyright owner
# see the NOTICE file and/or the repository
# https://github.com/boschresearch/pylife
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a co... | pd.testing.assert_series_equal(expected_obj, obj) | pandas.testing.assert_series_equal |
from __future__ import division
import os
import os.path as op
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import nibabel as nib
from nipype import (Workflow, Node, MapNode, JoinNode,
IdentityInterface, DataSink)
from nipype.interfaces.base import traits, TraitedSpec
fro... | pd.DataFrame(mc_data, columns=cols) | pandas.DataFrame |
import unittest
import xmlrunner
import pandas as pd
import numpy as np
class DummyUnitTest(unittest.TestCase):
def setUp(self):
self.int = 5
self.yes = True
self.no = False
self.float = 0.5
self.pi = 3.141592653589793238462643383279
self.string = "Miguel"
s... | pd.DataFrame(self.dict, index=[0]) | pandas.DataFrame |
import os
import random
import math
import numpy as np
import pandas as pd
import itertools
from functools import lru_cache
##########################
## Compliance functions ##
##########################
def delayed_ramp_fun(Nc_old, Nc_new, t, tau_days, l, t_start):
"""
t : timestamp
current date
... | pd.Timestamp('2022-09-21') | pandas.Timestamp |
import os
from sklearn import metrics
from . import data as D
import itertools
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
from . import plotting
SURROGATES = 'surrogates iaaft'.split() + [None]
TESTSET_VALS = (False, True)
LABELS = 'Wake S1 S2 S3 S4 REM'.split()
REDUCED_LABELS = 'Wake Ligh... | pd.Series(data=None, index=mindex) | pandas.Series |
import warnings
warnings.filterwarnings('ignore')
import matplotlib.pyplot as plt
import pandas as pd
import scipy.stats as stats
import numpy as np
import seaborn as sns
import os,sys
import itertools
import datetime
import scipy.signal as signal
import matplotlib.dates as mdates
from sklearn.neighbors._kde import Ke... | pd.DataFrame(gps_mean.loc[df_std.index]) | pandas.DataFrame |
import itertools
from collections import defaultdict
import numpy as np
import pandas as pd
import torch
from pytorch_toolbelt.utils.fs import id_from_fname
from pytorch_toolbelt.utils.torch_utils import to_numpy
from sklearn.metrics import cohen_kappa_score
from torch import nn
from torch.utils.data import DataLoader... | pd.DataFrame.from_dict(predictions) | pandas.DataFrame.from_dict |
from .test_dataset import CMD_CREATE_TEST_TABLE
import pytest
import pandas as pd
import numpy as np
import os
from ..dataset import sql_dataset
from .gen_rand_data import rand_df
CMD_DROP_TEST_TABLE_IF_EXISTS = "IF OBJECT_ID('test_table', 'U') IS NOT NULL DROP TABLE test_table;"
CMD_CREATE_TRUNCATED_TEST_TABLE = """... | pd.testing.assert_frame_equal(df_queried, df_orig, check_dtype=False, check_names=False) | pandas.testing.assert_frame_equal |
import pandas as pd
import numpy as np
from sklearn.model_selection import KFold
import statistics
import math
from sklearn.decomposition import PCA
COLUMNS_TO_USE = ["Home_type", "rooms", "home_size_m2", "lotsize_m2",
"expenses_dkk", "floor_as_int", "balcony", "zipcodes",
"m2_pr... | pd.concat([df, zipcodes_onehot], axis=1) | pandas.concat |
""" Format data """
from __future__ import division, print_function
import pandas as pd
import numpy as np
import re
from os.path import dirname, join
from copy import deepcopy
import lawstructural.lawstructural.constants as lc
import lawstructural.lawstructural.utils as lu
#TODO: Take out entrant stuff from lawData
... | pd.merge(self.data, self.data_sets[dset], how='outer') | pandas.merge |
import pandas as pd
import pytest
from pandas.util.testing import assert_frame_equal
from pyspark import sql
from pysparkhelpers import helpers
@pytest.mark.usefixtures("hive_context")
def test_single(hive_context):
df = pd.DataFrame({'user': ['a', 'a', 'a', 'b'],
'values': [1, 1, 1, 4]})
... | assert_frame_equal(expected_value, returned_value) | pandas.util.testing.assert_frame_equal |
# -*- coding: utf-8 -*-
"""
Created on Thu Jan 24 20:03:24 2019
@author: RV
"""
# Python(R)
# Modeules/packageslibraries
# OS - submodules/path/join
#eg. (os.path.join)
# pandas
# scipy
# onspy
#%% Setup
import os
projFld = "C:/Users/RV/Documents/Teaching/2019_01_Spring/ADEC7430_Spring2019/Lec... | pd.isnull(rawTrain) | pandas.isnull |
import datetime
from typing import List
import pandas as pd
import pytest
from ruamel.yaml import YAML
import great_expectations.exceptions as ge_exceptions
from great_expectations.core.batch import (
Batch,
BatchDefinition,
BatchSpec,
RuntimeBatchRequest,
)
from great_expectations.core.batch_spec imp... | pd.DataFrame(data={"col1": [1, 2], "col2": [3, 4]}) | pandas.DataFrame |
from dataProcessing import *
from sklearn.ensemble import RandomForestClassifier
from sklearn.ensemble import IsolationForest
import pandas as pd
import pickle
from RandomForestCounterFactual import *
def checkSamples(
datasetFileName,
unscaledFactualsFileName,
unscaledCounterFactualsFileName,
... | pd.DataFrame() | pandas.DataFrame |
# -*- coding: utf-8 -*-
"""
.. module:: courtship
:synopsis: Classes for Fly objects.
.. moduleauthor:: <NAME>
"""
import numpy as np
import pandas as pd
from .behavior import Behavior
class Point(object):
"""Stores 2D coordinates as attributes.
Attributes
----------
row : np... | pd.concat(to_concat, axis=1) | pandas.concat |
#!usr/local/bin
import pandas as pd
import numpy as np
from sys import argv
import subprocess
def sequence_splice_site(sp_junc, fasta_file):
part = sp_junc
part = part[['chr','first_base','last_base','motif','STAR_annotation','strand']]
part['first_base'] = part['first_base']
#part['chr'] = 'chr' + pa... | pd.concat([cat1,cat2]) | pandas.concat |
#https://www.youtube.com/watch?v=xKvffLRSyPk&list=PL3JVwFmb_BnSLFyVThMfEavAEZYHBpWEd&index=1
import os, io
from google.cloud import vision
import pandas as pd
#from google.cloud.vision import types -> 버전 업그레이드 되면서 types 사용 안함
os.environ['GOOGLE_APPLICATION_CREDENTIALS'] = r'ServiceAccountToken.json'
client ... | pd.DataFrame(columns=['locale', 'description']) | pandas.DataFrame |
"""
This is a place to create a python wrapper for the BASGRA fortran model in fortarn_BASGRA_NZ
Author: <NAME>
Created: 12/08/2020 9:32 AM
"""
import os
import ctypes as ct
import numpy as np
import pandas as pd
from subprocess import Popen
from copy import deepcopy
from input_output_keys import param_keys, out_co... | pd.to_datetime(strs, format='%Y-%j') | pandas.to_datetime |
import os
import uuid
from datetime import datetime
from time import sleep
import fsspec
import pandas as pd
import pytest
import v3iofs
from storey import EmitEveryEvent
import mlrun
import mlrun.feature_store as fs
from mlrun import store_manager
from mlrun.datastore.sources import CSVSource, ParquetSource
from mlr... | pd.Timestamp("2021-01-10 12:00:00") | pandas.Timestamp |
#modules to initialize the API
#the API will run on two endpoints:
# Students - to get the user's course code
# Lessons - to access the timetable details for which the alerts will be sent
#with the two endpoints, if the API is located at www.api.com
# comm. with the Students and Lessons classes will be at
# www.a... | pd.read_json('classCodes.json') | pandas.read_json |
#!/usr/bin/env python
import os
import sys
import pandas as pd
import argparse
import configparser
import multiprocessing
import time
import datetime
from sqlalchemy import create_engine
from sqlalchemy.pool import NullPool
import tqdm
import statsmodels.stats.multitest as multitest
import snps
import genes
import in... | pd.read_sql(sql, con=con) | pandas.read_sql |
################################################################################
# The contents of this file are Teradata Public Content and have been released
# to the Public Domain.
# <NAME> & <NAME> - April 2020 - v.1.1
# Copyright (c) 2020 by Teradata
# Licensed under BSD; see "license.txt" file in the bundle root ... | pd.to_numeric(df['tot_cust_years']) | pandas.to_numeric |
import numpy as np
import pandas as pd
import time
import cv2
import argparse
import torch
import torch.nn as nn
import torch.nn.functional as F
import torch.optim as optim
import torch.nn.init
from torch.autograd import Variable
from . models import Spatial_CNN
class TESLA(object):
def __init__(self):
super(TESLA,... | pd.Series(nbs) | pandas.Series |
from collections import abc, deque
from decimal import Decimal
from io import StringIO
from warnings import catch_warnings
import numpy as np
from numpy.random import randn
import pytest
from pandas.core.dtypes.dtypes import CategoricalDtype
import pandas as pd
from pandas import (
Categorical,
DataFrame,
... | pd.MultiIndex.from_tuples(tuples, names=["level2", "level1", None]) | pandas.MultiIndex.from_tuples |
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
from .plan_losses import PPC, PlanCost,get_leading_hint
from query_representation.utils import deterministic_hash,make_dir
from query_representation.viz import *
from matplotlib.backends.backend_pdf import PdfPages
import matplotlib.pyplot as plt
i... | pd.DataFrame(cur_costs) | pandas.DataFrame |
import cbsodata
import math
import pandas
import matplotlib.pyplot as plt
from functools import reduce
from pathlib import Path
from sklearn.preprocessing import MinMaxScaler
def download(identifier: str):
"""Download a dataset from the CBS odata portal."""
# Prepare download directory
download_director... | pandas.read_csv(source_path) | pandas.read_csv |
import numpy as np
import pandas as pd
import pytest
from numpy.testing import assert_array_equal
from pandas.testing import assert_series_equal
from src.shared import _create_group_ids
from src.shared import _determine_number_of_groups
from src.shared import _expand_or_contract_ids
from src.shared import create_group... | pd.Series({1: 20, 2: 5, 5: 2}) | pandas.Series |
#-*- coding:utf-8 -*-
from sklearn.pipeline import Pipeline
from sklearn.feature_extraction.text import CountVectorizer
import pandas as pd
import numpy as np
from sklearn import metrics
from sklearn.model_selection import train_test_split
from sklearn.linear_model import LogisticRegression
from sklearn.feature_extrac... | pd.read_csv('dataset/raw_training.csv') | pandas.read_csv |
import os
import pandas as pd
import numpy as np
from pandas.api.types import CategoricalDtype
from imblearn.over_sampling import SMOTE
DATA_PATH = '../cell-profiler/measurements'
def load_data(filename, data_path=DATA_PATH):
"""
Read a csv file.
"""
csv_path = os.path.join(data_path, filename)
... | pd.DataFrame(y_sm, columns=['stiffness']) | pandas.DataFrame |
from collections import OrderedDict
from datetime import timedelta
import numpy as np
import pytest
from pandas.core.dtypes.dtypes import CategoricalDtype, DatetimeTZDtype
import pandas as pd
from pandas import (
Categorical,
DataFrame,
Series,
Timedelta,
Timestamp,
_np_version_under1p14,
... | tm.assert_frame_equal(result, expected) | pandas.util.testing.assert_frame_equal |
# -*- coding: utf-8 -*-
# pylint: disable=E1101
# flake8: noqa
from datetime import datetime
import csv
import os
import sys
import re
import nose
import platform
from multiprocessing.pool import ThreadPool
from numpy import nan
import numpy as np
from pandas.io.common import DtypeWarning
from pandas import DataFr... | StringIO(test) | pandas.compat.StringIO |
# import Ipynb_importer
import pandas as pd
from .public_fun import *
# 全局变量
class glv:
def _init():
global _global_dict
_global_dict = {}
def set_value(key,value):
_global_dict[key] = value
def get_value(key,defValue=None):
try:
return _global_dict[key... | pd.merge(self.o.ol, self.c.ol, left_index=True, right_index=True) | pandas.merge |
import pandas as pd
import pymmwr as pm
import datetime
import warnings
import io
import requests
warnings.simplefilter(action='ignore')
def read_fips_codes(filepath):
# read file
fips_codes = pd.read_csv(filepath)
# take state code from all fips codes
fips_codes['state_abbr'] = fips_codes['location'].str[:2]
... | pd.to_datetime(df_byday['date']) | pandas.to_datetime |
import numpy as np
from numpy.random import randn
import pytest
from pandas import DataFrame, Series
import pandas._testing as tm
@pytest.mark.parametrize("name", ["var", "vol", "mean"])
def test_ewma_series(series, name):
series_result = getattr(series.ewm(com=10), name)()
assert isinstance(series_result, S... | tm.assert_series_equal(a, c) | pandas._testing.assert_series_equal |
import os
import json
import datetime
import multiprocessing
import random
import copy
import time
import warnings
import pandas as pd
import numpy as np
from datetime import date, timedelta
from pathlib import Path
from models.common.mit_buildings import MITBuildings
from models.common.to_precision import sig_fig_f... | pd.unique(self.all_boostrap_campus_daily_sample_occupancy['date']) | pandas.unique |
# last edited: 04/10/2021
#
# The functions pca_initial, pca_initial_, pca_final, and pca_final_ are adapted
# from a post by <NAME> here:
# https://nirpyresearch.com/classification-nir-spectra-principal-component-analysis-python/
#
# Retrieved in December 2020 and is licensed under Creative Commons Attribution 4.0
# I... | pd.DataFrame(data) | pandas.DataFrame |
import os
import shutil
import logging
import pandas as pd
import matplotlib
matplotlib.use("agg") # no need for tk
from supervised.automl import AutoML
from frameworks.shared.callee import call_run, result, output_subdir, utils
log = logging.getLogger(os.path.basename(__file__))
def run(dataset, config):
log... | pd.DataFrame(dataset.test.X, columns=column_names) | pandas.DataFrame |
"""
@author: <NAME>
"""
import os
import re
import numpy as np
import pandas as pd
import json
import pickle
from argparse import ArgumentParser
from sklearn.neighbors import KNeighborsClassifier
from sklearn.neural_network import MLPClassifier
from sklearn.svm import SVC
from sklearn.metrics import accuracy_score, cl... | pd.Series([], dtype=str) | pandas.Series |
import pandas as pd
def break_even_moneyline(probability):
if probability > 0.5:
x = -(100 / (1-probability))+100
else:
x = 100/probability - 100
return x
def predict_probs_and_moneylines(data):
predictions = | pd.DataFrame(columns=["Team1", "Seed1", "Team2", "Seed2", "Win%1", "Win%2", "ML1", "ML2"]) | pandas.DataFrame |
# -*- coding: utf-8 -*-
"""
Created on Wed Jun 13 10:15:43 2018
Build Neural network from perceptron,treating biases as part of weights and use
matrix computation to optimize the stochastic gradient descent method.
@author: Feng
"""
#### Libraries
import random
import numpy as np
class NeuralNetw... | pd.get_dummies(car_data['Auction'],prefix='Auction') | pandas.get_dummies |
import numpy as np
import pandas as pd
from sklearn import preprocessing, linear_model, metrics
import gc; gc.enable()
import random
import time, datetime
from sklearn.neural_network import MLPRegressor
from sklearn.linear_model import TheilSenRegressor, BayesianRidge
from sklearn.ensemble import BaggingRegressor
from... | pd.to_datetime(df['date']) | pandas.to_datetime |
import pysubgroup as ps
import pandas as pd
from subgroup_sem import SEMTarget, TestQF
############################################################################################
# imoport and preprocess data
############################################################################################
data = | pd.read_csv('artificial_data.csv') | pandas.read_csv |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.