prompt stringlengths 19 1.03M | completion stringlengths 4 2.12k | api stringlengths 8 90 |
|---|---|---|
# coding: utf-8
import os
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
from sklearn.metrics import roc_curve, auc
def cut_bins(x, bins=10, method='equal'):
"""
对x进行分bin,返回每个样本的bin值和每个bin的下界
Parameters
----------
x: numpy.ndarray or pandas.Series
变量
bins: int... | pd.DataFrame() | pandas.DataFrame |
'''
This file is part of PM4Py (More Info: https://pm4py.fit.fraunhofer.de).
PM4Py is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any late... | pd.concat([first_eve_df, last_eve_df], axis=1) | pandas.concat |
'''
tRNA Adaptation Index
'''
import collections
import os
import json
import logging
import pandas as pd
import numpy as np
import scipy.stats.mstats
from sqlalchemy import create_engine
from ..alphabet import CODON_REDUNDANCY
logger = logging.getLogger(__name__)
def main():
logging.basicConfig(level=logging... | pd.isnull(tpl.codon) | pandas.isnull |
import numpy as np
#import skimage.transform as sktransform
import random
import matplotlib.image as mpimg
import os
import pandas as pd
import matplotlib.pyplot as plt
import shutil
new_path = './track1/IMG/'
current_path = './out2rev/IMG/'
if not os.path.exists(new_path):
os.makedirs(new_path)
print('Fold... | pd.DataFrame() | pandas.DataFrame |
import pickle
import numpy as np
import pandas as pd
from sklearn.exceptions import ConvergenceWarning
from sklearn.mixture import BayesianGaussianMixture
from sklearn.preprocessing import OneHotEncoder
from sklearn.utils.testing import ignore_warnings
class DataTransformer(object):
"""Data Transformer.
Mod... | pd.DataFrame(output, columns=column_names) | pandas.DataFrame |
"""
Prepare training and testing datasets as CSV dictionaries
Created on 11/26/2018
@author: RH
"""
import os
import pandas as pd
import sklearn.utils as sku
import numpy as np
# get all full paths of images
def image_ids_in(root_dir, ignore=['.DS_Store','dict.csv', 'all.csv']):
ids = []
for id in os.listdi... | pd.concat(valist) | pandas.concat |
from sympy import *
import pandas as pd
from random import random
def random_optimization(xl, xu, n, function):
x = Symbol('x')
f = parse_expr(function)
iteration = 0
data = | pd.DataFrame(columns=['iteration','xl','xu','x','f(x)','max_x','max_f(x)']) | pandas.DataFrame |
import os
from datetime import date
from dask.dataframe import DataFrame as DaskDataFrame
from numpy import nan, ndarray
from numpy.testing import assert_allclose, assert_array_equal
from pandas import DataFrame, Series, Timedelta, Timestamp
from pandas.testing import assert_frame_equal, assert_series_equal
from pymo... | Timestamp('2008-10-23 05:53:11') | pandas.Timestamp |
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import tensorflow as tf
#-- fix for tensorflow 2.0 version ---
# import tensorflow.compat.v1 as tf
# tf.disable_v2_behavior()
import numpy as np
import os
import matplotlib.pyplot as plt
import traceback
import... | pd.plotting.scatter_matrix(df) | pandas.plotting.scatter_matrix |
import streamlit as st
import streamlit.components.v1 as components
from streamlit_folium import folium_static
# import folium
import pandas as pd
import matplotlib.pyplot as plt
import datetime
import immo
import ssl
# to avoid SSLCertVerificationError
ssl._create_default_https_context = ssl._create_unverified_contex... | pd.cut(df['prixm2'], bins=4,labels=['blue','green', 'yellow', 'red']) | pandas.cut |
''' ATP Matches Data Pipeline '''
import datetime as dt, pandas as pd
def run_pipeline(start_year = 1968, end_year = dt.datetime.now().year + 1):
# Extract match data
tour_files = []
for i in range(start_year, end_year):
url_base = r'https://raw.githubusercontent.com/JeffSackmann/tennis_atp/master... | pd.read_csv(tour_files[0]) | pandas.read_csv |
# -*- coding: utf-8 -*-
# pylint: disable=W0612,E1101
from datetime import datetime
import operator
import nose
from functools import wraps
import numpy as np
import pandas as pd
from pandas import Series, DataFrame, Index, isnull, notnull, pivot, MultiIndex
from pandas.core.datetools import bday
from pandas.core.n... | assert_almost_equal(casted2.values, exp_values) | pandas.util.testing.assert_almost_equal |
import pandas as pd
from schoonmaken.common import Person, Task
def save_data(people, save_path):
pass
def retrieve_data(data_path) -> list[tuple]:
'''
Layout of data:
[
(person_name: str, [TaskData, ...]),
...
]
'''
td = []
return td
def retrieve_p... | pd.DataFrame(task_table) | pandas.DataFrame |
import pkg_resources
import os
import sys
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
from windea_tool import weibull
from windea_tool import plotting
turbines_dict = {'Enercon E-70 (2300 kW)':'Enercon_E-70_2300kW.xlsx',
'Enercon E-115 (3000 kW)':'Enercon_E-115_3000kW.xlsx'}... | pd.DataFrame() | pandas.DataFrame |
import numpy as np
import pandas as pd
from os.path import join as joinPaths
from os.path import isdir
from os.path import isfile
from os import listdir as ls
from IPython.display import display, Markdown, Latex
import matplotlib.pyplot as plt
import matplotlib.lines as mlines
from matplotlib.pyplot import cm
from mult... | pd.to_datetime(df["time"], unit="s", utc=True) | pandas.to_datetime |
import numpy as np
import pandas as pd
import pytest
from pandas.util import hash_pandas_object
import dask.dataframe as dd
from dask.dataframe import _compat
from dask.dataframe._compat import tm
from dask.dataframe.utils import assert_eq
@pytest.mark.parametrize(
"obj",
[
pd.Series([1, 2, 3]),
... | pd.Series([1000, 2000, 3000, 4000]) | pandas.Series |
"""Tests for the sdv.constraints.tabular module."""
import numpy as np
import pandas as pd
import pytest
from sdv.constraints.errors import MissingConstraintColumnError
from sdv.constraints.tabular import (
ColumnFormula, CustomConstraint, GreaterThan, UniqueCombinations)
def dummy_transform():
pass
def d... | pd.to_datetime(['2020-01-01T00:00:01', '2020-01-02T00:00:01']) | pandas.to_datetime |
import matplotlib.dates as mdates
from tqdm import tqdm as tqdm
import datetime
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
from CometTS.CometTS import interpolate_gaps
import argparse
import os
sns.set(color_codes=True)
# Functions for a seasonal auto-regressive integ... | pd.read_csv(CometTSOutputCSV) | pandas.read_csv |
# -*- coding: utf-8 -*-
"""
/***************************************************************************
Summarizes raster statistics in parallel from soilgrids to estonian soil polygons
-------------------
copyright : (C) 2018-2020 by <NAME>
email ... | pd.DataFrame(outputs) | pandas.DataFrame |
import pandas as pd
lista_valores = [1,2,3]
lista_indices = ['a', 'b', 'c']
serie = pd.Series(lista_valores, index=lista_indices)
print(serie)
lista_notas = [[6,7,8],[8,9,5],[6,9,7]]
lista_indices2 = ['Matematicas', 'historia', 'fisica']
lista_nombres = ['Antonio', 'Maria', 'Pedro']
dataframe = | pd.DataFrame(lista_notas, index=lista_indices2, columns=lista_nombres) | pandas.DataFrame |
import numpy as np
import pandas as pd
from scipy import signal as ssig
from scipy import stats as spst
import os
import re
import string
from salishsea_tools import geo_tools
import netCDF4 as nc
class Cast:
def __init__(self,fpath):
mSta,mLat,mLon,df=readcnv(fpath)
self.sta=mSta
self.lat=... | pd.DataFrame(p,columns=['depth_m']) | pandas.DataFrame |
import datetime as dt
import streamlit as st
import pandas as pd
import numpy as np
import plotly.graph_objects as go
import numerapi
import plotly.express as px
from utils import *
# setup backend
napi = numerapi.SignalsAPI()
leaderboard_df = pd.DataFrame(napi.get_leaderboard(limit = 10_000))
MODELS_TO_CHECK = ... | pd.concat(rep_dfs) | pandas.concat |
from itertools import product as it_product
from typing import List, Dict
import numpy as np
import os
import pandas as pd
from scipy.stats import spearmanr, wilcoxon
from provided_code.constants_class import ModelParameters
from provided_code.data_loader import DataLoader
from provided_code.dose_evaluation_class imp... | pd.read_csv(consolidate_data_paths['weights'], index_col=[0, 1, 2, 3], header=[0, 1]) | pandas.read_csv |
import os
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sbn
import click
# define functions
def getUnique(df):
"""Calcualtes percentage of unique reads"""
return (
df.loc[df[0] == "total_nodups", 1].values[0]
/ df.loc[df[0] == "total_mapped", 1].values[0... | pd.melt(uniqueF) | pandas.melt |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Thu Jan 13 11:22:34 2022
@author: mariaolaru
"""
import os
import pandas as pd
import numpy as np
import statsmodels.api as sm
import scipy.stats as stat
import xarray as xr
from matplotlib import pyplot as plt
import math
from sklearn.preprocessing import ... | pd.Series(diff == 1) | pandas.Series |
from datetime import datetime
import numpy as np
import pytest
from pandas.core.dtypes.cast import find_common_type, is_dtype_equal
import pandas as pd
from pandas import DataFrame, Index, MultiIndex, Series
import pandas._testing as tm
class TestDataFrameCombineFirst:
def test_combine_first_mixed(self):
... | pd.Timestamp("2011-01-01") | pandas.Timestamp |
from io import StringIO
import pandas as pd
import numpy as np
import pytest
import bioframe
import bioframe.core.checks as checks
# import pyranges as pr
# def bioframe_to_pyranges(df):
# pydf = df.copy()
# pydf.rename(
# {"chrom": "Chromosome", "start": "Start", "end": "End"},
# axis="col... | pd.Int64Dtype() | pandas.Int64Dtype |
# Copyright 2021 <NAME>. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agree... | pd.DataFrame(results) | pandas.DataFrame |
# -*- coding: utf-8 -*-
import string
from collections import OrderedDict
from datetime import date, datetime
import numpy as np
import pandas as pd
import pandas.util.testing as pdt
import pytest
from kartothek.core.common_metadata import make_meta, store_schema_metadata
from kartothek.core.index import ExplicitSe... | pd.DataFrame({"a": [2]}) | pandas.DataFrame |
"""
author: <NAME> & <NAME>
Implementation of the climate data-utils for our training framework (i.e. on
synthetic SDE data).
This is mainly a copy of the data_utils.py file from the official implementation
of GRU-ODE-Bayes: https://github.com/edebrouwer/gru_ode_bayes.
"""
import torch
import pandas as pd
import num... | pd.read_csv(root_dir + "/" + label_file) | pandas.read_csv |
#!/usr/bin/env python3
"""Module to download cryptocurrency ohlc data"""
import time
import datetime as dt
import logging
import json
import requests
import pandas as pd
def from_datetime_to_unix(date):
'''in: datetime, out: unix_timestamp'''
return int(time.mktime(date.timetuple()))
def from_unix_to_date(da... | pd.concat(frames) | pandas.concat |
# -*- coding: utf-8 -*-
"""
Created on Fri Sep 3 09:59:23 2021
@author: <NAME>
Install Packages:
- pip install pandas
Draws a Gantt chart per trial
Each event is on a different y-tick
"""
# eNPHR 1.0 SS_6
#
from ganttChartDrawer import draw_group
import pandas as pd
import numpy as np
# Read and organize... | pd.read_excel('cleversys_events/Trial event export ' + filename + '.xlsx', skiprows=[0,1,2,3,4,5]) | pandas.read_excel |
"""
Name: shread.py
Author: <NAME>, Reclamation Technical Service Center
Description: Utilities for downloading and processing snow products
ADD CLASSES / FUNCTIONS DEFINED
ADD WHA
"""
import ftplib
import os
import tarfile
import gzip
from osgeo import gdal
import csv
import logging
import glob
from osgeo import osr
... | pd.concat(frames, axis=1) | pandas.concat |
import json
import csv
import numpy as np
from stockstats import StockDataFrame
import pandas as pd
import mplfinance as mpf
import seaborn as sn
import matplotlib.pyplot as plt
def load_secrets():
"""
Load data from secret.json as JSON
:return: Dict.
"""
try:
with open('secret.json', 'r')... | pd.DataFrame(data) | pandas.DataFrame |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Contains the Evaluator class.
Part of symenergy. Copyright 2018 authors listed in AUTHORS.
"""
import os
import sys
import gc
import py_compile
import sympy as sp
import numpy as np
from importlib import reload
from multiprocessing import current_process
import pandas... | pd.Series(True, index=df.index) | pandas.Series |
from __future__ import division
from datetime import datetime
import sys
if sys.version_info < (3, 3):
import mock
else:
from unittest import mock
import pandas as pd
import numpy as np
import random
from nose.tools import assert_almost_equal as aae
import bt
import bt.algos as algos
def test_algo_name():... | pd.DataFrame(index=dts, columns=['c1', 'c2'], data=100) | pandas.DataFrame |
import numpy as np
import pandas as pd
import pytest
from greykite.common.features.timeseries_lags import build_agg_lag_df
from greykite.common.features.timeseries_lags import build_autoreg_df
from greykite.common.features.timeseries_lags import build_autoreg_df_multi
from greykite.common.features.timeseries_lags impo... | pd.isnull(lag_df) | pandas.isnull |
import streamlit as st
import numpy as np
import pandas as pd
from matplotlib.image import imread
import matplotlib.pyplot as plt
import plotly.graph_objects as go
import seaborn as sns
import requests
import joblib
import shap
# import streamlit.components.v1 as components
shap.initjs()
st.set_option('deprecation.sho... | pd.DataFrame(X_train, columns=features_list_after_prepr) | pandas.DataFrame |
# Notebook to transform OSeMOSYS output to same format as EGEDA
# Import relevant packages
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
import os
from openpyxl import Workbook
import xlsxwriter
import pandas.io.formats.excel
import glob
import re
# Path for OSeMOSYS output
path_output = './d... | pd.DataFrame() | pandas.DataFrame |
import os
from tkinter import *
import pandas as pd
# UI set up
root = Tk()
root.title("Zoom Automator")
scroll = Scrollbar(root)
canvas = Canvas(root, width = 350, height = 350)
canvas.grid(columnspan = 7, rowspan = 7)
left_margin = Label(root, text = " ", padx = 7)
left_margin.grid(row = 0, column = 0)
bottom_... | pd.DataFrame(data=[], columns=colNames) | pandas.DataFrame |
#!/usr/bin/env python
# -*-coding:utf-8 -*-
'''
@File : Stress_detection_script.py
@Time : 2022/03/17 09:45:59
@Author : <NAME>
@Contact : <EMAIL>
'''
import os
import logging
import plotly.express as px
import numpy as np
import pandas as pd
import zipfile
import fnmatch
import flirt.reader.empatica
... | pd.DataFrame(ending_timestamp) | pandas.DataFrame |
# Copyright (c) 2013-2015 Siphon Contributors.
# Distributed under the terms of the BSD 3-Clause License.
# SPDX-License-Identifier: BSD-3-Clause
"""Read data from the National Data Buoy Center."""
from io import StringIO
import warnings
import numpy as np
import pandas as pd
import requests
from ..http_util import ... | pd.to_datetime(df[['year', 'month', 'day', 'hour', 'minute']], utc=True) | pandas.to_datetime |
# -*- coding: utf-8 -*-
"""
Load averaged results from csv, containing scores of all ckpts.
For each exp group, return the best ckpt (ranked by valid performance).
"""
import shutil
import configargparse
import os
import pandas as pd
__author__ = "<NAME>"
__email__ = "<EMAIL>"
dev_test_pairs = [
('kp20k_valid... | pd.concat([selfbest_dev_rows, dev_row]) | pandas.concat |
#!/usr/bin/env python
"""
Fuse csv files into one single file.
"""
## file_fuser.py
### fuse individual files into one giant file
import pandas as pd
import os
import argparse
from abc import ABCMeta, abstractmethod
class CsvFuserAbs(object, metaclass=ABCMeta):
## This object initialized from command line a... | pd.DataFrame() | pandas.DataFrame |
# Adapted from code written by <NAME>
import pandas as pd
from tqdm import tqdm
import numpy as np
import datetime
'''
Sources are currently human, mouse, and rat, in order of descending priority.
'''
sources = [
'ftp://ftp.ncbi.nih.gov/gene/DATA/GENE_INFO/Mammalia/Rattus_norvegicus.gene_info.gz',
'ftp://ftp.... | pd.read_csv(geneid_path, sep='\t', na_filter=False) | pandas.read_csv |
from flask import Flask, request, jsonify
import json
from sklearn.decomposition import PCA
from sklearn.preprocessing import StandardScaler
from sklearn.cluster import KMeans
# import matplotlib.pyplot as plt
import pandas as pd
import numpy as np
# from scipy.spatial import distance
from sklearn.manifold imp... | pd.read_csv(path) | pandas.read_csv |
# ---
# jupyter:
# jupytext:
# formats: ipynb,py
# text_representation:
# extension: .py
# format_name: light
# format_version: '1.5'
# jupytext_version: 1.9.1
# kernelspec:
# display_name: Python [conda env:core_acc] *
# language: python
# name: conda-env-core_acc-py
# -... | pd.read_csv(pa14_metadata_filename, header=0, index_col=0) | pandas.read_csv |
import warnings
import numpy as np
import pandas as pd
warnings.filterwarnings("ignore")
Data = pd.read_excel(io="Training_Data.xlsx", sheet_name="E0_Mod")
D = pd.DataFrame(Data[['HS','AS','HST','AST','HF','AF','HC','AC','HY','AY','HR','AR']].T)
D2 = pd.DataFrame(Data[['HS','AS','HST','AST','HF','AF','HC','AC',... | pd.DataFrame() | pandas.DataFrame |
"""
Tasks
-------
Search and transform jsonable structures, specifically to make it 'easy' to make tabular/csv output for other consumers.
Example
~~~~~~~~~~~~~
*give me a list of all the fields called 'id' in this stupid, gnarly
thing*
>>> Q('id',gnarly_data)
['id1','id2','id3']
Observations:
--... | u('name') | pandas.compat.u |
import itertools
import os
import random
import tempfile
from unittest import mock
import pandas as pd
import pytest
import pickle
import numpy as np
import string
import multiprocessing as mp
from copy import copy
import dask
import dask.dataframe as dd
from dask.dataframe._compat import tm, assert_categorical_equal... | pd.DataFrame({"tz": s_aware, "notz": s_naive}) | pandas.DataFrame |
#!/usr/bin/env python3
##############################################################
## <NAME> & <NAME> ##
## Copyright (C) 2020-2021 ##
##############################################################
'''
Created on 30 oct. 2020
@author: alba
Modified in Ma... | pd.DataFrame(data=None, columns=["length"]) | pandas.DataFrame |
import os
import ctypes
import math
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import matplotlib.cm as cm
from matplotlib.path import Path
import matplotlib.patches as patches
from matplotlib.collections import PolyCollection
import matplotlib.colors as colors
from pywfm import DLL_PATH, D... | pd.DataFrame({"NodeID": node_ids, "X": x, "Y": y}) | pandas.DataFrame |
import numpy
import pandas as pd
import math as m
#Moving Average
def MA(df, n):
MA = pd.Series(df['Close'].rolling(n).mean(), name = 'MA_' + str(n))
df = df.join(MA)
return df
def MACD(df, n_fast, n_slow):
"""Calculate MACD, MACD Signal and MACD difference
:param df: pandas.DataFr... | pd.Series(UpI) | pandas.Series |
from typing import List, Optional, Union
from pandas import DataFrame
from mstrio.api import documents
from mstrio.project_objects.document import Document
from mstrio.server.environment import Environment
from mstrio.utils import helper
from mstrio.connection import Connection
def list_dossiers(connection: Connec... | DataFrame(objects) | pandas.DataFrame |
from collections import deque
import numpy as np
import pandas as pd
from sunpy.util import SunpyUserWarning
__all__ = ['ELO']
class ELO:
"""
Recreating the ELO rating algirithm for Sunspotter.
"""
def __init__(self, score_board: pd.DataFrame, *, k_value=32, default_score=1400,
max... | pd.DataFrame([state_dict_0, state_dict_1]) | pandas.DataFrame |
import librosa
import numpy as np
import pandas as pd
from os import listdir
from os.path import isfile, join
from audioread import NoBackendError
def extract_features(path, label, emotionId, startid):
"""
提取path目录下的音频文件的特征,使用librosa库
:param path: 文件路径
:param label: 情绪类型
:param startid: 开始的序列号
... | pd.Series() | pandas.Series |
#!/usr/bin/env python2
# -*- coding: utf-8 -*-
"""
Created on Sat Jan 26 22:10:18 2019
@author: shashank
Takes a list of entries into a tournament you are running as:
ams.csv
amd.csv
awd.csv
axd.csv
bms.csv
etc.
and uses the rankings from the ranking.csv file to rank every entry in each
bracket against each other. ... | pd.read_csv("sandbagging/amd.csv",names=['Number','player']) | pandas.read_csv |
from difflib import SequenceMatcher
import functools
from typing import Optional
import pandas
__doc__ = """Get specialty codes and consolidate data from different sources in basic_data."""
COLUMNS = ['first_name', 'last_name', 'city', 'postal_code', 'state', 'specialty_code']
GENERIC_OPHTHALMOLOGY_CODE = '207W00000X... | pandas.isnull(value := row[col]) | pandas.isnull |
# -*- coding: utf-8 -*-
import numpy as np
import pandas as pd
from pandas.api.types import is_string_dtype
from pandas.api.types import is_numeric_dtype
import re
import warnings
import multiprocessing as mp
import matplotlib.pyplot as plt
import time
import os
import platform
from .condition_fun import *
from .info_... | pd.merge(bins_breakslist[['variable', 'breaks']], vars_class, how='left', on='variable') | pandas.merge |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Wed Mar 18 13:51:22 2020
@author: adiallo
<NAME>
Pans Project 2020
"""
import os #operating system import to get paths
import matplotlib.pyplot as plt #to plot graphs
import pandas as pd #to read/write data
#Reading the dataset
base_path = os.getcwd()
li... | pd.read_csv(path + 'iris.csv') | pandas.read_csv |
# -*- coding: utf-8 -*-
"""
bootstrapping based on event rate on one animal
num_bs_replicates=50000 takes too much time, I did 1000 instead.
Results with narrowest CIs: (space: 60%,75%,90% exceedance, 40-90 state threshold )
CASE: SpikerateCoact
animal min of exceedance & state threshold combination win... | pd.DataFrame(bsstats_all) | pandas.DataFrame |
"""
Generate figures for the DeepCytometer paper for v8 of the pipeline.
Environment: cytometer_tensorflow_v2.
We repeat the phenotyping from klf14_b6ntac_exp_0110_paper_figures_v8.py, but change the stratification of the data so
that we have Control (PATs + WT MATs) vs. Het MATs.
The comparisons we do are:
* Cont... | pd.read_pickle(dataframe_areas_filename) | pandas.read_pickle |
# -*- coding: utf-8 -*-
import os
import logging
import tempfile
import uuid
import shutil
import numpy as np
import pandas as pd
from rastertodataframe import util, tiling
log = logging.getLogger(__name__)
def raster_to_dataframe(raster_path, vector_path=None):
"""Convert a raster to a Pandas DataFrame.
... | pd.concat(tile_dfs) | pandas.concat |
# Copyright (c) 2021-2022, NVIDIA CORPORATION.
import numpy as np
import pandas as pd
import pytest
import cudf
from cudf.testing._utils import NUMERIC_TYPES, assert_eq
from cudf.utils.dtypes import np_dtypes_to_pandas_dtypes
def test_can_cast_safely_same_kind():
# 'i' -> 'i'
data = cudf.Series([1, 2, 3], d... | pd.to_numeric(ps, downcast=downcast) | pandas.to_numeric |
from __future__ import print_function, division
import GLM.constants, os, pdb, pandas, numpy, logging, crop_stats
import pygeoutil.util as util
class CropFunctionalTypes:
"""
"""
def __init__(self, res='q'):
"""
:param res: Resolution of output dataset: q=quarter, h=half, o=one
:r... | pandas.read_csv(GLM.constants.FAO_CONCOR) | pandas.read_csv |
# -*- coding: utf-8 -*-
"""
Created on Wed Sep 5 15:51:36 2018
@author: huangjin
"""
import pandas as pd
from tqdm import tqdm
import os
def gen_data(df, time_start, time_end):
df = df.sort_values(by=['code','pt'])
df = df[(df['pt']<=time_end)&(df['pt']>=time_start)]
col = [c for c in df.columns if c not ... | pd.DataFrame() | pandas.DataFrame |
import pandas as pd
from example_data_base import get_historical_data, get_connection, get_query
from sklearn.metrics.pairwise import cosine_similarity
STRONG_SIMILARITY_DIFFERENCE = 5
SMALL_SIMILARITY_DIFFERENCE = 10
FEATURES_AMOUNT = 6
# (strong similar features count, small similar features count)
records_similari... | pd.read_sql(sql, conn) | pandas.read_sql |
'''
This file includes all the locally differentially private mechanisms we designed for the SIGMOD work.
I am aware that this code can be cleaned a bit and there is a redundancy. But this helps keeping the code plug-n-play.
I can simply copy a class and use it in a different context.
http://dimacs.rutgers.edu/~graha... | pd.DataFrame(columns=["irr_l1_std", "mrr_l1_std", "iht_l1_std", "mht_l1_std", "ips_l1_std", "mps_l1_std","iolh_l1_std","icms_l1_std","icmsht_l1_std"]) | pandas.DataFrame |
import os
import numpy as np
import pandas as pd
from scipy import interp
from statsmodels.distributions import ECDF
import matplotlib
matplotlib.use('Agg')
import seaborn as sns
import matplotlib.pyplot as plt
from . import FIGURES_DIR
sns.set(context='notebook', font_scale=3.0, font='sans-serif')
sns.set_palette(s... | pd.DataFrame({'threshold': thresh[0], 'far': far, 'frr': frr}) | pandas.DataFrame |
import pandas as pd
import math
from sklearn.preprocessing import MinMaxScaler
class DataProcessor:
def __init__(self):
self.df_train = None
self.df_test = None
self.df_store = None
self.scale_y = None
'''importing data'''
def load_data(self, path):
... | pd.DataFrame() | pandas.DataFrame |
import os
import six
import inspect
import threading
import pandas as pd
import json
from tornado.gen import coroutine, Return, sleep
from tornado.httpclient import AsyncHTTPClient
from gramex.config import locate, app_log, merge, variables
from sklearn.externals import joblib
from sklearn.preprocessing import Standard... | pd.np.isnan(lo) | pandas.np.isnan |
import datetime
import hashlib
import os
import time
from warnings import (
catch_warnings,
simplefilter,
)
import numpy as np
import pytest
import pandas as pd
from pandas import (
DataFrame,
DatetimeIndex,
Index,
MultiIndex,
Series,
Timestamp,
concat,
date_range,
timedelt... | HDFStore(path) | pandas.io.pytables.HDFStore |
# -*- coding: utf-8 -*-
from __future__ import print_function
import pytest
import random
import numpy as np
import pandas as pd
from pandas.compat import lrange
from pandas.api.types import CategoricalDtype
from pandas import (DataFrame, Series, MultiIndex, Timestamp,
date_range, NaT, IntervalIn... | Timestamp("2016-01-01") | pandas.Timestamp |
#!/usr/bin/env python
# coding: utf-8
# <h2>Introduction:</h2>
# This is my First kernel, I have attempted to understand which features contribute to the Price of the houses.
# <br> A shoutout to SRK and Anisotropic from whom iv learned a lot about data visualisation</br>
# <h2>Lets import the libraries we need for n... | pd.to_datetime(data['date']) | pandas.to_datetime |
#!/usr/bin/env python
# coding: utf-8
# In[1]:
import requests
import numpy as np
import pandas as pd
from io import BytesIO
import PyPDF2
from bs4 import BeautifulSoup
from functools import reduce
from alphacast import Alphacast
from dotenv import dotenv_values
API_KEY = dotenv_values(".env").get("API_KEY")
alphac... | pd.to_numeric(df_aviar[df_aviar.columns[0]], errors='coerce') | pandas.to_numeric |
import pandas as pd
import streamlit as st
import yfinance as yf
@st.experimental_memo(max_entries=1000, show_spinner=False)
def get_asset_splits(ticker, cache_date):
return yf.Ticker(ticker).actions.loc[:, 'Stock Splits']
@st.experimental_memo(max_entries=50, show_spinner=False)
def get_historical_prices(ticke... | pd.Timestamp.now() | pandas.Timestamp.now |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from geoedfframework.utils.GeoEDFError import GeoEDFError
from geoedfframework.GeoEDFPlugin import GeoEDFPlugin
import pandas as pd
""" Module for implementing the DateTimeFilter. This supports a date time string pattern
that specifies the kinds of values that will ... | pd.to_datetime(self.end,format='%m/%d/%Y %H:%M:%S') | pandas.to_datetime |
"""Load remote meet data to DB."""
import copy
import datetime
import json
import logging
import os
import sys
from io import StringIO
from urllib.error import HTTPError
from urllib.request import Request, urlopen
import pandas as pd
from codetiming import Timer
from data.models.meets import Meet
from data.models.syn... | pd.DataFrame([df_meet]) | pandas.DataFrame |
# ---
# jupyter:
# jupytext:
# cell_metadata_filter: -all
# comment_magics: true
# formats: ipynb,py:percent
# text_representation:
# extension: .py
# format_name: percent
# format_version: '1.3'
# jupytext_version: 1.13.8
# kernelspec:
# display_name: Python 3 (ipykernel... | pd.read_csv(AltairSaver.path + f"/tables/{filename}.csv") | pandas.read_csv |
import random
from copy import deepcopy
import logging
import json, gzip
from tqdm import tqdm
import pandas as pd
from LeapOfThought.artiset import ArtiSet
from LeapOfThought.resources.teachai_kb import TeachAIKB
from LeapOfThought.common.data_utils import pandas_multi_column_agg
from LeapOfThought.common.file_utils i... | pd.DataFrame(preds) | pandas.DataFrame |
#!/usr/bin/env python
# coding: utf-8
# 1 Import libraries and Set path
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
import statsmodels.api as sm
import scipy.stats as scs
from scipy.stats.mstats import winsorize
from scipy.stats.mstats import gmean
from tabulate import ... | pd.isnull(df48['Market Index']) | pandas.isnull |
#!/usr/bin/env python
# -*- coding: utf-8; -*-
# Copyright (c) 2020, 2022 Oracle and/or its affiliates.
# Licensed under the Universal Permissive License v 1.0 as shown at https://oss.oracle.com/licenses/upl/
import logging
import time
import sys
import warnings
from abc import ABC, abstractmethod, abstractproperty
i... | pd.DataFrame(info) | pandas.DataFrame |
import itertools
import numpy as np
import pandas as pd
import pytest
from staircase import Stairs
def _expand_interval_definition(start, end=None, value=1):
return start, end, value
def _compare_iterables(it1, it2):
it1 = [i for i in it1 if i is not None]
it2 = [i for i in it2 if i is not None]
i... | pd.Series([1, 0, 1, 0, 1, 0]) | pandas.Series |
"""
Prelim script for looking at netcdf files and producing some trends
Broken into three parts
Part 1 pull out the NDVI from the relevant sites
"""
#==============================================================================
__title__ = "Time Series Chow Test"
__author__ = "<NAME>"
__version__ = "v1.0(27.02.2019... | pd.read_csv(infile+"AnnualMax.csv", index_col="sn") | pandas.read_csv |
# %% [markdown]
#
# # Comprehensive Exam
#
# ## Coding Artifact
#
# <NAME>
#
# Nov 20, 2020
# ## Model Selection
#
# Base selection of regressors is performed by fitting multiple regressors without
# performing any parameter tuning, then comparing the resulting errors across
# functional groups. Models with lower erro... | pd.set_option("display.max_rows", 120) | pandas.set_option |
import pandas as pd
import numpy as np
import git
import os
import sys
from pathlib import Path
import matplotlib.pyplot as plt
#-- Setup paths
# Get parent directory using git
repo = git.Repo("./", search_parent_directories=True)
homedir = repo.working_dir
# Change working directory to parent directory
os.chdir(ho... | pd.read_csv(cluster_ref_fln) | pandas.read_csv |
import numpy as np
import pandas as pd
import nibabel as nib
from nilearn import plotting
class SubjectAnalyzer:
def __init__(self,subject_nii_path,mean_nii_path,sd_nii_path,atlas_nii_path):
'''Get paths for files'''
self.subject_nii_path = subject_nii_path
self.mean_nii_path = mean_nii_p... | pd.Series([4, 2], index=['Values', 'Z-scores']) | pandas.Series |
# -*- coding: utf-8 -*-
# pylint: disable=W0612,E1101
from datetime import datetime
import operator
import nose
from functools import wraps
import numpy as np
import pandas as pd
from pandas import Series, DataFrame, Index, isnull, notnull, pivot, MultiIndex
from pandas.core.datetools import bday
from pandas.core.n... | assert_panel_equal(result, expected) | pandas.util.testing.assert_panel_equal |
# Copyright (c) 2020-2022, NVIDIA CORPORATION.
import datetime
import operator
import re
import cupy as cp
import numpy as np
import pandas as pd
import pytest
import cudf
from cudf.core._compat import PANDAS_GE_120
from cudf.testing import _utils as utils
from cudf.testing._utils import assert_eq, assert_exceptions... | pd.Timedelta(34765, unit="D") | pandas.Timedelta |
import requests
from bs4 import BeautifulSoup
import pandas as pd
import re
from datetime import timedelta
from datetime import date
scraped_job_titles = []
scraped_job_locations = []
scraped_company_names = []
scraped_salaries = []
scraped_ratings = []
scraped_apply_urls = []
scraped_days = []
scraped_d... | pd.DataFrame() | pandas.DataFrame |
"""
Copyright 2020 The Google Earth Engine Community Authors
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
https://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed ... | pd.DataFrame(ds, columns=pai_names) | pandas.DataFrame |
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not us... | pd.to_numeric(arg, errors=errors) | pandas.to_numeric |
from util import load_csv_as_dataframe
import pandas as pd
from feature_extractor import FeatureExtractor
import numpy as np
import pickle
from dateutil import parser
import monthdelta
import csv
from util import read_csv_file
from dateutil.relativedelta import relativedelta
import timeit
class LeaderBoard():
def... | pd.to_numeric(lb1_lb2['LB2']) | pandas.to_numeric |
# Copyright 2018-2019 QuantumBlack Visual Analytics Limited
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# THE SOFTWARE IS PROVIDED "AS IS"... | assert_frame_equal(result, expected) | pandas.util.testing.assert_frame_equal |
import datetime
import os
import re
import requests
import urllib.parse
import time
from bs4 import BeautifulSoup
import html2text
import numpy as np
import pandas
search_key_word = 'climate'
search_key = 's'
url = r'https://thebfd.co.nz/'
link_list_data_file_path = 'url-data.csv'
delay_time_min = 0.
delay_time_ma... | pandas.DataFrame(page_lists) | pandas.DataFrame |
import re
import numpy as np
import pandas as pd
import altair as alt
import streamlit as st
from vega_datasets import data
df = pd.read_csv('suicide_population.csv')
def getCountry(s):
# Get country name from country-year string
country = ""
return country.join(re.findall(r"\D",s))
def getYear(s):
#... | pd.read_csv('suicide_population.csv') | pandas.read_csv |
#Rule 24 - Description and text cannot be same.
def description_text(fle, fleName, target):
import re
import os
import sys
import json
import openpyxl
import pandas as pd
from pandas import ExcelWriter
from pandas import ExcelFile
file_name="Description_text_not_same.py"
configFile = 'https://s3.us-east.clou... | ExcelWriter(target, engine='openpyxl', mode='a') | pandas.ExcelWriter |
# pylint: disable-msg=E1101,W0612
from datetime import datetime, timedelta
import nose
import numpy as np
import pandas as pd
from pandas import (Index, Series, DataFrame, Timestamp, isnull, notnull,
bdate_range, date_range, _np_version_under1p7)
import pandas.core.common as com
from pandas.compa... | ct('10ms') | pandas.tseries.timedeltas._coerce_scalar_to_timedelta_type |
import json
import os
import dash
import dash_core_components as dcc
import dash_html_components as html
from dash.dependencies import Input, Output, State, MATCH
import plotly.express as px
import pandas as pd
## DATA FROM https://github.com/CSSEGISandData/COVID-19/tree/master/csse_covid_19_data/csse_covid_19_time_... | pd.read_csv(url) | pandas.read_csv |
import logging
import os
import os.path as op
import sys
from copy import copy, deepcopy
import cobra.flux_analysis
import cobra.manipulation
import numpy as np
import pandas as pd
from Bio import SeqIO
from cobra.core import DictList
from slugify import Slugify
import ssbio.core.modelpro
import ssbio.databases.ncbi
... | pd.DataFrame.from_records(appender, columns=cols) | pandas.DataFrame.from_records |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.