prompt
stringlengths
19
1.03M
completion
stringlengths
4
2.12k
api
stringlengths
8
90
#!/usr/bin/env/python import argparse import numpy as np import os import pandas as pd import yaml import micro_dl.inference.evaluation_metrics as metrics import micro_dl.utils.aux_utils as aux_utils import micro_dl.utils.preprocess_utils as preprocess_utils import micro_dl.utils.image_utils as image_utils import mic...
pd.DataFrame()
pandas.DataFrame
import numpy as np import pytest import pandas as pd from pandas import DataFrame, Series, date_range, timedelta_range import pandas._testing as tm class TestTimeSeries: def test_contiguous_boolean_preserve_freq(self): rng = date_range("1/1/2000", "3/1/2000", freq="B") mask = np.zeros(len(rng), ...
pd.date_range("2000", periods=2, tz=tz)
pandas.date_range
import pandas as pd import numpy as np from pathlib import Path from sklearn.manifold import TSNE from sklearn.cluster import KMeans import seaborn as sns import matplotlib.pyplot as plt import configparser from dateutil.parser import parse import os from sklearn.metrics import roc_auc_score, f1_score, precision_score,...
pd.read_csv(self.train_file)
pandas.read_csv
import numpy as np import pytest from pandas.core.dtypes.generic import ABCIndex import pandas as pd import pandas._testing as tm from pandas.core.arrays.integer import ( Int8Dtype, UInt32Dtype, ) def test_dtypes(dtype): # smoke tests on auto dtype construction if dtype.is_signed_integer: a...
pd.Series([1, 2, 3], dtype=dtype)
pandas.Series
import numpy as np import pandas as pd class DictUtil: ks = [] ds = [] def to_kv(self, src): for k, v in src.items(): self.ks.append(k) if type(v) == dict: self.to_kv(v) else: self.ds.append(np.array([".".join(sel...
pd.merge(m_df, right_df, on='key', how='left')
pandas.merge
import properties from sklearn.neighbors import NearestNeighbors from sklearn.metrics.pairwise import cosine_similarity import json import pandas as pd import numpy as np import utility import ast # Feature engineering family history def create_cols_family_hist(x): if x["tschq04-1"] == "YES": if isinstan...
pd.read_pickle(properties.simulate_hearing_file_location)
pandas.read_pickle
import matplotlib.pyplot as plt # %matplotlib inline # from utils import utils # import utils.utils as utils from utils.qedr.eval.hinton import hinton import os import numpy as np from utils.qedr.eval.regression import normalize, entropic_scores, print_table_pretty, nrmse from utils.qedr.zero_shot import get_gap_ids f...
pd.DataFrame(data=gts)
pandas.DataFrame
# import tabula import pandas as pd import numpy as np # !pip install tabula-py import camelot import os import string import pytz from datetime import datetime, timezone, timedelta from tzlocal import get_localzone from StatusMsg import StatusMsg from tqdm import tqdm from urllib.error import HTTPError im...
pd.read_csv(base_csv)
pandas.read_csv
# load libraries from sklearn import preprocessing from sklearn.pipeline import Pipeline import pandas as pd raw_data = {'first_name': ['Jason', 'Molly', 'Tina', 'Jake', 'Amy'], 'last_name': ['Miller', 'Jacobson', 'Ali', 'Milner', 'Cooze'], 'age': [42, 52, 36, 24, 73], 'city': ['San Francisco'...
pd.get_dummies(df['city'])
pandas.get_dummies
import pandas as pd #set the Quarter, suold be yearQn Quarter = "2020Q1" #read CSV file: quote and contract quotes =
pd.read_csv("./all quotes.csv")
pandas.read_csv
import argparse, pandas, os, random, seaborn, sys, re import numpy as np from unicodedata import name from numpy import median import matplotlib.pyplot as plt names_to_translate = { 'gflop_per_s_per_iter': 'Throughput [Gflop/s]', 'gbyte_per_s_per_iter': 'Bandwidth [GB/s]', 'runtime_problem_sizes_dict': 'P...
pandas.read_json(file)
pandas.read_json
"""add comment in script explaining what its for This is where the scripts to prepross the data go save files in data/targets/ """ import itertools import json import os import sys from datetime import datetime import numpy as np import pandas as pd from google_drive_downloader import GoogleDriveDownloader as gdd fro...
pd.read_csv(icu_dest)
pandas.read_csv
""" json 불러와서 캡션 붙이는 것 """ import json import pandas as pd path = './datasets/vqa/v2_OpenEnded_mscoco_train2014_questions.json' with open(path) as question: question = json.load(question) # question['questions'][0] # question['questions'][1] # question['questions'][2] df = pd.DataFrame(question['questions']) d...
pd.DataFrame(cap)
pandas.DataFrame
""" A warehouse for constant values required to initilize the PUDL Database. This constants module stores and organizes a bunch of constant values which are used throughout PUDL to populate static lists within the data packages or for data cleaning purposes. """ import importlib.resources import pandas as pd import ...
pd.StringDtype()
pandas.StringDtype
# 导入相关库 import requests import json import time import pandas as pd # import fool from PIL import Image,ImageSequence import numpy as np from wordcloud import WordCloud,ImageColorGenerator import matplotlib.pyplot as plt # 获取微博ID def getWeibo_id(): content_parameter = [] # 用来存放weibo_id值 # 获取每条微博的id值 url ...
pd.DataFrame(feature, columns=["性别", "年龄", "星座", "国家城市"])
pandas.DataFrame
import copy import os import pandas as pd import numpy as np import tempfile import skimage.io as io from toffy import rosetta import toffy.rosetta_test_cases as test_cases from ark.utils import test_utils from ark.utils.load_utils import load_imgs_from_tree from ark.utils.io_utils import list_folders, list_files f...
pd.read_csv(rosetta_path, index_col=0)
pandas.read_csv
"""This module contains the HydroMonitor object for reading groundwater head measurements from a HydroMonitor csv export file """ from collections import OrderedDict import warnings import os.path import errno import os import matplotlib as mpl import matplotlib.pyplot as plt from pandas import Series, DataFrame imp...
Series(data=heads,index=datetimes)
pandas.Series
import pandas as pd import transformer.result.generator as generator from transformer.result.result_config import ResultFormatterConfig, ResultFieldFormat class AbstractResultFormatter: def run(self, config:dict, frames: dict[str, pd.DataFrame]): pass class DefaultArrayResultFormatter(AbstractResultFormatter): ...
pd.concat(data, axis=1)
pandas.concat
#!/usr/bin/env python3 """Module to calculate reliability of samples of raw accelerometer files.""" import pandas as pd import matplotlib.pyplot as plt import datetime import argparse import os def main(): """ Application entry point responsible for parsing command line requests """ parser = argparse...
pd.read_csv(input_file)
pandas.read_csv
import os import numpy as np import pandas as pd from numpy import abs from numpy import log from numpy import sign from scipy.stats import rankdata import scipy as sp import statsmodels.api as sm from data_source import local_source from tqdm import tqdm as pb # region Auxiliary functions def ts_sum(df, window=10): ...
pd.DataFrame(na_lwma, index=df.index, columns=['CLOSE'])
pandas.DataFrame
import torch import json import pandas as pd import numpy as np from tqdm import tqdm import src.config as config import src.model_utils as mutils from src.dataset import CustomDataset def predict(df, model, device, label_list, description_col=config.TEXT_COLUMN): test_dataset = CustomDataset( desc...
pd.read_csv(data)
pandas.read_csv
from scipy import signal import numpy as np import pandas as pd from scipy.signal import filtfilt, butter import sympy as sp import math def interpolate(data): print("STATUS: Filling NaNs") unfixed = 0 for index in range(0, data.shape[0]): amount_before = data.loc[index, "interval_data"].isnull()...
pd.DataFrame(filtered_interval_data)
pandas.DataFrame
""" Provide a generic structure to support window functions, similar to how we have a Groupby object. """ from collections import defaultdict from datetime import timedelta from textwrap import dedent from typing import List, Optional, Set import warnings import numpy as np import pandas._libs.window as libwindow fro...
Substitution(name="rolling")
pandas.util._decorators.Substitution
from __future__ import unicode_literals, division, print_function import os import unittest import pandas as pd import numpy as np import warnings from itertools import product from pymatgen.core.structure import Structure from pymatgen.util.testing import PymatgenTest from sklearn.dummy import DummyRegressor, Dummy...
pd.DataFrame({'x': [1, 2, 3]})
pandas.DataFrame
#!/usr/bin/python3 # Functions to handle Input ############################################################################################# def read_csv(): # simple function to read data from a file data = pd.read_csv('out.csv', sep=';') return data def read_sacct(): # function to read the data directly fro...
pd.DataFrame()
pandas.DataFrame
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Dec 21 11:14:57 2021 @author: carlos """ import pandas as pd from wordcloud import WordCloud import matplotlib.pyplot as plt from parse_abstracts import * import seaborn as sns import numpy as np import json import re import nltk from nltk.corpus import...
pd.DataFrame(columns=masked_sepex_concepts)
pandas.DataFrame
import numpy as np import pandas as pd import datetime as dt import pickle import bz2 from .analyzer import summarize_returns DATA_PATH = '../backtest/' class Portfolio(): """ Portfolio is the core class for event-driven backtesting. It conducts the backtesting in the following order: 1. Initializati...
pd.Series()
pandas.Series
# -*- coding: UTF-8 -*- ''' @author: Andrewzhj @contact: <EMAIL> @file: comment_words_cloud.py @time: 10/16/18 3:53 PM @desc: 提取评论数据,进行热词展示 @note: ''' import jieba from wordcloud import WordCloud, ImageColorGenerator import pandas as pd from pymongo import MongoClient import numpy from PIL import Image import matplotl...
pd.DataFrame()
pandas.DataFrame
# Copyright 2018-2021 The Salish Sea NEMO Project and # The University of British Columbia # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # U...
pd.merge(Chl2017,staMap2017,how='inner', left_on=['Station'], right_on = ['Station'])
pandas.merge
# Dashboard Interativo com Streamlit, Folium e Plotly Para Monitoramento de Casos de Covid-19 em Tempo Real # Execute no terminal: streamlit run Mini-Projeto1.py # Imports import json import folium import requests import mimetypes import http.client import pandas as pd import streamlit as st import plotl...
pd.read_csv('dados/country-coordinates-world.csv')
pandas.read_csv
#### Master Script 5: Assess CPM_MNLR and CPM_POLR performance #### # # <NAME> # University of Cambridge # email address: <EMAIL> # ### Contents: # I. Initialisation # II. Create bootstrapping resamples (that will be used for all model performance evaluation) # III. Prepare compiled CPM_MNLR and CPM_POLR testing set pr...
pd.read_csv('../cross_validation_splits.csv')
pandas.read_csv
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Graficar rendimientos de los padres en periodos desconocidos. """ import pickle import numpy as np import pandas as pd import matplotlib.pyplot as plt from Simulacion import Optimizacion from Simulacion import Graficos from Simulacion import Genetico from Simulacion i...
pd.value_counts(padres[:,i+1])
pandas.value_counts
#!/usr/bin/env python3.7 # -*- coding: utf-8 -*- """ Created on Mon Nov 23 11:46:57 2020 @author: reideej1 :DESCRIPTION: Evaluate coaching data for the last 50 years of college football - the goal is to determine how coaches who struggle in their first 3 years fare over time at the same program :REQUIRES...
pd.DataFrame()
pandas.DataFrame
# coding=utf=8 import numpy as np import pandas as pd # from pandas.tools.plotting import bootstrap_plot import matplotlib.pyplot as plt import dataviz.utils as utils import matplotlib matplotlib.style.use('ggplot') def files2dataframe(root_dir: str, expression: str, offset: int, sort_columns: dict, size: int) -> p...
pd.DataFrame(errors, columns=column)
pandas.DataFrame
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ qualify donor data """ # %% REQUIRED LIBRARIES import os import argparse import json import ast import pandas as pd import datetime as dt import numpy as np # %% USER INPUTS (choices to be made in order to run the code) codeDescription = "qualify donor data" parser...
pd.DataFrame()
pandas.DataFrame
# -*- coding: utf-8 -*- """ Created on Tue May 3 10:49:58 2016 Auger peak finding and quantitative routines ... batch processing @author: tkc First get it working for single file. """ #%% import pandas as pd import numpy as np import os, sys, shutil, glob, re if 'C:\\Users\\tkc\\Documents\\Python_Scripts' n...
pd.read_csv('C:\\Users\\tkc\\Documents\\Python_Scripts\\AESquantparams.csv', encoding='utf-8')
pandas.read_csv
# -*- coding: utf-8 -*- """hood_event_scrape_module Authors: <NAME> <EMAIL> <NAME> <EMAIL> <NAME> <EMAIL> Imports to: WhatsUp_main_gui.py """ # !pip install rtree # !pip install geopandas # !pip install beautifulsoup4 # !pip install censusgeocode # Import libraries import rtree import geopandas as gp import num...
pd.DataFrame()
pandas.DataFrame
import datetime from datetime import timedelta from distutils.version import LooseVersion from io import BytesIO import os import re from warnings import catch_warnings, simplefilter import numpy as np import pytest from pandas.compat import is_platform_little_endian, is_platform_windows import pandas.util._test_deco...
tm.assert_class_equal(result.index, df.index, obj="dataframe index")
pandas.util.testing.assert_class_equal
#Copyright 2013 <NAME> (<EMAIL>) # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed t...
pd.DataFrame([data.extension, allcoords.ra, allcoords.dec, allcoords.l, allcoords.b, allcoords.ebv, data.z, data.flag_sdss])
pandas.DataFrame
import numpy as np import pandas as pd import time import sys def bill_calculator(load_profile, tariff): def pre_processing_load(load_profile): # placeholder for a quick quality check function for load profile # make sure it is kwh # make sure it is one year # make sure it doesn't...
pd.to_datetime(load_profile_f['READING_DATETIME'])
pandas.to_datetime
# NB: You have to run main_sampling.py in order for this script to function import numpy as np import pandas as pd import pickle import datetime from seir.sampling.model import SamplingNInfectiousModel import matplotlib.pyplot as plt import seaborn as sns import scipy.stats as st import logging logging.basicConfi...
pd.read_csv(f'data/sampling-runs/run{run:02}_resample.csv')
pandas.read_csv
import os import glob import pandas as pd # dict matching target to download link source_dict = {'Deaths': 'https://raw.githubusercontent.com/CSSEGISandData/COVID-19/master/csse_covid_19_data/' \ 'csse_covid_19_time_series/time_series_covid19_deaths_global.csv', 'Cases': 'https://raw....
pd.to_datetime(df.date)
pandas.to_datetime
import logging from pathlib import Path import altair as alt import pandas as pd import requests import streamlit as st import streamlit.components.v1 as components st.set_page_config(layout="wide") st.title("Bundes-Notbremse Ampel") pd.set_option('precision', 2) def is_covid_file_up_to_date(): covid_path = Path...
pd.DataFrame({'y': [100]})
pandas.DataFrame
""" inspiration from R Package - PerformanceAnalytics """ from collections import OrderedDict import pandas as pd import numpy as np from tia.analysis.util import per_series PER_YEAR_MAP = { 'BA': 1., 'BAS': 1., 'A': 1., 'AS': 1., 'BQ': 4., 'BQS': 4., 'Q': 4., 'QS': 4., 'D': 365....
pd.expanding_count(below)
pandas.expanding_count
# %% Import import numpy as np import pandas as pd import requests import os from bs4 import BeautifulSoup """ Takes a dictionary of relevant brands and their URLs and returns a raw csv file """ # %% Functions def outlets_crawl(brand, url): """ Returns a raw, unformatted df of outlets with it's brand from t...
pd.DataFrame(_ls)
pandas.DataFrame
import pandas as pd import numpy as np import seaborn as sb import base64 from io import BytesIO from flask import send_file from flask import request from napa import player_information as pi import matplotlib matplotlib.use('Agg') # required to solve multithreading issues with matplotlib within flask import matplotli...
pd.DataFrame(perms_joined, columns = ['permutation','r1','r2'])
pandas.DataFrame
""" Getting final clusters data :Author: <NAME> <<EMAIL>> :Date: 2019-09-30 :License: MIT """ # Import Libraries import pandas as pd import numpy as np import random from collections import Counter def main(): # clusters group user id clusters_users = pd.read_csv('clusters_final.csv') # clusters group user id ...
pd.read_csv('clusters_group_names.csv')
pandas.read_csv
# Making prediction about diagnostic labels of the subjects. Note that this file needs # the output of 'fit/gql_ml_pred.py'. from BD.sim.rnn_label_pred import finding_CV from actionflow.data.data_process import DataProcess from actionflow.qrl.gql import GQL from actionflow.qrl.opt_ml import OptML from actionflow.util ...
pd.DataFrame({'id': ids, 'train': 'train'})
pandas.DataFrame
"""The classes for specifying and compiling a declarative visualization.""" from __future__ import annotations import io import os import re import sys import inspect import itertools import textwrap from collections import abc from collections.abc import Callable, Generator, Hashable from typing import Any import pa...
pd.option_context("mode.use_inf_as_null", True)
pandas.option_context
import calendar from datetime import date, datetime, time import locale import unicodedata import numpy as np import pytest import pytz from pandas._libs.tslibs.timezones import maybe_get_tz from pandas.core.dtypes.common import is_integer_dtype, is_list_like import pandas as pd from pandas import ( DataFrame, ...
tm.assert_index_equal(result, expected)
pandas.util.testing.assert_index_equal
import pytest import sys import numpy as np import swan_vis as swan import networkx as nx import math import pandas as pd import anndata ########################################################################### ################# Related to input/error handling ######################### ##############################...
pd.DataFrame(data=data, columns=cols)
pandas.DataFrame
# coding: utf-8 # In[1]: import numpy as np import pandas as pd import matplotlib.pyplot as plt import math # ### 讀入資料 # - 取10~12月資料 # - 將NR轉換成0 # In[2]: data = pd.read_excel('hsinchu.xls') #10~12 data = data[data['日期'].between('2017/10/01','2017/12/31 ')] # NR->0 data.replace('NR',0, inplace=True) # In[3]: ...
pd.DataFrame(test_data_18)
pandas.DataFrame
from matplotlib import pyplot as plt import pandas as pd import numpy as np import os wd = os.chdir('/Users/larslarson/Documents/School/CU/Research/TADD/Data/2022-03-04') filepaths = [f for f in os.listdir(wd) if f.endswith('.csv')] df = pd.concat(map(pd.read_csv, filepaths),axis='columns') file_names = [] data_frame...
pd.concat(data_frames, axis=1)
pandas.concat
# -*- coding: utf-8 -*- """ Authors: <NAME> UNESCO-IHE 2016 Contact: <EMAIL> Repository: https://github.com/wateraccounting/wa Module: Sheets/sheet1 """ import os import pandas as pd import time import xml.etree.ElementTree as ET import subprocess def create_sheet3(basin, period, units, data, output, templat...
pd.isnull(lp_r04c07)
pandas.isnull
from typing import Any import numpy as np import numpy.testing as npt import pandas as pd import pytest from sklearn.preprocessing import PowerTransformer from etna.datasets import TSDataset from etna.transforms.power import BoxCoxTransform from etna.transforms.power import YeoJohnsonTransform @pytest.fixture def n...
pd.date_range("2021-06-01", "2021-07-01", freq="1d")
pandas.date_range
# Arithmetic tests for DataFrame/Series/Index/Array classes that should # behave identically. # Specifically for Period dtype import operator import numpy as np import pytest from pandas._libs.tslibs.period import IncompatibleFrequency from pandas.errors import PerformanceWarning import pandas as pd from pandas impo...
tm.box_expected(pi, box_with_array)
pandas.util.testing.box_expected
import argparse import json import logging import os import sys import warnings from itertools import product import numpy as np import pandas as pd import torch from paccmann_chemistry.models import (StackGRUDecoder, StackGRUEncoder, TeacherVAE) from paccmann_chemistry.utils import get_device from paccmann_generator...
pd.DataFrame({'loss': rl_losses, 'rewards': rewards})
pandas.DataFrame
# -*- coding: utf-8 -*- """ Created on Thu Nov 18 15:43:55 2021 @author: ZeitgeberH """ from pathlib import Path from PyQt5 import QtGui, QtCore, QtSvg from pyqtgraph.Qt import QtWidgets from PyQt5.QtWidgets import QMessageBox, QTableWidgetItem import pyqtgraph as pg import pyqtgraph.opengl as gl from pyqt...
pd.DataFrame()
pandas.DataFrame
import pandas as pd import numpy as np from datetime import date """ dataset split: (date_received) dateset3: 20160701~20160731 (113640),features3 from 20160315~20160630 (off_test) dateset2: 20160515~20160615 (258446),features2 from 20160201~2...
pd.merge(user2_feature,t13,on='user_id',how='left')
pandas.merge
import requests from bs4 import BeautifulSoup import pandas as pd pages=list(range(0,250,25)) def request_douban(url): htmls=[] headers = { 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0 Win64 x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/96.0.4664.110 Safari/537.36 Edg/96.0.1054.62' } try: ...
pd.DataFrame(results)
pandas.DataFrame
# coding: utf8 import os import numpy as np from tqdm import tqdm import pandas as pd import time from upsetplot import plot from matplotlib import pyplot from scipy import stats from intervaltree import IntervalTree def PeakOverlap(genesfile, peaksfile,tssdistance=[0,0],peakname='null'): LuckPeak, LuckGen, Luck...
pd.merge(peaktable, UTR3table, left_on='Gen_TransID', right_on='Gen_TransID', how='inner')
pandas.merge
import dataclasses from collections import namedtuple from copy import deepcopy, copy from typing import NoReturn import numpy as np import pandas as pd from numpy import datetime64 from sklearn.pipeline import Pipeline from sklearn.preprocessing import FunctionTransformer from IMLearn import BaseEstimator from chall...
pd.get_dummies(df, columns=cat_vars)
pandas.get_dummies
#! /usr/bin/env python3 """My Podcaster.""" import datetime import email.utils from subprocess import call, check_output import mimetypes import os import re import shutil import socket import urllib.error import urllib.request import requests import tqdm import random import signal from Podcast import Podcast import c...
pandas.DataFrame(columns=["Podcast", "Title"])
pandas.DataFrame
#!/usr/bin/python3 # -*- coding: utf-8 -*- # *****************************************************************************/ # * Authors: <NAME> # *****************************************************************************/ """transformCSV.py This module contains the basic functions for creating the content of...
pandas.StringDtype()
pandas.StringDtype
import sys import os import traceback from shapely.geometry import Point import core.download as dlf import pandas as pd import geopandas as gpd def err_to_parent(UDF): def handling(connection, load, message): try: UDF(connection, load, message) except Exception as e: ...
pd.read_pickle(filepath, **message['read_args'])
pandas.read_pickle
import os from distutils.util import strtobool import numpy as np import pytest import opendp.smartnoise.core as sn from tests import (TEST_PUMS_PATH, TEST_PUMS_NAMES) # Used to skip showing plots, etc. # IS_CI_BUILD = strtobool(os.environ.get('IS_CI_BUILD', 'False')) def test_multilayer_analysis(run=True): wi...
pd.DataFrame(non_dp_corr)
pandas.DataFrame
#%% # A. Importing packages, necessary datasets and concluding to our final dataset # i. Importing the packages import pandas as pd import numpy as np import matplotlib.pyplot as plt from pandas_datareader import wb import ipywidgets as widgets #%% # ii. Dowloading data from the World Bank (Countries, Years and GD...
pd.read_excel(country_codes)
pandas.read_excel
from collections import namedtuple import pandas as pd import numpy as np Scores = namedtuple('Scores', ['Benign', 'Likely_benign', 'Uncertain_significance', 'not_provided', 'Conflicting_interpretations_of_pathogenicity', 'Likely_pathogenic', 'Pathogenic', ...
pd.read_csv(clinvar_variant_summary_file, sep="\t")
pandas.read_csv
# # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not us...
pd.Series([100], index=["A"])
pandas.Series
def NMDS_analysis(TaXon_table_xlsx, meta_data_to_test, taxonomic_level, width, height, nmds_s, max_iter_val, n_init_val, path_to_outdirs, template, font_size, color_discrete_sequence, nmds_dissimilarity): import pandas as pd import numpy as np from skbio.diversity import beta_diversity from sklearn.man...
pd.DataFrame(nmds_results_dict[2]["nmds_results"], index=[samples])
pandas.DataFrame
#!/usr/bin/env python # -*- coding:utf-8 -*- """ Date: 2022/1/26 13:10 Desc: 申万指数-申万一级、二级和三级 http://www.swsindex.com/IdxMain.aspx https://legulegu.com/stockdata/index-composition?industryCode=851921.SI """ import time import json import pandas as pd from akshare.utils import demjson import requests from bs4 import Bea...
numeric(temp_df["昨收盘"])
pandas.to_numeric
import datareader import dataextractor import bandreader import numpy as np from _bisect import bisect import matplotlib.pyplot as plt import matplotlib.ticker as plticker import pandas as pd from scipy import stats from sklearn import metrics def full_signal_extract(path, ident): """Extract breathing and heartbe...
pd.DataFrame()
pandas.DataFrame
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue Mar 16 11:36:13 2021 @author: <NAME> (Finnish Meteorological Institute) """ import sys import pandas as pd import numpy as np import xarray as xr from satellitetools.biophys import SNAP_BIO_RMSE def xr_dataset_to_timeseries( xr_dataset, variab...
pd.to_datetime(xr_dataset.time.values)
pandas.to_datetime
# Module: Bachelor thesis # Theme: Detect malicious/unusual Login Events # Author: <NAME> <<EMAIL>> # Status: 28.07.2021 import datetime import pandas as pd import re import numpy as np from joblib import dump def read_features(data_path): features = pd.read_csv(data_...
pd.DataFrame(columns=features_without_scores.columns)
pandas.DataFrame
import igraph as Graph import pandas as pd import os import numpy as np import spacy from sklearn.cluster import KMeans from pylab import * import re import time import src.pickle_handler as ph import src.relation_creator as rc # the dataframe has been preprocessed by many other functions. However we only need a subs...
pd.concat(frames, sort=False)
pandas.concat
import pandas as pd import numpy as np import re import math import codecs import csv # 预计剩余电影总量220k到200k data=
pd.read_csv("Website_ETL.CSV")
pandas.read_csv
# -*- coding:utf-8 -*- import re import logging import pandas as pd from contrib.utils.DataCleanCheckTool import DataCleanCheckTool class CorpusFromEllisQTB(object): """ CorpusFromEllis, question_text with blank 整个程序是由大量函数构成的 主要的函数是final_process,其他在final_process中调度 final_process完成之...
pd.merge(data, data_packageid, on="exercise_id", how="left")
pandas.merge
# -*- coding: utf-8 -*- """Copy of rnn.ipynb Automatically generated by Colaboratory. Original file is located at https://colab.research.google.com/drive/1hw5VX0w03qnA-pD4YmOck-HAmzP9_fO8 # Recurrent Neural Network ## Part 1 - Data Preprocessing ### Importing the libraries """ import numpy as np import matplo...
pd.read_csv('Google_Stock_Price_Test.csv')
pandas.read_csv
import pandas as pd import numpy as np from sklearn.preprocessing import OneHotEncoder from sklearn.model_selection import train_test_split from sklearn.preprocessing import MinMaxScaler import tensorflow as tf from tensorflow import keras from tensorflow.keras.models import Sequential from tensorflow.keras.layers impo...
pd.read_csv('data/wind_table_07.csv')
pandas.read_csv
import logging import textwrap import pandas from sqlalchemy import text from triage.database_reflection import table_exists from triage.component.catwalk.storage import MatrixStore class ProtectedGroupsGeneratorNoOp(object): def generate_all_dates(self, *args, **kwargs): logging.warning( "N...
pandas.DataFrame()
pandas.DataFrame
import os from typing import cast import matplotlib.pyplot as plt import pandas as pd import pandera as pa import requests import seaborn as sns from dagster_pandera import pandera_schema_to_dagster_type from pandera.typing import Series # **************************************************************************** #...
pd.read_csv(path, parse_dates=["date"])
pandas.read_csv
import logging import pandas as pd from itertools import tee, izip from copy import deepcopy from modules.loggingFunctions import initialize_logging from modules.amr.aro import ARO_ACCESSIONS DF_ARO =
pd.DataFrame(ARO_ACCESSIONS)
pandas.DataFrame
from datetime import datetime import numpy as np import pandas as pd from pandas import ( Period, Series, date_range, period_range, to_datetime, ) import pandas._testing as tm class TestCombineFirst: def test_combine_first_period_datetime(self): # GH#3367 didx = date_range(st...
tm.assert_series_equal(ser, result)
pandas._testing.assert_series_equal
import pandas as pd import numpy as np import networkx as nx from sklearn.preprocessing import StandardScaler, normalize, MinMaxScaler import matplotlib.pyplot as plt from collections import defaultdict, Counter import urllib.request as request import json import os from scipy.sparse import csr_matrix as csr_matrix fro...
pd.DataFrame(edge_list)
pandas.DataFrame
""" Hold pandas dataframe of given excel sheet Performs various read operations which all return numpy arrays """ import numpy as np import pandas as pd def clean_vector(x): return x[~np.isnan(x)] class DataHandler: def __init__(self, filepath): self._path = filepath self._df = pd.read_exce...
pd.read_excel(self._path)
pandas.read_excel
'''reports details about a virtual boinc farm''' # standard library modules import argparse import collections #import contextlib #from concurrent import futures #import errno import datetime #import getpass #import json import logging #import math #import os #import re #import socket #import shutil #import signal impo...
pd.DataFrame()
pandas.DataFrame
import pytest from pandas._libs.tslibs.frequencies import INVALID_FREQ_ERR_MSG, _period_code_map from pandas.errors import OutOfBoundsDatetime from pandas import Period, Timestamp, offsets class TestFreqConversion: """Test frequency conversion of date objects""" @pytest.mark.parametrize("freq", ["A", "Q", ...
Period(freq="A-JAN", year=2008)
pandas.Period
# Copyright (c) 2019-2020, NVIDIA CORPORATION. import datetime as dt import re import cupy as cp import numpy as np import pandas as pd import pyarrow as pa import pytest from pandas.util.testing import ( assert_frame_equal, assert_index_equal, assert_series_equal, ) import cudf from cudf.core import Data...
pd.Index([1, 2, 3, 4])
pandas.Index
# -*- coding: utf-8 -*- """ Created on Wed Sep 25 16:14:12 2019 @author: <NAME> """ import pandas as pd import numpy as np import matplotlib.pyplot as plt #import graphviz import os import seaborn as sns from scipy.stats import chi2_contingency os.chdir("E:\PYTHON NOTES\projects\cab fare prediction") d...
pd.concat([dataset_train2,temp],axis=1)
pandas.concat
import os from typing import List try: from typing import Literal except ImportError: from typing_extensions import Literal # type: ignore from typing import Optional import numpy as np import pandas as pd import scanpy as sc from anndata import AnnData from rich import print WORKING_DIRECTORY = os.path.di...
pd.concat(all_markers)
pandas.concat
''' Author: <NAME> Date: May 1, 2019 Course: ISTA355 Final Project This file contains all the functions used for my ISTA355 Final Project. The purpose of the file is to accomplish the task of incorporating question answering features to a classifier in order to replicate a open ended question answer model or search e...
pd.DataFrame(data, index=index, columns=cols)
pandas.DataFrame
import pandas as pd import os CUR_PATH = os.path.abspath(os.path.dirname(__file__)) SYR = 2011 # calendar year used to normalize factors BEN_SYR = 2014 # calendar year used just for the benefit start year EYR = 2030 # last calendar year we have data for SOI_YR = 2014 # most recently available SOI estimates # defi...
pd.DataFrame(total_pop2)
pandas.DataFrame
#!/usr/bin/env python3.7 # -*- coding: utf-8 -*- """ Created on Mon May 10 15:36:24 2021 @author: reideej1 :DESCRIPTION: Rolls up situational statistics for individual player stats contained in CFBStats/teamXXX/individual folders. Totals will be generated for each player on a yearly and a career bas...
pd.DataFrame()
pandas.DataFrame
import pandas as pd import numpy as np import tensorflow as tf from tensorflow import keras tf.random.set_seed(2021) from models import DNMC, NMC, NSurv, MLP, train_model, evaluate_model FILL_VALUES = { 'alb': 3.5, 'pafi': 333.3, 'bili': 1.01, 'crea': 1.01, 'bun': 6.51, 'wblc': 9., 'urin...
pd.read_csv('../datasets/support2.csv')
pandas.read_csv
# -*- coding: utf-8 -*- """ Created on Wed May 24 16:15:24 2017 Sponsors Club messaging functions @author: tkc """ import pandas as pd import smtplib import numpy as np import datetime import tkinter as tk import glob import re import math import textwrap from tkinter import filedialog from email.mime.multipart impor...
pd.read_csv(cnf._INPUT_DIR+'\\master_signups.csv', encoding='cp437')
pandas.read_csv
import boto3 import json import os import requests import pandas as pd import warnings from pandas import json_normalize from github import Github warnings.filterwarnings('ignore') bucket = 'wmwaredata' fileName = 'gw_releases.json' s3 ...
pd.concat(data, ignore_index=True)
pandas.concat
import sys # sys.path.append("..") # sys.path.append("../..") import os, errno import core_models.parser_arguments as parser_arguments import warnings import numpy as np import pandas as pd import core_models.utils as utils import operator from sklearn.metrics import roc_auc_score as auc_compute from sklearn.metric...
pd.DataFrame([], index=['error_repair_dirtycells','error_repair_cleancells'], columns=attributes)
pandas.DataFrame
from typing import List import pytest import numpy as np import pandas as pd from obp.dataset import ( linear_reward_function, logistic_reward_function, linear_behavior_policy_logit, SyntheticSlateBanditDataset, ) from obp.types import BanditFeedback # n_unique_action, len_list, dim_context, reward_...
pd.DataFrame()
pandas.DataFrame
import itertools from sklearn.model_selection import train_test_split from challenge.agoda_cancellation_estimator import AgodaCancellationEstimator import matplotlib.pyplot as plt from sklearn import metrics import numpy as np import pandas as pd import re PATTERN = re.compile(r"((?P<days1>[1-9]\d*)D(?P<amount1>[1...
pd.read_csv(f'labels//l{i}.csv')
pandas.read_csv
import pandas as pd import numpy as np import matplotlib.pyplot as plt from linearmodels import PanelOLS import statsmodels.api as sm import econtools as econ import econtools.metrics as mt import math from statsmodels.stats.outliers_influence import variance_inflation_factor from auxiliary.prepare import * from auxil...
pd.DataFrame((ci_1,ci_2))
pandas.DataFrame