prompt
stringlengths
19
1.03M
completion
stringlengths
4
2.12k
api
stringlengths
8
90
import os import sys import pickle import pandas as pd import numpy as np import word2vec as wv from tqdm import tqdm def get_new_dataframe_names(df_path): file_name = df_path.rsplit('/')[-1] return 'sentence_' + file_name def create_sentence_df(df): sentences = {} indices = {} classes = {} ...
pd.Series(noun_class)
pandas.Series
import detailed_table import pandas as pd from locale import atof import json # KAP 19 COL_PROFIT_STOCKS = 'Aktien G/V' COL_DIVIDENDS_STOCKS = 'Aktien Dividende' COL_PROFIT_CFDS = 'CFD G/V' COL_FEES_CFDS = 'CFD Gebühren' # KAP 20 COL_PROFIT_ON_SALE_STOCKS = 'Aktien - Enthaltene Gewinne aus Aktienveräußerungen' COL_PR...
pd.DataFrame(columns=resultColumns)
pandas.DataFrame
import pandas as pd import random from src.func import tweet_utils from src.func import regex from src.func import labmtgen from src.scripts.process_tweets import * from labMTsimple.storyLab import * def get_tweets_timestamp(park_user_tweets): """ Take a list of lists (tweets by user) and assigns a random c...
pd.isnull(tweet['ParkID'])
pandas.isnull
"""Class for creating a Parallel Pipeline.""" from iguanas.exceptions.exceptions import DataFrameSizeError, NoRulesError from iguanas.pipeline._base_pipeline import _BasePipeline from iguanas.utils.typing import PandasDataFrameType, PandasSeriesType from iguanas.utils.types import PandasDataFrame, PandasSeries, Diction...
pd.DataFrame()
pandas.DataFrame
""" This script visualises the prevention parameters of the first and second COVID-19 waves. Arguments: ---------- -f: Filename of samples dictionary to be loaded. Default location is ~/data/interim/model_parameters/COVID19_SEIRD/calibrations/national/ Returns: -------- Example use: ------------ """ __author_...
pd.to_datetime(end_calibration)
pandas.to_datetime
import pandas as pd import numpy as np import lightgbm as lgb import xgboost as xgb from sklearn.model_selection import train_test_split from sklearn.preprocessing import OneHotEncoder from sklearn.model_selection import KFold, RepeatedKFold from scipy import sparse # 显示所有列
pd.set_option('display.max_columns', None)
pandas.set_option
import pandas as pd import os import itertools from collections import defaultdict """ Description: This script performs automatic filtering/aggregation of Qualys scans intended for analysis. Reads in a csv file and outputs a csv file. """ def main(): root = os.path.dirname(os.path.abspath(__file__)) ...
pd.isna(y)
pandas.isna
# https://github.com/bokeh/bokeh/issues/5701 # https://groups.google.com/a/continuum.io/forum/#!searchin/bokeh/selected/bokeh/ft2U4nX4fVo/srBMki9FAQAJ import pandas as pd import sys import io import os.path as op import MetaVisLauncherConfig as config from bokeh.io import show, output_file, save from bokeh.embed import...
pd.DataFrame({"Columns from Col Metadata": colmd_names})
pandas.DataFrame
try: from TACT import logger except ImportError: pass from TACT.readers.config import Config from future.utils import itervalues, iteritems import pandas as pd import re import sys from string import printable import numpy as np class Data(Config): """Class to hold data, derivative features, and metadata ...
pd.read_excel(self.input_filename)
pandas.read_excel
""" Script for the calculation of A_uv, A_dv & A_g, and the python graph for the xPDFs at initial scale. The parametrizations were taken from: <NAME>., & <NAME>. (2019). A new simple PDF parametrization: improved description of the HERA data. The European Physical Journal Plus, 134(10), 531....
pd.DataFrame({'x$u_{v}$': x_uv, 'x$d_{v}$': x_dv, 'x$\\overline{u}$': x_ubar, 'x$\\overline{d}$': x_dbar, 'x$gl$': xg}, index=x)
pandas.DataFrame
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Thu Jan 14 16:13:16 2021 @author: nicolasnavarre """ import pandas as pd import math data = 'data/' def crop_yield(POM_data, fish_products, meat_products, feed_list, crop_proxie, diet_div_crop, diet_source_crop): POM_crop_data = POM_data[~POM_data.Ite...
pd.merge(Weighted_item_tot, Weighted_item_ext['% of ext'], left_index = True, right_index = True)
pandas.merge
''' __author__=<NAME> MIT License Copyright (c) 2020 crewml Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, mer...
pd.to_datetime(df['DTY_REP_TM_UTC'], utc=True)
pandas.to_datetime
from flask import Flask, session, jsonify, request import pandas as pd import numpy as np import pickle import os from sklearn.metrics import f1_score from sklearn.model_selection import train_test_split from sklearn.linear_model import LogisticRegression import json import glob #################Load config.json and ...
pd.concat(df_from_each_file, ignore_index=True)
pandas.concat
from dataExtractor import reviewToList from dataExtractor200 import dataExtractor200 import numpy as np import random """IMPORTING FILES""" reviewList_p = dataExtractor200("positive.review") reviewList_n = dataExtractor200("negative.review") X_pos_test = reviewList_p[1055:] X_neg_test = reviewList_n[665:] """Y_trai...
pd.DataFrame(cnf_matrix)
pandas.DataFrame
import numpy as np import pandas as pd from sklearn.ensemble import ExtraTreesClassifier from cause.plotter import Plotter from cause.predictor import ClassificationSet class Breakdown(): def __init__(self, data, weights, algos, name): self.__data = data self.__weights = weights self.__...
pd.DataFrame(columns=["order", "value", "name", "error"])
pandas.DataFrame
'''Python script to generate Revenue Analysis given ARR by Customer''' '''Authors - <NAME> ''' import numpy as np import pandas as pd from datetime import datetime import collections from .helpers import * class RevAnalysis: def __init__(self, json): print("INIT REV ANALYSIS") self.arr = pd.Data...
pd.to_datetime(self.rev_cohorts['Cohort'])
pandas.to_datetime
import pandas as pd import numpy as np from datetime import timedelta import matplotlib.pyplot as plt from scipy.interpolate import griddata class spatial_mapping(): def __init__(self, data, gps, gps_utc=0): df=pd.DataFrame(data) df[0]=pd.to_datetime(df[0]-693962,unit='D',origin=pd.Timestamp('1900-01-01'),ut...
pd.to_datetime(slice_df['End_time'],utc=True)
pandas.to_datetime
#!/usr/bin/env python # -- coding: utf-8 -- # PAQUETES PARA CORRER OP. import netCDF4 import pandas as pd import numpy as np import datetime as dt import json import wmf.wmf as wmf import hydroeval import glob import MySQLdb #modulo pa correr modelo import hidrologia from sklearn.linear_model import LinearRegression ...
pd.Timedelta(stepback_start)
pandas.Timedelta
import pgpasslib from sqlalchemy import create_engine import pandas as pd from pymedextcore.document import Document from .med import MedicationAnnotator def get_engine(): password = pgpasslib.getpass('10.172.28.101', 5432, '<PASSWORD>', '<PASSWORD>') return create_engine(f'postgresql+psycopg2://coronascien...
pd.DataFrame.from_records(omop)
pandas.DataFrame.from_records
import argparse import six from tqdm import tqdm import pandas as pd import string, re from nltk.translate.bleu_score import corpus_bleu from typing import List, Tuple, Dict, Set, Union def compute_corpus_level_bleu_score(references: List[str], hypotheses: List[str]) -> float: """ Given decoding results and refere...
pd.DataFrame(data_list)
pandas.DataFrame
# -*- coding: utf-8 -*- """ Created on Wed Apr 04 18:33:27 2018 @author: Prodipta """ import pandas as pd import datetime as dt import numpy as np from pyfolio.utils import extract_rets_pos_txn_from_zipline from pyfolio.timeseries import perf_stats from empyrical.stats import cum_returns_final, aggregate_returns impo...
pd.concat(frames)
pandas.concat
import numpy as np import pandas as pd import os import argparse import json import tensorflow.keras as k def readData(tumorFileName, normalFileName): x_true = pd.read_csv(tumorFileName, sep='\t', header=0, index_col=0).T x_false = pd.read_csv(normalFileName, sep='\t', header=0, index_col=0).T # if this d...
pd.concat([x_true, x_false])
pandas.concat
from surf.script_tab import keytab from surf.surf_tool import regex2pairs import os, json, time, re, codecs, glob, shutil import matplotlib.pyplot as plt import matplotlib as mpl import logging.handlers import pandas as pd import itertools import numpy as np import random import tensorflow as tf from sklearn.model_sele...
pd.Series(index=pricepd.index)
pandas.Series
import pandas as pd import numpy as np from scipy.sparse.linalg import svds from skbio import OrdinationResults from skbio.stats.composition import clr import seaborn as sns import matplotlib.pyplot as plt plt.style.use("ggplot") def apca(df): """Performs Aitchison PCA on a feature table. Parameters -----...
pd.Series(p.T, index=cols)
pandas.Series
import numpy as np import pandas as pd from collections import namedtuple import matplotlib.pyplot as plt from sklearn import preprocessing from sklearn.cluster import KMeans from fcutils.maths.geometry import calc_distance_from_point from fcutils.maths.geometry import calc_distance_between_points_in_a_vector_2d, calc...
pd.DataFrame(bts)
pandas.DataFrame
from typing import Dict from typing import Union import numpy as np import pandas as pd import pytest from etna.datasets import TSDataset from etna.transforms import ResampleWithDistributionTransform DistributionDict = Dict[str, pd.DataFrame] @pytest.fixture def daily_exog_ts() -> Dict[str, Union[TSDataset, Distri...
pd.date_range(start="2020-01-05", freq="D", periods=3)
pandas.date_range
# # Copyright 2021 Grupo de Sistemas Inteligentes, DIT, Universidad Politecnica de Madrid (UPM) # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2...
pd.DataFrame(columns=['id', 'text', 'polarity'])
pandas.DataFrame
### preprocessing """ code is taken from tunguz - Surprise Me 2! https://www.kaggle.com/tunguz/surprise-me-2/code """ import glob, re import numpy as np import pandas as pd from sklearn import * from datetime import datetime import matplotlib.pyplot as plt data = { 'tra': pd.read_csv('../input/air_visit_data.csv'...
pd.read_csv('../input/sample_submission.csv')
pandas.read_csv
from IMLearn.utils import split_train_test from IMLearn.learners.regressors import LinearRegression from typing import NoReturn import numpy as np import pandas as pd import plotly.graph_objects as go import plotly.express as px import plotly.io as pio pio.templates.default = "simple_white" def load_data(filename: ...
pd.to_datetime(df.date, errors='coerce')
pandas.to_datetime
# -*- coding: utf-8 -*- # pylint: disable=W0612,E1101 from datetime import datetime import operator import nose from functools import wraps import numpy as np import pandas as pd from pandas import Series, DataFrame, Index, isnull, notnull, pivot, MultiIndex from pandas.core.datetools import bday from pandas.core.n...
Panel({'Item1': df})
pandas.core.panel.Panel
# %% import packages import numpy as np import pandas as pd import itertools import warnings import matplotlib.pyplot as plt import matplotlib.cm as cm from matplotlib.colors import Normalize from statsmodels.tsa.arima_model import ARIMA from statsmodels.tsa.stattools import acf, pacf from statsmodels.tsa.stattools i...
CategoricalDtype(categories=cats, ordered=True)
pandas.api.types.CategoricalDtype
""" GridFrame -- subclass of wx.Frame. Contains grid and buttons to manipulate it. GridBuilder -- data methods for GridFrame (add data to frame, save it, etc.) """ import wx import pandas as pd import numpy as np from dialogs import drop_down_menus3 as drop_down_menus from dialogs import pmag_widgets as pw from dialog...
pd.DataFrame.to_clipboard(self.df_slice, header=False, index=False)
pandas.DataFrame.to_clipboard
# coding=utf-8 """ PAT - the name of the current project. instrument.py - the name of the new file which you specify in the New File dialog box during the file creation. Hossein - the login name of the current user. 6 / 15 / 18 - the current system date. 8: 03 AM - the current system time. PyCharm - the name of the IDE...
pandas.DataFrame(new_row)
pandas.DataFrame
from typing import List import pandas as pd import plotly.graph_objs as go from dash import Dash import dash_core_components as dcc import dash_html_components as html import dash_table from dash.dependencies import Input, Output GOOGLE_SHEETS_URL = "https://docs.google.com/spreadsheets/d/e/{}&single=true...
pd.merge(points_data, schedule, on=[id_col, player_col], how="left")
pandas.merge
# -*- coding: utf-8 -*- """ Created on Sat Dec 2 23:17:22 2017 @author: roshi """ import pandas as pd import matplotlib.pyplot as plt import dash import dash_core_components as dcc import dash_html_components as html import plotly.graph_objs as go from app import app data =
pd.read_csv('./data/youth_tobacco_analysis.csv')
pandas.read_csv
import pandas as pd import pyomo.environ as pe import os import shutil class invsys: def __init__(self,inp_folder='',dshed_cost=1000000,rshed_cost=500,vmin=0.8,vmax=1.2,sbase=100,ref_bus=0): """Initialise the investment problem. :param str inp_folder: The input directory for the data. It expects...
pd.DataFrame(mat)
pandas.DataFrame
import pytest from pandas import ( Index, MultiIndex, Series, ) import pandas._testing as tm class TestSeriesRenameAxis: def test_rename_axis_mapper(self): # GH 19978 mi =
MultiIndex.from_product([["a", "b", "c"], [1, 2]], names=["ll", "nn"])
pandas.MultiIndex.from_product
""" inserindo dados com pandas C - CREATE R - READ U - UPDATE D - DELETE """ import pandas as pd BASE_PATH = 'base.csv' # CREATE def post(dados: dict): df_antigo = pd.DataFrame(get()) df_novo = pd.DataFrame(dados, index=[0]) df = df_antigo.append(df_novo) df.to_csv(BASE_PATH, sep=',', index=False) ...
pd.DataFrame(lista_dados_novos)
pandas.DataFrame
import numpy as np import pandas as pd from analysis.transform_fast import load_raw_cohort, transform def test_immuno_group(): raw_cohort = load_raw_cohort("tests/input.csv") cohort = transform(raw_cohort) for ix, row in cohort.iterrows(): # IF IMMRX_DAT <> NULL | Select | Next if pd...
pd.notnull(row["learndis_dat"])
pandas.notnull
import json import datetime import re import sys import ipdb import pandas as pd def logging(message): sys.stderr.write('\r') sys.stderr.write(message) sys.stderr.flush() def clean_company_name(name): company_token = [ '^\"', '\"$', 'Inc\W', 'Inc$', 'Co\W', 'Co$', 'Corp\W', 'Corp$', '...
pd.read_csv('webhose_data.csv')
pandas.read_csv
from os import path import os.path from datetime import datetime as dt import datetime # import plotly.express as px # from dash.dependencies import Input, Output, State # import dash_html_components as html # import dash_core_components as dcc import json import pandas as pd import numpy as np # from jupyter_dash impo...
pd.DataFrame(df, columns=['code', 'nom', 'color', 'custom_data'])
pandas.DataFrame
#!/usr/bin/env python # coding: utf-8 # In[1]: import pandas as pd import numpy as np from sklearn.preprocessing import LabelEncoder # In[2]: pd.set_option('display.max_rows', 1000) pd.set_option('display.max_columns', 500) pd.set_option('display.width', 1000) # In[3]: df = pd.read_csv('data.csv') # ## Gro...
pd.DataFrame(data=top_2_rated_players, columns=cols)
pandas.DataFrame
import pandas as pd import numpy as np import seaborn as sns import matplotlib.pyplot as plt sns.set(font_scale=2.2) plt.style.use("seaborn") from sklearn.preprocessing import LabelEncoder, MinMaxScaler, StandardScaler, OneHotEncoder from sklearn.model_selection import StratifiedKFold, train_test_split, ShuffleSplit...
pd.merge(df_train, agg_train, on="idhogar")
pandas.merge
# -*- coding: utf-8 -*- """ These the test the public routines exposed in types/common.py related to inference and not otherwise tested in types/test_common.py """ from warnings import catch_warnings, simplefilter import collections import re from datetime import datetime, date, timedelta, time from decimal import De...
pd.Timestamp('2011-01')
pandas.Timestamp
#!/usr/bin/env python from scipy import interpolate import numpy as np from numpy.lib.recfunctions import append_fields import scipy.signal as sig import scipy.stats as st import time, os import pandas as pd import math #import report_ctd import ctdcal.report_ctd as report_ctd import warnings import ctdcal.fit_ctd as f...
pd.Series(ssscc)
pandas.Series
import numpy as np import pandas as pd import logging DISTANCE_THRESHOLD = 1.4 #: max threshold for distnr SCORE_THRESHOLD = 0.4 #: max threshold for sgscore CHINR_THRESHOLD = 2 #: max threshold for chinr SHARPNR_MAX = 0.1 #: max value for sharpnr SHARPNR_MIN = -0.13 #: min value for sharpnr ZERO_MAG = 100. #: d...
pd.DataFrame(frame)
pandas.DataFrame
import pandas as pd import numpy as np import matplotlib.pyplot as plt import sys from sklearn.metrics import mean_squared_error from math import sqrt from statsmodels.tsa.api import ExponentialSmoothing, SimpleExpSmoothing, Holt # 1. 抽取2012年8月至2013年12月的数据,总共14个月 # Index 11856 marks the end of year 2013 df = pd.r...
pd.to_datetime(train.Datetime,format='%d-%m-%Y %H:%M')
pandas.to_datetime
from collections import OrderedDict import pydoc import warnings import numpy as np import pytest import pandas as pd from pandas import ( Categorical, DataFrame, DatetimeIndex, Index, Series, TimedeltaIndex, date_range, period_range, timedelta_range, ) from pandas.core.arrays impo...
tm.makeIntIndex(10)
pandas.util.testing.makeIntIndex
""" This file originated from the online analysis project at: https://github.com/OlafHaag/UCM-WebApp """ import itertools import pandas as pd import pingouin as pg import numpy as np from scipy.stats import wilcoxon from sklearn.decomposition import PCA from sklearn.covariance import EllipticEnvelope ...
pd.concat((df_stats, cov, length, df_synergies), axis='columns')
pandas.concat
from datetime import datetime, timedelta import dateutil import numpy as np import pytest import pytz from pandas._libs.tslibs.ccalendar import DAYS, MONTHS from pandas._libs.tslibs.period import IncompatibleFrequency from pandas.compat import lrange, range, zip import pandas as pd from pandas import DataFrame, Seri...
Series(1, index=expected_index)
pandas.Series
########################################################################### # Librairies import pandas as pd import os import unicodedata ########################################################################### # Fonctions def remove_accents(input_str): nfkd_form = unicodedata.normalize('NFKD', input_str) o...
pd.DataFrame(col_lab,columns=["col_lab7"])
pandas.DataFrame
# pylint: disable-msg=E1101,W0612 from datetime import datetime, timedelta import os import operator import unittest import cStringIO as StringIO import nose from numpy import nan import numpy as np import numpy.ma as ma from pandas import Index, Series, TimeSeries, DataFrame, isnull, notnull from pandas.core.index...
assert_series_equal(aa, ea)
pandas.util.testing.assert_series_equal
# -*- coding: utf-8 -*- # pylint: disable-msg=E1101,W0612 from datetime import datetime, timedelta import pytest import re from numpy import nan as NA import numpy as np from numpy.random import randint from pandas.compat import range, u import pandas.compat as compat from pandas import Index, Series, DataFrame, isn...
tm.assert_frame_equal(result, exp)
pandas.util.testing.assert_frame_equal
#definition of add_dataset that creates the meta-dataset import pandas as pd from pandas.core.dtypes.common import is_numeric_dtype from scipy.stats import pearsonr from sklearn.model_selection import train_test_split from supervised.automl import AutoML import os import pandas as pd from sklearn.preprocessing import L...
is_numeric_dtype(x)
pandas.core.dtypes.common.is_numeric_dtype
import gzip import os import sys import logging from collections import defaultdict import pandas as pd from dae.utils.regions import Region logger = logging.getLogger(__name__) # # Exon # class Exon: def __init__( self, start=None, stop=None, frame=None, number=None, ...
pd.read_csv(filename, sep="\t")
pandas.read_csv
# Copyright (c) 2018-2021, NVIDIA CORPORATION. import array as arr import datetime import io import operator import random import re import string import textwrap from copy import copy import cupy import numpy as np import pandas as pd import pyarrow as pa import pytest from numba import cuda import cudf from cudf.c...
pd.DataFrame()
pandas.DataFrame
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Quantify the variability of behavioral metrics within and between labs of mouse behavior. This script doesn't perform any analysis but plots summary statistics over labs. <NAME> 16 Jan 2020 """ import pandas as pd import matplotlib.pyplot as plt import numpy as np fr...
pd.DataFrame()
pandas.DataFrame
import itertools from typing import List, Optional, Union import numpy as np import pandas._libs.algos as libalgos import pandas._libs.reshape as libreshape from pandas._libs.sparse import IntIndex from pandas.util._decorators import cache_readonly from pandas.core.dtypes.cast import maybe_promote from pandas.core.d...
is_object_dtype(dtype)
pandas.core.dtypes.common.is_object_dtype
### Old import csv import pandas as pd import numpy as np import zipfile import os from datetime import datetime, timedelta import pickle import gzip import time import timeit def find_between( s, first, last ): try: start = s.index( first ) + len( first ) end = s.index( last, start ) retur...
pd.DataFrame(daily_data)
pandas.DataFrame
# -*- coding: utf-8 -*- # Copyright (c) 2015-2020, Exa Analytics Development Team # Distributed under the terms of the Apache License 2.0 from unittest import TestCase import h5py import numpy as np import pandas as pd from exatomic import Universe from exatomic.base import resource from exatomic.molcas.output import O...
pd.DataFrame(self.uo2sp.atom)
pandas.DataFrame
""" Module: libfmp.b.b_annotation Author: <NAME>, <NAME> License: The MIT license, https://opensource.org/licenses/MIT This file is part of the FMP Notebooks (https://www.audiolabs-erlangen.de/FMP) """ import numpy as np import pandas as pd import librosa import libfmp.b def read_csv(fn, header=True, add_label=Fal...
pd.read_csv(fn_in, sep=',', keep_default_na=False, header=None)
pandas.read_csv
import datetime as dt import matplotlib.pyplot as plt import lifetimes import numpy as np import os import pandas as pd import seaborn as sns def numcard(x): return x.nunique(), len(x) def todateclean(x): return
pd.to_datetime(x, errors='coerce')
pandas.to_datetime
""" PL Modeling Program -> Python file 1/3: InteractivePLFittingGUI.py (implements the GUI) Python file 2/3: PLModeling.py (implements the PL emission models) Python file 3/3: InterferenceFunction.py (implements the interference function models) Author: <NAME> Date: March 2021 """ import matplotlib matplotlib.use('TkA...
pd.DataFrame(output_dict)
pandas.DataFrame
#!/usr/bin python3 import pandas as pd import numpy as np import datetime as datetime def changeDate(): base_CSV =
pd.read_csv('../dataset/lpv_d.csv')
pandas.read_csv
import os import glob import datetime as dt import pandas as pd from forest import geo import bokeh.models class View(object): def __init__(self, loader): self.loader = loader self.source = bokeh.models.ColumnDataSource({ "x": [], "y": [], "date": [], ...
pd.concat(frames, ignore_index=True)
pandas.concat
import pandas as pd from sklearn.pipeline import Pipeline from sklearn.preprocessing import StandardScaler from sklearn.linear_model import SGDClassifier import argparse rate = "0.5" # 默认为6:4的正负样本比例,若要改为1:1则取rate=“0.5” class SGD: def __init__(self, trainfile, validfile, testfile): super(SGD, self).__in...
pd.read_csv(trainfile)
pandas.read_csv
# -*- coding: utf-8 -*- # pylint: disable-msg=E1101,W0612 from datetime import datetime, timedelta import pytest import re from numpy import nan as NA import numpy as np from numpy.random import randint from pandas.compat import range, u import pandas.compat as compat from pandas import Index, Series, DataFrame, isn...
tm.assert_frame_equal(result, exp)
pandas.util.testing.assert_frame_equal
import pandas as pd import os import seaborn as sns import numpy as np import matplotlib.pyplot as plt import warnings warnings.simplefilter(action='ignore', category=FutureWarning) cwd = os.getcwd() print(cwd) df =
pd.read_csv(cwd + '/' + 'HFI2021.csv')
pandas.read_csv
"""Functions for plotting system resource use.""" import logging from datetime import datetime, timedelta from pathlib import Path from typing import Optional, Union import numpy as np import pandas as pd import matplotlib.pyplot as plt import matplotlib.dates as mdates from matplotlib import gridspec async def _rea...
pd.concat([user_dataframes[a_user], missing_times_df], ignore_index=True)
pandas.concat
import pandas as pd import math import numpy as np def matchCheck(colNo, weightl, impactl): # print(colNo) # print(weightl) # print(impactl) if colNo != weightl or colNo != impactl or impactl != weightl: raise Exception("Number of weights, number of impacts and number of columns (from 2nd to ...
pd.DataFrame()
pandas.DataFrame
# -*- coding: utf-8 -*- # pylint: disable=E1101,E1103,W0232 import os import sys from datetime import datetime from distutils.version import LooseVersion import numpy as np import pandas as pd import pandas.compat as compat import pandas.core.common as com import pandas.util.testing as tm from pandas import (Categor...
pd.Categorical(idx)
pandas.Categorical
"""Covid Model""" __docformat__ = "numpy" import warnings import pandas as pd import numpy as np global_cases_time_series = ( "https://raw.githubusercontent.com/CSSEGISandData/COVID-19/master/csse_covid_19_data/csse_" "covid_19_time_series/time_series_covid19_confirmed_global.csv" ) global_deaths_time_series...
pd.read_csv(global_cases_time_series)
pandas.read_csv
from bs4 import BeautifulSoup import requests import pandas as pd import re def ensure_string_columns(df): newcols = [] for col in df.columns: strcol = str(col) if strcol[0]=="(": #if it's a tuple, instead of a string a = strcol b = a[a.find("("):a.find(",")] c = b[1:] d = c.replace("'","") strco...
pd.DataFrame(table)
pandas.DataFrame
# -*- coding: utf-8 -*- """ Created on Mon Jan 30 01:01:33 2017 @author: Flamingo """ #%% from bs4 import BeautifulSoup import urllib import pandas as pd import numpy as np CITY_NAME = pd.read_csv('CITY_NAME2.csv') PORT_NAME = CITY_NAME[['AIRPORT_CODE','PORT']].groupby('AIRPORT_CODE',as_index=False).count() for ind,v...
pd.DataFrame(Port, columns=['Port'])
pandas.DataFrame
from collections import ( abc, deque, ) from decimal import Decimal from warnings import catch_warnings import numpy as np import pytest import pandas as pd from pandas import ( DataFrame, Index, MultiIndex, PeriodIndex, Series, concat, date_range, ) import pandas._testing as tm fr...
tm.assert_frame_equal(result, expected)
pandas._testing.assert_frame_equal
import pandas as pd from argparse import ArgumentParser def save_new_labels(df_labels: pd.DataFrame, filename="labels_concatenated.csv"): df_labels.to_csv(filename, index_label="img_name") def main(args): first_labels_path = args.first_labels_path second_labels_path = args.second_labels_path df_fir...
pd.read_csv(first_labels_path, index_col="img_name")
pandas.read_csv
import pandas as pd import numpy as np import matplotlib.pyplot as plt import seaborn as sns import plotly.express as px import plotly.graph_objects as go from plotly.offline import plot,iplot from scipy.stats import norm, kurtosis import os from scipy.signal import butter, lfilter, freqz from scipy import signal from ...
pd.concat([acc_df,gyro_df],1)
pandas.concat
""" Copyright (c) 2021, Electric Power Research Institute All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: * Redistributions of source code must retain the above copyright notice, this li...
pd.DataFrame(index=new_index)
pandas.DataFrame
# -*- coding: utf-8 -*- # pylint: disable-msg=E1101,W0612 from datetime import datetime, timedelta import pytest import re from numpy import nan as NA import numpy as np from numpy.random import randint from pandas.compat import range, u import pandas.compat as compat from pandas import Index, Series, DataFrame, isn...
u(' bb')
pandas.compat.u
""" Usage: aggregate-makespan.py -i FOLDER [--output FOLDER] [--start-run INT] [--end-run INT] Required Options: -i FOLDER --input FOLDER where the experiments are Options: -o FOLDER --output FOLDER where the output should go [default: input] --start-run INT ...
pd.DataFrame()
pandas.DataFrame
import pandas as p#导入目前所需要的库并给与简称 data_train = '../homework/train.csv' #查看基本数据 data_train = p.read_csv(data_train)#导入训练模型 print(data_train.info())#查看数据类型 print(data_train.describe())#粗略查看基本数据 ###导入并且查看原始数据 import matplotlib.pyplot as pt import numpy as n pt.rcParams['font.sans-serif']=['Simhei'] #解决...
p.get_dummies(data_test[['Cabin','Sex','Embarked','Pclass']])
pandas.get_dummies
"""Miscellaneous internal PyJanitor helper functions.""" import fnmatch import functools import os import re import socket import sys import warnings from collections.abc import Callable as dispatch_callable from itertools import chain, combinations from typing import ( Callable, Dict, Iterable, List, ...
pd.MultiIndex.from_arrays([mapping, outcome])
pandas.MultiIndex.from_arrays
""" Module parse to/from Excel """ # --------------------------------------------------------------------- # ExcelFile class import abc from datetime import date, datetime, time, timedelta from distutils.version import LooseVersion from io import UnsupportedOperation import os from textwrap import fill import warnings...
is_integer(sheet_name)
pandas.core.dtypes.common.is_integer
import time import os import shutil import sys from sys import argv import pickle import csv from collections import defaultdict import cProfile import pstats import pandas as pd import numpy as np from sklearn.metrics import roc_auc_score from sklearn.pipeline import FeatureUnion, Pipeline from sklearn.preprocessin...
pd.DataFrame(dt_train_encoded)
pandas.DataFrame
import timeit import pandas as pd import numpy as np from typing import Dict,List loops=1000 inputfile:List[List[int]] = [[1,2,3,4,5,6] for x in range(0,1000)] # input arrives as a list of row lists # need to make columns ####################### # zip # 60us def i1() -> List[int]: return list(map(list, zip(*inpu...
pd.DataFrame(y2)
pandas.DataFrame
import logging import math import os import sys import geopandas as gpd import numpy as np import pandas as pd from _helpers import _sets_path_to_root from _helpers import _to_csv_nafix from _helpers import configure_logging # from shapely.geometry import LineString, Point, Polygon # from osm_data_config import AFRIC...
pd.concat([df_lines, df_cables])
pandas.concat
#!/usr/bin/env python3 # # Copyright, <NAME> 2020 # # MIT License # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, ...
pd.Series([truth_path, pred_path] + pc_iou, index=iou_df.columns)
pandas.Series
''' Author: <NAME> Utilities to summarize the outputs produced by the model i.e. the results.txt files spitted out by the evaluation scripts. ''' import os from sys import path import re import pandas as pd import numpy as np from scipy.stats import ttest_rel # res will need to be passed to the last function, as an...
pd.Index(index_list_verbose, name='labels')
pandas.Index
from functools import lru_cache from pyiso import client_factory from datetime import datetime, timedelta from funcy import compose, identity, retry from itertools import repeat from urllib.error import HTTPError import pandas as pd import numpy as np from app.model import RENEWABLES, NON_RENEWABLES from app.util impo...
pd.Timedelta("1s")
pandas.Timedelta
''' File: HAPT_Dataset.py Author: <NAME> Date: 03/10/2019 Version: 1.0 Description: utility functions to load the Human Activities and Postural Transitions (HAPT) dataset ''' import numpy as np import pandas as pd from os.path import expanduser from keras.utils import to_categorical from multiprocessing import P...
pd.concat([filtered_df,data_uuid], ignore_index=True)
pandas.concat
import os import pandas import numpy from ... import normalize from ... import convert def hellinger(aves, stds): """ Computes pairwise Hellinger distances between Gaussian distributions from lists of the means and standard deviations. Args: aves (numpy array): list of means (length n) ...
pandas.DataFrame()
pandas.DataFrame
"""Tradingview model""" __docformat__ = "numpy" import requests from tradingview_ta import TA_Handler import pandas as pd from gamestonk_terminal import config_terminal as cfg INTERVALS = { "1m": "1 min", "5m": "5 min", "15m": "15 min", "1h": "1 hour", "4h": "4 hours", "1d": "1 day", "1W"...
pd.DataFrame()
pandas.DataFrame
import datetime import logging import json import requests from pandas import json_normalize import pandas as pd from google.cloud import storage from anyway.parsers.waze.waze_db_functions import ( insert_waze_alerts, insert_waze_traffic_jams, enrich_waze_alerts_ended_at_timestamp, enrich_waze_traffic_...
json_normalize(waze_jams)
pandas.json_normalize
from __future__ import division, print_function, absolute_import import numpy as np import matplotlib.pyplot as plt import streamlit as st import pandas as pd import seaborn as sns import random from sklearn.model_selection import RepeatedKFold, train_test_split, cross_val_score, StratifiedKFold, RepeatedStratifiedKFo...
pd.DataFrame()
pandas.DataFrame
#!/usr/bin/env python """ coding=utf-8 Code Template """ from emblaze import app import logging import os import pandas import textract from emblaze.ResumeParser.bin import lib from emblaze.ResumeParser.bin import field_extraction import spacy def main(): """ Main function documentation template :retu...
pandas.DataFrame(data=candidate_file_agg, columns=['file_path'])
pandas.DataFrame
import scipy.signal import pandas as pd import numpy as np import peakutils from lmfit import models import chachifuncs as ccf import os import glob ################################ # OVERALL Wrapper Function ################################ def ML_generate(import_filepath): """Generates a dataframe containing c...
pd.concat([desc, desc_df], ignore_index=True)
pandas.concat
import json from datetime import datetime import pandas as pd def get_forecast(): first_row = True nrow = 0 with open('data/tmp/forecast_weather.json') as f: for line in f: print(nrow) nrow += 1 res = json.loads(line) timezone_offset = res['city']['timezone'] first_line = Tru...
pd.DataFrame.from_dict(data)
pandas.DataFrame.from_dict
# Input arguments flag import sys sys.path.append('..') _, *flag = sys.argv # Parse arguments import argparse parser = argparse.ArgumentParser(prog='hs_check', description='Check phase synchronization for selected BPMs and plane.') parser.add_argument('-p', '--plane', choices=('x', 'y'), help='data plane', default='x'...
pandas.DataFrame()
pandas.DataFrame
"""Utilities for parsing corpus tsv files into pandas DataFrames.""" import logging from glob import glob from pathlib import Path from typing import Dict, Iterable, Union import pandas as pd from tqdm import tqdm import harmonic_inference.utils.corpus_utils as cu from harmonic_inference.data.corpus_constants import ...
pd.DataFrame(files_dict)
pandas.DataFrame
import pysam import pandas as pd import numpy as np import re import os import sys import collections import scipy from scipy import stats import statsmodels from statsmodels.stats.multitest import fdrcorrection try: from . import global_para except ImportError: import global_para try: from .consensus_seq ...
pd.concat(list_df_transcript_merge)
pandas.concat