prompt
stringlengths
19
1.03M
completion
stringlengths
4
2.12k
api
stringlengths
8
90
# # Copyright (c) 2021 salesforce.com, inc. # All rights reserved. # SPDX-License-Identifier: BSD-3-Clause # For full license text, see the LICENSE file in the repo root or https://opensource.org/licenses/BSD-3-Clause # import logging import os import requests from tqdm import tqdm import pandas as pd from ts_datasets...
pd.DataFrame({"trainval": sequence.index < sequence.index[ntrain]}, index=sequence.index)
pandas.DataFrame
""" Fractional differentiation is a technique to make a time series stationary but also retain as much memory as possible. This is done by differencing by a positive real number. Fractionally differenced series can be used as a feature in machine learning process. """ import numpy as np import pandas as pd class Fr...
pd.Series(index=series.index)
pandas.Series
import numpy import matplotlib.pyplot as plt import tellurium as te from rrplugins import Plugin auto = Plugin("tel_auto2000") from te_bifurcation import model2te, run_bf import pandas as pd import numpy as np import matplotlib.pyplot as plt from matplotlib.ticker import ScalarFormatter sf = ScalarFormatter() sf.set_sc...
pd.DataFrame(binned_Rts)
pandas.DataFrame
# coding: utf-8 # --- # # _You are currently looking at **version 1.2** of this notebook. To download notebooks and datafiles, as well as get help on Jupyter notebooks in the Coursera platform, visit the [Jupyter Notebook FAQ](https://www.coursera.org/learn/python-social-network-analysis/resources/yPcBs) course reso...
pd.isnull(df['Future Connection'])
pandas.isnull
""" XGBoost regressor for construction machine price prediction """ import pandas as pd import matplotlib.pyplot as plt import seaborn as sns import xgboost as xgb from sklearn.model_selection import train_test_split from pandas_profiling import ProfileReport pd.set_option('display.max_rows', 500) plt.style.use('ggplo...
pd.api.types.is_string_dtype(content)
pandas.api.types.is_string_dtype
import requests import pandas as pd import numpy as np import pathlib import zipline as zl import logbook import datetime import os from azul import price_manager_registry, BasePriceManager log = logbook.Logger('PolygonPriceManager') @price_manager_registry.register('polygon') class PolygonPriceManager(BasePriceMana...
pd.DataFrame()
pandas.DataFrame
# -------------- import pandas as pd from sklearn import preprocessing #path : File path # Code starts here dataset =
pd.read_csv(path)
pandas.read_csv
#====================================================== # Model Utility Functions #====================================================== ''' Info: Utility functions for model building. Version: 2.0 Author: <NAME> Created: Saturday, 13 April 2019 ''' # Import modules import os import uuid import copy im...
pd.concat(dfs, axis=0)
pandas.concat
# -*- coding: utf-8 -*- """ Created on Wed May 30 14:47:20 2018 @author: Greydon """ import os import re import numpy as np import pandas as pd from scipy.signal import welch, hanning, butter, lfilter, resample import matplotlib.pyplot as plt from matplotlib.ticker import FormatStrFormatter import matplotlib.ticker as...
pd.DataFrame([])
pandas.DataFrame
import mongomanager import intriniowrapper import logging import inspect import copy import pandas as pd import commonqueries from datetime import datetime from dateutil.relativedelta import relativedelta import os import configwrapper class IntrinioUpdater(): def __init__(self,config_file,proxies=None,timeout=300,ma...
pd.DataFrame()
pandas.DataFrame
import unittest import numpy as np import pandas as pd import scipy.stats as st from ..analysis import determine_analysis_type from ..analysis.exc import NoDataError from ..data import Vector, Categorical class MyTestCase(unittest.TestCase): def test_small_float_array(self): np.random.seed(123456789) ...
pd.Series(['a'])
pandas.Series
# GetData.py # <NAME> # 29 November 2019 # # This program takes as input an output filename and a starting date and # returns song/artist names to the specified file. import csv import sys import time import pandas as pd from selenium import webdriver from datetime import datetime, timedelta # get the list of songs/...
pd.read_csv(fname)
pandas.read_csv
#!/usr/bin/env python """Script for generating figures of catalog statistics. Run `QCreport.py -h` for command line usage. """ import os import sys import errno import argparse from datetime import date, datetime from math import sqrt, radians, cos import markdown import numpy as np import pandas as pd import cartopy....
pd.Timedelta(seconds=16)
pandas.Timedelta
import os try: import fool except: print("缺少fool工具") import math import pandas as pd import numpy as np import random import tensorflow as tf import re np.random.seed(1) def add2vocab(path,word): vocab_data=pd.read_csv(path) idx_to_chars=list(vocab_data['vocabulary'])+[word] df_data = pd...
pd.DataFrame(datas)
pandas.DataFrame
""" Функции и классы для проведения WoE-преобразований """ import math import warnings import numpy as np import pandas as pd import sklearn as sk from IPython.display import display from matplotlib import pyplot as plt from sklearn.base import BaseEstimator, TransformerMixin from sklearn.model_selection import trai...
pd.concat([X_woe_num, X_woe_cat, X_woe_oth])
pandas.concat
import unittest import pandas as pd from pandas.core.dtypes.common import is_numeric_dtype, is_string_dtype from pandas.util.testing import assert_frame_equal from shift_detector.precalculations.store import InsufficientDataError, Store from shift_detector.utils.column_management import ColumnType class TestStore(u...
pd.DataFrame.from_dict(data)
pandas.DataFrame.from_dict
import numpy as np import matplotlib.pyplot as plt import pandas as pd import pickle met_df = pd.read_csv('../datasets/cleaned_dataset.csv', index_col=0) from sklearn.preprocessing import OneHotEncoder, LabelEncoder le = LabelEncoder() ohe = OneHotEncoder(sparse=False) val = le.fit_transform(met_df['rest_type']).res...
pd.concat([node,res], axis=1)
pandas.concat
import builtins from io import StringIO from itertools import product from string import ascii_lowercase import numpy as np import pytest from pandas.errors import UnsupportedFunctionCall import pandas as pd from pandas import ( DataFrame, Index, MultiIndex, Series, Timestamp, date_range, isna) import pandas.cor...
tm.assert_series_equal(result, expected)
pandas.util.testing.assert_series_equal
# -*- coding: utf-8 -*- """Interface for flopy's implementation for MODFLOW.""" __all__ = ["MfSfrNetwork"] import pickle from itertools import combinations, zip_longest from textwrap import dedent import geopandas import numpy as np import pandas as pd from shapely import wkt from shapely.geometry import LineString,...
pd.Series(dtype=int)
pandas.Series
import numpy as np import random import pandas as pd import korbinian import sys ##########parameters############# seq_len = 10000 number_seq = 50 number_mutations = 2000 subset_num = 12 ident = 100 * (seq_len - number_mutations) / seq_len List_rand_TM = r"D:\Databases\summaries\01\List01_rand\List01_rand_TM.csv" #L...
pd.Series.from_csv(List_rand_TM, sep="\t")
pandas.Series.from_csv
"""Legacy feature computation from depart.""" import itertools import re import numpy as np import pandas as pd from Bio.SeqUtils.ProtParam import ProteinAnalysis from Bio.SeqUtils.ProtParamData import kd from pyteomics import parser from sklearn.preprocessing import PolynomialFeatures from xirt import sequences d...
pd.DataFrame()
pandas.DataFrame
import unittest from yauber_algo.errors import * class IIFTestCase(unittest.TestCase): def test_iif(self): import yauber_algo.sanitychecks as sc from numpy import array, nan, inf import os import sys import pandas as pd import numpy as np from yauber_algo....
pd.Series(arr_false)
pandas.Series
import pandas as pd import matplotlib.pyplot as plt from matplotlib import cm # # Read in data frame # df =
pd.read_csv("mcs_interestrate_change.csv", skiprows=1)
pandas.read_csv
import argparse import os import matplotlib.pyplot as plt import numpy as np import pandas as pd import seaborn as sns parser = argparse.ArgumentParser(description="TF_diversity_rw") parser.add_argument( "file_names", type=str, help="Name of folder and filenames for the promoters extracted", ) parser.add_...
pd.merge(select_genes, df, on="AGI", how="left")
pandas.merge
"""Script of my solution to DrivenData Modeling Women's Health Care Decisions Use this script in the following way: python solution.py <name-of-submission> Argument is optional, the script will assign default name. """ from __future__ import division import sys import pdb import numpy as np import pandas as pd fr...
pd.read_csv('data/processed_train.csv')
pandas.read_csv
# -*- coding: utf-8 -*- # pylint: disable=E1101 # flake8: noqa from datetime import datetime import csv import os import sys import re import nose import platform from multiprocessing.pool import ThreadPool from numpy import nan import numpy as np from pandas.io.common import DtypeWarning from pandas import DataFr...
StringIO(data)
pandas.compat.StringIO
# -*- coding: utf-8 -*- """ Created on Tue Dec 17 17:28:40 2019 @author: Administrator """ import pdblp import pandas as pd import numpy as np import matplotlib.pyplot as plt import seaborn as sns con = pdblp.BCon(debug=False, port=8194, timeout=5000) con.start() from datetime import date start = '...
pd.concat( [swap_spread_de, swap_spread_uk, swap_spread_jp, swap_spread_ch, swap_spread_tr, swap_spread_br, swap_spread_ru, swap_spread_sa ],axis=1)
pandas.concat
import pandas as pd def full_describe(series: pd.Series, verbose=True): """ Calculates a pandas describe of series, plus a count of unique and NaN :param verbose: printing some other info :param series: Pandas Series :return: df with stats as cols """ stats_df =
pd.DataFrame()
pandas.DataFrame
# Pre-Process Text Data # Import Modules import os import pandas as pd import numpy as np import csv import matplotlib.pyplot as plt import nltk import string import re import datetime from tqdm import tqdm from textblob import TextBlob from nltk.corpus import wordnet as wn from nltk.stem.wordnet import WordNetLemmati...
pd.concat([posts,toplevelcomments,subcomments])
pandas.concat
""" Test output formatting for Series/DataFrame, including to_string & reprs """ from datetime import datetime from io import StringIO import itertools from operator import methodcaller import os from pathlib import Path import re from shutil import get_terminal_size import sys import textwrap import dateutil import ...
option_context("display.precision", 5)
pandas.option_context
import pytrec_eval from repro_eval.Evaluator import RplEvaluator from repro_eval.util import trim import pandas as pd from matplotlib import pyplot as plt import seaborn as sns sns.set() sns.set_style('whitegrid') palette = sns.color_palette("GnBu_d") sns.set_palette(palette) colors = sns.color_palette() ORIG_B = './d...
pd.DataFrame(df_content, index=['tf_1', 'tf_2', 'tf_3', 'tf_4', 'tf_5'])
pandas.DataFrame
import os, datetime, pymongo, configparser import pandas as pd from bson import json_util global_config = None global_client = None global_stocklist = None def getConfig(root_path): global global_config if global_config is None: #print("initial Config...") global_config = configparser.ConfigPa...
pd.DataFrame()
pandas.DataFrame
# -*- coding: utf-8 -*- """ Created on Sun May 3 18:37:20 2020 @author: Blackr The following is a script I wrote to automatically find the tafel data a .csv file with the appropriate I, Ewe, and time columns time = s I = mA E = V This is mostly used as a coding excerise for post processing. Re-evaluate ...
pd.Series(unique)
pandas.Series
# pylint: disable=redefined-outer-name import filecmp from io import StringIO from pathlib import Path from tempfile import TemporaryDirectory import pandas as pd import pytest from courier.config import get_config from courier.elements import CourierIssue, IssueStatistics, export_articles from courier.elements.expor...
pd.DataFrame(stats)
pandas.DataFrame
""" To test the quality of the estimators, we generate data both from a semilinear Choo and Siow model and from a semilinear nested logit model. We use both the Poisson estimator and the minimum-distance estimator on the former model, and only the minimum-distance estimator on the latter. """ from typing import List...
pd.read_csv(choo_siow_results_file)
pandas.read_csv
#! /usr/bin/env python # -*- coding: utf-8 -*- # vim:fenc=utf-8 # # Copyright © 2020 qizai <<EMAIL>> # # Distributed under terms of the MIT license. """ This script will search for all lp{0, 1, ..., N}_{right, left, middle}_100_region.csv file. Each line in the csv file is a complex, with at least two fragment interse...
pd.DataFrame()
pandas.DataFrame
import io import textwrap from collections import namedtuple import numpy as np import pandas as pd import statsmodels.api as sm from estimagic.config import EXAMPLE_DIR from estimagic.visualization.estimation_table import _convert_model_to_series from estimagic.visualization.estimation_table import _create_statistics...
ase(exp, res)
pandas.testing.assert_series_equal
# Streamlit live coding script import streamlit as st import pandas as pd import matplotlib.pyplot as plt import numpy as np from PIL import Image import plotly.express as px import plotly.graph_objects as go df = pd.read_csv('src/data/marketing_campaign_cleaned.csv', index_col=[0]) st.title("Customer personality ana...
pd.DataFrame(d)
pandas.DataFrame
import os import pandas as pd import datetime import numpy as np from talib import abstract from .crawler import check_monthly_revenue class Data(): def __init__(self): self.date = datetime.datetime.now().date() self.warrning = False self.col2table = {} t...
pd.to_numeric(s, errors='coerce')
pandas.to_numeric
from bs4 import BeautifulSoup as BS from selenium import webdriver from functools import reduce import pandas as pd import time import xport import pandas as pd def render_page(url): driver = webdriver.Chrome('/Users/cp/Downloads/chromedriver') driver.get(url) time.sleep(3) r = driver.p...
pd.merge(left, right, left_index=True, right_index=True)
pandas.merge
import pytest from datetime import datetime, timedelta import pytz import numpy as np from pandas import (NaT, Index, Timestamp, Timedelta, Period, DatetimeIndex, PeriodIndex, TimedeltaIndex, Series, isna) from pandas.util import testing as tm from pandas._libs.tslib import iNa...
Timestamp('NaT')
pandas.Timestamp
""" Testing that functions from rpy work as expected """ import pandas as pd import numpy as np import unittest import nose import pandas.util.testing as tm try: import pandas.rpy.common as com from rpy2.robjects import r import rpy2.robjects as robj except ImportError: raise nose.SkipTest('R not inst...
com.load_data(name)
pandas.rpy.common.load_data
from model.toolkits.parse_conf import parse_config_vina, parse_protein_vina, parse_ligand_vina import os import pandas as pd import numpy as np from pathlib import Path import argparse import rdkit from rdkit import Chem, DataStructs from rdkit.Chem import Descriptors, rdMolDescriptors, AllChem, QED try: from openb...
pd.read_csv(args.smi)
pandas.read_csv
import numpy as np from datetime import timedelta from distutils.version import LooseVersion import pandas as pd import pandas.util.testing as tm from pandas import to_timedelta from pandas.util.testing import assert_series_equal, assert_frame_equal from pandas import (Series, Timedelta, DataFrame, Timestamp, Timedelt...
Timestamp('20130102')
pandas.Timestamp
# -*- coding: utf-8 -*- # pylint: disable=E1101,E1103,W0232 import os import sys from datetime import datetime from distutils.version import LooseVersion import numpy as np import pandas as pd import pandas.compat as compat import pandas.core.common as com import pandas.util.testing as tm from pandas import (Categor...
Categorical([], ["a", "b", "c"])
pandas.Categorical
""" test the scalar Timedelta """ import numpy as np from datetime import timedelta import pandas as pd import pandas.util.testing as tm from pandas.tseries.timedeltas import _coerce_scalar_to_timedelta_type as ct from pandas import (Timedelta, TimedeltaIndex, timedelta_range, Series, to_timedelta,...
Timedelta(10.0, unit='d')
pandas.Timedelta
#coding=utf-8 #键盘分析 #(1)分别读取csdn和yahoo数据库中的passwd #(2)自定义了常见的14种键盘密码字符串 #(3)将从数据库中读取的passwd与定义的字符串进行子串匹配(忽略单个的字母和数字) #(4)只选择相对高频的密码,生成保存频率最高的密码和对应频率的csv import pandas as pd import numpy as np import csv np.set_printoptions(suppress=True) ############################################## #(1)读取数据 #########################...
pd.Series(yahoo_data['passwd'].values)
pandas.Series
"""Functions for transofrmation of films and books datasets. Functions --------- get_books_ratings - transform books dataset get_films_ratings - transform films dataset generate_datasets - generate films and books datasets """ from typing import Set import pandas as pd from pathlib im...
pd.read_csv(location_2, sep='\t', low_memory=False)
pandas.read_csv
import empyrical import pandas as pd def main(payload): port_vals_df = _convert_port_vals_to_df(payload["portVals"]) # Calculates per data point returns port_vals_returns = port_vals_df["value"].pct_change() cum_returns = empyrical.cum_returns(port_vals_returns, starting_value=0) # aggregate_retu...
pd.DataFrame.from_dict(port_vals, orient="columns")
pandas.DataFrame.from_dict
import pandas as pd from traja.dataset import dataset def test_category_wise_sampling_few_categories(): data = list() num_categories = 5 for category in range(num_categories): for sequence in range(40 + int(category / 14)): data.append([sequence, sequence, category]) df = pd.Dat...
pd.DataFrame(data, columns=['x', 'y', 'ID'])
pandas.DataFrame
import pandas as pd from bs4 import BeautifulSoup import requests import re from tqdm import tqdm def get_fund_holding(symbol): url = 'http://finance.sina.com.cn/fund/quotes/{}/bc.shtml'.format(symbol) html = requests.get(url) bs = BeautifulSoup(html.content, features="lxml") tbl = bs.find('tabl...
pd.concat(res, ignore_index=True)
pandas.concat
import pandas as pd from Datasets.utils import read_parquet, get_bib_info, clean import json import sys import time import csv import random from tqdm import tqdm sys.path.append("Models/") from Models import * from position_rank import get_weights def mask(text1, text2): """ a simple vectorization function ...
pd.DataFrame(all_scores_adjust)
pandas.DataFrame
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Computes broadband power, offset and slope of power spectrum Based on selected epochs (e.g. ASCIIS) in the list of files a power spectrum is computed. Based on this power spectrum the broadband power is calculated, followed by the offset and slope using the FOOOF algo...
pd.DataFrame(mean_pxx[index,:,:])
pandas.DataFrame
# 均分321个区间,统计落入格区间的数目 import pandas as pd import numpy as np df =
pd.read_csv('submit-final.csv', index_col=0)
pandas.read_csv
# Copyright (c) Microsoft Corporation. # Licensed under the MIT License. """ This file contains training and testing settings to be used in this benchmark, mainly: TRAIN_BASE_END: Base training end date common across all rounds TRAIN_ROUNDS_ENDS: a set of dates denoting end of training period for each ...
pd.to_datetime(("2017-04-01", "2017-05-01"))
pandas.to_datetime
# -*- coding: utf-8 -*- """ This script can be used to convert a .mat simulation result file into a .csv file with a subset of model variables, as defined in the "outputs" list. The .csv file is saved in the same directory as the .mat file, and is based on the case, climate, and scenario. """ from buildingspy.io.out...
pd.DataFrame(index=time,data=values,columns=[variable])
pandas.DataFrame
""" Module contains tools for processing Stata files into DataFrames The StataReader below was originally written by <NAME> as part of PyDTA. It has been extended and improved by <NAME> from the Statsmodels project who also developed the StataWriter and was finally added to pandas in a once again improved version. Yo...
DataFrame.from_records(data)
pandas.core.frame.DataFrame.from_records
# pylint: disable=E1101,E1103,W0232 from datetime import datetime, timedelta from pandas.compat import range, lrange, lzip, u, zip import operator import re import nose import warnings import os import numpy as np from numpy.testing import assert_array_equal from pandas import period_range, date_range from pandas.c...
Float64Index([1.0, np.nan])
pandas.core.index.Float64Index
import inspect import json import os import re from urllib.parse import quote from urllib.request import urlopen import pandas as pd import param from .configuration import DEFAULTS class TutorialData(param.Parameterized): label = param.String(allow_None=True) raw = param.Boolean() verbose = param.Bool...
pd.read_csv(self._data_url, **base_kwds)
pandas.read_csv
import ipyleaflet import ipywidgets import pandas as pd import geopandas as gpd from shapely.geometry import Polygon, Point import datetime import requests import xml.etree.ElementTree as ET import calendar import numpy as np import pathlib import os class ANA_interactive_map: def __init__(self, path_inventario)...
pd.date_range(start='2000-01-01',end='2020-01-01', freq='M')
pandas.date_range
import warnings warnings.simplefilter(action = 'ignore', category = UserWarning) # Front matter import os import glob import re import pandas as pd import numpy as np import scipy.constants as constants # Find the filepath of all .res NRIXS files resfilepath_list = [filepath for filepath in glob.glob('*/*.res')] # C...
pd.DataFrame({'Date': [month+' '+year], 'Folder': [folder], 'Index': [filename]})
pandas.DataFrame
def test_get_number_rows_cols_for_fig(): from mspypeline.helpers import get_number_rows_cols_for_fig assert get_number_rows_cols_for_fig([1, 1, 1, 1]) == (2, 2) assert get_number_rows_cols_for_fig(4) == (2, 2) def test_fill_dict(): from mspypeline.helpers import fill_dict def test_default_to_regular...
pd.DataFrame()
pandas.DataFrame
# This script helps to create TAble 1 (phenotypes per country) import pandas as pd from scipy.stats import chi2_contingency import matplotlib.pyplot as plt import numpy as np # Include all GENES, those containing Indels and SNVS (that's why I repeat this step of loading "alleles" dataframe) This prevents badly grouppi...
pd.read_csv('/path/to/phenotypes_20210107.csv',sep='\t')
pandas.read_csv
#%% import os try: os.chdir('/Volumes/GoogleDrive/My Drive/python_code/connectome_tools/') print(os.getcwd()) except: pass #%% import sys sys.path.append("/Volumes/GoogleDrive/My Drive/python_code/connectome_tools/") import pandas as pd import numpy as np import connectome_tools.process_matrix as promat...
pd.to_numeric(matrix_ad.columns)
pandas.to_numeric
""" Classes for comparing outputs of two RSMTool experiments. :author: <NAME> (<EMAIL>) :author: <NAME> (<EMAIL>) :author: <NAME> (<EMAIL>) :organization: ETS """ import warnings from collections import defaultdict from copy import deepcopy from os.path import exists, join import numpy as np import pandas as pd fro...
pd.concat(correlation_list, sort=True)
pandas.concat
from datetime import ( datetime, timedelta, timezone, ) import numpy as np import pytest import pytz from pandas import ( Categorical, DataFrame, DatetimeIndex, NaT, Period, Series, Timedelta, Timestamp, date_range, isna, ) import pandas._testing as tm class TestS...
Timestamp("2012-11-11 00:00:00+01:00")
pandas.Timestamp
from email.mime.multipart import MIMEMultipart from email.mime.text import MIMEText import smtplib import os from os.path import join import json import pandas as pd import subprocess USER_FIELDS = [ "created_at", #"description", #"entities", "id", "location", "name", #"pinned_tweet_id", ...
pd.concat([hour_tweets, tmp])
pandas.concat
### Filename = test_sfw_categorical_autotest_aggregated_functions.py ## ## # # # THIS TEST WAS AUTOGENERATED BY generator_categorical_unit_test.py # # ...
pd.DataFrame(test_class.data)
pandas.DataFrame
import pandas as pd import numpy as np import matplotlib.pyplot as plt import argparse from scipy.signal import butter, lfilter from scipy.signal import freqs def exponential_smooth(data, smooth_fac): """ :param data(np.array) :param smooth_fac(int): span_interval :return: """ ser = pd.Series(...
pd.read_csv(args.file[0])
pandas.read_csv
# coding=utf-8 # pylint: disable-msg=E1101,W0612 import numpy as np import pytest from pandas.compat import lrange, range import pandas as pd from pandas import DataFrame, Index, Series import pandas.util.testing as tm from pandas.util.testing import assert_series_equal def test_get(): # GH 6383 s = Series...
Series([2, np.nan], index=idx)
pandas.Series
import pandas as pd import numpy as np from datetime import datetime def transformData(RideWaits): RideWaits["RideId"] = pd.Categorical(RideWaits["RideId"]) #RideWaits["Status"] = pd.Categorical(RideWaits["Status"]) RideWaits["ParkId"] = pd.Categorical(RideWaits["ParkId"]) RideWaits["Tier"] = pd.Catego...
pd.Categorical(RideWaits["MagicHourType"])
pandas.Categorical
import json from types import SimpleNamespace import pandas as pd from hana_ml.dataframe import ConnectionContext from hana_ml.model_storage import ModelStorage from typing import List from hana_automl.algorithms.base_algo import BaseAlgorithm from hana_automl.algorithms.ensembles.blendcls import BlendingCls from han...
pd.DataFrame(res, columns=col_names)
pandas.DataFrame
import pandas as pd import numpy as np import psycopg2 from sklearn.model_selection import KFold import Constants import sys from pathlib import Path output_folder = Path(sys.argv[1]) output_folder.mkdir(parents=True, exist_ok=True) # update database credentials if MIMIC data stored in postgres database conn = psycop...
pd.merge(left=df, right=icds, on='hadm_id')
pandas.merge
# -*- coding: utf-8 -*- # Copyright (c) 2018-2021, earthobservations developers. # Distributed under the MIT License. See LICENSE for more info. import numpy as np import pandas as pd import pytest from pandas._testing import assert_frame_equal from wetterdienst import Settings from wetterdienst.exceptions import Inva...
assert_frame_equal(df, expected_df, check_categorical=False)
pandas._testing.assert_frame_equal
import unittest from enda.timeseries import TimeSeries import pandas as pd import pytz class TestTimeSeries(unittest.TestCase): def test_collapse_dt_series_into_periods(self): # periods is a list of (start, end) pairs. periods = [ (pd.to_datetime('2018-01-01 00:15:00+01:00'), pd.to_d...
pd.to_datetime('2018-01-01 00:00:00+01:00')
pandas.to_datetime
import re import numpy as np import pytest from pandas.core.dtypes.common import pandas_dtype from pandas import ( Float64Index, Index, Int64Index, ) import pandas._testing as tm class TestAstype: def test_astype_float64_to_object(self): float_index = Float64Index([0.0, 2.5...
Float64Index([1, 2, non_finite])
pandas.Float64Index
import csv import json import os import re from collections import OrderedDict from io import StringIO import pandas as pd import requests from django.core.management.base import BaseCommand from django.forms.models import model_to_dict from va_explorer.va_data_management.models import CauseCodingIssue from va_explor...
pd.DataFrame.from_records(va_data)
pandas.DataFrame.from_records
import pandas as pd import numpy as np from sklearn.model_selection import train_test_split from sklearn.model_selection import StratifiedKFold, KFold def cv_index(n_fold, feature, label): skf = KFold(n_fold, shuffle=True, random_state=7840) index_list = [] for i, j in skf.split(feature, label): i...
pd.get_dummies(for_dummy, prefix=col)
pandas.get_dummies
import pandas as pd import numpy as np import scipy.sparse as spl from concurrent.futures import ProcessPoolExecutor import sys threads = 4 all_tasks = [ [5, 8000, ['5t', '5nt'], 0.352], [10, 12000, ['10t', '10nt'], 0.38], [25, 40000, ['25f'], 0.43386578246281293], [25, 9000, ['25r'], 0.4], [100, 4...
pd.read_csv('data/challenge_set/playlists.csv')
pandas.read_csv
import nltk nltk.download('punkt') from newspaper import Article, Config from pygooglenews import GoogleNews import requests import pandas as pd from bs4 import BeautifulSoup import re from urllib.parse import urlparse from pathlib import Path class Newspaper_agent: TMP_DIRECTORY = Path("./tmp_data") if not T...
pd.read_csv("data_news/Custom_Websites_Tags.csv")
pandas.read_csv
""" Routines for casting. """ from contextlib import suppress from datetime import date, datetime, timedelta from typing import ( TYPE_CHECKING, Any, Dict, List, Optional, Sequence, Set, Sized, Tuple, Type, Union, ) import numpy as np from pandas._libs import lib, tslib, t...
is_timedelta64_ns_dtype(arr.dtype)
pandas.core.dtypes.common.is_timedelta64_ns_dtype
import matplotlib.pyplot as plt from sklearn.linear_model import LinearRegression from sklearn.preprocessing import PolynomialFeatures import numpy as np from pylab import rcParams ########################################################################################## # Designed and developed by <NAME> # Date : 11 ...
pd.DataFrame()
pandas.DataFrame
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Tue May 26 17:19:41 2020 @author: <NAME> """ import pandas as pd def int_br(x): return int(x.replace('.','')) def float_br(x): return float(x.replace('.', '').replace(',','.')) dia = '2805' file_HU = '~/ownCloud/sesab/exporta_bole...
pd.DataFrame(columns=colsutils)
pandas.DataFrame
import pandas as pd import numpy as np from multiprocessing import cpu_count from functools import partial from scipy.optimize import minimize from trading.accountcurve import accountCurve from core.utility import draw_sample, weight_forecast from multiprocessing_on_dill import Pool from contextlib import closing """...
pd.DataFrame()
pandas.DataFrame
import numpy as np import pandas as pd def get_market_list(client, *args): marketList = pd.DataFrame(client.get_products()['data']) if len(args)>0: quoteBase = args[0] marketList = marketList[marketList['quoteAsset']==quoteBase] marketList['volume_24h'] = marketList['tradedMoney'] marke...
pd.to_numeric(klines[10])
pandas.to_numeric
import pymorphy2 import re import string import os import time import collections as cl import numpy as np import matplotlib as mpl import matplotlib.pyplot as plt import seaborn as sns import pandas as pd from collections import defaultdict from matplotlib.backends.backend_pdf import PdfPages def numnum(y): retu...
pd.DataFrame(Ue, columns=['кол', 'п', 'чр'])
pandas.DataFrame
# # Copyright (c) 2018 <NAME> <<EMAIL>> # # See the file LICENSE for your rights. # """ Methods for processing VERIFICATION data. """ import os import re import numpy as np import pandas as pd from datetime import datetime, timedelta import pickle import requests from collections import OrderedDict ...
pd.Timestamp(date.year,date.month,date.day,hour2)
pandas.Timestamp
# Core functions # # this file contains reusable core functions like filtering on university # and adding year and month name info # these are functions which are generally used in every product # roadmap: I want to push all functions from loose function # to functions combined in classgroups from nlp_functions impo...
pd.read_csv(path_deals)
pandas.read_csv
#!/usr/bin/env python # -*- coding: utf-8 -*- # ======================================================================== # the metircs for docking power(AUC, Rp, success rate) just for ML-based models # ======================================================================== from warnings import simplefilter simplefi...
pd.concat([df, df_ref.loc[df.index][['rmsd']]], axis=1)
pandas.concat
# Copyright (c) 2021-2022, NVIDIA CORPORATION. import numpy as np import pandas as pd import pytest import cudf from cudf.testing._utils import NUMERIC_TYPES, assert_eq from cudf.utils.dtypes import np_dtypes_to_pandas_dtypes def test_can_cast_safely_same_kind(): # 'i' -> 'i' data = cudf.Series([1, 2, 3], d...
pd.Series(input_obj, dtype=np_dtypes_to_pandas_dtypes[dtype])
pandas.Series
import pandas as pd import glob import os import numpy as np import time import fastparquet import argparse from multiprocessing import Pool import multiprocessing as mp from os.path import isfile parser = argparse.ArgumentParser(description='Program to run google compounder for a particular file and setting') parse...
pd.DataFrame()
pandas.DataFrame
from __future__ import print_function, division from warnings import warn from nilmtk.disaggregate import Disaggregator from keras.layers import Conv1D, Dense, Dropout, Reshape, Flatten import os import pickle import pandas as pd import numpy as np from collections import OrderedDict from keras.optimizers import SGD fr...
pd.concat(train_main,axis=0)
pandas.concat
#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright (c) 2019 <NAME> <<EMAIL>> # and the Talkowski Laboratory # Distributed under terms of the MIT license. """ Collect features per gene for an input gtf """ import pybedtools as pbt import numpy as np import pandas as pd from pysam import faidx import csv fro...
pd.read_csv(exac_cnv_tsv, delimiter='\t')
pandas.read_csv
import os import re import sys import warnings from argparse import ArgumentParser warnings.filterwarnings('ignore', category=FutureWarning, module='rpy2.robjects.pandas2ri') import matplotlib.pyplot as plt import numpy as np import pandas as pd from pandas.api.types import is_object_dtype, is...
is_object_dtype(x)
pandas.api.types.is_object_dtype
import collections import dask from dask import delayed from dask.diagnostics import ProgressBar import logging import multiprocessing import pandas as pd import numpy as np import re import six import string import py_stringsimjoin as ssj from py_stringsimjoin.filter.overlap_filter import OverlapFilter from py_string...
pd.concat([ret_candset, missing_value_pairs], ignore_index=True, sort=False)
pandas.concat
import sys import pandas import numpy import math import numpy as np import networkx as nx from sklearn.preprocessing import normalize def ComputeRankWeightage(row): return (1+ row['max_actor_rank'] - row['actor_movie_rank'])/(1+ row['max_actor_rank'] - row['min_actor_rank']) def ComputeTimestampWeights(row, min_ti...
pandas.read_csv("movie-actor.csv")
pandas.read_csv
# This script gets the amount of funding gained within 9 months # of the first date import pandas as pd import numpy as np from datetime import datetime, timedelta from tqdm import tqdm # Need to merge two patreon stat CSV files df_patreon1 = pd.read_csv('files/20190714_github_patreon_stats.csv') df_patreon2 = pd.read_...
pd.concat([df_patreon1, df_patreon2], ignore_index=True)
pandas.concat
import os import numpy as np import pandas as pd from datetime import datetime from datetime import timedelta from dateutil import parser # from scipy.interpolate import NearestNDInterpolator import matplotlib.pyplot as plt # from emtracks.mapinterp import get_df_interp_func # copied interpolation here. FIXME! from ma...
pd.DataFrame(rows_list)
pandas.DataFrame
from typing import Tuple, Sequence, Mapping, Optional, Union try: from typing import Literal except ImportError: from typing_extensions import Literal from anndata import AnnData from copy import copy import pandas as pd import ntpath import numpy as np import matplotlib.pyplot as plt import scanpy as sc impor...
pd.read_table(psms, low_memory=False)
pandas.read_table
import collections import os import sys import joblib import numpy as np import pandas as pd import torch from sklearn.ensemble import GradientBoostingClassifier, RandomForestClassifier from sklearn.linear_model import LogisticRegression from sklearn.metrics import roc_auc_score from sklearn.model_selection import tra...
pd.DataFrame(columns=cols)
pandas.DataFrame