prompt
stringlengths
19
1.03M
completion
stringlengths
4
2.12k
api
stringlengths
8
90
#!/usr/local/bin/python """ Mess-around project to learn more python. I organize my card collection according to price. The $TRADE_BOX_THRESHOLD is for cards kept in my trade box. I adjust this whenever the trade box gets full. $1-$TRADE_BOX_THRESHOLD cards are kept in a separate less accessed box. Anything under $1...
pandas.notnull(row["OldCount"])
pandas.notnull
#!/usr/bin/env python # -*- coding: utf-8 -*- # @Time : 2019-05-22 17:45 # @Author : erwin import datetime import time import pandas as pd from component.demo_opentsdb.opentsdb_conn import OpenTSDBClient from machine_learning.similarity.dtw.hierarchical_helper import HierarchicalHelper from common.pickle_helper imp...
pd.set_option('display.max_colwidth', 1000)
pandas.set_option
#!/usr/bin/env python # ipdb> import os; os._exit(1) # call as: python convert_mat_to_excel.py # ======================================= # Version 0.1 # 30 March, 2019 # michael.taylor AT reading DOT ac DOT uk # ======================================= import os import os.path import glob import optparse from o...
pd.DataFrame(x, columns=x_cols, index=t)
pandas.DataFrame
import pandas as pd import numpy as np import sys, getopt import os from os import path import collections def process_kp_baseline_survey(data_dictionary_filename, data_filename, output_folder): print('input dd =', data_dictionary_filename) print('input df =', data_filename) print('output dir =', o...
pd.DataFrame(ipaq_scores)
pandas.DataFrame
""" Monte Carlo-type tests for the BM model Note that that the actual tests that run are just regression tests against previously estimated values with small sample sizes that can be run quickly for continuous integration. However, this file can be used to re-run (slow) large-sample Monte Carlo tests. """ import numpy...
pd.concat([endog1_M, f2, endog2_M], axis=1)
pandas.concat
''' Plots for my first-year (and beyond) figurative violence in the media project. Author: <NAME> <<EMAIL>> Date: April 01, 2017 ''' import matplotlib.pyplot as plt import matplotlib.dates as pltdates import pandas as pd import seaborn as sns from datetime import date, datetime, timedelta from .analysis import rela...
pd.DatetimeIndex([d0, d1, d2, d3, d4, d5])
pandas.DatetimeIndex
# -*- coding: utf-8 -*- """ These the test the public routines exposed in types/common.py related to inference and not otherwise tested in types/test_common.py """ from warnings import catch_warnings, simplefilter import collections import re from datetime import datetime, date, timedelta, time from decimal import De...
lib.infer_dtype(arr)
pandas._libs.lib.infer_dtype
import builtins from io import StringIO from itertools import product from string import ascii_lowercase import numpy as np import pytest from pandas.errors import UnsupportedFunctionCall import pandas as pd from pandas import ( DataFrame, Index, MultiIndex, Series, Timestamp, date_range, isna) import pandas.cor...
pd.DataFrame([{"a": 1, "b": i[0]}, {"a": 1, "b": i[1]}])
pandas.DataFrame
# -*- coding: utf-8 -*- """ @file:utils.py @time:2019/6/1 21:57 @author:Tangj @software:Pycharm @Desc """ import os import numpy as np from sklearn.base import BaseEstimator, TransformerMixin from sklearn.metrics import roc_auc_score from time import time import random import pandas as pd def frame_to_dict(train): ...
pd.concat([train1, train2])
pandas.concat
import altair as alt import pandas as pd from sys import argv import numpy as np df =
pd.read_csv(argv[1], keep_default_na=False)
pandas.read_csv
# -*- coding: utf-8 -*- # Global imports from __future__ import unicode_literals import os import glob import pickle import argparse as ap import pandas as pd import numpy as n import multiprocessing as mp # Plot imports import matplotlib.pyplot as plt import matplotlib matplotlib.use("tkagg") import seaborn as sns ...
pd.read_csv(report, sep=" ")
pandas.read_csv
# -*- coding: utf-8 -*- import datetime import time from sqlalchemy.sql import func import pandas as pd import math data=pd.read_csv('上证50_daily.csv',index_col=0) t=data.index f=0 jiange=30 res=[] import datetime # def get_day_nday_ago(date,n): # t = time.strptime(date, "%Y-%m-%d") # y, m, d = t[0:3] # D...
pd.DataFrame(res,index=[0])
pandas.DataFrame
# -*- coding: utf-8 -*- """ Created on Mon Jun 15 14:26:29 2020 @author: skyhigh """ import pandas as pd columns =['matth','english','science'] indexs = ['John','Julia',"Mary","Henry"] datas=[[80,70,90],[88,87,99],[77,66,76],[90,98,96]] df =
pd.DataFrame(datas,columns=columns,index=indexs)
pandas.DataFrame
# http://www.vdh.virginia.gov/coronavirus/ from bs4 import BeautifulSoup import csv from datetime import datetime from io import StringIO import os import requests import pandas as pd # Remove empty rows def filtered(rows): return [x for x in rows if "".join([(x[y] or "").strip() for y in x]) != ""] def run_VA(a...
pd.to_datetime(v_df[date_col])
pandas.to_datetime
import copy from GTac_Data import gtac_data from data_gen import raw_data_byts_checkout_2, collect_DataPoints # from data_collect_fingers_five import COLUMNS_RAW_FINGER_DATA, MAG_NUM, COL_INDEX from gtac_config import COLUMNS_RAW_FINGER_DATA, MAG_NUM, COL_INDEX # from Handover import collect_DataPoints, find_location,...
pd.DataFrame(columns=COLUMNS_RAW_FINGER_DATA)
pandas.DataFrame
import pandas as pd import json, os from datetime import timedelta f = open('data/game_def.json') game_def = json.load(f) def getHardwareLog(start_date, end_date, hwlogpath): df = pd.read_csv(hwlogpath, header=[1]) df['Time'] = pd.to_datetime(df['Time']).dt.tz_localize("America/New_York") mask = (df['Ti...
pd.Timedelta(seconds=trial['times'][idxLevel[1]])
pandas.Timedelta
import os import speedtest import time import sys import shutil import pandas as pd from pythonping import ping import argparse import logging import traceback import configparser from PyQt5.QtCore import pyqtSignal, QObject from modules.visuals import InteractivePlots class Communicate(QObject): GUI_signal = py...
pd.DataFrame(columns=["date", "min", "max", "avg", "url"])
pandas.DataFrame
# --- # jupyter: # jupytext: # formats: ipynb,py:percent # text_representation: # extension: .py # format_name: percent # format_version: '1.3' # jupytext_version: 1.9.1 # kernelspec: # display_name: udl # language: python # name: udl # --- # %% [markdown] # # Sentiment ...
pd.Series([0,0,0,0,1,1,1,1])
pandas.Series
from numpy import * import pandas as pd
pd.set_option('precision',2)
pandas.set_option
import numpy as np import pandas as pd from datetime import datetime, timedelta import pytest import vectorbt as vbt from vectorbt.utils.config import merge_dicts seed = 42 # ############# base.py ############# # class MyData(vbt.Data): @classmethod def download_symbol(cls, symbol, shape=(5, 3), start_dat...
pd.Int64Index([0, 1], dtype='int64', name='symbol')
pandas.Int64Index
import sys import pandas from decisionengine_modules.glideinwms.transforms.grid_figure_of_merit import GridFigureOfMerit grid_entries = ["g1", "g2", "g3", "g4", "g5"] running = [5, 10, 15, 20, 200] max_allowed = [10, 10, 10, 2000, 500] idle = [20, 3, 4, 5, 6] max_idle = [10, 10, 10, 10, 10] entries = { "EntryNa...
pandas.DataFrame(entries)
pandas.DataFrame
import pandas as pd import numpy as np import scipy import seaborn as sns import matplotlib.pyplot as plt import os from functools import reduce from statsmodels.tsa.stattools import coint sns.set(style='white') # Retrieve intraday price data and combine them into a DataFrame. # 1. Load downloaded prices from folder...
pd.DataFrame()
pandas.DataFrame
import numpy as np import pandas as pd import pdb import os import math import argparse if __name__ == '__main__': #edit the directory muat_dir = '/users/primasan/projects/muat/' metadata = pd.read_csv(muat_dir + 'extfile/metadata_icgc_pcawg.tsv',sep='\t',index_col=0) dictMutation = pd.read_csv(muat_...
pd.DataFrame(pd_all)
pandas.DataFrame
# -*- coding: utf-8 -*- """ Created on Sun Apr 11 22:34:12 2021 @author: orkun """ import numpy as np import pandas as pd from sklearn.impute import SimpleImputer from pandas.api.types import is_numeric_dtype from sklearn import preprocessing from PyQt5.QtWidgets import QMessageBox from PyQt5.QtCore import QSize from ...
pd.DataFrame(self.dataframe[col_name])
pandas.DataFrame
import pandas as pd import numpy as np import re #=============================================================================== def cleandots(x,mark): if str(x) == mark : # The mark was a '.' in HW 3 use case. return np.NaN else: return str (x) def cleandotsincolumn (series,mark): return ...
pd.reset_option('display.max_columns')
pandas.reset_option
# -*- coding: utf-8 -*- """ Tools for calculating the fatigue damage equivalent PSD. Adapted and enhanced from the CAM versions. """ from types import SimpleNamespace import itertools as it import multiprocessing as mp import numpy as np import scipy.signal as signal import pandas as pd from pyyeti import cyclecount, ...
pd.Series(SRSmax, index=index)
pandas.Series
import datetime import logging import json import requests from pandas import json_normalize import pandas as pd from google.cloud import storage from anyway.parsers.waze.waze_db_functions import ( insert_waze_alerts, insert_waze_traffic_jams, enrich_waze_alerts_ended_at_timestamp, enrich_waze_traffic_...
pd.to_datetime(waze_df["pubMillis"], unit="ms")
pandas.to_datetime
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Sat May 23 03:54:38 2020 @author: lukepinkel """ import numpy as np import scipy as sp import scipy.stats import pandas as pd from .linalg_operations import _check_shape def get_param_table(params, se_params, degfree=None, index=None, p...
pd.DataFrame(arr, index=index, columns=[parameter_label, 'SE'])
pandas.DataFrame
from datetime import datetime startTime = datetime.now() import json import glob import os import pandas as pd import tensorflow as tf import tensorflowjs as tfjs from tensorflow import keras from sklearn.model_selection import train_test_split import requests EPOCHS = 9 CLASSES = 2 """ Build and return the Keras mod...
pd.read_csv(path, skipinitialspace=True, low_memory=False)
pandas.read_csv
# coding: utf-8 # In[3]: import pandas as pd train = pd.read_csv('train.csv') test = pd.read_csv('test.csv') # In[4]: import matplotlib.pyplot as plt import seaborn as sns sns.set() # setting seaborn default for plots # In[5]: train_test_data = [train, test] # combining train and test dataset for dataset...
pd.DataFrame([Pclass1, Pclass2, Pclass3])
pandas.DataFrame
from bokeh.sampledata.us_states import data as stateBorders from bokeh.sampledata.us_counties import data as counties from COVID.extract import COVID_counts import pandas as pd import numpy as np import pickle # stateBorders['D.C.'] = stateBorders.pop('DC') stateBorders=
pd.DataFrame(stateBorders)
pandas.DataFrame
# -*- coding: utf-8 -*- """ Created on Tue Feb 4 22:47:56 2020 @author: nipunn """ def outlying_rows(filename): mydataset=
pd.read_csv(filename)
pandas.read_csv
import IPython import base64 import cv2 import json import numpy as np import pandas as pd import pravega.grpc_gateway as pravega from matplotlib import pyplot as plt import time def ignore_non_events(read_events): for read_event in read_events: if len(read_event.event) > 0: yield read_event ...
pd.to_datetime(df[self.timestamp_col], unit='ms', utc=True)
pandas.to_datetime
import csv from io import StringIO import os import numpy as np import pytest from pandas.errors import ParserError import pandas as pd from pandas import ( DataFrame, Index, MultiIndex, NaT, Series, Timestamp, date_range, read_csv, to_datetime, ) import pandas._testing as tm impo...
tm.assert_almost_equal(result, expected)
pandas._testing.assert_almost_equal
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import import json import os import torch import pandas as pd import numpy as np def save_csv_log(opt, head, value, is_create=False, file_name='test'): if len(value.shape) < 2: value = np.expand_dims(value, axis=0) df =
pd.DataFrame(value)
pandas.DataFrame
import streamlit as st import numpy as np import pandas as pd import plotly.graph_objects as go from datetime import datetime import requests class DataFetcher: def __init__(self): self.url_brazil_general = 'https://covid19-brazil-api.now.sh/api/report/v1/brazil/' self.url_brazil_states = 'https://...
pd.Timedelta(hours=3)
pandas.Timedelta
from pandas.util import hash_pandas_object import hashlib import pandas as pd import random random.seed(42) import numpy as np import psutil import time ROWS = 20000000 DATA = [random.random() for _ in range(ROWS)] def mem_use(): mem_profile = psutil.virtual_memory() print("Memory Usage = {} | percent ...
pd.concat(genera,axis=1)
pandas.concat
from scipy.stats import mannwhitneyu,wilcoxon import numpy as np from scipy.io import mmread import pandas as pd X = mmread('RFiles/all_data.mtx') X = X.tocsr() celllabels = np.load('Notebooks/meta/celllabels.npy') isCSF = np.load('Notebooks/meta/isCSF.npy') isMS = np.load('Notebooks/meta/isMS.npy') logX = np.log10...
pd.DataFrame([clusterid,stat,pvalue],index=['clusterid','stat','pvalue'])
pandas.DataFrame
from __future__ import annotations from collections import abc from datetime import datetime from functools import partial from itertools import islice from typing import ( TYPE_CHECKING, Callable, Hashable, List, Tuple, TypedDict, Union, cast, overload, ) import warnings import nu...
is_integer_dtype(values)
pandas.core.dtypes.common.is_integer_dtype
import json import sys import pandas as pd from pandas import DataFrame from db.sql import dal from flask import request import tempfile import tarfile import csv import shutil import subprocess from flask import send_from_directory from annotation.main import T2WMLAnnotation from db.sql.kgtk import import_kgtk_datafra...
pd.read_excel(request.files['file'], dtype=object, header=None)
pandas.read_excel
from sklearn.manifold import TSNE from clustering import silhouette as sil from data_processing import MulticlusteringExperimentUtils as expUtils # Keep the clustering experiments that involve outliers here from clustering.KMeansVariations import kMeans_baseline, kMeans_baseline_high_iteration, kMeans_baseline_random_...
pd.DataFrame(feature_set_outliers_removed)
pandas.DataFrame
from numpy import * import pandas as pd import datetime from datetime import timedelta def sum_duplicated(): fields = ['DATE', 'DAY_OFF', 'WEEK_END', 'DAY_WE_DS', 'ASS_ASSIGNMENT', 'CSPL_RECEIVED_CALLS' ] # selectionne les colonnes à lire x=pd.read_csv("data/train_2011_2012_2013.csv", sep=";", usecols=fields) ...
pd.DataFrame()
pandas.DataFrame
#%% # ANCHOR IMPORTS import sys import pandas as pd, numpy as np import pickle import re from sklearn import feature_extraction , feature_selection from scipy.sparse import csr_matrix, hstack from sklearn.feature_extraction.text import CountVectorizer from sklearn.feature_extraction import DictVectorizer from skl...
pd.concat(featList, axis=1)
pandas.concat
""" Copyright 2020 The Secure, Reliable, and Intelligent Systems Lab, ETH Zurich Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by appl...
pd.MultiIndex.from_tuples(product, names=["steps", "names", "prices", "samples"])
pandas.MultiIndex.from_tuples
#!/usr/bin/env python # coding: utf-8 # In[2]: import pandas as pd import numpy as np import matplotlib.pyplot as plt import seaborn as sns pd.options.display.max_columns = None # In[3]: df = pd.read_csv('full_data.csv') # In[4]: df.head() # Since i got the data from lolchess.gg, i only got informations o...
pd.set_option('display.max_rows', None)
pandas.set_option
# pylint: disable=E1101,E1103,W0232 from datetime import datetime, timedelta from pandas.compat import range, lrange, lzip, u, zip import operator import re import nose import warnings import os import numpy as np from numpy.testing import assert_array_equal from pandas import period_range, date_range from pandas.c...
Index([], name='Foo')
pandas.core.index.Index
from datetime import datetime, timedelta import inspect import numpy as np import pytest from pandas.core.dtypes.common import ( is_categorical_dtype, is_interval_dtype, is_object_dtype, ) from pandas import ( Categorical, DataFrame, DatetimeIndex, Index, IntervalIndex, MultiIndex...
tm.assert_frame_equal(result2, original)
pandas.util.testing.assert_frame_equal
# -*- coding: utf-8 -*- """ Created on Fri Aug 14 13:52:36 2020 @author: diego """ import os import sqlite3 import numpy as np import pandas as pd import plots as _plots import update_prices import update_companies_info pd.set_option("display.width", 400) pd.set_option("display.max_columns", 10) pd.options.mode.chai...
pd.DateOffset(months=12)
pandas.DateOffset
# Testing array.blend import utipy as ut import numpy as np import pandas as pd def test_blend_list(): x1 = [1, 2, 3, 4, 5, 6] x2 = [2, 3, 4, 5, 6, 7] blended0 = ut.blend(x1, x2, amount=0) blended1 = ut.blend(x1, x2, amount=1) blended05 = ut.blend(x1, x2, amount=.5) assert blended0 == x1 ...
pd.Series([2, 3, 4, 5, 6, 7])
pandas.Series
from dataProcessing import * from sklearn.ensemble import RandomForestClassifier from sklearn.ensemble import IsolationForest import pandas as pd import pickle from RandomForestCounterFactual import * def checkSamples( datasetFileName, unscaledFactualsFileName, unscaledCounterFactualsFileName, ...
pd.DataFrame()
pandas.DataFrame
# import hashlib # import random # import json import binascii import json import traceback import uuid import numpy as np import datetime import Crypto import Crypto.Random from Crypto.Random import get_random_bytes # from Crypto.Hash import SHA from Crypto.PublicKey import RSA from Crypto.Cipher import ...
pd.Series({'public_key':public_key,'name':name})
pandas.Series
"""Run the file manually to find all angle data.""" import math import numpy as np import pandas as pd from lilypadz.helper.constant import TOAD_HOP def convert_xyz_to_kinematic(xyz_data: pd.DataFrame) -> pd.DataFrame: """Calculate three kinematic variables from the XYZ data. :param xyz_data: xyz data of a ...
pd.read_csv(f"{name}/{hop}/xyz.csv")
pandas.read_csv
# # Copyright 2016 Quantopian, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in wr...
pd.Timestamp('2015-01-07 14:35', tz='UTC')
pandas.Timestamp
# -*- coding: utf-8 -*- from typing import Optional, Union import pandas as pd from mando import command from tstoolbox import tsutils try: from typing import Literal except ImportError: from typing_extensions import Literal try: from mando.rst_text_formatter import RSTHelpFormatter as HelpFormatter exce...
pd.Timestamp(DEFAULT_END_DATE)
pandas.Timestamp
from __future__ import absolute_import, division, print_function import pytest from datetime import datetime, timedelta import numpy as np import pandas as pd import pandas.util.testing as tm from pandas import DataFrame, Series from string import ascii_lowercase from blaze.compute.core import compute from blaze ...
Series(['1999-12-31', '2000-06-25'], dtype='M8[ns]', name='s')
pandas.Series
import pandas as pd import numpy as np import os from sklearn.linear_model import Ridge from sklearn.model_selection import cross_val_score from sklearn.model_selection import train_test_split from sklearn.metrics import mean_squared_error mos_data=pd.read_csv('data.csv')[['streaming_log','mos']] data=[] for l i...
pd.merge(mos_data, agg_data, on='streaming_log')
pandas.merge
from __future__ import absolute_import, division, print_function import json import logging import math import os import random import warnings from dataclasses import asdict from multiprocessing import cpu_count import numpy as np import pandas as pd import torch from scipy.stats import pearsonr from sklearn.metrics...
pd.DataFrame(training_progress_scores)
pandas.DataFrame
from datetime import ( datetime, timedelta, ) from importlib import reload import string import sys import numpy as np import pytest from pandas._libs.tslibs import iNaT import pandas.util._test_decorators as td from pandas import ( NA, Categorical, CategoricalDtype, Index, Interval, ...
CategoricalDtype(None, ordered=True)
pandas.CategoricalDtype
from db.DBConnector import execute_query from utils.log import log_init, log_close, log import time from os import path from pathlib import Path import pandas as pd import matplotlib.pyplot as plt import numpy as np # the ids are not included as they are the same for every table: id : long classMetricsEntities = ["cl...
pd.read_csv(file_path)
pandas.read_csv
#!/usr/bin/env python3 ''' FILE: nav01_parser.py DESCRIPTION: Nav01 parser class for raw output from a Furuno GP-90D GPS reciever Data file contains GGA/ZDA/VTG NMEA0183 sentences with no additional information added. BUGS: NOTES: AUTHOR: <NAME> COMPA...
pd.merge_asof(data, zda_df, on="lineno")
pandas.merge_asof
''''''''''''''''''''''''''''''''''''''''''''''''''''''''''''''''' RMDL: Random Multimodel Deep Learning for Classification * Copyright (C) 2018 <NAME> <<EMAIL>> * Last Update: May 3rd, 2018 * This file is part of RMDL project, University of Virginia. * Free to use, change, share and distribute source code of RMD...
pd.Series(shuffle_csv["class"])
pandas.Series
import os import logging from collections import defaultdict import pandas as pd from fol.foq_v2 import (concate_n_chains, copy_query, negation_sink, binary_formula_iterator, concate_iu_chains, parse_formula, ...
pd.DataFrame(data_dict)
pandas.DataFrame
import json import pandas as pd import numpy as np from sklearn.preprocessing import StandardScaler from sklearn.neighbors import NearestNeighbors def cluster(cohort_submissions: dict) -> list: """ Splits given dict into clusters of 4 based on their ranked complexity The 'remainder problem' of needing t...
pd.DataFrame.from_dict(cohort_submissions, orient="index")
pandas.DataFrame.from_dict
# To add a new cell, type '# %%' # To add a new markdown cell, type '# %% [markdown]' # %% # first, REMEMBER to activate cryptoalgowheel-S2 environment! # %% import datetime import os import sys import backtrader as bt import numpy as np import pandas as pd import matplotlib import PyQt5 # %% #*****WARNING: REVISE...
pd.to_datetime(from_datetime)
pandas.to_datetime
from torch.utils.data import DataLoader from torch.utils.tensorboard import SummaryWriter import torch from mbpert.loss import reg_loss_interaction, reg_loss_r, reg_loss_eps from mbpert.mbpert import MBPertDataset, MBPert import seaborn as sns import matplotlib.pyplot as plt import pandas as pd import numpy as np from ...
pd.DataFrame(data={'pred': x_pred_test, 'true': x_ss_test, 'value': 'x'})
pandas.DataFrame
from __future__ import print_function, absolute_import, division import pandas as pd import numpy as np import argparse import json import math import re import os import sys import csv import socket # -- ip checks import seaborn as sns import matplotlib.pyplot as plt from jinja2 import Environment, PackageLoader #...
pd.merge(dt, df_s1, on='_column', how='inner')
pandas.merge
### ETL script for generating input tables to model ### main point: ETL JHU covid-19 case and mortality data # todo: refactor ## HERE! -> not handled here in python -> Serial Interval Table -> would be worthwhile to reproduce the R for that here ## "serial interval table" <--> that discretized gamma distribution ## s...
pd.read_csv(jhu + deaths_csv)
pandas.read_csv
import json import logging import pandas as pd import requests from gamestonk_terminal import config_terminal as cfg from gamestonk_terminal.decorators import log_start_end from gamestonk_terminal.rich_config import console logger = logging.getLogger(__name__) api_url = "https://api.glassnode.com/v1/metrics/" GLAS...
pd.to_datetime(df["t"], unit="s")
pandas.to_datetime
import pandas as pd import matplotlib.pyplot as plt import statistics from datetime import datetime import copy import json import numpy as np from os import listdir from os.path import isfile, join import unknown import baseline import known_approx as kapprox import mary_optimal as mary dirs = ['/localdisk1/DOT/flig...
pd.concat(dfs)
pandas.concat
import pandas as pd import numpy as np d= pd.read_csv(snakemake.input[0], sep= '\t', header=0, compression= 'gzip') d= d.loc[~d['#chrom'].str.contains('_'), :] d['a1']= d.alts.str.split(',').str[0] d['a2']= d.alts.str.split(',').str[1] d['#chrom']= d['#chrom'].str.replace('chr', '') d['POS']= np.where(d.ref.str.len() ...
pd.concat([d, df])
pandas.concat
from context import dero import pandas as pd from pandas.util.testing import assert_frame_equal from pandas import Timestamp from numpy import nan import numpy class DataFrameTest: df = pd.DataFrame([ (10516, 'a', '1/1/2000', 1.01), (10516, 'a'...
Timestamp('2000-01-06 00:00:00')
pandas.Timestamp
"""Python Script Template.""" import os import pandas as pd H_PARAMS = "hparams.json" STATISTICS = "statistics.json" def get_name(h_params): """Get experiment name from hyper parameter json file.""" protagonist_name = h_params.protagonist_name[0] if protagonist_name in ["MVE", "Dyna", "STEVE"]: ...
pd.DataFrame()
pandas.DataFrame
import argparse import sys import pandas as pd def process_command_line(): """ Parse command line arguments `argv` is a list of arguments, or `None` for ``sys.argv[1:]``. Return a Namespace representing the argument list. """ # Create the parser parser = argparse.ArgumentParser(prog='mm...
pd.read_csv(perf_curve_pred_filename)
pandas.read_csv
import numpy as np import pytest from pandas import Categorical, Series import pandas._testing as tm @pytest.mark.parametrize( "keep, expected", [ ("first", Series([False, False, False, False, True, True, False])), ("last", Series([False, True, True, False, False, False, False])), (Fa...
Categorical(input1, categories=cat_array, ordered=ordered)
pandas.Categorical
from datetime import datetime import numpy as np import pytest import pandas.util._test_decorators as td from pandas.core.dtypes.base import _registry as ea_registry from pandas.core.dtypes.common import ( is_categorical_dtype, is_interval_dtype, is_object_dtype, ) from pandas.core.dtypes.dtypes import (...
tm.assert_index_equal(rs.index, rng)
pandas._testing.assert_index_equal
"""Assay information class""" import os import yaml import pandas as pd from .example_filetype_format import FileTypeFormat from . import process_functions class Assayinfo(FileTypeFormat): """Assay information file type""" _fileType = "assayinfo" _process_kwargs = ["newPath", "databaseSynId"] def...
pd.isnull(i)
pandas.isnull
import pandas as pd import xlsxwriter array = [['a1', 'a2', 'a3'], ['a4', 'a5', 'a6'], ['a7', 'a8', 'a9'], ['a10', 'a11', 'a12', 'a13', 'a14']] months = ['jan', 'feb', 'mar', 'apr', 'may'] df =
pd.DataFrame(array, columns=months)
pandas.DataFrame
import mne import pandas as pd from my_settings import * reject = dict(grad=4000e-13, # T / m (gradiometers) mag=4e-12, # T (magnetometers) eeg=180e-6 # ) result =
pd.DataFrame()
pandas.DataFrame
import json import pandas as pd import sys from tl.file_formats_validator import FFV from tl.exceptions import UnsupportTypeError from concurrent.futures import ThreadPoolExecutor from itertools import repeat class Utility(object): def __init__(self, es, output_column_name: str = 'retrieval_score', ...
pd.DataFrame(candidates_format)
pandas.DataFrame
# -*- coding: utf-8 -*- # pylint: disable-msg=W0612,E1101 import itertools import warnings from warnings import catch_warnings from datetime import datetime from pandas.types.common import (is_integer_dtype, is_float_dtype, is_scalar) from pandas.compat...
Series([0.4, np.nan, np.nan], index=exp_idx, name='s')
pandas.core.api.Series
## drive_path = 'c:/' import numpy as np import pandas as pd import os import sys import matplotlib.pyplot as plt from scipy.stats import ks_2samp from scipy.stats import anderson_ksamp from scipy.stats import kruskal from scipy.stats import variation from scipy import signal as sps import seaborn as sns import glob im...
pd.DataFrame([])
pandas.DataFrame
# -------------- # Importing header files import numpy as np import pandas as pd from scipy.stats import mode import warnings warnings.filterwarnings('ignore') #Reading file bank_data =
pd.read_csv(path)
pandas.read_csv
# import sys, os # sys.path.append( os.path.join( os.path.dirname( __file__ ), '..' ) ) import numpy as np import pandas as pd from . import ConfusionMatrix y_true = [1,1,3,1] y_pred = [1,2,2,1] labels = [1,2,3] names = ['foo','bar','baz'] def test_create_cmat(): ''' Check that constructing method, called w...
pd.Series([3,0,1], index=names )
pandas.Series
from collections import defaultdict import csv import pandas.compat as compat from pandas import DataFrame from pandas_datareader.base import _BaseReader _yahoo_codes = {'symbol': 's', 'last': 'l1', 'change_pct': 'p2', 'PE': 'r', 'time': 't1', 'short_ratio': 's7'} class YahooQuotesReader(_BaseReade...
compat.itervalues(_yahoo_codes)
pandas.compat.itervalues
#!/usr/bin/env python3 from argparse import ArgumentParser from collections import defaultdict import os import sys import matplotlib matplotlib.use('pdf') matplotlib.rcParams['pdf.fonttype'] = 42 matplotlib.rcParams['ps.fonttype'] = 42 matplotlib.rcParams['font.size'] = 12 import matplotlib.pyplot as plt import numpy...
pd.merge(df, check, how="left", on=["dataset", "id"])
pandas.merge
import streamlit as st import pandas as pd from pyvis.network import Network import networkx as nx import matplotlib.pyplot as plt import bz2 import pickle import _pickle as cPickle import pydot import math import numpy as num def decompress_pickle(file): data = bz2.BZ2File(file, 'rb') data = cPickle.load(data) re...
pd.DataFrame()
pandas.DataFrame
from datetime import datetime, timedelta, timezone import random from tabnanny import check import unittest import pandas as pd import pytz if __name__ == "__main__": from pathlib import Path import sys sys.path.insert(0, str(Path(__file__).resolve().parents[2])) from datatube.dtype import check_dtypes ...
pd.DataFrame(ALL_DATA)
pandas.DataFrame
import streamlit as st import pandas as pd import altair as alt import numpy as np from streamlit_option_menu import option_menu import xgboost as xgb from sklearn.model_selection import train_test_split from sklearn.metrics import mean_squared_error from sklearn.model_selection import cross_val_score, KFold from sklea...
pd.cut(x=state['avgSalary'], bins=[0,50,100,150,200, 250, 300], labels=category)
pandas.cut
""" This module contains all US-specific data loading and data cleaning routines. """ import datetime import requests import pandas as pd import numpy as np from .. import data idx = pd.IndexSlice def get_raw_covidtracking_data(run_date: pd.Timestamp): """ Gets the current daily CSV from COVIDTracking """ i...
pd.Timestamp("2020-06-03")
pandas.Timestamp
import numpy as np import pandas as pd import pytest from blocktorch.problem_types import ( ProblemTypes, detect_problem_type, handle_problem_types, is_binary, is_classification, is_multiclass, is_regression, is_time_series, ) @pytest.fixture def correct_problem_types(): # Unit te...
pd.Series([1, 0, 1, 0, 0, 1])
pandas.Series
# To add a new cell, type '#%%' # To add a new markdown cell, type '#%% [markdown]' #%% from IPython import get_ipython #%% import os import numpy as np import pandas as pd import matplotlib.pyplot as plt import io import base64 from matplotlib import animation from matplotlib import cm from matplotlib.pyplot import...
pd.DataFrame()
pandas.DataFrame
import pytest import pandas as pd from getdera import utils from pandas.testing import assert_frame_equal from getdera.dera import process # TESTCASES TEST_DATA_PATH = 'getdera/tests/data' TESTCASES = { 'process_tag': [ {'args': (f'{TEST_DATA_PATH}', 'risk', 'tag',...
assert_frame_equal(result, expected)
pandas.testing.assert_frame_equal
#!/usr/bin/env python # coding: utf-8 # # day6 宿題 # 作成:松島亮輔 # # 課題:住宅販売価格を予測する # In[1]: get_ipython().run_line_magic('matplotlib', 'inline') #グラフをnotebook内に描画させるための設定 import pandas as pd import numpy as np import matplotlib.pyplot as plt from sklearn import linear_model from sklearn.decomposition import PCA #主成分分...
pd.DataFrame(Y)
pandas.DataFrame
# -*- coding: utf-8 -*- import sys # import io from collections import OrderedDict from tabulate import tabulate import decimal from decimal import Decimal import itertools import numbers import string import numpy as np from scipy import stats import pandas as pd #import seaborn as sns import matplotlib as mpl impor...
pd.DataFrame(columns=['LATITUDE_SOUTH_180'])
pandas.DataFrame
# Copyright (c) 2018-2021, NVIDIA CORPORATION. import array as arr import datetime import io import operator import random import re import string import textwrap from copy import copy import cupy import numpy as np import pandas as pd import pyarrow as pa import pytest from numba import cuda import cudf from cudf.c...
pd.Index([0, 1, 2, 8, 11, 15])
pandas.Index
from sklearn.preprocessing import LabelEncoder from sklearn.feature_extraction.text import TfidfVectorizer from sklearn.model_selection import train_test_split from tensorflow.keras.models import Sequential from tensorflow.keras.callbacks import EarlyStopping, ModelCheckpoint from tensorflow.keras.layers import Dr...
pd.DataFrame(Xtrain)
pandas.DataFrame
from functools import partialmethod import pandas as pd from sqlalchemy.orm import sessionmaker from sqlalchemy import create_engine import sqlite3 import click import json import pkg_resources from itertools import combinations from q2_mlab.db.schema import RegressionScore from q2_mlab.plotting.components import ( ...
pd.crosstab(data[by], data[category])
pandas.crosstab
import plotly.graph_objects as go from plotly.subplots import make_subplots from parameters_cov import params import pandas as pd import numpy as np from math import ceil import datetime from data_constants import POPULATIONS month_len = 365/12 longname = {'S': 'Susceptible', 'I': 'Infected', 'R': '...
pd.Series(sol['y'][index[name],:])
pandas.Series
# -*- coding: utf-8 -*- """ Created on Mon Jan 27 13:30:31 2020 @author: User """ import sys import datetime as dt from collections import Counter import pprint import matplotlib import matplotlib.pyplot as plt import matplotlib.lines as mlines from matplotlib import cm from matplotlib import gridspe...
pd.read_pickle(IndexOVV_ORRpars_fn)
pandas.read_pickle
import numpy as np import pandas as pd from copy import copy, deepcopy from matplotlib import pyplot as plt from datetime import datetime, timedelta from matplotlib.backends.backend_pdf import PdfPages dfheight=
pd.read_csv('../data/raw/Results from Val_Roseg_Timelapse in µm per sec.csv')
pandas.read_csv