prompt
stringlengths
19
1.03M
completion
stringlengths
4
2.12k
api
stringlengths
8
90
import argparse import os import pandas as pd from af_dataset_builder import AFDatasetBuilder import plots import utils CLASS_NAMES = ['normal', 'af', 'other', 'noise'] SEED = 3 def summarize_metrics(models_path, test_set_path, target_record_len, batch_size, model_name=None, output_path=None): all_metrics =
pd.DataFrame()
pandas.DataFrame
import bs4 as bs import urllib.request import pandas as pd import random import pathlib import progressbar import numpy as np def schedule(names): schedule = [['YourScore', "OppScore","FGA","FGP","3PA","3PP","FTA","FTP","TRB","STL","BLK","TOV","OppFGA","OppFGP","Opp3PA","Opp3PP","OppFTA","OppFTP","OppTRB"...
pd.read_csv(url, names=names,encoding='utf-8')
pandas.read_csv
import argparse import pandas as pd import os from random import shuffle def parse_args(): parser = argparse.ArgumentParser(description="Takes the meta_data, l4, and l4_no_pca files for the train, val and test sets " "of students and returns them in libsvc format....
pd.DataFrame([])
pandas.DataFrame
import climetlab as cml from . import DATA_VERSION, PATTERN_GRIB, PATTERN_NCDF class Info: def __init__(self, dataset): import os import yaml self.dataset = dataset filename = self.dataset.replace("-", "_") + ".yaml" path = os.path.join(os.path.dirname(os.path.abspath(__...
pd.to_datetime(only_one_date)
pandas.to_datetime
import pandas as pd import numpy as np import math from openpyxl import load_workbook # Dictionary of expiry dates hard coded as historical expiry dates are not readily available expdct = {'10APR20': 1586505600000, '17APR20': 1587110400000, '24APR20': 1587715200000 } # Arbit...
pd.ExcelWriter(data_destination, engine="openpyxl", mode="a")
pandas.ExcelWriter
""" This script aggregates zugdata on a daily basis and uploads it in /live/aggdata """ import os import re import pandas as pd from datetime import datetime, date, timedelta # compatibility with ipython #os.chdir(os.path.dirname(__file__)) import json import boto3 from pathlib import Path from coords_to_kreis import c...
pd.DataFrame(json_content)
pandas.DataFrame
from functools import lru_cache import datetime from typing import Tuple, List, Callable, NamedTuple from collections import namedtuple import sqlalchemy import pandas as pd def get_securities(): return pd.read_sql('securities', con=sqlalchemy.create_engine('sqlite:///../data/jq.db')) @lru_cache(maxsize=1) def...
pd.DataFrame(data, columns=['ts_code', 'mg', 'mg_rank', 'ms', 'ms_rank'])
pandas.DataFrame
# %% import os import pandas as pd import numpy as np import threading import time base_dir = os.getcwd() # %% # 初始化表头 header = ['user', 'n_op', 'n_trans', 'op_type_0', 'op_type_1', 'op_type_2', 'op_type_3', 'op_type_4', 'op_type_5', 'op_type_6', 'op_type_7', 'op_type_8', 'op_type_9', 'op_type_perc', 'op_ty...
pd.read_csv(base_dir + '/dataset/dataset2/testset/test_a_op.csv')
pandas.read_csv
import os import multiprocessing as mp import numpy as np import matplotlib.pyplot as plt import pandas as pd import tqdm import cv2 # Specify HSV color range for detection lower = (25, 40, 200) upper = (30, 100, 255) data_dir = 'data/pufferfish-struggle' num_images = len([name for name in os.listdir(data_dir) if ...
pd.DataFrame(progress)
pandas.DataFrame
import nose import os import string from distutils.version import LooseVersion from datetime import datetime, date, timedelta from pandas import Series, DataFrame, MultiIndex, PeriodIndex, date_range from pandas.compat import range, lrange, StringIO, lmap, lzip, u, zip import pandas.util.testing as tm from pandas.uti...
tm.close()
pandas.util.testing.close
# -*- coding: utf-8 -*- import argparse import json import os import re from io import StringIO from pathlib import Path import dotenv import pandas as pd import requests from utils import get_gene_id2length DOTENV_KEY2VAL = dotenv.dotenv_values() def make_tissue2subtissue2sample_id(rawdir: str) -> pd.DataFrame: ...
pd.MultiIndex.from_frame(sample_id2tissue_type_subtype_df)
pandas.MultiIndex.from_frame
import unittest import pandas as pd import numpy as np from autopandas_v2.ml.featurization.featurizer import RelationGraph from autopandas_v2.ml.featurization.graph import GraphEdge, GraphEdgeType, GraphNodeType, GraphNode from autopandas_v2.ml.featurization.options import GraphOptions get_node_type = GraphNodeType.g...
pd.MultiIndex.from_tuples(tuples)
pandas.MultiIndex.from_tuples
# coding: utf-8 # CS FutureMobility Tool # See full license in LICENSE.txt. import numpy as np import pandas as pd #import openmatrix as omx from IPython.display import display from openpyxl import load_workbook,Workbook from time import strftime import os.path import mode_choice.model_defs as md import mode_choice.ma...
pd.concat([town_definition,zone_vmt_daily_o],axis=1,join='inner')
pandas.concat
#!/usr/bin/env python # -*- coding:utf-8 -*- """ Date: 2022/5/11 17:52 Desc: 加密货币 https://cn.investing.com/crypto/currencies 高频数据 https://bitcoincharts.com/about/markets-api/ """ import math import pandas as pd import requests from tqdm import tqdm from akshare.datasets import get_crypto_info_csv def crypto_name_ur...
pd.DataFrame()
pandas.DataFrame
import inspect from datetime import datetime from tralo.utils import filter_args, sha1_hash_object, valid_run, AttributeDict, get_attribute import yaml import os import json import re import torch from os.path import join, isfile, expanduser, realpath from tralo.log import log def load_model(checkpoint_id, weights_fi...
DataFrame(table)
pandas.DataFrame
# -*- coding: utf-8 -*- from unittest import TestCase from parameterized import parameterized import pandas as pd import numpy as np from numpy.testing.utils import assert_array_equal from pandas import (MultiIndex, Index) from pandas.util.testing import assert_frame_equal, assert_series_equal from...
pd.DataFrame({'001': [1, 2, 3], '002': [2, 3, 4]}, index=['2014', '2015', '2016'])
pandas.DataFrame
import torch from torch.utils.data import Dataset import pandas as pd import numpy as np class DeepInflammationDataset(Dataset): def __init__(self, c1_data, c2_data): """ - c1_data: pandas dataframe of the cell 1 data - c2_data: pandas dataframe of the cell 2 data """ supe...
pd.concat([self.c1_data['expr'], self.c2_data['expr']], axis=0, ignore_index=True)
pandas.concat
import textattack import textattack.datasets as datasets import random import pandas as pd from textattack.transformations.word_swap_embedding import WordSwapEmbedding as WordSwapEmbedding from textattack.constraints.semantics.word_embedding_distance import WordEmbeddingDistance as WordEmbeddingDistance NUM_NEAREST = ...
pd.DataFrame()
pandas.DataFrame
#!/usr/local/bin/python import argparse import os import sys import pandas as pd import numpy as np import time pd.options.mode.chained_assignment = None parser = argparse.ArgumentParser(prog='snvScore') parser.add_argument('SampleBED',type=str,help='Path to the mosdepth per-base BED output') parser.add_argument('SNVG...
pd.read_csv(args.SNVGermlineTXT,sep='\t')
pandas.read_csv
#!/usr/bin/python3 #By <NAME> import os import re import argparse import time import timeit import numpy as np import pandas as pd import pipeline_base def merge_window(intervals,vcf,ref_sequence,log_file=False,fullcheck=True,ignored=True,info_just_indel=False,drop_info=False,intervals_alignment_bool=False,interv...
pd.read_csv(interval,sep='\t')
pandas.read_csv
from __future__ import annotations import os import pandas as pd import streamlit as st st.set_page_config(layout="wide") class Pager: """Generates cycled stepper indices.""" def __init__(self, count) -> None: self.count = count self.current = 0 @property def next(self) -> int: """Fetches next index."...
pd.DataFrame(data)
pandas.DataFrame
#################################################### ## TO TRAIN AND TEST CLASSIFIER ## train : to train classifier with train dataset ## test : to predict labels of validation dataset ## submit: to predict labels of test dataset #################################################### import time import numpy a...
pd.concat([results_df, row_df])
pandas.concat
from flask import Flask, render_template, request, redirect, url_for, session import pandas as pd import pymysql import os import io #from werkzeug.utils import secure_filename from pulp import * import numpy as np import pymysql import pymysql.cursors from pandas.io import sql #from sqlalchemy import create...
pd.DataFrame(data['TotalDemand'])
pandas.DataFrame
from rdkit import Chem from rdkit.Chem import rdmolops, rdMolDescriptors, Crippen, GraphDescriptors import numpy as np import pandas as pd import pkg_resources import sys def crippenHContribs(mol,contribs): """Adds Crippen molar refractivity atomic contributions from attached H atoms to a heavy atom's contribution...
pd.read_csv(stream)
pandas.read_csv
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """This is to find out why my R peaks at a value around 2021-07-01, that is much higher than RIVM's. Created on Fri Jul 23 12:52:53 2021 @author: hk_nien """ import matplotlib.pyplot as plt import pandas as pd import numpy as np import tools import nlcovidstats as nlcs ...
pd.Timedelta(4, 'd')
pandas.Timedelta
''' This is a script for editing OS's ITN Road Network shapefiles so that: - the attributeds include an id for the 'to' and 'from' nodes - line strings are duplicated along links that are bidirectional Purpose of this script is to use the extracted orientation information from the gml data to edit the roads linesting ...
pd.concat([gdfORLink, gdfORLinkReversed])
pandas.concat
from sklearn.ensemble import ExtraTreesRegressor, RandomForestRegressor,RandomForestClassifier import pickle import scikitplot as skplt from sklearn import tree from sklearn.naive_bayes import GaussianNB from sklearn.neighbors import KNeighborsClassifier from sklearn.metrics import confusion_matrix,precision_score,rec...
pd.read_csv('Classification_winter_data.csv')
pandas.read_csv
import os import cv2 import pandas as pd import dataset_settings from util import insert_into_df, write_info, resize_image def prepare_subset_data(data_path, train_csv_path, test_csv_path, source_url): count = {'normal': 0, 'pneumonia': 0, 'covid-19': 0} train_csv = pd.read_csv(train_csv_path, nrows=None) ...
pd.read_csv(test_csv_path, nrows=None)
pandas.read_csv
import plotly.graph_objs as go import math import pandas as pd color_palette = ['#586BA4', '#324376', '#F5DD90', '#F68E5F', '#F76C5E'] def create_timedelta_graph(events_df): if events_df.empty: x_values = list() y_values = list()...
pd.to_datetime(events_df['timestamp'])
pandas.to_datetime
# coding: utf-8 # In[1]: import numpy as np from keras.models import Sequential from keras.layers import Dense, Dropout from sklearn import preprocessing from keras.optimizers import SGD import pandas as pd # In[4]: def xtrain_and_test(df_all): ''' 得到训练数据和测试数据 ''' df_label = pd.read_csv('../data...
pd.read_csv('../data/public/evaluation_public.csv')
pandas.read_csv
import pandas as pd import numpy as np from sklearn.metrics import mean_squared_error from sklearn.externals import joblib from collections import OrderedDict import json import argparse import os # 加上48小时前obs 信息 # 处理 RAIN 值 去除 35以上数值 target_list=['t2m','rh2m','w10m'] from datetime import timedelta from datetime import...
pd.to_datetime(opt.last_day)
pandas.to_datetime
import os import sys sys.path.append(os.path.abspath(os.path.dirname(__file__) + '/' + '../..')) import numpy as np import pandas as pd from python.tools import ( clean_folder ) # Formatters for LaTeX output def f1(x): return '%1.0f' % x def f2(x): return '%1.2f' % x ################ ## Parameters ## #...
pd.DataFrame(res)
pandas.DataFrame
import pandas as pd import matplotlib.pyplot as plt import glob import numpy # open btc csv clean and get ready btc = pd.read_csv('btc_dataset.csv') btc['Date'] = pd.to_datetime(btc['Date'])# btc= btc.set_index('Date') Btc =
pd.DataFrame(btc.ix['2015-01-01':, "Avg_price"])
pandas.DataFrame
# coding=utf-8 import pandas as pd from mock import MagicMock from sparkmagic.livyclientlib.exceptions import BadUserDataException from nose.tools import assert_raises, assert_equals from sparkmagic.livyclientlib.command import Command import sparkmagic.utils.constants as constants from sparkmagic.livyclientlib.sendpa...
pd.DataFrame({"A": [1], "B": [2]})
pandas.DataFrame
import pandas as pd # bookings_to_arr # # Accepts a pandas dataframe containing bookings data and returns a pandas # dataframe containing changes in ARR with the following columns: # - date - the date of the change # - type - the type of the change (new, upsell, downsell, and churn) # - customer_id - the id o...
pd.Timestamp(ts_input="9/25/2020", tz="UTC")
pandas.Timestamp
# coding: utf-8 import tweepy import json import os from datetime import datetime import pandas as pd import credentials.credentials_twitter as cred class Twitter_Analysis: """ copyright© 2019 — <NAME> - License MIT """ __consumer_key = cred.CONSUMER_KEY __token = cred.TOKEN __api = None def __in...
pd.DataFrame(new)
pandas.DataFrame
from copy import deepcopy from sklearn.model_selection import KFold import numpy as np import pandas as pd from .data_augmentation import DACombine from core.models.metrics import avg_loss, mse, rejection_ratio, avg_win_loss, avg_loss_ratio, loss_sum, invariance benchmark_functions = [avg_loss, mse, rejection_ratio,...
pd.DataFrame(results)
pandas.DataFrame
# -*- coding: utf-8 -*- """ Created on Fri Jun 24 00:52:56 2016 @author: ARM """ import numpy as np import pandas as pd import matplotlib.pyplot as plt from keras.models import Sequential, Graph from keras.layers.core import Dense, Dropout, Activation from keras.optimizers import SGD, Adam from sklearn.cross_validati...
pd.read_csv(name5,' ')
pandas.read_csv
import pandas as pd import numpy as np from sklearn.metrics import roc_curve, auc, confusion_matrix, precision_score, recall_score, f1_score from sklearn.metrics import average_precision_score, precision_recall_curve from ._woe_binning import woe_binning, woe_binning_2, woe_binning_3 class Metrics: def __init__(s...
pd.merge(dev, val, how='left', on=['var_name', 'var_cuts'], suffixes=['_dev','_val'])
pandas.merge
import unittest import qteasy as qt import pandas as pd from pandas import Timestamp import numpy as np from numpy import int64 import itertools import datetime from qteasy.utilfuncs import list_to_str_format, regulate_date_format, time_str_format, str_to_list from qteasy.utilfuncs import maybe_trade_day, is_market_tr...
pd.to_datetime(prev_seems_trade_day)
pandas.to_datetime
#!/usr/bin/env python3 # requirement: a unique train path for each given bitmap day and UID import app.solr as solr import json import pandas as pd DAY = pd.offsets.Day() MONDAY = pd.offsets.Week(weekday=0) WEEK = 7 * DAY N = 0 def days_str(n): return '{:b}'.format(n).zfill(7) def day_int(bitmap): return i...
pd.DataFrame(UPDATE)
pandas.DataFrame
# import libraries import requests import numpy as np import pandas as pd from bs4 import BeautifulSoup as bs import time import random import re import os # Important Note --- # change the value for which you want to scrape the data defaults to 2008-2019 year_list = [year for year in range(2019, 2007, -1)] # proje...
pd.to_numeric(df["Mat"], errors="coerce")
pandas.to_numeric
# -*- coding: utf-8 -*- """Provides programs to process and analyze GOES X-ray data.""" from __future__ import absolute_import import datetime import matplotlib.dates from matplotlib import pyplot as plt from astropy.io import fits as pyfits from numpy import nan from numpy import floor from pandas import DataFrame f...
DataFrame({'xrsa': newxrsa, 'xrsb': newxrsb}, index=times)
pandas.DataFrame
import viola import pandas as pd from io import StringIO import sys, os HERE = os.path.abspath(os.path.dirname(__file__)) data_expected = """vcf1_test1 0 small_del vcf2_test1 0 small_del vcf1_test2 0 small_del vcf2_test2 0 small_del vcf1_test3 0 large_del vcf2_test3 0 large_del vcf1_test4 0 large_del vcf2_test4 0 large...
pd.testing.assert_frame_equal(manual_sv_type, manual_sv_type_expected, check_like=True)
pandas.testing.assert_frame_equal
from os.path import abspath, dirname, join, isfile, normpath, relpath from pandas.testing import assert_frame_equal from numpy.testing import assert_allclose from scipy.interpolate import interp1d import matplotlib.pylab as plt from datetime import datetime import mhkit.wave as wave from io import StringIO import panda...
pd.Series(data['freqBinWidth'])
pandas.Series
import re import pandas as pd from .soup import get_soup, table_to_df TICKER_IN_PARENTHESIS_RE = re.compile(r'(?P<company_name>.+) \((?P<ticker>[A-Z]+)\)') def get_wiki_table_df(url, index_col=None, columns=None): """Returns the first table of a Wikipedia page as a DataFrame""" soup = get_soup(url) tabl...
pd.DataFrame(d)
pandas.DataFrame
import io import os from random import choice import pandas as pd import torch import torch.nn as nn from PIL import Image from torch.utils.data import DataLoader from torchvision import transforms as T from torchvision.datasets import ImageFolder from torchvision.models.resnet import BasicBlock, ResNet SANITY_DIR = ...
pd.DataFrame([{'accuracy': acc, 'loss': loss}])
pandas.DataFrame
import calendar import datetime import numpy as np import pandas as pd from pandas.util.testing import (assert_frame_equal, assert_series_equal, assert_index_equal) from numpy.testing import assert_allclose import pytest from pvlib.location import Location from pvlib import solarposi...
assert_frame_equal(expected_solpos, ephem_data[expected_solpos.columns])
pandas.util.testing.assert_frame_equal
from datetime import datetime, timedelta import dateutil import numpy as np import pytest import pytz from pandas._libs.tslibs.ccalendar import DAYS, MONTHS from pandas._libs.tslibs.period import IncompatibleFrequency from pandas.compat import lrange, range, zip import pandas as pd from pandas import DataFrame, Seri...
pd.DatetimeIndex([1457537600000000000, 1458059600000000000])
pandas.DatetimeIndex
import pandas as pd import numpy as np import seaborn as sns import matplotlib.pyplot as plt import random import pickle import missingno as msno from sklearn.experimental import enable_iterative_imputer from sklearn.impute import IterativeImputer from sklearn.preprocessing import StandardScaler from sklearn.decomposi...
pd.DataFrame(y_test)
pandas.DataFrame
import pandas as pd import numpy as np import matplotlib.pyplot as plt from sklearn.model_selection import KFold from sklearn.preprocessing import StandardScaler from scipy.stats import multivariate_normal as mvn import seaborn as sn import math import gc import tensorflow as tf from tensorflow.keras.models import Sequ...
pd.DataFrame(data=X_test_data, columns=X_ID2)
pandas.DataFrame
import matplotlib matplotlib.use('Agg') import tessreduce as tr import numpy as np import pandas as pd import matplotlib.pyplot as plt import lightkurve as lk from astropy.coordinates import SkyCoord from astropy import units as u import os dirname = os.path.dirname(__file__) #where we're going we dont need warnings...
pd.DataFrame(columns=['mjd','flux','err','trend1','trend2','zp','sector'])
pandas.DataFrame
import numpy as np import pandas as pd from insolver.frame import InsolverDataFrame from insolver.transforms import InsolverTransform, AutoFillNATransforms def test_fillna_numerical(): df_test = InsolverDataFrame(pd.DataFrame(data={'col1': [1, 2, np.nan]})) df_transformed = InsolverTransform(df_test, [ ...
pd.DataFrame(data={'col1': [np.nan, np.nan, np.nan]})
pandas.DataFrame
""" This module defines geometric methods that work in 3D and allow receiverpoints and observation objects to interact with a map """ # rays,to_crs used in observations # fresnel,to_crs,is_outside,ground_level used in sim # map_to_crs is a standalone map method from itertools import chain, compress, cycle, repeat f...
pd.Series((point.z for point in points), index=points.index)
pandas.Series
import json import pandas as pd from pandas import DataFrame import matplotlib.pyplot as plt import os import collections import nltk.classify import nltk.metrics import numpy as np import csv """ read all business id """ business=[] users=[] scores=[] rates=[] t=0 userdd=
pd.read_csv('users.tsv', sep="\t")
pandas.read_csv
import unittest from zeppos_bcpy.sql_statement import SqlStatement import pandas as pd import os class TestTheProjectMethods(unittest.TestCase): def test_get_table_create_statement_method(self): df =
pd.DataFrame({'seconds': [3600], 'minutes': [10]}, columns=['seconds', 'minutes'])
pandas.DataFrame
import shelve import numpy as np import re import pandas as pd import json import pickle import pdb from copy import copy with open('metadata/bacnet_devices.json', 'r') as fp: sensor_dict = json.load(fp) nae_dict = dict() nae_dict['bonner'] = ["607", "608", "609", "557", "610"] nae_dict['ap_m'] = ['514', '513','6...
pd.read_csv('metadata/bacnettype_mapping.csv')
pandas.read_csv
# Librairies print("Load Libraries") import os import hashlib import numpy as np import pandas as pd import tensorflow.keras.preprocessing.image as kpi import tensorflow.keras.models as km from tensorflow.python.client import device_lib MODE = "GPU" if "GPU" in [k.device_type for k in device_lib.list_local_devices(...
pd.DataFrame(array, columns=["filename","probabilities","classes"])
pandas.DataFrame
''' This program will calculate a timeseries of active users across the lifetime of a project (or a workflow id/version for a project). The inputs needed are: the classification export file (request & download from the Project Builder) [optional] the workflow id [optional] the workflow version (only the major (...
pd.to_datetime(t_temp, format='%Y-%m-%d %H:%M:%S')
pandas.to_datetime
# ---------------------------------------------------------------------------- # Copyright (c) 2013--, scikit-bio development team. # # Distributed under the terms of the Modified BSD License. # # The full license is in the file COPYING.txt, distributed with this software. # --------------------------------------------...
pd.Index([])
pandas.Index
#!/usr/bin/env python # coding: utf-8 # In[ ]: # This Python 3 environment comes with many helpful analytics libraries installed # It is defined by the kaggle/python docker image: https://github.com/kaggle/docker-python # For example, here's several helpful packages to load in import numpy as np # linear algebra i...
pd.get_dummies(df['Gender'], drop_first=True)
pandas.get_dummies
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Fri Jan 26 15:39:02 2018 @author: joyce """ import pandas as pd import numpy as np from numpy.matlib import repmat from stats import get_stockdata_from_sql,get_tradedate,Corr,Delta,Rank,Cross_max,\ Cross_min,Delay,Sum,Mean,STD,TsRank,TsMax,TsMin,DecayLinea...
pd.DataFrame(r['r1'] + r['r2'])
pandas.DataFrame
#!/usr/bin/env python # coding: utf-8 # In[ ]: import xlrd from sklearn.model_selection import train_test_split from sklearn.svm import SVC from sklearn.metrics import classification_report, confusion_matrix from sklearn.metrics import roc_curve, auc, accuracy_score import matplotlib.pyplot as plt import xgboost as...
DataFrame(X_test)
pandas.core.frame.DataFrame
__author__ = "<NAME>" __license__ = "Apache 2" __version__ = "1.0.0" __maintainer__ = "<NAME>" __website__ = "https://llp.berkeley.edu/asgari/" __git__ = "https://github.com/ehsanasgari/" __email__ = "<EMAIL>" __project__ = "1000Langs -- Super parallel project at CIS LMU" import requests from bs4 import BeautifulSoup ...
pd.read_table('../meta/massive_par_stat.tsv', sep='\t')
pandas.read_table
# -*- coding: utf-8 -*- """Creates folders and files with simulated data for various characterization techniques Notes ----- All data is made up and does not correspond to the materials listed. The data is meant to simply emulate real data and allow for basic analysis. @author: <NAME> Created on Jun 15, 2020 """ fr...
pd.DataFrame(data)
pandas.DataFrame
import pandas as pd import numpy as np import pandas as pd from datetime import datetime import matplotlib.pyplot as plt import itertools from dateutil.relativedelta import relativedelta import sklearn.tree as tree from sklearn.neural_network import MLPClassifier from imblearn.over_sampling import SMOTE, ADASY...
pd.DataFrame()
pandas.DataFrame
from __future__ import annotations from typing import Optional, List, Dict, Tuple import logging import textwrap import pandas as pd import numpy as np import h5py from tqdm import tqdm from .catmaid_interface import Catmaid, Bbox, ConnectorDetail from .utils import CoordZYX logger = logging.getLogger(__name__) de...
pd.HDFStore(fpath, "r")
pandas.HDFStore
import pandas as pd from collections import Counter import sklearn.preprocessing as preprocessing import numpy as np import os from pandas import Series import seaborn as sns from sklearn.preprocessing import LabelEncoder, OneHotEncoder from sklearn.feature_selection import SelectKBest column_names = ['age','workclass...
pd.read_csv(data,sep='\s*,\s*',encoding='ascii',names = column_names,engine='python')
pandas.read_csv
""" Functions and objects to work with mzML data and tabular data obtained from third party software used to process Mass Spectrometry data. Objects ------- MSData: reads raw MS data in the mzML format. Manages Chromatograms and MSSpectrum creation. Performs feature detection on centroid data. Functions ---...
pd.Series(df.columns)
pandas.Series
import os import argparse import itertools import numpy as np import pandas as pd import scipy.sparse as sp from tqdm import tqdm from shqod import ( read_trajec_csv, trajec, read_level_grid, od_matrix, calculate_field, field_to_dict, mobility_functional, fractalD, trajectory_lengt...
pd.concat(dfs)
pandas.concat
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Mon Jul 31 15:16:47 2017 @author: wasifaahmed """ from flask import Flask, flash,render_template, request, Response, redirect, url_for, send_from_directory,jsonify,session import json as json from datetime import datetime,timedelta,date from sklearn.cluste...
pd.Series(Tfirt_x)
pandas.Series
# --- # jupyter: # jupytext: # text_representation: # extension: .py # format_name: percent # format_version: '1.3' # jupytext_version: 1.4.2 # kernelspec: # display_name: Python 3 # language: python # name: python3 # --- # %% [markdown] # # P1 REST API # # - This Jupyter no...
pd.DataFrame.from_records(data["payload_data"])
pandas.DataFrame.from_records
from six import string_types, text_type, PY2 from docassemble.webapp.core.models import MachineLearning from docassemble.base.core import DAObject, DAList, DADict from docassemble.webapp.db_object import db from sqlalchemy import or_, and_ from sklearn.datasets import load_iris from sklearn.ensemble import RandomForest...
pd.DataFrame(data)
pandas.DataFrame
import os import warnings import contextlib import tempfile import zipfile from typing import Iterable import logging import pandas as pd from .base import BaseEndpoint from ..models.odata import Series, Task logger = logging.getLogger(__name__) @contextlib.contextmanager def _with_dir_path(dir_path=None): # p...
pd.DataFrame(df_data)
pandas.DataFrame
""" This package will create the simplified comix matrix needed by simple_network_sim basing itself in the following data: - The CoMix matrix: https://cmmid.github.io/topics/covid19/reports/20200327_comix_social_contacts.xlsx - The Scottish demographics (NRS): ftp://boydorr.gla.ac.uk/scrc/human/demographics/scotland/d...
pd.DataFrame(rows)
pandas.DataFrame
# -*- coding: utf-8 -*- """ Created on Tue Jun 12 08:47:38 2018 @author: cenv0574 """ import os import json import pandas as pd import geopandas as gpd from itertools import product def load_config(): # Define current directory and data directory config_path = os.path.realpath( os.path.join(os.path...
pd.DataFrame(tax_sub)
pandas.DataFrame
#!python3 # -*- coding:utf-8 -*- import os import numpy as np import pandas as pd ''' This source code is a sample of using pandas library. Series and DataFrame. ''' # Series Object: It's one dimension object. # It's not ndarray, list, and other sequense object. print("make instance of Series") ser =
pd.Series([10,20,30,40])
pandas.Series
""" A class to carry localization data. """ import copy import logging import time import warnings from itertools import accumulate import numpy as np import pandas as pd from google.protobuf import json_format, text_format try: from scipy.spatial import QhullError except ImportError: from scipy.spatial.qhu...
pd.concat([self.dataframe, new_df], axis=1)
pandas.concat
import baostock as bs import pandas as pd import numpy as np from IPython import embed class Data_Reader(): """ reading the data from the file """ def __init__(self, file="stock.csv"): self.file = file self.code_list = [] self.data = None def read_data(self, file="stock.c...
pd.read_csv(file, encoding="gbk")
pandas.read_csv
import pandas as pd import matplotlib.pyplot as plt import seaborn as sns import numpy as np def pyscript_diseases(): # measels measlesdf = pd.read_csv('https://docs.google.com/spreadsheets/d/1ogMiFRnX-N4lp1cqI0N22F9K9fFVVFfCWxw4T6W2iVw/export?format=csv&id') measlesdf['Total Measles Cases'] = measlesdf...
pd.read_csv("Data/COVID-19.csv")
pandas.read_csv
import os import unittest from builtins import range import matplotlib import mock import numpy as np import pandas as pd import root_numpy from mock import MagicMock, patch, mock_open import six from numpy.testing import assert_array_equal from pandas.util.testing import assert_frame_equal import ROOT from PyAnalysi...
pd.DataFrame({'var1': [1., 2.]})
pandas.DataFrame
# -*- coding: utf-8 -*- """ HornMT_Dataset_Preparation Created on Mon Dec 12 01:25:16 2021 @author: <NAME> """ # Import libs import pandas as pd # Load HornMT dataset file_path = '/data/HornMT.xlsx' HornMT = pd.read_excel(file_path) #HornMT.head(1) # Preprocess the dataframe eng = pd.DataFrame(HornMT...
pd.DataFrame(HornMT['tir'])
pandas.DataFrame
from datetime import timedelta from functools import partial from operator import attrgetter import dateutil import numpy as np import pytest import pytz from pandas._libs.tslibs import OutOfBoundsDatetime, conversion import pandas as pd from pandas import ( DatetimeIndex, Index, Timestamp, date_range, datetime,...
DatetimeIndex(['1-1-2000 00:00:01'])
pandas.DatetimeIndex
''' MIT License Copyright (c) 2020 <NAME> Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distri...
pd.read_csv(fte)
pandas.read_csv
# coding: utf-8 # In[6]: import numpy as np import pandas as pd import seaborn as sns import matplotlib.pyplot as plt # # 权利数据5right.csv提取特征 # 1. 企业拥有权利的个数,RIGHT_CNT # 2. 企业拥有权利类型的个数,RIGHT_TYPE_CNT # 3. 企业拥有权利类型的比例,RIGHT_TYPE_RATE # 4. 第一个获得的权利的类型,RIGHT_FIRST_TYPECODE # 5. 最后一个获得的权利的类型,RIGHT_EN...
pd.Series(row,columns)
pandas.Series
# Copyright (c) 2016. Mount Sinai School of Medicine # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law o...
pandas.Panel(dataframes)
pandas.Panel
import datetime as dt import os import unittest import numpy as np import pandas as pd import devicely class SpacelabsTestCase(unittest.TestCase): READ_PATH = "tests/SpaceLabs_test_data/spacelabs.abp" WRITE_PATH = "tests/SpaceLabs_test_data/spacelabs_written.abp" def __init__(self, *args, **kwargs): ...
pd.to_datetime("1.1.99 17:05:00")
pandas.to_datetime
import os import types import pandas as pd import data_go_kr as dgk def category_from_url(url:str) -> str: return os.path.basename( os.path.dirname(url) ) for k,v in dgk.api.__dict__.items(): if isinstance(v, types.ModuleType): print(k,v) lst_name = [] lst_desc = [] lst_url = [] lst_cat = [] lst_fl...
pd.set_option('display.colheader_justify', 'left')
pandas.set_option
#!/usr/bin/env python3 """tests.compare.compare.py: Auxiliary variables for tests.compare""" import pandas as pd from exfi.io.read_bed import read_bed3 from exfi.io.bed import BED3_COLS, BED3_DTYPES from exfi.compare import \ TP_DF_COLS, TP_DF_DTYPES, \ STATS_COLS, STATS_DTYPES BED3_EMPTY_FN = "tests/compa...
pd.DataFrame(columns=BED3_COLS)
pandas.DataFrame
# -*- coding: utf-8 -*- """ Created on Fri Nov 10 11:56:35 2017 @author: Madhav.L """ import pandas as pd from sklearn import tree from sklearn import model_selection import io import pydot import os os.environ["PATH"] += os.pathsep + 'D:/datascience/graphviz-2.38/release/bin/' #returns current workin...
pd.read_csv("test.csv")
pandas.read_csv
import csv import os import re import numpy as np import pandas as pd from functools import reduce from modules.classes.item_id_parser import ItemIDParser def map_dict(elem, dictionary): if elem in dictionary: return dictionary[elem] else: return np.nan def create_time_feature(series, wind...
pd.concat([df_shard[columns], gender_dummied], axis=1)
pandas.concat
import json from pathlib import Path from itertools import repeat from collections import OrderedDict import pandas as pd import os import warnings def check_input(text_list): """ 检查预测的输入list :param text_list: :return: """ # 一个str的话,转list if isinstance(text_list, str): text_list ...
pd.DataFrame(index=keys, columns=['total', 'counts', 'average'])
pandas.DataFrame
# %% from datetime import datetime, timedelta from pathlib import Path import random import pandas as pd # %% data = pd.read_csv("../data/base2020.csv", sep=";") # %% def report(state, date, last_date, last_state, age, sex): if last_state is not None: events.append(dict( from_state=last_state,...
pd.isna(confirm_date)
pandas.isna
# Author: <NAME>, PhD # # Email: <EMAIL> # # # Ref: https://scikit-learn.org/stable/modules/generated/sklearn.manifold.TSNE.html # Ref: https://docs.scipy.org/doc/scipy/reference/generated/scipy.spatial.distance.jaccard.html#scipy.spatial.distance.jaccard # Ref: https://docs.scipy.org/doc/scipy/reference/generated/scip...
pd.DataFrame ({'ID': ids, 'Dim_1': X_embedded[:,0], 'Dim_2': X_embedded[:,1]})
pandas.DataFrame
from matplotlib import pyplot as plt import matplotlib.ticker as mtick import pandas as pd import json from matplotlib import rcParams # Plots the qualifier statistics for the LC-QuAD 2.0 dataset and Wikidata line_width = 2 font_size = 15 rcParams.update({"figure.autolayout": True}) with open("./results/datasets_s...
pd.DataFrame(panda_data, index=["LC-QuAD 2.0"])
pandas.DataFrame
#!/usr/bin/env python #ADAPTED FROM #https://github.com/bio-ontology-research-group/deepgoplus/blob/master/evaluate_deepgoplus.py import numpy as np import pandas as pd import click as ck from sklearn.metrics import classification_report from sklearn.metrics.pairwise import cosine_similarity import sys from collection...
pd.read_pickle(terms_file)
pandas.read_pickle
""" This module contains transformers that apply string functions. """ import pandas as pd from tubular.base import BaseTransformer class SeriesStrMethodTransformer(BaseTransformer): """Tranformer that applies a pandas.Series.str method. Transformer assigns the output of the method to a new column. It is p...
pd.Series(["a"])
pandas.Series
import pandas as pd import numpy as np import matplotlib.pyplot as plt INPUT_DIR = "~/data/query-result/" OUTPUT_DIR = "~/data/summary-stats/" RES_LIST = ['cpu', 'mem', 'net_send', 'net_receive', 'disk_read', 'disk_write'] METRIC_LIST = ['_util_per_instance_95p', '_util_per_instance_max', '_util_per_pool', '_util_per_...
pd.DataFrame()
pandas.DataFrame
import pandas as pd from geopy.distance import great_circle #filter and drop duplcates rows data = pd.read_csv("AllStudent.csv") data = data.filter(['N°Ins','Adresse']).drop_duplicates() data.fillna("other", inplace=True)
pd.DataFrame(data)
pandas.DataFrame
import unittest import numpy as np import pandas as pd from pyalink.alink import * class TestPinyi(unittest.TestCase): def run_segment(self): # -*- coding=UTF-8 -*- data = np.array([ [0, u'二手旧书:医学电磁成像'], [1, u'二手美国文学选读( 下册 )李宜燮南开大学出版社 9787310003969'], [2, u'...
pd.DataFrame({"id": data[:, 0], "text": data[:, 1]})
pandas.DataFrame