prompt stringlengths 19 1.03M | completion stringlengths 4 2.12k | api stringlengths 8 90 |
|---|---|---|
import pandas as pd
import numpy as np
import holidays
import statsmodels.formula.api as sm
import time
from Helper import helper
import datetime
class DR(object):
def __init__(self, dataframe):
df = dataframe.copy()
self.lm_data = helper.DR_Temp_data_cleaning(df)
self.name = 'DR'
de... | pd.read_csv(path) | pandas.read_csv |
import json, os, sys
import pandas as pd
from urllib.request import urlopen
from xml.dom import minidom
from json import load
from pandas.io.json import json_normalize
def filterIQM(apidf, filter_list):
""" Loads the API info and filters based on user-provided
parameters. Filter parameters should be a l... | pd.concat([userdf,filtered_apidf], sort=True) | pandas.concat |
"""
"""
"""
>>> # ---
>>> # SETUP
>>> # ---
>>> import os
>>> import logging
>>> logger = logging.getLogger('PT3S.Rm')
>>> # ---
>>> # path
>>> # ---
>>> if __name__ == "__main__":
... try:
... dummy=__file__
... logger.debug("{0:s}{1:s}{2:s}".format('DOCTEST: __main__ Context: ','path = os.p... | pd.merge(vWBLZ_vKNOT,pFWVB,left_on='NAME_y',right_on='NAME_i') | pandas.merge |
import os
import lightgbm as lgb
import neptune
from neptunecontrib.monitoring.lightgbm import neptune_monitor
from neptunecontrib.versioning.data import log_data_version
from neptunecontrib.api.utils import get_filepaths
from neptunecontrib.monitoring.reporting import send_binary_classification_report
from neptunecon... | pd.merge(submission, test, on='TransactionID') | pandas.merge |
import pandas as pd
from multiprocessing import Pool
import logging
from src.helper import create_logger
# Create the logger object
logger = create_logger('Parser', 'logs/Hedging.log',
logging.DEBUG, logging.WARNING)
class Parser():
"""
Parser class that calls parent and divide the dat... | pd.DataFrame() | pandas.DataFrame |
__author__ = "unknow"
__copyright__ = "Sprace.org.br"
__version__ = "1.0.0"
import pandas as pd
import sys
from math import sqrt
import sys
import os
import ntpath
import scipy.stats
import seaborn as sns
from matplotlib import pyplot as plt
#sys.path.append('/home/silvio/git/track-ml-1/utils')
#sys.path.append('.... | pd.DataFrame() | pandas.DataFrame |
import pandas as pd
import numpy as np
import argparse
labels = [
"0",
"B-answer",
"I-answer",
]
def find_answer_start(answer, sent):
answer = [x.lower() for x in answer]
sent = [x.lower() for x in sent]
for idx, word in enumerate(sent):
if answer[0] in word:
is_match = Tru... | pd.read_pickle(args.data_path) | pandas.read_pickle |
import sys
import numpy as np
import pandas as pd
import wgdi.base as base
class karyotype_mapping():
def __init__(self, options):
self.position = 'order'
self.limit_length = 5
for k, v in options:
setattr(self, str(k), v)
print(str(k), ' = ', v)
def karyoty... | pd.read_csv(self.blockinfo, index_col='id') | pandas.read_csv |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
import sys, os, platform, copy
import logging, re
import pandas as pd
import numpy as np
import itertools
import time, random
from tqdm import tqdm
tqdm.pandas()
from conf import getworkdir, conf
from models import run_model, CoxNMF_initialization
from utils import feat... | pd.MultiIndex.from_arrays(columns) | pandas.MultiIndex.from_arrays |
import datetime
import integrationutils as ius
import numpy as np
import os
import pathlib
import pandas as pd
import sqlite3
import sys
import warnings
''' Direct questions and concerns regarding this script to <NAME>
<EMAIL>
'''
def find_abund_col(df):
clist = []
for c in df.columns:
... | pd.read_csv('WorkflowInputFile.txt',delimiter='\t') | pandas.read_csv |
import pandas as pd
import numpy as np
import sys
from tabulate import tabulate
import matplotlib.pyplot as plt
import seaborn as sns
from sklearn.preprocessing import StandardScaler
import matplotlib.ticker as ticker
# from pyutils import *
# import dtale
# dtale.show(df)
# from pandas_profiling import ProfileRepo... | pd.value_counts(s) | pandas.value_counts |
#!/usr/bin/env python
# -*- coding:utf-8 -*-
"""
Date: 2021/8/27 15:59
Desc: REITs 行情及信息
http://quote.eastmoney.com/center/gridlist.html#fund_reits_all
https://www.jisilu.cn/data/cnreits/#CnReits
"""
import pandas as pd
import requests
def reits_realtime_em() -> pd.DataFrame:
"""
东方财富网-行情中心-REITs-沪深 REITs
... | o_numeric(temp_df['涨幅']) | pandas.to_numeric |
# LIBRARIES
# set up backend for ssh -x11 figures
import matplotlib
matplotlib.use('Agg')
# read and write
import os
import sys
import glob
import re
import fnmatch
import csv
import shutil
from datetime import datetime
# maths
import numpy as np
import pandas as pd
import math
import random
# miscellaneous
import ... | pd.DataFrame({'version': versions, 'R2': r2s}) | pandas.DataFrame |
import time
from collections import defaultdict
from datetime import timedelta
import cvxpy as cp
import empiricalutilities as eu
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
from tqdm import tqdm
from transfer_entropy import TransferEntropy
plt.style.use('fivethirtyei... | pd.DataFrame({'returns': returns, 'vol': vols, 'ete': ete_out}) | pandas.DataFrame |
#Copyright 2021 <NAME>, <NAME>, <NAME>
#
#Licensed under the Apache License, Version 2.0 (the "License");
#you may not use this file except in compliance with the License.
#You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
#Unless required by applicable law or agreed to in writ... | is_numeric_dtype(dat[col]) | pandas.api.types.is_numeric_dtype |
########################################################
# <NAME> - drigols #
# Last update: 21/09/2021 #
########################################################
def OLS(dic):
import pandas as pd
df = | pd.DataFrame(dic) | pandas.DataFrame |
import numpy as np
import pandas as pd
from pyopenms import FeatureMap, FeatureXMLFile
def extractNamesAndIntensities(feature_dir, sample_names, database):
"""
This function takes .featureXML files, the output of SmartPeak
pre-processing, and extracts the metabolite's reference and its measured
intens... | pd.DataFrame.from_dict(extracted_data_dict, "index") | pandas.DataFrame.from_dict |
import warnings
import os
import numpy as np
import pandas as pd
from preprocessing.utils import remove_french_accents_and_cedillas_from_dataframe
columns_to_drop = ['nr', 'patient_id', 'eds_end_4digit', 'eds_manual', 'DOB', 'begin_date',
'end_date', 'death_date', 'death_hosp', 'eds_final_id',
... | pd.to_numeric(equalized_reorganised_lab_df['value'], errors='coerce') | pandas.to_numeric |
import pytest
import os
from mapping import util
from pandas.util.testing import assert_frame_equal, assert_series_equal
import pandas as pd
from pandas import Timestamp as TS
import numpy as np
@pytest.fixture
def price_files():
cdir = os.path.dirname(__file__)
path = os.path.join(cdir, 'data/')
files = ... | TS('2015-01-04') | pandas.Timestamp |
"""Electric grid models module."""
import cvxpy as cp
import itertools
from multimethod import multimethod
import natsort
import numpy as np
import opendssdirect
import pandas as pd
import scipy.sparse as sp
import scipy.sparse.linalg
import typing
import mesmo.config
import mesmo.data_interface
import mesmo.utils
l... | pd.DataFrame(columns=nodes, index=self.electric_grid_model.timesteps, dtype=float) | pandas.DataFrame |
import os
import argparse
import numpy as np
import pandas as pd
import nibabel as nib
from ukbb_cardiac.common.cardiac_utils import get_frames
from ukbb_cardiac.common.image_utils import np_categorical_dice
if __name__ == '__main__':
parser = argparse.ArgumentParser()
parser.add_argument('--output_c... | pd.DataFrame(init) | pandas.DataFrame |
# --------------
# import the libraries
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
from sklearn.model_selection import train_test_split
import warnings
warnings.filterwarnings('ignore')
# Code starts here
df = pd.read_json(path,lines=True)
df.columns=df.columns.str.st... | pd.get_dummies(data=X_test,columns=["category", "cup_size","length"],prefix=["category", "cup_size","length"]) | pandas.get_dummies |
"""
PRESSGRAPHS DASH CLIENT
WEB GUI interface for PressGraphs WebAPI
"""
###################################
# IMPORTS
###################################
#builtins
from datetime import datetime
from datetime import timedelta
#3rd party
import dash
import dash_core_components as dcc
import dash_html_components as html... | pd.DataFrame(s_2_content) | pandas.DataFrame |
from datetime import timedelta
from functools import partial
from operator import attrgetter
import dateutil
import numpy as np
import pytest
import pytz
from pandas._libs.tslibs import OutOfBoundsDatetime, conversion
import pandas as pd
from pandas import (
DatetimeIndex, Index, Timestamp, date_range, datetime,... | date_range(freq='D', start=start, end=end, tz=tz) | pandas.date_range |
import requests
import pandas as pd
import ftplib
import io
import re
import json
import datetime
try:
from requests_html import HTMLSession
except Exception:
print("""Warning - Certain functionality
requires requests_html, which is not installed.
Install ... | pd.Timestamp(start_date) | pandas.Timestamp |
import fire
from rest_api_asyncio import UniprotClient, get_db
import pandas as pd
from pandas import DataFrame
import gtfparse
import time
from tqdm import tqdm
from pathlib import Path
from functools import reduce
import glob
import sys
import urllib3
import asyncio
OUT_HEADER_BASE = [
'gene_id',
'gene_name... | DataFrame([None], columns=['protein_existence']) | pandas.DataFrame |
# -*- coding: utf-8 -*-
"""
@author: <NAME> <<EMAIL>>
Date: Oct 2019
"""
import numpy as np
import pandas as pd
from .pipe import Pipe
from .. import precomp_funs as _pf
class CHPPlant(Pipe):
"""
Construct a CHP plant.
Can be added to the simulation environment by using the following method:
.a... | pd.DataFrame(data=self.res_dQ[:array_length, ...], index=hdf_idx) | pandas.DataFrame |
import pandas as pd
from recalibrate.unarycalibration.singelsystematiccalibration import single_systematic_calibration
from pprint import pprint
from sklearn.metrics import brier_score_loss
from xgboost import XGBClassifier
# Illustrates calibration of a single set of model probabilities (user selecting a product)
if... | pd.read_csv('https://raw.githubusercontent.com/microprediction/recalibrate/main/examples/default_data/default.csv') | pandas.read_csv |
"""Step 1: Solving the problem in a deterministic manner."""
import cvxpy as cp
import fledge
import numpy as np
import os
import pandas as pd
import plotly.express as px
import plotly.graph_objects as go
import shutil
def main():
# Settings.
scenario_name = 'course_project_step_1'
results_path = os.pat... | pd.DataFrame(0.0, index=der_model_set.timesteps, columns=der_model_set.outputs) | pandas.DataFrame |
import numpy as np
import pandas as pd
import math
from elopackage.elo import Elo
from elopackage.player import Player
class ResultsTable:
def __init__(self, df):
"""
df - pd DataFrame of tournamenent results
"""
# self.df = df.sort_values(by='match_date_dt', ascending=True)
... | pd.concat([df_unique, df_tmp]) | pandas.concat |
import os, sys, platform, json, operator, multiprocessing, io, random, itertools, warnings, h5py, \
statistics, inspect, requests, validators, math, time, pprint, datetime, importlib, fsspec, scipy
# Python utils.
from textwrap import dedent
# External utils.
from tqdm import tqdm #progress bar.
from natsort import na... | pd.DataFrame(data=ndarray) | pandas.DataFrame |
#!/usr/bin/env python
# coding: utf-8
# # Avatar : The Last Airbender
# ### Machine Learning and Analysis of the show
# In[1]:
from IPython.display import Image
Image (filename = "images (1).jpg")
# ## Introduction :
#
# **Avatar: The Last Airbender (Avatar: The Legend of Aang in some regions)** is an American ... | pd.read_csv('series_names.csv') | pandas.read_csv |
# Copyright (c) 2019 Uber Technologies, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed... | pd.DataFrame(sig_errs) | pandas.DataFrame |
from sklearn.model_selection import train_test_split
import numpy as np
import pandas as pd
import sklearn
import json
from sklearn.preprocessing import LabelEncoder, OneHotEncoder
from sklearn.preprocessing import StandardScaler
from sklearn.metrics import confusion_matrix
from sklearn.linear_model import LogisticRegr... | pd.read_csv('./public/Python_Scripts/Dataset.csv', header=None) | pandas.read_csv |
import csv
from io import StringIO
import os
import numpy as np
import pytest
from pandas.errors import ParserError
import pandas as pd
from pandas import (
DataFrame,
Index,
MultiIndex,
NaT,
Series,
Timestamp,
date_range,
read_csv,
to_datetime,
)
import pandas._testing as tm
impo... | read_csv(path, index_col=0, encoding="UTF-8") | pandas.read_csv |
import numpy as np
import xarray as xr
import pandas as pd
import os
from collections import OrderedDict
# from astropy.time import Time
import logging
import copy
from typing import List, Dict, Union, Tuple
import pysagereader
class SAGEIILoaderV700(object):
"""
Class designed to load the v7.00 SAGE II spec ... | pd.Timestamp('1858-11-17') | pandas.Timestamp |
import sys
import pandas as pd
import numpy as np
import json
import os
from datetime import date
from scipy.stats import linregress
import yaml
from momentum_data import cfg
DIR = os.path.dirname(os.path.realpath(__file__))
pd.set_option('display.max_rows', None)
pd.set_option('display.width', None)
pd.set_option('d... | pd.Series(closes[-slope_days:]) | pandas.Series |
# kaggleのSMS Spam Collection Datasetでナイーブベイズを体験する
# コード:https://qiita.com/fujin/items/50fe0e0227ef8457a473
import matplotlib.pyplot as pyplot
import numpy as np
import pandas as pd
from sklearn.model_selection import train_test_split
from sklearn.feature_extraction.text import CountVectorizer
from sklearn.naiv... | pd.read_csv("./datasets/spam.csv", encoding="latin-1") | pandas.read_csv |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
from src.tasks.preprocessing_funcs import load_dataloaders
from src.tasks.trainer import train_and_fit
from src.tasks.infer import infer_from_trained
import logging
from argparse import ArgumentParser
from src.tasks.visualization import Graph
from src.tasks.pdf_to_txt imp... | pd.DataFrame(input_sents) | pandas.DataFrame |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import pandas as pd
import covsirphy as cs
def md(scenario, filename, name=None):
with open(filename, "w") as fh:
fh.write(scenario.summary(name=name).to_markdown())
def main():
print(cs.__version__)
# Data loading
data_loader = cs.DataLoader("i... | pd.DataFrame.from_dict(opt_dict, orient="index") | pandas.DataFrame.from_dict |
import re
import pandas as pd
import numpy as np
class Resampler(object):
"""Resamples time-series data from one frequency to another frequency.
"""
min_in_freqs = {
'MIN': 1,
'MINUTE': 1,
'DAILY': 1440,
'D': 1440,
'HOURLY': 60,
'HOUR': 60,
'H': 60,... | pd.infer_freq(idx) | pandas.infer_freq |
import re
import time
import argparse
import numpy as np
import pandas as pd
import util as ut
from collections import Counter
from collections import defaultdict
from sklearn.metrics.pairwise import cosine_similarity
from sklearn.feature_extraction.text import TfidfVectorizer
from sklearn.neighbors import NearestNeigh... | pd.read_csv(fname) | pandas.read_csv |
import time
import sorting
import bst
import timeit
import platform
import random
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import heapq
import copy
random.seed(521)
DEFAULT_NUMBER = 100000 # 100k
DEFAULT_POPULATION = range(100000) # 1m
DEFAULT_SIZES = [10, 100, 1000, 10000, 100000]
de... | pd.DataFrame.from_dict(self.test_result) | pandas.DataFrame.from_dict |
r"""
Baseline Calculation
"""
# Standard Library imports
import argparse
import cartopy.crs as ccrs
import datetime
import h5py
import json
import matplotlib.colors
import matplotlib.dates as mdates
import matplotlib.pyplot as plt
import netCDF4
import numpy as np
import os
import pandas as pd
impor... | pd.to_datetime("1990-01") | pandas.to_datetime |
# this class is aimed at generating data for the redundant and noisy contexts
import os
import numpy as np
import pandas as pd
from os.path import join
import random
from runs.experiments import Experiment
def generate_tr_vl_ts_splits(id,
source_path,
split_pa... | pd.concat([train_sentences.loc[p_], train_sentences.loc[n_]]) | pandas.concat |
import argparse
import json
import os
import re
from glob import glob
import matplotlib.pyplot as plt
import pandas as pd
import seaborn as sns
sns.set_style("dark")
def parse_args():
description = """Measure how you spend time by tracking key presses and
changes to window focus or title."""
p = argpars... | pd.Grouper(key="time", freq=freq) | pandas.Grouper |
from __future__ import print_function
import numpy as np
import time, os, sys
import matplotlib.pyplot as plt
from scipy import ndimage as ndi
from skimage import color, feature, filters, io, measure, morphology, segmentation, img_as_ubyte, transform
import warnings
import math
import pandas as pd
import argparse
impor... | pd.DataFrame() | pandas.DataFrame |
#Usage
# import sys
# sys.path.insert(0,'path to this file')
# import functions as f
import pickle
import pandas as pd
import os
import sys
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
from keras.preprocessing.text import Tokenizer
from keras.preprocessing.sequence import pad_s... | pd.read_pickle("C:/Users/nik00/py/proj/hyppi-independent.pkl") | pandas.read_pickle |
"""
The aim of this project was to build a classifier on the titanic kaggle dataset.
"""
### import libraries
import numpy as np
import pandas as pd
import matplotlib
matplotlib.use('Pdf')
import matplotlib.pyplot as plt
# import data preprocessing modules
from sklearn.preprocessing import Imputer
from sklearn.prepro... | pd.read_csv("test.csv") | pandas.read_csv |
#! /usr/bin/env python3
"""
Model Checker Collection for the Model Checking Contest.
"""
import argparse
import hashlib
import math
import logging
import os
import random
import statistics
# import getpass
import json
import pathlib
import pickle
import platform
import re
import sys
import tempfile
import tarfile
imp... | pandas.DataFrame([test]) | pandas.DataFrame |
import logging
import pandas as pd
from scipy.cluster.vq import kmeans2
from django.http import JsonResponse
from django.views import View
from . import forms
from . import models
logger = logging.getLogger(__name__)
class LocationListAPI(View):
def get(self, request, *args, **kwargs):
"""Summarize... | pd.concat(result, axis=1) | pandas.concat |
from IPython.display import display
import pandas as pd
import pyomo.environ as pe
import numpy as np
import csv
import os
import shutil
class inosys:
def __init__(self, inp_folder, ref_bus, dshed_cost = 1000000, rshed_cost = 500, phase = 3, vmin=0.85, vmax=1.15, sbase = 1, sc_fa = 1):
'''
... | pd.read_csv(inp_folder + os.sep + 'qrep_dist.csv') | pandas.read_csv |
from unittest.mock import ANY, MagicMock, patch
import pytest
import pandas as pd
from pandas._testing import assert_frame_equal
from muttlib.dbconn.base import BaseClient, EngineBaseClient
@pytest.fixture
def engine_baseClient():
client = EngineBaseClient(
database="database",
host="host",
... | pd.DataFrame({'col1': ['1'], 'col2': ['3.0']}) | pandas.DataFrame |
import pathlib
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
from prophet import Prophet
from sklearn import metrics
def get_prophet_data(stock_path):
with open(stock_path, 'r', encoding='utf-8') as f:
df = pd.read_json(f.read(), orient='records')
print(df)
# rename
... | pd.set_option('display.max_columns', None) | pandas.set_option |
from pandas import (
DataFrame,
Series,
)
import pandas._testing as tm
class TestToFrame:
def test_to_frame(self, datetime_series):
datetime_series.name = None
rs = datetime_series.to_frame()
xp = DataFrame(datetime_series.values, index=datetime_series.index)
tm... | tm.assert_frame_equal(result, expected) | pandas._testing.assert_frame_equal |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
__author__ = '<NAME>'
from pathlib import Path
import numpy as np
import pandas as pd
from sklearn.feature_extraction.text import CountVectorizer
from sklearn.model_selection import train_test_split
from gensim.models import LdaModel
from gensim.matutils import Sparse2Corp... | pd.DataFrame() | pandas.DataFrame |
from typing import Union, Optional
import pytest
import scanpy as sc
import cellrank.external as cre
from anndata import AnnData
from cellrank.tl.kernels import ConnectivityKernel
from cellrank.external.kernels._utils import MarkerGenes
from cellrank.external.kernels._wot_kernel import LastTimePoint
import numpy as ... | pd.Series(terminal_states) | pandas.Series |
#!/usr/bin/env python
"""Tests for `pubchem_api` package."""
import os
import numpy as np
import pandas as pd
import scipy
from scipy.spatial import distance
import unittest
# from click.testing import CliRunner
# from structure_prediction import cli
class TestDataPreprocessing(unittest.TestCase):
"""Tests for ... | pd.DataFrame(make_square) | pandas.DataFrame |
import numpy as np
import pandas as pd
import os
import pickle
from sklearn.model_selection import KFold
from sklearn.metrics import precision_recall_curve
import sklearn.metrics as metrics
from model import lightgbm_train
from glob import glob
from utils import *
import shap
from collections import defaultdict
def l... | pd.DataFrame.from_dict(t_shap) | pandas.DataFrame.from_dict |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
# Copyright 1999-2017 Alibaba Group Holding Ltd.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENS... | pd.DataFrame([['Col1', 1], ['Col2', 2]], columns=['Field1', 'Field2']) | pandas.DataFrame |
"""
Also test support for datetime64[ns] in Series / DataFrame
"""
from datetime import datetime, timedelta
import re
import numpy as np
import pytest
from pandas._libs import iNaT
import pandas._libs.index as _index
import pandas as pd
from pandas import DataFrame, DatetimeIndex, NaT, Series, Timestamp, date_range
... | tm.assert_series_equal(cp, expected) | pandas._testing.assert_series_equal |
# Copyright 2021 Fedlearn authors.
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writi... | pandas.merge(uid, g2.loc[:, ["uid"]], on="uid", how="inner") | pandas.merge |
import numpy as np
import pandas as pd
from numba import njit
import pytest
from vectorbt import defaults
from vectorbt.utils import checks, config, decorators, math, array
from tests.utils import hash
# ############# config.py ############# #
class TestConfig:
def test_config(self):
conf = config.Conf... | pd.Series([1, 2, 3], index=index) | pandas.Series |
import pandas
import os
import re
import numpy as np
import math
import warnings
from modin.error_message import ErrorMessage
from modin.engines.base.io import BaseIO
from modin.data_management.utils import compute_chunksize
from modin import __execution_engine__
if __execution_engine__ == "Ray":
import ray
PQ_... | pandas.DataFrame(index=index) | pandas.DataFrame |
import numpy as np
import pandas as pd
import pytest
from sklearn import metrics
from epiquark import ScoreCalculator
def test_non_case_imputation(shared_datadir, paper_example_score: ScoreCalculator) -> None:
cases = pd.read_csv(shared_datadir / "paper_example/cases_long.csv")
imputed = paper_example_score.... | pd.read_csv(shared_datadir / "paper_example/p_hat_di.csv") | pandas.read_csv |
# -*- coding: utf-8 -*-
"""
Created on Tue Sep 14 10:59:05 2021
@author: franc
"""
import os
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
from pathlib import Path
import json
from collections import Counter, OrderedDict
import math
import torchtext
from torchtext.data import get_tokenizer
... | pd.DataFrame({'spanish': ["orca"], 'english': ["killer_whale"]}) | pandas.DataFrame |
import uuid
import traceback
import os
import numpy as np
import pandas
import nrrd
import glob
import argparse
import random
from PIL import Image
import csv
from shutil import rmtree
from collections import defaultdict
from keras.preprocessing.image import ImageDataGenerator, Iterator
from keras.utils import to_categ... | pandas.concat(all_train) | pandas.concat |
"""
Created on Thu Jan 26 17:04:11 2017
@author: <NAME>, <EMAIL>
"""
#%matplotlib inline
import numpy as np
import pandas as pd
import dicom
import os
import scipy.ndimage as ndimage
import matplotlib.pyplot as plt
import scipy.ndimage # added for scaling
import cv2
import time
import glob
from skimage import me... | pd.read_csv(LUNA_ANNOTATIONS) | pandas.read_csv |
"""This module is meant to contain the Solscan class"""
from messari.dataloader import DataLoader
from messari.utils import validate_input
from string import Template
from typing import Union, List, Dict
from .helpers import unpack_dataframe_of_dicts
import pandas as pd
#### Block
BLOCK_LAST_URL = 'https://public-api... | pd.concat(df_list, keys=accounts, axis=1) | pandas.concat |
# TODO(*): Move to ib/medata and rename contract_metadata.py
import logging
import os
from typing import List
import ib_insync
import pandas as pd
import helpers.io_ as hio
import im.ib.data.extract.gateway.utils as videgu
_LOG = logging.getLogger(__name__)
class IbMetadata:
def __init__(self, file_name: str) ... | pd.concat(dfs, axis=0) | pandas.concat |
import pandas as pd
import numpy as np
import unittest
from dstools.preprocessing.Bucketizer import Bucketizer
class TestBucketizer(unittest.TestCase):
def compare_DataFrame(self, df_transformed, df_transformed_correct):
"""
helper function to compare the values of the transformed DataFrame with ... | pd.DataFrame({'x':[1,2,3]}) | pandas.DataFrame |
import numpy as np
import pandas as pd
from scipy import interpolate
import pickle # to serialise objects
from scipy import stats
import seaborn as sns
from sklearn import metrics
from sklearn.model_selection import train_test_split
sns.set(style='whitegrid', palette='muted', font_scale=1.5)
RANDOM_SEED = 42
dataset... | pd.DataFrame(training_set) | pandas.DataFrame |
# -*- coding: utf-8 -*-
"""
Created on Mon Jan 25 16:14:45 2021
@author: bdobson
"""
import os
import pandas as pd
import geopandas as gpd
from matplotlib import pyplot as plt
root = os.path.join("C:\\", "Users", "bdobson", "Documents", "GitHub", "cwsd_sewer","data")
catchment = "cranbrook"
cluster = 'cluster_Louv_... | pd.read_parquet(node_fid) | pandas.read_parquet |
import pandas as pd
import argparse
from matplotlib_venn import venn3
import matplotlib.pyplot as plt
import math
def get_args():
desc = 'Given sj files, see which splice junctions are shared/unique between datasets'
parser = argparse.ArgumentParser(description=desc)
parser.add_argument('-sj_1', dest='sj_1',
h... | pd.merge(dfa, dfc, how='inner', on=['chrom', 'start', 'stop', 'strand']) | pandas.merge |
#!/usr/bin/env python3
#libraries
import pandas as pd
import numpy as np
import re
import os
pd.set_option('display.max_rows',200)
pd.set_option('display.max_columns',200)
import matplotlib.pyplot as plt
import seaborn as sns
import pymysql
from sqlalchemy import create_engine
from sqlalchemy.orm import sessionmaker
f... | pd.merge(billing_avg,uspa_prod_dataframe, left_on = 'mobile',right_on='mobile_x',how = 'left') | pandas.merge |
# Author: <NAME> <<EMAIL>>
#
# License: BSD (3-clause)
import os.path as op
import pandas as pd
import numpy as np
import mne
from mne.transforms import apply_trans, _get_trans
from mne.utils import _validate_type, _check_fname
from mne.io import BaseRaw
def _read_fold_xls(fname, atlas="Juelich"):
"""Read fOLD... | pd.DataFrame() | pandas.DataFrame |
# -*- coding: utf-8 -*-
"""dlw9383-bandofthehawk-output.ipynb
Automatically generated by Colaboratory.
Original file is located at
https://colab.research.google.com/drive/154b5GvPxORu_mhpHDIsNlWvyBxMIwEw2
"""
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
import seaborn as sns
sns.set(rc=... | pd.DataFrame(y_kmeans, columns=['Clusters']) | pandas.DataFrame |
import pandas as pd
import plotly.express as px
# Passo 1 -Importar a basa de dados para o python
tabela = | pd.read_csv(r"C:\Users\jose_\OneDrive\Documentos\Estudos\arquivos_pyton\telecom_users.csv") | pandas.read_csv |
from datetime import (
datetime,
timedelta,
timezone,
)
import numpy as np
import pytest
import pytz
from pandas import (
Categorical,
DataFrame,
DatetimeIndex,
NaT,
Period,
Series,
Timedelta,
Timestamp,
date_range,
isna,
)
import pandas._testing as tm
class TestS... | isna(ser) | pandas.isna |
from typing import List
import pandas as pd
# Not covered: Essentially a script over other unit tested functions
def clean_data(df: pd.DataFrame) -> pd.DataFrame: # pragma: no cover
"""Parse
* location information to append city, state, zip code, and neighborhood columns
* salary information to append m... | pd.merge(df, df_expanded_salary, how="outer", on="link") | pandas.merge |
"""
Also test support for datetime64[ns] in Series / DataFrame
"""
from datetime import datetime, timedelta
import re
import numpy as np
import pytest
from pandas._libs import iNaT
import pandas._libs.index as _index
import pandas as pd
from pandas import DataFrame, DatetimeIndex, NaT, Series, Timestamp, date_range
... | tm.assert_series_equal(result, expected) | pandas._testing.assert_series_equal |
import numpy as np
np.random.seed(1337) # for reproducibility
import pandas as pd
from sklearn.model_selection import train_test_split
from sklearn.metrics._regression import r2_score, mean_squared_error
from sklearn.preprocessing import MinMaxScaler
from sklearn.metrics import accuracy_score
from dbn import Superv... | pd.to_datetime(df['Date']) | pandas.to_datetime |
import pandas
from copy import deepcopy
from palm.base.target_data import TargetData
from palm.discrete_state_trajectory import DiscreteStateTrajectory,\
DiscreteDwellSegment
class BlinkTargetData(TargetData):
"""
A dwell trajectory loaded from a file. The trajectory
... | pandas.read_csv(data_file, header=0) | pandas.read_csv |
import unittest
import pandas as pd
import numpy as np
from ..timeseries import TimeSeries
class TimeSeriesTestCase(unittest.TestCase):
times = pd.date_range('20130101', '20130110')
pd_series1 = pd.Series(range(10), index=times)
pd_series2 = pd.Series(range(5, 15), index=times)
pd_series3 = pd.Serie... | pd.Timestamp('20130107') | pandas.Timestamp |
#python imports
import os
import gc
import string
import random
import time
import pickle
import shutil
from datetime import datetime
#internal imports
from modules.Signal import Signal
from modules.Database import Database
from modules.Predictor import Classifier, ComplexBuilder
from modules.utils import calcula... | pd.DataFrame(out,index=combinedPeakModelsFiltered.index, columns = quantColumnNames) | pandas.DataFrame |
from __future__ import division
from datetime import datetime
import sys
if sys.version_info < (3, 3):
import mock
else:
from unittest import mock
import pandas as pd
import numpy as np
import random
from nose.tools import assert_almost_equal as aae
import bt
import bt.algos as algos
def test_algo_name():... | pd.DateOffset(months=3) | pandas.DateOffset |
#!/usr/bin/python -u
# +
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
import random
import os
SEED = 123
random.seed(SEED)
np.random.seed(SEED)
# +
#Load the train and test files with pchembl values (used for end-to-end deep learning)
train_with_label_df = pd.read_csv("../data/Train_Compound_... | pd.read_csv("../data/Train_Protein_LS.csv",header=None) | pandas.read_csv |
import os
from src.corpus.brat_writer import write_file
from typing import Dict, List
import pandas as pd
class DocumentMerger:
def __init__(
self,
ent_id2label,
rel_id2label,
true_doc_tokens: Dict[str, List[List[str]]],
save_dir="val"
) -> None:
super().__init... | pd.DataFrame() | pandas.DataFrame |
# author: <NAME>, <NAME>
# date: 2021-11-25
"""This script takes two file paths. It takes in the input path which includes the clean train and test data
and the output directory to store the results in. It performs machine learning analysis.
This script will have 4 outputs: 3 tables and 1 figure.
Usage: src/machine_... | pd.Series(data=out_col, index=mean_scores.index) | pandas.Series |
import numpy as np
import pandas as pd
import math
import matplotlib.pyplot as plt
import yfinance as yf
from pandas_datareader import data as web
import datetime as dt
from empyrical import*
import quantstats as qs
from darts.models import*
from darts import TimeSeries
from darts.utils.missing_values import... | pd.DataFrame(ret_data) | pandas.DataFrame |
import sys
import csv
import numpy as np
import gpflow
import os
os.environ["CUDA_DEVICE_ORDER"]="PCI_BUS_ID"
os.environ["CUDA_VISIBLE_DEVICES"]="0"
import pandas as pd
import h5py
from sklearn.model_selection import train_test_split
import tensorflow as tf
from scipy.cluster.vq import kmeans
tf.set_random_seed(1234)
i... | pd.DataFrame(targets) | pandas.DataFrame |
import pytest
import numpy as np
import pandas as pd
from six import StringIO
from dae.tools.generate_histogram import (
ScoreHistogramInfo,
GenerateScoresHistograms,
)
# pytestmark = pytest.mark.xfail
class MyStringIO(StringIO):
def __add__(self, other):
return ""
@pytest.fixture
def score_fi... | pd.DataFrame({"RANKSCORE_0": [10000], "start": [100], "end": [100]}) | pandas.DataFrame |
# -*- coding: utf-8 -*-
"""data_augmentation.py
# Notebook: Generate Training Dataset
In this Notebook, we want to simulate training dataset from the real world dataset. There are two steps in making such data:
* 1) Create pair of trajectories from the original set
* 2) Create label per pair of trajectories
# Requir... | pd.concat(data_list) | pandas.concat |
import numpy as np
import pandas as pd
import os
# http://archive.ics.uci.edu/ml/datasets/Statlog+%28German+Credit+Data%29
# Load .csv file
path = 'german/german_final.csv'
data = | pd.read_csv(path, header=None) | pandas.read_csv |
# coding: utf-8
"""
Aurelio_Amerio_Higgs_v4.py
In this analysis, I have used several MLP models, applied to the Kaggle Higgs dataset,
in order to distinguish signal from noise.
----------------------------------------------------------------------
author: <NAME> (<EMAIL>)
Student ID: QT08313
... | pd.read_csv(train_sig_path_sml, header=0) | pandas.read_csv |
# -*- coding: utf-8 -*-
import numpy as np
import pytest
from numpy.random import RandomState
from numpy import nan
from datetime import datetime
from itertools import permutations
from pandas import (Series, Categorical, CategoricalIndex,
Timestamp, DatetimeIndex, Index, IntervalIndex)
import pan... | algos.value_counts(s, bins=1) | pandas.core.algorithms.value_counts |
import numpy as np
import pandas as pd
import pytest
from sid.shared import boolean_choices
from src.create_initial_states.create_initial_immunity import (
_calculate_endog_immunity_prob,
)
from src.create_initial_states.create_initial_immunity import (
_calculate_exog_immunity_prob,
)
from src.create_initial_... | pd.DataFrame() | pandas.DataFrame |
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import pytest
import numpy as np
import pandas
from modin.pandas.utils import to_pandas
import modin.pandas as pd
from pathlib import Path
import pyarrow as pa
import os
import sys
from .utils import df_equals... | pandas.read_clipboard() | pandas.read_clipboard |
import os
import sys
import numpy as np
import pytest
import pandas as pd
from pandas import DataFrame, compat
from pandas.util import testing as tm
class TestToCSV:
@pytest.mark.xfail((3, 6, 5) > sys.version_info >= (3, 5),
reason=("Python csv library bug "
... | pd.Series([1], ind, name="data") | pandas.Series |
import requests
from bs4 import BeautifulSoup
import pandas as pd
def get_tables(urls, link=False):
"""Returns a dataframes list with the tables of the different groups.
Keyword arguments:
urls -- list with urls of the different groups
link -- indicates whether you want to include the url of ... | pd.concat(tables_list, axis=0) | pandas.concat |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.