prompt stringlengths 19 1.03M | completion stringlengths 4 2.12k | api stringlengths 8 90 |
|---|---|---|
#GiG
import numpy as np
import pandas as pd
from pathlib import Path
from deep_blocker import DeepBlocker
from tuple_embedding_models import AutoEncoderTupleEmbedding, CTTTupleEmbedding, HybridTupleEmbedding, SIFEmbedding
from vector_pairing_models import ExactTopKVectorPairing
import blocking_utils
from configurati... | pd.DataFrame(lines,columns=['full']) | pandas.DataFrame |
import os
import pandas as pd
import copy
import jieba
import numpy as np
import time
def readExcel(url):
df=pd.read_excel(url,na_values='')
return df
def writeExcel(df,url=''):
write = pd.ExcelWriter(url)
df.to_excel(write, sheet_name='Sheet1')
write.save()
def genDF(df_Base,df_IPC):
column... | pd.DataFrame(None,columns=['Company','Cat']) | pandas.DataFrame |
# Standard library imports
import pandas as pd
import numpy as np
from math import pi
# Local application/library specific imports
from .water_demand import (
set_cropland_share,
get_ky_list,
get_kc_list,
get_evap_i,
get_eto,
get_eff_rainfall_i,
get_effective_rainfall,
get_season_days,
... | pd.DataFrame() | pandas.DataFrame |
import os
import json
import random
import numpy as np
import pandas as pd
from copy import deepcopy
from string import ascii_uppercase, digits
from shutil import copyfile, rmtree, copytree
from datetime import datetime
#from ai.bot import Agent
from ai.bot import Agent
from tasks.games.chess.chess import Chess
from s... | pd.DataFrame() | pandas.DataFrame |
import numpy as np
import pytest
from pandas.errors import UnsupportedFunctionCall
from pandas import DataFrame, DatetimeIndex, Series
import pandas._testing as tm
from pandas.core.window import Expanding
def test_doc_string():
df = DataFrame({"B": [0, 1, 2, np.nan, 4]})
df
df.expanding(2... | Series([1, 2]) | pandas.Series |
import io
import json
import os
import re
import pickle
import subprocess
import pandas as pd
import numpy as np
from textblob import TextBlob, Blobber
from textblob_de import TextBlobDE as TextBlobDE
from textblob_fr import PatternTagger as PatternTaggerFR, PatternAnalyzer as PatternAnalyzerFR
import nltk
nltk.down... | pd.concat(df_list) | pandas.concat |
"""
Transfer applications.
|pic1|
.. |pic1| image:: ../images_source/transfer_tools/transfer.png
:width: 30%
"""
import os
import sys
from subprocess import Popen, PIPE
from pathlib import Path
import pandas as pd
import pexpect
import requests
import zipfile
from selenium.webdriver.chrome import webdriv... | pd.concat([dat_all, dat]) | pandas.concat |
import sys
import pandas as pd
import numpy as np
from scipy import stats
from itertools import compress
import statsmodels.stats.multitest as smt
import scikits.bootstrap as bootstrap
from sklearn.decomposition import PCA
from .scaler import scaler
from .imputeData import imputeData
class statistics:
usage = """G... | pd.DataFrame({mean_fold_change_name: [meanFoldChange], mean_fold_change_name_CIlower: CIs[0], mean_fold_change_name_CIupper: CIs[1], mean_fold_change_name_sig: [sigMeanFold]}) | pandas.DataFrame |
# coding=utf-8
# pylint: disable-msg=E1101,W0612
from datetime import datetime, timedelta
from numpy import nan
import numpy as np
import pandas as pd
from pandas.types.common import is_integer, is_scalar
from pandas import Index, Series, DataFrame, isnull, date_range
from pandas.core.index import MultiIndex
from pa... | pd.Series(['foo'], index=[0]) | pandas.Series |
from datetime import date
import unittest
import dolphindb as ddb
import pandas as pd
import numpy as np
from pandas.testing import assert_frame_equal
from setup import HOST, PORT, WORK_DIR, DATA_DIR
from numpy.testing import assert_array_equal, assert_array_almost_equal
import dolphindb.settings as keys
impor... | assert_frame_equal(re, expected) | pandas.testing.assert_frame_equal |
from itertools import product
import pandas as pd
from pandas.testing import assert_series_equal, assert_frame_equal
import pytest
from solarforecastarbiter.validation import quality_mapping
def test_ok_user_flagged():
assert quality_mapping.DESCRIPTION_MASK_MAPPING['OK'] == 0
assert quality_mapping.DESCR... | pd.Series([0, 0, 0, 1, 1]) | pandas.Series |
# -*- coding: utf-8 -*-
from __future__ import print_function
import pytest
import operator
from collections import OrderedDict
from datetime import datetime
from itertools import chain
import warnings
import numpy as np
from pandas import (notna, DataFrame, Series, MultiIndex, date_range,
Time... | pd.Timestamp('20130101') | pandas.Timestamp |
#!/usr/bin/env python
import pandas as pd
import argparse
import datetime
import time
import sys
import investpy
scrap_delay = 2
def main():
parser = argparse.ArgumentParser(description='scrap investing.com daily close')
parser.add_argument('-input_file', type=str, default='data_tickers/investing_stock_info.... | pd.concat(info_list) | pandas.concat |
##############################################################
# Author: <NAME>
##############################################################
'''
Module : create_kallisto_ec_count_matrix
Description : Create equivalence class matrix from kallisto.
Copyright : (c) <NAME>, Dec 2018
License : MIT
Maintainer :... | pd.merge(counts, tx_stack, left_on='ec_names', right_on='ec_names') | pandas.merge |
from datetime import timedelta
import pytest
import pandas.util._test_decorators as td
from pandas import (
DataFrame,
to_datetime,
)
@pytest.fixture(params=[True, False])
def raw(request):
"""raw keyword argument for rolling.apply"""
return request.param
@pytest.fixture(
params=[
"tr... | DataFrame([[2.0, 4.0], [1.0, 2.0], [5.0, 2.0], [8.0, 1.0]], columns=[1, 0.0]) | pandas.DataFrame |
from __future__ import division
import pytest
import numpy as np
from datetime import timedelta
from pandas import (
Interval, IntervalIndex, Index, isna, notna, interval_range, Timestamp,
Timedelta, compat, date_range, timedelta_range, DateOffset)
from pandas.compat import lzip
from pandas.tseries.offsets imp... | tm.assert_raises_regex(ValueError, msg) | pandas.util.testing.assert_raises_regex |
import submodels_module as modelbank
import numpy as np
from itertools import combinations
import matplotlib as mpl
mpl.use('Agg')
import matplotlib.pyplot as plt
import pandas as pd
import load_format_data
#Determine the most generalizable model from the top CV models
def get_loss_list(model_list):
model_los... | pd.DataFrame([model_name_list,model_loss_list,model_loss_std_list]) | pandas.DataFrame |
from datetime import (
datetime,
timedelta,
)
import re
import numpy as np
import pytest
from pandas._libs import iNaT
from pandas.errors import InvalidIndexError
import pandas.util._test_decorators as td
from pandas.core.dtypes.common import is_integer
import pandas as pd
from pandas import (
Categoric... | is_integer(result) | pandas.core.dtypes.common.is_integer |
import unittest
from context import grama as gr
from context import data
from numpy import NaN
from pandas import DataFrame, RangeIndex
from pandas.testing import assert_frame_equal
class TestPivotLonger(unittest.TestCase):
"""Test implementation of pivot_longer
"""
def test_pivot_longer(self):
"... | assert_frame_equal(long, expected) | pandas.testing.assert_frame_equal |
# -*- coding: utf-8 -*-
"""Supports Kp index values. Downloads data from ftp.gfz-potsdam.de or SWPC.
Parameters
----------
platform
'sw'
name
'kp'
tag
- '' : Standard Kp data
- 'forecast' : Grab forecast data from SWPC (next 3 days)
- 'recent' : Grab last 30 days of Kp data from SWPC
Note
----
Sta... | pds.date_range(forecast_date, periods=24, freq='3H') | pandas.date_range |
# Version 2.0 of the t-SNE Stock Market Example: with triggers and better defined functions.
# Try with simulated data so we know that they are clustered
#___________________________________________________________________________________________________ Import packages
import numpy as np
import pandas as pd
... | pd.read_csv('all_stocks_5yr.csv') | pandas.read_csv |
from nose.tools import with_setup
import pandas as pd
from ..widget import utils as utils
from ..widget.encoding import Encoding
df = None
encoding = None
def _setup():
global df, encoding
records = [
{
"buildingID": 0,
"date": "6/1/13",
"temp_diff": 12,
... | pd.date_range("2012", periods=3, freq="A") | pandas.date_range |
import pandas as pd
df = pd.read_csv('D:/5674-833_4th/part5/stock-data.csv')
#문자열인 날짜 데이터를 판다스 Timestamp로 변환
df['new_date']=pd.to_datetime(df['Date'])
df.set_index('new_date',inplace= True)
# print(df.loc['2018'].head())
# df_ym = df.loc['2018-07']
# print(df_ym)
#
today = | pd.to_datetime('2018-12-25') | pandas.to_datetime |
import pandas as pd
from src.features import build_features
from sklearn.model_selection import train_test_split
def make_dataset():
"""
This function laods the raw data, builds some features and saves the df.
It is not meant to be called but once to produce the dataset.
"""
raw_data = pd.read_csv(... | pd.DataFrame(df[mask]) | pandas.DataFrame |
import pandas as pd
from pandas._testing import assert_frame_equal
import pytest
import numpy as np
from scripts.normalize_data import (
remove_whitespace_from_column_names,
normalize_expedition_section_cols,
remove_bracket_text,
remove_whitespace,
ddm2dec,
remove_empty_unnamed_columns,
nor... | pd.DataFrame(data) | pandas.DataFrame |
"""
Script to processes basic data from all query
files to notebooks1.csv. After notebooks1.csv is created, files
can be downloaded with download.py.
"""
import time
import os
import datetime
import json
import sys
import argparse
import requests
import pandas as pd
from consts import (
URL,
COUNT_TRIGGER,... | pd.concat([repos_df, repos_done]) | pandas.concat |
"""Mock data for bwaw.insights tests."""
import pandas as pd
ACTIVE_BUSES = pd.DataFrame([
['213', 21.0921481, '1001', '2021-02-09 15:45:27', 52.224536, '2'],
['213', 21.0911025, '1001', '2021-02-09 15:46:22', 52.2223788, '2'],
['138', 21.0921481, '1001', '2021-02-09 15:45:27', 52.224536, '05'],
['138'... | pd.to_datetime(SPEED_INCIDENT['Time']) | pandas.to_datetime |
'''
MIT License
Copyright (c) 2020 <NAME>
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distri... | pd.merge(df, df_localidades, left_on='origen', right_on='Localidad') | pandas.merge |
"""Store the data in a nice big dataframe"""
import sys
from datetime import datetime, timedelta
import pandas as pd
import geopandas as gpd
import numpy as np
class Combine:
"""Combine defined countries together"""
THE_EU = [ 'Austria', 'Italy', 'Belgium', 'Latvia',
'Bulgaria', 'L... | pd.read_csv('data/belgium.csv', delimiter=',') | pandas.read_csv |
'''
Functions for the neural network
'''
import os
import numpy as np
import numba as nb
import pandas as pd
# Project imports
from utils import tickers, strategy, dates, fundamentals, io
# Other imports and type-hinting
from pandas import DataFrame as pandasDF
def main(nn_config: dict,
strat_config: dict)... | pd.read_csv(f'data/{ticker}.csv') | pandas.read_csv |
import pytest
import datetime
from pymapd._loaders import _build_input_rows
from pymapd import _pandas_loaders
from omnisci.mapd.MapD import TStringRow, TStringValue, TColumn, TColumnData
import pandas as pd
import numpy as np
from omnisci.mapd.ttypes import TColumnType
from omnisci.common.ttypes import TTypeInfo
def... | pd.Timestamp("2017") | pandas.Timestamp |
"""
Copyright 2019 <NAME>.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing,
software distribut... | pd.DataFrame(df) | pandas.DataFrame |
import pandas as pd
import numpy as np
import seaborn as sns
import matplotlib.pyplot as plt
from IPython.core.display import display
from scipy.stats import chi2_contingency
import glob
import os
import warnings
warnings.filterwarnings("ignore")
pd.options.display.float_format = '{:.4f}'.format
class DataExplorer()... | pd.to_datetime(date_df[colname]) | pandas.to_datetime |
import os.path
import json
import zipfile
import numpy as np
import pandas as pd
import requests
from openpyxl import load_workbook
import ukcensusapi.Nomisweb as Api
import ukpopulation.utils as utils
class SNPPData:
"""
Functionality for downloading and collating UK Subnational Population Projection (NPP) dat... | pd.DataFrame(data=females[1:, 1:], index=females[1:, 0], columns=females[0, 1:]) | pandas.DataFrame |
import unittest
import logging
import os
import numpy as np
import pandas as pd
import cmapPy.pandasGEXpress as GCToo
import cmapPy.pandasGEXpress.parse as parse
import broadinstitute_psp.utils.setup_logger as setup_logger
import broadinstitute_psp.tear.continuous_renormalization as renorm
# Setup logger
... | pd.Series([1.47, 1.42, 1.37, 1.31]) | pandas.Series |
import logging
import os
import unittest
import pandas as pd
import moneytrack as mt
logging.basicConfig(level=logging.DEBUG)
field_names = mt.Config.FieldNames
class DatasetTests(unittest.TestCase):
def test_a(self):
sim = mt.simulation.AccountSimulatorFixedRate(date=pd.to_datetime("2021-01-01"), ayr... | pd.to_datetime("2021-01-01") | pandas.to_datetime |
#!/usr/bin/env python
# -*- coding:utf-8 _*-
"""
@author: <NAME>
@file: grow_path.py
@time: 2021/01/19/13:42
"""
import os
import time
import argparse
import numpy as np
import pandas as pd
from rdkit import Chem
from rdkit.Chem import Descriptors
from rdkit.Chem.rdMolDescriptors import CalcExactMolWt
import subpro... | pd.read_csv(merged_df_path) | pandas.read_csv |
import pymysql
import pandas as pd
import numpy as np
import tushare as ts
from tqdm import tqdm
from sqlalchemy import create_engine
from getdata import read_mysql_and_insert
from datetime import date,timedelta,datetime
pro = ts.pro_api(token='???')
def strategy():
"""
auto strategy
"""
... | pd.DataFrame(index=all_stock_info.ts_code) | pandas.DataFrame |
"""
Functions and methods to extract statistics concerning the most used emojis in twitter datasets.
"""
from collections import Counter
import emoji
import seaborn as sns
import pandas as pd
import sys
import resource
from tqdm import tqdm
from IPython.core.debugger import set_trace
from tqdm import tqdm
import matpl... | pd.read_csv(tweet_path, chunksize=10000) | pandas.read_csv |
from __future__ import absolute_import
from __future__ import print_function
import os
import pandas as pd
import numpy as np
import sys
import shutil
from sklearn.preprocessing import MinMaxScaler
def dataframe_from_csv(path, header=0, index_col=False):
return pd.read_csv(path, header=header, index_col=index_... | pd.factorize(dx_type) | pandas.factorize |
# BSD 2-CLAUSE LICENSE
# Redistribution and use in source and binary forms, with or without modification,
# are permitted provided that the following conditions are met:
# Redistributions of source code must retain the above copyright notice, this
# list of conditions and the following disclaimer.
# Redistributions i... | pd.concat([runtimes_df, split_runtime_df]) | pandas.concat |
import pytz
import pytest
import dateutil
import warnings
import numpy as np
from datetime import timedelta
from itertools import product
import pandas as pd
import pandas._libs.tslib as tslib
import pandas.util.testing as tm
from pandas.errors import PerformanceWarning
from pandas.core.indexes.datetimes import cdate_... | Timestamp('2000-01-31 00:23:00') | pandas.Timestamp |
import pandas as pd
import snowflake.connector
import getpass as pwd
| pd.set_option('display.max_rows', None) | pandas.set_option |
# Copyright 2016 <NAME> and The Novo Nordisk Foundation Center for Biosustainability, DTU.
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
# http://www.apache.org/licenses/LICENSE-2.0
# Unle... | DataFrame(columns=["formula", "atoms", "bonds", "tanimoto_similarity", "structural_score"]) | pandas.DataFrame |
import pandas as pd
import os
def get_oneday_data(machine_path, machine_id, day):
data = | pd.read_csv(machine_path, header=None) | pandas.read_csv |
from zipline.api import symbol
from zipline import run_algorithm
import pandas as pd
def validate_single_stock(ticker):
def init(context):
symbol(ticker)
def handle_data(context, data):
pass
start = pd.to_datetime("2017-01-09").tz_localize('US/Eastern')
end = | pd.to_datetime("2017-01-11") | pandas.to_datetime |
import openpyxl
import pandas as pd
from datetime import datetime, timedelta
import xlsxwriter
now = datetime.now()
date_time = now.strftime("%m_%d_%Y %I_%M_%p")
federal_tax_rate_path = "./federaltaxrates.csv"
state_tax_rate_path = "./statetaxrates.csv"
city_tax_rate_path = "./NYCtaxrates.csv"
# calculate social s... | pd.DataFrame(analytics_table) | pandas.DataFrame |
# -*- coding: utf-8 -*-
import numpy as np
import pandas as pd
import json
import matplotlib.pyplot as plt
from datetime import datetime
from sys import stdout
from sklearn.preprocessing import scale
from sklearn.gaussian_process import GaussianProcessRegressor
from sklearn.gaussian_process.kernels import RBF, Constant... | pd.concat([X, bio_data], axis=1) | pandas.concat |
import multiprocessing
import os
from queue import Queue
from typing import List
from injector import inject
import pandas as pd
from pandas import DataFrame
from domain.operation.execution.services.OperationCacheService import OperationCacheService
from infrastructor.connection.adapters.ConnectionAdapter import Conn... | pd.notnull(df) | pandas.notnull |
import pandas as pd
import matplotlib.pyplot as plt
import numpy as np
#-------------read csv---------------------
df_2010_2011 = pd.read_csv("/mnt/nadavrap-students/STS/data/data_Shapira_20200911_2010_2011.csv")
df_2012_2013 = pd.read_csv("/mnt/nadavrap-students/STS/data/data_Shapira_20200911_2012_2013.csv")
df_2014_... | pd.merge(df1, df_2012, on='hospid') | pandas.merge |
#!/usr/bin/env python
#-*- coding:utf-8 -*-
"""
*.py: Description of what * does.
Last Modified:
"""
__author__ = "<NAME>"
__email__ = "<EMAIL>"
__version__ = "0.0.1"
# import gevent
from .dbManager import SQLiteWrapper, MongoDBWrapper
import pandas as pd
from . import GeoPoint, encode, blacklist, loc_defaul... | pd.DataFrame([i.__dict__ for i in ldist]) | pandas.DataFrame |
from netCDF4 import Dataset
import pandas as pd
import numpy as np
ncep_path = '/SubX/forecast/tas2m/daily/full/NCEP-CFSv2/' # the path where the raw data from NCEP-CFSv2 is saved
gmao_path = '/SubX/forecast/tas2m/daily/full/GMAO-GEOS_V2p1/'
for model in ['NCEP', 'GMAO']:
if model == 'NCEP':
path = ncep_... | pd.date_range('2017-07-01', '2019-12-31') | pandas.date_range |
import argparse
import math
import json
from tqdm import tqdm
from nltk.tag import pos_tag
import pandas as pd
import networkx as nx
import torch
import config
def get_relevant_tokens(word_count_path, threshold):
d = pd.read_csv(word_count_path, sep='\t', header=None, quotechar=None, quoting=3)
d.columns = ... | pd.read_csv(Seid_amharic_sentiment_path, header=None) | pandas.read_csv |
#!/usr/bin/python
# -*- coding: utf-8 -*-
"""
Created on Thu Jul 18 04:52:01 2019
@author: jamiesom
"""
import pandas as pd
from electricitylci.globals import data_dir, output_dir
import numpy as np
from electricitylci.eia860_facilities import eia860_generator_info
import re
def generate_power_plant_construction(year... | pd.read_csv(f"{data_dir}/plant_construction_inventory.csv") | pandas.read_csv |
from plotly.offline import plot, iplot, init_notebook_mode
from pandas.plotting import register_matplotlib_converters
import seaborn as sns
import matplotlib.pyplot as plt
from urllib.request import urlopen
from datetime import timedelta
import json
import numpy as np
import pandas as pd
import plotly.express as px
imp... | pd.read_csv('datasets/complete.csv', parse_dates=['Date']) | pandas.read_csv |
import pandas as pd
import datetime
import numpy as np
import icd
def get_age(row):
"""Calculate the age of patient by row
Arg:
row: the row of pandas dataframe.
return the patient age
"""
raw_age = row['DOD'].year - row['DOB'].year
if (row['DOD'].month < row['DOB'].month) or ((row['... | pd.read_csv(mimic_admissions) | pandas.read_csv |
import pandas as pd
# bookings_to_arr
#
# Accepts a pandas dataframe containing bookings data and returns a pandas
# dataframe containing changes in ARR with the following columns:
# - date - the date of the change
# - type - the type of the change (new, upsell, downsell, and churn)
# - customer_id - the id o... | pd.Timestamp(ts_input="10/1/2021", tz="UTC") | pandas.Timestamp |
import pandas as pd
def load_dataset(csv_path):
df_inflacao = | pd.read_csv(csv_path, sep=';', decimal=',') | pandas.read_csv |
""" Compute the accuracies required to compute the overlapping scores.
Namely, for each model m trained with data augmentation on one candidate corruption (trained with corruption_trainings.py), get the accuracy of m for each candidate corruption (using the ImageNet validation set corrupted with the considered corrupt... | pandas.DataFrame(res_array, index=list_models, columns=list_corruptions) | pandas.DataFrame |
"""Run 20newsgroups data experiment."""
import os
import numpy as np
import random
import pickle
import pandas as pd
import methods_20news
from methods_20news import Methods
from prep_20news import *
from utils_20news import *
from statistics import mean
from statistics import median
from statistics import stdev
rand... | pd.get_dummies(subcat_all) | pandas.get_dummies |
# -*- coding: utf-8 -*-
"""
Created on Wed Jul 14 09:27:05 2021
@author: vargh
"""
import numpy as np
import pandas as pd
from sympy import symbols, pi, Eq, integrate, diff, init_printing, solve
from scipy.optimize import curve_fit
from scipy.integrate import cumtrapz
from scipy.interpolate import interp1d, interp2d
... | pd.read_csv(filename) | pandas.read_csv |
# -*- coding: utf-8 -*-
import nose
import numpy as np
from datetime import datetime
from pandas.util import testing as tm
from pandas.core import config as cf
from pandas.compat import u
from pandas.tslib import iNaT
from pandas import (NaT, Float64Index, Series,
DatetimeIndex, TimedeltaIndex, da... | notnull(-np.inf) | pandas.types.missing.notnull |
import os
import gzip
import pandas as pd
import scipy.io as sio
import pathlib
from enum import Enum
import torch
from sklearn.preprocessing import MinMaxScaler, normalize, StandardScaler
class Btype(Enum):
Undefined = 0
Normal = 1
ESSV_aka_PAC = 2
Aberrated = 3
ESV_aka_PVC = 4... | pd.DataFrame(test_labels_subset, columns=['patient', 'segment', 'frame', 'bt_label', 'rt_label']) | pandas.DataFrame |
from kivy.uix.boxlayout import BoxLayout
from kivy.uix.label import Label
from kivy.uix.popup import Popup
import pandas as pd
class Ledger(BoxLayout):
"""
Ledger data structure:
x: first number in equation (float)
y: second number in equation (float)
op: operator ['+', "-", '*', '/'] (str)
z... | pd.DataFrame(columns=['x', 'y', 'op', 'z']) | pandas.DataFrame |
#!/usr/bin/env python3
"""
Author : <NAME>
Date : 2022-02-03
Purpose: Parse tracy JSON files and produce summary .xlsx sheet.
"""
import argparse
from typing import NamedTuple
import json, pathlib, time
import pandas as pd
class Args(NamedTuple):
""" Command-line arguments """
json_file_path: pathlib.Path
... | pd.DataFrame.from_dict(SNP_data, orient='index') | pandas.DataFrame.from_dict |
import pandas as pd
import matplotlib.pyplot as plt
import numpy as np
from sklearn.utils import resample
from scipy.ndimage import gaussian_filter
from scipy import signal
import cv2
## plot the data classes as a circle to view the unbalance between the classes
def plot_num_of_classes(labels):
plt.figure(figsize... | pd.DataFrame(df_4) | pandas.DataFrame |
import importlib
import pandas as pd
from pandas import compat
from .parser import Parser
import logging
class UberModel(object):
"""
Collection of static methods used across all the ubertool models.
"""
def __init__(self):
"""Main utility class for building Ubertool model classes for model e... | pd.Series([], dtype="object") | pandas.Series |
# %%
import pandas as pd
import numpy as np
import requests # http 요청 모듈
from bs4 import BeautifulSoup # 웹 크롤링 모듈
from urllib.request import urlopen # 웹 크롤링 모듈
from urllib.parse import quote_plus, urlencode
from pandas import DataFrame, Series # 시리즈, 데이터프레임 모듈
from pandas import ExcelFile, ExcelWriter # 엑셀 읽기, 쓰기... | pd.set_option("display.max_columns", 15) | pandas.set_option |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
#-----------------------------------------------------------------------------
# Copyright (c) 2015, IBM Corp.
# All rights reserved.
#
# Distributed under the terms of the BSD Simplified License.
#
# The full license is in the LICENSE file, distributed with this software.
... | pd.Categorical(dataframe['level_1'], agg_values) | pandas.Categorical |
from collections import abc, deque
from decimal import Decimal
from io import StringIO
from warnings import catch_warnings
import numpy as np
from numpy.random import randn
import pytest
from pandas.core.dtypes.dtypes import CategoricalDtype
import pandas as pd
from pandas import (
Categorical,
DataFrame,
... | Series([4, 5, 6]) | pandas.Series |
# Copyright 2017 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, ... | pd.Series(['', '']) | pandas.Series |
import math
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
def clean_portfolio(portfolio):
""" Clean the portfolio dataset.
- It makes columns for the channels
- Changes the name of the id column to offer_id
Input:
- portfolio: original dataset
Returns:
- portfolio_... | pd.merge(trans_prof, portfolio_clean, on='offer_id', how='left') | pandas.merge |
#!/usr/bin/python3
# import the module
import os
import glob
import pandas as pd
import csv
from sqlalchemy import create_engine
import psycopg2
import config #you need to create this config.py file and update the variables with your database, username and password
import subprocess
import sys
#Note: you need to ind... | pd.read_csv("/home/bmain/pesticide/chem_com.csv") | pandas.read_csv |
# -*- coding: utf-8 -*-
from __future__ import print_function
from datetime import datetime, timedelta
import functools
import itertools
import numpy as np
import numpy.ma as ma
import numpy.ma.mrecords as mrecords
from numpy.random import randn
import pytest
from pandas.compat import (
PY3, PY36, OrderedDict, ... | tm.assert_frame_equal(result, expected) | pandas.util.testing.assert_frame_equal |
# encoding: utf-8
import argparse
import os
import sys
import torch
from torch.backends import cudnn
import numpy as np
import random
sys.path.append('.')
from data import make_data_loader
from model import build_model
from engine.evaluator import do_inference
from config import cfg
from utils.logger import setup... | pd.ExcelWriter(xls_filename, engine="openpyxl", mode='a') | pandas.ExcelWriter |
from .nwb_interface import NWBDataset
from .chop import ChopInterface, chop_data, merge_chops
from itertools import product
import numpy as np
import pandas as pd
import h5py
import sys
import os
import logging
logger = logging.getLogger(__name__)
PARAMS = {
'mc_maze': {
'spk_field': 'spikes',
'h... | pd.concat([dataset.trial_info, align_jit], axis=1) | pandas.concat |
import sys
import itertools
from pathlib import Path
from matplotlib import pyplot as plt
import pandas as pd
import seaborn as sns
from matplotlib import cm
FIGURES_DIR = (
Path(__file__).resolve().parents[2]
/ "figures"
/ "ukbiobank"
/ Path(sys.argv[0]).stem
)
FIGURES_DIR.mkdir(exist_ok=True, paren... | pd.concat(scores, ignore_index=False) | pandas.concat |
"""Module to support machine learning of activity states from acc data"""
from accelerometer import utils
from accelerometer.models import MODELS
from io import BytesIO
import numpy as np
import os
import pandas as pd
from sklearn.ensemble import RandomForestClassifier
import sklearn.ensemble._forest as forest
import ... | pd.DataFrame(data=cnf_matrix, columns=labels, index=labels) | pandas.DataFrame |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import streamlit as st
import pandas as pd
import numpy as np
import geopandas as gpd
from pathlib import Path
from PIL import Image
import altair as alt
import pydeck as pdk
import numpy as np
from api_key import mapbox_key
import matplotlib.pyplot as plt
import plotly.exp... | pd.merge(filtered_drug,df_geometries[['id','lat','lon']],how = 'left',left_on = 'patientid',right_on = 'id') | pandas.merge |
#!/usr/bin/env python3
import pdb
import pandas as pd
from pylru import lrudecorator
import seaborn as sns
BII_URL = 'http://ipbes.s3.amazonaws.com/weighted/' \
'historical-BIIAb-npp-country-1880-2014.csv'
@lrudecorator(10)
def get_raw_bii_data():
return pd.read_csv(BII_URL)
def findt(ss):
rval = [None] * le... | pd.read_csv(url, encoding='utf-8') | pandas.read_csv |
import os
import pandas as pd
from Lib.get_texts import get_generated_lyrics, get_lyrics_dataset
from Lib.get_structure import get_lyrics_structure
from Lib.get_sentiment import calculate_sentiment_scores
from Lib.get_bagofwords import get_repetition_scores, combine_bag_of_words, lemmatize_lyrics
# This method perfo... | pd.concat([df, df2], axis=1) | pandas.concat |
"""
Download, transform and simulate various datasets.
"""
# Author: <NAME> <<EMAIL>>
# <NAME> <<EMAIL>>
# License: MIT
from os.path import join
from re import sub
from collections import Counter
from itertools import product
from urllib.parse import urljoin
from string import ascii_lowercase
from zipfile imp... | pd.read_csv(FETCH_URLS["volkert"]) | pandas.read_csv |
'''
Created on Sep 10, 2017
@author: twong
'''
import json
import logging
import random
import pandas as pd
import requests
_logger = logging.getLogger(__name__)
def _deserialize_roster_json(roster_json):
roster_cooked = json.loads(roster_json)
try:
players_json = roster_cooked['d'][0]
except ... | pd.concat([t.roster for t in teams], ignore_index=True) | pandas.concat |
import numpy as np
import pandas as pd
from scipy import stats
from ..base import AbstractDensity
class Multinomial:
def __init__(self, probs):
'''
Define a multinomial random variable object
:param probs: The probability of each class, with classes indexed as 0 to len(probs)-1
''... | pd.DataFrame({'bart_simpson': samples}) | pandas.DataFrame |
"""
This script reads all the bootstrap performance result files, plots histograms, and calculates averages.
t-tests are done to compute p-values and confidence intervals are computed
"""
import pandas as pd
import numpy as np
import os
import matplotlib.pyplot as plt
import matplotlib
from scipy import stat... | pd.read_csv(data) | pandas.read_csv |
import pandas as pd
import numpy as np
from sklearn.preprocessing import StandardScaler, LabelEncoder
from querying.tracking_query import get_play
def ft_in(x):
if '-' in x:
meas=x.split('-')
#this will be a list ['ft','in']
inches = int(meas[0])*12 + int(meas[1])
return inches
... | pd.merge(play_p, track19, left_on = ['gameId', 'playId'], right_on = ['gameId', 'playId']) | pandas.merge |
# pylint: disable=E1101,E1103,W0232
from datetime import datetime, timedelta
from pandas.compat import range, lrange, lzip, u, zip
import operator
import re
import nose
import warnings
import os
import numpy as np
from numpy.testing import assert_array_equal
from pandas import period_range, date_range
from pandas.c... | MultiIndex.from_tuples(tups) | pandas.core.index.MultiIndex.from_tuples |
from __future__ import division
import pandas as pd
import numpy as np
import scipy.stats
import argparse
import datetime
def fileToList(group_list):
with open(group_list, 'r') as fh:
return [line.strip() for line in fh.readlines()]
def cleanDF(df, sample_names):
'''
Convert string nans to np.nan ... | pd.read_csv(data_input, sep=',') | pandas.read_csv |
import numpy as np
import pandas as pd
from pandas.testing import assert_series_equal
import pytest
from ber_public.deap import dim
from ber_public.deap import fab
from ber_public.deap import vent
def test_calculate_fabric_heat_loss():
"""Output is equivalent to DEAP 4.2.0 example A"""
floor_area = pd.Series... | pd.Series([0.14]) | pandas.Series |
"""This module contains pyspark wrangler utility tests.
isort:skip_file
"""
import pytest
import pandas as pd
from pywrangler.pyspark.util import ColumnCacher
from pyspark.sql import functions as F
pytestmark = pytest.mark.pyspark # noqa: E402
pyspark = pytest.importorskip("pyspark") # noqa: E402
from pywrangler.... | pd.DataFrame(data) | pandas.DataFrame |
import os,sys
import pandas as pd
import numpy as np
import skbio.io
import gffpandas.gffpandas as gffpd
from statistics import stdev
def find_locs(kmer, blast_df):
"""
Finds the start and stop locations of this k-mer in each genome
"""
locs = []
# filter blast results to just our kmer of interest... | pd.read_csv("data/gene_labels.tsv",sep='\t') | pandas.read_csv |
# -*- coding: utf-8 -*-
"""
Created on Sun Sep 1 18:10:18 2019
@author: <NAME>
Code will plot the keypoint coordinates vs time in order to assign the maximum
value from this plot to the real-world distance measurement. This will be
the label.
Coding Improvement Note: Make use of functions for things lik... | pd.DataFrame(data=promsx_6[0][0:]) | pandas.DataFrame |
from collections import Counter
import sys
from sklearn.feature_extraction.text import CountVectorizer, TfidfVectorizer, HashingVectorizer
from sklearn.feature_extraction import DictVectorizer
import sklearn.cluster.k_means_
from sklearn.cluster.k_means_ import KMeans, MiniBatchKMeans
from sklearn.cluster import Spectr... | pd.DataFrame(m) | pandas.DataFrame |
# Copyright (c) 2018-2021, NVIDIA CORPORATION.
import array as arr
import datetime
import io
import operator
import random
import re
import string
import textwrap
from copy import copy
import cupy
import numpy as np
import pandas as pd
import pyarrow as pa
import pytest
from numba import cuda
import cudf
from cudf.c... | pd.DataFrame() | pandas.DataFrame |
import pandas as pd
import click
import os
@click.command()
@click.argument('input_folder')
@click.argument('output_folder')
def main(input_folder, output_folder):
if not os.path.exists(output_folder):
os.makedirs(output_folder)
files = [[(x[0] + '/' + y, x[0].split('/')[-1].replace('DATA_', '').rep... | pd.read_csv(all_mutations_filtered_mut_type_gene) | pandas.read_csv |
import pytest
import numpy as np
import pandas as pd
from systrade.trading.brokers import PaperBroker
T_START = pd.to_datetime('2019/07/10-09:30:00:000000', format='%Y/%m/%d-%H:%M:%S:%f')
T_END = pd.to_datetime('2019/07/10-10:00:00:000000', format='%Y/%m/%d-%H:%M:%S:%f')
TIMEINDEX = pd.date_range(start=T_START,en... | pd.to_datetime('2019/07/10-09:29:00:000000', format='%Y/%m/%d-%H:%M:%S:%f') | pandas.to_datetime |
import requests
import os
import json
import pandas as pd
import numpy as np
from requests.exceptions import HTTPError
import re
import matplotlib.pyplot as plt
# Part 1 Get Data With API
## Function 1-Get dataset with specific game "platform" and "type"
def api_game(platform = 'pc', type = 'game'):
"""
... | pd.DataFrame(g_list) | pandas.DataFrame |
# -*- coding: utf-8 -*-
import sys
import os
import pandas as pd
PROJECT_ID = "dots-stock" # @param {type:"string"}
REGION = "us-central1" # @param {type:"string"}
USER = "shkim01" # <---CHANGE THIS
BUCKET_NAME = "gs://pipeline-dots-stock" # @param {type:"string"}
PIPELINE_ROOT = f"{BUCKET_NAME}/pipeline_root/{USE... | pd.read_csv(bros_dataset.path, index_col=0) | pandas.read_csv |
# -*- coding: utf-8 -*-
"""
Created on Fri Jun 26 23:51:40 2020
@author: Narendrakumar
"""
# Part 1 - Data Preprocessing
# Importing the libraries
import numpy as np
import matplotlib.pyplot as plt
import pandas as pd
# Importing the dataset
dataset = | pd.read_csv('Churn_Modelling.csv') | pandas.read_csv |
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
scenario_filenames = ["OUTPUT_110011_20201117123025"]
scenario_labels =["Lockdown enabled,Self Isolation,Mask Compliance (0.5)"]
MAX_DAY = 250#250#120
POPULATION = 10000.0
FIGSIZE = [20,10]
plt.rcParams.update({'font.size': 22})
#### compari... | pd.to_datetime(df1["Date_Time"]) | pandas.to_datetime |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.