prompt stringlengths 19 1.03M | completion stringlengths 4 2.12k | api stringlengths 8 90 |
|---|---|---|
# ---
# jupyter:
# jupytext:
# formats: ipynb,py:percent
# text_representation:
# extension: .py
# format_name: percent
# format_version: '1.3'
# jupytext_version: 1.13.7
# kernelspec:
# display_name: Python [conda env:holoview]
# language: python
# name: conda-env-holovi... | pd.DataFrame(wrfout_vars) | pandas.DataFrame |
import requests
from pandas import DataFrame, Series
import pandas as pd
from utilities import (LICENSE_KEY, generate_token, master_player_lookup)
import numpy as np
pd.options.mode.chained_assignment = None
######################
# top level functions:
######################
def get_league_rosters(lookup, league_id... | pd.concat([starter_df2, bench_df], ignore_index=True) | pandas.concat |
import pandas as pd
import numpy as np
import random
# report confusion matrix with labels
def confusion_matrix(predicted, true):
if len(predicted) != len(true):
print("Error: lengths of labels do not match")
else:
d = {'predicted': predicted, 'true': true}
df = pd.DataFrame(data=d)... | pd.DataFrame(data=d) | pandas.DataFrame |
#!/usr/bin/env python
import sys, time
import numpy as np
from io import StringIO
import pickle as pickle
from pandas import DataFrame
from pandas import concat
from pandas import read_pickle
from pandas import cut
from pandas import concat
from sklearn.externals import joblib
from sklearn.cross_validation... | concat(blunder_cv_results, axis=1) | pandas.concat |
"""Tools for creating and manipulating neighborhood datasets."""
import os
import pathlib
from warnings import warn
import geopandas as gpd
import pandas as pd
from appdirs import user_data_dir
appname = "geosnap"
appauthor = "geosnap"
data_dir = user_data_dir(appname, appauthor)
def _fetcher(local_path, remote_pat... | pd.concat(blocks, sort=True) | pandas.concat |
import matplotlib.pyplot as plt
import cantools
import pandas as pd
import cv2
import numpy as np
import os
import glob
import re
import subprocess
import json
LOG_FOLDER = "/media/andrei/Samsung_T51/nemodrive/25_nov/session_2/1543155398_log"
CAN_FILE_PATH = os.path.join(LOG_FOLDER, "can_raw.log")
DBC_FILE = "logan.db... | pd.read_csv(OBD_SPEED_FILE, header=None) | pandas.read_csv |
from collections import OrderedDict
import timeit
import numpy as np
import pandas as pd
from randomstate.prng import (mt19937, sfmt, dsfmt, xoroshiro128plus,
xorshift1024, pcg64)
REPS = 3
SIZE = 100000
SETUP = """
import numpy
from numpy import array, random
from randomstate.prng impor... | pd.DataFrame(results) | pandas.DataFrame |
# -*- coding: utf-8 -*-
"""
Created on Tue Mar 19 23:34:57 2019
@author: reynaldo.espana.rey
Web scrapping algorithm to build data set for text generator
source: https://towardsdatascience.com/how-to-web-scrape-with-python-in-4-minutes-bc49186a8460
"""
# ============================================================... | pd.DataFrame({'poem': links}) | pandas.DataFrame |
import copy
import itertools
import os
import numpy as np
import pandas as pd
from pathlib import Path
from sklearn.preprocessing import PowerTransformer
from scipy.stats import yeojohnson
from tqdm import tqdm
import tensorflow as tf
import warnings
warnings.simplefilter("ignore")
n_wavelengths = 5... | pd.DataFrame(temp_storage_float) | pandas.DataFrame |
import random
import numpy as np
import pandas as pd
import seaborn as sns
import matplotlib.pyplot as plt
# Explanation "Where": Plot for explanation in fundamentals chapter
# 1. Generate data with gaussian, uniform and mixed distribution
n = 3000
var = 0.12
# Dimension 1
dim1_sequence_100percent_gaussian = np.rand... | pd.Series(dim1_sequence_100percent_gaussian) | pandas.Series |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
调用wset函数的部分
下载数据的方法
1.在时间上使用折半可以最少的下载数据,但已经下了一部分,要补下时如果挪了一位,又得全重下
2.在文件上,三个文件一组,三组一样,删中间一个,直到不能删了,退出
"""
import os
import pandas as pd
from .utils import asDateTime
def download_sectorconstituent(w, date, sector, windcode, field='wind_code'):
"""
板块成份
中信证... | pd.concat([df, curr_df]) | pandas.concat |
def report_classification(df_features,df_target,algorithms='default',test_size=0.3,scaling=None,
large_data=False,encode='dummy',average='binary',change_data_type = False,
threshold=8,random_state=None):
'''
df_features : Pandas DataFrame
... | pd.DataFrame(columns=["Algorithm_name",'R-Squared','Adj. R-Squared','Train-RMSE','Test-RMSE']) | pandas.DataFrame |
# -*- coding: utf-8 -*-
import os
import pandas as pd
from pandas.testing import assert_frame_equal
import camelot
from camelot.core import Table, TableList
from camelot.__version__ import generate_version
from .data import *
testdir = os.path.dirname(os.path.abspath(__file__))
testdir = os.path.join(testdir, "fil... | pd.DataFrame(data_stream_strip_text) | pandas.DataFrame |
# This script performs the statistical analysis for the pollution growth paper
# Importing required modules
import pandas as pd
import numpy as np
import statsmodels.api as stats
from ToTeX import restab
# Reading in the data
data = pd.read_csv('C:/Users/User/Documents/Data/Pollution/pollution_data.csv')... | pd.get_dummies(imp_ch4_rob['Year']) | pandas.get_dummies |
#!/usr/bin/env python
"""
coding=utf-8
Build model for a dataset by identifying type of column along with its
respective parameters.
"""
from __future__ import print_function
from __future__ import division
import copy
import random
import re
from collections import OrderedDict, defaultdict
import warnings
import pic... | pd.DataFrame(data, dtype=object) | pandas.DataFrame |
import numpy as np
import pandas as pd
import matplotlib as mpl
from matplotlib.colors import same_color, to_rgb, to_rgba
import pytest
from numpy.testing import assert_array_equal
from seaborn.external.version import Version
from seaborn._core.rules import categorical_order
from seaborn._core.scales import Nominal,... | pd.Series(["a", "b", "c"]) | pandas.Series |
import pytest
import numpy as np
from datetime import date, timedelta, time, datetime
import dateutil
import pandas as pd
import pandas.util.testing as tm
from pandas.compat import lrange
from pandas.compat.numpy import np_datetime64_compat
from pandas import (DatetimeIndex, Index, date_range, DataFrame,
... | DatetimeIndex(['2000-01-04', '2000-01-01', '2000-01-02']) | pandas.DatetimeIndex |
import itertools
import os
import random
import tempfile
from unittest import mock
import pandas as pd
import pytest
import pickle
import numpy as np
import string
import multiprocessing as mp
from copy import copy
import dask
import dask.dataframe as dd
from dask.dataframe._compat import tm, assert_categorical_equal... | pd.DataFrame({"a": [9, 8, 7], "b": [6, 5, 4], "c": [3, 2, 1]}) | pandas.DataFrame |
import json
import numpy
import pandas
import re
from datetime import timedelta
from GoogleSheetIOStream import GoogleSheetIOStream
class BonusProcessor (object):
def __init__(self, iostream, config_dir='config/', working_folder='Chouta Stand Payroll', input_folder='Input', config_folder='Config'):
self.i... | pandas.to_datetime(hours['Clock Out']) | pandas.to_datetime |
import os
import logging
import pandas as pd
from slackbot import licence_plate
log = logging.getLogger(__name__)
class CarOwners:
# Source data:
# https://intranet.xebia.com/display/XNL/Xebia+Group+Kenteken+Registratie
def __init__(self, csv_path='/data/car-owners.csv'):
self.csv_path = csv_pa... | pd.notnull(self.owners_df) | pandas.notnull |
import re
import numpy as np
import pandas as pd
import random as rd
from sklearn import preprocessing
from sklearn.cluster import KMeans
from sklearn.ensemble import RandomForestRegressor
from sklearn.decomposition import PCA
# Print options
np.set_printoptions(precision=4, threshold=10000, linewidth=160, edgeitems=9... | pd.get_dummies(df_titanic_data['TicketPrefix']) | pandas.get_dummies |
"""This module contains nodes for signal filtering."""
import numpy as np
import pandas as pd
from scipy import signal
from timeflux.core.branch import Branch
from timeflux.core.node import Node
from timeflux.nodes.window import Window
from timeflux_dsp.utils.filters import (
construct_fir_filter,
construct_i... | pd.concat([self._previous, self.i.data], axis=0) | pandas.concat |
""" test the scalar Timedelta """
from datetime import timedelta
import numpy as np
import pytest
from pandas._libs import lib
from pandas._libs.tslibs import (
NaT,
iNaT,
)
import pandas as pd
from pandas import (
Timedelta,
TimedeltaIndex,
offsets,
to_timedelta,
)
import pandas._testing as ... | Timedelta(" 10000D ") | pandas.Timedelta |
"""Contains plotting code used by the web server."""
from datetime import timedelta
from bokeh.models import (
ColumnDataSource,
CustomJS,
DataRange1d,
Range1d,
Whisker,
LabelSet,
HoverTool,
)
from bokeh.models.formatters import DatetimeTickFormatter
from bokeh.layouts import row, Row
from ... | pd.DataFrame(measurement_pairs) | pandas.DataFrame |
#==============================================================================
# Import packages
#==============================================================================
import numpy as np
import pandas as pd
# Utilities
from sklearn.utils import resample
# Transformer to select a subset of the Pandas DataFram... | pd.read_csv(DATAFILE, index_col=ID, header=0, nrows=NTRAINROWS) | pandas.read_csv |
from os.path import join
import numpy as np
import pandas as pd
import geopandas as gpd
import matplotlib.pyplot as plt
from src import utils as cutil
def convert_non_monotonic_to_nan(array):
"""Converts a numpy array to a monotonically increasing one.
Args:
array (numpy.ndarray [N,]): input array
... | pd.MultiIndex.from_tuples([("date", "")]) | pandas.MultiIndex.from_tuples |
"""SQL io tests
The SQL tests are broken down in different classes:
- `PandasSQLTest`: base class with common methods for all test classes
- Tests for the public API (only tests with sqlite3)
- `_TestSQLApi` base class
- `TestSQLApi`: test the public API with sqlalchemy engine
- `TestSQLiteFallbackApi`: t... | sql.read_sql_query("SELECT * FROM iris_view", self.conn) | pandas.io.sql.read_sql_query |
# -*- coding: utf-8 -*-
"""Device curtailment plots.
This module creates plots are related to the curtailment of generators.
@author: <NAME>
"""
import os
import logging
import pandas as pd
from collections import OrderedDict
import matplotlib.pyplot as plt
import matplotlib as mpl
import matplotlib.dates as mdates
... | pd.DataFrame() | pandas.DataFrame |
import pandas as pd
import numpy as np
from run import prediction
import tensorflow as tf
import time
import os
np.random.seed(12345)
def top_k_movies(users,ratings_df,k):
"""
Returns top k movies for respective user
INPUTS :
users : list of numbers or number , list of user ids
rating... | pd.read_pickle("user_item_table.pkl") | pandas.read_pickle |
from config import engine
import pandas as pd
import numpy as np
from datetime import datetime
from collections import Counter
def date_difference(my_date, max_date):
'''
This function takes in a single date from the donations dataframe (per row) and compares the difference between that date and the date in w... | pd.DataFrame(df, columns=['matching_id', 'amount', 'close_date']) | pandas.DataFrame |
import seaborn as sns
import matplotlib
import matplotlib.gridspec as gridspec
import matplotlib.pyplot as plt
import matplotlib.colors as mcolors
from matplotlib.ticker import ScalarFormatter
from matplotlib import lines
import pandas as pd
import numpy as np
from pathlib import Path
import os
import sys
import csv
im... | pd.read_csv(filename, delimiter='\t') | pandas.read_csv |
from itertools import permutations
from typing import List, Dict
import pandas as pd
import scipy.stats
import numpy as np
import json
import sys, os
from statsmodels import api as sm
from src.constants import AVG_SEED_WINS, ESPN_SCORES
from scipy import stats
myPath = os.path.dirname(os.path.abspath(__file__))
sys.p... | pd.DataFrame(self.simulation_results) | pandas.DataFrame |
import pandas as pd
def _performer_list():
performers = [
['FULL_NAME', 'SHORT_NAME'],
#['Test and Evaluation Team', 'te'],
['Accenture', 'acc'],
['ARA','ara'],
['Astra', 'ast'],
['BlackSky', 'bla'],
['Kitware', 'kit'],
['STR', 'str'],
]
retur... | pd.DataFrame(ssh_site_list) | pandas.DataFrame |
import numpy as np
import arviz as az
import pandas as pd
import seaborn as sns
import matplotlib.pylab as plt
from math import *
import json
import itertools
import os
import re
sns.set_style("whitegrid")
import tools
from tools import toVec
def jSonIterator(j):
yield j
if isinstance(j,dict):
... | pd.merge(data,xs,left_index=True,right_index=True) | pandas.merge |
import sys
import os
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
from matplotlib.ticker import FormatStrFormatter
plt.rcParams['font.size'] = 6
root_path = os.path.dirname(os.path.abspath('__file__'))
# root_path = os.path.abspath(os.path.join(root_path,os.path.pardir))
graphs_path = root_pat... | pd.read_csv(root_path+'/Huaxian/projects/lstm/7_ahead/optimal/opt_pred.csv') | pandas.read_csv |
# coding: utf-8
# In[1]:
import pandas as pd
import numpy as np
import sklearn as sk
import matplotlib.pyplot as plt
import gc
train = pd.read_csv("train.csv",parse_dates=["activation_date"])
test = | pd.read_csv("test.csv",parse_dates=["activation_date"]) | pandas.read_csv |
import numpy as np
import pandas as pd
from sklearn.model_selection import GroupShuffleSplit as sklearnGroupShuffleSplit
class Split():
def __init__(self, dataset=None):
self.dataset = dataset
def UnSplit(self):
"""Unsplit the dataset by setting all values of the split column to null."""
... | pd.DataFrame(group) | pandas.DataFrame |
from bs4 import BeautifulSoup
from datetime import datetime, timedelta
import warnings
import requests
import pandas as pd
import re
class RegionMatcher:
"""
Ironing out disrepances between RosPotrebNadzor labels and iso_alpha3 codes.
"""
def get_simplified_region(self, x):
x = x.lower()
... | pd.read_csv(self.regions_fname) | pandas.read_csv |
# Copyright 2019 WISE-PaaS/AFS
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by... | pd.DataFrame(data=data) | pandas.DataFrame |
# =============================================================================
# Pelote Network to Tabular Unit Tests
# =============================================================================
import networkx as nx
import pandas as pd
from pytest import raises
from pelote.exceptions import MissingPandasException... | pd.DataFrame(data={"source": [1], "target": [2], "age": [47]}) | pandas.DataFrame |
"""
Author: <NAME>
Modified: <NAME>
"""
import os
import warnings
import numpy as np
import pandas as pd
import scipy.stats
import pytest
from numpy.testing import assert_almost_equal, assert_allclose
from statsmodels.tools.sm_exceptions import EstimationWarning
from statsmodels.tsa.holtwinters import (ExponentialSmo... | pd.Series(data, index) | pandas.Series |
# coding: utf-8
import os
import re
import numpy as np
import pandas as pd
import ujson as json
patient_ids = []
for filename in os.listdir('./raw'):
# the patient data in PhysioNet contains 6-digits
match = re.search('\d{6}', filename)
if match:
id_ = match.group()
patient_ids.append(id_... | pd.DataFrame(values) | pandas.DataFrame |
# -*- coding: utf-8 -*-
"""
v17.csv (final submission) ... averaging model of v9s, v13 and v16
"""
from logging import getLogger, Formatter, StreamHandler, INFO, FileHandler
import subprocess
import importlib
import math
from pathlib import Path
import json
import re
import warnings
import tqdm
import click
import tab... | pd.read_csv(fn_valtest, index_col='ImageId') | pandas.read_csv |
import numpy as np
from numpy import nan
import pytest
from pandas._libs import groupby, lib, reduction
from pandas.core.dtypes.common import ensure_int64
from pandas import Index, isna
from pandas.core.groupby.ops import generate_bins_generic
import pandas.util.testing as tm
from pandas.util.testing import assert_a... | generate_bins_generic(values, [4], "right") | pandas.core.groupby.ops.generate_bins_generic |
import pandas as pd
import numpy as np
from sklearn.preprocessing import MinMaxScaler
import matplotlib as mpl
from sklearn.model_selection import TimeSeriesSplit
import matplotlib.pyplot as plt
from numpy import nan
cv_splits = 3 # time series cross validator
'''
Load dataset
'''
def load_data(path):
raw_data =... | pd.to_datetime(df_uni.index) | pandas.to_datetime |
#import serial
import keras
import pandas as pd
#import serial.tools.list_ports
import os
import numpy as np
from scipy import signal
import matplotlib.pyplot as plt
#from com_serial import *
#from filter import *
from model import *
from sklearn.metrics import confusion_matrix,classification_report
rawdata = []
pipi =... | pd.get_dummies(df['EMOSI']) | pandas.get_dummies |
#code will get the proper values like emyield, marketcap, cacl, etc, and supply a string and value to put back into the dataframe.
import pandas as pd
import numpy as np
import logging
import inspect
from scipy import stats
from dateutil.relativedelta import relativedelta
from datetime import datetime
from scipy import... | pd.isnull([sgaexpense1,sgaexpense2,totalrevenue1,totalrevenue2]) | pandas.isnull |
"""
sess_load_util.py
This module contains functions for loading data from files generated by the
Allen Institute OpenScope experiments for the Credit Assignment Project.
Authors: <NAME>
Date: August, 2018
Note: this code uses python 3.7.
"""
import copy
import logging
from pathlib import Path
import cv2
import... | pd.DataFrame() | pandas.DataFrame |
from config_chbp_eeg import bids_root, deriv_root, N_JOBS
import pandas as pd
from joblib import Parallel, delayed
import mne
import coffeine
from config_chbp_eeg import bids_root
subjects_list = list(bids_root.glob('*'))
subjects_list = list(bids_root.glob('*'))
subjects_df = | pd.read_csv(bids_root / "participants.tsv", sep='\t') | pandas.read_csv |
import os
import json
import pandas as pd
import datetime
import numpy as np
import itertools
from pprint import pprint
from tqdm import tqdm as pbar
from textblob import TextBlob
pd.plotting.register_matplotlib_converters()
#Input: N/A
#Return: A list of strings containing all chat names
def get_chats_names():
ch... | pd.io.json.json_normalize(messages) | pandas.io.json.json_normalize |
import os
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
from decouple import config
BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
def pandas_df(path_to_jsons):
json_files = [pos_json for pos_json in os.listdir(path_to_jsons) if pos_json.endswith('.json')]
js... | pd.read_json(json_file, lines=True) | pandas.read_json |
from utils.support_functions import calculate_rate_exact_day, calculate_rate_exact_day_cop, \
calculate_rate_exact_day_cop_inversed
from decimal import Decimal
import os
import pandas as pd
from settings import CURRENCIES, calc_categories
from statement_parser.preproccessing import get_category
def conversion(x, ... | pd.DataFrame() | pandas.DataFrame |
# -*- coding: utf-8 -*-
import os
import click
import logging
from pathlib import Path
from dotenv import find_dotenv, load_dotenv
import pandas as pd
import networkx as nx
from src.utils.utils_features import NetworkFeatureComputation
from src.data.financial_network import (
IndustryNetworkCreation,
Industry... | pd.DataFrame(adjacency_matrix, index=node_index, columns=node_index) | pandas.DataFrame |
# 국경일 : getHoliDeInfo
# 공휴일 : getRestDeInfo
import requests
from bs4 import BeautifulSoup
import csv
import json
import pandas as pd
import matplotlib.pyplot as plt
import numpy as np
# holiday_csv 파일 만드는 함수
def make_holiday_csv(years_lst=[2018, 2019, 2020, 2021], filename='C:/Users/km_mz/Desktop/dacon/daconcup/Data... | pd.pivot_table(df, index='year_month', columns='day', values=df.columns[1]) | pandas.pivot_table |
from mpl_toolkits import mplot3d
import sys, os
import numpy as np
import matplotlib.pyplot as plt
import pandas as pd
from plotnine import *
import copy, math
dist = 10
def find_min_discm_each_hyperparam(df):
x = df.sort_values(by=['Discm_percent', 'Points-Removed']).groupby("Model-count", as_index=False).first(... | pd.read_csv(f"{dataset}/results_{dataset}_method1.csv") | pandas.read_csv |
# AUTOGENERATED! DO NOT EDIT! File to edit: 02_process_duplicates_image_level.ipynb (unless otherwise specified).
__all__ = ['create_vocab', 'convert_category_lists_to_probability_vectors', 'get_test_csvs', 'get_train_csv',
'get_image_level_csvs']
# Cell
from fastcore.all import *
from .find_duplicates imp... | pd.DataFrame(labeled_test, columns=["image_name", "probability"]) | pandas.DataFrame |
import time
import sys
import datetime
import shutil
import os
import py7zr
from ftplib import FTP
import numpy as np
import pandas as pd
from stock.globalvar import *
from stock.utils.symbol_util import symbol_to_exsymbol, is_symbol_kc
from stock.utils.calc_price import get_zt_price
pd.set_option('max_columns', None)... | pd.DataFrame(columns=["max_matched", "max_unmatched", "zt_seconds", "open_incr"]) | pandas.DataFrame |
# -*- coding: utf-8 -*-
import pandas as pd
def traceZig(bars):
# size = len(bars.index)
last_bar = bars.iloc[0]
go_up = True
go_down = True
zigzags = []
# print(bars)
for t, bar in bars[1:].iterrows():
# 最低价是否低于last_bar True 继续 | False last_bar为zig
if go_down:
if... | pd.DataFrame(zigzags) | pandas.DataFrame |
import typing as T
from pathlib import Path
import defopt
import pandas as pd
import seaborn as sb
import matplotlib.pyplot as plt
import sklearn.metrics as metrics
from sklearn.calibration import calibration_curve
def plot_calibration_curves(dset, ax=None):
if ax is None:
_, ax = plt.subplots()
ax... | pd.Series(scores) | pandas.Series |
#
# Copyright 2015 Quantopian, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in wr... | assert_frame_equal(adjustment_dfs[action_name], exp) | pandas.testing.assert_frame_equal |
# import all the required files i.e. numpy , pandas and math library
from graphlib.financialGraph import Data
import numpy as np
import pandas as pd
from pandas import DataFrame , Series
import math
# All the indicators are defined and arranged in Alphabetical order
# ------------------> A <------------------------
... | pd.Series(s2, name="s2") | pandas.Series |
import pandas as pd
import pyomo.environ as pe
import pyomo.gdp as pyogdp
import os
import matplotlib.pyplot as plt
import matplotlib.cm as cm
from itertools import product
class TheatreScheduler:
def __init__(self, case_file_path, session_file_path):
"""
Read case and session data into Pandas Da... | pd.DataFrame(results) | pandas.DataFrame |
from googleapiclient.discovery import build
from datetime import datetime, timedelta
from pandas import DataFrame, Timedelta, to_timedelta
from structures import Structure
from networkdays import networkdays
from calendar import monthrange
class Timesheet:
def __init__(self, credentials, sheetid):
# The I... | Timedelta("00:00:00") | pandas.Timedelta |
"""
Used examples from SO:
https://stackoverflow.com/questions/22780563/group-labels-in-matplotlib-barchart-using-pandas-multiindex
"""
import pandas as pd
import matplotlib
import matplotlib.pyplot as plt
from itertools import groupby
import numpy as np
def add_line(ax, xpos, ypos):
line = plt.Line... | pd.DataFrame(index=index) | pandas.DataFrame |
# In[2]:
"""
Basic Configurations
"""
configs = {}
configs['cdsign'] = '\\'
configs['path_root'] = 'C:\\VQA_Project\\'
configs['path_datasets'] = configs['path_root'] + 'Datasets' ... | pd.DataFrame(temp) | pandas.DataFrame |
"""
Copyright 2018 <NAME>.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing,
software distribut... | pd.to_datetime('2019-01-20T01:08:00Z') | pandas.to_datetime |
import os
import pathlib
import pickle
import random
import numpy as np
import pandas as pd
from sklearn.decomposition import PCA
from S2S_load_sensor_data import read_data_datefolder_hourfile
from S2S_settings import settings
FPS = settings["FPS"]
FRAME_INTERVAL = settings["FRAME_INTERVAL"]
sample_counts = settings... | pd.concat(df_list, axis=1) | pandas.concat |
import os
import json
import requests
from pathlib import Path
import pandas as pd
from .formatUtil import formatTopStocks
from .crawler import Crawler
from datetime import datetime
import tushare as ts
class Plate(Crawler):
def __init__(self):
super().__init__()
self.__fileBasePath = str(os.path.a... | pd.read_csv(filePath) | pandas.read_csv |
"""
/*
* Copyright (C) 2019-2021 University of South Florida
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required b... | pd.DataFrame.from_dict(matches_dict) | pandas.DataFrame.from_dict |
from __future__ import absolute_import, division, print_function
import logging
import os
import matplotlib.pyplot as plt
import numpy as np
import pandas as pd
import seaborn as sns
import torch
from torch.distributions import constraints
import pyro
import pyro.distributions as dist
from pyro.infer import Empirica... | pd.DataFrame() | pandas.DataFrame |
"""
@<NAME>
==============================================
Training the different models by multiple sequential trials
==============================================
How to train mixtures and HMMs with various observation models on the same dataset.
"""
import bnpy
import numpy as np
import os
from matplotlib import ... | pd.read_csv(folders + '/' + dat_file, sep='\s+', header=None, skiprows=1) | pandas.read_csv |
from collections import OrderedDict
import datetime
from datetime import timedelta
from io import StringIO
import json
import os
import numpy as np
import pytest
from pandas.compat import is_platform_32bit, is_platform_windows
import pandas.util._test_decorators as td
import pandas as pd
from pandas import DataFrame... | read_json(path, encoding=encoding) | pandas.read_json |
import base64
import os, shutil, io, zipfile
from re import L, match
import json
from datetime import datetime, timedelta
from urllib.parse import urljoin
import requests
import pandas as pd
import pint
import numpy as np
#import geopandas as gpd
from django.views.decorators.csrf import csrf_protect, csrf_exempt
from ... | pd.isnull(row['tag']) | pandas.isnull |
# coding: utf-8
# # JDE ETL Source Design
# ## Goal: Generate source SQL with friendly names and built-in data Conversion
# 1. Pull *ALL* Field metadata based on QA 9.3: Name, Datatype, Decimals
# 2. Pull *Specific* Table fields
# 3. Create SQL mapiing pull with data-conversion
# In[254]:
import numpy as np
impor... | pd.read_sql_query(sql_table_fields, engine) | pandas.read_sql_query |
import pandas as pd
import matplotlib.pyplot as plt
import numpy as np
from scipy.stats import spearmanr as sr
from scipy.cluster import hierarchy as hc
from typing import List, Any, Union, Tuple, Optional, Dict
import random, math
# TODO: Remove Dependencies, starting with Sklearn
from sklearn.metrics import roc_curve... | pd.DataFrame(probs) | pandas.DataFrame |
#!/usr/bin/env python
import argparse
import pandas as pd
import numpy as np
import umap
import warnings
import sys
warnings.filterwarnings('ignore')
from bokeh.plotting import figure, output_file, show
from bokeh.layouts import column
from bokeh.core.properties import value
from bokeh.palettes import all_palettes
fro... | pd.concat([raw_data[['FID','IID']], umaped_df], axis=1) | pandas.concat |
from pycountry_convert import country_alpha2_to_continent_code, country_name_to_country_alpha2
import pandas as pd
from sklearn.linear_model import LinearRegression as LR
from datetime import datetime, timedelta
import matplotlib.pyplot as plt
import numpy as np
producers = [
# ' Qatar',
' United States (USA)... | pd.read_csv('./Data/SpotEur.csv', index_col='Date') | pandas.read_csv |
from src.report_generators.base_report_generator import BaseReportGenerator
from src.helpers.preprocess_text import extract_links_from_html
from src.helpers.preprocess_text import extract_from_path
from src.helpers.preprocess_text import extract_subtext
import os
import ast
import re
from bs4 import BeautifulSoup
imp... | pd.isna(content_item['details']) | pandas.isna |
import numpy as np
import pandas as pd
import matplotlib as mpl
import matplotlib.pyplot as plt
import tensorflow as tf
from tensorflow import keras
from sklearn.datasets import fetch_california_housing
from sklearn.model_selection import train_test_split
from sklearn.preprocessing import StandardScaler
housing = fe... | pd.DataFrame(history.history) | pandas.DataFrame |
import asyncio
import threading
import logging
import logging.handlers
import numpy as np
import os
import time
import tqdm
from datetime import datetime
from itertools import zip_longest
from collections import OrderedDict
import pandas as pd
import IPython
from ophyd import Device
from ophyd.status import Status
f... | pd.DataFrame(df_data) | pandas.DataFrame |
from flask import Flask
from flask_restful import Resource, Api, reqparse
import pandas as pd
import ast
app = Flask(__name__)
api = Api(app)
class Users(Resource):
def get(self):
data = pd.read_csv('users.csv') # read local CSV
data = data.to_dict() # convert dataframe to dict
return {... | pd.read_csv('locations.csv') | pandas.read_csv |
from selenium import webdriver
import pandas as pd
from bs4 import BeautifulSoup
import hashlib
import datetime
import multiprocessing as mp
import concurrent as cc
df = pd.DataFrame(columns=['Title','Location','Company','Salary','Sponsored','Description','Time'])
class Indeed():
@staticmethod
... | pd.DataFrame(columns=['Title','Location','Company','Salary','Sponsored','Description','Time']) | pandas.DataFrame |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Wed Aug 5 22:54:58 2020
@author: arti
"""
import pandas as pd
import numpy as np
student1 = pd.Series({'Korean': np.nan, 'English':80, 'Math':90})
student2 = | pd.Series({'Korean':80, 'Math':90}) | pandas.Series |
'''
This script provides code for training a neural network with entity embeddings
of the 'cat' variables. For more details on entity embedding, see:
https://github.com/entron/entity-embedding-rossmann
8-Fold training with 3 averaged runs per fold. Results may improve with more folds & runs.
'''
i... | pd.DataFrame(full_val_preds) | pandas.DataFrame |
import numpy as np
import pandas as pd
from sklearn.model_selection import train_test_split
from causalml.inference.tree import UpliftTreeClassifier
from causalml.inference.tree import UpliftRandomForestClassifier
from causalml.metrics import get_cumgain
from .const import RANDOM_SEED, N_SAMPLE, CONTROL_NAME, TREATME... | pd.DataFrame(y_pred) | pandas.DataFrame |
# Copyright 2020 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, ... | pd.DataFrame({'variable': [0, 1, 2, 3]}, index=[0, 1, 2, 3]) | pandas.DataFrame |
import numpy as np
import pandas as pd
import pymatgen as mg
# %% define filepath constant
TABLE_PATH = "./torrance_tabulated.xlsx"
# %% read in the tables
# read in the tabulated data of closed_shell oxides in table 2 in the original paper
closed_shell_oxides = pd.read_excel(TABLE_PATH, sheet_name="table_2")
# renam... | pd.read_excel(TABLE_PATH, sheet_name="table_3") | pandas.read_excel |
import argparse
import json
import os
import math
import heapq
from sklearn.cluster import KMeans
import pandas as pd
import matplotlib.pyplot as plt
import random
import time
from rich.progress import track
# --roadnetFile Shuanglong.json --dir .\tools\generator
# --roadnetFile roadnet_10_10.json --dir .\tools\genera... | pd.DataFrame(data=intersections, columns=['point']) | pandas.DataFrame |
"""Compilation of functions used for data processing."""
import os
import yaml
from itertools import compress
from datetime import datetime
import pandas as pd
import numpy as np
from ideotype.utils import get_filelist
from ideotype import DATA_PATH
def read_sims(path):
"""
Read and condense all maizsim raw... | pd.Series(issues, dtype='str') | pandas.Series |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
#
# Home Broker API - Market data downloader
# https://github.com/crapher/pyhomebroker.git
#
# Copyright 2020 <NAME>
#
# Licensed under the Apache License, Version 2.0 (the 'License');
# you may not use this file except in compliance with the License.
# You may obtain a cop... | pd.DataFrame(columns=result_columns) | pandas.DataFrame |
# -*- coding: utf-8 -*-
# @Time : 2018/10/3 下午2:36
# @Author : yidxue
import pandas as pd
from common.util_function import *
"""
http://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.concat.html
http://pandas.pydata.org/pandas-docs/stable/user_guide/merging.html
"""
df1 = | pd.DataFrame({'a': ['a', 'c', 'd'], 'b': [4, 6, 7]}) | pandas.DataFrame |
#!/usr/bin/env python
# coding: utf-8
import requests
import pandas as pd
from bs4 import BeautifulSoup
import numpy as np
from alphacast import Alphacast
from dotenv import dotenv_values
API_KEY = dotenv_values(".env").get("API_KEY")
alphacast = Alphacast(API_KEY)
page = requests.get('https://www.indec.gob.ar/Nive... | pd.read_excel(file_xls.content, sheet_name='Cuadro 18', skiprows=3) | pandas.read_excel |
import requests
import pandas as pd
import json
import os
from pandas.io.json import json_normalize #package for flattening json in pandas df
import flatjson
import warnings
warnings.filterwarnings('ignore')
def show_table():
jtoken = os.getenv('GITHUB_TOKEN', '')
ztoken ... | pd.concat([new, bak, prog, peer, gw]) | pandas.concat |
import numpy as np
import pandas as pd
import random
from rpy2.robjects.packages import importr
utils = importr('utils')
prodlim = importr('prodlim')
survival = importr('survival')
#KMsurv = importr('KMsurv')
#cvAUC = importr('pROC')
#utils.install_packages('pseudo')
#utils.install_packages('prodlim')
#utils... | pd.merge(left=long_test_df, right=test_clindata_all, how='left',left_on='ID' ,right_on='ID') | pandas.merge |
import math
from tqdm import tqdm
import pandas as pd
import numpy as np
from sklearn.preprocessing import LabelEncoder
from sklearn.preprocessing import OneHotEncoder, QuantileTransformer
from sklearn.neighbors import NearestNeighbors
from sklearn.compose import ColumnTransformer
from sklearn.pipeline import Pipeline
... | pd.read_excel(data) | pandas.read_excel |
# RESpost.py
import json
from numpy import product
import pandas as pd
import os
import glob
wd = os.getcwd()
simulations_raw = glob.glob(wd+'/cache/2030*.json')
extractor = lambda x: x.split('\\')[-1].replace(".json", '').replace("2030RES_", '')
simulations_keys = [extractor(x) for x in simulations_raw]
#%%
# sim_l... | pd.DataFrame(predf, index=simulations_keys) | pandas.DataFrame |
import argparse
import os
import shutil
import zipfile
import pathlib
import re
from datetime import datetime
import collections
import pandas as pd
import geohash
import math
import helpers
import plotly.express as px
ControlInfo = collections.namedtuple("ControlInfo", ["num_tracks", "date", "duration"])
def parse_... | pd.concat([pdf, edf]) | pandas.concat |
import pandas as pd
from sankeyview.sankey_definition import SankeyDefinition, Ordering, ProcessGroup, Waypoint, Bundle
from sankeyview.sankey_view import sankey_view
from sankeyview.partition import Partition
from sankeyview.dataset import Dataset
def test_sankey_view_accepts_dataframe_as_dataset():
nodes = {
... | pd.DataFrame({'id': ['a1', 'b1']}) | pandas.DataFrame |
import datetime
import pandas as pd
from local_group_support.config.config import get_config
from rebel_management_utilities.action_network import get_forms, query, query_all
FORMATION_DATE = datetime.date(2018, 4, 1)
def get_form(submission):
form_id = submission['action_network:form_id']
has_website = 'a... | pd.to_datetime(submission['created_date']) | pandas.to_datetime |
import wx.grid as gridlib
import pandas as pd
import numpy as np
import copy
import ciw
import re
import math
import statistics
import random
import imp
adapt = imp.load_source('adapt', 'src/adapt.py')
summary = imp.load_source('summary', 'src/Summary.py')
cluster = imp.load_source('cluster', 'src/clustering.py')
tra... | pd.DataFrame({'Activity': letters}) | pandas.DataFrame |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.