prompt stringlengths 19 1.03M | completion stringlengths 4 2.12k | api stringlengths 8 90 |
|---|---|---|
import dask.dataframe as dd
import numpy as np
import pandas as pd
from pandas.api.types import is_categorical_dtype
DEFAULT_WINDOW = 7
DEFAULT_TAKE_LOGS = True
DEFAULT_CENTER = False
DEFAULT_MIN_PERIODS = 1
def calculate_weekly_incidences_from_results(
results,
outcome,
groupby=None,
):
"""Create th... | pd.concat(period_outcomes) | pandas.concat |
"""Script to add interval on GVA and population file
Run script on 'data' folder in scenarios_not_extracted folder
"""
import os
import pandas as pd
import numpy as np
from energy_demand.basic import lookup_tables
from energy_demand.basic import basic_functions
def run(
path_to_folder,
path_MSOA_basel... | pd.read_csv(file_path) | pandas.read_csv |
""" Model for output of general/metadata data, useful for a batch """
import logging
from pathlib import Path
from typing import List, Optional, Union
import pandas as pd
from pydantic import BaseModel, Field, validator
from nowcasting_dataset.consts import SPATIAL_AND_TEMPORAL_LOCATIONS_OF_EACH_EXAMPLE_FILENAME
fro... | pd.read_csv(filename) | pandas.read_csv |
from datetime import datetime
import pandas as pd
import numpy as np
from extract import PreProcess
from shapely.geometry import Point
from shapely.geometry.polygon import Polygon
import geopandas as gpd
import fun_logger
log = fun_logger.init_log()
prepros = PreProcess()
fmt = '%Y-%m-%d %H:%M:%S'
class AnalyzeDf()... | pd.DataFrame() | pandas.DataFrame |
import numpy as np
import astropy.units as u
import pandas as pd
def deg2mas(value):
'''
Converts value from degree to milliarcseconds
value: a value in degree
'''
value_mas = (value * u.degree).to(u.mas).value
return value_mas
def time_diff(catalog):
"""
Calculates the time differ... | pd.DataFrame({'delta_days': delta_days}) | pandas.DataFrame |
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
from datetime import datetime
from sklearn.base import BaseEstimator
from sklearn.pipeline import Pipeline
from sklearn.metrics import (accuracy_score, precision_score, recall_score, roc_auc_score,
f1_score, roc_curve, ... | pd.DataFrame(y_test.values, columns=["True Label"]) | pandas.DataFrame |
import os
import logging
import copy
import numpy as np
import pandas as pd
from oemof.solph import EnergySystem, Bus, Sink, Source
import oemof.tabular.tools.postprocessing as pp
from oemof.tools.economics import annuity
from oemof_flexmex.helpers import delete_empty_subdirs, load_elements, load_scalar_input_data,\
... | pd.concat(carrier_cost) | pandas.concat |
import streamlit as st
import pandas as pd
import altair as alt
import numpy as np
from datetime import datetime, timedelta
from dateutil import parser
from utils import suffix, custom_strftime
population = 68134973
alt.themes.enable('fivethirtyeight')
latest_date = parser.parse("2021-04-07")
dose1 = pd.read_csv(f"d... | pd.to_datetime(dose1.date) | pandas.to_datetime |
#!/usr/bin/env python3
from pathlib import Path
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
from itertools import product
import statsmodels.api as sm
from statsmodels.tsa.arima.model import ARIMA
from statsmodels.tsa.arima_process import arma_generate_sample
from rom_plots import detrend
imp... | pd.unique(df.YEAR) | pandas.unique |
"""
Utils to plot graphs with arrows
"""
import matplotlib.transforms
import matplotlib.patches
import matplotlib.colors
import matplotlib.cm
import numpy as np
import pandas as pd
import logging
from tctx.util import plot
def _clip_arrows(arrows, tail_offset, head_offset):
"""
shorten head & tail so the ... | pd.DataFrame.from_dict(pos, orient='index', columns=['x', 'y']) | pandas.DataFrame.from_dict |
# -*- coding: utf-8 -*-
"""
Created on Tue Aug 28 22:50:43 2018
@author: kennedy
"""
"""
Credit:
https://www.quantopian.com/posts/technical-analysis-indicators-without-talib-code
Bug Fix by Kennedy:
Works fine for library import.
returns only column of the indicator r... | pd.Series(DoI) | pandas.Series |
import blpapi
import logging
import datetime
import pandas as pd
import contextlib
from collections import defaultdict
from pandas import DataFrame
@contextlib.contextmanager
def bopen(debug=False):
con = BCon(debug=debug)
con.start()
try:
yield con
finally:
con.stop()
class BCon(obj... | DataFrame(data) | pandas.DataFrame |
from __future__ import annotations
import numbers
from typing import TYPE_CHECKING
import warnings
import numpy as np
from pandas._libs import (
lib,
missing as libmissing,
)
from pandas._typing import (
ArrayLike,
Dtype,
type_t,
)
from pandas.compat.numpy import function as nv
from pandas.core.... | is_bool_dtype(result) | pandas.core.dtypes.common.is_bool_dtype |
import numpy as np
import pandas as pd
import pickle
import scipy.sparse
import tensorflow as tf
from typing import Union, List
import os
from tcellmatch.models.models_ffn import ModelBiRnn, ModelSa, ModelConv, ModelLinear, ModelNoseq
from tcellmatch.models.model_inception import ModelInception
from tcellmatch.estimat... | pd.DataFrame({"label": self.label_ids}) | pandas.DataFrame |
from stockscore.data import Stocks, return_top
import pandas as pd
import pytest
symbols = ["FB", "AAPL", "AMZN", "NFLX", "GOOGL"]
stocks = Stocks(symbols)
tdata = {
"Score": [6, 5, 4, 3, 2],
"Value Score": [1, 2, 3, 1, 0],
"Growth Score": [3, 2, 0, 1, 2],
"Momentum Score": [2, 1, 1, 1, 0],
}
tscores ... | pd.DataFrame(tdata, index=symbols) | pandas.DataFrame |
from xml.parsers.expat import model
import numpy as np
import pandas as pd
import plotly.express as px
import streamlit as st
import os
from joblib import dump, load
import sklearn
from sklearn.ensemble import RandomForestClassifier
from sklearn import metrics
import pickle
# Titulo do app
st.write("""
# Prevendo oco... | pd.read_csv('cardio_app2.csv') | pandas.read_csv |
# -*- coding: utf-8 -*-
# (c) <NAME>, see LICENSE.rst.
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import logging
from ciso8601 import parse_datetime
from lxml import etree
from pytz import FixedOffset
import num... | pd.DataFrame(meta_data) | pandas.DataFrame |
"""Dynamic file checks."""
from dataclasses import dataclass
from datetime import date, timedelta
from typing import Dict, Set
import re
import pandas as pd
import numpy as np
from .errors import ValidationFailure, APIDataFetchError
from .datafetcher import get_geo_signal_combos, threaded_api_calls
from .utils import r... | pd.isna(frame["ftstat"]) | pandas.isna |
# Importar librerias
import pandas # importar libreria pandas
import time # importar libreria time
import datetime ... | pandas.DataFrame(registros) | pandas.DataFrame |
from ..pyhrp.tools.distancematrices import CorrDistance, PortfolioDistance, LTDCDistance
from math import fabs, sqrt, log
import pandas as pd
import numpy as np
def test_CorrDistance_get_distance_matrix():
d = {'A': [0.5, 1, 0], 'B': [0, -1, 0.5], 'C':[-0.5, 1, 0]}
df = pd.DataFrame(data=d)
corr = df.corr(... | pd.DataFrame(data=d) | pandas.DataFrame |
'''
pyjade
A program to export, curate, and transform data from the MySQL database used by the Jane Addams Digital Edition.
'''
import os
import re
import sys
import json
import string
import datetime
import mysql.connector
from diskcache import Cache
import pandas as pd
import numpy as np
from bs4 import Beautiful... | pd.read_sql(statement,DB) | pandas.read_sql |
import pandas as pd
from assistants.compliance.util import fields
from assistants.compliance.util.input_data_utility import completed_to_due_vector
dt_new_mogl = pd.Timestamp(2020, 9, 15) # September 2020 MOGL changes
def fix_some_dates(data: pd.DataFrame) -> pd.DataFrame:
"""Fix Some Dates.
Make sure if ... | pd.Series(False, index=data.index) | pandas.Series |
#!/usr/bin/env python
# -*- coding: utf-8; -*-
# Copyright (c) 2020, 2022 Oracle and/or its affiliates.
# Licensed under the Universal Permissive License v 1.0 as shown at https://oss.oracle.com/licenses/upl/
from __future__ import print_function, absolute_import
import os
import re
import warnings
import oci
import... | pd.DataFrame() | pandas.DataFrame |
import matplotlib.pyplot as plt
from sklearn.linear_model import LinearRegression
from sklearn.preprocessing import PolynomialFeatures
import numpy as np
from pylab import rcParams
##########################################################################################
# Designed and developed by <NAME>
# Date : 11 ... | pd.DataFrame(df1) | pandas.DataFrame |
# importação de bibliotecas
import pandas as pd
# carrega um arquivo do HD para a memoria
data = pd.read_csv('data/kc_house_data.csv')
# mostrar na tela as primeiras 6 linhas
# print(data.head())
# fução que converte de object (string) -> date
data['date'] = | pd.to_datetime(data['date']) | pandas.to_datetime |
import sqlite3
import pandas as pd
import pandas.io.sql as psql
import ast
import hashlib
import sys
import random
from sqlalchemy import create_engine
import sqlalchemy.types as dtype
import requests
# message: |channel|user|text| <= ts
# coin : val <= username
slack_columns = ["ts", "text", "user", "channel"]
slack... | pd.DataFrame(data, columns=slack_columns, index=None) | pandas.DataFrame |
import numpy as np # linear algebra
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import seaborn as sns
import matplotlib.pyplot as plt
from sklearn.model_selection import train_test_split
from sklearn import preprocessing
from sklearn.utils import resample
from sklearn.metrics import accuracy... | pd.read_csv('thorp.csv') | pandas.read_csv |
import pandas as pd
import os
def prepare_legends(mean_models, models, interpretability_name):
bars = []
y_pos = []
index_bars = 0
for nb, i in enumerate(mean_models):
if nb % len(models) == int(len(models)/2):
bars.append(interpretability_name[index_bars])
index_bars +=... | pd.DataFrame(columns=self.columns_name_file3) | pandas.DataFrame |
"""
Data: Temperature and Salinity time series from SIO Scripps Pier
Salinity: measured in PSU at the surface (~0.5m) and at depth (~5m)
Temp: measured in degrees C at the surface (~0.5m) and at depth (~5m)
- Timestamp included beginning in 1990
"""
# imports
import sys,os
import pandas as pd
import numpy as np
im... | pd.read_csv('/Users/MMStoll/Python/Data/Ocean569_Data/SIO_Data/SIO_TEMP_1916_201905.txt', sep='\t', skiprows = 26) | pandas.read_csv |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
test_hydrofunctions
----------------------------------
Tests for `hydrofunctions` module.
"""
from __future__ import (
absolute_import,
print_function,
division,
unicode_literals,
)
from unittest import mock
import unittest
import warnings
from pandas... | pd.Timedelta("1 day 1 hour 2 minutes") | pandas.Timedelta |
import pandas as pd
import numpy as np
from preprocess import process_examples, get_requests_from_logs
from scipy.spatial import distance
import logging
log = logging.getLogger(__name__)
log.setLevel(logging.DEBUG)
def _process(data, max_attributes, restructure):
processed = process_examples(data, max_attributes,... | pd.get_dummies(processed) | pandas.get_dummies |
import pandas as pd
import sys,os,io,re
import numpy as np
path=sys.argv[1]
outName=sys.argv[2]
thresh=int(sys.argv[3])
anno_file=sys.argv[4]
anno_table= | pd.read_csv(anno_file) | pandas.read_csv |
#! /usr/bin/env python
import os
import pandas as pd
import numpy as np
import tensorflow as tf
from tensorflow.contrib import learn
from bson.objectid import ObjectId
from sentiment_analysis.tweet_preprocessing import preprocess_tweets
from sentiment_analysis.data_helpers import batch_iter
from config import config
f... | pd.DataFrame(posts) | pandas.DataFrame |
import pandas as pd
import numpy as np
import math
from datetime import datetime
from dateutil import parser
import csv
import urllib2
import sys
import pytz
url = 'https://www.alphavantage.co/query?function=TIME_SERIES_INTRADAY&symbol=NSE:RELIANCE&interval=1min&datatype=csv&outputsize=full&apikey=<KEY>'
response = ... | pd.read_hdf('StockMarketData_2020-02-14.h5', key='/RELIANCE__EQ__NSE__NSE__MINUTE') | pandas.read_hdf |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Tue Oct 26 14:26:51 2021
@author: michaeltown
"""
## beginning of module 1 MVP data analysis
import numpy as np
import pandas as pd
import os as os
import datetime as dt
import matplotlib.pyplot as plt
import seaborn as sns
## revised EDA project to look... | pd.read_csv(dataFileLoc1) | pandas.read_csv |
# -*- coding: utf-8 -*-
import pytest
import numpy as np
import pandas as pd
import pandas.util.testing as tm
import pandas.compat as compat
###############################################################
# Index / Series common tests which may trigger dtype coercions
###############################################... | pd.Timestamp('2011-01-04') | pandas.Timestamp |
import altair as alt
import pandas as pd
import streamlit as st
import coin_metrics
# Get data from coin metrics API
@st.cache
def get_data(asset_id, metrics, asset_name):
rates = coin_metrics.get_reference_rates_pandas(asset_id, metrics)
df = | pd.DataFrame(data=rates) | pandas.DataFrame |
# load in libraries
from bs4 import BeautifulSoup
import pandas as pd
import time
from selenium import webdriver
# %% set up selenium
from selenium import webdriver
driver = webdriver.Firefox()
url1 = 'https://freida.ama-assn.org/search/list?spec=43236&page=1'
driver.get(url1)
# %% define standard parse function
def ... | pd.DataFrame.from_dict(itemdict) | pandas.DataFrame.from_dict |
from enum import Enum
import sys
import os
import re
from typing import Any, Callable, Tuple
from pandas.core.frame import DataFrame
from tqdm import tqdm
import yaml
from icecream import ic
SCRIPT_DIR = os.path.dirname(os.path.abspath(__file__))
sys.path.append(os.path.dirname(SCRIPT_DIR))
from argpar... | pd.DataFrame.sparse.from_spmatrix(sprs, columns=columns) | pandas.DataFrame.sparse.from_spmatrix |
import numpy as np
import pandas as pd
import tensorflow as tf
import matplotlib.pyplot as plt
"""
Created by <NAME> on 4/2/18.
Email : <EMAIL>
Website: http://ce.sharif.edu/~naghipourfar
"""
import numpy as np
import pandas as pd
import tensorflow as tf
import matplotlib.pyplot as plt
import keras
from... | pd.get_dummies(all_data) | pandas.get_dummies |
#!/usr/bin/env python
# coding: utf-8
# # ReEDS Scenarios on PV ICE Tool
# To explore different scenarios for furture installation projections of PV (or any technology), ReEDS output data can be useful in providing standard scenarios. ReEDS installation projections are used in this journal as input data to the PV ICE... | pd.DataFrame() | pandas.DataFrame |
import pandas as pd
import numpy as np
from tqdm import tqdm
import matplotlib.pyplot as plt
import seaborn as sns
import shap
from sklearn.metrics import mean_squared_error
from sklearn.metrics import mean_absolute_error
from sklearn.metrics import r2_score
# from .utils import Boba_Utils as u
class Boba_Model_Diagn... | pd.qcut(y_temp['predicted'], 10) | pandas.qcut |
# %%
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
data = {
'bids_regdup': pd.read_csv('data/as_bids_REGUP.csv'),
'bids_regdown': pd.read_csv('data/as_bids_REGDOWN.csv'),
'plans': pd.read_csv('data/as_plan.csv'),
'energy_prices': | pd.read_csv('data/energy_price.csv') | pandas.read_csv |
# pylint: disable=E1101,E1103,W0232
from datetime import datetime, timedelta
from pandas.compat import range, lrange, lzip, u, zip
import operator
import re
import nose
import warnings
import os
import numpy as np
from numpy.testing import assert_array_equal
from pandas import period_range, date_range
from pandas.c... | Int64Index([1, 2, 5, 7, 12, 25]) | pandas.core.index.Int64Index |
from datetime import datetime, date
import sys
if sys.version_info >= (2, 7):
from nose.tools import assert_dict_equal
import xlwings as xw
try:
import numpy as np
from numpy.testing import assert_array_equal
def nparray_equal(a, b):
try:
assert_array_equal(a, b)
except Asse... | pd.MultiIndex.from_arrays([['a', 'a', 'b'], [1., 2., 1.]]) | pandas.MultiIndex.from_arrays |
import numpy as np
import pandas as pd
import pytest
import woodwork as ww
from evalml.data_checks import (
ClassImbalanceDataCheck,
DataCheckError,
DataCheckMessageCode,
DataCheckWarning,
)
class_imbalance_data_check_name = ClassImbalanceDataCheck.name
def test_class_imbalance_errors():
X = pd.... | pd.Series([200] * 10) | pandas.Series |
"""
Base class for a runnable script
"""
import pandas as pd
import numpy as np
from .. import api as mhapi
import os
from ..utility import logger
class Processor:
def __init__(self, verbose=True, violate=False, independent=True):
self.verbose = verbose
self.independent = independent
self.violate = violate
... | pd.DataFrame() | pandas.DataFrame |
"""--------------------------------------------------------------------------------------------------------------------
Copyright 2021 Market Maker Lite, LLC (MML)
Licensed under the Apache License, Version 2.0
THIS CODE IS PROVIDED AS IS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND
This file is part of the MML Op... | pd.DataFrame(symbol_df[0:][0]) | pandas.DataFrame |
#!/usr/bin/env python
# coding: utf-8
from sklearn.feature_extraction.text import CountVectorizer
import numpy as np
from scipy.stats import entropy
import pickle
import os
import json
from flask import Flask
from flask import request
from jinja2 import Template
import pandas as pd
from sklearn.ensemble import RandomF... | pd.concat([self.all_data, self.labelled]) | pandas.concat |
# -*- coding: utf-8 -*-
import csv
import os
import platform
import codecs
import re
import sys
from datetime import datetime
import pytest
import numpy as np
from pandas._libs.lib import Timestamp
import pandas as pd
import pandas.util.testing as tm
from pandas import DataFrame, Series, Index, MultiIndex
from pand... | tm.assert_frame_equal(df, expected) | pandas.util.testing.assert_frame_equal |
# -*- coding: utf-8 -*-
"""
Created on Fri Mar 1 13:37:10 2019
@author:Imarticus Machine Learning Team
"""
import numpy as np
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import matplotlib.pyplot as plt
import seaborn as sns
color = sns.color_palette()
pd.options.mode.chain... | pd.read_csv("order_products_prior.csv") | pandas.read_csv |
#! /usr/bin/env python
##! /usr/bin/arch -x86_64 /usr/bin/env python
from logging import error
import dash
import dash_table
import dash_core_components as dcc
import dash_html_components as html
import dash_bootstrap_components as dbc
import plotly.express as px
import plotly.graph_objects as go
import pandas as pd
f... | is_numeric_dtype(df[groups[1]]) | pandas.api.types.is_numeric_dtype |
from pymongo import MongoClient
import pandas as pd
from collections import Counter
# NLP libraries
from nltk.tokenize import TweetTokenizer
from nltk.corpus import stopwords
import string
import csv
import json
# from datetime import datetime
import datetime
from collections import deque
import pymongo
"""TIME SERI... | pd.Series(ones, index=idx) | pandas.Series |
#!/usr/bin/python
# <EMAIL>
#====================SET============================#
C_END = "\033[0m"
C_BOLD = "\033[1m"
C_RED = "\033[31m"
#==================================================#
import inspect
import sys
import os
import glob
import pandas as pd
import numpy as np
def DebugPrinter(arg):
... | pd.read_csv(result) | pandas.read_csv |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Mon Jul 10 04:11:27 2017
@author: konodera
nohup python -u 501_concat.py &
"""
import pandas as pd
import numpy as np
from tqdm import tqdm
import multiprocessing as mp
import gc
import utils
utils.start(__file__)
#======================================... | pd.merge(df, organic, on='product_id', how='left') | pandas.merge |
#!/usr/bin/env python
# coding: utf-8
# # OpenMC Program for BurnUp analysis and Benchmarking
#
# In[1]:
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Sun Apr 7 11:40:23 2020
@author: feryantama
"""
# ## 1. Initialization
#
# Initialize package we used in this program
# In[2]:
import numpy ... | pd.DataFrame(data=count_list,columns=['label'],index=Pebbledf.index) | pandas.DataFrame |
#!/usr/bin/env python
# coding: utf-8
import click
import datetime
import locale
import numpy as np
import os
from pathlib import Path
import pandas as pd
import plotly.graph_objects as go
CSV_NAME_MAP = {
"pl_PL": {
"p2.5": "dolne 2.5% modelowań",
"p25": "dolne 25% modelowań",
"p75": "gór... | pd.read_csv('https://raw.githubusercontent.com/KITmetricslab/covid19-forecast-hub-de/master/data-truth/MZ/truth_MZ-Incident%20Cases_Poland.csv') | pandas.read_csv |
"""
NAME : Molecular Arrangement and Fringe Identification Analysis from Molecular Dynamics (MAFIA-MD)
AUTHORS : <NAME>, Dr. <NAME> and Dr. <NAME>
MAFIA-MD is a post-processing utility to capture ring structures from molecular trajectory files (.xyz) generated by
reactive molecular dynamics simulation of Hydrocarbon... | pd.DataFrame({'Atoms. Timestep: 0'}, index=[0]) | pandas.DataFrame |
# -*- coding: utf-8 -*-
# ---
# jupyter:
# jupytext:
# text_representation:
# extension: .py
# format_name: light
# format_version: '1.5'
# jupytext_version: 1.7.1
# kernelspec:
# display_name: Python 3
# language: python
# name: python3
# ---
# + [markdown] papermill={"dura... | pd.read_gbq(query, dialect='standard') | pandas.read_gbq |
import pandas as pd
def process(feats, out_overlap, out_weights, out_feats):
#read features
overlap = pd.DataFrame(pd.np.zeros([0, 6]))
overlap.columns = ["drug1", "drug2", "mode", "overlap", "no_feats1", "no_feats2"]
weights = pd.DataFrame( | pd.np.zeros([0, 5]) | pandas.np.zeros |
import sys
import pandas as pd
import numpy as np
from sklearn.base import BaseEstimator
from sklearn.model_selection import ParameterGrid
class EVI(BaseEstimator):
"""Class for evaluating multi-modal data integration approaches for combining unspliced, spliced, and RNA velocity gene expression modalities
... | pd.DataFrame() | pandas.DataFrame |
""" Profile a single GConv layer """
import os
import sys
import argparse
import copy
import time
import shutil
import json
import logging
logging.getLogger().setLevel(logging.DEBUG)
import pandas as pd
import numpy as np
import torch
import torch.nn as nn
import torch.nn.parallel
import torch.back... | pd.DataFrame(data, columns=columns) | pandas.DataFrame |
""" test the scalar Timedelta """
from datetime import timedelta
import numpy as np
import pytest
from pandas._libs import lib
from pandas._libs.tslibs import (
NaT,
iNaT,
)
import pandas as pd
from pandas import (
Timedelta,
TimedeltaIndex,
offsets,
to_timedelta,
)
import pandas._testing as ... | Timedelta("-1 days, 10:11:12") | pandas.Timedelta |
from flask import Flask, render_template, request, session, redirect, url_for
from datetime import datetime, timedelta
import pandas as pd
import sqlite3, hashlib, os, random, os, dotenv
app = Flask(__name__)
app.secret_key = "super secret key"
dotenv.load_dotenv()
MAPBOX_TOKEN = os.getenv('MAPBOX_TOKEN')
conn = sqlit... | pd.read_sql("select * from w_sales_history", conn) | pandas.read_sql |
import numpy as np
import pandas as pd
from settings.config import RECOMMENDATION_LIST_SIZE, KL_LABEL, HE_LABEL, CHI_LABEL, FAIRNESS_METRIC_LABEL, \
VARIANCE_TRADE_OFF_LABEL, \
COUNT_GENRES_TRADE_OFF_LABEL, TRADE_OFF_LABEL, evaluation_label, MACE_LABEL, FIXED_LABEL, MAP_LABEL, \
MRR_LABEL, order_label, MC_... | pd.DataFrame() | pandas.DataFrame |
# %% load in libraries
from bs4 import BeautifulSoup
import pandas as pd
import time
from selenium import webdriver
import random
import numpy as np
# %% set up selenium
from selenium import webdriver
driver = webdriver.Firefox()
# %%
driver.get('https://www.doximity.com/residency/programs/009b631d-3390-4742-b583-820... | pd.read_csv('specialties_doximity.csv') | pandas.read_csv |
import re
import fnmatch
import os, sys, time
import pickle, uuid
from platform import uname
import pandas as pd
import numpy as np
import datetime
from math import sqrt
from datetime import datetime
import missingno as msno
import statsmodels.api as sm
from statsmodels.tsa.seasonal import seasonal_decompose
from stat... | pd.DataFrame.from_dict({'Actual': model.actual, 'Prediction': model.pred}) | pandas.DataFrame.from_dict |
import operator
import numpy as np
import pytest
import pandas as pd
import pandas._testing as tm
from pandas.core.arrays import FloatingArray
import pandas.core.ops as ops
# Basic test for the arithmetic array ops
# -----------------------------------------------------------------------------
@pytest.mark.paramet... | pd.Series([1, 2, 3], dtype="Int64") | pandas.Series |
import streamlit as st
import pandas as pd
import yfinance as yf
import datetime
import os
from pathlib import Path
import requests
import hvplot.pandas
import numpy as np
import matplotlib.pyplot as plt
from MCForecastTools_2Mod import MCSimulation
import plotly.express as px
from statsmodels.tsa.arima_model import ... | pd.Series(conf[:, 1], index=empty_df.index) | pandas.Series |
# -*- coding: utf-8 -*-
"""
Created on Wed Mar 1 16:06:36 2017
@author: Gonxo
This file is merges the different grid and solar feed into a single feed. Also
it deseasonalizes the data in hour of the daya, day of the week, and month of
the year. Finally it computes ANOVA tests to check seasonal variations significan... | pd.HDFStore(sourceFile) | pandas.HDFStore |
import os
import unittest
import random
import sys
import site # so that ai4water directory is in path
ai4_dir = os.path.dirname(os.path.dirname(os.path.abspath(sys.argv[0])))
site.addsitedir(ai4_dir)
import scipy
import numpy as np
import pandas as pd
from sklearn.model_selection import train_test_split
from ai4wa... | pd.date_range('20110101', periods=35, freq='D') | pandas.date_range |
import numpy as np
import pandas as pd
from lightgbm import LGBMClassifier, LGBMRegressor
from sklearn.metrics import roc_auc_score
from sklearn.model_selection import GroupKFold
from src.python.space_configs import space_lightgbm, tune_model, gp_minimize, forest_minimize
from ml_metrics import rmse
train_df = pd.rea... | pd.concat([importances, imp_df], axis=0, sort=False) | pandas.concat |
# 导入类库
import os.path
import os
import datetime
import lightgbm as lgb
from sklearn.model_selection import train_test_split
from gensim.models import Word2Vec
from pandas import read_csv
import re
import pandas as pd
import numpy as np
import itertools
import sys
def get_dict(file): ... | pd.read_csv('result/' + i) | pandas.read_csv |
# Author: <NAME> and <NAME>
# Plant and Food Research New Zealand and UNSW Sydney
#!/usr/bin/env python3
#! module load pfr-python3/3.6.5
print('Here we go...')
import sys
import pyqrcode
import os, os.path # For PATH etc. https://docs.python.org/2/library/os.html
import sys # For handling command line args
from... | pd.DataFrame() | pandas.DataFrame |
# -*- coding: utf-8 -*-
import click
import logging
from pathlib import Path
# from dotenv import find_dotenv, load_dotenv
import requests
from bs4 import BeautifulSoup
import numpy as np
import pandas as pd
import datetime
import yfinance as yf
from pandas_datareader import data as pdr
from flask import current_app
f... | pd.Series(df['log_ret_1d']) | pandas.Series |
from datetime import (
datetime,
timedelta,
)
import numpy as np
import pytest
from pandas.compat import (
pa_version_under2p0,
pa_version_under4p0,
)
from pandas.errors import PerformanceWarning
from pandas import (
DataFrame,
Index,
MultiIndex,
Series,
isna,
)
import pandas._tes... | Series(["foo", "bar"]) | pandas.Series |
# -*- coding: utf-8 -*-
import pytest
import numpy as np
import pandas as pd
import pandas.util.testing as tm
import pandas.compat as compat
###############################################################
# Index / Series common tests which may trigger dtype coercions
###############################################... | pd.Timestamp('2012-01-01') | pandas.Timestamp |
from typing import Any
from typing import Dict
from typing import Optional
import pandas
import pytest
from evidently.model_profile.sections.classification_performance_profile_section import \
ClassificationPerformanceProfileSection
from .helpers import calculate_section_results
from .helpers import check_profil... | pandas.DataFrame({'target': [1, 1, 3, 3], 'prediction': [1, 2, 1, 4]}) | pandas.DataFrame |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
@author: abhijit
"""
#%% preamble
import numpy as np
import pandas as pd
from glob import glob
#%% Tidy data
filenames = glob('data/table*.csv')
filenames = sorted(filenames)
table1, table2, table3, table4a, table4b, table5 = [pd.read_csv(f) for f in filenames] ... | pd.read_csv('data/pew.csv') | pandas.read_csv |
import json
import pandas as pd
from sklearn.compose import ColumnTransformer
from sklearn.feature_extraction.text import TfidfVectorizer
from sklearn.ensemble import RandomForestClassifier
from sklearn.metrics import f1_score
class SEIssueRF:
def __init__(self, data):
self.data = data
de... | pd.DataFrame({"corpus": fitting_data}) | pandas.DataFrame |
import pandas as pd
import numpy as np
from datetime import timedelta, datetime
from sys import argv
dates=("2020-04-01", "2020-04-08", "2020-04-15", "2020-04-22",
"2020-04-29" ,"2020-05-06", "2020-05-13","2020-05-20", "2020-05-27", "2020-06-03",
"2020-06-10", "2020-06-17", "2020-06-24", "2020-07-01", "2020-07-08",
... | pd.DataFrame.from_dict(sims_dict) | pandas.DataFrame.from_dict |
#!/usr/bin/env python
# coding: utf-8
# <b>Python Scraping of Book Information</b>
# In[1]:
get_ipython().system('pip install bs4')
# In[2]:
get_ipython().system('pip install splinter')
# In[3]:
get_ipython().system('pip install webdriver_manager')
# In[1]:
# Setup splinter
from splinter import Browser
... | pd.read_csv('greek-roman-clean.csv') | pandas.read_csv |
"""MVTecAd Dataset."""
# default packages
import dataclasses as dc
import enum
import logging
import pathlib
import shutil
import sys
import tarfile
import typing as t
import urllib.request as request
# third party packages
import pandas as pd
# my packages
import src.data.dataset as ds
import src.data.utils as ut
#... | pd.read_csv(self.train_list) | pandas.read_csv |
# -*- coding: utf-8 -*-
"""
Created on Fri Nov 17 14:10:58 2017
@author: tkc
"""
import os
import pandas as pd
from tkinter import filedialog
AESQUANTPARAMFILE='C:\\Users\\tkc\\Documents\\Python_Scripts\\Augerquant\\Params\\AESquantparams.csv'
class AESspectrum():
''' Single instance of AES spectra file created f... | pd.read_csv('Backfitlog.csv', encoding='cp437') | pandas.read_csv |
import pandas as pd
import datetime
import numpy as np
class get_result(object):
def __init__(self, data, material, start_time):
self.data = data
self.material = material
self.start_time = start_time
self.freq = 0.5 # 单位:小时
ele_struct = pd.read_excel("尖峰平谷电费结构.xlsx", ... | pd.DataFrame(index=idx, columns=['ele'], data=ygdn) | pandas.DataFrame |
import re
from datetime import datetime, timedelta
import numpy as np
import pandas.compat as compat
import pandas as pd
from pandas.compat import u, StringIO
from pandas.core.base import FrozenList, FrozenNDArray, DatetimeIndexOpsMixin
from pandas.util.testing import assertRaisesRegexp, assert_isinstance
from pandas i... | tm.assert_series_equal(hist, expected) | pandas.util.testing.assert_series_equal |
import time
import pandas as pd
import numpy as np
CITY_DATA = { 'chicago': 'chicago.csv',
'new york city': 'new_york_city.csv',
'washington': 'washington.csv' }
months = ['all','january','february','march','april','may','june','july','august','september','october','december']
def get_fi... | pd.read_csv(CITY_DATA[city]) | pandas.read_csv |
import pandas as pd
import networkx as nx
import logging
import math
import numpy as np
from statsmodels.stats.outliers_influence import variance_inflation_factor
def get_vif(df: pd.DataFrame, threshold: float = 5.0):
"""
Calculates the variance inflation factor (VIF) for each feature column. A VIF
value... | pd.DataFrame() | pandas.DataFrame |
#
# Copyright (C) 2019 Databricks, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to i... | pd.get_dummies(df.d) | pandas.get_dummies |
"""
This module allows to collect experimental variables from fits
to data that can then be used as input to simulations
"""
# Author: <NAME>, <NAME>, 2019
# License: MIT License
import numpy as np
import pandas as pd
import scipy.optimize
import scipy.stats
import colicycle.time_mat_operations as tmo
import coli... | pd.read_pickle(file_to_load) | pandas.read_pickle |
"""
library for simulating semi-analytic mock maps of CMB secondary anisotropies
"""
__author__ = ["<NAME>", "<NAME>"]
__email__ = ["<EMAIL>", "<EMAIL>"]
import os
import warnings
from sys import getsizeof
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
from matplotlib import cm
from warnings ... | pd.DataFrame(catalog) | pandas.DataFrame |
import os
os.environ['TF_CPP_MIN_LOG_LEVEL'] = '2'
import tensorflow as tf
from sklearn.model_selection import StratifiedKFold
import numpy as np
import csv
import re
import pickle
import time
from datetime import timedelta
import pandas as pd
from pathlib import Path
import sys
sys.path.insert(0,'/nfs/ghome/live/yas... | pd.DataFrame.from_dict(expdata) | pandas.DataFrame.from_dict |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
__author__ = 'ipetrash'
# SOURCE: https://stackoverflow.com/questions/7961363/removing-duplicates-in-lists
from collections import OrderedDict
from functools import reduce
# pip install pandas
import pandas as pd
# pip install numpy
import numpy as np
def remove_d... | pd.unique(items) | pandas.unique |
# standard libraries
import enum
import glob
import os
import warnings
import zipfile
# third-party libraries
import matplotlib.pyplot as plt
import natsort
import pandas
def get_align_count_pipelines():
return enum.Enum('align_count_pipeline', 'STAR_HTSeq Kallisto') # SAMstats')
# Below is the complete list ... | pandas.DataFrame(rows_list) | pandas.DataFrame |
from collections import OrderedDict
from datetime import datetime, timedelta
import numpy as np
import numpy.ma as ma
import pytest
from pandas._libs import iNaT, lib
from pandas.core.dtypes.common import is_categorical_dtype, is_datetime64tz_dtype
from pandas.core.dtypes.dtypes import (
CategoricalDtype,
Da... | Series([np.nan, np.nan]) | pandas.Series |
# -*- coding: utf-8 -*-
import re
import warnings
from datetime import timedelta
from itertools import product
import pytest
import numpy as np
import pandas as pd
from pandas import (CategoricalIndex, DataFrame, Index, MultiIndex,
compat, date_range, period_range)
from pandas.compat import PY... | pd.Timestamp('2011-01-01') | pandas.Timestamp |
from sklearn.model_selection import train_test_split
from sklearn.metrics import confusion_matrix
import pandas
TEST_DF = | pandas.DataFrame( [1,2,3]) | pandas.DataFrame |
# Licensed to Modin Development Team under one or more contributor license agreements.
# See the NOTICE file distributed with this work for additional information regarding
# copyright ownership. The Modin Development Team licenses this file to you under the
# Apache License, Version 2.0 (the "License"); you may not u... | assert_index_equal(renamed.columns, modin_renamed.columns) | pandas.testing.assert_index_equal |
import datetime
import inspect
import logging
import numpy.testing as npt
import os.path
import pandas as pd
import pkgutil
import sys
from tabulate import tabulate
import unittest
try:
from StringIO import StringIO
except ImportError:
from io import StringIO, BytesIO
# #find parent directory and import model
... | pd.read_csv(csv_transpose_path_in, index_col=0, engine='python') | pandas.read_csv |
'''
Created on 19 May 2018
@author: Ari-Tensors
Binary classification: Predict if an asset will fail within certain time frame (e.g. cycles)
'''
import keras
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
import os, traceback
import json
# Setting seed for reproducibility
np.random.seed(1234... | pd.read_csv('./server/Dataset/PM_test.txt', sep=" ", header=None) | pandas.read_csv |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.