prompt stringlengths 19 1.03M | completion stringlengths 4 2.12k | api stringlengths 8 90 |
|---|---|---|
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Mon Mar 8 22:58:48 2021
@author: laura.gf
"""
import requests
from requests.exceptions import HTTPError
import time
from dateutil.relativedelta import relativedelta
from datetime import datetime
import pandas as pd
import sys
def query_entry_pt(url):
... | pd.json_normalize(json_resp,record_path=record_path_field) | pandas.json_normalize |
from datetime import datetime, time, timedelta
from pandas.compat import range
import sys
import os
import nose
import numpy as np
from pandas import Index, DatetimeIndex, Timestamp, Series, date_range, period_range
import pandas.tseries.frequencies as frequencies
from pandas.tseries.tools import to_datetime
impor... | frequencies.get_freq_code((1000, 1)) | pandas.tseries.frequencies.get_freq_code |
import pandas as pd
import numpy as np
from itertools import chain
import requests
def read_cnv(inputfile):
"""Function to read CNV input file.
input: _somatic_cnv.tsv
output: dataframe"""
def convert_to_int(row):
if row['chr'].lower() in ["x", "y"]:
return row["chr"]
elif... | pd.DataFrame(reshaped_data) | pandas.DataFrame |
import pandas as pd
import numpy as np
from xgboost import XGBRegressor
from sklearn.preprocessing import StandardScaler
from sklearn.model_selection import GridSearchCV
from sklearn.model_selection import train_test_split
from sklearn import metrics
import os
import sys
import itertools
from pathlib import Path
import... | pd.DataFrame(training_overhead) | pandas.DataFrame |
## Packages.
import pandas, os, tqdm
## Group of table of data.
group = []
for mode in ['train', 'test']:
if(mode=='train'):
## Load table.
table = pandas.read_csv("../DATA/BMSMT/TRAIN/CSV/LABEL.csv")
table['mode'] = 'train'
## Information.
fol... | pandas.read_csv("../DATA/BMSMT/TEST/CSV/LABEL.csv") | pandas.read_csv |
import types
from functools import wraps
import numpy as np
import datetime
import collections
from pandas.compat import(
zip, builtins, range, long, lzip,
OrderedDict, callable
)
from pandas import compat
from pandas.core.base import PandasObject
from pandas.core.categorical import Categorical
from pandas.co... | _possibly_downcast_to_dtype(result, dtype) | pandas.core.common._possibly_downcast_to_dtype |
"""
Todo:
* Implement correct cfgstrs based on algorithm input
for cached computations.
* Go pandas all the way
Issues:
* errors when there is a word without any database vectors.
currently a weight of zero is hacked in
"""
from __future__ import absolute_import, division, print_function
import ib... | pd.DataFrame(kpts_list, index=aid_series, columns=['kpts']) | pandas.DataFrame |
import pandas as pd
import pickle
from sklearn.cluster import KMeans
import math
latlong_df = pd.read_csv("./latlongfinal.csv")
features = latlong_df.iloc[:, [2,3]]
new_features = | pd.DataFrame() | pandas.DataFrame |
#
# Copyright (C) 2019 Databricks, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to i... | pd.Series([1, 2, 3, 4, 3, 4, 3, 4]) | pandas.Series |
import os
import dash
from dash.dependencies import Input, Output, State
import dash_core_components as dcc
import dash_html_components as html
import dash_bootstrap_components as dbc
import plotly.graph_objects as go
from plotly import express as px
import pandas as pd
from layout import layout_1, layout_2, navbar, ... | pd.to_datetime(date) | pandas.to_datetime |
# -*- coding: utf-8 -*-
"""
Created on Mon Dec 13 17:52:00 2021
@author: SimenLab
"""
import pandas as pd
def Data_getter(file_location):
"""A function which gets and prepares data from CSV files, as well as
returning some additional params like an number of ROIs and their
corresponding labels.
Pa... | pd.DataFrame.to_numpy(averages_dataframe) | pandas.DataFrame.to_numpy |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Tue May 15 22:15:22 2018
@author: tknapen https://github.com/tknapen/hedfpy/blob/master/hedfpy/EyeSignalOperator.py
- Wrapper sacc detection
- sacc detection algorithm
- interpolate gaze function (for pl)
"""
import numpy as np
from scipy.interpolate im... | pd.DataFrame(vel_data) | pandas.DataFrame |
from wf_core_data_dashboard import core
import wf_core_data
import mefs_utils
import pandas as pd
import inflection
import urllib.parse
import os
def generate_mefs_table_data(
test_events_path,
student_info_path,
student_assignments_path
):
test_events = pd.read_pickle(test_events_path)
student_in... | pd.read_pickle(student_assignments_path) | pandas.read_pickle |
# To add a new cell, type '#%%'
# To add a new markdown cell, type '#%% [markdown]'
#%%
from IPython import get_ipython
#%%
import os
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import io
import base64
from matplotlib import animation
from matplotlib import cm
from matplotlib.pyplot import... | pd.read_csv('train.csv') | pandas.read_csv |
# 爬取智联 武汉地区 所有.net 相关的 职位信息
import requests
from bs4 import BeautifulSoup
import pandas
pos=[]
headers={
'Accept':'text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,image/apng,*/*;q=0.8',
'Accept-Encoding':'gzip, deflate',
'Accept-Language':'zh-CN,zh;q=0.9,en;q=0.8',
'Connection':'keep-... | pandas.DataFrame(pos) | pandas.DataFrame |
import pandas as pd
import pdb
import openpyxl
import html
import sys
import datetime
def main():
data = pd.read_excel('cancer_genes_with_pcgp_210921.xlsx',sheet_name='final',index_col=0)
# check args
try:
if sys.argv[1] == 'pcgp':
filterField = 'PCGP category (PMID 26580448)'
... | pd.ExcelWriter(directory + '/' + filename + '.xlsx') | pandas.ExcelWriter |
"""Tests for Table Schema integration."""
import json
from collections import OrderedDict
import numpy as np
import pandas as pd
import pytest
from pandas import DataFrame
from pandas.core.dtypes.dtypes import (
PeriodDtype, CategoricalDtype, DatetimeTZDtype)
from pandas.io.json.table_schema import (
as_json_... | pd.to_datetime(data, utc=True) | pandas.to_datetime |
# -*- coding: utf-8 -*-
"""
Created on Mon Sep 7 11:48:59 2020
@author: mazal
"""
"""
=========================================
Support functions of pydicom (Not sourced)
=========================================
Purpose: Create support functions for the pydicom project
"""
"""
Test mode 1 | Basics... | pd.read_csv(path_ProductType+filename) | pandas.read_csv |
# -*- coding: utf-8 -*-
from datetime import timedelta, time
import numpy as np
from pandas import (DatetimeIndex, Float64Index, Index, Int64Index,
NaT, Period, PeriodIndex, Series, Timedelta,
TimedeltaIndex, date_range, period_range,
timedelta_range, notnu... | Index(rng.asi8) | pandas.Index |
import streamlit as st
import pandas as pd
import numpy as np
import datetime
import plotly.express as px
import base64
def app():
LOGO_IMAGE_IBM = "apps/ibm.png"
LOGO_IMAGE_U_OF_F = "apps/u_of_f.svg.png"
LOGO_IMAGE_BRIGHTER = "apps/brighter_potential_logo.png"
st.markdown(
"""
<style>... | pd.to_datetime(area_stats['date_time']) | pandas.to_datetime |
# encoding=utf-8
""" gs_data centralizes all data import functions such as reading csv's
"""
import pandas as pd
import datetime as dt
from gs_datadict import *
def do_regions(ctydf: pd.DataFrame, mergef: str):
"""
do_regions assigns FIPS missing for multi-county regions, primarily occuring in UT where covid
data ... | pd.notnull(excl.iat[x, 5]) | pandas.notnull |
import collections
import logging
import multiprocessing
import os
import re
import warnings
import numpy as np
import pandas as pd
import tables
from trafficgraphnn.utils import (E1IterParseWrapper, E2IterParseWrapper,
TLSSwitchIterParseWrapper, _col_dtype_key,
... | pd.DataFrame(data) | pandas.DataFrame |
from unittest import TestCase
import pandas as pd
from pandas.api.types import is_numeric_dtype
import numpy as np
from scripts.utils import di
class TestDi(TestCase):
"""Test di() in utils.py"""
def test_di_nan_row(self):
"""Tests that correct distance is computed if NaNs occur in a row of a colum... | pd.Series({2: 0, 3: 0.05 * 0.05, 4: 0.1*0.1, 5: 0.0}) | pandas.Series |
import sys
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import config_local
import logging
_logger = logging.getLogger(__name__)
class PrepareData:
def __init__(self):
path_str = './data/' + config_local.data_clean['data_file_name'] + '.csv'
self.ds = pd.read_csv(path_s... | pd.to_datetime(self.ds['incident_date']) | pandas.to_datetime |
import pandas as pd
import matplotlib.pyplot as plt
import numpy as np
#-------------read csv---------------------
df_2010_2011 = pd.read_csv("/mnt/nadavrap-students/STS/data/data_Shapira_20200911_2010_2011.csv")
df_2012_2013 = pd.read_csv("/mnt/nadavrap-students/STS/data/data_Shapira_20200911_2012_2013.csv")
df_2014... | pd.merge(df5, df2016, on='siteid', how='outer') | pandas.merge |
# coding: utf-8
# In[ ]:
from __future__ import division
import numpy as np
import pandas as pd
from sklearn.model_selection import StratifiedKFold
from sklearn.feature_extraction.text import CountVectorizer
from sklearn.preprocessing import OneHotEncoder,LabelEncoder
from sklearn.model_selection import train_test_... | pd.concat([df_majority_downsampled, df_minority]) | pandas.concat |
from numbers import Number
from collections import Iterable
import re
import pandas as pd
from pandas.io.stata import StataReader
import numpy as np
pd.set_option('expand_frame_repr', False)
class hhkit(object):
def __init__(self, *args, **kwargs):
# if input data frame is specified as a stata data file or text ... | pd.crosstab(df[columns[0]], df[columns[1]], dropna=dropna) | pandas.crosstab |
'''
ASTGCN
'''
import sys
import math
import torch
import torch.nn as nn
import torch.nn.init as init
import torch.nn.functional as F
from torch.autograd import Variable
import numpy as np
import pandas as pd
from Param import *
from Param_ASTGCN import *
from scipy.sparse.linalg import eigs
from torchsummary import su... | pd.merge(result1,sensor_ids2,on='to') | pandas.merge |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Wed Nov 14 14:45:58 2018
@author: yasir
"""
import pandas as pd
import matplotlib.pyplot as plt
import numpy as np
from sklearn.preprocessing import LabelEncoder, OneHotEncoder, StandardScaler
from sklearn.compose import make_column_transformer
from sklea... | pd.read_csv("Churn_Modelling.csv") | pandas.read_csv |
# -*- coding: utf-8 -*-
"""
@author: dani
stable version as per 13 March, 2019
# Once per computer, Before the first run, install heteromotility by copying the following line into the console (and hit enter).
! pip install heteromotility
# Minor error is the vertical position of the text in the boxplots, which is c... | pd.read_csv(HM_outdir + HM_data) | pandas.read_csv |
import pandas as pd
import os
import dcase_util
import random
import tempfile
WORKSPACE = "/home/ccyoung/Downloads/dcase2018_task1-master"
def generate_new_meta_csv():
meta_csv_path = os.path.join(WORKSPACE, 'appendixes', 'meta.csv')
df = pd.read_csv(meta_csv_path, sep='\t')
data = df.groupby(['scene_la... | pd.concat([new_df, v]) | pandas.concat |
import json
import pandas as pd
import numpy as np
from sklearn.feature_extraction.text import TfidfVectorizer
from sklearn.preprocessing import LabelEncoder
from sklearn.model_selection import cross_val_score,train_test_split
from sklearn.linear_model import LogisticRegression, SGDClassifier
from sklearn.naive_bayes i... | pd.DataFrame() | pandas.DataFrame |
# -*- coding: utf-8 -*-
"""
Created on Mon Nov 21 10:24:15 2016
@author: <NAME>
In this file, you will find the main filters
and basics function for time series operations
"""
# Filters
# Every filter has to produce a filtered time serie of prices from a sequence of price relatives
"""
Documentation for all the fol... | pd.DataFrame() | pandas.DataFrame |
import re
import tempfile
from dataclasses import fields
from urllib.error import HTTPError
import click
import pandas as pd
from loguru import logger
from ..etl import collections
from ..etl.core import get_etl_sources
from .scrape import downloaded_pdf, extract_pdf_urls, get_driver
from .utils import RichClickComma... | pd.to_datetime(dt) | pandas.to_datetime |
# # # # # # # # # # # # # # # # # # # # # # # #
# #
# Module to run real time contingencies #
# By: <NAME> and <NAME> #
# 09-08-2018 #
# Version Aplha-0. 1 #
# ... | pd.concat([net.res_sgen.Type, net.res_storage.Type,net.res_ext_grid.Type], ignore_index=True) | pandas.concat |
from datetime import datetime, timedelta
import time
import pandas as pd
from email.mime.text import MIMEText
from smtplib import SMTP
import pytz
from utility import run_function_till_success
| pd.set_option('expand_frame_repr', False) | pandas.set_option |
# Importing necessary packages
import pandas as pd
import numpy as np
import datetime
import geocoder
from geopy.geocoders import Nominatim
from darksky.api import DarkSky, DarkSkyAsync
from darksky.types import languages, units, weather
# Reading monthly yellow taxi trip data for 2019
df1 = pd.read_csv("yellow_tripda... | pd.concat([df1, df2, df3, df4, df5, df6, df7, df8, df9, df10, df11, df12]) | pandas.concat |
import numpy as np
import pandas as pd
import os.path
import random
import collections
from bisect import bisect_right
from bisect import bisect_left
from .. import multimatch_gaze as mp
dtype = [
("onset", "<f8"),
("duration", "<f8"),
("label", "<U10"),
("start_x", "<f8"),
("start_y", "<f8"),
... | pd.DataFrame(data) | pandas.DataFrame |
import os
import sys
import pandas_alive
import pytest
import pandas as pd
import numpy as np
from datetime import datetime, timedelta
from PIL import Image
myPath = os.path.dirname(os.path.abspath(__file__))
sys.path.insert(0, "../..")
@pytest.fixture(scope="function")
def example_dataframe():
test_data = [
... | pd.DataFrame(data=test_data, columns=test_columns, index=test_index) | pandas.DataFrame |
import datetime
import numpy as np
import pytest
import pytz
import pandas as pd
from pandas import Timedelta, merge_asof, read_csv, to_datetime
import pandas._testing as tm
from pandas.core.reshape.merge import MergeError
class TestAsOfMerge:
def read_data(self, datapath, name, dedupe=False):
path = da... | tm.assert_frame_equal(result, expected) | pandas._testing.assert_frame_equal |
import unittest
import pandas as pd
import numpy as np
from econ_watcher_reader.reader import EconomyWatcherReader
import logging
logging.basicConfig()
logging.getLogger("econ_watcher_reader.reader").setLevel(level=logging.DEBUG)
class TestReaderCurrent(unittest.TestCase):
@classmethod
def setUpClass(cls):
... | pd.datetime(2018, 1, 1) | pandas.datetime |
from scapy.utils import RawPcapReader
from scapy.all import PcapReader, Packet
from scapy.layers.l2 import Ether
from scapy.layers.inet import IP, TCP
import scapy.contrib.modbus as mb
from scapy.fields import (
ConditionalField,
Emph,
)
from scapy.config import conf, _version_checker
from scapy.base_classes im... | pd.read_csv(file + "_1") | pandas.read_csv |
import pandas as pd
def generate_demand_csv(input_fn: str, user_data_dir: str):
# Demand
demand = pd.read_excel(input_fn, sheet_name='2.3 EUD', index_col=0, header=1, usecols=range(5))
demand.columns = [x.strip() for x in demand.columns]
demand.index = [x.strip() for x in demand.index]
# Add add... | pd.concat((header, updated_time_series)) | pandas.concat |
"""
Functions for building bokeh figure objects from dataframes.
"""
import datetime
import logging
import math
from typing import Optional, Tuple, Union
import numpy as np
import pandas as pd
from bokeh.layouts import Column, Row, gridplot
from bokeh.models import ColumnDataSource, LabelSet, Legend, LegendItem
from b... | pd.concat([df_bottom, df_top], ignore_index=True) | pandas.concat |
# Copyright (c) 2016 <NAME>
import numpy as np
import pandas as pd
from sklearn import decomposition
import json
import math
import pickle
### Load data
loadPrefix = "import/input/"
# Bins 1, 2, 3 of Up are to be removed later on
dirmagUpA = np.genfromtxt(loadPrefix+"MLM_adcpU_dirmag.csv", skip_header=3, delimite... | pd.DataFrame(data=dirmagDownA[:,1:(1+nBinsUnfilteredDown)], index=dirmagDownIndex) | pandas.DataFrame |
import numpy as np
np.random.seed(0)
import pandas as pd
import matplotlib.pyplot as plt
import gym
env = gym.make('Taxi-v3')
env.seed(0)
print('观察空间 = {}'.format(env.observation_space))
print('动作空间 = {}'.format(env.action_space))
print('状态数量 = {}'.format(env.observation_space.n))
print('动作数量 = {}'.format(env.action_s... | pd.DataFrame(agent.q) | pandas.DataFrame |
import pandas as pd
import pandas.testing as pdt
import qiime2
from qiime2.plugin.testing import TestPluginBase
from q2_types.feature_data import DNAFASTAFormat
from genome_sampler.subsample_diversity import subsample_diversity
class TestSubsampleDiversity(TestPluginBase):
package = 'genome_sampler.tests'
... | pdt.assert_series_equal(sel.inclusion, exp_inclusion) | pandas.testing.assert_series_equal |
from connections.mysql_connector import MySQL_Connector
from models.topic_modeling import Topic_Modeling
from connections.neo4j_connector import Neo4j_Connector
import os
from datetime import datetime
from gensim import corpora, models, similarities
from models.graph_generator import Graph_Generator
from models.tuple_e... | pd.DataFrame() | pandas.DataFrame |
"""Code for the bootstrap uncertainty quantification (BUQ) algorithm."""
import time
import logging
import numpy as np
import pandas as pd
import buq
import models
import tests
def import_time_series_data():
"""Import time series data for model, without any time slicing."""
ts_data = pd.read_csv('data/deman... | pd.to_datetime(ts_data.index) | pandas.to_datetime |
#!python
##################################################
# ACCESS QC Module
# Innovation Laboratory
# Center For Molecular Oncology
# Memorial Sloan Kettering Cancer Research Center
# maintainer: <NAME> (<EMAIL>)
#
#
# This module functions as an aggregation step to combine QC metrics
# across Waltz runs on differe... | pd.concat([gc_cov_int_table, unfilt[2], simplex[2], duplex[2]]) | pandas.concat |
import pytest
from pandas import (
DataFrame,
Index,
Series,
)
import pandas._testing as tm
@pytest.mark.parametrize("n, frac", [(2, None), (None, 0.2)])
def test_groupby_sample_balanced_groups_shape(n, frac):
values = [1] * 10 + [2] * 10
df = DataFrame({"a": values, "b": values})
... | tm.assert_frame_equal(result, expected) | pandas._testing.assert_frame_equal |
import numpy as np
import pandas as pd
from sklearn.decomposition import PCA
from sklearn.ensemble import ExtraTreesRegressor, RandomForestRegressor
from sklearn.feature_selection import SelectFromModel
from sklearn.model_selection import train_test_split
from sklearn.pipeline import make_pipeline
from sklearn.preproce... | pd.read_csv(POSTPROCESSED_DATAPATH, sep=",", header="infer") | pandas.read_csv |
# -*- coding: utf-8 -*-
"""
Created on Mon Oct 19 16:00:06 2020
@author: <NAME>, FINTECH CONSULTANCY
license: Apache 2.0,
Note: only tested on windows 10
"""
import pandas as pd
import sys
import re
import os
import win32com.client
from docx import *
# Hardcoded for now. Wondering where configurations should go for... | pd.isna(data['Revised Definition'].iloc[i]) | pandas.isna |
# Copyright 2019 DeepMind Technologies Limited
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law ... | pd.DataFrame(res, columns=names) | pandas.DataFrame |
# Related third party imports
import pandas as pd
import numpy as np
from imblearn.over_sampling import SMOTE
from sklearn.model_selection import train_test_split
from sklearn.linear_model import LogisticRegression
from sklearn.metrics import accuracy_score, f1_score, precision_score, recall_score, roc_auc_score
import... | pd.DataFrame() | pandas.DataFrame |
import logging
from pathlib import Path
import pandas as pd
from ..gov import Gov, Matcher
from ..const import FILENAME_GOV_TEST_SET
GOV_URL = "http://wiki-de.genealogy.net/Verlustlisten_Erster_Weltkrieg/Projekt/Ortsnamen"
logger = logging.getLogger(__name__)
class GovTestData:
def __init__(self, gov: Gov, url:... | pd.concat(correction_tables) | pandas.concat |
"""The noisemodels module contains all noisemodels available in Pastas.
Supported Noise Models
----------------------
.. autosummary::
:nosignatures:
:toctree: ./generated
NoiseModel
NoiseModel2
Examples
--------
By default, a noise model is added to Pastas. It is possible to replace the
default mod... | Series(index=res.index, data=a, name="Noise") | pandas.Series |
"""
Tasks
-------
Search and transform jsonable structures, specifically to make it 'easy' to make tabular/csv output for other consumers.
Example
~~~~~~~~~~~~~
*give me a list of all the fields called 'id' in this stupid, gnarly
thing*
>>> Q('id',gnarly_data)
['id1','id2','id3']
Observations:
--... | u('fxVersion') | pandas.compat.u |
import pytest
from pandas import Categorical, DataFrame, Series
import pandas.util.testing as tm
def _assert_series_equal_both(a, b, **kwargs):
"""
Check that two Series equal.
This check is performed commutatively.
Parameters
----------
a : Series
The first Series to compare.
b... | Series([1, 2, 3]) | pandas.Series |
#!/usr/bin/env python
#-*- coding:utf-8 -*-
"""Overview:
Classify cell candidates and determine true cells
Usage:
HDoG_classifier.py PARAM_FILE
Options:
-h --help Show this screen.
--version Show version.
"""
import matplotlib as mpl
mpl.use('Agg')
import matplotlib.pyplot as plt
import numpy ... | pd.Series(preds) | pandas.Series |
import os
import copy
import pytest
import numpy as np
import pandas as pd
import pyarrow as pa
from pyarrow import feather as pf
from pyarrow import parquet as pq
from time_series_transform.io.base import io_base
from time_series_transform.io.numpy import (
from_numpy,
to_numpy
)
from time_series_transfor... | pd.DataFrame(data) | pandas.DataFrame |
import os
from datetime import datetime, timedelta, timezone
import pandas as pd
from pandas.core.frame import DataFrame
from sklearn.linear_model import LinearRegression
def demand(exp_id, directory, threshold, warmup_sec):
raw_runs = []
# Compute SLI, i.e., lag trend, for each tested configuration
filen... | pd.DataFrame(raw_runs) | pandas.DataFrame |
#from scipy.stats import chi2
import argparse
import sys
import numpy as np
import pandas as pd
import itertools
#ARGS = None
pnames = ["PRIOR-0", "PRIOR-1", "LIK-0", "LIK-1", "LIK", "POST-0", "POST-1"]
def sigmoid(x, derivative=False):
return x*(1-x) if derivative else 1/(1+np.exp(-x))
def main(args):
par... | pd.read_csv(param_file, sep="\t") | pandas.read_csv |
from __future__ import absolute_import, print_function, division
import matplotlib.pyplot as plt
import matplotlib.dates as mdates
from matplotlib import rc
import pandas as pd
import logging
import json
import numpy as np
import datetime
from better.tools.indicator import max_drawdown, sharpe, positive_count, negative... | pd.DataFrame(results, index=labels) | pandas.DataFrame |
import pandas as pd
import numpy as np
from sklearn.model_selection import GridSearchCV, learning_curve,ShuffleSplit, train_test_split
import os
import time
import shap
import xgboost as xgb
areas = ['CE']
data_version = '2021-07-14_3'
#targets = ['g1','g2','q','r','D','mu_w_0','mu_a_0','RoCof','nadir','MeanDevInFirst... | pd.read_hdf(res_folder+'y_pred_cont.h5') | pandas.read_hdf |
# -*- coding: utf-8 -*-
"""
Created on Tue Apr 20 11:41:36 2021
@author: Koustav
"""
import os
import glob
import matplotlib.pyplot as plt
import seaborn as sea
import numpy as np
import pandas as pan
import math
import matplotlib.ticker as mtick
from scipy.optimize import curve_fit
def expo(x, a, b, c):
return ... | pan.DataFrame(blind, columns= ["p", "A", "SD(A)", "B", "SD(B)", "Decay Rate", "SD(C)"]) | pandas.DataFrame |
# The file will produce result based on closing rank
# -*- coding: utf-8 -*-
"""//@<NAME>
Automatically generated by Colaboratory.
Original file is located at
https://colab.research.google.com/github/priyanshgupta1998/Machine_learning/blob/master/Krisko_Assignmnet/coding.ipynb
"""
"""#Krisko Assi... | pd.DataFrame() | pandas.DataFrame |
import numpy as np
import pandas as pd
import pytest
from estimagic.benchmarking.process_benchmark_results import _clip_histories
from estimagic.benchmarking.process_benchmark_results import _find_first_converged
from estimagic.benchmarking.process_benchmark_results import (
_get_history_as_stacked_sr_from_results,... | pd.DataFrame() | pandas.DataFrame |
from pyspark.sql.functions import expr, col, lit, year
import matplotlib.pyplot as plt
import numpy as np
import pandas as pd
def violations_in_year(nyc_data, vyear):
return nyc_data.select('issue_date').filter(year('issue_date') == vyear).count()
def reduction_in_violations(nyc_data, enable_plot=True):
viola... | pd.DataFrame(seasonwise_violations, columns = ['Violation Code', 'Frequency', 'Season']) | pandas.DataFrame |
import os
import datetime
import random
import pandas as pd
import numpy as np
from calendar import monthrange
from dateutil.easter import easter
from utilities import get_path, get_config
DAYS = {"MON": 0, "Mon": 0, "Mo": 0, "Montag": 0, "Monday": 0,
"TUE": 1, "Tue": 1, "Di": 1, "Dienstag": 1, "Tuesday": 1,
... | pd.DataFrame(index=idx) | pandas.DataFrame |
import numpy as np
import networkx as nx
import pandas as pd
import random
import string
import scipy.stats
import network_prop
import sys
# for parallel processing
#from joblib import Parallel, delayed
#import multiprocessing
def main(num_reps=10, seed_gene_file='HC_genes/example_seed.tsv',int_file='../interactomes/... | pd.DataFrame({'min_degree':min_degree,'max_degree':max_degree,'genes_binned':genes_binned}) | pandas.DataFrame |
import numpy as np
import pandas as pd
from .base_test_class import DartsBaseTestClass
from darts.timeseries import TimeSeries
from darts.utils import timeseries_generation as tg
from darts.metrics import mape
from darts.models import (
NaiveSeasonal,
ExponentialSmoothing,
ARIMA,
Theta,
FourTheta,
... | pd.DataFrame({"V1": values}) | pandas.DataFrame |
import numpy as np
import pandas as pd
from cascade_at.core.log import get_loggers
from cascade_at.dismod.api.fill_extract_helpers import utils
from cascade_at.dismod.constants import DensityEnum, IntegrandEnum, \
INTEGRAND_TO_WEIGHT
LOG = get_loggers(__name__)
DEFAULT_DENSITY = ["uniform", 0, -np.inf, np.inf]
... | pd.DataFrame() | pandas.DataFrame |
import numpy as np
import pandas as pd
from pathlib import Path
from datetime import datetime
import random
import sys
from sklearn.model_selection import ParameterSampler
from scipy.stats import randint as sp_randint
from scipy.stats import uniform
from functions import (
under_over_sampler,
classifier_train... | pd.DataFrame.from_dict(classifier_parameters, orient="index") | pandas.DataFrame.from_dict |
# -*- coding: utf-8 -*-
from __future__ import print_function
from datetime import datetime
import itertools
import numpy as np
import pytest
from pandas.compat import u
import pandas as pd
from pandas import (
DataFrame, Index, MultiIndex, Period, Series, Timedelta, date_range)
from pandas.tests.frame.common ... | assert_frame_equal(result, expected) | pandas.util.testing.assert_frame_equal |
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
from time import time
from sklearn import metrics
from sklearn.cluster import KMeans
from sklearn.decomposition import PCA
from sklearn.preprocessing import scale, LabelEncoder
from sklearn.linear_model import LinearRegression
###################... | pd.DataFrame(kmeans_out, index=companies, columns=['ClusterID']) | pandas.DataFrame |
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
from sklearn.preprocessing import StandardScaler
def data_filling():
train_data = pd.read_csv("./data/train.csv")
test_data = pd.read_csv("./data/test.csv")
train_data = fill_holiday(train_data)
test_data = f... | pd.get_dummies(all_data['holiday']) | pandas.get_dummies |
import pandas as pd
from pandas import DataFrame
import pandas._testing as tm
class TestConcatSort:
def test_concat_sorts_columns(self, sort):
# GH-4588
df1 = DataFrame({"a": [1, 2], "b": [1, 2]}, columns=["b", "a"])
df2 = DataFrame({"a": [3, 4], "c": [5, 6]})
# for sort=True/None... | tm.assert_produces_warning(None) | pandas._testing.assert_produces_warning |
import os.path as osp
import matplotlib.pyplot as plt
# from bokeh.palettes import Category20
from sklearn.manifold import TSNE
import pandas as pd
def tsne(feature_map, results, component_num, dir_path):
# fig, ax = plt.subplots()
# y_pred, y, conf, img_name = results
y_pred, y = results
model_tsne ... | pd.DataFrame() | pandas.DataFrame |
from datetime import datetime
import numpy as np
import pytest
import pandas as pd
from pandas import (
Categorical,
CategoricalIndex,
DataFrame,
Index,
MultiIndex,
Series,
qcut,
)
import pandas._testing as tm
def cartesian_product_for_groupers(result, args, names, fill... | Index([0, 2], dtype="int64") | pandas.Index |
import pandas as pd
import numpy as np
from multiprocessing import Pool
import tqdm
import sys
import gzip as gz
from tango.prepare import init_sqlite_taxdb
def translate_taxids_to_names(res_df, reportranks, name_dict):
"""
Takes a pandas dataframe with ranks as columns and contigs as rows and taxids as value... | pd.DataFrame(taxidmap, index=["staxids"]) | pandas.DataFrame |
import pandas as pd
import numpy as np
from sklearn.preprocessing import LabelEncoder
from sklearn.model_selection import train_test_split
from sklearn.preprocessing import StandardScaler
import keras
from keras.models import Sequential
from keras.layers import Dense, Dropout
from sklearn.metrics import confusion_matri... | pd.DataFrame.from_dict(trainingDict, orient='index', columns=['DataSize']) | pandas.DataFrame.from_dict |
import pandas as pd
import numpy as np
from ..siu import create_sym_call, Symbolic
from functools import singledispatch
# TODO: move into siu
def register_symbolic(f):
@f.register(Symbolic)
def _dispatch_symbol(__data, *args, **kwargs):
return create_sym_call(f, __data.source, *args, **kwargs)
ret... | pd.Categorical.from_codes(new_codes, new_cats) | pandas.Categorical.from_codes |
from datetime import datetime
import numpy as np
from pandas import (
DataFrame,
Index,
MultiIndex,
Period,
Series,
period_range,
to_datetime,
)
import pandas._testing as tm
def test_multiindex_period_datetime():
# GH4861, using datetime in period of multiindex raise... | to_datetime("03/01/2020") | pandas.to_datetime |
#%%
path = '../../dataAndModel/data/o2o/'
import os, sys, pickle
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
from datetime import date
from sklearn.linear_model import SGDClassifier, LogisticRegression
dfoff = pd.read_csv(path+'ccf_offline_stage1_train.csv')
dftest = pd.read_csv(path+'ccf_... | pd.to_datetime(row['Date_received'], format='%Y%m%d') | pandas.to_datetime |
import os
import random
from io import BytesIO
from tempfile import TemporaryDirectory
import tensorflow as tf
from PIL import Image
from google.cloud import storage
import numpy as np
import glob
from tqdm import tqdm
import h5py
import json
from data.thor_constants import THOR_AFFORDANCES, THOR_OBJECT_TYPES, THOR_AC... | pd.DataFrame(df_rows) | pandas.DataFrame |
import functools
import numpy as np
import pandas as pd
import matplotlib as mpl
import matplotlib.transforms as transforms
import matplotlib.pyplot as plt
import scipy.interpolate as interp
import scipy.optimize as opt
from .stats import poisson_interval
__all__ = [
"cms_label", "legend_data_mc", "data_mc", "data... | pd.DataFrame(outdata) | pandas.DataFrame |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""Combine and normalize tweet.json files into a DataFrame dumped to csv
- Find json files (recursively) within the curent path
- Load those that look like tweets dumped by tweetget
- Expand columns that contain arrays, e.g. geo.coordinates -> geo.coordinates.lat and .lon
... | pd.json.load(fin) | pandas.json.load |
import pandas as pd
class TECRDB_compounds_data(object):
def __init__(self):
"""
A module that processes information of compounds in TECRDB
"""
self.TECRDB_compounds_data_dict = {}
self.TECRDB_compounds_pH7_species_id_dict = {}
self.TECRDB_compounds_least_H_sid_dict ... | pd.read_csv('data/TECRDB_compounds_data.csv') | pandas.read_csv |
# ClinVarome annotation functions
# Gather all genes annotations : gene, gene_id,
# (AF, FAF,) diseases, clinical features, mecanismes counts, nhomalt.
# Give score for genes according their confidence criteria
# Commented code is the lines needed to make the AgglomerativeClustering
import pandas as pd
import numpy as ... | pd.read_csv(compare_gene, sep="\t", compression="gzip") | pandas.read_csv |
from unittest import TestCase # or `from unittest import ...` if on Python 3.4+
from category_encoders.utils import convert_input_vector, convert_inputs
import pandas as pd
import numpy as np
class TestUtils(TestCase):
def test_convert_input_vector(self):
index = [2, 3, 4]
result = convert_input... | pd.Series(alist, aindex) | pandas.Series |
"""
Technical Analysis Library
Library of functions to compute various technical indicators.
@author: eyu
"""
import logging
import numpy as np
import pandas as pd
import math as math
import statistics as stats
import datetime
import constants as c
# create logger
logger = logging.getLogger("algo-trader")
def co... | pd.DateOffset(months=3) | pandas.DateOffset |
# -*- coding:utf-8 -*-
# /usr/bin/env python
"""
Date: 2021/7/8 22:08
Desc: 金十数据中心-经济指标-美国
https://datacenter.jin10.com/economic
"""
import json
import time
import pandas as pd
import demjson
import requests
from akshare.economic.cons import (
JS_USA_NON_FARM_URL,
JS_USA_UNEMPLOYMENT_RATE_URL,
JS_USA_EIA_... | pd.DataFrame(value_list) | pandas.DataFrame |
# -*- coding: utf-8 -*-
"""
单变量分析中常用工具,主要包含以下几类工具:
1、自动分箱(降基)模块:包括卡方分箱、Best-ks分箱
2、基本分析模块,单变量分析工具,以及woe编码工具,以及所有变量的分析报告
3、单变量分析绘图工具,如AUC,KS,分布相关的图
"""
# Author: <NAME>
import numpy as np
import pandas as pd
from abc import abstractmethod
from abc import ABCMeta
from sklearn.utils.multiclass import type_of_target
fro... | pd.concat(datas, axis=1) | pandas.concat |
import pandas as pd
import sqlite3
import sys
import datetime
_db_path = 'Database\Database.db'
class GetData(object):
@staticmethod
def Equity(Ticker, Start=None, End = None):
connection = sqlite3.connect(_db_path)
cursor = connection.cursor()
if (Start == None) & (End ==... | pd.read_csv(path) | pandas.read_csv |
import numpy as np
import pandas as pd
from pandas import DataFrame
import matplotlib.pyplot as plt
import seaborn as sns
from os.path import exists
from pathlib import Path
from scipy.constants import value
def change_width(ax, new_value) :
for patch in ax.patches :
current_width = patch.get_width()
... | pd.read_csv(stat_realization_csv_file) | pandas.read_csv |
# -*- coding: utf-8 -*-
# pylint: disable=E1101
# flake8: noqa
from datetime import datetime
import csv
import os
import sys
import re
import nose
import platform
from multiprocessing.pool import ThreadPool
from numpy import nan
import numpy as np
from pandas.io.common import DtypeWarning
from pandas import DataFr... | tm.assert_frame_equal(df, expected) | pandas.util.testing.assert_frame_equal |
#Merges two CSV files and saves the final result
import pandas as pd
import sys
df1 = pd.read_csv(sys.argv[1])
df2 = pd.read_csv(sys.argv[2])
df = | pd.concat([df1, df2], ignore_index=True) | pandas.concat |
import pandas as pd
import numpy as np
# personal csv reader module
import reader
def count_number(array, number):
"""
Counts the occurrence of number in array.
"""
count = 0
for entry in array:
if entry == number:
count += 1
return count
def count_numbers(array, numbers... | pd.DataFrame(data=out, dtype=np.int16) | pandas.DataFrame |
# Arithmetic tests for DataFrame/Series/Index/Array classes that should
# behave identically.
# Specifically for datetime64 and datetime64tz dtypes
from datetime import (
datetime,
time,
timedelta,
)
from itertools import (
product,
starmap,
)
import operator
import warnings
import numpy as np
impo... | Timestamp("2000-02-15", tz="US/Central") | pandas.Timestamp |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.