prompt stringlengths 19 1.03M | completion stringlengths 4 2.12k | api stringlengths 8 90 |
|---|---|---|
#! /usr/bin/env python3
import pandas as pd
import pathlib
import fire
import numpy as np
BEDHEADER = [
'chrom',
'start',
'end',
'tr_id',
'score',
'strand',
'thickStart',
'thickEnd',
'itemRgb',
'blockCount',
'blockSizes',
'blockStarts'
]
class Bed(object):
def _... | pd.read_table(tr_type_file) | pandas.read_table |
#!/usr/bin/env python
"""Units and constants for transforming into and out of SI units.
All data is sourced from :py:mod:`scipy.constants` and :py:attr:`scipy.constants.physical_constants`. Every quantity stored in :py:class:`~solarwindpy.core.plasma.Plasma` and contained objects should have a entry in :py:class:`Cons... | pd.Series(_kBoltzmann) | pandas.Series |
from __future__ import print_function
import logging
import pandas as pd
import numpy as np
import scipy.stats as stats
from matplotlib.backends.backend_pdf import PdfPages
import os.path
from .storemanager import StoreManager
from .condition import Condition
from .constants import WILD_TYPE_VARIANT
from .sfmap import ... | pd.isnull(bcm) | pandas.isnull |
from django.shortcuts import render
from django.views.generic import TemplateView
import pandas as pd
from .utils import clean_html
from form_submissions.models import FormResponse
from typeforms.models import Typeform
class DashboardView(TemplateView):
template_name = 'dashboard.html'
def get(self, reques... | pd.DataFrame(answers) | pandas.DataFrame |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
from sklearn.cluster.bicluster import SpectralCoclustering
from bokeh.plotting import figure, output_file, show
from bokeh.models import HoverTool, ColumnDataSource
from itertools import product
######... | pd.Series([6, 3, 8, 6], index=["q", "w", "e", "r"]) | pandas.Series |
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not us... | pd.Timestamp(1) | pandas.Timestamp |
import re
import os
import sys
import pandas as pd
from lxml import etree
from scipy import stats
import gzip
from sqlalchemy import create_engine
class Polite():
"""
MALLET parameters used: 'output-topic-keys', 'output-doc-topics',
'word-topic-counts-file', 'topic-word-weights-file',
'xml-topic-report... | pd.DataFrame(WORD, columns=['word_id', 'word_str']) | pandas.DataFrame |
#!/usr/bin/env python3
from __future__ import print_function
from collections import defaultdict as dd
from collections import Counter
import os
import pysam
import argparse
from operator import itemgetter
import pandas as pd
import numpy as np
import scipy.stats as ss
import matplotlib
# Force matplotlib to not ... | pd.to_numeric(meth_table['loc']) | pandas.to_numeric |
# -*- coding: utf-8 -*-
"""
Created on Fri Sep 4 10:46:32 2020
@author: OscarFlores-IFi
"""
#%%=========================================================================================================
# Librerías necesarias para correr el código
#==================================... | pd.to_datetime(Fest["Fecha"]) | pandas.to_datetime |
'''
Urban-PLUMBER processing code
Associated with the manuscript: Harmonized, gap-filled dataset from 20 urban flux tower sites
Copyright (c) 2021 <NAME>
Licensed under the Apache License, Version 2.0 (the "License").
You may obtain a copy of the License at: http://www.apache.org/licenses/LICENSE-2.0
'''
__title__ =... | pd.DateOffset(minutes=1) | pandas.DateOffset |
# -----------------------------------------------------------------------------
# WSDM Cup 2017 Classification and Evaluation
#
# Copyright (c) 2017 <NAME>, <NAME>, <NAME>, <NAME>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the ... | pd.DataFrame(result) | pandas.DataFrame |
import os
import numpy as np
import pandas as pd
import networkx as nx
def create_polarity_csv(neighbors_csv_path, mcmc_path, user_polarities_paths):
"""
Merge the neighbors csv with both the neighbourhood-based polarities and the
following-based polarities.
Input:
neighbors_csv_path : path... | pd.read_csv(neighbors_csv_path) | pandas.read_csv |
import os
import random
import math
import numpy as np
import pandas as pd
import itertools
from functools import lru_cache
##########################
## Compliance functions ##
##########################
def delayed_ramp_fun(Nc_old, Nc_new, t, tau_days, l, t_start):
"""
t : timestamp
current date
... | pd.Timedelta(days=1) | pandas.Timedelta |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
import os
import csv
import hashlib
from typing import ContextManager
import srt
import pandas
import functools
from pydub import AudioSegment
from datetime import datetime, timedelta
from pathlib import Path
from praatio import tgio
from .clean_transcript import clea... | pandas.DataFrame(columns=['wav_filename', 'wav_filesize', 'transcript']) | pandas.DataFrame |
# Copyright 2020 AstroLab Software
# Author: <NAME>
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or... | pd.DataFrame.from_dict(data, orient='index') | pandas.DataFrame.from_dict |
import datetime
from datetime import timedelta
from distutils.version import LooseVersion
from io import BytesIO
import os
import re
from warnings import catch_warnings, simplefilter
import numpy as np
import pytest
from pandas.compat import is_platform_little_endian, is_platform_windows
import pandas.util._test_deco... | pd.Timestamp("20130105") | pandas.Timestamp |
# ###########################################################################
#
# CLOUDERA APPLIED MACHINE LEARNING PROTOTYPE (AMP)
# (C) Cloudera, Inc. 2021
# All rights reserved.
#
# Applicable Open Source License: Apache 2.0
#
# NOTE: Cloudera open source products are modular software products
# made up of hun... | pd.Series(preds) | pandas.Series |
import os
import unittest
import pandas as pd
import pyarrow as pa
import pyarrow.parquet as pq
from datetime import datetime
from bqsqoop.utils.parquet_util import ParquetUtil
def sample_df():
_data = [
dict(colA="val1", colB=1),
dict(colA="val2", colB=2)
]
return | pd.DataFrame.from_dict(_data) | pandas.DataFrame.from_dict |
import string
import random
import pathlib
import numpy as np
import pandas as pd
from scipy import stats
path = pathlib.Path(
'~/dev/python/python1024/data/dataproc/006analysis/case').expanduser()
shop_path = path.joinpath('店铺基本数据.xlsx')
# 产品列表
product_list = [f'产品{c}' for c in string.ascii_uppercase]
# 产品价格/成本列... | eries(x_order_prod) | pandas.Series |
from typing import List
from typing import Optional
from typing import Callable
import numpy as np
import pandas as pd
import xarray as xr
from pathlib import Path
from sqlalchemy.orm import Session
import portfolio_management.paths as p
import portfolio_management.data.constants as c
from portfolio_management.io_ut... | pd.DataFrame(records) | pandas.DataFrame |
import numpy.testing as npt
import pandas as pd
import pandas.testing as pdt
import pytest
from message_ix import Scenario, make_df
from message_ix.testing import make_dantzig, make_westeros
def test_make_df():
# DataFrame prepared for the message_ix parameter 'input' has the correct
# shape
result = mak... | pd.DataFrame({"foo": "bar", "baz": [42, 43]}) | pandas.DataFrame |
#!/usr/bin/env python
# coding: utf-8
# ### - PCA and Clustering for Cell painting Level-4 profiles (per dose treament)
#
# #### - Use Silhouette and Davies Bouldin scores to assess the number of clusters from K-Means
# #### - Use BIC scores to assess the number of clusters from Gaussian Mixture Models (GMM)
#
# [r... | pd.read_csv(common_file, sep="\t") | pandas.read_csv |
import ntpath
from datetime import datetime as dt
import os
import pandas as pd
import numpy as np
import math
import sqlite3
# clean the original raw data by storing only the columns that we need, and removing the rest.
def clean(from_path, to_path, columns):
def convert_date(date):
if date == '':
... | pd.notnull(df['Date']) | pandas.notnull |
# -*- coding: utf-8 -*-
"""
Created on Sat Nov 7 22:13:43 2020
@author: <NAME>
"""
#==================================
#ARIMA
#==================================
import os
import warnings
warnings.filterwarnings('ignore')
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import... | pd.to_datetime(df_test.ds) | pandas.to_datetime |
from pickle import TRUE
from flask import *
import math
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
from sklearn.metrics import accuracy_score
import random
import socket
import os
import time
import rss23
pid = 3
def rssrd(r, xy,client):
f = {}
g = {}
R =... | pd.DataFrame(Origional_test_data) | pandas.DataFrame |
import functools
import numpy as np
from scipy.stats import norm as ndist
import regreg.api as rr
from selection.tests.instance import gaussian_instance
from selection.learning.utils import (partial_model_inference,
pivot_plot,
lee_inference)
fr... | pd.read_csv(csvfile) | pandas.read_csv |
import pandas as pd
filepath_dict = {'yelp': 'data/sentiment_analysis/yelp_labelled.txt',
'amazon': 'data/sentiment_analysis/amazon_cells_labelled.txt',
'imdb': 'data/sentiment_analysis/imdb_labelled.txt'}
df_list = []
for source, filepath in filepath_dict.items():
df = | pd.read_csv(filepath, names=['sentence', 'label'], sep='\t') | pandas.read_csv |
# -*- coding: utf-8 -*-
"""
Created on Mon Dec 17 19:51:21 2018
@author: Bob
"""
from sklearn.preprocessing import StandardScaler
from sklearn.cluster import DBSCAN
from nltk.tokenize import word_tokenize
from nltk.stem import PorterStemmer
from nltk.corpus import stopwords
from sqlalchemy import create_engine
from c... | pd.read_csv('country_land_data.csv', encoding='latin-1') | pandas.read_csv |
import numpy as np
from scipy.stats import ranksums
import pandas as pd
import csv
file = pd.read_csv('merged-file.txt', header=None, skiprows=0, delim_whitespace=True)
file.columns = ['Freq_allel','dpsnp','sift','polyphen','mutas','muaccessor','fathmm','vest3','CADD','geneName']
df = file.drop_duplicates(keep=False... | pd.read_csv('/encrypted/e3000/gatkwork/COREAD-ESCA-all-driver.tsv', header=None, skiprows=0, sep='\t') | pandas.read_csv |
#########################################################
### DNA variant annotation tool
### Version 1.0.0
### By <NAME>
### <EMAIL>
#########################################################
import pandas as pd
import numpy as np
import allel
import argparse
import subprocess
import sys
import os.path
import pickle... | pd.DataFrame() | pandas.DataFrame |
import preprocess
from sklearn.preprocessing import StandardScaler
from sklearn.decomposition import PCA
import pandas as pd
import plotly.express as px
from sklearn.decomposition import FastICA
import matplotlib.pyplot as plt
import numpy as np
dataPath = r"C:\Users\shalev\Desktop\Introduction_to_AI\Introduction-to-A... | pd.concat([self.reduced_X_for_plot, self.data[['odor']]], axis=1) | pandas.concat |
from dis import dis
import numpy as np
import pandas as pd
import warnings
from credoai.modules.credo_module import CredoModule
from credoai.utils.constants import MULTICLASS_THRESH
from credoai.utils.common import NotRunError, is_categorical
from credoai.utils.dataset_utils import ColumnTransformerUtil
from credoai.ut... | pd.DataFrame(prepared_arr, index=index) | pandas.DataFrame |
#!/usr/bin/env python3.9
import matplotlib.pyplot as plt
import pandas as pd
import subprocess
import copy
import re
import time
import argparse
import sys
class Log:
date=None
add_lines=0
del_lines=0
def reset(self):
self.date=None
self.add_lines=0
self.del_lines=0
def co... | pd.to_datetime(date, unit='s') | pandas.to_datetime |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import os
import pickle
import shutil
import sys
import tempfile
import numpy as np
from numpy import arange, nan
import pandas.testing as pdt
from pandas import DataFrame, MultiIndex, Series, to_datetime
# dependencies testing specific
import pytest
import recordlinka... | DataFrame({'col': ['abc', 'abc', 'abc', 'abc', 'abc']}) | pandas.DataFrame |
"""
Copyright 2019 <NAME>.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing,
software distribut... | pd.Series([49.2, 55.7, 98.4], index=_index * 3, name='esDisclosurePercentage') | pandas.Series |
import pandas as pd
df = pd.DataFrame({"A": [1, 2, 3, 4, 5]})
s = | pd.Series([1, 2, 3]) | pandas.Series |
import pandas as pd
import numpy as np
from sklearn.preprocessing import StandardScaler
from utils import sum_country_regions, get_change_rates, generate_df_change_rate, sliding_window, generate_COVID_input, generate_COVID_aux_input
import pickle
import copy
import os
import argparse
parser = argparse.ArgumentParse... | pd.read_csv(url_death, error_bad_lines=False) | pandas.read_csv |
# -*- coding: utf-8 -*-
"""
Created on Thu Aug 5 12:02:47 2021
@author: adarshpl7
"""
import pandas as pd
import numpy as np
# import matplotlib.pyplot as plt
import seaborn as sns
# import math
#Clears console and stored variables
try:
from IPython import get_ipython
get_ipython().mag... | pd.read_csv ("C:/Users/adars/Downloads/Laptop/Semester 4/RA work/Social/SocialIndicators_BroadUS_2020-10-01_2021-03-31/SocialIndicators_BroadUS_OpenToClose_2020-10-01_2021-03-31.tsv", sep = '\t') | pandas.read_csv |
"""
Classes for representing datasets of images and/or coordinates.
"""
from __future__ import print_function
import json
import copy
import logging
import os.path as op
import numpy as np
import pandas as pd
import nibabel as nib
from .base import NiMAREBase
from .utils import (tal2mni, mni2tal, mm2vox, get_template... | pd.merge(id_df, temp_df, left_index=True, right_index=True, how='outer') | pandas.merge |
import os
from typing import Text
from IPython.core.display import display, HTML
from jinja2 import Environment, FileSystemLoader
from numpy.lib.function_base import disp
import pandas as pd
import tensorflow_data_validation as tfdv
from tensorflow_data_validation.utils.display_util import (
get_anomalies_datafram... | pd.DataFrame(meta_table) | pandas.DataFrame |
# -*- coding: utf-8 -*-
''' This program takes a excel sheet as input where each row in first column of sheet represents a document. '''
import pandas as pd
import string
import numpy as np
from sklearn.feature_extraction.text import TfidfVectorizer
from sklearn.neighbors import KNeighborsClassifier
from sklearn.clu... | pd.DataFrame(classification_dic, index=[testing_ticket_numbers], columns=['Issue', 'Transformed Data', 'Machine Cluster']) | pandas.DataFrame |
'''
pandas demo -数据清洗 ( numpy-1.19.2 pandas-1.1.2 scikit-learn-0.23.2 )
'''
import numpy as np
from pandas import Series, DataFrame
import pandas as pd
from sqlalchemy import create_engine
def is_null():
df = pd.DataFrame(np.random.randn(10, 6))
df.iloc[:4, 1] = None
df.iloc[:2, 4:6] = None
df.iloc[6, 3:5] ... | pd.DataFrame(kmodel.cluster_centers_) | pandas.DataFrame |
# -*- coding: utf-8 -*-
"""
This file is part of the Shotgun Lipidomics Assistant (SLA) project.
Copyright 2020 <NAME> (UCLA), <NAME> (UCLA), <NAME> (UW).
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the Licen... | pd.read_excel(exp_temp_loc, sheet_name='POS', header=0, index_col=None, na_values='.') | pandas.read_excel |
from src.utility.DBUtils import get_engine
from src.model.BaseModel import BaseModel
import pandas as pd
import numpy as np
class Summary:
# use a sqlite database to save and fetch experiment results
def __init__(self, db_path):
self.__engine = get_engine(db_path)
self.table = None
sel... | pd.DataFrame() | pandas.DataFrame |
from secrets import IEX_CLOUD_API_TOKEN as iex_tkn
import pandas as pd
import requests
from statistics import mean
from scipy import stats
import math
stocks_file = 'sp_500_stocks.csv'
api_url = "https://sandbox.iexapis.com/stable/"
def get_info(symbol):
endpt = f"stock/{symbol}/stats?token={iex_tkn}"
respons... | pd.read_csv(filename) | pandas.read_csv |
from collections import abc, deque
from decimal import Decimal
from io import StringIO
from warnings import catch_warnings
import numpy as np
from numpy.random import randn
import pytest
from pandas.core.dtypes.dtypes import CategoricalDtype
import pandas as pd
from pandas import (
Categorical,
DataFrame,
... | tm.assert_frame_equal(result, df) | pandas._testing.assert_frame_equal |
"""
Tests for kf_lib_data_ingest/extract/operations.py
"""
import pandas
import pytest
from kf_lib_data_ingest.common.type_safety import function
from kf_lib_data_ingest.etl.extract import operations
from test_type_safety import type_exemplars
df = pandas.DataFrame({"COL_A": ["1", "2", "3"]})
other_df = pandas.DataF... | pandas.DataFrame({"OUT_COL": ["a", "b", "c"]}) | pandas.DataFrame |
from typing import Union, cast
import warnings
import numpy as np
from pandas._libs.lib import no_default
import pandas._libs.testing as _testing
from pandas.core.dtypes.common import (
is_bool,
is_categorical_dtype,
is_extension_array_dtype,
is_interval_dtype,
is_number,
is_numeric_dtype,
... | is_interval_dtype(right.dtype) | pandas.core.dtypes.common.is_interval_dtype |
import os
import pandas as pd
from tqdm import tqdm
import pipelines.p1_orca_by_stop as p1
from utils import constants, data_utils
NAME = 'p2_aggregate_orca'
WRITE_DIR = os.path.join(constants.PIPELINE_OUTPUTS_DIR, NAME)
def load_input():
path = os.path.join(constants.PIPELINE_OUTPUTS_DIR, f'{p1.NAME}.csv')
... | pd.read_csv(path) | pandas.read_csv |
# -*- coding:utf-8 -*-
# /usr/bin/env python
"""
Date: 2020/10/19 9:28
Desc: 新浪财经-A股-实时行情数据和历史行情数据(包含前复权和后复权因子)
"""
import re
import json
import demjson
from py_mini_racer import py_mini_racer
import pandas as pd
import requests
from tqdm import tqdm
from akshare.stock.cons import (zh_sina_a_stock_payload,
... | pd.to_datetime(temp_df["date"]) | pandas.to_datetime |
#!/usr/bin/env python
#
# -----------------------------------------------------------------------------
# Copyright (c) 2018 The Regents of the University of California
#
# This file is part of kevlar (http://github.com/dib-lab/kevlar) and is
# licensed under the MIT license: see LICENSE.
# ----------------------------... | pandas.DataFrame(columns=colnames) | pandas.DataFrame |
import sys,os
import pandas as pd
import numpy as np
from statsmodels.tsa.api import ARIMA, SARIMAX, ExponentialSmoothing, VARMAX
from statsmodels.tsa.arima.model import ARIMA as StateSpaceARIMA
import unittest
from nyoka import ExponentialSmoothingToPMML, StatsmodelsToPmml
class TestMethods(unittest.TestCase):
... | pd.Series(data, index) | pandas.Series |
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
from seir.sampling.model import SamplingNInfectiousModel
import logging
logging.basicConfig(level=logging.INFO)
if __name__ == '__main__':
logging.info('Loading data')
# read calibration data
actual_hospitalis... | pd.to_datetime('2020-03-27') | pandas.to_datetime |
import numpy as np
import pandas as pd
from relation import *
from connection import *
class RuleSet:
#############################
# Methods to build rule set #
#############################
def __init__(self,rules_list):
''' @rules_list: list of ordered dictionnaries each representing a rule... | pd.concat([self.set,rule],sort=False) | pandas.concat |
from time import time
from datetime import datetime
import os, sys
import numpy as np
from scipy.stats.mstats import gmean
import scipy.spatial.distance as ssd
import scipy.cluster.hierarchy as hc
import pandas as pd
import pickle
import gensim, data_nl_processing, data_nl_processing_v2
import spacy
import scispacy
fr... | pd.read_csv(data_path2) | pandas.read_csv |
#!/usr/bin/python3
# -*- coding: utf-8 -*-
import inspect
import pandas as pd
pd.options.display.colheader_justify = 'right'
| pd.set_option('display.unicode.east_asian_width', True) | pandas.set_option |
#This class defines the Scoreboard data structure and its associated methods
import pandas as pd
import os
from dotenv import load_dotenv
import boto3
class Scoreboard:
standard_display = ["Member","Score"]
all_time_display = ["Member","AllTime"]
commits_display = ["Member","Commits"]
#Load up S3 a... | pd.read_csv(obj["Body"], index_col=0) | pandas.read_csv |
# This scripts are possible solutions for the Database Tasks
# Packages that might have to be installed via pip/conda:
# conda install pandas
# conda install MySQL-python
# conda install mysqlclient
# conda install pymongo
# conda install sqlite3
# General Packages
import pandas as pd
# Packages for MySQL
... | pd.read_sql(query, conn) | pandas.read_sql |
import unittest
import pytest
from pyalink.alink import *
def print_value_and_type(v):
print(type(v), v)
class TestAkStream(unittest.TestCase):
def setUp(self) -> None:
self.lfs = LocalFileSystem()
self.hfs = HadoopFileSystem("2.8.3", "hdfs://xxx:9000")
self.ofs = OssFileSystem("3... | pd.DataFrame(arr) | pandas.DataFrame |
import argparse as ap
from itertools import product
from typing import List, Tuple
import bnet.utype as ut
import numpy as np
import yaml
from nptyping import NDArray
from pandas.core.frame import DataFrame
class Config(dict):
def __init__(self, path: str):
f = open(path, "r")
self.__path = path
... | DataFrame(result, columns=columns) | pandas.core.frame.DataFrame |
#!/usr/bin/env python3
import gc
import os
import pickle
import fire
import h5py
import matplotlib.pyplot as plt
import seaborn as sns
from hyperopt.fmin import generate_trials_to_calculate
from sklearn.preprocessing import StandardScaler
from sklearn.metrics import precision_recall_curve
from numpy import linalg as L... | pd.concat(temp, ignore_index=True) | pandas.concat |
"""
<NAME>017
Variational Autoencoder - Pan Cancer
scripts/vae_pancancer.py
Usage:
Run in command line with required command arguments:
python scripts/vae_pancancer.py --learning_rate
--batch_size
--epochs
... | pd.read_table(rnaseq_file, index_col=0) | pandas.read_table |
# -*- coding: utf-8 -*-
import csv
import os
import platform
import codecs
import re
import sys
from datetime import datetime
import pytest
import numpy as np
from pandas._libs.lib import Timestamp
import pandas as pd
import pandas.util.testing as tm
from pandas import DataFrame, Series, Index, MultiIndex
from pand... | StringIO(s) | pandas.compat.StringIO |
import streamlit as st
from alphapept.gui.utils import (
check_process,
init_process,
start_process,
escape_markdown,
)
from alphapept.paths import PROCESSED_PATH, PROCESS_FILE, QUEUE_PATH, FAILED_PATH
from alphapept.settings import load_settings_as_template, save_settings
import os
import psutil
import... | pd.DataFrame(failed_files) | pandas.DataFrame |
"""
Helper functions for loading, converting, reshaping data
"""
import pandas as pd
import json
from pandas.api.types import CategoricalDtype
FILLNA_VALUE_CAT = 'NaN'
CATEGORICAL = "categorical"
CONTINUOUS = "continuous"
ORDINAL = "ordinal"
COLUMN_CATEGORICAL = 'categorical_columns'
COLUMN_CONTINUOUS = 'continuous... | CategoricalDtype(categories=col['i2s'], ordered=True) | pandas.api.types.CategoricalDtype |
from numbers import Number
from collections import Iterable
import re
import pandas as pd
from pandas.io.stata import StataReader
import numpy as np
pd.set_option('expand_frame_repr', False)
class hhkit(object):
def __init__(self, *args, **kwargs):
# if input data frame is specified as a stata data file or text ... | pd.Series(right_using_on_key) | pandas.Series |
#IMPORTING LIBRARIES
import numpy as np
import pandas as pd
import os
from sklearn import preprocessing
from sklearn.model_selection import train_test_split
from statistics import mean
from sklearn.metrics import accuracy_score
import matplotlib.pyplot as plt
import seaborn as sns
from scipy.stats import norm
from skl... | pd.read_csv('../input/indian-liver-patient-records/indian_liver_patient.csv') | pandas.read_csv |
from typing import Tuple
import warnings
warnings.simplefilter("ignore", UserWarning)
from functools import partial
from multiprocessing.pool import Pool
import pandas as pd
import numpy as np
import numpy_groupies as npg
from cellphonedb.src.core.core_logger import core_logger
from cellphonedb.src.core.models.comple... | pd.concat([interactions_data_result, mean_pvalue_result], axis=1, join='inner', sort=False) | pandas.concat |
"""
Active Fairness Run through questions
"""
from sklearn.metrics import classification_report
from sklearn.metrics import accuracy_score
from sklearn.metrics import confusion_matrix
from sklearn.ensemble import RandomForestClassifier
from sklearn.tree import DecisionTreeClassifier
from sklearn.calibration import _Si... | pd.DataFrame(self.dataset_test.labels[:, 0]) | pandas.DataFrame |
#!/usr/bin/env python3 -u
# -*- coding: utf-8 -*-
__author__ = ["<NAME>"]
__all__ = [
"TEST_YS",
"TEST_SPS",
"TEST_ALPHAS",
"TEST_FHS",
"TEST_STEP_LENGTHS_INT",
"TEST_STEP_LENGTHS",
"TEST_INS_FHS",
"TEST_OOS_FHS",
"TEST_WINDOW_LENGTHS_INT",
"TEST_WINDOW_LENGTHS",
"TEST_INITI... | pd.Timedelta(-3, unit="D") | pandas.Timedelta |
import pytest
import numpy as np
import pandas
import pandas.util.testing as tm
from pandas.tests.frame.common import TestData
import matplotlib
import modin.pandas as pd
from modin.pandas.utils import to_pandas
from numpy.testing import assert_array_equal
from .utils import (
random_state,
RAND_LOW,
RAND_... | pandas.DataFrame([]) | pandas.DataFrame |
import os
import pandas as pd
import numpy as np
import pickle
import json
para = {
"window_size": 1,
"step_size": 0.5,
"structured_file": "BGL.log_structured.csv",
"BGL_sequence_train": "BGL_sequence_train.csv",
"BGL_sequence_test": "BGL_sequence_test.csv"
}
def load_BGL():
... | pd.DataFrame(columns=['sequence', 'label']) | pandas.DataFrame |
"""
Unit and regression test for the kissim.comparison.measures module.
"""
import pytest
import numpy as np
import pandas as pd
from kissim.comparison.utils import (
format_weights,
scaled_euclidean_distance,
scaled_cityblock_distance,
)
@pytest.mark.parametrize(
"feature_weights, feature_weights_f... | pd.Series([4, 3]) | pandas.Series |
#! /usr/bin/env python
# -*- coding: utf-8 -*-
import json
import datetime
import os
from os import listdir
from os.path import isfile, join
from shutil import copyfile
import logging
import pandas as pd
logger = logging.getLogger("root")
logging.basicConfig(
format="\033[1;36m%(levelname)s: %(filename)s (def %(f... | pd.read_csv(target) | pandas.read_csv |
###############################################################################
# Copyright (c) 2021, Lawrence Livermore National Security, LLC.
# Produced at the Lawrence Livermore National Laboratory
# Written by <NAME> <<EMAIL>>
#
# All rights reserved.
#
# Permission is hereby granted, free of charge, to any person... | pd.DataFrame(data=synth_data, columns=self.column_names) | pandas.DataFrame |
"""Run unit tests.
Use this to run tests and understand how tasks.py works.
Setup::
mkdir -p test-data/input
mkdir -p test-data/output
mysql -u root -p
CREATE DATABASE testdb;
CREATE USER 'testusr'@'localhost' IDENTIFIED BY 'test<PASSWORD>';
GRANT ALL PRIVILEGES ON testdb.* TO 'te... | pd.read_sql_table(tblname, sql_engine) | pandas.read_sql_table |
"""
utilities that are helpful in general model building
"""
import clickhouse_driver
from muti import chu
import numpy as np
import pandas as pd
import plotly.graph_objs as go
import plotly.io as pio
import scipy.stats as stats
import math
import os
def r_square(yh, y):
"""
find the r-square for the model i... | pd.Series(binary_variable) | pandas.Series |
"""
Tests that apply specifically to the Python parser. Unless specifically
stated as a Python-specific issue, the goal is to eventually move as many of
these tests out of this module as soon as the C parser can accept further
arguments when parsing.
"""
import csv
from io import BytesIO, StringIO
import py... | tm.assert_frame_equal(result, expected) | pandas._testing.assert_frame_equal |
#Imports
import requests
import json
import os
import sys
import pandas as pd
import numpy as np
import dash_core_components as dcc
import dash_html_components as html
import dash_bootstrap_components as dbc
from dash import Dash
from dash.dependencies import Output, Input, State
from dash.exceptions import PreventUpda... | pd.read_json(j_data, orient='split') | pandas.read_json |
"""Transformer for datetime data."""
import numpy as np
import pandas as pd
from pandas.core.tools.datetimes import _guess_datetime_format_for_array
from rdt.transformers.base import BaseTransformer
from rdt.transformers.null import NullTransformer
class UnixTimestampEncoder(BaseTransformer):
"""Transformer for ... | pd.Series(datetime_data) | pandas.Series |
import numpy as np
import pandas as pd
from numpy.testing import assert_array_equal
from pandas.testing import assert_frame_equal
from nose.tools import (assert_equal,
assert_almost_equal,
raises,
ok_,
eq_)
from rsmtool.p... | assert_frame_equal(df_new, df_new_expected) | pandas.testing.assert_frame_equal |
# ========== (c) <NAME> 3/8/21 ==========
import logging
import pandas as pd
import numpy as np
import plotly.express as px
logger = logging.getLogger(__name__)
root_logger = logging.getLogger()
root_logger.setLevel(logging.INFO)
sh = logging.StreamHandler()
formatter = logging.Formatter('%(asctime)s - %(name)s - %(... | pd.set_option('display.max_columns', 20) | pandas.set_option |
#Creo el dataset para la predicción del boosting
import gc
gc.collect()
import pandas as pd
import seaborn as sns
import numpy as np
#%% marzo
marzo = pd.read_csv(r'C:\Users\argomezja\Desktop\Data Science\MELI challenge\Project MELI\Dataset_limpios\marzo_limpio.csv.gz')
marzo = marzo.loc[marzo['day']>=4].r... | pd.merge(final, subtest7, left_index=True, right_index=True) | pandas.merge |
import unittest
from abc import ABC
import numpy as np
import pandas as pd
from toolbox.ml.ml_factor_calculation import ModelWrapper, calc_ml_factor, generate_indexes
from toolbox.utils.slice_holder import SliceHolder
class MyTestCase(unittest.TestCase):
def examples(self):
# index includes non trading... | pd.Timestamp(year=2010, month=1, day=1) | pandas.Timestamp |
from __future__ import division #brings in Python 3.0 mixed type calculations
import numpy as np
import os
import pandas as pd
import sys
#find parent directory and import model
parentddir = os.path.abspath(os.path.join(os.path.dirname(__file__), os.path.pardir))
sys.path.append(parentddir)
from base.uber_model impo... | pd.Series([], dtype="float") | pandas.Series |
import os
import copy
import pytest
import numpy as np
import pandas as pd
import pyarrow as pa
from pyarrow import feather as pf
from pyarrow import parquet as pq
from time_series_transform.io.base import io_base
from time_series_transform.io.numpy import (
from_numpy,
to_numpy
)
from time_series_transfor... | pd.testing.assert_frame_equal(x,expectedX,check_dtype=False) | pandas.testing.assert_frame_equal |
import glob
import numpy as np
import pandas as pd
from collections import OrderedDict
#from . import metrics
import metrics
from .csv_reader import csv_node
__all__ = ['tune_threshold',
'assemble_node',
'assemble_dev_threshold',
'metric_reading',
'Ensemble']
def tune_thres... | pd.DataFrame(df_dict) | pandas.DataFrame |
import re
import numpy as np
import pytest
import pandas as pd
import pandas._testing as tm
from pandas.core.arrays import IntervalArray
class TestSeriesReplace:
def test_replace_explicit_none(self):
# GH#36984 if the user explicitly passes value=None, give it to them
ser = pd.Series([0, 0, ""],... | pd.Series([1, 2, 3]) | pandas.Series |
import datetime
import hashlib
import os
import time
from warnings import (
catch_warnings,
simplefilter,
)
import numpy as np
import pytest
import pandas as pd
from pandas import (
DataFrame,
DatetimeIndex,
Index,
MultiIndex,
Series,
Timestamp,
concat,
date_range,
timedelt... | tm.assert_series_equal(obj, objs[leaf]) | pandas._testing.assert_series_equal |
# ============================================================================
# Piotroski f score implementation (data scraped from yahoo finance)
# Author - <NAME>
# Please report bugs/issues in the Q&A section
# =============================================================================
import requests
from bs4... | pd.DataFrame(financial_dir_cy) | pandas.DataFrame |
# -*- coding: utf-8 -*-
#!/usr/bin/python3
__author__ = "<NAME>"
__copyright__ = "Copyright 2019-2022"
__license__ = "MIT"
__version__ = "0.1.0"
__maintainer__ = "<NAME>, <NAME>"
__email__ = "<EMAIL>"
__status__ = "Dev"
import textwrap
import pandas as pd
import numpy as np
import json
import argparse
... | pd.read_csv('Ribosomes_newlist_F.txt', sep='\t') | pandas.read_csv |
#!/usr/bin/env python
# -*- coding: utf-8 -*-
# @Time : 2020/3/21 0021
# @Author : justin.郑 <EMAIL>
# @File : index_baidu.py
# @Desc : 获取百度指数
import json
import urllib.parse
import pandas as pd
import requests
def decrypt(t: str, e: str) -> str:
"""
解密函数
:param t:
:type t:
:param e:
... | pd.DataFrame(age_list) | pandas.DataFrame |
# Importing Data in Python (Part 1) on Data Camp
#######################################
# Part 1: Introduction and flat files
#######################################
## Importing entire text files
# Open a file: file
file = open('moby_dick.txt', mode='r')
# Print it
print(file.read())
# Check whether file is cl... | pd.DataFrame.hist(data[['Age']]) | pandas.DataFrame.hist |
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
from joblib import Memory
import datetime
from azure_table_interface import query_aq_data
# Set up caching for the Azure table access
memory = Memory('./_cache_')
ID_to_name = {'nesta-1': 'Priory Rd (South)',
'nesta-2': 'Priory Rd ... | pd.concat(dfs, axis=1) | pandas.concat |
#Se omiten tildes para evitar inconvenientes de codificacion
#Librerias requeridas
import sqlite3
from sqlite3 import Error
import pandas as pd
import numpy as np
import sys
import random
#Definir separador para la carga de los archivos CSV
separador = ";"
#Funciones para generacion de campos aleatorios... | pd.read_csv(rutaTablaCSV,sep=separador) | pandas.read_csv |
# -*- coding: utf-8 -*-
# Copyright (c) 2016-2017 by University of Kassel and Fraunhofer Institute for Wind Energy and
# Energy System Technology (IWES), Kassel. All rights reserved. Use of this source code is governed
# by a BSD-style license that can be found in the LICENSE file.
from math import pi
from numpy impo... | DataFrame(ppc_net['branch'][:, [0, 1, 8, 9]]) | pandas.DataFrame |
import pandas as pd
import numpy as np
def get_series(data: (pd.Series, pd.DataFrame), col='close') -> pd.DataFrame:
"""
Get close column from intraday data
Args:
data: intraday data
col: column to return
Returns:
pd.Series or pd.DataFrame
"""
if isinstance(data, pd.S... | pd.DataFrame(data) | pandas.DataFrame |
"""Tools used for clustering analysis"""
import csv
__author__ = "<NAME> (http://www.vmusco.com)"
import numpy
import os
import pandas
from mlperf.clustering.clusteringtoolkit import ClusteringToolkit
class DatasetFacts:
"""Object alternative to method read_dataset"""
def __init__(self, data):
self... | pandas.DataFrame(initial_clusters) | pandas.DataFrame |
# Object Oriented Programming Examples
import pandas as pd
df = | pd.DataFrame(['tree frog', 'white rhino', 'zebra']) | pandas.DataFrame |
import math
import glob
import os
import uuid
import itertools
import pandas as pd
import numpy as np
import datetime as dt
class GSTools(object):
@staticmethod
def load_csv_files(dir_str):
'''
This function reads all csv from the given directory, stores them in a dictionary and returns it.
... | pd.to_datetime(df['date'], format='%Y-%m-%d %H:%M:%S') | pandas.to_datetime |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.