prompt stringlengths 19 1.03M | completion stringlengths 4 2.12k | api stringlengths 8 90 |
|---|---|---|
import pandas as pd
import os
import time
from minder_utils.configurations import config
from .format_util import iter_dir
from minder_utils.download.download import Downloader
from minder_utils.util.decorators import load_save
from minder_utils.formatting.format_tihm import format_tihm_data
import numpy as np
from min... | pd.to_datetime(data_out.time, utc=True) | pandas.to_datetime |
# previously we looked into numpy and its ndarrayobject in particular. Here
# we build on that knowledge by looking at the data structures provided by the Pandas Library.
# Pandas is a newer package built on top of NumPz and proveides an efficient
# implementation of a DataFrame.
# Here we will focus on the mechani... | pd.Series([0.25, 0.5, 0.75, 1.0]) | pandas.Series |
"""
Data: Temperature and Salinity time series from SIO Scripps Pier
Salinity: measured in PSU at the surface (~0.5m) and at depth (~5m)
Temp: measured in degrees C at the surface (~0.5m) and at depth (~5m)
- Timestamp included beginning in 1990
"""
# imports
import sys,os
import pandas as pd
import numpy as np
im... | pd.to_datetime(sal_data[['YEAR', 'MONTH', 'DAY']]) | pandas.to_datetime |
# Arithmetic tests for DataFrame/Series/Index/Array classes that should
# behave identically.
# Specifically for datetime64 and datetime64tz dtypes
from datetime import (
datetime,
time,
timedelta,
)
from itertools import (
product,
starmap,
)
import operator
import warnings
import numpy as np
impo... | Series([NaT, NaT], dtype="datetime64[ns]") | pandas.Series |
#!/usr/bin/env python
# -*- coding:utf-8 -*-
"""
Date: 2022/2/14 18:19
Desc: 新浪财经-股票期权
https://stock.finance.sina.com.cn/option/quotes.html
期权-中金所-沪深 300 指数
https://stock.finance.sina.com.cn/futures/view/optionsCffexDP.php
期权-上交所-50ETF
期权-上交所-300ETF
https://stock.finance.sina.com.cn/option/quotes.html
"""
import json
i... | pd.DataFrame(data_json) | pandas.DataFrame |
import pandas as pd
import matplotlib.pyplot as plt
import numpy as np
import os
breakout_index = 0
prev_breakout_index = 0
num_samples = 90
month = "july"
num_samples_split = 10
path = str(num_samples_split) + "_normalized_refined_lfc/"
#filename_breakout = "normalized datasets 2/normalized_breakout_" + month + ".xl... | pd.DataFrame(data_required) | pandas.DataFrame |
from BoostInference_no_parallelization import Booster
import sys, pandas as pd, numpy as np
import glob, pickle
from sklearn.metrics import roc_auc_score, precision_recall_curve, auc
if len(sys.argv)<5:
print('python file.py df-val-PhyloPGM-input df-test-PhyloPGM-output info_tree fname_df_pgm_output')
exit(0)... | pd.read_csv(fname_dtrain, index_col=0) | pandas.read_csv |
# Import libraries
import pandas as pd
from bs4 import BeautifulSoup
import matplotlib.pyplot as plt
from urllib.request import urlopen, Request
from nltk.sentiment.vader import SentimentIntensityAnalyzer
# Parameters
n = 10 # the # of article headlines displayed per ticker
tickers = ['AAPL', 'TSLA', 'AMZN']
# Get D... | pd.DataFrame(scores) | pandas.DataFrame |
from sklearn.linear_model import LogisticRegression
from sklearn.preprocessing import LabelEncoder
from twitter.twitter_data_model import User
import pandas as pd
def get_most_likely_author(usernames, tweet_to_classify, nlp):
vects = []
# Puts vectorized tweets in dataframe for each user
for username in u... | pd.DataFrame([tweet.vect for tweet in user_0.tweets]) | pandas.DataFrame |
# -*- coding: utf-8 -*-
import numpy as np
import pytest
from numpy.random import RandomState
from numpy import nan
from datetime import datetime
from itertools import permutations
from pandas import (Series, Categorical, CategoricalIndex,
Timestamp, DatetimeIndex, Index, IntervalIndex)
import pan... | tm.assert_numpy_array_equal(result, expected) | pandas.util.testing.assert_numpy_array_equal |
# MIT License
#
# Copyright (c) 2021. <NAME> <<EMAIL>>
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, mo... | pd.isna(v) | pandas.isna |
import re
import numpy as np
import pandas as pd
import pytest
import cudf
from cudf import melt as cudf_melt
from cudf.core import DataFrame
from cudf.tests.utils import (
ALL_TYPES,
DATETIME_TYPES,
NUMERIC_TYPES,
assert_eq,
)
@pytest.mark.parametrize("num_id_vars", [0, 1, 2, 10])
@pytest.mark.para... | pd.MultiIndex.from_tuples([("c", 1), ("c", 2)], names=[None, None]) | pandas.MultiIndex.from_tuples |
"""Test attributing simple impact."""
import numpy as np
import pandas as pd
import pytest
from nbaspa.data.endpoints.pbp import EventTypes
from nbaspa.player_rating.tasks import SimplePlayerImpact
@pytest.mark.parametrize(
"evt",
[
EventTypes.REBOUND,
EventTypes.FREE_THROW,
EventTyp... | pd.Series([0.0, 0.0]) | pandas.Series |
# coding=utf-8
# pylint: disable-msg=E1101,W0612
from datetime import datetime, timedelta
import operator
from itertools import product, starmap
from numpy import nan, inf
import numpy as np
import pandas as pd
from pandas import (Index, Series, DataFrame, isnull, bdate_range,
NaT, date_range, ti... | tm.assert_frame_equal(res, exp) | pandas.util.testing.assert_frame_equal |
import os
import pytz
import logging
import pymongo
import multiprocessing
import pandas as pd
from datetime import datetime
from collections import Counter, defaultdict
from typing import List, Set, Tuple
# For non-docker use, change to your url (e.g., localhost:27017)
MONGO_URL = "mongodb://localhost:27... | pd.read_excel("data/rules.xlsx", engine="openpyxl") | pandas.read_excel |
import os
from multiprocessing import Value, context
import pandas as pd
import socket
import threading
import multiprocessing
import re
from contextlib import contextmanager
from pathlib import Path
import json
from portalocker import RLock, AlreadyLocked
import shutil
import pytest
from aljpy import humanhash
from fn... | pd.to_datetime(df['_created']) | pandas.to_datetime |
import json
import numpy as np
import pandas as pd
import requests
def get_state_fips_codes():
"""
Returns dataframe of state FIPS codes and state names
from the BLS JT series reference
"""
url = "https://download.bls.gov/pub/time.series/jt/jt.state"
data = requests.get(url)
data_fmt = da... | pd.concat(dfs) | pandas.concat |
"""
Msgpack serializer support for reading and writing pandas data structures
to disk
portions of msgpack_numpy package, by <NAME> were incorporated
into this module (and tests_packers.py)
License
=======
Copyright (c) 2013, <NAME>.
All rights reserved.
Redistribution and use in source and binary forms, with or wit... | Timestamp(obj['value'], tz=obj['tz'], freq=freq) | pandas.Timestamp |
import datetime
import logging
import os
import shutil
import geopandas as gpd
import numpy as np
import pandas as pd
import pytz
from berlin_hp import electricity
from demandlib import bdew as bdew
from demandlib import particular_profiles as profiles
from matplotlib import cm
from matplotlib import dates as mdates
f... | pd.to_datetime(re_en["utc_timestamp"], utc=True) | pandas.to_datetime |
"""
The MIT License (MIT)
Copyright (c) 2016 <NAME>
Copyright (c) 2019 <NAME>
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, ... | pd.concat([alldf, df], axis=0) | pandas.concat |
import os
import pandas as pd
import numpy as np
import copy
from pprint import pprint
def work(pres):
count = [0, 0]
for i in pres:
count[i] += 1
out = count.index(max(count))
return out
def simple_vote(model_name, date, dataset, pseudo=False):
if pseudo:
DATA_DIR = '../predict_... | pd.DataFrame(humor_final_data, columns=['ID', 'Speaker', 'Sentence', 'Label']) | pandas.DataFrame |
import re
import logging
from functools import reduce, partial
from concurrent import futures
from concurrent.futures import ThreadPoolExecutor, as_completed
import pandas as pd
import numpy as np
from pandas.api.types import is_numeric_dtype
from influxdb.resultset import ResultSet
from requests.exceptions import Re... | pd.to_datetime(right.time, unit='ms') | pandas.to_datetime |
import os
import pandas as pd
import mysql.connector as mysql
from mysql.connector import Error
def DBConnect(dbName=None):
"""
Parameters
----------
dbName :
Default value = None)
Returns
-------
"""
conn = mysql.connect(host='localhost', user='root', password='<PASSWORD>',
... | pd.read_csv('data/station_summary.csv') | pandas.read_csv |
import itertools
import warnings
import networkx as nx
import numpy as np
import pandas as pd
from tqdm import tqdm
from AppGenerator import AppGenerator
from ServerlessAppWorkflow import ServerlessAppWorkflow
warnings.filterwarnings("ignore")
class PerfOpt:
def __init__(self, Appworkflow, generate_perf_profile=... | pd.Series(aRow) | pandas.Series |
# Copyright 2021 Rikai Authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in wri... | pd.DataFrame(data) | pandas.DataFrame |
"""
# Part of localization phase
# suspected bug detection:
# 1. Tensorflow,Theano,CNTK
# 2. Tensorflow,Theano,MXNET
#
# voting process
# -> a. inconsistency -> error backend,error layer.
# b. check error backend in new container(whether inconsistency disappears).
# """
#
import numpy as np
import os
import sys
imp... | pd.DataFrame(columns=unsolved_columns) | pandas.DataFrame |
import numpy as np
from scipy.special import softmax
import torch
import torch.nn as nn
import torch.optim as optim
import torch.utils.data as tdata
import pandas as pd
import time
from tqdm import tqdm
from utils import validate, get_logits_targets, sort_sum
import pdb
# Conformalize a model with a calibration set.
#... | pd.DataFrame({'size': size, 'correct': correct}) | pandas.DataFrame |
import cv2
from collections import OrderedDict
import numpy as np
import itertools
from datetime import datetime
from configparser import ConfigParser, MissingSectionHeaderError, NoOptionError, NoSectionError
import pandas as pd
import os
import warnings
warnings.filterwarnings("ignore")
def shap_summary_... | pd.read_csv(feat_cat_csv_path, header=[0, 1]) | pandas.read_csv |
from qfengine.data.price.price_source import MySQLPriceDataSource as SQLTable
from qfengine.asset import assetClasses
import pandas as pd
from typing import Union, List, Dict
import os
import numpy as np
import logging
import functools
from qfengine import settings
import concurrent.futures
logger = logging.getLogge... | pd.Timedelta(hours=9, minutes=30) | pandas.Timedelta |
# cheat sheet https://share.streamlit.io/daniellewisdl/streamlit-cheat-sheet/app.py
# https://docs.streamlit.io/en/stable/index.html
import streamlit as st
import pandas as pd
import numpy as np
import time
# title
st.title('Uber pickups in NYC')
st.write(f" Streamlit version:{st.__version__}")
# get data
DATE_COLU... | pd.to_datetime(data[DATE_COLUMN]) | pandas.to_datetime |
# this program breaks down into a csv where phosphosite orthologs could be lost in the PhosphositeOrthology program
# PhosphositePlus is being used to verify that my orthologs are correct but PSP does not have everything which is the
# reason for using dbPAF
# we want to make sure that the UniprotIDs contained in BO... | pd.read_table('oma-uniprot_clean.txt', dtype=object) | pandas.read_table |
import seaborn as sns
import numpy as np
import matplotlib.pyplot as plt
import pandas as pd
import os
import argparse
from pathlib import Path
parser = argparse.ArgumentParser(description="Analyse the logs produced by torchbeast")
parser.add_argument("--dir", type=str, default="~/logs/torchbeast", help="Directory fo... | pd.concat(dfs) | pandas.concat |
import torch
import pandas as pd
import numpy as np
import os
#os.environ['CUDA_VISIBLE_DEVICES'] = '2'
import transformers as ppb
from transformers import BertForSequenceClassification, AdamW, BertConfig
from torch.utils.data import TensorDataset,DataLoader, RandomSampler, SequentialSampler
from keras.utils import to... | pd.DataFrame(data=training_stats) | pandas.DataFrame |
import pandas as pd
import pytest
from pandas.testing import assert_frame_equal
from sklearn.pipeline import Pipeline
from hcrystalball.feature_extraction import HolidayTransformer
@pytest.mark.parametrize(
"X_y_with_freq, country_code, country_code_column, country_code_column_value, extected_error",
[
... | pd.DataFrame(expected, index=X.index) | pandas.DataFrame |
"""Helper Functions to Support Pairs Trading
This file can be imported as a module and contains the following functions:
* create_and_save_historicals - returns a df with all coin information
* binance_data_to_df - historical information for a single coin
* two_coin_pricing - historical log pricing of two ... | pd.DataFrame(klines, columns = ['timestamp', 'open', 'high', 'low', 'close', 'volume', 'close_time', 'quote_av', 'trades', 'tb_base_av', 'tb_quote_av', 'ignore' ]) | pandas.DataFrame |
import numpy as np
import numpy.testing as npt
import pandas as pd
from stumpy import aampi, core, config
import pytest
import naive
substitution_locations = [(slice(0, 0), 0, -1, slice(1, 3), [0, 3])]
substitution_values = [np.nan, np.inf]
def test_aampi_int_input():
with pytest.raises(TypeError):
aampi... | pd.Series(T) | pandas.Series |
#!/usr/bin/env python
# coding: utf-8
# # Starbucks Capstone Challenge
#
# ### Introduction
#
# This data set contains simulated data that mimics customer behavior on the Starbucks rewards mobile app. Once every few days, Starbucks sends out an offer to users of the mobile app. An offer can be merely an advertisemen... | pd.read_json('data/profile.json', orient='records', lines=True) | pandas.read_json |
import inspect
import os
from unittest.mock import MagicMock, patch
import numpy as np
import pandas as pd
import pytest
import woodwork as ww
from evalml.model_understanding.graphs import visualize_decision_tree
from evalml.pipelines.components import ComponentBase
from evalml.utils.gen_utils import (
SEED_BOUND... | pd.Series([1, 2, 3, 4], dtype="Int64") | pandas.Series |
"""
Functions to scrape by season, games, and date range
"""
import hockey_scraper.json_schedule as json_schedule
import hockey_scraper.game_scraper as game_scraper
import hockey_scraper.shared as shared
import pandas as pd
import time
import random
# This hold the scraping errors in a string format.
# This may seem... | pd.concat(master_pbps) | pandas.concat |
# %%
import airtablecache.airtablecache as AC
import pandas as pd
import os
import sys
testDataLoc = os.path.join(os.path.dirname(sys.modules['airtablecache'].__file__), 'data/testData.csv')
sampleDF = pd.DataFrame([
['Alpha', 10, 'India'],
['Beta', 15, 'Australia']
], columns=['Name', 'Age', 'Country'])
new... | pd.testing.assert_frame_equal(df, sampleDF) | pandas.testing.assert_frame_equal |
import random
import timeit
from decimal import Decimal
import h5py
import hdf5plugin
import numpy as np
import pandas as pd
import gym
from gym import logger
from gym import spaces
import matplotlib.pyplot as plt
import os
from decimal import getcontext
os.environ['KMP_DUPLICATE_LIB_OK'] = 'True'
| pd.set_option('display.float_format', lambda x: '%.10f' % x) | pandas.set_option |
from datetime import datetime
from pandas.compat import range, long, zip
from pandas import compat
import re
import numpy as np
from pandas.core.algorithms import unique
from pandas.tseries.offsets import DateOffset
from pandas.util.decorators import cache_readonly
import pandas.tseries.offsets as offsets
import pand... | compat.iteritems(_reso_str_map) | pandas.compat.iteritems |
# Title: Weather Data Aggregator
# Description: Aggregates data from the weather station on Cockcroft from the OnCall API.
# Author: <NAME>
# Date: 17/12/2020
# Version: 1.0
# Import libraries
import pandas as pd
from pandas import json_normalize
import json
import requests
from datetime import datetime, timedelta
fr... | pd.DataFrame() | pandas.DataFrame |
import typer
import spotipy
import pandas as pd
import os
from loguru import logger
from spotify_smart_playlists.helpers import spotify_auth
from toolz import thread_last, mapcat, partition_all
from typing import List
def main(library_file: str, artists_file: str):
logger.info("Initializing Spotify client.")
... | pd.read_csv(artists_file) | pandas.read_csv |
from datetime import datetime, timedelta
import unittest
from pandas.core.datetools import (
bday, BDay, BQuarterEnd, BMonthEnd, BYearEnd, MonthEnd,
DateOffset, Week, YearBegin, YearEnd, Hour, Minute, Second,
format, ole2datetime, to_datetime, normalize_date,
getOffset, getOffsetName, inferTimeR... | Week(weekday=2) | pandas.core.datetools.Week |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
# This file is part of CbM (https://github.com/ec-jrc/cbm).
# Author : <NAME>
# Credits : GTCAP Team
# Copyright : 2021 European Commission, Joint Research Centre
# License : 3-Clause BSD
import pandas as pd
import matplotlib.dates as mdates
from matplotlib impor... | pd.read_csv(index_csv_file) | pandas.read_csv |
# -*- coding: utf-8 -*-
import os
import numpy as np
import pandas as pd
from dramkit.gentools import isnull
from dramkit.iotools import load_csv
from dramkit.datetimetools import today_date
from dramkit.datetimetools import get_date_format
from dramkit.datetimetools import date_reformat
from dramkit.datetimetools imp... | pd.DataFrame(df.iloc[0, :]) | pandas.DataFrame |
# from IPython.core.display import display, HTML
# display(HTML("<style>.container { width:100% !important; }</style>"))
_ = None
import argparse
import json as J
import os
import shutil
import tempfile
import joblib
import mlflow
import functools as F
from importlib import reload as rl
import copy
import pandas as ... | pd.Series(mp) | pandas.Series |
# coding=utf-8
# pylint: disable-msg=E1101,W0612
from datetime import timedelta
from numpy import nan
import numpy as np
import pandas as pd
from pandas import (Series, isnull, date_range,
MultiIndex, Index)
from pandas.tseries.index import Timestamp
from pandas.compat import range
from pandas.u... | assert_series_equal(result, expected) | pandas.util.testing.assert_series_equal |
import os
from os import listdir
from os.path import isfile, join
import re
from path import Path
import numpy as np
import pandas as pd
from poor_trader import utils
from poor_trader.utils import quotes_range
from poor_trader.config import INDICATORS_OUTPUT_PATH
def _true_range(df_quotes, indices):
cur = df_qu... | pd.DataFrame(columns=['top', 'mid', 'bottom'], index=df_quotes.index) | pandas.DataFrame |
# coding: utf-8
# In[ ]:
import numpy as np
import numpy.matlib as npml
import pandas as pd
import statistics as st
from copy import deepcopy
import networkx as nx
import simpy
import matplotlib.pyplot as plt
from simplekml import Kml, Style # for graph_kml
import math
import shapely.geometry
import pyproj
im... | pd.DataFrame() | pandas.DataFrame |
# ---
# jupyter:
# jupytext:
# text_representation:
# extension: .py
# format_name: light
# format_version: '1.5'
# jupytext_version: 1.3.4
# kernelspec:
# display_name: Python 3
# language: python
# name: python3
# ---
import matplotlib.pyplot as plt
import pandas as pd
imp... | pd.merge(average_combined_df,rev_drug_info.iloc[:,[4,5,6]],on='standard_inchi_key') | pandas.merge |
""" This file process the IO for the Text similarity index processor """
import math
import os
import pandas as pd
from similarity_processor.similarity_core import get_cosine
from similarity_processor.similarity_core import text_to_vector
import similarity_processor.similarity_logging as cl
LOG = cl.get_logger()
def ... | pd.ExcelWriter('%s.xlsx' % file_path, engine='xlsxwriter') | pandas.ExcelWriter |
#!/usr/bin/python3
# -*- coding: utf-8 -*-
# *****************************************************************************/
# * Authors: <NAME>
# *****************************************************************************/
"""transformCSV.py
This module contains the basic functions for creating the content of... | pandas.StringDtype() | pandas.StringDtype |
"""
"""
__version__='192.168.3.11.dev1'
import sys
import os
import logging
import pandas as pd
import re
import numpy as np
import matplotlib
import matplotlib.pyplot as plt
from matplotlib.backends.backend_pdf import PdfPages
logger = logging.getLogger('PT3S')
try:
from PT3S import Rm
except ImportError:... | pd.Timedelta('0 seconds') | pandas.Timedelta |
# -*- coding: utf-8 -*-
# pylint: disable=W0612,E1101
from datetime import datetime
import operator
import nose
from functools import wraps
import numpy as np
import pandas as pd
from pandas import Series, DataFrame, Index, isnull, notnull, pivot, MultiIndex
from pandas.core.datetools import bday
from pandas.core.n... | tm.assert_produces_warning(False) | pandas.util.testing.assert_produces_warning |
import numpy as np
import pandas as pd
from yews.cpic import pick
def detects2table(results_dict, wl, g, include_all=False, starttime=None):
'''
Converts dictionary of results from detect() function to pandas DataFrame
object. Columns include starttime, endtime, p probability and s probability
for each... | pd.DataFrame(data) | pandas.DataFrame |
import os, random
import pandas as pd
import re
import sys
import codecs
from shutil import copyfile
from datetime import datetime
def clean_text(text):
text = text.replace("<p>", "").replace("</p>", "\n")
return re.sub('\.+', ".", text)
def filecount(dir):
return len([f for f in os.listdir(dir)])
def... | pd.DataFrame(columns=['article_id', 'year']) | pandas.DataFrame |
import pandas as pd
import numpy as np
import warnings as w
import math
from scipy.signal import savgol_filter
from copy import deepcopy, copy
from fO2calculate import core
from fO2calculate import fO2bufferplotter
from fO2calculate import tavern as tv
def calc_dIW_from_fO2(P, T, fO2):
"""Translates from absolute... | pd.concat([_masses_data, fO2_data], axis=1) | pandas.concat |
import pandas as pd
import glob
import os
import numpy as np
import time
import fastparquet
import argparse
from multiprocessing import Pool
import multiprocessing as mp
from os.path import isfile
parser = argparse.ArgumentParser(description='Program to run google compounder for a particular file and setting')
parse... | pd.melt(cur_df,id_vars=['head','year','count'],value_vars=['modifier','w1','w2','w3'],value_name='context') | pandas.melt |
from django.shortcuts import render
from django.views.generic import FormView, UpdateView
from django.http import HttpResponse, JsonResponse, HttpResponseRedirect
from common.mixins import JSONResponseMixin, AdminUserRequiredMixin
from common.utils import get_object_or_none
from common.tools.data_fit import iter_regres... | pd.DataFrame(showjson) | pandas.DataFrame |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Python Script related to:
Deep Neural Network model to predict the electrostatic parameters in the polarizable classical Drude oscillator force field
<NAME>, <NAME>, <NAME> and <NAME>.
Copyright (c) 2022, University of Maryland Baltimore
"""
import numpy as np
impor... | pd.read_pickle('dgenff_dataset.2021/test_alphathole_target.pkl') | pandas.read_pickle |
# coding=utf-8
# pylint: disable-msg=E1101,W0612
from datetime import datetime, timedelta
import operator
from itertools import product, starmap
from numpy import nan, inf
import numpy as np
import pandas as pd
from pandas import (Index, Series, DataFrame, isnull, bdate_range,
NaT, date_range, ti... | Series([False, False, False]) | pandas.Series |
# -*- coding: utf-8 -*-
"""
Created on Fri Oct 29 09:20:13 2021
@author: bw98j
"""
import prose as pgx
import pandas as pd
import matplotlib.pyplot as plt
import matplotlib.colors as colors
import seaborn as sns
import numpy as np
import itertools
import glob
import os
import random
from tqdm import ... | pd.read_csv('interim_files/klijn_panel_spearmanCorr.tsv', sep='\t',index_col=0) | pandas.read_csv |
from distutils.util import execute
import sqlite3
from datetime import datetime, timedelta, tzinfo
from importlib import resources
from pathlib import Path
from sqlite3.dbapi2 import OperationalError, Row
from typing import Any, Dict, List
import numpy as np
import pandas as pd
from DailyData import __version__
from D... | pd.DataFrame(fetch, columns=columns) | pandas.DataFrame |
import numpy as np
import cvxpy as cp
import pandas as pd
from scoring import *
# %%
def main():
year = int(input('Enter Year: '))
week = int(input('Enter Week: '))
budget = int(input('Enter Budget: '))
source = 'NFL'
print(f'Source = {source}')
df = read_data(year=year, week=week, source=sourc... | pd.get_dummies(df['pos']) | pandas.get_dummies |
# -*- coding: utf-8 -*-
"""
Tests the usecols functionality during parsing
for all of the parsers defined in parsers.py
"""
import nose
import numpy as np
import pandas.util.testing as tm
from pandas import DataFrame, Index
from pandas.lib import Timestamp
from pandas.compat import StringIO
class UsecolsTests(obj... | StringIO(s) | pandas.compat.StringIO |
# -*- coding: utf-8 -*-
"""
@file:maketrain.py
@time:2019/5/6 16:42
@author:Tangj
@software:Pycharm
@Desc
"""
import pandas as pd
import numpy as np
import gc
import time
name = ['log_0_1999', 'log_2000_3999', 'log_4000_5999','log_6000_7999', 'log_8000_9999', 'log_10000_19999',
'log_20000_29999', 'log_30000_39... | pd.concat([Train, train2]) | pandas.concat |
import pandas as pd
import tkinter as tk
from tkinter import filedialog
from tkinter import messagebox
import os
import joblib
import json, codecs
import numpy as np
from sklearn.cross_decomposition import PLSRegression
from datetime import date
import Classes.Configurations as cfg
from Classes import Configurations
... | pd.DataFrame(df_y_resume) | pandas.DataFrame |
from bs4 import BeautifulSoup
import os
import pandas as pd
HTML_FOLDER = 'yeet/'
file_names = os.listdir(HTML_FOLDER)
final_list = [] # TODO: find some other solution other than a global variable
def extract_data(soup):
'''
Function that takes in a single HTML file and returns a DataFrame with the required... | pd.DataFrame(final_list, columns=headings) | pandas.DataFrame |
# -*- coding: utf-8 -*-
"""
Created on Tue Sep 1 15:26:36 2020
@author: <NAME>
"""
import pandas as pd
import numpy as np
data_ME = pd.read_csv("Portfolios_Formed_on_ME_monthly_EW.csv", index_col = 0)
def returns_anual(data_col):
returns = data_col / 100
n_months = returns.shape[0]
returns_annualized ... | pd.to_datetime(ind.index, format="%Y%m") | pandas.to_datetime |
# coding=utf-8
# pylint: disable-msg=E1101,W0612
import nose
import numpy as np
import pandas as pd
from pandas import (Index, Series, _np_version_under1p9)
from pandas.tseries.index import Timestamp
from pandas.types.common import is_integer
import pandas.util.testing as tm
from .common import TestData
class Test... | Series([np.nan], index=[0.5]) | pandas.Series |
from preprocessing.utils_communes import (
build_and_clean_df,
label_encoders_generator,
encode_df,
decode_df,
)
import pandas as pd
from preprocessing.preprocessing import (
standardize_education_level,
standardize_date,
standardize_tailmen,
standardize_bool_hors_nk,
standardize_mou... | pd.read_csv(aggregated_file, low_memory=False) | pandas.read_csv |
import numpy as np
import pandas as pd
from sklearn.model_selection import train_test_split
from sklearn.preprocessing import StandardScaler, MinMaxScaler
from sklearn import metrics
from log import logger
def split_train_test(x, y, test_size, seed):
idx_norm = y == 0
idx_out = y == 1
n_f = x.shape[1]
... | pd.DataFrame(matrix, columns=columns) | pandas.DataFrame |
import os
os.environ['TF_CPP_MIN_LOG_LEVEL']='2'
import numpy as np # linear algebra
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import matplotlib.pyplot as plt
import matplotlib.image as mpimg
import keras
from keras.datasets import mnist
from keras.models import Sequential
from keras.laye... | pd.get_dummies(train_ds['label']) | pandas.get_dummies |
# -*- coding:utf-8 -*-
# !/usr/bin/env python
"""
Date: 2021/11/2 21:08
Desc: 同花顺-数据中心-技术选股
http://data.10jqka.com.cn/rank/cxg/
"""
import pandas as pd
import requests
from bs4 import BeautifulSoup
from py_mini_racer import py_mini_racer
from tqdm import tqdm
from akshare.datasets import get_ths_js
def _get_file_co... | pd.DataFrame() | pandas.DataFrame |
# Authors: <NAME> <<EMAIL>>
# License: BSD 3 clause
import pandas as pd
import pytest
from categorical_encoder import CategoricalEncoder, RareLabelEncoder
def test_CategoricalEncoder_count():
df = {'category': ['A'] * 10 + ['B'] * 6 + ['C'] * 4,
'target' : [1,1,0,0,0,0,0,0,0,0,1,1,0,0,0,0, 1,1,0,0,]}
... | pd.DataFrame(transf_df) | pandas.DataFrame |
# -*- coding:utf-8 -*-
# /usr/bin/env python
"""
Date: 2021/7/12 15:47
Desc: 东方财富-沪深板块-概念板块
http://quote.eastmoney.com/center/boardlist.html#concept_board
"""
import requests
import pandas as pd
def stock_board_concept_name_em() -> pd.DataFrame:
"""
东方财富-沪深板块-概念板块-名称
http://quote.eastmoney.com/center/boar... | o_numeric(temp_df["开盘"]) | pandas.to_numeric |
import os
import pickle
import sys
from pathlib import Path
from typing import Union
import matplotlib.pyplot as plt
import numpy as np
import pandas as pd
from Bio import pairwise2
from scipy import interp
from scipy.stats import linregress
from sklearn.metrics import roc_curve, auc, precision_recall_curve
import th... | pd.Series(PR_AUC_dict) | pandas.Series |
import pandas as pd
from glob import glob
from collections import defaultdict
import csv
import time
import subprocess
import argparse
parser = argparse.ArgumentParser()
parser.add_argument('start_well_index', help='key to identify which wells to do')
args = parser.parse_args()
start_well_index = int(args.start_well_i... | pd.concat(dats) | pandas.concat |
import os
import pandas as pd
import pytest
from pandas.testing import assert_frame_equal
from .. import read_sql
@pytest.fixture(scope="module") # type: ignore
def mssql_url() -> str:
conn = os.environ["MSSQL_URL"]
return conn
@pytest.mark.xfail
def test_on_non_select(mssql_url: str) -> None:
query ... | pd.Series([None, -2.23e-308, 1.79e308], dtype="float") | pandas.Series |
import pandas as pd
class BarBase(object):
pass
class Current_bar(BarBase):
def __init__(self):
self._cur_bar_list = []
def add_new_bar(self, new_bar):
"添加新行情,会缓存第n条当前行情,和第n+1条行情,共两条"
self._cur_bar_list.pop(0) if len(self._cur_bar_list) == 2 else None
self._cur_bar_list.... | pd.DataFrame(self._bar_dict[self.instrument]) | pandas.DataFrame |
import pandas as pd
import numpy as np
from flask_socketio import SocketIO, emit
import time
import warnings
warnings.filterwarnings("ignore")
import pandas as pd
import numpy as np
import ast
from sklearn.metrics import mean_absolute_error,mean_squared_error
from statsmodels.tsa import arima_model
from statsmodels.ts... | pd.isnull(data) | pandas.isnull |
import streamlit as st
import pandas as pd
import requests
import numpy as np
import wordcloud
import itertools
# Try this?
# https://towardsdatascience.com/topic-modelling-in-python-with-nltk-and-gensim-4ef03213cd21 ... pyLDAvis
num_words = 150
document_limit = 5000
st_time_to_live = 4 * 3600
# Display options
st... | pd.read_json(js["topics"], orient="split") | pandas.read_json |
import pandas as pd
import numpy as np
DATA_PATH = 'rawdata/' #where the raw data files are
ALT_OUTPUT_PATH = 'alt_output/' #there will be many files produced -- the files that are not "the main ones"
# are placed in this directory
feasDF = pd.read_csv(DATA_PATH+"mipdev_feasibility.csv") #read ... | pd.read_csv(DATA_PATH+"regions_time_data.csv") | pandas.read_csv |
import numpy as np
import pandas as pd
import remixt.bamreader
import os
empty_data = {
'fragments': remixt.bamreader.create_fragment_table(0),
'alleles': remixt.bamreader.create_allele_table(0),
}
def _get_key(record_type, chromosome):
return '/{}/chromosome_{}'.format(record_type, chromosome)
def ... | pd.HDFStore(in_filename, 'r') | pandas.HDFStore |
# -*- coding: utf-8 -*-
"""
Created on Tue Apr 9 18:48:55 2019
@author: shday
"""
import math
from collections import namedtuple
import pandas as pd
import numpy as np
from dateutil import parser
import dash_html_components as html
import plotly.express as px
import pytz
import plotly.graph_objects as... | pd.to_datetime(tweet_data['created_at'], errors='coerce') | pandas.to_datetime |
import numpy as np
import pandas as pd
def load(path):
df = pd.read_csv(path,
encoding="utf-8",
delimiter=";",
quotechar="'").rename(
columns={
"Text": "text",
"Label": "label"
})
train, dev, test = split_df... | pd.get_dummies(train["label"]) | pandas.get_dummies |
# functions to run velocyto and scvelo
import numpy as np
import pandas as pd
# import velocyto as vcy
# import scvelo as scv
from scipy.sparse import csr_matrix
import matplotlib.pyplot as plt
from .moments import *
from anndata import AnnData
def vlm_to_adata(vlm, n_comps=30, basis="umap", trans_mats=None, cells_i... | pd.DataFrame.from_dict(scvelo_coef, orient="index") | pandas.DataFrame.from_dict |
#!/usr/bin/env python
# -*- coding:utf-8 -*-
"""
Date: 2021/12/30 11:31
Desc: 股票数据-总貌-市场总貌
股票数据-总貌-成交概括
http://www.szse.cn/market/overview/index.html
http://www.sse.com.cn/market/stockdata/statistic/
"""
import warnings
from io import BytesIO
from akshare.utils import demjson
import pandas as pd
import requests
warni... | o_numeric(temp_df['主板B'], errors="coerce") | pandas.to_numeric |
import numpy as np
import pytest
from pandas import DataFrame, Series, concat, isna, notna
import pandas._testing as tm
import pandas.tseries.offsets as offsets
@pytest.mark.parametrize(
"compare_func, roll_func, kwargs",
[
[np.mean, "mean", {}],
[np.nansum, "sum", {}],
[lambda x: np... | tm.assert_almost_equal(result0, result1) | pandas._testing.assert_almost_equal |
#!/usr/bin/env python3
import sys
sys.stderr = open(snakemake.log[0], "w")
import numpy as np
import pandas as pd
import allel
chroms = snakemake.params['chroms']
for chrom in chroms:
vcf = allel.read_vcf(f"results/variants/vcfs/annot.missense.{chrom}.vcf")
pos = vcf['variants/POS']
pos1 = pos+... | pd.DataFrame(data) | pandas.DataFrame |
import unittest
import pandas as pd
from data_profiler.profilers import OrderColumn
from . import test_utils
from unittest.mock import patch, MagicMock
from collections import defaultdict
# This is taken from: https://github.com/rlworkgroup/dowel/pull/36/files
# undo when cpython#4800 is merged.
unittest.case._Assert... | pd.Series(data3) | pandas.Series |
# -*- coding: utf-8 -*-
"""
These the test the public routines exposed in types/common.py
related to inference and not otherwise tested in types/test_common.py
"""
from warnings import catch_warnings, simplefilter
import collections
import re
from datetime import datetime, date, timedelta, time
from decimal import De... | lib.is_datetime64_array(arr) | pandas._libs.lib.is_datetime64_array |
import numpy as np
import pandas as pd
import csv
import os
from datetime import datetime
class Console_export(object):
def __init__(self, path):
self.path = path + "_sim_summary.txt"
def printLog(self, *args, **kwargs):
print(*args, **kwargs)
with open(self.path,'a') as file... | pd.concat([export_df, temp_df], axis=1) | pandas.concat |
# Copyright 2021 The TensorFlow Probability Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law o... | pd.to_datetime(['2012-12-25', '2013-01-01']) | pandas.to_datetime |
import datetime as dt
import pandas as pd
import numpy as np
import re
# Begin User Input Data
report_date = dt.datetime(2020, 8, 31)
wscf_market_value = 194719540.46
aqr_market_value = 182239774.63
delaware_market_value = 151551731.17
wellington_market_value = 149215529.22
qic_cash_market_value = 677011299.30
input... | pd.ExcelWriter(output_directory + 'CIO/#Data/output/holdings/top_holdings.xlsx', engine='xlsxwriter') | pandas.ExcelWriter |
import os
import csv
import pandas as pd
def main(dataset_path, dataset_mode):
data_label_pairs = []
data_path = 'leftImg8bit/'
label_path = 'gtFine/'
for subdirectory in os.listdir(dataset_path + '/' + data_path + dataset_mode):
for image_path in os.listdir(dataset_path + '/' + data_pat... | pd.DataFrame(data_label_pairs) | pandas.DataFrame |
import pandas as pd
from clr import AddReference
AddReference("System")
AddReference("QuantConnect.Algorithm")
AddReference("QuantConnect.Common")
AddReference("QuantConnect.Indicators")
from System import *
from QuantConnect import *
from QuantConnect.Algorithm import *
from QuantConnect.Algorithm.Framework import *... | pd.DataFrame(columns=self.colsU) | pandas.DataFrame |
import pandas as pd
import os
from upload_dados import *
import plotly.express as px
import numpy as np
os.system('cls')
#4. Qual a antecedência média das reservas?
# filtrando todos os dados dos anuncios alugados
data_df = data_df.loc[data_df['booked_on'] != 'blank']
data_df = pd.DataFrame(data_df)
#convertendo os... | pd.to_datetime(data_df['date']) | pandas.to_datetime |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.