prompt stringlengths 19 1.03M | completion stringlengths 4 2.12k | api stringlengths 8 90 |
|---|---|---|
import os
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
import numpy as np
from results.test_dicts import marker_styles, draw_order, metric_ylim, best_function, is_metric_increasing, \
metric_short_name, metric_en_name
excluded_methods = ['Simple split']
RESULTS_ROOT_DIR = 'detailed_r... | pd.concat(metrics_dfs) | pandas.concat |
import sys
from intopt_energy_mlp import intopt_energy
sys.path.insert(0,'../..')
sys.path.insert(0,"../../Interior")
sys.path.insert(0,"../../EnergyCost")
from intopt_energy_mlp import *
from KnapsackSolving import *
from get_energy import *
from ICON import *
import itertools
import scipy as sp
import numpy as np
im... | pd.DataFrame([two_stage_rslt, spo_rslt,qpt_rslt,intopt_rslt]) | pandas.DataFrame |
import json
import pandas as pd
import time
"""
需要一下文件:
1、预测的json:bbox_level{}_test_results.json
2、test集的json:test.json
3、sample_submission.csv
"""
LABLE_LEVEL = 4
SCORE_THRESHOLD = 0.001
def json_to_dict(json_file_dir):
with open(json_file_dir, "r") as json_file:
json_dict = json.load(json_file)
... | pd.Series(series_imageid) | pandas.Series |
"""
Copyright 2019 <NAME>.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing,
software distribut... | pd.Timestamp(state) | pandas.Timestamp |
'''
Created on Jan 11, 2016
@author: jch
'''
import numpy
import pandas
from collections import Mapping, OrderedDict
from blocks.log.log import TrainingLogBase
class _TimeSlice(Mapping):
def __init__(self, time, log):
self._time = time
self._columns = log._columns
assert isinstance(self... | pandas.Series(data, index=col['idx'], dtype=dtype) | pandas.Series |
### Twitter Data Tools
## <NAME>
## Created: 8/15/2018
## Updated: 8/23/2018
import os
import re
import sys
import math
import nltk
import errno
import tarfile
import unidecode
import numpy as np
import pandas as pd
import subprocess as sb
def get_id_sets(data):
parent = list(data['tweet']['tweet_id']['parent'].keys... | pd.read_csv(pathway_T,compression='gzip',sep=',',index_col=0,header=0,dtype=str) | pandas.read_csv |
import requests
from bs4 import BeautifulSoup
from datetime import datetime
import pandas as pd
import datetime
import os
def crawling(id_, page, lastupdate=None):
headers = {
'authority': 'feedback.aliexpress.com',
'cache-control': 'max-age=0',
'upgrade-insecure-requests': '1',
'origin': 'https://f... | pd.date_range(df.index[0], df.index[-1]) | pandas.date_range |
# Press Shift+F10 to execute it or replace it with your code.
# Press Double Shift to search everywhere for classes, files, tool windows, actions, and settings.
import numpy as np
import pandas as pd
import warnings
from sklearn.linear_model import LinearRegression
import scipy.cluster.hierarchy as sch
import datetime
... | pd.DataFrame([volume], columns=data.columns) | pandas.DataFrame |
import numpy as np
import pandas as pd
import random
import time
from sklearn.utils import shuffle
import torch
import torch.nn as nn
import torch.nn.functional as F
import torch.optim as optim
import random
from torch.utils.data import DataLoader
from torch.nn.functional import relu,leaky_relu
from torch.nn import Lin... | pd.DataFrame(scores) | pandas.DataFrame |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
qualifier.py
@author <NAME>
<EMAIL>
* qualifier * -> mpi_prep -> mpi_run -> mpi_read -> jackknifer
Program Use Instructions:
Instructions are found in qualifier.ini, as well as space for user input.
Program Description:
Prepares the random and data catal... | pd.DataFrame() | pandas.DataFrame |
import operator
import re
import numpy as np
import pandas as pd
import utils
def get_sites_from_kd_dict(transcript_id, sequence, kd_dict, overlap_dist):
if len(sequence) < 9:
return pd.DataFrame(None)
mir_info = {
'prev_loc': -100,
'prev_seq': '',
'prev_kd': 100,
'k... | pd.DataFrame({'loc': locs}) | pandas.DataFrame |
from dotenv import load_dotenv
load_dotenv()
import os, re, json, unicodedata
import tweepy
from tweepy import Stream,OAuthHandler
from datetime import datetime, timedelta
import nltk
# nltk.download('stopwords')
from nltk.corpus import stopwords
stop_words = stopwords.words('spanish')
from nltk.tokenize import Twe... | pd.DataFrame(topic_weights) | pandas.DataFrame |
from .indicator import Indicator
import sys
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
"""
Moving Average Crossover
-
Buy when short-term moving average > long-term moving average,
sell when short-term moving average < long-term moving average.
"""
class MovingAverageCrossover(Indicator):
... | pd.DataFrame(index=self.df.index) | pandas.DataFrame |
# -*- coding: utf-8 -*-
"""
Get historical information of a series from /mints. Eg: https://cryptoslam.io/cryptopunks/mints
@author: HP
"""
from selenium import webdriver
from selenium.webdriver.support.ui import Select
import pandas
import time
import requests
from bs4 import BeautifulSoup
from selenium.common.excepti... | pandas.read_html(browser.page_source) | pandas.read_html |
import pandas as pd
import numpy as np
from functools import wraps
import copy
# Pass through pd.DataFrame methods for a (1,1,o,d) shaped triangle:
df_passthru = ['to_clipboard', 'to_csv', 'to_pickle', 'to_excel', 'to_json',
'to_html', 'to_dict', 'unstack', 'pivot', 'drop_duplicates',
'de... | pd.DataFrame(val_array) | pandas.DataFrame |
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
from collections import Counter
from itertools import islice
from gensim.models.doc2vec import TaggedDocument, Doc2Vec
from gensim.parsing.preprocessing import preprocess_string
from sklearn.base import BaseEstimator
from sklearn import utils as ... | pd.read_csv('./whole.csv', index_col=False, header=0) | pandas.read_csv |
from Gridworld import Gridworld
from MonteCarlo import MonteCarlo
import numpy as np
import matplotlib.pyplot as plt
import pandas as pd
import csv
env = Gridworld(shape=[5,5], initialState=25)
print("------------------------------epsilon=0.01-------------------------------------")
MC_1 = MonteCarlo(grid_world = env,... | pd.concat(frames) | pandas.concat |
# -*- coding: utf-8 -*-
import pandas as pd
from flask import Flask, jsonify, render_template
from yahoofinancials import YahooFinancials
import numpy as np
from datetime import date
from sklearn.linear_model import LinearRegression
from sklearn.model_selection import train_test_split
from sklearn import preprocessing... | pd.DataFrame(data=result[ticker]['prices']) | pandas.DataFrame |
"""Основные метрики доходности на базе ML-модели"""
from functools import lru_cache
import numpy as np
import pandas as pd
from local import moex
from metrics.portfolio import CASH
from metrics.portfolio import PORTFOLIO
from metrics.portfolio import Portfolio
from metrics.returns_metrics import AbstractReturnsMetric... | pd.Timestamp(portfolio.date) | pandas.Timestamp |
from collections import namedtuple
import re
import time
import warnings
from geopy.distance import geodesic
from geopy.exc import GeocoderUnavailable
from geopy.geocoders import Nominatim
import pandas
import requests
Location = namedtuple("Location", ["latitude", "longitude"])
def clean_address(problem_address):... | pandas.isnull(in_row["latitude"]) | pandas.isnull |
import numpy as np
import pandas as pd
from sklearn.decomposition import PCA
df=pd.read_csv('TrainingData.csv')
#Change strings to numbers
from sklearn import preprocessing
lb = preprocessing.LabelBinarizer()
wi=lb.fit_transform(np.array(df.loc[:,['Working Ion']]))
cs=lb.fit_transform(np.array(df.loc[:,['Crystal Sys... | pd.DataFrame(newdata) | pandas.DataFrame |
'''
Download PDF files for a series of law chapter numbers
found in our training CSV file provided by partners.
Then extract the chapter texts from those PDFs.
Finally insert the chapter text into a new column 'Text' of our training CSV.
Note: each PDF contains a bit of the previous chapter and following one.
'''
impor... | pd.isna(first_line) | pandas.isna |
from __future__ import print_function
import pickle
import os.path
from googleapiclient.discovery import build
from google_auth_oauthlib.flow import InstalledAppFlow
from google.auth.transport.requests import Request
from collections import Counter
import pandas as pd
from datetime import datetime
import numpy as np
im... | pd.DataFrame(newValList, columns=headers) | pandas.DataFrame |
import pandas as pd
import numpy as np
from custom_stuff import Alone
from sklearn.pipeline import Pipeline
from sklearn.impute import SimpleImputer
from sklearn.preprocessing import OneHotEncoder
from sklearn.compose import ColumnTransformer
from sklearn.ensemble import RandomForestClassifier
from sklearn.model_selec... | pd.read_csv('./train.csv', index_col='PassengerId') | pandas.read_csv |
from __future__ import absolute_import
import datetime
from copy import deepcopy
import numpy as np
import pandas as pd
from scipy.stats import norm
import expan.core.statistics as statx
from expan.core.debugging import Dbg
from expan.core.version import __version__
class Results(object):
"""
A Results ins... | pd.Timestamp(v) | pandas.Timestamp |
"""
This file tests the utilities stored in cassiopeia/data/utilities.py
"""
import unittest
from typing import Dict, Optional
import networkx as nx
import numpy as np
import pandas as pd
from cassiopeia.data import CassiopeiaTree
from cassiopeia.data import utilities as data_utilities
from cassiopeia.preprocess imp... | pd.testing.assert_frame_equal(weight_matrix, expected_weight_matrix) | pandas.testing.assert_frame_equal |
from datetime import timedelta,datetime
import pandas as pd
from database.market import Market
class Analyzer(object):
@classmethod
def pv_analysis(self,portfolio):
stuff = []
total_cash = 100
trades = portfolio.trades
trades = trades[(trades["date"] >= portfolio.start) & (trade... | pd.DataFrame([{"message":"no trades..."}]) | pandas.DataFrame |
import numpy as np
from graspologic.utils import largest_connected_component
import pandas as pd
def get_paired_inds(meta, check_in=True, pair_key="pair", pair_id_key="pair_id"):
pair_meta = meta.copy()
pair_meta["_inds"] = range(len(pair_meta))
# remove any center neurons
pair_meta = pair_meta[pair_... | pd.DataFrame(node_rows) | pandas.DataFrame |
import numpy as np
import pandas as pd
import pytest
from folium import Map, FeatureGroup, Marker, Popup
from folium.utilities import (
validate_location,
validate_locations,
if_pandas_df_convert_to_numpy,
camelize,
deep_copy,
get_obj_in_upper_tree,
parse_options,
)
@pytest.mark.parametri... | pd.Series([5, 3]) | pandas.Series |
from sys import path
from os.path import expanduser
#path.append('/home/ubuntu/StatisticalClearSky/')
path.append('/Users/bennetmeyers/Documents/ClearSky/StatisticalClearSky/')
from statistical_clear_sky.algorithm.iterative_fitting import IterativeFitting
from solardatatools import standardize_time_axis, make_2d, fix_t... | pandas.read_csv(base + 'sys_meta.csv') | pandas.read_csv |
import os
import pathlib
import sys
import warnings
from functools import partial
from io import StringIO
from typing import Optional, TextIO
import click
import numpy as np # type: ignore
import pandas # type: ignore
import tomlkit as toml # type: ignore
from .compaction import compact as _compact
out = partial(... | pandas.read_csv(src, names=("dz", "porosity"), dtype=float, comment="#") | pandas.read_csv |
#!/home/mario/anaconda3/envs/project2_venv/bin python
"""
DESCRIPTION:
An script to retrieve the information generated during
the resquiggling from the fastq files.
"""
import h5py
import os
import pandas as pd
import csv
import numpy as np
from pytictoc import TicToc
from tombo import tombo_helper, tombo_stats, resq... | pd.DataFrame(reads_data, columns=columns) | pandas.DataFrame |
import pandas as pd
import numpy as np
import json
from tqdm import tqdm
from scipy.optimize import minimize
from utils import get_next_gw, time_decay
from ranked_probability_score import ranked_probability_score, match_outcome
class Bradley_Terry:
""" Model game outcomes using logistic distribution """
de... | pd.merge(self.test_games, idx, left_on="team1", right_on="team") | pandas.merge |
#!/usr/bin/env python
import os
import bisect
import sys
import logging
import math
import yaml
import numpy as np
import pandas as pd
import configparser
import shapefile
from collections import defaultdict
from shapely import geometry
from geopy.distance import geodesic
from scipy import stats
from wistl.constants ... | pd.DataFrame(None) | pandas.DataFrame |
import pandas as pd
from utils.save_data import write_csv
def filter_runs_not_us(data_subject):
data_subject['residence'] = data_subject['Current Country of Residence']
runs_not_us = data_subject.loc[
data_subject['residence'] != 'United States', 'run_id']
print(f"""{len(runs_not_us)} runs do n... | pd.isna(data_subject['logK']) | pandas.isna |
# Charting OSeMOSYS transformation data
# These charts won't necessarily need to be mapped back to EGEDA historical.
# Will effectively be base year and out
# But will be good to incorporate some historical generation before the base year eventually
import pandas as pd
import numpy as np
import matplotlib.pyplot as pl... | pd.Categorical(ref_powcap_df1['TECHNOLOGY'], prod_agg_tech[:-1]) | pandas.Categorical |
# Copyright 2021 The ProLoaF Authors. All Rights Reserved.
#
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the... | pd.get_dummies(df.index.month, prefix='month') | pandas.get_dummies |
# -*- coding: utf-8 -*-
"""
Created on Tue May 25 14:13:52 2021
@author: <NAME>
"""
import solarenergy as se
import geocoder
from datetime import datetime
from dateutil.relativedelta import relativedelta
import numpy as np
import timezonefinder
import pandas as pd
import streamlit as st
def getDaysList(start_date, t... | pd.DataFrame(data, index=[0]) | pandas.DataFrame |
import os, glob
import pandas as pd
from datetime import datetime as dt
from pathlib import Path
from emotion_recognition import EmotionRecognizer
from pylab import *
import numpy as np
import seaborn as sn
from progressbar import *
import pickle
import ntpath
from pathlib import Path
import shutil
from sklearn.svm imp... | pd.to_datetime(prob_table_by_sessionNmodel_df['Date'], format="%d/%m/%Y") | pandas.to_datetime |
from unittest import TestCase
import pandas as pd
from datamatch.filters import DissimilarFilter, NonOverlappingFilter
class DissimilarFilterTestCase(TestCase):
def test_valid(self):
f = DissimilarFilter('agency')
index = ['agency', 'uid']
self.assertFalse(f.valid(
pd.Series(... | pd.Series(['123', 0, 4], index=index) | pandas.Series |
import numpy as np
from numpy.core.numeric import zeros_like
import pandas as pd
# [TODO] This code was made in a hurry.
# It can be improved, someday I will. Please excuse me
data = {
"a3": [1.0, 6.0, 5.0, 4.0, 7.0, 3.0,8.0,7.0,5.0],
"class": ["CP", "CP", "CN", "CP", "CN", "CN", "CN", "CP", "CN"]
}
divisio... | pd.crosstab(dfi["a3"], dfi["class"], margins=True, margins_name="Total") | pandas.crosstab |
import pandas as pd
confirmed = 'https://raw.githubusercontent.com/CSSEGISandData/COVID-19/master/csse_covid_19_data' \
'/csse_covid_19_time_series/time_series_covid19_confirmed_global.csv '
recovered = 'https://raw.githubusercontent.com/CSSEGISandData/COVID-19/master/csse_covid_19_data' \
'/cs... | pd.to_datetime(df.index) | pandas.to_datetime |
import pandas as pd
import numpy as np
from sklearn.ensemble import IsolationForest
import matplotlib.pyplot as plt
from scipy import stats
from sklearn.preprocessing import MinMaxScaler
#read data
dataset = | pd.read_csv('raw_dataset.csv',engine='python') | pandas.read_csv |
from datetime import date as dt
import numpy as np
import pandas as pd
import pytest
import talib
import os
from finance_tools_py.simulation import Simulation
from finance_tools_py.simulation.callbacks import talib as cb_talib
from finance_tools_py.simulation import callbacks
@pytest.fixture
def init_global_data():
... | pd.Series.equals(mean, s.data[col_mean]) | pandas.Series.equals |
import pandas
import urllib
import json
import datetime
import pandas
import urllib2
import json
import pymongo
import time
import sys
import datetime
from datetime import date, timedelta
import pymongo
import collections
from bson.json_util import loads
import pandas as pd
import math as m
import nump... | pd.ewma(dataset['price'], span=12) | pandas.ewma |
# -*- coding: utf-8 -*-
"""
Classe for technical analysis of assets.
Created on Sat Oct 31 19:35:28 2020
@author: ryanar
"""
import math
import matplotlib.pyplot as plt
from stock_analysis import StockReader, StockVisualizer, Technical, AssetGroupVisualizer, StockAnalyzer, AssetGroupAnalyzer, StockModeler
from stock... | pd.Series(data) | pandas.Series |
import pandas as pd
import sys
from datetime import datetime
from dotenv import load_dotenv
from os import getcwd, getenv, startfile
from tqdm import tqdm
from tweepy import API, Cursor, OAuthHandler, TweepyException
# Loads .env file
load_dotenv()
cwd = getcwd()
today = datetime.now()
# Gets API Credentials for Tw... | pd.DataFrame(accounts_dict) | pandas.DataFrame |
# -*- coding: utf-8 -*-
'''
Created on Thu Jan 01 16:00:32 2015
@author: LukasHalim
Forked by @edridgedsouza
'''
import sqlite3
import os
import pandas as pd
from contextlib import closing
class Database():
def __init__(self, path='Godwin.db'):
self.path = os.path.abspath(path)
if... | pd.DataFrame(res) | pandas.DataFrame |
import pandas as pd
from statsmodels.stats.diagnostic import acorr_ljungbox
from statsmodels.stats.stattools import jarque_bera
from sklearn.metrics import (
mean_absolute_error,
r2_score,
median_absolute_error,
mean_squared_error,
mean_squared_log_error,
)
class AnnualTimeSeriesSplit():
"""... | pd.DataFrame.from_dict(some_dict[test_dict_name]) | pandas.DataFrame.from_dict |
from distutils.version import LooseVersion
from warnings import catch_warnings
import numpy as np
import pytest
from pandas._libs.tslibs import Timestamp
import pandas as pd
from pandas import (
DataFrame,
HDFStore,
Index,
MultiIndex,
Series,
_testing as tm,
bdate_range,
concat,
d... | tm.makeTimeDataFrame(100064, "S") | pandas._testing.makeTimeDataFrame |
from unittest import TestCase
import pandas as pd
import numpy as np
from moonstone.analysis.statistical_test import (
statistical_test_groups_comparison,
_compute_best_bins_values
)
class TestStatisticalTestFunction(TestCase):
def setUp(self):
self.test_df = pd.Series({
'sample1': ... | pd.testing.assert_frame_equal(matrix, expected_df, check_dtype=False) | pandas.testing.assert_frame_equal |
import csv
import requests
import pandas as pd
FRED_UNEMPLOY = 'https://www.quandl.com/api/v3/datasets/UNEMPLOY/GDPDEF/data.csv?api_key=<KEY>'
with requests.Session() as s:
download = s.get(FRED_UNEMPLOY)
decoded_content = download.content.decode('utf-8')
cr = csv.reader(decoded_content.splitlines(), de... | pd.DataFrame(UNEMPLOY_list) | pandas.DataFrame |
# Simplified and slightly modularized from:
# https://github.com/LeonardoL87/SARS-CoV-2-Model-with-and-without-temperature-dependence
import pickle
import datetime
import math as m
import numpy as np
import pandas as pd
from tqdm.auto import tqdm
from typing import Callable
from functools import partial
from scipy i... | pd.Series(TC) | pandas.Series |
import numpy
import pandas
import scipy
import sklearn.metrics as metrics
from sklearn.model_selection import train_test_split
import statsmodels.api as stats
# The SWEEP Operator
def SWEEPOperator (pDim, inputM, tol):
# pDim: dimension of matrix inputM, positive integer
# inputM: a square and sy... | pandas.get_dummies(thisVar) | pandas.get_dummies |
#%%
import argparse
import os
import tempfile
import mlflow
import mlflow.pytorch
import numpy as np
import optuna
import pandas as pd
import torch
import torch.optim as optim
import torch.utils.data as data
import yaml
from dlkit import models
from dlkit.criterions import Criterion
from estimator impo... | pd.DataFrame(value_list, columns=columns) | pandas.DataFrame |
# coding: utf-8
# # Bike Sharing Dataset Linear Modeling
#
# + Based on Bike Sharing dataset from [UCI Machine Learning Repository](https://archive.ics.uci.edu/ml/datasets/Bike+Sharing+Dataset)
# + This notebook is based upon the hourly data file, i.e. hour.csv
# + This notebook showcases linear modeling using linea... | pd.DataFrame(feature_arr, columns=feature_labels) | pandas.DataFrame |
import pandas as pd
def comparacao(a, b):
if (a == b):
return 'a linha E igual'
else:
return 'a linha NAO e igual'
df = | pd.read_csv('arq.csv', ';', header=0, usecols=["Titulo", "titulo"]) | pandas.read_csv |
# coding=utf-8
# Copyright 2021 The Google Research Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicab... | pd.DataFrame(labels) | pandas.DataFrame |
import pandas as pd
import numpy as np
import os
import itertools
def dados_teses():
diretorio = "/media/hdvm02/bd/007/002/007/002"
teses_anos = sorted(os.listdir(diretorio))
lista_dfs = []
for tese_ano in teses_anos:
csv = os.path.join(diretorio,tese_ano)
teses = pd.read_csv(csv, se... | pd.read_csv(csv_1987, sep=";", encoding='latin-1', on_bad_lines='skip', low_memory=False) | pandas.read_csv |
from collections import OrderedDict
from datetime import datetime, timedelta
import numpy as np
import numpy.ma as ma
import pytest
from pandas._libs import iNaT, lib
from pandas.core.dtypes.common import is_categorical_dtype, is_datetime64tz_dtype
from pandas.core.dtypes.dtypes import (
CategoricalDtype,
Da... | tm.assert_produces_warning(DeprecationWarning, check_stacklevel=False) | pandas._testing.assert_produces_warning |
# -*- coding: utf-8 -*-
from datetime import timedelta
import pytest
import numpy as np
import pandas as pd
import pandas.util.testing as tm
from pandas import (Timedelta,
period_range, Period, PeriodIndex,
_np_version_under1p10)
import pandas.core.indexes.period as period
cla... | pd.offsets.QuarterEnd(n=1, startingMonth=12) | pandas.offsets.QuarterEnd |
import pandas as pd
from product.anaiproduct import AnAIProduct
from datetime import timedelta
import pytz
from tqdm import tqdm
pd.options.mode.chained_assignment = None
from modeler.modeler import Modeler as m
from datetime import datetime, timedelta, timezone
import numpy as np
import math
import pickle
from sklearn... | pd.DataFrame([{}]) | pandas.DataFrame |
import os
import json
import pandas as pd
import numpy as np
from scipy.cluster.hierarchy import linkage, leaves_list
from installed_clients.DataFileUtilClient import DataFileUtil
from installed_clients.WorkspaceClient import Workspace
def get_statistics(df_metadata, result, wgs_dict, dist_col="containment_index",upa... | pd.concat(all_df) | pandas.concat |
from scipy.spatial import distance
import numpy as np
import pandas as pd
import scipy.stats
from utils import *
def alignstrategy(str1,str2,flag):
str1t = strpreprocess(str1,'intlist')
str2t = strpreprocess(str2,'intlist')
str1b = strpreprocess(str1,'bytelist')
str2b = strpreprocess(str2,'bytelist')
... | pd.Series(str2t) | pandas.Series |
# -*- coding: utf-8 -*-
###########################################################################
# we have searched for keywords in the original news
# for stemmed keywords in the stemmed news
# for lemmatized keywords int the lemmatized news
# now, want to merge... | pd.merge(selected_news_with_categories2,news[["message_header","trimmed_tag","doc_idx"]], how="left", on='doc_idx') | pandas.merge |
from sklearn.tree import DecisionTreeClassifier
import pandas as pd
import seaborn as sns
import matplotlib.pyplot as plt
def plot_feature_importance(X, Y):
clf = DecisionTreeClassifier()
clf.fit(X, Y)
features = X.columns.values
importances = clf.feature_importances_
df_plot = | pd.DataFrame() | pandas.DataFrame |
import os
import pandas as pd
pd.options.mode.chained_assignment = None
import numpy as np
import matplotlib.pyplot as plt
import matplotlib
import matplotlib.colors as mcolors
import numpy as np
import folium
import difflib
import geopandas as gpd
import unicodedata
#function to remove accents from states and munici... | pd.to_numeric(demand_melt_df['hour']) | pandas.to_numeric |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
from multiprocessing import cpu_count
import pandas as pd
from joblib import Parallel, delayed
from src.geo_mapping_helpers import map_district_to_community_area
def filter_single_year_range_departure_ca_format_datetimes(df):
"""
Replace single station name, ... | pd.to_datetime(df[c]) | pandas.to_datetime |
import numpy as np
import pytest
from pandas._libs import iNaT
from pandas.core.dtypes.common import (
is_datetime64tz_dtype,
needs_i8_conversion,
)
import pandas as pd
from pandas import NumericIndex
import pandas._testing as tm
from pandas.tests.base.common import allow_na_ops
def test_unique(index_or_se... | tm.assert_index_equal(result, expected, exact=True) | pandas._testing.assert_index_equal |
from typing import Optional
import numpy as np
import pandas as pd
import pytest
from pandas import testing as pdt
from rle_array.autoconversion import auto_convert_to_rle, decompress
from rle_array.dtype import RLEDtype
pytestmark = pytest.mark.filterwarnings("ignore:performance")
@pytest.mark.parametrize(
"o... | pd.Series([1], dtype=np.int32) | pandas.Series |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Wed Nov 4 00:13:06 2020
@author: sahand
"""
from rake_nltk import Rake
import pandas as pd
import re
import numpy as np
from sklearn.feature_extraction.text import TfidfVectorizer
import nltk
from nltk.corpus import stopwords
nltk.download('stopwords')
st... | pd.read_csv(path+'Corpus/AI 4k/embeddings/clustering/k10/Doc2Vec patent_wos_ai corpus DEC 200,500,10 k10 labels') | pandas.read_csv |
import requests
import pandas as pd
from datetime import datetime
def get_irradiance_next_hour():
# API parametros
baseurl = 'http://dataservice.accuweather.com/forecasts/v1/hourly/1hour/'
location_key = '310683'
apikey = '<KEY>'
parameters = {
'apikey': apikey,
'details': 'true'... | pd.DataFrame(new_row) | pandas.DataFrame |
#using TA-Lib to create technical analysis / charts / patterns
#package imports
import pandas as pd
import numpy as np
from pandas_datareader import DataReader
import math
import os
import path
import matplotlib
import matplotlib.pyplot as plt
from matplotlib import style
import time
import datetim... | pd.read_csv('TWTR.csv', parse_dates=True, index_col=0) | pandas.read_csv |
"""Test ir_dist._util utility functions"""
from scirpy.ir_dist._util import (
DoubleLookupNeighborFinder,
reduce_and,
reduce_or,
merge_coo_matrices,
)
import pytest
import numpy as np
import scipy.sparse as sp
import pandas as pd
import numpy.testing as npt
@pytest.fixture
def dlnf_square():
clon... | pd.DataFrame() | pandas.DataFrame |
from __future__ import division
from contextlib import contextmanager
from datetime import datetime
from functools import wraps
import locale
import os
import re
from shutil import rmtree
import string
import subprocess
import sys
import tempfile
import traceback
import warnings
import numpy as np
from numpy.random i... | Index([False, True] + [False] * (k - 2), name=name) | pandas.Index |
"""analysis.py: Collection of classes for performing analysis on Corpus"""
# <NAME> (<EMAIL>)
# DS 5001
# 6 May 2021
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
from pandas.core.algorithms import mode
import plotly.express as px
import scipy.cluster.hierarchy as sch
from gensim.models import... | pd.MultiIndex.from_product([work_ids, work_ids]) | pandas.MultiIndex.from_product |
import pandas as pd
import numpy as np
#import sys
#sys.path.append("F:\3RDSEM\DM\Assignment_1\DM-Project\Assignment-1\Code")
from Utility import getDataFrame
fileNames = ["./../DataFolder/CGMSeriesLunchPat1.csv", "./../DataFolder/CGMSeriesLunchPat2.csv",
"./../DataFolder/CGMSeriesLunchPat3.csv", "... | pd.concat([df, feature_1_df, feature_2_df, feature_3_df, feature_4_df, feature_5_df, feature_6_df], axis=1) | pandas.concat |
"""
Makes a figure providing an overview of our dataset with a focus on lineages
laid out as follows:
a - Patient metadata
b - Donut plot of our lineage distributions vs the world
c - Timeline of patient sampling vs lineages identified
d - Choropleth of lineages by region
"""
import matplotlib.pyplot as plt
import n... | pd.read_csv("data/external/pangolin2.csv") | pandas.read_csv |
#!/usr/bin/env python
# coding: utf-8
# In[ ]:
import pysam
import os
import pandas as pd
import numpy as np
import time
import argparse
import sys
from multiprocessing import Pool
# In[ ]:
# ##arguments for testing
# bam_file_path = '/fh/scratch/delete90/ha_g/realigned_bams/cfDNA_MBC_ULP_hg38/realign_bam_pa... | pd.DataFrame() | pandas.DataFrame |
import pytest
import numpy as np
import pandas as pd
import databricks.koalas as ks
from pandas.testing import assert_frame_equal
from gators.feature_generation.polynomial_features import PolynomialFeatures
ks.set_option('compute.default_index_type', 'distributed-sequence')
@pytest.fixture
def data_inter():
X = p... | assert_frame_equal(X_new, X_expected) | pandas.testing.assert_frame_equal |
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Fri Jan 14 10:52:33 2022
COVID-19 DEATHS IN US - COUNTY
Author: <NAME> (<EMAIL>)
"""
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
from pandas.plotting import register_matplotlib_converters
register_matplotlib_converters()
muertes_us... | pd.to_datetime(muertes_us.index,dayfirst=False,yearfirst=False) | pandas.to_datetime |
import time
import numpy as np
import pandas as pd
import logging
from scipy.sparse import issparse, csr_matrix
from scipy.stats import chi2
from sklearn.neighbors import NearestNeighbors
from anndata import AnnData
from joblib import effective_n_jobs
from typing import List, Tuple
from pegasus.tools import update_re... | pd.Series(attr_values[knn_indices[i, :]]) | pandas.Series |
import os
import string
import pandas as pd
from sklearn.feature_extraction.text import TfidfVectorizer
from common.utils import DATA_DIR
pos_tags_features = [
'noun',
'verb',
'adjective',
'adverb'
]
cos_sim_features = [
'cos_sim'
]
sentiment_features = [
'positive_count',
'negative_count... | pd.concat([features, cols], axis=1) | pandas.concat |
"""
:noindex:
preprocess.py
====================================
Script to convert provider datasets individual record dictionaries
Data Sources:
`https://www.acaps.org/covid-19-government-measures-dataset <https://www.acaps.org/covid-19-government-measures-dataset>`_
`https://www.cdc.gov/mmwr/preview/mmwrhtml/0000159... | pd.to_datetime(cdc["Date Entered"], format='%d/%m/%Y') | pandas.to_datetime |
import numpy as np
import pandas as pd
from sklearn.model_selection import cross_validate, train_test_split
from classification.utils import print_params
from sklearn import metrics
from config import CHANNEL_NAMES
from data.utils import prepare_dfs
# @print_params
def predict(lab, ba, cols, estimator, metapkl, gs=... | pd.DataFrame(y[y_pred != y]) | pandas.DataFrame |
# -*- coding: utf-8 -*-
"""
Functions to process raw respondent data from new survey
Convert that to habits
Find top n best matches from 'archetypes' of original survey of 10K people.
Generate habit lists to see overlap with best cluster, unique habits, etc.
Clean habit naming for consisntency.
"""
import sys
sys.pat... | pd.set_option('display.expand_frame_repr', False) | pandas.set_option |
"""
The :mod:`mlshells.model_selection.search` includes utilities to
optimize hyper-parameters.
:class:`mlshell.model_selection.Optimizer` class proposes unified interface to
arbitrary optimizer. Intended to be used in :class:`mlshell.Workflow` . For new
optimizer formats no need to edit `Workflow` class, just adapt i... | pd.DataFrame(optimizer.cv_results_) | pandas.DataFrame |
import pandas as pd
import pytest
| pd.set_option("display.max_rows", 500) | pandas.set_option |
import pandas as pd
import numpy as np
from tqdm import tqdm
from lib import config
from lib.logger import Log
logger = Log()
# Historical data columns
_DATE_COLUMN = 'DATE'
_N1 = 'N1'
_N2 = 'N2'
_N3 = 'N3'
_N4 = 'N4'
_N5 = 'N5'
_N6 = 'N6'
# Result columns
_DRAW_COLUMN = 'draw'
_MAX_SUCCESS_COLUMN = 'max_success'
_... | pd.DataFrame() | pandas.DataFrame |
#!/usr/bin/env python
# coding: utf-8
# In[2]:
import os
import sys
import pandas as pd
import numpy as np
# In[3]:
# In[4]:
#file 불러오기
#filepath = sys.argv[1]
#filename = sys.argv[2]
filepath = "/home/data/projects/rda/workspace/rda/files/"
filename = "input3.csv"
data = pd.read_csv(filepath + "/" + filena... | pd.Series(silhouette) | pandas.Series |
'''Module to assemble custom tensorflow procedures.'''
from dataclasses import dataclass, field
import tensorflow_datasets as tfds
import pandas as pd
from pyspark.sql import functions as F
from pyspark.sql import (
SparkSession,
functions as F,
DataFrame
)
from pyspark.sql.types import *
@dataclass
... | pd.DataFrame.from_dict(ex['data']) | pandas.DataFrame.from_dict |
import os
import numpy as np
import pandas as pd
import z5py
from mobie import add_segmentation
from mobie.metadata.image_dict import load_image_dict
ROOT = '/g/kreshuk/pape/Work/data/mito_em/data'
RESOLUTION = [.03, .008, .008]
SCALE_FACTORS = [[1, 2, 2], [1, 2, 2], [2, 2, 2], [2, 2, 2]]
def compute_object_scores(s... | pd.DataFrame(data, columns=columns) | pandas.DataFrame |
# -*- coding: utf-8 -*-
import numpy as np
import pandas as pd
import pytest
from ..testing_utils import make_ecommerce_entityset
from featuretools import Timedelta
from featuretools.computational_backends import PandasBackend
from featuretools.primitives import (
Absolute,
Add,
Count,
CumCount,
... | pd.isnull(a) | pandas.isnull |
# Copyright (c) 2016 <NAME> <<EMAIL>>
#
# This module is free software. You can redistribute it and/or modify it under
# the terms of the MIT License, see the file COPYING included with this
# distribution.
""" Module for motif activity prediction """
def warn(*args, **kwargs):
pass
import warnings
warnings.wa... | pd.read_feather(inputfile) | pandas.read_feather |
import numpy as np
import pandas as pd
import pickle
from sklearn.feature_selection import SelectKBest,f_regression
from sklearn.ensemble import RandomForestRegressor
import matplotlib.pyplot as plt
from sklearn import metrics
###########################
# Folder Name Setting
###########################
folder = 'J:... | pd.DataFrame(RESULT,columns=['ID','NUM_FEATURES','N_ESTIMATORS','MAX_FEATURES','MIN_SAMPLES_SPLIT','CV_FOLDS','AUC']) | pandas.DataFrame |
"""
SparseArray data structure
"""
from __future__ import division
import numbers
import operator
import re
from typing import Any, Callable, Union
import warnings
import numpy as np
from pandas._libs import index as libindex, lib
import pandas._libs.sparse as splib
from pandas._libs.sparse import BlockIndex, IntInd... | is_dtype_equal(ltype, rtype) | pandas.core.dtypes.common.is_dtype_equal |
#+ 数据科学常用工具
import matplotlib as mpl
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
import matplotlib.style as style
import seaborn as sns
from sklearn.preprocessing import PowerTransformer
import category_encoders as ce
from sklearn.model_selection import StratifiedKFold, KFold
from joblib impo... | pd.concat(retLst, axis=1) | pandas.concat |
import warnings
import numpy as np
import pytest
from pandas.errors import PerformanceWarning
import pandas as pd
from pandas import Index, MultiIndex
import pandas._testing as tm
def test_drop(idx):
dropped = idx.drop([("foo", "two"), ("qux", "one")])
index = MultiIndex.from_tuples([("foo", "two"), ("qux... | tm.assert_produces_warning(PerformanceWarning) | pandas._testing.assert_produces_warning |
import xml.etree.ElementTree as ET
import openpyxl
from openpyxl import Workbook, load_workbook
import pandas as pd
from pandas import ExcelWriter
import csv
import numpy as np
def get_armor(xml_file):
# CREATES TREE AND ROOT
tree = ET.parse(xml_file)
root = tree.getroot()
# head_armor = 'Head Armo... | pd.read_excel('MB2.xlsx', sheet_name='shoulder') | pandas.read_excel |
"""
Functions having to do with loading data from output of
files downloaded in scripts/download_data_glue.py
"""
import codecs
import csv
import json
import numpy as np
import pandas as pd
from allennlp.data import vocabulary
from jiant.utils.tokenizers import get_tokenizer
from jiant.utils.retokenize import realign... | pd.read_json(file_name, lines=True) | pandas.read_json |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.