index int64 | repo_name string | branch_name string | path string | content string | import_graph string |
|---|---|---|---|---|---|
87,859 | yl558/fake_news_detect | refs/heads/master | /crawl_twitter_user/get_user_id.py | import sys, os
sys.path.append('..')
import utils
def get_user_ids(line):
# return left uid, right uid
return eval(line.split('->')[0])[0], eval(line.split('->')[1])[0]
def main():
year = 16
root_folder = os.path.join('..', '..')
tree_folder = os.path.join(root_folder, 'dataset', 'twitter', 'twit... | {"/extract_feature/combine_feature_twitter.py": ["/utils.py"], "/model/base_model.py": ["/utils.py"], "/crawl_twitter_user/get_user_id.py": ["/utils.py"], "/crawl_twitter_user/crawl_user.py": ["/utils.py"], "/extract_feature/get_feature_weibo.py": ["/utils.py"], "/extract_feature/get_feature_twitter.py": ["/utils.py"]} |
87,860 | yl558/fake_news_detect | refs/heads/master | /crawl_twitter_user/crawl_user.py | import sys, os, time
sys.path.append('..')
import utils
import requests
from requests_oauthlib import OAuth1Session, OAuth1
def user_exists(user_id, dbc):
sql = 'select * from user where id = ?'
tup = (user_id, )
cursor = dbc.cursor()
cursor.execute(sql, tup)
rs = cursor.fetchall()
if rs:
... | {"/extract_feature/combine_feature_twitter.py": ["/utils.py"], "/model/base_model.py": ["/utils.py"], "/crawl_twitter_user/get_user_id.py": ["/utils.py"], "/crawl_twitter_user/crawl_user.py": ["/utils.py"], "/extract_feature/get_feature_weibo.py": ["/utils.py"], "/extract_feature/get_feature_twitter.py": ["/utils.py"]} |
87,861 | yl558/fake_news_detect | refs/heads/master | /utils.py | import sys, os, fnmatch, datetime, time
import sqlite3
import numpy as np
from scipy import misc
from sklearn import preprocessing
def read(file_path):
if not os.path.exists(file_path):
return None
f = open(file_path, 'r', encoding = 'utf-8', errors = 'ignore')
con = f.read()
f.close()
... | {"/extract_feature/combine_feature_twitter.py": ["/utils.py"], "/model/base_model.py": ["/utils.py"], "/crawl_twitter_user/get_user_id.py": ["/utils.py"], "/crawl_twitter_user/crawl_user.py": ["/utils.py"], "/extract_feature/get_feature_weibo.py": ["/utils.py"], "/extract_feature/get_feature_twitter.py": ["/utils.py"]} |
87,862 | yl558/fake_news_detect | refs/heads/master | /extract_feature/get_feature_weibo.py | import sys, os
sys.path.append('..')
import utils
import json, time, datetime, numpy
project_folder = os.path.join('..', '..')
def get_feature(post):
f = []
f.append(post['followers_count'])
f.append(post['bi_followers_count'])
f.append(post['friends_count'])
f.append(post['statuses_count'])
f.... | {"/extract_feature/combine_feature_twitter.py": ["/utils.py"], "/model/base_model.py": ["/utils.py"], "/crawl_twitter_user/get_user_id.py": ["/utils.py"], "/crawl_twitter_user/crawl_user.py": ["/utils.py"], "/extract_feature/get_feature_weibo.py": ["/utils.py"], "/extract_feature/get_feature_twitter.py": ["/utils.py"]} |
87,863 | yl558/fake_news_detect | refs/heads/master | /extract_feature/get_feature_twitter.py | import sys, os
sys.path.append('..')
import utils
import json, time, datetime, numpy
project_folder = os.path.join('..', '..')
def get_user(user_id, dbc):
sql = '''SELECT json FROM user where id = ?'''
tup = (user_id, )
cursor = dbc.cursor()
rs = cursor.execute(sql, tup).fetchall()
if rs:
r... | {"/extract_feature/combine_feature_twitter.py": ["/utils.py"], "/model/base_model.py": ["/utils.py"], "/crawl_twitter_user/get_user_id.py": ["/utils.py"], "/crawl_twitter_user/crawl_user.py": ["/utils.py"], "/extract_feature/get_feature_weibo.py": ["/utils.py"], "/extract_feature/get_feature_twitter.py": ["/utils.py"]} |
87,864 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/framework/topicdetector.py | import re, numpy as np, pandas as pd
import sys
from pprint import pprint
from gensim.test.utils import datapath
# Gensim
import gensim, spacy, logging, warnings
import gensim.corpora as corpora
import matplotlib
from gensim.utils import lemmatize, simple_preprocess
from gensim.models import CoherenceModel, LdaModel
i... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,865 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/framework/whitelistdetector.py | import json
from collections import Counter
import pandas as pd
from nltk.tokenize import TweetTokenizer
from sklearn.metrics import precision_score, recall_score, f1_score, roc_auc_score
class WhiteListDetector():
"""
Whitle List detector od semi-supervised module
"""
def __init__(self, whiteListUs... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,866 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/framework/batchupdate.py | import pickle
import sys
import numpy as np
import pandas as pd
from sklearn.ensemble import GradientBoostingClassifier, RandomForestClassifier, AdaBoostClassifier
from sklearn.metrics import roc_auc_score, precision_score, recall_score, f1_score, roc_curve, precision_recall_curve
from sklearn.model_selection import t... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,867 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/framework/main.py | import argparse
from logging import Logger
import pandas as pd
from tqdm import tqdm
import numpy as np
from src.collectData.twitterStreamer import TwitterStreamer
from src.framework.batchupdate import BatchUpdate
from src.framework.semiduperviseddetector import SemiSupervisedDetector
from src.framework.twitterStre... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,868 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/mturk/get_results.py | import boto3
import sys
import json
import xmltodict
import pandas as pd
from src.mturk import mturk_credentials
create_hits_in_production = False
environments = {
"production": {
"endpoint": "https://mturk-requester.us-east-1.amazonaws.com",
"preview": "https://www.mturk.com/mturk/preview"
},
"sandbo... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,869 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/framework/__init__.py | from .textdetector import TextDetector
from .superviseddetector import SupervisedDetector
from .clusterdetector import ClusterDetector
from .userdescriptiondetector import UserDescriptionDetector | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,870 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/collectData/activelearner.py | import os
import sys
import numpy
import pandas as pd
from scipy.sparse import csr_matrix
from sklearn.feature_extraction.text import TfidfVectorizer
from sklearn.metrics import precision_score, recall_score, f1_score, roc_auc_score, accuracy_score
from tqdm import tqdm
from ..framework.textdetector import TextDet... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,871 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/framework/evalutaor.py | from collections import Counter
import numpy as np
import pandas as pd
from numpy import loadtxt
from sklearn.metrics import roc_auc_score, roc_curve, cohen_kappa_score, precision_score, recall_score, \
precision_recall_curve
from mlens.visualization import corrmat
import matplotlib.pyplot as plt
import seaborn as... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,872 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/framework/textdetector.py | import pickle
import pandas as pd
from sklearn.ensemble import RandomForestClassifier
from sklearn.feature_extraction.text import CountVectorizer, TfidfTransformer
from sklearn.metrics import roc_auc_score, roc_curve, cohen_kappa_score, precision_score, recall_score, \
precision_recall_curve
from sklearn.pipeline ... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,873 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/framework/semiduperviseddetector.py | import pickle
from sklearn.metrics import roc_auc_score, precision_score, recall_score, f1_score
from src.framework.preprocessor import Preprocessor
from src.framework.blacklistdetector import BlackListUserDetector
from src.framework.textdetector import TextDetector
from src.framework.whitelistdetector import WhiteLi... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,874 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/mturk/create_tasks.py |
import boto3
import json
import pandas as pd
from src.mturk import mturk_credentials
"""
Class for creating Mechanical Mturk task for labeling tweets
"""
create_hits_in_production = False
environments = {
"production": {
"endpoint": "https://mturk-requester.us-east-1.amazonaws.com",
"preview": "https://ww... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,875 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/framework/userdescriptiondetector.py | import pickle
import pandas as pd
from sklearn.feature_extraction.text import CountVectorizer, TfidfVectorizer
from sklearn.linear_model import LogisticRegression
from sklearn.metrics import precision_score, recall_score, f1_score, roc_auc_score
from sklearn.model_selection import KFold, cross_val_score
from sklearn.p... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,876 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/collectData/noiseremover.py | import sys
import pandas as pd
import numpy as np
from collections import defaultdict
# from src.apricot import FeatureBasedSelection
from sklearn.linear_model import LogisticRegression
from sklearn.metrics import precision_score, recall_score, f1_score, roc_auc_score
from snorkel.labeling import labeling_function, L... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,877 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/collectData/collect_hystorical_data.py | import json
from searchtweets import ResultStream, gen_rule_payload, load_credentials
def __init__(self, api=None):
"""
Collect Historical data from Oct 2017 till Nov 2019
"""
self.file = open(OUTPUT_FILE, "w")
dev_env=[
"search_Popo_tweets_fullarchive_dev","search_Ali_tweets_fullarchive_... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,878 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/framework/clusterdetector.py | from sklearn.metrics import precision_score, recall_score, f1_score, roc_auc_score, auc
# from wordcloud import WordCloud
import pandas as pd
from boto import sns
from gensim.models import TfidfModel
from gensim.utils import simple_preprocess
import gensim.corpora as corpora
import gensim.models as models
import re
imp... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,879 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/apricot/__init__.py | # __init__.py
# Author: Jacob Schreiber <jmschreiber91@gmail.com>
from .functions import FeatureBasedSelection
from .functions import MaxCoverageSelection
from .functions import FacilityLocationSelection
from .functions import SaturatedCoverageSelection
from .functions import SumRedundancySelection
from .functions im... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,880 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/collectData/extractdata.py | import json
import re
import sys
import os
import pandas as pd
import ast
from pandas import json_normalize
from collections import Counter
output_path="../../data/"
def merge_Data(input_path,category):
"""
Merge all datas from different spammy hashtags category and assign category labels for them
"""
t... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,881 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/framework/preprocessor.py | import string
import nltk,sys
from nltk.corpus import stopwords
# nltk.download('stopwords')
# nltk.download('wordnet')
import numpy as np,pandas as pd
import nltk.data
from nltk.stem.porter import *
pd.options.mode.chained_assignment = None
class Preprocessor():
def load_data(self,filepath):
return pd.... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,882 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/framework/twitter_credentials.py |
###### Pooneh #######
consumer_key = "key"
consumer_secret = "key"
access_key = "key"
access_secret = "key" | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,883 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/collectData/twitterStreamer.py | import json
import logging
import sys
from pandas.io.json import json_normalize
import pandas as pd
from tweepy.streaming import StreamListener
from tweepy import OAuthHandler
from tweepy import Stream
import datetime
from tweepy import TweepError
consumer_key = ""
consumer_secret = ""
access_token = ""
access_t... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,884 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/mturk/mturk_credentials.py | AWSAccessKeyId="access Key Id"
AWSSecretKey="Secret Key" | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,885 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/framework/baseline.py |
from random import randrange
from sklearn.feature_extraction.text import TfidfVectorizer
from sklearn.metrics import precision_score, recall_score, f1_score, roc_auc_score
import pandas as pd
from string import digits
import nltk
import numpy as np
nltk.download('words')
words = set(nltk.corpus.words.words())
def... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,886 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/collectData/mergedara.py | import pandas as pandas
import pandas as pd
import json
import re
import tweepy
import numpy as np
import matplotlib.pyplot as plt
from matplotlib import style
from pandas.io.json import json_normalize
"""
Class for cleaning collected data
"""
ERR_FILE = ".new-data/error.txt"
SPAM_FILE = "new-data/spam.txt"
HAM_FILE... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,887 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/framework/twitterStreamer.py | import json
from pandas.io.json import json_normalize
import pandas as pd
from tweepy.streaming import StreamListener
from tweepy import OAuthHandler
from tweepy import Stream
import datetime
from tweepy import TweepError
from src.collectData import twitter_credentials
consumer_key = twitter_credentials.consumer_ke... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,888 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/collectData/collect.py | import argparse
import sys
import tweepy
import json
import pandas as pd
def read_parameters():
parser = argparse.ArgumentParser()
parser.add_argument('--inputpath', type=str,required=True,
help='dataset to load.')
parser.add_argument("--outputpath", type=str, required=True,
... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,889 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/framework/blacklistdetector.py | import pandas as pd
import json
from sklearn.metrics import precision_score, recall_score, f1_score, roc_auc_score
class BlackListUserDetector():
"""
BlackListUSer detector od semi-supervised module
"""
def __init__(self, blackListUser_filePath, userDictionary_filepath, threshhold):
self.blac... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,890 | poonehmousavi/Detecting-Hashtag-Hijacking-for-Hashtag-Activism | refs/heads/main | /src/framework/superviseddetector.py |
from sklearn.linear_model import LogisticRegression
from sklearn.metrics import precision_score, recall_score, f1_score, roc_auc_score
from sklearn.naive_bayes import GaussianNB
from sklearn.ensemble import RandomForestClassifier
from sklearn import preprocessing
import numpy as np
import pickle
from collections impor... | {"/src/framework/batchupdate.py": ["/src/framework/blacklistdetector.py", "/src/framework/clusterdetector.py", "/src/framework/preprocessor.py", "/src/framework/semiduperviseddetector.py", "/src/framework/superviseddetector.py", "/src/framework/textdetector.py", "/src/framework/userdescriptiondetector.py", "/src/framew... |
87,925 | anitalaw/api_unittest_scripts | refs/heads/master | /curl.py | import os
import pdb
test = """
# get date
curl -i -H "Content-Type: application/json" -X GET -d'{"full":true}' http://localhost:5000/dates
curl -i -H "Content-Type: application/json" -X GET -d'{"full":false}' http://localhost:5000/dates
curl http://localhost:5000/dates
curl -i -H "Content-Type: application/json" -X ... | {"/unit_tests.py": ["/app.py"]} |
87,926 | anitalaw/api_unittest_scripts | refs/heads/master | /app.py | from flask import Flask, request, make_response, jsonify, abort
from datetime import datetime
app = Flask(__name__)
items = [
{
'id': 1,
'name': 'laptop',
'value': 1000
},
{
'id': 2,
'name': 'chair',
'value': 300,
},
{
'id': 3,
'name... | {"/unit_tests.py": ["/app.py"]} |
87,927 | anitalaw/api_unittest_scripts | refs/heads/master | /unit_tests.py | import unittest
import json
from copy import deepcopy
import app
BASE_URL = 'http://127.0.0.1:5000'
class Testapp(unittest.TestCase):
def setUp(self):
self.backup_items = deepcopy(app.items) # recursively makes a copy of each element
self.client = app.app.test_client() # returns a new instance ... | {"/unit_tests.py": ["/app.py"]} |
87,931 | ShoukunYan/Tianchi_adversarial | refs/heads/master | /adv_gen.py | import os
import torch
from torch.autograd import Variable
import torch.nn as nn
import torch.optim as optim
import torchvision.transforms as transforms
import numpy as np
import torchvision.datasets as datasets
import pandas as pd
from torch.utils.data import DataLoader, Dataset
from skimage import io, color... | {"/adv_gen.py": ["/ImageLoad.py", "/attack.py"], "/clean.py": ["/ImageLoad.py"], "/train.py": ["/ImageLoad.py", "/attack.py"]} |
87,932 | ShoukunYan/Tianchi_adversarial | refs/heads/master | /attack.py | import os
import argparse
import time
import torch
import torch.optim as optim
import torch.nn as nn
import pandas as pd
from skimage import io, color, transform
import numpy as np
from torch.utils.data import Dataset, DataLoader
from torchvision import transforms, utils
import torchvision.models as models
device = to... | {"/adv_gen.py": ["/ImageLoad.py", "/attack.py"], "/clean.py": ["/ImageLoad.py"], "/train.py": ["/ImageLoad.py", "/attack.py"]} |
87,933 | ShoukunYan/Tianchi_adversarial | refs/heads/master | /clean.py | import os
import argparse
import torch
import torch.nn as nn
import pandas as pd
from skimage import io, transform
import numpy as np
from torch.utils.data import Dataset, DataLoader
from torchvision import transforms, utils
import torchvision.models as models
import warnings
from ImageLoad import TianchiDat... | {"/adv_gen.py": ["/ImageLoad.py", "/attack.py"], "/clean.py": ["/ImageLoad.py"], "/train.py": ["/ImageLoad.py", "/attack.py"]} |
87,934 | ShoukunYan/Tianchi_adversarial | refs/heads/master | /ImageLoad.py | import os
import argparse
import torch
import torch.nn as nn
import pandas as pd
from skimage import io, transform
import numpy as np
from torch.utils.data import Dataset, DataLoader
from torchvision import transforms, utils
import torchvision.models as models
import warnings
from skimage import io, color, t... | {"/adv_gen.py": ["/ImageLoad.py", "/attack.py"], "/clean.py": ["/ImageLoad.py"], "/train.py": ["/ImageLoad.py", "/attack.py"]} |
87,935 | ShoukunYan/Tianchi_adversarial | refs/heads/master | /train.py | import os
import argparse
import time
import torch
import torch.nn as nn
import pandas as pd
import numpy as np
from torch.utils.data import Dataset, DataLoader
from torchvision import transforms, utils
import torchvision.models as models
from ImageLoad import TianchiDataset, RandomCrop, Rescale, ToTensor
... | {"/adv_gen.py": ["/ImageLoad.py", "/attack.py"], "/clean.py": ["/ImageLoad.py"], "/train.py": ["/ImageLoad.py", "/attack.py"]} |
87,936 | GordonAnderson/MIPS-Python-Host-app | refs/heads/main | /RFdriver.py | import tkinter as tk
from tkinter import ttk
from tkinter import messagebox
import MIPSobjects
import Util
#
# RF channel objects
#
class RFchannel:
def __init__(self, parent, name, MIPSname, channel, comm, x, y):
self.master = parent
self.name = name
self.MIPSname = MIPSname
self.... | {"/RFdriver.py": ["/MIPSobjects.py"], "/System.py": ["/Comms.py"], "/Program.py": ["/Comms.py", "/Command.py"], "/PSG.py": ["/Comms.py", "/MIPSobjects.py"], "/DCBias.py": ["/MIPSobjects.py"], "/MIPS.py": ["/Comms.py", "/Command.py", "/Scripting.py", "/MIPSobjects.py", "/Program.py", "/DCBias.py", "/DIO.py", "/RFdriver.... |
87,937 | GordonAnderson/MIPS-Python-Host-app | refs/heads/main | /MIPSobjects.py | import tkinter as tk
from tkinter import messagebox
from tkinter import ttk
import Util
#
# Todo:
# 1.) Finish the DCBchannel and test, done
# 2.) Impliment DCBenable, done
# 3.) Impliment DCBoffset, done
# 4.) Impliment DCBiasGroung
# 5.) Build DCbias tab
# 6.) Impliment DIOchannel, done
# 7.) Implement... | {"/RFdriver.py": ["/MIPSobjects.py"], "/System.py": ["/Comms.py"], "/Program.py": ["/Comms.py", "/Command.py"], "/PSG.py": ["/Comms.py", "/MIPSobjects.py"], "/DCBias.py": ["/MIPSobjects.py"], "/MIPS.py": ["/Comms.py", "/Command.py", "/Scripting.py", "/MIPSobjects.py", "/Program.py", "/DCBias.py", "/DIO.py", "/RFdriver.... |
87,938 | GordonAnderson/MIPS-Python-Host-app | refs/heads/main | /JTAG.py | import tkinter as tk
from tkinter import filedialog
from tkinter import messagebox
import os
import string
import sys
import time
import serial
class Uploader:
def __init__(self, parent, cp, debug):
self.p = parent
self.cp = cp
self.debug = debug
# Close MIPS serial port and open fo... | {"/RFdriver.py": ["/MIPSobjects.py"], "/System.py": ["/Comms.py"], "/Program.py": ["/Comms.py", "/Command.py"], "/PSG.py": ["/Comms.py", "/MIPSobjects.py"], "/DCBias.py": ["/MIPSobjects.py"], "/MIPS.py": ["/Comms.py", "/Command.py", "/Scripting.py", "/MIPSobjects.py", "/Program.py", "/DCBias.py", "/DIO.py", "/RFdriver.... |
87,939 | GordonAnderson/MIPS-Python-Host-app | refs/heads/main | /System.py | # system imports
from __future__ import absolute_import
import tkinter as tk
from tkinter import ttk
# this applications imports
import Comms
class System:
def __init__(self, parent, mips):
self.parent = parent
self.mips = mips
self.photo = tk.PhotoImage(file=mips.application_path + '/MIPS... | {"/RFdriver.py": ["/MIPSobjects.py"], "/System.py": ["/Comms.py"], "/Program.py": ["/Comms.py", "/Command.py"], "/PSG.py": ["/Comms.py", "/MIPSobjects.py"], "/DCBias.py": ["/MIPSobjects.py"], "/MIPS.py": ["/Comms.py", "/Command.py", "/Scripting.py", "/MIPSobjects.py", "/Program.py", "/DCBias.py", "/DIO.py", "/RFdriver.... |
87,940 | GordonAnderson/MIPS-Python-Host-app | refs/heads/main | /Scripting.py | import tkinter as tk
from tkinter import filedialog
import code
import threading
import inspect
import ctypes
import time
import queue
import sys
import os
class Console(tk.Frame):
def __init__(self, parent, _locals, exit_callback):
tk.Frame.__init__(self, parent, bg="gray95")
self.parent = parent
... | {"/RFdriver.py": ["/MIPSobjects.py"], "/System.py": ["/Comms.py"], "/Program.py": ["/Comms.py", "/Command.py"], "/PSG.py": ["/Comms.py", "/MIPSobjects.py"], "/DCBias.py": ["/MIPSobjects.py"], "/MIPS.py": ["/Comms.py", "/Command.py", "/Scripting.py", "/MIPSobjects.py", "/Program.py", "/DCBias.py", "/DIO.py", "/RFdriver.... |
87,941 | GordonAnderson/MIPS-Python-Host-app | refs/heads/main | /Program.py | import tkinter as tk
from tkinter import filedialog
from tkinter import messagebox
import os
import time
import Comms
import Command
# This function is called when the program MIPS menu option is called. The following steps are
# performed:
# 1.) Make sure MIPS is connected, if not exit
# 2.) Ask user for the in... | {"/RFdriver.py": ["/MIPSobjects.py"], "/System.py": ["/Comms.py"], "/Program.py": ["/Comms.py", "/Command.py"], "/PSG.py": ["/Comms.py", "/MIPSobjects.py"], "/DCBias.py": ["/MIPSobjects.py"], "/MIPS.py": ["/Comms.py", "/Command.py", "/Scripting.py", "/MIPSobjects.py", "/Program.py", "/DCBias.py", "/DIO.py", "/RFdriver.... |
87,942 | GordonAnderson/MIPS-Python-Host-app | refs/heads/main | /PSG.py | # system imports
from __future__ import absolute_import
import tkinter as tk
from tkinter import ttk
# this applications imports
import Comms
import MIPSobjects
import Util
class psgPoint:
def __init__(self):
self.Name = None
self.TimePoint = None
self.Commands = []
self.Loop = Fal... | {"/RFdriver.py": ["/MIPSobjects.py"], "/System.py": ["/Comms.py"], "/Program.py": ["/Comms.py", "/Command.py"], "/PSG.py": ["/Comms.py", "/MIPSobjects.py"], "/DCBias.py": ["/MIPSobjects.py"], "/MIPS.py": ["/Comms.py", "/Command.py", "/Scripting.py", "/MIPSobjects.py", "/Program.py", "/DCBias.py", "/DIO.py", "/RFdriver.... |
87,943 | GordonAnderson/MIPS-Python-Host-app | refs/heads/main | /DCBias.py | import tkinter as tk
import MIPSobjects
import Util
#
# DCbias channel objects
#
# DCBchannel
# DCBenable
# DCBoffset
#
# todo:
# - Add the linked DCBias channels
class DCBchannel:
def __init__(self, parent, name, MIPSname, channel, comm, units, x, y):
self.master = parent
self.name = name
... | {"/RFdriver.py": ["/MIPSobjects.py"], "/System.py": ["/Comms.py"], "/Program.py": ["/Comms.py", "/Command.py"], "/PSG.py": ["/Comms.py", "/MIPSobjects.py"], "/DCBias.py": ["/MIPSobjects.py"], "/MIPS.py": ["/Comms.py", "/Command.py", "/Scripting.py", "/MIPSobjects.py", "/Program.py", "/DCBias.py", "/DIO.py", "/RFdriver.... |
87,944 | GordonAnderson/MIPS-Python-Host-app | refs/heads/main | /MIPS.py | #
# MIPS host application
#
# This application must be built using Python version 3.7.6 and pyinstaller version 4.0. Later versions
# of these tools will not work due to a code signing set of bugs that are not yet resolved.
#
# Todo:
# 1.) Impliment programming functions
# 2.) Impliment help functions
# 3.) Finis... | {"/RFdriver.py": ["/MIPSobjects.py"], "/System.py": ["/Comms.py"], "/Program.py": ["/Comms.py", "/Command.py"], "/PSG.py": ["/Comms.py", "/MIPSobjects.py"], "/DCBias.py": ["/MIPSobjects.py"], "/MIPS.py": ["/Comms.py", "/Command.py", "/Scripting.py", "/MIPSobjects.py", "/Program.py", "/DCBias.py", "/DIO.py", "/RFdriver.... |
87,945 | GordonAnderson/MIPS-Python-Host-app | refs/heads/main | /DIO.py | import tkinter as tk
from tkinter import ttk
import MIPSobjects
import Util
#
# DIO channel objects
#
class DIOchannel:
def __init__(self, parent, name, MIPSname, channel, comm, x, y):
self.master = parent
self.name = name
self.MIPSname = MIPSname
self.channel = channel
self... | {"/RFdriver.py": ["/MIPSobjects.py"], "/System.py": ["/Comms.py"], "/Program.py": ["/Comms.py", "/Command.py"], "/PSG.py": ["/Comms.py", "/MIPSobjects.py"], "/DCBias.py": ["/MIPSobjects.py"], "/MIPS.py": ["/Comms.py", "/Command.py", "/Scripting.py", "/MIPSobjects.py", "/Program.py", "/DCBias.py", "/DIO.py", "/RFdriver.... |
87,946 | GordonAnderson/MIPS-Python-Host-app | refs/heads/main | /Comms.py | import tkinter as tk
from tkinter import ttk
import time
from time import sleep
from tkinter import simpledialog
import serial
import io
import os
class RingBuffer:
def __init__(self, size):
self.size = size
self.head = 0
self.tail = 0
self.count = 0
self.lines = 0
s... | {"/RFdriver.py": ["/MIPSobjects.py"], "/System.py": ["/Comms.py"], "/Program.py": ["/Comms.py", "/Command.py"], "/PSG.py": ["/Comms.py", "/MIPSobjects.py"], "/DCBias.py": ["/MIPSobjects.py"], "/MIPS.py": ["/Comms.py", "/Command.py", "/Scripting.py", "/MIPSobjects.py", "/Program.py", "/DCBias.py", "/DIO.py", "/RFdriver.... |
87,947 | GordonAnderson/MIPS-Python-Host-app | refs/heads/main | /Command.py | import subprocess
import threading
class Command:
def __init__(self, parent, cmd):
self.parent = parent
self.cmd = cmd
self.dspstr = ""
self.indx = "1.0"
self.running = False
def process(self):
try:
self.parent.insert("end", "starting: " + ' '.join(se... | {"/RFdriver.py": ["/MIPSobjects.py"], "/System.py": ["/Comms.py"], "/Program.py": ["/Comms.py", "/Command.py"], "/PSG.py": ["/Comms.py", "/MIPSobjects.py"], "/DCBias.py": ["/MIPSobjects.py"], "/MIPS.py": ["/Comms.py", "/Command.py", "/Scripting.py", "/MIPSobjects.py", "/Program.py", "/DCBias.py", "/DIO.py", "/RFdriver.... |
87,970 | jupyterhub/binderhub | refs/heads/main | /binderhub/quota.py | """
Singleuser server quotas
"""
import asyncio
import json
import os
from collections import namedtuple
import kubernetes.config
from kubernetes import client
from traitlets import Any, Integer, Unicode, default
from traitlets.config import LoggingConfigurable
from .utils import KUBE_REQUEST_TIMEOUT
class LaunchQ... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,971 | jupyterhub/binderhub | refs/heads/main | /testing/local-binder-k8s-hub/binderhub_config_auth_additions.py | # A development config to test a BinderHub deployment that is relying on
# JupyterHub's as an OAuth2 based Identity Provider (IdP) for Authentication and
# Authorization. JupyterHub is configured with its own Authenticator.
# Deployment assumptions:
# - BinderHub: standalone local installation
# - JupyterHub: standal... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,972 | jupyterhub/binderhub | refs/heads/main | /binderhub/config.py | from .base import BaseHandler
class ConfigHandler(BaseHandler):
"""Serve config"""
def generate_config(self):
config = dict()
for repo_provider_class_alias, repo_provider_class in self.settings[
"repo_providers"
].items():
config[repo_provider_class_alias] = re... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,973 | jupyterhub/binderhub | refs/heads/main | /conftest.py | """top-level pytest config
options can only be defined here,
not in binderhub/tests/conftest.py
"""
def pytest_addoption(parser):
parser.addoption(
"--helm",
action="store_true",
default=False,
help="Run tests marked with pytest.mark.helm",
)
| {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,974 | jupyterhub/binderhub | refs/heads/main | /binderhub/builder.py | """
Handlers for working with version control services (i.e. GitHub) for builds.
"""
import asyncio
import hashlib
import json
import re
import string
import time
from http.client import responses
import docker
import escapism
from prometheus_client import Counter, Gauge, Histogram
from tornado import gen
from tornad... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,975 | jupyterhub/binderhub | refs/heads/main | /testing/local-binder-local-hub/jupyterhub_config.py | """
A development config to test BinderHub locally.
Run `jupyterhub --config=binderhub_config.py` terminal
Host IP is needed in a few places
"""
import os
import socket
from dockerspawner import DockerSpawner
from binderhub.binderspawner_mixin import BinderSpawnerMixin
s = socket.socket(socket.AF_INET, socket.SOCK_... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,976 | jupyterhub/binderhub | refs/heads/main | /binderhub/tests/utils.py | """Testing utilities"""
import asyncio
import io
from concurrent.futures import ThreadPoolExecutor
from urllib.parse import urlparse
import requests
from tornado import gen
from tornado.curl_httpclient import CurlAsyncHTTPClient
from tornado.httpclient import HTTPError, HTTPRequest, HTTPResponse
from tornado.httputil ... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,977 | jupyterhub/binderhub | refs/heads/main | /binderhub/ratelimit.py | """Rate limiting utilities"""
import time
from traitlets import Dict, Float, Integer, default
from traitlets.config import LoggingConfigurable
class RateLimitExceeded(Exception):
"""Exception raised when rate limit is exceeded"""
class RateLimiter(LoggingConfigurable):
"""Class representing a collection o... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,978 | jupyterhub/binderhub | refs/heads/main | /testing/local-binder-k8s-hub/binderhub_config.py | # A development config to test a BinderHub deployment generally. It can be
# combined with with the auth specific config.
# Deployment assumptions:
# - BinderHub: standalone local installation
# - JupyterHub: standalone k8s installation
import os
import subprocess
# We need to find out the IP at which BinderHub can... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,979 | jupyterhub/binderhub | refs/heads/main | /binderhub/tests/test_auth.py | """Test authentication"""
import ipaddress
import json
from unittest import mock
from urllib.parse import urlparse
import pytest
from .conftest import skip_remote
from .utils import async_requests
@pytest.fixture
def use_session():
# setup
async_requests.set_session()
yield "run the test function"
... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,980 | jupyterhub/binderhub | refs/heads/main | /binderhub/tests/conftest.py | """pytest fixtures for binderhub"""
import inspect
import json
import os
import subprocess
import time
from binascii import b2a_hex
from collections import defaultdict
from unittest import mock
from urllib.parse import urlparse
import kubernetes.client
import kubernetes.config
import pytest
import requests
from torna... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,981 | jupyterhub/binderhub | refs/heads/main | /binderhub/tests/test_eventlog.py | import json
import logging
import os
import tempfile
import jsonschema
import pytest
from binderhub.events import EventLog
def test_register_invalid():
"""
Test registering invalid schemas fails
"""
el = EventLog()
with pytest.raises(jsonschema.SchemaError):
el.register_schema(
... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,982 | jupyterhub/binderhub | refs/heads/main | /binderhub/build.py | """
Contains build of a docker image from a git repository.
"""
import datetime
import json
import os
import threading
import warnings
from collections import defaultdict
from enum import Enum
from typing import Union
from urllib.parse import urlparse
import kubernetes.config
from kubernetes import client, watch
from... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,983 | jupyterhub/binderhub | refs/heads/main | /setup.py | import os
import sys
from jupyter_packaging import npm_builder, wrap_installers
from setuptools import find_packages, setup
# ensure the current directory is on sys.path
# so versioneer can be imported when pip uses
# PEP 517/518 build rules.
# https://github.com/python-versioneer/python-versioneer/issues/193
sys.pat... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,984 | jupyterhub/binderhub | refs/heads/main | /binderhub/__init__.py | # next three lines were added by versioneer
from . import _version
__version__ = _version.get_versions()["version"]
| {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,985 | jupyterhub/binderhub | refs/heads/main | /binderhub/registry.py | """
Interaction with the Docker Registry
"""
import base64
import json
import os
import re
from urllib.parse import urlparse
from tornado import httpclient
from tornado.httputil import url_concat
from traitlets import Bool, Dict, Unicode, default
from traitlets.config import LoggingConfigurable
DEFAULT_DOCKER_REGISTR... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,986 | jupyterhub/binderhub | refs/heads/main | /binderhub/tests/test_builder.py | import pytest
from binderhub.builder import _generate_build_name, _get_image_basename_and_tag
@pytest.mark.parametrize(
"fullname,basename,tag",
[
(
"jupyterhub/k8s-binderhub:0.2.0-a2079a5",
"jupyterhub/k8s-binderhub",
"0.2.0-a2079a5",
),
("jupyterh... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,987 | jupyterhub/binderhub | refs/heads/main | /tools/generate-json-schema.py | #!/usr/bin/env python3
"""
This script reads schema.yaml and generates a values.schema.json that we can
package with the Helm chart, allowing helm the CLI to perform validation.
While we can directly generate a values.schema.json from schema.yaml, it
contains a lot of description text we use to generate our configurat... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,988 | jupyterhub/binderhub | refs/heads/main | /helm-chart/binderhub/files/binderhub_config.py | from functools import lru_cache
from urllib.parse import urlparse
from ruamel.yaml import YAML
yaml = YAML()
# memoize so we only load config once
@lru_cache
def _load_values():
"""Load configuration from disk
Memoized to only load once
"""
path = "/etc/binderhub/config/values.yaml"
print(f"Loa... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,989 | jupyterhub/binderhub | refs/heads/main | /testing/local-binder-mocked-hub/binderhub_config.py | # A development config to test BinderHub's UI. The image building or the
# subsequent launching of the built image in a JupyterHub is mocked so that
# users get stuck forever waiting for a build to complete.
# Deployment assumptions:
# - BinderHub: standalone local installation
# - JupyterHub: mocked
from binderhub.... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,990 | jupyterhub/binderhub | refs/heads/main | /binderhub/build_local.py | """
Contains build of a docker image from a git repository.
"""
import json
import os
# These methods are synchronous so don't use tornado.queue
import queue
import subprocess
from threading import Thread
from tornado.log import app_log
from traitlets import default
from .build import BuildExecutor, ProgressEvent
... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,991 | jupyterhub/binderhub | refs/heads/main | /examples/appendix/extra_notebook_config.py | import os
c.NotebookApp.extra_template_paths.append("/etc/jupyter/binder_templates")
c.NotebookApp.jinja_template_vars.update(
{"binder_url": os.environ.get("BINDER_URL", "https://mybinder.org")}
)
| {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,992 | jupyterhub/binderhub | refs/heads/main | /binderhub/tests/test_launcher.py | """Test launcher"""
import pytest
from tornado import web
from binderhub.launcher import Launcher
async def my_pre_launch_hook(launcher, *args):
raise web.HTTPError(
400, "Launch is not possible with parameters: " + ",".join(args)
)
async def test_pre_launch_hook():
launcher = Launcher(create_... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,993 | jupyterhub/binderhub | refs/heads/main | /binderhub/app.py | """
The binderhub application
"""
import asyncio
import ipaddress
import json
import logging
import os
import re
import secrets
import warnings
from binascii import a2b_hex
from concurrent.futures import ThreadPoolExecutor
from glob import glob
from urllib.parse import urlparse
import kubernetes.client
import kubernet... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,994 | jupyterhub/binderhub | refs/heads/main | /binderhub/tests/test_app.py | """Exercise the binderhub entrypoint"""
import sys
from subprocess import check_output
import pytest
from traitlets import TraitError
from binderhub.app import BinderHub
from binderhub.repoproviders import GitHubRepoProvider, GitLabRepoProvider, RepoProvider
def test_help():
check_output([sys.executable, "-m",... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,995 | jupyterhub/binderhub | refs/heads/main | /binderhub/tests/test_utils.py | import ipaddress
from unittest import mock
import pytest
from binderhub import utils
def test_rendezvous_rank():
# check that a key doesn't move if its assigned bucket remains but the
# other buckets are removed
key = "crazy frog is a crazy key"
first_round = utils.rendezvous_rank(["b1", "b2", "b3"]... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,996 | jupyterhub/binderhub | refs/heads/main | /docs/source/conf.py | # Configuration file for Sphinx to build our documentation to HTML.
#
# Configuration reference: https://www.sphinx-doc.org/en/master/usage/configuration.html
#
import datetime
import sys
from os.path import dirname
# -- Project information -----------------------------------------------------
# ref: https://www.sphin... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,997 | jupyterhub/binderhub | refs/heads/main | /binderhub/tests/test_health.py | """Test health handler"""
from .utils import async_requests
async def test_basic_health(app):
r = await async_requests.get(app.url + "/health")
assert r.status_code == 200
results = r.json()
assert results["ok"]
assert "checks" in results
checks = results["checks"]
assert {"service": ... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,998 | jupyterhub/binderhub | refs/heads/main | /examples/binder-api.py | """Launching a binder via API
The binder build API yields a sequence of messages via event-stream.
This example demonstrates how to consume events from the stream
and redirect to the URL when it is ready.
When the image is ready, your browser will open at the desired URL.
"""
import argparse
import json
import sys
i... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
87,999 | jupyterhub/binderhub | refs/heads/main | /binderhub/tests/test_quota.py | """Test launch quotas"""
import concurrent.futures
import json
from unittest import mock
import pytest
from binderhub.quota import KubernetesLaunchQuota, LaunchQuotaExceeded
@pytest.fixture
def mock_pod_list_resp():
r = mock.MagicMock()
r.read.return_value = json.dumps(
{
"items": [
... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
88,000 | jupyterhub/binderhub | refs/heads/main | /binderhub/launcher.py | """
Launch an image with a temporary user via JupyterHub
"""
import asyncio
import base64
import json
import os
import random
import re
import string
import uuid
from datetime import timedelta
from urllib.parse import quote, urlparse
from jupyterhub.traitlets import Callable
from jupyterhub.utils import maybe_future
f... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
88,001 | jupyterhub/binderhub | refs/heads/main | /binderhub/health.py | import asyncio
import json
import time
from functools import wraps
from tornado.httpclient import AsyncHTTPClient
from tornado.log import app_log
from .base import BaseHandler
from .builder import _get_image_basename_and_tag
from .utils import KUBE_REQUEST_TIMEOUT
def retry(_f=None, *, delay=1, attempts=3):
"""... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
88,002 | jupyterhub/binderhub | refs/heads/main | /binderhub/utils.py | """Miscellaneous utilities"""
import ipaddress
import time
from collections import OrderedDict
from hashlib import blake2b
from unittest.mock import Mock
from kubernetes.client import api_client
from traitlets import Integer, TraitError
# default _request_timeout for kubernetes api requests
# tuple of two timeouts: (... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
88,003 | jupyterhub/binderhub | refs/heads/main | /binderhub/repoproviders.py | """
Classes for Repo providers.
Subclass the base class, ``RepoProvider``, to support different version
control services and providers.
.. note:: When adding a new repo provider, add it to the allowed values for
repo providers in event-schemas/launch.json.
"""
import asyncio
import json
import os
import re
... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
88,004 | jupyterhub/binderhub | refs/heads/main | /binderhub/tests/test_build.py | """Test building repos"""
import json
import sys
from time import monotonic
from unittest import mock
from urllib.parse import quote
from uuid import uuid4
import docker
import pytest
from kubernetes import client
from tornado.httputil import url_concat
from tornado.queues import Queue
from binderhub.build import Ku... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
88,005 | jupyterhub/binderhub | refs/heads/main | /binderhub/main.py | """
Main handler classes for requests
"""
import time
import urllib.parse
import jwt
from tornado.httpclient import AsyncHTTPClient, HTTPRequest
from tornado.httputil import url_concat
from tornado.log import app_log
from tornado.web import HTTPError, authenticated
from .base import BaseHandler
SPEC_NAMES = {
"g... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
88,006 | jupyterhub/binderhub | refs/heads/main | /binderhub/metrics.py | from prometheus_client import CONTENT_TYPE_LATEST, REGISTRY, generate_latest
from .base import BaseHandler
class MetricsHandler(BaseHandler):
# demote logging of 200 responses to debug-level
log_success_debug = True
async def get(self):
self.set_header("Content-Type", CONTENT_TYPE_LATEST)
... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
88,007 | jupyterhub/binderhub | refs/heads/main | /binderhub/events.py | """
Emit structured, discrete events when various actions happen.
"""
import json
import logging
from datetime import datetime
import jsonschema
from jupyterhub.traitlets import Callable
from pythonjsonlogger import jsonlogger
from traitlets.config import Configurable
def _skip_message(record, **kwargs):
"""
... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
88,008 | jupyterhub/binderhub | refs/heads/main | /testing/local-binder-local-hub/binderhub_config.py | """
A development config to test BinderHub locally.
If you are running BinderHub manually (not via JupyterHub) run
`python -m binderhub -f binderhub_config.py`
Override the external access URL for JupyterHub by setting the
environment variable JUPYTERHUB_EXTERNAL_URL
Host IP is needed in a few places
"""
import os
im... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
88,009 | jupyterhub/binderhub | refs/heads/main | /examples/appendix/binderhub_config.py | c.BinderHub.appendix = """
USER root
ENV BINDER_URL={binder_url}
ENV REPO_URL={repo_url}
RUN cd /tmp \
&& wget -q https://github.com/jupyterhub/binderhub/archive/main.tar.gz -O binderhub.tar.gz \
&& tar --wildcards -xzf binderhub.tar.gz --strip 2 */examples/appendix\
&& ./appendix/run-appendix \
&& rm -rf binderhub... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
88,010 | jupyterhub/binderhub | refs/heads/main | /ci/check_embedded_chart_code.py | #!/usr/bin/env python
# FIXME: We currently have some code duplicated in
# binderhub/binderspawner_mixin.py and helm-chart/binderhub/values.yaml
# and we use a pre-commit hook to automatically update the values in
# values.yaml.
#
# We should remove the embedded code from values.yaml and in... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
88,011 | jupyterhub/binderhub | refs/heads/main | /binderhub/tests/test_ratelimit.py | from unittest import mock
import pytest
from binderhub.ratelimit import RateLimiter, RateLimitExceeded
def test_rate_limit():
r = RateLimiter(limit=10, period_seconds=60)
assert r._limits == {}
now = r.time()
reset = int(now) + 60
with mock.patch.object(r, "time", lambda: now):
limit = r... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
88,012 | jupyterhub/binderhub | refs/heads/main | /binderhub/binderspawner_mixin.py | """
Helpers for creating BinderSpawners
FIXME:
This file is defined in binderhub/binderspawner_mixin.py
and is copied to helm-chart/binderhub/values.yaml
by ci/check_embedded_chart_code.py
The BinderHub repo is just used as the distribution mechanism for this spawner,
BinderHub itself doesn't require this code.
Long... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
88,013 | jupyterhub/binderhub | refs/heads/main | /binderhub/tests/test_registry.py | """Tests for the registry"""
import base64
import json
import os
from random import randint
import pytest
from tornado import httpclient
from tornado.web import Application, HTTPError, RequestHandler
from binderhub.registry import DockerRegistry
def test_registry_defaults(tmpdir):
registry = DockerRegistry(dock... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
88,014 | jupyterhub/binderhub | refs/heads/main | /binderhub/base.py | """Base classes for request handlers"""
import json
import urllib.parse
from http.client import responses
import jwt
from jupyterhub.services.auth import HubOAuth, HubOAuthenticated
from tornado import web
from tornado.log import app_log
from . import __version__ as binder_version
from .ratelimit import RateLimitExc... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
88,015 | jupyterhub/binderhub | refs/heads/main | /binderhub/tests/test_repoproviders.py | import re
from unittest import TestCase
from urllib.parse import quote
import pytest
from tornado.ioloop import IOLoop
from binderhub.repoproviders import (
DataverseProvider,
FigshareProvider,
GistRepoProvider,
GitHubRepoProvider,
GitLabRepoProvider,
GitRepoProvider,
HydroshareProvider,
... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
88,016 | jupyterhub/binderhub | refs/heads/main | /binderhub/tests/test_main.py | """Test main handlers"""
import time
from urllib.parse import quote, urlparse
import jwt
import pytest
from bs4 import BeautifulSoup
from binderhub import __version__ as binder_version
from .utils import async_requests
@pytest.mark.parametrize(
"old_url, new_url",
[
(
"/repo/binderhub-... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
88,017 | jupyterhub/binderhub | refs/heads/main | /tools/validate-against-schema.py | #!/usr/bin/env python3
import os
import jsonschema
from ruamel.yaml import YAML
yaml = YAML()
here_dir = os.path.abspath(os.path.dirname(__file__))
schema_yaml = os.path.join(here_dir, os.pardir, "helm-chart/binderhub", "schema.yaml")
values_yaml = os.path.join(here_dir, os.pardir, "helm-chart/binderhub", "values.ya... | {"/binderhub/quota.py": ["/binderhub/utils.py"], "/binderhub/config.py": ["/binderhub/base.py"], "/binderhub/builder.py": ["/binderhub/base.py", "/binderhub/build.py", "/binderhub/quota.py"], "/testing/local-binder-local-hub/jupyterhub_config.py": ["/binderhub/binderspawner_mixin.py"], "/binderhub/tests/test_auth.py": ... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.