code stringlengths 2k 1.04M | repo_path stringlengths 5 517 | parsed_code stringlengths 0 1.04M | quality_prob float64 0.02 0.95 | learning_prob float64 0.02 0.93 |
|---|---|---|---|---|
import logging
import boto3
from botocore.stub import Stubber
from django.test import TestCase
from django.contrib.auth import get_user_model
from django.core.cache import caches
from rest_framework.response import Response
from rest_framework.test import APIRequestFactory
from rest_framework.views import APIView
fro... | tests/test_views.py | import logging
import boto3
from botocore.stub import Stubber
from django.test import TestCase
from django.contrib.auth import get_user_model
from django.core.cache import caches
from rest_framework.response import Response
from rest_framework.test import APIRequestFactory
from rest_framework.views import APIView
fro... | 0.437343 | 0.224029 |
from pprint import pformat
from six import iteritems
class V1Volume(object):
"""
NOTE: This class is auto generated by the swagger code generator program.
Do not edit the class manually.
"""
def __init__(self):
"""
Swagger model
:param dict swaggerTypes: The key is attribu... | magnum/common/pythonk8sclient/swagger_client/models/v1_volume.py | from pprint import pformat
from six import iteritems
class V1Volume(object):
"""
NOTE: This class is auto generated by the swagger code generator program.
Do not edit the class manually.
"""
def __init__(self):
"""
Swagger model
:param dict swaggerTypes: The key is attribu... | 0.750827 | 0.149035 |
from struct import pack, unpack
from time import time
from FS.Inode import Inode
class Root(object):
def __init__(self, superblock, fat, file):
self._superblock = superblock
self._fat = fat
self._file = file
self._init_files_list()
def add(self, file_name, inode):
sup... | FS/Root.py | from struct import pack, unpack
from time import time
from FS.Inode import Inode
class Root(object):
def __init__(self, superblock, fat, file):
self._superblock = superblock
self._fat = fat
self._file = file
self._init_files_list()
def add(self, file_name, inode):
sup... | 0.442877 | 0.215846 |
import tornado.gen
import tornado.httpserver
import tornado.web
from wotpy.codecs.enums import MediaTypes
from wotpy.protocols.enums import Protocols, InteractionVerbs
from wotpy.protocols.http.enums import HTTPSchemes
from wotpy.protocols.http.handlers.action import ActionInvokeHandler, PendingInvocationHandler
from ... | wotpy/protocols/http/server.py | import tornado.gen
import tornado.httpserver
import tornado.web
from wotpy.codecs.enums import MediaTypes
from wotpy.protocols.enums import Protocols, InteractionVerbs
from wotpy.protocols.http.enums import HTTPSchemes
from wotpy.protocols.http.handlers.action import ActionInvokeHandler, PendingInvocationHandler
from ... | 0.790126 | 0.113211 |
import numpy as np
import matplotlib.pyplot as plt
from acconeer_utils.clients import SocketClient, SPIClient, UARTClient
from acconeer_utils.clients import configs
from acconeer_utils import example_utils
def main():
args = example_utils.ExampleArgumentParser(num_sens=1).parse_args()
example_utils.config_lo... | examples/plotting/plot_with_matplotlib.py | import numpy as np
import matplotlib.pyplot as plt
from acconeer_utils.clients import SocketClient, SPIClient, UARTClient
from acconeer_utils.clients import configs
from acconeer_utils import example_utils
def main():
args = example_utils.ExampleArgumentParser(num_sens=1).parse_args()
example_utils.config_lo... | 0.525369 | 0.341335 |
import pygame
class JCSPyGm_Input:
"""Input hanlde."""
mouseDownThisFrame = False
mouseIsDown = False
mouseUpThisFrame = False
mousePosition = (0, 0)
keysPressedThisFrame = []
keysDown = []
keysReleasedThisFrame = []
@staticmethod
def clean_input_buffer():
... | jcspygm/util/JCSPyGm_Input.py |
import pygame
class JCSPyGm_Input:
"""Input hanlde."""
mouseDownThisFrame = False
mouseIsDown = False
mouseUpThisFrame = False
mousePosition = (0, 0)
keysPressedThisFrame = []
keysDown = []
keysReleasedThisFrame = []
@staticmethod
def clean_input_buffer():
... | 0.5144 | 0.300848 |
import numpy as np
import torch
import alf.utils.distributions as ad
import alf
class DistributionTest(alf.test.TestCase):
def _test_its(self, x, its: ad.InverseTransformSampling):
x.requires_grad_()
y = its.cdf(x)
x1 = its.icdf(y)
p = its.log_prob(x).exp()
step = x[1] - x... | alf/utils/distributions_test.py |
import numpy as np
import torch
import alf.utils.distributions as ad
import alf
class DistributionTest(alf.test.TestCase):
def _test_its(self, x, its: ad.InverseTransformSampling):
x.requires_grad_()
y = its.cdf(x)
x1 = its.icdf(y)
p = its.log_prob(x).exp()
step = x[1] - x... | 0.722821 | 0.761671 |
from .base_random_cell import BaseRandomCellTransform
import numpy as np
import pandas as pd
class ShuffleNoise(BaseRandomCellTransform):
"""A batch transformation for adding noise to data by randomly shuffling columns. The noise is added by mixing
incoming batch with its shuffled version using mask:
``... | keras_batchflow/base/batch_transformers/shuffle_noise.py | from .base_random_cell import BaseRandomCellTransform
import numpy as np
import pandas as pd
class ShuffleNoise(BaseRandomCellTransform):
"""A batch transformation for adding noise to data by randomly shuffling columns. The noise is added by mixing
incoming batch with its shuffled version using mask:
``... | 0.885792 | 0.902952 |
from pathlib import Path
from typing import Callable, Iterable, Dict
import re
from io import BytesIO
from collections import defaultdict
import nlzss11
from .bzs import ParsedBzs, parseBzs, buildBzs
from .msb import ParsedMsb, parseMSB, buildMSB
from .u8file import U8File
STAGE_REGEX = re.compile('(.+)_stg_l([0-9]+)... | sslib/patch.py | from pathlib import Path
from typing import Callable, Iterable, Dict
import re
from io import BytesIO
from collections import defaultdict
import nlzss11
from .bzs import ParsedBzs, parseBzs, buildBzs
from .msb import ParsedMsb, parseMSB, buildMSB
from .u8file import U8File
STAGE_REGEX = re.compile('(.+)_stg_l([0-9]+)... | 0.467575 | 0.194521 |
# standard and third party libraries
import cv2
import numpy as np
import sys
import time
import copy
import multiprocessing
import logging
import Image
# milovision libraries
from pipeline_modules import ContourFinder
from pipeline_modules import EllipseFitter
from pipeline_modules import PoseEstimatorA
from simulato... | pipeline.py | # standard and third party libraries
import cv2
import numpy as np
import sys
import time
import copy
import multiprocessing
import logging
import Image
# milovision libraries
from pipeline_modules import ContourFinder
from pipeline_modules import EllipseFitter
from pipeline_modules import PoseEstimatorA
from simulato... | 0.275812 | 0.160957 |
from sklearn.feature_extraction.text import TfidfVectorizer, CountVectorizer
from sklearn.decomposition import NMF, LatentDirichletAllocation
class LDA:
"""
Parameters
----------
n_topics : int
Number of topics
n_iter : int, default 2000
Number of sampling iterati... | src/models/LDA.py | from sklearn.feature_extraction.text import TfidfVectorizer, CountVectorizer
from sklearn.decomposition import NMF, LatentDirichletAllocation
class LDA:
"""
Parameters
----------
n_topics : int
Number of topics
n_iter : int, default 2000
Number of sampling iterati... | 0.853073 | 0.402921 |
import json
import warnings
import pulumi
import pulumi.runtime
from .. import utilities, tables
class AccessLevel(pulumi.CustomResource):
basic: pulumi.Output[dict]
description: pulumi.Output[str]
name: pulumi.Output[str]
parent: pulumi.Output[str]
title: pulumi.Output[str]
def __init__(__sel... | sdk/python/pulumi_gcp/accesscontextmanager/access_level.py |
import json
import warnings
import pulumi
import pulumi.runtime
from .. import utilities, tables
class AccessLevel(pulumi.CustomResource):
basic: pulumi.Output[dict]
description: pulumi.Output[str]
name: pulumi.Output[str]
parent: pulumi.Output[str]
title: pulumi.Output[str]
def __init__(__sel... | 0.505371 | 0.080177 |
import torch
import torch.nn as nn
import torch.nn.functional as F
from torch.autograd import Variable
import torch.optim as optim
import models
import util
if torch.cuda.is_available():
device = torch.device('cuda')
else:
device = torch.device('cpu')
def squared_l2_norm(x):
flattened = x.view(x.unsqueeze... | trades.py | import torch
import torch.nn as nn
import torch.nn.functional as F
from torch.autograd import Variable
import torch.optim as optim
import models
import util
if torch.cuda.is_available():
device = torch.device('cuda')
else:
device = torch.device('cpu')
def squared_l2_norm(x):
flattened = x.view(x.unsqueeze... | 0.927629 | 0.515864 |
import os
from abc import abstractmethod
from collections import OrderedDict
import matplotlib.pyplot as plt
import numpy as np
import pandas as pd
import torch
from hydra import compose, initialize
from ishtos_datasets import get_dataset
from ishtos_models import get_model
from ishtos_transforms import get_transforms... | {{cookiecutter.package_name}}/src/exp_000/ishtos_runner.py | import os
from abc import abstractmethod
from collections import OrderedDict
import matplotlib.pyplot as plt
import numpy as np
import pandas as pd
import torch
from hydra import compose, initialize
from ishtos_datasets import get_dataset
from ishtos_models import get_model
from ishtos_transforms import get_transforms... | 0.706596 | 0.214691 |
"""
TUPLAS
"""
# Una tupla es una colección de variables de cualquier tipo
tupla = (1, "dos", True)
#%%
# Aunque lo normal es que todos los objetos dentro de la colección sean del
# mismo tipo, como en este caso, que son variables de tipo texto
tupla = ('uno','dos','tres','cuatro','cinco','seis','siete','ocho','nueve... | 03_estructuras_de_datos.py | """
TUPLAS
"""
# Una tupla es una colección de variables de cualquier tipo
tupla = (1, "dos", True)
#%%
# Aunque lo normal es que todos los objetos dentro de la colección sean del
# mismo tipo, como en este caso, que son variables de tipo texto
tupla = ('uno','dos','tres','cuatro','cinco','seis','siete','ocho','nueve... | 0.196094 | 0.356587 |
from getratings.models.ratings import Ratings
class NA_Ziggs_Jng_Aatrox(Ratings):
pass
class NA_Ziggs_Jng_Ahri(Ratings):
pass
class NA_Ziggs_Jng_Akali(Ratings):
pass
class NA_Ziggs_Jng_Alistar(Ratings):
pass
class NA_Ziggs_Jng_Amumu(Ratings):
pass
class NA_Ziggs_Jng_Anivia(Ratings):
pass
... | loldib/getratings/models/NA/na_ziggs/na_ziggs_jng.py | from getratings.models.ratings import Ratings
class NA_Ziggs_Jng_Aatrox(Ratings):
pass
class NA_Ziggs_Jng_Ahri(Ratings):
pass
class NA_Ziggs_Jng_Akali(Ratings):
pass
class NA_Ziggs_Jng_Alistar(Ratings):
pass
class NA_Ziggs_Jng_Amumu(Ratings):
pass
class NA_Ziggs_Jng_Anivia(Ratings):
pass
... | 0.204938 | 0.097907 |
from torch.utils.data import Subset
from PIL import Image
from torchvision.datasets import CIFAR100
from base.torchvision_dataset import TorchvisionDataset
import numpy as np
import torch
import torchvision.transforms as transforms
import random
class CIFAR100_Dataset(TorchvisionDataset):
def __init__(self, roo... | src/datasets/cifar100.py | from torch.utils.data import Subset
from PIL import Image
from torchvision.datasets import CIFAR100
from base.torchvision_dataset import TorchvisionDataset
import numpy as np
import torch
import torchvision.transforms as transforms
import random
class CIFAR100_Dataset(TorchvisionDataset):
def __init__(self, roo... | 0.90692 | 0.594434 |
"""Server-side sessions."""
from datetime import datetime, timedelta
from flask import request
from flask.sessions import SessionInterface, SessionMixin
from .model import get_session_data, store_session_data, delete_session
from .util import random_string
class RedisSession(dict, SessionMixin):
"""The session o... | frost/session.py | """Server-side sessions."""
from datetime import datetime, timedelta
from flask import request
from flask.sessions import SessionInterface, SessionMixin
from .model import get_session_data, store_session_data, delete_session
from .util import random_string
class RedisSession(dict, SessionMixin):
"""The session o... | 0.839142 | 0.168754 |
import os
from unittest import mock
from unittest.mock import MagicMock
import pytest
from pytorch_lightning import Trainer
from pytorch_lightning.loggers import _MLFLOW_AVAILABLE, MLFlowLogger
from tests.helpers import BoringModel
def mock_mlflow_run_creation(logger, experiment_name=None, experiment_id=None, run_i... | tests/loggers/test_mlflow.py | import os
from unittest import mock
from unittest.mock import MagicMock
import pytest
from pytorch_lightning import Trainer
from pytorch_lightning.loggers import _MLFLOW_AVAILABLE, MLFlowLogger
from tests.helpers import BoringModel
def mock_mlflow_run_creation(logger, experiment_name=None, experiment_id=None, run_i... | 0.584153 | 0.452173 |
import numpy as np
import sys
from sklearn.metrics.pairwise import euclidean_distances
from scipy.optimize import linear_sum_assignment
def gen_gaus_mixture(centers, mixing_prop=None, noise_sd=0.1, M=5000):
# noise_sd=0.1
# M=5000
# mixing_prop=None
# centers = np.random.normal(0,10,(10,500))
K, D... | simulation/data_sampling.py | import numpy as np
import sys
from sklearn.metrics.pairwise import euclidean_distances
from scipy.optimize import linear_sum_assignment
def gen_gaus_mixture(centers, mixing_prop=None, noise_sd=0.1, M=5000):
# noise_sd=0.1
# M=5000
# mixing_prop=None
# centers = np.random.normal(0,10,(10,500))
K, D... | 0.276202 | 0.332026 |
import os
import sys
import warnings
import openstackdocstheme
sys.path.insert(0, os.path.abspath('../..'))
sys.path.insert(0, os.path.abspath('.'))
# -- General configuration ----------------------------------------------------
# Add any Sphinx extension module names here, as strings. They can be
# extensions com... | doc/source/conf.py |
import os
import sys
import warnings
import openstackdocstheme
sys.path.insert(0, os.path.abspath('../..'))
sys.path.insert(0, os.path.abspath('.'))
# -- General configuration ----------------------------------------------------
# Add any Sphinx extension module names here, as strings. They can be
# extensions com... | 0.242385 | 0.077762 |
import vtk
ROOT_SPLIT = 10
TARGET_LEVEL = 8
TARGET_LEVEL = 6
CUT_OFF = TARGET_LEVEL
# -----------------------------------------------------------------------------
# Helpers
# -----------------------------------------------------------------------------
def mandelbrotTest(x, y, timeStep=0):
count = 0
cReal ... | Common/DataModel/Testing/Python/TestHyperTreeGrid3DCursorsMandel.py | import vtk
ROOT_SPLIT = 10
TARGET_LEVEL = 8
TARGET_LEVEL = 6
CUT_OFF = TARGET_LEVEL
# -----------------------------------------------------------------------------
# Helpers
# -----------------------------------------------------------------------------
def mandelbrotTest(x, y, timeStep=0):
count = 0
cReal ... | 0.428712 | 0.395368 |
import cmath
import numpy as np
import seaborn as sns
import matplotlib.pyplot as plt
import scipy.signal as signal
from ..utils.Autocorrelation import Autocorrelation
sns.set()
class AutocorrelationMethod():
def __init__(self):
self.f = None
self.default_f = np.linspace(0, 0.5, 500)
self... | python/source/parametric/AutocorrelationMethod.py | import cmath
import numpy as np
import seaborn as sns
import matplotlib.pyplot as plt
import scipy.signal as signal
from ..utils.Autocorrelation import Autocorrelation
sns.set()
class AutocorrelationMethod():
def __init__(self):
self.f = None
self.default_f = np.linspace(0, 0.5, 500)
self... | 0.623835 | 0.358044 |
import uuid
import base64
import json
from flask import jsonify, make_response
from flask_restful import abort
import ast as type_evaluation
def is_valid_uuid(val):
try:
uuid.UUID(str(val), version=4)
return True
except ValueError:
return False
def get_or_create(session, model, keys,... | ros/lib/utils.py | import uuid
import base64
import json
from flask import jsonify, make_response
from flask_restful import abort
import ast as type_evaluation
def is_valid_uuid(val):
try:
uuid.UUID(str(val), version=4)
return True
except ValueError:
return False
def get_or_create(session, model, keys,... | 0.531696 | 0.177882 |
class AliKeyDecryptor(object):
def decrypt(self, key):
result = None
se = ord("a")
if(len(key) == 20):
r = key[0]
i = chr(r).lower()
a = int(i, base=36) % 7
# n = key[a]
# o = chr(n)
# s = key[a+1]
... | keydecryptor/ali.py | class AliKeyDecryptor(object):
def decrypt(self, key):
result = None
se = ord("a")
if(len(key) == 20):
r = key[0]
i = chr(r).lower()
a = int(i, base=36) % 7
# n = key[a]
# o = chr(n)
# s = key[a+1]
... | 0.112527 | 0.149967 |
import Tkinter
import tkFont
def mid(l1, l2):
return (int((l1[0]+l2[0]) / 2), int((l1[1]+l2[1]) / 2))
class SimulatorUI:
def __init__(self, settings):
self.settings = settings
self.square_size = 40
self.border_width = 1
self.padding = 0
self.arrow_width = 3
s... | rgsimulatorUI.py | import Tkinter
import tkFont
def mid(l1, l2):
return (int((l1[0]+l2[0]) / 2), int((l1[1]+l2[1]) / 2))
class SimulatorUI:
def __init__(self, settings):
self.settings = settings
self.square_size = 40
self.border_width = 1
self.padding = 0
self.arrow_width = 3
s... | 0.486332 | 0.296311 |
"""Resource for Transaction endpoints."""
from http import HTTPStatus
from flask import Response, current_app, jsonify, request
from flask_restplus import Namespace, Resource, cors
from pay_api.exceptions import BusinessException
from pay_api.schemas import utils as schema_utils
from pay_api.services import ReceiptS... | pay-api/src/pay_api/resources/invoice_receipt.py | """Resource for Transaction endpoints."""
from http import HTTPStatus
from flask import Response, current_app, jsonify, request
from flask_restplus import Namespace, Resource, cors
from pay_api.exceptions import BusinessException
from pay_api.schemas import utils as schema_utils
from pay_api.services import ReceiptS... | 0.68941 | 0.08819 |
if __name__ == "__main__":
from psana.pscalib.geometry.SegGeometryMatrixV1 import *
import sys
from time import time
import psana.pyalgos.generic.Graphics as gg
logging.basicConfig(format='[%(levelname).1s] L%(lineno)04d: %(message)s', level=logging.DEBUG)
def test_xyz_min_max():
w = segment_one
... | psana/psana/pscalib/geometry/test_SegGeometryMatrixV1.py |
if __name__ == "__main__":
from psana.pscalib.geometry.SegGeometryMatrixV1 import *
import sys
from time import time
import psana.pyalgos.generic.Graphics as gg
logging.basicConfig(format='[%(levelname).1s] L%(lineno)04d: %(message)s', level=logging.DEBUG)
def test_xyz_min_max():
w = segment_one
... | 0.152916 | 0.239527 |
import pytest
from ceph_deploy.cli import get_parser
from ceph_deploy.tests.util import assert_too_few_arguments
class TestParserPkg(object):
def setup(self):
self.parser = get_parser()
def test_pkg_help(self, capsys):
with pytest.raises(SystemExit):
self.parser.parse_args('pkg ... | ceph_deploy/tests/parser/test_pkg.py | import pytest
from ceph_deploy.cli import get_parser
from ceph_deploy.tests.util import assert_too_few_arguments
class TestParserPkg(object):
def setup(self):
self.parser = get_parser()
def test_pkg_help(self, capsys):
with pytest.raises(SystemExit):
self.parser.parse_args('pkg ... | 0.355104 | 0.429728 |
class GenomeBuilder:
def __init__(self, time_scale, time_per_task, min_task_time):
""" A class used to generate a Genome object."""
self.time_scale = time_scale
self.total_minutes = self.get_total_minutes()
if self.validate_minimum_task_time(min_task_time):
self.min_ta... | evoschedule/app/genome_builder.py | class GenomeBuilder:
def __init__(self, time_scale, time_per_task, min_task_time):
""" A class used to generate a Genome object."""
self.time_scale = time_scale
self.total_minutes = self.get_total_minutes()
if self.validate_minimum_task_time(min_task_time):
self.min_ta... | 0.921118 | 0.379551 |
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import oneflow as flow
def add_ofrecord_args(parser):
parser.add_argument("--image_size", type=int, default=224,
required=False, help="image size")
parser.add_argument("--resiz... | model_compress/ChannelSlimming/util/ofrecord_util.py | from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import oneflow as flow
def add_ofrecord_args(parser):
parser.add_argument("--image_size", type=int, default=224,
required=False, help="image size")
parser.add_argument("--resiz... | 0.609292 | 0.134804 |
from django import template
from django.core.urlresolvers import reverse
from django.template import Token, Parser, TOKEN_BLOCK
import fudge
import random
from .._utils import TestCase
from ... import context_processors
from ...templatetags.esi import EsiNode
from ...templatetags.esi import esi
def create_context():... | armstrong/esi/tests/templatetags/esi.py | from django import template
from django.core.urlresolvers import reverse
from django.template import Token, Parser, TOKEN_BLOCK
import fudge
import random
from .._utils import TestCase
from ... import context_processors
from ...templatetags.esi import EsiNode
from ...templatetags.esi import esi
def create_context():... | 0.545044 | 0.259222 |
from vue import *
def test_app_with_props_and_data(selenium):
def app_with_props_data(el):
class App(VueComponent):
text: str
template = """
<div id="el">{{ text }}</div>
"""
return App(el, props_data={"text": "TEXT"})
with selenium.app(app_wit... | tests/selenium/test_api.py | from vue import *
def test_app_with_props_and_data(selenium):
def app_with_props_data(el):
class App(VueComponent):
text: str
template = """
<div id="el">{{ text }}</div>
"""
return App(el, props_data={"text": "TEXT"})
with selenium.app(app_wit... | 0.538498 | 0.201518 |
from __future__ import unicode_literals
import datetime
import uuid
from dagster import ExecutionTargetHandle
from dagster.utils import script_relative_path
# pylint: disable=unused-import
from dagster_airflow.test_fixtures import (
dagster_airflow_docker_operator_pipeline,
dagster_airflow_python_operator_pi... | python_modules/dagster-airflow/dagster_airflow_tests/test_factory.py | from __future__ import unicode_literals
import datetime
import uuid
from dagster import ExecutionTargetHandle
from dagster.utils import script_relative_path
# pylint: disable=unused-import
from dagster_airflow.test_fixtures import (
dagster_airflow_docker_operator_pipeline,
dagster_airflow_python_operator_pi... | 0.430746 | 0.233553 |
import asyncio
import traceback
from nextcord.ext import commands
import nextcord
import os
import sys
from nextcord import Interaction, SlashOption, ChannelType
from cogs.debug import save
sys.path.append('../')
from util import admin_check, n_fc, eh
# join message system
# Copilotでちょっとだけ楽をした。
#loggingの設定
import lo... | cogs/welcome.py | n!welcome join %name%さんこんちゃ!!!!!
n!welcome leave %name%さんばいばい!!!
n!welcome join off
n!welcome leave off | 0.119216 | 0.363675 |
import os, sys
import shutil
import time
import logging
import collections
import json, yaml
from easydict import EasyDict
import pprint
from .dirs import create_dirs
from . import logging_utils
from . import config_utils
from . import torch_utils
from . import shutil_utils
from . import modelarts_utils
from . import ... | submodule/template_lib/utils/config.py | import os, sys
import shutil
import time
import logging
import collections
import json, yaml
from easydict import EasyDict
import pprint
from .dirs import create_dirs
from . import logging_utils
from . import config_utils
from . import torch_utils
from . import shutil_utils
from . import modelarts_utils
from . import ... | 0.145631 | 0.053108 |
from django.contrib.auth import get_user_model
from .testCases import RelayTestCase, DefaultTestCase
from graphql_auth.constants import Messages
from graphql_auth.utils import get_token, get_token_paylod
class PasswordResetTestCaseMixin:
def setUp(self):
self.user1 = self.register_user(
email... | tests/test_password_reset.py | from django.contrib.auth import get_user_model
from .testCases import RelayTestCase, DefaultTestCase
from graphql_auth.constants import Messages
from graphql_auth.utils import get_token, get_token_paylod
class PasswordResetTestCaseMixin:
def setUp(self):
self.user1 = self.register_user(
email... | 0.499268 | 0.29305 |
from datetime import datetime
from requests.auth import HTTPBasicAuth
from libqtile.widget import base
from libqtile import bar, images
from libqtile.log_utils import logger
from libqtile.popup import Popup
from .tvhlib import TVHJobServer, TVHTimeOut, ICON_PATH, ICONS
class TVHWidget(base._Widget, base.MarginMixin... | tvhwidget/tvhwidget.py | from datetime import datetime
from requests.auth import HTTPBasicAuth
from libqtile.widget import base
from libqtile import bar, images
from libqtile.log_utils import logger
from libqtile.popup import Popup
from .tvhlib import TVHJobServer, TVHTimeOut, ICON_PATH, ICONS
class TVHWidget(base._Widget, base.MarginMixin... | 0.635562 | 0.097005 |
import datetime
import logging
from pathlib import Path
from centralpy.client import CentralClient
from centralpy.responses import CsvZip
logger = logging.getLogger(__name__)
# pylint: disable=too-many-arguments
def pull_csv_zip(
client: CentralClient,
project: str,
form_id: str,
csv_dir: Path,
... | centralpy/use_cases/pull_csv_zip.py | import datetime
import logging
from pathlib import Path
from centralpy.client import CentralClient
from centralpy.responses import CsvZip
logger = logging.getLogger(__name__)
# pylint: disable=too-many-arguments
def pull_csv_zip(
client: CentralClient,
project: str,
form_id: str,
csv_dir: Path,
... | 0.435421 | 0.116764 |
from app import db
from app import login
from sqlalchemy.sql import func
from werkzeug.security import generate_password_hash, check_password_hash
from flask_login import UserMixin
class User(UserMixin, db.Model):
id = db.Column(db.Integer, primary_key=True)
username = db.Column(db.String(64), index=True, un... | app/models.py | from app import db
from app import login
from sqlalchemy.sql import func
from werkzeug.security import generate_password_hash, check_password_hash
from flask_login import UserMixin
class User(UserMixin, db.Model):
id = db.Column(db.Integer, primary_key=True)
username = db.Column(db.String(64), index=True, un... | 0.519278 | 0.059319 |
import requests as http
from aqt.addons import ConfigEditor
from aqt import mw
from aqt.qt import *
from aqt.utils import showWarning
from .config import get_credentials
from .const import LANGUAGE_CODES
# If user is authenticated return authorization token otherwise return None
def authenticate():
# Get username ... | sync_lingq/lingq.py | import requests as http
from aqt.addons import ConfigEditor
from aqt import mw
from aqt.qt import *
from aqt.utils import showWarning
from .config import get_credentials
from .const import LANGUAGE_CODES
# If user is authenticated return authorization token otherwise return None
def authenticate():
# Get username ... | 0.507812 | 0.091544 |
import argparse
import logging
import yaml
import datetime
import time
import urllib
import os
from kubernetes import client as k8s_client
from kubernetes.client import rest
from benchmark.test import deploy_utils
from kubeflow.testing import util
def parse_args():
parser = argparse.ArgumentParser()
parser.add_a... | src/benchmark/test/install_addon.py | import argparse
import logging
import yaml
import datetime
import time
import urllib
import os
from kubernetes import client as k8s_client
from kubernetes.client import rest
from benchmark.test import deploy_utils
from kubeflow.testing import util
def parse_args():
parser = argparse.ArgumentParser()
parser.add_a... | 0.568176 | 0.10325 |
import sys, os, time, traceback
from ixnetwork_restpy import SessionAssistant
apiServerIp = '172.16.101.3'
ixChassisIpList = ['172.16.102.5']
portList = [[ixChassisIpList[0], 1,1], [ixChassisIpList[0], 1, 2]]
# For Linux API server only
username = 'admin'
password = '<PASSWORD>'
# For linux and connection_manager o... | RestPy/SampleScripts/evpnNgpf2.py | import sys, os, time, traceback
from ixnetwork_restpy import SessionAssistant
apiServerIp = '172.16.101.3'
ixChassisIpList = ['172.16.102.5']
portList = [[ixChassisIpList[0], 1,1], [ixChassisIpList[0], 1, 2]]
# For Linux API server only
username = 'admin'
password = '<PASSWORD>'
# For linux and connection_manager o... | 0.183082 | 0.108378 |
from bs4 import BeautifulSoup
import json
import datetime
from pf2helpers import Pf2Helpers
import os
import re
data_holder = {}
data_holder['name'] = 'Pathfinder 2.0 ItemList v2'
data_holder['date'] = datetime.date.today().strftime("%B %d, %Y")
class ItemBuilder():
item_keywords = ['name','source', 'rarity', 'c... | buildItems.py | from bs4 import BeautifulSoup
import json
import datetime
from pf2helpers import Pf2Helpers
import os
import re
data_holder = {}
data_holder['name'] = 'Pathfinder 2.0 ItemList v2'
data_holder['date'] = datetime.date.today().strftime("%B %d, %Y")
class ItemBuilder():
item_keywords = ['name','source', 'rarity', 'c... | 0.082822 | 0.143938 |
__all__ = [
'tower_list',
'tower_receive',
'tower_send',
'log',
]
import json
import textwrap
import re
import click
import tower_cli
from tower_cli.cli.transfer.send import Sender
from tower_cli.cli.transfer.receive import Receiver
from tower_cli.cli.transfer.common import SEND_ORDER as ASSET_TYPES
i... | atacac/_utils.py | __all__ = [
'tower_list',
'tower_receive',
'tower_send',
'log',
]
import json
import textwrap
import re
import click
import tower_cli
from tower_cli.cli.transfer.send import Sender
from tower_cli.cli.transfer.receive import Receiver
from tower_cli.cli.transfer.common import SEND_ORDER as ASSET_TYPES
i... | 0.569374 | 0.12166 |
from tensorflow.keras.models import model_from_json, load_model
import os
from picamera.array import PiRGBArray
from picamera import PiCamera
import time
import cv2 as cv
from utils import *
import numpy as np
from tkinter import *
from tkinter.ttk import *
import PIL.Image
import PIL.ImageTk
REPODIR = "/home/pi/one-m... | prototype-scripts/src/test_network_ui.py | from tensorflow.keras.models import model_from_json, load_model
import os
from picamera.array import PiRGBArray
from picamera import PiCamera
import time
import cv2 as cv
from utils import *
import numpy as np
from tkinter import *
from tkinter.ttk import *
import PIL.Image
import PIL.ImageTk
REPODIR = "/home/pi/one-m... | 0.41324 | 0.228619 |
import pika
import sys
import time
import os
import multiprocessing
import ctypes
def create_test_msg(msg_size):
class TEST(ctypes.Structure):
_fields_ = [('data', ctypes.c_byte * msg_size)]
return TEST
def publisher_loop(pub_id=0, num_msgs=10000, msg_size=512, num_subscribers=1, ready_flag=None, ser... | rmq_bench_synchronous.py | import pika
import sys
import time
import os
import multiprocessing
import ctypes
def create_test_msg(msg_size):
class TEST(ctypes.Structure):
_fields_ = [('data', ctypes.c_byte * msg_size)]
return TEST
def publisher_loop(pub_id=0, num_msgs=10000, msg_size=512, num_subscribers=1, ready_flag=None, ser... | 0.327023 | 0.10942 |
from django import forms
from django.utils.translation import gettext as _
from dcim.models import DeviceRole, Platform, Region, Site, SiteGroup
from extras.forms import CustomFieldModelFilterForm, LocalConfigContextFilterForm
from tenancy.forms import TenancyFilterForm, ContactModelFilterForm
from utilities.forms imp... | netbox/virtualization/forms/filtersets.py | from django import forms
from django.utils.translation import gettext as _
from dcim.models import DeviceRole, Platform, Region, Site, SiteGroup
from extras.forms import CustomFieldModelFilterForm, LocalConfigContextFilterForm
from tenancy.forms import TenancyFilterForm, ContactModelFilterForm
from utilities.forms imp... | 0.538741 | 0.066721 |
import multiprocessing
import threading
import Queue
from uuid import uuid4
import numpy as np
import SharedArray
import data
def load_shared(args):
i, array_name, fname, kwargs = args
array = SharedArray.attach(array_name)
array[i] = data.load_augment(fname, **kwargs)
class BatchIterator(object):
... | iterator.py | import multiprocessing
import threading
import Queue
from uuid import uuid4
import numpy as np
import SharedArray
import data
def load_shared(args):
i, array_name, fname, kwargs = args
array = SharedArray.attach(array_name)
array[i] = data.load_augment(fname, **kwargs)
class BatchIterator(object):
... | 0.633977 | 0.223939 |
from typing import Tuple, List
import numpy as np
from torch.utils.data import Dataset
from pathlib import Path
from . import ModelCallback
from ...models.base import DetectionModel
from ...utils import image_utils, draw_bounding_boxes_with_name_tag
class GenerateDetectionImageCallback(ModelCallback):
def __init... | deepext/trainer/callbacks/object_detection.py | from typing import Tuple, List
import numpy as np
from torch.utils.data import Dataset
from pathlib import Path
from . import ModelCallback
from ...models.base import DetectionModel
from ...utils import image_utils, draw_bounding_boxes_with_name_tag
class GenerateDetectionImageCallback(ModelCallback):
def __init... | 0.875348 | 0.422564 |
import asyncio
import pytest
from dask_gateway_server.workqueue import WorkQueue, Backoff, WorkQueueClosed
from dask_gateway_server.utils import cancel_task
def test_backoff():
backoff = Backoff(base_delay=0.5, max_delay=5)
assert backoff.failures("foo") == 0
assert backoff.backoff("foo") == 0.5
a... | tests/test_workqueue.py | import asyncio
import pytest
from dask_gateway_server.workqueue import WorkQueue, Backoff, WorkQueueClosed
from dask_gateway_server.utils import cancel_task
def test_backoff():
backoff = Backoff(base_delay=0.5, max_delay=5)
assert backoff.failures("foo") == 0
assert backoff.backoff("foo") == 0.5
a... | 0.64579 | 0.576959 |
import glfw
from OpenGL.GL import *
from OpenGL.GL.shaders import compileProgram, compileShader
import numpy as np
vertex_src = """
# version 330
layout(location = 0) in vec3 a_position;
layout(location = 1) in vec3 a_color;
out vec3 v_color;
void main()
{
gl_Position = vec4(a_position, 1.0);
v_color = a_col... | py_src/Test(ing)/Chris_test.py | import glfw
from OpenGL.GL import *
from OpenGL.GL.shaders import compileProgram, compileShader
import numpy as np
vertex_src = """
# version 330
layout(location = 0) in vec3 a_position;
layout(location = 1) in vec3 a_color;
out vec3 v_color;
void main()
{
gl_Position = vec4(a_position, 1.0);
v_color = a_col... | 0.405213 | 0.271481 |
# Standard library modules.
import collections
import importlib
import os
# External dependencies.
from humanfriendly import concatenate, format, parse_path, pluralize
from pkg_resources import iter_entry_points
from property_manager import lazy_property, mutable_property
from sqlalchemy import func
from update_dotdee... | chat_archive/__init__.py | # Standard library modules.
import collections
import importlib
import os
# External dependencies.
from humanfriendly import concatenate, format, parse_path, pluralize
from pkg_resources import iter_entry_points
from property_manager import lazy_property, mutable_property
from sqlalchemy import func
from update_dotdee... | 0.792143 | 0.185504 |
import sys
_b=sys.version_info[0]<3 and (lambda x:x) or (lambda x:x.encode('latin1'))
from google.protobuf import descriptor as _descriptor
from google.protobuf import message as _message
from google.protobuf import reflection as _reflection
from google.protobuf import symbol_database as _symbol_database
from google.p... | third_party/gae_ts_mon/gae_ts_mon/protos/acquisition_task_pb2.py |
import sys
_b=sys.version_info[0]<3 and (lambda x:x) or (lambda x:x.encode('latin1'))
from google.protobuf import descriptor as _descriptor
from google.protobuf import message as _message
from google.protobuf import reflection as _reflection
from google.protobuf import symbol_database as _symbol_database
from google.p... | 0.202601 | 0.112527 |
from grr.lib import flags
from grr.lib import rdfvalue
from grr.lib.rdfvalues import flows as rdf_flows
from grr.server.grr_response_server import data_store
from grr.server.grr_response_server.hunts import results as hunts_results
from grr.test_lib import aff4_test_lib
from grr.test_lib import test_lib
class ResultT... | grr/server/grr_response_server/hunts/results_test.py | from grr.lib import flags
from grr.lib import rdfvalue
from grr.lib.rdfvalues import flows as rdf_flows
from grr.server.grr_response_server import data_store
from grr.server.grr_response_server.hunts import results as hunts_results
from grr.test_lib import aff4_test_lib
from grr.test_lib import test_lib
class ResultT... | 0.563858 | 0.33516 |
import os
import random
from imutils import paths
import numpy as np
import pandas as pd
import skimage as sk
import skimage.transform
from sklearn.model_selection import train_test_split
from tensorflow.keras.preprocessing.image import img_to_array, load_img
from tensorflow.keras.utils import to_categorical
import c... | src/data_operations/data_preprocessing.py | import os
import random
from imutils import paths
import numpy as np
import pandas as pd
import skimage as sk
import skimage.transform
from sklearn.model_selection import train_test_split
from tensorflow.keras.preprocessing.image import img_to_array, load_img
from tensorflow.keras.utils import to_categorical
import c... | 0.788909 | 0.663914 |
from __future__ import annotations
import logging
from datetime import datetime, timedelta
from time import gmtime, strftime
from typing import Any, Dict, Optional, Sequence, Tuple, Union
from wyze_sdk.errors import WyzeRequestError
from wyze_sdk.models import datetime_to_epoch
from wyze_sdk.models.devices import Dev... | wyze_sdk/service/api_service.py | from __future__ import annotations
import logging
from datetime import datetime, timedelta
from time import gmtime, strftime
from typing import Any, Dict, Optional, Sequence, Tuple, Union
from wyze_sdk.errors import WyzeRequestError
from wyze_sdk.models import datetime_to_epoch
from wyze_sdk.models.devices import Dev... | 0.845911 | 0.096663 |
from test_framework.mininode import *
from test_framework.test_framework import XazabTestFramework
from test_framework.util import *
'''
feature_llmq_signing.py
Checks LLMQs signing sessions
'''
class LLMQSigningTest(XazabTestFramework):
def set_test_params(self):
self.set_xazab_test_params(6, 5, fast_... | test/functional/feature_llmq_signing.py |
from test_framework.mininode import *
from test_framework.test_framework import XazabTestFramework
from test_framework.util import *
'''
feature_llmq_signing.py
Checks LLMQs signing sessions
'''
class LLMQSigningTest(XazabTestFramework):
def set_test_params(self):
self.set_xazab_test_params(6, 5, fast_... | 0.556159 | 0.422326 |
import numpy as np
from scipy.special import iv
def tapering_window(time,D,mywindow):
""" tapering_window returns the window for tapering a WOSA segment.
Inputs:
- time [1-dim numpy array of floats]: times along the WOSA segment.
- D [float]: Temporal length of the WOSA segment.
- mywindow [int]: Choice of ... | wavepal/tapering_window.py | import numpy as np
from scipy.special import iv
def tapering_window(time,D,mywindow):
""" tapering_window returns the window for tapering a WOSA segment.
Inputs:
- time [1-dim numpy array of floats]: times along the WOSA segment.
- D [float]: Temporal length of the WOSA segment.
- mywindow [int]: Choice of ... | 0.300746 | 0.595434 |
import pywikibot
from datetime import datetime
import re
site = pywikibot.Site('wikidata', 'wikidata')
# remove requests from Wikidata:Requests_for_permissions/Bot
def removeRequests(requests):
page = pywikibot.Page(site, 'Wikidata:Requests for permissions/Bot')
text = page.get()
for request in requests... | requestsForBotflagArchive.py |
import pywikibot
from datetime import datetime
import re
site = pywikibot.Site('wikidata', 'wikidata')
# remove requests from Wikidata:Requests_for_permissions/Bot
def removeRequests(requests):
page = pywikibot.Page(site, 'Wikidata:Requests for permissions/Bot')
text = page.get()
for request in requests... | 0.215433 | 0.079353 |
from __future__ import absolute_import, unicode_literals
import json
import tornado.web
import logging
import ldap
from .dtos import TrackDTO, DTOEncoder
from .base_request_handler import BaseRequestHandler
from .database_connection import DBConnection
logger = logging.getLogger(__package__)
class LoginRequestHan... | mopidy_bamp/mopidy_bamp/login_request_handler.py | from __future__ import absolute_import, unicode_literals
import json
import tornado.web
import logging
import ldap
from .dtos import TrackDTO, DTOEncoder
from .base_request_handler import BaseRequestHandler
from .database_connection import DBConnection
logger = logging.getLogger(__package__)
class LoginRequestHan... | 0.371821 | 0.056288 |
import matplotlib.pyplot as plt
import numpy as np
from ..utils.validation import validate_float, validate_int
class PoissonProcess(object):
"""Simulate a sample path of a Poisson process.
A Poisson process (with rate parameter λ) is continuous time stochastic
process (N(t) : t ≥ 0) such that
1) N(... | stattools/simulation/poisson.py |
import matplotlib.pyplot as plt
import numpy as np
from ..utils.validation import validate_float, validate_int
class PoissonProcess(object):
"""Simulate a sample path of a Poisson process.
A Poisson process (with rate parameter λ) is continuous time stochastic
process (N(t) : t ≥ 0) such that
1) N(... | 0.929216 | 0.82925 |
from __future__ import unicode_literals
RAW_URL = 'https://raw.githubusercontent.com/Kozea/tinycss/single-regex-tokenizer/tinycss/tokenizer2.py'
import re
import sys
import functools
import operator
import regex
from . import token_data
MACROS = {}
def macro(name, value):
MACROS[name] = '(?:%s)' % value.format(... | tinycss/tokenizer2.py | from __future__ import unicode_literals
RAW_URL = 'https://raw.githubusercontent.com/Kozea/tinycss/single-regex-tokenizer/tinycss/tokenizer2.py'
import re
import sys
import functools
import operator
import regex
from . import token_data
MACROS = {}
def macro(name, value):
MACROS[name] = '(?:%s)' % value.format(... | 0.39129 | 0.135318 |
from mcfunction.versions.mc_1_13.team import team, ParsedTeamCommand
from mcfunction.nodes import EntityNode
def test_team_add():
parsed = team.parse('team add testteam')
parsed: ParsedTeamCommand
assert parsed.action.value == 'add'
assert parsed.team.value == 'testteam'
assert str(parsed) == 't... | tests/commands/mc-1.13/test_team.py | from mcfunction.versions.mc_1_13.team import team, ParsedTeamCommand
from mcfunction.nodes import EntityNode
def test_team_add():
parsed = team.parse('team add testteam')
parsed: ParsedTeamCommand
assert parsed.action.value == 'add'
assert parsed.team.value == 'testteam'
assert str(parsed) == 't... | 0.724773 | 0.722576 |
import discord
from discord.ext import commands
from discord.ext import pages
from discord import Embed
import pycordSuperUtils
client = commands.Bot(command_prefix="-")
embedc = discord.Color.from_rgb(255, 255, 255)
@client.event
async def on_ready():
print("Page manager is ready.", client.user)
# The... | examples/paginator.py | import discord
from discord.ext import commands
from discord.ext import pages
from discord import Embed
import pycordSuperUtils
client = commands.Bot(command_prefix="-")
embedc = discord.Color.from_rgb(255, 255, 255)
@client.event
async def on_ready():
print("Page manager is ready.", client.user)
# The... | 0.413122 | 0.275739 |
import atexit
import signal
import sys
import os
import time
# Infoset imports
from infoset.utils import log
from infoset.utils import general
class Daemon(object):
"""A generic daemon class.
Usage: subclass the daemon class and override the run() method.
Modified from http://www.jejik.com/files/examp... | infoset/utils/daemon.py |
import atexit
import signal
import sys
import os
import time
# Infoset imports
from infoset.utils import log
from infoset.utils import general
class Daemon(object):
"""A generic daemon class.
Usage: subclass the daemon class and override the run() method.
Modified from http://www.jejik.com/files/examp... | 0.494873 | 0.122786 |
import imp
from generator.tree.builder import build_ast
from generator.tree.nodes.trivial.namespace import Namespace
from .generators.cpp import CppGenerator
from .generators.dot import DotGenerator
from .generators.go import GoGenerator
from .generators.python import PythonGenerator
from .generators.rust import Rust... | flatdata-py/generator/engine.py | import imp
from generator.tree.builder import build_ast
from generator.tree.nodes.trivial.namespace import Namespace
from .generators.cpp import CppGenerator
from .generators.dot import DotGenerator
from .generators.go import GoGenerator
from .generators.python import PythonGenerator
from .generators.rust import Rust... | 0.73678 | 0.297582 |
import collections
import datetime
import json
import multiprocessing.pool
import random
import time
import traceback
import types
from django import forms
from django.db import models
import django.core.mail
import django.urls
import django.forms.renderers
from cached_property import cached_property
from constance i... | src/web/modules/smartq/models.py | import collections
import datetime
import json
import multiprocessing.pool
import random
import time
import traceback
import types
from django import forms
from django.db import models
import django.core.mail
import django.urls
import django.forms.renderers
from cached_property import cached_property
from constance i... | 0.26923 | 0.106319 |
import json
import jwt
from fastapi import Depends, HTTPException, Path, Query
from fastapi.security import HTTPAuthorizationCredentials, HTTPBearer
from fastapi.security.utils import get_authorization_scheme_param
from pydantic import BaseModel
from starlette.requests import Request
from starlette.status import HTTP_... | fastapi_admin/depends.py | import json
import jwt
from fastapi import Depends, HTTPException, Path, Query
from fastapi.security import HTTPAuthorizationCredentials, HTTPBearer
from fastapi.security.utils import get_authorization_scheme_param
from pydantic import BaseModel
from starlette.requests import Request
from starlette.status import HTTP_... | 0.412057 | 0.066387 |
import numpy as np
import paddle
import paddle.nn as nn
from ..registry import BACKBONES
import sys
def get_hop_distance(num_node, edge, max_hop=1):
A = np.zeros((num_node, num_node))
for i, j in edge:
A[j, i] = 1
A[i, j] = 1
# compute hop steps
hop_dis = np.zeros((num_node, num_node)... | paddlevideo/modeling/backbones/ctrgcn2.py | import numpy as np
import paddle
import paddle.nn as nn
from ..registry import BACKBONES
import sys
def get_hop_distance(num_node, edge, max_hop=1):
A = np.zeros((num_node, num_node))
for i, j in edge:
A[j, i] = 1
A[i, j] = 1
# compute hop steps
hop_dis = np.zeros((num_node, num_node)... | 0.52683 | 0.348493 |
from __future__ import unicode_literals
from __future__ import print_function
from ...command import SubCommand
from ...wsgi import WSGIApplication
from ...console import Cell
from ...template.moyatemplates import Template
from datetime import datetime
from fs.path import join
import os.path
from collections import ... | moya/command/sub/extract.py | from __future__ import unicode_literals
from __future__ import print_function
from ...command import SubCommand
from ...wsgi import WSGIApplication
from ...console import Cell
from ...template.moyatemplates import Template
from datetime import datetime
from fs.path import join
import os.path
from collections import ... | 0.359701 | 0.080719 |
import unittest
import schrift
class SchriftTest(unittest.TestCase):
def setUp(self):
schrift.app.config["SQLALCHEMY_DATABASE_URI"] = "sqlite://"
self.app = schrift.app.test_client()
schrift.db.create_all()
self.setUpUsers()
def setUpUsers(self):
"""
Helper fu... | schrift_tests.py |
import unittest
import schrift
class SchriftTest(unittest.TestCase):
def setUp(self):
schrift.app.config["SQLALCHEMY_DATABASE_URI"] = "sqlite://"
self.app = schrift.app.test_client()
schrift.db.create_all()
self.setUpUsers()
def setUpUsers(self):
"""
Helper fu... | 0.446012 | 0.244242 |
from dataclasses import dataclass, field
from typing import Optional
from transformers import AutoModelForTokenClassification, AutoConfig
from transformers import HfArgumentParser
import transformers
import torch
import logging
import os
from collections import OrderedDict
from tqdm import tqdm
os.environ["WANDB_DISABL... | weaver.py | from dataclasses import dataclass, field
from typing import Optional
from transformers import AutoModelForTokenClassification, AutoConfig
from transformers import HfArgumentParser
import transformers
import torch
import logging
import os
from collections import OrderedDict
from tqdm import tqdm
os.environ["WANDB_DISABL... | 0.891782 | 0.376451 |
import numpy as np
from . import core
def mass(Q, T, M_T, Σ_T, trivial_idx=None, excl_zone=0, left=False, right=False):
"""
Compute "Mueen's Algorithm for Similarity Search" (MASS)
Parameters
----------
Q : ndarray
Query array or subsequence
T : ndarray
Time series array or s... | stumpy/stamp.py |
import numpy as np
from . import core
def mass(Q, T, M_T, Σ_T, trivial_idx=None, excl_zone=0, left=False, right=False):
"""
Compute "Mueen's Algorithm for Similarity Search" (MASS)
Parameters
----------
Q : ndarray
Query array or subsequence
T : ndarray
Time series array or s... | 0.941654 | 0.757862 |
from datetime import datetime
from kubernetes import client, config
import logging
import os
import subprocess
import time
ANNOTATION_PREFIX = 'time-sync.riasc.eu'
NODE_NAME = os.environ.get('NODE_NAME', 'infis-pi')
def get_status():
sources = {}
fields = {
'sources': sources
}
ret = subproc... | images/time-sync/chrony-monitor.py | from datetime import datetime
from kubernetes import client, config
import logging
import os
import subprocess
import time
ANNOTATION_PREFIX = 'time-sync.riasc.eu'
NODE_NAME = os.environ.get('NODE_NAME', 'infis-pi')
def get_status():
sources = {}
fields = {
'sources': sources
}
ret = subproc... | 0.405096 | 0.205217 |
from cosivina.base import *
elementSpec = []
class Element(object):
''' Base class for elements. '''
def __init__(self, label):
self.label = label
self.parameters = {}
self.components = {}
self.defaultOutputComponent = ''
self.inputElementLabels = []
self.inputC... | cosivina/Element.py | from cosivina.base import *
elementSpec = []
class Element(object):
''' Base class for elements. '''
def __init__(self, label):
self.label = label
self.parameters = {}
self.components = {}
self.defaultOutputComponent = ''
self.inputElementLabels = []
self.inputC... | 0.628065 | 0.193967 |
import pandas as pd
from sklearn import model_selection
from sklearn.tree import DecisionTreeClassifier
def predict(home_team, away_team, city, toss_winner, toss_decision):
matches_cleaned_data = pd.read_csv('./Dataset/matches_cleaned.csv')
matches_df = matches_cleaned_data[['team1', 'team2', 'city', '... | prediction.py | import pandas as pd
from sklearn import model_selection
from sklearn.tree import DecisionTreeClassifier
def predict(home_team, away_team, city, toss_winner, toss_decision):
matches_cleaned_data = pd.read_csv('./Dataset/matches_cleaned.csv')
matches_df = matches_cleaned_data[['team1', 'team2', 'city', '... | 0.303629 | 0.401453 |
import numpy as np
from model.content_types import ContentType
from model.group_chat import GroupChat
from model.message import Message
from model.reaction import Reaction
class JsonParser:
"""
Parses the facebook data files to a GroupChat object
"""
def json_to_group_chat(self, data_files):
... | json_parser.py | import numpy as np
from model.content_types import ContentType
from model.group_chat import GroupChat
from model.message import Message
from model.reaction import Reaction
class JsonParser:
"""
Parses the facebook data files to a GroupChat object
"""
def json_to_group_chat(self, data_files):
... | 0.444565 | 0.218065 |
import numpy as np
from scipy import linalg
from pressio4py import logger, solvers, ode, rom
np.set_printoptions(linewidth=140)
class MyTestApp:
def __init__(self, N):
self.N_ = N
self.callCount_ = 0
def createDiscreteTimeResidual(self):
return np.zeros(self.N_)
def createApplyDiscreteTimeJacobia... | tests_functional_small/test_rom_lspg_unsteady_discretetime_default_trivialapp_two_states.py | import numpy as np
from scipy import linalg
from pressio4py import logger, solvers, ode, rom
np.set_printoptions(linewidth=140)
class MyTestApp:
def __init__(self, N):
self.N_ = N
self.callCount_ = 0
def createDiscreteTimeResidual(self):
return np.zeros(self.N_)
def createApplyDiscreteTimeJacobia... | 0.591015 | 0.625095 |
import os
import platform
import shutil
import socket
import tempfile
import threading
import unittest
import ipaddr
import portpicker
import logging
from grr.lib import data_store
from grr.lib import data_store_test
from grr.lib import flags
from grr.lib import test_lib
from grr.lib import utils
from grr.lib.data... | grr/lib/data_stores/http_data_store_test.py | import os
import platform
import shutil
import socket
import tempfile
import threading
import unittest
import ipaddr
import portpicker
import logging
from grr.lib import data_store
from grr.lib import data_store_test
from grr.lib import flags
from grr.lib import test_lib
from grr.lib import utils
from grr.lib.data... | 0.423339 | 0.109753 |
from PyQt5.QtWidgets import QLabel
import constants
def get_label(text):
temp_label = QLabel(text)
if 'LEC' in text:
temp_label.setStyleSheet("background-color: lightskyblue; font-family: Arial; font-size: 22px")
elif 'TUT' in text:
temp_label.setStyleSheet("background-color: springgreen; ... | src/plannerMgr.py | from PyQt5.QtWidgets import QLabel
import constants
def get_label(text):
temp_label = QLabel(text)
if 'LEC' in text:
temp_label.setStyleSheet("background-color: lightskyblue; font-family: Arial; font-size: 22px")
elif 'TUT' in text:
temp_label.setStyleSheet("background-color: springgreen; ... | 0.347869 | 0.221898 |
from flask import Flask
import flask
import os
import base64
import markdown
from markupsafe import Markup
import json
app = Flask(__name__)
app.config.from_pyfile('settings.py')
blogs = []
with open('data/data.json', 'r') as f:
placeholder = (json.load(f))
for item in placeholder:
blogs.append({'titl... | A3.py | from flask import Flask
import flask
import os
import base64
import markdown
from markupsafe import Markup
import json
app = Flask(__name__)
app.config.from_pyfile('settings.py')
blogs = []
with open('data/data.json', 'r') as f:
placeholder = (json.load(f))
for item in placeholder:
blogs.append({'titl... | 0.228931 | 0.077692 |
import FWCore.ParameterSet.Config as cms
from DQMServices.Core.DQMEDHarvester import DQMEDHarvester
process = cms.Process("emdqm")
process.load('Configuration/StandardSequences/FrontierConditions_GlobalTag_cff')
process.GlobalTag.globaltag = 'START72_V1::All'
process.load("FWCore.MessageService.MessageLogger_cfi")
#... | HLTriggerOffline/Egamma/test/testEmDQM_cfg.py | import FWCore.ParameterSet.Config as cms
from DQMServices.Core.DQMEDHarvester import DQMEDHarvester
process = cms.Process("emdqm")
process.load('Configuration/StandardSequences/FrontierConditions_GlobalTag_cff')
process.GlobalTag.globaltag = 'START72_V1::All'
process.load("FWCore.MessageService.MessageLogger_cfi")
#... | 0.276105 | 0.10004 |
from standard_wallet.wallet import Wallet
import clvm
from chiasim.hashable import CoinSolution, Program, ProgramHash, SpendBundle
from clvm_tools import binutils
from chiasim.validation.Conditions import ConditionOpcode
from utilities.puzzle_utilities import puzzlehash_from_string
from utilities.keys import signature_... | atomic_swaps/as_wallet.py | from standard_wallet.wallet import Wallet
import clvm
from chiasim.hashable import CoinSolution, Program, ProgramHash, SpendBundle
from clvm_tools import binutils
from chiasim.validation.Conditions import ConditionOpcode
from utilities.puzzle_utilities import puzzlehash_from_string
from utilities.keys import signature_... | 0.562898 | 0.250947 |
import webapp2
import jinja2
from webapp2_extras import auth, sessions
# jinja environment
jinja_environment = jinja2.Environment(loader=jinja2.FileSystemLoader('static/templates'))
def user_required(handler):
""" decorator that checks if the user exist """
def check_login(self, *args, **kwargs):
aut... | prjGogisticsWINEVER/src/handlers/handler_webapp2_extra_auth.py | import webapp2
import jinja2
from webapp2_extras import auth, sessions
# jinja environment
jinja_environment = jinja2.Environment(loader=jinja2.FileSystemLoader('static/templates'))
def user_required(handler):
""" decorator that checks if the user exist """
def check_login(self, *args, **kwargs):
aut... | 0.33372 | 0.057812 |
# Importing the libraries
import numpy as np
import matplotlib.pyplot as plt
import pandas as pd
import pickle
# Importing the dataset
dataset = pd.read_csv('/home/mostafa/workarea_POSH/RF_FrontEnd/datasets/SH_Dataset_modified2.csv',header=None)
dataset=dataset.dropna()
X = np.array(dataset.iloc[1:, 0:5].values,dty... | Block/RF_FrontEnd/TSMC28_RF_FrontEnd_1.3.2020/make_reg/SHRegMaker.py |
# Importing the libraries
import numpy as np
import matplotlib.pyplot as plt
import pandas as pd
import pickle
# Importing the dataset
dataset = pd.read_csv('/home/mostafa/workarea_POSH/RF_FrontEnd/datasets/SH_Dataset_modified2.csv',header=None)
dataset=dataset.dropna()
X = np.array(dataset.iloc[1:, 0:5].values,dty... | 0.617282 | 0.498169 |
# Copyright (C) 2021 <NAME> <<EMAIL>>
# SPDX-License-Identifier: MIT
"""Day Twenty-One, Allergen Assessment."""
from sys import argv
from re import findall
from collections import defaultdict
from functools import reduce
from utils import open_file, arrange, usage_and_exit, freqd2
def gen_allergen(recipes):
... | src/day21.py |
# Copyright (C) 2021 <NAME> <<EMAIL>>
# SPDX-License-Identifier: MIT
"""Day Twenty-One, Allergen Assessment."""
from sys import argv
from re import findall
from collections import defaultdict
from functools import reduce
from utils import open_file, arrange, usage_and_exit, freqd2
def gen_allergen(recipes):
... | 0.696475 | 0.301658 |
_base_ = ['../../_base_/default_runtime.py']
dict_file = 'data/chineseocr/labels/dict_printed_chinese_english_digits.txt'
label_convertor = dict(
type='AttnConvertor', dict_file=dict_file, with_unknown=True)
model = dict(
type='SARNet',
backbone=dict(type='ResNet31OCR'),
encoder=dict(
type='SA... | configs/textrecog/sar/sar_r31_parallel_decoder_chinese.py | _base_ = ['../../_base_/default_runtime.py']
dict_file = 'data/chineseocr/labels/dict_printed_chinese_english_digits.txt'
label_convertor = dict(
type='AttnConvertor', dict_file=dict_file, with_unknown=True)
model = dict(
type='SARNet',
backbone=dict(type='ResNet31OCR'),
encoder=dict(
type='SA... | 0.612194 | 0.170802 |
import os
import math
import warnings
import time
import numpy as np
import random as rn
import tensorflow as tf
from tensorBNN.activationFunctions import Tanh
from tensorBNN.layer import DenseLayer
from tensorBNN.network import network
from tensorBNN.likelihood import GaussianLikelihood
from tensorBNN.metrics import... | Examples/trainRegression.py | import os
import math
import warnings
import time
import numpy as np
import random as rn
import tensorflow as tf
from tensorBNN.activationFunctions import Tanh
from tensorBNN.layer import DenseLayer
from tensorBNN.network import network
from tensorBNN.likelihood import GaussianLikelihood
from tensorBNN.metrics import... | 0.507324 | 0.349061 |
import math
from collections import defaultdict
import timeit
from wikiIndexer import removeStopWords, stem, tokenise
wlist = []
tlist = []
title_dic = defaultdict(int)
opt_dict = defaultdict(int)
global no_of_docs
def get_docNum():
global no_of_docs
with open("tmp/doc_count.txt", "r") as fp:
for line... | search.py | import math
from collections import defaultdict
import timeit
from wikiIndexer import removeStopWords, stem, tokenise
wlist = []
tlist = []
title_dic = defaultdict(int)
opt_dict = defaultdict(int)
global no_of_docs
def get_docNum():
global no_of_docs
with open("tmp/doc_count.txt", "r") as fp:
for line... | 0.111036 | 0.191914 |
from keras.models import Sequential, Model
from keras.layers import Dense, LSTM, Bidirectional, GRU, Embedding, Dropout, Lambda
from keras.layers import Input, concatenate
NUM_CLASSES = 3 # Predict [space, comma, period]
def blstm(hidden_units):
m = Sequential()
m.add(Bidirectional(LSTM(hidden_units, return... | rupunktor/model_zoo.py | from keras.models import Sequential, Model
from keras.layers import Dense, LSTM, Bidirectional, GRU, Embedding, Dropout, Lambda
from keras.layers import Input, concatenate
NUM_CLASSES = 3 # Predict [space, comma, period]
def blstm(hidden_units):
m = Sequential()
m.add(Bidirectional(LSTM(hidden_units, return... | 0.910869 | 0.427098 |
PATH_NONE = "NONE"
PATH_UNION = "UNION"
PATH_INTER = "INTERSECT"
PATH_EXCEPT = "EXCEPT"
PATH_WHERE = "WHERE"
PATH_HAVING = "HAVING"
PATH_PAR = "PARALLEL" # To represent the multiple selection clauses in a single WHERE clause.
VEC_AGGREGATORS = [ 'none', 'max', 'min', 'count', 'sum', 'avg' ]
VEC_OPERATORS ... | util/db_meta.py | PATH_NONE = "NONE"
PATH_UNION = "UNION"
PATH_INTER = "INTERSECT"
PATH_EXCEPT = "EXCEPT"
PATH_WHERE = "WHERE"
PATH_HAVING = "HAVING"
PATH_PAR = "PARALLEL" # To represent the multiple selection clauses in a single WHERE clause.
VEC_AGGREGATORS = [ 'none', 'max', 'min', 'count', 'sum', 'avg' ]
VEC_OPERATORS ... | 0.375248 | 0.10466 |
import html
from telegram import ParseMode, Update
from telegram.error import BadRequest
from telegram.ext import CallbackContext, CommandHandler, Filters, run_async
from telegram.utils.helpers import mention_html
from UltronRoBo import (
DEV_USERS,
LOGGER,
OWNER_ID,
DRAGONS,
DEMONS,
TIGERS,
... | UltronRoBo/modules/bans.py | import html
from telegram import ParseMode, Update
from telegram.error import BadRequest
from telegram.ext import CallbackContext, CommandHandler, Filters, run_async
from telegram.utils.helpers import mention_html
from UltronRoBo import (
DEV_USERS,
LOGGER,
OWNER_ID,
DRAGONS,
DEMONS,
TIGERS,
... | 0.184841 | 0.06216 |
import sys
import unittest
import timeit
from numpy.linalg import det
import numpy as np
sys.path.append("../src")
sys.path.append("src/")
from tools import EqualMatrices, AssertAlmostEqualMatrices
from gdft import dft_matrix, random_unitary_matrix, g_matrix, gdft_matrix, two_param_gdft_matrix, permutation_matrix
dft... | tests/testGDFT.py | import sys
import unittest
import timeit
from numpy.linalg import det
import numpy as np
sys.path.append("../src")
sys.path.append("src/")
from tools import EqualMatrices, AssertAlmostEqualMatrices
from gdft import dft_matrix, random_unitary_matrix, g_matrix, gdft_matrix, two_param_gdft_matrix, permutation_matrix
dft... | 0.347426 | 0.651937 |
import importlib
from types import LambdaType, ModuleType
from typing import Any
from typing import Callable as CallableType
from typing import Dict, Tuple, Union
from simple_di import (
VT,
Provider,
_inject_args,
_inject_kwargs,
_SentinelClass,
inject,
sentinel,
)
__all__ = [
"Static... | simple_di/providers.py | import importlib
from types import LambdaType, ModuleType
from typing import Any
from typing import Callable as CallableType
from typing import Dict, Tuple, Union
from simple_di import (
VT,
Provider,
_inject_args,
_inject_kwargs,
_SentinelClass,
inject,
sentinel,
)
__all__ = [
"Static... | 0.813794 | 0.208803 |
import numpy as np
import vel.api.base as base
import vel.util.intepolate as interp
from vel.api import BatchInfo, EpochInfo, TrainingInfo
class CycleCallback(base.Callback):
""" A callback that manages setting the proper learning rate """
def __init__(self, optimizer, max_lr, min_lr, cycles, cycle_len=1, ... | vel/phase/cycle.py | import numpy as np
import vel.api.base as base
import vel.util.intepolate as interp
from vel.api import BatchInfo, EpochInfo, TrainingInfo
class CycleCallback(base.Callback):
""" A callback that manages setting the proper learning rate """
def __init__(self, optimizer, max_lr, min_lr, cycles, cycle_len=1, ... | 0.775902 | 0.150216 |
import sys, os
import importlib
import re
import logging
import requests
from os import listdir
from subprocess import Popen
from urllib.parse import urljoin
from threading import Thread, Lock
from distutils.version import LooseVersion
from bs4 import BeautifulSoup
from artifactory import ArtifactoryPath
__version__ =... | jfrog2pypi.py | import sys, os
import importlib
import re
import logging
import requests
from os import listdir
from subprocess import Popen
from urllib.parse import urljoin
from threading import Thread, Lock
from distutils.version import LooseVersion
from bs4 import BeautifulSoup
from artifactory import ArtifactoryPath
__version__ =... | 0.283285 | 0.09947 |
from tempest.api.identity import base
from tempest.lib.common.utils import data_utils
from tempest.lib import decorators
class PoliciesTestJSON(base.BaseIdentityV3AdminTest):
def _delete_policy(self, policy_id):
self.policies_client.delete_policy(policy_id)
@decorators.idempotent_id('1a0ad286-2d06-... | tempest/api/identity/admin/v3/test_policies.py |
from tempest.api.identity import base
from tempest.lib.common.utils import data_utils
from tempest.lib import decorators
class PoliciesTestJSON(base.BaseIdentityV3AdminTest):
def _delete_policy(self, policy_id):
self.policies_client.delete_policy(policy_id)
@decorators.idempotent_id('1a0ad286-2d06-... | 0.489015 | 0.338023 |
import time
from typing import Callable, List, Optional, Sequence, cast
import hydra
import numpy as np
import omegaconf
import torch
import torch.distributions
import mbrl.models
import mbrl.types
import mbrl.util.math
from .core import Agent, complete_agent_cfg
class Optimizer:
def __init__(self):
pa... | mbrl/planning/trajectory_opt.py | import time
from typing import Callable, List, Optional, Sequence, cast
import hydra
import numpy as np
import omegaconf
import torch
import torch.distributions
import mbrl.models
import mbrl.types
import mbrl.util.math
from .core import Agent, complete_agent_cfg
class Optimizer:
def __init__(self):
pa... | 0.952153 | 0.564249 |
from google.protobuf import descriptor as _descriptor
from google.protobuf import message as _message
from google.protobuf import reflection as _reflection
from google.protobuf import symbol_database as _symbol_database
# @@protoc_insertion_point(imports)
_sym_db = _symbol_database.Default()
from google.api import ... | protos/gen/python/protos/public/uac/Workspace_pb2.py |
from google.protobuf import descriptor as _descriptor
from google.protobuf import message as _message
from google.protobuf import reflection as _reflection
from google.protobuf import symbol_database as _symbol_database
# @@protoc_insertion_point(imports)
_sym_db = _symbol_database.Default()
from google.api import ... | 0.264928 | 0.08772 |