code stringlengths 2k 1.04M | repo_path stringlengths 5 517 | parsed_code stringlengths 0 1.04M | quality_prob float64 0.02 0.95 | learning_prob float64 0.02 0.93 |
|---|---|---|---|---|
from enum import Enum
import numpy as np
import scipy as sp
from scipy import sparse
from . import rulsif
def sq_puc_tr_rulsif(xp_tr, xu_tr, xu_te, prior, lambda_list=np.logspace(-3, 0, num=11),
gamma_list=None, sigma_list=None, n_fold=5, n_basis=200, kertype='gauss'):
if gamma_list is None... | src/puc/pu.py | from enum import Enum
import numpy as np
import scipy as sp
from scipy import sparse
from . import rulsif
def sq_puc_tr_rulsif(xp_tr, xu_tr, xu_te, prior, lambda_list=np.logspace(-3, 0, num=11),
gamma_list=None, sigma_list=None, n_fold=5, n_basis=200, kertype='gauss'):
if gamma_list is None... | 0.354433 | 0.323821 |
# flake8: noqa
from builtins import _test_sink, _test_source
from typing import Awaitable, Callable, TypeVar
from pyre_extensions import ParameterSpecification
from pyre_extensions.type_variable_operators import Concatenate
P = ParameterSpecification("P")
def with_logging(f: Callable[[int], None]) -> Callable[[in... | source/interprocedural_analyses/taint/test/integration/decorator.py |
# flake8: noqa
from builtins import _test_sink, _test_source
from typing import Awaitable, Callable, TypeVar
from pyre_extensions import ParameterSpecification
from pyre_extensions.type_variable_operators import Concatenate
P = ParameterSpecification("P")
def with_logging(f: Callable[[int], None]) -> Callable[[in... | 0.763307 | 0.419232 |
import os
from time import sleep
import subprocess
import shlex
from pathlib import Path
from urllib.error import URLError
from urllib.request import urlopen
import common
import upload_file_manage
import upload_process
# ログの設定
logger = common.logger_setup(__name__, True)
# インターネットに接続出来るか確認する
def is_internet_access... | src/uploader.py |
import os
from time import sleep
import subprocess
import shlex
from pathlib import Path
from urllib.error import URLError
from urllib.request import urlopen
import common
import upload_file_manage
import upload_process
# ログの設定
logger = common.logger_setup(__name__, True)
# インターネットに接続出来るか確認する
def is_internet_access... | 0.268462 | 0.100348 |
import http.client
import json
import logging
import os
import fnmatch
from collections import Counter
from dataclasses import dataclass
from datetime import datetime, timedelta, timezone
from http.server import BaseHTTPRequestHandler, HTTPServer
from urllib.parse import parse_qs, urlparse
logging.basicConfig(level=lo... | main.py | import http.client
import json
import logging
import os
import fnmatch
from collections import Counter
from dataclasses import dataclass
from datetime import datetime, timedelta, timezone
from http.server import BaseHTTPRequestHandler, HTTPServer
from urllib.parse import parse_qs, urlparse
logging.basicConfig(level=lo... | 0.525369 | 0.066751 |
from .xdcrnewbasetests import XDCRNewBaseTest
import time
class XDCRFilterTests(XDCRNewBaseTest):
def setUp(self):
XDCRNewBaseTest.setUp(self)
def tearDown(self):
XDCRNewBaseTest.tearDown(self)
def get_cluster_objects_for_input(self, input):
"""returns a list of cluster objects f... | pytests/xdcr/filterXDCR.py | from .xdcrnewbasetests import XDCRNewBaseTest
import time
class XDCRFilterTests(XDCRNewBaseTest):
def setUp(self):
XDCRNewBaseTest.setUp(self)
def tearDown(self):
XDCRNewBaseTest.tearDown(self)
def get_cluster_objects_for_input(self, input):
"""returns a list of cluster objects f... | 0.428951 | 0.180811 |
from pliers import config
from pliers.filters import FrameSamplingFilter
from pliers.extractors import (GoogleVisionAPIFaceExtractor,
GoogleVisionAPILabelExtractor,
GoogleVisionAPIPropertyExtractor,
GoogleVisionAPISafeSearchExt... | pliers/tests/extractors/api/test_google_extractors.py | from pliers import config
from pliers.filters import FrameSamplingFilter
from pliers.extractors import (GoogleVisionAPIFaceExtractor,
GoogleVisionAPILabelExtractor,
GoogleVisionAPIPropertyExtractor,
GoogleVisionAPISafeSearchExt... | 0.514644 | 0.328556 |
from __future__ import absolute_import
import os
import sys
import pytest
from mock import MagicMock, patch, mock_open
FILE_DIR = os.path.dirname(os.path.realpath(__file__))
#FIXTURES_DIR = os.path.join(FILE_DIR, "fixtures")
REPO_DIR = os.path.join(FILE_DIR, "..", "..")
# Add environ.py into path for testing
sys.pat... | tests/small/test_environ.py | from __future__ import absolute_import
import os
import sys
import pytest
from mock import MagicMock, patch, mock_open
FILE_DIR = os.path.dirname(os.path.realpath(__file__))
#FIXTURES_DIR = os.path.join(FILE_DIR, "fixtures")
REPO_DIR = os.path.join(FILE_DIR, "..", "..")
# Add environ.py into path for testing
sys.pat... | 0.284775 | 0.345216 |
import logging
import gevent
from volttron.platform.vip.agent import Agent
from volttrontesting.utils.platformwrapper import start_wrapper_platform
from volttron.platform.agent import json
import pytest
import random
import requests
import os
import tempfile
from volttrontesting.fixtures.volttron_platform_fixtures im... | volttrontesting/platform/test_platform_web.py | import logging
import gevent
from volttron.platform.vip.agent import Agent
from volttrontesting.utils.platformwrapper import start_wrapper_platform
from volttron.platform.agent import json
import pytest
import random
import requests
import os
import tempfile
from volttrontesting.fixtures.volttron_platform_fixtures im... | 0.462716 | 0.115112 |
from collections import OrderedDict
import pandas as pd
import numpy as np
from models.detectors.base import BaseDetector
from scipy.stats import ttest_ind
class LODA(BaseDetector):
def __init__(self):
super().__init__()
self.projections_ = None
self.histograms_ = None
self.limits... | PredictiveOutlierExplanationBenchmark/src/models/detectors/Loda.py | from collections import OrderedDict
import pandas as pd
import numpy as np
from models.detectors.base import BaseDetector
from scipy.stats import ttest_ind
class LODA(BaseDetector):
def __init__(self):
super().__init__()
self.projections_ = None
self.histograms_ = None
self.limits... | 0.856677 | 0.319068 |
import time
import sys
import json
from create_merge_topo import *
from client import *
from util import *
from threading import Thread, Lock, Condition
cv = Condition()
lock = Lock()
count = 0
nt = None
def run(i, nh, hosts, lock, cv):
global count
global nt
if len(hosts) == 0:
hosts = get_hosts... | Miscellaneous/TOPOLOGIE_FUNZIONANTI/esperimenti/e4/start.py |
import time
import sys
import json
from create_merge_topo import *
from client import *
from util import *
from threading import Thread, Lock, Condition
cv = Condition()
lock = Lock()
count = 0
nt = None
def run(i, nh, hosts, lock, cv):
global count
global nt
if len(hosts) == 0:
hosts = get_hosts... | 0.165054 | 0.054626 |
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from absl.testing import absltest
from future.builtins import range # pylint: disable=redefined-builtin
from pysc2 import maps
from pysc2 import run_configs
from pysc2.lib import actions
from pysc2.lib import... | pysc2/tests/replay_obs_test.py | from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from absl.testing import absltest
from future.builtins import range # pylint: disable=redefined-builtin
from pysc2 import maps
from pysc2 import run_configs
from pysc2.lib import actions
from pysc2.lib import... | 0.677794 | 0.258613 |
import flask_sijax
from flask import render_template, g, session, request, redirect
import model
from apic_manager import cobra_apic_l2_tool
from app import app
from sijax_handlers.group_handler import group_handler
from sijax_handlers.network_handler import network_handler
from sijax_handlers.fabric_handler import fa... | app/views.py | import flask_sijax
from flask import render_template, g, session, request, redirect
import model
from apic_manager import cobra_apic_l2_tool
from app import app
from sijax_handlers.group_handler import group_handler
from sijax_handlers.network_handler import network_handler
from sijax_handlers.fabric_handler import fa... | 0.269902 | 0.056522 |
import os
import time
import glob
import sched
import multiprocessing
from superbench.common.utils import logger, run_command
from superbench.common.utils import device_manager as dm
from superbench.monitor.record import MonitorRecord
class Monitor(multiprocessing.Process):
"""The monitor class to collect system... | superbench/monitor/monitor.py | import os
import time
import glob
import sched
import multiprocessing
from superbench.common.utils import logger, run_command
from superbench.common.utils import device_manager as dm
from superbench.monitor.record import MonitorRecord
class Monitor(multiprocessing.Process):
"""The monitor class to collect system... | 0.61231 | 0.08617 |
# daal4py DBSCAN example for shared memory systems
import daal4py as d4p
import numpy as np
import os
from daal4py.oneapi import sycl_buffer
# let's try to use pandas' fast csv reader
try:
import pandas
read_csv = lambda f, c, t=np.float64: pandas.read_csv(f, usecols=c, delimiter=',', header=None, dtype=t)
e... | examples/sycl/dbscan_batch.py |
# daal4py DBSCAN example for shared memory systems
import daal4py as d4p
import numpy as np
import os
from daal4py.oneapi import sycl_buffer
# let's try to use pandas' fast csv reader
try:
import pandas
read_csv = lambda f, c, t=np.float64: pandas.read_csv(f, usecols=c, delimiter=',', header=None, dtype=t)
e... | 0.806586 | 0.365542 |
from __future__ import unicode_literals
from flask import jsonify, request
from indico.core.db import db
from indico.core.db.sqlalchemy.util.queries import preprocess_ts_string
from indico.modules.events.logs.models.entries import EventLogEntry, EventLogRealm
from indico.modules.events.logs.util import serialize_log... | indico/modules/events/logs/controllers.py |
from __future__ import unicode_literals
from flask import jsonify, request
from indico.core.db import db
from indico.core.db.sqlalchemy.util.queries import preprocess_ts_string
from indico.modules.events.logs.models.entries import EventLogEntry, EventLogRealm
from indico.modules.events.logs.util import serialize_log... | 0.729712 | 0.084568 |
# Copyright (c) Microsoft Corporation.
# Licensed under the MIT license.
import logging
from typing import Iterator, List, Optional
from asconnect.httpclient import HttpClient
from asconnect.models import (
AppStoreVersion,
Platform,
AppStoreVersionLocalization,
AppStoreReviewDetails,
IdfaDeclara... | asconnect/version_client.py |
# Copyright (c) Microsoft Corporation.
# Licensed under the MIT license.
import logging
from typing import Iterator, List, Optional
from asconnect.httpclient import HttpClient
from asconnect.models import (
AppStoreVersion,
Platform,
AppStoreVersionLocalization,
AppStoreReviewDetails,
IdfaDeclara... | 0.931299 | 0.101145 |
r'''
Copyright 2014 Google Inc. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to i... | nogotofail/mitm/connection/handlers/data/imap.py | r'''
Copyright 2014 Google Inc. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to i... | 0.612889 | 0.108142 |
"""Command line interface for extending feed effective dates."""
import argparse
import csv
from datetime import datetime, timedelta
import logging
import os
import shutil
import sys
import zipfile
DOWNLOAD_DIRECTORY = 'gtfs'
# extend feed effective date range this far into the past and future
EFFECTIVE_DAYS = 365
GTF... | extend_effective_dates.py | """Command line interface for extending feed effective dates."""
import argparse
import csv
from datetime import datetime, timedelta
import logging
import os
import shutil
import sys
import zipfile
DOWNLOAD_DIRECTORY = 'gtfs'
# extend feed effective date range this far into the past and future
EFFECTIVE_DAYS = 365
GTF... | 0.46952 | 0.147893 |
import os
import sys
import time
import ConfigParser
import pandas as pd
import numpy as np
import theano
import theano.tensor as T
import cPickle
theano.config.floatX = 'float32'
base_path = os.path.dirname(__file__)
sys.path.insert(1,os.path.join(base_path, '../external'))
sys.path.insert(2,os.path.join(base_path,... | code/model/deleteme/cnn_model.py |
import os
import sys
import time
import ConfigParser
import pandas as pd
import numpy as np
import theano
import theano.tensor as T
import cPickle
theano.config.floatX = 'float32'
base_path = os.path.dirname(__file__)
sys.path.insert(1,os.path.join(base_path, '../external'))
sys.path.insert(2,os.path.join(base_path,... | 0.330039 | 0.117471 |
import os
os.environ['TF_CPP_MIN_LOG_LEVEL'] = "2"
import wandb
import sys
import multiprocessing
import collections
import random
import warnings
import numpy as np
import tensorflow as tf
from tensorflow.keras.utils import to_categorical
from tensorflow.keras.preprocessing import sequence
from tensorflow.keras.mode... | lab/train-stable.py | import os
os.environ['TF_CPP_MIN_LOG_LEVEL'] = "2"
import wandb
import sys
import multiprocessing
import collections
import random
import warnings
import numpy as np
import tensorflow as tf
from tensorflow.keras.utils import to_categorical
from tensorflow.keras.preprocessing import sequence
from tensorflow.keras.mode... | 0.539954 | 0.194846 |
import tensorflow as tf
from tensorflow.keras import layers
from tensorflow.keras import models
class SiamModel(tf.keras.layers.Layer):
def __init__(self, input_shape=16, dense1_shape=32, dense2_shape=64, name='IntenalSiamModel', **kwargs):
super(SiamModel, self).__init__(name=name, **kwargs)
self... | src/siamese_model.py | import tensorflow as tf
from tensorflow.keras import layers
from tensorflow.keras import models
class SiamModel(tf.keras.layers.Layer):
def __init__(self, input_shape=16, dense1_shape=32, dense2_shape=64, name='IntenalSiamModel', **kwargs):
super(SiamModel, self).__init__(name=name, **kwargs)
self... | 0.926112 | 0.432842 |
from __future__ import print_function
import argparse
import os
import resource
import sys
USAGE_PROGRAM = ('%s -m oslo_concurrency.prlimit'
% os.path.basename(sys.executable))
RESOURCES = (
# argparse argument => resource
('as', resource.RLIMIT_AS),
('core', resource.RLIMIT_CORE),
... | oslo_concurrency/prlimit.py |
from __future__ import print_function
import argparse
import os
import resource
import sys
USAGE_PROGRAM = ('%s -m oslo_concurrency.prlimit'
% os.path.basename(sys.executable))
RESOURCES = (
# argparse argument => resource
('as', resource.RLIMIT_AS),
('core', resource.RLIMIT_CORE),
... | 0.454956 | 0.052062 |
import shutil
import os
import json
import re
import time
import hashlib
import uuid
from typing import List, Optional, Union, Tuple
from aim.__version__ import __version__ as aim_version
from aim.engine.configs import *
from aim.engine.utils import (
ls_dir,
deep_compare,
import_module,
clean_repo_pat... | aim/engine/repo/repo.py | import shutil
import os
import json
import re
import time
import hashlib
import uuid
from typing import List, Optional, Union, Tuple
from aim.__version__ import __version__ as aim_version
from aim.engine.configs import *
from aim.engine.utils import (
ls_dir,
deep_compare,
import_module,
clean_repo_pat... | 0.408631 | 0.078395 |
# use tdklib library,which provides a wrapper for tdk testcase script
import tdklib;
import time;
#Test component to be tested
obj = tdklib.TDKScriptingLibrary("cmhal","1");
obj1 = tdklib.TDKScriptingLibrary("tdkbtr181","1");
#IP and Port of box, No need to change,
#This will be replaced with correspoing Box Ip and p... | testscripts/RDKB/component/CMHAL/TS_CMHAL_ClearDocsisEventLog.py | # use tdklib library,which provides a wrapper for tdk testcase script
import tdklib;
import time;
#Test component to be tested
obj = tdklib.TDKScriptingLibrary("cmhal","1");
obj1 = tdklib.TDKScriptingLibrary("tdkbtr181","1");
#IP and Port of box, No need to change,
#This will be replaced with correspoing Box Ip and p... | 0.286169 | 0.3031 |
import json
import unittest
import boto3
from botocore.exceptions import (
ClientError,
)
from moto import (
mock_s3,
mock_sts,
)
from azul import (
cached_property,
config,
)
from azul.logging import (
configure_test_logging,
)
from azul.plugins import (
RepositoryPlugin,
)
from azul.port... | test/service/test_portal_service.py | import json
import unittest
import boto3
from botocore.exceptions import (
ClientError,
)
from moto import (
mock_s3,
mock_sts,
)
from azul import (
cached_property,
config,
)
from azul.logging import (
configure_test_logging,
)
from azul.plugins import (
RepositoryPlugin,
)
from azul.port... | 0.52342 | 0.204461 |
import machine
from time import sleep_us
class pca9865(object):
'''16 servo contoller. Use index 0-15 for the servo #.'''
_ADDRESS = 0x40
_MODE1 = 0
_PRESCALE = 0xFE
_LED0_ON_L = 0x6 # We only use LED0 and offset 0-16 from it.
# _LED0_ON_H = const(0x7)
# _LED0_OFF_L = const(0x8)
... | Projects/ESP32Micropython/pca9865.py |
import machine
from time import sleep_us
class pca9865(object):
'''16 servo contoller. Use index 0-15 for the servo #.'''
_ADDRESS = 0x40
_MODE1 = 0
_PRESCALE = 0xFE
_LED0_ON_L = 0x6 # We only use LED0 and offset 0-16 from it.
# _LED0_ON_H = const(0x7)
# _LED0_OFF_L = const(0x8)
... | 0.464659 | 0.222531 |
CSV-compare
-----------
Compare table data stored in CSV (comma seperated values) format.
"""
import re
import csv
import sys
import os
def _pr_list(l1, l2, replace_chars = '[\n ]'):
""" Calculate precision and recall regarding elements of a list.
When a 1:1 match canno... | script/csv-compare.py |
CSV-compare
-----------
Compare table data stored in CSV (comma seperated values) format.
"""
import re
import csv
import sys
import os
def _pr_list(l1, l2, replace_chars = '[\n ]'):
""" Calculate precision and recall regarding elements of a list.
When a 1:1 match canno... | 0.276007 | 0.466663 |
import numpy as np
from numpy.testing import assert_equal, assert_raises
from numpy.testing import assert_array_almost_equal
from scipy import sparse
from sklearn.utils.testing import assert_less
from sklearn.linear_model import LinearRegression, RANSACRegressor
from sklearn.linear_model.ransac import _dynamic_max_tri... | summary/sumy/sklearn/linear_model/tests/test_ransac.py | import numpy as np
from numpy.testing import assert_equal, assert_raises
from numpy.testing import assert_array_almost_equal
from scipy import sparse
from sklearn.utils.testing import assert_less
from sklearn.linear_model import LinearRegression, RANSACRegressor
from sklearn.linear_model.ransac import _dynamic_max_tri... | 0.850686 | 0.684468 |
from __future__ import division
import sys
import os
import unittest
import numpy as np
from numpy.testing import assert_allclose
from quantecon.lqnash import nnash
from quantecon.lqcontrol import LQ
class TestLQNash(unittest.TestCase):
def test_noninteractive(self):
"Test case for when agents don't inter... | quantecon/tests/test_lqnash.py | from __future__ import division
import sys
import os
import unittest
import numpy as np
from numpy.testing import assert_allclose
from quantecon.lqnash import nnash
from quantecon.lqcontrol import LQ
class TestLQNash(unittest.TestCase):
def test_noninteractive(self):
"Test case for when agents don't inter... | 0.392453 | 0.592195 |
"""Current-flow closeness centrality measures."""
import networkx as nx
from networkx.utils import not_implemented_for, reverse_cuthill_mckee_ordering
from networkx.algorithms.centrality.flow_matrix import *
__all__ = ['current_flow_closeness_centrality', 'information_centrality']
@not_implemented_for('directed')
d... | src/networkx/algorithms/centrality/current_flow_closeness.py | """Current-flow closeness centrality measures."""
import networkx as nx
from networkx.utils import not_implemented_for, reverse_cuthill_mckee_ordering
from networkx.algorithms.centrality.flow_matrix import *
__all__ = ['current_flow_closeness_centrality', 'information_centrality']
@not_implemented_for('directed')
d... | 0.918233 | 0.696391 |
from security_monkey.datastore import Account, AccountType, Technology
from security_monkey.tests import SecurityMonkeyTestCase
from security_monkey import db
from security_monkey.watchers.github.org import GitHubOrgItem
from security_monkey.auditors.github.repo import GitHubRepoAuditor
CONFIG_ONE = {
"id": 12962... | security_monkey/tests/auditors/github/test_repo_auditor.py | from security_monkey.datastore import Account, AccountType, Technology
from security_monkey.tests import SecurityMonkeyTestCase
from security_monkey import db
from security_monkey.watchers.github.org import GitHubOrgItem
from security_monkey.auditors.github.repo import GitHubRepoAuditor
CONFIG_ONE = {
"id": 12962... | 0.466359 | 0.372848 |
from django.utils.translation import ugettext_lazy as _
from django.contrib.sites.models import Site
from cms.plugin_base import CMSPluginBase
from cms.plugin_pool import plugin_pool
from .forms import LinkForm
from .models import Link
class LinkPlugin(CMSPluginBase):
model = Link
form = LinkForm
name =... | tech_project/lib/python2.7/site-packages/djangocms_link/cms_plugins.py | from django.utils.translation import ugettext_lazy as _
from django.contrib.sites.models import Site
from cms.plugin_base import CMSPluginBase
from cms.plugin_pool import plugin_pool
from .forms import LinkForm
from .models import Link
class LinkPlugin(CMSPluginBase):
model = Link
form = LinkForm
name =... | 0.418578 | 0.105487 |
from unittest import TestCase, skip # use TestCase and skip
from pathlib import Path # use Path
from itertools import zip_longest
from veniq.utils.ast_builder import build_ast
from veniq.ast_framework import AST, ASTNodeType
from veniq.ast_framework.ast import MemberReferenceParams, MethodInvocationParams
import os #... | test/ast_framework/test_ast.py | from unittest import TestCase, skip # use TestCase and skip
from pathlib import Path # use Path
from itertools import zip_longest
from veniq.utils.ast_builder import build_ast
from veniq.ast_framework import AST, ASTNodeType
from veniq.ast_framework.ast import MemberReferenceParams, MethodInvocationParams
import os #... | 0.492432 | 0.481149 |
# Copyright (c) Facebook, Inc. and its affiliates.
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
import contextlib
import os
import random
import time
import copy
import multiprocessing
import psutil
import socket
import warnings
from colle... | bps_nav/rl/ddppo/algo/ddppo_trainer.py |
# Copyright (c) Facebook, Inc. and its affiliates.
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
import contextlib
import os
import random
import time
import copy
import multiprocessing
import psutil
import socket
import warnings
from colle... | 0.776199 | 0.112429 |
import argparse
import os
import errno
import subprocess
import sys
import venv
from common_setup import running_on_ci, remote_cache_token, which
from torch_blade_build import TorchBladeBuild, get_fullpath_or_create
cwd = os.path.dirname(os.path.abspath(__file__))
def _make_executable(path):
mode = os.stat(path... | pytorch_blade/bazel_build.py |
import argparse
import os
import errno
import subprocess
import sys
import venv
from common_setup import running_on_ci, remote_cache_token, which
from torch_blade_build import TorchBladeBuild, get_fullpath_or_create
cwd = os.path.dirname(os.path.abspath(__file__))
def _make_executable(path):
mode = os.stat(path... | 0.34798 | 0.072933 |
import os
from pathlib import Path
from rpi.inputs2 import *
start_mtime=0
first_mtime=0
header="# timecode format v2"
pts_default="extracted"
pts_name=""
times=[]
start_time=0
min_start=0 # unit s: 1 µs
max_start=2592000 # unit s: 30 d
default_start=0
interval=0
min_interval=0.000001 # unit s: 1 µs
... | python/ptsextract.py | import os
from pathlib import Path
from rpi.inputs2 import *
start_mtime=0
first_mtime=0
header="# timecode format v2"
pts_default="extracted"
pts_name=""
times=[]
start_time=0
min_start=0 # unit s: 1 µs
max_start=2592000 # unit s: 30 d
default_start=0
interval=0
min_interval=0.000001 # unit s: 1 µs
... | 0.119434 | 0.09709 |
from typing import List, Any, Dict
import pandas as pd
import requests
from .settings import BASE_URL
import delta_sharing
class FidapClient:
"""
class for fidap client
"""
_api_key = None
_api_secret = None
_file_path = "https://fidap.s3-us-west-2.amazonaws.com/fidap_data.share"
_custom_... | fidap/fidap.py | from typing import List, Any, Dict
import pandas as pd
import requests
from .settings import BASE_URL
import delta_sharing
class FidapClient:
"""
class for fidap client
"""
_api_key = None
_api_secret = None
_file_path = "https://fidap.s3-us-west-2.amazonaws.com/fidap_data.share"
_custom_... | 0.827967 | 0.167934 |
import pyro
from ..gp import GP
from ._pyro_mixin import _PyroMixin
class PyroGP(GP, _PyroMixin):
"""
A :obj:`~gpytorch.models.ApproximateGP` designed to work with Pyro.
This module makes it possible to include GP models with more complex probablistic models,
or to use likelihood functions with add... | gpytorch/models/pyro/pyro_gp.py |
import pyro
from ..gp import GP
from ._pyro_mixin import _PyroMixin
class PyroGP(GP, _PyroMixin):
"""
A :obj:`~gpytorch.models.ApproximateGP` designed to work with Pyro.
This module makes it possible to include GP models with more complex probablistic models,
or to use likelihood functions with add... | 0.950371 | 0.702849 |
import requests
url = "http://192.168.86.192/targetcmd/"
ramRomPageWrBase = 0x70
ramRomPgenWrBase = 0x74
def takeControl():
req = requests.request("get", url + "rawBusControlOn")
req = requests.request("get", url + "rawBusWaitDisable")
req = requests.request("get", url + "rawBusWaitClear")
req = reque... | PiSw/examples/hardwareDebug/hwDebugRRRegSet.py | import requests
url = "http://192.168.86.192/targetcmd/"
ramRomPageWrBase = 0x70
ramRomPgenWrBase = 0x74
def takeControl():
req = requests.request("get", url + "rawBusControlOn")
req = requests.request("get", url + "rawBusWaitDisable")
req = requests.request("get", url + "rawBusWaitClear")
req = reque... | 0.06028 | 0.050894 |
from django.test import TestCase
from django.contrib.auth import get_user_model
from rest_framework.test import APIClient
from .models import Notion
# Create your tests here.
User = get_user_model()
class NotionTestCase(TestCase):
def setUp(self):
self.user = User.objects.create_user(username='testuser',... | myapp/tests.py | from django.test import TestCase
from django.contrib.auth import get_user_model
from rest_framework.test import APIClient
from .models import Notion
# Create your tests here.
User = get_user_model()
class NotionTestCase(TestCase):
def setUp(self):
self.user = User.objects.create_user(username='testuser',... | 0.348534 | 0.222352 |
import argparse
from corpus_readers import PandasBasedCorpus
import json
from cv_utils import CVManager, run_cv_evaluation
from global_constants import *
import os
from collections import defaultdict
from gram_matrix_extractors import compute_default_predefined_coling_gram_matrices
from corpus_readers import unpickle_... | scripts/emnlp2018/run_experiments.py | import argparse
from corpus_readers import PandasBasedCorpus
import json
from cv_utils import CVManager, run_cv_evaluation
from global_constants import *
import os
from collections import defaultdict
from gram_matrix_extractors import compute_default_predefined_coling_gram_matrices
from corpus_readers import unpickle_... | 0.480235 | 0.166879 |
from django.db import models
from cloudinary.models import CloudinaryField
from cloudinary.uploader import upload
from django.utils import timezone
from django.contrib.auth.models import AbstractBaseUser, PermissionsMixin, BaseUserManager
class UserManager(BaseUserManager):
def create_user(self, first_name, last_n... | apps/api/models.py | from django.db import models
from cloudinary.models import CloudinaryField
from cloudinary.uploader import upload
from django.utils import timezone
from django.contrib.auth.models import AbstractBaseUser, PermissionsMixin, BaseUserManager
class UserManager(BaseUserManager):
def create_user(self, first_name, last_n... | 0.558327 | 0.089018 |
from django.http import Http404
from django.contrib import messages
from django.core.paginator import Paginator
from django.contrib.auth.decorators import login_required
from django.shortcuts import get_object_or_404, redirect, render
from .models import Entry, Topic
from .forms import EntryForm, TopicForm
def check... | learning_log/views.py | from django.http import Http404
from django.contrib import messages
from django.core.paginator import Paginator
from django.contrib.auth.decorators import login_required
from django.shortcuts import get_object_or_404, redirect, render
from .models import Entry, Topic
from .forms import EntryForm, TopicForm
def check... | 0.66236 | 0.138812 |
import json
from json import JSONDecodeError
import argparse
import urllib.request
from colorama import Fore
from prettytable import PrettyTable
statistics = "/var/log/dystopia/statistics.json"
key_file = "/var/log/dystopia/ipstack.key"
def print_message(message):
print(Fore.GREEN + "[*] " + Fore.WHITE + messag... | tools/dstat.py | import json
from json import JSONDecodeError
import argparse
import urllib.request
from colorama import Fore
from prettytable import PrettyTable
statistics = "/var/log/dystopia/statistics.json"
key_file = "/var/log/dystopia/ipstack.key"
def print_message(message):
print(Fore.GREEN + "[*] " + Fore.WHITE + messag... | 0.287368 | 0.10725 |
import datetime
import jwt
from django import forms
from django.core.exceptions import PermissionDenied
from django.http import Http404
from django.shortcuts import redirect
from django.urls import reverse
from django.utils.decorators import method_decorator
from django.utils.timezone import now
from django.utils.tran... | pretix_venueless/views.py | import datetime
import jwt
from django import forms
from django.core.exceptions import PermissionDenied
from django.http import Http404
from django.shortcuts import redirect
from django.urls import reverse
from django.utils.decorators import method_decorator
from django.utils.timezone import now
from django.utils.tran... | 0.470737 | 0.064742 |
import json, urlparse, sys, os, signal
from BaseHTTPServer import BaseHTTPRequestHandler, HTTPServer
from subprocess import call
from BitbucketParse import BitbucketParse
class GitAutoDeploy(BaseHTTPRequestHandler):
CONFIG_FILEPATH = './GitAutoDeploy.conf.json'
config = None
quiet = False
daemon = Fa... | GitAutoDeploy.py |
import json, urlparse, sys, os, signal
from BaseHTTPServer import BaseHTTPRequestHandler, HTTPServer
from subprocess import call
from BitbucketParse import BitbucketParse
class GitAutoDeploy(BaseHTTPRequestHandler):
CONFIG_FILEPATH = './GitAutoDeploy.conf.json'
config = None
quiet = False
daemon = Fa... | 0.200088 | 0.049982 |
# Core imports
import os
import copy
import sys
from datetime import datetime
# Scipy/numpy imports
import numpy as np
# Astropy imports
from astropy.table import Table
import astropy.units as u
from astropy.stats import sigma_clipped_stats
# Import astroimage
import astroimage as ai
ai.set_instrument('Mimir')
#===... | 06b_photometricCalibration.py | # Core imports
import os
import copy
import sys
from datetime import datetime
# Scipy/numpy imports
import numpy as np
# Astropy imports
from astropy.table import Table
import astropy.units as u
from astropy.stats import sigma_clipped_stats
# Import astroimage
import astroimage as ai
ai.set_instrument('Mimir')
#===... | 0.40486 | 0.319626 |
import struct
from . import packet_base
from ryu.lib import addrconv
# Slow Protocol Multicast destination
SLOW_PROTOCOL_MULTICAST = '01:80:c2:00:00:02'
# Slow Protocol SubType
SLOW_SUBTYPE_LACP = 0x01
SLOW_SUBTYPE_MARKER = 0x02
SLOW_SUBTYPE_OAM = 0x03
SLOW_SUBTYPE_OSSP = 0x0a
class slow(packet_base.PacketBase):
... | ryu/lib/packet/slow.py |
import struct
from . import packet_base
from ryu.lib import addrconv
# Slow Protocol Multicast destination
SLOW_PROTOCOL_MULTICAST = '01:80:c2:00:00:02'
# Slow Protocol SubType
SLOW_SUBTYPE_LACP = 0x01
SLOW_SUBTYPE_MARKER = 0x02
SLOW_SUBTYPE_OAM = 0x03
SLOW_SUBTYPE_OSSP = 0x0a
class slow(packet_base.PacketBase):
... | 0.444565 | 0.379666 |
import sys
import os
from os.path import isfile, join, exists
from os import listdir, stat, makedirs
from datetime import datetime
from time import strftime
from platform import platform
def getProjectList(wd,inpfile):
with open(join(wd,inpfile),'r') as inp:
lines = inp.readlines()
projects = []
f... | python/joinCSVs.py | import sys
import os
from os.path import isfile, join, exists
from os import listdir, stat, makedirs
from datetime import datetime
from time import strftime
from platform import platform
def getProjectList(wd,inpfile):
with open(join(wd,inpfile),'r') as inp:
lines = inp.readlines()
projects = []
f... | 0.111241 | 0.067148 |
import posixpath
import json
from ._2to3 import STRTYPE, iteritems_
from .index_constants import JSON_INDEX_TYPE
from .index_constants import TEXT_INDEX_TYPE
from .index_constants import SPECIAL_INDEX_TYPE
from .index_constants import TEXT_INDEX_ARGS
from .errors import CloudantArgumentError, CloudantException
class ... | src/cloudant/indexes.py | import posixpath
import json
from ._2to3 import STRTYPE, iteritems_
from .index_constants import JSON_INDEX_TYPE
from .index_constants import TEXT_INDEX_TYPE
from .index_constants import SPECIAL_INDEX_TYPE
from .index_constants import TEXT_INDEX_ARGS
from .errors import CloudantArgumentError, CloudantException
class ... | 0.788624 | 0.283589 |
from __future__ import division
import sys
from pprint import pprint as pp
import requests
import re
import string
import operator
import getopt
import time
class hist():
def __init__(self, data):
self.mi = float(data['m'])
self.ma = float(data['M'])
self.num = int(data['n'])
self.d... | bin/hist_print.py | from __future__ import division
import sys
from pprint import pprint as pp
import requests
import re
import string
import operator
import getopt
import time
class hist():
def __init__(self, data):
self.mi = float(data['m'])
self.ma = float(data['M'])
self.num = int(data['n'])
self.d... | 0.410047 | 0.139016 |
import clr
# Import python sys module
import sys
# Import os module
import os
# Import System.IO for saving and opening files
from System.IO import *
# Import C compatible List and String
from System import String
from System.Collections.Generic import List
# Add needed dll references
sys.path.appe... | LFAutomation/Python/sequential_gating.py | import clr
# Import python sys module
import sys
# Import os module
import os
# Import System.IO for saving and opening files
from System.IO import *
# Import C compatible List and String
from System import String
from System.Collections.Generic import List
# Add needed dll references
sys.path.appe... | 0.356447 | 0.109992 |
import cv2
from PySide2.QtWidgets import QMainWindow, QFileDialog, QHBoxLayout
from main_interface import gui_main_interface
from PySide2.QtCore import QCoreApplication, Slot, Qt
from tools import add_tree_item, show_image_data, modify_graphics, widget_set
from opencv_function import function_warpaffine, function_cvtco... | src/main_interface/main_interface.py | import cv2
from PySide2.QtWidgets import QMainWindow, QFileDialog, QHBoxLayout
from main_interface import gui_main_interface
from PySide2.QtCore import QCoreApplication, Slot, Qt
from tools import add_tree_item, show_image_data, modify_graphics, widget_set
from opencv_function import function_warpaffine, function_cvtco... | 0.122418 | 0.114467 |
from sympy import oo
from sympy.core import igcd
from sympy.polys.monomials import monomial_min, monomial_div
from sympy.polys.orderings import monomial_key
import random
def poly_LC(f, K):
"""
Return leading coefficient of ``f``.
Examples
========
>>> from sympy.polys.domains import ZZ
>>> ... | sympy/polys/densebasic.py | from sympy import oo
from sympy.core import igcd
from sympy.polys.monomials import monomial_min, monomial_div
from sympy.polys.orderings import monomial_key
import random
def poly_LC(f, K):
"""
Return leading coefficient of ``f``.
Examples
========
>>> from sympy.polys.domains import ZZ
>>> ... | 0.861756 | 0.450239 |
import mmcv
import numpy as np
import trimesh
from os import path as osp
def _write_ply(points, out_filename):
"""Write points into ``ply`` format for meshlab visualization.
Args:
points (np.ndarray): Points in shape (N, dim).
out_filename (str): Filename to be saved.
"""
N = points.s... | mmdet3d/core/visualizer/show_result.py | import mmcv
import numpy as np
import trimesh
from os import path as osp
def _write_ply(points, out_filename):
"""Write points into ``ply`` format for meshlab visualization.
Args:
points (np.ndarray): Points in shape (N, dim).
out_filename (str): Filename to be saved.
"""
N = points.s... | 0.675551 | 0.591605 |
import asyncio
import os
import aiohttp
import discord
import orjson
import uvloop
from discord.ext import commands
from dotenv import load_dotenv
load_dotenv()
Tenor_API_Key = os.getenv("Tenor_API_Key")
class TenorV1(commands.Cog):
def __init__(self, bot):
self.bot = bot
@comma... | Bot/Cogs/tenor.py | import asyncio
import os
import aiohttp
import discord
import orjson
import uvloop
from discord.ext import commands
from dotenv import load_dotenv
load_dotenv()
Tenor_API_Key = os.getenv("Tenor_API_Key")
class TenorV1(commands.Cog):
def __init__(self, bot):
self.bot = bot
@comma... | 0.275909 | 0.080105 |
from io import BytesIO
import factory
import pytest
from django.core.management import call_command
from reversion.models import Version
from datahub.metadata.test.factories import SectorFactory
pytestmark = pytest.mark.django_db
def test_happy_path(s3_stubber):
"""Test that the command updates the specified r... | datahub/dbmaintenance/test/commands/test_update_sector_segment.py | from io import BytesIO
import factory
import pytest
from django.core.management import call_command
from reversion.models import Version
from datahub.metadata.test.factories import SectorFactory
pytestmark = pytest.mark.django_db
def test_happy_path(s3_stubber):
"""Test that the command updates the specified r... | 0.788054 | 0.310172 |
import collections
import json
import os
import subprocess
import sys
import urllib
import constants
import io_stats_parser
class DeviceStatsMonitor(object):
"""Class for collecting device stats such as IO/CPU usage.
Args:
adb: Instance of AndroidComannds.
hz: Frequency at which to sample device sta... | build/android/pylib/device_stats_monitor.py | import collections
import json
import os
import subprocess
import sys
import urllib
import constants
import io_stats_parser
class DeviceStatsMonitor(object):
"""Class for collecting device stats such as IO/CPU usage.
Args:
adb: Instance of AndroidComannds.
hz: Frequency at which to sample device sta... | 0.44553 | 0.176388 |
import functools
import numpy as np
from arch.api.proto.feature_scale_meta_pb2 import ScaleMeta
from arch.api.proto.feature_scale_param_pb2 import ScaleParam
from arch.api.proto.feature_scale_param_pb2 import ColumnScaleParam
from arch.api.utils import log_utils
from federatedml.feature.feature_scale.base_scale impor... | federatedml/feature/feature_scale/min_max_scale.py |
import functools
import numpy as np
from arch.api.proto.feature_scale_meta_pb2 import ScaleMeta
from arch.api.proto.feature_scale_param_pb2 import ScaleParam
from arch.api.proto.feature_scale_param_pb2 import ColumnScaleParam
from arch.api.utils import log_utils
from federatedml.feature.feature_scale.base_scale impor... | 0.792183 | 0.308835 |
from functools import partial
from kivy.clock import Clock
from kivy.graphics import Rectangle
from kivy.uix.boxlayout import BoxLayout
from kivy.uix.gridlayout import GridLayout
from kivy.uix.spinner import Spinner
from kivy.uix.scrollview import ScrollView
from kivy.uix.floatlayout import FloatLayout
from k... | ui/popups.py | from functools import partial
from kivy.clock import Clock
from kivy.graphics import Rectangle
from kivy.uix.boxlayout import BoxLayout
from kivy.uix.gridlayout import GridLayout
from kivy.uix.spinner import Spinner
from kivy.uix.scrollview import ScrollView
from kivy.uix.floatlayout import FloatLayout
from k... | 0.608245 | 0.172939 |
from __future__ import absolute_import
from __future__ import unicode_literals
import fluff
from casexml.apps.case.models import CommCareCase
from corehq.fluff.calculators.case import CasePropertyFilter
from custom.care_pathways import DOMAINS
from custom.care_pathways.utils import get_domain_configuration
# This calc... | custom/care_pathways/models.py | from __future__ import absolute_import
from __future__ import unicode_literals
import fluff
from casexml.apps.case.models import CommCareCase
from corehq.fluff.calculators.case import CasePropertyFilter
from custom.care_pathways import DOMAINS
from custom.care_pathways.utils import get_domain_configuration
# This calc... | 0.534612 | 0.172137 |
from context import _loss_func_semi_vectorized
from context import _loss_func_theano
import unittest
import sklearn.preprocessing
import theano
import numpy as np
class TheanoLossFunctionsTestSuite(unittest.TestCase):
"""Advanced test cases."""
def test_hinge_loss(self):
W = np.random.random((10, 10... | tests/test_theano_loss_functios.py |
from context import _loss_func_semi_vectorized
from context import _loss_func_theano
import unittest
import sklearn.preprocessing
import theano
import numpy as np
class TheanoLossFunctionsTestSuite(unittest.TestCase):
"""Advanced test cases."""
def test_hinge_loss(self):
W = np.random.random((10, 10... | 0.841403 | 0.665438 |
import asyncio
import audioop
import enum
import functools
import logging
from typing import Callable, Dict, Optional, Union
import discord
from concord.ext.audio.exceptions import AudioExtensionError
log = logging.getLogger(__name__)
class State:
"""Global state with guild's audio states."""
_audio_stat... | concord/ext/audio/state.py | import asyncio
import audioop
import enum
import functools
import logging
from typing import Callable, Dict, Optional, Union
import discord
from concord.ext.audio.exceptions import AudioExtensionError
log = logging.getLogger(__name__)
class State:
"""Global state with guild's audio states."""
_audio_stat... | 0.83767 | 0.154089 |
from zvt.domain import FinanceDebtpayingAbility
from zvt.recorders.emquantapi.finance.base_china_stock_finance_recorder import EmBaseChinaStockFinanceRecorder
from zvt.utils.utils import add_func_to_value, first_item_to_float
financial_debtpayingability_map = {
'debt_asset_ratio': 'LIBILITYTOASSET', # 资产负债率
... | zvt/recorders/emquantapi/finance/china_stock_finance_debtpayingability.py | from zvt.domain import FinanceDebtpayingAbility
from zvt.recorders.emquantapi.finance.base_china_stock_finance_recorder import EmBaseChinaStockFinanceRecorder
from zvt.utils.utils import add_func_to_value, first_item_to_float
financial_debtpayingability_map = {
'debt_asset_ratio': 'LIBILITYTOASSET', # 资产负债率
... | 0.310172 | 0.242923 |
import re
from collections import namedtuple
from itertools import chain
from pokertools import (
CANONICAL_HOLECARDS_NAMES,
SUIT_COMBINATIONS,
SUIT_PERMUATIONS,
SUITS,
get_numerical_rank,
get_string_rank,
holecards,
)
#----------------------------------------------------------------------... | examples/translation.py | import re
from collections import namedtuple
from itertools import chain
from pokertools import (
CANONICAL_HOLECARDS_NAMES,
SUIT_COMBINATIONS,
SUIT_PERMUATIONS,
SUITS,
get_numerical_rank,
get_string_rank,
holecards,
)
#----------------------------------------------------------------------... | 0.538498 | 0.310407 |
from django.shortcuts import render
from push_notifications.api.rest_framework import GCMDeviceSerializer
from push_notifications.models import GCMDevice
from rest_framework import status
from rest_framework.authentication import TokenAuthentication
from rest_framework.generics import CreateAPIView, DestroyAPIView, Li... | app/notification/views.py | from django.shortcuts import render
from push_notifications.api.rest_framework import GCMDeviceSerializer
from push_notifications.models import GCMDevice
from rest_framework import status
from rest_framework.authentication import TokenAuthentication
from rest_framework.generics import CreateAPIView, DestroyAPIView, Li... | 0.561215 | 0.055618 |
from abc import abstractmethod
from ..radiosonde import Radiosonde
from ..radiosonde import RadiosondeList
class BaseSondeLoader:
"""Base loader interface for radiosonde to define common loading
interface
To use:
>>> loader = SondeLoader(filepath=path)
>>> loader.list_vars()
['U','UDir', '... | src/radiosonde/loader/base_loader.py | from abc import abstractmethod
from ..radiosonde import Radiosonde
from ..radiosonde import RadiosondeList
class BaseSondeLoader:
"""Base loader interface for radiosonde to define common loading
interface
To use:
>>> loader = SondeLoader(filepath=path)
>>> loader.list_vars()
['U','UDir', '... | 0.429429 | 0.476823 |
import asyncio
import aiohttp
from rest_framework import status
from presqt.targets.osf.utilities import OSFForbiddenError, OSFNotFoundError
from presqt.targets.utilities import get_page_total, run_urls_async
from presqt.utilities import PresQTResponseException
from presqt.targets.utilities.utils.session import PresQ... | presqt/targets/osf/classes/base.py | import asyncio
import aiohttp
from rest_framework import status
from presqt.targets.osf.utilities import OSFForbiddenError, OSFNotFoundError
from presqt.targets.utilities import get_page_total, run_urls_async
from presqt.utilities import PresQTResponseException
from presqt.targets.utilities.utils.session import PresQ... | 0.62395 | 0.175786 |
import random
from typing import List
from unittest.mock import Mock, mock_open, patch
import pytest
from hypothesis import given
from hypothesis.strategies import builds, integers, lists
from rplugin.python3.ultest.handler.finder import TestFinder
from rplugin.python3.ultest.models.test import Test
from tests.mocks.... | tests/unit/handler/test_finder.py | import random
from typing import List
from unittest.mock import Mock, mock_open, patch
import pytest
from hypothesis import given
from hypothesis.strategies import builds, integers, lists
from rplugin.python3.ultest.handler.finder import TestFinder
from rplugin.python3.ultest.models.test import Test
from tests.mocks.... | 0.730482 | 0.649516 |
import shutil
import subprocess
from os import path, getenv
import requests
from requests.exceptions import ConnectionError
def is_responsive(url):
"""Check if something responds to ``url``."""
try:
response = requests.get(url)
if response.status_code == 204:
return True
excep... | tests/test_integration.py | import shutil
import subprocess
from os import path, getenv
import requests
from requests.exceptions import ConnectionError
def is_responsive(url):
"""Check if something responds to ``url``."""
try:
response = requests.get(url)
if response.status_code == 204:
return True
excep... | 0.482673 | 0.229417 |
"""Bound on integer range."""
from typing import List, TYPE_CHECKING
from chb.invariants.FnDictionaryRecord import FnXprDictionaryRecord, xprregistry
from chb.invariants.XNumerical import XNumerical
import chb.util.fileutil as UF
from chb.util.IndexedTable import IndexedTableValue
if TYPE_CHECKING:
from chb.i... | chb/invariants/XBound.py | """Bound on integer range."""
from typing import List, TYPE_CHECKING
from chb.invariants.FnDictionaryRecord import FnXprDictionaryRecord, xprregistry
from chb.invariants.XNumerical import XNumerical
import chb.util.fileutil as UF
from chb.util.IndexedTable import IndexedTableValue
if TYPE_CHECKING:
from chb.i... | 0.934932 | 0.37502 |
from bs4 import BeautifulSoup
from dotenv import load_dotenv
from email.message import EmailMessage
from email.utils import make_msgid
import chevron
import io
import json
import logging
import math
import os
import requests
import smtplib
import sys
import traceback
BASE_URL = "https://www.hasznaltauto.hu"
PAGE_SIZE... | hahu/main.py | from bs4 import BeautifulSoup
from dotenv import load_dotenv
from email.message import EmailMessage
from email.utils import make_msgid
import chevron
import io
import json
import logging
import math
import os
import requests
import smtplib
import sys
import traceback
BASE_URL = "https://www.hasznaltauto.hu"
PAGE_SIZE... | 0.209712 | 0.088072 |
import numpy as np
from ... import opcodes as OperandDef
from ...core import TilesError
from ...serialize import KeyField, BoolField
from ...utils import check_chunks_unknown_shape
from ..operands import TensorOperand, TensorOperandMixin
from ..datasource import tensor as astensor
from ..array_utils import as_same_de... | mars/tensor/base/isin.py |
import numpy as np
from ... import opcodes as OperandDef
from ...core import TilesError
from ...serialize import KeyField, BoolField
from ...utils import check_chunks_unknown_shape
from ..operands import TensorOperand, TensorOperandMixin
from ..datasource import tensor as astensor
from ..array_utils import as_same_de... | 0.878731 | 0.581184 |
import itertools
from fractions import Fraction
from io import BytesIO
from typing import Callable, Tuple
import cairosvg
import filetype
from PIL import Image, ImageSequence
from streamdeck_ui.display.filter import Filter
class ImageFilter(Filter):
"""
Represents a static image. It transforms the input ima... | streamdeck_ui/display/image_filter.py | import itertools
from fractions import Fraction
from io import BytesIO
from typing import Callable, Tuple
import cairosvg
import filetype
from PIL import Image, ImageSequence
from streamdeck_ui.display.filter import Filter
class ImageFilter(Filter):
"""
Represents a static image. It transforms the input ima... | 0.705176 | 0.264486 |
from .util import UrsadbTestContext, store_files, check_query, get_index_hash, UrsadbConfig
from .util import ursadb # noqa
import pytest
def test_indexing_small(ursadb: UrsadbTestContext):
store_files(ursadb, "gram3", {"kot": b"Ala ma kota ale czy kot ma Ale?"})
ursadb.check_request(
"topology;",
... | teste2e/test_indexing.py | from .util import UrsadbTestContext, store_files, check_query, get_index_hash, UrsadbConfig
from .util import ursadb # noqa
import pytest
def test_indexing_small(ursadb: UrsadbTestContext):
store_files(ursadb, "gram3", {"kot": b"Ala ma kota ale czy kot ma Ale?"})
ursadb.check_request(
"topology;",
... | 0.374562 | 0.47591 |
from jqdatapy.api import run_query
from zvt.contract.recorder import TimeSeriesDataRecorder
from zvt.domain import Index
from zvt.domain import StockSummary
from zvt.utils.time_utils import to_time_str
from zvt.utils.utils import multiple_number
# 聚宽编码
# 322001 上海市场
# 322002 上海A股
# 322003 上海B股
# 322004 深圳市场 该市场交易所未公布... | zvt/recorders/joinquant/overall/jq_stock_summary_recorder.py | from jqdatapy.api import run_query
from zvt.contract.recorder import TimeSeriesDataRecorder
from zvt.domain import Index
from zvt.domain import StockSummary
from zvt.utils.time_utils import to_time_str
from zvt.utils.utils import multiple_number
# 聚宽编码
# 322001 上海市场
# 322002 上海A股
# 322003 上海B股
# 322004 深圳市场 该市场交易所未公布... | 0.416085 | 0.169475 |
import cgi
import datetime
import time
from tempfile import NamedTemporaryFile
from fabric.api import *
from fabric import colors
@task
def update():
"""Requires code_root env variable. Does a git pull and restarts the web server"""
require('code_root')
git_pull()
restart_web_server()
@task
def g... | fabric_bolt/fabfile.py | import cgi
import datetime
import time
from tempfile import NamedTemporaryFile
from fabric.api import *
from fabric import colors
@task
def update():
"""Requires code_root env variable. Does a git pull and restarts the web server"""
require('code_root')
git_pull()
restart_web_server()
@task
def g... | 0.293607 | 0.158826 |
from __future__ import annotations
import asyncio
from typing import TYPE_CHECKING
from warnings import warn
from sanic.exceptions import SanicException
if TYPE_CHECKING:
from sanic import Sanic
class AsyncioServer:
"""
Wraps an asyncio server with functionality that might be useful to
a user who... | sanic/server/async_server.py | from __future__ import annotations
import asyncio
from typing import TYPE_CHECKING
from warnings import warn
from sanic.exceptions import SanicException
if TYPE_CHECKING:
from sanic import Sanic
class AsyncioServer:
"""
Wraps an asyncio server with functionality that might be useful to
a user who... | 0.822153 | 0.139602 |
from more_or_less import PageOfHeight
from more_or_less.fixed_size_screen import FixedSizeScreen
from more_or_less.input import Input
from more_or_less.more_page_builder import MorePageBuilder
from more_or_less.output import Output
from more_or_less.page_builder import StopOutput
from more_or_less.wrapped_page import W... | tests/test_more_page_builder.py | from more_or_less import PageOfHeight
from more_or_less.fixed_size_screen import FixedSizeScreen
from more_or_less.input import Input
from more_or_less.more_page_builder import MorePageBuilder
from more_or_less.output import Output
from more_or_less.page_builder import StopOutput
from more_or_less.wrapped_page import W... | 0.716615 | 0.312737 |
import os
import numpy as np
import pandas as pd
from keras.callbacks import ModelCheckpoint, TensorBoard, EarlyStopping, ReduceLROnPlateau
from keras.layers import Conv2D, Concatenate, MaxPooling2D, Conv2DTranspose, UpSampling2D, Dropout, BatchNormalization
from keras.models import Input, Model
from keras.optimizers ... | img_segmentation/model.py | import os
import numpy as np
import pandas as pd
from keras.callbacks import ModelCheckpoint, TensorBoard, EarlyStopping, ReduceLROnPlateau
from keras.layers import Conv2D, Concatenate, MaxPooling2D, Conv2DTranspose, UpSampling2D, Dropout, BatchNormalization
from keras.models import Input, Model
from keras.optimizers ... | 0.897907 | 0.419826 |
from __future__ import unicode_literals
import argparse
import json
import sys
import time
from googleapiclient import discovery
from googleapiclient import errors as apierrors
#pylint: disable=no-member
class SlaveManager(object):
"""Class for managing Jenkins Slaves."""
DEFAULT_SCOPES = ['https://www.google... | config/jenkins/start_slave.py |
from __future__ import unicode_literals
import argparse
import json
import sys
import time
from googleapiclient import discovery
from googleapiclient import errors as apierrors
#pylint: disable=no-member
class SlaveManager(object):
"""Class for managing Jenkins Slaves."""
DEFAULT_SCOPES = ['https://www.google... | 0.639511 | 0.179315 |
from PyQt4.QtCore import *
from PyQt4.QtGui import *
from froi.algorithm.imtool import merge
class ROIMergeDialog(QDialog):
"""A dialog for ROI selection and merging."""
def __init__(self, model, parent=None):
super(ROIMergeDialog, self).__init__(parent)
self._model = model
... | froi/gui/component/roimergedialog.py |
from PyQt4.QtCore import *
from PyQt4.QtGui import *
from froi.algorithm.imtool import merge
class ROIMergeDialog(QDialog):
"""A dialog for ROI selection and merging."""
def __init__(self, model, parent=None):
super(ROIMergeDialog, self).__init__(parent)
self._model = model
... | 0.527073 | 0.08061 |
from datetime import datetime
from six.moves import http_client
from django.core.urlresolvers import reverse
from freezegun import freeze_time
from common.test_utils import CassandraTestCase
from common.models import CassandraThingMultiplePK
@freeze_time('14-06-15 15:44:25')
def create_thing():
return Cassandra... | testproject/common/tests/test_views.py | from datetime import datetime
from six.moves import http_client
from django.core.urlresolvers import reverse
from freezegun import freeze_time
from common.test_utils import CassandraTestCase
from common.models import CassandraThingMultiplePK
@freeze_time('14-06-15 15:44:25')
def create_thing():
return Cassandra... | 0.613237 | 0.167355 |
import numpy as np
import torch
from torch.nn import BCEWithLogitsLoss as _BCEWithLogitsLoss
# pylint: disable=too-few-public-methods
class BCELoss:
"""
Applies a BCE Loss function to the model.
BCE Loss automatically applies a Sigmoid Layer
at the end of the model, so there is no need ... | neuralpy/loss_functions/bce_loss.py |
import numpy as np
import torch
from torch.nn import BCEWithLogitsLoss as _BCEWithLogitsLoss
# pylint: disable=too-few-public-methods
class BCELoss:
"""
Applies a BCE Loss function to the model.
BCE Loss automatically applies a Sigmoid Layer
at the end of the model, so there is no need ... | 0.915835 | 0.502563 |
"""Command for deleting a service."""
from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
from googlecloudsdk.calliope import base
from googlecloudsdk.command_lib.events import eventflow_operations
from googlecloudsdk.command_lib.events import exceptions
from... | google-cloud-sdk/lib/surface/events/triggers/delete.py | """Command for deleting a service."""
from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
from googlecloudsdk.calliope import base
from googlecloudsdk.command_lib.events import eventflow_operations
from googlecloudsdk.command_lib.events import exceptions
from... | 0.572603 | 0.088939 |
from redditimagespider.items import RedditImageFileItem
import scrapy
import json
class RedditSpider(scrapy.Spider):
name = 'reddit-spider'
start_urls = ["https://gateway.reddit.com/desktopapi/v1/subreddits/gifs?sort=new&allow_over18=1"]
page_limit = 10
i = 0
def parse(self, response):
... | redditimagespider/redditimagespider/spiders/redditspider.py | from redditimagespider.items import RedditImageFileItem
import scrapy
import json
class RedditSpider(scrapy.Spider):
name = 'reddit-spider'
start_urls = ["https://gateway.reddit.com/desktopapi/v1/subreddits/gifs?sort=new&allow_over18=1"]
page_limit = 10
i = 0
def parse(self, response):
... | 0.209793 | 0.129458 |
# setup the paths
from opentamiltests import *
import tamil.utf8 as utf8
from tamil.tscii import TSCII
import codecs
if PYTHON3:
class long(int):
pass
class NumeralStringLimitTests(unittest.TestCase):
def test_case_basic(self):
self.assertEqual(u"புள்ளி மூன்று மூன்று",tamil.nume... | tests/numeral_basic.py |
# setup the paths
from opentamiltests import *
import tamil.utf8 as utf8
from tamil.tscii import TSCII
import codecs
if PYTHON3:
class long(int):
pass
class NumeralStringLimitTests(unittest.TestCase):
def test_case_basic(self):
self.assertEqual(u"புள்ளி மூன்று மூன்று",tamil.nume... | 0.21892 | 0.336467 |
import json
import os
from pathlib import Path
from typing import List
from canvasxpress.util.example.generator import \
generate_canvasxpress_code_from_json_file
JSON_DIR_PATH = f"{os.getcwd()}/../../../tutorials/reproducible_json/"
JUPYTER_TEMPLATE_PATH = f"{os.getcwd()}/../../../canvasxpress/util/" \
... | canvasxpress/util/example/generate_tutorials.py | import json
import os
from pathlib import Path
from typing import List
from canvasxpress.util.example.generator import \
generate_canvasxpress_code_from_json_file
JSON_DIR_PATH = f"{os.getcwd()}/../../../tutorials/reproducible_json/"
JUPYTER_TEMPLATE_PATH = f"{os.getcwd()}/../../../canvasxpress/util/" \
... | 0.713531 | 0.357876 |
from __future__ import unicode_literals
from django.conf import settings
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
initial = True
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
]
operations ... | Campaign/migrations/0001_initial.py | from __future__ import unicode_literals
from django.conf import settings
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
initial = True
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
]
operations ... | 0.558327 | 0.11353 |
class SearchAlgorithm:
"""Interface of an event handler API for hyperparameter search.
Unlike TrialSchedulers, SearchAlgorithms will not have the ability
to modify the execution (i.e., stop and pause trials).
Trials added manually (i.e., via the Client API) will also notify
this class upon new eve... | python/ray/tune/suggest/search.py | class SearchAlgorithm:
"""Interface of an event handler API for hyperparameter search.
Unlike TrialSchedulers, SearchAlgorithms will not have the ability
to modify the execution (i.e., stop and pause trials).
Trials added manually (i.e., via the Client API) will also notify
this class upon new eve... | 0.891687 | 0.668218 |
from testlib.custom_exceptions import UICmdException
from testlib.linux import service_lib
from testlib.ui_onpss_shell.switch_driver import SwitchDriver
class Dcrpd(object):
SERVICE = 'dcrpd'
CONFIG_PATH = "/usr/lib/systemd/system/"
MANIFEST_FILE = CONFIG_PATH + "dcrpd.service"
def __init__(self, ru... | taf/testlib/linux/dcrpd/dcrpd.py | from testlib.custom_exceptions import UICmdException
from testlib.linux import service_lib
from testlib.ui_onpss_shell.switch_driver import SwitchDriver
class Dcrpd(object):
SERVICE = 'dcrpd'
CONFIG_PATH = "/usr/lib/systemd/system/"
MANIFEST_FILE = CONFIG_PATH + "dcrpd.service"
def __init__(self, ru... | 0.573678 | 0.126839 |
import cv2
import os
#Creating/Checking a Database Dir
if os.path.exists('database'):
pass
else:
os.mkdir('database')
#Reseting Counter and Users
dir = os.listdir('database')
if len(dir) == 0:
users = open('resources/user.txt','w')
ids = open('resources/count.txt','w')
users.write('N... | Recognition/database.py | import cv2
import os
#Creating/Checking a Database Dir
if os.path.exists('database'):
pass
else:
os.mkdir('database')
#Reseting Counter and Users
dir = os.listdir('database')
if len(dir) == 0:
users = open('resources/user.txt','w')
ids = open('resources/count.txt','w')
users.write('N... | 0.10917 | 0.075585 |
import pytest
from plenum.common.constants import STEWARD_STRING, VALIDATOR
from pytest import fixture
from plenum.common.throughput_measurements import RevivalSpikeResistantEMAThroughputMeasurement
from plenum.common.util import getMaxFailures
from plenum.test.helper import sdk_send_random_and_check, assertExp, sdk_g... | plenum/test/replica/test_catchup_after_replica_addition.py | import pytest
from plenum.common.constants import STEWARD_STRING, VALIDATOR
from pytest import fixture
from plenum.common.throughput_measurements import RevivalSpikeResistantEMAThroughputMeasurement
from plenum.common.util import getMaxFailures
from plenum.test.helper import sdk_send_random_and_check, assertExp, sdk_g... | 0.313525 | 0.35209 |
import logging
import textwrap
from datetime import datetime, timedelta
from airflow import DAG # noqa
from airflow import macros # noqa
from airflow.operators.python_operator import PythonOperator # noqa
from pyhocon import ConfigFactory
from databuilder.extractor.hive_table_metadata_extractor import HiveTableMet... | example/dags/sample_dag.py | import logging
import textwrap
from datetime import datetime, timedelta
from airflow import DAG # noqa
from airflow import macros # noqa
from airflow.operators.python_operator import PythonOperator # noqa
from pyhocon import ConfigFactory
from databuilder.extractor.hive_table_metadata_extractor import HiveTableMet... | 0.38122 | 0.082475 |
import numpy as np
from rlgym.utils import RewardFunction
from rlgym.utils.common_values import CEILING_Z, BALL_MAX_SPEED, CAR_MAX_SPEED, BLUE_TEAM, BLUE_GOAL_BACK, \
BLUE_GOAL_CENTER, ORANGE_GOAL_BACK, ORANGE_GOAL_CENTER, BALL_RADIUS, ORANGE_TEAM
from rlgym.utils.gamestates import GameState, PlayerData
from rlgym.... | training/reward.py | import numpy as np
from rlgym.utils import RewardFunction
from rlgym.utils.common_values import CEILING_Z, BALL_MAX_SPEED, CAR_MAX_SPEED, BLUE_TEAM, BLUE_GOAL_BACK, \
BLUE_GOAL_CENTER, ORANGE_GOAL_BACK, ORANGE_GOAL_CENTER, BALL_RADIUS, ORANGE_TEAM
from rlgym.utils.gamestates import GameState, PlayerData
from rlgym.... | 0.444565 | 0.284781 |
from pathlib import Path
import pytest
import xlwings as xw
this_dir = Path(__file__).resolve().parent
@pytest.fixture(scope="module")
def app():
with xw.App(visible=False) as app:
yield app
for f in Path(".").glob("tempfile*"):
f.unlink()
for f in Path("temp").glob("tempfile*"):
... | tests/test_fileformats.py | from pathlib import Path
import pytest
import xlwings as xw
this_dir = Path(__file__).resolve().parent
@pytest.fixture(scope="module")
def app():
with xw.App(visible=False) as app:
yield app
for f in Path(".").glob("tempfile*"):
f.unlink()
for f in Path("temp").glob("tempfile*"):
... | 0.226784 | 0.53868 |
from sklearn.linear_model.stochastic_gradient import SGDClassifier, SGDRegressor
from sklearn.linear_model.passive_aggressive import PassiveAggressiveClassifier
from sklearn.linear_model.perceptron import Perceptron
from skmultiflow.classification.perceptron import PerceptronMask
from skmultiflow.classification.lazy.kn... | src/skmultiflow/demos/_test_prequential.py | from sklearn.linear_model.stochastic_gradient import SGDClassifier, SGDRegressor
from sklearn.linear_model.passive_aggressive import PassiveAggressiveClassifier
from sklearn.linear_model.perceptron import Perceptron
from skmultiflow.classification.perceptron import PerceptronMask
from skmultiflow.classification.lazy.kn... | 0.825238 | 0.325534 |
from __future__ import unicode_literals
import codecs
import os
import sys
import re
sys.path.append(os.path.abspath(os.path.join(__file__, os.pardir, os.pardir, 'DropPy.Common')))
from file_tools import get_file_paths_from_directory
H1_SETEX_STYLE_REGEX = re.compile(r'^-+$')
H2_SETEX_STYLE_REGEX = re.compile(r'^=+... | Tasks/Markdown.AddToc/task.py |
from __future__ import unicode_literals
import codecs
import os
import sys
import re
sys.path.append(os.path.abspath(os.path.join(__file__, os.pardir, os.pardir, 'DropPy.Common')))
from file_tools import get_file_paths_from_directory
H1_SETEX_STYLE_REGEX = re.compile(r'^-+$')
H2_SETEX_STYLE_REGEX = re.compile(r'^=+... | 0.466846 | 0.273911 |
from ctypes import CFUNCTYPE, c_void_p, c_char_p
from objc_util import retain_global, ObjCInstance, UIApplication, c, ns, on_main_thread, sel, ObjCClass
from blackmamba.util.runtime import swizzle
from blackmamba.log import error, info
import blackmamba.system as system
from enum import Enum, IntEnum
from typing import... | blackmamba/uikit/keyboard.py | from ctypes import CFUNCTYPE, c_void_p, c_char_p
from objc_util import retain_global, ObjCInstance, UIApplication, c, ns, on_main_thread, sel, ObjCClass
from blackmamba.util.runtime import swizzle
from blackmamba.log import error, info
import blackmamba.system as system
from enum import Enum, IntEnum
from typing import... | 0.737725 | 0.150778 |
import miepy
import numpy as np
from .get_tmatrix import nfmds_solver, tmatrix_solvers
def tmatrix_sphere(radius, wavelength, eps, eps_m, lmax, conducting=False):
"""Compute the T-matrix of a sphere, using regular Mie theory
Arguments:
radius sphere radius
wavelength incident wavelength
... | miepy/tmatrix/common.py | import miepy
import numpy as np
from .get_tmatrix import nfmds_solver, tmatrix_solvers
def tmatrix_sphere(radius, wavelength, eps, eps_m, lmax, conducting=False):
"""Compute the T-matrix of a sphere, using regular Mie theory
Arguments:
radius sphere radius
wavelength incident wavelength
... | 0.87749 | 0.458531 |