code stringlengths 2k 1.04M | repo_path stringlengths 5 517 | parsed_code stringlengths 0 1.04M | quality_prob float64 0.02 0.95 | learning_prob float64 0.02 0.93 |
|---|---|---|---|---|
from oci.util import formatted_flat_dict, NONE_SENTINEL, value_allowed_none_or_none_sentinel # noqa: F401
from oci.decorators import init_model_state_from_kwargs
@init_model_state_from_kwargs
class Histogram(object):
"""
To capture all the histograms data related to profiling
"""
def __init__(self... | src/oci/data_connectivity/models/histogram.py |
from oci.util import formatted_flat_dict, NONE_SENTINEL, value_allowed_none_or_none_sentinel # noqa: F401
from oci.decorators import init_model_state_from_kwargs
@init_model_state_from_kwargs
class Histogram(object):
"""
To capture all the histograms data related to profiling
"""
def __init__(self... | 0.851027 | 0.430806 |
from ...flash.flash import Flash
from ...core.coresight_target import CoreSightTarget
from ...core.memory_map import (FlashRegion, RamRegion, MemoryMap)
from ...debug.svd.loader import SVDFile
import logging
FLASH_ALGO = { 'load_address' : 0x20000000,
'instructions' : [
0xE00ABE00, 0x062D780D, 0x24... | pyocd/target/builtin/target_MAX32630.py |
from ...flash.flash import Flash
from ...core.coresight_target import CoreSightTarget
from ...core.memory_map import (FlashRegion, RamRegion, MemoryMap)
from ...debug.svd.loader import SVDFile
import logging
FLASH_ALGO = { 'load_address' : 0x20000000,
'instructions' : [
0xE00ABE00, 0x062D780D, 0x24... | 0.39222 | 0.311126 |
"""Classes and functions related to Quality Control of incoming data."""
# Python imports
from __future__ import absolute_import
import logging
# weewx imports
import weeutil.weeutil
import weewx.units
from weeutil.weeutil import to_float
log = logging.getLogger(__name__)
# ========================================... | dist/weewx-4.6.0b7/bin/weewx/qc.py | """Classes and functions related to Quality Control of incoming data."""
# Python imports
from __future__ import absolute_import
import logging
# weewx imports
import weeutil.weeutil
import weewx.units
from weeutil.weeutil import to_float
log = logging.getLogger(__name__)
# ========================================... | 0.906091 | 0.650939 |
import re
from pygments.lexer import RegexLexer, bygroups, default
from pygments.token import Keyword, Punctuation, String, Number, Operator, \
Whitespace, Name, Literal, Comment, Text
__all__ = ['SparqlLexer']
class SparqlLexer(RegexLexer):
"""
Lexer for `SPARQL <http://www.w3.org/TR/rdf-sparql-query/>... | OmniMarkupLib/Renderers/libs/pygments/lexers/rdf.py | import re
from pygments.lexer import RegexLexer, bygroups, default
from pygments.token import Keyword, Punctuation, String, Number, Operator, \
Whitespace, Name, Literal, Comment, Text
__all__ = ['SparqlLexer']
class SparqlLexer(RegexLexer):
"""
Lexer for `SPARQL <http://www.w3.org/TR/rdf-sparql-query/>... | 0.460532 | 0.336481 |
from functools import partial
from typing import Iterator, List, Optional, Mapping
from preacher.compilation.argument import Arguments, inject_arguments
from preacher.compilation.error import on_key
from preacher.compilation.parameter import Parameter, compile_parameter
from preacher.compilation.util.functional impor... | preacher/compilation/scenario/scenario.py |
from functools import partial
from typing import Iterator, List, Optional, Mapping
from preacher.compilation.argument import Arguments, inject_arguments
from preacher.compilation.error import on_key
from preacher.compilation.parameter import Parameter, compile_parameter
from preacher.compilation.util.functional impor... | 0.914565 | 0.238018 |
import logging
from horizon import exceptions
from horizon import forms
from horizon import messages
from django.core.urlresolvers import reverse_lazy
from django.utils.translation import ugettext_lazy as _
from openstack_dashboard.dashboards.cdn.cdn_domain_manager.models import Domain, CdnBillMethod
import uuid
impor... | horizon/openstack_dashboard/dashboards/cdn/cdn_domain_manager/forms.py |
import logging
from horizon import exceptions
from horizon import forms
from horizon import messages
from django.core.urlresolvers import reverse_lazy
from django.utils.translation import ugettext_lazy as _
from openstack_dashboard.dashboards.cdn.cdn_domain_manager.models import Domain, CdnBillMethod
import uuid
impor... | 0.354768 | 0.063599 |
import pathlib
import typing
from os import environ as env
from typing import Any, Dict, Mapping, TypeVar
from urllib.parse import parse_qs, urlencode
import ipywidgets as w
import weldx
import weldx_widgets
import weldx_widgets.widget_base
import weldx_widgets.widget_factory
from weldx_widgets.translation_utils impo... | weldx_widgets/kisa/save.py | import pathlib
import typing
from os import environ as env
from typing import Any, Dict, Mapping, TypeVar
from urllib.parse import parse_qs, urlencode
import ipywidgets as w
import weldx
import weldx_widgets
import weldx_widgets.widget_base
import weldx_widgets.widget_factory
from weldx_widgets.translation_utils impo... | 0.507812 | 0.173989 |
from datetime import datetime
from json import load
from requests import get, head, post, put
from urllib.parse import quote_plus
from uuid import uuid4
class DaoElastic(object):
_INDEX = 'unfact'
_TYPE = 'news'
_MAX_FETCH_SIZE = 300
_BASE_HOST = 'http://127.0.0.1:9200'
_MAPPING_FILE = '../resourc... | python/scripts/dao_elastic.py | from datetime import datetime
from json import load
from requests import get, head, post, put
from urllib.parse import quote_plus
from uuid import uuid4
class DaoElastic(object):
_INDEX = 'unfact'
_TYPE = 'news'
_MAX_FETCH_SIZE = 300
_BASE_HOST = 'http://127.0.0.1:9200'
_MAPPING_FILE = '../resourc... | 0.640748 | 0.168515 |
number_string = """08 02 22 97 38 15 00 40 00 75 04 05 07 78 52 12 50 77 91 08
49 49 99 40 17 81 18 57 60 87 17 40 98 43 69 48 04 56 62 00
81 49 31 73 55 79 14 29 93 71 40 67 53 88 30 03 49 13 36 65
52 70 95 23 04 60 11 42 69 24 68 56 01 32 56 71 37 02 36 91
22 31 16 71 51 67 63 89 41 92 36 54 22 40 40 28 66 33 13 80
2... | python/problem11.py | number_string = """08 02 22 97 38 15 00 40 00 75 04 05 07 78 52 12 50 77 91 08
49 49 99 40 17 81 18 57 60 87 17 40 98 43 69 48 04 56 62 00
81 49 31 73 55 79 14 29 93 71 40 67 53 88 30 03 49 13 36 65
52 70 95 23 04 60 11 42 69 24 68 56 01 32 56 71 37 02 36 91
22 31 16 71 51 67 63 89 41 92 36 54 22 40 40 28 66 33 13 80
2... | 0.362066 | 0.335569 |
from __future__ import absolute_import
from __future__ import print_function
from __future__ import division
from six.moves import xrange
import logging
import numpy as np
from keras import backend as K
from keras import optimizers
from keras import objectives
from keras.layers import Input, Concatenate, MaxPooling1... | hyperion/keras/embed/seq_embed.py | from __future__ import absolute_import
from __future__ import print_function
from __future__ import division
from six.moves import xrange
import logging
import numpy as np
from keras import backend as K
from keras import optimizers
from keras import objectives
from keras.layers import Input, Concatenate, MaxPooling1... | 0.808029 | 0.176193 |
import torch
import torch.nn as nn
import torch.nn.functional as F
import torch.optim as optim
import utility
from utils.loss import discriminator
class Adversarial(nn.Module):
def __init__(self, args, gan_type):
super(Adversarial, self).__init__()
self.gan_type = gan_type
self.gan_k = arg... | utils/loss/adversarial.py | import torch
import torch.nn as nn
import torch.nn.functional as F
import torch.optim as optim
import utility
from utils.loss import discriminator
class Adversarial(nn.Module):
def __init__(self, args, gan_type):
super(Adversarial, self).__init__()
self.gan_type = gan_type
self.gan_k = arg... | 0.930616 | 0.292861 |
from glue.viewers.image.composite_array import CompositeArray
from bqplot_image_gl.viewlistener import ViewListener
from ...link import on_change
from ..common.viewer import BqplotBaseView
from ..scatter.layer_artist import BqplotScatterLayerArtist
from .layer_artist import BqplotImageLayerArtist, BqplotImageSubsetL... | glue_jupyter/bqplot/image/viewer.py | from glue.viewers.image.composite_array import CompositeArray
from bqplot_image_gl.viewlistener import ViewListener
from ...link import on_change
from ..common.viewer import BqplotBaseView
from ..scatter.layer_artist import BqplotScatterLayerArtist
from .layer_artist import BqplotImageLayerArtist, BqplotImageSubsetL... | 0.647241 | 0.354852 |
"""Command for creating target HTTP proxies."""
from googlecloudapis.compute.v1 import compute_v1_messages as messages
from googlecloudsdk.compute.lib import base_classes
class Create(base_classes.BaseAsyncMutator):
"""Create a target HTTP proxy."""
@staticmethod
def Args(parser):
parser.add_argument(
... | lib/googlecloudsdk/compute/subcommands/target_http_proxies/create.py | """Command for creating target HTTP proxies."""
from googlecloudapis.compute.v1 import compute_v1_messages as messages
from googlecloudsdk.compute.lib import base_classes
class Create(base_classes.BaseAsyncMutator):
"""Create a target HTTP proxy."""
@staticmethod
def Args(parser):
parser.add_argument(
... | 0.881341 | 0.16099 |
from __future__ import print_function
import time
import numpy as np
import sympy as sy
from bokeh.browserlib import view
from bokeh.document import Document
from bokeh.glyphs import Line
from bokeh.objects import Plot, DataRange1d, LinearAxis, ColumnDataSource, Grid, Legend
from bokeh.session import Session
from bo... | examples/glyphs/taylor_server.py | from __future__ import print_function
import time
import numpy as np
import sympy as sy
from bokeh.browserlib import view
from bokeh.document import Document
from bokeh.glyphs import Line
from bokeh.objects import Plot, DataRange1d, LinearAxis, ColumnDataSource, Grid, Legend
from bokeh.session import Session
from bo... | 0.57081 | 0.381969 |
import io
import sys
import pyxb
import pyxb.binding
# Import bindings for namespaces imported into schema
import pyxb.binding.datatypes
import pyxb.binding.saxer
import pyxb.utils.domutils
import pyxb.utils.six
import pyxb.utils.utility
from . import dataoneTypes_v1 as _ImportedBinding_dataoneTypes_v1
# Unique id... | lib_common/src/d1_common/types/generated/dataoneTypes_v1_1.py |
import io
import sys
import pyxb
import pyxb.binding
# Import bindings for namespaces imported into schema
import pyxb.binding.datatypes
import pyxb.binding.saxer
import pyxb.utils.domutils
import pyxb.utils.six
import pyxb.utils.utility
from . import dataoneTypes_v1 as _ImportedBinding_dataoneTypes_v1
# Unique id... | 0.551332 | 0.231593 |
import sys
import time
import curses
import argparse
import httplib2
import _thread
import colorama
from bs4 import BeautifulSoup
from urllib.parse import urljoin, urlparse
from threading import Thread
from colorama import Fore, Back, Style
class HttpRequest(Thread):
stop = False
request_de... | stress_http_server.py |
import sys
import time
import curses
import argparse
import httplib2
import _thread
import colorama
from bs4 import BeautifulSoup
from urllib.parse import urljoin, urlparse
from threading import Thread
from colorama import Fore, Back, Style
class HttpRequest(Thread):
stop = False
request_de... | 0.179638 | 0.044974 |
from datetime import datetime
from sqlalchemy.orm import sessionmaker
from sqlalchemy.ext.declarative import declarative_base
from sqlalchemy.ext.hybrid import hybrid_method, hybrid_property
from sqlalchemy import create_engine, UniqueConstraint, desc, Index
from sqlalchemy import Column, Integer, String, Date, Time, ... | data_alchemy/models.py | from datetime import datetime
from sqlalchemy.orm import sessionmaker
from sqlalchemy.ext.declarative import declarative_base
from sqlalchemy.ext.hybrid import hybrid_method, hybrid_property
from sqlalchemy import create_engine, UniqueConstraint, desc, Index
from sqlalchemy import Column, Integer, String, Date, Time, ... | 0.306423 | 0.290292 |
import asyncio
import logging
from abc import ABC, abstractmethod
from typing import Optional, Awaitable, Tuple, Union, Any, TYPE_CHECKING
import grpc
from google.protobuf import empty_pb2
from . import rpc
from .settings import Settings, configure, get_stack, get_project, get_root_resource
from .sync_await import _sy... | sdk/python/lib/pulumi/runtime/mocks.py | import asyncio
import logging
from abc import ABC, abstractmethod
from typing import Optional, Awaitable, Tuple, Union, Any, TYPE_CHECKING
import grpc
from google.protobuf import empty_pb2
from . import rpc
from .settings import Settings, configure, get_stack, get_project, get_root_resource
from .sync_await import _sy... | 0.827375 | 0.213869 |
from collections import namedtuple
from dataclasses import dataclass
from utils import to_form_url
@dataclass
class EntryInfo:
required: bool
prompt: bool
type: str
key: str
title: str
value: str
# See README's Config section for more info
TYPES = {
"words": ["w", "word", "te... | config.py |
from collections import namedtuple
from dataclasses import dataclass
from utils import to_form_url
@dataclass
class EntryInfo:
required: bool
prompt: bool
type: str
key: str
title: str
value: str
# See README's Config section for more info
TYPES = {
"words": ["w", "word", "te... | 0.751101 | 0.316079 |
import argparse
import asyncio
import fcntl
import json
import logging
import os
import pty
import shlex
import signal
import struct
import sys
import termios
import traceback
import zmq, zmq.asyncio
from .compat import current_loop
from .logging import BraceStyleAdapter
from .utils import safe_close_task
log = Brac... | src/ai/backend/kernel/terminal.py | import argparse
import asyncio
import fcntl
import json
import logging
import os
import pty
import shlex
import signal
import struct
import sys
import termios
import traceback
import zmq, zmq.asyncio
from .compat import current_loop
from .logging import BraceStyleAdapter
from .utils import safe_close_task
log = Brac... | 0.218669 | 0.072276 |
from consolemenu import ConsoleMenu
from consolemenu.items import SelectionItem
class SelectionMenu(ConsoleMenu):
"""
A menu that simplifies item creation, just give it a list of strings and it builds the menu for you
Args:
strings (:obj:`list` of :obj:`str`): The list of strings this menu shoul... | consolemenu/selection_menu.py | from consolemenu import ConsoleMenu
from consolemenu.items import SelectionItem
class SelectionMenu(ConsoleMenu):
"""
A menu that simplifies item creation, just give it a list of strings and it builds the menu for you
Args:
strings (:obj:`list` of :obj:`str`): The list of strings this menu shoul... | 0.549882 | 0.09899 |
import os
import sys
import re
import subprocess
import plistlib
import shutil
CLT_BINARY = os.path.dirname(os.path.realpath(__file__)) + '/platypus'
def profile_plist_for_args(args):
pnargs = [CLT_BINARY]
pnargs.extend(args)
pnargs.extend(['-O', '-'])
out = subprocess.check_output(pnargs)
return... | Tests/clt_tests.py |
import os
import sys
import re
import subprocess
import plistlib
import shutil
CLT_BINARY = os.path.dirname(os.path.realpath(__file__)) + '/platypus'
def profile_plist_for_args(args):
pnargs = [CLT_BINARY]
pnargs.extend(args)
pnargs.extend(['-O', '-'])
out = subprocess.check_output(pnargs)
return... | 0.151686 | 0.187411 |
import os
import matplotlib.pyplot as plt
import numpy as np
import plotly.express as px
import pandas as pd
import seaborn as sns
from dotenv import find_dotenv, load_dotenv
from IPython.core.interactiveshell import InteractiveShell
# Setting styles
InteractiveShell.ast_node_interactivity = "all"
sns.set(style="whi... | notebooks/data_exploraton.py |
import os
import matplotlib.pyplot as plt
import numpy as np
import plotly.express as px
import pandas as pd
import seaborn as sns
from dotenv import find_dotenv, load_dotenv
from IPython.core.interactiveshell import InteractiveShell
# Setting styles
InteractiveShell.ast_node_interactivity = "all"
sns.set(style="whi... | 0.474875 | 0.448185 |
from __future__ import division
from past.builtins import cmp
from future import standard_library
standard_library.install_aliases()
from builtins import object
import errno
import logging
import math
import os
import posixpath
import random
import subprocess
import sys
from django.utils.encoding import smart_str
fro... | desktop/libs/hadoop/src/hadoop/fs/hadoopfs.py | from __future__ import division
from past.builtins import cmp
from future import standard_library
standard_library.install_aliases()
from builtins import object
import errno
import logging
import math
import os
import posixpath
import random
import subprocess
import sys
from django.utils.encoding import smart_str
fro... | 0.361503 | 0.094887 |
import asyncio, base64, discord, requests, time, traceback
from utils.datautils import config, set_client
from utils.discordbot import BotClient, send
client = None
class TimerClient(BotClient):
def __init__(self):
BotClient.__init__(self, "")
self.name = "timer"
client = TimerClient()
timers = {}
@clie... | src/bots/timer/main.py | import asyncio, base64, discord, requests, time, traceback
from utils.datautils import config, set_client
from utils.discordbot import BotClient, send
client = None
class TimerClient(BotClient):
def __init__(self):
BotClient.__init__(self, "")
self.name = "timer"
client = TimerClient()
timers = {}
@clie... | 0.188847 | 0.058426 |
from dataclasses import dataclass
from typing import Final
from jupiter.domain.adate import ADate
from jupiter.domain.entity_name import EntityName
from jupiter.domain.storage_engine import StorageEngine
from jupiter.domain.vacations.infra.vacation_notion_manager import VacationNotionManager
from jupiter.framework.bas... | jupiter/use_cases/vacations/update.py | from dataclasses import dataclass
from typing import Final
from jupiter.domain.adate import ADate
from jupiter.domain.entity_name import EntityName
from jupiter.domain.storage_engine import StorageEngine
from jupiter.domain.vacations.infra.vacation_notion_manager import VacationNotionManager
from jupiter.framework.bas... | 0.904068 | 0.123736 |
from django.shortcuts import render
from django.http import JsonResponse
import os
import json
import time
from .api import GoogleAPI
from threpose.settings import BASE_DIR
from src.caching.caching_gmap import APICaching
from decouple import config
gapi = GoogleAPI()
api_caching = APICaching()
PLACE_IMG_PATH = os.p... | search/views.py | from django.shortcuts import render
from django.http import JsonResponse
import os
import json
import time
from .api import GoogleAPI
from threpose.settings import BASE_DIR
from src.caching.caching_gmap import APICaching
from decouple import config
gapi = GoogleAPI()
api_caching = APICaching()
PLACE_IMG_PATH = os.p... | 0.668123 | 0.118181 |
from __future__ import division, unicode_literals, print_function
import os
import sys
import hashlib
import time
import logging
import sqlite3
import numpy as np
from io import BytesIO
from ..debugging import DebugPlot
try:
import tqdm
except ImportError:
tqdm = None
logger = logging.getLogger(__name__)
... | molyso/generic/etc.py | from __future__ import division, unicode_literals, print_function
import os
import sys
import hashlib
import time
import logging
import sqlite3
import numpy as np
from io import BytesIO
from ..debugging import DebugPlot
try:
import tqdm
except ImportError:
tqdm = None
logger = logging.getLogger(__name__)
... | 0.540681 | 0.336604 |
def countComponents1(n: int, edges: list[list[int]]) -> int:
"""
quick find based implemenation
Args:
n (int): number of nodes
edges (list[list[int]]): list of edges
Returns:
int: number of connected components
"""
connections = [n for n in range(n)]
for edge in ed... | py/dcp/leetcode/graph/connected_components.py | def countComponents1(n: int, edges: list[list[int]]) -> int:
"""
quick find based implemenation
Args:
n (int): number of nodes
edges (list[list[int]]): list of edges
Returns:
int: number of connected components
"""
connections = [n for n in range(n)]
for edge in ed... | 0.796134 | 0.579876 |
from pydantic import BaseModel, IPvAnyAddress, Field, validator
from socialserver.constants import MAX_PIXEL_RATIO
from typing import Literal, Optional
class _ServerConfigNetwork(BaseModel):
host: IPvAnyAddress
# 1-65535 is the valid TCP port range, hence the limit.
port: int = Field(..., ge=1, le=65535)... | socialserver/resources/config/schema.py |
from pydantic import BaseModel, IPvAnyAddress, Field, validator
from socialserver.constants import MAX_PIXEL_RATIO
from typing import Literal, Optional
class _ServerConfigNetwork(BaseModel):
host: IPvAnyAddress
# 1-65535 is the valid TCP port range, hence the limit.
port: int = Field(..., ge=1, le=65535)... | 0.818519 | 0.282425 |
import os
import sys
import codecs
import re
import sem.importers
from sem.storage import Document, SEMCorpus, Annotation
from sem.exporters import BratExporter
lang2months = { # firt element is empty so index method returns values from 1 to 12
u"fr": [u"", u"janvier", u"février", u"mars", u"avril", u"mai", u"ju... | make_matches.py |
import os
import sys
import codecs
import re
import sem.importers
from sem.storage import Document, SEMCorpus, Annotation
from sem.exporters import BratExporter
lang2months = { # firt element is empty so index method returns values from 1 to 12
u"fr": [u"", u"janvier", u"février", u"mars", u"avril", u"mai", u"ju... | 0.143938 | 0.201263 |
from typing import Any, Dict, List, Optional, Tuple, Union
import numpy as np
import pandas as pd
import tensorflow as tf
import tensorflow_probability as tfp
tfd = tfp.distributions
tfb = tfp.bijectors
# K Local Level Prior Sample Size
# This is equal to the original [R package](https://github.com/google/CausalImpa... | causalimpact/model.py | from typing import Any, Dict, List, Optional, Tuple, Union
import numpy as np
import pandas as pd
import tensorflow as tf
import tensorflow_probability as tfp
tfd = tfp.distributions
tfb = tfp.bijectors
# K Local Level Prior Sample Size
# This is equal to the original [R package](https://github.com/google/CausalImpa... | 0.959126 | 0.604749 |
import backbone.support.configurations_variables as confv
import backbone.support.data_loading as dl
import backbone.support.data_analysis as da
import backbone.support.data_cleaning as dc
import backbone.support.configuration_classes as confc
import backbone.support.saving_loading as sl
import backbone.support.plots_a... | backbone/mains-dataset_wise_structural/main_emodb_male.py | import backbone.support.configurations_variables as confv
import backbone.support.data_loading as dl
import backbone.support.data_analysis as da
import backbone.support.data_cleaning as dc
import backbone.support.configuration_classes as confc
import backbone.support.saving_loading as sl
import backbone.support.plots_a... | 0.464416 | 0.202187 |
import pytest
from mock import MagicMock
from powerfulseal.node import Node, NodeInventory
@pytest.fixture
def nodes():
return [
Node(id="id1", ip="192.168.127.12", az="AZ1", no=1, name="node1"),
Node(id="id2", ip="172.16.31.10", az="AZ2", no=2, name="node2"),
Node(id="id3", ip="172.16... | tests/node/test_node_inventory.py |
import pytest
from mock import MagicMock
from powerfulseal.node import Node, NodeInventory
@pytest.fixture
def nodes():
return [
Node(id="id1", ip="192.168.127.12", az="AZ1", no=1, name="node1"),
Node(id="id2", ip="172.16.31.10", az="AZ2", no=2, name="node2"),
Node(id="id3", ip="172.16... | 0.495117 | 0.634713 |
from __future__ import print_function
import json
import os
import time
from chromite.cbuildbot.stages import generic_stages
from chromite.lib import buildbucket_lib
from chromite.lib import build_requests
from chromite.lib import constants
from chromite.lib import config_lib
from chromite.lib import cros_logging as ... | third_party/chromite/cbuildbot/stages/scheduler_stages.py | from __future__ import print_function
import json
import os
import time
from chromite.cbuildbot.stages import generic_stages
from chromite.lib import buildbucket_lib
from chromite.lib import build_requests
from chromite.lib import constants
from chromite.lib import config_lib
from chromite.lib import cros_logging as ... | 0.677687 | 0.110327 |
import torch
class Generator(torch.nn.Module):
"""
Simple Generator Network
"""
def __init__(
self,
latent_dim,
n_classes,
code_dim,
img_size,
num_channels):
"""
Parameters
----------
latent_dim :... | dlutils/models/gans/info/models.py | import torch
class Generator(torch.nn.Module):
"""
Simple Generator Network
"""
def __init__(
self,
latent_dim,
n_classes,
code_dim,
img_size,
num_channels):
"""
Parameters
----------
latent_dim :... | 0.947247 | 0.695441 |
import json
import uuid
from datetime import timedelta
from unittest.mock import patch
from alpaca_trade_api.entity import Account as AlpacaAccount
from alpaca_trade_api.entity import Order as AlpacaOrder
from alpaca_trade_api.entity import Position as AlpacaPosition
from assets.models import Asset, Bar
from assets.te... | server/core/tests/test_tasks.py | import json
import uuid
from datetime import timedelta
from unittest.mock import patch
from alpaca_trade_api.entity import Account as AlpacaAccount
from alpaca_trade_api.entity import Order as AlpacaOrder
from alpaca_trade_api.entity import Position as AlpacaPosition
from assets.models import Asset, Bar
from assets.te... | 0.574037 | 0.235614 |
from dash.dependencies import Input, Output
import dash_core_components as dcc
import dash_html_components as html
import pandas as pd
from textwrap import dedent
import dash_table
from tutorial import styles, tools
examples = {
example: tools.load_example('tutorial/examples/table/{}'.format(example))
for ex... | tutorial/table/table_callbacks_chapter.py | from dash.dependencies import Input, Output
import dash_core_components as dcc
import dash_html_components as html
import pandas as pd
from textwrap import dedent
import dash_table
from tutorial import styles, tools
examples = {
example: tools.load_example('tutorial/examples/table/{}'.format(example))
for ex... | 0.76769 | 0.478529 |
from django import forms
from .models import OwnerProfileInfo, RenterProfileInfo, User, Boat, RentContract, BoatCrew, RepairContract, Crew
from django.utils import timezone
class UserForm(forms.ModelForm):
password = forms.CharField(widget=forms.PasswordInput())
class Meta:
model = User
fields... | boats/forms.py | from django import forms
from .models import OwnerProfileInfo, RenterProfileInfo, User, Boat, RentContract, BoatCrew, RepairContract, Crew
from django.utils import timezone
class UserForm(forms.ModelForm):
password = forms.CharField(widget=forms.PasswordInput())
class Meta:
model = User
fields... | 0.563138 | 0.084041 |
# Common Imports
import rospy
import roslib
from harmoni_common_lib.constants import State
from harmoni_common_lib.service_server import HarmoniServiceServer
from harmoni_common_lib.service_manager import HarmoniServiceManager
import harmoni_common_lib.helper_functions as hf
# Specific Imports
from harmoni_common_li... | harmoni_actuators/harmoni_web/scripts/harmoni_web/web_service.py |
# Common Imports
import rospy
import roslib
from harmoni_common_lib.constants import State
from harmoni_common_lib.service_server import HarmoniServiceServer
from harmoni_common_lib.service_manager import HarmoniServiceManager
import harmoni_common_lib.helper_functions as hf
# Specific Imports
from harmoni_common_li... | 0.474388 | 0.103703 |
import tensorflow as tf
from tensorflow import keras
from tensorflow.python.keras.models import Model
from tensorflow.python.keras import layers
from tensorflow.python.keras.models import Sequential
from tensorflow.python.keras.optimizers import Adam, SGD, RMSprop
from tensorflow.python.keras.layers import concatenate,... | unet.py | import tensorflow as tf
from tensorflow import keras
from tensorflow.python.keras.models import Model
from tensorflow.python.keras import layers
from tensorflow.python.keras.models import Sequential
from tensorflow.python.keras.optimizers import Adam, SGD, RMSprop
from tensorflow.python.keras.layers import concatenate,... | 0.896733 | 0.569583 |
import arcanelaunch
from arcanelaunch import setenv,getenv
import sys
import os
import copy
import shutil
import optparse
import re
link_dirs = "@ARCANE_LINK_DIRECTORIES@"
#TODO: traiter correctement les espaces dans les chemins
link_dirs.replace(" ",os.pathsep)
#print "link_dirs=",link_dirs
path_bin = "@ARCANE_INST... | cmake/build-system/csharp/axl/axl2cc.py |
import arcanelaunch
from arcanelaunch import setenv,getenv
import sys
import os
import copy
import shutil
import optparse
import re
link_dirs = "@ARCANE_LINK_DIRECTORIES@"
#TODO: traiter correctement les espaces dans les chemins
link_dirs.replace(" ",os.pathsep)
#print "link_dirs=",link_dirs
path_bin = "@ARCANE_INST... | 0.052838 | 0.062445 |
import sys
import time
import boto3
import logging
import kraken.kubernetes.client as kubecli
import kraken.node_actions.common_node_functions as nodeaction
from kraken.node_actions.abstract_node_scenarios import abstract_node_scenarios
class AWS:
def __init__(self):
self.boto_client = boto3.client("ec2")... | kraken/node_actions/aws_node_scenarios.py | import sys
import time
import boto3
import logging
import kraken.kubernetes.client as kubecli
import kraken.node_actions.common_node_functions as nodeaction
from kraken.node_actions.abstract_node_scenarios import abstract_node_scenarios
class AWS:
def __init__(self):
self.boto_client = boto3.client("ec2")... | 0.349755 | 0.125708 |
# ----------------------------------------------------------------------
# Imports
# ----------------------------------------------------------------------
import numpy as np
import SUAVE
from SUAVE.Core import Units
from SUAVE.Methods.Propulsion.turbofan_sizing import turbofan_sizing
from SUAVE.Methods.Geometry.Two... | SUAVE/SUAVE-2.5.0/regression/scripts/Vehicles/Boeing_737.py | # ----------------------------------------------------------------------
# Imports
# ----------------------------------------------------------------------
import numpy as np
import SUAVE
from SUAVE.Core import Units
from SUAVE.Methods.Propulsion.turbofan_sizing import turbofan_sizing
from SUAVE.Methods.Geometry.Two... | 0.584271 | 0.254871 |
import logging
from newrelic_plugin_agent.plugins import base
LOGGER = logging.getLogger(__name__)
class Riak(base.JSONStatsPlugin):
DEFAULT_PATH = '/stats'
GUID = 'com.meetme.newrelic_riak_agent'
def add_datapoints(self, stats):
"""Add all of the data points for a node
:param dict st... | newrelic_plugin_agent/plugins/riak.py | import logging
from newrelic_plugin_agent.plugins import base
LOGGER = logging.getLogger(__name__)
class Riak(base.JSONStatsPlugin):
DEFAULT_PATH = '/stats'
GUID = 'com.meetme.newrelic_riak_agent'
def add_datapoints(self, stats):
"""Add all of the data points for a node
:param dict st... | 0.558327 | 0.104112 |
from django.db import models
from django.contrib.auth import get_user_model
from django.contrib.contenttypes.models import ContentType
from rest_framework.exceptions import NotAuthenticated
from baserow.core.user_files.models import UserFile
from .mixins import (
OrderableMixin,
PolymorphicContentTypeMixin,
... | backend/src/baserow/core/models.py | from django.db import models
from django.contrib.auth import get_user_model
from django.contrib.contenttypes.models import ContentType
from rest_framework.exceptions import NotAuthenticated
from baserow.core.user_files.models import UserFile
from .mixins import (
OrderableMixin,
PolymorphicContentTypeMixin,
... | 0.635788 | 0.158956 |
import decimal
from blingalytics import sources
DIVISION_BY_ZERO = (decimal.InvalidOperation, ZeroDivisionError)
class DerivedSource(sources.Source):
def post_process(self, row, clean_inputs):
# Compute derived values for all columns on this row
for name, column in self._columns:
row... | blingalytics/sources/derived.py | import decimal
from blingalytics import sources
DIVISION_BY_ZERO = (decimal.InvalidOperation, ZeroDivisionError)
class DerivedSource(sources.Source):
def post_process(self, row, clean_inputs):
# Compute derived values for all columns on this row
for name, column in self._columns:
row... | 0.75401 | 0.46308 |
from scikits.audiolab import wavread
from scikits.audiolab import wavwrite
import operator
def weighting_vector(size, sigma, mu, fs):
weights = []
mu = mu/(fs/2) * size
weights = []
for t in range(size):
a = 1/(sigma * np.sqrt(2 * math.pi *sigma))
val = a * math.exp( - math.pow(t - mu,... | scripts/sibilant_detector.py | from scikits.audiolab import wavread
from scikits.audiolab import wavwrite
import operator
def weighting_vector(size, sigma, mu, fs):
weights = []
mu = mu/(fs/2) * size
weights = []
for t in range(size):
a = 1/(sigma * np.sqrt(2 * math.pi *sigma))
val = a * math.exp( - math.pow(t - mu,... | 0.587352 | 0.484868 |
import numpy as np
from sklearn.svm import SVC
from sklearn.naive_bayes import GaussianNB
from sklearn.linear_model import LogisticRegression
from sklearn.tree import DecisionTreeClassifier
class StackingEnsemble:
def __init__(self, layers=None, final=None):
if layers == None:
self.layers = [... | autogoal/experimental/stacking.py | import numpy as np
from sklearn.svm import SVC
from sklearn.naive_bayes import GaussianNB
from sklearn.linear_model import LogisticRegression
from sklearn.tree import DecisionTreeClassifier
class StackingEnsemble:
def __init__(self, layers=None, final=None):
if layers == None:
self.layers = [... | 0.829699 | 0.393705 |
from utlis.rank import setrank,isrank,remrank,remsudos,setsudo, GPranks
from utlis.send import Name,Glang
from utlis.tg import Bot
from config import *
from pyrogram import ReplyKeyboardMarkup, InlineKeyboardMarkup, InlineKeyboardButton
import threading, requests, time, random, re,json
import importlib
def d... | handlers/delete.py | from utlis.rank import setrank,isrank,remrank,remsudos,setsudo, GPranks
from utlis.send import Name,Glang
from utlis.tg import Bot
from config import *
from pyrogram import ReplyKeyboardMarkup, InlineKeyboardMarkup, InlineKeyboardButton
import threading, requests, time, random, re,json
import importlib
def d... | 0.15219 | 0.116036 |
import paddle
import paddle.nn as nn
import paddle.fluid.layers as layers
from .builder import NECKS
from paddle.vision.models.resnet import BasicBlock, BottleneckBlock
from ...modules.init import init_backbone_weight, normal_init, kaiming_init, constant_, reset_parameters, xavier_init, init_backbone_weight_simclr
... | passl/modeling/necks/base_neck.py |
import paddle
import paddle.nn as nn
import paddle.fluid.layers as layers
from .builder import NECKS
from paddle.vision.models.resnet import BasicBlock, BottleneckBlock
from ...modules.init import init_backbone_weight, normal_init, kaiming_init, constant_, reset_parameters, xavier_init, init_backbone_weight_simclr
... | 0.887881 | 0.321021 |
__author__ = "<NAME> <<EMAIL>>, <NAME> <<EMAIL>>"
import os
import md5
import json
import Queue
import threading
import time
class Worker(object):
''' Worker thread for concurrent process of tasks from a queue using multiple threads.
This worker is designed to never die, always keeping num_threads threads a... | web-search-engine/final/utils.py | __author__ = "<NAME> <<EMAIL>>, <NAME> <<EMAIL>>"
import os
import md5
import json
import Queue
import threading
import time
class Worker(object):
''' Worker thread for concurrent process of tasks from a queue using multiple threads.
This worker is designed to never die, always keeping num_threads threads a... | 0.535827 | 0.194483 |
import typing as tp
import numpy as np
from static_frame.core.util import mloc
from static_frame.core.util import FilePathOrFileLike
from static_frame.core.util import write_optional_file
from static_frame.core.display import DisplayFormats
from static_frame.core.display import DisplayActive
from static_frame.core.di... | static_frame/core/index_base.py | import typing as tp
import numpy as np
from static_frame.core.util import mloc
from static_frame.core.util import FilePathOrFileLike
from static_frame.core.util import write_optional_file
from static_frame.core.display import DisplayFormats
from static_frame.core.display import DisplayActive
from static_frame.core.di... | 0.613005 | 0.315024 |
import struct
def recv_all(sock, size):
received = ""
while len(received) < size:
data = sock.recv(size - len(received))
if data == "":
raise Exception("Lost connection")
else:
received += data
return received
class basePacker(object):
@classmethod
def pack(cls, value):
return... | tyger.py | import struct
def recv_all(sock, size):
received = ""
while len(received) < size:
data = sock.recv(size - len(received))
if data == "":
raise Exception("Lost connection")
else:
received += data
return received
class basePacker(object):
@classmethod
def pack(cls, value):
return... | 0.451568 | 0.225961 |
import torch.nn as nn
import torch.nn.utils.spectral_norm as SN
import torchvision
import torch
def conv_block(in_channels, out_channels, kernel_size, stride, padding=1, bias=True, activation=nn.ReLU(), transpose=False, no_BN=False, all_tanh=False, spec_norm=False):
if(transpose):
block = [nn.ConvTranspo... | models.py | import torch.nn as nn
import torch.nn.utils.spectral_norm as SN
import torchvision
import torch
def conv_block(in_channels, out_channels, kernel_size, stride, padding=1, bias=True, activation=nn.ReLU(), transpose=False, no_BN=False, all_tanh=False, spec_norm=False):
if(transpose):
block = [nn.ConvTranspo... | 0.922731 | 0.345906 |
from mako import runtime, filters, cache
UNDEFINED = runtime.UNDEFINED
__M_dict_builtin = dict
__M_locals_builtin = locals
_magic_number = 9
_modified_time = 1396763868.373039
_enable_loop = True
_template_filename = 'C:\\Users\\<NAME>\\Desktop\\MyStuff\\account\\scripts/user.jsm'
_template_uri = 'user.jsm'
_source_enc... | account/cached_templates/scripts/user.jsm.py | from mako import runtime, filters, cache
UNDEFINED = runtime.UNDEFINED
__M_dict_builtin = dict
__M_locals_builtin = locals
_magic_number = 9
_modified_time = 1396763868.373039
_enable_loop = True
_template_filename = 'C:\\Users\\<NAME>\\Desktop\\MyStuff\\account\\scripts/user.jsm'
_template_uri = 'user.jsm'
_source_enc... | 0.218253 | 0.121009 |
class Item:
def __init__(self, itemID, modid='minecraft'):
super().__init__()
self.modid = modid
self.id = itemID
def __str__(self) -> str:
return f"{self.modid}:{self.id}"
WHITE_CANDLE = Item('white_candle')
ORANGE_CANDLE = Item('orange_candle')
MAGENTA_CANDLE = I... | MinecraftDataHelper/items.py | class Item:
def __init__(self, itemID, modid='minecraft'):
super().__init__()
self.modid = modid
self.id = itemID
def __str__(self) -> str:
return f"{self.modid}:{self.id}"
WHITE_CANDLE = Item('white_candle')
ORANGE_CANDLE = Item('orange_candle')
MAGENTA_CANDLE = I... | 0.381104 | 0.040541 |
import os
import shutil
import sys
from typing import List, Union, Callable
import functools
import importlib
def require(pkg_name) -> Callable:
"""Returns a decorator function, ensures pkg_name is available and can be imported.
Parameters
----------
pkg_name: str
Name of the package required.... | cmselemental/util/importing.py | import os
import shutil
import sys
from typing import List, Union, Callable
import functools
import importlib
def require(pkg_name) -> Callable:
"""Returns a decorator function, ensures pkg_name is available and can be imported.
Parameters
----------
pkg_name: str
Name of the package required.... | 0.61115 | 0.306929 |
import numpy as np
import cvxpy as cp
import scipy.linalg
import scipy.optimize
import matplotlib.pyplot as plt
from pytope import Polytope
if __name__ == '__main__':
from generate_invariant_set import invariant_set
else:
from envs.generate_invariant_set import invariant_set
import torch
#%%%
XX = np.array([[-... | Cyclic_projections/envs/generate_vertices.py | import numpy as np
import cvxpy as cp
import scipy.linalg
import scipy.optimize
import matplotlib.pyplot as plt
from pytope import Polytope
if __name__ == '__main__':
from generate_invariant_set import invariant_set
else:
from envs.generate_invariant_set import invariant_set
import torch
#%%%
XX = np.array([[-... | 0.539954 | 0.466481 |
import os
import unittest
from functools import partial
from textwrap import dedent
from typing import Dict, List, Optional
from pants.base.build_environment import get_buildroot
from pants.option.option_value_container import OptionValueContainer
from pants.option.options_bootstrapper import OptionsBootstrapper
from... | src/python/pants/option/options_bootstrapper_test.py |
import os
import unittest
from functools import partial
from textwrap import dedent
from typing import Dict, List, Optional
from pants.base.build_environment import get_buildroot
from pants.option.option_value_container import OptionValueContainer
from pants.option.options_bootstrapper import OptionsBootstrapper
from... | 0.787114 | 0.238151 |
import requests
import time
import pandas as pd
from selenium import webdriver
from selenium.webdriver.chrome.options import Options
from bs4 import BeautifulSoup
import json
url = 'https://stats.nba.com/players/traditional/?PerMode=Totals&Season=2019-20&SeasonType=Regular%20Season&sort=PLAYER_NAME&dir=-1'
options = ... | exemplo_02/Exemplo_02.py | import requests
import time
import pandas as pd
from selenium import webdriver
from selenium.webdriver.chrome.options import Options
from bs4 import BeautifulSoup
import json
url = 'https://stats.nba.com/players/traditional/?PerMode=Totals&Season=2019-20&SeasonType=Regular%20Season&sort=PLAYER_NAME&dir=-1'
options = ... | 0.250546 | 0.078184 |
from urllib.request import Request
from api.drivers.student import student_drivers
from api.middlewares import authentication_middleware
from api.schemas.admin.admin_request_schema import admin_request_schemas
from api.schemas.student.request_schemas import student_request_schemas
from api.schemas.student.response_sch... | api/routes/admin/admin_student_routes.py | from urllib.request import Request
from api.drivers.student import student_drivers
from api.middlewares import authentication_middleware
from api.schemas.admin.admin_request_schema import admin_request_schemas
from api.schemas.student.request_schemas import student_request_schemas
from api.schemas.student.response_sch... | 0.404743 | 0.060975 |
import pytest
from distutils.version import LooseVersion
from f5.bigip.resource import MissingRequiredCreationParameter
from f5.bigip.tm.security.nat import Destination_Translation
from f5.bigip.tm.security.nat import Policy
from f5.bigip.tm.security.nat import Rule
from f5.bigip.tm.security.nat import Source_Transla... | f5/bigip/tm/security/test/functional/test_nat.py |
import pytest
from distutils.version import LooseVersion
from f5.bigip.resource import MissingRequiredCreationParameter
from f5.bigip.tm.security.nat import Destination_Translation
from f5.bigip.tm.security.nat import Policy
from f5.bigip.tm.security.nat import Rule
from f5.bigip.tm.security.nat import Source_Transla... | 0.569853 | 0.322953 |
from unittest import TestCase
from profile_generator.unit import Point
from .gamma import (
_exp,
_inverse_exp,
_inverse_linear,
_inverse_sqrt,
_linear,
_sqrt,
exp,
inverse_exp,
inverse_linear,
inverse_sqrt,
linear,
sqrt,
)
_GREY = Point(87 / 255, 119 / 255)
class Ga... | profile_generator/model/gamma_test.py | from unittest import TestCase
from profile_generator.unit import Point
from .gamma import (
_exp,
_inverse_exp,
_inverse_linear,
_inverse_sqrt,
_linear,
_sqrt,
exp,
inverse_exp,
inverse_linear,
inverse_sqrt,
linear,
sqrt,
)
_GREY = Point(87 / 255, 119 / 255)
class Ga... | 0.890675 | 0.905322 |
'''
Created on 23.06.2010
@author: <NAME>
model description of a rope consisting of n pendulums as a benchmark test
'''
'''
Masse = 4.91 Gramm
Massenmittelpunkt: ( Millimeter )
X = -0.00
Y = 0.00
Z = -46.18
Trägheitsmomente: ( Gramm * QuadratMillimeter )
Bezogen auf den Massenmittelpunkt, ausgerichtet auf das ... | examples/misc/multi_rope.py | '''
Created on 23.06.2010
@author: <NAME>
model description of a rope consisting of n pendulums as a benchmark test
'''
'''
Masse = 4.91 Gramm
Massenmittelpunkt: ( Millimeter )
X = -0.00
Y = 0.00
Z = -46.18
Trägheitsmomente: ( Gramm * QuadratMillimeter )
Bezogen auf den Massenmittelpunkt, ausgerichtet auf das ... | 0.496094 | 0.316581 |
import logging
from genomic_operations.dts.single_pos import PSEQPos, GeminiPos, TwoColPos
class Sniff(object):
"""
Creates and ordered list of functions that are used to try and
sniff the datatype from arbitrary files.
"""
def __init__(self):
self.sniffer_list = []
de... | genomic_operations/sniff/sniffer.py | import logging
from genomic_operations.dts.single_pos import PSEQPos, GeminiPos, TwoColPos
class Sniff(object):
"""
Creates and ordered list of functions that are used to try and
sniff the datatype from arbitrary files.
"""
def __init__(self):
self.sniffer_list = []
de... | 0.672977 | 0.36886 |
import keras
import numpy as np
from skimage import io
import matplotlib.pyplot as plt
from keras.layers import Input, Dense, Reshape
from keras.layers import BatchNormalization, Activation, ZeroPadding2D
from keras.models import Sequential, Model
from keras.optimizers import Adam
# Load the target Model and make it... | 2_ExtractingInformation/solution_2_0.py | import keras
import numpy as np
from skimage import io
import matplotlib.pyplot as plt
from keras.layers import Input, Dense, Reshape
from keras.layers import BatchNormalization, Activation, ZeroPadding2D
from keras.models import Sequential, Model
from keras.optimizers import Adam
# Load the target Model and make it... | 0.792263 | 0.416322 |
from __future__ import print_function
import logging
import sys
import os
import cPickle
import numpy as np
from scipy.sparse import dok_matrix
from scipy.io import mmwrite, mmread
import text_entail.dictionary as td
import text_entail.io as tio
def w1Asfeature(d_triples, d_w1):
"""
"""
w1_mat = dok_matr... | src/text_entail/matrix.py | from __future__ import print_function
import logging
import sys
import os
import cPickle
import numpy as np
from scipy.sparse import dok_matrix
from scipy.io import mmwrite, mmread
import text_entail.dictionary as td
import text_entail.io as tio
def w1Asfeature(d_triples, d_w1):
"""
"""
w1_mat = dok_matr... | 0.340485 | 0.217961 |
import abc
from typing import Any
import torch
AGGREGATION_MODES = ["mean", "max", "min"]
class Metric(metaclass=abc.ABCMeta):
"""abstract class for Metric objects.
Example:
Simple usage of the Metric class::
class MyMetric(Metric):
def _update(self, predictions, truth):... | frarch/modules/metrics/base.py | import abc
from typing import Any
import torch
AGGREGATION_MODES = ["mean", "max", "min"]
class Metric(metaclass=abc.ABCMeta):
"""abstract class for Metric objects.
Example:
Simple usage of the Metric class::
class MyMetric(Metric):
def _update(self, predictions, truth):... | 0.946014 | 0.363195 |
from pprint import pprint
import sympy as sym
sym.init_printing(use_latex=True)
import numpy as np
from .benchmark import Benchmark
class Schubert(Benchmark):
def __init__(self, case: str):
super().__init__()
if case not in {'p3', 'p8', 'p16', 'p22'}:
raise ValueError('case must b... | zoo/schubert.py | from pprint import pprint
import sympy as sym
sym.init_printing(use_latex=True)
import numpy as np
from .benchmark import Benchmark
class Schubert(Benchmark):
def __init__(self, case: str):
super().__init__()
if case not in {'p3', 'p8', 'p16', 'p22'}:
raise ValueError('case must b... | 0.325092 | 0.329001 |
import logging
import os
import platform
import shutil
import sys
import unittest
import uuid
from copy import copy
import psutil
from psutil import AccessDenied, NoSuchProcess
from pyngrok.conf import PyngrokConfig
from pyngrok import ngrok, installer, conf
from pyngrok import process
__author__ = "<NAME>"
__copyri... | tests/testcase.py | import logging
import os
import platform
import shutil
import sys
import unittest
import uuid
from copy import copy
import psutil
from psutil import AccessDenied, NoSuchProcess
from pyngrok.conf import PyngrokConfig
from pyngrok import ngrok, installer, conf
from pyngrok import process
__author__ = "<NAME>"
__copyri... | 0.247532 | 0.058025 |
import os
import tempfile
import unittest
import pytorch_lightning as pl
import torch.utils.data
from hydra import compose, initialize_config_dir
from nuplan.planning.script.builders.model_builder import build_nn_model
from nuplan.planning.script.builders.scenario_building_builder import build_scenario_builder
from nu... | nuplan/planning/script/test/test_config_dataloader.py | import os
import tempfile
import unittest
import pytorch_lightning as pl
import torch.utils.data
from hydra import compose, initialize_config_dir
from nuplan.planning.script.builders.model_builder import build_nn_model
from nuplan.planning.script.builders.scenario_building_builder import build_scenario_builder
from nu... | 0.471467 | 0.301748 |
import numpy as np
from qutip import (
rand_ket, rand_dm, rand_herm, rand_unitary, rand_ket_haar, rand_dm_hs,
rand_super, rand_unitary_haar, rand_dm_ginibre, rand_super_bcsz, qeye,
rand_stochastic,
)
import pytest
@pytest.mark.repeat(5)
@pytest.mark.parametrize('func', [rand_unitary, rand_unitary_haar])
... | qutip/tests/test_random.py |
import numpy as np
from qutip import (
rand_ket, rand_dm, rand_herm, rand_unitary, rand_ket_haar, rand_dm_hs,
rand_super, rand_unitary_haar, rand_dm_ginibre, rand_super_bcsz, qeye,
rand_stochastic,
)
import pytest
@pytest.mark.repeat(5)
@pytest.mark.parametrize('func', [rand_unitary, rand_unitary_haar])
... | 0.627038 | 0.617657 |
import hmac
import hashlib
import sys
from ..errors import SignatureVerificationError
class Utility(object):
def __init__(self, client=None):
self.client = client
def verify_payment_signature(self, parameters):
order_id = str(parameters['razorpay_order_id'])
payment_id = str(paramet... | saleor/lib/python3.7/site-packages/razorpay/utility/utility.py | import hmac
import hashlib
import sys
from ..errors import SignatureVerificationError
class Utility(object):
def __init__(self, client=None):
self.client = client
def verify_payment_signature(self, parameters):
order_id = str(parameters['razorpay_order_id'])
payment_id = str(paramet... | 0.409575 | 0.176636 |
import falcon
from falcon.media.validators import jsonschema
from management_api.utils.logger import get_logger
from management_api.endpoints.endpoint_utils import create_endpoint, delete_endpoint, \
scale_endpoint, update_endpoint, view_endpoint, list_endpoints
from management_api.schemas.endpoints import endpo... | management/management_api/endpoints/endpoints.py |
import falcon
from falcon.media.validators import jsonschema
from management_api.utils.logger import get_logger
from management_api.endpoints.endpoint_utils import create_endpoint, delete_endpoint, \
scale_endpoint, update_endpoint, view_endpoint, list_endpoints
from management_api.schemas.endpoints import endpo... | 0.358802 | 0.048339 |
import streamlit as st
from PIL import Image
def app():
st.title("Vegetation Analysis")
st.markdown(
"""
The goal of this task is to discover the use of different vegetation indices to identify the level of
desertification in northern Iraq. Indices of interest include NDVI, NDWI, NDBI... | apps/vegetation_analysis.py | import streamlit as st
from PIL import Image
def app():
st.title("Vegetation Analysis")
st.markdown(
"""
The goal of this task is to discover the use of different vegetation indices to identify the level of
desertification in northern Iraq. Indices of interest include NDVI, NDWI, NDBI... | 0.795857 | 0.589894 |
"""Generated protocol buffer code."""
from google.protobuf import descriptor as _descriptor
from google.protobuf import message as _message
from google.protobuf import reflection as _reflection
from google.protobuf import symbol_database as _symbol_database
# @@protoc_insertion_point(imports)
_sym_db = _symbol_databas... | tianshu_serving/proto/inference_pb2.py | """Generated protocol buffer code."""
from google.protobuf import descriptor as _descriptor
from google.protobuf import message as _message
from google.protobuf import reflection as _reflection
from google.protobuf import symbol_database as _symbol_database
# @@protoc_insertion_point(imports)
_sym_db = _symbol_databas... | 0.224565 | 0.075585 |
import re
from comply.rules.rule import *
from comply.rules.patterns import FUNC_BODY_PATTERN
from comply.util.scope import depth
class ScopeTooDeep(Rule):
""" Don't write deeply nested code.
A deeply nested scope is often an indication of too high complexity and can be
difficult to read.
"""
... | comply/rules/standard/scope_too_deep.py |
import re
from comply.rules.rule import *
from comply.rules.patterns import FUNC_BODY_PATTERN
from comply.util.scope import depth
class ScopeTooDeep(Rule):
""" Don't write deeply nested code.
A deeply nested scope is often an indication of too high complexity and can be
difficult to read.
"""
... | 0.666714 | 0.246375 |
import pytest
import datetime
from pupa.scrape import Event
def event_obj():
e = Event(
name="get-together",
start_date=datetime.datetime.utcnow().isoformat().split('.')[0] + 'Z',
location_name="Joe's Place",
)
e.add_source(url='http://example.com/foobar')
return e
def test_b... | pupa/tests/scrape/test_event_scrape.py | import pytest
import datetime
from pupa.scrape import Event
def event_obj():
e = Event(
name="get-together",
start_date=datetime.datetime.utcnow().isoformat().split('.')[0] + 'Z',
location_name="Joe's Place",
)
e.add_source(url='http://example.com/foobar')
return e
def test_b... | 0.574753 | 0.518363 |
import unittest
from Multi_cell import *
class MultiCellTestCase(unittest.TestCase):
def test_multi_cell_INV_n_INV(self):
str_netlist_1 = "M0001 GND IN001 OUT01 GND NMOS\n" \
"M0002 OUT01 IN001 VDD VDD PMOS\n"
str_netlist_2 = "M0001 GND IN001 OUT01 GND NMOS\n" \
... | MultiCellTest.py | import unittest
from Multi_cell import *
class MultiCellTestCase(unittest.TestCase):
def test_multi_cell_INV_n_INV(self):
str_netlist_1 = "M0001 GND IN001 OUT01 GND NMOS\n" \
"M0002 OUT01 IN001 VDD VDD PMOS\n"
str_netlist_2 = "M0001 GND IN001 OUT01 GND NMOS\n" \
... | 0.500488 | 0.364778 |
import numpy as np
import pytest
import pandas as pd
import pandas._testing as tm
@pytest.mark.parametrize("align_axis", [0, 1, "index", "columns"])
def test_compare_axis(align_axis):
# GH#30429
df = pd.DataFrame(
{"col1": ["a", "b", "c"], "col2": [1.0, 2.0, np.nan], "col3": [1.0, 2.0, 3.0]},
... | pandas/tests/frame/methods/test_compare.py | import numpy as np
import pytest
import pandas as pd
import pandas._testing as tm
@pytest.mark.parametrize("align_axis", [0, 1, "index", "columns"])
def test_compare_axis(align_axis):
# GH#30429
df = pd.DataFrame(
{"col1": ["a", "b", "c"], "col2": [1.0, 2.0, np.nan], "col3": [1.0, 2.0, 3.0]},
... | 0.443841 | 0.648286 |
import json
import logging
from .base import WeTransferBase
from .file import File
LOG = logging.getLogger("wetransfer")
LOG.addHandler(logging.NullHandler())
LOG.setLevel(logging.INFO)
class WeTransfer(WeTransferBase):
WE_ENDPOINT_DEV = 'https://dev.wetransfer.com'
def __finalize_transfer(self, transfer_id... | wetransfer/transfer.py | import json
import logging
from .base import WeTransferBase
from .file import File
LOG = logging.getLogger("wetransfer")
LOG.addHandler(logging.NullHandler())
LOG.setLevel(logging.INFO)
class WeTransfer(WeTransferBase):
WE_ENDPOINT_DEV = 'https://dev.wetransfer.com'
def __finalize_transfer(self, transfer_id... | 0.40028 | 0.14013 |
from datetime import datetime
from dateutil.relativedelta import relativedelta
from django.contrib.contenttypes.fields import GenericForeignKey
from django.contrib.contenttypes.models import ContentType
from django.db import models, transaction
from django.db.models import Q
from django.utils import timezone
from djan... | gdpr/models.py | from datetime import datetime
from dateutil.relativedelta import relativedelta
from django.contrib.contenttypes.fields import GenericForeignKey
from django.contrib.contenttypes.models import ContentType
from django.db import models, transaction
from django.db.models import Q
from django.utils import timezone
from djan... | 0.804905 | 0.131982 |
import os
import glob
from PIL import Image
from resizeimage import resizeimage
import sys
from xml.etree.ElementTree import ElementTree
from xml.etree.ElementTree import Element
import xml.etree.ElementTree as etree
import xml.etree.cElementTree as ET
from yattag import Doc, indent
import shutil
import pand... | size_name_background_removal.py | import os
import glob
from PIL import Image
from resizeimage import resizeimage
import sys
from xml.etree.ElementTree import ElementTree
from xml.etree.ElementTree import Element
import xml.etree.ElementTree as etree
import xml.etree.cElementTree as ET
from yattag import Doc, indent
import shutil
import pand... | 0.193414 | 0.098296 |
from utils.utils import block_diag, stack_matrices, sum_sparse
from torch.nn.modules.module import Module
from torch.nn.parameter import Parameter
from torch import nn
import math
import torch
class RelationalGraphConvolution(Module):
""" Relational Graph Convolution (RGC) Layer (as described in https://arxiv.org... | torch_rgvae/layers/RGC_layers.py | from utils.utils import block_diag, stack_matrices, sum_sparse
from torch.nn.modules.module import Module
from torch.nn.parameter import Parameter
from torch import nn
import math
import torch
class RelationalGraphConvolution(Module):
""" Relational Graph Convolution (RGC) Layer (as described in https://arxiv.org... | 0.917052 | 0.622517 |
import logging
from rackattack.physical import logconfig
from rackattack.ssh import connection
connection.discardParamikoLogs()
connection.discardSSHDebugMessages()
logging.getLogger("pika").setLevel(logging.INFO)
import time
import argparse
from rackattack.physical import config
from rackattack.physical import network... | rackattack/physical/main.py | import logging
from rackattack.physical import logconfig
from rackattack.ssh import connection
connection.discardParamikoLogs()
connection.discardSSHDebugMessages()
logging.getLogger("pika").setLevel(logging.INFO)
import time
import argparse
from rackattack.physical import config
from rackattack.physical import network... | 0.32306 | 0.045058 |
import datetime
STRING_UNAVAILABLE = "spaceapi: N/A"
class Py3status:
"""
"""
# available configuration parameters
button_url = 3
cache_timeout = 60
format = "{state}[ {lastchanged}]"
format_lastchanged = "since %H:%M"
state_closed = "closed"
state_open = "open"
url = "https:... | py3status/modules/spaceapi.py | import datetime
STRING_UNAVAILABLE = "spaceapi: N/A"
class Py3status:
"""
"""
# available configuration parameters
button_url = 3
cache_timeout = 60
format = "{state}[ {lastchanged}]"
format_lastchanged = "since %H:%M"
state_closed = "closed"
state_open = "open"
url = "https:... | 0.426799 | 0.278045 |
from odoo import models, fields, api, tools, _
from odoo.http import request, content_disposition
from odoo.addons.hs_query.libs.query_libs import query_and_count_data, get_query_statement_by_code
import json
import traceback
ERROR_NO_STATEMENT_CODE = u"数据库查询代码[ %s ]不存在, 请联系管理员!!"
ERROR_SQL_QUERY = u"数据库查询异常, 请联系管理员!... | hs_query/models/query_adapter.py |
from odoo import models, fields, api, tools, _
from odoo.http import request, content_disposition
from odoo.addons.hs_query.libs.query_libs import query_and_count_data, get_query_statement_by_code
import json
import traceback
ERROR_NO_STATEMENT_CODE = u"数据库查询代码[ %s ]不存在, 请联系管理员!!"
ERROR_SQL_QUERY = u"数据库查询异常, 请联系管理员!... | 0.283285 | 0.087525 |
import atexit
import os
import random
import time
import traceback
from mpire.pool import WorkerPool
import pickle
from ditk import logging
import tempfile
import socket
from os import path
from typing import Callable, Dict, List, Optional, Tuple, Union, Set
from threading import Thread
from ding.framework.event_loop i... | ding/framework/parallel.py | import atexit
import os
import random
import time
import traceback
from mpire.pool import WorkerPool
import pickle
from ditk import logging
import tempfile
import socket
from os import path
from typing import Callable, Dict, List, Optional, Tuple, Union, Set
from threading import Thread
from ding.framework.event_loop i... | 0.722527 | 0.207014 |
import sys
import typing
from collections import OrderedDict
def raises(err, lamda):
try:
lamda()
return False
except err:
return True
def expand_tuples(L):
"""
>>> expand_tuples([1, (2, 3)])
[(1, 2), (1, 3)]
>>> expand_tuples([1, 2])
[(1, 2)]
"""
if no... | imperative/python/megengine/core/tensor/multipledispatch/utils.py |
import sys
import typing
from collections import OrderedDict
def raises(err, lamda):
try:
lamda()
return False
except err:
return True
def expand_tuples(L):
"""
>>> expand_tuples([1, (2, 3)])
[(1, 2), (1, 3)]
>>> expand_tuples([1, 2])
[(1, 2)]
"""
if no... | 0.596081 | 0.486697 |
"""Test materialized views"""
import pytest
from pyrseas.testutils import DatabaseToMapTestCase
from pyrseas.testutils import InputMapToSqlTestCase, fix_indent
CREATE_TABLE = "CREATE TABLE t1 (c1 INTEGER, c2 TEXT, c3 INTEGER)"
VIEW_STMT = "SELECT c1, c3 * 2 AS mc3 FROM t1"
CREATE_STMT = "CREATE MATERIALIZED VIEW sd.... | tests/dbobject/test_matview.py | """Test materialized views"""
import pytest
from pyrseas.testutils import DatabaseToMapTestCase
from pyrseas.testutils import InputMapToSqlTestCase, fix_indent
CREATE_TABLE = "CREATE TABLE t1 (c1 INTEGER, c2 TEXT, c3 INTEGER)"
VIEW_STMT = "SELECT c1, c3 * 2 AS mc3 FROM t1"
CREATE_STMT = "CREATE MATERIALIZED VIEW sd.... | 0.587943 | 0.428622 |
from django import forms
from django.contrib.localflavor.ar.forms import ARPostalCodeField
from django.contrib.localflavor.ar.forms import ARProvinceSelect
from django.forms.util import ErrorList
from ventas.models import DatosDeEnvio, GastosDeEnvio, ARP
class DumbSelect(forms.Select):
EMPTY_CHOICES = [('Otra', ... | ventas/forms.py | from django import forms
from django.contrib.localflavor.ar.forms import ARPostalCodeField
from django.contrib.localflavor.ar.forms import ARProvinceSelect
from django.forms.util import ErrorList
from ventas.models import DatosDeEnvio, GastosDeEnvio, ARP
class DumbSelect(forms.Select):
EMPTY_CHOICES = [('Otra', ... | 0.561936 | 0.128279 |
from hops import constants
from hops.featurestore_impl.dao.datasets.training_dataset import TrainingDataset
from hops.featurestore_impl.dao.featuregroups.featuregroup import Featuregroup
from hops.featurestore_impl.dao.featurestore.featurestore import Featurestore
from hops.featurestore_impl.dao.settings.featurestore_s... | hops/featurestore_impl/dao/common/featurestore_metadata.py | from hops import constants
from hops.featurestore_impl.dao.datasets.training_dataset import TrainingDataset
from hops.featurestore_impl.dao.featuregroups.featuregroup import Featuregroup
from hops.featurestore_impl.dao.featurestore.featurestore import Featurestore
from hops.featurestore_impl.dao.settings.featurestore_s... | 0.651355 | 0.315604 |
import numpy as np
from scipy.integrate import ode, odeint
import matplotlib.pyplot as plt
parsec = 3.086 * 1e16 #m
year = 3.156 * 1e7 #s
pi = 3.14159265358979323846
G = 6.67430 * 1e-11 #N * m2 / kg2
LO=93.016 * 1e9 * 9.461e15/2 #90 billions ly in m (diameter)
a0=1
omega_R = 4.8e-5
omega_lambda = 0.683-omega_R
omeg... | Rewinding the Universe to the Beginning of Time/Code/simulate_universe.py | import numpy as np
from scipy.integrate import ode, odeint
import matplotlib.pyplot as plt
parsec = 3.086 * 1e16 #m
year = 3.156 * 1e7 #s
pi = 3.14159265358979323846
G = 6.67430 * 1e-11 #N * m2 / kg2
LO=93.016 * 1e9 * 9.461e15/2 #90 billions ly in m (diameter)
a0=1
omega_R = 4.8e-5
omega_lambda = 0.683-omega_R
omeg... | 0.357007 | 0.497864 |
import unittest
from sliding_puzzle.algorithm.search import Search
from sliding_puzzle.representation.puzzle import Puzzle
class SolvableAtFirstTestCase(unittest.TestCase):
def test_is_unsolvable(self):
puzzle: Puzzle = Puzzle(
[
[0, 1, 2, 3, 4, 5],
[6, 7, 8, 9... | tests/test_solvable.py | import unittest
from sliding_puzzle.algorithm.search import Search
from sliding_puzzle.representation.puzzle import Puzzle
class SolvableAtFirstTestCase(unittest.TestCase):
def test_is_unsolvable(self):
puzzle: Puzzle = Puzzle(
[
[0, 1, 2, 3, 4, 5],
[6, 7, 8, 9... | 0.551815 | 0.689422 |
import random
from itertools import izip_longest, groupby, chain
from wa.framework.configuration.core import (MetaConfiguration, RunConfiguration,
JobGenerator, Status, settings)
from wa.framework.configuration.parsers import ConfigParser
from wa.framework.configuration.plu... | wa/framework/configuration/execution.py | import random
from itertools import izip_longest, groupby, chain
from wa.framework.configuration.core import (MetaConfiguration, RunConfiguration,
JobGenerator, Status, settings)
from wa.framework.configuration.parsers import ConfigParser
from wa.framework.configuration.plu... | 0.577853 | 0.120258 |
import apps.basics.op_drf.fields
from django.conf import settings
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
initial = True
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
('permission', '0001_... | backend/apps/projects/api/migrations/0001_initial.py |
import apps.basics.op_drf.fields
from django.conf import settings
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
initial = True
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
('permission', '0001_... | 0.385837 | 0.217504 |
from queue import PriorityQueue
Coordinate = tuple[int, int]
map: list[list[int]] = []
PART_TWO = True
# For each position, we store the lowest cost path to get there.
lowest_cost: list[list[None | tuple[int, list[Coordinate]]]] = []
with open('2021-12-15.txt') as f:
for line in (l.strip() for l in f):
... | 2021/2021-12-15.py | from queue import PriorityQueue
Coordinate = tuple[int, int]
map: list[list[int]] = []
PART_TWO = True
# For each position, we store the lowest cost path to get there.
lowest_cost: list[list[None | tuple[int, list[Coordinate]]]] = []
with open('2021-12-15.txt') as f:
for line in (l.strip() for l in f):
... | 0.302082 | 0.367015 |
import itertools
import pytest
from multpersist import OrderNotFound, compute_mp_order, \
efficient_candidate_generator, find_max_order, find_next, \
find_with_order, infinite_candidate_generator, is_in_order
def predetermined_number_generator():
for x in [10, 18, 237, 2777778888899, 277777788... | tests/test_compute_order.py |
import itertools
import pytest
from multpersist import OrderNotFound, compute_mp_order, \
efficient_candidate_generator, find_max_order, find_next, \
find_with_order, infinite_candidate_generator, is_in_order
def predetermined_number_generator():
for x in [10, 18, 237, 2777778888899, 277777788... | 0.426083 | 0.44059 |
from pprint import pformat
from six import iteritems
import re
class CouponDiscount(object):
"""
NOTE: This class is auto generated by the swagger code generator program.
Do not edit the class manually.
"""
def __init__(self, created=None, changed_by=None, updated=None, pricing_component=None, pri... | billforward/models/coupon_discount.py | from pprint import pformat
from six import iteritems
import re
class CouponDiscount(object):
"""
NOTE: This class is auto generated by the swagger code generator program.
Do not edit the class manually.
"""
def __init__(self, created=None, changed_by=None, updated=None, pricing_component=None, pri... | 0.757884 | 0.188567 |