input stringlengths 28 198k | output stringlengths 3 71k | file stringlengths 19 330 | input_tokens int64 5 159k | output_tokens int64 3 9.07k | __index_level_0__ int64 2 449k |
|---|---|---|---|---|---|
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME kernels
STORE_NAME kernels
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate any tests for kernel'
LOAD_CONST ('reason',)
CALL_FUNCTION
LOAD_CONST <code object test_pyawkward_IndexedA... | import pytest
import kernels
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_pyawkward_IndexedArray32_reduce_next_nonlocal_nextshifts_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-spec/test_pyawkward_IndexedArray32_reduce_next_nonlocal_nextshifts_64.py | 219 | 80 | 220,392 |
LOAD_CONST 0
LOAD_CONST ('absolute_import',)
IMPORT_NAME __future__
IMPORT_FROM absolute_import
STORE_NAME absolute_import
POP_TOP
LOAD_CONST 1
LOAD_CONST ('DatabaseToolsClient',)
IMPORT_NAME database_tools_client
IMPORT_FROM DatabaseToolsClient
STORE_NAME DatabaseToolsClient
POP_TOP
LOAD_CONST 1
LOAD_CONST ('Databas... | from __future__ import absolute_import
from .database_tools_client import DatabaseToolsClient
from .database_tools_client_composite_operations import (
DatabaseToolsClientCompositeOperations,
)
from . import models
__all__ = ["DatabaseToolsClient", "DatabaseToolsClientCompositeOperations", "models"]
| data/oci-2.122.0/src/oci/database_tools/__init__.py | 169 | 80 | 181,833 |
LOAD_CONST 0
LOAD_CONST ('re_path',)
IMPORT_NAME django.urls
IMPORT_FROM re_path
STORE_NAME re_path
POP_TOP
LOAD_CONST 1
LOAD_CONST ('views',)
IMPORT_NAME
IMPORT_FROM views
STORE_NAME views
POP_TOP
LOAD_NAME re_path
LOAD_CONST '^more/(?P<extra>\\w+)/$'
LOAD_NAME views
LOAD_ATTR empty_view
LOAD_CONST 'sub-extra'
LO... | from django.urls import re_path
from . import views
urlpatterns = [
re_path(
r"^more/(?P<extra>\w+)/$",
views.empty_view,
{"sub-extra": True},
name="inner-more",
),
]
| data/Django-5.0.2/tests/urlpatterns/more_urls.py | 126 | 80 | 325,659 |
LOAD_CONST 0
LOAD_CONST ('download_raw_files',)
IMPORT_NAME opendatasets.utils
IMPORT_FROM download_raw_files
STORE_NAME download_raw_files
POP_TOP
LOAD_CONST <code object download at 0x7f8e2fe58930, file "f.py", line 4>
LOAD_CONST 'download'
MAKE_FUNCTION
STORE_NAME download
LOAD_CONST None
RETURN_VALUE
LOAD_GLOBAL ... | from opendatasets.utils import download_raw_files
def download(dataset_id, data_dir, dry_run, **kwargs):
download_raw_files(
dataset_id, data_dir, dry_run, ["state-of-javascript-2018-responses.ndjson"]
)
| data/opendatasets-0.1.22/opendatasets/datasets/state-of-javascript-2018/__init__.py | 142 | 80 | 405,146 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME doctest
STORE_NAME doctest
LOAD_CONST 0
LOAD_CONST ('Path',)
IMPORT_NAME pathlib
IMPORT_FROM Path
STORE_NAME Path
POP_TOP
LOAD_CONST <code object test_readme at 0x7fab81f0b810, file "f.py", line 6>
LOAD_CONST 'test_readme'
MAKE_FUNCTION
STORE_NAME test_readme
LOAD_CONST None
R... | import doctest
from pathlib import Path
def test_readme():
failure_count, _ = doctest.testfile(
str(Path("..") / "README.rst"), verbose=False, optionflags=doctest.ELLIPSIS
)
assert failure_count == 0
| data/pg8000-1.30.4/test/test_readme.py | 197 | 80 | 196,078 |
LOAD_CONST '\nExpose public feature toggle API.\n'
STORE_NAME __doc__
LOAD_CONST 1
LOAD_CONST ('SettingDictToggle', 'SettingToggle')
IMPORT_NAME internal.setting_toggle
IMPORT_FROM SettingDictToggle
STORE_NAME SettingDictToggle
IMPORT_FROM SettingToggle
STORE_NAME SettingToggle
POP_TOP
LOAD_CONST 1
LOAD_CONST ('NonNa... | """
Expose public feature toggle API.
"""
from .internal.setting_toggle import SettingDictToggle, SettingToggle
from .internal.waffle.flag import NonNamespacedWaffleFlag, WaffleFlag
from .internal.waffle.switch import NonNamespacedWaffleSwitch, WaffleSwitch
| data/edx-toggles-5.1.1/edx_toggles/toggles/__init__.py | 207 | 80 | 169,910 |
LOAD_CONST '\nCore module exposes the document interfaces\nThese will be ingested via Drones, built by Builders, and served via the API.\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('PackageNotFoundError', 'version')
IMPORT_NAME importlib.metadata
IMPORT_FROM PackageNotFoundError
STORE_NAME PackageNotFoundError
IMPO... | """
Core module exposes the document interfaces
These will be ingested via Drones, built by Builders, and served via the API.
"""
from importlib.metadata import PackageNotFoundError, version
try:
__version__ = version("emmet-core")
except PackageNotFoundError: # pragma: no cover
pass
| data/emmet-core-0.77.1/emmet/core/__init__.py | 149 | 80 | 308,059 |
LOAD_CONST '\n An enumeration class for the message type field which describe party status\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('Enum',)
IMPORT_NAME enum
IMPORT_FROM Enum
STORE_NAME Enum
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object StatusType at 0x7faa5f1b6a50, file "f.py", line 9>
LOAD_CONST 'StatusTyp... | """
An enumeration class for the message type field which describe party status
"""
from enum import Enum
class StatusType(Enum):
"""
Status types for Party
"""
IDLE = 1
TRAINING = 2
EVALUATING = 3
STOPPING = 4
| data/ibm_watson_machine_learning-1.0.347/ibm_watson_machine_learning/libs/ibmfl/py38-tf24-pt17-sk23/ibmfl/party/status_type.py | 202 | 80 | 251,951 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME importlib
STORE_NAME importlib
SETUP_EXCEPT to 22
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME importlib.util
STORE_NAME importlib
POP_BLOCK
JUMP_FORWARD to 42
DUP_TOP
LOAD_NAME ImportError
COMPARE_OP exception match
POP_JUMP_IF_FALSE
POP_TOP
POP_TOP
POP_TOP
POP_EXCEPT
JUMP_FORW... | import importlib
try:
import importlib.util
except ImportError:
pass
try:
module_from_spec = importlib.util.module_from_spec
except AttributeError:
def module_from_spec(spec):
return spec.loader.load_module(spec.name)
| data/clvm_rs-0.6.0/venv/lib/python3.7/site-packages/setuptools/py34compat.py | 227 | 80 | 288,231 |
LOAD_CONST 0
LOAD_CONST ('UUID',)
IMPORT_NAME uuid
IMPORT_FROM UUID
STORE_NAME UUID
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME typer
STORE_NAME typer
LOAD_NAME UUID
LOAD_CONST ('user_id',)
BUILD_CONST_KEY_MAP
LOAD_CONST <code object main at 0x7fab641e7030, file "f.py", line 6>
LOAD_CONST 'main'
MAKE_FUNCTION
ST... | from uuid import UUID
import typer
def main(user_id: UUID):
print(f"USER_ID is {user_id}")
print(f"UUID version is: {user_id.version}")
if __name__ == "__main__":
typer.run(main)
| data/typer-0.9.0/docs_src/parameter_types/uuid/tutorial001.py | 190 | 80 | 224,499 |
LOAD_CONST 'Generated API for package: tensorflow_examples.lite.model_maker.third_party.recommendation.ml.data.example_generation_movielens.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME tensorflow_examples.lite.model_maker.third_party.recommendation.ml.data.example_generation_movielens
IMPORT_STAR
LO... | """Generated API for package: tensorflow_examples.lite.model_maker.third_party.recommendation.ml.data.example_generation_movielens."""
from tensorflow_examples.lite.model_maker.third_party.recommendation.ml.data.example_generation_movielens import *
| data/tflite-model-maker-nightly-0.4.4.dev202402210609/src/tflite_model_maker/python/third_party/recommendation/ml/data/example_generation_movielens.py | 100 | 80 | 389,286 |
LOAD_CONST 'Adds constants for Trafikverket Weather integration.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('Platform',)
IMPORT_NAME homeassistant.const
IMPORT_FROM Platform
STORE_NAME Platform
POP_TOP
LOAD_CONST 'trafikverket_weatherstation'
STORE_NAME DOMAIN
LOAD_CONST 'station'
STORE_NAME CONF_STATION
LOAD_NAM... | """Adds constants for Trafikverket Weather integration."""
from homeassistant.const import Platform
DOMAIN = "trafikverket_weatherstation"
CONF_STATION = "station"
PLATFORMS = [Platform.SENSOR]
ATTRIBUTION = "Data provided by Trafikverket"
| data/homeassistant-2024.2.2/homeassistant/components/trafikverket_weatherstation/const.py | 119 | 80 | 297,690 |
LOAD_CONST 4
STORE_NAME param
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pandas
STORE_NAME pd
LOAD_NAME pd
LOAD_ATTR DataFrame
LOAD_CONST 1
LOAD_CONST 2
BUILD_LIST
LOAD_CONST 3
LOAD_NAME param
BINARY_ADD
LOAD_CONST 4
BUILD_LIST
LOAD_CONST ('A', 'B')
BUILD_CONST_KEY_MAP
LOAD_NAME pd
LOAD_ATTR Index
LOAD_CONST 'x0'
LOAD... | param = 4
import pandas as pd
df = pd.DataFrame(
{"A": [1, 2], "B": [3 + param, 4]}, index=pd.Index(["x0", "x1"], name="x")
)
df
df.plot(kind="bar")
| data/jupytext-1.16.1/tests/data/notebooks/outputs/ipynb_to_percent/nteract_with_parameter.py | 146 | 80 | 312,556 |
LOAD_CONST 0
LOAD_CONST ('emit_usage',)
IMPORT_NAME whylogs.api.usage_stats
IMPORT_FROM emit_usage
STORE_NAME emit_usage
POP_TOP
LOAD_CONST 0
LOAD_CONST ('ConditionValidator',)
IMPORT_NAME whylogs.core.validators.condition_validator
IMPORT_FROM ConditionValidator
STORE_NAME ConditionValidator
POP_TOP
LOAD_CONST 0
LOA... | from whylogs.api.usage_stats import emit_usage
from whylogs.core.validators.condition_validator import ConditionValidator
from whylogs.core.validators.validator import Validator
__ALL__ = [
ConditionValidator,
Validator,
]
emit_usage("condition_validators")
| data/whylogs-1.3.24/whylogs/core/validators/__init__.py | 146 | 80 | 432,476 |
LOAD_CONST 0
LOAD_CONST ('absolute_import',)
IMPORT_NAME __future__
IMPORT_FROM absolute_import
STORE_NAME absolute_import
POP_TOP
LOAD_CONST 1
LOAD_CONST ('DesktopServiceClient',)
IMPORT_NAME desktop_service_client
IMPORT_FROM DesktopServiceClient
STORE_NAME DesktopServiceClient
POP_TOP
LOAD_CONST 1
LOAD_CONST ('Des... | from __future__ import absolute_import
from .desktop_service_client import DesktopServiceClient
from .desktop_service_client_composite_operations import (
DesktopServiceClientCompositeOperations,
)
from . import models
__all__ = ["DesktopServiceClient", "DesktopServiceClientCompositeOperations", "models"]
| data/oci-2.122.0/src/oci/desktops/__init__.py | 169 | 80 | 409,535 |
LOAD_CONST 0
LOAD_CONST ('nla',)
IMPORT_NAME pr2modules.netlink
IMPORT_FROM nla
STORE_NAME nla
POP_TOP
LOAD_CONST 0
LOAD_CONST ('TC_H_ROOT',)
IMPORT_NAME pr2modules.netlink.rtnl
IMPORT_FROM TC_H_ROOT
STORE_NAME TC_H_ROOT
POP_TOP
LOAD_NAME TC_H_ROOT
STORE_NAME parent
LOAD_BUILD_CLASS
LOAD_CONST <code object options a... | from pr2modules.netlink import nla
from pr2modules.netlink.rtnl import TC_H_ROOT
parent = TC_H_ROOT
class options(nla):
fields = (("limit", "i"),)
def get_parameters(kwarg):
return kwarg
| data/pyroute2.core-0.6.13/pr2modules/netlink/rtnl/tcmsg/sched_pfifo.py | 247 | 80 | 220,195 |
LOAD_CONST 0
LOAD_CONST ('__version__',)
IMPORT_NAME sp_api.__version__
IMPORT_FROM __version__
STORE_NAME __version__
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object BaseClient at 0x7fab41668810, file "f.py", line 4>
LOAD_CONST 'BaseClient'
MAKE_FUNCTION
LOAD_CONST 'BaseClient'
CALL_FUNCTION
STORE_NAME BaseClient
LO... | from sp_api.__version__ import __version__
class BaseClient:
scheme = "https://"
method = "GET"
content_type = "application/x-www-form-urlencoded;charset=UTF-8"
user_agent = f"python-sp-api-{__version__}"
| data/python-amazon-sp-api-1.4.0/sp_api/base/base_client.py | 193 | 80 | 358,247 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME coremltools
STORE_NAME ct
LOAD_CONST 0
LOAD_CONST ('backends_internal', 'clean_up_backends')
IMPORT_NAME coremltools.converters.mil.testing_reqs
IMPORT_FROM backends_internal
STORE_NAME backends_internal
IMPORT_FROM clean_up_backends
STORE_NAME clean_up_backends
POP_TOP
LOAD_N... | import coremltools as ct
from coremltools.converters.mil.testing_reqs import backends_internal, clean_up_backends
backends = clean_up_backends(
backends_internal, ct.target.iOS14, force_include_iOS15_test=True
)
| data/coremltools-7.1/coremltools/converters/mil/mil/ops/tests/iOS14/__init__.py | 140 | 80 | 447,240 |
LOAD_CONST 1
LOAD_CONST ('ComputeManagementClient',)
IMPORT_NAME _compute_management_client
IMPORT_FROM ComputeManagementClient
STORE_NAME ComputeManagementClient
POP_TOP
LOAD_CONST 'ComputeManagementClient'
BUILD_LIST
STORE_NAME __all__
SETUP_EXCEPT to 42
LOAD_CONST 1
LOAD_CONST ('patch_sdk',)
IMPORT_NAME _patch
IM... | from ._compute_management_client import ComputeManagementClient
__all__ = ["ComputeManagementClient"]
try:
from ._patch import patch_sdk # type: ignore
patch_sdk()
except ImportError:
pass
from ._version import VERSION
__version__ = VERSION
| data/azure-mgmt-compute-30.5.0/azure/mgmt/compute/__init__.py | 178 | 80 | 394,455 |
LOAD_CONST 0
LOAD_CONST ('Enum',)
IMPORT_NAME enum
IMPORT_FROM Enum
STORE_NAME Enum
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object ReportStatus at 0x7fab41668150, file "f.py", line 4>
LOAD_CONST 'ReportStatus'
MAKE_FUNCTION
LOAD_CONST 'ReportStatus'
LOAD_NAME str
LOAD_NAME Enum
CALL_FUNCTION
STORE_NAME ReportStatus
... | from enum import Enum
class ReportStatus(str, Enum):
CANCELLED = "CANCELLED"
DONE = "DONE"
FATAL = "FATAL"
IN_PROGRESS = "IN_PROGRESS"
IN_QUEUE = "IN_QUEUE"
| data/python-amazon-sp-api-1.4.0/sp_api/base/report_status.py | 190 | 80 | 358,246 |
LOAD_CONST 0
LOAD_CONST ('kernels', 'likelihoods', 'models', 'parameterized', 'util')
IMPORT_NAME pyro.contrib.gp
IMPORT_FROM kernels
STORE_NAME kernels
IMPORT_FROM likelihoods
STORE_NAME likelihoods
IMPORT_FROM models
STORE_NAME models
IMPORT_FROM parameterized
STORE_NAME parameterized
IMPORT_FROM util
STORE_NAME util... | from pyro.contrib.gp import kernels, likelihoods, models, parameterized, util
from pyro.contrib.gp.parameterized import Parameterized
__all__ = [
"Parameterized",
"kernels",
"likelihoods",
"models",
"parameterized",
"util",
]
| data/pyro-ppl-1.9.0/pyro/contrib/gp/__init__.py | 157 | 80 | 109,932 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME zmq.backend
STORE_NAME zmq
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME zmq.sugar
STORE_NAME zmq
LOAD_NAME print
LOAD_CONST 'from .backend import ('
CALL_FUNCTION
POP_TOP
SETUP_LOOP to 82
LOAD_NAME sorted
LOAD_NAME set
LOAD_NAME zmq
LOAD_ATTR backend
LOAD_ATTR __all__
CALL_FUNCTI... | import zmq.backend
import zmq.sugar
print("from .backend import (")
for name in sorted(set(zmq.backend.__all__).difference(zmq.sugar.__all__)):
print(f" {name} as {name},")
print(")")
| data/pyzmq-25.1.2/tools/backend_imports.py | 171 | 80 | 419,700 |
LOAD_CONST 0
LOAD_CONST ('absolute_import',)
IMPORT_NAME __future__
IMPORT_FROM absolute_import
STORE_NAME absolute_import
POP_TOP
LOAD_CONST 1
LOAD_CONST ('VbInstanceClient',)
IMPORT_NAME vb_instance_client
IMPORT_FROM VbInstanceClient
STORE_NAME VbInstanceClient
POP_TOP
LOAD_CONST 1
LOAD_CONST ('VbInstanceClientCom... | from __future__ import absolute_import
from .vb_instance_client import VbInstanceClient
from .vb_instance_client_composite_operations import VbInstanceClientCompositeOperations
from . import models
__all__ = ["VbInstanceClient", "VbInstanceClientCompositeOperations", "models"]
| data/oci-2.122.0/src/oci/visual_builder/__init__.py | 179 | 80 | 409,567 |
LOAD_CONST 0
LOAD_CONST ('download_raw_files',)
IMPORT_NAME opendatasets.utils
IMPORT_FROM download_raw_files
STORE_NAME download_raw_files
POP_TOP
LOAD_CONST <code object download at 0x7f8e2fe58930, file "f.py", line 4>
LOAD_CONST 'download'
MAKE_FUNCTION
STORE_NAME download
LOAD_CONST None
RETURN_VALUE
LOAD_GLOBAL ... | from opendatasets.utils import download_raw_files
def download(dataset_id, data_dir, dry_run, **kwargs):
download_raw_files(
dataset_id, data_dir, dry_run, ["state-of-javascript-2019-responses.ndjson"]
)
| data/opendatasets-0.1.22/opendatasets/datasets/state-of-javascript-2019/__init__.py | 142 | 80 | 405,152 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME functools
STORE_NAME functools
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME hashlib
STORE_NAME hashlib
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME sys
STORE_NAME sys
LOAD_NAME sys
LOAD_ATTR version_info
LOAD_CONST (3, 9)
COMPARE_OP >=
POP_JUMP_IF_FALSE
LOAD_CONST 'usedforsecurity'
L... | import functools
import hashlib
import sys
_kwargs = {"usedforsecurity": False} if sys.version_info >= (3, 9) else {}
md5 = functools.partial(hashlib.md5, **_kwargs)
sha1 = functools.partial(hashlib.sha1, **_kwargs)
| data/mlflow-skinny-2.10.2/mlflow/utils/insecure_hash.py | 156 | 80 | 291,625 |
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME azure.cli.core.aaz
IMPORT_STAR
LOAD_NAME register_command_group
LOAD_CONST 'afd log-analytic location'
CALL_FUNCTION
LOAD_BUILD_CLASS
LOAD_CONST <code object __CMDGroup at 0x7fab823bedb0, file "f.py", line 4>
LOAD_CONST '__CMDGroup'
MAKE_FUNCTION
LOAD_CONST '__CMDGroup'
LOA... | from azure.cli.core.aaz import *
@register_command_group(
"afd log-analytic location",
)
class __CMDGroup(AAZCommandGroup):
"""Manage available location names for AFD log analysis."""
pass
__all__ = ["__CMDGroup"]
| data/azure-cli-2.57.0/azure/cli/command_modules/cdn/aaz/latest/afd/log_analytic/location/__cmd_group.py | 188 | 80 | 375,223 |
LOAD_CONST 'InternalError.CacheErr'
STORE_NAME INTERNALERROR_CACHEERR
LOAD_CONST 'InternalError.LocalErr'
STORE_NAME INTERNALERROR_LOCALERR
LOAD_CONST 'InternalError.Timeout'
STORE_NAME INTERNALERROR_TIMEOUT
LOAD_CONST 'InvalidParameter'
STORE_NAME INVALIDPARAMETER
LOAD_CONST 'LimitExceeded'
STORE_NAME LIMITEXCEEDE... | INTERNALERROR_CACHEERR = "InternalError.CacheErr"
INTERNALERROR_LOCALERR = "InternalError.LocalErr"
INTERNALERROR_TIMEOUT = "InternalError.Timeout"
INVALIDPARAMETER = "InvalidParameter"
LIMITEXCEEDED = "LimitExceeded"
| data/tencentcloud-sdk-python-3.0.1092/tencentcloud/tics/v20181115/errorcodes.py | 93 | 80 | 119,044 |
LOAD_CONST 1
LOAD_CONST ('helps',)
IMPORT_NAME generated._help
IMPORT_FROM helps
STORE_NAME helps
POP_TOP
SETUP_EXCEPT to 30
LOAD_CONST 1
LOAD_CONST ('helps',)
IMPORT_NAME manual._help
IMPORT_FROM helps
STORE_NAME helps
POP_TOP
POP_BLOCK
JUMP_FORWARD to 82
DUP_TOP
LOAD_NAME ImportError
COMPARE_OP exception match
POP... | from .generated._help import helps # pylint: disable=reimported
try:
from .manual._help import helps # pylint: disable=reimported
except ImportError as e:
if e.name.endswith("manual._help"):
pass
else:
raise e
| data/azure-cli-2.57.0/azure/cli/command_modules/synapse/_help.py | 158 | 80 | 375,456 |
LOAD_CONST 'Provide cached_property for Python <=3.8 programs.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME functools
STORE_NAME functools
SETUP_EXCEPT to 30
LOAD_CONST 0
LOAD_CONST ('cached_property',)
IMPORT_NAME functools
IMPORT_FROM cached_property
STORE_NAME cached_property
POP_TOP
POP_BLOCK
JUM... | """Provide cached_property for Python <=3.8 programs."""
import functools
try:
from functools import cached_property # pylint: disable=unused-import
except ImportError:
def cached_property(fn):
return property(functools.lru_cache()(fn))
| data/podman-4.9.0/podman/api/cached_property.py | 195 | 80 | 190,403 |
LOAD_CONST 'AUTOGENERATED. DO NOT EDIT.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('MobileNet',)
IMPORT_NAME tf_keras.src.applications.mobilenet
IMPORT_FROM MobileNet
STORE_NAME MobileNet
POP_TOP
LOAD_CONST 0
LOAD_CONST ('decode_predictions',)
IMPORT_NAME tf_keras.src.applications.mobilenet
IMPORT_FROM decode_predi... | """AUTOGENERATED. DO NOT EDIT."""
from tf_keras.src.applications.mobilenet import MobileNet
from tf_keras.src.applications.mobilenet import decode_predictions
from tf_keras.src.applications.mobilenet import preprocess_input
| data/tf_keras-nightly-2.17.0.dev2024022110/tf_keras/api/_v1/keras/applications/mobilenet/__init__.py | 155 | 80 | 165,202 |
LOAD_CONST 'RGB endpoint module.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('annotations',)
IMPORT_NAME __future__
IMPORT_FROM annotations
STORE_NAME annotations
POP_TOP
LOAD_CONST 0
LOAD_CONST ('Any',)
IMPORT_NAME typing
IMPORT_FROM Any
STORE_NAME Any
POP_TOP
LOAD_CONST 'str'
LOAD_CONST 'dict[str, Any]'
LOAD_CONS... | """RGB endpoint module."""
from __future__ import annotations
from typing import Any
def read(content: str) -> dict[str, Any]: # pylint: disable=unused-argument
"""Read RGB data"""
rgb: dict[str, Any] = {}
return rgb
| data/asusrouter-1.7.0/asusrouter/modules/endpoint/rgb.py | 149 | 80 | 10,624 |
LOAD_CONST 0
STORE_NAME VERSION_MAJOR
LOAD_CONST 9
STORE_NAME VERSION_MINOR
LOAD_CONST 0
STORE_NAME VERSION_BUILD
LOAD_NAME VERSION_MAJOR
LOAD_NAME VERSION_MINOR
LOAD_NAME VERSION_BUILD
BUILD_TUPLE
STORE_NAME VERSION_INFO
LOAD_CONST '%d.%d.%d'
LOAD_NAME VERSION_INFO
BINARY_MODULO
STORE_NAME VERSION_STRING
LOAD_NAM... | VERSION_MAJOR = 0
VERSION_MINOR = 9
VERSION_BUILD = 0
VERSION_INFO = (VERSION_MAJOR, VERSION_MINOR, VERSION_BUILD)
VERSION_STRING = "%d.%d.%d" % VERSION_INFO
__version__ = VERSION_INFO
| data/wandb-0.16.3/wandb/vendor/watchdog_0_9_0/wandb_watchdog/version.py | 120 | 80 | 360,726 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME abc
STORE_NAME abc
LOAD_CONST 0
LOAD_CONST ('Optional',)
IMPORT_NAME typing
IMPORT_FROM Optional
STORE_NAME Optional
POP_TOP
LOAD_CONST 2
LOAD_CONST ('UserOptions',)
IMPORT_NAME user_options
IMPORT_FROM UserOptions
STORE_NAME UserOptions
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <c... | import abc
from typing import Optional
from ..user_options import UserOptions
class UserOptionReader(abc.ABC):
@abc.abstractmethod
def read_options(
self, config_filename: Optional[str] = None
) -> Optional[UserOptions]:
raise NotImplementedError
| data/import-linter-2.0/src/importlinter/application/ports/user_options.py | 279 | 80 | 136,486 |
LOAD_CONST 'For when pip wants to check the date or time.\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME datetime
STORE_NAME datetime
LOAD_NAME int
LOAD_NAME int
LOAD_NAME int
LOAD_NAME bool
LOAD_CONST ('year', 'month', 'day', 'return')
BUILD_CONST_KEY_MAP
LOAD_CONST <code object today_is_later_than a... | """For when pip wants to check the date or time.
"""
import datetime
def today_is_later_than(year: int, month: int, day: int) -> bool:
today = datetime.date.today()
given = datetime.date(year, month, day)
return today > given
| data/pigar-2.1.3/pigar/_vendor/pip/_internal/utils/datetime.py | 178 | 80 | 241,422 |
LOAD_CONST 'For when pip wants to check the date or time.\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME datetime
STORE_NAME datetime
LOAD_NAME int
LOAD_NAME int
LOAD_NAME int
LOAD_NAME bool
LOAD_CONST ('year', 'month', 'day', 'return')
BUILD_CONST_KEY_MAP
LOAD_CONST <code object today_is_later_than a... | """For when pip wants to check the date or time.
"""
import datetime
def today_is_later_than(year: int, month: int, day: int) -> bool:
today = datetime.date.today()
given = datetime.date(year, month, day)
return today > given
| data/pipenv-2023.12.1/pipenv/patched/pip/_internal/utils/datetime.py | 178 | 80 | 402,311 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME sys
STORE_NAME sys
LOAD_CONST 0
LOAD_CONST ('set_default_config',)
IMPORT_NAME prettyprinter
IMPORT_FROM set_default_config
STORE_NAME set_default_config
POP_TOP
LOAD_NAME set_default_config
LOAD_CONST True
LOAD_CONST ('sort_dict_keys',)
CALL_FUNCTION
POP_TOP
LOAD_CONST <code... | import sys
from prettyprinter import set_default_config
set_default_config(sort_dict_keys=True)
def pytest_ignore_collect(path, config):
if "test_dataclasses" in str(path) and sys.version_info < (3, 7):
return True
| data/prettyprinter-0.18.0/tests/conftest.py | 193 | 80 | 70,055 |
LOAD_CONST <code object test_get_url at 0x7fab821841e0, file "f.py", line 1>
LOAD_CONST 'test_get_url'
MAKE_FUNCTION
STORE_NAME test_get_url
LOAD_CONST None
RETURN_VALUE
LOAD_FAST make_dataset
LOAD_CONST False
LOAD_CONST True
LOAD_CONST False
LOAD_CONST False
LOAD_CONST False
LOAD_CONST ('cache', 'files', 'dvcfile', ... | def test_get_url(bench_dvc, tmp_dir, scm, dvc, make_dataset):
dataset = make_dataset(
cache=False, files=True, dvcfile=False, commit=False, remote=False
)
bench_dvc("get-url", str(dataset), "new")
| data/dvc-3.47.0/dvc/testing/benchmarks/cli/commands/test_get_url.py | 141 | 80 | 318,078 |
LOAD_CONST 0
LOAD_CONST ('unicode_literals', 'division', 'absolute_import', 'print_function')
IMPORT_NAME __future__
IMPORT_FROM unicode_literals
STORE_NAME unicode_literals
IMPORT_FROM division
STORE_NAME division
IMPORT_FROM absolute_import
STORE_NAME absolute_import
IMPORT_FROM print_function
STORE_NAME print_functi... | from __future__ import unicode_literals, division, absolute_import, print_function
from powerline.bindings.vim import list_tabpages
def single_tab(pl, segment_info, mode):
"""Returns True if Vim has only one tab opened"""
return len(list_tabpages()) == 1
| data/powerline-status-2.7/powerline/selectors/vim.py | 189 | 80 | 343,296 |
LOAD_CONST 0
LOAD_CONST ('Any',)
IMPORT_NAME typing
IMPORT_FROM Any
STORE_NAME Any
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object ExceptionWithDisplayMessage at 0x7fa6d804aed0, file "f.py", line 4>
LOAD_CONST 'ExceptionWithDisplayMessage'
MAKE_FUNCTION
LOAD_CONST 'ExceptionWithDisplayMessage'
LOAD_NAME Exception
CAL... | from typing import Any
class ExceptionWithDisplayMessage(Exception):
"""
Exception that can be used to display a custom message to the user.
"""
def __init__(self, display_message: str, **kwargs: Any):
super().__init__(**kwargs)
self.display_message = display_message
| data/airbyte-cdk-0.64.0/airbyte_cdk/sources/streams/concurrent/exceptions.py | 277 | 80 | 52,035 |
LOAD_CONST '\nModule to store compatiblity imports to prevent Django deprecation warnings.\n'
STORE_NAME __doc__
SETUP_EXCEPT to 22
LOAD_CONST 0
LOAD_CONST ('url_has_allowed_host_and_scheme',)
IMPORT_NAME django.utils.http
IMPORT_FROM url_has_allowed_host_and_scheme
STORE_NAME url_has_allowed_host_and_scheme
POP_TOP
... | """
Module to store compatiblity imports to prevent Django deprecation warnings.
"""
try:
from django.utils.http import url_has_allowed_host_and_scheme
except ImportError:
from django.utils.http import is_safe_url as url_has_allowed_host_and_scheme
| data/django-contrib-comments-2.2.0/django_comments/compat.py | 179 | 80 | 418,960 |
LOAD_CONST 0
LOAD_CONST ('LivySession',)
IMPORT_NAME livy.session
IMPORT_FROM LivySession
STORE_NAME LivySession
POP_TOP
LOAD_CONST 0
LOAD_CONST ('LivyBatch',)
IMPORT_NAME livy.batch
IMPORT_FROM LivyBatch
STORE_NAME LivyBatch
POP_TOP
LOAD_CONST 0
LOAD_CONST ('SessionKind', 'SessionState', 'SparkRuntimeError')
IMPORT_... | from livy.session import LivySession # noqa: F401
from livy.batch import LivyBatch # noqa: F401
from livy.models import ( # noqa: F401
SessionKind,
SessionState,
SparkRuntimeError,
)
| data/livy-0.8.0/livy/__init__.py | 129 | 80 | 121,137 |
LOAD_CONST "Dataset definition for asset.\n\nDEPRECATED!\nIf you want to use the Asset dataset builder class, use:\ntfds.builder_cls('asset')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_builder_import
POP_TO... | """Dataset definition for asset.
DEPRECATED!
If you want to use the Asset dataset builder class, use:
tfds.builder_cls('asset')
"""
from tensorflow_datasets.core import lazy_builder_import
Asset = lazy_builder_import.LazyBuilderImport("asset")
| data/tfds-nightly-4.9.4.dev202402210044/tensorflow_datasets/text_simplification/asset/asset.py | 124 | 80 | 238,344 |
LOAD_CONST 0
LOAD_CONST ('NotImported',)
IMPORT_NAME great_expectations.compatibility.not_imported
IMPORT_FROM NotImported
STORE_NAME NotImported
POP_TOP
LOAD_NAME NotImported
LOAD_CONST "py4j is not installed, please 'pip install py4j'"
CALL_FUNCTION
STORE_NAME PY4J_NOT_IMPORTED
SETUP_EXCEPT to 38
LOAD_CONST 0
LOAD... | from great_expectations.compatibility.not_imported import NotImported
PY4J_NOT_IMPORTED = NotImported("py4j is not installed, please 'pip install py4j'")
try:
from py4j import protocol
except ImportError:
protocol = PY4J_NOT_IMPORTED
| data/great_expectations-0.18.9/great_expectations/compatibility/py4j.py | 163 | 80 | 18,881 |
LOAD_CONST '\nTask for Scheduler and AsyncScheduler\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('Any', 'Callable', 'Optional')
IMPORT_NAME typing
IMPORT_FROM Any
STORE_NAME Any
IMPORT_FROM Callable
STORE_NAME Callable
IMPORT_FROM Optional
STORE_NAME Optional
POP_TOP
LOAD_CONST 0
LOAD_CONST ('dataclass', 'field')
I... | """
Task for Scheduler and AsyncScheduler
"""
from typing import Any, Callable, Optional
from dataclasses import dataclass, field
@dataclass(order=True)
class Task:
time: float
action: Optional[Callable[[], Any]] = field(compare=False)
| data/psycopg-pool-3.2.1/psycopg_pool/_task.py | 274 | 80 | 235,965 |
LOAD_CONST <code object _get_system_version at 0x7fab4106cb70, file "f.py", line 1>
LOAD_CONST '_get_system_version'
MAKE_FUNCTION
STORE_NAME _get_system_version
LOAD_CONST None
RETURN_VALUE
LOAD_GLOBAL __file__
LOAD_CONST None
COMPARE_OP is
POP_JUMP_IF_FALSE
SETUP_EXCEPT to 18
LOAD_CONST 5
STORE_FAST m
POP_BLOCK
JU... | def _get_system_version():
if __file__ is None:
try:
m = 5
except IOError:
pass
else:
try:
m = 10
finally:
m = 15
if m is not None:
m = 20
return m
| data/xdis-6.0.5/test/simple_source/bug33/01_try_except.py | 187 | 80 | 85,408 |
LOAD_CONST 1
LOAD_CONST ('libsvm',)
IMPORT_NAME
IMPORT_FROM libsvm
STORE_NAME libsvm
POP_TOP
LOAD_CONST 1
LOAD_CONST ('sklearn',)
IMPORT_NAME
IMPORT_FROM sklearn
STORE_NAME sklearn
POP_TOP
LOAD_CONST 1
LOAD_CONST ('xgboost',)
IMPORT_NAME
IMPORT_FROM xgboost
STORE_NAME xgboost
POP_TOP
LOAD_CONST 1
LOAD_CONST ('conver... | from . import libsvm
from . import sklearn
from . import xgboost
from ._converters_entry import convert
from .mil import (
ClassifierConfig,
ColorLayout,
TensorType,
ImageType,
RangeDim,
Shape,
EnumeratedShapes,
)
| data/coremltools-7.1/coremltools/converters/__init__.py | 218 | 80 | 447,099 |
LOAD_CONST 0
LOAD_CONST ('Function',)
IMPORT_NAME pypika.terms
IMPORT_FROM Function
STORE_NAME Function
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object Random at 0x7fab8234aed0, file "f.py", line 4>
LOAD_CONST 'Random'
MAKE_FUNCTION
LOAD_CONST 'Random'
LOAD_NAME Function
CALL_FUNCTION
STORE_NAME Random
LOAD_CONST Non... | from pypika.terms import Function
class Random(Function): # type: ignore
"""
Generate random number.
:samp:`Random()`
"""
def __init__(self, alias=None) -> None:
super().__init__("RANDOM", alias=alias)
| data/tortoise_orm-0.20.0/tortoise/contrib/sqlite/functions.py | 259 | 80 | 144,352 |
LOAD_CONST 0
LOAD_CONST ('auto',)
IMPORT_NAME enum
IMPORT_FROM auto
STORE_NAME auto
POP_TOP
LOAD_CONST 0
LOAD_CONST ('JsonProperty',)
IMPORT_NAME spdx_tools.spdx.jsonschema.json_property
IMPORT_FROM JsonProperty
STORE_NAME JsonProperty
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object ExternalPackageRefProperty at 0x7... | from enum import auto
from spdx_tools.spdx.jsonschema.json_property import JsonProperty
class ExternalPackageRefProperty(JsonProperty):
COMMENT = auto()
REFERENCE_CATEGORY = auto()
REFERENCE_LOCATOR = auto()
REFERENCE_TYPE = auto()
| data/spdx-tools-0.8.2/src/spdx_tools/spdx/jsonschema/external_package_ref_properties.py | 217 | 80 | 286,060 |
LOAD_BUILD_CLASS
LOAD_CONST <code object OnlineMeetingProviderType at 0x7fab701acc00, file "f.py", line 1>
LOAD_CONST 'OnlineMeetingProviderType'
MAKE_FUNCTION
LOAD_CONST 'OnlineMeetingProviderType'
CALL_FUNCTION
STORE_NAME OnlineMeetingProviderType
LOAD_CONST None
RETURN_VALUE
LOAD_NAME __name__
STORE_NAME __module__... | class OnlineMeetingProviderType:
"""Specifies the type of a principal."""
def __init__(self):
pass
unknown = 0
""""""
skypeForBusiness = 1
""""""
skypeForConsumer = 2
""""""
teamsForBusiness = 3
""""""
| data/Office365-REST-Python-Client-2.5.5/office365/communications/onlinemeetings/provider_type.py | 218 | 80 | 188,651 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME sys
STORE_NAME sys
LOAD_CONST 0
LOAD_CONST ('dx_build_app',)
IMPORT_NAME dxpy.scripts
IMPORT_FROM dx_build_app
STORE_NAME dx_build_app
POP_TOP
LOAD_CONST <code object main at 0x7faa5e9e9150, file "f.py", line 5>
LOAD_CONST 'main'
MAKE_FUNCTION
STORE_NAME main
LOAD_NAME __name... | import sys
from dxpy.scripts import dx_build_app
def main():
sys.argv = [sys.argv[0]] + ["--create-applet"] + sys.argv[1:]
dx_build_app.main()
if __name__ == "__main__":
main()
| data/dxpy-0.369.1/dxpy/scripts/dx_build_applet.py | 200 | 80 | 441,121 |
LOAD_CONST 'faster-rcnn_r50_fpn_32xb2-1x_openimages-challenge.py'
BUILD_LIST
STORE_NAME _base_
LOAD_NAME dict
LOAD_NAME dict
LOAD_CONST True
LOAD_CONST 'ClassAwareSampler'
LOAD_CONST 1
LOAD_CONST ('_delete_', 'type', 'num_sample_class')
CALL_FUNCTION
LOAD_CONST ('sampler',)
CALL_FUNCTION
STORE_NAME train_dataloader
L... | _base_ = ["faster-rcnn_r50_fpn_32xb2-1x_openimages-challenge.py"]
train_dataloader = dict(
sampler=dict(_delete_=True, type="ClassAwareSampler", num_sample_class=1)
)
| data/mmdet-3.3.0/mmdet/.mim/configs/openimages/faster-rcnn_r50_fpn_32xb2-cas-1x_openimages-challenge.py | 106 | 80 | 151,352 |
LOAD_CONST 2
LOAD_CONST ('_utilities',)
IMPORT_NAME
IMPORT_FROM _utilities
STORE_NAME _utilities
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME typing
STORE_NAME typing
LOAD_CONST 1
LOAD_CONST ('*',)
IMPORT_NAME get_tensorflow_versions
IMPORT_STAR
LOAD_CONST 1
LOAD_CONST ('*',)
IMPORT_NAME get_v2_accelerator_types... | from .. import _utilities
import typing
from .get_tensorflow_versions import *
from .get_v2_accelerator_types import *
from .get_v2_runtime_versions import *
from .node import *
from .v2_vm import *
from ._inputs import *
from . import outputs
| data/pulumi_gcp-7.9.0/pulumi_gcp/tpu/__init__.py | 190 | 80 | 445,731 |
LOAD_CONST 0
LOAD_CONST ('current_audit_info',)
IMPORT_NAME prowler.providers.aws.lib.audit_info.audit_info
IMPORT_FROM current_audit_info
STORE_NAME current_audit_info
POP_TOP
LOAD_CONST 0
LOAD_CONST ('WAFv2',)
IMPORT_NAME prowler.providers.aws.services.wafv2.wafv2_service
IMPORT_FROM WAFv2
STORE_NAME WAFv2
POP_TOP
... | from prowler.providers.aws.lib.audit_info.audit_info import current_audit_info
from prowler.providers.aws.services.wafv2.wafv2_service import WAFv2
wafv2_client = WAFv2(current_audit_info)
| data/prowler-3.14.0/prowler/providers/aws/services/wafv2/wafv2_client.py | 137 | 80 | 183,060 |
LOAD_CONST <code object _get_system_version at 0x7fab4106cc00, file "f.py", line 1>
LOAD_CONST '_get_system_version'
MAKE_FUNCTION
STORE_NAME _get_system_version
LOAD_CONST None
RETURN_VALUE
LOAD_GLOBAL __file__
LOAD_CONST None
COMPARE_OP is
POP_JUMP_IF_FALSE
SETUP_EXCEPT to 18
LOAD_CONST 5
STORE_FAST m
POP_BLOCK
JU... | def _get_system_version():
if __file__ is None:
try:
m = 5
except IOError:
pass
else:
try:
m = 10
finally:
m = 15
if m is not None:
m = 20
return m
| data/xdis-6.0.5/test/simple_source/bug33/01_if_try_except.py | 187 | 80 | 85,414 |
LOAD_CONST 1
LOAD_CONST ('LFSClient',)
IMPORT_NAME client
IMPORT_FROM LFSClient
STORE_NAME LFSClient
POP_TOP
LOAD_CONST 1
LOAD_CONST ('LFSError',)
IMPORT_NAME exceptions
IMPORT_FROM LFSError
STORE_NAME LFSError
POP_TOP
LOAD_CONST 1
LOAD_CONST ('fetch',)
IMPORT_NAME fetch
IMPORT_FROM fetch
STORE_NAME fetch
POP_TOP
LO... | from .client import LFSClient
from .exceptions import LFSError
from .fetch import fetch
from .pointer import Pointer
from .smudge import smudge
from .storage import LFSStorage
__all__ = ["LFSClient", "LFSError", "LFSStorage", "Pointer", "fetch", "smudge"]
| data/scmrepo-3.1.0/src/scmrepo/git/lfs/__init__.py | 209 | 80 | 187,877 |
LOAD_CONST 0
LOAD_CONST ('Enum',)
IMPORT_NAME enum
IMPORT_FROM Enum
STORE_NAME Enum
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object VulnerabilitySeverity at 0x7fab70027660, file "f.py", line 4>
LOAD_CONST 'VulnerabilitySeverity'
MAKE_FUNCTION
LOAD_CONST 'VulnerabilitySeverity'
LOAD_NAME str
LOAD_NAME Enum
CALL_FUNCTI... | from enum import Enum
class VulnerabilitySeverity(str, Enum):
None_ = ("none",)
Low = ("low",)
Medium = ("medium",)
High = ("high",)
Critical = ("critical",)
UnknownFutureValue = ("unknownFutureValue",)
| data/msgraph-sdk-1.1.0/msgraph/generated/models/security/vulnerability_severity.py | 212 | 80 | 262,277 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME os
STORE_NAME os
SETUP_EXCEPT to 34
LOAD_CONST 0
LOAD_CONST ('DeckGLWidget',)
IMPORT_NAME pydeck.widget
IMPORT_FROM DeckGLWidget
STORE_NAME DeckGLWidget
POP_TOP
POP_BLOCK
JUMP_FORWARD to 72
DUP_TOP
LOAD_NAME M... | import pytest
import os
try:
from pydeck.widget import DeckGLWidget
except ModuleNotFoundError:
import warnings
warnings.warn("Widget test will fail")
def test_example_creation_blank():
w = DeckGLWidget()
assert w.json_input == ""
| data/pydeck-0.8.0/tests/test_deckgl_widget.py | 253 | 80 | 347,997 |
LOAD_CONST <code object method1 at 0x7fa6dbb668a0, file "f.py", line 1>
LOAD_CONST 'method1'
MAKE_FUNCTION
STORE_NAME method1
LOAD_CONST <code object method2 at 0x7fa6a68f0ae0, file "f.py", line 5>
LOAD_CONST 'method2'
MAKE_FUNCTION
STORE_NAME method2
LOAD_NAME method1
LOAD_NAME method2
LOAD_CONST ('method1', 'metho... | def method1():
pass
def method2():
pass
METHODS = {
"method1": method1,
"method2": method2,
}
class Class:
pass
for name, method in METHODS.items():
setattr(Class, name, method)
| data/pytkdocs-0.16.1/tests/fixtures/dynamic_members.py | 269 | 80 | 46,835 |
LOAD_CONST 0
LOAD_CONST ('Enum',)
IMPORT_NAME enum
IMPORT_FROM Enum
STORE_NAME Enum
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object S3DataNaming at 0x7f8ed39ed6f0, file "f.py", line 4>
LOAD_CONST 'S3DataNaming'
MAKE_FUNCTION
LOAD_CONST 'S3DataNaming'
LOAD_NAME Enum
CALL_FUNCTION
STORE_NAME S3DataNaming
LOAD_CONST Non... | from enum import Enum
class S3DataNaming(Enum):
UNIQUE = "unique"
TABLE = "table"
TABLE_UNIQUE = "table_unique"
SCHEMA_TABLE = "schema_table"
SCHEMA_TABLE_UNIQUE = "schema_table_unique"
| data/dbt-athena-community-1.7.2/dbt/adapters/athena/s3.py | 201 | 80 | 113,168 |
LOAD_CONST 0
LOAD_CONST ('SSOError',)
IMPORT_NAME botocore.exceptions
IMPORT_FROM SSOError
STORE_NAME SSOError
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object PendingAuthorizationExpiredError at 0x7f8e2fcbf300, file "f.py", line 4>
LOAD_CONST 'PendingAuthorizationExpiredError'
MAKE_FUNCTION
LOAD_CONST 'PendingAuthori... | from botocore.exceptions import SSOError
class PendingAuthorizationExpiredError(SSOError):
fmt = (
"The pending authorization to retrieve an Identity Center token has expired."
"The device authorization flow to retrieve an Identity Center token must be "
"restarted."
)
| data/aws_sso_lib-1.14.0/aws_sso_lib/vendored_botocore/exceptions.py | 189 | 80 | 8,838 |
LOAD_CONST 0
LOAD_CONST ('TestClient',)
IMPORT_NAME fastapi.testclient
IMPORT_FROM TestClient
STORE_NAME TestClient
POP_TOP
LOAD_CONST 0
LOAD_CONST ('app',)
IMPORT_NAME docs_src.custom_response.tutorial009c
IMPORT_FROM app
STORE_NAME app
POP_TOP
LOAD_NAME TestClient
LOAD_NAME app
CALL_FUNCTION
STORE_NAME client
LOAD... | from fastapi.testclient import TestClient
from docs_src.custom_response.tutorial009c import app
client = TestClient(app)
def test_get():
response = client.get("/")
assert response.content == b'{\n "message": "Hello World"\n}'
| data/fastapi-0.109.2/tests/test_tutorial/test_custom_response/test_tutorial009c.py | 187 | 80 | 141,961 |
LOAD_CONST 0
LOAD_CONST ('print_function',)
IMPORT_NAME __future__
IMPORT_FROM print_function
STORE_NAME print_function
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pyeapi
STORE_NAME pyeapi
LOAD_NAME pyeapi
LOAD_ATTR connect
LOAD_CONST '192.168.1.16'
LOAD_CONST ('host',)
CALL_FUNCTION
STORE_NAME connection
LOAD_... | from __future__ import print_function
import pyeapi
connection = pyeapi.connect(host="192.168.1.16")
output = connection.execute(["enable", "show version"])
print(("My system MAC address is", output["result"][1]["systemMacAddress"]))
| data/pyeapi-1.0.2/examples/simple.py | 157 | 80 | 439,765 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME unittest
STORE_NAME unittest
LOAD_CONST 0
LOAD_CONST ('AzureConfiguration',)
IMPORT_NAME msrestazure.azure_configuration
IMPORT_FROM AzureConfiguration
STORE_NAME AzureConfiguration
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object TestAzConfiguration at 0x7faa74652d20, file "f... | import unittest
from msrestazure.azure_configuration import AzureConfiguration
class TestAzConfiguration(unittest.TestCase):
def test_config_basic(self):
AzureConfiguration("http://management.something.com")
if __name__ == "__main__":
unittest.main()
| data/msrestazure-0.6.4/tests/test_configuration.py | 269 | 80 | 244,471 |
LOAD_CONST 0
LOAD_CONST ('print_function',)
IMPORT_NAME __future__
IMPORT_FROM print_function
STORE_NAME print_function
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pyeapi
STORE_NAME pyeapi
LOAD_NAME pyeapi
LOAD_METHOD load_config
LOAD_CONST 'nodes.conf'
CALL_METHOD
POP_TOP
LOAD_NAME pyeapi
LOAD_METHOD connect_t... | from __future__ import print_function
import pyeapi
pyeapi.load_config("nodes.conf")
node = pyeapi.connect_to("veos01")
output = node.enable("show version")
print(("My System MAC address is", output[0]["result"]["systemMacAddress"]))
| data/pyeapi-1.0.2/examples/sysmac.py | 160 | 80 | 439,762 |
LOAD_CONST '\nShared test utilities.\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('NAME_TO_LEVEL',)
IMPORT_NAME structlog._log_levels
IMPORT_FROM NAME_TO_LEVEL
STORE_NAME NAME_TO_LEVEL
POP_TOP
LOAD_CONST <code object <listcomp> at 0x7f8ab6189930, file "f.py", line 8>
LOAD_CONST '<listcomp>'
MAKE_FUNCTION
LOAD_NAME ... | """
Shared test utilities.
"""
from structlog._log_levels import NAME_TO_LEVEL
stdlib_log_methods = [m for m in NAME_TO_LEVEL if m != "notset"]
class CustomError(Exception):
"""
Custom exception for testing purposes.
"""
| data/structlog-24.1.0/tests/utils.py | 276 | 80 | 430,098 |
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME azure.cli.core.aaz
IMPORT_STAR
LOAD_NAME register_command_group
LOAD_CONST 'network vpn-connection ipsec-policy'
CALL_FUNCTION
LOAD_BUILD_CLASS
LOAD_CONST <code object __CMDGroup at 0x7fab82197d20, file "f.py", line 4>
LOAD_CONST '__CMDGroup'
MAKE_FUNCTION
LOAD_CONST '__CMDG... | from azure.cli.core.aaz import *
@register_command_group(
"network vpn-connection ipsec-policy",
)
class __CMDGroup(AAZCommandGroup):
"""Manage VPN connection IPSec policies."""
pass
__all__ = ["__CMDGroup"]
| data/azure-cli-2.57.0/azure/cli/command_modules/network/aaz/profile_2018_03_01_hybrid/network/vpn_connection/ipsec_policy/__cmd_group.py | 190 | 80 | 378,831 |
LOAD_CONST <code object test_assert2 at 0x7fab42324d20, file "f.py", line 1>
LOAD_CONST 'test_assert2'
MAKE_FUNCTION
STORE_NAME test_assert2
LOAD_NAME test_assert2
LOAD_CONST 5
CALL_FUNCTION
POP_TOP
SETUP_LOOP to 42
LOAD_CONST (2, 4, 6)
GET_ITER
FOR_ITER to 40
STORE_NAME x
LOAD_NAME x
LOAD_NAME x
COMPARE_OP ==
POP_J... | def test_assert2(c):
if c < 2:
raise SyntaxError("Oops")
test_assert2(5)
for x in (2, 4, 6):
assert x == x
for x in (1, 3, 5):
assert x == x, "foo"
| data/uncompyle6-3.9.0/test/simple_source/bug37/01_assert2.py | 215 | 80 | 442,131 |
LOAD_CONST ' A source file '
STORE_NAME __doc__
LOAD_NAME importlib
LOAD_ATTR metadata
LOAD_METHOD version
LOAD_CONST 'mypackage'
CALL_METHOD
STORE_NAME __version__
LOAD_NAME __version__
LOAD_CONST 'UNKNOWN'
COMPARE_OP ==
POP_JUMP_IF_FALSE
LOAD_CONST (0, 0, 0)
STORE_NAME __version_info__
JUMP_FORWARD to 50
LOAD_NAM... | """ A source file """
__version__ = importlib.metadata.version("mypackage")
if __version__ == "UNKNOWN":
__version_info__ = (0, 0, 0)
else:
__version_info__ = tuple(map(int, __version__.split(".")))
| data/versioningit-3.0.0/test/data/replace-version/replacement.py | 122 | 80 | 46,898 |
LOAD_CONST 0
LOAD_CONST ('current_audit_info',)
IMPORT_NAME prowler.providers.aws.lib.audit_info.audit_info
IMPORT_FROM current_audit_info
STORE_NAME current_audit_info
POP_TOP
LOAD_CONST 0
LOAD_CONST ('SageMaker',)
IMPORT_NAME prowler.providers.aws.services.sagemaker.sagemaker_service
IMPORT_FROM SageMaker
STORE_NAME... | from prowler.providers.aws.lib.audit_info.audit_info import current_audit_info
from prowler.providers.aws.services.sagemaker.sagemaker_service import SageMaker
sagemaker_client = SageMaker(current_audit_info)
| data/prowler-3.14.0/prowler/providers/aws/services/sagemaker/sagemaker_client.py | 136 | 80 | 183,305 |
LOAD_CONST 0
LOAD_CONST ('dataclass',)
IMPORT_NAME dataclasses
IMPORT_FROM dataclass
STORE_NAME dataclass
POP_TOP
LOAD_CONST 0
LOAD_CONST ('Optional',)
IMPORT_NAME typing
IMPORT_FROM Optional
STORE_NAME Optional
POP_TOP
LOAD_NAME dataclass
LOAD_BUILD_CLASS
LOAD_CONST <code object DacsParams at 0x7fab81f0b930, file "f... | from dataclasses import dataclass
from typing import Optional
@dataclass
class DacsParams:
"""DacsParams object."""
deployed_platform_username: str = "user"
dacs_application_id: str = "256"
dacs_position: Optional[str] = None
| data/refinitiv-data-1.6.0/refinitiv/data/_core/session/_dacs_params.py | 275 | 80 | 195,141 |
LOAD_CONST 1
LOAD_CONST ('RunCoordinator', 'SubmitRunContext')
IMPORT_NAME base
IMPORT_FROM RunCoordinator
STORE_NAME RunCoordinator
IMPORT_FROM SubmitRunContext
STORE_NAME SubmitRunContext
POP_TOP
LOAD_CONST 1
LOAD_CONST ('DefaultRunCoordinator',)
IMPORT_NAME default_run_coordinator
IMPORT_FROM DefaultRunCoordinator
... | from .base import (
RunCoordinator as RunCoordinator,
SubmitRunContext as SubmitRunContext,
)
from .default_run_coordinator import DefaultRunCoordinator as DefaultRunCoordinator
from .queued_run_coordinator import QueuedRunCoordinator as QueuedRunCoordinator
| data/dagster-1.6.5/dagster/_core/run_coordinator/__init__.py | 136 | 80 | 54,909 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME kernels
STORE_NAME kernels
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate any tests for kernel'
LOAD_CONST ('reason',)
CALL_FUNCTION
LOAD_CONST <code object test_pyawkward_IndexedA... | import pytest
import kernels
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_pyawkward_IndexedArray64_reduce_next_nonlocal_nextshifts_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-spec/test_pyawkward_IndexedArray64_reduce_next_nonlocal_nextshifts_64.py | 218 | 80 | 220,634 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME StoreKit
STORE_NAME StoreKit
LOAD_CONST 0
LOAD_CONST ('TestCase',)
IMPORT_NAME PyObjCTools.TestSupport
IMPORT_FROM TestCase
STORE_NAME TestCase
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object TestStoreKitDefines at 0x7fab820d6a50, file "f.py", line 5>
LOAD_CONST 'TestStoreKit... | import StoreKit
from PyObjCTools.TestSupport import TestCase
class TestStoreKitDefines(TestCase):
def test_defines(self):
self.assertIsInstance(StoreKit.StoreKitBundle, StoreKit.NSBundle)
self.assertIsInstance(StoreKit.SKLocalizedString("hello"), str)
| data/pyobjc-framework-StoreKit-10.1/PyObjCTest/test_storekitdefines.py | 279 | 80 | 41,068 |
LOAD_CONST '\nThis is a URLconf to be loaded by tests.py. Add any URLs needed for tests only.\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('path',)
IMPORT_NAME django.urls
IMPORT_FROM path
STORE_NAME path
POP_TOP
LOAD_CONST 1
LOAD_CONST ('TestForm',)
IMPORT_NAME forms
IMPORT_FROM TestForm
STORE_NAME TestForm
POP_TO... | """
This is a URLconf to be loaded by tests.py. Add any URLs needed for tests only.
"""
from django.urls import path
from .forms import TestForm
from .tests import TestFormPreview
urlpatterns = [
path("preview/", TestFormPreview(TestForm)),
]
| data/django-formtools-2.5.1/tests/urls.py | 152 | 80 | 204,686 |
LOAD_CONST "Dataset definition for squad.\n\nDEPRECATED!\nIf you want to use the Squad dataset builder class, use:\ntfds.builder_cls('squad')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_builder_import
POP_TO... | """Dataset definition for squad.
DEPRECATED!
If you want to use the Squad dataset builder class, use:
tfds.builder_cls('squad')
"""
from tensorflow_datasets.core import lazy_builder_import
Squad = lazy_builder_import.LazyBuilderImport("squad")
| data/tfds-nightly-4.9.4.dev202402210044/tensorflow_datasets/question_answering/squad/squad.py | 124 | 80 | 238,371 |
LOAD_CONST '\nMost of the implementation of the package is here and is internal-only.\n\nPublic API will be in the ``edx_event_bus_redis`` module for the most part.\n\nSee ADR ``docs/decisions/0006-public-api-and-app-organization.rst`` for the reasoning.\n'
STORE_NAME __doc__
LOAD_CONST None
RETURN_VALUE | """
Most of the implementation of the package is here and is internal-only.
Public API will be in the ``edx_event_bus_redis`` module for the most part.
See ADR ``docs/decisions/0006-public-api-and-app-organization.rst`` for the reasoning.
"""
| data/edx_event_bus_redis-0.3.3/edx_event_bus_redis/internal/__init__.py | 92 | 80 | 127,468 |
LOAD_CONST <code object _get_system_version at 0x7fab0028b810, file "f.py", line 1>
LOAD_CONST '_get_system_version'
MAKE_FUNCTION
STORE_NAME _get_system_version
LOAD_CONST None
RETURN_VALUE
LOAD_GLOBAL __file__
LOAD_CONST None
COMPARE_OP is
POP_JUMP_IF_FALSE
SETUP_EXCEPT to 18
LOAD_CONST 5
STORE_FAST m
POP_BLOCK
JU... | def _get_system_version():
if __file__ is None:
try:
m = 5
except IOError:
pass
else:
try:
m = 10
finally:
m = 15
if m is not None:
m = 20
return m
| data/uncompyle6-3.9.0/test/simple_source/bug33/01_if_try_except.py | 188 | 80 | 442,113 |
LOAD_CONST 0
LOAD_CONST ('date',)
IMPORT_NAME datetime
IMPORT_FROM date
STORE_NAME date
POP_TOP
LOAD_CONST 'product'
LOAD_NAME date
LOAD_METHOD today
CALL_METHOD
BUILD_TUPLE
LOAD_CONST 'the database dsn'
LOAD_CONST ('dsn',)
BUILD_CONST_KEY_MAP
LOAD_CONST <code object main at 0x7fae471296f0, file "f.py", line 4>
LOAD_C... | from datetime import date
def main(dsn: "the database dsn", table="product", today=date.today()):
"Do something on the database"
print(dsn, table, today)
if __name__ == "__main__":
import plac
plac.call(main)
| data/plac-1.4.2/doc/example5_.py | 170 | 80 | 112,221 |
LOAD_CONST 0
LOAD_CONST ('path',)
IMPORT_NAME django.urls
IMPORT_FROM path
STORE_NAME path
POP_TOP
LOAD_CONST 1
LOAD_CONST ('views',)
IMPORT_NAME
IMPORT_FROM views
STORE_NAME views
POP_TOP
LOAD_CONST 'example_app'
STORE_NAME app_name
LOAD_NAME path
LOAD_CONST ''
LOAD_NAME views
LOAD_ATTR index
LOAD_CONST 'index'
LOA... | from django.urls import path
from . import views
app_name = "example_app"
urlpatterns = [
path(route="", view=views.index, name="index"),
path(route="create", view=views.ExampleCreateView.as_view(), name="create"),
]
| data/django-silk-5.1.0/project/example_app/urls.py | 144 | 80 | 20,707 |
LOAD_CONST 0
LOAD_CONST ('BaseSettings', 'Field')
IMPORT_NAME pydantic
IMPORT_FROM BaseSettings
STORE_NAME BaseSettings
IMPORT_FROM Field
STORE_NAME Field
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object Settings at 0x7fab82363030, file "f.py", line 4>
LOAD_CONST 'Settings'
MAKE_FUNCTION
LOAD_CONST 'Settings'
LOAD_NAM... | from pydantic import BaseSettings, Field
class Settings(BaseSettings):
db_url: str = Field(
"postgresql://codecarbon-user:supersecret@localhost:5432/codecarbon_db",
env="DATABASE_URL",
)
settings = Settings()
| data/codecarbon-2.3.4/carbonserver/carbonserver/config.py | 208 | 80 | 379,633 |
LOAD_CONST '\ndataprep\n========\n\nDataprep let you prepare your data using a single library with a few lines of code.\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME logging
STORE_NAME logging
LOAD_CONST 1
STORE_NAME DEFAULT_PARTITIONS
LOAD_NAME logging
LOAD_ATTR basicConfig
LOAD_NAME logging
LOAD_A... | """
dataprep
========
Dataprep let you prepare your data using a single library with a few lines of code.
"""
import logging
DEFAULT_PARTITIONS = 1
logging.basicConfig(level=logging.INFO, format="%(message)s")
__version__ = "0.4.4"
| data/dataprep-0.4.5/dataprep/__init__.py | 121 | 80 | 199,084 |
LOAD_CONST 0
LOAD_CONST ('absolute_import', 'division', 'print_function')
IMPORT_NAME __future__
IMPORT_FROM absolute_import
STORE_NAME absolute_import
IMPORT_FROM division
STORE_NAME division
IMPORT_FROM print_function
STORE_NAME print_function
POP_TOP
LOAD_NAME type
STORE_NAME __metaclass__
LOAD_CONST 0
LOAD_CONST ... | from __future__ import absolute_import, division, print_function
__metaclass__ = type
import os
def abspath(file):
return os.path.abspath(file)
class FilterModule(object):
def filters(self):
return {"abspath": abspath}
| data/ansible-9.2.0/ansible_collections/f5networks/f5_modules/plugins/filter/abspath.py | 301 | 80 | 6,204 |
LOAD_CONST '\nURLs for edx_api_doc_tools.\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('path',)
IMPORT_NAME django.urls
IMPORT_FROM path
STORE_NAME path
POP_TOP
LOAD_CONST 0
LOAD_CONST ('TemplateView',)
IMPORT_NAME django.views.generic
IMPORT_FROM TemplateView
STORE_NAME TemplateView
POP_TOP
LOAD_NAME path
LOAD_CO... | """
URLs for edx_api_doc_tools.
"""
from django.urls import path
from django.views.generic import TemplateView
urlpatterns = [
path("", TemplateView.as_view(template_name="edx_api_doc_tools/base.html")),
]
| data/edx-api-doc-tools-1.7.0/edx_api_doc_tools/urls.py | 135 | 80 | 13,684 |
LOAD_CONST 0
LOAD_CONST ('DEBUG', 'getLogger')
IMPORT_NAME logging
IMPORT_FROM DEBUG
STORE_NAME DEBUG
IMPORT_FROM getLogger
STORE_NAME getLogger
POP_TOP
LOAD_CONST 0
LOAD_CONST ('configure_azure_monitor',)
IMPORT_NAME azure.monitor.opentelemetry
IMPORT_FROM configure_azure_monitor
STORE_NAME configure_azure_monitor
PO... | from logging import DEBUG, getLogger
from azure.monitor.opentelemetry import configure_azure_monitor
configure_azure_monitor()
logger = getLogger(__name__)
logger.setLevel(DEBUG)
logger.debug("DEBUG: Debug with properties", extra={"debug": "true"})
input()
| data/azure-monitor-opentelemetry-1.2.0/samples/logging/custom_properties.py | 172 | 80 | 443,765 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME json
STORE_NAME json
LOAD_CONST 0
LOAD_CONST ('IO',)
IMPORT_NAME typing
IMPORT_FROM IO
STORE_NAME IO
POP_TOP
LOAD_CONST 0
LOAD_CONST ('ConfigurationMechanism',)
IMPORT_NAME datahub.configuration
IMPORT_FROM ConfigurationMechanism
STORE_NAME ConfigurationMechanism
POP_TOP
LOAD... | import json
from typing import IO
from datahub.configuration import ConfigurationMechanism
class JsonConfigurationMechanism(ConfigurationMechanism):
"""Ability to load configuration from json files"""
def load_config(self, config_fp: IO) -> dict:
return json.load(config_fp)
| data/acryl-datahub-0.12.1.5/src/datahub/configuration/json_loader.py | 288 | 80 | 435,306 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME kfactory
STORE_NAME kf
LOAD_BUILD_CLASS
LOAD_CONST <code object LAYER at 0x7fab41499db0, file "f.py", line 4>
LOAD_CONST 'LAYER'
MAKE_FUNCTION
LOAD_CONST 'LAYER'
LOAD_NAME kf
LOAD_ATTR LayerEnum
CALL_FUNCTION
STORE_NAME LAYER
LOAD_NAME kf
LOAD_ATTR enclosure
LOAD_METHOD LayerE... | import kfactory as kf
class LAYER(kf.LayerEnum):
SI = (1, 0)
SIEXCLUDE = (1, 1)
si_enc = kf.enclosure.LayerEnclosure([(LAYER.SIEXCLUDE, 2000)])
| data/kfactory-0.11.2/docs/source/layers.py | 187 | 80 | 360,092 |
LOAD_CONST 'For when pip wants to check the date or time.\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME datetime
STORE_NAME datetime
LOAD_NAME int
LOAD_NAME int
LOAD_NAME int
LOAD_NAME bool
LOAD_CONST ('year', 'month', 'day', 'return')
BUILD_CONST_KEY_MAP
LOAD_CONST <code object today_is_later_than a... | """For when pip wants to check the date or time.
"""
import datetime
def today_is_later_than(year: int, month: int, day: int) -> bool:
today = datetime.date.today()
given = datetime.date(year, month, day)
return today > given
| data/clvm_rs-0.6.0/venv/lib/python3.7/site-packages/pip/_internal/utils/datetime.py | 177 | 80 | 288,681 |
LOAD_CONST 'Doctest module for XML comparison.\n\nUsage::\n\n >>> import lxml.usedoctest\n >>> # now do your XML doctests ...\n\nSee `lxml.doctestcompare`\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('doctestcompare',)
IMPORT_NAME lxml
IMPORT_FROM doctestcompare
STORE_NAME doctestcompare
POP_TOP
LOAD_NAME doctestco... | """Doctest module for XML comparison.
Usage::
>>> import lxml.usedoctest
>>> # now do your XML doctests ...
See `lxml.doctestcompare`
"""
from lxml import doctestcompare
doctestcompare.temp_install(del_module=__name__)
| data/lxml-5.1.0/src/lxml/usedoctest.py | 130 | 80 | 58,460 |
LOAD_CONST 'Vocabulary.'
STORE_NAME __doc__
LOAD_CONST 1
LOAD_CONST ('bert', 'elmo', 'subwords', 'vocab')
IMPORT_NAME
IMPORT_FROM bert
STORE_NAME bert
IMPORT_FROM elmo
STORE_NAME elmo
IMPORT_FROM subwords
STORE_NAME subwords
IMPORT_FROM vocab
STORE_NAME vocab
POP_TOP
LOAD_CONST 1
LOAD_CONST ('*',)
IMPORT_NAME bert
IM... | """Vocabulary."""
from . import bert, elmo, subwords, vocab
from .bert import *
from .elmo import *
from .subwords import *
from .vocab import *
__all__ = vocab.__all__ + subwords.__all__ + elmo.__all__ + bert.__all__
| data/gluonnlp-0.10.0/src/gluonnlp/vocab/__init__.py | 192 | 80 | 370,898 |
LOAD_CONST 'default'
LOAD_CONST 'ENGINE'
LOAD_CONST 'django.db.backends.sqlite3'
BUILD_MAP
BUILD_MAP
STORE_NAME DATABASES
LOAD_CONST 'django.db.models.BigAutoField'
STORE_NAME DEFAULT_AUTO_FIELD
LOAD_CONST 'secrekey'
STORE_NAME SECRET_KEY
LOAD_CONST 'phonenumber_field'
LOAD_CONST 'tests'
BUILD_LIST
STORE_NAME INSTAL... | DATABASES = {"default": {"ENGINE": "django.db.backends.sqlite3"}}
DEFAULT_AUTO_FIELD = "django.db.models.BigAutoField"
SECRET_KEY = "secrekey"
INSTALLED_APPS = ["phonenumber_field", "tests"]
| data/django-phonenumber-field-7.3.0/tests/settings.py | 105 | 80 | 247,105 |
LOAD_CONST 'https://s3.amazonaws.com/fast-ai-sample/mnist_tiny.tgz'
LOAD_CONST (342207, '56143e8f24db90d925d82a5a74141875')
BUILD_MAP
POP_TOP
LOAD_CONST None
RETURN_VALUE | {
"https://s3.amazonaws.com/fast-ai-sample/mnist_tiny.tgz": (
342207,
"56143e8f24db90d925d82a5a74141875",
)
}
| data/fastdownload-0.0.7/fastdownload/download_checks.py | 80 | 80 | 394,189 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME SoundAnalysis
STORE_NAME SoundAnalysis
LOAD_CONST 0
LOAD_CONST ('TestCase', 'min_sdk_level')
IMPORT_NAME PyObjCTools.TestSupport
IMPORT_FROM TestCase
STORE_NAME TestCase
IMPORT_FROM min_sdk_level
STORE_NAME min_sdk_level
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object TestSNR... | import SoundAnalysis # noqa: F401
from PyObjCTools.TestSupport import TestCase, min_sdk_level
class TestSNRequest(TestCase):
@min_sdk_level("10.15")
def test_protocols(self):
self.assertProtocolExists("SNRequest")
| data/pyobjc-framework-SoundAnalysis-10.1/PyObjCTest/test_snrequest.py | 266 | 80 | 309,166 |
LOAD_CONST 'Init file.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('ATOM_PUB_ENTRY_URL', 'Article', 'HatenaBlogReader')
IMPORT_NAME llama_hub.hatena_blog.base
IMPORT_FROM ATOM_PUB_ENTRY_URL
STORE_NAME ATOM_PUB_ENTRY_URL
IMPORT_FROM Article
STORE_NAME Article
IMPORT_FROM HatenaBlogReader
STORE_NAME HatenaBlogReader
PO... | """Init file."""
from llama_hub.hatena_blog.base import (
ATOM_PUB_ENTRY_URL,
Article,
HatenaBlogReader,
)
__all__ = ["ATOM_PUB_ENTRY_URL", "Article", "HatenaBlogReader"]
| data/llama_hub-0.0.79.post1/llama_hub/hatena_blog/__init__.py | 145 | 80 | 359,138 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME kernels
STORE_NAME kernels
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate any tests for kernel'
LOAD_CONST ('reason',)
CALL_FUNCTION
LOAD_CONST <code object test_pyawkward_ListArra... | import pytest
import kernels
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_pyawkward_ListArrayU32_getitem_next_range_spreadadvanced_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-spec/test_pyawkward_ListArrayU32_getitem_next_range_spreadadvanced_64.py | 218 | 80 | 220,637 |
LOAD_CONST 0
LOAD_CONST ('App', 'ComposeResult')
IMPORT_NAME textual.app
IMPORT_FROM App
STORE_NAME App
IMPORT_FROM ComposeResult
STORE_NAME ComposeResult
POP_TOP
LOAD_CONST 0
LOAD_CONST ('Header',)
IMPORT_NAME textual.widgets
IMPORT_FROM Header
STORE_NAME Header
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object Heade... | from textual.app import App, ComposeResult
from textual.widgets import Header
class HeaderApp(App):
def compose(self) -> ComposeResult:
yield Header()
if __name__ == "__main__":
app = HeaderApp()
app.run()
| data/textual-0.52.1/docs/examples/widgets/header.py | 278 | 80 | 26,330 |
LOAD_CONST 'Module to handle cached_property version dependent imports.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME sys
STORE_NAME sys
LOAD_NAME sys
LOAD_ATTR version_info
LOAD_CONST (3, 8)
COMPARE_OP >=
POP_JUMP_IF_FALSE
LOAD_CONST 0
LOAD_CONST ('cached_property',)
IMPORT_NAME functools
IMPORT_FROM... | """Module to handle cached_property version dependent imports."""
import sys
if sys.version_info >= (3, 8):
from functools import cached_property
else: # pragma: no cover
from backports.cached_property import cached_property
__all__ = ("cached_property",)
| data/sqlfluff-2.3.5/src/sqlfluff/core/cached_property.py | 147 | 80 | 351,393 |
LOAD_CONST 0
LOAD_CONST ('BaseModel', 'Field')
IMPORT_NAME pydantic
IMPORT_FROM BaseModel
STORE_NAME BaseModel
IMPORT_FROM Field
STORE_NAME Field
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object Reference at 0x7f8e2fb9ced0, file "f.py", line 4>
LOAD_CONST 'Reference'
MAKE_FUNCTION
LOAD_CONST 'Reference'
LOAD_NAME Base... | from pydantic import BaseModel, Field
class Reference(BaseModel):
"""
https://spec.openapis.org/oas/v3.1.0#reference-object
"""
ref: str = Field(..., alias="$ref")
model_config = {"extra": "allow"}
| data/flask_openapi3-3.0.2/flask_openapi3/models/reference.py | 217 | 80 | 405,881 |
LOAD_CONST <code object test_import_url at 0x7fab82184660, file "f.py", line 1>
LOAD_CONST 'test_import_url'
MAKE_FUNCTION
STORE_NAME test_import_url
LOAD_CONST None
RETURN_VALUE
LOAD_FAST make_dataset
LOAD_CONST False
LOAD_CONST True
LOAD_CONST False
LOAD_CONST False
LOAD_CONST False
LOAD_CONST ('cache', 'files', 'd... | def test_import_url(bench_dvc, tmp_dir, scm, dvc, make_dataset):
dataset = make_dataset(
cache=False, files=True, dvcfile=False, commit=False, remote=False
)
bench_dvc("import-url", str(dataset), "new")
| data/dvc-3.47.0/dvc/testing/benchmarks/cli/commands/test_import_url.py | 141 | 80 | 318,081 |
LOAD_CONST 'Constants for the ROMY integration.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('timedelta',)
IMPORT_NAME datetime
IMPORT_FROM timedelta
STORE_NAME timedelta
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME logging
STORE_NAME logging
LOAD_CONST 0
LOAD_CONST ('Platform',)
IMPORT_NAME homeassistant.const
... | """Constants for the ROMY integration."""
from datetime import timedelta
import logging
from homeassistant.const import Platform
DOMAIN = "romy"
PLATFORMS = [Platform.VACUUM]
UPDATE_INTERVAL = timedelta(seconds=5)
LOGGER = logging.getLogger(__package__)
| data/homeassistant-2024.2.2/homeassistant/components/romy/const.py | 161 | 80 | 406,757 |
LOAD_CONST 0
LOAD_CONST ('backend_to_check', 'gnome', 'kde', 'prog_check')
IMPORT_NAME bt
IMPORT_FROM backend_to_check
STORE_NAME backend_to_check
IMPORT_FROM gnome
STORE_NAME gnome
IMPORT_FROM kde
STORE_NAME kde
IMPORT_FROM prog_check
STORE_NAME prog_check
POP_TOP
LOAD_NAME gnome
CALL_FUNCTION
POP_JUMP_IF_FALSE
LOAD... | from bt import backend_to_check, gnome, kde, prog_check
if gnome():
if prog_check(["gnome-screenshot", "--version"]):
def test_gnome_screenshot():
assert not kde()
backend_to_check("gnome-screenshot", delay=1)
| data/pyscreenshot-3.1/tests/test_x_gnome_screenshot.py | 221 | 80 | 12,427 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.