input stringlengths 28 198k | output stringlengths 3 71k | file stringlengths 19 330 | input_tokens int64 5 159k | output_tokens int64 3 9.07k | __index_level_0__ int64 2 449k |
|---|---|---|---|---|---|
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME azure.cli.core.aaz
IMPORT_STAR
LOAD_NAME register_command_group
LOAD_CONST 'network lb inbound-nat-rule'
CALL_FUNCTION
LOAD_BUILD_CLASS
LOAD_CONST <code object __CMDGroup at 0x7fab42062ed0, file "f.py", line 4>
LOAD_CONST '__CMDGroup'
MAKE_FUNCTION
LOAD_CONST '__CMDGroup'
L... | from azure.cli.core.aaz import *
@register_command_group(
"network lb inbound-nat-rule",
)
class __CMDGroup(AAZCommandGroup):
"""Manage inbound NAT rules of a load balancer."""
pass
__all__ = ["__CMDGroup"]
| data/azure-cli-2.57.0/azure/cli/command_modules/network/aaz/latest/network/lb/inbound_nat_rule/__cmd_group.py | 193 | 84 | 378,138 |
LOAD_CONST 0
LOAD_CONST ('auto',)
IMPORT_NAME enum
IMPORT_FROM auto
STORE_NAME auto
POP_TOP
LOAD_CONST 0
LOAD_CONST ('JsonProperty',)
IMPORT_NAME spdx_tools.spdx.jsonschema.json_property
IMPORT_FROM JsonProperty
STORE_NAME JsonProperty
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object RelationshipProperty at 0x7fab540... | from enum import auto
from spdx_tools.spdx.jsonschema.json_property import JsonProperty
class RelationshipProperty(JsonProperty):
SPDX_ELEMENT_ID = auto()
COMMENT = auto()
RELATED_SPDX_ELEMENT = auto()
RELATIONSHIP_TYPE = auto()
| data/spdx-tools-0.8.2/src/spdx_tools/spdx/jsonschema/relationship_properties.py | 219 | 84 | 286,055 |
LOAD_CONST 'Module of model pipeline and pipenet.\n\nPipeline or Pipenet connects multiple components (transformers, detectors,\nand/or aggregators) into a model that may perform complex anomaly detection\nprocess.\n\n'
STORE_NAME __doc__
LOAD_CONST 1
LOAD_CONST ('Pipeline', 'Pipenet')
IMPORT_NAME _pipe
IMPORT_FROM Pi... | """Module of model pipeline and pipenet.
Pipeline or Pipenet connects multiple components (transformers, detectors,
and/or aggregators) into a model that may perform complex anomaly detection
process.
"""
from ._pipe import Pipeline, Pipenet
__all__ = ["Pipeline", "Pipenet"]
| data/adtk-0.6.2/src/adtk/pipe/__init__.py | 132 | 84 | 345,454 |
LOAD_CONST 0
LOAD_CONST ('absolute_import', 'print_function', 'division')
IMPORT_NAME __future__
IMPORT_FROM absolute_import
STORE_NAME absolute_import
IMPORT_FROM print_function
STORE_NAME print_function
IMPORT_FROM division
STORE_NAME division
POP_TOP
LOAD_CONST 0
LOAD_CONST ('version_info',)
IMPORT_NAME sys
IMPORT_... | from __future__ import absolute_import, print_function, division
from sys import version_info
py_2 = version_info[0] < 3
long = py_2 and (long) or (int)
range = py_2 and (xrange) or (range)
unicode = py_2 and (unicode) or (str)
| data/snakebite-py3-3.0.5/snakebite/compat.py | 170 | 84 | 244,088 |
LOAD_CONST 0
LOAD_CONST ('current_audit_info',)
IMPORT_NAME prowler.providers.aws.lib.audit_info.audit_info
IMPORT_FROM current_audit_info
STORE_NAME current_audit_info
POP_TOP
LOAD_CONST 0
LOAD_CONST ('SecretsManager',)
IMPORT_NAME prowler.providers.aws.services.secretsmanager.secretsmanager_service
IMPORT_FROM Secre... | from prowler.providers.aws.lib.audit_info.audit_info import current_audit_info
from prowler.providers.aws.services.secretsmanager.secretsmanager_service import (
SecretsManager,
)
secretsmanager_client = SecretsManager(current_audit_info)
| data/prowler-3.14.0/prowler/providers/aws/services/secretsmanager/secretsmanager_client.py | 131 | 84 | 183,366 |
LOAD_CONST '\n'
STORE_NAME __doc__
LOAD_CONST 2
LOAD_CONST ('Config',)
IMPORT_NAME config
IMPORT_FROM Config
STORE_NAME Config
POP_TOP
LOAD_CONST 1
LOAD_CONST ('torch',)
IMPORT_NAME
IMPORT_FROM torch
STORE_NAME torch
POP_TOP
LOAD_NAME Config
LOAD_ATTR zero_gpu
POP_JUMP_IF_FALSE
LOAD_NAME torch
LOAD_METHOD is_in_bad... | """
"""
from ..config import Config
from . import torch
if Config.zero_gpu:
if torch.is_in_bad_fork():
raise RuntimeError(
"CUDA has been initialized before importing the `spaces` package"
)
torch.patch() # pragma: no cover
| data/spaces-0.23.2/spaces/zero/__init__.py | 132 | 84 | 93,390 |
LOAD_CONST 0
LOAD_CONST ('Optional', 'TypedDict')
IMPORT_NAME typing
IMPORT_FROM Optional
STORE_NAME Optional
IMPORT_FROM TypedDict
STORE_NAME TypedDict
POP_TOP
LOAD_CONST 0
LOAD_CONST ('DatasourceType',)
IMPORT_NAME superset.utils.core
IMPORT_FROM DatasourceType
STORE_NAME DatasourceType
POP_TOP
LOAD_BUILD_CLASS
LOA... | from typing import Optional, TypedDict
from superset.utils.core import DatasourceType
class TemporaryExploreState(TypedDict):
owner: Optional[int]
datasource_id: int
datasource_type: DatasourceType
chart_id: Optional[int]
form_data: str
| data/apache-superset-3.1.1/superset/commands/explore/form_data/state.py | 296 | 84 | 393,401 |
LOAD_CONST "Dataset definition for billsum.\n\nDEPRECATED!\nIf you want to use the Billsum dataset builder class, use:\ntfds.builder_cls('billsum')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_builder_import
... | """Dataset definition for billsum.
DEPRECATED!
If you want to use the Billsum dataset builder class, use:
tfds.builder_cls('billsum')
"""
from tensorflow_datasets.core import lazy_builder_import
Billsum = lazy_builder_import.LazyBuilderImport("billsum")
| data/tfds-nightly-4.9.4.dev202402210044/tensorflow_datasets/summarization/billsum.py | 127 | 84 | 211,345 |
LOAD_CONST 'DO NOT EDIT.\n\nThis file was autogenerated. Do not edit it by hand,\nsince your modifications would be overwritten.\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('deserialize_keras_object',)
IMPORT_NAME keras.src.legacy.saving.serialization
IMPORT_FROM deserialize_keras_object
STORE_NAME deserialize_kera... | """DO NOT EDIT.
This file was autogenerated. Do not edit it by hand,
since your modifications would be overwritten.
"""
from keras.src.legacy.saving.serialization import deserialize_keras_object
from keras.src.legacy.saving.serialization import serialize_keras_object
| data/keras-3.0.5/keras/utils/legacy/__init__.py | 154 | 84 | 349,097 |
LOAD_CONST 0
LOAD_CONST ('Enum',)
IMPORT_NAME enum
IMPORT_FROM Enum
STORE_NAME Enum
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object DelegatedAdminRelationshipRequestAction at 0x7fab701208a0, file "f.py", line 4>
LOAD_CONST 'DelegatedAdminRelationshipRequestAction'
MAKE_FUNCTION
LOAD_CONST 'DelegatedAdminRelationshipR... | from enum import Enum
class DelegatedAdminRelationshipRequestAction(str, Enum):
LockForApproval = ("lockForApproval",)
Approve = ("approve",)
Terminate = ("terminate",)
UnknownFutureValue = ("unknownFutureValue",)
Reject = ("reject",)
| data/msgraph-sdk-1.1.0/msgraph/generated/models/delegated_admin_relationship_request_action.py | 222 | 84 | 261,328 |
LOAD_CONST 1441195
LOAD_CONST 'Dallas 10'
LOAD_CONST 'dal10'
LOAD_CONST ('id', 'longName', 'name')
BUILD_CONST_KEY_MAP
LOAD_CONST 449494
LOAD_CONST 'Dallas 9'
LOAD_CONST 'dal09'
LOAD_CONST ('id', 'longName', 'name')
BUILD_CONST_KEY_MAP
BUILD_LIST
STORE_NAME getDatacenters
BUILD_LIST
STORE_NAME getHardwareRouters
LOAD... | getDatacenters = [
{"id": 1441195, "longName": "Dallas 10", "name": "dal10"},
{"id": 449494, "longName": "Dallas 9", "name": "dal09"},
]
getHardwareRouters = []
| data/SoftLayer-6.1.11/SoftLayer/fixtures/SoftLayer_Location_Datacenter.py | 103 | 84 | 291,510 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME os
STORE_NAME os
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME salt.modules.ini_manage
STORE_NAME salt
LOAD_CONST <code object test_section_req at 0x7fab70078c00, file "f.py", line 6>
LOAD_CONST 'test_section_req'
MAKE_FUNCTION
STORE_NAME test_section_req
LOAD_CONST None
RETURN_VAL... | import os
import salt.modules.ini_manage
def test_section_req():
"""
Test the __repr__ in the _Section class
"""
expected = "_Section(){}{{}}".format(os.linesep)
assert repr(salt.modules.ini_manage._Section("test")) == expected
| data/salt-3006.6/tests/pytests/unit/modules/test_ini_manage.py | 170 | 84 | 96,537 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME unittest
STORE_NAME unittest
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME zope.testrunner.huh
STORE_NAME zope
LOAD_BUILD_CLASS
LOAD_CONST <code object Test at 0x7f8e2ff25b70, file "f.py", line 6>
LOAD_CONST 'Test'
MAKE_FUNCTION
LOAD_CONST 'Test'
LOAD_NAME unittest
LOAD_ATTR TestCa... | import unittest
import zope.testrunner.huh # noqa: F401
class Test(unittest.TestCase):
def test(self):
self.assertEqual(1, 0)
def test_suite():
return unittest.defaultTestLoader.loadTestsFromTestCase(Test)
| data/zope.testrunner-6.3.1/src/zope/testrunner/tests/testrunner-ex/sample2/sample21/sampletests_i.py | 279 | 84 | 115,805 |
LOAD_BUILD_CLASS
LOAD_CONST <code object Foo at 0x7fab800abdb0, file "f.py", line 1>
LOAD_CONST 'Foo'
MAKE_FUNCTION
LOAD_CONST 'Foo'
CALL_FUNCTION
STORE_NAME Foo
LOAD_CONST None
RETURN_VALUE
LOAD_NAME __name__
STORE_NAME __module__
LOAD_CONST 'Foo'
STORE_NAME __qualname__
LOAD_CONST 'docstring of Foo.'
STORE_NAME __d... | class Foo:
"""docstring of Foo."""
def meth(self):
"""docstring of meth."""
pass
def skipmeth(self):
"""docstring of skipmeth."""
pass
def _privatemeth(self):
"""docstring of _privatemeth."""
pass
| data/sphinx-7.2.6/tests/roots/test-ext-autosummary-skip-member/target.py | 273 | 84 | 381,157 |
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME azure.cli.core.aaz
IMPORT_STAR
LOAD_NAME register_command_group
LOAD_CONST 'monitor log-analytics workspace data-export'
CALL_FUNCTION
LOAD_BUILD_CLASS
LOAD_CONST <code object __CMDGroup at 0x7fab64192270, file "f.py", line 4>
LOAD_CONST '__CMDGroup'
MAKE_FUNCTION
LOAD_CONST... | from azure.cli.core.aaz import *
@register_command_group(
"monitor log-analytics workspace data-export",
)
class __CMDGroup(AAZCommandGroup):
"""Manage data export ruls for log analytics workspace."""
pass
__all__ = ["__CMDGroup"]
| data/azure-cli-2.57.0/azure/cli/command_modules/monitor/aaz/latest/monitor/log_analytics/workspace/data_export/__cmd_group.py | 194 | 84 | 374,989 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME polars
STORE_NAME pl
LOAD_CONST None
LOAD_CONST ('return',)
BUILD_CONST_KEY_MAP
LOAD_CONST <code object test_base_class at 0x7f8b2000b300, file "f.py", line 6>
LOAD_CONST 'test_base_class'
MAKE_FUNCTION
STORE_NA... | import pytest
import polars as pl
def test_base_class() -> None:
assert isinstance(pl.ComputeError("msg"), pl.PolarsError)
msg = "msg"
with pytest.raises(pl.PolarsError, match=msg):
raise pl.OutOfBoundsError(msg)
| data/polars-0.20.10/py-polars/tests/unit/test_exceptions.py | 218 | 84 | 333,766 |
LOAD_CONST 1
LOAD_CONST ('FormatColumn',)
IMPORT_NAME intcolumn
IMPORT_FROM FormatColumn
STORE_NAME FormatColumn
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object NullColumn at 0x7fab8223f270, file "f.py", line 4>
LOAD_CONST 'NullColumn'
MAKE_FUNCTION
LOAD_CONST 'NullColumn'
LOAD_NAME FormatColumn
CALL_FUNCTION
STORE_N... | from .intcolumn import FormatColumn
class NullColumn(FormatColumn):
ch_type = "Null"
format = "B"
@property
def size(self):
return 1
def after_read_items(self, items, nulls_map=None):
return (None,) * len(items)
| data/clickhouse-driver-0.2.7/clickhouse_driver/columns/nullcolumn.py | 278 | 84 | 305,060 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME os
STORE_NAME os
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME portalocker
STORE_NAME portalocker
LOAD_CONST <code object test_temporary_file_lock at 0x7fab802345d0, file "f.py", line 6>
LOAD_CONST 'test_temporary_file_lock'
MAKE_FUNCTION
STORE_NAME test_temporary_file_lock
LOAD_CO... | import os
import portalocker
def test_temporary_file_lock(tmpfile):
with portalocker.TemporaryFileLock(tmpfile):
pass
assert not os.path.isfile(tmpfile)
lock = portalocker.TemporaryFileLock(tmpfile)
lock.acquire()
del lock
| data/portalocker-2.8.2/portalocker_tests/temporary_file_lock.py | 201 | 84 | 162,175 |
LOAD_CONST 1
LOAD_CONST ('*',)
IMPORT_NAME __cmd_group
IMPORT_STAR
LOAD_CONST 1
LOAD_CONST ('*',)
IMPORT_NAME _convert
IMPORT_STAR
LOAD_CONST 1
LOAD_CONST ('*',)
IMPORT_NAME _deallocate
IMPORT_STAR
LOAD_CONST 1
LOAD_CONST ('*',)
IMPORT_NAME _generalize
IMPORT_STAR
LOAD_CONST 1
LOAD_CONST ('*',)
IMPORT_NAME _list_si... | from .__cmd_group import *
from ._convert import *
from ._deallocate import *
from ._generalize import *
from ._list_sizes import *
from ._list_vm_resize_options import *
from ._perform_maintenance import *
from ._redeploy import *
from ._start import *
| data/azure-cli-2.57.0/azure/cli/command_modules/vm/aaz/profile_2019_03_01_hybrid/vm/__init__.py | 177 | 84 | 376,443 |
LOAD_CONST 1
LOAD_CONST ('nested_lookup', 'get_all_keys', 'get_occurrence_of_key', 'get_occurrence_of_value', 'get_occurrences_and_values')
IMPORT_NAME nested_lookup
IMPORT_FROM nested_lookup
STORE_NAME nested_lookup
IMPORT_FROM get_all_keys
STORE_NAME get_all_keys
IMPORT_FROM get_occurrence_of_key
STORE_NAME get_occur... | from .nested_lookup import (
nested_lookup,
get_all_keys,
get_occurrence_of_key,
get_occurrence_of_value,
get_occurrences_and_values,
)
from .lookup_api import nested_update, nested_delete, nested_alter
| data/nested-lookup-0.2.25/nested_lookup/__init__.py | 223 | 84 | 207,707 |
LOAD_CONST 1
LOAD_CONST ('FormatColumn',)
IMPORT_NAME intcolumn
IMPORT_FROM FormatColumn
STORE_NAME FormatColumn
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object NothingColumn at 0x7fae2f4044b0, file "f.py", line 4>
LOAD_CONST 'NothingColumn'
MAKE_FUNCTION
LOAD_CONST 'NothingColumn'
LOAD_NAME FormatColumn
CALL_FUNCTIO... | from .intcolumn import FormatColumn
class NothingColumn(FormatColumn):
ch_type = "Nothing"
format = "B"
@property
def size(self):
return 1
def after_read_items(self, items, nulls_map=None):
return (None,) * len(items)
| data/clickhouse-driver-0.2.7/clickhouse_driver/columns/nothingcolumn.py | 278 | 84 | 305,044 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME numpy
STORE_NAME np
LOAD_CONST 0
LOAD_CONST ('Softmax',)
IMPORT_NAME onnx.reference.ops.op_softmax
IMPORT_FROM Softmax
STORE_NAME Softmax
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object LogSoftmax at 0x7f8ece8a8c90, file "f.py", line 6>
LOAD_CONST 'LogSoftmax'
MAKE_FUNCTION
L... | import numpy as np
from onnx.reference.ops.op_softmax import Softmax
class LogSoftmax(Softmax):
def _run(self, X): # type: ignore
Y = Softmax._run(self, X)[0]
np.log(Y, out=Y)
return (Y,)
| data/onnxsim-0.4.35/third_party/onnx-optimizer/third_party/onnx/onnx/reference/ops/op_log_softmax.py | 254 | 84 | 114,187 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME matplotlib.pyplot
IMPORT_FROM pyplot
STORE_NAME plt
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME numpy
STORE_NAME np
LOAD_NAME np
LOAD_METHOD array
LOAD_CONST 2
LOAD_CONST 1
LOAD_CONST 5
LOAD_CONST 7
LOAD_CONST 4
LOAD_CONST 6
LOAD_CONST 8
LOAD_CONST 14
LOAD_CONST 10
LOAD_C... | import matplotlib.pyplot as plt
import numpy as np
a = np.array([2, 1, 5, 7, 4, 6, 8, 14, 10, 9, 18, 20, 22])
plt.plot(a)
plt.show()
| data/numpy-1.26.4/doc/source/user/plots/matplotlib1.py | 143 | 84 | 363,602 |
LOAD_CONST 0
LOAD_CONST ('Checkpointer',)
IMPORT_NAME allennlp.training.checkpointer
IMPORT_FROM Checkpointer
STORE_NAME Checkpointer
POP_TOP
LOAD_CONST 0
LOAD_CONST ('NoOpTrainer',)
IMPORT_NAME allennlp.training.no_op_trainer
IMPORT_FROM NoOpTrainer
STORE_NAME NoOpTrainer
POP_TOP
LOAD_CONST 0
LOAD_CONST ('TrainerCal... | from allennlp.training.checkpointer import Checkpointer
from allennlp.training.no_op_trainer import NoOpTrainer
from allennlp.training.callbacks import TrainerCallback
from allennlp.training.trainer import Trainer
from allennlp.training.gradient_descent_trainer import GradientDescentTrainer
| data/allennlp-2.10.1/allennlp/training/__init__.py | 202 | 84 | 82,585 |
LOAD_CONST './panoptic-fpn_r50_fpn_1x_coco.py'
STORE_NAME _base_
LOAD_NAME dict
LOAD_NAME dict
LOAD_CONST 101
LOAD_NAME dict
LOAD_CONST 'Pretrained'
LOAD_CONST 'torchvision://resnet101'
LOAD_CONST ('type', 'checkpoint')
CALL_FUNCTION
LOAD_CONST ('depth', 'init_cfg')
CALL_FUNCTION
LOAD_CONST ('backbone',)
CALL_FUNCT... | _base_ = "./panoptic-fpn_r50_fpn_1x_coco.py"
model = dict(
backbone=dict(
depth=101,
init_cfg=dict(type="Pretrained", checkpoint="torchvision://resnet101"),
)
)
| data/mmdet-3.3.0/mmdet/.mim/configs/panoptic_fpn/panoptic-fpn_r101_fpn_1x_coco.py | 109 | 84 | 151,210 |
LOAD_CONST 0
LOAD_CONST ('annotations',)
IMPORT_NAME __future__
IMPORT_FROM annotations
STORE_NAME annotations
POP_TOP
LOAD_CONST 0
LOAD_CONST ('dataclass',)
IMPORT_NAME dataclasses
IMPORT_FROM dataclass
STORE_NAME dataclass
POP_TOP
LOAD_CONST 0
LOAD_CONST ('TYPE_CHECKING',)
IMPORT_NAME typing
IMPORT_FROM TYPE_CHECKI... | from __future__ import annotations
from dataclasses import dataclass
from typing import TYPE_CHECKING
if TYPE_CHECKING:
from pathlib import Path
@dataclass(frozen=True)
class Location:
file: Path
line: int | None = None
column: int | None = None
| data/deptry-0.12.0/deptry/imports/location.py | 298 | 84 | 45,202 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ometa
STORE_NAME ometa
LOAD_CONST 0
LOAD_CONST ('OMetaGrammarBase',)
IMPORT_NAME ometa.runtime
IMPORT_FROM OMetaGrammarBase
STORE_NAME OMetaGrammarBase
POP_TOP
LOAD_CONST 0
LOAD_CONST ('OMeta',)
IMPORT_NAME ometa.grammar
IMPORT_FROM OMeta
STORE_NAME OMeta
POP_TOP
LOAD_CONST 0... | import ometa
from ometa.runtime import OMetaGrammarBase
from ometa.grammar import OMeta
from ometa.grammar import loadGrammar
from terml.nodes import termMaker as t
OMeta1 = loadGrammar(ometa, "pymeta_v1", globals(), OMetaGrammarBase)
| data/Parsley-1.3/ometa/compat.py | 195 | 84 | 193,500 |
LOAD_CONST 0
LOAD_CONST ('can_import_module', 'collect_submodules')
IMPORT_NAME PyInstaller.utils.hooks
IMPORT_FROM can_import_module
STORE_NAME can_import_module
IMPORT_FROM collect_submodules
STORE_NAME collect_submodules
POP_TOP
LOAD_NAME can_import_module
LOAD_CONST 'cffi'
CALL_FUNCTION
POP_JUMP_IF_FALSE
LOAD_NAM... | from PyInstaller.utils.hooks import can_import_module, collect_submodules
if can_import_module("cffi"):
hiddenimports = collect_submodules("win32ctypes.core.cffi")
else:
hiddenimports = collect_submodules("win32ctypes.core.ctypes")
| data/pyinstaller-6.4.0/PyInstaller/hooks/hook-win32ctypes.core.py | 144 | 84 | 185,287 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME os
STORE_NAME os
LOAD_CONST 0
LOAD_CONST ('Celery',)
IMPORT_NAME celery
IMPORT_FROM Celery
STORE_NAME Celery
POP_TOP
LOAD_NAME os
LOAD_ATTR environ
LOAD_METHOD setdefault
LOAD_CONST 'DJANGO_SETTINGS_MODULE'
LOAD_CONST 't.proj.settings'
CALL_METHOD
POP_TOP
LOAD_NAME Celery
LOA... | import os
from celery import Celery
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "t.proj.settings")
app = Celery("proj")
app.config_from_object("django.conf:settings", namespace="CELERY")
app.autodiscover_tasks()
| data/django-celery-beat-2.5.0/t/proj/celery.py | 153 | 84 | 117,906 |
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME azure.cli.core.aaz
IMPORT_STAR
LOAD_NAME register_command_group
LOAD_CONST 'network application-gateway routing-rule'
CALL_FUNCTION
LOAD_BUILD_CLASS
LOAD_CONST <code object __CMDGroup at 0x7fab42062030, file "f.py", line 4>
LOAD_CONST '__CMDGroup'
MAKE_FUNCTION
LOAD_CONST '_... | from azure.cli.core.aaz import *
@register_command_group(
"network application-gateway routing-rule",
)
class __CMDGroup(AAZCommandGroup):
"""Evaluate probe information and define TCP/TLS routing rules."""
pass
__all__ = ["__CMDGroup"]
| data/azure-cli-2.57.0/azure/cli/command_modules/network/aaz/latest/network/application_gateway/routing_rule/__cmd_group.py | 194 | 84 | 378,438 |
LOAD_CONST '\nA module to implement finitary relations [1] as predicate.\n\nReferences\n==========\n\n.. [1] https://en.wikipedia.org/wiki/Finitary_relation\n\n'
STORE_NAME __doc__
LOAD_CONST 'BinaryRelation'
LOAD_CONST 'AppliedBinaryRelation'
BUILD_LIST
STORE_NAME __all__
LOAD_CONST 1
LOAD_CONST ('BinaryRelation', '... | """
A module to implement finitary relations [1] as predicate.
References
==========
.. [1] https://en.wikipedia.org/wiki/Finitary_relation
"""
__all__ = ["BinaryRelation", "AppliedBinaryRelation"]
from .binrel import BinaryRelation, AppliedBinaryRelation
| data/sympy-1.12/sympy/assumptions/relation/__init__.py | 140 | 84 | 228,861 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME logging
STORE_NAME logging
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME sys
STORE_NAME sys
LOAD_NAME logging
LOAD_METHOD getLogger
LOAD_NAME __name__
CALL_METHOD
STORE_NAME LOGGER
SETUP_EXCEPT to 70
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ljdata
STORE_NAME _data
LOAD_NAME _dat... | import logging
import sys
LOGGER = logging.getLogger(__name__)
try:
import ljdata as _data
globals()["data"] = _data
sys.modules["pylibjpeg.data"] = _data
LOGGER.debug("pylibjpeg-data module loaded")
except ImportError:
pass
| data/pylibjpeg-2.0.0/pylibjpeg/tests/__init__.py | 178 | 84 | 57,568 |
LOAD_CONST 0
LOAD_CONST ('assert_type',)
IMPORT_NAME typing_extensions
IMPORT_FROM assert_type
STORE_NAME assert_type
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME click
STORE_NAME click
LOAD_NAME click
LOAD_METHOD command
CALL_METHOD
LOAD_NAME click
LOAD_METHOD help_option
LOAD_CONST '-h'
LOAD_CONST '--help'
CAL... | from typing_extensions import assert_type
import click
@click.command()
@click.help_option("-h", "--help")
def hello() -> None:
"""Simple program that greets NAME for a total of COUNT times."""
click.echo("Hello!")
assert_type(hello, click.Command)
| data/asyncclick-8.1.7.1/tests/typing/typing_help_option.py | 179 | 84 | 249,504 |
LOAD_CONST 0
LOAD_CONST ('track_dag', 'track_task')
IMPORT_NAME dbnd_airflow.tracking.dbnd_dag_tracking
IMPORT_FROM track_dag
STORE_NAME track_dag
IMPORT_FROM track_task
STORE_NAME track_task
POP_TOP
LOAD_CONST 0
LOAD_CONST ('track_operator',)
IMPORT_NAME dbnd_airflow.tracking.execute_tracking
IMPORT_FROM track_operat... | from dbnd_airflow.tracking.dbnd_dag_tracking import track_dag, track_task
from dbnd_airflow.tracking.execute_tracking import track_operator
__all__ = ["track_dag", "track_task", "track_operator"]
__version__ = "1.0.21.5"
| data/dbnd-airflow-1.0.21.5/src/dbnd_airflow/__init__.py | 148 | 84 | 366,510 |
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME azure.cli.core.aaz
IMPORT_STAR
LOAD_NAME register_command_group
LOAD_CONST 'network lb inbound-nat-rule'
CALL_FUNCTION
LOAD_BUILD_CLASS
LOAD_CONST <code object __CMDGroup at 0x7fab8204a390, file "f.py", line 4>
LOAD_CONST '__CMDGroup'
MAKE_FUNCTION
LOAD_CONST '__CMDGroup'
L... | from azure.cli.core.aaz import *
@register_command_group(
"network lb inbound-nat-rule",
)
class __CMDGroup(AAZCommandGroup):
"""Manage inbound NAT rules of a load balancer."""
pass
__all__ = ["__CMDGroup"]
| data/azure-cli-2.57.0/azure/cli/command_modules/network/aaz/profile_2018_03_01_hybrid/network/lb/inbound_nat_rule/__cmd_group.py | 194 | 84 | 378,950 |
LOAD_CONST '\nSome\nmulti-line\ndocstring\nhere\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('absolute_import',)
IMPORT_NAME __future__
IMPORT_FROM absolute_import
STORE_NAME absolute_import
POP_TOP
LOAD_CONST 0
LOAD_CONST ('division',)
IMPORT_NAME __future__
IMPORT_FROM division
STORE_NAME division
POP_TOP
LOAD_C... | """
Some
multi-line
docstring
here
"""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import html
def fixture_function_str():
return str("test")
html_escape = html.escape
| data/ddtrace-2.6.3/tests/appsec/iast/fixtures/aspects/str/future_import_function_str.py | 249 | 84 | 269,728 |
LOAD_CONST 0
LOAD_CONST ('version_info',)
IMPORT_NAME sys
IMPORT_FROM version_info
STORE_NAME v
POP_TOP
BUILD_LIST
STORE_NAME collect_ignore
LOAD_NAME v
LOAD_CONST 0
BINARY_SUBSCR
LOAD_CONST 3
COMPARE_OP >=
POP_JUMP_IF_FALSE
LOAD_NAME v
LOAD_CONST 1
BINARY_SUBSCR
LOAD_CONST 5
COMPARE_OP >=
POP_JUMP_IF_TRUE
LOAD_NAME... | from sys import version_info as v
collect_ignore = []
if not (v[0] >= 3 and v[1] >= 5):
collect_ignore.append("test_async.py")
if not (v[0] >= 3 and v[1] >= 7):
collect_ignore.append("test_trio.py")
| data/pyee-11.1.0/tests/conftest.py | 157 | 84 | 362,074 |
LOAD_CONST 'Top-level package for Glean parser.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('get_distribution', 'DistributionNotFound')
IMPORT_NAME pkg_resources
IMPORT_FROM get_distribution
STORE_NAME get_distribution
IMPORT_FROM DistributionNotFound
STORE_NAME DistributionNotFound
POP_TOP
SETUP_EXCEPT to 36
LOAD_... | """Top-level package for Glean parser."""
from pkg_resources import get_distribution, DistributionNotFound
try:
__version__ = get_distribution(__name__).version
except DistributionNotFound:
pass
__author__ = """The Glean Team"""
__email__ = "glean-team@mozilla.com"
| data/glean_parser-12.0.1/glean_parser/__init__.py | 169 | 84 | 234,216 |
LOAD_CONST 'Project wide exception classes.'
STORE_NAME __doc__
LOAD_BUILD_CLASS
LOAD_CONST <code object PrometheusApiClientException at 0x7fab540c5810, file "f.py", line 4>
LOAD_CONST 'PrometheusApiClientException'
MAKE_FUNCTION
LOAD_CONST 'PrometheusApiClientException'
LOAD_NAME Exception
CALL_FUNCTION
STORE_NAME Pr... | """Project wide exception classes."""
class PrometheusApiClientException(Exception):
"""API client exception, raises when response status code != 200."""
pass
class MetricValueConversionError(Exception):
"""Raises when we find a metric that is a string where we fail to convert it to a float."""
pa... | data/prometheus-api-client-0.5.4/prometheus_api_client/exceptions.py | 289 | 84 | 108,372 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME hubspot.cms.blogs.tags
IMPORT_FROM cms
ROT_TWO
POP_TOP
IMPORT_FROM blogs
ROT_TWO
POP_TOP
IMPORT_FROM tags
STORE_NAME api_client
POP_TOP
LOAD_CONST 4
LOAD_CONST ('DiscoveryBase',)
IMPORT_NAME discovery_base
IMPORT_FROM DiscoveryBase
STORE_NAME DiscoveryBase
POP_TOP
LOAD_BUILD_C... | import hubspot.cms.blogs.tags as api_client
from ....discovery_base import DiscoveryBase
class Discovery(DiscoveryBase):
@property
def tag_api(self) -> api_client.BlogTagsApi:
return self._configure_api_client(api_client, "BlogTagsApi")
| data/hubspot-api-client-8.2.1/hubspot/discovery/cms/blogs/tags/discovery.py | 273 | 84 | 320,803 |
LOAD_CONST 2
LOAD_CONST ('and_', 'not_', 'or_')
IMPORT_NAME blocking_rule_composition
IMPORT_FROM and_
STORE_NAME and_
IMPORT_FROM not_
STORE_NAME not_
IMPORT_FROM or_
STORE_NAME or_
POP_TOP
LOAD_CONST 1
LOAD_CONST ('block_on', 'exact_match_rule')
IMPORT_NAME postgres_helpers.postgres_blocking_rule_imports
IMPORT_FROM... | from ..blocking_rule_composition import ( # noqa: F401
and_,
not_,
or_,
)
from .postgres_helpers.postgres_blocking_rule_imports import ( # noqa: F401
block_on,
exact_match_rule,
)
| data/splink-3.9.12/splink/postgres/blocking_rule_library.py | 119 | 84 | 15,969 |
LOAD_CONST 'Google Benchmark tooling'
STORE_NAME __doc__
LOAD_CONST 'Eric Fiselier'
STORE_NAME __author__
LOAD_CONST 'eric@efcs.ca'
STORE_NAME __email__
LOAD_CONST (0, 5, 0)
STORE_NAME __versioninfo__
LOAD_CONST '.'
LOAD_METHOD join
LOAD_CONST <code object <genexpr> at 0x7f8ed39ed6f0, file "f.py", line 6>
LOAD_CONS... | """Google Benchmark tooling"""
__author__ = "Eric Fiselier"
__email__ = "eric@efcs.ca"
__versioninfo__ = (0, 5, 0)
__version__ = ".".join(str(v) for v in __versioninfo__) + "dev"
__all__ = []
| data/onnxsim-0.4.35/third_party/onnx-optimizer/third_party/protobuf/third_party/benchmark/tools/gbench/__init__.py | 187 | 84 | 114,751 |
LOAD_CONST 0
LOAD_CONST ('Response',)
IMPORT_NAME pyramid.response
IMPORT_FROM Response
STORE_NAME Response
POP_TOP
LOAD_CONST 0
LOAD_CONST ('view_config',)
IMPORT_NAME pyramid.view
IMPORT_FROM view_config
STORE_NAME view_config
POP_TOP
LOAD_NAME view_config
LOAD_CONST 'hello'
LOAD_CONST ('route_name',)
CALL_FUNCTION... | from pyramid.response import Response
from pyramid.view import view_config
@view_config(route_name="hello")
def hello_world(request):
body = "<h1>Hi %(first)s %(last)s!</h1>" % request.matchdict
return Response(body)
| data/pyramid-2.0.2/docs/quick_tour/routing/views.py | 184 | 84 | 19,985 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME hubspot.conversations.visitor_identification
IMPORT_FROM conversations
ROT_TWO
POP_TOP
IMPORT_FROM visitor_identification
STORE_NAME api_client
POP_TOP
LOAD_CONST 3
LOAD_CONST ('DiscoveryBase',)
IMPORT_NAME discovery_base
IMPORT_FROM DiscoveryBase
STORE_NAME DiscoveryBase
POP_T... | import hubspot.conversations.visitor_identification as api_client
from ...discovery_base import DiscoveryBase
class Discovery(DiscoveryBase):
@property
def generate_api(self) -> api_client.GenerateApi:
return self._configure_api_client(api_client, "GenerateApi")
| data/hubspot-api-client-8.2.1/hubspot/discovery/conversations/visitor_identification/discovery.py | 269 | 84 | 320,829 |
LOAD_CONST 'DO NOT EDIT.\n\nThis file was autogenerated. Do not edit it by hand,\nsince your modifications would be overwritten.\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('deserialize_keras_object',)
IMPORT_NAME keras.src.legacy.saving.serialization
IMPORT_FROM deserialize_keras_object
STORE_NAME deserialize_kera... | """DO NOT EDIT.
This file was autogenerated. Do not edit it by hand,
since your modifications would be overwritten.
"""
from keras.src.legacy.saving.serialization import deserialize_keras_object
from keras.src.legacy.saving.serialization import serialize_keras_object
| data/keras-3.0.5/keras/_tf_keras/keras/legacy/saving/__init__.py | 154 | 84 | 349,518 |
LOAD_CONST 0
LOAD_CONST ('TestCase',)
IMPORT_NAME PyObjCTools.TestSupport
IMPORT_FROM TestCase
STORE_NAME TestCase
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME PHASE
STORE_NAME PHASE
LOAD_BUILD_CLASS
LOAD_CONST <code object TestPHASEEngine at 0x7faa5017fe40, file "f.py", line 6>
LOAD_CONST 'TestPHASEEngine'
MAKE_... | from PyObjCTools.TestSupport import TestCase
import PHASE
class TestPHASEEngine(TestCase):
def test_methods(self):
self.assertResultIsBOOL(PHASE.PHASEEngine.startAndReturnError_)
self.assertArgIsOut(PHASE.PHASEEngine.startAndReturnError_, 0)
| data/pyobjc-framework-PHASE-10.1/PyObjCTest/test_phaseengine.py | 271 | 84 | 135,230 |
LOAD_CONST './boxinst_r50_fpn_ms-90k_coco.py'
STORE_NAME _base_
LOAD_NAME dict
LOAD_NAME dict
LOAD_CONST 101
LOAD_NAME dict
LOAD_CONST 'Pretrained'
LOAD_CONST 'torchvision://resnet101'
LOAD_CONST ('type', 'checkpoint')
CALL_FUNCTION
LOAD_CONST ('depth', 'init_cfg')
CALL_FUNCTION
LOAD_CONST ('backbone',)
CALL_FUNCTI... | _base_ = "./boxinst_r50_fpn_ms-90k_coco.py"
model = dict(
backbone=dict(
depth=101,
init_cfg=dict(type="Pretrained", checkpoint="torchvision://resnet101"),
)
)
| data/mmdet-3.3.0/mmdet/.mim/configs/boxinst/boxinst_r101_fpn_ms-90k_coco.py | 108 | 84 | 151,675 |
LOAD_CONST 0
LOAD_CONST ('Enum',)
IMPORT_NAME enum
IMPORT_FROM Enum
STORE_NAME Enum
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object ManagedAppClipboardSharingLevel at 0x7fab4298ec00, file "f.py", line 4>
LOAD_CONST 'ManagedAppClipboardSharingLevel'
MAKE_FUNCTION
LOAD_CONST 'ManagedAppClipboardSharingLevel'
LOAD_NAME ... | from enum import Enum
class ManagedAppClipboardSharingLevel(str, Enum):
AllApps = ("allApps",)
ManagedAppsWithPasteIn = ("managedAppsWithPasteIn",)
ManagedApps = ("managedApps",)
Blocked = ("blocked",)
| data/msgraph-sdk-1.1.0/msgraph/generated/models/managed_app_clipboard_sharing_level.py | 219 | 84 | 260,736 |
LOAD_CONST 0
LOAD_CONST ('Enum',)
IMPORT_NAME enum
IMPORT_FROM Enum
STORE_NAME Enum
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object ActionState at 0x7fab4298e420, file "f.py", line 4>
LOAD_CONST 'ActionState'
MAKE_FUNCTION
LOAD_CONST 'ActionState'
LOAD_NAME str
LOAD_NAME Enum
CALL_FUNCTION
STORE_NAME ActionState
LOAD... | from enum import Enum
class ActionState(str, Enum):
None_ = ("none",)
Pending = ("pending",)
Canceled = ("canceled",)
Active = ("active",)
Done = ("done",)
Failed = ("failed",)
NotSupported = ("notSupported",)
| data/msgraph-sdk-1.1.0/msgraph/generated/models/action_state.py | 198 | 84 | 260,739 |
LOAD_CONST 0
LOAD_CONST ('datetime',)
IMPORT_NAME datetime
IMPORT_FROM datetime
STORE_NAME datetime
POP_TOP
LOAD_CONST 0
LOAD_CONST ('UUID',)
IMPORT_NAME uuid
IMPORT_FROM UUID
STORE_NAME UUID
POP_TOP
LOAD_CONST 0
LOAD_CONST ('BaseModel',)
IMPORT_NAME fiddler3.schemas.base
IMPORT_FROM BaseModel
STORE_NAME BaseModel
PO... | from datetime import datetime
from uuid import UUID
from fiddler3.schemas.base import BaseModel
class OrganizationCompactResp(BaseModel):
id: UUID
name: str
class OrganizationResp(BaseModel):
id: UUID
name: str
created_at: datetime
updated_at: datetime
| data/fiddler-client-2.4.0/fiddler3/schemas/organization.py | 394 | 84 | 422,819 |
LOAD_CONST <code object func_a at 0x7fab823a6ed0, file "f.py", line 1>
LOAD_CONST 'func_a'
MAKE_FUNCTION
STORE_NAME func_a
LOAD_CONST <code object func_b at 0x7fab703bbb70, file "f.py", line 8>
LOAD_CONST 'func_b'
MAKE_FUNCTION
STORE_NAME func_b
LOAD_CONST <code object func_c at 0x7fa98c2c7d20, file "f.py", line 15>
... | def func_a():
"""func_a[^1].
[^1]: Footnote\x20A
"""
def func_b():
"""func_b[^x].
[^x]: Footnote\x20B
"""
def func_c():
"""func_c."""
| data/mkdocstrings-0.24.0/tests/fixtures/footnotes.py | 168 | 84 | 209,127 |
LOAD_CONST 2
LOAD_CONST ('and_', 'not_', 'or_')
IMPORT_NAME blocking_rule_composition
IMPORT_FROM and_
STORE_NAME and_
IMPORT_FROM not_
STORE_NAME not_
IMPORT_FROM or_
STORE_NAME or_
POP_TOP
LOAD_CONST 1
LOAD_CONST ('block_on', 'exact_match_rule')
IMPORT_NAME sqlite_helpers.sqlite_blocking_rule_imports
IMPORT_FROM blo... | from ..blocking_rule_composition import ( # noqa: F401
and_,
not_,
or_,
)
from .sqlite_helpers.sqlite_blocking_rule_imports import ( # noqa: F401
block_on,
exact_match_rule,
)
| data/splink-3.9.12/splink/sqlite/blocking_rule_library.py | 120 | 84 | 15,939 |
LOAD_CONST 0
LOAD_CONST ('absolute_import', 'division', 'print_function')
IMPORT_NAME __future__
IMPORT_FROM absolute_import
STORE_NAME absolute_import
IMPORT_FROM division
STORE_NAME division
IMPORT_FROM print_function
STORE_NAME print_function
POP_TOP
LOAD_NAME type
STORE_NAME __metaclass__
LOAD_CONST None
RETURN_V... | from __future__ import absolute_import, division, print_function
__metaclass__ = type
"""
Compat library for ansible. This contains compatibility definitions for older python
When we need to import a module differently depending on python version, do it
here. Then in the code we can simply import from compat in ord... | data/ansible-core-2.16.3/lib/ansible/compat/__init__.py | 70 | 84 | 436,957 |
LOAD_CONST '\nSphinx Gallery\n==============\n\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME os
STORE_NAME os
LOAD_CONST '0.15.0'
STORE_NAME __version__
LOAD_CONST <code object glr_path_static at 0x7fab823756f0, file "f.py", line 12>
LOAD_CONST 'glr_path_static'
MAKE_FUNCTION
STORE_NAME glr_path_sta... | """
Sphinx Gallery
==============
"""
import os
__version__ = "0.15.0"
def glr_path_static():
"""Returns path to packaged static files"""
return os.path.abspath(os.path.join(os.path.dirname(__file__), "_static"))
| data/sphinx-gallery-0.15.0/sphinx_gallery/__init__.py | 165 | 84 | 391,142 |
LOAD_CONST 2
LOAD_CONST ('and_', 'not_', 'or_')
IMPORT_NAME blocking_rule_composition
IMPORT_FROM and_
STORE_NAME and_
IMPORT_FROM not_
STORE_NAME not_
IMPORT_FROM or_
STORE_NAME or_
POP_TOP
LOAD_CONST 1
LOAD_CONST ('block_on', 'exact_match_rule')
IMPORT_NAME spark_helpers.spark_blocking_rule_imports
IMPORT_FROM block... | from ..blocking_rule_composition import ( # noqa: F401
and_,
not_,
or_,
)
from .spark_helpers.spark_blocking_rule_imports import ( # noqa: F401
block_on,
exact_match_rule,
)
| data/splink-3.9.12/splink/spark/blocking_rule_library.py | 119 | 84 | 15,924 |
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME azure.cli.core.aaz
IMPORT_STAR
LOAD_NAME register_command_group
LOAD_CONST 'relay'
CALL_FUNCTION
LOAD_BUILD_CLASS
LOAD_CONST <code object __CMDGroup at 0x7fab42062a50, file "f.py", line 4>
LOAD_CONST '__CMDGroup'
MAKE_FUNCTION
LOAD_CONST '__CMDGroup'
LOAD_NAME AAZCommandGro... | from azure.cli.core.aaz import *
@register_command_group(
"relay",
)
class __CMDGroup(AAZCommandGroup):
"""Manage Azure Relay Service namespaces, WCF relays, hybrid connections, and rules."""
pass
__all__ = ["__CMDGroup"]
| data/azure-cli-2.57.0/azure/cli/command_modules/relay/aaz/latest/relay/__cmd_group.py | 194 | 84 | 375,610 |
LOAD_CONST 0
LOAD_CONST ('TestCase',)
IMPORT_NAME PyObjCTools.TestSupport
IMPORT_FROM TestCase
STORE_NAME TestCase
POP_TOP
LOAD_CONST 0
LOAD_CONST ('simd',)
IMPORT_NAME objc
IMPORT_FROM simd
STORE_NAME simd
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME PHASE
STORE_NAME PHASE
LOAD_BUILD_CLASS
LOAD_CONST <code obje... | from PyObjCTools.TestSupport import TestCase
from objc import simd
import PHASE
class TestPHASEEnvelope(TestCase):
def test_methods(self):
self.assertResultHasType(
PHASE.PHASEEnvelope.startPoint, simd.simd_double2.__typestr__
)
| data/pyobjc-framework-PHASE-10.1/PyObjCTest/test_phaseenvelope.py | 286 | 84 | 135,235 |
LOAD_CONST 'cartesian'
LOAD_CONST 'color'
LOAD_CONST 'animal'
BUILD_LIST
LOAD_CONST ('type', 'lists')
BUILD_CONST_KEY_MAP
LOAD_CONST 'words'
LOAD_CONST 'белая'
LOAD_CONST 'чёрная'
BUILD_LIST
LOAD_CONST ('type', 'words')
BUILD_CONST_KEY_MAP
LOAD_CONST 'words'
LOAD_CONST 'кошка'
LOAD_CONST 'собака'
BUILD_LIST
LOAD_CONS... | config = {
"all": {"type": "cartesian", "lists": ["color", "animal"]},
"color": {"type": "words", "words": ["белая", "чёрная"]},
"animal": {"type": "words", "words": ["кошка", "собака"]},
}
| data/coolname-2.2.0/examples/russian_module/__init__.py | 118 | 84 | 156,476 |
SETUP_ANNOTATIONS
LOAD_CONST 'Provides I/O modules using dask dataframe.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('Any',)
IMPORT_NAME typing
IMPORT_FROM Any
STORE_NAME Any
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME lazy_loader
STORE_NAME lazy
LOAD_NAME Any
LOAD_NAME __annotations__
LOAD_CONST 'ParquetDatas... | """Provides I/O modules using dask dataframe."""
from typing import Any
import lazy_loader as lazy
ParquetDataset: Any
__getattr__, __dir__, __all__ = lazy.attach(
__name__, submod_attrs={"parquet_dataset": ["ParquetDataset"]}
)
| data/kedro-datasets-2.0.0/kedro_datasets/dask/__init__.py | 148 | 84 | 61,499 |
LOAD_CONST 'Test the command line interfae explicitely'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME subprocess
STORE_NAME subprocess
LOAD_CONST 'x-wr-timezone'
STORE_NAME CMD
LOAD_CONST <code object test_help at 0x7faa7c1f6d20, file "f.py", line 9>
LOAD_CONST 'test_help'
MAKE_FUNCTION
STORE_NAME test... | """Test the command line interfae explicitely"""
import subprocess
CMD = "x-wr-timezone"
def test_help():
"""Test that a help is being displayed."""
help = subprocess.check_output([CMD, "--help"])
assert b"x-wr-timezone" in help
| data/x_wr_timezone-0.0.6/test/test_command_line.py | 162 | 84 | 209,025 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME asyncio
STORE_NAME asyncio
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME os
STORE_NAME os
LOAD_CONST '<IPY-INPUT>'
LOAD_NAME os
LOAD_ATTR environ
LOAD_CONST 'IPYKERNEL_CELL_NAME'
STORE_SUBSCR
LOAD_NAME os
LOAD_ATTR name
LOAD_CONST 'nt'
COMPARE_OP ==
POP_JUMP_IF_FALSE
LOAD_NAME has... | import asyncio
import os
os.environ["IPYKERNEL_CELL_NAME"] = "<IPY-INPUT>"
if os.name == "nt" and hasattr(asyncio, "WindowsSelectorEventLoopPolicy"):
asyncio.set_event_loop_policy(asyncio.WindowsSelectorEventLoopPolicy())
| data/nbclient-0.9.0/tests/conftest.py | 139 | 84 | 404,202 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME sys
STORE_NAME sys
LOAD_CONST 0
LOAD_CONST ('Widget',)
IMPORT_NAME widget_module
IMPORT_FROM Widget
STORE_NAME Widget
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object DerivedWidget at 0x7fab823b2d20, file "f.py", line 6>
LOAD_CONST 'DerivedWidget'
MAKE_FUNCTION
LOAD_CONST 'Der... | import sys
from widget_module import Widget
class DerivedWidget(Widget):
def __init__(self, message):
super(DerivedWidget, self).__init__(message)
def the_answer(self):
return 42
def argv0(self):
return sys.argv[0]
| data/onnx-simplifier-0.4.35/third_party/pybind11/tests/test_embed/test_interpreter.py | 369 | 84 | 309,432 |
LOAD_CONST 0
LOAD_CONST ('tzinfo', 'datetime')
IMPORT_NAME datetime
IMPORT_FROM tzinfo
STORE_NAME tzinfo
IMPORT_FROM datetime
STORE_NAME datetime
POP_TOP
LOAD_CONST 0
LOAD_CONST ('parse',)
IMPORT_NAME dateutil.parser
IMPORT_FROM parse
STORE_NAME parse
POP_TOP
LOAD_CONST <code object _utc_string_to_datetime at 0x7fab4... | from datetime import tzinfo, datetime
from dateutil.parser import parse
def _utc_string_to_datetime(value):
return parse(value)
def _datetime_to_utc_string(value):
return value.strftime("%Y-%m-%dT%H:%M:%SZ")
| data/vsts-client-1.2.1/vstsclient/_conversion.py | 241 | 84 | 438,253 |
LOAD_CONST 0
LOAD_CONST ('Dict', 'Protocol', 'Union')
IMPORT_NAME typing
IMPORT_FROM Dict
STORE_NAME Dict
IMPORT_FROM Protocol
STORE_NAME Protocol
IMPORT_FROM Union
STORE_NAME Union
POP_TOP
LOAD_CONST 0
LOAD_CONST ('RemoteConnection',)
IMPORT_NAME selenium.webdriver.remote.remote_connection
IMPORT_FROM RemoteConnectio... | from typing import Dict, Protocol, Union
from selenium.webdriver.remote.remote_connection import RemoteConnection
class CanExecuteCommands(Protocol):
command_executor: RemoteConnection
def execute(self, driver_command: str, params: Union[Dict, None] = None) -> Dict:
...
| data/Appium-Python-Client-3.1.1/appium/protocols/webdriver/can_execute_commands.py | 301 | 84 | 146,141 |
LOAD_CONST 0
LOAD_CONST ('Enum',)
IMPORT_NAME enum
IMPORT_FROM Enum
STORE_NAME Enum
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object SimulationAutomationRunStatus at 0x7fab4298e270, file "f.py", line 4>
LOAD_CONST 'SimulationAutomationRunStatus'
MAKE_FUNCTION
LOAD_CONST 'SimulationAutomationRunStatus'
LOAD_NAME str
LO... | from enum import Enum
class SimulationAutomationRunStatus(str, Enum):
Unknown = ("unknown",)
Running = ("running",)
Succeeded = ("succeeded",)
Failed = ("failed",)
Skipped = ("skipped",)
UnknownFutureValue = ("unknownFutureValue",)
| data/msgraph-sdk-1.1.0/msgraph/generated/models/simulation_automation_run_status.py | 216 | 84 | 260,663 |
LOAD_CONST '1.1.0'
STORE_NAME __version__
LOAD_CONST 0
LOAD_CONST ('Consul',)
IMPORT_NAME consul.std
IMPORT_FROM Consul
STORE_NAME Consul
POP_TOP
LOAD_CONST 0
LOAD_CONST ('Check',)
IMPORT_NAME consul.base
IMPORT_FROM Check
STORE_NAME Check
POP_TOP
LOAD_CONST 0
LOAD_CONST ('ConsulException',)
IMPORT_NAME consul.base
... | __version__ = "1.1.0"
from consul.std import Consul
from consul.base import Check
from consul.base import ConsulException
from consul.base import ACLPermissionDenied
from consul.base import ACLDisabled
from consul.base import NotFound
from consul.base import Timeout
| data/python-consul-1.1.0/consul/__init__.py | 229 | 84 | 344,819 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME typer
STORE_NAME typer
LOAD_CONST 0
LOAD_CONST ('Annotated',)
IMPORT_NAME typing_extensions
IMPORT_FROM Annotated
STORE_NAME Annotated
POP_TOP
LOAD_CONST ('World',)
LOAD_NAME Annotated
LOAD_NAME str
LOAD_NAME typer
LOAD_ATTR Argument
LOAD_CONST '✨username✨'
LOAD_CONST ('metava... | import typer
from typing_extensions import Annotated
def main(name: Annotated[str, typer.Argument(metavar="✨username✨")] = "World"):
print(f"Hello {name}")
if __name__ == "__main__":
typer.run(main)
| data/typer-0.9.0/docs_src/arguments/help/tutorial006_an.py | 206 | 84 | 224,556 |
LOAD_CONST 0
LOAD_CONST ('TestCase', 'min_os_level')
IMPORT_NAME PyObjCTools.TestSupport
IMPORT_FROM TestCase
STORE_NAME TestCase
IMPORT_FROM min_os_level
STORE_NAME min_os_level
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME Quartz
STORE_NAME Quartz
LOAD_BUILD_CLASS
LOAD_CONST <code object TestCAEDRMetadata at 0x7... | from PyObjCTools.TestSupport import TestCase, min_os_level
import Quartz
class TestCAEDRMetadata(TestCase):
@min_os_level("13.0")
def test_methods13_0(self):
self.assertResultIsBOOL(Quartz.CAEDRMetadata.isAvailable)
| data/pyobjc-framework-Quartz-10.1/PyObjCTest/test_caedrmetadata.py | 291 | 84 | 133,596 |
LOAD_CONST 1
LOAD_CONST ('BenchmarkResult',)
IMPORT_NAME _result
IMPORT_FROM BenchmarkResult
STORE_NAME BenchmarkResult
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object ComputeTime at 0x7fab001365d0, file "f.py", line 4>
LOAD_CONST 'ComputeTime'
MAKE_FUNCTION
LOAD_CONST 'ComputeTime'
CALL_FUNCTION
STORE_NAME ComputeTi... | from ._result import BenchmarkResult
class ComputeTime:
"""Extracts a runtime benchmark result from the passed Explanation."""
def __call__(self, explanation, name):
return BenchmarkResult(
"compute time", name, value=explanation.compute_time / explanation.shape[0]
)
| data/shap-0.44.1/shap/benchmark/_compute.py | 249 | 84 | 92,709 |
LOAD_CONST 0
LOAD_CONST ('Gumbel',)
IMPORT_NAME jaxopt._src.perturbations
IMPORT_FROM Gumbel
STORE_NAME Gumbel
POP_TOP
LOAD_CONST 0
LOAD_CONST ('Normal',)
IMPORT_NAME jaxopt._src.perturbations
IMPORT_FROM Normal
STORE_NAME Normal
POP_TOP
LOAD_CONST 0
LOAD_CONST ('make_perturbed_argmax',)
IMPORT_NAME jaxopt._src.pertu... | from jaxopt._src.perturbations import Gumbel
from jaxopt._src.perturbations import Normal
from jaxopt._src.perturbations import make_perturbed_argmax
from jaxopt._src.perturbations import make_perturbed_max
from jaxopt._src.perturbations import make_perturbed_fun
| data/jaxopt-0.8.3/jaxopt/perturbations.py | 208 | 84 | 252,424 |
LOAD_CONST '__version__'
BUILD_LIST
STORE_NAME __all__
LOAD_CONST 'django-bootstrap-v5'
STORE_NAME PACKAGE_NAME
SETUP_EXCEPT to 28
LOAD_CONST 0
LOAD_CONST ('metadata',)
IMPORT_NAME importlib.metadata
IMPORT_FROM metadata
STORE_NAME metadata
POP_TOP
POP_BLOCK
JUMP_FORWARD to 60
DUP_TOP
LOAD_NAME ImportError
COMPARE_... | __all__ = [
"__version__",
]
PACKAGE_NAME = "django-bootstrap-v5"
try:
from importlib.metadata import metadata
except ImportError:
from importlib_metadata import metadata
package_metadata = metadata(PACKAGE_NAME)
__version__ = package_metadata["Version"]
| data/django-bootstrap-v5-1.0.11/src/bootstrap5/__init__.py | 167 | 84 | 41,407 |
LOAD_CONST 0
LOAD_CONST ('Entity',)
IMPORT_NAME office365.entity
IMPORT_FROM Entity
STORE_NAME Entity
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object UserFlowLanguagePage at 0x7fab700add20, file "f.py", line 4>
LOAD_CONST 'UserFlowLanguagePage'
MAKE_FUNCTION
LOAD_CONST 'UserFlowLanguagePage'
LOAD_NAME Entity
CALL_FUN... | from office365.entity import Entity
class UserFlowLanguagePage(Entity):
"""
Determines the user flow language pages that are shown to users during a user flow.
These language pages include both the default language translations provided by Microsoft,
or custom pages that can be created to customize th... | data/Office365-REST-Python-Client-2.5.5/office365/directory/identities/userflows/language_page.py | 189 | 84 | 188,903 |
LOAD_CONST 'Google Benchmark tooling'
STORE_NAME __doc__
LOAD_CONST 'Eric Fiselier'
STORE_NAME __author__
LOAD_CONST 'eric@efcs.ca'
STORE_NAME __email__
LOAD_CONST (0, 5, 0)
STORE_NAME __versioninfo__
LOAD_CONST '.'
LOAD_METHOD join
LOAD_CONST <code object <genexpr> at 0x7f8ab706f930, file "f.py", line 6>
LOAD_CONS... | """Google Benchmark tooling"""
__author__ = "Eric Fiselier"
__email__ = "eric@efcs.ca"
__versioninfo__ = (0, 5, 0)
__version__ = ".".join(str(v) for v in __versioninfo__) + "dev"
__all__ = []
| data/onnxsim-0.4.35/third_party/onnx-optimizer/third_party/onnx/third_party/benchmark/tools/gbench/__init__.py | 188 | 84 | 114,611 |
LOAD_CONST 1
LOAD_CONST ('SYNTAXES', 'extract')
IMPORT_NAME _extruct
IMPORT_FROM SYNTAXES
STORE_NAME SYNTAXES
IMPORT_FROM extract
STORE_NAME extract
POP_TOP
LOAD_CONST 1
LOAD_CONST ('JsonLdExtractor',)
IMPORT_NAME jsonld
IMPORT_FROM JsonLdExtractor
STORE_NAME JsonLdExtractor
POP_TOP
LOAD_CONST 1
LOAD_CONST ('Microfor... | from ._extruct import SYNTAXES, extract
from .jsonld import JsonLdExtractor
from .microformat import MicroformatExtractor
from .opengraph import OpenGraphExtractor
from .rdfa import RDFaExtractor
from .w3cmicrodata import MicrodataExtractor
from .xmldom import XmlDomHTMLParser
| data/extruct-0.16.0/extruct/__init__.py | 257 | 84 | 17,501 |
LOAD_CONST 0
LOAD_CONST ('Enum',)
IMPORT_NAME enum
IMPORT_FROM Enum
STORE_NAME Enum
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object DelegatedAdminAccessAssignmentStatus at 0x7fae471296f0, file "f.py", line 4>
LOAD_CONST 'DelegatedAdminAccessAssignmentStatus'
MAKE_FUNCTION
LOAD_CONST 'DelegatedAdminAccessAssignmentSta... | from enum import Enum
class DelegatedAdminAccessAssignmentStatus(str, Enum):
Pending = ("pending",)
Active = ("active",)
Deleting = ("deleting",)
Deleted = ("deleted",)
Error = ("error",)
UnknownFutureValue = ("unknownFutureValue",)
| data/msgraph-sdk-1.1.0/msgraph/generated/models/delegated_admin_access_assignment_status.py | 220 | 84 | 260,578 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME cx_Oracle
STORE_NAME cx_Oracle
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME db_config
STORE_NAME db_config
LOAD_NAME cx_Oracle
LOAD_METHOD connect
LOAD_NAME db_config
LOAD_ATTR user
LOAD_NAME db_config
LOAD_ATTR pw
LOAD_NAME db_config
LOAD_ATTR dsn
CALL_METHOD
STORE_NAME con
LOAD... | import cx_Oracle
import db_config
con = cx_Oracle.connect(db_config.user, db_config.pw, db_config.dsn)
print(cx_Oracle.version)
print("Database version:", con.version)
print("Client version:", cx_Oracle.clientversion())
| data/cx_Oracle-8.3.0/samples/tutorial/solutions/versions.py | 151 | 84 | 335,300 |
LOAD_CONST 1
LOAD_CONST ('bilinear_upsampling', 'causal_conv1d', 'conv2d', 'conv2d_transpose', 'conv3d', 'frac_bilinear_upsampling', 'separable_conv2d', 'separable_conv3d')
IMPORT_NAME abstract_conv
IMPORT_FROM bilinear_upsampling
STORE_NAME bilinear_upsampling
IMPORT_FROM causal_conv1d
STORE_NAME causal_conv1d
IMPORT_... | from .abstract_conv import (
bilinear_upsampling,
causal_conv1d,
conv2d,
conv2d_transpose,
conv3d,
frac_bilinear_upsampling,
separable_conv2d,
separable_conv3d,
)
| data/pytensor-2.18.6/pytensor/tensor/conv/__init__.py | 211 | 84 | 88,393 |
LOAD_CONST 0
LOAD_CONST ('Union',)
IMPORT_NAME typing
IMPORT_FROM Union
STORE_NAME Union
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME torch
STORE_NAME torch
LOAD_NAME torch
LOAD_ATTR Tensor
LOAD_NAME Union
LOAD_NAME str
LOAD_NAME torch
LOAD_ATTR dtype
BUILD_TUPLE
BINARY_SUBSCR
LOAD_NAME torch
LOAD_ATTR Tensor
LO... | from typing import Union
import torch
def _convert_fp_tensor(
tensor: torch.Tensor, dst_type: Union[str, torch.dtype]
) -> torch.Tensor:
return tensor.to(dst_type) if torch.is_floating_point(tensor) else tensor
| data/lightning-lite-1.8.6/src/lightning_lite/plugins/precision/utils.py | 192 | 84 | 323,123 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME typer
STORE_NAME typer
LOAD_NAME typer
LOAD_METHOD Option
LOAD_CONST True
LOAD_CONST ' /--demo'
LOAD_CONST ' /-d'
CALL_METHOD
BUILD_TUPLE
LOAD_NAME bool
LOAD_CONST ('in_prod',)
BUILD_CONST_KEY_MAP
LOAD_CONST <code object main at 0x7fab700db390, file "f.py", line 4>
LOAD_CONST '... | import typer
def main(in_prod: bool = typer.Option(True, " /--demo", " /-d")):
if in_prod:
print("Running in production")
else:
print("Running demo")
if __name__ == "__main__":
typer.run(main)
| data/typer-0.9.0/docs_src/parameter_types/bool/tutorial004.py | 185 | 84 | 224,489 |
LOAD_CONST '\nInternal FiftyOne Brain package.\n\nContains all non-public code powering the ``fiftyone.brain`` public namespace.\n\n| Copyright 2017-2023, Voxel51, Inc.\n| `voxel51.com <https://voxel51.com/>`_\n|\n'
STORE_NAME __doc__
LOAD_CONST None
RETURN_VALUE | """
Internal FiftyOne Brain package.
Contains all non-public code powering the ``fiftyone.brain`` public namespace.
| Copyright 2017-2023, Voxel51, Inc.
| `voxel51.com <https://voxel51.com/>`_
|
"""
| data/fiftyone-brain-0.16.1/fiftyone/brain/internal/__init__.py | 97 | 84 | 416,747 |
LOAD_CONST 1
LOAD_CONST ('AttributeValue', 'Module', 'Span', 'StackTrace', 'TruncatableString')
IMPORT_NAME trace
IMPORT_FROM AttributeValue
STORE_NAME AttributeValue
IMPORT_FROM Module
STORE_NAME Module
IMPORT_FROM Span
STORE_NAME Span
IMPORT_FROM StackTrace
STORE_NAME StackTrace
IMPORT_FROM TruncatableString
STORE_NA... | from .trace import AttributeValue, Module, Span, StackTrace, TruncatableString
from .tracing import BatchWriteSpansRequest
__all__ = (
"AttributeValue",
"Module",
"Span",
"StackTrace",
"TruncatableString",
"BatchWriteSpansRequest",
)
| data/google-cloud-trace-1.13.1/google/cloud/trace_v2/types/__init__.py | 156 | 84 | 191,474 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME argparse
STORE_NAME argparse
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME doctest
STORE_NAME doctest
LOAD_NAME argparse
LOAD_ATTR ArgumentParser
LOAD_CONST 'Run a test by name'
LOAD_CONST ('description',)
CALL_FUNCTION
STORE_NAME parser
LOAD_NAME parser
LOAD_METHOD add_argument
L... | import argparse
import doctest
parser = argparse.ArgumentParser(description="Run a test by name")
parser.add_argument("test_name")
args = parser.parse_args()
doctest.testfile(args.test_name, optionflags=doctest.REPORT_ONLY_FIRST_FAILURE)
| data/boto-2.49.0/tests/mturk/run-doctest.py | 150 | 84 | 346,544 |
LOAD_CONST 'array'
BUILD_LIST
STORE_NAME __tags__
LOAD_CONST False
STORE_NAME exclude
SETUP_EXCEPT to 24
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME numpy
STORE_NAME numpy
POP_BLOCK
JUMP_FORWARD to 48
DUP_TOP
LOAD_NAME ImportError
COMPARE_OP exception match
POP_JUMP_IF_FALSE
POP_TOP
POP_TOP
POP_TOP
LOAD_CONST True
ST... | __tags__ = ["array"]
exclude = False
try:
import numpy
except ImportError:
exclude = True
else:
try:
import pygame.pixelcopy
except ImportError:
exclude = True
if exclude:
__tags__.extend(("ignore", "subprocess_ignore"))
| data/pygame-2.5.2/test/surfarray_tags.py | 194 | 84 | 124,934 |
LOAD_CONST 0
LOAD_CONST ('absolute_import', 'division', 'print_function')
IMPORT_NAME __future__
IMPORT_FROM absolute_import
STORE_NAME absolute_import
IMPORT_FROM division
STORE_NAME division
IMPORT_FROM print_function
STORE_NAME print_function
POP_TOP
LOAD_NAME type
STORE_NAME __metaclass__
LOAD_BUILD_CLASS
LOAD_CO... | from __future__ import absolute_import, division, print_function
__metaclass__ = type
class AnsibleISEException(Exception):
"""Base class for all Ansible ISE package exceptions."""
pass
class InconsistentParameters(AnsibleISEException):
"""Provided parameters are not consistent."""
pass
| data/ansible-9.2.0/ansible_collections/cisco/ise/plugins/plugin_utils/exceptions.py | 306 | 84 | 2,682 |
LOAD_CONST 'DO NOT EDIT.\n\nThis file was autogenerated. Do not edit it by hand,\nsince your modifications would be overwritten.\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('deserialize_keras_object',)
IMPORT_NAME keras.src.legacy.saving.serialization
IMPORT_FROM deserialize_keras_object
STORE_NAME deserialize_kera... | """DO NOT EDIT.
This file was autogenerated. Do not edit it by hand,
since your modifications would be overwritten.
"""
from keras.src.legacy.saving.serialization import deserialize_keras_object
from keras.src.legacy.saving.serialization import serialize_keras_object
| data/keras-3.0.5/keras/_tf_keras/keras/utils/legacy/__init__.py | 154 | 84 | 349,511 |
LOAD_CONST "Dataset definition for lambada.\n\nDEPRECATED!\nIf you want to use the Lambada dataset builder class, use:\ntfds.builder_cls('lambada')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_builder_import
... | """Dataset definition for lambada.
DEPRECATED!
If you want to use the Lambada dataset builder class, use:
tfds.builder_cls('lambada')
"""
from tensorflow_datasets.core import lazy_builder_import
Lambada = lazy_builder_import.LazyBuilderImport("lambada")
| data/tensorflow-datasets-4.9.4/tensorflow_datasets/text/lambada/lambada.py | 127 | 84 | 256,238 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('Step',)
IMPORT_NAME frictionless
IMPORT_FROM Step
STORE_NAME Step
POP_TOP
LOAD_CONST <code object test_step_from_descriptor_type_framework_v4 at 0x7fab429cd5d0, file "f.py", line 6>
LOAD_CONST 'test_step_from_descriptor_type_f... | import pytest
from frictionless import Step
def test_step_from_descriptor_type_framework_v4():
with pytest.warns(UserWarning):
step = Step.from_descriptor({"code": "table-print"})
assert step.to_descriptor() == {"type": "table-print"}
| data/frictionless-5.16.1/tests/pipeline/step/test_convert.py | 234 | 84 | 41,475 |
LOAD_CONST 0
LOAD_CONST ('template',)
IMPORT_NAME django
IMPORT_FROM template
STORE_NAME template
POP_TOP
LOAD_CONST 0
LOAD_CONST ('WARN_AFTER', 'EXPIRE_AFTER')
IMPORT_NAME session_security.settings
IMPORT_FROM WARN_AFTER
STORE_NAME WARN_AFTER
IMPORT_FROM EXPIRE_AFTER
STORE_NAME EXPIRE_AFTER
POP_TOP
LOAD_NAME templat... | from django import template
from session_security.settings import WARN_AFTER, EXPIRE_AFTER
register = template.Library()
@register.filter
def expire_after(request):
return EXPIRE_AFTER
@register.filter
def warn_after(request):
return WARN_AFTER
| data/django-session-security-2.6.7/session_security/templatetags/session_security_tags.py | 238 | 84 | 146,135 |
LOAD_CONST '../cascade_rcnn/cascade-mask-rcnn_x101-32x4d_fpn_1x_coco.py'
STORE_NAME _base_
LOAD_NAME dict
LOAD_NAME dict
LOAD_NAME dict
LOAD_CONST 'SyncBN'
LOAD_CONST True
LOAD_CONST ('type', 'requires_grad')
CALL_FUNCTION
LOAD_CONST False
LOAD_CONST ('norm_cfg', 'norm_eval')
CALL_FUNCTION
LOAD_CONST ('backbone',)
CA... | _base_ = "../cascade_rcnn/cascade-mask-rcnn_x101-32x4d_fpn_1x_coco.py"
model = dict(
backbone=dict(norm_cfg=dict(type="SyncBN", requires_grad=True), norm_eval=False)
)
| data/mmdet-3.3.0/mmdet/.mim/configs/gcnet/cascade-mask-rcnn_x101-32x4d-syncbn_fpn_1x_coco.py | 114 | 84 | 151,758 |
LOAD_CONST '\nExample module that is used for a regression test where a module with\na name ending with "io" was skipped from patching (see #569).\n'
STORE_NAME __doc__
LOAD_CONST <code object file_contents at 0x7f8af0471ed0, file "f.py", line 7>
LOAD_CONST 'file_contents'
MAKE_FUNCTION
STORE_NAME file_contents
LOAD_C... | """
Example module that is used for a regression test where a module with
a name ending with "io" was skipped from patching (see #569).
"""
def file_contents(path):
"""Return the contents of the given path as byte array."""
with open(path, "rb") as f:
return f.read()
| data/pyfakefs-5.3.5/pyfakefs/tests/logsio.py | 144 | 84 | 8,890 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME smartsheet
STORE_NAME smartsheet
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_METHOD usefixtures
LOAD_CONST 'smart_setup'
CALL_METHOD
LOAD_BUILD_CLASS
LOAD_CONST <code object TestServerInfo at 0x7fab001364b0, file "f.py... | import pytest
import smartsheet
@pytest.mark.usefixtures("smart_setup")
class TestServerInfo:
def test_server_info(self, smart_setup):
smart = smart_setup["smart"]
info = smart.Server.server_info()
assert "en_US" in info.supported_locales
| data/smartsheet-python-sdk-3.0.2/tests/integration/test_server_info.py | 274 | 84 | 92,814 |
LOAD_CONST 0
LOAD_CONST ('annotations',)
IMPORT_NAME __future__
IMPORT_FROM annotations
STORE_NAME annotations
POP_TOP
LOAD_CONST 0
LOAD_CONST ('Executable', 'setup')
IMPORT_NAME cx_Freeze
IMPORT_FROM Executable
STORE_NAME Executable
IMPORT_FROM setup
STORE_NAME setup
POP_TOP
LOAD_NAME Executable
LOAD_CONST 'relimpor... | from __future__ import annotations
from cx_Freeze import Executable, setup
executables = [Executable("relimport.py")]
setup(
name="relimport",
version="0.1",
description="Sample cx_Freeze script for relative imports",
executables=executables,
)
| data/cx_Freeze-6.15.15/samples/relimport/setup.py | 148 | 84 | 175,515 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME webview
STORE_NAME webview
LOAD_CONST 1
LOAD_CONST ('run_test',)
IMPORT_NAME util
IMPORT_FROM run_test
STORE_NAME run_test
POP_TOP
LOAD_CONST <code object test_set_title at 0x7fab542adc00, file "f.py", line 6>
LOAD_CONST 'test_set_title'
MAKE_FUNCTION
STORE_NAME test_set_title... | import webview
from .util import run_test
def test_set_title():
window = webview.create_window("Set title test", "https://www.example.org")
run_test(webview, window, set_title)
def set_title(window):
window.set_title("New title")
| data/pywebview-4.4.1/tests/test_set_title.py | 236 | 84 | 390,109 |
LOAD_CONST 1
LOAD_CONST ('List',)
IMPORT_NAME abstract_nodes
IMPORT_FROM List
STORE_NAME AbstractList
POP_TOP
LOAD_CONST 1
LOAD_CONST ('Token',)
IMPORT_NAME ast
IMPORT_FROM Token
STORE_NAME Token
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object List at 0x7fadb5811c00, file "f.py", line 5>
LOAD_CONST 'List'
MAKE_FUNCT... | from .abstract_nodes import List as AbstractList
from .ast import Token
class List(AbstractList):
pass
class NumExprEvaluate(Token):
"""represents a call to :class:`numexpr`s :func:`evaluate`"""
__slots__ = _fields = ("expr",)
| data/sympy-1.12/sympy/codegen/pynodes.py | 290 | 84 | 228,317 |
LOAD_BUILD_CLASS
LOAD_CONST <code object Controller at 0x7faa5f6db390, file "f.py", line 1>
LOAD_CONST 'Controller'
MAKE_FUNCTION
LOAD_CONST 'Controller'
CALL_FUNCTION
STORE_NAME Controller
LOAD_CONST None
RETURN_VALUE
LOAD_NAME __name__
STORE_NAME __module__
LOAD_CONST 'Controller'
STORE_NAME __qualname__
LOAD_CONST... | class Controller:
def __init__(self, http_client, schema_client):
self.http_client = http_client
self.schema_client = schema_client
def get_usage(self, **kwargs):
resp, body = self.http_client.get("/v2/info/usage")
return body["usage"]
| data/python-glanceclient-4.4.0/glanceclient/v2/info.py | 259 | 84 | 248,746 |
LOAD_CONST 1
LOAD_CONST ('*',)
IMPORT_NAME hub_interface
IMPORT_STAR
LOAD_CONST 1
LOAD_CONST ('*',)
IMPORT_NAME model
IMPORT_STAR
LOAD_CONST 1
LOAD_CONST ('*',)
IMPORT_NAME enc_dec
IMPORT_STAR
LOAD_CONST 1
LOAD_CONST ('*',)
IMPORT_NAME model_camembert
IMPORT_STAR
LOAD_CONST 1
LOAD_CONST ('*',)
IMPORT_NAME model_got... | from .hub_interface import * # noqa
from .model import * # noqa
from .enc_dec import * # noqa
from .model_camembert import * # noqa
from .model_gottbert import * # noqa
from .model_xlmr import * # noqa
| data/fairseq-0.12.2/fairseq/models/roberta/__init__.py | 116 | 84 | 120,079 |
LOAD_CONST 'DO NOT EDIT.\n\nThis file was autogenerated. Do not edit it by hand,\nsince your modifications would be overwritten.\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('Xception',)
IMPORT_NAME keras.src.applications.xception
IMPORT_FROM Xception
STORE_NAME Xception
POP_TOP
LOAD_CONST 0
LOAD_CONST ('decode_pre... | """DO NOT EDIT.
This file was autogenerated. Do not edit it by hand,
since your modifications would be overwritten.
"""
from keras.src.applications.xception import Xception
from keras.src.applications.xception import decode_predictions
from keras.src.applications.xception import preprocess_input
| data/keras-3.0.5/keras/applications/xception/__init__.py | 160 | 84 | 349,111 |
LOAD_CONST 'Copyright (C) 2020 Nidhal Baccouri'
STORE_NAME __copyright__
LOAD_CONST 0
LOAD_CONST ('BaseTranslator',)
IMPORT_NAME deep_translator.base
IMPORT_FROM BaseTranslator
STORE_NAME BaseTranslator
POP_TOP
LOAD_CONST <code object <dictcomp> at 0x7fab8237fdb0, file "f.py", line 5>
LOAD_CONST '<dictcomp>'
MAKE_FUN... | __copyright__ = "Copyright (C) 2020 Nidhal Baccouri"
from deep_translator.base import BaseTranslator
__engines__ = {
translator.__name__.replace("Translator", "").lower(): translator
for translator in BaseTranslator.__subclasses__()
}
| data/deep_translator-1.11.4/deep_translator/engines.py | 194 | 84 | 57,398 |
LOAD_CONST 0
LOAD_CONST ('Optional',)
IMPORT_NAME typing
IMPORT_FROM Optional
STORE_NAME Optional
POP_TOP
LOAD_CONST 0
LOAD_CONST ('dataclass',)
IMPORT_NAME pydantic.dataclasses
IMPORT_FROM dataclass
STORE_NAME dataclass
POP_TOP
LOAD_NAME dataclass
LOAD_BUILD_CLASS
LOAD_CONST <code object GITModel at 0x7fab700b4660, ... | from typing import Optional
from pydantic.dataclasses import dataclass
@dataclass
class GITModel:
branch: Optional[str] = None
tag: Optional[str] = None
commit: Optional[str] = None
dirty: Optional[bool] = None
origin: Optional[str] = None
| data/safety_schemas-0.0.2/safety_schemas/models/git.py | 300 | 84 | 444,161 |
LOAD_CONST 1
LOAD_CONST ('ContainerRegistryManagementClient',)
IMPORT_NAME _container_registry_management_client
IMPORT_FROM ContainerRegistryManagementClient
STORE_NAME ContainerRegistryManagementClient
POP_TOP
LOAD_CONST 'ContainerRegistryManagementClient'
BUILD_LIST
STORE_NAME __all__
SETUP_EXCEPT to 42
LOAD_CONS... | from ._container_registry_management_client import ContainerRegistryManagementClient
__all__ = ["ContainerRegistryManagementClient"]
try:
from ._patch import patch_sdk # type: ignore
patch_sdk()
except ImportError:
pass
from ._version import VERSION
__version__ = VERSION
| data/azure-mgmt-containerregistry-10.3.0/azure/mgmt/containerregistry/__init__.py | 183 | 84 | 83,624 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.