input stringlengths 28 198k | output stringlengths 3 71k | file stringlengths 19 330 | input_tokens int64 5 159k | output_tokens int64 3 9.07k | __index_level_0__ int64 2 449k |
|---|---|---|---|---|---|
LOAD_CONST 'Constants for the Whirlpool Appliances integration.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('Region',)
IMPORT_NAME whirlpool.backendselector
IMPORT_FROM Region
STORE_NAME Region
POP_TOP
LOAD_CONST 'whirlpool'
STORE_NAME DOMAIN
LOAD_NAME Region
LOAD_ATTR EU
LOAD_NAME Region
LOAD_ATTR US
LOAD_CONST (... | """Constants for the Whirlpool Appliances integration."""
from whirlpool.backendselector import Region
DOMAIN = "whirlpool"
CONF_REGIONS_MAP = {
"EU": Region.EU,
"US": Region.US,
}
| data/homeassistant-2024.2.2/homeassistant/components/whirlpool/const.py | 98 | 69 | 407,161 |
LOAD_CONST 0
LOAD_CONST ('absolute_import', 'division', 'print_function')
IMPORT_NAME __future__
IMPORT_FROM absolute_import
STORE_NAME absolute_import
IMPORT_FROM division
STORE_NAME division
IMPORT_FROM print_function
STORE_NAME print_function
POP_TOP
LOAD_NAME type
STORE_NAME __metaclass__
LOAD_CONST 5
LOAD_CONST ... | from __future__ import absolute_import, division, print_function
__metaclass__ = type
from .....plugins.module_utils.my_util import hello
def test_hello():
assert hello("Ansibull") == "Hello Ansibull"
| data/ansible-core-2.16.3/test/integration/targets/ansible-test-docker/ansible_collections/ns/col/tests/unit/plugins/module_utils/test_my_util.py | 190 | 69 | 437,355 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME typer
STORE_NAME typer
LOAD_NAME typer
LOAD_ATTR Option
LOAD_CONST Ellipsis
LOAD_CONST True
LOAD_CONST ('prompt',)
CALL_FUNCTION
BUILD_TUPLE
LOAD_NAME str
LOAD_NAME str
LOAD_CONST ('name', 'lastname')
BUILD_CONST_KEY_MAP
LOAD_CONST <code object main at 0x7fab641e7030, file "f.p... | import typer
def main(name: str, lastname: str = typer.Option(..., prompt=True)):
print(f"Hello {name} {lastname}")
if __name__ == "__main__":
typer.run(main)
| data/typer-0.9.0/docs_src/options/prompt/tutorial001.py | 174 | 69 | 224,410 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME unittest
STORE_NAME unittest
LOAD_CONST 0
LOAD_CONST ('install_tests_in_module_dict',)
IMPORT_NAME multiprocess.tests
IMPORT_FROM install_tests_in_module_dict
STORE_NAME install_tests_in_module_dict
POP_TOP
LOAD_NAME install_tests_in_module_dict
LOAD_NAME globals
CALL_FUNCTION... | import unittest
from multiprocess.tests import install_tests_in_module_dict
install_tests_in_module_dict(globals(), "forkserver", exclude_types=True)
if __name__ == "__main__":
unittest.main()
| data/multiprocess-0.70.16/py3.13/multiprocess/tests/test_multiprocessing_forkserver/test_misc.py | 144 | 69 | 144,771 |
LOAD_CONST 0
LOAD_CONST ('Entity',)
IMPORT_NAME office365.sharepoint.entity
IMPORT_FROM Entity
STORE_NAME Entity
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object SiteDesignPrincipal at 0x7fab700e40c0, file "f.py", line 4>
LOAD_CONST 'SiteDesignPrincipal'
MAKE_FUNCTION
LOAD_CONST 'SiteDesignPrincipal'
LOAD_NAME Entity
... | from office365.sharepoint.entity import Entity
class SiteDesignPrincipal(Entity):
@property
def entity_type_name(self):
return (
"Microsoft.SharePoint.Utilities.WebTemplateExtensions.SiteDesignPrincipal"
)
| data/Office365-REST-Python-Client-2.5.5/office365/sharepoint/sitedesigns/principal.py | 223 | 69 | 189,156 |
LOAD_CONST 1
LOAD_CONST ('AsymmetricLossMultiLabel', 'AsymmetricLossSingleLabel')
IMPORT_NAME asymmetric_loss
IMPORT_FROM AsymmetricLossMultiLabel
STORE_NAME AsymmetricLossMultiLabel
IMPORT_FROM AsymmetricLossSingleLabel
STORE_NAME AsymmetricLossSingleLabel
POP_TOP
LOAD_CONST 1
LOAD_CONST ('BinaryCrossEntropy',)
IMPOR... | from .asymmetric_loss import AsymmetricLossMultiLabel, AsymmetricLossSingleLabel
from .binary_cross_entropy import BinaryCrossEntropy
from .cross_entropy import LabelSmoothingCrossEntropy, SoftTargetCrossEntropy
from .jsd import JsdCrossEntropy
| data/timm-0.9.16/timm/loss/__init__.py | 203 | 69 | 22,440 |
LOAD_CONST 0
LOAD_CONST ('Package', 'package')
IMPORT_NAME localstack.packages
IMPORT_FROM Package
STORE_NAME Package
IMPORT_FROM package
STORE_NAME package
POP_TOP
LOAD_NAME package
LOAD_CONST 'dynamodb-local'
LOAD_CONST ('name',)
CALL_FUNCTION
LOAD_NAME Package
LOAD_CONST ('return',)
BUILD_CONST_KEY_MAP
LOAD_CONST ... | from localstack.packages import Package, package
@package(name="dynamodb-local")
def dynamodb_local_package() -> Package:
from localstack.services.dynamodb.packages import dynamodblocal_package
return dynamodblocal_package
| data/localstack-core-3.1.0/localstack/services/dynamodb/plugins.py | 187 | 69 | 225,475 |
LOAD_CONST 1
LOAD_CONST ('ffi', 'dlls')
IMPORT_NAME _util
IMPORT_FROM ffi
STORE_NAME ffi
IMPORT_FROM dlls
STORE_NAME dlls
POP_TOP
LOAD_NAME ffi
LOAD_METHOD cdef
LOAD_CONST '\n\nDWORD WINAPI GetTickCount(void);\n\n'
CALL_METHOD
POP_TOP
LOAD_CONST <code object _GetTickCount at 0x7f8aac1faf60, file "f.py", line 12>
LOA... | from ._util import ffi, dlls
ffi.cdef(
"""
DWORD WINAPI GetTickCount(void);
"""
)
def _GetTickCount():
return dlls.kernel32.GetTickCount()
| data/pywin32-ctypes-0.2.2/win32ctypes/core/cffi/_time.py | 156 | 69 | 427,069 |
LOAD_CONST 0
LOAD_CONST ('annotations',)
IMPORT_NAME __future__
IMPORT_FROM annotations
STORE_NAME annotations
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ibis
STORE_NAME ibis
LOAD_CONST <code object test_string_length at 0x7fab823ac1e0, file "f.py", line 6>
LOAD_CONST 'test_string_length'
MAKE_FUNCTION
STORE_NA... | from __future__ import annotations
import ibis
def test_string_length(con):
t = ibis.memtable({"s": ["aaa", "a", "aa"]})
assert con.execute(t.s.length()).gt(0).all()
| data/ibis_framework-8.0.0/ibis/backends/datafusion/tests/test_string.py | 189 | 69 | 161,469 |
LOAD_CONST 0
LOAD_CONST ('add_qt6_dependencies', 'pyqt6_library_info')
IMPORT_NAME PyInstaller.utils.hooks.qt
IMPORT_FROM add_qt6_dependencies
STORE_NAME add_qt6_dependencies
IMPORT_FROM pyqt6_library_info
STORE_NAME pyqt6_library_info
POP_TOP
LOAD_NAME add_qt6_dependencies
LOAD_NAME __file__
CALL_FUNCTION
UNPACK_SEQU... | from PyInstaller.utils.hooks.qt import add_qt6_dependencies, pyqt6_library_info
hiddenimports, binaries, datas = add_qt6_dependencies(__file__)
binaries += pyqt6_library_info.collect_qtnetwork_files()
| data/pyinstaller-6.4.0/PyInstaller/hooks/hook-PyQt6.QtNetwork.py | 140 | 69 | 185,444 |
LOAD_CONST 0
LOAD_CONST ('base',)
IMPORT_NAME keystoneclient.v3.contrib.federation
IMPORT_FROM base
STORE_NAME federation_base
POP_TOP
LOAD_CONST 0
LOAD_CONST ('projects',)
IMPORT_NAME keystoneclient.v3
IMPORT_FROM projects
STORE_NAME projects
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object ProjectManager at 0x7f8e4... | from keystoneclient.v3.contrib.federation import base as federation_base
from keystoneclient.v3 import projects
class ProjectManager(federation_base.EntityManager):
object_type = "projects"
resource_class = projects.Project
| data/python-keystoneclient-5.3.0/keystoneclient/v3/contrib/federation/projects.py | 190 | 69 | 421,238 |
LOAD_CONST 0
LOAD_CONST ('DirectoryObject',)
IMPORT_NAME office365.directory.object
IMPORT_FROM DirectoryObject
STORE_NAME DirectoryObject
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object GroupSettingTemplate at 0x7fab64179e40, file "f.py", line 4>
LOAD_CONST 'GroupSettingTemplate'
MAKE_FUNCTION
LOAD_CONST 'GroupSetti... | from office365.directory.object import DirectoryObject
class GroupSettingTemplate(DirectoryObject):
"""Group setting templates represent system-defined settings available to the tenant. Group settings can be
created based on the available groupSettingTemplates, and values changed from their preset defaults.
... | data/Office365-REST-Python-Client-2.5.5/office365/directory/groups/setting_template.py | 178 | 69 | 188,879 |
LOAD_CONST nan
STORE_NAME a
LOAD_CONST nan
STORE_NAME b
LOAD_CONST inf
STORE_NAME c
LOAD_CONST -inf
STORE_NAME d
LOAD_CONST None
RETURN_VALUE | a = 1e300 * 1e300 * 0
b = -1e300 * 1e300 * 0
c = 1e300 * 1e300
d = -1e300 * 1e300
| data/uncompyle6-3.9.0/test/simple_source/expression/01_float.py | 33 | 69 | 442,201 |
LOAD_CONST 0
LOAD_CONST ('annotations',)
IMPORT_NAME __future__
IMPORT_FROM annotations
STORE_NAME annotations
POP_TOP
LOAD_CONST 0
LOAD_CONST ('setup',)
IMPORT_NAME setuptools
IMPORT_FROM setup
STORE_NAME setup
POP_TOP
LOAD_NAME setup
LOAD_CONST 'small_fake_with_unpinned_deps'
LOAD_CONST 0.1
LOAD_CONST 'small-fak... | from __future__ import annotations
from setuptools import setup
setup(
name="small_fake_with_unpinned_deps",
version=0.1,
install_requires=["small-fake-a", "small-fake-b"],
)
| data/pip-tools-7.4.0/tests/test_data/packages/small_fake_with_unpinned_deps/setup.py | 121 | 69 | 438,741 |
LOAD_CONST 1
LOAD_CONST ('Error', 'Result', 'Results', 'SecretScanCollection')
IMPORT_NAME secret_scan_collection
IMPORT_FROM Error
STORE_NAME Error
IMPORT_FROM Result
STORE_NAME Result
IMPORT_FROM Results
STORE_NAME Results
IMPORT_FROM SecretScanCollection
STORE_NAME SecretScanCollection
POP_TOP
LOAD_CONST 1
LOAD_CON... | from .secret_scan_collection import Error, Result, Results, SecretScanCollection
from .secret_scanner import SecretScanner
__all__ = [
"Error",
"Result",
"Results",
"SecretScanCollection",
"SecretScanner",
]
| data/ggshield-1.24.0/ggshield/verticals/secret/__init__.py | 134 | 69 | 401,112 |
LOAD_CONST 0
LOAD_CONST ('add_qt5_dependencies', 'pyqt5_library_info')
IMPORT_NAME PyInstaller.utils.hooks.qt
IMPORT_FROM add_qt5_dependencies
STORE_NAME add_qt5_dependencies
IMPORT_FROM pyqt5_library_info
STORE_NAME pyqt5_library_info
POP_TOP
LOAD_NAME add_qt5_dependencies
LOAD_NAME __file__
CALL_FUNCTION
UNPACK_SEQU... | from PyInstaller.utils.hooks.qt import add_qt5_dependencies, pyqt5_library_info
hiddenimports, binaries, datas = add_qt5_dependencies(__file__)
binaries += pyqt5_library_info.collect_qtnetwork_files()
| data/pyinstaller-6.4.0/PyInstaller/hooks/hook-PyQt5.QtNetwork.py | 140 | 69 | 185,405 |
LOAD_CONST 0
LOAD_CONST ('GiModuleInfo',)
IMPORT_NAME PyInstaller.utils.hooks.gi
IMPORT_FROM GiModuleInfo
STORE_NAME GiModuleInfo
POP_TOP
LOAD_NAME GiModuleInfo
LOAD_CONST 'GstGLEGL'
LOAD_CONST '1.0'
CALL_FUNCTION
STORE_NAME module_info
LOAD_NAME module_info
LOAD_ATTR available
POP_JUMP_IF_FALSE
LOAD_NAME module_inf... | from PyInstaller.utils.hooks.gi import GiModuleInfo
module_info = GiModuleInfo("GstGLEGL", "1.0")
if module_info.available:
binaries, datas, hiddenimports = module_info.collect_typelib_data()
| data/pyinstaller-6.4.0/PyInstaller/hooks/hook-gi.repository.GstGLEGL.py | 113 | 69 | 185,402 |
LOAD_CONST 0
LOAD_CONST ('Enum', 'unique', 'auto')
IMPORT_NAME enum
IMPORT_FROM Enum
STORE_NAME Enum
IMPORT_FROM unique
STORE_NAME unique
IMPORT_FROM auto
STORE_NAME auto
POP_TOP
LOAD_NAME unique
LOAD_BUILD_CLASS
LOAD_CONST <code object StreamCxnState at 0x7fab640fa540, file "f.py", line 4>
LOAD_CONST 'StreamCxnState'... | from enum import Enum, unique, auto
@unique
class StreamCxnState(Enum):
Initial = auto()
Connecting = auto()
MessageProcessing = auto()
Disconnecting = auto()
Disconnected = auto()
Disposed = auto()
| data/refinitiv-data-1.6.0/refinitiv/data/delivery/_stream/stream_cxn_state.py | 214 | 69 | 195,926 |
LOAD_BUILD_CLASS
LOAD_CONST <code object PyiBlockCipher at 0x7fab700adf60, file "f.py", line 1>
LOAD_CONST 'PyiBlockCipher'
MAKE_FUNCTION
LOAD_CONST 'PyiBlockCipher'
CALL_FUNCTION
STORE_NAME PyiBlockCipher
LOAD_CONST None
RETURN_VALUE
LOAD_NAME __name__
STORE_NAME __module__
LOAD_CONST 'PyiBlockCipher'
STORE_NAME __qu... | class PyiBlockCipher:
def __init__(self, key=None):
from PyInstaller.exceptions import RemovedCipherFeatureError
raise RemovedCipherFeatureError(
"Please remove cipher and block_cipher parameters from your spec file."
)
| data/pyinstaller-6.4.0/PyInstaller/archive/pyz_crypto.py | 240 | 69 | 185,119 |
LOAD_CONST 1
LOAD_CONST ('ParameterStoreResource', 'ParameterStoreTag', 'SSMResource', 'parameter_store_resource', 'ssm_resource')
IMPORT_NAME resources
IMPORT_FROM ParameterStoreResource
STORE_NAME ParameterStoreResource
IMPORT_FROM ParameterStoreTag
STORE_NAME ParameterStoreTag
IMPORT_FROM SSMResource
STORE_NAME SSMR... | from .resources import (
ParameterStoreResource as ParameterStoreResource,
ParameterStoreTag as ParameterStoreTag,
SSMResource as SSMResource,
parameter_store_resource as parameter_store_resource,
ssm_resource as ssm_resource,
)
| data/dagster-aws-0.22.5/dagster_aws/ssm/__init__.py | 104 | 69 | 154,884 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME typer
STORE_NAME typer
LOAD_NAME typer
LOAD_ATTR Argument
LOAD_CONST 'World'
LOAD_CONST '✨username✨'
LOAD_CONST ('metavar',)
CALL_FUNCTION
BUILD_TUPLE
LOAD_NAME str
LOAD_CONST ('name',)
BUILD_CONST_KEY_MAP
LOAD_CONST <code object main at 0x7fab641e74b0, file "f.py", line 4>
LOA... | import typer
def main(name: str = typer.Argument("World", metavar="✨username✨")):
print(f"Hello {name}")
if __name__ == "__main__":
typer.run(main)
| data/typer-0.9.0/docs_src/arguments/help/tutorial006.py | 168 | 69 | 224,557 |
LOAD_CONST '\nUtility functions and classes.\n\nYou have to import sub-modules of `healpy.utils` directly,\nthe `healpy.utils` namespace is empty.\n\nExamples::\n from healpy.utils.deprecation import HealpyDeprecationWarning\n'
STORE_NAME __doc__
LOAD_CONST None
RETURN_VALUE | """
Utility functions and classes.
You have to import sub-modules of `healpy.utils` directly,
the `healpy.utils` namespace is empty.
Examples::
from healpy.utils.deprecation import HealpyDeprecationWarning
"""
| data/healpy-1.16.6/healpy/utils/__init__.py | 82 | 69 | 405,800 |
LOAD_CONST 0
LOAD_CONST ('absolute_import', 'division', 'print_function')
IMPORT_NAME __future__
IMPORT_FROM absolute_import
STORE_NAME absolute_import
IMPORT_FROM division
STORE_NAME division
IMPORT_FROM print_function
STORE_NAME print_function
POP_TOP
LOAD_NAME type
STORE_NAME __metaclass__
SETUP_EXCEPT to 38
LOAD... | from __future__ import absolute_import, division, print_function
__metaclass__ = type
try:
import __builtin__
except ImportError:
BUILTINS = "builtins"
else:
BUILTINS = "__builtin__"
| data/ansible-9.2.0/ansible_collections/cisco/aci/tests/unit/compat/builtins.py | 161 | 69 | 4,419 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME typer
STORE_NAME typer
LOAD_CONST 1
LOAD_CONST ('get_cpu_limit',)
IMPORT_NAME co_utils
IMPORT_FROM get_cpu_limit
STORE_NAME get_cpu_limit
POP_TOP
LOAD_NAME typer
LOAD_METHOD Typer
CALL_METHOD
STORE_NAME app
LOAD_NAME app
LOAD_METHOD command
CALL_METHOD
LOAD_CONST <code object... | import typer
from .co_utils import get_cpu_limit
app = typer.Typer()
@app.command()
def main():
"""This function returns an integer corresponding to the number of cores available"""
typer.echo(get_cpu_limit())
| data/code_ocean_aux_tools-1.1.3/src/co_tools/get_cpu_count.py | 149 | 69 | 199,671 |
LOAD_CONST 0
LOAD_CONST ('File',)
IMPORT_NAME twisted.web.static
IMPORT_FROM File
STORE_NAME File
POP_TOP
LOAD_CONST 0
LOAD_CONST ('Klein',)
IMPORT_NAME klein
IMPORT_FROM Klein
STORE_NAME Klein
POP_TOP
LOAD_NAME Klein
CALL_FUNCTION
STORE_NAME app
LOAD_NAME app
LOAD_ATTR route
LOAD_CONST '/'
LOAD_CONST True
LOAD_CONS... | from twisted.web.static import File
from klein import Klein
app = Klein()
@app.route("/", branch=True)
def pg_index(request):
return File("./")
app.run("localhost", 8080)
| data/klein-23.12.0/docs/introduction/codeexamples/staticFiles.py | 172 | 69 | 444,919 |
LOAD_CONST 0
LOAD_CONST ('annotations',)
IMPORT_NAME __future__
IMPORT_FROM annotations
STORE_NAME annotations
POP_TOP
LOAD_CONST 0
LOAD_CONST ('AirflowPlugin',)
IMPORT_NAME airflow.plugins_manager
IMPORT_FROM AirflowPlugin
STORE_NAME AirflowPlugin
POP_TOP
LOAD_CONST 0
LOAD_CONST ('events_listener',)
IMPORT_NAME list... | from __future__ import annotations
from airflow.plugins_manager import AirflowPlugin
from listener import events_listener
import os
class AcceldataListenerPlugin(AirflowPlugin):
name = "AcceldataListenerPlugin"
listeners = [events_listener]
| data/acceldata_airflow_sdk-3.0.0/adoc_listener_plugin/plugins/listener_plugin.py | 238 | 69 | 422,501 |
LOAD_CONST 1
LOAD_CONST ('AzureMachineLearningWorkspaces',)
IMPORT_NAME _azure_machine_learning_workspaces
IMPORT_FROM AzureMachineLearningWorkspaces
STORE_NAME AzureMachineLearningWorkspaces
POP_TOP
LOAD_CONST 1
LOAD_CONST ('VERSION',)
IMPORT_NAME _version
IMPORT_FROM VERSION
STORE_NAME VERSION
POP_TOP
LOAD_NAME VER... | from ._azure_machine_learning_workspaces import AzureMachineLearningWorkspaces
from ._version import VERSION
__version__ = VERSION
__all__ = ["AzureMachineLearningWorkspaces"]
from ._patch import patch_sdk
patch_sdk()
| data/azure-ai-ml-1.13.0/azure/ai/ml/_restclient/v2023_02_01_preview/__init__.py | 147 | 69 | 425,596 |
LOAD_CONST 1
LOAD_CONST ('AzureMachineLearningWorkspaces',)
IMPORT_NAME _azure_machine_learning_workspaces
IMPORT_FROM AzureMachineLearningWorkspaces
STORE_NAME AzureMachineLearningWorkspaces
POP_TOP
LOAD_CONST 1
LOAD_CONST ('VERSION',)
IMPORT_NAME _version
IMPORT_FROM VERSION
STORE_NAME VERSION
POP_TOP
LOAD_NAME VER... | from ._azure_machine_learning_workspaces import AzureMachineLearningWorkspaces
from ._version import VERSION
__version__ = VERSION
__all__ = ["AzureMachineLearningWorkspaces"]
from ._patch import patch_sdk
patch_sdk()
| data/azure-ai-ml-1.13.0/azure/ai/ml/_restclient/v2022_10_01/__init__.py | 147 | 69 | 425,326 |
LOAD_BUILD_CLASS
LOAD_CONST <code object ClientTypeError at 0x7fab8239a9c0, file "f.py", line 1>
LOAD_CONST 'ClientTypeError'
MAKE_FUNCTION
LOAD_CONST 'ClientTypeError'
LOAD_NAME Exception
CALL_FUNCTION
STORE_NAME ClientTypeError
LOAD_BUILD_CLASS
LOAD_CONST <code object IgnoredResultWarning at 0x7fab8239a8a0, file "f.... | class ClientTypeError(Exception):
"""Raised when trying to open a Client on a non-catalog STAC Object."""
pass
class IgnoredResultWarning(RuntimeWarning):
"""
Warning raised when a 'modifier' callable returns a result.
"""
| data/pystac-client-0.7.5/pystac_client/errors.py | 249 | 69 | 224,847 |
LOAD_CONST 'Provide the RedditorList class.'
STORE_NAME __doc__
LOAD_CONST 1
LOAD_CONST ('BaseList',)
IMPORT_NAME base
IMPORT_FROM BaseList
STORE_NAME BaseList
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object RedditorList at 0x7fab82343b70, file "f.py", line 6>
LOAD_CONST 'RedditorList'
MAKE_FUNCTION
LOAD_CONST 'Redd... | """Provide the RedditorList class."""
from .base import BaseList
class RedditorList(BaseList):
"""A list of :class:`.Redditor` objects. Works just like a regular list."""
CHILD_ATTRIBUTE = "children"
| data/praw-7.7.1/praw/models/list/redditor.py | 189 | 69 | 205,112 |
LOAD_CONST 0
LOAD_CONST ('warn',)
IMPORT_NAME warnings
IMPORT_FROM warn
STORE_NAME warn
POP_TOP
LOAD_NAME warn
LOAD_CONST 'IPython.utils.daemonize has moved to ipyparallel.apps.daemonize since IPython 4.0'
LOAD_NAME DeprecationWarning
LOAD_CONST 2
LOAD_CONST ('stacklevel',)
CALL_FUNCTION
POP_TOP
LOAD_CONST 0
LOAD_... | from warnings import warn
warn(
"IPython.utils.daemonize has moved to ipyparallel.apps.daemonize since IPython 4.0",
DeprecationWarning,
stacklevel=2,
)
from ipyparallel.apps.daemonize import daemonize
| data/ipython-8.21.0/IPython/utils/daemonize.py | 117 | 69 | 154,504 |
LOAD_CONST <code object do_commands at 0x7faa75305a50, file "f.py", line 1>
LOAD_CONST 'do_commands'
MAKE_FUNCTION
STORE_NAME do_commands
LOAD_CONST None
RETURN_VALUE
LOAD_FAST arg
POP_JUMP_IF_TRUE
LOAD_CONST 1
STORE_FAST bnum
JUMP_FORWARD to 44
SETUP_EXCEPT to 24
LOAD_GLOBAL int
LOAD_FAST arg
CALL_FUNCTION
STORE_F... | def do_commands(self, arg):
if not arg:
bnum = 1
else:
try:
bnum = int(arg)
except:
self.error("Usage:")
return
self.commands_bnum = bnum
| data/uncompyle6-3.9.0/test/simple_source/bug35/06_try_return.py | 153 | 69 | 442,154 |
LOAD_CONST 0
LOAD_CONST ('notfound_view_config',)
IMPORT_NAME pyramid.view
IMPORT_FROM notfound_view_config
STORE_NAME notfound_view_config
POP_TOP
LOAD_NAME notfound_view_config
LOAD_CONST 'myproject:templates/404.jinja2'
LOAD_CONST ('renderer',)
CALL_FUNCTION
LOAD_CONST <code object notfound_view at 0x7faa51f69e40, ... | from pyramid.view import notfound_view_config
@notfound_view_config(renderer="myproject:templates/404.jinja2")
def notfound_view(request):
request.response.status = 404
return {}
| data/pyramid-2.0.2/docs/narr/myproject/myproject/views/notfound.py | 155 | 69 | 20,035 |
LOAD_CONST 1
LOAD_CONST ('AzureMachineLearningWorkspaces',)
IMPORT_NAME _azure_machine_learning_workspaces
IMPORT_FROM AzureMachineLearningWorkspaces
STORE_NAME AzureMachineLearningWorkspaces
POP_TOP
LOAD_CONST 1
LOAD_CONST ('VERSION',)
IMPORT_NAME _version
IMPORT_FROM VERSION
STORE_NAME VERSION
POP_TOP
LOAD_NAME VER... | from ._azure_machine_learning_workspaces import AzureMachineLearningWorkspaces
from ._version import VERSION
__version__ = VERSION
__all__ = ["AzureMachineLearningWorkspaces"]
from ._patch import patch_sdk
patch_sdk()
| data/azure-ai-ml-1.13.0/azure/ai/ml/_restclient/v2022_10_01_preview/__init__.py | 147 | 69 | 425,396 |
LOAD_CONST 1
LOAD_CONST ('DynamicModelInterface',)
IMPORT_NAME interfaces.model_interface
IMPORT_FROM DynamicModelInterface
STORE_NAME DynamicModelInterface
POP_TOP
LOAD_CONST 1
LOAD_CONST ('TimeSeriesData',)
IMPORT_NAME data.series_data
IMPORT_FROM TimeSeriesData
STORE_NAME TimeSeriesData
POP_TOP
LOAD_CONST 1
LOAD_C... | from .interfaces.model_interface import DynamicModelInterface
from .data.series_data import TimeSeriesData
from .data.interval_data import IntervalData
from .data.scalar_data import ScalarData
from .data.get_cuid import get_indexed_cuid
| data/Pyomo-6.7.0/pyomo/contrib/mpc/__init__.py | 179 | 69 | 23,455 |
LOAD_CONST <code object unique at 0x7fab802486f0, file "f.py", line 1>
LOAD_CONST 'unique'
MAKE_FUNCTION
STORE_NAME unique
LOAD_CONST None
RETURN_VALUE
BUILD_LIST
STORE_FAST values
LOAD_GLOBAL list
LOAD_FAST d
LOAD_METHOD keys
CALL_METHOD
CALL_FUNCTION
STORE_FAST keys
SETUP_LOOP to 74
LOAD_FAST keys
GET_ITER
FOR_ITE... | def unique(d):
values = []
keys = list(d.keys())
for key in keys:
value = d.get(key, None)
if value in values:
d.pop(key, None)
continue
values.append(value)
| data/python-benedict-0.33.1/benedict/core/unique.py | 164 | 69 | 211,418 |
LOAD_CONST 1
LOAD_CONST ('compose',)
IMPORT_NAME compose
IMPORT_FROM compose
STORE_NAME compose
POP_TOP
LOAD_CONST 1
LOAD_CONST ('initialize', 'initialize_config_dir', 'initialize_config_module')
IMPORT_NAME initialize
IMPORT_FROM initialize
STORE_NAME initialize
IMPORT_FROM initialize_config_dir
STORE_NAME initialize... | from .compose import compose
from .initialize import initialize, initialize_config_dir, initialize_config_module
__all__ = [
"compose",
"initialize",
"initialize_config_module",
"initialize_config_dir",
]
| data/hydra-core-1.3.2/hydra/experimental/__init__.py | 139 | 69 | 434,504 |
LOAD_CONST 'Implements protocols on top of lower level libraries to talk to instruments.\n\n\n:copyright: 2014-2020 by PyVISA-py Authors, see AUTHORS for more details.\n:license: MIT, see LICENSE for more details.\n\n'
STORE_NAME __doc__
LOAD_CONST None
RETURN_VALUE | """Implements protocols on top of lower level libraries to talk to instruments.
:copyright: 2014-2020 by PyVISA-py Authors, see AUTHORS for more details.
:license: MIT, see LICENSE for more details.
"""
| data/PyVISA-py-0.7.1/pyvisa_py/protocols/__init__.py | 82 | 69 | 421,743 |
LOAD_CONST 0
LOAD_CONST ('List',)
IMPORT_NAME typing
IMPORT_FROM List
STORE_NAME List
POP_TOP
LOAD_CONST 0
LOAD_CONST ('BaseModel',)
IMPORT_NAME pydantic
IMPORT_FROM BaseModel
STORE_NAME BaseModel
POP_TOP
LOAD_CONST ('Tag', 'TagsResult')
STORE_NAME __all__
LOAD_BUILD_CLASS
LOAD_CONST <code object Tag at 0x7faa74b64c... | from typing import List
from pydantic import BaseModel
__all__ = ("Tag", "TagsResult")
class Tag(BaseModel):
tag_name: str
class TagsResult(BaseModel):
tags: List[str]
tagsCount: int
| data/pornhub_api-0.3.0/pornhub_api/schemas/tag.py | 304 | 69 | 22,522 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME yatest.common
IMPORT_FROM common
STORE_NAME yc
POP_TOP
LOAD_CONST <code object test_export_metrics at 0x7fabdc1d0c00, file "f.py", line 4>
LOAD_CONST 'test_export_metrics'
MAKE_FUNCTION
STORE_NAME test_export_metrics
LOAD_CONST None
RETURN_VALUE
LOAD_FAST metrics
LOAD_METHOD s... | import yatest.common as yc
def test_export_metrics(metrics):
metrics.set_benchmark(
yc.execute_benchmark(
"library/cpp/testing/benchmark/examples/examples", threads=8
)
)
| data/catboost-1.2.2/catboost_all_src/library/cpp/testing/benchmark/examples/metrics/main.py | 145 | 69 | 177,254 |
LOAD_CONST 1
LOAD_CONST ('AzureMachineLearningWorkspaces',)
IMPORT_NAME _azure_machine_learning_workspaces
IMPORT_FROM AzureMachineLearningWorkspaces
STORE_NAME AzureMachineLearningWorkspaces
POP_TOP
LOAD_CONST 1
LOAD_CONST ('VERSION',)
IMPORT_NAME _version
IMPORT_FROM VERSION
STORE_NAME VERSION
POP_TOP
LOAD_NAME VER... | from ._azure_machine_learning_workspaces import AzureMachineLearningWorkspaces
from ._version import VERSION
__version__ = VERSION
__all__ = ["AzureMachineLearningWorkspaces"]
from ._patch import patch_sdk
patch_sdk()
| data/azure-ai-ml-1.13.0/azure/ai/ml/_restclient/workspace_dataplane/__init__.py | 147 | 69 | 424,630 |
LOAD_CONST 0
LOAD_CONST ('GiModuleInfo',)
IMPORT_NAME PyInstaller.utils.hooks.gi
IMPORT_FROM GiModuleInfo
STORE_NAME GiModuleInfo
POP_TOP
LOAD_NAME GiModuleInfo
LOAD_CONST 'GstTranscoder'
LOAD_CONST '1.0'
CALL_FUNCTION
STORE_NAME module_info
LOAD_NAME module_info
LOAD_ATTR available
POP_JUMP_IF_FALSE
LOAD_NAME modul... | from PyInstaller.utils.hooks.gi import GiModuleInfo
module_info = GiModuleInfo("GstTranscoder", "1.0")
if module_info.available:
binaries, datas, hiddenimports = module_info.collect_typelib_data()
| data/pyinstaller-6.4.0/PyInstaller/hooks/hook-gi.repository.GstTranscoder.py | 113 | 69 | 185,455 |
LOAD_CONST 0
LOAD_CONST ('FunctionName',)
IMPORT_NAME localstack.services.stepfunctions.asl.component.intrinsic.functionname.function_name
IMPORT_FROM FunctionName
STORE_NAME FunctionName
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object CustomFunctionName at 0x7fab703068a0, file "f.py", line 6>
LOAD_CONST 'CustomFunct... | from localstack.services.stepfunctions.asl.component.intrinsic.functionname.function_name import (
FunctionName,
)
class CustomFunctionName(FunctionName):
def __init__(self, name: str):
super().__init__(name=name)
| data/localstack-core-3.1.0/localstack/services/stepfunctions/asl/component/intrinsic/functionname/custom_function_name.py | 253 | 69 | 225,254 |
LOAD_CONST 'Legacy public TF-Keras utilities.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('deserialize_keras_object',)
IMPORT_NAME tf_keras.src.saving.legacy.serialization
IMPORT_FROM deserialize_keras_object
STORE_NAME deserialize_keras_object
POP_TOP
LOAD_CONST 0
LOAD_CONST ('serialize_keras_object',)
IMPORT_NAME ... | """Legacy public TF-Keras utilities."""
from tf_keras.src.saving.legacy.serialization import deserialize_keras_object
from tf_keras.src.saving.legacy.serialization import serialize_keras_object
| data/tf_keras-nightly-2.17.0.dev2024022110/tf_keras/src/utils/legacy/__init__.py | 137 | 69 | 164,861 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME sys
STORE_NAME sys
LOAD_CONST True
STORE_NAME is_py3
LOAD_CONST False
STORE_NAME is_py2
LOAD_NAME sys
LOAD_ATTR version_info
LOAD_ATTR major
LOAD_CONST 3
COMPARE_OP <
POP_JUMP_IF_FALSE
LOAD_NAME AssertionError
LOAD_CONST 'Python 2 has reached end-of-life, and dbus-python no... | import sys
is_py3 = True
is_py2 = False
if sys.version_info.major < 3:
raise AssertionError(
"Python 2 has reached end-of-life, and dbus-python no longer " "supports it."
)
| data/dbus-python-1.3.2/dbus/_compat.py | 94 | 69 | 145,629 |
LOAD_CONST 1
LOAD_CONST ('AzureMachineLearningWorkspaces',)
IMPORT_NAME _azure_machine_learning_workspaces
IMPORT_FROM AzureMachineLearningWorkspaces
STORE_NAME AzureMachineLearningWorkspaces
POP_TOP
LOAD_CONST 1
LOAD_CONST ('VERSION',)
IMPORT_NAME _version
IMPORT_FROM VERSION
STORE_NAME VERSION
POP_TOP
LOAD_NAME VER... | from ._azure_machine_learning_workspaces import AzureMachineLearningWorkspaces
from ._version import VERSION
__version__ = VERSION
__all__ = ["AzureMachineLearningWorkspaces"]
from ._patch import patch_sdk
patch_sdk()
| data/azure-ai-ml-1.13.0/azure/ai/ml/_restclient/v2022_01_01_preview/__init__.py | 147 | 69 | 424,648 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME os
STORE_NAME os
LOAD_CONST <code object should_run_eventlet_tests at 0x7faa5e8158a0, file "f.py", line 4>
LOAD_CONST 'should_run_eventlet_tests'
MAKE_FUNCTION
STORE_NAME should_run_eventlet_tests
LOAD_NAME should_run_eventlet_tests
CALL_FUNCTION
POP_JUMP_IF_FALSE
LOAD_CONST ... | import os
def should_run_eventlet_tests():
return bool(int(os.environ.get("TEST_EVENTLET") or "0"))
if should_run_eventlet_tests():
import eventlet
eventlet.monkey_patch()
| data/oslo.db-14.1.0/oslo_db/tests/__init__.py | 168 | 69 | 152,263 |
LOAD_BUILD_CLASS
LOAD_CONST <code object UnsupportedAPIWarning at 0x7fab8233d780, file "f.py", line 1>
LOAD_CONST 'UnsupportedAPIWarning'
MAKE_FUNCTION
LOAD_CONST 'UnsupportedAPIWarning'
LOAD_NAME Warning
CALL_FUNCTION
STORE_NAME UnsupportedAPIWarning
LOAD_CONST None
RETURN_VALUE
LOAD_NAME __name__
STORE_NAME __module... | class UnsupportedAPIWarning(Warning):
"""
Warn users on Pyro operations that do not have a meaningful interpretation
in NumPyro. Unlike raising NotImplementedError, it might be possible in
such cases to return a dummy object and recover.
"""
pass
| data/numpyro-0.13.2/numpyro/compat/util.py | 162 | 69 | 141,060 |
LOAD_CONST 0
LOAD_CONST ('RPCClient', 'EmptyRPCHandler', 'RPCFunc', 'RPCHandler', 'RPCServer', 'make_rpc_server', 'to_rpc_handler')
IMPORT_NAME fugue.rpc.base
IMPORT_FROM RPCClient
STORE_NAME RPCClient
IMPORT_FROM EmptyRPCHandler
STORE_NAME EmptyRPCHandler
IMPORT_FROM RPCFunc
STORE_NAME RPCFunc
IMPORT_FROM RPCHandler
S... | from fugue.rpc.base import (
RPCClient,
EmptyRPCHandler,
RPCFunc,
RPCHandler,
RPCServer,
make_rpc_server,
to_rpc_handler,
)
| data/fugue-0.8.7/fugue/rpc/__init__.py | 163 | 69 | 8,012 |
LOAD_CONST 0
LOAD_CONST ('absolute_import',)
IMPORT_NAME __future__
IMPORT_FROM absolute_import
STORE_NAME absolute_import
POP_TOP
LOAD_CONST 1
LOAD_CONST ('KafkaProducer',)
IMPORT_NAME kafka
IMPORT_FROM KafkaProducer
STORE_NAME KafkaProducer
POP_TOP
LOAD_CONST 1
LOAD_CONST ('SimpleProducer',)
IMPORT_NAME simple
IMPO... | from __future__ import absolute_import
from .kafka import KafkaProducer
from .simple import SimpleProducer
from .keyed import KeyedProducer
__all__ = ["KafkaProducer", "SimpleProducer", "KeyedProducer"] # deprecated
| data/kafka-1.3.5/kafka/producer/__init__.py | 163 | 69 | 399,474 |
LOAD_CONST 0
LOAD_CONST ('Package', 'package')
IMPORT_NAME localstack.packages
IMPORT_FROM Package
STORE_NAME Package
IMPORT_FROM package
STORE_NAME package
POP_TOP
LOAD_NAME package
LOAD_CONST 'kinesis-mock'
LOAD_CONST ('name',)
CALL_FUNCTION
LOAD_NAME Package
LOAD_CONST ('return',)
BUILD_CONST_KEY_MAP
LOAD_CONST <c... | from localstack.packages import Package, package
@package(name="kinesis-mock")
def kinesismock_package() -> Package:
from localstack.services.kinesis.packages import kinesismock_package
return kinesismock_package
| data/localstack-core-3.1.0/localstack/services/kinesis/plugins.py | 186 | 69 | 225,466 |
LOAD_CONST 1
LOAD_CONST ('AzureMachineLearningWorkspaces',)
IMPORT_NAME _azure_machine_learning_workspaces
IMPORT_FROM AzureMachineLearningWorkspaces
STORE_NAME AzureMachineLearningWorkspaces
POP_TOP
LOAD_CONST 1
LOAD_CONST ('VERSION',)
IMPORT_NAME _version
IMPORT_FROM VERSION
STORE_NAME VERSION
POP_TOP
LOAD_NAME VER... | from ._azure_machine_learning_workspaces import AzureMachineLearningWorkspaces
from ._version import VERSION
__version__ = VERSION
__all__ = ["AzureMachineLearningWorkspaces"]
from ._patch import patch_sdk
patch_sdk()
| data/azure-ai-ml-1.13.0/azure/ai/ml/_restclient/runhistory/__init__.py | 147 | 69 | 424,684 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST <code object test_extracting_tokens at 0x7f8af1683540, file "f.py", line 4>
LOAD_CONST 'test_extracting_tokens'
MAKE_FUNCTION
STORE_NAME test_extracting_tokens
LOAD_CONST None
RETURN_VALUE
LOAD_FAST root
LOAD_ATTR _session_token
LOAD_CONST N... | import pytest
def test_extracting_tokens(root):
if root._session_token is None:
pytest.fail("session token should not be None")
if root._master_token is None:
pytest.fail("master token should not be None")
| data/snowflake_core-0.6.0/tests/integ/test_root.py | 166 | 69 | 400,147 |
LOAD_CONST 0
LOAD_CONST ('unicode_literals',)
IMPORT_NAME __future__
IMPORT_FROM unicode_literals
STORE_NAME unicode_literals
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object DeleteNotAllowed at 0x7f8e2fcbfb70, file "f.py", line 4>
LOAD_CONST 'DeleteNotAllowed'
MAKE_FUNCTION
LOAD_CONST 'DeleteNotAllowed'
LOAD_NAME Exc... | from __future__ import unicode_literals
class DeleteNotAllowed(Exception):
pass
class BadValueError(Exception):
"""raised when a value can't be validated or is required"""
class WrappingAttributeError(AttributeError):
pass
| data/jsonobject-2.1.0/jsonobject/exceptions.py | 352 | 69 | 10,212 |
LOAD_CONST 0
LOAD_CONST ('annotations',)
IMPORT_NAME __future__
IMPORT_FROM annotations
STORE_NAME annotations
POP_TOP
LOAD_CONST 0
LOAD_CONST ('import_optional_dependency',)
IMPORT_NAME pandas.compat._optional
IMPORT_FROM import_optional_dependency
STORE_NAME import_optional_dependency
POP_TOP
LOAD_NAME import_optio... | from __future__ import annotations
from pandas.compat._optional import import_optional_dependency
ne = import_optional_dependency("numexpr", errors="warn")
NUMEXPR_INSTALLED = ne is not None
__all__ = ["NUMEXPR_INSTALLED"]
| data/pandas-2.2.0/pandas/core/computation/check.py | 140 | 70 | 171,358 |
LOAD_CONST 'Amazon Clean Rooms Module.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('read_sql_query',)
IMPORT_NAME awswrangler.cleanrooms._read
IMPORT_FROM read_sql_query
STORE_NAME read_sql_query
POP_TOP
LOAD_CONST 0
LOAD_CONST ('wait_query',)
IMPORT_NAME awswrangler.cleanrooms._utils
IMPORT_FROM wait_query
STORE_NA... | """Amazon Clean Rooms Module."""
from awswrangler.cleanrooms._read import read_sql_query
from awswrangler.cleanrooms._utils import wait_query
__all__ = [
"read_sql_query",
"wait_query",
]
| data/awswrangler-3.6.0/awswrangler/cleanrooms/__init__.py | 125 | 70 | 416,148 |
LOAD_CONST 1
LOAD_CONST ('init', 'deinit', 'reinit', 'colorama_text', 'just_fix_windows_console')
IMPORT_NAME initialise
IMPORT_FROM init
STORE_NAME init
IMPORT_FROM deinit
STORE_NAME deinit
IMPORT_FROM reinit
STORE_NAME reinit
IMPORT_FROM colorama_text
STORE_NAME colorama_text
IMPORT_FROM just_fix_windows_console
STOR... | from .initialise import init, deinit, reinit, colorama_text, just_fix_windows_console
from .ansi import Fore, Back, Style, Cursor
from .ansitowin32 import AnsiToWin32
__version__ = "0.4.6"
| data/pip-24.0/src/pip/_vendor/colorama/__init__.py | 203 | 70 | 324,977 |
LOAD_NAME __name__
LOAD_CONST '__main"'
COMPARE_OP !=
POP_JUMP_IF_TRUE
LOAD_GLOBAL AssertionError
LOAD_CONST 'Indent decreased below 0.'
CALL_FUNCTION
RAISE_VARARGS
SETUP_LOOP to 82
SETUP_EXCEPT to 28
LOAD_CONST 10
STORE_NAME more
POP_BLOCK
JUMP_FORWARD to 50
DUP_TOP
LOAD_NAME EOFError
COMPARE_OP exception match
PO... | assert __name__ != '__main"', "Indent decreased below 0."
while 1:
try:
more = 10
except EOFError:
break
more = len(__file__)
assert not more, "FOO"
assert not more
| data/uncompyle6-3.9.0/test/simple_source/bug_pypy27/01_assert2.py | 157 | 70 | 442,121 |
LOAD_CONST 1
LOAD_CONST ('GitHubParser',)
IMPORT_NAME github_parser
IMPORT_FROM GitHubParser
STORE_NAME GitHubParser
POP_TOP
LOAD_CONST 1
LOAD_CONST ('GitLabParser',)
IMPORT_NAME gitlab_parser
IMPORT_FROM GitLabParser
STORE_NAME GitLabParser
POP_TOP
LOAD_CONST 1
LOAD_CONST ('parse_kindle_text_to_list',)
IMPORT_NAME k... | from .github_parser import GitHubParser
from .gitlab_parser import GitLabParser
from .kindle_parser import parse_kindle_text_to_list
__all__ = ("GitHubParser", "GitLabParser", "parse_kindle_text_to_list")
| data/github_heatmap-1.0.9/github_heatmap/html_parser/__init__.py | 152 | 70 | 416,955 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME unittest
STORE_NAME unittest
LOAD_CONST 0
LOAD_CONST ('install_tests_in_module_dict',)
IMPORT_NAME multiprocess.tests
IMPORT_FROM install_tests_in_module_dict
STORE_NAME install_tests_in_module_dict
POP_TOP
LOAD_NAME install_tests_in_module_dict
LOAD_NAME globals
CALL_FUNCTION... | import unittest
from multiprocess.tests import install_tests_in_module_dict
install_tests_in_module_dict(globals(), "forkserver", only_type="manager")
if __name__ == "__main__":
unittest.main()
| data/multiprocess-0.70.16/py3.13/multiprocess/tests/test_multiprocessing_forkserver/test_manager.py | 145 | 70 | 144,772 |
LOAD_CONST 2
LOAD_CONST ('Provider',)
IMPORT_NAME en_PH
IMPORT_FROM Provider
STORE_NAME EnPhAutomotiveProvider
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object Provider at 0x7faa53e70f60, file "f.py", line 4>
LOAD_CONST 'Provider'
MAKE_FUNCTION
LOAD_CONST 'Provider'
LOAD_NAME EnPhAutomotiveProvider
CALL_FUNCTION
STORE... | from ..en_PH import Provider as EnPhAutomotiveProvider
class Provider(EnPhAutomotiveProvider):
"""Implement automotive provider for ``fil_PH`` locale.
There is no difference from the ``en_PH`` implementation.
"""
pass
| data/Faker-23.2.1/faker/providers/automotive/fil_PH/__init__.py | 164 | 70 | 16,574 |
LOAD_CONST 0
LOAD_CONST ('abstractmethod',)
IMPORT_NAME abc
IMPORT_FROM abstractmethod
STORE_NAME abstractmethod
POP_TOP
LOAD_CONST 0
LOAD_CONST ('Any', 'Mapping')
IMPORT_NAME typing
IMPORT_FROM Any
STORE_NAME Any
IMPORT_FROM Mapping
STORE_NAME Mapping
POP_TOP
LOAD_CONST 0
LOAD_CONST ('BaseOperator',)
IMPORT_NAME bea... | from abc import abstractmethod
from typing import Any, Mapping
from beanie.odm.operators import BaseOperator
class BaseUpdateOperator(BaseOperator):
@property
@abstractmethod
def query(self) -> Mapping[str, Any]:
...
| data/beanie-1.25.0/beanie/odm/operators/update/__init__.py | 276 | 70 | 323,805 |
LOAD_CONST 0
LOAD_CONST ('ClientValue',)
IMPORT_NAME office365.runtime.client_value
IMPORT_FROM ClientValue
STORE_NAME ClientValue
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object CommunicationSiteCreationResponse at 0x7fab700add20, file "f.py", line 4>
LOAD_CONST 'CommunicationSiteCreationResponse'
MAKE_FUNCTION
LOAD... | from office365.runtime.client_value import ClientValue
class CommunicationSiteCreationResponse(ClientValue):
def __init__(self, site_status=None, site_url=None):
self.SiteStatus = site_status
self.SiteUrl = site_url
| data/Office365-REST-Python-Client-2.5.5/office365/sharepoint/publishing/sites/communication/creation_response.py | 240 | 70 | 189,102 |
LOAD_CONST 1
LOAD_CONST ('FileDataSplitter',)
IMPORT_NAME file_data_splitter
IMPORT_FROM FileDataSplitter
STORE_NAME FileDataSplitter
POP_TOP
LOAD_CONST 1
LOAD_CONST ('GridType',)
IMPORT_NAME grid_type
IMPORT_FROM GridType
STORE_NAME GridType
POP_TOP
LOAD_CONST 1
LOAD_CONST ('IndexType',)
IMPORT_NAME index_type
IMPOR... | from .file_data_splitter import FileDataSplitter
from .grid_type import GridType
from .index_type import IndexType
from .join_build_side import JoinBuildSide
__all__ = ["FileDataSplitter", "GridType", "IndexType", "JoinBuildSide"]
| data/apache-sedona-1.5.1/sedona/core/enums/__init__.py | 161 | 70 | 335,166 |
LOAD_CONST 'Contains utilities for coverage analysis.'
STORE_NAME __doc__
LOAD_BUILD_CLASS
LOAD_CONST <code object CoverageResult at 0x7fab64195780, file "f.py", line 4>
LOAD_CONST 'CoverageResult'
MAKE_FUNCTION
LOAD_CONST 'CoverageResult'
CALL_FUNCTION
STORE_NAME CoverageResult
LOAD_CONST None
RETURN_VALUE
LOAD_NAME... | """Contains utilities for coverage analysis."""
class CoverageResult:
def __init__(self, cov_obj):
self.covered = cov_obj.covered()
self.total = cov_obj.total()
self.percentage = cov_obj.percentage()
| data/spark-nlp-5.2.3/sparknlp/common/coverage_result.py | 221 | 70 | 418,127 |
LOAD_CONST 0
LOAD_CONST ('annotations',)
IMPORT_NAME __future__
IMPORT_FROM annotations
STORE_NAME annotations
POP_TOP
LOAD_CONST 0
LOAD_CONST ('event_listener',)
IMPORT_NAME airflow.example_dags.plugins
IMPORT_FROM event_listener
STORE_NAME event_listener
POP_TOP
LOAD_CONST 0
LOAD_CONST ('AirflowPlugin',)
IMPORT_NAM... | from __future__ import annotations
from airflow.example_dags.plugins import event_listener
from airflow.plugins_manager import AirflowPlugin
class MetadataCollectionPlugin(AirflowPlugin):
name = "MetadataCollectionPlugin"
listeners = [event_listener]
| data/apache_airflow-2.8.1/airflow/example_dags/plugins/listener_plugin.py | 221 | 70 | 254,449 |
LOAD_CONST 0
LOAD_CONST ('BaseMessageConverter', 'DefaultMessageConverter', 'SQLChatMessageHistory')
IMPORT_NAME langchain_community.chat_message_histories.sql
IMPORT_FROM BaseMessageConverter
STORE_NAME BaseMessageConverter
IMPORT_FROM DefaultMessageConverter
STORE_NAME DefaultMessageConverter
IMPORT_FROM SQLChatMessa... | from langchain_community.chat_message_histories.sql import (
BaseMessageConverter,
DefaultMessageConverter,
SQLChatMessageHistory,
)
__all__ = [
"BaseMessageConverter",
"DefaultMessageConverter",
"SQLChatMessageHistory",
]
| data/langchain-0.1.8/langchain/memory/chat_message_histories/sql.py | 108 | 70 | 368,799 |
LOAD_CONST 2
LOAD_CONST ('Provider',)
IMPORT_NAME en_PH
IMPORT_FROM Provider
STORE_NAME EnPhAutomotiveProvider
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object Provider at 0x7faa53e70420, file "f.py", line 4>
LOAD_CONST 'Provider'
MAKE_FUNCTION
LOAD_CONST 'Provider'
LOAD_NAME EnPhAutomotiveProvider
CALL_FUNCTION
STORE... | from ..en_PH import Provider as EnPhAutomotiveProvider
class Provider(EnPhAutomotiveProvider):
"""Implement automotive provider for ``tl_PH`` locale.
There is no difference from the ``en_PH`` implementation.
"""
pass
| data/Faker-23.2.1/faker/providers/automotive/tl_PH/__init__.py | 164 | 70 | 16,565 |
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME azure.cli.core.aaz
IMPORT_STAR
LOAD_NAME register_command_group
LOAD_CONST 'image'
CALL_FUNCTION
LOAD_BUILD_CLASS
LOAD_CONST <code object __CMDGroup at 0x7fab420620c0, file "f.py", line 4>
LOAD_CONST '__CMDGroup'
MAKE_FUNCTION
LOAD_CONST '__CMDGroup'
LOAD_NAME AAZCommandGro... | from azure.cli.core.aaz import *
@register_command_group(
"image",
)
class __CMDGroup(AAZCommandGroup):
"""Manage custom virtual machine images."""
pass
__all__ = ["__CMDGroup"]
| data/azure-cli-2.57.0/azure/cli/command_modules/vm/aaz/profile_2020_09_01_hybrid/image/__cmd_group.py | 180 | 70 | 376,700 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME warnings
STORE_NAME warnings
LOAD_CONST 1
LOAD_CONST ('BarsExperimentalWarning',)
IMPORT_NAME std
IMPORT_FROM BarsExperimentalWarning
STORE_NAME BarsExperimentalWarning
POP_TOP
LOAD_NAME warnings
LOAD_METHOD catch_warnings
CALL_METHOD
SETUP_WITH to 64
POP_TOP
LOAD_NAME warnin... | import warnings
from .std import BarsExperimentalWarning
with warnings.catch_warnings():
warnings.simplefilter("ignore", category=BarsExperimentalWarning)
from .autonotebook import tqdm, brange
__all__ = ["tqdm", "brange"]
| data/os_sys-2.1.4/bars/bars/auto.py | 176 | 70 | 248,329 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME unittest
STORE_NAME unittest
LOAD_CONST 0
LOAD_CONST ('install_tests_in_module_dict',)
IMPORT_NAME multiprocess.tests
IMPORT_FROM install_tests_in_module_dict
STORE_NAME install_tests_in_module_dict
POP_TOP
LOAD_NAME install_tests_in_module_dict
LOAD_NAME globals
CALL_FUNCTION... | import unittest
from multiprocess.tests import install_tests_in_module_dict
install_tests_in_module_dict(globals(), "fork", only_type="threads")
if __name__ == "__main__":
unittest.main()
| data/multiprocess-0.70.16/py3.12/multiprocess/tests/test_multiprocessing_fork/test_threads.py | 145 | 70 | 144,917 |
LOAD_CONST 0
LOAD_CONST ('Protocol',)
IMPORT_NAME typing
IMPORT_FROM Protocol
STORE_NAME Protocol
POP_TOP
LOAD_CONST 0
LOAD_CONST ('TableHandle',)
IMPORT_NAME atc.tables
IMPORT_FROM TableHandle
STORE_NAME TableHandle
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object ThMaker at 0x7fab7002bdb0, file "f.py", line 6>
LOAD... | from typing import Protocol
from atc.tables import TableHandle
class ThMaker(Protocol):
"""Use this as a type when you need either DeltaHandle or a SqlServer instance."""
def from_tc(self, id: str) -> TableHandle:
pass
| data/atc-dataplatform-1.1.69/src/atc/tables/ThMaker.py | 251 | 70 | 193,665 |
LOAD_CONST 1
LOAD_CONST ('init', 'deinit', 'reinit', 'colorama_text', 'just_fix_windows_console')
IMPORT_NAME initialise
IMPORT_FROM init
STORE_NAME init
IMPORT_FROM deinit
STORE_NAME deinit
IMPORT_FROM reinit
STORE_NAME reinit
IMPORT_FROM colorama_text
STORE_NAME colorama_text
IMPORT_FROM just_fix_windows_console
STOR... | from .initialise import init, deinit, reinit, colorama_text, just_fix_windows_console
from .ansi import Fore, Back, Style, Cursor
from .ansitowin32 import AnsiToWin32
__version__ = "0.4.6"
| data/clvm_rs-0.6.0/venv/lib/python3.7/site-packages/pip/_vendor/colorama/__init__.py | 203 | 70 | 288,631 |
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME azure.cli.core.aaz
IMPORT_STAR
LOAD_NAME register_command_group
LOAD_CONST 'image'
CALL_FUNCTION
LOAD_BUILD_CLASS
LOAD_CONST <code object __CMDGroup at 0x7fab700555d0, file "f.py", line 4>
LOAD_CONST '__CMDGroup'
MAKE_FUNCTION
LOAD_CONST '__CMDGroup'
LOAD_NAME AAZCommandGro... | from azure.cli.core.aaz import *
@register_command_group(
"image",
)
class __CMDGroup(AAZCommandGroup):
"""Manage custom virtual machine images."""
pass
__all__ = ["__CMDGroup"]
| data/azure-cli-2.57.0/azure/cli/command_modules/vm/aaz/profile_2018_03_01_hybrid/image/__cmd_group.py | 180 | 70 | 376,839 |
LOAD_CONST 'C-style "raw" API for low-level ctypes-specific access to OpenGL\n\nThis sub-package is autogenerated using a customised version\nof the ctypes codegenerator package (see src/openglgenerator.py and\nsrc/generateraw.py).\n'
STORE_NAME __doc__
LOAD_CONST None
RETURN_VALUE | """C-style "raw" API for low-level ctypes-specific access to OpenGL
This sub-package is autogenerated using a customised version
of the ctypes codegenerator package (see src/openglgenerator.py and
src/generateraw.py).
"""
| data/PyOpenGL-3.1.7/OpenGL/raw/__init__.py | 85 | 70 | 59,677 |
LOAD_CONST 0
LOAD_CONST ('gapic_version',)
IMPORT_NAME google.events.cloud.scheduler
IMPORT_FROM gapic_version
STORE_NAME package_version
POP_TOP
LOAD_NAME package_version
LOAD_ATTR __version__
STORE_NAME __version__
LOAD_CONST 0
LOAD_CONST ('SchedulerJobData',)
IMPORT_NAME google.events.cloud.scheduler_v1.types.data... | from google.events.cloud.scheduler import gapic_version as package_version
__version__ = package_version.__version__
from google.events.cloud.scheduler_v1.types.data import SchedulerJobData
__all__ = ("SchedulerJobData",)
| data/google-events-0.11.0/src/google/events/cloud/scheduler/__init__.py | 119 | 70 | 13,327 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME unittest
STORE_NAME unittest
LOAD_CONST 0
LOAD_CONST ('install_tests_in_module_dict',)
IMPORT_NAME multiprocess.tests
IMPORT_FROM install_tests_in_module_dict
STORE_NAME install_tests_in_module_dict
POP_TOP
LOAD_NAME install_tests_in_module_dict
LOAD_NAME globals
CALL_FUNCTION... | import unittest
from multiprocess.tests import install_tests_in_module_dict
install_tests_in_module_dict(globals(), "fork", only_type="processes")
if __name__ == "__main__":
unittest.main()
| data/multiprocess-0.70.16/py3.13/multiprocess/tests/test_multiprocessing_fork/test_processes.py | 145 | 70 | 144,779 |
LOAD_CONST 2
LOAD_CONST ('register_shape_calculator',)
IMPORT_NAME common._registration
IMPORT_FROM register_shape_calculator
STORE_NAME register_shape_calculator
POP_TOP
LOAD_CONST 2
LOAD_CONST ('calculate_linear_classifier_output_shapes',)
IMPORT_NAME common.shape_calculator
IMPORT_FROM calculate_linear_classifier_o... | from ..common._registration import register_shape_calculator
from ..common.shape_calculator import calculate_linear_classifier_output_shapes
register_shape_calculator(
"SklearnOneVsOneClassifier", calculate_linear_classifier_output_shapes
)
| data/skl2onnx-1.16.0/skl2onnx/shape_calculators/one_vs_one_classifier.py | 140 | 70 | 345,244 |
LOAD_CONST 0
LOAD_CONST ('absolute_import', 'division', 'print_function')
IMPORT_NAME __future__
IMPORT_FROM absolute_import
STORE_NAME absolute_import
IMPORT_FROM division
STORE_NAME division
IMPORT_FROM print_function
STORE_NAME print_function
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME re
STORE_NAME re
LOAD_N... | from __future__ import absolute_import, division, print_function
import re
_canonicalize_regex = re.compile(r"[-_.]+")
def canonicalize_name(name):
return _canonicalize_regex.sub("-", name).lower()
| data/clvm_rs-0.6.0/venv/lib/python3.7/site-packages/pkg_resources/_vendor/packaging/utils.py | 178 | 70 | 288,308 |
LOAD_CONST 0
LOAD_CONST ('private', 'public')
IMPORT_NAME public
IMPORT_FROM private
STORE_NAME private
IMPORT_FROM public
STORE_NAME public
POP_TOP
LOAD_NAME public
LOAD_NAME int
LOAD_NAME int
LOAD_CONST ('x', 'return')
BUILD_CONST_KEY_MAP
LOAD_CONST <code object one at 0x7fab420184b0, file "f.py", line 4>
LOAD_CONS... | from public import private, public
@public
def one(x: int) -> int:
return x * 2
one(4)
@private
def two(x: int) -> int:
return x * 3
two(4)
| data/atpublic-4.0/test/test_mypy.py | 213 | 70 | 363,532 |
LOAD_CONST 1
LOAD_CONST ('visualize_alignment',)
IMPORT_NAME vis_alignment
IMPORT_FROM visualize_alignment
STORE_NAME visualize_alignment
POP_TOP
LOAD_CONST 1
LOAD_CONST ('visualize_detection',)
IMPORT_NAME vis_detection
IMPORT_FROM visualize_detection
STORE_NAME visualize_detection
POP_TOP
LOAD_CONST 1
LOAD_CONST ('... | from .vis_alignment import visualize_alignment
from .vis_detection import visualize_detection
from .vis_headpose import visualize_headpose
__all__ = ["visualize_detection", "visualize_alignment", "visualize_headpose"]
| data/facexlib-0.3.0/facexlib/visualization/__init__.py | 156 | 70 | 345,961 |
LOAD_CONST 0
LOAD_CONST ('annotations',)
IMPORT_NAME __future__
IMPORT_FROM annotations
STORE_NAME annotations
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME tokenize
STORE_NAME tokenize
LOAD_CONST 0
LOAD_CONST ('StringIO',)
IMPORT_NAME io
IMPORT_FROM StringIO
STORE_NAME StringIO
POP_TOP
LOAD_CONST 0
LOAD_CONST ('... | from __future__ import annotations
import tokenize
from io import StringIO
from tokenize import TokenInfo
def _tokenize_str(code: str) -> list[TokenInfo]:
return list(tokenize.generate_tokens(StringIO(code).readline))
| data/pylint-3.0.3/pylint/testutils/tokenize_str.py | 211 | 70 | 310,616 |
LOAD_CONST '\nFiles in this directory are installed as executable scripts when the `dxpy` module is installed.\n\nSee the file ../../Readme.md (https://github.com/dnanexus/dx-toolkit/blob/master/src/python/Readme.md) for more info.\n'
STORE_NAME __doc__
LOAD_CONST None
RETURN_VALUE | """
Files in this directory are installed as executable scripts when the `dxpy` module is installed.
See the file ../../Readme.md (https://github.com/dnanexus/dx-toolkit/blob/master/src/python/Readme.md) for more info.
"""
| data/dxpy-0.369.1/dxpy/scripts/__init__.py | 81 | 70 | 441,124 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME aesara.tensor.random.rewriting
STORE_NAME aesara
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME aesara.tensor.random.utils
STORE_NAME aesara
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME aesara.tensor.random.basic
IMPORT_STAR
LOAD_CONST 0
LOAD_CONST ('RandomState', 'default_rng')
IMPO... | import aesara.tensor.random.rewriting
import aesara.tensor.random.utils
from aesara.tensor.random.basic import *
from aesara.tensor.random.op import RandomState, default_rng
from aesara.tensor.random.utils import RandomStream
| data/aesara-2.9.3/aesara/tensor/random/__init__.py | 160 | 70 | 323,656 |
LOAD_CONST 0
LOAD_CONST ('BaseCustomProperty',)
IMPORT_NAME office365.sharepoint.publishing.customproperties.base
IMPORT_FROM BaseCustomProperty
STORE_NAME BaseCustomProperty
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object BigIntCustomProperty at 0x7fab700e40c0, file "f.py", line 4>
LOAD_CONST 'BigIntCustomProperty'
... | from office365.sharepoint.publishing.customproperties.base import BaseCustomProperty
class BigIntCustomProperty(BaseCustomProperty):
@property
def entity_type_name(self):
return "Microsoft.SharePoint.Publishing.RestOnly.BigIntCustomProperty"
| data/Office365-REST-Python-Client-2.5.5/office365/sharepoint/publishing/customproperties/bigint.py | 230 | 70 | 189,138 |
LOAD_CONST 0
LOAD_CONST ('rpc', 'RpcProxy')
IMPORT_NAME nameko.rpc
IMPORT_FROM rpc
STORE_NAME rpc
IMPORT_FROM RpcProxy
STORE_NAME RpcProxy
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object Service at 0x7fab8002f150, file "f.py", line 4>
LOAD_CONST 'Service'
MAKE_FUNCTION
LOAD_CONST 'Service'
CALL_FUNCTION
STORE_NAME Se... | from nameko.rpc import rpc, RpcProxy
class Service:
name = "service"
other_rpc = RpcProxy("another_service")
@rpc # `method` is exposed over RPC
def method(self):
pass
| data/nameko-2.14.1/docs/examples/anatomy.py | 213 | 70 | 43,619 |
LOAD_CONST 0
LOAD_CONST ('Enum',)
IMPORT_NAME enum
IMPORT_FROM Enum
STORE_NAME Enum
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object DefenderCloudBlockLevelType at 0x7fab70027660, file "f.py", line 4>
LOAD_CONST 'DefenderCloudBlockLevelType'
MAKE_FUNCTION
LOAD_CONST 'DefenderCloudBlockLevelType'
LOAD_NAME str
LOAD_NAM... | from enum import Enum
class DefenderCloudBlockLevelType(str, Enum):
NotConfigured = ("notConfigured",)
High = ("high",)
HighPlus = ("highPlus",)
ZeroTolerance = ("zeroTolerance",)
| data/msgraph-sdk-1.1.0/msgraph/generated/models/defender_cloud_block_level_type.py | 194 | 70 | 260,221 |
LOAD_BUILD_CLASS
LOAD_CONST <code object SourceFile at 0x7faa760206f0, file "f.py", line 1>
LOAD_CONST 'SourceFile'
MAKE_FUNCTION
LOAD_CONST 'SourceFile'
CALL_FUNCTION
STORE_NAME SourceFile
LOAD_CONST None
RETURN_VALUE
LOAD_NAME __name__
STORE_NAME __module__
LOAD_CONST 'SourceFile'
STORE_NAME __qualname__
LOAD_CONST... | class SourceFile:
"""The base class for any source file that provides translation contexts."""
def __init__(self, filename, **kwargs):
"""Initialise the object."""
super().__init__(**kwargs)
self.filename = filename
self.contexts = []
| data/PyQt6-6.6.1/lupdate/source_file.py | 223 | 70 | 209,901 |
LOAD_CONST ('WandbMetricsLogger', 'WandbModelCheckpoint', 'WandbEvalCallback')
STORE_NAME __all__
LOAD_CONST 1
LOAD_CONST ('WandbMetricsLogger',)
IMPORT_NAME metrics_logger
IMPORT_FROM WandbMetricsLogger
STORE_NAME WandbMetricsLogger
POP_TOP
LOAD_CONST 1
LOAD_CONST ('WandbModelCheckpoint',)
IMPORT_NAME model_checkpoi... | __all__ = ("WandbMetricsLogger", "WandbModelCheckpoint", "WandbEvalCallback")
from .metrics_logger import WandbMetricsLogger
from .model_checkpoint import WandbModelCheckpoint
from .tables_builder import WandbEvalCallback
| data/wandb-0.16.3/wandb/integration/keras/callbacks/__init__.py | 150 | 70 | 360,594 |
LOAD_CONST 1
LOAD_CONST ('LinearRegression',)
IMPORT_NAME linear_regression
IMPORT_FROM LinearRegression
STORE_NAME LinearRegression
POP_TOP
LOAD_CONST 1
LOAD_CONST ('StackingCVRegressor',)
IMPORT_NAME stacking_cv_regression
IMPORT_FROM StackingCVRegressor
STORE_NAME StackingCVRegressor
POP_TOP
LOAD_CONST 1
LOAD_CONS... | from .linear_regression import LinearRegression
from .stacking_cv_regression import StackingCVRegressor
from .stacking_regression import StackingRegressor
__all__ = ["LinearRegression", "StackingRegressor", "StackingCVRegressor"]
| data/mlxtend-0.23.1/mlxtend/regressor/__init__.py | 152 | 70 | 196,367 |
LOAD_CONST 1
LOAD_CONST ('PromptEncoderConfig', 'PromptEncoderReparameterizationType')
IMPORT_NAME config
IMPORT_FROM PromptEncoderConfig
STORE_NAME PromptEncoderConfig
IMPORT_FROM PromptEncoderReparameterizationType
STORE_NAME PromptEncoderReparameterizationType
POP_TOP
LOAD_CONST 1
LOAD_CONST ('PromptEncoder',)
IMPO... | from .config import PromptEncoderConfig, PromptEncoderReparameterizationType
from .model import PromptEncoder
__all__ = [
"PromptEncoder",
"PromptEncoderConfig",
"PromptEncoderReparameterizationType",
]
| data/peft-0.8.2/src/peft/tuners/p_tuning/__init__.py | 141 | 70 | 11,725 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME unittest
STORE_NAME unittest
LOAD_CONST 0
LOAD_CONST ('install_tests_in_module_dict',)
IMPORT_NAME multiprocess.tests
IMPORT_FROM install_tests_in_module_dict
STORE_NAME install_tests_in_module_dict
POP_TOP
LOAD_NAME install_tests_in_module_dict
LOAD_NAME globals
CALL_FUNCTION... | import unittest
from multiprocess.tests import install_tests_in_module_dict
install_tests_in_module_dict(globals(), "spawn", only_type="threads")
if __name__ == "__main__":
unittest.main()
| data/multiprocess-0.70.16/py3.13/multiprocess/tests/test_multiprocessing_spawn/test_threads.py | 145 | 70 | 144,775 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME cherrypy
STORE_NAME cherrypy
LOAD_BUILD_CLASS
LOAD_CONST <code object Root at 0x7fab81f2f9c0, file "f.py", line 4>
LOAD_CONST 'Root'
MAKE_FUNCTION
LOAD_CONST 'Root'
LOAD_NAME object
CALL_FUNCTION
STORE_NAME Root
LOAD_NAME cherrypy
LOAD_ATTR config
LOAD_METHOD update
LOAD_CONST... | import cherrypy
class Root(object):
@cherrypy.expose
def index(self):
return "Hello World!"
cherrypy.config.update({"environment": "embedded"})
app = cherrypy.tree.mount(Root())
| data/gunicorn-21.2.0/examples/frameworks/cherryapp.py | 219 | 70 | 343,512 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME unittest
STORE_NAME unittest
LOAD_CONST 0
LOAD_CONST ('install_tests_in_module_dict',)
IMPORT_NAME multiprocess.tests
IMPORT_FROM install_tests_in_module_dict
STORE_NAME install_tests_in_module_dict
POP_TOP
LOAD_NAME install_tests_in_module_dict
LOAD_NAME globals
CALL_FUNCTION... | import unittest
from multiprocess.tests import install_tests_in_module_dict
install_tests_in_module_dict(globals(), "spawn", only_type="processes")
if __name__ == "__main__":
unittest.main()
| data/multiprocess-0.70.16/py3.13/multiprocess/tests/test_multiprocessing_spawn/test_processes.py | 145 | 70 | 144,774 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pyomo.contrib.example
STORE_NAME pyomo
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pyomo.common.unittest
IMPORT_FROM common
ROT_TWO
POP_TOP
IMPORT_FROM unittest
STORE_NAME unittest
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object Tests at 0x7fab4290d300, file "f.py", line 5>
LOAD... | import pyomo.contrib.example
import pyomo.common.unittest as unittest
class Tests(unittest.TestCase):
def test1(self):
pass
if __name__ == "__main__":
unittest.main()
| data/Pyomo-6.7.0/pyomo/contrib/example/tests/test_example.py | 239 | 70 | 23,676 |
LOAD_CONST 0
LOAD_CONST ('AnyDocArray',)
IMPORT_NAME docarray.array.any_array
IMPORT_FROM AnyDocArray
STORE_NAME AnyDocArray
POP_TOP
LOAD_CONST 0
LOAD_CONST ('DocList',)
IMPORT_NAME docarray.array.doc_list.doc_list
IMPORT_FROM DocList
STORE_NAME DocList
POP_TOP
LOAD_CONST 0
LOAD_CONST ('DocVec',)
IMPORT_NAME docarray... | from docarray.array.any_array import AnyDocArray
from docarray.array.doc_list.doc_list import DocList
from docarray.array.doc_vec.doc_vec import DocVec
__all__ = ["DocList", "DocVec", "AnyDocArray"]
| data/docarray-0.40.0/docarray/array/__init__.py | 142 | 70 | 250,030 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME numpy
STORE_NAME np
LOAD_CONST 0
LOAD_CONST ('OpRunBinary',)
IMPORT_NAME onnx.reference.ops._op
IMPORT_FROM OpRunBinary
STORE_NAME OpRunBinary
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object BitwiseXor at 0x7fab540c5db0, file "f.py", line 6>
LOAD_CONST 'BitwiseXor'
MAKE_FUNCT... | import numpy as np
from onnx.reference.ops._op import OpRunBinary
class BitwiseXor(OpRunBinary):
def _run(self, x, y): # type: ignore
return (np.bitwise_xor(x, y),)
| data/onnx-1.15.0/onnx/reference/ops/op_bitwise_xor.py | 226 | 70 | 111,235 |
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME azure.cli.core.aaz
IMPORT_STAR
LOAD_NAME register_command_group
LOAD_CONST 'image'
CALL_FUNCTION
LOAD_BUILD_CLASS
LOAD_CONST <code object __CMDGroup at 0x7fab82132390, file "f.py", line 4>
LOAD_CONST '__CMDGroup'
MAKE_FUNCTION
LOAD_CONST '__CMDGroup'
LOAD_NAME AAZCommandGro... | from azure.cli.core.aaz import *
@register_command_group(
"image",
)
class __CMDGroup(AAZCommandGroup):
"""Manage custom virtual machine images."""
pass
__all__ = ["__CMDGroup"]
| data/azure-cli-2.57.0/azure/cli/command_modules/vm/aaz/latest/image/__cmd_group.py | 180 | 70 | 376,515 |
LOAD_CONST 0
LOAD_CONST ('distutils',)
IMPORT_NAME setuptools
IMPORT_FROM distutils
STORE_NAME distutils
POP_TOP
LOAD_CONST 0
LOAD_CONST ('cc',)
IMPORT_NAME source_module
IMPORT_FROM cc
STORE_NAME cc
POP_TOP
LOAD_NAME distutils
LOAD_ATTR core
LOAD_ATTR setup
STORE_NAME setup
LOAD_CONST <code object run_setup at 0x7f... | from setuptools import distutils
from source_module import cc
setup = distutils.core.setup
def run_setup():
setup(ext_modules=[cc.distutils_extension()])
if __name__ == "__main__":
run_setup()
| data/numba-0.59.0/numba/tests/pycc_distutils_usecase/setup_distutils.py | 178 | 70 | 134,249 |
LOAD_CONST 0
LOAD_CONST ('ValidationError',)
IMPORT_NAME pyhanko_certvalidator.errors
IMPORT_FROM ValidationError
STORE_NAME ValidationError
POP_TOP
LOAD_CONST 'PastValidatePrecheckFailure'
LOAD_CONST 'TimeSlideFailure'
BUILD_LIST
STORE_NAME __all__
LOAD_BUILD_CLASS
LOAD_CONST <code object PastValidatePrecheckFailure... | from pyhanko_certvalidator.errors import ValidationError
__all__ = ["PastValidatePrecheckFailure", "TimeSlideFailure"]
class PastValidatePrecheckFailure(ValidationError):
pass
class TimeSlideFailure(ValidationError):
pass
| data/pyhanko-certvalidator-0.26.3/pyhanko_certvalidator/ltv/errors.py | 279 | 70 | 51,454 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.