input
stringlengths
28
198k
output
stringlengths
3
71k
file
stringlengths
19
330
input_tokens
int64
5
159k
output_tokens
int64
3
9.07k
__index_level_0__
int64
2
449k
LOAD_CONST "Dataset definition for siscore.\n\nDEPRECATED!\nIf you want to use the Siscore dataset builder class, use:\ntfds.builder_cls('siscore')\n" STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('lazy_builder_import',) IMPORT_NAME tensorflow_datasets.core IMPORT_FROM lazy_builder_import STORE_NAME lazy_builder_import ...
"""Dataset definition for siscore. DEPRECATED! If you want to use the Siscore dataset builder class, use: tfds.builder_cls('siscore') """ from tensorflow_datasets.core import lazy_builder_import Siscore = lazy_builder_import.LazyBuilderImport("siscore")
data/tfds-nightly-4.9.4.dev202402210044/tensorflow_datasets/image_classification/siscore/siscore.py
130
86
238,536
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME ctypes STORE_NAME ctypes LOAD_CONST 0 LOAD_CONST None IMPORT_NAME pytest STORE_NAME pytest LOAD_CONST 0 LOAD_CONST ('lib',) IMPORT_NAME awkward_cpp.cpu_kernels IMPORT_FROM lib STORE_NAME lib POP_TOP LOAD_NAME pytest LOAD_ATTR mark LOAD_ATTR skip LOAD_CONST 'Unable to generate...
import ctypes import pytest from awkward_cpp.cpu_kernels import lib @pytest.mark.skip(reason="Unable to generate any tests for kernel") def test_cpuawkward_reduce_argmax_uint32_64_1(): raise NotImplementedError("Unable to generate any tests for kernel")
data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmax_uint32_64.py
226
86
221,413
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME unittest STORE_NAME unittest LOAD_CONST 2 LOAD_CONST ('Five9',) IMPORT_NAME five9 IMPORT_FROM Five9 STORE_NAME Five9 POP_TOP LOAD_BUILD_CLASS LOAD_CONST <code object Common at 0x7fab8212bd20, file "f.py", line 6> LOAD_CONST 'Common' MAKE_FUNCTION LOAD_CONST 'Common' LOAD_NAME ...
import unittest from ..five9 import Five9 class Common(unittest.TestCase): def setUp(self): super(Common, self).setUp() self.user = "username@something.com" self.password = "password" self.five9 = Five9(self.user, self.password)
data/five9-0.0.3b192/five9/tests/common.py
280
86
175,960
LOAD_CONST "Dataset definition for so2sat.\n\nDEPRECATED!\nIf you want to use the So2sat dataset builder class, use:\ntfds.builder_cls('so2sat')\n" STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('lazy_builder_import',) IMPORT_NAME tensorflow_datasets.core IMPORT_FROM lazy_builder_import STORE_NAME lazy_builder_import POP...
"""Dataset definition for so2sat. DEPRECATED! If you want to use the So2sat dataset builder class, use: tfds.builder_cls('so2sat') """ from tensorflow_datasets.core import lazy_builder_import So2sat = lazy_builder_import.LazyBuilderImport("so2sat")
data/tensorflow-datasets-4.9.4/tensorflow_datasets/image_classification/so2sat.py
130
86
255,372
LOAD_CONST '\nErrors around Submission Locking\n' STORE_NAME __doc__ LOAD_BUILD_CLASS LOAD_CONST <code object SubmissionLockContestedError at 0x7f8e2ff254b0, file "f.py", line 6> LOAD_CONST 'SubmissionLockContestedError' MAKE_FUNCTION LOAD_CONST 'SubmissionLockContestedError' LOAD_NAME Exception CALL_FUNCTION STORE_NA...
""" Errors around Submission Locking """ class SubmissionLockContestedError(Exception): """ Error indicating trying to modify a lock that the user does not have access to modify. """ error_code = "ERR_LOCK_CONTESTED" def get_error_code(self): return self.error_code
data/ora2-6.1.0/openassessment/staffgrader/errors/submission_lock.py
249
86
116,308
LOAD_BUILD_CLASS LOAD_CONST <code object Positional at 0x7faa7c009420, file "f.py", line 1> LOAD_CONST 'Positional' MAKE_FUNCTION LOAD_CONST 'Positional' LOAD_NAME object CALL_FUNCTION STORE_NAME Positional LOAD_CONST None RETURN_VALUE LOAD_NAME __name__ STORE_NAME __module__ LOAD_CONST 'Positional' STORE_NAME __qualn...
class Positional(object): def __init__(self, help="", type=None, choices=None, metavar=None): self.help = help self.kind = "positional" self.abbrev = None self.type = type self.choices = choices self.metavar = metavar
data/plac-1.4.2/doc/annotations.py
226
86
112,196
LOAD_CONST 1 LOAD_CONST ('DeepSpeedTransformerLayer', 'DeepSpeedTransformerConfig') IMPORT_NAME transformer IMPORT_FROM DeepSpeedTransformerLayer STORE_NAME DeepSpeedTransformerLayer IMPORT_FROM DeepSpeedTransformerConfig STORE_NAME DeepSpeedTransformerConfig POP_TOP LOAD_CONST 1 LOAD_CONST ('DeepSpeedInferenceConfig'...
from .transformer import DeepSpeedTransformerLayer, DeepSpeedTransformerConfig from .inference.config import DeepSpeedInferenceConfig from ...model_implementations.transformers.ds_transformer import ( DeepSpeedTransformerInference, ) from .inference.moe_inference import DeepSpeedMoEInferenceConfig, DeepSpeedMoEInfe...
data/deepspeed-0.13.2/deepspeed/ops/transformer/__init__.py
218
86
84,706
LOAD_CONST 0 LOAD_CONST ('reactor',) IMPORT_NAME twisted.internet IMPORT_FROM reactor STORE_NAME reactor POP_TOP LOAD_CONST 0 LOAD_CONST ('WSGIResource',) IMPORT_NAME twisted.web.wsgi IMPORT_FROM WSGIResource STORE_NAME WSGIResource POP_TOP LOAD_CONST 2 LOAD_CONST ('exposition', 'REGISTRY') IMPORT_NAME IMPORT_FROM ex...
from twisted.internet import reactor from twisted.web.wsgi import WSGIResource from .. import exposition, REGISTRY MetricsResource = lambda registry=REGISTRY: WSGIResource( reactor, reactor.getThreadPool(), exposition.make_wsgi_app(registry) )
data/prometheus_client-0.20.0/prometheus_client/twisted/_exposition.py
217
86
154,254
LOAD_CONST '\nVersion 3 classes\n=================\nThis class is a iterator for version 3 API call\n\n.. autoclass:: APIResultIterator\n :members:\n' STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('APIIterator',) IMPORT_NAME restfly.iterator IMPORT_FROM APIIterator STORE_NAME APIIterator POP_TOP LOAD_BUILD_CLASS LOAD_C...
""" Version 3 classes ================= This class is a iterator for version 3 API call .. autoclass:: APIResultIterator :members: """ from restfly.iterator import APIIterator class APIResultIterator(APIIterator): """ Iterator class for version 3 API """ pass
data/pyTenable-1.4.20/tenable/io/v3/base/iterators/iterator.py
206
86
175,191
LOAD_BUILD_CLASS LOAD_CONST <code object Cloud at 0x7fab42018660, file "f.py", line 1> LOAD_CONST 'Cloud' MAKE_FUNCTION LOAD_CONST 'Cloud' CALL_FUNCTION STORE_NAME Cloud LOAD_CONST None RETURN_VALUE LOAD_NAME __name__ STORE_NAME __module__ LOAD_CONST 'Cloud' STORE_NAME __qualname__ LOAD_CONST <code object create_pars...
class Cloud: def create_parser(self, argparser): return NotImplemented def set_cloud_config_from_arguments(self, arguments): return NotImplemented def set_composition_arguments(self, comp_group): pass def get_composition(self, **kwargs): pass
data/qds_sdk-1.16.1/qds_sdk/cloud/cloud.py
347
86
358,218
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME ctypes STORE_NAME ctypes LOAD_CONST 0 LOAD_CONST None IMPORT_NAME pytest STORE_NAME pytest LOAD_CONST 0 LOAD_CONST ('lib',) IMPORT_NAME awkward_cpp.cpu_kernels IMPORT_FROM lib STORE_NAME lib POP_TOP LOAD_NAME pytest LOAD_ATTR mark LOAD_ATTR skip LOAD_CONST 'Unable to generate...
import ctypes import pytest from awkward_cpp.cpu_kernels import lib @pytest.mark.skip(reason="Unable to generate any tests for kernel") def test_cpuawkward_reduce_argmax_uint16_64_1(): raise NotImplementedError("Unable to generate any tests for kernel")
data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmax_uint16_64.py
225
86
221,418
LOAD_CONST '\nversion info\n' STORE_NAME __doc__ LOAD_CONST '1.4.7' STORE_NAME VERSION LOAD_CONST 'Steve McGrath <steve@mcgrath.sh>' STORE_NAME AUTHOR LOAD_CONST 'REST API library framework' STORE_NAME DESCRIPTION LOAD_NAME tuple LOAD_CONST <code object <genexpr> at 0x7f8e2ff5e420, file "f.py", line 8> LOAD_CONST '...
""" version info """ VERSION = "1.4.7" AUTHOR = "Steve McGrath <steve@mcgrath.sh>" DESCRIPTION = "REST API library framework" version_info = tuple(int(d) for d in VERSION.split("-", maxsplit=1)[0].split("."))
data/restfly-1.4.7/restfly/version.py
208
86
405,022
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME not_existing STORE_NAME not_existing LOAD_CONST 0 LOAD_CONST None IMPORT_NAME os STORE_NAME os LOAD_CONST 0 LOAD_CONST ('abspath',) IMPORT_NAME os.path IMPORT_FROM abspath STORE_NAME abspath POP_TOP LOAD_CONST 0 LOAD_CONST ('not_existing',) IMPORT_NAME os.path IMPORT_FROM not...
import not_existing import os from os.path import abspath from os.path import not_existing from datetime import date date.today date.not_existing_attribute from datetime.date import today import datetime.datetime import not_existing_nested.date import os.path
data/jedi-0.19.1/test/static_analysis/imports.py
231
86
421,043
LOAD_CONST 0 LOAD_CONST ('IS_PYDANTIC_V2',) IMPORT_NAME beanie.odm.utils.pydantic IMPORT_FROM IS_PYDANTIC_V2 STORE_NAME IS_PYDANTIC_V2 POP_TOP LOAD_NAME IS_PYDANTIC_V2 POP_JUMP_IF_FALSE LOAD_CONST 0 LOAD_CONST ('DecimalAnnotation',) IMPORT_NAME beanie.odm.custom_types.decimal IMPORT_FROM DecimalAnnotation STORE_NAME ...
from beanie.odm.utils.pydantic import IS_PYDANTIC_V2 if IS_PYDANTIC_V2: from beanie.odm.custom_types.decimal import DecimalAnnotation else: from decimal import Decimal as DecimalAnnotation __all__ = [ "DecimalAnnotation", ]
data/beanie-1.25.0/beanie/odm/custom_types/__init__.py
160
86
323,825
LOAD_CONST 1 LOAD_CONST ('post_processor_catalogue', 'register_post_processor', 'remove_post_processor') IMPORT_NAME catalogue IMPORT_FROM post_processor_catalogue STORE_NAME post_processor_catalogue IMPORT_FROM register_post_processor STORE_NAME register_post_processor IMPORT_FROM remove_post_processor STORE_NAME remo...
from .catalogue import ( post_processor_catalogue, register_post_processor, remove_post_processor, ) from .base import PostProcessor from .filth_replacer import FilthReplacer from .prefix_suffix import PrefixSuffixReplacer from .remover import FilthRemover
data/scrubadub-2.0.1/scrubadub/post_processors/__init__.py
227
86
236,920
LOAD_CONST 0 LOAD_CONST ('TestCase', 'min_os_level') IMPORT_NAME PyObjCTools.TestSupport IMPORT_FROM TestCase STORE_NAME TestCase IMPORT_FROM min_os_level STORE_NAME min_os_level POP_TOP LOAD_CONST 0 LOAD_CONST None IMPORT_NAME objc STORE_NAME objc LOAD_CONST 0 LOAD_CONST None IMPORT_NAME MapKit STORE_NAME MapKit LO...
from PyObjCTools.TestSupport import TestCase, min_os_level import objc import MapKit class TestMKPointAnnotation(TestCase): @min_os_level("10.9") def testClasses(self): self.assertIsInstance(MapKit.MKPointAnnotation, objc.objc_class)
data/pyobjc-framework-MapKit-10.1/PyObjCTest/test_mkpointannotation.py
298
86
405,960
LOAD_CONST 0 LOAD_CONST ('BaseHook', 'DatahubGenericHook', 'DatahubKafkaHook', 'DatahubRestHook') IMPORT_NAME datahub_airflow_plugin.hooks.datahub IMPORT_FROM BaseHook STORE_NAME BaseHook IMPORT_FROM DatahubGenericHook STORE_NAME DatahubGenericHook IMPORT_FROM DatahubKafkaHook STORE_NAME DatahubKafkaHook IMPORT_FROM Da...
from datahub_airflow_plugin.hooks.datahub import ( BaseHook, DatahubGenericHook, DatahubKafkaHook, DatahubRestHook, ) __all__ = ["DatahubRestHook", "DatahubKafkaHook", "DatahubGenericHook", "BaseHook"]
data/acryl-datahub-0.12.1.5/src/datahub_provider/hooks/datahub.py
154
86
435,790
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME ctypes STORE_NAME ctypes LOAD_CONST 0 LOAD_CONST None IMPORT_NAME pytest STORE_NAME pytest LOAD_CONST 0 LOAD_CONST ('lib',) IMPORT_NAME awkward_cpp.cpu_kernels IMPORT_FROM lib STORE_NAME lib POP_TOP LOAD_NAME pytest LOAD_ATTR mark LOAD_ATTR skip LOAD_CONST 'Unable to generate...
import ctypes import pytest from awkward_cpp.cpu_kernels import lib @pytest.mark.skip(reason="Unable to generate any tests for kernel") def test_cpuawkward_reduce_argmax_int64_64_1(): raise NotImplementedError("Unable to generate any tests for kernel")
data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmax_int64_64.py
225
86
221,498
LOAD_CONST 0 LOAD_CONST ('Enum',) IMPORT_NAME enum IMPORT_FROM Enum STORE_NAME Enum POP_TOP LOAD_BUILD_CLASS LOAD_CONST <code object WifiBand at 0x7fab4298ec00, file "f.py", line 4> LOAD_CONST 'WifiBand' MAKE_FUNCTION LOAD_CONST 'WifiBand' LOAD_NAME str LOAD_NAME Enum CALL_FUNCTION STORE_NAME WifiBand LOAD_CONST None ...
from enum import Enum class WifiBand(str, Enum): Unknown = ("unknown",) Frequency24GHz = ("frequency24GHz",) Frequency50GHz = ("frequency50GHz",) Frequency60GHz = ("frequency60GHz",) UnknownFutureValue = ("unknownFutureValue",)
data/msgraph-sdk-1.1.0/msgraph/generated/models/call_records/wifi_band.py
203
86
262,402
LOAD_CONST 0 LOAD_CONST ('FileCache', 'SeparateBodyFileCache') IMPORT_NAME pigar._vendor.pip._vendor.cachecontrol.caches.file_cache IMPORT_FROM FileCache STORE_NAME FileCache IMPORT_FROM SeparateBodyFileCache STORE_NAME SeparateBodyFileCache POP_TOP LOAD_CONST 0 LOAD_CONST ('RedisCache',) IMPORT_NAME pigar._vendor.pip...
from pigar._vendor.pip._vendor.cachecontrol.caches.file_cache import ( FileCache, SeparateBodyFileCache, ) from pigar._vendor.pip._vendor.cachecontrol.caches.redis_cache import RedisCache __all__ = ["FileCache", "SeparateBodyFileCache", "RedisCache"]
data/pigar-2.1.3/pigar/_vendor/pip/_vendor/cachecontrol/caches/__init__.py
144
86
241,309
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME ctypes STORE_NAME ctypes LOAD_CONST 0 LOAD_CONST None IMPORT_NAME pytest STORE_NAME pytest LOAD_CONST 0 LOAD_CONST ('lib',) IMPORT_NAME awkward_cpp.cpu_kernels IMPORT_FROM lib STORE_NAME lib POP_TOP LOAD_NAME pytest LOAD_ATTR mark LOAD_ATTR skip LOAD_CONST 'Unable to generate...
import ctypes import pytest from awkward_cpp.cpu_kernels import lib @pytest.mark.skip(reason="Unable to generate any tests for kernel") def test_cpuawkward_reduce_argmin_uint32_64_1(): raise NotImplementedError("Unable to generate any tests for kernel")
data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmin_uint32_64.py
225
86
221,391
LOAD_CONST 'DO NOT EDIT.\n\nThis file was autogenerated. Do not edit it by hand,\nsince your modifications would be overwritten.\n' STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('ResNet50',) IMPORT_NAME keras.src.applications.resnet IMPORT_FROM ResNet50 STORE_NAME ResNet50 POP_TOP LOAD_CONST 0 LOAD_CONST ('decode_predi...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.applications.resnet import ResNet50 from keras.src.applications.resnet import decode_predictions from keras.src.applications.resnet import preprocess_input
data/keras-nightly-3.1.0.dev2024022103/keras/applications/resnet50/__init__.py
166
86
284,938
LOAD_CONST 0 LOAD_CONST ('SubProcessCompressor',) IMPORT_NAME pipeline.compressors IMPORT_FROM SubProcessCompressor STORE_NAME SubProcessCompressor POP_TOP LOAD_CONST 0 LOAD_CONST ('settings',) IMPORT_NAME pipeline.conf IMPORT_FROM settings STORE_NAME settings POP_TOP LOAD_BUILD_CLASS LOAD_CONST <code object CSSMinCo...
from pipeline.compressors import SubProcessCompressor from pipeline.conf import settings class CSSMinCompressor(SubProcessCompressor): def compress_css(self, css): command = (settings.CSSMIN_BINARY, settings.CSSMIN_ARGUMENTS) return self.execute_command(command, css)
data/django-pipeline-3.0.0/pipeline/compressors/cssmin.py
278
86
76,419
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME ctypes STORE_NAME ctypes LOAD_CONST 0 LOAD_CONST None IMPORT_NAME pytest STORE_NAME pytest LOAD_CONST 0 LOAD_CONST ('lib',) IMPORT_NAME awkward_cpp.cpu_kernels IMPORT_FROM lib STORE_NAME lib POP_TOP LOAD_NAME pytest LOAD_ATTR mark LOAD_ATTR skip LOAD_CONST 'Unable to generate...
import ctypes import pytest from awkward_cpp.cpu_kernels import lib @pytest.mark.skip(reason="Unable to generate any tests for kernel") def test_cpuawkward_reduce_argmax_uint64_64_1(): raise NotImplementedError("Unable to generate any tests for kernel")
data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmax_uint64_64.py
225
86
221,399
LOAD_CONST 0 LOAD_CONST ('TestCase', 'min_os_level') IMPORT_NAME PyObjCTools.TestSupport IMPORT_FROM TestCase STORE_NAME TestCase IMPORT_FROM min_os_level STORE_NAME min_os_level POP_TOP LOAD_CONST 0 LOAD_CONST None IMPORT_NAME Quartz STORE_NAME Quartz LOAD_BUILD_CLASS LOAD_CONST <code object TestIKFilterBrowserView ...
from PyObjCTools.TestSupport import TestCase, min_os_level import Quartz class TestIKFilterBrowserView(TestCase): @min_os_level("10.5") def testMethods(self): self.assertArgIsBOOL(Quartz.IKFilterBrowserView.setPreviewState_, 0)
data/pyobjc-framework-Quartz-10.1/PyObjCTest/test_IKFilterBrowserView.py
291
86
133,565
LOAD_CONST 0 LOAD_CONST ('*',) IMPORT_NAME azure.cli.core.aaz IMPORT_STAR LOAD_NAME register_command_group LOAD_CONST 'eventhubs namespace authorization-rule keys' CALL_FUNCTION LOAD_BUILD_CLASS LOAD_CONST <code object __CMDGroup at 0x7fab415056f0, file "f.py", line 4> LOAD_CONST '__CMDGroup' MAKE_FUNCTION LOAD_CONST...
from azure.cli.core.aaz import * @register_command_group( "eventhubs namespace authorization-rule keys", ) class __CMDGroup(AAZCommandGroup): """Manage Azure EventHubs Authorizationrule connection strings for Namespace.""" pass __all__ = ["__CMDGroup"]
data/azure-cli-2.57.0/azure/cli/command_modules/eventhubs/aaz/profile_2019_03_01_hybrid/eventhubs/namespace/authorization_rule/keys/__cmd_group.py
196
86
374,726
LOAD_CONST 0 LOAD_CONST ('ClientValue',) IMPORT_NAME office365.runtime.client_value IMPORT_FROM ClientValue STORE_NAME ClientValue POP_TOP LOAD_BUILD_CLASS LOAD_CONST <code object SocialLink at 0x7fab64179e40, file "f.py", line 4> LOAD_CONST 'SocialLink' MAKE_FUNCTION LOAD_CONST 'SocialLink' LOAD_NAME ClientValue CALL...
from office365.runtime.client_value import ClientValue class SocialLink(ClientValue): """The SocialLink class defines a link that includes a URI and text representation. This class is used to represent the location of a web site.""" @property def entity_type_name(self): return "SP.Social.Soci...
data/Office365-REST-Python-Client-2.5.5/office365/sharepoint/social/link.py
241
86
189,644
LOAD_CONST "Dataset definition for dsprites.\n\nDEPRECATED!\nIf you want to use the Dsprites dataset builder class, use:\ntfds.builder_cls('dsprites')\n" STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('lazy_builder_import',) IMPORT_NAME tensorflow_datasets.core IMPORT_FROM lazy_builder_import STORE_NAME lazy_builder_impo...
"""Dataset definition for dsprites. DEPRECATED! If you want to use the Dsprites dataset builder class, use: tfds.builder_cls('dsprites') """ from tensorflow_datasets.core import lazy_builder_import Dsprites = lazy_builder_import.LazyBuilderImport("dsprites")
data/tensorflow-datasets-4.9.4/tensorflow_datasets/image/dsprites.py
130
86
255,260
LOAD_CONST 'The Reader Component of a Source.' STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('FileReader',) IMPORT_NAME butterfree.extract.readers.file_reader IMPORT_FROM FileReader STORE_NAME FileReader POP_TOP LOAD_CONST 0 LOAD_CONST ('KafkaReader',) IMPORT_NAME butterfree.extract.readers.kafka_reader IMPORT_FROM Kaf...
"""The Reader Component of a Source.""" from butterfree.extract.readers.file_reader import FileReader from butterfree.extract.readers.kafka_reader import KafkaReader from butterfree.extract.readers.table_reader import TableReader __all__ = ["FileReader", "KafkaReader", "TableReader"]
data/butterfree-1.2.3/butterfree/extract/readers/__init__.py
164
86
9,059
LOAD_CONST './libra-faster-rcnn_r50_fpn_1x_coco.py' STORE_NAME _base_ LOAD_NAME dict LOAD_NAME dict LOAD_CONST 101 LOAD_NAME dict LOAD_CONST 'Pretrained' LOAD_CONST 'torchvision://resnet101' LOAD_CONST ('type', 'checkpoint') CALL_FUNCTION LOAD_CONST ('depth', 'init_cfg') CALL_FUNCTION LOAD_CONST ('backbone',) CALL_...
_base_ = "./libra-faster-rcnn_r50_fpn_1x_coco.py" model = dict( backbone=dict( depth=101, init_cfg=dict(type="Pretrained", checkpoint="torchvision://resnet101"), ) )
data/mmdet-3.3.0/mmdet/.mim/configs/libra_rcnn/libra-faster-rcnn_r101_fpn_1x_coco.py
111
86
151,457
LOAD_CONST 0 LOAD_CONST ('setup',) IMPORT_NAME setuptools IMPORT_FROM setup STORE_NAME setup POP_TOP LOAD_NAME RuntimeError LOAD_CONST 'This mimics the build error on unmatched platform' CALL_FUNCTION RAISE_VARARGS LOAD_NAME setup LOAD_CONST 'demo' LOAD_CONST '0.0.1' LOAD_CONST 'test demo' LOAD_CONST 'demo' BUILD...
from setuptools import setup if True: raise RuntimeError("This mimics the build error on unmatched platform") setup( name="demo", version="0.0.1", description="test demo", py_modules=["demo"], python_requires=">=3.3", )
data/pdm-2.12.3/tests/fixtures/projects/demo-failure-no-dep/setup.py
123
86
354,242
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME gto.log STORE_NAME gto LOAD_CONST 0 LOAD_CONST ('api',) IMPORT_NAME gto IMPORT_FROM api STORE_NAME api POP_TOP LOAD_CONST 0 LOAD_CONST ('__version__',) IMPORT_NAME gto._version IMPORT_FROM __version__ STORE_NAME __version__ POP_TOP LOAD_CONST 0 LOAD_CONST ('CONFIG',) IMPORT_N...
import gto.log # noqa from gto import api from gto._version import __version__ from gto.config import CONFIG from gto.index import RepoIndexManager from gto.registry import GitRegistry __all__ = ["api", "CONFIG", "RepoIndexManager", "GitRegistry", "__version__"]
data/gto-1.7.0/gto/__init__.py
207
86
354,192
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME typing STORE_NAME t LOAD_CONST 0 LOAD_CONST ('expressions',) IMPORT_NAME sqlglot IMPORT_FROM expressions STORE_NAME exp POP_TOP LOAD_NAME t LOAD_ATTR Dict LOAD_NAME exp LOAD_ATTR Identifier LOAD_NAME exp LOAD_ATTR Identifier BUILD_TUPLE BINARY_SUBSCR LOAD_CONST ('replacement_m...
import typing as t from sqlglot import expressions as exp def replace_id_value(node, replacement_mapping: t.Dict[exp.Identifier, exp.Identifier]): if isinstance(node, exp.Identifier) and node in replacement_mapping: node = node.replace(replacement_mapping[node].copy()) return node
data/acryl-sqlglot-21.1.2.dev10/sqlglot/dataframe/sql/transforms.py
198
86
12,748
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME sys STORE_NAME sys LOAD_CONST 0 LOAD_CONST None IMPORT_NAME os STORE_NAME os LOAD_CONST 0 LOAD_CONST None IMPORT_NAME PIL.Image STORE_NAME PIL LOAD_CONST '/' LOAD_NAME PIL LOAD_ATTR Image STORE_ATTR __file__ LOAD_NAME sys LOAD_ATTR _MEIPASS STORE_NAME basedir LOAD_NAME PIL ...
import sys import os import PIL.Image PIL.Image.__file__ = "/" basedir = sys._MEIPASS im = PIL.Image.open(os.path.join(basedir, "tinysample.tiff")) im.save(os.path.join(basedir, "tinysample.png"))
data/pyinstaller-6.4.0/tests/functional/scripts/pyi_lib_PIL_img_conversion.py
159
86
184,854
LOAD_CONST 'A module for compatibility with the docutils>=0.17 `include` directive, in RST documents:\n\nFor example::\n\n .. include:: path/to/file.md\n :parser: myst_parser.docutils_\n' STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('Parser',) IMPORT_NAME myst_parser.parsers.docutils_ IMPORT_FROM Parser STORE_NAME Pars...
"""A module for compatibility with the docutils>=0.17 `include` directive, in RST documents: For example:: .. include:: path/to/file.md :parser: myst_parser.docutils_ """ from myst_parser.parsers.docutils_ import Parser # noqa: F401
data/myst_parser-2.0.0/myst_parser/docutils_.py
105
86
192,736
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME pyperf STORE_NAME pyperf LOAD_NAME pyperf LOAD_METHOD Runner CALL_METHOD STORE_NAME runner LOAD_NAME runner LOAD_ATTR timeit LOAD_CONST 'sorted(list(range(1000)), key=lambda x: x)' LOAD_CONST 'sorted(s, key=f)' LOAD_CONST 'f = lambda x: x; s = list(range(1000))' LOAD_CONST ...
import pyperf runner = pyperf.Runner() runner.timeit( "sorted(list(range(1000)), key=lambda x: x)", stmt="sorted(s, key=f)", setup="f = lambda x: x; s = list(range(1000))", )
data/pyperf-2.6.2/doc/examples/bench_timeit.py
117
86
123,054
LOAD_CONST "Dataset definition for bigearthnet.\n\nDEPRECATED!\nIf you want to use the Bigearthnet dataset builder class, use:\ntfds.builder_cls('bigearthnet')\n" STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('lazy_builder_import',) IMPORT_NAME tensorflow_datasets.core IMPORT_FROM lazy_builder_import STORE_NAME lazy_bui...
"""Dataset definition for bigearthnet. DEPRECATED! If you want to use the Bigearthnet dataset builder class, use: tfds.builder_cls('bigearthnet') """ from tensorflow_datasets.core import lazy_builder_import Bigearthnet = lazy_builder_import.LazyBuilderImport("bigearthnet")
data/tensorflow-datasets-4.9.4/tensorflow_datasets/image_classification/bigearthnet.py
130
86
255,386
LOAD_CONST 'train' STORE_NAME TRAIN_LOGGER_NAME LOAD_CONST 'hyperparameter-optimization' STORE_NAME HYPEROPT_LOGGER_NAME LOAD_CONST 'model.pt' STORE_NAME MODEL_FILE_NAME LOAD_CONST 'test_scores.csv' STORE_NAME TEST_SCORES_FILE_NAME LOAD_CONST 'hyperopt_seeds.txt' STORE_NAME HYPEROPT_SEED_FILE_NAME LOAD_CONST None R...
TRAIN_LOGGER_NAME = "train" HYPEROPT_LOGGER_NAME = "hyperparameter-optimization" MODEL_FILE_NAME = "model.pt" TEST_SCORES_FILE_NAME = "test_scores.csv" HYPEROPT_SEED_FILE_NAME = "hyperopt_seeds.txt"
data/chemprop-1.6.1/chemprop/constants.py
106
86
367,595
LOAD_CONST 0 LOAD_CONST ('setup', 'Extension') IMPORT_NAME setuptools IMPORT_FROM setup STORE_NAME setup IMPORT_FROM Extension STORE_NAME Extension POP_TOP LOAD_NAME Extension LOAD_CONST 'lru._lru' LOAD_CONST 'src/lru/_lru.c' BUILD_LIST CALL_FUNCTION BUILD_LIST STORE_NAME extensions LOAD_CONST True LOAD_CONST '' LOA...
from setuptools import setup, Extension extensions = [ Extension("lru._lru", ["src/lru/_lru.c"]), ] args = { "include_package_data": True, "exclude_package_data": {"": ["*.c"]}, } setup(ext_modules=extensions, **args)
data/lru-dict-1.3.0/setup.py
137
86
416,002
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME unittest STORE_NAME unittest LOAD_CONST 0 LOAD_CONST ('StringIO',) IMPORT_NAME gmplot.utility IMPORT_FROM StringIO STORE_NAME StringIO POP_TOP LOAD_BUILD_CLASS LOAD_CONST <code object StringIOTest at 0x7fab702a6660, file "f.py", line 5> LOAD_CONST 'StringIOTest' MAKE_FUNCTION ...
import unittest from gmplot.utility import StringIO class StringIOTest(unittest.TestCase): def test_enter_exit(self): with StringIO() as f: f.write("Content") self.assertEqual(f.getvalue(), "Content") self.assertTrue(f.closed)
data/gmplot-1.4.1/tests/test_utility.py
306
86
352,870
LOAD_CONST 1 LOAD_CONST ('ray_triangle',) IMPORT_NAME IMPORT_FROM ray_triangle STORE_NAME ray_triangle POP_TOP SETUP_EXCEPT to 34 LOAD_CONST 1 LOAD_CONST ('ray_pyembree',) IMPORT_NAME IMPORT_FROM ray_pyembree STORE_NAME ray_pyembree POP_TOP LOAD_CONST True STORE_NAME has_embree POP_BLOCK JUMP_FORWARD to 94 DUP_TOP ...
from . import ray_triangle try: from . import ray_pyembree has_embree = True except BaseException as E: from .. import exceptions ray_pyembree = exceptions.ExceptionWrapper(E) has_embree = False __all__ = ["ray_triangle", "ray_pyembree"]
data/trimesh-4.1.4/trimesh/ray/__init__.py
217
86
197,148
LOAD_CONST 1 LOAD_CONST ('Missing', 'ResultBackend') IMPORT_NAME backend IMPORT_FROM Missing STORE_NAME Missing IMPORT_FROM ResultBackend STORE_NAME ResultBackend POP_TOP LOAD_CONST 1 LOAD_CONST ('ResultError', 'ResultFailure', 'ResultMissing', 'ResultTimeout') IMPORT_NAME errors IMPORT_FROM ResultError STORE_NAME Res...
from .backend import Missing, ResultBackend from .errors import ResultError, ResultFailure, ResultMissing, ResultTimeout from .middleware import Results __all__ = [ "Missing", "ResultBackend", "ResultError", "ResultFailure", "ResultTimeout", "ResultMissing", "Results", ]
data/dramatiq-1.16.0/dramatiq/results/__init__.py
180
86
145,483
LOAD_CONST '\nTest cases for testing if it is possible to pickle\nObjective-C objects\n' STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('TestCase',) IMPORT_NAME PyObjCTools.TestSupport IMPORT_FROM TestCase STORE_NAME TestCase POP_TOP LOAD_BUILD_CLASS LOAD_CONST <code object TestPickleObjC at 0x7fab4106cae0, file "f.py",...
""" Test cases for testing if it is possible to pickle Objective-C objects """ from PyObjCTools.TestSupport import TestCase class TestPickleObjC(TestCase): pass class TestPickleMixedGraph(TestCase): pass class TestPicklePythonNSObject(TestCase): pass
data/pyobjc-core-10.1/PyObjCTest/test_pickling_objc.py
403
86
95,141
LOAD_CONST "Dataset definition for bccd.\n\nDEPRECATED!\nIf you want to use the BCCD dataset builder class, use:\ntfds.builder_cls('bccd')\n" STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('lazy_builder_import',) IMPORT_NAME tensorflow_datasets.core IMPORT_FROM lazy_builder_import STORE_NAME lazy_builder_import POP_TOP ...
"""Dataset definition for bccd. DEPRECATED! If you want to use the BCCD dataset builder class, use: tfds.builder_cls('bccd') """ from tensorflow_datasets.core import lazy_builder_import BCCD = lazy_builder_import.LazyBuilderImport("bccd")
data/tensorflow-datasets-4.9.4/tensorflow_datasets/image/bccd/bccd.py
130
86
255,280
LOAD_CONST 0 LOAD_CONST ('BaseModel',) IMPORT_NAME langchain_core.pydantic_v1 IMPORT_FROM BaseModel STORE_NAME BaseModel POP_TOP LOAD_BUILD_CLASS LOAD_CONST <code object AttributeInfo at 0x7faa755caf60, file "f.py", line 4> LOAD_CONST 'AttributeInfo' MAKE_FUNCTION LOAD_CONST 'AttributeInfo' LOAD_NAME BaseModel CALL_FU...
from langchain_core.pydantic_v1 import BaseModel class AttributeInfo(BaseModel): """Information about a data source attribute.""" name: str description: str type: str class Config: """Configuration for this pydantic object.""" arbitrary_types_allowed = True frozen = True...
data/langchain-0.1.8/langchain/chains/query_constructor/schema.py
314
86
369,010
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME abc STORE_NAME abc LOAD_CONST 0 LOAD_CONST ('MetricsContext',) IMPORT_NAME aws_embedded_metrics.logger.metrics_context IMPORT_FROM MetricsContext STORE_NAME MetricsContext POP_TOP LOAD_CONST 0 LOAD_CONST ('List',) IMPORT_NAME typing IMPORT_FROM List STORE_NAME List POP_TOP LO...
import abc from aws_embedded_metrics.logger.metrics_context import MetricsContext from typing import List class Serializer(abc.ABC): @staticmethod @abc.abstractmethod def serialize(context: MetricsContext) -> List[str]: """Flushes the metrics context to the sink."""
data/aws-embedded-metrics-3.2.0/aws_embedded_metrics/serializers/__init__.py
266
86
367,098
LOAD_CONST 0 LOAD_CONST ('*',) IMPORT_NAME azure.cli.core.aaz IMPORT_STAR LOAD_NAME register_command_group LOAD_CONST 'relay wcfrelay authorization-rule' CALL_FUNCTION LOAD_BUILD_CLASS LOAD_CONST <code object __CMDGroup at 0x7fab4130c8a0, file "f.py", line 4> LOAD_CONST '__CMDGroup' MAKE_FUNCTION LOAD_CONST '__CMDGro...
from azure.cli.core.aaz import * @register_command_group( "relay wcfrelay authorization-rule", ) class __CMDGroup(AAZCommandGroup): """Manage Azure Relay Service WCF Relay Authorization Rule.""" pass __all__ = ["__CMDGroup"]
data/azure-cli-2.57.0/azure/cli/command_modules/relay/aaz/latest/relay/wcfrelay/authorization_rule/__cmd_group.py
196
86
375,643
LOAD_CONST 0 LOAD_CONST ('TestCase', 'min_os_level') IMPORT_NAME PyObjCTools.TestSupport IMPORT_FROM TestCase STORE_NAME TestCase IMPORT_FROM min_os_level STORE_NAME min_os_level POP_TOP LOAD_CONST 0 LOAD_CONST None IMPORT_NAME ContactsUI STORE_NAME ContactsUI LOAD_BUILD_CLASS LOAD_CONST <code object TestCNContactPic...
from PyObjCTools.TestSupport import TestCase, min_os_level import ContactsUI # noqa: F401 class TestCNContactPickerDelegate(TestCase): @min_os_level("10.11") def testProtocols(self): self.assertProtocolExists("CNContactPickerDelegate")
data/pyobjc-framework-ContactsUI-10.1/PyObjCTest/test_cncontactpickerdelegate.py
286
86
354,949
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME ctypes STORE_NAME ctypes LOAD_CONST 0 LOAD_CONST None IMPORT_NAME pytest STORE_NAME pytest LOAD_CONST 0 LOAD_CONST ('lib',) IMPORT_NAME awkward_cpp.cpu_kernels IMPORT_FROM lib STORE_NAME lib POP_TOP LOAD_NAME pytest LOAD_ATTR mark LOAD_ATTR skip LOAD_CONST 'Unable to generate...
import ctypes import pytest from awkward_cpp.cpu_kernels import lib @pytest.mark.skip(reason="Unable to generate any tests for kernel") def test_cpuawkward_reduce_countnonzero_int8_64_1(): raise NotImplementedError("Unable to generate any tests for kernel")
data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_countnonzero_int8_64.py
226
86
221,331
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME re STORE_NAME re LOAD_NAME re LOAD_METHOD compile LOAD_CONST '^arn:aws(?:-[a-z]+)?:dynamodb:(?P<region>[a-z\\d\\-]+-\\d):(?P<accountid>[\\d]{12}):table/(?P<id>[\\S]+)$' CALL_METHOD STORE_NAME TABLE_ARN_RE LOAD_CONST None RETURN_VALUE
import re TABLE_ARN_RE = re.compile( r"^arn:aws(?:-[a-z]+)?:dynamodb:(?P<region>[a-z\d\-]+-\d):(?P<accountid>[\d]{12}):table/(?P<id>[\S]+)$" )
data/compose_x_common-1.4.6/src/compose_x_common/aws/dynamodb.py
106
86
345,344
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME ctypes STORE_NAME ctypes LOAD_CONST 0 LOAD_CONST None IMPORT_NAME pytest STORE_NAME pytest LOAD_CONST 0 LOAD_CONST ('lib',) IMPORT_NAME awkward_cpp.cpu_kernels IMPORT_FROM lib STORE_NAME lib POP_TOP LOAD_NAME pytest LOAD_ATTR mark LOAD_ATTR skip LOAD_CONST 'Unable to generate...
import ctypes import pytest from awkward_cpp.cpu_kernels import lib @pytest.mark.skip(reason="Unable to generate any tests for kernel") def test_cpuawkward_reduce_argmin_uint64_64_1(): raise NotImplementedError("Unable to generate any tests for kernel")
data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmin_uint64_64.py
226
86
221,278
LOAD_CONST 0 LOAD_CONST ('Enum',) IMPORT_NAME enum IMPORT_FROM Enum STORE_NAME Enum POP_TOP LOAD_BUILD_CLASS LOAD_CONST <code object AlertStatus at 0x7fab70027660, file "f.py", line 4> LOAD_CONST 'AlertStatus' MAKE_FUNCTION LOAD_CONST 'AlertStatus' LOAD_NAME str LOAD_NAME Enum CALL_FUNCTION STORE_NAME AlertStatus LOAD...
from enum import Enum class AlertStatus(str, Enum): Unknown = ("unknown",) NewAlert = ("newAlert",) InProgress = ("inProgress",) Resolved = ("resolved",) Dismissed = ("dismissed",) UnknownFutureValue = ("unknownFutureValue",)
data/msgraph-sdk-1.1.0/msgraph/generated/models/alert_status.py
202
86
261,913
LOAD_CONST 0 LOAD_CONST ('DownloadFactory',) IMPORT_NAME pyomo.common.download IMPORT_FROM DownloadFactory STORE_NAME DownloadFactory POP_TOP LOAD_CONST 0 LOAD_CONST ('ExtensionBuilderFactory',) IMPORT_NAME pyomo.common.extensions IMPORT_FROM ExtensionBuilderFactory STORE_NAME ExtensionBuilderFactory POP_TOP LOAD_CON...
from pyomo.common.download import DownloadFactory from pyomo.common.extensions import ExtensionBuilderFactory from .getMCPP import get_mcpp from .build import MCPPBuilder def load(): DownloadFactory.register("mcpp")(get_mcpp) ExtensionBuilderFactory.register("mcpp")(MCPPBuilder)
data/Pyomo-6.7.0/pyomo/contrib/mcpp/plugins.py
234
86
23,652
LOAD_CONST "Dataset definition for mctaco.\n\nDEPRECATED!\nIf you want to use the Mctaco dataset builder class, use:\ntfds.builder_cls('mctaco')\n" STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('lazy_builder_import',) IMPORT_NAME tensorflow_datasets.core IMPORT_FROM lazy_builder_import STORE_NAME lazy_builder_import POP...
"""Dataset definition for mctaco. DEPRECATED! If you want to use the Mctaco dataset builder class, use: tfds.builder_cls('mctaco') """ from tensorflow_datasets.core import lazy_builder_import Mctaco = lazy_builder_import.LazyBuilderImport("mctaco")
data/tensorflow-datasets-4.9.4/tensorflow_datasets/question_answering/mctaco.py
130
86
255,294
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME ctypes STORE_NAME ctypes LOAD_CONST 0 LOAD_CONST None IMPORT_NAME pytest STORE_NAME pytest LOAD_CONST 0 LOAD_CONST ('lib',) IMPORT_NAME awkward_cpp.cpu_kernels IMPORT_FROM lib STORE_NAME lib POP_TOP LOAD_NAME pytest LOAD_ATTR mark LOAD_ATTR skip LOAD_CONST 'Unable to generate...
import ctypes import pytest from awkward_cpp.cpu_kernels import lib @pytest.mark.skip(reason="Unable to generate any tests for kernel") def test_cpuawkward_IndexedArrayU32_numnull_parents_1(): raise NotImplementedError("Unable to generate any tests for kernel")
data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_IndexedArrayU32_numnull_parents.py
225
86
221,467
LOAD_CONST "Dataset definition for siscore.\n\nDEPRECATED!\nIf you want to use the Siscore dataset builder class, use:\ntfds.builder_cls('siscore')\n" STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('lazy_builder_import',) IMPORT_NAME tensorflow_datasets.core IMPORT_FROM lazy_builder_import STORE_NAME lazy_builder_import ...
"""Dataset definition for siscore. DEPRECATED! If you want to use the Siscore dataset builder class, use: tfds.builder_cls('siscore') """ from tensorflow_datasets.core import lazy_builder_import Siscore = lazy_builder_import.LazyBuilderImport("siscore")
data/tensorflow-datasets-4.9.4/tensorflow_datasets/image_classification/siscore/siscore.py
130
86
255,475
LOAD_CONST 0 LOAD_CONST ('annotations',) IMPORT_NAME __future__ IMPORT_FROM annotations STORE_NAME annotations POP_TOP LOAD_CONST 0 LOAD_CONST ('NotebookImportExtractor',) IMPORT_NAME deptry.imports.extractors.notebook_import_extractor IMPORT_FROM NotebookImportExtractor STORE_NAME NotebookImportExtractor POP_TOP LOA...
from __future__ import annotations from deptry.imports.extractors.notebook_import_extractor import NotebookImportExtractor from deptry.imports.extractors.python_import_extractor import PythonImportExtractor __all__ = ( "NotebookImportExtractor", "PythonImportExtractor", )
data/deptry-0.12.0/deptry/imports/extractors/__init__.py
150
86
45,206
LOAD_CONST 0 LOAD_CONST ('CallbackReader',) IMPORT_NAME pysmi.reader.callback IMPORT_FROM CallbackReader STORE_NAME CallbackReader POP_TOP LOAD_CONST 0 LOAD_CONST ('FtpReader',) IMPORT_NAME pysmi.reader.ftpclient IMPORT_FROM FtpReader STORE_NAME FtpReader POP_TOP LOAD_CONST 0 LOAD_CONST ('HttpReader',) IMPORT_NAME py...
from pysmi.reader.callback import CallbackReader from pysmi.reader.ftpclient import FtpReader from pysmi.reader.httpclient import HttpReader from pysmi.reader.zipreader import ZipReader from pysmi.reader.localfile import FileReader from pysmi.reader.url import getReadersFromUrls
data/pysmi-0.3.4/pysmi/reader/__init__.py
216
86
155,038
LOAD_CONST 0 LOAD_CONST ('App', 'ComposeResult') IMPORT_NAME textual.app IMPORT_FROM App STORE_NAME App IMPORT_FROM ComposeResult STORE_NAME ComposeResult POP_TOP LOAD_CONST 0 LOAD_CONST ('LoadingIndicator',) IMPORT_NAME textual.widgets IMPORT_FROM LoadingIndicator STORE_NAME LoadingIndicator POP_TOP LOAD_BUILD_CLASS...
from textual.app import App, ComposeResult from textual.widgets import LoadingIndicator class LoadingApp(App): def compose(self) -> ComposeResult: yield LoadingIndicator() if __name__ == "__main__": app = LoadingApp() app.run()
data/textual-0.52.1/docs/examples/widgets/loading_indicator.py
289
86
26,301
LOAD_CONST 0 LOAD_CONST ('ClientValue',) IMPORT_NAME office365.runtime.client_value IMPORT_FROM ClientValue STORE_NAME ClientValue POP_TOP LOAD_BUILD_CLASS LOAD_CONST <code object Certification at 0x7fab700ad930, file "f.py", line 4> LOAD_CONST 'Certification' MAKE_FUNCTION LOAD_CONST 'Certification' LOAD_NAME ClientV...
from office365.runtime.client_value import ClientValue class Certification(ClientValue): """Represents the certification details of an application.""" def __init__(self, certification_details_url=None): """ :param str certification_details_url: """ self.certificationDetailsUrl...
data/Office365-REST-Python-Client-2.5.5/office365/directory/certificates/certification.py
220
86
188,871
LOAD_CONST 0 LOAD_CONST ('exceptions',) IMPORT_NAME keystonemiddleware IMPORT_FROM exceptions STORE_NAME exceptions POP_TOP LOAD_NAME exceptions LOAD_ATTR ConfigurationError STORE_NAME ConfigurationError LOAD_BUILD_CLASS LOAD_CONST <code object InvalidToken at 0x7f8e2fe3cb70, file "f.py", line 7> LOAD_CONST 'InvalidT...
from keystonemiddleware import exceptions ConfigurationError = exceptions.ConfigurationError class InvalidToken(exceptions.KeystoneMiddlewareException): pass class ServiceError(exceptions.KeystoneMiddlewareException): pass class RevocationListError(exceptions.KeystoneMiddlewareException): pass
data/keystonemiddleware-10.5.0/keystonemiddleware/auth_token/_exceptions.py
346
86
397,969
LOAD_CONST './panoptic-fpn_r50_fpn_ms-3x_coco.py' STORE_NAME _base_ LOAD_NAME dict LOAD_NAME dict LOAD_CONST 101 LOAD_NAME dict LOAD_CONST 'Pretrained' LOAD_CONST 'torchvision://resnet101' LOAD_CONST ('type', 'checkpoint') CALL_FUNCTION LOAD_CONST ('depth', 'init_cfg') CALL_FUNCTION LOAD_CONST ('backbone',) CALL_FU...
_base_ = "./panoptic-fpn_r50_fpn_ms-3x_coco.py" model = dict( backbone=dict( depth=101, init_cfg=dict(type="Pretrained", checkpoint="torchvision://resnet101"), ) )
data/mmdet-3.3.0/mmdet/.mim/configs/panoptic_fpn/panoptic-fpn_r101_fpn_ms-3x_coco.py
111
86
151,212
LOAD_CONST '\nrequests_kerberos.exceptions\n~~~~~~~~~~~~~~~~~~~\n\nThis module contains the set of exceptions.\n\n' STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('RequestException',) IMPORT_NAME requests.exceptions IMPORT_FROM RequestException STORE_NAME RequestException POP_TOP LOAD_BUILD_CLASS LOAD_CONST <code object...
""" requests_kerberos.exceptions ~~~~~~~~~~~~~~~~~~~ This module contains the set of exceptions. """ from requests.exceptions import RequestException class MutualAuthenticationError(RequestException): """Mutual Authentication Error""" class KerberosExchangeError(RequestException): """Kerberos Exchange Fai...
data/pywinrm-0.4.3/winrm/vendor/requests_kerberos/exceptions.py
315
86
269,488
LOAD_CONST 0 LOAD_CONST ('TestCase', 'main') IMPORT_NAME unittest IMPORT_FROM TestCase STORE_NAME TestCase IMPORT_FROM main STORE_NAME main POP_TOP LOAD_CONST 0 LOAD_CONST ('ApiNinjas',) IMPORT_NAME random_word IMPORT_FROM ApiNinjas STORE_NAME ApiNinjas POP_TOP LOAD_BUILD_CLASS LOAD_CONST <code object TestRandomWord ...
from unittest import TestCase, main from random_word import ApiNinjas class TestRandomWord(TestCase): r = ApiNinjas() def test_random_word(self): return self.r.get_random_word() if __name__ == "__main__": main()
data/Random Word-1.0.11/random_word/tests/test_apininjas.py
286
86
134,980
LOAD_CONST "A lil' TOML parser." STORE_NAME __doc__ LOAD_CONST ('loads', 'load', 'TOMLDecodeError') STORE_NAME __all__ LOAD_CONST '1.2.0' STORE_NAME __version__ LOAD_CONST 1 LOAD_CONST ('TOMLDecodeError', 'load', 'loads') IMPORT_NAME _parser IMPORT_FROM TOMLDecodeError STORE_NAME TOMLDecodeError IMPORT_FROM load STO...
"""A lil' TOML parser.""" __all__ = ("loads", "load", "TOMLDecodeError") __version__ = "1.2.0" # DO NOT EDIT THIS LINE MANUALLY. LET bump2version UTILITY DO IT from ._parser import TOMLDecodeError, load, loads
data/isort-5.13.2/isort/_vendored/tomli/__init__.py
116
86
105,664
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME os STORE_NAME os LOAD_CONST 0 LOAD_CONST ('Dict', 'List', 'Tuple', 'TypeVar', 'Union') IMPORT_NAME typing IMPORT_FROM Dict STORE_NAME Dict IMPORT_FROM List STORE_NAME List IMPORT_FROM Tuple STORE_NAME Tuple IMPORT_FROM TypeVar STORE_NAME TypeVar IMPORT_FROM Union STORE_NAME Uni...
import os from typing import Dict, List, Tuple, TypeVar, Union T = TypeVar("T") ListLike = Union[List[T], Tuple[T, ...]] NestedDataStructureLike = Union[T, List[T], Dict[str, T]] PathLike = Union[str, bytes, os.PathLike]
data/datasets-2.17.1/src/datasets/utils/typing.py
202
86
52,869
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME pytest STORE_NAME pytest LOAD_CONST 0 LOAD_CONST ('Check',) IMPORT_NAME frictionless IMPORT_FROM Check STORE_NAME Check POP_TOP LOAD_CONST <code object test_check_from_descriptor_type_framework_v4 at 0x7fab429cdae0, file "f.py", line 6> LOAD_CONST 'test_check_from_descriptor_t...
import pytest from frictionless import Check def test_check_from_descriptor_type_framework_v4(): with pytest.warns(UserWarning): check = Check.from_descriptor({"code": "table-dimensions"}) assert check.to_descriptor() == {"type": "table-dimensions"}
data/frictionless-5.16.1/tests/checklist/check/test_convert.py
236
86
41,479
LOAD_CONST "Dataset definition for titanic.\n\nDEPRECATED!\nIf you want to use the Titanic dataset builder class, use:\ntfds.builder_cls('titanic')\n" STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('lazy_builder_import',) IMPORT_NAME tensorflow_datasets.core IMPORT_FROM lazy_builder_import STORE_NAME lazy_builder_import ...
"""Dataset definition for titanic. DEPRECATED! If you want to use the Titanic dataset builder class, use: tfds.builder_cls('titanic') """ from tensorflow_datasets.core import lazy_builder_import Titanic = lazy_builder_import.LazyBuilderImport("titanic")
data/tensorflow-datasets-4.9.4/tensorflow_datasets/structured/titanic.py
130
86
255,197
LOAD_CONST 0 LOAD_CONST ('Any', 'MutableMapping') IMPORT_NAME typing IMPORT_FROM Any STORE_NAME Any IMPORT_FROM MutableMapping STORE_NAME MutableMapping POP_TOP SETUP_EXCEPT to 34 LOAD_CONST 0 LOAD_CONST ('Context',) IMPORT_NAME airflow.utils.context IMPORT_FROM Context STORE_NAME Context POP_TOP POP_BLOCK JUMP_FORWA...
from typing import Any, MutableMapping try: from airflow.utils.context import Context except ModuleNotFoundError: class Context(MutableMapping[str, Any]): # type: ignore[no-redef] """Placeholder typing class for ``airflow.utils.context.Context``.""" __all__ = ["Context"]
data/astronomer-providers-1.19.0/astronomer/providers/utils/typing_compat.py
244
86
252,524
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME hubspot.cms.url_redirects IMPORT_FROM cms ROT_TWO POP_TOP IMPORT_FROM url_redirects STORE_NAME api_client POP_TOP LOAD_CONST 3 LOAD_CONST ('DiscoveryBase',) IMPORT_NAME discovery_base IMPORT_FROM DiscoveryBase STORE_NAME DiscoveryBase POP_TOP LOAD_BUILD_CLASS LOAD_CONST <code ...
import hubspot.cms.url_redirects as api_client from ...discovery_base import DiscoveryBase class Discovery(DiscoveryBase): @property def redirects_api(self) -> api_client.RedirectsApi: return self._configure_api_client(api_client, "RedirectsApi")
data/hubspot-api-client-8.2.1/hubspot/discovery/cms/url_redirects/discovery.py
276
86
320,806
LOAD_CONST 1 LOAD_CONST ('check_kwargs',) IMPORT_NAME check_kwargs IMPORT_FROM check_kwargs STORE_NAME check_kwargs POP_TOP LOAD_CONST 1 LOAD_CONST ('__version__',) IMPORT_NAME version IMPORT_FROM __version__ STORE_NAME __version__ POP_TOP LOAD_CONST 1 LOAD_CONST ('get_user_agent',) IMPORT_NAME user_agent IMPORT_FROM...
from .check_kwargs import check_kwargs from .version import __version__ from .user_agent import get_user_agent from .deprecation_notice import warn_deprecated from .fix_tuple_length import fix_tuple_length from .convert_to_list import convert_to_list from .normalize_host import normalize_host
data/pinecone_client-3.0.3/pinecone/utils/__init__.py
245
86
251,330
SETUP_ANNOTATIONS LOAD_CONST 'Constants for the mütesync integration.' STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('timedelta',) IMPORT_NAME datetime IMPORT_FROM timedelta STORE_NAME timedelta POP_TOP LOAD_CONST 0 LOAD_CONST ('Final',) IMPORT_NAME typing IMPORT_FROM Final STORE_NAME Final POP_TOP LOAD_CONST 'mutesyn...
"""Constants for the mütesync integration.""" from datetime import timedelta from typing import Final DOMAIN: Final = "mutesync" UPDATE_INTERVAL_NOT_IN_MEETING: Final = timedelta(seconds=10) UPDATE_INTERVAL_IN_MEETING: Final = timedelta(seconds=10)
data/homeassistant-2024.2.2/homeassistant/components/mutesync/const.py
219
86
295,976
LOAD_CONST 0 LOAD_CONST ('profile',) IMPORT_NAME memory_profiler IMPORT_FROM profile STORE_NAME profile POP_TOP LOAD_NAME profile LOAD_CONST <code object test_with_profile at 0x7faa74ae3ed0, file "f.py", line 4> LOAD_CONST 'test_with_profile' MAKE_FUNCTION CALL_FUNCTION STORE_NAME test_with_profile LOAD_NAME __name__...
from memory_profiler import profile @profile def test_with_profile(arg1): """dummy doc""" return None if __name__ == "__main__": assert test_with_profile.__doc__ == "dummy doc" assert test_with_profile.__name__ == "test_with_profile"
data/memory_profiler-0.61.0/test/test_attributes.py
178
86
249,400
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME ctypes STORE_NAME ctypes LOAD_CONST 0 LOAD_CONST None IMPORT_NAME pytest STORE_NAME pytest LOAD_CONST 0 LOAD_CONST ('lib',) IMPORT_NAME awkward_cpp.cpu_kernels IMPORT_FROM lib STORE_NAME lib POP_TOP LOAD_NAME pytest LOAD_ATTR mark LOAD_ATTR skip LOAD_CONST 'Unable to generate...
import ctypes import pytest from awkward_cpp.cpu_kernels import lib @pytest.mark.skip(reason="Unable to generate any tests for kernel") def test_cpuawkward_reduce_sum_bool_int8_64_1(): raise NotImplementedError("Unable to generate any tests for kernel")
data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_sum_bool_int8_64.py
225
86
221,603
LOAD_CONST "Dataset definition for penguins.\n\nDEPRECATED!\nIf you want to use the Penguins dataset builder class, use:\ntfds.builder_cls('penguins')\n" STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('lazy_builder_import',) IMPORT_NAME tensorflow_datasets.core IMPORT_FROM lazy_builder_import STORE_NAME lazy_builder_impo...
"""Dataset definition for penguins. DEPRECATED! If you want to use the Penguins dataset builder class, use: tfds.builder_cls('penguins') """ from tensorflow_datasets.core import lazy_builder_import Penguins = lazy_builder_import.LazyBuilderImport("penguins")
data/tensorflow-datasets-4.9.4/tensorflow_datasets/structured/penguins/penguins.py
130
86
255,208
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME ctypes STORE_NAME ctypes LOAD_CONST 0 LOAD_CONST None IMPORT_NAME pytest STORE_NAME pytest LOAD_CONST 0 LOAD_CONST ('lib',) IMPORT_NAME awkward_cpp.cpu_kernels IMPORT_FROM lib STORE_NAME lib POP_TOP LOAD_NAME pytest LOAD_ATTR mark LOAD_ATTR skip LOAD_CONST 'Unable to generate...
import ctypes import pytest from awkward_cpp.cpu_kernels import lib @pytest.mark.skip(reason="Unable to generate any tests for kernel") def test_cpuawkward_reduce_argmax_int16_64_1(): raise NotImplementedError("Unable to generate any tests for kernel")
data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmax_int16_64.py
225
86
221,461
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME json STORE_NAME json LOAD_CONST 0 LOAD_CONST ('tool',) IMPORT_NAME promptflow IMPORT_FROM tool STORE_NAME tool POP_TOP LOAD_NAME tool LOAD_NAME str LOAD_CONST ('input_str',) BUILD_CONST_KEY_MAP LOAD_CONST <code object convert_to_dict at 0x7f8ab63eaed0, file "f.py", line 6> LO...
import json from promptflow import tool @tool def convert_to_dict(input_str: str): try: return json.loads(input_str) except Exception as e: print("input is not valid, error: {}".format(e)) return {"category": "None", "evidence": "None"}
data/azure-ai-ml-1.13.0/tests/test_configs/flows/additional_includes/convert_to_dict.py
239
86
424,218
LOAD_CONST 0 LOAD_CONST ('E2BDataAnalysisTool', 'E2BDataAnalysisToolArguments', 'UploadedFile') IMPORT_NAME langchain_community.tools.e2b_data_analysis.tool IMPORT_FROM E2BDataAnalysisTool STORE_NAME E2BDataAnalysisTool IMPORT_FROM E2BDataAnalysisToolArguments STORE_NAME E2BDataAnalysisToolArguments IMPORT_FROM Uploade...
from langchain_community.tools.e2b_data_analysis.tool import ( E2BDataAnalysisTool, E2BDataAnalysisToolArguments, UploadedFile, ) __all__ = [ "UploadedFile", "E2BDataAnalysisToolArguments", "E2BDataAnalysisTool", ]
data/langchain-0.1.8/langchain/tools/e2b_data_analysis/tool.py
137
86
369,342
LOAD_CONST 0 LOAD_CONST ('TType', 'TMessageType', 'TFrozenDict', 'TException', 'TApplicationException') IMPORT_NAME thrift.Thrift IMPORT_FROM TType STORE_NAME TType IMPORT_FROM TMessageType STORE_NAME TMessageType IMPORT_FROM TFrozenDict STORE_NAME TFrozenDict IMPORT_FROM TException STORE_NAME TException IMPORT_FROM TA...
from thrift.Thrift import ( TType, TMessageType, TFrozenDict, TException, TApplicationException, ) from thrift.protocol.TProtocol import TProtocolException import sys import jaeger.ttypes import zipkincore.ttypes from thrift.transport import TTransport
data/opentelemetry_exporter_jaeger_thrift-1.21.0/src/opentelemetry/exporter/jaeger/thrift/gen/agent/ttypes.py
216
86
246,316
LOAD_CONST 0 LOAD_CONST ('unique',) IMPORT_NAME enum IMPORT_FROM unique STORE_NAME unique POP_TOP LOAD_CONST 4 LOAD_CONST ('StrEnum',) IMPORT_NAME _base_enum IMPORT_FROM StrEnum STORE_NAME StrEnum POP_TOP LOAD_NAME unique LOAD_BUILD_CLASS LOAD_CONST <code object IndexConvexityAdjustmentType at 0x7fab701ac780, file "f...
from enum import unique from ...._base_enum import StrEnum @unique class IndexConvexityAdjustmentType(StrEnum): NONE = "None" BLACK_SCHOLES = "BlackScholes" REPLICATION = "Replication" LIBOR_SWAP_METHOD = "LiborSwapMethod"
data/refinitiv-data-1.6.0/refinitiv/data/content/ipa/_enums/_index_convexity_adjustment_type.py
236
86
195,529
LOAD_CONST 0 LOAD_CONST ('Optional',) IMPORT_NAME typing IMPORT_FROM Optional STORE_NAME Optional POP_TOP LOAD_CONST 0 LOAD_CONST None IMPORT_NAME typer STORE_NAME typer LOAD_NAME typer LOAD_ATTR Argument LOAD_CONST None LOAD_CONST ('default',) CALL_FUNCTION BUILD_TUPLE LOAD_NAME Optional LOAD_NAME str BINARY_SUBSCR ...
from typing import Optional import typer def main(name: Optional[str] = typer.Argument(default=None)): if name is None: print("Hello World!") else: print(f"Hello {name}") if __name__ == "__main__": typer.run(main)
data/typer-0.9.0/docs_src/arguments/optional/tutorial002.py
212
86
224,543
LOAD_CONST 0 LOAD_CONST ('annotations',) IMPORT_NAME __future__ IMPORT_FROM annotations STORE_NAME annotations POP_TOP LOAD_CONST 0 LOAD_CONST ('TYPE_CHECKING',) IMPORT_NAME typing IMPORT_FROM TYPE_CHECKING STORE_NAME TYPE_CHECKING POP_TOP LOAD_CONST 0 LOAD_CONST ('ArgumentParser',) IMPORT_NAME argparse IMPORT_FROM A...
from __future__ import annotations from typing import TYPE_CHECKING from argparse import ArgumentParser from . import completions if TYPE_CHECKING: from argparse import _SubParsersAction def register(subparser: _SubParsersAction[ArgumentParser]) -> None: completions.register(subparser)
data/openai-1.12.0/src/openai/cli/_api/chat/__init__.py
255
86
63,612
LOAD_CONST "Dataset definition for bigearthnet.\n\nDEPRECATED!\nIf you want to use the Bigearthnet dataset builder class, use:\ntfds.builder_cls('bigearthnet')\n" STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('lazy_builder_import',) IMPORT_NAME tensorflow_datasets.core IMPORT_FROM lazy_builder_import STORE_NAME lazy_bui...
"""Dataset definition for bigearthnet. DEPRECATED! If you want to use the Bigearthnet dataset builder class, use: tfds.builder_cls('bigearthnet') """ from tensorflow_datasets.core import lazy_builder_import Bigearthnet = lazy_builder_import.LazyBuilderImport("bigearthnet")
data/tfds-nightly-4.9.4.dev202402210044/tensorflow_datasets/image_classification/bigearthnet.py
130
86
238,447
LOAD_CONST 4 STORE_NAME param LOAD_CONST 0 LOAD_CONST None IMPORT_NAME pandas STORE_NAME pd LOAD_NAME pd LOAD_ATTR DataFrame LOAD_CONST 1 LOAD_CONST 2 BUILD_LIST LOAD_CONST 3 LOAD_NAME param BINARY_ADD LOAD_CONST 4 BUILD_LIST LOAD_CONST ('A', 'B') BUILD_CONST_KEY_MAP LOAD_NAME pd LOAD_ATTR Index LOAD_CONST 'x0' LOAD...
param = 4 "" import pandas as pd "" df = pd.DataFrame( {"A": [1, 2], "B": [3 + param, 4]}, index=pd.Index(["x0", "x1"], name="x") ) df "" df.plot(kind="bar")
data/jupytext-1.16.1/tests/data/notebooks/outputs/ipynb_to_sphinx/nteract_with_parameter.py
146
86
312,595
LOAD_CONST 0 LOAD_CONST ('pass_environment',) IMPORT_NAME complex.cli IMPORT_FROM pass_environment STORE_NAME pass_environment POP_TOP LOAD_CONST 0 LOAD_CONST None IMPORT_NAME click STORE_NAME click LOAD_NAME click LOAD_ATTR command LOAD_CONST 'status' LOAD_CONST 'Shows file changes.' LOAD_CONST ('short_help',) CALL_...
from complex.cli import pass_environment import click @click.command("status", short_help="Shows file changes.") @pass_environment def cli(ctx): """Shows file changes in the current working directory.""" ctx.log("Changed files: none") ctx.vlog("bla bla bla, debug info")
data/click-8.1.7/examples/complex/complex/commands/cmd_status.py
180
86
117,111
LOAD_CONST 1 LOAD_CONST ('ChangeTrueDivToMulByInverse', 'FuseBatchNorm1dInLinear', 'FuseBatchNorm2dInConv2d', 'FuseBiasInLinear', 'MergeLinears', 'ReversibleTransformation', 'Transformation', 'compose') IMPORT_NAME transformations IMPORT_FROM ChangeTrueDivToMulByInverse STORE_NAME ChangeTrueDivToMulByInverse IMPORT_FRO...
from .transformations import ( # noqa ChangeTrueDivToMulByInverse, FuseBatchNorm1dInLinear, FuseBatchNorm2dInConv2d, FuseBiasInLinear, MergeLinears, ReversibleTransformation, Transformation, compose, )
data/optimum-1.17.1/optimum/fx/optimization/__init__.py
205
86
150,301
LOAD_CONST "Dataset definition for asqa.\n\nDEPRECATED!\nIf you want to use the Asqa dataset builder class, use:\ntfds.builder_cls('asqa')\n" STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('lazy_builder_import',) IMPORT_NAME tensorflow_datasets.core IMPORT_FROM lazy_builder_import STORE_NAME lazy_builder_import POP_TOP ...
"""Dataset definition for asqa. DEPRECATED! If you want to use the Asqa dataset builder class, use: tfds.builder_cls('asqa') """ from tensorflow_datasets.core import lazy_builder_import Asqa = lazy_builder_import.LazyBuilderImport("asqa")
data/tensorflow-datasets-4.9.4/tensorflow_datasets/question_answering/asqa/asqa.py
130
86
255,315
LOAD_CONST 0 LOAD_CONST ('annotations',) IMPORT_NAME __future__ IMPORT_FROM annotations STORE_NAME annotations POP_TOP LOAD_CONST 0 LOAD_CONST ('public',) IMPORT_NAME public IMPORT_FROM public STORE_NAME public POP_TOP LOAD_CONST 0 LOAD_CONST ('Column', 'Scalar', 'Value') IMPORT_NAME ibis.expr.types.generic IMPORT_FR...
from __future__ import annotations from public import public from ibis.expr.types.generic import Column, Scalar, Value @public class UUIDValue(Value): pass @public class UUIDScalar(Scalar, UUIDValue): pass @public class UUIDColumn(Column, UUIDValue): pass
data/ibis_framework-8.0.0/ibis/expr/types/uuid.py
400
86
161,642
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME subprocess STORE_NAME subprocess LOAD_CONST 0 LOAD_CONST None IMPORT_NAME argparse STORE_NAME argparse LOAD_CONST 0 LOAD_CONST None IMPORT_NAME sys STORE_NAME sys LOAD_NAME __name__ LOAD_CONST '__main__' COMPARE_OP == POP_JUMP_IF_FALSE LOAD_NAME argparse LOAD_METHOD Argument...
import subprocess import argparse import sys if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("prog") args = parser.parse_args() res = subprocess.run(args.prog) sys.exit(res.returncode - 42)
data/meson-1.3.2/test cases/windows/13 test argument extra paths/test/test_run_exe.py
161
86
81,935
LOAD_CONST 'DO NOT EDIT.\n\nThis file was autogenerated. Do not edit it by hand,\nsince your modifications would be overwritten.\n' STORE_NAME __doc__ LOAD_CONST 0 LOAD_CONST ('decode_predictions',) IMPORT_NAME keras_core.src.applications.mobilenet_v3 IMPORT_FROM decode_predictions STORE_NAME decode_predictions POP_TO...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras_core.src.applications.mobilenet_v3 import decode_predictions from keras_core.src.applications.mobilenet_v3 import preprocess_input
data/keras-core-0.1.7/keras_core/applications/mobilenet_v3/__init__.py
143
86
385,621
LOAD_CONST './cspnext-s_8xb256-rsb-a1-600e_in1k.py' STORE_NAME _base_ LOAD_NAME dict LOAD_NAME dict LOAD_CONST 0.67 LOAD_CONST 0.75 LOAD_CONST ('deepen_factor', 'widen_factor') CALL_FUNCTION LOAD_NAME dict LOAD_CONST 768 LOAD_CONST ('in_channels',) CALL_FUNCTION LOAD_CONST ('backbone', 'head') CALL_FUNCTION STORE_NAM...
_base_ = "./cspnext-s_8xb256-rsb-a1-600e_in1k.py" model = dict( backbone=dict(deepen_factor=0.67, widen_factor=0.75), head=dict(in_channels=768) )
data/mmdet-3.3.0/mmdet/.mim/configs/rtmdet/classification/cspnext-m_8xb256-rsb-a1-600e_in1k.py
115
86
151,261
LOAD_CONST 0 LOAD_CONST ('Enum',) IMPORT_NAME enum IMPORT_FROM Enum STORE_NAME Enum POP_TOP LOAD_BUILD_CLASS LOAD_CONST <code object IncidentStatus at 0x7fab70027420, file "f.py", line 4> LOAD_CONST 'IncidentStatus' MAKE_FUNCTION LOAD_CONST 'IncidentStatus' LOAD_NAME str LOAD_NAME Enum CALL_FUNCTION STORE_NAME Inciden...
from enum import Enum class IncidentStatus(str, Enum): Active = ("active",) Resolved = ("resolved",) InProgress = ("inProgress",) Redirected = ("redirected",) UnknownFutureValue = ("unknownFutureValue",) AwaitingAction = ("awaitingAction",)
data/msgraph-sdk-1.1.0/msgraph/generated/models/security/incident_status.py
205
86
262,105
LOAD_CONST 0 LOAD_CONST ('annotations',) IMPORT_NAME __future__ IMPORT_FROM annotations STORE_NAME annotations POP_TOP LOAD_CONST 0 LOAD_CONST ('public',) IMPORT_NAME public IMPORT_FROM public STORE_NAME public POP_TOP LOAD_CONST 0 LOAD_CONST ('Column', 'Scalar', 'Value') IMPORT_NAME ibis.expr.types.generic IMPORT_FR...
from __future__ import annotations from public import public from ibis.expr.types.generic import Column, Scalar, Value @public class SetValue(Value): pass @public class SetScalar(Scalar, SetValue): pass @public class SetColumn(Column, SetValue): pass
data/ibis_framework-8.0.0/ibis/expr/types/collections.py
399
86
161,643
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME ctypes STORE_NAME ctypes LOAD_CONST 0 LOAD_CONST None IMPORT_NAME pytest STORE_NAME pytest LOAD_CONST 0 LOAD_CONST ('lib',) IMPORT_NAME awkward_cpp.cpu_kernels IMPORT_FROM lib STORE_NAME lib POP_TOP LOAD_NAME pytest LOAD_ATTR mark LOAD_ATTR skip LOAD_CONST 'Unable to generate...
import ctypes import pytest from awkward_cpp.cpu_kernels import lib @pytest.mark.skip(reason="Unable to generate any tests for kernel") def test_cpuawkward_reduce_argmin_int32_64_1(): raise NotImplementedError("Unable to generate any tests for kernel")
data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmin_int32_64.py
226
86
221,592
LOAD_CONST 0 LOAD_CONST ('SparkSession',) IMPORT_NAME pyspark.sql IMPORT_FROM SparkSession STORE_NAME SparkSession POP_TOP LOAD_CONST 1 LOAD_CONST ('MosaicDataFrameReader',) IMPORT_NAME mosaic_data_frame_reader IMPORT_FROM MosaicDataFrameReader STORE_NAME MosaicDataFrameReader POP_TOP LOAD_CONST <code object read at ...
from pyspark.sql import SparkSession from .mosaic_data_frame_reader import MosaicDataFrameReader def read(): """ Returns a MosaicDataFrameReader for reading MosaicDataFrames. """ spark = SparkSession.builder.getOrCreate() return MosaicDataFrameReader()
data/databricks-mosaic-0.4.0/mosaic/readers/readers.py
156
86
52,977
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME re STORE_NAME re LOAD_CONST 0 LOAD_CONST ('join',) IMPORT_NAME os.path IMPORT_FROM join STORE_NAME join POP_TOP LOAD_CONST 0 LOAD_CONST None IMPORT_NAME setuptools STORE_NAME setuptools LOAD_NAME setuptools LOAD_ATTR setup LOAD_NAME re LOAD_METHOD search LOAD_CONST "__versi...
import re from os.path import join import setuptools setuptools.setup( version=re.search( "__version__ = '([^']+)'", open(join("mock", "__init__.py")).read() ).group(1), long_description=open("README.rst").read(), )
data/mock-5.1.0/setup.py
163
86
427,474
LOAD_CONST 0 LOAD_CONST ('print_function', 'division', 'absolute_import', 'annotations') IMPORT_NAME __future__ IMPORT_FROM print_function STORE_NAME print_function IMPORT_FROM division STORE_NAME division IMPORT_FROM absolute_import STORE_NAME absolute_import IMPORT_FROM annotations STORE_NAME annotations POP_TOP LOA...
from __future__ import print_function, division, absolute_import, annotations from ..message import BackendMessage class EndOfBatchResponse(BackendMessage): message_id = b"J" def __init__(self, data): BackendMessage.__init__(self) BackendMessage.register(EndOfBatchResponse)
data/vertica-python-1.3.8/vertica_python/vertica/messages/backend_messages/end_of_batch_response.py
304
86
287,015
LOAD_CONST 0 LOAD_CONST ('re_path',) IMPORT_NAME django.urls IMPORT_FROM re_path STORE_NAME re_path POP_TOP LOAD_CONST 0 LOAD_CONST ('include',) IMPORT_NAME django.conf.urls IMPORT_FROM include STORE_NAME include POP_TOP LOAD_CONST 0 LOAD_CONST ('admin',) IMPORT_NAME grappelli.tests IMPORT_FROM admin STORE_NAME admin...
from django.urls import re_path from django.conf.urls import include from grappelli.tests import admin urlpatterns = [ re_path(r"^admin/", admin.site.urls), re_path(r"^grappelli/", include("grappelli.urls")), ]
data/django-grappelli-3.0.8/grappelli/tests/urls.py
154
86
20,649
LOAD_CONST 'Constants for QVR Pro component.' STORE_NAME __doc__ LOAD_CONST 'qvr_pro' STORE_NAME DOMAIN LOAD_CONST 'QVR' STORE_NAME SHORT_NAME LOAD_CONST 'exclude_channels' STORE_NAME CONF_EXCLUDE_CHANNELS LOAD_CONST 'stop_record' STORE_NAME SERVICE_STOP_RECORD LOAD_CONST 'start_record' STORE_NAME SERVICE_START_RE...
"""Constants for QVR Pro component.""" DOMAIN = "qvr_pro" SHORT_NAME = "QVR" CONF_EXCLUDE_CHANNELS = "exclude_channels" SERVICE_STOP_RECORD = "stop_record" SERVICE_START_RECORD = "start_record"
data/homeassistant-2024.2.2/homeassistant/components/qvr_pro/const.py
108
86
406,799
LOAD_CONST 0 LOAD_CONST None IMPORT_NAME uwsgi STORE_NAME uwsgi LOAD_CONST 5 STORE_NAME sleepvalue LOAD_CONST <code object application at 0x7f8e44ccfc00, file "f.py", line 6> LOAD_CONST 'application' MAKE_FUNCTION STORE_NAME application LOAD_CONST None RETURN_VALUE LOAD_FAST start_response LOAD_CONST '200 Ok' LOAD_C...
import uwsgi sleepvalue = 5 def application(env, start_response): start_response("200 Ok", [("Content-type", "text/html")]) yield uwsgi.async_sleep(sleepvalue) yield "<h1>Hello World after %d seconds</h1>" % sleepvalue
data/uwsgi-2.0.24/tests/sleeping_async.py
164
86
398,187