input stringlengths 28 198k | output stringlengths 3 71k | file stringlengths 19 330 | input_tokens int64 5 159k | output_tokens int64 3 9.07k | __index_level_0__ int64 2 449k |
|---|---|---|---|---|---|
LOAD_CONST "Dataset definition for siscore.\n\nDEPRECATED!\nIf you want to use the Siscore dataset builder class, use:\ntfds.builder_cls('siscore')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_builder_import
... | """Dataset definition for siscore.
DEPRECATED!
If you want to use the Siscore dataset builder class, use:
tfds.builder_cls('siscore')
"""
from tensorflow_datasets.core import lazy_builder_import
Siscore = lazy_builder_import.LazyBuilderImport("siscore")
| data/tfds-nightly-4.9.4.dev202402210044/tensorflow_datasets/image_classification/siscore/siscore.py | 130 | 86 | 238,536 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_reduce_argmax_uint32_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmax_uint32_64.py | 226 | 86 | 221,413 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME unittest
STORE_NAME unittest
LOAD_CONST 2
LOAD_CONST ('Five9',)
IMPORT_NAME five9
IMPORT_FROM Five9
STORE_NAME Five9
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object Common at 0x7fab8212bd20, file "f.py", line 6>
LOAD_CONST 'Common'
MAKE_FUNCTION
LOAD_CONST 'Common'
LOAD_NAME ... | import unittest
from ..five9 import Five9
class Common(unittest.TestCase):
def setUp(self):
super(Common, self).setUp()
self.user = "username@something.com"
self.password = "password"
self.five9 = Five9(self.user, self.password)
| data/five9-0.0.3b192/five9/tests/common.py | 280 | 86 | 175,960 |
LOAD_CONST "Dataset definition for so2sat.\n\nDEPRECATED!\nIf you want to use the So2sat dataset builder class, use:\ntfds.builder_cls('so2sat')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_builder_import
POP... | """Dataset definition for so2sat.
DEPRECATED!
If you want to use the So2sat dataset builder class, use:
tfds.builder_cls('so2sat')
"""
from tensorflow_datasets.core import lazy_builder_import
So2sat = lazy_builder_import.LazyBuilderImport("so2sat")
| data/tensorflow-datasets-4.9.4/tensorflow_datasets/image_classification/so2sat.py | 130 | 86 | 255,372 |
LOAD_CONST '\nErrors around Submission Locking\n'
STORE_NAME __doc__
LOAD_BUILD_CLASS
LOAD_CONST <code object SubmissionLockContestedError at 0x7f8e2ff254b0, file "f.py", line 6>
LOAD_CONST 'SubmissionLockContestedError'
MAKE_FUNCTION
LOAD_CONST 'SubmissionLockContestedError'
LOAD_NAME Exception
CALL_FUNCTION
STORE_NA... | """
Errors around Submission Locking
"""
class SubmissionLockContestedError(Exception):
"""
Error indicating trying to modify a lock that the user does not have access to modify.
"""
error_code = "ERR_LOCK_CONTESTED"
def get_error_code(self):
return self.error_code
| data/ora2-6.1.0/openassessment/staffgrader/errors/submission_lock.py | 249 | 86 | 116,308 |
LOAD_BUILD_CLASS
LOAD_CONST <code object Positional at 0x7faa7c009420, file "f.py", line 1>
LOAD_CONST 'Positional'
MAKE_FUNCTION
LOAD_CONST 'Positional'
LOAD_NAME object
CALL_FUNCTION
STORE_NAME Positional
LOAD_CONST None
RETURN_VALUE
LOAD_NAME __name__
STORE_NAME __module__
LOAD_CONST 'Positional'
STORE_NAME __qualn... | class Positional(object):
def __init__(self, help="", type=None, choices=None, metavar=None):
self.help = help
self.kind = "positional"
self.abbrev = None
self.type = type
self.choices = choices
self.metavar = metavar
| data/plac-1.4.2/doc/annotations.py | 226 | 86 | 112,196 |
LOAD_CONST 1
LOAD_CONST ('DeepSpeedTransformerLayer', 'DeepSpeedTransformerConfig')
IMPORT_NAME transformer
IMPORT_FROM DeepSpeedTransformerLayer
STORE_NAME DeepSpeedTransformerLayer
IMPORT_FROM DeepSpeedTransformerConfig
STORE_NAME DeepSpeedTransformerConfig
POP_TOP
LOAD_CONST 1
LOAD_CONST ('DeepSpeedInferenceConfig'... | from .transformer import DeepSpeedTransformerLayer, DeepSpeedTransformerConfig
from .inference.config import DeepSpeedInferenceConfig
from ...model_implementations.transformers.ds_transformer import (
DeepSpeedTransformerInference,
)
from .inference.moe_inference import DeepSpeedMoEInferenceConfig, DeepSpeedMoEInfe... | data/deepspeed-0.13.2/deepspeed/ops/transformer/__init__.py | 218 | 86 | 84,706 |
LOAD_CONST 0
LOAD_CONST ('reactor',)
IMPORT_NAME twisted.internet
IMPORT_FROM reactor
STORE_NAME reactor
POP_TOP
LOAD_CONST 0
LOAD_CONST ('WSGIResource',)
IMPORT_NAME twisted.web.wsgi
IMPORT_FROM WSGIResource
STORE_NAME WSGIResource
POP_TOP
LOAD_CONST 2
LOAD_CONST ('exposition', 'REGISTRY')
IMPORT_NAME
IMPORT_FROM ex... | from twisted.internet import reactor
from twisted.web.wsgi import WSGIResource
from .. import exposition, REGISTRY
MetricsResource = lambda registry=REGISTRY: WSGIResource(
reactor, reactor.getThreadPool(), exposition.make_wsgi_app(registry)
)
| data/prometheus_client-0.20.0/prometheus_client/twisted/_exposition.py | 217 | 86 | 154,254 |
LOAD_CONST '\nVersion 3 classes\n=================\nThis class is a iterator for version 3 API call\n\n.. autoclass:: APIResultIterator\n :members:\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('APIIterator',)
IMPORT_NAME restfly.iterator
IMPORT_FROM APIIterator
STORE_NAME APIIterator
POP_TOP
LOAD_BUILD_CLASS
LOAD_C... | """
Version 3 classes
=================
This class is a iterator for version 3 API call
.. autoclass:: APIResultIterator
:members:
"""
from restfly.iterator import APIIterator
class APIResultIterator(APIIterator):
"""
Iterator class for version 3 API
"""
pass
| data/pyTenable-1.4.20/tenable/io/v3/base/iterators/iterator.py | 206 | 86 | 175,191 |
LOAD_BUILD_CLASS
LOAD_CONST <code object Cloud at 0x7fab42018660, file "f.py", line 1>
LOAD_CONST 'Cloud'
MAKE_FUNCTION
LOAD_CONST 'Cloud'
CALL_FUNCTION
STORE_NAME Cloud
LOAD_CONST None
RETURN_VALUE
LOAD_NAME __name__
STORE_NAME __module__
LOAD_CONST 'Cloud'
STORE_NAME __qualname__
LOAD_CONST <code object create_pars... | class Cloud:
def create_parser(self, argparser):
return NotImplemented
def set_cloud_config_from_arguments(self, arguments):
return NotImplemented
def set_composition_arguments(self, comp_group):
pass
def get_composition(self, **kwargs):
pass
| data/qds_sdk-1.16.1/qds_sdk/cloud/cloud.py | 347 | 86 | 358,218 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_reduce_argmax_uint16_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmax_uint16_64.py | 225 | 86 | 221,418 |
LOAD_CONST '\nversion info\n'
STORE_NAME __doc__
LOAD_CONST '1.4.7'
STORE_NAME VERSION
LOAD_CONST 'Steve McGrath <steve@mcgrath.sh>'
STORE_NAME AUTHOR
LOAD_CONST 'REST API library framework'
STORE_NAME DESCRIPTION
LOAD_NAME tuple
LOAD_CONST <code object <genexpr> at 0x7f8e2ff5e420, file "f.py", line 8>
LOAD_CONST '... | """
version info
"""
VERSION = "1.4.7"
AUTHOR = "Steve McGrath <steve@mcgrath.sh>"
DESCRIPTION = "REST API library framework"
version_info = tuple(int(d) for d in VERSION.split("-", maxsplit=1)[0].split("."))
| data/restfly-1.4.7/restfly/version.py | 208 | 86 | 405,022 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME not_existing
STORE_NAME not_existing
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME os
STORE_NAME os
LOAD_CONST 0
LOAD_CONST ('abspath',)
IMPORT_NAME os.path
IMPORT_FROM abspath
STORE_NAME abspath
POP_TOP
LOAD_CONST 0
LOAD_CONST ('not_existing',)
IMPORT_NAME os.path
IMPORT_FROM not... | import not_existing
import os
from os.path import abspath
from os.path import not_existing
from datetime import date
date.today
date.not_existing_attribute
from datetime.date import today
import datetime.datetime
import not_existing_nested.date
import os.path
| data/jedi-0.19.1/test/static_analysis/imports.py | 231 | 86 | 421,043 |
LOAD_CONST 0
LOAD_CONST ('IS_PYDANTIC_V2',)
IMPORT_NAME beanie.odm.utils.pydantic
IMPORT_FROM IS_PYDANTIC_V2
STORE_NAME IS_PYDANTIC_V2
POP_TOP
LOAD_NAME IS_PYDANTIC_V2
POP_JUMP_IF_FALSE
LOAD_CONST 0
LOAD_CONST ('DecimalAnnotation',)
IMPORT_NAME beanie.odm.custom_types.decimal
IMPORT_FROM DecimalAnnotation
STORE_NAME ... | from beanie.odm.utils.pydantic import IS_PYDANTIC_V2
if IS_PYDANTIC_V2:
from beanie.odm.custom_types.decimal import DecimalAnnotation
else:
from decimal import Decimal as DecimalAnnotation
__all__ = [
"DecimalAnnotation",
]
| data/beanie-1.25.0/beanie/odm/custom_types/__init__.py | 160 | 86 | 323,825 |
LOAD_CONST 1
LOAD_CONST ('post_processor_catalogue', 'register_post_processor', 'remove_post_processor')
IMPORT_NAME catalogue
IMPORT_FROM post_processor_catalogue
STORE_NAME post_processor_catalogue
IMPORT_FROM register_post_processor
STORE_NAME register_post_processor
IMPORT_FROM remove_post_processor
STORE_NAME remo... | from .catalogue import (
post_processor_catalogue,
register_post_processor,
remove_post_processor,
)
from .base import PostProcessor
from .filth_replacer import FilthReplacer
from .prefix_suffix import PrefixSuffixReplacer
from .remover import FilthRemover
| data/scrubadub-2.0.1/scrubadub/post_processors/__init__.py | 227 | 86 | 236,920 |
LOAD_CONST 0
LOAD_CONST ('TestCase', 'min_os_level')
IMPORT_NAME PyObjCTools.TestSupport
IMPORT_FROM TestCase
STORE_NAME TestCase
IMPORT_FROM min_os_level
STORE_NAME min_os_level
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME objc
STORE_NAME objc
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME MapKit
STORE_NAME MapKit
LO... | from PyObjCTools.TestSupport import TestCase, min_os_level
import objc
import MapKit
class TestMKPointAnnotation(TestCase):
@min_os_level("10.9")
def testClasses(self):
self.assertIsInstance(MapKit.MKPointAnnotation, objc.objc_class)
| data/pyobjc-framework-MapKit-10.1/PyObjCTest/test_mkpointannotation.py | 298 | 86 | 405,960 |
LOAD_CONST 0
LOAD_CONST ('BaseHook', 'DatahubGenericHook', 'DatahubKafkaHook', 'DatahubRestHook')
IMPORT_NAME datahub_airflow_plugin.hooks.datahub
IMPORT_FROM BaseHook
STORE_NAME BaseHook
IMPORT_FROM DatahubGenericHook
STORE_NAME DatahubGenericHook
IMPORT_FROM DatahubKafkaHook
STORE_NAME DatahubKafkaHook
IMPORT_FROM Da... | from datahub_airflow_plugin.hooks.datahub import (
BaseHook,
DatahubGenericHook,
DatahubKafkaHook,
DatahubRestHook,
)
__all__ = ["DatahubRestHook", "DatahubKafkaHook", "DatahubGenericHook", "BaseHook"]
| data/acryl-datahub-0.12.1.5/src/datahub_provider/hooks/datahub.py | 154 | 86 | 435,790 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_reduce_argmax_int64_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmax_int64_64.py | 225 | 86 | 221,498 |
LOAD_CONST 0
LOAD_CONST ('Enum',)
IMPORT_NAME enum
IMPORT_FROM Enum
STORE_NAME Enum
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object WifiBand at 0x7fab4298ec00, file "f.py", line 4>
LOAD_CONST 'WifiBand'
MAKE_FUNCTION
LOAD_CONST 'WifiBand'
LOAD_NAME str
LOAD_NAME Enum
CALL_FUNCTION
STORE_NAME WifiBand
LOAD_CONST None
... | from enum import Enum
class WifiBand(str, Enum):
Unknown = ("unknown",)
Frequency24GHz = ("frequency24GHz",)
Frequency50GHz = ("frequency50GHz",)
Frequency60GHz = ("frequency60GHz",)
UnknownFutureValue = ("unknownFutureValue",)
| data/msgraph-sdk-1.1.0/msgraph/generated/models/call_records/wifi_band.py | 203 | 86 | 262,402 |
LOAD_CONST 0
LOAD_CONST ('FileCache', 'SeparateBodyFileCache')
IMPORT_NAME pigar._vendor.pip._vendor.cachecontrol.caches.file_cache
IMPORT_FROM FileCache
STORE_NAME FileCache
IMPORT_FROM SeparateBodyFileCache
STORE_NAME SeparateBodyFileCache
POP_TOP
LOAD_CONST 0
LOAD_CONST ('RedisCache',)
IMPORT_NAME pigar._vendor.pip... | from pigar._vendor.pip._vendor.cachecontrol.caches.file_cache import (
FileCache,
SeparateBodyFileCache,
)
from pigar._vendor.pip._vendor.cachecontrol.caches.redis_cache import RedisCache
__all__ = ["FileCache", "SeparateBodyFileCache", "RedisCache"]
| data/pigar-2.1.3/pigar/_vendor/pip/_vendor/cachecontrol/caches/__init__.py | 144 | 86 | 241,309 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_reduce_argmin_uint32_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmin_uint32_64.py | 225 | 86 | 221,391 |
LOAD_CONST 'DO NOT EDIT.\n\nThis file was autogenerated. Do not edit it by hand,\nsince your modifications would be overwritten.\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('ResNet50',)
IMPORT_NAME keras.src.applications.resnet
IMPORT_FROM ResNet50
STORE_NAME ResNet50
POP_TOP
LOAD_CONST 0
LOAD_CONST ('decode_predi... | """DO NOT EDIT.
This file was autogenerated. Do not edit it by hand,
since your modifications would be overwritten.
"""
from keras.src.applications.resnet import ResNet50
from keras.src.applications.resnet import decode_predictions
from keras.src.applications.resnet import preprocess_input
| data/keras-nightly-3.1.0.dev2024022103/keras/applications/resnet50/__init__.py | 166 | 86 | 284,938 |
LOAD_CONST 0
LOAD_CONST ('SubProcessCompressor',)
IMPORT_NAME pipeline.compressors
IMPORT_FROM SubProcessCompressor
STORE_NAME SubProcessCompressor
POP_TOP
LOAD_CONST 0
LOAD_CONST ('settings',)
IMPORT_NAME pipeline.conf
IMPORT_FROM settings
STORE_NAME settings
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object CSSMinCo... | from pipeline.compressors import SubProcessCompressor
from pipeline.conf import settings
class CSSMinCompressor(SubProcessCompressor):
def compress_css(self, css):
command = (settings.CSSMIN_BINARY, settings.CSSMIN_ARGUMENTS)
return self.execute_command(command, css)
| data/django-pipeline-3.0.0/pipeline/compressors/cssmin.py | 278 | 86 | 76,419 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_reduce_argmax_uint64_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmax_uint64_64.py | 225 | 86 | 221,399 |
LOAD_CONST 0
LOAD_CONST ('TestCase', 'min_os_level')
IMPORT_NAME PyObjCTools.TestSupport
IMPORT_FROM TestCase
STORE_NAME TestCase
IMPORT_FROM min_os_level
STORE_NAME min_os_level
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME Quartz
STORE_NAME Quartz
LOAD_BUILD_CLASS
LOAD_CONST <code object TestIKFilterBrowserView ... | from PyObjCTools.TestSupport import TestCase, min_os_level
import Quartz
class TestIKFilterBrowserView(TestCase):
@min_os_level("10.5")
def testMethods(self):
self.assertArgIsBOOL(Quartz.IKFilterBrowserView.setPreviewState_, 0)
| data/pyobjc-framework-Quartz-10.1/PyObjCTest/test_IKFilterBrowserView.py | 291 | 86 | 133,565 |
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME azure.cli.core.aaz
IMPORT_STAR
LOAD_NAME register_command_group
LOAD_CONST 'eventhubs namespace authorization-rule keys'
CALL_FUNCTION
LOAD_BUILD_CLASS
LOAD_CONST <code object __CMDGroup at 0x7fab415056f0, file "f.py", line 4>
LOAD_CONST '__CMDGroup'
MAKE_FUNCTION
LOAD_CONST... | from azure.cli.core.aaz import *
@register_command_group(
"eventhubs namespace authorization-rule keys",
)
class __CMDGroup(AAZCommandGroup):
"""Manage Azure EventHubs Authorizationrule connection strings for Namespace."""
pass
__all__ = ["__CMDGroup"]
| data/azure-cli-2.57.0/azure/cli/command_modules/eventhubs/aaz/profile_2019_03_01_hybrid/eventhubs/namespace/authorization_rule/keys/__cmd_group.py | 196 | 86 | 374,726 |
LOAD_CONST 0
LOAD_CONST ('ClientValue',)
IMPORT_NAME office365.runtime.client_value
IMPORT_FROM ClientValue
STORE_NAME ClientValue
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object SocialLink at 0x7fab64179e40, file "f.py", line 4>
LOAD_CONST 'SocialLink'
MAKE_FUNCTION
LOAD_CONST 'SocialLink'
LOAD_NAME ClientValue
CALL... | from office365.runtime.client_value import ClientValue
class SocialLink(ClientValue):
"""The SocialLink class defines a link that includes a URI and text representation. This class is used to represent
the location of a web site."""
@property
def entity_type_name(self):
return "SP.Social.Soci... | data/Office365-REST-Python-Client-2.5.5/office365/sharepoint/social/link.py | 241 | 86 | 189,644 |
LOAD_CONST "Dataset definition for dsprites.\n\nDEPRECATED!\nIf you want to use the Dsprites dataset builder class, use:\ntfds.builder_cls('dsprites')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_builder_impo... | """Dataset definition for dsprites.
DEPRECATED!
If you want to use the Dsprites dataset builder class, use:
tfds.builder_cls('dsprites')
"""
from tensorflow_datasets.core import lazy_builder_import
Dsprites = lazy_builder_import.LazyBuilderImport("dsprites")
| data/tensorflow-datasets-4.9.4/tensorflow_datasets/image/dsprites.py | 130 | 86 | 255,260 |
LOAD_CONST 'The Reader Component of a Source.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('FileReader',)
IMPORT_NAME butterfree.extract.readers.file_reader
IMPORT_FROM FileReader
STORE_NAME FileReader
POP_TOP
LOAD_CONST 0
LOAD_CONST ('KafkaReader',)
IMPORT_NAME butterfree.extract.readers.kafka_reader
IMPORT_FROM Kaf... | """The Reader Component of a Source."""
from butterfree.extract.readers.file_reader import FileReader
from butterfree.extract.readers.kafka_reader import KafkaReader
from butterfree.extract.readers.table_reader import TableReader
__all__ = ["FileReader", "KafkaReader", "TableReader"]
| data/butterfree-1.2.3/butterfree/extract/readers/__init__.py | 164 | 86 | 9,059 |
LOAD_CONST './libra-faster-rcnn_r50_fpn_1x_coco.py'
STORE_NAME _base_
LOAD_NAME dict
LOAD_NAME dict
LOAD_CONST 101
LOAD_NAME dict
LOAD_CONST 'Pretrained'
LOAD_CONST 'torchvision://resnet101'
LOAD_CONST ('type', 'checkpoint')
CALL_FUNCTION
LOAD_CONST ('depth', 'init_cfg')
CALL_FUNCTION
LOAD_CONST ('backbone',)
CALL_... | _base_ = "./libra-faster-rcnn_r50_fpn_1x_coco.py"
model = dict(
backbone=dict(
depth=101,
init_cfg=dict(type="Pretrained", checkpoint="torchvision://resnet101"),
)
)
| data/mmdet-3.3.0/mmdet/.mim/configs/libra_rcnn/libra-faster-rcnn_r101_fpn_1x_coco.py | 111 | 86 | 151,457 |
LOAD_CONST 0
LOAD_CONST ('setup',)
IMPORT_NAME setuptools
IMPORT_FROM setup
STORE_NAME setup
POP_TOP
LOAD_NAME RuntimeError
LOAD_CONST 'This mimics the build error on unmatched platform'
CALL_FUNCTION
RAISE_VARARGS
LOAD_NAME setup
LOAD_CONST 'demo'
LOAD_CONST '0.0.1'
LOAD_CONST 'test demo'
LOAD_CONST 'demo'
BUILD... | from setuptools import setup
if True:
raise RuntimeError("This mimics the build error on unmatched platform")
setup(
name="demo",
version="0.0.1",
description="test demo",
py_modules=["demo"],
python_requires=">=3.3",
)
| data/pdm-2.12.3/tests/fixtures/projects/demo-failure-no-dep/setup.py | 123 | 86 | 354,242 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME gto.log
STORE_NAME gto
LOAD_CONST 0
LOAD_CONST ('api',)
IMPORT_NAME gto
IMPORT_FROM api
STORE_NAME api
POP_TOP
LOAD_CONST 0
LOAD_CONST ('__version__',)
IMPORT_NAME gto._version
IMPORT_FROM __version__
STORE_NAME __version__
POP_TOP
LOAD_CONST 0
LOAD_CONST ('CONFIG',)
IMPORT_N... | import gto.log # noqa
from gto import api
from gto._version import __version__
from gto.config import CONFIG
from gto.index import RepoIndexManager
from gto.registry import GitRegistry
__all__ = ["api", "CONFIG", "RepoIndexManager", "GitRegistry", "__version__"]
| data/gto-1.7.0/gto/__init__.py | 207 | 86 | 354,192 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME typing
STORE_NAME t
LOAD_CONST 0
LOAD_CONST ('expressions',)
IMPORT_NAME sqlglot
IMPORT_FROM expressions
STORE_NAME exp
POP_TOP
LOAD_NAME t
LOAD_ATTR Dict
LOAD_NAME exp
LOAD_ATTR Identifier
LOAD_NAME exp
LOAD_ATTR Identifier
BUILD_TUPLE
BINARY_SUBSCR
LOAD_CONST ('replacement_m... | import typing as t
from sqlglot import expressions as exp
def replace_id_value(node, replacement_mapping: t.Dict[exp.Identifier, exp.Identifier]):
if isinstance(node, exp.Identifier) and node in replacement_mapping:
node = node.replace(replacement_mapping[node].copy())
return node
| data/acryl-sqlglot-21.1.2.dev10/sqlglot/dataframe/sql/transforms.py | 198 | 86 | 12,748 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME sys
STORE_NAME sys
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME os
STORE_NAME os
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME PIL.Image
STORE_NAME PIL
LOAD_CONST '/'
LOAD_NAME PIL
LOAD_ATTR Image
STORE_ATTR __file__
LOAD_NAME sys
LOAD_ATTR _MEIPASS
STORE_NAME basedir
LOAD_NAME PIL
... | import sys
import os
import PIL.Image
PIL.Image.__file__ = "/"
basedir = sys._MEIPASS
im = PIL.Image.open(os.path.join(basedir, "tinysample.tiff"))
im.save(os.path.join(basedir, "tinysample.png"))
| data/pyinstaller-6.4.0/tests/functional/scripts/pyi_lib_PIL_img_conversion.py | 159 | 86 | 184,854 |
LOAD_CONST 'A module for compatibility with the docutils>=0.17 `include` directive, in RST documents:\n\nFor example::\n\n .. include:: path/to/file.md\n :parser: myst_parser.docutils_\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('Parser',)
IMPORT_NAME myst_parser.parsers.docutils_
IMPORT_FROM Parser
STORE_NAME Pars... | """A module for compatibility with the docutils>=0.17 `include` directive, in RST documents:
For example::
.. include:: path/to/file.md
:parser: myst_parser.docutils_
"""
from myst_parser.parsers.docutils_ import Parser # noqa: F401
| data/myst_parser-2.0.0/myst_parser/docutils_.py | 105 | 86 | 192,736 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pyperf
STORE_NAME pyperf
LOAD_NAME pyperf
LOAD_METHOD Runner
CALL_METHOD
STORE_NAME runner
LOAD_NAME runner
LOAD_ATTR timeit
LOAD_CONST 'sorted(list(range(1000)), key=lambda x: x)'
LOAD_CONST 'sorted(s, key=f)'
LOAD_CONST 'f = lambda x: x; s = list(range(1000))'
LOAD_CONST ... | import pyperf
runner = pyperf.Runner()
runner.timeit(
"sorted(list(range(1000)), key=lambda x: x)",
stmt="sorted(s, key=f)",
setup="f = lambda x: x; s = list(range(1000))",
)
| data/pyperf-2.6.2/doc/examples/bench_timeit.py | 117 | 86 | 123,054 |
LOAD_CONST "Dataset definition for bigearthnet.\n\nDEPRECATED!\nIf you want to use the Bigearthnet dataset builder class, use:\ntfds.builder_cls('bigearthnet')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_bui... | """Dataset definition for bigearthnet.
DEPRECATED!
If you want to use the Bigearthnet dataset builder class, use:
tfds.builder_cls('bigearthnet')
"""
from tensorflow_datasets.core import lazy_builder_import
Bigearthnet = lazy_builder_import.LazyBuilderImport("bigearthnet")
| data/tensorflow-datasets-4.9.4/tensorflow_datasets/image_classification/bigearthnet.py | 130 | 86 | 255,386 |
LOAD_CONST 'train'
STORE_NAME TRAIN_LOGGER_NAME
LOAD_CONST 'hyperparameter-optimization'
STORE_NAME HYPEROPT_LOGGER_NAME
LOAD_CONST 'model.pt'
STORE_NAME MODEL_FILE_NAME
LOAD_CONST 'test_scores.csv'
STORE_NAME TEST_SCORES_FILE_NAME
LOAD_CONST 'hyperopt_seeds.txt'
STORE_NAME HYPEROPT_SEED_FILE_NAME
LOAD_CONST None
R... | TRAIN_LOGGER_NAME = "train"
HYPEROPT_LOGGER_NAME = "hyperparameter-optimization"
MODEL_FILE_NAME = "model.pt"
TEST_SCORES_FILE_NAME = "test_scores.csv"
HYPEROPT_SEED_FILE_NAME = "hyperopt_seeds.txt"
| data/chemprop-1.6.1/chemprop/constants.py | 106 | 86 | 367,595 |
LOAD_CONST 0
LOAD_CONST ('setup', 'Extension')
IMPORT_NAME setuptools
IMPORT_FROM setup
STORE_NAME setup
IMPORT_FROM Extension
STORE_NAME Extension
POP_TOP
LOAD_NAME Extension
LOAD_CONST 'lru._lru'
LOAD_CONST 'src/lru/_lru.c'
BUILD_LIST
CALL_FUNCTION
BUILD_LIST
STORE_NAME extensions
LOAD_CONST True
LOAD_CONST ''
LOA... | from setuptools import setup, Extension
extensions = [
Extension("lru._lru", ["src/lru/_lru.c"]),
]
args = {
"include_package_data": True,
"exclude_package_data": {"": ["*.c"]},
}
setup(ext_modules=extensions, **args)
| data/lru-dict-1.3.0/setup.py | 137 | 86 | 416,002 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME unittest
STORE_NAME unittest
LOAD_CONST 0
LOAD_CONST ('StringIO',)
IMPORT_NAME gmplot.utility
IMPORT_FROM StringIO
STORE_NAME StringIO
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object StringIOTest at 0x7fab702a6660, file "f.py", line 5>
LOAD_CONST 'StringIOTest'
MAKE_FUNCTION
... | import unittest
from gmplot.utility import StringIO
class StringIOTest(unittest.TestCase):
def test_enter_exit(self):
with StringIO() as f:
f.write("Content")
self.assertEqual(f.getvalue(), "Content")
self.assertTrue(f.closed)
| data/gmplot-1.4.1/tests/test_utility.py | 306 | 86 | 352,870 |
LOAD_CONST 1
LOAD_CONST ('ray_triangle',)
IMPORT_NAME
IMPORT_FROM ray_triangle
STORE_NAME ray_triangle
POP_TOP
SETUP_EXCEPT to 34
LOAD_CONST 1
LOAD_CONST ('ray_pyembree',)
IMPORT_NAME
IMPORT_FROM ray_pyembree
STORE_NAME ray_pyembree
POP_TOP
LOAD_CONST True
STORE_NAME has_embree
POP_BLOCK
JUMP_FORWARD to 94
DUP_TOP
... | from . import ray_triangle
try:
from . import ray_pyembree
has_embree = True
except BaseException as E:
from .. import exceptions
ray_pyembree = exceptions.ExceptionWrapper(E)
has_embree = False
__all__ = ["ray_triangle", "ray_pyembree"]
| data/trimesh-4.1.4/trimesh/ray/__init__.py | 217 | 86 | 197,148 |
LOAD_CONST 1
LOAD_CONST ('Missing', 'ResultBackend')
IMPORT_NAME backend
IMPORT_FROM Missing
STORE_NAME Missing
IMPORT_FROM ResultBackend
STORE_NAME ResultBackend
POP_TOP
LOAD_CONST 1
LOAD_CONST ('ResultError', 'ResultFailure', 'ResultMissing', 'ResultTimeout')
IMPORT_NAME errors
IMPORT_FROM ResultError
STORE_NAME Res... | from .backend import Missing, ResultBackend
from .errors import ResultError, ResultFailure, ResultMissing, ResultTimeout
from .middleware import Results
__all__ = [
"Missing",
"ResultBackend",
"ResultError",
"ResultFailure",
"ResultTimeout",
"ResultMissing",
"Results",
]
| data/dramatiq-1.16.0/dramatiq/results/__init__.py | 180 | 86 | 145,483 |
LOAD_CONST '\nTest cases for testing if it is possible to pickle\nObjective-C objects\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('TestCase',)
IMPORT_NAME PyObjCTools.TestSupport
IMPORT_FROM TestCase
STORE_NAME TestCase
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object TestPickleObjC at 0x7fab4106cae0, file "f.py",... | """
Test cases for testing if it is possible to pickle
Objective-C objects
"""
from PyObjCTools.TestSupport import TestCase
class TestPickleObjC(TestCase):
pass
class TestPickleMixedGraph(TestCase):
pass
class TestPicklePythonNSObject(TestCase):
pass
| data/pyobjc-core-10.1/PyObjCTest/test_pickling_objc.py | 403 | 86 | 95,141 |
LOAD_CONST "Dataset definition for bccd.\n\nDEPRECATED!\nIf you want to use the BCCD dataset builder class, use:\ntfds.builder_cls('bccd')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_builder_import
POP_TOP
... | """Dataset definition for bccd.
DEPRECATED!
If you want to use the BCCD dataset builder class, use:
tfds.builder_cls('bccd')
"""
from tensorflow_datasets.core import lazy_builder_import
BCCD = lazy_builder_import.LazyBuilderImport("bccd")
| data/tensorflow-datasets-4.9.4/tensorflow_datasets/image/bccd/bccd.py | 130 | 86 | 255,280 |
LOAD_CONST 0
LOAD_CONST ('BaseModel',)
IMPORT_NAME langchain_core.pydantic_v1
IMPORT_FROM BaseModel
STORE_NAME BaseModel
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object AttributeInfo at 0x7faa755caf60, file "f.py", line 4>
LOAD_CONST 'AttributeInfo'
MAKE_FUNCTION
LOAD_CONST 'AttributeInfo'
LOAD_NAME BaseModel
CALL_FU... | from langchain_core.pydantic_v1 import BaseModel
class AttributeInfo(BaseModel):
"""Information about a data source attribute."""
name: str
description: str
type: str
class Config:
"""Configuration for this pydantic object."""
arbitrary_types_allowed = True
frozen = True... | data/langchain-0.1.8/langchain/chains/query_constructor/schema.py | 314 | 86 | 369,010 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME abc
STORE_NAME abc
LOAD_CONST 0
LOAD_CONST ('MetricsContext',)
IMPORT_NAME aws_embedded_metrics.logger.metrics_context
IMPORT_FROM MetricsContext
STORE_NAME MetricsContext
POP_TOP
LOAD_CONST 0
LOAD_CONST ('List',)
IMPORT_NAME typing
IMPORT_FROM List
STORE_NAME List
POP_TOP
LO... | import abc
from aws_embedded_metrics.logger.metrics_context import MetricsContext
from typing import List
class Serializer(abc.ABC):
@staticmethod
@abc.abstractmethod
def serialize(context: MetricsContext) -> List[str]:
"""Flushes the metrics context to the sink."""
| data/aws-embedded-metrics-3.2.0/aws_embedded_metrics/serializers/__init__.py | 266 | 86 | 367,098 |
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME azure.cli.core.aaz
IMPORT_STAR
LOAD_NAME register_command_group
LOAD_CONST 'relay wcfrelay authorization-rule'
CALL_FUNCTION
LOAD_BUILD_CLASS
LOAD_CONST <code object __CMDGroup at 0x7fab4130c8a0, file "f.py", line 4>
LOAD_CONST '__CMDGroup'
MAKE_FUNCTION
LOAD_CONST '__CMDGro... | from azure.cli.core.aaz import *
@register_command_group(
"relay wcfrelay authorization-rule",
)
class __CMDGroup(AAZCommandGroup):
"""Manage Azure Relay Service WCF Relay Authorization Rule."""
pass
__all__ = ["__CMDGroup"]
| data/azure-cli-2.57.0/azure/cli/command_modules/relay/aaz/latest/relay/wcfrelay/authorization_rule/__cmd_group.py | 196 | 86 | 375,643 |
LOAD_CONST 0
LOAD_CONST ('TestCase', 'min_os_level')
IMPORT_NAME PyObjCTools.TestSupport
IMPORT_FROM TestCase
STORE_NAME TestCase
IMPORT_FROM min_os_level
STORE_NAME min_os_level
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ContactsUI
STORE_NAME ContactsUI
LOAD_BUILD_CLASS
LOAD_CONST <code object TestCNContactPic... | from PyObjCTools.TestSupport import TestCase, min_os_level
import ContactsUI # noqa: F401
class TestCNContactPickerDelegate(TestCase):
@min_os_level("10.11")
def testProtocols(self):
self.assertProtocolExists("CNContactPickerDelegate")
| data/pyobjc-framework-ContactsUI-10.1/PyObjCTest/test_cncontactpickerdelegate.py | 286 | 86 | 354,949 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_reduce_countnonzero_int8_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_countnonzero_int8_64.py | 226 | 86 | 221,331 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME re
STORE_NAME re
LOAD_NAME re
LOAD_METHOD compile
LOAD_CONST '^arn:aws(?:-[a-z]+)?:dynamodb:(?P<region>[a-z\\d\\-]+-\\d):(?P<accountid>[\\d]{12}):table/(?P<id>[\\S]+)$'
CALL_METHOD
STORE_NAME TABLE_ARN_RE
LOAD_CONST None
RETURN_VALUE | import re
TABLE_ARN_RE = re.compile(
r"^arn:aws(?:-[a-z]+)?:dynamodb:(?P<region>[a-z\d\-]+-\d):(?P<accountid>[\d]{12}):table/(?P<id>[\S]+)$"
)
| data/compose_x_common-1.4.6/src/compose_x_common/aws/dynamodb.py | 106 | 86 | 345,344 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_reduce_argmin_uint64_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmin_uint64_64.py | 226 | 86 | 221,278 |
LOAD_CONST 0
LOAD_CONST ('Enum',)
IMPORT_NAME enum
IMPORT_FROM Enum
STORE_NAME Enum
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object AlertStatus at 0x7fab70027660, file "f.py", line 4>
LOAD_CONST 'AlertStatus'
MAKE_FUNCTION
LOAD_CONST 'AlertStatus'
LOAD_NAME str
LOAD_NAME Enum
CALL_FUNCTION
STORE_NAME AlertStatus
LOAD... | from enum import Enum
class AlertStatus(str, Enum):
Unknown = ("unknown",)
NewAlert = ("newAlert",)
InProgress = ("inProgress",)
Resolved = ("resolved",)
Dismissed = ("dismissed",)
UnknownFutureValue = ("unknownFutureValue",)
| data/msgraph-sdk-1.1.0/msgraph/generated/models/alert_status.py | 202 | 86 | 261,913 |
LOAD_CONST 0
LOAD_CONST ('DownloadFactory',)
IMPORT_NAME pyomo.common.download
IMPORT_FROM DownloadFactory
STORE_NAME DownloadFactory
POP_TOP
LOAD_CONST 0
LOAD_CONST ('ExtensionBuilderFactory',)
IMPORT_NAME pyomo.common.extensions
IMPORT_FROM ExtensionBuilderFactory
STORE_NAME ExtensionBuilderFactory
POP_TOP
LOAD_CON... | from pyomo.common.download import DownloadFactory
from pyomo.common.extensions import ExtensionBuilderFactory
from .getMCPP import get_mcpp
from .build import MCPPBuilder
def load():
DownloadFactory.register("mcpp")(get_mcpp)
ExtensionBuilderFactory.register("mcpp")(MCPPBuilder)
| data/Pyomo-6.7.0/pyomo/contrib/mcpp/plugins.py | 234 | 86 | 23,652 |
LOAD_CONST "Dataset definition for mctaco.\n\nDEPRECATED!\nIf you want to use the Mctaco dataset builder class, use:\ntfds.builder_cls('mctaco')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_builder_import
POP... | """Dataset definition for mctaco.
DEPRECATED!
If you want to use the Mctaco dataset builder class, use:
tfds.builder_cls('mctaco')
"""
from tensorflow_datasets.core import lazy_builder_import
Mctaco = lazy_builder_import.LazyBuilderImport("mctaco")
| data/tensorflow-datasets-4.9.4/tensorflow_datasets/question_answering/mctaco.py | 130 | 86 | 255,294 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_IndexedArrayU32_numnull_parents_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_IndexedArrayU32_numnull_parents.py | 225 | 86 | 221,467 |
LOAD_CONST "Dataset definition for siscore.\n\nDEPRECATED!\nIf you want to use the Siscore dataset builder class, use:\ntfds.builder_cls('siscore')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_builder_import
... | """Dataset definition for siscore.
DEPRECATED!
If you want to use the Siscore dataset builder class, use:
tfds.builder_cls('siscore')
"""
from tensorflow_datasets.core import lazy_builder_import
Siscore = lazy_builder_import.LazyBuilderImport("siscore")
| data/tensorflow-datasets-4.9.4/tensorflow_datasets/image_classification/siscore/siscore.py | 130 | 86 | 255,475 |
LOAD_CONST 0
LOAD_CONST ('annotations',)
IMPORT_NAME __future__
IMPORT_FROM annotations
STORE_NAME annotations
POP_TOP
LOAD_CONST 0
LOAD_CONST ('NotebookImportExtractor',)
IMPORT_NAME deptry.imports.extractors.notebook_import_extractor
IMPORT_FROM NotebookImportExtractor
STORE_NAME NotebookImportExtractor
POP_TOP
LOA... | from __future__ import annotations
from deptry.imports.extractors.notebook_import_extractor import NotebookImportExtractor
from deptry.imports.extractors.python_import_extractor import PythonImportExtractor
__all__ = (
"NotebookImportExtractor",
"PythonImportExtractor",
)
| data/deptry-0.12.0/deptry/imports/extractors/__init__.py | 150 | 86 | 45,206 |
LOAD_CONST 0
LOAD_CONST ('CallbackReader',)
IMPORT_NAME pysmi.reader.callback
IMPORT_FROM CallbackReader
STORE_NAME CallbackReader
POP_TOP
LOAD_CONST 0
LOAD_CONST ('FtpReader',)
IMPORT_NAME pysmi.reader.ftpclient
IMPORT_FROM FtpReader
STORE_NAME FtpReader
POP_TOP
LOAD_CONST 0
LOAD_CONST ('HttpReader',)
IMPORT_NAME py... | from pysmi.reader.callback import CallbackReader
from pysmi.reader.ftpclient import FtpReader
from pysmi.reader.httpclient import HttpReader
from pysmi.reader.zipreader import ZipReader
from pysmi.reader.localfile import FileReader
from pysmi.reader.url import getReadersFromUrls
| data/pysmi-0.3.4/pysmi/reader/__init__.py | 216 | 86 | 155,038 |
LOAD_CONST 0
LOAD_CONST ('App', 'ComposeResult')
IMPORT_NAME textual.app
IMPORT_FROM App
STORE_NAME App
IMPORT_FROM ComposeResult
STORE_NAME ComposeResult
POP_TOP
LOAD_CONST 0
LOAD_CONST ('LoadingIndicator',)
IMPORT_NAME textual.widgets
IMPORT_FROM LoadingIndicator
STORE_NAME LoadingIndicator
POP_TOP
LOAD_BUILD_CLASS... | from textual.app import App, ComposeResult
from textual.widgets import LoadingIndicator
class LoadingApp(App):
def compose(self) -> ComposeResult:
yield LoadingIndicator()
if __name__ == "__main__":
app = LoadingApp()
app.run()
| data/textual-0.52.1/docs/examples/widgets/loading_indicator.py | 289 | 86 | 26,301 |
LOAD_CONST 0
LOAD_CONST ('ClientValue',)
IMPORT_NAME office365.runtime.client_value
IMPORT_FROM ClientValue
STORE_NAME ClientValue
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object Certification at 0x7fab700ad930, file "f.py", line 4>
LOAD_CONST 'Certification'
MAKE_FUNCTION
LOAD_CONST 'Certification'
LOAD_NAME ClientV... | from office365.runtime.client_value import ClientValue
class Certification(ClientValue):
"""Represents the certification details of an application."""
def __init__(self, certification_details_url=None):
"""
:param str certification_details_url:
"""
self.certificationDetailsUrl... | data/Office365-REST-Python-Client-2.5.5/office365/directory/certificates/certification.py | 220 | 86 | 188,871 |
LOAD_CONST 0
LOAD_CONST ('exceptions',)
IMPORT_NAME keystonemiddleware
IMPORT_FROM exceptions
STORE_NAME exceptions
POP_TOP
LOAD_NAME exceptions
LOAD_ATTR ConfigurationError
STORE_NAME ConfigurationError
LOAD_BUILD_CLASS
LOAD_CONST <code object InvalidToken at 0x7f8e2fe3cb70, file "f.py", line 7>
LOAD_CONST 'InvalidT... | from keystonemiddleware import exceptions
ConfigurationError = exceptions.ConfigurationError
class InvalidToken(exceptions.KeystoneMiddlewareException):
pass
class ServiceError(exceptions.KeystoneMiddlewareException):
pass
class RevocationListError(exceptions.KeystoneMiddlewareException):
pass
| data/keystonemiddleware-10.5.0/keystonemiddleware/auth_token/_exceptions.py | 346 | 86 | 397,969 |
LOAD_CONST './panoptic-fpn_r50_fpn_ms-3x_coco.py'
STORE_NAME _base_
LOAD_NAME dict
LOAD_NAME dict
LOAD_CONST 101
LOAD_NAME dict
LOAD_CONST 'Pretrained'
LOAD_CONST 'torchvision://resnet101'
LOAD_CONST ('type', 'checkpoint')
CALL_FUNCTION
LOAD_CONST ('depth', 'init_cfg')
CALL_FUNCTION
LOAD_CONST ('backbone',)
CALL_FU... | _base_ = "./panoptic-fpn_r50_fpn_ms-3x_coco.py"
model = dict(
backbone=dict(
depth=101,
init_cfg=dict(type="Pretrained", checkpoint="torchvision://resnet101"),
)
)
| data/mmdet-3.3.0/mmdet/.mim/configs/panoptic_fpn/panoptic-fpn_r101_fpn_ms-3x_coco.py | 111 | 86 | 151,212 |
LOAD_CONST '\nrequests_kerberos.exceptions\n~~~~~~~~~~~~~~~~~~~\n\nThis module contains the set of exceptions.\n\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('RequestException',)
IMPORT_NAME requests.exceptions
IMPORT_FROM RequestException
STORE_NAME RequestException
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object... | """
requests_kerberos.exceptions
~~~~~~~~~~~~~~~~~~~
This module contains the set of exceptions.
"""
from requests.exceptions import RequestException
class MutualAuthenticationError(RequestException):
"""Mutual Authentication Error"""
class KerberosExchangeError(RequestException):
"""Kerberos Exchange Fai... | data/pywinrm-0.4.3/winrm/vendor/requests_kerberos/exceptions.py | 315 | 86 | 269,488 |
LOAD_CONST 0
LOAD_CONST ('TestCase', 'main')
IMPORT_NAME unittest
IMPORT_FROM TestCase
STORE_NAME TestCase
IMPORT_FROM main
STORE_NAME main
POP_TOP
LOAD_CONST 0
LOAD_CONST ('ApiNinjas',)
IMPORT_NAME random_word
IMPORT_FROM ApiNinjas
STORE_NAME ApiNinjas
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object TestRandomWord ... | from unittest import TestCase, main
from random_word import ApiNinjas
class TestRandomWord(TestCase):
r = ApiNinjas()
def test_random_word(self):
return self.r.get_random_word()
if __name__ == "__main__":
main()
| data/Random Word-1.0.11/random_word/tests/test_apininjas.py | 286 | 86 | 134,980 |
LOAD_CONST "A lil' TOML parser."
STORE_NAME __doc__
LOAD_CONST ('loads', 'load', 'TOMLDecodeError')
STORE_NAME __all__
LOAD_CONST '1.2.0'
STORE_NAME __version__
LOAD_CONST 1
LOAD_CONST ('TOMLDecodeError', 'load', 'loads')
IMPORT_NAME _parser
IMPORT_FROM TOMLDecodeError
STORE_NAME TOMLDecodeError
IMPORT_FROM load
STO... | """A lil' TOML parser."""
__all__ = ("loads", "load", "TOMLDecodeError")
__version__ = "1.2.0" # DO NOT EDIT THIS LINE MANUALLY. LET bump2version UTILITY DO IT
from ._parser import TOMLDecodeError, load, loads
| data/isort-5.13.2/isort/_vendored/tomli/__init__.py | 116 | 86 | 105,664 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME os
STORE_NAME os
LOAD_CONST 0
LOAD_CONST ('Dict', 'List', 'Tuple', 'TypeVar', 'Union')
IMPORT_NAME typing
IMPORT_FROM Dict
STORE_NAME Dict
IMPORT_FROM List
STORE_NAME List
IMPORT_FROM Tuple
STORE_NAME Tuple
IMPORT_FROM TypeVar
STORE_NAME TypeVar
IMPORT_FROM Union
STORE_NAME Uni... | import os
from typing import Dict, List, Tuple, TypeVar, Union
T = TypeVar("T")
ListLike = Union[List[T], Tuple[T, ...]]
NestedDataStructureLike = Union[T, List[T], Dict[str, T]]
PathLike = Union[str, bytes, os.PathLike]
| data/datasets-2.17.1/src/datasets/utils/typing.py | 202 | 86 | 52,869 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('Check',)
IMPORT_NAME frictionless
IMPORT_FROM Check
STORE_NAME Check
POP_TOP
LOAD_CONST <code object test_check_from_descriptor_type_framework_v4 at 0x7fab429cdae0, file "f.py", line 6>
LOAD_CONST 'test_check_from_descriptor_t... | import pytest
from frictionless import Check
def test_check_from_descriptor_type_framework_v4():
with pytest.warns(UserWarning):
check = Check.from_descriptor({"code": "table-dimensions"})
assert check.to_descriptor() == {"type": "table-dimensions"}
| data/frictionless-5.16.1/tests/checklist/check/test_convert.py | 236 | 86 | 41,479 |
LOAD_CONST "Dataset definition for titanic.\n\nDEPRECATED!\nIf you want to use the Titanic dataset builder class, use:\ntfds.builder_cls('titanic')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_builder_import
... | """Dataset definition for titanic.
DEPRECATED!
If you want to use the Titanic dataset builder class, use:
tfds.builder_cls('titanic')
"""
from tensorflow_datasets.core import lazy_builder_import
Titanic = lazy_builder_import.LazyBuilderImport("titanic")
| data/tensorflow-datasets-4.9.4/tensorflow_datasets/structured/titanic.py | 130 | 86 | 255,197 |
LOAD_CONST 0
LOAD_CONST ('Any', 'MutableMapping')
IMPORT_NAME typing
IMPORT_FROM Any
STORE_NAME Any
IMPORT_FROM MutableMapping
STORE_NAME MutableMapping
POP_TOP
SETUP_EXCEPT to 34
LOAD_CONST 0
LOAD_CONST ('Context',)
IMPORT_NAME airflow.utils.context
IMPORT_FROM Context
STORE_NAME Context
POP_TOP
POP_BLOCK
JUMP_FORWA... | from typing import Any, MutableMapping
try:
from airflow.utils.context import Context
except ModuleNotFoundError:
class Context(MutableMapping[str, Any]): # type: ignore[no-redef]
"""Placeholder typing class for ``airflow.utils.context.Context``."""
__all__ = ["Context"]
| data/astronomer-providers-1.19.0/astronomer/providers/utils/typing_compat.py | 244 | 86 | 252,524 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME hubspot.cms.url_redirects
IMPORT_FROM cms
ROT_TWO
POP_TOP
IMPORT_FROM url_redirects
STORE_NAME api_client
POP_TOP
LOAD_CONST 3
LOAD_CONST ('DiscoveryBase',)
IMPORT_NAME discovery_base
IMPORT_FROM DiscoveryBase
STORE_NAME DiscoveryBase
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code ... | import hubspot.cms.url_redirects as api_client
from ...discovery_base import DiscoveryBase
class Discovery(DiscoveryBase):
@property
def redirects_api(self) -> api_client.RedirectsApi:
return self._configure_api_client(api_client, "RedirectsApi")
| data/hubspot-api-client-8.2.1/hubspot/discovery/cms/url_redirects/discovery.py | 276 | 86 | 320,806 |
LOAD_CONST 1
LOAD_CONST ('check_kwargs',)
IMPORT_NAME check_kwargs
IMPORT_FROM check_kwargs
STORE_NAME check_kwargs
POP_TOP
LOAD_CONST 1
LOAD_CONST ('__version__',)
IMPORT_NAME version
IMPORT_FROM __version__
STORE_NAME __version__
POP_TOP
LOAD_CONST 1
LOAD_CONST ('get_user_agent',)
IMPORT_NAME user_agent
IMPORT_FROM... | from .check_kwargs import check_kwargs
from .version import __version__
from .user_agent import get_user_agent
from .deprecation_notice import warn_deprecated
from .fix_tuple_length import fix_tuple_length
from .convert_to_list import convert_to_list
from .normalize_host import normalize_host
| data/pinecone_client-3.0.3/pinecone/utils/__init__.py | 245 | 86 | 251,330 |
SETUP_ANNOTATIONS
LOAD_CONST 'Constants for the mütesync integration.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('timedelta',)
IMPORT_NAME datetime
IMPORT_FROM timedelta
STORE_NAME timedelta
POP_TOP
LOAD_CONST 0
LOAD_CONST ('Final',)
IMPORT_NAME typing
IMPORT_FROM Final
STORE_NAME Final
POP_TOP
LOAD_CONST 'mutesyn... | """Constants for the mütesync integration."""
from datetime import timedelta
from typing import Final
DOMAIN: Final = "mutesync"
UPDATE_INTERVAL_NOT_IN_MEETING: Final = timedelta(seconds=10)
UPDATE_INTERVAL_IN_MEETING: Final = timedelta(seconds=10)
| data/homeassistant-2024.2.2/homeassistant/components/mutesync/const.py | 219 | 86 | 295,976 |
LOAD_CONST 0
LOAD_CONST ('profile',)
IMPORT_NAME memory_profiler
IMPORT_FROM profile
STORE_NAME profile
POP_TOP
LOAD_NAME profile
LOAD_CONST <code object test_with_profile at 0x7faa74ae3ed0, file "f.py", line 4>
LOAD_CONST 'test_with_profile'
MAKE_FUNCTION
CALL_FUNCTION
STORE_NAME test_with_profile
LOAD_NAME __name__... | from memory_profiler import profile
@profile
def test_with_profile(arg1):
"""dummy doc"""
return None
if __name__ == "__main__":
assert test_with_profile.__doc__ == "dummy doc"
assert test_with_profile.__name__ == "test_with_profile"
| data/memory_profiler-0.61.0/test/test_attributes.py | 178 | 86 | 249,400 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_reduce_sum_bool_int8_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_sum_bool_int8_64.py | 225 | 86 | 221,603 |
LOAD_CONST "Dataset definition for penguins.\n\nDEPRECATED!\nIf you want to use the Penguins dataset builder class, use:\ntfds.builder_cls('penguins')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_builder_impo... | """Dataset definition for penguins.
DEPRECATED!
If you want to use the Penguins dataset builder class, use:
tfds.builder_cls('penguins')
"""
from tensorflow_datasets.core import lazy_builder_import
Penguins = lazy_builder_import.LazyBuilderImport("penguins")
| data/tensorflow-datasets-4.9.4/tensorflow_datasets/structured/penguins/penguins.py | 130 | 86 | 255,208 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_reduce_argmax_int16_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmax_int16_64.py | 225 | 86 | 221,461 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME json
STORE_NAME json
LOAD_CONST 0
LOAD_CONST ('tool',)
IMPORT_NAME promptflow
IMPORT_FROM tool
STORE_NAME tool
POP_TOP
LOAD_NAME tool
LOAD_NAME str
LOAD_CONST ('input_str',)
BUILD_CONST_KEY_MAP
LOAD_CONST <code object convert_to_dict at 0x7f8ab63eaed0, file "f.py", line 6>
LO... | import json
from promptflow import tool
@tool
def convert_to_dict(input_str: str):
try:
return json.loads(input_str)
except Exception as e:
print("input is not valid, error: {}".format(e))
return {"category": "None", "evidence": "None"}
| data/azure-ai-ml-1.13.0/tests/test_configs/flows/additional_includes/convert_to_dict.py | 239 | 86 | 424,218 |
LOAD_CONST 0
LOAD_CONST ('E2BDataAnalysisTool', 'E2BDataAnalysisToolArguments', 'UploadedFile')
IMPORT_NAME langchain_community.tools.e2b_data_analysis.tool
IMPORT_FROM E2BDataAnalysisTool
STORE_NAME E2BDataAnalysisTool
IMPORT_FROM E2BDataAnalysisToolArguments
STORE_NAME E2BDataAnalysisToolArguments
IMPORT_FROM Uploade... | from langchain_community.tools.e2b_data_analysis.tool import (
E2BDataAnalysisTool,
E2BDataAnalysisToolArguments,
UploadedFile,
)
__all__ = [
"UploadedFile",
"E2BDataAnalysisToolArguments",
"E2BDataAnalysisTool",
]
| data/langchain-0.1.8/langchain/tools/e2b_data_analysis/tool.py | 137 | 86 | 369,342 |
LOAD_CONST 0
LOAD_CONST ('TType', 'TMessageType', 'TFrozenDict', 'TException', 'TApplicationException')
IMPORT_NAME thrift.Thrift
IMPORT_FROM TType
STORE_NAME TType
IMPORT_FROM TMessageType
STORE_NAME TMessageType
IMPORT_FROM TFrozenDict
STORE_NAME TFrozenDict
IMPORT_FROM TException
STORE_NAME TException
IMPORT_FROM TA... | from thrift.Thrift import (
TType,
TMessageType,
TFrozenDict,
TException,
TApplicationException,
)
from thrift.protocol.TProtocol import TProtocolException
import sys
import jaeger.ttypes
import zipkincore.ttypes
from thrift.transport import TTransport
| data/opentelemetry_exporter_jaeger_thrift-1.21.0/src/opentelemetry/exporter/jaeger/thrift/gen/agent/ttypes.py | 216 | 86 | 246,316 |
LOAD_CONST 0
LOAD_CONST ('unique',)
IMPORT_NAME enum
IMPORT_FROM unique
STORE_NAME unique
POP_TOP
LOAD_CONST 4
LOAD_CONST ('StrEnum',)
IMPORT_NAME _base_enum
IMPORT_FROM StrEnum
STORE_NAME StrEnum
POP_TOP
LOAD_NAME unique
LOAD_BUILD_CLASS
LOAD_CONST <code object IndexConvexityAdjustmentType at 0x7fab701ac780, file "f... | from enum import unique
from ...._base_enum import StrEnum
@unique
class IndexConvexityAdjustmentType(StrEnum):
NONE = "None"
BLACK_SCHOLES = "BlackScholes"
REPLICATION = "Replication"
LIBOR_SWAP_METHOD = "LiborSwapMethod"
| data/refinitiv-data-1.6.0/refinitiv/data/content/ipa/_enums/_index_convexity_adjustment_type.py | 236 | 86 | 195,529 |
LOAD_CONST 0
LOAD_CONST ('Optional',)
IMPORT_NAME typing
IMPORT_FROM Optional
STORE_NAME Optional
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME typer
STORE_NAME typer
LOAD_NAME typer
LOAD_ATTR Argument
LOAD_CONST None
LOAD_CONST ('default',)
CALL_FUNCTION
BUILD_TUPLE
LOAD_NAME Optional
LOAD_NAME str
BINARY_SUBSCR
... | from typing import Optional
import typer
def main(name: Optional[str] = typer.Argument(default=None)):
if name is None:
print("Hello World!")
else:
print(f"Hello {name}")
if __name__ == "__main__":
typer.run(main)
| data/typer-0.9.0/docs_src/arguments/optional/tutorial002.py | 212 | 86 | 224,543 |
LOAD_CONST 0
LOAD_CONST ('annotations',)
IMPORT_NAME __future__
IMPORT_FROM annotations
STORE_NAME annotations
POP_TOP
LOAD_CONST 0
LOAD_CONST ('TYPE_CHECKING',)
IMPORT_NAME typing
IMPORT_FROM TYPE_CHECKING
STORE_NAME TYPE_CHECKING
POP_TOP
LOAD_CONST 0
LOAD_CONST ('ArgumentParser',)
IMPORT_NAME argparse
IMPORT_FROM A... | from __future__ import annotations
from typing import TYPE_CHECKING
from argparse import ArgumentParser
from . import completions
if TYPE_CHECKING:
from argparse import _SubParsersAction
def register(subparser: _SubParsersAction[ArgumentParser]) -> None:
completions.register(subparser)
| data/openai-1.12.0/src/openai/cli/_api/chat/__init__.py | 255 | 86 | 63,612 |
LOAD_CONST "Dataset definition for bigearthnet.\n\nDEPRECATED!\nIf you want to use the Bigearthnet dataset builder class, use:\ntfds.builder_cls('bigearthnet')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_bui... | """Dataset definition for bigearthnet.
DEPRECATED!
If you want to use the Bigearthnet dataset builder class, use:
tfds.builder_cls('bigearthnet')
"""
from tensorflow_datasets.core import lazy_builder_import
Bigearthnet = lazy_builder_import.LazyBuilderImport("bigearthnet")
| data/tfds-nightly-4.9.4.dev202402210044/tensorflow_datasets/image_classification/bigearthnet.py | 130 | 86 | 238,447 |
LOAD_CONST 4
STORE_NAME param
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pandas
STORE_NAME pd
LOAD_NAME pd
LOAD_ATTR DataFrame
LOAD_CONST 1
LOAD_CONST 2
BUILD_LIST
LOAD_CONST 3
LOAD_NAME param
BINARY_ADD
LOAD_CONST 4
BUILD_LIST
LOAD_CONST ('A', 'B')
BUILD_CONST_KEY_MAP
LOAD_NAME pd
LOAD_ATTR Index
LOAD_CONST 'x0'
LOAD... | param = 4
""
import pandas as pd
""
df = pd.DataFrame(
{"A": [1, 2], "B": [3 + param, 4]}, index=pd.Index(["x0", "x1"], name="x")
)
df
""
df.plot(kind="bar")
| data/jupytext-1.16.1/tests/data/notebooks/outputs/ipynb_to_sphinx/nteract_with_parameter.py | 146 | 86 | 312,595 |
LOAD_CONST 0
LOAD_CONST ('pass_environment',)
IMPORT_NAME complex.cli
IMPORT_FROM pass_environment
STORE_NAME pass_environment
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME click
STORE_NAME click
LOAD_NAME click
LOAD_ATTR command
LOAD_CONST 'status'
LOAD_CONST 'Shows file changes.'
LOAD_CONST ('short_help',)
CALL_... | from complex.cli import pass_environment
import click
@click.command("status", short_help="Shows file changes.")
@pass_environment
def cli(ctx):
"""Shows file changes in the current working directory."""
ctx.log("Changed files: none")
ctx.vlog("bla bla bla, debug info")
| data/click-8.1.7/examples/complex/complex/commands/cmd_status.py | 180 | 86 | 117,111 |
LOAD_CONST 1
LOAD_CONST ('ChangeTrueDivToMulByInverse', 'FuseBatchNorm1dInLinear', 'FuseBatchNorm2dInConv2d', 'FuseBiasInLinear', 'MergeLinears', 'ReversibleTransformation', 'Transformation', 'compose')
IMPORT_NAME transformations
IMPORT_FROM ChangeTrueDivToMulByInverse
STORE_NAME ChangeTrueDivToMulByInverse
IMPORT_FRO... | from .transformations import ( # noqa
ChangeTrueDivToMulByInverse,
FuseBatchNorm1dInLinear,
FuseBatchNorm2dInConv2d,
FuseBiasInLinear,
MergeLinears,
ReversibleTransformation,
Transformation,
compose,
)
| data/optimum-1.17.1/optimum/fx/optimization/__init__.py | 205 | 86 | 150,301 |
LOAD_CONST "Dataset definition for asqa.\n\nDEPRECATED!\nIf you want to use the Asqa dataset builder class, use:\ntfds.builder_cls('asqa')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_builder_import
POP_TOP
... | """Dataset definition for asqa.
DEPRECATED!
If you want to use the Asqa dataset builder class, use:
tfds.builder_cls('asqa')
"""
from tensorflow_datasets.core import lazy_builder_import
Asqa = lazy_builder_import.LazyBuilderImport("asqa")
| data/tensorflow-datasets-4.9.4/tensorflow_datasets/question_answering/asqa/asqa.py | 130 | 86 | 255,315 |
LOAD_CONST 0
LOAD_CONST ('annotations',)
IMPORT_NAME __future__
IMPORT_FROM annotations
STORE_NAME annotations
POP_TOP
LOAD_CONST 0
LOAD_CONST ('public',)
IMPORT_NAME public
IMPORT_FROM public
STORE_NAME public
POP_TOP
LOAD_CONST 0
LOAD_CONST ('Column', 'Scalar', 'Value')
IMPORT_NAME ibis.expr.types.generic
IMPORT_FR... | from __future__ import annotations
from public import public
from ibis.expr.types.generic import Column, Scalar, Value
@public
class UUIDValue(Value):
pass
@public
class UUIDScalar(Scalar, UUIDValue):
pass
@public
class UUIDColumn(Column, UUIDValue):
pass
| data/ibis_framework-8.0.0/ibis/expr/types/uuid.py | 400 | 86 | 161,642 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME subprocess
STORE_NAME subprocess
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME argparse
STORE_NAME argparse
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME sys
STORE_NAME sys
LOAD_NAME __name__
LOAD_CONST '__main__'
COMPARE_OP ==
POP_JUMP_IF_FALSE
LOAD_NAME argparse
LOAD_METHOD Argument... | import subprocess
import argparse
import sys
if __name__ == "__main__":
parser = argparse.ArgumentParser()
parser.add_argument("prog")
args = parser.parse_args()
res = subprocess.run(args.prog)
sys.exit(res.returncode - 42)
| data/meson-1.3.2/test cases/windows/13 test argument extra paths/test/test_run_exe.py | 161 | 86 | 81,935 |
LOAD_CONST 'DO NOT EDIT.\n\nThis file was autogenerated. Do not edit it by hand,\nsince your modifications would be overwritten.\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('decode_predictions',)
IMPORT_NAME keras_core.src.applications.mobilenet_v3
IMPORT_FROM decode_predictions
STORE_NAME decode_predictions
POP_TO... | """DO NOT EDIT.
This file was autogenerated. Do not edit it by hand,
since your modifications would be overwritten.
"""
from keras_core.src.applications.mobilenet_v3 import decode_predictions
from keras_core.src.applications.mobilenet_v3 import preprocess_input
| data/keras-core-0.1.7/keras_core/applications/mobilenet_v3/__init__.py | 143 | 86 | 385,621 |
LOAD_CONST './cspnext-s_8xb256-rsb-a1-600e_in1k.py'
STORE_NAME _base_
LOAD_NAME dict
LOAD_NAME dict
LOAD_CONST 0.67
LOAD_CONST 0.75
LOAD_CONST ('deepen_factor', 'widen_factor')
CALL_FUNCTION
LOAD_NAME dict
LOAD_CONST 768
LOAD_CONST ('in_channels',)
CALL_FUNCTION
LOAD_CONST ('backbone', 'head')
CALL_FUNCTION
STORE_NAM... | _base_ = "./cspnext-s_8xb256-rsb-a1-600e_in1k.py"
model = dict(
backbone=dict(deepen_factor=0.67, widen_factor=0.75), head=dict(in_channels=768)
)
| data/mmdet-3.3.0/mmdet/.mim/configs/rtmdet/classification/cspnext-m_8xb256-rsb-a1-600e_in1k.py | 115 | 86 | 151,261 |
LOAD_CONST 0
LOAD_CONST ('Enum',)
IMPORT_NAME enum
IMPORT_FROM Enum
STORE_NAME Enum
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object IncidentStatus at 0x7fab70027420, file "f.py", line 4>
LOAD_CONST 'IncidentStatus'
MAKE_FUNCTION
LOAD_CONST 'IncidentStatus'
LOAD_NAME str
LOAD_NAME Enum
CALL_FUNCTION
STORE_NAME Inciden... | from enum import Enum
class IncidentStatus(str, Enum):
Active = ("active",)
Resolved = ("resolved",)
InProgress = ("inProgress",)
Redirected = ("redirected",)
UnknownFutureValue = ("unknownFutureValue",)
AwaitingAction = ("awaitingAction",)
| data/msgraph-sdk-1.1.0/msgraph/generated/models/security/incident_status.py | 205 | 86 | 262,105 |
LOAD_CONST 0
LOAD_CONST ('annotations',)
IMPORT_NAME __future__
IMPORT_FROM annotations
STORE_NAME annotations
POP_TOP
LOAD_CONST 0
LOAD_CONST ('public',)
IMPORT_NAME public
IMPORT_FROM public
STORE_NAME public
POP_TOP
LOAD_CONST 0
LOAD_CONST ('Column', 'Scalar', 'Value')
IMPORT_NAME ibis.expr.types.generic
IMPORT_FR... | from __future__ import annotations
from public import public
from ibis.expr.types.generic import Column, Scalar, Value
@public
class SetValue(Value):
pass
@public
class SetScalar(Scalar, SetValue):
pass
@public
class SetColumn(Column, SetValue):
pass
| data/ibis_framework-8.0.0/ibis/expr/types/collections.py | 399 | 86 | 161,643 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_reduce_argmin_int32_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmin_int32_64.py | 226 | 86 | 221,592 |
LOAD_CONST 0
LOAD_CONST ('SparkSession',)
IMPORT_NAME pyspark.sql
IMPORT_FROM SparkSession
STORE_NAME SparkSession
POP_TOP
LOAD_CONST 1
LOAD_CONST ('MosaicDataFrameReader',)
IMPORT_NAME mosaic_data_frame_reader
IMPORT_FROM MosaicDataFrameReader
STORE_NAME MosaicDataFrameReader
POP_TOP
LOAD_CONST <code object read at ... | from pyspark.sql import SparkSession
from .mosaic_data_frame_reader import MosaicDataFrameReader
def read():
"""
Returns a MosaicDataFrameReader for reading MosaicDataFrames.
"""
spark = SparkSession.builder.getOrCreate()
return MosaicDataFrameReader()
| data/databricks-mosaic-0.4.0/mosaic/readers/readers.py | 156 | 86 | 52,977 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME re
STORE_NAME re
LOAD_CONST 0
LOAD_CONST ('join',)
IMPORT_NAME os.path
IMPORT_FROM join
STORE_NAME join
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME setuptools
STORE_NAME setuptools
LOAD_NAME setuptools
LOAD_ATTR setup
LOAD_NAME re
LOAD_METHOD search
LOAD_CONST "__versi... | import re
from os.path import join
import setuptools
setuptools.setup(
version=re.search(
"__version__ = '([^']+)'", open(join("mock", "__init__.py")).read()
).group(1),
long_description=open("README.rst").read(),
)
| data/mock-5.1.0/setup.py | 163 | 86 | 427,474 |
LOAD_CONST 0
LOAD_CONST ('print_function', 'division', 'absolute_import', 'annotations')
IMPORT_NAME __future__
IMPORT_FROM print_function
STORE_NAME print_function
IMPORT_FROM division
STORE_NAME division
IMPORT_FROM absolute_import
STORE_NAME absolute_import
IMPORT_FROM annotations
STORE_NAME annotations
POP_TOP
LOA... | from __future__ import print_function, division, absolute_import, annotations
from ..message import BackendMessage
class EndOfBatchResponse(BackendMessage):
message_id = b"J"
def __init__(self, data):
BackendMessage.__init__(self)
BackendMessage.register(EndOfBatchResponse)
| data/vertica-python-1.3.8/vertica_python/vertica/messages/backend_messages/end_of_batch_response.py | 304 | 86 | 287,015 |
LOAD_CONST 0
LOAD_CONST ('re_path',)
IMPORT_NAME django.urls
IMPORT_FROM re_path
STORE_NAME re_path
POP_TOP
LOAD_CONST 0
LOAD_CONST ('include',)
IMPORT_NAME django.conf.urls
IMPORT_FROM include
STORE_NAME include
POP_TOP
LOAD_CONST 0
LOAD_CONST ('admin',)
IMPORT_NAME grappelli.tests
IMPORT_FROM admin
STORE_NAME admin... | from django.urls import re_path
from django.conf.urls import include
from grappelli.tests import admin
urlpatterns = [
re_path(r"^admin/", admin.site.urls),
re_path(r"^grappelli/", include("grappelli.urls")),
]
| data/django-grappelli-3.0.8/grappelli/tests/urls.py | 154 | 86 | 20,649 |
LOAD_CONST 'Constants for QVR Pro component.'
STORE_NAME __doc__
LOAD_CONST 'qvr_pro'
STORE_NAME DOMAIN
LOAD_CONST 'QVR'
STORE_NAME SHORT_NAME
LOAD_CONST 'exclude_channels'
STORE_NAME CONF_EXCLUDE_CHANNELS
LOAD_CONST 'stop_record'
STORE_NAME SERVICE_STOP_RECORD
LOAD_CONST 'start_record'
STORE_NAME SERVICE_START_RE... | """Constants for QVR Pro component."""
DOMAIN = "qvr_pro"
SHORT_NAME = "QVR"
CONF_EXCLUDE_CHANNELS = "exclude_channels"
SERVICE_STOP_RECORD = "stop_record"
SERVICE_START_RECORD = "start_record"
| data/homeassistant-2024.2.2/homeassistant/components/qvr_pro/const.py | 108 | 86 | 406,799 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME uwsgi
STORE_NAME uwsgi
LOAD_CONST 5
STORE_NAME sleepvalue
LOAD_CONST <code object application at 0x7f8e44ccfc00, file "f.py", line 6>
LOAD_CONST 'application'
MAKE_FUNCTION
STORE_NAME application
LOAD_CONST None
RETURN_VALUE
LOAD_FAST start_response
LOAD_CONST '200 Ok'
LOAD_C... | import uwsgi
sleepvalue = 5
def application(env, start_response):
start_response("200 Ok", [("Content-type", "text/html")])
yield uwsgi.async_sleep(sleepvalue)
yield "<h1>Hello World after %d seconds</h1>" % sleepvalue
| data/uwsgi-2.0.24/tests/sleeping_async.py | 164 | 86 | 398,187 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.