input stringlengths 28 198k | output stringlengths 3 71k | file stringlengths 19 330 | input_tokens int64 5 159k | output_tokens int64 3 9.07k | __index_level_0__ int64 2 449k |
|---|---|---|---|---|---|
LOAD_CONST 1
LOAD_CONST ('Element',)
IMPORT_NAME element
IMPORT_FROM Element
STORE_NAME Element
POP_TOP
LOAD_CONST 1
LOAD_CONST ('Envelope', 'GroupEnvelope', 'InterchangeEnvelope', 'TransactionSetEnvelope')
IMPORT_NAME envelope
IMPORT_FROM Envelope
STORE_NAME Envelope
IMPORT_FROM GroupEnvelope
STORE_NAME GroupEnvelope... | from .element import Element
from .envelope import (
Envelope,
GroupEnvelope,
InterchangeEnvelope,
TransactionSetEnvelope,
)
from .errors import FieldValidationError, IDMismatchError, SegmentCountError
from .interchange import Interchange, InterchangeHeader, InterchangeTrailer
from .segment import Segme... | data/badx12-0.2.2/badx12/utils/__init__.py | 249 | 86 | 136,915 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME Naked.settings
STORE_NAME Naked
LOAD_CONST 0
LOAD_CONST ('exit_success',)
IMPORT_NAME Naked.toolshed.system
IMPORT_FROM exit_success
STORE_NAME exit_success
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object Help at 0x7fab800c7030, file "f.py", line 5>
LOAD_CONST 'Help'
MAKE_FUN... | import Naked.settings
from Naked.toolshed.system import exit_success
class Help:
def __init__(self):
self.help = Naked.settings.help
def print_help(self):
print(self.help)
exit_success()
if __name__ == "__main__":
pass
| data/Naked-0.1.32/lib/Naked/commands/help.py | 306 | 86 | 178,035 |
LOAD_CONST 'Logic for selecting examples to include in prompts.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('NGramOverlapExampleSelector', 'ngram_overlap_score')
IMPORT_NAME langchain_community.example_selectors.ngram_overlap
IMPORT_FROM NGramOverlapExampleSelector
STORE_NAME NGramOverlapExampleSelector
IMPORT_FROM n... | """Logic for selecting examples to include in prompts."""
from langchain_community.example_selectors.ngram_overlap import (
NGramOverlapExampleSelector,
ngram_overlap_score,
)
__all__ = [
"NGramOverlapExampleSelector",
"ngram_overlap_score",
]
| data/langchain_community-0.0.21/langchain_community/example_selectors/__init__.py | 133 | 86 | 15,081 |
LOAD_CONST 2
LOAD_CONST ('and_', 'not_', 'or_')
IMPORT_NAME blocking_rule_composition
IMPORT_FROM and_
STORE_NAME and_
IMPORT_FROM not_
STORE_NAME not_
IMPORT_FROM or_
STORE_NAME or_
POP_TOP
LOAD_CONST 1
LOAD_CONST ('block_on', 'exact_match_rule')
IMPORT_NAME athena_helpers.athena_blocking_rule_imports
IMPORT_FROM blo... | from ..blocking_rule_composition import ( # noqa: F401
and_,
not_,
or_,
)
from .athena_helpers.athena_blocking_rule_imports import ( # noqa: F401
block_on,
exact_match_rule,
)
| data/splink-3.9.12/splink/athena/blocking_rule_library.py | 122 | 86 | 15,950 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME array
STORE_NAME array
LOAD_CONST <code object application at 0x7f8e2fe3cb70, file "f.py", line 4>
LOAD_CONST 'application'
MAKE_FUNCTION
STORE_NAME application
LOAD_CONST None
RETURN_VALUE
LOAD_FAST sr
LOAD_CONST '200 OK'
LOAD_CONST ('Content-Type', 'text/html')
BUILD_LIST
CA... | import array
def application(e, sr):
sr("200 OK", [("Content-Type", "text/html")])
a = array.array("b", [54, 55, 56, 57])
yield a
yield bytearray(b"abcdef")
yield b"ciao"
| data/uwsgi-2.0.24/t/python/testba.py | 173 | 86 | 398,206 |
LOAD_CONST 0
LOAD_CONST ('Enum',)
IMPORT_NAME enum
IMPORT_FROM Enum
STORE_NAME Enum
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object ProvisioningAction at 0x7fab4298ed20, file "f.py", line 4>
LOAD_CONST 'ProvisioningAction'
MAKE_FUNCTION
LOAD_CONST 'ProvisioningAction'
LOAD_NAME str
LOAD_NAME Enum
CALL_FUNCTION
STORE_... | from enum import Enum
class ProvisioningAction(str, Enum):
Other = ("other",)
Create = ("create",)
Delete = ("delete",)
Disable = ("disable",)
Update = ("update",)
StagedDelete = ("stagedDelete",)
UnknownFutureValue = ("unknownFutureValue",)
| data/msgraph-sdk-1.1.0/msgraph/generated/models/provisioning_action.py | 207 | 86 | 260,718 |
LOAD_CONST 0
LOAD_CONST ('TextResource',)
IMPORT_NAME frictionless.resources
IMPORT_FROM TextResource
STORE_NAME TextResource
POP_TOP
LOAD_CONST <code object test_resource_write_text at 0x7fab429cda50, file "f.py", line 4>
LOAD_CONST 'test_resource_write_text'
MAKE_FUNCTION
STORE_NAME test_resource_write_text
LOAD_CON... | from frictionless.resources import TextResource
def test_resource_write_text(tmpdir):
source = TextResource(path="data/article.md")
target = source.write_text(path=str(tmpdir.join("article.md")))
assert target.read_text() == "# Article\n\nContents\n"
| data/frictionless-5.16.1/tests/resources/text/test_write.py | 199 | 86 | 41,519 |
LOAD_BUILD_CLASS
LOAD_CONST <code object BadFilterFormat at 0x7fab702a6db0, file "f.py", line 1>
LOAD_CONST 'BadFilterFormat'
MAKE_FUNCTION
LOAD_CONST 'BadFilterFormat'
LOAD_NAME Exception
CALL_FUNCTION
STORE_NAME BadFilterFormat
LOAD_BUILD_CLASS
LOAD_CONST <code object BadSortFormat at 0x7fab702a6ed0, file "f.py", li... | class BadFilterFormat(Exception):
pass
class BadSortFormat(Exception):
pass
class BadLoadFormat(Exception):
pass
class BadSpec(Exception):
pass
class FieldNotFound(Exception):
pass
class BadQuery(Exception):
pass
class InvalidPage(Exception):
pass
| data/sqlalchemy-filters-0.13.0/sqlalchemy_filters/exceptions.py | 637 | 86 | 351,914 |
LOAD_CONST "Package metadata for dm_env.\n\nThis is kept in a separate module so that it can be imported from setup.py, at\na time when dm_env's dependencies may not have been installed yet.\n"
STORE_NAME __doc__
LOAD_CONST '1.6'
STORE_NAME __version__
LOAD_CONST None
RETURN_VALUE | """Package metadata for dm_env.
This is kept in a separate module so that it can be imported from setup.py, at
a time when dm_env's dependencies may not have been installed yet.
"""
__version__ = "1.6" # https://www.python.org/dev/peps/pep-0440/
| data/dm-env-1.6/dm_env/_metadata.py | 76 | 86 | 309,149 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_IndexedArray_numnull_unique_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_IndexedArray_numnull_unique_64.py | 225 | 86 | 221,325 |
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME azure.cli.core.aaz
IMPORT_STAR
LOAD_NAME register_command_group
LOAD_CONST 'relay hyco authorization-rule keys'
CALL_FUNCTION
LOAD_BUILD_CLASS
LOAD_CONST <code object __CMDGroup at 0x7fab42062030, file "f.py", line 4>
LOAD_CONST '__CMDGroup'
MAKE_FUNCTION
LOAD_CONST '__CMDGr... | from azure.cli.core.aaz import *
@register_command_group(
"relay hyco authorization-rule keys",
)
class __CMDGroup(AAZCommandGroup):
"""Manage Azure Authorization Rule keys for Relay Service Hybrid Connection."""
pass
__all__ = ["__CMDGroup"]
| data/azure-cli-2.57.0/azure/cli/command_modules/relay/aaz/latest/relay/hyco/authorization_rule/keys/__cmd_group.py | 196 | 86 | 375,665 |
LOAD_CONST 1
LOAD_CONST ('AsyncIOScheduler',)
IMPORT_NAME asyncioscheduler
IMPORT_FROM AsyncIOScheduler
STORE_NAME AsyncIOScheduler
POP_TOP
LOAD_CONST 1
LOAD_CONST ('AsyncIOThreadSafeScheduler',)
IMPORT_NAME asynciothreadsafescheduler
IMPORT_FROM AsyncIOThreadSafeScheduler
STORE_NAME AsyncIOThreadSafeScheduler
POP_TOP... | from .asyncioscheduler import AsyncIOScheduler
from .asynciothreadsafescheduler import AsyncIOThreadSafeScheduler
from .eventletscheduler import EventletScheduler
from .geventscheduler import GEventScheduler
from .ioloopscheduler import IOLoopScheduler
from .twistedscheduler import TwistedScheduler
| data/Rx-3.2.0/rx/scheduler/eventloop/__init__.py | 229 | 86 | 187,086 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_reduce_countnonzero_uint8_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_countnonzero_uint8_64.py | 226 | 86 | 221,281 |
LOAD_CONST 'DO NOT EDIT.\n\nThis file was autogenerated. Do not edit it by hand,\nsince your modifications would be overwritten.\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('decode_predictions',)
IMPORT_NAME keras_core.src.applications.mobilenet_v3
IMPORT_FROM decode_predictions
STORE_NAME decode_predictions
POP_TO... | """DO NOT EDIT.
This file was autogenerated. Do not edit it by hand,
since your modifications would be overwritten.
"""
from keras_core.src.applications.mobilenet_v3 import decode_predictions
from keras_core.src.applications.mobilenet_v3 import preprocess_input
| data/keras-core-0.1.7/keras_core/_tf_keras/applications/mobilenet_v3/__init__.py | 143 | 86 | 386,019 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_reduce_argmax_float32_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmax_float32_64.py | 226 | 86 | 221,275 |
LOAD_CONST "Dataset definition for summscreen.\n\nDEPRECATED!\nIf you want to use the Summscreen dataset builder class, use:\ntfds.builder_cls('summscreen')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_builde... | """Dataset definition for summscreen.
DEPRECATED!
If you want to use the Summscreen dataset builder class, use:
tfds.builder_cls('summscreen')
"""
from tensorflow_datasets.core import lazy_builder_import
Summscreen = lazy_builder_import.LazyBuilderImport("summscreen")
| data/tensorflow-datasets-4.9.4/tensorflow_datasets/summarization/summscreen/summscreen.py | 130 | 86 | 256,391 |
LOAD_CONST 2
LOAD_CONST ('and_', 'not_', 'or_')
IMPORT_NAME blocking_rule_composition
IMPORT_FROM and_
STORE_NAME and_
IMPORT_FROM not_
STORE_NAME not_
IMPORT_FROM or_
STORE_NAME or_
POP_TOP
LOAD_CONST 1
LOAD_CONST ('block_on', 'exact_match_rule')
IMPORT_NAME duckdb_helpers.duckdb_blocking_rule_imports
IMPORT_FROM blo... | from ..blocking_rule_composition import ( # noqa: F401
and_,
not_,
or_,
)
from .duckdb_helpers.duckdb_blocking_rule_imports import ( # noqa: F401
block_on,
exact_match_rule,
)
| data/splink-3.9.12/splink/duckdb/blocking_rule_library.py | 122 | 86 | 15,981 |
LOAD_CONST './reppoints-moment_r50_fpn_1x_coco.py'
STORE_NAME _base_
LOAD_NAME dict
LOAD_CONST 'GN'
LOAD_CONST 32
LOAD_CONST True
LOAD_CONST ('type', 'num_groups', 'requires_grad')
CALL_FUNCTION
STORE_NAME norm_cfg
LOAD_NAME dict
LOAD_NAME dict
LOAD_NAME norm_cfg
LOAD_CONST ('norm_cfg',)
CALL_FUNCTION
LOAD_NAME dict
... | _base_ = "./reppoints-moment_r50_fpn_1x_coco.py"
norm_cfg = dict(type="GN", num_groups=32, requires_grad=True)
model = dict(neck=dict(norm_cfg=norm_cfg), bbox_head=dict(norm_cfg=norm_cfg))
| data/mmdet-3.3.0/mmdet/.mim/configs/reppoints/reppoints-moment_r50_fpn-gn_head-gn_1x_coco.py | 130 | 86 | 151,824 |
LOAD_CONST 'Constants for the Rollease Acmeda Automate integration.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME logging
STORE_NAME logging
LOAD_NAME logging
LOAD_METHOD getLogger
LOAD_NAME __package__
CALL_METHOD
STORE_NAME LOGGER
LOAD_CONST 'acmeda'
STORE_NAME DOMAIN
LOAD_CONST 'acmeda_hub_update_... | """Constants for the Rollease Acmeda Automate integration."""
import logging
LOGGER = logging.getLogger(__package__)
DOMAIN = "acmeda"
ACMEDA_HUB_UPDATE = "acmeda_hub_update_{}"
ACMEDA_ENTITY_REMOVE = "acmeda_entity_remove_{}"
| data/homeassistant-2024.2.2/homeassistant/components/acmeda/const.py | 121 | 86 | 406,744 |
LOAD_CONST 0
LOAD_CONST ('absolute_import',)
IMPORT_NAME __future__
IMPORT_FROM absolute_import
STORE_NAME absolute_import
POP_TOP
LOAD_CONST 0
LOAD_CONST ('AppConfig',)
IMPORT_NAME django.apps
IMPORT_FROM AppConfig
STORE_NAME AppConfig
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object RavenConfig at 0x7fab82197e40, f... | from __future__ import absolute_import
from django.apps import AppConfig
class RavenConfig(AppConfig):
name = "raven.contrib.django"
label = "raven_contrib_django"
verbose_name = "Raven"
def ready(self):
from .models import initialize
initialize()
| data/raven-6.10.0/raven/contrib/django/apps.py | 280 | 86 | 373,921 |
LOAD_CONST 0
LOAD_CONST ('Any', 'Dict', 'List')
IMPORT_NAME typing
IMPORT_FROM Any
STORE_NAME Any
IMPORT_FROM Dict
STORE_NAME Dict
IMPORT_FROM List
STORE_NAME List
POP_TOP
LOAD_CONST 1
LOAD_CONST ('YDocExtension',)
IMPORT_NAME app
IMPORT_FROM YDocExtension
STORE_NAME YDocExtension
POP_TOP
LOAD_CONST '0.8.0'
STORE_NAM... | from typing import Any, Dict, List
from .app import YDocExtension
__version__ = "0.8.0"
def _jupyter_server_extension_points() -> List[Dict[str, Any]]:
return [{"module": "jupyter_server_ydoc", "app": YDocExtension}]
| data/jupyter_server_ydoc-0.8.0/jupyter_server_ydoc/__init__.py | 220 | 86 | 308,415 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME typer
STORE_NAME typer
LOAD_NAME typer
LOAD_ATTR Argument
LOAD_CONST Ellipsis
LOAD_CONST 'The name of the user to greet'
LOAD_CONST ('help',)
CALL_FUNCTION
BUILD_TUPLE
LOAD_NAME str
LOAD_CONST ('name',)
BUILD_CONST_KEY_MAP
LOAD_CONST <code object main at 0x7fab641e74b0, file "f... | import typer
def main(name: str = typer.Argument(..., help="The name of the user to greet")):
"""
Say hi to NAME very gently, like Dirk.
"""
print(f"Hello {name}")
if __name__ == "__main__":
typer.run(main)
| data/typer-0.9.0/docs_src/arguments/help/tutorial002.py | 166 | 86 | 224,552 |
LOAD_CONST 2
LOAD_CONST ('ShopifyResource',)
IMPORT_NAME base
IMPORT_FROM ShopifyResource
STORE_NAME ShopifyResource
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object Webhook at 0x7fab427d34b0, file "f.py", line 4>
LOAD_CONST 'Webhook'
MAKE_FUNCTION
LOAD_CONST 'Webhook'
LOAD_NAME ShopifyResource
CALL_FUNCTION
STORE_NAM... | from ..base import ShopifyResource
class Webhook(ShopifyResource):
def __get_format(self):
return self.attributes.get("format")
def __set_format(self, data):
self.attributes["format"] = data
format = property(__get_format, __set_format, None, "Format attribute")
| data/ShopifyAPI-12.4.0/shopify/resources/webhook.py | 312 | 86 | 216,666 |
LOAD_CONST 'Constants for integration.'
STORE_NAME __doc__
LOAD_CONST 'gogogate2'
STORE_NAME DOMAIN
LOAD_CONST 'data_update_coordinator'
STORE_NAME DATA_UPDATE_COORDINATOR
LOAD_CONST 'gogogate2'
STORE_NAME DEVICE_TYPE_GOGOGATE2
LOAD_CONST 'ismartgate'
STORE_NAME DEVICE_TYPE_ISMARTGATE
LOAD_CONST 'Remsol'
STORE_NAM... | """Constants for integration."""
DOMAIN = "gogogate2"
DATA_UPDATE_COORDINATOR = "data_update_coordinator"
DEVICE_TYPE_GOGOGATE2 = "gogogate2"
DEVICE_TYPE_ISMARTGATE = "ismartgate"
MANUFACTURER = "Remsol"
| data/homeassistant-2024.2.2/homeassistant/components/gogogate2/const.py | 111 | 86 | 406,909 |
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME networkx.algorithms.shortest_paths.generic
IMPORT_STAR
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME networkx.algorithms.shortest_paths.unweighted
IMPORT_STAR
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME networkx.algorithms.shortest_paths.weighted
IMPORT_STAR
LOAD_CONST 0
LOAD_C... | from networkx.algorithms.shortest_paths.generic import *
from networkx.algorithms.shortest_paths.unweighted import *
from networkx.algorithms.shortest_paths.weighted import *
from networkx.algorithms.shortest_paths.astar import *
from networkx.algorithms.shortest_paths.dense import *
| data/networkx-3.2.1/networkx/algorithms/shortest_paths/__init__.py | 144 | 86 | 130,900 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_reduce_argmin_uint16_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmin_uint16_64.py | 225 | 86 | 221,457 |
LOAD_CONST 0
LOAD_CONST ('Enum',)
IMPORT_NAME enum
IMPORT_FROM Enum
STORE_NAME Enum
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object KubernetesServiceType at 0x7fab70027420, file "f.py", line 4>
LOAD_CONST 'KubernetesServiceType'
MAKE_FUNCTION
LOAD_CONST 'KubernetesServiceType'
LOAD_NAME str
LOAD_NAME Enum
CALL_FUNCTI... | from enum import Enum
class KubernetesServiceType(str, Enum):
Unknown = ("unknown",)
ClusterIP = ("clusterIP",)
ExternalName = ("externalName",)
NodePort = ("nodePort",)
LoadBalancer = ("loadBalancer",)
UnknownFutureValue = ("unknownFutureValue",)
| data/msgraph-sdk-1.1.0/msgraph/generated/models/security/kubernetes_service_type.py | 210 | 86 | 262,212 |
LOAD_CONST "Dataset definition for reddit.\n\nDEPRECATED!\nIf you want to use the Reddit dataset builder class, use:\ntfds.builder_cls('reddit')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_builder_import
POP... | """Dataset definition for reddit.
DEPRECATED!
If you want to use the Reddit dataset builder class, use:
tfds.builder_cls('reddit')
"""
from tensorflow_datasets.core import lazy_builder_import
Reddit = lazy_builder_import.LazyBuilderImport("reddit")
| data/tensorflow-datasets-4.9.4/tensorflow_datasets/summarization/reddit.py | 130 | 86 | 256,376 |
LOAD_CONST 'Reading functions for freesurfer files\n'
STORE_NAME __doc__
LOAD_CONST 1
LOAD_CONST ('read_annot', 'read_geometry', 'read_label', 'read_morph_data', 'write_annot', 'write_geometry', 'write_morph_data')
IMPORT_NAME io
IMPORT_FROM read_annot
STORE_NAME read_annot
IMPORT_FROM read_geometry
STORE_NAME read_ge... | """Reading functions for freesurfer files
"""
from .io import (
read_annot,
read_geometry,
read_label,
read_morph_data,
write_annot,
write_geometry,
write_morph_data,
)
from .mghformat import MGHImage, load, save
| data/nibabel-5.2.0/nibabel/freesurfer/__init__.py | 212 | 86 | 68,738 |
LOAD_CONST 0
LOAD_CONST ('List', 'Optional')
IMPORT_NAME typing
IMPORT_FROM List
STORE_NAME List
IMPORT_FROM Optional
STORE_NAME Optional
POP_TOP
LOAD_CONST 0
LOAD_CONST ('BaseModel',)
IMPORT_NAME pydantic
IMPORT_FROM BaseModel
STORE_NAME BaseModel
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object XaiParams at 0x7f8e2... | from typing import List, Optional
from pydantic import BaseModel
class XaiParams(BaseModel):
custom_explain_methods: List[str] = []
"""User-defined explain_custom method of the model object defined in package.py"""
default_explain_method: Optional[str] = None
"""Default explanation method"""
| data/fiddler-client-2.4.0/fiddler3/schemas/xai_params.py | 232 | 86 | 422,818 |
LOAD_CONST 0
LOAD_CONST ('absolute_import',)
IMPORT_NAME __future__
IMPORT_FROM absolute_import
STORE_NAME absolute_import
POP_TOP
LOAD_CONST 1
LOAD_CONST ('LogAnalyticsClient',)
IMPORT_NAME log_analytics_client
IMPORT_FROM LogAnalyticsClient
STORE_NAME LogAnalyticsClient
POP_TOP
LOAD_CONST 1
LOAD_CONST ('LogAnalytic... | from __future__ import absolute_import
from .log_analytics_client import LogAnalyticsClient
from .log_analytics_client_composite_operations import (
LogAnalyticsClientCompositeOperations,
)
from . import models
__all__ = ["LogAnalyticsClient", "LogAnalyticsClientCompositeOperations", "models"]
| data/oci-2.122.0/src/oci/log_analytics/__init__.py | 179 | 86 | 179,977 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_IndexedArray64_index_of_nulls_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_IndexedArray64_index_of_nulls.py | 225 | 86 | 221,446 |
LOAD_CONST 0
LOAD_CONST ('Optional',)
IMPORT_NAME typing
IMPORT_FROM Optional
STORE_NAME Optional
POP_TOP
LOAD_CONST 0
LOAD_CONST ('Entity',)
IMPORT_NAME office365.entity
IMPORT_FROM Entity
STORE_NAME Entity
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object ColumnLink at 0x7fab700add20, file "f.py", line 6>
LOAD_CONST... | from typing import Optional
from office365.entity import Entity
class ColumnLink(Entity):
"""A columnLink on a contentType attaches a site columnDefinition to that content type."""
@property
def name(self):
"""The name of the column in this content type."""
return self.properties.get("na... | data/Office365-REST-Python-Client-2.5.5/office365/onedrive/columns/column_link.py | 240 | 86 | 188,556 |
LOAD_CONST 2
LOAD_CONST ('Table',)
IMPORT_NAME db_models.table
IMPORT_FROM Table
STORE_NAME Table
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object TemplateTable at 0x7fab703bb8a0, file "f.py", line 4>
LOAD_CONST 'TemplateTable'
MAKE_FUNCTION
LOAD_CONST 'TemplateTable'
CALL_FUNCTION
STORE_NAME TemplateTable
LOAD_CONST ... | from ..db_models.table import Table
class TemplateTable:
def __init__(self, table: Table) -> None:
for attr in dir(table):
if not attr.startswith("__"):
setattr(self, attr, getattr(table, attr))
self.table = table
self.comments = ""
| data/dataprep-0.4.5/dataprep/eda/create_db_report/template_models/table.py | 280 | 86 | 199,368 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_reduce_argmax_float64_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_argmax_float64_64.py | 225 | 86 | 221,289 |
LOAD_CONST "Dataset definition for assin2.\n\nDEPRECATED!\nIf you want to use the Assin2 dataset builder class, use:\ntfds.builder_cls('assin2')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_builder_import
POP... | """Dataset definition for assin2.
DEPRECATED!
If you want to use the Assin2 dataset builder class, use:
tfds.builder_cls('assin2')
"""
from tensorflow_datasets.core import lazy_builder_import
Assin2 = lazy_builder_import.LazyBuilderImport("assin2")
| data/tensorflow-datasets-4.9.4/tensorflow_datasets/text/assin2/assin2.py | 130 | 86 | 256,272 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME subprocess
STORE_NAME subprocess
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME argparse
STORE_NAME argparse
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME sys
STORE_NAME sys
LOAD_NAME __name__
LOAD_CONST '__main__'
COMPARE_OP ==
POP_JUMP_IF_FALSE
LOAD_NAME argparse
LOAD_METHOD Argument... | import subprocess
import argparse
import sys
if __name__ == "__main__":
parser = argparse.ArgumentParser()
parser.add_argument("prog")
args = parser.parse_args()
res = subprocess.run(args.prog)
sys.exit(res.returncode - 42)
| data/numpy-1.26.4/vendored-meson/meson/test cases/windows/13 test argument extra paths/test/test_run_exe.py | 161 | 86 | 363,664 |
SETUP_EXCEPT to 14
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME some_module
STORE_NAME some_module
POP_BLOCK
JUMP_FORWARD to 42
DUP_TOP
LOAD_NAME ImportError
COMPARE_OP exception match
POP_JUMP_IF_FALSE
POP_TOP
POP_TOP
POP_TOP
LOAD_NAME print
LOAD_CONST 'Cannot import wrongly cased module.'
CALL_FUNCTION
POP_TOP
POP_EXC... | try:
import some_module
except ImportError:
print("Cannot import wrongly cased module.")
else:
print("OK, imported wrongly cased module.")
try:
import some_package
except ImportError:
print("Cannot import wrongly cased package.")
else:
print("OK, imported wrongly cased package.")
| data/Nuitka-2.0.3/tests/programs/case_imports3/CasedImportingMain.py | 207 | 86 | 148,278 |
LOAD_CONST 0
LOAD_CONST ('absolute_import',)
IMPORT_NAME __future__
IMPORT_FROM absolute_import
STORE_NAME absolute_import
POP_TOP
LOAD_CONST 1
LOAD_CONST ('DataLabelingClient',)
IMPORT_NAME data_labeling_client
IMPORT_FROM DataLabelingClient
STORE_NAME DataLabelingClient
POP_TOP
LOAD_CONST 1
LOAD_CONST ('DataLabelin... | from __future__ import absolute_import
from .data_labeling_client import DataLabelingClient
from .data_labeling_client_composite_operations import (
DataLabelingClientCompositeOperations,
)
from . import models
__all__ = ["DataLabelingClient", "DataLabelingClientCompositeOperations", "models"]
| data/oci-2.122.0/src/oci/data_labeling_service_dataplane/__init__.py | 179 | 86 | 178,690 |
LOAD_CONST 'Tsuyoshi Hombashi'
STORE_NAME __author__
LOAD_CONST 'Copyright 2016, '
LOAD_NAME __author__
FORMAT_VALUE
BUILD_STRING
STORE_NAME __copyright__
LOAD_CONST 'MIT License'
STORE_NAME __license__
LOAD_CONST '1.1.3'
STORE_NAME __version__
LOAD_NAME __author__
STORE_NAME __maintainer__
LOAD_CONST 'tsuyoshi.ho... | __author__ = "Tsuyoshi Hombashi"
__copyright__ = f"Copyright 2016, {__author__}"
__license__ = "MIT License"
__version__ = "1.1.3"
__maintainer__ = __author__
__email__ = "tsuyoshi.hombashi@gmail.com"
| data/mbstrdecoder-1.1.3/mbstrdecoder/__version__.py | 112 | 87 | 45,231 |
LOAD_CONST 0
LOAD_CONST ('Union',)
IMPORT_NAME typing
IMPORT_FROM Union
STORE_NAME Union
POP_TOP
LOAD_NAME str
LOAD_NAME Union
LOAD_NAME int
LOAD_NAME str
BUILD_TUPLE
BINARY_SUBSCR
LOAD_CONST ('s', 'return')
BUILD_CONST_KEY_MAP
LOAD_CONST <code object _to_int at 0x7fab823600c0, file "f.py", line 4>
LOAD_CONST '_to_int... | from typing import Union
def _to_int(s: str) -> Union[int, str]:
try:
return int(s)
except ValueError:
return s
__version__ = "2.10.4"
version_info = tuple(_to_int(s) for s in __version__.split("."))
| data/onnx-1.15.0/third_party/pybind11/pybind11/_version.py | 281 | 87 | 111,597 |
LOAD_CONST '../_base_/models/faster-rcnn_r50_fpn.py'
LOAD_CONST '../_base_/datasets/coco_detection.py'
LOAD_CONST '../_base_/schedules/schedule_2x.py'
LOAD_CONST '../_base_/default_runtime.py'
BUILD_LIST
STORE_NAME _base_
LOAD_CONST None
RETURN_VALUE | _base_ = [
"../_base_/models/faster-rcnn_r50_fpn.py",
"../_base_/datasets/coco_detection.py",
"../_base_/schedules/schedule_2x.py",
"../_base_/default_runtime.py",
]
| data/mmdet-3.3.0/mmdet/.mim/configs/faster_rcnn/faster-rcnn_r50_fpn_2x_coco.py | 93 | 87 | 151,117 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME benchling_api_client.v2.types
IMPORT_FROM v2
ROT_TWO
POP_TOP
IMPORT_FROM types
STORE_NAME root_types
POP_TOP
LOAD_NAME root_types
LOAD_ATTR Unset
STORE_NAME Unset
LOAD_NAME root_types
LOAD_ATTR UNSET
STORE_NAME UNSET
LOAD_NAME root_types
LOAD_ATTR FileJsonType
STORE_NAME File... | import benchling_api_client.v2.types as root_types
Unset = root_types.Unset
UNSET = root_types.UNSET
FileJsonType = root_types.FileJsonType
File = root_types.File
T = root_types.T
Response = root_types.Response
__all__ = ["File", "Response"]
| data/benchling_api_client-2.0.271/benchling_api_client/webhooks/v0/beta/types.py | 141 | 87 | 70,875 |
LOAD_CONST 'Define Notion model mixins.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('dataclass',)
IMPORT_NAME dataclasses
IMPORT_FROM dataclass
STORE_NAME dataclass
POP_TOP
LOAD_CONST 0
LOAD_CONST ('ListenerKind',)
IMPORT_NAME aionotion.sensor.models
IMPORT_FROM ListenerKind
STORE_NAME ListenerKind
POP_TOP
LOAD_NAM... | """Define Notion model mixins."""
from dataclasses import dataclass
from aionotion.sensor.models import ListenerKind
@dataclass(frozen=True, kw_only=True)
class NotionEntityDescription:
"""Define an description for Notion entities."""
listener_kind: ListenerKind
| data/homeassistant-2024.2.2/homeassistant/components/notion/model.py | 248 | 87 | 298,095 |
LOAD_CONST 0
LOAD_CONST ('Enum',)
IMPORT_NAME enum
IMPORT_FROM Enum
STORE_NAME Enum
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object LogonType at 0x7fab70027660, file "f.py", line 4>
LOAD_CONST 'LogonType'
MAKE_FUNCTION
LOAD_CONST 'LogonType'
LOAD_NAME str
LOAD_NAME Enum
CALL_FUNCTION
STORE_NAME LogonType
LOAD_CONST N... | from enum import Enum
class LogonType(str, Enum):
Unknown = ("unknown",)
Interactive = ("interactive",)
RemoteInteractive = ("remoteInteractive",)
Network = ("network",)
Batch = ("batch",)
Service = ("service",)
UnknownFutureValue = ("unknownFutureValue",)
| data/msgraph-sdk-1.1.0/msgraph/generated/models/logon_type.py | 205 | 87 | 259,926 |
LOAD_CONST 0
LOAD_CONST ('Union',)
IMPORT_NAME typing
IMPORT_FROM Union
STORE_NAME Union
POP_TOP
LOAD_NAME str
LOAD_NAME Union
LOAD_NAME int
LOAD_NAME str
BUILD_TUPLE
BINARY_SUBSCR
LOAD_CONST ('s', 'return')
BUILD_CONST_KEY_MAP
LOAD_CONST <code object _to_int at 0x7fab8212bb70, file "f.py", line 4>
LOAD_CONST '_to_int... | from typing import Union
def _to_int(s: str) -> Union[int, str]:
try:
return int(s)
except ValueError:
return s
__version__ = "2.11.1"
version_info = tuple(_to_int(s) for s in __version__.split("."))
| data/osmium-3.7.0/contrib/pybind11/pybind11/_version.py | 279 | 87 | 170,033 |
LOAD_CONST 1
LOAD_CONST ('DeleteBulkAction',)
IMPORT_NAME delete
IMPORT_FROM DeleteBulkAction
STORE_NAME DeleteBulkAction
POP_TOP
LOAD_CONST 1
LOAD_CONST ('MoveBulkAction',)
IMPORT_NAME move
IMPORT_FROM MoveBulkAction
STORE_NAME MoveBulkAction
POP_TOP
LOAD_CONST 1
LOAD_CONST ('PublishBulkAction',)
IMPORT_NAME publish... | from .delete import DeleteBulkAction
from .move import MoveBulkAction
from .publish import PublishBulkAction
from .unpublish import UnpublishBulkAction
__all__ = [
"DeleteBulkAction",
"MoveBulkAction",
"PublishBulkAction",
"UnpublishBulkAction",
]
| data/wagtail-6.0.1/wagtail/admin/views/pages/bulk_actions/__init__.py | 185 | 87 | 200,373 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME benchling_api_client.v2.extensions
STORE_NAME benchling_api_client
LOAD_NAME benchling_api_client
LOAD_ATTR v2
LOAD_ATTR extensions
LOAD_ATTR NotPresentError
STORE_NAME NotPresentError
LOAD_NAME benchling_api_client
LOAD_ATTR v2
LOAD_ATTR extensions
LOAD_ATTR UnknownType
STORE... | import benchling_api_client.v2.extensions
NotPresentError = benchling_api_client.v2.extensions.NotPresentError
UnknownType = benchling_api_client.v2.extensions.UnknownType
T = benchling_api_client.v2.extensions.T
Enums = benchling_api_client.v2.extensions.Enums
| data/benchling_api_client-2.0.271/benchling_api_client/webhooks/v0/beta/extensions.py | 129 | 87 | 70,877 |
LOAD_CONST 'Tsuyoshi Hombashi'
STORE_NAME __author__
LOAD_CONST 'Copyright 2016, '
LOAD_NAME __author__
FORMAT_VALUE
BUILD_STRING
STORE_NAME __copyright__
LOAD_CONST 'MIT License'
STORE_NAME __license__
LOAD_CONST '2.2.0'
STORE_NAME __version__
LOAD_NAME __author__
STORE_NAME __maintainer__
LOAD_CONST 'tsuyoshi.ho... | __author__ = "Tsuyoshi Hombashi"
__copyright__ = f"Copyright 2016, {__author__}"
__license__ = "MIT License"
__version__ = "2.2.0"
__maintainer__ = __author__
__email__ = "tsuyoshi.hombashi@gmail.com"
| data/DateTimeRange-2.2.0/datetimerange/__version__.py | 112 | 87 | 232,683 |
LOAD_CONST 1
LOAD_CONST ('ascii_value',)
IMPORT_NAME ascii_value
IMPORT_FROM ascii_value
STORE_NAME ascii_value
POP_TOP
LOAD_CONST 1
LOAD_CONST ('deviated_cell',)
IMPORT_NAME deviated_cell
IMPORT_FROM deviated_cell
STORE_NAME deviated_cell
POP_TOP
LOAD_CONST 1
LOAD_CONST ('deviated_value',)
IMPORT_NAME deviated_value... | from .ascii_value import ascii_value
from .deviated_cell import deviated_cell
from .deviated_value import deviated_value
from .forbidden_value import forbidden_value
from .required_value import required_value
from .sequential_value import sequential_value
from .truncated_value import truncated_value
| data/frictionless-5.16.1/frictionless/checks/cell/__init__.py | 244 | 87 | 41,810 |
LOAD_CONST 0
LOAD_CONST ('exceptions',)
IMPORT_NAME selenium.common
IMPORT_FROM exceptions
STORE_NAME exceptions
POP_TOP
LOAD_CONST 0
LOAD_CONST ('Keys',)
IMPORT_NAME selenium.webdriver.common.keys
IMPORT_FROM Keys
STORE_NAME Keys
POP_TOP
LOAD_CONST 0
LOAD_CONST ('Select',)
IMPORT_NAME selenium.webdriver.support.ui
I... | from selenium.common import exceptions # noqa: F401
from selenium.webdriver.common.keys import Keys # noqa: F401
from selenium.webdriver.support.ui import Select # noqa: F401
from .requestium import Session # noqa: F401
| data/requestium-0.4.0/requestium/__init__.py | 118 | 87 | 249,283 |
LOAD_CONST 0
LOAD_CONST ('CMSPluginBase',)
IMPORT_NAME cms.plugin_base
IMPORT_FROM CMSPluginBase
STORE_NAME CMSPluginBase
POP_TOP
LOAD_CONST 0
LOAD_CONST ('plugin_pool',)
IMPORT_NAME cms.plugin_pool
IMPORT_FROM plugin_pool
STORE_NAME plugin_pool
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object NoCustomModel at 0x7fab... | from cms.plugin_base import CMSPluginBase
from cms.plugin_pool import plugin_pool
class NoCustomModel(CMSPluginBase):
name = "NoCustomModel"
render_plugin = True
render_template = "plugins/no_custom_model.html"
plugin_pool.register_plugin(NoCustomModel)
| data/django-cms-4.1.0/cms/test_utils/project/pluginapp/plugins/no_custom_model/cms_plugins.py | 225 | 87 | 87,008 |
LOAD_BUILD_CLASS
LOAD_CONST <code object NumericalWarning at 0x7fab8223fc90, file "f.py", line 1>
LOAD_CONST 'NumericalWarning'
MAKE_FUNCTION
LOAD_CONST 'NumericalWarning'
LOAD_NAME RuntimeWarning
CALL_FUNCTION
STORE_NAME NumericalWarning
LOAD_BUILD_CLASS
LOAD_CONST <code object PerformanceWarning at 0x7fab8223f930, f... | class NumericalWarning(RuntimeWarning):
"""
Warning thrown when convergence criteria are not met, or when comptuations require extra stability.
"""
pass
class PerformanceWarning(RuntimeWarning):
"""
Warning thrown when LinearOperators are used in a way that may incur large performance / memor... | data/linear_operator-0.5.2/linear_operator/utils/warnings.py | 257 | 87 | 299,990 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_reduce_sum_bool_int32_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_sum_bool_int32_64.py | 228 | 87 | 221,433 |
LOAD_CONST 0
LOAD_CONST ('migrations',)
IMPORT_NAME django.db
IMPORT_FROM migrations
STORE_NAME migrations
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object Migration at 0x7fa954a8ced0, file "f.py", line 4>
LOAD_CONST 'Migration'
MAKE_FUNCTION
LOAD_CONST 'Migration'
LOAD_NAME migrations
LOAD_ATTR Migration
CALL_FUNCTIO... | from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
("core", "0016_charfield_to_textfield"),
]
operations = [
migrations.RemoveField(
model_name="task",
name="parent",
),
]
| data/pulpcore-3.46.0/pulpcore/app/migrations/0017_remove_task_parent.py | 186 | 87 | 218,139 |
LOAD_CONST 0
LOAD_CONST ('symbols', 'S')
IMPORT_NAME sympy.core
IMPORT_FROM symbols
STORE_NAME symbols
IMPORT_FROM S
STORE_NAME S
POP_TOP
LOAD_NAME symbols
LOAD_CONST 'x,y'
CALL_FUNCTION
UNPACK_SEQUENCE
STORE_NAME x
STORE_NAME y
LOAD_CONST <code object timeit_Symbol_meth_lookup at 0x7fab8239ab70, file "f.py", line 6>... | from sympy.core import symbols, S
x, y = symbols("x,y")
def timeit_Symbol_meth_lookup():
x.diff # no call, just method lookup
def timeit_S_lookup():
S.Exp1
def timeit_Symbol_eq_xy():
x == y
| data/sympy-1.12/sympy/core/benchmarks/bench_basic.py | 301 | 87 | 228,954 |
LOAD_CONST 0
LOAD_CONST ('OpRun',)
IMPORT_NAME onnx.reference.op_run
IMPORT_FROM OpRun
STORE_NAME OpRun
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object Mean at 0x7f8e447b2c00, file "f.py", line 4>
LOAD_CONST 'Mean'
MAKE_FUNCTION
LOAD_CONST 'Mean'
LOAD_NAME OpRun
CALL_FUNCTION
STORE_NAME Mean
LOAD_CONST None
RETURN_VA... | from onnx.reference.op_run import OpRun
class Mean(OpRun):
def _run(self, *args): # type: ignore
res = args[0].copy()
for m in args[1:]:
res += m
return ((res / len(args)).astype(args[0].dtype),)
| data/onnxsim-0.4.35/third_party/onnx-optimizer/third_party/onnx/onnx/reference/ops/op_mean.py | 278 | 87 | 114,153 |
LOAD_CONST 0
LOAD_CONST ('AccessTokenDatabase',)
IMPORT_NAME fastapi_users.authentication.strategy.db.adapter
IMPORT_FROM AccessTokenDatabase
STORE_NAME AccessTokenDatabase
POP_TOP
LOAD_CONST 0
LOAD_CONST ('AP', 'AccessTokenProtocol')
IMPORT_NAME fastapi_users.authentication.strategy.db.models
IMPORT_FROM AP
STORE_NAM... | from fastapi_users.authentication.strategy.db.adapter import AccessTokenDatabase
from fastapi_users.authentication.strategy.db.models import AP, AccessTokenProtocol
from fastapi_users.authentication.strategy.db.strategy import DatabaseStrategy
__all__ = ["AP", "AccessTokenDatabase", "AccessTokenProtocol", "DatabaseStr... | data/fastapi_users-12.1.3/fastapi_users/authentication/strategy/db/__init__.py | 167 | 87 | 109,538 |
LOAD_BUILD_CLASS
LOAD_CONST <code object ODataMethod at 0x7fab541051e0, file "f.py", line 1>
LOAD_CONST 'ODataMethod'
MAKE_FUNCTION
LOAD_CONST 'ODataMethod'
LOAD_NAME object
CALL_FUNCTION
STORE_NAME ODataMethod
LOAD_CONST None
RETURN_VALUE
LOAD_NAME __name__
STORE_NAME __module__
LOAD_CONST 'ODataMethod'
STORE_NAME __... | class ODataMethod(object):
def __init__(
self, name=None, is_beta=None, parameters=None, return_type_full_name=None
):
self.Name = name
self.Parameters = parameters
self.IsBeta = is_beta
self.ReturnTypeFullName = return_type_full_name
| data/Office365-REST-Python-Client-2.5.5/office365/runtime/odata/method.py | 218 | 87 | 189,817 |
LOAD_CONST 0
LOAD_CONST ('absolute_import', 'division', 'print_function')
IMPORT_NAME __future__
IMPORT_FROM absolute_import
STORE_NAME absolute_import
IMPORT_FROM division
STORE_NAME division
IMPORT_FROM print_function
STORE_NAME print_function
POP_TOP
LOAD_NAME type
STORE_NAME __metaclass__
LOAD_BUILD_CLASS
LOAD_CO... | from __future__ import absolute_import, division, print_function
__metaclass__ = type
class ModuleDocFragment(object):
DOCUMENTATION = """
options:
name:
description: Use O(name) along with the other options to return only resources that match the given name
exact... | data/ansible-9.2.0/ansible_collections/community/general/plugins/doc_fragments/oracle_name_option.py | 211 | 87 | 275,301 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_reduce_sum_bool_int64_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_sum_bool_int64_64.py | 229 | 87 | 221,443 |
LOAD_CONST 'yarg'
STORE_NAME __title__
LOAD_CONST '0.1.9'
STORE_NAME __version__
LOAD_CONST 'Kura'
STORE_NAME __author__
LOAD_CONST 'kura@kura.io'
STORE_NAME __email__
LOAD_CONST 'https://yarg.readthedocs.org/'
STORE_NAME __url__
LOAD_CONST 'MIT'
STORE_NAME __license__
LOAD_CONST 'Copyright 2014 Kura'
STORE_NAME ... | __title__ = "yarg"
__version__ = "0.1.9"
__author__ = "Kura"
__email__ = "kura@kura.io"
__url__ = "https://yarg.readthedocs.org/"
__license__ = "MIT"
__copyright__ = "Copyright 2014 Kura"
| data/yarg-0.1.9/yarg/__about__.py | 112 | 87 | 272,535 |
LOAD_CONST 'Models.'
STORE_NAME __doc__
LOAD_CONST 1
LOAD_CONST ('Model', 'ModelOptions', 'maybe_model', 'registry')
IMPORT_NAME base
IMPORT_FROM Model
STORE_NAME Model
IMPORT_FROM ModelOptions
STORE_NAME ModelOptions
IMPORT_FROM maybe_model
STORE_NAME maybe_model
IMPORT_FROM registry
STORE_NAME registry
POP_TOP
LOAD... | """Models."""
from .base import Model, ModelOptions, maybe_model, registry
from .fields import FieldDescriptor, StringField
from .record import Record
__all__ = [
"FieldDescriptor",
"Model",
"ModelOptions",
"Record",
"StringField",
"maybe_model",
"registry",
]
| data/faust-streaming-0.10.22/faust/models/__init__.py | 186 | 87 | 77,424 |
LOAD_CONST 3
LOAD_CONST ('register_converter',)
IMPORT_NAME common._registration
IMPORT_FROM register_converter
STORE_NAME register_converter
POP_TOP
LOAD_CONST <code object convert_identity at 0x7fab823c25d0, file "f.py", line 4>
LOAD_CONST 'convert_identity'
MAKE_FUNCTION
STORE_NAME convert_identity
LOAD_NAME regis... | from ...common._registration import register_converter
def convert_identity(scope, operator, container):
container.add_node(
"Identity",
operator.input_full_names,
operator.output_full_names,
name=operator.full_name,
)
register_converter("identity", convert_identity)
| data/onnxmltools-1.12.0/onnxmltools/convert/coreml/operator_converters/Identity.py | 169 | 87 | 289,032 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME os
STORE_NAME os
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME spacy
STORE_NAME spacy
LOAD_CONST <code object test_all_entry_points at 0x7fadb5811c00, file "f.py", line 5>
LOAD_CONST 'test_all_entry_points'
MAKE_FUNCTION
STORE_NAME test_all_entry_points
LOAD_CONST None
RETURN_VALUE... | import os
import spacy
def test_all_entry_points():
entry_points = spacy.registry.lookups.get_all()
for lang, tables in spacy.registry.lookups.get_all().items():
for name, filename in tables.items():
spacy.util.load_language_data(filename)
| data/spacy_lookups_data-1.0.5/spacy_lookups_data/tests/test_entry_points.py | 231 | 87 | 98,302 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_UnionArray64_flatten_length_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_UnionArray64_flatten_length_64.py | 229 | 87 | 221,436 |
LOAD_CONST './cascade-mask-rcnn_r50_fpn_20e_coco.py'
STORE_NAME _base_
LOAD_NAME dict
LOAD_NAME dict
LOAD_CONST 101
LOAD_NAME dict
LOAD_CONST 'Pretrained'
LOAD_CONST 'torchvision://resnet101'
LOAD_CONST ('type', 'checkpoint')
CALL_FUNCTION
LOAD_CONST ('depth', 'init_cfg')
CALL_FUNCTION
LOAD_CONST ('backbone',)
CALL... | _base_ = "./cascade-mask-rcnn_r50_fpn_20e_coco.py"
model = dict(
backbone=dict(
depth=101,
init_cfg=dict(type="Pretrained", checkpoint="torchvision://resnet101"),
)
)
| data/mmdet-3.3.0/mmdet/.mim/configs/cascade_rcnn/cascade-mask-rcnn_r101_fpn_20e_coco.py | 112 | 87 | 151,181 |
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME azure.cli.core.aaz
IMPORT_STAR
LOAD_NAME register_command_group
LOAD_CONST 'network vnet-gateway aad'
CALL_FUNCTION
LOAD_BUILD_CLASS
LOAD_CONST <code object __CMDGroup at 0x7fab54036930, file "f.py", line 4>
LOAD_CONST '__CMDGroup'
MAKE_FUNCTION
LOAD_CONST '__CMDGroup'
LOAD... | from azure.cli.core.aaz import *
@register_command_group(
"network vnet-gateway aad",
)
class __CMDGroup(AAZCommandGroup):
"""Manage AAD(Azure Active Directory) authentication of a virtual network gateway."""
pass
__all__ = ["__CMDGroup"]
| data/azure-cli-2.57.0/azure/cli/command_modules/network/aaz/latest/network/vnet_gateway/aad/__cmd_group.py | 197 | 87 | 378,313 |
LOAD_BUILD_CLASS
LOAD_CONST <code object A at 0x7fab4106c0c0, file "f.py", line 1>
LOAD_CONST 'A'
MAKE_FUNCTION
LOAD_CONST 'A'
CALL_FUNCTION
STORE_NAME A
LOAD_BUILD_CLASS
LOAD_CONST <code object B at 0x7fab4106c930, file "f.py", line 5>
LOAD_CONST 'B'
MAKE_FUNCTION
LOAD_CONST 'B'
LOAD_NAME Exception
CALL_FUNCTION
STOR... | class A:
pass
class B(Exception):
pass
class MyInt(int):
class MyInt(int):
__slots__ = ()
try:
(1).__class__ = MyInt
assert False, "builtin types don't support __class__ assignment."
except TypeError:
pass
| data/xdis-6.0.5/test/simple_source/def/01_class.py | 443 | 87 | 85,461 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME os
STORE_NAME os
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME sys
STORE_NAME sys
LOAD_NAME sys
LOAD_ATTR path
LOAD_METHOD insert
LOAD_CONST 0
LOAD_NAME os
LOAD_METHOD getcwd
CALL_METHOD
CALL_METHOD
POP_TOP
LOAD_CONST 'amqp://guest:guest@localhost:5672//'
STORE_NAME broker_url
LO... | import os
import sys
sys.path.insert(0, os.getcwd())
broker_url = "amqp://guest:guest@localhost:5672//"
worker_disable_rate_limits = True
result_expires = 30 * 60
imports = ("tasks", "webcrawler")
| data/celery-5.3.6/examples/eventlet/celeryconfig.py | 131 | 87 | 69,818 |
LOAD_CONST 'Benchmarking utilities\n'
STORE_NAME __doc__
LOAD_CONST 2
LOAD_CONST ('get_info',)
IMPORT_NAME
IMPORT_FROM get_info
STORE_NAME get_info
POP_TOP
LOAD_CONST <code object print_git_title at 0x7fab700cb300, file "f.py", line 7>
LOAD_CONST 'print_git_title'
MAKE_FUNCTION
STORE_NAME print_git_title
LOAD_CONST N... | """Benchmarking utilities
"""
from .. import get_info
def print_git_title(title):
"""Prints title string with git hash if possible, and underline"""
title = f"{title} for git revision {get_info()['commit_hash']}"
print(title)
print("-" * len(title))
| data/nibabel-5.2.0/nibabel/benchmarks/butils.py | 177 | 87 | 68,618 |
LOAD_CONST './cspnext-s_8xb256-rsb-a1-600e_in1k.py'
STORE_NAME _base_
LOAD_NAME dict
LOAD_NAME dict
LOAD_CONST 1.33
LOAD_CONST 1.25
LOAD_CONST ('deepen_factor', 'widen_factor')
CALL_FUNCTION
LOAD_NAME dict
LOAD_CONST 1280
LOAD_CONST ('in_channels',)
CALL_FUNCTION
LOAD_CONST ('backbone', 'head')
CALL_FUNCTION
STORE_NA... | _base_ = "./cspnext-s_8xb256-rsb-a1-600e_in1k.py"
model = dict(
backbone=dict(deepen_factor=1.33, widen_factor=1.25), head=dict(in_channels=1280)
)
| data/mmdet-3.3.0/mmdet/.mim/configs/rtmdet/classification/cspnext-x_8xb256-rsb-a1-600e_in1k.py | 116 | 87 | 151,262 |
LOAD_CONST 0
LOAD_CONST ('*',)
IMPORT_NAME azure.cli.core.aaz
IMPORT_STAR
LOAD_NAME register_command_group
LOAD_CONST 'monitor log-analytics query-pack'
LOAD_CONST True
LOAD_CONST ('is_preview',)
CALL_FUNCTION
LOAD_BUILD_CLASS
LOAD_CONST <code object __CMDGroup at 0x7fab416c65d0, file "f.py", line 4>
LOAD_CONST '__C... | from azure.cli.core.aaz import *
@register_command_group(
"monitor log-analytics query-pack",
is_preview=True,
)
class __CMDGroup(AAZCommandGroup):
"""Manage Azure log analytics query pack."""
pass
__all__ = ["__CMDGroup"]
| data/azure-cli-2.57.0/azure/cli/command_modules/monitor/aaz/latest/monitor/log_analytics/query_pack/__cmd_group.py | 201 | 87 | 374,925 |
LOAD_CONST 0
LOAD_CONST ('run_once',)
IMPORT_NAME doit.tools
IMPORT_FROM run_once
STORE_NAME run_once
POP_TOP
LOAD_CONST <code object task_get_pylogo at 0x7fab540c5c00, file "f.py", line 4>
LOAD_CONST 'task_get_pylogo'
MAKE_FUNCTION
STORE_NAME task_get_pylogo
LOAD_CONST None
RETURN_VALUE
LOAD_CONST 'http://python.org... | from doit.tools import run_once
def task_get_pylogo():
url = "http://python.org/images/python-logo.gif"
return {
"actions": ["wget %s" % url],
"targets": ["python-logo.gif"],
"uptodate": [run_once],
}
| data/doit-0.36.0/doc/samples/download.py | 168 | 87 | 109,332 |
LOAD_CONST '\nThis module provide Python 3 compatability\n'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME sys
STORE_NAME sys
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME re
STORE_NAME re
SETUP_EXCEPT to 32
LOAD_NAME re
LOAD_ATTR Pattern
STORE_NAME re_pattern
POP_BLOCK
JUMP_FORWARD to 58
DUP_TOP
LOAD_NAME... | """
This module provide Python 3 compatability
"""
import sys
import re
try:
re_pattern = re.Pattern # >= Python 3.7
except AttributeError:
re_pattern = re._pattern_type
RecursionError = RuntimeError if sys.version_info < (3, 5) else RecursionError
| data/radish-bdd-0.17.1/radish/compat.py | 170 | 87 | 243,463 |
LOAD_CONST 0
LOAD_CONST ('annotations',)
IMPORT_NAME __future__
IMPORT_FROM annotations
STORE_NAME annotations
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME sys
STORE_NAME sys
LOAD_NAME sys
LOAD_ATTR version_info
LOAD_CONST (3, 11)
COMPARE_OP <
POP_JUMP_IF_FALSE
LOAD_CONST 0
LOAD_CONST ('NotRequired', 'Self', 'as... | from __future__ import annotations
import sys
if sys.version_info < (3, 11):
from typing_extensions import NotRequired, Self, assert_never
else:
from typing import NotRequired, Self, assert_never
__all__ = (
"assert_never",
"NotRequired",
"Self",
)
| data/cibuildwheel-2.16.5/cibuildwheel/_compat/typing.py | 192 | 87 | 155,633 |
LOAD_CONST 0
LOAD_CONST ('migrations',)
IMPORT_NAME django.db
IMPORT_FROM migrations
STORE_NAME migrations
POP_TOP
LOAD_BUILD_CLASS
LOAD_CONST <code object Migration at 0x7fab54330f60, file "f.py", line 4>
LOAD_CONST 'Migration'
MAKE_FUNCTION
LOAD_CONST 'Migration'
LOAD_NAME migrations
LOAD_ATTR Migration
CALL_FUNCTIO... | from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
("cornerstone", "0031_cornerstoneapirequestlogs"),
]
operations = [
migrations.DeleteModel(
name="CornerstoneAPIRequestLogs",
),
]
| data/edx-enterprise-4.12.1/integrated_channels/cornerstone/migrations/0032_delete_cornerstoneapirequestlogs.py | 187 | 87 | 91,745 |
LOAD_CONST 0
LOAD_CONST ('initializer',)
IMPORT_NAME google.cloud.aiplatform.training_utils.cloud_profiler
IMPORT_FROM initializer
STORE_NAME initializer
POP_TOP
LOAD_NAME initializer
LOAD_ATTR initialize
STORE_NAME init
LOAD_CONST None
RETURN_VALUE | from google.cloud.aiplatform.training_utils.cloud_profiler import initializer
"""
Initialize the cloud profiler for tensorflow.
Usage:
from google.cloud.aiplatform.training_utils import cloud_profiler
cloud_profiler.init(profiler='tensorflow')
"""
init = initializer.initialize
| data/google-cloud-aiplatform-1.42.1/google/cloud/aiplatform/training_utils/cloud_profiler/__init__.py | 55 | 87 | 14,016 |
LOAD_CONST 'Utilities to get elements of generated spec'
STORE_NAME __doc__
LOAD_CONST <code object get_definitions at 0x7f8e4475c9c0, file "f.py", line 4>
LOAD_CONST 'get_definitions'
MAKE_FUNCTION
STORE_NAME get_definitions
LOAD_CONST <code object get_paths at 0x7f8e4475ced0, file "f.py", line 10>
LOAD_CONST 'get_p... | """Utilities to get elements of generated spec"""
def get_definitions(spec):
if spec.openapi_version.major < 3:
return spec.to_dict()["definitions"]
return spec.to_dict()["components"]["schemas"]
def get_paths(spec):
return spec.to_dict()["paths"]
| data/apispec_webframeworks-1.0.0/tests/utils.py | 214 | 87 | 10,880 |
LOAD_CONST 'Holds data loaders for historical and online feature store.'
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('HistoricalFeatureStoreWriter',)
IMPORT_NAME butterfree.load.writers.historical_feature_store_writer
IMPORT_FROM HistoricalFeatureStoreWriter
STORE_NAME HistoricalFeatureStoreWriter
POP_TOP
LOAD_CONST ... | """Holds data loaders for historical and online feature store."""
from butterfree.load.writers.historical_feature_store_writer import (
HistoricalFeatureStoreWriter,
)
from butterfree.load.writers.online_feature_store_writer import OnlineFeatureStoreWriter
__all__ = ["HistoricalFeatureStoreWriter", "OnlineFeature... | data/butterfree-1.2.3/butterfree/load/writers/__init__.py | 144 | 87 | 9,032 |
LOAD_CONST 0
LOAD_CONST ('LangChainTracer', 'get_client', 'log_error_once', 'wait_for_all_tracers')
IMPORT_NAME langchain_core.tracers.langchain
IMPORT_FROM LangChainTracer
STORE_NAME LangChainTracer
IMPORT_FROM get_client
STORE_NAME get_client
IMPORT_FROM log_error_once
STORE_NAME log_error_once
IMPORT_FROM wait_for_a... | from langchain_core.tracers.langchain import (
LangChainTracer,
get_client,
log_error_once,
wait_for_all_tracers,
)
__all__ = ["log_error_once", "wait_for_all_tracers", "get_client", "LangChainTracer"]
| data/langchain-0.1.8/langchain/schema/callbacks/tracers/langchain.py | 160 | 87 | 369,138 |
LOAD_CONST 0
LOAD_CONST ('partial',)
IMPORT_NAME functools
IMPORT_FROM partial
STORE_NAME partial
POP_TOP
LOAD_CONST <code object func at 0x7fab800ab810, file "f.py", line 4>
LOAD_CONST 'func'
MAKE_FUNCTION
STORE_NAME func
LOAD_CONST <code object coroutinefunc at 0x7fab800ab390, file "f.py", line 8>
LOAD_CONST 'corou... | from functools import partial
def func():
pass
async def coroutinefunc():
pass
async def asyncgenerator():
yield
partial_func = partial(func)
partial_coroutinefunc = partial(coroutinefunc)
builtin_func = print
partial_builtin_func = partial(print)
| data/sphinx-7.2.6/tests/roots/test-ext-autodoc/target/functions.py | 253 | 87 | 381,174 |
LOAD_CONST 'Exception types that may be thrown by this library.'
STORE_NAME __doc__
LOAD_BUILD_CLASS
LOAD_CONST <code object DeserializationException at 0x7faa8c22a420, file "f.py", line 4>
LOAD_CONST 'DeserializationException'
MAKE_FUNCTION
LOAD_CONST 'DeserializationException'
LOAD_NAME Exception
CALL_FUNCTION
STORE... | """Exception types that may be thrown by this library."""
class DeserializationException(Exception):
"""Encountered an unexpected byte sequence or missing element type."""
class SessionException(Exception):
"""Session request exception."""
class CommandException(Exception):
"""Command result exception... | data/vici-5.9.11.post1/vici/exception.py | 446 | 87 | 444,691 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME sys
STORE_NAME sys
LOAD_NAME sys
LOAD_ATTR stdin
STORE_NAME fi
LOAD_NAME sys
LOAD_ATTR stdout
STORE_NAME fo
SETUP_LOOP to 104
LOAD_NAME fi
GET_ITER
FOR_ITER to 102
STORE_NAME line
LOAD_NAME line
LOAD_METHOD strip
LOAD_CONST '\n'
CALL_METHOD
STORE_NAME line
LOAD_NAME line
PO... | import sys
fi = sys.stdin
fo = sys.stdout
for line in fi:
line = line.strip("\n")
if not line:
fo.write("\n")
fields = line.split("\t")
fo.write("%s %s\n" % (" ".join(fields[1:]), fields[0]))
| data/python-crfsuite-0.9.10/crfsuite/bench/crfsuite_to_mallet.py | 179 | 87 | 215,221 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pathlib
STORE_NAME pathlib
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('RUNTIME_VARS',)
IMPORT_NAME tests.support.runtests
IMPORT_FROM RUNTIME_VARS
STORE_NAME RUNTIME_VARS
POP_TOP
LOAD_NAME pytest
LOAD_ATTR fixture
LOAD_CONST 'se... | import pathlib
import pytest
from tests.support.runtests import RUNTIME_VARS
@pytest.fixture(scope="session", autouse=True)
def _create_old_tempdir():
pathlib.Path(RUNTIME_VARS.TMP).mkdir(exist_ok=True, parents=True)
| data/salt-3006.6/tests/unit/conftest.py | 218 | 87 | 96,817 |
LOAD_CONST 'Country.'
STORE_NAME __doc__
LOAD_CONST 1
LOAD_CONST ('es_cif', 'es_doi', 'es_nie', 'es_nif')
IMPORT_NAME es
IMPORT_FROM es_cif
STORE_NAME es_cif
IMPORT_FROM es_doi
STORE_NAME es_doi
IMPORT_FROM es_nie
STORE_NAME es_nie
IMPORT_FROM es_nif
STORE_NAME es_nif
POP_TOP
LOAD_CONST 1
LOAD_CONST ('fi_business_id'... | """Country."""
from .es import es_cif, es_doi, es_nie, es_nif
from .fi import fi_business_id, fi_ssn
__all__ = ("fi_business_id", "fi_ssn", "es_cif", "es_doi", "es_nie", "es_nif")
| data/validators-0.22.0/src/validators/i18n/__init__.py | 191 | 87 | 406,051 |
LOAD_CONST "Dataset definition for tedlium.\n\nDEPRECATED!\nIf you want to use the Tedlium dataset builder class, use:\ntfds.builder_cls('tedlium')\n"
STORE_NAME __doc__
LOAD_CONST 0
LOAD_CONST ('lazy_builder_import',)
IMPORT_NAME tensorflow_datasets.core
IMPORT_FROM lazy_builder_import
STORE_NAME lazy_builder_import
... | """Dataset definition for tedlium.
DEPRECATED!
If you want to use the Tedlium dataset builder class, use:
tfds.builder_cls('tedlium')
"""
from tensorflow_datasets.core import lazy_builder_import
Tedlium = lazy_builder_import.LazyBuilderImport("tedlium")
| data/tfds-nightly-4.9.4.dev202402210044/tensorflow_datasets/audio/tedlium.py | 130 | 87 | 238,220 |
LOAD_CONST 0
LOAD_CONST ('IO',)
IMPORT_NAME typing
IMPORT_FROM IO
STORE_NAME IO
POP_TOP
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME yaml
STORE_NAME yaml
LOAD_CONST 0
LOAD_CONST ('ConfigurationMechanism',)
IMPORT_NAME datahub.configuration
IMPORT_FROM ConfigurationMechanism
STORE_NAME ConfigurationMechanism
POP_TOP
LOAD... | from typing import IO
import yaml
from datahub.configuration import ConfigurationMechanism
class YamlConfigurationMechanism(ConfigurationMechanism):
"""Ability to load configuration from yaml files"""
def load_config(self, config_fp: IO) -> dict:
return yaml.safe_load(config_fp)
| data/acryl-datahub-0.12.1.5/src/datahub/configuration/yaml.py | 300 | 87 | 435,291 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_NAME pytest
LOAD_ATTR fixture
LOAD_CONST <code object spam at 0x7f8ab7985c90, file "f.py", line 4>
LOAD_CONST 'spam'
MAKE_FUNCTION
CALL_FUNCTION
STORE_NAME spam
LOAD_BUILD_CLASS
LOAD_CONST <code object TestSpam at 0x7f8ab7617660, file "f.py", line... | import pytest
@pytest.fixture
def spam():
return "spam"
class TestSpam:
@pytest.fixture
def spam(self, spam):
return spam * 2
def test_spam(self, spam):
assert spam == "spamspam"
| data/pytest-8.0.1/testing/example_scripts/fixtures/fill_fixtures/test_extend_fixture_module_class.py | 335 | 87 | 427,690 |
LOAD_CONST 'Exceptions for marking tests as skipped or deprecated.\n\nThis module exists to provide backwards compatibility with previous\nversions of nose where skipped and deprecated tests were core\nfunctionality, rather than being provided by plugins. It may be\nremoved in a future release.\n'
STORE_NAME __doc__
L... | """Exceptions for marking tests as skipped or deprecated.
This module exists to provide backwards compatibility with previous
versions of nose where skipped and deprecated tests were core
functionality, rather than being provided by plugins. It may be
removed in a future release.
"""
from nose.plugins.skip import Ski... | data/nose-1.3.7/nose/exc.py | 141 | 87 | 401,687 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_reduce_sum_int32_bool_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_sum_int32_bool_64.py | 228 | 87 | 221,381 |
LOAD_CONST 0
LOAD_CONST ('JsonResponse',)
IMPORT_NAME django.http
IMPORT_FROM JsonResponse
STORE_NAME JsonResponse
POP_TOP
LOAD_CONST 0
LOAD_CONST ('require_POST',)
IMPORT_NAME django.views.decorators.http
IMPORT_FROM require_POST
STORE_NAME require_POST
POP_TOP
LOAD_CONST 0
LOAD_CONST ('fmt_sql',)
IMPORT_NAME explor... | from django.http import JsonResponse
from django.views.decorators.http import require_POST
from explorer.utils import fmt_sql
@require_POST
def format_sql(request):
sql = request.POST.get("sql", "")
formatted = fmt_sql(sql)
return JsonResponse({"formatted": formatted})
| data/django-sql-explorer-4.0.2/explorer/views/format_sql.py | 213 | 87 | 386,217 |
LOAD_CONST 0
LOAD_CONST ('App', 'ComposeResult')
IMPORT_NAME textual.app
IMPORT_FROM App
STORE_NAME App
IMPORT_FROM ComposeResult
STORE_NAME ComposeResult
POP_TOP
LOAD_CONST 0
LOAD_CONST ('Header', 'Footer')
IMPORT_NAME textual.widgets
IMPORT_FROM Header
STORE_NAME Header
IMPORT_FROM Footer
STORE_NAME Footer
POP_TOP
... | from textual.app import App, ComposeResult
from textual.widgets import Header, Footer
class ExampleApp(App):
def compose(self) -> ComposeResult:
yield Header()
yield Footer()
if __name__ == "__main__":
app = ExampleApp()
app.run()
| data/textual-0.52.1/docs/examples/guide/dom2.py | 297 | 87 | 26,368 |
LOAD_CONST 0
LOAD_CONST ('Attacker',)
IMPORT_NAME allennlp.interpret.attackers.attacker
IMPORT_FROM Attacker
STORE_NAME Attacker
POP_TOP
LOAD_CONST 0
LOAD_CONST ('SaliencyInterpreter',)
IMPORT_NAME allennlp.interpret.saliency_interpreters.saliency_interpreter
IMPORT_FROM SaliencyInterpreter
STORE_NAME SaliencyInterpre... | from allennlp.interpret.attackers.attacker import Attacker
from allennlp.interpret.saliency_interpreters.saliency_interpreter import (
SaliencyInterpreter,
)
from allennlp.interpret.influence_interpreters.influence_interpreter import (
InfluenceInterpreter,
)
| data/allennlp-2.10.1/allennlp/interpret/__init__.py | 150 | 87 | 82,790 |
LOAD_CONST 0
LOAD_CONST ('annotations',)
IMPORT_NAME __future__
IMPORT_FROM annotations
STORE_NAME annotations
POP_TOP
LOAD_CONST 0
LOAD_CONST ('TypedDict',)
IMPORT_NAME typing_extensions
IMPORT_FROM TypedDict
STORE_NAME TypedDict
POP_TOP
LOAD_CONST 'JobListParams'
BUILD_LIST
STORE_NAME __all__
LOAD_BUILD_CLASS
LOAD... | from __future__ import annotations
from typing_extensions import TypedDict
__all__ = ["JobListParams"]
class JobListParams(TypedDict, total=False):
after: str
"""Identifier for the last job from the previous pagination request."""
limit: int
"""Number of fine-tuning jobs to retrieve."""
| data/openai-1.12.0/src/openai/types/fine_tuning/job_list_params.py | 224 | 87 | 63,729 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME os
STORE_NAME os
LOAD_CONST 0
LOAD_CONST ('Path',)
IMPORT_NAME pathlib
IMPORT_FROM Path
STORE_NAME Path
POP_TOP
LOAD_CONST 'file://'
LOAD_NAME str
LOAD_NAME Path
LOAD_NAME Path
LOAD_NAME os
LOAD_ATTR path
LOAD_METHOD realpath
LOAD_NAME __file__
CALL_METHOD
CALL_FUNCTION
LOAD_... | import os
from pathlib import Path
docs_path = "file://" + str(
Path(Path(os.path.realpath(__file__)).parent, "html", "index.html").absolute()
)
try:
import webbrowser
webbrowser.open(docs_path)
except ImportError:
print(docs_path)
| data/benchling_sdk-1.10.0/benchling_sdk/docs/__main__.py | 203 | 87 | 157,240 |
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME ctypes
STORE_NAME ctypes
LOAD_CONST 0
LOAD_CONST None
IMPORT_NAME pytest
STORE_NAME pytest
LOAD_CONST 0
LOAD_CONST ('lib',)
IMPORT_NAME awkward_cpp.cpu_kernels
IMPORT_FROM lib
STORE_NAME lib
POP_TOP
LOAD_NAME pytest
LOAD_ATTR mark
LOAD_ATTR skip
LOAD_CONST 'Unable to generate... | import ctypes
import pytest
from awkward_cpp.cpu_kernels import lib
@pytest.mark.skip(reason="Unable to generate any tests for kernel")
def test_cpuawkward_reduce_countnonzero_int16_64_1():
raise NotImplementedError("Unable to generate any tests for kernel")
| data/awkward-cpp-29/tests-cpu-kernels/test_cpuawkward_reduce_countnonzero_int16_64.py | 229 | 87 | 221,313 |
LOAD_CONST '#6d2f20'
STORE_NAME COPPER
LOAD_CONST '#b75347'
STORE_NAME RED
LOAD_CONST '#df7e66'
STORE_NAME SALMON
LOAD_CONST '#e09351'
STORE_NAME ORANGE
LOAD_CONST '#edc775'
STORE_NAME YELLOW
LOAD_CONST '#94b594'
STORE_NAME GREEN
LOAD_CONST '#224b5e'
STORE_NAME BLUE
LOAD_CONST None
RETURN_VALUE | COPPER = "#6d2f20"
RED = "#b75347"
SALMON = "#df7e66"
ORANGE = "#e09351"
YELLOW = "#edc775"
GREEN = "#94b594"
BLUE = "#224b5e"
| data/gluonts-0.14.4/src/gluonts/meta/colors.py | 111 | 87 | 140,192 |
LOAD_CONST 'default'
LOAD_CONST 'ENGINE'
LOAD_CONST 'django.db.backends.sqlite3'
BUILD_MAP
BUILD_MAP
STORE_NAME DATABASES
LOAD_CONST 'tests.django'
BUILD_LIST
STORE_NAME INSTALLED_APPS
LOAD_CONST 'tests.django.urls'
STORE_NAME ROOT_URLCONF
LOAD_CONST 'commenter_tests_secret_key'
STORE_NAME SECRET_KEY
LOAD_CONST Non... | DATABASES = {
"default": {
"ENGINE": "django.db.backends.sqlite3",
},
}
INSTALLED_APPS = ["tests.django"]
ROOT_URLCONF = "tests.django.urls"
SECRET_KEY = "commenter_tests_secret_key"
| data/google-cloud-sqlcommenter-2.0.0/tests/django/settings.py | 100 | 87 | 300,697 |
LOAD_CONST ('__version__', 'get_version', 'SingleSourceError', 'VersionNotFoundError')
STORE_NAME __all__
LOAD_CONST 0
LOAD_CONST ('Path',)
IMPORT_NAME pathlib
IMPORT_FROM Path
STORE_NAME Path
POP_TOP
LOAD_CONST 1
LOAD_CONST ('SingleSourceError', 'VersionNotFoundError')
IMPORT_NAME errors
IMPORT_FROM SingleSourceErro... | __all__ = (
"__version__",
"get_version",
"SingleSourceError",
"VersionNotFoundError",
)
from pathlib import Path
from .errors import SingleSourceError, VersionNotFoundError
from .version import get_version
__version__ = get_version(__name__, Path(__file__).parent.parent)
| data/single-source-0.3.0/single_source/__init__.py | 159 | 87 | 157,224 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.