language
stringclasses
1 value
repo
stringclasses
346 values
path
stringlengths
6
201
class_span
dict
source
stringlengths
21
2.38M
target
stringlengths
1
96
python
kamyu104__LeetCode-Solutions
Python/minimum-relative-loss-after-buying-chocolates.py
{ "start": 84, "end": 985 }
class ____(object): def minimumRelativeLosses(self, prices, queries): """ :type prices: List[int] :type queries: List[List[int]] :rtype: List[int] """ def binary_search(left, right, check): while left <= right: mid = left + (right-left)//2 if check(mid): right = mid-1 else: left = mid+1 return left prices.sort() prefix = [0]*(len(prices)+1) for i in xrange(len(prices)): prefix[i+1] = prefix[i]+prices[i] result = [] for k, m in queries: cnt = binary_search(0, m-1, lambda x: k-(prices[-(m-x)]-k) <= prices[(x+1)-1]-0) a = prefix[-1]-prefix[-1-(m-cnt)]-(m-cnt)*k b = prefix[cnt]+(m-cnt)*k result.append(b-a) return result
Solution
python
sqlalchemy__sqlalchemy
lib/sqlalchemy/sql/crud.py
{ "start": 45948, "end": 61090 }
class ____(elements.ColumnElement[Any]): _is_multiparam_column = True def __init__(self, original, index): self.index = index self.key = "%s_m%d" % (original.key, index + 1) self.original = original self.default = original.default self.type = original.type def compare(self, other, **kw): raise NotImplementedError() def _copy_internals(self, **kw): raise NotImplementedError() def __eq__(self, other): return ( isinstance(other, _multiparam_column) and other.key == self.key and other.original == self.original ) @util.memoized_property def _default_description_tuple(self) -> _DefaultDescriptionTuple: """used by default.py -> _process_execute_defaults()""" return _DefaultDescriptionTuple._from_column_default(self.default) @util.memoized_property def _onupdate_description_tuple(self) -> _DefaultDescriptionTuple: """used by default.py -> _process_execute_defaults()""" return _DefaultDescriptionTuple._from_column_default(self.onupdate) def _process_multiparam_default_bind( compiler: SQLCompiler, stmt: ValuesBase, c: KeyedColumnElement[Any], index: int, kw: Dict[str, Any], ) -> str: if not c.default: raise exc.CompileError( "INSERT value for column %s is explicitly rendered as a bound" "parameter in the VALUES clause; " "a Python-side value or SQL expression is required" % c ) elif default_is_clause_element(c.default): return compiler.process(c.default.arg.self_group(), **kw) elif c.default.is_sequence: # these conditions would have been established # by append_param_insert_(?:hasdefault|pk_returning|pk_no_returning) # in order for us to be here, so these don't need to be # checked # assert compiler.dialect.supports_sequences and ( # not c.default.optional # or not compiler.dialect.sequences_optional # ) return compiler.process(c.default, **kw) else: col = _multiparam_column(c, index) assert isinstance(stmt, dml.Insert) return _create_insert_prefetch_bind_param( compiler, col, process=True, **kw ) def _get_update_multitable_params( compiler, stmt, compile_state, stmt_parameter_tuples, check_columns, _col_bind_name, _getattr_col_key, values, kw, ): normalized_params = { coercions.expect(roles.DMLColumnRole, c): param for c, param in stmt_parameter_tuples or () } include_table = compile_state.include_table_with_column_exprs affected_tables = set() for t in compile_state._extra_froms: # extra gymnastics to support the probably-shouldnt-have-supported # case of "UPDATE table AS alias SET table.foo = bar", but it's # supported we_shouldnt_be_here_if_columns_found = ( not include_table and not compile_state.dml_table.is_derived_from(t) ) for c in t.c: if c in normalized_params: if we_shouldnt_be_here_if_columns_found: raise exc.CompileError( "Backend does not support additional tables " "in the SET " "clause; cannot include columns from table(s) " f"'{t.description}' in " "SET clause" ) affected_tables.add(t) check_columns[_getattr_col_key(c)] = c value = normalized_params[c] col_value = compiler.process(c, include_table=include_table) if coercions._is_literal(value): value = _create_bind_param( compiler, c, value, required=value is REQUIRED, name=_col_bind_name(c), **kw, # TODO: no test coverage for literal binds here ) accumulated_bind_names: Iterable[str] = (c.key,) elif value._is_bind_parameter: cbn = _col_bind_name(c) value = _handle_values_anonymous_param( compiler, c, value, name=cbn, **kw ) accumulated_bind_names = (cbn,) else: compiler.postfetch.append(c) value = compiler.process(value.self_group(), **kw) accumulated_bind_names = () values.append((c, col_value, value, accumulated_bind_names)) # determine tables which are actually to be updated - process onupdate # and server_onupdate for these for t in affected_tables: for c in t.c: if c in normalized_params: continue elif c.onupdate is not None and not c.onupdate.is_sequence: if c.onupdate.is_clause_element: values.append( ( c, compiler.process(c, include_table=include_table), compiler.process( c.onupdate.arg.self_group(), **kw ), (), ) ) compiler.postfetch.append(c) else: values.append( ( c, compiler.process(c, include_table=include_table), _create_update_prefetch_bind_param( compiler, c, name=_col_bind_name(c), **kw ), (c.key,), ) ) elif c.server_onupdate is not None: compiler.postfetch.append(c) def _extend_values_for_multiparams( compiler: SQLCompiler, stmt: ValuesBase, compile_state: DMLState, initial_values: Sequence[_CrudParamElementStr], _column_as_key: Callable[..., str], kw: Dict[str, Any], ) -> List[Sequence[_CrudParamElementStr]]: values_0 = initial_values values = [initial_values] has_visiting_cte = kw.get("visiting_cte") is not None mp = compile_state._multi_parameters assert mp is not None for i, row in enumerate(mp[1:]): extension: List[_CrudParamElementStr] = [] row = {_column_as_key(key): v for key, v in row.items()} for col, col_expr, param, accumulated_names in values_0: if col.key in row: key = col.key if coercions._is_literal(row[key]): new_param = _create_bind_param( compiler, col, row[key], name=("%s_m%d" % (col.key, i + 1)), force_anonymous=has_visiting_cte, **kw, ) else: new_param = compiler.process(row[key].self_group(), **kw) else: new_param = _process_multiparam_default_bind( compiler, stmt, col, i, kw ) extension.append((col, col_expr, new_param, accumulated_names)) values.append(extension) return values def _get_stmt_parameter_tuples_params( compiler, compile_state, parameters, stmt_parameter_tuples, _column_as_key, values, kw, ): for k, v in stmt_parameter_tuples: colkey = _column_as_key(k) if colkey is not None: parameters.setdefault(colkey, v) else: # a non-Column expression on the left side; # add it to values() in an "as-is" state, # coercing right side to bound param # note one of the main use cases for this is array slice # updates on PostgreSQL, as the left side is also an expression. col_expr = compiler.process( k, include_table=compile_state.include_table_with_column_exprs ) if coercions._is_literal(v): v = compiler.process( elements.BindParameter(None, v, type_=k.type), **kw ) else: if v._is_bind_parameter and v.type._isnull: # either unique parameter, or other bound parameters that # were passed in directly # set type to that of the column unconditionally v = v._with_binary_element_type(k.type) v = compiler.process(v.self_group(), **kw) # TODO: not sure if accumulated_bind_names applies here values.append((k, col_expr, v, ())) def _get_returning_modifiers(compiler, stmt, compile_state, toplevel): """determines RETURNING strategy, if any, for the statement. This is where it's determined what we need to fetch from the INSERT or UPDATE statement after it's invoked. """ dialect = compiler.dialect need_pks = ( toplevel and _compile_state_isinsert(compile_state) and not stmt._inline and ( not compiler.for_executemany or (dialect.insert_executemany_returning and stmt._return_defaults) ) and not stmt._returning # and (not stmt._returning or stmt._return_defaults) and not compile_state._has_multi_parameters ) # check if we have access to simple cursor.lastrowid. we can use that # after the INSERT if that's all we need. postfetch_lastrowid = ( need_pks and dialect.postfetch_lastrowid and stmt.table._autoincrement_column is not None ) # see if we want to add RETURNING to an INSERT in order to get # primary key columns back. This would be instead of postfetch_lastrowid # if that's set. implicit_returning = ( # statement itself can veto it need_pks # the dialect can veto it if it just doesnt support RETURNING # with INSERT and dialect.insert_returning # user-defined implicit_returning on Table can veto it and compile_state._primary_table.implicit_returning # the compile_state can veto it (SQlite uses this to disable # RETURNING for an ON CONFLICT insert, as SQLite does not return # for rows that were updated, which is wrong) and compile_state._supports_implicit_returning and ( # since we support MariaDB and SQLite which also support lastrowid, # decide if we should use lastrowid or RETURNING. for insert # that didnt call return_defaults() and has just one set of # parameters, we can use lastrowid. this is more "traditional" # and a lot of weird use cases are supported by it. # SQLite lastrowid times 3x faster than returning, # Mariadb lastrowid 2x faster than returning (not postfetch_lastrowid or dialect.favor_returning_over_lastrowid) or compile_state._has_multi_parameters or stmt._return_defaults ) ) if implicit_returning: postfetch_lastrowid = False if _compile_state_isinsert(compile_state): should_implicit_return_defaults = ( implicit_returning and stmt._return_defaults ) explicit_returning = ( should_implicit_return_defaults or stmt._returning or stmt._supplemental_returning ) use_insertmanyvalues = ( toplevel and compiler.for_executemany and dialect.use_insertmanyvalues and ( explicit_returning or dialect.use_insertmanyvalues_wo_returning ) ) use_sentinel_columns = None if ( use_insertmanyvalues and explicit_returning and stmt._sort_by_parameter_order ): use_sentinel_columns = compiler._get_sentinel_column_for_table( stmt.table ) elif compile_state.isupdate: should_implicit_return_defaults = ( stmt._return_defaults and compile_state._primary_table.implicit_returning and compile_state._supports_implicit_returning and dialect.update_returning ) use_insertmanyvalues = False use_sentinel_columns = None elif compile_state.isdelete: should_implicit_return_defaults = ( stmt._return_defaults and compile_state._primary_table.implicit_returning and compile_state._supports_implicit_returning and dialect.delete_returning ) use_insertmanyvalues = False use_sentinel_columns = None else: should_implicit_return_defaults = False # pragma: no cover use_insertmanyvalues = False use_sentinel_columns = None if should_implicit_return_defaults: if not stmt._return_defaults_columns: # TODO: this is weird. See #9685 where we have to # take an extra step to prevent this from happening. why # would this ever be *all* columns? but if we set to blank, then # that seems to break things also in the ORM. So we should # try to clean this up and figure out what return_defaults # needs to do w/ the ORM etc. here implicit_return_defaults = set(stmt.table.c) else: implicit_return_defaults = set(stmt._return_defaults_columns) else: implicit_return_defaults = None return ( need_pks, implicit_returning or should_implicit_return_defaults, implicit_return_defaults, postfetch_lastrowid, use_insertmanyvalues, use_sentinel_columns, ) def _warn_pk_with_no_anticipated_value(c): msg = ( "Column '%s.%s' is marked as a member of the " "primary key for table '%s', " "but has no Python-side or server-side default generator indicated, " "nor does it indicate 'autoincrement=True' or 'nullable=True', " "and no explicit value is passed. " "Primary key columns typically may not store NULL." % (c.table.fullname, c.name, c.table.fullname) ) if len(c.table.primary_key) > 1: msg += ( " Note that as of SQLAlchemy 1.1, 'autoincrement=True' must be " "indicated explicitly for composite (e.g. multicolumn) primary " "keys if AUTO_INCREMENT/SERIAL/IDENTITY " "behavior is expected for one of the columns in the primary key. " "CREATE TABLE statements are impacted by this change as well on " "most backends." ) util.warn(msg)
_multiparam_column
python
ansible__ansible
test/lib/ansible_test/_internal/host_configs.py
{ "start": 7843, "end": 8419 }
class ____(PosixConfig): """Configuration for a POSIX SSH host.""" user: t.Optional[str] = None host: t.Optional[str] = None port: t.Optional[int] = None def get_defaults(self, context: HostContext) -> PosixSshCompletionConfig: """Return the default settings.""" return PosixSshCompletionConfig( user=self.user, host=self.host, ) @property def have_root(self) -> bool: """True if root is available, otherwise False.""" return self.user == 'root' @dataclasses.dataclass
PosixSshConfig
python
airbytehq__airbyte
airbyte-integrations/connectors/source-zendesk-support/unit_tests/integrations/zs_responses/articles_response_builder.py
{ "start": 384, "end": 715 }
class ____(HttpResponseBuilder): @classmethod def response(cls, next_page_url: Optional[HttpRequest] = None) -> "ArticlesResponseBuilder": return cls( find_template("articles", __file__), FieldPath("articles"), NextPagePaginationStrategy(http_request_to_str(next_page_url)) )
ArticlesResponseBuilder
python
keras-team__keras
keras/src/ops/math_test.py
{ "start": 40354, "end": 40847 }
class ____(testing.TestCase): def test_in_top_k_call(self): targets = np.array([2, 0, 1], dtype=np.int32) predictions = np.array( [[0.1, 0.2, 0.7], [1.0, 0.2, 0.3], [0.2, 0.6, 0.2]], dtype=np.float32, ) k = 2 in_top_k_op = kmath.InTopK(k=k) output = in_top_k_op.call(targets, predictions) expected_output = np.array([True, True, True], dtype=bool) self.assertAllEqual(output, expected_output)
InTopKTest
python
Lightning-AI__lightning
src/lightning/pytorch/callbacks/batch_size_finder.py
{ "start": 1111, "end": 8912 }
class ____(Callback): """Finds the largest batch size supported by a given model before encountering an out of memory (OOM) error. All you need to do is add it as a callback inside Trainer and call ``trainer.{fit,validate,test,predict}``. Internally, it calls the respective step function ``steps_per_trial`` times for each batch size until one of the batch sizes generates an OOM error. .. warning:: This is an :ref:`experimental <versioning:Experimental API>` feature. Args: mode: search strategy to update the batch size: - ``'power'``: Keep multiplying the batch size by 2, until we get an OOM error. - ``'binsearch'``: Initially keep multiplying by 2 and after encountering an OOM error do a binary search between the last successful batch size and the batch size that failed. steps_per_trial: number of steps to run with a given batch size. Ideally 1 should be enough to test if an OOM error occurs, however in practice a few are needed. init_val: initial batch size to start the search with. max_trials: max number of increases in batch size done before algorithm is terminated batch_arg_name: name of the attribute that stores the batch size. It is expected that the user has provided a model or datamodule that has a hyperparameter with that name. We will look for this attribute name in the following places - ``model`` - ``model.hparams`` - ``trainer.datamodule`` (the datamodule passed to the tune method) margin: Margin to reduce the found batch size by to provide a safety buffer. Only applied when using 'binsearch' mode. Should be a float between 0 and 1. Defaults to 0.05 (5% reduction). max_val: Maximum batch size limit, defaults to 8192. Helps prevent testing unrealistically large or inefficient batch sizes (e.g., 2**25) when running on CPU or when automatic OOM detection is not available. Example:: # 1. Customize the BatchSizeFinder callback to run at different epochs. This feature is # useful while fine-tuning models since you can't always use the same batch size after # unfreezing the backbone. from lightning.pytorch.callbacks import BatchSizeFinder class FineTuneBatchSizeFinder(BatchSizeFinder): def __init__(self, milestones, *args, **kwargs): super().__init__(*args, **kwargs) self.milestones = milestones def on_fit_start(self, *args, **kwargs): return def on_train_epoch_start(self, trainer, pl_module): if trainer.current_epoch in self.milestones or trainer.current_epoch == 0: self.scale_batch_size(trainer, pl_module) trainer = Trainer(callbacks=[FineTuneBatchSizeFinder(milestones=(5, 10))]) trainer.fit(...) Example:: # 2. Run batch size finder for validate/test/predict. from lightning.pytorch.callbacks import BatchSizeFinder class EvalBatchSizeFinder(BatchSizeFinder): def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) def on_fit_start(self, *args, **kwargs): return def on_test_start(self, trainer, pl_module): self.scale_batch_size(trainer, pl_module) trainer = Trainer(callbacks=[EvalBatchSizeFinder()]) trainer.test(...) """ SUPPORTED_MODES = ("power", "binsearch") def __init__( self, mode: str = "power", steps_per_trial: int = 3, init_val: int = 2, max_trials: int = 25, batch_arg_name: str = "batch_size", margin: float = 0.05, max_val: int = 8192, ) -> None: mode = mode.lower() if mode not in self.SUPPORTED_MODES: raise ValueError(f"`mode` should be either of {self.SUPPORTED_MODES}") assert 0.0 <= margin < 1.0, f"`margin` should be between 0 and 1. Found {margin=}" self.optimal_batch_size: Optional[int] = init_val self._mode = mode self._steps_per_trial = steps_per_trial self._init_val = init_val self._max_trials = max_trials self._batch_arg_name = batch_arg_name self._margin = margin self._max_val = max_val self._early_exit = False @override def setup(self, trainer: "pl.Trainer", pl_module: "pl.LightningModule", stage: Optional[str] = None) -> None: if trainer._accelerator_connector.is_distributed: raise MisconfigurationException("The Batch size finder is not supported with distributed strategies.") # TODO: check if this can be enabled (#4040) if not trainer.fit_loop._data_source.is_module(): raise MisconfigurationException( "The Batch size finder cannot be used with dataloaders passed directly to `.fit()`. Please disable" " the feature or incorporate the dataloader into your LightningModule or LightningDataModule." ) # TODO: Add support for multiple eval dataloader if stage != "fit": loop = trainer._active_loop assert loop is not None loop.setup_data() combined_loader = loop._combined_loader assert combined_loader is not None if len(combined_loader.flattened) > 1: stage = trainer.state.stage assert stage is not None raise MisconfigurationException( f"The Batch size finder cannot be used with multiple {stage.dataloader_prefix} dataloaders." ) if not lightning_hasattr(pl_module, self._batch_arg_name): raise MisconfigurationException( f"Field {self._batch_arg_name} not found in `model`, `datamodule`, nor their `hparams` attributes." ) if ( hasattr(pl_module, self._batch_arg_name) and hasattr(pl_module, "hparams") and self._batch_arg_name in pl_module.hparams ): rank_zero_warn( f"Field `model.{self._batch_arg_name}` and `model.hparams.{self._batch_arg_name}` are mutually" f" exclusive! `model.{self._batch_arg_name}` will be used as the initial batch size for scaling." " If this is not the intended behavior, please remove either one." ) def scale_batch_size(self, trainer: "pl.Trainer", pl_module: "pl.LightningModule") -> None: new_size = _scale_batch_size( trainer, self._mode, self._steps_per_trial, self._init_val, self._max_trials, self._batch_arg_name, self._margin, self._max_val, ) self.optimal_batch_size = new_size if self._early_exit: raise _TunerExitException() @override def on_fit_start(self, trainer: "pl.Trainer", pl_module: "pl.LightningModule") -> None: self.scale_batch_size(trainer, pl_module) @override def on_validation_start(self, trainer: "pl.Trainer", pl_module: "pl.LightningModule") -> None: if trainer.sanity_checking or trainer.state.fn != "validate": return self.scale_batch_size(trainer, pl_module) @override def on_test_start(self, trainer: "pl.Trainer", pl_module: "pl.LightningModule") -> None: self.scale_batch_size(trainer, pl_module) @override def on_predict_start(self, trainer: "pl.Trainer", pl_module: "pl.LightningModule") -> None: self.scale_batch_size(trainer, pl_module)
BatchSizeFinder
python
google__jax
jax/_src/interpreters/partial_eval.py
{ "start": 5706, "end": 5778 }
class ____: parents : list[Tracer] recipe : JaxprEqnRecipe
EffectHandle
python
sqlalchemy__sqlalchemy
test/orm/test_events.py
{ "start": 111062, "end": 115511 }
class ____( RemoveORMEventsGlobally, _fixtures.FixtureTest, AssertsCompiledSQL, testing.AssertsExecutionResults, ): __dialect__ = "default" @classmethod def setup_mappers(cls): User = cls.classes.User users = cls.tables.users cls.mapper_registry.map_imperatively(User, users) def test_before_compile(self): @event.listens_for(query.Query, "before_compile", retval=True) def no_deleted(query): for desc in query.column_descriptions: if desc["type"] is User: entity = desc["expr"] query = query.filter(entity.id != 10) return query User = self.classes.User s = fixture_session() q = s.query(User).filter_by(id=7) self.assert_compile( q, "SELECT users.id AS users_id, users.name AS users_name " "FROM users " "WHERE users.id = :id_1 AND users.id != :id_2", checkparams={"id_2": 10, "id_1": 7}, ) def test_before_compile_no_retval(self): counter = [0] @event.listens_for(query.Query, "before_compile") def count(query): counter[0] += 1 User = self.classes.User s = fixture_session() q = s.query(User).filter_by(id=7) str(q) str(q) eq_(counter, [2]) def test_alters_entities(self): User = self.classes.User @event.listens_for(query.Query, "before_compile", retval=True) def fn(query): return query.add_columns(User.name) s = fixture_session() q = s.query(User.id).filter_by(id=7) self.assert_compile( q, "SELECT users.id AS users_id, users.name AS users_name " "FROM users " "WHERE users.id = :id_1", checkparams={"id_1": 7}, ) eq_(q.all(), [(7, "jack")]) def test_before_compile_update(self): @event.listens_for(query.Query, "before_compile_update", retval=True) def no_deleted(query, update_context): assert update_context.query is query for desc in query.column_descriptions: if desc["type"] is User: entity = desc["expr"] query = query.filter(entity.id != 10) update_context.values["name"] = ( update_context.values["name"] + "_modified" ) return query User = self.classes.User s = fixture_session() with self.sql_execution_asserter() as asserter: s.query(User).filter_by(id=7).update({"name": "ed"}) asserter.assert_( CompiledSQL( "UPDATE users SET name=:name WHERE " "users.id = :id_1 AND users.id != :id_2", [{"name": "ed_modified", "id_1": 7, "id_2": 10}], ) ) def test_before_compile_delete(self): @event.listens_for(query.Query, "before_compile_delete", retval=True) def no_deleted(query, delete_context): assert delete_context.query is query for desc in query.column_descriptions: if desc["type"] is User: entity = desc["expr"] query = query.filter(entity.id != 10) return query User = self.classes.User s = fixture_session() # note this deletes no rows with self.sql_execution_asserter() as asserter: s.query(User).filter_by(id=10).delete() asserter.assert_( CompiledSQL( "DELETE FROM users WHERE " "users.id = :id_1 AND users.id != :id_2", [{"id_1": 10, "id_2": 10}], ) ) def test_before_compile_execution_options(self): User = self.classes.User @event.listens_for(query.Query, "before_compile", retval=True) def _before_compile(query): return query.execution_options(my_option=True) opts = {} @event.listens_for(testing.db, "before_cursor_execute") def _before_cursor_execute( conn, cursor, statement, parameters, context, executemany ): opts.update(context.execution_options) sess = fixture_session() sess.query(User).first() eq_(opts["my_option"], True)
QueryEventsTest
python
spyder-ide__spyder
spyder/plugins/help/widgets.py
{ "start": 4928, "end": 6782 }
class ____(QWidget, SpyderWidgetMixin): """ WebView widget with find dialog """ sig_link_clicked = Signal(QUrl) def __init__(self, parent): if not PYSIDE2: super().__init__(parent, class_parent=parent) else: QWidget.__init__(self, parent) SpyderWidgetMixin.__init__(self, class_parent=parent) from qtpy.QtWebEngineWidgets import QWebEnginePage from spyder.widgets.browser import FrameWebView self.webview = FrameWebView(self) self.webview.setup() if WEBENGINE: self.webview.web_widget.page().setBackgroundColor( QColor(MAIN_BG_COLOR)) else: self.webview.web_widget.setStyleSheet( "background:{}".format(MAIN_BG_COLOR)) self.webview.page().setLinkDelegationPolicy( QWebEnginePage.DelegateAllLinks) self.find_widget = FindReplace(self) self.find_widget.set_editor(self.webview.web_widget) self.find_widget.hide() # Layout layout = QVBoxLayout() layout.setSpacing(0) layout.setContentsMargins(0, 0, 0, 0) layout.addWidget(self.webview) layout.addWidget(self.find_widget) self.setLayout(layout) # Signals self.webview.linkClicked.connect(self.sig_link_clicked) def set_font(self, font, fixed_font=None): """Set font""" self.webview.set_font(font, fixed_font=fixed_font) def set_html(self, html_text, base_url): """Set html text""" self.webview.setHtml(html_text, base_url) def load_url(self, url): if isinstance(url, QUrl): qurl = url else: qurl = QUrl(url) self.webview.load(qurl) def clear(self): self.set_html('', self.webview.url())
RichText
python
airbytehq__airbyte
airbyte-integrations/connectors/source-kyriba/source_kyriba/source.py
{ "start": 8696, "end": 10435 }
class ____(IncrementalKyribaStream): primary_key = "uuid" def path(self, **kwargs) -> str: return "cash-flows" def stream_slices( self, cursor_field: List[str] = None, stream_state: Mapping[str, Any] = None, **kwargs ) -> Iterable[Optional[Mapping[str, Any]]]: end_date = self.end_date or date.today() if stream_state and stream_state.get(self.cursor_field): latest = stream_state.get(self.cursor_field) start = datetime.strptime(latest, "%Y-%m-%dT%H:%M:%SZ").date() # produce at least one slice with abnormal state start = start if start <= end_date else end_date else: start = self.start_date slices = [] while start <= end_date: # use small slices to maintain state since the API is unreliable end = start if start < end_date else end_date slices.append({"startDate": start.isoformat(), "endDate": end.isoformat()}) start = end + timedelta(days=1) return slices def request_params(self, stream_slice: Optional[Mapping[str, Any]], **kwargs) -> MutableMapping[str, Any]: params = super().request_params(**kwargs) or {} params["dateType"] = "UPDATE" params["page.limit"] = 1000 params = {**params, **stream_slice} return params def parse_response(self, response: requests.Response, **kwargs) -> Iterable[Mapping]: # the updatedDateTime is unnecessarily nested under date # Airbyte cannot accomodate nested cursors, so this needs to be fixed results = response.json().get("results") for result in results: yield self.unnest("date", result) # Source
CashFlows
python
urllib3__urllib3
src/urllib3/exceptions.py
{ "start": 5526, "end": 5648 }
class ____(ValueError, HTTPError): """Raised when there is something wrong with a given URL input."""
LocationValueError
python
apache__airflow
providers/amazon/src/airflow/providers/amazon/aws/sensors/athena.py
{ "start": 1226, "end": 3661 }
class ____(AwsBaseSensor[AthenaHook]): """ Poll the state of the Query until it reaches a terminal state; fails if the query fails. .. seealso:: For more information on how to use this sensor, take a look at the guide: :ref:`howto/sensor:AthenaSensor` :param query_execution_id: query_execution_id to check the state of :param max_retries: Number of times to poll for query state before returning the current state, defaults to None :param sleep_time: Time in seconds to wait between two consecutive call to check query status on athena, defaults to 10 :param aws_conn_id: The Airflow connection used for AWS credentials. If this is ``None`` or empty then the default boto3 behaviour is used. If running Airflow in a distributed manner and aws_conn_id is None or empty, then default boto3 configuration would be used (and must be maintained on each worker node). :param region_name: AWS region_name. If not specified then the default boto3 behaviour is used. :param verify: Whether or not to verify SSL certificates. See: https://boto3.amazonaws.com/v1/documentation/api/latest/reference/core/session.html :param botocore_config: Configuration dictionary (key-values) for botocore client. See: https://botocore.amazonaws.com/v1/documentation/api/latest/reference/config.html """ INTERMEDIATE_STATES = ( "QUEUED", "RUNNING", ) FAILURE_STATES = ( "FAILED", "CANCELLED", ) SUCCESS_STATES = ("SUCCEEDED",) aws_hook_class = AthenaHook template_fields: Sequence[str] = aws_template_fields( "query_execution_id", ) ui_color = "#66c3ff" def __init__( self, *, query_execution_id: str, max_retries: int | None = None, sleep_time: int = 10, **kwargs: Any, ) -> None: super().__init__(**kwargs) self.query_execution_id = query_execution_id self.sleep_time = sleep_time self.max_retries = max_retries def poke(self, context: Context) -> bool: state = self.hook.poll_query_status(self.query_execution_id, self.max_retries, self.sleep_time) if state in self.FAILURE_STATES: raise AirflowException("Athena sensor failed") if state in self.INTERMEDIATE_STATES: return False return True
AthenaSensor
python
django__django
tests/admin_views/tests.py
{ "start": 323861, "end": 329878 }
class ____(TestCase): @classmethod def setUpTestData(cls): cls.superuser = User.objects.create_superuser( username="super", password="secret", email="super@example.com" ) cls.s1 = Section.objects.create(name="Test section") cls.a1 = Article.objects.create( content="<p>Middle content</p>", date=datetime.datetime(2008, 3, 18, 11, 54, 58), section=cls.s1, ) cls.a2 = Article.objects.create( content="<p>Oldest content</p>", date=datetime.datetime(2000, 3, 18, 11, 54, 58), section=cls.s1, ) cls.a3 = Article.objects.create( content="<p>Newest content</p>", date=datetime.datetime(2009, 3, 18, 11, 54, 58), section=cls.s1, ) cls.p1 = PrePopulatedPost.objects.create( title="A Long Title", published=True, slug="a-long-title" ) def setUp(self): self.client.force_login(self.superuser) def test_field_prefix_css_classes(self): """ Fields have a CSS class name with a 'field-' prefix. """ response = self.client.get(reverse("admin:admin_views_post_add")) # The main form self.assertContains(response, 'class="form-row field-title"') self.assertContains(response, 'class="form-row field-content"') self.assertContains(response, 'class="form-row field-public"') self.assertContains(response, 'class="form-row field-awesomeness_level"') self.assertContains(response, 'class="form-row field-coolness"') self.assertContains(response, 'class="form-row field-value"') self.assertContains(response, 'class="form-row"') # The lambda function # The tabular inline self.assertContains(response, '<td class="field-url">') self.assertContains(response, '<td class="field-posted">') def test_index_css_classes(self): """ CSS class names are used for each app and model on the admin index pages (#17050). """ # General index page response = self.client.get(reverse("admin:index")) self.assertContains(response, '<div class="app-admin_views module') self.assertContains( response, '<thead class="visually-hidden"><tr><th scope="col">Model name</th>' '<th scope="col">Add link</th><th scope="col">Change or view list link</th>' "</tr></thead>", html=True, ) self.assertContains(response, '<tr class="model-actor">') self.assertContains(response, '<tr class="model-album">') # App index page response = self.client.get(reverse("admin:app_list", args=("admin_views",))) self.assertContains(response, '<div class="app-admin_views module') self.assertContains( response, '<thead class="visually-hidden"><tr><th scope="col">Model name</th>' '<th scope="col">Add link</th><th scope="col">Change or view list link</th>' "</tr></thead>", html=True, ) self.assertContains(response, '<tr class="model-actor">') self.assertContains(response, '<tr class="model-album">') def test_app_model_in_form_body_class(self): """ Ensure app and model tag are correctly read by change_form template """ response = self.client.get(reverse("admin:admin_views_section_add")) self.assertContains(response, '<body class=" app-admin_views model-section ') def test_app_model_in_list_body_class(self): """ Ensure app and model tag are correctly read by change_list template """ response = self.client.get(reverse("admin:admin_views_section_changelist")) self.assertContains(response, '<body class=" app-admin_views model-section ') def test_app_model_in_delete_confirmation_body_class(self): """ Ensure app and model tag are correctly read by delete_confirmation template """ response = self.client.get( reverse("admin:admin_views_section_delete", args=(self.s1.pk,)) ) self.assertContains(response, '<body class=" app-admin_views model-section ') def test_app_model_in_app_index_body_class(self): """ Ensure app and model tag are correctly read by app_index template """ response = self.client.get(reverse("admin:app_list", args=("admin_views",))) self.assertContains(response, '<body class=" dashboard app-admin_views') def test_app_model_in_delete_selected_confirmation_body_class(self): """ Ensure app and model tag are correctly read by delete_selected_confirmation template """ action_data = { ACTION_CHECKBOX_NAME: [self.s1.pk], "action": "delete_selected", "index": 0, } response = self.client.post( reverse("admin:admin_views_section_changelist"), action_data ) self.assertContains(response, '<body class=" app-admin_views model-section ') def test_changelist_field_classes(self): """ Cells of the change list table should contain the field name in their class attribute. """ Podcast.objects.create(name="Django Dose", release_date=datetime.date.today()) response = self.client.get(reverse("admin:admin_views_podcast_changelist")) self.assertContains(response, '<th class="field-name">') self.assertContains(response, '<td class="field-release_date nowrap">') self.assertContains(response, '<td class="action-checkbox">') try: import docutils except ImportError: docutils = None @unittest.skipUnless(docutils, "no docutils installed.") @override_settings(ROOT_URLCONF="admin_views.urls") @modify_settings( INSTALLED_APPS={"append": ["django.contrib.admindocs", "django.contrib.flatpages"]} )
CSSTest
python
Netflix__metaflow
metaflow/plugins/cards/card_modules/test_cards.py
{ "start": 358, "end": 692 }
class ____(MetaflowCard): type = "test_pathspec_card" def render(self, task): import random import string return "%s %s" % ( task.pathspec, "".join( random.choice(string.ascii_uppercase + string.digits) for _ in range(6) ), )
TestPathSpecCard
python
allegroai__clearml
clearml/backend_api/services/v2_23/dataviews.py
{ "start": 13662, "end": 15669 }
class ____(NonStrictDataModel): """ :param version: Version id of a version belonging to the dataset :type version: str :param dataset: Existing Dataset id :type dataset: str :param merge_with: Version ID to merge with :type merge_with: str """ _schema = { "properties": { "dataset": {"description": "Existing Dataset id", "type": "string"}, "merge_with": {"description": "Version ID to merge with", "type": "string"}, "version": { "description": "Version id of a version belonging to the dataset", "type": "string", }, }, "required": ["dataset", "version"], "type": "object", } def __init__(self, version, dataset, merge_with=None, **kwargs): super(DataviewEntry, self).__init__(**kwargs) self.version = version self.dataset = dataset self.merge_with = merge_with @schema_property("version") def version(self): return self._property_version @version.setter def version(self, value): if value is None: self._property_version = None return self.assert_isinstance(value, "version", six.string_types) self._property_version = value @schema_property("dataset") def dataset(self): return self._property_dataset @dataset.setter def dataset(self, value): if value is None: self._property_dataset = None return self.assert_isinstance(value, "dataset", six.string_types) self._property_dataset = value @schema_property("merge_with") def merge_with(self): return self._property_merge_with @merge_with.setter def merge_with(self, value): if value is None: self._property_merge_with = None return self.assert_isinstance(value, "merge_with", six.string_types) self._property_merge_with = value
DataviewEntry
python
fluentpython__example-code-2e
06-obj-ref/twilight_bus.py
{ "start": 224, "end": 632 }
class ____: """A bus model that makes passengers vanish""" def __init__(self, passengers=None): if passengers is None: self.passengers = [] # <1> else: self.passengers = passengers #<2> def pick(self, name): self.passengers.append(name) def drop(self, name): self.passengers.remove(name) # <3> # end::TWILIGHT_BUS_CLASS[]
TwilightBus
python
joke2k__faker
tests/providers/test_company.py
{ "start": 21564, "end": 21874 }
class ____: """Test nl_BE company provider methods""" def test_company_suffix(self, faker, num_samples): for _ in range(num_samples): suffix = faker.company_suffix() assert isinstance(suffix, str) assert suffix in NlBeCompanyProvider.company_suffixes
TestNlBe
python
tensorflow__tensorflow
tensorflow/python/kernel_tests/sparse_ops/sparse_reshape_op_test.py
{ "start": 1165, "end": 15279 }
class ____(test.TestCase): def _SparseTensorPlaceholder(self): return sparse_tensor.SparseTensor( array_ops.placeholder(dtypes.int64), array_ops.placeholder(dtypes.float64), array_ops.placeholder(dtypes.int64)) def _SparseTensorValue_5x6(self): ind = np.array([[0, 0], [1, 0], [1, 3], [1, 4], [3, 2], [3, 3]]).astype(np.int64) val = np.array([0, 10, 13, 14, 32, 33]).astype(np.float64) shape = np.array([5, 6]).astype(np.int64) return sparse_tensor.SparseTensorValue(ind, val, shape) def _SparseTensorValue_2x3x4(self): ind = np.array([[0, 0, 1], [0, 1, 0], [0, 1, 2], [1, 0, 3], [1, 1, 1], [1, 1, 3], [1, 2, 2]]) val = np.array([1, 10, 12, 103, 111, 113, 122]) shape = np.array([2, 3, 4]) return sparse_tensor.SparseTensorValue(ind, val, shape) def testStaticShapeInfoPreserved(self): sp_input = sparse_tensor.SparseTensor.from_value( self._SparseTensorValue_5x6()) self.assertAllEqual((5, 6), sp_input.get_shape()) sp_output = sparse_ops.sparse_reshape(sp_input, shape=(1, 5, 2, 3)) self.assertAllEqual((1, 5, 2, 3), sp_output.get_shape()) def testStaticShapeInfoPreservedWithInferredDims(self): sp_input = sparse_tensor.SparseTensor.from_value( self._SparseTensorValue_2x3x4()) self.assertAllEqual((2, 3, 4), sp_input.get_shape()) sp_output = sparse_ops.sparse_reshape(sp_input, shape=(2, -1)) self.assertAllEqual((2, 3 * 4), sp_output.get_shape()) @test_util.run_deprecated_v1 def testRaisesIfMoreThanOneInferredDim(self): sp_input = sparse_tensor.SparseTensor.from_value( self._SparseTensorValue_2x3x4()) with self.assertRaisesRegex(ValueError, "At most one dimension can"): sparse_ops.sparse_reshape(sp_input, shape=(-1, 2, -1)) @test_util.run_deprecated_v1 def testRaisesIfInferredShapeNotPossible(self): sp_input = sparse_tensor.SparseTensor.from_value( self._SparseTensorValue_2x3x4()) with self.assertRaisesRegex(ValueError, "Cannot reshape"): sparse_ops.sparse_reshape(sp_input, shape=(-1, 7)) @test_util.run_deprecated_v1 def testPropagatesFullyKnownDenseShapeWhenShapePartiallyKnown(self): sp_input = sparse_tensor.SparseTensor.from_value( self._SparseTensorValue_2x3x4()) self.assertAllEqual((2, 3, 4), sp_input.shape) sp_output = sparse_ops.sparse_reshape( sp_input, shape=array_ops.concat( (constant_op.constant([2], dtype=dtypes.int64), array_ops.placeholder(dtype=dtypes.int64, shape=[1])), axis=0)) self.assertAllEqual((2, 3 * 4), sp_output.shape) def testSameShape(self): with self.session() as sess: input_val = self._SparseTensorValue_5x6() sp_output = sparse_ops.sparse_reshape(input_val, [5, 6]) output_val = self.evaluate(sp_output) self.assertAllEqual(output_val.indices, input_val.indices) self.assertAllEqual(output_val.values, input_val.values) self.assertAllEqual(output_val.dense_shape, input_val.dense_shape) def testReshapeIntegeroverflow(self): with self.session(): with self.assertRaises(errors.InvalidArgumentError): sp_output = sparse_ops.gen_sparse_ops.sparse_reshape( input_indices=[[0, 0]], input_shape=[2**32, 2**31], new_shape=[1, 1], name=None) self.evaluate(sp_output) def testReshapeNegativeShape(self): with self.session(): with self.assertRaises(errors.InvalidArgumentError): sp_output = sparse_ops.gen_sparse_ops.sparse_reshape( input_indices=[[0, 0]], input_shape=[1, -1], new_shape=[1, 1], name=None) self.evaluate(sp_output) @test_util.run_deprecated_v1 def testFeedSameShape(self): with self.session() as sess: sp_input = self._SparseTensorPlaceholder() input_val = self._SparseTensorValue_5x6() sp_output = sparse_ops.sparse_reshape(sp_input, [5, 6]) output_val = sess.run(sp_output, {sp_input: input_val}) self.assertAllEqual(output_val.indices, input_val.indices) self.assertAllEqual(output_val.values, input_val.values) self.assertAllEqual(output_val.dense_shape, input_val.dense_shape) @test_util.run_deprecated_v1 def testWorksWellWithTfShape(self): with self.session() as sess: sp_input = self._SparseTensorPlaceholder() input_val = self._SparseTensorValue_5x6() shape = array_ops.shape(sp_input) # tf.shape generates int32 output sp_output = sparse_ops.sparse_reshape(sp_input, shape) output_val = sess.run(sp_output, {sp_input: input_val}) self.assertAllEqual(output_val.indices, input_val.indices) self.assertAllEqual(output_val.values, input_val.values) self.assertAllEqual(output_val.dense_shape, input_val.dense_shape) @test_util.run_deprecated_v1 def testFeedSameShapeWithInferredDim(self): with self.session() as sess: sp_input = self._SparseTensorPlaceholder() input_val = self._SparseTensorValue_5x6() sp_output = sparse_ops.sparse_reshape(sp_input, [-1, 6]) output_val = sess.run(sp_output, {sp_input: input_val}) self.assertAllEqual(output_val.indices, input_val.indices) self.assertAllEqual(output_val.values, input_val.values) self.assertAllEqual(output_val.dense_shape, input_val.dense_shape) @test_util.run_deprecated_v1 def testFeedNewShapeSameRank(self): with self.session() as sess: sp_input = self._SparseTensorPlaceholder() input_val = self._SparseTensorValue_5x6() sp_output = sparse_ops.sparse_reshape(sp_input, [3, 10]) output_val = sess.run(sp_output, {sp_input: input_val}) self.assertAllEqual(output_val.indices, np.array([[0, 0], [0, 6], [0, 9], [1, 0], [2, 0], [2, 1]])) self.assertAllEqual(output_val.values, input_val.values) self.assertAllEqual(output_val.dense_shape, [3, 10]) @test_util.run_deprecated_v1 def testFeedNewShapeSameRankWithInferredDim(self): with self.session() as sess: sp_input = self._SparseTensorPlaceholder() input_val = self._SparseTensorValue_5x6() sp_output = sparse_ops.sparse_reshape(sp_input, [3, -1]) output_val = sess.run(sp_output, {sp_input: input_val}) self.assertAllEqual(output_val.indices, np.array([[0, 0], [0, 6], [0, 9], [1, 0], [2, 0], [2, 1]])) self.assertAllEqual(output_val.values, input_val.values) self.assertAllEqual(output_val.dense_shape, [3, 10]) def testUpRank(self): with self.session() as sess: input_val = self._SparseTensorValue_5x6() sp_output = sparse_ops.sparse_reshape(input_val, [2, 3, 5]) output_val = self.evaluate(sp_output) self.assertAllEqual(output_val.indices, np.array([[0, 0, 0], [0, 1, 1], [0, 1, 4], [0, 2, 0], [1, 1, 0], [1, 1, 1]])) self.assertAllEqual(output_val.values, input_val.values) self.assertAllEqual(output_val.dense_shape, [2, 3, 5]) @test_util.run_deprecated_v1 def testFeedUpRank(self): with self.session() as sess: sp_input = self._SparseTensorPlaceholder() input_val = self._SparseTensorValue_5x6() sp_output = sparse_ops.sparse_reshape(sp_input, [2, 3, 5]) output_val = sess.run(sp_output, {sp_input: input_val}) self.assertAllEqual(output_val.indices, np.array([[0, 0, 0], [0, 1, 1], [0, 1, 4], [0, 2, 0], [1, 1, 0], [1, 1, 1]])) self.assertAllEqual(output_val.values, input_val.values) self.assertAllEqual(output_val.dense_shape, [2, 3, 5]) @test_util.run_deprecated_v1 def testFeedUpRankWithInferredDim(self): with self.session() as sess: sp_input = self._SparseTensorPlaceholder() input_val = self._SparseTensorValue_5x6() sp_output = sparse_ops.sparse_reshape(sp_input, [2, -1, 5]) output_val = sess.run(sp_output, {sp_input: input_val}) self.assertAllEqual(output_val.indices, np.array([[0, 0, 0], [0, 1, 1], [0, 1, 4], [0, 2, 0], [1, 1, 0], [1, 1, 1]])) self.assertAllEqual(output_val.values, input_val.values) self.assertAllEqual(output_val.dense_shape, [2, 3, 5]) @test_util.run_deprecated_v1 def testFeedDownRank(self): with self.session() as sess: sp_input = self._SparseTensorPlaceholder() input_val = self._SparseTensorValue_2x3x4() sp_output = sparse_ops.sparse_reshape(sp_input, [6, 4]) output_val = sess.run(sp_output, {sp_input: input_val}) self.assertAllEqual(output_val.indices, np.array([[0, 1], [1, 0], [1, 2], [3, 3], [4, 1], [4, 3], [5, 2]])) self.assertAllEqual(output_val.values, input_val.values) self.assertAllEqual(output_val.dense_shape, [6, 4]) @test_util.run_deprecated_v1 def testFeedDownRankWithInferredDim(self): with self.session() as sess: sp_input = self._SparseTensorPlaceholder() input_val = self._SparseTensorValue_2x3x4() sp_output = sparse_ops.sparse_reshape(sp_input, [6, -1]) output_val = sess.run(sp_output, {sp_input: input_val}) self.assertAllEqual(output_val.indices, np.array([[0, 1], [1, 0], [1, 2], [3, 3], [4, 1], [4, 3], [5, 2]])) self.assertAllEqual(output_val.values, input_val.values) self.assertAllEqual(output_val.dense_shape, [6, 4]) @test_util.run_deprecated_v1 def testFeedMultipleInferredDims(self): with self.session() as sess: sp_input = self._SparseTensorPlaceholder() input_val = self._SparseTensorValue_5x6() sp_output = sparse_ops.sparse_reshape(sp_input, [4, -1, -1]) with self.assertRaisesOpError("only one output dimension may be -1"): sess.run(sp_output, {sp_input: input_val}) @test_util.run_deprecated_v1 def testProvideStaticallyMismatchedSizes(self): input_val = self._SparseTensorValue_5x6() sp_input = sparse_tensor.SparseTensor.from_value(input_val) with self.assertRaisesRegex(ValueError, "Cannot reshape"): sparse_ops.sparse_reshape(sp_input, [4, 7]) @test_util.run_deprecated_v1 def testFeedMismatchedSizes(self): with self.session() as sess: sp_input = self._SparseTensorPlaceholder() input_val = self._SparseTensorValue_5x6() sp_output = sparse_ops.sparse_reshape(sp_input, [4, 7]) with self.assertRaisesOpError( "Input to reshape is a tensor with 30 dense values"): sess.run(sp_output, {sp_input: input_val}) @test_util.run_deprecated_v1 def testFeedMismatchedSizesWithInferredDim(self): with self.session() as sess: sp_input = self._SparseTensorPlaceholder() input_val = self._SparseTensorValue_5x6() sp_output = sparse_ops.sparse_reshape(sp_input, [4, -1]) with self.assertRaisesOpError("requested shape requires a multiple"): sess.run(sp_output, {sp_input: input_val}) @test_util.run_deprecated_v1 def testFeedPartialShapes(self): with self.session(): # Incorporate new rank into shape information if known sp_input = self._SparseTensorPlaceholder() sp_output = sparse_ops.sparse_reshape(sp_input, [2, 3, 5]) self.assertListEqual(sp_output.indices.get_shape().as_list(), [None, 3]) self.assertListEqual(sp_output.dense_shape.get_shape().as_list(), [3]) # Incorporate known shape information about input indices in output # indices sp_input = self._SparseTensorPlaceholder() sp_input.indices.set_shape([5, None]) sp_output = sparse_ops.sparse_reshape(sp_input, [2, 3, 5]) self.assertListEqual(sp_output.indices.get_shape().as_list(), [5, 3]) self.assertListEqual(sp_output.dense_shape.get_shape().as_list(), [3]) # Even if new_shape has no shape information, we know the ranks of # output indices and shape sp_input = self._SparseTensorPlaceholder() sp_input.indices.set_shape([5, None]) new_shape = array_ops.placeholder(dtypes.int64) sp_output = sparse_ops.sparse_reshape(sp_input, new_shape) self.assertListEqual(sp_output.indices.get_shape().as_list(), [5, None]) self.assertListEqual(sp_output.dense_shape.get_shape().as_list(), [None]) @test_util.run_deprecated_v1 def testFeedDenseReshapeSemantics(self): with self.session() as sess: # Compute a random rank-5 initial shape and new shape, randomly sparsify # it, and check that the output of SparseReshape has the same semantics # as a dense reshape. factors = np.array([2] * 4 + [3] * 4 + [5] * 4) # 810k total elements orig_rank = np.random.randint(2, 7) orig_map = np.random.randint(orig_rank, size=factors.shape) orig_shape = [np.prod(factors[orig_map == d]) for d in range(orig_rank)] new_rank = np.random.randint(2, 7) new_map = np.random.randint(new_rank, size=factors.shape) new_shape = [np.prod(factors[new_map == d]) for d in range(new_rank)] orig_dense = np.random.uniform(size=orig_shape) orig_indices = np.transpose(np.nonzero(orig_dense < 0.5)) orig_values = orig_dense[orig_dense < 0.5] new_dense = np.reshape(orig_dense, new_shape) new_indices = np.transpose(np.nonzero(new_dense < 0.5)) new_values = new_dense[new_dense < 0.5] sp_input = self._SparseTensorPlaceholder() input_val = sparse_tensor.SparseTensorValue(orig_indices, orig_values, orig_shape) sp_output = sparse_ops.sparse_reshape(sp_input, new_shape) output_val = sess.run(sp_output, {sp_input: input_val}) self.assertAllEqual(output_val.indices, new_indices) self.assertAllEqual(output_val.values, new_values) self.assertAllEqual(output_val.dense_shape, new_shape)
SparseReshapeTest
python
ansible__ansible
lib/ansible/_internal/_templating/_marker_behaviors.py
{ "start": 236, "end": 525 }
class ____(metaclass=abc.ABCMeta): """Base class to support custom handling of `Marker` values encountered during concatenation or finalization.""" @abc.abstractmethod def handle_marker(self, value: Marker) -> t.Any: """Handle the given `Marker` value."""
MarkerBehavior
python
modin-project__modin
modin/experimental/core/storage_formats/pandas/parsers.py
{ "start": 5701, "end": 6294 }
class ____(PandasParser): @staticmethod @doc(_doc_parse_func, parameters=_doc_parse_parameters_common) def parse(fname, **kwargs): warnings.filterwarnings("ignore") num_splits = 1 single_worker_read = kwargs.pop("single_worker_read", None) df = pandas.read_xml(fname, **kwargs) if single_worker_read: return df length = len(df) width = len(df.columns) return _split_result_for_readers(1, num_splits, df) + [length, width] @doc(_doc_pandas_parser_class, data_type="custom text")
ExperimentalPandasXmlParser
python
facebook__pyre-check
client/commands/pyre_language_server.py
{ "start": 2964, "end": 4864 }
class ____(abc.ABC): @abc.abstractmethod async def write_telemetry( self, parameters: Dict[str, object], activity_key: Optional[Dict[str, object]], ) -> None: raise NotImplementedError() @abc.abstractmethod def get_language_server_features(self) -> features.LanguageServerFeatures: raise NotImplementedError() @abc.abstractmethod async def update_overlay_if_needed(self, document_path: Path) -> float: raise NotImplementedError() @abc.abstractmethod async def process_open_request( self, parameters: lsp.DidOpenTextDocumentParameters, activity_key: Optional[Dict[str, object]] = None, ) -> None: raise NotImplementedError() @abc.abstractmethod async def process_close_request( self, parameters: lsp.DidCloseTextDocumentParameters ) -> None: raise NotImplementedError() @abc.abstractmethod async def process_did_change_request( self, parameters: lsp.DidChangeTextDocumentParameters, activity_key: Optional[Dict[str, object]] = None, ) -> None: raise NotImplementedError() @abc.abstractmethod async def process_did_save_request( self, parameters: lsp.DidSaveTextDocumentParameters, activity_key: Optional[Dict[str, object]] = None, ) -> None: raise NotImplementedError() @abc.abstractmethod async def process_type_coverage_request( self, parameters: lsp.TypeCoverageParameters, request_id: Union[int, str, None], activity_key: Optional[Dict[str, object]] = None, ) -> None: raise NotImplementedError() @abc.abstractmethod async def process_shutdown_request(self, request_id: Union[int, str, None]) -> None: raise NotImplementedError() @dataclasses.dataclass(frozen=True)
PyreLanguageServerApi
python
Textualize__rich
tests/test_repr.py
{ "start": 1617, "end": 2164 }
class ____: def __init__( self, name, eats, fly=True, another=StupidClass(2), extinct=NotStupid() ): self.name = name self.eats = eats self.fly = fly self.another = another self.extinct = extinct def test_rich_repr() -> None: assert (repr(Foo("hello"))) == "Foo('hello', 'hello', egg=1)" assert (repr(Foo("hello", bar=3))) == "Foo('hello', 'hello', bar=3, egg=1)" def test_rich_repr_positional_only() -> None: _locals = locals().copy() exec( """\ @rich.repr.auto
Bird
python
kamyu104__LeetCode-Solutions
Python/check-if-word-is-valid-after-substitutions.py
{ "start": 29, "end": 448 }
class ____(object): def isValid(self, S): """ :type S: str :rtype: bool """ stack = [] for i in S: if i == 'c': if stack[-2:] == ['a', 'b']: stack.pop() stack.pop() else: return False else: stack.append(i) return not stack
Solution
python
openai__openai-python
src/openai/types/audio/transcription.py
{ "start": 1373, "end": 1685 }
class ____(BaseModel): seconds: float """Duration of the input audio in seconds.""" type: Literal["duration"] """The type of the usage object. Always `duration` for this variant.""" Usage: TypeAlias = Annotated[Union[UsageTokens, UsageDuration], PropertyInfo(discriminator="type")]
UsageDuration
python
django__django
tests/invalid_models_tests/test_custom_fields.py
{ "start": 151, "end": 666 }
class ____(SimpleTestCase): def test_none_column(self): class NoColumnField(models.AutoField): def db_type(self, connection): # None indicates not to create a column in the database. return None class Model(models.Model): field = NoColumnField(primary_key=True, db_column="other_field") other_field = models.IntegerField() field = Model._meta.get_field("field") self.assertEqual(field.check(), [])
CustomFieldTest
python
encode__starlette
starlette/websockets.py
{ "start": 576, "end": 8004 }
class ____(HTTPConnection): def __init__(self, scope: Scope, receive: Receive, send: Send) -> None: super().__init__(scope) assert scope["type"] == "websocket" self._receive = receive self._send = send self.client_state = WebSocketState.CONNECTING self.application_state = WebSocketState.CONNECTING async def receive(self) -> Message: """ Receive ASGI websocket messages, ensuring valid state transitions. """ if self.client_state == WebSocketState.CONNECTING: message = await self._receive() message_type = message["type"] if message_type != "websocket.connect": raise RuntimeError(f'Expected ASGI message "websocket.connect", but got {message_type!r}') self.client_state = WebSocketState.CONNECTED return message elif self.client_state == WebSocketState.CONNECTED: message = await self._receive() message_type = message["type"] if message_type not in {"websocket.receive", "websocket.disconnect"}: raise RuntimeError( f'Expected ASGI message "websocket.receive" or "websocket.disconnect", but got {message_type!r}' ) if message_type == "websocket.disconnect": self.client_state = WebSocketState.DISCONNECTED return message else: raise RuntimeError('Cannot call "receive" once a disconnect message has been received.') async def send(self, message: Message) -> None: """ Send ASGI websocket messages, ensuring valid state transitions. """ if self.application_state == WebSocketState.CONNECTING: message_type = message["type"] if message_type not in {"websocket.accept", "websocket.close", "websocket.http.response.start"}: raise RuntimeError( 'Expected ASGI message "websocket.accept", "websocket.close" or "websocket.http.response.start", ' f"but got {message_type!r}" ) if message_type == "websocket.close": self.application_state = WebSocketState.DISCONNECTED elif message_type == "websocket.http.response.start": self.application_state = WebSocketState.RESPONSE else: self.application_state = WebSocketState.CONNECTED await self._send(message) elif self.application_state == WebSocketState.CONNECTED: message_type = message["type"] if message_type not in {"websocket.send", "websocket.close"}: raise RuntimeError( f'Expected ASGI message "websocket.send" or "websocket.close", but got {message_type!r}' ) if message_type == "websocket.close": self.application_state = WebSocketState.DISCONNECTED try: await self._send(message) except OSError: self.application_state = WebSocketState.DISCONNECTED raise WebSocketDisconnect(code=1006) elif self.application_state == WebSocketState.RESPONSE: message_type = message["type"] if message_type != "websocket.http.response.body": raise RuntimeError(f'Expected ASGI message "websocket.http.response.body", but got {message_type!r}') if not message.get("more_body", False): self.application_state = WebSocketState.DISCONNECTED await self._send(message) else: raise RuntimeError('Cannot call "send" once a close message has been sent.') async def accept( self, subprotocol: str | None = None, headers: Iterable[tuple[bytes, bytes]] | None = None, ) -> None: headers = headers or [] if self.client_state == WebSocketState.CONNECTING: # pragma: no branch # If we haven't yet seen the 'connect' message, then wait for it first. await self.receive() await self.send({"type": "websocket.accept", "subprotocol": subprotocol, "headers": headers}) def _raise_on_disconnect(self, message: Message) -> None: if message["type"] == "websocket.disconnect": raise WebSocketDisconnect(message["code"], message.get("reason")) async def receive_text(self) -> str: if self.application_state != WebSocketState.CONNECTED: raise RuntimeError('WebSocket is not connected. Need to call "accept" first.') message = await self.receive() self._raise_on_disconnect(message) return cast(str, message["text"]) async def receive_bytes(self) -> bytes: if self.application_state != WebSocketState.CONNECTED: raise RuntimeError('WebSocket is not connected. Need to call "accept" first.') message = await self.receive() self._raise_on_disconnect(message) return cast(bytes, message["bytes"]) async def receive_json(self, mode: str = "text") -> Any: if mode not in {"text", "binary"}: raise RuntimeError('The "mode" argument should be "text" or "binary".') if self.application_state != WebSocketState.CONNECTED: raise RuntimeError('WebSocket is not connected. Need to call "accept" first.') message = await self.receive() self._raise_on_disconnect(message) if mode == "text": text = message["text"] else: text = message["bytes"].decode("utf-8") return json.loads(text) async def iter_text(self) -> AsyncIterator[str]: try: while True: yield await self.receive_text() except WebSocketDisconnect: pass async def iter_bytes(self) -> AsyncIterator[bytes]: try: while True: yield await self.receive_bytes() except WebSocketDisconnect: pass async def iter_json(self) -> AsyncIterator[Any]: try: while True: yield await self.receive_json() except WebSocketDisconnect: pass async def send_text(self, data: str) -> None: await self.send({"type": "websocket.send", "text": data}) async def send_bytes(self, data: bytes) -> None: await self.send({"type": "websocket.send", "bytes": data}) async def send_json(self, data: Any, mode: str = "text") -> None: if mode not in {"text", "binary"}: raise RuntimeError('The "mode" argument should be "text" or "binary".') text = json.dumps(data, separators=(",", ":"), ensure_ascii=False) if mode == "text": await self.send({"type": "websocket.send", "text": text}) else: await self.send({"type": "websocket.send", "bytes": text.encode("utf-8")}) async def close(self, code: int = 1000, reason: str | None = None) -> None: await self.send({"type": "websocket.close", "code": code, "reason": reason or ""}) async def send_denial_response(self, response: Response) -> None: if "websocket.http.response" in self.scope.get("extensions", {}): await response(self.scope, self.receive, self.send) else: raise RuntimeError("The server doesn't support the Websocket Denial Response extension.")
WebSocket
python
huggingface__transformers
src/transformers/models/roc_bert/modeling_roc_bert.py
{ "start": 23755, "end": 24420 }
class ____(nn.Module): def __init__(self, config): super().__init__() self.dense = nn.Linear(config.hidden_size, config.hidden_size) self.activation = nn.Tanh() def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: # We "pool" the model by simply taking the hidden state corresponding # to the first token. first_token_tensor = hidden_states[:, 0] pooled_output = self.dense(first_token_tensor) pooled_output = self.activation(pooled_output) return pooled_output # Copied from transformers.models.bert.modeling_bert.BertPredictionHeadTransform with Bert->RoCBert
RoCBertPooler
python
kamyu104__LeetCode-Solutions
Python/count-the-number-of-inversions.py
{ "start": 2319, "end": 3138 }
class ____(object): def numberOfPermutations(self, n, requirements): """ :type n: int :type requirements: List[List[int]] :rtype: int """ MOD = 10**9+7 lookup = [-1]*n for i, c in requirements: lookup[i] = c dp = [0]*(lookup[-1]+1) dp[0] = 1 for i in xrange(n): new_dp = [0]*len(dp) curr = 0 for j in xrange(len(dp)): curr = (curr+dp[j])%MOD if j-(i+1) >= 0: curr = (curr-dp[j-(i+1)])%MOD new_dp[j] = curr if lookup[i] == -1 or lookup[i] == j else 0 dp = new_dp return dp[-1] # Time: O(n^2 * k), k = max(cnt for _, cnt in requirements) # Space: O(n + k) # knapsack dp, combinatorics
Solution3
python
getsentry__sentry
tests/sentry/api/endpoints/test_organization_search_details.py
{ "start": 3560, "end": 8995 }
class ____(APITestCase): endpoint = "sentry-api-0-organization-search-details" method = "put" def setUp(self) -> None: self.login_as(user=self.user) @cached_property def member(self): user = self.create_user("test@test.com") self.create_member(organization=self.organization, user=user) return user def test_owner_can_edit_shared_search(self) -> None: search = SavedSearch.objects.create( organization=self.organization, owner_id=self.create_user().id, name="foo", query="", visibility=Visibility.ORGANIZATION, ) response = self.get_response( self.organization.slug, search.id, type=SearchType.ISSUE.value, name="foo", query="test", visibility=Visibility.ORGANIZATION, ) assert response.status_code == 200, response.content updated_obj = SavedSearch.objects.get(id=search.id) assert updated_obj.name == "foo" assert updated_obj.query == "test" def test_member_cannot_edit_org_search(self) -> None: search = SavedSearch.objects.create( organization=self.organization, owner_id=self.user.id, name="foo", query="", visibility=Visibility.ORGANIZATION, ) self.login_as(user=self.member) response = self.get_response( self.organization.slug, search.id, type=SearchType.ISSUE.value, name="foo", query="test", visibility=Visibility.ORGANIZATION, ) assert response.status_code == 403, response.content def test_member_can_edit_personal_search(self) -> None: search = SavedSearch.objects.create( organization=self.organization, owner_id=self.member.id, name="foo", query="", visibility=Visibility.OWNER, ) self.login_as(user=self.member) response = self.get_response( self.organization.slug, search.id, type=SearchType.ISSUE.value, name="foo", query="test", visibility=Visibility.OWNER, ) assert response.status_code == 200, response.content updated_obj = SavedSearch.objects.get(id=search.id) assert updated_obj.name == "foo" assert updated_obj.query == "test" def test_member_cannot_switch_personal_search_to_org(self) -> None: search = SavedSearch.objects.create( organization=self.organization, owner_id=self.member.id, name="foo", query="", visibility=Visibility.OWNER, ) self.login_as(user=self.member) response = self.get_response( self.organization.slug, search.id, type=SearchType.ISSUE.value, name="foo", query="test", visibility=Visibility.ORGANIZATION, ) assert response.status_code == 400, response.content def test_exists(self) -> None: SavedSearch.objects.create( type=SearchType.ISSUE.value, name="Some global search", query="is:unresolved", is_global=True, visibility=Visibility.ORGANIZATION, ) SavedSearch.objects.create( owner_id=self.member.id, type=SearchType.ISSUE.value, name="Some other users search", query="is:mine", visibility=Visibility.OWNER, ) search = SavedSearch.objects.create( organization=self.organization, owner_id=self.user.id, name="foo", query="", visibility=Visibility.OWNER, ) # Ignores other users searches when updateing response = self.get_response( self.organization.slug, search.id, type=SearchType.ISSUE.value, name="foo", query="is:mine", visibility=Visibility.OWNER, ) assert response.status_code == 200, response.content # 400 if it matches the global search response = self.get_response( self.organization.slug, search.id, type=SearchType.ISSUE.value, name="foo", query="is:unresolved", visibility=Visibility.OWNER, ) assert response.status_code == 400, response.content assert "already exists" in response.data["detail"] def test_can_edit_without_changing_query(self) -> None: search = SavedSearch.objects.create( organization=self.organization, owner_id=self.create_user().id, name="foo", query="test123", visibility=Visibility.ORGANIZATION, ) response = self.get_response( self.organization.slug, search.id, type=SearchType.ISSUE.value, name="bar", query="test123", visibility=Visibility.ORGANIZATION, ) assert response.status_code == 200, response.content updated_obj = SavedSearch.objects.get(id=search.id) assert updated_obj.name == "bar" assert updated_obj.query == "test123"
PutOrganizationSearchTest
python
kamyu104__LeetCode-Solutions
Python/subtree-inversion-sum.py
{ "start": 1928, "end": 2939 }
class ____(object): def subtreeInversionSum(self, edges, nums, k): """ :type edges: List[List[int]] :type nums: List[int] :type k: int :rtype: int """ def dfs(u, p): dp.append([0]*2) total, pos, neg = nums[u], 0, 0 for v in adj[u]: if v == p: continue new_total, new_pos, new_neg = dfs(v, u) total += new_total pos += new_pos neg += new_neg pos = max(pos, dp[-1][1]-2*total) neg = max(neg, dp[-1][0]+2*total) dp.pop() if len(dp)-k >= 0: dp[len(dp)-k][0] += pos dp[len(dp)-k][1] += neg return total, pos, neg adj = [[] for _ in xrange(len(nums))] for u, v in edges: adj[u].append(v) adj[v].append(u) dp = [] total, pos, neg = dfs(0, -1) return total+pos
Solution2
python
getsentry__sentry
src/sentry/integrations/discord/message_builder/base/base.py
{ "start": 638, "end": 2202 }
class ____: """ Base DiscordMessageBuilder class. Should be extended to provide more abstracted message builders for specific types of messages (e.g. DiscordIssuesMessageBuilder). """ def __init__( self, content: str = "", embeds: list[DiscordMessageEmbed] | None = None, components: list[DiscordMessageComponent] | None = None, flags: DiscordMessageFlags | None = None, ) -> None: self.content = content self.embeds = embeds self.components = components self.flags = flags def build(self, notification_uuid: str | None = None) -> DiscordMessage: return self._build( self.content, self.embeds, self.components, self.flags, ) def _build( self, content: str = "", embeds: list[DiscordMessageEmbed] | None = None, components: list[DiscordMessageComponent] | None = None, flags: DiscordMessageFlags | None = None, ) -> DiscordMessage: """ Helper method for building arbitrary Discord messages. """ embeds_list = [] if embeds is None else [embed.build() for embed in embeds] components_list = ( [] if components is None else [component.build() for component in components] ) message = DiscordMessage(content=content, embeds=embeds_list, components=components_list) if flags is not None: message["flags"] = flags.value return message
DiscordMessageBuilder
python
bokeh__bokeh
src/bokeh/core/property/struct.py
{ "start": 1503, "end": 3768 }
class ____(ParameterizedProperty[T]): """ Accept values that are structures. """ _fields: dict[str, Property[Any]] _optional: set[str] def __init__(self, **fields: Any) -> None: default = fields.pop("default", None) help = fields.pop("help", None) if not fields: raise ValueError("expected specification of fields, got nothing") self._fields = {} self._optional = set() for name, type in fields.items(): if isinstance(type, Optional): self._optional.add(name) type = type.type_param self._fields[name] = self._validate_type_param(type, help_allowed=True) # XXX super().__init__(default=default, help=help) def __eq__(self, other: object) -> bool: if isinstance(other, self.__class__): return super().__eq__(other) and self._fields == other._fields and self._optional == other._optional else: return False @property def type_params(self): return list(self._fields.values()) def validate(self, value: Any, detail: bool = True): super().validate(value, detail) if isinstance(value, SimpleNamespace): value = value.__dict__ if isinstance(value, dict) and len(value) <= len(self._fields): for name, type in self._fields.items(): if name not in value: if name not in self._optional: break elif not type.is_valid(value[name]): break else: for name in value.keys(): if name not in self._fields: break else: return msg = "" if not detail else f"expected an element of {self}, got {value!r}" raise ValueError(msg) def __str__(self) -> str: class_name = self.__class__.__name__ fields = ", ".join(f"{name}={typ}" for name, typ in self._fields.items()) return f"{class_name}({fields})" #----------------------------------------------------------------------------- # Dev API #-----------------------------------------------------------------------------
Struct
python
kamyu104__LeetCode-Solutions
Python/print-immutable-linked-list-in-reverse.py
{ "start": 49, "end": 1001 }
class ____(object): def printLinkedListInReverse(self, head): """ :type head: ImmutableListNode :rtype: None """ def print_nodes(head, count): nodes = [] while head and len(nodes) != count: nodes.append(head) head = head.getNext() for node in reversed(nodes): node.printValue() count = 0 curr = head while curr: curr = curr.getNext() count += 1 bucket_count = int(math.ceil(count**0.5)) buckets = [] count = 0 curr = head while curr: if count % bucket_count == 0: buckets.append(curr) curr = curr.getNext() count += 1 for node in reversed(buckets): print_nodes(node, bucket_count) # Time: O(n) # Space: O(n)
Solution
python
redis__redis-py
redis/_parsers/base.py
{ "start": 2054, "end": 4122 }
class ____(ABC): EXCEPTION_CLASSES = { "ERR": { "max number of clients reached": ConnectionError, "invalid password": AuthenticationError, # some Redis server versions report invalid command syntax # in lowercase "wrong number of arguments " "for 'auth' command": AuthenticationWrongNumberOfArgsError, # some Redis server versions report invalid command syntax # in uppercase "wrong number of arguments " "for 'AUTH' command": AuthenticationWrongNumberOfArgsError, MODULE_LOAD_ERROR: ModuleError, MODULE_EXPORTS_DATA_TYPES_ERROR: ModuleError, NO_SUCH_MODULE_ERROR: ModuleError, MODULE_UNLOAD_NOT_POSSIBLE_ERROR: ModuleError, **NO_AUTH_SET_ERROR, **EXTERNAL_AUTH_PROVIDER_ERROR, }, "OOM": OutOfMemoryError, "WRONGPASS": AuthenticationError, "EXECABORT": ExecAbortError, "LOADING": BusyLoadingError, "NOSCRIPT": NoScriptError, "READONLY": ReadOnlyError, "NOAUTH": AuthenticationError, "NOPERM": NoPermissionError, "ASK": AskError, "TRYAGAIN": TryAgainError, "MOVED": MovedError, "CLUSTERDOWN": ClusterDownError, "CROSSSLOT": ClusterCrossSlotError, "MASTERDOWN": MasterDownError, } @classmethod def parse_error(cls, response): "Parse an error response" error_code = response.split(" ")[0] if error_code in cls.EXCEPTION_CLASSES: response = response[len(error_code) + 1 :] exception_class = cls.EXCEPTION_CLASSES[error_code] if isinstance(exception_class, dict): exception_class = exception_class.get(response, ResponseError) return exception_class(response) return ResponseError(response) def on_disconnect(self): raise NotImplementedError() def on_connect(self, connection): raise NotImplementedError()
BaseParser
python
facebookresearch__faiss
contrib/big_batch_search.py
{ "start": 5556, "end": 17640 }
class ____: """ computation within one bucket """ def __init__( self, index, method="knn_function", pairwise_distances=faiss.pairwise_distances, knn=faiss.knn): self.index = index if index.__class__ == faiss.IndexIVFFlat: index_help = faiss.IndexFlat(index.d, index.metric_type) decode_func = lambda x: x.view("float32") by_residual = False elif index.__class__ == faiss.IndexIVFPQ: index_help = faiss.IndexPQ( index.d, index.pq.M, index.pq.nbits, index.metric_type) index_help.pq = index.pq decode_func = index_help.pq.decode index_help.is_trained = True by_residual = index.by_residual elif index.__class__ == faiss.IndexIVFScalarQuantizer: index_help = faiss.IndexScalarQuantizer( index.d, index.sq.qtype, index.metric_type) index_help.sq = index.sq decode_func = index_help.sq.decode index_help.is_trained = True by_residual = index.by_residual else: raise RuntimeError(f"index type {index.__class__} not supported") self.index_help = index_help self.decode_func = None if method == "index" else decode_func self.by_residual = by_residual self.method = method self.pairwise_distances = pairwise_distances self.knn = knn def block_search(self, xq_l, xb_l, list_ids, k, **extra_args): metric_type = self.index.metric_type if xq_l.size == 0 or xb_l.size == 0: D = I = None elif self.method == "index": faiss.copy_array_to_vector(xb_l, self.index_help.codes) self.index_help.ntotal = len(list_ids) D, I = self.index_help.search(xq_l, k) elif self.method == "pairwise_distances": # TODO implement blockwise to avoid mem blowup D = self.pairwise_distances(xq_l, xb_l, metric=metric_type) I = None elif self.method == "knn_function": D, I = self.knn(xq_l, xb_l, k, metric=metric_type, **extra_args) return D, I def big_batch_search( index, xq, k, method="knn_function", pairwise_distances=faiss.pairwise_distances, knn=faiss.knn, verbose=0, threaded=0, use_float16=False, prefetch_threads=1, computation_threads=1, q_assign=None, checkpoint=None, checkpoint_freq=7200, start_list=0, end_list=None, crash_at=-1 ): """ Search queries xq in the IVF index, with a search function that collects batches of query vectors per inverted list. This can be faster than the regular search indexes. Supports IVFFlat, IVFPQ and IVFScalarQuantizer. Supports three computation methods: method = "index": build a flat index and populate it separately for each index method = "pairwise_distances": decompress codes and compute all pairwise distances for the queries and index and add result to heap method = "knn_function": decompress codes and compute knn results for the queries threaded=0: sequential execution threaded=1: prefetch next bucket while computing the current one threaded=2: prefetch prefetch_threads buckets at a time. compute_threads>1: the knn function will get an additional thread_no that tells which worker should handle this. In threaded mode, the computation is tiled with the bucket preparation and the writeback of results (useful to maximize GPU utilization). use_float16: convert all matrices to float16 (faster for GPU gemm) q_assign: override coarse assignment, should be a matrix of size nq * nprobe checkpointing (only for threaded > 1): checkpoint: file where the checkpoints are stored checkpoint_freq: when to perform checkpointing. Should be a multiple of threaded start_list, end_list: process only a subset of invlists """ nprobe = index.nprobe assert method in ("index", "pairwise_distances", "knn_function") mem_queries = xq.nbytes mem_assign = len(xq) * nprobe * np.dtype('int32').itemsize mem_res = len(xq) * k * ( np.dtype('int64').itemsize + np.dtype('float32').itemsize ) mem_tot = mem_queries + mem_assign + mem_res if verbose > 0: logging.info( f"memory: queries {mem_queries} assign {mem_assign} " f"result {mem_res} total {mem_tot} = {mem_tot / (1<<30):.3f} GiB" ) bbs = BigBatchSearcher( index, xq, k, verbose=verbose, use_float16=use_float16 ) comp = BlockComputer( index, method=method, pairwise_distances=pairwise_distances, knn=knn ) bbs.decode_func = comp.decode_func bbs.by_residual = comp.by_residual if q_assign is None: bbs.coarse_quantization() else: bbs.q_assign = q_assign bbs.reorder_assign() if end_list is None: end_list = index.nlist completed = set() if checkpoint is not None: assert (start_list, end_list) == (0, index.nlist) if os.path.exists(checkpoint): logging.info(f"recovering checkpoint: {checkpoint}") completed = bbs.read_checkpoint(checkpoint) logging.info(f" already completed: {len(completed)}") else: logging.info("no checkpoint: starting from scratch") if threaded == 0: # simple sequential version for l in range(start_list, end_list): bbs.report(l) q_subset, xq_l, list_ids, xb_l = bbs.prepare_bucket(l) t0i = time.time() D, I = comp.block_search(xq_l, xb_l, list_ids, k) bbs.t_accu[2] += time.time() - t0i bbs.add_results_to_heap(q_subset, D, list_ids, I) elif threaded == 1: # parallel version with granularity 1 def add_results_and_prefetch(to_add, l): """ perform the addition for the previous bucket and prefetch the next (if applicable) """ if to_add is not None: bbs.add_results_to_heap(*to_add) if l < index.nlist: return bbs.prepare_bucket(l) prefetched_bucket = bbs.prepare_bucket(start_list) to_add = None pool = ThreadPool(1) for l in range(start_list, end_list): bbs.report(l) prefetched_bucket_a = pool.apply_async( add_results_and_prefetch, (to_add, l + 1)) q_subset, xq_l, list_ids, xb_l = prefetched_bucket bbs.start_t_accu() D, I = comp.block_search(xq_l, xb_l, list_ids, k) bbs.stop_t_accu(2) to_add = q_subset, D, list_ids, I bbs.start_t_accu() prefetched_bucket = prefetched_bucket_a.get() bbs.stop_t_accu(4) bbs.add_results_to_heap(*to_add) pool.close() else: def task_manager_thread( task, pool_size, start_task, end_task, completed, output_queue, input_queue, ): try: with ThreadPool(pool_size) as pool: res = [pool.apply_async( task, args=(i, output_queue, input_queue)) for i in range(start_task, end_task) if i not in completed] for r in res: r.get() pool.close() pool.join() output_queue.put(None) except: traceback.print_exc() _thread.interrupt_main() raise def task_manager(*args): task_manager = threading.Thread( target=task_manager_thread, args=args, ) task_manager.daemon = True task_manager.start() return task_manager def prepare_task(task_id, output_queue, input_queue=None): try: logging.info(f"Prepare start: {task_id}") q_subset, xq_l, list_ids, xb_l = bbs.prepare_bucket(task_id) output_queue.put((task_id, q_subset, xq_l, list_ids, xb_l)) logging.info(f"Prepare end: {task_id}") except: traceback.print_exc() _thread.interrupt_main() raise def compute_task(task_id, output_queue, input_queue): try: logging.info(f"Compute start: {task_id}") t_wait_out = 0 while True: t0 = time.time() logging.info(f'Compute input: task {task_id}') input_value = input_queue.get() t_wait_in = time.time() - t0 if input_value is None: # signal for other compute tasks input_queue.put(None) break centroid, q_subset, xq_l, list_ids, xb_l = input_value logging.info(f'Compute work: task {task_id}, centroid {centroid}') t0 = time.time() if computation_threads > 1: D, I = comp.block_search( xq_l, xb_l, list_ids, k, thread_id=task_id ) else: D, I = comp.block_search(xq_l, xb_l, list_ids, k) t_compute = time.time() - t0 logging.info(f'Compute output: task {task_id}, centroid {centroid}') t0 = time.time() output_queue.put( (centroid, t_wait_in, t_wait_out, t_compute, q_subset, D, list_ids, I) ) t_wait_out = time.time() - t0 logging.info(f"Compute end: {task_id}") except: traceback.print_exc() _thread.interrupt_main() raise prepare_to_compute_queue = Queue(2) compute_to_main_queue = Queue(2) compute_task_manager = task_manager( compute_task, computation_threads, 0, computation_threads, set(), compute_to_main_queue, prepare_to_compute_queue, ) prepare_task_manager = task_manager( prepare_task, prefetch_threads, start_list, end_list, completed, prepare_to_compute_queue, None, ) t_checkpoint = time.time() while True: logging.info("Waiting for result") value = compute_to_main_queue.get() if not value: break centroid, t_wait_in, t_wait_out, t_compute, q_subset, D, list_ids, I = value # to test checkpointing if centroid == crash_at: 1 / 0 bbs.t_accu[2] += t_compute bbs.t_accu[4] += t_wait_in bbs.t_accu[5] += t_wait_out logging.info(f"Adding to heap start: centroid {centroid}") bbs.add_results_to_heap(q_subset, D, list_ids, I) logging.info(f"Adding to heap end: centroid {centroid}") completed.add(centroid) bbs.report(centroid) if checkpoint is not None: if time.time() - t_checkpoint > checkpoint_freq: logging.info("writing checkpoint") bbs.write_checkpoint(checkpoint, completed) t_checkpoint = time.time() prepare_task_manager.join() compute_task_manager.join() bbs.tic("finalize heap") bbs.rh.finalize() bbs.toc() return bbs.rh.D, bbs.rh.I
BlockComputer
python
gevent__gevent
examples/udp_server.py
{ "start": 268, "end": 618 }
class ____(DatagramServer): def handle(self, data, address): # pylint:disable=method-hidden print('%s: got %r' % (address[0], data)) self.socket.sendto(('Received %s bytes' % len(data)).encode('utf-8'), address) if __name__ == '__main__': print('Receiving datagrams on :9000') EchoServer(':9000').serve_forever()
EchoServer
python
getsentry__sentry
tests/snuba/api/endpoints/test_organization_events_stats_trace_metrics.py
{ "start": 204, "end": 7321 }
class ____(OrganizationEventsEndpointTestBase): dataset = "tracemetrics" viewname = "sentry-api-0-organization-events-stats" def setUp(self) -> None: super().setUp() self.login_as(user=self.user) self.start = self.day_ago = before_now(days=1).replace( hour=10, minute=0, second=0, microsecond=0 ) self.end = self.start + timedelta(hours=4) def test_simple_deprecated(self) -> None: metric_values = [1, 2, 3, 4] trace_metrics = [ self.create_trace_metric( metric_name, metric_value, "counter", timestamp=self.start + timedelta(hours=i) ) for metric_name in ["foo", "bar"] for i, metric_value in enumerate(metric_values) ] self.store_trace_metrics(trace_metrics) response = self.do_request( { "start": self.start, "end": self.end, "interval": "1h", "yAxis": "sum(value)", "query": "metric.name:foo metric.type:counter", "project": self.project.id, "dataset": self.dataset, } ) assert response.status_code == 200, response.content assert [bucket for _, bucket in response.data["data"]] == [ [{"count": value}] for value in metric_values ] def test_simple(self) -> None: metric_values = [1, 2, 3, 4] trace_metrics = [ self.create_trace_metric( metric_name, metric_value, "counter", timestamp=self.start + timedelta(hours=i) ) for metric_name in ["foo", "bar"] for i, metric_value in enumerate(metric_values) ] self.store_trace_metrics(trace_metrics) response = self.do_request( { "metricName": "foo", "metricType": "counter", "start": self.start, "end": self.end, "interval": "1h", "yAxis": "sum(value)", "project": self.project.id, "dataset": self.dataset, } ) assert response.status_code == 200, response.content assert [bucket for _, bucket in response.data["data"]] == [ [{"count": value}] for value in metric_values ] def test_per_second_function(self) -> None: metrics = [] # Only place events in the first bucket to save time in the test for _ in range(36): metrics.append( self.create_trace_metric( "test_metric", 1.0, "counter", timestamp=self.start + timedelta(hours=0) ) ) self.store_trace_metrics(metrics) response = self.do_request( { "metricName": "test_metric", "metricType": "counter", "start": self.start, "end": self.end, "interval": "1h", "yAxis": "per_second(value, test_metric, counter, -)", "project": self.project.id, "dataset": self.dataset, } ) assert response.status_code == 200, response.content assert len(response.data["data"]) == 4 assert response.data["data"][0][1][0]["count"] == pytest.approx(0.01, abs=0.001) def test_top_events(self) -> None: metric_values = [6, 0, 6, 3, 0, 3] trace_metrics = [ self.create_trace_metric( "foo", 1, "counter", timestamp=self.start + timedelta(hours=1), attributes={"key": "value1"}, ), self.create_trace_metric( "bar", 1, "counter", timestamp=self.start + timedelta(hours=1), attributes={"key": "value3"}, ), ] for hour, count in enumerate(metric_values): trace_metrics.append( self.create_trace_metric( "bar", count, "counter", timestamp=self.start + timedelta(hours=hour), attributes={"key": "value1"}, ) ) trace_metrics.append( self.create_trace_metric( "bar", count, "counter", timestamp=self.start + timedelta(hours=hour), attributes={"key": "value2"}, ) ) self.store_trace_metrics(trace_metrics) response = self.do_request( { "metricName": "bar", "metricType": "counter", "start": self.day_ago, "end": self.day_ago + timedelta(hours=6), "interval": "1h", "yAxis": "sum(value)", "field": ["key", "sum(value)"], "orderby": ["-sum(value)"], "project": self.project.id, "dataset": self.dataset, "excludeOther": 0, "topEvents": 2, } ) assert response.status_code == 200, response.content for key in ["value1", "value2"]: rows = response.data[key]["data"][0:6] for expected, result in zip(metric_values, rows): assert result[1][0]["count"] == expected, key rows = response.data["Other"]["data"][0:6] for expected, result in zip([0, 1, 0, 0, 0, 0], rows): assert result[1][0]["count"] == expected, "Other" assert response.data["value1"]["meta"]["dataset"] == "tracemetrics" assert response.data["value2"]["meta"]["dataset"] == "tracemetrics" assert response.data["Other"]["meta"]["dataset"] == "tracemetrics" def test_meta_accuracy(self): metric_values = [1, 2, 3, 4] trace_metrics = [ self.create_trace_metric( metric_name, metric_value, "counter", timestamp=self.start + timedelta(hours=i) ) for metric_name in ["foo", "bar"] for i, metric_value in enumerate(metric_values) ] self.store_trace_metrics(trace_metrics) response = self.do_request( { "metricName": "foo", "metricType": "counter", "start": self.start, "end": self.end, "interval": "1h", "yAxis": ["sum(value)", "per_second(value)"], "project": self.project.id, "dataset": self.dataset, } ) assert response.status_code == 200, response.content for y_axis in ["sum(value)", "per_second(value)"]: assert [ bucket["value"] for bucket in response.data[y_axis]["meta"]["accuracy"]["sampleCount"] ] == [1 for _ in metric_values]
OrganizationEventsStatsTraceMetricsEndpointTest
python
ipython__ipython
IPython/extensions/deduperreload/deduperreload.py
{ "start": 4481, "end": 5554 }
class ____: """ Recursive data structure to keep track of reloadable functions/methods. Each object corresponds to a specific scope level. children: classes inside given scope, maps class name to autoreload tree for that class's scope funcs_to_autoreload: list of function names that can be autoreloaded in given scope. new_nested_classes: Classes getting added in new autoreload cycle """ def __init__(self) -> None: self.children: dict[str, AutoreloadTree] = {} self.defs_to_reload: list[tuple[tuple[str, ...], ast.AST]] = [] self.defs_to_delete: set[str] = set() self.new_nested_classes: dict[str, ast.AST] = {} def traverse_prefixes(self, prefixes: list[str]) -> AutoreloadTree: """ Return ref to the AutoreloadTree at the namespace specified by prefixes """ cur = self for prefix in prefixes: if prefix not in cur.children: cur.children[prefix] = AutoreloadTree() cur = cur.children[prefix] return cur
AutoreloadTree
python
scipy__scipy
benchmarks/benchmarks/go_benchmark_functions/go_funcs_D.py
{ "start": 8492, "end": 9752 }
class ____(Benchmark): r""" Deckkers-Aarts objective function. This class defines the Deckkers-Aarts [1]_ global optimization problem. This is a multimodal minimization problem defined as follows: .. math:: f_{\text{DeckkersAarts}}(x) = 10^5x_1^2 + x_2^2 - (x_1^2 + x_2^2)^2 + 10^{-5}(x_1^2 + x_2^2)^4 with :math:`x_i \in [-20, 20]` for :math:`i = 1, 2`. *Global optimum*: :math:`f(x) = -24776.518242168` for :math:`x = [0, \pm 14.9451209]` .. [1] Jamil, M. & Yang, X.-S. A Literature Survey of Benchmark Functions For Global Optimization Problems Int. Journal of Mathematical Modelling and Numerical Optimisation, 2013, 4, 150-194. TODO: jamil solution and global minimum are slightly wrong. """ def __init__(self, dimensions=2): Benchmark.__init__(self, dimensions) self._bounds = list(zip([-20.0] * self.N, [20.0] * self.N)) self.custom_bounds = ([-1, 1], [14, 16]) self.global_optimum = [[0.0, 14.9451209]] self.fglob = -24776.518342168 def fun(self, x, *args): self.nfev += 1 return (1.e5 * x[0] ** 2 + x[1] ** 2 - (x[0] ** 2 + x[1] ** 2) ** 2 + 1.e-5 * (x[0] ** 2 + x[1] ** 2) ** 4)
DeckkersAarts
python
ray-project__ray
python/ray/tune/examples/pbt_dcgan_mnist/common.py
{ "start": 2566, "end": 3291 }
class ____(nn.Module): def __init__(self): super(Discriminator, self).__init__() self.main = nn.Sequential( nn.Conv2d(nc, ndf, 4, 2, 1, bias=False), nn.LeakyReLU(0.2, inplace=True), nn.Conv2d(ndf, ndf * 2, 4, 2, 1, bias=False), nn.BatchNorm2d(ndf * 2), nn.LeakyReLU(0.2, inplace=True), nn.Conv2d(ndf * 2, ndf * 4, 4, 2, 1, bias=False), nn.BatchNorm2d(ndf * 4), nn.LeakyReLU(0.2, inplace=True), nn.Conv2d(ndf * 4, 1, 4, 1, 0, bias=False), nn.Sigmoid(), ) def forward(self, input): return self.main(input) # __GANmodel_end__ # __INCEPTION_SCORE_begin__
Discriminator
python
huggingface__transformers
src/transformers/models/dac/configuration_dac.py
{ "start": 826, "end": 4581 }
class ____(PreTrainedConfig): r""" This is the configuration class to store the configuration of an [`DacModel`]. It is used to instantiate a Dac model according to the specified arguments, defining the model architecture. Instantiating a configuration with the defaults will yield a similar configuration to that of the [descript/dac_16khz](https://huggingface.co/descript/dac_16khz) architecture. Configuration objects inherit from [`PreTrainedConfig`] and can be used to control the model outputs. Read the documentation from [`PreTrainedConfig`] for more information. Args: encoder_hidden_size (`int`, *optional*, defaults to 64): Intermediate representation dimension for the encoder. downsampling_ratios (`list[int]`, *optional*, defaults to `[2, 4, 8, 8]`): Ratios for downsampling in the encoder. These are used in reverse order for upsampling in the decoder. decoder_hidden_size (`int`, *optional*, defaults to 1536): Intermediate representation dimension for the decoder. n_codebooks (`int`, *optional*, defaults to 9): Number of codebooks in the VQVAE. codebook_size (`int`, *optional*, defaults to 1024): Number of discrete codes in each codebook. codebook_dim (`int`, *optional*, defaults to 8): Dimension of the codebook vectors. If not defined, uses `encoder_hidden_size`. quantizer_dropout (`bool`, *optional*, defaults to 0): Whether to apply dropout to the quantizer. commitment_loss_weight (float, *optional*, defaults to 0.25): Weight of the commitment loss term in the VQVAE loss function. codebook_loss_weight (float, *optional*, defaults to 1.0): Weight of the codebook loss term in the VQVAE loss function. sampling_rate (`int`, *optional*, defaults to 16000): The sampling rate at which the audio waveform should be digitalized expressed in hertz (Hz). Example: ```python >>> from transformers import DacModel, DacConfig >>> # Initializing a "descript/dac_16khz" style configuration >>> configuration = DacConfig() >>> # Initializing a model (with random weights) from the "descript/dac_16khz" style configuration >>> model = DacModel(configuration) >>> # Accessing the model configuration >>> configuration = model.config ```""" model_type = "dac" def __init__( self, encoder_hidden_size=64, downsampling_ratios=[2, 4, 8, 8], decoder_hidden_size=1536, n_codebooks=9, codebook_size=1024, codebook_dim=8, quantizer_dropout=0, commitment_loss_weight=0.25, codebook_loss_weight=1.0, sampling_rate=16000, **kwargs, ): self.encoder_hidden_size = encoder_hidden_size self.downsampling_ratios = downsampling_ratios self.decoder_hidden_size = decoder_hidden_size self.upsampling_ratios = downsampling_ratios[::-1] self.n_codebooks = n_codebooks self.codebook_size = codebook_size self.codebook_dim = codebook_dim self.quantizer_dropout = quantizer_dropout self.sampling_rate = sampling_rate self.hidden_size = encoder_hidden_size * (2 ** len(downsampling_ratios)) self.hop_length = int(np.prod(downsampling_ratios)) self.commitment_loss_weight = commitment_loss_weight self.codebook_loss_weight = codebook_loss_weight super().__init__(**kwargs) @property def frame_rate(self) -> int: hop_length = np.prod(self.upsampling_ratios) return math.ceil(self.sampling_rate / hop_length) __all__ = ["DacConfig"]
DacConfig
python
Textualize__textual
docs/examples/guide/widgets/checker01.py
{ "start": 174, "end": 1107 }
class ____(Widget): """Render an 8x8 checkerboard.""" def render_line(self, y: int) -> Strip: """Render a line of the widget. y is relative to the top of the widget.""" row_index = y // 4 # A checkerboard square consists of 4 rows if row_index >= 8: # Generate blank lines when we reach the end return Strip.blank(self.size.width) is_odd = row_index % 2 # Used to alternate the starting square on each row white = Style.parse("on white") # Get a style object for a white background black = Style.parse("on black") # Get a style object for a black background # Generate a list of segments with alternating black and white space characters segments = [ Segment(" " * 8, black if (column + is_odd) % 2 else white) for column in range(8) ] strip = Strip(segments, 8 * 8) return strip
CheckerBoard
python
apache__airflow
providers/google/src/airflow/providers/google/cloud/operators/cloud_sql.py
{ "start": 42620, "end": 46939 }
class ____(CloudSQLBaseOperator): """ Import data into a Cloud SQL instance from Cloud Storage. CSV IMPORT `````````` This operator is NOT idempotent for a CSV import. If the same file is imported multiple times, the imported data will be duplicated in the database. Moreover, if there are any unique constraints the duplicate import may result in an error. SQL IMPORT `````````` This operator is idempotent for a SQL import if it was also exported by Cloud SQL. The exported SQL contains 'DROP TABLE IF EXISTS' statements for all tables to be imported. If the import file was generated in a different way, idempotence is not guaranteed. It has to be ensured on the SQL file level. .. seealso:: For more information on how to use this operator, take a look at the guide: :ref:`howto/operator:CloudSQLImportInstanceOperator` :param instance: Cloud SQL instance ID. This does not include the project ID. :param body: The request body, as described in https://cloud.google.com/sql/docs/mysql/admin-api/v1beta4/instances/import#request-body :param project_id: Optional, Google Cloud Project ID. If set to None or missing, the default project_id from the Google Cloud connection is used. :param gcp_conn_id: The connection ID used to connect to Google Cloud. :param api_version: API version used (e.g. v1beta4). :param validate_body: Whether the body should be validated. Defaults to True. :param impersonation_chain: Optional service account to impersonate using short-term credentials, or chained list of accounts required to get the access_token of the last account in the list, which will be impersonated in the request. If set as a string, the account must grant the originating account the Service Account Token Creator IAM role. If set as a sequence, the identities from the list must grant Service Account Token Creator IAM role to the directly preceding identity, with first account from the list granting this role to the originating account (templated). """ # [START gcp_sql_import_template_fields] template_fields: Sequence[str] = ( "project_id", "instance", "body", "gcp_conn_id", "api_version", "impersonation_chain", ) # [END gcp_sql_import_template_fields] ui_color = "#D3EDFB" operator_extra_links = (CloudSQLInstanceLink(), FileDetailsLink()) def __init__( self, *, instance: str, body: dict, project_id: str = PROVIDE_PROJECT_ID, gcp_conn_id: str = "google_cloud_default", api_version: str = "v1beta4", validate_body: bool = True, impersonation_chain: str | Sequence[str] | None = None, **kwargs, ) -> None: self.body = body self.validate_body = validate_body super().__init__( project_id=project_id, instance=instance, gcp_conn_id=gcp_conn_id, api_version=api_version, impersonation_chain=impersonation_chain, **kwargs, ) def _validate_inputs(self) -> None: super()._validate_inputs() if not self.body: raise AirflowException("The required parameter 'body' is empty") def _validate_body_fields(self) -> None: if self.validate_body: GcpBodyFieldValidator(CLOUD_SQL_IMPORT_VALIDATION, api_version=self.api_version).validate( self.body ) def execute(self, context: Context) -> None: self._validate_body_fields() hook = CloudSQLHook( gcp_conn_id=self.gcp_conn_id, api_version=self.api_version, impersonation_chain=self.impersonation_chain, ) CloudSQLInstanceLink.persist( context=context, project_id=self.project_id or hook.project_id, ) FileDetailsLink.persist( context=context, uri=self.body["importContext"]["uri"][5:], project_id=self.project_id or hook.project_id, ) return hook.import_instance(project_id=self.project_id, instance=self.instance, body=self.body)
CloudSQLImportInstanceOperator
python
bokeh__bokeh
src/bokeh/core/property/dataspec.py
{ "start": 8521, "end": 8678 }
class ____(DataSpec): def __init__(self, default, *, help: str | None = None) -> None: super().__init__(Float, default=default, help=help)
FloatSpec
python
apache__airflow
providers/google/src/airflow/providers/google/cloud/links/bigquery.py
{ "start": 1803, "end": 2022 }
class ____(BaseGoogleLink): """Helper class for constructing BigQuery Job Detail Link.""" name = "BigQuery Job Detail" key = "bigquery_job_detail" format_str = BIGQUERY_JOB_DETAIL_LINK
BigQueryJobDetailLink
python
fastapi__sqlmodel
docs_src/tutorial/relationship_attributes/back_populates/tutorial001.py
{ "start": 317, "end": 4538 }
class ____(SQLModel, table=True): id: Optional[int] = Field(default=None, primary_key=True) name: str = Field(index=True) secret_name: str age: Optional[int] = Field(default=None, index=True) team_id: Optional[int] = Field(default=None, foreign_key="team.id") team: Optional[Team] = Relationship() sqlite_file_name = "database.db" sqlite_url = f"sqlite:///{sqlite_file_name}" engine = create_engine(sqlite_url, echo=True) def create_db_and_tables(): SQLModel.metadata.create_all(engine) def create_heroes(): with Session(engine) as session: team_preventers = Team(name="Preventers", headquarters="Sharp Tower") team_z_force = Team(name="Z-Force", headquarters="Sister Margaret's Bar") hero_deadpond = Hero( name="Deadpond", secret_name="Dive Wilson", team=team_z_force ) hero_rusty_man = Hero( name="Rusty-Man", secret_name="Tommy Sharp", age=48, team=team_preventers ) hero_spider_boy = Hero(name="Spider-Boy", secret_name="Pedro Parqueador") session.add(hero_deadpond) session.add(hero_rusty_man) session.add(hero_spider_boy) session.commit() session.refresh(hero_deadpond) session.refresh(hero_rusty_man) session.refresh(hero_spider_boy) print("Created hero:", hero_deadpond) print("Created hero:", hero_rusty_man) print("Created hero:", hero_spider_boy) hero_spider_boy.team = team_preventers session.add(hero_spider_boy) session.commit() session.refresh(hero_spider_boy) print("Updated hero:", hero_spider_boy) hero_black_lion = Hero(name="Black Lion", secret_name="Trevor Challa", age=35) hero_sure_e = Hero(name="Princess Sure-E", secret_name="Sure-E") team_wakaland = Team( name="Wakaland", headquarters="Wakaland Capital City", heroes=[hero_black_lion, hero_sure_e], ) session.add(team_wakaland) session.commit() session.refresh(team_wakaland) print("Team Wakaland:", team_wakaland) hero_tarantula = Hero(name="Tarantula", secret_name="Natalia Roman-on", age=32) hero_dr_weird = Hero(name="Dr. Weird", secret_name="Steve Weird", age=36) hero_cap = Hero( name="Captain North America", secret_name="Esteban Rogelios", age=93 ) team_preventers.heroes.append(hero_tarantula) team_preventers.heroes.append(hero_dr_weird) team_preventers.heroes.append(hero_cap) session.add(team_preventers) session.commit() session.refresh(hero_tarantula) session.refresh(hero_dr_weird) session.refresh(hero_cap) print("Preventers new hero:", hero_tarantula) print("Preventers new hero:", hero_dr_weird) print("Preventers new hero:", hero_cap) def select_heroes(): with Session(engine) as session: statement = select(Team).where(Team.name == "Preventers") result = session.exec(statement) team_preventers = result.one() print("Preventers heroes:", team_preventers.heroes) def update_heroes(): with Session(engine) as session: hero_spider_boy = session.exec( select(Hero).where(Hero.name == "Spider-Boy") ).one() preventers_team = session.exec( select(Team).where(Team.name == "Preventers") ).one() print("Hero Spider-Boy:", hero_spider_boy) print("Preventers Team:", preventers_team) print("Preventers Team Heroes:", preventers_team.heroes) hero_spider_boy.team = None print("Spider-Boy without team:", hero_spider_boy) print("Preventers Team Heroes again:", preventers_team.heroes) session.add(hero_spider_boy) session.commit() print("After committing") session.refresh(hero_spider_boy) print("Spider-Boy after commit:", hero_spider_boy) print("Preventers Team Heroes after commit:", preventers_team.heroes) def main(): create_db_and_tables() create_heroes() select_heroes() update_heroes() if __name__ == "__main__": main()
Hero
python
dagster-io__dagster
python_modules/dagster-test/dagster_test/dg_defs/composites/python/component.py
{ "start": 147, "end": 487 }
class ____(Component, Model, Resolvable): asset: ResolvedAssetSpec def build_defs(self, context): return Definitions(assets=[self.asset]) @component_instance def first(_): return PyComponent(asset=AssetSpec("first_py")) @component_instance def second(_): return PyComponent(asset=AssetSpec("second_py"))
PyComponent
python
RaRe-Technologies__gensim
gensim/corpora/textcorpus.py
{ "start": 2224, "end": 15917 }
class ____(interfaces.CorpusABC): """Helper class to simplify the pipeline of getting BoW vectors from plain text. Notes ----- This is an abstract base class: override the :meth:`~gensim.corpora.textcorpus.TextCorpus.get_texts` and :meth:`~gensim.corpora.textcorpus.TextCorpus.__len__` methods to match your particular input. Given a filename (or a file-like object) in constructor, the corpus object will be automatically initialized with a dictionary in `self.dictionary` and will support the :meth:`~gensim.corpora.textcorpus.TextCorpus.__iter__` corpus method. You have a few different ways of utilizing this class via subclassing or by construction with different preprocessing arguments. The :meth:`~gensim.corpora.textcorpus.TextCorpus.__iter__` method converts the lists of tokens produced by :meth:`~gensim.corpora.textcorpus.TextCorpus.get_texts` to BoW format using :meth:`gensim.corpora.dictionary.Dictionary.doc2bow`. :meth:`~gensim.corpora.textcorpus.TextCorpus.get_texts` does the following: #. Calls :meth:`~gensim.corpora.textcorpus.TextCorpus.getstream` to get a generator over the texts. It yields each document in turn from the underlying text file or files. #. For each document from the stream, calls :meth:`~gensim.corpora.textcorpus.TextCorpus.preprocess_text` to produce a list of tokens. If metadata=True, it yields a 2-`tuple` with the document number as the second element. Preprocessing consists of 0+ `character_filters`, a `tokenizer`, and 0+ `token_filters`. The preprocessing consists of calling each filter in `character_filters` with the document text. Unicode is not guaranteed, and if desired, the first filter should convert to unicode. The output of each character filter should be another string. The output from the final filter is fed to the `tokenizer`, which should split the string into a list of tokens (strings). Afterwards, the list of tokens is fed through each filter in `token_filters`. The final output returned from :meth:`~gensim.corpora.textcorpus.TextCorpus.preprocess_text` is the output from the final token filter. So to use this class, you can either pass in different preprocessing functions using the `character_filters`, `tokenizer`, and `token_filters` arguments, or you can subclass it. If subclassing: override :meth:`~gensim.corpora.textcorpus.TextCorpus.getstream` to take text from different input sources in different formats. Override :meth:`~gensim.corpora.textcorpus.TextCorpus.preprocess_text` if you must provide different initial preprocessing, then call the :meth:`~gensim.corpora.textcorpus.TextCorpus.preprocess_text` method to apply the normal preprocessing. You can also override :meth:`~gensim.corpora.textcorpus.TextCorpus.get_texts` in order to tag the documents (token lists) with different metadata. The default preprocessing consists of: #. :func:`~gensim.parsing.preprocessing.lower_to_unicode` - lowercase and convert to unicode (assumes utf8 encoding) #. :func:`~gensim.utils.deaccent`- deaccent (asciifolding) #. :func:`~gensim.parsing.preprocessing.strip_multiple_whitespaces` - collapse multiple whitespaces into one #. :func:`~gensim.utils.simple_tokenize` - tokenize by splitting on whitespace #. :func:`~gensim.parsing.preprocessing.remove_short_tokens` - remove words less than 3 characters long #. :func:`~gensim.parsing.preprocessing.remove_stopword_tokens` - remove stopwords """ def __init__(self, input=None, dictionary=None, metadata=False, character_filters=None, tokenizer=None, token_filters=None): """ Parameters ---------- input : str, optional Path to top-level directory (file) to traverse for corpus documents. dictionary : :class:`~gensim.corpora.dictionary.Dictionary`, optional If a dictionary is provided, it will not be updated with the given corpus on initialization. If None - new dictionary will be built for the given corpus. If `input` is None, the dictionary will remain uninitialized. metadata : bool, optional If True - yield metadata with each document. character_filters : iterable of callable, optional Each will be applied to the text of each document in order, and should return a single string with the modified text. For Python 2, the original text will not be unicode, so it may be useful to convert to unicode as the first character filter. If None - using :func:`~gensim.parsing.preprocessing.lower_to_unicode`, :func:`~gensim.utils.deaccent` and :func:`~gensim.parsing.preprocessing.strip_multiple_whitespaces`. tokenizer : callable, optional Tokenizer for document, if None - using :func:`~gensim.utils.simple_tokenize`. token_filters : iterable of callable, optional Each will be applied to the iterable of tokens in order, and should return another iterable of tokens. These filters can add, remove, or replace tokens, or do nothing at all. If None - using :func:`~gensim.parsing.preprocessing.remove_short_tokens` and :func:`~gensim.parsing.preprocessing.remove_stopword_tokens`. Examples -------- .. sourcecode:: pycon >>> from gensim.corpora.textcorpus import TextCorpus >>> from gensim.test.utils import datapath >>> from gensim import utils >>> >>> >>> class CorpusMiislita(TextCorpus): ... stopwords = set('for a of the and to in on'.split()) ... ... def get_texts(self): ... for doc in self.getstream(): ... yield [word for word in utils.to_unicode(doc).lower().split() if word not in self.stopwords] ... ... def __len__(self): ... self.length = sum(1 for _ in self.get_texts()) ... return self.length >>> >>> >>> corpus = CorpusMiislita(datapath('head500.noblanks.cor.bz2')) >>> len(corpus) 250 >>> document = next(iter(corpus.get_texts())) """ self.input = input self.metadata = metadata self.character_filters = character_filters if self.character_filters is None: self.character_filters = [lower_to_unicode, deaccent, strip_multiple_whitespaces] self.tokenizer = tokenizer if self.tokenizer is None: self.tokenizer = simple_tokenize self.token_filters = token_filters if self.token_filters is None: self.token_filters = [remove_short_tokens, remove_stopword_tokens] self.length = None self.dictionary = None self.init_dictionary(dictionary) def init_dictionary(self, dictionary): """Initialize/update dictionary. Parameters ---------- dictionary : :class:`~gensim.corpora.dictionary.Dictionary`, optional If a dictionary is provided, it will not be updated with the given corpus on initialization. If None - new dictionary will be built for the given corpus. Notes ----- If self.input is None - make nothing. """ self.dictionary = dictionary if dictionary is not None else Dictionary() if self.input is not None: if dictionary is None: logger.info("Initializing dictionary") metadata_setting = self.metadata self.metadata = False self.dictionary.add_documents(self.get_texts()) self.metadata = metadata_setting else: logger.info("Input stream provided but dictionary already initialized") else: logger.warning("No input document stream provided; assuming dictionary will be initialized some other way.") def __iter__(self): """Iterate over the corpus. Yields ------ list of (int, int) Document in BoW format (+ metadata if self.metadata). """ if self.metadata: for text, metadata in self.get_texts(): yield self.dictionary.doc2bow(text, allow_update=False), metadata else: for text in self.get_texts(): yield self.dictionary.doc2bow(text, allow_update=False) def getstream(self): """Generate documents from the underlying plain text collection (of one or more files). Yields ------ str Document read from plain-text file. Notes ----- After generator end - initialize self.length attribute. """ num_texts = 0 with utils.file_or_filename(self.input) as f: for line in f: yield line num_texts += 1 self.length = num_texts def preprocess_text(self, text): """Apply `self.character_filters`, `self.tokenizer`, `self.token_filters` to a single text document. Parameters --------- text : str Document read from plain-text file. Return ------ list of str List of tokens extracted from `text`. """ for character_filter in self.character_filters: text = character_filter(text) tokens = self.tokenizer(text) for token_filter in self.token_filters: tokens = token_filter(tokens) return tokens def step_through_preprocess(self, text): """Apply preprocessor one by one and generate result. Warnings -------- This is useful for debugging issues with the corpus preprocessing pipeline. Parameters ---------- text : str Document text read from plain-text file. Yields ------ (callable, object) Pre-processor, output from pre-processor (based on `text`) """ for character_filter in self.character_filters: text = character_filter(text) yield (character_filter, text) tokens = self.tokenizer(text) yield (self.tokenizer, tokens) for token_filter in self.token_filters: yield (token_filter, token_filter(tokens)) def get_texts(self): """Generate documents from corpus. Yields ------ list of str Document as sequence of tokens (+ lineno if self.metadata) """ lines = self.getstream() if self.metadata: for lineno, line in enumerate(lines): yield self.preprocess_text(line), (lineno,) else: for line in lines: yield self.preprocess_text(line) def sample_texts(self, n, seed=None, length=None): """Generate `n` random documents from the corpus without replacement. Parameters ---------- n : int Number of documents we want to sample. seed : int, optional If specified, use it as a seed for local random generator. length : int, optional Value will used as corpus length (because calculate length of corpus can be costly operation). If not specified - will call `__length__`. Raises ------ ValueError If `n` less than zero or greater than corpus size. Notes ----- Given the number of remaining documents in a corpus, we need to choose n elements. The probability for the current element to be chosen is `n` / remaining. If we choose it, we just decrease the `n` and move to the next element. Yields ------ list of str Sampled document as sequence of tokens. """ random_generator = random if seed is None else random.Random(seed) if length is None: length = len(self) if not n <= length: raise ValueError("n {0:d} is larger/equal than length of corpus {1:d}.".format(n, length)) if not 0 <= n: raise ValueError("Negative sample size n {0:d}.".format(n)) i = 0 for i, sample in enumerate(self.getstream()): if i == length: break remaining_in_corpus = length - i chance = random_generator.randint(1, remaining_in_corpus) if chance <= n: n -= 1 if self.metadata: yield self.preprocess_text(sample[0]), sample[1] else: yield self.preprocess_text(sample) if n != 0: # This means that length was set to be greater than number of items in corpus # and we were not able to sample enough documents before the stream ended. raise ValueError("length {0:d} greater than number of documents in corpus {1:d}".format(length, i + 1)) def __len__(self): """Get length of corpus Warnings -------- If self.length is None - will read all corpus for calculate this attribute through :meth:`~gensim.corpora.textcorpus.TextCorpus.getstream`. Returns ------- int Length of corpus. """ if self.length is None: # cache the corpus length self.length = sum(1 for _ in self.getstream()) return self.length
TextCorpus
python
mlflow__mlflow
mlflow/genai/optimize/types.py
{ "start": 537, "end": 1353 }
class ____: """ Parameters for configuring a LLM Model. Args: model_name: Name of the model in the format `<provider>:/<model name>` or `<provider>/<model name>`. For example, "openai:/gpt-4o", "anthropic:/claude-4", or "openai/gpt-4o". base_uri: Optional base URI for the API endpoint. If not provided, the default endpoint for the provider will be used. temperature: Optional sampling temperature for the model's outputs. Higher values (e.g., 0.8) make the output more random, while lower values (e.g., 0.2) make it more deterministic. """ model_name: str base_uri: str | None = None temperature: float | None = None @deprecated( since="3.5.0", ) @experimental(version="3.0.0") @dataclass
LLMParams
python
dagster-io__dagster
python_modules/dagster/dagster/_core/definitions/auto_materialize_rule_impls.py
{ "start": 29154, "end": 33648 }
class ____( AutoMaterializeRule, NamedTuple( "_SkipOnNotAllParentsUpdatedRule", [("require_update_for_all_parent_partitions", bool)] ), ): """An auto-materialize rule that enforces that an asset can only be materialized if all parents have been materialized since the asset's last materialization. Args: require_update_for_all_parent_partitions (Optional[bool]): Applies only to an unpartitioned asset or an asset partition that depends on more than one partition in any upstream asset. If true, requires all upstream partitions in each upstream asset to be materialized since the downstream asset's last materialization in order to update it. If false, requires at least one upstream partition in each upstream asset to be materialized since the downstream asset's last materialization in order to update it. Defaults to false. """ @property def decision_type(self) -> AutoMaterializeDecisionType: return AutoMaterializeDecisionType.SKIP @property def description(self) -> str: if self.require_update_for_all_parent_partitions is False: return "waiting on upstream data to be updated" else: return "waiting until all upstream partitions are updated" def evaluate_for_asset( self, context: "AutomationContext", ) -> "AutomationResult": from dagster._core.definitions.declarative_automation.automation_condition import ( AutomationResult, ) asset_partitions_by_evaluation_data = defaultdict(set) # only need to evaluate net-new candidates and candidates whose parents have changed subset_to_evaluate = ( context.legacy_context.candidates_not_evaluated_on_previous_tick_subset | context.legacy_context.candidate_parent_has_or_will_update_subset ) for candidate in subset_to_evaluate.asset_partitions: parent_partitions = context.legacy_context.asset_graph.get_parents_partitions( context.legacy_context.asset_key, candidate.partition_key, ).parent_partitions updated_parent_partitions = ( context.legacy_context.instance_queryer.get_parent_asset_partitions_updated_after_child( asset_partition=candidate, parent_asset_partitions=parent_partitions, respect_materialization_data_versions=context.legacy_context.respect_materialization_data_versions, ignored_parent_keys=set(), ) | context.legacy_context.parent_will_update_subset.asset_partitions ) if self.require_update_for_all_parent_partitions: # All upstream partitions must be updated in order for the candidate to be updated non_updated_parent_keys = { parent.asset_key for parent in parent_partitions - updated_parent_partitions } else: # At least one upstream partition in each upstream asset must be updated in order # for the candidate to be updated parent_asset_keys = context.legacy_context.asset_graph.get( context.legacy_context.asset_key ).parent_keys updated_parent_keys = {ap.asset_key for ap in updated_parent_partitions} non_updated_parent_keys = parent_asset_keys - updated_parent_keys # do not require past partitions of this asset to be updated non_updated_parent_keys -= {context.legacy_context.asset_key} if non_updated_parent_keys: asset_partitions_by_evaluation_data[ WaitingOnAssetsRuleEvaluationData( frozenset(non_updated_parent_keys) ).frozen_metadata ].add(candidate) true_subset, subsets_with_metadata = ( context.legacy_context.add_evaluation_data_from_previous_tick( asset_partitions_by_evaluation_data, ignore_subset=subset_to_evaluate ) ) true_subset = context.asset_graph_view.legacy_get_asset_subset_from_valid_subset( true_subset ) return AutomationResult(context, true_subset, subsets_with_metadata=subsets_with_metadata) @whitelist_for_serdes
SkipOnNotAllParentsUpdatedRule
python
PyCQA__pylint
tests/functional/i/invalid/invalid_overridden_method.py
{ "start": 117, "end": 426 }
class ____(metaclass=abc.ABCMeta): @property @abc.abstractmethod def prop(self): pass @abc.abstractmethod async def async_method(self): pass @abc.abstractmethod def method_a(self): pass @abc.abstractmethod def method_b(self): pass
SuperClass
python
keon__algorithms
tests/test_stack.py
{ "start": 295, "end": 2621 }
class ____(unittest.TestCase): def test_is_consecutive(self): self.assertTrue(first_is_consecutive([3, 4, 5, 6, 7])) self.assertFalse(first_is_consecutive([3, 4, 6, 7])) self.assertFalse(first_is_consecutive([3, 2, 1])) self.assertTrue(second_is_consecutive([3, 4, 5, 6, 7])) self.assertFalse(second_is_consecutive([3, 4, 6, 7])) self.assertFalse(second_is_consecutive([3, 2, 1])) def test_is_sorted(self): # Test case: bottom [6, 3, 5, 1, 2, 4] top self.assertFalse(is_sorted([6, 3, 5, 1, 2, 4])) self.assertTrue(is_sorted([1, 2, 3, 4, 5, 6])) self.assertFalse(is_sorted([3, 4, 7, 8, 5, 6])) def test_remove_min(self): # Test case: bottom [2, 8, 3, -6, 7, 3] top self.assertEqual([2, 8, 3, 7, 3], remove_min([2, 8, 3, -6, 7, 3])) # Test case: 2 smallest value [2, 8, 3, 7, 3] self.assertEqual([4, 8, 7], remove_min([4, 8, 3, 7, 3])) def test_stutter(self): # Test case: bottom [3, 7, 1, 14, 9] top self.assertEqual([3, 3, 7, 7, 1, 1, 14, 14, 9, 9], first_stutter([3, 7, 1, 14, 9])) self.assertEqual([3, 3, 7, 7, 1, 1, 14, 14, 9, 9], second_stutter([3, 7, 1, 14, 9])) def test_switch_pairs(self): # Test case: even number of values in stack # bottom [3, 8, 17, 9, 1, 10] top self.assertEqual([8, 3, 9, 17, 10, 1], first_switch_pairs([3, 8, 17, 9, 1, 10])) self.assertEqual([8, 3, 9, 17, 10, 1], second_switch_pairs([3, 8, 17, 9, 1, 10])) # Test case: odd number of values in stack # bottom [3, 8, 17, 9, 1] top self.assertEqual([8, 3, 9, 17, 1], first_switch_pairs([3, 8, 17, 9, 1])) self.assertEqual([8, 3, 9, 17, 1], second_switch_pairs([3, 8, 17, 9, 1])) def test_is_valid_parenthesis(self): self.assertTrue(is_valid("[]")) self.assertTrue(is_valid("[]()[]")) self.assertFalse(is_valid("[[[]]")) self.assertTrue(is_valid("{([])}")) self.assertFalse(is_valid("(}")) def test_simplify_path(self): p = '/my/name/is/..//keon' self.assertEqual('/my/name/keon', simplify_path(p))
TestSuite
python
tensorflow__tensorflow
tensorflow/python/framework/ops_test.py
{ "start": 119838, "end": 120669 }
class ____(test_util.TensorFlowTestCase): def testRegisteredNode(self): graph = ops.Graph() node = ops._NodeDef("a", "an_a") flops = ops.get_stats_for_node_def(graph, node, "flops") self.assertEqual(20, flops.value) missing_stat = ops.get_stats_for_node_def(graph, node, "missing_stat") self.assertEqual(None, missing_stat.value) def testUnregisteredNode(self): graph = ops.Graph() node = ops._NodeDef("b", "a_b") weight_params = ops.get_stats_for_node_def(graph, node, "weight_params") self.assertEqual(None, weight_params.value) def testAccumulateStatistics(self): flops_total = ops.OpStats("flops") self.assertEqual(None, flops_total.value) second_flops = ops.OpStats("flops", 3) flops_total += second_flops self.assertEqual(3, flops_total.value)
StatisticsTest
python
google__jax
tests/pallas/export_pallas_test.py
{ "start": 3021, "end": 3664 }
class ____(ExportTestWithTriton): def setUp(self): # TODO(b/432678342): remove once this is fixed. if jtu.is_device_cuda() and not jtu.is_cuda_compute_capability_at_least("9.0"): self.skipTest( "LLVM seems to care about the compute capability if a GPU is present" ) super().setUp() self.enter_context(pallas_call_lib._PALLAS_USE_MOSAIC_GPU(True)) def _check_cuda_export(self, exp): self.assertRegex( exp.mlir_module(), r"stablehlo.custom_call @mosaic_gpu_v2.*my_custom_kernel_name") if __name__ == '__main__': absltest.main(testLoader=jtu.JaxTestLoader())
ExportTestWithMosaicGpu
python
plotly__plotly.py
plotly/graph_objs/scattergeo/marker/_colorbar.py
{ "start": 233, "end": 61703 }
class ____(_BaseTraceHierarchyType): _parent_path_str = "scattergeo.marker" _path_str = "scattergeo.marker.colorbar" _valid_props = { "bgcolor", "bordercolor", "borderwidth", "dtick", "exponentformat", "labelalias", "len", "lenmode", "minexponent", "nticks", "orientation", "outlinecolor", "outlinewidth", "separatethousands", "showexponent", "showticklabels", "showtickprefix", "showticksuffix", "thickness", "thicknessmode", "tick0", "tickangle", "tickcolor", "tickfont", "tickformat", "tickformatstopdefaults", "tickformatstops", "ticklabeloverflow", "ticklabelposition", "ticklabelstep", "ticklen", "tickmode", "tickprefix", "ticks", "ticksuffix", "ticktext", "ticktextsrc", "tickvals", "tickvalssrc", "tickwidth", "title", "x", "xanchor", "xpad", "xref", "y", "yanchor", "ypad", "yref", } @property def bgcolor(self): """ Sets the color of padded area. The 'bgcolor' property is a color and may be specified as: - A hex string (e.g. '#ff0000') - An rgb/rgba string (e.g. 'rgb(255,0,0)') - An hsl/hsla string (e.g. 'hsl(0,100%,50%)') - An hsv/hsva string (e.g. 'hsv(0,100%,100%)') - A named CSS color: see https://plotly.com/python/css-colors/ for a list Returns ------- str """ return self["bgcolor"] @bgcolor.setter def bgcolor(self, val): self["bgcolor"] = val @property def bordercolor(self): """ Sets the axis line color. The 'bordercolor' property is a color and may be specified as: - A hex string (e.g. '#ff0000') - An rgb/rgba string (e.g. 'rgb(255,0,0)') - An hsl/hsla string (e.g. 'hsl(0,100%,50%)') - An hsv/hsva string (e.g. 'hsv(0,100%,100%)') - A named CSS color: see https://plotly.com/python/css-colors/ for a list Returns ------- str """ return self["bordercolor"] @bordercolor.setter def bordercolor(self, val): self["bordercolor"] = val @property def borderwidth(self): """ Sets the width (in px) or the border enclosing this color bar. The 'borderwidth' property is a number and may be specified as: - An int or float in the interval [0, inf] Returns ------- int|float """ return self["borderwidth"] @borderwidth.setter def borderwidth(self, val): self["borderwidth"] = val @property def dtick(self): """ Sets the step in-between ticks on this axis. Use with `tick0`. Must be a positive number, or special strings available to "log" and "date" axes. If the axis `type` is "log", then ticks are set every 10^(n*dtick) where n is the tick number. For example, to set a tick mark at 1, 10, 100, 1000, ... set dtick to 1. To set tick marks at 1, 100, 10000, ... set dtick to 2. To set tick marks at 1, 5, 25, 125, 625, 3125, ... set dtick to log_10(5), or 0.69897000433. "log" has several special values; "L<f>", where `f` is a positive number, gives ticks linearly spaced in value (but not position). For example `tick0` = 0.1, `dtick` = "L0.5" will put ticks at 0.1, 0.6, 1.1, 1.6 etc. To show powers of 10 plus small digits between, use "D1" (all digits) or "D2" (only 2 and 5). `tick0` is ignored for "D1" and "D2". If the axis `type` is "date", then you must convert the time to milliseconds. For example, to set the interval between ticks to one day, set `dtick` to 86400000.0. "date" also has special values "M<n>" gives ticks spaced by a number of months. `n` must be a positive integer. To set ticks on the 15th of every third month, set `tick0` to "2000-01-15" and `dtick` to "M3". To set ticks every 4 years, set `dtick` to "M48" The 'dtick' property accepts values of any type Returns ------- Any """ return self["dtick"] @dtick.setter def dtick(self, val): self["dtick"] = val @property def exponentformat(self): """ Determines a formatting rule for the tick exponents. For example, consider the number 1,000,000,000. If "none", it appears as 1,000,000,000. If "e", 1e+9. If "E", 1E+9. If "power", 1x10^9 (with 9 in a super script). If "SI", 1G. If "B", 1B. "SI" uses prefixes from "femto" f (10^-15) to "tera" T (10^12). *SI extended* covers instead the full SI range from "quecto" q (10^-30) to "quetta" Q (10^30). If "SI" or *SI extended* is used and the exponent is beyond the above ranges, the formatting rule will automatically be switched to the power notation. The 'exponentformat' property is an enumeration that may be specified as: - One of the following enumeration values: ['none', 'e', 'E', 'power', 'SI', 'B', 'SI extended'] Returns ------- Any """ return self["exponentformat"] @exponentformat.setter def exponentformat(self, val): self["exponentformat"] = val @property def labelalias(self): """ Replacement text for specific tick or hover labels. For example using {US: 'USA', CA: 'Canada'} changes US to USA and CA to Canada. The labels we would have shown must match the keys exactly, after adding any tickprefix or ticksuffix. For negative numbers the minus sign symbol used (U+2212) is wider than the regular ascii dash. That means you need to use −1 instead of -1. labelalias can be used with any axis type, and both keys (if needed) and values (if desired) can include html- like tags or MathJax. The 'labelalias' property accepts values of any type Returns ------- Any """ return self["labelalias"] @labelalias.setter def labelalias(self, val): self["labelalias"] = val @property def len(self): """ Sets the length of the color bar This measure excludes the padding of both ends. That is, the color bar length is this length minus the padding on both ends. The 'len' property is a number and may be specified as: - An int or float in the interval [0, inf] Returns ------- int|float """ return self["len"] @len.setter def len(self, val): self["len"] = val @property def lenmode(self): """ Determines whether this color bar's length (i.e. the measure in the color variation direction) is set in units of plot "fraction" or in *pixels. Use `len` to set the value. The 'lenmode' property is an enumeration that may be specified as: - One of the following enumeration values: ['fraction', 'pixels'] Returns ------- Any """ return self["lenmode"] @lenmode.setter def lenmode(self, val): self["lenmode"] = val @property def minexponent(self): """ Hide SI prefix for 10^n if |n| is below this number. This only has an effect when `tickformat` is "SI" or "B". The 'minexponent' property is a number and may be specified as: - An int or float in the interval [0, inf] Returns ------- int|float """ return self["minexponent"] @minexponent.setter def minexponent(self, val): self["minexponent"] = val @property def nticks(self): """ Specifies the maximum number of ticks for the particular axis. The actual number of ticks will be chosen automatically to be less than or equal to `nticks`. Has an effect only if `tickmode` is set to "auto". The 'nticks' property is a integer and may be specified as: - An int (or float that will be cast to an int) in the interval [0, 9223372036854775807] Returns ------- int """ return self["nticks"] @nticks.setter def nticks(self, val): self["nticks"] = val @property def orientation(self): """ Sets the orientation of the colorbar. The 'orientation' property is an enumeration that may be specified as: - One of the following enumeration values: ['h', 'v'] Returns ------- Any """ return self["orientation"] @orientation.setter def orientation(self, val): self["orientation"] = val @property def outlinecolor(self): """ Sets the axis line color. The 'outlinecolor' property is a color and may be specified as: - A hex string (e.g. '#ff0000') - An rgb/rgba string (e.g. 'rgb(255,0,0)') - An hsl/hsla string (e.g. 'hsl(0,100%,50%)') - An hsv/hsva string (e.g. 'hsv(0,100%,100%)') - A named CSS color: see https://plotly.com/python/css-colors/ for a list Returns ------- str """ return self["outlinecolor"] @outlinecolor.setter def outlinecolor(self, val): self["outlinecolor"] = val @property def outlinewidth(self): """ Sets the width (in px) of the axis line. The 'outlinewidth' property is a number and may be specified as: - An int or float in the interval [0, inf] Returns ------- int|float """ return self["outlinewidth"] @outlinewidth.setter def outlinewidth(self, val): self["outlinewidth"] = val @property def separatethousands(self): """ If "true", even 4-digit integers are separated The 'separatethousands' property must be specified as a bool (either True, or False) Returns ------- bool """ return self["separatethousands"] @separatethousands.setter def separatethousands(self, val): self["separatethousands"] = val @property def showexponent(self): """ If "all", all exponents are shown besides their significands. If "first", only the exponent of the first tick is shown. If "last", only the exponent of the last tick is shown. If "none", no exponents appear. The 'showexponent' property is an enumeration that may be specified as: - One of the following enumeration values: ['all', 'first', 'last', 'none'] Returns ------- Any """ return self["showexponent"] @showexponent.setter def showexponent(self, val): self["showexponent"] = val @property def showticklabels(self): """ Determines whether or not the tick labels are drawn. The 'showticklabels' property must be specified as a bool (either True, or False) Returns ------- bool """ return self["showticklabels"] @showticklabels.setter def showticklabels(self, val): self["showticklabels"] = val @property def showtickprefix(self): """ If "all", all tick labels are displayed with a prefix. If "first", only the first tick is displayed with a prefix. If "last", only the last tick is displayed with a suffix. If "none", tick prefixes are hidden. The 'showtickprefix' property is an enumeration that may be specified as: - One of the following enumeration values: ['all', 'first', 'last', 'none'] Returns ------- Any """ return self["showtickprefix"] @showtickprefix.setter def showtickprefix(self, val): self["showtickprefix"] = val @property def showticksuffix(self): """ Same as `showtickprefix` but for tick suffixes. The 'showticksuffix' property is an enumeration that may be specified as: - One of the following enumeration values: ['all', 'first', 'last', 'none'] Returns ------- Any """ return self["showticksuffix"] @showticksuffix.setter def showticksuffix(self, val): self["showticksuffix"] = val @property def thickness(self): """ Sets the thickness of the color bar This measure excludes the size of the padding, ticks and labels. The 'thickness' property is a number and may be specified as: - An int or float in the interval [0, inf] Returns ------- int|float """ return self["thickness"] @thickness.setter def thickness(self, val): self["thickness"] = val @property def thicknessmode(self): """ Determines whether this color bar's thickness (i.e. the measure in the constant color direction) is set in units of plot "fraction" or in "pixels". Use `thickness` to set the value. The 'thicknessmode' property is an enumeration that may be specified as: - One of the following enumeration values: ['fraction', 'pixels'] Returns ------- Any """ return self["thicknessmode"] @thicknessmode.setter def thicknessmode(self, val): self["thicknessmode"] = val @property def tick0(self): """ Sets the placement of the first tick on this axis. Use with `dtick`. If the axis `type` is "log", then you must take the log of your starting tick (e.g. to set the starting tick to 100, set the `tick0` to 2) except when `dtick`=*L<f>* (see `dtick` for more info). If the axis `type` is "date", it should be a date string, like date data. If the axis `type` is "category", it should be a number, using the scale where each category is assigned a serial number from zero in the order it appears. The 'tick0' property accepts values of any type Returns ------- Any """ return self["tick0"] @tick0.setter def tick0(self, val): self["tick0"] = val @property def tickangle(self): """ Sets the angle of the tick labels with respect to the horizontal. For example, a `tickangle` of -90 draws the tick labels vertically. The 'tickangle' property is a angle (in degrees) that may be specified as a number between -180 and 180. Numeric values outside this range are converted to the equivalent value (e.g. 270 is converted to -90). Returns ------- int|float """ return self["tickangle"] @tickangle.setter def tickangle(self, val): self["tickangle"] = val @property def tickcolor(self): """ Sets the tick color. The 'tickcolor' property is a color and may be specified as: - A hex string (e.g. '#ff0000') - An rgb/rgba string (e.g. 'rgb(255,0,0)') - An hsl/hsla string (e.g. 'hsl(0,100%,50%)') - An hsv/hsva string (e.g. 'hsv(0,100%,100%)') - A named CSS color: see https://plotly.com/python/css-colors/ for a list Returns ------- str """ return self["tickcolor"] @tickcolor.setter def tickcolor(self, val): self["tickcolor"] = val @property def tickfont(self): """ Sets the color bar's tick label font The 'tickfont' property is an instance of Tickfont that may be specified as: - An instance of :class:`plotly.graph_objs.scattergeo.marker.colorbar.Tickfont` - A dict of string/value properties that will be passed to the Tickfont constructor Returns ------- plotly.graph_objs.scattergeo.marker.colorbar.Tickfont """ return self["tickfont"] @tickfont.setter def tickfont(self, val): self["tickfont"] = val @property def tickformat(self): """ Sets the tick label formatting rule using d3 formatting mini- languages which are very similar to those in Python. For numbers, see: https://github.com/d3/d3-format/tree/v1.4.5#d3-format. And for dates see: https://github.com/d3/d3-time- format/tree/v2.2.3#locale_format. We add two items to d3's date formatter: "%h" for half of the year as a decimal number as well as "%{n}f" for fractional seconds with n digits. For example, *2016-10-13 09:15:23.456* with tickformat "%H~%M~%S.%2f" would display "09~15~23.46" The 'tickformat' property is a string and must be specified as: - A string - A number that will be converted to a string Returns ------- str """ return self["tickformat"] @tickformat.setter def tickformat(self, val): self["tickformat"] = val @property def tickformatstops(self): """ The 'tickformatstops' property is a tuple of instances of Tickformatstop that may be specified as: - A list or tuple of instances of plotly.graph_objs.scattergeo.marker.colorbar.Tickformatstop - A list or tuple of dicts of string/value properties that will be passed to the Tickformatstop constructor Returns ------- tuple[plotly.graph_objs.scattergeo.marker.colorbar.Tickformatstop] """ return self["tickformatstops"] @tickformatstops.setter def tickformatstops(self, val): self["tickformatstops"] = val @property def tickformatstopdefaults(self): """ When used in a template (as layout.template.data.scattergeo.mar ker.colorbar.tickformatstopdefaults), sets the default property values to use for elements of scattergeo.marker.colorbar.tickformatstops The 'tickformatstopdefaults' property is an instance of Tickformatstop that may be specified as: - An instance of :class:`plotly.graph_objs.scattergeo.marker.colorbar.Tickformatstop` - A dict of string/value properties that will be passed to the Tickformatstop constructor Returns ------- plotly.graph_objs.scattergeo.marker.colorbar.Tickformatstop """ return self["tickformatstopdefaults"] @tickformatstopdefaults.setter def tickformatstopdefaults(self, val): self["tickformatstopdefaults"] = val @property def ticklabeloverflow(self): """ Determines how we handle tick labels that would overflow either the graph div or the domain of the axis. The default value for inside tick labels is *hide past domain*. In other cases the default is *hide past div*. The 'ticklabeloverflow' property is an enumeration that may be specified as: - One of the following enumeration values: ['allow', 'hide past div', 'hide past domain'] Returns ------- Any """ return self["ticklabeloverflow"] @ticklabeloverflow.setter def ticklabeloverflow(self, val): self["ticklabeloverflow"] = val @property def ticklabelposition(self): """ Determines where tick labels are drawn relative to the ticks. Left and right options are used when `orientation` is "h", top and bottom when `orientation` is "v". The 'ticklabelposition' property is an enumeration that may be specified as: - One of the following enumeration values: ['outside', 'inside', 'outside top', 'inside top', 'outside left', 'inside left', 'outside right', 'inside right', 'outside bottom', 'inside bottom'] Returns ------- Any """ return self["ticklabelposition"] @ticklabelposition.setter def ticklabelposition(self, val): self["ticklabelposition"] = val @property def ticklabelstep(self): """ Sets the spacing between tick labels as compared to the spacing between ticks. A value of 1 (default) means each tick gets a label. A value of 2 means shows every 2nd label. A larger value n means only every nth tick is labeled. `tick0` determines which labels are shown. Not implemented for axes with `type` "log" or "multicategory", or when `tickmode` is "array". The 'ticklabelstep' property is a integer and may be specified as: - An int (or float that will be cast to an int) in the interval [1, 9223372036854775807] Returns ------- int """ return self["ticklabelstep"] @ticklabelstep.setter def ticklabelstep(self, val): self["ticklabelstep"] = val @property def ticklen(self): """ Sets the tick length (in px). The 'ticklen' property is a number and may be specified as: - An int or float in the interval [0, inf] Returns ------- int|float """ return self["ticklen"] @ticklen.setter def ticklen(self, val): self["ticklen"] = val @property def tickmode(self): """ Sets the tick mode for this axis. If "auto", the number of ticks is set via `nticks`. If "linear", the placement of the ticks is determined by a starting position `tick0` and a tick step `dtick` ("linear" is the default value if `tick0` and `dtick` are provided). If "array", the placement of the ticks is set via `tickvals` and the tick text is `ticktext`. ("array" is the default value if `tickvals` is provided). The 'tickmode' property is an enumeration that may be specified as: - One of the following enumeration values: ['auto', 'linear', 'array'] Returns ------- Any """ return self["tickmode"] @tickmode.setter def tickmode(self, val): self["tickmode"] = val @property def tickprefix(self): """ Sets a tick label prefix. The 'tickprefix' property is a string and must be specified as: - A string - A number that will be converted to a string Returns ------- str """ return self["tickprefix"] @tickprefix.setter def tickprefix(self, val): self["tickprefix"] = val @property def ticks(self): """ Determines whether ticks are drawn or not. If "", this axis' ticks are not drawn. If "outside" ("inside"), this axis' are drawn outside (inside) the axis lines. The 'ticks' property is an enumeration that may be specified as: - One of the following enumeration values: ['outside', 'inside', ''] Returns ------- Any """ return self["ticks"] @ticks.setter def ticks(self, val): self["ticks"] = val @property def ticksuffix(self): """ Sets a tick label suffix. The 'ticksuffix' property is a string and must be specified as: - A string - A number that will be converted to a string Returns ------- str """ return self["ticksuffix"] @ticksuffix.setter def ticksuffix(self, val): self["ticksuffix"] = val @property def ticktext(self): """ Sets the text displayed at the ticks position via `tickvals`. Only has an effect if `tickmode` is set to "array". Used with `tickvals`. The 'ticktext' property is an array that may be specified as a tuple, list, numpy array, or pandas Series Returns ------- numpy.ndarray """ return self["ticktext"] @ticktext.setter def ticktext(self, val): self["ticktext"] = val @property def ticktextsrc(self): """ Sets the source reference on Chart Studio Cloud for `ticktext`. The 'ticktextsrc' property must be specified as a string or as a plotly.grid_objs.Column object Returns ------- str """ return self["ticktextsrc"] @ticktextsrc.setter def ticktextsrc(self, val): self["ticktextsrc"] = val @property def tickvals(self): """ Sets the values at which ticks on this axis appear. Only has an effect if `tickmode` is set to "array". Used with `ticktext`. The 'tickvals' property is an array that may be specified as a tuple, list, numpy array, or pandas Series Returns ------- numpy.ndarray """ return self["tickvals"] @tickvals.setter def tickvals(self, val): self["tickvals"] = val @property def tickvalssrc(self): """ Sets the source reference on Chart Studio Cloud for `tickvals`. The 'tickvalssrc' property must be specified as a string or as a plotly.grid_objs.Column object Returns ------- str """ return self["tickvalssrc"] @tickvalssrc.setter def tickvalssrc(self, val): self["tickvalssrc"] = val @property def tickwidth(self): """ Sets the tick width (in px). The 'tickwidth' property is a number and may be specified as: - An int or float in the interval [0, inf] Returns ------- int|float """ return self["tickwidth"] @tickwidth.setter def tickwidth(self, val): self["tickwidth"] = val @property def title(self): """ The 'title' property is an instance of Title that may be specified as: - An instance of :class:`plotly.graph_objs.scattergeo.marker.colorbar.Title` - A dict of string/value properties that will be passed to the Title constructor Returns ------- plotly.graph_objs.scattergeo.marker.colorbar.Title """ return self["title"] @title.setter def title(self, val): self["title"] = val @property def x(self): """ Sets the x position with respect to `xref` of the color bar (in plot fraction). When `xref` is "paper", defaults to 1.02 when `orientation` is "v" and 0.5 when `orientation` is "h". When `xref` is "container", defaults to 1 when `orientation` is "v" and 0.5 when `orientation` is "h". Must be between 0 and 1 if `xref` is "container" and between "-2" and 3 if `xref` is "paper". The 'x' property is a number and may be specified as: - An int or float Returns ------- int|float """ return self["x"] @x.setter def x(self, val): self["x"] = val @property def xanchor(self): """ Sets this color bar's horizontal position anchor. This anchor binds the `x` position to the "left", "center" or "right" of the color bar. Defaults to "left" when `orientation` is "v" and "center" when `orientation` is "h". The 'xanchor' property is an enumeration that may be specified as: - One of the following enumeration values: ['left', 'center', 'right'] Returns ------- Any """ return self["xanchor"] @xanchor.setter def xanchor(self, val): self["xanchor"] = val @property def xpad(self): """ Sets the amount of padding (in px) along the x direction. The 'xpad' property is a number and may be specified as: - An int or float in the interval [0, inf] Returns ------- int|float """ return self["xpad"] @xpad.setter def xpad(self, val): self["xpad"] = val @property def xref(self): """ Sets the container `x` refers to. "container" spans the entire `width` of the plot. "paper" refers to the width of the plotting area only. The 'xref' property is an enumeration that may be specified as: - One of the following enumeration values: ['container', 'paper'] Returns ------- Any """ return self["xref"] @xref.setter def xref(self, val): self["xref"] = val @property def y(self): """ Sets the y position with respect to `yref` of the color bar (in plot fraction). When `yref` is "paper", defaults to 0.5 when `orientation` is "v" and 1.02 when `orientation` is "h". When `yref` is "container", defaults to 0.5 when `orientation` is "v" and 1 when `orientation` is "h". Must be between 0 and 1 if `yref` is "container" and between "-2" and 3 if `yref` is "paper". The 'y' property is a number and may be specified as: - An int or float Returns ------- int|float """ return self["y"] @y.setter def y(self, val): self["y"] = val @property def yanchor(self): """ Sets this color bar's vertical position anchor This anchor binds the `y` position to the "top", "middle" or "bottom" of the color bar. Defaults to "middle" when `orientation` is "v" and "bottom" when `orientation` is "h". The 'yanchor' property is an enumeration that may be specified as: - One of the following enumeration values: ['top', 'middle', 'bottom'] Returns ------- Any """ return self["yanchor"] @yanchor.setter def yanchor(self, val): self["yanchor"] = val @property def ypad(self): """ Sets the amount of padding (in px) along the y direction. The 'ypad' property is a number and may be specified as: - An int or float in the interval [0, inf] Returns ------- int|float """ return self["ypad"] @ypad.setter def ypad(self, val): self["ypad"] = val @property def yref(self): """ Sets the container `y` refers to. "container" spans the entire `height` of the plot. "paper" refers to the height of the plotting area only. The 'yref' property is an enumeration that may be specified as: - One of the following enumeration values: ['container', 'paper'] Returns ------- Any """ return self["yref"] @yref.setter def yref(self, val): self["yref"] = val @property def _prop_descriptions(self): return """\ bgcolor Sets the color of padded area. bordercolor Sets the axis line color. borderwidth Sets the width (in px) or the border enclosing this color bar. dtick Sets the step in-between ticks on this axis. Use with `tick0`. Must be a positive number, or special strings available to "log" and "date" axes. If the axis `type` is "log", then ticks are set every 10^(n*dtick) where n is the tick number. For example, to set a tick mark at 1, 10, 100, 1000, ... set dtick to 1. To set tick marks at 1, 100, 10000, ... set dtick to 2. To set tick marks at 1, 5, 25, 125, 625, 3125, ... set dtick to log_10(5), or 0.69897000433. "log" has several special values; "L<f>", where `f` is a positive number, gives ticks linearly spaced in value (but not position). For example `tick0` = 0.1, `dtick` = "L0.5" will put ticks at 0.1, 0.6, 1.1, 1.6 etc. To show powers of 10 plus small digits between, use "D1" (all digits) or "D2" (only 2 and 5). `tick0` is ignored for "D1" and "D2". If the axis `type` is "date", then you must convert the time to milliseconds. For example, to set the interval between ticks to one day, set `dtick` to 86400000.0. "date" also has special values "M<n>" gives ticks spaced by a number of months. `n` must be a positive integer. To set ticks on the 15th of every third month, set `tick0` to "2000-01-15" and `dtick` to "M3". To set ticks every 4 years, set `dtick` to "M48" exponentformat Determines a formatting rule for the tick exponents. For example, consider the number 1,000,000,000. If "none", it appears as 1,000,000,000. If "e", 1e+9. If "E", 1E+9. If "power", 1x10^9 (with 9 in a super script). If "SI", 1G. If "B", 1B. "SI" uses prefixes from "femto" f (10^-15) to "tera" T (10^12). *SI extended* covers instead the full SI range from "quecto" q (10^-30) to "quetta" Q (10^30). If "SI" or *SI extended* is used and the exponent is beyond the above ranges, the formatting rule will automatically be switched to the power notation. labelalias Replacement text for specific tick or hover labels. For example using {US: 'USA', CA: 'Canada'} changes US to USA and CA to Canada. The labels we would have shown must match the keys exactly, after adding any tickprefix or ticksuffix. For negative numbers the minus sign symbol used (U+2212) is wider than the regular ascii dash. That means you need to use −1 instead of -1. labelalias can be used with any axis type, and both keys (if needed) and values (if desired) can include html-like tags or MathJax. len Sets the length of the color bar This measure excludes the padding of both ends. That is, the color bar length is this length minus the padding on both ends. lenmode Determines whether this color bar's length (i.e. the measure in the color variation direction) is set in units of plot "fraction" or in *pixels. Use `len` to set the value. minexponent Hide SI prefix for 10^n if |n| is below this number. This only has an effect when `tickformat` is "SI" or "B". nticks Specifies the maximum number of ticks for the particular axis. The actual number of ticks will be chosen automatically to be less than or equal to `nticks`. Has an effect only if `tickmode` is set to "auto". orientation Sets the orientation of the colorbar. outlinecolor Sets the axis line color. outlinewidth Sets the width (in px) of the axis line. separatethousands If "true", even 4-digit integers are separated showexponent If "all", all exponents are shown besides their significands. If "first", only the exponent of the first tick is shown. If "last", only the exponent of the last tick is shown. If "none", no exponents appear. showticklabels Determines whether or not the tick labels are drawn. showtickprefix If "all", all tick labels are displayed with a prefix. If "first", only the first tick is displayed with a prefix. If "last", only the last tick is displayed with a suffix. If "none", tick prefixes are hidden. showticksuffix Same as `showtickprefix` but for tick suffixes. thickness Sets the thickness of the color bar This measure excludes the size of the padding, ticks and labels. thicknessmode Determines whether this color bar's thickness (i.e. the measure in the constant color direction) is set in units of plot "fraction" or in "pixels". Use `thickness` to set the value. tick0 Sets the placement of the first tick on this axis. Use with `dtick`. If the axis `type` is "log", then you must take the log of your starting tick (e.g. to set the starting tick to 100, set the `tick0` to 2) except when `dtick`=*L<f>* (see `dtick` for more info). If the axis `type` is "date", it should be a date string, like date data. If the axis `type` is "category", it should be a number, using the scale where each category is assigned a serial number from zero in the order it appears. tickangle Sets the angle of the tick labels with respect to the horizontal. For example, a `tickangle` of -90 draws the tick labels vertically. tickcolor Sets the tick color. tickfont Sets the color bar's tick label font tickformat Sets the tick label formatting rule using d3 formatting mini-languages which are very similar to those in Python. For numbers, see: https://github.com/d3/d3-format/tree/v1.4.5#d3-format. And for dates see: https://github.com/d3/d3-time- format/tree/v2.2.3#locale_format. We add two items to d3's date formatter: "%h" for half of the year as a decimal number as well as "%{n}f" for fractional seconds with n digits. For example, *2016-10-13 09:15:23.456* with tickformat "%H~%M~%S.%2f" would display "09~15~23.46" tickformatstops A tuple of :class:`plotly.graph_objects.scattergeo.mark er.colorbar.Tickformatstop` instances or dicts with compatible properties tickformatstopdefaults When used in a template (as layout.template.data.scatte rgeo.marker.colorbar.tickformatstopdefaults), sets the default property values to use for elements of scattergeo.marker.colorbar.tickformatstops ticklabeloverflow Determines how we handle tick labels that would overflow either the graph div or the domain of the axis. The default value for inside tick labels is *hide past domain*. In other cases the default is *hide past div*. ticklabelposition Determines where tick labels are drawn relative to the ticks. Left and right options are used when `orientation` is "h", top and bottom when `orientation` is "v". ticklabelstep Sets the spacing between tick labels as compared to the spacing between ticks. A value of 1 (default) means each tick gets a label. A value of 2 means shows every 2nd label. A larger value n means only every nth tick is labeled. `tick0` determines which labels are shown. Not implemented for axes with `type` "log" or "multicategory", or when `tickmode` is "array". ticklen Sets the tick length (in px). tickmode Sets the tick mode for this axis. If "auto", the number of ticks is set via `nticks`. If "linear", the placement of the ticks is determined by a starting position `tick0` and a tick step `dtick` ("linear" is the default value if `tick0` and `dtick` are provided). If "array", the placement of the ticks is set via `tickvals` and the tick text is `ticktext`. ("array" is the default value if `tickvals` is provided). tickprefix Sets a tick label prefix. ticks Determines whether ticks are drawn or not. If "", this axis' ticks are not drawn. If "outside" ("inside"), this axis' are drawn outside (inside) the axis lines. ticksuffix Sets a tick label suffix. ticktext Sets the text displayed at the ticks position via `tickvals`. Only has an effect if `tickmode` is set to "array". Used with `tickvals`. ticktextsrc Sets the source reference on Chart Studio Cloud for `ticktext`. tickvals Sets the values at which ticks on this axis appear. Only has an effect if `tickmode` is set to "array". Used with `ticktext`. tickvalssrc Sets the source reference on Chart Studio Cloud for `tickvals`. tickwidth Sets the tick width (in px). title :class:`plotly.graph_objects.scattergeo.marker.colorbar .Title` instance or dict with compatible properties x Sets the x position with respect to `xref` of the color bar (in plot fraction). When `xref` is "paper", defaults to 1.02 when `orientation` is "v" and 0.5 when `orientation` is "h". When `xref` is "container", defaults to 1 when `orientation` is "v" and 0.5 when `orientation` is "h". Must be between 0 and 1 if `xref` is "container" and between "-2" and 3 if `xref` is "paper". xanchor Sets this color bar's horizontal position anchor. This anchor binds the `x` position to the "left", "center" or "right" of the color bar. Defaults to "left" when `orientation` is "v" and "center" when `orientation` is "h". xpad Sets the amount of padding (in px) along the x direction. xref Sets the container `x` refers to. "container" spans the entire `width` of the plot. "paper" refers to the width of the plotting area only. y Sets the y position with respect to `yref` of the color bar (in plot fraction). When `yref` is "paper", defaults to 0.5 when `orientation` is "v" and 1.02 when `orientation` is "h". When `yref` is "container", defaults to 0.5 when `orientation` is "v" and 1 when `orientation` is "h". Must be between 0 and 1 if `yref` is "container" and between "-2" and 3 if `yref` is "paper". yanchor Sets this color bar's vertical position anchor This anchor binds the `y` position to the "top", "middle" or "bottom" of the color bar. Defaults to "middle" when `orientation` is "v" and "bottom" when `orientation` is "h". ypad Sets the amount of padding (in px) along the y direction. yref Sets the container `y` refers to. "container" spans the entire `height` of the plot. "paper" refers to the height of the plotting area only. """ def __init__( self, arg=None, bgcolor=None, bordercolor=None, borderwidth=None, dtick=None, exponentformat=None, labelalias=None, len=None, lenmode=None, minexponent=None, nticks=None, orientation=None, outlinecolor=None, outlinewidth=None, separatethousands=None, showexponent=None, showticklabels=None, showtickprefix=None, showticksuffix=None, thickness=None, thicknessmode=None, tick0=None, tickangle=None, tickcolor=None, tickfont=None, tickformat=None, tickformatstops=None, tickformatstopdefaults=None, ticklabeloverflow=None, ticklabelposition=None, ticklabelstep=None, ticklen=None, tickmode=None, tickprefix=None, ticks=None, ticksuffix=None, ticktext=None, ticktextsrc=None, tickvals=None, tickvalssrc=None, tickwidth=None, title=None, x=None, xanchor=None, xpad=None, xref=None, y=None, yanchor=None, ypad=None, yref=None, **kwargs, ): """ Construct a new ColorBar object Parameters ---------- arg dict of properties compatible with this constructor or an instance of :class:`plotly.graph_objs.scattergeo.marker.ColorBar` bgcolor Sets the color of padded area. bordercolor Sets the axis line color. borderwidth Sets the width (in px) or the border enclosing this color bar. dtick Sets the step in-between ticks on this axis. Use with `tick0`. Must be a positive number, or special strings available to "log" and "date" axes. If the axis `type` is "log", then ticks are set every 10^(n*dtick) where n is the tick number. For example, to set a tick mark at 1, 10, 100, 1000, ... set dtick to 1. To set tick marks at 1, 100, 10000, ... set dtick to 2. To set tick marks at 1, 5, 25, 125, 625, 3125, ... set dtick to log_10(5), or 0.69897000433. "log" has several special values; "L<f>", where `f` is a positive number, gives ticks linearly spaced in value (but not position). For example `tick0` = 0.1, `dtick` = "L0.5" will put ticks at 0.1, 0.6, 1.1, 1.6 etc. To show powers of 10 plus small digits between, use "D1" (all digits) or "D2" (only 2 and 5). `tick0` is ignored for "D1" and "D2". If the axis `type` is "date", then you must convert the time to milliseconds. For example, to set the interval between ticks to one day, set `dtick` to 86400000.0. "date" also has special values "M<n>" gives ticks spaced by a number of months. `n` must be a positive integer. To set ticks on the 15th of every third month, set `tick0` to "2000-01-15" and `dtick` to "M3". To set ticks every 4 years, set `dtick` to "M48" exponentformat Determines a formatting rule for the tick exponents. For example, consider the number 1,000,000,000. If "none", it appears as 1,000,000,000. If "e", 1e+9. If "E", 1E+9. If "power", 1x10^9 (with 9 in a super script). If "SI", 1G. If "B", 1B. "SI" uses prefixes from "femto" f (10^-15) to "tera" T (10^12). *SI extended* covers instead the full SI range from "quecto" q (10^-30) to "quetta" Q (10^30). If "SI" or *SI extended* is used and the exponent is beyond the above ranges, the formatting rule will automatically be switched to the power notation. labelalias Replacement text for specific tick or hover labels. For example using {US: 'USA', CA: 'Canada'} changes US to USA and CA to Canada. The labels we would have shown must match the keys exactly, after adding any tickprefix or ticksuffix. For negative numbers the minus sign symbol used (U+2212) is wider than the regular ascii dash. That means you need to use −1 instead of -1. labelalias can be used with any axis type, and both keys (if needed) and values (if desired) can include html-like tags or MathJax. len Sets the length of the color bar This measure excludes the padding of both ends. That is, the color bar length is this length minus the padding on both ends. lenmode Determines whether this color bar's length (i.e. the measure in the color variation direction) is set in units of plot "fraction" or in *pixels. Use `len` to set the value. minexponent Hide SI prefix for 10^n if |n| is below this number. This only has an effect when `tickformat` is "SI" or "B". nticks Specifies the maximum number of ticks for the particular axis. The actual number of ticks will be chosen automatically to be less than or equal to `nticks`. Has an effect only if `tickmode` is set to "auto". orientation Sets the orientation of the colorbar. outlinecolor Sets the axis line color. outlinewidth Sets the width (in px) of the axis line. separatethousands If "true", even 4-digit integers are separated showexponent If "all", all exponents are shown besides their significands. If "first", only the exponent of the first tick is shown. If "last", only the exponent of the last tick is shown. If "none", no exponents appear. showticklabels Determines whether or not the tick labels are drawn. showtickprefix If "all", all tick labels are displayed with a prefix. If "first", only the first tick is displayed with a prefix. If "last", only the last tick is displayed with a suffix. If "none", tick prefixes are hidden. showticksuffix Same as `showtickprefix` but for tick suffixes. thickness Sets the thickness of the color bar This measure excludes the size of the padding, ticks and labels. thicknessmode Determines whether this color bar's thickness (i.e. the measure in the constant color direction) is set in units of plot "fraction" or in "pixels". Use `thickness` to set the value. tick0 Sets the placement of the first tick on this axis. Use with `dtick`. If the axis `type` is "log", then you must take the log of your starting tick (e.g. to set the starting tick to 100, set the `tick0` to 2) except when `dtick`=*L<f>* (see `dtick` for more info). If the axis `type` is "date", it should be a date string, like date data. If the axis `type` is "category", it should be a number, using the scale where each category is assigned a serial number from zero in the order it appears. tickangle Sets the angle of the tick labels with respect to the horizontal. For example, a `tickangle` of -90 draws the tick labels vertically. tickcolor Sets the tick color. tickfont Sets the color bar's tick label font tickformat Sets the tick label formatting rule using d3 formatting mini-languages which are very similar to those in Python. For numbers, see: https://github.com/d3/d3-format/tree/v1.4.5#d3-format. And for dates see: https://github.com/d3/d3-time- format/tree/v2.2.3#locale_format. We add two items to d3's date formatter: "%h" for half of the year as a decimal number as well as "%{n}f" for fractional seconds with n digits. For example, *2016-10-13 09:15:23.456* with tickformat "%H~%M~%S.%2f" would display "09~15~23.46" tickformatstops A tuple of :class:`plotly.graph_objects.scattergeo.mark er.colorbar.Tickformatstop` instances or dicts with compatible properties tickformatstopdefaults When used in a template (as layout.template.data.scatte rgeo.marker.colorbar.tickformatstopdefaults), sets the default property values to use for elements of scattergeo.marker.colorbar.tickformatstops ticklabeloverflow Determines how we handle tick labels that would overflow either the graph div or the domain of the axis. The default value for inside tick labels is *hide past domain*. In other cases the default is *hide past div*. ticklabelposition Determines where tick labels are drawn relative to the ticks. Left and right options are used when `orientation` is "h", top and bottom when `orientation` is "v". ticklabelstep Sets the spacing between tick labels as compared to the spacing between ticks. A value of 1 (default) means each tick gets a label. A value of 2 means shows every 2nd label. A larger value n means only every nth tick is labeled. `tick0` determines which labels are shown. Not implemented for axes with `type` "log" or "multicategory", or when `tickmode` is "array". ticklen Sets the tick length (in px). tickmode Sets the tick mode for this axis. If "auto", the number of ticks is set via `nticks`. If "linear", the placement of the ticks is determined by a starting position `tick0` and a tick step `dtick` ("linear" is the default value if `tick0` and `dtick` are provided). If "array", the placement of the ticks is set via `tickvals` and the tick text is `ticktext`. ("array" is the default value if `tickvals` is provided). tickprefix Sets a tick label prefix. ticks Determines whether ticks are drawn or not. If "", this axis' ticks are not drawn. If "outside" ("inside"), this axis' are drawn outside (inside) the axis lines. ticksuffix Sets a tick label suffix. ticktext Sets the text displayed at the ticks position via `tickvals`. Only has an effect if `tickmode` is set to "array". Used with `tickvals`. ticktextsrc Sets the source reference on Chart Studio Cloud for `ticktext`. tickvals Sets the values at which ticks on this axis appear. Only has an effect if `tickmode` is set to "array". Used with `ticktext`. tickvalssrc Sets the source reference on Chart Studio Cloud for `tickvals`. tickwidth Sets the tick width (in px). title :class:`plotly.graph_objects.scattergeo.marker.colorbar .Title` instance or dict with compatible properties x Sets the x position with respect to `xref` of the color bar (in plot fraction). When `xref` is "paper", defaults to 1.02 when `orientation` is "v" and 0.5 when `orientation` is "h". When `xref` is "container", defaults to 1 when `orientation` is "v" and 0.5 when `orientation` is "h". Must be between 0 and 1 if `xref` is "container" and between "-2" and 3 if `xref` is "paper". xanchor Sets this color bar's horizontal position anchor. This anchor binds the `x` position to the "left", "center" or "right" of the color bar. Defaults to "left" when `orientation` is "v" and "center" when `orientation` is "h". xpad Sets the amount of padding (in px) along the x direction. xref Sets the container `x` refers to. "container" spans the entire `width` of the plot. "paper" refers to the width of the plotting area only. y Sets the y position with respect to `yref` of the color bar (in plot fraction). When `yref` is "paper", defaults to 0.5 when `orientation` is "v" and 1.02 when `orientation` is "h". When `yref` is "container", defaults to 0.5 when `orientation` is "v" and 1 when `orientation` is "h". Must be between 0 and 1 if `yref` is "container" and between "-2" and 3 if `yref` is "paper". yanchor Sets this color bar's vertical position anchor This anchor binds the `y` position to the "top", "middle" or "bottom" of the color bar. Defaults to "middle" when `orientation` is "v" and "bottom" when `orientation` is "h". ypad Sets the amount of padding (in px) along the y direction. yref Sets the container `y` refers to. "container" spans the entire `height` of the plot. "paper" refers to the height of the plotting area only. Returns ------- ColorBar """ super().__init__("colorbar") if "_parent" in kwargs: self._parent = kwargs["_parent"] return if arg is None: arg = {} elif isinstance(arg, self.__class__): arg = arg.to_plotly_json() elif isinstance(arg, dict): arg = _copy.copy(arg) else: raise ValueError("""\ The first argument to the plotly.graph_objs.scattergeo.marker.ColorBar constructor must be a dict or an instance of :class:`plotly.graph_objs.scattergeo.marker.ColorBar`""") self._skip_invalid = kwargs.pop("skip_invalid", False) self._validate = kwargs.pop("_validate", True) self._set_property("bgcolor", arg, bgcolor) self._set_property("bordercolor", arg, bordercolor) self._set_property("borderwidth", arg, borderwidth) self._set_property("dtick", arg, dtick) self._set_property("exponentformat", arg, exponentformat) self._set_property("labelalias", arg, labelalias) self._set_property("len", arg, len) self._set_property("lenmode", arg, lenmode) self._set_property("minexponent", arg, minexponent) self._set_property("nticks", arg, nticks) self._set_property("orientation", arg, orientation) self._set_property("outlinecolor", arg, outlinecolor) self._set_property("outlinewidth", arg, outlinewidth) self._set_property("separatethousands", arg, separatethousands) self._set_property("showexponent", arg, showexponent) self._set_property("showticklabels", arg, showticklabels) self._set_property("showtickprefix", arg, showtickprefix) self._set_property("showticksuffix", arg, showticksuffix) self._set_property("thickness", arg, thickness) self._set_property("thicknessmode", arg, thicknessmode) self._set_property("tick0", arg, tick0) self._set_property("tickangle", arg, tickangle) self._set_property("tickcolor", arg, tickcolor) self._set_property("tickfont", arg, tickfont) self._set_property("tickformat", arg, tickformat) self._set_property("tickformatstops", arg, tickformatstops) self._set_property("tickformatstopdefaults", arg, tickformatstopdefaults) self._set_property("ticklabeloverflow", arg, ticklabeloverflow) self._set_property("ticklabelposition", arg, ticklabelposition) self._set_property("ticklabelstep", arg, ticklabelstep) self._set_property("ticklen", arg, ticklen) self._set_property("tickmode", arg, tickmode) self._set_property("tickprefix", arg, tickprefix) self._set_property("ticks", arg, ticks) self._set_property("ticksuffix", arg, ticksuffix) self._set_property("ticktext", arg, ticktext) self._set_property("ticktextsrc", arg, ticktextsrc) self._set_property("tickvals", arg, tickvals) self._set_property("tickvalssrc", arg, tickvalssrc) self._set_property("tickwidth", arg, tickwidth) self._set_property("title", arg, title) self._set_property("x", arg, x) self._set_property("xanchor", arg, xanchor) self._set_property("xpad", arg, xpad) self._set_property("xref", arg, xref) self._set_property("y", arg, y) self._set_property("yanchor", arg, yanchor) self._set_property("ypad", arg, ypad) self._set_property("yref", arg, yref) self._process_kwargs(**dict(arg, **kwargs)) self._skip_invalid = False
ColorBar
python
huggingface__transformers
src/transformers/models/depth_pro/modeling_depth_pro.py
{ "start": 29067, "end": 30441 }
class ____(nn.Module): def __init__(self, config: DepthProConfig, use_deconv: bool = True): super().__init__() self.config = config self.use_deconv = use_deconv self.residual_layer1 = DepthProPreActResidualLayer(config) self.residual_layer2 = DepthProPreActResidualLayer(config) if self.use_deconv: self.deconv = nn.ConvTranspose2d( in_channels=config.fusion_hidden_size, out_channels=config.fusion_hidden_size, kernel_size=2, stride=2, padding=0, bias=False, ) self.projection = nn.Conv2d(config.fusion_hidden_size, config.fusion_hidden_size, kernel_size=1, bias=True) def forward(self, hidden_state: torch.Tensor, residual: Optional[torch.Tensor] = None) -> torch.Tensor: if residual is not None: residual = self.residual_layer1(residual) hidden_state = hidden_state + residual hidden_state = self.residual_layer2(hidden_state) if self.use_deconv: hidden_state = self.deconv(hidden_state) hidden_state = self.projection(hidden_state) return hidden_state # Modified from transformers.models.dpt.modeling_dpt.DPTFeatureFusionStage with DPT->DepthPro # with deconv and reversed layers
DepthProFeatureFusionLayer
python
dagster-io__dagster
docs/sphinx/_ext/sphinx-mdx-builder/tests/dummy_module.py
{ "start": 2020, "end": 3039 }
class ____: """Base vehicle class. This is a base class that demonstrates class documentation with attributes and methods. Attributes: make: The vehicle manufacturer model: The vehicle model year: The manufacturing year """ def __init__(self, make: str, model: str, year: int): """Initialize a new vehicle. Args: make: The vehicle manufacturer model: The vehicle model year: The manufacturing year """ self.make = make self.model = model self.year = year def start_engine(self) -> bool: """Start the vehicle engine. Returns: True if engine started successfully, False otherwise """ return True def get_info(self) -> dict[str, Any]: """Get vehicle information. Returns: Dictionary containing vehicle details """ return {"make": self.make, "model": self.model, "year": self.year}
Vehicle
python
huggingface__transformers
src/transformers/models/audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py
{ "start": 9328, "end": 9931 }
class ____(nn.Module): def __init__(self, config: ASTConfig): super().__init__() self.dense = nn.Linear(config.intermediate_size, config.hidden_size) self.dropout = nn.Dropout(config.hidden_dropout_prob) def forward(self, hidden_states: torch.Tensor, input_tensor: torch.Tensor) -> torch.Tensor: hidden_states = self.dense(hidden_states) hidden_states = self.dropout(hidden_states) hidden_states = hidden_states + input_tensor return hidden_states # Copied from transformers.models.vit.modeling_vit.ViTLayer with ViT->AST,VIT->AST
ASTOutput
python
huggingface__transformers
src/transformers/generation/logits_process.py
{ "start": 21539, "end": 24693 }
class ____(LogitsProcessor): """ [`LogitsProcessor`] that performs top-p, i.e. restricting to top tokens summing to prob_cut_off <= prob_cut_off. Often used together with [`TemperatureLogitsWarper`] and [`TopKLogitsWarper`]. Args: top_p (`float`): If set to < 1, only the smallest set of most probable tokens with probabilities that add up to `top_p` or higher are kept for generation. filter_value (`float`, *optional*, defaults to -inf): All filtered values will be set to this float value. min_tokens_to_keep (`int`, *optional*, defaults to 1): Minimum number of tokens that cannot be filtered. Examples: ```python >>> from transformers import AutoTokenizer, AutoModelForCausalLM, set_seed >>> set_seed(1) >>> model = AutoModelForCausalLM.from_pretrained("distilbert/distilgpt2") >>> tokenizer = AutoTokenizer.from_pretrained("distilbert/distilgpt2") >>> inputs = tokenizer("A sequence: 1, 2", return_tensors="pt") >>> # With sampling, the output is unexpected -- sometimes too unexpected. >>> outputs = model.generate(**inputs, do_sample=True) >>> print(tokenizer.batch_decode(outputs, skip_special_tokens=True)[0]) A sequence: 1, 2, 3 | < 4 (left-hand pointer) ; <BLANKLINE> <BLANKLINE> >>> # With `top_p` sampling, the output gets restricted to high-probability tokens. >>> # Pro tip: In practice, LLMs use `top_p` in the 0.9-0.95 range. >>> outputs = model.generate(**inputs, do_sample=True, top_p=0.1) >>> print(tokenizer.batch_decode(outputs, skip_special_tokens=True)[0]) A sequence: 1, 2, 3, 4, 5, 6, 7, 8, 9 ``` """ def __init__(self, top_p: float, filter_value: float = -float("Inf"), min_tokens_to_keep: int = 1): top_p = float(top_p) if top_p < 0 or top_p > 1.0: raise ValueError(f"`top_p` has to be a float > 0 and < 1, but is {top_p}") if not isinstance(min_tokens_to_keep, int) or (min_tokens_to_keep < 1): raise ValueError(f"`min_tokens_to_keep` has to be a positive integer, but is {min_tokens_to_keep}") self.top_p = top_p self.filter_value = filter_value self.min_tokens_to_keep = min_tokens_to_keep @add_start_docstrings(LOGITS_PROCESSOR_INPUTS_DOCSTRING) def __call__(self, input_ids: torch.LongTensor, scores: torch.FloatTensor) -> torch.FloatTensor: sorted_logits, sorted_indices = torch.sort(scores, descending=False) cumulative_probs = sorted_logits.softmax(dim=-1).cumsum(dim=-1) # Remove tokens with cumulative top_p above the threshold (token with 0 are kept) sorted_indices_to_remove = cumulative_probs <= (1 - self.top_p) # Keep at least min_tokens_to_keep sorted_indices_to_remove[..., -self.min_tokens_to_keep :] = 0 # scatter sorted tensors to original indexing indices_to_remove = sorted_indices_to_remove.scatter(1, sorted_indices, sorted_indices_to_remove) scores_processed = scores.masked_fill(indices_to_remove, self.filter_value) return scores_processed
TopPLogitsWarper
python
apache__airflow
providers/google/tests/unit/google/cloud/operators/test_vertex_ai.py
{ "start": 109012, "end": 109923 }
class ____: @mock.patch(VERTEX_AI_PATH.format("model_service.ModelServiceHook")) def test_execute(self, mock_hook): op = DeleteModelOperator( task_id=TASK_ID, gcp_conn_id=GCP_CONN_ID, impersonation_chain=IMPERSONATION_CHAIN, region=GCP_LOCATION, project_id=GCP_PROJECT, model_id=TEST_MODEL_ID, retry=RETRY, timeout=TIMEOUT, metadata=METADATA, ) op.execute(context={}) mock_hook.assert_called_once_with(gcp_conn_id=GCP_CONN_ID, impersonation_chain=IMPERSONATION_CHAIN) mock_hook.return_value.delete_model.assert_called_once_with( region=GCP_LOCATION, project_id=GCP_PROJECT, model=TEST_MODEL_ID, retry=RETRY, timeout=TIMEOUT, metadata=METADATA, )
TestVertexAIDeleteModelOperator
python
scikit-learn__scikit-learn
sklearn/model_selection/_split.py
{ "start": 8067, "end": 11057 }
class ____(_UnsupportedGroupCVMixin, BaseCrossValidator): """Leave-P-Out cross-validator. Provides train/test indices to split data in train/test sets. This results in testing on all distinct samples of size p, while the remaining n - p samples form the training set in each iteration. Note: ``LeavePOut(p)`` is NOT equivalent to ``KFold(n_splits=n_samples // p)`` which creates non-overlapping test sets. Due to the high number of iterations which grows combinatorically with the number of samples this cross-validation method can be very costly. For large datasets one should favor :class:`KFold`, :class:`StratifiedKFold` or :class:`ShuffleSplit`. Read more in the :ref:`User Guide <leave_p_out>`. Parameters ---------- p : int Size of the test sets. Must be strictly less than the number of samples. Examples -------- >>> import numpy as np >>> from sklearn.model_selection import LeavePOut >>> X = np.array([[1, 2], [3, 4], [5, 6], [7, 8]]) >>> y = np.array([1, 2, 3, 4]) >>> lpo = LeavePOut(2) >>> lpo.get_n_splits(X) 6 >>> print(lpo) LeavePOut(p=2) >>> for i, (train_index, test_index) in enumerate(lpo.split(X)): ... print(f"Fold {i}:") ... print(f" Train: index={train_index}") ... print(f" Test: index={test_index}") Fold 0: Train: index=[2 3] Test: index=[0 1] Fold 1: Train: index=[1 3] Test: index=[0 2] Fold 2: Train: index=[1 2] Test: index=[0 3] Fold 3: Train: index=[0 3] Test: index=[1 2] Fold 4: Train: index=[0 2] Test: index=[1 3] Fold 5: Train: index=[0 1] Test: index=[2 3] """ def __init__(self, p): self.p = p def _iter_test_indices(self, X, y=None, groups=None): n_samples = _num_samples(X) if n_samples <= self.p: raise ValueError( "p={} must be strictly less than the number of samples={}".format( self.p, n_samples ) ) for combination in combinations(range(n_samples), self.p): yield np.array(combination) def get_n_splits(self, X, y=None, groups=None): """Returns the number of splitting iterations in the cross-validator. Parameters ---------- X : array-like of shape (n_samples, n_features) Training data, where `n_samples` is the number of samples and `n_features` is the number of features. y : array-like of shape (n_samples,), default=None Always ignored, exists for API compatibility. groups : array-like of shape (n_samples,), default=None Always ignored, exists for API compatibility. """ if X is None: raise ValueError("The 'X' parameter should not be None.") return int(comb(_num_samples(X), self.p, exact=True))
LeavePOut
python
pypa__warehouse
tests/unit/admin/views/test_emails.py
{ "start": 8663, "end": 9175 }
class ____: def test_existing_email(self, db_session): em = EmailMessageFactory.create() request = pretend.stub(matchdict={"email_id": em.id}, db=db_session) assert views.email_detail(request) == {"email": em} def test_nonexistent_email(self, db_session): EmailMessageFactory.create() request = pretend.stub(matchdict={"email_id": str(uuid.uuid4())}, db=db_session) with pytest.raises(HTTPNotFound): views.email_detail(request)
TestEmailDetail
python
ansible__ansible
test/integration/targets/config/lookup_plugins/broken.py
{ "start": 1110, "end": 1379 }
class ____(LookupBase): def run(self, terms, variables=None, **kwargs): """Generate list.""" self.set_options(var_options=variables, direct=kwargs) return [self.get_option("some_option"), *self.get_option_and_origin("some_option")]
LookupModule
python
walkccc__LeetCode
solutions/2306. Naming a Company/2306.py
{ "start": 0, "end": 498 }
class ____: def distinctNames(self, ideas: list[str]) -> int: ans = 0 # suffixes[i] := the set of strings omitting the first letter, where the # first letter is ('a' + i) suffixes = [set() for _ in range(26)] for idea in ideas: suffixes[ord(idea[0]) - ord('a')].add(idea[1:]) for i, j in itertools.combinations(range(26), 2): count = len(suffixes[i] & suffixes[j]) ans += 2 * (len(suffixes[i]) - count) * (len(suffixes[j]) - count) return ans
Solution
python
numba__numba
numba/tests/npyufunc/test_errors.py
{ "start": 2018, "end": 5497 }
class ____(TestCase, CheckWarningsMixin): """ Test floating-point exceptions inside ufuncs. Note the warnings emitted by Numpy reflect IEEE-754 semantics. """ def check_truediv_real(self, dtype): """ Test 1 / 0 and 0 / 0. """ f = vectorize(nopython=True)(truediv) a = np.array([5., 6., 0., 8.], dtype=dtype) b = np.array([1., 0., 0., 4.], dtype=dtype) expected = np.array([5., float('inf'), float('nan'), 2.]) with self.check_warnings(["divide by zero encountered", "invalid value encountered"]): res = f(a, b) self.assertPreciseEqual(res, expected) def test_truediv_float(self): self.check_truediv_real(np.float64) def test_truediv_integer(self): self.check_truediv_real(np.int32) def check_divmod_float(self, pyfunc, values, messages): """ Test 1 // 0 and 0 // 0. """ f = vectorize(nopython=True)(pyfunc) a = np.array([5., 6., 0., 9.]) b = np.array([1., 0., 0., 4.]) expected = np.array(values) with self.check_warnings(messages): res = f(a, b) self.assertPreciseEqual(res, expected) def test_floordiv_float(self): self.check_divmod_float(floordiv, [5.0, float('inf'), float('nan'), 2.0], ["divide by zero encountered", "invalid value encountered"]) @skip_macos_fenv_errors def test_remainder_float(self): self.check_divmod_float(remainder, [0.0, float('nan'), float('nan'), 1.0], ["invalid value encountered"]) def check_divmod_int(self, pyfunc, values): """ Test 1 % 0 and 0 % 0. """ f = vectorize(nopython=True)(pyfunc) a = np.array([5, 6, 0, 9]) b = np.array([1, 0, 0, 4]) expected = np.array(values) # No warnings raised because LLVM makes it difficult with self.check_warnings([]): res = f(a, b) self.assertPreciseEqual(res, expected) def test_floordiv_int(self): self.check_divmod_int(floordiv, [5, 0, 0, 2]) def test_remainder_int(self): self.check_divmod_int(remainder, [0, 0, 0, 1]) def test_power_float(self): """ Test 0 ** -1 and 2 ** <big number>. """ f = vectorize(nopython=True)(power) a = np.array([5., 0., 2., 8.]) b = np.array([1., -1., 1e20, 4.]) expected = np.array([5., float('inf'), float('inf'), 4096.]) with self.check_warnings(["divide by zero encountered", "overflow encountered"]): res = f(a, b) self.assertPreciseEqual(res, expected) def test_power_integer(self): """ Test 0 ** -1. Note 2 ** <big number> returns an undefined value (depending on the algorithm). """ dtype = np.int64 f = vectorize(["int64(int64, int64)"], nopython=True)(power) a = np.array([5, 0, 6], dtype=dtype) b = np.array([1, -1, 2], dtype=dtype) expected = np.array([5, -2**63, 36], dtype=dtype) with self.check_warnings([]): res = f(a, b) self.assertPreciseEqual(res, expected) if __name__ == "__main__": unittest.main()
TestFloatingPointExceptions
python
kamyu104__LeetCode-Solutions
Python/subarray-sums-divisible-by-k.py
{ "start": 50, "end": 444 }
class ____(object): def subarraysDivByK(self, A, K): """ :type A: List[int] :type K: int :rtype: int """ count = collections.defaultdict(int) count[0] = 1 result, prefix = 0, 0 for a in A: prefix = (prefix+a) % K result += count[prefix] count[prefix] += 1 return result
Solution
python
jmcnamara__XlsxWriter
xlsxwriter/table.py
{ "start": 286, "end": 7088 }
class ____(xmlwriter.XMLwriter): """ A class for writing the Excel XLSX Table file. """ ########################################################################### # # Public API. # ########################################################################### def __init__(self) -> None: """ Constructor. """ super().__init__() self.properties = {} ########################################################################### # # Private API. # ########################################################################### def _assemble_xml_file(self) -> None: # Assemble and write the XML file. # Write the XML declaration. self._xml_declaration() # Write the table element. self._write_table() # Write the autoFilter element. self._write_auto_filter() # Write the tableColumns element. self._write_table_columns() # Write the tableStyleInfo element. self._write_table_style_info() # Write the extLst element for alt text/title. self._write_ext_lst() # Close the table tag. self._xml_end_tag("table") # Close the file. self._xml_close() def _set_properties(self, properties) -> None: # Set the document properties. self.properties = properties ########################################################################### # # XML methods. # ########################################################################### def _write_table(self) -> None: # Write the <table> element. schema = "http://schemas.openxmlformats.org/" xmlns = schema + "spreadsheetml/2006/main" table_id = self.properties["id"] name = self.properties["name"] display_name = self.properties["name"] ref = self.properties["range"] totals_row_shown = self.properties["totals_row_shown"] header_row_count = self.properties["header_row_count"] attributes = [ ("xmlns", xmlns), ("id", table_id), ("name", name), ("displayName", display_name), ("ref", ref), ] if not header_row_count: attributes.append(("headerRowCount", 0)) if totals_row_shown: attributes.append(("totalsRowCount", 1)) else: attributes.append(("totalsRowShown", 0)) self._xml_start_tag("table", attributes) def _write_auto_filter(self) -> None: # Write the <autoFilter> element. autofilter = self.properties.get("autofilter", 0) if not autofilter: return attributes = [ ( "ref", autofilter, ) ] self._xml_empty_tag("autoFilter", attributes) def _write_table_columns(self) -> None: # Write the <tableColumns> element. columns = self.properties["columns"] count = len(columns) attributes = [("count", count)] self._xml_start_tag("tableColumns", attributes) for col_data in columns: # Write the tableColumn element. self._write_table_column(col_data) self._xml_end_tag("tableColumns") def _write_table_column(self, col_data) -> None: # Write the <tableColumn> element. attributes = [ ("id", col_data["id"]), ("name", col_data["name"]), ] if col_data.get("total_string"): attributes.append(("totalsRowLabel", col_data["total_string"])) elif col_data.get("total_function"): attributes.append(("totalsRowFunction", col_data["total_function"])) if "format" in col_data and col_data["format"] is not None: attributes.append(("dataDxfId", col_data["format"])) if col_data.get("formula") or col_data.get("custom_total"): self._xml_start_tag("tableColumn", attributes) if col_data.get("formula"): # Write the calculatedColumnFormula element. self._write_calculated_column_formula(col_data["formula"]) if col_data.get("custom_total"): # Write the totalsRowFormula element. self._write_totals_row_formula(col_data.get("custom_total")) self._xml_end_tag("tableColumn") else: self._xml_empty_tag("tableColumn", attributes) def _write_table_style_info(self) -> None: # Write the <tableStyleInfo> element. props = self.properties attributes = [] name = props["style"] show_first_column = 0 + props["show_first_col"] show_last_column = 0 + props["show_last_col"] show_row_stripes = 0 + props["show_row_stripes"] show_column_stripes = 0 + props["show_col_stripes"] if name is not None and name != "" and name != "None": attributes.append(("name", name)) attributes.append(("showFirstColumn", show_first_column)) attributes.append(("showLastColumn", show_last_column)) attributes.append(("showRowStripes", show_row_stripes)) attributes.append(("showColumnStripes", show_column_stripes)) self._xml_empty_tag("tableStyleInfo", attributes) def _write_calculated_column_formula(self, formula) -> None: # Write the <calculatedColumnFormula> element. self._xml_data_element("calculatedColumnFormula", formula) def _write_totals_row_formula(self, formula) -> None: # Write the <totalsRowFormula> element. self._xml_data_element("totalsRowFormula", formula) def _write_ext_lst(self) -> None: # Write the <extLst> element for the alt text/description. props = self.properties if not props.get("description") and not props.get("title"): return attributes = [ ("uri", "{504A1905-F514-4f6f-8877-14C23A59335A}"), ( "xmlns:x14", "http://schemas.microsoft.com/office/spreadsheetml/2009/9/main", ), ] self._xml_start_tag("extLst") self._xml_start_tag("ext", attributes) # Write the x14:table element. self._write_x14_table() self._xml_end_tag("ext") self._xml_end_tag("extLst") def _write_x14_table(self): # Write the <x14:table> element. props = self.properties attributes = [] if props.get("title"): attributes.append(("altText", props["title"])) if props.get("description"): attributes.append(("altTextSummary", props["description"])) self._xml_empty_tag("x14:table", attributes)
Table
python
run-llama__llama_index
llama-index-integrations/embeddings/llama-index-embeddings-yandexgpt/llama_index/embeddings/yandexgpt/util.py
{ "start": 132, "end": 2397 }
class ____: """Class to handle authentication using folder ID and API key.""" def __init__(self, folder_id: str = None, api_key: str = None) -> None: """ Initialize the YAuth object with folder_id and api_key. :param folder_id: The folder ID for authentication. :param api_key: The API key for authentication. """ self.folder_id = folder_id self.api_key = api_key @property def headers(self) -> Optional[dict]: """ Generate authentication headers. :return: A dictionary containing the authorization headers. """ if self.folder_id is not None and self.api_key is not None: return { "Authorization": f"Api-key {self.api_key}", "x-folder-id": self.folder_id, } return None @staticmethod def from_dict(js: Dict[str, str]) -> Optional["YAuth"]: """ Create a YAuth instance from a dictionary. :param js: A dictionary containing 'folder_id' and 'api_key'. :return: A YAuth instance. :raises YException: If 'folder_id' or 'api_key' is not provided. """ if js.get("folder_id") is not None and js.get("api_key") is not None: return YAuth(js["folder_id"], api_key=js["api_key"]) raise YException( "Cannot create valid authentication object: you need to provide folder_id and either iam token or api_key fields" ) @staticmethod def from_config_file(fn: str) -> "YAuth": """ Create a YAuth instance from a configuration file. :param fn: Path to the JSON configuration file. :return: A YAuth instance. """ with open(fn, encoding="utf-8") as f: js = json.load(f) return YAuth.from_dict(js) @staticmethod def from_params(kwargs) -> "YAuth": """ Create a YAuth instance from parameters. :param kwargs: A dictionary containing either a 'config' file path or direct 'folder_id' and 'api_key'. :return: A YAuth instance. """ if kwargs.get("config") is not None: return YAuth.from_config_file(kwargs["config"]) return YAuth.from_dict(kwargs)
YAuth
python
huggingface__transformers
src/transformers/models/roberta/modeling_roberta.py
{ "start": 18070, "end": 20849 }
class ____(GradientCheckpointingLayer): def __init__(self, config, layer_idx=None): super().__init__() self.chunk_size_feed_forward = config.chunk_size_feed_forward self.seq_len_dim = 1 self.attention = RobertaAttention(config, is_causal=config.is_decoder, layer_idx=layer_idx) self.is_decoder = config.is_decoder self.add_cross_attention = config.add_cross_attention if self.add_cross_attention: if not self.is_decoder: raise ValueError(f"{self} should be used as a decoder model if cross attention is added") self.crossattention = RobertaAttention( config, is_causal=False, layer_idx=layer_idx, is_cross_attention=True, ) self.intermediate = RobertaIntermediate(config) self.output = RobertaOutput(config) def forward( self, hidden_states: torch.Tensor, attention_mask: Optional[torch.FloatTensor] = None, encoder_hidden_states: Optional[torch.FloatTensor] = None, encoder_attention_mask: Optional[torch.FloatTensor] = None, past_key_values: Optional[Cache] = None, cache_position: Optional[torch.Tensor] = None, **kwargs: Unpack[TransformersKwargs], ) -> tuple[torch.Tensor]: self_attention_output, _ = self.attention( hidden_states, attention_mask, past_key_values=past_key_values, cache_position=cache_position, **kwargs, ) attention_output = self_attention_output if self.is_decoder and encoder_hidden_states is not None: if not hasattr(self, "crossattention"): raise ValueError( f"If `encoder_hidden_states` are passed, {self} has to be instantiated with cross-attention layers" " by setting `config.add_cross_attention=True`" ) cross_attention_output, _ = self.crossattention( self_attention_output, None, # attention_mask encoder_hidden_states, encoder_attention_mask, past_key_values=past_key_values, **kwargs, ) attention_output = cross_attention_output layer_output = apply_chunking_to_forward( self.feed_forward_chunk, self.chunk_size_feed_forward, self.seq_len_dim, attention_output ) return layer_output def feed_forward_chunk(self, attention_output): intermediate_output = self.intermediate(attention_output) layer_output = self.output(intermediate_output, attention_output) return layer_output @auto_docstring
RobertaLayer
python
apache__airflow
providers/google/tests/unit/google/cloud/hooks/test_datafusion.py
{ "start": 3057, "end": 23395 }
class ____: @staticmethod def mock_endpoint(get_conn_mock): return get_conn_mock.return_value.projects.return_value.locations.return_value.instances.return_value def test_name(self, hook): expected = f"projects/{PROJECT_ID}/locations/{LOCATION}/instances/{INSTANCE_NAME}" assert hook._name(PROJECT_ID, LOCATION, INSTANCE_NAME) == expected def test_parent(self, hook): expected = f"projects/{PROJECT_ID}/locations/{LOCATION}" assert hook._parent(PROJECT_ID, LOCATION) == expected @mock.patch(HOOK_STR.format("build")) @mock.patch(HOOK_STR.format("DataFusionHook._authorize")) def test_get_conn(self, mock_authorize, mock_build, hook): mock_authorize.return_value = "test" hook.get_conn() mock_build.assert_called_once_with("datafusion", hook.api_version, http="test", cache_discovery=False) @mock.patch(HOOK_STR.format("DataFusionHook.get_conn")) def test_restart_instance(self, get_conn_mock, hook): method_mock = self.mock_endpoint(get_conn_mock).restart method_mock.return_value.execute.return_value = "value" result = hook.restart_instance(instance_name=INSTANCE_NAME, location=LOCATION, project_id=PROJECT_ID) assert result == "value" method_mock.assert_called_once_with(name=hook._name(PROJECT_ID, LOCATION, INSTANCE_NAME)) @mock.patch(HOOK_STR.format("DataFusionHook.get_conn")) def test_delete_instance(self, get_conn_mock, hook): method_mock = self.mock_endpoint(get_conn_mock).delete method_mock.return_value.execute.return_value = "value" result = hook.delete_instance(instance_name=INSTANCE_NAME, location=LOCATION, project_id=PROJECT_ID) assert result == "value" method_mock.assert_called_once_with(name=hook._name(PROJECT_ID, LOCATION, INSTANCE_NAME)) @mock.patch(HOOK_STR.format("DataFusionHook.get_conn")) def test_create_instance(self, get_conn_mock, hook): method_mock = self.mock_endpoint(get_conn_mock).create method_mock.return_value.execute.return_value = "value" result = hook.create_instance( instance_name=INSTANCE_NAME, instance=INSTANCE, location=LOCATION, project_id=PROJECT_ID, ) assert result == "value" method_mock.assert_called_once_with( parent=hook._parent(PROJECT_ID, LOCATION), body=INSTANCE, instanceId=INSTANCE_NAME, ) @mock.patch(HOOK_STR.format("DataFusionHook.get_conn")) def test_patch_instance(self, get_conn_mock, hook): method_mock = self.mock_endpoint(get_conn_mock).patch method_mock.return_value.execute.return_value = "value" result = hook.patch_instance( instance_name=INSTANCE_NAME, instance=INSTANCE, update_mask="instance.name", location=LOCATION, project_id=PROJECT_ID, ) assert result == "value" method_mock.assert_called_once_with( name=hook._name(PROJECT_ID, LOCATION, INSTANCE_NAME), body=INSTANCE, updateMask="instance.name", ) @mock.patch(HOOK_STR.format("DataFusionHook.get_conn")) def test_get_instance(self, get_conn_mock, hook): method_mock = self.mock_endpoint(get_conn_mock).get method_mock.return_value.execute.return_value = "value" result = hook.get_instance(instance_name=INSTANCE_NAME, location=LOCATION, project_id=PROJECT_ID) assert result == "value" method_mock.assert_called_once_with(name=hook._name(PROJECT_ID, LOCATION, INSTANCE_NAME)) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_get_instance_artifacts(self, mock_request, hook): scope = "SYSTEM" artifact = { "name": "test-artifact", "version": "1.2.3", "scope": scope, } mock_request.return_value = mock.MagicMock(status=200, data=json.dumps([artifact])) hook.get_instance_artifacts(instance_url=INSTANCE_URL, scope=scope) mock_request.assert_called_with( url=f"{INSTANCE_URL}/v3/namespaces/default/artifacts", method="GET", params={"scope": scope}, ) @mock.patch("google.auth.transport.requests.Request") @mock.patch(HOOK_STR.format("DataFusionHook.get_credentials")) def test_cdap_request(self, get_credentials_mock, mock_request, hook): url = "test_url" headers = {"Content-Type": "application/json"} method = "POST" request = mock_request.return_value request.return_value = mock.MagicMock() body = {"data": "value"} params = {"param_key": "param_value"} result = hook._cdap_request(url=url, method=method, body=body, params=params) mock_request.assert_called_once_with() get_credentials_mock.assert_called_once_with() get_credentials_mock.return_value.before_request.assert_called_once_with( request=request, method=method, url=url, headers=headers ) request.assert_called_once_with( method=method, url=url, headers=headers, body=json.dumps(body), params=params ) assert result == request.return_value @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_create_pipeline(self, mock_request, hook): mock_request.return_value.status = 200 hook.create_pipeline(pipeline_name=PIPELINE_NAME, pipeline=PIPELINE, instance_url=INSTANCE_URL) mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/apps/{PIPELINE_NAME}", method="PUT", body=PIPELINE, ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_create_pipeline_should_fail_if_empty_data_response(self, mock_request, hook): mock_request.return_value.status = 200 mock_request.return_value.data = None with pytest.raises( AirflowException, match=r"Empty response received. Please, check for possible root causes " r"of this behavior either in DAG code or on Cloud DataFusion side", ): hook.create_pipeline(pipeline_name=PIPELINE_NAME, pipeline=PIPELINE, instance_url=INSTANCE_URL) mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/apps/{PIPELINE_NAME}", method="PUT", body=PIPELINE, ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_create_pipeline_should_fail_if_status_not_200(self, mock_request, hook): mock_request.return_value.status = 404 with pytest.raises(AirflowException, match=r"Creating a pipeline failed with code 404"): hook.create_pipeline(pipeline_name=PIPELINE_NAME, pipeline=PIPELINE, instance_url=INSTANCE_URL) mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/apps/{PIPELINE_NAME}", method="PUT", body=PIPELINE, ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_delete_pipeline(self, mock_request, hook): mock_request.return_value.status = 200 hook.delete_pipeline(pipeline_name=PIPELINE_NAME, instance_url=INSTANCE_URL) mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/apps/{PIPELINE_NAME}", method="DELETE", body=None, ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_delete_pipeline_should_fail_if_empty_data_response(self, mock_request, hook): mock_request.return_value.status = 200 mock_request.return_value.data = None with pytest.raises( AirflowException, match=r"Empty response received. Please, check for possible root causes " r"of this behavior either in DAG code or on Cloud DataFusion side", ): hook.delete_pipeline(pipeline_name=PIPELINE_NAME, instance_url=INSTANCE_URL) mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/apps/{PIPELINE_NAME}", method="DELETE", body=None, ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_delete_pipeline_should_fail_if_status_not_200(self, mock_request, hook): mock_request.return_value.status = 404 with pytest.raises(AirflowException, match=r"Deleting a pipeline failed with code 404"): hook.delete_pipeline(pipeline_name=PIPELINE_NAME, instance_url=INSTANCE_URL) mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/apps/{PIPELINE_NAME}", method="DELETE", body=None, ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_delete_pipeline_should_fail_if_status_409(self, mock_request, hook, caplog): caplog.set_level(logging.INFO) mock_request.side_effect = [ MockResponse(status=409, data="Conflict: Resource is still in use."), MockResponse(status=200, data="Success"), ] hook.delete_pipeline(pipeline_name=PIPELINE_NAME, instance_url=INSTANCE_URL) assert mock_request.call_count == 2 assert "Conflict: Resource is still in use." in caplog.text mock_request.assert_called_with( url=f"{INSTANCE_URL}/v3/namespaces/default/apps/{PIPELINE_NAME}", method="DELETE", body=None, ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_list_pipelines(self, mock_request, hook): data = {"data": "test"} mock_request.return_value.status = 200 mock_request.return_value.data = json.dumps(data) result = hook.list_pipelines(instance_url=INSTANCE_URL) mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/apps", method="GET", body=None ) assert result == data @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_list_pipelines_should_fail_if_empty_data_response(self, mock_request, hook): mock_request.return_value.status = 200 mock_request.return_value.data = None with pytest.raises( AirflowException, match=r"Empty response received. Please, check for possible root causes " r"of this behavior either in DAG code or on Cloud DataFusion side", ): hook.list_pipelines(instance_url=INSTANCE_URL) mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/apps", method="GET", body=None ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_list_pipelines_should_fail_if_status_not_200(self, mock_request, hook): mock_request.return_value.status = 404 with pytest.raises(AirflowException, match=r"Listing pipelines failed with code 404"): hook.list_pipelines(instance_url=INSTANCE_URL) mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/apps", method="GET", body=None ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_start_pipeline(self, mock_request, hook): run_id = 1234 mock_request.return_value = mock.MagicMock(status=200, data=f'[{{"runId":{run_id}}}]') hook.start_pipeline(pipeline_name=PIPELINE_NAME, instance_url=INSTANCE_URL, runtime_args=RUNTIME_ARGS) body = [ { "appId": PIPELINE_NAME, "programType": "workflow", "programId": "DataPipelineWorkflow", "runtimeargs": RUNTIME_ARGS, } ] mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/start", method="POST", body=body ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_start_pipeline_stream(self, mock_request, hook): run_id = 1234 mock_request.return_value = mock.MagicMock(status=200, data=f'[{{"runId":{run_id}}}]') hook.start_pipeline( pipeline_name=PIPELINE_NAME, instance_url=INSTANCE_URL, runtime_args=RUNTIME_ARGS, pipeline_type=DataFusionPipelineType.STREAM, ) body = [ { "appId": PIPELINE_NAME, "programType": "spark", "programId": "DataStreamsSparkStreaming", "runtimeargs": RUNTIME_ARGS, } ] mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/start", method="POST", body=body ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_start_pipeline_should_fail_if_empty_data_response(self, mock_request, hook): mock_request.return_value.status = 200 mock_request.return_value.data = None with pytest.raises( AirflowException, match=r"Empty response received. Please, check for possible root causes " r"of this behavior either in DAG code or on Cloud DataFusion side", ): hook.start_pipeline( pipeline_name=PIPELINE_NAME, instance_url=INSTANCE_URL, runtime_args=RUNTIME_ARGS ) body = [ { "appId": PIPELINE_NAME, "programType": "workflow", "programId": "DataPipelineWorkflow", "runtimeargs": RUNTIME_ARGS, } ] mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/start", method="POST", body=body ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_start_pipeline_should_fail_if_status_not_200(self, mock_request, hook): mock_request.return_value.status = 404 with pytest.raises(AirflowException, match=r"Starting a pipeline failed with code 404"): hook.start_pipeline( pipeline_name=PIPELINE_NAME, instance_url=INSTANCE_URL, runtime_args=RUNTIME_ARGS ) body = [ { "appId": PIPELINE_NAME, "programType": "workflow", "programId": "DataPipelineWorkflow", "runtimeargs": RUNTIME_ARGS, } ] mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/start", method="POST", body=body ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_stop_pipeline(self, mock_request, hook): mock_request.return_value.status = 200 hook.stop_pipeline(pipeline_name=PIPELINE_NAME, instance_url=INSTANCE_URL) mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/apps/{PIPELINE_NAME}/" f"workflows/DataPipelineWorkflow/stop", method="POST", ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_stop_pipeline_should_fail_if_empty_data_response(self, mock_request, hook): mock_request.return_value.status = 200 mock_request.return_value.data = None with pytest.raises( AirflowException, match=r"Empty response received. Please, check for possible root causes " r"of this behavior either in DAG code or on Cloud DataFusion side", ): hook.stop_pipeline(pipeline_name=PIPELINE_NAME, instance_url=INSTANCE_URL) mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/apps/{PIPELINE_NAME}/" f"workflows/DataPipelineWorkflow/stop", method="POST", ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_stop_pipeline_should_fail_if_status_not_200(self, mock_request, hook): mock_request.return_value.status = 404 with pytest.raises(AirflowException, match=r"Stopping a pipeline failed with code 404"): hook.stop_pipeline(pipeline_name=PIPELINE_NAME, instance_url=INSTANCE_URL) mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/apps/{PIPELINE_NAME}/" f"workflows/DataPipelineWorkflow/stop", method="POST", ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_get_pipeline_workflow(self, mock_request, hook): run_id = 1234 mock_request.return_value = mock.MagicMock(status=200, data=f'[{{"runId":{run_id}}}]') hook.get_pipeline_workflow( pipeline_name=PIPELINE_NAME, instance_url=INSTANCE_URL, pipeline_id=PIPELINE_ID ) mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/apps/{PIPELINE_NAME}/" f"workflows/DataPipelineWorkflow/runs/{PIPELINE_ID}", method="GET", ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_get_pipeline_workflow_stream(self, mock_request, hook): run_id = 1234 mock_request.return_value = mock.MagicMock(status=200, data=f'[{{"runId":{run_id}}}]') hook.get_pipeline_workflow( pipeline_name=PIPELINE_NAME, instance_url=INSTANCE_URL, pipeline_id=PIPELINE_ID, pipeline_type=DataFusionPipelineType.STREAM, ) mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/apps/{PIPELINE_NAME}/" f"sparks/DataStreamsSparkStreaming/runs/{PIPELINE_ID}", method="GET", ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_get_pipeline_workflow_should_fail_if_empty_data_response(self, mock_request, hook): mock_request.return_value.status = 200 mock_request.return_value.data = None with pytest.raises( AirflowException, match=r"Empty response received. Please, check for possible root causes " r"of this behavior either in DAG code or on Cloud DataFusion side", ): hook.get_pipeline_workflow( pipeline_name=PIPELINE_NAME, instance_url=INSTANCE_URL, pipeline_id=PIPELINE_ID ) mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/apps/{PIPELINE_NAME}/" f"workflows/DataPipelineWorkflow/runs/{PIPELINE_ID}", method="GET", ) @mock.patch(HOOK_STR.format("DataFusionHook._cdap_request")) def test_get_pipeline_workflow_should_fail_if_status_not_200(self, mock_request, hook): mock_request.return_value.status = 404 with pytest.raises(AirflowException, match=r"Retrieving a pipeline state failed with code 404"): hook.get_pipeline_workflow( pipeline_name=PIPELINE_NAME, instance_url=INSTANCE_URL, pipeline_id=PIPELINE_ID ) mock_request.assert_called_once_with( url=f"{INSTANCE_URL}/v3/namespaces/default/apps/{PIPELINE_NAME}/" f"workflows/DataPipelineWorkflow/runs/{PIPELINE_ID}", method="GET", ) @pytest.mark.parametrize( ("pipeline_type", "expected_program_type"), [ (DataFusionPipelineType.BATCH, "workflow"), (DataFusionPipelineType.STREAM, "spark"), ("non existing value", ""), ], ) def test_cdap_program_type(self, pipeline_type, expected_program_type): assert DataFusionHook.cdap_program_type(pipeline_type) == expected_program_type @pytest.mark.parametrize( ("pipeline_type", "expected_program_id"), [ (DataFusionPipelineType.BATCH, "DataPipelineWorkflow"), (DataFusionPipelineType.STREAM, "DataStreamsSparkStreaming"), ("non existing value", ""), ], ) def test_cdap_program_id(self, pipeline_type, expected_program_id): assert DataFusionHook.cdap_program_id(pipeline_type) == expected_program_id
TestDataFusionHook
python
walkccc__LeetCode
solutions/1959. Minimum Total Space Wasted With K Resizing Operations/1959.py
{ "start": 0, "end": 642 }
class ____: def minSpaceWastedKResizing(self, nums: list[int], k: int) -> int: MAX = 200_000_000 @functools.lru_cache(None) def dp(i: int, k: int) -> int: """ Returns the minimum space wasted for nums[i..n) if you can resize k times. """ if i == len(nums): return 0 if k == -1: return MAX res = MAX summ = 0 maxNum = nums[i] for j in range(i, len(nums)): summ += nums[j] maxNum = max(maxNum, nums[j]) wasted = maxNum * (j - i + 1) - summ res = min(res, dp(j + 1, k - 1) + wasted) return res return dp(0, k)
Solution
python
apache__airflow
providers/amazon/src/airflow/providers/amazon/aws/triggers/ecs.py
{ "start": 4475, "end": 9306 }
class ____(BaseTrigger): """ Waits for an ECS task to be done, while eventually polling logs. :param cluster: short name or full ARN of the cluster where the task is running. :param task_arn: ARN of the task to watch. :param waiter_delay: The amount of time in seconds to wait between attempts. :param waiter_max_attempts: The number of times to ping for status. Will fail after that many unsuccessful attempts. :param aws_conn_id: The Airflow connection used for AWS credentials. :param region: The AWS region where the cluster is located. """ def __init__( self, cluster: str, task_arn: str, waiter_delay: int, waiter_max_attempts: int, aws_conn_id: str | None, region: str | None, log_group: str | None = None, log_stream: str | None = None, ): self.cluster = cluster self.task_arn = task_arn self.waiter_delay = waiter_delay self.waiter_max_attempts = waiter_max_attempts self.aws_conn_id = aws_conn_id self.region = region self.log_group = log_group self.log_stream = log_stream def serialize(self) -> tuple[str, dict[str, Any]]: return ( self.__class__.__module__ + "." + self.__class__.__qualname__, { "cluster": self.cluster, "task_arn": self.task_arn, "waiter_delay": self.waiter_delay, "waiter_max_attempts": self.waiter_max_attempts, "aws_conn_id": self.aws_conn_id, "region": self.region, "log_group": self.log_group, "log_stream": self.log_stream, }, ) async def run(self) -> AsyncIterator[TriggerEvent]: async with ( await EcsHook( aws_conn_id=self.aws_conn_id, region_name=self.region ).get_async_conn() as ecs_client, await AwsLogsHook( aws_conn_id=self.aws_conn_id, region_name=self.region ).get_async_conn() as logs_client, ): waiter = ecs_client.get_waiter("tasks_stopped") logs_token = None while self.waiter_max_attempts: self.waiter_max_attempts -= 1 try: await waiter.wait( cluster=self.cluster, tasks=[self.task_arn], WaiterConfig={"MaxAttempts": 1} ) # we reach this point only if the waiter met a success criteria yield TriggerEvent( {"status": "success", "task_arn": self.task_arn, "cluster": self.cluster} ) return except WaiterError as error: if "terminal failure" in str(error): raise self.log.info("Status of the task is %s", error.last_response["tasks"][0]["lastStatus"]) await asyncio.sleep(int(self.waiter_delay)) finally: if self.log_group and self.log_stream: logs_token = await self._forward_logs(logs_client, logs_token) raise AirflowException("Waiter error: max attempts reached") async def _forward_logs(self, logs_client, next_token: str | None = None) -> str | None: """ Read logs from the cloudwatch stream and print them to the task logs. :return: the token to pass to the next iteration to resume where we started. """ while True: if next_token is not None: token_arg: dict[str, str] = {"nextToken": next_token} else: token_arg = {} try: response = await logs_client.get_log_events( logGroupName=self.log_group, logStreamName=self.log_stream, startFromHead=True, **token_arg, ) except ClientError as ce: if ce.response["Error"]["Code"] == "ResourceNotFoundException": self.log.info( "Tried to get logs from stream %s in group %s but it didn't exist (yet). " "Will try again.", self.log_stream, self.log_group, ) return None raise events = response["events"] for log_event in events: self.log.info(AwsTaskLogFetcher.event_to_str(log_event)) if len(events) == 0 or next_token == response["nextForwardToken"]: return response["nextForwardToken"] next_token = response["nextForwardToken"]
TaskDoneTrigger
python
tensorflow__tensorflow
tensorflow/python/util/dispatch_test.py
{ "start": 5359, "end": 16191 }
class ____(test_util.TensorFlowTestCase): def testAddDispatchForTypes_With_CppOp(self): original_handlers = gen_math_ops.atan2._tf_fallback_dispatchers[:] # Override the behavior of gen_math_ops.atan2 and make it look like add. @dispatch.dispatch_for_types(gen_math_ops.atan2, CustomTensor) def custom_atan2(y, x, name=None): # pylint: disable=unused-variable return CustomTensor( gen_math_ops.add(y.tensor, x.tensor, name), (x.score + y.score) / 2.0) self.assertEqual( len(math_ops.atan2._tf_fallback_dispatchers), len(original_handlers) + 1) # Test that we see the overridden behavior when using CustomTensors. x = CustomTensor([1., 2., 3.], 2.0) y = CustomTensor([7., 8., 2.], 0.0) x_plus_y = gen_math_ops.atan2(y, x) self.assertAllEqual(self.evaluate(x_plus_y.tensor), [8, 10, 5]) self.assertNear(x_plus_y.score, 1.0, 0.001) # Test that we still get the right behavior when using normal Tensors. a = [1., 2., 3.] b = [7., 8., 2.] a_plus_b = gen_math_ops.atan2(a, b) self.assertAllClose(a_plus_b, [0.14189707, 0.24497867, 0.98279375]) # Test that we still get a TypeError or ValueError if we pass some # type that's not supported by any dispatcher. with self.assertRaises((TypeError, ValueError)): gen_math_ops.atan2(a, None) # Clean up gen_math_ops.atan2._tf_fallback_dispatchers = original_handlers def testAddDispatchForTypes_With_PythonOp(self): original_handlers = test_op._tf_fallback_dispatchers[:] def override_for_test_op(x, y, z): # pylint: disable=unused-variable return CustomTensor( test_op(x.tensor, y.tensor, z.tensor), (x.score + y.score + z.score) / 3.0) override = dispatch.dispatch_for_types(test_op, CustomTensor)( override_for_test_op ) self.assertIs(override, override_for_test_op) x = CustomTensor([1, 2, 3], 0.2) y = CustomTensor([7, 8, 2], 0.4) z = CustomTensor([0, 1, 2], 0.6) result = test_op(x, y, z) self.assertAllEqual(self.evaluate(result.tensor), [15, 21, 13]) self.assertNear(result.score, 0.4, 0.001) # Clean up test_op._tf_fallback_dispatchers = original_handlers def testDispatchForTypes_MissingArgs(self): original_handlers = test_op_with_optional._tf_fallback_dispatchers[:] def override_for_test_op(x, y, z): # pylint: disable=unused-variable return CustomTensor( test_op(x.tensor, y.tensor, z.tensor), (x.score + y.score + z.score) / 3.0, ) override = dispatch.dispatch_for_types(test_op_with_optional, CustomTensor)( override_for_test_op ) self.assertIs(override, override_for_test_op) x = CustomTensor([1, 2, 3], 0.2) y = CustomTensor([7, 8, 2], 0.4) z = CustomTensor([0, 1, 2], 0.6) result = test_op_with_optional(x, y, z) self.assertAllEqual(self.evaluate(result.tensor), [15, 21, 13]) self.assertNear(result.score, 0.4, 0.001) # Clean up test_op_with_optional._tf_fallback_dispatchers = original_handlers def testDispatchForTypes_ProvidingMissingArgs(self): original_handlers = test_op_with_optional._tf_fallback_dispatchers[:] @dispatch.dispatch_for_types(test_op_with_optional, CustomTensor) def override_for_test_op(x, y, z): # pylint: disable=unused-variable return CustomTensor( test_op(x.tensor, y.tensor, z.tensor), (x.score + y.score + z.score) / 3.0, ) x = CustomTensor([1, 2, 3], 0.2) y = CustomTensor([7, 8, 2], 0.4) z = CustomTensor([0, 1, 2], 0.6) with self.assertRaisesRegex( AssertionError, "Dispatched op is called with argument `optional` set to a non-default" " value, which is not supported by the decorated function", ): test_op_with_optional(x, y, z, optional=3) # Clean up test_op_with_optional._tf_fallback_dispatchers = original_handlers def testDispatchForTypes_NewArgs(self): original_handlers = test_op_with_optional._tf_fallback_dispatchers[:] @dispatch.dispatch_for_types(test_op_with_optional, CustomTensor) def override_for_test_op(x, y, z, u=None): # pylint: disable=unused-variable del u return CustomTensor( test_op(x.tensor, y.tensor, z.tensor), (x.score + y.score + z.score) / 3.0, ) x = CustomTensor([1, 2, 3], 0.2) y = CustomTensor([7, 8, 2], 0.4) z = CustomTensor([0, 1, 2], 0.6) result = test_op_with_optional(x, y, z) self.assertAllEqual(self.evaluate(result.tensor), [15, 21, 13]) self.assertNear(result.score, 0.4, 0.001) # Clean up test_op_with_optional._tf_fallback_dispatchers = original_handlers def testDispatchForTypes_SignatureMismatchOrder(self): with self.assertRaisesRegex( AssertionError, "The decorated function's non-default arguments must be identical to" " that of the overridden op.", ): @dispatch.dispatch_for_types(test_op, CustomTensor) def override_for_test_op(x, z, y): # pylint: disable=unused-variable return CustomTensor( test_op(x.tensor, y.tensor, z.tensor), (x.score + y.score + z.score) / 3.0, ) def testDispatchForTypes_MissingKwOnly(self): with self.assertRaisesRegex( AssertionError, "The decorated function's non-default arguments must be identical to" " that of the overridden op.", ): @dispatch.dispatch_for_types(test_op_with_kwonly, CustomTensor) def override_for_test_op(x, z, y): # pylint: disable=unused-variable return CustomTensor( test_op(x.tensor, y.tensor, z.tensor), (x.score + y.score + z.score) / 3.0, ) def testDispatchForTypes_SignatureMismatchNames(self): with self.assertRaisesRegex( AssertionError, "The decorated function's non-default arguments must be identical to" " that of the overridden op.", ): @dispatch.dispatch_for_types(test_op, CustomTensor) def override_for_test_op(a, b, c): # pylint: disable=unused-variable return CustomTensor( test_op(a.tensor, b.tensor, c.tensor), (a.score + b.score + c.score) / 3.0) def testDispatchForTypes_OpDoesNotSupportDispatch(self): def some_op(x, y): return x + y with self.assertRaisesRegex(AssertionError, "Dispatching not enabled for"): @dispatch.dispatch_for_types(some_op, CustomTensor) def override_for_some_op(x, y): # pylint: disable=unused-variable return x if x.score > 0 else y @test.mock.patch.object(tf_logging, "warning", autospec=True) def testInteractionWithDeprecationWarning(self, mock_warning): @deprecation.deprecated(date=None, instructions="Instructions") @dispatch.add_dispatch_support def some_op(x): return x some_op(5) message = mock_warning.call_args[0][0] % mock_warning.call_args[0][1:] self.assertRegex( message, r".*some_op \(from __main__\) is deprecated and will be " "removed in a future version.*") def testGlobalDispatcher(self): original_global_dispatchers = dispatch._GLOBAL_DISPATCHERS try: TensorTracerOpDispatcher().register() x = TensorTracer("x") y = TensorTracer("y") trace = math_ops.reduce_sum(math_ops.add(math_ops.abs(x), y), axis=3) self.assertEqual( str(trace), "math.reduce_sum(math.add(math.abs(x), y), axis=3)") proto_val = TensorTracer("proto") trace = decode_proto(proto_val, "message_type", ["field"], ["float32"]) self.assertIn("io.decode_proto(bytes=proto,", str(trace)) finally: # Clean up. dispatch._GLOBAL_DISPATCHERS = original_global_dispatchers def testGlobalDispatcherConvertToTensor(self): original_global_dispatchers = dispatch._GLOBAL_DISPATCHERS try: TensorTracerOpDispatcher().register() x = TensorTracer("x") y = TensorTracer("y") trace = math_ops.add( math_ops.abs(tensor_conversion.convert_to_tensor_v2_with_dispatch(x)), y, ) self.assertEqual( str(trace), "math.add(math.abs(convert_to_tensor(x)), y)") finally: # Clean up. dispatch._GLOBAL_DISPATCHERS = original_global_dispatchers def testGlobalDispatcherGetItem(self): original_global_dispatchers = dispatch._GLOBAL_DISPATCHERS try: TensorTracerOpDispatcher().register() x = TensorTracer("x") trace = x[0] self.assertEqual(str(trace), "__operators__.getitem(x, 0)") x = TensorTracer("x") y = TensorTracer("y") trace = x[y] self.assertEqual(str(trace), "__operators__.getitem(x, y)") x = TensorTracer("x") y = TensorTracer("y") trace = x[:y] # pylint: disable=invalid-slice-index self.assertEqual( str(trace), "__operators__.getitem(x, slice(None, y, None))") x = array_ops.ones(shape=(3, 3)) y = TensorTracer("y") trace = x[y] self.assertEqual(str(trace), "__operators__.getitem(%s, y)" % x) trace = x[:y] # pylint: disable=invalid-slice-index self.assertEqual( str(trace), "__operators__.getitem(%s, slice(None, y, None))" % x) finally: # Clean up. dispatch._GLOBAL_DISPATCHERS = original_global_dispatchers def testGlobalDispatcherLinearOperators(self): original_global_dispatchers = dispatch._GLOBAL_DISPATCHERS try: TensorTracerOpDispatcher().register() x = TensorTracer("x") # To grab the eigenvalues the diag operator just calls convert_to_tensor # (twice) in this case. trace = linear_operator_diag.LinearOperatorDiag(x).eigvals() self.assertEqual( str(trace), "convert_to_tensor(convert_to_tensor(x, dtype=None, dtype_hint=None, " "name=diag))") # The diagonal tensor addition gets traced even though the linear_operator # API only uses dispatchable ops instead of directly exposing dispatching. trace = linear_operator_diag.LinearOperatorDiag(x).add_to_tensor(x) self.assertIn( "linalg.set_diag(convert_to_tensor(x, name=x), __operators__.add(" "convert_to_tensor(x, dtype=None, dtype_hint=None, name=diag), " "linalg.diag_part(convert_to_tensor(x, name=x)), " "name=", str(trace)) # The dispatch-supporting ops the non-singular check calls out to # get traced. trace = linear_operator_diag.LinearOperatorDiag(x).assert_non_singular() self.assertIn("debugging.assert_less", str(trace)) self.assertIn( "message=Singular operator: Diagonal contained zero values.", str(trace)) finally: # Clean up. dispatch._GLOBAL_DISPATCHERS = original_global_dispatchers
DispatchTest
python
huggingface__transformers
tests/quantization/bnb/test_mixed_int8.py
{ "start": 4044, "end": 17608 }
class ____(BaseMixedInt8Test): def setUp(self): super().setUp() # Models and tokenizer self.model_fp16 = AutoModelForCausalLM.from_pretrained(self.model_name, dtype=torch.float16, device_map="auto") self.model_8bit = AutoModelForCausalLM.from_pretrained( self.model_name, quantization_config=BitsAndBytesConfig(load_in_8bit=True), device_map="auto" ) def tearDown(self): r""" TearDown function needs to be called at the end of each test to free the GPU memory and cache, also to avoid unexpected behaviors. Please see: https://discuss.pytorch.org/t/how-can-we-release-gpu-memory-cache/14530/27 """ del self.model_fp16 del self.model_8bit gc.collect() backend_empty_cache(torch_device) def test_get_keys_to_not_convert(self): r""" Test the `get_keys_to_not_convert` function. """ from accelerate import init_empty_weights from transformers import AutoModelForMaskedLM, Blip2ForConditionalGeneration, MptForCausalLM, OPTForCausalLM from transformers.quantizers.base import get_keys_to_not_convert model_id = "mosaicml/mpt-7b" config = AutoConfig.from_pretrained(model_id, revision="72e5f594ce36f9cabfa2a9fd8f58b491eb467ee7") with init_empty_weights(): model = MptForCausalLM(config) # The order of the keys does not matter, so we sort them before comparing, same for the other tests. self.assertEqual(get_keys_to_not_convert(model).sort(), ["lm_head", "transformer.wte"].sort()) model_id = "Salesforce/blip2-opt-2.7b" config = AutoConfig.from_pretrained(model_id, revision="1ef7f63a8f0a144c13fdca8103eb7b4691c74cec") with init_empty_weights(): model = Blip2ForConditionalGeneration(config) self.assertEqual( get_keys_to_not_convert(model).sort(), ["language_model.lm_head", "language_model.model.decoder.embed_tokens"].sort(), ) model_id = "facebook/opt-350m" config = AutoConfig.from_pretrained(model_id, revision="cb32f77e905cccbca1d970436fb0f5e6b58ee3c5") with init_empty_weights(): model = OPTForCausalLM(config) self.assertEqual(get_keys_to_not_convert(model).sort(), ["lm_head", "model.decoder.embed_tokens"].sort()) model_id = "FacebookAI/roberta-large" config = AutoConfig.from_pretrained(model_id, revision="716877d372b884cad6d419d828bac6c85b3b18d9") with init_empty_weights(): model = AutoModelForMaskedLM.from_config(config) self.assertEqual( get_keys_to_not_convert(model).sort(), ["'roberta.embeddings.word_embeddings', 'lm_head', 'lm_head.decoder"].sort(), ) def test_quantization_config_json_serialization(self): r""" A simple test to check if the quantization config is correctly serialized and deserialized """ config = self.model_8bit.config self.assertTrue(hasattr(config, "quantization_config")) _ = config.to_dict() _ = config.to_diff_dict() _ = config.to_json_string() def test_original_dtype(self): r""" A simple test to check if the model successfully stores the original dtype """ self.assertTrue(hasattr(self.model_8bit.config, "_pre_quantization_dtype")) self.assertFalse(hasattr(self.model_fp16.config, "_pre_quantization_dtype")) self.assertTrue(self.model_8bit.config._pre_quantization_dtype == torch.float16) def test_memory_footprint(self): r""" A simple test to check if the model conversion has been done correctly by checking on the memory footprint of the converted model and the class type of the linear layers of the converted models """ from bitsandbytes.nn import Int8Params mem_fp16 = self.model_fp16.get_memory_footprint() mem_8bit = self.model_8bit.get_memory_footprint() self.assertAlmostEqual(mem_fp16 / mem_8bit, self.EXPECTED_RELATIVE_DIFFERENCE, delta=1e-5) self.assertTrue(get_some_linear_layer(self.model_8bit).weight.__class__ == Int8Params) def test_linear_are_8bit(self): r""" A simple test to check if the model conversion has been done correctly by checking on the memory footprint of the converted model and the class type of the linear layers of the converted models """ from transformers import T5PreTrainedModel self.model_fp16.get_memory_footprint() self.model_8bit.get_memory_footprint() for name, module in self.model_8bit.named_modules(): if isinstance(module, torch.nn.Linear): if name not in ["lm_head"] + T5PreTrainedModel._keep_in_fp32_modules: self.assertTrue(module.weight.dtype == torch.int8) def test_llm_skip(self): r""" A simple test to check if `llm_int8_skip_modules` works as expected """ quantization_config = BitsAndBytesConfig(load_in_8bit=True, llm_int8_skip_modules=["classifier"]) seq_classification_model = AutoModelForSequenceClassification.from_pretrained( "FacebookAI/roberta-large-mnli", quantization_config=quantization_config ) self.assertTrue(seq_classification_model.roberta.encoder.layer[0].output.dense.weight.dtype == torch.int8) self.assertTrue( isinstance(seq_classification_model.roberta.encoder.layer[0].output.dense, bnb.nn.Linear8bitLt) ) self.assertTrue(isinstance(seq_classification_model.classifier.dense, nn.Linear)) self.assertTrue(seq_classification_model.classifier.dense.weight.dtype != torch.int8) self.assertTrue(isinstance(seq_classification_model.classifier.out_proj, nn.Linear)) self.assertTrue(seq_classification_model.classifier.out_proj != torch.int8) def test_generate_quality(self): r""" Test the generation quality of the quantized model and see that we are matching the expected output. Given that we are operating on small numbers + the testing model is relatively small, we might not get the same output across GPUs. So we'll generate few tokens (5-10) and check their output. """ encoded_input = self.tokenizer(self.input_text, return_tensors="pt") output_sequences = self.model_8bit.generate( input_ids=encoded_input["input_ids"].to(self.model_8bit.device), max_new_tokens=10 ) self.assertIn(self.tokenizer.decode(output_sequences[0], skip_special_tokens=True), self.EXPECTED_OUTPUTS) def test_generate_quality_config(self): r""" Test that loading the model with the config is equivalent """ bnb_config = BitsAndBytesConfig() bnb_config.load_in_8bit = True model_8bit_from_config = AutoModelForCausalLM.from_pretrained( self.model_name, quantization_config=bnb_config, device_map="auto" ) encoded_input = self.tokenizer(self.input_text, return_tensors="pt") output_sequences = model_8bit_from_config.generate( input_ids=encoded_input["input_ids"].to(model_8bit_from_config.device), max_new_tokens=10 ) self.assertIn(self.tokenizer.decode(output_sequences[0], skip_special_tokens=True), self.EXPECTED_OUTPUTS) def test_generate_quality_dequantize(self): r""" Test that loading the model and dequantizing it produce correct results """ bnb_config = BitsAndBytesConfig(load_in_8bit=True) model_8bit = AutoModelForCausalLM.from_pretrained( self.model_name, quantization_config=bnb_config, device_map="auto" ) model_8bit.dequantize() encoded_input = self.tokenizer(self.input_text, return_tensors="pt") output_sequences = model_8bit.generate( input_ids=encoded_input["input_ids"].to(model_8bit.device), max_new_tokens=10 ) self.assertIn(self.tokenizer.decode(output_sequences[0], skip_special_tokens=True), self.EXPECTED_OUTPUTS) def test_device_and_dtype_assignment(self): r""" Test whether attempting to change the device or cast the dtype of a model after converting it to 8-bit precision will raise an appropriate error. The test ensures that such operations are prohibited on 8-bit models to prevent invalid conversions. """ with self.assertRaises(ValueError): # Tries with `str` self.model_8bit.to("cpu") with self.assertRaises(ValueError): # Tries with a `dtype`` self.model_8bit.to(torch.float16) with self.assertRaises(ValueError): # Tries with a `device` self.model_8bit.to(torch.device(torch_device)) with self.assertRaises(ValueError): # Tries to cast the 8-bit model to float32 using `float()` self.model_8bit.float() with self.assertRaises(ValueError): # Tries to cast the 4-bit model to float16 using `half()` self.model_8bit.half() # Test if we did not break anything encoded_input = self.tokenizer(self.input_text, return_tensors="pt") self.model_fp16 = self.model_fp16.to(torch.float32) _ = self.model_fp16.generate( input_ids=encoded_input["input_ids"].to(self.model_fp16.device), max_new_tokens=10 ) # Check this does not throw an error _ = self.model_fp16.to("cpu") # Check this does not throw an error _ = self.model_fp16.half() # Check this does not throw an error _ = self.model_fp16.float() def test_fp32_int8_conversion(self): r""" Test whether it is possible to mix both `int8` and `fp32` weights when using `keep_in_fp32_modules` correctly. """ model = AutoModelForSeq2SeqLM.from_pretrained( "google-t5/t5-small", quantization_config=BitsAndBytesConfig(load_in_8bit=True), device_map="auto" ) self.assertTrue(model.decoder.block[0].layer[2].DenseReluDense.wo.weight.dtype == torch.float32) def test_int8_serialization(self): r""" Test whether it is possible to serialize a model in 8-bit. """ from bitsandbytes.nn import Int8Params with tempfile.TemporaryDirectory() as tmpdirname: self.model_8bit.save_pretrained(tmpdirname) # check that the file `quantization_config` is present config = AutoConfig.from_pretrained(tmpdirname) self.assertTrue(hasattr(config, "quantization_config")) model_from_saved = AutoModelForCausalLM.from_pretrained( tmpdirname, quantization_config=BitsAndBytesConfig(load_in_8bit=True), device_map="auto" ) linear = get_some_linear_layer(model_from_saved) self.assertTrue(linear.weight.__class__ == Int8Params) self.assertTrue(hasattr(linear.weight, "SCB")) # generate encoded_input = self.tokenizer(self.input_text, return_tensors="pt") output_sequences = model_from_saved.generate( input_ids=encoded_input["input_ids"].to(model_from_saved.device), max_new_tokens=10 ) self.assertIn(self.tokenizer.decode(output_sequences[0], skip_special_tokens=True), self.EXPECTED_OUTPUTS) def test_int8_serialization_sharded(self): r""" Test whether it is possible to serialize a model in 8-bit - sharded version. """ from bitsandbytes.nn import Int8Params with tempfile.TemporaryDirectory() as tmpdirname: self.model_8bit.save_pretrained(tmpdirname, max_shard_size="200MB") # check that the file `quantization_config` is present config = AutoConfig.from_pretrained(tmpdirname) self.assertTrue(hasattr(config, "quantization_config")) model_from_saved = AutoModelForCausalLM.from_pretrained(tmpdirname) linear = get_some_linear_layer(model_from_saved) self.assertTrue(linear.weight.__class__ == Int8Params) self.assertTrue(hasattr(linear.weight, "SCB")) # generate encoded_input = self.tokenizer(self.input_text, return_tensors="pt") output_sequences = model_from_saved.generate( input_ids=encoded_input["input_ids"].to(torch_device), max_new_tokens=10 ) self.assertIn(self.tokenizer.decode(output_sequences[0], skip_special_tokens=True), self.EXPECTED_OUTPUTS) def test_int8_from_pretrained(self): r""" Test whether loading a 8bit model from the Hub works as expected """ from bitsandbytes.nn import Int8Params model_id = "ybelkada/bloom-1b7-8bit" model = AutoModelForCausalLM.from_pretrained(model_id) linear = get_some_linear_layer(model) self.assertTrue(linear.weight.__class__ == Int8Params) self.assertTrue(hasattr(linear.weight, "SCB")) # generate encoded_input = self.tokenizer(self.input_text, return_tensors="pt") output_sequences = model.generate(input_ids=encoded_input["input_ids"].to(torch_device), max_new_tokens=10) self.assertIn(self.tokenizer.decode(output_sequences[0], skip_special_tokens=True), self.EXPECTED_OUTPUTS) @require_bitsandbytes @require_accelerate @require_torch @slow
MixedInt8Test
python
django__django
tests/admin_scripts/tests.py
{ "start": 27219, "end": 31880 }
class ____(AdminScriptTestCase): """ A series of tests for django-admin when the settings file is in a directory. (see #9751). """ def setUp(self): super().setUp() self.write_settings("settings", is_dir=True) def test_setup_environ(self): "directory: startapp creates the correct directory" args = ["startapp", "settings_test"] app_path = os.path.join(self.test_dir, "settings_test") out, err = self.run_django_admin(args, "test_project.settings") self.assertNoOutput(err) self.assertTrue(os.path.exists(app_path)) with open(os.path.join(app_path, "apps.py")) as f: content = f.read() self.assertIn("class SettingsTestConfig(AppConfig)", content) self.assertInAfterFormatting("name = 'settings_test'", content) def test_setup_environ_custom_template(self): """ directory: startapp creates the correct directory with a custom template """ template_path = os.path.join(custom_templates_dir, "app_template") args = ["startapp", "--template", template_path, "custom_settings_test"] app_path = os.path.join(self.test_dir, "custom_settings_test") out, err = self.run_django_admin(args, "test_project.settings") self.assertNoOutput(err) self.assertTrue(os.path.exists(app_path)) self.assertTrue(os.path.exists(os.path.join(app_path, "api.py"))) def test_startapp_unicode_name(self): """startapp creates the correct directory with Unicode characters.""" args = ["startapp", "こんにちは"] app_path = os.path.join(self.test_dir, "こんにちは") out, err = self.run_django_admin(args, "test_project.settings") self.assertNoOutput(err) self.assertTrue(os.path.exists(app_path)) with open(os.path.join(app_path, "apps.py"), encoding="utf8") as f: content = f.read() self.assertIn("class こんにちはConfig(AppConfig)", content) self.assertInAfterFormatting("name = 'こんにちは'", content) def test_builtin_command(self): """ directory: django-admin builtin commands fail with an error when no settings provided. """ args = ["check", "admin_scripts"] out, err = self.run_django_admin(args) self.assertNoOutput(out) self.assertOutput(err, "settings are not configured") def test_builtin_with_bad_settings(self): """ directory: django-admin builtin commands fail if settings file (from argument) doesn't exist. """ args = ["check", "--settings=bad_settings", "admin_scripts"] out, err = self.run_django_admin(args) self.assertOutput(err, "No module named '?bad_settings'?", regex=True) def test_builtin_with_bad_environment(self): """ directory: django-admin builtin commands fail if settings file (from environment) doesn't exist. """ args = ["check", "admin_scripts"] out, err = self.run_django_admin(args, "bad_settings") self.assertNoOutput(out) self.assertOutput(err, "No module named '?bad_settings'?", regex=True) def test_custom_command(self): """ directory: django-admin can't execute user commands unless settings are provided. """ args = ["noargs_command"] out, err = self.run_django_admin(args) self.assertNoOutput(out) self.assertOutput(err, "No Django settings specified") self.assertOutput(err, "Unknown command: 'noargs_command'") def test_builtin_with_settings(self): """ directory: django-admin builtin commands succeed if settings are provided as argument. """ args = ["check", "--settings=test_project.settings", "admin_scripts"] out, err = self.run_django_admin(args) self.assertNoOutput(err) self.assertOutput(out, SYSTEM_CHECK_MSG) def test_builtin_with_environment(self): """ directory: django-admin builtin commands succeed if settings are provided in the environment. """ args = ["check", "admin_scripts"] out, err = self.run_django_admin(args, "test_project.settings") self.assertNoOutput(err) self.assertOutput(out, SYSTEM_CHECK_MSG) ########################################################################## # MANAGE.PY TESTS # This next series of test classes checks the environment processing # of the generated manage.py script ##########################################################################
DjangoAdminSettingsDirectory
python
joke2k__faker
faker/providers/color/fa_IR/__init__.py
{ "start": 80, "end": 5883 }
class ____(ColorProvider): """Implement color provider for ``fa_IR`` locale. Sources: - https://www.seyedrezabazyar.com/fa/name-and-code-of-colors/ - https://bit.ly/353BBiY """ all_colors = OrderedDict( ( ("نیلی محو", "#F0F8FF"), ("بژ تیره", "#FAEBD7"), ("فیروزه‌ای", "#00FFFF"), ("یشمی", "#7FFFD4"), ("لاجوردی", "#F0FFFF"), ("بژ", "#F5F5DC"), ("کرم", "#FFE4C4"), ("مشکی", "#000000"), ("کاهگلی", "#FFEBCD"), ("آبی", "#0000FF"), ("آبی-بنفش سیر", "#8A2BE2"), ("قهوه‌ای", "#A52A2A"), ("خاکی", "#DEB887"), ("آبی لجنی", "#5F9EA0"), ("سبز روشن", "#7FFF00"), ("شوکولاتی", "#D2691E"), ("مرجانی", "#FF7F50"), ("آبی کدر", "#6495ED"), ("کاهی", "#FFF8DC"), ("زرشکی", "#DC143C"), ("فیروزه‌ای", "#00FFFF"), ("سرمه‌ای", "#00008B"), ("سبز کبریتی تیره", "#008B8B"), ("ماشی سیر", "#B8860B"), ("خاکستری سیر", "#A9A9A9"), ("سبز آووکادو", "#006400"), ("ماشی", "#BDB76B"), ("مخملی", "#8B008B"), ("زیتونی سیر", "#556B2F"), ("نارنجی سیر", "#FF8C00"), ("ارکیده بنفش", "#9932CC"), ("عنابی تند", "#8B0000"), ("قهوه‌ایِ حنایی", "#E9967A"), ("سبز دریایی تیره", "#8FBC8F"), ("آبی دودی", "#483D8B"), ("لجنی تیره", "#2F4F4F"), ("فیروزه‌ای سیر", "#00CED1"), ("بنفش باز", "#9400D3"), ("شفقی", "#FF1493"), ("آبی کمرنگ", "#00BFFF"), ("دودی", "#696969"), ("نیلی", "#1E90FF"), ("شرابی", "#B22222"), ("پوست پیازی", "#FFFAF0"), ("شویدی", "#228B22"), ("سرخابی", "#FF00FF"), ("خاکستری مات", "#DCDCDC"), ("سفید بنفشه", "#F8F8FF"), ("کهربایی باز", "#FFD700"), ("خردلی", "#DAA520"), ("خاکستری", "#808080"), ("سبز", "#008000"), ("مغزپسته‌ای کمرنگ", "#ADFF2F"), ("یشمی محو", "#F0FFF0"), ("سرخابی", "#FF69B4"), ("جگری", "#CD5C5C"), ("نیلی سیر", "#4B0082"), ("استخوانی", "#FFFFF0"), ("خاکی روشن", "#F0E68C"), ("نیلی کمرنگ", "#E6E6FA"), ("صورتی مات", "#FFF0F5"), ("مغزپسته‌ای پررنگ", "#7CFC00"), ("شیرشکری", "#FFFACD"), ("آبی کبریتی", "#ADD8E6"), ("بژ تیره", "#F08080"), ("آبی آسمانی", "#E0FFFF"), ("لیمویی روشن", "#FAFAD2"), ("خاکستری روشن", "#D3D3D3"), ("سبز روشن", "#90EE90"), ("صورتی روشن", "#FFB6C1"), ("کرم نارنجی", "#FFA07A"), ("سبز کبریتی روشن", "#20B2AA"), ("آبی آسمانی روشن", "#87CEFA"), ("سربی", "#778899"), ("بنفش مایل به آبی", "#B0C4DE"), ("شیری", "#FFFFE0"), ("مغزپسته‌ای روشن", "#00FF00"), ("سبز چمنی", "#32CD32"), ("كتانی", "#FAF0E6"), ("سرخ آبی", "#FF00FF"), ("آلبالویی", "#800000"), ("سبز دریایی", "#66CDAA"), ("آبی سیر", "#0000CD"), ("ارکیده سیر", "#BA55D3"), ("سرخ آبی سیر", "#9370DB"), ("خزه‌ای", "#3CB371"), ("آبی متالیک روشن", "#7B68EE"), ("یشمی سیر", "#00FA9A"), ("فیروزه‌ای تیره", "#48D1CC"), ("ارغوانی", "#C71585"), ("آبی نفتی", "#191970"), ("سفید نعنائی", "#F5FFFA"), ("بژ", "#FFE4E1"), ("هلویی", "#FFE4B5"), ("کرم سیر", "#FFDEAD"), ("لاجوردی", "#000080"), ("بژ روشن", "#FDF5E6"), ("زیتونی", "#808000"), ("سبز ارتشی", "#6B8E23"), ("نارنجی", "#FFA500"), ("قرمز-نارنجی", "#FF4500"), ("ارکیده", "#DA70D6"), ("نخودی", "#EEE8AA"), ("سبز کمرنگ", "#98FB98"), ("فیروزه‌ای کدر", "#AFEEEE"), ("شرابی روشن", "#DB7093"), ("هلویی روشن", "#FFEFD5"), ("هلویی پررنگ", "#FFDAB9"), ("بادامی سیر", "#CD853F"), ("صورتی", "#FFC0CB"), ("بنفش کدر", "#DDA0DD"), ("آبی کبریتی روشن", "#B0E0E6"), ("بنفش", "#800080"), ("قرمز", "#FF0000"), ("بادمجانی", "#BC8F8F"), ("فیروزه‌ای فسفری", "#4169E1"), ("کاکائویی", "#8B4513"), ("سالمحناییِ روشنوني", "#FA8072"), ("هلویی سیر", "#F4A460"), ("خزه‌ای پررنگ", "#2E8B57"), ("صدفی", "#FFF5EE"), ("قهوه‌ای متوسط", "#A0522D"), ("طوسی", "#C0C0C0"), ("آبی آسمانی", "#87CEEB"), ("آبی فولادی", "#6A5ACD"), ("سربی تیره", "#708090"), ("صورتی محو", "#FFFAFA"), ("یشمی کمرنگ", "#00FF7F"), ("نیلی متالیک", "#4682B4"), ("برنزه کدر", "#D2B48C"), ("سبز دودی", "#008080"), ("بادمجانی روشن", "#D8BFD8"), ("قرمز گوجه‌ای", "#FF6347"), ("سبز دریایی روشن", "#40E0D0"), ("بنفش روشن", "#EE82EE"), ("گندمی", "#F5DEB3"), ("سفید", "#FFFFFF"), ("خاکستری محو", "#F5F5F5"), ("زرد", "#FFFF00"), ("سبز لجنی", "#9ACD32"), ) ) safe_colors = ( "سیاه", "عنابی", "سبز", "آبی کاربنی", "زیتونی", "بنفش", "سبز دودی", "آهکی", "آبی", "نقره‌ای", "خاکستری", "زرد", "ارغوانی", "فیروزه‌ای", "سفید", )
Provider
python
airbytehq__airbyte
airbyte-integrations/connectors/source-github/source_github/github_schema.py
{ "start": 279334, "end": 279804 }
class ____(sgqlc.types.Input): """Autogenerated input type of RemoveStar""" __schema__ = github_schema __field_names__ = ("starrable_id", "client_mutation_id") starrable_id = sgqlc.types.Field(sgqlc.types.non_null(ID), graphql_name="starrableId") """The Starrable ID to unstar.""" client_mutation_id = sgqlc.types.Field(String, graphql_name="clientMutationId") """A unique identifier for the client performing the mutation."""
RemoveStarInput
python
wandb__wandb
wandb/vendor/pygments/token.py
{ "start": 236, "end": 6167 }
class ____(tuple): parent = None def split(self): buf = [] node = self while node is not None: buf.append(node) node = node.parent buf.reverse() return buf def __init__(self, *args): # no need to call super.__init__ self.subtypes = set() def __contains__(self, val): return self is val or ( type(val) is self.__class__ and val[:len(self)] == self ) def __getattr__(self, val): if not val or not val[0].isupper(): return tuple.__getattribute__(self, val) new = _TokenType(self + (val,)) setattr(self, val, new) self.subtypes.add(new) new.parent = self return new def __repr__(self): return 'Token' + (self and '.' or '') + '.'.join(self) def __copy__(self): # These instances are supposed to be singletons return self def __deepcopy__(self, memo): # These instances are supposed to be singletons return self Token = _TokenType() # Special token types Text = Token.Text Whitespace = Text.Whitespace Escape = Token.Escape Error = Token.Error # Text that doesn't belong to this lexer (e.g. HTML in PHP) Other = Token.Other # Common token types for source code Keyword = Token.Keyword Name = Token.Name Literal = Token.Literal String = Literal.String Number = Literal.Number Punctuation = Token.Punctuation Operator = Token.Operator Comment = Token.Comment # Generic types for non-source code Generic = Token.Generic # String and some others are not direct children of Token. # alias them: Token.Token = Token Token.String = String Token.Number = Number def is_token_subtype(ttype, other): """ Return True if ``ttype`` is a subtype of ``other``. exists for backwards compatibility. use ``ttype in other`` now. """ return ttype in other def string_to_tokentype(s): """ Convert a string into a token type:: >>> string_to_token('String.Double') Token.Literal.String.Double >>> string_to_token('Token.Literal.Number') Token.Literal.Number >>> string_to_token('') Token Tokens that are already tokens are returned unchanged: >>> string_to_token(String) Token.Literal.String """ if isinstance(s, _TokenType): return s if not s: return Token node = Token for item in s.split('.'): node = getattr(node, item) return node # Map standard token types to short names, used in CSS class naming. # If you add a new item, please be sure to run this file to perform # a consistency check for duplicate values. STANDARD_TYPES = { Token: '', Text: '', Whitespace: 'w', Escape: 'esc', Error: 'err', Other: 'x', Keyword: 'k', Keyword.Constant: 'kc', Keyword.Declaration: 'kd', Keyword.Namespace: 'kn', Keyword.Pseudo: 'kp', Keyword.Reserved: 'kr', Keyword.Type: 'kt', Name: 'n', Name.Attribute: 'na', Name.Builtin: 'nb', Name.Builtin.Pseudo: 'bp', Name.Class: 'nc', Name.Constant: 'no', Name.Decorator: 'nd', Name.Entity: 'ni', Name.Exception: 'ne', Name.Function: 'nf', Name.Function.Magic: 'fm', Name.Property: 'py', Name.Label: 'nl', Name.Namespace: 'nn', Name.Other: 'nx', Name.Tag: 'nt', Name.Variable: 'nv', Name.Variable.Class: 'vc', Name.Variable.Global: 'vg', Name.Variable.Instance: 'vi', Name.Variable.Magic: 'vm', Literal: 'l', Literal.Date: 'ld', String: 's', String.Affix: 'sa', String.Backtick: 'sb', String.Char: 'sc', String.Delimiter: 'dl', String.Doc: 'sd', String.Double: 's2', String.Escape: 'se', String.Heredoc: 'sh', String.Interpol: 'si', String.Other: 'sx', String.Regex: 'sr', String.Single: 's1', String.Symbol: 'ss', Number: 'm', Number.Bin: 'mb', Number.Float: 'mf', Number.Hex: 'mh', Number.Integer: 'mi', Number.Integer.Long: 'il', Number.Oct: 'mo', Operator: 'o', Operator.Word: 'ow', Punctuation: 'p', Comment: 'c', Comment.Hashbang: 'ch', Comment.Multiline: 'cm', Comment.Preproc: 'cp', Comment.PreprocFile: 'cpf', Comment.Single: 'c1', Comment.Special: 'cs', Generic: 'g', Generic.Deleted: 'gd', Generic.Emph: 'ge', Generic.Error: 'gr', Generic.Heading: 'gh', Generic.Inserted: 'gi', Generic.Output: 'go', Generic.Prompt: 'gp', Generic.Strong: 'gs', Generic.Subheading: 'gu', Generic.Traceback: 'gt', }
_TokenType
python
pyqtgraph__pyqtgraph
pyqtgraph/dockarea/DockDrop.py
{ "start": 68, "end": 2807 }
class ____: """Provides dock-dropping methods""" def __init__(self, dndWidget): self.dndWidget = dndWidget self.allowedAreas = {'center', 'right', 'left', 'top', 'bottom'} self.dndWidget.setAcceptDrops(True) self.dropArea = None self.overlay = DropAreaOverlay(dndWidget) self.overlay.raise_() def addAllowedArea(self, area): self.allowedAreas.update(area) def removeAllowedArea(self, area): self.allowedAreas.discard(area) def resizeOverlay(self, size): self.overlay.resize(size) def raiseOverlay(self): self.overlay.raise_() def dragEnterEvent(self, ev): src = ev.source() if hasattr(src, 'implements') and src.implements('dock'): #print "drag enter accept" ev.accept() else: #print "drag enter ignore" ev.ignore() def dragMoveEvent(self, ev): #print "drag move" # QDragMoveEvent inherits QDropEvent which provides posF() # PyQt6 provides only position() width, height = self.dndWidget.width(), self.dndWidget.height() posF = ev.posF() if hasattr(ev, 'posF') else ev.position() ld = posF.x() rd = width - ld td = posF.y() bd = height - td mn = min(ld, rd, td, bd) if mn > 30: self.dropArea = "center" elif (ld == mn or td == mn) and mn > height/3: self.dropArea = "center" elif (rd == mn or ld == mn) and mn > width/3: self.dropArea = "center" elif rd == mn: self.dropArea = "right" elif ld == mn: self.dropArea = "left" elif td == mn: self.dropArea = "top" elif bd == mn: self.dropArea = "bottom" if ev.source() is self.dndWidget and self.dropArea == 'center': #print " no self-center" self.dropArea = None ev.ignore() elif self.dropArea not in self.allowedAreas: #print " not allowed" self.dropArea = None ev.ignore() else: #print " ok" ev.accept() self.overlay.setDropArea(self.dropArea) def dragLeaveEvent(self, ev): self.dropArea = None self.overlay.setDropArea(self.dropArea) def dropEvent(self, ev): area = self.dropArea if area is None: return if area == 'center': area = 'above' self.dndWidget.area.moveDock(ev.source(), area, self.dndWidget) self.dropArea = None self.overlay.setDropArea(self.dropArea)
DockDrop
python
getsentry__sentry
src/sentry/incidents/endpoints/organization_alert_rule_details.py
{ "start": 16393, "end": 20825 }
class ____(OrganizationAlertRuleEndpoint): owner = ApiOwner.ISSUES publish_status = { "DELETE": ApiPublishStatus.PUBLIC, "GET": ApiPublishStatus.PUBLIC, "PUT": ApiPublishStatus.PUBLIC, } @extend_schema( operation_id="Retrieve a Metric Alert Rule for an Organization", parameters=[GlobalParams.ORG_ID_OR_SLUG, MetricAlertParams.METRIC_RULE_ID], responses={ 200: AlertRuleSerializer, 401: RESPONSE_UNAUTHORIZED, 403: RESPONSE_FORBIDDEN, 404: RESPONSE_NOT_FOUND, }, examples=MetricAlertExamples.GET_METRIC_ALERT_RULE, ) @track_alert_endpoint_execution("GET", "sentry-api-0-organization-alert-rule-details") @_check_project_access def get(self, request: Request, organization: Organization, alert_rule: AlertRule) -> Response: """ Return details on an individual metric alert rule. A metric alert rule is a configuration that defines the conditions for triggering an alert. It specifies the metric type, function, time interval, and threshold values that determine when an alert should be triggered. Metric alert rules are used to monitor and notify you when certain metrics, like error count, latency, or failure rate, cross a predefined threshold. These rules help you proactively identify and address issues in your project. """ return fetch_alert_rule(request, organization, alert_rule) @extend_schema( operation_id="Update a Metric Alert Rule", parameters=[GlobalParams.ORG_ID_OR_SLUG, MetricAlertParams.METRIC_RULE_ID], request=OrganizationAlertRuleDetailsPutSerializer, responses={ 200: AlertRuleSerializer, 401: RESPONSE_UNAUTHORIZED, 403: RESPONSE_FORBIDDEN, 404: RESPONSE_NOT_FOUND, }, examples=MetricAlertExamples.UPDATE_METRIC_ALERT_RULE, ) @track_alert_endpoint_execution("PUT", "sentry-api-0-organization-alert-rule-details") @_check_project_access def put(self, request: Request, organization: Organization, alert_rule: AlertRule) -> Response: """ Updates a metric alert rule. See **Metric Alert Rule Types** under [Create a Metric Alert Rule for an Organization](/api/alerts/create-a-metric-alert-rule-for-an-organization/#metric-alert-rule-types) to see valid request body configurations for different types of metric alert rule types. > Warning: Calling this endpoint fully overwrites the specified metric alert. A metric alert rule is a configuration that defines the conditions for triggering an alert. It specifies the metric type, function, time interval, and threshold values that determine when an alert should be triggered. Metric alert rules are used to monitor and notify you when certain metrics, like error count, latency, or failure rate, cross a predefined threshold. These rules help you proactively identify and address issues in your project. """ return update_alert_rule(request, organization, alert_rule) @extend_schema( operation_id="Delete a Metric Alert Rule", parameters=[GlobalParams.ORG_ID_OR_SLUG, MetricAlertParams.METRIC_RULE_ID], responses={ 202: RESPONSE_ACCEPTED, 401: RESPONSE_UNAUTHORIZED, 403: RESPONSE_FORBIDDEN, 404: RESPONSE_NOT_FOUND, }, ) @track_alert_endpoint_execution("DELETE", "sentry-api-0-organization-alert-rule-details") @_check_project_access def delete( self, request: Request, organization: Organization, alert_rule: AlertRule ) -> Response: """ Delete a specific metric alert rule. A metric alert rule is a configuration that defines the conditions for triggering an alert. It specifies the metric type, function, time interval, and threshold values that determine when an alert should be triggered. Metric alert rules are used to monitor and notify you when certain metrics, like error count, latency, or failure rate, cross a predefined threshold. These rules help you proactively identify and address issues in your project. """ return remove_alert_rule(request, organization, alert_rule)
OrganizationAlertRuleDetailsEndpoint
python
pandas-dev__pandas
pandas/tests/scalar/timestamp/test_timestamp.py
{ "start": 8029, "end": 14032 }
class ____: @pytest.mark.parametrize("tz", [None, zoneinfo.ZoneInfo("US/Pacific")]) def test_disallow_setting_tz(self, tz): # GH#3746 ts = Timestamp("2010") msg = "Cannot directly set timezone" with pytest.raises(AttributeError, match=msg): ts.tz = tz def test_default_to_stdlib_utc(self): msg = "Timestamp.utcnow is deprecated" with tm.assert_produces_warning(Pandas4Warning, match=msg): assert Timestamp.utcnow().tz is timezone.utc assert Timestamp.now("UTC").tz is timezone.utc assert Timestamp("2016-01-01", tz="UTC").tz is timezone.utc def test_tz(self): tstr = "2014-02-01 09:00" ts = Timestamp(tstr) local = ts.tz_localize("Asia/Tokyo") assert local.hour == 9 assert local == Timestamp(tstr, tz="Asia/Tokyo") conv = local.tz_convert("US/Eastern") assert conv == Timestamp("2014-01-31 19:00", tz="US/Eastern") assert conv.hour == 19 # preserves nanosecond ts = Timestamp(tstr) + offsets.Nano(5) local = ts.tz_localize("Asia/Tokyo") assert local.hour == 9 assert local.nanosecond == 5 conv = local.tz_convert("US/Eastern") assert conv.nanosecond == 5 assert conv.hour == 19 def test_utc_z_designator(self): assert get_timezone(Timestamp("2014-11-02 01:00Z").tzinfo) is timezone.utc def test_asm8(self): ns = [Timestamp.min._value, Timestamp.max._value, 1000] for n in ns: assert ( Timestamp(n).asm8.view("i8") == np.datetime64(n, "ns").view("i8") == n ) assert Timestamp("nat").asm8.view("i8") == np.datetime64("nat", "ns").view("i8") def test_class_ops(self): def compare(x, y): assert int((Timestamp(x)._value - Timestamp(y)._value) / 1e9) == 0 compare(Timestamp.now(), datetime.now()) compare(Timestamp.now("UTC"), datetime.now(timezone.utc)) compare(Timestamp.now("UTC"), datetime.now(tzutc())) msg = "Timestamp.utcnow is deprecated" with tm.assert_produces_warning(Pandas4Warning, match=msg): compare(Timestamp.utcnow(), datetime.now(timezone.utc)) compare(Timestamp.today(), datetime.today()) current_time = calendar.timegm(datetime.now().utctimetuple()) msg = "Timestamp.utcfromtimestamp is deprecated" with tm.assert_produces_warning(Pandas4Warning, match=msg): ts_utc = Timestamp.utcfromtimestamp(current_time) assert ts_utc.timestamp() == current_time compare( Timestamp.fromtimestamp(current_time), datetime.fromtimestamp(current_time) ) compare( # Support tz kwarg in Timestamp.fromtimestamp Timestamp.fromtimestamp(current_time, "UTC"), datetime.fromtimestamp(current_time, timezone.utc), ) compare( # Support tz kwarg in Timestamp.fromtimestamp Timestamp.fromtimestamp(current_time, tz="UTC"), datetime.fromtimestamp(current_time, timezone.utc), ) date_component = datetime.now(timezone.utc) time_component = (date_component + timedelta(minutes=10)).time() compare( Timestamp.combine(date_component, time_component), datetime.combine(date_component, time_component), ) def test_basics_nanos(self): val = np.int64(946_684_800_000_000_000).view("M8[ns]") stamp = Timestamp(val.view("i8") + 500) assert stamp.year == 2000 assert stamp.month == 1 assert stamp.microsecond == 0 assert stamp.nanosecond == 500 # GH 14415 val = np.iinfo(np.int64).min + 80_000_000_000_000 stamp = Timestamp(val) assert stamp.year == 1677 assert stamp.month == 9 assert stamp.day == 21 assert stamp.microsecond == 145224 assert stamp.nanosecond == 192 def test_roundtrip(self): # test value to string and back conversions # further test accessors base = Timestamp("20140101 00:00:00").as_unit("ns") result = Timestamp(base._value + Timedelta("5ms")._value) assert result == Timestamp(f"{base}.005000") assert result.microsecond == 5000 result = Timestamp(base._value + Timedelta("5us")._value) assert result == Timestamp(f"{base}.000005") assert result.microsecond == 5 result = Timestamp(base._value + Timedelta("5ns")._value) assert result == Timestamp(f"{base}.000000005") assert result.nanosecond == 5 assert result.microsecond == 0 result = Timestamp(base._value + Timedelta("6ms 5us")._value) assert result == Timestamp(f"{base}.006005") assert result.microsecond == 5 + 6 * 1000 result = Timestamp(base._value + Timedelta("200ms 5us")._value) assert result == Timestamp(f"{base}.200005") assert result.microsecond == 5 + 200 * 1000 def test_hash_equivalent(self): d = {datetime(2011, 1, 1): 5} stamp = Timestamp(datetime(2011, 1, 1)) assert d[stamp] == 5 @pytest.mark.parametrize( "timezone, year, month, day, hour", [["America/Chicago", 2013, 11, 3, 1], ["America/Santiago", 2021, 4, 3, 23]], ) def test_hash_timestamp_with_fold(self, timezone, year, month, day, hour): # see gh-33931 test_timezone = gettz(timezone) transition_1 = Timestamp( year=year, month=month, day=day, hour=hour, minute=0, fold=0, tzinfo=test_timezone, ) transition_2 = Timestamp( year=year, month=month, day=day, hour=hour, minute=0, fold=1, tzinfo=test_timezone, ) assert hash(transition_1) == hash(transition_2)
TestTimestamp
python
joke2k__faker
tests/providers/test_ssn.py
{ "start": 29354, "end": 30167 }
class ____: @pytest.mark.parametrize( "digits,expected", [ ("22500105", "CHE225001055"), ("60362354", "CHE603623540"), ("36806684", "CHE368066842"), ], ids=[ "checksum_remainder_11", "checksum_remainder_10", "checksum_remainder_other", ], ) def test_checksum(self, digits, expected): """The checksum of the Swiss UID number is calculated correctly given a certain input of 8 digits.""" fake = Faker("fr_CH") Faker.seed(0) with mock.patch( "faker.providers.ssn.fr_CH.Provider.numerify", return_value=digits, autospec=True, ): result = fake.vat_id() assert result == expected
TestFrCH
python
yaml__pyyaml
lib/yaml/tokens.py
{ "start": 1, "end": 511 }
class ____(object): def __init__(self, start_mark, end_mark): self.start_mark = start_mark self.end_mark = end_mark def __repr__(self): attributes = [key for key in self.__dict__ if not key.endswith('_mark')] attributes.sort() arguments = ', '.join(['%s=%r' % (key, getattr(self, key)) for key in attributes]) return '%s(%s)' % (self.__class__.__name__, arguments) #class BOMToken(Token): # id = '<byte order mark>'
Token
python
apache__airflow
providers/amazon/tests/unit/amazon/aws/operators/test_emr_notebook_execution.py
{ "start": 1794, "end": 9319 }
class ____: @mock.patch("botocore.waiter.get_service_module_name", return_value="emr") @mock.patch.object(EmrHook, "conn") @mock.patch.object(Waiter, "wait") def test_start_notebook_execution_wait_for_completion(self, mock_waiter, mock_conn, _): test_execution_id = "test-execution-id" mock_conn.start_notebook_execution.return_value = { "NotebookExecutionId": test_execution_id, "ResponseMetadata": { "HTTPStatusCode": 200, }, } op = EmrStartNotebookExecutionOperator( task_id="test-id", editor_id=PARAMS["EditorId"], relative_path=PARAMS["RelativePath"], cluster_id=PARAMS["ExecutionEngine"]["Id"], service_role=PARAMS["ServiceRole"], notebook_execution_name=PARAMS["NotebookExecutionName"], notebook_params=PARAMS["NotebookParams"], notebook_instance_security_group_id=PARAMS["NotebookInstanceSecurityGroupId"], master_instance_security_group_id=PARAMS["ExecutionEngine"]["MasterInstanceSecurityGroupId"], tags=PARAMS["Tags"], wait_for_completion=True, ) op_response = op.execute(None) mock_conn.start_notebook_execution.assert_called_once_with(**PARAMS) mock_waiter.assert_called_once_with( mock.ANY, NotebookExecutionId=test_execution_id, WaiterConfig=mock.ANY ) assert_expected_waiter_type(mock_waiter, "notebook_running") assert op_response == test_execution_id @mock.patch("airflow.providers.amazon.aws.hooks.emr.EmrHook.conn") def test_start_notebook_execution_no_wait_for_completion(self, mock_conn): test_execution_id = "test-execution-id" mock_conn.start_notebook_execution.return_value = { "NotebookExecutionId": test_execution_id, "ResponseMetadata": { "HTTPStatusCode": 200, }, } op = EmrStartNotebookExecutionOperator( task_id="test-id", editor_id=PARAMS["EditorId"], relative_path=PARAMS["RelativePath"], cluster_id=PARAMS["ExecutionEngine"]["Id"], service_role=PARAMS["ServiceRole"], notebook_execution_name=PARAMS["NotebookExecutionName"], notebook_params=PARAMS["NotebookParams"], notebook_instance_security_group_id=PARAMS["NotebookInstanceSecurityGroupId"], master_instance_security_group_id=PARAMS["ExecutionEngine"]["MasterInstanceSecurityGroupId"], tags=PARAMS["Tags"], ) op_response = op.execute(None) mock_conn.start_notebook_execution.assert_called_once_with(**PARAMS) assert op.wait_for_completion is False assert not mock_conn.describe_notebook_execution.called assert op_response == test_execution_id @mock.patch("airflow.providers.amazon.aws.hooks.emr.EmrHook.conn") def test_start_notebook_execution_http_code_fail(self, mock_conn): test_execution_id = "test-execution-id" mock_conn.start_notebook_execution.return_value = { "NotebookExecutionId": test_execution_id, "ResponseMetadata": { "HTTPStatusCode": 400, }, } op = EmrStartNotebookExecutionOperator( task_id="test-id", editor_id=PARAMS["EditorId"], relative_path=PARAMS["RelativePath"], cluster_id=PARAMS["ExecutionEngine"]["Id"], service_role=PARAMS["ServiceRole"], notebook_execution_name=PARAMS["NotebookExecutionName"], notebook_params=PARAMS["NotebookParams"], notebook_instance_security_group_id=PARAMS["NotebookInstanceSecurityGroupId"], master_instance_security_group_id=PARAMS["ExecutionEngine"]["MasterInstanceSecurityGroupId"], tags=PARAMS["Tags"], ) with pytest.raises(AirflowException, match=r"Starting notebook execution failed:"): op.execute(None) mock_conn.start_notebook_execution.assert_called_once_with(**PARAMS) @mock.patch("botocore.waiter.get_service_module_name", return_value="emr") @mock.patch("time.sleep", return_value=None) @mock.patch.object(EmrHook, "conn") @mock.patch.object(Waiter, "wait") def test_start_notebook_execution_wait_for_completion_multiple_attempts(self, mock_waiter, mock_conn, *_): test_execution_id = "test-execution-id" mock_conn.start_notebook_execution.return_value = { "NotebookExecutionId": test_execution_id, "ResponseMetadata": { "HTTPStatusCode": 200, }, } op = EmrStartNotebookExecutionOperator( task_id="test-id", editor_id=PARAMS["EditorId"], relative_path=PARAMS["RelativePath"], cluster_id=PARAMS["ExecutionEngine"]["Id"], service_role=PARAMS["ServiceRole"], notebook_execution_name=PARAMS["NotebookExecutionName"], notebook_params=PARAMS["NotebookParams"], notebook_instance_security_group_id=PARAMS["NotebookInstanceSecurityGroupId"], master_instance_security_group_id=PARAMS["ExecutionEngine"]["MasterInstanceSecurityGroupId"], tags=PARAMS["Tags"], wait_for_completion=True, ) op_response = op.execute(None) mock_conn.start_notebook_execution.assert_called_once_with(**PARAMS) mock_waiter.assert_called_once_with( mock.ANY, NotebookExecutionId=test_execution_id, WaiterConfig=mock.ANY ) assert_expected_waiter_type(mock_waiter, "notebook_running") assert op_response == test_execution_id @mock.patch("botocore.waiter.get_service_module_name", return_value="emr") @mock.patch.object(EmrHook, "conn") @mock.patch.object(Waiter, "wait") def test_start_notebook_execution_wait_for_completion_fail_state(self, mock_waiter, mock_conn, _): test_execution_id = "test-execution-id" mock_conn.start_notebook_execution.return_value = { "NotebookExecutionId": test_execution_id, "ResponseMetadata": { "HTTPStatusCode": 200, }, } mock_conn.describe_notebook_execution.return_value = {"NotebookExecution": {"Status": "FAILED"}} op = EmrStartNotebookExecutionOperator( task_id="test-id", editor_id=PARAMS["EditorId"], relative_path=PARAMS["RelativePath"], cluster_id=PARAMS["ExecutionEngine"]["Id"], service_role=PARAMS["ServiceRole"], notebook_execution_name=PARAMS["NotebookExecutionName"], notebook_params=PARAMS["NotebookParams"], notebook_instance_security_group_id=PARAMS["NotebookInstanceSecurityGroupId"], master_instance_security_group_id=PARAMS["ExecutionEngine"]["MasterInstanceSecurityGroupId"], tags=PARAMS["Tags"], wait_for_completion=True, ) with pytest.raises(AirflowException, match=r"Notebook Execution reached failure state FAILED\."): op.execute(None) mock_waiter.assert_called_once_with( mock.ANY, NotebookExecutionId=test_execution_id, WaiterConfig=mock.ANY ) assert_expected_waiter_type(mock_waiter, "notebook_running") mock_conn.start_notebook_execution.assert_called_once_with(**PARAMS)
TestEmrStartNotebookExecutionOperator
python
spyder-ide__spyder
spyder/plugins/run/api.py
{ "start": 19174, "end": 20226 }
class ____(TypedDict): """Run execution configuration metadata.""" # File extension or identifier of the input context. input_extension: str # The context for the given input extension, e.g. file, selection, cell or # others. # The context can be compared against the values of `RunContext`. e.g., # `info['context'] == RunContext.File` context: Context # The output formats available for the given context and extension. output_formats: List[OutputFormat] # The configuration widget used to set the options for the current # input extension and context combination. If None or missing then no # configuration widget will be shown on the Run dialog for that # combination. configuration_widget: NotRequired[ Optional[RunExecutorConfigurationGroupCreator]] # True if the executor requires a path in order to work. False otherwise requires_cwd: bool # Priority number used to select the executor. The lower, the better. priority: int
SupportedExecutionRunConfiguration
python
pytorch__pytorch
torch/testing/_internal/autocast_test_lists.py
{ "start": 23926, "end": 28405 }
class ____(TestCase): def args_maybe_kwargs(self, op_with_args): if len(op_with_args) == 2: return op_with_args[0], op_with_args[1], {} else: return op_with_args[0], op_with_args[1], op_with_args[2] def _run_autocast_outofplace( self, op, args, run_as_type, device, out_type=None, module=torch, add_kwargs=None, amp_dtype=torch.bfloat16, ): # helper to cast args def cast(val, to_type): if isinstance(val, torch.Tensor): return val.to(to_type) if val.is_floating_point() else val elif isinstance(val, collections.abc.Iterable): return type(val)(cast(v, to_type) for v in val) else: return val if add_kwargs is None: add_kwargs = {} self.assertFalse(torch.is_autocast_enabled(device_type=device)) with torch.amp.autocast(device_type=device, dtype=amp_dtype): self.assertTrue(torch.is_autocast_enabled(device_type=device)) out_type = out_type if out_type is not None else run_as_type output = output_method = None # Try module.* variant, if requested: if module is not None and hasattr(module, op): output = getattr(module, op)(*args, **add_kwargs) if isinstance(output, torch.Tensor): self.assertTrue( out_type == output.dtype, f"autocast for torch.{op} produced {output.dtype}, should produce {out_type}", ) # Try Tensor.* variant: if hasattr(torch.Tensor, op): output_method = getattr(args[0], op)(*args[1:], **add_kwargs) if isinstance(output_method, torch.Tensor): self.assertTrue( out_type == output_method.dtype, f"autocast for torch.{op} produced {output_method.dtype}, should produce torch.{out_type}", ) self.assertTrue( (output is not None) or (output_method is not None), f"{op} not found as an attribute on either Tensor or the requested module {module}", ) # Accounts for ops that return Tensors, iterables, and other non-Tensors. # For example, lstm_cell returns a tuple and equal returns bool. def compare(first, second): if isinstance(first, torch.Tensor): return torch.equal(first, second) elif isinstance(first, collections.abc.Iterable): return all(compare(f, s) for f, s in zip(first, second, strict=False)) else: return first == second # If both torch.* and Tensor.* variants were found, check outputs are identical if (output is not None) and (output_method is not None): self.assertTrue(type(output) is type(output_method)) comparison = compare(output, output_method) self.assertTrue( comparison, f"torch.{op} result did not match Tensor.{op} result" ) # Compare numerics to Python-side "autocasting" that (we expect) does the same thing # as the C++-side autocasting, and should be bitwise accurate. output_to_compare = output if output is not None else output_method with torch.amp.autocast(device_type=device, enabled=False): self.assertFalse( torch.is_autocast_enabled(device_type=device) ) if module is not None and hasattr(module, op): control = getattr(module, op)( *cast(args, run_as_type), **add_kwargs ) else: control = getattr(args[0].to(run_as_type), op)( *cast(args[1:], run_as_type), **add_kwargs ) self.assertTrue(type(output_to_compare) is type(control)) comparison = compare(output_to_compare, control) self.assertTrue(comparison, f"torch.{op} result did not match control") self.assertTrue(torch.is_autocast_enabled(device_type=device)) self.assertFalse(torch.is_autocast_enabled(device_type=device))
TestAutocast
python
Pylons__pyramid
tests/test_url.py
{ "start": 45597, "end": 46885 }
class ____(unittest.TestCase): def _callFUT(self, path, request, **kw): from pyramid.url import static_path return static_path(path, request, **kw) def _makeRequest(self): class Request: def static_path(self, path, **kw): self.path = path self.kw = kw return 'static path' return Request() def test_it_abs(self): request = self._makeRequest() result = self._callFUT('/foo/bar/abc', request, _anchor='anchor') self.assertEqual(result, 'static path') self.assertEqual(request.path, '/foo/bar/abc') self.assertEqual(request.kw, {'_anchor': 'anchor'}) def test_it_absspec(self): request = self._makeRequest() result = self._callFUT('foo:abc', request, _anchor='anchor') self.assertEqual(result, 'static path') self.assertEqual(request.path, 'foo:abc') self.assertEqual(request.kw, {'_anchor': 'anchor'}) def test_it_rel(self): request = self._makeRequest() result = self._callFUT('abc', request, _app_url='') self.assertEqual(result, 'static path') self.assertEqual(request.path, 'tests:abc') self.assertEqual(request.kw, {'_app_url': ''})
Test_static_path
python
wandb__wandb
wandb/_pydantic/base.py
{ "start": 4793, "end": 5116 }
class ____(GQLBase, ABC): model_config = ConfigDict( alias_generator=to_camel, # Assume JSON names are camelCase, by default frozen=True, # Keep the actual response data immutable ) # Base class for GraphQL input types, i.e. prepared variables or input objects # for queries and mutations.
GQLResult
python
pytorch__pytorch
torch/distributed/tensor/examples/convnext_example.py
{ "start": 2515, "end": 3529 }
class ____(nn.Module): def __init__(self, dim_in=3, dim_out=2, down_scale=4, norm_first=False): super().__init__() self.norm_first = norm_first if norm_first: self.norm = LayerNorm(dim_in, eps=1e-6, data_format=torch.contiguous_format) self.conv = nn.Conv2d( dim_in, dim_out, kernel_size=down_scale, stride=down_scale ) else: self.conv = nn.Conv2d( dim_in, dim_out, kernel_size=down_scale, stride=down_scale ) self.norm = LayerNorm( dim_out, eps=1e-6, data_format=torch.contiguous_format ) def forward(self, x): if self.norm_first: return self.conv(self.norm(x)) else: return self.norm(self.conv(x)) @torch.no_grad() def init_weights(m): if type(m) is nn.Conv2d or type(m) is nn.Linear: nn.init.ones_(m.weight) if m.bias is not None: nn.init.zeros_(m.bias)
DownSampling
python
tensorflow__tensorflow
tensorflow/python/ops/gradients_util.py
{ "start": 36857, "end": 42797 }
class ____: """A class listing aggregation methods used to combine gradients. Computing partial derivatives can require aggregating gradient contributions. This class lists the various methods that can be used to combine gradients in the graph. The following aggregation methods are part of the stable API for aggregating gradients: * `ADD_N`: All of the gradient terms are summed as part of one operation using the "AddN" op (see `tf.add_n`). This method has the property that all gradients must be ready and buffered separately in memory before any aggregation is performed. * `DEFAULT`: The system-chosen default aggregation method. The following aggregation methods are experimental and may not be supported in future releases: * `EXPERIMENTAL_TREE`: Gradient terms are summed in pairs using the "AddN" op. This method of summing gradients may reduce performance, but it can improve memory utilization because the gradients can be released earlier. * `EXPERIMENTAL_ACCUMULATE_N`: Same as `EXPERIMENTAL_TREE`. Example usage when computing gradient: >>> @tf.function ... def example(): ... x = tf.constant(1.0) ... y = x * 2.0 ... z = y + y + y + y ... return tf.gradients(z, [x, y], ... aggregation_method=tf.AggregationMethod.EXPERIMENTAL_ACCUMULATE_N) >>> example() [<tf.Tensor: shape=(), dtype=float32, numpy=8.0>, <tf.Tensor: shape=(), dtype=float32, numpy=4.0>] """ ADD_N = 0 DEFAULT = ADD_N # The following are experimental and may not be supported in future releases. EXPERIMENTAL_TREE = 1 EXPERIMENTAL_ACCUMULATE_N = 2 # An alias for EXPERIMENTAL_ADD_N = 1 def _AggregatedGrads(grads, op, gradient_uid, loop_state, aggregation_method=None): """Get the aggregated gradients for op. Args: grads: The map of memoized gradients. op: The op to get gradients for. gradient_uid: A unique identifier within the graph indicating which invocation of gradients is being executed. Used to cluster ops for compilation. loop_state: An object for maintaining the state of the while loops in the graph. It is of type ControlFlowState. None if the graph contains no while loops. aggregation_method: Specifies the method used to combine gradient terms. Accepted values are constants defined in the class `AggregationMethod`. Returns: A list of gradients, one per each output of `op`. If the gradients for a particular output is a list, this function aggregates it before returning. Raises: TypeError: if the incoming grads are not Tensors or IndexedSlices. ValueError: if the arguments are invalid. """ if aggregation_method is None: aggregation_method = AggregationMethod.DEFAULT valid_aggregation_methods = [ AggregationMethod.ADD_N, AggregationMethod.EXPERIMENTAL_TREE, AggregationMethod.EXPERIMENTAL_ACCUMULATE_N] if aggregation_method not in valid_aggregation_methods: raise ValueError( f"Invalid `aggregation_method` specified {aggregation_method}. " f"Accepted values are {valid_aggregation_methods}.") out_grads = _GetGrads(grads, op) for i, out_grad in enumerate(out_grads): if loop_state: if isinstance( out_grad, (tensor_lib.Tensor, indexed_slices.IndexedSlices)): assert control_flow_util.IsLoopSwitch(op) continue # Grads have to be Tensors or IndexedSlices if (isinstance(out_grad, collections_abc.Sequence) and not all( isinstance(g, (tensor_lib.Tensor, indexed_slices.IndexedSlices)) for g in out_grad if g is not None)): raise TypeError(f"Invalid gradient {out_grad} [index = {i}]. Gradients " "have to be either all Tensors or all IndexedSlices") # Aggregate multiple gradients, and convert [] to None. if out_grad: if len(out_grad) < 2: used = "nop" out_grads[i] = out_grad[0] elif all( isinstance(g, tensor_lib.Tensor) for g in out_grad if g is not None): tensor_shape = _AccumulatorShape(out_grad) if aggregation_method in [ AggregationMethod.EXPERIMENTAL_TREE, AggregationMethod.EXPERIMENTAL_ACCUMULATE_N ]: # Aggregate all gradients by doing pairwise sums: this may # reduce performance, but it can improve memory because the # gradients can be released earlier. # # TODO(vrv): Consider replacing this with a version of # tf.AddN() that eagerly frees its inputs as soon as they are # ready, so the order of this tree does not become a problem. used = "tree" with ops.name_scope(op.name + "_gradient_sum"): running_sum = out_grad[0] for grad in out_grad[1:]: running_sum = math_ops.add_n([running_sum, grad]) out_grads[i] = running_sum else: used = "add_n" out_grads[i] = _MultiDeviceAddN(out_grad, gradient_uid) logging.vlog(2, " _AggregatedGrads %d x %s using %s", len(out_grad), tensor_shape, used) else: out_grads[i] = backprop_util.AggregateIndexedSlicesGradients(out_grad) # pylint: disable=protected-access else: # not out_grad # out_grads[i] is [], thus its aggregation is simply None. out_grads[i] = None return out_grads # Represents the output of TFE_Py_TapeSetPossibleGradientTypes. Real enums are # unfortunately too slow to use here. POSSIBLE_GRADIENT_TYPES_NONE = 0 POSSIBLE_GRADIENT_TYPES_FIRST_ORDER = 1 POSSIBLE_GRADIENT_TYPES_HIGHER_ORDER = 2 def PossibleTapeGradientTypes(tensors): """Determines whether and how `args` may require tape gradients.""" return pywrap_tfe.TFE_Py_TapeSetPossibleGradientTypes(tensors)
AggregationMethod
python
astropy__astropy
astropy/stats/spatial.py
{ "start": 408, "end": 13929 }
class ____: """ Estimators for Ripley's K function for two-dimensional spatial data. See [1]_, [2]_, [3]_, [4]_, [5]_ for detailed mathematical and practical aspects of those estimators. Parameters ---------- area : float Area of study from which the points where observed. x_max, y_max : float, float, optional Maximum rectangular coordinates of the area of study. Required if ``mode == 'translation'`` or ``mode == ohser``. x_min, y_min : float, float, optional Minimum rectangular coordinates of the area of study. Required if ``mode == 'variable-width'`` or ``mode == ohser``. Examples -------- >>> import numpy as np >>> from matplotlib import pyplot as plt # doctest: +SKIP >>> from astropy.stats import RipleysKEstimator >>> z = np.random.uniform(low=5, high=10, size=(100, 2)) >>> Kest = RipleysKEstimator(area=25, x_max=10, y_max=10, ... x_min=5, y_min=5) >>> r = np.linspace(0, 2.5, 100) >>> plt.plot(r, Kest.poisson(r)) # doctest: +SKIP >>> plt.plot(r, Kest(data=z, radii=r, mode='none')) # doctest: +SKIP >>> plt.plot(r, Kest(data=z, radii=r, mode='translation')) # doctest: +SKIP >>> plt.plot(r, Kest(data=z, radii=r, mode='ohser')) # doctest: +SKIP >>> plt.plot(r, Kest(data=z, radii=r, mode='var-width')) # doctest: +SKIP >>> plt.plot(r, Kest(data=z, radii=r, mode='ripley')) # doctest: +SKIP References ---------- .. [1] Peebles, P.J.E. *The large scale structure of the universe*. <https://ui.adsabs.harvard.edu/abs/1980lssu.book.....P> .. [2] Spatial descriptive statistics. <https://en.wikipedia.org/wiki/Spatial_descriptive_statistics> .. [3] Package spatstat. <https://cran.r-project.org/web/packages/spatstat/spatstat.pdf> .. [4] Cressie, N.A.C. (1991). Statistics for Spatial Data, Wiley, New York. .. [5] Stoyan, D., Stoyan, H. (1992). Fractals, Random Shapes and Point Fields, Akademie Verlag GmbH, Chichester. """ def __init__( self, area: float, x_max: float | None = None, y_max: float | None = None, x_min: float | None = None, y_min: float | None = None, ) -> None: self.area = area self.x_max = x_max self.y_max = y_max self.x_min = x_min self.y_min = y_min @property def area(self) -> float: return self._area @area.setter def area(self, value: float) -> None: if isinstance(value, (float, int)) and value > 0: self._area = value else: raise ValueError(f"area is expected to be a positive number. Got {value}.") @property def y_max(self) -> float | None: return self._y_max @y_max.setter def y_max(self, value: float | None) -> None: if value is None or isinstance(value, (float, int)): self._y_max = value else: raise ValueError( f"y_max is expected to be a real number or None. Got {value}." ) @property def x_max(self) -> float | None: return self._x_max @x_max.setter def x_max(self, value: float | None) -> None: if value is None or isinstance(value, (float, int)): self._x_max = value else: raise ValueError( f"x_max is expected to be a real number or None. Got {value}." ) @property def y_min(self) -> float | None: return self._y_min @y_min.setter def y_min(self, value: float | None) -> None: if value is None or isinstance(value, (float, int)): self._y_min = value else: raise ValueError(f"y_min is expected to be a real number. Got {value}.") @property def x_min(self) -> float | None: return self._x_min @x_min.setter def x_min(self, value: float | None) -> None: if value is None or isinstance(value, (float, int)): self._x_min = value else: raise ValueError(f"x_min is expected to be a real number. Got {value}.") def __call__( self, data: NDArray[float], radii: NDArray[float], mode: _ModeOps = "none", ) -> NDArray[float]: return self.evaluate(data=data, radii=radii, mode=mode) def _pairwise_diffs(self, data: NDArray[float]) -> NDArray[float]: npts = len(data) diff = np.zeros(shape=(npts * (npts - 1) // 2, 2), dtype=np.double) k = 0 for i in range(npts - 1): size = npts - i - 1 diff[k : k + size] = abs(data[i] - data[i + 1 :]) k += size return diff def poisson(self, radii: NDArray[float]) -> NDArray[float]: """ Evaluates the Ripley K function for the homogeneous Poisson process, also known as Complete State of Randomness (CSR). Parameters ---------- radii : 1D array Set of distances in which Ripley's K function will be evaluated. Returns ------- output : 1D array Ripley's K function evaluated at ``radii``. """ return np.pi * radii * radii def Lfunction( self, data: NDArray[float], radii: NDArray[float], mode: _ModeOps = "none", ) -> NDArray[float]: """ Evaluates the L function at ``radii``. For parameter description see ``evaluate`` method. """ return np.sqrt(self.evaluate(data, radii, mode=mode) / np.pi) def Hfunction( self, data: NDArray[float], radii: NDArray[float], mode: _ModeOps = "none", ) -> NDArray[float]: """ Evaluates the H function at ``radii``. For parameter description see ``evaluate`` method. """ return self.Lfunction(data, radii, mode=mode) - radii def evaluate( self, data: NDArray[float], radii: NDArray[float], mode: _ModeOps = "none", ) -> NDArray[float]: """ Evaluates the Ripley K estimator for a given set of values ``radii``. Parameters ---------- data : 2D array Set of observed points in as a n by 2 array which will be used to estimate Ripley's K function. radii : 1D array Set of distances in which Ripley's K estimator will be evaluated. Usually, it's common to consider max(radii) < (area/2)**0.5. mode : str Keyword which indicates the method for edge effects correction. Available methods are 'none', 'translation', 'ohser', 'var-width', and 'ripley'. * 'none' this method does not take into account any edge effects whatsoever. * 'translation' computes the intersection of rectangular areas centered at the given points provided the upper bounds of the dimensions of the rectangular area of study. It assumes that all the points lie in a bounded rectangular region satisfying x_min < x_i < x_max; y_min < y_i < y_max. A detailed description of this method can be found on ref [4]. * 'ohser' this method uses the isotropized set covariance function of the window of study as a weight to correct for edge-effects. A detailed description of this method can be found on ref [4]. * 'var-width' this method considers the distance of each observed point to the nearest boundary of the study window as a factor to account for edge-effects. See [3] for a brief description of this method. * 'ripley' this method is known as Ripley's edge-corrected estimator. The weight for edge-correction is a function of the proportions of circumferences centered at each data point which crosses another data point of interest. See [3] for a detailed description of this method. Returns ------- ripley : 1D array Ripley's K function estimator evaluated at ``radii``. """ data = np.asarray(data) if not data.shape[1] == 2: raise ValueError( "data must be an n by 2 array, where n is the " "number of observed points." ) npts = len(data) ripley = np.zeros(len(radii)) if mode == "none": diff = self._pairwise_diffs(data) distances = np.hypot(diff[:, 0], diff[:, 1]) for r in range(len(radii)): ripley[r] = (distances < radii[r]).sum() ripley = self.area * 2.0 * ripley / (npts * (npts - 1)) # eq. 15.11 Stoyan book page 283 elif mode == "translation": diff = self._pairwise_diffs(data) distances = np.hypot(diff[:, 0], diff[:, 1]) intersec_area = ((self.x_max - self.x_min) - diff[:, 0]) * ( (self.y_max - self.y_min) - diff[:, 1] ) for r in range(len(radii)): dist_indicator = distances < radii[r] ripley[r] = ((1 / intersec_area) * dist_indicator).sum() ripley = (self.area**2 / (npts * (npts - 1))) * 2 * ripley # Stoyan book page 123 and eq 15.13 elif mode == "ohser": diff = self._pairwise_diffs(data) distances = np.hypot(diff[:, 0], diff[:, 1]) a = self.area b = max( (self.y_max - self.y_min) / (self.x_max - self.x_min), (self.x_max - self.x_min) / (self.y_max - self.y_min), ) x = distances / math.sqrt(a / b) u = np.sqrt((x * x - 1) * (x > 1)) v = np.sqrt((x * x - b**2) * (x < math.sqrt(b**2 + 1)) * (x > b)) c1 = np.pi - 2 * x * (1 + 1 / b) + x * x / b c2 = 2 * np.arcsin((1 / x) * (x > 1)) - 1 / b - 2 * (x - u) c3 = ( 2 * np.arcsin( ((b - u * v) / (x * x)) * (x > b) * (x < math.sqrt(b**2 + 1)) ) + 2 * u + 2 * v / b - b - (1 + x * x) / b ) cov_func = (a / np.pi) * ( c1 * (x >= 0) * (x <= 1) + c2 * (x > 1) * (x <= b) + c3 * (b < x) * (x < math.sqrt(b**2 + 1)) ) for r in range(len(radii)): dist_indicator = distances < radii[r] ripley[r] = ((1 / cov_func) * dist_indicator).sum() ripley = (self.area**2 / (npts * (npts - 1))) * 2 * ripley # Cressie book eq 8.2.20 page 616 elif mode == "var-width": lt_dist = np.minimum( np.minimum(self.x_max - data[:, 0], self.y_max - data[:, 1]), np.minimum(data[:, 0] - self.x_min, data[:, 1] - self.y_min), ) for r in range(len(radii)): for i in range(npts): for j in range(npts): if i != j: diff = abs(data[i] - data[j]) dist = math.sqrt((diff * diff).sum()) if dist < radii[r] < lt_dist[i]: ripley[r] = ripley[r] + 1 lt_dist_sum = (lt_dist > radii[r]).sum() if not lt_dist_sum == 0: ripley[r] = ripley[r] / lt_dist_sum ripley = self.area * ripley / npts # Cressie book eq 8.4.22 page 640 elif mode == "ripley": hor_dist = np.zeros(shape=(npts * (npts - 1)) // 2, dtype=np.double) ver_dist = np.zeros(shape=(npts * (npts - 1)) // 2, dtype=np.double) for k in range(npts - 1): min_hor_dist = min(self.x_max - data[k][0], data[k][0] - self.x_min) min_ver_dist = min(self.y_max - data[k][1], data[k][1] - self.y_min) start = (k * (2 * (npts - 1) - (k - 1))) // 2 end = ((k + 1) * (2 * (npts - 1) - k)) // 2 hor_dist[start:end] = min_hor_dist * np.ones(npts - 1 - k) ver_dist[start:end] = min_ver_dist * np.ones(npts - 1 - k) diff = self._pairwise_diffs(data) dist = np.hypot(diff[:, 0], diff[:, 1]) dist_ind = dist <= np.hypot(hor_dist, ver_dist) w1 = ( 1 - ( np.arccos(np.minimum(ver_dist, dist) / dist) + np.arccos(np.minimum(hor_dist, dist) / dist) ) / np.pi ) w2 = ( 3 / 4 - 0.5 * ( np.arccos(ver_dist / dist * ~dist_ind) + np.arccos(hor_dist / dist * ~dist_ind) ) / np.pi ) weight = dist_ind * w1 + ~dist_ind * w2 for r in range(len(radii)): ripley[r] = ((dist < radii[r]) / weight).sum() ripley = self.area * 2.0 * ripley / (npts * (npts - 1)) else: raise ValueError(f"mode {mode} is not implemented.") return ripley
RipleysKEstimator
python
huggingface__transformers
src/transformers/models/got_ocr2/modular_got_ocr2.py
{ "start": 10004, "end": 10110 }
class ____(SamVisionEncoder, GotOcr2PreTrainedModel): input_modalities = ("image",)
GotOcr2VisionEncoder
python
pytorch__pytorch
torch/nn/modules/rnn.py
{ "start": 71147, "end": 74952 }
class ____(RNNCellBase): r"""A gated recurrent unit (GRU) cell. .. math:: \begin{array}{ll} r = \sigma(W_{ir} x + b_{ir} + W_{hr} h + b_{hr}) \\ z = \sigma(W_{iz} x + b_{iz} + W_{hz} h + b_{hz}) \\ n = \tanh(W_{in} x + b_{in} + r \odot (W_{hn} h + b_{hn})) \\ h' = (1 - z) \odot n + z \odot h \end{array} where :math:`\sigma` is the sigmoid function, and :math:`\odot` is the Hadamard product. Args: input_size: The number of expected features in the input `x` hidden_size: The number of features in the hidden state `h` bias: If ``False``, then the layer does not use bias weights `b_ih` and `b_hh`. Default: ``True`` Inputs: input, hidden - **input** : tensor containing input features - **hidden** : tensor containing the initial hidden state for each element in the batch. Defaults to zero if not provided. Outputs: h' - **h'** : tensor containing the next hidden state for each element in the batch Shape: - input: :math:`(N, H_{in})` or :math:`(H_{in})` tensor containing input features where :math:`H_{in}` = `input_size`. - hidden: :math:`(N, H_{out})` or :math:`(H_{out})` tensor containing the initial hidden state where :math:`H_{out}` = `hidden_size`. Defaults to zero if not provided. - output: :math:`(N, H_{out})` or :math:`(H_{out})` tensor containing the next hidden state. Attributes: weight_ih: the learnable input-hidden weights, of shape `(3*hidden_size, input_size)` weight_hh: the learnable hidden-hidden weights, of shape `(3*hidden_size, hidden_size)` bias_ih: the learnable input-hidden bias, of shape `(3*hidden_size)` bias_hh: the learnable hidden-hidden bias, of shape `(3*hidden_size)` .. note:: All the weights and biases are initialized from :math:`\mathcal{U}(-\sqrt{k}, \sqrt{k})` where :math:`k = \frac{1}{\text{hidden\_size}}` On certain ROCm devices, when using float16 inputs this module will use :ref:`different precision<fp16_on_mi200>` for backward. Examples:: >>> rnn = nn.GRUCell(10, 20) >>> input = torch.randn(6, 3, 10) >>> hx = torch.randn(3, 20) >>> output = [] >>> for i in range(6): ... hx = rnn(input[i], hx) ... output.append(hx) """ def __init__( self, input_size: int, hidden_size: int, bias: bool = True, device=None, dtype=None, ) -> None: factory_kwargs = {"device": device, "dtype": dtype} super().__init__(input_size, hidden_size, bias, num_chunks=3, **factory_kwargs) def forward(self, input: Tensor, hx: Optional[Tensor] = None) -> Tensor: if input.dim() not in (1, 2): raise ValueError( f"GRUCell: Expected input to be 1D or 2D, got {input.dim()}D instead" ) if hx is not None and hx.dim() not in (1, 2): raise ValueError( f"GRUCell: Expected hidden to be 1D or 2D, got {hx.dim()}D instead" ) is_batched = input.dim() == 2 if not is_batched: input = input.unsqueeze(0) if hx is None: hx = torch.zeros( input.size(0), self.hidden_size, dtype=input.dtype, device=input.device ) else: hx = hx.unsqueeze(0) if not is_batched else hx ret = _VF.gru_cell( input, hx, self.weight_ih, self.weight_hh, self.bias_ih, self.bias_hh, ) if not is_batched: ret = ret.squeeze(0) return ret
GRUCell
python
sqlalchemy__sqlalchemy
lib/sqlalchemy/sql/ddl.py
{ "start": 34859, "end": 35397 }
class ____(_DropBase["Index"]): """Represent a DROP INDEX statement.""" __visit_name__ = "drop_index" def __init__(self, element: Index, if_exists: bool = False) -> None: """Create a :class:`.DropIndex` construct. :param element: a :class:`_schema.Index` that's the subject of the DROP. :param if_exists: if True, an IF EXISTS operator will be applied to the construct. .. versionadded:: 1.4.0b2 """ super().__init__(element, if_exists=if_exists)
DropIndex
python
PrefectHQ__prefect
src/integrations/prefect-databricks/prefect_databricks/models/jobs.py
{ "start": 31444, "end": 33828 }
class ____(BaseModel): """ See source code for the fields' description. """ model_config = ConfigDict(extra="allow", frozen=True) base_parameters: Optional[Dict[str, Any]] = Field( default=None, description=( "Base parameters to be used for each run of this job. If the run is" " initiated by a call to" " [`run-now`](https://docs.databricks.com/dev-tools/api/latest/jobs.html#operation/JobsRunNow)" " with parameters specified, the two parameters maps are merged. If the" " same key is specified in `base_parameters` and in `run-now`, the value" " from `run-now` is used.\n\nUse [Task parameter" " variables](https://docs.databricks.com/jobs.html#parameter-variables) to" " set parameters containing information about job runs.\n\nIf the notebook" " takes a parameter that is not specified in the job’s `base_parameters` or" " the `run-now` override parameters, the default value from the notebook is" " used.\n\nRetrieve these parameters in a notebook using" " [dbutils.widgets.get](https://docs.databricks.com/dev-tools/databricks-utils.html#dbutils-widgets)." ), examples=[{"age": 35, "name": "John Doe"}], ) notebook_path: str = Field( default=..., description=( "The path of the notebook to be run in the Databricks workspace or remote" " repository. For notebooks stored in the Databricks workspace, the path" " must be absolute and begin with a slash. For notebooks stored in a remote" " repository, the path must be relative. This field is required." ), examples=["/Users/user.name@databricks.com/notebook_to_run"], ) source: Optional[Literal["WORKSPACE", "GIT"]] = Field( default=None, description=( "Optional location type of the notebook. When set to `WORKSPACE`, the" " notebook will be retrieved from the local Databricks workspace. When set" " to `GIT`, the notebook will be retrieved from a Git repository defined in" " `git_source`. If the value is empty, the task will use `GIT` if" " `git_source` is defined and `WORKSPACE` otherwise." ), examples=["WORKSPACE"], )
NotebookTask
python
cython__cython
Cython/Compiler/Nodes.py
{ "start": 331535, "end": 335223 }
class ____(LoopNode, StatNode): # Base class of 'for-in' statements. # # target ExprNode # iterator IteratorNode | AIterAwaitExprNode(AsyncIteratorNode) # body StatNode # else_clause StatNode # item NextNode | AwaitExprNode(AsyncNextNode) # is_async boolean true for 'async for' statements child_attrs = ["target", "item", "iterator", "body", "else_clause"] item = None is_async = False def _create_item_node(self): raise NotImplementedError("must be implemented by subclasses") def analyse_declarations(self, env): self.target.analyse_target_declaration(env) self.body.analyse_declarations(env) if self.else_clause: self.else_clause.analyse_declarations(env) self._create_item_node() def analyse_expressions(self, env): self.target = self.target.analyse_target_types(env) self.iterator = self.iterator.analyse_expressions(env) self._create_item_node() # must rewrap self.item after analysis self.item = self.item.analyse_expressions(env) if (not self.is_async and (self.iterator.type.is_ptr or self.iterator.type.is_array) and self.target.type.assignable_from(self.iterator.type)): # C array slice optimization. pass else: self.item = self.item.coerce_to(self.target.type, env) self.body = self.body.analyse_expressions(env) if self.else_clause: self.else_clause = self.else_clause.analyse_expressions(env) return self def generate_execution_code(self, code): code.mark_pos(self.pos) old_loop_labels = code.new_loop_labels() self.iterator.generate_evaluation_code(code) code.put("for (") self.iterator.generate_for_loop_header(code) code.putln(") {") self.item.generate_evaluation_code(code) self.target.generate_assignment_code(self.item, code) code.write_trace_line(self.pos) self.body.generate_execution_code(code) code.mark_pos(self.pos) code.put_label(code.continue_label) code.putln("}") # clean up before we enter the 'else:' branch self.iterator.generate_disposal_code(code) else_label = code.new_label("for_else") if self.else_clause else None end_label = code.new_label("for_end") label_intercepts = code.label_interceptor( [code.break_label], [end_label], skip_to_label=else_label or end_label, pos=self.pos, ) code.mark_pos(self.pos) for _ in label_intercepts: self.iterator.generate_disposal_code(code) code.set_loop_labels(old_loop_labels) self.iterator.free_temps(code) if self.else_clause: code.putln("/*else*/ {") code.put_label(else_label) self.else_clause.generate_execution_code(code) code.putln("}") code.put_label(end_label) def generate_function_definitions(self, env, code): self.target.generate_function_definitions(env, code) self.iterator.generate_function_definitions(env, code) self.body.generate_function_definitions(env, code) if self.else_clause is not None: self.else_clause.generate_function_definitions(env, code) def annotate(self, code): self.target.annotate(code) self.iterator.annotate(code) self.body.annotate(code) if self.else_clause: self.else_clause.annotate(code) self.item.annotate(code)
_ForInStatNode
python
ray-project__ray
rllib/utils/framework.py
{ "start": 8755, "end": 9044 }
class ____: def __init__(self, *a, **kw): # Fake nn.functional module within torch.nn. self.functional = None self.Module = _FakeTorchClassStub self.parallel = _ParallelStub() # Fake class for e.g. torch.nn.Module to allow it to be inherited from.
_NNStub
python
realpython__materials
python-async-iterators/counter.py
{ "start": 44, "end": 863 }
class ____: def __init__(self, name="", end=5): self.counter = 0 self.name = name self.end = end def __aiter__(self): return self async def __anext__(self): if self.counter >= self.end: raise StopAsyncIteration self.counter += 1 await asyncio.sleep(randint(1, 3) / 10) return self.counter async def task(iterator): async for item in iterator: print(item, f"from iterator {iterator.name}") async def main(): # This code runs sequentially: # await task(AsyncCounterIterator("#1")) # await task(AsyncCounterIterator("#2")) # This is concurrent: await asyncio.gather( task(AsyncCounterIterator("#1")), task(AsyncCounterIterator("#2")), ) asyncio.run(main())
AsyncCounterIterator
python
coleifer__peewee
tests/libs/mock.py
{ "start": 74959, "end": 75527 }
class ____(Mock): """ A mock intended to be used as a property, or other descriptor, on a class. `PropertyMock` provides `__get__` and `__set__` methods so you can specify a return value when it is fetched. Fetching a `PropertyMock` instance from an object calls the mock, with no args. Setting it calls the mock with the value being set. """ def _get_child_mock(self, **kwargs): return MagicMock(**kwargs) def __get__(self, obj, obj_type): return self() def __set__(self, obj, val): self(val)
PropertyMock
python
great-expectations__great_expectations
contrib/great_expectations_semantic_types_expectations/great_expectations_semantic_types_expectations/expectations/expect_column_values_ip_asn_country_code_in_set.py
{ "start": 2029, "end": 4905 }
class ____(ColumnMapExpectation): """Expect the provided IP address ASN country code in set.""" # These examples will be shown in the public gallery. # They will also be executed as unit tests for your Expectation. examples = [ { "data": { "all_hu": [ "213.181.199.16", "213.181.199.16", "213.181.199.16", "213.181.199.16", "213.181.199.16", ], "some_other": [ "213.181.199.16", "213.181.199.16", "213.181.199.16", "213.181.199.16", "142.250.180.206", ], }, "tests": [ { "title": "basic_positive_test", "exact_match_out": False, "include_in_gallery": True, "in": {"column": "all_hu", "country_codes": ["hu"]}, "out": { "success": True, }, }, { "title": "basic_negative_test", "exact_match_out": False, "include_in_gallery": True, "in": { "column": "some_other", "country_codes": ["hu"], "mostly": 0.9, }, "out": { "success": False, }, }, ], } ] # This is the id string of the Metric used by this Expectation. # For most Expectations, it will be the same as the `condition_metric_name` defined in your Metric class above. map_metric = "column_values.ip_asn_country_code_in_set" # This is a list of parameter names that can affect whether the Expectation evaluates to True or False success_keys = ("mostly",) # This dictionary contains default values for any parameters that should have default values default_kwarg_values = {} # This object contains metadata for display in the public Gallery library_metadata = { "maturity": "experimental", "tags": [ "hackathon-22", "experimental", "typed-entities", ], # Tags for this Expectation in the Gallery "contributors": [ # Github handles for all contributors to this Expectation. "@szecsip", # Don't forget to add your github handle here! ], "requirements": ["dnspython", "ipwhois"], } success_keys = ( "country_codes", "mostly", ) if __name__ == "__main__": ExpectColumnValuesIpAsnCountryCodeInSet().print_diagnostic_checklist()
ExpectColumnValuesIpAsnCountryCodeInSet