patch
stringlengths
18
160k
callgraph
stringlengths
4
179k
summary
stringlengths
4
947
msg
stringlengths
6
3.42k
@@ -0,0 +1,17 @@ +package watch + +import etcd "github.com/coreos/etcd/clientv3" + +// OpOption is a simple typedef for etcd.OpOption. +type OpOption etcd.OpOption + +// WithPrevKV gets the previous key-value pair before the event happens. If the previous KV is already compacted, +// nothing will be returned. +func WithPrevKV() OpOption { + return OpOption(etcd.WithPrevKV()) +} + +// WithFilterPut discards PUT events from the watcher. +func WithFilterPut() OpOption { + return OpOption(etcd.WithFilterPut()) +}
[No CFG could be retrieved]
No Summary Found.
Is there a reason to lift these types rather than just using them directly? Seems like this is just some extra typing that's not needed.
@@ -19,8 +19,8 @@ def get_all_state_changes(log): def get_all_state_events(log): - """ Returns a list of tuples of event id, state_change_id and events""" + """ Returns a list of tuples of event id, state_change_id, block_number and events""" return [ - (res[0], res[1], log.serializer.deserialize(res[2])) + (res[0], res[1], res[2], log.serializer.deserialize(res[3])) for res in get_db_state_changes(log.storage, 'state_events') ]
[get_all_state_events->[get_db_state_changes],get_all_state_changes->[get_db_state_changes]]
Returns a list of tuples of event id state change id and events.
Please, use a `namedtuple`
@@ -51,6 +51,12 @@ class CouetteFlowTest(KratosUnittest.TestCase): with KratosUnittest.WorkFolderScope(self.work_folder, __file__): KratosUtilities.DeleteFileIfExisting('couette_flow_test.time') + def _ExecuteCouetteFlowTest(self): + self.setUp() + self._CustomizeSimulationSettings() + self.runTest() + self.tearDown() + def _CustomizeSimulationSettings(self): # Customize simulation settings self.parameters["solver_settings"]["solver_type"].SetString(self.solver_type)
[testCouetteFlow2DSymbolicNavierStokes,runTest,setUp,tearDown,CouetteFlowTest]
Tear down the object.
Hi, just curious... Is there a specific reason why you use a seperate function to call it? Wouldn't it be possible by using default cals of unit tests to call `setUp`, `tearDown` methods?
@@ -41,7 +41,7 @@ util.inherits(Generator, yeoman.Base); */ Generator.prototype.addElementToMenu = function (routerName, glyphiconName, enableTranslation) { try { - var fullPath = CLIENT_MAIN_SRC_DIR + 'app/layouts/navbar/navbar.html'; + var fullPath = CLIENT_MAIN_SRC_DIR + 'app/layouts/navbar/navbar.component.html'; jhipsterUtils.rewriteFile({ file: fullPath, needle: 'jhipster-needle-add-element-to-menu',
[No CFG could be retrieved]
Generates a menu element that can be added to the menu. Adds a new menu element to the admin menu.
this shouldnt be changed now. It will be have to work for both ng1 and ng2 hence will be changed after entity migration
@@ -66,13 +66,12 @@ namespace Kratos { mListOfCoordinates.clear(); mListOfRadii.clear(); - if (mpIntegrationScheme!=NULL) { + if (mpIntegrationScheme != NULL) { delete mpIntegrationScheme; } } - void Cluster3D::Initialize(ProcessInfo& r_process_info) { if (GetGeometry()[0].GetDof(VELOCITY_X).IsFixed()) GetGeometry()[0].Set(DEMFlags::FIXED_VEL_X, true);
[No CFG could be retrieved]
Creates a new Cluster3D object. Sets the fixed flags for the ANGLE geometry.
These operations can be moved to the base class RigidBodyElement3D
@@ -112,8 +112,11 @@ LocalVideo.prototype.changeVideo = function (stream) { localVideoContainer.removeChild(localVideo); // when removing only the video element and we are on stage // update the stage - if(this.isCurrentlyOnLargeVideo()) - this.VideoLayout.updateLargeVideo(this.id); + if (this.isCurrentlyOnLargeVideo()) { + this.VideoLayout.updateLargeVideo( + this.id, + true /* force - stream removed for the same user ID */); + } stream.off(TrackEvents.LOCAL_TRACK_STOPPED, endedHandler); }; stream.on(TrackEvents.LOCAL_TRACK_STOPPED, endedHandler);
[No CFG could be retrieved]
Creates a video element and attaches it to the video container. Check if the video is in the hidden state and if it is flipX state.
Why are the changes to add "force" needed? Looking at the implementation of updateLargeVideo, updates happen if the video is on large or if force is true. (Maybe the change here should be to not do any force/currentlyOnLarge checks and instead let VideoLayout handle what it needs to. That seems ideal but not sure if it's possible.)
@@ -95,7 +95,15 @@ public class Worker { currentActorId = returnId; } } finally { - runtime.getWorkerContext().setCurrentTask(null, null); + if (spec.isActorTask() || spec.isActorCreationTask()) { + // Don't need to reset current driver id if the worker is an actor. + // Because the following tasks should all have the same driver id. + runtime.getWorkerContext().setCurrentTask(null, spec.driverId, null); + } else { + // For normal task, we should set current task id to null as well. + runtime.getWorkerContext().setCurrentTask(null, null, null); + } + Thread.currentThread().setContextClassLoader(oldLoader); } }
[Worker->[loop->[info,getTask,getName,execute],execute->[RayException,checkState,debug,newInstance,equals,error,getContextClassLoader,unwrap,invoke,setCurrentTask,put,info,isActorCreationTask,isConstructor,getFunction,isActorTask,setContextClassLoader],getLogger]]
Executes the task.
Sorry I'm not clear on what `classLoader` is, but it seems like you could get rid of this line and pass in `oldLoader` in the `setCurrentTask` lines above?
@@ -115,4 +115,9 @@ class Seq2SeqDatasetReader(DatasetReader): params.assert_empty(cls.__name__) return Seq2SeqDatasetReader(source_tokenizer, target_tokenizer, source_token_indexers, target_token_indexers, - source_add_start_token, lazy) + source_add_start_token, + source_max_sequence_length, + source_truncate_sequence_length, + target_max_sequence_length, + target_truncate_sequence_length, + lazy)
[Seq2SeqDatasetReader->[from_params->[Seq2SeqDatasetReader,from_params]]]
Create a new Seq2SeqDatasetReader from params.
This predates this PR, but it'd be nice if you could pass all of these parameters by name (`source_add_start_token=source_add_start_token`), one per line. Makes it easier to maintain later if parameters change or move around. It's fine if you don't want to bother with this now, though, as it's not a problem you're introducing here.
@@ -66,7 +66,11 @@ EOT */ protected function execute(InputInterface $input, OutputInterface $output) { - $file = $input->getArgument('file'); + if ($input->hasArgument('file')) { + $file = $input->getArgument('file'); + } else { + $file = Factory::getComposerFile(); + } $io = $this->getIO(); if (!file_exists($file)) {
[ValidateCommand->[execute->[outputResult,getName,getOption,getLocker,getArgument,validate,isLocked,getInstallPath,writeError,isFresh,getPackages,getPrettyName,getLocalRepository,getIO,dispatch],outputResult->[writeError,write],configure->[setHelp]]]
Executes the command. Checks if a node is valid and returns its exit code.
`$input->hasArgument` returns `true` if the argument is defined, not if it is passed to the command. This means it will always be `true` here.
@@ -282,6 +282,7 @@ class TestPipEnvironment(TestFileEnvironment): environ['PIP_NO_INPUT'] = '1' environ['PIP_LOG_FILE'] = str(self.root_path/'pip-log.txt') + environ['PIP_USE_MIRRORS'] = 'false' super(TestPipEnvironment, self).__init__( self.root_path, ignore_hidden=False,
[FastTestPipEnvironment->[__init__->[_use_cached_pypi_server,demand_dirs,run,__init__,relpath,install_setuptools,create_virtualenv,clear_environ]],mkdir->[mkdir],TestPipEnvironment->[run->[TestPipResult],__init__->[demand_dirs,relpath,install_setuptools,create_virtualenv,clear_environ]],_change_test_package_version->[run,write_file],TestPipResult->[assert_installed->[TestFailure]],get_env->[reset_env],run_pip->[run],assert_all_changes->[TestFailure,diff_states],diff_states->[prefix_match,any],_create_test_package->[mkdir,write_file,run]]
Initialize a new environment with a specific . Create a in the user_site_path and add it to the environment.
Is there a reason the tests shouldn't use mirrors? Isn't that going to leave a lot of code paths untested?
@@ -41,7 +41,9 @@ class Repository < ApplicationRecord repository = nil, options = {} ) - repositories = repository || Repository.where(team: user.teams) + teams = user.teams.select(:id) + repositories = repository || + Repository.accessible_by_teams(teams) includes_json = { repository_cells: Extends::REPOSITORY_SEARCH_INCLUDES } searchable_attributes = ['repository_rows.name', 'users.full_name'] +
[Repository->[available_columns_ids->[pluck],search->[teams,group,offset,where_attributes_like,where],importable_repository_fields->[id,each,t,name,importable?],destroy_discarded->[discarded_by_id],import_records->[new,run],name_like->[where],copy->[call,id,created_by,find_each,transaction,save!,repository,team,dup,name],shared_with?->[any?],viewable_by_user->[where],auto_strip_attributes,include,belongs_to,default_scope,scope,validates,handle_asynchronously,has_many,attribute,lambda,sort_by]]
This method is used to search for a single node in a repository.
just `user.teams.pluck(:id)` instead of assigning to teams?
@@ -709,14 +709,6 @@ class SequenceLoadedParameter(LoadedParameter): super(SequenceLoadedParameter, self).__init__( name, value, key_flag, value_flags, validation) - def __eq__(self, other): - if type(other) is type(self): - return self.value == other.value - return False - - def __hash__(self): - return hash(self.value) - def collect_errors(self, instance, typed_value, source="<<merged>>"): errors = super(SequenceLoadedParameter, self).collect_errors( instance, typed_value, self.value)
[PrimitiveParameter->[load->[PrimitiveLoadedParameter,keyflag,value,valueflags]],ParameterLoader->[raw_parameters_from_single_source->[MultipleKeysError],__get__->[get_all_matches,load,expand,typify,collect_errors,raise_errors]],SequenceLoadedParameter->[collect_errors->[collect_errors],merge->[_first_important_matches,SequenceLoadedParameter,InvalidTypeError,get_marked_lines]],MapLoadedParameter->[collect_errors->[collect_errors],merge->[merge,InvalidTypeError,key_is_important,MapLoadedParameter,_first_important_matches]],raise_errors->[MultiValidationError],EnvRawParameter->[value->[EnvRawParameter]],ArgParseRawParameter->[value->[ArgParseRawParameter]],MapParameter->[load->[valueflags,keyflag,InvalidTypeError,load,value,MapLoadedParameter]],Configuration->[_set_argparse_args->[make_raw_parameters],check_source->[collect_errors,_raw_parameters_from_single_source,typify,load],_set_search_path->[load_file_configs],_set_env_vars->[make_raw_parameters],validate_configuration->[raise_errors,_collect_validation_error],collect_all->[check_source,raise_errors],describe_parameter->[typify],validate_all->[check_source,raise_errors],typify_parameter->[typify]],Parameter->[typify->[CustomValidationError],default->[DefaultValueRawParameter]],pretty_list->[pretty_list],InvalidTypeError->[__init__->[pretty_list]],DefaultValueRawParameter->[__init__->[DefaultValueRawParameter]],LoadedParameter->[expand->[expand_environment_variables,expand],typify->[CustomValidationError],_first_important_matches->[_match_key_is_important],collect_errors->[InvalidTypeError,ValidationError,CustomValidationError],_typify_data_structure->[typify]],ConfigurationType->[__init__->[_set_name]],MultipleKeysError->[__init__->[pretty_list]],load_file_configs->[_file_yaml_loader->[make_raw_parameters_from_file],_dir_yaml_loader->[make_raw_parameters_from_file],_get_st_mode],ParameterFlag->[from_string->[from_value]],InvalidElementTypeError->[__init__->[pretty_list]],YamlRawParameter->[make_raw_parameters_from_file->[ConfigurationLoadError,make_raw_parameters],keyflag->[from_string],make_raw_parameters->[_get_yaml_key_comment],__init__->[from_string,YamlRawParameter]],SequenceParameter->[load->[valueflags,keyflag,InvalidTypeError,load,value,SequenceLoadedParameter]]]
Initialize a sequence loaded parameter.
Are these needed for comparison elsewhere?
@@ -309,10 +309,8 @@ class Message(object): self._metadata_json = json_encode(self.metadata) return self._metadata_json + # buffer properties + @property def buffers(self): return self._buffers - - @buffers.setter - def buffers(self, value): - self._buffers = list(value)
[Message->[send->[write_buffers,Return,len,ValueError,write_message,acquire],buffers->[list],write_buffers->[write_message,ValueError,Return,len],metadata_json->[json_encode],assemble_buffer->[str,append,ProtocolError,len],assemble->[json_decode,cls,MessageError],add_buffer->[add],header_json->[json_encode],complete->[get,len],content_json->[json_encode],create_header->[make_id]]]
Returns the JSON - encoded metadata.
unused, dangerous, removed
@@ -49,7 +49,7 @@ public class OutboundAttachmentMapContext extends AbstractMapContext<DataHandler @Override public void clear() { - event.setMessage(MuleMessage.builder(event.getMessage()).clearOutboundAttachments().build()); + event.setMessage(MuleMessage.builder(event.getMessage()).outboundAttachments(emptyMap()).build()); } }
[OutboundAttachmentMapContext->[doRemove->[build,setMessage],doGet->[getOutboundAttachment],doPut->[build,setMessage],clear->[build,setMessage],keySet->[getOutboundAttachmentNames]]]
clear - clear any unhandled message.
did we remove the clear... method from the builder?
@@ -3340,4 +3340,4 @@ int opt_legacy_okay(void) if (provider_options || libctx) return 0; return 1; -} +} \ No newline at end of file
[No CFG could be retrieved]
return 0 if provider_options or libctx.
new line here too please.
@@ -64,6 +64,10 @@ def compute_csd(epochs, mode='multitaper', fmin=0, fmax=np.inf, tmin=None, Minimum frequency of interest. fmax : float | np.inf Maximum frequency of interest. + fsum : bool + Sum CSD values for the frequencies of interest. If True, a single CSD + matrix will be returned. If False, the output will be a list of CSD + matrices. tmin : float | None Minimum time instant to consider. If None start at first sample. tmax : float | None
[compute_csd->[_psd_from_mt_adaptive,any,hanning,sqrt,_csd_from_mt,sum,where,transpose,warn,info,int,_mt_spectra,zeros,CrossSpectralDensity,len,ValueError,dpss_windows,float,deepcopy,tile,slice,pick_types,array],CrossSpectralDensity->[__init__->[deepcopy,len]],getLogger]
Compute cross - spectral density from a set of epochs. Compute cross - spectral density from epochs. Compute CSD and CSD for all windows in the system. This function can be used to calculate cross - spectral density of a missing node in a cross.
Did we prefer summing over averaging? If so, the identifier used should be renamed accordingly: `csd_mean` -> `csd_sum`
@@ -1,6 +1,7 @@ // Licensed to the .NET Foundation under one or more agreements. // The .NET Foundation licenses this file to you under the MIT license. +#if NETCOREAPP using System.Diagnostics; using System.Numerics; using System.Runtime.CompilerServices;
[BitHelper->[GetIndexOfFirstNeedToEscape->[TrailingZeroCount,AggressiveInlining,Assert]]]
This class is used to find the index of the first non - zero byte that needs to.
Why this ifdef rather than conditionally including/excluding the file in the csproj?
@@ -117,7 +117,9 @@ def get_abi_tag(): d = 'd' if get_flag('WITH_PYMALLOC', lambda: impl == 'cp', - warn=(impl == 'cp')): + warn=(impl == 'cp' and + sys.version_info < (3, 8))) \ + and sys.version_info < (3, 8): m = 'm' if get_flag('Py_UNICODE_SIZE', lambda: sys.maxunicode == 0x10ffff,
[get_impl_tag->[get_impl_ver,get_abbr_impl],get_abi_tag->[get_impl_ver,get_flag,get_abbr_impl,get_config_var],get_impl_ver->[get_abbr_impl,get_config_var],get_platform->[get_platform,_is_running_32bit],is_manylinux2010_compatible->[get_platform],get_darwin_arches->[_supports_arch->[_supports_arch],_supports_arch],get_supported->[get_abi_tag,get_platform,get_darwin_arches,is_manylinux2010_compatible,get_all_minor_versions_as_strings,is_manylinux1_compatible,get_impl_version_info,get_abbr_impl],get_flag->[get_config_var],is_manylinux1_compatible->[get_platform],get_config_var->[get_config_var],get_impl_version_info->[get_abbr_impl],get_impl_tag]
Return the ABI tag based on SOABI or emulate SOABI.
Question: is there a reason the `sys.version_info < (3, 8)` has to go after the call to `get_flag()`? It seems like this would be simpler with it before because then the `sys.version_info < (3, 8)` wouldn't need to be duplicated.
@@ -130,6 +130,7 @@ func (s *balanceRegionScheduler) IsScheduleAllowed(cluster opt.Cluster) bool { func (s *balanceRegionScheduler) Schedule(cluster opt.Cluster) []*operator.Operator { schedulerCounter.WithLabelValues(s.GetName(), "schedule").Inc() stores := cluster.GetStores() + s.filters[1] = s.hitsCounter.buildSourceFilter(s.GetName(), cluster) stores = filter.SelectSourceStores(stores, s.filters, cluster) sort.Slice(stores, func(i, j int) bool { return stores[i].RegionScore(cluster.GetHighSpaceRatio(), cluster.GetLowSpaceRatio(), 0) > stores[j].RegionScore(cluster.GetHighSpaceRatio(), cluster.GetLowSpaceRatio(), 0)
[Schedule->[GetName],EncodeConfig->[EncodeConfig],transferPeer->[GetName]]
Schedule returns a list of regions that can be scheduled to be run. This function is called to transfer a region from a cluster to another cluster.
kind strange to reset filters[1] every time. I prefer `filters := append(s.filters, s.hitsCounter.buildSourceFilter(s.GetName(), cluster))`
@@ -175,8 +175,9 @@ def model_pcollection(argv): 'Or to take arms against a sea of troubles, ']) | beam.io.WriteToText(my_options.output)) - p.run() + result = p.run() # [END model_pcollection] + result.wait_until_finish() def pipeline_options_remote(argv):
[examples_wordcount_debugging->[FilterTextFn,RenameFiles],pipeline_monitoring->[CountWords->[expand->[FormatCountsFn,ExtractWordsFn]],CountWords,RenameFiles],examples_wordcount_minimal->[RenameFiles],construct_pipeline->[ReverseWords,RenameFiles],pipeline_logging->[ExtractWordsFn],examples_wordcount_wordcount->[FormatAsTextFn,CountWords,RenameFiles],model_custom_source->[ReadFromCountingSource->[expand->[_CountingSource]],ReadFromCountingSource,CountingSource],model_composite_transform_example->[CountWords],model_textio->[RenameFiles],model_multiple_pcollections_partition->[partition_fn->[get_percentile]],model_custom_sink->[WriteToKVSink->[expand->[_SimpleKVSink]],SimpleKVSink,WriteToKVSink]]
Creates a PCollection from data in local memory and a Pipeline using a PipelineOptions object for missing - node - id - v1. 0. 0.
Seems a lot of these snippets could be updated to use TestPipeline, right? (At least as long as it's part of the setup/teardown code, not the actual snippet.) They're never run not under a test...
@@ -148,11 +148,12 @@ const EXPERIMENTS = [ cleanupIssue: 'https://github.com/ampproject/amphtml/pull/6351', }, { - id: 'ios-embed-wrapper', + id: 'ios-embed-sd', name: 'A new iOS embedded viewport model that wraps the body into' + - ' a synthetic root (launched)', - spec: '', - cleanupIssue: 'https://github.com/ampproject/amphtml/issues/5639', + ' shadow root', + spec: 'https://medium.com/@dvoytenko/amp-ios-scrolling-redo-2-the' + + '-shadow-wrapper-approach-experimental-3362ed3c2fa2', + cleanupIssue: 'https://github.com/ampproject/amphtml/issues/16640', }, { id: 'chunked-amp',
[No CFG could be retrieved]
A Vega grammar for AMP Visualization. Requirements for a specific extension that is not the main JS binary.
I think this has to be changed in a clean (no code changes) PR.
@@ -86,8 +86,7 @@ namespace Microsoft.Win32 public static void SetValue(string keyName, string valueName, object value, RegistryValueKind valueKind) { - string subKeyName; - RegistryKey basekey = GetBaseKeyFromKeyName(keyName, out subKeyName); + RegistryKey basekey = GetBaseKeyFromKeyName(keyName, out string subKeyName); using (RegistryKey key = basekey.CreateSubKey(subKeyName)) {
[Registry->[SetValue->[CreateSubKey,SetValue,Unknown,Assert,GetBaseKeyFromKeyName],RegistryKey->[Arg_RegInvalidKeyName,Format,nameof,Name,Empty,IndexOf,Substring,StartsWith,OrdinalIgnoreCase,Length,ToUpperInvariant],GetValue->[GetBaseKeyFromKeyName,OpenSubKey],CurrentConfig,PerformanceData,LocalMachine,CurrentUser,OpenBaseKey,Default,ClassesRoot,Users]]
This method will try to set the value of the given key name to the given value.
Same here and above L82 `SetValue(string keyName, string? valueName, object value)` the `valueName` is nullable
@@ -39,8 +39,17 @@ func (p *Proxy) ServeTCP(conn WriteCloser) { // maybe not needed, but just in case defer connBackend.Close() - errChan := make(chan error) + + switch p.proxyProtocolVersion { + case "1", "2": + version, _ := strconv.Atoi(p.proxyProtocolVersion) + header := proxyproto.HeaderProxyFromAddrs(byte(version), conn.RemoteAddr(), conn.LocalAddr()) + _, err := header.WriteTo(connBackend) + if err != nil { + errChan <- err + } + } go p.connCopy(conn, connBackend, errChan) go p.connCopy(connBackend, conn, errChan)
[ServeTCP->[RemoteAddr,DialTCP,Close,connCopy,Errorf,WithoutContext,Debugf],connCopy->[Copy,Add,CloseWrite,Now,WithoutContext,SetReadDeadline,Debugf],ResolveTCPAddr]
ServeTCP handles a TCP connection.
Sending a message into the `errChan` is dangerous here. Since values will be pulled only 2 times, the last `p.connCopy` writing to it will be stuck. We get this error before we even start copying, so we could directly log it and return.
@@ -1527,13 +1527,14 @@ dmu_recv_begin_sync(void *arg, dmu_tx_t *tx) const char *tofs = drba->drba_cookie->drc_tofs; dsl_dataset_t *ds, *newds; uint64_t dsobj; + int flags = DS_HOLD_FLAG_DECRYPT; int error; uint64_t crflags = 0; if (drrb->drr_flags & DRR_FLAG_CI_DATA) crflags |= DS_FLAG_CI_DATASET; - error = dsl_dataset_hold(dp, tofs, FTAG, &ds); + error = dsl_dataset_hold_flags(dp, tofs, flags, FTAG, &ds); if (error == 0) { /* create temporary clone */ dsl_dataset_t *snap = NULL;
[No CFG could be retrieved]
The function to estimate the size of a record associated with a dataset. Get compressed and uncompressed size estimates of changed data.
maybe name this `dsflags` for consistency with `dmu_recv_begin_check()`?
@@ -40,7 +40,7 @@ REGISTER(SimpleEnrollPlugin, "enroll", "test_simple"); TEST_F(EnrollTests, test_enroll_secret_retrieval) { // Write an example secret (deploy key). - FLAGS_enroll_secret_path = kTestWorkingDirectory + "secret.txt"; + FLAGS_enroll_secret_path = (fs::path(kTestWorkingDirectory) / "secret.txt").make_preferred().string(); writeTextFile(FLAGS_enroll_secret_path, "test_secret\n", 0600, false); // Make sure the file content was read and trimmed. auto secret = getEnrollSecret();
[EXPECT_EQ->[EXPECT_EQ],getNodeKey->[getNodeKey,EXPECT_EQ]]
This is a test method to test if the secret is not available in the deploy.
nit, this is 80+ lines.
@@ -205,6 +205,7 @@ type PatchUpdateCheckpointRequest struct { } // AppendUpdateLogEntryRequest defines the body of a request to the append update log entry endpoint of the service API. +// No longer sent from, but the type definition is still required for backwards compat with older clients. type AppendUpdateLogEntryRequest struct { Kind string `json:"kind"` Fields map[string]interface{} `json:"fields"`
[No CFG could be retrieved]
AppendUpdateLogEntryRequest defines the body of a request to the append update log entry endpoint of.
NIT: "No longer sent from" read a little weird. Did you drop something like "the CLI"
@@ -67,7 +67,13 @@ func (r *Renewer) Close() (retErr error) { } }() ctx := context.Background() - _, err := r.tracker.SetTTLPrefix(ctx, r.id+"/", ExpireNow) + _, n, err := r.tracker.SetTTLPrefix(ctx, r.id+"/", ExpireNow) + if err != nil { + return err + } + if n != r.n { + return errors.Errorf("renewer prefix has wrong count HAVE: %d WANT: %d", n, r.n) + } return err }
[Close->[Close,Background,SetTTLPrefix],Add->[Sprintf,nextInt],nextInt->[Lock,Unlock],Background,NewRenewer,NewWithoutDashes,SetTTLPrefix]
Close closes the renewer and sets the TTL prefix to expire the renewer.
We should add a TODO to make it such that we can fail earlier when this happens. A long upload will fail right at the end when this issue is hit.
@@ -47,6 +47,17 @@ func ensureMigrated(db *sql.DB) { panic(err) } + // Look for the squashed migration. If not present, the db needs to be migrated on an earlier release first + found := false + for _, name := range names { + if name == "1611847145" { + found = true + } + } + if !found { + panic("Database state is too old. Need to migrate to chainlink version 0.9.10 first before upgrading to this version") + } + // insert records for existing migrations sql := fmt.Sprintf(`INSERT INTO %s (version_id, is_applied) VALUES ($1, true);`, goose.TableName()) err = postgres.SqlTransaction(context.Background(), db, func(tx *sqlx.Tx) error {
[TableName,Status,Exec,Index,EnsureDBVersion,SetBaseFS,Up,ParseBool,New,Select,SetTableName,SetSequential,GetDBVersion,ParseInt,DownTo,Create,WrapDbWithSqlx,SqlTransaction,Sprintf,Background,SetVerbose,Getenv,Down]
requires that all migrations have been run in the same database Migration functions - Migration functions - Rollback migrations - Up migrations - Rollback migrations - Rollback migrations.
Do we rather want to panic or dirty exit here?
@@ -28,7 +28,7 @@ $is_master_user = $current_user->ID == Jetpack_Options::get_option( 'master_u <a href="http://jetpack.me/contact-support/" title="<?php esc_attr_e( 'Contact the Jetpack Happiness Squad.', 'jetpack' ); ?>"><?php _e( 'Support', 'jetpack' ); ?></a> <a href="http://jetpack.me/survey/?rel=<?php echo JETPACK__VERSION; ?>" title="<?php esc_attr_e( 'Take a survey. Tell us how we&#8217;re doing.', 'jetpack' ); ?>"><?php _e( 'Give Us Feedback', 'jetpack' ); ?></a> - <?php if ( $is_active && current_user_can( 'jetpack_disconnect' ) ) : ?> + <?php if ( $is_active && current_user_can( 'jetpack_disconnect' ) && ! Jetpack::is_development_mode() ) : ?> <a href="<?php echo wp_nonce_url( Jetpack::admin_url( 'action=disconnect' ), 'jetpack-disconnect' ); ?>" onclick="return confirm('<?php echo htmlspecialchars( __('Are you sure you want to disconnect from WordPress.com?', 'jetpack'), ENT_QUOTES ); ?>');"><?php esc_html_e( 'Disconnect from WordPress.com', 'jetpack' ); ?></a> <?php endif; ?> <?php if ( $is_active && $is_user_connected && ! $is_master_user ) : ?>
[No CFG could be retrieved]
Displays a list of all the nodes of a single node. Show a hidden hidden menu with a link to the user s account.
Wouldn't this be unnecessary if we modified $is_active so it's never showing as active in dev mode?
@@ -3426,13 +3426,12 @@ void ssl_update_cache(SSL *s, int mode) /* auto flush every 255 connections */ if ((!(i & SSL_SESS_CACHE_NO_AUTO_CLEAR)) && ((i & mode) == mode)) { - int *stat, val; + TSAN_QUALIFIER int *stat; if (mode & SSL_SESS_CACHE_CLIENT) stat = &s->session_ctx->stats.sess_connect_good; else stat = &s->session_ctx->stats.sess_accept_good; - if (CRYPTO_atomic_read(stat, &val, s->session_ctx->lock) - && (val & 0xff) == 0xff) + if ((tsan_load(stat) & 0xff) == 0xff) SSL_CTX_flush_sessions(s->session_ctx, (unsigned long)time(NULL)); } }
[No CFG could be retrieved]
The method for caching a session id. Private functions for SSL_CTX_flush_sessions.
On related note this is kind of a head scratcher. I mean the chances that if misses `*stat & 0xff == 0xff`...
@@ -16,12 +16,12 @@ * through the world wide web, please send an email to * licensing@ellislab.com so we can send you a copy immediately. * - * @package CodeIgniter - * @author EllisLab Dev Team - * @copyright Copyright (c) 2008 - 2013, EllisLab, Inc. (http://ellislab.com/) - * @license http://opensource.org/licenses/AFL-3.0 Academic Free License (AFL 3.0) - * @link http://codeigniter.com - * @since Version 1.0 + * @package CodeIgniter + * @author EllisLab Dev Team + * @copyright Copyright (c) 2008 - 2013, EllisLab, Inc. (http://ellislab.com/) + * @license http://opensource.org/licenses/AFL-3.0 Academic Free License (AFL 3.0) + * @link http://codeigniter.com + * @since Version 1.0 * @filesource */
[No CFG could be retrieved]
This function is used to provide a description of a specific .
DO NOT, under any reasoning change those lines. Even I got bashed for doing it. :)
@@ -28,7 +28,7 @@ namespace System.Runtime.InteropServices.Tests { int errorExpected = 123; Marshal.SetLastPInvokeError(errorExpected); - Assert.Equal(errorExpected, Marshal.GetLastWin32Error()); + Assert.Equal(errorExpected, Marshal.GetLastPInvokeError()); } [Fact]
[LastErrorTests->[SetLastPInvokeError_GetLastWin32Error->[GetLastWin32Error,Equal,SetLastPInvokeError],SetLastSystemError_PInvokeErrorUnchanged->[Equal,GetLastPInvokeError,SetLastSystemError,SetLastPInvokeError,NotEqual,GetLastWin32Error],LastSystemError_RoundTrip->[GetLastSystemError,SetLastSystemError,Equal],LastPInvokeError_RoundTrip->[GetLastPInvokeError,Equal,SetLastPInvokeError]]]
Set last pinvoke error and last system error.
This change should be reverted. (Otherwise, the test would be redundant with `LastPInvokeError_RoundTrip`.)
@@ -7,7 +7,6 @@ import KratosMultiphysics.KratosUnittest as KratosUnittest # Import the tests o test_classes to create the suits from generalTests import KratosMetisGeneralTests -from test_metis_submodelpart_list import TestMetisSubModelPartList def AssembleTestSuites(): ''' Populates the test suites to run.
[AssembleTestSuites->[KratosMetisGeneralTests,addTest,TestLoader,addTests],AssembleTestSuites,runTests]
Assemble a test suites with the specified test_cases.
could you delete this file as it does nothing? also `test_MetisApplication.py` is unused so can be removed too I would suggest
@@ -255,10 +255,16 @@ def process_options(args: List[str], parser.add_argument('-V', '--version', action='version', version='%(prog)s ' + __version__) parser.add_argument('--python-version', type=parse_version, metavar='x.y', - help='use Python x.y') + help='use Python x.y', dest='special-opts:python_version') + parser.add_argument('--python-executable', action='store', metavar='EXECUTABLE', + help="Python executable which will be used in typechecking.", + dest='special-opts:python_executable') + parser.add_argument('--no-site-packages', action='store_true', + dest='special-opts:no_executable', + help="Do not search for installed PEP 561 compliant packages.") parser.add_argument('--platform', action='store', metavar='PLATFORM', help="typecheck special-cased code for the given OS platform " - "(defaults to sys.platform).") + "(defaults to sys.platform).") parser.add_argument('-2', '--py2', dest='python_version', action='store_const', const=defaults.PYTHON2_VERSION, help="use Python 2 mode") parser.add_argument('--ignore-missing-imports', action='store_true',
[expand_dir->[expand_dir],process_options->[add_invertible_flag->[invert_flag_name],add_invertible_flag,SplitNamespace],crawl_up->[InvalidPackageName]]
Parse command line arguments and return a tuple of build sources and options. Add command line options to typecheck. Add flags that can be used to check for missing type annotations. Parses command line options and returns a single node ID.
The help text is a bit vague. What about something like "Python executable used for finding PEP 561 compliant installed packages and stubs"? Remove dot from the end of the help text.
@@ -156,7 +156,10 @@ func (r *LocalRuntime) LoadFromArchiveReference(ctx context.Context, srcRef type } // New calls into local storage to look for an image in local storage or to pull it -func (r *LocalRuntime) New(ctx context.Context, name, signaturePolicyPath, authfile string, writer io.Writer, dockeroptions *image.DockerRegistryOptions, signingoptions image.SigningOptions, forcePull bool) (*ContainerImage, error) { +func (r *LocalRuntime) New(ctx context.Context, name, signaturePolicyPath, authfile string, writer io.Writer, dockeroptions *image.DockerRegistryOptions, signingoptions image.SigningOptions, forcePull bool, label *string) (*ContainerImage, error) { + if label != nil { + return nil, errors.New("the remote client function does not support checking a remote image for a label") + } // TODO Creds needs to be figured out here too, like above tlsBool := dockeroptions.DockerInsecureSkipTLSVerify // Remember SkipTlsVerify is the opposite of tlsverify
[LoadFromArchiveReference->[NewImageFromLocal],New->[NewImageFromLocal],TagImage->[ID,TagImage],Dangling->[Names],RemoveImage->[RemoveImage]]
New creates a new container image from the local docker registry.
Why is it OK to just silently ignore the `label` parameter?
@@ -225,7 +225,13 @@ type RemoveKeyRule struct { } // Apply applies remove key rule. -func (r *RemoveKeyRule) Apply(_ AgentInfo, ast *AST) error { +func (r *RemoveKeyRule) Apply(_ AgentInfo, ast *AST) (err error) { + defer func() { + if err != nil { + err = errors.New(err, "RemoveKeyRule") + } + }() + sourceMap, ok := ast.root.(*Dict) if !ok { return nil
[Apply->[ReplaceAllString,Value,MatchString,Sprintf,Version,Apply,New,Find,ValueOf,AgentID,String,Clone,Errorf,InjectItem,Snapshot,Inject,Map],MarshalYAML->[String,Errorf,MarshalYAML],UnmarshalYAML->[Marshal,New,Unmarshal,Compile,Errorf],Value,Apply,Find,String,Errorf]
Apply removes the key rule from the source map.
Pretty cryptic error message. It will make it easy for the software engineer to find where the error happened but will not be useful for an end user I think.
@@ -52,7 +52,7 @@ func newConfigCmd() *cobra.Command { cmd.PersistentFlags().StringVarP( &stack, "stack", "s", "", - "Choose an stack other than the currently selected one") + "Target a specific stack instead of all stacks") cmd.PersistentFlags().BoolVar( &unset, "unset", false, "Unset a configuration value")
[Strings,Printf,Sprintf,Contains,RunFunc,StringVarP,String,Errorf,ParseModuleMember,ModuleMember,Assert,Wrap,HasPrefix,PersistentFlags,BoolVar]
newConfigCmd returns a new command that can be used to set or delete a configuration value prettyKey returns the name of the last known node in the stack or an error if the.
`... instead of all of this project's stacks`?
@@ -24,7 +24,6 @@ with this program; if not, write to the Free Software Foundation, Inc., #include "log.h" #include "itemdef.h" #include "util/strfnd.h" -#include "content_mapnode.h" // For loading legacy MaterialItems #include "nameidmapping.h" // For loading legacy MaterialItems #include "util/serialize.h" #include "util/string.h"
[removeItem->[addItem,takeItem],addItem->[addItem],takeItem->[ItemStack],clearContents->[getSize,deleteItem],roomForItem->[itemFits],serialize->[serialize,getSize],getFreeSlots->[getUsedSlots,getSize],deSerialize->[clearItems,deSerialize,clear],clear->[clear],moveItem->[ItemStack,changeItem,takeItem,addItem],getItemString->[serialize],moveItemSomewhere->[takeItem,addItem,getSize,ItemStack,changeItem]]
Replies the object that holds the content of a node in the system. - - - - - - - - - - - - - - - - - -.
can this include be removed?
@@ -54,6 +54,9 @@ class FreqtradeBot: self._heartbeat_msg = 0 + self._sell_rate_cache = TTLCache(maxsize=100, ttl=5) + self._buy_rate_cache = TTLCache(maxsize=100, ttl=5) + self.heartbeat_interval = self.config.get('internals', {}).get('heartbeat_interval', 60) self.strategy: IStrategy = StrategyResolver.load_strategy(self.config)
[FreqtradeBot->[execute_buy->[get_buy_rate,_get_min_pair_stake_amount],_check_available_stake_amount->[_get_available_stake_amount],_notify_buy_cancel->[get_buy_rate],_notify_sell->[get_sell_rate],_notify_sell_cancel->[get_sell_rate],cleanup->[cleanup],handle_trailing_stoploss_on_exchange->[create_stoploss_order],handle_stoploss_on_exchange->[create_stoploss_order],_calculate_unlimited_stake_amount->[get_free_open_trades,_get_available_stake_amount],execute_sell->[_safe_sell_amount],handle_trade->[get_sell_rate],create_trade->[get_free_open_trades,get_trade_stake_amount],check_handle_timedout->[_notify_buy_cancel,_check_timed_out]]]
Initialize FreqtradeBot object with base class objects This method is called when the object is initialized.
this can later be moved to the dataprovider... what do you think?
@@ -2710,7 +2710,7 @@ class Facture extends CommonInvoice $txtva = preg_replace('/\s*\(.*\)/', '', $txtva); // Remove code into vatrate. } - $tabprice=calcul_price_total($qty, $pu, $remise_percent, $txtva, $txlocaltax1, $txlocaltax2, 0, $price_base_type, $info_bits, $type, $mysoc, $localtaxes_type, $situation_percent, $this->multicurrency_tx, $pu_ht_devise); + $tabprice=calcul_price_total($qty, $pu, $remise_percent, $txtva, $txlocaltax1, $txlocaltax2, $type, $price_base_type, $info_bits, $type, $mysoc, $localtaxes_type, $situation_percent, $this->multicurrency_tx, $pu_ht_devise); $total_ht = $tabprice[0]; $total_tva = $tabprice[1];
[Facture->[is_erasable->[getNextNumRef,getVentilExportCompta],getLinesArray->[fetch_lines],deleteline->[delete,fetch],insert_discount->[fetch],getSumDepositsUsed->[getSumDepositsUsed],update_percent->[update],info->[fetch],createFromOrder->[create],getSumCreditNotesUsed->[getSumCreditNotesUsed],get_prev_sits->[fetch],demande_prelevement->[getSumCreditNotesUsed,fetch,getSumDepositsUsed],updateline->[fetch,update],fetchPreviousNextSituationInvoice->[fetch],createFromCurrent->[create],validate->[set_canceled,fetch,fetch_lines],createFromClone->[create],delete->[fetch_lines],addline->[fetch]],FactureLigne->[insert->[fetch],get_prev_progress->[fetch]]]
Updates a line in the cart. This function initializes the object with all the necessary data. This function is called by the facture_ligne class when a new object is This function is called when a new product is added to the invoice.
The field type is alread at position 10. At position 7, we should have 0, it is a field no more used.
@@ -1637,7 +1637,7 @@ public abstract class PluginManager extends AbstractModelObject implements OnMas @RequirePOST public HttpResponse doUploadPlugin(StaplerRequest req) throws IOException, ServletException { try { - Jenkins.get().checkPermission(UPLOAD_PLUGINS); + Jenkins.get().checkPermission(Jenkins.ADMINISTER); ServletFileUpload upload = new ServletFileUpload(new DiskFileItemFactory());
[PluginManager->[createDefault->[create],doPlugins->[getDisplayName],install->[start,install,getPlugin],dynamicLoad->[dynamicLoad],getPlugin->[getPlugin,getPlugins],PluginUpdateMonitor->[ifPluginOlderThenReport->[getPlugin]],addDependencies->[addDependencies],start->[run],loadDetachedPlugins->[loadPluginsFromWar],getPluginVersion->[getPluginVersion],loadPluginsFromWar->[loadPluginsFromWar],PluginCycleDependenciesMonitor->[isActivated->[getPlugins]],doPrevalidateConfig->[getPlugin],getPlugins->[getPlugin,getPlugins],stop->[stop],doInstallNecessaryPlugins->[prevalidateConfig],prevalidateConfig->[getPlugin],disablePlugins->[getPlugin],doCheckUpdatesServer->[start],create->[doCreate],parseRequestedPlugins->[resolveEntity->[resolveEntity]],initTasks->[run->[]]]]
Upload a plugin to the server. finds an unused update center and returns a redirect to the update center.
Strictly speaking, none of the permission checks in this file matter, but if System Read is about to land, it probably makes sense to keep them.
@@ -20,9 +20,9 @@ module Api fields = fields_to_render(@@default_fields) respond_to do |format| - format.xml{render :xml => test_results.to_xml(:only => fields, :root => - 'test_results', :skip_types => 'true')} - format.json{render :json => test_results.to_json(:only => fields)} + format.xml{render xml: test_results.to_xml(only: fields, root: + 'test_results', skip_types: 'true')} + format.json{render json: test_results.to_json(only: fields)} end end
[TestResultsController->[destroy->[destroy],create->[create]]]
This view shows the nag - test result for a given assignment group.
`end` at 24, 45 is not aligned with `format.xml{render xml: test_results.to_xml(only: fields, root:` at 23, 8<br>Align the elements of a hash literal if they span more than one line.
@@ -138,7 +138,16 @@ public class KsqlConfigTest { } @Test - public void shouldCleanStreamsProperties() { + public void shouldSetMonitoringInterceptorConfigProperties() { + final KsqlConfig ksqlConfig = new KsqlConfig(Collections.singletonMap( + "confluent.monitoring.interceptor.topic", "foo")); + final Object result + = ksqlConfig.getKsqlStreamConfigProps().get("confluent.monitoring.interceptor.topic"); + assertThat(result, equalTo("foo")); + } + + @Test + public void shouldObfuscateSecretStreamsProperties() { final String password = "super-secret-password"; final KsqlConfig ksqlConfig = new KsqlConfig(Collections.singletonMap( SslConfigs.SSL_KEY_PASSWORD_CONFIG, password
[KsqlConfigTest->[shouldSetStreamsConfigProducerPrefixedProperties->[singletonMap,equalTo,get,KsqlConfig,assertThat],shouldSetStreamsConfigProperties->[singletonMap,equalTo,get,KsqlConfig,assertThat],shouldCloneWithStreamPropertyOverwrite->[singletonMap,get,cloneWithPropertyOverwrite,equalTo,KsqlConfig,assertThat],shouldSetLogAndContinueExceptionHandlerWhenFailOnDeserializationErrorFalse->[singletonMap,equalTo,get,KsqlConfig,assertThat],shouldPreserveOriginalCompatibilitySensitiveConfigs->[overrideBreakingConfigsWithOriginalValues,of,equalTo,getString,KsqlConfig,emptyMap,assertThat],shouldSetLogAndContinueExceptionHandlerByDefault->[equalTo,get,KsqlConfig,emptyMap,assertThat],shouldSetInitialValuesCorrectly->[getShort,equalTo,KsqlConfig,getInt,assertThat,put],shouldSetStreamsConfigConsumerPrefixedProperties->[singletonMap,equalTo,get,KsqlConfig,assertThat],shouldSetStreamsConfigProducerUnprefixedProperties->[singletonMap,equalTo,get,KsqlConfig,assertThat],shouldCloneWithKsqlPropertyOverwrite->[singletonMap,equalTo,cloneWithPropertyOverwrite,getString,KsqlConfig,assertThat],shouldCleanStreamsProperties->[value,singletonMap,get,equalTo,KsqlConfig,assertThat,not],shouldNotSetDeserializationExceptionHandlerWhenFailOnDeserializationErrorTrue->[singletonMap,get,KsqlConfig,nullValue,assertThat],shouldSetStreamsConfigAdminClientProperties->[singletonMap,equalTo,get,KsqlConfig,assertThat],shouldSetStreamsConfigConsumerUnprefixedProperties->[singletonMap,equalTo,get,KsqlConfig,assertThat],shouldCloneWithPrefixedStreamPropertyOverwrite->[singletonMap,get,cloneWithPropertyOverwrite,equalTo,KsqlConfig,assertThat],shouldUseCurrentValueForCompatibilityInsensitiveConfigs->[overrideBreakingConfigsWithOriginalValues,singletonMap,getBoolean,KsqlConfig,is,assertThat],shouldSetPrefixedStreamsConfigProperties->[singletonMap,equalTo,get,KsqlConfig,assertThat]]]
Checks if the streams properties should be cleaned.
Would be good to use the actual monitoring interceptor config. It is something like `producer.interceptor.classes` and `consumer.interceptor.classes`.
@@ -75,6 +75,9 @@ public class EsWFSFeatureIndexer { public static final String TREE_FIELD_SUFFIX = "_tree"; public static final String FEATURE_FIELD_PREFIX = "ft_"; + @Value("es.index.features") + private String index = "features"; + @Autowired private EsClient client;
[EsWFSFeatureIndexer->[deleteFeatures->[deleteFeatures],indexFeatures->[saveHarvesterReport]]]
Creates a new feature type with the specified properties. Gets the configuration of the WFS harvester.
Will this work with multinode setup?
@@ -3095,11 +3095,9 @@ namespace System.Diagnostics.Tracing if (attr != null) { - Type t = attr.GetType(); - foreach (CustomAttributeNamedArgument namedArgument in data.NamedArguments) { - PropertyInfo p = t.GetProperty(namedArgument.MemberInfo.Name, BindingFlags.Public | BindingFlags.Instance)!; + PropertyInfo p = attributeType.GetProperty(namedArgument.MemberInfo.Name, BindingFlags.Public | BindingFlags.Instance)!; object value = namedArgument.TypedValue.Value!; if (p.PropertyType.IsEnum)
[No CFG could be retrieved]
Replies if two related EventSource - domain types should be considered the same. is the running EventSource type.
I think this can potentially break the existing code. The attribute match above is name based (`AttributeTypeNamesMatch`) so there's no guarantee that `attributeType` and `attr.GetType()` will be type-system-related. They can be two separate types which happen to have the same name. After this change the `p.SetValue` below may fail if the types were different.
@@ -7,6 +7,10 @@ http.cpp HTTP server handling ***************************************************************************/ +#ifdef __sun +#define ASIO_DISABLE_DEV_POLL +#define ASIO_HAS_EPOLL +#endif #include "emu.h" #include "server_ws_impl.hpp"
[No CFG could be retrieved]
Structure of all of the types of a file in the system. Returns an array of media types that can be used by the application.
I get a C4603 warning in VS2017 if this is defined before including `emu.h`. @despair86 can it be safely moved below the includes?
@@ -72,7 +72,7 @@ from apache_beam.transforms.display import DisplayDataItem from apache_beam.transforms.display import HasDisplayData from apache_beam.typehints import native_type_compatibility from apache_beam.typehints import typehints -from apache_beam.typehints.decorators import TypeCheckError +from apache_beam.typehints.decorators import TypeCheckError, IOTypeHints, get_type_hints from apache_beam.typehints.decorators import WithTypeHints from apache_beam.typehints.decorators import get_signature from apache_beam.typehints.decorators import getcallargs_forhints
[_FinalizeMaterialization->[visit->[visit_nested]],_ZipPValues->[visit->[visit],visit_dict->[visit],visit_sequence->[visit]],_AddMaterializationTransforms->[_materialize_transform->[_allocate_materialized_result,_MaterializeValuesDoFn],visit->[visit_nested,_MaterializedDoOutputsTuple,_materialize_transform,visit]],PTransform->[register_urn->[register->[register],register],type_check_inputs_or_outputs->[_ZipPValues],__ror__->[_release_materialized_pipeline,_FinalizeMaterialization,_AddMaterializationTransforms,_allocate_materialized_pipeline,_SetInputPValues],_extract_input_pvalues->[_dict_tuple_leaves->[_dict_tuple_leaves],_dict_tuple_leaves]],_ChainedPTransform->[__or__->[_ChainedPTransform]],_NamedPTransform->[__ror__->[__ror__]],_create_transform->[PTransform],get_named_nested_pvalues->[get_named_nested_pvalues],PTransformWithSideInputs->[type_check_inputs->[element_type],with_input_types->[with_input_types],default_label->[default_label]],_SetInputPValues->[visit->[visit_nested]],ptransform_fn->[callable_ptransform_factory->[_PTransformFnPTransform]],label_from_callable->[default_label],register_urn]
Imports a single object from a list of objects Visitor for PValueish objects.
Please stay consistent with existing code: one import per line.
@@ -189,7 +189,8 @@ class DataInputOperation(RunnerIOOperation): self.name_context.step_name, 0, next(iter(itervalues(consumers))), - self.windowed_coder) + self.windowed_coder, + get_perf_runtime_type_hints(self)) ] self.splitting_lock = threading.Lock() self.index = -1
[create_combine_per_key_precombine->[get_only_output_coder,augment_oldstyle_op,get_input_windowing],StateBackedSideInputMap->[__getitem__->[MultiMap->[],MultiMap,_StateBackedIterable]],create_read_from_impulse_python->[get_only_output_coder],ReadModifyWriteRuntimeState->[commit->[commit],read->[read],clear->[clear]],create_source_java->[get_only_output_coder,augment_oldstyle_op],create_assign_windows->[WindowIntoDoFn,_create_simple_pardo_operation],create_deprecated_read->[augment_oldstyle_op],CombiningValueRuntimeState->[commit->[commit],add->[_read_accumulator,add,clear],clear->[clear],_read_accumulator->[read,clear],read->[_read_accumulator]],BundleProcessor->[monitoring_infos->[monitoring_infos],reset->[reset],create_execution_tree->[topological_height->[topological_height],get_operation->[get_operation],is_side_input,get_operation],finalize_bundle->[finalize_bundle],process_bundle->[set_output_stream,finish,start],try_split->[try_split,add]],DataInputOperation->[_compute_split->[is_valid_split_point,try_split]],_create_pardo_operation->[augment_oldstyle_op,StateBackedSideInputMap,get_output_coders,get_input_coders,FnApiUserStateContext,set,get_windowed_coder],create_identity_dofn->[get_only_output_coder,augment_oldstyle_op],SynchronousBagRuntimeState->[commit->[clear],read->[_StateBackedIterable,_ConcatIterable]],BeamTransformFactory->[extract_timers_info->[TimerInfo],get_input_windowing->[only_element],get_only_output_coder->[get_output_coders,only_element],get_output_coders->[get_windowed_coder],get_input_coders->[get_windowed_coder],__init__->[_StateBackedIterable],get_only_input_coder->[get_input_coders,only_element],get_windowed_coder->[get_coder]],create_flatten->[get_only_output_coder,augment_oldstyle_op],_create_combine_phase_operation->[get_only_output_coder,augment_oldstyle_op],FnApiUserStateContext->[commit->[commit],_create_state->[SynchronousSetRuntimeState,CombiningValueRuntimeState,ReadModifyWriteRuntimeState,SynchronousBagRuntimeState],get_timer->[OutputTimer]],create_source_runner->[DataInputOperation,get_coder],_create_simple_pardo_operation->[_create_pardo_operation],create_map_windows->[MapWindows,_create_simple_pardo_operation],SynchronousSetRuntimeState->[_compact_data->[_StateBackedIterable,_ConcatIterable,clear],commit->[clear],read->[_compact_data],add->[_compact_data,add,clear]],create_sink_runner->[DataOutputOperation,get_coder],register_urn]
Initialize a DataInputOperation with a sequence of missing values.
This should just be None (or better, omitted) for operations without specs. We should probably just have a method on ParDoOperation to extract the type hints as this is the one case we support.
@@ -474,6 +474,9 @@ public class IncrementalIndex implements Iterable<Row>, Closeable } } in.set(null); + if(row.getTimestamp().isAfter(lastIngestedEventTime)) { + lastIngestedEventTime = row.getTimestamp(); + } return numEntries.get(); }
[IncrementalIndex->[makeDimensionSelector->[lookupName->[get],getRow->[size->[size],get->[get],iterator->[apply->[],iterator]]],getMaxTime->[getMaxTimeMillis,isEmpty],close->[close],isEmpty->[get],getDimensionIndex->[get],DimDimImpl->[size->[size],getId->[get],getValue->[get],add->[size],sort->[size,sort]],makeObjectColumnSelector->[get->[get]],size->[get],getMetricType->[get],getDimVals->[size,get,add],getDimension->[get,isEmpty],getCapabilities->[get],DimensionHolder->[get->[get],add->[get]],getMetricBuffer->[get],getMetricIndex->[get],iterableWithPostAggregations->[iterator->[apply->[getMetricPosition,get],iterator]],TimeAndDims->[compareTo->[compareTo]],getInterval->[getMaxTimeMillis,isEmpty],add->[formatRow,size,get,add],getMinTime->[isEmpty,getMinTimeMillis],iterator->[apply->[],iterator]]]
Add a row to the aggregation. This method is called when a new row is added to the buffer.
This has a race condition, I think the compare-then-set doesn't appear to be synchronized on anything.
@@ -236,11 +236,6 @@ func (a *imagePolicyPlugin) admit(attr admission.Attributes, mutationAllowed boo return nil } - // This will convert any non-legacy Origin resource to a legacy resource, so specifying - // a 'builds.build.openshift.io' is converted to 'builds'. - // TODO: denormalize this at config time, or write a migration for user's config - attr = mutateAttributesToLegacyResources(attr) - policy := resolutionConfig{a.config} schemagr := attr.GetResource().GroupResource()
[SetDefaultRegistryFunc->[RegistryNameMatcher],Mutate->[DeepCopy,Mutate,Infof,DeepEqual,V,Errorf,ObjectGoPrintSideBySide],ValidateInitialization->[Errorf],Validate->[admit],resolveImageReference->[Before,Exact,Images,resolveImageStreamTag,Matches,Errorf,Get,Add],SetRESTClientConfig->[NewForConfig,HandleError],resolveImageStreamTag->[ImageStreams,ParseDockerImageReference,Infof,JoinImageStreamTag,V,Errorf,ImageStreamTags,Get,Add],admit->[GetAnnotationAccessor,GetObject,GroupResource,GetImageReferenceMutator,NewForbidden,GetName,GetResource,Errorf,NewString,GetNamespace,GetSubresource,Infof,GetOldObject,V,Get,Covers,Split,GetKind,GetOperation],ResolveObjectReference->[SplitImageStreamImage,ParseDockerImageReference,resolveImageReference,resolveImageStreamTag,Errorf,resolveImageStreamImage,SplitImageStreamTag],resolveImageStreamImage->[ParseDockerImageReference,Errorf,ImageStreamImages,Get,Add,JoinImageStreamImage],Admit->[admit],SetExternalKubeInformerFactory->[Core,V1,Lister,Namespaces],Status,SetDefaults_ImagePolicyConfig,NewRegistryMatcher,GetObject,NewHandler,Validate,New,ValidationInterface,NewCodecFactory,GetName,GetResource,GetNamespace,GetSubresource,Infof,MutationInterface,NewAttributesRecord,GetOldObject,Must,WantsRESTClientConfig,ToAggregate,V,Register,DecodeInto,WantsDefaultRegistryFunc,NewScheme,UniversalDecoder,GetKind,NewExecutionRulesAccepter,Install,GetUserInfo,WantsExternalKubeInformerFactory,GetOperation,IsNotFound,ReadAll]
admit performs the admission admission for the given attributes. Get - get - policy - rules - annotation - if present - accept it.
@bparees this was also suspicious
@@ -313,7 +313,6 @@ const files = { path: TEST_SRC_DIR, templates: [ 'jest.conf.js', - 'spec/enzyme-setup.ts', 'spec/storage-mock.ts', 'spec/app/utils.ts', 'spec/app/config/axios-interceptor.spec.ts',
[No CFG could be retrieved]
This module provides a list of all possible configuration files that can be used to configure a layout Component spec for all components.
Cleanup this file.
@@ -194,6 +194,11 @@ func createRouter(ctx context.Context, prefix string, svr *server.Server) (*mux. apiRouter.Handle("/debug/pprof/block", pprof.Handler("block")) apiRouter.Handle("/debug/pprof/goroutine", pprof.Handler("goroutine")) + // swagger docs + swaggerHandler := swaggerserver.Handler() + apiRouter.PathPrefix("/swagger/").Handler(swaggerHandler) + apiRouter.Path("/swagger").Subrouter().StrictSlash(true).Handle("/", swaggerHandler) + // Deprecated rootRouter.Handle("/health", newHealthHandler(svr, rd)).Methods("GET") // Deprecated
[Handle,Vars,WithInsecure,GetConfig,GetMessage,ToTLSConfig,RegisterConfigHandlerFromEndpoint,PathPrefix,WithTransportCredentials,Handler,Error,NewRouter,New,NewServeMux,GetAddr,JSON,HandleFunc,GetSecurityConfig,Methods,NewTLS,GetAllComponentIDs,GetConfigManager,WithMarshalerOption,GetHandler,GetComponentIDs,Write,WithUnaryInterceptor,GetStatus,Use,WithForwardResponseOption,Subrouter,NewRoute,Parse]
Initialize the router for all the Nagios - specific routes. lazyComponentRouter returns a list of component IDs.
Is it better to add a `HandlerBuilder` to `main.go` and access it directly via `/swagger`?
@@ -724,7 +724,7 @@ function bb_CleanPictureLinks($text) { } function bb_highlight($match) { - if(in_array(strtolower($match[1]),['php','css','mysql','sql','abap','diff','html','perl','ruby', + if (in_array(strtolower($match[1]),['php','css','mysql','sql','abap','diff','html','perl','ruby', 'vbscript','avrc','dtd','java','xml','cpp','python','javascript','js','sh'])) return text_highlight($match[2],strtolower($match[1])); return $match[0];
[bbcode->[saveHTML,loadHTML],bb_RemovePictureLinks->[save_timestamp,loadHTML,query,get_useragent],bb_CleanPictureLinksSub->[save_timestamp,loadHTML,query,get_useragent]]
Highlight bb code.
Standards: Please add braces to this condition and a space after commas.
@@ -75,7 +75,7 @@ <% end %> <% if @article.video_state == "PROGRESSING" %> - <div class="crayons-notice crayons-notice--warning mb-4" aria-live="polite">⏳ Video Transcoding In Progress ⏳</div> + <div class="crayons-notice crayons-notice--warning mb-4" aria-live="polite"><%= t("views.editor.video.progress.heading") %></div> <% end %> <article class="crayons-card crayons-article mb-4"
[No CFG could be retrieved]
Displays a single unique key in the network. Renders the crayons - hierarchy hierarchy of the n - ary.
Isn't there a similar key in a different PR? Maybe we should re-use that to ensure we don't get copy drift between the two locations.
@@ -114,9 +114,13 @@ async function main() { // List of contracts the listener watches events from. const contracts = { IdentityEvents: contractsContext.identityEvents, - Marketplace: contractsContext.marketplaces['000'].contract, ProxyFactory: contractsContext.ProxyFactory } + // Listen to all versions of marketplace + Object.keys(contractsContext.marketplaces).forEach(key => { + contracts[`V${key}_Marketplace`] = + contractsContext.marketplaces[key].contract + }) if (context.config.concurrency > 1) { logger.warn(`Backfill mode: concurrency=${context.config.concurrency}`)
[No CFG could be retrieved]
The main function of the tracking loop. Get the for a given contract.
One thing I just realized was `key` is 3 characters long (eg '000') but the config is 2 (eg 'V00'). Not sure if that affects this. Also might need to say `Object.keys(contractsContext.marketplaces || {})` incase no marketplaces are set yet for whatever reason
@@ -85,13 +85,11 @@ final class NativeDatagramPacketArray implements ChannelOutboundBuffer.MessagePr void clear() { this.count = 0; + this.iovArray.clear(); } void release() { - // Release all packets - for (NativeDatagramPacket datagramPacket : packets) { - datagramPacket.release(); - } + iovArray.release(); } /**
[NativeDatagramPacketArray->[processMessage->[add],NativeDatagramPacket->[init->[count,add,clear],release->[release]],release->[release]]]
clear - clear all packets.
Should we be using cleaner on the allocated byte buffer in clean?
@@ -509,6 +509,14 @@ func (b *Beat) Setup(bt beat.Creator, template, setupDashboards, machineLearning fmt.Println("Loaded Ingest pipelines") } + if policy { + err := b.loadILMPolicy() + if err != nil { + return err + } + fmt.Println("ILM policy loaded for Beat") + } + return nil }()) }
[launch->[InitWithSettings,createBeater],TestConfig->[createBeater,Init],Setup->[createBeater,Init],createBeater->[BeatConfig],configure->[BeatConfig],Init->[InitWithSettings]]
Setup initializes the beat and loads the template and dashboards Load the machine learning job configurations and load the ingest pipelines if they are available.
Nitpick - could you align with the other loading outputs to print `Loaded Index Lifecycle Management (ILM) policies`.
@@ -64,7 +64,7 @@ public final class LocalWriteCommand extends LocalFileCommand implements WriteCo throw new FileAccessDeniedException(format("Could not write to file '%s' because access was denied by the operating system", path), e); - } catch (org.mule.extension.file.common.api.exceptions.FileAlreadyExistsException e) { + } catch (ModuleException e) { throw e; } catch (Exception e) { throw exception(format("Exception was found writing to file '%s'", path), e);
[LocalWriteCommand->[getOutputStream->[FileAlreadyExistsException,format,newOutputStream],getOpenOptions->[IllegalArgumentException],write->[resolvePath,exception,getOutputStream,lock,copy,format,FileAccessDeniedException,release,assureParentFolderExists,NullPathLock,getOpenOptions]]]
Writes the content to the given file.
why this? Is there a test which verifies that the associated error type is still catchable?
@@ -108,7 +108,8 @@ public class HoodieWriteConfig extends DefaultHoodieConfig { } public int getWriteBufferLimitBytes() { - return Integer.parseInt(props.getProperty(WRITE_BUFFER_LIMIT_BYTES, DEFAULT_WRITE_BUFFER_LIMIT_BYTES)); + return Integer + .parseInt(props.getProperty(WRITE_BUFFER_LIMIT_BYTES, DEFAULT_WRITE_BUFFER_LIMIT_BYTES)); } public boolean shouldCombineBeforeInsert() {
[HoodieWriteConfig->[shouldUseTempFolderForCopyOnWrite->[shouldUseTempFolderForCopyOnWriteForMerge,shouldUseTempFolderForCopyOnWriteForCreate],Builder->[build->[HoodieWriteConfig,getBasePath,build]]]]
This method returns the maximum size of the write buffer in bytes.
Can we avoid additional formatting changes across the diff and trim it down to only real changes?
@@ -213,6 +213,12 @@ public abstract class HoodieTable<T extends HoodieRecordPayload> implements Seri public abstract HoodieWriteMetadata bulkInsertPrepped(JavaSparkContext jsc, String instantTime, JavaRDD<HoodieRecord<T>> preppedRecords, Option<BulkInsertPartitioner> bulkInsertPartitioner); + /** + * Logically delete all existing records and Insert a batch of new records into Hoodie table at the supplied instantTime. + */ + public abstract HoodieWriteMetadata insertOverwrite(JavaSparkContext jsc, String instantTime, + JavaRDD<HoodieRecord<T>> records); + public HoodieWriteConfig getConfig() { return config; }
[HoodieTable->[getSliceView->[getFileSystemView],requireSortedRecords->[getBaseFileFormat],getBaseFileOnlyView->[getFileSystemView],getHoodieView->[getFileSystemView],validateSchema->[getMetaClient],create->[create],validateUpsertSchema->[validateSchema],validateInsertSchema->[validateSchema],getActiveTimeline->[getActiveTimeline],reconcileAgainstMarkers->[deleteInvalidFilesByPartitions],getLogFileFormat->[getLogFileFormat],getHadoopConf->[getHadoopConf],getLogDataBlockFormat->[getBaseFileFormat],getBaseFileFormat->[getBaseFileFormat]]]
Bulk insert the given records into the table.
I think at the HoodieTable level, the API has to be about replacing file groups and not insertOverwrite (which can be limited to the WriteClient level). This way clustering can also use the same method, to build on top.
@@ -6,8 +6,8 @@ package submodule1 import ( "fmt" + "dash-named-schema/foo" "github.com/blang/semver" - "github.com/pulumi/pulumi/pkg/v3/codegen/internal/test/testdata/dash-named-schema/go/foo" "github.com/pulumi/pulumi/sdk/v3/go/pulumi" )
[Construct->[URN_,RegisterResource,Errorf],Printf,PkgVersion,RegisterResourceModule]
Construct creates a new resource object from a given type and name.
This now is guaranteed not to grab things off github.
@@ -324,7 +324,7 @@ namespace Microsoft.Xna.Framework.Content { //MonoGame try to load as a non-content file - assetName = TitleContainer.GetFilename(Path.Combine(RootDirectory, assetName)); + assetName = TitleContainer.GetAbsoluteFilename(Path.Combine(RootDirectory, assetName)); assetName = Normalize<T>(assetName);
[ContentManager->[ReloadAsset->[Dispose],T->[Dispose],ContentReader->[Dispose],Normalize->[Normalize],Unload->[Dispose],Dispose->[RemoveContentManager,Dispose],AddContentManager]]
This method tries to load an asset from the system. Catch all the objects in the asset.
This issue only affects PSM, so the use of absolute filename should be wrapped in #if PSM
@@ -1,4 +1,6 @@ # -*- coding: utf-8 -*- +from gevent import monkey +monkey.patch_all() from ui.cli import run
[run]
- - - - - - - - - - - - - - - - - -.
while you're at it, fix the import here: `from raiden.ui.cli `
@@ -23544,9 +23544,16 @@ void Player::UpdateObjectVisibility(bool forced, bool fromUpdate) void Player::UpdateVisibilityForPlayer(bool mapChange) { + // After added to map seer must be a player - there is no possibility to still have different seer (all charm auras must be already removed) + if (mapChange && m_seer != this) + { + m_seer = this; + } + Acore::VisibleNotifier notifierNoLarge(*this, mapChange, false); // visit only objects which are not large; default distance Cell::VisitAllObjects(m_seer, notifierNoLarge, GetSightRange() + VISIBILITY_INC_FOR_GOBJECTS); notifierNoLarge.SendToSelf(); + Acore::VisibleNotifier notifierLarge(*this, mapChange, true); // visit only large objects; maximum distance Cell::VisitAllObjects(m_seer, notifierLarge, GetSightRange()); notifierLarge.SendToSelf();
[No CFG could be retrieved]
Updates visibility of a specific object. ModifyMoney - Modify amount of money in the game.
TODO: add an error log here (I've tried doing it via GitHub but the file `Player.cpp` is too big and it does not work)
@@ -6,13 +6,13 @@ from django.http import HttpResponse from django.shortcuts import get_object_or_404, redirect from django.template.context_processors import csrf from django.template.response import TemplateResponse -from django.urls import reverse from django.utils.translation import pgettext_lazy from django_prices.templatetags.prices_i18n import gross from payments import PaymentStatus from prices import Price from satchless.item import InsufficientStock +from saleor.product.models import StockLocation from .filters import OrderFilter from .forms import ( AddressForm, AddVariantToDeliveryGroupForm, CancelGroupForm,
[refund_payment->[create_history_entry,pgettext_lazy,gross,redirect,refund,RefundPaymentForm,success,is_valid,TemplateResponse,get_object_or_404],cancel_order->[atomic,create_history_entry,pgettext_lazy,TemplateResponse,cancel_order,redirect,success,is_valid,CancelOrderForm,get_object_or_404],order_details->[list,all,any,get_total,get_captured_price,TemplateResponse,Price,get_total_price,exclude,last,select_related,get_object_or_404],ship_delivery_group->[all,pgettext_lazy,create_history_entry,TemplateResponse,ShipGroupForm,save,redirect,success,is_valid,atomic,select_related,get_object_or_404],orderline_split->[create_history_entry,pgettext_lazy,TemplateResponse,MoveLinesForm,redirect,get,move_lines,filter,success,is_valid,atomic,get_object_or_404],order_list->[OrderFilter,get,get_paginator_items,exists,prefetch_related,TemplateResponse],capture_payment->[CapturePaymentForm,create_history_entry,pgettext_lazy,get_total,gross,capture,redirect,success,is_valid,TemplateResponse,get_object_or_404],release_payment->[get_object_or_404,pgettext_lazy,create_history_entry,redirect,success,is_valid,TemplateResponse,release,ReleasePaymentForm],remove_order_voucher->[create_history_entry,pgettext_lazy,TemplateResponse,remove_voucher,RemoveVoucherForm,redirect,success,is_valid,atomic,get_object_or_404],add_variant_to_group->[all,pgettext_lazy,create_history_entry,TemplateResponse,save,redirect,get,AddVariantToDeliveryGroupForm,success,warning,is_valid,atomic,get_object_or_404],cancel_delivery_group->[CancelGroupForm,all,pgettext_lazy,create_history_entry,TemplateResponse,save,redirect,success,is_valid,atomic,get_object_or_404],order_invoice->[get_statics_absolute_url,create_invoice_pdf,prefetch_related,HttpResponse,get_object_or_404],address_view->[create_history_entry,pgettext_lazy,AddressForm,TemplateResponse,redirect,get,success,is_valid,save],order_packing_slip->[get_statics_absolute_url,create_packing_slip_pdf,prefetch_related,HttpResponse,get_object_or_404],orderline_change_stock->[TemplateResponse,ChangeStockForm,pgettext_lazy,success,is_valid,save,get_object_or_404],order_add_note->[create_history_entry,pgettext_lazy,TemplateResponse,send_confirmation_email,update,OrderNote,success,is_valid,save,OrderNoteForm,csrf,get_object_or_404],orderline_change_quantity->[ChangeQuantityForm,create_history_entry,pgettext_lazy,TemplateResponse,save,redirect,filter,success,is_valid,atomic,get_object_or_404],orderline_cancel->[create_history_entry,pgettext_lazy,TemplateResponse,cancel_line,redirect,filter,success,CancelOrderLineForm,is_valid,atomic,get_object_or_404],permission_required]
View a list of orders. A view that returns details about an order object.
This should be a relative import as we don't use absolutes outside of our tests suite.
@@ -34,9 +34,13 @@ func NewRenewer(ctx context.Context, ttl time.Duration, renewFunc Func) *Renewer done: make(chan struct{}), } go func() { - r.err = r.renewLoop(ctx) - r.cancel() - close(r.done) + defer close(r.done) + defer r.cancel() + err := r.renewLoop(ctx) + if errors.Is(err, ctx.Err()) { + err = nil + } + r.err = err }() return r }
[renewLoop->[Stop,renewFunc,NewTicker,WithTimeout,Is,Err,Done],Close->[cancel],cancel,renewLoop,WithCancel]
renew import creates a renewer from a given . renewFunc is the main entry point for renewing the lease. It is called by.
Could we make this `errors.Is(ctx.Err(), context.Canceled)`.
@@ -5,6 +5,7 @@ return [ [ 'type' => 'object', 'subtype' => 'file', + 'class' => '\ElggFile', 'searchable' => true, ], ],
[No CFG could be retrieved]
<?php SHOW ELEGABS >.
is this needed because it will be overriden otherwise?
@@ -14,7 +14,7 @@ class Snakemake(PythonPackage): version('3.11.2', sha256='f7a3b586bc2195f2dce4a4817b7ec828b6d2a0cff74a04e0f7566dcd923f9761') - depends_on('python@3.3:') + depends_on('python@3.3:3.6.999') depends_on('py-requests', type=('build', 'run')) depends_on('py-setuptools', type=('build', 'run')) depends_on('py-wrapt', type=('build', 'run'))
[Snakemake->[depends_on,version]]
version - 3. 11. 2.
I guess this works for now, but needs to be updated as soon as we'll get more recent versions of this package.
@@ -140,6 +140,9 @@ namespace Microsoft.NET.HostModel.AppHost { throw new Win32Exception(Marshal.GetLastWin32Error(), $"Could not set file permission {filePermissionOctal} for {appHostDestinationFilePath}."); } + + if (enableMacOSCodeSign && RuntimeInformation.IsOSPlatform(OSPlatform.OSX)) + CodeSign(appHostDestinationFilePath); } } catch (Exception ex)
[HostWriter->[IsBundle->[RetryOnIOError,SearchInFile,CreateFromFile,ReadInt64,CreateViewAccessor],CreateAppHost->[RetryOnWin32Error,Delete,SearchAndReplace,CreateViewAccessor,IsOSPlatform,Windows,GetBytes,CreateFromFile,IsSupportedOS,Create,ToInt32,None,RetryOnIOError,RewriteAppHost,WriteToStream,Read,GetLastWin32Error,Update,CopyOnWrite,chmod,IsPEImage,SetWindowsGraphicalUserInterfaceBit,Open,RemoveSignature,Length],SetAsBundle->[UtcNow,SetLastWriteTimeUtc,AdjustHeadersForBundle,RetryOnIOError,SearchAndReplace,GetBytes],GetBytes]]
Create an app host with the given parameters. This method writes the data from the source apphost file to the destination file.
Either here or in the SDK we should handle the case where we're on macOS but we're building for a different platform (Windows or Linux). In that case we should not attempt signing (as it will fail). I "think" this should be done in the SDK since here we don't have a good way to detect the target platform (we kind of rely on the SDK to tell us the right things).
@@ -306,10 +306,10 @@ class UsersController < ApplicationController def determine_follow_suggestions(current_user) return default_suggested_users if SiteConfig.prefer_manual_suggested_users? && default_suggested_users - recent_suggestions = Suggester::Users::Recent.new( + recent_suggestions = Users::SuggestRecent.call( current_user, attributes_to_select: INDEX_ATTRIBUTES_FOR_SERIALIZATION, - ).suggest + ) recent_suggestions.presence || default_suggested_users end
[UsersController->[add_org_admin->[update],remove_org_admin->[update],remove_identity->[update]]]
Returns the user that follows the given node index.
I changed this to follow our normal service pattern with a `call` class method.
@@ -1489,6 +1489,7 @@ void vga_device::crtc_reg_write(uint8_t index, uint8_t data) break; case 0x0c: case 0x0d: + //if(machine().first_screen()->vpos() < (vga.crtc.vert_retrace_start)) break; vga.crtc.start_addr_latch &= ~(0xff << (((index & 1)^1) * 8)); vga.crtc.start_addr_latch |= (data << (((index & 1)^1) * 8)); break;
[No CFG could be retrieved]
This function is used to set the values of the HSCR - C CRTC properties This function is used to read the values from the HVDC.
Please remove stuff like this before committing.
@@ -204,7 +204,7 @@ public final class ProxyConfiguration extends AbstractDescribableImpl<ProxyConfi SaveableListener.fireOnChange(this, config); } - public Object readResolve() { + private Object readResolve() { if (secretPassword == null) // backward compatibility : get scrambled password and store it encrypted secretPassword = Secret.fromString(Scrambler.descramble(password));
[ProxyConfiguration->[load->[getXmlFile],decorate->[decorate],getInputStream->[getPasswordAuthentication->[getUserName],createProxy,getUserName],createProxy->[getNoProxyHostPatterns,createProxy],getNoProxyHostPatterns->[getNoProxyHostPatterns],DescriptorImpl->[isNoProxyHost->[getNoProxyHostPatterns]],open->[getPasswordAuthentication->[getUserName],createProxy,getUserName]]]
Save the configuration to the file.
Sounds weird, but formally it's a compatibility breaking change. It could be a real case when you override the method in the inherited class, but this class is actually final. So I'm not sure it is important. I would just add Restricted
@@ -7300,7 +7300,12 @@ void Sedp::match_continue(const GUID_t& writer, const GUID_t& reader) } else { reader_type_name = dsi->second.get_type_name(); } - consistent = writer_type_name == reader_type_name; + if (writer_type_name == "" || reader_type_name == "") { + // force consistency with replayer or recorder + consistent = true; + } else { + consistent = writer_type_name == reader_type_name; + } } }
[No CFG could be retrieved]
Check if the data types of the topics are consistent. Get the object from the DCPS which could be used to construct a new.
I think this makes sense when 1 peer is a recorder because it can receive any type from the other peer. But if 1 peer is a replayer, then I'm not sure if the other peer would consider its type to be consistent with the type that is sent by the replayer, which can be any type. If the other peer is just a regular reader (not a recorder), then it doesn't seem right to me.
@@ -1,3 +1,8 @@ +<% if @current_user.ta? %> + <% manage_marking_schemes = GraderPermissions.find_by(user_id: @current_user.id).manage_marking_schemes %> + <% download_grades_report = GraderPermissions.find_by(user_id: @current_user.id).download_grades_report %> +<% end %> + <ul class='main'> <li id='logo'> <a href='<%= root_url %>' id='logo-img'></a>
[No CFG could be retrieved]
This method shows the list of all items in the nag object. Displays a link to the user s group or results page if the user is a admin or.
use policies here and below instead of setting variables in a view
@@ -478,7 +478,11 @@ void lmdb_resized(MDB_env *env) mdb_env_info(env, &mei); uint64_t old = mei.me_mapsize; + if (isactive) + mdb_txn_safe::increment_txns(-1); mdb_txn_safe::wait_no_active_txns(); + if (isactive) + mdb_txn_safe::increment_txns(1); int result = mdb_env_set_mapsize(env, 0); if (result)
[No CFG could be retrieved]
private int m_txn_safe = 0 ; This function is called from the txn_safe library. It is called from the constructor of.
Just for the sake of nitpicking, instead of an ugly if, something like this can be done: `mdb_txn_safe::increment_txns(-isactive);` `mdb_txn_safe::increment_txns(isactive);` Or to ensure that no bad value is passed, since bools and ints play nice: `mdb_txn_safe::increment_txns(-!!isactive);` `mdb_txn_safe::increment_txns(!!isactive);` If you don't like any of this, a ternary operator still looks better since the expression is very small: `mdb_txn_safe::increment_txns(isactive ? -1 : 0);` `mdb_txn_safe::increment_txns(isactive ? 1 : 0);`
@@ -305,8 +305,10 @@ int EVP_Cipher(EVP_CIPHER_CTX *ctx, unsigned char *out, return ctx->cipher->ccipher(ctx->provctx, out, &outl, inl + (blocksize == 1 ? 0 : blocksize), - in, (size_t)inl); - return 0; + in, (size_t)inl) + ? (int)outl + : -1; + return -1; } return ctx->cipher->do_cipher(ctx, out, in, inl);
[EVP_Cipher->[EVP_CIPHER_CTX_block_size],EVP_CIPHER_CTX_block_size->[EVP_CIPHER_block_size],EVP_CIPHER_CTX_iv_length->[EVP_CIPHER_iv_length,EVP_CIPHER_flags],EVP_MD_meth_dup->[EVP_MD_meth_new],EVP_CIPHER_mode->[EVP_CIPHER_flags],char->[EVP_CIPHER_nid],EVP_CIPHER_is_a->[EVP_CIPHER_nid]]
EVP_Cipher checks if the given buffer is encrypted or not.
Note that pre-3.0 `EVP_Cipher()` doesn't check if `ctx->cipher->do_cipher` is NULL or not. We should probably add an error message here...
@@ -74,7 +74,7 @@ class MechanicalSolver(PythonSolver): "residual_absolute_tolerance": 1.0e-9, "max_iteration": 10, "linear_solver_settings":{ - "solver_type": "SuperLUSolver", + "solver_type": "SkylineLUFactorizationSolver", "max_iteration": 500, "tolerance": 1e-9, "scaling": false,
[MechanicalSolver->[_create_convergence_criterion->[_get_convergence_criterion_settings],_create_builder_and_solver->[GetComputingModelPart,get_linear_solver],_execute_after_reading->[import_constitutive_laws],Initialize->[Initialize],_create_line_search_strategy->[GetComputingModelPart,get_solution_scheme,get_builder_and_solver,get_linear_solver,get_convergence_criterion],InitializeSolutionStep->[Initialize],get_mechanical_solution_strategy->[GetComputingModelPart],get_builder_and_solver->[GetComputingModelPart],_create_newton_raphson_strategy->[GetComputingModelPart,get_solution_scheme,get_builder_and_solver,get_linear_solver,get_convergence_criterion],_set_and_fill_buffer->[GetMinimumBufferSize],_create_linear_strategy->[GetComputingModelPart,get_builder_and_solver,get_solution_scheme,get_linear_solver]]]
Initialize the MechanicalSolver with a model and custom settings. Missing key in the model. Creates a KratosMultiphysics model part.
uff, your correction is correct, however this is fantastically slow. I would put a bit more of sofistication here, for example to verify that if a solver is not provided in the settings, than we verify if PastixSolver is availble and we use it as default, if that is not available than the Eigen direct solver, otherwise superlu ... and so on and so forth...
@@ -495,6 +495,9 @@ describes.realWin('amp-video', { */ function getVideoWithBlur(addPlaceholder, addBlurClass) { const v = doc.createElement('amp-video'); + v.setAttribute('layout', 'fixed'); + v.setAttribute('width', '300px'); + v.setAttribute('height', '250px'); const img = doc.createElement('img'); if (addPlaceholder) { img.setAttribute('placeholder', '');
[No CFG could be retrieved]
Creates an amp - video with an image child that could potentially potentially be a blurred image Get video with blur.
`layout` is added to depress the failures of "contain not supported"
@@ -223,11 +223,12 @@ public class CommonPanacheQueryImpl<Entity> { return PanacheJpaUtil.getCountQuery(selectQuery); } - @SuppressWarnings("unchecked") + @SuppressWarnings({ "unchecked" }) public <T extends Entity> List<T> list() { Query jpaQuery = createQuery(); try (NonThrowingCloseable c = applyFilters()) { - return jpaQuery.getResultList(); + List<T> results = jpaQuery.getResultList(); + return mapRows(results); } }
[CommonPanacheQueryImpl->[previousPage->[page],nextPage->[page],firstResultOptional->[firstResult],createBaseQuery->[createQuery],lastPage->[page],firstPage->[page],page->[page]]]
countQuery - count query.
why does this not check `rowMapper` but `stream()` does?
@@ -34,7 +34,7 @@ final class CodestartProjectGeneration { log.debug("processed shared-data: %s" + data); - final Codestart projectCodestart = projectDefinition.getRequiredCodestart(CodestartType.PROJECT); + projectDefinition.getRequiredCodestart(CodestartType.PROJECT); final List<CodestartFileStrategy> strategies = buildStrategies(mergeStrategies(projectDefinition));
[CodestartProjectGeneration->[generateProject->[CodestartProcessor,process,getCodestarts,info,getDepsData,getSharedData,of,checkTargetDir,log,joining,deepMerge,debug,writeFiles,buildStrategies,getLanguageName,getRequiredCodestart,collect,getCodestartProjectData,mergeStrategies],mergeStrategies->[deepMerge,map]]]
Generate a project with the specified codestart.
@glefloch why removing that line?
@@ -1,10 +1,14 @@ # frozen_string_literal: true class MyModuleRepositoriesController < ApplicationController + include ApplicationHelper + before_action :load_my_module - before_action :load_repository, except: %i(repositories_dropdown_list) + before_action :load_repository, except: %i(repositories_dropdown_list repositories_list_html) before_action :check_my_module_view_permissions - before_action :check_repository_view_permissions, except: %i(repositories_dropdown_list) + before_action :check_repository_view_permissions, except: %i(repositories_dropdown_list repositories_list_html) + + before_action :check_assign_repository_records_permissions, only: :update def index_dt @draw = params[:draw].to_i
[MyModuleRepositoriesController->[index_dt->[new,to_i,render,per,mappings,all_count],load_my_module->[find_by],check_repository_view_permissions->[can_read_repository?],full_view_table->[render_to_string,render],check_my_module_view_permissions->[experiment,can_read_experiment?],repositories_dropdown_list->[accessible_by_teams,render_to_string,render],load_repository->[find_by],before_action]]
Displays a list of all records of a .
Why empty line?
@@ -318,8 +318,13 @@ abstract class Publicize_Base { // Did this request happen via wp-admin? $from_web = 'post' == strtolower( $_SERVER['REQUEST_METHOD'] ) && isset( $_POST[$this->ADMIN_PAGE] ); - if ( ( $from_web || defined( 'POST_BY_EMAIL' ) ) && !empty( $_POST['wpas_title'] ) ) - update_post_meta( $post_id, $this->POST_MESS, trim( stripslashes( $_POST['wpas_title'] ) ) ); + if ( ( $from_web || defined( 'POST_BY_EMAIL' ) ) && isset( $_POST['wpas_title'] ) ) { + if ( empty( $_POST['wpas_title'] ) ) { + delete_post_meta( $post_id, $this->POST_MESS ); + } else { + update_post_meta( $post_id, $this->POST_MESS, trim( stripslashes( $_POST['wpas_title'] ) ) ); + } + } // change current user to provide context for get_services() if we're running during cron if ( defined( 'DOING_CRON' ) && DOING_CRON ) {
[Publicize_Base->[test_publicize_conns->[get_services,get_connection_id,test_connection],save_meta->[get_services],is_enabled->[blog_id,get_connections,user_id],get_profile_link->[get_connection_meta],get_display_name->[get_connection_meta],show_options_popup->[get_connection_meta]]]
Save meta data for a post This function is used to handle the post - related stuff. Function to handle post meta This function is called after all the meta data has been written and before the post is processed.
I know it's an edge case, but do you think it would make sense to do `trim` and `stripslashes` before checking if it's empty?
@@ -352,7 +352,7 @@ def lcmv(evoked, forward, noise_cov, data_cov, reg=0.01, label=None, @verbose -def lcmv_epochs(epochs, forward, noise_cov, data_cov, reg=0.01, label=None, +def lcmv_epochs(epochs, forward, noise_cov, data_cov, reg=0.1, label=None, pick_ori=None, return_generator=False, picks=None, rank=None, verbose=None): """Linearly Constrained Minimum Variance (LCMV) beamformer.
[_lcmv_source_power->[_prepare_beamformer_input],lcmv_raw->[_setup_picks,_apply_lcmv],lcmv->[_setup_picks,_apply_lcmv],tf_lcmv->[_setup_picks,_lcmv_source_power],lcmv_epochs->[_setup_picks,_apply_lcmv]]
Linearly Constrained Minimum Variance on a single trial. Returns a generator of the non - linearly constrained minimum - variance non - spatial features of.
@Eric89GXL here you use 0.1 not 1.
@@ -2710,6 +2710,7 @@ class SectionEditingResourceTests(UserTestCase, WikiTestCase): <dd>Type: <em>integer</em></dd> <dd>Przykłady 例 예제 示例</dd> </dl> + <p><iframe></iframe></p> <div class="noinclude"> <p>{{ languages( { &quot;ja&quot;: &quot;ja/XUL/Attribute/maxlength&quot; } ) }}</p> </div>
[DocumentSEOTests->[test_seo_title->[_make_doc],test_seo_script->[make_page_and_compare_seo]],ViewTests->[test_children_view->[_make_doc,_depth_test]],DocumentEditingTests->[test_translation_spam_ajax->[test_edit_spam_ajax],test_translation_midair_collission_ajax->[test_edit_midair_collisions],test_email_for_first_edits->[_check_message_for_headers],test_multiple_translation_edits_ajax->[test_multiple_edits_ajax],test_discard_location->[_create_doc],test_edit_midair_collisions_ajax->[test_edit_midair_collisions],test_translation_midair_collission->[test_edit_midair_collisions]]]
Test that raw include option is correct.
I was going to give feedback that the ``<iframe>`` test should be in it's own test, but 1) I don't want to make you convert to ``pytest`` as well, and 2) pyquery will return ``None`` if you just serialize an empty ``<iframe>``.
@@ -637,6 +637,11 @@ def plot_evoked_topo(evoked, layout=None, layout_scale=0.945, color=None, See matplotlib documentation for more details. axes : instance of matplotlib Axes | None Axes to plot into. If None, axes will be created. + dark_background : bool + Plot on dark background (black) or not (white). + It will be set to False by default in v0.16 + + .. versionadded:: 0.15.0 show : bool Show figure if True.
[plot_evoked_joint->[plot_evoked_joint,_connection_line,_plot_evoked],plot_evoked_image->[_plot_evoked],_handle_spatial_colors->[_plot_legend],_plot_lines->[_rgb],plot_compare_evokeds->[_truncate_yaxis,_ci,plot_compare_evokeds,_setup_styles],_plot_evoked_white->[whitened_gfp],plot_evoked->[_plot_evoked]]
Plots a 2D topography of evoked responses. Plots the evoked responses at sensor locations .
I would rather have a generic `background_color` argument (or whatever we call it elsewhere). Then to decide the font color you just use the `np.mean(matplotlib.colors.colorConverter.to_rgb(background_color)) > 0.5` to decide whether to use black/white font
@@ -430,10 +430,10 @@ static int ipc_stream_trigger(uint32_t header) cmd = COMP_TRIGGER_STOP; break; case SOF_IPC_STREAM_TRIG_PAUSE: - cmd = COMP_TRIGGER_PAUSE; + cmd = COMP_TRIGGER_STOP; break; case SOF_IPC_STREAM_TRIG_RELEASE: - cmd = COMP_TRIGGER_PRE_RELEASE; + cmd = COMP_TRIGGER_PRE_START; break; /* XRUN is special case- TODO */ case SOF_IPC_STREAM_TRIG_XRUN:
[No CFG could be retrieved]
get the message header and the message position and timestamp This function handles the trigger of a component.
any issues with this ?
@@ -324,7 +324,14 @@ module Discourse ActionView::Base.precompiled_asset_checker = -> logical_path do default_checker[logical_path] || - %w{qunit.js qunit.css test_helper.css test_helper.js wizard/test/test_helper.js}.include?(logical_path) + %w{qunit.js + qunit.css + test_helper.css + test_helper.js + wizard/test/test_helper.js + }.include?(logical_path) || + logical_path =~ /\/node_modules/ || + logical_path =~ /\/dist/ end end end
[Application->[config->[production?,database_config],pbkdf2_algorithm,new,register_mime_type,ember_location,register_postprocessor,paths,activate_plugins!,ignore,after_initialize,autoloader,load_path,lambda,pbkdf2_iterations,insert_after,env,register_transformer,root,time_zone,plugin_initialization_guard,generators,autoload_paths,require,templates_root,initializer,clear_active_connections!,to_s,enabled,mode,start_with?,relative_url_root,skip_minification,test?,include?,new_redis_store,schema_format,swap,precompiled_asset_checker,each,encoding,match,present?,development?,try,delete,cache_store,image_optim,rack_cache,test_framework,public_path,on_load,variant,filter_parameters,exceptions_app,store,handlebars_location,version,precompile,extname,raw_template_namespace,require_dependency],test?,configure!,match?,concat,groups,skip_db,require_relative,skip_redis,require,load_defaults,cdn_url,production?,present?,exit,puts,development?,expand_path]
Initialize a new instance of the object.
I don't understand what is this file doing in this commit? how is this related?
@@ -36,6 +36,7 @@ namespace Dynamo.Controls { public object Convert(object value, Type targetType, object parameter, CultureInfo culture) { + if (value == null) return ""; string incomingString = value as string; return incomingString.Split(new[] { '\r', '\n' }, 2)[0].Trim(); }
[TransformOriginStatConverter->[Convert->[ConverterMessageTransformOrigin,Y,X,Format]],ToolTipAllLinesButFirst->[Convert->[Trim]],WorkspaceTypeConverter->[Convert->[GetType]],AttachmentToPathConverter->[Convert->[Left,Top,Bottom,AttachmentSide,IndexOf,Substring,Right,Equals]],LinterIssueCountToVisibilityConverter->[Convert->[Collapsed,Visible]],DependencyListToStringConverter->[Convert->[Length,IndexOf,Empty,Remove],ToString],StringDisplay->[Convert->[ToString],ConvertBack->[ToString]],ConnectionStateToColorConverter->[Convert->[ExecutionPreview,None,Selection]],InOutParamTypeConverter->[Convert->[Concat,IsNullOrEmpty,Equals],NoneString],NumberFormatConverter->[Convert->[DynamoViewSettingMenuNumber00000,DynamoViewSettingMenuNumber0,DynamoViewSettingMenuNumber000,DynamoViewSettingMenuNumber0000,DynamoViewSettingMenuNumber00]],TabSizeConverter->[Convert->[Count,MinTabsBeforeClipping,TabControlMenuWidth,ActualWidth,TabDefaultWidth]],FilePathDisplayConverter->[ShortenNestedFilePath->[Root,Join,Name,FullName,IsNullOrEmpty,Insert,Parent,GetFileName,GetDirectoryName,Add],Convert->[IsNullOrEmpty,ShortenNestedFilePath,ToString,FilePathConverterNoFileSelected,Length],ConvertBack->[ToString]],PackageDownloadStateToStringConverter->[Convert->[PackageDownloadStateError,PackageStateUnknown,Error,Installed,Installing,PackageDownloadStateDownloaded,PackageDownloadStateStarting,Uninitialized,Downloading,PackageDownloadStateInstalled,Downloaded,State,PackageDownloadStateInstalling,PackageDownloadStateDownloading]],SearchResultsToVisibilityConverter->[Convert->[Collapsed,Visible,IsNullOrEmpty]],LacingToTooltipConverter->[Convert->[Shortest,LacingAutoToolTip,Auto,LacingCrossProductToolTip,Disabled,LacingDisabledToolTip,LacingShortestToolTip,LacingLongestToolTip,CrossProduct,Longest]],BoolToVisibilityCollapsedConverter->[Convert->[Collapsed,Visible]],RadianToDegreesConverter->[Convert->[Any,PI,TryParse],ConvertBack->[Any,PI,TryParse]],ViewModeToVisibilityConverter->[Convert->[ToString,Collapsed,Visible]],PathToSaveStateConverter->[Convert->[IsNullOrEmpty]],ShowHideClassicNavigatorMenuItemConverter->[Convert->[HideClassicNodeLibrary,ShowClassicNodeLibrary]],RunPreviewToolTipConverter->[Convert->[ShowRunPreviewDisableToolTip,ShowRunPreviewEnableToolTip]],Watch3DBackgroundColorConverter->[Convert->[DynamoColorsAndBrushesDictionary]],ZoomStatConverter->[Convert->[ToString,ConverterMessageZoom,Format]],CompareToParameterConverter->[Convert->[ToString]],LibraryTreeItemsHostVisibilityConverter->[Convert->[Collapsed,Visible]],NonEmptyStringToCollapsedConverter->[Convert->[Collapsed,Visible,IsNullOrEmpty]],IntegerDisplay->[Convert->[ToString,InvariantCulture],ConvertBack->[Any,TryParse,MinValue,MaxValue,ToString,IsDigit,InvariantCulture,Length]],ZoomToVisibilityConverter->[Convert->[Visible,Message,WriteLine,InvariantCulture,ToDouble,Hidden]],IsUpdateAvailableToTextConverter->[Convert->[AboutWindowCannotGetVersion,AvailableVersion,AboutWindowUpToDate,IsUpdateAvailable,ToString]],GroupFontSizeToEditorEnabledConverter->[Convert->[WriteLine,Message,InvariantCulture,ToDouble]],PortTypeToGridColumnConverter->[Convert->[Input]],IsUpdateAvailableBrushConverter->[Convert->[DynamoColorsAndBrushesDictionary]],PortToAttachmentConverter->[Convert->[Input,Right,Left]],BoolToFullscreenWatchVisibilityConverter->[Convert->[Visible,Hidden]],AttachmentToRowColumnConverter->[Convert->[MessageFailedToAttachToRowColumn,Left,Top,AttachmentSide,Bottom,Right,Equals]],ConnectionStateToVisibilityCollapsedConverter->[Convert->[ExecutionPreview,None,Collapsed,Selection,Visible]],ElementTypeToBoolConverter->[Convert->[Any]],NullValueToCollapsedConverter->[Convert->[Collapsed,Visible]],PortCountToHeightConverter->[Convert->[Count,Max]],BoolToCanvasCursorConverter->[Convert->[AppStarting]],BrushColorToStringConverter->[Convert->[Replace,Substring]],LacingToVisibilityConverter->[Convert->[Disabled,Collapsed,Visible]],SearchHighlightMarginConverter->[Convert->[ActualHeight,CurrentCultureIgnoreCase,SearchText,ActualWidth,IndexOf,ComputeTextWidth,Substring,RegularTypeface,Length,Text],ComputeTextWidth->[CurrentUICulture,LeftToRight,Foreground,FontSize,Width]],NodeOriginalNameToMarginConverter->[Convert->[ToString]],PackageSearchStateToStringConverter->[Convert->[PackageSearchStateSyncingWithServer,PackageSearchState,PackageStateUnknown,PackageSearchStateNoResult,Syncing,PackageSearchStateSearching,NoResults,Searching,Results]],AutoLacingToVisibilityConverter->[Convert->[Collapsed,Auto,Visible]],MenuItemCheckConverter->[Convert->[WriteLine,Message,InvariantCulture,ToDouble]],ListHasItemsToBoolConverter->[Convert->[Count]],TreeViewVLineMarginConverter->[Convert->[Left,Top,Parse,ToString,Right]],MeasureConverter->[Convert->[ToString],ConvertBack->[ToString,Value,SetValueFromString]],InverseBooleanToVisibilityCollapsedConverter->[Convert->[Collapsed,Visible]],BinaryRadioButtonCheckedConverter->[Convert->[ToString,Parse,Equals],ConvertBack->[DoNothing,ToString,Parse,Equals]],PackageUploadStateToStringConverter->[Convert->[Copying,PackageStateUnknown,Error,PackageUploadStateCopying,PackageUploadStateUploading,Ready,PackageUploadStateUploaded,Compressing,Uploaded,PackageUploadStateCompressing,PackageUploadStateReady,PackageUploadStateError,State,Uploading]],ShowHideFullscreenWatchMenuItemConverter->[Convert->[DynamoViewViewMenuShowBackground3DPreview]],BoolToScrollBarVisibilityConverter->[Convert->[ToString,Auto,Hidden,Disabled]],FullyQualifiedNameToDisplayConverter->[Convert->[ReduceRowCount,Join,Count,TruncateRows,ToList,MaxLengthRowClassButtonTitle,Insert,LastIndexOf,MaxRowNumber,ToString,WrapText,Length,MaxLengthTooltipCode]],AnnotationTextConverter->[Convert->[ToString,Empty],ConvertBack->[ToString]],DoubleDisplay->[Convert->[ToString,InvariantCulture],ConvertBack->[ToString,Any,InvariantCulture,TryParse]],EmptyStringToCollapsedConverter->[Convert->[Collapsed,Visible,IsNullOrEmpty]],TreeViewHLineMarginConverter->[Convert->[ItemsControlFromItemContainer,GetParent,Left,Count,Name,Right,IndexFromContainer,TreeView]],ConnectionStateToBrushConverter->[Convert->[ExecutionPreview,None,Selection]],TreeViewLineMarginConverter->[Convert->[ItemsControlFromItemContainer,GetParent,Left,Name,Right]],RgbaStringToBrushConverter->[Convert->[Count,Parse,RemoveEmptyEntries,FromRgb,Split]],ToolTipFirstLineOnly->[Convert->[Trim]],WorkspaceContextMenuHeightConverter->[Convert->[InCanvasSearchTextBoxHeight]],RadioButtonCheckedConverter->[Convert->[Equals],ConvertBack->[DoNothing,Equals]],ConsoleHeightConverter->[ConvertBack->[Value]],PortTypeToClipConverter->[Convert->[Input]],PathToFileNameConverter->[Convert->[IsReadOnlyPath,TabFileNameReadOnlyPrefix,IsNullOrEmpty,GetFileName]],BoolToVisibilityConverter->[Convert->[ToString,Collapsed,Visible,Hidden]],PortTypeToTextAlignmentConverter->[Convert->[Input,Right,Left]],PrettyDateConverter->[Convert->[PrettyDate,UnknowDateFormat],PrettyDate->[CreateSpecificCulture,ToString,Parse]],FullCategoryNameToMarginConverter->[Convert->[CategoryDelimiterString,Count,IsNullOrEmpty]],IntToVisibilityConverter->[Convert->[Collapsed,Visible]],LacingToAbbreviationConverter->[Convert->[First,Shortest,Auto,Disabled,CrossProduct,Longest]],WarningLevelToColorConverter->[Convert->[Error,Tomato,Gold,Gray,Mild,Moderate]],GroupTitleVisibilityConverter->[Convert->[ToString,Collapsed,Visible]],TooltipLengthTruncater->[Convert->[LastIndexOf,Length,Remove]],EmptyDepStringToCollapsedConverter->[Convert->[Collapsed,IndexOf,Visible],ToString],InverseBoolToVisibilityConverter->[Convert->[Visible,Hidden]],PortNameToWidthConverter->[Convert->[ToString,NaN,IsNullOrEmpty]],TreeViewMarginCheck->[Convert->[Right]],ElementGroupToColorConverter->[Convert->[DynamoColorsAndBrushesDictionary,Create,Action,Query]],ElementTypeToShorthandConverter->[Convert->[Package,CustomNode,ElementTypeShorthandPackage,CustomDll,ElementTypeShorthandImportedDll,ElementTypeShorthandCategory]],ElementTypeToShortConverter->[Convert->[CustomNode,None,Empty,PackageTypeShortString,CustomNodeTypeShortString,ZeroTouchTypeShortString,BuiltIn,Packaged,ZeroTouch]],StringToBrushColorConverter->[Convert->[ConvertFrom]],SnapRegionMarginConverter->[Convert->[HasFlag,Output,Left,Input,Top,Bottom,PortType,Right]],ClassViewMarginConverter->[Convert->[DataContext,Left,IsClassButton,Count,WpfUtilities,Child,Children,Margin]],ExpandersBindingConverter->[Convert->[Equals,IsNullOrEmpty],ConvertBack->[Empty]],CurrentOffsetStatConverter->[Convert->[Y,X,Format,ConverterMessageCurrentOffset]],TreeViewLineConverter->[Convert->[ItemsControlFromItemContainer,IndexFromContainer,Count]],NestedContentMarginConverter->[Convert->[Right,Left]],EnumToBooleanConverter->[Convert->[GetType,IsDefined,UnsetValue,Parse,Equals],ConvertBack->[UnsetValue,Parse]],ZoomToBooleanConverter->[Convert->[ChangeType]],LibraryViewModeToBoolConverter->[Convert->[LibraryView]],NullToVisibiltyConverter->[Convert->[Visible,Hidden]],DoubleInputDisplay->[Convert->[Any,NumberFormat,TryParse,ToString,InvariantCulture],ConvertBack->[Any,InvariantCulture,TryParse]]]
Convert the given object to the target type.
Make sense, would you update to use `String.Empty` where applies, thanks!
@@ -171,7 +171,7 @@ func (b *BadgeState) UpdateWithGregor(ctx context.Context, gstate gregor.State) b.state.RevokedDevices = []keybase1.DeviceID{} b.state.NewTeams = nil b.state.DeletedTeams = nil - b.state.NewTeamAccessRequests = nil + b.state.NewTeamAccessRequests = 0 b.state.HomeTodoItems = 0 b.state.TeamsWithResetUsers = nil b.state.ResetState = keybase1.ResetState{}
[Export->[ApplyLocalChatState],ConversationBadge->[ConversationBadgeStr]]
UpdateWithGregor updates the badges in the state with the given gregor. State This function is called when a new message arrives from a gregor server. This function is called when a new device is received from the server.
how would you feel about renaming this to something like NewTeamAccessRequestsCount since almost all of the other things are actually what they hold
@@ -207,12 +207,8 @@ security: client-authentication-scheme: header <%_ } _%> resource: + <%_ if (applicationType === 'gateway' || applicationType === 'monolith') { _%> user-info-uri: http://localhost:9080/auth/realms/jhipster/protocol/openid-connect/userinfo - token-info-uri: http://localhost:9080/auth/realms/jhipster/protocol/openid-connect/token/introspect - prefer-token-info: false - <%_ if (applicationType === 'gateway' || applicationType === 'microservice') { _%> - jwk: - key-set-uri: http://localhost:9080/auth/realms/jhipster/protocol/openid-connect/certs <%_ } _%> <%_ } _%>
[No CFG could be retrieved]
Private methods for handling authentication tokens. Private method for showing the specific .
Additional note : client-authentication-scheme should be the same in gateway/microservice applicationTypes ( I don't know if Keycloak/Okta accept both scheme )
@@ -654,9 +654,6 @@ func (mc *MeshCatalog) buildOutboundPermissiveModePolicies(srcServices []service destServices = append(destServices, utils.K8sSvcToMeshSvc(k8sService)) } - // For permissive mode build an outbound policy to every service in the mesh except to itself - destServices = difference(destServices, srcServices) - for _, destService := range destServices { hostnames, err := mc.getServiceHostnames(destService, false) if err != nil {
[buildOutboundPermissiveModePolicies->[getServiceHostnames],routesFromRules->[getHTTPPathsPerRoute,getTrafficSpecName],listOutboundPoliciesForTrafficTargets->[buildOutboundPolicies],ListPoliciesForPermissiveMode->[buildOutboundPermissiveModePolicies,buildInboundPermissiveModePolicies],ListAllowedInboundServices->[getAllowedDirectionalServices],buildInboundPermissiveModePolicies->[getServiceHostnames],buildOutboundPolicies->[getServiceHostnames,getDestinationServicesFromTrafficTarget],getAllowedDirectionalServices->[ListTrafficPolicies],listInboundPoliciesFromTrafficTargets->[buildInboundPolicies],buildInboundPolicies->[routesFromRules,getServiceHostnames],getTrafficSpecName]
buildOutboundPermissiveModePolicies creates a list of outbound policy for each service in the.
previously in permissive mode we were preventing services from building routes to itself. However to keep consistency with SMI we will allow this going forward.
@@ -925,7 +925,7 @@ const serverFiles = { ] }, { - condition: generator => generator.applicationType === 'monolith' && generator.authenticationType !== 'oauth2', + condition: generator => !shouldSkipUserManagement(generator) && generator.authenticationType !== 'oauth2', path: SERVER_TEST_SRC_DIR, templates: [ // Create auth config test files
[No CFG could be retrieved]
The base directory where the test files are created. Missing package - level conditions.
This uses the same condition for generating `DomainUserDetailsServiceIntTest` as it does for generating `DomainUserDetailsService`
@@ -1320,6 +1320,9 @@ class SimpleRNNCell(DropoutRNNCellMixin, Layer): else: self._enable_caching_device = kwargs.pop('enable_caching_device', False) super(SimpleRNNCell, self).__init__(**kwargs) + if units < 0: + raise ValueError("Received a negative value for `units`, ", + "expected a positive value.") self.units = units self.activation = activations.get(activation) self.use_bias = use_bias
[SimpleRNN->[__init__->[SimpleRNNCell]],GRU->[__init__->[GRUCell]],SimpleRNNCell->[call->[get_recurrent_dropout_mask_for_cell,get_dropout_mask_for_cell]],LSTM->[__init__->[LSTMCell]],LSTMCell->[call->[get_dropout_mask_for_cell,get_recurrent_dropout_mask_for_cell,_compute_carry_and_output_fused,_compute_carry_and_output,activation],_compute_carry_and_output_fused->[recurrent_activation,activation],_compute_carry_and_output->[recurrent_activation,activation],build->[bias_initializer->[bias_initializer]]],_standardize_args->[to_list_or_none],_generate_dropout_mask->[dropped_inputs->[dropout]],_generate_zero_filled_state->[create_zeros],GRUCell->[call->[get_recurrent_dropout_mask_for_cell,get_dropout_mask_for_cell,activation]],StackedRNNCells->[build->[bias_initializer->[],build]],RNN->[reset_states->[get_initial_state],__init__->[StackedRNNCells],compute_output_shape->[_get_output_shape],_process_inputs->[get_initial_state],build->[bias_initializer->[],build,get_input_spec,get_step_input_shape]],PeepholeLSTMCell->[_compute_carry_and_output->[recurrent_activation,activation],_compute_carry_and_output_fused->[recurrent_activation,activation]]]
Initialize a SimpleRNNCell with a sequence of missing parameters.
Please use the message in core.py, which contains the invalid value: raise ValueError(f"Received an invalid value for `units`, expected f"a positive integer, got {units}.")
@@ -948,7 +948,6 @@ class UCF101TestCase(datasets_utils.VideoDatasetTestCase): other_annotations.remove(current_annotation) for name in other_annotations: self._create_annotation_file(root, name, other_videos) - return len(current_videos) def _annotation_file_name(self, fold, train):
[Flickr30kTestCase->[_create_annotations_file->[_create_captions]],MNISTTestCase->[inject_fake_data->[_create_binary_file]],STL10Tester->[test_download_preexisting->[mocked_dataset],test_folds->[mocked_dataset],test_transforms->[mocked_dataset],test_unlabeled->[mocked_dataset],test_not_found->[mocked_dataset],test_invalid_folds2->[mocked_dataset],test_splits->[mocked_dataset,generic_classification_dataset_test],mocked_dataset->[mocked_root],test_invalid_folds1->[mocked_dataset],test_repr_smoke->[mocked_dataset]],VOCSegmentationTestCase->[_create_annotation_file->[add_name->[add_child],add_bndbox->[add_child],add_name,add_child,add_bndbox],_create_annotation_files->[_create_annotation_file]],UCF101TestCase->[_create_videos->[file_name_fn],inject_fake_data->[_create_annotation_files],_create_annotation_files->[_create_annotation_file]],ImageFolderTestCase->[test_classes->[create_dataset]],CityScapesTestCase->[inject_fake_data->[make_image]],Flickr8kTestCase->[test_captions->[create_dataset],inject_fake_data->[_create_images]],QMNISTTestCase->[_labels_file->[_prefix],test_num_examples_test50k->[create_dataset]],DatasetFolderTestCase->[test_is_valid_file->[create_dataset],test_classes->[create_dataset]],SBDatasetTestCase->[inject_fake_data->[_create_split_files]],Tester->[test_places365->[generic_classification_dataset_test]],HMDB51TestCase->[_create_videos->[file_name_fn],inject_fake_data->[_create_videos]],CocoDetectionTestCase->[inject_fake_data->[_create_annotation_file]]]
Create annotation files for the given list of video files.
Please revert as this is unrelated.
@@ -56,6 +56,8 @@ import com.twosigma.beaker.cpp.utils.Extractor; import com.twosigma.beaker.cpp.autocomplete.CPP14Lexer; import com.twosigma.beaker.cpp.autocomplete.CPP14Parser; import com.twosigma.beaker.cpp.autocomplete.CPP14Listener; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; public class CppEvaluator { protected final String shellId;
[CppEvaluator->[exit->[cancelExecution],setShellOptions->[resetEnvironment],resetEnvironment->[killAllThreads],evaluate->[jobDescriptor],workerThread->[MyRunnable->[run->[createMainCaller]]]]]
Creates a new instance of CppEvaluator. Construct a jobDescriptor object from the given SimpleEvaluationObject codeToBeExecuted cellId and job.
please don't import if not used. this applies to many files not just this one.
@@ -279,7 +279,7 @@ class Grouping < ActiveRecord::Base end if self.inviter == user errors.add(:base, I18n.t('invite_student.fail.inviting_self', - :user_name => user.user_name)) + user_name: user.user_name)) m_logger.log("Student failed to invite '#{user.user_name}'. Tried to invite " + 'himself.', MarkusLogger::ERROR)
[Grouping->[deletable_by?->[is_valid?],update_repository_permissions->[is_valid?],write_repo_permissions?->[repository_external_commits_only?],revoke_repository_permissions_for_membership->[write_repo_permissions?],remove_member->[membership_status],remove_rejected->[membership_status],assign_tas_by_csv->[add_tas_by_user_name_array],past_due_date?->[section],marking_completed?->[has_submission?],revoke_repository_permissions_for_students->[write_repo_permissions?],can_invite?->[pending?],membership_status->[membership_status],revoke_repository_permissions->[write_repo_permissions?],decline_invitation->[membership_status],grant_repository_permissions->[write_repo_permissions?]]]
Checks if the given user can invite this student. Check if the user is already part of another group and if not add an error message.
Align the parameters of a method call if they span more than one line.
@@ -104,8 +104,16 @@ public class ZooKeeperStateServer extends ZooKeeperServerMain { embeddedZkServer.setMinSessionTimeout(config.getMinSessionTimeout()); embeddedZkServer.setMaxSessionTimeout(config.getMaxSessionTimeout()); + final InetSocketAddress port = config.getClientPortAddress(); + final InetSocketAddress secPort = config.getSecureClientPortAddress(); connectionFactory = ServerCnxnFactory.createFactory(); - connectionFactory.configure(config.getClientPortAddress(), config.getMaxClientCnxns()); + if (secPort != null) { + logger.info("Configuring secure connections for embedded server on port: " + secPort); + connectionFactory.configure(secPort, config.getMaxClientCnxns(), true); + } else { + logger.info("Configuring insecure connections for embedded server on port: " + port); + connectionFactory.configure(port, config.getMaxClientCnxns(), false); + } connectionFactory.startup(embeddedZkServer); } catch (InterruptedException e) { Thread.currentThread().interrupt();
[ZooKeeperStateServer->[start->[start],startDistributed->[start],create->[ZooKeeperStateServer],shutdown->[shutdown]]]
Start the standalone server.
Since the insecure client port and secure client port aren't mutually exclusive you should add the checks for each port independently and setup two connection factories with their own startup and shutdown.
@@ -39,11 +39,12 @@ import org.springframework.util.StringUtils; * @author Gary Russell */ @IntegrationManagedResource -public class NullChannel implements PollableChannel, MessageChannelMetrics, BeanNameAware, NamedComponent { +public class NullChannel implements PollableChannel, MessageChannelMetrics, ConfigurableMetricsAware, + BeanNameAware, NamedComponent { private final Log logger = LogFactory.getLog(this.getClass()); - private volatile ChannelSendMetrics channelMetrics = new ChannelSendMetrics("nullChannel"); + private volatile AbstractMessageChannelMetrics channelMetrics = new DefaultMessageChannelMetrics("nullChannel"); private volatile boolean countsEnabled;
[NullChannel->[getSendCountLong->[getSendCountLong],getErrorRate->[getErrorRate],receive->[receive],getSendErrorCount->[getSendErrorCount],getSendDuration->[getSendDuration],reset->[reset],getMeanSendDuration->[getMeanSendDuration],getMeanErrorRatio->[getMeanErrorRatio],getMeanErrorRate->[getMeanErrorRate],getStandardDeviationSendDuration->[getStandardDeviationSendDuration],getMaxSendDuration->[getMaxSendDuration],getSendErrorCountLong->[getSendErrorCountLong],getMinSendDuration->[getMinSendDuration],getTimeSinceLastSend->[getTimeSinceLastSend],getMeanSendRate->[getMeanSendRate],getSendRate->[getSendRate],send->[send],getSendCount->[getSendCount]]]
Creates a new instance of a NullChannel. Resets the metrics.
Generic type has been missed. Will be fixed on merge.
@@ -146,6 +146,8 @@ public class TestCommentsMigrator { // This file will be commented and then removed to simulate comments with removed parent protected DocumentModel fileToCommentAndRemove; + protected DocumentModel placeLessFileToComment; + protected DocumentModel proxyFileToComment; protected static void sleep(long millis) {
[TestCommentsMigrator->[migrateFromRelationToProperty->[runMigration],runMigrationStep->[runMigration,sleep],sleep->[sleep],testProbe->[runMigration],testProbeWithInConsistentComments->[runMigration],migrateFromPropertyToSecured->[runMigration]]]
Sleeps for the specified number of milliseconds.
placeless is a single word, not capital L.
@@ -82,8 +82,12 @@ class AccessControlManager implements AccessControlManagerInterface /** * {@inheritdoc} */ - public function getUserPermissions(SecurityCondition $securityCondition, UserInterface $user) + public function getUserPermissions(SecurityCondition $securityCondition, $user) { + if (!$user instanceof UserInterface) { + return; + } + $objectPermissions = $this->getUserObjectPermission($securityCondition, $user); $checkPermissionType = empty($objectPermissions);
[AccessControlManager->[getUserObjectPermission->[getPermissions],cumulatePermissions->[mapPermissions],getPermissions->[getPermissions],getUserRoleSecurityContextPermission->[getPermissions],setPermissions->[setPermissions],restrictPermissions->[mapPermissions]]]
Returns the permissions for the given user.
You can leave the type hint if you make it nullable, can't you? Like `?UserInterface`.