patch
stringlengths
18
160k
callgraph
stringlengths
4
179k
summary
stringlengths
4
947
msg
stringlengths
6
3.42k
@@ -74,7 +74,8 @@ class Account(object): # pylint: disable=too-few-public-methods self.meta.creation_dt), self.meta.creation_host, self.id[:4]) def __repr__(self): - return "<{0}({1})>".format(self.__class__.__name__, self.id) + return "<{0}({1}, {2}, {3})>".format( + self.__class__.__name__, self.id, self.regr, self.meta) def __eq__(self, other): return (isinstance(other, self.__class__) and
[AccountFileStorage->[delete->[_account_dir_path],find_all->[load],load->[_account_dir_path,_key_path,Account,_metadata_path,_regr_path],_save->[_account_dir_path,_key_path,_metadata_path,_regr_path]],Account->[__init__->[Meta]]]
Return a string representation of the object.
nit: As you probably already know, normally it's good practice to make the `repr` of an object be as close to the constructor call necessary to recreate that object as possible. While we don't want to do exactly that here because the `repr` would include the private key, we can keep the order the same as the constructor by swapping `id` and `regr` here.
@@ -51,8 +51,8 @@ class Article < ApplicationRecord validates :body_markdown, length: { minimum: 0, allow_nil: false }, uniqueness: { scope: %i[user_id title] } validates :cached_tag_list, length: { maximum: 126 } - validates :canonical_url, url: { allow_blank: true, no_local: true, - schemes: %w[https http] }, uniqueness: { allow_blank: true } + validates :canonical_url, uniqueness: { allow_nil: true } + validates :canonical_url, url: { allow_blank: true, no_local: true, schemes: %w[https http] } validates :feed_source_url, uniqueness: { allow_blank: true } validates :main_image, url: { allow_blank: true, schemes: %w[https http] } validates :main_image_background_hex_color, format: /\A#([A-Fa-f0-9]{6}|[A-Fa-f0-9]{3})\z/
[Article->[username->[username],evaluate_front_matter->[set_tag_list],update_notifications->[update_notifications],readable_edit_date->[edited?]]]
The Notifiable model. video_source_url is the source url of the video.
split the validators in 2 lines and replaced `allow_blank` with `allow_nil` to avoid the empty string ;)
@@ -724,7 +724,7 @@ module AutomatedTestsHelper # Run script test_harness_name = File.basename(test_harness_path) - stdout, stderr, status = Open3.capture3("ssh #{server} \"cd #{test_box_path}; ruby #{test_harness_name} #{arg_list}\"") + stdout, stderr, status = Open3.capture3("cd #{test_box_path}; ruby #{test_harness_name} #{arg_list}") if !(status.success?) return [stderr, stdout, false]
[export_repository->[export,markus_config_automated_tests_repository,exist?,rm_rf,message,repo_name,join,mkdir],create_test_repo->[short_identifier,markus_config_automated_tests_repository,exist?,join,mkdir],launch_test->[capture3,markus_ate_test_run_directory,scripts_to_run,basename,to_i,write,repository_folder,markus_config_automated_tests_repository,gsub,halts_testing,success?,join,each,markus_ate_test_runner_script_name,mkdir_p],process_test_form->[nil?,enable_test,size,original_filename,unlimited_tokens,include?,raise,find_by_id,test_scripts_attributes,tokens_per_day,test_support_files_attributes,t,each,respond_to?,each_key],delete_repo->[rm_rf,exists?],export_group_repo->[exists?,export,markus_config_automated_tests_repository,mkdir,delete_repo],can_run_test?->[admin?,tokens,token,student?,include?,raise,ta?,decrease_tokens,t],export_configuration_files->[short_identifier,write,group_name,markus_config_automated_tests_repository,touch,join,close,open,api_key],perform->[markus_ate_test_runner_script_name,group,markus_ate_test_run_directory,find,instance,process_result,launch_test,markus_config_automated_tests_repository,raise,log,repo_name,join,repository_folder,assignment],process_result->[nil?,repo,get_latest_revision,to_i,create,is_a?,to_json,each,revision_number,find_by,first,from_xml,id],request_a_test_run->[group,find,markus_config_automated_tests_repository,repo_name,join,scripts_to_run,assignment,export_group_repo,async_test_request],files_available?->[short_identifier,empty?,exists?,markus_config_automated_tests_repository,raise,where,join,t,repository_folder,length,id,entries],run_ant_file->[short_identifier,system,each_line,new,create,cd,id,close,raise,pwd,now,join,exist?,parse_test_output,exitstatus,t,open,l],delete_test_repo->[markus_config_automated_tests_repository,rm_rf,repo_name,join,exist?],has_permission?->[nil?,admin?,tokens,token,unlimited_tokens,student?,include?,raise,ta?,decrease_tokens,t],add_parser_file_link->[new,render,link_to_function,escape_javascript,t],add_test_support_file_link->[new,render,link_to_function,escape_javascript,t],parse_test_output->[system,each_line,new,cd,find_by_filetype,pwd,join,close,open],scripts_to_run->[new,run_on_submission,insert,sort_by!,run_on_request,where,each,length,id],async_test_request->[enqueue,files_available?],add_lib_file_link->[new,render,link_to_function,escape_javascript,t],add_test_file_link->[new,render,link_to_function,escape_javascript,t],copy_ant_files->[short_identifier,mkdir,cd,student?,mv,markus_config_automated_tests_repository,filename,raise,pwd,glob,cp,join,exist?,each,t,cp_r],create_ant_test_files->[short_identifier,new,empty?,filetype,markus_config_automated_tests_repository,filename,join,reload,save,assignment,makedirs],require]
Launches a single test of the n - node node. capture_nack - capture nack.
Line is too long. [105/80]
@@ -127,6 +127,11 @@ class Conv(Layer): if filters is not None and not isinstance(filters, int): filters = int(filters) self.filters = filters + self.groups = groups + if filters and filters % self.groups != 0: + raise ValueError( + 'The number of filters is not divisible by the number groups. ' + '{} % {} = {}'.format(filters, groups, filters % groups)) self.kernel_size = conv_utils.normalize_tuple( kernel_size, rank, 'kernel_size') if not all(self.kernel_size):
[SeparableConv->[build->[_get_channel_axis]],Conv->[_get_input_channel->[_get_channel_axis]],SeparableConv1D->[call->[_compute_causal_padding]],Conv3DTranspose->[call->[compute_output_shape],build->[_get_channel_axis]],DepthwiseConv2D->[build->[_get_channel_axis]],Conv2DTranspose->[call->[compute_output_shape],build->[_get_channel_axis]]]
Initialize a new Conv1D object with a single . MissingNodeException - MissingNodeException.
"The number of filters must be evenly divisible by the number of groups."
@@ -427,8 +427,9 @@ namespace Content.Server.GameObjects.Components.Buckle { drawDepth = BuckledTo.SpriteComponent.DrawDepth - 1; } + - return new BuckleComponentState(Buckled, drawDepth); + return new BuckleComponentState(Buckled, drawDepth, EntityBuckledTo, DontCollide); } bool IInteractHand.InteractHand(InteractHandEventArgs eventArgs)
[BuckleComponent->[TryBuckle->[CanBuckle,ReAttach,UpdateBuckleStatus],TryUnbuckle->[UpdateBuckleStatus],InteractHand->[TryUnbuckle],OnRemove->[OnRemove,TryUnbuckle,UpdateBuckleStatus],ExposeData->[ExposeData],ToggleBuckle->[TryBuckle,TryUnbuckle],BuckleVerb->[Activate->[TryUnbuckle]],Startup->[UpdateBuckleStatus,Startup]]]
override GetComponentState method.
You shouldn't spam WakeBody like this if the body ain't moving because it's expensive for physics.
@@ -75,6 +75,7 @@ module.exports = LanguagesGenerator.extend({ this.websocket = this.config.get('websocket') === 'no' ? false : this.config.get('websocket'); this.databaseType = this.config.get('databaseType'); this.searchEngine = this.config.get('searchEngine') === 'no' ? false : this.config.get('searchEngine'); + this.messageBroker = this.config.get('messageBroker') === 'no' ? false : this.config.get('messageBroker'); this.env.options.appPath = this.config.get('appPath') || constants.CLIENT_MAIN_SRC_DIR; this.enableTranslation = this.config.get('enableTranslation'); this.enableSocialSignIn = this.config.get('enableSocialSignIn');
[No CFG could be retrieved]
The main function that is called when the user presses enter to enter a list of supported Get all supported languages.
Not required as this is a new option and doesnt need backward compatibility
@@ -98,9 +98,10 @@ public class ExplodedGraphWalkerTest { } private void reportIssuesFor(JavaFileScannerContext context, JavaCheck check) { - Multimap<Tree, String> issues = ((DefaultJavaFileScannerContext) context).getSEIssues((Class<? extends SECheck>) check.getClass()); - for (Map.Entry<Tree, String> issue : issues.entries()) { - context.reportIssue(check, issue.getKey(), issue.getValue()); + Multimap<Tree, DefaultJavaFileScannerContext.SEIssue> issues = ((DefaultJavaFileScannerContext) context).getSEIssues((Class<? extends SECheck>) check.getClass()); + for (Map.Entry<Tree, DefaultJavaFileScannerContext.SEIssue> issue : issues.entries()) { + DefaultJavaFileScannerContext.SEIssue seIssue = issue.getValue(); + context.reportIssue(this, seIssue.getTree(), seIssue.getMessage(), seIssue.getSecondary(), null); } } }
[ExplodedGraphWalkerTest->[reproducer->[IssueVisitor,verify],IssueVisitor->[scanFile->[LocksNotUnlockedCheck,UnclosedResourcesCheck,NullDereferenceCheck,reportIssuesFor,ConditionAlwaysTrueOrFalseCheck],reportIssuesFor->[getValue,reportIssue,getSEIssues,getKey,getClass,entries]],test2->[visitNode->[ExplodedGraphWalker,accept,fail],SymbolicExecutionVisitor,verifyNoIssue],test_cleanup_state->[visitNode->[ExplodedGraphWalker,accept],SymbolicExecutionVisitor,verifyNoIssue,isGreaterThan,isPositive],test->[IssueVisitor,verify]]]
Reports all issues for the given check.
The test fails to show what the end user will see on the screen. So, the effect of the change remains quite theoretical.
@@ -12,7 +12,7 @@ // See the License for the specific language governing permissions and // limitations under the License. -//go:generate ../../../../hack/generate-controller-registration.sh provider-local . ../../../../VERSION ../../../../example/provider-local/base/controller-registration.yaml DNSProvider:local DNSRecord:local ControlPlane:local Infrastructure:local Network:local OperatingSystemConfig:local Worker:local +//go:generate ../../../../hack/generate-controller-registration.sh provider-local . '' ../../../../example/provider-local/base/controller-registration.yaml DNSProvider:local DNSRecord:local ControlPlane:local Infrastructure:local Network:local OperatingSystemConfig:local Worker:local //go:generate cp ../../../../example/provider-local/base/controller-registration.yaml ../../../../charts/gardener/provider-local/registration/templates/controller-registration.yaml //go:generate sh -c "sed -i 's/ image:/{{ toYaml .Values.values | indent 4 }}/g' ../../../../charts/gardener/provider-local/registration/templates/controller-registration.yaml" //go:generate sh -c "sed -i 's/ tag: .*//g' ../../../../charts/gardener/provider-local/registration/templates/controller-registration.yaml"
[No CFG could be retrieved]
Package chart enables go to generate controller registration.
In general this looks odd. And error-prone - if there is a typo in the VERSION file path it will wrongly assume wrong version/tag, right? Anyways `eu.gcr.io/gardener-project/gardener/extensions/provider-local` does not seem to be build and pushed by Gardener CI/CD. WDYT about modifying the `generate-controller-registration.sh` to accept a `<version>` instead of a `<version-file>`? This will be a small breaking change for extensions (they will have to pass the VERSION file content instead of the VERSION file path). But I think this approach looks much better and cleaner. We would hard-code the version here to some stable one for this extension.
@@ -25,8 +25,8 @@ /* #define ENGINE_DEVCRYPTO_DEBUG */ -#ifdef CRYPTO_ALGORITHM_MIN -# define CHECK_BSD_STYLE_MACROS +#if CRYPTO_ALGORITHM_MIN < CRYPTO_ALGORITHM_MAX +#define CHECK_BSD_STYLE_MACROS #endif #define engine_devcrypto_id "devcrypto"
[No CFG could be retrieved]
ONE global file descriptor for all sessions. Get the default values of driver_info_st.
This should be indented by adding a space between `#` and `define`
@@ -284,14 +284,14 @@ void MultiTopicDataReaderBase::data_available(DDS::DataReader_ptr reader) gen.info_[i].instance_state); } } else { - ACE_ERROR((LM_ERROR, ACE_TEXT("(%P|%t) MultiTopicDataReaderBase::data_available:" - " failed to obtain DataReaderImpl."))); + ACE_ERROR((LM_ERROR, ACE_TEXT("(%P|%t) MultiTopicDataReaderBase::data_available:") + ACE_TEXT(" failed to obtain DataReaderImpl."))); } } } } catch (const std::runtime_error& e) { if (OpenDDS::DCPS::DCPS_debug_level) { - ACE_ERROR((LM_ERROR, "(%P|%t) MultiTopicDataReaderBase::data_available: " + ACE_ERROR((LM_ERROR, ACE_TEXT("(%P|%t) MultiTopicDataReaderBase::data_available: ") "%C", e.what())); } }
[No CFG could be retrieved]
missing data for the topic protected static final int MAX_DEPTH = 0 ;.
The wchar problem is still present on this line. I suggest making the %C part of line 294 instead of 295.
@@ -33,7 +33,8 @@ const ( // Standard locations for the secrets mounted in pods StandardMasterCaPath = "/var/run/secrets/kubernetes.io/serviceaccount/ca.crt" StandardTokenPath = "/var/run/secrets/kubernetes.io/serviceaccount/token" - StandardMasterUrl = "https://kubernetes.default.svc.cluster.local" + // TODO: use dnsDomain instead of cluster.local + StandardMasterUrl = "https://kubernetes.default.svc.cluster.local" ) const longPodDiagDescription = `
[BuildAndRunDiagnostics->[NewAggregate,buildPodDiagnostics,Errorf,Stack],buildPodDiagnostics->[List,Errorf],Complete->[List,NewLogger],RecommendedLoggerOptionFlags,LogEntry,Errors,Exit,Stack,Error,CanRun,Description,Errorf,NewString,Complete,Notice,CheckErr,Name,Summary,Logs,Warnings,Sprintf,Intersection,BuildAndRunDiagnostics,SetOutput,BindLoggerOptionFlags,List,EvalTemplate,Check,Flags]
NewCommandPodDiagnostics creates a new Command that will run on the given command. Command for running a single unit of work.
Or omit the cluster domain and rely on the DNS lookup
@@ -524,6 +524,12 @@ vdev_config_generate(spa_t *spa, vdev_t *vd, boolean_t getstats, fnvlist_add_uint64(nv, ZPOOL_CONFIG_VDEV_TOP_ZAP, vd->vdev_top_zap); } + + if (vd->vdev_resilver_deferred) { + ASSERT(vd->vdev_ops->vdev_op_leaf); + ASSERT(spa->spa_resilver_deferred); + fnvlist_add_boolean(nv, ZPOOL_CONFIG_RESILVER_DEFER); + } } if (getstats) {
[No CFG could be retrieved]
region ResourcePool. *** - The number of segments in the system that are not in the system.
can we also assert that the feature is active here?
@@ -320,9 +320,14 @@ class Jetpack_Sync_Sender { $module->reset_data(); } + // leave if the user was running having an older version of Jetpack (4.2) delete_option( self::SYNC_THROTTLE_OPTION_NAME ); delete_option( self::NEXT_SYNC_TIME_OPTION_NAME ); + foreach ( array( 'sync', 'full_sync' ) as $queue_name ) { + delete_option( self::NEXT_SYNC_TIME_OPTION_NAME . '_' . $queue_name ); + } + Jetpack_Sync_Settings::reset_data(); }
[Jetpack_Sync_Sender->[set_defaults->[set_upload_max_rows,set_sync_wait_time,set_max_dequeue_time,set_sync_wait_threshold,set_dequeue_max_bytes,set_upload_max_bytes],do_sync_and_set_delays->[set_next_sync_time,get_next_sync_time],uninstall->[reset_data],reset_data->[reset_data,reset_sync_queue]]]
reset data of all modules.
I think we can delete this line now \o/
@@ -1260,10 +1260,12 @@ public class ReferenceCountedOpenSslEngine extends SSLEngine implements Referenc // This is needed so we ensure close_notify etc is correctly send to the remote peer. // See https://github.com/netty/netty/issues/3900 if (SSL.bioLengthNonApplication(networkBIO) > 0) { - if (handshakeException == null && handshakeState != HandshakeState.FINISHED) { + if (pendingException == null) { // we seems to have data left that needs to be transferred and so the user needs // call wrap(...). Store the error so we can pick it up later. - handshakeException = new SSLHandshakeException(SSL.getErrorString(stackError)); + String message = SSL.getErrorString(stackError); + pendingException = handshakeState == HandshakeState.FINISHED ? + new SSLException(message) : new SSLHandshakeException(message); } // We need to clear all errors so we not pick up anything that was left on the stack on the next // operation. Note that shutdownWithError(...) will cleanup the stack as well so its only needed here.
[ReferenceCountedOpenSslEngine->[handshakeException->[shutdown],needPendingStatus->[isOutboundDone,isInboundDone],getSSLParameters->[getSSLParameters],retain->[retain],sslReadErrorResult->[shutdownWithError],getOcspResponse->[getOcspResponse],setKeyMaterial->[setKeyMaterial],getDelegatedTask->[run->[run]],wrap->[release,writePlaintextData,isBytesAvailableEnoughForWrap,wrap,resetSingleSrcBuffer,singleSrcBuffer],closeInbound->[shutdown],newResult->[shutdown,newResult],writeEncryptedData->[release],beginHandshake->[calculateMaxWrapOverhead],checkSniHostnameMatch->[checkSniHostnameMatch],calculateMaxWrapOverhead->[maxEncryptedPacketLength0],rejectRemoteInitiatedRenegotiation->[shutdown],release->[release],setSSLParameters->[setVerify,setSSLParameters],newResultMayFinishHandshake->[newResult],setVerify->[setVerify],handshake->[shutdownWithError,pendingStatus,handshakeException,checkEngineClosed],closeOutbound->[shutdown],unwrap->[readPlaintextData,release,sslPending0,newResultMayFinishHandshake,singleDstBuffer,newResult,unwrap,resetSingleDstBuffer,resetSingleSrcBuffer,writeEncryptedData,singleSrcBuffer],doSSLShutdown->[shutdown],getHandshakeStatus->[pendingStatus],readPlaintextData->[release],shutdownWithError->[shutdownWithError,shutdown],setOcspResponse->[setOcspResponse],toJavaCipherSuitePrefix->[isEmpty],isEndPointVerificationEnabled->[isEmpty],writePlaintextData->[release],refCnt->[refCnt],sslPending0->[sslPending],mayFinishHandshake->[handshake],touch->[touch],DefaultOpenSslSession->[handshakeFinished->[isDestroyed,calculateMaxWrapOverhead,toJavaCipherSuite],getPeerCertificateChain->[isEmpty],initPeerCerts->[isEmpty],getPeerPort->[getPeerPort],invalidate->[isDestroyed],getCreationTime->[isDestroyed],getProtocol->[isDestroyed],getPeerPrincipal->[getPeerCertificates],getLastAccessedTime->[getCreationTime],notifyUnbound->[newSSLSessionBindingEvent],selectApplicationProtocol->[selectApplicationProtocol],getPacketBufferSize->[maxEncryptedPacketLength],getValueNames->[isEmpty],putValue->[newSSLSessionBindingEvent],isValid->[isDestroyed],getPeerCertificates->[isEmpty],getPeerHost->[getPeerHost]],setClientAuth->[setVerify]]]
This method checks if there is a pending handshakeException and if so if so it will throw.
I hardly know the control flow here but since this is not just `handshakeException` now, do we need to also handle a case when `pendingException != null`?
@@ -430,7 +430,7 @@ module Engine simple_logo: '1830/PRR.alt', tokens: [60, 100, 60, 100, 60, 100, 60, 100, 60, 100], coordinates: 'K9', - color: '#32763f', + color: '#40b1b9', type: 'groupA', reservation_color: nil, },
[Game->[company_header->[id,find],after_buy_company->[find,corporation,buy_shares,after_par,set_par,president,each,price,abilities],optional_short_game->[include?],include_meta,register_colors,freeze,merge],require_relative]
A list of all possible components of a system - wide header that are defined in the section Color color = color of all possible tokens of a residue.
change to snake_case
@@ -64,7 +64,13 @@ module Search TERM_KEYS.map do |term_key, search_key| next unless @params.key? term_key - { terms: { search_key => Array.wrap(@params[term_key]) } } + values = Array.wrap(@params[term_key]) + + if params[:tag_boolean_mode] == "all" + values.map { |val| { term: { search_key => val } } } + else + { terms: { search_key => values } } + end end.compact end
[Reaction->[terms_keys_present?->[present?,detect],build_queries->[terms_keys_present?,query_keys_present?],initialize->[deep_symbolize_keys],terms_keys->[compact],freeze,attr_accessor]]
Returns a hash of missing terms or missing fields.
Ideally, I would like to add this to all search query builders but I will save that for a future refactor. Then in the view we can allow users to toggle between AND and OR options for searching multiple items.
@@ -293,6 +293,7 @@ def run_single_test(name: str, test: Any) -> Tuple[bool, bool]: test.run() except BaseException as e: if isinstance(e, KeyboardInterrupt): + test.tear_down() raise exc_type, exc_value, exc_traceback = sys.exc_info() test.tear_down() # FIX: check exceptions
[run_test_recursive->[run_test_recursive,cases],assert_equal->[good_repr,AssertionFailure],fail->[AssertionFailure],assert_raises->[assert_equal,AssertionFailure],assert_false->[AssertionFailure],run_single_test->[tear_down,run,set_up],assert_type->[AssertionFailure],match_pattern->[match_pattern],main->[ListSuite,add_suites_from_module],assert_not_equal->[good_repr,AssertionFailure],Suite->[init->[TestCase],skip->[SkipTestCaseException]],assert_true->[AssertionFailure],TestCase->[tear_down->[tear_down],set_up->[set_up]]]
Run a single test in a single thread.
How about let's move this line into a `finally` block? Then we only have to say `tear_down` in one place, and it's easier to see that it'll always be called.
@@ -53,7 +53,10 @@ public final class DefaultEventExecutorChooserFactory implements EventExecutorCh @Override public EventExecutor next() { - return executors[idx.getAndIncrement() & executors.length - 1]; + final AtomicInteger idx = this.idx; + final EventExecutor[] executors = this.executors; + final int executorsMask = executors.length - 1; + return executors[idx.getAndIncrement() & executorsMask]; } }
[DefaultEventExecutorChooserFactory->[DefaultEventExecutorChooserFactory]]
Returns the next event executor in the chain.
Is there a subtle performance reason for the additional verbosity here that I'm not seeing? :)
@@ -25,10 +25,14 @@ can predict accurately over time and on a second set of conditions. # License: BSD (3-clause) import numpy as np +import matplotlib.pyplot as plt import mne from mne.datasets import sample -from mne.decoding import GeneralizationAcrossTime +from mne.decoding.search_light import GeneralizationLight +from sklearn.pipeline import make_pipeline +from sklearn.preprocessing import StandardScaler +from sklearn.linear_model import LogisticRegression print(__doc__)
[,read_events,plot,dict,print,GeneralizationAcrossTime,unique,data_path,read_raw_fif,filter,score,in1d,Epochs,pick_types,fit]
This example runs the analysis of a single . Train the classifier on all the events in the raw data.
so we deprecate GeneralizationAcrossTime ? and if not the class GeneralizationAcrossTime should use the new GeneralizationLight code.
@@ -1426,6 +1426,14 @@ Sedp::update_topic_qos(const RepoId& topicId, const DDS::TopicQos& qos) DDS::ReturnCode_t Sedp::remove_publication_i(const RepoId& publicationId, LocalPublication& pub) { + if (!(spdp_.available_builtin_endpoints() & (DISC_BUILTIN_ENDPOINT_PUBLICATION_ANNOUNCER +#ifdef OPENDDS_SECURITY + | DDS::Security::SEDP_BUILTIN_PUBLICATIONS_SECURE_WRITER +#endif + ))) { + return DDS::RETCODE_PRECONDITION_NOT_MET; + } + #ifdef OPENDDS_SECURITY ICE::Endpoint* endpoint = pub.publication_->get_ice_endpoint(); if (endpoint) {
[No CFG could be retrieved]
This function checks if the TOPIC_DATA QoS changed on the local endpoints and if Reads a specific from the local cache and writes it to the DDS.
I think we don't need this block (and the corresponding one in `remove_subscription_i`) because the entity being removed couldn't have been added in the first place.
@@ -1221,6 +1221,7 @@ function $CompileProvider($provide, $$sanitizeUriProvider) { var terminalPriority = -Number.MAX_VALUE, newScopeDirective, controllerDirectives = previousCompileContext.controllerDirectives, + controllers, newIsolateScopeDirective = previousCompileContext.newIsolateScopeDirective, templateDirective = previousCompileContext.templateDirective, nonTlbTranscludeDirective = previousCompileContext.nonTlbTranscludeDirective,
[No CFG could be retrieved]
Applies the given directives to the current node and returns a link function that can be used to Parse the group node.
having a separate object for this is suspect... I think elementControllers could probably be used
@@ -61,6 +61,14 @@ export class AbstractHDWallet extends LegacyWallet { return this; } + setPassphrase(passphrase) { + this.passphrase = passphrase; + } + + getPassphrase() { + return this.passphrase; + } + /** * @return {Boolean} is mnemonic in `this.secret` valid */
[No CFG could be retrieved]
Returns the next free address in the chain. finds the next free address in the chain and updates the cache.
hmm, weird. its not used anywhere..?
@@ -8,11 +8,14 @@ import ( "github.com/elastic/beats/filebeat/harvester" "github.com/elastic/beats/filebeat/input/file" "github.com/elastic/beats/libbeat/logp" + + file_helper "github.com/elastic/beats/libbeat/common/file" ) // Log contains all log related data type Log struct { fs harvester.Source + stateOS file_helper.StateOS offset int64 config LogConfig lastTimeRead time.Time
[wait->[After,Duration],errorChecks->[Size,Stat,Continuable,Name,IsSameFile,Since,Err,Debug],Read->[Name,Now,errorChecks,wait,Read,Debug],Seek,Now]
NewLogimport imports and returns a new log instance from the given log file. errorChecks reads n bytes from the log and returns the number of bytes read and the first.
Please use a name without underscores, and don't leave an empty line between this import and the ones for other beats packages.
@@ -23,7 +23,7 @@ def safe_rmdir(dirname: str) -> None: class GenericTest(unittest.TestCase): # The path module to be tested - pathmodule = genericpath # type: Any + pathmodule = genericpath # type: Any common_attributes = ['commonprefix', 'getsize', 'getatime', 'getctime', 'getmtime', 'exists', 'isdir', 'isfile'] attributes = [] # type: List[str]
[GenericTest->[test_isfile->[safe_rmdir],test_isdir->[safe_rmdir]],CommonTest->[test_nonascii_abspath->[test_abspath]],test_main]
Test that path module does not raise TypeError.
``Any`` is the most precise _currently existing_ type that reflects the semantics of this variable. If we had dependent types/literal types, then this should be something like ``Union[Literal[genericpath], Literal[posixpath]]``.
@@ -343,11 +343,13 @@ public abstract class AbstractHoodieLogRecordReader { } } - protected HoodieRecord<?> createHoodieRecord(IndexedRecord rec) { - if (!simpleKeyGenFields.isPresent()) { - return SpillableMapUtils.convertToHoodieRecordPayload((GenericRecord) rec, this.payloadClassFQN, this.preCombineField, this.withOperationField); + private HoodieRecord<?> createHoodieRecord(IndexedRecord rec) { + if (!virtualKeysEnabled) { + return SpillableMapUtils.convertToHoodieRecordPayload((GenericRecord) rec, this.payloadClassFQN, + this.preCombineField, this.withOperationField); } else { - return SpillableMapUtils.convertToHoodieRecordPayload((GenericRecord) rec, this.payloadClassFQN, this.preCombineField, this.simpleKeyGenFields.get(), this.withOperationField); + return SpillableMapUtils.convertToHoodieRecordPayload((GenericRecord) rec, this.payloadClassFQN, + this.preCombineField, simpleKeyGenFields.get(), this.withOperationField, this.partitionName); } }
[AbstractHoodieLogRecordReader->[processQueuedBlocksForInstant->[processDataBlock],scan->[scan]]]
Process the data block.
I feel we should not touch the methods here in AbstractHoodieLogRecordReader. for a regular data table w/ virutal keys enabled, existing code should handle. Its only the metadata table which will have issues. So, may be we should override createHoodieRecord() within HoodieMetadataMergedLogRecordReader and handle the special case for partitionName. Also, thinking if we should confine the the new argument for this class (partitionName) to HoodieMetadataMergedLogRecordReader.
@@ -70,7 +70,6 @@ public class ConditionalOperationPrimaryOwnerFailTest extends MultipleCacheManag log.debugf("wrapEntryForWriting invoked with %s", context); Object mvccEntry = invocation.callRealMethod(); - assertNull(mvccEntry, "Entry should not be wrapped!"); assertNull(context.lookupEntry(key), "Entry should not be wrapped!"); return mvccEntry; }).when(spyEntryFactory).wrapEntryForWriting(any(InvocationContext.class), any(), anyInt(),
[ConditionalOperationPrimaryOwnerFailTest->[spyEntryFactory->[extractComponent,replaceComponent,spy],spyInvocationHandler->[wrapInboundInvocationHandler],createCacheManagers->[createClusteredCaches,fetchInMemoryState,getDefaultClusteredCacheConfig],testEntryNotWrapped->[killMember,any,fork,MagicKey,countDown,get,await,anyInt,handle,spyEntryFactory,anyBoolean,assertClusterSize,spyInvocationHandler,CountDownLatch,put,wrapEntryForWriting,cache]]]
Test if the entry is not wrapped.
If it's an instance of `NullCacheEntry` now, maybe you should assert that.
@@ -231,7 +231,9 @@ int safe_open(const char *pathname, int flags, ...) next_component = strchr(component + 1, '/'); if (next_component) { - *(next_component++) = '\0'; + *next_component = '\0'; + // Eliminate double slashes. + while (*(++next_component) == '/') { /*noop*/ } } struct stat stat_before, stat_after;
[No CFG could be retrieved]
Opens a file system object and returns - 1 if the object is not found. POINT if - 1 ;.
/me tends to use { /\* skip */ } as the empty body, just to make it clear ...
@@ -276,8 +276,8 @@ export class LegacyWallet extends AbstractWallet { if (!changeAddress) throw new Error('No change address provided'); let algo = coinSelectAccumulative; - if (targets.length === 1 && targets[0] && !targets[0].value) { - // we want to send MAX + // if targets has output without a value, we want send MAX to it + if (targets.some(i => !('value' in i))) { algo = coinSelectSplit; }
[No CFG could be retrieved]
Broadcasts a single transaction to a single address. This function creates a transaction.
this change should apply to multisig wallet as well
@@ -0,0 +1,17 @@ +const resolveFrom = require('resolve-from') +const path = require('path') +const fs = require('fs') + +const jestPath = require.resolve('jest') +const jestCliPath = resolveFrom(jestPath, 'jest-cli') +const jestUtilPath = resolveFrom(jestCliPath, 'jest-util') +const patchFilePath = path.resolve( + path.dirname(jestUtilPath), + 'installCommonGlobals.js', +) +const lines = fs.readFileSync(patchFilePath, 'utf-8').split('\n') + +lines[78] = ' globalObject.process = process' + +const result = lines.join('\n') +fs.writeFileSync(patchFilePath, result)
[No CFG could be retrieved]
No Summary Found.
`patchJest` may not be needed in the sdk, try removing it
@@ -151,8 +151,7 @@ class EventHubConsumerClient(EventHubClient): else: self._event_processors[(consumer_group, "-1")] = event_processor - event_processor.start() - return Task(event_processor) + event_processor.start() def close(self): # type: () -> None
[Task->[cancel->[info,_stop_eventprocessor]],EventHubConsumerClient->[receive->[keys,start,ValueError,format,EventPosition,EventProcessor,Task,warning],close->[stop,len,super,range,popitem],__init__->[dict,super,ValueError,get],_stop_eventprocessor->[stop]],getLogger]
Receive events from a specific partition. Create a new Task that will handle a single event.
`self._event_processors[[consumer_group, partition_id or "-1")] = event_processor`
@@ -1009,6 +1009,9 @@ OpenDDS::XTypes::MemberId BE_GlobalData::compute_id(AST_Field* field, AutoidKind mid = member_id++; } + if (mid > OpenDDS::DCPS::Serializer::MEMBER_ID_MAX) { + be_util::misc_error_and_abort("Member id exceeds the maximum allowed value"); + } member_id_map_[field] = mid; return mid; }
[No CFG could be retrieved]
Get the next unique member id for a field.
Could this have more context for the user? Which type and member is the problem?
@@ -188,6 +188,13 @@ module T::Private::Methods end T::Private::DeclState.current.reset! + if method_name == :method_added || method_name == :singleton_method_added + raise( + "Putting a `sig` on `#{method_name}` is not supported" + + " (sorbet-runtime uses this method internally to perform `sig` validation logic)" + ) + end + original_method = mod.instance_method(method_name) sig_block = lambda do T::Private::Methods.run_sig(hook_mod, method_name, original_method, current_declaration)
[_on_method_added->[add_module_with_final,_check_final_ancestors,add_final_method],set_final_checks_on_hooks->[_hook_impl],install_hooks->[_on_method_added],_hook_impl->[add_module_with_final,_check_final_ancestors,module_with_final?],run_all_sig_blocks->[run_sig_block_for_key],build_sig->[signature_for_method],method_to_key->[method_owner_and_name_to_key],maybe_run_sig_block_for_key->[has_sig_block_for_key],install_singleton_method_added_hook->[_on_method_added]]
Checks if a method has a and if so invokes it. If so it calls the Returns a new object if it can be found.
How does this interact with the `skip_on_method_added` above? Should this be before that?
@@ -209,9 +209,9 @@ const expectedFiles = { ], userManagementServer: [ - `${SERVER_MAIN_RES_DIR}config/liquibase/authorities.csv`, - `${SERVER_MAIN_RES_DIR}config/liquibase/users.csv`, - `${SERVER_MAIN_RES_DIR}config/liquibase/users_authorities.csv`, + `${SERVER_MAIN_RES_DIR}config/liquibase/authority.csv`, + `${SERVER_MAIN_RES_DIR}config/liquibase/user.csv`, + `${SERVER_MAIN_RES_DIR}config/liquibase/user_authority.csv`, `${SERVER_MAIN_RES_DIR}templates/mail/activationEmail.html`, `${SERVER_MAIN_RES_DIR}templates/mail/passwordResetEmail.html`, `${SERVER_MAIN_SRC_DIR}com/mycompany/myapp/domain/Authority.java`,
[No CFG could be retrieved]
This is a list of all possible errors that can be handled by the security framework. list of all possible logback resources.
Those 3 csv files are expected to be in the `config/liquibase/data/` directory no?
@@ -313,8 +313,8 @@ func (k *Keyrings) GetSecretKeyWithPassphrase(me *User, passphrase string, secre k.G().Log.Debug("- GetSecretKeyWithPassphrase() -> %s", ErrToOk(err)) }() ska := SecretKeyArg{ - All: true, - Me: me, + Me: me, + KeyType: AllSecretKeyTypes, } var skb *SKB skb, _, err = k.GetSecretKeyLocked(ska)
[GetSecretKeyWithPassphrase->[GetSecretKeyLocked],GetSecretKeyWithPrompt->[GetSecretKeyLocked],GetLockedLocalSecretKey->[LoadSKBKeyring],LoadSKBKeyring->[SKBFilenameForUser],GetSecretKeyWithStoredSecret->[GetSecretKeyLocked]]
GetSecretKeyWithPassphrase returns a GenericKey that is encrypted with the given passphrase. If the.
I think this can remain AllSecretKeyTypes for the same reason as btc.go
@@ -54,6 +54,7 @@ class TrainerPieces(NamedTuple): "are ignoring them, using instead the loaded model parameters." ) + # TODO(mattg): This should be updated now that directory_path no longer exists. if vocabulary_params.get("directory_path", None): logger.warning( "You passed `directory_path` in parameters for the vocabulary in "
[TrainerPieces->[from_params->[from_params],create_or_extend_vocab->[from_params]]]
Construct a new instance of the critical part of the network from the given parameters. Initialize a from the current model.
This whole class is getting removed in #3625
@@ -262,6 +262,11 @@ class Article < ApplicationRecord end end + def webhook_data + serializer = destroyed? ? DestroyedArticleSerializer : ArticleForWebhooksSerializer + serializer.new(self).serializable_hash + end + def remove_algolia_index remove_from_index! delete_related_objects
[Article->[username->[username],update_notifications->[update_notifications],update_cached_user->[username],readable_edit_date->[edited?]]]
This method will either index or remove the object from the index depending on the published state.
I wonder if this shouldn't be completely outside the model in a method that receive the article as an argument... This way the model and its serialization are completely detached and it doesn't have to know anything about this.
@@ -902,6 +902,9 @@ interface Function final String s = args.get(i).eval(bindings).asString(); if (s != null) { builder.append(s); + } else { + // Result of concatenation is null if any of the Values is null. + return ExprEval.of(null); } } return ExprEval.of(builder.toString());
[CaseSearchedFunc->[apply->[eval,name]],UpperFunc->[apply->[name]],NvlFunc->[apply->[eval,name]],StrlenFunc->[apply->[name]],ConditionFunc->[apply->[eval,name]],SingleParamMath->[eval->[eval]],DoubleParam->[apply->[eval,name]],CaseSimpleFunc->[apply->[eval,name]],TimestampFromEpochFunc->[apply->[eval,name]],ReplaceFunc->[apply->[name]],LowerFunc->[apply->[name]],SubstringFunc->[apply->[name]],DoubleParamMath->[eval->[eval]],SingleParam->[apply->[name]]]
Evaluate the .
Please refer to SQL spec or some other doc that specifies this
@@ -409,7 +409,7 @@ public class ComponentManagerImpl implements ComponentManager { } void sendEvent(ComponentEvent event) { - log.debug("Dispatching event: " + event); + log.trace("Dispatching event: {}", event); Object[] listeners = this.compListeners.getListeners(); for (Object listener : listeners) { ((ComponentListener) listener).handleEvent(event);
[ComponentManagerImpl->[getStartFailureRegistrations->[getRegistrations],startComponent->[getComponent],getService->[getComponentProvidingService],refresh->[start,reset,refresh,stop],unregisterExtension->[unregisterExtension,sendEvent,getComponent],getComponentProvidingService->[getComponent],unstash->[isRunning,applyStash],getComponent->[getComponent],standby->[stopComponents],restart->[start,stop],reset->[stop],activateComponent->[size,registerExtension,registerServices,getComponent],Listeners->[afterStart->[afterStart],beforeActivation->[beforeActivation],afterActivation->[afterActivation],remove->[remove],beforeStart->[beforeStart],add->[add],afterStop->[afterStop],beforeDeactivation->[beforeDeactivation],afterDeactivation->[afterDeactivation],beforeStop->[beforeStop]],start->[activateComponents,startComponents],runWihtinTimeout->[shutdown],unregisterByLocation->[unregister],size->[size],applyStash->[unregister,register],applyStashWhenRunning->[applyStash,isStarted,isStandby,startComponent,stopComponent,activateComponent,deactivateComponent],resume->[startComponents],getActivatingRegistrations->[getRegistrations],getServices->[size],unregister->[unregister],stop->[stopComponents,deactivateComponents],stopComponents->[size],RIApplicationStartedComparator->[compare->[compare]],registerExtension->[registerExtension,sendEvent,getComponent],loadContributions->[loadContributions],stopComponent->[getComponent],shouldStash->[isRunning],deactivateComponent->[getComponent,unregisterServices,unregisterExtension],Stash->[isEmpty->[isEmpty],getRegistrationsToRemove->[getComponent,add],remove->[add],add->[add]]]]
Dispatches the given event to all registered listeners.
Here `trace` is fine as we potentially sends lots of those per transaction
@@ -236,14 +236,14 @@ abstract class AbstractEpollChannel extends AbstractChannel { return localReadAmount; } - protected final int doWriteBytes(ByteBuf buf) throws Exception { + protected final int doWriteBytes(ByteBuf buf, int writeSpinCount) throws Exception { int readableBytes = buf.readableBytes(); int writtenBytes = 0; if (buf.hasMemoryAddress()) { long memoryAddress = buf.memoryAddress(); int readerIndex = buf.readerIndex(); int writerIndex = buf.writerIndex(); - for (;;) { + for (int i = writeSpinCount - 1; i >= 0; i--) { int localFlushedAmount = Native.writeAddress( fileDescriptor.intValue(), memoryAddress, readerIndex, writerIndex); if (localFlushedAmount > 0) {
[AbstractEpollChannel->[modifyEvents->[isOpen],newDirectBuffer->[newDirectBuffer],doDisconnect->[doClose],doBeginRead->[setFlag],doWriteBytes->[setFlag],AbstractEpollUnsafe->[clearEpollIn0->[clearFlag],epollOutReady->[flush0],flush0->[isFlagSet,flush0]]]]
This method is used to read and write bytes from the specified buffer. Returns EPOLLOUT if EAGAIN need to set EPOLLOUT otherwise.
The initial value has to be `writeSpinCount - 1`. Please see `NioSocketChannel.doWrite()`.
@@ -82,12 +82,13 @@ class AccountController extends AbstractContactController protected $accountContactFieldDescriptors; protected $accountAddressesFieldDescriptors; - public function __construct() + public function setContainer(ContainerInterface $container = null) { + parent::setContainer($container); + $this->initFieldDescriptors(); $this->initAccountContactFieldDescriptors(); $this->initAccountAddressesFieldDescriptors(); - } /**
[AccountController->[setResponsiblePerson->[setResponsiblePerson],setParent->[setParent],putAction->[setResponsiblePerson]]]
Initialize the object.
Why did you change this to `setContainer` instead of using the constructor? Adding these methods is not what this method is thought for.
@@ -2113,6 +2113,7 @@ describe('$compile', function() { var iscope; beforeEach(module(function() { + var fakeModule = {name: 'fakeModule'}; forEach(['', 'a', 'b'], function(name) { directive('scope' + uppercase(name), function(log) { return {
[No CFG could be retrieved]
Requires angular 1. 8. 0 directive - directive tscope and stscope.
Faking out a module to pass to the directive function so the test will show the result of passing a module.
@@ -252,7 +252,7 @@ namespace System.Net.Quic.Implementations.MsQuic state.ShutdownTcs.SetResult(MsQuicStatusCodes.Success); // Stop accepting new streams. - state.AcceptQueue.Writer.Complete(); + state.AcceptQueue.Writer.TryComplete(); // Stop notifying about available streams. TaskCompletionSource? unidirectionalTcs = null;
[MsQuicConnection->[Task->[Dispose],NativeCallbackHandler->[HandleEventConnected,HandleEventStreamsAvailable,HandleEventShutdownInitiatedByTransport,HandleEventPeerCertificateReceived,HandleEventShutdownComplete,HandleEventShutdownInitiatedByPeer,HandleEventNewStream],Dispose->[Dispose,SetClosing],HandleEventNewStream->[TryQueueNewStream],Dispose]]
HandleEventShutdownComplete - This method is called when a shutdown is complete.
It seems like there are a bunch of places now where we either call Writer.Complete or Writer.TryComplete. Are all of these necessary? I would assume we only have to do this in a limited number of places, e.g. HandleEventShutdownInitiatedByTransport and HandleEventShutdownInitiatedByPeer, and maybe one other for when we initiate shutdown ourselves. I always get nervous when we do stuff like call TryComplete instead of Complete, because it seems like we don't have a clean handling of exactly when the Complete operation needs to happen. I'd prefer to see the TryComplete calls changed to either Complete or to assert that it's already completed.
@@ -18,7 +18,7 @@ <%= auto_discovery_link_tag(:rss, app_url("/feed/#{@user.username}"), title: "#{community_name} RSS Feed") %> <% end %> -<% if @user.suspended? %> +<% if @user.score < 0 @user.suspended? %> <meta name="robots" content="noindex"> <meta name="robots" content="nofollow"> <% end %>
[No CFG could be retrieved]
Returns a link tag to the user s RSS feed.
Typo and syntax error, missing a `||` or a `&`
@@ -47,12 +47,10 @@ extern "C" { #include "include/UdpContext.h" //#define DEBUG_SSDP Serial -#define SSDP_INTERVAL 1200 #define SSDP_PORT 1900 #define SSDP_METHOD_SIZE 10 #define SSDP_URI_SIZE 2 #define SSDP_BUFFER_SIZE 64 -#define SSDP_MULTICAST_TTL 2 // ssdp ipv6 is FF05::C // lwip-v2's igmp_joingroup only supports IPv4
[No CFG could be retrieved]
THIS IS A PUREURELY REALLY INSIDE THE SOFTWARE XML - RPC - Response - Response - Response - Notifier - Response - Notifier -.
I would leave `SSDP_INTERVAL` and `SSDP_MULTICAST_TTL` defined and used in member initializers for clarity (maybe also adding the unity: `SSDP_INTERVAL_SECONDS`)
@@ -37,8 +37,8 @@ export const UrlReplacementPolicy = { * @param {!Element} element * @param {string=} opt_expr Dot-syntax reference to subdata of JSON result * to return. If not specified, entire JSON result is returned. - * @param {UrlReplacementPolicy=} opt_urlReplacement If ALL, replaces all URL vars. - * If OPT_IN, replaces whitelisted URL vars. Otherwise, don't expand. + * @param {UrlReplacementPolicy=} opt_urlReplacement If ALL, replaces all URL + * vars. If OPT_IN, replaces whitelisted URL vars. Otherwise, don't expand. * @return {!Promise<!JsonObject|!Array<JsonObject>>} Resolved with JSON * result or rejected if response is invalid. */
[No CFG could be retrieved]
Batch fetches the JSON object for the given element. If the element is whitelisted add the unwhitelisted values to the data - amp - replace.
Whoa, I don't remember this file at all. I think really bad naming, `SERVICEfor` generally just gets the installed service, but this adds in a bunch of element specific code and URL replacements just for fun.
@@ -21,7 +21,7 @@ import {guaranteeSrcForSrcsetUnsupportedBrowsers} from '../src/utils/img'; import {isExperimentOn} from '../src/experiments'; import {listen} from '../src/event-helper'; import {registerElement} from '../src/service/custom-element-registry'; -import {setImportantStyles} from '../src/style'; +import {installObjectCropStyles, setImportantStyles} from '../src/style'; /** * Attributes to propagate to internal image when changed externally.
[No CFG could be retrieved]
Creates an object that represents a single unique identifier. Replies the attributes of an AMP image.
I suggest `propagateObjectFitStyles` - `propagate` for consistency with `propagateAttributes` - `ObjectFit` eventhough it is less accurate, it is more clear what it does because it matches one of the CSS properties related to crop.
@@ -879,3 +879,13 @@ def hash_file(path, blocksize=1 << 20): length += len(block) h.update(block) return (h, length) # type: ignore + + +def interpreter_name(): + # type: () -> str + try: + name = sys.implementation.name # type: ignore + except AttributeError: # pragma: no cover + # Python 2.7 compatibility. + name = platform.python_implementation().lower() + return name
[hide_value->[HiddenText],hide_url->[redact_auth_from_url,HiddenText],get_installed_distributions->[editables_only_test->[dist_is_editable],editable_test->[dist_is_editable],user_test,editables_only_test,local_test,editable_test],captured_stdout->[captured_output],redact_netloc->[split_auth_from_netloc],split_auth_netloc_from_url->[_transform_url],captured_stderr->[captured_output],dist_location->[egg_link_path,normalize_path],ask_input->[_check_no_input],dist_in_site_packages->[normalize_path],captured_output->[from_stream],_get_netloc->[split_auth_from_netloc],is_local->[normalize_path],dist_is_local->[is_local],rmtree->[rmtree],parse_netloc->[build_url_from_netloc],remove_auth_from_url->[_transform_url],hash_file->[read_chunks],redact_auth_from_url->[_transform_url],ask->[_check_no_input],splitext->[splitext],dist_in_usersite->[normalize_path],_redact_netloc->[redact_netloc],ask_password->[_check_no_input]]
Return hash and length of file with block size.
We should create an issue to replace this with `packaging.tags.interpreter_name()` once we pick up the as-yet released version.
@@ -289,12 +289,16 @@ describe('session module', () => { res.end(mockPDF) downloadServer.close() }) + + const isPathEqual = (path1, path2) => { + return path.relative(path1, path2) === '' + } const assertDownload = (event, state, url, mimeType, receivedBytes, totalBytes, disposition, filename, port, savePath, isCustom) => { assert.equal(state, 'completed') assert.equal(filename, 'mock.pdf') - assert.equal(savePath, path.join(__dirname, 'fixtures', 'mock.pdf')) + assert.equal(isPathEqual(savePath, path.join(__dirname, 'fixtures', 'mock.pdf')), true) if (isCustom) { assert.equal(url, `${protocolName}://item`) } else {
[No CFG could be retrieved]
Example of how to download an item from a web page. link to download file.
I know it's kinda an edge case but this will glitch out if `path.relative` receives a non-absolute path. `path` will automatically try resolve it based on the CWD. Can we just `toLowerCase()` to solve your issue?
@@ -73,7 +73,10 @@ class ACRExchangeClient(object): def exchange_aad_token_for_refresh_token(self, service=None, **kwargs): # type: (str, Dict[str, Any]) -> str refresh_token = self._client.authentication.exchange_aad_access_token_for_acr_refresh_token( - service=service, access_token=self._credential.get_token(*self.credential_scopes).token, **kwargs + grant_type=PostContentSchemaGrantType.ACCESS_TOKEN, + service=service, + access_token=self._credential.get_token(*self.credential_scopes).token, + **kwargs ) return refresh_token.refresh_token
[ACRExchangeClient->[close->[close],__exit__->[__exit__],__enter__->[__enter__],__init__->[ExchangeClientAuthenticationPolicy]]]
Exchange an AAD access token for a refresh token.
it seems that the service allows for users to pass in more than just "access_token" for grant_type, and it doesn't look like we're exposing this functionality to users. Is this expected?
@@ -379,7 +379,7 @@ def plot_raw(raw, events=None, duration=10.0, start=0.0, n_channels=20, if show_options is True: _toggle_options(None, params) # initialize the first selection set - if order in ['selection', 'position']: + if order in ['selection', 'position', 'lasso']: _radio_clicked(fig_selection.radio.labels[0]._text, params) callback_selection_key = partial(_selection_key_press, params=params) callback_selection_scroll = partial(_selection_scroll, params=params)
[_pick_bad_channels->[_plot_update_raw_proj],_label_clicked->[_plot_update_raw_proj],plot_raw_psd->[_set_psd_plot_params]]
Plot a raw block of data with a . A function to plot a single object. Plots a ButterKnea - Butter - Andering - House and Missing key - value block in the system. This function is called when a channel is classified and the user has clicked a missing key.
You're using this set ('selection', 'position', 'lasso') at least 3x, maybe just initialize it as a list once and refer to it.
@@ -147,7 +147,7 @@ def _append_revision(data_points, revision, **kwargs): class PlotData: REVISION_FIELD = "rev" - INDEX_FIELD = "index" + INDEX_FIELD = "step" def __init__(self, filename, revision, content, **kwargs): self.filename = filename
[plot_data->[PlotMetricTypeError],_find_data->[PlotDataStructureError,_lists],_lists->[_lists],_apply_path->[PlotDataStructureError],PlotData->[to_datapoints->[PlotParsingError,raw,_processors]]]
Initialize a new object with the given filename revision and content.
Seems like `step` is generic enough to replace `index` - in both cases they do not make too much sense when plotting anything, so we agreed with @dmpetrov that we can replace it for the sake of "additional config"-less integration with logs.
@@ -76,7 +76,8 @@ public abstract class DynamicIFrame extends Frame { if (loadFrame.execute()) { - Scheduler.get().scheduleFixedDelay(loadFrame, 50); + if (retryInterval_ < 500) retryInterval_ += 50; + Scheduler.get().scheduleFixedDelay(loadFrame, retryInterval_); } } });
[DynamicIFrame->[pollForLoad->[execute->[execute]],getDocument->[getDocument],setUrl->[setUrl]]]
pollForLoad - Poll for a window object to become available.
I think the `retryInterval_` should be reset whenever we set the URL (i.e. at the start of `pollForLoad`) so that each load gets its own backoff ramp.
@@ -15,8 +15,16 @@ */ import {CommonSignals} from '../../../src/common-signals'; +import {Services} from '../../../src/services'; +import {user} from '../../../src/log'; import {whenUpgradedToCustomElement} from '../../../src/dom'; +/** + * Maximum milliseconds to wait for service to load before logging a warning. + * @const + */ +const LOAD_TIMEOUT = 3000; + /** @implements {../../../src/render-delaying-services.RenderDelayingService} */ export class AmpStoryRenderService { /**
[No CFG could be retrieved]
Provides a function to return a promise for when it is finished delaying render and is ready.
This needs to be shorter than the render delay timeout to account for the latency downloading and executing the amp-story js.
@@ -17,13 +17,15 @@ class AddUserEmailFingerprint < ActiveRecord::Migration rename_column :users, :email_plain, :email decrypt_user_emails change_column_null :users, :email, false + add_index :users, :email, unique: true remove_column :users, :email_fingerprint remove_column :users, :encrypted_email end def encrypt_user_emails User.where(encrypted_email: '').each do |user| - ee = EncryptedEmail.new_from_email(user.email) + email_address = user.email.present? ? user.email : user.id.to_s + ee = EncryptedEmail.new_from_email(email_address) # must use raw SQL here to change data during the migration transaction. execute "UPDATE users SET encrypted_email='#{ee.encrypted}', email_fingerprint='#{ee.fingerprint}' WHERE id=#{user.id}" end
[AddUserEmailFingerprint->[decrypt_user_emails->[new,quote,decrypted,encrypted_email,each,id,execute],encrypt_user_emails->[email,fingerprint,new_from_email,each,encrypted,id,execute],up->[rename_column,change_column_null,add_column,remove_index],down->[rename_column,change_column_null,remove_column]]]
Remove any user with a lease that has no lease.
Where is the code that would decrypt the email field and read the user ID? Or is this just to have something unique before it gets dropped?
@@ -86,6 +86,7 @@ ica_picks = fiff.pick_types(raw.info, meg=True, eeg=False, stim=False, ecg=False, eog=False, exclude='bads') + def test_plot_topo(): """Test plotting of ERP topography """
[test_plot_sparse_source_estimates->[,any,mne_analyze_colormap,plot_source_estimates,SourceEstimate,len,sum,assert_raises,empty,where,plot_sparse_source_estimates,zeros],test_plot_topo->[dict,plot_topo,pick_channels_evoked,assert_raises],test_plot_raw_psds->[pick_types,plot_psds,assert_raises,axes],test_plot_connectivity_circle->[randn,circular_layout,plot_connectivity_circle],test_plot_topomap->[plot_topomap,catch_warnings,repeat,plot_projs_topomap,plot_evoked_topomap,assert_raises,read_proj,read_evoked],test_plot_cov->[read_cov,plot_cov],test_plot_topo_power->[randn,arange,plot_topo_power,abs,plot_topo_phase_lock],test_plot_drop_log->[plot_drop_log],test_plot_evoked->[dict,apply_proj,assert_raises,plot],test_plot_epochs->[epochs,plot],test_plot_ica_panel->[read_cov,decompose_raw,ICA,plot_sources_raw],test_plot_image_epochs->[plot_image_epochs],test_plot_topo_image_epochs->[plot_topo_image_epochs],test_plot_raw->[plot],test_compare_fiff->[compare_fiff],test_plot_topo_tfr->[randn,arange,plot_topo_tfr,len],requires_sklearn->[dec->[function,check_sklearn_version,SkipTest],wraps],test_plot_ica_topomap->[plot_topomap,read_cov,assert_raises,decompose_raw,ICA],skipif,round,Raw,read_events,average,dict,use,dirname,data_path,read_source_spaces,read_layout,join,Epochs,simplefilter,pick_types]
Test plotting of ERP topography .
Too many newlines
@@ -273,7 +273,7 @@ class Nit(object): def __str__(self): """convert ascii for safe terminal output""" flat = list(self.flatten_lines([self.message], self.lines)) - return '\n |'.join(flat).encode('ascii', errors='replace') + return '\n |'.join(flat).encode('ascii', errors='replace').decode('ascii') @property def line_number(self):
[Nit->[__str__->[flatten_lines]],PythonFile->[enumerate->[enumerate],__init__->[OffByOneList,_remove_coding_header],parse->[_parse],from_statement->[_parse],iter_logical_lines->[iter_tokens,translate_logical_line],tokens->[iter_tokens]],CheckstylePlugin->[__iter__->[nits],warning->[nit],comment->[nit],error->[nit],nit->[line_range,Nit]],OffByOneList->[index->[index]],CheckSyntaxError->[as_nit->[Nit,OffByOneList]]]
convert ascii for safe terminal output.
`__str__` must return unicode in Py3 no matter what. This function will still strip out any non-ascii, though, and then spit it back out as unicode.
@@ -356,8 +356,10 @@ OPENSSL_INIT_SETTINGS *OPENSSL_INIT_new(void) { OPENSSL_INIT_SETTINGS *ret = malloc(sizeof(*ret)); - if (ret != NULL) - memset(ret, 0, sizeof(*ret)); + if (ret == NULL) + return; + + memset(ret, 0, sizeof(*ret)); ret->flags = DEFAULT_CONF_MFLAGS; return ret;
[CONF_dump_bio->[CONF_set_nconf],NCONF_get_number_e->[NCONF_get_string],CONF_load_bio->[CONF_set_nconf],CONF_get_string->[CONF_set_nconf],CONF_get_section->[CONF_set_nconf],CONF_free->[CONF_set_nconf],CONF_get_number->[CONF_set_nconf]]
This method creates a new openssl_init_settings object if it is not NULL and.
That should be return NULL
@@ -119,6 +119,15 @@ export default class JitsiStreamBackgroundEffect { this._inputVideoElement.width, this._inputVideoElement.height ); + } + if (this._options.virtualBackground.backgroundType === 'desktop-share') { + this._outputCanvasCtx.drawImage( + this._virtualVideo, + 0, + 0, + this._inputVideoElement.width, + this._inputVideoElement.height + ); } else { this._outputCanvasCtx.filter = `blur(${this._options.virtualBackground.blurValue}px)`; this._outputCanvasCtx.drawImage(this._inputVideoElement, 0, 0);
[No CFG could be retrieved]
Draws the segmentation mask image video and the background. Loop function to render the background mask.
This can be combined with the if statement above that checks for `backgroundType === 'image' `as only the input element that is to be drawn into the context will change depending on the `backgroundType`.
@@ -131,6 +131,7 @@ public class TimedShutoffInputSourceReader implements InputSourceReader shutoffTime.getMillis() - System.currentTimeMillis(), TimeUnit.MILLISECONDS ); + exec.shutdown(); return wrappingIterator; }
[TimedShutoffInputSourceReader->[decorateShutdownTimeout->[close->[close],hasNext->[hasNext],close],sample->[sample],read->[read]]]
Decorates the shutdown timeout with a new CloseableIterator.
What's the purpose of this line? (It doesn't seem necessary to me)
@@ -127,6 +127,7 @@ def test_generate_sample_customer_payload(customer_user): def test_generate_sample_product_payload(variant): + variant.product.refresh_from_db() payload = generate_sample_payload(WebhookEventType.PRODUCT_CREATED) assert payload == json.loads(generate_product_payload(variant.product))
[test_generate_sample_payload_order->[_remove_anonymized_order_data],test_generate_sample_payload_fulfillment_created->[_remove_anonymized_order_data],test_generate_sample_checkout_payload->[_remove_anonymized_checkout_data]]
Test generate sample product payload.
Why we do this?
@@ -45,7 +45,7 @@ CONTAINER; '<div class="pd-embed" data-settings="' . esc_attr( json_encode( $settings ) ) . '"></div>'; - if ( $type === 'button' ) { + if ( $settings['type'] === 'button' ) { $placeholder = '<a class="pd-embed" href="' . esc_attr( $survey_link )
[PolldaddyShortcode->[polldaddy_shortcode->[compress_it,get_async_code]]]
Get the javascript code that embeds the missing javascript in the main window.
:yoda: style please! `'button' === $settings['type']` (same for below changeset)
@@ -156,9 +156,6 @@ class Statement(Node): class Expression(Node): """An expression node.""" - literal = LITERAL_NO - literal_hash = None # type: Key - def accept(self, visitor: ExpressionVisitor[T]) -> T: raise RuntimeError('Not implemented')
[ClassDef->[serialize->[serialize],is_generic->[is_generic],deserialize->[ClassDef,deserialize]],merge->[MroError],FuncDef->[serialize->[serialize],deserialize->[FuncDef]],Decorator->[fullname->[fullname],serialize->[serialize],name->[name],deserialize->[Decorator,deserialize]],TypeVarExpr->[serialize->[serialize],deserialize->[TypeVarExpr]],SymbolTable->[serialize->[serialize],deserialize->[deserialize,SymbolTable]],FuncItem->[set_line->[set_line],__init__->[name]],MypyFile->[serialize->[serialize],deserialize->[deserialize,MypyFile]],NameExpr->[serialize->[serialize],deserialize->[deserialize,NameExpr]],TypeInfo->[__repr__->[fullname],has_readable_member->[get],serialize->[fullname,serialize],__getitem__->[get],get->[get],has_base->[fullname],deserialize->[TypeInfo,deserialize],_calculate_is_enum->[fullname],is_metaclass->[fullname],__str__->[fullname]],Argument->[_initialization_statement->[name],set_line->[set_line]],get_member_expr_fullname->[get_member_expr_fullname],Var->[serialize->[serialize],deserialize->[Var]],SymbolTableNode->[fullname->[fullname],serialize->[fullname,serialize],deserialize->[deserialize,SymbolTableNode]],linearize_hierarchy->[direct_base_classes,fullname,linearize_hierarchy],OverloadedFuncDef->[serialize->[serialize],deserialize->[deserialize,OverloadedFuncDef],__init__->[set_line]]]
Visit this node and return the next node.
@ilevkivskyi isn't this a violation of strict optional? This value is actually read, it's not just a declaration. I assume this should be `Optional[Key]`? (This is why the build for this PR breaks; the new `literal_hash()` returns None when there's no match, which is what the client code expects).
@@ -908,6 +908,9 @@ class TestUnicode(BaseTest): ] for sep, parts in CASES: + print(sep) + print(parts) + print(cfunc(sep, parts)) self.assertEqual(pyfunc(sep, parts), cfunc(sep, parts), "'%s'.join('%s')?" % (sep, parts))
[title->[title],TestUnicodeInTuple->[test_ascii_flag_getitem->[f],test_ascii_flag_join->[f],test_const_unicode_in_tuple->[f],test_ascii_flag_add_mul->[f],test_const_unicode_in_hetero_tuple->[f],test_ascii_flag_unbox->[f,isascii]],TestUnicode->[test_pointless_slice->[pyfunc],test_istitle->[title],test_rfind_wrong_start_end_optional->[try_compile_wrong_end_optional,try_compile_wrong_start_optional],test_lower->[pyfunc],test_count_optional_arg_type_check->[try_compile_bad_optional],test_literal_xyzwith->[pyfunc],test_le->[_check_ordering_op],test_literal_in->[pyfunc],test_gt->[_check_ordering_op],test_literal_getitem->[pyfunc],test_basic_lt->[pyfunc],test_ge->[_check_ordering_op],test_not->[pyfunc],test_literal_comparison->[pyfunc],test_title->[pyfunc],test_comparison->[pyfunc],test_basic_gt->[pyfunc],test_literal_concat->[pyfunc],test_isupper->[pyfunc],test_islower->[pyfunc],test_walk_backwards->[pyfunc],test_lt->[_check_ordering_op],test_stride_slice->[pyfunc],test_upper->[pyfunc],test_literal_find->[pyfunc],test_literal_len->[pyfunc]],TestUnicodeIteration->[test_unicode_stopiteration_iter->[f],test_unicode_literal_stopiteration_iter->[f],test_unicode_iter->[pyfunc],test_unicode_enumerate_iter->[pyfunc],test_unicode_literal_iter->[pyfunc]]]
Test join of strings.
please remove these print statements.
@@ -75,7 +75,7 @@ Rails.application.configure do config.assets.raise_runtime_errors = true # Only log info and higher on development - config.log_level = :info + config.log_level = :debug # Only allow Better Errors to work on trusted ip, use ifconfig to see which # one you use and put it into application.yml!
[perform_deliveries,mailer_reply_to,default_options,raise_delivery_errors,enable_recaptcha,mailer_authentication,mailer_port,headers,enable_email_confirmations,new_team_on_signup,consider_all_requests_local,cache_classes,mailer_address,deprecation,allow_ip!,exist?,enabled,enable_user_registration,migration_error,debug,delivery_method,to_i,eager_load,default_url_options,linkedin_signin_enabled,raise_on_missing_translations,mailer_from,cache_store,perform_caching,mailer_password,raise_runtime_errors,mailer_domain,digest,log_level,configure,mailer_user_name,smtp_settings,mail_server_url]
Configuration of a single asset. Configure the linked - in signin feature.
I also use this locally a lot, but I'm not sure if this is needed to be changed?
@@ -357,6 +357,9 @@ define([ if (directions && (directions.length >= 3)) { this._va = CustomSensorVolume._createVertexArray(context, directions, this.radius, this.bufferUsage); } + + this.boundingVolume.center = Cartesian3.fromCartesian4(this.modelMatrix.getColumn(3)); + this.boundingVolume.radius = isFinite(this.radius) ? this.radius : FAR; } } };
[No CFG could be retrieved]
Creates a vertex buffer when material changes and directions changes. Adds a shader to the object that will be used to pick the sensor.
As we discussed, we can also do more tightly fit bounding spheres for all sensor types by adjusting the center point. For complex conic, we need to consider the outer angle, which can be greater than 180 degrees. For custom sensors, we can just put a sphere around the computed points. This will be a nice optimization because sensors tend to be large, so a loose bounding sphere would lead to lots of drawing in other frustums.
@@ -84,6 +84,9 @@ namespace Revit.Elements { ElementBinder.SetElementForTrace(this.InternalElement); } + + // necessary so the points in the topography are valid + DocumentManager.Regenerate(); } [SupressImportIntoVM]
[Topography->[InternalSetTopographySurface->[UniqueId,Id,InternalUniqueId,InternalElementId],Regenerate,GetPoints,Create,CleanupAndSetElementForTrace,First,ElementBinder,InternalElement,EnsureInTransaction,ToList,SetElementForTrace,TransactionTaskDone,CurrentDBDocument,InternalSetTopographySurface]]
Gets the underlying triangular mesh from the Topography surface. Method to set the TopographySurface object.
Just a small fix. I'd prefer to avoid side-effects like this in a property, which might be invoked in debugging or from the UI.
@@ -126,6 +126,9 @@ func (a *apiServer) workerPodSpec(options *workerOptions, pipelineInfo *pps.Pipe }, { Name: client.PPSPipelineNameEnv, Value: pipelineInfo.Pipeline.Name, + }, { + Name: "JAEGER_ENDPOINT", + Value: os.Getenv("JAEGER_ENDPOINT"), }, // These are set explicitly below to prevent kubernetes from setting them to the service host and port. {
[createWorkerSvcAndRc->[createWorkerPachctlSecret,workerPodSpec,getWorkerOptions]]
workerPodSpec returns the pod spec that will be used to create a worker container. This function is used to setup environment variables for a node that can be found in the cluster The environment variables that are passed to the daemon are set in the environment variables.
Could the workers rely on `JAEGER_COLLECTOR_SERVICE_HOST`? I think that would eliminate the need to set any env vars explicitly, and might make deploying Jaeger into an existing cluster easier (just deploy Jaeger and restart the worker pods, and they'll start talking to Jaeger automatically)
@@ -367,6 +367,6 @@ function BaseCarouselWithRef( ); } -const BaseCarousel = forwardRef(BaseCarouselWithRef); -BaseCarousel.displayName = 'BaseCarousel'; // Make findable for tests. -export {BaseCarousel}; +const BentoBaseCarousel = forwardRef(BaseCarouselWithRef); +BentoBaseCarousel.displayName = 'BentoBaseCarousel'; // Make findable for tests. +export {BentoBaseCarousel};
[No CFG could be retrieved]
Get the base carousel.
please also name this `BentoBaseCarouselWithRef`
@@ -284,4 +284,14 @@ export class WatchOnlyWallet extends LegacyWallet { if (this._hdWalletInstance) return this._hdWalletInstance.setUTXOMetadata(...args); return super.setUTXOMetadata(...args); } + + getDerivationPath() { + if (!this.isHd()) throw new Error('Not initialized'); + return this._derivationPath; + } + + setDerivationPath(path) { + if (!this.isHd()) throw new Error('Not initialized'); + this._derivationPath = path; + } }
[No CFG could be retrieved]
Set UTXOMetadata on the HDWalletInstance.
shouldnt we proxy this call to `this._hdWalletInstance` ? "Not initialized" is not a correct error message here
@@ -63,6 +63,12 @@ if ($user->socid > 0) { $socid = $user->socid; } +// list of limit id for invoice +$facids = GETPOST('facids','alpha'); +if(empty($facid) && !empty($facids)) { + $r =explode(',', $facids); + $facid = $r[0]; +} $object = new Facture($db);
[fetch,addPaymentToBank,create,formconfirm,fetch_object,getSommePaiement,jdate,getNomUrl,rollback,begin,initHooks,select_comptes,load,plimit,getSumCreditNotesUsed,transnoentities,executeHooks,loadLangs,select_types_paiements,close,selectDate,free,query,trans,num_rows,hasDelay,fetch_thirdparty,commit,getSumDepositsUsed]
Load and initialize a single user object. This function checks if a given node has an action and if so executes it.
Should use GETPOST('facids', 'intcomma'); or a SQL injection will be possible later in the $sql .= ' AND f.rowid IN('.$facids.')';
@@ -56,6 +56,12 @@ RSpec::Matchers.define_negated_matcher :not_change, :change Rack::Attack.enabled = false +# `browser`, a dependency of `field_test`, starting from version 3.0 +# considers the empty user agent a bot, with will fail tests as we +# explicitly configure field tests to exclude bots +# see https://github.com/fnando/browser/blob/master/CHANGELOG.md#300 +Browser::Bot.matchers.delete(Browser::Bot::EmptyUserAgentMatcher) + RSpec.configure do |config| config.use_transactional_fixtures = true config.fixture_path = "#{::Rails.root}/spec/fixtures"
[disable_net_connect!,clear_cache,define_negated_matcher,info,fixture_path,recreate_indexes,infer_spec_type_from_file_location!,filter_rails_from_backtrace!,clear_all,root,toggle_live,before,logger,require,to_rack,enabled,include,key?,use_transactional_fixtures,include?,to_json,production?,around,each,turned_off,test_mode,abort,after,run,configure,allow_net_connect!,to_return,maintain_test_schema!,expand_path]
Requires all required files and registers necessary functions. end of class.
`which will fail tests as we`
@@ -16,7 +16,7 @@ export default class WalletDetails extends Component { title: loc.wallets.details.title, headerRight: ( <TouchableOpacity - style={{ marginHorizontal: 16 }} + style={{ marginHorizontal: 16, height: 40, width: 40, justifyContent: 'center', alignItems: 'center' }} onPress={() => { navigation.getParam('saveAction')(); }}
[No CFG could be retrieved]
The base class for the wallets that are used to display the details of a single user Displays the UI.
why even add this setting in wallet details when it can just be saved from wallet/transactions when user taps the number and it changes to desired unit..?
@@ -79,3 +79,18 @@ class StateSampler(statesampler_impl.StateSampler): for state in self._states_by_name.values(): state_msecs = int(1e-6 * state.nsecs) state.counter.update(state_msecs - state.counter.value()) + + +def simple_tracker(): + """Create, register and return a simple StateSampler.""" + sampler = StateSampler('', CounterFactory()) + sampler.register() + return sampler + + +def make_full_tracker(prefix, counter_factory, + sampling_period_ms=DEFAULT_SAMPLING_PERIOD_MS): + """Create, register and return a StateSampler.""" + sampler = StateSampler(prefix, counter_factory, sampling_period_ms) + sampler.register() + return sampler
[StateSampler->[stop_if_still_running->[stop],commit_counters->[int,values,update,value],scoped_state->[super,CounterName,get_counter],get_info->[StateSamplerInfo,current_state],__init__->[super]],namedtuple]
Updates output counters with latest state statistics.
If this is just used for tests, put it in a test file. I would rather force the callers to provide a prefix and counter factory manually.
@@ -318,6 +318,9 @@ def make_bem_solution(surfs, verbose=None): logger.info('Homogeneous model surface loaded.') else: raise RuntimeError('Only 1- or 3-layer BEM computations supported') + if bem['surfs'][0]['np'] > 10000: + warn('The bem surface has %s data points. 5120 should be enough.' + % bem['surfs'][0]['np']) _fwd_bem_linear_collocation_solution(bem) logger.info('BEM geometry computations complete.') return bem
[make_sphere_model->[ConductorModel,_fwd_eeg_fit_berg_scherg],_lin_pot_coeff->[_calc_beta],fit_sphere_to_headshape->[fit_sphere_to_headshape],make_bem_model->[_surfaces_to_bem],_one_step->[_compose_linear_fitting_data],_check_surfaces->[_assert_complete_surface,_assert_inside],write_bem_solution->[_write_bem_surfaces_block],_fwd_bem_linear_collocation_solution->[_fwd_bem_lin_pot_coeff,_fwd_bem_ip_modify_solution,_fwd_bem_homog_solution,_fwd_bem_multi_solution],_check_origin->[fit_sphere_to_headshape],read_bem_solution->[ConductorModel,read_bem_surfaces],_fwd_bem_lin_pot_coeff->[_correct_auto_elements,_lin_pot_coeff],_surfaces_to_bem->[_ico_downsample,_check_surfaces,_check_thicknesses,_order_surfaces,_check_surface_size],make_bem_solution->[ConductorModel,_fwd_bem_linear_collocation_solution],_fwd_bem_homog_solution->[_fwd_bem_multi_solution],_fwd_eeg_fit_berg_scherg->[_compute_linear_parameters,_fwd_eeg_get_multi_sphere_model_coeffs],_compute_linear_parameters->[_compose_linear_fitting_data],convert_flash_mris->[_prepare_env],make_flash_bem->[_prepare_env,_extract_volume_info]]
Create a BEM solution using the linear collocation approach.
Should probably warn that it might not save properly, and to DRY it this can be a `_check_bem_size` private function
@@ -84,6 +84,12 @@ public class NioEventLoopGroup extends MultithreadEventLoopGroup { super(nThreads, executor, selectorProvider, selectStrategyFactory); } + public NioEventLoopGroup(int nThreads, Executor executor, EventExecutorChooserFactory chooserFactory, + final SelectorProvider selectorProvider, + final SelectStrategyFactory selectStrategyFactory) { + super(nThreads, executor, chooserFactory, selectorProvider, selectStrategyFactory); + } + /** * Sets the percentage of the desired amount of time spent for I/O in the child event loops. The default value is * {@code 50}, which means the event loop will try to spend the same amount of time for I/O as for non-I/O tasks.
[NioEventLoopGroup->[setIoRatio->[setIoRatio],rebuildSelectors->[rebuildSelector],newChild->[NioEventLoop,newSelectStrategy],provider]]
Sets the io ratio for all NioEventLoop in this executor.
Will also add these to the other EvnetLoop impls once @trustin reviewed.
@@ -86,6 +86,7 @@ public class Note implements JsonSerializable { /********************************** transient fields ******************************************/ private transient boolean loaded = false; + private transient boolean executionAborted = false; private transient String path; private transient InterpreterFactory interpreterFactory; private transient InterpreterSettingManager interpreterSettingManager;
[Note->[snapshotAngularObjectRegistry->[getId],equals->[equals],hashCode->[hashCode],insertParagraph->[fireParagraphCreateEvent],toJson->[toJson],clearParagraphOutput->[clearParagraphOutputFields],setRunning->[getInfo],run->[getParagraph,run],populateParagraphInfo->[getId],addCloneParagraph->[getId],clearPersonalizedParagraphOutput->[clearParagraphOutputFields],runAll->[getId],removeParagraph->[fireParagraphRemoveEvent,getId],clearUserParagraphs->[clearUserParagraphs],moveParagraph->[moveParagraph],setPath->[getName],completion->[getParagraph,completion],generateId->[generateId],setCronSupported->[isCronSupported],getUserNote->[getName,getId,Note,getAngularObjects],removeAllAngularObjectInParagraph->[getId],convertOldInput->[convertOldInput],isLastParagraph->[getId],fromJson->[fromJson]]]
Creates a new Note object.
oh, it's non-volatile object. It doesnt looks like thread safe code. You should use volatile w synchronized or something like java.util.concurrent.atomic.AtomicBoolean.
@@ -307,6 +307,10 @@ public class DefaultMuleContext implements MuleContextWithRegistry, PrivilegedMu lifecycleStrategy.initialise(this); + // if there is a correlation id generator, check validity. If there is an error in the generation + // we want to catch it in deploy time, and avoid waiting for the first event to appear in runtime + getConfiguration().getDefaultCorrelationIdGenerator().ifPresent(generator -> generator.validateGenerator()); + } catch (InitialisationException e) { dispose(); throw e;
[DefaultMuleContext->[fireNotification->[fireNotification],isInitialising->[isInitialising],isStarting->[isStarting],isDisposing->[isDisposing],disposeManagers->[dispose],isDisposed->[isDisposed],stop->[stop],initialise->[initialise],dispose->[dispose,stop],start->[start],isStopped->[isStopped],isStopping->[isStopping],handleException->[handleException],DefaultLifecycleStrategy->[initialise->[initialise]],getClusterNodeId->[getClusterNodeId],getClusterId->[getClusterId],overrideClusterConfiguration->[getClusterNodeId],isInitialised->[isInitialised]]]
Initialises the context.
add a TODO to move this validation to the validations, when the expressionManager is available there.
@@ -45,6 +45,13 @@ func maxUint64(a, b uint64) uint64 { return b } +func maxFloat64(a, b float64) float64 { + if a > b { + return a + } + return b +} + func minDuration(a, b time.Duration) time.Duration { if a < b { return a
[StdDeviation->[GetScore,Mean,GetPairs,Len],Mean->[Len],GetMin->[Sort],Sort->[Sort],GetPairs,Add,SetScore,Sort,GetStoreID,GetScore,Mean,GetMin]
requires that the caller has read lock adjustTolerantRatio returns the maximum number of regions in the cluster that are in the.
`math.Max` in standard lib
@@ -414,6 +414,9 @@ class RepositoryCache(RepositoryNoCache): Caches Repository GET operations using a local temporary Repository. """ + # maximum object size that will be cached, 64 kiB. + THRESHOLD = 2**16 + def __init__(self, repository): super().__init__(repository) tmppath = tempfile.mkdtemp(prefix='borg-tmp')
[cache_if_remote->[RepositoryNoCache,RepositoryCache],RemoteRepository->[destroy->[call],delete->[call],break_lock->[call],__len__->[call],load_key->[call],save_key->[call],commit->[call],get_many->[call_many],call_many->[PathNotAllowed,InvalidRPCMethod,fetch_from_cache,ConnectionClosed,RPCError],__init__->[ConnectionClosedWithHint],close->[close],put->[call],check->[call],list->[call],rollback->[call]],RepositoryCache->[close->[destroy],get_many->[get,get_many,put],__init__->[__enter__]],RepositoryServer->[open->[PathNotAllowed],serve->[InvalidRPCMethod]],RepositoryNoCache->[__exit__->[close],get->[get_many],get_many->[get_many]]]
Initialize a cache repository with a temporary file.
ok, so this caches just small files and item metadata chunks (always, as max chunk size for metadata is `2**16`), but no file content data for files `>2**16` (compressed) length. so, this will re-request content chunks from remote (even popular chunks like all-zero) all the time? hmm, if compression is on, that will save us at least for the all-x cases.
@@ -2787,7 +2787,9 @@ def _plot_masked_image(ax, data, times, mask=None, yvals=None, linewidths=[.75], corner_mask=False, antialiased=False, levels=[.5]) time_lims = times[[0, -1]] - ylim = yvals[0], yvals[-1] + 1 + time_lims = [extent[0], extent[1]] + if ylim is None: + ylim = [extent[2], extent[3]] ax.set_xlim(time_lims[0], time_lims[-1]) ax.set_ylim(ylim)
[_annotation_radio_clicked->[_get_active_radiobutton],tight_layout->[tight_layout],_handle_change_selection->[_set_radio_button],_annotate_select->[_get_active_radiobutton],_onclick_new_label->[_setup_annotation_fig,_set_annotation_radio_button],_helper_raw_resize->[_layout_figure],_mouse_click->[_plot_raw_time,_handle_change_selection],_setup_ax_spines->[log_fix],_change_channel_group->[_set_radio_button],_setup_butterfly->[_radio_clicked,_setup_browser_offsets,_get_active_radiobutton],_annotation_modify->[_plot_annotations,_remove_segment_line],_change_annotation_description->[_onclick_new_label],_plot_sensors->[plt_show],ClickableImage->[plot_clicks->[plt_show],__init__->[plt_show]],_process_times->[_find_peaks],_onclick_help->[plt_show,_get_help_text,figure_nobar],_setup_annotation_colors->[_get_color_list],DraggableLine->[on_motion->[drag_callback]],_draw_proj_checkbox->[plt_show],_toggle_proj->[_events_off],_setup_annotation_fig->[plt_show,figure_nobar],_plot_raw_onkey->[_channels_changed,_change_channel_group,_plot_raw_time],_check_cov->[_check_sss],_set_radio_button->[_radio_clicked],_triage_rank_sss->[_check_sss],_select_bads->[_find_channel_idx,_draw_vert_line,_handle_topomap_bads]]
Plot a potentially masked image. Compute a color - based on the data. Draw a critical block. missing points in the image.
forgot to delete the first definition of `time_lims`
@@ -33,7 +33,10 @@ namespace Content.Server.GameObjects.Components.Mobs.State if (entity.TryGetComponent(out IPhysBody? physics)) { - physics.CanCollide = false; + // remove mob from layers where they can be hit by ranged weapon + // other layers will stay unchanged + foreach (var fixture in physics.Fixtures) + fixture.CollisionLayer &= (int) ~CollisionGroup.WeaponMask; } }
[DeadMobState->[ExitState->[ExitState],EnterState->[EnterState]]]
Override state to handle dead damage.
Should get a specific fixture IMO.
@@ -291,7 +291,7 @@ func (h *IdentifyHandler) resolveIdentifyImplicitTeamDoIdentifies(ctx context.Co eng := engine.NewIdentify2WithUID(h.G(), &id2arg) m := libkb.NewMetaContext(subctx, h.G()).WithUIs(uis) err := engine.RunEngine2(m, eng) - idRes := eng.Result() + idRes, _ := eng.Result() if err != nil { h.G().Log.CDebugf(subctx, "identify failed (IDres %v, TrackBreaks %v): %v", idRes != nil, idRes != nil && idRes.TrackBreaks != nil, err) if idRes != nil && idRes.TrackBreaks != nil {
[DisplayTLFCreateWithInvite->[newContext,DisplayTLFCreateWithInvite],DisplayCryptocurrency->[newContext,DisplayCryptocurrency],DisplayKey->[newContext,DisplayKey],DisplayUserCard->[newContext,DisplayUserCard],IdentifyLite->[IdentifyLite],FinishWebProofCheck->[FinishWebProofCheck,newContext],DisplayTrackStatement->[newContext,DisplayTrackStatement],Cancel->[newContext,Cancel],Finish->[newContext,Finish],Start->[Start,newContext],FinishSocialProofCheck->[newContext,FinishSocialProofCheck],ReportLastTrack->[newContext,ReportLastTrack],ReportTrackToken->[newContext,ReportTrackToken],Dismiss->[newContext,Dismiss],Confirm->[Confirm],LaunchNetworkChecks->[newContext,LaunchNetworkChecks]]
Resolve the implicit assertions in a given session. identify returns a response with a masked error if any.
it seems like you should check the error here
@@ -60,15 +60,15 @@ func createTargetMap(targets []resource.URN) map[resource.URN]bool { // checkTargets validates that all the targets passed in refer to existing resources. Diagnostics // are generated for any target that cannot be found. The target must either have existed in the stack // prior to running the operation, or it must be the urn for a resource that was created. -func (pe *planExecutor) checkTargets(targets []resource.URN) result.Result { +func (pe *planExecutor) checkTargets(targets []resource.URN, op StepOp) result.Result { if len(targets) == 0 { return nil } olds := pe.plan.olds var news map[resource.URN]bool - if pe.stepGen != nil && pe.stepGen.creates != nil { - news = pe.stepGen.creates + if pe.stepGen != nil { + news = pe.stepGen.urns } hasUnknownTarget := false
[retirePendingDeletes->[reportExecResult],Execute->[reportExecResult,checkTargets,reportError],refresh->[reportExecResult,checkTargets]]
checkTargets checks if the given targets are not in the stack and if so it will return.
This fixes a bug that caused a "resource not found" error if a specified target was replaced rather than created or updated.
@@ -18,6 +18,7 @@ import java.util.concurrent.TimeoutException; * @author Kohsuke Kawaguchi * @since 1.557 */ +@Deprecated public abstract class InterceptingExecutorService extends ForwardingExecutorService { private final ExecutorService base;
[InterceptingExecutorService->[invokeAny->[invokeAny,wrap],execute->[execute,wrap],invokeAll->[invokeAll,wrap],wrap->[wrap],submit->[wrap,submit]]]
Creates a ExecutorService that wraps all the tasks that run inside. Invokes the given tasks in parallel until either the last task in the collection is executed or the.
`@Restricted(NoExternalUse.class) @RestrictedSince("TODO")`? Then plugins cannot trivially ignore a deprecation warning.
@@ -41,7 +41,8 @@ OUTS_MAP = { # so when a few types of outputs share the same name, we only need # specify it once. CHECKSUM_SCHEMA = { - schema.Optional(RemoteLOCAL.PARAM_CHECKSUM): schema.Or(str, None), + schema.Optional(utils.CHECKSUM_MD5): schema.Or(str, None), + schema.Optional(utils.CHECKSUM_SHA256): schema.Or(str, None), schema.Optional(RemoteS3.PARAM_CHECKSUM): schema.Or(str, None), schema.Optional(RemoteHDFS.PARAM_CHECKSUM): schema.Or(str, None), }
[_get->[Remote,get_remote_settings,supported,o,OutputLOCAL,urlparse],loadd_from->[append,_get,pop],loads_from->[append,_get],copy,Optional,Or]
Get the base output of a single n - tuple. Get a single object from the output.
should we move hashing related stuff to a separate module? so, that it looks like `hash.CHECKSUM_MD5`. `utils` should contain really generic stuff, everything else should be moving into packages with some meaningful names
@@ -94,6 +94,8 @@ var migrations = []Migration{ NewMigration("rewrite authorized_keys file via new format", useNewPublickeyFormat), // v21 -> v22 NewMigration("generate and migrate wiki Git hooks", generateAndMigrateWikiGitHooks), + // v23 -> v24 + NewMigration("generate and migrate repo and wiki Git hooks", generateAndMigrateGitHookChains), } // Migrate database to current version
[Migrate->[migrate],IsFile,Dir,Commit,Exec,Find,Close,GetRandomString,SQL,IsZero,Info,WriteFile,Migrate,Rollback,Error,Marshal,MustInt64,Description,DeleteSection,NewV4,SaveTo,Errorf,And,Bytes,Trace,Sync2,SaveToIndent,Join,Unix,ToStr,Contains,Id,Where,Get,Split,NewSession,MkdirAll,Query,Update,InsertOne,Begin,Sprintf,Sync,Unmarshal,TrimPrefix,Replace,Fatal,Insert,String,WriteString,StrTo,Rename,Load,IsExist]
NewMigration migrations to current version currentVersion. ID = 0 ;.
No migrations can be allowed in 1.1 branch
@@ -66,9 +66,10 @@ def discover_plugins() -> Iterable[str]: """ Returns an iterable of the plugins found. """ - for module_info in discover_namespace_plugins(): - yield module_info.name - yield from discover_file_plugins() + with push_python_path("."): + for module_info in discover_namespace_plugins(): + yield module_info.name + yield from discover_file_plugins() def import_plugins() -> None:
[discover_plugins->[discover_file_plugins,discover_namespace_plugins],import_plugins->[discover_plugins]]
Returns an iterable of all plugins found.
What does this do when I run something from some unexpected working directory?
@@ -3,7 +3,8 @@ declare var JitsiMeetJS: Object; declare var APP: Object; -import uuid from 'uuid'; +import 'react-native-get-random-values'; +import { v4 as uuidv4 } from 'uuid'; import { getDialOutStatusUrl, getDialOutUrl, updateConfig } from '../base/config'; import { isIosMobileBrowser } from '../base/environment/utils';
[No CFG could be retrieved]
Imports a single object. Status of a network connection.
Does this run on web?
@@ -392,10 +392,17 @@ void commit_params(struct dt_iop_module_t *self, dt_iop_params_t *p1, dt_dev_pix float DT_ALIGNED_ARRAY input_matrix[3][4]; mat3mul4((float *)input_matrix, (float *)XYZ_to_gradingRGB, (float *)RGB_to_XYZ); + // Test white point of the current space in grading RGB + const float white_pipe_RGB[4] = { 1.f, 1.f, 1.f }; + float white_grading_RGB[4] = { 0.f }; + dot_product(white_pipe_RGB, input_matrix, white_grading_RGB); + const float sum_white = white_grading_RGB[0] + white_grading_RGB[1] + white_grading_RGB[2]; + for(size_t c = 0; c < 4; c++) white_grading_RGB[c] /= sum_white; + // make RGB values vary between [0; 1] in working space, convert to Ych and get the max(c(h))) #ifdef _OPENMP #pragma omp parallel for default(none) \ - dt_omp_firstprivate(input_matrix) schedule(static) dt_omp_sharedconst(LUT) + dt_omp_firstprivate(input_matrix, white_grading_RGB) schedule(static) dt_omp_sharedconst(LUT) #endif for(size_t r = 0; r < STEPS; r++) for(size_t g = 0; g < STEPS; g++)
[No CFG could be retrieved]
dt_iop_order_iccprofile_info_t * d = self - - - - - - - - - - - - - - - - - -.
for_four_channels(c) white_grading_RGB[c] /= sum_white;
@@ -477,8 +477,9 @@ class Invoices extends DolibarrApi * * @param int $id Id of invoice to update * @param int $rowid Row key of the contact in the array contact_ids. + * @param string $type Type of the contact (BILLING, SHIPPING, CUSTOMER). * - * @url DELETE {id}/contact/{rowid} + * @url DELETE {id}/contact/{rowid}/{type} * * @return array *
[Invoices->[validate->[validate],putLine->[get],put->[get],delete->[delete],deleteLine->[get]]]
Add a contact to an invoice Delete a line of a given invoice.
Why adding this parameter $type ? Delete should receive the rowid to delete to be able to delete ?
@@ -32,13 +32,17 @@ module GobiertoAdmin def gobierto_budgets_params params.require(:gobierto_budgets_options).permit(:elaboration_enabled, :budget_lines_feedback_enabled, :feedback_emails, :receipt_enabled, :receipt_configuration, :comparison_tool_enabled, :comparison_context_table_enabled, :comparison_show_widget, :comparison_compare_municipalities_enabled, - :providers_enabled, :indicators_enabled, comparison_compare_municipalities: []) + :providers_enabled, :indicators_enabled, :budgets_guide_page, + comparison_compare_municipalities: []) end def get_services_config OpenStruct.new(APP_CONFIG["services"]) end + def get_available_pages + @site.pages.active.sorted + end end end end
[OptionsController->[update->[redirect_to,new,save,merge,t,to_sentence],get_services_config->[new],update_annual_data->[redirect_to,perform_later,find,object,t,id],index->[new],gobierto_budgets_params->[permit],module_enabled!,before_action,module_allowed!]]
Get the configuration parameters for the gobierto budgets.
Do not prefix reader method names with get_.
@@ -15,7 +15,7 @@ from conans.util.log import logger def export_pkg(conanfile, package_id, src_package_folder, package_folder, hook_manager, conanfile_path, ref): mkdir(package_folder) - conanfile.package_folder = src_package_folder + conanfile.package_folder = package_folder output = conanfile.output output.info("Exporting to cache existing package from user folder") output.info("Package folder %s" % package_folder)
[update_package_metadata->[update_metadata],export_pkg->[create,_report_files_from_manifest,FileCopier,copier,dumps,join,mkdir,info,success,save,execute],_report_files_from_manifest->[list,report_copied_files,files,warn,remove],_create_aux_files->[create,debug,ConanException,copy,join,dumps,save],create_package->[ConanException,conanfile_exception_formatter,package,FileCopier,warn,info,rmdir,success,_create_aux_files,highlight,error,_report_files_from_manifest,ScopedOutput,isinstance,execute,basename,str,chdir,mkdir]]
Export a package from a folder to a folder in a conan.
We will break someone.
@@ -100,9 +100,6 @@ func (s *testLeaderServerSuite) TestLeader(c *C) { svr.Close() delete(s.svrs, leader1.GetAddr()) - // now, another two servers must select a leader - s.setUpClient(c) - // wait leader changes for i := 0; i < 50; i++ { leader, _ := getLeader(s.client, s.leaderPath)
[SetUpSuite->[getLeaderPath,setUpClient,Assert],TestLeader->[setUpClient,Close,Assert,GetAddr,Sleep,Run],setUpClient->[New,Assert,GetEndpoints],TearDownSuite->[Close],Assert]
TestLeader tests that the server is the leader.
Should we remove any other redundant codes?
@@ -0,0 +1,18 @@ +class CreateDbActivities < ActiveRecord::Migration + def change + create_table :db_activities do |t| + t.integer :user_id, null: false + t.integer :recipient_id + t.string :recipient_type + t.integer :trackable_id, null: false + t.string :trackable_type, null: false + t.string :action, null: false + t.timestamps null: false + end + + add_index :db_activities, [:recipient_id, :recipient_type] + add_index :db_activities, [:trackable_id, :trackable_type] + + add_foreign_key :db_activities, :users + end +end
[No CFG could be retrieved]
No Summary Found.
Unnecessary spacing detected.<br>Put one space between the method name and the first argument.
@@ -186,6 +186,7 @@ export const StyledSingleValue = styled('div', props => { whiteSpace: 'nowrap', maxWidth: '100%', ...getControlPadding($size, sizing, $type), + ...ellipsisText(), }; });
[No CFG could be retrieved]
Create styled components for the n - node control. Create a styled input with the given props.
maybe in-line if the properties are not dynamic?
@@ -32,7 +32,7 @@ public class DeclarativeAutoConfigTest extends AbstractInfinispanTest { assertEquals(properties.getProperty("hibernate.search.default.exclusive_index_use"), "true"); assertEquals(properties.getProperty("hibernate.search.default.reader.strategy"), "shared"); assertEquals(properties.getProperty("hibernate.search.default.indexmanager"), "near-real-time"); - assertEquals(properties.getProperty("hibernate.search.backends.infinispan_backend.directory.type"), "filesystem"); + assertEquals(properties.getProperty("hibernate.search.backends.infinispan_backend.directory.type"), "local-filesystem"); cacheConfiguration = cm.getCacheConfiguration("dist-with-default"); properties = cacheConfiguration.indexing().properties();
[DeclarativeAutoConfigTest->[testAutoConfig->[call->[getProperty,getCacheConfiguration,properties,isEmpty,assertFalse,assertEquals],withCacheManager,fromXml,CacheManagerCallable]]]
TestAutoConfig test.
could we put these strings into constants ?
@@ -451,6 +451,7 @@ type KubernetesMasterConfig struct { ServicesSubnet string `json:"servicesSubnet"` StaticNodeNames []string `json:"staticNodeNames"` SchedulerConfigFile string `json:"schedulerConfigFile"` + PodEvictionTimeout string `json:"podEvictionTimeout"` } type CertInfo struct {
[No CFG could be retrieved]
TenantConfig is a list of all the possible components of the system that are expected to be.
need to update types_test.go to include the new field, and open a doc and config change issue with details
@@ -484,6 +484,7 @@ func ViewPullFiles(ctx *context.Context) { return } + ctx.Data["IsOwner"] = ctx.Repo.IsWriter() || (ctx.IsSigned && issue.IsPoster(ctx.User.ID)) ctx.Data["IsImageFile"] = commit.IsImageFile ctx.Data["SourcePath"] = setting.AppSubURL + "/" + path.Join(headTarget, "src", "commit", endCommitID) ctx.Data["BeforeSourcePath"] = setting.AppSubURL + "/" + path.Join(headTarget, "src", "commit", startCommitID)
[NumFiles,Status,GetIssueByIndex,GetHeadRepo,Merge,Redirect,IsErrRepoNotExist,GetOwner,GetPatch,IsErrIssueNotExist,ValidateCommitsWithEmails,GetRepositoryByID,IsErrPullRequestNotExist,GetAssignees,Len,IsErrReviewNotExist,IsErrUserDoesNotHaveAccessToRepo,ServerError,ParamsInt64,ToStr,GetBaseRepo,GetWorkInProgressPrefix,IsWriterOfRepo,CanBeForked,GetUserByName,Split,IsErrRepoAlreadyExist,QueryUnescape,GetDiffRange,GetCommit,AddDeletePRBranchComment,Add,RenderWithErr,RepoPath,Copy,ParseCommitsWithSignature,Link,LoadComments,IsErrNameReserved,JSON,IsOrganization,GetDiffRangeWithWhitespaceBehavior,GetPullRequestByIssueID,NotifyIssue,IsOwnedBy,IsBranchExist,IsErrUserNotExist,ParseCommitsWithStatus,HasForkedRepo,GetRefCommitID,PatchPath,Success,IssueNoDependenciesLeft,GetUnmergedPullRequest,GetUserByID,GetBranches,AddTestPullRequestTask,NotFound,IsWorkInProgress,GetBranchCommitID,Trace,TrimSpace,Tr,AssignForm,ServeFileContent,GetDefaultMergeMessage,CanAutoMerge,GetOwnedOrganizations,Query,MergeStyle,HasAccess,GetCurrentReview,Sprintf,IsProtectedBranch,PushToBaseRepo,QueryInt64,NextIssueIndex,GetFormatPatch,IsErrInvalidMergeStyle,DeleteBranch,HTML,Error,OpenRepository,GetPullRequestInfo,Errorf,HasError,ReadBy,ForkRepository,Join,Written,GetGitRefName,IsErrNamePatternNotAllowed,Contains,EncodeMD5,GetDefaultSquashMessage,NewPullRequest,Params]
Get a specific commit in the branch. MergePullRequest returns response for merging pull request.
Should this just be `ctx.IsSigned && issue.IsPoster(ctx.User.ID)`
@@ -249,6 +249,12 @@ final class XmlExtensionLoaderDelegate { } } + private XmlDslModel getXmlDslModel(ComponentModel moduleModel, String name, String version) { + final Optional<String> prefix = ofNullable(moduleModel.getParameters().get(MODULE_PREFIX_ATTRIBUTE)); + final Optional<String> namespace = ofNullable(moduleModel.getParameters().get(MODULE_NAMESPACE_ATTRIBUTE)); + return createXmlLanguageModel(prefix, namespace, name, version); + } + private String getDescription(ComponentModel componentModel) { return componentModel.getParameters().getOrDefault(DOC_DESCRIPTION, ""); }
[XmlExtensionLoaderDelegate->[extractOperationExtension->[getDescription],loadCustomTypes->[getResource],extractOperationParameters->[getRole],extractType->[getCustomTypeFilename],loadPropertiesFrom->[extractGlobalElementsFrom],getResource->[getResource]]]
loadModuleExtension - Load extension of a module. Checks if there is a child with the same identifier as the current one.
just read identical code in JavaXmlDeclarationEnricher.java. Doesn't look like doing the same here is necessary at all
@@ -40,9 +40,15 @@ import org.springframework.stereotype.Component; import java.io.FileNotFoundException; import java.io.IOException; +import java.util.ArrayList; import java.util.Arrays; +import java.util.Collections; +import java.util.Comparator; +import java.util.HashMap; +import java.util.HashSet; import java.util.List; import java.util.Map; +import java.util.Set; /** * CLI commands to operate on the Metadata Table.
[MetadataCommand->[init->[getMetadataTableBasePath,create],create->[getMetadataTableBasePath,create],stats->[getMetadataTableBasePath,stats],set->[setMetadataBaseDirectory],getMetadataTableBasePath->[getMetadataTableBasePath],delete->[delete,getMetadataTableBasePath]]]
Package for Hudi - CLI This is a utility method that is used to set the base directory for the Metadata Table.
Can you add an example here. for eg: connect set spark env variables sample metadata command.