patch stringlengths 18 160k | callgraph stringlengths 4 179k | summary stringlengths 4 947 | msg stringlengths 6 3.42k |
|---|---|---|---|
@@ -575,7 +575,7 @@ insert_entry(daos_handle_t oh, daos_handle_t th, const char *name, size_t len,
if (rc) {
/** don't log error if conditional failed */
if (rc != -DER_EXIST)
- D_ERROR("Failed to insert entry %s (%d)\n", name, rc);
+ D_ERROR("Failed to insert entry %s, "DF_RC"\n", name, DP_RC(rc));
return daos_der2errno(rc);
}
| [dfs_mount_root_cont->[dfs_cont_create,dfs_mount],dfs_umount_root_cont->[dfs_umount],dfs_obj_local2global->[dfs_get_chunk_size],dfs_chmod->[dfs_release],dfs_access->[dfs_release]] | This function is called from the init function of the dfs_entry_t. insert a new entry in the list of entries. | (style) line over 80 characters |
@@ -12,6 +12,8 @@ namespace System.Data.Common
private readonly DbSchemaTable _schemaTable;
private readonly DataRow _dataRow;
+ [UnconditionalSuppressMessage("ReflectionAnalysis", "IL2026:UnrecognizedReflectionPattern",
+ Justification = "Filter expression is null.")]
internal static DbSchemaRow[] GetSortedSchemaRows(DataTable dataTable, bool returnProviderSpecificTypes)
{
DataColumn? sortindex = dataTable.Columns[SchemaMappingUnsortedIndex];
| [DbSchemaRow->[GetSortedSchemaRows->[ModifiedCurrent,Length,Count,Columns,Added,Select,Rows,Assert,Add,Unchanged],Size,IsUnique,IsRowVersion,IsReadOnly,BaseSchemaName,IsExpression,BaseTableName,IsHidden,Empty,DataType,UnsortedIndex,Assert,ColumnName,BaseServerName,ToInt32,Default,IsAutoIncrement,AllowDBNull,IsKey,ToString,IsDBNull,BaseColumnName,BaseCatalogName,ToBoolean,IsLong,InvariantCulture]] | Get sorted schema rows. | I don't understand this justification. Maybe it would be better to refactor the troubled code into a separate method and just suppress the one spot that is being warned about. |
@@ -89,10 +89,10 @@ class AbstractCrowdsourcingTest:
self.config = compose(
config_name="example",
overrides=[
- f'+mephisto.blueprint._blueprint_type={blueprint_type}',
- f'+mephisto.blueprint.link_task_source=False',
- f'+mephisto/architect=mock',
- f'+mephisto/provider=mock',
+ f'mephisto.blueprint._blueprint_type={blueprint_type}',
+ f'++mephisto.blueprint.link_task_source=False',
+ f'mephisto/architect=mock',
+ f'mephisto/provider=mock',
f'+task_dir={task_directory}',
f'+current_time={int(time.time())}',
]
| [AbstractParlAIChatTest->[_test_agent_states->[_register_mock_agents]],AbstractOneTurnCrowdsourcingTest->[_get_agent_state->[_register_mock_agents]]] | Set up the configuration and database for a specific . Missing configuration entry. | Ooof, fiddling with the pluses is painful -__- Hmm what does the double-plus mean? Does that allow for either adding *or* overwriting the param? |
@@ -31,10 +31,10 @@ def details(request, token):
orders = orders.select_related(
'billing_address', 'shipping_address', 'user')
order = get_object_or_404(orders, token=token)
- notes = order.notes.filter(is_public=True)
+ user = request.user if request.user.is_authenticated else None
+ notes = order.notes.filter(user=user)
ctx = {'order': order, 'notes': notes}
if order.is_open():
- user = request.user if request.user.is_authenticated else None
note = OrderNote(order=order, user=user)
note_form = OrderNoteForm(request.POST or None, instance=note)
ctx.update({'note_form': note_form})
| [checkout_success->[authenticate,PasswordForm,save,attach_order_to_user,get,update,redirect,copy,login,filter,LoginForm,is_valid,TemplateResponse,get_object_or_404],connect_order_with_user->[pgettext_lazy,attach_order_to_user,get,redirect,success,warning],payment_success->[reverse,redirect],payment->[all,is_fully_paid,PaymentMethodsForm,is_pre_authorized,redirect,get,confirmed,prefetch_related,TemplateResponse,PaymentDeleteForm,is_valid,select_related,get_object_or_404],cancel_payment->[save,redirect,is_valid,atomic,PaymentDeleteForm,HttpResponseForbidden],start_payment->[TemplateResponse,get_client_ip,pgettext_lazy,exception,error,get_form,redirect,Http404,str,filter,get_or_create,atomic,change_status],details->[TemplateResponse,is_open,redirect,update,OrderNote,filter,confirmed,is_valid,save,OrderNoteForm,select_related,get_object_or_404],getLogger] | Show a list of orders and notes. | I see some trouble here, let's say if staff user added some note, and his account was deleted after, then `OrderNote` he added will be preserved, but `user` will be set to `None`. This may cause some data leakage. I wonder if we shouldn't leave `is_public` on the `OrderNote` to determine if the note was added by the customer or not, without a possibility to create a public note from the dashboard Or shall we leave `OrderNote` to customers only, and introduce another way for staff to add notes about orders/ fulfillments etc. cc @maarcingebala @NyanKiyoshi |
@@ -1302,6 +1302,8 @@ def test_epoch_eq():
cond1, cond2 = ['a', ['b', 'b/y']], [['a/x', 'a/y'], 'x']
for c in (cond1, cond2): # error b/c tag and id mix/non-orthogonal tags
assert_raises(ValueError, epochs.equalize_event_counts, c)
+ assert_raises(ValueError, epochs.equalize_event_counts,
+ ["a/no_match", "b"])
def test_access_by_name():
| [test_epochs_bad_baseline->[_get_data],test_drop_channels_mixin->[_get_data],test_event_ordering->[_get_data],test_resample->[_get_data],test_detrend->[_get_data],test_illegal_event_id->[_get_data],test_epochs_proj->[_get_data],test_drop_epochs_mult->[_get_data],test_epoch_eq->[_get_data],test_decim->[_get_data],test_epoch_combine_ids->[_get_data],test_base_epochs->[_get_data],test_reject->[_get_data],test_evoked_arithmetic->[_get_data],test_to_data_frame->[_get_data],test_delayed_epochs->[_get_data],test_access_by_name->[_get_data],test_add_channels_epochs->[make_epochs,_get_data],test_preload_epochs->[_get_data],test_epochs_hash->[_get_data],test_subtract_evoked->[_get_data],test_read_write_epochs->[_get_data],test_reject_epochs->[_get_data],test_epochs_proj_mixin->[_get_data],test_evoked_io_from_epochs->[_get_data],test_pick_channels_mixin->[_get_data],test_comparision_with_c->[_get_data],test_concatenate_epochs->[_get_data],test_equalize_channels->[_get_data],test_savgol_filter->[_get_data],test_add_channels->[_get_data],test_epochs_copy->[_get_data],test_drop_epochs->[_get_data],test_indexing_slicing->[_get_data],test_epoch_multi_ids->[_get_data],test_read_epochs_bad_events->[_get_data],test_crop->[_get_data],test_bootstrap->[_get_data],test_iter_evoked->[_get_data],test_evoked_standard_error->[_get_data],test_contains->[_get_data]] | Test epoch count equality and condition combining. rejects missing missing keys in epochs no - op if new_shapes == 2 Test access by event name and event count. find all epochs with a specific event_id Check if all epochs are equal and have the same name. | can you actually differentiate both conditionals by catching the error and testing for the according message? would be good to have specific tests. |
@@ -16,8 +16,14 @@ namespace Microsoft.Xna.Framework.Graphics
private bool _isDisposed;
private BlendState _blendState;
- private BlendState _actualBlendState;
- private DepthStencilState _depthStencilState = DepthStencilState.Default;
+ private BlendState _actualBlendState;
+
+ private DepthStencilState _depthStencilState;
+ private DepthStencilState _actualDepthStencilState;
+
+ private DepthStencilState _depthStencilStateDefault;
+ private DepthStencilState _depthStencilStateDepthRead;
+ private DepthStencilState _depthStencilStateNone;
private RasterizerState _rasterizerState;
private RasterizerState _actualRasterizerState;
| [GraphicsDevice->[ApplyRenderTargets->[Clear],DrawUserPrimitives->[DrawUserPrimitives],Dispose->[Dispose],SetRenderTarget->[SetRenderTarget],SetIndexBuffer,Initialize]] | Construct a new GraphicsDevice object that represents a single . The active vertex shader for a specific object. | Do you mind cleaning up the blend state declarations? Like here you have the `_blendState` and `_actualBlendState`, but have all the `_blendStateAdditive`/`_blendStateAlphaBlend`/etc down around line 300. |
@@ -330,7 +330,8 @@ main(int argc, char **argv)
{"subtests", required_argument, NULL, 'u'},
{"exclude", required_argument, NULL, 'E'},
{"filter", required_argument, NULL, 'f'},
- {"dfs", no_argument, NULL, 'F'},
+ {"dfs_unit", no_argument, NULL, 'I'},
+ {"dfs_funct", no_argument, NULL, 'F'},
{"work_dir", required_argument, NULL, 'W'},
{"workload_file", required_argument, NULL, 'w'},
{"help", no_argument, NULL, 'h'},
| [No CFG could be retrieved] | This function returns a list of all the methods that are defined in the spec. Initialization of the n - tuple. | (style) please, no space before tabs |
@@ -84,6 +84,7 @@ class Feature
['multi_profiles', L10n::t('Multiple Profiles'), L10n::t('Ability to create multiple profiles'), false, Config::get('feature_lock', 'multi_profiles', false)],
['photo_location', L10n::t('Photo Location'), L10n::t("Photo metadata is normally stripped. This extracts the location \x28if present\x29 prior to stripping metadata and links it to a map."), false, Config::get('feature_lock', 'photo_location', false)],
['export_calendar', L10n::t('Export Public Calendar'), L10n::t('Ability for visitors to download the public calendar'), false, Config::get('feature_lock', 'export_calendar', false)],
+ ['trending_tags', L10n::t('Trending Tags'), L10n::t('Show a community page widget with a list of the most popular tags in recent public posts.'), false, Config::get('feature_lock', 'trending_tags', false)],
],
// Post composition
| [No CFG could be retrieved] | Get all the items in the configuration. Filters all network posts based on a list of selected protocols. get all items in an array. | I guess the artist who started the indentation of the `L10n` to be aligned would be happy if you could align your new entries with the rest of the list. |
@@ -201,6 +201,12 @@ class Product(models.Model, ItemRange, index.Indexed):
def set_attribute(self, pk, value_pk):
self.attributes[smart_text(pk)] = smart_text(value_pk)
+ def get_price_range(self, **kwargs):
+ if not self.variants.exists():
+ return PriceRange(self.price, self.price)
+ else:
+ return super(Product, self).get_price_range(**kwargs)
+
@python_2_unicode_compatible
class ProductVariant(models.Model, Item):
| [ProductVariant->[get_absolute_url->[get_slug],get_cost_price->[select_stockrecord],as_data->[get_price_per_item],get_first_image->[get_first_image]],Category->[CategoryManager],Product->[is_in_stock->[is_in_stock],ProductManager],Stock->[StockManager]] | Set the value of an attribute. | This needs to mimic the discount logic from `ProductVariant.get_unit_price` (even if it may seem silly to care about discounts for unavailable products) |
@@ -135,7 +135,6 @@ class BookKeeping extends CommonObject
/**
* @var float FEC:Amount (Not necessary)
- * @deprecated No more used
*/
public $amount;
| [BookKeeping->[transformTransaction->[getNextNumMvt],createFromClone->[fetch,create]]] | Creates a new object from a base - object. Creates an object in the database. | The amount is same value than debit or credit right ? What is rule of sign for this field ? For example for Common invoice: debit= ? credit = ? amount= ? Credit note : debit= ? credit = ? amount= ? |
@@ -418,7 +418,6 @@ class RaidenService(Runnable):
self._initialize_transactions_queues(chain_state)
self._initialize_messages_queues(chain_state)
self._initialize_whitelists(chain_state)
- self._initialize_channel_fees()
self._initialize_monitoring_services_queue(chain_state)
self._initialize_ready_to_process_events()
| [RaidenService->[_callback_new_block->[handle_and_track_state_changes],handle_and_track_state_changes->[add_pending_greenlet],on_message->[on_message],stop->[stop],_start_alarm_task->[start],_start_transport->[start],_initialize_payment_statuses->[PaymentStatus],mediate_mediated_transfer->[mediator_init,handle_and_track_state_changes],set_channel_reveal_timeout->[handle_and_track_state_changes],_initialize_transactions_queues->[add_pending_greenlet,handle_event],sign->[sign],target_mediated_transfer->[target_init,start_health_check_for,handle_and_track_state_changes],set_node_network_state->[handle_and_track_state_changes],withdraw->[handle_and_track_state_changes],_initialize_channel_fees->[handle_event],add_pending_greenlet->[remove->[remove]],start_mediated_transfer_with_secret->[PaymentStatus,handle_and_track_state_changes,matches,initiator_init,start_health_check_for]]] | Start the node synchronously. This method is called when a new block is known. It is called by the raiden Initialize all the components of the Raft service. | I see, you're removing setting channel fees at startup / restart. How do you persist the channel fee settings? How do you make sure, that channel fee settings are (or have been!) communicated to the selected PFS? |
@@ -423,7 +423,7 @@ WasmBinaryReader::ModuleHeader()
ThrowDecodingError(_u("Malformed WASM module header!"));
}
- if (version != 10)
+ if (version != 0xb)
{
ThrowDecodingError(_u("Invalid WASM version!"));
}
| [CheckBytesLeft->[ThrowDecodingError],CallImportNode->[ThrowDecodingError],ReadSectionHeader->[ThrowDecodingError],GetWasmType->[ThrowDecodingError],SLEB128->[LEB128],ReadImportEntries->[ReadInlineName,ThrowDecodingError],ModuleHeader->[ThrowDecodingError],Offset->[ThrowDecodingError],ASTNode->[ThrowDecodingError],ReadExportTable->[ThrowDecodingError],ReadNamesSection->[ThrowDecodingError],va_start->[va_start],CallIndirectNode->[ThrowDecodingError],ReadWasmType->[GetWasmType,ThrowDecodingError],ReadFunctionBodies->[ThrowDecodingError],ReadNextSection->[ReadNextSection,ThrowDecodingError],ReadFunctionsSignatures->[ThrowDecodingError],CallNode->[ThrowDecodingError],ThrowDecodingError->[va_start],char->[ThrowDecodingError],ReadIndirectFunctionTable->[ThrowDecodingError],LEB128->[ThrowDecodingError]] | ModuleHeader - read WASM module header if found. | Changing from decimal 10 (0xA) to 11 (0xB) intentional? |
@@ -353,7 +353,7 @@ func getUpgradeCommand() string {
return "$ brew upgrade pulumi"
}
- if filepath.Dir(exe) != filepath.Join(curUser.HomeDir, ".pulumi", "bin") {
+ if filepath.Dir(exe) != filepath.Join(curUser.HomeDir, workspace.BookkeepingDir, "bin") {
return ""
}
| [StringVar,RawMessage,Dir,Colorize,Executable,Warningf,Now,Encode,HasPrefix,Add,IsTruthy,IgnoreClose,Flush,Stat,ParseTolerant,NewEncoder,New,RunFunc,Diag,StringVarP,NewClient,InitLogging,Run,After,Chdir,AddCommand,LookPath,GetCachedVersionFilePath,Bytes,TrimSpace,AssertNoError,InitProfiling,Join,Wrapf,Infof,Current,EvalSymlinks,ReadString,CloseTracing,V,GetCLIVersionInfo,NewDecoder,StdStreams,CloseProfiling,SetGlobalColorization,GT,ModTime,Flag,Command,NewReader,Sprintf,Decode,IntVarP,DefaultURL,Print,InitTracing,String,Parse,DoWithRetry,OpenFile,Getenv,BoolVarP,PersistentFlags,BoolVar] | getUpgradeMessage returns a message that will be used to upgrade from the current version to the isBrewInstall returns true if the current running executable is on macOS and has a. | Why bookkeepingdir here but homedir elsewhere? |
@@ -34,10 +34,10 @@ func setUp() {
// register sample plugin into test
log.SetLevel(log.DebugLevel)
trace.Logger.Level = log.DebugLevel
- version.MaxPluginVersion = 1
+ version.MaxPluginVersion = version.MaxPluginVersion + 1
if err := manager.Migrator.Register(1, manager.ApplianceConfigure, &plugin1.ApplianceStopSignalRename{}); err != nil {
- log.Errorf("Failed to register plugin %s:%d, %s", manager.ApplianceConfigure, 1, err)
+ log.Errorf("Failed to register plugin %s:%d, %s", manager.ApplianceConfigure, version.MaxPluginVersion, err)
}
}
| [Equal,MapSource,Contains,MapSink,Decode,False,String,Errorf,SetLevel,Fail,Register,Encode,Logf,True] | This function initializes the required components of the virtual container host configuration. Attach is a test function that tests that the session config has a reserved sequence number. | This looks like a critical failure - should be a panic or `t.Fatal(...)` |
@@ -170,7 +170,7 @@ public class MuleMetadataManager implements MetadataManager, Initialisable {
try {
return producer.get();
} catch (InvalidComponentIdException e) {
- return failure(e);
+ return failure(null, e.getMessage(), FailureCode.COMPONENT_NOT_FOUND, e.getDetailedMessage());
} catch (Exception e) {
return failure(null, format("%s: %s", failureMessage, e.getMessage()), e);
}
| [MuleMetadataManager->[getEntityKeys->[getEntityKeys],getEntityMetadata->[getEntityMetadata],getMetadata->[getMetadata],getMetadataKeys->[getMetadataKeys]]] | Exception handled by MetadataFetch. | add a new failure builder that accepts only an exception and an overrided failure code |
@@ -653,6 +653,8 @@ func (p *Properties) setAgentProfileDefaults(isUpgrade, isScale bool, cloudName
} else if profile.Distro == AKS1804Deprecated {
profile.Distro = AKSUbuntu1804
}
+ } else if p.OrchestratorProfile.KubernetesConfig.CustomHyperkubeImage != "" {
+ profile.Distro = Ubuntu
}
// The AKS Distro is not available in Azure German Cloud.
if cloudName == AzureGermanCloud {
| [SetPropertiesDefaults->[setVMSSDefaultsForMasters,setAgentProfileDefaults,setVMSSDefaultsForAgents,setStorageDefaults,SetDefaultCerts,GetCloudSpecConfig,setCustomCloudProfileDefaults,HasVMSSAgentPool,setExtensionDefaults,setHostedMasterProfileDefaults,setMasterProfileDefaults,IsVirtualMachineScaleSets,setWindowsProfileDefaults,setOrchestratorDefaults],setVMSSDefaultsForAgents->[BoolPtr,HasAvailabilityZones],setVMSSDefaultsForMasters->[BoolPtr,HasAvailabilityZones],setAgentProfileDefaults->[BoolPtr,IsKubernetes,Sprintf,IsAzureStackCloud,IsUbuntu1804,IsVHDDistro,IntPtr,IsAzureCNI,AcceleratedNetworkingSupported,IsVirtualMachineScaleSets,Atoi,IsCustomVNET],setStorageDefaults->[IsKubernetesVersionGe],SetDefaultCerts->[Uint32,To4,PutUint32,CreatePkiKeyCertPair,GetCustomCloudName,Split,CidrStringFirstIP,Errorf,CreatePki,IsFeatureEnabled,ParseIP,GetLocations,IsVirtualMachineScaleSets],setMasterProfileDefaults->[GetFirstConsecutiveStaticIPAddress,IsKubernetes,BoolPtr,IsAzureStackCloud,IsVHDDistro,IntPtr,IsVirtualMachineScaleSets,HasWindows,IsUbuntu1804,IsCustomVNET,Atoi,IsAzureCNI],setOrchestratorDefaults->[SetCloudProviderBackoffDefaults,Warnf,setAPIServerConfig,GetMinVersion,setAddonsConfig,setCloudControllerManagerConfig,SetCloudProviderRateLimitDefaults,BoolPtr,IsRBACEnabled,IsKubernetesVersionGe,IsAzureStackCloud,ParseCIDR,HasAvailabilityZones,Make,Bool,To4,setKubeletConfig,Join,TotalNodes,setSchedulerConfig,PrivateJumpboxProvision,GetCloudSpecConfig,ToLower,HasWindows,IsHostedMasterProfile,IsFeatureEnabled,Split,IsAzureCNI,LT,setControllerManagerConfig,GetValidPatchVersion],Strings,Read,TrimSuffix,Sprintf,IsKubernetesVersionGe,String,WriteString,EncodeToString,Split,Trim] | setAgentProfileDefaults sets the default values for all agent profiles This function is used to set the default values for all the configuration options. This function is used to configure the AKS configuration. | we don't want to override a non-empty distro otherwise the instructions in the doc above to set it to 18.04 won't work... this needs to be inside the `if profile.Distro == "" && profile.ImageRef == nil` if statement |
@@ -14,7 +14,7 @@ from typing import Union, Any, TYPE_CHECKING
from uamqp import types # type: ignore
from azure.eventhub import __version__
-from azure.eventhub.configuration import _Configuration
+from .configuration import Configuration
from .common import EventHubSharedKeyCredential, EventHubSASTokenCredential, _Address
try:
| [EventHubClientAbstract->[from_connection_string->[_parse_conn_str]]] | Creates a token object from the given connection string. Create a shared access signiture token as a string literal. | Should `configuration` be made internal? |
@@ -656,9 +656,6 @@ class TextAnalyticsClient(TextAnalyticsClientBase):
:class:`~azure.ai.textanalytics.SentenceSentiment` objects
will have property `mined_opinions` containing the result of this analysis. Only available for
API version v3.1-preview and up.
- :keyword str string_index_type: Specifies the method used to interpret string offsets. Possible values are
- 'UnicodeCodePoint', 'TextElements_v8', or 'Utf16CodeUnit'. The default value is 'UnicodeCodePoint'.
- Only available for API version v3.1-preview and up.
:keyword str language: The 2 letter ISO 639-1 representation of language for the
entire batch. For example, use "en" for English; "es" for Spanish etc.
If not set, uses "en" for English as default. Per-document language will
| [TextAnalyticsClient->[recognize_entities->[_check_string_index_type_arg,process_http_response_error,pop,update,_validate_input,entities_recognition_general],analyze_sentiment->[_check_string_index_type_arg,process_http_response_error,pop,update,str,sentiment,ValueError,_validate_input],extract_key_phrases->[key_phrases,_validate_input,pop,process_http_response_error],begin_analyze_batch_actions->[process_http_response_error,partial,_determine_action_type,begin_analyze,get,pop,str,ValueError,AnalyzeBatchActionsLROPollingMethod,models,to_generated,_validate_input,TextAnalyticsOperationResourcePolling],detect_language->[process_http_response_error,_validate_input,languages,pop],_healthcare_result_callback->[healthcare_paged_result,models,_deserialize],begin_analyze_healthcare_entities->[begin_health,process_http_response_error,partial,get,pop,AnalyzeHealthcareEntitiesLROPollingMethod,str,ValueError,_validate_input,TextAnalyticsOperationResourcePolling],recognize_pii_entities->[_check_string_index_type_arg,entities_recognition_pii,process_http_response_error,pop,update,str,ValueError,_validate_input],__init__->[pop,super,get,_get_deserialize],_analyze_result_callback->[analyze_paged_result,models,_deserialize],recognize_linked_entities->[_check_string_index_type_arg,process_http_response_error,pop,update,_validate_input,entities_linking]]] | Analyzes sentiment for a batch of documents. This function returns a list of all language - specific tags used in the sentiment of a Get a single chunk of the n - ary sentiment index. | we have `string_index_type` duplicated in the docstring so I deleted one |
@@ -1143,7 +1143,7 @@ namespace Js
if (nestedCount > 0)
{
nestedArray = RecyclerNewPlusZ(m_scriptContext->GetRecycler(),
- nestedCount*sizeof(FunctionProxy), NestedArray, nestedCount);
+ (nestedCount-1)*sizeof(FunctionProxy*), NestedArray, nestedCount);
}
else
{
| [No CFG could be retrieved] | Creates a new object of type n_nan and initializes it with the values of the n Construct a parseable function - based parseable function - based parseable function - based parse. | Is it possible for NestedArray to have 0 length array to avoid this -1 here? |
@@ -388,6 +388,12 @@ func RawDiff(ctx *context.Context) {
git.RawDiffType(ctx.Params(":ext")),
ctx.Resp,
); err != nil {
+ if git.IsErrNotExist(err) {
+ ctx.JSON(404, map[string]interface{}{
+ "message": "commit " + ctx.Params(":sha") + " does not exist.",
+ })
+ return
+ }
ctx.ServerError("GetRawDiff", err)
return
}
| [QueryBool,GetCommitGraphsCount,Warn,ValidateCommitWithEmail,GetLatestCommitStatus,CommitsByRange,GetRawDiff,AddParam,GetCommit,GetCommitGraph,HasPrefix,Redirect,ParentID,RawDiffType,RepoPath,HTML,Error,ValidateCommitsWithEmails,OpenRepository,ParseCommitsWithSignature,FileCommitsCount,NotFound,SetDefaultParams,LoadAndProcessCommits,ShortSha,ToUTF8WithFallback,Len,GetBranchName,GetRefs,CommitsByFileAndRange,Tr,ServerError,GetTagName,Join,ParentCount,QueryTrim,AddParamString,CalculateTrustStatus,ToLower,IsErrNotExist,Summary,NewSearchCommitsOptions,QueryStrings,GetDiffCommitWithWhitespaceBehavior,CalcCommitStatus,BranchNameSubURL,GetWhitespaceFlag,ParseCommitsWithStatus,Query,ParseCommitWithSignature,SearchCommits,GetCommitsCount,WikiPath,String,GetNote,QueryInt,NewPagination,Params,Trim] | GetRawDiff - get raw diff of a commit. | GH just send a text reply - what's the reasoning behind the json? |
@@ -130,8 +130,10 @@ module Users
begin
cacher.save(auth_params[:password], profile)
rescue Encryption::EncryptionError => err
- profile.deactivate(:encryption_error)
- analytics.track_event(Analytics::PROFILE_ENCRYPTION_INVALID, error: err.message)
+ if profile
+ profile.deactivate(:encryption_error)
+ analytics.track_event(Analytics::PROFILE_ENCRYPTION_INVALID, error: err.message)
+ end
end
end
| [SessionsController->[cache_active_profile->[new],track_authentication_attempt->[new],process_locked_out_user->[new],now->[now]]] | Returns a that can be used to find a user in the active or pending profiles. | Interesting that with this change, but no change to the specs, everything passes. I thought about going down this route originally, but the test was looking for a particular decryption error to get thrown, which couldn't come from this code since it gets rescued. |
@@ -241,6 +241,16 @@ taskq_thread(void *arg)
if (!prealloc)
task_free(tq, t);
}
+ /*
+ * This thread is on its way out, so just drop the lock temporarily
+ * in order to call the shutdown callback. This allows the callback
+ * to look at the taskqueue, even just before it dies.
+ */
+ if (tq->tq_dtor != NULL) {
+ mutex_exit(&tq->tq_lock);
+ tq->tq_dtor(tq);
+ mutex_enter(&tq->tq_lock);
+ }
tq->tq_nthreads--;
cv_broadcast(&tq->tq_wait_cv);
mutex_exit(&tq->tq_lock);
| [No CFG could be retrieved] | This function blocks until all tasks in queue have been processed. - - - - - - - - - - - - - - - - - - A wrapper around kmem_al. | Is there a reason, since this is all new code and the dtor callback would be new, to not simply have it deal with any locking if it needs to? |
@@ -3191,7 +3191,7 @@ Sedp::Writer::write_participant_message(const ParticipantMessageData& pmd,
(ser << pmd);
if (ok) {
- send_sample(payload, size, reader, sequence);
+ send_sample(payload, size + padding, reader, sequence, reader != GUID_UNKNOWN);
} else {
result = DDS::RETCODE_ERROR;
| [No CFG could be retrieved] | Reads and writes a message from the DCPS. Reads the header of the last N - tuple. | If this uses `size+padding`, does line 3184 need to change as well? |
@@ -2635,6 +2635,9 @@ public class ActiveMQServerControlImpl extends AbstractControl implements Active
if (addressSettings.getDeadLetterAddress() != null) {
settings.add("DLA", addressSettings.getDeadLetterAddress().toString());
}
+ if (addressSettings.getDeadLetterAddressPrefix() != null) {
+ settings.add("DLAPrefix", addressSettings.getDeadLetterAddressPrefix().toString());
+ }
if (addressSettings.getExpiryAddress() != null) {
settings.add("expiryAddress", addressSettings.getExpiryAddress().toString());
}
| [ActiveMQServerControlImpl->[listBindingsForAddress->[listBindingsForAddress],listConnectionIDs->[listConnectionIDs],getConnectorsAsJSON->[getConnectorsAsJSON],getDiskScanPeriod->[getDiskScanPeriod],listSessions->[listSessions],getTransactionTimeout->[getTransactionTimeout],getJournalType->[getJournalType],getJournalBufferTimeout->[getJournalBufferTimeout],isPersistIDCache->[isPersistIDCache],deployQueue->[deployQueue],isPersistenceEnabled->[isPersistenceEnabled],getLargeMessagesDirectory->[getLargeMessagesDirectory],isCreateJournalDir->[isCreateJournalDir],listNetworkTopology->[listNetworkTopology],getMessageExpiryThreadPriority->[getMessageExpiryThreadPriority],createAddress->[createAddress],destroyQueue->[destroyQueue],getManagementNotificationAddress->[getManagementNotificationAddress],getBindingsDirectory->[getBindingsDirectory],getManagementAddress->[getManagementAddress],isClustered->[isClustered],getTotalConsumerCount->[getTotalConsumerCount],createQueue->[createQueue],getConnectors->[getConnectors],createDivert->[createDivert],isJournalSyncNonTransactional->[isJournalSyncNonTransactional],closeConnectionsForUser->[closeConnectionsForUser],isJournalSyncTransactional->[isJournalSyncTransactional],getOutgoingInterceptorClassNames->[getOutgoingInterceptorClassNames],isSecurityEnabled->[isSecurityEnabled],getJournalCompactMinFiles->[getJournalCompactMinFiles],getConnectionTTLOverride->[getConnectionTTLOverride],getGlobalMaxSize->[getGlobalMaxSize],isReplicaSync->[isReplicaSync],addSecuritySettings->[addSecuritySettings],removeSecuritySettings->[removeSecuritySettings],setMessageCounterEnabled->[isMessageCounterEnabled,isStarted,setMessageCounterEnabled],getBridgeNames->[getBridgeNames],getJournalCompactPercentage->[getJournalCompactPercentage],getNotificationInfo->[getNotificationInfo],isMessageCounterEnabled->[isMessageCounterEnabled],setFailoverOnServerShutdown->[setFailoverOnServerShutdown],getSecurityInvalidationInterval->[getSecurityInvalidationInterval],checkStarted->[isStarted],getTotalMessageCount->[getTotalMessageCount],destroyBridge->[destroyBridge],getUptimeMillis->[getUptimeMillis],isPersistDeliveryCountBeforeDelivery->[isPersistDeliveryCountBeforeDelivery],getVersion->[getVersion],listConsumersAsJSON->[listConsumersAsJSON],listProducersInfoAsJSON->[listProducersInfoAsJSON],isStarted->[isStarted],listConnectionsAsJSON->[listConnectionsAsJSON],isCreateBindingsDir->[isCreateBindingsDir],closeConsumerConnectionsForAddress->[closeConsumerConnectionsForAddress],getRolesAsJSON->[getRolesAsJSON],listAllConsumersAsJSON->[listAllConsumersAsJSON],getDivertNames->[getDivertNames],destroyConnectorService->[destroyConnectorService],listHeuristicRolledBackTransactions->[listHeuristicRolledBackTransactions],getMaxDiskUsage->[getMaxDiskUsage],freezeReplication->[freezeReplication],sendQueueInfoToQueue->[sendQueueInfoToQueue],getPropertiesLoginModuleConfigurator->[getSecurityDomain],addUser->[addUser],isAsyncConnectionExecutionEnabled->[isAsyncConnectionExecutionEnabled],getNodeID->[getNodeID],listAddresses->[listAddresses,getAddressInfo],getJournalBufferSize->[getJournalBufferSize],getJournalDirectory->[getJournalDirectory],listRemoteAddresses->[listRemoteAddresses],resetAllMessageCounterHistories->[resetAllMessageCounterHistories],closeConsumerWithID->[closeConsumerWithID],resetUser->[resetUser],getAddressInfo->[getAddressInfo],listProducers->[listProducers],closeConnectionWithID->[closeConnectionWithID],disableMessageCounters->[disableMessageCounters],getAddressSettingsAsJSON->[getAddressSettingsAsJSON],closeSessionWithID->[closeSessionWithID],updateQueue->[updateQueue],deleteAddress->[deleteAddress],listSessionsAsJSON->[listSessionsAsJSON],getThreadPoolMaxSize->[getThreadPoolMaxSize],getAddressMemoryUsage->[getAddressMemoryUsage],internaListUser->[listUser],addAddressSettings->[addAddressSettings],scaleDown->[scaleDown],getTotalMessagesAdded->[getTotalMessagesAdded],getQueueNames->[getQueueNames],resetAllMessageCounters->[resetAllMessageCounters],getAddressMemoryUsagePercentage->[getGlobalMaxSize,getAddressMemoryUsage,getAddressMemoryUsagePercentage],listQueues->[listQueues],getMessageCounterSamplePeriod->[getMessageCounterSamplePeriod],getConnectorServices->[getConnectorServices],createConnectorService->[createConnectorService],isFailoverOnServerShutdown->[isFailoverOnServerShutdown],getJournalMaxIO->[getJournalMaxIO],getMessageCounterMaxDayCount->[getMessageCounterMaxDayCount],listConnections->[listConnections],listUser->[listUser],isWildcardRoutingEnabled->[isWildcardRoutingEnabled],getTotalMessagesAcknowledged->[getTotalMessagesAcknowledged],updateAddress->[updateAddress],getScheduledThreadPoolMaxSize->[getScheduledThreadPoolMaxSize],enableMessageCounters->[enableMessageCounters],listPreparedTransactions->[format,listPreparedTransactions],listConsumers->[listConsumers],getJournalFileSize->[getJournalFileSize],removeUser->[removeUser],getIncomingInterceptorClassNames->[getIncomingInterceptorClassNames],getUptime->[getUptime],getIDCacheSize->[getIDCacheSize],commitPreparedTransaction->[commitPreparedTransaction],addNotificationListener->[addNotificationListener],isSharedStore->[isSharedStore],closeConnectionsForAddress->[closeConnectionsForAddress],getMessageExpiryScanPeriod->[getMessageExpiryScanPeriod],internalRemoveUser->[removeUser],rollbackPreparedTransaction->[rollbackPreparedTransaction],setMessageCounterMaxDayCount->[setMessageCounterMaxDayCount],forceFailover->[forceFailover],removeAddressSettings->[removeAddressSettings],getAddressNames->[getAddressNames],getConnectionCount->[getConnectionCount],isBackup->[isBackup],listPreparedTransactionDetailsAsJSON->[listPreparedTransactionDetailsAsJSON],listPreparedTransactionDetailsAsHTML->[listPreparedTransactionDetailsAsHTML],setMessageCounterSamplePeriod->[setMessageCounterSamplePeriod],getTotalConnectionCount->[getTotalConnectionCount],removeNotificationListener->[removeNotificationListener],getTransactionTimeoutScanPeriod->[getTransactionTimeoutScanPeriod],getRoles->[getRoles],listAllSessionsAsJSON->[listAllSessionsAsJSON],getPagingDirectory->[getPagingDirectory],destroyDivert->[destroyDivert],getJournalMinFiles->[getJournalMinFiles],getClusterConnectionNames->[getClusterConnectionNames],listHeuristicCommittedTransactions->[listHeuristicCommittedTransactions],createBridge->[createBridge]]] | Get the address settings as JSON. This method returns a sequence of keys that can be used to determine if a node should be. | Should we avoid proliferating acronymn and use deadLetterAddressPrefix instead. Similar to expiryAddress. Is then clearer going forwards And in line with xml based configuration which isnt acronymd |
@@ -255,9 +255,6 @@ namespace System.Diagnostics
/// </summary>
public IEnumerable<KeyValuePair<string, object?>> TagObjects
{
-#if ALLOW_PARTIALLY_TRUSTED_CALLERS
- [System.Security.SecuritySafeCriticalAttribute]
-#endif
get => _tags ?? s_emptyTagObjects;
}
| [ActivitySpanId->[ToString->[ToHexString],GetHashCode->[GetHashCode],CopyTo->[SetSpanFromHexChars],ToString],Activity->[TrySetTraceIdFromParent->[IsW3CId],TryConvertIdToContext->[IsW3CId],TrySetTraceFlagsFromParent->[IsW3CId],ValidateSetCurrent->[NotifyError],TagsLinkedList->[Set->[Remove]],Dispose->[Stop,Dispose]],ActivityTraceId->[SetToRandomBytes->[CopyTo],ToString->[ToHexString],GetHashCode->[GetHashCode],ToString]] | - A key - value pair that represent the ID of the object that will be logged This is the only way to pass along to children of this activity. | @tarekgh I removed the `SecuritySafeCritical`. Assuming this is no longer needed because we removed the `Unsafe.As` on the last PR? |
@@ -32,6 +32,12 @@ var RingOp = ring.NewOp([]ring.InstanceState{ring.ACTIVE}, func(s ring.InstanceS
return s != ring.ACTIVE
})
+// UserOwnedRingOp is the operation used for checking if a user is owned by an alertmanager.
+var UserOwnedRingOp = ring.NewOp([]ring.InstanceState{ring.ACTIVE, ring.JOINING}, func(s ring.InstanceState) bool {
+ // A user is owned by an alertmanager in both ACTIVE and JOINING states.
+ return (s != ring.ACTIVE && s != ring.JOINING)
+})
+
// RingConfig masks the ring lifecycler config which contains
// many options not really required by the alertmanager ring. This config
// is used to strip down the config to the minimum, and avoid confusion
| [ToRingConfig->[DefaultValues],ToLifecyclerConfig->[Sprintf,GetInstancePort,GetInstanceAddr],RegisterFlags->[StringVar,Error,RegisterFlagsWithPrefix,Hostname,Log,DurationVar,Exit,Var,IntVar],NewOp] | "github. com. godocs. org / showup. docs. RegisterFlags registers the flags with the given FlagSet. | Given the "is owned" logic is conceptually used in other places (eg. `ReadFullStateForUser()` also gets the set of alertmanagers owning a tenant) I'm wondering if we should call it `SyncRingOp` given this operation is specific to the sync operation done by the alertmanager itself. |
@@ -29,8 +29,9 @@ var KnownValidationExceptions = []reflect.Type{
// MissingValidationExceptions is the list of types that were missing validation methods when I started
// You should never add to this list
var MissingValidationExceptions = []reflect.Type{
- reflect.TypeOf(&buildapi.BuildLogOptions{}), // TODO, looks like this one should have validation
- reflect.TypeOf(&imageapi.DockerImage{}), // TODO, I think this type is ok to skip validation (internal), but needs review
+ reflect.TypeOf(&buildapi.BuildLogOptions{}), // TODO, looks like this one should have validation
+ reflect.TypeOf(&buildapi.BinaryBuildRequestOptions{}), // TODO, looks like this one should have validation
+ reflect.TypeOf(&imageapi.DockerImage{}), // TODO, I think this type is ok to skip validation (internal), but needs review
}
func TestCoverage(t *testing.T) {
| [TypeOf,PkgPath,Contains,PtrTo,KnownTypes,Errorf,HasSuffix] | TestCoverage tests if a given object has a matching validation method. Invite to the validator that a pointer type is not valid. | so why not add validation? |
@@ -9,8 +9,11 @@ import os
import sys
from UM.Platform import Platform
+from cura import ApplicationMetadata
from cura.ApplicationMetadata import CuraAppName
+import sentry_sdk
+
parser = argparse.ArgumentParser(prog = "cura",
add_help = False)
parser.add_argument("--debug",
| [exceptHook->[show,exec_,CrashHandler,getInstance,QApplication,removePostedEvents,close,exit],get_cura_dir_path->[expanduser,isOSX,getenv,isWindows,join,isLinux],CuraApplication,hasattr,expanduser,getattr,vars,isWindows,join,CDLL,isLinux,remove,insert,getInstallPrefix,isOSX,makedirs,add_argument,append,ArgumentParser,get_cura_dir_path,enable,open,split,environ,basename,keys,parse_known_args,get,find_library,chdir,realpath,run,reverse] | Creates an object that represents a single unique identifier for a given application. This workaround is only needed on Windows and OSX. | Is this going to stay the same for the forseaable future? How does sentry know its us (or our builds of Cura) sending the data? |
@@ -51,7 +51,7 @@ type manifestSchema1Handler struct {
var _ ManifestHandler = &manifestSchema1Handler{}
-func (h *manifestSchema1Handler) FillImageMetadata(ctx context.Context, image *imageapi.Image) error {
+func (h *manifestSchema1Handler) FillImageMetadata(ctx context.Context, image *imageapiv1.Image) error {
signatures, err := h.manifest.Signatures()
if err != nil {
return err
| [FillImageMetadata->[Has,Stat,Blobs,Signatures,Insert,Errorf,NewString,ImageWithMetadata,References,GetLogger],Payload->[Payload],Verify->[Join,Error,Stat,MatchString,Blobs,Errorf,Verify,References],Digest->[FromBytes],NewJSONSignature,PrettySignature,ParsePrettySignature,Unmarshal] | FillImageMetadata fills in the image metadata with the latest metadata from the manifest. | pre-existing, but mutation of args always rubs me the wrong way. |
@@ -51,12 +51,12 @@ func (d *sccExecRestrictions) Validate(a admission.Attributes, o admission.Objec
pod, err := d.client.CoreV1().Pods(a.GetNamespace()).Get(a.GetName(), metav1.GetOptions{})
if err != nil {
- return admission.NewForbidden(a, err)
+ return err
}
// we have to convert to the internal pod because admission uses internal types for now
internalPod := &coreapi.Pod{}
if err := coreapiv1conversions.Convert_v1_Pod_To_core_Pod(pod, internalPod, nil); err != nil {
- return admission.NewForbidden(a, err)
+ return err
}
// TODO, if we want to actually limit who can use which service account, then we'll need to add logic here to make sure that
| [SetSecurityInformers->[SetSecurityInformers],SetExternalKubeClientSet->[SetExternalKubeClientSet],ValidateInitialization->[ValidateInitialization],SetAuthorizer->[SetAuthorizer]] | Validate is the main admission function that will be called when an SCC is requested. Returns a list of all the possible values for the current request. | how does a non status error end up with this? as internal error? |
@@ -1335,6 +1335,7 @@ class XmlLegacyLoaderTest extends \PHPUnit_Framework_TestCase
'controller' => 'SuluContentBundle:Default:index',
'cacheLifetime' => ['type' => CacheLifetimeResolverInterface::TYPE_SECONDS, 'value' => 2400],
'tags' => [],
+ 'areas' => [],
'meta' => [
'title' => [
'de' => 'Das ist das Template 1',
| [XmlLegacyLoaderTest->[testSections->[assertEquals,loadFixture],testNestingParams->[assertEquals,loadFixture],testReservedName->[setExpectedException,loadFixture],testCacheLifeTimeZero->[assertEquals,loadFixture],loadFixture->[getResourceDirectory,willReturn,prophesize,reveal,load],testWithoutRlpTagTypeHomeInternal->[assertNotNull,getResourceDirectory,willReturn,prophesize,reveal,load],testReadTypesEmptyProperties->[setExpectedException,loadFixture,assertEquals],testReadTemplate->[assertEquals,loadFixture],testReadTemplateWithXInclude->[assertEquals,loadFixture],testWithoutRlpTagTypePageInternal->[assertNotNull,getResourceDirectory,willReturn,prophesize,reveal,load],testLoadCacheLifetimeInvalidExpression->[setExpectedException,getResourceDirectory,willReturn,prophesize,reveal,load],testWithoutTitle->[setExpectedException,loadFixture],testReadTypesInvalidPath->[loadFixture],testBlockMultipleTypes->[assertEquals,loadFixture],testReadTitleInSection->[assertEquals,loadFixture],testWithoutRlpTagTypePage->[setExpectedException,getResourceDirectory,willReturn,prophesize,reveal,load],testLoadCacheLifetimeExpression->[getResourceDirectory,assertEquals,willReturn,prophesize,reveal,load],testWithoutRlpTagTypeHome->[setExpectedException,getResourceDirectory,willReturn,prophesize,reveal,load],testWithoutRlpTagTypeSnippet->[assertNotNull,getResourceDirectory,willReturn,prophesize,reveal,load],testReadTypesMissingMandatory->[loadFixture],testMetaParams->[assertEquals,loadFixture],testReadBlockTemplate->[assertEquals,loadFixture],testWithoutRlpTagTypeSnippetInternal->[assertNotNull,getResourceDirectory,willReturn,prophesize,reveal,load]]] | Tests if a template with xinclude is available. | How about a test where some areas are loaded? |
@@ -230,6 +230,9 @@ export const createEntity: ICrudPutAction<I<%= entityReactName %>> = entity => a
type: ACTION_TYPES.CREATE_<%= entityActionName %>,
payload: axios.post(apiUrl, cleanEntity(entity))
});
+ <%_ if (pagination === 'infinite-scroll') { _%>
+ dispatch(reset());
+ <%_ } _%>
dispatch(getEntities());
return result;
};
| [No CFG could be retrieved] | Create an action that fetches a list of entities and dispatches to the appropriate action. Create action that updates an entity by ID. | why are these calls needed? |
@@ -525,9 +525,12 @@ public class PersistenceManagerImpl implements PersistenceManager {
<K, V> SegmentedAdvancedLoadWriteStore<K, V> getFirstSegmentedStore(AccessMode mode) {
storesMutex.readLock().lock();
try {
- for (CacheLoader loader : loaders) {
- if (mode.canPerform(getStoreConfig(loader)) && loader instanceof SegmentedAdvancedLoadWriteStore) {
- return ((SegmentedAdvancedLoadWriteStore<K, V>) loader);
+ for (CacheLoader l : loaders) {
+ StoreConfiguration storeConfiguration;
+ if (l instanceof SegmentedAdvancedLoadWriteStore &&
+ (storeConfiguration = getStoreConfig(l)) != null && storeConfiguration.segmented() &&
+ mode.canPerform(storeConfiguration)) {
+ return ((SegmentedAdvancedLoadWriteStore<K, V>) l);
}
}
} finally {
| [PersistenceManagerImpl->[prepareAllTxStores->[checkStoreAvailability],deleteBatchFromAllNonTxStores->[checkStoreAvailability],performOnAllTxStores->[checkStoreAvailability],deleteFromAllStores->[checkStoreAvailability],addSegments->[addSegments],indexShareable->[indexShareable],preload->[preload],size->[size,checkStoreAvailability],startWriter->[start,undelegate],stop->[stop],clearAllStores->[checkStoreAvailability],StoreStatus->[availabilityChanged->[isAvailable]],publishKeys->[getFirstSegmentedStore,getFirstAdvancedCacheLoader,publishKeys],publishEntries->[getFirstSegmentedStore,publishEntries,getFirstAdvancedCacheLoader],disableStore->[pollStoreAvailability],purgeExpired->[checkStoreAvailability],writeBatchToAllNonTxStores->[checkStoreAvailability],undelegate->[undelegate],doRemove->[stop],removeCacheWriter->[undelegate],removeSegments->[removeSegments],loadFromAllStores->[checkStoreAvailability],removeCacheLoader->[undelegate],getMaxEntries->[size],startLoader->[start,undelegate],getCacheForStateInsertion->[size],writeToAllNonTxStores->[checkStoreAvailability,writeToAllNonTxStores],getStateTransferProvider->[checkStoreAvailability]]] | Gets the first SegmentedAdvancedLoadWriteStore that can perform the given access mode. | I assume you added this check as you were encountering NPE, does this fix need to be backported? |
@@ -21,6 +21,7 @@ import games.strategy.triplea.ui.mapdata.MapData;
public class TerritoryNameDrawable implements IDrawable {
private final String territoryName;
private final UiContext uiContext;
+ private Rectangle territoryBounds;
public TerritoryNameDrawable(final String territoryName, final UiContext uiContext) {
this.territoryName = territoryName;
| [TerritoryNameDrawable->[draw->[getAscent,shouldDrawTerritoryName,drawNamesFromTopLeft,drawTerritoryNames,getOriginalOwner,getPuImage,getWidth,getHeight,getPropertyTerritoryNameAndPuAndCommentColor,drawImage,toString,getPuPlacementPoint,getSize,draw,getTerritory,stringWidth,getProduction,defaultNamedToTextList,isEmpty,isWater,setFont,getBestTerritoryNameRect,getConvoyAttached,getNamePlacementPoint,drawSeaZoneNames,isPresent,drawString,getConvoyRoute,getWhatTerritoriesThisIsUsedInConvoysFor,drawComments,getFontMetrics,getLeading,getName,getPropertyMapFont,get,getCommentMarkerLocation,setColor,drawResources],getBestTerritoryNameRect->[getAscent,getName,stringWidth,Rectangle,isRectangleContainedInTerritory,getBoundingRect,ceil,abs],isRectangleContainedInTerritory->[getPolygons,contains,getName]]] | Create a drawable which draws a territory name. Comments = true ;. | You might want to wrap this in a SoftReference. This way it still can be garbage collected when the machine absolutely requires some memory. |
@@ -74,8 +74,6 @@ class MaskRCNNAdapter(Adapter):
def configure(self):
box_outputs = ['classes_out', 'scores_out', 'boxes_out']
detection_out = 'detection_out'
- if contains_all(self.launcher_config, [*box_outputs, detection_out]):
- raise ConfigError('only detection output or [{}] should be provided'.format(', '.join(box_outputs)))
self.detection_out = self.get_value_from_config(detection_out)
if not self.detection_out:
if not contains_all(self.launcher_config, box_outputs):
| [MaskRCNNAdapter->[_process_pytorch_outputs->[segm_postprocess,astype,append,squeeze,len,CoCocInstanceSegmentationPrediction,zip,next,DetectionPrediction,ContainerPrediction,shape],segm_postprocess->[astype,resize,im_mask,pad,maximum,expand_boxes,ascontiguousarray,clip,encoder,zeros,array],parameters->[StringField,super,update],process->[_extract_predictions,realisation],_process_detection_output->[segm_postprocess,get_coeff_x_y_from_metadata,int,append,CoCocInstanceSegmentationPrediction,where,zip,enumerate,DetectionPrediction,ContainerPrediction],expand_boxes->[zeros],_process_tf_obj_detection_api_outputs->[segm_postprocess,astype,int,append,CoCocInstanceSegmentationPrediction,zip,DetectionPrediction,ContainerPrediction,array],configure->[contains_all,get_value_from_config,format,join,ConfigError],__init__->[ImportError,super]]] | Configures the object based on the configuration values. | but we still need to check that both output formats are not provided together What if `detection_out` and `['classes_out', 'scores_out', 'boxes_out']` are not empty at the same time? |
@@ -266,6 +266,8 @@ class Openmpi(AutotoolsPackage):
depends_on('pkgconfig', type='build')
+ depends_on('libevent@2.0:')
+
depends_on('hwloc@2.0:', when='@4:')
# ompi@:3.0.0 doesn't support newer hwloc releases:
# "configure: error: OMPI does not currently support hwloc v2 API"
| [Openmpi->[setup_dependent_build_environment->[setup_run_environment],filter_rpaths->[filter_lang_rpaths],test->[_test_examples,_test_bin_ops,_test_check_versions]]] | Adds a default value for the given sequence number. Adds a dependency on all of the system - wide configuration options. | We can also make this conditional on `@4:` since earlier releases just used the bundled libevent if no external one is specified. Not sure how many people rely on the very old tagged versions in this package, I did not test the libevent-Open MPI compatibility matrix for those `version`s (libevent in Spack generally ships version 2.x) |
@@ -77,6 +77,9 @@ namespace MonoGame.Tools.Pipeline
_controller.OnProjectLoading += invokeUpdateMenus;
_controller.OnProjectLoaded += invokeUpdateMenus;
+ _controller.OnProjectLoading += OnProjectLoading;
+ _controller.OnProjectLoaded += OnProjectLoaded;
+
var updateUndoRedo = new CanUndoRedoChanged(UpdateUndoRedo);
var invokeUpdateUndoRedo = new CanUndoRedoChanged((u, r) => Invoke(updateUndoRedo, u, r));
| [MainView->[UpdateMenus->[UpdateRecentProjectList],TreeViewOnKeyDown->[TreeViewShowContextMenu]]] | Attaches the controller to this object. | Since they are really never removed. You could do this as an inline delegate and make the intent more clear. |
@@ -340,10 +340,7 @@ namespace System.Net.Http
private void CheckDisposed()
{
- if (_disposed)
- {
- throw new ObjectDisposedException(this.GetType().FullName);
- }
+ ObjectDisposedException.ThrowIf(_disposed, this);
}
// The only way to abort pending async operations in WinHTTP is to close the request handle.
| [WinHttpResponseStream->[Dispose->[Dispose],IAsyncResult->[ReadAsync],ReadAsync->[Task]]] | Checks if this object is disposed. | This case needs to be reverted since this library builds down-level. |
@@ -146,6 +146,12 @@ public interface EndpointLogger extends BasicLogger {
@Message(id = 10021, value = "Invalid Strength value: %s")
IllegalStateException invalidStrength(String strengthValue);
- @Message(id = 10022, value = "Endpoint '%s' requires Client Certificates, but no Trust Store is available in realm '%s'")
+ @Message(id = 10022, value = "Cannot retrieve authorization information for user %s")
+ SecurityException cannotRetrieveAuthorizationInformation(@Cause Throwable cause, String user);
+
+ @Message(id = 10023, value = "Endpoint '%s' requires Client Certificates, but no Trust Store is available in realm '%s'")
StartException noSSLTrustStore(String endpoint, String realm);
+
+ @Message(id = 10024, value = "Invalid authorizationId %s")
+ IllegalArgumentException invalidAuthorizationId(String authorizationId);
}
| [getMessageLogger,getName] | Missing Trust Store. | can you keep the old id in the old message and set the new id to the new message? |
@@ -24,6 +24,7 @@ class Gromacs(CMakePackage):
git = 'https://github.com/gromacs/gromacs.git'
version('develop', branch='master')
+ version('2018.3', 'c82634a31d0ec7dc8a128f404149440e')
version('2018.2', '7087462bb08393aec4ce3192fa4cd8df')
version('2018.1', '7ee393fa3c6b7ae351d47eae2adf980e')
version('2018', '6467ffb1575b8271548a13abfba6374c')
| [Gromacs->[cmake_args->[append],patch->[spec],variant,depends_on,version]] | Creates a new object containing all the attributes of a single object. Procedure to enable or disable a sequence number. | could you provide the sha256 hash? |
@@ -1305,7 +1305,7 @@ def _compute_mt_params(n_times, sfreq, bandwidth, low_bias, adaptive):
if bandwidth is not None:
half_nbw = float(bandwidth) * n_times / (2. * sfreq)
else:
- half_nbw = 2.
+ half_nbw = 4.
# Compute DPSS windows
n_tapers_max = int(2 * half_nbw)
| [csd_epochs->[sum],pick_channels_csd->[CrossSpectralDensity,_vector_to_sym_mat,_sym_mat_to_vector],csd_fourier->[get_data],_vector_to_sym_mat->[_n_dims_from_triu],csd_morlet->[get_data],_csd_fourier->[_sym_mat_to_vector],csd_array->[sum],read_csd->[CrossSpectralDensity],CrossSpectralDensity->[get_data->[_get_frequency_index],mean->[sum],__getitem__->[CrossSpectralDensity],save->[__getstate__],pick_frequency->[_get_frequency_index],sum->[CrossSpectralDensity,sum]],_csd_multitaper->[_sym_mat_to_vector],_execute_csd_function->[CrossSpectralDensity,sum],_prepare_csd->[copy],_csd_morlet->[mean],csd_multitaper->[get_data]] | Compute windowing and multitaper parameters. | @larsoner can you explain this change? |
@@ -18,7 +18,7 @@ namespace NServiceBus.Timeout.Core
{
if (timeout.Time.AddSeconds(-1) <= DateTime.UtcNow)
{
- var sendOptions = new DispatchOptions(timeout.Destination,new AtomicWithReceiveOperation(), new List<DeliveryConstraint>());
+ var sendOptions = new DispatchOptions(timeout.Destination,new AtomicWithReceiveOperation(), new List<DeliveryConstraint>(), new ContextBag());
var message = new OutgoingMessage(timeout.Headers[Headers.MessageId],timeout.Headers, timeout.State);
MessageSender.Dispatch(message, sendOptions);
| [DefaultTimeoutManager->[RemoveTimeoutBy->[RemoveTimeoutBy]]] | Push timeout. | We should flow the context from the incoming `TimeoutMessageProcessorBehavior` I'll create a separate pull for that |
@@ -87,7 +87,7 @@ class SecurityEventForm
end
def check_public_key_error(public_key)
- return false if public_key.present?
+ return false if service_provider&.ssl_certs.present?
errors.add(:jwt, t('risc.security_event.errors.no_public_key'))
@error_code = ErrorCodes::JWS
| [SecurityEventForm->[validate_jwt->[check_public_key_error],user->[user]]] | Checks if the public key is missing and adds an error if it is. | **Question:** This argument is no longer being used, right? |
@@ -277,13 +277,16 @@ class WebPubSubServiceClient(GeneratedWebPubSubServiceClient):
def __init__(
self,
endpoint, # type: str
+ hub, # type: str
credential, # type: Union[TokenCredential, AzureKeyCredential]
**kwargs # type: Any
):
# type: (...) -> None
+ if kwargs.get("port") and endpoint:
+ endpoint = endpoint.rstrip("/") + ":{}".format(kwargs.pop('port'))
kwargs['origin_endpoint'] = endpoint
_endpoint = '{Endpoint}'
- self._config = WebPubSubServiceClientConfiguration(endpoint, credential, **kwargs)
+ self._config = WebPubSubServiceClientConfiguration(hub=hub, endpoint=endpoint, credential=credential, **kwargs)
self._client = PipelineClient(base_url=_endpoint, config=self._config, **kwargs)
self._serialize = Serializer()
| [JwtCredentialPolicy->[_encode->[_UTC_TZ]],WebPubSubServiceClientConfiguration->[_configure->[JwtCredentialPolicy,ApiManagementProxy]],WebPubSubServiceClient->[from_connection_string->[_parse_connection_string],get_client_access_token->[_get_token_by_key],__init__->[WebPubSubServiceClientConfiguration]],_get_token_by_key->[_UTC_TZ]] | Initialize a new object with the given configuration. | What if the user already specified the port in the `endpoint`? |
@@ -92,7 +92,9 @@ export function _mapDispatchToProps(dispatch: Dispatch<any>) {
*/
_onSendMessage(text: string) {
dispatch(sendMessage(text));
- }
+ },
+
+ dispatch
};
}
| [No CFG could be retrieved] | Provides a function to map redux state to props. | why is this necessary? When you connect() you shoudl already have dispatch(), otherwise you can just compose a new mapDispatchToProps |
@@ -50,6 +50,11 @@ define([
owner : this
});
this._clearCommand = clearCommand;
+
+ this._qualityPreset = 39;
+ this._qualitySubPix = 0.5;
+ this._qualityEdgeThreshold = 0.125;
+ this._qualityEdgeThresholdMin = 0.0833;
}
function destroyResources(fxaa) {
| [No CFG could be retrieved] | Creates a FXAA object from a set of objects. Private functions - Texture is not defined in FAA. | Should these be packed into a `vec4` or just constants in the shader? |
@@ -2545,8 +2545,12 @@ class Report(object):
return html
- def _render_raw_butterfly_segments(self, *, raw: BaseRaw, image_format,
- tags):
+ def _render_raw_butterfly_segments(
+ self, *, raw: BaseRaw, image_format, tags
+ ):
+ orig_annotations = raw.annotations.copy()
+ raw.set_annotations(None)
+
# Pick 10 1-second time slices
times = np.linspace(raw.times[0], raw.times[-1], 12)[1:-1]
figs = []
| [Report->[add_inverse_operator->[_check_tags],add_image->[_get_dom_id,_add_or_replace,_html_image_element,_check_tags],add_epochs->[_html_epochs_element,_get_dom_id,_check_tags],_render_bem->[_render_one_bem_axis],add_figure->[_html_image_element,_add_or_replace,_get_dom_id,_fig_to_img,_check_tags],__getstate__->[_get_state_params],add_html->[_get_dom_id,_add_or_replace,_check_tags,_html_element],add_covariance->[_get_dom_id,_html_cov_element,_check_tags],_render_evoked_gfp->[_get_dom_id,_fig_to_img,_html_image_element],_render_evoked->[_render_evoked_joint,_render_evoked_whitened,_render_evoked_topomap_slider,_ssp_projs_html,_get_ch_types,_render_evoked_gfp],_render_ica_properties->[_get_dom_id,_fig_to_img,_html_image_element,_plot_ica_properties_as_arrays],add_evokeds->[_get_dom_id,_html_evoked_element,_check_tags],add_forward->[_check_tags],_render_ica_components->[_get_dom_id,_fig_to_img,_html_image_element],add_bem->[_get_dom_id,_add_or_replace,_check_tags,_html_element],_render_trans->[_get_dom_id,_iterate_trans_views,_html_image_element],add_trans->[_check_tags],_render_evoked_topomap_slider->[_validate_topomap_kwargs,_render_slider],_render_ica_overlay->[_get_dom_id,_fig_to_img,_html_image_element],__setstate__->[_get_state_params,_ContentElement],_render_ica->[_html_ica_element,_render_ica_artifact_sources,_render_ica_properties,_get_dom_id,_render_ica_components,_render_ica_artifact_scores,_render_ica_overlay,_html_element],_iterate_files->[add_inverse_operator,add_raw,add_evokeds,add_forward,add_epochs,_endswith,add_events,add_trans,add_projs,add_covariance,add_stc],_render_events->[_get_dom_id,_fig_to_img,_html_image_element],add_ica->[_render_ica,_check_tags],_render_ica_artifact_sources->[_get_dom_id,_fig_to_img,_html_image_element],_render_slider->[_get_dom_id,_fig_to_img,_html_slider_element],_add_or_replace->[_ContentElement],add_raw->[_html_raw_element,_get_dom_id,_check_tags],parse_folder->[_endswith,add_bem,_check_image_format],_render_evoked_joint->[_get_dom_id,_fig_to_img,_html_image_element],__init__->[_check_image_format],_render_one_bem_axis->[_get_bem_contour_figs_as_arrays,_render_slider],_render_ica_artifact_scores->[_get_dom_id,_fig_to_img,_html_image_element],save->[_html_footer_element,_html_toc_element,_html_header_element,__getstate__],_render_evoked_whitened->[_get_dom_id,_fig_to_img,_html_image_element],_render_cov->[_get_dom_id,_fig_to_img,_html_image_element],__exit__->[save],add_code->[_add_or_replace,_check_tags,_render_code],_render_raw_butterfly_segments->[_render_slider],_render_forward->[_html_forward_sol_element,_get_dom_id],_render_stc->[_render_slider],_render_inverse_operator->[_get_dom_id,_html_inverse_operator_element],_render_raw->[_html_image_element,_get_dom_id,_fig_to_img,_render_raw_butterfly_segments,_html_element],add_events->[_check_tags],_render_code->[_get_dom_id,_html_code_element],_render_epochs->[_html_image_element,_get_dom_id,_fig_to_img,_ssp_projs_html,_get_ch_types,_html_element],_render_ssp_projs->[_fig_to_img,_validate_topomap_kwargs,_html_image_element,_get_dom_id],add_sys_info->[add_code,_check_tags],add_projs->[_check_tags],add_stc->[_check_tags]],_itv->[_fig_to_img]] | Render one axis of bem contours. Renders the HTML tag. | Since `raw` can be passed directly rather than via filename, this should really be wrapped in a `try/finally` (or equivalently, a `@contextmanager`) that ensures that, no matter what, `raw.set_annotations(orig_annotations` gets called |
@@ -76,8 +76,8 @@ public class HiveAvroToOrcConverter
// TODO: Add num of buckets (from config)
// Avro table name and location
- String avroTableName = conversionEntity.getHiveUnit().getTableName();
- String avroTableLocation = conversionEntity.getHiveUnit().getLocation().get();
+ String avroTableName = conversionEntity.getHiveTable().getTableName();
+ String avroTableLocation = conversionEntity.getHiveTable().getSd().getLocation();
// ORC table name and location
String orcTableName = avroTableName + "_orc";
| [HiveAvroToOrcConverter->[convertRecord->[generateTableMappingDML,splitToList,size,checkArgument,checkNotNull,info,isBlank,put,getTableName,IllegalArgumentException,getProp,of,getDbName,endsWith,flatten,isNotBlank,generateCreateTableDDL,isPresent,substring,absent,format,getConversionQuery,length,appendQuery,get,getId],AvroFlattener]] | Convert record with given sequence number in output schema. Populates optional partition info and partition info map. Create DDL statement and DML statement for conversionEntity. | In case that converter is running for a partition, the location will be wrong (of table). We should check if this conversion entity has a partition, if so use its location Also, the avroTableLocation can be changed to avroDataLocation or similar, since it can be table or partition location based on what the conversion is happening |
@@ -58,10 +58,15 @@ const PAGE_LOADED_CLASS_NAME = 'i-amphtml-story-page-loaded';
/**
- * Selector for which media to wait for on page layout.
- * @const {string}
+ * Selectors for media elements
+ * @enum {string}
*/
-const PAGE_MEDIA_SELECTOR = 'amp-audio, amp-video, amp-img, amp-anim';
+const SELECTORS = {
+ // which media to wait for on page layout.
+ ALL_AMP_MEDIA: 'amp-audio, amp-video, amp-img, amp-anim',
+ ALL_MEDIA: 'audio, video',
+ ALL_VIDEO: 'video',
+};
/** @private @const {string} */
| [No CFG could be retrieved] | A custom element which represents a single page of an amp - story. Private methods for the page. | Nit: ``Selectors`` now that it's an enum |
@@ -50,8 +50,8 @@ public class CronTabDayOfWeekLocaleTest {
* HUDSON-8656.
*/
@Test
- @Url("http://issues.hudson-ci.org/browse/HUDSON-8656")
- public void hudson8658() throws Exception {
+ @Issue("HUDSON-8656") // This is _not_ JENKINS-8656
+ public void hudson8656() throws Exception {
final Calendar cal = Calendar.getInstance(locale);
cal.set(2011, 0, 16, 0, 0, 0); // Sunday, Jan 16th 2011, 00:00
final String cronStr = "0 23 * * 1-5"; // execute on weekdays @23:00
| [CronTabDayOfWeekLocaleTest->[parameters->[getAvailableLocales,equals,getInstance,add,getClass],isSundayAndPreviousRunIsSaturdayAsterisk->[compare,getInstance,set,CronTab,floor],isSundayAndNextRunIsWednesday->[compare,getInstance,set,ceil,CronTab],isSundayAndPreviousRunIsThursday->[compare,getInstance,set,CronTab,floor],isSundayAndPreviousRunIsPreviousSunday7->[compare,getInstance,set,CronTab,floor],isSundayAndNextRunIsTuesday->[compare,getInstance,set,ceil,CronTab],isSundayAndPreviousRunIsMonday->[compare,getInstance,set,CronTab,floor],isSundayAndPreviousRunIsWednesday->[compare,getInstance,set,CronTab,floor],isSundayAndNextRunIsNextSunday7->[compare,getInstance,set,ceil,CronTab],isSundayAndNextRunIsMonday->[compare,getInstance,set,ceil,CronTab],isSundayAndNextRunIsNextSunday->[compare,getInstance,set,ceil,CronTab],isSundayAndPreviousRunIsSaturday->[compare,getInstance,set,CronTab,floor],isSundayAndPreviousRunIsTuesday->[compare,getInstance,set,CronTab,floor],isSundayAndPreviousRunIsFriday->[compare,getInstance,set,CronTab,floor],isSaturdayAndNextRunIsSundayAsterisk->[compare,getInstance,set,ceil,CronTab],isSundayAndNextRunIsFriday->[compare,getInstance,set,ceil,CronTab],isSundayAndNextRunIsSaturday->[compare,getInstance,set,ceil,CronTab],hudson8658->[compare,getInstance,set,ceil,CronTab],compare->[getDateTimeInstance,getTime,get,getFirstDayOfWeek,format,assertEquals],isSundayAndPreviousRunIsPreviousSunday->[compare,getInstance,set,CronTab,floor],isSundayAndNextRunIsThursday->[compare,getInstance,set,ceil,CronTab]]] | Checks if a node in the system has a lease on the next day. | Why do you rename the test? |
@@ -1563,6 +1563,7 @@ namespace System.ComponentModel.Design
protected virtual bool ShowDialog(System.ComponentModel.Design.DesignerOptionService.DesignerOptionCollection options, object optionObject) { throw null; }
object System.ComponentModel.Design.IDesignerOptionService.GetOptionValue(string pageName, string valueName) { throw null; }
void System.ComponentModel.Design.IDesignerOptionService.SetOptionValue(string pageName, string valueName, object value) { }
+ [System.ComponentModel.EditorAttribute("", "System.Drawing.Design.UITypeEditor, System.Drawing, Version=4.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a")]
public sealed partial class DesignerOptionCollection : System.Collections.ICollection, System.Collections.IEnumerable, System.Collections.IList
{
internal DesignerOptionCollection() { }
| [BindableAttribute->[All],PropertyTabAttribute->[All],MemberRelationship->[Sequential],DataObjectFieldAttribute->[Property],RunInstallerAttribute->[Class],AttributeProviderAttribute->[Property],TypeDescriptor->[RemoveProviderTransparent->[Advanced],RemoveAssociations->[Advanced],GetAssociation->[Advanced],AddAttributes->[Advanced],GetProvider->[Advanced],GetConverter->[Advanced],GetEditor->[Advanced],CreateAssociation->[Advanced],GetEvents->[Advanced],RemoveAssociation->[Advanced],GetDefaultProperty->[Advanced],GetAttributes->[Advanced],GetClassName->[Advanced],AddProviderTransparent->[Advanced],GetDefaultEvent->[Advanced],GetProperties->[Advanced],RemoveProvider->[Advanced],AddEditorTable->[Advanced],AddProvider->[Advanced],GetReflectionType->[Advanced],GetComponentName->[Advanced],Advanced,PublicParameterlessConstructor],HelpKeywordAttribute->[All],TimersDescriptionAttribute->[All],DefaultBindingPropertyAttribute->[Class],ToolboxItemFilterAttribute->[Class],DefaultEventAttribute->[Class],ListBindableAttribute->[All],ToolboxItemAttribute->[All],DataObjectMethodAttribute->[Method],InheritanceAttribute->[Event,Field,Property],LicenseProviderAttribute->[Class],EditorAttribute->[All],ComplexBindingPropertiesAttribute->[Class],InstallerTypeAttribute->[Class],RecommendedAsConfigurableAttribute->[Property],PasswordPropertyTextAttribute->[All],DefaultSerializationProviderAttribute->[Class],DesignTimeVisibleAttribute->[Interface,Class],DefaultPropertyAttribute->[Class],RootDesignerSerializerAttribute->[Interface,Class],LookupBindingPropertiesAttribute->[Class],AmbientValueAttribute->[All],ProvidePropertyAttribute->[Class],DataObjectAttribute->[Class],SettingsBindableAttribute->[Property],ExtenderProvidedPropertyAttribute->[All],MarshalByValueComponent->[Hidden]] | This method is used to get or set a option value. | Empty string looks odd but that seems consistent with .NETFramework |
@@ -442,7 +442,8 @@ func main() {
nodeConfig := createGlobalConfig()
currentNode := setupConsensusAndNode(nodeConfig)
- if currentNode.Blockchain().ShardID() != 0 {
+ if nodeConfig.ShardID != 0 {
+ utils.GetLogInstance().Info("SupportBeaconSyncing", "shardID", currentNode.Blockchain().ShardID(), "shardID1", nodeConfig.ShardID)
go currentNode.SupportBeaconSyncing()
}
| [Warn,SerializeToHexStr,Info,RunServices,GetMetricsFlag,Serialize,SupportSyncing,GetMemProfiling,New,GetLogInstance,Bool,CommitCommittee,SetBeaconGroupID,FindAccount,GetHandler,Seed,UpdateConsensusInformation,Network,Println,SetPushgatewayIP,AddLogFile,ReshardingEpoch,GetPublicKey,SetMode,NewClientGroupIDByShardID,NewLegacySyncingPeerProvider,GetDefaultConfig,LoadBlsKeyWithPassPhrase,Now,Blockchain,GetBeaconGroupID,Exit,Add,NewLDBDatabase,NewGroupIDByShardID,Nanosecond,SupportBeaconSyncing,SetNetworkType,Logger,NumberU64,GetVersion,NewDNSSyncingPeerProvider,SetVersion,SetRole,Str,Printf,GetClientGroupID,Pretty,InstanceForEpoch,NewInt,NewFixedSchedule,Fatal,ParseDuration,ShardID,SetClientGroupID,GOMAXPROCS,NewConnLogger,NewHost,Duration,SetHandler,GetP2PHost,LoadKeyFromFile,Crit,NumCPU,Start,SetLogVerbosity,Lvl,SetViewID,Var,Role,Fprintln,NumShards,Base,ParseUint,Err,InitShardState,NumNodesPerShard,DisableViewChangeForTestingOnly,StringsToAddrs,MaybeCallGCPeriodically,Msg,MustAddressToBech32,Sprintf,Uint,String,Parse,StartServer,StartRPC,EncodeToString,SetCommitDelay,RemoveAll,GetShardConfig,NewInstance,GetShardGroupID,SetMetricsFlag,Error,Int,GetID,CollectMetrics,Notify,NewLocalSyncingPeerProvider,GetPassphraseFromSource,ServiceManagerSetup,GetNetworkType,SetLogContext,Config,NetworkType,Fprintf,CurrentBlock,SetShardGroupID,MatchFilterHandler,SetPushgatewayPort,ParseAddr,GetSyncingPort,GetLogger] | Private functions for handling devnet related functions. This function is called from the main loop to start the node. | are these supposed to be/can be different? |
@@ -18,7 +18,9 @@ const jobName = 'bundle-size-module-build.js';
* Steps to run during push builds.
*/
function pushBuildWorkflow() {
- timedExecOrDie('amp dist --noconfig --esm --version_override 0000000000000');
+ timedExecOrDie(
+ 'amp dist --noconfig --esm --version_override 0000000000000 --nomanglecache'
+ );
storeModuleBuildToWorkspace();
}
| [No CFG could be retrieved] | Script that builds the module AMP runtime for bundle - size during CI. | nit: it could make sense to instead make a new flag `--for_bundle_size` which encapsulates both of these flags |
@@ -321,8 +321,8 @@ public class MustFightBattle extends AbstractBattle implements BattleStepStrings
if (m_stack.isExecuting()) {
final ITripleADisplay display = getDisplay(bridge);
display.showBattle(m_battleID, m_battleSite, getBattleTitle(),
- removeNonCombatants(m_attackingUnits, true, m_attacker, false, false, false),
- removeNonCombatants(m_defendingUnits, false, m_defender, false, false, false), m_killed,
+ removeNonCombatants(m_attackingUnits, true, false, false, false),
+ removeNonCombatants(m_defendingUnits, false, false, false, false), m_killed,
m_attackingWaitingToDie, m_defendingWaitingToDie, m_dependentUnits, m_attacker, m_defender, isAmphibious(),
getBattleType(), m_amphibiousLandAttackers);
display.listBattleSteps(m_battleID, m_stepStrings);
| [MustFightBattle->[transporting->[transporting],showCasualties->[isEmpty],attackerRetreatSubs->[getAttackerRetreatTerritories,canAttackerRetreatSubs],defenderRetreatSubs->[canDefenderRetreatSubs],attackAirOnNonSubs->[fire],getBattleExecutables->[execute->[determineStepStrings,shouldEndBattleDueToMaxRounds,updateDefendingAAUnits,pushFightLoopOnStack,updateOffensiveAAUnits,isEmpty],addFightStartToStack],remove->[isEmpty],attackerRetreat->[canAttackerRetreat,getAttackerRetreatTerritories],isEmpty->[isEmpty],attackSubs->[isEmpty,fire],determineStepStrings->[canSubsSubmerge,isEmpty],submergeSubsVsOnlyAir->[submergeUnits],checkDefendingPlanesCanLand->[match,isEmpty,remove],fireNavalBombardment->[fire],checkUndefendedTransports->[isEmpty],fireSuicideUnitsDefend->[fire],FireAA->[execute->[]],attackerWins->[match,isEmpty,getTransportDependents],retreatFromNonCombat->[isEmpty],retreatUnits->[retreatFromDependents,isEmpty,retreatFromNonCombat],clearWaitingToDie->[remove],unitsLostInPrecedingBattle->[isEmpty,remove],queryRetreat->[canSubsSubmerge,isEmpty],attackNonSubs->[canAirAttackSubs,isEmpty,fire],endBattle->[clearTransportedByForAlliedAirOnCarrier,isEmpty,clearWaitingToDie],canDefenderRetreatSubs->[canSubsSubmerge],defendNonSubs->[isEmpty,fire],cancelBattle->[endBattle],writeUnitsToHistory->[isEmpty],defendSubs->[isEmpty,fire],canFireDefendingAA->[updateDefendingAAUnits],canFireOffensiveAA->[updateOffensiveAAUnits],fireSuicideUnitsAttack->[fire],clearTransportedByForAlliedAirOnCarrier->[match,isEmpty],canAttackerRetreatSubs->[canAttackerRetreat,canSubsSubmerge],retreatFromDependents->[removeAttack],removeCasualties->[isEmpty],markNoMovementLeft->[isEmpty],checkForUnitsThatCanRollLeft->[match,isEmpty],retreatPlanes->[isEmpty],addFightStepsNonEditMode->[DefendSubs,defenderSubsFireFirst,AttackSubs],submergeUnits->[isEmpty],addAttackChange->[isEmpty],retreatUnitsAndPlanes->[retreatFromDependents,isEmpty,retreatFromNonCombat],fire->[isEmpty],defendAirOnNonSubs->[canAirAttackSubs,isEmpty,fire],removeNonCombatants->[isEmpty,removeNonCombatants],removeFromNonCombatLandings->[getTransportDependents,isEmpty,remove],fight->[addDependentUnits,updateDefendingAAUnits,getBattleTitle,updateOffensiveAAUnits],canAttackerRetreat->[getAttackerRetreatTerritories],defenderWins->[isEmpty],attackerRetreatNonAmphibUnits->[getAttackerRetreatTerritories],landParatroops->[isEmpty]]] | This method is called when a battle is fighting. It is called from Add dependent units. if there is no match in the stack execute it. | This file contains the largest number of changes by far and might be a candidate for extraction to a separate PR. |
@@ -271,6 +271,7 @@ public class GobblinHelixJobScheduler extends JobScheduler implements StandardMe
Instrumented.updateTimer(Optional.of(metrics.timeForJobFailure), System.nanoTime() - startTime, TimeUnit.NANOSECONDS);
} else {
metrics.totalJobsCommitted.incrementAndGet();
+ Instrumented.updateTimer(Optional.of(metrics.timeForJobCommit), System.nanoTime() - startTime, TimeUnit.NANOSECONDS);
}
}
}
| [GobblinHelixJobScheduler->[startUp->[startUp],scheduleJob->[scheduleJob],handleNewJobConfigArrival->[scheduleJob,updateTimeBeforeJobScheduling,MetricsTrackingListener],NonScheduledJobRunner->[run->[runJob,updateTimeBeforeJobLaunching]],Metrics->[getName->[getName]],runJob->[runJob],handleUpdateJobConfigArrival->[handleNewJobConfigArrival],MetricsTrackingListener->[onJobCompletion->[onJobCompletion,isInstrumentationEnabled],onJobPrepare->[onJobPrepare,isInstrumentationEnabled],onJobCancellation->[onJobCancellation,isInstrumentationEnabled]]]] | onJobCompletion - override to record total jobs completed and cancelling. | Call this timeForCommittedJobs and timeForFailedJobs to be clearer. |
@@ -1437,9 +1437,9 @@ vos_fetch_begin(daos_handle_t coh, daos_unit_oid_t oid, daos_epoch_t epoch,
struct vos_io_context *ioc;
int i, rc;
- D_DEBUG(DB_TRACE, "Fetch "DF_UOID", desc_nr %d, epoch "DF_X64"\n",
+ D_DEBUG(DB_IO, "Fetch "DF_UOID", desc_nr %d, epoch "DF_X64"\n",
DP_UOID(oid), iod_nr, epoch);
-
+
rc = vos_ioc_create(coh, oid, true, epoch, iod_nr, iods,
NULL, vos_flags, shadows, false, 0,
dth, &ioc);
| [No CFG could be retrieved] | Fetches all missing keys from the DOS and adds them to the DOS s list of vos_flags shadows = 0 shadows_ioc = 0 shadows_ - DER_NONEX = DER_NONEX. | (style) trailing whitespace |
@@ -10,12 +10,13 @@ class Zeromq(AutotoolsPackage):
"""The ZMQ networking/concurrency library and core API"""
homepage = "http://zguide.zeromq.org/"
- url = "http://download.zeromq.org/zeromq-4.1.2.tar.gz"
git = "https://github.com/zeromq/libzmq.git"
version('develop', branch='master')
- version('4.2.5', 'a1c95b34384257e986842f4d006957b8',
- url='https://github.com/zeromq/libzmq/releases/download/v4.2.5/zeromq-4.2.5.tar.gz')
+ version('4.3.2', sha256='ebd7b5c830d6428956b67a0454a7f8cbed1de74b3b01e5c33c5378e22740f763')
+ version('4.3.1', sha256='bcbabe1e2c7d0eec4ed612e10b94b112dd5f06fcefa994a0c79a45d835cd21eb')
+ version('4.3.0', sha256='8e9c3af6dc5a8540b356697081303be392ade3f014615028b3c896d0148397fd')
+ version('4.2.5', 'a1c95b34384257e986842f4d006957b8')
version('4.2.2', '52499909b29604c1e47a86f1cb6a9115')
version('4.1.4', 'a611ecc93fffeb6d058c0e6edf4ad4fb')
version('4.1.2', '159c0c56a895472f02668e692d122685')
| [Zeromq->[autoreconf->[which,bash],configure_args->[append],depends_on,conflicts,version,when,variant]] | Creates a new object. Provide a list of command line options for the autogen command. | You should still keep the URL here (maybe update to the newer URL). Spack uses this URL when you run `spack versions` and `spack checksum`. |
@@ -72,6 +72,8 @@ class AtisDatasetReader(DatasetReader):
self._token_indexers = token_indexers or {'tokens': SingleIdTokenIndexer()}
self._tokenizer = tokenizer or WordTokenizer(SpacyWordSplitter(pos_tags=True))
self._database_directory = database_directory
+ # TODO(kevin): Add a keep_if_no_dpd flag so that during validation, we do not skip queries that
+ # cannot be parsed.
@overrides
def _read(self, file_path: str):
| [AtisDatasetReader->[text_to_instance->[get_action_sequence,split,debug,ArrayField,all_possible_actions,AtisWorld,append,tokenize,ProductionRuleField,IndexField,join,enumerate,ListField,Instance,lower,TextField,MetadataField],__init__->[SingleIdTokenIndexer,super,WordTokenizer,SpacyWordSplitter],_read->[text_to_instance,append,cached_path,_lazy_parse,read,info,open]],_lazy_parse->[split,loads],register,getLogger] | Initialize a new instance of the class with a lazy load of the ATIS file. | I wouldn't name this `keep_if_no_dpd`, because what you really want is `keep_unparseable_utterances`, or something. DPD isn't relevant here. |
@@ -915,13 +915,7 @@ namespace Internal.JitInterface
private bool canTailCall(CORINFO_METHOD_STRUCT_* callerHnd, CORINFO_METHOD_STRUCT_* declaredCalleeHnd, CORINFO_METHOD_STRUCT_* exactCalleeHnd, bool fIsTailPrefix)
{
- if (fIsTailPrefix)
- {
- // FUTURE: Delay load fixups for tailcalls
- throw new RequiresRuntimeJitException(nameof(fIsTailPrefix));
- }
-
- return false;
+ return true;
}
private MethodWithToken ComputeMethodWithToken(MethodDesc method, ref CORINFO_RESOLVED_TOKEN pResolvedToken, TypeDesc constrainedType, bool unboxing)
| [RequiresRuntimeJitException->[ToString],CorInfoImpl->[ceeInfoGetCallInfo->[ToString,IsGenericTooDeeplyNested,IsTypeSpecForTypicalInstantiation],embedMethodHandle->[ToString],getCallInfo->[ToString,ceeInfoGetCallInfo,UpdateConstLookupWithRequiresRuntimeJitSymbolIfNeeded,VerifyMethodSignatureIsStable],ISymbolNode->[ToString],classMustBeLoadedBeforeCodeIsRun->[classMustBeLoadedBeforeCodeIsRun],EncodeFieldBaseOffset->[ToString,HasLayoutMetadata,PreventRecursiveFieldInlinesOutsideVersionBubble],ComputeRuntimeLookupForSharedGenericToken->[ToString],UpdateConstLookupWithRequiresRuntimeJitSymbolIfNeeded->[MethodSignatureIsUnstable],embedClassHandle->[ToString],embedGenericHandle->[ToString,ceeInfoEmbedGenericHandle],ceeInfoEmbedGenericHandle->[ComputeRuntimeLookupForSharedGenericToken,ToString],CompileMethod->[FunctionJustThrows,ShouldSkipCompilation],getReadyToRunHelper->[ToString],VerifyMethodSignatureIsStable->[MethodSignatureIsUnstable],getFieldInfo->[IsClassPreInited],getFunctionEntryPoint->[ToString],IsGenericTooDeeplyNested->[IsGenericTooDeeplyNested],setEHinfo->[classMustBeLoadedBeforeCodeIsRun],PreventRecursiveFieldInlinesOutsideVersionBubble->[ToString]],MethodWithToken->[ToString->[ToString],Equals->[Equals],GetHashCode->[GetHashCode],CompareTo->[CompareTo]],GenericContext->[GetHashCode->[GetHashCode]]] | Checks if a method can be called on a tail call. | This cannot be just true. It must respect the same set of rules that are implemented in canTailCall in jitinterface.cpp. Notably if there is no tail prefix the following situations are forbidden from tail call optimization - If the Caller method is NoInline - If the Caller method is the entrypoint to the application - If the Callee method has the RequireSecObject bit set on the MethodDef |
@@ -74,6 +74,18 @@ func Compute(staked shard.SlotList) *Roster {
roster.Voters[staked[i].BlsPublicKey] = member
}
+ // NOTE Enforce voting power sums to one, give diff (expect tiny amt) to last staked voter
+ if diff := numeric.OneDec().Sub(
+ ourPercentage.Add(theirPercentage),
+ ); diff.GT(numeric.ZeroDec()) && lastStakedVoter != nil {
+ lastStakedVoter.EffectivePercent = lastStakedVoter.EffectivePercent.Add(diff)
+ theirPercentage = theirPercentage.Add(diff)
+ utils.Logger().Info().
+ Str("diff", diff.String()).
+ Str("bls-public-key-of-receipent", lastStakedVoter.Identity.Hex()).
+ Msg("sum of voting power of hmy & staked slots not equal to 1, gave diff to staked slot")
+ }
+
roster.OurVotingPowerTotalPercentage = ourPercentage
roster.TheirVotingPowerTotalPercentage = theirPercentage
| [MustNewDecFromStr,Quo,NewDec,ZeroDec,Add,Mul] | NewRoster creates a roster object for the given member. | I believe the diff can be negative, because the Quo uses banker's rounding. |
@@ -26,6 +26,7 @@ from sickbeard import logger
from sickbeard.providers import btn, newznab, womble, thepiratebay, torrentleech, kat, iptorrents, torrentz, \
omgwtfnzbs, scc, hdtorrents, torrentday, hdbits, hounddawgs, nextgen, speedcd, nyaatorrents, animenzb, bluetigers, cpasbien, fnt, xthor, torrentbytes, \
freshontv, titansoftv, libertalia, morethantv, bitsoup, t411, tokyotoshokan, shazbat, rarbg, alpharatio, tntvillage, binsearch, torrentproject, extratorrent, \
+<<<<<<< HEAD
scenetime, btdigg, strike, transmitthenet, tvchaosuk, bitcannon, pretome, gftracker, hdspace, newpct, elitetorrent, bitsnoop
__all__ = [
| [getProviderClass->[len,getID],makeTorrentRssProvider->[TorrentRssProvider,log,split,len],sortedProviderList->[dict,append,providerDict,shuffle,zip,getID],makeNewznabProvider->[log,split,NewznabProvider,len],getTorrentRssProviderList->[split,append,filter,makeTorrentRssProvider,set,add],makeProviderList->[getProviderModule],getNewznabProviderList->[split,append,dict,zip,filter,set,makeNewznabProvider,getDefaultNewznabProviders,add],getProviderModule->[lower,Exception]] | This module is used to provide a full description of a single unique integer in a GN Returns a list of modules sorted by their priority. | @JohnDooe This file cant be touched ,it should only inclued your provider.. |
@@ -4,6 +4,11 @@ module LambdaCallback
dcs = DocumentCaptureSession.find_by(result_id: result_id_parameter)
if dcs
+ analytics.track_event(
+ Analytics::LAMBDA_RESULT_ADDRESS_PROOF_RESULT,
+ result: address_result_parameter,
+ )
+
dcs.store_proofing_result(address_result_parameter.to_h)
track_exception_in_result(address_result_parameter)
| [AddressProofResultController->[result_id_parameter->[require],create->[to_h,track_exception_in_result,find_by,head,store_proofing_result,notice_error],track_exception_in_result->[nil?,notify_exception,notice_error],address_result_parameter->[permit],config_auth_token->[address_proof_result_lambda_token]]] | create a new missing result_id. | are we trying to log the whole result? or just the result ID? |
@@ -603,6 +603,18 @@ describe('angular', function() {
/* eslint-enable */
});
+ it('should copy media stream objects', function() {
+ var source = new MediaStream();
+ var destination = copy(source);
+
+ expect(destination.id).toMatch(/^((\w+)-){4}(\w+)$/);
+ expect(typeof destination.active).toBe('boolean');
+ expect(source.id).not.toBe(destination.id);
+ expect(source.active).toBe(destination.active);
+
+
+ });
+
it('should copy source until reaching a given max depth', function() {
var source = {a1: 1, b1: {b2: {b3: 1}}, c1: [1, {c2: 1}], d1: {d2: 1}};
var dest;
| [No CFG could be retrieved] | Checks that the destination object is not null and not equal to the object passed to it. Private functions - > Assertion that the source object has a unique key and that the destination object. | Not all browsers support `MediaStream` (which seems to cause Travis to fail). |
@@ -39,6 +39,8 @@ import org.springframework.integration.MessageDeliveryException;
import org.springframework.integration.channel.DirectChannel;
import org.springframework.integration.channel.PublishSubscribeChannel;
import org.springframework.integration.channel.QueueChannel;
+import org.springframework.integration.channel.interceptor.ChannelInterceptorAdapter;
+import org.springframework.integration.channel.interceptor.GlobalChannelInterceptorTests;
import org.springframework.integration.config.TestChannelInterceptor;
import org.springframework.integration.core.PollableChannel;
import org.springframework.integration.dispatcher.RoundRobinLoadBalancingStrategy;
| [ChannelParserTests->[testChannelWithoutId->[getClass,ClassPathXmlApplicationContext],testChannelInteceptorInnerBean->[send,receive,getBean,getPayload,assertEquals,getClass,ClassPathXmlApplicationContext],testMultipleDatatypeChannelWithIncorrectType->[getBean,getClass,send,ClassPathXmlApplicationContext],testChannelInteceptorRef->[send,receive,getSendCount,getBean,assertEquals,getClass,getReceiveCount,ClassPathXmlApplicationContext],testChannelWithCapacity->[assertTrue,send,getBean,assertFalse,getClass,ClassPathXmlApplicationContext],channelWithFailoverDispatcherAttribute->[DirectFieldAccessor,getPropertyValue,assertNull,getBean,is,instanceOf,assertThat,assertEquals,getClass,ClassPathXmlApplicationContext],testDirectChannelByDefault->[DirectFieldAccessor,getPropertyValue,getBean,is,instanceOf,assertThat,assertEquals,getClass,ClassPathXmlApplicationContext],testDatatypeChannelWithIncorrectType->[getBean,getClass,send,ClassPathXmlApplicationContext],testMultipleDatatypeChannelWithCorrectTypes->[assertTrue,send,getBean,getClass,ClassPathXmlApplicationContext],testPriorityChannelWithIntegerDatatypeEnforced->[assertTrue,send,getBean,getPayload,assertEquals,getClass,ClassPathXmlApplicationContext],testPublishSubscribeChannelWithTaskExecutorReference->[assertNotNull,DirectFieldAccessor,getPropertyValue,getBean,assertEquals,getClass,ClassPathXmlApplicationContext],channelWithCustomQueue->[getPropertyValue,getBean,assertSame,assertEquals,getClass,ClassPathXmlApplicationContext],testDatatypeChannelWithAssignableSubTypes->[assertTrue,send,getBean,getClass,ClassPathXmlApplicationContext],testPublishSubscribeChannel->[getBean,assertEquals,getClass,ClassPathXmlApplicationContext],testDatatypeChannelWithCorrectType->[assertTrue,send,getBean,getClass,ClassPathXmlApplicationContext],testPriorityChannelWithDefaultComparator->[send,build,receive,getBean,getPayload,assertEquals,getClass,ClassPathXmlApplicationContext],testPriorityChannelWithCustomComparator->[send,receive,getBean,getPayload,assertEquals,getClass,ClassPathXmlApplicationContext]]] | Test method for channel without id. Test channel with capacity. | And remove this useless import too, please |
@@ -282,6 +282,11 @@ class ModelSerializer(BaseSerializer):
crit_dict['fields'] = [self._translate(model, field) for field in crit.fields]
return Criteria.from_dict(crit_dict)
+ # expose these "private" methods as public, pending a more in-depth refactor
+ # https://pulp.plan.io/issues/1555
+ translate_field = _translate
+ translate_filters = _translate_filters
+
class Repository(ModelSerializer):
"""
| [BaseSerializer->[_mask_field->[_mask_field],_to_representation->[to_representation,get_href],_remove_excluded->[_remove_excluded],data->[_to_representation]],ModelSerializer->[translate_criteria->[_translate,_translate_filters]]] | Translate a with object to the internal representation. | Seems reasonable in the short term. |
@@ -3324,8 +3324,16 @@ public final class OzoneManager extends ServiceRuntimeInfoImpl
omRatisServer.getOmStateMachine().unpause(lastAppliedIndex, term);
LOG.info("Reloaded OM state with Term: {} and Index: {}", term,
lastAppliedIndex);
- } catch (IOException ex) {
+ } catch (Exception ex) {
String errorMsg = "Failed to reload OM state and instantiate services.";
+ // Delete the backup DB if exists and then terminate OM.
+ try {
+ if (dbBackup != null) {
+ FileUtils.deleteFully(dbBackup);
+ }
+ } catch (Exception e) {
+ LOG.error("Failed to delete the backup of the original DB {}", dbBackup);
+ }
exitManager.exitSystem(1, errorMsg, ex, LOG);
}
| [OzoneManager->[completeMultipartUpload->[buildAuditMessageForFailure,buildAuditMessageForSuccess,completeMultipartUpload],getVolumeInfo->[checkAcls,getVolumeInfo],startTrashEmptier->[start],loginOMUserIfSecurityEnabled->[loginOMUser],join->[join],hasAcls->[checkAcls],allocateBlock->[checkAcls,allocateBlock],deleteBucket->[checkAcls,deleteBucket],initiateMultipartUpload->[buildAuditMessageForFailure,buildAuditMessageForSuccess,initiateMultipartUpload],abortMultipartUpload->[buildAuditMessageForSuccess,abortMultipartUpload,buildAuditMessageForFailure],getAcl->[getAcl,getResourceType,auditAcl,checkAcls],getOMNodeId->[getOMNodeId],isLeader->[isLeader],renameKey->[checkAcls,renameKey],deleteKey->[deleteKey,checkAcls],renewDelegationToken->[isAllowedDelegationTokenOp],getRemoteUser->[getRemoteUser],createDirectory->[buildAuditMessageForSuccess,buildAuditMessageForFailure,createDirectory],resolveBucketLink->[resolveBucketLink,getVolumeOwner,getBucketInfo,checkAcls,getRemoteUser],commitMultipartUploadPart->[buildAuditMessageForSuccess,commitMultipartUploadPart,buildAuditMessageForFailure],openKey->[checkAcls,openKey],restart->[instantiateServices,ScheduleOMMetricsWriteTask,start,getMetricsStorageFile,buildRpcServerStartMessage],listVolumeByUser->[hasAcls,getRemoteUser],setOwner->[checkAcls,setOwner],start->[start,getMetricsStorageFile,ScheduleOMMetricsWriteTask,buildRpcServerStartMessage],getDelegationToken->[isAllowedDelegationTokenOp,getRemoteUser],listStatus->[getClientAddress,buildAuditMessageForFailure,listStatus,buildAuditMessageForSuccess,checkAcls,getResourceType],listMultipartUploads->[buildAuditMessageForSuccess,listMultipartUploads,buildAuditMessageForFailure],checkAcls->[checkAcls,getRemoteUser],getBucketInfo->[checkAcls,getBucketInfo],listParts->[buildAuditMessageForFailure,buildAuditMessageForSuccess,listParts],commitKey->[checkAcls,commitKey],getFileStatus->[buildAuditMessageForSuccess,getFileStatus,buildAuditMessageForFailure,getClientAddress],getS3Secret->[getS3Secret,getRemoteUser],createFile->[buildAuditMessageForSuccess,createFile,buildAuditMessageForFailure],getObjectIdFromTxId->[getObjectIdFromTxId],omInit->[loginOMUserIfSecurityEnabled],removeAcl->[removeAcl,getResourceType,auditAcl,checkAcls],lookupFile->[getClientAddress,lookupFile,buildAuditMessageForFailure,buildAuditMessageForSuccess,checkAcls],setBucketProperty->[setBucketProperty,checkAcls],stop->[stop,stopSecretManager],getOMServiceId->[getOMServiceId],getVolumeOwner->[getVolumeOwner],listAllVolumes->[checkAcls],startJVMPauseMonitor->[start],getRpcServer->[startRpcServer],createOm->[OzoneManager],listTrash->[listTrash,checkAcls],createVolume->[createVolume],startSecretManagerIfNecessary->[startSecretManager,isOzoneSecurityEnabled],stopServices->[stop,stopSecretManager],listBuckets->[checkAcls,listBuckets],createBucket->[createBucket,checkAcls],getScmInfo->[getScmInfo],auditAcl->[buildAuditMessageForSuccess,buildAuditMessageForFailure],replaceOMDBWithCheckpoint->[createFile],checkVolumeAccess->[checkVolumeAccess,checkAcls],listKeys->[listKeys,checkAcls],setAcl->[getResourceType,auditAcl,checkAcls,setAcl],reloadOMState->[start,instantiateServices,saveOmMetrics],deleteVolume->[checkAcls,deleteVolume],getServiceInfo->[getServiceList],lookupKey->[lookupKey,checkAcls],installCheckpoint->[start,installCheckpoint],addAcl->[getResourceType,auditAcl,checkAcls,addAcl]]] | Install a checkpoint from the DB. This method is called when a new term is requested. It reloads the state machine updates. | Bharat, in this scenario where the OM state was not reloaded with new checkpoint, it would be better to keep the backup also in place. Just in case the OM needs to be manually reversed back to old state. |
@@ -75,6 +75,9 @@ class H2OJob(object):
if self.status == "CANCELLED":
raise H2OJobCancelled("Job<%s> was cancelled by the user." % self.job_key)
if self.status == "FAILED":
+ illegal_arg_ex_prefix = 'water.exceptions.H2OIllegalArgumentException: '
+ if self.exception and self.exception.startswith(illegal_arg_ex_prefix):
+ raise H2OValueError(self.exception[len(illegal_arg_ex_prefix):])
if (isinstance(self.job, dict)) and ("stacktrace" in list(self.job)):
raise EnvironmentError("Job with key {} failed with an exception: {}\nstacktrace: "
"\n{}".format(self.job_key, self.exception, self.job["stacktrace"]))
| [H2OJob->[poll_once->[poll],_refresh_job_status->[_query_job_status_safe]]] | Poll for a specific key in the job. | `H2OValueError` may be confusing for both dev and user, maybe we can introduce something like `H2OBackendValueError` to make things clear to everyone? |
@@ -141,7 +141,17 @@ App.WizardStep9Controller = App.WizardStepController.extend(App.ReloadPopupMixin
* Computed property to determine if the Retry button should be made visible on the page.
* @type {bool}
*/
- showRetry: Em.computed.equal('content.cluster.status', 'INSTALL FAILED'),
+ showRetry: function () {
+ const status = this.get('content.cluster.status');
+ switch (status) {
+ case 'INSTALL FAILED':
+ case 'INSTALLED':
+ case 'START FAILED':
+ return true;
+ }
+
+ return false;
+ }.property('content.cluster.status'),
/**
* Observer function: Calls {hostStatusUpdates} function once with change in a host status from any registered hosts.
| [No CFG could be retrieved] | Property to enable or disable the next step link if install task failed in installer wizard. The object that holds the state of the current open task. | currently BE is not correctly handling the state transition from "INSTALLED" to "STARTED", "START_FAILED" or "START_SKIPPED" so this is a temporary workaround. Once that is fixed this will need to be reverted. @mradha25 will open a jira for that. |
@@ -282,10 +282,11 @@ class Server:
if self.options.use_fine_grained_cache:
# Pull times and hashes out of the saved_cache and stick them into
# the fswatcher, so we pick up the changes.
- for meta, mypyfile, type_map in manager.saved_cache.values():
- if meta.mtime is None: continue
+ for state in self.fine_grained_manager.graph.values():
+ meta = state.meta
+ if meta is None: continue
self.fswatcher.set_file_data(
- mypyfile.path,
+ state.xpath,
FileData(st_mtime=float(meta.mtime), st_size=meta.size, md5=meta.hash))
# Run an update
| [Server->[check_default->[stats_summary,get_meminfo,build,append,get_stats,update,join,GcLogger],cmd_hang->[sleep],run_command->[getattr,method],check_fine_grained->[fine_grained_increment,initialize_fine_grained],cmd_status->[get_meminfo,update],cmd_check->[StringIO,create_source_list,getvalue,check],cmd_recheck->[check],update_sources->[add_watched_paths],check->[check_fine_grained,check_default],serve->[sendall,dump,write,getsockname,run_command,close,print_exception,pop,exc_info,getpid,unlink,receive,exit,isinstance,dumps,open,create_listening_socket,accept],initialize_fine_grained->[set_file_data,FileSystemCache,build,FineGrainedBuildManager,flush,update,FileData,update_sources,join,float,find_changed,values,FileSystemWatcher],fine_grained_increment->[time,flush,update,log,update_sources,format,join,find_changed],find_changed->[append,find_changed],create_listening_socket->[abspath,listen,bind,unlink,exists,socket],__init__->[unlink,isfile,process_options,exit]],get_meminfo->[getpid,memory_info,getrusage,Process],daemonize->[fork,umask,print,flush,waitpid,setsid,dup2,_exit,close,open,fileno,func]] | Initialize the fine - grained cache and manager. Returns a dict with the output of the last . | Using xpath is dubious. It may be `'<string>'`. I assume you need something that's not Optional, so I'd use path and add an assert that it's not None. |
@@ -103,6 +103,14 @@ class RunTracker:
def goals(self) -> list[str]:
return self._all_options.goals if self._all_options else []
+ @property
+ def active_standard_backends(self) -> list[str]:
+ return [
+ backend
+ for backend in self._all_options.for_global_scope().backend_packages
+ if backend.starts_with("pants.backend.")
+ ]
+
def start(self, run_start_time: float, specs: list[str]) -> None:
"""Start tracking this pants run."""
if self._has_started:
| [RunTracker->[end_run->[has_ended],get_anonymous_telemetry_data->[maybe_hash_with_repo_id_prefix]]] | Return a list of all goals that have been added to the run. | Should this be compared instead against an explicit list of backends to ensure that users that have custom backends on the PYTHONPATH and chose to put it in `pants.backend.*` do not have those backends reported? |
@@ -458,7 +458,8 @@ def run_rust_tests() -> None:
command = [
"build-support/bin/native/cargo",
"test",
- "--all",
+ # Until we can run fuse tests in osx CI, we omit --all.
+ #"--all",
# We pass --tests to skip doc tests, because our generated protos contain invalid doc tests in
# their comments.
"--tests",
| [run_doc_gen_tests->[_run_command],_run_command->[check_pants_pex_exists],run_jvm_tests->[_run_command],run_platform_specific_tests->[_run_command],TestTargetSets->[calculate->[get_listed_targets,TestTargetSets]],run_clippy->[_run_command],run_integration_tests_v2->[pants_command,_run_command,calculate],run_plugin_tests->[_run_command],run_lint->[_run_command],run_githooks->[_run_command],run_sanity_checks->[run_check,check_pants_pex_exists],run_unit_tests->[pants_command,_run_command,calculate],run_integration_tests_v1->[pants_command,_run_command,calculate],main] | Run rust tests. | (I know you're still experimenting at this stage) Would feel better if we could parameterize this function so that Linux shard still uses `--all`. |
@@ -101,6 +101,17 @@ public class CompositeFileListFilter<F> implements ReversibleFileListFilter<F>,
return this;
}
+ @Override
+ public void addDiscardCallback(DiscardCallback<F> discardCallback) {
+ this.discardCallback = discardCallback;
+ if (this.discardCallback != null) {
+ this.fileFilters
+ .stream()
+ .filter(DiscardAwareFileListFilter.class::isInstance)
+ .map(f -> (DiscardAwareFileListFilter<F>) f)
+ .forEach(f -> f.addDiscardCallback(discardCallback));
+ }
+ }
@Override
public List<F> filterFiles(F[] files) {
| [CompositeFileListFilter->[filterFiles->[filterFiles],remove->[remove],close->[close],addFilters->[addFilters],rollback->[rollback]]] | Add a collection of filters to this list. | We must only call the callback once per file - remember the composite filter passes all files to all filters. Or, change the queue to linked set. |
@@ -483,6 +483,8 @@ const (
DefaultNonMasqueradeCIDR = "0.0.0.0/0"
// DefaultKubeProxyMode is the default KubeProxyMode value
DefaultKubeProxyMode KubeProxyMode = KubeProxyModeIPTables
+ // DefaultWindowsSshEnabled is the default windowsProfile.sshEnabled value
+ DefaultWindowsSshEnabled = true
)
const (
| [No CFG could be retrieved] | DefaultContainerdVersion specifies the default containerd version of the container. A sequence of strings representing the name of the Azure Stack Cloud. | nit: prefer `DefaultWindowsSSHEnabled` (SSH instead of Ssh) |
@@ -26,6 +26,10 @@ from pants.util.frozendict import FrozenDict
from pants.util.meta import frozen_after_init
from pants.util.ordered_set import FrozenOrderedSet
+# -----------------------------------------------------------------------------------------------
+# Core Field abstractions
+# -----------------------------------------------------------------------------------------------
+
# Type alias to express the intent that the type should be immutable and hashable. There's nothing
# to actually enforce this, outside of convention. Maybe we could develop a MyPy plugin?
ImmutableValue = Any
| [Sources->[sanitize_raw_value->[InvalidFieldTypeException]],IntField->[compute_value->[InvalidFieldTypeException]],hydrate_sources->[prefix_glob_with_address,validate_snapshot,HydratedSources],StringOrStringSequenceField->[compute_value->[InvalidFieldTypeException]],StringSequenceField->[compute_value->[InvalidFieldTypeException]],StringField->[compute_value->[InvalidFieldTypeException]],BoolField->[compute_value->[InvalidFieldTypeException]],Target->[has_fields->[_has_fields],class_field_types->[_find_plugin_fields],__getitem__->[_maybe_get],get->[_maybe_get],_maybe_get->[_find_registered_field_subclass],class_has_fields->[_has_fields,class_field_types],_has_fields->[_find_registered_field_subclass]],FloatField->[compute_value->[InvalidFieldTypeException]]] | The base class constructor for all objects. A Field that does not need the engine in order to be hydrated. | `target.py` has grown pretty large. But, I'd prefer we keep it all in one file for now because it's all related and it greatly improves discoverability in call sites, imo - you don't have to guess which file something is defined in. These section headers are meant to help readers to quickly find things. |
@@ -163,6 +163,18 @@ public class SCMPipelineManager implements PipelineManager {
metrics.incNumPipelineCreated();
metrics.createPerPipelineMetrics(pipeline);
}
+ Pipeline overlapPipeline = RatisPipelineUtils
+ .checkPipelineContainSameDatanodes(stateManager, pipeline);
+ if (overlapPipeline != null) {
+ metrics.incNumPipelineContainSameDatanodes();
+ //TODO remove until pipeline allocation is proved equally distributed.
+ LOG.info("Pipeline: " + pipeline.getId().toString() +
+ " contains same datanodes as previous pipeline: " +
+ overlapPipeline.getId().toString() + " nodeIds: " +
+ pipeline.getNodes().get(0).getUuid().toString() +
+ ", " + pipeline.getNodes().get(1).getUuid().toString() +
+ ", " + pipeline.getNodes().get(2).getUuid().toString());
+ }
return pipeline;
} catch (IOException ex) {
metrics.incNumPipelineCreationFailed();
| [SCMPipelineManager->[addContainerToPipeline->[addContainerToPipeline],deactivatePipeline->[deactivatePipeline],waitPipelineReady->[getPipeline],removePipeline->[removePipeline],openPipeline->[openPipeline],scrubPipeline->[finalizeAndDestroyPipeline],triggerPipelineCreation->[triggerPipelineCreation],getPipeline->[getPipeline],destroyPipeline->[triggerPipelineCreation],close->[close],activatePipeline->[activatePipeline],finalizePipeline->[finalizePipeline],removeContainerFromPipeline->[removeContainerFromPipeline],getPipelines->[getPipelines],getNumberOfContainers->[getNumberOfContainers]]] | Creates a new pipeline of the given type and factor. | can we 1) put this with if (LOG.isDebugEnabled()) 2) change the log level to DEBUG 3) change to use parameterized log formatting to avoid performance penalty from logging as this is could affect write? |
@@ -33,6 +33,11 @@ Rails.application.configure do
:ses
end
+ if IdentityConfig.store.rails_mailer_previews_enabled
+ config.action_mailer.show_previews = true
+ config.action_mailer.preview_path ||= Rails.root.join('spec/mailers/previews')
+ end
+
routes.default_url_options[:protocol] = :https
# turn off IP spoofing protection since the network configuration in the production environment
| [new,check_yarn_integrity,compile,raise_delivery_errors,domain_name,consider_all_requests_local,asset_host,cache_classes,deprecation,dump_schema_after_migration,logger,redis_url,fallbacks,log_to_stdout,delivery_method,disable_email_sending,eager_load,ignore_actions,default_url_options,presence,cache_store,perform_caching,digest,proc,log_level,ip_spoofing_check,configure,mailer_domain_name,js_compressor] | This method is called by the action processor when the user is not authorized to access the resource. | we could guard this from prod even more with `&& Identity::Hostdata.domain != 'login.gov'` but maybe we'd want check in `dm` or `staging`? |
@@ -22,8 +22,8 @@ class MyPySourcePlugin(Target):
`pants-plugins/mypy_plugins/custom_plugin.py`, and you set `pants-plugins` as a source root,
then set `plugins = mypy_plugins.custom_plugin`. Set the `config`
option in the `[mypy]` scope to point to your MyPy config file.
- 5. Set the option `source_plugins` in the `[mypy]` scope to include this target's
- address, e.g. `source_plugins = ["build-support/mypy_plugins:plugin"]`.
+ 4. Set the option `source_plugins` in the `[mypy]` scope to include this target's
+ address, e.g. `source_plugins = ["pants-plugins/mypy_plugins:plugin"]`.
To instead load a third-party plugin, set the option `extra_requirements` in the `[mypy]`
scope (see https://www.pantsbuild.org/v2.0/docs/python-typecheck-goal). Set `plugins` in
| [No CFG could be retrieved] | A class that can be used to load a plugin from a source code. Get the current page. | Typos I realized while writing the MyPy docs. |
@@ -498,15 +498,7 @@ namespace System.Net.Http
_sendBuffer.Commit(2);
HttpMethod normalizedMethod = HttpMethod.Normalize(request.Method);
- if (normalizedMethod.Http3EncodedBytes != null)
- {
- BufferBytes(normalizedMethod.Http3EncodedBytes);
- }
- else
- {
- BufferLiteralHeaderWithStaticNameReference(H3StaticTable.MethodGet, normalizedMethod.Method);
- }
-
+ BufferBytes(normalizedMethod.Http3EncodedBytes);
BufferIndexedHeader(H3StaticTable.SchemeHttps);
if (request.HasHeaders && request.Headers.Host != null)
| [Http3RequestStream->[ReadNextDataFrameAsync->[CopyTrailersToResponseMessage],DisposeSyncHelper->[Dispose],OnHeader->[OnHeader],Http3ReadStream->[Read->[ReadResponseContent],Dispose->[Dispose],ReadAsync->[ValueTask,ReadResponseContentAsync],Dispose],Trace->[Trace],Http3WriteStream->[Dispose->[Dispose]],ReadFrameEnvelopeAsync->[GoAway],OnStaticIndexedHeader->[OnHeader],Dispose->[Dispose]]] | BufferHeaders - buffers the headers of the given request. Private methods. | @scalablecory, is this an ok change? |
@@ -17,7 +17,7 @@ def cli():
"""
pass
-
+# TODO: This may need to be depricated
@cli.command()
@click.argument("environment_file", type=click.Path(exists=True))
@click.option("--runner_kwargs", default={})
| [create_environment->[EnvironmentSchema,setup,loads,load,execute],run->[EnvironmentSchema,run,open,load,echo],Path,group,argument,command,option] | A CLI for running a flow from an environment file. | Two TODOs in this file, these CLI functions can probably be removed after refactor |
@@ -88,11 +88,13 @@ class WPSEO_Utils {
* @return bool
*/
public static function is_apache() {
- if ( isset( $_SERVER['SERVER_SOFTWARE'] ) && stristr( $_SERVER['SERVER_SOFTWARE'], 'apache' ) !== false ) {
- return true;
+ if ( ! isset( $_SERVER['SERVER_SOFTWARE'] ) ) {
+ return '';
}
- return false;
+ $software = sanitize_text_field( wp_unslash( $_SERVER['SERVER_SOFTWARE'] ) );
+
+ return stripos( $software, 'apache' ) !== false;
}
/**
| [WPSEO_Utils->[get_roles->[get_names],translate_score->[get_label,get_css_class],get_title_separator->[get_separator_options]]] | Check if the current server is Apache. | This is not a boolean return value |
@@ -331,6 +331,9 @@ class Jetpack_React_Page extends Jetpack_Admin_Page {
'setupWizardStatus' => Jetpack_Options::get_option( 'setup_wizard_status', 'not-started' ),
'isSafari' => $is_safari,
'doNotUseConnectionIframe' => Constants::is_true( 'JETPACK_SHOULD_NOT_USE_CONNECTION_IFRAME' ),
+ 'licensing' => array(
+ 'error' => Jetpack_Options::get_option( 'licensing_error', '' ),
+ ),
);
}
| [Jetpack_React_Page->[get_initial_state->[get_dismissed_jetpack_notices],add_noscript_head_meta->[add_fallback_head_meta]]] | Get the initial state of the API. Get all the information about the current theme and all of its themes that Infinite Scroll provides This function is used to build a JSON - ready array of all the necessary data. | It might make sense to move the `get_option()` calls into the `Licensing` class to encapsulate the storage. |
@@ -14,9 +14,11 @@ has_ssvep_data = partial(has_dataset, name='ssvep')
def data_path(
path=None, force_update=False, update_path=True,
download=True, verbose=None): # noqa: D103
- return _data_path(path=path, force_update=force_update,
- update_path=update_path, name='ssvep',
- download=download)
+ dataset_params = dict(name=ssvep)
+ processor = ssvep['processor'](extract_dir=path)
+ return fetch_dataset(dataset_params=dataset_params, processor=processor,
+ path=path, force_update=force_update,
+ update_path=update_path, download=download)
data_path.__doc__ = _data_path_doc.format(name='ssvep',
| [get_version->[_get_version],data_path->[_data_path],partial,format] | Return the path to the data directory of a given lease. | Option 2: Processor defined directly in `data_path()` |
@@ -201,11 +201,13 @@ feature 'IdV session' do
fill_out_phone_form_ok(good_phone_value)
click_idv_continue
+ page.find('.accordion').click
+
# success advances to next step
- expect(page).to have_content(t('idv.titles.review'))
+ expect(page).to have_content(t('idv.titles.session.review'))
expect(page).to have_content(second_ssn_value)
expect(page).to_not have_content(first_ssn_value)
- expect(page).to have_content(second_ccn_value)
+ expect(page).to_not have_content(second_ccn_value)
expect(page).to_not have_content(mortgage_value)
expect(page).to_not have_content(first_ccn_value)
expect(page).to have_content(good_phone_formatted)
| [complete_idv_profile_with_phone->[fill_in,t,fill_out_phone_form_ok,click_button],complete_idv_profile_fail->[click_button],visit,create,phone,let,to_not,feature,fill_out_phone_form_ok,it,complete_idv_profile_ok,idv_max_attempts,to,have_content,have_selector,before,click_button,click_link,scenario,select,t,require,it_behaves_like,include,sign_in_and_2fa_user,have_link,times,enter_correct_otp_code_for_user,have_css,fill_in,context,be_a,complete_idv_profile_with_phone,send_keys,eq,reload,and_return] | click on the page with the neccessary input field and check if the user has idv_finance_form_ok idv_finance_form_fin. | why is this spec expectation changing? |
@@ -106,6 +106,8 @@ func NewValidator(ctx context.Context, vch *config.VirtualContainerHostConfigSpe
v.LicenseIssues = template.HTML(fmt.Sprintf("%s<span class=\"error-message\">%s</span>\n", v.LicenseIssues, err))
}
}
+ log.Info(fmt.Sprintf("LicenseStatus set to: %s", v.LicenseStatus))
+ log.Info(fmt.Sprintf("LicenseIssues set to: %s", v.LicenseIssues))
//Network Connection Check
hosts := []string{
| [QueryVCHStatus->[Title,HTML,Join,Infof,Begin,ReadFile,Sprintf,PIDFileDir,End,Errorf,Base,Split],QueryDatastore->[Reference,DatastoreOrDefault,HTML,DefaultCollector,Infof,Sprintf,Sort,Errorf,Retrieve],Title,Close,Hostname,Info,HTML,Dial,End,CheckLicense,IsNil,QueryVCHStatus,QueryDatastore,ClearIssues,CheckFirewall,Infof,CreateFromVCHConfig,GetIssues,Begin,Sprintf,String] | log. Info - Setting version to v. Version - Setting hostname to v. Hostname - v. GetNetworkIssues = template. HTML. | You can replace `log.Info(fmt.Sprintf(...))` with `log.Infof(...)`. |
@@ -220,12 +220,11 @@
<% cache("article-show-scripts", expires_in: 8.hours) do %>
<script async>
- <%= TweetTag.script.html_safe %>
- <%= YoutubeTag.script.html_safe %>
+ <%# we consider these script safe for embedding as they come from our code %>
<%= PodcastTag.script.html_safe %>
- <%= GistTag.script.html_safe %>
- <%= RunkitTag.script.html_safe %>
<%= PollTag.script.html_safe %>
+ <%= RunkitTag.script.html_safe %>
+ <%= TweetTag.script.html_safe %>
</script>
<% end %>
| [No CFG could be retrieved] | This is a hack to work around the fact that the page doesn t have any script that. | this has no script at all |
@@ -475,6 +475,13 @@ export class ManualAdvancement extends AdvancementConfig {
return true;
}
+ if (
+ el.classList.contains('i-amphtml-story-screen-reader-back-button')
+ ) {
+ shouldHandleEvent = true;
+ return true;
+ }
+
return false;
},
/* opt_stopAt */ this.element_
| [No CFG could be retrieved] | Checks if an event should be handled by ManualAdvancement or ManualAdvance Checks if the can be shown on the tooltip. | Insider tip: you don't need this block if you add a `role="button"` to your button |
@@ -94,7 +94,6 @@ class Controls:
cp_bytes = self.CP.to_bytes()
params.put("CarParams", cp_bytes)
put_nonblocking("CarParamsCache", cp_bytes)
- put_nonblocking("LongitudinalControl", "1" if self.CP.openpilotLongitudinalControl else "0")
self.CC = car.CarControl.new_message()
self.AM = AlertManager()
| [Controls->[controlsd_thread->[step],step->[state_transition,update_events,publish_logs,state_control,data_sample]],main->[Controls,controlsd_thread],main] | Initialize object with basic configuration. Initialize a single node - level object. Initialize the object variables. | Should we remove the param from params.py. As far as I can tell it's unused now. |
@@ -110,6 +110,7 @@ class PotentialFlowSolver(FluidSolver):
# Kratos variables
self.main_model_part.AddNodalSolutionStepVariable(KratosMultiphysics.FLAG_VARIABLE) # Required for variational_distance_process
self.main_model_part.AddNodalSolutionStepVariable(KratosMultiphysics.DISTANCE)
+ self.main_model_part.AddNodalSolutionStepVariable(KratosMultiphysics.REACTION)
self.main_model_part.AddNodalSolutionStepVariable(KratosMultiphysics.DISTANCE_GRADIENT)
self.main_model_part.AddNodalSolutionStepVariable(KratosMultiphysics.NODAL_H) # Required for modify_distance_process
self.main_model_part.AddNodalSolutionStepVariable(KratosMultiphysics.VELOCITY) # Required for modify_distance_process
| [PotentialFlowSolver->[__init__->[PotentialFlowFormulation]]] | Adds variables to the main model part. | Do we need it as historical variable? |
@@ -391,6 +391,14 @@ public class WorkManagerImpl extends DefaultComponent implements WorkManager {
for (String id : workQueueConfig.getQueueIds()) {
deactivateQueue(workQueueConfig.get(id));
}
+ try {
+ if (!shutdown(1, TimeUnit.MINUTES)) {
+ log.error("Some processors are still active");
+ }
+ } catch (InterruptedException cause) {
+ Thread.currentThread().interrupt();
+ throw new NuxeoException("Interrupted while stopping", cause);
+ }
} else if (event.id == RuntimeServiceEvent.RUNTIME_IS_STANDBY) {
Framework.removeListener(this);
}
| [WorkManagerImpl->[registerWorkQueueDescriptor->[registerWorkQueueDescriptor],getExecutor->[init],schedule->[schedule,isQueuingEnabled,getCategoryQueueId,removeScheduled],WorkThreadPoolExecutor->[afterExecute->[signalCompletedWork],getScheduledOrRunningSize->[getWorkQueueIds],shutdownAndSuspend->[deactivateQueueMetrics],removeScheduled->[removeScheduled]],find->[find],isProcessingEnabled->[isProcessingEnabled],listWork->[listWork],enableProcessing->[enableProcessing,deactivateQueue,activateQueue],shutdown->[shutdownExecutors],listWorkIds->[listWorkIds],init->[handleEvent->[deactivateQueue,activateQueue],newWorkQueuing,WorkCompletionSynchronizer,initializeQueue],isQueuingEnabled->[isQueuingEnabled],shutdownQueue->[getExecutor],awaitCompletion->[waitForCompletedWork,signalCompletedWork,awaitCompletion],getWorkState->[getWorkState],getQueueSize->[getMetrics],scheduleAfterCommit->[WorkScheduling],noScheduledOrRunningWork->[getWorkQueueIds,getQueueSize,isProcessingEnabled,noScheduledOrRunningWork]]] | Initializes the queue. | later we should remove the ambiguous "processor" term which is not used in Nuxeo (only one occurence in workmanager) |
@@ -61,14 +61,11 @@ public class JTextAreaOptionPane {
windowFrame.getContentPane().add(m_label, BorderLayout.NORTH);
windowFrame.getContentPane().add(new JScrollPane(editor), BorderLayout.CENTER);
windowFrame.getContentPane().add(okButton, BorderLayout.SOUTH);
- okButton.addActionListener(new ActionListener() {
- @Override
- public void actionPerformed(final ActionEvent e) {
- if (JTextAreaOptionPane.this.countDownLatch != null) {
- JTextAreaOptionPane.this.countDownLatch.countDown();
- }
- dispose();
+ okButton.addActionListener(e -> {
+ if (JTextAreaOptionPane.this.countDownLatch != null) {
+ JTextAreaOptionPane.this.countDownLatch.countDown();
}
+ dispose();
});
}
| [JTextAreaOptionPane->[appendNewLine->[append],dispose->[dispose],append->[append],countDown->[setWidgetActivation]]] | This method is called when the user clicks on the count down button. | `JTextAreaOptionPane.this` is not needed when using lambda expressions |
@@ -0,0 +1,13 @@
+package games.strategy.triplea.delegate;
+
+import java.io.Serializable;
+import java.util.List;
+
+class AbstractMoveExtendedDelegateState implements Serializable {
+ private static final long serialVersionUID = -4072966724295569322L;
+ Serializable superState;
+ // add other variables here:
+ public boolean m_nonCombat;
+ public List<UndoableMove> m_movesToUndo;
+ public MovePerformer m_tempMovePerformer;
+}
| [No CFG could be retrieved] | No Summary Found. | @DanVanAtta Shouldn't this class be public then? |
@@ -56,7 +56,7 @@ class DolibarrApi
$this->db = $db;
$production_mode = ( empty($conf->global->API_PRODUCTION_MODE) ? false : true );
$this->r = new Restler($production_mode, $refreshCache);
-
+ $this->r->setBaseUrls(DOL_MAIN_URL_ROOT, $urlwithroot);
$this->r->setAPIVersion(1);
}
| [DolibarrApi->[_cleanObjectDatas->[_cleanObjectDatas]]] | Constructor for a missing node. | Sure, but $urlwithroot should be defined :-) |
@@ -578,4 +578,16 @@ class Featured_Content {
Featured_Content::setup();
+function jetpack_update_featured_content_for_split_terms( $old_term_id, $new_term_id, $term_taxonomy_id, $taxonomy ) {
+ $fc_settings = get_option( 'featured-content', array() );
+
+ // Check to see whether the stored tag ID is the one that's just been split.
+ if ( isset( $fc_settings['tag-id'] ) && $old_term_id == $fc_settings['tag-id'] && 'post_tag' == $taxonomy ) {
+ // We have a match, so we swap out the old tag ID for the new one and resave the option.
+ $fc_settings['tag-id'] = $new_term_id;
+ update_option( 'featured-content', $fc_settings );
+ }
+}
+add_action( 'split_shared_term', 'jetpack_update_featured_content_for_split_terms', 10, 4 );
+
} // end if ( ! class_exists( 'Featured_Content' ) && isset( $GLOBALS['pagenow'] ) && 'plugins.php' !== $GLOBALS['pagenow'] ) {
| [Featured_Content->[customize_register->[add_control,add_setting,add_section],pre_get_posts->[get,is_main_query,set,is_home]]] | Setup featured_content. | I would spell out fc. |
@@ -15,7 +15,7 @@ namespace System.Windows.Forms {
/// for this dll have been localized to a RTL language.
/// </para>
/// </devdoc>
- internal sealed class RTLAwareMessageBox {
+ public sealed class RTLAwareMessageBox {
/// <include file='doc\MessageBox.uex' path='docs/doc[@for="MessageBox.Show6"]/*' />
/// <devdoc>
| [RTLAwareMessageBox->[DialogResult->[Show,RtlReading,IsRTLResources,RightAlign],RTL]] | Creates a new MessageBox with the specified parameters. | Can you explain why this had to be / why you wanted to make this public? |
@@ -2233,8 +2233,9 @@ public:
| FunctionInfo::Attributes::CapturesThis
| FunctionInfo::Attributes::Generator
| FunctionInfo::Attributes::ClassConstructor
- | FunctionInfo::Attributes::ClassMethod)) == 0,
- "Only the ErrorOnNew|SuperReference|Lambda|CapturesThis|Generator|ClassConstructor|Async|ClassMember attributes should be set on a serialized function");
+ | FunctionInfo::Attributes::ClassMethod
+ | FunctionInfo::Attributes::EnclosedByGlobalFunc)) == 0,
+ "Only the ErrorOnNew|SuperReference|Lambda|CapturesThis|Generator|ClassConstructor|Async|ClassMember|EnclosedByGlobalFunc attributes should be set on a serialized function");
PrependInt32(builder, _u("Offset Into Source"), sourceDiff);
if (function->GetNestedCount() > 0)
| [No CFG could be retrieved] | Adds a builder to the given function. Adds a serializable field to the defined fields. | I would really like for there to be some static FunctionInfo::SerializableFunctionAttributes FunctionInfo::NonSerializableFunctionAttributes that would break at compile time when we add a new attribute, but that's out of the scope of this. |
@@ -56,13 +56,16 @@ func newConfigCmd() *cobra.Command {
return err
}
- return listConfig(stack, showSecrets)
+ return listConfig(stack, showSecrets, jsonOut)
}),
}
cmd.Flags().BoolVar(
&showSecrets, "show-secrets", false,
"Show secret values when listing config instead of displaying blinded values")
+ cmd.Flags().BoolVarP(
+ &jsonOut, "json", "j", false,
+ "Emit output as JSON")
cmd.PersistentFlags().StringVarP(
&stack, "stack", "s", "",
"The name of the stack to operate on. Defaults to the current stack")
| [StringVar,Value,GetLatestConfiguration,NewBlindingDecrypter,PrintTable,Fd,RemoveTralingNewline,SaveProjectStack,PasswordStrength,DetectProjectStackPath,Wrap,Secure,HasSecureValue,ReadConsoleNoEcho,IsNotExist,SpecificArgs,Stat,Save,MatchString,RunFunc,StringVarP,Errorf,Namespace,Ref,MustCompile,AddCommand,GetGlobalColorization,NewPanicCrypter,Contains,Name,PersistentFlags,ReadConsole,GetStackCrypter,DetectProjectStack,Sort,DetectProject,ParseKey,NewSecureValue,NewValue,Printf,Sprintf,EncryptValue,RangeArgs,IsTerminal,LoadProjectStack,ReadAll,BoolVarP,Rename,Flags,BoolVar] | newConfigCmd returns a command to manage a specific configuration key. newConfigGetCmd returns a command to get a single configuration value. | Super nitpick (sorry): I'm OK with `"Emit output as JSON"`, but just wanted to point out that we're slightly inconsistent with the wording used in other commands. - `pulumi stack output` uses `"Emit outputs as JSON"`, which makes sense because it's all about "outputs". I believe this was the first command that had support for emitting as JSON. - `pulumi logs` and `pulumi stack ls` use `"Emit outputs as JSON"`, which looks like it was just copy/pasted from `pulumi stack output`. - `pulumi stack tag` I chose to use `"Emit stack tags as JSON"`. It might be worth standardizing on `"Emit output as JSON"` across the board, or using the appropriate noun for each type of thing that is being outputted, e.g. `"Emit config as JSON"`, `"Emit logs as JSON"`, `"Emit stacks as JSON"`, etc. Edit: I see you updated it for `pulumi stack ls`. |
@@ -54,11 +54,13 @@ var mainCommands = []*cobra.Command{
podCommand.Command,
_pullCommand,
_pushCommand,
+ _rmCommand,
&_rmiCommand,
_runCommand,
_saveCommand,
_stopCommand,
_tagCommand,
+ _umountCommand,
_versionCommand,
_waitCommand,
imageCommand.Command,
| [NewSyslogHook,StringVar,Umask,StartCPUProfile,IsRootless,StopCPUProfile,StringSliceVar,Close,SetGlobalTracer,MarkHidden,GetRunningContainers,Info,HasPrefix,Exit,Atoi,IntVar,Setrlimit,Error,ReadFile,Init,ContextWithSpan,BecomeRootInUserNS,BoolVar,ParseLevel,GetRuntime,StartSpan,Errorf,AddCommand,Wrapf,Create,Execute,Geteuid,Shutdown,AddHook,Config,JoinUserAndMountNS,Flag,Finish,Background,SetLevel,Getrlimit,PersistentFlags,OnInitialize,SetXdgRuntimeDir] | Initialization of the cobra. Deprecated due to conflict with - c. | Should unmount be moved as well? |
@@ -649,7 +649,7 @@ int ScriptApiSecurity::sl_g_loadfile(lua_State *L)
lua_pop(L, 1);
if (script->getType() == ScriptingType::Client) {
- std:: string display_path = lua_tostring(L, 1);
+ std:: string display_path = readParam<std::string>(L, 1);
const std::string *path = script->getClient()->getModFile(display_path);
if (!path) {
std::string error_msg = "Coudln't find script called:" + display_path;
| [No CFG could be retrieved] | This function checks if a chunk of code is available in lua lua file load - load function. | Space after `::` |
@@ -50,7 +50,9 @@ namespace Dynamo.ViewModels
}
private readonly IPreferences setting;
-
+ private readonly PackageLoader packageLoader;
+ private readonly LoadPackageParams loadPackageParams;
+ private readonly CustomNodeManager customNodeManager;
public DelegateCommand AddPathCommand { get; private set; }
public DelegateCommand DeletePathCommand { get; private set; }
| [PackagePathViewModel->[RaiseCanExecuteChanged->[RaiseCanExecuteChanged],ShowFileDialog->[OnRequestShowFileDialog],UpdatePathAt->[ShowFileDialog],RemovePathAt->[RaiseCanExecuteChanged],InsertPath->[RaiseCanExecuteChanged]]] | A view model that displays the path to the custom packages folder. CanDelete - checks if the object can be deleted or moved up. | Can you convert this setting to a property to return the value from LoadedPackageParams. |
@@ -131,8 +131,8 @@ public class ObserverMethodImpl<T, X> implements ObserverMethod<T> {
}
injectionPoints.add(injectionPoint);
}
- this.injectionPoints = immutableGuavaSet(injectionPoints);
- this.newInjectionPoints = immutableGuavaSet(newInjectionPoints);
+ this.injectionPoints = immutableSet(injectionPoints);
+ this.newInjectionPoints = immutableSet(newInjectionPoints);
this.isStatic = observer.isStatic();
this.eventMetadataRequired = initMetadataRequired(this.injectionPoints);
this.notificationStrategy = MethodInvocationStrategy.forObserver(observerMethod, beanManager);
| [ObserverMethodImpl->[equals->[getId,equals],initialize->[checkObserverMethod],hashCode->[hashCode],toString->[toString],sendEvent->[sendEvent]]] | Creates a unique id for the given observer method. Returns the unique id of the object. | `ImmutableSet.Builder` for `injectionPoints`. `newInjectionPoints` seem to be unsued - remove pls. |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.