patch
stringlengths
18
160k
callgraph
stringlengths
4
179k
summary
stringlengths
4
947
msg
stringlengths
6
3.42k
@@ -1315,7 +1315,15 @@ namespace System.Runtime.Serialization return (param == null) ? SR.Format(message, nestedMessage) : SR.Format(message, nestedMessage, param); } - private static void FindCollectionMethodsOnInterface(Type type, Type interfaceType, ref MethodInfo? addMethod, ref MethodInfo? getEnumeratorMethod) + [UnconditionalSuppressMessage("ReflectionAnalysis", "IL2075:MakeGenericMethod", + Justification = "Warnings about calling getMethod on the interface type are safe since we are annotating both the type and the " + + "interface with DynamicallyAccessedMembers.")] + private static void FindCollectionMethodsOnInterface( + [DynamicallyAccessedMembers(DynamicallyAccessedMemberTypes.All)] + Type type, + [DynamicallyAccessedMembers(DynamicallyAccessedMemberTypes.PublicMethods)] + Type interfaceType, + ref MethodInfo? addMethod, ref MethodInfo? getEnumeratorMethod) { Type? t = type.GetInterfaces().Where(it => it.Equals(interfaceType)).FirstOrDefault(); if (t != null)
[CollectionDataContract->[IsCollection->[IsCollection],IncrementCollectionCount->[IncrementCollectionCount],GetCollectionMethods->[GetTargetMethodWithName,FindCollectionMethodsOnInterface],IsCollectionOrTryCreate->[IsArraySegment,IsCollection],GenericDictionaryEnumerator->[MoveNext->[MoveNext],Reset->[Reset]],DictionaryEnumerator->[MoveNext->[MoveNext],Reset->[Reset]],ReadXmlValue->[XmlFormatGetOnlyCollectionReaderDelegate],CollectionDataContractCriticalHelper->[IncrementCollectionCount->[IncrementCollectionCount],Init],XmlFormatGetOnlyCollectionReaderDelegate]]
Get invalid collection message.
This shouldn't be `MakeGenericMethod` #Resolved
@@ -353,6 +353,16 @@ func (g *GlobalContext) ConfigureConfig() error { func (g *GlobalContext) ConfigReload() error { err := g.ConfigureConfig() g.ConfigureUpdaterConfig() + g.ConfigureDeviceCloneState() + return err +} + +func (g *GlobalContext) ConfigureDeviceCloneState() error { + c := NewDeviceCloneStateJSONFile(g) + err := c.Load(false) + if err != nil { + g.Log.Debug("Failed to open device clone state json file: %s\n", err) + } return err }
[GetDataDir->[GetDataDir],GetCacheDir->[GetCacheDir],LogoutSelfCheck->[Logout],BustLocalUserCache->[GetFullSelfer,GetUPAKLoader],Configure->[SetCommandLine,ConfigureLogging],Shutdown->[shutdownCachesLocked,Shutdown],GetMeUV->[GetUPAKLoader],ConfigReload->[ConfigureConfig],GetLogDir->[GetLogDir],GetStoredSecretServiceName->[GetStoredSecretServiceName],GetRuntimeDir->[GetRuntimeDir],StartStandaloneChat->[StartStandaloneChat],SetCommandLine->[SetCommandLine],IsOneshot->[IsOneshot],CallLoginHooks->[GetFullSelfer],KeyfamilyChanged->[BustLocalUserCache],ConfigureUsage->[ConfigureKeyring,ConfigureAPI,ConfigureCaches,UseKeyring,ConfigureMerkleClient,ConfigureTimers,ConfigureExportedStreams,Configure,ConfigReload],configureMemCachesLocked->[shutdownCachesLocked],GetServerURI->[GetServerURI],ConfigureCaches->[configureMemCachesLocked],GetKBFSInfoPath->[GetKBFSInfoPath],GetRunMode->[GetRunMode],GetServiceInfoPath->[GetServiceInfoPath],GetMountDir->[GetMountDir],UserChanged->[BustLocalUserCache],GetAppType->[GetAppType],GetConfigDir->[GetConfigDir],FlushCaches->[configureMemCachesLocked],GetStoredSecretAccessGroup->[GetStoredSecretAccessGroup],GetUsersWithStoredSecrets->[GetUsersWithStoredSecrets],Init]
ConfigReload reloads the configuration from the file system.
why open this here? why not just open it whenever you need to access it?
@@ -272,13 +272,13 @@ class ResultsController < ApplicationController def expand_criteria @assignment = Assignment.find(params[:aid]) @rubric_criteria = @assignment.rubric_criteria - render :partial => 'results/marker/expand_criteria', :locals => {:rubric_criteria => @rubric_criteria} + render partial: 'results/marker/expand_criteria', locals: {rubric_criteria: @rubric_criteria} end def collapse_criteria @assignment = Assignment.find(params[:aid]) @rubric_criteria = @assignment.rubric_criteria - render :partial => 'results/marker/collapse_criteria', :locals => {:rubric_criteria => @rubric_criteria} + render partial: 'results/marker/collapse_criteria', locals: {rubric_criteria: @rubric_criteria} end def expand_unmarked_criteria
[ResultsController->[create->[redirect_to,id,find,get_result_used,new,has_result?,submission,result_version_used,save,marking_state],add_extra_mark->[new,find,unit,render,update_attributes,result,post?,save],view_marks->[submission_files,find,percentage,accepted_grouping_for,first,rubric_criteria,released_to_students,get_remark_result,group,nil?,find_or_create_by_result_id_and_rubric_criterion_id,points,remark_submitted?,get_submission_used,each,marking_state,id,redirect_to,render,annotation_categories,has_submission?,has_result?,get_original_result,save],edit->[submission_files,admin?,find,percentage,first,rubric_criteria,released_to_students,assignment,group,find_or_create_by_result_id_and_rubric_criterion_id,points,groupings,submission,grouping,ta?,each,id,index,push,annotation_categories,save],next_grouping->[redirect_to,has_submission?,id,find],codeviewer->[call,nil?,find,render,get_latest_result,student?,message,raise,retrieve_file,get_file_type,annotations],update_overall_comment->[overall_comment,save,find],expand_criteria->[rubric_criteria,render,find],download->[redirect_to,send_data,find,message,filename,retrieve_file,id],collapse_criteria->[rubric_criteria,render,find],expand_unmarked_criteria->[rubric_criteria,render,all,find],retrieve_file->[group,nil?,repo,path,repository_name,download_as_string,filename,files_at_path,raise,get_revision,revision_number],update_extra_mark->[find,render,get_bonus_marks,to_json,calculate_total,replace_html,total_mark,result_id,valid?,save,id,get_deductions],update_mark->[call,to_s,render,find,rubric_criterion,weight,replace_html,mark,total_mark,get_subtotal,errors,save,id],remove_extra_mark->[render,result,find,destroy],before_filter]]
This method expands the criteria and the unmarked ones into a single single node.
Line is too long. [97/80]<br>Space inside { missing.<br>Space inside } missing.
@@ -18,7 +18,16 @@ limitations under the License. -%> package <%= packageName %>.config; +<%_ + let springFox3 = false; + if (reactive && (applicationType === 'gateway' || applicationType === 'monolith')) { + springFox3 = true; + } +_%> + +<%_ if (!springFox3) { _%> import com.google.common.base.Predicates; +<%_ } _%> import io.github.jhipster.config.JHipsterConstants; import io.github.jhipster.config.JHipsterProperties; import io.github.jhipster.config.apidoc.customizer.SwaggerCustomizer;
[No CFG could be retrieved]
Imports a single from the JHipster project. OpenAPI configuration for the first .
This will generate an extra empty line, could you remove it?
@@ -78,10 +78,12 @@ class _CloudXNSLexiconClient(dns_common_lexicon.LexiconClient): self.provider = cloudxns.Provider(config) - def _handle_http_error(self, e, domain_name): + def _handle_http_error(self, e: HTTPError, domain_name: str) -> errors.PluginError: hint = None if str(e).startswith('400 Client Error:'): hint = 'Are your API key and Secret key values correct?' - return errors.PluginError('Error determining zone identifier for {0}: {1}.{2}' - .format(domain_name, e, ' ({0})'.format(hint) if hint else '')) + hint_disp = f' ({hint})' if hint else '' + + return errors.PluginError(f'Error determining zone identifier for {domain_name}: ' + f'{e}.{hint_disp}')
[_CloudXNSLexiconClient->[__init__->[build_lexicon_config,super,Provider],_handle_http_error->[str,,PluginError,format]],Authenticator->[add_parser_arguments->[add,super],_perform->[_get_cloudxns_client],_setup_credentials->[_configure_credentials,format],_get_cloudxns_client->[conf,Error,_CloudXNSLexiconClient],_cleanup->[_get_cloudxns_client],__init__->[super]],getLogger]
Initialize a object.
`LexiconClient._handle_http_error` is annotated with `Optional[errors.PluginError]`. How does this work for Python typing? Does an overridden function returning`T` satisfy the original function's `Optional[T]`?
@@ -1545,8 +1545,9 @@ public class MoveValidator { data.getMap().getRoute_IgnoreEnd(start, end, Match.allOf(Matches.TerritoryIsWater, noImpassable)); if (waterRoute != null && ((waterRoute.getLargestMovementCost(unitsWhichAreNotBeingTransportedOrDependent) <= defaultRoute - .getLargestMovementCost(unitsWhichAreNotBeingTransportedOrDependent)) || (forceLandOrSeaRoute && Match - .anyMatch(unitsWhichAreNotBeingTransportedOrDependent, Matches.UnitIsSea)))) { + .getLargestMovementCost(unitsWhichAreNotBeingTransportedOrDependent)) || (forceLandOrSeaRoute + && Match + .anyMatch(unitsWhichAreNotBeingTransportedOrDependent, Matches.UnitIsSea)))) { defaultRoute = waterRoute; mustGoSea = true; }
[MoveValidator->[validateTransport->[onlyIgnoredUnitsOnPath,getTerritoryTransportHasUnloadedTo,getEditMode,isLoad],nonParatroopersPresent->[allLandUnitsAreBeingParatroopered],getNeutralCharge->[getNeutralCharge],isNeutralsBlitzable->[isNeutralsImpassable],carrierMustMoveWith->[carrierMustMoveWith],validateParatroops->[getParatroopsRequiringTransport,getEditMode],validateCanal->[validateCanal],getEditMode->[getEditMode]]]
Get the best route for the given territory. Returns a route with the most likely impassable conditions. This method checks if the route is in the water route or if it is in the water Returns the route that matches the given tests.
It might be easier to read if lines 1553 and 1554 were combined (doesn't look like it will exceed max line length).
@@ -523,4 +523,14 @@ defineSuite([ s.destroy(); }).toThrow(); }); + + it('fails with built-in function circular dependency', function() { + var vs = 'void main() { }'; + var fs = 'void main() { czm_circularDependency1(); gl_FragColor = vec4(1.0); }'; + + + expect(function() { + sp = context.createShaderProgram(vs, fs); + }).toThrow(); + }); }, 'WebGL'); \ No newline at end of file
[No CFG could be retrieved]
Destroy WebGL Object.
It kinda doesn't matter, but this should contain `gl_Position = vec4(0.0);`, otherwise I believe the compile/link will fail, which would throw an exception.
@@ -49,7 +49,7 @@ export class AmpSlides extends BaseCarousel { this.slides_.forEach((slide, i) => { this.setAsOwner(slide); // Only the first element is initially visible. - slide.style.display = i > 0 ? 'none' : 'block'; + slide.style.visibility = i > 0 ? 'hidden' : 'visible'; this.applyFillContent(slide); });
[No CFG could be retrieved]
Creates an AmpSlides containing the children of the current element. Callback for the next .
I just need some context here: `display: none` is much better from a performance pov. But if we do want to render them, then this change is good.
@@ -159,6 +159,18 @@ class ConstantInitializer(Initializer): Returns: the initialization op """ + if in_dygraph_mode(): + out_dtype = var.dtype + attrs = { + "shape": var.shape, + "dtype": int(out_dtype), + "value": float(self._value), + 'force_cpu': self._force_cpu or force_init_on_cpu() + } + outputs = {"Out": [var]} + core.ops.fill_constant({}, attrs, outputs) + return None + assert isinstance(var, framework.Variable) assert isinstance(block, framework.Block)
[init_on_cpu->[force_init_on_cpu],MSRAInitializer->[__call__->[_compute_fans]],ConstantInitializer->[__call__->[force_init_on_cpu]],XavierInitializer->[__call__->[_compute_fans]]]
Adds constant initialization ops for a variable in a given block.
in_dygraph_mode __init__, self.is_dygraph_mode = in_dygraph_mode, if self.is_dygraph_mode
@@ -6,10 +6,12 @@ import ( "time" kapi "k8s.io/kubernetes/pkg/api" + "k8s.io/kubernetes/pkg/api/unversioned" kcoreclient "k8s.io/kubernetes/pkg/client/clientset_generated/internalclientset/typed/core/internalversion" "k8s.io/kubernetes/pkg/fields" "k8s.io/kubernetes/pkg/watch" + "github.com/golang/glog" "github.com/openshift/origin/pkg/deploy/api" deployutil "github.com/openshift/origin/pkg/deploy/util" )
[Until,AsSelector,Stop,New,ReplicationControllers,Watch,Errorf,DeploymentStatusFor]
WaitForRunningDeployment waits until the specified object is no longer New or Pending or an error is Returns true if the deployment phase is observed.
Isn't this error typed in the watch library?
@@ -203,7 +203,7 @@ class mailing_fraise extends MailingTargets $sql.= " a.lastname, a.firstname,"; $sql.= " a.datefin, a.civility as civility_id, a.login, a.societe"; // Other fields $sql.= " FROM ".MAIN_DB_PREFIX."adherent as a, ".MAIN_DB_PREFIX."adherent_type as ta"; - $sql.= " WHERE a.email <> ''"; // Note that null != '' is false + $sql.= " WHERE a.email <> '' AND a.entity IN (".getEntity('member').")"; // Note that null != '' is false $sql.= " AND a.email NOT IN (SELECT email FROM ".MAIN_DB_PREFIX."mailing_cibles WHERE fk_mailing=".$mailing_id.")"; if (isset($_POST["filter"]) && $_POST["filter"] == '-1') $sql.= " AND a.statut=-1"; if (isset($_POST["filter"]) && $_POST["filter"] == '1a') $sql.= " AND a.statut=1 AND a.datefin >= '".$this->db->idate($now)."'";
[mailing_fraise->[add_to_target->[query,error,transnoentities,num_rows,fetch_object,jdate,url,idate,load],formFilter->[query,trans,num_rows,fetch_object,select_date,load],getSqlArrayForStats->[trans,escape,load]]]
Add a member to a target add_to_target add_cibles add_civilers add_ add_to_target add_to_target function.
hello @ptibogxiv it's better to filter with entity before all for mysql performance `$sql.= " WHERE a.entity IN (".getEntity('member').") AND a.email <> ''";`
@@ -2136,6 +2136,7 @@ public class StorageManagerImpl extends ManagerBase implements StorageManager, C if (hostIds.isEmpty()) { return null; } + Collections.shuffle(hostIds); for (Long hostId : hostIds) { Host host = _hostDao.findById(hostId);
[StorageManagerImpl->[createChildDatastoreVO->[getStoragePoolTags],updateStoragePool->[enablePrimaryStoragePool,disablePrimaryStoragePool,updateStoragePool],sendToPool->[getUpHostsInPool,sendToPool],canHostAccessStoragePool->[canHostAccessStoragePool],discoverImageStore->[getName],createCapacityEntry->[createCapacityEntry,getStorageOverProvisioningFactor],syncDatastoreClusterStoragePool->[getStoragePoolTags],getUsedSize->[canPoolProvideStorageStats,getStoragePoolStats],cleanupStorage->[cleanupInactiveTemplates],getVolumeStats->[getVolumeStats],setDiskProfileThrottling->[getDiskIopsReadRate,getDiskBytesWriteRate,getDiskIopsWriteRate,getDiskBytesReadRate],getBytesRequiredForTemplate->[getBytesRequiredForTemplate],setVolumeObjectTOThrottling->[getDiskIopsReadRate,getDiskBytesWriteRate,getDiskIopsWriteRate,getDiskBytesReadRate],migrateToObjectStore->[discoverImageStore,migrateToObjectStore],checkPoolforSpace->[getStorageOverProvisioningFactor],getHypervisorType->[getHypervisorType],updateSecondaryStorage->[getHost],getDataObjectSizeIncludingHypervisorSnapshotReserve->[getDataObjectSizeIncludingHypervisorSnapshotReserve],createSecondaryStagingStore->[getName],checkUsagedSpace->[canPoolProvideStorageStats],isStoragePoolCompliantWithStoragePolicy->[getUpHostsInPool],StorageGarbageCollector->[runInContext->[cleanupStorage]],registerSystemVmTemplateOnFirstNfsStore->[doInTransactionWithoutResult->[getValidTemplateName]],updateStorageCapabilities->[getHypervisorType],storagePoolHasEnoughSpace->[checkUsagedSpace,storagePoolHasEnoughSpace],storagePoolHasEnoughSpaceForResize->[checkUsagedSpace]]]
Finds up and enabled host with access to storage pools.
I think we could add some log to this method, about which host was selected (or not).
@@ -1695,8 +1695,7 @@ public class LibvirtComputingResource extends ServerResourceBase implements Serv if (nic.getBrName().equalsIgnoreCase(_linkLocalBridgeName)) { broadcastUriAllocatedToVM.put("LinkLocal", nicPos); } else { - if (nic.getBrName().equalsIgnoreCase(_publicBridgeName) || nic.getBrName().equalsIgnoreCase(_privBridgeName) || - nic.getBrName().equalsIgnoreCase(_guestBridgeName)) { + if (nic.getBrName() == null) { broadcastUriAllocatedToVM.put(BroadcastDomainType.Vlan.toUri(Vlan.UNTAGGED).toString(), nicPos); } else { final String broadcastUri = getBroadcastUriFromBridge(nic.getBrName());
[LibvirtComputingResource->[getInterfaces->[getInterfaces],isSnapshotSupported->[executeBashScript],getNetworkStats->[networkUsage],cleanupVMNetworks->[getAllVifDrivers],cleanupNetworkElementCommand->[vifHotUnPlug,VifHotPlug,getBroadcastUriFromBridge],getVPCNetworkStats->[configureVPCNetworkUsage],getVmNetworkStat->[getInterfaces,getDomain],configure->[getDefaultKvmScriptsDir,getDefaultNetworkScriptsDir,configure,getDeveloperProperties,getDefaultStorageScriptsDir,getDefaultHypervisorScriptsDir,getDefaultDomrScriptsDir],getVifDriverClass->[configure],destroyNetworkRulesForVM->[getInterfaces],createVMFromSpec->[getUuid],vifHotUnPlug->[getAllVifDrivers],post_default_network_rules->[getInterfaces],checkBridgeNetwork->[matchPifFileInDirectory],configureTunnelNetwork->[findOrCreateTunnelNetwork],getDisks->[getDisks],defaultNetworkRules->[getInterfaces],rebootVM->[getPif,startVM],attachOrDetachISO->[cleanupDisk],prepareNetworkElementCommand->[VifHotPlug,getBroadcastUriFromBridge],findOrCreateTunnelNetwork->[checkNetwork],getVmState->[convertToPowerState],getVersionStrings->[KeyValueInterpreter,getKeyValues],attachOrDetachDisk->[getUuid],initialize->[getVersionStrings,getUuid],executeInVR->[executeInVR],getHostVmStateReport->[convertToPowerState,getHostVmStateReport],getVmDiskStat->[getDomain,getDisks],getDeveloperProperties->[getEndIpFromStartIp],getVncPort->[getVncPort],getVmStat->[getInterfaces,getDomain,VmStats,getDisks],createVbd->[getUuid],syncNetworkGroups->[getRuleLogsForVms],getBroadcastUriFromBridge->[matchPifFileInDirectory]]]
This method is called when a network element command is to be executed. Checks if there is a node in the chain.
@konstantintrushin can you consider checking using `Strings.isNullOrEmpty`?
@@ -1179,7 +1179,8 @@ class PostProcessor(object): # pylint: disable=too-many-instance-attributes with cur_ep.lock: cur_ep.location = ek(os.path.join, dest_path, new_file_name) # download subtitles - if sickbeard.USE_SUBTITLES and ep_obj.show.subtitles: + if sickbeard.USE_SUBTITLES and ep_obj.show.subtitles \ + and (cur_ep.season != 0 or cur_ep.season == 0 and sickbeard.SUBTITLES_INCLUDE_SPECIALS): cur_ep.refreshSubtitles() cur_ep.download_subtitles(force=True) sql_l.append(cur_ep.get_sql())
[PostProcessor->[_find_info->[_analyze_name,_log],_get_quality->[_log],process->[_get_quality,_find_info,_move,_checkForExistingFile,_get_ep_obj,_log,_copy,_hardlink,_moveAndSymlink,_symlink,_run_extra_scripts,_delete,_is_priority,_add_to_anidb_mylist],_move->[_int_move->[_log],_combined_file_operation],list_associated_files->[_log,recursive_glob],_checkForExistingFile->[_log],_hardlink->[_int_hard_link->[_log],_combined_file_operation],_copy->[_int_copy->[_log],_combined_file_operation],_get_ep_obj->[_log],_combined_file_operation->[list_associated_files,_log],_moveAndSymlink->[_int_move_and_sym_link->[_log],_combined_file_operation],_symlink->[_int_sym_link->[_log],_combined_file_operation],_history_lookup->[_log],_delete->[list_associated_files,_log],_analyze_name->[_finalize],_is_priority->[_log],_run_extra_scripts->[_log],_add_to_anidb_mylist->[_build_anidb_episode,_log]]]
Post - process a given file and return True if the file is a valid one Check if a is available for this episode. check if the file has a specific season and if so delete it. Get a from the show and relatedEps.
`and (cur_ep.season != 0 or sickbeard.SUBTITLES_INCLUDE_SPECIALS):`
@@ -18,7 +18,7 @@ RSpec.describe RepositoryListValue, type: :model do it { should accept_nested_attributes_for(:repository_cell) } end - describe '#data' do + describe '#formatted' do let!(:repository) { create :repository } let!(:repository_column) { create :repository_column, name: 'My column' } let!(:repository_column) do
[class_name,to,create,have_db_column,accept_nested_attributes_for,to_not,describe,repository_list_item,save,eq,let!,it,require,belong_to,should]
Describe the list of items cell_attributes - cell attributes.
Block has too many lines. [47/25]
@@ -315,6 +315,12 @@ static GenericAgentConfig *CheckOpts(int argc, char **argv) char **argv_new = TranslateOldBootstrapOptionsConcatenated(argc_new, argv_tmp); FreeStringArray(argc_new, argv_tmp); + /* true if cf-agent is executed by cf-serverd in response to a cf-runagent invocation. + * In that case, prohibit file-input, no-lock and dry-runs to prevent remote execution of + * arbitrary policy and DoS attacks. + */ + bool cfruncommand = false; + while ((c = getopt_long(argc_new, argv_new, "dvnKIf:D:N:VxMB:b:hlC::E", OPTIONS, NULL)) != EOF) { switch ((char) c)
[No CFG could be retrieved]
Parses command line options and returns the configuration object. Destroy the list of bundles and remove the list of all the bundles.
I don't see dry-runs being blocked by the code changes below. What am I missing ?
@@ -25,8 +25,6 @@ class ApplicationSummaryPaged(Paged): def __init__(self, *args, **kwargs): super(ApplicationSummaryPaged, self).__init__(*args, **kwargs) - - class PoolUsageMetricsPaged(Paged): """ A paging container for iterating over a list of :class:`PoolUsageMetrics <azure.batch.models.PoolUsageMetrics>` object
[ComputeNodePaged->[__init__->[super]],CertificatePaged->[__init__->[super]],ApplicationSummaryPaged->[__init__->[super]],CloudJobSchedulePaged->[__init__->[super]],CloudPoolPaged->[__init__->[super]],CloudTaskPaged->[__init__->[super]],PoolUsageMetricsPaged->[__init__->[super]],JobPreparationAndReleaseTaskExecutionInformationPaged->[__init__->[super]],ImageInformationPaged->[__init__->[super]],NodeFilePaged->[__init__->[super]],PoolNodeCountsPaged->[__init__->[super]],CloudJobPaged->[__init__->[super]]]
Initialize ApplicationSummaryPaged with a sequence of unique IDs.
Don't remove the lines.
@@ -116,7 +116,7 @@ class SelfAttention(TransformerModule, FromParams): mask_reshp = (batch_size, 1, 1, k_length) attention_mask = (attention_mask == 0).view(mask_reshp).expand_as( attention_scores - ) * -10e5 + ) * nn_util.min_value_of_dtype(attention_scores.dtype) attention_scores = attention_scores + attention_mask # Normalize the attention scores to probabilities.
[SelfAttention->[forward->[_transpose_for_scores]]]
Forward computation of the attention scores. Outputs a list of tuples of the context_layer and the attention_probs.
Is that really what we use? Let's use whatever huggingface uses. `min_value` seems risky, because of numeric weirdnesses at the edge of floating point precision.
@@ -262,8 +262,11 @@ class TableServiceClient(AsyncStorageAccountHostsMixin, TableServiceClientBase): page_iterator_class=TablePropertiesPaged ) - def get_table_client(self, table, **kwargs): - # type: (Union[TableProperties, str], Optional[Any]) -> TableClient + def get_table_client( + self, table_name, # type: Union[TableProperties, str] + **kwargs # type: Optional[Any] + ): + # type: (...) -> TableClient """Get a client to interact with the specified table. The table need not already exist.
[TableServiceClient->[get_service_properties->[get_properties,pop,process_table_error,service_properties_deserialize],delete_table->[_validate_table_name,delete],get_table_client->[AsyncPipeline,AsyncTransportWrapper,TableClient],get_service_stats->[service_stats_deserialize,get_statistics,pop,process_table_error],set_service_properties->[set_properties,TableServiceProperties,process_table_error],create_table->[get_table_client,create,_validate_table_name,TableProperties],list_tables->[partial,pop,AsyncItemPaged,isinstance,join,QueryOptions],__init__->[ExponentialRetry,AzureTable,pop,get,super],query_tables->[_parameter_filter_substitution,partial,pop,AsyncItemPaged,isinstance,join,QueryOptions]]]
Queries tables under the given account. Creates a new table client for the given .
it looks like this only takes a `table_name` now. does the type hint need to be updated?
@@ -119,8 +119,9 @@ public class Fingerprint implements ModelObject, Saveable { * Gets the {@link Job} that this pointer points to, * or null if such a job no longer exists. */ - public AbstractProject getJob() { - return Jenkins.getInstance().getItemByFullName(name,AbstractProject.class); + @WithBridgeMethods(value=AbstractProject.class, castRequired=true) + public Job<?,?> getJob() { + return Jenkins.getInstance().getItemByFullName(name, Job.class); } /**
[Fingerprint->[getActions->[getFacets],RangeSet->[retainAll->[addAll,intersect,add,equals],includes->[includes],listNumbersReverse->[iterator->[expand->[iterator]]],addAll->[add],equals->[equals],fromString->[contains,Range,add,RangeSet],listNumbers->[iterator->[expand->[iterator]]],isSmallerThan->[isEmpty,isSmallerThan],hashCode->[hashCode],add->[expandRight,includes,addAll,isIndependent,expandLeft,combine,Range,add,isBiggerThan],checkCollapse->[Range,isAdjacentTo],removeAll->[contains,addAll,isDisjoint,Range,add],toString->[toString],isEmpty->[isEmpty],ConverterImpl->[unmarshal->[unmarshal,fromString,RangeSet],serialize->[toString,isSingle]]],messageOfParseException->[messageOfParseException],getSortedFacets->[compare->[getTimestamp],getFacets],rename->[save,setName,remove,equals],BuildPtr->[is->[getNumber],isAlive->[getRun],getRun->[getJob]],load->[getConfigFile,getFingerprintFile,load],getFacet->[getFacets],getRangeSet->[getRangeSet,RangeSet],save->[contains,serialize,save,toString,isEmpty],getFacets->[size->[size],contains->[contains],remove->[remove],add->[add],iterator->[expand->[],iterator]],_getUsages->[RangeItem,add],addWithoutSaving->[add,RangeSet],trim->[retainAll,getNumber,RangeSet,Range,add,removeAll,isEmpty],isAlive->[isSmallerThan,getNumber,isAlive],getJobs->[addAll],Range->[expandRight->[Range],intersect->[Range,isDisjoint],expandLeft->[Range],combine->[isIndependent,Range]],add->[add,getNumber],toString->[getHashString,toString],getName,BuildPtr,ConverterImpl]]
Get the job with the given name.
Or would it be safer to introduce a new method (and update the usages in `getRun` and `index.jelly`)?
@@ -184,11 +184,11 @@ public class SpringRegistry extends AbstractRegistry { // FBE is a result of a broken config, propagate it (see MULE-3297 for more details) String message = String.format("Failed to lookup beans of type %s from the Spring registry", type); - throw new MuleRuntimeException(MessageFactory.createStaticMessage(message), fbex); + throw new MuleRuntimeException(createStaticMessage(message), fbex); } catch (Exception e) { - logger.debug(e); + logger.debug(e.getMessage(), e); return Collections.emptyMap(); } }
[SpringRegistry->[internalLookupByTypeWithoutAncestors->[MuleRuntimeException,getBeansOfType,debug,emptyMap,createStaticMessage,format],doDispose->[set,get,isActive,close],lookupLocalObjects->[values],lookupObject->[debug,equals,createStaticMessage,getBean,fillInStackTrace,warn,isBlank],unregisterObject->[UnsupportedOperationException],createLifecycleManager->[SpringRegistryLifecycleManager,getRegistryId],internalLookupByType->[MuleRuntimeException,debug,beansOfTypeIncludingAncestors,emptyMap,createStaticMessage,format],registerObjects->[UnsupportedOperationException],lookupByType->[internalLookupByType],doInitialise->[refresh,set],lookupObjects->[values],lookupObjectsForLifecycle->[values],registerObject->[UnsupportedOperationException],setParent,AtomicBoolean]]
Internal method to lookup by type.
"he previous object will be overwritten" -> "The previous object will be overwritten"
@@ -181,7 +181,7 @@ class ChatChannelsController < ApplicationController def create_channel chat_channel_params = params[:chat_channel] - chat_channel_name = chat_channel_params[:channel_name].split(" ").join("-") + chat_channel_name = chat_channel_params[:channel_name].split.join("-") chat_channel = ChatChannel.new( channel_type: "invite_only", channel_name: chat_channel_params[:channel_name],
[ChatChannelsController->[send_chat_action_message->[create],moderate->[update],update->[create],create->[create],update_channel->[create,update],open->[update]]]
method called from the bot when it wants to create a new nack channel.
you'll notice a bunch of these throught the PR. `split(" ")` and `.split` are the same thing. Splitting over space char is the default behavior
@@ -52,7 +52,7 @@ const ORIGINAL_VALUE_PROPERTY = 'amp-original-value'; * @param {*} val * @return {string} */ -function encodeValue(val) { +export function encodeValue(val) { if (val == null) { return ''; }
[GlobalVariableSource->[getVairiantsValue_->[getter,user,variantForOrNull],setTimingResolver_->[getTimingDataAsync,getTimingDataSync],constructor->[accessServiceForDocOrNull],getShareTrackingValue_->[shareTrackingForOrNull,user,getter],getAccessValue_->[getter,user],getQueryParamData_->[user,search,parseQueryString,parseUrl],getDocInfoValue_->[getter,documentInfoForDoc],getStoryValue_->[getter,user,storyVariableServiceForOrNull],initialize->[create,userNotificationManagerForDoc,outgoingFragment,language,dev,sourceUrl,getTotalEngagedTime,browserLanguage,join,charset,hostname,getSize,pageViewId,getAccessReaderId,getElementById,videoManagerForDoc,host,characterSet,activityForDoc,viewportForDoc,canonicalUrl,getAuthdataField,extractClientIdFromGaCookie,user,pathname,getHostname,getTrackImpressionPromise,getScrollLeft,getNavigationData,getScrollTop,userLanguage,now,getScrollHeight,pageIndex,userAgent,incomingFragment,push,resolve,keys,pageId,getSourceUrl,get,random,removeFragment,cidForDoc,getScrollWidth,getTimingDataSync,parseUrl,viewerForDoc,getTimingDataAsync]],UrlReplacements->[isAllowedOrigin_->[getAttribute,hasAttribute,origin,documentInfoForDoc,sourceUrl,parseUrl,length,canonicalUrl],getWhitelistForElement_->[getAttribute,user,trim,hasOwnProperty],expandInputValue_->[getAttribute,then,resolve,value,dev,tagName],expand_->[then,split,resolve,user,encodeValue,apply,replace,async,rethrowAsync,catch,sync],collectVars->[create],maybeExpandLink->[getAttribute,addParamsToUrl,tagName,dev,user,href,parseQueryString,parseUrl],ensureProtocolMatches_->[user,isProtocolValid,parseUrl]],encodeURIComponent,installServiceInEmbedScope,registerServiceBuilderForDoc,replace]
Imports a variable from the top level AMP window. A utility function for setting timing data that supports sync and async.
is this used outside?
@@ -366,6 +366,7 @@ class ShareClient(StorageAccountHostsMixin): :dedent: 12 :caption: Deletes the share and any snapshots. """ + access_conditions = get_access_conditions(kwargs.pop('lease', None)) timeout = kwargs.pop('timeout', None) delete_include = None if delete_snapshots:
[ShareClient->[delete_directory->[delete_directory,get_directory_client],create_directory->[create_directory,get_directory_client],create_permission_for_share->[_create_permission_for_share_options],list_directories_and_files->[list_directories_and_files,get_directory_client],create_snapshot->[create_snapshot]]]
Marks the specified managed managed managed managed managed managed managed share for deletion.
missing `lease` doc for all these apis
@@ -140,7 +140,7 @@ class PythonBinaryCreate(Task): if is_python_target(tgt): constraint_tgts.append(tgt) - # Add target's interpreter compatibility constraints to pex info. + # Add target-level and possibly global interpreter compatibility constraints to pex info. pex_builder.add_interpreter_constraints_from(constraint_tgts) # Dump everything into the builder's chroot.
[PythonBinaryCreate->[implementation_version->[super],prepare->[optional_product,optional_data,require_data],_python_native_code_settings->[scoped_instance],is_binary->[isinstance],subsystem_dependencies->[super,scoped],__init__->[get_options,super],_create_binary->[targets,create,add_sources_from,has_resources,add_requirement_libs_from,join,info,check_build_for_current_platform_only,get_data,temporary_dir,is_python_target,copy,get_as_dict,debug,add_interpreter_constraints_from,append,update,make_build_properties,format,closure,set_shebang,build,has_python_requirements,has_python_sources,PEXBuilder],execute->[relpath,targets,atomic_copy,debug,basename,get_buildroot,info,get,invalidated,format,join,safe_mkdir_for,add,_create_binary,TaskError]]]
Create a. pex file for the specified binary target. Get the path to the. pex file.
Where are the global constraints added?
@@ -194,9 +194,8 @@ module.exports = class extends BaseBlueprintGenerator { return this._configuring(); } - _loadPlatformConfig(config = _.defaults({}, this.jhipsterConfig, defaultConfig), dest = this) { - super.loadPlatformConfig(config, dest); - dest.cicdIntegrationsSnyk = config.cicdIntegrations || []; + _loadCiCdConfig(config = _.defaults({}, this.jhipsterConfig, defaultConfig), dest = this) { + dest.cicdIntegrations = dest.cicdIntegrations || config.cicdIntegrations || []; dest.cicdIntegrationsSnyk = dest.cicdIntegrations.includes('snyk'); dest.cicdIntegrationsSonar = dest.cicdIntegrations.includes('sonar'); dest.cicdIntegrationsHeroku = dest.cicdIntegrations.includes('heroku');
[No CFG could be retrieved]
Public API method used by the getter and also by Blueprints Public API method used by the getter and by the blueprints.
Better move to after priorities to don't mix priority and api.
@@ -96,7 +96,7 @@ class CarController(): # 20 Hz LFA MFA message - if frame % 5 == 0 and self.car_fingerprint in [CAR.SONATA, CAR.PALISADE]: + if frame % 5 == 0 and self.car_fingerprint in [CAR.SONATA, CAR.PALISADE, CAR.SANTA_FE]: can_sends.append(create_lfa_mfa(self.packer, frame, enabled)) return can_sends
[CarController->[__init__->[CANPacker],update->[create_clu11,process_hud_alert,create_lkas11,append,apply_std_steer_torque_limits,create_lfa_mfa,abs]]]
Update the state of the car. check if a message is a lease and create the necessary MFA messages.
Would this be a problem for the 2019 Santa FE that doesn't have LFA? Otherwise we need to split them into two different cars. Maybe you can find someone on discord with a 2019 one and have them try this PR.
@@ -26,7 +26,12 @@ module Users else result = personal_key_form.submit - analytics.track_event(Analytics::PERSONAL_KEY_REACTIVATION_SUBMITTED, result.to_h) + analytics_result = FormResponse.new( + success: result.success?, + errors: result.errors, + extra: result.extra.except(:decrypted_pii), + ) + analytics.track_event(Analytics::PERSONAL_KEY_REACTIVATION_SUBMITTED, analytics_result.to_h) if result.success? handle_success(result) else
[VerifyPersonalKeyController->[personal_key_form->[new],new->[new]]]
Creates a new node in the system.
This would always be empty (as of today), so we could just omit `extra` altogether.
@@ -2,6 +2,9 @@ import os import random from django.conf import settings +from django.contrib.admin.views.decorators import ( + staff_member_required as _staff_member_required, +) from django.core.files import File from ..checkout import AddressType
[change_user_default_address->[set_user_default_billing_address,set_user_default_shipping_address]]
Create a user object from a list of user objects. Protected get_user_first_name get_user_first_name get_user.
What about naming it `django_staff_member_required` so we explicitly know it is not internal, but django's?
@@ -250,6 +250,18 @@ public final class MetaStoreImpl implements MetaStore { return functionRegistry.listAggregateFunctions(); } + private Stream<SourceInfo> streamSources(final Set<String> sourceNames) { + return sourceNames.stream() + .map(sourceName -> { + final SourceInfo sourceInfo = dataSources.get(sourceName); + if (sourceInfo == null) { + throw new KsqlException("Unknown source: " + sourceName); + } + + return sourceInfo; + }); + } + private static final class SourceInfo { private final StructuredDataSource source;
[MetaStoreImpl->[listFunctions->[listFunctions],copy->[MetaStoreImpl],listAggregateFunctions->[listAggregateFunctions],getAggregate->[getAggregate],getUdfFactory->[getUdfFactory],getAggregateFactory->[getAggregateFactory],SourceInfo->[copy->[SourceInfo],copy],addFunction->[addFunction],isAggregate->[isAggregate],addFunctionFactory->[addFunctionFactory],addAggregateFunctionFactory->[addAggregateFunctionFactory]]]
List all aggregate functions that are registered in the registry.
I'm always hesitant to throw exceptions within streams because they're lazily evaluated and that can cause somewhat unexpected behavior (and prevents things from getting cleaned up). Instead, can we just validate the entire set of source names up front? That also has the added benefit of not needing to validate the source names four times (one for each call to streamSources) and having the error message contain all invalid sources if multiple are invalid: `Set<String> unknownKeys = Sets.difference(sourceNames, dataSources.keySet());`
@@ -15,12 +15,18 @@ package statistics import ( "math/rand" + "testing" + "time" . "github.com/pingcap/check" "github.com/pingcap/kvproto/pkg/metapb" "github.com/tikv/pd/server/core" ) +func Test(t *testing.T) { + TestingT(t) +} + var _ = Suite(&testHotPeerCache{}) type testHotPeerCache struct{}
[TestStoreTimeUnsync->[SetWrittenBytes,RegionStats,NewRegionInfo,Assert,SetReportInterval],GetPeers,SetReportInterval,GetID,getOldHotPeerStat,CheckRegionFlow,SetWrittenBytes,SetReadBytes,GetMeta,NewRegionInfo,Assert,Intn,GetLeader,Update]
TestStoreTimeUnsync tests that the time of a key is not synced with the cache.
I've added in another PR.
@@ -64,7 +64,7 @@ class OntonotesSentence: word_senses: List[Optional[float]], speakers: List[Optional[str]], named_entities: List[str], - srl_frames: Dict[str, List[str]], + srl_frames: List[Tuple[str, List[str]]], coref_spans: Set[TypedSpan]) -> None: self.document_id = document_id
[Ontonotes->[_conll_rows_to_sentence->[OntonotesSentence]]]
Initialize a new object from a sequence of missing values.
Just curious, what's the motivation for switching from a Dict to a list of tuples?
@@ -1,7 +1,9 @@ import os import platform +import pytest import textwrap import unittest + from textwrap import dedent from nose.plugins.attrib import attr
[CMakeFlagsTest->[test_transitive_targets_flags->[_get_line],test_targets_own_flags->[_get_line],test_standard_20_as_cxx_flag->[conan_set_std_branch],test_flags->[_get_line],test_transitive_flags->[_get_line],test_targets_flags->[_get_line]]]
Package info for the given node. CmakeFlagsTest - A test to check if a C - option is present in a.
pytest is not a builtin python module, imports shouldn't be here (but not a big problem)
@@ -125,10 +125,13 @@ func Run(name, version string, bt beat.Creator) error { } // NewBeat creates a new beat instance -func NewBeat(name, v string) (*Beat, error) { +func NewBeat(name, indexPrefix, v string) (*Beat, error) { if v == "" { v = version.GetDefaultVersion() } + if indexPrefix == "" { + indexPrefix = name + } hostname, err := os.Hostname() if err != nil {
[launch->[createBeater,Init],TestConfig->[createBeater,Init],Setup->[createBeater,Init],createBeater->[BeatConfig],configure->[Init,BeatConfig]]
NewBeat creates a new instance of the beat. Init returns a reference to the beat object.
I suggest if `indexPrefix` is empty, we set it to `name`. Like this the param can be set to `""`.
@@ -11,8 +11,8 @@ <title>Laravel Application</title> <!-- Bootstrap CSS --> - <link href="/css/app.css" rel="stylesheet"> - <link href="/css/vendor/font-awesome.css" rel="stylesheet"> + <link href="{!! url('css/app.css') !!}" rel="stylesheet"> + <link href="{!! url('css/vendor/font-awesome.css') !!}" rel="stylesheet"> <!-- Web Fonts --> <link href='http://fonts.googleapis.com/css?family=Lato:300,400,700,300italic,400italic,700italic' rel='stylesheet' type='text/css'>
[No CFG could be retrieved]
Displays a single - element sequence in a nice way. Displays a navbar with a menu for the n - node .
Probably better to use the asset helper here?
@@ -22,6 +22,11 @@ namespace Kratos { /////////////////////////////////////////////////////////////////////////////////////////////////// // Public Operations +template <int TDim, int TNumNodes> +void TransonicPerturbationPotentialFlowElement<TDim, TNumNodes>::Initialize(const ProcessInfo& rCurrentProcessInfo) +{ + FindUpwindElement(rCurrentProcessInfo); +} template <int TDim, int TNumNodes> Element::Pointer TransonicPerturbationPotentialFlowElement<TDim, TNumNodes>::Create(
[No CFG could be retrieved]
Create a single object from a list of objects. - - - - - - - - - - - - - - - - - -.
We usually implement Initialize after clone
@@ -458,13 +458,13 @@ void io992_device::device_start() READ8_MEMBER(io992_device::cruread) { - int address = offset << 4; + int address = offset << 1; uint8_t value = 0x7f; // All Hexbus lines high double inp = 0; int i; uint8_t bit = 1; - switch (address) + switch (address & 0xf800) { case 0xe000: // CRU E000-E7fE: Keyboard
[No CFG could be retrieved]
region Device start Reads the next 16 - bit value from the keyboard and the cassette.
We're using the software address, so the "<<1" is already done in the CPU.
@@ -242,6 +242,12 @@ func resourceAwsRamResourceShareGetInvitation(conn *ram.RAM, resourceShareARN, s return nil, fmt.Errorf("Error reading RAM resource share invitation %s: %s", resourceShareARN, err) } + if *invitation.ReceiverAccountId != client.accountid { + // Report an error, this won't work on the long term since once the invitation has disappeared + // the receiver_account_id is populated with the account_id (and same during import) + return nil, fmt.Errorf("Unexpected behavior while reading RAM resource share invitation %s: receiver_account_id is %s while current account_id is %s", resourceShareARN, *invitation.ReceiverAccountId, client.accountid) + } + return invitation, nil }
[LastIndex,UniqueId,StringSlice,AcceptResourceShareInvitation,Set,Errorf,SetId,GetResourceShareInvitations,Timeout,Id,Int64,Get,Printf,DefaultTimeout,StringValue,GetResourceShareInvitationsPages,DisassociateResourceShare,WaitForState,String,ListResourcesPages,Replace]
This function returns a resource. StateRefreshFunc that is used to watch a resource share invitation.
I see no reason why this could happen, but I prefer to have the check in place to be notified if this happens under a corner-case condition, since this would cause the bugs fixed by this change to reappear under this corner-case condition.
@@ -56,7 +56,7 @@ class AVSClient(SDKClient): super(AVSClient, self).__init__(self.config.credentials, self.config) client_models = {k: v for k, v in models.__dict__.items() if isinstance(v, type)} - self.api_version = '2020-03-20' + self.api_version = '2020-07-17-preview' self._serialize = Serializer(client_models) self._deserialize = Deserializer(client_models)
[AVSClient->[__init__->[PrivateCloudsOperations,AVSClientConfiguration,LocationsOperations,Operations,HcxEnterpriseSitesOperations,super,isinstance,Serializer,items,Deserializer,ClustersOperations,AuthorizationsOperations]]]
Initialize an AVS client with the specified credentials subscription_id and base_url.
It would still be using the 2020-03-20 API.
@@ -1220,7 +1220,8 @@ class Setup extends DolibarrApi } /** - * Get the list of shipping methods. + * Get the list of shipping methods. This operation is deprecated, use + * /setup/dictionary/shipment_methods instead. * * @param int $limit Number of items per page * @param int $page Page number {@min 0}
[Setup->[getModules->[_cleanObjectDatas],getCompany->[_cleanObjectDatas]]]
Get list of shipping methods. Get c_shipment_mode list.
This url suggested seems the same than this one. ?
@@ -760,6 +760,7 @@ ActiveRecord::Schema.define(version: 20180305102456) do t.integer "visibility_level", default: 0, null: false t.text "css" t.datetime "archived_at" + t.jsonb "footer_translations" t.index ["archived_at"], name: "index_gplan_plans_on_archived_at" t.index ["plan_type_id"], name: "index_gplan_plans_on_plan_type_id" t.index ["site_id"], name: "index_gplan_plans_on_site_id"
[jsonb,bigint,decimal,string,text,date,hstore,add_foreign_key,datetime,integer,enable_extension,create_table,float,boolean,index,define,inet]
Table for creating Gplan objects. create a table in the issues table.
Prefer single-quoted strings when you don't need string interpolation or special symbols.
@@ -158,6 +158,7 @@ public class UtilTest { " \"#%/:;<>?", "%20%22%23%25%2F%3A%3B%3C%3E%3F", "[\\]^`{|}~", "%5B%5C%5D%5E%60%7B%7C%7D%7E", "d\u00E9velopp\u00E9s", "d%C3%A9velopp%C3%A9s", + "Foo \uD800\uDF98 Foo", "Foo%20%F0%90%8E%98%20Foo" }; for (int i = 0; i < data.length; i += 2) { assertEquals("test " + i, data[i + 1], Util.rawEncode(data[i]));
[UtilTest->[resolveSymlinkToFile->[resolveSymlinkToFile],loadProperties->[loadProperties]]]
test raw encode.
U+10398 UGARITIC LETTER THANNA? Heavy.
@@ -52,10 +52,12 @@ public class TaskSpec { // The task's resource demands. public final Map<String, Double> resources; - // Function descriptor is a list of strings that can uniquely identify a function. - // It will be sent to worker and used to load the target callable function. + // Descriptor of the target function. This field is only valid for Java tasks. public final FunctionDescriptor functionDescriptor; + // Descriptor of the target Python function. This field is only valid for Python tasks. + public final PyFunctionDescriptor pyFunctionDescriptor; + private List<UniqueId> executionDependencies; public boolean isActorTask() {
[TaskSpec->[isActorCreationTask->[isNil],toString->[toString,getResourcesStringFromMap],isActorTask->[isNil]]]
is actor task.
Consider merging these two fields to avoid nulls.
@@ -2254,6 +2254,15 @@ class Archiver: '\*/.bundler/gems' to get the same effect. See ``borg help patterns`` for more information. + In addition to using ``--exclude`` patterns, it is possible to use + ``--exclude-if-present`` to specify the name of a filesystem object (e.g. a file + or folder name) which, when contained within another folder, will prevent the + containing folder from being backed up. By default, the containing folder and + all of its contents will be omitted from the backup. If, however, you wish to + only include the objects specified by ``--exclude-if-present`` in your backup, + and not include any other contents of the containing folder, this can be enabled + through using the ``--keep-exclude-tags`` option. + Item flags ++++++++++
[main->[get_args,run,Archiver],with_repository->[decorator->[wrapper->[argument]]],Archiver->[do_prune->[print_error,write],do_mount->[print_error],do_check->[print_error],do_extract->[build_filter,print_warning,build_matcher],_list_archive->[_list_inner->[write],_list_inner,build_matcher],do_debug_get_obj->[write],run->[_setup_topic_debugging,prerun_checks,_setup_implied_logging],do_debug_dump_archive->[output->[write,do_indent],output],do_recreate->[print_error,write,build_matcher],_list_repository->[write],do_key_export->[print_error],do_upgrade->[write],do_benchmark_crud->[measurement_run,test_files],_info_archives->[format_cmdline],_process->[print_file_status,_process,print_warning],do_debug_dump_archive_items->[write],build_parser->[process_epilog],do_key_import->[print_error],do_diff->[contents_changed->[sum_chunk_size,fetch_and_compare_chunks],compare_archives->[update_hardlink_masters->[is_hardlink_master],compare_or_defer->[update_hardlink_masters,hardlink_master_seen,compare_items],hardlink_master_seen,print_output,update_hardlink_masters,compare_items,compare_or_defer],fetch_and_compare_chunks->[compare_chunk_contents],compare_content->[sum_chunk_size,contents_changed],compare_mode->[get_mode],compare_owner->[get_owner],compare_items->[compare_directory,get_mode,has_hardlink_master,compare_link,compare_content,compare_mode,compare_owner],build_matcher,compare_archives,print_warning],do_debug_dump_repo_objs->[write],parse_args->[parse_args,build_parser,preprocess_args],do_list->[write->[write]],do_change_passphrase_deprecated->[do_change_passphrase],do_create->[create_inner->[print_file_status,print_warning],create_inner],with_repository],main]
Build an argument parser for the borg - cli command. command line interface to run borg command command to serve a specific key file or directory This function is called when a key is not encrypted with a passphrase.
well, this is explaining what it does, but not much the motivation to do so. - for tag files, that is just not to lose the tags (the information about not wanting to back up this specific dir) in case you have to recover from a backup. - for tag directories with actual content, there is the .git/ usecase: you do not want to back up all your workdir checkouts, but you do want to backup your repos.
@@ -23,13 +23,13 @@ import org.jgroups.util.Util; * @since 13.0 **/ public class NamedSocketFactory implements SocketFactory { - private final javax.net.SocketFactory socketFactory; - private final ServerSocketFactory serverSocketFactory; + private final Supplier<javax.net.SocketFactory> socketFactory; + private final Supplier<ServerSocketFactory> serverSocketFactory; private String name; private BiConsumer<String, Socket> socketConfigurator = (c, s) -> {}; private BiConsumer<String, ServerSocket> serverSocketConfigurator = (c, s) -> {}; - public NamedSocketFactory(javax.net.SocketFactory socketFactory, ServerSocketFactory serverSocketFactory) { + public NamedSocketFactory(Supplier<javax.net.SocketFactory> socketFactory, Supplier<ServerSocketFactory> serverSocketFactory) { this.socketFactory = socketFactory; this.serverSocketFactory = serverSocketFactory; }
[NamedSocketFactory->[createServerSocket->[createServerSocket,configureSocket],createSocket->[createSocket,configureSocket],close->[close]]]
Creates a named socket factory which allows to configure the sockets using a supplied name. Configure a socket with a server socket.
This constructor is never used.
@@ -283,8 +283,7 @@ export function selectGptExperiment(data) { export function writeAdScript(global, data, gptFilename) { const url = `https://www.googletagservices.com/tag/js/${gptFilename || 'gpt.js'}`; - if (gptFilename || data.useSameDomainRenderingUntilDeprecated != undefined - || data.multiSize) { + if (gptFilename || data.multiSize) { doubleClickWithGpt(global, data, GladeExperiment.GLADE_OPT_OUT, url); } else { const experimentFraction = 0.1;
[No CFG could be retrieved]
Writes an ad script to the specified file.
We still get quite a few ad requests from the GLADE_OPT_OUT branch, unfortunately we cannot tell for what reason very easily due to all of these being grouped together as opt out. How do we know that this is safe to just turn off? 1) What if we split these out (or at least the explicit opt-out), so that we can verify the actual amount of traffic from it? and/or 2) What if rather than flipping it off 100% which is slow to revert in an emergency we setup an experiment which ignores this some fraction of the time. We can gradually increase it as a way to urge pubs to change without breaking all of their ads at once. This also gives us a faster mechanism to change things if anything goes wrong.
@@ -278,6 +278,16 @@ describes.sandboxed('UrlReplacements', {}, () => { }); }); + it('should replace CLIENT_ID with opt_cookieName', () => { + setCookie(window, 'url-abc', 'cid-for-abc'); + // Make sure cookie does not exist + setCookie(window, 'url-xyz', ''); + return expandAsync('?a=CLIENT_ID(abc,,url-abc)&b=CLIENT_ID(xyz,,url-xyz)', + /*opt_bindings*/undefined, {withCid: true}).then(res => { + expect(res).to.match(/^\?a=cid-for-abc\&b=amp-([a-zA-Z0-9_-]+){10,}/); + }); + }); + it('should replace CLIENT_ID synchronously when available', () => { return getReplacements({withCid: true}).then(urlReplacements => { setCookie(window, 'url-abc', 'cid-for-abc');
[No CFG could be retrieved]
Generates a sequence of tokens that can be used to generate a unique identifier. This function checks that the result of the feature is exactly the same as the result of the.
What does scope do when there's a cookie name provided?
@@ -94,7 +94,9 @@ def einsum(axes, *inputs): """ A generalized contraction between tensors of arbitrary dimension. - Like numpy.einsum. + Like numpy.einsum, but does not support: + -- ellipses (subscripts like 'ij...,jk...->ik...') + -- subscripts that reduce to scalars """ match = re.match('([a-z,]+)->([a-z]+)', axes)
[lbeta->[nonempty_lbeta->[reduce_sum,lgamma],empty_lbeta->[squeeze,assert_rank_at_least,control_dependencies],convert_to_tensor,size,cond,equal,with_dependencies,get_shape,empty_lbeta,nonempty_lbeta,name_scope,assert_rank_at_least],einsum->[reduce_sum,group,list,reshape,find,append,get_shape,len,sorted,set,transpose,join,shapes,zip,enumerate,isinstance,match]]
A generalized contraction between tensors of arbitrary dimension. Compute the total along a given axis.
Can you turn those apostrophes into backticks? Then it'll render nicely as code. Also, please put `numpy.einsum` into backticks. All docstrings get rendered as markdown.
@@ -60,15 +60,15 @@ func createTargetMap(targets []resource.URN) map[resource.URN]bool { // checkTargets validates that all the targets passed in refer to existing resources. Diagnostics // are generated for any target that cannot be found. The target must either have existed in the stack // prior to running the operation, or it must be the urn for a resource that was created. -func (pe *planExecutor) checkTargets(targets []resource.URN) result.Result { +func (pe *planExecutor) checkTargets(targets []resource.URN, op StepOp) result.Result { if len(targets) == 0 { return nil } olds := pe.plan.olds var news map[resource.URN]bool - if pe.stepGen != nil && pe.stepGen.creates != nil { - news = pe.stepGen.creates + if pe.stepGen != nil { + news = pe.stepGen.urns } hasUnknownTarget := false
[retirePendingDeletes->[reportExecResult],Execute->[reportExecResult,checkTargets,reportError],refresh->[reportExecResult,checkTargets]]
checkTargets checks if the given targets are not in the stack and if so it will return.
This fixes a bug that caused a "resource not found" error if a specified target was replaced rather than created or updated.
@@ -831,7 +831,7 @@ void RAND_seed(const void *buf, int num) { const RAND_METHOD *meth = RAND_get_rand_method(); - if (meth->seed != NULL) + if (meth != NULL && meth->seed != NULL) meth->seed(buf, num); }
[RAND_bytes->[rand_bytes_ex],RAND_priv_bytes->[rand_priv_bytes_ex],RAND_set_rand_engine->[RAND_set_rand_method],rand_pool_bytes_needed->[rand_pool_entropy_needed]]
RAND_seed - Seed random number generator.
I'm feeling a bit uncomfortable about this location here, because there is no way to indicate an error to the caller. OTOH: if the CSPRNG is already initialized, the failure will not be disastrous, and if it is not, the generate call will fail later on.
@@ -40,6 +40,10 @@ const ( // This file will be ignored when copying from the template cache to // a project directory. pulumiTemplateManifestFile = ".pulumi.template.yaml" + + // pulumiLocalTemplatePathEnvVar is a path to the folder where template are stored. + // It is used in sandboxed environments where the classic template folder may not be writable. + pulumiLocalTemplatePathEnvVar = "PULUMI_TEMPLATE_PATH" ) // Template represents a project template.
[CopyTemplateFiles->[ReadFile,IsDir,Mkdir,Base,IsExist],CopyTemplateFilesDryRun->[Base,Stat,IsDir],RemoveAll,FileMode,Wrap,IsNotExist,Copy,IgnoreClose,Stat,ReadFile,New,ReadDir,Errorf,Assert,Wrapf,Join,Current,Next,Name,Base,Require,MkdirAll,Write,IsDir,NewReader,Sprintf,Unmarshal,IsPackageName,Replace,OpenFile,IsExist]
LoadLocalTemplate loads a template from a local directory. returns the template with the name of the template in the templateDir.
Nit: template => templates (e.g. "... where template**s** are stored.")
@@ -268,7 +268,7 @@ class GradientChecker(unittest.TestCase): :param input_vars: numpy value of input variable. The following computation will use these variables. :param inputs_to_check: inputs var names that should check gradient. - :param output_name: output name that used to + :param output_name: the final output variable name. :param max_relative_error: The relative tolerance parameter. :param no_grad_set: used when create backward ops :param only_cpu: only compute and check gradient on cpu kernel.
[get_numeric_gradient->[get_output,product,restore_inputs],GradientChecker->[compare_grad->[empty_var_name,__get_gradient],__assert_is_close->[err_msg],check_grad->[__assert_is_close,grad_var_name,__get_gradient,get_numeric_gradient]]]
This function computes the gradient of a single . Assert that all the gradients are close.
the final output variable name. => the output variable name.
@@ -53,6 +53,14 @@ public class NumberedShardedFileTest { @Mock private PipelineResult pResult = Mockito.mock(PipelineResult.class); private final BackOff backOff = NumberedShardedFile.BACK_OFF_FACTORY.backoff(); + private String filePattern; + + @Before + public void setup() throws IOException { + filePattern = FileSystems.matchSingleFileSpec( + tmpFolder.getRoot().getPath()).resourceId().resolve( + "*", StandardResolveOptions.RESOLVE_FILE).toString(); + } @Test public void testPreconditionFilePathIsNull() {
[NumberedShardedFileTest->[testReadWithRetriesFailsWhenTemplateIncorrect->[newFile,readFilesWithRetries,resolve,expect,write,NumberedShardedFile,expectMessage,compile,containsString,getPath],testPreconditionFilePathIsNull->[expectMessage,containsString,expect,NumberedShardedFile],testReadEmpty->[newFile,readFilesWithRetries,resolve,write,NumberedShardedFile,empty,assertThat,getPath],testReadCustomTemplate->[newFile,readFilesWithRetries,resolve,write,NumberedShardedFile,compile,assertThat,getPath,containsInAnyOrder],testReadWithRetriesFailsWhenRedundantFileLoaded->[newFile,readFilesWithRetries,resolve,expect,NumberedShardedFile,expectMessage,containsString,getPath],testReadMultipleShards->[newFile,readFilesWithRetries,resolve,write,NumberedShardedFile,assertThat,getPath,containsInAnyOrder],testReadWithRetriesFailsSinceFilesystemError->[newFile,readFilesWithRetries,readLines,resolve,any,write,NumberedShardedFile,expect,expectMessage,getPath,containsString,spy,anyCollection],testReadWithRetriesFailsWhenOutputDirEmpty->[readFilesWithRetries,resolve,expect,NumberedShardedFile,expectMessage,containsString,getPath],testPreconditionFilePathIsEmpty->[expectMessage,containsString,expect,NumberedShardedFile],FastNanoClockAndSleeper,TemporaryFolder,mock,backoff,none]]
Checks whether the file path is null.
`LocalResources` could help here.
@@ -36,9 +36,10 @@ class ContentMapper extends ContainerAware implements ContentMapperInterface * @param $data array The data to be saved * @param $language string Save data for given language * @param $templateKey string name of template + * @param $userId int The id of the user who saves * @return StructureInterface */ - public function save($data, $language, $templateKey) + public function save($data, $language, $templateKey, $userId) { // TODO localize $structure = $this->getStructure($templateKey);
[ContentMapper->[getSession->[getSession],save->[save],getStructure->[getStructure]]]
Save a node in the system. This method will return a structure object with all properties of a node.
@drotter languageCode please, and order $data, $templateKey, $languageCode, $userId
@@ -45,7 +45,10 @@ </div> <%_ } _%> <br/> - <div class="table-responsive" *ngIf="<%=entityInstancePlural %>"> + <div class="alert alert-warning" *ngIf="<%=entityInstancePlural %>?.length === 0"> + No <%=entityInstancePlural %> found + </div> + <div class="table-responsive" *ngIf="<%=entityInstancePlural %>?.length > 0"> <table class="table table-striped"> <thead> <tr<% if (pagination !== 'no') { %> jhiSort [(predicate)]="predicate" [(ascending)]="reverse" [callback]="<%=pagination !== 'infinite-scroll' ? 'transition.bind(this)' : 'reset.bind(this)'%>"<% } %>>
[No CFG could be retrieved]
Renders the JHI - related NGUI elements. Renders a JHI sort field if the relationship is a many - to - many relationship.
Shouldn't it be in json file, when there is i18n ?
@@ -164,7 +164,11 @@ abstract class PartialSegmentGenerateTask<T extends GeneratedPartitionsReport> e final PartitionsSpec partitionsSpec = tuningConfig.getGivenOrDefaultPartitionsSpec(); final long pushTimeout = tuningConfig.getPushTimeout(); - final IndexTaskSegmentAllocator segmentAllocator = createSegmentAllocator(toolbox); + final CachingSegmentAllocator segmentAllocator = createSegmentAllocator(toolbox, taskClient); + final SequenceNameFunction sequenceNameFunction = new NonLinearlyPartitionedSequenceNameFunction( + getId(), + segmentAllocator.getShardSpecs() + ); final Appenderator appenderator = BatchAppenderators.newAppenderator( getId(),
[PartialSegmentGenerateTask->[generateSegments->[createSegmentAllocator]]]
Generate segments for the given input source and temporary directory. returns pushed segments.
Why is this always `NonLinearlyPartitionedSequenceNameFunction` shouldn't we check the partitionsSpec type to determine the `sequenceNameFunction` here? If it is, I think we should , make the constructors package private and expose the function name creation through a factory that accepts a `PartitionsSpec`
@@ -1521,7 +1521,10 @@ namespace Dynamo.Nodes internal override IEnumerable<AssociativeNode> BuildAst(List<AssociativeNode> inputAstNodes) { - var rhs = AstFactory.BuildStringNode(this.Value); + string content = this.Value; + content = content.Replace("\r\n", "\n"); + + var rhs = AstFactory.BuildStringNode(content); var assignment = AstFactory.BuildAssignment(GetAstIdentifierForOutputIndex(0), rhs); return new[] { assignment };
[StringInput->[NodeMigrationData->[Value],BuildAst->[Value],LoadNode->[DeserializeValue,Value],Value],BasicInteractive->[SerializeCore->[SerializeValue,Value,SerializeCore],DeserializeCore->[DeserializeValue,Value,DeserializeCore],LoadNode->[DeserializeValue,Value]],Sublists->[NodeMigrationData->[Value],SerializeCore->[SerializeCore],DeserializeCore->[DeserializeCore],LoadNode->[LoadNode]],DoubleInput->[DeserializeCore->[Value,DeserializeCore],OneNumber->[GetValue->[GetValue],AssociativeNode->[Value]],Sequence->[GetValue->[GetValue],GetValue],SerializeCore->[SerializeCore],CountRange->[Process->[Process]],IDoubleInputToken->[Value],Range->[GetValue->[GetValue],AssociativeNode->[GetRangeExpressionOperator],Process->[Range],GetValue],ApproxRange->[Process->[Process]],LoadNode->[Value]],Apply1->[LoadNode->[Value],RemoveInput->[RemoveInput]],VariableInput->[SerializeCore->[SerializeCore],AddInput->[GetInputNameIndex,GetInputRootName,GetTooltipRootName],DeserializeCore->[DeserializeCore]],AbstractString->[SerializeCore->[SerializeCore],DeserializeCore->[DeserializeCore],PrintExpression->[PrintExpression]],DropDrownBase->[combo_DropDownOpened->[PopulateItems],SaveNode->[ToString],LoadNode->[Value]],DynamoDropDownItem->[CompareTo->[CompareTo]],VariableInputAndOutput->[SerializeCore->[SerializeCore],AddInput->[GetInputNameIndex,GetInputRootName,GetTooltipRootName,GetOutputRootName],DeserializeCore->[DeserializeCore]]]
Build the expression and return an array of expressions.
Hi, do we need this conversion for \r\n to \n. I made it because there are some test cases expecting \n instead of \r\n. Shall I change the test cases instead of changing the implementation?
@@ -153,9 +153,12 @@ public class DeployVMCmd extends BaseAsyncCreateCustomIdCmd implements SecurityG @Parameter(name = ApiConstants.USER_DATA, type = CommandType.STRING, description = "an optional binary data that can be sent to the virtual machine upon a successful deployment. This binary data must be base64 encoded before adding it to the request. Using HTTP GET (via querystring), you can send up to 2KB of data after base64 encoding. Using HTTP POST(via POST body), you can send up to 32K of data after base64 encoding.", length = 32768) private String userData; - @Parameter(name = ApiConstants.SSH_KEYPAIR, type = CommandType.STRING, description = "name of the ssh key pair used to login to the virtual machine") + @Parameter(name = ApiConstants.SSH_KEYPAIR, type = CommandType.LIST, collectionType = CommandType.STRING, description = "name of the ssh key pairs used to login to the virtual machine") private String sshKeyPairName; + @Parameter(name = ApiConstants.SSH_KEYPAIRS, type = CommandType.LIST, collectionType = CommandType.STRING, description = "name of the ssh key pairs used to login to the virtual machine") + private List<String> sshKeyPairNames; + @Parameter(name = ApiConstants.HOST_ID, type = CommandType.UUID, entityType = HostResponse.class, description = "destination Host ID to deploy the VM to - parameter available for root admin only") private Long hostId;
[DeployVMCmd->[execute->[getStartVm,getCommandName],getAccountName->[getAccountName],getDetails->[getBootType],getDomainId->[getDomainId]]]
Creates a virtual machine group and a list of security groups to be deployed to.
This should not be changed
@@ -216,6 +216,8 @@ class WeightNormParamAttr(ParamAttr): It is recommended to use ``minimize(loss, grad_clip=clip)`` to clip gradient. There are three clipping strategies: :ref:`api_fluid_clip_GradientClipByGlobalNorm` , :ref:`api_fluid_clip_GradientClipByNorm` , :ref:`api_fluid_clip_GradientClipByValue` . + + Please use 'paddle.nn.utils.weight_norm' in dygraph mode. Args: dim(int): Dimension over which to compute the norm. Dim is a non-negative
[ParamAttr->[_set_default_bias_initializer->[_set_default_initializer],_to_attr->[ParamAttr,_to_attr],_set_default_param_initializer->[_set_default_initializer]]]
A class which represents a parameter of weight normalization. Optional. If regularizer is set in optimizer it will be ignored.
move this line to the beginning of the docstring would be better. (i.e.: Line 206)
@@ -145,7 +145,7 @@ func IsRegistryDockerHub(registry string) bool { func ParseDockerImageReference(spec string) (DockerImageReference, error) { var ref DockerImageReference - namedRef, err := reference.ParseNamedDockerImageReference(spec) + namedRef, err := parseNamedDockerImageReference(spec) if err != nil { return ref, err }
[Exact->[NameString],String->[Exact],RegistryHostPort->[DockerClientDefaults],DaemonMinimal->[Minimal],Exact,Equal,MostSpecific]
ParseImageStreamTagName parses a string into its name and tag components. It returns an error if Equal returns true if the receiver has the same default values as the other.
@deads2k We can have it in the versioned API because registry also uses `DockerImageReference` ?
@@ -101,7 +101,7 @@ func NewDefaultMasterArgs() *MasterArgs { MasterAddr: flagtypes.Addr{Value: "localhost:8443", DefaultScheme: "https", DefaultPort: 8443, AllowPrefix: true}.Default(), EtcdAddr: flagtypes.Addr{Value: "0.0.0.0:4001", DefaultScheme: "https", DefaultPort: 4001}.Default(), MasterPublicAddr: flagtypes.Addr{Value: "localhost:8443", DefaultScheme: "https", DefaultPort: 8443, AllowPrefix: true}.Default(), - DNSBindAddr: flagtypes.Addr{Value: "0.0.0.0:53", DefaultScheme: "tcp", DefaultPort: 53, AllowPrefix: true}.Default(), + DNSBindAddr: flagtypes.Addr{Value: "0.0.0.0:8053", DefaultScheme: "tcp", DefaultPort: 8053, AllowPrefix: true}.Default(), ConfigDir: &util.StringFlag{},
[GetAssetPublicAddress->[GetMasterPublicAddress],BuildSerializeableMasterConfig->[GetPolicyFile],GetMasterPublicAddress->[GetMasterAddress],GetEtcdPeerAddress->[GetEtcdAddress],Validate->[Validate],GetEtcdAddress->[GetMasterAddress]]
NewDefaultMasterArgs creates a new master args object with default values set. GetConfigFileToWrite returns the configuration filepath for the master .
This broke `TEST_END_TO_END=direct hack/test-end-to-end.sh` because it longer starts DNS on 53, so our lookups fail.
@@ -161,7 +161,8 @@ void Sky::render() float wn = nightlength / 2; float wicked_time_of_day = 0; if (m_time_of_day > wn && m_time_of_day < 1.0 - wn) - wicked_time_of_day = (m_time_of_day - wn) / (1.0 - wn * 2) * 0.5 + 0.25; + wicked_time_of_day = (m_time_of_day - wn) / (1.0 - wn * 2) * 0.5 + + 0.25; else if (m_time_of_day < 0.5) wicked_time_of_day = m_time_of_day / wn * 0.25; else
[OnRegisterSceneNode->[registerNodeForRendering],ISceneNode->[normalize,myrand_range,setAutomaticCulling,get_scene_manager,getTextureForMesh,getBool,set,v3f,getTexture,isKnownSourceImage,m_materials],render->[setAlpha,fabs,getInterpolated,rotateVect,getNearValue,setScale,unlock,getAbsolutePosition,drawIndexedTriangleFan,getFarValue,drawVertexPrimitiveList,setMaterial,rotateXYBy,getGreen,setTransform,lock,setTranslation,getRed,v3f,easeCurve,getVideoDriver,getActiveCamera,getBlue,MYMAX,buildRotateFromTo,sin,toSColor,rotateXZBy,MYMIN,drawIndexedTriangleList],update->[wrapDegrees_0_360,getInterpolated,m_mix_scolor,getAlpha,getGreen,m_horizon_blend,update,getBlue,m_mix_scolorf,getRed,toSColor,rangelim,MYMIN,m_materials]]
This is the main entry point for the skeleton rendering. Standard frame in the FITS format This is a low level API that can be used to draw a cloudy fog with Draw all the horizons of the three cloudy fog segments Draws a network segment of a single network segment.
Lines only need splitting if over 90 columns (with tab size 4). If <=90 columns the previous format was better.
@@ -24,6 +24,7 @@ import io.confluent.ksql.function.udf.UdfDescription; + " Default masking rules will replace all upper-case characters with 'X', all lower-case" + " characters with 'x', all digits with 'n', and any other character with '-'.") public class MaskKeepLeftKudf { + private final String udfName = Masker.getMaskUdfName(this); @Udf(description = "Returns a masked version of the input string. All characters except for the" + " first n will be replaced according to the default masking rules.")
[MaskKeepLeftKudf->[doMask->[validateParams,append,min,substring,mask,toString,StringBuilder,length],validateParams->[KsqlFunctionException],mask->[doMask,getMaskCharacter,Masker]]]
Returns a masked version of the input string. All characters except for the last n will be.
This variable cannot be uppercase nor static because 1. the UDF name uses the class annotation which cannot be obtained statically, and 2. checkstyle fails with more than 3 uppercase letters in non-static variables. Same for the other udfName variables.
@@ -21,12 +21,13 @@ bool process_record_kb(uint16_t keycode, keyrecord_t *record) { return process_record_user(keycode, record); } +__attribute__ ((weak)) void led_set_user(uint8_t usb_led) { if (usb_led & (1 << USB_LED_CAPS_LOCK)) { - DDRB |= (1 << 7); + DDRB |= (1 << 7); PORTB &= ~(1 << 7); } else { - DDRB &= ~(1 << 7); + DDRB &= ~(1 << 7); PORTB &= ~(1 << 7); } }
[matrix_init_kb->[matrix_init_user],process_record_kb->[process_record_user],matrix_scan_kb->[matrix_scan_user]]
This function processes a keycode and a keyrecord_t object.
Well, we don't need the DDRB for each here. .... it should only be called in a `keyboard_pre_init_kb` function And also, it may be best to convert these to GPIO commands.
@@ -10,9 +10,11 @@ from ..product.utils import allocate_stock, deallocate_stock, increase_stock def check_order_status(func): - """Prevent execution of decorated function if order is fully paid. + """Check if order meets preconditions of payment process. - Instead redirects to order details page. + Order can not have draft status or be fully paid. Billing address + must be provided. + If not, redirect to order details page. """ # pylint: disable=cyclic-import from .models import Order
[recalculate_order->[save,get_total,sum],restock_fulfillment_lines->[increase_stock],cancel_order->[save,all,restock_order_lines],cancel_fulfillment->[restock_fulfillment_lines,update_order_status,save],attach_order_to_user->[store_user_address,save],merge_duplicates_into_order_line->[save,sum,filter,exclude,count],add_variant_to_existing_lines->[allocate_stock,filter,F,save],update_order_status->[save,get_total_quantity],add_variant_to_order->[allocate_stock,create,display_product,add_variant_to_existing_lines,InsufficientStock,refresh_from_db,get_price_per_item,select_stockrecord],change_order_line_quantity->[save,delete],check_order_status->[decorator->[is_fully_paid,redirect,pop,func,get_object_or_404],wraps],restock_order_lines->[save,increase_stock,deallocate_stock]]
Prevent execution of decorated function if order is fully paid. Instead redirects to order details page.
I'd assign it to some sort of variable and then pass it to `min()` function, it would be more readable
@@ -4,6 +4,9 @@ import static com.google.common.base.Preconditions.checkNotNull; import games.strategy.util.IntegerMap; +/** + * A repair rule. + */ public class RepairRule extends DefaultNamed { private static final long serialVersionUID = -45646671022993959L; private final IntegerMap<Resource> m_cost;
[RepairRule->[addCost->[put],getCosts->[copy],addResult->[IllegalArgumentException,put,getName],toString->[getName],checkNotNull,copy]]
Creates a new instance of RepairRule.
Restating class name, describing the scope or responsibility of this class, what it represents, what is a 'repair rule' would be good information to have.
@@ -61,8 +61,9 @@ public class DBScanner implements Callable<Void>, SubcommandWithParent { private String tableName; @CommandLine.Option(names = {"--with-keys"}, + required = true, description = "List Key -> Value instead of just Value.", - defaultValue = "false", + defaultValue = "true", showDefaultValue = CommandLine.Help.Visibility.ALWAYS) private static boolean withKey;
[DBScanner->[printAppropriateTable->[displayTable,getColumnFamilyHandle,constructColumnFamilyMap]]]
Imports a single - line sub - command for scanning a table. Reads the next N object from the given iterator and displays it in a table.
Thanks @sky76093016 for working on this. Do we need to change the default value of this option `--with-keys` to true?
@@ -81,7 +81,11 @@ func (c *CmdGitCreate) Run() error { } dui := c.G().UI.GetDumbOutputUI() - dui.Printf("Repo created! You can clone it with:\n git clone %s\n", urlString) + dui.Printf(`Repo created! You can clone it with: + git clone %s +Or add it as a remote to an existing repo with: + git remote add origin %s +`, urlString, urlString) return nil }
[Run->[Printf,G,GetDumbOutputUI,runPersonal,String,runTeam],runPersonal->[GetUsername,CreatePersonalRepo,Sprintf,Background,G],ParseArgv->[Args,New,GitRepoName,String,TeamNameFromString,Bool],runTeam->[Sprintf,Background,CreateTeamRepo],NewContextified,ChooseCommand]
Run executes the git command.
Not sure we should be recommending `origin` to people for existing repos, since that could easily fail if there's already an `origin` and it would probably be too verbose to explain all the subtleties here. But I really don't know for sure, and I think @malgorithms is the best choice of reviewer for user-facing wording issues.
@@ -188,6 +188,8 @@ public class HttpContentDecoderTest { @Test public void testResponseBrotliDecompression() throws Throwable { Brotli.ensureAvailability(); + // Failing on windows atm + Assume.assumeFalse(PlatformDependent.isWindows()); HttpResponseDecoder decoder = new HttpResponseDecoder(); HttpContentDecoder decompressor = new HttpContentDecompressor(); HttpObjectAggregator aggregator = new HttpObjectAggregator(Integer.MAX_VALUE);
[HttpContentDecoderTest->[testCleanupThrows->[channelInactive->[channelInactive]]]]
Test response brotli decompression.
@slandelle @hyperxpro these two tests fail on windows... I wonder if the native code not works as expected on windows.
@@ -216,12 +216,12 @@ def get_cmake_lib_files(name, version, package_name="Pkg"): package_name=package_name), "src/{}.cpp".format(name): source_cpp.format(name=name, version=version), "src/{}.h".format(name): source_h.format(name=name, version=version), - "src/CMakeLists.txt": cmake_v2.format(name=name, version=version), + "CMakeLists.txt": cmake_v2.format(name=name, version=version), "test_package/conanfile.py": test_conanfile_v2.format(name=name, version=version, package_name=package_name), "test_package/src/example.cpp": test_main.format(name=name), - "test_package/src/CMakeLists.txt": test_cmake_v2.format(name=name)} + "test_package/CMakeLists.txt": test_cmake_v2.format(name=name)} return files
[get_cmake_lib_files->[format],get_cmake_exe_files->[format]]
Get a list of files that can be found in the CMake library.
Maybe we want to follow CMake common pattern that is the top root uses a ``add_subdirectory(src)`` and keep a CMakeLists.txt inside "src" with the heavy logic.
@@ -145,7 +145,7 @@ class SubmissionsController < ApplicationController # we need to give them a list of all Groupings for this Assignment. if current_user.ta? groupings = [] - assignment.ta_memberships.find_all_by_user_id(current_user.id).each do |membership| + assignment.ta_memberships.where(user_id: current_user.id).each do |membership| groupings.push(membership.grouping) end elsif current_user.admin?
[SubmissionsController->[populate_repo_browser->[directories_at_path,repo,find,to_i,render,construct_repo_browser_directory_table_row,files_at_path,get_revision,join,construct_repo_browser_table_row,first,each,repository_folder,assignment,id],update_submissions->[find,size,group_name,post?,first,released_to_students,nil?,t,get_latest_result,get_submission_used,update_results_stats,each,marking_state,push,redirect_to,has_submission?,unrelease_results,has_result?,save],update_files->[find,can_collect_now?,original_filename,commit_after_collection_message,accepted_grouping_for,join,remove,has_jobs?,repo,nil?,conflicts,message,replace,sanitize_file_name,read,get_transaction,each,repository_folder,redirect_to,is_valid?,user_name,raise,add,commit,content_type],unrelease->[redirect_to,nil?,unrelease_results,post?,each],populate_submissions_table->[admin?,find,groupings,render,construct_submissions_table_row,grouping,ta?,each,id,push],download->[nil?,repo,find,get_latest_revision,to_i,render,send_data,download_as_string,find_appropriate_grouping,files_at_path,message,get_revision,join,is_binary?,repository_folder,id],download_svn_export_commands->[get_svn_commands,short_identifier,send_data,find],browse->[find],manually_collect_and_begin_grading->[redirect_to,to_i,apply_submission_rule,find,get_latest_result,assignment,id,create_by_revision_number],repo_browser->[find,timestamp,repository_name,message,revision_number,first,assignment],download_simple_csv_report->[short_identifier,all,nil?,find,user_name,send_data,generate,has_submission?,accepted_grouping_for,total_mark,get_submission_used,each,id,push],populate_file_manager->[directories_at_path,find,accepted_grouping_for,join,first,construct_file_manager_table_row,group,repo,nil?,construct_file_manager_dir_table_row,get_revision,to_i,repository_external_commits_only?,each,repository_folder,id,get_latest_revision,render,files_at_path],collect_and_begin_grading->[redirect_to,find,apply_submission_rule,can_collect_now?,get_latest_result,group_name,localtime,create_by_timestamp,id],file_manager->[redirect_to,group,nil?,repo,find,get_latest_revision,path_exists?,filename,files_at_path,accepted_grouping_for,join,each,repository_folder,id,push],index->[all,render],download_detailed_csv_report->[find_by_membership_id,all,find,accepted_grouping_for,rubric_criteria,get_total_extra_points,nil?,total_mark,mark,find_by_user_id,generate,get_submission_used,get_total_extra_percentage,each,id,push,short_identifier,send_data,user_name,find_by_rubric_criterion_id,has_submission?,weight],include,before_filter]]
This view populates the submission table with all the groupings that the current user has assigned to.
Line is too long. [84/80]
@@ -106,6 +106,8 @@ void FWRetract::retract(const bool retracting SERIAL_ECHOLNPAIR("hop_amount ", hop_amount); //*/ + stepper.synchronize(); // Wait for buffered moves to complete + const bool has_zhop = retract_zlift > 0.01; // Is there a hop set? const float old_feedrate_mm_s = feedrate_mm_s;
[No CFG could be retrieved]
Determines the Z height of the next frame in the table. This method is called from the base class when G - code includes a header. It will.
This is the only change needed.
@@ -2318,3 +2318,16 @@ func WithPodSlirp4netns(networkOptions map[string][]string) PodCreateOption { return nil } } + +// WithVolatile sets the volatile flag for the container storage. +// The option can potentially cause data loss when used on a container that must survive a machine reboot. +func WithVolatile() CtrCreateOption { + return func(ctr *Container) error { + if ctr.valid { + return define.ErrCtrFinalized + } + + ctr.config.Volatile = true + return nil + } +}
[WithPod->[ID],IsRootless,GetSecretsStorageDir,AddLinuxGIDMapping,StringInSlice,Wrap,GetRootlessUID,Stat,MatchString,ClearLinuxUIDMappings,New,NewManager,InfraContainerID,IsValidEventer,AddLinuxUIDMapping,Debugf,ID,Wrapf,Join,ClearLinuxGIDMappings,Lookup,ToLower,ProcessOptions,ParseIP,NetworkMode,DefaultStoreOptions,IsDir]
null - > .
Can you add a warning here that this is potentially dangerous to use and can result in data loss?
@@ -343,8 +343,8 @@ func newNewCmd() *cobra.Command { &yes, "yes", "y", false, "Skip prompts and proceed with default values") cmd.PersistentFlags().StringVar( - &secretsProvider, "secrets-provider", "", "The name of the provider that should be used to encrypt and "+ - "decrypt secrets.") + &secretsProvider, "secrets-provider", "default", "The type of the provider that should be used to encrypt and "+ + "decrypt secrets (possible choices: default, passpharse)") return cmd }
[StringVar,Value,Colorize,SetHelpFunc,Warningf,Encrypter,SaveProject,Delete,ValueOrDefaultProjectDescription,Wrap,Interactive,AskOne,CopyTemplateFilesDryRun,ReadConsoleNoEcho,IsNotExist,Strings,Stat,IgnoreError,New,RunFunc,StringArrayVarP,StringVarP,ReadDir,Errorf,RetrieveTemplates,Ref,Run,Chdir,GetStack,SplitN,TrimSpace,Wrapf,GetGlobalColorization,BoolVarP,Tags,Name,SetCurrentStack,HelpFunc,Sort,ReadConsole,PackageName,IsSpace,Highlight,Base,Rel,MkdirAll,ValueOrSanitizedDefaultProjectName,Command,NewValue,Printf,Templates,EqualFold,Decrypter,NewSecureValue,Println,CopyTemplateFiles,Sprintf,EncryptValue,ValidateProjectName,ParseStackReference,Print,String,Getwd,MaximumNArgs,PersistentFlags,EmojiOr]
Config to save getStack returns the stack name and description of the given stack.
passpharse => passphrase (throughout)
@@ -42,6 +42,8 @@ from nilearn.plotting import plot_glass_brain print(__doc__) +print(len(warnings.filters)) + ############################################################################### # Setup paths sample_dir_raw = sample.data_path()
[read_evokeds,compute_source_morph,show,print,data_path,apply,add_overlay,apply_inverse,read_inverse_operator,crop,join,plot_glass_brain,load]
Plots the data of a single . x_glr_auto_examples_inverse_plot_compute_mne_.
this is temporary, right?
@@ -118,11 +118,12 @@ void ossl_statem_set_renegotiate(SSL *s) void ossl_statem_fatal(SSL *s, int al, int func, int reason, const char *file, int line) { + ERR_put_error(ERR_LIB_SSL, func, reason, file, line); /* We shouldn't call SSLfatal() twice. Once is enough */ - assert(s->statem.state != MSG_FLOW_ERROR); + if (s->statem.in_init && s->statem.state == MSG_FLOW_ERROR) + return; s->statem.in_init = 1; s->statem.state = MSG_FLOW_ERROR; - ERR_put_error(ERR_LIB_SSL, func, reason, file, line); if (al != SSL_AD_NO_ALERT && s->statem.enc_write_state != ENC_WRITE_STATE_INVALID) ssl3_send_alert(s, SSL3_AL_FATAL, al);
[No CFG could be retrieved]
The function below is called from the server side of the SSL handshake process. This function is called by the server when the current connection is in the error state. It.
Oops, I have accidentally used 2-spaces GCC-indentation style...
@@ -144,7 +144,11 @@ Status getProcList(std::set<long>& pids) { return Status(0, "Ok"); } -void genProcess(const WmiResultItem& result, QueryData& results_data) { +void genProcess( + const WmiResultItem& result, + QueryData& results_data, + QueryContext& context, + std::map<std::int32_t, std::map<std::string, std::int64_t>> perfData) { Row r; Status s; long pid;
[genProcessMemoryMap->[genMemoryMap,getProcList],genProcesses->[genProcess]]
Get the list of processes that match the query. region PrivatePageCount GetModuleFileName GetProcessTimes GetSystemTime GetProcessTime GetSystem Get the process UID and GID from its SID handle.
Make this a `std::map<std::int32_t, std::map<std::string, std::int64_t>>&`, and I'd say also make it the second parameter of the call, as we tend to keep the context and result params at the end.
@@ -222,13 +222,9 @@ func testAccAWSVpnConnectionDisappears(connection *ec2.VpnConnection) resource.T _, err := conn.DeleteVpnConnection(&ec2.DeleteVpnConnectionInput{ VpnConnectionId: connection.VpnConnectionId, }) + if err != nil { - if ec2err, ok := err.(awserr.Error); ok && ec2err.Code() == "InvalidVpnConnectionID.NotFound" { - return nil - } - if err != nil { - return err - } + return err } return resource.Retry(40*time.Minute, func() *resource.RetryError {
[ParallelTest,NonRetryableError,Code,RandInt,RandIntRange,RootModule,ComposeTestCheckFunc,Errorf,MustCompile,RetryableError,RandStringFromCharSet,DescribeVpnConnections,DeleteVpnConnection,Fatalf,Meta,Sprintf,TestCheckResourceAttr,String,Retry]
TestAccAWSVpnConnectionDisappears tests if the given connection has a specific TestAccAwsVpnConnectionDestroy - destroy a VPN Connection if it is in the correct state.
The acceptance test should fail if this error is returned here.
@@ -781,6 +781,11 @@ func NewIssue(ctx *context.Context) { ctx.Data["TitleQuery"] = title body := ctx.Query("body") ctx.Data["BodyQuery"] = body + + // get permalink query + permalink := ctx.Query("permalink") + ctx.Data["Permalink"] = permalink + ctx.Data["IsProjectsEnabled"] = ctx.Repo.CanRead(models.UnitTypeProjects) ctx.Data["IsAttachmentEnabled"] = setting.Attachment.Enabled upload.AddUploadContext(ctx, "comment")
[TeamReviewRequest,Status,LoadTime,GetIssueByIndex,Warn,IsPoster,Int64sToMap,CreateCommentReaction,ChangeProjectAssign,CanUseTimetracker,GetRefEndNamesAndURLs,CanEnablePulls,AddParam,IsCollaborator,Redirect,IsErrDependenciesLeft,Info,GetOwner,GetTreeEntryByPath,IsErrIssueNotExist,ChangeIssueRef,IsValidTeamReviewRequest,GetRepositoryByID,MergeBlockedByOutdatedBranch,ChangeStatus,GetErrMsg,New,IsErrPullRequestNotExist,GetLabelsByOrgID,GetUserRepoPermission,GetAssignees,NewGhostUser,ChangeMilestoneAssign,IsErrReviewNotExist,ParseInt,IsErrUserDoesNotHaveAccessToRepo,GetIsRead,ServerError,ParamsInt64,LoadBaseRepo,LoadLabel,MergeBlockedByRejectedReview,GetForm,GetUserTeams,Render,IsOwner,MergeBlockedByOfficialReviewRequests,CalcCommitStatus,Split,UpdateAttachments,BlockingDependencies,IsUserRealRepoAdmin,IsTrace,LoadResolveDoer,CheckIssueWatch,LoadRepo,GetReviewers,CanMarkConversation,GetMilestones,TotalTimes,DataAsync,HTMLString,DeleteComment,Trim,RelAvatarLink,LoadReview,ComposeMetas,LoadRepositories,Close,IsTimetrackerEnabled,RefEndName,ChangeContent,CanReadIssuesOrPulls,IsAdmin,GetReviewerTeams,DeleteAttachment,RenderWithErr,RepoPath,LoadSelectedLabelsAfterClick,LoadMilestone,LoadReviewerTeam,ToReleaseAttachment,LoadIssue,JSON,IsOrganization,IsUserAllowedToMerge,IsEmptyString,LoadHeadRepo,GroupByType,GetLastCommitStatus,Current,StateType,IsOwnedBy,GetReviewersFromOriginalAuthorsByIssueID,LoadReviewer,CommentTypeIsRef,IsErrUserNotExist,IsBranchExist,DeleteNotPassedAssignee,HasEnoughApprovals,ChangeTitle,HasUserStopwatch,OptionalBoolOf,LoadAssigneeUserAndTeam,EqualFold,HasForkedRepo,IsOfficialReviewer,GetIssueStats,GetReviewersByIssueID,GetApprovalCounts,ExtractMetadata,GetIssuesByIDs,NewIssue,GetIssueByID,GetAttachmentsByIssueID,NewPagination,GetLabelsByRepoID,GetProjects,CanWrite,Size,CanWriteIssuesOrPulls,NotFoundOrServerError,GetUnmergedPullRequest,LoadProject,UpdateComment,GetUserByID,StringsToInt64s,GetBranches,ExternalTrackerConfig,GetBranchCommit,LoadDepIssueDetails,GetTeams,GetAttachmentsByCommentID,Sprint,GetTeamByID,NotFound,Trace,LoadAttachments,LoadProtectedBranch,AddUploadContext,Tr,LoadReactions,QueryTrim,ToggleAssignee,MustHeadUserName,AddToTaskQueue,GetGrantedApprovalsCount,NewColoredIDValue,QueryStrings,DeleteCommentReaction,Query,CanCreateIssueDependencies,Expand,IsValidReviewRequest,IsMergeStyleAllowed,Sprintf,LoadPushCommits,BlockedByDependencies,IsProtectedBranch,IsErrForbiddenIssueReaction,QueryInt,Issues,ReadAll,QueryInt64,ReviewRequest,SearchIssuesByKeyword,IsErrWontSign,GetUnit,LoadAttributes,LoadCodeComments,HTML,Error,HashTag,GetProjectByID,Errorf,StopwatchExists,CanRead,IssueList,LoadPullRequest,HasError,ReadBy,IsStringInSlice,Debug,Join,Written,LoadPoster,GetGitRefName,Contains,CanBeAssigned,GetCommentByID,CreateIssueReaction,IssueTemplatesFromDefaultBranch,SignMerge,DeleteIssueReaction,CreateIssueComment,GetMilestoneByID,PullRequestsConfig,HTMLURL,FullName,IsErrUnitTypeNotExist,Blob,IsErrTeamNotExist,IsErrNotValidReviewRequest,Params]
NewIssue render creates a new issue page with all the necessary data GetProjectByID returns the project with the given ID.
Why don't you just use the body query parameter we already have here?
@@ -203,7 +203,12 @@ class HyperoptTuner(Tuner): self.json = None self.total_data = {} self.rval = None + self.CL_rval = None self.supplement_data_num = 0 + self.parallel = parallel_optimize + self.constant_liar_type = constant_liar_type + self.running_data = [] + self.optimal_y = 0 def _choose_tuner(self, algorithm_name): """
[_split_index->[_split_index],json2vals->[json2vals],HyperoptTuner->[get_suggestion->[json2parameter],import_data->[_add_index,receive_trial_result],generate_parameters->[_split_index],update_search_space->[_choose_tuner,json2space],__init__->[OptimizeMode],receive_trial_result->[json2vals]],_add_index->[_add_index],json2space->[json2space],json2parameter->[json2parameter]]
Initialize the object with the specified n - node algorithm.
what's the "CL" means? What's the "self.CL_rval" represent?
@@ -34,7 +34,7 @@ class PluginGridRow extends PKPPluginGridRow { * @param $plugin Plugin * @return boolean */ - function _canEdit(&$plugin) { + protected function _canEdit($plugin) { if ($plugin->isSitePlugin()) { if (in_array(ROLE_ID_SITE_ADMIN, $this->_userRoles)) { return true;
[PluginGridRow->[_canEdit->[isSitePlugin]]]
Checks if the user can edit the resource.
I see now why underscore
@@ -200,8 +200,15 @@ func (logger *taggedLogger) Write(p []byte) (_ int, retErr error) { } func (logger *taggedLogger) Close() (*pfs.Object, int64, error) { + close(logger.msgCh) if logger.putObjClient != nil { + if err := logger.eg.Wait(); err != nil { + return nil, 0, err + } object, err := logger.putObjClient.CloseAndRecv() + // we set putObjClient to nil so that future calls to Logf won't send + // msg down logger.msgCh as we've just closed that channel. + logger.putObjClient = nil return object, logger.objSize, err } return nil, 0, nil
[userLogger->[clone],Write->[Logf,Write],downloadData->[Logf],uploadOutput->[Logf,Close],Process->[getTaggedLogger,Logf,downloadData,uploadOutput,Close,runUserCode],runUserCode->[Logf,userLogger],Write]
Close closes the object associated with the logger.
How could there be future calls to `Logf` after `Close` is called?
@@ -64,6 +64,10 @@ public final class InterfaceTestUtils })); try { + if (!defines(forwardingInstance, actualMethod)) { + continue; + } + actualMethod.invoke(forwardingInstance, actualArguments); } catch (Exception e) {
[InterfaceTestUtils->[assertProperForwardingMethodsAreCalled->[RuntimeException,getName,getDeclaredMethods,newProxy,apply,getParameterCount,invoke,format,getParameterTypes,getReturnType,assertEquals],assertAllMethodsOverridden->[fail,getDeclaredMethod,of,getMethods,getName,copyOf,getInterfaces,format,getParameterTypes,getReturnType,assertEquals,isAssignableFrom]]]
Assert that the forwarding methods of the given interface are called.
I think this would be easier to read with shorter variable names. Also, introduce a variable ```java Class<?> forwardingClass = forwardingInstance.getClass(); Method forwardingMethod = forwardingClass.getMethod(method.getName().method.getParameterTypes()); return forwardingClass == forwardingMethod.getDeclaringClass();
@@ -43,6 +43,10 @@ class PretrainedTransformerIndexer(TokenIndexer[int]): self._added_to_vocabulary = False self._padding_value = self.tokenizer.convert_tokens_to_ids([self.tokenizer.pad_token])[0] logger.info(f"Using token indexer padding value of {self._padding_value}") + self._add_special_tokens = add_special_tokens + self._max_length = max_length + self._stride = stride + self._truncation_strategy = truncation_strategy @overrides def count_vocab_items(self, token: Token, counter: Dict[str, Dict[str, int]]):
[PretrainedTransformerIndexer->[tokens_to_indices->[_add_encoding_to_vocabulary]]]
Initialize a with the given model name and namespace.
Did you figure this out? Seems like they should be?
@@ -829,6 +829,14 @@ class MessageBuilder: self.fail('Overloaded function signatures {} and {} overlap with ' 'incompatible return types'.format(index1, index2), context) + def overloaded_signatures_arg_specific(self, index1: int, context: Context) -> None: + self.fail('Overloaded function implementation cannot accept all possible arguments ' + 'of signature {}'.format(index1), context) + + def overloaded_signatures_ret_specific(self, index1: int, context: Context) -> None: + self.fail('Overloaded function implementation cannot produce return type ' + 'of signature {}'.format(index1), context) + def operator_method_signatures_overlap( self, reverse_class: str, reverse_method: str, forward_class: str, forward_method: str, context: Context) -> None:
[temp_message_builder->[MessageBuilder],pretty_or->[format],MessageBuilder->[return_type_incompatible_with_supertype->[format,fail],read_only_property->[format,fail],forward_operator_not_callable->[format,fail],unsupported_left_operand->[format,fail],too_many_arguments->[format,fail],unexpected_keyword_argument->[note,format,fail],undefined_in_superclass->[format,fail],string_interpolation_mixing_key_and_non_keys->[fail],unsupported_type_type->[format,fail],check_unusable_type->[does_not_return_value],yield_from_invalid_operand_type->[format,fail],requires_int_or_char->[fail],invalid_signature->[format,fail],key_not_in_mapping->[fail],cannot_determine_type_in_base->[fail],too_few_arguments->[format,fail],format_distinctly->[format],invalid_keyword_var_arg->[fail],has_no_attr->[format,fail],copy->[copy,MessageBuilder],invalid_index_type->[format,fail],incompatible_type_application->[fail],overloaded_signatures_overlap->[format,fail],redundant_cast->[format,note],string_interpolation_with_star_and_key->[fail],warn->[report],incompatible_conditional_function_def->[fail],format_simple->[format_simple,format],reveal_type->[format,fail],not_callable->[format,fail],incompatible_argument->[format,fail,unsupported_operand_types,format_simple,format_distinctly],no_variant_matches_arguments->[format,fail],could_not_infer_type_arguments->[format,fail],incompatible_typevar_value->[format,fail],untyped_function_call->[format,fail],cant_assign_to_method->[fail],typeddict_item_name_not_found->[format,fail],too_many_positional_arguments->[format,fail],override_target->[format],format->[format],fail->[report],invalid_var_arg->[fail],too_few_string_formatting_arguments->[fail],too_many_string_formatting_arguments->[fail],does_not_return_value->[format,fail],is_errors->[is_errors],unsupported_operand_types->[format,fail],missing_named_argument->[format,fail],wrong_number_values_to_unpack->[fail],type_not_iterable->[format,fail],signature_incompatible_with_supertype->[format,fail],operator_method_signatures_overlap->[format,fail],signatures_incompatible->[format,fail],typeddict_instantiated_with_unexpected_items->[format,fail],deleted_as_rvalue->[format,fail],typeddict_item_name_must_be_string_literal->[format,fail],incompatible_operator_assignment->[format,fail],deleted_as_lvalue->[format,fail],invalid_class_method_type->[fail],report->[report],invalid_method_type->[fail],base_class_definitions_incompatible->[format,fail],duplicate_argument_value->[format,fail],cannot_determine_type->[fail],note->[report],unsupported_placeholder->[fail],invalid_cast->[format,fail],argument_incompatible_with_supertype->[format,fail],cannot_instantiate_abstract_class->[fail]]]
Overloaded function signatures and overlap with a given return type.
"cannot" -> "does not" or "should". (But the next one is correct IMO.)
@@ -2035,6 +2035,11 @@ class Jetpack { /* Jetpack Options API */ + /** + * @param string $type Jetpack option type. + * + * @return array + */ public static function get_option_names( $type = 'compact' ) { return Jetpack_Options::get_option_names( $type ); }
[Jetpack->[generate_secrets->[generate_secrets],stat->[initialize_stats],do_server_side_stat->[do_server_side_stat],get_locale->[guess_locale_from_lang],disconnect_user->[disconnect_user],jetpack_show_user_connected_icon->[is_user_connected],admin_page_load->[try_registration,disconnect_user,is_user_connected],jetpack_track_last_sync_callback->[jetpack_track_last_sync_callback],do_stats->[do_stats,initialize_stats],build_stats_url->[build_stats_url],handle_unique_registrations_stats->[do_stats,stat],build_connect_url->[build_connect_url],is_user_connected->[is_user_connected],get_connected_user_data->[get_connected_user_data],try_registration->[try_registration],authorize_starting->[do_stats,stat],register->[try_registration],is_active->[is_active]]]
Get the names of all options.
Going to come back to this one until I fix everything under it so I don't mess up my spreadsheet of fixes. The CI does not show this as blocking.
@@ -94,8 +94,13 @@ class WindmillTimerInternals implements TimerInternals { @Override public void setTimer( - StateNamespace namespace, String timerId, Instant timestamp, TimeDomain timeDomain) { - timers.put(timerId, namespace, TimerData.of(timerId, namespace, timestamp, timeDomain)); + StateNamespace namespace, + String timerId, + String timerFamilyId, + Instant timestamp, + TimeDomain timeDomain) { + timers.put( + timerId, namespace, TimerData.of(timerId, timerFamilyId, namespace, timestamp, timeDomain)); timerStillPresent.put(timerId, namespace, true); }
[WindmillTimerInternals->[withPrefix->[WindmillTimerInternals]]]
Method to add a timer to the timers map.
This isn't quite right. The idea of having the timer namespace is that two timers with the same id but different namespaces should not conflict, but right now they will override each other. We need to make sure that they don't override each other in every XXXTimerInternals class. Here this involves making them a key in the table, as well as making it a part of the timerTag returned. We also must make sure that the default namespace results in the old tag being generated.
@@ -61,7 +61,7 @@ class UserBadgesController < ApplicationController if params[:reason].present? unless is_badge_reason_valid? params[:reason] - return render json: failed_json.merge(message: I18n.t('invalid_grant_badge_reason_link')), status: 400 + return render json: failed_json.merge(message: I18n.t("invalid_grant_badge_reason_link")), status: 400 end if route = Discourse.route_for(params[:reason])
[UserBadgesController->[can_assign_badge_to_user?->[is_api?,can_grant_badges?,nil?],create->[can_assign_badge_to_user?,render,to_i,merge,is_badge_reason_valid?,grant,render_serialized,route_for,present?,t,require,id],ensure_badges_enabled->[raise,enable_badges?],is_badge_reason_valid?->[route_for],destroy->[can_assign_badge_to_user?,render,find,revoke,user,require],fetch_badge_from_params->[nil?,raise,permit,find_by,blank?,require],username->[fetch_user_from_params,can_see_profile?,user_badges,show_inactive_accounts,raise,render_serialized,permit,select,includes,try,map],index->[limit,offset,to_i,new,pluck_first,where,render_serialized,permit,includes,count],before_action]]
Creates a new object.
no formatting change please
@@ -283,7 +283,7 @@ class Info(dict, MontageMixin): gantry_angle : float | None Tilt angle of the gantry in degrees. lowpass : float - Lowpass corner frequency in Hertz. + Lowpass corner frequency in Hertz. It is automatically set to half the sampling rate if there is otherwise no low-pass applied to the data. meas_date : datetime The time (UTC) of the recording.
[write_info->[write_meas_info],write_meas_info->[_check_dates,_rename_comps,_check_consistency],create_info->[_update_redundant,_check_consistency],anonymize_info->[_add_timedelta_to_stamp,_check_dates,_check_consistency],_empty_info->[Info,_update_redundant,_check_consistency],Info->[__deepcopy__->[copy],_check_consistency->[_unique_channel_names,_check_ch_keys],pick_channels->[pick_channels],__init__->[_format_trans]],_simplify_info->[Info,_update_redundant],read_meas_info->[_read_bad_channels,Info,_check_consistency,_update_redundant],_merge_info_values->[_check_isinstance,_flatten,_where_isinstance],_merge_info->[Info,_update_redundant,_merge_info_values,_check_consistency],_write_ch_infos->[copy],_make_ch_names_mapping->[copy,_unique_channel_names]]
Returns a list of events for the given MaxFilter object. Required fields from the object.
this will yield a "line too long" error from our style checker. Max line length is 79 characters. Your IDE should have a "rulers" setting that will add a vertical line after 79 characters, this can help you notice such problems before pushing. Also running `make flake` from the root of the local MNE-Python clone would have caught this.
@@ -60,8 +60,8 @@ class Container { } $key = self::SHARED_DEPENDENCY_KEYS[ Hook_Manager::class ]; + require_once __DIR__ . '/class-hook-manager.php'; if ( ! isset( $jetpack_autoloader_container_shared[ $key ] ) ) { - require_once __DIR__ . '/class-hook-manager.php'; $jetpack_autoloader_container_shared[ $key ] = new Hook_Manager(); } $this->dependencies[ Hook_Manager::class ] = &$jetpack_autoloader_container_shared[ $key ];
[Container->[register_dependencies->[get],__construct->[initialize_globals,register_dependencies,register_shared_dependencies]]]
Register shared dependencies.
Hi @reatang! Thanks for the PR. The `Container` class only needs to load the `class-hook-manager.php` file when the condition in the if statement below is true. So, I'm not sure that it makes sense to always load the `class-hook-manager.php` file here. Also, I don't think you're encountering a bug in the `jetpack-autoloader` package. It looks like you would like to use an object of the `Hook_Manager` class in your plugin. If your plugin needs to use a file that isn't already loaded, it's probably best to have the plugin load the file it needs. Is there a reason that your plugin needs to use a `Hook_Manager` object from a specific namespace?
@@ -0,0 +1,12 @@ +module DataUpdateScripts + class RemoveProRoles + def run + pro_role = Role.find_by(name: "pro") + + return unless pro_role + + pro_role.users.find_each { |u| u.remove_role(:pro) } + pro_role.destroy + end + end +end
[No CFG could be retrieved]
No Summary Found.
Do we have any concerns about running this on Forem instances other than DEV? Should we add a guard clause if so?
@@ -461,6 +461,11 @@ public class HoodieTestDataGenerator { () -> UUID.randomUUID().toString()); } + public Stream<HoodieRecord> generatePartialUpdateInsertsStream(String commitTime, Integer n, boolean isFlattened, String schemaStr, boolean containsAllPartitions) { + return generatePartialUpdateInsertsStream(commitTime, n, isFlattened, schemaStr, containsAllPartitions, + () -> partitionPaths[RAND.nextInt(partitionPaths.length)]); + } + /** * Generates new inserts, uniformly across the partition paths above. It also updates the list of existing keys. */
[HoodieTestDataGenerator->[generateUpdates->[generateUpdateRecord],generateUpdateRecord->[generateRandomValue],generateInsertsWithHoodieAvroPayload->[incrementNumExistingKeysBySchema,populateKeysBySchema,generateAvroPayload],generateUpdatesWithHoodieAvroPayload->[generateAvroPayload],createReplaceFile->[createMetadataFile],generateUniqueDeleteRecordStream->[generateRandomDeleteValue],generateUpdatesWithDiffPartition->[generateUpdateRecord],generateDeleteRecord->[generateDeleteRecord],generateRandomValue->[generateRandomValue],generateInserts->[generateInserts],generateUniqueUpdatesStream->[generateRandomValueAsPerSchema],generateGenericRecords->[generateGenericRecord],generateInsertsStream->[generateInsertsStream,generateRandomValueAsPerSchema],createCommitFile->[createCommitFile],generateUpdatesForAllRecords->[generateUpdateRecord],generateDeletes->[generateInserts],generateSameKeyInserts->[generateRandomValue],generateGenericRecord->[generateGenericRecord]]]
Generates a stream of records that will be inserted.
please split the parameters into two lines.
@@ -94,7 +94,8 @@ export class Platform { * @return {boolean} */ isWebKit() { - return /WebKit/i.test(this.navigator_.userAgent) && !this.isEdge(); + return /WebKit/i.test(this.navigator_.userAgent) && !this.isEdge() + && !(this.isOpera() && this.isAndroid()); } /**
[No CFG could be retrieved]
Determines if the current browser is a Safari Chrome Chrome Chrome Chrome Chrome Chrome Chrome Chrome Chrome Chrome Determines if a user agent is able to determine the major version of a page.
hmm, Opera is Webkit, any particular reason this was added?
@@ -122,7 +122,8 @@ public final class ScmBlockLocationProtocolClientSideTranslatorPB if (response.getStatus() == ScmBlockLocationProtocolProtos.Status.SCM_NOT_LEADER) { failoverProxyProvider - .performFailoverToAssignedLeader(response.getLeaderSCMNodeId()); + .performFailoverToAssignedLeader(response.getLeaderSCMNodeId(), + null); } return response; } catch (ServiceException e) {
[ScmBlockLocationProtocolClientSideTranslatorPB->[addSCM->[submitRequest,handleError],getScmInfo->[submitRequest,handleError],allocateBlock->[submitRequest,handleError],sortDatanodes->[submitRequest,handleError],deleteKeyBlocks->[submitRequest,handleError],close->[close]]]
Submits a request to the server and checks if the response is a leader or not.
Do we need this at all, as server is only sending exception for NotLeader
@@ -273,8 +273,8 @@ class SubmissionsController < ApplicationController revision = nil end if revision - @revisions_history << {:num => revision.revision_number, - :date => revision.timestamp} + @revisions_history << {num: revision.revision_number, + date: revision.timestamp} end end repo.close
[SubmissionsController->[downloads->[find,find_entry,group_name,join,first,send_file,nil?,get_output_stream,download_as_string,message,repo_name,get_revision,revision_number,t,last,count,to_i,access_repo,open,repository_folder,each,id,puts,short_identifier,get_latest_revision,render,files_at_path,mkdir,find_appropriate_grouping],update_submissions->[call,redirect_to,short_identifier,nil?,find,empty?,set_results_statistics,instance,raise,log,collect_submissions_for_section,post?,set_release_on_results,blank?,ta?,t,id,push],download_groupings_files->[submission_files,find,find_entry,join,send_file,map,get_output_stream,message,tr,repo_name,current_submission_used,exist?,t,filename,blank?,each,open,puts,delete,redirect_to,short_identifier,render,user_name,retrieve_file,mkdir],set_filebrowser_vars->[get_latest_revision,path_exists?,access_repo,filename,files_at_path,join,each,repository_folder,push],server_time->[render],populate_repo_browser->[directories_at_path,find,to_i,respond_to,js,render,access_repo,construct_repo_browser_directory_table_row,files_at_path,object_id,get_revision,join,construct_repo_browser_table_row,first,each,repository_folder,assignment],download_svn_repo_list->[get_svn_repo_list,short_identifier,send_data,find],collect_ta_submissions->[redirect_to,short_identifier,all,find,can_collect_now?,instance,push_groupings_to_queue,t,id],file_manager->[redirect_to,group,nil?,find,accepted_grouping_for,set_filebrowser_vars,id],download_svn_export_commands->[short_identifier,send_data,find,join,get_svn_export_commands],all_assignments_marked?->[where,size],collect_all_submissions->[redirect_to,short_identifier,groupings,find,can_collect_now?,instance,push_groupings_to_queue,t,id],browse->[get_filters,error_collecting,handle_paginate_event,to_s,all,find,to_i,detect,any?,blank?,ta?,present?,t,id,map],collect_and_begin_grading->[redirect_to,find,group_name,can_collect_grouping_now?,t,push_grouping_to_priority_queue,id],update_converted_pdfs->[nil?,find,is_converted,current_submission_used,each,is_pdf?],index->[all,render],update_files->[find,new,original_filename,can_collect_now?,commit_after_collection_message,accepted_grouping_for,set_filebrowser_vars,join,remove,has_jobs?,group,nil?,merge!,conflicts,message,replace,log,sanitize_file_name,t,to_i,instance,repository_external_commits_only?,access_repo,read,get_transaction,each,repository_folder,push,short_identifier,is_valid?,render,user_name,raise,rewind,add,commit,content_type],unrelease->[redirect_to,short_identifier,nil?,find,instance,unrelease_results,log,post?,each,t,length,id],download->[nil?,get_latest_revision,find,to_i,render,send_data,access_repo,download_as_string,find_appropriate_grouping,files_at_path,message,get_revision,join,is_binary?,t,repository_folder,escapeHTML,id],manually_collect_and_begin_grading->[redirect_to,to_i,find,manually_collect_submission,id],repo_browser->[repo,path_exists?,find,to_i,timestamp,close,repository_name,message,raise,get_revision,revision_number,join,first,each,repository_folder,assignment],populate_file_manager->[directories_at_path,find,respond_to,accepted_grouping_for,join,first,construct_file_manager_table_row,group,nil?,js,construct_file_manager_dir_table_row,get_revision,to_i,repository_external_commits_only?,access_repo,object_id,each,repository_folder,id,get_latest_revision,files_at_path],download_simple_csv_report->[get_simple_csv_report,send_data,find,short_identifier],download_detailed_csv_report->[short_identifier,send_data,find,get_detailed_csv_report],all,include,section,released_to_students,downcase,has_submission?,revision_timestamp,total_mark,remark_submitted?,grace_period_deduction_single,collect,before_filter,lambda,grouping,t,select,marking_state,helper_method],require]
This method retrieves a from the repository and creates a hash with the necessary information.
Space inside { missing.
@@ -12,8 +12,7 @@ elgg_push_context('owner_block'); // groups and other users get owner block $owner = elgg_get_page_owner_entity(); -if ($owner instanceof ElggGroup || - ($owner instanceof ElggUser && $owner->getGUID() != elgg_get_logged_in_user_guid())) { +if ($owner instanceof ElggGroup || $owner instanceof ElggUser) { $header = elgg_view_entity($owner, array('full_view' => false));
[getGUID]
Displays page owner block.
There's a tab after `||`. Change it to space and I'll merge.
@@ -641,7 +641,9 @@ namespace System.Drawing { // We threw this way on NetFX if (outputStream == null) +#pragma warning disable CA2208 // Instantiate argument exceptions correctly throw new ArgumentNullException("dataStream"); +#pragma warning restore CA2208 picture.SaveAsFile(new GPStream(outputStream, makeSeekable: false), -1, out int temp); }
[Icon->[DrawUnstretched->[DrawIcon],Bitmap->[BitmapHasAlpha,Draw,CopyBitmapData,Dispose],Draw->[Draw,DrawIcon],ExtractAssociatedIcon->[ExtractAssociatedIcon],Dispose->[Dispose,DestroyHandle],Dispose]]
Save the icon in the specified output stream.
Why not `nameof(outputStream)`?
@@ -713,9 +713,9 @@ export class AmpList extends AMP.BaseElement { .then(elements => this.render_(elements, current.append)); if (!isSSR) { const payload = /** @type {!JsonObject} */ (current.payload); - renderPromise = renderPromise - .then(() => this.maybeRenderLoadMoreTemplates_(payload)) - .then(() => this.maybeSetLoadMore_()); + renderPromise = renderPromise.then(() => + this.maybeRenderLoadMoreTemplates_(payload) + ); } renderPromise.then(onFulfilledCallback, onRejectedCallback); }
[AmpList->[undoLayout_->[FLEX_ITEM,RESPONSIVE,parseLayout,FLUID,FIXED_HEIGHT,getLayoutClass,FIXED,devAssert,setStyles,INTRINSIC],ssrTemplate_->[setupJsonFetchInit,fetchOpt,dict,userAssert,user,requestForBatchFetch,setupAMPCors,xhrUrl,setupInput],getPolicy_->[getSourceOrigin,ALL,OPT_IN],constructor->[templatesFor,HIGH],setLoadMore_->[setStyles],changeToLayoutContainer_->[CONTAINER,resolve,toggle],updateBindings_->[some,resolve,isArray,hasAttribute,querySelector,rescan,updateWith,bindForDocOrNull],diff_->[setDOM,dev,length],doRenderPass_->[then,resolver,append,data,rejecter,devAssert,scheduleNextPass,payload],addElementsToContainer_->[setAttribute,hasAttribute,dev,appendChild,forEach],mutatedAttributesCallback->[isArray,dev,user,getMode,renderLocalData],attemptToFit_->[CONTAINER,/*OK*/],attemptToFitLoadMoreElement_->[CONTAINER,dev,/*OK*/,setStyles],buildCallback->[viewerForDoc,user,bindForDocOrNull,scopedQuerySelector],prepareAndSendFetch_->[getViewerAuthTokenIfAvailable],showFallbackOrThrow_->[childElementByAttr],truncateToMaxLen_->[slice,length,parseInt],layoutCallback->[dev],render_->[removeChildren,dev,hasChildNodes,resetPendingChangeSize,DOM_UPDATE,createCustomEvent],maybeResizeListToFitItems_->[dev],manuallyDiffElement_->[nodeName,getAttribute,some,hasAttribute,setAttribute,classList,devAssert,parentElement,startsWith],maybeLoadMoreItems_->[dev,bottom],fetchList_->[trigger,resolve,getValueForExpr,isArray,userAssert,user,dict,actionServiceForDoc,response,catch,LOW,createCustomEvent],isTabbable_->[matches],maybeRenderLoadMoreTemplates_->[resolve,all,push],isLayoutSupported->[isLayoutSizeDefined],maybeRenderLoadMoreElement_->[dev,removeChildren,resolve,appendChild],resetIfNecessary_->[dev,removeChildren,toArray],markContainerForDiffing_->[forEach,querySelectorAll,String,markElementForDiffing],loadMoreCallback_->[dev,resolve,tryFocus],firstTabbableChild_->[scopedQuerySelector],adjustContainerForLoadMoreButton_->[dev,setStyles,px],initializeLoadMoreElements_->[toggle,listen],fetch_->[batchFetchJsonFor],createContainer_->[setAttribute],updateLoadMoreSrc_->[getValueForExpr],handleLoadMoreFailed_->[dev],lastTabbableChild_->[scopedQuerySelectorAll,length],maybeSetLoadMore_->[resolve],BaseElement],registerElement,extension]
Render the next render pass.
Does it matter that the order changed here?
@@ -384,12 +384,16 @@ def set_action_env_var(environ_cp, def convert_version_to_int(version): """Convert a version number to a integer that can be used to compare. + Version strings of the form X.YZ and X.Y.Z-xxxxx are supported. The + 'xxxxx' part, for instance 'homebrew' on OS/X, is ignored. + Args: - version: a version to be covnerted + version: a version to be converted Returns: An integer if converted successfully, otherwise return None. """ + version = version.split('-')[0] version_segments = version.split('.') for seg in version_segments: if not seg.isdigit():
[set_cc_opt_flags->[write_to_bazelrc,is_ppc64le],set_build_var->[write_to_bazelrc,get_var],set_computecpp_toolkit_path->[is_linux,write_action_env_to_bazelrc,get_from_env_or_user_or_default],set_clang_cuda_compiler_path->[run_shell,write_action_env_to_bazelrc,get_from_env_or_user_or_default],set_tf_cunn_version->[run_shell,cygpath,is_windows,is_macos,get_from_env_or_user_or_default,is_linux,write_action_env_to_bazelrc],set_tf_cuda_clang->[set_action_env_var],setup_python->[cygpath,is_windows,get_input,write_to_bazelrc,get_python_path,write_action_env_to_bazelrc],set_mkl->[write_to_bazelrc],cygpath->[run_shell],set_host_c_compiler->[run_shell,write_action_env_to_bazelrc,get_from_env_or_user_or_default],main->[set_cc_opt_flags,is_windows,set_build_var,set_computecpp_toolkit_path,is_macos,set_clang_cuda_compiler_path,set_tf_cunn_version,set_tf_cuda_clang,cleanup_makefile,setup_python,set_mkl,set_host_c_compiler,set_gcc_host_compiler_path,set_other_mpi_vars,set_tf_cuda_compute_capabilities,set_action_env_var,set_other_cuda_vars,run_gen_git_source,check_bazel_version,reset_tf_configure_bazelrc,set_tf_cuda_version,set_host_cxx_compiler,set_mpi_home],set_gcc_host_compiler_path->[run_shell,write_action_env_to_bazelrc,get_from_env_or_user_or_default],get_from_env_or_user_or_default->[get_input],set_other_mpi_vars->[sed_in_place,symlink_force],set_tf_cuda_compute_capabilities->[get_from_env_or_user_or_default,write_action_env_to_bazelrc,get_native_cuda_compute_capabilities],write_action_env_to_bazelrc->[write_to_bazelrc],set_action_env_var->[get_var,write_action_env_to_bazelrc],set_other_cuda_vars->[write_to_bazelrc,write_action_env_to_bazelrc,is_windows],get_var->[get_input],check_bazel_version->[run_shell,convert_version_to_int],reset_tf_configure_bazelrc->[remove_line_with],set_tf_cuda_version->[cygpath,is_windows,is_macos,get_from_env_or_user_or_default,is_linux,write_action_env_to_bazelrc],get_native_cuda_compute_capabilities->[run_shell],set_host_cxx_compiler->[run_shell,write_action_env_to_bazelrc,get_from_env_or_user_or_default],set_mpi_home->[run_shell,get_from_env_or_user_or_default],main]
Convert a version number to a integer that can be used to compare. covnerted.
Should we check that the leading parts are not zero? Otherwise the version number will not be unique, e.g. 0.23 and 0.0.23 would results in the same integer.
@@ -12,11 +12,11 @@ import KratosMultiphysics.KratosUnittest as KratosUnittest import KratosMultiphysics.kratos_utilities as kratos_utilities - +@KratosUnittest.skipIfApplicationsNotAvailable("StructuralMechanicsApplication") class ROMDynamicStruct(KratosUnittest.TestCase): ######################################################################################### - @KratosUnittest.skipIf(numpy_available == False, "numpy is required for RomApplication") + @KratosUnittest.skipUnless(numpy_available,"numpy is required for RomApplication") def test_Struct_Dynamic_ROM_2D(self): with KratosUnittest.WorkFolderScope(".", __file__):
[ROMDynamicStruct->[test_Struct_Dynamic_ROM_2D->[EvaluateQuantityOfInterest2,TestStructuralMechanicsDynamicROM,sqrt,sum,assertLess,read,EvaluateQuantityOfInterest,Parameters,DeleteDirectoryIfExisting,Run,open,range,WorkFolderScope,Model,load,shape],skipIf],main,GetDefaultOutput]
Tests if a structure is dynamic - rom with 2D parameters.
maybe you also need the `LinearSolversApp`?