patch stringlengths 18 160k | callgraph stringlengths 4 179k | summary stringlengths 4 947 | msg stringlengths 6 3.42k |
|---|---|---|---|
@@ -5072,9 +5072,14 @@ namespace System.Windows.Forms
/// <summary>
/// Called by ToolTip to poke in that Tooltip into this ComCtl so that the Native ChildToolTip is not exposed.
/// </summary>
- internal void SetToolTip(ToolTip toolTip, string toolTipCaption)
+ protected override void SetToolTip(ToolTip toolTip, string toolTipText)
{
- this.toolTipCaption = toolTipCaption;
+ if (toolTip == null)
+ {
+ return;
+ }
+
+ this.toolTipCaption = toolTipText;
// native ListView expects tooltip HWND as a wParam and ignores lParam
IntPtr oldHandle = User32.SendMessageW(this, (User32.WM)LVM.SETTOOLTIPS, toolTip.Handle, IntPtr.Zero);
| [ListView->[InsertGroupInListView->[EnsureDefaultGroup],OnSystemColorsChanged->[OnSystemColorsChanged],Clear->[Clear],LargeImageListRecreateHandle->[ForceCheckBoxUpdate],OnFontChanged->[EndUpdate,BeginUpdate,OnFontChanged,InvalidateColumnHeaders],WmNotify->[SetDisplayIndices,OnColumnWidthChanging,CompensateColumnHeaderResize,CheckDisplayIndices,OnColumnClick,ToString,OnColumnWidthChanged,OnColumnReordered,OnDrawColumnHeader],SelectedIndexCollection->[IndexOf->[IndexOf],Clear->[SetItemState,Clear],Add->[SetItemState,Add],Contains->[Contains],Remove->[SetItemState,Remove]],Sort->[ApplyUpdateCachedItems],GenerateRandomName->[Dispose],WmReflectNotify->[CustomDraw,OnBeforeLabelEdit,OnItemChecked,OnRetrieveVirtualItem,OnItemActivate,OnSelectedIndexChanged,WmNmClick,GetIndexOfClickedItem,OnVirtualItemsSelectionRangeChanged,OnAfterLabelEdit,OnItemDrag,WmNmDblClick,OnSearchForVirtualItem,OnItemSelectionChanged,OnCacheVirtualItems,OnItemCheck],OnHandleDestroyed->[OnHandleDestroyed],OnMouseLeave->[OnMouseLeave],SmallImageListRecreateHandle->[ForceCheckBoxUpdate],ToString->[ToString],OnHandleCreated->[InsertGroupNative,InsertItemsNative,InsertColumnNative,ArrangeIcons,OnHandleCreated],ForceCheckBoxUpdate->[ArrangeIcons],ColumnHeaderCollection->[IndexOf->[IndexOf],Clear->[RecreateHandleInternal],Insert->[Insert],Add->[Add],ContainsKey->[IsValidIndex,IndexOfKey],AddRange->[ContainsKey,Add,SetDisplayIndices],IndexOfKey->[IsValidIndex],RemoveAt->[RecreateHandleInternal,SetDisplayIndices],RemoveByKey->[IndexOfKey,RemoveAt,IsValidIndex],Contains->[IndexOf,Contains],Remove->[IndexOf,Remove,RemoveAt],IndexOfKey,IsValidIndex],ListViewItem->[CreateHandle],ListViewItemCollection->[IndexOf->[IndexOf],Clear->[Clear],CopyTo->[CopyTo],ListViewItem->[Insert,Add],Insert->[ToString,Insert],Add->[IndexOf,ToString,Add],ContainsKey->[IsValidIndex,IndexOfKey],AddRange->[CopyTo,AddRange],IndexOfKey->[IsValidIndex],RemoveAt->[RemoveAt],Contains->[Contains],RemoveByKey->[IndexOfKey,RemoveAt,IsValidIndex],ArrayList->[Add],Remove->[Remove],Find->[CopyTo],IndexOfKey,ToString,IsValidIndex],ArrangeIcons->[ArrangeIcons],GetDisplayIndex->[ApplyUpdateCachedItems],RemoveGroupFromListView->[EnsureDefaultGroup,ListView],EndUpdate->[ApplyUpdateCachedItems],SetBackgroundImage->[DeleteFileName,GenerateRandomName],SetColumnInfo->[InvalidateColumnHeaders],GetNativeGroupId->[EnsureDefaultGroup],InsertItems->[GenerateUniqueID,ArrangeIcons],OnResize->[OnResize],ListViewNativeItemCollection->[Clear->[ApplyUpdateCachedItems,ToString,Clear],ListViewItem->[ApplyUpdateCachedItems,UpdateSavedCheckedItems,InsertItems],AddRange->[BeginUpdate,InsertItems,Sort,UpdateSavedCheckedItems,EndUpdate],IEnumerator->[CopyTo],RemoveAt->[Remove,RemoveAt,DisplayIndexToID,ApplyUpdateCachedItems,UpdateSavedCheckedItems],Contains->[ApplyUpdateCachedItems,Contains],Remove->[IndexOf,RemoveAt],ApplyUpdateCachedItems,Insert,RemoveAt,OnRetrieveVirtualItem],OnBackgroundImageChanged->[OnBackgroundImageChanged],OnMouseHover->[OnMouseHover],IsInputKey->[IsInputKey],SetItemText->[SetItemText,SetColumnWidth],OnParentChanged->[OnParentChanged],SetColumnWidth->[CompensateColumnHeaderResize],CheckedListViewItemCollection->[IndexOf->[IndexOf],Contains->[ListView,Contains]],CreateHandle->[CreateHandle],ListViewAccessibleObject->[IsIAccessibleExSupported->[IsIAccessibleExSupported],GetPropertyValue->[GetPropertyValue]],CompensateColumnHeaderResize->[CompensateColumnHeaderResize],InsertItemsNative->[GetNativeGroupId],CheckedIndexCollection->[IndexOf->[IndexOf],Contains->[Contains]],ColumnHeader->[CheckDisplayIndices],UpdateListViewItemsLocations->[BeginUpdate,EndUpdate],IconComparer->[Compare->[Compare]],UpdateColumnWidths->[SetColumnWidth],Dispose->[Clear,Dispose],WndProc->[WmNotify,GetIndexOfClickedItem,OnMouseHover,WmReflectNotify,WmPrint,WmMouseDown,WndProc],SelectedListViewItemCollection->[IndexOf->[IndexOf],ContainsKey->[IsValidIndex,IndexOfKey],Contains->[IndexOf,Contains],IndexOfKey->[IsValidIndex],IndexOfKey,IsValidIndex],ListView]] | Sets the tooltip of the ListView. | It seems that call `SetToolTip()` with `toolTip = null` is not correct and we should throw `ArgumentNullException`. Maybe argument checks can be added to the base implementation and overridden methods should call `base.SetToolTip()` before the remaining implementation. |
@@ -423,9 +423,14 @@ def api(*, since, **kwargs_decorator):
"""Check version requirements and use self.call to do the remote method call.
<since> specifies the version in which borg introduced this method,
- calling this method when connected to an older version will fail without transmiting
+ calling this method when connected to an older version will fail without transmitting
anything to the server.
+ <previously> is the value resulting in the behaviour before introducing the new version.
+
+ <dontcare> is a flag to set the behaviour if an old version is called the new way. If set to True, the method is
+ called as defined by <previously>. If False, an exception is generated.
+
Further kwargs can be used to encode version specific restrictions.
If a previous hardcoded behaviour is parameterized in a version, this allows calls that
use the previously hardcoded behaviour to pass through and generates an error if another
| [handle_remote_line->[write],cache_if_remote->[pack->[pack],unpack->[unpack],RepositoryNoCache,RepositoryCache],RemoteRepository->[get_many->[call_many],call_many->[handle_error->[InvalidRPCMethod,PathNotAllowed,RPCError],UnexpectedRPCDataFormatFromServer,pop_preload_msgid,named_to_positional,handle_error,write,ConnectionClosed],__init__->[do_open->[open],SleepingBandwidthLimiter,do_open,ConnectionClosedWithHint,write],close->[close],async_response->[call_many],RPCError->[exception_full->[get_message]],api],RepositoryCache->[close->[log_instrumentation],add_entry->[key_filename,write,backoff,open],get_many->[add_entry,get_many,key_filename,get,open],backoff->[key_filename,query_size_limit]],SleepingBandwidthLimiter->[write->[write]],RepositoryServer->[inject_exception->[InvalidRPCMethod,PathNotAllowed],open->[PathNotAllowed,_resolve_path],serve->[decode_keys,positional_to_named,UnexpectedRPCDataFormatFromClient,InvalidRPCMethod,os_write,filter_args]],RepositoryNoCache->[__exit__->[close],get->[get_many],get_many->[get_many]]] | Decorator to call a method on the server. decorator for . | ... the new parameter. |
@@ -152,6 +152,9 @@ export class Resource {
/** @const @private {!./resources-impl.Resources} */
this.resources_ = resources;
+ /** @const @private {!./viewport/viewport-impl.Viewport} */
+ this.viewport_ = Services.viewportForDoc(element);
+
/** @const @private {boolean} */
this.isPlaceholder_ = element.hasAttribute('placeholder');
| [No CFG could be retrieved] | Constructor for a resource object. Replies the type of a given map. | This might introduce a race condition. Can we just move this down into the measure method? |
@@ -214,6 +214,11 @@ class RemoteTheme < ActiveRecord::Base
"https://github.com/#{org_repo}"
end
end
+
+ def is_git
+ return true if remote_url.present?
+ false
+ end
end
# == Schema Information
| [RemoteTheme->[update_theme_color_schemes->[normalize_override]]] | Returns the url of the github repo if it exists. | Minor but in ruby we prefer boolean methods to end with `?`. Also the body could be `remote_url.present?` |
@@ -80,14 +80,12 @@ public class SchemaRepoBasedAvroBytesDecoder<SUBJECT, ID> implements AvroBytesDe
DatumReader<GenericRecord> reader = new GenericDatumReader<>(schema);
try (ByteBufferInputStream inputStream = new ByteBufferInputStream(Collections.singletonList(bytes))) {
return reader.read(null, DecoderFactory.get().binaryDecoder(inputStream, null));
- }
- catch (EOFException eof) {
+ } catch (EOFException eof) {
// waiting for avro v1.9.0 (#AVRO-813)
throw new ParseException(
eof, "Avro's unnecessary EOFException, detail: [%s]", "https://issues.apache.org/jira/browse/AVRO-813"
);
- }
- catch (IOException e) {
+ } catch (IOException e) {
throw new ParseException(e, "Fail to decode avro message!");
}
}
| [SchemaRepoBasedAvroBytesDecoder->[hashCode->[hashCode],equals->[equals]]] | Checks if this decoder has a record with the given bytes. | Actually this violate Druid code style. Former version of code is correct |
@@ -0,0 +1,15 @@
+package com.baeldung.supertypecompilerexception;
+
+public class MyClassSolutionI {
+
+ private int myField1 = 10;
+ private int myField2;
+
+ public MyClassSolutionI() {
+ myField2 = myField1;
+ }
+
+ public MyClassSolutionI(int i) {
+ myField2 = i;
+ }
+}
| [No CFG could be retrieved] | No Summary Found. | I think it may be clearer to use `1`, `2` etc, rather than roman numerals. `I`, `II` etc. |
@@ -198,7 +198,7 @@ public class HoodieTableMetaClient implements Serializable {
Properties properties = new Properties();
properties.setProperty(HoodieTableConfig.HOODIE_TABLE_NAME_PROP_NAME, tableName);
properties.setProperty(HoodieTableConfig.HOODIE_TABLE_TYPE_PROP_NAME, tableType.name());
- if (tableType == HoodieTableType.MERGE_ON_READ) {
+ if (tableType == MERGE_ON_READ) {
properties.setProperty(HoodieTableConfig.HOODIE_PAYLOAD_CLASS_PROP_NAME, payloadClassName);
}
return HoodieTableMetaClient.initializePathAsHoodieDataset(hadoopConf, basePath, properties);
| [HoodieTableMetaClient->[initializePathAsHoodieDataset->[getFs,HoodieTableMetaClient,getTableType],equals->[equals],getFs->[getFs],getArchivePath->[getMetaPath],toString->[toString]]] | Initialize Hoodie table type. | should we persist the index type into table properties.. I think it might be good to do that? |
@@ -305,12 +305,14 @@ export class AmpA4A extends AMP.BaseElement {
this.isRelayoutNeededFlag = false;
/**
- * Used as a signal in some of the CSI pings. Canary is shortened to 'ca'
- * and production to 'pr'. TODO(@glevitzky, #10060) Include other release
- * types (control) when available.
+ * Used as a signal in some of the CSI pings.
* @private @const {string}
*/
- this.releaseType_ = isCanary(this.win) ? 'ca' : 'pr';
+ this.releaseType_ = {
+ 'production': 'pr',
+ 'control': 'co',
+ 'canary': 'ca',
+ }[getBinaryType(this.win)] || 'un';
}
/** @override */
| [AmpA4A->[extractSize->[user,get,Number],constructor->[AmpAdUIHandler,AmpAdXOriginIframeHandler,isCanary,dev,generateSentinel,platformFor,now,SAFEFRAME],tearDownSlot->[dev,platformFor,SAFEFRAME],renderViaNameAttrOfXOriginIframe_->[NAMEFRAME,utf8Decode,dev,checkStillCurrent,user,getContextMetadata,dict,getDefaultBootstrapBaseUrl,stringify,SAFEFRAME,reject,length],preconnectCallback->[forEach,getDefaultBootstrapBaseUrl],attemptToRenderCreative->[round,resolve,dev,checkStillCurrent,user,cancellation],renderNonAmpCreative_->[NAMEFRAME,resolve,incrementLoadingAds,user,assertHttpsUrl,SAFEFRAME],viewportCallback->[setFriendlyIframeEmbedVisible],refresh->[dev,refreshEndCallback,timerFor,resourcesForDoc],buildCallback->[loadKeyset,round,signatureVerifierFor,AmpAdUIHandler,dev,protectFunctionWrapper,viewerForDoc],renderAmpCreative_->[round,setFriendlyIframeEmbedVisible,dev,whenIniLoaded,createElementWithAttributes,installFriendlyIframeEmbed,dict,body,protectFunctionWrapper,minifiedCreative,win,setStyle,customStylesheets,iframe,push,customElementExtensions,checkStillCurrent,installAnchorClickInterceptor,getTimingDataAsync,installUrlReplacementsForEmbed],getAmpAdMetadata_->[customElementExtensions,isArray,dev,customStylesheets,slice,lastIndexOf,minifiedCreative,parseJson,length,isObject],onCrossDomainIframeCreated->[dev],verifyStillCurrent->[cancellation],sendXhrRequest->[xhrFor,dev,adUrl,frameGetDisabled],renderOutsideViewport->[is3pThrottled,getAmpAdRenderOutsideViewport],isLayoutSupported->[isLayoutSizeDefined],forceCollapse->[dev],renderViaCachedContentIframe_->[dict,stringify,getContextMetadata,xhrFor],registerAlpHandler_->[viewerForDoc,document,handleClick,isExperimentOn],iframeRenderHelper_->[AmpAdXOriginIframeHandler,dict,dev,protectFunctionWrapper,assign,createElementWithAttributes],shouldInitializePromiseChain_->[height,dev,user,isAdPositionAllowed,cryptoFor,width],getSigningServiceNames->[getMode],initiateAdRequest->[signatureVerifierFor,status,dev,getMode,isEnumValue,headers,ERROR_SIGNATURE_MISMATCH,reject,CLIENT_CACHE,arrayBuffer,utf8Decode,byteLength,user,extensionsFor,UNVERIFIED,OK,customElementExtensions,resolve,checkStillCurrent,ERROR_KEY_NOT_FOUND,preloadExtension,viewerForDoc],promiseErrorHandler_->[isCancellation,duplicateErrorIfNecessary,user,assignAdUrlToError,dev,ignoreStack,message,getMode,random],resumeCallback->[isMeasureRequested,hasBeenMeasured],BaseElement],dict,indexOf,apply,unshift,args,substring] | Initializes a new keyset. Protected version of protected methods. Priority used for scheduling preload and layout callback. | Do we really need two different mappings for the same thing? |
@@ -26,8 +26,8 @@ from tensorflow.python.util.tf_export import tf_export
# pylint: disable=protected-access
@tf_export('roll', v1=['roll', 'manip.roll'])
@deprecation.deprecated_endpoints('manip.roll')
-def roll(input, shift, axis): # pylint: disable=redefined-builtin
- return _gen_manip_ops.roll(input, shift, axis)
+def roll(input, shift, axis, name=None): # pylint: disable=redefined-builtin
+ return _gen_manip_ops.roll(input, shift, axis, name)
roll.__doc__ = _gen_manip_ops.roll.__doc__
| [roll->[roll],tf_export,deprecated_endpoints] | Roll an array along a given axis. | The gen_manip_ops.roll should already have a name argument you can just use it. |
@@ -106,7 +106,7 @@ func resourceComputeSharedVpcUpdate(d *schema.ResourceData, meta interface{}) er
for project, _ := range oldMap {
if _, ok := newMap[project]; !ok {
// The project is in the old config but not the new one, disable it
- if err := disableResource(config, hostProject, project); err != nil {
+ if err := disableXpnResource(config, hostProject, project); err != nil {
return fmt.Errorf("Error disabling Shared VPC service project %q: %s", project, err)
}
}
| [EnableXpnResource,Printf,EnableXpnHost,GetChange,DisableXpnResource,GetOk,Do,Sprintf,Background,HasChange,List,Pages,Errorf,SetId,Get,DisableXpnHost,Set,GetXpnResources] | resourceComputeSharedVpcUpdate - updates the Shared VPC service projects for a given host project. disableXpnHost disables Shared VPC Resource for all XPN host projects. | Renamed `disableResource` to `disableXpnResource`. |
@@ -141,7 +141,12 @@ namespace NServiceBus.Hosting.Helpers
try
{
+#if NETCOREAPP2_1
+ var context = AssemblyLoadContext.GetLoadContext(Assembly.GetExecutingAssembly());
+ assembly = context.LoadFromAssemblyPath(assemblyPath);
+#else
assembly = Assembly.LoadFrom(assemblyPath);
+#endif
return true;
}
| [AssemblyScanner->[ScanAssembly->[ScanAssembly],AddTypesToResult->[FormatReflectionTypeLoadException],ShouldScanDependencies->[IsExcluded]]] | TryLoadScannableAssembly tries to load an assembly from the given path. If the. | at least for functions this seems the way to go to ensure that everything is loaded into the function execution specific context. I'm mostly wondering whether there is a case where this behavior would be undesired and this behavior would be breaking (so it should be made opt-in instead), but I can't really come up with a specific scenario though. |
@@ -111,12 +111,14 @@ public class PutKinesisStream extends AbstractKinesisStreamProcessor {
for (int i = 0; i < flowFiles.size(); i++) {
FlowFile flowFile = flowFiles.get(i);
+ String streamName = context.getProperty(KINESIS_STREAM_NAME).evaluateAttributeExpressions(flowFile).getValue();;
+
final ByteArrayOutputStream baos = new ByteArrayOutputStream();
session.exportTo(flowFile, baos);
PutRecordsRequestEntry record = new PutRecordsRequestEntry().withData(ByteBuffer.wrap(baos.toByteArray()));
String partitionKey = context.getProperty(PutKinesisStream.KINESIS_PARTITION_KEY)
- .evaluateAttributeExpressions(flowFiles.get(i)).getValue();
+ .evaluateAttributeExpressions(flowFiles.get(i)).getValue();
if ( ! StringUtils.isBlank(partitionKey) ) {
record.setPartitionKey(partitionKey);
| [PutKinesisStream->[onTrigger->[getValue,longValue,size,setStreamName,ByteArrayOutputStream,toString,withData,PutRecordsRequest,isBlank,put,getErrorMessage,setRecords,yield,putAllAttributes,exportTo,getClient,setPartitionKey,nextInt,getRecords,getErrorCode,transfer,error,debug,getShardId,asInteger,getSequenceNumber,putRecords,filterMessagesByMaxSize,wrap,toByteArray,get,add,clear],build,unmodifiableList,asList,Random]] | Trigger the event. private HashMap attributes = new HashMap();. | Would you also please fix these `if`s? I completely understand they were already that way. |
@@ -223,10 +223,13 @@ class BaseLower(object):
self.builder.branch(self.blkmap[self.firstblk])
if config.DUMP_LLVM:
- print(("LLVM DUMP %s" % self.fndesc).center(80,'-'))
+ print(("LLVM DUMP %s" % self.fndesc).center(80, '-'))
print(self.module)
print('=' * 80)
self.module.verify()
+ # Run function-level optimize to reduce memory usage and improve
+ # module-level optimization
+ self.context.optimize_function(self.function)
def init_argument(self, arg):
return arg
| [PyLower->[lower_expr->[storevar,loadvar],getvar->[alloca],lower_inst->[lower_assign,storevar,loadvar],is_null->[is_null],incref->[incref],alloca->[alloca],storevar->[getvar],lower_const->[lower_const],decref->[decref],loadvar->[getvar],lower_assign->[lower_expr,loadvar],lower_global->[ForbiddenConstruct]],BaseLower->[lower_block->[LoweringError],lower->[post_lower]],Lower->[getvar->[typeof],lower_inst->[typeof],alloca->[alloca],storevar->[getvar],loadvar->[getvar],lower_assign->[typeof],lower_expr->[typeof,ExternalFunctionDescriptor]],FunctionDescriptor->[_from_python_function->[_get_function_info]],PythonFunctionDescriptor->[from_object_mode_function->[_from_python_function],from_specialized_function->[_from_python_function]]] | Lower all functions in the function. | If we are debugging a LLVM dump, do we want to dump before or after the function optimization step? |
@@ -230,10 +230,9 @@ stc_diff = stc_aud - stc_vis
brain = stc_diff.plot(subjects_dir=subjects_dir, initial_time=0.1,
hemi='split', views=['lat', 'med'])
+
###############################################################################
# References
# ----------
-# .. [1] Destrieux C, Fischl B, Dale A, Halgren E (2010). Automatic
-# parcellation of human cortical gyri and sulci using standard
-# anatomical nomenclature, vol. 53(1), 1-15, NeuroImage.
+# .. footbibliography::
\ No newline at end of file
| [data_fun->[sin,rand,exp],read_events,read_cov,where,join,make_inverse_operator,enumerate,add_ecg,read_forward_solution,add_eog,read_info,epochs,set,get_stc,range,simulate_raw,values,read_labels_from_annot,list,add_data,plot,print,data_path,sorted,RandomState,data_fun,SourceSimulator,split,keys,arange,set_eeg_reference,apply_inverse,Epochs,add_noise] | Plots a single using the standard anatomical magnetic cortical gyri. | missing newline at end of file. Our style checker will complain about this. |
@@ -79,9 +79,11 @@ bool operator==(const TimeStamp& lhs, const TimeStamp& rhs)
std::ostream&
operator<<(std::ostream& out, const TimeStamp& ts)
{
- std::streamsize ssize = out.precision();
+ const std::ios_base::fmtflags out_flags(out.flags());
+ const std::streamsize ssize = out.precision();
out << std::setprecision(3) << std::fixed <<
static_cast<double>(ts.sec) + (static_cast<double>(ts.nsec) / 1.0e9) << std::setprecision(ssize) << std::flush;
+ out.flags(out_flags);
return out;
}
| [get_or_create_property->[c_str,seq,str,PropertyIndex,length],operator<<->[precision],get_property->[length,ConstPropertyIndex,seq],get_time->[time_since_epoch,,count]] | << - - - - - - - - - - - - - - - - -. | It'd be nice if there was an RAII way of saving and restoring the format state of the `ostream`. |
@@ -28,7 +28,7 @@ namespace Microsoft.Xna.Framework.Graphics
private bool _shared;
private bool _mipmap;
- protected SampleDescription _sampleDescription;
+ private SampleDescription _sampleDescription;
private SharpDX.Direct3D11.Texture2D _cachedStagingTexture;
| [Texture2D->[Dispose->[Dispose],PlatformSetData->[Alloc,Left,Top,Bottom,Front,Free,Right,IsCompressedFormat,Width,CalculateSubresourceIndex,Back,_d3dContext,GetSizeForLevel,Get,GetTexture,GetPitch,ToInt64,Pinned,UpdateSubresource],PlatformGetData->[Seek,MipLevels,Left,Top,stream,Bottom,_d3dDevice,SampleDescription,ToFormat,CpuAccessFlags,Right,Max,Format,GetSize,RowPitch,Usage,ReadRange,IsCompressedFormat,ArraySize,Width,Height,Current,BindFlags,None,CalculateSubresourceIndex,UnmapSubresource,MapSubresource,Dispose,_d3dContext,CopySubresourceRegion,Read,Get,Staging,GetTexture,Begin,OptionFlags],Texture2DDescription->[MipLevels,Height,Format,BindFlags,None,OptionFlags,ShaderResource,Usage,Shared,Default,SampleDescription,ArraySize,ToFormat,CpuAccessFlags,Width],CreateTexture->[_d3dDevice,GetTexture2DDescription]]] | PlatformConstruct - Constructs the object from the parameters. | make _sampleDescription private. This field was shadowed by RenderTarget2D..DirectX.cs |
@@ -247,6 +247,8 @@ int ec_GFp_mont_group_set_curve(EC_GROUP *group, const BIGNUM *p,
BN_CTX_free(new_ctx);
if (mont != NULL)
BN_MONT_CTX_free(mont);
+ if (one != NULL)
+ BN_free(one);
return ret;
}
| [ec_GFp_mont_field_encode->[BN_to_montgomery,ECerr],ec_GFp_mont_group_set_curve->[BN_MONT_CTX_free,ec_GFp_simple_group_set_curve,BN_CTX_free,BN_to_montgomery,BN_MONT_CTX_new,BN_CTX_new,BN_value_one,BN_MONT_CTX_set,BN_free,ECerr,BN_new],ec_GFp_mont_field_mul->[ECerr,BN_mod_mul_montgomery],ec_GFp_mont_group_clear_finish->[BN_MONT_CTX_free,ec_GFp_simple_group_clear_finish,BN_clear_free],ec_GFp_mont_group_finish->[ec_GFp_simple_group_finish,BN_MONT_CTX_free,BN_free],ec_GFp_mont_field_decode->[BN_from_montgomery,ECerr],ec_GFp_mont_field_sqr->[ECerr,BN_mod_mul_montgomery],ec_GFp_mont_group_copy->[ec_GFp_simple_group_copy,BN_MONT_CTX_free,BN_MONT_CTX_copy,BN_clear_free,BN_MONT_CTX_new,BN_dup],ec_GFp_mont_field_set_to_one->[BN_copy,ECerr],ec_GFp_mont_group_init->[ec_GFp_simple_group_init],EC_METHOD->[FIPS_mode,fips_ec_gfp_mont_method]] | EC_GFp_montgomery_set_curve - Set the EC_GFp_montgomery - Montgomery multiplication. | FYI, we did change the XXX_free routines to ignore NULL in 1.0.2 to make cherry-picking fixes easier. So this if isn't needed. But the one above has it for "mont" so I guess parallel construction is better. |
@@ -267,6 +267,7 @@ public class UntilSuccessful extends AbstractOutboundRouter implements UntilSucc
public Processor getRoute() {
final DefaultMessageProcessorChain chain = DefaultMessageProcessorChain.from(muleContext, routes.get(0));
chain.setTemplateMuleContext(muleContext);
+ chain.setFlowConstruct(flowConstruct);
return chain;
}
| [UntilSuccessful->[initialise->[initialise],start->[start],stop->[stop],route->[route]]] | Get the next route. | This shouldn't be required. |
@@ -752,6 +752,11 @@ public class PutDatabaseRecord extends AbstractSessionFactoryProcessor {
if (DELETE_TYPE.equalsIgnoreCase(statementType)) {
ps.setObject(i * 2 + 1, currentValue, sqlType);
ps.setObject(i * 2 + 2, currentValue, sqlType);
+ } else if (UPSERT_TYPE.equalsIgnoreCase(statementType)) {
+ final int timesToAddObjects = databaseAdapter.getTimesToAddColumnObjectsForUpsert();
+ for (int j = 0; j < timesToAddObjects; j++) {
+ ps.setObject(i + (fieldIndexes.size() * j) + 1, currentValue, sqlType);
+ }
} else {
ps.setObject(i + 1, currentValue, sqlType);
}
| [PutDatabaseRecord->[checkValuesForRequiredColumns->[getNormalizedColumnNames],normalizeKeyColumnNamesAndCheckForValues->[getNormalizedColumnNames],customValidate->[customValidate],TableSchema->[from->[from,getColumns,normalizeColumnName,TableSchema],normalizeColumnName],SchemaKey->[hashCode->[hashCode],equals->[equals]],generateDelete->[getNormalizedColumnNames],ColumnDescription->[from->[getColumnName,ColumnDescription]],onTrigger->[FunctionContext,onTrigger]]] | Executes a DML statement. generate update or upsert. if there s no index map execute the query and add it to the index map. | `fieldIndexes` in this place is NULL |
@@ -883,11 +883,9 @@ public class UnitAttachment extends DefaultAttachment {
}
public boolean canInvadeFrom(final Unit transport) {
- if (m_canInvadeOnlyFrom == null || Arrays.asList(m_canInvadeOnlyFrom).isEmpty() || m_canInvadeOnlyFrom[0].isEmpty()
- || m_canInvadeOnlyFrom[0].equals("all")) {
- return true;
- }
- return Arrays.asList(m_canInvadeOnlyFrom).contains(transport.getType().getName());
+ return m_canInvadeOnlyFrom == null || Arrays.asList(m_canInvadeOnlyFrom).isEmpty() || m_canInvadeOnlyFrom[0]
+ .isEmpty() || m_canInvadeOnlyFrom[0].equals("all") || Arrays.asList(m_canInvadeOnlyFrom)
+ .contains(transport.getType().getName());
}
private void resetCanInvadeOnlyFrom() {
| [UnitAttachment->[getAllowedBombingTargetsIntersection->[get,getBombingTargets],getAllOfTypeAas->[getTypeAa],setIsTwoHit->[setIsTwoHit],getMaximumNumberOfThisUnitTypeToReachStackingLimit->[getPlacementLimit,getAttackingLimit,getMovementLimit,get,getIsAaForCombatOnly,getIsAaForBombingThisUnitOnly],getUnitsWhichReceivesAbilityWhenWith->[getReceivesAbilityWhenWithMap],setIsAaMovement->[setIsAaMovement],get->[get],getReceivesAbilityWhenWithMap->[getReceivesAbilityWhenWith,getUnitTypesFromUnitList],toStringShortAndOnlyImportantDifferences->[getCreatesResourcesList,getFuelFlatCost,getIsStrategicBomber,getCanOnlyBePlacedInTerritoryValuedAtX,playerHasMechInf,getAirAttack,getBombingMaxDieSides,getCreatesUnitsList,getAttackAa,getIsMarine,getMaxBuiltPerPlayer,get,getRequiresUnitsToMove,getIsAir,getArtillerySupportable,getIsRocket,getCanBeDamaged,getMaxAaAttacks,getBombard,getIsSea,getIsSuicide,getIsInfantry,getCarrierCost,getIsCombatTransport,getReceivesAbilityWhenWith,getCanAirBattle,getAirDefense,getAttackRolls,getTransportCost,getPlacementLimit,getRequiresUnits,getIsAaForFlyOverOnly,getIsSuicideOnHit,getDefense,getIsDestroyer,getTransportCapacity,getIsSub,getBlockade,getIsInfrastructure,getRepairsUnits,getIsLandTransport,getAttack,getOffensiveAttackAa,playerHasRockets,getBombingBonus,getMaxDamage,getFuelCost,getCanBlitz,getCarrierCapacity,getHitPoints,getIsAirBase,getIsAaForCombatOnly,getIsKamikaze,getIsAirTransport,getConsumesUnits,getIsAirTransportable,getMovement,getCanProduceXUnits,playerHasParatroopers,getArtillery,getCanProduceUnits,getMaxOperationalDamage,getIsLandTransportable,getAttackingLimit,getUnitPlacementRestrictions,getCanNotMoveDuringCombatMove,getCanDieFromReachingMaxDamage,getGivesMovement,getCanScramble,getMovementLimit,getIsConstruction,getCanIntercept,getIsAaForBombingThisUnitOnly,getDefenseRolls,getCanBombard,getCanEscort],setIsFactory->[setIsFactory],setIsAa->[setIsAa,setIsInfrastructure],setDestroyedWhenCapturedFrom->[setDestroyedWhenCapturedBy]]] | Checks if the given transport is in the list of can invade - only - from units. | formatting ,please put each or condition on its own line. |
@@ -402,6 +402,10 @@ def disable(request, addon_id, addon):
@post_required
def unlist(request, addon_id, addon):
addon.update(is_listed=False, disabled_by_user=False)
+ # In https://github.com/mozilla/addons-server/issues/3471 this view will
+ # no longer be global, but in the meantime we need to set the channel
+ # property on all versions to stay consistent.
+ addon.versions.update(channel=amo.RELEASE_CHANNEL_UNLISTED)
amo.log(amo.LOG.ADDON_UNLISTED, addon)
if addon.latest_version.is_unreviewed:
| [upload->[handle_upload],upload_image->[ajax_upload_image],upload_detail->[_compat_result,json_upload_detail],_resume->[_step_url],auto_sign_version->[auto_sign_file],submit_media->[_step_url],upload_for_addon->[upload],version_add_file->[delete,auto_sign_file,check_validation_override],submit->[_step_url],submit_bump->[_step_url],ajax_dependencies->[AddonDependencySearch],remove_locale->[remove_locale],dashboard->[addon_listing],delete->[delete],version_add->[auto_sign_version,check_validation_override],submit_addon->[_step_url,auto_sign_version,check_validation_override],version_delete->[delete],ownership->[delete,mail_user_changes],version_edit->[delete],submit_describe->[delete,_step_url],feed->[_get_addons,_get_activities,_get_rss_feed,_get_items],submit_step] | Unlist an add - on. | Does `.versions` include deleted versions? |
@@ -350,6 +350,7 @@ class CheckoutLinesAdd(BaseMutation):
)
variant_ids = [line.get("variant_id") for line in lines]
+ info.context.channel_slug = checkout.channel.slug
variants = cls.get_nodes_or_error(variant_ids, "variant_id", ProductVariant)
quantities = [line.get("quantity") for line in lines]
| [CheckoutBillingAddressUpdate->[perform_mutation->[CheckoutBillingAddressUpdate,save]],CheckoutCreate->[process_checkout_lines->[check_lines_quantity],clean_input->[retrieve_billing_address,retrieve_shipping_address,process_checkout_lines],Arguments->[CheckoutCreateInput],perform_mutation->[save,CheckoutCreate,clean_input],save->[save,save_addresses]],CheckoutShippingAddressUpdate->[process_checkout_lines->[check_lines_quantity],perform_mutation->[update_checkout_shipping_method_if_invalid,save,process_checkout_lines,CheckoutShippingAddressUpdate]],update_checkout_shipping_method_if_invalid->[clean_shipping_method],CheckoutLineDelete->[perform_mutation->[update_checkout_shipping_method_if_invalid,CheckoutLineDelete]],CheckoutShippingMethodUpdate->[perform_mutation->[CheckoutShippingMethodUpdate,save,clean_shipping_method]],CheckoutRemovePromoCode->[perform_mutation->[CheckoutRemovePromoCode]],CheckoutAddPromoCode->[perform_mutation->[CheckoutAddPromoCode]],CheckoutEmailUpdate->[perform_mutation->[save,CheckoutEmailUpdate]],CheckoutLinesAdd->[perform_mutation->[CheckoutLinesAdd,update_checkout_shipping_method_if_invalid,check_lines_quantity]],CheckoutComplete->[perform_mutation->[CheckoutComplete]],CheckoutCustomerAttach->[perform_mutation->[save,CheckoutCustomerAttach]],CheckoutCustomerDetach->[perform_mutation->[save,CheckoutCustomerDetach]]] | Perform a single mutation of the n - node checkout. | Why do we need this? This value comes from middleware and IMO it should be always set by the middleware, otherwise we'll end up with magic which is really difficult to debug. |
@@ -47,10 +47,15 @@ export default class WalletTransactions extends Component {
this.props.navigation.setParams({ wallet: wallet });
this.state = {
isLoading: true,
- isTransactionsLoading: false,
+ showShowFlatListRefreshControl: false,
wallet: wallet,
dataSource: wallet.getTransactions(),
walletPreviousPreferredUnit: wallet.getPreferredBalanceUnit(),
+ walletHeaderLatestTransaction: '...',
+ showSendButton:
+ (wallet.allowSend() && wallet.type === LightningCustodianWallet.type && wallet.balance > 0) ||
+ (wallet.allowSend() && wallet.type !== LightningCustodianWallet.type),
+ showReceiveButton: wallet.allowReceive(),
};
}
| [No CFG could be retrieved] | The WalletTransactions component that displays the transactions of a wallet. This function is used to determine if the wallet should be displayed or not. | I dont get this condition. Is there an error here..? |
@@ -1306,9 +1306,11 @@ class BonPrelevement extends CommonObject
// SEPA Initialisation
$CrLf = "\n";
$date_actu = dol_now();
+ $dateTime_ECMA = dol_print_date($date_actu, '%Y-%m-%dT%H:%M:%S');
+
+ if(!empty($executiondate)) $date_actu=$executiondate;
$dateTime_YMD = dol_print_date($date_actu, '%Y%m%d');
$dateTime_YMDHMS = dol_print_date($date_actu, '%Y%m%d%H%M%S');
- $dateTime_ECMA = dol_print_date($date_actu, '%Y-%m-%dT%H:%M:%S');
$fileDebiteurSection = '';
$fileEmetteurSection = '';
$i = 0;
| [BonPrelevement->[set_infocredit->[fetch],Create->[fetch,AddFacture],set_credite->[fetch],EnregDestinataireSEPA->[buildRumNumber],EnregEmetteurSEPA->[fetch],AddNotification->[DeleteNotification]]] | generate - generate a build file for European countries This method is used to get all the data for all prelevency lignes Diese Methode d ajout d un nuevo Sepa. Diagnose fichier de la fichier. This method is used to generate a withdrawal file for all withdrawal prele check if the file is writeable and set the main umask. | Are you sure the if(!empty($executiondate)) $date_actu=$executiondate; is after $dateTime_ECMA = dol_print_date($date_actu, '%Y-%m-%dT%H:%M:%S'); and not before ? |
@@ -316,6 +316,17 @@ func MustMakeDuration(d time.Duration) Duration {
return rv
}
+func MustMakeDurationPtr(d *time.Duration) *Duration {
+ if d == nil {
+ return nil
+ }
+ rv, err := MakeDuration(*d)
+ if err != nil {
+ panic(err)
+ }
+ return &rv
+}
+
// Duration returns the value as the standard time.Duration value.
func (d Duration) Duration() time.Duration {
return d.d
| [AsMap->[Value],Scan->[Duration],MarshalText->[String,Duration],Delete->[Delete],IsZero->[Duration],PrependAtArrayKey->[Value,Add],MarshalJSON->[String],Before->[Add,Duration],CBOR->[Value],MultiAdd->[AsMap],Duration->[Duration],Value->[ToStrings,String,Duration],String->[String,Duration],Value] | Duration returns the duration of the duration. | This is a bit weird, why would I ever call a `Must` function with `nil` ? |
@@ -627,8 +627,6 @@ process_setup_request(Drpc__Call *drpc_req, Drpc__Response *drpc_resp)
/* Response status is populated with SUCCESS on init. */
mgmt__daos_resp__init(resp);
- D_DEBUG(DB_MGMT, "Received request to set up server\n");
-
dss_init_state_set(DSS_INIT_STATE_SET_UP);
pack_daos_response(resp, drpc_resp);
| [No CFG could be retrieved] | This function processes the response of a specific type of pool request. This function handles all of the request related to the DRPC method. | Don't want to keep this one at all? |
@@ -167,7 +167,10 @@ class IdeaPluginGen(ConsoleTask):
generator.write(output)
return output.name
- def execute(self):
+ def console_output(self, _targets):
+ if not self.context.target_roots:
+ raise TaskError("No targets specified.")
+
# Heuristics to guess whether user tries to load a python project,
# in which case intellij project sdk has to be set up manually.
jvm_target_num = len([x for x in self.context.target_roots if isinstance(x, JvmTarget)])
| [IdeaPluginGen->[execute->[generate_project],generate_project->[gen_file]]] | Applies the specified generator to a temp file and returns the path to that file. | I'm not sure how to handle this `targets` argument to `console_output`. Originally, the task was using `self.context.target_roots` and `self.context.options.target_specs`. Both of these things are not the same as the `self.context.targets` that `ConsoleTask` passes on in its call to `console_output()`. Is it fine as is or is there a way to make use of the `targets` param? |
@@ -236,18 +236,6 @@ public class MessagePropertiesContext implements Serializable
}
getScopedProperties(scope).put(key, value);
- keySet.add(key);
- }
-
- /**
- * @deprecated use {@link #getPropertyNames(org.mule.api.transport.PropertyScope)}
- */
- @Deprecated
- public Set<String> getPropertyNames()
- {
- Set<String> allProps = new HashSet<String>();
- allProps.addAll(keySet);
- return allProps;
}
/**
| [MessagePropertiesContext->[getStringProperty->[getStringProperty,getProperty],getDoubleProperty->[getProperty],clearProperties->[clearProperties,getScopedProperties],getProperty->[getProperty],getFloatProperty->[getProperty],getIntProperty->[getProperty],getShortProperty->[getProperty],removeProperty->[getProperty,removeProperty],getLongProperty->[getProperty],toString->[toString],getBooleanProperty->[getProperty],getByteProperty->[getProperty]]] | Sets a property in this session. | If this was already tagged as deprecated instead of being deleted, why are you deleting it right now? Make it expensive by calculating the set here, and ensure Mule doesn't call this method, but don't remove it if it was previously decided to be kept |
@@ -45,8 +45,13 @@ final class ConnectionProviderSchemaDelegate
this.builder = builder;
}
- public void registerConnectionProviderElement(Schema schema, ConnectionProviderModel providerModel)
+ public void registerConnectionProviderElement(ConnectionProviderModel providerModel)
{
+ if (trackElement(providerModel.getName()))
+ {
+ return;
+ }
+
Element providerElement = new TopLevelElement();
providerElement.setName(providerModel.getName());
providerElement.setSubstitutionGroup(MULE_EXTENSION_CONNECTION_PROVIDER_ELEMENT);
| [ConnectionProviderSchemaDelegate->[addValidationFlag->[add,createAttribute,load],registerConnectionProviderElement->[setBase,ComplexContent,registerParameters,setComplexType,ExplicitGroup,setName,ExtensionType,addValidationFlag,TopLevelElement,addRetryPolicy,setMaxOccurs,setMinOccurs,setSubstitutionGroup,setComplexContent,addConnectionProviderPoolingProfile,getParameterModels,setExtension,isPooled,ifPresent,getName,isCached,add,LocalComplexType],addConnectionProviderPoolingProfile->[add,getPoolingSupport,createElement,createRefElement],createTypeLoader,ObjectFactory]] | Registers a connection provider element. | Should these changes also be in ConfigurationSchemaDelegate? |
@@ -550,6 +550,13 @@ public class KsqlConfig extends AbstractConfig {
+ " runtimes, if true, new queries may share streams instances.";
+ public static final String KSQL_SHARED_RUNTIMES_COUNT = "ksql.shared.runtimes.count";
+ public static final Integer KSQL_SHARED_RUNTIMES_COUNT_DEFAULT = 8;
+ public static final String KSQL_SHARED_RUNTIMES_COUNT_DOC =
+ "Controls how many runtimes queries are allocated over initially."
+ + "this is only used when ksql.runtime.feature.shared.enabled is true.";
+
+
public static final String KSQL_SUPPRESS_BUFFER_SIZE_BYTES = "ksql.suppress.buffer.size.bytes";
public static final Long KSQL_SUPPRESS_BUFFER_SIZE_BYTES_DEFAULT = -1L;
public static final String KSQL_SUPPRESS_BUFFER_SIZE_BYTES_DOC =
| [KsqlConfig->[ConfigValue->[isResolved->[isResolved]],buildConfigDef->[defineLegacy,defineCurrent],getKsqlStreamConfigProps->[getKsqlStreamConfigProps],buildStreamingConfig->[applyStreamsConfig],streamTopicConfigNames->[getName],getAllConfigPropsWithSecretsObfuscated->[getKsqlConfigPropsWithSecretsObfuscated],overrideBreakingConfigsWithOriginalValues->[KsqlConfig],empty->[KsqlConfig],cloneWithPropertyOverwrite->[KsqlConfig,getKsqlStreamConfigProps,buildStreamingConfig],getKsqlStreamConfigPropsWithSecretsObfuscated->[convertToObfuscatedString],resolveStreamsConfig->[ConfigValue,empty],CompatibilityBreakingConfigDef->[defineCurrent->[define],defineLegacy->[define],define->[define]],configDef,buildStreamingConfig]] | The feature flag that enables table materialization on source tables. Default is true. This method is used to set the maximum amount of time to wait before attempting to retry a. | nit: include the word "initial" or similar, eg `num.initial.shared.runtimes` |
@@ -3452,7 +3452,7 @@ long ssl3_ctrl(SSL *s, int cmd, long larg, void *parg)
case SSL_CTRL_GET_FLAGS:
ret = (int)(s->s3.flags);
break;
-#if !defined(OPENSSL_NO_DH) && !defined(OPENSSL_NO_DEPRECATED_3_0)
+#if !defined(OPENSSL_NO_DEPRECATED_3_0)
case SSL_CTRL_SET_TMP_DH:
{
EVP_PKEY *pkdh = NULL;
| [No CFG could be retrieved] | This function is called by the server when a new certificate is requested. This function is called by the server when the server receives a DH key. It is called. | I wonder if this particular change makes sense as the CTRL requires the DH objects to be built in libcrypto to be functional. |
@@ -879,6 +879,10 @@ void writePlayerPos(LocalPlayer *myplayer, ClientMap *clientMap, NetworkPacket *
u8 wanted_range = MYMIN(255,
std::ceil(clientMap->getControl().wanted_range / MAP_BLOCKSIZE));
+ if (clientMap->getCameraFrontView()) {
+ yaw += 180 * 100;
+ pitch = -pitch;
+ }
v3s32 position(pf.X, pf.Y, pf.Z);
v3s32 speed(sf.X, sf.Y, sf.Z);
| [No CFG could be retrieved] | Handle a specific packet. u8 action u8 item u8 position u8 pitch yaw yaw. | can you add a define for 180 to define what is this and use it in the multiple places ? |
@@ -103,7 +103,7 @@ export class DomWriterStreamer {
/** @private {boolean} */
this.eof_ = false;
- /** @private {?Element} */
+ /** @private {?Promise<!Element>} */
this.targetBody_ = null;
}
| [No CFG could be retrieved] | A utility class for writing a DOM node with a single . This method is used to write a chunk of data to the underlying parser. It is called. | Shouldn't this be something like `{!Promise<?Element>}`? Right now the element is optional but it appears you are making it required. Or are you only resolving the promise if there is an element? |
@@ -329,13 +329,13 @@ class Experiment:
return [TrialResult(**trial_result) for trial_result in resp]
def _get_query_type(self, key: str):
- if key == 'trial_concurrency':
+ if key == 'trialConcurrency':
return '?update_type=TRIAL_CONCURRENCY'
- if key == 'max_experiment_duration':
+ if key == 'maxExecDuration':
return '?update_type=MAX_EXEC_DURATION'
- if key == 'search_space':
+ if key == 'searchSpace':
return '?update_type=SEARCH_SPACE'
- if key == 'max_trial_number':
+ if key == 'maxTrialNum':
return '?update_type=MAX_TRIAL_NUM'
def _update_experiment_profile(self, key: str, value: Any):
| [Experiment->[get_trial_job->[_experiment_rest_get],get_job_statistics->[_experiment_rest_get],update_max_trial_number->[_update_experiment_profile],run->[start,stop],update_max_experiment_duration->[_update_experiment_profile],update_search_space->[_update_experiment_profile],start->[start],export_data->[_experiment_rest_get],get_job_metrics->[_experiment_rest_get],update_trial_concurrency->[_update_experiment_profile],get_status->[_experiment_rest_get],get_experiment_profile->[_experiment_rest_get],_update_experiment_profile->[_experiment_rest_put,get_experiment_profile,_get_query_type],list_trial_jobs->[_experiment_rest_get]]] | Get the query type for a given key. | so the previous version is a bug? |
@@ -87,12 +87,11 @@ class CI_Hooks {
*/
public function __construct()
{
- $CFG =& load_class('Config', 'core');
log_message('info', 'Hooks Class Initialized');
// If hooks are not enabled in the config file
// there is nothing else to do
- if ($CFG->item('enable_hooks') === FALSE)
+ if (config_item('enable_hooks') === FALSE)
{
return;
}
| [CI_Hooks->[call_hook->[_run_hook],__construct->[item]]] | Initialize the hooks class. | You've said this in #5323: > *note: using the config_item('enable_hooks') makes more sense than calling the load_class('Config', 'core') class registry here. But that's not really true ... `config_item()` is easier to write, true, but there's nothing that determines what "makes more sense". I could go in detail about how `load_class(...)->item()` makes perfect sense, but that's a moot point. What really shoud happen here is this: public function __construct(CI_Config $CFG) { // Use $CFG->item() // ... } And this is how you pass the `CI_Config` instance from the bootstrap script: // $CFG already being declared: $EXT =& load_class('Hooks', 'core', $CFG); |
@@ -2216,6 +2216,8 @@ void Player::SetGMVisible(bool on)
m_ExtraFlags |= PLAYER_EXTRA_GM_INVISIBLE; //add flag
SetAcceptWhispers(false);
+
+ SetAcceptMail(false);
SetGameMaster(true);
m_serverSideVisibility.SetValue(SERVERSIDE_VISIBILITY_GM, GetSession()->GetSecurity());
| [No CFG could be retrieved] | Add or remove a specific flag from the system. - If the player is in the same group as p - if the player is invited. | Seems kind of wrong to me to block recieving mails when the GM uses the invisibility command. |
@@ -1,7 +1,9 @@
"""Let's Encrypt CLI."""
# TODO: Sanity check all input. Be sure to avoid shell code etc...
+
import argparse
import atexit
+import configobj
import functools
import logging
import logging.handlers
| [_plugins_parsing->[add,add_group,add_plugin_args],_create_subparsers->[add_subparser,add_argument,flag_default],auth->[_init_le_client,_find_domains],HelpfulArgumentParser->[add_plugin_args->[add_group],add->[add_argument],__init__->[SilentParser,flag_default]],SilentParser->[add_argument->[add_argument]],install->[_init_le_client,_find_domains],run->[_init_le_client,_find_domains],create_parser->[HelpfulArgumentParser,flag_default,add,config_help,add_group],main->[_setup_logging,create_parser],_paths_parser->[config_help,add,add_group,flag_default],_init_le_client->[_determine_account],rollback->[rollback],main] | The Let s Encrypt CLI. | should be removed |
@@ -104,7 +104,17 @@ function compile(entryModuleFilename, outputDir,
// Don't include tests.
'!**_test.js',
'!**/test-*.js',
- ])
+ ];
+ // Many files include the polyfills, but we only want to deliver them
+ // once. Since all files automatically wait for the main binary to load
+ // this works fine.
+ if (options.includePolyfills) {
+ srcs.push('!build/fake-module/src/polyfills.js');
+ } else {
+ srcs.push('!src/polyfills.js');
+ }
+ /*eslint "google-camelcase/google-camelcase": 0*/
+ return gulp.src(srcs)
.pipe(closureCompiler({
// Temporary shipping with our own compiler that has a single patch
// applied
| [No CFG could be retrieved] | This is the main entry point for the compiler. It compiles the JS files into a Construct a new object. | I think the conditional is backwards? |
@@ -43,6 +43,7 @@ KNOWN_TASK_TYPES = {
'object_attributes',
'optical_character_recognition',
'semantic_segmentation',
+ 'mono_depth',
}
RE_MODEL_NAME = re.compile(r'[0-9a-zA-Z._-]+')
| [PostprocRegexReplace->[apply->[print_section_heading],deserialize->[validate_string,validate_nonnegative_int,validate_relative_path]],FileSourceGoogleDrive->[deserialize->[validate_string]],FileSourceHttp->[deserialize->[validate_string]],PostprocUnpackArchive->[apply->[print_section_heading],deserialize->[validate_string,validate_relative_path]],load_models_or_die->[print,load_models],ModelFile->[deserialize->[DeserializationError,validate_string,deserialization_context,validate_nonnegative_int,validate_relative_path,deserialize]],load_models->[add_model->[DeserializationError],print,DeserializationError,add_model,validate_string,deserialization_context,validate_relative_path,deserialize],validate_string->[DeserializationError],Model->[deserialize->[DeserializationError,validate_string,deserialization_context,validate_string_enum,deserialize]],deserialization_context->[DeserializationError],validate_string_enum->[validate_string,DeserializationError],validate_relative_path->[validate_string,DeserializationError],validate_nonnegative_int->[DeserializationError],TaggedBase->[deserialize->[DeserializationError]],load_models_from_args->[print,load_models_or_die],Reporter->[log_error->[print],with_event_context->[Reporter],print->[print],log_details->[print],log_warning->[print],emit_event->[print]]] | Creates a basic report of the given n - tuple. Print a section of the object. | Please: 1. Keep the list sorted. 2. Update the downloader README accordingly. |
@@ -48,6 +48,12 @@ class PythonSolver(object):
"echo_level" : 0
}""")
+ @classmethod
+ def ListOfParametersNotRecursivelyValidated(cls):
+ """This function skips the
+ """
+ return []
+
def AddVariables(self):
"""This function add the Variables needed by this PythonSolver to the the ModelPart
It has to be called BEFORE the ModelPart is read!
| [PythonSolver->[validate_and_transfer_matching_settings->[validate_and_transfer_matching_settings],Solve->[FinalizeSolutionStep,Initialize,Predict,InitializeSolutionStep,SolveSolutionStep]]] | Returns the default settings for a KratosMultiphysics model. | this does not really make sense I would say. This would grow infinitely due to different settings Also it is difficult to check if it works for all cases |
@@ -62,11 +62,18 @@ class MSCOCOorigBaseMetric(FullDatasetEvaluationMetric):
'threshold': BaseField(optional=True, default='.50:.05:.95', description='threshold for metric calculation')
})
+ parameters.update({
+ 'add_1_to_bbox_side': BoolField(
+ optional=True, default=True,
+ description="Whether to add '1' when computing height and width from bounding box coordinates."),
+ })
+
return parameters
def configure(self):
threshold = process_threshold(self.get_value_from_config('threshold'))
self.threshold = get_or_parse_value(threshold, COCO_THRESHOLDS)
+ self.box_side_delta = int(self.get_value_from_config('add_1_to_bbox_side'))
if not self.dataset.metadata:
raise ConfigError('coco orig metrics require dataset_meta'
'Please provide dataset meta file or regenerate annotation')
| [MSCOCOorigRecall->[evaluate->[compute_precision_recall]],MSCOCOorigBaseMetric->[_convert_data_to_generated_coco_format->[_iou_type_data_to_coco],_prepare_original_coco_structures->[generate_map_pred_label_id_to_coco_cat_id],compute_precision_recall->[_debug_printing_and_displaying_predictions,_convert_data_to_generated_coco_format,_prepare_original_coco_structures,_prepare_generated_coco_structures,_reload_results_to_coco_class,_convert_data_to_original_coco_format,_run_coco_evaluation],_iou_type_specific_coco_annotation->[_iou_type_data_to_coco],_prepare_data_for_annotation_file->[_iou_type_specific_coco_annotation],_reload_results_to_coco_class->[_create_json],_convert_data_to_original_coco_format->[_iou_type_data_to_coco]],MSCOCOorigAveragePrecision->[evaluate->[compute_precision_recall]]] | Return a dictionary of parameters for the coco orig metric calculation. | for me `include_boundaries` sounds better |
@@ -11,6 +11,7 @@
namespace Sulu\Bundle\CoreBundle\DependencyInjection;
+use Oro\ORM\Query\AST\Functions\Cast;
use Oro\ORM\Query\AST\Functions\String\GroupConcat;
use Sulu\Bundle\ContactBundle\Entity\Account;
use Sulu\Bundle\ContactBundle\Entity\AccountInterface;
| [SuluCoreExtension->[initListBuilder->[load],load->[load],initCache->[load],initWebspace->[load],initContent->[load]]] | This extension is a class that loads and manages the configuration of a specific node. Resolves configuration for . | should we mark the `Sulu\Component\Rest\DQL\Cast` as deprecated? |
@@ -1,7 +1,7 @@
# -*- encoding : utf-8 -*-
require 'digest/sha1'
-class AddUserBounceColumns < ActiveRecord::Migration
+class AddUserBounceColumns < !rails5? ? ActiveRecord::Migration : ActiveRecord::Migration[4.2] # 2.3
def self.up
add_column :users, :email_bounced_at, :datetime
add_column :users, :email_bounce_message, :text, :default => "", :null => false
| [AddUserBounceColumns->[up->[add_column],down->[remove_column]],require] | Adds the missing columns to the model so that users can be notified when they enter a confirmation. | Line is too long. [100/80] |
@@ -22,6 +22,9 @@ foreach (dbFetchRows('SELECT * FROM `ports` WHERE `device_id` = ? AND `disabled`
if (is_array(\LibreNMS\Config::get('device_traffic_descr'))) {
foreach (\LibreNMS\Config::get('device_traffic_descr') as $ifdescr) {
+ if (is_null($ifdesc) || !is_string($ifdesc) || $ifdesc == '') {
+ continue; // or pgreg match may be a false positive
+ }
if (preg_match($ifdescr.'i', $port['ifDescr']) || preg_match($ifdescr.'i', $port['ifName'])) {
$ignore = 1;
}
| [No CFG could be retrieved] | Generate a list of ports and return a list of multi bits grapher. This function generates the list of all possible network interfaces for the given port. | Shouldn't `empty()` be enough? |
@@ -79,6 +79,13 @@ class KeyVaultClientBase(object):
+ "Supported versions: {}".format(", ".join(v.value for v in ApiVersion))
)
+ def _get_attributes(self, enabled, not_before, expires_on, exportable=None):
+ """Return a KeyAttributes object if none-None attributes are provided, or None otherwise"""
+ if enabled is not None or not_before is not None or expires_on is not None or exportable is not None:
+ return self._models.KeyAttributes(
+ enabled=enabled, not_before=not_before, expires=expires_on, exportable=exportable
+ )
+ return None
@property
def vault_url(self):
| [KeyVaultClientBase->[close->[close],__exit__->[__exit__],__enter__->[__enter__]]] | Initialize a new object with a single key in the vault. | This should be on `KeyClient` so `_shared` remains identical across azure-keyvault-* |
@@ -564,6 +564,17 @@ class FromParams:
else:
# This is not a base class, so convert our params and extras into a dict of kwargs.
+ # If there was a "type" given, we probably didn't want to end up here.
+ if "type" in params:
+ value = params.pop("type")
+ message = (
+ f"{value} not in acceptable choices for {params.history}type. "
+ "You should either use the --include-package flag to make sure the correct module "
+ "is loaded, or use a fully qualified class name in your config file like "
+ """{"model": "my_module.models.MyModel"} to have it imported automatically."""
+ )
+ raise ConfigurationError(message)
+
# See the docstring for an explanation of what's going on here.
if not constructor_to_inspect:
constructor_to_inspect = cls.__init__
| [create_kwargs->[remove_optional,infer_params],infer_params->[infer_params],FromParams->[from_params->[create_extras,from_params,create_kwargs]],create_extras->[takes_arg,takes_kwargs],can_construct_from_params->[can_construct_from_params],construct_arg->[create_extras,can_construct_from_params,construct_arg]] | Instantiates a new object from a list of parameters. indicates malformed parameters in a configuration file where something that s actually a list or something else This is a recursive method that creates a object. | Missing space before "type"? |
@@ -539,6 +539,15 @@ define([
*/
this.fog = new Fog();
+ /**
+ * Determines the uniform depth size in meters of each frustum of the multifrustum in 2D. If a primitive or model close
+ * to the surface shows z-fighting, decreasing this will eliminate the artifact, but decrease performance. On the
+ * other hand, increasing this will increase performance but may cause z-fighting among primitives close to thesurface.
+ * @type {Number}
+ * @default 1.75e6
+ */
+ this.frustumSize2D = 1.75e6;
+
this._terrainExaggeration = defaultValue(options.terrainExaggeration, 1.0);
this._performanceDisplay = undefined;
| [No CFG could be retrieved] | Displays a single . Initialize the object with the values of the n - tuple. | Put this under `farToNearRatio` and perhaps name it `nearToFarDistance2D` for better symmetry. Update CHANGES.md. |
@@ -645,8 +645,12 @@ def cosine_distance(
ValueError: If `predictions` shape doesn't match `labels` shape, or
`weights` is `None`.
"""
- if dim is None:
- raise ValueError("`dim` cannot be None.")
+ if dim is not None:
+ if axis is not None:
+ raise ValueError("Cannot specify both 'axis' and 'dim'")
+ axis = dim
+ if axis is None and dim is None:
+ raise ValueError("`axis` and `dim` cannot both be None.")
with ops.name_scope(scope, "cosine_distance_loss",
[predictions, labels, weights]) as scope:
predictions.get_shape().assert_is_compatible_with(labels.get_shape())
| [_safe_mean->[_safe_div],mean_pairwise_squared_error->[_scale_losses,add_loss,_safe_div,_num_present],compute_weighted_loss->[_safe_mean,_scale_losses],get_total_loss->[get_losses,get_regularization_losses],absolute_difference->[compute_weighted_loss],cosine_distance->[compute_weighted_loss],sigmoid_cross_entropy->[compute_weighted_loss],softmax_cross_entropy->[compute_weighted_loss],sparse_softmax_cross_entropy->[compute_weighted_loss],log_loss->[compute_weighted_loss],mean_squared_error->[compute_weighted_loss]] | Adds a cosine - distance loss to the training procedure. | Can you change the error message here to "You must specify 'axis'." Since dim is deprecated, no need to mention it. |
@@ -42,6 +42,12 @@ const (
// NsRunDir is the default directory in which running network namespaces
// are stored
NsRunDir = "/var/run/netns"
+
+ // JournaldLogging is the string conmon expects to specify journald logging
+ JournaldLogging = "journald"
+
+ // KubernetesLogging is the string conmon expects when specifying to use the kubernetes logging format
+ KubernetesLogging = "k8s-file"
)
// OCIRuntime represents an OCI-compatible runtime that libpod can call into
| [killContainer->[ID,Wrapf,Sprintf,GetRootlessRuntimeDir,ExecCmdWithStdStreams,Debugf],pauseContainer->[ExecCmdWithStdStreams,Sprintf,ID,GetRootlessRuntimeDir],unpauseContainer->[ExecCmdWithStdStreams,Sprintf,ID,GetRootlessRuntimeDir],updateContainerStatus->[Now,Close,StderrPipe,removeConmonFiles,IsNotExist,Stat,exitFilePath,GetRootlessRuntimeDir,Start,Errorf,Wait,handleExitFile,ID,NewBuffer,Wrapf,Contains,StdoutPipe,NewDecoder,Command,Sprintf,Decode,ReadAll],startContainer->[ID,Sprintf,Now,GetRootlessRuntimeDir,LookupEnv,ExecCmdWithStdStreams],featureCheckCheckpointing->[Start,Wait,Command],checkpointContainer->[ID,SetSocketLabel,bundlePath,ProcessLabel,ExecCmdWithStdStreams,CheckpointPath,Debugf],execContainer->[LogPath,ID,Wrapf,Sprintf,GetRootlessRuntimeDir,Start,Close,execPidPath,Command,NewFile,Debugf],deleteContainer->[ExecCmdWithStdStreams,Sprintf,ID,GetRootlessRuntimeDir],Wrapf,Join,ListenTCP,Sprintf,ListenUDP,Warnf,ResolveTCPAddr,ResolveUDPAddr,Errorf,File,MkdirAll,IsExist] | Requirements for a single node create an OCI runtime object. | I think these constants are probably better in container.go |
@@ -121,6 +121,10 @@ public class JoinNodeTest {
@Rule
public final ExpectedException expectedException = ExpectedException.none();
+ @Mock
+ private DataSource<?> leftSource;
+ @Mock
+ private DataSource<?> rightSource;
@Mock
private DataSourceNode left;
@Mock
| [JoinNodeTest->[buildJoin->[buildJoin],setupTable->[setupTable]]] | Creates a new instance of the KsqlStreamBuilder and KsqlStreamBuilder. If all the components in the builder are available return the configuration. | do we need a test for serde options of internal join topics? |
@@ -66,7 +66,7 @@ class TestPythonSources(TestBase):
multiple_sources = PythonBinarySources(["f1.py", "f2.py"], address=address)
with pytest.raises(ExecutionError) as exc:
self.request_single_product(SourcesResult, multiple_sources.request)
- assert "//:binary had 2 sources" in str(exc)
+ assert "has 2 sources" in str(exc)
def test_python_library_sources_default_globs(self) -> None:
self.create_files(path="", files=[*self.PYTHON_SRC_FILES, *self.PYTHON_TEST_FILES])
| [TestPythonSources->[test_python_library_sources_default_globs->[PythonLibrarySources,create_files,request_single_product,parse],rules->[target_rules,RootRule],test_python_binary_sources_validation->[parse,str,create_files,PythonBinarySources,raises,request_single_product],test_python_tests_sources_default_globs->[PythonTestsSources,parse,create_files,set,request_single_product],test_python_sources_validation->[parse,str,create_files,PythonSources,raises,request_single_product]],test_timeout_validation->[parse,raises,Timeout]] | Test if the binary sources are valid. | Can you also `assert "//:binary" in str(exc)`? |
@@ -98,11 +98,15 @@ public class Extension {
}
public String[] getLabels() {
- return labels;
+ List<String> keywords = (List<String>) getMetadata().get("keywords");
+ if(keywords != null) {
+ return (String[]) keywords.toArray(new String[keywords.size()]);
+ }
+ return null;
}
public Extension setLabels(String[] labels) {
- this.labels = labels;
+ getMetadata().put("keywords", Arrays.asList(labels));
return this;
}
| [Extension->[managementKey->[getArtifactId,getGroupId],equals->[equals],hashCode->[hashCode],toString->[gav],toDependency->[setVersion,setArtifactId,setScope,setGroupId,setClassifier],gav->[managementKey]]] | Gets the labels of this extension. | This looks like not been formatted. |
@@ -11,7 +11,7 @@ class ServiceProviderRequestProxy
cattr_accessor :redis_last_uuid
def self.from_uuid(uuid)
- find_by(uuid: uuid) || NullServiceProviderRequest.new
+ find_by(uuid: uuid.to_s) || NullServiceProviderRequest.new
rescue ArgumentError # a null byte in the uuid will raise this
NullServiceProviderRequest.new
end
| [ServiceProviderRequestProxy->[last->[find_by],create!->[create],write->[write],find_or_create_by->[find_by],delete->[delete]]] | Returns a object for the given uuid. If no such object exists a new one is. | Thanks for the fix and even more, the tests to ensure we don't regress! |
@@ -2098,8 +2098,14 @@ static void modflags_changed(GtkWidget *widget, gpointer user_data)
void gui_changed(dt_iop_module_t *self, GtkWidget *w, void *previous)
{
dt_iop_lensfun_params_t *p = (dt_iop_lensfun_params_t *)self->params;
+ dt_iop_lensfun_gui_data_t *g = (dt_iop_lensfun_gui_data_t *)self->gui_data;
- if(p->tca_r != 1.0 || p->tca_b != 1.0) p->tca_override = 1;
+ // enable tca override toggle if tca_b or tca_r is changed
+ if(w == g->tca_b || w == g->tca_r)
+ {
+ p->tca_override = 1;
+ gui_update(self);
+ }
p->modified = 1;
}
| [No CFG could be retrieved] | This is the entry point for the lighttable. develop class. - - - - - - - - - - - - - - - - - -. | As this is now a boolean we also want to clean the setting : 1 -> TRUE Also please double check any other setting in the code as I suppose we have 0 or 1 instead of TRUE or FALSE. |
@@ -7,7 +7,7 @@ Ecu = car.CarParams.Ecu
# Steer torque limits
class SteerLimitParams:
def __init__(self, CP):
- if CP.carFingerprint in [CAR.SONATA, CAR.PALISADE]:
+ if CP.carFingerprint in [CAR.SONATA, CAR.PALISADE, CAR.IONIQ]:
self.STEER_MAX = 384
else:
self.STEER_MAX = 255
| [dbc_dict,set] | Initialize the object with the values of the missing values. | Both Ioniq's should be here. 2018-19 does 409 no problem, 384 is also fine. |
@@ -667,10 +667,12 @@ class GradeEntryFormsControllerTest < AuthenticatedControllerTest
@new_grade = -5
@original_grade = Grade.find_by_grade_entry_student_id_and_grade_entry_item_id(@grade_entry_student_with_some_grades.id,
@grade_entry_items[0].id).grade
- post_as @admin, :update_grade, {:grade_entry_item_id => @grade_entry_items[0].id,
- :student_id => @grade_entry_student_with_some_grades.user_id,
- :updated_grade => @new_grade,
- :id => @grade_entry_form_with_grade_entry_items.id}
+ post_as @admin, :update_grade,
+ format: :js,
+ grade_entry_item_id: @grade_entry_items[0].id,
+ student_id: @grade_entry_student_with_some_grades.user_id,
+ updated_grade: @new_grade,
+ id: @grade_entry_form_with_grade_entry_items.id
assert_not_nil assigns :grade
assert render_template :update_grade
assert_response :success
| [GradeEntryFormsControllerTest->[position,update_total_grade,new,find,post_as,fixture_file_upload,assert_raise,errors,name,should,description,find_by_name,assert_equal,body,assert,message,returns,find_by_user_id,find_by_short_identifier,assigns,t,released_to_student,assert_not_nil,date,find_by_grade_entry_student_id_and_grade_entry_item_id,find_by_id,assert_nil,from_now,to_s,empty?,out_of,find_by_user_name,times,show_total,get_as,each,make,id,put_as,length,user_id,short_identifier,context,grade,grade_entry_items,assert_response,reload,assert_match,render_template,setup,save],join,require,expand_path,dirname] | this is a hack to make sure that the user has a grade and that the user has POST on update_grade when the Grade does not have an existing value and the Gra. | Line is too long. [82/80] |
@@ -351,11 +351,12 @@ public class JoinNode extends PlanNode {
.getKsqlTopic()
.getValueSerdeFactory();
- final KsqlSchema schema = dataSource.getSchema();
-
return builder.buildGenericRowSerde(
valueSerdeFactory,
- schema.getSchema(),
+ PhysicalSchema.from(
+ dataSource.getSchema(),
+ dataSource.getSerdeOptions()
+ ),
contextStacker.getQueryContext()
);
}
| [JoinNode->[StreamToTableJoiner->[join->[getRight,getRightJoinFieldName,getLeftJoinFieldName,getKeyField,getSerDeForNode,getJoinedKeyField,join,getRightAlias,buildStream,buildTable,getLeft]],TableToTableJoiner->[join->[getRight,getRightJoinFieldName,getLeftJoinFieldName,getOuterJoinedKeyField,getKeyField,getJoinedKeyField,join,getRightAlias,getLeftAlias,buildTable,getLeft]],ensureMatchingPartitionCounts->[getPartitions],StreamToStreamJoiner->[join->[getRight,getRightJoinFieldName,getLeftJoinFieldName,getOuterJoinedKeyField,getKeyField,getSerDeForNode,join,buildStream,getJoinedKeyField,getLeft]],getPartitions->[getPartitions],Joiner->[buildTable->[getSchema,buildStream],getSerDeForNode->[getSchema],maybeRePartitionByKey->[getSchema],buildStream->[buildStream]]]] | Gets the Serde for the given node. | in the description you mentioned that JoinNode now uses `SerdeOption.none()` for internal topics - is that being done here? Should it? |
@@ -996,7 +996,10 @@ size_t ossl_statem_client_max_message_size(SSL *s)
return CCS_MAX_LENGTH;
case TLS_ST_CR_SESSION_TICKET:
- return SSL3_RT_MAX_PLAIN_LENGTH;
+ if (SSL_IS_TLS13(s))
+ return SESSION_TICKET_MAX_LENGTH_TLS13;
+ else
+ return SESSION_TICKET_MAX_LENGTH_TLS12;
case TLS_ST_CR_FINISHED:
return FINISHED_MAX_LENGTH;
| [No CFG could be retrieved] | Reads the next key from the state machine and updates the cipher state according to the state machine Get the construction function and message type for sending from the client. | really hate the else after return. maybe use `?:` ? |
@@ -313,7 +313,6 @@ const files = {
path: TEST_SRC_DIR,
templates: [
'jest.conf.js',
- 'spec/enzyme-setup.ts',
'spec/storage-mock.ts',
'spec/app/utils.ts',
'spec/app/config/axios-interceptor.spec.ts',
| [No CFG could be retrieved] | This module provides a list of all possible configuration files that can be used to configure a layout Component spec for all components. | Cleanup this file. |
@@ -19,6 +19,8 @@
package io.druid.query.filter;
+import com.metamx.collections.bitmap.BitmapFactory;
+import com.metamx.collections.bitmap.ImmutableBitmap;
import com.metamx.collections.spatial.ImmutableRTree;
import io.druid.segment.data.Indexed;
import it.uniroma3.mat.extendedset.intset.ImmutableConciseSet;
| [No CFG could be retrieved] | This class is used to provide a basic description of a specific header of a specific header of. | This was removed from POM, should it still be here? |
@@ -145,7 +145,10 @@ public abstract class AbstractIntegrationNamespaceHandler implements NamespaceHa
alreadyRegistered = parserContext.getRegistry().isBeanNameInUse(CHANNEL_INITIALIZER_BEAN_NAME);
}
if (!alreadyRegistered) {
- BeanDefinitionBuilder channelDef = BeanDefinitionBuilder.genericBeanDefinition(ChannelInitializer.class);
+ String channelsAutoCreateExpression = "#{@" +IntegrationContextUtils.INTEGRATION_PROPERTIES_BEAN_NAME +
+ "['" + IntegrationProperties.CHANNELS_AUTOCREATE + "']}";
+ BeanDefinitionBuilder channelDef = BeanDefinitionBuilder.genericBeanDefinition(ChannelInitializer.class)
+ .addPropertyValue("autoCreate", channelsAutoCreateExpression);
BeanDefinitionHolder channelCreatorHolder = new BeanDefinitionHolder(channelDef.getBeanDefinition(), CHANNEL_INITIALIZER_BEAN_NAME);
BeanDefinitionReaderUtils.registerBeanDefinition(channelCreatorHolder, parserContext.getRegistry());
}
| [AbstractIntegrationNamespaceHandler->[parse->[parse],NamespaceHandlerDelegate->[init->[init],doRegisterBeanDefinitionParser->[registerBeanDefinitionParser],doRegisterBeanDefinitionDecorator->[registerBeanDefinitionDecorator],doRegisterBeanDefinitionDecoratorForAttribute->[registerBeanDefinitionDecoratorForAttribute]],decorate->[decorate]]] | Register the implicit channel creator bean. This method registers the channel registry bean with the parser context. | Since the CI is `BeanFactoryAware`, I think I would prefer for it to retrieve the property itself (if the field was not set by an explicit declaration). My reasoning is that this `channels.autoCreate=junk` equates to `false`, with no error or warning. If the CI tests it itself, it can still treat it as false, but emit a WARN log. WDYT? |
@@ -121,9 +121,14 @@ export default Controller.extend(ModalFunctionality, {
actions: {
onChangeStatusType(value) {
- if (value !== CLOSE_STATUS_TYPE) {
+ if (![CLOSE_AFTER_LAST_POST_STATUS_TYPE].includes(value)) {
this.set("topicTimer.based_on_last_post", false);
}
+
+ if (value === CLOSE_AFTER_LAST_POST_STATUS_TYPE) {
+ this.set("topicTimer.based_on_last_post", true);
+ }
+
this.set("topicTimer.status_type", value);
},
| [No CFG could be retrieved] | This method is called when the user clicks on the topic_timer modal. Method to update the topic status. | It's a bit odd to use `includes` for one value. I'd use `value !== CLOSE_AFTER_LAST_POST_STATUS_TYPE` |
@@ -69,6 +69,9 @@ def plot_epochs_image(epochs, picks=None, sigma=0., vmin=None,
Figure instance to draw the image to. Figure must contain two axes for
drawing the single trials and evoked responses. If None a new figure is
created. Defaults to None.
+ logscale : bool
+ Plots logarithmic colorscale (base 2) of epochs if True. Defaults to
+ False.
Returns
-------
| [_plot_onkey->[_plot_traces,_plot_window],_mouse_click->[plot_epochs_image,_pick_bad_epochs,_plot_window],_toggle_labels->[_plot_vert_lines],_epochs_navigation_onclick->[_draw_epochs_axes]] | Plots a series of epochs on a single - channel image. Plots a list of matplotlib figures and plots a single channel displayed . Plot the nanoseconds of the data. Plots the missing - block block layout. | your colorscale is linear but data is logarithmic, right? |
@@ -139,9 +139,8 @@ module.exports = options => ({
}),
new ForkTsCheckerWebpackPlugin({ eslint: true }),
new CopyWebpackPlugin([
- { from: './node_modules/swagger-ui/dist/css', to: 'swagger-ui/dist/css' },
- { from: './node_modules/swagger-ui/dist/lib', to: 'swagger-ui/dist/lib' },
- { from: './node_modules/swagger-ui/dist/swagger-ui.min.js', to: 'swagger-ui/dist/swagger-ui.min.js' },
+ { from: './node_modules/swagger-ui-dist/*.{js,css,html,png}', to: 'swagger-ui', flatten: true, ignore: 'index.html'},
+ { from: './node_modules/axios/dist/axios.min.js', to: 'swagger-ui'},
{ from: './<%= MAIN_SRC_DIR %>/swagger-ui/', to: 'swagger-ui' },
{ from: './<%= MAIN_SRC_DIR %>content/', to: 'content' },
{ from: './<%= MAIN_SRC_DIR %>favicon.ico', to: 'favicon.ico' },
| [No CFG could be retrieved] | Plugin that defines the necessary dependencies for the Swagger UI. includes favicon. ico. | the ignore should be: `ignore: ['index.html']` |
@@ -163,6 +163,12 @@ func getK8sMasterVars(cs *api.ContainerService) (map[string]interface{}, error)
}
masterVars["environmentJSON"] = environmentJSON
masterVars["provisionConfigsCustomCloud"] = getBase64EncodedGzippedCustomScript(kubernetesCSECustomCloud)
+
+ if cs.Properties.OrchestratorProfile.IsAzureCNI() {
+ masterVars["serviceManagementEndpoint"] = cs.Properties.CustomCloudProfile.Environment.ServiceManagementEndpoint
+ masterVars["resourceManagerEndpoint"] = cs.Properties.CustomCloudProfile.Environment.ResourceManagerEndpoint
+ masterVars["activeDirectoryEndpoint"] = cs.Properties.CustomCloudProfile.Environment.ActiveDirectoryEndpoint
+ }
}
if kubernetesConfig != nil {
| [IsAvailabilitySets,AnyAgentUsesVirtualMachineScaleSets,IsAddonEnabled,AreAgentProfilesCustomVNET,HasMultipleNodes,IsCustomVNET,GetAgentVMPrefix,FormatBool,GetCustomEnvironmentJSON,GetPrimaryScaleSetName,IsAzureStackCloud,HasStorageAccountDisks,IsStorageAccount,IsWindows,IsVirtualMachineScaleSets,Bool,PrivateJumpboxProvision,IsHostedMasterProfile,HasWindows,IsFeatureEnabled,GetCustomCloudIdentitySystem,IsPrivateCluster,HasCosmosEtcd,Sprintf,GetSizeMap,Unmarshal,GetMasterVMPrefix,IsLowPriorityScaleSet,GetAgentPoolIndexByName,EncodeToString,HasDisks,GetCustomCloudAuthenticationMethod] | get the master variables for a specific version of Cosmos Configure the private key for the ETCD server. | Add cs.Properties.IsAzureStackCloud() condition as well. |
@@ -268,6 +268,17 @@ function adoptShared(global, callback) {
setStyle(global.document.documentElement, 'cursor', 'pointer');
}
+ // Some deferred polyfills.
+ if (
+ // eslint-disable-next-line no-undef
+ (INTERSECTION_OBSERVER_POLYFILL || getMode().localDev || getMode().test) &&
+ shouldLoadInObPolyfill(global)
+ ) {
+ Services.extensionsFor(global).preloadExtension(
+ 'amp-intersection-observer-polyfill'
+ );
+ }
+
return iniPromise;
}
| [No CFG could be retrieved] | Preloads a single extension if it s not already loaded. Load a unique identifier from the extension cache. | Wonder if it is possible to have cache preload the extension based on user agent? |
@@ -161,8 +161,8 @@ public class DeckAdapter extends RecyclerView.Adapter<DeckAdapter.ViewHolder> {
// Update views for this node
Sched.DeckDueTreeNode node = mDeckList.get(position);
- // Create the right expander for this deck
- setDeckExpander(holder.deckExpander, node);
+ // Create the correct expander for this deck
+ setDeckExpander(holder.deckExpander, holder.indentView, node, mHasSubdecks);
if (node.children.size() > 0) {
holder.deckExpander.setTag(node.did);
| [DeckAdapter->[findDeckPosition->[findDeckPosition],onCreateViewHolder->[ViewHolder],processNodes->[processNodes]]] | Bind the view holder with the given position of the deck. LongClickListener - Called when the user clicks on the long click. | Can't you just use `node.children.size() > 0` instead of storing and fetching `mHasSubdecks`? Also you should never need to pass a private variable as a parameter since you can use it directly in the method. |
@@ -284,7 +284,10 @@ namespace Microsoft.Xna.Framework.Content.Pipeline.Audio
// Loop start and length in number of samples. Defaults to entire sound
loopStart = 0;
- loopLength = data.Count / ((bitsPerSample / 8) * channelCount);
+ if (data != null && bitsPerSample > 0 && channelCount > 0)
+ loopLength = data.Count / ((bitsPerSample / 8) * channelCount);
+ else
+ loopLength = 0;
}
finally
{
| [AudioContent->[ConvertFormat->[QualityToBitRate],Read->[Read]]] | Convert the AudioContent to the specified format. External tool for reading and parsing vorbis files. This function parses the output of the ffprobe command and returns the missing information if the command. | Under what conditions can all this data... the array, the sample bits, and the channel count.... be null? Are these not required to correctly generate audio content? |
@@ -185,7 +185,7 @@ static void eq_fir_s16_passthrough(struct fir_state_32x16 fir[],
struct audio_stream *sink,
int frames, int nch)
{
- audio_stream_copy_s16(source, sink, frames * nch);
+ audio_stream_copy_s16(source, 0, sink, 0, frames * nch);
}
#endif /* CONFIG_FORMAT_S16LE */
| [No CFG could be retrieved] | pass - through functions to replace the fir core while not configured for SOF_IP Function to select pass - through depending on PCM format. | would it be better to introduce the offset into the source/sink structures ? This means it can be more easily used by other users without having to modify further APIs ? |
@@ -313,7 +313,7 @@ void mock_daemon::mine_blocks(size_t num_blocks, const std::string &miner_addres
{
bool blocks_mined = false;
const uint64_t start_height = get_height();
- const auto mining_timeout = std::chrono::seconds(30);
+ const auto mining_timeout = std::chrono::seconds(120);
MDEBUG("Current height before mining: " << start_height);
start_mining(miner_address);
| [try_init_and_run->[init_and_run,load_params,set_ports],deinit->[deinit],init->[init],init_and_run->[init],run_main->[run],stop_and_deinit->[deinit]] | Mock daemon. | What is the trezor daemon ? I'm surprised to see mining stuff in there. |
@@ -51,6 +51,11 @@ const buildNavigationEl = element => {
`;
};
+/** @enum {string} */
+const Screen = {
+ ONBOARDING_NAVIGATION: 'on',
+};
+
/** @enum */
export const State = {
HIDDEN: 0,
| [No CFG could be retrieved] | Generates a navigation education template for an element. The main entry point for the object. | nit: maybe be more descriptive about what this screen is? These names (the minified `on` version) are permanent since they'll be embedded in other viewers that we can't change, but I wonder if e.g. we experimented with multiple versions of onboarding, how we would tell them apart |
@@ -26,8 +26,7 @@ func canonicalizeServiceScope(scope string) string {
"storage-ro": "https://www.googleapis.com/auth/devstorage.read_only",
"storage-rw": "https://www.googleapis.com/auth/devstorage.read_write",
"taskqueue": "https://www.googleapis.com/auth/taskqueue",
- "trace-append": "https://www.googleapis.com/auth/trace.append",
- "trace-ro": "https://www.googleapis.com/auth/trace.readonly",
+ "trace": "https://www.googleapis.com/auth/trace.append",
"useraccounts-ro": "https://www.googleapis.com/auth/cloud.useraccounts.readonly",
"useraccounts-rw": "https://www.googleapis.com/auth/cloud.useraccounts",
"userinfo-email": "https://www.googleapis.com/auth/userinfo.email",
| [HashString] | Return a list of all urls that are available in the current network. canonicalizeServiceScopes - canonicalizes a list of scopes to a canonical list of scopes. | Thanks for upstreaming this, sorry for the late feedback but I'm hesitant to make this change as we might break the users of trace-append and trace-ro. Do these fields not work at all as is? @rileykarson |
@@ -57,9 +57,11 @@ public class CompactionGMCEPublishingAction implements CompactionCompleteAction<
public static final String ICEBERG_ID_ATTRIBUTE = "iceberg.id";
public static final String ICEBERG_REQUIRED_ATTRIBUTE = "iceberg.required";
+ public final static String GMCE_EMITTED_KEY = "GMCE.emitted";
private final State state;
private final CompactionJobConfigurator configurator;
private final Configuration conf;
+ private InputRecordCountHelper helper;
private EventSubmitter eventSubmitter;
public CompactionGMCEPublishingAction(State state, CompactionJobConfigurator configurator) {
| [CompactionGMCEPublishingAction->[getNewFileMetrics->[getAttribute,setAttribute,setProp,create,get,convert,getMaximumId,getDstNewFiles,toString,warn,getDatasetName,fromString,getIcebergFormat,put,getMetrics],onCompactionJobComplete->[setProp,sendGMCE,newArrayList,isVirtual,parse,clear,join,getDatasetName,getGobblinMCEProducer,getNewFileMetrics,getDstBaseDir,getOldFiles],getConfFromState,UnsupportedOperationException,getName]] | Creates a GMCE action that can be used to publish a GMCE file. Join the destination directory and dataset name to the dataset directory. | This class may deserve a better name given the fact that it is not only concerning for recordCount anymore. |
@@ -66,9 +66,17 @@ public class KubernetesSupport {
protected void watchLog(String podName, Consumer<InputStream> handler, String sinceTime, Executor executor) throws IOException {
try {
PodOperationsImpl pod = (PodOperationsImpl) client.pods().withName(podName);
+
+ List<Container> containers = pod.get().getSpec().getContainers();
+ String containerFilter = getSpecificUserContainer(containers)
+ .map(n -> "&container=" + n)
+ .orElse("");
+
StringBuilder url = new StringBuilder()
.append(pod.getResourceUrl().toString())
- .append("/log?pretty=false&follow=true×tamps=true");
+ .append("/log?pretty=false&follow=true×tamps=true")
+ .append(containerFilter);
+
if (sinceTime != null) {
url.append("&sinceTime=").append(sinceTime);
}
| [KubernetesSupport->[watchLog->[onFailure->[info,accept],onResponse->[string,debug,error,code,byteStream,warn,setName,info,accept,execute],append,build,Callback,toString,IOException,setName,withName,enqueue],cancelAllRequests->[cancelAll,dispatcher],getLogger,getConfiguration,createHttpClient,ofMinutes]] | Watch the log stream for changes. This method is called when the activity is not tracking an activity. It is called when the. | is this used somewhere else where the changes can alter current behavior? |
@@ -273,6 +273,9 @@ class PyNumpy(PythonPackage):
blas = 'atlas'
elif spec['blas'].name == 'veclibfort':
blas = 'accelerate'
+ elif spec['blas'].name == 'fujitsu-ssl2':
+ blas = 'blas'
+ env.append_flags('LDFLAGS', self.spec["blas"].libs.ld_flags)
else:
blas = 'blas'
| [PyNumpy->[set_blas_lapack->[write_library_dirs]]] | Sets up the environment variables for the BLAS and LAPACK libraries. | Maybe we need to add a `[fujitsu]` section that knows which libs to search for? |
@@ -174,6 +174,10 @@ public abstract class AbstractRegistry implements Registry {
}
}
} catch (Throwable e) {
+ if (e instanceof IOException) { // may not be recoverable when IOException throws, give up retrying
+ logger.warn("Failed to save registry cache file, cause: " + e.getMessage(), e);
+ return;
+ }
if (version < lastCacheChanged.get()) {
return;
} else {
| [AbstractRegistry->[destroy->[getRegistered,getSubscribed,unregister,getUrl,unsubscribe],saveProperties->[doSaveProperties],recover->[subscribe,getRegistered,getSubscribed,register],notify->[filterEmpty,notify],SaveProperties->[run->[doSaveProperties]],toString->[toString]]] | Saves the latest version of the properties file. | Why not catch IOException directly? |
@@ -1602,3 +1602,16 @@ def test_invalid_uid_fails(wiki_user, client):
assert response.status_code == 200
assert_no_cache_header(response)
assert b'This link is no longer valid.' in response.content
+
+
+def test_signin_landing(db, client, settings):
+ settings.MULTI_AUTH_ENABLED = True
+ response = client.get(reverse('socialaccount_signin'))
+ assert response.status_code == 200
+ doc = pq(response.content)
+ assert 'Create your MDN Web Docs Account' in doc.find('h1').text()
+
+
+def test_signin_landing_multi_auth_disabled(db, client):
+ response = client.get(reverse('socialaccount_signin'))
+ assert response.status_code == 404
| [test_user_edit_beta->[_get_current_form_field_values],test_user_edit_websites->[_get_current_form_field_values],test_user_edit_github_is_public->[_get_current_form_field_values],test_bug_709938_interests->[_get_current_form_field_values],KumaGitHubTests->[test_signup_private_github->[test_signup_public_github]],BanUserAndCleanupSummaryTestCase->[test_post_submits_no_revisions_to_akismet_when_no_user_revisions->[enable_akismet_and_mock_requests],test_post_submits_no_revisions_to_akismet_when_revisions_not_in_request->[enable_akismet_and_mock_requests],test_post_submits_revisions_to_akismet_as_spam->[enable_akismet_and_mock_requests],test_post_submits_no_revisions_to_akismet_when_wrong_revisions_in_request->[enable_akismet_and_mock_requests]],test_user_edit_interests->[_get_current_form_field_values]] | Test whether a user has invalid UID. | How is this test different from the one in `test_templates.py`? Actually, I'm not entirely sure what `test_TEMPLATES.py` mean because what really matters is what users SEE and that API is the view. Not the template. I'd vote you *move* that nice `test_signin_landing` into this file here. |
@@ -131,6 +131,18 @@ void genNetlinkRoutes(const struct nlmsghdr* netlink_msg, QueryData& results) {
case RTA_PRIORITY:
r["metric"] = INTEGER(*(int*)RTA_DATA(attr));
break;
+ case RTA_METRICS:
+ struct rtattr* xattr = static_cast<struct rtattr*> RTA_DATA(attr);
+ uint32_t xattr_size = RTA_PAYLOAD(attr);
+ while (RTA_OK(xattr, xattr_size)) {
+ switch (xattr->rta_type) {
+ case RTAX_HOPLIMIT:
+ r["hopcount"] = INTEGER(*(int*)RTA_DATA(xattr));
+ break;
+ }
+ xattr = RTA_NEXT(xattr, xattr_size);
+ }
+ break;
}
attr = RTA_NEXT(attr, attr_size);
}
| [genRoutes->[readNetlink,genNetlinkRoutes],genNetlinkRoutes->[getNetlinkIP]] | Generate the routes in the netlink message. Route type determination. | if(xattr->rta_type == RTAX_HOPLIMIT) |
@@ -1175,8 +1175,8 @@ namespace Dynamo.Controls
var customNodeManager = dynamoViewModel.Model.CustomNodeManager;
var packageLoader = dynamoViewModel.Model.GetPackageManagerExtension().PackageLoader;
var viewModel = new PackagePathViewModel(packageLoader, loadPackagesParams, customNodeManager);
- var view = new PackagePathView(viewModel) { Owner = this };
- view.ShowDialog();
+ //var view = new PackagePathView(viewModel) { Owner = this };
+ //view.ShowDialog();
}
private InstalledPackagesView _installedPkgsView;
| [DynamoView->[SaveExtensionWindowsState->[SaveExtensionWindowSettings],UndockExtension->[CloseExtensionTab],SlideWindowToIncludeTab->[GetSlidingWindowSize],Log->[Log],ExtensionHandle_MouseLeave->[UpdateHandleUnhoveredStyle],WindowClosing->[PerformShutdownSequenceOnViewModel],OnCollapsedLeftSidebarClick->[UpdateLibraryCollapseIcon],WorkspaceTabs_SizeChanged->[ToggleWorkspaceTabVisibility,UpdateLibraryCollapseIcon],DynamoView_Loaded->[DynamoLoadedViewExtensionHandler,LoadNodeViewCustomizations,SubscribeNodeViewCustomizationEvents,InitializeLogin,InitializeStartPage,InitializeShortcutBar],LibraryHandle_MouseLeave->[UpdateLibraryCollapseIcon],DynamoView_Unloaded->[UnsubscribeNodeViewCustomizationEvents],DynamoViewModelRequestShowPackageManagerSearch->[DisplayTermsOfUseForAcceptance],WindowClosed->[PerformShutdownSequenceOnViewModel],Window_PreviewMouseLeftButtonUp->[HidePopupWhenWindowDeactivated],CloseExtensionTab->[CloseExtensionTab],ToggleExtensionBarCollapseStatus->[UpdateLibraryCollapseIcon],WorkspaceTabs_TargetUpdated->[ToggleWorkspaceTabVisibility],ExtensionHandle_MouseEnter->[UpdateHandleHoveredStyle],Dispose->[OnCollectionChanged,Dispose]]] | Request that the package paths are loaded. | I guess I should delete this whole method and the debug menu item for raising the manage node and package path dialog as it's now shown under preferences - |
@@ -13,7 +13,7 @@ namespace DynamoCoreWpfTests
#region Regular Connector Tests
/// <summary>
- /// Check to see a pin can be added to a connector
+ /// Check to see a connector is visible after pre 2.13 graph open
/// </summary>
[Test]
public void ConnectorVisibilityForLegacyGraphTest()
| [ConnectorViewModelTests->[CanUndoDragPin->[CenterY,EndDrag,X,Execute,First,CenterX,ExecuteCommand,PanelX,AreEqual,Count,BeginDrag,FlipOnConnectorAnchor,Y,Open,PanelY,Undo,Add,AreNotEqual],PlaceWatchNodeAndRewireCorrectly->[Execute,Index,PanelX,FirstOrDefault,AreEqual,FlipOnConnectorAnchor,Parse,Open,PanelY,IsNotNull,GUID],PinPlacedOnCorrectWireToWatchNode->[ConnectorGuid,Execute,Index,PanelX,FirstOrDefault,Count,AreEqual,FlipOnConnectorAnchor,Parse,Open,PanelY,GUID],CanUndoDeletePin->[Execute,First,ExecuteCommand,PanelX,Count,AreEqual,FlipOnConnectorAnchor,Open,PanelY,Undo,GUID],CanSelectConnectedNodes->[Execute,First,AreEqual,Count,Open],CanBreakConnection->[Execute,First,AreEqual,Count,Open],CanHideConnector->[Execute,IsHidden,First,AreEqual,Open],CanUndoUnpin->[Execute,First,ExecuteCommand,PanelX,Count,Empty,AreEqual,FlipOnConnectorAnchor,Open,OnRequestRemove,PanelY,Undo],CanPlaceWatchNode->[Execute,First,Count,AreEqual,FlipOnConnectorAnchor,Open],CanUnPinFromConnector->[Execute,First,PanelX,Count,Empty,AreEqual,FlipOnConnectorAnchor,Open,OnRequestRemove,PanelY],CanPinConnector->[Execute,First,PanelX,Count,AreEqual,FlipOnConnectorAnchor,Open,PanelY],CanUndoPin->[Execute,First,ExecuteCommand,PanelX,Count,AreEqual,FlipOnConnectorAnchor,Open,PanelY,Undo],CanUnhideConnector->[Execute,IsHidden,First,AreEqual,Open],ConnectorVisibilityForLegacyGraphTest->[AreEqual,IsHidden,Open,First],CanUndoPlaceWatchNode->[Execute,ExecuteCommand,PanelX,FirstOrDefault,Count,AreEqual,FlipOnConnectorAnchor,Parse,Open,PanelY,Undo,GUID]]] | The connector is hidden when opening a legacy graph. | /// Check to see **if** a connector is visible after pre 2.13 graph open |
@@ -38,7 +38,11 @@ module View
text = tile.preprinted ? '' : '#'
text += name
text += "-#{rotation}" unless rotations == [0]
- text += " × #{num}" if num
+ if tile.unlimited
+ text += ' x ∞'
+ elsif num
+ text += " × #{num}"
+ end
hex = Engine::Hex.new(hex_coordinates || 'A1',
layout: layout,
| [Tiles->[render_tile_blocks->[x,h,new,preprinted,y,location_name,rotate!,any?,for,map]],require] | Renders a block of objects. | It looks like the multiplication sign is different sizes? Should they not be the same? |
@@ -12,3 +12,13 @@ class H2OCoxPHModel(ModelBase):
def coefficients_table(self):
"""Coefficients table."""
return self._model_json["output"]["coefficients_table"]
+
+ def summary(self):
+ """Prints summary information about this model."""
+ print("Call: ")
+ print(self.formula())
+ print(self.coefficients_table())
+ output = self._model_json["output"]
+ print("Likelihood ratio test=%f" % (output["loglik_test"]))
+ print("n=%d, number of events=%d" % (output["n"], output["total_event"]))
+
| [No CFG could be retrieved] | Coefficients table. | Is it ok, it overloads the `summary` method from the `ModelBase`? If there is nothing in `output["model_summary"]`, it is ok. :+1: |
@@ -953,6 +953,12 @@ module Engine
end
def payout_companies
+ # Functionality extracted so it can be over-ridden by subclasses
+ set_private_revenues
+ super
+ end
+
+ def set_private_revenues
# Set the correct revenue of P15-HR, P20-C&WR and P9-M&GNR
@companies.each do |c|
next unless c.owner
| [Game->[setup_companies->[optional_plus_expansion?,find_and_remove_train_by_id],company_made_choice_chpr->[add_exchange_token],bidbox_privates->[bank_companies],add_interest_player_loans!->[format_currency],payout_companies->[format_currency],starting_companies->[optional_plus_expansion?],on_acquired_train->[can_gain_extra_train?],status_str->[format_currency],company_choices_osth->[format_currency],company_made_choice_egr->[company_choices],company_made_choice_osth->[format_currency,float_corporation],extra_train_permanent_count->[extra_train_permanent?],payoff_player_loan->[format_currency],setup_bidboxes->[bidbox_minors_refill!],take_player_loan->[player_loan_interest],revenue_str->[format_currency],on_acquired_phase_revenue->[format_currency],company_tax_haven_payout->[format_currency]]] | Payout companies have a n - ary node with the correct revenue. | I don't mind removing it, but I thought it would be better to explain why a method contained only a single call along with "super". |
@@ -784,9 +784,11 @@ public class HoodieWriteClient<T extends HoodieRecordPayload> extends AbstractHo
/**
* NOTE : This action requires all writers (ingest and compact) to a dataset to be stopped before proceeding. Revert
- * the (inflight/committed) record changes for all commits after the provided @param. Three steps: (1) Atomically
+ * the (inflight/committed) record changes for all commits after the provided @param. Four steps: (1) Atomically
* unpublish this commit (2) clean indexing data, (3) clean new generated parquet/log files and/or append rollback to
* existing log files. (4) Finally delete .commit, .inflight, .compaction.inflight or .compaction.requested file
+ *
+ * @param instantTime Instant time
*/
public void restoreToInstant(final String instantTime) throws HoodieRollbackException {
| [HoodieWriteClient->[doRollbackAndGetStats->[rollback],rollbackInflightCompaction->[rollback],startCommitWithTime->[startCommit],runCompaction->[compact,commitCompaction],rollbackInternal->[doRollbackAndGetStats,startContext,finishRollback],startCommit->[startCommit],rollbackPendingCommits->[rollback],bulkInsert->[bulkInsert],close->[close],savepoint->[savepoint],forceCompact->[compact,scheduleCompaction],scheduleCompactionAtInstant->[scheduleCompaction],compact->[compact],commitCompaction->[commitCompaction],clean->[clean],doCompactionCommit->[finalizeWrite],commit->[commit],upsertRecordsInternal->[saveWorkloadProfileMetadataToInflight],updateIndexAndCommitIfNeeded->[commitOnAutoCommit],finalizeWrite->[finalizeWrite]]] | Restore the current state of the Hoodie timeline to the given time. rollback instants that were not committed. | minor: instant time to which restoration is requested |
@@ -118,7 +118,8 @@ class ClientBase:
:rtype: `.RegistrationResource`
"""
- return self.update_registration(regr, update={'status': 'deactivated'})
+ return self.update_registration(regr, messages.Registration.from_json(
+ {"status": "deactivated"}))
def deactivate_authorization(self,
authzr: messages.AuthorizationResource
| [ClientV2->[new_order->[_authzr_from_response,_post],revoke->[_revoke],poll->[_authzr_from_response],finalize_order->[_post],poll_authorizations->[_authzr_from_response],_get_v2_account->[_post],new_account->[_regr_from_response,_post],query_registration->[_regr_from_response,_post],_post_as_get->[_post]],Client->[revoke->[_revoke],agree_to_tos->[update_registration],poll->[_authzr_from_response],request_challenges->[_authzr_from_response,_post],check_cert->[_get_cert],fetch_chain->[_get_cert],refresh->[check_cert],query_registration->[_send_recv_regr],request_domain_challenges->[request_challenges],poll_and_request_issuance->[retry_after,poll,request_issuance],request_issuance->[_post],register->[_regr_from_response,_post]],ClientNetwork->[_get_nonce->[_check_response,_add_nonce,head],_post_once->[_check_response,_wrap_in_jws,_send_request,_add_nonce,_get_nonce],get->[_check_response,_send_request],head->[_send_request]],BackwardsCompatibleClientV2->[new_order->[request_domain_challenges],revoke->[revoke],finalize_order->[fetch_chain,request_issuance],new_account_and_tos->[agree_to_tos,_assess_tos,new_account,register],__init__->[ClientV2,Client]],ClientBase->[_revoke->[_post],update_registration->[_send_recv_regr],deactivate_registration->[update_registration],_send_recv_regr->[_regr_from_response],deactivate_authorization->[_post],answer_challenge->[_post]],_ClientDeprecationModule] | Deactivate a registration or authorization resource. | Ah, sorry, I jumped the gun. This change was not effective; `contact` is still serialized. |
@@ -2283,7 +2283,7 @@ func (a *apiServer) StartPipeline(ctx context.Context, request *pps.StartPipelin
return nil, err
}
- if err := a.updatePipelineState(pachClient, request.Pipeline.Name, pps.PipelineState_PIPELINE_RUNNING); err != nil {
+ if err := a.markPipelineRunning(pachClient, request.Pipeline.Name); err != nil {
return nil, err
}
return &types.Empty{}, nil
| [validateJob->[validateInput],getPachClient->[sudo],FlushJob->[listJob,InspectJob],GetLogs->[GetLogs,authorizePipelineOp],ActivateAuth->[fixPipelineInputRepoACLs,ListPipeline,sudo],makePipelineInfoCommit->[sudo],RestartDatum->[InspectJob],ListJobStream->[listJob],ListJob->[listJob],GarbageCollect->[ListPipeline],deletePipeline->[hardStopPipeline,inspectPipeline,authorizePipelineOp,fixPipelineInputRepoACLs,DeleteJob,sudo],ListDatum->[listDatum],CreatePipeline->[hardStopPipeline,makePipelineInfoCommit,authorizePipelineOp,fixPipelineInputRepoACLs,validatePipeline,sudo],StartPipeline->[inspectPipeline,authorizePipelineOp],fixPipelineInputRepoACLs->[sudo],DeleteAll->[DeletePipeline,DeleteJob,ListJob,ListPipeline],StopPipeline->[inspectPipeline,authorizePipelineOp],ListDatumStream->[listDatum],InspectDatum->[getDatum,InspectJob],listDatum->[InspectJob,authorizePipelineOp],validatePipeline->[validateInput]] | StartPipeline starts a pipeline. | After creating a helper to mark the pipeline as stopped, we only used this helper (`updatePipelineState`) here to mark a pipeline as running, so I simplified the helper and named it as such. |
@@ -88,6 +88,9 @@ std::string bstrToString(const BSTR src) {
LONGLONG cimDatetimeToUnixtime(const std::string& src) {
// First init the SWbemDateTime class
ISWbemDateTime* pCimDateTime = nullptr;
+ auto pCimDateTimeManager =
+ scope_guard::create([&pCimDateTime]() { pCimDateTime->Release(); });
+
auto hres = CoCreateInstance(CLSID_SWbemDateTime,
nullptr,
CLSCTX_INPROC_SERVER,
| [wstringToString->[to_bytes],cimDatetimeToUnixtime->[stringToWstring],bstrToString->[wstringToString],stringToWstring->[from_bytes]] | Convert a CIM Datetime string into a FILETIME and return the corresponding UNIX epoch timestamp. | Need to move this down below the error check otherwise we have the same potential segfault with the nullptr. |
@@ -66,6 +66,9 @@ final class ItemNormalizer extends AbstractItemNormalizer
$resourceMetadata = $this->resourceMetadataFactory->create($resourceClass);
$data = $this->addJsonLdContext($this->contextBuilder, $resourceClass, $context);
+ // Use resolved resource class instead of given resource class to support multiple inheritance child types
+ $context['resource_class'] = $resourceClass;
+
$rawData = parent::normalize($object, $format, $context);
if (!is_array($rawData)) {
return $rawData;
| [ItemNormalizer->[normalize->[create,addJsonLdContext,getShortName,getIriFromItem,getResourceClass,getIri],denormalize->[getItemFromIri]]] | Normalizes an object into a JSON - LD array. | ~~Is this true for classes without inheritance?~~ |
@@ -12,8 +12,8 @@ const { GitProcess } = require('dugite')
const GitHub = require('github')
const pass = '\u2713'.green
const path = require('path')
-const pkg = require('../package.json')
const readline = require('readline')
+const ReleaseNotes = require('./release-notes/index.js')
const versionType = args._[0]
const targetRepo = versionType === 'nightly' ? 'nightlies' : 'electron'
| [No CFG could be retrieved] | Creates a new version number based on the given version type. Bumps the version number and returns the new version number. | capitalizing `ReleaseNotes` here indicates to me that this is a class, but it looks like its actually a function called getReleaseNotes. I'd redo this as `const releaseNotesGenerator = require('./release-notes/index.js')` |
@@ -86,7 +86,7 @@ func (vm *VirtualMachine) FolderName(ctx context.Context) (string, error) {
func (vm *VirtualMachine) DSPath(ctx context.Context) (url.URL, error) {
var mvm mo.VirtualMachine
- if err := vm.Properties(ctx, vm.Reference(), []string{"runtime.host", "config"}, &mvm); err != nil {
+ if err := vm.Properties(ctx, vm.Reference(), []string{"config.files.vmPathName"}, &mvm); err != nil {
log.Errorf("Unable to get managed config for VM: %s", err)
return url.URL{}, err
}
| [WaitForKeyInExtraConfig->[WaitForExtraConfig,Error],fixVM->[LeaveFixingState,EnterFixingState,registerVM],Properties->[fixVM,Properties],WaitForResult->[WaitForResult,fixVM,needsFix],Parent->[Properties],UpgradeInProgress->[GetCurrentSnapshotTree],DatastoreReference->[Properties],bfsSnapshotTree->[bfsSnapshotTree]] | DSPath returns the datastore path of the managed VM. | Note that this is not a required change for this PR, but in the process notice we can reduce the data collected quite a bit here. |
@@ -35,8 +35,9 @@ import java.util.List;
public class MergeOnReadLazyInsertIterable<T extends HoodieRecordPayload> extends CopyOnWriteLazyInsertIterable<T> {
public MergeOnReadLazyInsertIterable(Iterator<HoodieRecord<T>> sortedRecordItr, HoodieWriteConfig config,
- String instantTime, HoodieTable<T> hoodieTable, String idPfx) {
- super(sortedRecordItr, config, instantTime, hoodieTable, idPfx);
+ String instantTime, HoodieTable<T> hoodieTable, String idPfx,
+ Suppliers suppliers) {
+ super(sortedRecordItr, config, instantTime, hoodieTable, idPfx, suppliers);
}
@Override
| [MergeOnReadLazyInsertIterable->[getInsertHandler->[MergeOnReadInsertHandler],MergeOnReadInsertHandler->[consumeOneRecord->[canWrite,getWriteStatus,write,getNextFileId,close,add,getPartitionPath,HoodieAppendHandle]]]] | Returns a copy - on - write insert handler. | side point: we should fix method arg formatting consistently between intellij and checkstyle. Keep seeing these sort of whitespace changes in PRs. |
@@ -385,6 +385,11 @@ def test_plot_compare_evokeds():
plot_compare_evokeds(evoked_dict, cmap=('cmap title', 'inferno'),
linestyles=['-', ':', '--'])
plt.close('all')
+ # test combine
+ match = 'must be an instance of None, callable, or str'
+ with pytest.raises(TypeError, match=match):
+ plot_compare_evokeds(evoked, combine=["mean", "gfp"])
+ plt.close('all')
# test warnings
with pytest.warns(RuntimeWarning, match='in "picks"; cannot combine'):
plot_compare_evokeds(evoked, picks=[0], combine='median')
| [test_plot_evoked_reconstruct->[_get_epochs,_get_amplitudes],test_plot_evoked->[_get_epochs,_get_epochs_delayed_ssp],test_plot_evoked_cov->[_get_epochs],test_plot_ctf->[get_axes_midpoints],test_plot_evoked_image->[_get_epochs],test_plot_compare_evokeds->[_get_epochs],test_plot_white->[_get_epochs]] | Test plot_compare_evokeds. Plot a single plot. Plot a bunch of colors and evokes and return the best fit. Plots a single in the plot. finds a specific n - th channel in the red object. | Can the test check for the item_name that you set to combine too? |
@@ -284,7 +284,7 @@ func ReapResult(r *resource.Result, f cmdutil.Factory, out io.Writer, isDefaultD
return nil
}
-func DeleteResult(r *resource.Result, out io.Writer, ignoreNotFound bool, shortOutput bool, mapper meta.RESTMapper) error {
+func DeleteResult(r *resource.Result, out io.Writer, ignoreNotFound bool, gracePeriod int, shortOutput bool, mapper meta.RESTMapper) error {
found := 0
if ignoreNotFound {
r = r.IgnoreErrors(errors.IsNotFound)
| [Validate->[Fprintf,Lookup,Flags,Errorf],Complete->[DefaultNamespace,NamespaceParam,ContinueOnError,Do,FilenameParam,ClientMapperFunc,ResourceTypeOrNameArgs,UnstructuredObject,NewBuilder,Flatten,SelectorParam,Err,SelectAllParam,RequireObject],AddOutputVarFlagsForMutation,Examples,Printer,IntVar,AddFilenameOptionFlags,Error,Stop,LongDesc,Validate,Reaper,StringVarP,HandledResources,DeleteWithOptions,Complete,DurationVar,CheckErr,ResourceAliases,AddInclude3rdPartyVarFlags,NewHelper,PollImmediate,NewDeleteOptions,ValidateOutputArgs,AddSourceToErr,IgnoreErrors,Get,PrintSuccess,Fprintf,IsNoSuchReaperError,Visit,UsageError,RunDelete,IsNotFound,T,Flags,BoolVar] | DeleteResult returns an error if no reaper is found. deleteResource deletes a resource by name. | as in upstream, a godoc would be nice about this int. time.Duration would be obvious, int is not. |
@@ -991,9 +991,14 @@ def gru_unit(input,
param_attr=None,
bias_attr=None,
activation='tanh',
- gate_activation='sigmoid'):
+ gate_activation='sigmoid',
+ origin_mode=False):
"""
- GRU unit layer. The equation of a gru step is:
+ **GRU unit layer**
+
+ if origin_mode is True, then the equation of a gru step is from paper
+ `Learning Phrase Representations using RNN Encoder-Decoder for Statistical
+ Machine Translation <https://arxiv.org/pdf/1406.1078.pdf>`_
.. math::
u_t & = actGate(xu_{t} + W_u h_{t-1} + b_u)
| [ctc_greedy_decoder->[topk],py_func->[PyFuncRegistry],image_resize->[_is_list_or_turple_],sequence_first_step->[sequence_pool],logical_xor->[_logical_op],elementwise_pow->[_elementwise_op],elementwise_min->[_elementwise_op],elementwise_max->[_elementwise_op],logical_not->[_logical_op],conv2d->[_get_default_param_initializer],logical_and->[_logical_op],sequence_last_step->[sequence_pool],elementwise_add->[_elementwise_op],elementwise_sub->[_elementwise_op],lstm_unit->[fc],elementwise_mul->[_elementwise_op],logical_or->[_logical_op],elementwise_div->[_elementwise_op],dice_loss->[reduce_sum,one_hot,reduce_mean],matmul->[__check_input],resize_bilinear->[image_resize],resize_nearest->[image_resize],image_resize_short->[image_resize],conv3d->[_get_default_param_initializer]] | This function creates a non - linear GRU unit layer. The hidden value reset - hidden value and the hidden state of the hidden state of the GR Adds a base on the base network for a . | > origin_mode Is there any other exact name? 1. ? 2. fine-tune? ?GPUcuDNN? @ @Xreki |
@@ -731,8 +731,7 @@ namespace ProtoFFI
return marshaler.UnMarshal(dsObject, context, dsi, expectedCLRType);
//The dsObject must be of pointer type
- Validity.Assert(dsObject.IsPointer || dsObject.IsFunctionPointer,
- string.Format("Operand type {0} not supported for marshalling", dsObject.optype));
+ Validity.Assert(dsObject.IsPointer || dsObject.IsFunctionPointer, "Operand type not supported for marshaling");
if (dsObject.IsFunctionPointer)
{
| [PointerValueComparer->[GetHashCode->[GetHashCode]],CharMarshaler->[CreateType],BoolMarshaler->[CreateType],CLRObjectMarshaler->[UnMarshal->[UnMarshal],CreateCLRObject->[BindObjects],GetUserDefinedType->[CreateType],GeneratePrimaryPropertiesAsXml->[GeneratePrimaryPropertiesAsXml,GetPrimaryProperties],FFIObjectMarshaler->[IsAssignableFromDictionary,CreateType],GetTypeName->[GetTypeName],StackValue->[GetTypeName],GetStringValue->[GeneratePrimaryPropertiesAsXml],ComputeDSType->[GetMarshaledType,GetProtoCoreType,CreateType],GetProtoCoreType->[CreateType],PopulatePrimaryProperties->[GetMarshaledType]],IntMarshaler->[CreateType],StringMarshaler->[CreateType],DictionaryMarshaler->[GetApproxDSType->[CreateType],StackValue->[GetApproxDSType]],PrimitiveMarshaler->[IsPrimitiveDSType->[IsPrimitiveRange]],FloatMarshaler->[CreateType],ReferenceEqualityComparer->[GetHashCode->[GetHashCode]],ArrayMarshaler->[StackValue->[GetApproxDSType],ICollection->[UnMarshal],GetApproxDSType->[CreateType],UnMarshal->[UnMarshal]]] | Unmarshals a given object. | Could we add the message to the string resources in the project. |
@@ -98,7 +98,11 @@ public class ParametersDefinitionProperty extends OptionalJobProperty<Job<?, ?>>
public List<String> getParameterDefinitionNames() {
return new AbstractList<String>() {
public String get(int index) {
- return parameterDefinitions.get(index).getName();
+ if (parameterDefinitions.get(index) != null) {
+ return parameterDefinitions.get(index).getName();
+ } else {
+ return null;
+ }
}
public int size() {
| [ParametersDefinitionProperty->[_doBuild->[get,getJob,_doBuild],getJobActions->[getJobActions],buildWithParameters->[buildWithParameters],getParameterDefinitionNames->[size->[size]]]] | Returns a list of parameter names that are not in the parameter list. | :ant: better to check for null elements in the constructor and `readResolve` |
@@ -137,10 +137,16 @@ func (cs *ContainerService) setOrchestratorDefaults(isUpdate bool) {
switch o.KubernetesConfig.ContainerRuntime {
case Docker:
if o.KubernetesConfig.MobyVersion == "" || isUpdate {
+ if isUpdate && o.KubernetesConfig.MobyVersion != DefaultMobyVersion {
+ log.Warnf("moby will be upgraded to version %s\n", DefaultMobyVersion)
+ }
o.KubernetesConfig.MobyVersion = DefaultMobyVersion
}
case Containerd, ClearContainers, KataContainers:
if o.KubernetesConfig.ContainerdVersion == "" || isUpdate {
+ if isUpdate && o.KubernetesConfig.ContainerdVersion != DefaultContainerdVersion {
+ log.Warnf("containerd will be upgraded to version %s\n", DefaultContainerdVersion)
+ }
o.KubernetesConfig.ContainerdVersion = DefaultContainerdVersion
}
}
| [SetPropertiesDefaults->[setVMSSDefaultsForMasters,setAgentProfileDefaults,setVMSSDefaultsForAgents,setStorageDefaults,SetDefaultCerts,GetCloudSpecConfig,setCustomCloudProfileDefaults,HasVMSSAgentPool,setExtensionDefaults,setHostedMasterProfileDefaults,setMasterProfileDefaults,IsVirtualMachineScaleSets,setWindowsProfileDefaults,setOrchestratorDefaults],setVMSSDefaultsForAgents->[BoolPtr,HasAvailabilityZones],setVMSSDefaultsForMasters->[BoolPtr,HasAvailabilityZones],setAgentProfileDefaults->[BoolPtr,IsKubernetes,Sprintf,IsUbuntu1804,IsVHDDistro,IsAzureCNI,AcceleratedNetworkingSupported,IsVirtualMachineScaleSets,Atoi,IsCustomVNET],setStorageDefaults->[IsKubernetesVersionGe],SetDefaultCerts->[Uint32,To4,PutUint32,CreatePkiKeyCertPair,GetCustomCloudName,CidrStringFirstIP,Errorf,CreatePki,ParseIP,GetLocations,IsVirtualMachineScaleSets],setMasterProfileDefaults->[GetFirstConsecutiveStaticIPAddress,IsKubernetes,BoolPtr,IsVHDDistro,IsVirtualMachineScaleSets,HasWindows,IsUbuntu1804,IsCustomVNET,Atoi,IsAzureCNI],setOrchestratorDefaults->[SetCloudProviderBackoffDefaults,setAPIServerConfig,setCloudControllerManagerConfig,setAddonsConfig,SetCloudProviderRateLimitDefaults,BoolPtr,IsRBACEnabled,IsKubernetesVersionGe,HasAvailabilityZones,Make,Bool,setKubeletConfig,TotalNodes,setSchedulerConfig,PrivateJumpboxProvision,GetCloudSpecConfig,HasWindows,IsAzureCNI,LT,setControllerManagerConfig,GetValidPatchVersion],Strings,Read,TrimSuffix,Sprintf,IsKubernetesVersionGe,String,WriteString,EncodeToString,Split,Trim] | setOrchestratorDefaults sets the default values for the Orchestrator profile and the default This function is used to set the default values for all configuration options. This function is called by DockerBridge to set default values for all configuration options. | let's uppercase `Moby` |
@@ -669,7 +669,7 @@ rebuild_offline_pool_connect_internal(void **state, unsigned int fail_loc)
/* hang the rebuild */
if (arg->myrank == 0)
- daos_mgmt_set_params(arg->group, -1, DMG_KEY_FAIL_LOC, fail_loc,
+ daos_debug_set_params(arg->group, -1, DMG_KEY_FAIL_LOC, fail_loc,
0, NULL);
MPI_Barrier(MPI_COMM_WORLD);
| [int->[uuid_is_null,rebuild_add_back_tgts,test_get_leader,MPI_Barrier,daos_cont_destroy,daos_cont_close,rebuild_close_container_cb,dmg_pool_destroy,daos_mgmt_set_params,DP_UUID,daos_handle_is_inval,rebuild_pool_connect_internal,rebuild_io_validate,print_message,sleep,MPI_Bcast,rebuild_io,MPI_Allreduce,uuid_clear,rebuild_pool_disconnect_internal],run_daos_rebuild_test->[MPI_Barrier,ARRAY_SIZE,run_daos_sub_tests_only],rebuild_test_setup->[MPI_Barrier,test_setup,daos_mgmt_set_params],rebuild_test_teardown->[MPI_Barrier,daos_mgmt_set_params,test_teardown],void->[daos_obj_layout_get,rebuild_add_back_tgts,rebuild_change_leader_cb,daos_kill_server,get_rank_by_oid_shard,test_get_leader,MPI_Barrier,reintegrate_single_pool_target,D_ASSERT,test_teardown,rebuild_destroy_pool_internal,test_pool_get_info,dts_oid_gen,reintegrate_single_pool_rank,memcmp,daos_mgmt_set_params,test_runable,rebuild_pools_ranks,test_rebuild_wait,assert_int_equal,rebuild_io_validate,print_message,sleep,rebuild_single_pool_rank,daos_exclude_server,rebuild_pool_create,rebuild_io,dts_oid_set_rank,rebuild_offline_pool_connect_internal,rebuild_pool_destroy,rebuild_tgt_pool_disconnect_internal,memset,rebuild_single_pool_target,reintegrate_pools_ranks,daos_obj_layout_free,daos_reint_server,dts_oid_set_tgt]] | Internal function to handle the rebuild of offline pools. | (style) line over 80 characters |
@@ -49,7 +49,7 @@ def _clang_compiler(output, compiler_exe="clang"):
compiler = "apple-clang"
elif "clang version" in out:
compiler = "clang"
- installed_version = re.search("([0-9]\.[0-9])", out).group()
+ installed_version = re.search("version ([0-9]\.[0-9])", out).group(1)
if installed_version:
output.success("Found %s %s" % (compiler, installed_version))
return compiler, installed_version
| [detect_defaults_settings->[_detect_os_arch,_detect_compiler_version],_sun_cc_compiler->[_execute],_visual_compiler_last->[_visual_compiler],_detect_os_arch->[detected_os],_gcc_compiler->[_execute],_get_default_compiler->[_visual_compiler_last,_clang_compiler,_sun_cc_compiler,_gcc_compiler],_detect_compiler_version->[_get_default_compiler],_visual_compiler->[_visual_compiler_cygwin],_clang_compiler->[_execute]] | Determine the compiler and installed version. | This change in the regular expression worries me a little. If some specific clang is not printing the "version" word we could be breaking. Why did you need it? If there is a good reason I'm ok with it, if don't, better keep it like it was. |
@@ -61,7 +61,8 @@ namespace System.Net.WebSockets
Abort();
- if (exc is WebSocketException)
+ if (exc is WebSocketException ||
+ (exc is OperationCanceledException && cancellationToken.IsCancellationRequested))
{
throw;
}
| [WebSocketHandle->[Task->[Closed,ThrowIfCancellationRequested,Token,CreateLinkedTokenSource,CanBeCanceled,Abort,ConfigureAwait,net_webstatus_ConnectFailure],Dispose->[Closed],Connecting]] | Connects to the given URI asynchronously. | I'm not sure what the optimal/expected behavior here is but this check should probably match the BrowserWebSocket case |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.