patch stringlengths 18 160k | callgraph stringlengths 4 179k | summary stringlengths 4 947 | msg stringlengths 6 3.42k |
|---|---|---|---|
@@ -47,8 +47,13 @@ import {isFormDataWrapper} from '../form-data-wrapper';
*/
export let FetchInitJsonDef;
-/** @private @const {!Array<function(*):boolean>} */
-const allowedJsonBodyTypes_ = [isArray, isObject];
+/**
+ * @typedef {{
+ * xhrUrl: string,
+ * fetchOpt: !FetchInitJsonDef
+ * }}
+ */
+export let FetchData;
/**
* A service that polyfills Fetch API for use within AMP.
| [No CFG could be retrieved] | Create a XHR object that can be used to fetch a single - document . The fetch_ method is called when the user requests a . | "FetchData" is a very confusing name. It sounds like the fetched response data. |
@@ -338,6 +338,10 @@ class DecodeError(HttpResponseError):
"""Error raised during response deserialization."""
+class IncompleteReadError(DecodeError):
+ """Error raised if peer closes connection without sending complete message body."""
+
+
class ResourceExistsError(HttpResponseError):
"""An error response with status code 4xx.
This will not be raised directly by the Azure core pipeline."""
| [ErrorMap->[get->[get]],ODataV4Format->[__init__->[get]],HttpResponseError->[__init__->[get]],AzureError->[__init__->[get]],map_error->[get],ODataV4Error->[__init__->[get]]] | This method is used to parse the response body of an HTTP GET or POST request to retrieve Initialize a TooManyRedirectsError with a message. | I don't think this should be a `DecodeError`. |
@@ -45,6 +45,12 @@ public interface ApplicationArchive extends Closeable {
*/
Path getArchiveRoot();
+ /**
+ *
+ * @return <code>true</code> if this archive represents part of the platform, i.e. an existing jar file that is not part of the current application
+ */
+ boolean isPlatformArchive();
+
/**
* Convenience method, returns the child path if it exists, otherwise null.
*
| [getChildPath->[exists,resolve]] | Returns the root of the archive. | So if I understand it correctly we don't specify additional app (non-platform) archives yet. It's only used in legacy launcher, right? |
@@ -6830,7 +6830,7 @@ namespace System.Windows.Forms
private protected void AdjustWindowRectEx(ref RECT rect, int style, bool bMenu, int exStyle)
{
- if (DpiHelper.IsPerMonitorV2Awareness)
+ if (DpiHelper.IsPerMonitorV2Awareness || DpiHelper.IsScalingRequired)
{
User32.AdjustWindowRectExForDpi(ref rect, style, bMenu.ToBOOL(), exStyle, (uint)_deviceDpi);
}
| [Control->[OnSystemColorsChanged->[OnSystemColorsChanged,Invalidate],UpdateRoot->[GetTopLevel],OnFontChanged->[GetAnyDisposingInHierarchy,Font,DisposeFontHandle,GetStyle,Invalidate],AccessibilityNotifyClients->[AccessibilityNotifyClients],OnParentFontChanged->[OnFontChanged],AutoValidate->[AutoValidate],OnParentBackColorChanged->[OnBackColorChanged],WmKeyChar->[ProcessKeyMessage,DefWndProc],WmWindowPosChanging->[ActiveXUpdateBounds,DefWndProc],WmMouseHover->[OnMouseHover,DefWndProc],AdjustWindowRectEx->[AdjustWindowRectEx],ScaleBitmapLogicalToDevice->[ScaleBitmapLogicalToDevice],OnDragOver->[OnDragOver],GetNeighboringToolsRectangles->[GetNeighboringToolsRectangles],CreateControl->[CreateHandle,CreateControl],WmParentNotify->[ReflectMessage,DefWndProc],ScaleFont->[DisposeFontHandle],OnParentEnabledChanged->[OnEnabledChanged,GetState],Refresh->[Invalidate],WmNotifyFormat->[ReflectMessage,DefWndProc],ResetPadding->[ResetPadding],DefWndProc->[DefWndProc],SetVisibleCore->[SelectNextIfFocused,SetState,OnVisibleChanged,GetTopLevel,CreateControl],CanShowToolTipsNow->[CanShowToolTipsNow],OnForeColorChanged->[GetAnyDisposingInHierarchy,Invalidate],OnRightToLeftChanged->[GetAnyDisposingInHierarchy],OnDragLeave->[OnDragLeave],SetAutoSizeMode->[SetAutoSizeMode],WmCtlColorControl->[InitializeDCForWmCtlColor,DefWndProc],SetAcceptDrops->[GetState],PreProcessControlState->[GetExtendedState,IsInputKey,PreProcessMessage,IsInputChar,OnPreviewKeyDown],WmPrintClient->[OnPrint],UpdateStyles->[OnStyleChanged],ProcessKeyMessage->[ProcessKeyEventArgs],OnDragDrop->[OnDragDrop],InitLayout->[InitLayout],WmPaint->[Graphics,PaintWithErrorHandling,GetStyle],OnParentRightToLeftChanged->[OnRightToLeftChanged],PerformContainerValidation->[PerformControlValidation,GetStyle,PerformContainerValidation],SelectNextControl->[Select],WmCaptureChanged->[OnMouseCaptureChanged,DefWndProc],OnVisibleChanged->[GetAnyDisposingInHierarchy,OnParentBecameInvisible,OnParentVisibleChanged,CreateControl],OnParentChanged->[OnTopMostActiveXParentChanged],ShouldPerformContainerValidation->[GetStyle],WmKillFocus->[InvokeLostFocus,DefWndProc],CreateHandle->[CreateHandle],Invoke->[Invoke],ProcessKeyPreview->[ProcessKeyPreview],WmDpiChangedAfterParent->[OnDpiChangedAfterParent,DefWndProc],SendToBack->[GetTopLevel],WmGetControlName->[MarshalStringToMessage],WmOwnerDraw->[ReflectMessage,DefWndProc],Dispose->[DestroyHandle,Dispose,ResetBindings],PrintToMetaFile_SendPrintMessage->[GetStyle],WndProc->[WmDestroy,WmGetControlType,ReflectMessage,WmNotify,WmKeyChar,WmGetControlName,WmEraseBkgnd,WmWindowPosChanging,WmClose,WmMouseUp,WmMouseHover,WmUpdateUIState,WmCommand,WmMove,WmParentNotify,WmNotifyFormat,InvokeMarshaledCallbacks,WmHelp,WmMouseMove,DefWndProc,SetState,OnNotifyMessage,GetStyle,WmCtlColorControl,WmShowWindow,WmCreate,WmMouseDown,WmPrintClient,WmPaint,WmCaptureChanged,WmMouseLeave,WmKillFocus,WmDisplayChange,WmGetObject,WmSetCursor,WmMouseWheel,WmQueryNewPalette,WmSetFocus,WmContextMenu,WmDpiChangedBeforeParent,WmDpiChangedAfterParent,WmWindowPosChanged,WmMouseEnter,WmOwnerDraw],ResumeLayout->[InitLayout,ResumeLayout,PerformLayout,OnLayoutResuming,GetState],Load->[Load],GetCaptionForTool->[GetCaptionForTool],PaintTransparentBackground->[InvokePaintBackground,Control,PaintTransparentBackground,InvokePaint,Graphics],WmMouseUp->[OnMouseDoubleClick,DefWndProc,SetState,GetStyle,OnMouseClick,OnClick,OnMouseUp,OnDoubleClick,GetState],DisposeAxControls->[DisposeAxControls],OnHandleCreated->[GetExtendedState,SetRegion,GetStyle,ListenToUserPreferenceChanged,GetTopLevel,GetState],ProcessKeyEventArgs->[OnKeyDown,OnKeyUp,OnKeyPress],WmMove->[UpdateBounds,DefWndProc],OnMove->[Invalidate],WmHelp->[OnHelpRequested,DefWndProc],HRESULT->[OnFrameWindowActivate,Load,IsInputKey,GetStyle,ProcessMnemonic,OnHelpRequested],OnParentBindingContextChanged->[OnBindingContextChanged],ScaleCore->[AssertLayoutSuspendCount,Scale,ResumeLayout],PreProcessMessage->[IsInputKey,GetExtendedState,IsInputChar],UpdateStylesCore->[Invalidate,SetState],OnTopMostActiveXParentChanged->[OnTopMostActiveXParentChanged],CanProcessMnemonic->[CanProcessMnemonic,TraceCanProcessMnemonic],WmShowWindow->[DefWndProc,SetState,OnVisibleChanged,GetState,GetTopLevel,CreateControl],ListenToUserPreferenceChanged->[GetExtendedState],WmMouseDown->[GetExtendedState,DefWndProc,SetState,OnMouseDown,GetStyle,Focus],OnPrint->[DefWndProc,GetStyle],OnResize->[Invalidate,GetState],UnhookMouseEvent->[SetState],OnBackgroundImageChanged->[GetAnyDisposingInHierarchy,Invalidate],UserPreferenceChanged->[OnSystemColorsChanged],WmMouseLeave->[OnMouseLeave,DefWndProc],Select->[Select],SetTopLevelInternal->[GetExtendedState,SetParentHandle,SetState,ListenToUserPreferenceChanged,GetTopLevel,CreateControl],OnBackgroundImageLayoutChanged->[GetAnyDisposingInHierarchy,Invalidate],WmSetFocus->[InvokeGotFocus,DefWndProc],IsHoveredWithMouse->[IsHoveredWithMouse],OnParentBackgroundImageChanged->[OnBackgroundImageChanged],Scale->[AssertLayoutSuspendCount,Scale,ResumeLayout],WmWindowPosChanged->[UpdateChildControlIndex,UpdateBounds,DefWndProc],UpdateBounds->[OnSizeChanged,OnClientSizeChanged,AdjustWindowRectEx,OnLocationChanged,UpdateBounds,GetTopLevel],WmDestroy->[DefWndProc,UnhookMouseEvent,SetState,ReleaseUiaProvider,OnHandleDestroyed,OnMouseLeave,GetState],WmGetControlType->[MarshalStringToMessage],MarshaledInvoke->[WaitForWaitHandle,InvokeMarshaledCallbacks],HookMouseEvent->[GetState],WmNotify->[ReflectMessage,DefWndProc],OnQueryContinueDrag->[OnQueryContinueDrag],ResetMouseEventArgs->[HookMouseEvent,GetState],CheckParentingCycle->[CheckParentingCycle],LogicalToDeviceUnits->[LogicalToDeviceUnits],InitializeDCForWmCtlColor->[GetStyle],OnHandleDestroyed->[GetAnyDisposingInHierarchy,ListenToUserPreferenceChanged,GetState],WmCommand->[ReflectMessage,DefWndProc],OnParentInvalidated->[Invalidate],OnSizeChanged->[OnResize],InvokeMarshaledCallbacks->[InvokeMarshaledCallback],PaintWithErrorHandling->[PaintException,GetStyle,OnPaint,Invalidate,OnPaintBackground,GetState],ScaleChildControls->[Scale],OnDragEnter->[OnDragEnter],OnChildLayoutResuming->[OnChildLayoutResuming],WmCreate->[DefWndProc,GetStyle,UpdateChildZOrder,UpdateBounds,OnHandleCreated],DrawToBitmap->[CreateHandle],OnEnabledChanged->[GetAnyDisposingInHierarchy,Invalidate,GetStyle],WmDisplayChange->[DefWndProc,Invalidate],OnInvalidated->[ActiveXViewChanged],WmGetObject->[DefWndProc],WmMouseWheel->[OnMouseWheel,DefWndProc],Font->[Font],ProcessCmdKey->[ProcessCmdKey],WmQueryNewPalette->[DefWndProc,Invalidate],ScaleControl->[Size,ScaleControl,AdjustWindowRectEx],PaintBackground->[RenderColorTransparent,PaintBackground],OnGiveFeedback->[OnGiveFeedback],WmContextMenu->[WmContextMenu,DefWndProc,Contains],PerformLayout->[GetAnyDisposingInHierarchy,GetExtendedState,OnLayout,PerformLayout,GetState],ShouldSerializeEnabled->[GetState],RecreateHandleCore->[OnParentHandleRecreated,DestroyHandle,OnParentHandleRecreating,Focus,CreateControl,GetState,CreateHandle],ChildGotFocus->[ChildGotFocus,ActiveXOnFocus],SetBounds->[Size,SetBoundsCore,InitScaling],ProcessDialogChar->[ProcessDialogChar],DestroyHandle->[DestroyHandle],OnLayout->[ActiveXViewChanged],WmEraseBkgnd->[PaintWithErrorHandling,DefWndProc,GetStyle],OnParentCursorChanged->[OnCursorChanged],WmClose->[DefWndProc],WmUpdateUIState->[OnChangeUICues,Invalidate,DefWndProc],EndUpdateInternal->[EndUpdateInternal],SetBoundsCore->[InitScaling,InitLayout,ResumeLayout,GetState],Save->[Save],SetParentHandle->[GetTopLevel,RecreateHandle],OnGotFocus->[ChildGotFocus,ActiveXOnFocus],WmMouseMove->[OnMouseMove,DefWndProc,GetStyle],SetClientSizeCore->[OnClientSizeChanged],OnPaddingChanged->[Invalidate,GetStyle],OnLostFocus->[ActiveXOnFocus],ShowsOwnToolTip->[ShowsOwnToolTip],Rectangle->[Size],OnParentBecameInvisible->[OnParentBecameInvisible],OnBackColorChanged->[GetAnyDisposingInHierarchy,Invalidate,GetState],PrintToMetaFileRecursive->[PrintToMetaFileRecursive],PerformControlValidation->[NotifyValidated,NotifyValidating],SuspendLayout->[OnLayoutSuspended],AllowsToolTip->[AllowsToolTip],AllowsChildrenToShowToolTips->[AllowsChildrenToShowToolTips],WmSetCursor->[DefWndProc],SelectNextIfFocused->[SelectNextControl],ProcessDialogKey->[ProcessDialogKey],EndInvoke->[WaitForWaitHandle],Size->[Size,LogicalToDeviceUnits,GetStyle,AdjustWindowRectEx],OnParentForeColorChanged->[OnForeColorChanged],UpdateWindowFontIfNeeded->[GetStyle],WmMouseEnter->[DefWndProc,OnMouseEnter],WmDpiChangedBeforeParent->[OnDpiChangedBeforeParent,Size,RescaleConstantsForDpi,DefWndProc],Invalidate->[Invalidate],AddReflectChild,AssertLayoutSuspendCount,Control,Size,DisposeFontHandle,GetState]] | Adjust window rect with style and display flags. | This API was introduced on 1607, you need APIHelper here. |
@@ -4102,7 +4102,7 @@ namespace ProtoAssociative
ProtoCore.DSASM.SymbolNode symnode = Allocate(classIndex, classScope, ProtoCore.DSASM.Constants.kGlobalScope, name, ptrType, datasize, isStatic, access);
if (null == symnode)
{
- buildStatus.LogSemanticError("Member variable '" + name + "' is already defined in class " + core.ClassTable.ClassNodes[classIndex].name);
+ buildStatus.LogSemanticError(Resource.memberVariable + name + Resource.alreadyDefined + core.ClassTable.ClassNodes[classIndex].name);
return ProtoCore.DSASM.Constants.kInvalidIndex;
}
| [CodeGen->[EmitDependency->[SetEntry,EmitJumpDependency],GetFirstSymbolFromIdentList->[GetFirstSymbolFromIdentList],DFSEmitSSA_AST->[EmitSSALHS,DFSEmitSSA_AST,SSAIdentList,GetReplicationGuides,EmitSSAArrayIndex],CyclicDependencyTest->[StrongConnectComponent],DfsSSAIeentList->[DfsSSAIeentList],EmitGetterSetterForIdentList->[EmitIdentifierNode,EmitBinaryExpressionNode],TraverseDotCallArguments->[PushSymbolAsDependent],EmitSSALHS->[EmitSSAArrayIndexRetainDimension],IdentifierNode->[EmitBinaryExpressionNode],EmitClassDeclNode->[EmitMemberVariables,FunctionDefinitionNode],EmitLHSIdentifierListForBinaryExpr->[AutoGenerateUpdateArgumentReference,AutoGenerateUpdateReference,EmitDependency],BuildSSA->[SetExecutionFlagForNode,DFSEmitSSA_AST,BuildSSA],InsertThisPointerAtBody->[BinaryExpressionNode,TraverseAndAppendThisPtrArg],EmitBinaryExpressionNode->[EmitDependency,Allocate,EmitLHSThisDotProperyForBinaryExpr,EmitLHSIdentifierListForBinaryExpr,GetUpdatedNodeRef,AutoGenerateUpdateReference,CyclicDependencyTest,EmitPostFixNode],ProcedureNode->[TraverseDotCallArguments,GetProcedureFromInstance,PushSymbolAsDependent,EmitFunctionCall],InferDFSTraverse->[InferDFSTraverse],AllocateMemberVariable->[Allocate],DfsTraverse->[EmitFunctionCallNode,EmitForLoopNode,EmitGroupExpressionNode,EmitLanguageBlockNode,EmitRangeExprNode,EmitClassDeclNode,EmitBinaryExpressionNode,EmitExceptionHandlingNode,EmitModifierStackNode,EmitFunctionDefinitionNode,EmitInlineConditionalNode,EmitConstructorDefinitionNode,EmitImportNode,EmitPostFixNode,EmitIdentifierNode,EmitIfStatementNode,EmitThrowNode,EmitUnaryExpressionNode,EmitDynamicBlockNode],EmitFunctionDefinitionNode->[AllocateArg],EmitSSAArrayIndexRetainDimension->[EmitSSAforArrayDimension],StrongConnectComponent->[IsDependentSubNode,StrongConnectComponent],EmitMemberVariables->[EmitGetterForProperty,EmitMemberVariables,AllocateMemberVariable,EmitSetterForProperty],EmitInlineConditionalNode->[EmitFunctionCallNode,EmitDynamicLanguageBlockNode,EmitIfStatementNode],AllocateContextGlobals->[Allocate],TraverseAndAppendThisPtrArg->[BinaryExpressionNode,TraverseAndAppendThisPtrArg],EmitSSA->[BuildSSA,SplitMulitpleAssignment],Emit->[SetDeltaCompilePC,AllocateContextGlobals,SplitMulitpleAssignment,EmitExpressionInterpreter,BuildSSA],GetReplicationGuides->[GetReplicationGuides],EmitConstructorDefinitionNode->[EmitRetc,EmitAllocc,AllocateArg,EmitCallingForBaseConstructor],EmitImportNode->[BuildSSA,SplitMulitpleAssignment],SSAIdentList->[SSAIdentList],EmitIdentifierNode->[Allocate,EmitDependency,PushSymbolAsDependent],EmitAllocc->[SetEntry],SplitMulitpleAssignment->[SplitMulitpleAssignment],RemoveReplicationGuides->[RemoveReplicationGuides]]] | Allocate a member variable. | MemberVariableRedefinitionError = "Member variable '{0}' is already defined in class {1}" |
@@ -953,7 +953,7 @@ _c;
}
[Test]
- [Category("Failing")]
+ [Category("Failure")]
[Category("SmokeTest")]
public void T46_TestBooleanOperationOnNull()
{
| [Assignment->[T66_Imperative_1467368_negative->[RunScriptSource,Throws],T17_TestInvalidSyntax->[RunScriptSource,Throws],T23_TestUsingMathAndLogicalExpr->[RunScriptSource,Payload,IsTrue],T40_Defect_1450552->[kIdUnboundIdentifier,IsTrue,RunScriptSource,VerifyBuildWarning,IsNull],T44__Defect_1452423_4->[RunScriptSource,IsNull,IsTrue],T56_Defect_1454691->[RunScriptSource,Verify],T60_Defect_1455590_2->[RunScriptSource,Verify],T63_Defect_1452643->[RunScriptSource,Throws],T62_Defect_1456721->[RunScriptSource,IsNull,Verify,IsTrue],T61_TestBooleanOperationOnNull->[RunScriptSource,Verify],T67_DNL_1467458_4->[RunScriptSource,Verify],T68_DNL_1467523->[RunScriptSource,Verify],T04_TestAssignmentStmtInExpression_negative->[RunScriptSource,Throws],T42__Defect_1452423_2->[RunScriptSource,IsNull,IsTrue],T01_SampleTestUsingCodeWithinTestFunction->[RunScriptSource,GetValue,Payload,IsTrue],T66_Associative_1467368_negative_2->[RunScriptSource,Throws],T43__Defect_1452423_3->[RunScriptSource,Payload,IsTrue],T31_Defect_1449877->[RunScriptSource,ToInt64,Payload,IsTrue],T29_Defect_1449887->[RunScriptSource,Payload,IsTrue],T14_TestUsingMathAndLogicalExpr->[Payload,IsTrue,RunScriptSource,ToInt64,ToBoolean],T46_TestBooleanOperationOnNull->[RunScriptSource,IsNull,Payload,IsTrue],T48_MultipleAssignments->[RunScriptSource,Verify],T19_TestAssignmentToCollection->[RunScriptSource,Payload,IsTrue],T16_TestInvalidSyntax->[RunScriptSource,Throws],T64_Defect_1450715->[RunScriptSource,Verify],T52_Defect_1449889->[RunScriptSource,IsNull,Verify,IsTrue],T57_Defect_1454691_2->[RunScriptSource,Verify],T27_Defect_1450847->[RunScriptSource,IsNull,Payload,IsTrue],T39_Defect_1449704->[RunScriptSource,IsNull,IsTrue],T20_TestInvalidSyntax->[RunScriptSource,Payload,IsTrue],T10_TestInFunctionScope->[RunScriptSource,Payload,IsTrue],T50_Defect_1449889->[RunScriptSource,Verify],T15_TestInRecursiveFunctionScope->[RunScriptSource,Payload,IsTrue],T12_TestUsingMathAndLogicalExpr->[RunScriptSource,Payload,ToDouble,IsTrue],T53_Defect_1454691->[RunScriptSource,Verify],T11_TestInClassScope->[RunScriptSource,Payload,IsTrue],T05_TestRepeatedAssignment_negative->[RunScriptSource,Payload,IsTrue],T37_TestOperationOnNullAndBool->[RunScriptSource,IsNull,IsTrue],T54_Defect_1454691->[RunScriptSource,Verify],T26_Defect_1450854->[RunScriptSource,Payload,IsTrue],T38_Defect_1449928->[RunScriptSource,Payload,IsTrue],T59_Defect_1455590->[RunScriptSource,Verify],T36_Defect_1450555->[RunScriptSource,Payload,IsTrue],T32_Defect_1449877_2->[RunScriptSource,Payload,IsTrue],T47_TestBooleanOperationOnNull->[RunScriptSource,Payload,IsTrue],T67_DNL_1467458->[RunScriptSource,Verify],T35_Defect_1450727_2->[RunScriptSource,Payload,IsTrue],T02_SampleTestUsingCodeFromExternalFile->[RunScriptSource,Payload,IsTrue],T58_Defect_1454691_3->[RunScriptSource,Verify],T45__Defect_1452423_5->[RunScriptSource,IsNull,IsTrue],T18_TestMethodCallInExpr->[RunScriptSource,Payload,IsTrue],T09_TestInNestedBlock->[RunScriptSource,IsNull,Payload,IsTrue],T51_Assignment_Using_Negative_Index->[RunScriptSource,IsNull,Verify,IsTrue],T06_TestInUnnamedBlock_negative->[RunScriptSource,Throws],T55_Defect_1454691->[RunScriptSource,Verify],T07_TestOutsideBlock->[RunScriptSource,Payload,IsTrue],T67_DNL_1467458_3->[RunScriptSource,Verify],T65_Operation_On_Null->[RunScriptSource,Verify],T22_TestAssignmentToNegativeNumbers->[RunScriptSource,Payload,IsTrue],T08_TestCyclicReference->[VerifyRunScriptSource,Verify],T34_Defect_1450727->[RunScriptSource,Payload,IsTrue],T41__Defect_1452423->[RunScriptSource,IsNull,IsTrue],T49_TestForStringObjectType->[RunScriptSource,Verify],T24_TestUsingMathematicalExpr->[RunScriptSource,Verify],T13_TestUsingMathAndLogicalExpr->[RunScriptSource,Payload,IsTrue],T03_TestAssignmentToUndefinedVariables_negative->[kIdUnboundIdentifier,IsTrue,RunScriptSource,VerifyBuildWarning,IsNull],T30_Defect_1449887_2->[RunScriptSource,Payload,IsTrue],T21_TestAssignmentToBool->[RunScriptSource,Payload,ToBoolean,IsTrue],T33_Defect_1450003->[RunScriptSource,Verify],T25_TestUsingMathematicalExpr->[RunScriptSource,Payload,IsTrue]]] | Tests if a condition is true on null. | Failing tests can't be categorized as Smoke Test, consider removing this category from all failing tests. In some tests we are using "SmokeTests" as category and we need to unify this as well. You might need to update "SmokeTest/Smoke Test" to "SmokeTests" |
@@ -176,12 +176,14 @@ func initState(log *logp.Logger, c loginp.Cursor, s fileSource) state {
return state
}
-func (inp *filestream) open(log *logp.Logger, canceler input.Canceler, path string, offset int64) (reader.Reader, error) {
+func (inp *filestream) open(log *logp.Logger, canceler input.Canceler, srcID, path string, offset int64) (reader.Reader, error) {
f, err := inp.openFile(log, path, offset)
if err != nil {
return nil, err
}
+ inp.startMonitoring(log, canceler, srcID, path, f)
+
log.Debug("newLogFileReader with config.MaxBytes:", inp.readerConfig.MaxBytes)
// TODO: NewLineReader uses additional buffering to deal with encoding and testing
| [Test->[open,Close,Errorf],initFileOffset->[Seek],isDroppedLine->[Debug],readFromSource->[Infof,IsEmpty,Next,isDroppedLine,eventFromMessage,Errorf,Publish,Info,Err],open->[AppendReaders,NewLimitReader,NewStripNewline,Close,NewEncodeReader,NewFilemeta,openFile,Debug],Run->[Name,WithFunc,Close,readFromSource,Errorf,With,open,Debug],openFile->[FileMode,IfNot,Size,encodingFactory,Stat,Infof,IgnoreError,Name,initFileOffset,Close,Errorf,OpenFile,Mode],Unpack,Error,MatchString,IsRegular,Name,FindEncoding,Errorf,IsNew,Mode] | open opens a file and returns a reader that reads the file and reads the file meta data readfile. Reader. | alternatively to actively probing the file with `fstat`, we could define a function based metric to save CPU. The function based metric will query the size only on demand, and only if the last file size update is >10s. |
@@ -285,6 +285,12 @@ bool itl_generator::gen_struct(AST_Structure*, UTL_ScopedName* /* name */,
<< Open(this)
<< Indent(this) << "\"kind\" : \"alias\",\n"
<< Indent(this) << "\"name\" : \"" << repoid << "\",\n"
+
+ // Check if this is defined as a primary data type
+ << Indent(this) << "\"note\" : { \"is_dcps_data_type\" : "
+ << ((idl_global->is_dcps_type(name) == 0) ? "false" : "true")
+ << " },\n"
+
<< Indent(this) << "\"type\" :\n"
<< Open(this)
<< Indent(this) << "{\n"
| [gen_typedef->[new_type,InlineType],gen_union->[new_type,InlineType],gen_struct->[new_type,InlineType],gen_enum->[new_type]] | Generate an ITL for a structure with a specific type. Returns a new object that represents the result of a call to the method. | This could be simplified by swapping true/false so you don't need == 0 and extra parens |
@@ -453,6 +453,9 @@ class ProductVariant(ModelWithMetadata):
images = list(self.images.all())
return images[0] if images else self.product.get_first_image()
+ def get_ordering_queryset(self):
+ return self.product.variants.all()
+
class ProductVariantTranslation(models.Model):
language_code = models.CharField(max_length=10)
| [ProductVariant->[is_digital->[is_shipping_required],get_first_image->[get_first_image]],DigitalContent->[create_new_url->[create]],BaseAttributeQuerySet->[get_visible_to_user->[user_has_access_to_all,get_public_attributes]],ProductsQueryset->[visible_to_user->[published_with_variants]],AttributeQuerySet->[variant_attributes_sorted->[_get_sorted_m2m_field],product_attributes_sorted->[_get_sorted_m2m_field]]] | Get first image in the product. | Why do we add this function which is never called? |
@@ -1,5 +1,7 @@
module Constants
module SiteConfig
+ IMAGE_PLACEHOLDER = "https://url/image.png".freeze
+
DETAILS = {
allow_email_password_registration: {
description: "People can sign up using their email and password",
| [freeze] | The configuration for the user s account. Missing parameters for missing parameters. | I love this change! Also love that this now sets the precedent for us to abstract _more_ of this file into constants, too! |
@@ -38,7 +38,10 @@ class Spfft(CMakePackage):
if self.spec.satisfies('+gpu_direct'):
args += ["-DSPFFT_GPU_DIRECT=On"]
if self.spec.satisfies('+cuda'):
- args += ["-DSPFFT_BACKEND=CUDA"]
- if self.spec.satisfies('+cuda'):
- args += ["-DSPFFT_BACKEND=CUDA"]
+ args += ["-DSPFFT_GPU_BACKEND=CUDA"]
+ if self.spec.satisfies('+fortran'):
+ args += ["-DSPFFT_FORTAN=On"]
+ if self.spec.satisfies('+static'):
+ args += ["-DSPFFT_STATIC=On"]
+
return args
| [Spfft->[cmake_args->[satisfies],variant,depends_on,version]] | Return a list of arguments to pass to the DSPFFT command. | Was this wrong before or is the option dependent on the version of the package used (`-DSPFFT_BACKEND` vs. `-DSPFFT_GPU_BACKEND`? |
@@ -83,3 +83,14 @@ export function setPrivateMessageRecipient(participant: Object) {
type: SET_PRIVATE_MESSAGE_RECIPIENT
};
}
+
+/**
+ * Toggles display of the chat panel.
+ *
+ * @returns {Function}
+ */
+export function toggleChat() {
+ return function(dispatch: (Object) => Object) {
+ dispatch({ type: TOGGLE_CHAT });
+ };
+}
| [No CFG could be retrieved] | The set_private_message_reCIPIENT element. | Why did you make this action async? |
@@ -66,7 +66,6 @@ class AnnotationCategoriesController < ApplicationController
def update
@assignment = Assignment.find(params[:assignment_id])
@annotation_category = @assignment.annotation_categories.find(params[:id])
-
if @annotation_category.update(annotation_category_params)
flash_message(:success, t('.success'))
else
| [AnnotationCategoriesController->[create_annotation_text->[new],destroy->[destroy],destroy_annotation_text->[destroy],update->[update],create->[new],update_annotation_text->[update],update_positions->[update]]] | update - update the nag - item. | Revert this change. |
@@ -193,6 +193,8 @@ class SearchTest(unittest.TestCase):
arch: x86
compiler: gcc
compiler.version: 4.3
+ [recipe hash]
+ Not present (created with conan < 0.15)
""", self.client.user_io.out)
| [SearchTest->[package_search_properties_filter_test->[assertNotIn,assertEquals,assertIn,run],package_search_with_invalid_reference_test->[str,assertIn,run],setUp->[TestClient,save,assertIn,run],package_search_with_empty_query_test->[assertIn,run],recipe_search_test->[assertEquals,run],package_search_with_invalid_query_test->[assertIn,run]]] | This test tests that the user s properties are not in the correct order. Checks that the existing packages for the given reference are not already in the system. Find the first package that matches the query. | I would say that this is a bit verbose and uninformative. What can I do with such hash?. What does it means?. I'd say that a single line something like: `outdated from recipe: True/False` could be better. |
@@ -184,6 +184,7 @@ class Jetpack_Tracks_Client {
}
return array(
+ 'blogid' => Jetpack_Options::get_option( 'id', 0 ),
'userid' => $user_data['ID'],
'username' => $user_data['login'],
);
| [Jetpack_Tracks_Client->[build_pixel_url->[build_pixel_url]]] | Get the connected user s tracks identity. | Nit: now that we're here, should we align those `=>` vertically according to our coding standards? |
@@ -171,7 +171,7 @@ class SplitPdfJob < ApplicationJob
split_page.update(status: status)
end
- group.access_repo do |repo|
+ grouping.access_repo do |repo|
assignment_folder = exam_template.assignment.repository_folder
txn = repo.get_transaction(Admin.first.user_name)
| [SplitPdfJob->[perform->[base_path,create,new,from_blob,decode,join,mkdir_p,rows,name,each_index,nil?,exists?,save_pages,quality,log,crop,pages,group_name_for,load,to_i,instance,find_or_create_by,filename,update,read,density,each,length,id,match,delete,write,raise,columns,total,increment,save],match_student->[find_by],save_pages->[base_path,crop_x,new,to_pdf,from_blob,start,any?,join,first,capture2,python,mkdir_p,name,assignment,rows,nil?,exists?,crop_width,root,sort_by!,end,replace,quality,crop,group_name_for,reject,commit,label,empty?,to_i,dirname,access_repo,find_or_create_by,update,each_with_index,automatic_parsing,density,num_pages,get_transaction,each,repository_folder,length,id,num_cover_fields,split,write,user_name,match_student,crop_height,crop_y,columns,find_by,collect,add,save],show_status->[t],get_num_groups_in_dir->[exists?,foreach,directory?,join,start_with?],group_name_for->[name],queue_as,before_enqueue,split_pdf,update,arguments,name]] | save_pages saves the pages of a template that have a missing page number in the group add page if page_num already exists finds a in the assignment_folder that contains the extra_pdf cover_pdf find_n_complete - finds n_complete - finds n_complete . | Metrics/BlockLength: Block has too many lines. [106/25] |
@@ -87,7 +87,7 @@ class ArticlesController < ApplicationController
begin
fixed_body_markdown = MarkdownFixer.fix_for_preview(params[:article_body])
parsed = FrontMatterParser::Parser.new(:md).call(fixed_body_markdown)
- parsed_markdown = MarkdownParser.new(parsed.content, source: Article.new(user: current_user))
+ parsed_markdown = MarkdownParser.new(parsed.content, source: Article.new(user: current_user), user: current_user)
processed_html = parsed_markdown.finalize
rescue StandardError => e
@article = Article.new(body_markdown: params[:article_body])
| [ArticlesController->[preview->[new],update->[update]]] | preview a single tag. | Here's context in action! The `source` is **where** the liquid tag will be used and `user` is **who** is using it. |
@@ -133,10 +133,13 @@ namespace epee
template<typename T>
byte_slice::byte_slice(const adapt_buffer, T&& buffer)
- : storage_(nullptr), portion_(to_byte_span(to_span(buffer)))
+ : storage_(nullptr), portion_(nullptr)
{
if (!buffer.empty())
+ {
storage_ = allocate_slice<adapted_byte_slice<T>>(0, std::move(buffer));
+ portion_ = to_byte_span(to_span(static_cast<adapted_byte_slice<T> *>(storage_.get())->buffer));
+ }
}
byte_slice::byte_slice(std::initializer_list<span<const std::uint8_t>> sources)
| [remove_prefix->[remove_prefix],call->[~byte_slice_data],byte_buffer_increase->[byte_buffer_resize],take_slice->[remove_prefix]] | byte_slice - > byte_slice. | Explicit initialize `portion_(nullptr)`. Per C++ rules, its going to be initialized anyway to the same value, but it doesn't "look" possibly uninitialized that way. |
@@ -1051,6 +1051,17 @@ class GlobalOptions(Subsystem):
help="Number of times to retry any RPC to the remote store before giving up.",
)
+ register(
+ "--remote-cache-warnings",
+ type=RemoteCacheWarningsBehavior,
+ default=DEFAULT_EXECUTION_OPTIONS.remote_cache_warnings,
+ advanced=True,
+ help=(
+ "Whether to log remote cache failures at the `warn` log level.\n\n"
+ "All errors not logged at the `warn` level will instead be logged at the "
+ "`debug` level."
+ ),
+ )
register(
"--remote-cache-eager-fetch",
type=bool,
| [OwnersNotFoundBehavior->[to_glob_match_error_behavior->[GlobMatchErrorBehavior]],FilesNotFoundBehavior->[to_glob_match_error_behavior->[GlobMatchErrorBehavior]],GlobalOptions->[register_options->[register_bootstrap_options],compute_pants_ignore->[add],validate_instance->[validate_remote_headers,validate_remote_address]],LocalStoreOptions,ExecutionOptions] | Register options that are required to create a Scheduler. Register a warning or error level for a single log - level logging target. Registers a sequence number with the command line interface. Register a sequence number that can be used to satisfy a specific constraint. | I debated explaining what option does, but `ignore` and `first_only` seem pretty self-explanatory. `backoff` could maybe use an explanation, although I doubt we want to mention the implementation detail of using the exponent of 2. |
@@ -3,6 +3,7 @@ class ChannelWorksController < ApplicationController
def index(page: nil)
@works = current_user.works.wanna_watch_and_watching
+ .published
.program_registered
.includes(:episodes)
.order(released_at: :desc)
| [ChannelWorksController->[index->[page],before_action]] | index all works that have been wanna watch and have been watched. | Place the . on the previous line, together with the method call receiver.<br>Align the operands of an expression in an assignment spanning multiple lines. |
@@ -27,6 +27,9 @@ class Mesa(MesonPackage):
depends_on('meson@0.45:', type='build')
depends_on('binutils', type='build')
+ depends_on('bison', type='build')
+ depends_on('flex', type='build')
+ depends_on('gettext', type='build')
depends_on('pkgconfig', type='build')
depends_on('python@3:', type='build')
depends_on('py-mako@0.8.0:', type='build')
| [Mesa->[meson_args->[append,SpecError,join,extend],depends_on,any_combination_of,patch,startswith,version,provides,variant]] | Creates a new object. returns a variant object for the given node. | I see the bison and flex dependency in the meson build file but where is the gettext dependency coming in? (trying to figure out if it's explicitly called for in the meson build or if ti's implicitly required, which in turn would men a bug in upstream for it not being explicit. |
@@ -488,7 +488,7 @@ class UM3OutputDevicePlugin(OutputDevicePlugin):
return True
## Check if the prerequsites are in place to start the cloud flow
- def checkCloudFlowIsPossible(self) -> None:
+ def checkCloudFlowIsPossible(self, cluster: Optional["CloudOutputDevice"]) -> None:
Logger.log("d", "Checking if cloud connection is possible...")
# Pre-Check: Skip if active machine already has been cloud connected or you said don't ask again
| [UM3OutputDevicePlugin->[associateActiveMachineWithPrinterDevice->[refreshConnections],_onMachineSwitched->[checkCloudFlowIsPossible],stop->[stop],_onNetworkRequestFinished->[removeManualDevice],removeManualDevice->[resetLastManualDevice],start->[start],startDiscovery->[resetLastManualDevice],_createMachineFromDiscoveredPrinter->[refreshConnections]]] | Checks if a cloud connection is possible. | You'll need to import `CloudOutputDevice` at the top of the file. |
@@ -27,6 +27,9 @@ switch ($op) {
case 'submission':
case 'submissionRegrets':
case 'submissionReview':
+ /* https://github.com/ubiquitypress/OJS-Draft-Editorial/commit/9135893848e9d40923c105a2cafc59b0529659c9 */
+ case 'draftDecision':
+ case 'viewDraftDecision':
case 'submissionEditing':
case 'submissionHistory':
case 'submissionCitations':
| [No CFG could be retrieved] | Provides a function to handle all of the functions that are defined in the page editor. Private methods for all of the methods that are defined in the module. | I'd rather not leave references like this scattered throughout the code. |
@@ -788,7 +788,7 @@ public class ClassicPluginStrategy implements PluginStrategy {
if(p!=null)
try {
return p.classLoader.loadClass(name);
- } catch (ClassNotFoundException _) {
+ } catch (ClassNotFoundException ex) {
// try next
}
}
| [ClassicPluginStrategy->[createPluginWrapper->[loadLinkedManifest,isLinked],updateDependency->[getShortName],load->[getShortName],parseClassPath->[resolve],createClassLoader->[createClassLoader],AntClassLoader2->[defineClassFromData->[defineClassFromData]],createClassJarFromWebInfClasses->[putNextEntry->[putNextEntry],zipDir->[zipDir]],DependencyClassLoader->[findClass->[getTransitiveDependencies],findResources->[getTransitiveDependencies],findResource->[getTransitiveDependencies]],findAncestorDependencyClassLoader->[findAncestorDependencyClassLoader]]] | Find a class by name. | Would prefer `ignored` so IntelliJ understands it. |
@@ -184,15 +184,15 @@ func (eb *ethBroadcaster) processUnstartedEthTxs(fromAddress gethCommon.Address)
return nil
}
n++
- attempt, err := newAttempt(eb.store, *etx, eb.config.EthGasPriceDefault())
+ a, err := newAttempt(eb.store, *etx, eb.config.EthGasPriceDefault())
if err != nil {
return errors.Wrap(err, "processUnstartedEthTxs failed")
}
- if err := eb.saveInProgressTransaction(etx, &attempt); err != nil {
+ if err := eb.saveInProgressTransaction(etx, &a); err != nil {
return errors.Wrap(err, "processUnstartedEthTxs failed")
}
- if err := eb.handleInProgressEthTx(*etx, attempt, true); err != nil {
+ if err := eb.handleInProgressEthTx(*etx, a, true); err != nil {
return errors.Wrap(err, "processUnstartedEthTxs failed")
}
}
| [tryAgainWithHigherGasPrice->[handleInProgressEthTx]] | processUnstartedEthTxs processes any unstarted transactions that are not yet started. | Same here, the initial name is better! |
@@ -47,6 +47,11 @@ public class AgentInitializer {
private static ClassLoader AGENT_CLASSLOADER = null;
public static void initialize(Instrumentation inst, URL bootstrapUrl) {
+ if (AGENT_CLASSLOADER != null) {
+ // don't run initialization multiple times, initialize can be called multiple times via dynamic attach
+ return;
+ }
+
startAgent(inst, bootstrapUrl);
boolean appUsingCustomLogManager = isAppUsingCustomLogManager();
| [AgentInitializer->[isJavaBefore9WithJfr->[isJavaBefore9]]] | Initializes the agent. | Just confirming, do we need the field to be volatile if accessed multiple times? With dynamic attach, is `initialize` guaranteed to read the memory from a previous update? |
@@ -1039,6 +1039,8 @@ func (cache *DiskBlockCacheLocal) evictLocked(ctx context.Context,
// blockIDs is a slice of blocks from which evictions will be selected.
blockIDs := make(blockIDsByTime, 0, numElements)
+ cache.log.CDebugf(ctx, "There are %d TLFs in priority \"%s\" for eviction", len(shuffledSlice), priorityToEvict)
+
// For each TLF until we get enough elements to select among,
// add its blocks to the eviction slice.
for _, tlfIDStruct := range shuffledSlice {
| [GetLastUnrefRev->[checkCacheLocked],evictUntilBytesAvailable->[useLimiter],evictSomeBlocks->[deleteLocked],PutLastUnrefRev->[Put,encodeLastUnref,checkCacheLocked],evictLocked->[evictSomeBlocks,getLRULocked,shuffleTLFsAtPriorityWeighted,getRandomBlockID],Shutdown->[getCurrBytes,Shutdown,useLimiter],Mark->[updateMetadataLocked,getMetadataLocked,checkCacheLocked],DeleteUnmarked->[tlfKey,deleteNextUnmarkedBatchFromTlf],deleteLocked->[subCurrBytes,tlfKey,Get,useLimiter],deleteNextBatchFromClearedTlf->[evictFromTLFLocked,checkCacheLocked],UpdateMetadata->[updateMetadataLocked,getMetadataLocked,checkCacheLocked],getLRULocked->[getMetadataLocked],Get->[updateMetadataLocked,getMetadataLocked,Get,checkCacheLocked,decodeBlockCacheEntry],Delete->[deleteLocked,checkCacheLocked],Status->[getCurrBytes],Put->[getMetadataLocked,evictUntilBytesAvailable,useLimiter,checkCacheLocked,updateMetadataLocked,tlfKey,Put,encodeBlockCacheEntry,addCurrBytes],GetMetadata->[getMetadataLocked,checkCacheLocked],evictFromTLFLocked->[evictSomeBlocks,getLRULocked,getRandomBlockID],ClearAllTlfBlocks->[deleteNextBatchFromClearedTlf],syncBlockCountsAndUnrefsFromDb->[decodeLastUnref,setCurrBytes],deleteNextUnmarkedBatchFromTlf->[deleteLocked,getMetadataLocked,checkCacheLocked]] | evictLocked evicts the specified number of blocks from the cache. It returns the number This function is used to iterate over all the elements in the cache and add them to the. | If the cache is full and we're evicting each time we insert something, this line will be printed a lot of times. Are we okay with that? Is that going to pollute the logs too much? |
@@ -341,7 +341,7 @@ class VersionSpec(object):
elif spec.startswith(('=', '<', '>', '!')):
m = version_relation_re.match(spec)
if m is None:
- raise CondaRuntimeError('Invalid version spec: %s' % spec)
+ raise CondaDependencyError('Invalid version spec: %s' % spec)
op, b = m.groups()
self.op = opdict[op]
self.cmp = VersionOrder(b)
| [VersionSpec->[__new__->[VersionSpec,startswith,__new__,VersionOrder],__repr__->[str],__or__->[VersionSpec],__and__->[VersionSpec],str->[str],veval_match_->[VersionOrder],__str__->[str]],VersionOrder->[startswith->[_eq,startswith],__new__->[__new__],__eq__->[_eq]]] | Create a new object from a version spec. self. match = self. exact_match_ . | We're updating the MatchSpec object in another pull request. Just make this `CondaError` for now. |
@@ -33,6 +33,10 @@ func TestYarnInstall(t *testing.T) {
}
func testInstall(t *testing.T, expectedBin string) {
+ // Skip for community PRs
+ if os.Getenv("COMMUNITY_PR") == "true" {
+ t.Skip("Skipped for community PRs")
+ }
// Skip during short test runs since this test involves downloading dependencies.
if testing.Short() {
t.Skip("Skipped in short test run")
| [RemoveAll,Join,Equal,Short,Setenv,TempDir,NoError,Mkdir,Skip,Chdir,WriteFile] | Test the given and test if it is installed. Package. json is a JSON file with the dependencies installed on the local machine. | Related to the other feedback - it would be great if we could keep the codebase itself independt of the notion of "community PR" (which is really a test/CI construct). Ideally the tests refer to semantically meaningful environment variables or parameterization - and the CI environment maps things that make sense at that layer into these in a way that makes sense. What exactly is it that this test depends on? Can we make the skip behaviour here more directly tied to that? |
@@ -24,7 +24,7 @@ public class ResolverSetupCleanup {
public void setup() throws Exception {
workDir = initWorkDir();
repoHome = IoUtils.mkdirs(workDir.resolve("repo"));
- resolver = initResolver();
+ resolver = initResolver(null);
repo = TsRepoBuilder.getInstance(resolver, workDir);
}
| [ResolverSetupCleanup->[install->[install],initWorkDir->[createRandomTmpDir],cleanup->[cleanWorkDir,recursiveDelete],newJar->[resolve,toString,TsJar],initResolver->[build,BootstrapAppModelResolver],setup->[resolve,initWorkDir,mkdirs,getInstance,initResolver]]] | Initializes the repository and the work directory. | We really want to pass null here? |
@@ -138,9 +138,8 @@ def is_overlapping_types(t: Type, s: Type, use_promotions: bool = False) -> bool
# We must check for TupleTypes before Instances, since Tuple[A, ...]
# is an Instance
- tup_overlap = is_overlapping_tuples(t, s, use_promotions)
- if tup_overlap is not None:
- return tup_overlap
+ if is_overlapping_tuples(t, s, use_promotions):
+ return True
if isinstance(t, Instance):
if isinstance(s, Instance):
| [is_overlapping_types->[is_overlapping_types],is_overlapping_tuples->[is_overlapping_types],narrow_declared_type->[narrow_declared_type,meet_types],meet_similar_callables->[meet_types],is_partially_overlapping_types->[is_object],TypeMeetVisitor->[meet->[meet_types],visit_union_type->[meet_types],visit_tuple_type->[meet_types],visit_instance->[meet_types]]] | Checks if t and s are overlapping. Check if a type t or type s has a cunk. Checks if the type s is not an instance union or tuple type and if so returns True. | I think this change is now unnecessary (and the corresponding change below) after you added the check at the top of this function. |
@@ -48,9 +48,8 @@
// If we're redirected to login, our
// previousState is already set in the authExpiredInterceptor. When login succesful go to stored state
- if ($rootScope.redirected && $rootScope.previousStateName) {
- $state.go($rootScope.previousStateName, $rootScope.previousStateParams);
- $rootScope.redirected = false;
+ if ($localStorage.previousStateName) {
+ $state.go($localStorage.previousStateName, $localStorage.previousStateParams);
} else {
$rootScope.$broadcast('authenticationSuccess');
}
| [No CFG could be retrieved] | Login to the user and go to the previous state if login is successful. | I forgot why I added this but it was done for some reason. Are you sure everything works fine without this? |
@@ -223,11 +223,17 @@ public class HoodieTableMetadataUtil {
return convertFilesToRecords(partitionToDeletedFiles, partitionToAppendedFiles, instantTime, "Restore");
}
- public static List<HoodieRecord> convertMetadataToRecords(HoodieRollbackMetadata rollbackMetadata, String instantTime, Option<String> lastSyncTs) {
+ public static List<HoodieRecord> convertMetadataToRecords(HoodieRollbackMetadata rollbackMetadata, String instantTime,
+ Option<String> lastSyncTs, boolean wasSynced) {
Map<String, Map<String, Long>> partitionToAppendedFiles = new HashMap<>();
Map<String, List<String>> partitionToDeletedFiles = new HashMap<>();
processRollbackMetadata(rollbackMetadata, partitionToDeletedFiles, partitionToAppendedFiles, lastSyncTs);
+ if (!wasSynced) {
+ // Since the instant-being-rolled-back was never committed to the metadata table, the files added there
+ // need not be deleted. For MOR Table, the rollback appends logBlocks so we need to keep the appended files.
+ partitionToDeletedFiles.clear();
+ }
return convertFilesToRecords(partitionToDeletedFiles, partitionToAppendedFiles, instantTime, "Rollback");
}
| [HoodieTableMetadataUtil->[convertFilesToRecords->[remove,checkState,getOrDefault,of,size,removeAll,createPartitionFilesRecord,info,add,empty,emptyList,forEach,containsKey],convertMetadataToRecords->[size,equals,createPartitionFilesRecord,getTotalWriteBytes,createPartitionListRecord,warn,info,containsKey,put,forEach,of,getOperationType,processRollbackMetadata,empty,getSuccessDeleteFiles,checkState,substring,length,convertFilesToRecords,toList,collect,add,getPath],convertInstantToMetaRecords->[getAction,convertMetadataToRecords,of,fromBytes,HoodieException,checkArgument,isCompleted,getTimestamp,get,deserializeHoodieRestoreMetadata,empty,getActiveTimeline,getCleanerMetadata,deserializeHoodieRollbackMetadata],processRollbackMetadata->[getName,getRollbackLogFiles,toList,isPresent,info,merge,get,compareTimestamps,sum,getWrittenLogFiles,format,isEmpty,collect,forEach,addAll,getPartitionPath,containsKey,put],getLogger]] | Convert HoodieRestoreMetadata to records. | nts: need to revisit again with rollback/restore issues fixed |
@@ -436,7 +436,7 @@ class EnvironmentModifications(object):
env.prepend_path(x, item)
else:
# We just need to set the variable to the new value
- env.set(x, env_after[x])
+ env.set(x, after)
return env
| [EnvironmentModifications->[append_path->[AppendPath,_get_outside_caller_attributes],append_flags->[AppendFlagsEnv,_get_outside_caller_attributes],from_sourcing_file->[return_separator_if_any,append_path,prepend_path,remove_path,EnvironmentModifications,set,unset,extend],set_path->[SetPath,_get_outside_caller_attributes],prepend_path->[PrependPath,_get_outside_caller_attributes],remove_path->[RemovePath,_get_outside_caller_attributes],set->[SetEnv,_get_outside_caller_attributes],unset->[UnsetEnv,_get_outside_caller_attributes],clear->[clear],extend->[extend],apply_modifications->[group_by_name,execute]],validate->[group_by_name,set_or_unset_not_first],inspect_path->[EnvironmentModifications,prepend_path]] | Creates a modifications that would be made by sourcing a file. This function checks if a file in the system is sourcing a and if so path - > path - > path - > path - > path - > path - > Concatenates an iterable of paths into a string of paths separated by separator defaulting to. | OK, but between reading the messages, it's already done and pushed. |
@@ -354,6 +354,8 @@ function baseExtend(dst, objs, deep) {
if (deep && isObject(src)) {
if (isDate(src)) {
dst[key] = new Date(src.valueOf());
+ } else if (isElement(src)) {
+ dst[key] = src[0] ? jqLite(src).clone()[0] : jqLite(src).clone();
} else {
if (!isObject(dst[key])) dst[key] = isArray(src) ? [] : {};
baseExtend(dst[key], [src], true);
| [No CFG could be retrieved] | Creates a unique alpha - numeric string based on the hashkey of an object. Extend src object with dst object. | A little late to the party, but shouldn't it be the other way around? Namely: `!src[0] ? jqLite(src).clone()[0] : jqLite(src).clone();` (In fact, I would rather expect: `!src[0] ? jqLite(src).clone()[0] : src.clone();`) |
@@ -0,0 +1,12 @@
+from django.conf.urls import url
+
+from . import views
+
+lang_urlpatterns = [
+ url(r'^confirmation/success/?$',
+ views.contribute_confirmation, {'status': 'succeeded'},
+ name='contribute_confirmation_succeeded'),
+ url(r'^confirmation/error/?$',
+ views.contribute_confirmation, {'status': 'error'},
+ name='contribute_confirmation_error')
+]
| [No CFG could be retrieved] | No Summary Found. | I think a confirmation screen used to be part of the process, but it isn't anymore. The URL should simplify from ``/contribute/confirmation/success`` to ``/contribute/success``. |
@@ -399,8 +399,13 @@ def show_experiments(
include_params=include_params,
exclude_params=exclude_params,
)
- metric_headers = _normalize_headers(metric_names)
- param_headers = _normalize_headers(param_names)
+
+ names = {**metric_names, **param_names}
+ counter = Counter(
+ name for path in names for name in names[path] for path in names
+ )
+ metric_headers = _normalize_headers(metric_names, counter)
+ param_headers = _normalize_headers(param_names, counter)
td = experiments_table(
all_experiments,
| [_sort_exp->[_sort->[_sort]],add_parser->[add_parser],CmdExperimentsShow->[run->[show_experiments]],show_experiments->[_parse_filter_list,experiments_table,_collect_names],_filter_names->[_filter_name],CmdExperimentsRun->[run->[run]],experiments_table->[_collect_rows],CmdExperimentsGC->[run->[_raise_error_if_all_disabled]],_collect_names->[_filter_names,_update_names]] | Display a table of all the non - negative negative values in a single row of the table Displays a table with the availables metrics. | @pmrowla, Excuse me $ names = {"a":{"b":1, "c":2}} $ [name for path in names for name in names[path] for path in names] > ['b', 'c'] $ names = {"a":{"b":1, "c":2}, "d":{}} $ [name for path in names for name in names[path] for path in names] > ['b', 'b', 'c', 'c'] $ names = {'a': {'b': 1, 'c': 2}, 'd': {'b': 2}} $ [name for path in names for name in names[path] for path in names] > ['b', 'b', 'c', 'c', 'b', 'b'] A bug or feature? I do not quite understand this line. Is it for counting the keys in the nested dict. |
@@ -23,7 +23,6 @@ public final class KsqlConstants {
public static final String CONFLUENT_AUTHOR = "Confluent";
public static final String KSQL_INTERNAL_TOPIC_PREFIX = "_confluent-ksql-";
- public static final String CONFLUENT_INTERNAL_TOPIC_PREFIX = "__confluent";
public static final String STREAMS_CHANGELOG_TOPIC_SUFFIX = "-changelog";
public static final String STREAMS_REPARTITION_TOPIC_SUFFIX = "-repartition";
| [No CFG could be retrieved] | Creates a new instance of the ksql. constants. Missing AVRO schema name. | Consider moving `KSQL_INTERNAL_TOPIC_PREFIX` into the new `ReservedInternalTopics` class too. Just expose it as a constant or static method. |
@@ -94,6 +94,16 @@ func parseUrl(url string) (uv urlVars, err error) {
if len(parts) > 3 {
uv.path = strings.Join(parts[3:], "/")
}
+
+ // TODO for now, we pull namespace from query parameter, but according to spec, it must go in resource path in future PR
+ // if a namespace if specified, it's always used.
+ // for list/watch operations, a namespace is not required if omitted.
+ // for all other operations, if namespace is omitted, we will default to default namespace.
+ namespace := req.URL.Query().Get("namespace")
+ if len(namespace) > 0 {
+ ctx = kubeapi.WithNamespace(ctx, namespace)
+ }
+
return
}
| [ServeHTTP->[Error,CreateBuild,Extract,GetBuildConfig,NewContext],Join,Error,Errorf,Split,Trim] | parseUrl parses the url and returns a list of unique id for the build. | So this is copy/paste from upstream, seems like we want to put this in an upstream utils function that we can reuse. |
@@ -114,4 +114,17 @@ public class ProxyServiceFactoryBean extends ReflectionServiceFactoryBean
}
+ public void setSoapVersion(String soapVersion)
+ {
+ for (AbstractServiceConfiguration serviceConfiguration : getServiceConfigurations())
+ {
+ if (serviceConfiguration instanceof ProxyServiceConfiguration)
+ {
+ ProxyServiceConfiguration proxyServiceConfiguration = ((ProxyServiceConfiguration) serviceConfiguration);
+ proxyServiceConfiguration.setSoapVersion(soapVersion);
+ return;
+ }
+ }
+ }
+
}
| [ProxyServiceFactoryBean->[initializeWSDLOperations->[initializeWSDLOperations]]] | Initialize the WSDL - based bindings for the provider. | add log saying that soapVersion could not be determined and what are you defaulting to. Is there no other way to determine that other than this instanceof??? Sounds like the kind of things to be merged on closed eyes |
@@ -1123,7 +1123,7 @@ public class RulesAttachment extends AbstractPlayerRulesAttachment {
}
}
if (count == 0) {
- return count == found;
+ return 0 == found;
}
if (getCountEach()) {
m_eachMultiple = found;
| [RulesAttachment->[isSatisfied->[getAlliedOwnershipTerritories,getDirectPresenceTerritories,getAtWarPlayers,getEnemyPresenceTerritories,getDirectExclusionTerritories,getEnemySurfaceExclusionTerritories,getAtWarCount,get,getDirectOwnershipTerritories,getAlliedExclusionTerritories,getEnemyExclusionTerritories,getAlliedPresenceTerritories],checkUnitPresence->[getUnitPresence],checkRelationships->[getRelationship],checkUnitExclusions->[getUnitPresence],get->[get]]] | Check if the player is at war. | The argument and the 0 should be switched |
@@ -35,7 +35,6 @@ function writeFiles() {
this.copy('gitignore', '.gitignore');
this.copy('gitattributes', '.gitattributes');
this.copy('editorconfig', '.editorconfig');
- this.template('_travis.yml', '.travis.yml', this, {});
this.template('_Jenkinsfile', 'Jenkinsfile', this, {});
},
| [No CFG could be retrieved] | Create a list of files that can be written to the server. Missing package - level variables. | Shouldn't _Jenkinsfile be removed from main generator and from here then also? If yes then also remove from expected test files. |
@@ -1361,9 +1361,7 @@ zvol_open(struct block_device *bdev, fmode_t flag)
int error = 0;
boolean_t drop_suspend = B_TRUE;
- ASSERT(!MUTEX_HELD(&zvol_state_lock));
-
- mutex_enter(&zvol_state_lock);
+ rw_enter(&zvol_state_lock, RW_READER);
/*
* Obtain a copy of private_data under the zvol_state_lock to make
* sure that either the result of zvol free code path setting
| [No CFG could be retrieved] | Creates a new object of type zvol. zvol_open - open of zvol. | Can we assert that zvol_state_lock is not held here ? |
@@ -359,6 +359,9 @@ class ExplicitStrategy():
def AddVariables(self):
pass
+ def AddAdditionalHistoricalVariables(self,list_additional_variables):
+ pass
+
def BeforeInitialize(self):
self.CreateCPlusPlusStrategy()
self.RebuildListOfDiscontinuumSphericParticles()
| [ExplicitStrategy->[GetTranslationalScheme->[GetTranslationalSchemeInstance,TranslationalIntegrationSchemeTranslator],CreateCPlusPlusStrategy->[SetVariablesAndOptions],GetRotationalScheme->[GetRotationalSchemeInstance,RotationalIntegrationSchemeTranslator],RootByBisection->[coeff_of_rest_diff],SetVariablesAndOptions->[SetOneOrZeroInProcessInfoAccordingToBoolValue],ModifySubProperties->[SinAlphaConicalDamage,GammaForHertzThornton,RootByBisection],Initialize->[CheckMomentumConservation,Initialize],ModifyProperties->[GetTranslationalScheme,GetRotationalScheme],RotationalIntegrationSchemeTranslator->[TranslationalIntegrationSchemeTranslator],BeforeInitialize->[CreateCPlusPlusStrategy]]] | Add variables to the model. | I would only add it if it is actually used |
@@ -259,7 +259,7 @@ class Nit(object):
def __init__(self, code, severity, filename, message, line_range=None, lines=None):
if severity not in self.SEVERITY:
- raise ValueError('Severity should be one of {}'.format(' '.join(self.SEVERITY.values())))
+ raise ValueError('Severity should be one of {}'.format(' '.join(list(self.SEVERITY.values()))))
if not re.match(r'[A-Z]\d{3}', code):
raise ValueError('Code must contain a prefix letter followed by a 3 digit number')
self.filename = filename
| [Nit->[__str__->[flatten_lines]],PythonFile->[enumerate->[enumerate],__init__->[OffByOneList,_remove_coding_header],parse->[_parse],from_statement->[_parse],iter_logical_lines->[iter_tokens,translate_logical_line],tokens->[iter_tokens]],CheckstylePlugin->[__iter__->[nits],warning->[nit],comment->[nit],error->[nit],nit->[line_range,Nit]],OffByOneList->[index->[index]],CheckSyntaxError->[as_nit->[Nit,OffByOneList]]] | Initialize a new Error object. | Likewise I think you don't need list(). The str.format() will call the values when it needs them, avoiding the intermediate call to list(). |
@@ -18,6 +18,7 @@ const (
SubnetworkLinkRegex = "projects/(" + ProjectRegex + ")/regions/(" + RegionRegex + ")/subnetworks/(" + SubnetworkRegex + ")$"
RFC1035NameTemplate = "[a-z](?:[-a-z0-9]{%d,%d}[a-z0-9])"
+ CloudIoTIdRegex = "[a-zA-Z][-a-zA-Z0-9._+~%]{2,254}"
)
var (
| [MatchString,Sprintf,CIDRNetwork,Contains,Errorf,ParseCIDR,MustCompile,ParseUint] | Imports a single object. validateRFC1918Network returns a schema. SchemaValidateFunc that checks if a given value. | Is this well-tested? The formatter is trying to complain about the % as if it's a format-string specifier and I'm pretty sure it's a formatter bug and not a bug in your code. :) |
@@ -239,11 +239,11 @@ public:
pInstance->SetData(TYPE_HODIR, NOT_STARTED);
}
- if (GameObject* go = me->FindNearestGameObject(GO_HODIR_FRONTDOOR, 300.0f))
+ if (GameObject* go = me->FindNearestGameObject(GO_HODIR_FRONTDOOR, 900.0f))
{
go->SetGoState(GO_STATE_ACTIVE);
}
-
+
if (pInstance && pInstance->GetData(TYPE_HODIR) != DONE)
{
pInstance->SetData(TYPE_SPAWN_HODIR_CACHE, 0);
| [No CFG could be retrieved] | The boss_hodir_AI class. region GameObject HODIR Functions. | revert this please |
@@ -94,6 +94,7 @@ if __name__ == '__main__':
PARSER.add_argument('--pai_hdfs_host', type=str, help='the host of hdfs')
PARSER.add_argument('--pai_user_name', type=str, help='the username of hdfs')
PARSER.add_argument('--nni_hdfs_exp_dir', type=str, help='nni experiment directory in hdfs')
+ PARSER.add_argument('--disable_log', type=bool, default=False, help='disable trial log')
args, unknown = PARSER.parse_known_args()
if args.trial_command is None:
exit(1)
| [trial_keeper_help_info->[print],main_loop->[,split,HdfsClient,poll,copyDirectoryToHdfs,sleep,copyHdfsDirectoryToLocal,str,nni_log,format,Popen,exists,open,makedirs,exit,getcwd,read_experiment_metrics],add_argument,main_loop,ArgumentParser,getLogger,parse_known_args,str,set_defaults,nni_log,format,exit] | Main function of the trial keeper. | This causes confusion, --disable_log False or --disable 'False' are evaluated as True. Typical solution for bool argument are : PARSER.add_argument('--disable_log', action='store_true'), so that you do not need to pass 'True' or 'False' |
@@ -230,8 +230,10 @@ export class <%= entityReactName %> extends React.Component<I<%= entityReactName
</Button>
</td>
<%_ for (idx in fields) {
+ console.log(fields[idx]);
const fieldType = fields[idx].fieldType;
const fieldName = fields[idx].fieldName;
+ const fieldIsEnum = fields[idx].fieldIsEnum;
_%>
<td>
<%_ if (fieldType === 'Boolean') { _%>
| [No CFG could be retrieved] | Renders a list of entities that can be queried by the user. XML - based field that can be used to generate a unique identifier. | please remove this |
@@ -15,7 +15,6 @@ from parlai.core.worlds import create_task
def main(opt):
# Check options
opt['datatype'] = 'valid'
- opt['batchsize'] = 1
assert('pretrained_model' in opt)
# Load document reader
| [main->[time,Timer,format,create_task,items,join,info,parley,report,DocReaderAgent],addHandler,ParlaiParser,setLevel,StreamHandler,parse_args,getLogger,Formatter,is_available,set_device,main,info,add_cmdline_args,setFormatter] | Main entry point for the Task. | ooh for batch testing? weeeeee~~ |
@@ -1669,6 +1669,14 @@ func (a *AgentPoolProfile) GetKubernetesLabels(rg string, deprecated bool) strin
return buf.String()
}
+// IsCSIProxyEnabled returns true if csi proxy service should be enable for Windows nodes
+func (w *WindowsProfile) IsCSIProxyEnabled() bool {
+ if w.EnableCSIProxy != nil {
+ return *w.EnableCSIProxy
+ }
+ return DefaultEnableCSIProxyWindows
+}
+
// HasSecrets returns true if the customer specified secrets to install
func (w *WindowsProfile) HasSecrets() bool {
return len(w.Secrets) > 0
| [GetCustomEnvironmentJSON->[IsAzureStackCloud],HasAvailabilityZones->[HasAvailabilityZones],GetUserAssignedID->[UserAssignedIDEnabled],SetCloudProviderBackoffDefaults->[IsAzureStackCloud],GetVMType->[HasVMSSAgentPool],HasImageGallery->[IsGalleryImage],GetSecondaryNonMasqueradeCIDR->[IsHostedMasterProfile],GetFirstConsecutiveStaticIPAddress->[IsVirtualMachineScaleSets],IsUbuntu->[IsUbuntu1804,IsUbuntu1604],IsPrivateCluster->[IsKubernetes],GetUserAssignedClientID->[UserAssignedClientIDEnabled],GetCustomCloudAuthenticationMethod->[IsAzureStackCloud],GetAzureProdFQDN->[GetCustomCloudName],GetVirtualNetworkName->[K8sOrchestratorName,IsHostedMasterProfile],GetRouteTableName->[GetResourcePrefix],GetClusterMetadata->[GetVirtualNetworkName,GetRouteTableName,GetPrimaryScaleSetName,GetResourcePrefix,GetPrimaryAvailabilitySetName,GetSubnetName,GetVNetResourceGroupName,GetNSGName],HasImageRef->[IsValid],GetAddonScript->[GetAddonByName],IsAzurePolicyEnabled->[IsAddonEnabled],GetCustomCloudName->[IsAzureStackCloud],GetNonMasqueradeCIDR->[IsHostedMasterProfile],GetProvisionScriptParametersCommon->[GetVirtualNetworkName,GetCloudSpecConfig,GetRouteTableName,GetPrimaryScaleSetName,GetKubernetesVersion,GetAzureCNIURLLinux,IsHostedMasterProfile,GetPrimaryAvailabilitySetName,GetKubernetesHyperkubeSpec,GetSubnetName,GetVMType,IsFeatureEnabled,GetCustomCloudIdentitySystem,GetVNetResourceGroupName,GetCustomCloudAuthenticationMethod,GetNSGName],RequireRouteTable->[IsAzureCNI],IsNVIDIADevicePluginEnabled->[IsAddonEnabled],IsIPMasqAgentEnabled->[IsAddonEnabled,IsIPMasqAgentEnabled],GetCustomCloudIdentitySystem->[IsAzureStackCloud],IsAKSBillingEnabled->[GetCloudSpecConfig],GetVNetResourceGroupName->[IsHostedMasterProfile],GetCosmosEndPointURI->[HasCosmosEtcd],SetCloudProviderRateLimitDefaults->[IsAzureStackCloud],GetPrimaryScaleSetName->[GetAgentVMPrefix],GetMasterFQDN->[IsHostedMasterProfile],GetCloudSpecConfig->[GetCustomCloudName],IsClusterAutoscalerEnabled->[IsAddonEnabled],GetResourcePrefix->[K8sOrchestratorName],GetSubnetName->[K8sOrchestratorName,IsHostedMasterProfile],GetLocations->[IsAzureStackCloud],GetAgentVMPrefix->[K8sOrchestratorName],IsContainerMonitoringAddonEnabled->[IsAddonEnabled],IsComponentEnabled->[GetComponentByName,IsEnabled],IsReschedulerEnabled->[IsAddonEnabled],IsIPMasqAgentDisabled->[IsAddonDisabled,IsIPMasqAgentDisabled],IsNvidiaDevicePluginCapable->[HasNSeriesSKU],IsUbuntuNonVHD->[IsUbuntu,IsVHDDistro],GetNSGName->[GetResourcePrefix],GetMasterVMPrefix->[K8sOrchestratorName],IsAddonDisabled->[IsDisabled,GetAddonByName],IsAADPodIdentityEnabled->[IsAddonEnabled],GetComponentData->[GetComponentByName],IsAppGWIngressEnabled->[IsAddonEnabled],IsAddonEnabled->[IsEnabled,GetAddonByName]] | GetKubernetesLabels returns the labels for the agentpool. | this needs to be `common.DefaultEnableCSIProxyWindows` (assuming `common` is a reference to the api/common package) |
@@ -1588,7 +1588,8 @@ public class SearchManager {
Classifier classifier = dimension.getClassifier();
for (CategoryPath categoryPath: classifier.classify(value)) {
- if (!dimension.isLocalized() || dimension.getLocales().contains(locale)) {
+ result.add(new FacetField(dimension.getName(), categoryPath.components));
+ if (dimension.isLocalized() && dimension.getLocales().contains(locale)) {
result.add(new FacetField(dimension.getName(locale), categoryPath.components));
}
}
| [SearchManager->[newDocument->[toString,equals],getFacetFieldsFor->[getFacetFieldsFor],initHardCodedAnalyzers->[createHardCodedPerFieldAnalyzerWrapper],reloadLuceneConfiguration->[createDocumentBoost,createAnalyzer],indexGeometry->[index],transform->[transform],getTermsFequency->[getAnalyzer],getNewIndexReader->[getIndexReader],Spatial->[filter->[toString],getFilterParser->[equals],setupIndex],getIndexFields->[toString,allText,addField],allText->[allText],createAnalyzer->[initHardCodedAnalyzers],TermFrequency->[hashCode->[hashCode],compareTo->[compareTo],equals->[compareTo]],rebuildIndex->[init,setupIndex],init->[createAnalyzer],addNumericField->[equals],delete->[delete],mergeDefaultLang->[compare->[compare,compareTo],equals],createHardCodedPerFieldAnalyzerWrapper->[createGeoNetworkAnalyzer],configurePerFieldAnalyzerWrapper->[createAnalyzerFromLuceneConfig],end->[end],index->[forceIndexChanges],createAnalyzerFromLuceneConfig->[createGeoNetworkAnalyzer]]] | This method returns a list of FacetFields for the given locale and dimension. | In fact it breaks indexing because a field and a facet have the same name. I will fix that. |
@@ -176,6 +176,7 @@ func createPodOptions(p *specgen.PodSpecGenerator, rt *libpod.Runtime, infraSpec
options = append(options, libpod.WithPodUser())
}
}
+ options = append(options, libpod.WithPodCgroups())
if len(p.CgroupParent) > 0 {
options = append(options, libpod.WithPodCgroupParent(p.CgroupParent))
}
| [Pull,LibimageRuntime,TempFile,NewPod,Close,AddInfra,WithPodUser,IsDefault,IsHost,Validate,WithPodLabels,NamespaceMode,CgroupParent,Errorf,FindHelperBinary,WithPodCreateCommand,Debugf,ID,SavePod,Wrapf,WithPod,WithInfraContainer,Name,Remove,GetVersion,GetConfigNoCopy,Build,Sprintf,Background,WithPodHostname,WithPodCgroupParent,WriteString,LookupImage,WithPodName] | createPodOptions creates a new pod object with the given options. MapSpec creates a specgenerator that maps the given pod spec to the infra container spec. | `WithPodCgroup` or `Cgroups`? |
@@ -90,6 +90,8 @@ s.serve_forever()" """
def add_parser_arguments(cls, add):
add("test-mode", action="store_true",
help="Test mode. Executes the manual command in subprocess.")
+ add("public-ip-logging-ok", action="store_true",
+ help="Automatically allows public IP logging.")
def prepare(self): # pylint: disable=missing-docstring,no-self-use
pass # pragma: no cover
| [Authenticator->[_perform_single->[_test_mode_busy_wait]]] | Adds parser arguments for . | Small suggestion: This becomes the flag `--manual-public-ip-logging-ok`. This is very verbose. What about `--manual-ip-logging-ok` or `--manual-public-log-ok`? |
@@ -95,6 +95,8 @@ public class CoordinatorDynamicConfig
this.killDataSourceWhitelist = parseJsonStringOrArray(killDataSourceWhitelist);
this.killPendingSegmentsSkipList = parseJsonStringOrArray(killPendingSegmentsSkipList);
this.maxSegmentsInNodeLoadingQueue = maxSegmentsInNodeLoadingQueue;
+ this.maintenanceList = parseJsonStringOrArray(maintenanceList);
+ this.maintenanceModeSegmentsPriority = Math.min(10, Math.max(0, maintenanceModeSegmentsPriority));
if (this.killAllDataSources && !this.killDataSourceWhitelist.isEmpty()) {
throw new IAE("can't have killAllDataSources and non-empty killDataSourceWhitelist");
| [CoordinatorDynamicConfig->[Builder->[build->[getMaxSegmentsToMove,isKillAllDataSources,getMaxSegmentsInNodeLoadingQueue,getKillDataSourceWhitelist,getMergeSegmentsLimit,getBalancerComputeThreads,getReplicantLifetime,emitBalancingStats,getMergeBytesLimit,getMillisToWaitBeforeDeleting,getReplicationThrottleLimit,getKillPendingSegmentsSkipList,CoordinatorDynamicConfig]],equals->[equals]]] | This method is used to parse the JSON string of a . Json serializable Set of . | I think it should rather check bounds than saturating, to avoid silent consumption of configuration errors |
@@ -43,11 +43,8 @@ func NewTaskQueue(ctx context.Context, etcdClient *etcd.Client, etcdPrefix strin
taskQueue: newTaskQueue(ctx),
}
// Clear etcd key space.
- // TODO: Multiple storage task queues are setup, so deleting the existing tasks is problematic.
- if taskNamespace != "storage" {
- if err := tq.deleteAllTasks(); err != nil {
- return nil, err
- }
+ if err := tq.deleteAllTasks(); err != nil {
+ return nil, err
}
return tq, nil
}
| [taskFunc->[subtaskFunc,Unmarshal,ReadOnly,Close,Watch,WatchOne,runSubtask,Err,Done],TaskCount->[ReadOnly,CountRev],RunSubtasksChan->[AddInt64,Printf,createSubtask,LoadInt64,deleteSubtasks,New,WithCancel,Go,ReadOnly,Unmarshal,WatchOneF,Err,runSubtaskBlock,Is,Wait],subtaskFunc->[Printf,ReadWrite,Error,NewSTM,Claim,IsErrNotFound,Get,Is,Err,Update],deleteSubtasks->[ReadWrite,Background,NewSTM,DeleteAllPrefix],RunSubtasks->[RunSubtasksChan,Go,Err,Wait,Done],Run->[WatchF,Printf,runTask,Unmarshal,ReadOnly,deleteTask,taskFunc,Is,Err],deleteTask->[ReadWrite,NewSTM,Background,Delete,DeleteAllPrefix],RunTask->[Printf,ReadWrite,runTask,NewSTM,deleteTask,NewWithoutDashes,Put],createSubtask->[ReadWrite,Join,NewSTM,NewWithoutDashes,Put],RunTaskBlock->[RunTask],deleteAllTasks->[ReadWrite,Background,NewSTM,DeleteAll],NewEtcdCollection,Join,deleteAllTasks] | New const creates a new taskQueue based on a set of parallel tasks and provides RunTask runs a task in the task queue. | What else constructs a NewTaskQueue? Will there be a race between different components. e.g. Component A creates a task queue and starts queuing items. Component B creates a task queue and wipes out component A's tasks. |
@@ -321,8 +321,8 @@ public class ConversionServiceImpl extends DefaultComponent implements Conversio
if (config.isCacheEnabled()) {
ConversionCacheHolder.addToCache(cacheKey, result);
}
- } else {
- // we need to reset the filename if result came from cache because it's just a hash
+ } else if (result.getBlobs() != null && result.getBlobs().size() == 1) {
+ // we need to reset the filename if result is a single file from the cache because the name is just a hash
result.getBlob().setFilename(null);
}
| [ConversionServiceImpl->[getConverterNames->[getConverterNames],isCacheEnabled->[isCacheEnabled],getAdapter->[getAdapter],isConverterAvailable->[isConverterAvailable],startGC->[start],getConverterName->[getConverterNames],convertToMimeType->[convertThroughHTML,convert],convert->[isCacheEnabled,convert],adjustBlobName->[adjustBlobName]]] | This method is called by the converter to convert the given blobHolder with the given parameters. | Space between `)` and `{` |
@@ -74,7 +74,16 @@ func (u *UserSessionStore) UserSession(id string) *UserSession {
// Get logs into vSphere and returns a vSphere session object. Caller responsible for error handling/logout
func (u *UserSessionStore) VSphere(id string) (vSphereSession *session.Session, err error) {
- return vSphereSessionGet(u.UserSession(id).config)
+ us := u.UserSession(id)
+ if us.vsphere == nil {
+ log.Infof("Creating session for %s", id)
+ s, err := vSphereSessionGet(us.config)
+ if err != nil {
+ return nil, err
+ }
+ us.vsphere = s
+ }
+ return us.vsphere, nil
}
// reaper takes abandoned sessions to a farm upstate so they don't build up forever
| [reaper->[Since,Delete,Infof],VSphere->[UserSession],Delete->[Lock,Unlock],UserSession->[RUnlock,RLock],Add->[Lock,Unlock,Now],GenerateRandomKey,NewTicker,Seconds,reaper,NewCookieStore] | VSphere returns a vSphere session. | This should also check to make sure the `UserSession` is not nil itself. |
@@ -446,12 +446,8 @@ define([
this._primitives = primitives;
this._geometryUpdater = geometryUpdater;
this._positions = [];
- };
- var generateCartesianArcOptions = {
- positions : undefined,
- granularity : undefined,
- height : undefined
+ generateCartesianArcOptions.ellipsoid = geometryUpdater._scene.globe.ellipsoid;
};
DynamicGeometryUpdater.prototype.update = function(time) {
| [No CFG could be retrieved] | Creates a new DynamicGeometryUpdater instance. Extract heights from the positions and create a 2D polygon. | `scene.globe` is not guaranteed to exist, generally we prefer `scene.mapProjection.ellipsoid` for this kind of thing. |
@@ -257,7 +257,7 @@ public class ApkBuilder
string? buildTools = Directory.GetDirectories(Path.Combine(androidSdkDir, "build-tools"))
.Select(Path.GetFileName)
.Where(file => !file!.Contains("-"))
- .Select(file => Version.TryParse(Path.GetFileName(file), out Version? version) ? version : default)
+ .Select(file => Version.TryParse(Path.GetFileName(file), out Version? version) ? (version ?? default) : default)
.OrderByDescending(v => v)
.FirstOrDefault()?.ToString();
| [ApkBuilder->[BuildApk->[GetLatestBuildTools,Exists,Delete,StartsWith,Add,Copy,GetRid,Combine,GetFiles,DirectoryCopy,TryParse,Contains,IsNullOrEmpty,GetFileNameWithoutExtension,CreateDirectory,GetFileName,RunProcess,GetEnvironmentVariable,AddRange,LogInfo,WriteAllText,GetLatestApiLevel,Replace,GetExtension,GetEmbeddedResource,Length],GetLatestBuildTools->[FirstOrDefault,IsNullOrEmpty],GetLatestApiLevel->[ToString]]] | Get latest build - tools in Android SDK. | The original code looks correct. Is this working around a Roslyn bug? |
@@ -34,12 +34,12 @@ class TaskNotFound(MissingResource):
return _('Task Not Found: %(id)s') % {'id': self.args[0]}
-class TaskCancelNotImplemented(PulpExecutionException):
+class TaskCompleteException(PulpExecutionException):
- http_status_code = httplib.NOT_IMPLEMENTED
+ http_status_code = httplib.CONFLICT
def __str__(self):
- return _('Cancel Not Implemented for Task: %(id)s') % {'id': self.args[0]}
+ return _('Task is already in a complete state: %(id)s') % {'id': self.args[0]}
# task controllers -------------------------------------------------------------
| [TaskResource->[GET->[TaskNotFound],DELETE->[TaskCancelNotImplemented]]] | Return a string representation of the task. | This should be updated as a PulpCoded Exception with an appropriate error code & message |
@@ -888,14 +888,8 @@ func checkIfRoleIntersectsProjectsFilter(ctx context.Context, q Querier,
// Return true or false if there is intersection between iam_role_projects and projectsFilter,
// assuming '{(unassigned)}' in the case that iam_role_projects is empty. If a role of id
- // doesn't exist, this will return 0 rows which will bubble up to NotFoundErr when passed to processError.
- row := q.QueryRowContext(ctx,
- `SELECT COALESCE(array_agg(rp.project_id)
- FILTER (WHERE rp.project_id IS NOT NULL), '{(unassigned)}') && $2 AS intersection
- FROM iam_roles AS r
- LEFT OUTER JOIN iam_role_projects AS rp ON rp.role_id=r.db_id
- WHERE r.id = $1;`,
- id, pq.Array(projectsFilter))
+ // doesn't exist, this will return a proper SQL "no rows" error when passed to processError.
+ row := q.QueryRowContext(ctx, "SELECT projects_match(role_projects($1), $2)", id, pq.Array(projectsFilter))
var result bool
err := row.Scan(&result)
| [AddPolicyMembers->[notifyPolicyChange,queryPolicy],Reset->[Reset],CreateRole->[notifyPolicyChange],RemovePolicyMembers->[notifyPolicyChange,queryPolicy],recordMigrationStatusAndNotifyPG->[notifyPolicyChange],Success->[recordMigrationStatusAndNotifyPG],UpdateRole->[notifyPolicyChange],ApplyStagedRules->[notifyPolicyChange],ListPolicyMembers->[queryPolicy],Close->[Close],SuccessBeta1->[recordMigrationStatusAndNotifyPG],DeleteRole->[notifyPolicyChange],ReplacePolicyMembers->[notifyPolicyChange,queryPolicy]] | checkIfRoleIntersectsProjectsFilter checks if a role in the specified projectsFilter is in finds the next unique id for the given role. | would make sense to coalesce this into `{(unassigned)}`, but the tests pass regardless... |
@@ -68,6 +68,11 @@ public class HeadlessUiContext extends AbstractUiContext {
return null;
}
+ @Override
+ public TerritoryEffectImageFactory getTerritoryEffectImageFactory() {
+ return null;
+ }
+
@Override
public MapImage getMapImage() {
return null;
| [HeadlessUiContext->[internalSetMapDir->[Stopwatch,done,getMapResourceLoader,getDouble]]] | Gets the resource image factory. | @Override returning null usaully indicates an abstraction problem. It's not much of a surprise given headless vs headed was done in a massive hurry to get the bots up and running. In this case it seems like `getTerritoryEffectImageFactory` should only be a method on the `HeadedContext`, not the headless. If a caller is not aware of the difference, they can get a NPE. Seems like a caller of this code would need to know which instance type, so moving the method to the right instance type I suspect is appropriate and would avoid the error and abstraction problem. |
@@ -107,6 +107,7 @@ public class CardTemplateEditor extends AnkiActivity implements DeckSelectionDia
// the current editor view among front/style/back
private HashMap<Integer, Integer> mEditorViewId;
private int mStartingOrdId;
+ private static SelectableDeck selectedDeck;
private static final String EDITOR_POSITION_KEY = "editorPosition";
private static final String EDITOR_VIEW_ID_KEY = "editorViewId";
| [CardTemplateEditor->[onCollectionLoaded->[onCollectionLoaded],onKeyUp->[onKeyUp],onBackPressed->[onBackPressed],onSaveInstanceState->[onSaveInstanceState],onOptionsItemSelected->[onOptionsItemSelected,onBackPressed],onCreate->[onCreate],CardTemplateFragment->[newInstance->[CardTemplateFragment],onResume->[onResume],ActionModeCallback->[onActionItemClicked->[getItemId]],setCurrentEditorView->[setCurrentEditorViewId],displayDeckOverrideDialog->[newInstance],performPreview->[getTempModel],addNewTemplate->[addNewTemplate],onOptionsItemSelected->[getTempModel,getItemId,onOptionsItemSelected,showInsertFieldDialog,modelHasChanged],deleteTemplate->[ordinalShift],onCreateView->[getTempModel,getItemId],onCardBrowserAppearanceResult->[getCurrentTemplate],onCreateOptionsMenu->[onCreateOptionsMenu],modelHasChanged->[modelHasChanged],ordinalShift],TemplatePagerAdapter->[containsItem->[getItemCount],ordinalShift->[getItemCount],getItemCount->[getTempModel]]]] | The CardTemplateEditor class This is called when the bundle is being edited. | I had to create this as static as the method from which we are calling the DeckSelectionDialog is also static |
@@ -1168,7 +1168,9 @@ def plot_evoked_topomap(evoked, times="auto", ch_type=None, layout=None,
fig = axes[0].get_figure()
fig.subplots_adjust(left=w_frame, right=1 - w_frame, bottom=0,
top=1 - top_frame)
- time_idx = [np.where(evoked.times >= t)[0][0] for t in times]
+ # find first index that's >= (to rounding error) to each time point
+ time_idx = [np.where(_time_mask(evoked.times, t, None))[0][0]
+ for t in times]
if proj is True and evoked.proj is not True:
data = evoked.copy().apply_proj().data
| [_init_anim->[_griddata,_make_image_mask,_check_outlines,_hide_frame],plot_ica_components->[_make_image_mask,plot_ica_components,_check_outlines,_prepare_topo_plot,plot_topomap],_plot_topomap_multi_cbar->[plot_topomap],plot_evoked_topomap->[_make_image_mask,_prepare_topo_plot,plot_topomap,_check_outlines],plot_tfr_topomap->[_prepare_topo_plot,plot_topomap],plot_topomap->[_griddata,_show_names,_check_outlines,_plot_sensors],plot_psds_topomap->[_plot_topomap_multi_cbar],_onselect->[_check_outlines],plot_epochs_psd_topomap->[_prepare_topo_plot],_animate->[_hide_frame],_topomap_animation->[_prepare_topo_plot]] | Plot a topographic map of specific time points of evoked data. Plots a single topomap. Plots a single . Plots the n - th pick and n - th merge grads. Plots the non - zero nanoseconds in the data. Plots a 2D image of the missing missing values. | you could actually pass sfreq too |
@@ -2719,6 +2719,18 @@ public class LibvirtComputingResource extends ServerResourceBase implements Serv
return storagePool.getPhysicalDisk(data.getPath());
}
+ /**
+ * Set Disk IO Driver, if supported by the Libvirt/Qemu version.
+ * IO Driver works for:
+ * (i) Qemu >= 5.0;
+ * (ii) Libvirt >= 6.3.0
+ */
+ protected void setDiskIoDriver(DiskDef disk) {
+ if (getHypervisorLibvirtVersion() >= HYPERVISOR_LIBVIRT_VERSION_SUPPORTS_IO_URING && getHypervisorQemuVersion() >= HYPERVISOR_QEMU_VERSION_SUPPORTS_IO_URING) {
+ disk.setIoDriver(DiskDef.IoDriver.IOURING);
+ }
+ }
+
private KVMPhysicalDisk getPhysicalDiskFromNfsStore(String dataStoreUrl, DataTO data) {
final String volPath = dataStoreUrl + File.separator + data.getPath();
final int index = volPath.lastIndexOf("/");
| [LibvirtComputingResource->[getInterfaces->[getInterfaces],isSnapshotSupported->[executeBashScript],getNetworkStats->[networkUsage],getVifDriver->[getVifDriver],cleanupVMNetworks->[getAllVifDrivers],checkNetwork->[getVifDriver],cleanupNetworkElementCommand->[vifHotUnPlug,VifHotPlug],getVPCNetworkStats->[configureVPCNetworkUsage],getVmNetworkStat->[getInterfaces,getDomain],configure->[getDefaultDirectDownloadTemporaryPath,getDefaultKvmScriptsDir,getDefaultNetworkScriptsDir,configure,getDeveloperProperties,getDefaultCachePath,getDefaultStorageScriptsDir,getDefaultHypervisorScriptsDir,getDefaultDomrScriptsDir],getVifDriverClass->[configure],destroyNetworkRulesForVM->[getInterfaces],applyDefaultNetworkRules->[getType],createVMFromSpec->[enlightenWindowsVm,setQuotaAndPeriod,getUuid],vifHotUnPlug->[getAllVifDrivers],post_default_network_rules->[getInterfaces],getPhysicalDiskPrimaryStore->[getUuid],getOvsPifs->[isGuestBridge],checkBridgeNetwork->[matchPifFileInDirectory],configureTunnelNetwork->[findOrCreateTunnelNetwork],getDisks->[getDisks],defaultNetworkRules->[getInterfaces],configureHostParams->[configure],destroyNetworkRulesForNic->[getInterfaces],attachOrDetachISO->[cleanupDisk],prepareNetworkElementCommand->[VifHotPlug,getBroadcastUriFromBridge],findOrCreateTunnelNetwork->[checkNetwork],getVmState->[convertToPowerState],getVersionStrings->[KeyValueInterpreter,getKeyValues],attachOrDetachDisk->[getUuid],initialize->[getVersionStrings,getUuid],executeInVR->[executeInVR],getHostVmStateReport->[convertToPowerState,getHostVmStateReport],getVmDiskStat->[getDomain,getDisks],enlightenWindowsVm->[getUuid],getDeveloperProperties->[getEndIpFromStartIp],setQuotaAndPeriod->[getUuid],getVncPort->[getVncPort],getVmStat->[getInterfaces,getDomain,VmStats,getDisks],createVbd->[getUuid,getConfigPath],syncNetworkGroups->[getRuleLogsForVms],getVolumePath->[getVolumePath,getConfigPath],getBroadcastUriFromBridge->[matchPifFileInDirectory]]] | getPhysicalDiskPrimaryStore - Get the physical disk from the primary store. | LGTM and tests have passed, but this enables io_uring as default for the matching version of qemu and libvirt, could that cause an issue @GabrielBrascher for different type of guest OS? |
@@ -40,12 +40,12 @@ class MechanicalSolver(object):
def __init__(self, main_model_part, custom_settings):
default_settings = KratosMultiphysics.Parameters("""
{
- "solver_type": "Static",
+ "solver_type": "static",
"echo_level": 0,
"buffer_size": 2,
- "analysis_type": "Linear",
- "time_integration_method": "Implicit",
- "scheme_type": "Newmark",
+ "analysis_type": "non_linear",
+ "time_integration_method": "implicit",
+ "scheme_type": "newmark",
"model_import_settings": {
"input_type": "mdpa",
"input_filename": "unknown_name",
| [MechanicalSolver->[validate_and_transfer_matching_settings->[validate_and_transfer_matching_settings],_create_builder_and_solver->[get_linear_solver],_execute_after_reading->[import_constitutive_laws],Initialize->[Initialize],_create_newton_raphson_strategy->[GetComputingModelPart,get_solution_scheme,get_builder_and_solver,get_linear_solver,get_convergence_criterion],_set_and_fill_buffer->[GetMinimumBufferSize],_create_linear_strategy->[GetComputingModelPart,get_builder_and_solver,get_solution_scheme,get_linear_solver],Solve->[Solve]]] | Initialize the KratosMultiphysics object with default parameters. This method is called when the model is implemented by the MechanicalSolver. It is. | I had overseen this, thx! |
@@ -865,14 +865,14 @@ public class HoodieWriteClient<T extends HoodieRecordPayload> extends AbstractHo
"Found commits after time :" + lastCommit + ", please rollback greater commits first");
}
- List<String> inflights =
- inflightCommitTimeline.getInstants().map(HoodieInstant::getTimestamp).collect(Collectors.toList());
+ List<String> inflights = inflightAndRequestedCommitTimeline.getInstants().map(HoodieInstant::getTimestamp)
+ .collect(Collectors.toList());
if ((lastCommit != null) && !inflights.isEmpty() && (inflights.indexOf(lastCommit) != inflights.size() - 1)) {
throw new HoodieRollbackException(
"Found in-flight commits after time :" + lastCommit + ", please rollback greater commits first");
}
- List<HoodieRollbackStat> stats = table.rollback(jsc, commitToRollback, true);
+ List<HoodieRollbackStat> stats = table.rollback(jsc, instantToRollback, true);
LOG.info("Deleted inflight commits " + commitToRollback);
| [HoodieWriteClient->[doRollbackAndGetStats->[rollback],rollbackInflightCompaction->[rollback],startCommitWithTime->[startCommit],runCompaction->[compact,commitCompaction],rollbackInternal->[doRollbackAndGetStats,startContext,startInstant,finishRollback],startCommit->[startCommit],rollbackInflightCommits->[rollback],bulkInsert->[bulkInsert],close->[close],savepoint->[savepoint],forceCompact->[compact,scheduleCompaction],scheduleCompactionAtInstant->[scheduleCompaction],compact->[compact],commitCompaction->[commitCompaction],clean->[clean],doCompactionCommit->[finalizeWrite],commit->[commit],upsertRecordsInternal->[saveWorkloadProfileMetadataToInflight],updateIndexAndCommitIfNeeded->[commitOnAutoCommit],finalizeWrite->[finalizeWrite]]] | Rolls back the given commit and returns the stats. This method is called when the index changes have been rolled back. | Does rollback need inflight and requested both ? The restore command probably needs both but rollback should only consider inflight ? |
@@ -4,13 +4,13 @@ from setuptools import find_packages
from setuptools import setup
from setuptools.command.test import test as TestCommand
-version = '1.1.0.dev0'
+version = '1.2.0.dev0'
# Remember to update local-oldest-requirements.txt when changing the minimum
# acme/certbot version.
install_requires = [
'acme>=0.31.0',
- 'certbot>=1.0.0.dev0',
+ 'certbot>=1.0.0',
'dns-lexicon>=2.2.1', # Support for >1 TXT record per name
'mock',
'setuptools',
| [PyTest->[initialize_options->[initialize_options]]] | Creates a test class for the given version of the package. - > Python - > Python - > Python - > Python - > Python - > Python. | Weird that the previous release didn't catch this |
@@ -903,11 +903,15 @@ def enhance(config, plugins):
if not domains:
raise errors.Error("User cancelled the domain selection. No domains "
"defined, exiting.")
+
+ lineage = cert_manager.lineage_for_certname(config, config.certname)
if not config.chain_path:
- lineage = cert_manager.lineage_for_certname(config, config.certname)
config.chain_path = lineage.chain_path
- le_client = _init_le_client(config, authenticator=None, installer=installer)
- le_client.enhance_config(domains, config.chain_path, ask_redirect=False)
+ if oldstyle_enh:
+ le_client = _init_le_client(config, authenticator=None, installer=installer)
+ le_client.enhance_config(domains, config.chain_path, ask_redirect=False)
+ if enhancements.are_requested(config):
+ enhancements.enable(lineage, domains, installer, config)
def rollback(config, plugins):
| [enhance->[_init_le_client],revoke->[revoke,_delete_if_appropriate,_determine_account],renew_cert->[_init_le_client,_get_and_save_cert],_ask_user_to_confirm_new_names->[_format_list,_get_added_removed],install->[_init_le_client,_install_cert,_find_domains_or_certname],unregister->[_determine_account],certificates->[certificates],run->[_suggest_donation_if_appropriate,_find_domains_or_certname,_get_and_save_cert,_report_new_cert,_install_cert,_init_le_client,_find_cert],_report_new_cert->[_report_successful_dry_run],certonly->[_suggest_donation_if_appropriate,_find_domains_or_certname,_get_and_save_cert,_report_new_cert,_csr_get_and_save_cert,_init_le_client,_find_cert],rollback->[rollback],main->[set_displayer,make_or_verify_needed_dirs],_find_lineage_for_domains->[_handle_identical_cert_request,_handle_subset_cert_request],_find_lineage_for_domains_and_certname->[_handle_identical_cert_request,_find_lineage_for_domains],delete->[delete],_init_le_client->[_determine_account],register->[_determine_account],main] | Add security enhancements to existing configuration object. Rollback server configuration changes made during install. | Oh Python. Nice catch moving this outside of the `if not config.chain_path` branch to avoid an ugly crash down below when calling `enhancements.enable`. |
@@ -28,12 +28,8 @@ def send_set_password_email(staff):
@shared_task
def send_promote_customer_to_staff_email(staff):
- site = Site.objects.get_current()
- ctx = {
- 'protocol': 'https' if settings.ENABLE_SSL else 'http',
- 'domain': site.domain,
- 'url': build_absolute_uri(reverse('dashboard:index')),
- 'site_name': site.name}
+ ctx = get_email_base_context()
+ ctx['url'] = build_absolute_uri(reverse('dashboard:index'))
send_templated_mail(
template_name='dashboard/staff/promote_customer',
from_email=settings.DEFAULT_FROM_EMAIL,
| [send_set_password_email->[send_templated_mail,get_current,make_token,urlsafe_base64_encode],send_promote_customer_to_staff_email->[get_current,send_templated_mail,reverse,build_absolute_uri]] | Sends a confirmation email to the user that the user is promotioned to the staff. | I think this URL could use more meaningful name |
@@ -41,7 +41,7 @@ static int mocked_fn_cnt;
bool verbose_unit_test_output = FALSE;
/*
- * Mimick requires some functions setup that are otherwise optional, but doesn't
+ * Mimic requires some functions setup that are otherwise optional, but doesn't
* see to provide an include file with them in so define them here and
* initialize them in main below.
*/
| [main->[cull_tests,putenv,_cmocka_run_group_tests,print_usage,getopt,print_error,generated_cmocka_tests],char->[free,strstr,malloc,snprintf,strncpy],assert_message->[vsnprintf,param_replace,va_start],clear_mocked_functions->[mmk_reset,assert_true_msg],void->[memcpy,strstr,print_message],add_mocked_function->[assert,ARRAY_SIZE,assert_true_msg,mmk_mock_create_internal]] | Creates a mocked version of the that can be used to test the type of Set the given arguments for the given function. | I don't know if this one is correct or not. |
@@ -553,9 +553,12 @@ class Epochs(object):
while not is_good:
if self._current >= len(self.events):
raise StopIteration
- epoch = self._get_epoch_from_disk(self._current)
+ epoch, epoch_raw = self._get_epoch_from_disk(self._current, proj=proj)
self._current += 1
is_good = self._is_good_epoch(epoch)[0]
+ # If in delayed-ssp mode, read 'virgin' data after rejection decision.
+ if self._delayed_ssp:
+ epoch = self._preprocess_epoch(epoch_raw, 'reload')
return epoch
| [read_epochs->[Epochs],combine_event_ids->[copy],equalize_epoch_counts->[drop_epochs,drop_bad_epochs],Epochs->[equalize_event_counts->[drop_epochs,copy,_key_match,drop_bad_epochs],_get_data_from_disk->[_get_epoch_from_disk],to_nitime->[copy,get_data],resample->[resample],get_data->[_get_data_from_disk],next->[_is_good_epoch,_get_epoch_from_disk],__getitem__->[_key_match],save->[get_data],as_data_frame->[copy,get_data]],bootstrap->[copy]] | Get next from the sequence. | that seems weird if you have epoch and epoch_raw above |
@@ -22,6 +22,7 @@ class TestDecomposableAttention(ModelTestCase):
output_dict = self.model.forward(**training_arrays)
assert_almost_equal(numpy.sum(output_dict["label_probs"][0].data.numpy(), -1), 1, decimal=6)
+ @flaky
def test_model_can_train_save_and_load(self):
self.ensure_model_can_train_save_and_load(self.param_file)
| [TestDecomposableAttention->[test_forward_pass_runs_correctly->[forward,output_dict,arrays_to_variables,sum,as_array_dict,assert_almost_equal],test_model_can_train_save_and_load->[ensure_model_can_train_save_and_load],test_mismatched_dimensions_raise_configuration_errors->[raises,from_params,pop,from_file],test_batch_predictions_are_consistent->[ensure_batch_predictions_are_consistent],test_model_load->[isinstance,load,from_file],setUp->[set_up_model,super]]] | Test forward pass runs correctly. | Have you actually seen this one fail? It's much smaller than BiDAF. But I guess it does have weighted sums that use `bmm`... |
@@ -68,6 +68,11 @@ public class FakeBigQueryServices implements BigQueryServices {
return datasetService;
}
+ @Override
+ public DatasetService getDatasetService(BigQueryOptions bqOptions, Histogram histogram) {
+ return datasetService;
+ }
+
@Override
public StorageClient getStorageClient(BigQueryOptions bqOptions) {
return storageClient;
| [FakeBigQueryServices->[FakeBigQueryServerStream->[iterator->[iterator]],encodeQueryResult->[encodeQueryResult]]] | Gets the dataset service. | I don't think we should be passing in the histogram to the factory for building the service. |
@@ -22,7 +22,7 @@ namespace Microsoft.Extensions.Hosting
public const string CreateHostBuilder = nameof(CreateHostBuilder);
// The amount of time we wait for the diagnostic source events to fire
- private static readonly TimeSpan s_defaultWaitTimeout = Debugger.IsAttached ? Timeout.InfiniteTimeSpan : TimeSpan.FromSeconds(5);
+ private static readonly TimeSpan s_defaultWaitTimeout = Debugger.IsAttached ? Timeout.InfiniteTimeSpan : TimeSpan.FromMinutes(1);
public static Func<string[], TWebHost>? ResolveWebHostFactory<TWebHost>(Assembly assembly)
{
| [HostFactoryResolver->[ResolveServiceProviderFactory->[ResolveHostFactory]]] | Resolve web host factory. | Assuming we improve the error message to make it clear that there was a timeout, are we worried that people will just kill the migration before seeing the error if we make this a full minute? |
@@ -18,6 +18,9 @@ import java.net.URL;
public class DefaultSingleLogoutServiceLogoutUrlBuilder implements SingleLogoutServiceLogoutUrlBuilder {
private static final Logger LOGGER = LoggerFactory.getLogger(DefaultSingleLogoutServiceLogoutUrlBuilder.class);
+ @Autowired
+ private UrlValidator urlValidator;
+
@Override
public URL determineLogoutUrl(final RegisteredService registeredService, final WebApplicationService singleLogoutService) {
try {
| [DefaultSingleLogoutServiceLogoutUrlBuilder->[determineLogoutUrl->[IllegalArgumentException,isValid,debug,getLogoutUrl,URL,getOriginalUrl],getLogger]] | Determine logout url. | Please don't autowire. Pass the instance directly when the builder is constructed in its own bean. |
@@ -896,10 +896,13 @@ class Command(object):
parser.add_argument("-l", "--lockfile", action=OnceArgument, nargs='?', const=".",
help="Path to a lockfile or folder containing 'conan.lock' file. "
"Lockfile will be updated with the exported package")
+ parser.add_argument("-id", "--ignore-dirty", default=False, action='store_true',
+ help='When using the "scm" feature with "auto" values, capture the'
+ 'commit and url even if there are uncommitted changes')
args = parser.parse_args(*args)
- self._warn_python2()
+ self._warn_python_version()
name, version, user, channel, _ = get_reference_fields(args.reference,
user_channel_input=True)
cwd = os.getcwd()
| [Command->[export->[export],info->[info],install->[install],editable->[info],source->[source],remove->[remove,info],new->[new],imports->[imports],upload->[upload],copy->[copy],download->[download],run->[_commands,_warn_python2,_show_help],export_pkg->[export_pkg],test->[test],inspect->[inspect],package->[package],_show_help->[check_all_commands_listed],create->[create],build->[build]],main->[Command,run]] | Exports a package from the specified path. Handles command line options for the conan package. | -id might not be the best short expression. |
@@ -581,6 +581,7 @@ void ContentFeatures::deSerialize(std::istream &is)
// read legacy properties
legacy_facedir_simple = readU8(is);
legacy_wallmounted = readU8(is);
+ node_dig_prediction = deSerializeString(is);
}
#ifndef SERVER
| [removeNode->[getId],addNameIdMapping->[set],getIdFromNrBacklog->[getId],getNodeBoxUnion->[boxVectorUnion],applyTextureOverrides->[getId],allocateId->[ContentFeatures],getId->[getId],getIdsFromNrBacklog->[getIds,getId],nodeResolveInternal->[clear],reserve->[reserve],serializeOld->[serialize],updateTextures->[clear,updateTextures,fillTileAttribs,readSettings,correctAlpha,getPalette], reset->[reset],allocateDummy->[set],deSerialize->[getNodeBoxUnion,reset,getId,deSerialize,clear],getIds->[getId],updateAliases->[getId,clear],runNodeResolveCallbacks->[clear],serialize->[serialize],ContentFeatures->[getId],set->[allocateId,getId,getNodeBoxUnion],clear->[reset,clear],resetNodeResolveState->[clear],mapNodeboxConnections->[getIds],deSerializeOld->[deSerialize,clear]] | Reads a content feature section from the stream. Reads all network related objects from the stream. Private method for reading a single object from a file - like object region Animation Frame. | shouldn't this be in a try-catch? |
@@ -28,6 +28,11 @@ func TestGenProgram(t *testing.T) {
continue
}
+ if filepath.Base(f.Name()) == "azure-native.pp" {
+ // The generated code fails to compile
+ continue
+ }
+
t.Run(f.Name(), func(t *testing.T) {
path := filepath.Join(testdataPath, f.Name())
contents, err := ioutil.ReadFile(path)
| [NewHost,SortedValues,BindProgram,ReadFile,ReadDir,Join,Equal,collectImports,Name,NewParser,NoError,Fatalf,HasErrors,Ext,Base,PluginHost,NewFormatter,ParseFile,NewStringSet,NewReader,Run] | TestGenProgram imports the given and runs it through the given t. TestCollectImports tests that the given program has the same name as the main. go file. | This Go program generation would fail with the code in master. This PR doesn't fix it but it doesn't make it worse either, so I suggest filing a separate issue for later. |
@@ -37,6 +37,10 @@ public class Unit extends GameDataComponent {
public PlayerID getOwner() {
return m_owner;
}
+
+ public UnitAttachment getUnitAttachment() {
+ return (UnitAttachment) m_type.getAttachment("unitAttachment");
+ }
/**
* @deprecated DO NOT USE THIS METHOD if at all possible. It is very slow.
| [Unit->[hashCode->[hashCode,getID,getTerritoryUnitIsIn],isEquivalent->[getHits,getOwner,getType,equals],toString->[getID,getTerritoryUnitIsIn],equals->[equals]]] | Get the owner of the player. | This is good, information hiding, the clients of this class do not have to know about quite as many details. Changes like this add up quickly to allow a person to use high level of abstractions :+1: |
@@ -47,6 +47,8 @@ class ConfigBase:
They will be converted to snake_case automatically.
If a field is missing and don't have default value, it will be set to `dataclasses.MISSING`.
"""
+ if 'basepath' in kwargs:
+ _base_path = kwargs.pop('basepath')
kwargs = {util.case_insensitive(key): value for key, value in kwargs.items()}
if _base_path is None:
_base_path = Path()
| [ConfigBase->[validate->[_is_missing,validate,canonical],load->[load],canonical->[canonical],__init__->[_is_missing]]] | Initialize a config object and set some fields. | what is basepath used for? |
@@ -342,9 +342,13 @@ class PostMover
notifications_changed_at, notifications_reason_id)
SELECT tu.user_id,
:new_topic_id AS topic_id,
- CASE
- WHEN p.user_id IS NULL THEN FALSE
- ELSE TRUE END AS posted,
+ EXISTS(
+ SELECT 1
+ FROM posts p
+ WHERE p.topic_id = :new_topic_id
+ AND p.user_id = tu.user_id
+ LIMIT 1
+ ) AS posted,
(
SELECT MAX(lr.new_post_number)
FROM moved_posts lr
| [PostMover->[to_new_topic->[move_types],to_topic->[move_types],update_statistics->[update_statistics],enqueue_jobs->[enqueue_jobs]]] | Copy TopicUsers . region NotificationLevelNotifications interface. last_emailed_post_number first_visited_at last_visited_at. | I think it auto optimises, but you might as well do a `limit 1` once done feel free to merge |
@@ -1199,6 +1199,10 @@ namespace System
throw new ArgumentOutOfRangeException(nameof(typeToken),
SR.Format(SR.Argument_InvalidToken, typeToken, new ModuleHandle(module)));
+ // defensive copy of user-provided array, per CopyRuntimeTypeHandles contract
+ typeInstantiationContext = (RuntimeTypeHandle[]?)typeInstantiationContext?.Clone();
+ methodInstantiationContext = (RuntimeTypeHandle[]?)methodInstantiationContext?.Clone();
+
IntPtr[]? typeInstantiationContextHandles = RuntimeTypeHandle.CopyRuntimeTypeHandles(typeInstantiationContext, out int typeInstCount);
IntPtr[]? methodInstantiationContextHandles = RuntimeTypeHandle.CopyRuntimeTypeHandles(methodInstantiationContext, out int methodInstCount);
| [RuntimeFieldHandle->[MdUtf8String->[_GetUtf8Name],Equals],RuntimeTypeHandle->[GetConstraints->[GetConstraints],SatisfiesConstraints->[SatisfiesConstraints,CopyRuntimeTypeHandles],GetInstantiationPublic->[GetInstantiation],GetHashCode->[GetHashCode],GetGenericVariableIndex->[IsGenericVariable,GetGenericVariableIndex],GetTypeByName->[GetTypeByName],IsVisible->[_IsVisible],VerifyInterfaceIsImplemented->[VerifyInterfaceIsImplemented],GetInstantiationInternal->[GetInstantiation],MdUtf8String->[_GetUtf8Name],RuntimeType->[Instantiate,MakeByRef,GetTypeByNameUsingCARules,GetGenericTypeDefinition,HasInstantiation,MakePointer,MakeSZArray,MakeArray,CopyRuntimeTypeHandles],ConstructName->[ConstructName],ContainsGenericVariables->[ContainsGenericVariables]],RuntimeMethodHandle->[GetName->[GetName],GetSlot->[GetSlot],IRuntimeMethodInfo->[IsTypicalMethodDefinition,StripMethodInstantiation,GetTypicalMethodDefinition],GetMethodInstantiationInternal->[GetMethodInstantiation],MdUtf8String->[_GetUtf8Name],GetCurrentMethod->[_GetCurrentMethod],IsGenericMethodDefinition->[IsGenericMethodDefinition],ConstructInstantiation->[ConstructInstantiation],HasMethodInstantiation->[HasMethodInstantiation],GetGenericParameterCount->[GetGenericParameterCount],GetMethodInstantiationPublic->[GetMethodInstantiation],Equals],Signature->[GetSignature],ModuleHandle->[RuntimeMethodHandleInternal->[ValidateModulePointer],GetHashCode->[GetHashCode],IRuntimeMethodInfo->[CopyRuntimeTypeHandles],IRuntimeFieldInfo->[ValidateModulePointer,CopyRuntimeTypeHandles],GetPEKind->[GetPEKind],ContainsPropertyMatchingHash->[_ContainsPropertyMatchingHash],RuntimeType->[GetModuleType,ValidateModulePointer,CopyRuntimeTypeHandles],Equals,GetMDStreamVersion]] | Internal method to resolve a type handle. | Was this a potential GC bug in the original? i.e. no clone of `methodInstantiationContext`? |
@@ -77,7 +77,11 @@ func (c *Command) RunInDirTimeoutEnvFullPipeline(env []string, timeout time.Dura
defer cancel()
cmd := exec.CommandContext(ctx, c.name, c.args...)
- cmd.Env = env
+ if env == nil {
+ cmd.Env = append(os.Environ(), fmt.Sprintf("LC_ALL=%s", DefaultLCALL))
+ } else {
+ cmd.Env = append(env, fmt.Sprintf("LC_ALL=%s", DefaultLCALL))
+ }
cmd.Dir = dir
cmd.Stdout = stdout
cmd.Stderr = stderr
| [RunInDirWithEnv->[RunInDirTimeoutEnv],RunInDirFullPipeline->[RunInDirTimeoutFullPipeline],RunInDirTimeoutEnv->[String,RunInDirTimeoutEnvPipeline],RunInDirTimeoutEnvFullPipeline->[String],RunInDirTimeoutPipeline->[RunInDirTimeoutEnvPipeline],Run->[RunTimeout],RunInDirTimeoutFullPipeline->[RunInDirTimeoutEnvFullPipeline],RunInDirBytes->[RunInDirTimeout],RunTimeout->[RunInDirTimeout]] | RunInDirTimeoutEnvFullPipeline runs the command in the specified directory with the specified environment. | In this case `env` is expected to include `os.Environ()`? |
@@ -56,10 +56,16 @@ public class CompressorHttp2ConnectionEncoder extends DecoratingHttp2ConnectionE
public CompressorHttp2ConnectionEncoder(Http2ConnectionEncoder delegate, int compressionLevel, int windowBits,
int memLevel) {
+ this(delegate, compressionLevel, windowBits, memLevel, new Encoder.Parameters().setQuality(4));
+ }
+
+ public CompressorHttp2ConnectionEncoder(Http2ConnectionEncoder delegate, int compressionLevel, int windowBits,
+ int memLevel, Encoder.Parameters parameters) {
super(delegate);
this.compressionLevel = ObjectUtil.checkInRange(compressionLevel, 0, 9, "compressionLevel");
this.windowBits = ObjectUtil.checkInRange(windowBits, 9, 15, "windowBits");
this.memLevel = ObjectUtil.checkInRange(memLevel, 1, 9, "memLevel");
+ this.parameters = ObjectUtil.checkNotNull(parameters, "Parameters");
propertyKey = connection().newKey();
connection().addListener(new Http2ConnectionAdapter() {
| [CompressorHttp2ConnectionEncoder->[writeHeaders->[writeHeaders],newCompressor->[getTargetContentEncoding,newContentCompressor],writeData->[writeData]]] | On stream removed remove any orphan channels. | Do you think it makes sense to accept a list of `CompressionOptions` instead? |
@@ -44,7 +44,7 @@ namespace System.Numerics
Array.Copy(value, _bits, _length);
}
- public unsafe void MultiplySelf(ref BitsBuffer value,
+ public void MultiplySelf(ref BitsBuffer value,
ref BitsBuffer temp)
{
Debug.Assert(temp._length == 0);
| [BigIntegerCalculator->[BitsBuffer->[Reduce->[Reduce]]]] | Multiplies this and value to temp and writes the result to temp. | Nit: align argument with line above. |
@@ -108,7 +108,7 @@ public class GenericServiceTest {
FullServiceDefinition fullServiceDefinition = ServiceDefinitionBuilder.buildFullDefinition(DemoService.class);
MethodDefinition methodDefinition = getMethod("complexCompute", fullServiceDefinition.getMethods());
- Map parm2= createComplextObject(fullServiceDefinition,var1, var2, l, var3, var4, testEnum);
+ Map parm2= createComplexObject(fullServiceDefinition,var1, var2, l, var3, var4, testEnum);
ComplexObject complexObject = map2bean(parm2);
Invoker<GenericService> invoker = protocol.refer(GenericService.class, url);
| [GenericServiceTest->[bean2Map->[toJSONString,parseObject],map2bean->[toJSONString,parseObject],getMethod->[equals],testGenericComplexCompute4FullServiceMetadata->[map2bean,toString,getParameterTypes,DemoServiceImpl,getMethod,getMethods,destroy,asList,valueOf,unexport,export,createComplextObject,$invoke,getProxy,assertEquals,refer,getName,getInvoker,buildFullDefinition,getAdaptiveExtension],testGeneric->[DemoServiceImpl,getName,getInvoker,unexport,export,$invoke,destroy,refer,getProxy,getAdaptiveExtension,assertEquals,valueOf],createComplextObject->[getName,equals,getType,contains,assertSame,add,name,getTypes,put,assertEquals],createComplexObject->[ComplexObject],testGenericFindComplexObject4FullServiceMetadata->[DemoServiceImpl,assertNotNull,map2bean,getName,getInvoker,asList,unexport,export,createComplexObject,$invoke,destroy,refer,getProxy,getAdaptiveExtension,assertEquals,valueOf],testGeneric2->[DemoServiceImpl,getName,getInvoker,unexport,export,$invoke,destroy,refer,getProxy,getAdaptiveExtension,assertEquals,valueOf]]] | Test generic complex compute 4 full service metadata. Destroy the invoker and unexport all exporters. | The name `parm2` looks weird to me. Could you improve it? And there should be a space before the equal sign. |
@@ -244,9 +244,9 @@ class PluginTestBase(WebTest):
unit['_storage_path'] = storage_path
if n == 0: # 1st one is a directory of files
os.makedirs(storage_path)
- path = os.path.join(storage_path, 'abc.rpm')
- with open(path, 'w+') as fp:
- fp.write(path)
+ tb = tarfile.open('data/distribution.tar')
+ tb.extractall(path=storage_path)
+ tb.close()
else:
with open(storage_path, 'w+') as fp:
fp.write(unit_id)
| [TestDistributor->[test_entry_point->[Repository,validate_config],test_payload->[Repository,dist_conf,populate],test_metadata->[metadata],test_config_missing_http_protocol->[Repository,validate_config],test_config_missing_alias->[Repository,validate_config],test_config_missing_protocol->[Repository,validate_config],test_config_missing_invalid_alias->[Repository,validate_config],test_publish->[Repository,publish_repo,dist_conf,populate],test_valid_config->[Repository,validate_config]],TestEndToEnd->[clean->[clean,add_units,define_plugins],test_handler_mirror_repository_scope->[test_handler->[dist_conf,publish_repo,Repository,AgentConduit,populate],verify,test_handler,AdditiveTestStrategy],populate->[dist_conf,populate],test_importer_exception->[test_handler->[dist_conf,publish_repo,Repository,AgentConduit,populate],verify,MirrorTestStrategy,test_handler],test_missing_plugins->[test_handler->[dist_conf,publish_repo,Repository,AgentConduit,populate],MirrorTestStrategy,test_handler],test_handler_additive->[test_handler->[dist_conf,publish_repo,Repository,AgentConduit,populate],verify,test_handler,AdditiveTestStrategy],test_handler_merge->[test_handler->[dist_conf,publish_repo,Repository,AgentConduit,populate],verify,MirrorTestStrategy,test_handler],test_handler_nothing_updated->[test_handler->[dist_conf,publish_repo,Repository,AgentConduit,populate],test_handler],test_handler_merge_and_delete_extra_units->[test_handler->[dist_conf,publish_repo,Repository,AgentConduit,populate],verify,MirrorTestStrategy,test_handler],test_handler_merge_and_delete_repositories->[test_handler->[dist_conf,publish_repo,Repository,AgentConduit,populate],verify,MirrorTestStrategy,test_handler],test_handler_unit_errors->[test_handler->[dist_conf,publish_repo,Repository,AgentConduit,populate],verify,MirrorTestStrategy,test_handler],test_repository_handler->[test_handler->[dist_conf,publish_repo,Repository,AgentConduit,populate],verify,MirrorTestStrategy,test_handler],test_handler_mirror->[test_handler->[dist_conf,publish_repo,Repository,AgentConduit,populate],verify,MirrorTestStrategy,test_handler],test_handler_merge_dist_changed->[test_handler->[dist_conf,publish_repo,Repository,AgentConduit,populate],verify,MirrorTestStrategy,test_handler],test_handler_cancelled->[test_handler->[dist_conf,publish_repo,Repository,AgentConduit,populate],MirrorTestStrategy,test_handler],test_handler_content_skip->[test_handler->[dist_conf,publish_repo,Repository,AgentConduit,populate],MirrorTestStrategy,test_handler]],ImporterTest->[test_entry_point->[Repository,validate_config],test_import_cached_manifest_matched->[dist_conf,define_plugins,publish_repo,Repository,populate],test_import_cached_manifest_missing_units->[dist_conf,define_plugins,publish_repo,Repository,populate],test_metadata->[metadata],test_import->[dist_conf,define_plugins,publish_repo,Repository,populate],test_import_cached_manifest_units_invalid->[dist_conf,define_plugins,publish_repo,Repository,populate],test_import_unit_files_already_exist_size_mismatch->[dist_conf,define_plugins,publish_repo,Repository,populate],test_invalid_strategy->[Repository,validate_config],test_import_unit_files_already_exist->[dist_conf,define_plugins,publish_repo,Repository,populate],test_import_modified_units->[Repository,dist_conf,publish_repo,populate],test_config_missing_properties->[Repository,validate_config],test_valid_config->[Repository,validate_config]],PluginTestBase->[tearDown->[tearDown],setUp->[tmpdir,setUp]],TestProfiler->[test_update_units->[node_configuration]],AgentHandlerTest->[test_repository_handler_get_bindings_failed->[AgentConduit],test_node_handler_get_bindings_failed->[AgentConduit]],BadDownloadRequest->[__init__->[__init__]]] | Add units to the database. | this should probably be in a try/finally block in case there's an error during extraction, so we guarantee the file will be closed. |
@@ -12,6 +12,7 @@
#include <errno.h>
#include <stdio.h>
#include <stdlib.h>
+#include <unistd.h>
#include <string.h>
#include <openssl/crypto.h>
| [RAND_load_file->[OPENSSL_cleanse,openssl_fopen,fclose,S_ISREG,RANDerr,RAND_add,fread,stat,ERR_add_error_data],RAND_write_file->[OPENSSL_cleanse,fwrite,openssl_fopen,fclose,S_ISREG,chmod,RANDerr,RAND_bytes,fdopen,open,stat,ERR_add_error_data],char->[strcat,strlen,getenv,GetEnvironmentVariableW,_alloca,strcpy,OPENSSL_issetugid,WideCharToMultiByte]] | Creates a new object. This function is a hack to get around vms s extension to fopen. It is a hack. | Did anyone already point out that Windows doesn't have `unistd.h`? |
@@ -25,7 +25,7 @@ namespace Microsoft.Extensions.FileProviders.Physical
/// </summary>
public class PhysicalFilesWatcher : IDisposable
{
- private static readonly Action<object> _cancelTokenSource = state => ((CancellationTokenSource)state).Cancel();
+ private static readonly Action<object?> _cancelTokenSource = state => ((CancellationTokenSource?)state)?.Cancel();
internal static TimeSpan DefaultPollingInterval = TimeSpan.FromSeconds(4);
| [PhysicalFilesWatcher->[OnError->[Keys,ReportChangeForMatchedEntries],RaiseChangeEvents->[TryRemove,HasChanged,Cancel,Key],OnFileSystemEntryChange->[IsExcluded,Length,ReportChangeForMatchedEntries,Substring],TryEnableFileSystemWatcher->[EnableRaisingEvents,IsEmpty],OnRenamed->[Exists,OnFileSystemEntryChange,EnumerateFileSystemEntries,OldFullPath,FullPath,AllDirectories,Substring,Combine,Length],Dispose->[SuppressFinalize,Dispose],ReportChangeForMatchedEntries->[Match,TryRemove,IsNullOrEmpty,TryDisableFileSystemWatcher,HasMatches,CancelToken,NormalizePath,Key],TryDisableFileSystemWatcher->[EnableRaisingEvents,IsEmpty],IsDirectoryPath->[DirectorySeparatorChar,AltDirectorySeparatorChar,Length],CancelToken->[TokenSource,IsCancellationRequested,None,Default,StartNew,DenyChildAttach],IChangeToken->[CancellationTokenSource,EnsureInitialized,PathNavigatesAboveRoot,GetOrAddChangeToken,ChangeToken,NormalizePath,GetOrAddWildcardChangeToken,Token,TryGetValue,IsPathRooted,IndexOf,GetOrAddFilePathChangeToken,Combine,TryAdd,GetOrAdd,ActiveChangeCallbacks,nameof,TryEnableFileSystemWatcher,IsDirectoryPath,OrdinalIgnoreCase,AddInclude,Singleton],NormalizePath->[Replace],OnChanged->[FullPath,OnFileSystemEntryChange],FileSystemWatcher_PlatformNotSupported,Deleted,Error_FileSystemWatcherRequiredWithoutPolling,Create,Error,Format,nameof,IncludeSubdirectories,Changed,Renamed,IsBrowser,FromSeconds,Created,Cancel,OrdinalIgnoreCase,Zero,Sensitive]] | Creates a new instance of a file watcher that watches a physical filesystem for changes and then invokes Checks if a node is watching changes. | avoid any product changes here. Please do them separately (I'd suggest filing an issue, ideally with a repro if you think this is a product bug), otherwise `!` if you feel confident enough this will not be null (or if this will always NRE and you've filed an issue) |
@@ -69,11 +69,15 @@ public class JobExecutionInfoServerTest {
Properties properties = new Properties();
properties.setProperty(ConfigurationKeys.JOB_HISTORY_STORE_JDBC_DRIVER_KEY, "org.apache.derby.jdbc.EmbeddedDriver");
properties.setProperty(ConfigurationKeys.JOB_HISTORY_STORE_URL_KEY, "jdbc:derby:memory:gobblin;create=true");
+
+ String randomPort = chooseRandomPort();
+ properties.setProperty(ConfigurationKeys.REST_SERVER_PORT_KEY, randomPort);
+
prepareJobHistoryStoreDatabase(properties);
Injector injector = Guice.createInjector(new MetaStoreModule(properties));
this.jobHistoryStore = injector.getInstance(JobHistoryStore.class);
- this.client = new JobExecutionInfoClient("http://localhost:8080/");
+ this.client = new JobExecutionInfoClient(String.format("%s:%s/", "http://localhost", randomPort));
this.server = new JobExecutionInfoServer(properties);
this.server.startUp();
| [JobExecutionInfoServerTest->[assertMetric->[getValue,getName,getType,getGroup,assertEquals],testBatchGet->[JobExecutionQuery,create,newArrayList,getJobId,size,equals,assertJobExecution,get,setId,getJobName,batchGet,assertEquals,setIdType],assertJobExecution->[assertMetric,longValue,getJobId,getState,getCompletedTasks,size,getDuration,getLaunchedTasks,getJobProperties,get,hasDuration,getJobName,assertTaskExecution,assertEquals,getMetrics],testBadGet->[assertTrue,JobExecutionQuery,create,getJobId,get,setId,isEmpty,setIdType],testGet->[JobExecutionQuery,create,getJobId,getJobExecutions,size,assertJobExecution,get,setId,assertEquals,setIdType],prepareJobHistoryStoreDatabase->[prepareStatement,readLines,split,newArrayList,MetaStoreModule,of,isPresent,get,createInjector,getInstance,absent,getConnection,join,close,add,File,startsWith,execute],createJobExecutionInfo->[setState,setJobName,StringMap,JobExecutionInfo,TaskExecutionInfoArray,setName,put,setHighWatermark,setGroup,setTaskProperties,getJobId,setType,setCompletedTasks,setJobId,setJobProperties,setValue,MetricArray,setTable,setMetrics,getJobName,currentTimeMillis,TaskExecutionInfo,setTaskExecutions,Metric,newHashMap,setStartTime,setTaskId,setLowWatermark,Table,setNamespace,add,setLaunchedTasks],assertTaskExecution->[getLowWatermark,getTable,assertMetric,longValue,getTaskProperties,getJobId,getState,size,getDuration,get,hasDuration,getTaskId,getHighWatermark,assertEquals,getMetrics],setUp->[MetaStoreModule,JobExecutionInfoServer,startUp,createInjector,JobExecutionInfoClient,getInstance,prepareJobHistoryStoreDatabase,setProperty,Properties,createJobExecutionInfo,put],tearDown->[getConnection,shutDown,close]]] | Sets up the necessary properties. | I think doing `"http://%s:%s/"` is sufficient here. |
@@ -15,6 +15,7 @@
CALL_WAITING_STATE = 'waiting'
CALL_SKIPPED_STATE = 'skipped'
+CALL_ACCEPTED_STATE = 'accepted'
CALL_RUNNING_STATE = 'running'
CALL_SUSPENDED_STATE = 'suspended'
CALL_FINISHED_STATE = 'finished'
| [No CFG could be retrieved] | - - - - - - - - - - - - - - - - - - Group distributor. | Agent timeout story: need to support 'accepted' status so we can update the task status. That way REST callers can make decisions when agent related tasks dont progress. |
@@ -64,11 +64,12 @@ const (
BasicUserRoleName = "basic-user"
StatusCheckerRoleName = "cluster-status"
- BuildControllerRoleName = "system:build-controller"
- ReplicationControllerRoleName = "system:replication-controller"
- DeploymentControllerRoleName = "system:deployment-controller"
- JobControllerRoleName = "system:job-controller"
- HPAControllerRoleName = "system:hpa-controller"
+ BuildControllerRoleName = "system:build-controller"
+ ReplicationControllerRoleName = "system:replication-controller"
+ DeploymentControllerRoleName = "system:deployment-controller"
+ JobControllerRoleName = "system:job-controller"
+ HPAControllerRoleName = "system:hpa-controller"
+ PersistentVolumeControllerRoleName = "system:pv-controller"
ImagePullerRoleName = "system:image-puller"
ImageBuilderRoleName = "system:image-builder"
| [No CFG could be retrieved] | The names of the groups that are defined in the System module. This function is used to set the required role bindings for the Nagios node. | you need to define this role in `bootstrappolicy/policy.go` (see `JobControllerRoleName` definition as an example). Include all API calls, verbs and objects the persistent volume controller's client will need to be permitted to do. |
@@ -4606,6 +4606,15 @@ Sedp::DiscoveryReader::data_received_i(const DCPS::ReceivedDataSample& sample,
wdata.ice_agent_info_ = pos->second;
}
#endif
+
+ // Cache non-empty TypeInformation
+ if (wdata.type_info_.minimal.typeid_with_size.type_id.kind() != XTypes::TK_NONE ||
+ wdata.type_info_.complete.typeid_with_size.type_id.kind() != XTypes::TK_NONE) {
+ const GUID_t& remote_guid = wdata.writer_data_.writerProxy.remoteWriterGuid;
+ DDS::BuiltinTopicKey_t key = DCPS::repo_id_to_bit_key(remote_guid);
+ sedp_.type_lookup_service_->cache_type_info(key, wdata.type_info_);
+ }
+
sedp_.data_received(id, wdata);
#ifdef OPENDDS_SECURITY
| [No CFG could be retrieved] | - > Extract data from ParameterList - > DiscoveredWriterData - > DiscoveredWriter list - list of parameters. | Should this and the similar following sections of code be put into a function? They seem to follow the same pattern. |
@@ -231,10 +231,6 @@ final class ShowQueriesRewrite
catalogName = qualifiedTableName.getCatalogName();
- accessControl.checkCanShowTablesMetadata(
- session.toSecurityContext(),
- new CatalogSchemaName(catalogName, qualifiedTableName.getSchemaName()));
-
predicate = Optional.of(combineConjuncts(
metadata,
equal(identifier("table_schema"), new StringLiteral(qualifiedTableName.getSchemaName())),
| [ShowQueriesRewrite->[Visitor->[buildProperties->[getExpression]]]] | Visit a ShowGrants node. This method is called when the table is not found in the catalog. | Can you file an issue for this one? I believe we should have some kind of access control on `SHOW GRANTS FOR TABLE`. My guess would be you can always see your own grants, or if you are the owner, you can see all grants... or maybe the "with grant option". We'll need to look it up in the spec. |
@@ -333,7 +333,8 @@ public class UpdateCenter extends AbstractModelObject implements Saveable, OnMas
}
return HttpResponses.okJSON(checkJob.connectionStates);
} else {
- return HttpResponses.errorJSON(String.format("Unknown site '%s'.", siteId));
+ return HttpResponses.errorJSON(String.format("Cannot check connection status of the update site with ID='%s'"
+ + ". This update center cannot be resolved", siteId));
}
} catch (Exception e) {
return HttpResponses.errorJSON(String.format("ERROR: %s", e.getMessage()));
| [UpdateCenter->[getConnectionCheckJob->[getSite,getConnectionCheckJob],getAvailables->[getAvailables],getCategorizedAvailables->[getAvailables],UpdateCenterJob->[submit->[submit]],isRestartScheduled->[getJobs],HudsonDowngradeJob->[getURL->[getData],onSuccess->[Success],run->[onSuccess,_run,getName,Success,Failure],_run->[install,getDestination]],DownloadJob->[_run->[getURL,getDestination,download,preValidate,install,postValidate],run->[getName,onSuccess,Failure],Pending],getPlugin->[getPlugin],PluginEntry->[compareTo->[compareTo]],doUpgrade->[getCoreSource],doRestart->[getCoreSource],doInstallStatus->[getJobs],doInvalidateData->[doInvalidateData],getUpdates->[getUpdates],getHudsonJob->[getJobs],init->[load],doDowngrade->[getCoreSource,isDowngradable],CoreUpdateMonitor->[getData->[getCoreSource,getData]],EnableJob->[run->[getDisplayName]],doSafeRestart->[getCoreSource],HudsonUpgradeJob->[getURL->[getData],onSuccess->[Success],replace->[verifyChecksums,getComputedSHA1]],updateAllSites->[getSites],getJob->[getJobs],persistInstallStatus->[getJobs,persistInstallStatus],PluginDowngradeJob->[getName->[getDisplayName],run->[onSuccess,_run,getName,Success,Failure],toString->[toString],_run->[install,getDestination,getBackup]],InstallationJob->[_run->[_run,getDestination,SuccessButRequiresRestart,getDisplayName,Skipped],getName->[getDisplayName],toString->[toString],wasInstalled->[getJobs,getDisplayName],replace->[verifyChecksums,getLegacyDestination,getComputedSHA1]],ConnectionCheckJob->[run->[getConnectionCheckUrl,checkUpdateCenter,submit]],getName]] | Checks connection status. | NIT: missing closing period |
@@ -177,10 +177,12 @@ public class HoodieMetadataPayload implements HoodieRecordPayload<HoodieMetadata
/**
* Returns the files added as part of this record.
*/
- public FileStatus[] getFileStatuses(Path partitionPath) {
+ public FileStatus[] getFileStatuses(Configuration hadoopConf, Path partitionPath) throws IOException {
+ FileSystem fs = partitionPath.getFileSystem(hadoopConf);
+ long blockSize = fs.getDefaultBlockSize(partitionPath);
return filterFileInfoEntries(false)
- .map(e -> new FileStatus(e.getValue().getSize(), false, 0, 0, 0, 0, null, null, null,
- new Path(partitionPath, e.getKey())))
+ .map(e -> new FileStatus(e.getValue().getSize(), false, 0, blockSize, 0, 0,
+ null, null, null, new Path(partitionPath, e.getKey())))
.toArray(FileStatus[]::new);
}
| [HoodieMetadataPayload->[createPartitionFilesRecord->[HoodieMetadataPayload],combineAndGetUpdateValue->[preCombine,HoodieMetadataPayload],preCombine->[HoodieMetadataPayload],toString->[toString],createPartitionListRecord->[HoodieMetadataPayload]]] | Returns an array of FileStatus objects for all files in the partition. | @umehrot2 Since we are fixing 0 -> blockSize, just curious, are the other `0` arguments valid or should they also have some dynamic value ? We don't have to fix it as part of the PR but would be great if you can take a look.. |
@@ -1,7 +1,7 @@
<?php
/**
* Module Name: WP.me Shortlinks
- * Module Description: Generates shorter links so you can have more space to write on social media sites.
+ * Module Description: Generates shorter links using the wp.me domain.
* Sort Order: 8
* First Introduced: 1.1
* Requires Connection: Yes
| [wpme_get_shortlink->[get_queried_object_id]] | Shortlink link generator Returns the URL of the node with the link to the post. | Noting that the translation for this get auto-generated by a build script by the looks of it - let me know if I need to run that as part of the PR. |
@@ -39,6 +39,15 @@ TrussElementLinear3D2N::Create(IndexType NewId,
NewId, rGeom.Create(rThisNodes), pProperties);
}
+Element::Pointer
+TrussElementLinear3D2N::Create(IndexType NewId,
+ GeometryType::Pointer pGeom,
+ PropertiesType::Pointer pProperties) const {
+ const GeometryType &rGeom = this->GetGeometry();
+ return Kratos::make_shared<TrussElementLinear3D2N>(
+ NewId, pGeom, pProperties);
+}
+
TrussElementLinear3D2N::~TrussElementLinear3D2N() {}
BoundedMatrix<double, TrussElement3D2N::msLocalSize,
| [No CFG could be retrieved] | The functions below are defined in the System module. - - - - - - - - - - - - - - - - - -. | is this intended? It is not used... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.