id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
2087104585
Work around SwiftPM build regression affecting packages with C++ targets This is a workaround to the SwiftPM issue being fixed here: https://github.com/apple/swift-package-manager/pull/7267 @swift-ci please test No longer relevant, the SwiftPM bug was fixed!
gharchive/pull-request
2024-01-17T22:22:01
2025-04-01T04:33:30.749645
{ "authors": [ "stmontgomery" ], "repo": "apple/swift-testing", "url": "https://github.com/apple/swift-testing/pull/188", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1909296675
Freestanding declaration macro fails inside classes: Expected '{' in body of function declaration Description Using a freestanding declaration macro inside a class produces a build error: Expected '{' in body of function declaration Here is the expanded macro: func myFunction() { } The same code compiles if the macro is used inside a struct. Steps to reproduce Create a swift package using the macro template. Client code: import MyMacro struct MyStruct { #blankFunction // works ok here } class MyClass { #blankFunction // fails here } Macro declaration: @freestanding(declaration, names: named(myFunction)) public macro blankFunction() = #externalMacro( module: "MyMacroMacros", type: "BlankFunctionMacro" ) Macro implementation: import SwiftCompilerPlugin import SwiftSyntax import SwiftSyntaxBuilder import SwiftSyntaxMacros public struct BlankFunctionMacro: DeclarationMacro { public static func expansion( of node: some FreestandingMacroExpansionSyntax, in context: some MacroExpansionContext ) throws -> [DeclSyntax] { return ["func myFunction() {}"] } } @main struct MyMacroPlugin: CompilerPlugin { let providingMacros: [Macro.Type] = [ BlankFunctionMacro.self, ] } Expected behavior swift run MyMacroClient should compile and run without any errors. Environment swift-driver version: 1.87.1 Apple Swift version 5.9 (swiftlang-5.9.0.128.108 clang-1500.0.40.1) Target: arm64-apple-macosx13.0 Xcode 15.0 Build version 15A240d Deployment target: macOS 13.6 (22G120) Hi facing the same issue now. Just a follow-up. What's the progress regarding this issue? Thank you! This also fails in the same way when adding a 'static func...' to anything, structs included. Also here. Thanks for reporting, hoping for a fix... It's also current blocking me on a macros-heavy project. Same here, blocking awesome ideas realisation. Any updates on this? Tried on Swift 5.9.2 and 5.10 - doesn't work. 🙅‍♂️ Tried on Swift 6.0 DEVELOPMENT-SNAPSHOT-2024-04-30-a-osx and it works! 🚀
gharchive/issue
2023-09-22T17:43:40
2025-04-01T04:33:30.755919
{ "authors": [ "MihaelIsaev", "drseg", "huyaoyu", "juozasvalancius", "regexident" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/issues/68704", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
246486426
[sil-combine] Make sure that (apply (partial_apply)) -> (apply') bails if we have an argument with a dependent type. [sil-combine] Make sure that (apply (partial_apply)) -> (apply') bails if we have an argument with a dependent type. What is going on here is that currently this optimization if itneeds to perform lfietime extension always creates an alloc_stack at the beginning/end of a function. If the object whose lifetime is being extended has a dependent type, then the alloc_stack will be created before the dependent type exists resulting in the compiler crashing. rdar://33595317 @swift-ci smoke test and merge
gharchive/pull-request
2017-07-29T00:23:52
2025-04-01T04:33:30.758128
{ "authors": [ "gottesmm" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/11253", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
307319870
[RemoteMirrors] Interop header fixes and library lookup caching. Some deficiencies in the interop API were identified when an actual client started trying to use it. typeinfo and childinfo needed interop variants that used interop variants for their fields. The client needed to be able to look things up starting from a metadata address directly, rather than starting from an object. Figuring out which library owns a particular metadata address is not particularly fast, but is done frequently, so it needed a cache. rdar://problem/37538580 @swift-ci please test
gharchive/pull-request
2018-03-21T16:25:03
2025-04-01T04:33:30.760198
{ "authors": [ "mikeash" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/15395", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
141752602
[Coverage] Respect function linkage in PGO name variables What's in this pull request? The Ubuntu bots have a problem with the ordered IR check lines in test/IRGen/coverage.swift. The intended match appears later in the file, so try again with CHECK-DAG lines. Resolved bug number: (SR-) Before merging this pull request to apple/swift repository: [ ] Test pull request on Swift continuous integration. Triggering Swift CI The swift-ci is triggered by writing a comment on this PR addressed to the GitHub user @swift-ci. Different tests will run depending on the specific comment that you use. The currently available comments are: Smoke Testing Platform | Comment ------------ | ------------- All supported platforms | @swift-ci Please smoke test OS X platform | @swift-ci Please smoke test OS X platform Linux platform | @swift-ci Please smoke test Linux platform Validation Testing Platform | Comment ------------ | ------------- All supported platforms | @swift-ci Please test OS X platform | @swift-ci Please test OS X platform Linux platform | @swift-ci Please test Linux platform Note: Only members of the Apple organization can trigger swift-ci. @swift-ci Please test Linux platform __DATA doesn't exist on ELF. Fixed the test case in dc11f8bf.
gharchive/pull-request
2016-03-18T01:54:41
2025-04-01T04:33:30.764530
{ "authors": [ "vedantk" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/1729", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
377504583
Revert "Revert "[Build System: CMake] make add_swift_library a wrappe… …r to add_llvm_library"" This reverts commit 103f9a8246beb385ca6d8a33003a0ebac035095d. Replace this paragraph with a description of your changes and rationale. Provide links to external references/discussions if appropriate. Resolves SR-NNNN. @gottesmm kindly volunteered to help sort out the internal CI conflicts. @swift-ci please test Build failed Swift Test OS X Platform Git Sha - 1efdb2954778d9baac05201dfc3b857e77612219 Build failed Swift Test Linux Platform Git Sha - 1efdb2954778d9baac05201dfc3b857e77612219 @swift-ci please test @swift-ci please test Build failed Swift Test OS X Platform Git Sha - 1efdb2954778d9baac05201dfc3b857e77612219 Build failed Swift Test Linux Platform Git Sha - 1efdb2954778d9baac05201dfc3b857e77612219 @swift-ci please test Build failed Swift Test OS X Platform Git Sha - 0347eff30b84fbd1d34f30c48c4c2e12cb82a132 Build failed Swift Test Linux Platform Git Sha - 0347eff30b84fbd1d34f30c48c4c2e12cb82a132 @gottesmm - I suppose that this should wait until you're ready to deal with the internal merge? Thanks for waiting. I'll take care of it.
gharchive/pull-request
2018-11-05T17:30:03
2025-04-01T04:33:30.772444
{ "authors": [ "compnerd", "gottesmm", "swift-ci" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/20325", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
378978885
[ABI] Retrieve all associated conformances via a runtime function. Introduce a new runtime entry point, swift_getAssociatedConformanceWitness(), which extracts an associated conformance witness from a witness table. Teach IRGen to use this entry point rather than loading the witness from the witness table and calling it directly. There’s no advantage to doing this now, but it is staging for changing the representation of associated conformances in witness tables. @swift-ci please smoke test @swift-ci please benchmark Build comment file: Performance: -O TEST OLD NEW DELTA RATIO Regression Dictionary2 958 1123 +17.2% 0.85x IterateData 1524 1675 +9.9% 0.91x StringEqualPointerComparison 600 657 +9.5% 0.91x Performance: -Osize TEST OLD NEW DELTA RATIO Regression StringEqualPointerComparison 571 628 +10.0% 0.91x How to read the data The tables contain differences in performance which are larger than 8% and differences in code size which are larger than 1%. If you see any unexpected regressions, you should consider fixing the regressions before you merge the PR. Noise: Sometimes the performance results (not code size!) contain false alarms. Unexpected regressions which are marked with '(?)' are probably noise. If you see regressions which you cannot explain you can try to run the benchmarks again. If regressions still show up, please consult with the performance team (@eeckstein). Hardware Overview Model Name: Mac Pro Model Identifier: MacPro6,1 Processor Name: 12-Core Intel Xeon E5 Processor Speed: 2.7 GHz Number of Processors: 1 Total Number of Cores: 12 L2 Cache (per Core): 256 KB L3 Cache: 30 MB Memory: 64 GB --------------
gharchive/pull-request
2018-11-09T00:49:12
2025-04-01T04:33:30.781608
{ "authors": [ "DougGregor", "swift-ci" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/20449", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
390493764
Exclude/fix certain tests for swift-evolve This pull request adds a swift_evolve feature to our lit configuration and marks several tests as unsupported with it. swift-evolve is a tool currently in development which automatically modifies source code in ways that should be source- and ABI-compatible; we will be using it to test resilience. Currently it simply shuffles declarations. The tests I'm marking as unsupported in this pull request either intentionally depend on the order of declarations, or expose bugs in our tools that we've filed but haven't fixed. This PR also fixes one more order-dependent test in SourceKit. Part of rdar://problem/44426013. Tagging @nkcsgexi because he wrote a few of these tests and @jrose-apple because he suggested using a lit feature. @swift-ci please smoke test @swift-ci please smoke test
gharchive/pull-request
2018-12-13T02:39:38
2025-04-01T04:33:30.784065
{ "authors": [ "brentdax" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/21276", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
451500190
IRGen: getSingletonAggregateFieldType must not return field if its ac… …cess level does not match rdar://50554717 @swift-ci Please test Build failed Swift Test Linux Platform Git Sha - 12b27a1332b852480a529bf6bb5b00d3353df504 Build failed Swift Test OS X Platform Git Sha - 12b27a1332b852480a529bf6bb5b00d3353df504 @swift-ci Please test Build failed Swift Test Linux Platform Git Sha - 12b27a1332b852480a529bf6bb5b00d3353df504 Build failed Swift Test OS X Platform Git Sha - 12b27a1332b852480a529bf6bb5b00d3353df504 Nice catch!
gharchive/pull-request
2019-06-03T13:59:23
2025-04-01T04:33:30.788663
{ "authors": [ "aschwaighofer", "slavapestov", "swift-ci" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/25216", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
120602039
Fix lexer to properly handle single quote strings In response to a FIXME comment, the lexer now suggests replacing double quotes inside single quote strings with a backslash followed by a single quote. Please add tests for the fixits. The file test/type/array.swift at the end shows how to test fixits: typealias FixIt1 = Int[][] // expected-error{{array types are now written with the brackets around the element type}}{{20-20=[}}{{25-26=}} Updated the test in expressions.swift to expect escaped double quotes! Just tested compilation, tests are passing. @gribozavr Can this be merged? @Jumhyn Would you mind squashing everything into one commit and doing a push -f to the pull request? There's one case where I think this will fail: it will suggest converting 'a\"b' to "a\\"b". You need to check to see whether the quote mark you're escaping is already escaped. It's subtle because you still want to convert 'a\\"b' to "a\\\"b". I think it will be sufficient to step through the string, skipping all escaped sequences and replacing unescaped " with \". You're right. This approach also makes it easier to address un-escaping escaped single quotes. Will push a new patch shortly. This should work now -- added new tests for the cases noted above as well. Done. Are the style rules available on Swift.org (or otherwise)? I haven't seen an explicit style guide mentioned anywhere. All of the C++ code in the swift compiler should follow the LLVM coding standards: http://llvm.org/docs/CodingStandards.html Sweet, thanks for putting this together! No problem, and thanks @lattner for the link to the coding standards! Those should be mentioned somewhere in reference to Swift in the contributing guide on Swift.org. Currently, the only line mentioning these is in the LLVM and Swift section, which says, "Contributions to Swift’s LLVM or Clang clones are governed by the LLVM Developer Policy and should follow the appropriate licensing and coding standards," but makes no mention of contributions to the Swift code itself. Probably worth a reference earlier on the page under the Contributing Code heading.
gharchive/pull-request
2015-12-06T02:06:17
2025-04-01T04:33:30.795625
{ "authors": [ "Jumhyn", "cwillmor", "gribozavr", "lattner" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/258", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
159753206
Don't treat Swift methods named "init" as ObjC ARC init methods. Under ARC, methods in the "init" family are considered to have NS_REPLACES_RECEIVER semantics ("consumes" self and returning a value at +1). This is correct for Objective-C "init methods", which are equivalent for Swift's initializers, but almost never correct for any other methods that happen to start with the word "init". Note that Swift still follows all the other ARC conventions, so if you name a method or property, say, "newItemController", the value will be returned at +1. For methods this is probably desirable, but for properties maybe not. We could do something similar for property accessors to make sure they always have the default "no method family" semantics in Objective-C. rdar://problem/25759260 Before merging this pull request to apple/swift repository: [ ] Test pull request on Swift continuous integration. Triggering Swift CI The swift-ci is triggered by writing a comment on this PR addressed to the GitHub user @swift-ci. Different tests will run depending on the specific comment that you use. The currently available comments are: Smoke Testing Platform | Comment ------------ | ------------- All supported platforms | @swift-ci Please smoke test All supported platforms | @swift-ci Please smoke test and merge OS X platform | @swift-ci Please smoke test OS X platform Linux platform | @swift-ci Please smoke test Linux platform Validation Testing Platform | Comment ------------ | ------------- All supported platforms | @swift-ci Please test All supported platforms | @swift-ci Please test and merge OS X platform | @swift-ci Please test OS X platform OS X platform | @swift-ci Please benchmark Linux platform | @swift-ci Please test Linux platform Lint Testing Language | Comment ------------ | ------------- Python | @swift-ci Please Python lint Note: Only members of the Apple organization can trigger swift-ci. @rjmccall, @jckarter, @slavapestov, can one of you review the SILGen part? @swift-ci Please test Seems reasonable. Is there a good way to test the round-trip case, where we try to call the ObjC method exported through the bridging header, to ensure we see the attribute and get the right calling convention on the caller side? Good idea. (Also, clearly I missed something; the OS X failure is related.) @swift-ci Please test Note: not quite ready to merge yet because I haven't written the test Joe suggested. There we go. @swift-ci Please test Linux failure is unrelated, merging.
gharchive/pull-request
2016-06-11T03:27:12
2025-04-01T04:33:30.802110
{ "authors": [ "jckarter", "jrose-apple" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/2989", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
706241382
[linux] remove absolute rpath of /usr/lib/swift/linux added to many shared libraries This was presumably added as a backup, in case the libraries in a toolchain couldn't be found, but will not work well, so take it out. Here's a full list of shared library runpaths from the just-released official Swift 5.3 toolchain for linux: > find swift-5.3-RELEASE-ubuntu20.04/ -name "lib*.so"| xargs readelf -d | ag "File:|runpath" File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/libsourcekitdInProc.so 0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN/../lib/swift/linux:/usr/lib/swift/linux] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/pm/4_2/libPackageDescription.so 0x000000000000001d (RUNPATH) Library runpath: [/home/build-user/swift-nightly-install/usr/lib/swift/linux] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/pm/4/libPackageDescription.so 0x000000000000001d (RUNPATH) Library runpath: [/home/build-user/swift-nightly-install/usr/lib/swift/linux] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/pm/llbuild/libllbuildSwift.so 0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN:$ORIGIN/../../linux] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/pm/llbuild/libllbuild.so File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libicuucswift.so File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libswiftRemoteMirror.so 0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN:/usr/lib/swift/linux] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libdispatch.so 0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libBlocksRuntime.so File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libFoundationNetworking.so 0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libicudataswift.so File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libXCTest.so 0x000000000000001d (RUNPATH) Library runpath: [/home/build-user/build/buildbot_linux/swift-linux-x86_64/lib/swift/linux] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libswiftDispatch.so 0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libswift_Differentiation.so 0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN:/usr/lib/swift/linux] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/lib_InternalSwiftSyntaxParser.so 0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN:/usr/lib/swift/linux] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libswiftGlibc.so 0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN:/usr/lib/swift/linux] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libFoundationXML.so 0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libFoundation.so 0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libswiftSwiftOnoneSupport.so 0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN:/usr/lib/swift/linux] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libicui18nswift.so File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libswiftCore.so 0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN:/usr/lib/swift/linux] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/liblldb.so 0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN/../lib:/home/build-user/build/buildbot_linux/llvm-linux-x86_64/./lib] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/libIndexStore.so 0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN/../lib] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/clang/10.0.0/lib/linux/libclang_rt.hwasan-x86_64.so File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/clang/10.0.0/lib/linux/libclang_rt.ubsan_standalone-x86_64.so File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/clang/10.0.0/lib/linux/libclang_rt.scudo_minimal-x86_64.so File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/clang/10.0.0/lib/linux/libclang_rt.scudo-x86_64.so File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/clang/10.0.0/lib/linux/libclang_rt.ubsan_minimal-x86_64.so 0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN/../lib:/home/build-user/build/buildbot_linux/llvm-linux-x86_64/./lib] File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/clang/10.0.0/lib/linux/libclang_rt.dyndd-x86_64.so File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/clang/10.0.0/lib/linux/libclang_rt.asan-x86_64.so File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/libswiftDemangle.so 0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN:/usr/lib/swift/linux] If there's any difference between the toolchain being used and the one installed in /usr/lib/swift/linux, say if you download this 5.3 toolchain for linux into your home directory but 5.2.5 is installed in /usr/lib/swift/linux, using that system runpath won't work properly. I checked and all these libraries already have a runpath relative to $ORIGIN, ie locally, so all this absolute path will do is supply the system Swift library if the local one is missing (obviously if the used toolchain is installed in the system, ie /usr/lib/swift/linux, this absolute path is redundant). But since using the system Swift libraries will not work properly if the Swift versions don't match, causing subtle bugs in the worst case, I think it's better not to add this absolute runpath at all. It was added five years ago by @bitjammer, ad95b5f, and copied to SourceKit a couple years later, a93bddf. I brought up this issue earlier this year, but got no response at the time, so I let it go. I think this is worth getting in and into the 5.3 branch for the next patch release. @gottesmm and @compnerd, I think this should be removed, similar to the other ELF runpath issues we discussed. @tachoknight, let us know what you think from your Fedora packaging perspective. As we discussed offline, can you create a PR against swift-integration-test that validates that the snapshots maintain this property? Then I think you can do cross-repo testing to test it with this PR. Then first merge this and then the integration test one. Will do. @drexin, let me know what you think, I'd like to get this into the 5.3.1 patch release. @swift-ci smoke test Thanks, sometimes adding a comment seems to get the CI to run? I think the CI needs a swift kick. 😉 @swift-ci smoke test Passes CI, ready to merge. apple/swift-integration-tests#72 @swift-ci test I validated in the CI job that swift-ci found the appropriate PR on swift-integration-tests. Confirmed that the CI checked out the new integration test and that those tests were run, but the log doesn't list them individually so my new test isn't listed. The Windows CI failure appears unrelated.
gharchive/pull-request
2020-09-22T10:01:04
2025-04-01T04:33:30.809488
{ "authors": [ "buttaface", "drexin", "gottesmm" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/34023", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
797203978
[Build] Fix macOS version passed in -target parameter. This code was added in commit b4368259480ed2c895e8c855bbfde20c1805acd1, but the macOS variable expansion had an extra "SWIFT_" at the front. Among other potential problems, this causes classes in the stdlib to always be built with the legacy is-swift bit, which causes havoc when libswiftCore is in the shared cache where those bits aren't fixed up. rdar://73767816 Karoy, am I right in understanding that the SWIFT_-prefixed versions are for the stdlib, runtime and overlays, whereas the non-SWIFT_ prefixed variables are for the host tools? In that case, I suspect what's going wrong is that the SWIFT_-prefixed variable is not being set correctly for macOS. Karoy, am I right in understanding that the SWIFT_-prefixed versions are for the stdlib, runtime and overlays, whereas the non-SWIFT_ prefixed variables are for the host tools? In that case, I suspect what's going wrong is that the SWIFT_-prefixed variable is not being set correctly for macOS. After further discussion with Varun, we think this may be fine, and I just ran into trouble due to some bad state in my build directory. I had to do a clean build to test this change.... I'm going to do another clean build without this supposed fix to double-check. After further discussion with Varun, we think this may be fine, and I just ran into trouble due to some bad state in my build directory. I had to do a clean build to test this change.... I'm going to do another clean build without this supposed fix to double-check. This was indeed a false alarm. Sorry if I got anybody worried! This was indeed a false alarm. Sorry if I got anybody worried!
gharchive/pull-request
2021-01-29T21:01:41
2025-04-01T04:33:30.813944
{ "authors": [ "mikeash", "varungandhi-apple" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/35657", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
815691134
[SIL Optimizer] Replace VarDeclUsageChecker AST Walker with DiagnoseVarUsage SIL Pass SIL Diagnostics for Var Usage This mandatory SIL optimization pass, DiagnoseVarUsage, replaces VarDeclUsageChecker as the system for generating usage diagnostics. Why a SIL pass? With access to the def-use graph, we have a more accurate picture of the lifetime of a variable declaration for more robust diagnostics. Ultimately, the former AST walker missed many cases, shown in the bugs below. Additionally, this implementation has many possibilities for future features. Addresses Bugs: Mutating Inouts func increment() { var x = 0 // no warning, but should warn x was mutated but never read. x += 1 } This bug (and similar bugs) are addressed by improving the way usage of inout parameters is diagnosed. If the function is local to the SIL module, we can follow the SILArgument to look for a read of the parameter. For external SILFunctions, an inout parameter is marked as used if the function produces some used output, or if the function has a closure parameter (where the inout argument may be used to produce other parameters within the closure). Closures Missing Diagnostics SR-11992, SR-13821 class Foo { lazy var x: Int = { var y = 42 // no warning, but should warn unused var z = 42 // no warning, but should warn never mutated return z }() } var missesAllDiagsInClosures: () -> Int = { var unused = 42 // no warning, but should warn unused return 12 } Bugs like this are addressed because we are no longer using the AST walker, which inconsistently missed certain nodes. Instead, we can see every SILInstruction that originates from a VarDecl and then decide whether to inspect it for diagnostics. Misleading Unused Setter Param Diagnostics var x : Int { get { 0 } set { } } x // warning: setter argument 'newValue' was never used, but the property was accessed // note: did you mean to use 'newValue' instead of accessing the property's current value? These diagnostics are misleading because newValue is not in scope for the diagnostic location, and the setter function was never called. Now, a case like this does not produce diagnostics. Cases like the following still produce diagnostics as expected. var x : Int { get { 0 } set { print(x) // warning: setter argument 'newValue' was never used, but the property was accessed // note: did you mean to use 'newValue' instead of accessing the property's current value? } } For some more information on this decision, this gist details reasoning and a possible alternative diagnostic. And many other similar cases are addressed... After switching to the SIL implementation, we now get consistent diagnostics in many of these common cases. Modifications and Limitations Test files This diagnostic occurs during the mandatory diagnostics SIL optimizations, and therefore the diagnostics are only emitted after errors are addressed in the file (because code with errors does not reach SILGen). DISCLAIMER TO REVIEWERS Nearly every test case regarding variable usage is touched by this PR. I apologize for the nightmare this review may entail. If I removed expected warnings, it is because the file never reaches SIL optimizations. Rest assured, every case I edited is either now checked in SILOptimizer/diagnose_var_usage or moved from the original <testname>.swift to <testname>_usage.swift. '_' Assignments for Constants Constants have no getter function, so _ = x no longer counts as a use of the variable. This source code is invisible after SILGen. let x = 1 _ = x /* SIL is identical to: */ let x = 1 This is because the underscore assignments have no effect, and the value is never accessed. Therefore, diagnosing a constant is unused (even after an underscore assignment) is more correct because the value is never actually used. Future work may be done to change SILGen to make underscore assignments visible to this diagnostic pass and removed in a SIL Optimization pass. Underscore assignments are still useful for variables because the right-hand side may produce side effects from creating this value. For variables, because an actual function is called, there is still no warning for underscore assignments. var y = 1 _ = y // this is still a read TL;DR: we produce an 'unused' diagnostic if the rhs of an underscore assignment has no effects (accessing a constant value). #if declarations Because inactive clauses are not generated in SIL, we can not track usage here. var x = 12 // we end up diagnosing this is never mutated #if false x = 1 #endif print(x) Future work may be done in SILGen to generate useful information about variable usage in inactive clauses if this is important. Or maybe some information can be inspected in the AST. Otherwise, this current diagnostic might encourage Swift users to write code with better usage enforcement across all builds. Tracking Value Usage In this SIL pass, we track the use of a value, not the declaration. This means if we have multiple constants referencing the same value, we have no way of knowing which reference to the value we used: func duplicateReference() { let x = 12 let y = x print(x) // we don't know if x or y was used here, because they are identical in SIL } To a small extent, we can check up through the AST to find if a user of the value corresponds to a certain reference, and we do this to at least know x was assigned to y. To handle diagnostics for this situation, instead of generating possibly misleading usage diagnostics, we introduce a new diagnostic: func duplicateReference() { let x = 12 // note: initially referenced here let y = x // warning: immutable value referenced by 'y' is already referenced by 'x'; consider removing it and replacing all uses of 'y' with 'x'}} } Future work may be done here to improve this diagnostic to include fix-its for replacement. Or, if this diagnostic is to be reverted, we would need to produce separate values for every variable declaration. What currently looks like this: debug_value %3 : $Int, let, name "x" // id: %4 debug_value %3 : $Int, let, name "y" // id: %5 Would need to do something like this: debug_value %3 : $Int, let, name "x" ... copy_value %3 to %4 debug_value %4 : $Int, let, name "y" However, in the handful of times this pattern appears in the stlib, the suggestion of the diagnostic made more sense, was quick to address, and lead to more straight-forward code. Tracking Value Usage - Exceptions Capture Lists - redeclare an identical value, so we don't want to produce incorrect diagnostics here. We perform an explicit check for declarations belonging to capture lists to make sure this declaration of the value is captured. Case Statements - binding declarations in switch statements may redeclare an identical value. Again, we explicitly check that a 'use' of the value is in SILBlocks accessible to the case statement. Future Extensions While the focus of this PR is to lay the groundwork to replace the previous AST Walker implementation, I noticed this SIL pass opens up plenty of further diagnostics. Unused Parameter diags See SR-2819 This pass already performs a check for the newValue parameter of setter functions, and the logic could easily be transferred to diagnosing other unused parameters or unmutated inout parameters. If the parameter is not required (such as when overriding a parent function), then these diagnostics could be helpful to simplify code. Unused Local Functions Using def-use traversals in a similar manner for functions inaccessible to outside modules can produce useful diagnostics. Links SR-11992 and SR-13821 rdar://59636530 Superceeds PR #34617 Thanks to @CodaFi and @augusto2112 for the brainstorming sessions and motivation! This is really cool! func increment() { var x = 0 // no warning, but should warn x was mutated but never read. x += 1 } Well, only if it’s a built-in += do we know that. If I create a custom implementation that logs the previous value before incrementing by one, then the value is both read and mutated. That is, admittedly, far-fetched for +=, but not for arbitrary inout functions. @xwu that's a great point. For custom inouts local to the SILModule, this code actually follows the argument to look for a read in the function body. But your comment does make me think about custom functions that are declared externally, and you're right that we can't verify it was read. I'd like to point out that if this inout function does not produce any used result, I'm not sure if the mutating function ever provided value to the use of the variable, and the warning might still make sense. But maybe this discussion is out of scope for this PR. Perhaps, for now, I can change this to assume external inout functions read the parameter. How do you think we should approach this? Constants have no getter function, so _ = x no longer counts as a use of the variable. This source code is invisible after SILGen. let x = 1 _ = x /* SIL is identical to: */ let x = 1 This is because the underscore assignments have no effect, and the value is never accessed. While understandable, this does have some impact on the user experience, as one of the most obvious ways currently to silence the warning is to write _ = x, and it would be nice to preserve that for both mutable and immutable bindings. Perhaps, for now, I can change this to assume external inout functions read the parameter. How do you think we should approach this? Neat that you’re actually looking for a read in the function body where possible. I think it’s reasonable to scope the diagnostic so that it’s emitted only where the function body is available (either because it’s inlinable from the standard library or otherwise) or in the same module, and so you can be sure if there’s no read. That would cover a lot of scenarios, I think, without giving false positives. @xwu okay, thank you. I will make and test the appropriate changes for this. I think it may be interesting to expand upon this in the future. Down the line, it would be great to provide these diagnostics for purely mutating functions, like the built-in += operator. Down the line, it would be great to provide these diagnostics for purely mutating functions, like the built-in += operator. Well, the built-in operator implementations are inlinable if not transparent, so you should be able to look through those and make the diagnosis, no? @xwu thanks for the conversation! I made updates to inspect usage through transparent functions, and I think this is a much better approach to these diagnostics! Summary of changes local or transparent functions are inspected for a read of inout parameters other external functions will assume a read of inout parameters local or transparent functions Is it too expensive to look through any functions that have their function body serialized? I believe you can just use isSerialized() for that. Okay, I am running tests for that right now. This is my first time working with SIL, so I never knew that was possible. Thanks for the suggestion, and I'll tag you with an update after I test it! @xwu, this works! Thank you. Sorry I just noticed your comment about underscore assignments, @xwu. I realize this is not great behavior for these diagnostics. This change won't affect function results on the rhs, it only affects if an immutable value is directly used. let x = 0 _ = x // this is a problem _ = x + 1 // ok _ = someFunctionResult() // ok I will look into modifying SILGen to produce a debug_value instruction for underscore assignments to mark this as a read, but I have no idea what level of effort this entails. Will look into this tomorrow. Thanks again! I will look into modifying SILGen to produce a debug_value instruction for underscore assignments to mark this as a read, but I have no idea what level of effort this entails. Will look into this tomorrow. Admittedly, I don't think it's a huge deal since this is limited to constants; if it's that narrowly tailored, I don't know that many people are deliberately creating a binding and then not using it, but one thing I've learned is that if it's possible, someone is relying on it. @xwu, I made a fix to look for a DiscardAssignmentExpr, restoring the expected functionality of underscore assignments. We do collect enough information at this stage to know the discarded assignment produces no effects, so it may be interesting to add a diagnostic here in the future. @swift-ci Please test @swift-ci Please test source compatibility @maustinstar Looking into callee function bodies sounds tricky because those functions might in turn call other functions, which might be recursive, etc. There is existing infrastructure in the SIL optimizer that could handle this if you decide to go down this route, but I wonder if its worth the added complexity. Also, it would be nice if we could avoid triggering deserialization of functions, since it can add a lot of unnecessary overhead at -Onone. @slavapestov might in turn call other functions, which might be recursive If your concern is that we might find ourselves in an infinite loop of searching through recursive functions, this won't happen because instructions that have already been seen are not searched again. Also, this def-use traversal will keep stepping into functions until a read is found. I don't know if performance is a huge concern here because this will only happen in the case that a variable gets used in multiple inout functions without ever reading the value. There is existing infrastructure in the SIL optimizer that could handle this if you decide to go down this route. Great! Can you give me some pointers for where to look and learn? it would be nice if we could avoid triggering deserialization of functions Can you explain a little further? In regards to inspecting callees, we only search serialized and locally defined SILFunctions. Thank you for your feedback! Build failed Swift Test OS X Platform Git Sha - bc4db0c92fc2309b76a19d3dac6cfd3164543ae2 To handle diagnostics for this situation, instead of generating possibly misleading usage diagnostics, we introduce a new diagnostic: func duplicateReference() { let x = 12 // note: initially referenced here let y = x // warning: immutable value referenced by 'y' is already referenced by 'x'; consider removing it and replacing all uses of 'y' with 'x'}} } I like that you detect this; it seems very much in keeping with detecting unused bindings. One caveat and one nit: There is a reasonable use case for this, so there should be some way to silence the warning at the point of use. Consider, for instance: let physicalConstantA = /* some large number */ let physicalConstantB = physicalConstantA Suppose physicalConstantA serves a different purpose and is used in different kinds of calculation than physicalConstantB; however, it is fixed to be the same value. With the warning above, I'm encouraged either to (A) refer to this value all the time as physicalConstantA, even where at the point of use this could be confusing because readers are accustomed to reasoning about a calculation in terms of physicalConstantB; or to (B) rewrite the large number twice, risking a second typo or the two values falling out of sync, which they should not. In other words, the use case here would be analogous to type aliases, only in this case for values. I should be allowed to alias immutable values, and so in the case where this is not being done accidentally, there should be a way to shut off this warning. Usually, parens are offered as a way to silence warnings like this. If this produces a concern about how downstream diagnostics, we could work out some way of referring to such "value aliases" in diagnostics (for example, an imaginary error that the value is too amazing to be used might read: 'x' (aka 'y') is too amazing to be used). The wording of the diagnostic might be somewhat confusing. The user-facing language distinguishes "reference types" from "value types," and semantically, let y = x creates a second copy of the same value. Now, of course, users familiar with compiler internals know that this "copy" is entirely notional and always elided when you've got immutable bindings of a constant value. However, calling these bindings "references" and telling the user that they've "referenced" the same value twice would be confusing to a user who's working with the mental model that Swift explicitly teaches them to use. Because inactive clauses are not generated in SIL, we can not track usage here. var x = 12 // we end up diagnosing this is never mutated #if false x = 1 #endif print(x) Future work may be done in SILGen to generate useful information about variable usage in inactive clauses if this is important. Or maybe some information can be inspected in the AST. Otherwise, this current diagnostic might encourage Swift users to write code with better usage enforcement across all builds. Given the nature of the diagnostic in question, we should prefer false negatives (not showing the diagnostic when a variable isn't mutated) over false positives (showing the diagnostic incorrectly when a variable really is mutated). Therefore--and I apologize if I misunderstand your text and this is already the behavior you've implemented--in the case where there's code that's conditionally not compiled, I think the diagnostic should just not be shown. This doesn't preclude the future work you mention where variable usage could be determined within these conditional compilation blocks, but until that time, I would fail open. @xwu I really like that idea for using parenthesis to explicitly note a copy of the reference. I will look into this. For #if statements, you are correct in assuming that the diagnostic is shown if a variable is used within the conditional compilation blocks. let a = 0 // diagnoses never used #if false print(a) #endif The code for the inactive clause never goes through SILGen, so it's invisible to the SIL pass for diagnostics. So, I'm not sure how to find that a value is used in a case like this. let a = 0 // ok #if true print(a) #endif This code does pass through SILGen, so we can see this use in the diagnostic pass. The only way I can imagine doing this is by changing SILGen (perhaps create some debug instruction to inform us that a value is captured in an inactive clause). Unfortunately, I'm not familiar with the SILGen process, but I will try my best to figure this out. Any ideas here? The only way I can imagine doing this is by changing SILGen (perhaps create some debug instruction to inform us that a value is captured in an inactive clause). Unfortunately, I'm not familiar with the SILGen process, but I will try my best to figure this out. Any ideas here? This does sound like it'd be a bit of a tough thing to tackle. To be clear, I'm not suggesting that it needs to be done as part of this PR even though it'd be a very nice improvement later. I'm only suggesting that, for now, we don't diagnose unused variables in cases where there are conditional clauses (that is, let the presence of a conditional clause cause all in-scope variables to be considered "used" for now). Leaving an update so this thread doesn't go stale: The reason _ = foo produces no code is because SILGen peepholes it away. Perhaps we should just not do that, and allow dead code elimination to remove the discarded result after the diagnostics pass? I started working on a change to SILGen to produce debug values for discarded assignment expressions. Otherwise, the use of the value is invisible by the time we get to this diagnostic pass. I am considering introducing this as an incremental change in a separate PR, since it seems other aspects of this project need further consideration and may go on for a long while. In the meantime, I will convert this PR to a draft. Thank you for the continuous feedback and support. Update: was able to remove SILGen peephole for underscore assignments. This restores functionality for underscore assignments. Tagging @varungandhi-apple, thanks for the help. @xwu, the regressions we discussed earlier in this thread have now been addressed. I would appreciate if you could run the test suite on CI. @swift-ci test [Okay if other people who have studied this PR in more detail would like to overrule this but] Given the large scale of this change, I think there is a greater risk of things going wrong and that this may need to be reverted. Should that happen, the size of this PR creates a greater chance of merge conflicts, making reverting and relanding this work tricky. If possible, I recommend breaking it up into at least 3 PRs, spaced out by 1-2 weeks: PR 1. Add the new checker under a new frontend flag. Make sure any tests pass with both by default and with the new flag. (Alternately, this flag takes a boolean which is false by default). This is also a good opportunity to squash some of the WIP commits. PR 2. Flip the default. PR 3. Remove the old code and the flag. Build failed Swift Test Linux Platform Git Sha - ba95da4d2738f5b4ca8c0dfedaca47d1e5d3405a Build failed Swift Test OS X Platform Git Sha - ba95da4d2738f5b4ca8c0dfedaca47d1e5d3405a Okay, thanks @varungandhi-apple, I'll split this up and add the flag
gharchive/pull-request
2021-02-24T17:26:42
2025-04-01T04:33:30.855889
{ "authors": [ "maustinstar", "slavapestov", "swift-ci", "varungandhi-apple", "xwu" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/36134", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
948390520
[Distributed] fix all distributed runtime tests; they are currently not running on CI These have an issue on some specific CI configuration so they are currently disabled on CI. They pass again all locally though: Passed Tests (7): Swift(macosx-x86_64) :: Distributed/Runtime/distributed_actor_remote_functions.swift Swift(macosx-x86_64) :: Distributed/Runtime/distributed_actor_deinit.swift Swift(macosx-x86_64) :: Distributed/Runtime/distributed_actor_isRemote.swift Swift(macosx-x86_64) :: Distributed/Runtime/distributed_actor_dynamic_remote_func.swift Swift(macosx-x86_64) :: Distributed/Runtime/distributed_no_transport_boom.swift Swift(macosx-x86_64) :: Distributed/Runtime/distributed_actor_init_local.swift Swift(macosx-x86_64) :: Distributed/Runtime/distributed_actor_local.swift Tests blocked by rdar://78290608 @swift-ci please smoke test and merge @swift-ci please smoke test and merge @swift-ci please build toolchain Linux Toolchain (Ubuntu 16.04) Download Toolchain Git Sha - 2fd09b32f2343064e4e9bfffb848ee8b1c4002a6 Install command tar zxf swift-PR-38502-630-ubuntu16.04.tar.gz More info @swift-ci please smoke test macOS macOS Toolchain Download Toolchain Git Sha - 2fd09b32f2343064e4e9bfffb848ee8b1c4002a6 Install command tar -zxf swift-PR-38502-1059-osx.tar.gz --directory ~/ @swift-ci please smoke test Linux and merge @swift-ci please smoke test linux platform
gharchive/pull-request
2021-07-20T07:50:11
2025-04-01T04:33:30.863105
{ "authors": [ "ktoso", "swift-ci" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/38502", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
192685364
Removing more calls to ValueDecl::getType() Almost ready to push it down to VarDecl. @swift-ci Please smoke test @swift-ci Please smoke test @swift-ci Please smoke test @swift-ci Please smoke test @swift-ci Please smoke test @swift-ci Please smoke test @swift-ci Please smoke test @swift-ci Please smoke test This broke two iOS tests. Reverting in #6019. Never mind, Slava pushed followups before my revert. We'll just fix it elsewhere.
gharchive/pull-request
2016-11-30T21:31:58
2025-04-01T04:33:30.866276
{ "authors": [ "gparker42", "slavapestov" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/5990", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
195670998
Revert "[docs] DebuggingTheCompiler.rst => DebuggingTheCompiler.md." This reverts commit 444920f2b67dea7350a6098b75078edf1be8b64f. Caused the build failure: toctree contains reference to nonexisting document u'DebuggingTheCompiler' @swift-ci Please smoke test and merge Just an FYI, I spoke with Erik about this. The reason that this is being reverted is that there is a reference to this file in the main toc file. It errors since it is not there. That being said, the main toc file is in *.rst and we are converting away from that. Perhaps the right solution here is to convert all of the files in the main toc file to markdown and then change the toc to markdown as well.
gharchive/pull-request
2016-12-14T23:03:14
2025-04-01T04:33:30.868370
{ "authors": [ "eeckstein", "gottesmm" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/6281", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1642843392
DNM: Add -Rmodule-interface-rebuild to startOnMainActor Add remark to see why we're rebuilding from the swiftinterface occasionally, instead of using the perfectly good swiftmodule for the stdlib. This is starting to affect the Windows CI now. Hopefully I get lucky and can trigger the failure first try. @swift-ci please test We can merge if I don't catch it the first time so that we get some parallel data collection action going. @swift-ci please smoke test Windows failure: FAIL: Swift(windows-x86_64) :: Concurrency/Runtime/startOnMainActor.swift (547 of 9197) ******************** TEST 'Swift(windows-x86_64) :: Concurrency/Runtime/startOnMainActor.swift' FAILED ******************** Script: -- : 'RUN: at line 1'; rm -rf "T:\swift\test-windows-x86_64\Concurrency\Runtime\Output\startOnMainActor.swift.tmp" && mkdir -p "T:\swift\test-windows-x86_64\Concurrency\Runtime\Output\startOnMainActor.swift.tmp" : 'RUN: at line 2'; 't:\\swift\\bin\\swiftc.exe' -target x86_64-unknown-windows-msvc -vfsoverlay T:/swift\stdlib\windows-vfs-overlay.yaml -swift-version 4 -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 9999:macOS 9999, iOS 9999, watchOS 9999, tvOS 9999' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.0:macOS 10.14.4, iOS 12.2, watchOS 5.2, tvOS 12.2' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.1:macOS 10.15, iOS 13.0, watchOS 6.0, tvOS 13.0' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.2:macOS 10.15.4, iOS 13.4, watchOS 6.2, tvOS 13.4' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.3:macOS 11.0, iOS 14.0, watchOS 7.0, tvOS 14.0' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.4:macOS 11.3, iOS 14.5, watchOS 7.4, tvOS 14.5' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.5:macOS 12.0, iOS 15.0, watchOS 8.0, tvOS 15.0' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.6:macOS 12.3, iOS 15.4, watchOS 8.5, tvOS 15.4' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.7:macOS 13.0, iOS 16.0, watchOS 9.0, tvOS 16.0' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.8:macOS 13.3, iOS 16.4, watchOS 9.4, tvOS 16.4' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.9:macOS 9999, iOS 9999, watchOS 9999, tvOS 9999' -libc MD -Xfrontend -Rmodule-interface-rebuild -Xfrontend -disable-availability-checking C:\Users\swift-ci\jenkins\workspace\swift-PR-windows\swift\test\Concurrency\Runtime\startOnMainActor.swift -o T:\swift\test-windows-x86_64\Concurrency\Runtime\Output\startOnMainActor.swift.tmp/a.out 1>&2 : 'RUN: at line 3'; echo T:\swift\test-windows-x86_64\Concurrency\Runtime\Output\startOnMainActor.swift.tmp/a.out : 'RUN: at line 4'; T:\swift\test-windows-x86_64\Concurrency\Runtime\Output\startOnMainActor.swift.tmp/a.out -- Exit Code: 127 Command Output (stdout): -- $ ":" "RUN: at line 1" $ "rm" "-rf" "T:\swift\test-windows-x86_64\Concurrency\Runtime\Output\startOnMainActor.swift.tmp" $ "mkdir" "-p" "T:\swift\test-windows-x86_64\Concurrency\Runtime\Output\startOnMainActor.swift.tmp" $ ":" "RUN: at line 2" $ "t:\\swift\\bin\\swiftc.exe" "-target" "x86_64-unknown-windows-msvc" "-vfsoverlay" "T:/swift\stdlib\windows-vfs-overlay.yaml" "-swift-version" "4" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 9999:macOS 9999, iOS 9999, watchOS 9999, tvOS 9999" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.0:macOS 10.14.4, iOS 12.2, watchOS 5.2, tvOS 12.2" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.1:macOS 10.15, iOS 13.0, watchOS 6.0, tvOS 13.0" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.2:macOS 10.15.4, iOS 13.4, watchOS 6.2, tvOS 13.4" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.3:macOS 11.0, iOS 14.0, watchOS 7.0, tvOS 14.0" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.4:macOS 11.3, iOS 14.5, watchOS 7.4, tvOS 14.5" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.5:macOS 12.0, iOS 15.0, watchOS 8.0, tvOS 15.0" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.6:macOS 12.3, iOS 15.4, watchOS 8.5, tvOS 15.4" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.7:macOS 13.0, iOS 16.0, watchOS 9.0, tvOS 16.0" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.8:macOS 13.3, iOS 16.4, watchOS 9.4, tvOS 16.4" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.9:macOS 9999, iOS 9999, watchOS 9999, tvOS 9999" "-libc" "MD" "-Xfrontend" "-Rmodule-interface-rebuild" "-Xfrontend" "-disable-availability-checking" "C:\Users\swift-ci\jenkins\workspace\swift-PR-windows\swift\test\Concurrency\Runtime\startOnMainActor.swift" "-o" "T:\swift\test-windows-x86_64\Concurrency\Runtime\Output\startOnMainActor.swift.tmp/a.out" # command stderr: Unsupported redirect: (('>&', 1), '2') error: command failed with exit status: 127 --
gharchive/pull-request
2023-03-27T21:32:12
2025-04-01T04:33:30.872080
{ "authors": [ "etcwilde" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/64653", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1692823975
[cxx-interop] Update a test on Linux libstdc++ currently cannot be split into submodules due to the way some types are defined, e.g. there are multiple headers that define size_t, so those headers must be a single (sub-)module. @swift-ci please smoke test
gharchive/pull-request
2023-05-02T17:37:44
2025-04-01T04:33:30.873340
{ "authors": [ "egorzhdan" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/65588", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1860559856
[5.9] [Macros] Use outermost source file for establishing (file)private access scope Fixes rdar://114048069. Cherry-picked from https://github.com/apple/swift/pull/68057 @swift-ci please test
gharchive/pull-request
2023-08-22T05:04:19
2025-04-01T04:33:30.874680
{ "authors": [ "DougGregor" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/68058", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1903803379
[5.10][build] Move libdispatch earlier in the build so it can be used by the compiler validation suite Cherrypick of #65829 and #68565 Explanation: The libdispatch and several concurrency tests in the validation suite were inadvertently disabled a couple years ago. This enables them again by making sure the target libdispatch is built first and using the new dispatch-vfs-overlay.yaml it now generates. Scope: Enabling compiler tests again Issue: #53973 Risk: negative, as it is only testing more code, thus lowering risk Testing: Passes all CI on trunk for the last couple days, and I've been doing this manually on Android for years without a problem. Reviewer: @gottesmm @bnbarham, please run the CI on this. @swift-ci please test @gottesmm, should I try to enable these for the 5.9 branch also? @gottesmm, should I try to enable these for the 5.9 branch also? There should be very few changes going to 5.9 at this point, so probably not worth it in my mind. There should be very few changes going to 5.9 at this point, so probably not worth it in my mind. @bnbarham, this changes zero behavior in the toolchain though, only slightly modifying the build order and some tests to enable several tests on the CI again. That's why I thought it might be worth it for 5.9 too, as it might catch 5.9 regressions better by having several dozen more tests run. Pinging @tbkka, ready for review. Ping @DougGregor, tiny pull that re-enables several tests on the linux CI. Ping @bnbarham, can you review and merge? I didn't know you could get pulls into 5.10.
gharchive/pull-request
2023-09-19T22:19:10
2025-04-01T04:33:30.880301
{ "authors": [ "bnbarham", "finagolfin" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/68640", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
201903791
Remove dead witness tables, including dead witness functions. A witness table is dead if it is not used outside the module (private/internal) and it’s not used by any instruction or other witness table in the module. Also the meta-type of the conforming type must not escape, because it’s possible to test any opaque type if it conforms to a protocol. This gives about 2.5% code size reduction on stdlib. Also some benchmarks see code size reductions, but this depends on the benchmark. For most benchmarks (especially our micro-benchmarks, where we don't define any types) there is no improvement. For some larger benchmarks there are significant size reductions. On average it's 2.4%. @swift-ci Please test Build failed Jenkins build - Swift Test Linux Platform Git Commit - a7782567e8a28829dc4b6689d0912809a457d857 Test requested by - @eeckstein Build failed Jenkins build - Swift Test OS X Platform Git Commit - a7782567e8a28829dc4b6689d0912809a457d857 Test requested by - @eeckstein @swift-ci Please test Build failed Jenkins build - Swift Test Linux Platform Git Commit - a7782567e8a28829dc4b6689d0912809a457d857 Test requested by - @eeckstein Build failed Jenkins build - Swift Test OS X Platform Git Commit - a7782567e8a28829dc4b6689d0912809a457d857 Test requested by - @eeckstein @swift-ci Please test linux Build failed Jenkins build - Swift Test Linux Platform Git Commit - 44f35b916089a46042368b535ece4805f063ce64 Test requested by - @eeckstein "Also the meta-type of the conforming type must not escape" What if an instance of the type escapes? Depends. An instance could escape into a concrete context that can handle the value without metadata. Anywhere we need to escape an instance into a context that also needs the type metadata to work with the instance, that should be tracked independently. It's fuzzy but I can imagine this: internal class Impl: SomeProto { /* … */ } public func produce() -> AnyObject { return Impl() } public func consume(_ value: AnyObject) { guard let interestingValue = value as? SomeProto else { return } // … } I'm pretty sure in this case the witness table is never referenced statically and the metatype does not escape (because we're using AnyObject rather than Any), but we still need the witness table at run-time dynamically. What if an instance of the type escapes? @jrose-apple The question is if someone outside the translation unit can dynamically check if a value conforms to a protocol. This is done by looking up the metatype of that value in the protocol_conformances section. So if the metatype does not escape we are good. @swift-ci Please test linux Why does it matter whether the lookup happens outside the translation unit? I'm pretty sure in this case the witness table is never referenced statically and the metatype does not escape (because we're using AnyObject rather than Any), but we still need the witness table at run-time dynamically. The metatype does escape because it's stored in the Impl() instance header. Why does it matter whether the lookup happens outside the translation unit? It doesn't. My mistake Ah, right. I didn't think about the metatype being either passed as a constructor argument or directly stored in the type.
gharchive/pull-request
2017-01-19T16:12:33
2025-04-01T04:33:30.890821
{ "authors": [ "eeckstein", "jckarter", "jrose-apple", "swift-ci" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/6912", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1961967921
[swift-vfe][swift-wme] Stop disabling function mergers for Swift VFE / WME The Swift and LLVM function mergers were disabled when Swift VFE or WME are enabled because the function merger did not respect metadata on calls to llvm.type.checked.load. This is no longer the case, so we can turn these passes back on. https://github.com/apple/llvm-project/pull/7679 @swift-ci Please test Linux failures look unrelated to me. https://github.com/apple/llvm-project/pull/7679 @swift-ci test linux @eeckstein Do you mind merging / testing the LLVM side of things? https://github.com/apple/llvm-project/pull/7679. I can merge Swift. I have merged the LLVM PR @NuriAmari feel free to merge
gharchive/pull-request
2023-10-25T18:11:25
2025-04-01T04:33:30.894340
{ "authors": [ "NuriAmari", "aschwaighofer", "eeckstein" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/69407", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
212182421
Revert "[build-script-impl] Add support for building benchmarks on linux but disable it in all currently defined presets for linux." Reverts apple/swift#7860 This broke the OSS - LLDB incremental linux bots. @swift-ci please smoke test Thanks! On Mar 6, 2017, at 9:39 AM, Ben Langmuir notifications@github.com wrote: Merged #7931 https://github.com/apple/swift/pull/7931. — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/apple/swift/pull/7931#event-988152162, or mute the thread https://github.com/notifications/unsubscribe-auth/AAee35h43uQLRgkYGnYc6h8N_V9HrZl-ks5rjETagaJpZM4MUZy7.
gharchive/pull-request
2017-03-06T16:50:49
2025-04-01T04:33:30.897434
{ "authors": [ "benlangmuir", "gottesmm" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/7931", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
213498021
[Constraint solver] More updates for using the constraint solver type… … map. Cache types in a few more places and ensure that we're reading from the map in another spot (but interestingly stop reading from it in a different spot - this is going to need another round of clean-ups at some point on some of these Expr APIs that like to read types from other nodes). @swift-ci Please smoke test and merge
gharchive/pull-request
2017-03-11T02:23:46
2025-04-01T04:33:30.899344
{ "authors": [ "rudkx" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/8029", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
221115050
[GSB] Delay requirements on recursive types Delaying requirements that form new, recursively-defined types, providing an alternative to the current detect-recursion-complain-and-bail approach that bans recursive protocol requirements entirely. This is along the path toward recursive protocol conformances (which would require us to reprocess those requirements when we need to later), and eliminates some extraneous bookkeeping in the GenericSignatureBuilder. Go, Doug! @swift-ci please smoke test and merge @swift-ci please smoke test and merge @swift-ci please smoke test Linux LLDB failure. Ugh. @swift-ci please smoke test Linux patched up the LLDB issue... @swift-ci please smoke test Linux
gharchive/pull-request
2017-04-12T00:07:41
2025-04-01T04:33:30.902104
{ "authors": [ "DougGregor", "dabrahams" ], "repo": "apple/swift", "url": "https://github.com/apple/swift/pull/8714", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
313649126
Plot objects display twice in Jupyter Notebook The plot object (when yielded as the output value of a cell) shows twice. Repro steps (in Jupyter Notebook): import turicreate as tc sf = tc.SFrame.read_csv('https://docs-assets.developer.apple.com/turicreate/datasets/tc-clang-format-results.csv') tc.visualization.scatter(sf['Lines Added'], sf['Lines Removed']) Maybe this has been fixed, but it doesn't seem to happen for me, though I am using JupyterLab, not the older Notebook interface. Cannot repro.
gharchive/issue
2018-04-12T09:36:44
2025-04-01T04:33:30.903667
{ "authors": [ "abhishekpratapa", "theavey", "znation" ], "repo": "apple/turicreate", "url": "https://github.com/apple/turicreate/issues/520", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
708538149
New pip dependency resolver Add --use-feature=2020-resolver to our Python build/test setup to ensure we stay compatible with the new dependency resolver. Also in our Python build/test setup, pin TensorFlow version (for a few Python versions and OS combinations). Without this one of our packages used for unit testing causes a SciPy version conflict. Internal tests pass
gharchive/pull-request
2020-09-24T23:39:58
2025-04-01T04:33:30.905169
{ "authors": [ "TobyRoseman" ], "repo": "apple/turicreate", "url": "https://github.com/apple/turicreate/pull/3328", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
604774598
deprecate Event.last and Span.last Event.last => ao.lastEvent Span.last => ao.lastSpan the base of this branch is p9p-noti+minimal; to minimize conflicts this should be reviewed and merged only after PR 117 has been merged. there will be 20 files changed after that. other than the deprecation code itself, all the changes are to replace last = Event.last and {last} = Event with last = ao.lastEvent, and analogous changes for Span.last good catch on the api - it generate it automatically from the sources but it does nedd to be updated. i'll add that to a subsequent PR.
gharchive/pull-request
2020-04-22T13:50:43
2025-04-01T04:33:30.913680
{ "authors": [ "bmacnaughton" ], "repo": "appoptics/appoptics-apm-node", "url": "https://github.com/appoptics/appoptics-apm-node/pull/118", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
475783724
Transaction logs may be reporting illegal actions When running a batch training learning curves sometimes show more opportunities than are possible in the problem set. This seems to be from times that agents gets a step A correct, then does some step B, then attempts to go back to attempt step A, which is marked as incorrect. In a human using CTAT this would never happen because if you try to take a previously correct steps the field will be locked and so it will never appear. Unsure how we want to handle this long term, because its kind of interesting, but in the short term it violates the assumptions of transaction roll ups for learning curve analysis. It is possible this is desirable behavior and the rollup process we were using from PyAFM is actually wrong. We are pretty sure this is an issue related to a particular tutor and not the framework itself. Closing for now unless it comes up again.
gharchive/issue
2019-08-01T16:42:35
2025-04-01T04:33:30.961012
{ "authors": [ "eharpste" ], "repo": "apprenticelearner/AL_HTML", "url": "https://github.com/apprenticelearner/AL_HTML/issues/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1238148854
[Bug]: Very long values in APPSMITH_ENCRYPTION_PASSWORD and APPSMITH_ENCRYPTION_SALT makes it skip these variables Is there an existing issue for this? [X] I have searched the existing issues Description A user flagged this issue. It seems like that very long values for APPSMITH_ENCRYPTION_PASSWORD and APPSMITH_ENCRYPTION_SALT break it somehow to skip email and signup variables. Now I use shorter ones and everything is working again. Steps To Reproduce Needs to be investigated and reproduced https://discord.com/channels/725602949748752515/975730370689499136/975740299005558784 Public Sample App No response Version v1.6.24 @sub1983 @sum35h could one of you please have a look?
gharchive/issue
2022-05-17T06:42:09
2025-04-01T04:33:30.972336
{ "authors": [ "Nikhil-Nandagopal", "danciaclara" ], "repo": "appsmithorg/appsmith", "url": "https://github.com/appsmithorg/appsmith/issues/13889", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1270335943
[Task]: Redesign action selector Is there an existing issue for this? [X] I have searched the existing issues SubTasks [ ] PRD [ ] Solution exploration [ ] Refinement [ ] Feedback [ ] Handoff Estimated due date: June 24th, 2022
gharchive/issue
2022-06-14T06:49:28
2025-04-01T04:33:30.974387
{ "authors": [ "ayushi-appsmith" ], "repo": "appsmithorg/appsmith", "url": "https://github.com/appsmithorg/appsmith/issues/14515", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1747205325
[Bug]: Appsmith server is taking too long to respond Is there an existing issue for this? [X] I have searched the existing issues Description Installed on Ubuntu 20.04 sudo docker ps CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 76284bf62841 appsmith/appsmith-ce:latest "/opt/appsmith/entry…" 23 hours ago Up 23 hours (healthy) 0.0.0.0:80->80/tcp, :::80->80/tcp, 0.0.0.0:443->443/tcp, :::443->443/tcp appsmith d7a2e3340eed containrrr/watchtower "/watchtower --sched…" 41 hours ago Up 40 hours 8080/tcp appsmith_auto_update_1 When I try to edit my first App Logs: editor stdout | 10.96.26.140 - - [08/Jun/2023:06:51:48 +0000] "GET /api/v1/pages?pageId=647f441bb9009d5e1ccb6841&mode=EDIT HTTP/1.1" 200 497 "http://10.96.32.161/app/my-first-application/page1-647f441bb9009d5e1ccb6841/edit" "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36 OPR/99.0.0.0" editor stdout | rts stdout | ns:/# socket _uR-aKx-o11UwoKEAAAB joined the room app:647f441bb9009d5e1ccb683e rts stdout | redis stdout | 1676:M 08 Jun 2023 06:51:48.280 * Background saving terminated with success redis stdout | backend stdout | [2023-06-08 06:51:48,320] userEmail=paolo@ftgm.it, sessionId=b80a75a0-0c6e-4a3e-9eed-889424b3bdea, thread=lettuce-epollEventLoop-6-1, requestId=2350f68b-1c57-4481-8a1c-44e5910607ea - Going to get all unpublished action collections with params: {applicationId=[647f441bb9009d5e1ccb683e]}, branch: null backend stdout | editor stdout | 10.96.26.140 - - [08/Jun/2023:06:51:48 +0000] "GET /api/v1/collections/actions?applicationId=647f441bb9009d5e1ccb683e HTTP/1.1" 200 101 "http://10.96.32.161/app/my-first-application/page1-647f441bb9009d5e1ccb6841/edit" "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36 OPR/99.0.0.0" editor stdout | editor stdout | 10.96.26.140 - - [08/Jun/2023:06:51:48 +0000] "GET /api/v1/themes/applications/647f441bb9009d5e1ccb683e/current?mode=EDIT HTTP/1.1" 200 1442 "http://10.96.32.161/app/my-first-application/page1-647f441bb9009d5e1ccb6841/edit" "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36 OPR/99.0.0.0" editor stdout | editor stdout | 10.96.26.140 - - [08/Jun/2023:06:51:48 +0000] "GET /api/v1/pages/647f441bb9009d5e1ccb6841 HTTP/1.1" 200 518 "http://10.96.32.161/app/my-first-application/page1-647f441bb9009d5e1ccb6841/edit" "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36 OPR/99.0.0.0" editor stdout | rts stdout | ns:/# socket _uR-aKx-o11UwoKEAAAB left the room app:647f441bb9009d5e1ccb683e rts stdout | editor stdout | 10.96.26.140 - - [08/Jun/2023:06:51:48 +0000] "GET /api/v1/themes/applications/647f441bb9009d5e1ccb683e HTTP/1.1" 200 4624 "http://10.96.32.161/app/my-first-application/page1-647f441bb9009d5e1ccb6841/edit" "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36 OPR/99.0.0.0" editor stdout | rts stdout | ns:/# socket _uR-aKx-o11UwoKEAAAB joined the room app:647f441bb9009d5e1ccb683e rts stdout | backend stdout | [2023-06-08 06:51:48,397] userEmail=paolo@ftgm.it, sessionId=b80a75a0-0c6e-4a3e-9eed-889424b3bdea, thread=lettuce-epollEventLoop-6-1, requestId=4910e92d-c69a-4077-bf8b-da19a27cc71d - Going to get snapshot with application id: 647f441bb9009d5e1ccb683e, branch: null backend stdout | backend stdout | [2023-06-08 06:51:48,404] userEmail=paolo@ftgm.it, sessionId=b80a75a0-0c6e-4a3e-9eed-889424b3bdea, thread=nioEventLoopGroup-3-5, requestId=4910e92d-c69a-4077-bf8b-da19a27cc71d - backend stdout | com.appsmith.server.exceptions.AppsmithException: Unable to find application 647f441bb9009d5e1ccb683e backend stdout | at com.appsmith.server.services.ce.ApplicationSnapshotServiceCEImpl.getWithoutDataByApplicationId(ApplicationSnapshotServiceCEImpl.java:73) backend stdout | Suppressed: reactor.core.publisher.FluxOnAssembly$OnAssemblyException: backend stdout | Error has been observed at the following site(s): backend stdout | *__checkpoint ⇢ Handler com.appsmith.server.controllers.ApplicationController#getSnapshotWithoutApplicationJson(String, String) [DispatcherHandler] backend stdout | Original Stack Trace: backend stdout | at com.appsmith.server.services.ce.ApplicationSnapshotServiceCEImpl.getWithoutDataByApplicationId(ApplicationSnapshotServiceCEImpl.java:73) backend stdout | at com.appsmith.server.controllers.ce.ApplicationControllerCE.getSnapshotWithoutApplicationJson(ApplicationControllerCE.java:216) backend stdout | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) backend stdout | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) backend stdout | at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) backend stdout | at java.base/java.lang.reflect.Method.invoke(Method.java:568) backend stdout | at org.springframework.web.reactive.result.method.InvocableHandlerMethod.lambda$invoke$0(InvocableHandlerMethod.java:145) backend stdout | at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:132) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onNext(MDCConfig.java:53) backend stdout | at reactor.core.publisher.MonoZip$ZipCoordinator.signal(MonoZip.java:293) backend stdout | at reactor.core.publisher.MonoZip$ZipInner.onNext(MonoZip.java:474) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onNext(MDCConfig.java:53) backend stdout | at reactor.core.publisher.MonoPeekTerminal$MonoTerminalPeekSubscriber.onNext(MonoPeekTerminal.java:180) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onNext(MDCConfig.java:53) backend stdout | at reactor.core.publisher.Operators$BaseFluxToMonoOperator.completePossiblyEmpty(Operators.java:2071) backend stdout | at reactor.core.publisher.FluxDefaultIfEmpty$DefaultIfEmptySubscriber.onComplete(FluxDefaultIfEmpty.java:134) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onComplete(MDCConfig.java:63) backend stdout | at reactor.core.publisher.FluxSwitchIfEmpty$SwitchIfEmptySubscriber.onComplete(FluxSwitchIfEmpty.java:85) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onComplete(MDCConfig.java:63) backend stdout | at reactor.core.publisher.MonoSupplier$MonoSupplierSubscription.request(MonoSupplier.java:148) backend stdout | at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.set(Operators.java:2341) backend stdout | at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.onSubscribe(Operators.java:2215) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onSubscribe(MDCConfig.java:47) backend stdout | at reactor.core.publisher.MonoSupplier.subscribe(MonoSupplier.java:48) backend stdout | at reactor.core.publisher.Mono.subscribe(Mono.java:4485) backend stdout | at reactor.core.publisher.FluxSwitchIfEmpty$SwitchIfEmptySubscriber.onComplete(FluxSwitchIfEmpty.java:82) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onComplete(MDCConfig.java:63) backend stdout | at reactor.core.publisher.Operators.complete(Operators.java:137) backend stdout | at reactor.core.publisher.FluxFlatMap.trySubscribeScalarMap(FluxFlatMap.java:145) backend stdout | at reactor.core.publisher.MonoFlatMap.subscribeOrReturn(MonoFlatMap.java:53) backend stdout | at reactor.core.publisher.InternalMonoOperator.subscribe(InternalMonoOperator.java:57) backend stdout | at reactor.core.publisher.MonoZip$ZipCoordinator.request(MonoZip.java:216) backend stdout | at reactor.core.publisher.MonoFlatMap$FlatMapMain.request(MonoFlatMap.java:194) backend stdout | at reactor.core.publisher.MonoIgnoreThen$ThenIgnoreMain.onSubscribe(MonoIgnoreThen.java:134) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onSubscribe(MDCConfig.java:47) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onSubscribe(MDCConfig.java:47) backend stdout | at reactor.core.publisher.MonoFlatMap$FlatMapMain.onSubscribe(MonoFlatMap.java:117) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onSubscribe(MDCConfig.java:47) backend stdout | at reactor.core.publisher.MonoZip.subscribe(MonoZip.java:125) backend stdout | at reactor.core.publisher.InternalMonoOperator.subscribe(InternalMonoOperator.java:64) backend stdout | at reactor.core.publisher.MonoDefer.subscribe(MonoDefer.java:52) backend stdout | at reactor.core.publisher.InternalMonoOperator.subscribe(InternalMonoOperator.java:64) backend stdout | at reactor.core.publisher.MonoIgnoreThen$ThenIgnoreMain.subscribeNext(MonoIgnoreThen.java:240) backend stdout | at reactor.core.publisher.MonoIgnoreThen$ThenIgnoreMain.onComplete(MonoIgnoreThen.java:203) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onComplete(MDCConfig.java:63) backend stdout | at reactor.core.publisher.MonoFlatMap$FlatMapMain.onComplete(MonoFlatMap.java:189) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onComplete(MDCConfig.java:63) backend stdout | at reactor.core.publisher.Operators.complete(Operators.java:137) backend stdout | at reactor.core.publisher.MonoZip.subscribe(MonoZip.java:121) backend stdout | at reactor.core.publisher.Mono.subscribe(Mono.java:4485) backend stdout | at reactor.core.publisher.MonoIgnoreThen$ThenIgnoreMain.subscribeNext(MonoIgnoreThen.java:263) backend stdout | at reactor.core.publisher.MonoIgnoreThen.subscribe(MonoIgnoreThen.java:51) backend stdout | at reactor.core.publisher.InternalMonoOperator.subscribe(InternalMonoOperator.java:64) backend stdout | at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:165) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onNext(MDCConfig.java:53) backend stdout | at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:79) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onNext(MDCConfig.java:53) Steps To Reproduce login go to first app Public Sample App No response Environment Deploy Preview Issue video log No response Version Self hosted with docker and autoupdate @marcheschi When do encounter this error? Is it during installation or when you try to open some app? Also, please use this document to provide detailed container logs - https://docs.appsmith.com/learning-and-resources/how-to-guides/how-to-get-container-logs#docusaurus_skipToContent_fallback @marcheschi, sorry you're facing trouble here. Can you confirm the system capacity of the server you're running with? Like vCPU count, available memory etc., and if there's any other applications also running on this server, along with Appsmith please? Also, please share full logs as indicated in prev. message. Thanks! Thank you @pranavkanade @sharat87 I used a dedicated VM with 8 CPU and 16 GB of ram with 20 GB of disk appsmith is the only app on the server. The installation runs fine, and I can login but when I try to edit my first app appears the message and the editor is not shown. Logs: https://drive.google.com/file/d/1rAleXljvo0edsNP8c8GPWarirWCt35ev/view?usp=sharing @marcheschi, thank you for the details. It looks like your server isn't connected to the Internet. Appsmith is unable to reach cs.appsmith.com, which is required for it's operation. More details about this in the callout at https://docs.appsmith.com/getting-started/setup#self-hosting-benefits. Can you please confirm if your server has access to reach external endpoints in the Internet? If not, can we allow that access and restart Appsmith please? The server has internet access, via proxy . It does not work either. I tried to recreate the container twice. Too difficult and is a nonsense it needs a connection to the internet. I'm sorry, Thank you for your assistance. Hey @marcheschi, sorry you feel that way. We need Internet access to make some features possible, and some others convenient. I hope you managed to resolve the issue, if not, we can help you out if you can share the logs and what's failing now. Thank you! Thanks @marcheschi for confirming. We actually have a provision to allow you to configure proxy in upper or lower case. We accept both. I'm wondering if uppercase env variables wasn't working for you? 🤔 Relevant code is in this function: https://github.com/appsmithorg/appsmith/blob/c4b9e272dbc0b08f307931c71c24e897435975a1/deploy/docker/entrypoint.sh#L97-L113
gharchive/issue
2023-06-08T06:55:52
2025-04-01T04:33:30.989645
{ "authors": [ "marcheschi", "pranavkanade", "sharat87" ], "repo": "appsmithorg/appsmith", "url": "https://github.com/appsmithorg/appsmith/issues/24231", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1906454455
[Bug]: Unable to import attached json file in the airgapped image Is there an existing issue for this? [X] I have searched the existing issues Description When I try to import the attached application, I get an error to the effect unable to connect to server. Steps To Reproduce Take the pull of the airgapped image, and set it up without internet Try to import the attached file [rename it to json] Note that import does not work and error pops up. OldApp_DSTesting1.9.24.txt Note: The attached app has Airtable / GSheets and other SAAS plugins. Since Airgapped does not support these plugins, if thats the reason for this failure, then an appropriate message needs to be given - or there should be some way of handling that portion in the import modal. Public Sample App No response Environment Release Issue video log No response Version Self Hosted It's because of the unsupported DS In the current airgapped regression also we are seeing the above issue, and also, sometimes we see a broken UI, which is not such a great user experience. Though it is a case with unsupported datasources, it should be handled gracefully. https://github.com/appsmithorg/appsmith/assets/101863839/036d9f61-a8c9-486d-8031-2c03d2341095 @btsgh - GSheets is not supported and should not exist in Airgapped apps. Any failues because of that can be ignored.
gharchive/issue
2023-09-21T08:52:43
2025-04-01T04:33:30.995660
{ "authors": [ "btsgh", "infinitetrooper" ], "repo": "appsmithorg/appsmith", "url": "https://github.com/appsmithorg/appsmith/issues/27539", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2403654993
[Bug]: Cross referencing - environment change does not auto refresh the response for awslambda ds. Is there an existing issue for this? [X] I have searched the existing issues Description In the cross referencing DP https://ee-4056.dp.appsmith.com/, for awslambda datasource, changing the environment does not refresh the response section automatically Steps To Reproduce https://jam.dev/c/45ace210-8d31-4225-9401-8c1c17f68b3b Create an awslambda datasource in a workspace with custom environments Create new query and run it against prod environment Now change the environment to say staging or any custom env. The response tab does not get refreshed until we click Run manually Now change it back to production, and here also note that the response tab does not get refreshed and show data until we click run manually Public Sample App No response Environment Deploy Preview Severity Medium (Frustrating UX) Issue video log No response Version Cloud This is the behavior even on release if no widget is bound. Looks like this is expected behavior.
gharchive/issue
2024-07-11T17:03:29
2025-04-01T04:33:31.000022
{ "authors": [ "btsgh" ], "repo": "appsmithorg/appsmith", "url": "https://github.com/appsmithorg/appsmith/issues/34888", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2434928432
[Bug]: MySql datasource connection error - Authentication failed Is there an existing issue for this? [X] I have searched the existing issues Description I deploy appsmith in my local env and have running a long time stably. Recently, I create a new connection to mysql and display the error: Error Message: Authentication failed Report Class: o.m.r.message.flow.AuthenticationFlow Stack: j.l.IndexOutOfBoundsException: readerIndex(63) + length(1) exceeds writerIndex(63): PooledSlicedByteBuf(ridx: 63, widx: 63, cap: 63/63, unwrapped: PooledUnsafeDirectByteBuf(ridx: 66, widx: 66, cap: 2048)) at i.n.b.AbstractByteBuf.checkReadableBytes0(AbstractByteBuf.java:1442) at i.n.b.AbstractByteBuf.checkReadableBytes(AbstractByteBuf.java:1428) at i.n.b.AbstractByteBuf.skipBytes(AbstractByteBuf.java:971) at o.m.r.m.s.InitialHandshakePacket.decode(InitialHandshakePacket.java:100) at o.m.r.c.DecoderState$1.decode(DecoderState.java:25) at o.m.r.c.MariadbFrameDecoder.decode(MariadbFrameDecoder.java:104) at o.m.r.c.MariadbFrameDecoder.decode(MariadbFrameDecoder.java:93) at i.n.h.c.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:529) at i.n.h.c.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:468) ... 15 common frames omitted Wrapped by: i.n.h.c.DecoderException: java.lang.IndexOutOfBoundsException: readerIndex(63) + length(1) exceeds writerIndex(63): PooledSlicedByteBuf(ridx: 63, widx: 63, cap: 63/63, unwrapped: PooledUnsafeDirectByteBuf(ridx: 66, widx: 66, cap: 2048)) at i.n.h.c.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:499) at i.n.h.c.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:290) at i.n.c.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444) ... 13 common frames omitted Wrapped by: i.r.s.R2dbcNonTransientResourceException: Connection error at o.m.r.c.SimpleClient.handleConnectionError(SimpleClient.java:187) at o.m.r.c.SimpleClient.sendResumeError(SimpleClient.java:196) at o.m.r.c.SimpleClient.receiveResumeError(SimpleClient.java:203) at r.c.p.FluxOnErrorResume$ResumeSubscriber.onError(FluxOnErrorResume.java:94) at c.t.w.p.c.l.MdcContextLifter.onError(MdcContextLifter.java:35) at r.c.p.FluxMap$MapSubscriber.onError(FluxMap.java:134) at r.n.c.FluxReceive.terminateReceiver(FluxReceive.java:478) at r.n.c.FluxReceive.drainReceiver(FluxReceive.java:273) at r.n.c.FluxReceive.onInboundError(FluxReceive.java:466) at r.n.c.ChannelOperations.onInboundError(ChannelOperations.java:495) at r.n.c.ChannelOperationsHandler.exceptionCaught(ChannelOperationsHandler.java:144) at i.n.c.AbstractChannelHandlerContext.invokeExceptionCaught(AbstractChannelHandlerContext.java:346) at i.n.c.AbstractChannelHandlerContext.invokeExceptionCaught(AbstractChannelHandlerContext.java:325) at i.n.c.AbstractChannelHandlerContext.fireExceptionCaught(AbstractChannelHandlerContext.java:317) at i.n.c.ChannelInboundHandlerAdapter.exceptionCaught(ChannelInboundHandlerAdapter.java:143) at i.n.c.AbstractChannelHandlerContext.invokeExceptionCaught(AbstractChannelHandlerContext.java:346) at i.n.c.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:447) at i.n.c.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420) at i.n.c.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412) at i.n.c.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) at i.n.c.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:440) at i.n.c.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420) at i.n.c.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) at i.n.c.e.AbstractEpollStreamChannel$EpollStreamUnsafe.epollInReady(AbstractEpollStreamChannel.java:800) at i.n.c.e.EpollEventLoop.processReady(EpollEventLoop.java:499) at i.n.c.e.EpollEventLoop.run(EpollEventLoop.java:397) at i.n.u.c.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at i.n.u.i.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at i.n.u.c.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.lang.Thread.run(Thread.java:840) It sames like a netty error when connect by r2dbc Steps To Reproduce Create mysql connection by Datasources Public Sample App No response Environment Production Severity High (Blocker to building or releasing) Issue video log No response Version v1.9.21 @alpacaca can you please upgrade to the latest to see if that help? Also, a screen recording would help in understanding exact steps to reproduce the issue. Is it happening during connection test or saving or creation etc. @alpacaca can you please upgrade to the latest to see if that help? Also, a screen recording would help in understanding exact steps to reproduce the issue. Is it happening during connection test or saving or creation etc. @pranavkanade sorry, I have repackaged the front-end page and just use appsmith plugin backend, so screen recording cannot help. Actually, it is happening during connection test. I will try the latest version instead and reply soon. I try on app.appsmith.com/app in my app, it also doesn't work ... here is the screenshot and request curl copy: curl 'https://app.appsmith.com/api/v1/datasources/test' \ -H 'Accept: application/json, text/plain, */*' \ -H 'Accept-Language: zh-CN,zh;q=0.9,en-US;q=0.8,en;q=0.7' \ -H 'Connection: keep-alive' \ -H 'Content-Type: application/json' \ -H 'Cookie: ajs_anonymous_id=598de59d-162e-41a3-a710-6804b59db718; _hjSessionUser_2952959=eyJpZCI6Ijc5MTRhZTMyLWJkOGMtNTY1ZS05MjI0LWJhNTA0NTU1NGM3MiIsImNyZWF0ZWQiOjE3MjIyNDU2NjUwMTAsImV4aXN0aW5nIjp0cnVlfQ==; _gcl_au=1.1.1654232351.1722245666; _ga=GA1.1.356015106.1722245666; SESSION=7bf1d513-42fa-4b41-91d2-8f6c61793265; intercom-device-id-y10e7138=4982bf1f-7c74-417a-90eb-1e6279eec1dc; ajs_user_id=craig.sese@gmail.com; _hjSession_2952959=eyJpZCI6IjMxYWJjN2I5LTczY2EtNGRhZC05OTY5LTkyYWM1OGUwOWEwZCIsImMiOjE3MjIyNTM4NDg0MzQsInMiOjEsInIiOjEsInNiIjowLCJzciI6MCwic2UiOjAsImZzIjowLCJzcCI6MH0=; _ga_D1VS24CQXE=GS1.1.1722253846.2.0.1722253849.0.0.0; SL_C_23361dd035530_SID={"c370af0df0edf38360adbefbdc47d2b42ea137c9":{"sessionId":"meFwPGahFI_J9BcwRaEar","visitorId":"8UHRmHLrX0FD6AlJD6pG3"}}; intercom-session-y10e7138=UkZ4SHUrOWRUSUtTZkZHUzJxSzE4Z2NSNzEyc3JHK2tVeU0yTC9PR21qTEx6V1lScy92UlhGZVBrdU51QlQ1by0tSDBsbEROS2t4b2s0MlRsUElSWE9JZz09--53ad53c4e84f13df15c4f90144be0a4f2856508d' \ -H 'Origin: https://app.appsmith.com' \ -H 'Referer: https://app.appsmith.com/app/my-first-application/page1-64f7e9e149cf293c0e2f9a78/edit/datasource/temp-id-0?from=datasources&pluginId=5f192a44a2d05225d010db9d' \ -H 'Sec-Fetch-Dest: empty' \ -H 'Sec-Fetch-Mode: cors' \ -H 'Sec-Fetch-Site: same-origin' \ -H 'User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/126.0.0.0 Safari/537.36' \ -H 'X-Requested-By: Appsmith' \ -H 'newrelic: eyJ2IjpbMCwxXSwiZCI6eyJ0eSI6IkJyb3dzZXIiLCJhYyI6IjQyMDk1NDgiLCJhcCI6IjExMzQzODgzMDkiLCJpZCI6IjhlOWVmYzliNzhiOWZmM2EiLCJ0ciI6IjcxNDJiNGI0NzQ0MDU5OWE2MDAwMjM2M2UwNWU1MjdmIiwidGkiOjE3MjIyNTM5NDc1MzN9fQ==' \ -H 'sec-ch-ua: "Not/A)Brand";v="8", "Chromium";v="126", "Google Chrome";v="126"' \ -H 'sec-ch-ua-mobile: ?0' \ -H 'sec-ch-ua-platform: "macOS"' \ -H 'sentry-trace: 970bad5e3cf349b4864562131ed65ade-8826e65a5515406e-0' \ -H 'traceparent: 00-7142b4b47440599a60002363e05e527f-8e9efc9b78b9ff3a-01' \ -H 'tracestate: 4209548@nr=0-1-4209548-1134388309-8e9efc9b78b9ff3a----1722253947533' \ -H 'x-anonymous-user-id: 598de59d-162e-41a3-a710-6804b59db718' \ --data-raw '{"datasourceId":"","environmentId":"64f7e9df49cf293c0e2f9a72","datasourceConfiguration":{"properties":[null,{"key":"Connection method","value":"STANDARD"}],"connection":{"mode":"READ_WRITE","ssl":{"authType":"DEFAULT"}},"endpoints":[{"port":"3307","host":"146.56.222.244"}],"authentication":{"databaseName":"information_schema","username":"test","password":"123456"}},"pluginId":"5f192a44a2d05225d010db9d","workspaceId":"64f7e9df49cf293c0e2f9a6e"}' @Nikhil-Nandagopal can reopen this issue? Sample app: https://app.appsmith.com/app/my-first-application/page1-64f7e9e149cf293c0e2f9a78/edit/datasource/temp-id-0?from=datasources&pluginId=5f192a44a2d05225d010db9d
gharchive/issue
2024-07-29T09:33:59
2025-04-01T04:33:31.014357
{ "authors": [ "Nikhil-Nandagopal", "alpacaca", "pranavkanade" ], "repo": "appsmithorg/appsmith", "url": "https://github.com/appsmithorg/appsmith/issues/35253", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2492049115
[Bug]: Dropdown option invisible but visible in preview Is there an existing issue for this? [X] I have searched the existing issues Description https://github.com/user-attachments/assets/38afc570-f7b9-416d-a89a-086d57281785 Steps To Reproduce Label key Js is not wokring which was working earlier Public Sample App No response Environment Production Severity Critical (Broken Production apps) Issue video log No response Version 1.38.0 @gokaco please can I see what your source data looks like so I can debug further, thanks. @gokaco please can I see what your source data looks like so I can debug further, thanks. Something like [ { "liveClassThumbnailURL": "https://res.cloudinary.com", "cover": { "vidURL": "https://res.cloudinary.com" }, "email": "dfsdfsdf@adf.com", "name": "Kartikeya Goswami", "imgURL": "https://res.cloudinary.com", "password": "abcdefg", "roles": [ { "roleId": "uuid" }, { "roleId": "uuid" } ], "subCategory": "COACH", "expertise": [], "trainerId": "uuid", "category": "category", "description": "description" } ] Thanks @gokaco. The label key cannot be the joined string you are using in the video, the label key needs to be one of the keys from your object above, e.g name. Add. name as the label key and see if it works. @gokaco Did above comment help resolve your issue? Thanks @gokaco. The label key cannot be the joined string you are using in the video, the label key needs to be one of the keys from your object above, e.g name. Add. name as the label key and see if it works. @jacquesikot But earlier this was working. And this issue started happening from some days only. How it was working earlier? Have any changes been made in the new release? @gokaco We don't think anything has changed in the recent release. The key field should be one of the keys in your JSON. This should work for your use case, please let us know if there's anything that is not working.
gharchive/issue
2024-08-28T13:14:55
2025-04-01T04:33:31.020763
{ "authors": [ "Nikhil-Nandagopal", "gokaco", "jacquesikot", "rahulbarwal" ], "repo": "appsmithorg/appsmith", "url": "https://github.com/appsmithorg/appsmith/issues/35953", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1388285451
feat: Moved UserProfileImagePicker from ads folder Description This PR have moved UserProfileImagePicker component from ads folder to pages/UserProfile. Fixes #17110 Type of change Bug fix (non-breaking change which fixes an issue) How Has This Been Tested? Tested in local system Checklist: [x] My code follows the style guidelines of this project [x] I have performed a self-review of my own code [ ] I have commented my code, particularly in hard-to-understand areas [ ] I have made corresponding changes to the documentation [x] My changes generate no new warnings [ ] I have added tests that prove my fix is effective or that my feature works [x] New and existing unit tests pass locally with my changes /ok-to-test sha=626569f PR is tested, no issues found. /ok-to-test sha=3237a55 /ok-to-test sha=5ce24ef /ok-to-test sha=831b5d7 /ok-to-test sha=1def66f /ok-to-test sha=86243b5 /ok-to-test sha=0de5497 /ok-to-test sha=8357e2d
gharchive/pull-request
2022-09-27T19:53:43
2025-04-01T04:33:31.026254
{ "authors": [ "albinAppsmith", "shadabbuchh" ], "repo": "appsmithorg/appsmith", "url": "https://github.com/appsmithorg/appsmith/pull/17111", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1582157450
chore: type check callback functions in setInterval and setTimeout before invoking. Description We have recently changed the implementation for setInterval in https://github.com/appsmithorg/appsmith/pull/20168. This has the potential to blow up our sentry quota if someone sets up an empty setInterval function with no callback in it and execute it. This PR adds a type check before invoking the callback to ensure that the passed in callback is indeed, a function. Fixes https://github.com/appsmithorg/appsmith/issues/20591 https://appsmith.sentry.io/issues/3928721303/?project=1546547&query=is%3Aunresolved+worker&referrer=issue-stream&statsPeriod=14d Media A video or a GIF is preferred. when using Loom, don’t embed because it looks like it’s a GIF. instead, just link to the video Type of change Chore (housekeeping or task changes that don't impact user perception) How Has This Been Tested? Manual Test Plan Add Testsmith test cases links that relate to this PR Issues raised during DP testing Link issues raised during DP testing for better visiblity and tracking (copy link from comments dropped on this PR) Checklist: Dev activity [x] My code follows the style guidelines of this project [x] I have performed a self-review of my own code [x] I have commented my code, particularly in hard-to-understand areas [ ] I have made corresponding changes to the documentation [x] My changes generate no new warnings [ ] I have added tests that prove my fix is effective or that my feature works [ ] New and existing unit tests pass locally with my changes [ ] PR is being merged under a feature flag QA activity: [ ] Test plan has been approved by relevant developers [ ] Test plan has been peer reviewed by QA [ ] Cypress test cases have been added and approved by either SDET or manual QA [ ] Organized project review call with relevant stakeholders after Round 1/2 of QA [ ] Added Test Plan Approved label after reveiwing all Cypress test /ok-to-test sha=2386038 /ok-to-test sha=eed1615 /ok-to-test sha=a80807d
gharchive/pull-request
2023-02-13T11:16:53
2025-04-01T04:33:31.033830
{ "authors": [ "arunvjn" ], "repo": "appsmithorg/appsmith", "url": "https://github.com/appsmithorg/appsmith/pull/20592", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1652397012
fix: set initialState from metaProps Description Fix the meta value code to set the initialMetaState correctly. Type of change Bug fix (non-breaking change which fixes an issue) How Has This Been Tested? Manual Jest Cypress Test Plan Issues raised during DP testing Checklist: Dev activity [ ] My code follows the style guidelines of this project [ ] I have performed a self-review of my own code [ ] I have commented my code, particularly in hard-to-understand areas [ ] I have made corresponding changes to the documentation [ ] My changes generate no new warnings [ ] I have added tests that prove my fix is effective or that my feature works [ ] New and existing unit tests pass locally with my changes [ ] PR is being merged under a feature flag QA activity: [ ] Test plan has been approved by relevant developers [ ] Test plan has been peer reviewed by QA [ ] Cypress test cases have been added and approved by either SDET or manual QA [ ] Organized project review call with relevant stakeholders after Round 1/2 of QA [ ] Added Test Plan Approved label after reveiwing all Cypress test /ok-to-test sha=2213c10 /ok-to-test sha=11adecf /ok-to-test sha=186a194 /ok-to-test sha=1ebf63f /ok-to-test sha=1ebf63f /ok-to-test sha=1ff88a1 /ci-merge-check
gharchive/pull-request
2023-04-03T16:29:51
2025-04-01T04:33:31.040277
{ "authors": [ "Aishwarya-U-R", "Rishabh-Rathod", "arunvjn" ], "repo": "appsmithorg/appsmith", "url": "https://github.com/appsmithorg/appsmith/pull/22049", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1904586187
fix: shifting migrations to shared dsl package Description Shifting migration logic to @shared/dsl package PR fixes following issue(s) Fixes #26783 Media A video or a GIF is preferred. when using Loom, don’t embed because it looks like it’s a GIF. instead, just link to the video Type of change Please delete options that are not relevant. Bug fix (non-breaking change which fixes an issue) New feature (non-breaking change which adds functionality) Breaking change (fix or feature that would cause existing functionality to not work as expected) Chore (housekeeping or task changes that don't impact user perception) This change requires a documentation update Testing How Has This Been Tested? Please describe the tests that you ran to verify your changes. Also list any relevant details for your test configuration. Delete anything that is not relevant [ ] Manual [ ] JUnit [ ] Jest [ ] Cypress Test Plan Add Testsmith test cases links that relate to this PR Issues raised during DP testing Link issues raised during DP testing for better visiblity and tracking (copy link from comments dropped on this PR) Checklist: Dev activity [ ] My code follows the style guidelines of this project [ ] I have performed a self-review of my own code [ ] I have commented my code, particularly in hard-to-understand areas [ ] I have made corresponding changes to the documentation [ ] My changes generate no new warnings [ ] I have added tests that prove my fix is effective or that my feature works [ ] New and existing unit tests pass locally with my changes [ ] PR is being merged under a feature flag QA activity: [ ] Speedbreak features have been covered [ ] Test plan covers all impacted features and areas of interest [ ] Test plan has been peer reviewed by project stakeholders and other QA members [ ] Manually tested functionality on DP [ ] We had an implementation alignment call with stakeholders post QA Round 2 [ ] Cypress test cases have been added and approved by SDET/manual QA [ ] Added Test Plan Approved label after Cypress tests were reviewed [ ] Added Test Plan Approved label after JUnit tests were reviewed /ok-to-test sha=1b162e2 /ok-to-test sha=2cdad4c /ok-to-test sha=dfb9efe @sbalaji1192 can you please re-review and add your approval, I have added the changes /build-deploy-preview /ok-to-test sha=69a175b Merged this PR with #28518 /build-deploy-preview
gharchive/pull-request
2023-09-20T09:32:20
2025-04-01T04:33:31.051221
{ "authors": [ "Parthvi12", "brayn003" ], "repo": "appsmithorg/appsmith", "url": "https://github.com/appsmithorg/appsmith/pull/27490", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2028226196
feat: rest, graphql and curl fixed for onboarding Description This PR adds: functionality to REST, GraphQL and Curl import options when using these during onboarding start with data flow. On clicking on any of these options would take us inside application. PR fixes following issue(s) Fixes #29303 Media A video or a GIF is preferred. when using Loom, don’t embed because it looks like it’s a GIF. instead, just link to the video Type of change New feature (non-breaking change which adds functionality) Testing How Has This Been Tested? Please describe the tests that you ran to verify your changes. Also list any relevant details for your test configuration. Delete anything that is not relevant [x] Manual [ ] JUnit [ ] Jest [ ] Cypress Test Plan Add Testsmith test cases links that relate to this PR Issues raised during DP testing Link issues raised during DP testing for better visiblity and tracking (copy link from comments dropped on this PR) Checklist: Dev activity [x] My code follows the style guidelines of this project [x] I have performed a self-review of my own code [x] I have commented my code, particularly in hard-to-understand areas [ ] I have made corresponding changes to the documentation [x] My changes generate no new warnings [ ] I have added tests that prove my fix is effective or that my feature works [x] New and existing unit tests pass locally with my changes [ ] PR is being merged under a feature flag QA activity: [ ] Speedbreak features have been covered [ ] Test plan covers all impacted features and areas of interest [ ] Test plan has been peer reviewed by project stakeholders and other QA members [ ] Manually tested functionality on DP [ ] We had an implementation alignment call with stakeholders post QA Round 2 [ ] Cypress test cases have been added and approved by SDET/manual QA [ ] Added Test Plan Approved label after Cypress tests were reviewed [ ] Added Test Plan Approved label after JUnit tests were reviewed Summary by CodeRabbit Refactor Improved the logic for identifying the current page during the onboarding process to enhance user experience. /build-deploy-preview skip-tests=true /ok-to-test Progress on this is on halt due to critical issue: https://github.com/appsmithorg/appsmith/issues/29373 /ok-to-test
gharchive/pull-request
2023-12-06T10:27:13
2025-04-01T04:33:31.061887
{ "authors": [ "sneha122" ], "repo": "appsmithorg/appsmith", "url": "https://github.com/appsmithorg/appsmith/pull/29377", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2329983955
fix: Wrong evaluated value after binding checkbox group widget with… Description In this PR I have fixed wrong evaluated value after binding checkbox group widget with query Output: Loom [!TIP] Add a TL;DR when the description is longer than 500 words or extremely technical (helps the content, marketing, and DevRel team). Please also include relevant motivation and context. List any dependencies that are required for this change. Add links to Notion, Figma or any other documents that might be relevant to the PR. Fixes #24620 or Fixes https://github.com/appsmithorg/appsmith/issues/24620 [!WARNING] If no issue exists, please create an issue first, and check with the maintainers if the issue is valid. Automation /ok-to-test tags="" :mag: Cypress test results [!CAUTION] If you modify the content in this section, you are likely to disrupt the CI result for your PR. Communication Should the DevRel and Marketing teams inform users about this change? [ ] Yes [x] No Summary by CodeRabbit New Features Introduced test cases for various Anvil widgets, covering Canvas, Preview, and Deploy modes. Added filtering functionality for Table widgets in the UI. Bug Fixes Updated the Checkbox Group Widget to automatically validate and update selected values. Resolved issues with Docker image builds and workflows by updating configurations and scripts. Fixed URL redirection logic for legacy to slug URLs. Chores Updated GitHub Actions workflows for improved CI/CD processes. Adjusted Dockerfile permissions and configurations for better compatibility. Tests Added numerous Cypress end-to-end tests for Anvil widgets and other UI components. Enhanced test scripts to handle Cypress snapshots on failure. /build-deploy-preview Fix works on fixed mode, needs to be propagated to Anvil too cc: @jsartisan @KelvinOm @Shivam-z The solution does not work properly. Check the video. A better approach would be a custom validator for the defaultSelectedValues property. https://github.com/appsmithorg/appsmith/assets/6636360/4962a637-8919-408f-9437-0ed1cec838ef Closing the PR as the solution does not solve it properly. @jsartisan sorry, I was bit busy with the other work , i.e could not able to looked into this PR. can you give me some time to look into this issue and for a better solution. Please reopen this PR. @jsartisan I have modified defaultSelectedValuesValidation could you please review that. The solution looks good to me. Two things the validation is always valid. even when the value is not an array. When the options change, it does not trigger the validation again. The solution to this is to add depencyMap: static getDependencyMap(): Record<string, string[]> { return { defaultSelectedValues: ["options"], }; } Hi @jsartisan , I fixed the second suggestion you gave me. can you give me more clarification on first suggestion "the validation is always valid. even when the value is not an array." here which value you are talking about ? is it defaultSelectedValues ? If it is defaultSelectedValues , then in case of checkBoxGroup Widget it should be an array only right ? Hey, couldn't get time to check again. Regarding this"the validation is always valid. even when the value is not an array.", i think it's fine. @ramsaptami can you check this PR as well? Codewise it looks okay. Need some help in checking validations for queries. /build-deploy-preview skip-tests=true default value when different from derived value is not retained in selectedValues property binding a different query refreshes the 'selectedValues' property cc: @jsartisan @jsartisan can you please look into this PR and tell me whether it is good or I need to rectify something. @Shivam-z Can you check the comments by Saptami? The solution is failing in a few cases. @jsartisan The comments given by @ramsaptami is the expected behaviour not the bug. In the previous version it was showing the defaultValues as selectedValues even though the derived did not contain those default values . For ex: derivedValues: {1,2,3} , defaultValues:{2,3,4} . so selectedValues should be {2,3} but previously it was {2,3,4} i.e 4 was not retained. In the previous version when binding a different query it was not refreshing the 'selectedValues' property but now it is fixed. @Shivam-z Code looks good to me now. I ran the tests on a separate PR and found there is one test that is constantly failing cypress/e2e/Regression/ClientSide/Widgets/Switch/SwitchGroup1_spec.ts https://github.com/appsmithorg/appsmith/assets/6636360/c9d1efbd-a662-42fb-966c-7823d2cb992c Check the video. Btw Sorry for the back and forth, I know it's taking time. Hi @jsartisan , I have updated some code in the cypress test case for SwitchGroup1_spec.ts file and it is running fine in my local. can you please check this PR. @Shivam-z something wrong with your PR. It's now showing 616 files. Can you rebase latest release into your branch I want to create a PR from your PR and run tests on it like we did previously but now your PR is showing 616 files changed. @jsartisan updated file changes @Shivam-z all tests passing now. What was the issue? Hi @jsartisan , some assertions were failing in Switch/SwitchGroup1_spec.ts and it was expecting different color value because of the functionality added in this PR and due to this the test cases were failing earlier , but now it is fixed. Regarding test to cover this functionality , I have already added one test case in : app/client/src/widgets/CheckboxGroupWidget/widget/index.test.tsx
gharchive/pull-request
2024-06-03T01:25:42
2025-04-01T04:33:31.081474
{ "authors": [ "Shivam-z", "jsartisan", "ramsaptami" ], "repo": "appsmithorg/appsmith", "url": "https://github.com/appsmithorg/appsmith/pull/33906", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2408004204
feat:Restrict Phone Number Length Based on User Selection of Max Leng… …th (#34787) Description [!TIP] Add a TL;DR when the description is longer than 500 words or extremely technical (helps the content, marketing, and DevRel team). Please also include relevant motivation and context. List any dependencies that are required for this change. Add links to Notion, Figma or any other documents that might be relevant to the PR. Fixes #34787 or Fixes https://github.com/appsmithorg/appsmith/issues/34787 [!WARNING] If no issue exists, please create an issue first, and check with the maintainers if the issue is valid. Automation /ok-to-test tags="" :mag: Cypress test results [!CAUTION] If you modify the content in this section, you are likely to disrupt the CI result for your PR. Communication Should the DevRel and Marketing teams inform users about this change? [ ] Yes [ ] No VIDEOS VIDEO:SHOWING THE FEATURE VIDEO:CYPRESS TESTING Summary by CodeRabbit New Features Added maxLength property to Phone Input widget, allowing users to specify a maximum character limit. Implemented validation for the Phone Input widget to display an error message when the input exceeds the set maximum character limit. Tests Introduced a new test scenario for validating the maximum character limit enforcement on the Phone Input widget. @Harshithazemoso Can we please jest test cases instead of cypress tests ? cc : @ApekshaBhosale @yatinappsmith @sagar-qa007 Left comments hello @rajatagrawal added a test file and added a test case can you please check it now hello @rajatagrawal can you please review this pr cc : @rahulbarwal Can you please help review the PR ? hello @rahulbarwal ,could you please review this pr . Thank you
gharchive/pull-request
2024-07-15T06:58:45
2025-04-01T04:33:31.090508
{ "authors": [ "Harshithazemoso", "rajatagrawal" ], "repo": "appsmithorg/appsmith", "url": "https://github.com/appsmithorg/appsmith/pull/34925", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2530349482
fix:updated searchbar component to return result with client side sea… Description: modified the code the return the result when client side search is enable and not to when client side search is disabled. fixes: #15386 updates in PR: 1.passed enableClientSide prop to the search component and applied conditional rendering to return result. 2.added test file for the changed code. snapshorts: Summary by CodeRabbit New Features Introduced enableClientSideSearch property across multiple components, allowing for enhanced client-side searching capabilities. Updated SearchComponent, Table, and TableHeader to support dynamic search functionalities based on user input. Bug Fixes Improved state management in the SearchComponent to reset search values appropriately. Tests Added comprehensive unit tests for the SearchComponent to validate its functionality and user interactions. Hii @shadabbuchh @carinanfonseca , I have raised the PR, could you please assign the reviewer to this PR. thank you. Hii @shadabbuchh @carinanfonseca @NilanshBansal ,could you please assign the reviewer to this PR. thank you. Adding @rahulbarwal @jacquesikot from the widgets team to check this out. Hii @rahulbarwal , I have updated the test case names, could you check the changes. thank you. @Naveen-Goud CI is failing for your changes: https://github.com/appsmithorg/appsmith/actions/runs/11230421118/job/31217757085?pr=36745 Please check and fix. @Naveen-Goud can you please fix the failing CI issues. Hii @rahulbarwal , I have resolved the lint issue , can you re-run the checks. thank you. @Naveen-Goud client lint is still failing on your changes: https://github.com/appsmithorg/appsmith/actions/runs/11321677640/job/31481100928?pr=36745#step:9:379 Please run yarn run lint:ci in your local to confirm after fixing. @Naveen-Goud are you still working on this?
gharchive/pull-request
2024-09-17T07:59:44
2025-04-01T04:33:31.099782
{ "authors": [ "Naveen-Goud", "NilanshBansal", "rahulbarwal" ], "repo": "appsmithorg/appsmith", "url": "https://github.com/appsmithorg/appsmith/pull/36360", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2552536384
fix: fixing modal height by removing min-content Description min-content css property is behaving weirdly in the latest update of the chrome browser Fixes https://github.com/appsmithorg/appsmith/issues/36586 Automation /ok-to-test tags="" :mag: Cypress test results [!CAUTION] If you modify the content in this section, you are likely to disrupt the CI result for your PR. Communication Should the DevRel and Marketing teams inform users about this change? [ ] Yes [ ] No /build-deploy-preview
gharchive/pull-request
2024-09-27T10:07:23
2025-04-01T04:33:31.103061
{ "authors": [ "brayn003" ], "repo": "appsmithorg/appsmith", "url": "https://github.com/appsmithorg/appsmith/pull/36589", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
355842359
not working after update my flutter to 2.0 error in get package Same Does this fix your problem? https://github.com/apptreesoftware/flutter_google_map_view/pull/115 @MathieuAuclair Could not find a file named "pubspec.yaml" in "/home/hossein/AndroidStudioProjects/mahout/packages/flutter_google_map_view-master/lib/dependencies/uri.dart". Same, Flutter 0.6.0 • channel beta • https://github.com/flutter/flutter.git Framework • revision 9299c02cf7 (3 weeks ago) • 2018-08-16 00:35:12 +0200 Engine • revision e3687f70c7 Tools • Dart 2.1.0-dev.0.0.flutter-be6309690f thanks anyway for your work. Map_view is depending on uri 0.11.1 which is outdated. Someone should upgrade the dependencies immediately.
gharchive/issue
2018-08-31T05:16:04
2025-04-01T04:33:31.125779
{ "authors": [ "MathieuAuclair", "andreagr", "dbbd59", "elricym", "hojjabr", "sudipta213" ], "repo": "apptreesoftware/flutter_google_map_view", "url": "https://github.com/apptreesoftware/flutter_google_map_view/issues/114", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2446442033
[Feature Request]: add smooth scrolling Is there an existing issue for this? [X] I have searched the existing issues Feature Description The ui is very laggy .Adding a smooth scroll effect can make it slight better. Use Case Enhancing User Experience. Benefits No response Add ScreenShots No response Priority High Record [X] I have read the Contributing Guidelines [X] I'm a GSSOC'24 contributor [ ] I'm a VSOC'24 contributor [X] I have starred the repository Hii, I want to tackle this issue, So please assign to me. Hii, I want to tackle this issue, So please assign to me. u are assigned
gharchive/issue
2024-08-03T16:14:19
2025-04-01T04:33:31.177017
{ "authors": [ "Ranj8521Kumar", "Subhams-GIT", "apu52" ], "repo": "apu52/Travel_Website", "url": "https://github.com/apu52/Travel_Website/issues/1308", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2349727709
Email text modification Problem * The email text is not properly positioned like other text. * I marked the problem with a green marker. [X] I am contributing to an open-source program in GSSOC'24. @Shashanktriathi1703 please assign this issue to me under gssoc @Shashanktriathi1703 I have assigned u this time .But please from next time try to raise the issues using the issue templates in the "NEW ISSUE" section
gharchive/issue
2024-06-12T21:40:21
2025-04-01T04:33:31.179506
{ "authors": [ "Salma-Mamdoh", "Shashanktriathi1703", "apu52" ], "repo": "apu52/Travel_Website", "url": "https://github.com/apu52/Travel_Website/issues/859", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1528257263
Feature/tables view Task details: https://quotron.notion.site/Tables-view-7d00ce65a4294cc2904860277ff2c351 merged
gharchive/pull-request
2023-01-11T01:37:33
2025-04-01T04:33:31.180605
{ "authors": [ "apurvak", "whitewolfdev410" ], "repo": "apurvak/superset", "url": "https://github.com/apurvak/superset/pull/42", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1638095044
chore: call bpf_program__attach_iter from go This changes libbpfgo to call bpf_program__attach_iter directly from go instead of doing this via a C wrapper. What this means is that we maintained opts creation in C, but the actual libbpf attach call is done in go. This creates two new functions in libbpfgo.h as helpers: bpf_iter_attach_opts_new() bpf_iter_attach_opts_free() As questioned in https://github.com/aquasecurity/libbpfgo/pull/254#discussion_r1144230843. Thanks for that @geyslan. My OCD sent compliments!
gharchive/pull-request
2023-03-23T18:37:41
2025-04-01T04:33:31.201095
{ "authors": [ "geyslan" ], "repo": "aquasecurity/libbpfgo", "url": "https://github.com/aquasecurity/libbpfgo/pull/304", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1506332436
[Feature] Add support for severity and ignore-unfixed parameters At the moment it's not possible to specify which severities should be scannen for. Also unfixable severities can't be hidden. This is all possible with existing arguments for the trivy command. duplicate of #16
gharchive/issue
2022-12-21T13:54:21
2025-04-01T04:33:31.202225
{ "authors": [ "ferronsw" ], "repo": "aquasecurity/trivy-azure-pipelines-task", "url": "https://github.com/aquasecurity/trivy-azure-pipelines-task/issues/26", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1954066962
Missing ServiceAccount, some Role rules and RoleBinding with OLM installation Description Hi all, I'm Following the procedure described in the official docs, and after creating the Subscription resource the status of the cluster service version CR is Pending: $ kubectl get clusterserviceversions.operators.coreos.com -n trivy-system NAME DISPLAY VERSION REPLACES PHASE trivy-operator.v0.16.3 Trivy Operator 0.16.3 trivy-operator.v0.16.2 Pending The initial reason is the absence of the ServiceAccount resource named trivy-operator: $ kubectl describe clusterserviceversions.operators.coreos.com -n trivy-system ... ... Group: Kind: ServiceAccount Message: Service account does not exist Name: trivy-operator Status: NotPresent Version: v1 Events: Type Reason Age From Message ---- ------ ---- ---- ------- Normal RequirementsUnknown 10m operator-lifecycle-manager requirements not yet checked Normal RequirementsNotMet 10m operator-lifecycle-manager one or more requirements couldn't be found But even after creating it, some permissions are missing in the ClusterRole named trivy-operator: $ kubectl describe clusterserviceversions.operators.coreos.com -n trivy-system ... ... Group: rbac.authorization.k8s.io Kind: PolicyRule Message: cluster rule:{"verbs":["get","list","watch","create","update","delete"],"apiGroups":["aquasecurity.github.io"],"resources":["vulnerabilityreports","configauditreports","clustercompliancereports","clusterconfigauditreports","exposedsecretreports","sbomreports","rbacassessmentreports","infraassessmentreports","clusterrbacassessmentreports","clusterinfraassessmentreports"]} Status: NotSatisfied Version: v1 Group: rbac.authorization.k8s.io Kind: PolicyRule Message: cluster rule:{"verbs":["create","get","update"],"apiGroups":["coordination.k8s.io"],"resources":["leases"]} Status: NotSatisfied Version: v1 Group: Kind: ServiceAccount Message: Policy rule not satisfied for service account Name: trivy-operator Status: PresentNotSatisfied Version: v1 Events: Type Reason Age From Message ---- ------ ---- ---- ------- Normal RequirementsUnknown 0s operator-lifecycle-manager requirements not yet checked Normal RequirementsNotMet 0s operator-lifecycle-manager one or more requirements couldn't be found After manually adding everything, the Operator gets installed, so I'm using this yaml that contains what is missing: # ServiceAccount apiVersion: v1 kind: ServiceAccount metadata: name: trivy-operator namespace: trivy-system labels: app.kubernetes.io/name: trivy-operator app.kubernetes.io/instance: trivy-operator app.kubernetes.io/version: "0.1.4" app.kubernetes.io/managed-by: kubectl --- # ClusterRole apiVersion: rbac.authorization.k8s.io/v1 kind: ClusterRole metadata: creationTimestamp: null name: trivy-operator rules: #################################################### # These are the installed rules: - apiGroups: - "" resources: - configmaps verbs: - get - list - watch - apiGroups: - "" resources: - limitranges verbs: - get - list - watch - apiGroups: - "" resources: - pods verbs: - get - list - watch - apiGroups: - "" resources: - pods/log verbs: - get - list - apiGroups: - "" resources: - replicationcontrollers verbs: - get - list - watch - apiGroups: - "" resources: - resourcequotas verbs: - get - list - watch - apiGroups: - "" resources: - secrets verbs: - create - get - apiGroups: - "" resources: - serviceaccounts verbs: - get - apiGroups: - "" resources: - services verbs: - get - list - watch - apiGroups: - apiextensions.k8s.io resources: - customresourcedefinitions verbs: - get - list - watch - apiGroups: - apps resources: - daemonsets verbs: - get - list - watch - apiGroups: - apps resources: - replicasets verbs: - get - list - watch - apiGroups: - apps resources: - statefulsets verbs: - get - list - watch - apiGroups: - aquasecurity.github.io resources: - clustercompliancedetailreports verbs: - create - delete - get - list - patch - update - watch - apiGroups: - aquasecurity.github.io resources: - clustercompliancereports verbs: - create - delete - get - list - patch - update - watch - apiGroups: - aquasecurity.github.io resources: - clustercompliancereports/status verbs: - get - patch - update - apiGroups: - aquasecurity.github.io resources: - clusterconfigauditreports verbs: - create - delete - get - list - patch - update - watch - apiGroups: - aquasecurity.github.io resources: - clusterrbacassessmentreports verbs: - create - delete - get - list - patch - update - watch - apiGroups: - aquasecurity.github.io resources: - configauditreports verbs: - create - delete - get - list - patch - update - watch - apiGroups: - aquasecurity.github.io resources: - exposedsecretreports verbs: - create - delete - get - list - patch - update - watch - apiGroups: - aquasecurity.github.io resources: - rbacassessmentreports verbs: - create - delete - get - list - patch - update - watch - apiGroups: - aquasecurity.github.io resources: - vulnerabilityreports verbs: - create - delete - get - list - patch - update - watch - apiGroups: - batch resources: - cronjobs verbs: - get - list - watch - apiGroups: - batch resources: - jobs verbs: - create - delete - get - list - watch - apiGroups: - networking.k8s.io resources: - ingresses verbs: - get - list - watch - apiGroups: - networking.k8s.io resources: - networkpolicies verbs: - get - list - watch - apiGroups: - rbac.authorization.k8s.io resources: - clusterrolebindings verbs: - get - list - watch - apiGroups: - rbac.authorization.k8s.io resources: - clusterroles verbs: - get - list - watch - apiGroups: - rbac.authorization.k8s.io resources: - rolebindings verbs: - get - list - watch - apiGroups: - rbac.authorization.k8s.io resources: - roles verbs: - get - list - watch #################################################### # These are the missing rules that I manually added: - verbs: - get - list - watch apiGroups: - "" resources: - pods - pods/log - replicationcontrollers - services - resourcequotas - limitranges - verbs: - get - list - watch apiGroups: - "" resources: - nodes - verbs: - list - watch - get - create - update apiGroups: - "" resources: - configmaps - secrets - serviceaccounts - verbs: - delete apiGroups: - "" resources: - secrets - verbs: - create apiGroups: - "" resources: - events - verbs: - get - list - watch apiGroups: - apps resources: - replicasets - statefulsets - daemonsets - deployments - verbs: - get - list - watch apiGroups: - policy resources: - podsecuritypolicies - verbs: - get - list - watch - create - update - delete apiGroups: - aquasecurity.github.io resources: - vulnerabilityreports - configauditreports - clustercompliancereports - clusterconfigauditreports - exposedsecretreports - sbomreports - rbacassessmentreports - infraassessmentreports - clusterrbacassessmentreports - clusterinfraassessmentreports - verbs: - create - get - update apiGroups: - coordination.k8s.io resources: - leases --- # ClusterRoleBinding apiVersion: rbac.authorization.k8s.io/v1 kind: ClusterRoleBinding metadata: name: trivy-operator labels: app.kubernetes.io/name: trivy-operator app.kubernetes.io/instance: trivy-operator app.kubernetes.io/version: "0.1.4" app.kubernetes.io/managed-by: kubectl roleRef: apiGroup: rbac.authorization.k8s.io kind: ClusterRole name: trivy-operator subjects: - kind: ServiceAccount name: trivy-operator namespace: trivy-system Issue I was expecting everything to be covered by the subscription, but this also makes me feel like I'm missing something. Is this a normal behavior? Is it expected that the admin must create the SA/Role/RoleBinding by himself before creating the Subscription? This doesn't happen while installing everything using just kubectl. Environment Trivy-Operator version (use trivy-operator version): 0.16.3 Kubernetes version (use kubectl version): v1.27.4 OS (macOS 10.15, Windows 10, Ubuntu 19.10 etc): AlmaLinux release 8.8 (Sapphire Caracal) @rascasoft has is a ref for operator I'll take a look at it as well. also please review operator install instruction Thanks for your answer @chen-keinan , but I didn't understand what you're trying to tell me. The first link is the manifest of the cluster service version of the trivy-operator in the operatorhub sources, and I'm not able to understand from here what I want to know, which is if it's on purpose that service account and role/rolebindings are not created by the operator installer. The second link takes me to the instructions which I already followed and don't tell me much more compared to what is the doc I followed (this one doesn't cover the operatorgroup resource creation for instance). Am I missing something? Many thanks! Thanks for your answer @chen-keinan , but I didn't understand what you're trying to tell me. The first link is the manifest of the cluster service version of the trivy-operator in the operatorhub sources, and I'm not able to understand from here what I want to know, which is if it's on purpose that service account and role/rolebindings are not created by the operator installer. The second link takes me to the instructions which I already followed and don't tell me much more compared to what is the doc I followed (this one doesn't cover the operatorgroup resource creation for instance). Am I missing something? Many thanks! its just a ref for operator hub OLM official install instructions, I'll have another look myself and see if I can find the issue Many thanks @chen-keinan. There's one main thing I would like to understand (and maybe you can help here): when I configure a subscription for OLM, are all the resources that are going to be used (so Service Accounts, ClusterRole, ClusterRoleBinding, etc.) expected to be created by the subscription itself? I ask this because if you use OLM in OpenShift, then everything works flawlessly, but in K8s things seem different. Many thanks! Many thanks @chen-keinan. There's one main thing I would like to understand (and maybe you can help here): when I configure a subscription for OLM, are all the resources that are going to be used (so Service Accounts, ClusterRole, ClusterRoleBinding, etc.) expected to be created by the subscription itself? I ask this because if you use OLM in OpenShift, then everything works flawlessly, but in K8s things seem different. Many thanks! yes , this is how it works based on my understanding I've got an addition to made at this, after further testing: it seems that if you install OLM using operator-sdk the behavior changes to what you expect. On a clean environment I installed OLM by using: $ operator-sdk olm install And then I created OperaotrGroup and Subscription as follows: apiVersion: operators.coreos.com/v1 kind: OperatorGroup metadata: name: trivy-operator-group namespace: trivy-system spec: targetNamespaces: - trivy-system - mions1 --- apiVersion: operators.coreos.com/v1alpha1 kind: Subscription metadata: name: trivy-operator-subscription namespace: trivy-system spec: channel: alpha name: trivy-operator source: operatorhubio-catalog sourceNamespace: olm installPlanApproval: Automatic config: env: - name: OPERATOR_EXCLUDE_NAMESPACES value: "kube-system,trivy-system The CSV installed correctly: $ kubectl get csv -n trivy-system NAME DISPLAY VERSION REPLACES PHASE trivy-operator.v0.16.4 Trivy Operator 0.16.4 trivy-operator.v0.16.3 Succeeded The only remaining problem was the absence of the trivy-operator-trivy-config config map inside the trivy-system namespace, which was causing this message to appear in the pod logs: {"level":"error","ts":"2023-11-14T16:53:09Z","msg":"Reconciler error","controller":"clusterrole","controllerGroup":"rbac.authorization.k8s.io","controllerKind":"ClusterRole","ClusterRole":{"name":"rbacassessmentreports.aquasecurity.github.io-v1alpha1-crdview"},"namespace":"","name":"rbacassessmentreports.aquasecurity.github.io-v1alpha1-crdview","reconcileID":"4f4e747e-6947-416c-bada-22e270b1f9fe","error":"ConfigMap \"trivy-operator-trivy-config\" not found","stacktrace":"sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/home/runner/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/internal/controller/controller.go:329\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/home/runner/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/home/runner/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/internal/controller/controller.go:227"} So I created it manually, using the code from the static manifest: apiVersion: v1 kind: ConfigMap metadata: name: trivy-operator-trivy-config namespace: trivy-system labels: app.kubernetes.io/name: trivy-operator app.kubernetes.io/instance: trivy-operator app.kubernetes.io/version: "0.16.4" app.kubernetes.io/managed-by: kubectl data: trivy.repository: "ghcr.io/aquasecurity/trivy" trivy.tag: "0.47.0" trivy.imagePullPolicy: "IfNotPresent" trivy.additionalVulnerabilityReportFields: "" trivy.severity: "UNKNOWN,LOW,MEDIUM,HIGH,CRITICAL" trivy.slow: "true" trivy.skipJavaDBUpdate: "false" trivy.imageScanCacheDir: "/tmp/trivy/.cache" trivy.filesystemScanCacheDir: "/var/trivyoperator/trivy-db" trivy.dbRepository: "ghcr.io/aquasecurity/trivy-db" trivy.javaDbRepository: "ghcr.io/aquasecurity/trivy-java-db" trivy.command: "image" trivy.dbRepositoryInsecure: "false" trivy.useBuiltinRegoPolicies: "true" trivy.supportedConfigAuditKinds: "Workload,Service,Role,ClusterRole,NetworkPolicy,Ingress,LimitRange,ResourceQuota" trivy.timeout: "5m0s" trivy.mode: "Standalone" trivy.resources.requests.cpu: "100m" trivy.resources.requests.memory: "100M" trivy.resources.limits.cpu: "500m" trivy.resources.limits.memory: "500M" Even if I'm still getting some weird logs inside the Pod: {"level":"error","ts":"2023-11-14T16:55:40Z","msg":"Reconciler error","controller":"job","controllerGroup":"batch","controllerKind":"Job","Job":{"name":"scan-vulnerabilityreport-75dc95d75c","namespace":"trivy-system"},"namespace":"trivy-system","name":"scan-vulnerabilityreport-75dc95d75c","reconcileID":"94e371e7-4f8e-4499-b918-39945792221f","error":"SbomReport.aquasecurity.github.io \"replicaset-nginx-8bc664746-nginx\" is invalid: report.components.specVersion: Invalid value: \"string\": report.components.specVersion in body must be of type integer: \"string\"","stacktrace":"sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/home/runner/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/internal/controller/controller.go:329\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/home/runner/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/home/runner/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/internal/controller/controller.go:227"} The configauditreport and vulnerabilityreport are generated. So I see two topics here: Understand what is the difference between manual and operator-sdk OLM installation (not to be addressed here). Fix the ConfigMap generation inside the trivy-operator manifest (which becomes the main issue of this bug). Thanks, i'll have a look at this feedback @chen-keinan I have created a patch , it will be released with trivy-operator v0.17.0
gharchive/issue
2023-10-20T11:10:19
2025-04-01T04:33:31.220988
{ "authors": [ "chen-keinan", "rascasoft" ], "repo": "aquasecurity/trivy-operator", "url": "https://github.com/aquasecurity/trivy-operator/issues/1584", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1266988904
Eng 588 allow users to upload csvs to demo be https://github.com/aqueducthq/aqueduct/pull/82 -- BE API endpoint to create table in demo DB from CSV https://github.com/aqueducthq/aqueduct/pull/83 -- FE UI to allow upload of CSV to demo DB https://github.com/aqueducthq/aqueduct/pull/55 -- FE Integrations Details page where this functionality lives A small request on PR naming @eunice-chan -- when we're stacking multiple PRs like this, it would be great if we included the system component in the title. Otherwise, it's kinda difficult to tell which one is which. e.g., #83 could be titled "ENG 588: [UI] Allow users to upload CSVs to demo DB" and #82 could be titled "ENG 588: [Server] Add endpoint for accepting and uploading new table". Does that make sense?
gharchive/pull-request
2022-06-10T04:03:39
2025-04-01T04:33:31.229658
{ "authors": [ "eunice-chan", "vsreekanti" ], "repo": "aqueducthq/aqueduct", "url": "https://github.com/aqueducthq/aqueduct/pull/82", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
912749764
Decision Transformer: Reinforcement Learning via Sequence Modeling 一言でいうと Transformerを強化学習に応用した研究。State/Action/Rewardの系列を入力して次の行動を予測させる。収録済みの軌跡から学習するオフライン強化学習で、既存の手法を上回る精度(オンラインの強化学習ではまだ検証されていない)。 論文リンク https://arxiv.org/abs/2106.01345 著者/所属機関 Lili Chen, Kevin Lu, Aravind Rajeswaran, Kimin Lee, Aditya Grover, Michael Laskin, Pieter Abbeel, Aravind Srinivas, Igor Mordatch UC Berkeley Facebook AI Research Google Brain 投稿日付(yyyy/MM/dd) 2021/6/2 概要 新規性・差分 手法 結果 コメント 解説資料 https://www.slideshare.net/yasunoriozaki12/decision-transformer-reinforcement-learning-via-sequence-modeling
gharchive/issue
2021-06-06T09:40:16
2025-04-01T04:33:31.234226
{ "authors": [ "icoxfog417" ], "repo": "arXivTimes/arXivTimes", "url": "https://github.com/arXivTimes/arXivTimes/issues/2063", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
383394455
script hangs at "checking major 3rd party components" the issue could very well be on my end, however I am using a fresh intall of ubuntu minimal 14.04.1. upon running rtinst, it hangs at the xmlrpc-c check. here is terminal output: user@server:~# rtinst Ubuntu 14.04.1 LTS Checking the web sites we will need are accessible Checking Ubuntu mirrors http://archive.ubuntu.com/ubuntu: OK http://security.ubuntu.com/ubuntu: OK http://archive.canonical.com/ubuntu: OK Checking major 3rd party components Rtorrent: OK xmlrpc-c: it's odd because the pages referenced by xmlrpc_url and xmlrpc_url_alt both seem to resolve for me. i also noticed there's some error messages to display if the urls don't resolve, but the script doesn't seem to get that far. i git cloned the repo and changed the xmlrpc-c URL to the one referenced in commit #29a3ad9 but that didn't seem to fix it. i'm not actually sure what else i could try but i'm gonna play with it some more. line #293 and #295 of /scripts/rtinst reference xmlrpc_repo=1 but this doesn't seem to be set, looks like it belongs at line #278. i added this, i alo switched xmlrpc_url and xmlrpc_alt because maybe there was a resolving error there? and i added a failure message similar to the other checks, but the script still hangs for me. # check required web repos are accessible os_repo=0 major_repo=0 xmlrpc_repo=0 sed -i "s/\/debian\s/\/debian\/ /g" /etc/apt/sources.list echo echo "Checking the web sites we will need are accessible" echo "Checking $osname mirrors" for i in $(cat /etc/apt/sources.list | grep "^deb http" | cut -d' ' -f2 | uniq ); do echo -n $i": " check_url $i && echo "OK" || { echo "FAIL"; os_repo=1; } done echo echo "Checking major 3rd party components" echo -n "Rtorrent: "; check_url $rt_url && echo "OK" || { echo "FAIL"; major_repo=1; } echo -n "xmlrpc-c: "; check_url $xmlrpc_url && echo "OK" || { echo "FAIL"; xmlrpc_repo=1; } if [[ $xmlrpc_repo = 1 ]]; then xmlrpc_url=$xmlrpc_url_alt check_url $xmlrpc_url && echo "OK" || { echo "FAIL"; major_repo=1; } fi Same here. hang at xmlrpc-c: any solutions? Is this issue still present? If so, please add the following to rtinst script after #!/usr/bin/env bash on the top: set -e -x And remember to make the script executable again with: chmod +x /etc/rtinst/scripts/rtinst Re-run the script and post the whole output.
gharchive/issue
2018-11-22T06:17:01
2025-04-01T04:33:31.262815
{ "authors": [ "V33m", "ieffedmybffjill", "lowprofileusername" ], "repo": "arakasi72/rtinst", "url": "https://github.com/arakasi72/rtinst/issues/381", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1772140021
Client component translations are coming one step back. What version of this package are you using? 2.3.1 What operating system, Node.js, and npm version? NodeJs=> 16.17.1 Npm=> 8.15.0 Next=>13.4.4 What happened? If I change the language using the language switcher component, Client components' translations are coming one step back. If I choose French as a default language it works as expected at first. Then if I change the language to Dutch, as you can see the buttons translations are still French Then If I change the language to English, now buttons translations are Dutch(one step back) My language switcher component 'use client'; import Link from 'next/link'; import { usePathname } from 'next/navigation'; const LanguageOptions = [ { label: 'English', value: 'en' }, { label: 'Français', value: 'fr' }, { label: 'Nederlands', value: 'nl' }, ]; export default function LanguageSwitcher() { const pathname = usePathname(); const pathArray = pathname.split('/'); //Remove the first two elements from an array const pathWithoutLang = pathArray.slice(2); // join the array back to a string const newPath = pathWithoutLang.join('/'); return ( <div className='px-1 bg-gray-200 rounded-lg outline-0'> {LanguageOptions.map((locale) => ( <> <Link key={locale.label} href={`/?lang=${locale.value + '/' + newPath}`} as={`/${locale.value + '/' + newPath}`} > {locale.value} </Link> <br /> </> ))} </div> ); } app/[lang]/layout.tsx import { Inter } from 'next/font/google'; import { redirect } from 'next/navigation'; import React from 'react'; import i18n from '../../i18n'; import './globals.css'; const inter = Inter({ subsets: ['latin'] }); export const metadata = { title: 'Create Next App', description: 'Generated by creating next app', }; export default function RootLayout({ children, params: { lang }, }: { children: React.ReactNode; params: { lang: string }; }) { // Redirect to default locale if lang is not supported. /second-page -> /en/second-page if (!i18n.locales.includes(lang)) { redirect(`/${i18n.defaultLocale}/${lang}`); } return ( <html lang={lang}> <body className={inter.className}>{children}</body> </html> ); } i18n.js module.exports = { locales: ['en', 'fr', 'nl'], defaultLocale: 'en', pages: { '*': ['common'], '/[lang]': ['form', 'login'], '/[lang]/deneme': ['login', 'form'], '/[lang]/onboarding': ['common'], '/[lang]/signup': ['common', 'signup', 'login', 'form'], }, }; // middleware.ts import type { NextRequest } from 'next/server'; import { NextResponse } from 'next/server'; import i18n from './i18n'; // /es/page-name -> rewrites to -> /es/page-name?lang=es export function middleware(request: NextRequest) { const locale = request.nextUrl.locale || i18n.defaultLocale; request.nextUrl.searchParams.set('lang', locale); return NextResponse.rewrite(request.nextUrl); } What did you expect to happen? I want to use Next js' new app directory with next-translate. I expect that rendering client/server components together with multiple languages Are you willing to submit a pull request to fix this bug? Try the latest 2.4.0 version please BTW; you can remove the middleware Thank you it fixed the issue. 🥳 I appreciate your help.
gharchive/issue
2023-06-23T20:55:42
2025-04-01T04:33:31.269988
{ "authors": [ "aralroca", "kod-man" ], "repo": "aralroca/next-translate", "url": "https://github.com/aralroca/next-translate/issues/1084", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
503440402
Feature/aql subquery single input tests This PR adds a set of disabled tests for the AQL subquery refactoring. The plan is that we can continue implementing the missing parts in the code and one after the other activate the tests within this suite. Contains tests for PassThrough blocks Sort Nested empty collect http://jenkins01.arangodb.biz:8080/view/PR/job/arangodb-matrix-pr/6615/ The suite might not be complete yet, please feel free to suggest more test cases / ideas. LGTM. Jenkins green besides the known Skip in Example http://jenkins.arangodb.biz:8080/job/arangodb-matrix-pr/6632/
gharchive/pull-request
2019-10-07T13:06:28
2025-04-01T04:33:31.279365
{ "authors": [ "markuspf", "mchacki" ], "repo": "arangodb/arangodb", "url": "https://github.com/arangodb/arangodb/pull/10175", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
463682934
@maierlars :heart_eyes: Scope & Purpose Port some small VelocyPack improvements from arangodb/velocypack to arangodb/arangodb [x] Bug-Fix for devel-branch (i.e. no need for backports?) [x] Strictly new functionality (i.e. a new feature / new option, no need for porting) [x] The behaviour change can only be verified via automatic tests Testing & Verification This change is already covered by existing tests, such as velocypack tests. [x] There are tests in an external testing repository (velocypack) [x] I ensured this code runs with ASan / TSan or other static verification tools https://jenkins01.arangodb.biz/view/PR/job/arangodb-matrix-pr/5041/ https://jenkins01.arangodb.biz/view/PR/job/arangodb-matrix-pr/5049/ https://jenkins01.arangodb.biz/view/PR/job/arangodb-matrix-pr/5059/
gharchive/pull-request
2019-07-03T10:51:31
2025-04-01T04:33:31.283148
{ "authors": [ "jsteemann" ], "repo": "arangodb/arangodb", "url": "https://github.com/arangodb/arangodb/pull/9394", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
299374159
Transaction: throw custom exception, with custom code Hi, Im' trying to throw a custom exception in transaction's action, but I can not set the errorNumber. arangojs: 6.1.1 ArangoDB: 3.3.3 [linux] 64bit RocksDB 5.6.0 Code: const action = String((params) => { var error = new Error('Conflict'); error.code = 409; error.errorNum = 409; error.statusCode = 409; error.errorNumber = 409; throw error; }); try { await DB.db.transaction( { read: 'test', write: 'test' }, action, { id: 1 } ); } catch (error) { console.log(error); } Output: { error: true, errorMessage: 'Error: Conflict', code: 500, errorNum: 4 } I expect to catch an exception with my custom error code. Did I do something wrong? Thanks! Hi, its working the following way: const arangodb = require('@arangodb') var err = new arangodb.ArangoError(); err.errorNum = arangodb.ERROR_BAD_PARAMETER; // <usually some well defined arango error err.errorNum = 12345; /// but here some random value err.errorMessage = "who's bad?"; throw err; It works! Thanks Ok, I've fixed the documentation, and we now will also forward errorNum in other cases, plus added tests. Did you find the examples in the arangojs documentation or just from https://docs.arangodb.com/devel/Manual/Transactions/TransactionInvocation.html#custom-exceptions ?
gharchive/issue
2018-02-22T14:28:53
2025-04-01T04:33:31.287131
{ "authors": [ "AntoineAA", "dothebart" ], "repo": "arangodb/arangojs", "url": "https://github.com/arangodb/arangojs/issues/508", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
454755361
Add docs for so far undocumented Aardvark API Also see https://github.com/arangodb/planning/issues/3897 Aardvark Swagger JSON is available via Web UI (using documentation router?): http://localhost:8529/_db/_system/_admin/aardvark/foxxes/docs/swagger.json?mount=/_admin/aardvark No intention to document the internal Aardvark API.
gharchive/pull-request
2019-06-11T15:22:38
2025-04-01T04:33:31.288818
{ "authors": [ "Simran-B" ], "repo": "arangodb/docs", "url": "https://github.com/arangodb/docs/pull/74", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1843146828
MPI job aborts if MPI_Finalize is called in the same scope as a DistributedTree instance is destroyed The following snippet makes MPI aborts the current job. int main(int argc, char *argv[]) { MPI_Init(&argc, &argv); Kokkos::ScopeGuard guard(argc, argv); Kokkos::View<ArborX::Point *, Kokkos::HostSpace> data("MpiError::data", 1); ArborX::DistributedTree<Kokkos::HostSpace> tree(MPI_COMM_WORLD, Kokkos::Serial{}, data); MPI_Finalize(); return 0; } *** The MPI_Comm_free() function was called after MPI_FINALIZE was invoked. *** This is disallowed by the MPI standard. *** Your MPI job will now abort. [mac122579:07159] Local abort after MPI_FINALIZE started completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed! Moving the DistributedTree instance in another scope removes this error messages. I am guessing this error is due to the destructor of the internal comm pointer of DistributedTree. Adding a check to free the comm pointer if and only if MPI is still running does the trick. [](MPI_Comm *p) { int mpi_finalized; MPI_Finalized(&mpi_finalized); if (!mpi_finalized) MPI_Comm_free(p); delete p; }); In other words, this is expected. We just need to make sure that this behavior is documented. Related to https://github.com/arborx/Wiki/pull/53.
gharchive/issue
2023-08-09T12:36:47
2025-04-01T04:33:31.302866
{ "authors": [ "dalg24", "masterleinad", "mrlag31" ], "repo": "arborx/ArborX", "url": "https://github.com/arborx/ArborX/issues/925", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1640477561
useFormSplitter while triggering handlerSubmit validates the whole form and does not update (React Native) Description While creating a wizard form with multiple steps BUG 1 - While doing handleSubmit from useFormSplitter the form doest not update if is not touched, BUG 2- While doing handleSubmit from useFormSplitter is receiving all errors for the whole form instead of the separed step. Steps to reproduce 1- Create a form with multiple steps using useFormSplitter 2 - On de step 1 trigger the handlerSubmit (if not touched fields will not rerender the component since the bug is clearing the changedKeys every time) 3 - If you change to the second step will appear the errors too even if we didn't do anything on the second step. Expected result 2 - Update the screen with the correct errors 3 - Changing the step from 1 to 2 should not appear de errors on the 2 step Actual result Step 1 does not show the errors Step 2 show errors event we didn't do anything Environment Node -v v14.21.2 "@resourge/react-form": "1.12.0" "@resourge/schema": "1.9.4" "react-native": "0.71.3" I already made a fix for both of the bugs but needs a review. Here is the patch with production bundle in RN PATCH The original fille changed with the fixes`() @resourge\react-form\dist\umd\react-form.development.js Next week i can create a PR if needed
gharchive/issue
2023-03-25T11:07:30
2025-04-01T04:33:31.374529
{ "authors": [ "archie-rp" ], "repo": "archie-rp/resourge-react-form", "url": "https://github.com/archie-rp/resourge-react-form/issues/29", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1006539829
Add isLocal argument to fromJsDate method of Datetime Sometimes we want to transform a JavaScript Date to our Datetime wrapper AND keep the local time zone. The time zone is always UTC in the current implementation: static fromJsDate(date: Date) { return new Datetime(LuxonDatetime.fromJSDate(date)) } Because the Datetime constructor already supports the isLocal argument, I think the following implementation change is the more accurate: static fromJsDate(date: Date, isLocal: boolean = false) { return new Datetime(LuxonDatetime.fromJSDate(date), isLocal) } Maybe the second parameter should be an object in order to support more options in the future. Will that suffice @esuca? This will be released in v2!
gharchive/issue
2021-09-24T14:43:30
2025-04-01T04:33:31.379770
{ "authors": [ "cesalberca", "esuca" ], "repo": "archimedes-projects/archimedes-js", "url": "https://github.com/archimedes-projects/archimedes-js/issues/7", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
902717071
Add and improve Inventory validation Thank you for helping out! ✨ We really appreciate your commitment to improving Architect To maintain a high standard of quality in our releases, before merging every pull request we ask that you've completed the following: [x] Forked the repo and created your branch from master [x] Made sure tests pass (run npm it from the repo root) [x] Expanded test coverage related to your changes: [x] Added and/or updated unit tests (if appropriate) [x] Added and/or updated integration tests (if appropriate) [x] Updated relevant documentation: [x] Internal to this repo (e.g. readme.md, help docs, inline docs & comments, etc.) [x] Architect docs (arc.codes) [ ] Summarized your changes in changelog.md [ ] Linked to any related issues, PRs, etc. below that may relate to, consume, or necessitate these changes Please also be sure to completed the CLA (if you haven't already). Learn more about contributing to Architect here. Thanks again! 😮 Ok, this wound up being a bit bigger than expected, but it was some necessary tech debt to pay down. This PR should do the following: Totally refactors error handling to support aggregation of many validation errors (instead of just one at a time) Adds comprehensive validation to all pragmas that require validation Gets unit test coverage to 💯% Fixes a bunch of smol bugs along the way Pretty much all the test changes were to failure cases. Since all the normal cases that were tested before are still passing now, this should be a drop-in – but given how much changed and how important Inventory is, I'd like to extensively test anyway! Should fix: https://github.com/architect/architect/issues/710 https://github.com/architect/architect/issues/951
gharchive/pull-request
2021-05-26T17:25:25
2025-04-01T04:33:31.388867
{ "authors": [ "filmaj", "ryanblock" ], "repo": "architect/inventory", "url": "https://github.com/architect/inventory/pull/19", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1883129068
Input加一个props,可以在不同移动端设备上聚焦的时候只改变网页高度 [ ] I'm sure this does not appear in the issue list of the repository Basic Info What are the similar cases of this feature none What problem does this feature solve? 目前IOS设备上输入框聚焦会把网页弹起,想着如果可以聚焦的时候动态改变网页高度为屏幕高度减去键盘高度就更方便了 哈哈哈,如果可以的话,加一个这个props,true的时候有这个表现,false的时候就默认就好了 这个做不到,无法拿到键盘高度哈
gharchive/issue
2023-09-06T03:45:55
2025-04-01T04:33:31.427210
{ "authors": [ "2513483494", "TinaPeach" ], "repo": "arco-design/arco-design-mobile", "url": "https://github.com/arco-design/arco-design-mobile/issues/162", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
355561263
When watching diff's with vimagit or fugitive it's bugged Hello I am using nord-vim with Hyper terminal (also nord themed), and everything is great - I love it! But I just started using vimagit (tried fugitive too) and when watching diff's the highlight is green/red and so is the text, so can't actually see what's changed. Can someone help me fix this is the core or maybe a hotfix at my machine, since the theme is unusable together with those 2 plugins. Here's a screenshot of the actual deal: Hi @sebastianks and @tomgeorge :wave: Thanks for your patience, I've been too busy with my "real" life and the development of the official Nord website to work on the massive amount of issues/PRs that were submitted each day for all of Nord's port projects. @arcticicestudio I think mine was due to not using Nord in Hyper as I thought I did. There's no issue for me when using Hyper+Nord and Vim+Nord. @tomgeorge If the solution I suggested above doesn't work please open a new issue. This way it makes it easier for me to keep track of all the currently open issues.
gharchive/issue
2018-08-30T12:24:08
2025-04-01T04:33:31.430546
{ "authors": [ "arcticicestudio", "sebastianks" ], "repo": "arcticicestudio/nord-vim", "url": "https://github.com/arcticicestudio/nord-vim/issues/136", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
412724709
Do not highlight function arguments in Python I feel like arguments to functions should not be colored (especially with the same color as the function itself!), as this can lead to confusion. For example, here's some code to demonstrate the issue: At a glance, it is unclear that prev_x is a variable while babylon() is a function. I have found the fix to the problem... removing "source.python meta.function-call" from line 1004 in /themes/nord.json. If you would like I can submit a PR sometime this weekend. @Al2Me6 Sorry for the long stale state. Thanks again for your report and fix :+1: It works fine so I've created PR #113 to get this into the next release version.
gharchive/issue
2019-02-21T02:54:18
2025-04-01T04:33:31.433117
{ "authors": [ "Al2Me6", "arcticicestudio" ], "repo": "arcticicestudio/nord-visual-studio-code", "url": "https://github.com/arcticicestudio/nord-visual-studio-code/issues/108", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2699689833
feat: Add option InjectAsBase64 for ARM Unfortunately, I don't know how to run the test locally. Also don't know how to update the version number. Hi Martin, great thank you for your addition and work on Arcus. Couple of things: [ ] Please read the contribution guide to answer already couple of your questions; [ ] Please revert the changes in the v1.3.1 folder, it is not up to contributors to determine when we release. If changes are needed to the docs, please change them in the /preview folder; [ ] After that, we can start reviewing. Thx again! I cannot run the tests locally for some reason: PS C:\Users\twinp> Invoke-Pester C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Arcus.Scripting.ARM.tests.ps1 WARNING: The names of some imported commands from the module 'Arcus.Scripting.ARM' include unapproved verbs that might make them less discoverable. To find the commands with unapproved verbs, run the Import-Module command again with the Verbose parameter. For a list of approved verbs, type Get-Verb. Describing Arcus ARM integration tests Context ARM injection Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-inline-value.json' for injection tokens Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-inline.json' for injection tokens Finished script 'Inject-ArmContent.ps1' on path 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-inline.json' [-] Replaces file path with inline file contents 773ms RuntimeException: '-Be' is not a valid Should operator. at Get-TestResult, C:\Program Files\WindowsPowerShell\Modules\Pester\3.4.0\Functions\Assertions\Should.ps1: line 42 at , C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Arcus.Scripting.ARM.tests.ps1: line 18 Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-object-value (windows).json' for injection tokens I cannot run the tests locally for some reason: RuntimeException: '-Be' is not a valid Should operator. at Get-TestResult, C:\Program Files\WindowsPowerShell\Modules\Pester\3.4.0\Functions\Assertions\Should.ps1: line 42 at , C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Arcus.Scripting.ARM.tests.ps1: line 18 This shows a conflicting local PS Pester package version as there were breaking changes regarding the -Be operator in some of their versions. /az run CI - Arcus.Scripting I was able to run the test locally after upgrading Pester to 5.6.1 /azp run CI - Arcus.Scripting You can delete this PR. I'm giving it up. Takes me too much time. Test is succesful locally, but the same test fails in the pipeline. Will find another solution. PS C:\Windows\System32> Invoke-Pester C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Arcus.Scripting.ARM.tests.ps1 Starting discovery in 1 files. WARNING: The names of some imported commands from the module 'Arcus.Scripting.ARM' include unapproved verbs that might make them less discoverable. To find the commands with unapproved verbs, run the Import-Module command again with the Verbose parameter. For a list of approved verbs, type Get-Verb. Discovery found 5 tests in 1.99s. Running tests. Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-inline-value.json' for injection tokens Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-inline.json' for injection tokens Finished script 'Inject-ArmContent.ps1' on path 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-inline.json' Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-object-value (windows).json' for injection tokens WARNING: Content to inject into ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-object (windows).json' cannot be parsed as a JSON object! Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-object (windows).json' for injection tokens Finished script 'Inject-ArmContent.ps1' on path 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-object (windows).json' Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-object-value (windows).json' for injection tokens WARNING: Content to inject into ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-object-absolutepath (windows).json' cannot be parsed as a JSON object! Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-object-absolutepath (windows).json' for injection tokens Finished script 'Inject-ArmContent.ps1' on path 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-object-absolutepath (windows).json' Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-escape-value.xml' for injection tokens Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-escape.json' for injection tokens Finished script 'Inject-ArmContent.ps1' on path 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-escape.json' Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\codit.eu.crt' for injection tokens Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-certificate.json' for injection tokens Finished script 'Inject-ArmContent.ps1' on path 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-certificate.json' [+] C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Arcus.Scripting.ARM.tests.ps1 7.62s (2.09s|3.77s) Tests completed in 7.78s Tests Passed: 5, Failed: 0, Skipped: 0, Inconclusive: 0, NotRun: 0 You can delete this PR. I'm giving it up. Takes me too much time. Test is succesful locally, but the same test fails in the pipeline. Will find another solution. It's probably a different test remotely, as it succeeds remotely on windows but fails on linux (and I'm guessting that you work with a windows system locally). Putting @pim-simons in CC. /azp run CI - Arcus.Scripting Turns out base64 encoding on Linux returns a bit of a different string than on Windows. Decoding both base64 strings from Linux and Windows returns the same correct file so I dont really think it is an issue. Created seperate Windows and Linux tests and all working fine now 👍🏻 Thanks Stijn and Pim for finding and fixing this. Thanks Stijn and Pim for finding and fixing this. Thanks for contributing! 🥇 @twinpiloot PR doesn't have to be closed, will wait for @stijnmoreels to review and then we can squash and merge to main 👍🏻 Its the public certificate of the codit website. Just used for the test Verzonden vanaf mijn Galaxy -------- Oorspronkelijk bericht -------- Van: Stijn Moreels @.> Datum: 29-11-2024 21:19 (GMT+01:00) Aan: "arcus-azure/arcus.scripting" @.> Cc: Martin Peters @.>, Mention @.> Onderwerp: Re: [arcus-azure/arcus.scripting] feat: Add option InjectAsBase64 for ARM (PR #438) @stijnmoreels approved this pull request. Fine for me as well. On src/Arcus.Scripting.Tests.Integration/Files/codit.eu.crthttps://github.com/arcus-azure/arcus.scripting/pull/438#discussion_r1863909719: Is this some self-signed public certificate? — Reply to this email directly, view it on GitHubhttps://github.com/arcus-azure/arcus.scripting/pull/438#pullrequestreview-2470477816, or unsubscribehttps://github.com/notifications/unsubscribe-auth/ABRMHWDUD5E4APGBABE3IZ32DDD5HAVCNFSM6AAAAABSTUHFKOVHI2DSMVQWIX3LMV43YUDVNRWFEZLROVSXG5CSMV3GSZLXHMZDINZQGQ3TOOBRGY. You are receiving this because you were mentioned.Message ID: @.***>
gharchive/pull-request
2024-11-27T20:10:07
2025-04-01T04:33:31.458133
{ "authors": [ "pim-simons", "stijnmoreels", "twinpiloot" ], "repo": "arcus-azure/arcus.scripting", "url": "https://github.com/arcus-azure/arcus.scripting/pull/438", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2757885457
fix(server): add waitUntil to the server context in Fastify integration While calling handleNodeRequest or handleNodeRequestAndResponse, waitUntil is not added automatically as in requestListener for Node.js integration. This change adds waitUntil into the serverContext if not present. Fixes the issue with Fastify integration that uses the mentioned methods @ardatan do we now know that all other integrations supports this and we have similar tests for all of the other integrations?
gharchive/pull-request
2024-12-24T14:41:49
2025-04-01T04:33:31.461807
{ "authors": [ "Urigo", "ardatan" ], "repo": "ardatan/whatwg-node", "url": "https://github.com/ardatan/whatwg-node/pull/1926", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1710102150
Error saat ingin convert Inferencing bang pas saya convert error terus Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/gradio/routes.py", line 414, in run_predict output = await app.get_blocks().process_api( File "/usr/local/lib/python3.10/dist-packages/gradio/blocks.py", line 1323, in process_api data = self.postprocess_data(fn_index, result["prediction"], state) File "/usr/local/lib/python3.10/dist-packages/gradio/blocks.py", line 1257, in postprocess_data prediction_value = block.postprocess(prediction_value) File "/usr/local/lib/python3.10/dist-packages/gradio/components.py", line 2531, in postprocess processing_utils.audio_to_file(sample_rate, data, file.name) File "/usr/local/lib/python3.10/dist-packages/gradio/processing_utils.py", line 163, in audio_to_file data = convert_to_16_bit_wav(data) File "/usr/local/lib/python3.10/dist-packages/gradio/processing_utils.py", line 177, in convert_to_16_bit_wav if data.dtype in [np.float64, np.float32, np.float16]: AttributeError: 'NoneType' object has no attribute 'dtype' Traceback (most recent call last): File "/content/Retrieval-based-Voice-Conversion-WebUI/infer-web.py", line 146, in vc_single if_f0 = cpt.get("f0", 1) NameError: name 'cpt' is not defined Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/gradio/routes.py", line 414, in run_predict output = await app.get_blocks().process_api( File "/usr/local/lib/python3.10/dist-packages/gradio/blocks.py", line 1323, in process_api data = self.postprocess_data(fn_index, result["prediction"], state) File "/usr/local/lib/python3.10/dist-packages/gradio/blocks.py", line 1257, in postprocess_data prediction_value = block.postprocess(prediction_value) File "/usr/local/lib/python3.10/dist-packages/gradio/components.py", line 2531, in postprocess processing_utils.audio_to_file(sample_rate, data, file.name) File "/usr/local/lib/python3.10/dist-packages/gradio/processing_utils.py", line 163, in audio_to_file data = convert_to_16_bit_wav(data) File "/usr/local/lib/python3.10/dist-packages/gradio/processing_utils.py", line 177, in convert_to_16_bit_wav if data.dtype in [np.float64, np.float32, np.float16]: AttributeError: 'NoneType' object has no attribute 'dtype' Kirim ssnya Itu inferencing timbrenya masih kosong Lah iya ga nyadar wkwkwk, langsung saya close aja
gharchive/issue
2023-05-15T13:25:51
2025-04-01T04:33:31.465014
{ "authors": [ "Reno121", "ardha27" ], "repo": "ardha27/AI-Song-Cover-RVC", "url": "https://github.com/ardha27/AI-Song-Cover-RVC/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1800619148
Wire: fix setClock clock source is fixed for RA4M1, so the rates will be wrong for C33. This will be fixed in the next commit @facchinm Have a question on the settings that you are using for 400khz. If I attach a Logic analyzer to SDA/SCL while running a sketch pulling data from a MPU-9250 I am seeing: which I believe is showing a frequency of about 257khz not close to the 400khz. However if I change your settings to: m_i2c_extend.clock_settings.brl_value = 12; m_i2c_extend.clock_settings.brh_value = 11; m_i2c_extend.clock_settings.cks_value = 0; I get closer to 400khz (357khz): @facchinm These settings work better - gives 400khz with 60% duty cycle case I2C_MASTER_RATE_FAST: m_i2c_extend.clock_settings.brl_value = 16; m_i2c_extend.clock_settings.brh_value = 15; m_i2c_extend.clock_settings.cks_value = 0; setClock should not only allow the three values, 100kHz, 400kHz, 1Mhz. See this issue. @mjs513 @greiman I pushed a new version of the patch which should take care of the clocks for both UNO and C33. In this patch I explicitly disabled FAST_MODE_PLUS for platforms not supporting it (eg. RA4M1) so we are in the specifications of the chip. Let me know if it works for you, then we can merge it safely. @facchinm Got some strange things happening when testing the changes at 400khz. Using a 5v liquidCrystal_i2c device on Wire with your current settings of 16/15 I am seeing 500khz using a LA: If I change the settings for 400khz to 22/21 I am seeing 400khz, again for 5v device on Wire: But if I use those settings 22/21 for a MPU-9250 running at 3.3v on Wire I am seeing something like 257khz Using your settings, again on wire for a MPU-9250 I am getting 294khz: Confused!!!!!! And if I move the 9250 over to Wire1 using your settings of 16/15 I am seeing that 526khz again and if I use the 22/21 settings I am seeing 416khz at 52% duty cycle. So now I am confused. Please note I do have the Micros propose PR incorporated as well as a couple of others. @facchinm out of curiosity I pulled out a T4.1 (had it handy) and check 100khz and 400khz. At 100khz seeing 68khz with 50% while at 400khz I am seeing 357khz clock with 50% duty cycle for my MPU-9250. Never one to leave enough alone: Adafruit Metro M0 express: 97khz with 50% duty at 100Khz; 357khz with 50% duty at 400khz Uno R3 Clone: 370khz with 52% duty at 400khz; 100khz with 49% duty at 100khz setclock. Wire1 has level shifters and works correctly for 3V3 devices. Wire is marginal for 3V3 devices. For 3V3 devices on Wire, the speed depends on pull-up resistance, pull-up voltage, and VCC for the RA4M1. See this. @facchinm - just wanted to know if there was anything else you wanted me to try with this? @facchinm - maybe its a good idea to go ahead and incorporate the current fix (this pr) into the core. Seems to be resolve some issues people are having: https://forum.arduino.cc/t/i2c-setclock-does-not-appear-to-be-standard-way-to-setclock/1147122/19 As @facchinm is currently out-of-office I'll move ahead and merge this one.
gharchive/pull-request
2023-07-12T09:59:39
2025-04-01T04:33:31.526222
{ "authors": [ "aentinger", "facchinm", "greiman", "mjs513" ], "repo": "arduino/ArduinoCore-renesas", "url": "https://github.com/arduino/ArduinoCore-renesas/pull/46", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1857665184
🛑 U3A Raspberry Pis is down In 183cf9e, U3A Raspberry Pis (http://u3a.epizy.com/rpis) was down: HTTP code: 0 Response time: 0 ms Resolved: U3A Raspberry Pis is back up in 3782af6 after 750 days, 22 hours, 25 minutes.
gharchive/issue
2023-08-19T10:26:09
2025-04-01T04:33:31.550310
{ "authors": [ "ardus-uk" ], "repo": "ardus-uk/upptime", "url": "https://github.com/ardus-uk/upptime/issues/297", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
346832380
An error occurred when I ran the command 'php artisan twill:setup' Hi @linqiuyu! Thanks for reporting. We are aware of this one and have a fix coming up but you don't have to worry about it. This is old code trying to locate a folder that does not exist anymore. As long as the rest of the setup worked, you're good! Hi everybody! Good job @ifox I have save same problem today. Thanks!
gharchive/issue
2018-08-02T02:06:46
2025-04-01T04:33:31.554941
{ "authors": [ "dehboris", "ifox", "linqiuyu" ], "repo": "area17/twill", "url": "https://github.com/area17/twill/issues/25", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2725061306
Question: potential bug while testing AREG_BUILD_TESTS Hi @aregtech I am not familiar with cmake, but there seems to be a bug in conf/cmake/setup.cmake while testing AREG_BUILD_TESTS if (NOT AREG_BUILD_TESTS) set(AREG_GTEST_PACKAGE OFF) endif() Although, I could't find any information in the main README how to build/run/disable unittest BR /biwiki Hi @biwiki, It is not a bug. It says that if "build of unit tests are disabled, to not use GTest package". It plays no real role, simply set it, because at the end of configuration it types a report. Previously, it unit tests were disabled, it was typing "Unit tests == OFF, use GTest package == ON", which I found a little confusing. Regarding disabling building Unit Tests, actually, there is a documentation CMake Configuration Options for Building AREG SDK. But you are right, such simple things like "disable unit tests" or "disable examples" are simple and they should be clearly described. Thank you for the note, I'll keep in mind and update documents. To keep simple, Disable unit test build:cmake -B ./build -DAREG_BUILD_TESTS=OFF Disable examples:cmake -B ./build -DAREG_BUILD_EXAMPLES=OFF I keep this ticket opened to update documents. Thamks @aregtech , that would be so helpful! @biwiki , I appreciate very much if you report uncomfortable / confusing codes or descriptions. I like your notes. They are reasonable. Thanks @aregtech . Basically I knew how to build and run the tests, but it is a bit more convenience to have instructions on the main README since I have seen that quite often in other repos. @biwiki, this is what I mean. You notice the details I might miss. Even the most experienced developers can overlook certain use cases, which might lead to bugs -- it’s just part of the process. Your experience with frameworks and repositories really adds value here, and your feedback is helpful. My goal is to make this framework as useful and user-friendly as possible, so your insights are truly appreciated 🤝 Good point, this can be added in the HOWTO.md document. Or do you think i should create a separate document like "Tests and examples"? To me, it should be in a separate doco but I am not sure if that would be the best
gharchive/issue
2024-12-08T09:01:44
2025-04-01T04:33:31.566855
{ "authors": [ "aregtech", "biwiki" ], "repo": "aregtech/areg-sdk", "url": "https://github.com/aregtech/areg-sdk/issues/466", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1307144511
Gabe's review of the checklist Hi Gabe, this is intended as a practice PR to get you familiar with the writing checklist and part of the code review process. What we're looking for you to do is, after completing the MOOC writing course, go over the existing writing checklist and see if there's anything to add, clarify, or change about it. If you have any questions @ mention me in this issue, and we can work on a PR or hash it out. This issue can be closed when you confirm that you've gone through the writing checklist and a PR has been merged to add anything you think should be added, or you sign off saying it's all good! I have a suggestion for a correction that you might pick up when you read through. If you click Hidden Suggestion, you'll see it. Hidden Suggestion The line, "Abbreviations including et (e.g., "et al.") should not have a period after "et" because it is a whole word," doesn't really make sense. If we're saying "et" isn't actually an abbreviation, this should be rewritten!
gharchive/issue
2022-07-17T16:21:34
2025-04-01T04:33:31.775519
{ "authors": [ "nsryan2" ], "repo": "arfc/arfc.github.io", "url": "https://github.com/arfc/arfc.github.io/issues/264", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
1601162033
Documentation update: tutorial for text classification models comparison Description Context: https://github.com/argilla-io/argilla/issues/2068 In this PR a new tutorial is added: model-comparison for text classification. It is the follow up of the work done during PyConFr in Bordeaux. Closes #2068 Type of change (Please delete options that are not relevant. Remember to title the PR according to the type of change) [ ] Bug fix (non-breaking change which fixes an issue) [ ] New feature (non-breaking change which adds functionality) [ ] Breaking change (fix or feature that would cause existing functionality to not work as expected) [ ] Refactor (change restructuring the codebase without changing functionality) [ ] Improvement (change adding some improvement to an existing functionality) [x] Documentation update How Has This Been Tested (Please describe the tests that you ran to verify your changes. And ideally, reference tests) [ ] Test A [ ] Test B Checklist [ ] I have merged the original branch into my forked branch [ ] I added relevant documentation [ ] follows the style guidelines of this project [ ] I did a self-review of my code [ ] I made corresponding changes to the documentation [ ] My changes generate no new warnings [ ] I have added tests that prove my fix is effective or that my feature works This PR is a WIP, because I haven't figured out how to add the notebook in docs/source/tutorials Hi @embonhomme this is super cool and useful!! In order to make it even more useful, would it be possible to use SetFit's zeroshot model instead of the fewshot classy? We've just published a tutorial to show how easy is to use SetFit and many people are asking about comparisons with the zeroshot HF pipeline, so this tutorial would be awesome and make for a better comparison? https://docs.argilla.io/en/latest/tutorials/notebooks/labelling-textclassification-setfit-zeroshot.html#🔫-Zero-shot-predictions-with-SetFit We'd be happy to walk you through if you have questions. Hello @dvsrepo :) Thank you for the feedback, you can find in the new commit the comparaison with SetFit zero-shot. Tell me if it is relevant. This is looking just perfect! The only one remaining change would be to review the remaining mentions of few-shot and classy-classification and replace them with zero-shot and SetFit. Then we are good to go! We'd love to share this next week via LinkedIn and Twitter, if you'd like us to mention you as the author, send me an email to daniel @ argilla.io Thank you! Yes sorry I totally forgot to change the description part. It should be better now. Also here I have just added a Jupyter Notebook, I didn't figure out how it works with the modal.md, dvc.md,... I will send you an email with my LinkedIn :) Great stuff @embonhomme! @embonhomme, in case you can tackle this, the process is: Create a folder here with the filename, like this one: https://github.com/argilla-io/argilla/tree/develop/docs/_source/_static/tutorials/training-textclassification-setfit-fewshot Add the reference to the tutorial: here https://github.com/argilla-io/argilla/blob/develop/docs/_source/tutorials/libraries/setfit.md, here: https://github.com/argilla-io/argilla/blob/develop/docs/_source/tutorials/steps/4_monitoring.md , here: https://github.com/argilla-io/argilla/blob/develop/docs/_source/tutorials/tasks/text_classification.md and here: https://github.com/argilla-io/argilla/blob/develop/docs/_source/tutorials/techniques/few_shot.md I would say this tutorial is about Monitoring, TextClassification, few-shot Otherwise, let us know and @davidberenstein1957 might be able to help Thank you, I did the integration :) Hi @davidberenstein1957 I renamed everything :) Yes, I would like to participate in the LinkedIn shoutout! Lovely!
gharchive/pull-request
2023-02-27T13:11:48
2025-04-01T04:33:31.823678
{ "authors": [ "davidberenstein1957", "dvsrepo", "embonhomme" ], "repo": "argilla-io/argilla", "url": "https://github.com/argilla-io/argilla/pull/2426", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1167517517
chore: Bump Kubernetes dependencies to version 1.22.4 Signed-off-by: jannfis jann@mistrust.net Codecov Report Merging #394 (383ebe3) into master (f2eb28b) will not change coverage. The diff coverage is n/a. @@ Coverage Diff @@ ## master #394 +/- ## ======================================= Coverage 65.60% 65.60% ======================================= Files 21 21 Lines 2035 2035 ======================================= Hits 1335 1335 Misses 571 571 Partials 129 129 :mega: Codecov can now indicate which changes are the most critical in Pull Requests. Learn more
gharchive/pull-request
2022-03-13T08:36:32
2025-04-01T04:33:31.833531
{ "authors": [ "codecov-commenter", "jannfis" ], "repo": "argoproj-labs/argocd-image-updater", "url": "https://github.com/argoproj-labs/argocd-image-updater/pull/394", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2165601173
Support defining custom cluster roles for cluster scoped instances Is your feature request related to a problem? Please describe. When using a cluster scoped instance of Argo CD the operator will create a separate and hard coded set of clusterroles and clusterrolebindings. Users that want to lock down the permissions of their instances need the ability to use an alternate set of clusterroles. As an example, the new feature in Argo CD 2.10 to auto-respect RBAC (https://argo-cd.readthedocs.io/en/stable/operator-manual/declarative-setup/#auto-respect-rbac-for-controller) enables users to remove the cluster-wide view all permissions required by a cluster scoped instance and provide a much more tailored set of permissions. However the Argo CD Operator hard codes this cluster role to define view all permissions. Describe the solution you'd like Enable users to specify their own roles using the environment variables CONTROLLER_CLUSTER_SCOPE_ROLE and SERVER_CLUSTER_SCOPE_ROLE Describe alternatives you've considered Updates to the cluster role created by the operator get overridden by the operator so it's not a suitable approach. PR is in progress for this. Reopening since PR was reverted and there is a new PR raised to address this.
gharchive/issue
2024-03-03T22:35:58
2025-04-01T04:33:31.836738
{ "authors": [ "gnunn1", "jparsai" ], "repo": "argoproj-labs/argocd-operator", "url": "https://github.com/argoproj-labs/argocd-operator/issues/1275", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1139156120
chore: add missing bundle versions Signed-off-by: John Pitman jpitman@redhat.com What type of PR is this? /kind chore What does this PR do / why we need it: After releasing 0.2.0 of the argocd-operator, we didn't set up the master branch for the next version. That means on master, any changes to the CRD have gone into the 0.2.0 bundle folder. Since we've already split off the branch for 0.3.0, the next version should be 0.4.0. This PR is to fix this situation Have you updated the necessary documentation? [x] Documentation update is required by this PR. [x] Documentation has been updated. Which issue(s) this PR fixes: Had a call/discussion with @jopit on why the operator version is updated as 0.4.0 instead of 0.3.0. Since we already branched out 0.3.0(release-0.3) and we know some changes that are in master are not going into release-0.3, it is better to update the operator version to 0.4.0 in master. This way when we create the next release branch we dont have to update the operator version and manifests in the release branch. However, I see an overhead of merging back the 0.3.0 bundle folder in release-0.3 branch back into master after/during v0.3.0 release. @jopit and I have decided that we shall merge this PR for now, if not the 0.2.0 bundle folder in the master would keep updating. we can decide about this overhead in the next cabal, if there is a change in the decision we can act accordingly with a new PR if required.
gharchive/pull-request
2022-02-15T19:59:31
2025-04-01T04:33:31.842318
{ "authors": [ "iam-veeramalla", "jopit" ], "repo": "argoproj-labs/argocd-operator", "url": "https://github.com/argoproj-labs/argocd-operator/pull/563", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2050918782
quick start fails out of the box Pre-requisites [X] I have double-checked my configuration [X] I can confirm the issues exists when I tested with :latest [ ] I'd like to contribute the fix myself (see contributing guide) What happened/what did you expect to happen? Able to run the quick start CLI with success. My environment was a macOS M2 machine running Docker Desktop with kind. The steps are straight from https://argoproj.github.io/argo-workflows/quick-start/ kind create cluster # cluster created with 1.27.3 kubectl create namespace argo kubectl apply -n argo -f https://github.com/argoproj/argo-workflows/releases/download/v3.5.2/install.yaml kubectl patch deployment \ argo-server \ --namespace argo \ --type='json' \ -p='[{"op": "replace", "path": "/spec/template/spec/containers/0/args", "value": [ "server", "--auth-mode=server" ]}]' At this point I ran: argo submit -n argo --watch https://raw.githubusercontent.com/argoproj/argo-workflows/main/examples/hello-world.yaml And the output was: Name: hello-world-g2csp Namespace: argo ServiceAccount: unset (will run with the default ServiceAccount) Status: Error Message: Error (exit code 1): pods "hello-world-g2csp" is forbidden: User "system:serviceaccount:argo:default" cannot patch resource "pods" in API group "" in the namespace "argo" Conditions: PodRunning False Completed True Created: Wed Dec 20 10:25:35 -0600 (9 minutes ago) Started: Wed Dec 20 10:25:35 -0600 (9 minutes ago) Finished: Wed Dec 20 10:26:05 -0600 (9 minutes ago) Duration: 30 seconds Progress: 0/1 ResourcesDuration: 13s*(1 cpu),13s*(100Mi memory) STEP TEMPLATE PODNAME DURATION MESSAGE ⚠ hello-world-g2csp whalesay hello-world-g2csp 21s Error (exit code 1): pods "hello-world-g2csp" is forbidden: User "system:serviceaccount:argo:default" cannot patch resource "pods" in API group "" in the namespace "argo" Version v3.5.2 Paste a small workflow that reproduces the issue. We must be able to run the workflow; don't enter a workflows that uses private images. apiVersion: argoproj.io/v1alpha1 kind: Workflow metadata: generateName: hello-world- labels: workflows.argoproj.io/archive-strategy: "false" annotations: workflows.argoproj.io/description: | This is a simple hello world example. spec: entrypoint: whalesay templates: - name: whalesay container: image: docker/whalesay:latest command: [cowsay] args: ["hello world"] Logs from the workflow controller time="2023-12-20T16:25:35.731Z" level=info msg="Processing workflow" namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:25:35.736Z" level=info msg="Task-result reconciliation" namespace=argo numObjs=0 workflow=hello-world-g2csp time="2023-12-20T16:25:35.736Z" level=info msg="Updated phase -> Running" namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:25:35.736Z" level=warning msg="Node was nil, will be initialized as type Skipped" namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:25:35.736Z" level=info msg="was unable to obtain node for , letting display name to be nodeName" namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:25:35.736Z" level=info msg="Pod node hello-world-g2csp initialized Pending" namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:25:35.745Z" level=info msg="Created pod: hello-world-g2csp (hello-world-g2csp)" namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:25:35.746Z" level=info msg="TaskSet Reconciliation" namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:25:35.746Z" level=info msg=reconcileAgentPod namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:25:35.752Z" level=info msg="Workflow update successful" namespace=argo phase=Running resourceVersion=2676 workflow=hello-world-g2csp time="2023-12-20T16:25:45.751Z" level=info msg="Processing workflow" namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:25:45.752Z" level=info msg="Task-result reconciliation" namespace=argo numObjs=0 workflow=hello-world-g2csp time="2023-12-20T16:25:45.752Z" level=info msg="node changed" namespace=argo new.message=PodInitializing new.phase=Pending new.progress=0/1 nodeID=hello-world-g2csp old.message= old.phase=Pending old.progress=0/1 workflow=hello-world-g2csp time="2023-12-20T16:25:45.752Z" level=info msg="TaskSet Reconciliation" namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:25:45.752Z" level=info msg=reconcileAgentPod namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:25:45.765Z" level=info msg="Workflow update successful" namespace=argo phase=Running resourceVersion=2703 workflow=hello-world-g2csp time="2023-12-20T16:25:55.768Z" level=info msg="Processing workflow" namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:25:55.768Z" level=info msg="Task-result reconciliation" namespace=argo numObjs=0 workflow=hello-world-g2csp time="2023-12-20T16:25:55.768Z" level=info msg="node unchanged" namespace=argo nodeID=hello-world-g2csp workflow=hello-world-g2csp time="2023-12-20T16:25:55.769Z" level=info msg="TaskSet Reconciliation" namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:25:55.769Z" level=info msg=reconcileAgentPod namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:26:05.933Z" level=info msg="Processing workflow" namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:26:05.933Z" level=info msg="Task-result reconciliation" namespace=argo numObjs=0 workflow=hello-world-g2csp time="2023-12-20T16:26:05.934Z" level=info msg="Pod failed: Error (exit code 1): pods \"hello-world-g2csp\" is forbidden: User \"system:serviceaccount:argo:default\" cannot patch resource \"pods\" in API group \"\" in the namespace \"argo\"" displayName=hello-world-g2csp namespace=argo pod=hello-world-g2csp templateName=whalesay workflow=hello-world-g2csp time="2023-12-20T16:26:05.934Z" level=info msg="node changed" namespace=argo new.message="Error (exit code 1): pods \"hello-world-g2csp\" is forbidden: User \"system:serviceaccount:argo:default\" cannot patch resource \"pods\" in API group \"\" in the namespace \"argo\"" new.phase=Error new.progress=0/1 nodeID=hello-world-g2csp old.message=PodInitializing old.phase=Pending old.progress=0/1 workflow=hello-world-g2csp time="2023-12-20T16:26:05.934Z" level=info msg="TaskSet Reconciliation" namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:26:05.934Z" level=info msg=reconcileAgentPod namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:26:05.934Z" level=info msg="Updated phase Running -> Error" namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:26:05.934Z" level=info msg="Updated message -> Error (exit code 1): pods \"hello-world-g2csp\" is forbidden: User \"system:serviceaccount:argo:default\" cannot patch resource \"pods\" in API group \"\" in the namespace \"argo\"" namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:26:05.934Z" level=info msg="Marking workflow completed" namespace=argo workflow=hello-world-g2csp time="2023-12-20T16:26:05.941Z" level=info msg="cleaning up pod" action=deletePod key=argo/hello-world-g2csp-1340600742-agent/deletePod time="2023-12-20T16:26:05.951Z" level=info msg="Workflow update successful" namespace=argo phase=Error resourceVersion=2745 workflow=hello-world-g2csp time="2023-12-20T16:26:05.959Z" level=info msg="cleaning up pod" action=labelPodCompleted key=argo/hello-world-g2csp/labelPodCompleted Logs from in your workflow's wait container time="2023-12-20T16:25:56.956Z" level=info msg="No output parameters" time="2023-12-20T16:25:56.956Z" level=info msg="No output artifacts" time="2023-12-20T16:25:56.967Z" level=warning msg="failed to patch task set, falling back to legacy/insecure pod patch, see https://argoproj.github.io/argo-workflows/workflow-rbac/" error="workflowtaskresults.argoproj.io is forbidden: User \"system:serviceaccount:argo:default\" cannot create resource \"workflowtaskresults\" in API group \"argoproj.io\" in the namespace \"argo\"" time="2023-12-20T16:25:56.968Z" level=warning msg="Non-transient error: pods \"hello-world-g2csp\" is forbidden: User \"system:serviceaccount:argo:default\" cannot patch resource \"pods\" in API group \"\" in the namespace \"argo\"" time="2023-12-20T16:25:56.968Z" level=error msg="executor error: pods \"hello-world-g2csp\" is forbidden: User \"system:serviceaccount:argo:default\" cannot patch resource \"pods\" in API group \"\" in the namespace \"argo\"" time="2023-12-20T16:25:56.968Z" level=info msg="Alloc=7218 TotalAlloc=13045 Sys=24677 NumGC=4 Goroutines=8" time="2023-12-20T16:25:56.969Z" level=warning msg="failed to patch task set, falling back to legacy/insecure pod patch, see https://argoproj.github.io/argo-workflows/workflow-rbac/" error="workflowtaskresults.argoproj.io \"hello-world-g2csp\" is forbidden: User \"system:serviceaccount:argo:default\" cannot patch resource \"workflowtaskresults\" in API group \"argoproj.io\" in the namespace \"argo\"" time="2023-12-20T16:25:56.969Z" level=warning msg="Non-transient error: pods \"hello-world-g2csp\" is forbidden: User \"system:serviceaccount:argo:default\" cannot patch resource \"pods\" in API group \"\" in the namespace \"argo\"" time="2023-12-20T16:25:56.969Z" level=error msg="executor error: pods \"hello-world-g2csp\" is forbidden: User \"system:serviceaccount:argo:default\" cannot patch resource \"pods\" in API group \"\" in the namespace \"argo\"" time="2023-12-20T16:25:56.970Z" level=fatal msg="pods \"hello-world-g2csp\" is forbidden: User \"system:serviceaccount:argo:default\" cannot patch resource \"pods\" in API group \"\" in the namespace \"argo\"" I failed to mention that adding --serviceaccount argo to the argo submit made it work. Would you like to submit a PR to help improve the docs? Would you like to submit a PR to help improve the docs? Would be glad to. I'm just not sure if that's what the maintainers want from the quick start. If it is then I'll happily update the docs and include a reference to the workflow RBAC for further reading and make a note that when a service account isn't specified then the default account is used which often doesn't have sufficient permissions. Looks like the quick start and walk-through guides all use default service account which does not have the permission. We can probably just give default service-account enough permission https://github.com/argoproj/argo-workflows/tree/76f4d1f9ca53b8ac48965e12c15454e544936c49/manifests/quick-start/base Looks like the quick start and walk-through guides all use default service account which does not have the permission. The default SA actually has a RoleBinding to the executor Role and both are included in the kustomization.yaml so the quick start should work with the default SA already. kubectl apply -n argo -f https://github.com/argoproj/argo-workflows/releases/download/v3.5.2/install.yaml I think the problem is that the Quick Start docs are referencing install.yaml instead of quick-start-minimal.yaml @cardoe would you like to submit a PR to modify the docs to reference quick-start-minimal.yaml instead of install.yaml? Yep will do. I found this issue after stumbling upon the same problem across our workflows when upgrading from v3.5.0 to v3.5.4 using the released manifest files. It seems like this issue was closed after a doc change to the quick start page, but the installation docs and release pages seemingly suggest that one should be able to use install.yaml without issues: To install Argo Workflows, navigate to the releases page and find the release you wish to use (the latest full release is preferred). Scroll down to the Controller and Server section and execute the kubectl commands. It seems like this issue was closed after a doc change to the quick start page, but the installation docs and release pages seemingly suggest that one should be able to use install.yaml without issue Yes, this is specific to the quick start, which explicitly states it should not be used in production. Yes, I see that the quick start page links to the install page: These instructions are intended to help you get started quickly. They are not suitable for production. For production installs, please refer to the installation documentation. :point_up: refers to the release pages which then tells users to run: kubectl create namespace argo kubectl apply -n argo -f https://github.com/argoproj/argo-workflows/releases/download/v3.5.4/install.yaml In any case, perhaps it's user / reader error on my part :man_shrugging: . Given quick-start-minimal.yaml is not intended for production, and install.yaml doesn't work "out of the box" anymore, is there documentation for what users that were previously using install.yaml should do when updating to v3.5.2 and beyond? and install.yaml doesn't work "out of the box" anymore, is there documentation for what users that were previously using install.yaml should do when updating to v3.5.2 and beyond? This wasn't listed as a regression, is there a regression in the manifests that you noticed for v3.5.2 vs an older version? You mention v3.5.0 as your prior, the only changes to the manifests between v3.5.0 and v3.5.2 were #11839 and #11947 Ah, yes, sorry, I wasn't clear about that :grimacing: . So, for better or worse, we have been using the release doc steps to deploy/update argo via install.yaml since v3.3.9 or so. It's always worked fine without issues. However, when updating from v3.5.0 to 3.5.4 workflows began failing with the message cannot patch resource "pods" in API group "". I did the same diff check between the manifests and the only differences I saw were the image tags: - args: - server env: [] - image: quay.io/argoproj/argocli:v3.5.0 + image: quay.io/argoproj/argocli:v3.5.4 name: argo-server ports: - containerPort: 2746 @@ -1323,7 +1323,7 @@ fieldRef: apiVersion: v1 fieldPath: metadata.name - image: quay.io/argoproj/workflow-controller:v3.5.0 + image: quay.io/argoproj/workflow-controller:v3.5.4 livenessProbe: failureThreshold: 3 httpGet: That's why we were so surprised to hit the problem. For now we've downgraded to the https://github.com/argoproj/argo-workflows/releases/download/v3.5.1/install.yaml manifest :man_shrugging: . Any thoughts / suggestions would be very much appreciated :pray: . Thanks, @agilgur5! I will look more into the Workflow RBAC to see what changes might be needed for our workflows to work with v3.5.4. Just to help clarify my understanding, is the reason why we're seeing existing workflows work when downgrading just the version tag to v3.5.1 in the manifest because we've been historically lucking into some default behavior that no longer applies starting v3.5.2, and the newer versions are simply surfacing a latent RBAC issue on our end? For example, it's why the following works fine: $ minikube start $ kubectl create namespace argo $ kubectl apply -n argo -f https://github.com/argoproj/argo-workflows/releases/download/v3.5.1/install.yaml $ argo submit --watch https://raw.githubusercontent.com/argoproj/argo-workflows/main/examples/hello-world.yaml Output: STEP TEMPLATE PODNAME DURATION MESSAGE ✔ hello-world-45qtp whalesay hello-world-45qtp 18s But the following will fail (as will versions 3.5.2 and 3.5.3): $ minikube start $ kubectl create namespace argo $ kubectl apply -n argo -f https://github.com/argoproj/argo-workflows/releases/download/v3.5.4/install.yaml $ argo submit --watch https://raw.githubusercontent.com/argoproj/argo-workflows/main/examples/hello-world.yaml Output: STEP TEMPLATE PODNAME DURATION MESSAGE ⚠ hello-world-jdsk6 whalesay hello-world-jdsk6 22s Error (exit code 1): pods "hello-world-jdsk6" is forbidden: User "system:serviceaccount:default:default" cannot patch resource "pods" in API group "" in the namespace "default" For example, it's why the following works fine: 👀 That's an interesting reproduction. I thought your prod cluster wasn't fresh, so could be that something else changed. If this is a fresh minikube cluster each time, then that suggests a change in the Executor is requiring that permission when it previously didn't. Or, more specifically, the simple hello-world example managed to work without any additional RBAC, but now that is required. I'd need to look through the Executor diffs for 3.5.2 to see why that might happen. From a real quick glance, this might be from #11947 -- this change to wait.go ensure it properly reports results. If that's the case, it may be that some really short, simple steps were not recorded entirely correctly previously. cc @Garett-MacGowan who probably knows this part of the codebase better than me at this point. I only skimmed this, so im not sure if I'm missing something, but have you configured RBAC according to this? I can take a better look at this tomorrow. I only skimmed this, so im not sure if I'm missing something, but have you configured RBAC according to this? I linked that above -- but apparently it wasn't necessary for simple steps like hello-world prior to 3.5.2. #11947 had some Executor changes that seemed like they might've changed some of that logic. It might've been a long-standing bug that didn't have much user-facing impact that you ended up fixing or maybe there's a new request that isn't necessary for simpler steps? I'm not sure, hence why I tagged you. I can take a better look at this tomorrow. Take your time! @agilgur5 here's my take. TaskResultReconciliation() which uses workflowtaskresult, for which we need RBAC, is now (necessarily) called before if woc.wf.Labels[common.LabelKeyCompleted] == "true" { // abort now, we do not want to perform any more processing on a complete workflow because we could corrupt it Before my PR, if the workflow completed quick enough, reconciliation wouldn't have a chance to run and therefore you would avoid the RBAC issue. This is closely related to #11947 The quickstart docs should probably mention the need for workflowtaskresult RBAC per https://argo-workflows.readthedocs.io/en/latest/workflow-rbac/ Before my PR, if the workflow completed quick enough, reconciliation wouldn't have a chance to run and therefore you would avoid the RBAC issue. Thought that could be the case per above, thanks for confirming Garett! The quickstart docs should probably mention the need for workflowtaskresult RBAC per https://argo-workflows.readthedocs.io/en/latest/workflow-rbac/ The PR that closed this issue, #12445 fixed the quickstart docs to actually use the quick-start-*.yaml at my suggestion above, which has the appropriate permissions. But the Workflow RBAC page should be linked in the production installation page where it is apparently missing. @rnathuji would you like to submit a PR to add a link to the Workflow RBAC page to the production installation page? Perhaps it could be made as the first bullet under "Additional installation considerations" @agilgur5 ahh, yes, you're right. The real reason is here. In this PR, we switched to always creating a task result (a placeholder, if no outputs exist) in order to track reconciliation status. The implementation has changed slightly since this PR, but it's effectively doing the same thing. Thanks for the context! Thought it might be something like that. Re-wording, that would mean that previously, as hello-world had no outputs, it would made no task result and hence technically did not need the RBAC. That led to some race conditions however, so now it does require that. The docs have always stated this, so it was actually an undocumented technicality that simpler steps like hello-world didn't need the RBAC in the past. Thanks for the context! Thought it might be something like that. Re-wording, that would mean that previously, as hello-world had no outputs, it would made no task result and hence technically did not need the RBAC. That led to some race conditions however, so now it does create a task result and therefore require the RBAC. The docs have always stated this, so it was actually an undocumented technicality that simpler steps like hello-world didn't need the RBAC in the past. That's right. @agilgur5 - Sure, I can spin up a doc PR. Thanks all for taking a look and helping to clarify the reason behind the behavioral change that we observed. :tada: See PR in #12680
gharchive/issue
2023-12-20T16:41:12
2025-04-01T04:33:32.210376
{ "authors": [ "Garett-MacGowan", "agilgur5", "cardoe", "rnathuji", "terrytangyuan" ], "repo": "argoproj/argo-workflows", "url": "https://github.com/argoproj/argo-workflows/issues/12391", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1127810528
Support archived workflows logs in UI from archived logs Summary Currently, if you open an archived workflow, you are not able to see logs in the log viewer, even if the logs were archived to 'main.log'. We could fix this. Use Cases When would you use this? Message from the maintainers: Love this enhancement proposal? Give it a 👍. We prioritise the proposals with the most 👍. Hi @alexec , would you mind providing more details on this. I couldn't get my head around what you meant by "Support archived workflows logs in UI from archived logs". All I'm thinking right now is we can view archived log from archived workflows If I understand correctly - the current UI does not make the request to get the logs (output artifacts) by the workflow's UID? because the server's endpoint for it exist. Does this relate to the Log UI "All" option not showing nothing after pods are garbage collected? Or, should that be considered a separate issue? On a side note, what does this recommendation in the docs mean?: ⚠️ We do not recommend you rely on Argo Workflows to archive logs. Instead, use a conventional Kubernetes logging facility. Are there some known issues? Also, does kubernetes provide any facility for collecting and returning logs for garbage collected pods, and then can argo workflows use this facility? I ask because viewing logs in the argo workflows UI is very convenient, and one of the main "selling points" of the application in my mind Also, does kubernetes provide any facility for collecting and returning logs for garbage collected pods, and then can argo workflows use this facility? A pretty conventional cluster set-up is to have fluentd / fluent-bit as a DaemonSet collecting logs from all nodes and sending those a log provider (e.g. Splunk, Logstash, etc). That would be an interesting feature if Workflows could use it, as it is indeed convenient. As a workaround to that, you can add an external link to a query in your log provider. Not sure if #11121 ended up fixing this as well It's supported. Re-open if not.
gharchive/issue
2022-02-08T22:09:16
2025-04-01T04:33:32.217927
{ "authors": [ "agilgur5", "alexec", "asfaltboy", "noam-codefresh", "tczhao", "terrytangyuan" ], "repo": "argoproj/argo-workflows", "url": "https://github.com/argoproj/argo-workflows/issues/7830", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1139938418
Request help, can realize drawing the way to generate workflow? Can Argo UI add workflow by drawing? The flow chart drawn by Activiti is imported into Argo to implement the workflow. This issue template is only for mentoring. Not feature request. Also this is a duplicate of https://github.com/argoproj/argo-workflows/issues/5610 so I am closing this. Asking for help, can you now implement the way drawing generates workflow? This issue template is only for mentoring. Not feature request. Also this is a duplicate of #5610 so I am closing this. Asking for help, can you now implement the way drawing generates workflow? No that's not supported which is why the linked issue is still open. No that's not supported which is why the linked issue is still open. Is there any requirement that can be realized: Flowchart drawing tools such as Activiti and Camunda can be used to design a flowchart and then be connected to or imported into Argo Workflow for execution
gharchive/issue
2022-02-16T12:17:13
2025-04-01T04:33:32.221760
{ "authors": [ "LingClassmate", "LingHuaHui", "terrytangyuan" ], "repo": "argoproj/argo-workflows", "url": "https://github.com/argoproj/argo-workflows/issues/7892", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1174727619
I think my Artifact repository is being interpreted as a volume(?) Checklist [ ] Double-checked my configuration. [ ] Tested using the latest version. [ ] Used the Emissary executor. Summary This is a follow-on to this issue (https://github.com/argoproj/argo-workflows/issues/8168). I have implemented my artifact repository per these instructions (https://argoproj.github.io/argo-workflows/configure-artifact-repository/). Here is my abstracted implementation artifactRepository: | gcs: bucket: my-bucket region: us-east-1 keyFormat: /tmp/ serviceAccountKeySecret: name: account@developer.gserviceaccount.com key: my-64bit encoded-key I expect that the following workflow should save the output of whalesay to the repository so that it is available for the print pod. --- { "apiVersion": "argoproj.io/v1alpha1", "kind": "Workflow", "metadata": { "annotations": { "workflows.argoproj.io/description": "testing a linter", "workflows.argoproj.io/version": ">= 3.1.0" }, "labels": { "workflows.argoproj.io/archive-strategy": "false" }, "name": "hello-world", }, "spec": { "serviceaccount": "725104128071-compute@developer.gserviceaccount.com ", "arguments": { "parameters": [ { "name": "msg", "value": "Hello there" } ] }, "entrypoint": "entrypoint", "securityContext": { "fsGroup": 2000, "fsGroupChangePolicy": "OnRootMismatch", "runAsGroup": 3000, "runAsNonRoot": true, "runAsUser": 1000 }, "templates": [ { "container": { "args": [ "cowsay Hello Test >> /tmp/output" ], "command": [ "sh", "-c" ], "image": "docker/whalesay:latest", "imagePullPolicy": "IfNotPresent", }, "name": "whalesay", "outputs": { "artifacts": [ { "name": "message", "path": "/tmp/output", "s3": { "key": "whalesay", } } ] }, "retryStrategy": { "limit": "10" }, "securityContext": { "fsGroup": 2000, "fsGroupChangePolicy": "OnRootMismatch", "runAsGroup": 3000, "runAsNonRoot": true, "runAsUser": 1000 } }, { "inputs": { "artifacts": [ { "s3": { "key": "whalesay", }, "name": "data", "path": "/tmp/input" } ] }, "name": "print", "retryStrategy": { "limit": "10" }, "script": { "command": [ "python" ], "image": "python:alpine3.6", "imagePullPolicy": "IfNotPresent", "source": "import sys \nsys.stdout.write(\"{{inputs.artifacts.data}}\")\n\n" }, "securityContext": { "fsGroup": 2000, "fsGroupChangePolicy": "OnRootMismatch", "runAsGroup": 3000, "runAsNonRoot": true, "runAsUser": 1000 } }, { "dag": { "tasks": [ { "name": "whalesay", "template": "whalesay" }, { "arguments": { "artifacts": [ { "from": "{{whalesay.outputs.artifacts.message}}", "name": "data" } ] }, "dependencies": [ "whalesay" ], "name": "print", "template": "print" } ] }, "name": "entrypoint" } ] } } ... Instead, I am getting an error which suggests it is expecting a volumeClaim Pod "hello-world2xr6z-2749602831" is invalid: [spec.volumes[1].name: Invalid value: "725104128071-compute@developer.gserviceaccount.com": a lowercase RFC 1123 label must consist of lower case alphanumeric characters or '-', and must start and end with an alphanumeric character (e.g. 'my-name', or '123-abc', regexused for validation is '[a-z0-9]([-a-z0-9]*[a-z0-9])?'), spec.containers[0].volumeMounts[0].name: Not found: "725104128071-compute@developer.gserviceaccount.com", spec.initContainers[0].volumeMounts[0].name: Not found: "725104128071-compute@developer.gserviceaccount.com"] What version are you running? Argo 3.3.0 I am able to successfully run the quickstart example, so I think it is the artifact repository implementation. i.e. Argo itself is working ok. No pods are generated in the case of the problematic workflow (above) Message from the maintainers: Impacted by this bug? Give it a 👍. We prioritise the issues with the most 👍. Hypothesis: you have executor set in your config map. Can you please confirm? Yes. containerRuntimeExecutor: emissary sorry. I meant the specifically data.executor. Here you go! apiVersion: v1 kind: ConfigMap metadata: name: workflow-controller-configmap data: artifactRepository: | gcs: endpoint: storage.googleapis.com bucket: my-bucket region: us-east-1 serviceAccountKeySecret: name: '' key: '' useSDKCreds: false containerRuntimeExecutor: emissary workflowDefaults: | spec: ttlStrategy: secondsAfterSuccess: 5 parallelism: 3 entrypoint: 'entrypoint' securityContext: runAsNonRoot: true runAsUser: 1000 fsGroup: 2000 runAsGroup: 3000 fsGroupChangePolicy: "OnRootMismatch" You're configuration is incorrect. It is impossible for there to be a secret with an @ in the name. Fix this serviceAccountKeySecret: name: name-of-secret key: key-within-secret
gharchive/issue
2022-03-20T23:31:43
2025-04-01T04:33:32.230989
{ "authors": [ "alexec", "wdma" ], "repo": "argoproj/argo-workflows", "url": "https://github.com/argoproj/argo-workflows/issues/8197", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1219428303
Full CRDs are incomplete and causes json2jsii to fail Summary The generated CustomResourceDefinitions seem to be incomplete because of the steps property. https://github.com/argoproj/argo-workflows/blob/ecd91b1c4215a2ab8742f7c43eaade98a1d47eba/manifests/base/crds/full/argoproj.io_workflowtemplates.yaml#L7673-L7676 The items property has a type property but does not have an accompanying items property. This causes json2jsii to fail. https://github.com/cdklabs/json2jsii/blob/60e891447e35a8bb1a9dac3496d6adce87d73292/src/type-generator.ts#L534-L540 This is an issue for tools that depend on json2jsii. https://github.com/cdk8s-team/cdk8s-cli/issues/93 However, the json schema has these missing properties. https://github.com/argoproj/argo-workflows/blob/ecd91b1c4215a2ab8742f7c43eaade98a1d47eba/api/jsonschema/schema.json#L7315-L7321 https://github.com/argoproj/argo-workflows/blob/ecd91b1c4215a2ab8742f7c43eaade98a1d47eba/api/jsonschema/schema.json#L6449-L6454 It looks like this problem was fixed at one point with #4828, but was later reverted by @JPZ13 here #4810 because of the CRDs becoming too long. Is there way we can get the full CRDs for use with json2jsii as I do not believe it has any limit on CRD size like Kubernetes does. I'd like to have all the types for steps so that I can get strong typing with cdk8s. Message from the maintainers: Impacted by this bug? Give it a 👍. We prioritise the issues with the most 👍. Would you like to submit a PR to fix this? I would need some guidance, I spent some time looking into the issue yesterday on a fork https://github.com/adrianord/argo-workflows, but I believe the fix I've done there causes the CRDs to be too large for Kubernetes, which is why it was reverted before. Share you PR and maybe we can take a look at it. Also getting an issue here with crds and json2jsii. Error: unsupported array type undefined at TypeGenerator.typeForArray (/snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:404:19) at TypeGenerator.emitArray (/snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:260:25) at TypeGenerator.emitTypeInternal (/snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:186:29) at TypeGenerator.typeForProperty (/snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:383:21) at TypeGenerator.typeForArray (/snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:406:21) at TypeGenerator.emitArray (/snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:260:25) at TypeGenerator.emitTypeInternal (/snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:186:29) at TypeGenerator.typeForProperty (/snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:383:21) at TypeGenerator.emitProperty (/snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:322:35) at /snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:302:22
gharchive/issue
2022-04-28T23:10:45
2025-04-01T04:33:32.238327
{ "authors": [ "adrianord", "alexec", "sstaley-hioscar" ], "repo": "argoproj/argo-workflows", "url": "https://github.com/argoproj/argo-workflows/issues/8532", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
673147256
Surface K8s Events about Nodes to UI We run argo on GKE with cluster-autoscaler. Most of the time all our workflows trigger a scaleUp event since powerful machines are usually down when not used. Usually, the scale-up takes a few minutes. During this time the workflow is in pending state with a message similar to this: Once the node has been created the workflow starts and all is good! But sometimes the cluster is not able to scale up (often due to an error in the resources and nodeSelector configuration). Admin have access to cluster event logs and we can spot the issue quickly: But most of our staff only use the Argo dashboard and so can only rely on the workflow status message to understand what is going on. And scaling up error reported by the cluster-autoscaler is not propagated to it: pod didn't trigger scale-up (it wouldn't fit if a new node is added): 16 Insufficient memory, 1 node(s) didn't match node selector, 15 Insufficient cpu Would that be possible to report cluster events related to a specific workflow or pods managed by a workflow in the status field? Not sure it's technically possible (because of ServiceAccount and permissions) but I am asking just in case. We watch pods of a workflow to compute state. If this information is available on the pod’s status we could. Are you able to attach the YAML of a pod that we involved? Or is this only available on events? (Please attach the event YAML). Indeed the only thing I see in the pod YAML is what Argo report in the UI: status: conditions: - lastProbeTime: null lastTransitionTime: "2020-08-05T16:53:27Z" message: '0/21 nodes are available: 21 Insufficient cpu.' reason: Unschedulable status: "False" type: PodScheduled phase: Pending qosClass: Burstable and the source is default-scheduler for the message. But the message "pod didn't trigger scale-up (it wouldn't fit if a new node is added): 17 Insufficient cpu" coming from the source cluster-autoscaler is not shown. $ kubectl get event -n argo --field-selector involvedObject.name=wonderful-tiger LAST SEEN TYPE REASON OBJECT MESSAGE 25s Normal WorkflowRunning workflow/wonderful-tiger Workflow Running 25s Warning FailedScheduling pod/wonderful-tiger 0/21 nodes are available: 21 Insufficient cpu. 22s Normal NotTriggerScaleUp pod/wonderful-tiger pod didn't trigger scale-up (it wouldn't fit if a new node is added): 17 Insufficient cpu and kubectl get event -n argo --field-selector involvedObject.name=wonderful-tiger -o yaml apiVersion: v1 items: - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2020-08-05T17:01:09Z" involvedObject: apiVersion: argoproj.io/v1alpha1 kind: Workflow name: wonderful-tiger namespace: argo resourceVersion: "14645697" uid: f4bb68ae-84a2-4e40-8d5b-71f0a9ab91f3 kind: Event lastTimestamp: "2020-08-05T17:01:09Z" message: Workflow Running metadata: creationTimestamp: "2020-08-05T17:01:09Z" name: wonderful-tiger.16286dde6befd70c namespace: argo resourceVersion: "135777" selfLink: /api/v1/namespaces/argo/events/wonderful-tiger.16286dde6befd70c uid: 57c2bd09-11c6-4dc8-8ada-b33346764566 reason: WorkflowRunning reportingComponent: "" reportingInstance: "" source: component: workflow-controller type: Normal - apiVersion: v1 count: 3 eventTime: null firstTimestamp: "2020-08-05T17:01:09Z" involvedObject: apiVersion: v1 kind: Pod name: wonderful-tiger namespace: argo resourceVersion: "14645699" uid: a6c471b1-a31e-46fa-b53e-27554a81d328 kind: Event lastTimestamp: "2020-08-05T17:02:33Z" message: '0/21 nodes are available: 21 Insufficient cpu.' metadata: creationTimestamp: "2020-08-05T17:01:09Z" name: wonderful-tiger.16286dde6d37eca9 namespace: argo resourceVersion: "135781" selfLink: /api/v1/namespaces/argo/events/wonderful-tiger.16286dde6d37eca9 uid: d256c9fe-40fd-4b97-a7e6-62ec0470eb4f reason: FailedScheduling reportingComponent: "" reportingInstance: "" source: component: default-scheduler type: Warning - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2020-08-05T17:01:12Z" involvedObject: apiVersion: v1 kind: Pod name: wonderful-tiger namespace: argo resourceVersion: "14645700" uid: a6c471b1-a31e-46fa-b53e-27554a81d328 kind: Event lastTimestamp: "2020-08-05T17:01:12Z" message: 'pod didn''t trigger scale-up (it wouldn''t fit if a new node is added): 17 Insufficient cpu' metadata: creationTimestamp: "2020-08-05T17:01:12Z" name: wonderful-tiger.16286ddf00db261b namespace: argo resourceVersion: "135780" selfLink: /api/v1/namespaces/argo/events/wonderful-tiger.16286ddf00db261b uid: ac79183e-aaff-4f8d-a604-10a8c4b254f6 reason: NotTriggerScaleUp reportingComponent: "" reportingInstance: "" source: component: cluster-autoscaler type: Normal kind: List metadata: resourceVersion: "" selfLink: "" So maybe it's possible that the cluster-autoscaler event was too fast to be displayed? If that's the case then is there is to access the previous status messages of a workflow? Would it be enough to list events related to the pod in the UI? I think it would do the job yeah. And actually this will be useful for probably other things than this specific issue to be able to see the history of events. Similar to what's GKE is doing: I guess if this is available in the UI it will also be available in the status field of a Workflow YAML spec. Then we could also use this information from a machine for automatic processing. So yeah that would a very nice feature! As an MVP, I don't think we would make it available in the YAML. Instead, we would just make it available in the UI only. That would exclude the UI and the CLI. Do you think this should be MVP? For us, the most important is to have this information available on the UI. Having it in the YAML is just a bonus and a cool enhancement IMO. I like the idea of having all the information related to a workflow in a single YAML object. Its state, spec, and history of status. So my answer is yes it should go in MVP but if this requires too much work then having the information on the UI only is fine. Available for testing in v2.10.0-rc1.
gharchive/issue
2020-08-04T23:32:44
2025-04-01T04:33:32.248791
{ "authors": [ "alexec", "hadim" ], "repo": "argoproj/argo", "url": "https://github.com/argoproj/argo/issues/3673", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1447112945
References not correctly added to the model README There's functionality to auto add the values from the datasets "reference" field to the .argosmodel package READMEs. There's a bug and it's been broken recently and I haven't been able to figure out why. I've experimented with trying to fix this but something is still broken. https://github.com/argosopentech/argos-train/commit/cc2d7752dcb527e6a35a09c249acacb422231cef
gharchive/issue
2022-11-13T22:18:13
2025-04-01T04:33:32.250777
{ "authors": [ "PJ-Finlay", "argosopentech" ], "repo": "argosopentech/argos-train", "url": "https://github.com/argosopentech/argos-train/issues/25", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1202890802
Fix typo in German translation The German word for precision, "Ge­nau­ig­keit", was spelled wrong. Also, it seems that Github automatically converted the file to UTF-8. @jstuyts is this repo still being maintained?
gharchive/pull-request
2022-04-13T07:46:37
2025-04-01T04:33:32.252042
{ "authors": [ "tsaglam" ], "repo": "argparse4j/argparse4j", "url": "https://github.com/argparse4j/argparse4j/pull/139", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2122415999
Failed to launch without user gesture. This is a repo reproducing this question: https://github.com/daGaiGuanYu/show-me-bugs/tree/nostalgist240207 warn message in console: The AudioContext was not allowed to start. It must be resumed (or created) after a user gesture on the page. I don't know whether this can be fixed. If not, it would be nice to mention it in the documentation. Developer could show a dialog to tell users "rom download, start now?". And user's clicking will offer a "user gesture", and this limit are gone. I'm making a little APP (not completed): load nes games from local and the internet. Installable, and offline playing(not completed), and launching by clicking a nes file. Make a user gesture before launch: // ... await make_usergesture() Nostalgist.launch(...) function make_usergesture () { const div = document.createElement('div') div.style.position = 'fixed' div.style.left = '0' div.style.top = '0' div.style.right = '0' div.style.bottom = '0' div.style.zIndex = '1000' div.style.background = '#333' div.style.color = '#fff' div.style.display = 'grid' div.style.placeItems = 'center' div.innerHTML = 'Please click on the screen to continue' return new Promise<void>(res => { div.addEventListener('click', () => { document.body.removeChild(div) res() }) document.body.appendChild(div) }) } It's a limitation from the browser (see https://developer.chrome.com/blog/autoplay) so we cannot bypass it. We have to ensure the launch method is invoked after a user click event or a keyboard event. In fact, there is an undocumented parameter waitForInteraction working like this: await Nostalgist.nes({ rom: 'flappybird.nes', waitForInteraction({ done }) { const button = document.createElement('button') button.textContent = 'click to start' document.body.append(button) button.addEventListener('click', () => { done() button.remove() }) }, }) It still haven't been added to the documentation now, since it's a bit complex to explain, and I also think the API design may have room to improve to achieve better clarity. I think your approach for making a user gesture before launching is quite straightforward and you could just keep using it.
gharchive/issue
2024-02-07T08:01:07
2025-04-01T04:33:32.261500
{ "authors": [ "arianrhodsandlot", "daGaiGuanYu" ], "repo": "arianrhodsandlot/nostalgist", "url": "https://github.com/arianrhodsandlot/nostalgist/issues/18", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
305175583
build sample/initial pattern funtcion check for sample that are uploaded if not , take initial sample in real time done
gharchive/issue
2018-03-14T14:13:43
2025-04-01T04:33:32.263893
{ "authors": [ "arik-le", "gaiMaimoni" ], "repo": "arik-le/Chips-Bits", "url": "https://github.com/arik-le/Chips-Bits/issues/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1763252159
Feat(eos_cli_config_gen): add global logging event storm-control Enhancement summary global logging event for storm-control is the only one being disabled by default in eos. Add this global configuration option to eos_cli_config_gen to be able to enable it. access#show run all | i logging event no logging event storm-control discards global no logging event storm-control discards interval default logging event congestion-drops logging event port-channel member-status global no logging event login root logging event link-status global logging event spanning-tree global logging event flowcontrol received syslog interval 10 seconds access(config)#logging event ? congestion-drops Drops due to congestion flowcontrol Configure syslogging for flowcontrol link-status UPDOWN messages login Login activities port-channel Configure port-channel messages spanning-tree Spanning tree messages storm-control Configure storm-control access(config)#logging event storm-control ? discards Discards due to storm control access(config)#logging event storm-control discards ? global Configure global storm control discard logging interval Logging interval Which component of AVD is impacted eos_cli_config_gen Use case example logging: event: - storm_control_discards: enable: true interval: 10 logging event storm-control discards global logging event storm-control discards interval 10 Describe the solution you would like logging: event: - storm_control_discards: enable: <bool> interval: <int> should render the following cli: logging event storm-control discards global logging event storm-control discards interval <int> Describe alternatives you have considered use of eos_cli Additional context No response Contributing Guide [X] I agree to follow this project's Code of Conduct I think you have a typo in your data model logging: event: storm_control_discards: enable: <bool> interval: <int> Thanks
gharchive/issue
2023-06-19T10:36:03
2025-04-01T04:33:32.269322
{ "authors": [ "ClausHolbechArista", "kmueller68" ], "repo": "aristanetworks/ansible-avd", "url": "https://github.com/aristanetworks/ansible-avd/issues/2993", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
820194450
Validate role fails when NTP is not configured Issue Type [x] Bug Report [ ] EOS Configuration rendering error [x] Documentation Issue Report Summary While executing eos validation role on a device without ntp configured, ansible gets an unexpected value type and this breaks the MD documentation file (which is broken and not possible to read). CSV can still be properly interpreted. Component Name eos_validate_state role Steps to reproduce Execute EOS validation role on a device without NTP configuration. Expected results $ Actual results TASK [eos_validate_state : Validate ntp status] ************************************************************************************************************************************* fatal: [DC1-SUPER-SPINE1]: FAILED! => {"msg": "The conditional check 'ntp_status.stdout[0] | regex_search(\"synchronised to NTP server\")' failed. The error was: Unexpected templating type error occurred on ({% if ntp_status.stdout[0] | regex_search(\"synchronised to NTP server\") %} True {% else %} False {% endif %}): expected string or bytes-like object"} ...ignoring fatal: [DC1-SUPER-SPINE2]: FAILED! => {"msg": "The conditional check 'ntp_status.stdout[0] | regex_search(\"synchronised to NTP server\")' failed. The error was: Unexpected templating type error occurred on ({% if ntp_status.stdout[0] | regex_search(\"synchronised to NTP server\") %} True {% else %} False {% endif %}): expected string or bytes-like object"} ...ignoring Fixed by #763
gharchive/issue
2021-03-02T16:46:19
2025-04-01T04:33:32.273095
{ "authors": [ "ClausHolbechArista", "xaviramon" ], "repo": "aristanetworks/ansible-avd", "url": "https://github.com/aristanetworks/ansible-avd/issues/761", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
201134262
Splat v.2.0.0 compatibility as in the title Please read the contrib guidelines before filing in the future (the ones you deleted to post your message)
gharchive/issue
2017-01-16T22:44:05
2025-04-01T04:33:32.293836
{ "authors": [ "aritchie", "belandsoft" ], "repo": "aritchie/userdialogs", "url": "https://github.com/aritchie/userdialogs/issues/317", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
836894744
Tag sessions with the capturing interface Is your feature request related to a problem? Please describe. No problem, but it would be useful if we could easily filter on the capturing interface. Describe the solution you'd like I don't know if this would be best as just a tag, as a discrete column in the UI, or modifying the node name. Describe alternatives you've considered N/A Additional context In my experimentation, I was able to make this work by creating an instance of each of the following files for each interface I was monitoring: /data/moloch/etc/config.ini /data/moloch/bin/moloch_config_interfaces.sh /etc/systemd/system/molochcapture.service The specific config.ini and moloch_config_interfaces.sh each was changed to only reference an individual interface. I also made discrete pcap directories for each interface and specified those in the config.ini files. The molochcapture.service files I created were specified with -n ORIGINALNODENAE-INTERFACENAME, -t INTERFACENAME, a unique log file name, and finally the unique config.ini file for ExecStart, and the unique moloch_config_interfaces.sh for ExecStartPre. I specified a unique node name, otherwise I would get errors from elasticsearch referring to the stats index, specifically looking like this: {"error":{"root_cause":[{"type":"version_conflict_engine_exception","reason":"[molochVM]: version conflict, current version [22070] is higher or equal to the one provided [22070]","index_uuid":"NMAYhwULQvahWDcaYKhl1A","shard":"0","index":"stats_v4"}],"type":"version_conflict_engine_exception","reason":"[molochVM]: version conflict, current version [22070] is higher or equal to the one provided [22070]","index_uuid":"NMAYhwULQvahWDcaYKhl1A","shard":"0","index":"stats_v4"},"status":409} I'd guess this is due to having two moloch_capture processes running on the same VM at the same time. I added the tag since I wasn't sure which would work best, so I figured I'd try with a tag as well. Generally that seems to be working and I'm seeing both the unique node name, which includes the interface name, as well as the interface name as a tag I'm now realizing this is probably what the [node1] and [node2] sections are at the bottom of the config.ini file... This is already mostly possible if you are ok using the tags field: If using 1 interface per capture just either use the -t interfacename option or the --op tags=interfacename to capture If using multple interfaces per capture then look at the https://arkime.com/settings#interfaceops setting If you wanted a new field called "interface" that would have to be added using https://arkime.com/settings#custom-fields, and then you would use --op interface=interfacename or again the interfaceops setting. I'd guess this is due to having two moloch_capture processes running on the same VM at the same time. yes node name should be unique. Instead of running multiple, you should just have capture listen to 2 interfaces I'm now realizing this is probably what the [node1] and [node2] sections are at the bottom of the config.ini file... not sure what you mean, but order in the file shouldn't matter. Just make sure settings you are adding actually end up in the right section. It seems i missed https://arkime.com/faq#how-do-i-monitor-multiple-interfaces Most of the info there is what I ended up doing, but I did differ by adding the tag and having two config files. One question on the existing instructions: When the two different systemd services start, each one would run the moloch_config_interfaces.sh script. If the single script makes changes to all of the interfaces, wouldn't that potentially cause issues with other interfaces that are already capturing? For example if you're capturing on eth1 and eth2, and restart the eth2 process, since both eth1 and eth2 are referenced in the common moloch_config_interfaces.sh, would that impact eth1? not sure what you mean, but order in the file shouldn't matter. Just make sure settings you are adding actually end up in the right section. To clear up what I meant: I was mixing up a few things, and didn't realize the use of creating unique sections in the config.ini file. Hey, I'm having a bit of an issue understanding how to specify interfaceops in the config.ini file. Anything you can give me to point me in the right direction would be appreciated! Updated the FAQ, otherwise come on over to slack for questions
gharchive/issue
2021-03-20T19:36:42
2025-04-01T04:33:32.309022
{ "authors": [ "awick", "copeland3300" ], "repo": "arkime/arkime", "url": "https://github.com/arkime/arkime/issues/1638", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
253946981
ClickOutsideModule not working inside an anchor I import ClickOutsideModule in the following module: @NgModule({ imports: [ClickOutsideModule], declarations: [TopBarComponent, MyComponent] }) export class TopBarModule { } MyComponent is as follows: @Component({ moduleId: module.id, selector: 'my-component', styleUrls: ['my-component.css'], templateUrl: 'my-component.component.html' }) export class MyComponent { public onClickedOutside(event: Event) { console.log('Clicked outside:', event); } } with the template being: < ul (clickOutside)="onClickedOutside($event)">Hello</ul > Nothing is printed on the console when I click outside the unsorted list. Works fine for me. Only notable thing is that you can't have spaces in your tags (although that should've given you an error): <ul (clickOutside)="onClickedOutside($event)">Hello</ul>
gharchive/issue
2017-08-30T10:12:38
2025-04-01T04:33:32.313925
{ "authors": [ "arkon", "jacopolanzoni" ], "repo": "arkon/ng-click-outside", "url": "https://github.com/arkon/ng-click-outside/issues/15", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1030826652
Split permutation from sponge construction Work towards #29; this doesn't touch the constraint system implementation yet, in order to be able to get design feedback on the software part. The poseidon::PoseidonParameters struct is renamed to poseidon::Parameters but otherwise remains unchanged. The poseidon::PoseidonSpongeState struct is renamed to poseidon::State and redefined to hold just the state itself, as well as the parameters needed to run the permutation. It exposes a permute(&mut self) method, rate() and capacity() accessors, as well as Index, IndexMut, AsRef, and AsMut impls that allow access to the state. The poseidon::PoseidonSponge struct is renamed to poseidon::Sponge and holds a State and a DuplexSpongeMode. In other words, it consists of the state, together with the extra data tracking how that state is being used to implement a higher-level duplex construction. The CryptographicSponge trait is changed so that new() takes an owned, Self::Parameters, not a borrowed one. This allows the caller to decide where to copy data, instead of forcing the sponge implementation to clone internally. Or, a CryptographicSponge implementation could declare the associated Parameters type to be some shared type (like an Arc wrapper) that avoids the need to copy at all. The SpongeExt trait that allows converting back and forth between a state and a sponge is deleted; it's not safe to pass between abstraction layers that way. Before we can merge this PR, please make sure that all the following items have been checked off. If any of the checklist items are not applicable, please leave them but write a little note why. [x] Targeted PR against correct branch (master) [x] Linked to Github issue with discussion and accepted design OR have an explanation in the PR that describes this work. [ ] Wrote unit tests [x] Updated relevant documentation in the code [ ] Added a relevant changelog entry to the Pending section in CHANGELOG.md [ ] Re-reviewed Files changed in the Github PR explorer (A rendered copy of the docs is available here: https://rustdoc.penumbra.zone/main/ark_sponge/poseidon/index.html) Hey, just bumping this -- if this approach seems good I can also apply it to the constraint implementations. Feel free to apply to the constraints implementations.
gharchive/pull-request
2021-10-19T23:31:26
2025-04-01T04:33:32.324364
{ "authors": [ "hdevalence", "weikengchen" ], "repo": "arkworks-rs/sponge", "url": "https://github.com/arkworks-rs/sponge/pull/30", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1255088981
Arlon cli context management support Currently arlon cli interacted with 3 entities: Git repo ArgoCD Server K8s API Server of Management Cluster It would be very beneficial to have a context management similar to how its done by argocd cli. A context can be defined as combination of git repo (with credentials if there), ArgoCD Server connection information, k8s API Server connection information. All the information about multiple contexts can be saved in ~/.arlon folder on user machine. User should be able to perform following activities with the contexts: Create new contexts Delete contexts Shift between contexts Renew tokens in the context (good to have) Once a context is set by the user, the individual commands having parameters about above 3 (git, k8s, argocd) can be treated as optional. If present, they override the set context. Commands can be see as follows: Create new context arlon context add <new_context_name> --argocd-server <argocd_addr> --gitrepo ..... Delete context arlon context delete <context_name> Shift between context arlon context set <context_name> Impact of other commands because of context. Below given is an example command: arlon create clusterspec [--context <context_to_use>] [--gitrepo <repo url>].. Notice in above command that git repo and context are optional parameters. Individual command to retrieve the git repo in following order: a. overridden param in command b. context present in the command c. command set globally Similarly do the same for argocd server & mgmt cluster also. @bcle who can scope this work? Can you label this with a t-shirt size?
gharchive/issue
2022-06-01T06:19:43
2025-04-01T04:33:32.362808
{ "authors": [ "bcle", "chinmaykolhatkar", "cre8minus1" ], "repo": "arlonproj/arlon", "url": "https://github.com/arlonproj/arlon/issues/52", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1619638900
Service Port Alignment in Operator objects We need to grab http, grpc, and metrics ports from the App Config of each service, and use it for the Deployment, Service, and Ingress configs created by the operator. https://github.com/armadaproject/armada-operator/issues/96 duplicate
gharchive/issue
2023-03-10T21:21:23
2025-04-01T04:33:32.364879
{ "authors": [ "kannon92", "suprjinx" ], "repo": "armadaproject/armada", "url": "https://github.com/armadaproject/armada/issues/2256", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1710749078
How to setup on Lazy? I'm trying with: { "arnamak/stay-centered.nvim", lazy = false, }, And it is listed as "loded" but it does nothing. Nevermind, just created an autocmd local augroup = vim.api.nvim_create_augroup local autocmd = vim.api.nvim_create_autocmd -- Keep cusor always centered local cursor_group = augroup("cursor", { clear = true }) autocmd({ "CursorMoved"}, { desc = "Keep cursor always centered", group = cursor_group, callback = function() vim.api.nvim_exec("norm zz", false) baseevent "CursorCentered" end, }) SOLVED: -- stay-centered.nvim [cursor centered] -- https://github.com/arnamak/stay-centered.nvim { "arnamak/stay-centered.nvim", lazy=false, opts = { skip_filetypes = { } } },
gharchive/issue
2023-05-15T20:06:17
2025-04-01T04:33:32.417984
{ "authors": [ "Zeioth" ], "repo": "arnamak/stay-centered.nvim", "url": "https://github.com/arnamak/stay-centered.nvim/issues/8", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1453013459
[FEATURE] Update to choco 4.10.10 Is your feature request related to a problem? Please describe. Update choco to benefit from the latest enhancements and to seek support more easily. Describe the solution you'd like Update to the latest choco release. Additional context A first attempt has shown that the expression handling has changed : some tests will fail. We must determine if the tests fails because of a regression. It requires to isolate the failing tests, hard-code the models and post issues in the choco project if needed. JUnit tests that shows a case where choco found results in 4.10.6 but not in 4.10.10 import org.junit.Before; import org.junit.Test; import org.chocosolver.solver.Model; import org.chocosolver.solver.Solver; import org.chocosolver.solver.variables.IntVar; public class LittleTest { public static Model model; public static IntVar copper, neon, iron, silver, result; @Before /*** * Tests if choco can solve "copper*neon=iron*silver" via various way of writing constraints */ public void setup(){ model = new Model("copper*neon=iron*silver"); //model.getSettings().getWelcomeMessage(); copper = model.intVar("copper",100000,999999); neon = model.intVar("neon",1000,9999); iron = model.intVar("iron",1000,9999); silver = model.intVar("silver",100000,999999); //result is an intermediary to compare copper*neon and iron*silver //all test would fail if ub was equal to 'org.chocosolver.solver.variables.IntVar.MAX_INT_BOUND' result = model.intVar("result",0,org.chocosolver.solver.variables.IntVar.MAX_INT_BOUND*10); model.allDifferent(new IntVar[]{copper,neon,iron,silver}).post(); } @Test /*** * Using times dirrectly to make our constraint * will fail to find any solution because value needed is out of result's bound */ public void model1(){ result = model.intVar("result",0,org.chocosolver.solver.variables.IntVar.MAX_INT_BOUND); model.times(copper, neon, result).post(); model.times(silver, iron, result).post(); Solver solver = model.getSolver(); //solver.showStatistics(); //solver.showSolutions(); solver.findSolution(); //solver.printVersion(); assertEquals(solver.getSolutionCount(),0); } @Test /*** * Working version of the previous test. */ public void model1Success(){ model.times(copper, neon, result).post(); model.times(silver, iron, result).post(); Solver solver = model.getSolver(); //solver.showStatistics(); //solver.showSolutions(); solver.findSolution(); //solver.printVersion(); assertEquals(solver.getSolutionCount(),1); } @Test /*** * Creating a model using arithm instead of times directly. */ public void model2(){ model.arithm(copper, "*",neon,"=",result).post(); model.arithm(iron, "*",silver,"=",result).post(); Solver solver = model.getSolver(); solver.findSolution(); assertEquals(solver.getSolutionCount(),1); } @Test /*** * Creating a model using the operators function given by IntVar */ public void model3(){ copper.mul(neon).eq(result).post(); iron.mul(silver).eq(result).post(); System.out.println(model.toString()); Solver solver = model.getSolver(); solver.findSolution(); assertEquals(solver.getSolutionCount(),1); } @Test /*** * Creating a model using the operators function given by IntVar without using result as an intermediary */ public void model4(){ copper.mul(neon).eq(iron.mul(silver)).post(); System.out.println(model.toString()); Solver solver = model.getSolver(); solver.findSolution(); assertEquals(solver.getSolutionCount(),1); } }``` My conclusions are that: The use of MAX_INT_BOUND is unclear, because it is not checked when creating a variable (see setupResult). The introduction of auxiliary variable is an issue (see for instance model1 and model2). The issue is that the domain of the right member (result) is not used for defining the domain of the auxiliary variable. For instance, it is the case for the arithm constraint : see here or just below. // v1 * v2 OP v3 int[] bounds = VariableUtils.boundsForMultiplication(var1, var2); IntVar var4 = ref().intVar(bounds[0], bounds[1]); ref().times(var1, var2, var4).post(); return arithm(var4, op2, var3); In this case, it is unclear to me why the variable v4 is introduced. I have slightly improved the code. You can find it below. import org.chocosolver.solver.Model; import org.chocosolver.solver.Solver; import org.chocosolver.solver.variables.IntVar; import org.junit.Assert; import org.junit.Before; import org.junit.Test; public class ProductTest { private Model model; private IntVar copper, neon, iron, silver, result; @Before /*** * Tests if choco can solve "copper*neon=iron*silver" via various way of writing constraints */ public void setup(){ model = new Model("copper*neon=iron*silver"); copper = model.intVar("copper",100000,999999); neon = model.intVar("neon",1000,9999); iron = model.intVar("iron",1000,9999); silver = model.intVar("silver",100000,999999); model.allDifferent(new IntVar[]{copper,neon,iron,silver}).post(); } public void setupResult(int factor) { //result is an intermediary to compare copper*neon and iron*silver //all test would fail if factor = 1. result = model.intVar("result",0, factor * IntVar.MAX_INT_BOUND); } public void solve(int solutionCount) { Solver solver = model.getSolver(); System.out.println(model); solver.printVersion(); solver.showStatistics(); solver.showSolutions(); solver.findSolution(); Assert.assertEquals(solutionCount, solver.getSolutionCount()); } @Test /*** * Using times directly to make our constraint * will fail to find any solution because value needed is greater than the result's upper bound. */ public void model1NoSol(){ setupResult(1); model.times(copper, neon, result).post(); model.times(silver, iron, result).post(); solve(0); } @Test /*** * Working version of the previous test. */ public void model1(){ setupResult(10); model.times(copper, neon, result).post(); model.times(silver, iron, result).post(); solve(1); } @Test /*** * Creating a model using arithm instead of times directly. */ public void model2(){ setupResult(10); model.arithm(copper, "*",neon,"=",result).post(); model.arithm(iron, "*",silver,"=",result).post(); solve(1); } @Test /*** * Creating a model using the operators function given by IntVar */ public void model3(){ setupResult(10); copper.mul(neon).eq(result).post(); iron.mul(silver).eq(result).post(); solve(1); } @Test /*** * Creating a model using the operators function given by IntVar without using result as an intermediary */ public void model4(){ copper.mul(neon).eq(iron.mul(silver)).post(); solve(1); } } https://github.com/chocoteam/choco-solver/issues/978 https://github.com/chocoteam/choco-solver/issues/979 https://github.com/chocoteam/choco-solver/issues/980 You can now move the tests in Failing.java with a comment.
gharchive/issue
2022-11-17T09:34:13
2025-04-01T04:33:32.424192
{ "authors": [ "Imp95", "arnaud-m" ], "repo": "arnaud-m/cryptator", "url": "https://github.com/arnaud-m/cryptator/issues/53", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }