id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
2087104585 | Work around SwiftPM build regression affecting packages with C++ targets
This is a workaround to the SwiftPM issue being fixed here:
https://github.com/apple/swift-package-manager/pull/7267
@swift-ci please test
No longer relevant, the SwiftPM bug was fixed!
| gharchive/pull-request | 2024-01-17T22:22:01 | 2025-04-01T04:33:30.749645 | {
"authors": [
"stmontgomery"
],
"repo": "apple/swift-testing",
"url": "https://github.com/apple/swift-testing/pull/188",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1909296675 | Freestanding declaration macro fails inside classes: Expected '{' in body of function declaration
Description
Using a freestanding declaration macro inside a class produces a build error:
Expected '{' in body of function declaration
Here is the expanded macro:
func myFunction() {
}
The same code compiles if the macro is used inside a struct.
Steps to reproduce
Create a swift package using the macro template.
Client code:
import MyMacro
struct MyStruct {
#blankFunction // works ok here
}
class MyClass {
#blankFunction // fails here
}
Macro declaration:
@freestanding(declaration, names: named(myFunction))
public macro blankFunction() = #externalMacro(
module: "MyMacroMacros",
type: "BlankFunctionMacro"
)
Macro implementation:
import SwiftCompilerPlugin
import SwiftSyntax
import SwiftSyntaxBuilder
import SwiftSyntaxMacros
public struct BlankFunctionMacro: DeclarationMacro {
public static func expansion(
of node: some FreestandingMacroExpansionSyntax,
in context: some MacroExpansionContext
) throws -> [DeclSyntax] {
return ["func myFunction() {}"]
}
}
@main
struct MyMacroPlugin: CompilerPlugin {
let providingMacros: [Macro.Type] = [
BlankFunctionMacro.self,
]
}
Expected behavior
swift run MyMacroClient should compile and run without any errors.
Environment
swift-driver version: 1.87.1 Apple Swift version 5.9 (swiftlang-5.9.0.128.108 clang-1500.0.40.1)
Target: arm64-apple-macosx13.0
Xcode 15.0
Build version 15A240d
Deployment target: macOS 13.6 (22G120)
Hi facing the same issue now. Just a follow-up. What's the progress regarding this issue? Thank you!
This also fails in the same way when adding a 'static func...' to anything, structs included.
Also here. Thanks for reporting, hoping for a fix...
It's also current blocking me on a macros-heavy project.
Same here, blocking awesome ideas realisation. Any updates on this?
Tried on Swift 5.9.2 and 5.10 - doesn't work. 🙅♂️
Tried on Swift 6.0 DEVELOPMENT-SNAPSHOT-2024-04-30-a-osx and it works! 🚀
| gharchive/issue | 2023-09-22T17:43:40 | 2025-04-01T04:33:30.755919 | {
"authors": [
"MihaelIsaev",
"drseg",
"huyaoyu",
"juozasvalancius",
"regexident"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/issues/68704",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
246486426 | [sil-combine] Make sure that (apply (partial_apply)) -> (apply') bails if we have an argument with a dependent type.
[sil-combine] Make sure that (apply (partial_apply)) -> (apply') bails if we have an argument with a dependent type.
What is going on here is that currently this optimization if itneeds to perform
lfietime extension always creates an alloc_stack at the beginning/end of a
function. If the object whose lifetime is being extended has a dependent type,
then the alloc_stack will be created before the dependent type exists resulting
in the compiler crashing.
rdar://33595317
@swift-ci smoke test and merge
| gharchive/pull-request | 2017-07-29T00:23:52 | 2025-04-01T04:33:30.758128 | {
"authors": [
"gottesmm"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/11253",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
307319870 | [RemoteMirrors] Interop header fixes and library lookup caching.
Some deficiencies in the interop API were identified when an actual client started trying to use it.
typeinfo and childinfo needed interop variants that used interop variants for their fields.
The client needed to be able to look things up starting from a metadata address directly, rather than starting from an object.
Figuring out which library owns a particular metadata address is not particularly fast, but is done frequently, so it needed a cache.
rdar://problem/37538580
@swift-ci please test
| gharchive/pull-request | 2018-03-21T16:25:03 | 2025-04-01T04:33:30.760198 | {
"authors": [
"mikeash"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/15395",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
141752602 | [Coverage] Respect function linkage in PGO name variables
What's in this pull request?
The Ubuntu bots have a problem with the ordered IR check lines in test/IRGen/coverage.swift. The intended match appears later in the file, so try again with CHECK-DAG lines.
Resolved bug number: (SR-)
Before merging this pull request to apple/swift repository:
[ ] Test pull request on Swift continuous integration.
Triggering Swift CI
The swift-ci is triggered by writing a comment on this PR addressed to the GitHub user @swift-ci. Different tests will run depending on the specific comment that you use. The currently available comments are:
Smoke Testing
Platform | Comment
------------ | -------------
All supported platforms | @swift-ci Please smoke test
OS X platform | @swift-ci Please smoke test OS X platform
Linux platform | @swift-ci Please smoke test Linux platform
Validation Testing
Platform | Comment
------------ | -------------
All supported platforms | @swift-ci Please test
OS X platform | @swift-ci Please test OS X platform
Linux platform | @swift-ci Please test Linux platform
Note: Only members of the Apple organization can trigger swift-ci.
@swift-ci Please test Linux platform
__DATA doesn't exist on ELF. Fixed the test case in dc11f8bf.
| gharchive/pull-request | 2016-03-18T01:54:41 | 2025-04-01T04:33:30.764530 | {
"authors": [
"vedantk"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/1729",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
377504583 | Revert "Revert "[Build System: CMake] make add_swift_library a wrappe…
…r to add_llvm_library""
This reverts commit 103f9a8246beb385ca6d8a33003a0ebac035095d.
Replace this paragraph with a description of your changes and rationale. Provide links to external references/discussions if appropriate.
Resolves SR-NNNN.
@gottesmm kindly volunteered to help sort out the internal CI conflicts.
@swift-ci please test
Build failed
Swift Test OS X Platform
Git Sha - 1efdb2954778d9baac05201dfc3b857e77612219
Build failed
Swift Test Linux Platform
Git Sha - 1efdb2954778d9baac05201dfc3b857e77612219
@swift-ci please test
@swift-ci please test
Build failed
Swift Test OS X Platform
Git Sha - 1efdb2954778d9baac05201dfc3b857e77612219
Build failed
Swift Test Linux Platform
Git Sha - 1efdb2954778d9baac05201dfc3b857e77612219
@swift-ci please test
Build failed
Swift Test OS X Platform
Git Sha - 0347eff30b84fbd1d34f30c48c4c2e12cb82a132
Build failed
Swift Test Linux Platform
Git Sha - 0347eff30b84fbd1d34f30c48c4c2e12cb82a132
@gottesmm - I suppose that this should wait until you're ready to deal with the internal merge?
Thanks for waiting. I'll take care of it.
| gharchive/pull-request | 2018-11-05T17:30:03 | 2025-04-01T04:33:30.772444 | {
"authors": [
"compnerd",
"gottesmm",
"swift-ci"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/20325",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
378978885 | [ABI] Retrieve all associated conformances via a runtime function.
Introduce a new runtime entry point, swift_getAssociatedConformanceWitness(),
which extracts an associated conformance witness from a witness table.
Teach IRGen to use this entry point rather than loading the witness
from the witness table and calling it directly.
There’s no advantage to doing this now, but it is staging for changing the
representation of associated conformances in witness tables.
@swift-ci please smoke test
@swift-ci please benchmark
Build comment file:
Performance: -O
TEST
OLD
NEW
DELTA
RATIO
Regression
Dictionary2
958
1123
+17.2%
0.85x
IterateData
1524
1675
+9.9%
0.91x
StringEqualPointerComparison
600
657
+9.5%
0.91x
Performance: -Osize
TEST
OLD
NEW
DELTA
RATIO
Regression
StringEqualPointerComparison
571
628
+10.0%
0.91x
How to read the data
The tables contain differences in performance which are larger than 8% and
differences in code size which are larger than 1%.
If you see any unexpected regressions, you should consider fixing the
regressions before you merge the PR.
Noise: Sometimes the performance results (not code size!) contain false
alarms. Unexpected regressions which are marked with '(?)' are probably noise.
If you see regressions which you cannot explain you can try to run the
benchmarks again. If regressions still show up, please consult with the
performance team (@eeckstein).
Hardware Overview
Model Name: Mac Pro
Model Identifier: MacPro6,1
Processor Name: 12-Core Intel Xeon E5
Processor Speed: 2.7 GHz
Number of Processors: 1
Total Number of Cores: 12
L2 Cache (per Core): 256 KB
L3 Cache: 30 MB
Memory: 64 GB
--------------
| gharchive/pull-request | 2018-11-09T00:49:12 | 2025-04-01T04:33:30.781608 | {
"authors": [
"DougGregor",
"swift-ci"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/20449",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
390493764 | Exclude/fix certain tests for swift-evolve
This pull request adds a swift_evolve feature to our lit configuration and marks several tests as unsupported with it.
swift-evolve is a tool currently in development which automatically modifies source code in ways that should be source- and ABI-compatible; we will be using it to test resilience. Currently it simply shuffles declarations. The tests I'm marking as unsupported in this pull request either intentionally depend on the order of declarations, or expose bugs in our tools that we've filed but haven't fixed.
This PR also fixes one more order-dependent test in SourceKit.
Part of rdar://problem/44426013.
Tagging @nkcsgexi because he wrote a few of these tests and @jrose-apple because he suggested using a lit feature.
@swift-ci please smoke test
@swift-ci please smoke test
| gharchive/pull-request | 2018-12-13T02:39:38 | 2025-04-01T04:33:30.784065 | {
"authors": [
"brentdax"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/21276",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
451500190 | IRGen: getSingletonAggregateFieldType must not return field if its ac…
…cess level does not match
rdar://50554717
@swift-ci Please test
Build failed
Swift Test Linux Platform
Git Sha - 12b27a1332b852480a529bf6bb5b00d3353df504
Build failed
Swift Test OS X Platform
Git Sha - 12b27a1332b852480a529bf6bb5b00d3353df504
@swift-ci Please test
Build failed
Swift Test Linux Platform
Git Sha - 12b27a1332b852480a529bf6bb5b00d3353df504
Build failed
Swift Test OS X Platform
Git Sha - 12b27a1332b852480a529bf6bb5b00d3353df504
Nice catch!
| gharchive/pull-request | 2019-06-03T13:59:23 | 2025-04-01T04:33:30.788663 | {
"authors": [
"aschwaighofer",
"slavapestov",
"swift-ci"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/25216",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
120602039 | Fix lexer to properly handle single quote strings
In response to a FIXME comment, the lexer now suggests
replacing double quotes inside single quote strings
with a backslash followed by a single quote.
Please add tests for the fixits. The file test/type/array.swift at the end shows how to test fixits:
typealias FixIt1 = Int[][] // expected-error{{array types are now written with the brackets around the element type}}{{20-20=[}}{{25-26=}}
Updated the test in expressions.swift to expect escaped double quotes!
Just tested compilation, tests are passing.
@gribozavr Can this be merged?
@Jumhyn Would you mind squashing everything into one commit and doing a push -f to the pull request?
There's one case where I think this will fail: it will suggest converting 'a\"b' to "a\\"b". You need to check to see whether the quote mark you're escaping is already escaped. It's subtle because you still want to convert 'a\\"b' to "a\\\"b".
I think it will be sufficient to step through the string, skipping all escaped sequences and replacing unescaped " with \".
You're right. This approach also makes it easier to address un-escaping escaped single quotes. Will push a new patch shortly.
This should work now -- added new tests for the cases noted above as well.
Done. Are the style rules available on Swift.org (or otherwise)? I haven't seen an explicit style guide mentioned anywhere.
All of the C++ code in the swift compiler should follow the LLVM coding standards: http://llvm.org/docs/CodingStandards.html
Sweet, thanks for putting this together!
No problem, and thanks @lattner for the link to the coding standards! Those should be mentioned somewhere in reference to Swift in the contributing guide on Swift.org. Currently, the only line mentioning these is in the LLVM and Swift section, which says, "Contributions to Swift’s LLVM or Clang clones are governed by the LLVM Developer Policy and should follow the appropriate licensing and coding standards," but makes no mention of contributions to the Swift code itself. Probably worth a reference earlier on the page under the Contributing Code heading.
| gharchive/pull-request | 2015-12-06T02:06:17 | 2025-04-01T04:33:30.795625 | {
"authors": [
"Jumhyn",
"cwillmor",
"gribozavr",
"lattner"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/258",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
159753206 | Don't treat Swift methods named "init" as ObjC ARC init methods.
Under ARC, methods in the "init" family are considered to have NS_REPLACES_RECEIVER semantics ("consumes" self and returning a value at +1). This is correct for Objective-C "init methods", which are equivalent for Swift's initializers, but almost never correct for any other methods that happen to start with the word "init".
Note that Swift still follows all the other ARC conventions, so if you name a method or property, say, "newItemController", the value will be returned at +1. For methods this is probably desirable, but for properties maybe not. We could do something similar for property accessors to make sure they always have the default "no method family" semantics in Objective-C.
rdar://problem/25759260
Before merging this pull request to apple/swift repository:
[ ] Test pull request on Swift continuous integration.
Triggering Swift CI
The swift-ci is triggered by writing a comment on this PR addressed to the GitHub user @swift-ci. Different tests will run depending on the specific comment that you use. The currently available comments are:
Smoke Testing
Platform | Comment
------------ | -------------
All supported platforms | @swift-ci Please smoke test
All supported platforms | @swift-ci Please smoke test and merge
OS X platform | @swift-ci Please smoke test OS X platform
Linux platform | @swift-ci Please smoke test Linux platform
Validation Testing
Platform | Comment
------------ | -------------
All supported platforms | @swift-ci Please test
All supported platforms | @swift-ci Please test and merge
OS X platform | @swift-ci Please test OS X platform
OS X platform | @swift-ci Please benchmark
Linux platform | @swift-ci Please test Linux platform
Lint Testing
Language | Comment
------------ | -------------
Python | @swift-ci Please Python lint
Note: Only members of the Apple organization can trigger swift-ci.
@rjmccall, @jckarter, @slavapestov, can one of you review the SILGen part?
@swift-ci Please test
Seems reasonable. Is there a good way to test the round-trip case, where we try to call the ObjC method exported through the bridging header, to ensure we see the attribute and get the right calling convention on the caller side?
Good idea. (Also, clearly I missed something; the OS X failure is related.)
@swift-ci Please test
Note: not quite ready to merge yet because I haven't written the test Joe suggested.
There we go.
@swift-ci Please test
Linux failure is unrelated, merging.
| gharchive/pull-request | 2016-06-11T03:27:12 | 2025-04-01T04:33:30.802110 | {
"authors": [
"jckarter",
"jrose-apple"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/2989",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
706241382 | [linux] remove absolute rpath of /usr/lib/swift/linux added to many shared libraries
This was presumably added as a backup, in case the libraries in a toolchain couldn't be found, but will not work well, so take it out.
Here's a full list of shared library runpaths from the just-released official Swift 5.3 toolchain for linux:
> find swift-5.3-RELEASE-ubuntu20.04/ -name "lib*.so"| xargs readelf -d | ag "File:|runpath"
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/libsourcekitdInProc.so
0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN/../lib/swift/linux:/usr/lib/swift/linux]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/pm/4_2/libPackageDescription.so
0x000000000000001d (RUNPATH) Library runpath: [/home/build-user/swift-nightly-install/usr/lib/swift/linux]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/pm/4/libPackageDescription.so
0x000000000000001d (RUNPATH) Library runpath: [/home/build-user/swift-nightly-install/usr/lib/swift/linux]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/pm/llbuild/libllbuildSwift.so
0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN:$ORIGIN/../../linux]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/pm/llbuild/libllbuild.so
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libicuucswift.so
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libswiftRemoteMirror.so
0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN:/usr/lib/swift/linux]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libdispatch.so
0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libBlocksRuntime.so
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libFoundationNetworking.so
0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libicudataswift.so
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libXCTest.so
0x000000000000001d (RUNPATH) Library runpath: [/home/build-user/build/buildbot_linux/swift-linux-x86_64/lib/swift/linux]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libswiftDispatch.so
0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libswift_Differentiation.so
0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN:/usr/lib/swift/linux]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/lib_InternalSwiftSyntaxParser.so
0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN:/usr/lib/swift/linux]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libswiftGlibc.so
0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN:/usr/lib/swift/linux]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libFoundationXML.so
0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libFoundation.so
0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libswiftSwiftOnoneSupport.so
0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN:/usr/lib/swift/linux]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libicui18nswift.so
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/swift/linux/libswiftCore.so
0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN:/usr/lib/swift/linux]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/liblldb.so
0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN/../lib:/home/build-user/build/buildbot_linux/llvm-linux-x86_64/./lib]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/libIndexStore.so
0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN/../lib]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/clang/10.0.0/lib/linux/libclang_rt.hwasan-x86_64.so
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/clang/10.0.0/lib/linux/libclang_rt.ubsan_standalone-x86_64.so
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/clang/10.0.0/lib/linux/libclang_rt.scudo_minimal-x86_64.so
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/clang/10.0.0/lib/linux/libclang_rt.scudo-x86_64.so
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/clang/10.0.0/lib/linux/libclang_rt.ubsan_minimal-x86_64.so
0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN/../lib:/home/build-user/build/buildbot_linux/llvm-linux-x86_64/./lib]
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/clang/10.0.0/lib/linux/libclang_rt.dyndd-x86_64.so
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/clang/10.0.0/lib/linux/libclang_rt.asan-x86_64.so
File: swift-5.3-RELEASE-ubuntu20.04/usr/lib/libswiftDemangle.so
0x000000000000001d (RUNPATH) Library runpath: [$ORIGIN:/usr/lib/swift/linux]
If there's any difference between the toolchain being used and the one installed in /usr/lib/swift/linux, say if you download this 5.3 toolchain for linux into your home directory but 5.2.5 is installed in /usr/lib/swift/linux, using that system runpath won't work properly. I checked and all these libraries already have a runpath relative to $ORIGIN, ie locally, so all this absolute path will do is supply the system Swift library if the local one is missing (obviously if the used toolchain is installed in the system, ie /usr/lib/swift/linux, this absolute path is redundant). But since using the system Swift libraries will not work properly if the Swift versions don't match, causing subtle bugs in the worst case, I think it's better not to add this absolute runpath at all.
It was added five years ago by @bitjammer, ad95b5f, and copied to SourceKit a couple years later, a93bddf. I brought up this issue earlier this year, but got no response at the time, so I let it go. I think this is worth getting in and into the 5.3 branch for the next patch release.
@gottesmm and @compnerd, I think this should be removed, similar to the other ELF runpath issues we discussed. @tachoknight, let us know what you think from your Fedora packaging perspective.
As we discussed offline, can you create a PR against swift-integration-test that validates that the snapshots maintain this property? Then I think you can do cross-repo testing to test it with this PR. Then first merge this and then the integration test one.
Will do.
@drexin, let me know what you think, I'd like to get this into the 5.3.1 patch release.
@swift-ci smoke test
Thanks, sometimes adding a comment seems to get the CI to run?
I think the CI needs a swift kick. 😉
@swift-ci smoke test
Passes CI, ready to merge.
apple/swift-integration-tests#72
@swift-ci test
I validated in the CI job that swift-ci found the appropriate PR on swift-integration-tests.
Confirmed that the CI checked out the new integration test and that those tests were run, but the log doesn't list them individually so my new test isn't listed. The Windows CI failure appears unrelated.
| gharchive/pull-request | 2020-09-22T10:01:04 | 2025-04-01T04:33:30.809488 | {
"authors": [
"buttaface",
"drexin",
"gottesmm"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/34023",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
797203978 | [Build] Fix macOS version passed in -target parameter.
This code was added in commit b4368259480ed2c895e8c855bbfde20c1805acd1, but the macOS variable expansion had an extra "SWIFT_" at the front.
Among other potential problems, this causes classes in the stdlib to always be built with the legacy is-swift bit, which causes havoc when libswiftCore is in the shared cache where those bits aren't fixed up.
rdar://73767816
Karoy, am I right in understanding that the SWIFT_-prefixed versions are for the stdlib, runtime and overlays, whereas the non-SWIFT_ prefixed variables are for the host tools? In that case, I suspect what's going wrong is that the SWIFT_-prefixed variable is not being set correctly for macOS.
Karoy, am I right in understanding that the SWIFT_-prefixed versions are for the stdlib, runtime and overlays, whereas the non-SWIFT_ prefixed variables are for the host tools? In that case, I suspect what's going wrong is that the SWIFT_-prefixed variable is not being set correctly for macOS.
After further discussion with Varun, we think this may be fine, and I just ran into trouble due to some bad state in my build directory. I had to do a clean build to test this change.... I'm going to do another clean build without this supposed fix to double-check.
After further discussion with Varun, we think this may be fine, and I just ran into trouble due to some bad state in my build directory. I had to do a clean build to test this change.... I'm going to do another clean build without this supposed fix to double-check.
This was indeed a false alarm. Sorry if I got anybody worried!
This was indeed a false alarm. Sorry if I got anybody worried!
| gharchive/pull-request | 2021-01-29T21:01:41 | 2025-04-01T04:33:30.813944 | {
"authors": [
"mikeash",
"varungandhi-apple"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/35657",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
815691134 | [SIL Optimizer] Replace VarDeclUsageChecker AST Walker with DiagnoseVarUsage SIL Pass
SIL Diagnostics for Var Usage
This mandatory SIL optimization pass, DiagnoseVarUsage, replaces VarDeclUsageChecker as the system for generating usage diagnostics.
Why a SIL pass?
With access to the def-use graph, we have a more accurate picture of the lifetime of a variable declaration for more robust diagnostics. Ultimately, the former AST walker missed many cases, shown in the bugs below. Additionally, this implementation has many possibilities for future features.
Addresses Bugs:
Mutating Inouts
func increment() {
var x = 0 // no warning, but should warn x was mutated but never read.
x += 1
}
This bug (and similar bugs) are addressed by improving the way usage of inout parameters is diagnosed. If the function is local to the SIL module, we can follow the SILArgument to look for a read of the parameter. For external SILFunctions, an inout parameter is marked as used if the function produces some used output, or if the function has a closure parameter (where the inout argument may be used to produce other parameters within the closure).
Closures Missing Diagnostics
SR-11992, SR-13821
class Foo {
lazy var x: Int = {
var y = 42 // no warning, but should warn unused
var z = 42 // no warning, but should warn never mutated
return z
}()
}
var missesAllDiagsInClosures: () -> Int = {
var unused = 42 // no warning, but should warn unused
return 12
}
Bugs like this are addressed because we are no longer using the AST walker, which inconsistently missed certain nodes. Instead, we can see every SILInstruction that originates from a VarDecl and then decide whether to inspect it for diagnostics.
Misleading Unused Setter Param Diagnostics
var x : Int {
get { 0 }
set { }
}
x
// warning: setter argument 'newValue' was never used, but the property was accessed
// note: did you mean to use 'newValue' instead of accessing the property's current value?
These diagnostics are misleading because newValue is not in scope for the diagnostic location, and the setter function was never called. Now, a case like this does not produce diagnostics.
Cases like the following still produce diagnostics as expected.
var x : Int {
get { 0 }
set {
print(x)
// warning: setter argument 'newValue' was never used, but the property was accessed
// note: did you mean to use 'newValue' instead of accessing the property's current value?
}
}
For some more information on this decision, this gist details reasoning and a possible alternative diagnostic.
And many other similar cases are addressed...
After switching to the SIL implementation, we now get consistent diagnostics in many of these common cases.
Modifications and Limitations
Test files
This diagnostic occurs during the mandatory diagnostics SIL optimizations, and therefore the diagnostics are only emitted after errors are addressed in the file (because code with errors does not reach SILGen).
DISCLAIMER TO REVIEWERS
Nearly every test case regarding variable usage is touched by this PR. I apologize for the nightmare this review may entail.
If I removed expected warnings, it is because the file never reaches SIL optimizations. Rest assured, every case I edited is either now checked in SILOptimizer/diagnose_var_usage or moved from the original <testname>.swift to <testname>_usage.swift.
'_' Assignments for Constants
Constants have no getter function, so _ = x no longer counts as a use of the variable. This source code is invisible after SILGen.
let x = 1
_ = x
/* SIL is identical to: */
let x = 1
This is because the underscore assignments have no effect, and the value is never accessed. Therefore, diagnosing a constant is unused (even after an underscore assignment) is more correct because the value is never actually used.
Future work may be done to change SILGen to make underscore assignments visible to this diagnostic pass and removed in a SIL Optimization pass.
Underscore assignments are still useful for variables because the right-hand side may produce side effects from creating this value. For variables, because an actual function is called, there is still no warning for underscore assignments.
var y = 1
_ = y // this is still a read
TL;DR: we produce an 'unused' diagnostic if the rhs of an underscore assignment has no effects (accessing a constant value).
#if declarations
Because inactive clauses are not generated in SIL, we can not track usage here.
var x = 12 // we end up diagnosing this is never mutated
#if false
x = 1
#endif
print(x)
Future work may be done in SILGen to generate useful information about variable usage in inactive clauses if this is important. Or maybe some information can be inspected in the AST.
Otherwise, this current diagnostic might encourage Swift users to write code with better usage enforcement across all builds.
Tracking Value Usage
In this SIL pass, we track the use of a value, not the declaration. This means if we have multiple constants referencing the same value, we have no way of knowing which reference to the value we used:
func duplicateReference() {
let x = 12
let y = x
print(x) // we don't know if x or y was used here, because they are identical in SIL
}
To a small extent, we can check up through the AST to find if a user of the value corresponds to a certain reference, and we do this to at least know x was assigned to y.
To handle diagnostics for this situation, instead of generating possibly misleading usage diagnostics, we introduce a new diagnostic:
func duplicateReference() {
let x = 12 // note: initially referenced here
let y = x // warning: immutable value referenced by 'y' is already referenced by 'x'; consider removing it and replacing all uses of 'y' with 'x'}}
}
Future work may be done here to improve this diagnostic to include fix-its for replacement. Or, if this diagnostic is to be reverted, we would need to produce separate values for every variable declaration.
What currently looks like this:
debug_value %3 : $Int, let, name "x" // id: %4
debug_value %3 : $Int, let, name "y" // id: %5
Would need to do something like this:
debug_value %3 : $Int, let, name "x"
...
copy_value %3 to %4
debug_value %4 : $Int, let, name "y"
However, in the handful of times this pattern appears in the stlib, the suggestion of the diagnostic made more sense, was quick to address, and lead to more straight-forward code.
Tracking Value Usage - Exceptions
Capture Lists - redeclare an identical value, so we don't want to produce incorrect diagnostics here. We perform an explicit check for declarations belonging to capture lists to make sure this declaration of the value is captured.
Case Statements - binding declarations in switch statements may redeclare an identical value. Again, we explicitly check that a 'use' of the value is in SILBlocks accessible to the case statement.
Future Extensions
While the focus of this PR is to lay the groundwork to replace the previous AST Walker implementation, I noticed this SIL pass opens up plenty of further diagnostics.
Unused Parameter diags
See SR-2819
This pass already performs a check for the newValue parameter of setter functions, and the logic could easily be transferred to diagnosing other unused parameters or unmutated inout parameters. If the parameter is not required (such as when overriding a parent function), then these diagnostics could be helpful to simplify code.
Unused Local Functions
Using def-use traversals in a similar manner for functions inaccessible to outside modules can produce useful diagnostics.
Links
SR-11992 and SR-13821
rdar://59636530
Superceeds PR #34617
Thanks to @CodaFi and @augusto2112 for the brainstorming sessions and motivation!
This is really cool!
func increment() {
var x = 0 // no warning, but should warn x was mutated but never read.
x += 1
}
Well, only if it’s a built-in += do we know that. If I create a custom implementation that logs the previous value before incrementing by one, then the value is both read and mutated. That is, admittedly, far-fetched for +=, but not for arbitrary inout functions.
@xwu that's a great point. For custom inouts local to the SILModule, this code actually follows the argument to look for a read in the function body.
But your comment does make me think about custom functions that are declared externally, and you're right that we can't verify it was read.
I'd like to point out that if this inout function does not produce any used result, I'm not sure if the mutating function ever provided value to the use of the variable, and the warning might still make sense. But maybe this discussion is out of scope for this PR.
Perhaps, for now, I can change this to assume external inout functions read the parameter. How do you think we should approach this?
Constants have no getter function, so _ = x no longer counts as a use of the variable. This source code is invisible after SILGen.
let x = 1
_ = x
/* SIL is identical to: */
let x = 1
This is because the underscore assignments have no effect, and the value is never accessed.
While understandable, this does have some impact on the user experience, as one of the most obvious ways currently to silence the warning is to write _ = x, and it would be nice to preserve that for both mutable and immutable bindings.
Perhaps, for now, I can change this to assume external inout functions read the parameter. How do you think we should approach this?
Neat that you’re actually looking for a read in the function body where possible. I think it’s reasonable to scope the diagnostic so that it’s emitted only where the function body is available (either because it’s inlinable from the standard library or otherwise) or in the same module, and so you can be sure if there’s no read. That would cover a lot of scenarios, I think, without giving false positives.
@xwu okay, thank you. I will make and test the appropriate changes for this.
I think it may be interesting to expand upon this in the future. Down the line, it would be great to provide these diagnostics for purely mutating functions, like the built-in += operator.
Down the line, it would be great to provide these diagnostics for purely mutating functions, like the built-in += operator.
Well, the built-in operator implementations are inlinable if not transparent, so you should be able to look through those and make the diagnosis, no?
@xwu thanks for the conversation! I made updates to inspect usage through transparent functions, and I think this is a much better approach to these diagnostics!
Summary of changes
local or transparent functions are inspected for a read of inout parameters
other external functions will assume a read of inout parameters
local or transparent functions
Is it too expensive to look through any functions that have their function body serialized? I believe you can just use isSerialized() for that.
Okay, I am running tests for that right now. This is my first time working with SIL, so I never knew that was possible. Thanks for the suggestion, and I'll tag you with an update after I test it!
@xwu, this works! Thank you.
Sorry I just noticed your comment about underscore assignments, @xwu. I realize this is not great behavior for these diagnostics. This change won't affect function results on the rhs, it only affects if an immutable value is directly used.
let x = 0
_ = x // this is a problem
_ = x + 1 // ok
_ = someFunctionResult() // ok
I will look into modifying SILGen to produce a debug_value instruction for underscore assignments to mark this as a read, but I have no idea what level of effort this entails. Will look into this tomorrow.
Thanks again!
I will look into modifying SILGen to produce a debug_value instruction for underscore assignments to mark this as a read, but I have no idea what level of effort this entails. Will look into this tomorrow.
Admittedly, I don't think it's a huge deal since this is limited to constants; if it's that narrowly tailored, I don't know that many people are deliberately creating a binding and then not using it, but one thing I've learned is that if it's possible, someone is relying on it.
@xwu, I made a fix to look for a DiscardAssignmentExpr, restoring the expected functionality of underscore assignments.
We do collect enough information at this stage to know the discarded assignment produces no effects, so it may be interesting to add a diagnostic here in the future.
@swift-ci Please test
@swift-ci Please test source compatibility
@maustinstar Looking into callee function bodies sounds tricky because those functions might in turn call other functions, which might be recursive, etc. There is existing infrastructure in the SIL optimizer that could handle this if you decide to go down this route, but I wonder if its worth the added complexity. Also, it would be nice if we could avoid triggering deserialization of functions, since it can add a lot of unnecessary overhead at -Onone.
@slavapestov
might in turn call other functions, which might be recursive
If your concern is that we might find ourselves in an infinite loop of searching through recursive functions, this won't happen because instructions that have already been seen are not searched again.
Also, this def-use traversal will keep stepping into functions until a read is found. I don't know if performance is a huge concern here because this will only happen in the case that a variable gets used in multiple inout functions without ever reading the value.
There is existing infrastructure in the SIL optimizer that could handle this if you decide to go down this route.
Great! Can you give me some pointers for where to look and learn?
it would be nice if we could avoid triggering deserialization of functions
Can you explain a little further? In regards to inspecting callees, we only search serialized and locally defined SILFunctions.
Thank you for your feedback!
Build failed
Swift Test OS X Platform
Git Sha - bc4db0c92fc2309b76a19d3dac6cfd3164543ae2
To handle diagnostics for this situation, instead of generating possibly misleading usage diagnostics, we introduce a new diagnostic:
func duplicateReference() {
let x = 12 // note: initially referenced here
let y = x // warning: immutable value referenced by 'y' is already referenced by 'x'; consider removing it and replacing all uses of 'y' with 'x'}}
}
I like that you detect this; it seems very much in keeping with detecting unused bindings. One caveat and one nit:
There is a reasonable use case for this, so there should be some way to silence the warning at the point of use. Consider, for instance:
let physicalConstantA = /* some large number */
let physicalConstantB = physicalConstantA
Suppose physicalConstantA serves a different purpose and is used in different kinds of calculation than physicalConstantB; however, it is fixed to be the same value. With the warning above, I'm encouraged either to (A) refer to this value all the time as physicalConstantA, even where at the point of use this could be confusing because readers are accustomed to reasoning about a calculation in terms of physicalConstantB; or to (B) rewrite the large number twice, risking a second typo or the two values falling out of sync, which they should not.
In other words, the use case here would be analogous to type aliases, only in this case for values. I should be allowed to alias immutable values, and so in the case where this is not being done accidentally, there should be a way to shut off this warning. Usually, parens are offered as a way to silence warnings like this.
If this produces a concern about how downstream diagnostics, we could work out some way of referring to such "value aliases" in diagnostics (for example, an imaginary error that the value is too amazing to be used might read: 'x' (aka 'y') is too amazing to be used).
The wording of the diagnostic might be somewhat confusing. The user-facing language distinguishes "reference types" from "value types," and semantically, let y = x creates a second copy of the same value. Now, of course, users familiar with compiler internals know that this "copy" is entirely notional and always elided when you've got immutable bindings of a constant value. However, calling these bindings "references" and telling the user that they've "referenced" the same value twice would be confusing to a user who's working with the mental model that Swift explicitly teaches them to use.
Because inactive clauses are not generated in SIL, we can not track usage here.
var x = 12 // we end up diagnosing this is never mutated
#if false
x = 1
#endif
print(x)
Future work may be done in SILGen to generate useful information about variable usage in inactive clauses if this is important. Or maybe some information can be inspected in the AST.
Otherwise, this current diagnostic might encourage Swift users to write code with better usage enforcement across all builds.
Given the nature of the diagnostic in question, we should prefer false negatives (not showing the diagnostic when a variable isn't mutated) over false positives (showing the diagnostic incorrectly when a variable really is mutated).
Therefore--and I apologize if I misunderstand your text and this is already the behavior you've implemented--in the case where there's code that's conditionally not compiled, I think the diagnostic should just not be shown. This doesn't preclude the future work you mention where variable usage could be determined within these conditional compilation blocks, but until that time, I would fail open.
@xwu I really like that idea for using parenthesis to explicitly note a copy of the reference. I will look into this.
For #if statements, you are correct in assuming that the diagnostic is shown if a variable is used within the conditional compilation blocks.
let a = 0 // diagnoses never used
#if false
print(a)
#endif
The code for the inactive clause never goes through SILGen, so it's invisible to the SIL pass for diagnostics. So, I'm not sure how to find that a value is used in a case like this.
let a = 0 // ok
#if true
print(a)
#endif
This code does pass through SILGen, so we can see this use in the diagnostic pass.
The only way I can imagine doing this is by changing SILGen (perhaps create some debug instruction to inform us that a value is captured in an inactive clause). Unfortunately, I'm not familiar with the SILGen process, but I will try my best to figure this out. Any ideas here?
The only way I can imagine doing this is by changing SILGen (perhaps create some debug instruction to inform us that a value is captured in an inactive clause). Unfortunately, I'm not familiar with the SILGen process, but I will try my best to figure this out. Any ideas here?
This does sound like it'd be a bit of a tough thing to tackle. To be clear, I'm not suggesting that it needs to be done as part of this PR even though it'd be a very nice improvement later. I'm only suggesting that, for now, we don't diagnose unused variables in cases where there are conditional clauses (that is, let the presence of a conditional clause cause all in-scope variables to be considered "used" for now).
Leaving an update so this thread doesn't go stale:
The reason _ = foo produces no code is because SILGen peepholes it away. Perhaps we should just not do that, and allow dead code elimination to remove the discarded result after the diagnostics pass?
I started working on a change to SILGen to produce debug values for discarded assignment expressions. Otherwise, the use of the value is invisible by the time we get to this diagnostic pass. I am considering introducing this as an incremental change in a separate PR, since it seems other aspects of this project need further consideration and may go on for a long while.
In the meantime, I will convert this PR to a draft.
Thank you for the continuous feedback and support.
Update: was able to remove SILGen peephole for underscore assignments. This restores functionality for underscore assignments.
Tagging @varungandhi-apple, thanks for the help.
@xwu, the regressions we discussed earlier in this thread have now been addressed. I would appreciate if you could run the test suite on CI.
@swift-ci test
[Okay if other people who have studied this PR in more detail would like to overrule this but] Given the large scale of this change, I think there is a greater risk of things going wrong and that this may need to be reverted. Should that happen, the size of this PR creates a greater chance of merge conflicts, making reverting and relanding this work tricky.
If possible, I recommend breaking it up into at least 3 PRs, spaced out by 1-2 weeks:
PR 1. Add the new checker under a new frontend flag. Make sure any tests pass with both by default and with the new flag. (Alternately, this flag takes a boolean which is false by default). This is also a good opportunity to squash some of the WIP commits.
PR 2. Flip the default.
PR 3. Remove the old code and the flag.
Build failed
Swift Test Linux Platform
Git Sha - ba95da4d2738f5b4ca8c0dfedaca47d1e5d3405a
Build failed
Swift Test OS X Platform
Git Sha - ba95da4d2738f5b4ca8c0dfedaca47d1e5d3405a
Okay, thanks @varungandhi-apple, I'll split this up and add the flag
| gharchive/pull-request | 2021-02-24T17:26:42 | 2025-04-01T04:33:30.855889 | {
"authors": [
"maustinstar",
"slavapestov",
"swift-ci",
"varungandhi-apple",
"xwu"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/36134",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
948390520 | [Distributed] fix all distributed runtime tests; they are currently not running on CI
These have an issue on some specific CI configuration so they are currently disabled on CI.
They pass again all locally though:
Passed Tests (7):
Swift(macosx-x86_64) :: Distributed/Runtime/distributed_actor_remote_functions.swift
Swift(macosx-x86_64) :: Distributed/Runtime/distributed_actor_deinit.swift
Swift(macosx-x86_64) :: Distributed/Runtime/distributed_actor_isRemote.swift
Swift(macosx-x86_64) :: Distributed/Runtime/distributed_actor_dynamic_remote_func.swift
Swift(macosx-x86_64) :: Distributed/Runtime/distributed_no_transport_boom.swift
Swift(macosx-x86_64) :: Distributed/Runtime/distributed_actor_init_local.swift
Swift(macosx-x86_64) :: Distributed/Runtime/distributed_actor_local.swift
Tests blocked by rdar://78290608
@swift-ci please smoke test and merge
@swift-ci please smoke test and merge
@swift-ci please build toolchain
Linux Toolchain (Ubuntu 16.04)
Download Toolchain
Git Sha - 2fd09b32f2343064e4e9bfffb848ee8b1c4002a6
Install command
tar zxf swift-PR-38502-630-ubuntu16.04.tar.gz
More info
@swift-ci please smoke test macOS
macOS Toolchain
Download Toolchain
Git Sha - 2fd09b32f2343064e4e9bfffb848ee8b1c4002a6
Install command
tar -zxf swift-PR-38502-1059-osx.tar.gz --directory ~/
@swift-ci please smoke test Linux and merge
@swift-ci please smoke test linux platform
| gharchive/pull-request | 2021-07-20T07:50:11 | 2025-04-01T04:33:30.863105 | {
"authors": [
"ktoso",
"swift-ci"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/38502",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
192685364 | Removing more calls to ValueDecl::getType()
Almost ready to push it down to VarDecl.
@swift-ci Please smoke test
@swift-ci Please smoke test
@swift-ci Please smoke test
@swift-ci Please smoke test
@swift-ci Please smoke test
@swift-ci Please smoke test
@swift-ci Please smoke test
@swift-ci Please smoke test
This broke two iOS tests. Reverting in #6019.
Never mind, Slava pushed followups before my revert. We'll just fix it elsewhere.
| gharchive/pull-request | 2016-11-30T21:31:58 | 2025-04-01T04:33:30.866276 | {
"authors": [
"gparker42",
"slavapestov"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/5990",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
195670998 | Revert "[docs] DebuggingTheCompiler.rst => DebuggingTheCompiler.md."
This reverts commit 444920f2b67dea7350a6098b75078edf1be8b64f.
Caused the build failure: toctree contains reference to nonexisting document u'DebuggingTheCompiler'
@swift-ci Please smoke test and merge
Just an FYI, I spoke with Erik about this. The reason that this is being reverted is that there is a reference to this file in the main toc file. It errors since it is not there.
That being said, the main toc file is in *.rst and we are converting away from that. Perhaps the right solution here is to convert all of the files in the main toc file to markdown and then change the toc to markdown as well.
| gharchive/pull-request | 2016-12-14T23:03:14 | 2025-04-01T04:33:30.868370 | {
"authors": [
"eeckstein",
"gottesmm"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/6281",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1642843392 | DNM: Add -Rmodule-interface-rebuild to startOnMainActor
Add remark to see why we're rebuilding from the swiftinterface occasionally, instead of using the perfectly good swiftmodule for the stdlib.
This is starting to affect the Windows CI now. Hopefully I get lucky and can trigger the failure first try.
@swift-ci please test
We can merge if I don't catch it the first time so that we get some parallel data collection action going.
@swift-ci please smoke test
Windows failure:
FAIL: Swift(windows-x86_64) :: Concurrency/Runtime/startOnMainActor.swift (547 of 9197)
******************** TEST 'Swift(windows-x86_64) :: Concurrency/Runtime/startOnMainActor.swift' FAILED ********************
Script:
--
: 'RUN: at line 1'; rm -rf "T:\swift\test-windows-x86_64\Concurrency\Runtime\Output\startOnMainActor.swift.tmp" && mkdir -p "T:\swift\test-windows-x86_64\Concurrency\Runtime\Output\startOnMainActor.swift.tmp"
: 'RUN: at line 2'; 't:\\swift\\bin\\swiftc.exe' -target x86_64-unknown-windows-msvc -vfsoverlay T:/swift\stdlib\windows-vfs-overlay.yaml -swift-version 4 -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 9999:macOS 9999, iOS 9999, watchOS 9999, tvOS 9999' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.0:macOS 10.14.4, iOS 12.2, watchOS 5.2, tvOS 12.2' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.1:macOS 10.15, iOS 13.0, watchOS 6.0, tvOS 13.0' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.2:macOS 10.15.4, iOS 13.4, watchOS 6.2, tvOS 13.4' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.3:macOS 11.0, iOS 14.0, watchOS 7.0, tvOS 14.0' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.4:macOS 11.3, iOS 14.5, watchOS 7.4, tvOS 14.5' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.5:macOS 12.0, iOS 15.0, watchOS 8.0, tvOS 15.0' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.6:macOS 12.3, iOS 15.4, watchOS 8.5, tvOS 15.4' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.7:macOS 13.0, iOS 16.0, watchOS 9.0, tvOS 16.0' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.8:macOS 13.3, iOS 16.4, watchOS 9.4, tvOS 16.4' -Xfrontend -define-availability -Xfrontend 'SwiftStdlib 5.9:macOS 9999, iOS 9999, watchOS 9999, tvOS 9999' -libc MD -Xfrontend -Rmodule-interface-rebuild -Xfrontend -disable-availability-checking C:\Users\swift-ci\jenkins\workspace\swift-PR-windows\swift\test\Concurrency\Runtime\startOnMainActor.swift -o T:\swift\test-windows-x86_64\Concurrency\Runtime\Output\startOnMainActor.swift.tmp/a.out 1>&2
: 'RUN: at line 3'; echo T:\swift\test-windows-x86_64\Concurrency\Runtime\Output\startOnMainActor.swift.tmp/a.out
: 'RUN: at line 4'; T:\swift\test-windows-x86_64\Concurrency\Runtime\Output\startOnMainActor.swift.tmp/a.out
--
Exit Code: 127
Command Output (stdout):
--
$ ":" "RUN: at line 1"
$ "rm" "-rf" "T:\swift\test-windows-x86_64\Concurrency\Runtime\Output\startOnMainActor.swift.tmp"
$ "mkdir" "-p" "T:\swift\test-windows-x86_64\Concurrency\Runtime\Output\startOnMainActor.swift.tmp"
$ ":" "RUN: at line 2"
$ "t:\\swift\\bin\\swiftc.exe" "-target" "x86_64-unknown-windows-msvc" "-vfsoverlay" "T:/swift\stdlib\windows-vfs-overlay.yaml" "-swift-version" "4" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 9999:macOS 9999, iOS 9999, watchOS 9999, tvOS 9999" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.0:macOS 10.14.4, iOS 12.2, watchOS 5.2, tvOS 12.2" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.1:macOS 10.15, iOS 13.0, watchOS 6.0, tvOS 13.0" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.2:macOS 10.15.4, iOS 13.4, watchOS 6.2, tvOS 13.4" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.3:macOS 11.0, iOS 14.0, watchOS 7.0, tvOS 14.0" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.4:macOS 11.3, iOS 14.5, watchOS 7.4, tvOS 14.5" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.5:macOS 12.0, iOS 15.0, watchOS 8.0, tvOS 15.0" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.6:macOS 12.3, iOS 15.4, watchOS 8.5, tvOS 15.4" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.7:macOS 13.0, iOS 16.0, watchOS 9.0, tvOS 16.0" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.8:macOS 13.3, iOS 16.4, watchOS 9.4, tvOS 16.4" "-Xfrontend" "-define-availability" "-Xfrontend" "SwiftStdlib 5.9:macOS 9999, iOS 9999, watchOS 9999, tvOS 9999" "-libc" "MD" "-Xfrontend" "-Rmodule-interface-rebuild" "-Xfrontend" "-disable-availability-checking" "C:\Users\swift-ci\jenkins\workspace\swift-PR-windows\swift\test\Concurrency\Runtime\startOnMainActor.swift" "-o" "T:\swift\test-windows-x86_64\Concurrency\Runtime\Output\startOnMainActor.swift.tmp/a.out"
# command stderr:
Unsupported redirect: (('>&', 1), '2')
error: command failed with exit status: 127
--
| gharchive/pull-request | 2023-03-27T21:32:12 | 2025-04-01T04:33:30.872080 | {
"authors": [
"etcwilde"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/64653",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1692823975 | [cxx-interop] Update a test on Linux
libstdc++ currently cannot be split into submodules due to the way some types are defined, e.g. there are multiple headers that define size_t, so those headers must be a single (sub-)module.
@swift-ci please smoke test
| gharchive/pull-request | 2023-05-02T17:37:44 | 2025-04-01T04:33:30.873340 | {
"authors": [
"egorzhdan"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/65588",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1860559856 | [5.9] [Macros] Use outermost source file for establishing (file)private access scope
Fixes rdar://114048069. Cherry-picked from https://github.com/apple/swift/pull/68057
@swift-ci please test
| gharchive/pull-request | 2023-08-22T05:04:19 | 2025-04-01T04:33:30.874680 | {
"authors": [
"DougGregor"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/68058",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1903803379 | [5.10][build] Move libdispatch earlier in the build so it can be used by the compiler validation suite
Cherrypick of #65829 and #68565
Explanation: The libdispatch and several concurrency tests in the validation suite were inadvertently disabled a couple years ago. This enables them again by making sure the target libdispatch is built first and using the new dispatch-vfs-overlay.yaml it now generates.
Scope: Enabling compiler tests again
Issue: #53973
Risk: negative, as it is only testing more code, thus lowering risk
Testing: Passes all CI on trunk for the last couple days, and I've been doing this manually on Android for years without a problem.
Reviewer: @gottesmm
@bnbarham, please run the CI on this.
@swift-ci please test
@gottesmm, should I try to enable these for the 5.9 branch also?
@gottesmm, should I try to enable these for the 5.9 branch also?
There should be very few changes going to 5.9 at this point, so probably not worth it in my mind.
There should be very few changes going to 5.9 at this point, so probably not worth it in my mind.
@bnbarham, this changes zero behavior in the toolchain though, only slightly modifying the build order and some tests to enable several tests on the CI again. That's why I thought it might be worth it for 5.9 too, as it might catch 5.9 regressions better by having several dozen more tests run.
Pinging @tbkka, ready for review.
Ping @DougGregor, tiny pull that re-enables several tests on the linux CI.
Ping @bnbarham, can you review and merge? I didn't know you could get pulls into 5.10.
| gharchive/pull-request | 2023-09-19T22:19:10 | 2025-04-01T04:33:30.880301 | {
"authors": [
"bnbarham",
"finagolfin"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/68640",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
201903791 | Remove dead witness tables, including dead witness functions.
A witness table is dead if it is not used outside the module (private/internal) and it’s not used by any instruction or other witness table in the module.
Also the meta-type of the conforming type must not escape, because it’s possible to test any opaque type if it conforms to a protocol.
This gives about 2.5% code size reduction on stdlib.
Also some benchmarks see code size reductions, but this depends on the benchmark. For most benchmarks (especially our micro-benchmarks, where we don't define any types) there is no improvement. For some larger benchmarks there are significant size reductions. On average it's 2.4%.
@swift-ci Please test
Build failed
Jenkins build - Swift Test Linux Platform
Git Commit - a7782567e8a28829dc4b6689d0912809a457d857
Test requested by - @eeckstein
Build failed
Jenkins build - Swift Test OS X Platform
Git Commit - a7782567e8a28829dc4b6689d0912809a457d857
Test requested by - @eeckstein
@swift-ci Please test
Build failed
Jenkins build - Swift Test Linux Platform
Git Commit - a7782567e8a28829dc4b6689d0912809a457d857
Test requested by - @eeckstein
Build failed
Jenkins build - Swift Test OS X Platform
Git Commit - a7782567e8a28829dc4b6689d0912809a457d857
Test requested by - @eeckstein
@swift-ci Please test linux
Build failed
Jenkins build - Swift Test Linux Platform
Git Commit - 44f35b916089a46042368b535ece4805f063ce64
Test requested by - @eeckstein
"Also the meta-type of the conforming type must not escape"
What if an instance of the type escapes?
Depends. An instance could escape into a concrete context that can handle the value without metadata. Anywhere we need to escape an instance into a context that also needs the type metadata to work with the instance, that should be tracked independently.
It's fuzzy but I can imagine this:
internal class Impl: SomeProto { /* … */ }
public func produce() -> AnyObject { return Impl() }
public func consume(_ value: AnyObject) {
guard let interestingValue = value as? SomeProto else { return }
// …
}
I'm pretty sure in this case the witness table is never referenced statically and the metatype does not escape (because we're using AnyObject rather than Any), but we still need the witness table at run-time dynamically.
What if an instance of the type escapes?
@jrose-apple The question is if someone outside the translation unit can dynamically check if a value conforms to a protocol. This is done by looking up the metatype of that value in the protocol_conformances section. So if the metatype does not escape we are good.
@swift-ci Please test linux
Why does it matter whether the lookup happens outside the translation unit?
I'm pretty sure in this case the witness table is never referenced statically and the metatype does not escape (because we're using AnyObject rather than Any), but we still need the witness table at run-time dynamically.
The metatype does escape because it's stored in the Impl() instance header.
Why does it matter whether the lookup happens outside the translation unit?
It doesn't. My mistake
Ah, right. I didn't think about the metatype being either passed as a constructor argument or directly stored in the type.
| gharchive/pull-request | 2017-01-19T16:12:33 | 2025-04-01T04:33:30.890821 | {
"authors": [
"eeckstein",
"jckarter",
"jrose-apple",
"swift-ci"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/6912",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1961967921 | [swift-vfe][swift-wme] Stop disabling function mergers for Swift VFE / WME
The Swift and LLVM function mergers were disabled when Swift VFE or WME are enabled because the function merger did not respect metadata on calls to llvm.type.checked.load. This is no longer the case, so we can turn these passes back on.
https://github.com/apple/llvm-project/pull/7679 @swift-ci Please test
Linux failures look unrelated to me.
https://github.com/apple/llvm-project/pull/7679 @swift-ci test linux
@eeckstein Do you mind merging / testing the LLVM side of things? https://github.com/apple/llvm-project/pull/7679. I can merge Swift.
I have merged the LLVM PR
@NuriAmari feel free to merge
| gharchive/pull-request | 2023-10-25T18:11:25 | 2025-04-01T04:33:30.894340 | {
"authors": [
"NuriAmari",
"aschwaighofer",
"eeckstein"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/69407",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
212182421 | Revert "[build-script-impl] Add support for building benchmarks on linux but disable it in all currently defined presets for linux."
Reverts apple/swift#7860
This broke the OSS - LLDB incremental linux bots.
@swift-ci please smoke test
Thanks!
On Mar 6, 2017, at 9:39 AM, Ben Langmuir notifications@github.com wrote:
Merged #7931 https://github.com/apple/swift/pull/7931.
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub https://github.com/apple/swift/pull/7931#event-988152162, or mute the thread https://github.com/notifications/unsubscribe-auth/AAee35h43uQLRgkYGnYc6h8N_V9HrZl-ks5rjETagaJpZM4MUZy7.
| gharchive/pull-request | 2017-03-06T16:50:49 | 2025-04-01T04:33:30.897434 | {
"authors": [
"benlangmuir",
"gottesmm"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/7931",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
213498021 | [Constraint solver] More updates for using the constraint solver type…
… map.
Cache types in a few more places and ensure that we're reading from the map in another spot (but interestingly stop reading from it in a different spot - this is going to need another round of clean-ups at some point on some of these Expr APIs that like to read types from other nodes).
@swift-ci Please smoke test and merge
| gharchive/pull-request | 2017-03-11T02:23:46 | 2025-04-01T04:33:30.899344 | {
"authors": [
"rudkx"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/8029",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
221115050 | [GSB] Delay requirements on recursive types
Delaying requirements that form new, recursively-defined types, providing an alternative to the current detect-recursion-complain-and-bail approach that bans recursive protocol requirements entirely. This is along the path toward recursive protocol conformances (which would require us to reprocess those requirements when we need to later), and eliminates some extraneous bookkeeping in the GenericSignatureBuilder.
Go, Doug!
@swift-ci please smoke test and merge
@swift-ci please smoke test and merge
@swift-ci please smoke test Linux
LLDB failure. Ugh.
@swift-ci please smoke test Linux
patched up the LLDB issue...
@swift-ci please smoke test Linux
| gharchive/pull-request | 2017-04-12T00:07:41 | 2025-04-01T04:33:30.902104 | {
"authors": [
"DougGregor",
"dabrahams"
],
"repo": "apple/swift",
"url": "https://github.com/apple/swift/pull/8714",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
313649126 | Plot objects display twice in Jupyter Notebook
The plot object (when yielded as the output value of a cell) shows twice.
Repro steps (in Jupyter Notebook):
import turicreate as tc
sf = tc.SFrame.read_csv('https://docs-assets.developer.apple.com/turicreate/datasets/tc-clang-format-results.csv')
tc.visualization.scatter(sf['Lines Added'], sf['Lines Removed'])
Maybe this has been fixed, but it doesn't seem to happen for me, though I am using JupyterLab, not the older Notebook interface.
Cannot repro.
| gharchive/issue | 2018-04-12T09:36:44 | 2025-04-01T04:33:30.903667 | {
"authors": [
"abhishekpratapa",
"theavey",
"znation"
],
"repo": "apple/turicreate",
"url": "https://github.com/apple/turicreate/issues/520",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
708538149 | New pip dependency resolver
Add --use-feature=2020-resolver to our Python build/test setup to ensure we stay compatible with the new dependency resolver.
Also in our Python build/test setup, pin TensorFlow version (for a few Python versions and OS combinations). Without this one of our packages used for unit testing causes a SciPy version conflict.
Internal tests pass
| gharchive/pull-request | 2020-09-24T23:39:58 | 2025-04-01T04:33:30.905169 | {
"authors": [
"TobyRoseman"
],
"repo": "apple/turicreate",
"url": "https://github.com/apple/turicreate/pull/3328",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
604774598 | deprecate Event.last and Span.last
Event.last => ao.lastEvent
Span.last => ao.lastSpan
the base of this branch is p9p-noti+minimal; to minimize conflicts this should be reviewed and merged only after PR 117 has been merged. there will be 20 files changed after that. other than the deprecation code itself, all the changes are to replace last = Event.last and {last} = Event with last = ao.lastEvent, and analogous changes for Span.last
good catch on the api - it generate it automatically from the sources but it does nedd to be updated. i'll add that to a subsequent PR.
| gharchive/pull-request | 2020-04-22T13:50:43 | 2025-04-01T04:33:30.913680 | {
"authors": [
"bmacnaughton"
],
"repo": "appoptics/appoptics-apm-node",
"url": "https://github.com/appoptics/appoptics-apm-node/pull/118",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
475783724 | Transaction logs may be reporting illegal actions
When running a batch training learning curves sometimes show more opportunities than are possible in the problem set. This seems to be from times that agents gets a step A correct, then does some step B, then attempts to go back to attempt step A, which is marked as incorrect. In a human using CTAT this would never happen because if you try to take a previously correct steps the field will be locked and so it will never appear. Unsure how we want to handle this long term, because its kind of interesting, but in the short term it violates the assumptions of transaction roll ups for learning curve analysis.
It is possible this is desirable behavior and the rollup process we were using from PyAFM is actually wrong.
We are pretty sure this is an issue related to a particular tutor and not the framework itself. Closing for now unless it comes up again.
| gharchive/issue | 2019-08-01T16:42:35 | 2025-04-01T04:33:30.961012 | {
"authors": [
"eharpste"
],
"repo": "apprenticelearner/AL_HTML",
"url": "https://github.com/apprenticelearner/AL_HTML/issues/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1238148854 | [Bug]: Very long values in APPSMITH_ENCRYPTION_PASSWORD and APPSMITH_ENCRYPTION_SALT makes it skip these variables
Is there an existing issue for this?
[X] I have searched the existing issues
Description
A user flagged this issue.
It seems like that very long values for APPSMITH_ENCRYPTION_PASSWORD and APPSMITH_ENCRYPTION_SALT break it somehow to skip email and signup variables.
Now I use shorter ones and everything is working again.
Steps To Reproduce
Needs to be investigated and reproduced
https://discord.com/channels/725602949748752515/975730370689499136/975740299005558784
Public Sample App
No response
Version
v1.6.24
@sub1983 @sum35h could one of you please have a look?
| gharchive/issue | 2022-05-17T06:42:09 | 2025-04-01T04:33:30.972336 | {
"authors": [
"Nikhil-Nandagopal",
"danciaclara"
],
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/issues/13889",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1270335943 | [Task]: Redesign action selector
Is there an existing issue for this?
[X] I have searched the existing issues
SubTasks
[ ] PRD
[ ] Solution exploration
[ ] Refinement
[ ] Feedback
[ ] Handoff
Estimated due date: June 24th, 2022
| gharchive/issue | 2022-06-14T06:49:28 | 2025-04-01T04:33:30.974387 | {
"authors": [
"ayushi-appsmith"
],
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/issues/14515",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1747205325 | [Bug]: Appsmith server is taking too long to respond
Is there an existing issue for this?
[X] I have searched the existing issues
Description
Installed on Ubuntu 20.04
sudo docker ps
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
76284bf62841 appsmith/appsmith-ce:latest "/opt/appsmith/entry…" 23 hours ago Up 23 hours (healthy) 0.0.0.0:80->80/tcp, :::80->80/tcp, 0.0.0.0:443->443/tcp, :::443->443/tcp appsmith
d7a2e3340eed containrrr/watchtower "/watchtower --sched…" 41 hours ago Up 40 hours 8080/tcp appsmith_auto_update_1
When I try to edit my first App
Logs:
editor stdout | 10.96.26.140 - - [08/Jun/2023:06:51:48 +0000] "GET /api/v1/pages?pageId=647f441bb9009d5e1ccb6841&mode=EDIT HTTP/1.1" 200 497 "http://10.96.32.161/app/my-first-application/page1-647f441bb9009d5e1ccb6841/edit" "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36 OPR/99.0.0.0" editor stdout | rts stdout | ns:/# socket _uR-aKx-o11UwoKEAAAB joined the room app:647f441bb9009d5e1ccb683e rts stdout | redis stdout | 1676:M 08 Jun 2023 06:51:48.280 * Background saving terminated with success redis stdout | backend stdout | [2023-06-08 06:51:48,320] userEmail=paolo@ftgm.it, sessionId=b80a75a0-0c6e-4a3e-9eed-889424b3bdea, thread=lettuce-epollEventLoop-6-1, requestId=2350f68b-1c57-4481-8a1c-44e5910607ea - Going to get all unpublished action collections with params: {applicationId=[647f441bb9009d5e1ccb683e]}, branch: null backend stdout | editor stdout | 10.96.26.140 - - [08/Jun/2023:06:51:48 +0000] "GET /api/v1/collections/actions?applicationId=647f441bb9009d5e1ccb683e HTTP/1.1" 200 101 "http://10.96.32.161/app/my-first-application/page1-647f441bb9009d5e1ccb6841/edit" "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36 OPR/99.0.0.0" editor stdout | editor stdout | 10.96.26.140 - - [08/Jun/2023:06:51:48 +0000] "GET /api/v1/themes/applications/647f441bb9009d5e1ccb683e/current?mode=EDIT HTTP/1.1" 200 1442 "http://10.96.32.161/app/my-first-application/page1-647f441bb9009d5e1ccb6841/edit" "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36 OPR/99.0.0.0" editor stdout | editor stdout | 10.96.26.140 - - [08/Jun/2023:06:51:48 +0000] "GET /api/v1/pages/647f441bb9009d5e1ccb6841 HTTP/1.1" 200 518 "http://10.96.32.161/app/my-first-application/page1-647f441bb9009d5e1ccb6841/edit" "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36 OPR/99.0.0.0" editor stdout | rts stdout | ns:/# socket _uR-aKx-o11UwoKEAAAB left the room app:647f441bb9009d5e1ccb683e rts stdout | editor stdout | 10.96.26.140 - - [08/Jun/2023:06:51:48 +0000] "GET /api/v1/themes/applications/647f441bb9009d5e1ccb683e HTTP/1.1" 200 4624 "http://10.96.32.161/app/my-first-application/page1-647f441bb9009d5e1ccb6841/edit" "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36 OPR/99.0.0.0" editor stdout | rts stdout | ns:/# socket _uR-aKx-o11UwoKEAAAB joined the room app:647f441bb9009d5e1ccb683e rts stdout | backend stdout | [2023-06-08 06:51:48,397] userEmail=paolo@ftgm.it, sessionId=b80a75a0-0c6e-4a3e-9eed-889424b3bdea, thread=lettuce-epollEventLoop-6-1, requestId=4910e92d-c69a-4077-bf8b-da19a27cc71d - Going to get snapshot with application id: 647f441bb9009d5e1ccb683e, branch: null backend stdout | backend stdout | [2023-06-08 06:51:48,404] userEmail=paolo@ftgm.it, sessionId=b80a75a0-0c6e-4a3e-9eed-889424b3bdea, thread=nioEventLoopGroup-3-5, requestId=4910e92d-c69a-4077-bf8b-da19a27cc71d - backend stdout | com.appsmith.server.exceptions.AppsmithException: Unable to find application 647f441bb9009d5e1ccb683e backend stdout | at com.appsmith.server.services.ce.ApplicationSnapshotServiceCEImpl.getWithoutDataByApplicationId(ApplicationSnapshotServiceCEImpl.java:73) backend stdout | Suppressed: reactor.core.publisher.FluxOnAssembly$OnAssemblyException: backend stdout | Error has been observed at the following site(s): backend stdout | *__checkpoint ⇢ Handler com.appsmith.server.controllers.ApplicationController#getSnapshotWithoutApplicationJson(String, String) [DispatcherHandler] backend stdout | Original Stack Trace: backend stdout | at com.appsmith.server.services.ce.ApplicationSnapshotServiceCEImpl.getWithoutDataByApplicationId(ApplicationSnapshotServiceCEImpl.java:73) backend stdout | at com.appsmith.server.controllers.ce.ApplicationControllerCE.getSnapshotWithoutApplicationJson(ApplicationControllerCE.java:216) backend stdout | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) backend stdout | at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) backend stdout | at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) backend stdout | at java.base/java.lang.reflect.Method.invoke(Method.java:568) backend stdout | at org.springframework.web.reactive.result.method.InvocableHandlerMethod.lambda$invoke$0(InvocableHandlerMethod.java:145) backend stdout | at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:132) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onNext(MDCConfig.java:53) backend stdout | at reactor.core.publisher.MonoZip$ZipCoordinator.signal(MonoZip.java:293) backend stdout | at reactor.core.publisher.MonoZip$ZipInner.onNext(MonoZip.java:474) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onNext(MDCConfig.java:53) backend stdout | at reactor.core.publisher.MonoPeekTerminal$MonoTerminalPeekSubscriber.onNext(MonoPeekTerminal.java:180) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onNext(MDCConfig.java:53) backend stdout | at reactor.core.publisher.Operators$BaseFluxToMonoOperator.completePossiblyEmpty(Operators.java:2071) backend stdout | at reactor.core.publisher.FluxDefaultIfEmpty$DefaultIfEmptySubscriber.onComplete(FluxDefaultIfEmpty.java:134) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onComplete(MDCConfig.java:63) backend stdout | at reactor.core.publisher.FluxSwitchIfEmpty$SwitchIfEmptySubscriber.onComplete(FluxSwitchIfEmpty.java:85) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onComplete(MDCConfig.java:63) backend stdout | at reactor.core.publisher.MonoSupplier$MonoSupplierSubscription.request(MonoSupplier.java:148) backend stdout | at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.set(Operators.java:2341) backend stdout | at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.onSubscribe(Operators.java:2215) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onSubscribe(MDCConfig.java:47) backend stdout | at reactor.core.publisher.MonoSupplier.subscribe(MonoSupplier.java:48) backend stdout | at reactor.core.publisher.Mono.subscribe(Mono.java:4485) backend stdout | at reactor.core.publisher.FluxSwitchIfEmpty$SwitchIfEmptySubscriber.onComplete(FluxSwitchIfEmpty.java:82) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onComplete(MDCConfig.java:63) backend stdout | at reactor.core.publisher.Operators.complete(Operators.java:137) backend stdout | at reactor.core.publisher.FluxFlatMap.trySubscribeScalarMap(FluxFlatMap.java:145) backend stdout | at reactor.core.publisher.MonoFlatMap.subscribeOrReturn(MonoFlatMap.java:53) backend stdout | at reactor.core.publisher.InternalMonoOperator.subscribe(InternalMonoOperator.java:57) backend stdout | at reactor.core.publisher.MonoZip$ZipCoordinator.request(MonoZip.java:216) backend stdout | at reactor.core.publisher.MonoFlatMap$FlatMapMain.request(MonoFlatMap.java:194) backend stdout | at reactor.core.publisher.MonoIgnoreThen$ThenIgnoreMain.onSubscribe(MonoIgnoreThen.java:134) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onSubscribe(MDCConfig.java:47) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onSubscribe(MDCConfig.java:47) backend stdout | at reactor.core.publisher.MonoFlatMap$FlatMapMain.onSubscribe(MonoFlatMap.java:117) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onSubscribe(MDCConfig.java:47) backend stdout | at reactor.core.publisher.MonoZip.subscribe(MonoZip.java:125) backend stdout | at reactor.core.publisher.InternalMonoOperator.subscribe(InternalMonoOperator.java:64) backend stdout | at reactor.core.publisher.MonoDefer.subscribe(MonoDefer.java:52) backend stdout | at reactor.core.publisher.InternalMonoOperator.subscribe(InternalMonoOperator.java:64) backend stdout | at reactor.core.publisher.MonoIgnoreThen$ThenIgnoreMain.subscribeNext(MonoIgnoreThen.java:240) backend stdout | at reactor.core.publisher.MonoIgnoreThen$ThenIgnoreMain.onComplete(MonoIgnoreThen.java:203) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onComplete(MDCConfig.java:63) backend stdout | at reactor.core.publisher.MonoFlatMap$FlatMapMain.onComplete(MonoFlatMap.java:189) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onComplete(MDCConfig.java:63) backend stdout | at reactor.core.publisher.Operators.complete(Operators.java:137) backend stdout | at reactor.core.publisher.MonoZip.subscribe(MonoZip.java:121) backend stdout | at reactor.core.publisher.Mono.subscribe(Mono.java:4485) backend stdout | at reactor.core.publisher.MonoIgnoreThen$ThenIgnoreMain.subscribeNext(MonoIgnoreThen.java:263) backend stdout | at reactor.core.publisher.MonoIgnoreThen.subscribe(MonoIgnoreThen.java:51) backend stdout | at reactor.core.publisher.InternalMonoOperator.subscribe(InternalMonoOperator.java:64) backend stdout | at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:165) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onNext(MDCConfig.java:53) backend stdout | at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:79) backend stdout | at com.appsmith.server.configurations.MDCConfig$MdcContextLifter.onNext(MDCConfig.java:53)
Steps To Reproduce
login
go to first app
Public Sample App
No response
Environment
Deploy Preview
Issue video log
No response
Version
Self hosted with docker and autoupdate
@marcheschi When do encounter this error? Is it during installation or when you try to open some app?
Also, please use this document to provide detailed container logs - https://docs.appsmith.com/learning-and-resources/how-to-guides/how-to-get-container-logs#docusaurus_skipToContent_fallback
@marcheschi, sorry you're facing trouble here. Can you confirm the system capacity of the server you're running with? Like vCPU count, available memory etc., and if there's any other applications also running on this server, along with Appsmith please? Also, please share full logs as indicated in prev. message. Thanks!
Thank you @pranavkanade @sharat87
I used a dedicated VM with 8 CPU and 16 GB of ram with 20 GB of disk
appsmith is the only app on the server.
The installation runs fine, and I can login
but when I try to edit my first app appears the message and the editor is not shown.
Logs:
https://drive.google.com/file/d/1rAleXljvo0edsNP8c8GPWarirWCt35ev/view?usp=sharing
@marcheschi, thank you for the details. It looks like your server isn't connected to the Internet. Appsmith is unable to reach cs.appsmith.com, which is required for it's operation. More details about this in the callout at https://docs.appsmith.com/getting-started/setup#self-hosting-benefits.
Can you please confirm if your server has access to reach external endpoints in the Internet? If not, can we allow that access and restart Appsmith please?
The server has internet access, via proxy .
It does not work either.
I tried to recreate the container twice.
Too difficult and is a nonsense it needs a connection to the internet.
I'm sorry, Thank you for your assistance.
Hey @marcheschi, sorry you feel that way. We need Internet access to make some features possible, and some others convenient. I hope you managed to resolve the issue, if not, we can help you out if you can share the logs and what's failing now. Thank you!
Thanks @marcheschi for confirming. We actually have a provision to allow you to configure proxy in upper or lower case. We accept both. I'm wondering if uppercase env variables wasn't working for you? 🤔
Relevant code is in this function: https://github.com/appsmithorg/appsmith/blob/c4b9e272dbc0b08f307931c71c24e897435975a1/deploy/docker/entrypoint.sh#L97-L113
| gharchive/issue | 2023-06-08T06:55:52 | 2025-04-01T04:33:30.989645 | {
"authors": [
"marcheschi",
"pranavkanade",
"sharat87"
],
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/issues/24231",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1906454455 | [Bug]: Unable to import attached json file in the airgapped image
Is there an existing issue for this?
[X] I have searched the existing issues
Description
When I try to import the attached application, I get an error to the effect unable to connect to server.
Steps To Reproduce
Take the pull of the airgapped image, and set it up without internet
Try to import the attached file [rename it to json]
Note that import does not work and error pops up.
OldApp_DSTesting1.9.24.txt
Note: The attached app has Airtable / GSheets and other SAAS plugins. Since Airgapped does not support these plugins, if thats the reason for this failure, then an appropriate message needs to be given - or there should be some way of handling that portion in the import modal.
Public Sample App
No response
Environment
Release
Issue video log
No response
Version
Self Hosted
It's because of the unsupported DS
In the current airgapped regression also we are seeing the above issue, and also, sometimes we see a broken UI, which is not such a great user experience. Though it is a case with unsupported datasources, it should be handled gracefully.
https://github.com/appsmithorg/appsmith/assets/101863839/036d9f61-a8c9-486d-8031-2c03d2341095
@btsgh - GSheets is not supported and should not exist in Airgapped apps. Any failues because of that can be ignored.
| gharchive/issue | 2023-09-21T08:52:43 | 2025-04-01T04:33:30.995660 | {
"authors": [
"btsgh",
"infinitetrooper"
],
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/issues/27539",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2403654993 | [Bug]: Cross referencing - environment change does not auto refresh the response for awslambda ds.
Is there an existing issue for this?
[X] I have searched the existing issues
Description
In the cross referencing DP https://ee-4056.dp.appsmith.com/, for awslambda datasource, changing the environment does not refresh the response section automatically
Steps To Reproduce
https://jam.dev/c/45ace210-8d31-4225-9401-8c1c17f68b3b
Create an awslambda datasource in a workspace with custom environments
Create new query and run it against prod environment
Now change the environment to say staging or any custom env. The response tab does not get refreshed until we click Run manually
Now change it back to production, and here also note that the response tab does not get refreshed and show data until we click run manually
Public Sample App
No response
Environment
Deploy Preview
Severity
Medium (Frustrating UX)
Issue video log
No response
Version
Cloud
This is the behavior even on release if no widget is bound. Looks like this is expected behavior.
| gharchive/issue | 2024-07-11T17:03:29 | 2025-04-01T04:33:31.000022 | {
"authors": [
"btsgh"
],
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/issues/34888",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2434928432 | [Bug]: MySql datasource connection error - Authentication failed
Is there an existing issue for this?
[X] I have searched the existing issues
Description
I deploy appsmith in my local env and have running a long time stably.
Recently, I create a new connection to mysql and display the error:
Error Message: Authentication failed
Report Class: o.m.r.message.flow.AuthenticationFlow
Stack:
j.l.IndexOutOfBoundsException: readerIndex(63) + length(1) exceeds writerIndex(63): PooledSlicedByteBuf(ridx: 63, widx: 63, cap: 63/63, unwrapped: PooledUnsafeDirectByteBuf(ridx: 66, widx: 66, cap: 2048))
at i.n.b.AbstractByteBuf.checkReadableBytes0(AbstractByteBuf.java:1442)
at i.n.b.AbstractByteBuf.checkReadableBytes(AbstractByteBuf.java:1428)
at i.n.b.AbstractByteBuf.skipBytes(AbstractByteBuf.java:971)
at o.m.r.m.s.InitialHandshakePacket.decode(InitialHandshakePacket.java:100)
at o.m.r.c.DecoderState$1.decode(DecoderState.java:25)
at o.m.r.c.MariadbFrameDecoder.decode(MariadbFrameDecoder.java:104)
at o.m.r.c.MariadbFrameDecoder.decode(MariadbFrameDecoder.java:93)
at i.n.h.c.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:529)
at i.n.h.c.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:468)
... 15 common frames omitted
Wrapped by: i.n.h.c.DecoderException: java.lang.IndexOutOfBoundsException: readerIndex(63) + length(1) exceeds writerIndex(63): PooledSlicedByteBuf(ridx: 63, widx: 63, cap: 63/63, unwrapped: PooledUnsafeDirectByteBuf(ridx: 66, widx: 66, cap: 2048))
at i.n.h.c.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:499)
at i.n.h.c.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:290)
at i.n.c.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:444)
... 13 common frames omitted
Wrapped by: i.r.s.R2dbcNonTransientResourceException: Connection error
at o.m.r.c.SimpleClient.handleConnectionError(SimpleClient.java:187)
at o.m.r.c.SimpleClient.sendResumeError(SimpleClient.java:196)
at o.m.r.c.SimpleClient.receiveResumeError(SimpleClient.java:203)
at r.c.p.FluxOnErrorResume$ResumeSubscriber.onError(FluxOnErrorResume.java:94)
at c.t.w.p.c.l.MdcContextLifter.onError(MdcContextLifter.java:35)
at r.c.p.FluxMap$MapSubscriber.onError(FluxMap.java:134)
at r.n.c.FluxReceive.terminateReceiver(FluxReceive.java:478)
at r.n.c.FluxReceive.drainReceiver(FluxReceive.java:273)
at r.n.c.FluxReceive.onInboundError(FluxReceive.java:466)
at r.n.c.ChannelOperations.onInboundError(ChannelOperations.java:495)
at r.n.c.ChannelOperationsHandler.exceptionCaught(ChannelOperationsHandler.java:144)
at i.n.c.AbstractChannelHandlerContext.invokeExceptionCaught(AbstractChannelHandlerContext.java:346)
at i.n.c.AbstractChannelHandlerContext.invokeExceptionCaught(AbstractChannelHandlerContext.java:325)
at i.n.c.AbstractChannelHandlerContext.fireExceptionCaught(AbstractChannelHandlerContext.java:317)
at i.n.c.ChannelInboundHandlerAdapter.exceptionCaught(ChannelInboundHandlerAdapter.java:143)
at i.n.c.AbstractChannelHandlerContext.invokeExceptionCaught(AbstractChannelHandlerContext.java:346)
at i.n.c.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:447)
at i.n.c.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at i.n.c.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:412)
at i.n.c.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
at i.n.c.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:440)
at i.n.c.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:420)
at i.n.c.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
at i.n.c.e.AbstractEpollStreamChannel$EpollStreamUnsafe.epollInReady(AbstractEpollStreamChannel.java:800)
at i.n.c.e.EpollEventLoop.processReady(EpollEventLoop.java:499)
at i.n.c.e.EpollEventLoop.run(EpollEventLoop.java:397)
at i.n.u.c.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at i.n.u.i.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at i.n.u.c.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.lang.Thread.run(Thread.java:840)
It sames like a netty error when connect by r2dbc
Steps To Reproduce
Create mysql connection by Datasources
Public Sample App
No response
Environment
Production
Severity
High (Blocker to building or releasing)
Issue video log
No response
Version
v1.9.21
@alpacaca can you please upgrade to the latest to see if that help? Also, a screen recording would help in understanding exact steps to reproduce the issue. Is it happening during connection test or saving or creation etc.
@alpacaca can you please upgrade to the latest to see if that help? Also, a screen recording would help in understanding exact steps to reproduce the issue. Is it happening during connection test or saving or creation etc.
@pranavkanade sorry, I have repackaged the front-end page and just use appsmith plugin backend, so screen recording cannot help.
Actually, it is happening during connection test.
I will try the latest version instead and reply soon.
I try on app.appsmith.com/app in my app, it also doesn't work ... here is the screenshot
and request curl copy:
curl 'https://app.appsmith.com/api/v1/datasources/test' \
-H 'Accept: application/json, text/plain, */*' \
-H 'Accept-Language: zh-CN,zh;q=0.9,en-US;q=0.8,en;q=0.7' \
-H 'Connection: keep-alive' \
-H 'Content-Type: application/json' \
-H 'Cookie: ajs_anonymous_id=598de59d-162e-41a3-a710-6804b59db718; _hjSessionUser_2952959=eyJpZCI6Ijc5MTRhZTMyLWJkOGMtNTY1ZS05MjI0LWJhNTA0NTU1NGM3MiIsImNyZWF0ZWQiOjE3MjIyNDU2NjUwMTAsImV4aXN0aW5nIjp0cnVlfQ==; _gcl_au=1.1.1654232351.1722245666; _ga=GA1.1.356015106.1722245666; SESSION=7bf1d513-42fa-4b41-91d2-8f6c61793265; intercom-device-id-y10e7138=4982bf1f-7c74-417a-90eb-1e6279eec1dc; ajs_user_id=craig.sese@gmail.com; _hjSession_2952959=eyJpZCI6IjMxYWJjN2I5LTczY2EtNGRhZC05OTY5LTkyYWM1OGUwOWEwZCIsImMiOjE3MjIyNTM4NDg0MzQsInMiOjEsInIiOjEsInNiIjowLCJzciI6MCwic2UiOjAsImZzIjowLCJzcCI6MH0=; _ga_D1VS24CQXE=GS1.1.1722253846.2.0.1722253849.0.0.0; SL_C_23361dd035530_SID={"c370af0df0edf38360adbefbdc47d2b42ea137c9":{"sessionId":"meFwPGahFI_J9BcwRaEar","visitorId":"8UHRmHLrX0FD6AlJD6pG3"}}; intercom-session-y10e7138=UkZ4SHUrOWRUSUtTZkZHUzJxSzE4Z2NSNzEyc3JHK2tVeU0yTC9PR21qTEx6V1lScy92UlhGZVBrdU51QlQ1by0tSDBsbEROS2t4b2s0MlRsUElSWE9JZz09--53ad53c4e84f13df15c4f90144be0a4f2856508d' \
-H 'Origin: https://app.appsmith.com' \
-H 'Referer: https://app.appsmith.com/app/my-first-application/page1-64f7e9e149cf293c0e2f9a78/edit/datasource/temp-id-0?from=datasources&pluginId=5f192a44a2d05225d010db9d' \
-H 'Sec-Fetch-Dest: empty' \
-H 'Sec-Fetch-Mode: cors' \
-H 'Sec-Fetch-Site: same-origin' \
-H 'User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/126.0.0.0 Safari/537.36' \
-H 'X-Requested-By: Appsmith' \
-H 'newrelic: eyJ2IjpbMCwxXSwiZCI6eyJ0eSI6IkJyb3dzZXIiLCJhYyI6IjQyMDk1NDgiLCJhcCI6IjExMzQzODgzMDkiLCJpZCI6IjhlOWVmYzliNzhiOWZmM2EiLCJ0ciI6IjcxNDJiNGI0NzQ0MDU5OWE2MDAwMjM2M2UwNWU1MjdmIiwidGkiOjE3MjIyNTM5NDc1MzN9fQ==' \
-H 'sec-ch-ua: "Not/A)Brand";v="8", "Chromium";v="126", "Google Chrome";v="126"' \
-H 'sec-ch-ua-mobile: ?0' \
-H 'sec-ch-ua-platform: "macOS"' \
-H 'sentry-trace: 970bad5e3cf349b4864562131ed65ade-8826e65a5515406e-0' \
-H 'traceparent: 00-7142b4b47440599a60002363e05e527f-8e9efc9b78b9ff3a-01' \
-H 'tracestate: 4209548@nr=0-1-4209548-1134388309-8e9efc9b78b9ff3a----1722253947533' \
-H 'x-anonymous-user-id: 598de59d-162e-41a3-a710-6804b59db718' \
--data-raw '{"datasourceId":"","environmentId":"64f7e9df49cf293c0e2f9a72","datasourceConfiguration":{"properties":[null,{"key":"Connection method","value":"STANDARD"}],"connection":{"mode":"READ_WRITE","ssl":{"authType":"DEFAULT"}},"endpoints":[{"port":"3307","host":"146.56.222.244"}],"authentication":{"databaseName":"information_schema","username":"test","password":"123456"}},"pluginId":"5f192a44a2d05225d010db9d","workspaceId":"64f7e9df49cf293c0e2f9a6e"}'
@Nikhil-Nandagopal can reopen this issue?
Sample app:
https://app.appsmith.com/app/my-first-application/page1-64f7e9e149cf293c0e2f9a78/edit/datasource/temp-id-0?from=datasources&pluginId=5f192a44a2d05225d010db9d
| gharchive/issue | 2024-07-29T09:33:59 | 2025-04-01T04:33:31.014357 | {
"authors": [
"Nikhil-Nandagopal",
"alpacaca",
"pranavkanade"
],
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/issues/35253",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2492049115 | [Bug]: Dropdown option invisible but visible in preview
Is there an existing issue for this?
[X] I have searched the existing issues
Description
https://github.com/user-attachments/assets/38afc570-f7b9-416d-a89a-086d57281785
Steps To Reproduce
Label key Js is not wokring which was working earlier
Public Sample App
No response
Environment
Production
Severity
Critical (Broken Production apps)
Issue video log
No response
Version
1.38.0
@gokaco please can I see what your source data looks like so I can debug further, thanks.
@gokaco please can I see what your source data looks like so I can debug further, thanks.
Something like
[
{
"liveClassThumbnailURL": "https://res.cloudinary.com",
"cover": {
"vidURL": "https://res.cloudinary.com"
},
"email": "dfsdfsdf@adf.com",
"name": "Kartikeya Goswami",
"imgURL": "https://res.cloudinary.com",
"password": "abcdefg",
"roles": [
{
"roleId": "uuid"
},
{
"roleId": "uuid"
}
],
"subCategory": "COACH",
"expertise": [],
"trainerId": "uuid",
"category": "category",
"description": "description"
}
]
Thanks @gokaco. The label key cannot be the joined string you are using in the video, the label key needs to be one of the keys from your object above, e.g name. Add. name as the label key and see if it works.
@gokaco Did above comment help resolve your issue?
Thanks @gokaco. The label key cannot be the joined string you are using in the video, the label key needs to be one of the keys from your object above, e.g name. Add. name as the label key and see if it works.
@jacquesikot But earlier this was working. And this issue started happening from some days only. How it was working earlier? Have any changes been made in the new release?
@gokaco We don't think anything has changed in the recent release. The key field should be one of the keys in your JSON. This should work for your use case, please let us know if there's anything that is not working.
| gharchive/issue | 2024-08-28T13:14:55 | 2025-04-01T04:33:31.020763 | {
"authors": [
"Nikhil-Nandagopal",
"gokaco",
"jacquesikot",
"rahulbarwal"
],
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/issues/35953",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1388285451 | feat: Moved UserProfileImagePicker from ads folder
Description
This PR have moved UserProfileImagePicker component from ads folder to pages/UserProfile.
Fixes #17110
Type of change
Bug fix (non-breaking change which fixes an issue)
How Has This Been Tested?
Tested in local system
Checklist:
[x] My code follows the style guidelines of this project
[x] I have performed a self-review of my own code
[ ] I have commented my code, particularly in hard-to-understand areas
[ ] I have made corresponding changes to the documentation
[x] My changes generate no new warnings
[ ] I have added tests that prove my fix is effective or that my feature works
[x] New and existing unit tests pass locally with my changes
/ok-to-test sha=626569f
PR is tested, no issues found.
/ok-to-test sha=3237a55
/ok-to-test sha=5ce24ef
/ok-to-test sha=831b5d7
/ok-to-test sha=1def66f
/ok-to-test sha=86243b5
/ok-to-test sha=0de5497
/ok-to-test sha=8357e2d
| gharchive/pull-request | 2022-09-27T19:53:43 | 2025-04-01T04:33:31.026254 | {
"authors": [
"albinAppsmith",
"shadabbuchh"
],
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/pull/17111",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1582157450 | chore: type check callback functions in setInterval and setTimeout before invoking.
Description
We have recently changed the implementation for setInterval in https://github.com/appsmithorg/appsmith/pull/20168.
This has the potential to blow up our sentry quota if someone sets up an empty setInterval function with no callback in it and execute it. This PR adds a type check before invoking the callback to ensure that the passed in callback is indeed, a function.
Fixes https://github.com/appsmithorg/appsmith/issues/20591
https://appsmith.sentry.io/issues/3928721303/?project=1546547&query=is%3Aunresolved+worker&referrer=issue-stream&statsPeriod=14d
Media
A video or a GIF is preferred. when using Loom, don’t embed because it looks like it’s a GIF. instead, just link to the video
Type of change
Chore (housekeeping or task changes that don't impact user perception)
How Has This Been Tested?
Manual
Test Plan
Add Testsmith test cases links that relate to this PR
Issues raised during DP testing
Link issues raised during DP testing for better visiblity and tracking (copy link from comments dropped on this PR)
Checklist:
Dev activity
[x] My code follows the style guidelines of this project
[x] I have performed a self-review of my own code
[x] I have commented my code, particularly in hard-to-understand areas
[ ] I have made corresponding changes to the documentation
[x] My changes generate no new warnings
[ ] I have added tests that prove my fix is effective or that my feature works
[ ] New and existing unit tests pass locally with my changes
[ ] PR is being merged under a feature flag
QA activity:
[ ] Test plan has been approved by relevant developers
[ ] Test plan has been peer reviewed by QA
[ ] Cypress test cases have been added and approved by either SDET or manual QA
[ ] Organized project review call with relevant stakeholders after Round 1/2 of QA
[ ] Added Test Plan Approved label after reveiwing all Cypress test
/ok-to-test sha=2386038
/ok-to-test sha=eed1615
/ok-to-test sha=a80807d
| gharchive/pull-request | 2023-02-13T11:16:53 | 2025-04-01T04:33:31.033830 | {
"authors": [
"arunvjn"
],
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/pull/20592",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1652397012 | fix: set initialState from metaProps
Description
Fix the meta value code to set the initialMetaState correctly.
Type of change
Bug fix (non-breaking change which fixes an issue)
How Has This Been Tested?
Manual
Jest
Cypress
Test Plan
Issues raised during DP testing
Checklist:
Dev activity
[ ] My code follows the style guidelines of this project
[ ] I have performed a self-review of my own code
[ ] I have commented my code, particularly in hard-to-understand areas
[ ] I have made corresponding changes to the documentation
[ ] My changes generate no new warnings
[ ] I have added tests that prove my fix is effective or that my feature works
[ ] New and existing unit tests pass locally with my changes
[ ] PR is being merged under a feature flag
QA activity:
[ ] Test plan has been approved by relevant developers
[ ] Test plan has been peer reviewed by QA
[ ] Cypress test cases have been added and approved by either SDET or manual QA
[ ] Organized project review call with relevant stakeholders after Round 1/2 of QA
[ ] Added Test Plan Approved label after reveiwing all Cypress test
/ok-to-test sha=2213c10
/ok-to-test sha=11adecf
/ok-to-test sha=186a194
/ok-to-test sha=1ebf63f
/ok-to-test sha=1ebf63f
/ok-to-test sha=1ff88a1
/ci-merge-check
| gharchive/pull-request | 2023-04-03T16:29:51 | 2025-04-01T04:33:31.040277 | {
"authors": [
"Aishwarya-U-R",
"Rishabh-Rathod",
"arunvjn"
],
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/pull/22049",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1904586187 | fix: shifting migrations to shared dsl package
Description
Shifting migration logic to @shared/dsl package
PR fixes following issue(s)
Fixes #26783
Media
A video or a GIF is preferred. when using Loom, don’t embed because it looks like it’s a GIF. instead, just link to the video
Type of change
Please delete options that are not relevant.
Bug fix (non-breaking change which fixes an issue)
New feature (non-breaking change which adds functionality)
Breaking change (fix or feature that would cause existing functionality to not work as expected)
Chore (housekeeping or task changes that don't impact user perception)
This change requires a documentation update
Testing
How Has This Been Tested?
Please describe the tests that you ran to verify your changes. Also list any relevant details for your test configuration.
Delete anything that is not relevant
[ ] Manual
[ ] JUnit
[ ] Jest
[ ] Cypress
Test Plan
Add Testsmith test cases links that relate to this PR
Issues raised during DP testing
Link issues raised during DP testing for better visiblity and tracking (copy link from comments dropped on this PR)
Checklist:
Dev activity
[ ] My code follows the style guidelines of this project
[ ] I have performed a self-review of my own code
[ ] I have commented my code, particularly in hard-to-understand areas
[ ] I have made corresponding changes to the documentation
[ ] My changes generate no new warnings
[ ] I have added tests that prove my fix is effective or that my feature works
[ ] New and existing unit tests pass locally with my changes
[ ] PR is being merged under a feature flag
QA activity:
[ ] Speedbreak features have been covered
[ ] Test plan covers all impacted features and areas of interest
[ ] Test plan has been peer reviewed by project stakeholders and other QA members
[ ] Manually tested functionality on DP
[ ] We had an implementation alignment call with stakeholders post QA Round 2
[ ] Cypress test cases have been added and approved by SDET/manual QA
[ ] Added Test Plan Approved label after Cypress tests were reviewed
[ ] Added Test Plan Approved label after JUnit tests were reviewed
/ok-to-test sha=1b162e2
/ok-to-test sha=2cdad4c
/ok-to-test sha=dfb9efe
@sbalaji1192 can you please re-review and add your approval, I have added the changes
/build-deploy-preview
/ok-to-test sha=69a175b
Merged this PR with #28518
/build-deploy-preview
| gharchive/pull-request | 2023-09-20T09:32:20 | 2025-04-01T04:33:31.051221 | {
"authors": [
"Parthvi12",
"brayn003"
],
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/pull/27490",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2028226196 | feat: rest, graphql and curl fixed for onboarding
Description
This PR adds:
functionality to REST, GraphQL and Curl import options when using these during onboarding start with data flow. On clicking on any of these options would take us inside application.
PR fixes following issue(s)
Fixes #29303
Media
A video or a GIF is preferred. when using Loom, don’t embed because it looks like it’s a GIF. instead, just link to the video
Type of change
New feature (non-breaking change which adds functionality)
Testing
How Has This Been Tested?
Please describe the tests that you ran to verify your changes. Also list any relevant details for your test configuration.
Delete anything that is not relevant
[x] Manual
[ ] JUnit
[ ] Jest
[ ] Cypress
Test Plan
Add Testsmith test cases links that relate to this PR
Issues raised during DP testing
Link issues raised during DP testing for better visiblity and tracking (copy link from comments dropped on this PR)
Checklist:
Dev activity
[x] My code follows the style guidelines of this project
[x] I have performed a self-review of my own code
[x] I have commented my code, particularly in hard-to-understand areas
[ ] I have made corresponding changes to the documentation
[x] My changes generate no new warnings
[ ] I have added tests that prove my fix is effective or that my feature works
[x] New and existing unit tests pass locally with my changes
[ ] PR is being merged under a feature flag
QA activity:
[ ] Speedbreak features have been covered
[ ] Test plan covers all impacted features and areas of interest
[ ] Test plan has been peer reviewed by project stakeholders and other QA members
[ ] Manually tested functionality on DP
[ ] We had an implementation alignment call with stakeholders post QA Round 2
[ ] Cypress test cases have been added and approved by SDET/manual QA
[ ] Added Test Plan Approved label after Cypress tests were reviewed
[ ] Added Test Plan Approved label after JUnit tests were reviewed
Summary by CodeRabbit
Refactor
Improved the logic for identifying the current page during the onboarding process to enhance user experience.
/build-deploy-preview skip-tests=true
/ok-to-test
Progress on this is on halt due to critical issue: https://github.com/appsmithorg/appsmith/issues/29373
/ok-to-test
| gharchive/pull-request | 2023-12-06T10:27:13 | 2025-04-01T04:33:31.061887 | {
"authors": [
"sneha122"
],
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/pull/29377",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2329983955 | fix: Wrong evaluated value after binding checkbox group widget with…
Description
In this PR I have fixed wrong evaluated value after binding checkbox group widget with query
Output: Loom
[!TIP]
Add a TL;DR when the description is longer than 500 words or extremely technical (helps the content, marketing, and DevRel team).
Please also include relevant motivation and context. List any dependencies that are required for this change. Add links to Notion, Figma or any other documents that might be relevant to the PR.
Fixes #24620
or
Fixes https://github.com/appsmithorg/appsmith/issues/24620
[!WARNING]
If no issue exists, please create an issue first, and check with the maintainers if the issue is valid.
Automation
/ok-to-test tags=""
:mag: Cypress test results
[!CAUTION]
If you modify the content in this section, you are likely to disrupt the CI result for your PR.
Communication
Should the DevRel and Marketing teams inform users about this change?
[ ] Yes
[x] No
Summary by CodeRabbit
New Features
Introduced test cases for various Anvil widgets, covering Canvas, Preview, and Deploy modes.
Added filtering functionality for Table widgets in the UI.
Bug Fixes
Updated the Checkbox Group Widget to automatically validate and update selected values.
Resolved issues with Docker image builds and workflows by updating configurations and scripts.
Fixed URL redirection logic for legacy to slug URLs.
Chores
Updated GitHub Actions workflows for improved CI/CD processes.
Adjusted Dockerfile permissions and configurations for better compatibility.
Tests
Added numerous Cypress end-to-end tests for Anvil widgets and other UI components.
Enhanced test scripts to handle Cypress snapshots on failure.
/build-deploy-preview
Fix works on fixed mode, needs to be propagated to Anvil too
cc: @jsartisan @KelvinOm
@Shivam-z The solution does not work properly. Check the video. A better approach would be a custom validator for the defaultSelectedValues property.
https://github.com/appsmithorg/appsmith/assets/6636360/4962a637-8919-408f-9437-0ed1cec838ef
Closing the PR as the solution does not solve it properly.
@jsartisan sorry, I was bit busy with the other work , i.e could not able to looked into this PR.
can you give me some time to look into this issue and for a better solution.
Please reopen this PR.
@jsartisan I have modified defaultSelectedValuesValidation
could you please review that.
The solution looks good to me. Two things
the validation is always valid. even when the value is not an array.
When the options change, it does not trigger the validation again.
The solution to this is to add depencyMap:
static getDependencyMap(): Record<string, string[]> {
return {
defaultSelectedValues: ["options"],
};
}
Hi @jsartisan , I fixed the second suggestion you gave me.
can you give me more clarification on first suggestion "the validation is always valid. even when the value is not an array."
here which value you are talking about ? is it defaultSelectedValues ?
If it is defaultSelectedValues , then in case of checkBoxGroup Widget it should be an array only right ?
Hey, couldn't get time to check again. Regarding this"the validation is always valid. even when the value is not an array.", i think it's fine.
@ramsaptami can you check this PR as well? Codewise it looks okay. Need some help in checking validations for queries.
/build-deploy-preview skip-tests=true
default value when different from derived value is not retained in selectedValues property
binding a different query refreshes the 'selectedValues' property
cc: @jsartisan
@jsartisan can you please look into this PR and tell me whether it is good or I need to rectify something.
@Shivam-z Can you check the comments by Saptami? The solution is failing in a few cases.
@jsartisan The comments given by @ramsaptami is the expected behaviour not the bug.
In the previous version it was showing the defaultValues as selectedValues even though the derived did not contain those default values . For ex: derivedValues: {1,2,3} , defaultValues:{2,3,4} . so selectedValues should be {2,3} but previously it was {2,3,4} i.e 4 was not retained.
In the previous version when binding a different query it was not refreshing the 'selectedValues' property but now it is fixed.
@Shivam-z Code looks good to me now. I ran the tests on a separate PR and found there is one test that is constantly failing
cypress/e2e/Regression/ClientSide/Widgets/Switch/SwitchGroup1_spec.ts
https://github.com/appsmithorg/appsmith/assets/6636360/c9d1efbd-a662-42fb-966c-7823d2cb992c
Check the video.
Btw Sorry for the back and forth, I know it's taking time.
Hi @jsartisan , I have updated some code in the cypress test case for SwitchGroup1_spec.ts file and it is running fine in my local. can you please check this PR.
@Shivam-z something wrong with your PR. It's now showing 616 files. Can you rebase latest release into your branch
I want to create a PR from your PR and run tests on it like we did previously but now your PR is showing 616 files changed.
@jsartisan updated file changes
@Shivam-z all tests passing now. What was the issue?
Hi @jsartisan , some assertions were failing in Switch/SwitchGroup1_spec.ts and it was expecting different color value because of the functionality added in this PR and due to this the test cases were failing earlier , but now it is fixed.
Regarding test to cover this functionality , I have already added one test case in : app/client/src/widgets/CheckboxGroupWidget/widget/index.test.tsx
| gharchive/pull-request | 2024-06-03T01:25:42 | 2025-04-01T04:33:31.081474 | {
"authors": [
"Shivam-z",
"jsartisan",
"ramsaptami"
],
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/pull/33906",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2408004204 | feat:Restrict Phone Number Length Based on User Selection of Max Leng…
…th (#34787)
Description
[!TIP]
Add a TL;DR when the description is longer than 500 words or extremely technical (helps the content, marketing, and DevRel team).
Please also include relevant motivation and context. List any dependencies that are required for this change. Add links to Notion, Figma or any other documents that might be relevant to the PR.
Fixes #34787
or
Fixes https://github.com/appsmithorg/appsmith/issues/34787
[!WARNING]
If no issue exists, please create an issue first, and check with the maintainers if the issue is valid.
Automation
/ok-to-test tags=""
:mag: Cypress test results
[!CAUTION]
If you modify the content in this section, you are likely to disrupt the CI result for your PR.
Communication
Should the DevRel and Marketing teams inform users about this change?
[ ] Yes
[ ] No
VIDEOS
VIDEO:SHOWING THE FEATURE
VIDEO:CYPRESS TESTING
Summary by CodeRabbit
New Features
Added maxLength property to Phone Input widget, allowing users to specify a maximum character limit.
Implemented validation for the Phone Input widget to display an error message when the input exceeds the set maximum character limit.
Tests
Introduced a new test scenario for validating the maximum character limit enforcement on the Phone Input widget.
@Harshithazemoso Can we please jest test cases instead of cypress tests ?
cc : @ApekshaBhosale @yatinappsmith @sagar-qa007
Left comments
hello @rajatagrawal added a test file and added a test case can you please check it now
hello @rajatagrawal can you please review this pr
cc : @rahulbarwal Can you please help review the PR ?
hello @rahulbarwal ,could you please review this pr .
Thank you
| gharchive/pull-request | 2024-07-15T06:58:45 | 2025-04-01T04:33:31.090508 | {
"authors": [
"Harshithazemoso",
"rajatagrawal"
],
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/pull/34925",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2530349482 | fix:updated searchbar component to return result with client side sea…
Description:
modified the code the return the result when client side search is enable and not to when client side search is disabled.
fixes: #15386
updates in PR:
1.passed enableClientSide prop to the search component and applied conditional rendering to return result.
2.added test file for the changed code.
snapshorts:
Summary by CodeRabbit
New Features
Introduced enableClientSideSearch property across multiple components, allowing for enhanced client-side searching capabilities.
Updated SearchComponent, Table, and TableHeader to support dynamic search functionalities based on user input.
Bug Fixes
Improved state management in the SearchComponent to reset search values appropriately.
Tests
Added comprehensive unit tests for the SearchComponent to validate its functionality and user interactions.
Hii @shadabbuchh @carinanfonseca , I have raised the PR, could you please assign the reviewer to this PR.
thank you.
Hii @shadabbuchh @carinanfonseca @NilanshBansal ,could you please assign the reviewer to this PR.
thank you.
Adding @rahulbarwal @jacquesikot from the widgets team to check this out.
Hii @rahulbarwal , I have updated the test case names, could you check the changes.
thank you.
@Naveen-Goud CI is failing for your changes: https://github.com/appsmithorg/appsmith/actions/runs/11230421118/job/31217757085?pr=36745
Please check and fix.
@Naveen-Goud can you please fix the failing CI issues.
Hii @rahulbarwal , I have resolved the lint issue , can you re-run the checks.
thank you.
@Naveen-Goud client lint is still failing on your changes: https://github.com/appsmithorg/appsmith/actions/runs/11321677640/job/31481100928?pr=36745#step:9:379
Please run yarn run lint:ci in your local to confirm after fixing.
@Naveen-Goud are you still working on this?
| gharchive/pull-request | 2024-09-17T07:59:44 | 2025-04-01T04:33:31.099782 | {
"authors": [
"Naveen-Goud",
"NilanshBansal",
"rahulbarwal"
],
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/pull/36360",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2552536384 | fix: fixing modal height by removing min-content
Description
min-content css property is behaving weirdly in the latest update of the chrome browser
Fixes https://github.com/appsmithorg/appsmith/issues/36586
Automation
/ok-to-test tags=""
:mag: Cypress test results
[!CAUTION]
If you modify the content in this section, you are likely to disrupt the CI result for your PR.
Communication
Should the DevRel and Marketing teams inform users about this change?
[ ] Yes
[ ] No
/build-deploy-preview
| gharchive/pull-request | 2024-09-27T10:07:23 | 2025-04-01T04:33:31.103061 | {
"authors": [
"brayn003"
],
"repo": "appsmithorg/appsmith",
"url": "https://github.com/appsmithorg/appsmith/pull/36589",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
355842359 | not working after update my flutter to 2.0
error in get package
Same
Does this fix your problem?
https://github.com/apptreesoftware/flutter_google_map_view/pull/115
@MathieuAuclair
Could not find a file named "pubspec.yaml" in "/home/hossein/AndroidStudioProjects/mahout/packages/flutter_google_map_view-master/lib/dependencies/uri.dart".
Same,
Flutter 0.6.0 • channel beta • https://github.com/flutter/flutter.git
Framework • revision 9299c02cf7 (3 weeks ago) • 2018-08-16 00:35:12 +0200
Engine • revision e3687f70c7
Tools • Dart 2.1.0-dev.0.0.flutter-be6309690f
thanks anyway for your work.
Map_view is depending on uri 0.11.1 which is outdated. Someone should upgrade the dependencies immediately.
| gharchive/issue | 2018-08-31T05:16:04 | 2025-04-01T04:33:31.125779 | {
"authors": [
"MathieuAuclair",
"andreagr",
"dbbd59",
"elricym",
"hojjabr",
"sudipta213"
],
"repo": "apptreesoftware/flutter_google_map_view",
"url": "https://github.com/apptreesoftware/flutter_google_map_view/issues/114",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2446442033 | [Feature Request]: add smooth scrolling
Is there an existing issue for this?
[X] I have searched the existing issues
Feature Description
The ui is very laggy .Adding a smooth scroll effect can make it slight better.
Use Case
Enhancing User Experience.
Benefits
No response
Add ScreenShots
No response
Priority
High
Record
[X] I have read the Contributing Guidelines
[X] I'm a GSSOC'24 contributor
[ ] I'm a VSOC'24 contributor
[X] I have starred the repository
Hii, I want to tackle this issue, So please assign to me.
Hii, I want to tackle this issue, So please assign to me.
u are assigned
| gharchive/issue | 2024-08-03T16:14:19 | 2025-04-01T04:33:31.177017 | {
"authors": [
"Ranj8521Kumar",
"Subhams-GIT",
"apu52"
],
"repo": "apu52/Travel_Website",
"url": "https://github.com/apu52/Travel_Website/issues/1308",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2349727709 | Email text modification
Problem
* The email text is not properly positioned like other text.
* I marked the problem with a green marker.
[X] I am contributing to an open-source program in GSSOC'24.
@Shashanktriathi1703 please assign this issue to me under gssoc
@Shashanktriathi1703 I have assigned u this time .But please from next time try to raise the issues using the issue templates in the "NEW ISSUE" section
| gharchive/issue | 2024-06-12T21:40:21 | 2025-04-01T04:33:31.179506 | {
"authors": [
"Salma-Mamdoh",
"Shashanktriathi1703",
"apu52"
],
"repo": "apu52/Travel_Website",
"url": "https://github.com/apu52/Travel_Website/issues/859",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1528257263 | Feature/tables view
Task details: https://quotron.notion.site/Tables-view-7d00ce65a4294cc2904860277ff2c351
merged
| gharchive/pull-request | 2023-01-11T01:37:33 | 2025-04-01T04:33:31.180605 | {
"authors": [
"apurvak",
"whitewolfdev410"
],
"repo": "apurvak/superset",
"url": "https://github.com/apurvak/superset/pull/42",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1638095044 | chore: call bpf_program__attach_iter from go
This changes libbpfgo to call bpf_program__attach_iter directly from go instead of doing this via a C wrapper. What this means is that we maintained opts creation in C, but the actual libbpf attach call is done in go.
This creates two new functions in libbpfgo.h as helpers:
bpf_iter_attach_opts_new()
bpf_iter_attach_opts_free()
As questioned in https://github.com/aquasecurity/libbpfgo/pull/254#discussion_r1144230843.
Thanks for that @geyslan. My OCD sent compliments!
| gharchive/pull-request | 2023-03-23T18:37:41 | 2025-04-01T04:33:31.201095 | {
"authors": [
"geyslan"
],
"repo": "aquasecurity/libbpfgo",
"url": "https://github.com/aquasecurity/libbpfgo/pull/304",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1506332436 | [Feature] Add support for severity and ignore-unfixed parameters
At the moment it's not possible to specify which severities should be scannen for.
Also unfixable severities can't be hidden.
This is all possible with existing arguments for the trivy command.
duplicate of #16
| gharchive/issue | 2022-12-21T13:54:21 | 2025-04-01T04:33:31.202225 | {
"authors": [
"ferronsw"
],
"repo": "aquasecurity/trivy-azure-pipelines-task",
"url": "https://github.com/aquasecurity/trivy-azure-pipelines-task/issues/26",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1954066962 | Missing ServiceAccount, some Role rules and RoleBinding with OLM installation
Description
Hi all,
I'm Following the procedure described in the official docs, and after creating the Subscription resource the status of the cluster service version CR is Pending:
$ kubectl get clusterserviceversions.operators.coreos.com -n trivy-system
NAME DISPLAY VERSION REPLACES PHASE
trivy-operator.v0.16.3 Trivy Operator 0.16.3 trivy-operator.v0.16.2 Pending
The initial reason is the absence of the ServiceAccount resource named trivy-operator:
$ kubectl describe clusterserviceversions.operators.coreos.com -n trivy-system
...
...
Group:
Kind: ServiceAccount
Message: Service account does not exist
Name: trivy-operator
Status: NotPresent
Version: v1
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal RequirementsUnknown 10m operator-lifecycle-manager requirements not yet checked
Normal RequirementsNotMet 10m operator-lifecycle-manager one or more requirements couldn't be found
But even after creating it, some permissions are missing in the ClusterRole named trivy-operator:
$ kubectl describe clusterserviceversions.operators.coreos.com -n trivy-system
...
...
Group: rbac.authorization.k8s.io
Kind: PolicyRule
Message: cluster rule:{"verbs":["get","list","watch","create","update","delete"],"apiGroups":["aquasecurity.github.io"],"resources":["vulnerabilityreports","configauditreports","clustercompliancereports","clusterconfigauditreports","exposedsecretreports","sbomreports","rbacassessmentreports","infraassessmentreports","clusterrbacassessmentreports","clusterinfraassessmentreports"]}
Status: NotSatisfied
Version: v1
Group: rbac.authorization.k8s.io
Kind: PolicyRule
Message: cluster rule:{"verbs":["create","get","update"],"apiGroups":["coordination.k8s.io"],"resources":["leases"]}
Status: NotSatisfied
Version: v1
Group:
Kind: ServiceAccount
Message: Policy rule not satisfied for service account
Name: trivy-operator
Status: PresentNotSatisfied
Version: v1
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal RequirementsUnknown 0s operator-lifecycle-manager requirements not yet checked
Normal RequirementsNotMet 0s operator-lifecycle-manager one or more requirements couldn't be found
After manually adding everything, the Operator gets installed, so I'm using this yaml that contains what is missing:
# ServiceAccount
apiVersion: v1
kind: ServiceAccount
metadata:
name: trivy-operator
namespace: trivy-system
labels:
app.kubernetes.io/name: trivy-operator
app.kubernetes.io/instance: trivy-operator
app.kubernetes.io/version: "0.1.4"
app.kubernetes.io/managed-by: kubectl
---
# ClusterRole
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRole
metadata:
creationTimestamp: null
name: trivy-operator
rules:
####################################################
# These are the installed rules:
- apiGroups:
- ""
resources:
- configmaps
verbs:
- get
- list
- watch
- apiGroups:
- ""
resources:
- limitranges
verbs:
- get
- list
- watch
- apiGroups:
- ""
resources:
- pods
verbs:
- get
- list
- watch
- apiGroups:
- ""
resources:
- pods/log
verbs:
- get
- list
- apiGroups:
- ""
resources:
- replicationcontrollers
verbs:
- get
- list
- watch
- apiGroups:
- ""
resources:
- resourcequotas
verbs:
- get
- list
- watch
- apiGroups:
- ""
resources:
- secrets
verbs:
- create
- get
- apiGroups:
- ""
resources:
- serviceaccounts
verbs:
- get
- apiGroups:
- ""
resources:
- services
verbs:
- get
- list
- watch
- apiGroups:
- apiextensions.k8s.io
resources:
- customresourcedefinitions
verbs:
- get
- list
- watch
- apiGroups:
- apps
resources:
- daemonsets
verbs:
- get
- list
- watch
- apiGroups:
- apps
resources:
- replicasets
verbs:
- get
- list
- watch
- apiGroups:
- apps
resources:
- statefulsets
verbs:
- get
- list
- watch
- apiGroups:
- aquasecurity.github.io
resources:
- clustercompliancedetailreports
verbs:
- create
- delete
- get
- list
- patch
- update
- watch
- apiGroups:
- aquasecurity.github.io
resources:
- clustercompliancereports
verbs:
- create
- delete
- get
- list
- patch
- update
- watch
- apiGroups:
- aquasecurity.github.io
resources:
- clustercompliancereports/status
verbs:
- get
- patch
- update
- apiGroups:
- aquasecurity.github.io
resources:
- clusterconfigauditreports
verbs:
- create
- delete
- get
- list
- patch
- update
- watch
- apiGroups:
- aquasecurity.github.io
resources:
- clusterrbacassessmentreports
verbs:
- create
- delete
- get
- list
- patch
- update
- watch
- apiGroups:
- aquasecurity.github.io
resources:
- configauditreports
verbs:
- create
- delete
- get
- list
- patch
- update
- watch
- apiGroups:
- aquasecurity.github.io
resources:
- exposedsecretreports
verbs:
- create
- delete
- get
- list
- patch
- update
- watch
- apiGroups:
- aquasecurity.github.io
resources:
- rbacassessmentreports
verbs:
- create
- delete
- get
- list
- patch
- update
- watch
- apiGroups:
- aquasecurity.github.io
resources:
- vulnerabilityreports
verbs:
- create
- delete
- get
- list
- patch
- update
- watch
- apiGroups:
- batch
resources:
- cronjobs
verbs:
- get
- list
- watch
- apiGroups:
- batch
resources:
- jobs
verbs:
- create
- delete
- get
- list
- watch
- apiGroups:
- networking.k8s.io
resources:
- ingresses
verbs:
- get
- list
- watch
- apiGroups:
- networking.k8s.io
resources:
- networkpolicies
verbs:
- get
- list
- watch
- apiGroups:
- rbac.authorization.k8s.io
resources:
- clusterrolebindings
verbs:
- get
- list
- watch
- apiGroups:
- rbac.authorization.k8s.io
resources:
- clusterroles
verbs:
- get
- list
- watch
- apiGroups:
- rbac.authorization.k8s.io
resources:
- rolebindings
verbs:
- get
- list
- watch
- apiGroups:
- rbac.authorization.k8s.io
resources:
- roles
verbs:
- get
- list
- watch
####################################################
# These are the missing rules that I manually added:
- verbs:
- get
- list
- watch
apiGroups:
- ""
resources:
- pods
- pods/log
- replicationcontrollers
- services
- resourcequotas
- limitranges
- verbs:
- get
- list
- watch
apiGroups:
- ""
resources:
- nodes
- verbs:
- list
- watch
- get
- create
- update
apiGroups:
- ""
resources:
- configmaps
- secrets
- serviceaccounts
- verbs:
- delete
apiGroups:
- ""
resources:
- secrets
- verbs:
- create
apiGroups:
- ""
resources:
- events
- verbs:
- get
- list
- watch
apiGroups:
- apps
resources:
- replicasets
- statefulsets
- daemonsets
- deployments
- verbs:
- get
- list
- watch
apiGroups:
- policy
resources:
- podsecuritypolicies
- verbs:
- get
- list
- watch
- create
- update
- delete
apiGroups:
- aquasecurity.github.io
resources:
- vulnerabilityreports
- configauditreports
- clustercompliancereports
- clusterconfigauditreports
- exposedsecretreports
- sbomreports
- rbacassessmentreports
- infraassessmentreports
- clusterrbacassessmentreports
- clusterinfraassessmentreports
- verbs:
- create
- get
- update
apiGroups:
- coordination.k8s.io
resources:
- leases
---
# ClusterRoleBinding
apiVersion: rbac.authorization.k8s.io/v1
kind: ClusterRoleBinding
metadata:
name: trivy-operator
labels:
app.kubernetes.io/name: trivy-operator
app.kubernetes.io/instance: trivy-operator
app.kubernetes.io/version: "0.1.4"
app.kubernetes.io/managed-by: kubectl
roleRef:
apiGroup: rbac.authorization.k8s.io
kind: ClusterRole
name: trivy-operator
subjects:
- kind: ServiceAccount
name: trivy-operator
namespace: trivy-system
Issue
I was expecting everything to be covered by the subscription, but this also makes me feel like I'm missing something. Is this a normal behavior? Is it expected that the admin must create the SA/Role/RoleBinding by himself before creating the Subscription?
This doesn't happen while installing everything using just kubectl.
Environment
Trivy-Operator version (use trivy-operator version): 0.16.3
Kubernetes version (use kubectl version): v1.27.4
OS (macOS 10.15, Windows 10, Ubuntu 19.10 etc): AlmaLinux release 8.8 (Sapphire Caracal)
@rascasoft has is a ref for operator I'll take a look at it as well.
also please review operator install instruction
Thanks for your answer @chen-keinan , but I didn't understand what you're trying to tell me.
The first link is the manifest of the cluster service version of the trivy-operator in the operatorhub sources, and I'm not able to understand from here what I want to know, which is if it's on purpose that service account and role/rolebindings are not created by the operator installer.
The second link takes me to the instructions which I already followed and don't tell me much more compared to what is the doc I followed (this one doesn't cover the operatorgroup resource creation for instance).
Am I missing something?
Many thanks!
Thanks for your answer @chen-keinan , but I didn't understand what you're trying to tell me. The first link is the manifest of the cluster service version of the trivy-operator in the operatorhub sources, and I'm not able to understand from here what I want to know, which is if it's on purpose that service account and role/rolebindings are not created by the operator installer. The second link takes me to the instructions which I already followed and don't tell me much more compared to what is the doc I followed (this one doesn't cover the operatorgroup resource creation for instance).
Am I missing something?
Many thanks!
its just a ref for operator hub OLM official install instructions, I'll have another look myself and see if I can find the issue
Many thanks @chen-keinan. There's one main thing I would like to understand (and maybe you can help here): when I configure a subscription for OLM, are all the resources that are going to be used (so Service Accounts, ClusterRole, ClusterRoleBinding, etc.) expected to be created by the subscription itself?
I ask this because if you use OLM in OpenShift, then everything works flawlessly, but in K8s things seem different.
Many thanks!
Many thanks @chen-keinan. There's one main thing I would like to understand (and maybe you can help here): when I configure a subscription for OLM, are all the resources that are going to be used (so Service Accounts, ClusterRole, ClusterRoleBinding, etc.) expected to be created by the subscription itself?
I ask this because if you use OLM in OpenShift, then everything works flawlessly, but in K8s things seem different.
Many thanks!
yes , this is how it works based on my understanding
I've got an addition to made at this, after further testing: it seems that if you install OLM using operator-sdk the behavior changes to what you expect.
On a clean environment I installed OLM by using:
$ operator-sdk olm install
And then I created OperaotrGroup and Subscription as follows:
apiVersion: operators.coreos.com/v1
kind: OperatorGroup
metadata:
name: trivy-operator-group
namespace: trivy-system
spec:
targetNamespaces:
- trivy-system
- mions1
---
apiVersion: operators.coreos.com/v1alpha1
kind: Subscription
metadata:
name: trivy-operator-subscription
namespace: trivy-system
spec:
channel: alpha
name: trivy-operator
source: operatorhubio-catalog
sourceNamespace: olm
installPlanApproval: Automatic
config:
env:
- name: OPERATOR_EXCLUDE_NAMESPACES
value: "kube-system,trivy-system
The CSV installed correctly:
$ kubectl get csv -n trivy-system
NAME DISPLAY VERSION REPLACES PHASE
trivy-operator.v0.16.4 Trivy Operator 0.16.4 trivy-operator.v0.16.3 Succeeded
The only remaining problem was the absence of the trivy-operator-trivy-config config map inside the trivy-system namespace, which was causing this message to appear in the pod logs:
{"level":"error","ts":"2023-11-14T16:53:09Z","msg":"Reconciler error","controller":"clusterrole","controllerGroup":"rbac.authorization.k8s.io","controllerKind":"ClusterRole","ClusterRole":{"name":"rbacassessmentreports.aquasecurity.github.io-v1alpha1-crdview"},"namespace":"","name":"rbacassessmentreports.aquasecurity.github.io-v1alpha1-crdview","reconcileID":"4f4e747e-6947-416c-bada-22e270b1f9fe","error":"ConfigMap \"trivy-operator-trivy-config\" not found","stacktrace":"sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/home/runner/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/internal/controller/controller.go:329\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/home/runner/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/home/runner/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/internal/controller/controller.go:227"}
So I created it manually, using the code from the static manifest:
apiVersion: v1
kind: ConfigMap
metadata:
name: trivy-operator-trivy-config
namespace: trivy-system
labels:
app.kubernetes.io/name: trivy-operator
app.kubernetes.io/instance: trivy-operator
app.kubernetes.io/version: "0.16.4"
app.kubernetes.io/managed-by: kubectl
data:
trivy.repository: "ghcr.io/aquasecurity/trivy"
trivy.tag: "0.47.0"
trivy.imagePullPolicy: "IfNotPresent"
trivy.additionalVulnerabilityReportFields: ""
trivy.severity: "UNKNOWN,LOW,MEDIUM,HIGH,CRITICAL"
trivy.slow: "true"
trivy.skipJavaDBUpdate: "false"
trivy.imageScanCacheDir: "/tmp/trivy/.cache"
trivy.filesystemScanCacheDir: "/var/trivyoperator/trivy-db"
trivy.dbRepository: "ghcr.io/aquasecurity/trivy-db"
trivy.javaDbRepository: "ghcr.io/aquasecurity/trivy-java-db"
trivy.command: "image"
trivy.dbRepositoryInsecure: "false"
trivy.useBuiltinRegoPolicies: "true"
trivy.supportedConfigAuditKinds: "Workload,Service,Role,ClusterRole,NetworkPolicy,Ingress,LimitRange,ResourceQuota"
trivy.timeout: "5m0s"
trivy.mode: "Standalone"
trivy.resources.requests.cpu: "100m"
trivy.resources.requests.memory: "100M"
trivy.resources.limits.cpu: "500m"
trivy.resources.limits.memory: "500M"
Even if I'm still getting some weird logs inside the Pod:
{"level":"error","ts":"2023-11-14T16:55:40Z","msg":"Reconciler error","controller":"job","controllerGroup":"batch","controllerKind":"Job","Job":{"name":"scan-vulnerabilityreport-75dc95d75c","namespace":"trivy-system"},"namespace":"trivy-system","name":"scan-vulnerabilityreport-75dc95d75c","reconcileID":"94e371e7-4f8e-4499-b918-39945792221f","error":"SbomReport.aquasecurity.github.io \"replicaset-nginx-8bc664746-nginx\" is invalid: report.components.specVersion: Invalid value: \"string\": report.components.specVersion in body must be of type integer: \"string\"","stacktrace":"sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/home/runner/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/internal/controller/controller.go:329\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/home/runner/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/internal/controller/controller.go:266\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/home/runner/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.16.3/pkg/internal/controller/controller.go:227"}
The configauditreport and vulnerabilityreport are generated.
So I see two topics here:
Understand what is the difference between manual and operator-sdk OLM installation (not to be addressed here).
Fix the ConfigMap generation inside the trivy-operator manifest (which becomes the main issue of this bug).
Thanks, i'll have a look at this feedback
@chen-keinan I have created a patch , it will be released with trivy-operator v0.17.0
| gharchive/issue | 2023-10-20T11:10:19 | 2025-04-01T04:33:31.220988 | {
"authors": [
"chen-keinan",
"rascasoft"
],
"repo": "aquasecurity/trivy-operator",
"url": "https://github.com/aquasecurity/trivy-operator/issues/1584",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1266988904 | Eng 588 allow users to upload csvs to demo be
https://github.com/aqueducthq/aqueduct/pull/82 -- BE API endpoint to create table in demo DB from CSV
https://github.com/aqueducthq/aqueduct/pull/83 -- FE UI to allow upload of CSV to demo DB
https://github.com/aqueducthq/aqueduct/pull/55 -- FE Integrations Details page where this functionality lives
A small request on PR naming @eunice-chan -- when we're stacking multiple PRs like this, it would be great if we included the system component in the title. Otherwise, it's kinda difficult to tell which one is which. e.g., #83 could be titled "ENG 588: [UI] Allow users to upload CSVs to demo DB" and #82 could be titled "ENG 588: [Server] Add endpoint for accepting and uploading new table". Does that make sense?
| gharchive/pull-request | 2022-06-10T04:03:39 | 2025-04-01T04:33:31.229658 | {
"authors": [
"eunice-chan",
"vsreekanti"
],
"repo": "aqueducthq/aqueduct",
"url": "https://github.com/aqueducthq/aqueduct/pull/82",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
912749764 | Decision Transformer: Reinforcement Learning via Sequence Modeling
一言でいうと
Transformerを強化学習に応用した研究。State/Action/Rewardの系列を入力して次の行動を予測させる。収録済みの軌跡から学習するオフライン強化学習で、既存の手法を上回る精度(オンラインの強化学習ではまだ検証されていない)。
論文リンク
https://arxiv.org/abs/2106.01345
著者/所属機関
Lili Chen, Kevin Lu, Aravind Rajeswaran, Kimin Lee, Aditya Grover, Michael Laskin, Pieter Abbeel, Aravind Srinivas, Igor Mordatch
UC Berkeley
Facebook AI Research
Google Brain
投稿日付(yyyy/MM/dd)
2021/6/2
概要
新規性・差分
手法
結果
コメント
解説資料
https://www.slideshare.net/yasunoriozaki12/decision-transformer-reinforcement-learning-via-sequence-modeling
| gharchive/issue | 2021-06-06T09:40:16 | 2025-04-01T04:33:31.234226 | {
"authors": [
"icoxfog417"
],
"repo": "arXivTimes/arXivTimes",
"url": "https://github.com/arXivTimes/arXivTimes/issues/2063",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
383394455 | script hangs at "checking major 3rd party components"
the issue could very well be on my end, however I am using a fresh intall of ubuntu minimal 14.04.1.
upon running rtinst, it hangs at the xmlrpc-c check. here is terminal output:
user@server:~# rtinst
Ubuntu 14.04.1 LTS
Checking the web sites we will need are accessible
Checking Ubuntu mirrors
http://archive.ubuntu.com/ubuntu: OK
http://security.ubuntu.com/ubuntu: OK
http://archive.canonical.com/ubuntu: OK
Checking major 3rd party components
Rtorrent: OK
xmlrpc-c:
it's odd because the pages referenced by xmlrpc_url and xmlrpc_url_alt both seem to resolve for me.
i also noticed there's some error messages to display if the urls don't resolve, but the script doesn't seem to get that far.
i git cloned the repo and changed the xmlrpc-c URL to the one referenced in commit #29a3ad9 but that didn't seem to fix it. i'm not actually sure what else i could try but i'm gonna play with it some more.
line #293 and #295 of /scripts/rtinst reference xmlrpc_repo=1 but this doesn't seem to be set, looks like it belongs at line #278. i added this, i alo switched xmlrpc_url and xmlrpc_alt because maybe there was a resolving error there? and i added a failure message similar to the other checks, but the script still hangs for me.
# check required web repos are accessible
os_repo=0
major_repo=0
xmlrpc_repo=0
sed -i "s/\/debian\s/\/debian\/ /g" /etc/apt/sources.list
echo
echo "Checking the web sites we will need are accessible"
echo "Checking $osname mirrors"
for i in $(cat /etc/apt/sources.list | grep "^deb http" | cut -d' ' -f2 | uniq ); do
echo -n $i": "
check_url $i && echo "OK" || { echo "FAIL"; os_repo=1; }
done
echo
echo "Checking major 3rd party components"
echo -n "Rtorrent: "; check_url $rt_url && echo "OK" || { echo "FAIL"; major_repo=1; }
echo -n "xmlrpc-c: "; check_url $xmlrpc_url && echo "OK" || { echo "FAIL"; xmlrpc_repo=1; }
if [[ $xmlrpc_repo = 1 ]]; then
xmlrpc_url=$xmlrpc_url_alt
check_url $xmlrpc_url && echo "OK" || { echo "FAIL"; major_repo=1; }
fi
Same here. hang at xmlrpc-c:
any solutions?
Is this issue still present? If so, please add the following to rtinst script after #!/usr/bin/env bash on the top:
set -e -x
And remember to make the script executable again with:
chmod +x /etc/rtinst/scripts/rtinst
Re-run the script and post the whole output.
| gharchive/issue | 2018-11-22T06:17:01 | 2025-04-01T04:33:31.262815 | {
"authors": [
"V33m",
"ieffedmybffjill",
"lowprofileusername"
],
"repo": "arakasi72/rtinst",
"url": "https://github.com/arakasi72/rtinst/issues/381",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1772140021 | Client component translations are coming one step back.
What version of this package are you using? 2.3.1
What operating system, Node.js, and npm version? NodeJs=> 16.17.1 Npm=> 8.15.0 Next=>13.4.4
What happened?
If I change the language using the language switcher component, Client components' translations are coming one step back.
If I choose French as a default language it works as expected at first.
Then if I change the language to Dutch, as you can see the buttons translations are still French
Then If I change the language to English, now buttons translations are Dutch(one step back)
My language switcher component
'use client';
import Link from 'next/link';
import { usePathname } from 'next/navigation';
const LanguageOptions = [
{ label: 'English', value: 'en' },
{ label: 'Français', value: 'fr' },
{ label: 'Nederlands', value: 'nl' },
];
export default function LanguageSwitcher() {
const pathname = usePathname();
const pathArray = pathname.split('/');
//Remove the first two elements from an array
const pathWithoutLang = pathArray.slice(2);
// join the array back to a string
const newPath = pathWithoutLang.join('/');
return (
<div className='px-1 bg-gray-200 rounded-lg outline-0'>
{LanguageOptions.map((locale) => (
<>
<Link
key={locale.label}
href={`/?lang=${locale.value + '/' + newPath}`}
as={`/${locale.value + '/' + newPath}`}
>
{locale.value}
</Link>
<br />
</>
))}
</div>
);
}
app/[lang]/layout.tsx
import { Inter } from 'next/font/google';
import { redirect } from 'next/navigation';
import React from 'react';
import i18n from '../../i18n';
import './globals.css';
const inter = Inter({ subsets: ['latin'] });
export const metadata = {
title: 'Create Next App',
description: 'Generated by creating next app',
};
export default function RootLayout({
children,
params: { lang },
}: {
children: React.ReactNode;
params: { lang: string };
}) {
// Redirect to default locale if lang is not supported. /second-page -> /en/second-page
if (!i18n.locales.includes(lang)) {
redirect(`/${i18n.defaultLocale}/${lang}`);
}
return (
<html lang={lang}>
<body className={inter.className}>{children}</body>
</html>
);
}
i18n.js
module.exports = {
locales: ['en', 'fr', 'nl'],
defaultLocale: 'en',
pages: {
'*': ['common'],
'/[lang]': ['form', 'login'],
'/[lang]/deneme': ['login', 'form'],
'/[lang]/onboarding': ['common'],
'/[lang]/signup': ['common', 'signup', 'login', 'form'],
},
};
// middleware.ts
import type { NextRequest } from 'next/server';
import { NextResponse } from 'next/server';
import i18n from './i18n';
// /es/page-name -> rewrites to -> /es/page-name?lang=es
export function middleware(request: NextRequest) {
const locale = request.nextUrl.locale || i18n.defaultLocale;
request.nextUrl.searchParams.set('lang', locale);
return NextResponse.rewrite(request.nextUrl);
}
What did you expect to happen?
I want to use Next js' new app directory with next-translate. I expect that rendering client/server components together with multiple languages
Are you willing to submit a pull request to fix this bug?
Try the latest 2.4.0 version please
BTW; you can remove the middleware
Thank you it fixed the issue. 🥳 I appreciate your help.
| gharchive/issue | 2023-06-23T20:55:42 | 2025-04-01T04:33:31.269988 | {
"authors": [
"aralroca",
"kod-man"
],
"repo": "aralroca/next-translate",
"url": "https://github.com/aralroca/next-translate/issues/1084",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
503440402 | Feature/aql subquery single input tests
This PR adds a set of disabled tests for the AQL subquery refactoring.
The plan is that we can continue implementing the missing parts in the code and one after the other activate the tests within this suite.
Contains tests for
PassThrough blocks
Sort
Nested empty collect
http://jenkins01.arangodb.biz:8080/view/PR/job/arangodb-matrix-pr/6615/
The suite might not be complete yet, please feel free to suggest more test cases / ideas.
LGTM.
Jenkins green besides the known Skip in Example
http://jenkins.arangodb.biz:8080/job/arangodb-matrix-pr/6632/
| gharchive/pull-request | 2019-10-07T13:06:28 | 2025-04-01T04:33:31.279365 | {
"authors": [
"markuspf",
"mchacki"
],
"repo": "arangodb/arangodb",
"url": "https://github.com/arangodb/arangodb/pull/10175",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
463682934 | @maierlars :heart_eyes:
Scope & Purpose
Port some small VelocyPack improvements from arangodb/velocypack to arangodb/arangodb
[x] Bug-Fix for devel-branch (i.e. no need for backports?)
[x] Strictly new functionality (i.e. a new feature / new option, no need for porting)
[x] The behaviour change can only be verified via automatic tests
Testing & Verification
This change is already covered by existing tests, such as velocypack tests.
[x] There are tests in an external testing repository (velocypack)
[x] I ensured this code runs with ASan / TSan or other static verification tools
https://jenkins01.arangodb.biz/view/PR/job/arangodb-matrix-pr/5041/
https://jenkins01.arangodb.biz/view/PR/job/arangodb-matrix-pr/5049/
https://jenkins01.arangodb.biz/view/PR/job/arangodb-matrix-pr/5059/
| gharchive/pull-request | 2019-07-03T10:51:31 | 2025-04-01T04:33:31.283148 | {
"authors": [
"jsteemann"
],
"repo": "arangodb/arangodb",
"url": "https://github.com/arangodb/arangodb/pull/9394",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
299374159 | Transaction: throw custom exception, with custom code
Hi,
Im' trying to throw a custom exception in transaction's action, but I can not set the errorNumber.
arangojs: 6.1.1
ArangoDB: 3.3.3 [linux] 64bit
RocksDB 5.6.0
Code:
const action = String((params) => {
var error = new Error('Conflict');
error.code = 409;
error.errorNum = 409;
error.statusCode = 409;
error.errorNumber = 409;
throw error;
});
try {
await DB.db.transaction(
{ read: 'test', write: 'test' },
action,
{ id: 1 }
);
} catch (error) {
console.log(error);
}
Output:
{
error: true,
errorMessage: 'Error: Conflict',
code: 500,
errorNum: 4
}
I expect to catch an exception with my custom error code.
Did I do something wrong?
Thanks!
Hi,
its working the following way:
const arangodb = require('@arangodb')
var err = new arangodb.ArangoError();
err.errorNum = arangodb.ERROR_BAD_PARAMETER; // <usually some well defined arango error
err.errorNum = 12345; /// but here some random value
err.errorMessage = "who's bad?";
throw err;
It works!
Thanks
Ok, I've fixed the documentation, and we now will also forward errorNum in other cases, plus added tests. Did you find the examples in the arangojs documentation or just from https://docs.arangodb.com/devel/Manual/Transactions/TransactionInvocation.html#custom-exceptions ?
| gharchive/issue | 2018-02-22T14:28:53 | 2025-04-01T04:33:31.287131 | {
"authors": [
"AntoineAA",
"dothebart"
],
"repo": "arangodb/arangojs",
"url": "https://github.com/arangodb/arangojs/issues/508",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
454755361 | Add docs for so far undocumented Aardvark API
Also see https://github.com/arangodb/planning/issues/3897
Aardvark Swagger JSON is available via Web UI (using documentation router?):
http://localhost:8529/_db/_system/_admin/aardvark/foxxes/docs/swagger.json?mount=/_admin/aardvark
No intention to document the internal Aardvark API.
| gharchive/pull-request | 2019-06-11T15:22:38 | 2025-04-01T04:33:31.288818 | {
"authors": [
"Simran-B"
],
"repo": "arangodb/docs",
"url": "https://github.com/arangodb/docs/pull/74",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1843146828 | MPI job aborts if MPI_Finalize is called in the same scope as a DistributedTree instance is destroyed
The following snippet makes MPI aborts the current job.
int main(int argc, char *argv[])
{
MPI_Init(&argc, &argv);
Kokkos::ScopeGuard guard(argc, argv);
Kokkos::View<ArborX::Point *, Kokkos::HostSpace> data("MpiError::data", 1);
ArborX::DistributedTree<Kokkos::HostSpace> tree(MPI_COMM_WORLD,
Kokkos::Serial{}, data);
MPI_Finalize();
return 0;
}
*** The MPI_Comm_free() function was called after MPI_FINALIZE was invoked.
*** This is disallowed by the MPI standard.
*** Your MPI job will now abort.
[mac122579:07159] Local abort after MPI_FINALIZE started completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
Moving the DistributedTree instance in another scope removes this error messages. I am guessing this error is due to the destructor of the internal comm pointer of DistributedTree. Adding a check to free the comm pointer if and only if MPI is still running does the trick.
[](MPI_Comm *p) {
int mpi_finalized;
MPI_Finalized(&mpi_finalized);
if (!mpi_finalized)
MPI_Comm_free(p);
delete p;
});
In other words, this is expected. We just need to make sure that this behavior is documented.
Related to https://github.com/arborx/Wiki/pull/53.
| gharchive/issue | 2023-08-09T12:36:47 | 2025-04-01T04:33:31.302866 | {
"authors": [
"dalg24",
"masterleinad",
"mrlag31"
],
"repo": "arborx/ArborX",
"url": "https://github.com/arborx/ArborX/issues/925",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1640477561 | useFormSplitter while triggering handlerSubmit validates the whole form and does not update (React Native)
Description
While creating a wizard form with multiple steps
BUG 1 - While doing handleSubmit from useFormSplitter the form doest not update if is not touched,
BUG 2- While doing handleSubmit from useFormSplitter is receiving all errors for the whole form instead of the separed step.
Steps to reproduce
1- Create a form with multiple steps using useFormSplitter
2 - On de step 1 trigger the handlerSubmit (if not touched fields will not rerender the component since the bug is clearing the changedKeys every time)
3 - If you change to the second step will appear the errors too even if we didn't do anything on the second step.
Expected result
2 - Update the screen with the correct errors
3 - Changing the step from 1 to 2 should not appear de errors on the 2 step
Actual result
Step 1 does not show the errors
Step 2 show errors event we didn't do anything
Environment
Node -v v14.21.2
"@resourge/react-form": "1.12.0"
"@resourge/schema": "1.9.4"
"react-native": "0.71.3"
I already made a fix for both of the bugs but needs a review.
Here is the patch with production bundle in RN
PATCH
The original fille changed with the fixes`()
@resourge\react-form\dist\umd\react-form.development.js
Next week i can create a PR if needed
| gharchive/issue | 2023-03-25T11:07:30 | 2025-04-01T04:33:31.374529 | {
"authors": [
"archie-rp"
],
"repo": "archie-rp/resourge-react-form",
"url": "https://github.com/archie-rp/resourge-react-form/issues/29",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1006539829 | Add isLocal argument to fromJsDate method of Datetime
Sometimes we want to transform a JavaScript Date to our Datetime wrapper AND keep the local time zone.
The time zone is always UTC in the current implementation:
static fromJsDate(date: Date) {
return new Datetime(LuxonDatetime.fromJSDate(date))
}
Because the Datetime constructor already supports the isLocal argument, I think the following implementation change is the more accurate:
static fromJsDate(date: Date, isLocal: boolean = false) {
return new Datetime(LuxonDatetime.fromJSDate(date), isLocal)
}
Maybe the second parameter should be an object in order to support more options in the future. Will that suffice @esuca?
This will be released in v2!
| gharchive/issue | 2021-09-24T14:43:30 | 2025-04-01T04:33:31.379770 | {
"authors": [
"cesalberca",
"esuca"
],
"repo": "archimedes-projects/archimedes-js",
"url": "https://github.com/archimedes-projects/archimedes-js/issues/7",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
902717071 | Add and improve Inventory validation
Thank you for helping out! ✨
We really appreciate your commitment to improving Architect
To maintain a high standard of quality in our releases, before merging every pull request we ask that you've completed the following:
[x] Forked the repo and created your branch from master
[x] Made sure tests pass (run npm it from the repo root)
[x] Expanded test coverage related to your changes:
[x] Added and/or updated unit tests (if appropriate)
[x] Added and/or updated integration tests (if appropriate)
[x] Updated relevant documentation:
[x] Internal to this repo (e.g. readme.md, help docs, inline docs & comments, etc.)
[x] Architect docs (arc.codes)
[ ] Summarized your changes in changelog.md
[ ] Linked to any related issues, PRs, etc. below that may relate to, consume, or necessitate these changes
Please also be sure to completed the CLA (if you haven't already).
Learn more about contributing to Architect here.
Thanks again!
😮
Ok, this wound up being a bit bigger than expected, but it was some necessary tech debt to pay down.
This PR should do the following:
Totally refactors error handling to support aggregation of many validation errors (instead of just one at a time)
Adds comprehensive validation to all pragmas that require validation
Gets unit test coverage to 💯%
Fixes a bunch of smol bugs along the way
Pretty much all the test changes were to failure cases. Since all the normal cases that were tested before are still passing now, this should be a drop-in – but given how much changed and how important Inventory is, I'd like to extensively test anyway!
Should fix:
https://github.com/architect/architect/issues/710
https://github.com/architect/architect/issues/951
| gharchive/pull-request | 2021-05-26T17:25:25 | 2025-04-01T04:33:31.388867 | {
"authors": [
"filmaj",
"ryanblock"
],
"repo": "architect/inventory",
"url": "https://github.com/architect/inventory/pull/19",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1883129068 | Input加一个props,可以在不同移动端设备上聚焦的时候只改变网页高度
[ ] I'm sure this does not appear in the issue list of the repository
Basic Info
What are the similar cases of this feature
none
What problem does this feature solve?
目前IOS设备上输入框聚焦会把网页弹起,想着如果可以聚焦的时候动态改变网页高度为屏幕高度减去键盘高度就更方便了 哈哈哈,如果可以的话,加一个这个props,true的时候有这个表现,false的时候就默认就好了
这个做不到,无法拿到键盘高度哈
| gharchive/issue | 2023-09-06T03:45:55 | 2025-04-01T04:33:31.427210 | {
"authors": [
"2513483494",
"TinaPeach"
],
"repo": "arco-design/arco-design-mobile",
"url": "https://github.com/arco-design/arco-design-mobile/issues/162",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
355561263 | When watching diff's with vimagit or fugitive it's bugged
Hello
I am using nord-vim with Hyper terminal (also nord themed), and everything is great - I love it! But I just started using vimagit (tried fugitive too) and when watching diff's the highlight is green/red and so is the text, so can't actually see what's changed.
Can someone help me fix this is the core or maybe a hotfix at my machine, since the theme is unusable together with those 2 plugins.
Here's a screenshot of the actual deal:
Hi @sebastianks and @tomgeorge :wave:
Thanks for your patience, I've been too busy with my "real" life and the development of the official Nord website to work on the massive amount of issues/PRs that were submitted each day for all of Nord's port projects.
@arcticicestudio I think mine was due to not using Nord in Hyper as I thought I did. There's no issue for me when using Hyper+Nord and Vim+Nord.
@tomgeorge If the solution I suggested above doesn't work please open a new issue. This way it makes it easier for me to keep track of all the currently open issues.
| gharchive/issue | 2018-08-30T12:24:08 | 2025-04-01T04:33:31.430546 | {
"authors": [
"arcticicestudio",
"sebastianks"
],
"repo": "arcticicestudio/nord-vim",
"url": "https://github.com/arcticicestudio/nord-vim/issues/136",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
412724709 | Do not highlight function arguments in Python
I feel like arguments to functions should not be colored (especially with the same color as the function itself!), as this can lead to confusion.
For example, here's some code to demonstrate the issue:
At a glance, it is unclear that prev_x is a variable while babylon() is a function.
I have found the fix to the problem... removing "source.python meta.function-call" from line 1004 in /themes/nord.json.
If you would like I can submit a PR sometime this weekend.
@Al2Me6 Sorry for the long stale state.
Thanks again for your report and fix :+1:
It works fine so I've created PR #113 to get this into the next release version.
| gharchive/issue | 2019-02-21T02:54:18 | 2025-04-01T04:33:31.433117 | {
"authors": [
"Al2Me6",
"arcticicestudio"
],
"repo": "arcticicestudio/nord-visual-studio-code",
"url": "https://github.com/arcticicestudio/nord-visual-studio-code/issues/108",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2699689833 | feat: Add option InjectAsBase64 for ARM
Unfortunately, I don't know how to run the test locally. Also don't know how to update the version number.
Hi Martin, great thank you for your addition and work on Arcus.
Couple of things:
[ ] Please read the contribution guide to answer already couple of your questions;
[ ] Please revert the changes in the v1.3.1 folder, it is not up to contributors to determine when we release. If changes are needed to the docs, please change them in the /preview folder;
[ ] After that, we can start reviewing.
Thx again!
I cannot run the tests locally for some reason:
PS C:\Users\twinp> Invoke-Pester C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Arcus.Scripting.ARM.tests.ps1 WARNING: The names of some imported commands from the module 'Arcus.Scripting.ARM' include unapproved verbs that might make them less discoverable. To find the commands with unapproved verbs, run the Import-Module command again with the Verbose parameter. For a list of approved verbs, type Get-Verb. Describing Arcus ARM integration tests
Context ARM injection
Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-inline-value.json' for injection tokens Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-inline.json' for injection tokens Finished script 'Inject-ArmContent.ps1' on path 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-inline.json' [-] Replaces file path with inline file contents 773ms RuntimeException: '-Be' is not a valid Should operator. at Get-TestResult, C:\Program Files\WindowsPowerShell\Modules\Pester\3.4.0\Functions\Assertions\Should.ps1: line 42 at , C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Arcus.Scripting.ARM.tests.ps1: line 18 Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-object-value (windows).json' for injection tokens
I cannot run the tests locally for some reason:
RuntimeException: '-Be' is not a valid Should operator. at Get-TestResult, C:\Program Files\WindowsPowerShell\Modules\Pester\3.4.0\Functions\Assertions\Should.ps1: line 42 at , C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Arcus.Scripting.ARM.tests.ps1: line 18
This shows a conflicting local PS Pester package version as there were breaking changes regarding the -Be operator in some of their versions.
/az run CI - Arcus.Scripting
I was able to run the test locally after upgrading Pester to 5.6.1
/azp run CI - Arcus.Scripting
You can delete this PR. I'm giving it up. Takes me too much time. Test is succesful locally, but the same test fails in the pipeline.
Will find another solution.
PS C:\Windows\System32> Invoke-Pester C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Arcus.Scripting.ARM.tests.ps1
Starting discovery in 1 files.
WARNING: The names of some imported commands from the module 'Arcus.Scripting.ARM' include unapproved verbs that might
make them less discoverable. To find the commands with unapproved verbs, run the Import-Module command again with the
Verbose parameter. For a list of approved verbs, type Get-Verb.
Discovery found 5 tests in 1.99s.
Running tests.
Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-inline-value.json' for injection tokens
Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-inline.json' for injection tokens
Finished script 'Inject-ArmContent.ps1' on path 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-inline.json'
Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-object-value (windows).json' for injection tokens
WARNING: Content to inject into ARM template file
'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-object (windows).json' cannot be
parsed as a JSON object!
Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-object (windows).json' for injection tokens
Finished script 'Inject-ArmContent.ps1' on path 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-object (windows).json'
Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-object-value (windows).json' for injection tokens
WARNING: Content to inject into ARM template file
'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-object-absolutepath (windows).json'
cannot be parsed as a JSON object!
Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-object-absolutepath (windows).json' for injection tokens
Finished script 'Inject-ArmContent.ps1' on path 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-object-absolutepath (windows).json'
Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-escape-value.xml' for injection tokens
Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-escape.json' for injection tokens
Finished script 'Inject-ArmContent.ps1' on path 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-escape.json'
Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\codit.eu.crt' for injection tokens
Done checking ARM template file 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-certificate.json' for injection tokens
Finished script 'Inject-ArmContent.ps1' on path 'C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Files\arm-template-certificate.json'
[+] C:\Repos\arcus.scripting\src\Arcus.Scripting.Tests.Integration\Arcus.Scripting.ARM.tests.ps1 7.62s (2.09s|3.77s)
Tests completed in 7.78s
Tests Passed: 5, Failed: 0, Skipped: 0, Inconclusive: 0, NotRun: 0
You can delete this PR. I'm giving it up. Takes me too much time. Test is succesful locally, but the same test fails in the pipeline. Will find another solution.
It's probably a different test remotely, as it succeeds remotely on windows but fails on linux (and I'm guessting that you work with a windows system locally).
Putting @pim-simons in CC.
/azp run CI - Arcus.Scripting
Turns out base64 encoding on Linux returns a bit of a different string than on Windows. Decoding both base64 strings from Linux and Windows returns the same correct file so I dont really think it is an issue. Created seperate Windows and Linux tests and all working fine now 👍🏻
Thanks Stijn and Pim for finding and fixing this.
Thanks Stijn and Pim for finding and fixing this.
Thanks for contributing! 🥇
@twinpiloot PR doesn't have to be closed, will wait for @stijnmoreels to review and then we can squash and merge to main 👍🏻
Its the public certificate of the codit website. Just used for the test
Verzonden vanaf mijn Galaxy
-------- Oorspronkelijk bericht --------
Van: Stijn Moreels @.>
Datum: 29-11-2024 21:19 (GMT+01:00)
Aan: "arcus-azure/arcus.scripting" @.>
Cc: Martin Peters @.>, Mention @.>
Onderwerp: Re: [arcus-azure/arcus.scripting] feat: Add option InjectAsBase64 for ARM (PR #438)
@stijnmoreels approved this pull request.
Fine for me as well.
On src/Arcus.Scripting.Tests.Integration/Files/codit.eu.crthttps://github.com/arcus-azure/arcus.scripting/pull/438#discussion_r1863909719:
Is this some self-signed public certificate?
—
Reply to this email directly, view it on GitHubhttps://github.com/arcus-azure/arcus.scripting/pull/438#pullrequestreview-2470477816, or unsubscribehttps://github.com/notifications/unsubscribe-auth/ABRMHWDUD5E4APGBABE3IZ32DDD5HAVCNFSM6AAAAABSTUHFKOVHI2DSMVQWIX3LMV43YUDVNRWFEZLROVSXG5CSMV3GSZLXHMZDINZQGQ3TOOBRGY.
You are receiving this because you were mentioned.Message ID: @.***>
| gharchive/pull-request | 2024-11-27T20:10:07 | 2025-04-01T04:33:31.458133 | {
"authors": [
"pim-simons",
"stijnmoreels",
"twinpiloot"
],
"repo": "arcus-azure/arcus.scripting",
"url": "https://github.com/arcus-azure/arcus.scripting/pull/438",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2757885457 | fix(server): add waitUntil to the server context in Fastify integration
While calling handleNodeRequest or handleNodeRequestAndResponse, waitUntil is not added automatically as in requestListener for Node.js integration.
This change adds waitUntil into the serverContext if not present.
Fixes the issue with Fastify integration that uses the mentioned methods
@ardatan do we now know that all other integrations supports this and we have similar tests for all of the other integrations?
| gharchive/pull-request | 2024-12-24T14:41:49 | 2025-04-01T04:33:31.461807 | {
"authors": [
"Urigo",
"ardatan"
],
"repo": "ardatan/whatwg-node",
"url": "https://github.com/ardatan/whatwg-node/pull/1926",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1710102150 | Error saat ingin convert Inferencing
bang pas saya convert error terus
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/gradio/routes.py", line 414, in run_predict
output = await app.get_blocks().process_api(
File "/usr/local/lib/python3.10/dist-packages/gradio/blocks.py", line 1323, in process_api
data = self.postprocess_data(fn_index, result["prediction"], state)
File "/usr/local/lib/python3.10/dist-packages/gradio/blocks.py", line 1257, in postprocess_data
prediction_value = block.postprocess(prediction_value)
File "/usr/local/lib/python3.10/dist-packages/gradio/components.py", line 2531, in postprocess
processing_utils.audio_to_file(sample_rate, data, file.name)
File "/usr/local/lib/python3.10/dist-packages/gradio/processing_utils.py", line 163, in audio_to_file
data = convert_to_16_bit_wav(data)
File "/usr/local/lib/python3.10/dist-packages/gradio/processing_utils.py", line 177, in convert_to_16_bit_wav
if data.dtype in [np.float64, np.float32, np.float16]:
AttributeError: 'NoneType' object has no attribute 'dtype'
Traceback (most recent call last):
File "/content/Retrieval-based-Voice-Conversion-WebUI/infer-web.py", line 146, in vc_single
if_f0 = cpt.get("f0", 1)
NameError: name 'cpt' is not defined
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/gradio/routes.py", line 414, in run_predict
output = await app.get_blocks().process_api(
File "/usr/local/lib/python3.10/dist-packages/gradio/blocks.py", line 1323, in process_api
data = self.postprocess_data(fn_index, result["prediction"], state)
File "/usr/local/lib/python3.10/dist-packages/gradio/blocks.py", line 1257, in postprocess_data
prediction_value = block.postprocess(prediction_value)
File "/usr/local/lib/python3.10/dist-packages/gradio/components.py", line 2531, in postprocess
processing_utils.audio_to_file(sample_rate, data, file.name)
File "/usr/local/lib/python3.10/dist-packages/gradio/processing_utils.py", line 163, in audio_to_file
data = convert_to_16_bit_wav(data)
File "/usr/local/lib/python3.10/dist-packages/gradio/processing_utils.py", line 177, in convert_to_16_bit_wav
if data.dtype in [np.float64, np.float32, np.float16]:
AttributeError: 'NoneType' object has no attribute 'dtype'
Kirim ssnya
Itu inferencing timbrenya masih kosong
Lah iya ga nyadar wkwkwk, langsung saya close aja
| gharchive/issue | 2023-05-15T13:25:51 | 2025-04-01T04:33:31.465014 | {
"authors": [
"Reno121",
"ardha27"
],
"repo": "ardha27/AI-Song-Cover-RVC",
"url": "https://github.com/ardha27/AI-Song-Cover-RVC/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1800619148 | Wire: fix setClock
clock source is fixed for RA4M1, so the rates will be wrong for C33.
This will be fixed in the next commit
@facchinm
Have a question on the settings that you are using for 400khz. If I attach a Logic analyzer to SDA/SCL while running a sketch pulling data from a MPU-9250 I am seeing:
which I believe is showing a frequency of about 257khz not close to the 400khz.
However if I change your settings to:
m_i2c_extend.clock_settings.brl_value = 12;
m_i2c_extend.clock_settings.brh_value = 11;
m_i2c_extend.clock_settings.cks_value = 0;
I get closer to 400khz (357khz):
@facchinm
These settings work better - gives 400khz with 60% duty cycle
case I2C_MASTER_RATE_FAST:
m_i2c_extend.clock_settings.brl_value = 16;
m_i2c_extend.clock_settings.brh_value = 15;
m_i2c_extend.clock_settings.cks_value = 0;
setClock should not only allow the three values, 100kHz, 400kHz, 1Mhz. See this issue.
@mjs513 @greiman I pushed a new version of the patch which should take care of the clocks for both UNO and C33.
In this patch I explicitly disabled FAST_MODE_PLUS for platforms not supporting it (eg. RA4M1) so we are in the specifications of the chip.
Let me know if it works for you, then we can merge it safely.
@facchinm
Got some strange things happening when testing the changes at 400khz.
Using a 5v liquidCrystal_i2c device on Wire with your current settings of 16/15 I am seeing 500khz using a LA:
If I change the settings for 400khz to 22/21 I am seeing 400khz, again for 5v device on Wire:
But if I use those settings 22/21 for a MPU-9250 running at 3.3v on Wire I am seeing something like 257khz
Using your settings, again on wire for a MPU-9250 I am getting 294khz:
Confused!!!!!!
And if I move the 9250 over to Wire1 using your settings of 16/15 I am seeing that 526khz again
and if I use the 22/21 settings I am seeing 416khz at 52% duty cycle.
So now I am confused. Please note I do have the Micros propose PR incorporated as well as a couple of others.
@facchinm
out of curiosity I pulled out a T4.1 (had it handy) and check 100khz and 400khz. At 100khz seeing 68khz with 50% while at 400khz I am seeing 357khz clock with 50% duty cycle for my MPU-9250.
Never one to leave enough alone:
Adafruit Metro M0 express: 97khz with 50% duty at 100Khz; 357khz with 50% duty at 400khz
Uno R3 Clone: 370khz with 52% duty at 400khz; 100khz with 49% duty at 100khz setclock.
Wire1 has level shifters and works correctly for 3V3 devices. Wire is marginal for 3V3 devices. For 3V3 devices on Wire, the speed depends on pull-up resistance, pull-up voltage, and VCC for the RA4M1. See this.
@facchinm - just wanted to know if there was anything else you wanted me to try with this?
@facchinm - maybe its a good idea to go ahead and incorporate the current fix (this pr) into the core. Seems to be resolve some issues people are having: https://forum.arduino.cc/t/i2c-setclock-does-not-appear-to-be-standard-way-to-setclock/1147122/19
As @facchinm is currently out-of-office I'll move ahead and merge this one.
| gharchive/pull-request | 2023-07-12T09:59:39 | 2025-04-01T04:33:31.526222 | {
"authors": [
"aentinger",
"facchinm",
"greiman",
"mjs513"
],
"repo": "arduino/ArduinoCore-renesas",
"url": "https://github.com/arduino/ArduinoCore-renesas/pull/46",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1857665184 | 🛑 U3A Raspberry Pis is down
In 183cf9e, U3A Raspberry Pis (http://u3a.epizy.com/rpis) was down:
HTTP code: 0
Response time: 0 ms
Resolved: U3A Raspberry Pis is back up in 3782af6 after 750 days, 22 hours, 25 minutes.
| gharchive/issue | 2023-08-19T10:26:09 | 2025-04-01T04:33:31.550310 | {
"authors": [
"ardus-uk"
],
"repo": "ardus-uk/upptime",
"url": "https://github.com/ardus-uk/upptime/issues/297",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
346832380 | An error occurred when I ran the command 'php artisan twill:setup'
Hi @linqiuyu!
Thanks for reporting. We are aware of this one and have a fix coming up but you don't have to worry about it. This is old code trying to locate a folder that does not exist anymore. As long as the rest of the setup worked, you're good!
Hi everybody!
Good job @ifox
I have save same problem today.
Thanks!
| gharchive/issue | 2018-08-02T02:06:46 | 2025-04-01T04:33:31.554941 | {
"authors": [
"dehboris",
"ifox",
"linqiuyu"
],
"repo": "area17/twill",
"url": "https://github.com/area17/twill/issues/25",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2725061306 | Question: potential bug while testing AREG_BUILD_TESTS
Hi @aregtech
I am not familiar with cmake, but there seems to be a bug in conf/cmake/setup.cmake while testing AREG_BUILD_TESTS
if (NOT AREG_BUILD_TESTS)
set(AREG_GTEST_PACKAGE OFF)
endif()
Although, I could't find any information in the main README how to build/run/disable unittest
BR /biwiki
Hi @biwiki,
It is not a bug. It says that if "build of unit tests are disabled, to not use GTest package". It plays no real role, simply set it, because at the end of configuration it types a report. Previously, it unit tests were disabled, it was typing "Unit tests == OFF, use GTest package == ON", which I found a little confusing.
Regarding disabling building Unit Tests, actually, there is a documentation CMake Configuration Options for Building AREG SDK. But you are right, such simple things like "disable unit tests" or "disable examples" are simple and they should be clearly described. Thank you for the note, I'll keep in mind and update documents.
To keep simple,
Disable unit test build:cmake -B ./build -DAREG_BUILD_TESTS=OFF
Disable examples:cmake -B ./build -DAREG_BUILD_EXAMPLES=OFF
I keep this ticket opened to update documents.
Thamks @aregtech , that would be so helpful!
@biwiki , I appreciate very much if you report uncomfortable / confusing codes or descriptions.
I like your notes. They are reasonable.
Thanks @aregtech .
Basically I knew how to build and run the tests, but it is a bit more convenience to have instructions on the main README since I have seen that quite often in other repos.
@biwiki, this is what I mean. You notice the details I might miss. Even the most experienced developers can overlook certain use cases, which might lead to bugs -- it’s just part of the process. Your experience with frameworks and repositories really adds value here, and your feedback is helpful. My goal is to make this framework as useful and user-friendly as possible, so your insights are truly appreciated 🤝
Good point, this can be added in the HOWTO.md document. Or do you think i should create a separate document like "Tests and examples"?
To me, it should be in a separate doco but I am not sure if that would be the best
| gharchive/issue | 2024-12-08T09:01:44 | 2025-04-01T04:33:31.566855 | {
"authors": [
"aregtech",
"biwiki"
],
"repo": "aregtech/areg-sdk",
"url": "https://github.com/aregtech/areg-sdk/issues/466",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1307144511 | Gabe's review of the checklist
Hi Gabe, this is intended as a practice PR to get you familiar with the writing checklist and part of the code review process.
What we're looking for you to do is, after completing the MOOC writing course, go over the existing writing checklist and see if there's anything to add, clarify, or change about it.
If you have any questions @ mention me in this issue, and we can work on a PR or hash it out.
This issue can be closed when you confirm that you've gone through the writing checklist and a PR has been merged to add anything you think should be added, or you sign off saying it's all good!
I have a suggestion for a correction that you might pick up when you read through. If you click Hidden Suggestion, you'll see it.
Hidden Suggestion
The line, "Abbreviations including et (e.g., "et al.") should not have a period after "et" because it is a whole word," doesn't really make sense. If we're saying "et" isn't actually an abbreviation, this should be rewritten!
| gharchive/issue | 2022-07-17T16:21:34 | 2025-04-01T04:33:31.775519 | {
"authors": [
"nsryan2"
],
"repo": "arfc/arfc.github.io",
"url": "https://github.com/arfc/arfc.github.io/issues/264",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1601162033 | Documentation update: tutorial for text classification models comparison
Description
Context: https://github.com/argilla-io/argilla/issues/2068
In this PR a new tutorial is added: model-comparison for text classification. It is the follow up of the work done during PyConFr in Bordeaux.
Closes #2068
Type of change
(Please delete options that are not relevant. Remember to title the PR according to the type of change)
[ ] Bug fix (non-breaking change which fixes an issue)
[ ] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
[ ] Refactor (change restructuring the codebase without changing functionality)
[ ] Improvement (change adding some improvement to an existing functionality)
[x] Documentation update
How Has This Been Tested
(Please describe the tests that you ran to verify your changes. And ideally, reference tests)
[ ] Test A
[ ] Test B
Checklist
[ ] I have merged the original branch into my forked branch
[ ] I added relevant documentation
[ ] follows the style guidelines of this project
[ ] I did a self-review of my code
[ ] I made corresponding changes to the documentation
[ ] My changes generate no new warnings
[ ] I have added tests that prove my fix is effective or that my feature works
This PR is a WIP, because I haven't figured out how to add the notebook in docs/source/tutorials
Hi @embonhomme this is super cool and useful!!
In order to make it even more useful, would it be possible to use SetFit's zeroshot model instead of the fewshot classy? We've just published a tutorial to show how easy is to use SetFit and many people are asking about comparisons with the zeroshot HF pipeline, so this tutorial would be awesome and make for a better comparison? https://docs.argilla.io/en/latest/tutorials/notebooks/labelling-textclassification-setfit-zeroshot.html#🔫-Zero-shot-predictions-with-SetFit
We'd be happy to walk you through if you have questions.
Hello @dvsrepo :) Thank you for the feedback, you can find in the new commit the comparaison with SetFit zero-shot.
Tell me if it is relevant.
This is looking just perfect!
The only one remaining change would be to review the remaining mentions of few-shot and classy-classification and replace them with zero-shot and SetFit. Then we are good to go!
We'd love to share this next week via LinkedIn and Twitter, if you'd like us to mention you as the author, send me an email to daniel @ argilla.io
Thank you! Yes sorry I totally forgot to change the description part. It should be better now.
Also here I have just added a Jupyter Notebook, I didn't figure out how it works with the modal.md, dvc.md,...
I will send you an email with my LinkedIn :)
Great stuff @embonhomme!
@embonhomme, in case you can tackle this, the process is:
Create a folder here with the filename, like this one: https://github.com/argilla-io/argilla/tree/develop/docs/_source/_static/tutorials/training-textclassification-setfit-fewshot
Add the reference to the tutorial: here https://github.com/argilla-io/argilla/blob/develop/docs/_source/tutorials/libraries/setfit.md, here: https://github.com/argilla-io/argilla/blob/develop/docs/_source/tutorials/steps/4_monitoring.md , here: https://github.com/argilla-io/argilla/blob/develop/docs/_source/tutorials/tasks/text_classification.md and here: https://github.com/argilla-io/argilla/blob/develop/docs/_source/tutorials/techniques/few_shot.md
I would say this tutorial is about Monitoring, TextClassification, few-shot
Otherwise, let us know and @davidberenstein1957 might be able to help
Thank you, I did the integration :)
Hi @davidberenstein1957 I renamed everything :)
Yes, I would like to participate in the LinkedIn shoutout!
Lovely!
| gharchive/pull-request | 2023-02-27T13:11:48 | 2025-04-01T04:33:31.823678 | {
"authors": [
"davidberenstein1957",
"dvsrepo",
"embonhomme"
],
"repo": "argilla-io/argilla",
"url": "https://github.com/argilla-io/argilla/pull/2426",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1167517517 | chore: Bump Kubernetes dependencies to version 1.22.4
Signed-off-by: jannfis jann@mistrust.net
Codecov Report
Merging #394 (383ebe3) into master (f2eb28b) will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #394 +/- ##
=======================================
Coverage 65.60% 65.60%
=======================================
Files 21 21
Lines 2035 2035
=======================================
Hits 1335 1335
Misses 571 571
Partials 129 129
:mega: Codecov can now indicate which changes are the most critical in Pull Requests. Learn more
| gharchive/pull-request | 2022-03-13T08:36:32 | 2025-04-01T04:33:31.833531 | {
"authors": [
"codecov-commenter",
"jannfis"
],
"repo": "argoproj-labs/argocd-image-updater",
"url": "https://github.com/argoproj-labs/argocd-image-updater/pull/394",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2165601173 | Support defining custom cluster roles for cluster scoped instances
Is your feature request related to a problem? Please describe.
When using a cluster scoped instance of Argo CD the operator will create a separate and hard coded set of clusterroles and clusterrolebindings. Users that want to lock down the permissions of their instances need the ability to use an alternate set of clusterroles.
As an example, the new feature in Argo CD 2.10 to auto-respect RBAC (https://argo-cd.readthedocs.io/en/stable/operator-manual/declarative-setup/#auto-respect-rbac-for-controller) enables users to remove the cluster-wide view all permissions required by a cluster scoped instance and provide a much more tailored set of permissions. However the Argo CD Operator hard codes this cluster role to define view all permissions.
Describe the solution you'd like
Enable users to specify their own roles using the environment variables CONTROLLER_CLUSTER_SCOPE_ROLE and SERVER_CLUSTER_SCOPE_ROLE
Describe alternatives you've considered
Updates to the cluster role created by the operator get overridden by the operator so it's not a suitable approach.
PR is in progress for this.
Reopening since PR was reverted and there is a new PR raised to address this.
| gharchive/issue | 2024-03-03T22:35:58 | 2025-04-01T04:33:31.836738 | {
"authors": [
"gnunn1",
"jparsai"
],
"repo": "argoproj-labs/argocd-operator",
"url": "https://github.com/argoproj-labs/argocd-operator/issues/1275",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1139156120 | chore: add missing bundle versions
Signed-off-by: John Pitman jpitman@redhat.com
What type of PR is this?
/kind chore
What does this PR do / why we need it:
After releasing 0.2.0 of the argocd-operator, we didn't set up the master branch for the next version. That means on master, any changes to the CRD have gone into the 0.2.0 bundle folder. Since we've already split off the branch for 0.3.0, the next version should be 0.4.0. This PR is to fix this situation
Have you updated the necessary documentation?
[x] Documentation update is required by this PR.
[x] Documentation has been updated.
Which issue(s) this PR fixes:
Had a call/discussion with @jopit on why the operator version is updated as 0.4.0 instead of 0.3.0.
Since we already branched out 0.3.0(release-0.3) and we know some changes that are in master are not going into release-0.3, it is better to update the operator version to 0.4.0 in master. This way when we create the next release branch we dont have to update the operator version and manifests in the release branch.
However, I see an overhead of merging back the 0.3.0 bundle folder in release-0.3 branch back into master after/during v0.3.0 release.
@jopit and I have decided that
we shall merge this PR for now, if not the 0.2.0 bundle folder in the master would keep updating.
we can decide about this overhead in the next cabal, if there is a change in the decision we can act accordingly with a new PR if required.
| gharchive/pull-request | 2022-02-15T19:59:31 | 2025-04-01T04:33:31.842318 | {
"authors": [
"iam-veeramalla",
"jopit"
],
"repo": "argoproj-labs/argocd-operator",
"url": "https://github.com/argoproj-labs/argocd-operator/pull/563",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2050918782 | quick start fails out of the box
Pre-requisites
[X] I have double-checked my configuration
[X] I can confirm the issues exists when I tested with :latest
[ ] I'd like to contribute the fix myself (see contributing guide)
What happened/what did you expect to happen?
Able to run the quick start CLI with success.
My environment was a macOS M2 machine running Docker Desktop with kind. The steps are straight from https://argoproj.github.io/argo-workflows/quick-start/
kind create cluster
# cluster created with 1.27.3
kubectl create namespace argo
kubectl apply -n argo -f https://github.com/argoproj/argo-workflows/releases/download/v3.5.2/install.yaml
kubectl patch deployment \
argo-server \
--namespace argo \
--type='json' \
-p='[{"op": "replace", "path": "/spec/template/spec/containers/0/args", "value": [
"server",
"--auth-mode=server"
]}]'
At this point I ran:
argo submit -n argo --watch https://raw.githubusercontent.com/argoproj/argo-workflows/main/examples/hello-world.yaml
And the output was:
Name: hello-world-g2csp
Namespace: argo
ServiceAccount: unset (will run with the default ServiceAccount)
Status: Error
Message: Error (exit code 1): pods "hello-world-g2csp" is forbidden: User "system:serviceaccount:argo:default" cannot patch resource "pods" in API group "" in the namespace "argo"
Conditions:
PodRunning False
Completed True
Created: Wed Dec 20 10:25:35 -0600 (9 minutes ago)
Started: Wed Dec 20 10:25:35 -0600 (9 minutes ago)
Finished: Wed Dec 20 10:26:05 -0600 (9 minutes ago)
Duration: 30 seconds
Progress: 0/1
ResourcesDuration: 13s*(1 cpu),13s*(100Mi memory)
STEP TEMPLATE PODNAME DURATION MESSAGE
⚠ hello-world-g2csp whalesay hello-world-g2csp 21s Error (exit code 1): pods "hello-world-g2csp" is forbidden: User "system:serviceaccount:argo:default" cannot patch resource "pods" in API group "" in the namespace "argo"
Version
v3.5.2
Paste a small workflow that reproduces the issue. We must be able to run the workflow; don't enter a workflows that uses private images.
apiVersion: argoproj.io/v1alpha1
kind: Workflow
metadata:
generateName: hello-world-
labels:
workflows.argoproj.io/archive-strategy: "false"
annotations:
workflows.argoproj.io/description: |
This is a simple hello world example.
spec:
entrypoint: whalesay
templates:
- name: whalesay
container:
image: docker/whalesay:latest
command: [cowsay]
args: ["hello world"]
Logs from the workflow controller
time="2023-12-20T16:25:35.731Z" level=info msg="Processing workflow" namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:25:35.736Z" level=info msg="Task-result reconciliation" namespace=argo numObjs=0 workflow=hello-world-g2csp
time="2023-12-20T16:25:35.736Z" level=info msg="Updated phase -> Running" namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:25:35.736Z" level=warning msg="Node was nil, will be initialized as type Skipped" namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:25:35.736Z" level=info msg="was unable to obtain node for , letting display name to be nodeName" namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:25:35.736Z" level=info msg="Pod node hello-world-g2csp initialized Pending" namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:25:35.745Z" level=info msg="Created pod: hello-world-g2csp (hello-world-g2csp)" namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:25:35.746Z" level=info msg="TaskSet Reconciliation" namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:25:35.746Z" level=info msg=reconcileAgentPod namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:25:35.752Z" level=info msg="Workflow update successful" namespace=argo phase=Running resourceVersion=2676 workflow=hello-world-g2csp
time="2023-12-20T16:25:45.751Z" level=info msg="Processing workflow" namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:25:45.752Z" level=info msg="Task-result reconciliation" namespace=argo numObjs=0 workflow=hello-world-g2csp
time="2023-12-20T16:25:45.752Z" level=info msg="node changed" namespace=argo new.message=PodInitializing new.phase=Pending new.progress=0/1 nodeID=hello-world-g2csp old.message= old.phase=Pending old.progress=0/1 workflow=hello-world-g2csp
time="2023-12-20T16:25:45.752Z" level=info msg="TaskSet Reconciliation" namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:25:45.752Z" level=info msg=reconcileAgentPod namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:25:45.765Z" level=info msg="Workflow update successful" namespace=argo phase=Running resourceVersion=2703 workflow=hello-world-g2csp
time="2023-12-20T16:25:55.768Z" level=info msg="Processing workflow" namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:25:55.768Z" level=info msg="Task-result reconciliation" namespace=argo numObjs=0 workflow=hello-world-g2csp
time="2023-12-20T16:25:55.768Z" level=info msg="node unchanged" namespace=argo nodeID=hello-world-g2csp workflow=hello-world-g2csp
time="2023-12-20T16:25:55.769Z" level=info msg="TaskSet Reconciliation" namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:25:55.769Z" level=info msg=reconcileAgentPod namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:26:05.933Z" level=info msg="Processing workflow" namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:26:05.933Z" level=info msg="Task-result reconciliation" namespace=argo numObjs=0 workflow=hello-world-g2csp
time="2023-12-20T16:26:05.934Z" level=info msg="Pod failed: Error (exit code 1): pods \"hello-world-g2csp\" is forbidden: User \"system:serviceaccount:argo:default\" cannot patch resource \"pods\" in API group \"\" in the namespace \"argo\"" displayName=hello-world-g2csp namespace=argo pod=hello-world-g2csp templateName=whalesay workflow=hello-world-g2csp
time="2023-12-20T16:26:05.934Z" level=info msg="node changed" namespace=argo new.message="Error (exit code 1): pods \"hello-world-g2csp\" is forbidden: User \"system:serviceaccount:argo:default\" cannot patch resource \"pods\" in API group \"\" in the namespace \"argo\"" new.phase=Error new.progress=0/1 nodeID=hello-world-g2csp old.message=PodInitializing old.phase=Pending old.progress=0/1 workflow=hello-world-g2csp
time="2023-12-20T16:26:05.934Z" level=info msg="TaskSet Reconciliation" namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:26:05.934Z" level=info msg=reconcileAgentPod namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:26:05.934Z" level=info msg="Updated phase Running -> Error" namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:26:05.934Z" level=info msg="Updated message -> Error (exit code 1): pods \"hello-world-g2csp\" is forbidden: User \"system:serviceaccount:argo:default\" cannot patch resource \"pods\" in API group \"\" in the namespace \"argo\"" namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:26:05.934Z" level=info msg="Marking workflow completed" namespace=argo workflow=hello-world-g2csp
time="2023-12-20T16:26:05.941Z" level=info msg="cleaning up pod" action=deletePod key=argo/hello-world-g2csp-1340600742-agent/deletePod
time="2023-12-20T16:26:05.951Z" level=info msg="Workflow update successful" namespace=argo phase=Error resourceVersion=2745 workflow=hello-world-g2csp
time="2023-12-20T16:26:05.959Z" level=info msg="cleaning up pod" action=labelPodCompleted key=argo/hello-world-g2csp/labelPodCompleted
Logs from in your workflow's wait container
time="2023-12-20T16:25:56.956Z" level=info msg="No output parameters"
time="2023-12-20T16:25:56.956Z" level=info msg="No output artifacts"
time="2023-12-20T16:25:56.967Z" level=warning msg="failed to patch task set, falling back to legacy/insecure pod patch, see https://argoproj.github.io/argo-workflows/workflow-rbac/" error="workflowtaskresults.argoproj.io is forbidden: User \"system:serviceaccount:argo:default\" cannot create resource \"workflowtaskresults\" in API group \"argoproj.io\" in the namespace \"argo\""
time="2023-12-20T16:25:56.968Z" level=warning msg="Non-transient error: pods \"hello-world-g2csp\" is forbidden: User \"system:serviceaccount:argo:default\" cannot patch resource \"pods\" in API group \"\" in the namespace \"argo\""
time="2023-12-20T16:25:56.968Z" level=error msg="executor error: pods \"hello-world-g2csp\" is forbidden: User \"system:serviceaccount:argo:default\" cannot patch resource \"pods\" in API group \"\" in the namespace \"argo\""
time="2023-12-20T16:25:56.968Z" level=info msg="Alloc=7218 TotalAlloc=13045 Sys=24677 NumGC=4 Goroutines=8"
time="2023-12-20T16:25:56.969Z" level=warning msg="failed to patch task set, falling back to legacy/insecure pod patch, see https://argoproj.github.io/argo-workflows/workflow-rbac/" error="workflowtaskresults.argoproj.io \"hello-world-g2csp\" is forbidden: User \"system:serviceaccount:argo:default\" cannot patch resource \"workflowtaskresults\" in API group \"argoproj.io\" in the namespace \"argo\""
time="2023-12-20T16:25:56.969Z" level=warning msg="Non-transient error: pods \"hello-world-g2csp\" is forbidden: User \"system:serviceaccount:argo:default\" cannot patch resource \"pods\" in API group \"\" in the namespace \"argo\""
time="2023-12-20T16:25:56.969Z" level=error msg="executor error: pods \"hello-world-g2csp\" is forbidden: User \"system:serviceaccount:argo:default\" cannot patch resource \"pods\" in API group \"\" in the namespace \"argo\""
time="2023-12-20T16:25:56.970Z" level=fatal msg="pods \"hello-world-g2csp\" is forbidden: User \"system:serviceaccount:argo:default\" cannot patch resource \"pods\" in API group \"\" in the namespace \"argo\""
I failed to mention that adding --serviceaccount argo to the argo submit made it work.
Would you like to submit a PR to help improve the docs?
Would you like to submit a PR to help improve the docs?
Would be glad to. I'm just not sure if that's what the maintainers want from the quick start. If it is then I'll happily update the docs and include a reference to the workflow RBAC for further reading and make a note that when a service account isn't specified then the default account is used which often doesn't have sufficient permissions.
Looks like the quick start and walk-through guides all use default service account which does not have the permission. We can probably just give default service-account enough permission https://github.com/argoproj/argo-workflows/tree/76f4d1f9ca53b8ac48965e12c15454e544936c49/manifests/quick-start/base
Looks like the quick start and walk-through guides all use default service account which does not have the permission.
The default SA actually has a RoleBinding to the executor Role and both are included in the kustomization.yaml so the quick start should work with the default SA already.
kubectl apply -n argo -f https://github.com/argoproj/argo-workflows/releases/download/v3.5.2/install.yaml
I think the problem is that the Quick Start docs are referencing install.yaml instead of quick-start-minimal.yaml
@cardoe would you like to submit a PR to modify the docs to reference quick-start-minimal.yaml instead of install.yaml?
Yep will do.
I found this issue after stumbling upon the same problem across our workflows when upgrading from v3.5.0 to v3.5.4 using the released manifest files. It seems like this issue was closed after a doc change to the quick start page, but the installation docs and release pages seemingly suggest that one should be able to use install.yaml without issues:
To install Argo Workflows, navigate to the releases page and find the release you wish to use (the latest full release is preferred). Scroll down to the Controller and Server section and execute the kubectl commands.
It seems like this issue was closed after a doc change to the quick start page, but the installation docs and release pages seemingly suggest that one should be able to use install.yaml without issue
Yes, this is specific to the quick start, which explicitly states it should not be used in production.
Yes, I see that the quick start page links to the install page:
These instructions are intended to help you get started quickly. They are not suitable for production. For production installs, please refer to the installation documentation.
:point_up: refers to the release pages which then tells users to run:
kubectl create namespace argo
kubectl apply -n argo -f https://github.com/argoproj/argo-workflows/releases/download/v3.5.4/install.yaml
In any case, perhaps it's user / reader error on my part :man_shrugging: . Given quick-start-minimal.yaml is not intended for production, and install.yaml doesn't work "out of the box" anymore, is there documentation for what users that were previously using install.yaml should do when updating to v3.5.2 and beyond?
and install.yaml doesn't work "out of the box" anymore, is there documentation for what users that were previously using install.yaml should do when updating to v3.5.2 and beyond?
This wasn't listed as a regression, is there a regression in the manifests that you noticed for v3.5.2 vs an older version? You mention v3.5.0 as your prior, the only changes to the manifests between v3.5.0 and v3.5.2 were #11839 and #11947
Ah, yes, sorry, I wasn't clear about that :grimacing: . So, for better or worse, we have been using the release doc steps to deploy/update argo via install.yaml since v3.3.9 or so. It's always worked fine without issues. However, when updating from v3.5.0 to 3.5.4 workflows began failing with the message cannot patch resource "pods" in API group "". I did the same diff check between the manifests and the only differences I saw were the image tags:
- args:
- server
env: []
- image: quay.io/argoproj/argocli:v3.5.0
+ image: quay.io/argoproj/argocli:v3.5.4
name: argo-server
ports:
- containerPort: 2746
@@ -1323,7 +1323,7 @@
fieldRef:
apiVersion: v1
fieldPath: metadata.name
- image: quay.io/argoproj/workflow-controller:v3.5.0
+ image: quay.io/argoproj/workflow-controller:v3.5.4
livenessProbe:
failureThreshold: 3
httpGet:
That's why we were so surprised to hit the problem. For now we've downgraded to the https://github.com/argoproj/argo-workflows/releases/download/v3.5.1/install.yaml manifest :man_shrugging: . Any thoughts / suggestions would be very much appreciated :pray: .
Thanks, @agilgur5! I will look more into the Workflow RBAC to see what changes might be needed for our workflows to work with v3.5.4.
Just to help clarify my understanding, is the reason why we're seeing existing workflows work when downgrading just the version tag to v3.5.1 in the manifest because we've been historically lucking into some default behavior that no longer applies starting v3.5.2, and the newer versions are simply surfacing a latent RBAC issue on our end? For example, it's why the following works fine:
$ minikube start
$ kubectl create namespace argo
$ kubectl apply -n argo -f https://github.com/argoproj/argo-workflows/releases/download/v3.5.1/install.yaml
$ argo submit --watch https://raw.githubusercontent.com/argoproj/argo-workflows/main/examples/hello-world.yaml
Output:
STEP TEMPLATE PODNAME DURATION MESSAGE
✔ hello-world-45qtp whalesay hello-world-45qtp 18s
But the following will fail (as will versions 3.5.2 and 3.5.3):
$ minikube start
$ kubectl create namespace argo
$ kubectl apply -n argo -f https://github.com/argoproj/argo-workflows/releases/download/v3.5.4/install.yaml
$ argo submit --watch https://raw.githubusercontent.com/argoproj/argo-workflows/main/examples/hello-world.yaml
Output:
STEP TEMPLATE PODNAME DURATION MESSAGE
⚠ hello-world-jdsk6 whalesay hello-world-jdsk6 22s Error (exit code 1): pods "hello-world-jdsk6" is forbidden: User "system:serviceaccount:default:default" cannot patch resource "pods" in API group "" in the namespace "default"
For example, it's why the following works fine:
👀 That's an interesting reproduction.
I thought your prod cluster wasn't fresh, so could be that something else changed. If this is a fresh minikube cluster each time, then that suggests a change in the Executor is requiring that permission when it previously didn't. Or, more specifically, the simple hello-world example managed to work without any additional RBAC, but now that is required.
I'd need to look through the Executor diffs for 3.5.2 to see why that might happen. From a real quick glance, this might be from #11947 -- this change to wait.go ensure it properly reports results. If that's the case, it may be that some really short, simple steps were not recorded entirely correctly previously.
cc @Garett-MacGowan who probably knows this part of the codebase better than me at this point.
I only skimmed this, so im not sure if I'm missing something, but have you configured RBAC according to this?
I can take a better look at this tomorrow.
I only skimmed this, so im not sure if I'm missing something, but have you configured RBAC according to this?
I linked that above -- but apparently it wasn't necessary for simple steps like hello-world prior to 3.5.2. #11947 had some Executor changes that seemed like they might've changed some of that logic. It might've been a long-standing bug that didn't have much user-facing impact that you ended up fixing or maybe there's a new request that isn't necessary for simpler steps? I'm not sure, hence why I tagged you.
I can take a better look at this tomorrow.
Take your time!
@agilgur5 here's my take.
TaskResultReconciliation() which uses workflowtaskresult, for which we need RBAC, is now (necessarily) called before
if woc.wf.Labels[common.LabelKeyCompleted] == "true" { // abort now, we do not want to perform any more processing on a complete workflow because we could corrupt it
Before my PR, if the workflow completed quick enough, reconciliation wouldn't have a chance to run and therefore you would avoid the RBAC issue.
This is closely related to #11947
The quickstart docs should probably mention the need for workflowtaskresult RBAC per https://argo-workflows.readthedocs.io/en/latest/workflow-rbac/
Before my PR, if the workflow completed quick enough, reconciliation wouldn't have a chance to run and therefore you would avoid the RBAC issue.
Thought that could be the case per above, thanks for confirming Garett!
The quickstart docs should probably mention the need for workflowtaskresult RBAC per https://argo-workflows.readthedocs.io/en/latest/workflow-rbac/
The PR that closed this issue, #12445 fixed the quickstart docs to actually use the quick-start-*.yaml at my suggestion above, which has the appropriate permissions.
But the Workflow RBAC page should be linked in the production installation page where it is apparently missing.
@rnathuji would you like to submit a PR to add a link to the Workflow RBAC page to the production installation page? Perhaps it could be made as the first bullet under "Additional installation considerations"
@agilgur5 ahh, yes, you're right. The real reason is here. In this PR, we switched to always creating a task result (a placeholder, if no outputs exist) in order to track reconciliation status.
The implementation has changed slightly since this PR, but it's effectively doing the same thing.
Thanks for the context! Thought it might be something like that.
Re-wording, that would mean that previously, as hello-world had no outputs, it would made no task result and hence technically did not need the RBAC. That led to some race conditions however, so now it does require that. The docs have always stated this, so it was actually an undocumented technicality that simpler steps like hello-world didn't need the RBAC in the past.
Thanks for the context! Thought it might be something like that.
Re-wording, that would mean that previously, as hello-world had no outputs, it would made no task result and hence technically did not need the RBAC. That led to some race conditions however, so now it does create a task result and therefore require the RBAC. The docs have always stated this, so it was actually an undocumented technicality that simpler steps like hello-world didn't need the RBAC in the past.
That's right.
@agilgur5 - Sure, I can spin up a doc PR.
Thanks all for taking a look and helping to clarify the reason behind the behavioral change that we observed. :tada:
See PR in #12680
| gharchive/issue | 2023-12-20T16:41:12 | 2025-04-01T04:33:32.210376 | {
"authors": [
"Garett-MacGowan",
"agilgur5",
"cardoe",
"rnathuji",
"terrytangyuan"
],
"repo": "argoproj/argo-workflows",
"url": "https://github.com/argoproj/argo-workflows/issues/12391",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1127810528 | Support archived workflows logs in UI from archived logs
Summary
Currently, if you open an archived workflow, you are not able to see logs in the log viewer, even if the logs were archived to 'main.log'. We could fix this.
Use Cases
When would you use this?
Message from the maintainers:
Love this enhancement proposal? Give it a 👍. We prioritise the proposals with the most 👍.
Hi @alexec , would you mind providing more details on this.
I couldn't get my head around what you meant by "Support archived workflows logs in UI from archived logs".
All I'm thinking right now is we can view archived log from archived workflows
If I understand correctly - the current UI does not make the request to get the logs (output artifacts) by the workflow's UID? because the server's endpoint for it exist.
Does this relate to the Log UI "All" option not showing nothing after pods are garbage collected? Or, should that be considered a separate issue?
On a side note, what does this recommendation in the docs mean?:
⚠️ We do not recommend you rely on Argo Workflows to archive logs. Instead, use a conventional Kubernetes logging facility.
Are there some known issues? Also, does kubernetes provide any facility for collecting and returning logs for garbage collected pods, and then can argo workflows use this facility? I ask because viewing logs in the argo workflows UI is very convenient, and one of the main "selling points" of the application in my mind
Also, does kubernetes provide any facility for collecting and returning logs for garbage collected pods, and then can argo workflows use this facility?
A pretty conventional cluster set-up is to have fluentd / fluent-bit as a DaemonSet collecting logs from all nodes and sending those a log provider (e.g. Splunk, Logstash, etc).
That would be an interesting feature if Workflows could use it, as it is indeed convenient. As a workaround to that, you can add an external link to a query in your log provider.
Not sure if #11121 ended up fixing this as well
It's supported. Re-open if not.
| gharchive/issue | 2022-02-08T22:09:16 | 2025-04-01T04:33:32.217927 | {
"authors": [
"agilgur5",
"alexec",
"asfaltboy",
"noam-codefresh",
"tczhao",
"terrytangyuan"
],
"repo": "argoproj/argo-workflows",
"url": "https://github.com/argoproj/argo-workflows/issues/7830",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1139938418 | Request help, can realize drawing the way to generate workflow?
Can Argo UI add workflow by drawing?
The flow chart drawn by Activiti is imported into Argo to implement the workflow.
This issue template is only for mentoring. Not feature request.
Also this is a duplicate of https://github.com/argoproj/argo-workflows/issues/5610 so I am closing this.
Asking for help, can you now implement the way drawing generates workflow?
This issue template is only for mentoring. Not feature request.
Also this is a duplicate of #5610 so I am closing this.
Asking for help, can you now implement the way drawing generates workflow?
No that's not supported which is why the linked issue is still open.
No that's not supported which is why the linked issue is still open.
Is there any requirement that can be realized:
Flowchart drawing tools such as Activiti and Camunda can be used to design a flowchart and then be connected to or imported into Argo Workflow for execution
| gharchive/issue | 2022-02-16T12:17:13 | 2025-04-01T04:33:32.221760 | {
"authors": [
"LingClassmate",
"LingHuaHui",
"terrytangyuan"
],
"repo": "argoproj/argo-workflows",
"url": "https://github.com/argoproj/argo-workflows/issues/7892",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1174727619 | I think my Artifact repository is being interpreted as a volume(?)
Checklist
[ ] Double-checked my configuration.
[ ] Tested using the latest version.
[ ] Used the Emissary executor.
Summary
This is a follow-on to this issue (https://github.com/argoproj/argo-workflows/issues/8168).
I have implemented my artifact repository per these instructions (https://argoproj.github.io/argo-workflows/configure-artifact-repository/). Here is my abstracted implementation
artifactRepository: |
gcs:
bucket: my-bucket
region: us-east-1
keyFormat: /tmp/
serviceAccountKeySecret:
name: account@developer.gserviceaccount.com
key: my-64bit encoded-key
I expect that the following workflow should save the output of whalesay to the repository so that it is available for the print pod.
---
{
"apiVersion": "argoproj.io/v1alpha1",
"kind": "Workflow",
"metadata": {
"annotations": {
"workflows.argoproj.io/description": "testing a linter",
"workflows.argoproj.io/version": ">= 3.1.0"
},
"labels": {
"workflows.argoproj.io/archive-strategy": "false"
},
"name": "hello-world",
},
"spec": {
"serviceaccount": "725104128071-compute@developer.gserviceaccount.com ",
"arguments": {
"parameters": [
{
"name": "msg",
"value": "Hello there"
}
]
},
"entrypoint": "entrypoint",
"securityContext": {
"fsGroup": 2000,
"fsGroupChangePolicy": "OnRootMismatch",
"runAsGroup": 3000,
"runAsNonRoot": true,
"runAsUser": 1000
},
"templates": [
{
"container": {
"args": [
"cowsay Hello Test >> /tmp/output"
],
"command": [
"sh",
"-c"
],
"image": "docker/whalesay:latest",
"imagePullPolicy": "IfNotPresent",
},
"name": "whalesay",
"outputs": {
"artifacts": [
{
"name": "message",
"path": "/tmp/output",
"s3": {
"key": "whalesay",
}
}
]
},
"retryStrategy": {
"limit": "10"
},
"securityContext": {
"fsGroup": 2000,
"fsGroupChangePolicy": "OnRootMismatch",
"runAsGroup": 3000,
"runAsNonRoot": true,
"runAsUser": 1000
}
},
{
"inputs": {
"artifacts": [
{
"s3": {
"key": "whalesay",
},
"name": "data",
"path": "/tmp/input"
}
]
},
"name": "print",
"retryStrategy": {
"limit": "10"
},
"script": {
"command": [
"python"
],
"image": "python:alpine3.6",
"imagePullPolicy": "IfNotPresent",
"source": "import sys \nsys.stdout.write(\"{{inputs.artifacts.data}}\")\n\n"
},
"securityContext": {
"fsGroup": 2000,
"fsGroupChangePolicy": "OnRootMismatch",
"runAsGroup": 3000,
"runAsNonRoot": true,
"runAsUser": 1000
}
},
{
"dag": {
"tasks": [
{
"name": "whalesay",
"template": "whalesay"
},
{
"arguments": {
"artifacts": [
{
"from": "{{whalesay.outputs.artifacts.message}}",
"name": "data"
}
]
},
"dependencies": [
"whalesay"
],
"name": "print",
"template": "print"
}
]
},
"name": "entrypoint"
}
]
}
}
...
Instead, I am getting an error which suggests it is expecting a volumeClaim
Pod "hello-world2xr6z-2749602831" is invalid: [spec.volumes[1].name: Invalid value: "725104128071-compute@developer.gserviceaccount.com": a lowercase RFC 1123 label must consist of lower case alphanumeric characters or '-', and must start and end with an alphanumeric character (e.g. 'my-name', or '123-abc', regexused for validation is '[a-z0-9]([-a-z0-9]*[a-z0-9])?'), spec.containers[0].volumeMounts[0].name: Not found: "725104128071-compute@developer.gserviceaccount.com", spec.initContainers[0].volumeMounts[0].name: Not found: "725104128071-compute@developer.gserviceaccount.com"]
What version are you running?
Argo 3.3.0
I am able to successfully run the quickstart example, so I think it is the artifact repository implementation. i.e. Argo itself is working ok.
No pods are generated in the case of the problematic workflow (above)
Message from the maintainers:
Impacted by this bug? Give it a 👍. We prioritise the issues with the most 👍.
Hypothesis: you have executor set in your config map. Can you please confirm?
Yes.
containerRuntimeExecutor: emissary
sorry. I meant the specifically data.executor.
Here you go!
apiVersion: v1
kind: ConfigMap
metadata:
name: workflow-controller-configmap
data:
artifactRepository: |
gcs:
endpoint: storage.googleapis.com
bucket: my-bucket
region: us-east-1
serviceAccountKeySecret:
name: ''
key: ''
useSDKCreds: false
containerRuntimeExecutor: emissary
workflowDefaults: |
spec:
ttlStrategy:
secondsAfterSuccess: 5
parallelism: 3
entrypoint: 'entrypoint'
securityContext:
runAsNonRoot: true
runAsUser: 1000
fsGroup: 2000
runAsGroup: 3000
fsGroupChangePolicy: "OnRootMismatch"
You're configuration is incorrect. It is impossible for there to be a secret with an @ in the name. Fix this
serviceAccountKeySecret:
name: name-of-secret
key: key-within-secret
| gharchive/issue | 2022-03-20T23:31:43 | 2025-04-01T04:33:32.230989 | {
"authors": [
"alexec",
"wdma"
],
"repo": "argoproj/argo-workflows",
"url": "https://github.com/argoproj/argo-workflows/issues/8197",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1219428303 | Full CRDs are incomplete and causes json2jsii to fail
Summary
The generated CustomResourceDefinitions seem to be incomplete because of the steps property.
https://github.com/argoproj/argo-workflows/blob/ecd91b1c4215a2ab8742f7c43eaade98a1d47eba/manifests/base/crds/full/argoproj.io_workflowtemplates.yaml#L7673-L7676
The items property has a type property but does not have an accompanying items property.
This causes json2jsii to fail.
https://github.com/cdklabs/json2jsii/blob/60e891447e35a8bb1a9dac3496d6adce87d73292/src/type-generator.ts#L534-L540
This is an issue for tools that depend on json2jsii. https://github.com/cdk8s-team/cdk8s-cli/issues/93
However, the json schema has these missing properties.
https://github.com/argoproj/argo-workflows/blob/ecd91b1c4215a2ab8742f7c43eaade98a1d47eba/api/jsonschema/schema.json#L7315-L7321
https://github.com/argoproj/argo-workflows/blob/ecd91b1c4215a2ab8742f7c43eaade98a1d47eba/api/jsonschema/schema.json#L6449-L6454
It looks like this problem was fixed at one point with #4828, but was later reverted by @JPZ13 here #4810 because of the CRDs becoming too long.
Is there way we can get the full CRDs for use with json2jsii as I do not believe it has any limit on CRD size like Kubernetes does. I'd like to have all the types for steps so that I can get strong typing with cdk8s.
Message from the maintainers:
Impacted by this bug? Give it a 👍. We prioritise the issues with the most 👍.
Would you like to submit a PR to fix this?
I would need some guidance, I spent some time looking into the issue yesterday on a fork https://github.com/adrianord/argo-workflows, but I believe the fix I've done there causes the CRDs to be too large for Kubernetes, which is why it was reverted before.
Share you PR and maybe we can take a look at it.
Also getting an issue here with crds and json2jsii.
Error: unsupported array type undefined
at TypeGenerator.typeForArray (/snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:404:19)
at TypeGenerator.emitArray (/snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:260:25)
at TypeGenerator.emitTypeInternal (/snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:186:29)
at TypeGenerator.typeForProperty (/snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:383:21)
at TypeGenerator.typeForArray (/snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:406:21)
at TypeGenerator.emitArray (/snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:260:25)
at TypeGenerator.emitTypeInternal (/snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:186:29)
at TypeGenerator.typeForProperty (/snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:383:21)
at TypeGenerator.emitProperty (/snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:322:35)
at /snapshot/node_modules/cdk8s-cli/node_modules/json2jsii/lib/type-generator.js:302:22
| gharchive/issue | 2022-04-28T23:10:45 | 2025-04-01T04:33:32.238327 | {
"authors": [
"adrianord",
"alexec",
"sstaley-hioscar"
],
"repo": "argoproj/argo-workflows",
"url": "https://github.com/argoproj/argo-workflows/issues/8532",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
673147256 | Surface K8s Events about Nodes to UI
We run argo on GKE with cluster-autoscaler. Most of the time all our workflows trigger a scaleUp event since powerful machines are usually down when not used.
Usually, the scale-up takes a few minutes. During this time the workflow is in pending state with a message similar to this:
Once the node has been created the workflow starts and all is good!
But sometimes the cluster is not able to scale up (often due to an error in the resources and nodeSelector configuration).
Admin have access to cluster event logs and we can spot the issue quickly:
But most of our staff only use the Argo dashboard and so can only rely on the workflow status message to understand what is going on. And scaling up error reported by the cluster-autoscaler is not propagated to it:
pod didn't trigger scale-up (it wouldn't fit if a new node is added): 16 Insufficient memory, 1 node(s) didn't match node selector, 15 Insufficient cpu
Would that be possible to report cluster events related to a specific workflow or pods managed by a workflow in the status field?
Not sure it's technically possible (because of ServiceAccount and permissions) but I am asking just in case.
We watch pods of a workflow to compute state. If this information is available on the pod’s status we could.
Are you able to attach the YAML of a pod that we involved? Or is this only available on events? (Please attach the event YAML).
Indeed the only thing I see in the pod YAML is what Argo report in the UI:
status:
conditions:
- lastProbeTime: null
lastTransitionTime: "2020-08-05T16:53:27Z"
message: '0/21 nodes are available: 21 Insufficient cpu.'
reason: Unschedulable
status: "False"
type: PodScheduled
phase: Pending
qosClass: Burstable
and the source is default-scheduler for the message. But the message "pod didn't trigger scale-up (it wouldn't fit if a new node is added): 17 Insufficient cpu" coming from the source cluster-autoscaler is not shown.
$ kubectl get event -n argo --field-selector involvedObject.name=wonderful-tiger
LAST SEEN TYPE REASON OBJECT MESSAGE
25s Normal WorkflowRunning workflow/wonderful-tiger Workflow Running
25s Warning FailedScheduling pod/wonderful-tiger 0/21 nodes are available: 21 Insufficient cpu.
22s Normal NotTriggerScaleUp pod/wonderful-tiger pod didn't trigger scale-up (it wouldn't fit if a new node is added): 17 Insufficient cpu
and
kubectl get event -n argo --field-selector involvedObject.name=wonderful-tiger -o yaml
apiVersion: v1
items:
- apiVersion: v1
count: 1
eventTime: null
firstTimestamp: "2020-08-05T17:01:09Z"
involvedObject:
apiVersion: argoproj.io/v1alpha1
kind: Workflow
name: wonderful-tiger
namespace: argo
resourceVersion: "14645697"
uid: f4bb68ae-84a2-4e40-8d5b-71f0a9ab91f3
kind: Event
lastTimestamp: "2020-08-05T17:01:09Z"
message: Workflow Running
metadata:
creationTimestamp: "2020-08-05T17:01:09Z"
name: wonderful-tiger.16286dde6befd70c
namespace: argo
resourceVersion: "135777"
selfLink: /api/v1/namespaces/argo/events/wonderful-tiger.16286dde6befd70c
uid: 57c2bd09-11c6-4dc8-8ada-b33346764566
reason: WorkflowRunning
reportingComponent: ""
reportingInstance: ""
source:
component: workflow-controller
type: Normal
- apiVersion: v1
count: 3
eventTime: null
firstTimestamp: "2020-08-05T17:01:09Z"
involvedObject:
apiVersion: v1
kind: Pod
name: wonderful-tiger
namespace: argo
resourceVersion: "14645699"
uid: a6c471b1-a31e-46fa-b53e-27554a81d328
kind: Event
lastTimestamp: "2020-08-05T17:02:33Z"
message: '0/21 nodes are available: 21 Insufficient cpu.'
metadata:
creationTimestamp: "2020-08-05T17:01:09Z"
name: wonderful-tiger.16286dde6d37eca9
namespace: argo
resourceVersion: "135781"
selfLink: /api/v1/namespaces/argo/events/wonderful-tiger.16286dde6d37eca9
uid: d256c9fe-40fd-4b97-a7e6-62ec0470eb4f
reason: FailedScheduling
reportingComponent: ""
reportingInstance: ""
source:
component: default-scheduler
type: Warning
- apiVersion: v1
count: 1
eventTime: null
firstTimestamp: "2020-08-05T17:01:12Z"
involvedObject:
apiVersion: v1
kind: Pod
name: wonderful-tiger
namespace: argo
resourceVersion: "14645700"
uid: a6c471b1-a31e-46fa-b53e-27554a81d328
kind: Event
lastTimestamp: "2020-08-05T17:01:12Z"
message: 'pod didn''t trigger scale-up (it wouldn''t fit if a new node is added):
17 Insufficient cpu'
metadata:
creationTimestamp: "2020-08-05T17:01:12Z"
name: wonderful-tiger.16286ddf00db261b
namespace: argo
resourceVersion: "135780"
selfLink: /api/v1/namespaces/argo/events/wonderful-tiger.16286ddf00db261b
uid: ac79183e-aaff-4f8d-a604-10a8c4b254f6
reason: NotTriggerScaleUp
reportingComponent: ""
reportingInstance: ""
source:
component: cluster-autoscaler
type: Normal
kind: List
metadata:
resourceVersion: ""
selfLink: ""
So maybe it's possible that the cluster-autoscaler event was too fast to be displayed? If that's the case then is there is to access the previous status messages of a workflow?
Would it be enough to list events related to the pod in the UI?
I think it would do the job yeah. And actually this will be useful for probably other things than this specific issue to be able to see the history of events. Similar to what's GKE is doing:
I guess if this is available in the UI it will also be available in the status field of a Workflow YAML spec. Then we could also use this information from a machine for automatic processing. So yeah that would a very nice feature!
As an MVP, I don't think we would make it available in the YAML. Instead, we would just make it available in the UI only.
That would exclude the UI and the CLI.
Do you think this should be MVP?
For us, the most important is to have this information available on the UI.
Having it in the YAML is just a bonus and a cool enhancement IMO. I like the idea of having all the information related to a workflow in a single YAML object. Its state, spec, and history of status.
So my answer is yes it should go in MVP but if this requires too much work then having the information on the UI only is fine.
Available for testing in v2.10.0-rc1.
| gharchive/issue | 2020-08-04T23:32:44 | 2025-04-01T04:33:32.248791 | {
"authors": [
"alexec",
"hadim"
],
"repo": "argoproj/argo",
"url": "https://github.com/argoproj/argo/issues/3673",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1447112945 | References not correctly added to the model README
There's functionality to auto add the values from the datasets "reference" field to the .argosmodel package READMEs. There's a bug and it's been broken recently and I haven't been able to figure out why.
I've experimented with trying to fix this but something is still broken.
https://github.com/argosopentech/argos-train/commit/cc2d7752dcb527e6a35a09c249acacb422231cef
| gharchive/issue | 2022-11-13T22:18:13 | 2025-04-01T04:33:32.250777 | {
"authors": [
"PJ-Finlay",
"argosopentech"
],
"repo": "argosopentech/argos-train",
"url": "https://github.com/argosopentech/argos-train/issues/25",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1202890802 | Fix typo in German translation
The German word for precision, "Genauigkeit", was spelled wrong.
Also, it seems that Github automatically converted the file to UTF-8.
@jstuyts is this repo still being maintained?
| gharchive/pull-request | 2022-04-13T07:46:37 | 2025-04-01T04:33:32.252042 | {
"authors": [
"tsaglam"
],
"repo": "argparse4j/argparse4j",
"url": "https://github.com/argparse4j/argparse4j/pull/139",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2122415999 | Failed to launch without user gesture.
This is a repo reproducing this question: https://github.com/daGaiGuanYu/show-me-bugs/tree/nostalgist240207
warn message in console:
The AudioContext was not allowed to start. It must be resumed (or created) after a user gesture on the page.
I don't know whether this can be fixed.
If not, it would be nice to mention it in the documentation.
Developer could show a dialog to tell users "rom download, start now?".
And user's clicking will offer a "user gesture", and this limit are gone.
I'm making a little APP (not completed):
load nes games from local and the internet.
Installable, and offline playing(not completed), and launching by clicking a nes file.
Make a user gesture before launch:
// ...
await make_usergesture()
Nostalgist.launch(...)
function make_usergesture () {
const div = document.createElement('div')
div.style.position = 'fixed'
div.style.left = '0'
div.style.top = '0'
div.style.right = '0'
div.style.bottom = '0'
div.style.zIndex = '1000'
div.style.background = '#333'
div.style.color = '#fff'
div.style.display = 'grid'
div.style.placeItems = 'center'
div.innerHTML = 'Please click on the screen to continue'
return new Promise<void>(res => {
div.addEventListener('click', () => {
document.body.removeChild(div)
res()
})
document.body.appendChild(div)
})
}
It's a limitation from the browser (see https://developer.chrome.com/blog/autoplay) so we cannot bypass it. We have to ensure the launch method is invoked after a user click event or a keyboard event.
In fact, there is an undocumented parameter waitForInteraction working like this:
await Nostalgist.nes({
rom: 'flappybird.nes',
waitForInteraction({ done }) {
const button = document.createElement('button')
button.textContent = 'click to start'
document.body.append(button)
button.addEventListener('click', () => {
done()
button.remove()
})
},
})
It still haven't been added to the documentation now, since it's a bit complex to explain, and I also think the API design may have room to improve to achieve better clarity.
I think your approach for making a user gesture before launching is quite straightforward and you could just keep using it.
| gharchive/issue | 2024-02-07T08:01:07 | 2025-04-01T04:33:32.261500 | {
"authors": [
"arianrhodsandlot",
"daGaiGuanYu"
],
"repo": "arianrhodsandlot/nostalgist",
"url": "https://github.com/arianrhodsandlot/nostalgist/issues/18",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
305175583 | build sample/initial pattern funtcion
check for sample that are uploaded
if not , take initial sample in real time
done
| gharchive/issue | 2018-03-14T14:13:43 | 2025-04-01T04:33:32.263893 | {
"authors": [
"arik-le",
"gaiMaimoni"
],
"repo": "arik-le/Chips-Bits",
"url": "https://github.com/arik-le/Chips-Bits/issues/14",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1763252159 | Feat(eos_cli_config_gen): add global logging event storm-control
Enhancement summary
global logging event for storm-control is the only one being disabled by default in eos.
Add this global configuration option to eos_cli_config_gen to be able to enable it.
access#show run all | i logging event
no logging event storm-control discards global
no logging event storm-control discards interval
default logging event congestion-drops
logging event port-channel member-status global
no logging event login root
logging event link-status global
logging event spanning-tree global
logging event flowcontrol received syslog interval 10 seconds
access(config)#logging event ?
congestion-drops Drops due to congestion
flowcontrol Configure syslogging for flowcontrol
link-status UPDOWN messages
login Login activities
port-channel Configure port-channel messages
spanning-tree Spanning tree messages
storm-control Configure storm-control
access(config)#logging event storm-control ?
discards Discards due to storm control
access(config)#logging event storm-control discards ?
global Configure global storm control discard logging
interval Logging interval
Which component of AVD is impacted
eos_cli_config_gen
Use case example
logging:
event:
- storm_control_discards:
enable: true
interval: 10
logging event storm-control discards global
logging event storm-control discards interval 10
Describe the solution you would like
logging:
event:
- storm_control_discards:
enable: <bool>
interval: <int>
should render the following cli:
logging event storm-control discards global
logging event storm-control discards interval <int>
Describe alternatives you have considered
use of eos_cli
Additional context
No response
Contributing Guide
[X] I agree to follow this project's Code of Conduct
I think you have a typo in your data model
logging:
event:
storm_control_discards:
enable: <bool>
interval: <int>
Thanks
| gharchive/issue | 2023-06-19T10:36:03 | 2025-04-01T04:33:32.269322 | {
"authors": [
"ClausHolbechArista",
"kmueller68"
],
"repo": "aristanetworks/ansible-avd",
"url": "https://github.com/aristanetworks/ansible-avd/issues/2993",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
820194450 | Validate role fails when NTP is not configured
Issue Type
[x] Bug Report
[ ] EOS Configuration rendering error
[x] Documentation Issue Report
Summary
While executing eos validation role on a device without ntp configured, ansible gets an unexpected value type and this breaks the MD documentation file (which is broken and not possible to read). CSV can still be properly interpreted.
Component Name
eos_validate_state role
Steps to reproduce
Execute EOS validation role on a device without NTP configuration.
Expected results
$
Actual results
TASK [eos_validate_state : Validate ntp status] *************************************************************************************************************************************
fatal: [DC1-SUPER-SPINE1]: FAILED! => {"msg": "The conditional check 'ntp_status.stdout[0] | regex_search(\"synchronised to NTP server\")' failed. The error was: Unexpected templating type error occurred on ({% if ntp_status.stdout[0] | regex_search(\"synchronised to NTP server\") %} True {% else %} False {% endif %}): expected string or bytes-like object"}
...ignoring
fatal: [DC1-SUPER-SPINE2]: FAILED! => {"msg": "The conditional check 'ntp_status.stdout[0] | regex_search(\"synchronised to NTP server\")' failed. The error was: Unexpected templating type error occurred on ({% if ntp_status.stdout[0] | regex_search(\"synchronised to NTP server\") %} True {% else %} False {% endif %}): expected string or bytes-like object"}
...ignoring
Fixed by #763
| gharchive/issue | 2021-03-02T16:46:19 | 2025-04-01T04:33:32.273095 | {
"authors": [
"ClausHolbechArista",
"xaviramon"
],
"repo": "aristanetworks/ansible-avd",
"url": "https://github.com/aristanetworks/ansible-avd/issues/761",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
201134262 | Splat v.2.0.0 compatibility
as in the title
Please read the contrib guidelines before filing in the future (the ones you deleted to post your message)
| gharchive/issue | 2017-01-16T22:44:05 | 2025-04-01T04:33:32.293836 | {
"authors": [
"aritchie",
"belandsoft"
],
"repo": "aritchie/userdialogs",
"url": "https://github.com/aritchie/userdialogs/issues/317",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
836894744 | Tag sessions with the capturing interface
Is your feature request related to a problem? Please describe.
No problem, but it would be useful if we could easily filter on the capturing interface.
Describe the solution you'd like
I don't know if this would be best as just a tag, as a discrete column in the UI, or modifying the node name.
Describe alternatives you've considered
N/A
Additional context
In my experimentation, I was able to make this work by creating an instance of each of the following files for each interface I was monitoring:
/data/moloch/etc/config.ini
/data/moloch/bin/moloch_config_interfaces.sh
/etc/systemd/system/molochcapture.service
The specific config.ini and moloch_config_interfaces.sh each was changed to only reference an individual interface.
I also made discrete pcap directories for each interface and specified those in the config.ini files.
The molochcapture.service files I created were specified with -n ORIGINALNODENAE-INTERFACENAME, -t INTERFACENAME, a unique log file name, and finally the unique config.ini file for ExecStart, and the unique moloch_config_interfaces.sh for ExecStartPre. I specified a unique node name, otherwise I would get errors from elasticsearch referring to the stats index, specifically looking like this:
{"error":{"root_cause":[{"type":"version_conflict_engine_exception","reason":"[molochVM]: version conflict, current version [22070] is higher or equal to the one provided [22070]","index_uuid":"NMAYhwULQvahWDcaYKhl1A","shard":"0","index":"stats_v4"}],"type":"version_conflict_engine_exception","reason":"[molochVM]: version conflict, current version [22070] is higher or equal to the one provided [22070]","index_uuid":"NMAYhwULQvahWDcaYKhl1A","shard":"0","index":"stats_v4"},"status":409}
I'd guess this is due to having two moloch_capture processes running on the same VM at the same time.
I added the tag since I wasn't sure which would work best, so I figured I'd try with a tag as well.
Generally that seems to be working and I'm seeing both the unique node name, which includes the interface name, as well as the interface name as a tag
I'm now realizing this is probably what the [node1] and [node2] sections are at the bottom of the config.ini file...
This is already mostly possible if you are ok using the tags field:
If using 1 interface per capture just either use the -t interfacename option or the --op tags=interfacename to capture
If using multple interfaces per capture then look at the https://arkime.com/settings#interfaceops setting
If you wanted a new field called "interface" that would have to be added using https://arkime.com/settings#custom-fields, and then you would use --op interface=interfacename or again the interfaceops setting.
I'd guess this is due to having two moloch_capture processes running on the same VM at the same time.
yes node name should be unique. Instead of running multiple, you should just have capture listen to 2 interfaces
I'm now realizing this is probably what the [node1] and [node2] sections are at the bottom of the config.ini file...
not sure what you mean, but order in the file shouldn't matter. Just make sure settings you are adding actually end up in the right section.
It seems i missed https://arkime.com/faq#how-do-i-monitor-multiple-interfaces
Most of the info there is what I ended up doing, but I did differ by adding the tag and having two config files.
One question on the existing instructions: When the two different systemd services start, each one would run the moloch_config_interfaces.sh script. If the single script makes changes to all of the interfaces, wouldn't that potentially cause issues with other interfaces that are already capturing? For example if you're capturing on eth1 and eth2, and restart the eth2 process, since both eth1 and eth2 are referenced in the common moloch_config_interfaces.sh, would that impact eth1?
not sure what you mean, but order in the file shouldn't matter. Just make sure settings you are adding actually end up in the right section.
To clear up what I meant: I was mixing up a few things, and didn't realize the use of creating unique sections in the config.ini file.
Hey, I'm having a bit of an issue understanding how to specify interfaceops in the config.ini file. Anything you can give me to point me in the right direction would be appreciated!
Updated the FAQ, otherwise come on over to slack for questions
| gharchive/issue | 2021-03-20T19:36:42 | 2025-04-01T04:33:32.309022 | {
"authors": [
"awick",
"copeland3300"
],
"repo": "arkime/arkime",
"url": "https://github.com/arkime/arkime/issues/1638",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
253946981 | ClickOutsideModule not working inside an anchor
I import ClickOutsideModule in the following module:
@NgModule({
imports: [ClickOutsideModule],
declarations: [TopBarComponent, MyComponent]
})
export class TopBarModule { }
MyComponent is as follows:
@Component({
moduleId: module.id,
selector: 'my-component',
styleUrls: ['my-component.css'],
templateUrl: 'my-component.component.html'
})
export class MyComponent {
public onClickedOutside(event: Event) {
console.log('Clicked outside:', event);
}
}
with the template being:
< ul (clickOutside)="onClickedOutside($event)">Hello</ul >
Nothing is printed on the console when I click outside the unsorted list.
Works fine for me. Only notable thing is that you can't have spaces in your tags (although that should've given you an error):
<ul (clickOutside)="onClickedOutside($event)">Hello</ul>
| gharchive/issue | 2017-08-30T10:12:38 | 2025-04-01T04:33:32.313925 | {
"authors": [
"arkon",
"jacopolanzoni"
],
"repo": "arkon/ng-click-outside",
"url": "https://github.com/arkon/ng-click-outside/issues/15",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1030826652 | Split permutation from sponge construction
Work towards #29; this doesn't touch the constraint system implementation yet,
in order to be able to get design feedback on the software part.
The poseidon::PoseidonParameters struct is renamed to poseidon::Parameters
but otherwise remains unchanged.
The poseidon::PoseidonSpongeState struct is renamed to poseidon::State
and redefined to hold just the state itself, as well as the parameters needed
to run the permutation. It exposes a permute(&mut self) method, rate() and
capacity() accessors, as well as Index, IndexMut, AsRef, and AsMut
impls that allow access to the state.
The poseidon::PoseidonSponge struct is renamed to poseidon::Sponge and
holds a State and a DuplexSpongeMode. In other words, it consists of the
state, together with the extra data tracking how that state is being used to
implement a higher-level duplex construction.
The CryptographicSponge trait is changed so that new() takes an owned,
Self::Parameters, not a borrowed one. This allows the caller to decide
where to copy data, instead of forcing the sponge implementation to clone
internally. Or, a CryptographicSponge implementation could declare the
associated Parameters type to be some shared type (like an Arc wrapper)
that avoids the need to copy at all.
The SpongeExt trait that allows converting back and forth between a state
and a sponge is deleted; it's not safe to pass between abstraction layers
that way.
Before we can merge this PR, please make sure that all the following items have been
checked off. If any of the checklist items are not applicable, please leave them but
write a little note why.
[x] Targeted PR against correct branch (master)
[x] Linked to Github issue with discussion and accepted design OR have an explanation in the PR that describes this work.
[ ] Wrote unit tests
[x] Updated relevant documentation in the code
[ ] Added a relevant changelog entry to the Pending section in CHANGELOG.md
[ ] Re-reviewed Files changed in the Github PR explorer
(A rendered copy of the docs is available here: https://rustdoc.penumbra.zone/main/ark_sponge/poseidon/index.html)
Hey, just bumping this -- if this approach seems good I can also apply it to the constraint implementations.
Feel free to apply to the constraints implementations.
| gharchive/pull-request | 2021-10-19T23:31:26 | 2025-04-01T04:33:32.324364 | {
"authors": [
"hdevalence",
"weikengchen"
],
"repo": "arkworks-rs/sponge",
"url": "https://github.com/arkworks-rs/sponge/pull/30",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1255088981 | Arlon cli context management support
Currently arlon cli interacted with 3 entities:
Git repo
ArgoCD Server
K8s API Server of Management Cluster
It would be very beneficial to have a context management similar to how its done by argocd cli.
A context can be defined as combination of git repo (with credentials if there), ArgoCD Server connection information, k8s API Server connection information.
All the information about multiple contexts can be saved in ~/.arlon folder on user machine.
User should be able to perform following activities with the contexts:
Create new contexts
Delete contexts
Shift between contexts
Renew tokens in the context (good to have)
Once a context is set by the user, the individual commands having parameters about above 3 (git, k8s, argocd) can be treated as optional. If present, they override the set context.
Commands can be see as follows:
Create new context
arlon context add <new_context_name> --argocd-server <argocd_addr> --gitrepo .....
Delete context
arlon context delete <context_name>
Shift between context
arlon context set <context_name>
Impact of other commands because of context. Below given is an example command:
arlon create clusterspec [--context <context_to_use>] [--gitrepo <repo url>]..
Notice in above command that git repo and context are optional parameters. Individual command to retrieve the git repo in following order:
a. overridden param in command
b. context present in the command
c. command set globally
Similarly do the same for argocd server & mgmt cluster also.
@bcle who can scope this work?
Can you label this with a t-shirt size?
| gharchive/issue | 2022-06-01T06:19:43 | 2025-04-01T04:33:32.362808 | {
"authors": [
"bcle",
"chinmaykolhatkar",
"cre8minus1"
],
"repo": "arlonproj/arlon",
"url": "https://github.com/arlonproj/arlon/issues/52",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1619638900 | Service Port Alignment in Operator objects
We need to grab http, grpc, and metrics ports from the App Config of each service, and use it for the Deployment, Service, and Ingress configs created by the operator.
https://github.com/armadaproject/armada-operator/issues/96 duplicate
| gharchive/issue | 2023-03-10T21:21:23 | 2025-04-01T04:33:32.364879 | {
"authors": [
"kannon92",
"suprjinx"
],
"repo": "armadaproject/armada",
"url": "https://github.com/armadaproject/armada/issues/2256",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1710749078 | How to setup on Lazy?
I'm trying with:
{
"arnamak/stay-centered.nvim",
lazy = false,
},
And it is listed as "loded" but it does nothing.
Nevermind, just created an autocmd
local augroup = vim.api.nvim_create_augroup
local autocmd = vim.api.nvim_create_autocmd
-- Keep cusor always centered
local cursor_group = augroup("cursor", { clear = true })
autocmd({ "CursorMoved"}, {
desc = "Keep cursor always centered",
group = cursor_group,
callback = function()
vim.api.nvim_exec("norm zz", false)
baseevent "CursorCentered"
end,
})
SOLVED:
-- stay-centered.nvim [cursor centered]
-- https://github.com/arnamak/stay-centered.nvim
{
"arnamak/stay-centered.nvim",
lazy=false,
opts = {
skip_filetypes = { }
}
},
| gharchive/issue | 2023-05-15T20:06:17 | 2025-04-01T04:33:32.417984 | {
"authors": [
"Zeioth"
],
"repo": "arnamak/stay-centered.nvim",
"url": "https://github.com/arnamak/stay-centered.nvim/issues/8",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1453013459 | [FEATURE] Update to choco 4.10.10
Is your feature request related to a problem? Please describe.
Update choco to benefit from the latest enhancements and to seek support more easily.
Describe the solution you'd like
Update to the latest choco release.
Additional context
A first attempt has shown that the expression handling has changed : some tests will fail.
We must determine if the tests fails because of a regression.
It requires to isolate the failing tests, hard-code the models and post issues in the choco project if needed.
JUnit tests that shows a case where choco found results in 4.10.6 but not in 4.10.10
import org.junit.Before;
import org.junit.Test;
import org.chocosolver.solver.Model;
import org.chocosolver.solver.Solver;
import org.chocosolver.solver.variables.IntVar;
public class LittleTest {
public static Model model;
public static IntVar copper, neon, iron, silver, result;
@Before
/***
* Tests if choco can solve "copper*neon=iron*silver" via various way of writing constraints
*/
public void setup(){
model = new Model("copper*neon=iron*silver");
//model.getSettings().getWelcomeMessage();
copper = model.intVar("copper",100000,999999);
neon = model.intVar("neon",1000,9999);
iron = model.intVar("iron",1000,9999);
silver = model.intVar("silver",100000,999999);
//result is an intermediary to compare copper*neon and iron*silver
//all test would fail if ub was equal to 'org.chocosolver.solver.variables.IntVar.MAX_INT_BOUND'
result = model.intVar("result",0,org.chocosolver.solver.variables.IntVar.MAX_INT_BOUND*10);
model.allDifferent(new IntVar[]{copper,neon,iron,silver}).post();
}
@Test
/***
* Using times dirrectly to make our constraint
* will fail to find any solution because value needed is out of result's bound
*/
public void model1(){
result = model.intVar("result",0,org.chocosolver.solver.variables.IntVar.MAX_INT_BOUND);
model.times(copper, neon, result).post();
model.times(silver, iron, result).post();
Solver solver = model.getSolver();
//solver.showStatistics();
//solver.showSolutions();
solver.findSolution();
//solver.printVersion();
assertEquals(solver.getSolutionCount(),0);
}
@Test
/***
* Working version of the previous test.
*/
public void model1Success(){
model.times(copper, neon, result).post();
model.times(silver, iron, result).post();
Solver solver = model.getSolver();
//solver.showStatistics();
//solver.showSolutions();
solver.findSolution();
//solver.printVersion();
assertEquals(solver.getSolutionCount(),1);
}
@Test
/***
* Creating a model using arithm instead of times directly.
*/
public void model2(){
model.arithm(copper, "*",neon,"=",result).post();
model.arithm(iron, "*",silver,"=",result).post();
Solver solver = model.getSolver();
solver.findSolution();
assertEquals(solver.getSolutionCount(),1);
}
@Test
/***
* Creating a model using the operators function given by IntVar
*/
public void model3(){
copper.mul(neon).eq(result).post();
iron.mul(silver).eq(result).post();
System.out.println(model.toString());
Solver solver = model.getSolver();
solver.findSolution();
assertEquals(solver.getSolutionCount(),1);
}
@Test
/***
* Creating a model using the operators function given by IntVar without using result as an intermediary
*/
public void model4(){
copper.mul(neon).eq(iron.mul(silver)).post();
System.out.println(model.toString());
Solver solver = model.getSolver();
solver.findSolution();
assertEquals(solver.getSolutionCount(),1);
}
}```
My conclusions are that:
The use of MAX_INT_BOUND is unclear, because it is not checked when creating a variable (see setupResult).
The introduction of auxiliary variable is an issue (see for instance model1 and model2).
The issue is that the domain of the right member (result) is not used for defining the domain of the auxiliary variable.
For instance, it is the case for the arithm constraint : see here or just below.
// v1 * v2 OP v3
int[] bounds = VariableUtils.boundsForMultiplication(var1, var2);
IntVar var4 = ref().intVar(bounds[0], bounds[1]);
ref().times(var1, var2, var4).post();
return arithm(var4, op2, var3);
In this case, it is unclear to me why the variable v4 is introduced.
I have slightly improved the code. You can find it below.
import org.chocosolver.solver.Model;
import org.chocosolver.solver.Solver;
import org.chocosolver.solver.variables.IntVar;
import org.junit.Assert;
import org.junit.Before;
import org.junit.Test;
public class ProductTest {
private Model model;
private IntVar copper, neon, iron, silver, result;
@Before
/***
* Tests if choco can solve "copper*neon=iron*silver" via various way of writing constraints
*/
public void setup(){
model = new Model("copper*neon=iron*silver");
copper = model.intVar("copper",100000,999999);
neon = model.intVar("neon",1000,9999);
iron = model.intVar("iron",1000,9999);
silver = model.intVar("silver",100000,999999);
model.allDifferent(new IntVar[]{copper,neon,iron,silver}).post();
}
public void setupResult(int factor) {
//result is an intermediary to compare copper*neon and iron*silver
//all test would fail if factor = 1.
result = model.intVar("result",0, factor * IntVar.MAX_INT_BOUND);
}
public void solve(int solutionCount) {
Solver solver = model.getSolver();
System.out.println(model);
solver.printVersion();
solver.showStatistics();
solver.showSolutions();
solver.findSolution();
Assert.assertEquals(solutionCount, solver.getSolutionCount());
}
@Test
/***
* Using times directly to make our constraint
* will fail to find any solution because value needed is greater than the result's upper bound.
*/
public void model1NoSol(){
setupResult(1);
model.times(copper, neon, result).post();
model.times(silver, iron, result).post();
solve(0);
}
@Test
/***
* Working version of the previous test.
*/
public void model1(){
setupResult(10);
model.times(copper, neon, result).post();
model.times(silver, iron, result).post();
solve(1);
}
@Test
/***
* Creating a model using arithm instead of times directly.
*/
public void model2(){
setupResult(10);
model.arithm(copper, "*",neon,"=",result).post();
model.arithm(iron, "*",silver,"=",result).post();
solve(1);
}
@Test
/***
* Creating a model using the operators function given by IntVar
*/
public void model3(){
setupResult(10);
copper.mul(neon).eq(result).post();
iron.mul(silver).eq(result).post();
solve(1);
}
@Test
/***
* Creating a model using the operators function given by IntVar without using result as an intermediary
*/
public void model4(){
copper.mul(neon).eq(iron.mul(silver)).post();
solve(1);
}
}
https://github.com/chocoteam/choco-solver/issues/978
https://github.com/chocoteam/choco-solver/issues/979
https://github.com/chocoteam/choco-solver/issues/980
You can now move the tests in Failing.java with a comment.
| gharchive/issue | 2022-11-17T09:34:13 | 2025-04-01T04:33:32.424192 | {
"authors": [
"Imp95",
"arnaud-m"
],
"repo": "arnaud-m/cryptator",
"url": "https://github.com/arnaud-m/cryptator/issues/53",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.