id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
2434015555
🛑 FX blog (kty) is down In 1ad90f9, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in f964205 after 7 minutes.
gharchive/issue
2024-07-28T16:29:18
2025-04-01T06:44:41.107814
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/18886", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1292001146
🛑 Portal tem poder quem age app is down In e75c082, Portal tem poder quem age app (https://app.portaltempoderquemage.com.br) was down: HTTP code: 403 Response time: 59 ms Resolved: Portal tem poder quem age app is back up in 6d544a0.
gharchive/issue
2022-07-02T08:20:13
2025-04-01T06:44:41.110210
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/1906", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2487614436
🛑 FX corretor (kty) is down In 4ea495d, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: FX corretor (kty) is back up in 5cf83a6 after 6 minutes.
gharchive/issue
2024-08-26T19:31:51
2025-04-01T06:44:41.112738
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/25268", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2574614103
🛑 FX corretor (kty) is down In d47428c, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: FX corretor (kty) is back up in b95f909 after 10 minutes.
gharchive/issue
2024-10-09T03:03:32
2025-04-01T06:44:41.115166
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/33429", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2601191203
🛑 FX blog (kty) is down In 95fbd6e, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in 4861c78 after 11 minutes.
gharchive/issue
2024-10-21T03:27:28
2025-04-01T06:44:41.117587
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/35456", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2609835305
🛑 FX blog (kty) is down In a8b57ad, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in c291831 after 8 minutes.
gharchive/issue
2024-10-23T20:59:33
2025-04-01T06:44:41.120242
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/35980", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2627539300
🛑 FX blog (kty) is down In de1ca2d, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in b4b4ea5 after 13 minutes.
gharchive/issue
2024-10-31T18:17:44
2025-04-01T06:44:41.122803
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/37433", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2646755768
🛑 FX corretor (kty) is down In 7b08035, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: FX corretor (kty) is back up in 4ad09be after 20 minutes.
gharchive/issue
2024-11-10T03:22:35
2025-04-01T06:44:41.125240
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/39172", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2647528520
🛑 FX blog (kty) is down In 03ef2b4, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in da0de24 after 6 minutes.
gharchive/issue
2024-11-10T18:51:50
2025-04-01T06:44:41.127638
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/39301", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2659145666
🛑 FX blog (kty) is down In f7f936e, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in f22117b after 9 minutes.
gharchive/issue
2024-11-14T15:19:32
2025-04-01T06:44:41.130137
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/40014", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2688290937
🛑 FX corretor (kty) is down In e7242a4, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: FX corretor (kty) is back up in 88e8161 after 6 minutes.
gharchive/issue
2024-11-24T19:54:06
2025-04-01T06:44:41.132791
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/41906", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2728497717
🛑 FX blog (kty) is down In f47c9cd, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in a94f634 after 42 minutes.
gharchive/issue
2024-12-09T23:16:18
2025-04-01T06:44:41.135106
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/44663", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2753028643
🛑 FX blog (kty) is down In 9854725, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in cf4bf1c after 10 minutes.
gharchive/issue
2024-12-20T16:08:27
2025-04-01T06:44:41.137462
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/46616", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2760997028
🛑 FX blog (kty) is down In 7966bd6, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in 38c2cfb after 9 minutes.
gharchive/issue
2024-12-27T15:15:43
2025-04-01T06:44:41.139769
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/47871", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2762188334
🛑 FX corretor (kty) is down In 538923f, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: FX corretor (kty) is back up in bab8722 after 9 minutes.
gharchive/issue
2024-12-29T10:14:43
2025-04-01T06:44:41.142132
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/48201", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1966772783
🛑 krystal site (kty) is down In 651185e, krystal site (kty) (https://krystal.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: krystal site (kty) is back up in 280344d after 11 minutes.
gharchive/issue
2023-10-28T23:41:42
2025-04-01T06:44:41.144693
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/6531", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
893349022
Recheck sizes of all blue team sites Blue team GTmetrix reports in domain alphabetical order https://gtmetrix.com/reports/allaboutberlin.com/ZQ3eIZzQ/ https://gtmetrix.com/reports/anibalsolon.com/lpbPbNPA/ https://gtmetrix.com/reports/appbeat.io/g4lV7tHq/ https://gtmetrix.com/reports/aryan.app/Kh0LtgZO/ https://gtmetrix.com/reports/block.sunappu.net/50nO9CZx/ https://gtmetrix.com/reports/blog.khaleelgibran.com/29w9TeUW/ https://gtmetrix.com/reports/calibratemedia.ca/3l0FMm5e/ https://gtmetrix.com/reports/coopermatt.com/pYvj6XyV/ https://gtmetrix.com/reports/dariusz.wieckiewicz.org/vyHJFMt1/ https://gtmetrix.com/reports/dn15.de/QbohdC4f/ https://gtmetrix.com/reports/drwho.virtadpt.net/lz5lWrje/ https://gtmetrix.com/reports/dusansimic.me/bMAO1wj3/ https://gtmetrix.com/reports/ebass.uk/AMVNsOFr/ https://gtmetrix.com/reports/emadmasroor.github.io/8Qii5REh/ https://gtmetrix.com/reports/evantravers.com/zOowMygS/ https://gtmetrix.com/reports/havenweb.org/vJ5XpDWc/ https://gtmetrix.com/reports/jorp.xyz/2mjjYn24/ https://gtmetrix.com/reports/joshkasuboski.com/IJ6zNmoC/ https://gtmetrix.com/reports/jpdb.io/JE5AlCWd/ https://gtmetrix.com/reports/justalecu.ro/QNPjgSXA/ https://gtmetrix.com/reports/koknut.xyz/PDAMk8ai/ https://gtmetrix.com/reports/leedsbabybank.org/eyWGU1XQ/ https://gtmetrix.com/reports/lexicon.ga/LPnN3Ubg/ https://gtmetrix.com/reports/lopeztel.xyz/q8EoQL7U/ https://gtmetrix.com/reports/marioseijo.com/I4CcCBjU/ https://gtmetrix.com/reports/mcgillij.dev/Kn0QUBT5/ https://gtmetrix.com/reports/melindawyers.com/5IjsrXag/ https://gtmetrix.com/reports/nixnet.services/hD1xBckx/ https://gtmetrix.com/reports/noblogo.org/QhOiEmEJ/ https://gtmetrix.com/reports/nymbus.xyz/GDHivFXc/ https://gtmetrix.com/reports/pablosgweb.com/txDgnjhi/ https://gtmetrix.com/reports/paramdeo.com/Xsq3x0Hi/ https://gtmetrix.com/reports/playerone.kevincox.ca/WdHS9WQT/ https://gtmetrix.com/reports/qunitjs.com/bcqJP6qD/ https://gtmetrix.com/reports/robbie.antenesse.net/cJRQ0C9D/ https://gtmetrix.com/reports/roelbazuin.nl/5MvEBOCu/ https://gtmetrix.com/reports/scott.mortimer.name/MwIj5oL4/ https://gtmetrix.com/reports/sebastian.graphics/tFKaK5GL/ https://gtmetrix.com/reports/simpleblogs.org/F1J9JnTS/ https://gtmetrix.com/reports/slashdev.space/8MDAwEmx/ https://gtmetrix.com/reports/steamosaic.com/nkENDAsy/ https://gtmetrix.com/reports/textpattern.com/69dX9HlU/ https://gtmetrix.com/reports/theandrewbailey.com/BOAfCz5A/ https://gtmetrix.com/reports/thecrow.uk/UN82epNZ/ https://gtmetrix.com/reports/vreeman.com/xjHysLWD/ https://gtmetrix.com/reports/wolfgang.lol/FZ8becIa/ https://gtmetrix.com/reports/wor.do/ZfXcDkcG/ https://gtmetrix.com/reports/www.matthewgraybosch.com/n5zJvvml/ https://gtmetrix.com/reports/www.yeetpc.com/890xOKcx/ https://gtmetrix.com/reports/xanny.family/kOCF4Pa4/ https://gtmetrix.com/reports/zakr.es/64KshF4A/ https://gtmetrix.com/reports/zinzy.website/tw2apXcN/ Are they still in size order once the change is made? If not, it will mess up the list. Alternatively, do you know how we can automatically sort the list by size on build? @kevquirk They are now in order. This is my process Open all sites in the browser and check which ones are still live (Done) Create a new GTmetrix report and fix size (Blue done) Move sites that are in the wrong team Reorder based on size I have no finished the Blue Team.
gharchive/pull-request
2021-05-17T13:44:47
2025-04-01T06:44:41.150922
{ "authors": [ "kevquirk", "kmutahar" ], "repo": "kevquirk/512kb.club", "url": "https://github.com/kevquirk/512kb.club/pull/375", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
63943323
Re-enable support for download-status On 4.x.x, support for download-status was removed, so now there's no way to have a feedback of the current progress. Since now several files can be downloaded at once, it would be ok, if the download progress is the total of all the files, but a better option is to have support for an independent progress for each file if they don't interfere in the screen. Yup, just need https://github.com/sindresorhus/got/pull/41 to be merged. Thank you.
gharchive/issue
2015-03-24T09:04:58
2025-04-01T06:44:41.152722
{ "authors": [ "kevva", "piranna" ], "repo": "kevva/download", "url": "https://github.com/kevva/download/issues/56", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2425467498
problem with tracker hit associations (VXDTrackerHitRelations)? using the nightlies, the last CLDConfig and CLD_o2_v06, VXDTrackerHitRelations (as well as other relation collections for the tracker hits) rec and sim index are filled with -2s an example file can be found here: /eos/experiment/fcc/ee/datasets/CLD_tracking/condor/Pythia_eval/Zcard_CLD_v3/1/out_reco_edm4hep_edm4hep.root I would suspect this check https://github.com/key4hep/k4EDM4hep2LcioConv/blob/73793abf6794c48e27c0a584e645d2f0a0377f48/k4EDM4hep2LcioConv/include/k4EDM4hep2LcioConv/k4Lcio2EDM4hepConv.ipp#L856 Still need to figure out why it fails though, maybe something in the translation mappings has changed unexpectedly recently @tmadlener? When did it start failing? My suspicion would be that it's an issue related to TrackerHitPlane vs TrackerHit in the filling of the association. We have removed the association from TrackerHitPlane to SimTrackerHit in EDM4hep (https://github.com/key4hep/EDM4hep/pull/331). The corresponding changes in the converter are: https://github.com/key4hep/k4EDM4hep2LcioConv/pull/73 and https://github.com/key4hep/k4EDM4hep2LcioConv/pull/76 I think I found one suspicious bit but I don't think that it was the source of this error https://github.com/key4hep/k4EDM4hep2LcioConv/pull/83 As there are still associations created but just default initialised (which I would think from the -2) this check should not have been the problem. Does it maybe just not work because the typeMapping.trackerHits that is also passed for TrackerHitPlanes contains a mapping to TrackerHit3D and not to the TrackerHit interface? I think that would fit with the above error :) Does it maybe just not work because the typeMapping.trackerHits that is also passed for TrackerHitPlanes contains a mapping to TrackerHit3D and not to the TrackerHit interface? I think that would fit with the above error :) indeed
gharchive/issue
2024-07-23T15:12:44
2025-04-01T06:44:41.162405
{ "authors": [ "Zehvogel", "andresailer", "doloresgarcia", "tmadlener" ], "repo": "key4hep/CLDConfig", "url": "https://github.com/key4hep/CLDConfig/issues/48", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
585044705
Feature Request: autobuild to dockerhub It would be really nice to have an autobuild process that publishes a docker image of the Dockerfile-ca file to dockerhub. This would make deploying a lot easier and faster Especially so as I found it impossible on several ARM models to successfully complete the builds... I resorted to multi-arch builds and transfer the image. With an official multi arch docker image, that would be a problem no more... Maybe, first step is to make the tooling multi-arch, so that it's easier to figure out how to cross build the image. The best resource I found for doing this manually was: https://medium.com/@artur.klauser/building-multi-architecture-docker-images-with-buildx-27d80f7e2408 Resource found for circleci: https://github.com/varnishcache/varnish-cache/pull/3263/files (support seems still behind the concurrence) github: https://github.blog/changelog/2019-12-03-github-actions-self-hosted-runners-on-arm-architectures/ There is a little elevated container which seems be doing all the qemu related stuff for people who don't want to know the details docker run --rm --privileged multiarch/qemu-user-static --reset -p yes referencedin the cited medium article... Looks like this https://github.com/jitesoft/docker-keybase-sshca is the best "mirror" with aparent arm64/aarch64 support...
gharchive/issue
2020-03-20T12:42:47
2025-04-01T06:44:41.166937
{ "authors": [ "blaggacao", "krezreb" ], "repo": "keybase/bot-sshca", "url": "https://github.com/keybase/bot-sshca/issues/83", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
220397799
keybase pgp select --no-import asks for a passphrase for "import of key" I'm trying to get my public pgp key into keybase, and I don't want to import my secret key. When I run keybase pgp select --no-import I'm asked for my passphrase twice. The first time is by the standard dialog that gpg uses on my system. After I comply I'm asked a second time by keybase, which says "Reason: Import of key into Keybase keyring". I'm now feeling confused and suspicious: I used --no-import to say I didn't want to import, but the wording of the dialog seems to contradict that. And I don't know why I'm being asked for my passphrase a second time. I was under the impression that the passphrase was just used to sign the public key, and so it would only be needed once, by gpg (and not keybase). Have I misunderstood something? Is this second passphrase prompt necessary, and if so could it be made to explain itself more clearly? (I found this comment which said: There's one instance in which we prompt for your PGP password directly [...] that's adding the username you@keybase.io to your public key if it's not already there and so I had already made a uid for my key ("keybase.io/mickbrooks mickbrooks@keybase.io") matching the pattern I found from another user. Did I miss anything?) Thanks for the report, we'll take a look. CC: @zapu can you make a ticket? Internal id: CORE-4898 Thank you for the report! Had the same issue. It only imports the public key, you can verify with keybase pgp export -s which should not print you private key afterwards. It should not request the pgp password for the public key... any updates here? even temporary clarification would be nice (I didn't even know about --no-import and was still suspicious as to why I'd want to type my passphrase anywhere other than the gpg toolchain) Sorry, this is one of several GitHub issues and internal tickets about this problem. A fix was merged a while ago: https://github.com/keybase/client/pull/7414 that will shell out to GPG in order to make a reverse-signature of PGP key. We reverse-sign to prove that user uploading a key is also the owner of said key. We used to do this with go-crypto library, but now the secret key will not leave gpg keyring anymore. Let me know if you have any further questions! Thanks. @zapu thanks for the update! So, to confirm, if my download of the keybard .deb would have had that pull request in it, then I'd not be prompted by the keybase app (that screenshot above) as I was? Right, you would not see a Keybase prompt for key passphrase during keybase pgp select --no-import, but you would still see GPG prompting you for the passphrase. Because now we do that: 2017-07-12T21:53:34.072737+02:00 ▶ [DEBU keybase log.go:30] 0d4 | GPGKey reverse delegate to CLI 2017-07-12T21:53:34.072996+02:00 ▶ [DEBU keybase gpg_cli.go:225] 0d5 GPG version: gpg (GnuPG) 2.1.15 libgcrypt 1.7.2-beta Copyright (C) 2016 Free Software Foundation, Inc. License GPLv3+: GNU GPL version 3 or later <https://gnu.org/licenses/gpl.html> This is free software: you are free to change and redistribute it. There is NO WARRANTY, to the extent permitted by law. Home: /tmp/tmp.1exRTUE0z3/.gnupg Supported algorithms: Pubkey: RSA, ELG, DSA, ECDH, ECDSA, EDDSA Cipher: IDEA, 3DES, CAST5, BLOWFISH, AES, AES192, AES256, TWOFISH, CAMELLIA128, CAMELLIA192, CAMELLIA256 Hash: SHA1, RIPEMD160, SHA256, SHA384, SHA512, SHA224 Compression: Uncompressed, ZIP, ZLIB, BZIP2 2017-07-12T21:53:34.073071+02:00 ▶ [DEBU keybase gpg_cli.go:353] 0d6 | running Gpg: /usr/bin/gpg2 --armor --sign -u ab1059a2529b7619986480c8c9293dacd499084d 2017-07-12T21:53:34.073104+02:00 ▶ [DEBU keybase gpg_cli.go:360] 0d7 | setting GPG_TTY=/dev/pts/18 2017-07-12T21:53:53.662046+02:00 ▶ [DEBU keybase log.go:30] 0d8 - GPGKey Signing -> %!s(<nil>) so the /usr/bin/gpg2 --armor --sign -u ab1059a2529b7619986480c8c9293dacd499084d command needs to unlock key, but otherwise the secret key is not exported. Thanks for the update. The new version worked as I expected, so I'll close this issue now. I just ran into the same dialog before learning about --no-import. I had just generated a new keypair to replace an expiring one and was trying to upload only the public part to Keybase. Called without --no-import, keybase pgp select creates a dialog box asking me for the secret part's passphrase. Reason: Import of key into Keybase keyring isn't quiiiite specific enough. Which key, and why? Same here. Killed my great 1st impression somewhat. Because some people like to use keybase instead of pgp for decrypt/sign. Note that it's just importing to a locally stored keychain, encrypted using the same policies as the reset of your private keybase credentials (like your device private key). keybase help gpg has more specifics. I guess Keybase can be use as a GPG keyring app if you don't have already one. And since most users do not use this kind of app, it is primarily made for them ... Might be an explanation @irogaro exactly
gharchive/issue
2017-04-08T12:48:37
2025-04-01T06:44:41.177920
{ "authors": [ "JustATrick", "irogaro", "jzacsh", "maxtaco", "ransford", "shredding", "victortoso", "zapu" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/issues/6549", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
354786487
Don't black bar for 'EOF from server' error r? @keybase/react-hackers Does this EOF error resolve itself after a while? @songgao yeah, they happen when the service restarts, and are usually transient. Before this you'd see the reachability screen and then a black bar on reconnecting. This removes the black bar. Cool; thanks. I'll be able close some KBFS ticket too then! @chrisnojima I moved the check to a new function isErrorTransient, PTAL
gharchive/pull-request
2018-08-28T15:33:36
2025-04-01T06:44:41.180287
{ "authors": [ "buoyad", "songgao" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/pull/13470", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
117913568
no parens around single param arrows @keybase/react-hackers (blah) => {} must now be blah => {} :thumbsup:
gharchive/pull-request
2015-11-19T21:39:30
2025-04-01T06:44:41.181245
{ "authors": [ "chrisnojima", "cjb" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/pull/1367", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
172568817
Mobile edit proof popup DESKTOP-1655 Implements a mobile version of the desktop edit proof popup menu. See below: The mobile version is currently following the design spec of the iOS Activity Sheet (with expectations that this will eventually be changed to something platform agnostic). I've followed the design relatively closely, though I omitted elements such as: I used the Keybase standard font instead of the iOS font in the desing; and I did not add the blurring effect. I have matched all the proof states that the desktop version covers with implementations that work from the design docs when their present and educated guesses when they're not. Here are some screenshots: Standard Popup Popup for Errored Proof Popup for valid proof Here is a demo of the touch interaction (notice how the header does not respond to touches because it lacks an onClick handler): @keybase/react-hackers @chrisnojima I've removed visible from the props 👍
gharchive/pull-request
2016-08-22T22:19:34
2025-04-01T06:44:41.185819
{ "authors": [ "awendland", "chrisnojima" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/pull/3996", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2492329078
Module '"keycloakify/login"' has no exported member 'createGetKcContext' in Keycloakify 10.0.3 Hi, recently we've been trying to bump keycloakify version from 9.1.4 to 10.0.3 and we've encountered two issues: Module '"keycloakify/login"' has no exported member 'createGetKcContext' and in fact after checking out lib files I don't see that export anymore, however it's absence is not mentioned in migration guide We cannot import KcContext type because there seems to be casing issue. Funny, that for account, import statement is exact the same as it was before (notice kcContext starting with lowercase) . I've fixed that by changing casing in application components, however I wanted to let you know as this is undocumented change (at least in migration guide) while 2nd issue is easily fixable, the first one we don't know how to tackle, any ideas? Hello @PrzemyslawPolrolniczak, I'm sorry to have to inform you that updating form v9 to v10 isn't as easy as bumping the package version number 😞 https://docs.keycloakify.dev/faq-and-help/migration-guides/v9-greater-than-v10
gharchive/issue
2024-08-28T14:48:46
2025-04-01T06:44:41.287957
{ "authors": [ "PrzemyslawPolrolniczak", "garronej" ], "repo": "keycloakify/keycloakify", "url": "https://github.com/keycloakify/keycloakify/issues/626", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
761845871
Client Allocation Request for: [chenxi] Client Allocation Questions Core Information Name: wufei Website / Social Media: http://blog.sina.com.cn/u/1662458344 DataCap Requested: 4TiB Addresses to be Notarized: f132avo3voj34w4ddxbevr77o4p5mdddrcxhgmfcq Notary Requested: philippbanhardt Hey @wufeivsko - thanks for your request. Can you kindly add more information / context here on: yourself , e.g. include links to social media / news mentions etc your history working on IPFS / Filecoin, e.g. notable contributions to the community, type of engagement shown over time etc type, size, source and your relationship with the data you (and by extension other Discover participants) are looking to store on Filecoin, e.g. public data vs private data etc how you intend to store the data on Filecoin, e.g. redundancy, location / miner preferences, etc Thanks Introduce to myself: my social media: https://www.linkedin.com/in/飞-吴-726847152/ About my Github: https://github.com/wufeivsko I am a senior blockchain operation and maintenance developer and also the specialist of data processing filesystem, contrainers, system supervisor, etc. I have a great intrest on Filecoin/IPFS and planning to store our chain data onto the Filecoin/IPFS network, the community that I am keeping watching on is http://ipfs.cn/. For my working experience, I have deployed networks for ethereum and cosmos for my team project, etc. Due to the day-by-day increasing exchange transactions and big amount of contract data stored on our chain, there became a large requirement for data store. We are planning to put our historical data onto ipfs, and I'm planning to build my project based on my fork tree of https://github.com/wufeivsko/ipfs-webui.git. Due to the huge amount of contract-user-data space requirement, we need about 2-6 TB disk space to save the contract data and the exchange transactions. To us, the Filecoin should be an economical and practical solution which data handling can be elastic and scalable, also keeping privacy and security. Hey @wufeivsko can you kindly clarify What data you are looking to store on the network? The way I read your answer above is that you are looking to store chain data? How you are intending to store the data on the network, e.g. how / which miners to select etc Also - can you please add more information / links about yourself? Can also be in Chinese web, but the LinkedIn profile you shared has very little social reputation Thanks Hi, Here is the keypoints how we interact with IPFS/Filecoin based on the questions. What data we are looking to store on the network: We aimed to store both the detailed data from user transaction( not the raw transaction from block, we unpack the trx and parsed the data and decrypt the data already)---- the data will be multiple kinds based on our system , like: some trainning trainning models' dataset collected from data-feed (some of them we treat as the payload of one transaction), some key graphs for feature matching, also we need put some important transactions to IPFS due to we need keep some logical sequence in our system, etc. How you are intending to store the data on the network, e.g. how / which miners to select etc I plan to store our ipfs data with about 3 replica at least. Currently our users are located in Asia, then about the miners I'd like to choose in Asia Areas for data store testing and verifications . Thanks. @wufeivsko is the data public / non encrypted and will be accessible to everyone? Thanks Yes, all transaction data are public for everyone as everyone can get the transactions. And we got the public graph from public sharing resources such from shopping web, online-public datasets, or sharing feature data training models, etc. (which has no personal privacy sensitive data) trainning for data-feed to keep the dataset safe and available for everyone. Request Approved Your Datacap Allocation Request has been approved by the Notary Message sent to Filecoin Network bafy2bzacecgzhyf2igjxzcvuchazmsadcxwhj6p7r73k2hb57ywavypgdvjyq Address f132avo3voj34w4ddxbevr77o4p5mdddrcxhgmfcq Datacap Allocated 4 TiB
gharchive/issue
2020-12-11T03:24:59
2025-04-01T06:44:41.300296
{ "authors": [ "philippbanhardt", "wufeivsko" ], "repo": "keyko-io/filecoin-clients-onboarding", "url": "https://github.com/keyko-io/filecoin-clients-onboarding/issues/123", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
408062601
Add test objects to gitignore Add the following left after tests: swtpm2_scripts/NVChip test/.auditing-0 test/NVChip test/testdata.sqlite Fixes #46 @jetwhiz , a simple one to review if you don't mind - I can then merge / close it. thanks! FYI @jetwhiz is out on vacation until Friday. Looks good to me.
gharchive/pull-request
2019-02-08T09:09:14
2025-04-01T06:44:41.302418
{ "authors": [ "lukehinds", "nabilschear" ], "repo": "keylime/python-keylime", "url": "https://github.com/keylime/python-keylime/pull/51", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
47335898
LICENSE.md too complicated and hard to maintain Do we really need to list all those software at the bottom of the LICENSE.md page? Some of those are probably part of K modules that have either disappeared or will disappear from the kernel. Also, we say: The K software contains code written by third parties. Such software will have its own individual LICENSE file in the directory in which it appears. So then, why bother listing all the following on the main License page? The following pieces of software have additional or alternate copyrights, licenses, and/or restrictions: Maude (http://maude.cs.uiuc.edu) is licensed under the GPL license. Stratego (http://strategoxt.org) is licensed under the LGPL license. Xstream (http://xstream.codehaus.org/) is licensed under the BSD license. XML Pull (http://www.xmlpull.org) API is free. Xpp3 (http://www.extreme.indiana.edu/xgws/xsoap/xpp) is licensed under Indiana University Extreme! Lab Software License. KXML2 (http://kxml.sourceforge.net/kxml2/) is licensed under the BSD license. SSCTarjan (https://github.com/indy256/codelibrary) is licensed under the Unlicense. Alphanum (http://www.davekoelle.com/alphanum.html) is licensed under the LGPL license. I would indeed make sure that the licenses of these software are included in the directories where they are used, and then replace the last section of the K License with the following simpler paragraph: The K software may contain code written by third parties. Such software will have its own individual LICENSE file in the directory in which it appears. The disclaimer of warranty in the University of Illinois Open Source License applies to all code in the K Distribution, and nothing in any of the other licenses gives permission to use the names of the institutions or companies forming the K Team to endorse or promote products derived from this Software. Since we are at this, I would also change the paragraph above, Neither the names of the K Team, the University of Illinois at Urbana-Champaign, the University Alexandru-Ioan Cuza, nor the names of its contributors may be used to endorse or promote products derived from this Software without specific prior written permission. which does not mention RV anyway, as follows: Neither of the names of the institutions or companies forming the K Team, nor the names of its contributors may be used to endorse or promote products derived from this Software without specific prior written permission. @dwightguth feel free to change the above however you wish to make it all sound good. Oncall should probably go through this and update the license so that we are complying with all the open source licenses that we import in this project. Basically that means picking through every transitive dependency and making sure that LICENSE.md contains any information which the license in question requires us to replicate.
gharchive/issue
2014-10-30T22:10:48
2025-04-01T06:44:41.364454
{ "authors": [ "dwightguth", "grosu" ], "repo": "kframework/k", "url": "https://github.com/kframework/k/issues/1136", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
121898128
Merged PR is gone I cannot find the merged PR #1754 in the latest master. There is no history regarding neither merge nor revert of the PR. I suspect it is somehow rebased. Or, what am I missing? Any idea? The PR is still there. Only the branch is gone -- Brandon deleted it after the merge. There's usually no point in keeping merged branches around. What do you need it for? Seem like github added a "Restore branch" button -- see towards the end of #1754. What I'm saying is that the PR was merged into master, but the change is not found in the current master. I'm trying to find who reverted the change, but I cannot find the attempt to revert in history. Even, I cannot find the merge history neither. That's why I suspect someone accidently rebased it, but I have no idea how to find it. Sent from my iPhone On Dec 13, 2015, at 9:53 AM, Cosmin Radoi notifications@github.com wrote: Seem like github added a "Restore branch" button -- see towards the end of #1754. — Reply to this email directly or view it on GitHub. This is indeed strange. I'll look into it. It might be something like this: http://stackoverflow.com/questions/13468027/the-mystery-of-the-missing-commit-across-merges @daejunpark, it's of your own making: https://github.com/kframework/k/commit/daa1978488c3f9c7c0a040c22462e8d5c784bc7a :) @cos Oh, it was due to me. Thanks a lot! What happened is that I cherry-picked commits from a pending PR, but reverted them later because the pending PR had not been merged before my PR was merged. I thought that merging the pending PR will bring the changes back, but it didn't. This is counter-intuitive because my reverting the commits preceded merging the pending PR. That is, merging a PR could have no effect. Is there any way to prevent this kind of mistakes later? But, still the strange thing is that 'git log' didn't show such a history. For example, the following directory history does not show the 'merge-revert' history: https://github.com/kframework/k/commits/master/k-distribution/tests/regression/kore-cell-fragments This is counter-intuitive because my reverting the commits preceded merging the pending PR. That is, merging a PR could have no effect. Is there any way to prevent this kind of mistakes later? I don't think so. I think the principle is that git prioritizes the latest commits. If you don't make explicit commits that re-add the content, git considers the reverts as authoritative. But, still the strange thing is that 'git log' didn't show such a history. For example, the following directory history does not show the 'merge-revert' history: https://github.com/kframework/k/commits/master/k-distribution/tests/regression/kore-cell-fragments I had some trouble finding it also. I don't know why they disappear. I eventually found them with gitk. What happened is that I cherry-picked commits from a pending PR, but reverted them later because the pending PR had not been merged before my PR was merged. I thought that merging the pending PR will bring the changes back, but it didn't. If cherry-pick actually made a new commit, and then the pr was merged like with the merge command, that's exactly how it should have worked. Other things like rebase (and pull may use merge or rebase depending on settings) might act differently. For education purposes, understand that the tip of a branch and its history are different things. Thus there is a distinction between rebasing a commit out of the history, and reverting it. Both lead to the same working copy, but one preserves history and adds a new revert commit, the other refactors history and removes the existing commit. When you merge changes, what actually happens is that the changes of both branches are combined using a merge strategy to create a new commit that is the child of both commits. Changes (deltas) applied in both parents are simply redundant and merge cleanly. So what actually happened was you took one branch with the feature commits and one branch with both those commits and the revert commits, and created a new commit that also had both and thus was the reverted form. Something to bear in mind in the future so you can avoid this happening again... On Dec 14, 2015 9:35 PM, "bmmoore" notifications@github.com wrote: What happened is that I cherry-picked commits from a pending PR, but reverted them later because the pending PR had not been merged before my PR was merged. I thought that merging the pending PR will bring the changes back, but it didn't. If cherry-pick actually made a new commit, and then the pr was merged like with the merge command, that's exactly how it should have worked. Other things like rebase (and pull may use merge or rebase depending on settings) might act differently. — Reply to this email directly or view it on GitHub https://github.com/kframework/k/issues/1905#issuecomment-164635184. Merges generally only take into account the state at the tip of the tree and at the common ancestor. In particular, cherry-picking a commit onto master, reverting the cherry-picked commit, and then merging the pull request should result in the change being applied: $ git init Initialized empty Git repository in /home/brandon/test/git/.git/ $ echo 'Hello' > foo $ git add foo $ git commit -a -m starting $ git checkout -b feature Switched to a new branch 'feature' $ echo 'Goodbye' > bar $ git add bar $ git commit -a -m "Feature work" [feature 1db360b] Feature work 1 file changed, 1 insertion(+) create mode 100644 bar $ git checkout master Switched to branch 'master' $ git cherry-pick 1db360b [master d0f1caf] Feature work Date: Tue Dec 15 12:37:22 2015 -0600 1 file changed, 1 insertion(+) create mode 100644 bar $ echo ' World' >> foo $ git commit -a -m 'More master work' [master edbddf1] More master work 1 file changed, 1 insertion(+) $ git revert --no-edit d0f1caf [master 0121bf6] Revert "Feature work" 1 file changed, 1 deletion(-) delete mode 100644 bar $ git merge --no-edit feature Merge made by the 'recursive' strategy. bar | 1 + 1 file changed, 1 insertion(+) create mode 100644 bar $ cat foo bar Hello World Goodbye $ git log --graph --oneline * ec3522c Merge branch 'feature' |\ | * 1db360b Feature work * | 0121bf6 Revert "Feature work" * | edbddf1 More master work * | d0f1caf Feature work |/ * 49e44dd starting The revert would only take precedence in a merge if the reverted commit was already part of the common ancestor of the two commits. understand that the tip of a branch and its history are different things. That's certainly true, but git merges are only interested at the state of the code at the tips of branches and at certain common ancestor points, not at all in how they got to that state (as opposed to VCS like darcs which consider sets of changes primitive instead of tree states).
gharchive/issue
2015-12-13T04:19:21
2025-04-01T06:44:41.380005
{ "authors": [ "bmmoore", "cos", "daejunpark", "dwightguth" ], "repo": "kframework/k", "url": "https://github.com/kframework/k/issues/1905", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
168059226
Circular dependencies in k-distribution/include/builtins E.g. bool.k -> ../k-prelude.k -> builtins/builtins.k -> bool.k int.k -> k-equal.k -> list.k -> k-equal.k `-> int.k These are used by K3.x. We're trying to move towards K4, which uses files only from /include/builtin. Hopefully, this won't be an issue then since we're planning to remove them completely. Cool, thanks Radu. Any timeframe for when you’ll move to K4? On 22 Aug 2016, at 4:20 PM, Radu Mereuta notifications@github.com wrote: These are used by K3.x. We're trying to move towards K4, which uses files only from /include/builtin. Hopefully, this won't be an issue then since we're planning to remove them completely. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/kframework/k/issues/2192#issuecomment-241427762, or mute the thread https://github.com/notifications/unsubscribe-auth/AIOhkhJHVz_jSGnHlofyBJL1ivsmR5rJks5qibAbgaJpZM4JXDpT. @cos might be able to answer that better, but in the meantime, you can use it in the master branch of the K framework. To call it K4, we just have to iron out some kinks, and clean out old files. @adrianherrera, we have already released K4: https://github.com/kframework/k/releases/tag/v4.0.0 We'll do another release in about a month with new features and cleaning up the old files. related to deprecated code so closing
gharchive/issue
2016-07-28T10:07:03
2025-04-01T06:44:41.385730
{ "authors": [ "adrianherrera", "cos", "radumereuta" ], "repo": "kframework/k", "url": "https://github.com/kframework/k/issues/2192", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
274310471
add power-modulo operator to UIUC K This is needed to execute the power-modulo operation in the modexp precompiled contract in the EVM semantics. when we need it we can just represent it with a macro in the prelude that does power and modulo directly. but until we verify contracts that call precompiled contracts it won't be an issue.
gharchive/pull-request
2017-11-15T21:28:04
2025-04-01T06:44:41.387786
{ "authors": [ "dwightguth" ], "repo": "kframework/k", "url": "https://github.com/kframework/k/pull/2372", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1661964947
sftp 上传文件网速较慢的问题 我上传一个jar包到云服务器上,一般情况下不会这么慢(使用termious时),但是使用 electern 网速明显下降,约 300 kb / s,但是非常稳定(点赞)。 请问是否有什么办法可以优化一下提高速度呢 ~ 此致, 敬上 ~ It sounds like you're saying you noticed your network upload speeds drop while electern is open? I haven't noticed anything like that myself so I can't really debug this, but I just pushed a prerelease that updates electern to the latest Electron version. Maybe try that and see if it doesn't have the same issue? I've gone ahead and updated Electron again in light of the recent libwebp CVE and released 0.2.0. Given no further response here and my inability to reproduce the issue, I'm closing it at this time.
gharchive/issue
2023-04-11T08:00:20
2025-04-01T06:44:41.390272
{ "authors": [ "heyanyidui", "kfranqueiro" ], "repo": "kfranqueiro/electern", "url": "https://github.com/kfranqueiro/electern/issues/6", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
516153248
having some issues with lures pls help i did this lures edit redirect_url <1> https://login.live.com and i got this [err] lures: edit: strconv.Atoi: parsing "<1>": invalid syntax what may be the problem i did this lures edit redirect_url <1> https://login.live.com and i got this [err] lures: edit: strconv.Atoi: parsing "<1>": invalid syntax what may be the problem i can help you with the issue..rickorkins@gmail.com i did this lures edit redirect_url <1> https://login.live.com and i got this [err] lures: edit: strconv.Atoi: parsing "<1>": invalid syntax what may be the problem remove the <>
gharchive/issue
2019-11-01T14:39:22
2025-04-01T06:44:41.396516
{ "authors": [ "DSUN01", "arinze12345", "horllste" ], "repo": "kgretzky/evilginx2", "url": "https://github.com/kgretzky/evilginx2/issues/398", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2165306597
🛑 AU is down In 1fcd457, AU ($AU) was down: HTTP code: 521 Response time: 431 ms Resolved: AU is back up in 65f6b22 after 8 minutes.
gharchive/issue
2024-03-03T11:13:20
2025-04-01T06:44:41.398636
{ "authors": [ "khadanja" ], "repo": "khadanja/upptime", "url": "https://github.com/khadanja/upptime/issues/475", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
278718697
Fast blink when the transition starts Hello I have got a small issue. When the source is updated, i have a white blink of the image just before the animation starts. I just made an update within your code to create a "fullscreen" image. I replaced : <div style={{ ...defaultStyle, ...{ position: "relative" } }}> with <div style={{ ...defaultStyle, ...{ position: "absolute" } }}> Using React 16 Thanks for opening the issue, feel free to submit a PR. I'll have a look. Sorry i did not explained well. I still have the blink and no way to remove it. The update i made within the code is for my specific use. However even without the code change, i have the blink/flash before each transition this is still an issue. I think it happens when a user has the cache disabled and the image has to load 1st?
gharchive/issue
2017-12-02T21:19:18
2025-04-01T06:44:41.402444
{ "authors": [ "khanglu", "pihomeserver", "rcolepeterson" ], "repo": "khanglu/react-crossfade-image", "url": "https://github.com/khanglu/react-crossfade-image/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2447437458
🛑 jellyfin is down In 3068264, jellyfin (https://jellyfin.hmuy.tech) was down: HTTP code: 502 Response time: 3743 ms Resolved: jellyfin is back up in 6453b91 after 2 hours, 46 minutes.
gharchive/issue
2024-08-05T01:54:51
2025-04-01T06:44:41.413938
{ "authors": [ "khanhmuy" ], "repo": "khanhmuy/webstatus", "url": "https://github.com/khanhmuy/webstatus/issues/359", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1245541740
Where is the Debug Logfile I recently switched to a mac. I installed that extention like I did on my previous thinkpad with the exact same settings and it doesn't work (either on firefox, chromium or chrome). I enabled the debug logs but I cannot find those. Any thing? Probably the logs are shown in the browser's debug tool/console (press F12 in Firefox). pon., 23 maj 2022 o 20:20 Arnold @.***> napisał(a): I recently switched to a mac. I installed that extention like I did on my previous thinkpad with the exact same settings and it doesn't work (either on firefox, chromium or chrome). I enabled the debug logs but I cannot find those. Any thing? — Reply to this email directly, view it on GitHub https://github.com/khloke/play-to-xbmc-chrome/issues/187, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAFN6OEJ3C3XGNTF7C5K55LVLPEAPANCNFSM5WWVZ6WQ . You are receiving this because you are subscribed to this thread.Message ID: @.***> -- Maciej Sitarz
gharchive/issue
2022-05-23T18:20:40
2025-04-01T06:44:41.421941
{ "authors": [ "altf4arnold", "maciex" ], "repo": "khloke/play-to-xbmc-chrome", "url": "https://github.com/khloke/play-to-xbmc-chrome/issues/187", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1216326557
Pluralsight Clone Project title Pluralsight Clone Project Description It's just a UI of Pluralsight Stack: And in order to tick the check box just but x inside them for example - [x] like this. Please delete options that are not relevant. [x] Html [x] CSS [x] JavaScript /assign
gharchive/issue
2022-04-26T18:26:30
2025-04-01T06:44:41.430258
{ "authors": [ "gurjeetsinghvirdee" ], "repo": "khushi-purwar/WebDev-ProjectKart", "url": "https://github.com/khushi-purwar/WebDev-ProjectKart/issues/1010", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2190847363
Update README.md Remove rarechems.com due to seized domain Removing a link to site that has been taken down
gharchive/pull-request
2024-03-17T20:18:20
2025-04-01T06:44:41.442910
{ "authors": [ "darnocer" ], "repo": "kianenigma/awesome-psychedelics", "url": "https://github.com/kianenigma/awesome-psychedelics/pull/4", "license": "Unlicense", "license_type": "permissive", "license_source": "github-api" }
303404701
NPE fix fix npe https://github.com/kicksolutions/swagger2puml/issues/37 @msantosh1188 thank you. It seems I have no rights to merge PR..
gharchive/pull-request
2018-03-08T09:06:05
2025-04-01T06:44:41.443954
{ "authors": [ "kolisko" ], "repo": "kicksolutions/swagger2puml", "url": "https://github.com/kicksolutions/swagger2puml/pull/38", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1006284430
[prod nighlty] 7.13.0 Thank you for submitting this pull request JIRA: (please edit the JIRA link if it exists) link referenced Pull Requests: (please edit the URLs of referenced pullrequests if they exist) referenced pull request 1 referenced pull request 2 referenced pull request 2 etc. How to retest a PR or trigger a specific build: * a pull request please add comment: Jenkins retest this looks good so that https://github.com/kiegroup/kie-jenkins-scripts/pull/1118/files#diff-9d55ce9075a526a3684617ae54e1f15ee81673ba5c89794b916086a8ffacc4d9R40 can be updated always on the top of job, which creates the new branches looks good so that https://github.com/kiegroup/kie-jenkins-scripts/pull/1118/files#diff-9d55ce9075a526a3684617ae54e1f15ee81673ba5c89794b916086a8ffacc4d9R40 can be updated always on the top of job, which creates the new branches https://issues.redhat.com/browse/BXMSPROD-1483
gharchive/pull-request
2021-09-24T09:48:33
2025-04-01T06:44:41.467777
{ "authors": [ "Ginxo", "mareknovotny" ], "repo": "kiegroup/kie-jenkins-scripts", "url": "https://github.com/kiegroup/kie-jenkins-scripts/pull/1118", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
594985968
[master][KOGITO-1593] Cucumber tests: Add Cucumber test for resource requests … …and limits JIRA Ticket: https://issues.redhat.com/browse/KOGITO-1593 Description: Add a Cucumber tests making sure that if resource requests and limits are defined in CR/using CLI then they are applied to Kogito application. Check both build resources and runtime resources. How to run tests: ../hack/run-tests.sh --feature "features/deploy_service_resources.feature" Many thanks for submiting your Pull Request :heart:! Please make sure that your PR meets the following requirements: [x] You have read the contributors guide [x] Pull Request title is properly formatted: [KOGITO-XYZ] Subject [x] Pull Request contains link to the JIRA issue [x] Pull Request contains description of the issue [ ] Pull Request does not include fixes for issues other than the main ticket [ ] Your feature/bug fix has a unit test that verifies it [x] You've tested the new feature/bug fix in an actual OpenShift cluster Depends on https://github.com/kiegroup/kogito-cloud-operator/pull/263 Rebase done. Apart from minor comments looks good.
gharchive/pull-request
2020-04-06T10:15:33
2025-04-01T06:44:41.507368
{ "authors": [ "Sgitario", "ricardozanini", "sutaakar" ], "repo": "kiegroup/kogito-cloud-operator", "url": "https://github.com/kiegroup/kogito-cloud-operator/pull/275", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
972435526
KOGITO-5705 rename of ddl files and their usage Jira https://issues.redhat.com/browse/KOGITO-5705 Related PR: kogito-app (will update pull request link shortly) files like the exists_table.sql files removed from kogito-ddl.zip artifact (since not ddl) ddl files renamed acc. to repo/module that uses them Run all builds Please add comment: Jenkins retest this Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? Can one of the admins verify this patch? @cristianonicolai pls review ok to test
gharchive/pull-request
2021-08-17T08:30:19
2025-04-01T06:44:41.519085
{ "authors": [ "cristianonicolai", "kie-ci", "uteegozi" ], "repo": "kiegroup/kogito-runtimes", "url": "https://github.com/kiegroup/kogito-runtimes/pull/1531", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
711233730
[KOGITO-3460] Create native daily build Job that clone and compiles all the repos in native mode @radtriste @cristianonicolai Related PR: https://github.com/kiegroup/kogito-examples/pull/375 Jenkins retest this jenkins retest this please @danielezonca Seems Jenkins is having problem starting new job. Once back, I will restart this job. jenkins retest this @danielezonca @cristianonicolai can be merged
gharchive/pull-request
2020-09-29T15:36:01
2025-04-01T06:44:41.521841
{ "authors": [ "danielezonca", "radtriste" ], "repo": "kiegroup/kogito-runtimes", "url": "https://github.com/kiegroup/kogito-runtimes/pull/793", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
732164743
KOGITO-3732 - monitoring refactoring: decouple from Prometheus JIRA: https://issues.redhat.com/browse/KOGITO-3732 Still a draft, description to be done This PR aims to decouple the monitoring addon from prometheus: the user will have the possibility to specify what is the tecnology to use to export the metrics generated by the kogito runtime. The existing monitoring-prometheus-addon will remain available with exactly the same interfaces that already exist: it is backcompatible. A new artifact monitoring-core will be available, and the user can import that so to inject the micrometer registy to use. For example, if the user would like to export the metrics to elastic, he/she will have to import the dependencies: <dependency> <groupId>org.kie.kogito</groupId> <artifactId>monitoring-core</artifactId> </dependency> <dependency> <groupId>io.micrometer</groupId> <artifactId>micrometer-registry-elastic</artifactId> </dependency> Then register and start the micrometer registry like import java.util.concurrent.Executors; import javax.annotation.PostConstruct; import javax.inject.Singleton; import io.micrometer.core.lang.Nullable; import io.micrometer.elastic.ElasticConfig; import io.micrometer.elastic.ElasticMeterRegistry; import io.quarkus.runtime.Startup; import org.kie.kogito.monitoring.core.MonitoringRegistry; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @Singleton @Startup public class ElasticProvider { private ElasticMeterRegistry registry; private static final Logger logger = LoggerFactory.getLogger(ElasticProvider.class); private ElasticProvider() { } @PostConstruct public void setUp() { ElasticConfig elasticConfig = new ElasticConfig() { @Override @Nullable public String get(String k) { return null; } }; registry = ElasticMeterRegistry.builder(elasticConfig).build(); MonitoringRegistry.addRegistry(registry); logger.info("Elastic registry added to monitoring addon."); runMe(); } private void runMe() { logger.info("Initializing elastic metrics publisher"); registry.start(Executors.defaultThreadFactory()); } } If the user would like to use Prometheus, the module <dependency> <groupId>org.kie.kogito</groupId> <artifactId>monitoring-prometheus-addon</artifactId> </dependency> is still available and the behaviour/interfaces are the same. The files PrometheusProcessEventListener, PrometheusMetrics, PrometheusMetricsDroolsListener have been moved to ProcessEventListener, RuleMetrics, RuleMetricsDroolsListener keeping the history (here on Github they are shown as new files, but if you git blame them you will see the history, as well as you can check the single commits). Then they were recreated to implement the wrapper. Many thanks for submitting your Pull Request :heart:! Please make sure that your PR meets the following requirements: [x] You have read the contributors guide [x] Pull Request title is properly formatted: KOGITO-XYZ Subject [x] Pull Request title contains the target branch if not targeting master: [0.9.x] KOGITO-XYZ Subject [x] Pull Request contains link to the JIRA issue [x] Pull Request contains link to any dependent or related Pull Request [x] Pull Request contains description of the issue [x] Pull Request does not include fixes for issues other than the main ticket Hi @cristianonicolai , I'm including you since I'm refactoring also the monitoring of processes Jenkins retest this please Hi @danielezonca @cristianonicolai @MarianMacik , I've some other PRs in the pipeline for the monitoring addon. Do you have any additional comment on this one? The PR overall is fine for me, some additional points not directly related with the code: Have you already verified if native compilation is still working (you can use the metrics example if you want) Have you verified if the operator is working fine (I expect so but just to be sure)? Do you expect to have any additional impact on operator side now that monitoring and dashboard generations are sort of independent? Jenkins retest this please Hi @danielezonca , my comments inline Have you already verified if native compilation is still working (you can use the metrics example if you want) I had to include a couple of classes for reflection. Now it works properly. Have you verified if the operator is working fine (I expect so but just to be sure)? Everything should be fine since we did not touch the generation of the dashboards. The operator simply looks for the /monitoring/dashboards/list.json file and if it exists, it fetches the dashboards. Do you expect to have any additional impact on operator side now that monitoring and dashboard generations are sort of independent? The generation of the dashboards is enabled if and only if the prometheus addon is enabled. The consequence is that the operator will fetch the dashboards only if the prometheus addon is enabled: currently this is the correct behaviour. Jenkins retest this please Jenkins retest this please jenkins retest this please jenkins retest this please Hi @danielezonca @cristianonicolai @MarianMacik , once the CI is green are we good to go with this? Jenkins retest this please Jenkins retest this please Jenkins retest this please Jenkins retest this please Jenkins retest this please Hi @cristianonicolai @MarianMacik, CI is green, could you please have a look and review/approve? Thanks! Hi @MarianMacik , I've renamed the files and removed the unused variables; thanks for the tips! Let me know if we are good to go Jenkins retest this please Hi @danielezonca @cristianonicolai @MarianMacik , can any of you merge this one together with its example https://github.com/kiegroup/kogito-examples/pull/433 ?
gharchive/pull-request
2020-10-29T09:51:10
2025-04-01T06:44:41.536219
{ "authors": [ "danielezonca", "r00ta" ], "repo": "kiegroup/kogito-runtimes", "url": "https://github.com/kiegroup/kogito-runtimes/pull/859", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2492784315
Unable to load weights from checkpoint file CogVideoX-5b doesn't work, it seem to be a file issue : There are actually 2 files on HuggingFace not 1 : Here's my attempt to rename the files using a hardlink but no luck. The smaller file (02 of 02) was already downloaded automatically so I tried that too. Am I supposed to combine them in some way maybe ? Note, I have 3090/24GB and the requirements are successfully installed : It's diffusers format so the model is the folder itself, not any of the files. There shouldn't be a .bin file at all, just what you see in the online folder. It's diffusers format so the model is the folder itself, not any of the files. There shouldn't be a .bin file at all, just what you see in the online folder. I understand but then why is it asking for a .bin file ? At first there was only 'diffusion_pytorch_model-00002-of-00002.safetensors' file with config.json, downloaded automatically. I added manually the other json and the first part 'diffusion_pytorch_model-00001-of-00002', hoping that it would help but it didn't. I removed the folder so that it can be downloaded again. Let's see what happens
gharchive/issue
2024-08-28T18:53:17
2025-04-01T06:44:41.548356
{ "authors": [ "kijai", "tetsuoo-online" ], "repo": "kijai/ComfyUI-CogVideoXWrapper", "url": "https://github.com/kijai/ComfyUI-CogVideoXWrapper/issues/27", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
816223435
The label is not evasive I ran this code and found that the annotations were not evasive yes, annotation is bad and you should use cesium to achieve this function
gharchive/issue
2021-02-25T08:50:47
2025-04-01T06:44:41.559791
{ "authors": [ "kikitte", "yiwenzhang666999" ], "repo": "kikitte/MVTImageryProvider", "url": "https://github.com/kikitte/MVTImageryProvider/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
105004727
Mark TabbyChat mod as clientSideOnly ...so it doesn't crash the server with SideOnly issues when installed on a server. Makes TabbyChat less of a pain to use in a dev environment too, as I then don't have to remove the mod every time I want to test my server build. Is that not covered by this? In a deployed .jar inside the mods directory possibly, but when using it as a library (for example when grabbing it via gradle dependency), and running it in a dev environment, the manifest isn't loaded.
gharchive/pull-request
2015-09-05T05:15:27
2025-04-01T06:44:41.562527
{ "authors": [ "blay09", "killjoy1221" ], "repo": "killjoy1221/TabbyChat-2", "url": "https://github.com/killjoy1221/TabbyChat-2/pull/33", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1244177686
Price 부분과 주사용처, 관련 site link 수정안 Price 부분에는 실물 가격이나 해당 상품의 선물 가격을 표기하고 sidebar애서 남는 부분은 전부 해당 상품과 관련된 여러 곁 가지 정보들 ex) 무역, 수출, 관세 추이 등을 보여주는 것이 좋을 것 같습니다. 주 사용처와 관련 site link는 저희가 중요시하는 직관성을 되려 해칠 수 있기에 곁 가지 정보들을 차트로 보여줘서 추세가 이렇게 된다는 사실을 보여주는 것이 더 좋다고 생각이 들었습니다. 다른 분들은 어떻게 생각하시나요? 만약 찬성하시면 제가 구현해 보겠습니다. 오프라인에서 이야기 한 대로 진행하겠습니다.
gharchive/issue
2022-05-22T08:05:26
2025-04-01T06:44:41.617720
{ "authors": [ "poeo1877" ], "repo": "kimnamhyeon0112/Open_Source_Project", "url": "https://github.com/kimnamhyeon0112/Open_Source_Project/issues/23", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
241112688
thank you i searched a long time for this, thanks for your work ! 👍 @michabbb Glad you like it! :smile:
gharchive/issue
2017-07-06T23:13:11
2025-04-01T06:44:41.628741
{ "authors": [ "fruitl00p", "michabbb" ], "repo": "kingsquare/docker-tunnel", "url": "https://github.com/kingsquare/docker-tunnel/issues/6", "license": "ISC", "license_type": "permissive", "license_source": "github-api" }
21507248
Create a page where a user can customize & download config.php With support for: Editor selection for IDE:// links in called from Permanent link a-la jQueryUI embeded so one can download the latest version at any time Framework selection for path replacement in called from display Customize shorthand functions (when they're implemented properly) Not applicable to 2.0
gharchive/issue
2013-10-24T08:06:56
2025-04-01T06:44:41.630407
{ "authors": [ "jnvsor", "raveren" ], "repo": "kint-php/kint", "url": "https://github.com/kint-php/kint/issues/87", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
630856279
Eliminate manual installation of Terraform dependencies Right now users have to install Terraform manually as well as the correct version of the ct provider in order for lokoctl to be able to manage clusters. This isn't very nice at the UX level and is also error-prone. We should consider shipping the correct Terraform binary with lokoctl and safely pin a known set of Terraform providers so that the correct providers are downloaded during the terraform init phase. The end result should be that a user can download a lokoctl binary and simply run lokoctl cluster apply to get a working cluster. This can be easily solved after we upgrade to Terraform 0.13, which will allow pulling providers from 3rd part registries. See https://github.com/hashicorp/terraform/blob/mildwonkey/ps-docs/website/docs/configuration/providers.html.md#provider-source for more details. ct provider is now automatically pulled from the registry. However, some platforms may not have a provider published (e.g. libvirt), then manual installation is still needed.
gharchive/issue
2020-06-04T14:05:37
2025-04-01T06:44:41.698586
{ "authors": [ "invidian", "johananl" ], "repo": "kinvolk/lokomotive", "url": "https://github.com/kinvolk/lokomotive/issues/556", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1132323612
:root output target does not work in Gutenberg editor Issue description: Any css that targets :root element works fine in frontend, but due to the prefixing of the target with .editor-styles-wrapper this functionality breaks in gutenberg editor. Rendered string will be: .editor-styles-wrapper :root{ in css found at: ?action=kirki-styles&editor=1&ver=4.0. Version used: 4.0.20 Using theme_mods or options? theme_mod Code to reproduce the issue (config + field(s)) This does not work \Kirki::add_field( \Municipio\Customizer::KIRKI_CONFIG, [ 'type' => 'multicolor', 'settings' => 'color_palette_primary', 'label' => esc_html__( 'Primary colors', 'municipio' ), 'section' => self::SECTION_ID, 'priority' => 10, 'transport' => 'auto', 'choices' => [ 'base' => esc_html__( 'Base', 'municipio' ), 'dark' => esc_html__( 'Dark', 'municipio' ), 'light' => esc_html__( 'Light', 'municipio' ), 'contrasting' => esc_html__( 'Contrastring', 'municipio' ), ], 'default' => [ 'base' => '#ae0b05', 'dark' => '#770000', 'light' => '#e84c31', 'contrasting' => '#ffffff', ], 'output' => [ [ 'choice' => 'base', 'element' => ':root', 'property' => '--color-primary', ], ], ] ); This do work: \Kirki::add_field( \Municipio\Customizer::KIRKI_CONFIG, [ 'type' => 'multicolor', 'settings' => 'color_palette_primary', 'label' => esc_html__( 'Primary colors', 'municipio' ), 'section' => self::SECTION_ID, 'priority' => 10, 'transport' => 'auto', 'choices' => [ 'base' => esc_html__( 'Base', 'municipio' ), 'dark' => esc_html__( 'Dark', 'municipio' ), 'light' => esc_html__( 'Light', 'municipio' ), 'contrasting' => esc_html__( 'Contrastring', 'municipio' ), ], 'default' => [ 'base' => '#ae0b05', 'dark' => '#770000', 'light' => '#e84c31', 'contrasting' => '#ffffff', ], 'output' => [ [ 'choice' => 'base', 'element' => 'div', 'property' => '--color-primary', ], ], ] ); Not the output array change. I did a quickfix with a filter, solution looks like this: ` /** * Fixes issue when using :root selector in * output args for Gutenberg editor. */ add_filter('kirki_municipio_config_dynamic_css', function ($styles) { $isEditor = (isset($_GET['editor']) && $_GET['editor'] == '1'); $isStyles = (isset($_GET['action']) && $_GET['action'] == 'kirki-styles'); if ($isEditor && $isStyles) { return str_replace( '.editor-styles-wrapper :root', '.editor-styles-wrapper', $styles ); } return $styles; }, 20); ` nice! I thought of something similar in core. We'll look into this. Definitely needs to be addressed. Hi @sebastianthulin , thanks for helping us finding the issue and the solution. It's implemented in this commit It will be shipped to the next release. Let me close this issue :) For me this is still an issue, I'm on 4.0.23. Generated output is .editor-styles-wrapper :root,.editor-styles-wrapper div
gharchive/issue
2022-02-11T10:41:01
2025-04-01T06:44:41.732221
{ "authors": [ "DeoThemes", "MapSteps", "contactjavas", "sebastianthulin" ], "repo": "kirki-framework/kirki", "url": "https://github.com/kirki-framework/kirki/issues/2461", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
64160039
Deployment workflow from Desktop to Server running Docker Is it possible to deploy a docker from you mac app? That would be pretty cool as I have no idea how to do this manually. Hi there! Thanks for the issue. There isn't a way to deploy containers to servers via Kitematic (yet!, there's an issue tracked here: #231). This seems to be similar to that issue, but with a spin on having a workflow from Laptop => Server so I'll keep this open :smiley: it would be great if kitematic GUI have the ability to install/create/deploy container to cloud servers, that would be awesome feature, everything would be automated from the GUI panel... is it possible to install kitematic into the cloud server? for installing/creating/deploying containers closing in favor of your other issue
gharchive/issue
2015-03-25T03:10:28
2025-04-01T06:44:41.756980
{ "authors": [ "08ski11s", "FrenchBen", "chovy", "jeffdm" ], "repo": "kitematic/kitematic", "url": "https://github.com/kitematic/kitematic/issues/330", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
308241378
Gem based theme problems So I tried once again to use the gem based version of your theme(latest 1.5.0), however it has two big problems which I can't seem to find a solution to(and I don't understand where things went wrong): Archive page is broken for some reason (tags are grayed out and remain unfocused). Files that I have changed that could lead to this issue are: _includes/components/tags.html(only translated Show All to Italian, shouldn't be an issue); _data/locale.yml (added Italian localization); _data/variables.yml (changed lang: it); /all.html (added Italian title). TOC is not showing up: I have an article with h3 headers(three #) and toc isn't showing up anymore. I've uncommented the selectors: 'h1,h2,h3' in _config.yml however it's not working. Everything works flawlessly with non-gem based theme version 1.4.3. I resolved the tags problem, it was a wrong indentation in the sources option. The TOC problem still remains tho. sources options? It's a bug ? No, it was a human error on my part, when I downloaded the new _config.yml for some reasons safari broke line breaks and didn't notice this one as well. Sorry about it.
gharchive/issue
2018-03-24T08:59:05
2025-04-01T06:44:41.761586
{ "authors": [ "kitian616", "maddovr" ], "repo": "kitian616/jekyll-TeXt-theme", "url": "https://github.com/kitian616/jekyll-TeXt-theme/issues/44", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1491194284
Switch to unprivileged user for nginx Further Notes Reviewing the Dockerfile, I found that the production nginx container is using the root user; for the sake of best practices this should be avoided, see https://github.com/nodejs/docker-node/blob/main/docs/BestPractices.md#non-root-user New Features / Enhancements [x] Avoiding root user privileges for production docker image => increase security [x] Exclude unnecessary files from docker image => decrease docker image file size After Merge Checklist Thanks for this pull request! I've still one more wish: Could you please add restart=always to docker compose file?
gharchive/pull-request
2022-12-12T07:31:35
2025-04-01T06:44:41.769984
{ "authors": [ "gerardo-navarro", "tim-krieger" ], "repo": "kitsteam/pictosearch", "url": "https://github.com/kitsteam/pictosearch/pull/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1174627257
Missing text MDFloatingActionButtonSpeedDial Description of the Bug Hi can some body help me please solve this mystery? :) When is hint_animation: True it will show only and always setting text but if is hint_animation: False the text appears but I don't know how to change the color Hintu . Code and Logs from kivymd.app import MDApp from kivy.lang import Builder class MossyApp(MDApp): buttons = { 'Materials': 'atom', 'Calculations': 'calculator-variant', 'Notes': 'fountain-pen-tip', 'Settings': "bitcoin" } def build(self): self.theme_cls.theme_style = 'Dark' # self.theme_cls.primary_colore = (0.968,.576,.101,1) return Builder.load_file('Mossy.kv') MossyApp().run() kv = """ MDBoxLayout: orientation: 'vertical' MDScreen: MDLabel: id: my_lable text: "Mossy" halign: "center" MDFloatingActionButtonSpeedDial: data: app.buttons root_button_anim: True hint_animation: True bg_color_stack_button: 0.968,.576,.101,1 bg_color_root_button: 0.968,.576,.101,1 bg_hint_color: 0.968,.576,.101,1 color_icon_root_button: .1,.1,.1,1 color_icon_stack_button: .1,.1,.1,1 label_text_color: .1,.9,.1,1 """ Screenshots Versions OS: win 10 Python: 3.9.7 Kivy: 2.1.0 KivyMD: 0.104.2 @MrPanco We do not support KivyMD version 0.104.2. Use the master branch of the library.
gharchive/issue
2022-03-20T17:44:01
2025-04-01T06:44:41.803192
{ "authors": [ "HeaTTheatR", "MrPanco" ], "repo": "kivymd/KivyMD", "url": "https://github.com/kivymd/KivyMD/issues/1210", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
346655610
Add optional serde support Two open questions: Right now for the sake of feature clarity, the feature is called serde, but that means people have to explicitly enable std as well if they disable all features, it ends up working out if they just use the default features tho. Should the deserializer be smarter and check the value is within bounds and fail deserialization? If you would have asked me what the benefits of being able to (de)serialize these types directly with serde instead of converting from native types i would assume the built in bounds checking was the main reason. Would you mind mentioning other reasons (if you have any) it would be nice to have serde support? It's just because I have structs that contain uX types, and I want to be able to use serde with them, checking the bounds on deserialization seems to be a must to maintain the invariants anyway. Perhaps this is clear when familiar with serde, but is it trivial to accept non byte aligned values for deserialization with serde from a binary format? I assume it's expected that the following struct should require 2 bytes of payload for deserialization (not 3 where a lot of the bits were padding which i suspect is the case when using derive). I guess being able to tightly pack bit fields would be another reason to want serde support for this library? Personally I'm just using these types to maintain invariants about the actual size of the fields, I'm probably going to end up using MessagePack, bincode and probably JSON to serialize/deserialize these structs, so I don't think it would be possible to tightly pack them right now. I'm also not entirely sure if serde even supports this kind of thing as an optional hint for serializers/deserializers. Another concern that needs to be addressed is stability. Even though serde is very stable, i guess its possible that a version 2 would be released some time in the future. Since this crate is mimicking the std library and aiming to be completely stable once a good way to handle construction is found we should have a strategy of avoiding to bump the version number even if serde needs to. Do you know how other crates plan for this? Having a serde2 would probably be catastrophic for the ecosystem, so I don't think it's going to happen any time soon, if ever. I'm going to add a custom deserializer that checks bounds. Okay, implemented Serialize as a forwarding serializer to the primitive type, and Deserialize as a forwarder with an additional bound check. I will be traveling from Aug 11. to Aug 18. I might be able to check in on github a few times but expect communication to be slow this week and the days after my return. I've created #17 to discuss everything related to serde support. I think we should be able to figure out how to solve the unresolved questions in the mentioned issue pretty quick. I have a few solutions in mind that i think would well. I hope that in the mean time, the above suggestion is acceptable for you. Even though we're hopefully able to bring the serde feature out of nightly pretty quick, we've discussed unstable features before (https://github.com/kjetilkjeka/uX/issues/5) and having the infrastructure there would be nice. Hope you don't mind implementing this. I'm not really sure what we can expect when deriving serde traits on a struct containing these fields. I think it would be beneficial of having a way to derive packed serialization code. There's no support for this in serde at all, and most serialization formats do not have a concept of bitfields either, it might work for bincode but most formats don't support that. Create a nightly cfg attribute using build.rs and https://crates.io/crates/rustc_version What is the purpose of making it nightly only? serde works on stable. There's no support for this in serde at all, and most serialization formats do not have a concept of bitfields either, it might work for bincode but most formats don't support that. Does that mean that result from bincode is unspecified except that a serialized value will be deserialized back to the same value? Does that also mean that new format where binary encoding is important will be impossible to add for serde? Or is it possible to opt out of supporting formats based on binary representations for the time being? Leaving it to be figured out in the future would be totally acceptable, implementing it in an unfortunate way is not. What is the purpose of making it nightly only? serde works on stable. If we're going to publish this crate it must adhere to the rust/semver properties, meaning that the serde feature must be compatible to future versions. Well the former is true when compiled on stable, when using nightly rust all bets are off. This allow us to publish a crate with an unstable feature as long as it's only accessible on nightly. I think it's a great idea to have the serde feature, but I want to make sure that we "get it right". Making the nightly only feature was an offer to get this PR merged early since I already wanted to have a nightly only options for this crate. If you prefer to sort these things out before merging that's totally fine as well. I also want to leave it hanging for some amount of time to allow more people (that have more experience with serde than me) to have a look at it before merging, and giving me some time to experiment with serde as well. I understand that you probably need this feature for your work, and the parts you're going to use should not change after this PR is merged anyway. This means that if you're able to use nightly for the part of your work using this feature you should be able to not do anything after the feature is stabilized (except moving over to stable again). If you prefer to track your own fork rather than crates.io for the time being, you may discard my review comments relating to the nightly feature. Seems like a great add! @kjetilkjeka: Do you intend to merge it? I didn't understand any of your points against it... @kjetilkjeka: Do you intend to merge it? I didn't understand any of your points against it.. Absolutely! I suspect my concerns are related to me not being familiar with serde, and either way I'm sure they're absolutely solvable. But since I'm away from 11. Aug to 18. Aug I didn't really have the time to resolve them before going away. I will try to get to this as soon as possible when I'm back. If I cannot resolve my concerns related to binary formats quickly when I'm back home I will write a post to explain them further. I still want the following minor changes before merging: I think we should change the custom error to invalid error with an informative expected, am I wrong here? Implement serde in a seperate file. Then the serde feature can be used on the mod declaration instead of "everywhere". Document the serde feature in the README.md or as doc comments in lib.rs ( If @meh prefers to leave this to me, this is fine) Implement test that confirms that deserialization of out of bounds numbers throw the correct error. Create entry in changelog ( If @meh prefers to leave this to me, this is fine ) I think it's really great to get contributions to this library, and serde support seems quite important in making it ergonomic to use this library. I'm sorry my unfamiliarity with serde and my trip is delaying this PR. Can this issue be revisited? ux is being consumed by x86_64_types and we'd like to use it in rust-vmm. Having derived serde traits is super helpful to us. For reference, I've published a fork with just this PR applied as ux_serde: https://crates.io/crates/ux_serde For reference, I've published a fork with just this PR applied as ux_serde: https://crates.io/crates/ux_serde
gharchive/pull-request
2018-08-01T15:39:07
2025-04-01T06:44:41.958899
{ "authors": [ "GuillaumeGomez", "kjetilkjeka", "meh", "npmccallum", "trevyn" ], "repo": "kjetilkjeka/uX", "url": "https://github.com/kjetilkjeka/uX/pull/16", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
927473525
Parsing fails with NullPointerException Hi Kim, I have a very strange problem here with: zprint <<< '(is (= #::sut{:groups #{"FOO_ADMIN" "FOO_USER"}} (sut/map->user {:groups ["FOO_USER" "FOO_ADMIN"]})))' (is (= #::sut{:groups #{"FOO_ADMIN" "FOO_USER"}} (sut/map->user {:groups ["FOO_USER" "FOO_ADMIN"]}))) Failed to zprint: java.lang.NullPointerException Now if I remove the #::sut everything is good: zprint <<< '(is (= {:groups #{"FOO_ADMIN" "FOO_USER"}} (sut/map->user {:groups ["FOO_USER" "FOO_ADMIN"]})))' But: zprint <<< '#::sut{:groups #{"ADMIN" "USER"}}' #::sut{:groups #{"ADMIN" "USER"}} Therefore I am not sure it is the namespaced map syntax that's breaking there. I have a workaround - not to use that syntax solves it in some place but I thought I should report it :smile: Thanks as usual for your hard work on zprint! Thanks for noticing this! Namespaced map syntax does not play well with the currently shipping version of rewrite-clj, but the new version (V1) by @lread has fixed this, so it will work in the next zprint release: 1.2.0. Assuming that I work up the ... to release zprint with an alpha version of rewrite-clj V1. But it runs all 1200+ zprint tests, so why not? Go for it! 🙂 I don't expect issues to be caused by rewrite-clj v1, but if we find some, I'll be happy address them. This is now fixed by 1.2.0, which includes rewrite-clj V1!
gharchive/issue
2021-06-22T17:29:31
2025-04-01T06:44:41.985828
{ "authors": [ "arichiardi", "kkinnear", "lread" ], "repo": "kkinnear/zprint", "url": "https://github.com/kkinnear/zprint/issues/199", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1028883769
Malformed parsing for non quoted attributes. In instances where attributes are not within quotes it breaks the parsing. For example the src attribute: <img class="img-fluid" src=https://www.example.com/banner-1.jpg > I understand this is invalid html, but is quite common on the web to have quoteless data attributes. Thank you for trying parsercher. I also appreciate the information provided. I have released a modified v3.1.5. I've tested it, but if it doesn't work, please let me know again. github: https://github.com/kkmtyyz/parsercher/releases/tag/v3.1.5 crates.io: https://crates.io/crates/parsercher/3.1.5 Thanks so much for the quick fix much appreciated.
gharchive/issue
2021-10-18T09:44:02
2025-04-01T06:44:41.992739
{ "authors": [ "JSONhilder", "kkmtyyz" ], "repo": "kkmtyyz/parsercher", "url": "https://github.com/kkmtyyz/parsercher/issues/2", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1070883103
Histogram functions are not found. let image = vips.Image.newFromBuffer(arrayBuff); image = image.hist_norm() image.hist_norm() results in error is not a function. Never mind. My mistake I needed to use javascript function image.histNorm()
gharchive/issue
2021-12-03T19:20:41
2025-04-01T06:44:42.026001
{ "authors": [ "delebash" ], "repo": "kleisauke/wasm-vips", "url": "https://github.com/kleisauke/wasm-vips/issues/11", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
748678323
Session data is too long for default String type when using mysql Hi, love the project! Was very happy to discover I could keep my session store within the prisma ecosystem. I wanted to flag a problem I had getting set up with mysql: I was finding the following error when creating a session: Invalid `prisma.session.update()` invocation: The provided value for the column is too long for the column's type. Column: data The prisma String type defaults to a VarChar of length 191 which is what I suspect the problem was. If I manually update the Session.data column to LongText then the problem goes away: ALTER TABLE Session MODIFY data LONGTEXT; Prisma are currently working on adding schema annotations that can be used to control which underlying type your schema will map to: https://www.prisma.io/docs/concepts/components/preview-features/native-types/. However using prisma migrate with the new annotations is not currently supported, they are working on it though: https://github.com/prisma/prisma-engines/issues/72. Anyways I thought it would be good to bring this up in case you wanted to add some notes to your README (or I could make a PR for my first open-source contribution 🤭). Even just having this issue to refer to might be enough to help mysql users on their way, up to you 😄. If you have any suggestions for the type to use for the data column that would be great, or if there is a way of changing the schema so that no manual intervention is needed that would be amazing (but it looks like it's probably not worth it long-term 🤷‍♂️ ). Again, thanks for the project, after the above it's working like a charm. 👍 If you could make a PR under Setup > MYSQL in the README.md that would be great. If you could make a PR under Setup > MYSQL in the README.md that would be great. Brilliant, I'll get on that tonight 👍 Hi Oliver, Glad your finding this project beneficial. I'm on the road at the moment, but will tend to this when I can. -K On Nov 23, 2020, at 2:35 AM, Olivier Wilkinson notifications@github.com wrote:  Hi, love the project! Was very happy to discover I could keep my session store within the prisma ecosystem. I wanted to flag a problem I had getting set up with mysql: I was finding the following error when creating a session: Invalid prisma.session.update() invocation: The provided value for the column is too long for the column's type. Column: data The prisma String type defaults to a VarChar of length 191 which is what I suspect the problem was. If I manually update the Session.data column to LongText then the problem goes away: ALTER TABLE Session MODIFY data LONGTEXT; Prisma are currently working on adding schema annotations that can be used to control which underlying type your schema will map to: https://www.prisma.io/docs/concepts/components/preview-features/native-types/. However using prisma migrate with the new annotations is not currently supported, they are working on it though: prisma/prisma-engines#72. Anyways I thought it would be good to bring this up in case you wanted to add some notes to your README and I figured even just having this issue to refer to might be helpful. If there is a way of changing the schema so that no manual intervention is needed that would be amazing but it looks like it's probably not worth it long-term. Again, thanks for the project, after the above it's working like a charm. 👍 — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub, or unsubscribe. Woops! Just saw wSedlacek's reply. -K On Nov 23, 2020, at 8:24 AM, William Sedlacek notifications@github.com wrote:  If you could make a PR under Setup > MYSQL in the README.md that would be great. — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub, or unsubscribe. :tada: This issue has been resolved in version 1.1.1 :tada: The release is available on: npm package (@latest dist-tag) GitHub release Your semantic-release bot :package::rocket:
gharchive/issue
2020-11-23T10:33:26
2025-04-01T06:44:42.049119
{ "authors": [ "kleydon", "olivierwilkinson", "wSedlacek" ], "repo": "kleydon/prisma-session-store", "url": "https://github.com/kleydon/prisma-session-store/issues/21", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1680094188
Missing X11 Window Flags I noticed some missing window flags such as SDL_WINDOW_VULKAN which is presumably due to not supporting Vulkan (yet?) but the following are X11 and should probably be supported: SDL_WINDOW_ALWAYS_ON_TOP SDL_WINDOW_SKIP_TASKBAR SDL_WINDOW_UTILITY SDL_WINDOW_TOOLTIP SDL_WINDOW_POPUP_MENU Additionally, there doesn't seem to be a helper function (ie. Window.getFlags()) to easily get the flags as an object. I would love seeing those added. Thanks for the great work so far! @Blade67 Thanks for the request. In v0.7.1 I have added the alwaysOnTop, skipTaskbar, popupMenu, tooltip, and utility flags to createWindow(). For getting the various flags, use the corresponding properties of the window object. I'm not sure how vulkan can be supported from javascript. My plan was to just use WebGPU. I have made some progress over here.
gharchive/issue
2023-04-23T16:58:32
2025-04-01T06:44:42.058918
{ "authors": [ "Blade67", "kmamal" ], "repo": "kmamal/node-sdl", "url": "https://github.com/kmamal/node-sdl/issues/29", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1586816448
add load areas to PJM get_load Hi all, Excited to have found this project! I hope I can contribute as I learn the code better, but for now, a quick upvote on adding the "load areas" underneath PJM, as noted in the code here: https://github.com/kmax12/gridstatus/blob/main/gridstatus/pjm.py#L111 Thanks again! this is been on my todo list, so thanks for the nudge. started an implementation here: #160. will likely need to polish off tomorrow before merging. keep the suggestions coming and I hope you can contribute sometime as well!
gharchive/issue
2023-02-16T00:52:17
2025-04-01T06:44:42.074209
{ "authors": [ "kmax12", "mconlow" ], "repo": "kmax12/gridstatus", "url": "https://github.com/kmax12/gridstatus/issues/158", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1698055971
Update sbt-sonatype to 3.9.20 About this PR 📦 Updates org.xerial.sbt:sbt-sonatype from 3.8.1 to 3.9.20 📜 GitHub Release Notes - Release Notes Usage ✅ Please merge! I'll automatically update this PR to resolve conflicts as long as you don't change it yourself. If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below. Configure Scala Steward for your repository with a .scala-steward.conf file. Have a fantastic day writing Scala! ⚙ Adjust future updates Add this to your .scala-steward.conf file to ignore future updates of this dependency: updates.ignore = [ { groupId = "org.xerial.sbt", artifactId = "sbt-sonatype" } ] Or, add this to slow down future updates of this dependency: dependencyOverrides = [{ pullRequests = { frequency = "30 days" }, dependency = { groupId = "org.xerial.sbt", artifactId = "sbt-sonatype" } }] labels: sbt-plugin-update, early-semver-minor, semver-spec-minor, commit-count:1 Superseded by #69.
gharchive/pull-request
2023-05-05T18:57:42
2025-04-01T06:44:42.095259
{ "authors": [ "scala-steward" ], "repo": "kmizu/pegex", "url": "https://github.com/kmizu/pegex/pull/68", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2459298406
Fixed typos in comments and docs Some instances of "kmonad" have been left unchanged (capitalization) since they (probably) refer to the executable. Thanks!
gharchive/pull-request
2024-08-10T21:42:02
2025-04-01T06:44:42.103135
{ "authors": [ "jokesper", "slotThe" ], "repo": "kmonad/kmonad", "url": "https://github.com/kmonad/kmonad/pull/870", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
743755602
Several ATECC608A manifest Hello: I have bought 100 ATECC608A for 100 diferent board, I have a one manifest for all chips, and you scripts get the pem for each chip. I 'm not very clear but upload each .pem file is the right way for works with this chips? There is no better way for example upload the full manifest file to Azure? Best Regards Thank you for your checking. This was created for testing a few devices, and I think the program on the manufacturer's site will help you register many certificates. It seems that it is possible to automatically enroll multiple certificates listed in the manifest file with the AWS CLI. https://github.com/MicrochipTech/cryptoauth_trustplatform_designsuite/blob/master/TrustnGO/05_cloud_connect/notebook/aws/Microchip_manifest_handler.py Regards, Kentaro Mitsuyasu Sorry, I had a mistake. for Azure, I'm looking at the code below for auto-registration of multiple devices. https://github.com/MicrochipTech/cryptoauth_trustplatform_designsuite/blob/master/TrustnGO/05_cloud_connect/notebook/azure/helper_azure.py Hello I I have finally managed to upload the manifest to Azure with the helper_azure.py program and with the .crt file that Microchip has published on the page of the new ATECC608B device. When trying to connect to Azure I get error But if I do individual enrollment as you explain in your document, everything works correctly, what can be the difference between the two ways of registering the device? Thanks Hello I found at the Device Explorer utility two diferent line for the my device. very strange. Hello I think that the problem is that the helper_azure.py register the devices at the IoThub not at the DPS service, the I need to use iotHubClientHandle = IoTHubClient_LL_CreateFromConnectionString(connectionString, MQTT_Protocol)) and not use the DPS service device_ll_handle = IoTHubDeviceClient_LL_CreateFromDeviceAuth(user_ctx.iothub_uri, user_ctx.device_id, iothub_transport) Then, I need to remove de DPS service for you program and look for the device_id and device_key to build the connectionstrint Hello: Sorry I have wrong. I have to use the device_ll_handle = IoTHubDeviceClient_LL_CreateFromDeviceAuth(user_ctx.iothub_uri, user_ctx.device_id, iothub_transport) but the url is the IotHub url, not de DPS service url, and the device_id is the chip serial number. Ah, and also remove the DPS service from your program. Thanks Hello. Yes, as you've noticed, my project first registers the certificate with Azure DPS and provisions it to Azure IoT Hub. If the provisioning is successful, discard the connection handle and try connecting to Azure IoT Hub again. I misunderstood the Microchip sample. It is the operation of registering the certificate directly to Azure IoT Hub. In that case, you should connect directly to Azure IoT Hub instead of DPS.
gharchive/issue
2020-11-16T11:23:23
2025-04-01T06:44:42.112596
{ "authors": [ "JAHTKELD", "kmwebnet" ], "repo": "kmwebnet/ECC608-TNG-Azure-DPS-Connect", "url": "https://github.com/kmwebnet/ECC608-TNG-Azure-DPS-Connect/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1909311697
Added UX WG to automation Changes Add knative/ux to the knobots automation, specifically for community file sync, markdown checks, and spell checks Currently, the community files are not being synced to knative/ux, which makes it hard to keep the approvers synced with knative/community Release Note Community files are now synced to the UX WG. /cc @dprotaso /lgtm /approve If you have issues with this reach out to the productivity folks in #knative-productivity channel
gharchive/pull-request
2023-09-22T17:56:18
2025-04-01T06:44:42.134949
{ "authors": [ "Cali0707", "dprotaso" ], "repo": "knative-extensions/knobots", "url": "https://github.com/knative-extensions/knobots/pull/335", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
705812003
Codify the relationship between knative-sandbox/.github and knative/.github This PR does three things: Update the README to describe the relationship between knative-sandbox/.github and downstream .github repositories. Add a PR check that fails when PRs are opened against downstream .github repositories. Add a cron to mirror upstream into downstream. Fixes: https://github.com/knative-sandbox/.github/issues/38 /hold When this same PR is attempted against downstream it yields: /unhold Alright, I have pushed this (and the Go 1.15 change) to knative/.github so from here on out, everything should be automatic.
gharchive/pull-request
2020-09-21T18:12:52
2025-04-01T06:44:42.138481
{ "authors": [ "mattmoor" ], "repo": "knative-sandbox/.github", "url": "https://github.com/knative-sandbox/.github/pull/39", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
960462543
[main] Update actions Cron -knative-prow-robot /cc knative-sandbox/source-wg-leads /assign knative-sandbox/source-wg-leads Produced by: knative-sandbox/knobots/actions/update-actions Details: /lgtm /approve
gharchive/pull-request
2021-08-04T13:34:58
2025-04-01T06:44:42.140108
{ "authors": [ "knative-automation", "lionelvillard" ], "repo": "knative-sandbox/discovery", "url": "https://github.com/knative-sandbox/discovery/pull/210", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1085350233
[main] Update actions Cron -knative-prow-robot /cc knative-sandbox/channel-wg-leads /assign knative-sandbox/channel-wg-leads Produced by: knative-sandbox/knobots/actions/update-actions Details: /lgtm /approve
gharchive/pull-request
2021-12-21T01:38:58
2025-04-01T06:44:42.141663
{ "authors": [ "benmoss", "knative-automation" ], "repo": "knative-sandbox/eventing-rabbitmq", "url": "https://github.com/knative-sandbox/eventing-rabbitmq/pull/557", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1154755810
[main] Update actions Cron -knative-prow-robot /cc knative-sandbox/networking-wg-leads /assign knative-sandbox/networking-wg-leads Produced by: knative-sandbox/knobots/actions/update-actions Details: /lgtm /approve
gharchive/pull-request
2022-03-01T01:56:25
2025-04-01T06:44:42.143203
{ "authors": [ "knative-automation", "nak3" ], "repo": "knative-sandbox/net-certmanager", "url": "https://github.com/knative-sandbox/net-certmanager/pull/364", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
643875820
[master] Auto-update dependencies Produced via: ./hack/update-deps.sh --upgrade && ./hack/update-codegen.sh /assign davidor markusthoemmes /cc davidor markusthoemmes /lgtm
gharchive/pull-request
2020-06-23T14:12:05
2025-04-01T06:44:42.177310
{ "authors": [ "jmprusi", "mattmoor" ], "repo": "knative/net-kourier", "url": "https://github.com/knative/net-kourier/pull/108", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1369157308
[main] Upgrade to latest dependencies Cron -knative-prow-robot /cc knative/operations-wg-leads /assign knative/operations-wg-leads Produced by: knative-sandbox/knobots/actions/update-deps Details: bumping k8s.io/api f89a761...44d27eb: > 44d27eb Update dependencies to v0.24.4 tag > 0bf1867 Revert "Introduce APIs to support multiple ClusterCIDRs (# 108290)" > 2de6996 Merge pull request # 109241 from ravisantoshgudimetla/sts-ar-optional > 7734d26 [sts] api: Make available replicas optional > 38ec09a [sts] Generated: Make available replicas optional > ec84bcb Merge pull request # 109178 from liggitt/openapi-master > e3797f2 Drop enum tag from certificate request condition > 02c2207 Merge pull request # 109151 from Argh4k/r-101566 > 1eb735b Revert "Field `status.hostIPs` added for Pod (# 101566)" > 290a349 Introduce APIs to support multiple ClusterCIDRs (# 108290) > b754a94 Merge pull request # 109031 from Jefftree/openapiv3beta > a729cc1 generated: Update kube-openapi and vendor > 6031be5 Merge pull request # 108736 from NetApp/any-volume-data-source-beta > e78ec14 Merge pull request # 108032 from deejross/kep3140-cronjob-timezone > 5b36872 Update AnyVolumeDataSource feature gate to beta > 6519413 Merge pull request # 107476 from alculquicondor/job-ready > 4d603a5 make update after timeZone support for CronJob > ed1c913 Field `status.hostIPs` added for Pod (# 101566) > 135d041 Graduate JobReadyPods to beta > 3bd5d7d TimeZone support for CronJobs > f457c96 Implementation on Network Policy Status (# 107963) > b091752 Merge pull request # 82162 from krmayankk/maxun > e43e442 Merge pull request # 108282 from sftim/20220222_reword_types_container_runtime_neutral > 6135a28 API: maxUnavailable for StatefulSet > 9431395 Merge pull request # 108996 from cici37/errUpdate > 0d1146f Update generated docs > b0daa7a Bump kube-openapi > 9d3e6aa Avoid using Docker Engine as example for container runtime > 92515b8 Handle Non-graceful Node Shutdown (# 108486) > 24067bf Update types to be neutral about container runtime > df53a95 Update kube-openapi (# 108895) > 252596f Merge pull request # 108725 from pohly/klog-flush-and-exit > eba027b Merge pull request # 107859 from ravisantoshgudimetla/promote-PodOS-beta > 31a68ac klog v2.60.1 > b14ef06 Merge pull request # 108522 from SergeyKanzhelev/grpcToBeta > 2a29e32 Generated: PodOS field to beta > 20a98db Merge pull request # 107395 from alculquicondor/indexed-job > b956cca promote grpcProbes to beta > c3242c8 Merge pull request # 108445 from pohly/storage-capacity-ga > 1864916 Graduate IndexedJob to stable > 1f158ba Merge pull request # 108782 from cfryanr/expirationseconds_conformance > b78eb95 storage capacity: generated files > e8dffa9 Promote CertificateSigningRequest's Spec.ExpirationSeconds field to GA > e9a2988 storage capacity: clarify relationship between capacity fields > 0f1a9d7 Merge pull request # 108441 from pacoxu/pod-overload-ga > 4d4dca7 storage capacity: v1 API > b9830ac Merge pull request # 108644 from Jefftree/googleapis-gnostic > e8cba0c mark PodOverhead to GA in v1.24; remove in v1.26 > 045528e storage capacity: GA, always enabled, remove feature check > 939bda2 googleapis/gnostic -> google/gnostic > 17c9f73 Merge pull request # 108639 from jiahuif-forks/feature/openapi/node-condition-type-non-enum > fcdc547 polish comments of non-enum values. > c1d9c2b unmark non-validated types as enums. > ee4a762 Merge pull request # 107674 from sanposhiho/api-min-domains > 99de566 Add MinDomains API to TopologySpreadConstraints field > b8c40e0 Merge pull request # 108365 from liggitt/fix-proto-gen > a7a04f7 Regenerate protobuf > 860906f Merge pull request # 103061 from SergeyKanzhelev/removeAlphaRuntimeClass > cf872a4 Merge pull request # 108280 from liggitt/secrets > 2cb3498 Remove RuntimeClass feature gate as it was GA-ed > 35b1d68 Update secrets field API doc > 1b1f1b7 Merge pull request # 107105 from justaugustus/go118 > fedc50b Merge pull request # 108219 from liggitt/api-compatibility-fuzzer > 6a7745d generated: Run hack/lint-dependencies.sh and hack/update-vendor.sh > ef99526 Merge pull request # 108136 from ahg-g/ahg-affinity > b9d60f7 Regenerate HEAD compatibility fixtures > 5ab0434 Merge pull request # 106630 from howardjohn/protogen/full-package-name > f4d2900 Graduate PodAffinityNamespaceSelector to GA > 42578ae Merge pull request # 108129 from ahg-g/ahg-suspend > 402c109 go-to-protobuf: regenerate with full go_package > 4097bdc Merge pull request # 108138 from liggitt/v1beta1-selector > 52c1bbc generated files > 0ce2a88 Revert v1beta1 PodDisruptionBudget select patchStrategy > fc54d6a Graduate SuspendJob to GA > 29fd43e Merge pull request # 108059 from RaghavRoy145/k8s-utils-update > 9f54b3e Updated k8s.io/utils dependency > 64a9060 Merge pull request # 108055 from liggitt/api-compatibility > b2746bc Remove unused API compatibility fixture files > 816e830 Add hpa v2 compatibility fixtures > 8fb9658 Merge pull request # 107691 from SubhasmitaSw/issue_31393 > 67a3496 Merge pull request # 107432 from denkensk/graduate-nonpreemptingpriority-to-ga > 85611ae removed references to Docker in Kubernetes API > 8a75781 Merge pull request # 107656 from dims/add-labels-when-there-are-sig-aliases-used-in-approvers-reviewers > 583fc80 graduate nonpreemptingpriority to ga > b2d630a Merge pull request # 106932 from SergeyKanzhelev/removeDynamicKubeletConfig > dc08a5d Add labels when there sig aliases used in approvers/reviewers > e6d62dd Merge pull request # 105632 from xens/fix/kubectl-socks5-proxy2 > 107b310 remove DynamicKubeletConfig logic from kubelet > 1d22b5b feat: add missing SOCKS5 features > 35d41aa Merge pull request # 105142 from pacoxu/cespare-2.1.2 > 0ade255 upgrade prometheus/client_golang to v1.12.0(common to v0.32.1) > 94676c7 Merge pull request # 107603 from jayesh-srivastava/IANA-links > 5be450b Update IANA links > 37c9308 Merge pull request # 107565 from jiahuif-forks/deps/structured-merged-diff > 13861f0 generated: ./hack/update-vendor.sh > 688ca8d Merge pull request # 107529 from humblec/vol-fields > a75b0b5 Merge pull request # 107293 from dims/jan-1-owners-cleanup > 2240bf0 storage: correct struct fields in volume plugins > 6b8e940 Merge pull request # 107466 from humblec/new-field > 15e2172 Add mwielgus back > 58b7590 correct struct fields in various core storage structs > 204fd46 OWNERS cleanup - Jan 2021 Week 1 > 0bfaff8 Merge pull request # 107441 from humblec/more-pv-source > 5b5d5de correct struct fields in VolumeSource and PersistentVolumeSource > 37748cc Merge pull request # 107385 from humblec/csipvsource > 274b9af csi: correct struct fields in CSI{Persistent}VolumeSource. > 7289fed Merge pull request # 107235 from Nordix/deprecate-loadbalancerip > db00d7c Update generated files > 38141dd Deprecate service.spec.LoadBalancerIP > e869828 Merge pull request # 106643 from cyclinder/addresses_in_endpointslice > 18d22ba Merge pull request # 107103 from pohly/log-klog-update > f45808c fix: documenting that kube-proxy only use the first address in each endpoint > 236866c Merge pull request # 107027 from Nordix/disable-nodeport-ga > 4f891e5 dependencies: update klog to v2.40.1 > fcebee2 Update generated files after ServiceLBNodePortControl->GA > 6c48d34 Mark ServiceLBNodePortControl as GA > 038a002 Merge pull request # 106436 from dims/cleanup-owners-files-no-activity-in-a-year > b7112c1 Merge pull request # 107004 from liggitt/flowcontrol-beta > f1dacd8 Cleanup OWNERS files (No Activity in the last year) > 3dfa338 Point flowcontrol users at v1beta2 > 0820d15 Merge pull request # 106568 from liggitt/json-fieldpath > d992e83 Merge pull request # 106887 from liggitt/hpa-beta > 248e099 bump sigs.k8s.io/json > b3fcc5e Merge pull request # 106949 from cpanato/update-golang-xnet > a7f63e3 Direct v2betaX users to migrate to HPA v2 > 217eb4a Merge pull request # 106458 from dims/lint-yaml-in-owners-files > 32d2b15 dependencies: Update golang.org/x/net to v0.0.0-20211209124913-491a49abca63 > e092618 Merge pull request # 106921 from dims/update-x/tools-to-v0.1.8 > 7ee1208 Check in OWNERS modified by update-yamlfmt.sh > ef29486 Update golang.org/x/tools to a specific tag and avoid SHA > f3ee229 Merge pull request # 106850 from MadhavJivrajani/deprecate-clock-pkg > 969edbc Bump k8s.io/utils > ec2f4f7 Merge pull request # 106473 from Dingshujie/fix_inotify_memory_leak > cc1590f Merge pull request # 106868 from pacoxu/v1.23.0-api-testdata > b155431 update k/utils to v0.0.0-20211116205334-6203023598ed > 669b4af Remove 1.21.0 API test data > 23fb2ef Add v1.23.0 API testdata > 25b7aa9 staging: add dummy commit to trigger gomod update (# 106794) > 1a73729 Merge pull request # 106774 from SergeyKanzhelev/grpcFieldRename > ed38fc4 generated files for the grpc field rename > baa6f4b update the grpc field name for consistency > 4c321cf Merge pull request # 106660 from liggitt/smd-merge > 3b613a0 Revert sigs.k8s.io/structured-merge-diff/v4 to v4.1.2 bumping k8s.io/apiextensions-apiserver cf300a6...b993e22: > b993e22 Update dependencies to v0.24.4 tag > f6dccd8 Merge pull request # 110042 from Jefftree/automated-cherry-pick-of-# 109880-upstream-release-1.24 > b82ec0b Remove warning log for merging meta and scale type > b34b607 Merge pull request # 109242 from cici37/addTest > 1ad114b Merge pull request # 109238 from jpbetz/reduce-time-cel-validation-tests > a205029 Add separate cost limit test. > 8ec5404 Only validate each expression once in primary CEL correctness suite > b94fca5 Merge pull request # 109128 from Jefftree/openapiv3-metrics > 16d550b Merge pull request # 109122 from jpbetz/runtime-limit-tightening > 62f765a Add metrics for OpenAPI v3 generation > 1221011 Merge pull request # 109031 from Jefftree/openapiv3beta > dfe9350 Reduce CEL runtime cost limits by 1/2 based on latency goals > f350dab Separate OpenAPI V2 and V3 Config > d05111e Update tests for kube-openapi upgrade > c671030 generated: Update kube-openapi and vendor > 426e20e Merge pull request # 108126 from sanposhiho/doc/generatedname > 810bf54 Merge pull request # 108990 from jpbetz/per-crd > 205af0a fix the doc about generateName conflict > ae7a078 Merge pull request # 108629 from skitt/blang-semver-v4 > 9240dba Enfoce per-CRD estimated cost limit > b5908bd Merge pull request # 108612 from DangerOnTheRanger/cel-crd-maxlength > 068f21d Upgrade to blang/semver/v4 v4.0.0 > a17b9ba Add per-CRD cost evaluation. > 6b99f74 Merge pull request # 108996 from cici37/errUpdate > 97bfd37 Update err handling > abe11f3 Bump kube-openapi > 76c7ff3 Merge pull request # 109040 from jpbetz/revert-cel-bump > 21e0a14 Merge pull request # 108859 from cici37/placeholder > 5021b3f Revert "Initialize a base CEL env and share it to avoid repeated function declaration validation" > 8f35bcf Merge pull request # 108954 from jpbetz/cel-0.11.1 > a18d2aa Update due to inconsistent message got from kube-openapi > 6ae144f Revert "Bump CEL to 0.11.2" > be2a6a0 Initialize a base CEL env and share it to avoid repeated function declaration validation > d3fef43 Update err handling > f8dc6df Bump CEL to 0.11.2 > 899817b Return a placeholder error for blocking failure before CEL validation. > 6c890e6 Merge pull request # 108073 from benluddy/cel-transition-rule-oldself-plumbing > fc90687 Add validation rule tests for transition rules > 6fe18a4 Allow and enforce CEL CRD validation rules that reference oldSelf. > 6ab3fd1 Update kube-openapi (# 108895) > afe8ea9 Merge pull request # 107090 from cici37/reduceNoise > 808447f Merge pull request # 106591 from Kartik494/bumpetcd > e0c3d86 Return type instead of object > 5799d5a Merge pull request # 108725 from pohly/klog-flush-and-exit > b4c2bc2 Updated Etcd Version to 3.5.1 in go.mod > 8f25f22 Merge pull request # 108717 from lavalamp/remove-clustername > 0bf7876 klog v2.60.1 > 39229d8 Merge pull request # 108950 from jpbetz/fix-cel-test > 6e966cf generated files > b9cbe6c Remove flaky CEL cost tests that rely on map iteration order > a599743 Merge pull request # 108907 from Jefftree/crd-openapi-bug > 4812e51 Merge pull request # 108529 from cici37/celCostStability > 0a88d0b Fix OpenAPI v3 bug with #/definitions > b776ad9 Merge pull request # 103516 from ykakarap/kubectl-subresources-apiserver > fab2d1a Adjust cost after library func introduced. > faa1359 Merge pull request # 108617 from jpbetz/function-ext-costs > c65c485 kubectl: API changes to support --subresource in kubectl > 91bff26 Add cel cost stability test. > 5772ec0 Add cost calculator for extensions functions, enable regex pre-compilation > 31c15bc Merge pull request # 108772 from aojea/user_agent > 2b7cf69 client-go: update generated > 855f37e Merge pull request # 108646 from cici37/context > 3385c39 Merge pull request # 108644 from Jefftree/googleapis-gnostic > eff5693 Wire in request context > adf04b6 Merge pull request # 108419 from DangerOnTheRanger/cel-maxlength-integration > e2b78f2 googleapis/gnostic -> google/gnostic > c907672 Merge pull request # 108695 from cici37/errMessage > 1b50e07 Add maxLength/maxItems/maxProperties support to cel.Compile. > ab086dd Fix err handling due to cel update > 4846ad6 Merge pull request # 108595 from cici37/updateLimit > a2327b5 Update cost budget > b111587 Merge pull request # 108482 from cici37/vendorCEL > feee012 Merge pull request # 108643 from liggitt/cobra > 963365e Wire runtime cost into validation. > 824a45c Merge pull request # 108647 from cici37/bumpCEL > d6b1060 Update cobra to v1.4.0 > 21a13eb Bump cel-go to v0.10.1 > 58b0992 Merge pull request # 108204 from kevindelgado/field-validation-crd-unit-tests > f5b06c9 Merge pull request # 108576 from jpbetz/cel-0_10_0 > 68670ad PR feedback > 685b1c8 Add tests demonstrating numeric comparisons and int-or-string behavior > 97d9a2b WIP: start debuging crd handler testing > df72649 Bump cel-go to v0.10.0 > e429598 Merge pull request # 108437 from andrewsykim/bump-konnectivity-v0.0.29 > 7066a88 bump sigs.k8s.io/apiserver-network-proxy/konnectivity-client@v0.0.30 > cca509a Merge pull request # 108013 from benluddy/cel-transition-rule-schema-validation > e2c64f5 Forbid CEL transition rules on unmergeable CRD subschemas. > 96339dc Merge pull request # 108312 from jpbetz/stdlib > fa462a9 Add stdlib of CEL functions to Kubernetes that extends the standard library provided by CEL > e446098 Merge pull request # 108365 from liggitt/fix-proto-gen > f985f3a Merge pull request # 108339 from cheftako/anp-release > e7aa232 Regenerate protobuf > 59c870e Bump konnectivity-client to v0.0.28 > d5c90c4 Merge pull request # 108328 from dgrisonnet/fix-go-collector > 73189fb bump client_golang to v1.12.1 > b80398f Merge pull request # 94637 from liggitt/namespace-before-admission > f3ea2cc Merge pull request # 108263 from deads2k/more-resthandlers > 6e79810 Update tests to have namespace set in req context > 15285c5 migrate more rest handlers to select by resource enablement > 8b3b1d9 Merge pull request # 107105 from justaugustus/go118 > fd2dd05 generated: Run hack/lint-dependencies.sh and hack/update-vendor.sh > f250253 Merge pull request # 106630 from howardjohn/protogen/full-package-name > 83dafd2 Merge pull request # 108163 from wojtek-t/update_selflink_doc > c54016b go-to-protobuf: regenerate with full go_package > 2a6193b Autogenerated > 1d3fa8d Merge pull request # 107527 from wojtek-t/remove_selflink_ga > d325580 Autogenerated > 50ae234 Remove Selflink from convertors > ba8efeb Remove unused selflink parameters from ContextBasedNaming > c854e56 Relax to using namer instead of selflinker in API groupversion > 00b1c08 Remove selflink integration tests > 392ce15 Merge pull request # 108059 from RaghavRoy145/k8s-utils-update > 7c6ef62 Merge pull request # 107650 from apelisse/update-managedfields-time-field-description > 2d699db Updated k8s.io/utils dependency > 72daf89 Merge pull request # 107970 from liggitt/validations-round-trip > 181de81 Run hack/update-all.sh > ed8b997 Use serializable struct for x-kubernetes-validations in openapi > b475f27 Make JSON schema round tripping test more strict > 29400d7 Merge pull request # 107956 from benluddy/xvalidations-deepcopy > 8ad7e6e Deeply copy JSONSchemaProps.XValidations. > 5e47baf Merge pull request # 107688 from liggitt/prune-array > 871b919 Merge pull request # 105632 from xens/fix/kubectl-socks5-proxy2 > 0c5d09f Fix regression pruning array fields with x-kubernetes-preserve-unknown-fields: true > 2cfc338 feat: add missing SOCKS5 features > bf40ed2 Merge pull request # 105142 from pacoxu/cespare-2.1.2 > 0e92d7a upgrade prometheus/client_golang to v1.12.0(common to v0.32.1) > c718707 upgrade cespare/xxhash/v2 to v2.1.2 > 003be15 Merge pull request # 107551 from wojtek-t/remove_unused_selflink_test_references > 1f308a5 Merge pull request # 107565 from jiahuif-forks/deps/structured-merged-diff > 538d83d Remove selflink references in different testing-related files > 528ea3b generated: ./hack/update-vendor.sh > 05f9509 upgrade sigs.k8s.io/structured-merge-diff/v4 to v4.2.1 > 38cf7f6 Merge pull request # 107176 from Iceber/fix-request-info > 3c274a4 Merge pull request # 107195 from caesarxuchao/blockOwner > 4a15d17 if the ok returned by RequestInfoFrom is false, the requestInfo is nil > ad20064 Refers to the GC doc in the blockOwnerDeletion API doc > 1ef4cd2 Merge pull request # 107103 from pohly/log-klog-update > 62d72d9 Merge pull request # 107061 from alexzielenski/oav3-bug-incorrect-definitions-version > 306b66d dependencies: update klog to v2.40.1 > 6abfddc Merge pull request # 104733 from thaJeztah/bump_moby_term > b2e5f2b fix bug which causes openapi builder to lookup cached possibly incorrect schema > b4a81f5 Merge pull request # 106436 from dims/cleanup-owners-files-no-activity-in-a-year > 0fdde31 go.mod: github.com/moby/term v0.0.0-20210619224110-3f7ff695adc6 > b0c4213 Merge pull request # 106601 from liggitt/gomodule-gengo-normalize > 1d7b071 Cleanup OWNERS files (No Activity in the last year) > 982d384 Merge pull request # 106568 from liggitt/json-fieldpath > 6c4c420 bump k8s.io/gengo > 27fee24 bump sigs.k8s.io/json > a09f760 Merge pull request # 106964 from ncdc/portable-code-generation > dc0fc6d Merge pull request # 106922 from cheftako/anp-release > 4944e93 Bump k8s.io/gengo to v0.0.0-20211115164449-b448ea381d54 > e6d99b1 Merge pull request # 106949 from cpanato/update-golang-xnet > 3497c54 Bump konnectivity network proxy to v0.0.27. > 39a622b Merge pull request # 106458 from dims/lint-yaml-in-owners-files > ebb93b9 dependencies: Update golang.org/x/net to v0.0.0-20211209124913-491a49abca63 > ff3cd97 Merge pull request # 106921 from dims/update-x/tools-to-v0.1.8 > 6dba7a6 Check in OWNERS modified by update-yamlfmt.sh > 0d43231 Update golang.org/x/tools to a specific tag and avoid SHA > aa82854 Merge pull request # 106850 from MadhavJivrajani/deprecate-clock-pkg > ba11ada Bump k8s.io/utils > f3eac92 Merge pull request # 106473 from Dingshujie/fix_inotify_memory_leak > 65b3c17 Merge pull request # 106722 from kevindelgado/field-validation-apirefdoc > 6e695a1 update k/utils to v0.0.0-20211116205334-6203023598ed > b0680dd staging: add dummy commit to trigger gomod update (# 106794) > dba3122 Update fieldValidation godoc > 6ca57ba Merge pull request # 106660 from liggitt/smd-merge > 4ac31f8 Revert sigs.k8s.io/structured-merge-diff/v4 to v4.1.2 bumping knative.dev/eventing fae8544...e46f4d5: > e46f4d5 Quarterly release cadence (# 6520) bumping knative.dev/serving 8eb8629...67255a6: > 67255a6 Update net-gateway-api nightly (# 13296) > f38229d Use new k8s.io/util/clock package (# 13300) > a92d508 Update net-kourier nightly (# 13297) > 50bec63 Update net-istio nightly (# 13295) > 8b76bfe Update net-contour nightly (# 13294) > 1c6a05d Update net-certmanager nightly (# 13291) > c5c5ea2 use GKE 1.23 clusters (# 13289) > 540e947 upgrade to latest dependencies (# 13290) bumping k8s.io/kube-openapi 4241196...3ee0da9: > 3ee0da9 Merge pull request # 299 from Jefftree/def-optimize > 3a31a64 Merge pull request # 292 from Jefftree/v2v3conv > ba6f675 Optimize Definition generation > e69a86c Merge pull request # 298 from Jefftree/allOf > 2cbad14 Add OpenAPIConv > 9f9c01d Merge pull request # 297 from cici37/errFix > 207f0c7 Wrap Refs with AllOf > 29d7264 Merge pull request # 295 from Jefftree/fix-api-json > d2a55e8 Align maxProperties/minProperties err value with others > 52feaf8 Merge pull request # 290 from jiahuif/feature/prune-v2-when-serving-v3 > 352e363 Fix json encoding for OpenAPI v3 Discovery > 662cbb0 Return valid for kubernetes to handle err properly > 13192b9 Merge pull request # 289 from jiahuif/feature/openapi-oneof-type > 2580423 remove embedded v2 from expected schema > 8a7ee80 Merge pull request # 293 from alexzielenski/cache-busting > 79a5a7d generated: update integration test. > d436835 add test for pruning v2 schema. > 6a7b704 Merge pull request # 291 from Jefftree/builder3-test > de6f8b0 split openapi mock class into Fake and FakeV3 > da7e27a add integration test for v3 OneOf > eca4f62 do not include v2 schema when serving v3. > 89ac9db Merge pull request # 284 from Jefftree/cache-busting > 7afa56c Add test for builder3 > 3ca26f0 address comments > 697543f add support for custom v3 oneOf types. > ddc6692 Merge pull request # 288 from Jefftree/gnostic-fix > b0d7548 Cache busting > 130db2b add wrapper around gnostic v3 document > a9dfcaa Merge pull request # 286 from Jefftree/leading-newline > 91ab739 Update gnostic to drop jsonschema dependency > 3f90b8c Merge pull request # 282 from alexzielenski/header_vendorext_marshal > c7e0de3 Merge pull request # 285 from Jefftree/proto-upgrade > 7626df5 Trim leading newlines in Descriptions > 1cd4920 add vendor extensions to marshal openapi v2 header > 86aaf54 Upgrade protobuf and change references of googleapis/gnostic -> google/gnostic bumping k8s.io/code-generator 31c00a6...65c70a5: > 65c70a5 generated: Update kube-openapi and vendor > 15a5f67 Merge pull request # 108126 from sanposhiho/doc/generatedname > c771500 Merge pull request # 108992 from alexzielenski/cache-busting-client-go > 4a52dcf fix the doc about generateName conflict > 5a00495 update vendor > a207ba4 Merge pull request # 108996 from cici37/errUpdate > 9f21e2b Bump kube-openapi > 671bd3e Update kube-openapi (# 108895) > 9929ef4 Merge pull request # 108725 from pohly/klog-flush-and-exit > c5df488 Merge pull request # 108717 from lavalamp/remove-clustername > 206a8cc klog v2.60.1 > a23a1f1 generated files > f85d174 teach code generator about ugly fields > 3e00677 Merge pull request # 108772 from aojea/user_agent > 306583c client-go: update generated > 9dbaade default kubernetes agent for generated clients > cdbee2b Merge pull request # 108644 from Jefftree/googleapis-gnostic > 9763ef4 googleapis/gnostic -> google/gnostic > 8d7923d Merge pull request # 108365 from liggitt/fix-proto-gen > f1858f0 Fix protobuf generation > 0c4a6d9 Merge pull request # 107105 from justaugustus/go118 > 25e8fa3 generated: Run hack/lint-dependencies.sh and hack/update-vendor.sh > 32a7a63 [go1.18] Bump golang.org/x/... dependencies > 5770f4d Merge pull request # 106630 from howardjohn/protogen/full-package-name > 5915ef0 Merge pull request # 108163 from wojtek-t/update_selflink_doc > 6769cf5 Remove double boilerplate > 6538e15 Autogenerated > 87d0e9f go-to-protobuf: use full path for go_package > cad66f3 Merge pull request # 107527 from wojtek-t/remove_selflink_ga > 52348e7 Autogenerated > 547d6d2 Remove SelfLink from autogenerating applyconfigurations > bbecdaa Merge pull request # 108059 from RaghavRoy145/k8s-utils-update > ed2f643 Merge pull request # 107650 from apelisse/update-managedfields-time-field-description > 836551d Updated k8s.io/utils dependency > 2df02cb Merge pull request # 107834 from SataQiu/fix-applyconfiguration > 2515885 Run hack/update-all.sh > f3a897a code-generator: fix the bug that ApplyConfiguration constructor missing WithKind/WithAPIVersion methods > 9c1c6b4 Merge pull request # 105632 from xens/fix/kubectl-socks5-proxy2 > ca4cdc1 feat: add missing SOCKS5 features > e4455d9 Merge pull request # 105142 from pacoxu/cespare-2.1.2 > 76ef341 upgrade prometheus/client_golang to v1.12.0(common to v0.32.1) > 46b6323 Merge pull request # 107565 from jiahuif-forks/deps/structured-merged-diff > 03c1f21 generated: ./hack/update-vendor.sh > 197714c upgrade sigs.k8s.io/structured-merge-diff/v4 to v4.2.1 > 7e38d57 Merge pull request # 106978 from pohly/log-message-verbosity > 1ad454e avoid klog Info calls without verbosity > ae5cafc Merge pull request # 107195 from caesarxuchao/blockOwner > b64c36c Refers to the GC doc in the blockOwnerDeletion API doc > ce8fd65 Merge pull request # 107103 from pohly/log-klog-update > c367d78 dependencies: update klog to v2.40.1 > 08f5f5b Merge pull request # 106601 from liggitt/gomodule-gengo-normalize > 5a604ed bump k8s.io/gengo > 05a01c0 Merge pull request # 106964 from ncdc/portable-code-generation > f1a3d4f Bump k8s.io/gengo to v0.0.0-20211115164449-b448ea381d54 > 4dd26a2 Merge pull request # 106949 from cpanato/update-golang-xnet > 194e9e9 Merge pull request # 106458 from dims/lint-yaml-in-owners-files > f7b51b0 dependencies: Update golang.org/x/net to v0.0.0-20211209124913-491a49abca63 > f1673b2 Merge pull request # 106921 from dims/update-x/tools-to-v0.1.8 > 742f451 Check in OWNERS modified by update-yamlfmt.sh > d001e99 Update golang.org/x/tools to a specific tag and avoid SHA > 77d70b0 Merge pull request # 106850 from MadhavJivrajani/deprecate-clock-pkg > 83c01e1 Bump k8s.io/utils > 901b633 Merge pull request # 106473 from Dingshujie/fix_inotify_memory_leak > 81b6111 Merge pull request # 106722 from kevindelgado/field-validation-apirefdoc > 62630d3 update k/utils to v0.0.0-20211116205334-6203023598ed > 993e1ce staging: add dummy commit to trigger gomod update (# 106794) > 036d4df Update fieldValidation godoc > 1685c04 Merge pull request # 106660 from liggitt/smd-merge > 0f2dbc9 Revert sigs.k8s.io/structured-merge-diff/v4 to v4.1.2 > 999c27b Merge pull request # 105916 from kevindelgado/validation-unify-all bumping knative.dev/networking 215bac9...aded181: > aded181 upgrade to latest dependencies (# 709) bumping golang.org/x/mod 49f84bc...9b9b3d8: > 9b9b3d8 zip: always populate known keys in haveVCS map in TestVCS > c2adaff zip: treat plan9 git as unsupported > 5f2f164 zip: skip tests that run 'git init' and 'git config' on plan9 > 3a5865c modfile: rename directory directive to use > c96bc14 x/mod: update requirement on x/crypto > ecfafd6 semver: remove unused err field > dd30a60 module: accept trailing slash in MatchPrefixPattern > 37dd689 modfile: in SetRequireSeparateIndirect, convert lines to blocks > 4be982b modfile: in SetRequireSeparateIndirect, arrange requirements consistently > 1b1db11 zip: add ErrUnrecognizedVCS error, allowing fallback behavior > 4029241 [dev.cmdgo] modfile: parser changes for workfile proposal > 607370a module: use strings.ContainsRune in fileNameOK > 98d530a zip: add CreateFromVCS, which creates a module zip from vcs > 154199d zip: add bit more information about what checkFiles does bumping k8s.io/client-go 11f68c5...3e73df6: > 3e73df6 Update dependencies to v0.24.4 tag > 5e66e32 Merge pull request # 111274 from Abirdcfly/automated-cherry-pick-of-# 111235-upstream-release-1.24 > 37577f3 fix a possible panic because of taking the address of nil > 686b396 Revert "Introduce APIs to support multiple ClusterCIDRs (# 108290)" > af4295f Merge pull request # 109114 from tkashem/client-go-retry-thread-safe > d8531f5 client-go: make retry in Request thread safe > 33011f1 Merge pull request # 109050 from MadhavJivrajani/client-go-retry > 8a672f0 Merge pull request # 109151 from Argh4k/r-101566 > 01ab7fb client-go: reset request body after response is read and closed > f699049 Revert "Field `status.hostIPs` added for Pod (# 101566)" > 092a109 Introduce APIs to support multiple ClusterCIDRs (# 108290) > 488e9bb Merge pull request # 109137 from wojtek-t/fix_multiple_values_indexer > e540ebe Addresses the issue which caused # 109115 > 1cab689 Add test for indexer with multiple values > b1e85f6 Merge pull request # 109031 from Jefftree/openapiv3beta > 4fdf361 generated: Update kube-openapi and vendor > 9cfda74 Merge pull request # 108032 from deejross/kep3140-cronjob-timezone > 0c14d0c Field `status.hostIPs` added for Pod (# 101566) > 400b1be make update after timeZone support for CronJob > 94a1081 Implementation on Network Policy Status (# 107963) > 11ca265 Merge pull request # 108992 from alexzielenski/cache-busting-client-go > 92adc4d Merge pull request # 82162 from krmayankk/maxun > fa08fc2 adjust comments > 1a9591b API: maxUnavailable for StatefulSet > 27f72fc update vendor > 402aa66 Merge pull request # 108996 from cici37/errUpdate > 018cf8a add fetching into discovery client for OpenAPI v3 > ef44f11 Bump kube-openapi > 1933a66 Merge pull request # 108505 from rikatz/kubectl-debug-not-showing-response > e0aaa37 Merge pull request # 107180 from ardaguclu/remove-deprecated-serverresource > 9c04962 Reintroduce response status and header on kubectl verbose debug > 23ab4ae Merge pull request # 108698 from tkashem/last-error > 7e062f8 Remove deprecated discovery/ServerResources function > ed28381 Merge pull request # 108531 from tallclair/redirects > 7c9347d client-go: wrap error from previous attempt to provide more context > b307538 Merge pull request # 108159 from pohly/logcheck > 8c38cf3 Don't follow redirects with spdy > af0252e Update kube-openapi (# 108895) > 9175c47 enhance and fix log calls > f4f092d Merge pull request # 108981 from liggitt/skip-race > 97bcbe7 sync: initially remove files BUILD */BUILD BUILD.bazel */BUILD.bazel Gopkg.toml */.gitattributes > 8302916 Temporarily disable TestCheckRetryClosesBody > cf06297 Merge pull request # 108725 from pohly/klog-flush-and-exit > d82e1fd Merge pull request # 108717 from lavalamp/remove-clustername > 2bba973 klog v2.60.1 > c012194 Merge pull request # 108342 from tkashem/rate-limit-error > a34beeb generated files > a1a133b Merge pull request # 108445 from pohly/storage-capacity-ga > 147848c client-go: chain the error returned by rate limiter > 14e253c remove unneeded references > c55e66b Merge pull request # 108680 from kidlj/fix_cache_index_comments > 6ece0de storage capacity: generated files > dedf7ce Merge pull request # 108852 from mikedanese/auth-plugin-doc > 6b59aa0 make comments of updateIndices optimization code more accurate > f624f8e add better link for gcp auth plugin doc > 3ac142e Merge pull request # 108616 from margocrawf/delete-execcredential-v1alpha1 > c47b802 Merge pull request # 108772 from aojea/user_agent > 7088906 Generated code for deleting exec credential v1alpha1 api > 0f3c791 client-go: update generated > 28b9e26 Remove v1alpha1 of the execcredential > 7390aac default kubernetes agent for generated clients > 6889a6b Merge pull request # 108644 from Jefftree/googleapis-gnostic > 186c332 googleapis/gnostic -> google/gnostic > ab732f5 Merge pull request # 107674 from sanposhiho/api-min-domains > b32b6a7 Add MinDomains API to TopologySpreadConstraints field > 2f52a10 Merge pull request # 107410 from margocrawf/master > 0a04e97 Merge pull request # 108426 from deads2k/e2e-check > 2b59dab Check whether static cert is already configured in UpdateTransportConfig > eb103e0 Merge pull request # 108347 from tkashem/refactor > 7d7fd49 add resource enablement check for e2e tests of beta APIs > 34f3aff client-go: refactor retry logic for backoff, rate limiter and metric > 8e46da3 Merge pull request # 108296 from aojea/client_go_size_metrics > e2c62ff client-go: add request and response size metrics > cc43a70 Merge pull request # 108262 from tkashem/retry-unit-test > 5463dac Merge pull request # 106911 from aojea/client_go_metrics > e4ecde2 client-go: add unit test to verify order of calls > 83bb1e3 client-go: remove no longer used finalURLTemplate > 41e0447 Merge pull request # 107105 from justaugustus/go118 > dade88b generated: Run hack/lint-dependencies.sh and hack/update-vendor.sh > 50aa9bb [go1.18] Bump golang.org/x/... dependencies > a7d2e01 Merge pull request # 107527 from wojtek-t/remove_selflink_ga > dd71ff2 Autogenerated > b74c541 Merge pull request # 108059 from RaghavRoy145/k8s-utils-update > 6f7b0ae Updated k8s.io/utils dependency > 6964e37 Merge pull request # 107834 from SataQiu/fix-applyconfiguration > 4c892f8 Merge pull request # 107812 from diyi0926/add-field-path-back-to-event-logs > 1f90b31 code-generator: fix the bug that ApplyConfiguration constructor missing WithKind/WithAPIVersion methods > 8f44946 Merge pull request # 107876 from jlsong01/fix-concurrent-scheme-usage > 5b56e4a resolve casing issue > 470c93d allocate a unique scheme for each test to fix concurrent usage issue > 93a6315 add fieldPath back to event logs > af150e1 Merge pull request # 107904 from sabbey37/update_azure_auth > f4bf759 Update azure auth plugin deprecation to warning > 7f04550 Merge pull request # 107507 from alexzielenski/client-preprocess-handler > 3255cdc Merge pull request # 105632 from xens/fix/kubectl-socks5-proxy2 > 54928ee modify SharedIndexInformer to use newInformer constructor which supports transformers > f157a09 feat: add missing SOCKS5 features > b70ee0a Merge pull request # 105142 from pacoxu/cespare-2.1.2 > 909a459 Merge pull request # 107559 from liggitt/invalid-selectors > ae5b2b8 upgrade prometheus/client_golang to v1.12.0(common to v0.32.1) > 664b1a6 client-go: refactor: Fix styling issues (# 107248) > aab0bb8 Handle invalid selectors properly > 3618ec5 Merge pull request # 107551 from wojtek-t/remove_unused_selflink_test_references > 092b54e Merge pull request # 107565 from jiahuif-forks/deps/structured-merged-diff > a806c6e Remove selflink references in different testing-related files > 61ce20e generated: ./hack/update-vendor.sh > 3c9082a upgrade sigs.k8s.io/structured-merge-diff/v4 to v4.2.1 > 07be351 Merge pull request # 105387 from howardjohn/metadata-client/tracker > 23e2d9e Merge pull request # 107293 from dims/jan-1-owners-cleanup > f582d43 Make metadata fake client implement testing.FakeClient > 3bf0eac OWNERS cleanup - Jan 2021 Week 1 > b025aba Merge pull request # 107420 from 21kyu/fix-typos > 3480737 fix typos in comment > 255c796 Merge pull request # 93709 from ash2k/ash2k/interruptible-sleep > 6524542 Merge pull request # 106547 from astraw99/fix-log-typo > 7a08020 Make sleep interruptible > 8344b1c Merge pull request # 107311 from fasaxc/fix-resource-ver-match > a8ff96d fix log typo > 5754d8f client-go: Clear the ResourceVersionMatch on paged list calls > 782ff78 Merge pull request # 107103 from pohly/log-klog-update > a6257fd Enable setting proxyurl in kubeconfig via kubectl config (# 105566) > 416ac20 dependencies: update klog to v2.40.1 > d7ed940 Merge pull request # 106436 from dims/cleanup-owners-files-no-activity-in-a-year > 6096096 Merge pull request # 106934 from spencer-p/spencer-master > 73f2731 Cleanup OWNERS files (No Activity in the last year) > 056a9de Merge pull request # 106568 from liggitt/json-fieldpath > 5a53787 Document when workqueue metrics are dropped > df79851 bump sigs.k8s.io/json > 80a3c86 Merge pull request # 106949 from cpanato/update-golang-xnet > 6919fc5 Merge pull request # 106458 from dims/lint-yaml-in-owners-files > b630899 dependencies: Update golang.org/x/net to v0.0.0-20211209124913-491a49abca63 > b945f6e Merge pull request # 106921 from dims/update-x/tools-to-v0.1.8 > 70f09c4 Check in OWNERS modified by update-yamlfmt.sh > 096f1fd Merge pull request # 106852 from wojtek-t/remove_legacy_leader_election > 464c8e2 Update golang.org/x/tools to a specific tag and avoid SHA > cbd965e Merge pull request # 106850 from MadhavJivrajani/deprecate-clock-pkg > 276ea3e Remove support for Endpoints and ConfigMaps lock from leader election > 1106390 Bump k8s.io/utils > 8bffd27 Merge pull request # 106473 from Dingshujie/fix_inotify_memory_leak > c713f38 Merge pull request # 106153 from hyschumi/refactor_thread_safe_store > c8a8ad5 update k/utils to v0.0.0-20211116205334-6203023598ed > d85d0d9 Merge pull request # 106774 from SergeyKanzhelev/grpcFieldRename > 547d6c8 refactor: remove dup code > 2f5ae78 generated files for the grpc field rename > 91177d7 Merge pull request # 106716 from aojea/http1_flake_timeout > 115ee0e bump TestHTTP1DoNotReuseRequestAfterTimeout timeout > 1218545 Merge pull request # 106660 from liggitt/smd-merge > 38ad836 Revert sigs.k8s.io/structured-merge-diff/v4 to v4.1.2 > 0e6a989 Merge pull request # 99728 from mattcary/control bumping knative.dev/caching d274c12...8bbae3e: > 8bbae3e upgrade to latest dependencies (# 679) > 5443dbe upgrade to latest dependencies (# 678) bumping knative.dev/hack 36b2b3c...7556452: > 7556452 :gift: Use modscope tool to get info about Go modules (# 223) > 103cdb3 Default `SIGN_IMAGES` variable to `false` (# 225) bumping knative.dev/pkg 93b66e6...3d24b84: > 3d24b84 fix lint warnings (# 2589) > cc6e435 upgrade to latest dependencies (# 2587) > 20aabd5 Bump to k8s v0.24.4 client packages (# 2586) > 2a05469 updates boilerplate action to use go v1.18 (# 2588) bumping k8s.io/apimachinery 276a8a7...97e5df2: > 97e5df2 fix remove implicit copy of a lock > 6550efd Merge pull request # 109102 from liggitt/darwin-tls > 00f0711 Merge pull request # 109031 from Jefftree/openapiv3beta > 53a85ef Tolerate additional error messages in TLS unit tests > 9b5b68c generated: Update kube-openapi and vendor > 31e52c9 Merge pull request # 108126 from sanposhiho/doc/generatedname > 3b8fb46 Merge pull request # 108713 from jiahuif-forks/feature/openapi/intstr-any-of > dd2f21c fix the doc about generateName conflict > 2866f23 oneOf types for IntOrString > 7b6c37e oneOf types for Quantity > 1992dcc Merge pull request # 109044 from liggitt/encoder-log > 4d8ad18 Merge pull request # 108996 from cici37/errUpdate > 9015177 Raise verbosity of EncoderWithAllocator log > ea355d9 Bump kube-openapi > 1a82635 Merge pull request # 108859 from cici37/placeholder > 27a96d8 Merge pull request # 109019 from liggitt/null-fix > 2c106ec Remove unused func > df7fbf8 Merge pull request # 108531 from tallclair/redirects > 6b8ae0d Make strict json unstructured decoding consistent with non-strict decoding > 611d116 Update err handling > b68ae5e Update kube-openapi (# 108895) > 30f68c4 Don't follow redirects with spdy > 955b30c Expand testcases around untyped data in json decoding > aacfc8a Return a placeholder error for blocking failure before CEL validation. > 7300632 Merge pull request # 108725 from pohly/klog-flush-and-exit > ff4eb2c Merge pull request # 108717 from lavalamp/remove-clustername > aaf4b2e klog v2.60.1 > a475f7c generated files > a8dbe81 review comments > f992b21 remove unneeded references > 5556187 find and replace > e1c5512 change field name > c68a4df Merge pull request # 108644 from Jefftree/googleapis-gnostic > eeb7a8d Merge pull request # 108359 from ArangoGutierrez/devel/108355 > b8c3709 googleapis/gnostic -> google/gnostic > aa72564 Merge pull request # 108252 from wojtek-t/avoid_object_deep_copies > a324a2a Fix logging statement with missing parameter > 5f1f3f0 Prepare apiserver for operating on cached objects by not modifying them > d81a7ed Merge pull request # 106525 from dgrisonnet/sanitize-forwarded-uri > da0995b Merge pull request # 106830 from pacoxu/StreamingProxyRedirects-remove > 2b952f7 pkg/util/proxy: escape redirect URL > 2936d3f Merge pull request # 108365 from liggitt/fix-proto-gen > 96ac948 remove unused parameter: intercceptRedirects & RequireSameHostRedirects > b0775b9 apimachinery/pkg/util/proxy: escape forwarded URI > 57893b8 Merge pull request # 108186 from p0lyn0mial/watch-list-reduce-allocations-in-watch-server > 1657375 Regenerate protobuf > 4f3ae9f Merge pull request # 107105 from justaugustus/go118 > 5901cfa introduces a new streaming encoder that utilizes a memory allocator during objects serialization > 86d2c9e Merge pull request # 108219 from liggitt/api-compatibility-fuzzer > dc51050 generated: Run hack/lint-dependencies.sh and hack/update-vendor.sh > ce93857 fixes TestNestedEncodeError test > f356a33 Merge pull request # 106630 from howardjohn/protogen/full-package-name > 7d61d87 Fill API compatibility data with identifying values rather than random data > 60477b4 [go1.18] Bump golang.org/x/... dependencies > 94e26b2 introduces a memory allocator > ddce29f Merge pull request # 107697 from kevindelgado/nested-decoding > 08e82e3 go-to-protobuf: regenerate with full go_package > a65a9ec provides EncodeWithAllocator method for the protobuf encoder > ec22c43 Merge pull request # 107705 from ardaguclu/filter-strategic-patch-nonexist > d603de9 Nested decoders handle strict decoding errors > 92a83a4 codec: exposes EncodeWithAllocator method > a995449 Merge pull request # 108163 from wojtek-t/update_selflink_doc > cfd9b3f Simplify casting in discardNullValuesFromPatch > e7b1571 codec interfaces > 20500e5 Autogenerated > 1b70fc8 Discard null values in complex objects in strategic patch > f5001e0 Update SelfLink field documentation > 1414335 Merge pull request # 107527 from wojtek-t/remove_selflink_ga > cafff28 Autogenerated > eebcb96 Update SelfLink OpenAPI documentation > 7e3486b Relax to using namer instead of selflinker in API groupversion > 4475671 Introduce Namer interface > cec36f7 Remove selflink setting from apiserver > 1a1682d Merge pull request # 108059 from RaghavRoy145/k8s-utils-update > a8d1671 Merge pull request # 107650 from apelisse/update-managedfields-time-field-description > 0ff68d0 Updated k8s.io/utils dependency > 96c98da Merge pull request # 108055 from liggitt/api-compatibility > ce3db85 Run hack/update-all.sh > e7458ea Add check for unused API compatibility fixture files > 04356ed Update ManagedFields Time description to be more accurate > df99359 Merge pull request # 107841 from aojea/fix_race_socks5 > 1dd05ad deflake TestRoundTripSocks5AndNewConnection unit test > 322368c Merge pull request # 107758 from aojea/flake_socks5 > 3b49665 disable socks5 flake test > 3d7c63b Merge pull request # 105632 from xens/fix/kubectl-socks5-proxy2 > e787e55 feat: add missing SOCKS5 features > 73cb564 Merge pull request # 105142 from pacoxu/cespare-2.1.2 > e537f0c upgrade prometheus/client_golang to v1.12.0(common to v0.32.1) > 162a22f Merge pull request # 107606 from xens/feat/propagate_req_context_into_proxyreq > 91a5267 feat: propagate req context into proxyReq > 3c16f3d Merge pull request # 107565 from jiahuif-forks/deps/structured-merged-diff > d54bd08 generated: ./hack/update-vendor.sh > d1e4f60 upgrade sigs.k8s.io/structured-merge-diff/v4 to v4.2.1 > 80d954b Merge pull request # 107293 from dims/jan-1-owners-cleanup > d8a436c OWNERS cleanup - Jan 2021 Week 1 > e9b426b Merge pull request # 106472 from ash2k/ash2k/remove-return-value > 1646622 Merge pull request # 106647 from kidlj/add_unit_tests_for_SetNestedSlice > 489a8d5 Remove return value from MaybeResetRESTMapper() > 621c98e Merge pull request # 106915 from MikeSpreitzer/deprecate-interval-clock > 9225110 add unit tests for TestSetNestedStringSlice, TestSetNestedSlice etc. > b76ef24 Merge pull request # 104819 from justinsb/less_cmp > 3750a41 Note dishonesty of IntervalClock > fbbdcc6 Merge pull request # 106688 from kidlj/fix_obsolete_runtime_schema_comments > f495b04 Use utils string-slice comparison > ea11419 Merge pull request # 107195 from caesarxuchao/blockOwner > 22bcbe5 fix obsolete comments in runtime/schema struct > 9584bbb Merge pull request # 105808 from yuanjize/master > 6976816 Refers to the GC doc in the blockOwnerDeletion API doc > de7147d Merge pull request # 107103 from pohly/log-klog-update > f75369a Fix function comment typo > 7e097b1 dependencies: update klog to v2.40.1 > e7b02e6 Merge pull request # 106436 from dims/cleanup-owners-files-no-activity-in-a-year > e65876e Merge pull request # 106568 from liggitt/json-fieldpath > 4e5ef7a Cleanup OWNERS files (No Activity in the last year) > 90b4bfb bump sigs.k8s.io/json > 6df201a Merge pull request # 106949 from cpanato/update-golang-xnet > 32abfd8 Merge pull request # 106458 from dims/lint-yaml-in-owners-files > 4bfcafb dependencies: Update golang.org/x/net to v0.0.0-20211209124913-491a49abca63 > ee64e46 Merge pull request # 106921 from dims/update-x/tools-to-v0.1.8 > 1edfc96 Check in OWNERS modified by update-yamlfmt.sh > 57e7b49 Update golang.org/x/tools to a specific tag and avoid SHA > 0cb2c3d Merge pull request # 106850 from MadhavJivrajani/deprecate-clock-pkg > 5a5dd29 clock: Deprecate types to use k8s.io/utils/clock > 573b50c Bump k8s.io/utils > 32df714 Merge pull request # 106473 from Dingshujie/fix_inotify_memory_leak > 6cf7a67 Merge pull request # 106797 from kidlj/fix_mapToUnstructured_todo > 3b691e5 update k/utils to v0.0.0-20211116205334-6203023598ed > 15885e2 Merge pull request # 106722 from kevindelgado/field-validation-apirefdoc > a58f9b5 remove TODOs in mapToUnstructured and sliceToUnstructured > 15877ef Merge pull request # 101719 from SergeyKanzhelev/removeReallyCrashForTesting > 8fa9c2c Update fieldValidation godoc > a50b011 Merge pull request # 106786 from kidlj/fix_mapToUnstructured_error_message > 73ee90c remove ReallyCrashForTesting and cleaned up some references to HandleCrash behavior > 0b80a01 Merge pull request # 106785 from kidlj/cleanup_typeFrom_made_twice > 5052b16 fix mapToUnstructured error message > 67ffd82 Merge pull request # 106646 from pohly/api-errors-nil > 32ad59f cleanup: use present typeFrom variable to avoid another reflect.TypeOf call > 7d2b743 Merge pull request # 103863 from tanjing2020/time > 995f5f8 api/errors: explicitly allow nil error parameters > cdd1abb Merge pull request # 106452 from MadhavJivrajani/watch-test-coverage > 15b0056 Replace with > 5f07275 staging: add dummy commit to trigger gomod update (# 106794) > 418d00d Add unit test for Recorder > ffb9472 Merge pull request # 106660 from liggitt/smd-merge > a9c6a30 Revert sigs.k8s.io/structured-merge-diff/v4 to v4.1.2 > 7149480 Merge pull request # 105916 from kevindelgado/validation-unify-all bumping github.com/prometheus/client_golang 989baa3...2e1c481: > 2e1c481 Cut v1.12.1 (# 978) > 77626d6 Reduce granularity of histogram buckets for Go 1.17 collector (# 974) > 4dd3cbb API client: make http reads more efficient (# 976) > 8520671 Use simpler locking in the Go 1.17 collector (# 975) > f63e219 Make the Go 1.17 collector thread-safe (# 969) > 0108796 Cut v1.12.0 (# 966) > 5bd9ee5 go.*: Update dependencies (# 965) > 9075cdf promhttp: Check validity of method and code label values (# 962) > 22da949 Use the runtime/metrics package for the Go collector for 1.17+ (# 955) > dc1559e Update /api/v1/status/tsdb to include headStats (# 925) > e6e54e8 Merge pull request # 944 from yeya24/add-wal-replay-status > 98fbd99 Merge pull request # 952 from prometheus/repo_sync > 440c09d support wal replay status api > dbf420e Merge pull request # 928 from prometheus/repo_sync > 70a41d5 Update common Prometheus files > 6d5cf25 Merge pull request # 939 from gozeloglu/fix-typo > 1d09783 Merge pull request # 950 from mrueg/go-1.17 > a7919ad Update common Prometheus files > 1b145ca Merge pull request # 935 from prometheus/beorn7/examples > 55320aa Fix typo > 8535b62 Add new go:build lines for go 1.17 > 679eb0d Update cespare/xxhash dependency (# 913) > f78dc19 Move flags and metrics into main() > d28a147 chore: adds support for go 1.17 on CI. > 35ee299 Merge pull request # 909 from prometheus/repo_sync > 2261d5c add ExponentialBucketsRange function (# 899) > 6cd53c4 Update common Prometheus files > 20eef74 Update common Prometheus files (# 892) > cb5c8ff Merge pull request # 881 from sivabalan/master > 26d4b4c Merge pull request # 885 from prometheus/superq/badgers > 5426b51 Updating dependency versions > c7175cc Merge pull request # 888 from prometheus/repo_sync > 6d946b5 Update status badgets > 1965365 Merge pull request # 879 from kakkoyun/fix_changelog > e23b9ce Update common Prometheus files > 64a9c51 Address minor issues on Changelog bumping golang.org/x/tools c20fd7c...897bd77: > 897bd77 internal/gocommand: remove support for -workfile > e6a7e13 go/analysis/tools/internal/checker: add support for RunDespiteError > adcfb65 internal/lsp/source: use the object as the hover source for type params > 411d040 internal/lsp/source: clean up the interface to hover information > 45aeaf7 internal/lsp/source: improve the heuristics for linkable identifiers > a317113 internal/lsp/source: fix hover on generic type declarations > 1f3875c internal/lsp/source: begin to refactor hovering with a HoverContext type > fd59bdf internal/lsp/source: adjust object position when formatting full AST > cda4201 internal/lsp/source: simplify Identifier.enclosing > 59f1f2c cmd/goyacc: reduce array sizes using smaller base types > c677677 internal/lsp/cache: let Session.getGoEnv query GOWORK > 9095d10 go/analysis/passes/tests: fix a missed rename from CL 374495 > 70c3ea2 gopls,internal/lsp: Implement method stubbing via CodeAction > 2ff4db7 go/analysis/passes/tests: Check malformed fuzz target. > 11109f6 go/ssa/ssautil: Initialize Instances field. > 33002ea go/loader: Initialize (types/Info).Instances field > be40034 internal/lsp: add support for formatting go.work files > 2405dce internal/lsp: use placeholders with prepare rename > c6fca02 godoc: handle type parameters correctly in LinkifyText > c0b9fb5 internal/lsp/analysis/undeclaredname: suppress impossible quick fixes > 9b156ee internal/lsp/source: provide full documentation of builtin types > caecc2b internal/typeparams: add a generic form of types.AssignableTo > 164402d internal/lsp/cache: set types.Config.GoVersion > 6aaba77 internal/lsp/debug: add go version to gopls version output > 414ec9c internal/lsp: add ListImports > d55d892 gopls/doc: how to request a particular gopls w/ sublime text > cf66aec gopls: update coc.nvim documentation for using go.work > 597b165 internal/lsp/cache: use -workfile on 1.18 > 8305295 gopls: recognize the go.work file type > 962260a internal/lsp/source: remove the unused source.DetectLanguage > 3e30e21 gopls: remove the workspace_metadata command > 25d2ab2 internal/lsp/cache: fixes for workspace invalidation > 8d915b1 internal/lsp/cache: simplify workspace construction > 77f5fa5 internal/lsp/cache: replace old go.work parser with x/mod > 4da8ceb internal/lsp/source: remove the unused UpdateUserModFile invocation flag > 727dfdb internal/lsp/source/completion: add conventional acronyms for type names > c009ab4 internal/lsp/completion: the inserted text should be what is shown to the user > eeb95ae gopls: make 'fastfuzzy' the default fuzzy matcher for workspace symbols > ea5e1dc internal/typeparams: add a helper to return the origin method > a739c97 internal/lsp/fake: retry ioutil.ReadFile on plan9 if it fails due to exclusive use > 82366c6 internal/lsp/fake: in WriteFileData, retry writes that fail with ERROR_LOCK_VIOLATION on Windows > f6067dc gopls: cleanup the main help pages > dfda270 gopls: use the = form in flag parameter printing rather than a space > 5ad7054 gopls: add long forms for all the 1 rune flags > 217a9fc gopls: take control of the flag printing > afd524f gopls: clean up the help text > ec9a7de gopls: write out the raw usage text > 461d130 internal/lsp: include receiver type parameters when completing receivers > 263785d gopls: update dependencies post-release > 8e9bdc5 gopls: document that v0.7.5 is the final version to support Go 1.12 > 492cf3a doc: update the instruction to use beta2 > 939c2c0 godoc/redirect: close HTTP response bodies in TestRedirect > 0f0bbfd go/callgraph/vta: avoid panic on missing function definitions > 845bb90 go/types/objectpath: update objectpath tests for TypeString changes > be5769c go/types/objectpath: update objectpath tests for TypeString changes > 135791d go/callgraph/vta: move test code to testdata/src > fe74b5f internal/template: return available semantic tokens even on template error > 772a399 internal/template: return a Source with template diagnostics /test eventing-upgrade-tests_operator_main /test upgrade-tests_operator_main /test serving-upgrade-tests_operator_main /lgtm /approve
gharchive/pull-request
2022-09-12T01:46:58
2025-04-01T06:44:42.198355
{ "authors": [ "houshengbo", "knative-automation" ], "repo": "knative/operator", "url": "https://github.com/knative/operator/pull/1202", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1771458271
[main] Upgrade to latest dependencies Cron -knative-prow-robot /cc knative/operations-wg-leads /assign knative/operations-wg-leads Produced by: knative-sandbox/knobots/actions/update-deps /lgtm /approve
gharchive/pull-request
2023-06-23T13:22:20
2025-04-01T06:44:42.204997
{ "authors": [ "houshengbo", "knative-automation" ], "repo": "knative/operator", "url": "https://github.com/knative/operator/pull/1481", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
601046242
Update the manifest of serving to 0.14.0 Fixes # Proposed Changes This PR updates the manifest of serving to 0.14 version. Release Note /lgtm
gharchive/pull-request
2020-04-16T13:18:24
2025-04-01T06:44:42.207078
{ "authors": [ "houshengbo", "jcrossley3" ], "repo": "knative/serving-operator", "url": "https://github.com/knative/serving-operator/pull/403", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
212246903
Request for feedback! Hi! I have changed so that i can get all messages during a period 2017-01-01 -> 2017-01-05. So this fixes my problem i am know wondering what your thoughts are on my solution. Do you think i am on the right track? Do you feel there are missing tests ? My hope is to merge this so that i can get rid of my custom version of elktail. Best Regards Björn Blomqvist Thanks for PR! I did a quick look at the code and it looks good. I will test it out as soon as I get a chance. How does this solution interact with -n option? I'm assuming it's currently ignoring it? I suppose it kind of makes sense to ignore -n if a full date range is specified... What is your opinion? The quality of this code is quite bad so will be closing this. I hope to return with new and more polished pull requests. /Björn
gharchive/pull-request
2017-03-06T20:54:48
2025-04-01T06:44:42.226964
{ "authors": [ "bjornblomqvist", "knes1" ], "repo": "knes1/elktail", "url": "https://github.com/knes1/elktail/pull/21", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
664213527
knex insert json object to multiple tables I have a json object with nested key/value properties that needs to be inserted to multiple tables based on the key name. It's not a straightforward insert as the output of one insert needs to be piped as input to the next insert. How do I do conditional inserts using knex to mssql? Here is the object that I'm trying to insert { "activity": { "activityType": "Marketing", "activityDate": "2020-07-19", "activityStartTime": "2020-07-19 03:27:45.000", "activityEndTime": "2020-07-19 04:27:45.000", "activityUrl": "https://www.sdss.com", "isPublished": false, "activityDuration": 30, "eventId": "03361624-3964-4f2e-9ab0-34ea48027f09", "activityName": "Customer Delight" }, "marketing": { "Name": "sales booster", "meetingUrl": "http://www.zzz.com", "clientViewName": "sales boost", "linkOneName": "test1", "linkTwoName": "test2", "linkOneNameUrl": "www.gggg.com", "linkTwoNameUrl": "www.gggg.com", "participants": [ { "email": "bbb@bbb.com" }, { "email": "aaa@aaa.com" }, { "email": "ccc@ccc.com" } ] } } In the above, the output of activity insert should be piped to marketing and the output of the same should be piped to participants. Since participants is a separate table, I can't just do a loop as it would involve many db calls. I'd like to club this all and then do a single write. Do we have a better way to achieve this using knex? Closing as knex usage questions should go to gitter / stackoverflow. You need to create a separate insert for each table and setup foreign keys correctly. Most probably you want to use objection.js for that which can insert nested data with correct foreign keys automatically if realation mappings has been set correctly for models.
gharchive/issue
2020-07-23T05:32:25
2025-04-01T06:44:42.229957
{ "authors": [ "csnarain", "elhigu" ], "repo": "knex/knex", "url": "https://github.com/knex/knex/issues/3942", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
218685308
zookeeper服务器断开 您好,在实践中当客户端压力较大时可能会和zookeeper服务器心跳超时,这个时候有的可以自动恢复 ,有的不可以,我想问下什么情况下是不可恢复的 同求 是啊,我们这边也遇到这种问题,时间长了就同步不成功。 我们也经常遇到这个问题。感觉压力也不大, 就是长时间跑后就有这问题 同问!公司的生产环境出现这个问题的频率好高,一直解决不了,打算转Apollo了! 看代码,理论上ConnectionWatcher类里的process 方法中: } else if (event.getState().equals(KeeperState.Disconnected)) { // 这时收到断开连接的消息,这里其实无能为力,因为这时已经和ZK断开连接了,只能等ZK再次开启了 LOGGER.warn("zk Disconnected"); } 加一下reconnect就应该可以,大家试试
gharchive/issue
2017-04-01T09:58:02
2025-04-01T06:44:42.232464
{ "authors": [ "18099538715", "fiyang", "jackiechan1981", "jzja123", "onlyforzhuce", "tatumsu" ], "repo": "knightliao/disconf", "url": "https://github.com/knightliao/disconf/issues/291", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
614711490
Improving DataLayerService read performance [Environment]: Kubernetes, bare metal [Known affected releases]: master (includes all releases) The current DataLayerService asks users to specify whether they want to retrieve the data from local or global data layer. In the future, it'd be valuable to maintain each data layer server keeps additional metadata for each item in the cache. This metadata contains the expiry date timestamp. when a new item is requested from the data layer server, it does the following: 2.1) if it is in the cache and it hasn't expired yet, it returns the value 2.2.) if it is in the cache but expired OR if it is not in the cache, it retrieves from the global data layer, caches it with the new expiry date timestamp and returns that value. The 'lease duration' is then a configurable value. As another improvement, at step 2.1 (when the value in the cache hasn't expired yet), after returning to the client, the data layer server may make another asynchronous request to the global data layer server around the expiry time to retrieve the latest value and update the expiry date timestamp accordingly. This approach would try to keep the caches synchronized and may increase the overhead a bit, but it will only be for the items that have been accessed recently. For others, the cache will expire as normal. (So for a cached item that is accessed once, the time in the cache will be extended at least once more lease duration. Afterwards, if it hasn't been accessed again, it will expire normally.) Then we could have a really lazy thread running once every in a while to clean up the cache to remove the expired values (simply go over the cache and remove the expired ones). This is an optimization to reduce resource consumption and may not be critical. The write synchronization stays the same: it is written to the local cache and sent to the global data layer asynchronously. We can apply the same principle here as well: after writing to the local cache, the expiry date will be set accordingly (hopefully, by that time the synchronization to the global data layer will have finished). Perhaps, we can use this criterion for having an idea about the lease duration. created a new branch "datalayer_lease" for this feature. It's not a lease, because that would imply the local data layer [cache] has a temporal ownership of the global data, which it doesn't - any other local data layer can write the same identifier to the global backend, which renders the local cache content invalid. It's an expiration timer typically labelled TTL. As an alternative to the above solution, one could utilize Redis as a caching layer in front of Riak. See https://riak.com/products/riak-kv/redis-integration/index.html. In this alternative, the DataLayerService would become a translation layer for storage operations: If the operation is a KV operation, then the DataLayerService will redirect it to Redis (which has its own synchronization with Riak). If the operation is a CRDT operation, then the DataLayerService will redirect it to Riak. Other operations related to other features (e.g., triggerable tables) need to be figured out, but will most likely use a combination of Redis and Riak. As a result of these modifications, the DataLayerService will not cache anything; thus, offloading cache-related problems like invalidation to Redis. Furthermore, its logic will be simplified, so that it can become part of each sandbox. As an implementation optimization regarding JVMs, it can be merged with the queue service. @ruichuan, @paarijaat: Did I miss anything? As far as I can tell, the data layer service is only used by sandboxes (including management) and the storage frontend, which is going to be removed (see #28). After that, the data layer service will be the only remaining component in the bare metal setup that requires the fluentbit server to send the logs to elasticsearch (in fact, the fluentbit server is not utilized in the knative setup anyway). If the data layer service moves inside the sandbox, we can also get rid of the extra fluentbit server in bare metal, because the sandbox logs are already sent to elasticsearch. That move would affect the way the management service is bootstrapped, though. @iakkus thanks, it captures our discussion well. The data layer service (DSL) is a bit more than a simple translation layer of redirecting requests to riak/redis for processing. DSL needs to use respective client libraries to organize operations in certain ways for correct processing. I agree that the new `riak_redis' approach can simplify DSL. Regarding where to put DSL (either standalone or inside a sandbox), we can decide later. It anyways is a simple move. switched this work to "feature/redis_caching" branch
gharchive/issue
2020-05-08T12:26:24
2025-04-01T06:44:42.244385
{ "authors": [ "iakkus", "manuelstein", "ruichuan" ], "repo": "knix-microfunctions/knix", "url": "https://github.com/knix-microfunctions/knix/issues/10", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
629134136
Perform storage operations for GUI and CLI via management service (and remove the Jetty frontend) Describe the feature Currently a jetty based frontend handles storage requests from the GUI and the CLI. The frontend queries the management service to authenticate the token received along with the request. Once the token is verified it contacts a Datalayer cache over the thrift interface to perform the necessary storage operation. As the actions performed by the frontend are stateless they can be safely moved into the management service as an additional action - "storageOperation". Benefits Removes an entire component from the system, hence decreases the susceptibility of the system to faults Additional roundtrip to the management service (to verify tokens) is saved. Requires changing: Management service, GUI, CLI. As the HttpFrontend component is getting removed, we'd also need to move the HttpFrontend/nginx_config/ somewhere else and update the relevant ansible/helm files. @paarijaat do you have a branch for this? Branch feature/storage_ops_via_management_remove_httpfrontend. So far, we had always allowed access to the KV store in the GUI. Do you think we should also supply the functions to access the rest of the data structures (i.e., maps, sets, counters) @paarijaat @abeckn? (I guess after the issue related to the storage triggers (#31), we can also add those as well, but those should be handled perhaps from that branch and not this one.) By the way, do we envision a single function with different parameters (e.g., "op"="getMap", params={"name": mapname, "is_private": true}, "op"="getSet", params={"name": setname, "is_private": false})? Or do we envision a separate function for each operation? I think a single function would suffice; otherwise, there would be 30+ functions just for storage access, which might a bit inefficient, assuming the access to such data is not really the most frequent ones (compared to other management actions).
gharchive/issue
2020-06-02T11:40:30
2025-04-01T06:44:42.249935
{ "authors": [ "iakkus", "paarijaat" ], "repo": "knix-microfunctions/knix", "url": "https://github.com/knix-microfunctions/knix/issues/28", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
778138810
Prevent obtaining custom items in some regions Currently, custom items can be obtained and crafted in all worlds of a server. For some servers, this is not desirable. It would be nice if it were possible to add a whitelist or blacklist of worlds in which custom items can be crafted or dropped. This could be made even more powerful by allowing not only worlds, but also biomes or WorldGuard regions to affect drops. I have thought about this, and I don't think WorldGuard regions are very useful for this. Biomes can be interesting for block drops & mob drops, but I don't think they make much sense for blocking crafting.
gharchive/issue
2021-01-04T14:36:26
2025-04-01T06:44:42.251496
{ "authors": [ "knokko" ], "repo": "knokko/custom-items-gradle", "url": "https://github.com/knokko/custom-items-gradle/issues/33", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
149857494
"Attempting MQTT connection...connected" in an endless loop I am using the basic "mqtt_esp8266.ino" example as it is, no changes, except: I connect to my local mosquito instance running on a raspberry pi 2 instead of "broker.mqtt-dashboard.com" and I added different nodeIDs to the topic-string of each device. This example works fine for each individual device, unless I run two esp8266 devices (Wemos D1 mini) with the same software (different topics) in parallel. The Serial output is "Attempting MQTT connection...connected" in an endless loop and the broker logs messages from both devices permanently in an endless loop. Seemed to that both devices are calling each other and vice versa. Has anyone else experience with two esp8266 running in parallel with this mqtt-library? Any advice how to deal with it? It sounds like you haven't changed the client ID to be unique for each device? If so, then when the second one connects, it will cause the first to be kicked off - and so on as they each reconnect. thanks very much, you are perfectly right. Too obvious, should have see it myself
gharchive/issue
2016-04-20T19:05:06
2025-04-01T06:44:42.254713
{ "authors": [ "dirkclemens", "knolleary" ], "repo": "knolleary/pubsubclient", "url": "https://github.com/knolleary/pubsubclient/issues/155", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
213873754
Can't subscribe to Runtime.consoleAPICalled event I have a page with javascript that logs to the console. When I navigate to the page using chromdp with logging enabled, I can see that the Runtime.consoleAPICalled event is being fired. I'd like to subscribe to this event, but my naive approach isn't working. Any tips? c.Run(ctxt, cdp.ActionFunc(func(z context.Context, h cdptypes.Handler) error { ch3 := h.Listen(cdptypes.EventRuntimeConsoleAPICalled) go func() { x := <-ch3 fmt.Println("\n\nGOT A THING3", x) }() return nil })) c.Run(ctxt, cdp.Navigate(file)) time.Sleep(time.Second) I'm using Listen as well, and it doesn't look like any events are being fired? I've put a pull request in for this (#64). You might be interested in the discussion in #79. It would be good to know the context you're using the console events in - are you trying to just print them, or react to them in the execution path? If the latter, it is important to understand when and how you want to use the result, so that race conditions (of the form that go build -race cannot find) can be avoided. So if you could provide a real world example of the problem you're trying to solve - it would be very useful for getting the design and usage patterns for Listen correct. This will be fixed in a big rollout coming in the next couple days.
gharchive/issue
2017-03-13T19:33:33
2025-04-01T06:44:42.262379
{ "authors": [ "MrSaints", "kenshaw", "pwaller", "sandro" ], "repo": "knq/chromedp", "url": "https://github.com/knq/chromedp/issues/41", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
906865904
macros: match unused vars if logging is disabled Fixes #494 I am a bit unsure how to test this. Running the snapshot tests with --no-default-features would probably work, since it then would emit the unused-vars-warnings which should mismatch the snapshots, but this seems a bit over the top. bors r+
gharchive/pull-request
2021-05-30T19:57:36
2025-04-01T06:44:42.267291
{ "authors": [ "Urhengulas", "jonas-schievink" ], "repo": "knurling-rs/defmt", "url": "https://github.com/knurling-rs/defmt/pull/497", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1294950678
'xml.etree.ElementTree.Element' object has no attribute 'getchildren' Python 버전 3.9.7 KoalaNLP 버전 koalanlp-2.1.7 KoalaNLP를 사용한 코드 부분을 보여주세요. from koalanlp.Util import initialize, finalize from koalanlp.proc import * from koalanlp import API initialize(java_options="-Xmx4g -Dfile.encoding=utf-8", KKMA="2.0.2", EUNJEON="2.0.2", ETRI="2.0.2") 안녕하세요, 패키지를 써보고 싶은데 initialize 하면 에러 메세지가 뜹니다, 수정할 수 있는 방법이 있을까요? 'xml.etree.ElementTree.Element' object has no attribute 'getchildren' initialize 에 다른 옵션을 써도 마찬가지입니다 (예: initialize(java_options="-Xmx4g", ETRI="LATEST")) 저도 같은 에러가 나네요 https://stackoverflow.com/questions/72203390/xml-etree-elementtree-element-object-has-no-attribute-getchildren-and-relativ getchildren을 iter로 바꾸면 되네용 실행 환경 : Python 3.10.9 찾아보니, getchildren() is deprecated since Python version 2.7 이라고 하네요. 문서 참고해서, koalanlp/jip/maven.py>Pom 클래스에서 에러 발생하는 줄 코드 iter()로 수정하니 실행됩니다. iter(tag=None) 루트 엘리먼트에 대한 트리 이터레이터를 만들고 반환합니다. 이터레이터는 이 트리의 모든 엘리먼트를 섹션 순서대로 루핑합니>다. tag는 찾을 태그입니다 (기본값은 모든 엘리먼트를 반환하는 것입니다). if properties_ele is not None: # prop_eles = properties_ele.getchildren() prop_eles = properties_ele.iter()
gharchive/issue
2022-07-06T00:55:34
2025-04-01T06:44:42.273724
{ "authors": [ "June-Jo", "aaajeong", "ehsong" ], "repo": "koalanlp/koalanlp", "url": "https://github.com/koalanlp/koalanlp/issues/32", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2235082446
More pics I just noticed that the sidebar is big and there is actually space for more pictures. I would maybe align it to middle or make it possible to add more pics? check: https://github.com/kodadot/nft-gallery/issues/9634 it's possible to do it, Viki told me how to do it.
gharchive/issue
2024-04-10T08:53:33
2025-04-01T06:44:42.282719
{ "authors": [ "JustLuuuu", "exezbcz" ], "repo": "kodadot/nft-gallery", "url": "https://github.com/kodadot/nft-gallery/issues/10043", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1556991673
redesign: icons in wallet dialog redesign During the wallet redesign, I found that some icons in the Figma are not supported by our icon packagefontawesome v5. icons in figma: icons in fontawesome v5 possible solution: use custom svg file fontawesome v6 / pro version ref: #4788 related with #796
gharchive/issue
2023-01-25T17:06:02
2025-04-01T06:44:42.285068
{ "authors": [ "Jarsen136", "roiLeo" ], "repo": "kodadot/nft-gallery", "url": "https://github.com/kodadot/nft-gallery/issues/4830", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1688575259
QR code modal is not fully displayed on profile page What happened? QR code modal is not fully displayed in user page. Please reproduce in steps go to https://canary.kodadot.xyz/bsx/u/bXjrHpyJzSczJuuqmtN6nqCnd5b7KFa5LK1rBtF1tUHDYzp8N?page=1 click qrcode button Expected Behavior The screenshot below shows what the modal looks like on kodadot.xyz. What browsers are you seeing the problem on? Desktop Chrome At which address did you encounter bug? canary.kodadot.xyz, beta.kodadot.xyz Are you logged in? Yes Which wallet you are using? PolkadotJS At which chain did you encounter bug? No response Screenshots Relevant log output No response Payment link for reward https://beta.kodadot.xyz/transfer?target=GqheyaQ3c8SFBJYorTXyLW6jvLXKNAXzbad77NXoLYJBXXs Code of Conduct [X] I agree to follow this project's Code of Conduct legit @Gavin-Gong if you want to fix it :) 👋
gharchive/issue
2023-04-28T13:37:19
2025-04-01T06:44:42.290500
{ "authors": [ "Gavin-Gong", "yangwao" ], "repo": "kodadot/nft-gallery", "url": "https://github.com/kodadot/nft-gallery/issues/5829", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1720440754
Add attribute button almost invisible What happened? The add attribute button is almost invisible when trying to create RMRK2 NFTs Please reproduce in steps Chose RMRK2 as chain Go to Create Click on "Show On Chain Attributes" Expected Behavior Button should be more visible What browsers are you seeing the problem on? Firefox At which address did you encounter bug? canary.kodadot.xyz, beta.kodadot.xyz Are you logged in? Yes Which wallet you are using? Talisman Screenshots https://github.com/kodadot/nft-gallery/assets/36627808/1eb456ec-6c7b-4868-831b-9d5bdcf38e6b Relevant log output No response Payment link for reward No response Code of Conduct [X] I agree to follow this project's Code of Conduct DarkMode ftw!
gharchive/issue
2023-05-22T20:16:18
2025-04-01T06:44:42.294654
{ "authors": [ "prury", "roiLeo" ], "repo": "kodadot/nft-gallery", "url": "https://github.com/kodadot/nft-gallery/issues/6073", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2201296823
fix: Drop cards order on drop page Thank you for your contribution to the KodaDot - One Stop Shop for Polkadot NFTs. 👇 __ Let's make a quick check before the contribution. PR Type [x] Bugfix Needs Design check @exezbcz please review Context [x] Closes #9844 new order would be the same as on the landing page - newest minting live drop to the oldest minting live drop put scheduled - under 24hours that get to the top part first Did your issue had any of the "$" label on it? [x] My DOT address: Payout Screenshot 📸 [x] My fix has changed UI Copilot Summary copilot:summary copilot:poem thank you! thanks! pay 20 usd 😍 Perfect, I’ve sent the payout 💵 $20 @ 9.11 USD/DOT ~ 2.195 $DOT 🧗 16SjUbGKSdjCdWTy3NNT3JxbRVGGqD4mwkHpc6BD9U2Rp29Z 🔗 0x78c902aeaa1d514ae8b79da7d2f2eb93c8e31a6d93ca5a6e6f83c68cc3bd204e 🪅 Let’s grab another issue and get rewarded! 🪄 github.com/kodadot/nft-gallery/issues
gharchive/pull-request
2024-03-21T22:24:16
2025-04-01T06:44:42.301602
{ "authors": [ "Jarsen136", "exezbcz", "yangwao" ], "repo": "kodadot/nft-gallery", "url": "https://github.com/kodadot/nft-gallery/pull/9846", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
348629571
Toggle Offset not working after follow-focus event I updated chunkwm + plugins to 0.4.5 today. My last update was probably a month ago. I noticed a change in behaviour which I couldn't find a note in the changelog for: When I send a window to a different desktop with follow focus (chunkc tiling::window --send-to-monitor 2; chunkc tiling::monitor -f 2) and try to toggle offset immediately afterwards (lalt - a : chunkc tiling::desktop --toggle offset) nothing will happen until I change focus to a different window and re-focus the widow I sent to the desktop beforehand. Can someone reproduce this behaviour? Samples taken from .skhdrc: lctrl + lcmd - w: chunkc tiling::window --send-to-monitor 2; chunkc tiling::monitor -f 2 lalt - a : chunkc tiling::desktop --toggle offset Note to self: Change how we detect active space to prioritize space of active window, if a window is actually in focus; fallback to current method. LOC: https://github.com/koekeishiya/chunkwm/blob/master/src/plugins/tiling/controller.cpp#L1429 Should probably be done for all methods where we interact with desktops separately from windows.. fixed on master.
gharchive/issue
2018-08-08T08:41:29
2025-04-01T06:44:42.322951
{ "authors": [ "Beulenyoshi", "koekeishiya" ], "repo": "koekeishiya/chunkwm", "url": "https://github.com/koekeishiya/chunkwm/issues/484", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1250705489
How could I make it detect the "Semicolon Key" or other Punctuation? Take the semicolon as an example : I have tried to edit it in my skrdrc like this : alt - ";" : <commands> and alt - ';' : <commands> But it couldn't work. I don't know if this is because the semicolon and other punctuation symbols are keywords. Forgive my poor English, please. @Ushio-Clover alt - 0x29 : <command> works fine. This would be nice to have in the readme 😄. @syuni Thank you very much for your help! I don't know exactly what type hotkey needs to be set. I look stupid. 😂
gharchive/issue
2022-05-27T12:22:12
2025-04-01T06:44:42.326087
{ "authors": [ "Ushio-Clover", "jjant", "syuni" ], "repo": "koekeishiya/skhd", "url": "https://github.com/koekeishiya/skhd/issues/205", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1015735852
Active border stopped working. Active border stopped working after making an update of yabai. Only work once when I restart Yabai. Then it disappears after switching between windows and never works back. Only shows a little corner border at the top. Please look at the picture below. https://ibb.co/3N36VyS Window borders require the scripting-addition. Make sure it is loaded properly on startup: https://github.com/koekeishiya/yabai/wiki/Installing-yabai-(latest-release)#macos-big-sur---automatically-load-scripting-addition-on-startup
gharchive/issue
2021-10-04T23:45:08
2025-04-01T06:44:42.328034
{ "authors": [ "ahmed-rezk-dev", "koekeishiya" ], "repo": "koekeishiya/yabai", "url": "https://github.com/koekeishiya/yabai/issues/1021", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1183464160
Sharing desktop in Webex: cannot see windows Configuraitons $ sw_vers Mon Mar 28 15:40:55 2022 ProductName: macOS ProductVersion: 12.3 BuildVersion: 21E230 $ yabai --version Mon Mar 28 15:40:58 2022 yabai-v4.0.0 MBP 16inch 2019, intel CPU. SIP disabled and "sa" is installed. Issue When sharing the desktop by using Webex, the host who sharing the desktop can see the windows, but the other participants could not see any window. When disabling yabai and sharing desktop by using Webex, the participants can see all the windows. Same problem here as well Don't think there is a way to resolve this. Disable window borders in your config or disable them using a hotkey before sharing screen: yabai -m config window_border off The following one-liner will toggle: if [[ $(yabai -m config window_border) = "off" ]]; then yabai -m config window_border on; else yabai -m config window_border off; fi works like a charm! thank you @koekeishiya !!
gharchive/issue
2022-03-28T13:42:29
2025-04-01T06:44:42.331030
{ "authors": [ "junwei-wang", "koekeishiya", "sollymay" ], "repo": "koekeishiya/yabai", "url": "https://github.com/koekeishiya/yabai/issues/1213", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
803852558
Cannot convert a symbolic Tensor (dialogue_encoder/strided_slice_4:0) to a numpy array Starting from Rasa 1.10.21, training fails with the following error: pi@pi-yellow:~/test_rasa $ docker-compose up -d Recreating test_rasa_rasa_1 ... done pi@pi-yellow:~/test_rasa $ docker-compose run rasa init Welcome to Rasa! 🤖 To get started quickly, an initial project will be created. If you need some help, check out the documentation at https://rasa.com/docs/rasa. Now let's start! 👇🏽 ? Please enter a path where the project will be crea ted [default: current directory] . ? Directory '/app' is not empty. Continue? Yes Created project directory at '/app'. Finished creating project structure. ? Do you want to train an initial model? 💪🏽 Yes Training an initial model... 2021-02-08 19:15:03 INFO rasa.model - Data (version) for Core model section changed. 2021-02-08 19:15:03 INFO rasa.model - Data (version) for NLU model section changed. Training Core model... Processed Story Blocks: 100%|█| 5/5 [00:00<00:00, 59 Processed Story Blocks: 100%|█| 5/5 [00:00<00:00, 30 Processed Story Blocks: 100%|█| 5/5 [00:00<00:00, 71 Processed Story Blocks: 100%|█| 5/5 [00:00<00:00, 49 Processed trackers: 100%|█| 5/5 [00:00<00:00, 249.91 Processed actions: 16it [00:00, 1342.53it/s, # examples=16] Processed trackers: 100%|█| 231/231 [00:02<00:00, 11 Traceback (most recent call last): File "/opt/venv/bin/rasa", line 8, in <module> sys.exit(main()) File "/opt/venv/lib/python3.7/site-packages/rasa/__main__.py", line 92, in main cmdline_arguments.func(cmdline_arguments) File "/opt/venv/lib/python3.7/site-packages/rasa/cli/scaffold.py", line 206, in run init_project(args, path) File "/opt/venv/lib/python3.7/site-packages/rasa/cli/scaffold.py", line 118, in init_project print_train_or_instructions(args, path) File "/opt/venv/lib/python3.7/site-packages/rasa/cli/scaffold.py", line 58, in print_train_or_instructions args.model = rasa.train(domain, config, training_files, output) File "/opt/venv/lib/python3.7/site-packages/rasa/train.py", line 50, in train additional_arguments=additional_arguments, File "uvloop/loop.pyx", line 1456, in uvloop.loop.Loop.run_until_complete File "/opt/venv/lib/python3.7/site-packages/rasa/train.py", line 101, in train_async additional_arguments, File "/opt/venv/lib/python3.7/site-packages/rasa/train.py", line 188, in _train_async_internal additional_arguments=additional_arguments, File "/opt/venv/lib/python3.7/site-packages/rasa/train.py", line 223, in _do_training additional_arguments=additional_arguments, File "/opt/venv/lib/python3.7/site-packages/rasa/train.py", line 361, in _train_core_with_validated_data additional_arguments=additional_arguments, File "/opt/venv/lib/python3.7/site-packages/rasa/core/train.py", line 66, in train agent.train(training_data, **additional_arguments) File "/opt/venv/lib/python3.7/site-packages/rasa/core/agent.py", line 752, in train self.policy_ensemble.train(training_trackers, self.domain, **kwargs) [7/1116] File "/opt/venv/lib/python3.7/site-packages/rasa/core/policies/ensemble.py", line 124, in train policy.train(training_trackers, domain, **kwargs) File "/opt/venv/lib/python3.7/site-packages/rasa/core/policies/ted_policy.py", line 325, in train batch_strategy=self.config[BATCH_STRATEGY], File "/opt/venv/lib/python3.7/site-packages/rasa/utils/tensorflow/models.py", line 131, in fit ) = self._get_tf_train_functions(eager, model_data, batch_strategy) File "/opt/venv/lib/python3.7/site-packages/rasa/utils/tensorflow/models.py", line 348, in _get_tf_train_functions train_dataset_function, self.train_on_batch, eager, "train" File "/opt/venv/lib/python3.7/site-packages/rasa/utils/tensorflow/models.py", line 330, in _get_tf_call_model_function tf_call_model_function(next(iter(init_dataset))) File "/opt/venv/lib/python3.7/site-packages/tensorflow_core/python/eager/def_function.py", line 568, in __call__ result = self._call(*args, **kwds) File "/opt/venv/lib/python3.7/site-packages/tensorflow_core/python/eager/def_function.py", line 615, in _call self._initialize(args, kwds, add_initializers_to=initializers) File "/opt/venv/lib/python3.7/site-packages/tensorflow_core/python/eager/def_function.py", line 497, in _initialize *args, **kwds)) File "/opt/venv/lib/python3.7/site-packages/tensorflow_core/python/eager/function.py", line 2389, in _get_concrete_function_internal_garbage_collected graph_function, _, _ = self._maybe_define_function(args, kwargs) File "/opt/venv/lib/python3.7/site-packages/tensorflow_core/python/eager/function.py", line 2703, in _maybe_define_function graph_function = self._create_graph_function(args, kwargs) File "/opt/venv/lib/python3.7/site-packages/tensorflow_core/python/eager/function.py", line 2593, in _create_graph_function capture_by_value=self._capture_by_value), File "/opt/venv/lib/python3.7/site-packages/tensorflow_core/python/framework/func_graph.py", line 978, in func_graph_from_py_func func_outputs = python_func(*func_args, **func_kwargs) File "/opt/venv/lib/python3.7/site-packages/tensorflow_core/python/eager/def_function.py", line 439, in wrapped_fn return weak_wrapped_fn().__wrapped__(*args, **kwds) File "/opt/venv/lib/python3.7/site-packages/tensorflow_core/python/framework/func_graph.py", line 968, in wrapper raise e.ag_error_metadata.to_exception(e) NotImplementedError: in converted code: /opt/venv/lib/python3.7/site-packages/rasa/utils/tensorflow/models.py:197 train_on_batch * prediction_loss = self.batch_loss(batch_in) /opt/venv/lib/python3.7/site-packages/rasa/core/policies/ted_policy.py:610 batch_loss * dialogue_embed, mask = self._emebed_dialogue(dialogue_in) /opt/venv/lib/python3.7/site-packages/rasa/core/policies/ted_policy.py:578 _emebed_dialogue * dialogue_transformed = self._tf_layers["transformer"]( /opt/venv/lib/python3.7/site-packages/tensorflow_core/python/keras/engine/base_layer.py:778 __call__ outputs = call_fn(cast_inputs, *args, **kwargs) /opt/venv/lib/python3.7/site-packages/rasa/utils/tensorflow/transformer.py:619 call * pad_mask = tf.minimum( /opt/venv/lib/python3.7/site-packages/rasa/utils/tensorflow/transformer.py:586 _look_ahead_pad_mask * pad_mask = 1 - tf.linalg.band_part(tf.ones((max_position, max_position)), -1, 0) /opt/venv/lib/python3.7/site-packages/tensorflow_core/python/ops/array_ops.py:2659 ones output = _constant_if_small(one, shape, dtype, name) /opt/venv/lib/python3.7/site-packages/tensorflow_core/python/ops/array_ops.py:2391 _constant_if_small if np.prod(shape) < 1000: <__array_function__ internals>:6 prod /opt/venv/lib/python3.7/site-packages/numpy/core/fromnumeric.py:3031 prod keepdims=keepdims, initial=initial, where=where) /opt/venv/lib/python3.7/site-packages/numpy/core/fromnumeric.py:87 _wrapreduction return ufunc.reduce(obj, axis, dtype, out, **passkwargs) /opt/venv/lib/python3.7/site-packages/tensorflow_core/python/framework/ops.py:728 __array__ " array.".format(self.name)) NotImplementedError: Cannot convert a symbolic Tensor (dialogue_encoder/strided_slice_4:0) to a numpy array. If anyone has a clue about this, please let me know. Installing numpy==1.19.5 fixed the issue when running Rasa==1.10.22 locally. Many thanks for the help, @melindaloubser1! Fixing the numpy version to 1.19.5 results in a working 1.20.21 container on my Raspberry Pi, now building 1.20.22. 1.20.22 works too, I published both images to Docker Hub. Keep up the good work with Rasa, I hope to be able to catch up with the Rasa 2.x releases for ARM soon.
gharchive/issue
2021-02-08T19:21:33
2025-04-01T06:44:42.336602
{ "authors": [ "koenvervloesem", "melindaloubser1" ], "repo": "koenvervloesem/rasa-docker-arm", "url": "https://github.com/koenvervloesem/rasa-docker-arm/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
909716009
kohey604のファイルを追加 変更する 変更する
gharchive/pull-request
2021-06-02T17:16:46
2025-04-01T06:44:42.347948
{ "authors": [ "kohey604" ], "repo": "kohey604/github-slideshow", "url": "https://github.com/kohey604/github-slideshow/pull/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2674377186
ROS1 catkin build error report Describe the bug When building with catkin_make, there are no issues, but when using the catkin build command, the following error occurs, and the glim_ros1 build fails. To Reproduce Steps to reproduce the behavior: cd ~/ws catkin build Screenshots // line 17 find_package(glim REQUIRED) Environment (please complete the following information): CPU/GPU: i9-9900K/RTX4090] OS: Ubuntu 20.04 ROS: ROS1 noetic CUDA: 11.8 Build with CUDA: OFF In ROS1, catkin build is commonly used for building rather than catkin_make, so it would be great if there is a way to build using catkin build. I have spent 4 hours trying to resolve this issue but was unable to find a solution, so I am reporting this issue. Thank you. To my understanding, catkin build builds packages independently, resulting in dependency issues. While it should be resolved with minor modifications, I'm not keen to do it as ROS1 is mostly outdated (Any PR from the community will be appreciated). I think building glim and glim_ros in another workspace with catkin_make could be a workaround.
gharchive/issue
2024-11-20T04:44:51
2025-04-01T06:44:42.354253
{ "authors": [ "JongRok-Lee", "koide3" ], "repo": "koide3/glim", "url": "https://github.com/koide3/glim/issues/114", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
226825123
Reduce albums view cards margins Currently, cards margin is (if I'm not wrong) 16dp and that's looks a bit too big, so please reduce it to just 8dp. And also, please add an option to make those cards squared. I know I just requested a squared grid for albums view, but that's a flat view, and this one has cards, and some people might like it. Thanks again. Already done that. I also made the marings equal. They look better, but (from screenshot) I think margins are 8dp still :/ They look better, but (from screenshot) I think margins are 16dp still :/ I actually like the 16dp margins better than the 8dp. @kollerlukas Can you please keep it with 8dp? Or maybe as an option in settings? I think the smaller margin results in best utilized space. I think 8dp looks better as well, so an option might be a good idea.
gharchive/issue
2017-05-07T03:44:31
2025-04-01T06:44:42.854407
{ "authors": [ "TGNYC", "jahirfiquitiva", "kollerlukas" ], "repo": "kollerlukas/Camera-Roll-Android-App", "url": "https://github.com/kollerlukas/Camera-Roll-Android-App/issues/8", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
385968313
Can't get to run torchnlp.ner properly This is the result I get when following installation and running instructions: >>> train('ner-conll2003', TransformerTagger, conll2003) Traceback (most recent call last): File "<stdin>", line 1, in <module> File "torchnlp/tasks/sequence_tagging/main.py", line 46, in train dataset = dataset_fn() File "torchnlp/data/conll.py", line 67, in conll2003_dataset fields=tuple(fields)) File "/usr/local/lib/python2.7/dist-packages/torchtext/data/dataset.py", line 78, in splits os.path.join(path, train), **kwargs) File "/usr/local/lib/python2.7/dist-packages/torchtext/datasets/sequence_tagging.py", line 33, in __init__ examples.append(data.Example.fromlist(columns, fields)) File "/usr/local/lib/python2.7/dist-packages/torchtext/data/example.py", line 50, in fromlist setattr(ex, n, f.preprocess(val)) File "/usr/local/lib/python2.7/dist-packages/torchtext/data/field.py", line 181, in preprocess x = Pipeline(six.text_type.lower)(x) File "/usr/local/lib/python2.7/dist-packages/torchtext/data/pipeline.py", line 37, in __call__ x = pipe.call(x, *args) File "/usr/local/lib/python2.7/dist-packages/torchtext/data/pipeline.py", line 52, in call return [self.convert_token(tok, *args) for tok in x] TypeError: descriptor 'lower' requires a 'unicode' object but received a 'str' Took conll2003 dataset files from THIS REPO ENV: Distributor ID: Ubuntu Description: Ubuntu 16.04.4 LTS Release: 16.04 Codename: xenial -------------------------------- torch.__version__: 0.4.1 torchtext.__version__: 0.3.1 -------------------------------- python: 2.7.12 Hey you should try running with PyTorch version 3.5 or greater, that might fix it Sorry I meant Python 3.5! You are right. I'm running python 3.6.7 and it works for me Even worse :/ With py3 getting always ENV Linux version 4.15.0-39-generic buildd@lcy01-amd64-012 gcc version 5.4.0 20160609 Ubuntu 5.4.0-6ubuntu1~16.04.10 #42~16.04.1-Ubuntu SMP Wed Oct 24 17:09:54 UTC 2018 ----------------- Python: 3.6.7 torch.__version__: 1.0.0.dev20181130 also torch.__version__: 0.4.1 root@37cafe66fa00:/workspace/torchnlp# pytest ... /workspace/torchnlp/tests/tasks/test_sequence_tagging.py:19: in udpos_dataset fields=(('inputs_word', inputs), ('labels', tags), (None, None))) /opt/conda/lib/python3.6/site-packages/torchtext/datasets/sequence_tagging.py:66: in splits test=test, **kwargs) /opt/conda/lib/python3.6/site-packages/torchtext/data/dataset.py:78: in splits os.path.join(path, train), **kwargs) /opt/conda/lib/python3.6/site-packages/torchtext/datasets/sequence_tagging.py:29: in __init__ for line in input_file: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <encodings.ascii.IncrementalDecoder object at 0x7f9df029a780> input = b'\ntake\tVERB\tVB\nover\tADP\tRP\nAfghanistan\tPROPN\tNNP\n.\tPUNCT\t.\n\nResponse\tNOUN\tNN\n:\tPUNCT\t:\nThe\tDET\t...one\tPRON\tPRP\n\'s\tPART\tPOS\neyes\tNOUN\tNNS\noff\tADP\tIN\nthe\tDET\tDT\nformer\tADJ\tJJ\nball\tNOUN\tNN\n.\tPUNCT' final = False def decode(self, input, final=False): > return codecs.ascii_decode(input, self.errors)[0] E UnicodeDecodeError: 'ascii' codec can't decode byte 0xc3 in position 6271: ordinal not in range(128) /opt/conda/lib/python3.6/encodings/ascii.py:26: UnicodeDecodeError ... @Alexmac22347 could you please provide a docker on which this package works? @aleksas are you input files in UTF-8 format? I'm running pytest on cleanly cloned repo according to the installation instructions, On py2.7 pytest runs without a problem. But training raises exception as shown in initial post. On py3.6.5 pytest fails on tests/tasks/test_sequence_tagging.py tests. Took the CoNLL-2003 data from HERE as it's not clear how to get the dataset from provided link. Tried also nyt_ingredients_ner dataset for training with same exception. @aleksas Ive ran into the same unicode error. Might be a bit of a hack, running export LC_CTYPE=en_US.UTF-8 fixed it for me. https://stackoverflow.com/questions/18649512/unicodedecodeerror-ascii-codec-cant-decode-byte-0xe2-in-position-13-ordinal Let me know if that works/doesnt work @Alexmac22347 that did the trick. Made a PR with docker with appropriate locale settings. Thanks @aleksas! Merged your PR. Hope this solves your problem :) It did solve the problem, thanks!
gharchive/issue
2018-11-29T23:14:29
2025-04-01T06:44:42.862742
{ "authors": [ "Alexmac22347", "aleksas", "kolloldas" ], "repo": "kolloldas/torchnlp", "url": "https://github.com/kolloldas/torchnlp/issues/8", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }